query
stringlengths 7
9.55k
| document
stringlengths 10
363k
| metadata
dict | negatives
listlengths 0
101
| negative_scores
listlengths 0
101
| document_score
stringlengths 3
10
| document_rank
stringclasses 102
values |
|---|---|---|---|---|---|---|
Use callbacks to share common setup or constraints between actions.
|
def set_sales_forecast
@sales_forecast = SalesForecast.find(params[:id])
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def set_required_actions\n # TODO: check what fields change to asign required fields\n end",
"def action_hook; end",
"def run_actions; end",
"def define_action_hook; end",
"def actions; end",
"def define_action_helpers\n if super && action == :save\n @instance_helper_module.class_eval do\n define_method(:valid?) do |*args|\n self.class.state_machines.fire_event_attributes(self, :save, false) { super(*args) }\n end\n end\n end\n end",
"def add_actions; end",
"def callbacks; end",
"def callbacks; end",
"def setup *actions, &proc\n (@setup_procs ||= []) << [proc, actions.size > 0 ? actions : [:*]]\n end",
"def define_action_helpers; end",
"def post_setup\n end",
"def action_methods; end",
"def action_methods; end",
"def action_methods; end",
"def before_setup; end",
"def action_run\n end",
"def execute(setup)\n @action.call(setup)\n end",
"def define_action_helpers?; end",
"def set_actions\n actions :all\n end",
"def action_done(action)\n dispatch = { :migrate => :done_migrating, :map => :done_mapping, :reduce =>\n :done_reducing, :finalize => :done_finalizing } \n self.send dispatch[action[:action]], action\n end",
"def dependencies action, &block\n @actions.each do |other|\n if action[:requires].include? other[:provide]\n block.call other\n end\n end\n end",
"def setup!\n return unless @setup_procs\n http_actions = actions\n @setup_procs.each do |setup_proc|\n proc, actions = setup_proc\n @setup__actions = actions.map do |action|\n\n action.is_a?(Regexp) ?\n http_actions.select { |a| a.to_s =~ action } :\n action.is_a?(String) && action =~ /\\A\\./ ?\n http_actions.map { |a| a.to_s << action if format?(a).include?(action) }.compact :\n action\n\n end.flatten\n self.class_exec &proc\n @setup__actions = nil\n end\n @setup_procs = nil\n end",
"def before_actions(*logic)\n self.before_actions = logic\n end",
"def setup_handler\n end",
"def set_action(opts)\n opts = check_params(opts,[:actions])\n super(opts)\n end",
"def setup(action)\n @targets.clear\n unless action.item.target_filters.empty?\n @targets = SES::TargetManager.make_targets(action)\n else\n item = action.item\n if item.for_opponent?\n @targets = $game_troop.alive_members\n elsif item.for_dead_friend?\n @targets = $game_party.battle_members.select { |actor| actor.dead? }\n else\n $game_party.battle_members.select { |actor| actor.alive? }\n end\n end\n @item_max = @targets.size\n create_contents\n refresh\n show\n activate\n end",
"def action; end",
"def action; end",
"def action; end",
"def action; end",
"def action; end",
"def workflow\n end",
"def revisable_shared_setup(args, block)\n class << self\n attr_accessor :revisable_options\n end\n options = args.extract_options!\n self.revisable_options = Options.new(options, &block)\n \n self.send(:include, Common)\n self.send(:extend, Validations) unless self.revisable_options.no_validation_scoping?\n self.send(:include, WithoutScope::QuotedColumnConditions)\n end",
"def setup\n @action = SampleActionAndroid.new(os_name: 'android',\n app_name: APP_PATH)\n end",
"def before(action)\n invoke_callbacks *self.class.send(action).before\n end",
"def process_action(...)\n send_action(...)\n end",
"def before_dispatch(env); end",
"def after_actions(*logic)\n self.after_actions = logic\n end",
"def setup\n # override and do something appropriate\n end",
"def setup(client)\n return unless @setup\n actions = @setup['setup'].select { |action| action['do'] }.map { |action| Action.new(action['do']) }\n actions.each do |action|\n action.execute(client)\n end\n self\n end",
"def setup(_context)\n end",
"def setup(resources) ; end",
"def validate_actions\n errors.add(:base, :should_give_at_least_one_action) if !manage? && !forecasting? && !read? && !api?\n end",
"def setup\n @resource_config = {\n :callbacks => {\n :before_create => nil,\n :after_create => nil,\n :before_update => nil,\n :after_update => nil,\n :before_destroy => nil,\n :after_destroy => nil,\n },\n :child_assoc => nil,\n :model => nil,\n :parent => nil,\n :path => nil,\n :permission => {},\n :properties => {},\n :relation => {\n :create => nil,\n :delete => nil,\n },\n :roles => nil,\n }\n end",
"def determine_valid_action\n\n end",
"def process_shared\n handle_taxes\n handle_shippings\n create_adjustments_from_params\n handle_status\n handle_inventory_refunds\n handle_payment_transactions\n order.updater.update\n end",
"def startcompany(action)\n @done = true\n action.setup\n end",
"def init_actions\n am = action_manager()\n am.add_action(Action.new(\"&Disable selection\") { @selection_mode = :none; unbind_key(32); bind_key(32, :scroll_forward); } )\n am.add_action(Action.new(\"&Edit Toggle\") { @edit_toggle = !@edit_toggle; $status_message.value = \"Edit toggle is #{@edit_toggle}\" })\n end",
"def event_callbacks(event, metadata={})\n case event\n when :reset, :review\n if confirmed\n update_attributes(confirmed: false)\n end\n when :confirm\n confirm\n # trigger :order for all applicable items\n # NOTE: :order event is common to both physical and digital items\n items.each do |i|\n if i.event_permitted(:order)\n user_id = last_transition.user_id\n i.trigger!(:order, { order_id: id, user_id: user_id })\n end\n end\n when :complete_work\n request = metadata[:request]\n work_complete_notification(request)\n when :close\n close\n end\n if event != :close && !open\n reopen\n end\n end",
"def setup_action\n return unless PONY::ERRNO::check_sequence(current_act)\n new_sequence = @action_sequence[@sequence_index+1...@action_sequence.size]\n @sequence_index = 0\n new_sequence = DND::SkillSequence::ACTS[@acts[1]] + new_sequence\n execute_sequence\n end",
"def define_tasks\n define_weave_task\n connect_common_tasks\n end",
"def setup(&block)\n define_method(:setup, &block)\n end",
"def setup\n transition_to(:setup)\n end",
"def setup\n transition_to(:setup)\n end",
"def action\n end",
"def setup( *args )\n\t\t\tself.class.setupBlocks.each {|sblock|\n\t\t\t\tdebugMsg \"Calling setup block method #{sblock}\"\n\t\t\t\tself.send( sblock )\n\t\t\t}\n\t\t\tsuper( *args )\n\t\tend",
"def config(action, *args); end",
"def setup\n @setup_proc.call(self) if @setup_proc\n end",
"def before_action \n end",
"def setup_callbacks\n defined_callbacks.each do |meth|\n unless respond_to?(\"call_#{meth}_callbacks\".to_sym)\n self.class.module_eval <<-EOE\n def call_#{meth}_callbacks(*args)\n plugin_store.each {|a| a.call_#{meth}_callbacks(*args) } if respond_to?(:plugin_store) && plugin_store\n self.send :#{meth}, *args if respond_to?(:#{meth})\n end\n EOE\n end\n end\n end",
"def action\n end",
"def matt_custom_action_begin(label); end",
"def setup\n # override this if needed\n end",
"def setup\n\t\t\t\t\t\t# Do nothing\n\t\t\t\tend",
"def setup\n\t\t\t\t\t\t# Do nothing\n\t\t\t\tend",
"def action(options,&callback)\n new_action = Action===options ? options : Action.new(options,&callback)\n # replace any with (shared name/alias or both default) + same arity\n @actions.delete_if do |existing_action|\n ((existing_action.names & new_action.names).size > 0 ||\n existing_action.default? && new_action.default?) &&\n existing_action.required.size == new_action.required.size &&\n existing_action.optional.size <= new_action.optional.size\n end\n @actions = (@actions + [new_action]).sort\n new_action\n end",
"def set_target_and_action target, action\n self.target = target\n self.action = 'sugarcube_handle_action:'\n @sugarcube_action = action\n end",
"def after(action)\n invoke_callbacks *options_for(action).after\n end",
"def pre_task\n end",
"def setup(server)\n server.on('beforeMethod', method(:before_method), 10)\n end",
"def add_actions\n attribute = machine.attribute\n name = self.name\n \n owner_class.class_eval do\n define_method(name) {self.class.state_machines[attribute].events[name].fire(self)}\n define_method(\"#{name}!\") {self.class.state_machines[attribute].events[name].fire!(self)}\n define_method(\"can_#{name}?\") {self.class.state_machines[attribute].events[name].can_fire?(self)}\n end\n end",
"def init_actions\n @select_action = SelectAction.new\n @endpoint_mouse_action = EndpointMouseAction.new\n @move_action = MoveAction.new\n end",
"def setup_signals; end",
"def after_created\r\n return unless compile_time\r\n Array(action).each do |action|\r\n run_action(action)\r\n end\r\nend",
"def after_created\r\n return unless compile_time\r\n Array(action).each do |action|\r\n run_action(action)\r\n end\r\nend",
"def set_target_and_action target, action\n self.target = target\n self.action = 'sugarcube_handle_action:'\n @sugarcube_action = action.respond_to?('weak!') ? action.weak! : action\n end",
"def initialize(*args)\n super\n @action = :set\nend",
"def after_set_callback; end",
"def setup\n #implement in subclass;\n end",
"def lookup_action; end",
"def setup &block\n if block_given?\n @setup = block\n else\n @setup.call\n end\n end",
"def setup_action\n return TSBS.error(@acts[0], 1, @used_sequence) if @acts.size < 2\n actions = TSBS::AnimLoop[@acts[1]]\n if actions.nil?\n show_action_error(@acts[1])\n end\n @sequence_stack.push(@acts[1])\n @used_sequence = @acts[1]\n actions.each do |acts|\n @acts = acts\n execute_sequence\n break if @break_action\n end\n @sequence_stack.pop\n @used_sequence = @sequence_stack[-1]\n end",
"def release_actions; end",
"def around_hooks; end",
"def save_action; end",
"def setup(easy)\n super\n easy.customrequest = @verb\n end",
"def action_target()\n \n end",
"def setup\n callback(:setup) do\n notify(:setup)\n migration_check.last_deployed_commit\n end\n end",
"def setup\n return unless @setup\n\n actions = @setup['setup'].select { |action| action['do'] }.map { |action| Action.new(action['do']) }\n run_actions_and_retry(actions)\n self\n end",
"def before_setup\n # do nothing by default\n end",
"def my_actions(options)\n @setup = false\n get_template_part(\"custom_used\",\"action_users\",true)\n end",
"def default_action; end",
"def setup(&blk)\n @setup_block = blk\n end",
"def callback_phase\n super\n end",
"def advice\n end",
"def _handle_action_missing(*args); end",
"def duas1(action)\n action.call\n action.call\nend",
"def shared_action(name, &block)\n @controller.shared_actions[name] = block\n end",
"def before_action action, &block\n @audience[:before][action] ||= Set.new\n @audience[:before][action] << block\n end",
"def setup_initial_state\n\n state_a = State.new(\"a\", 0)\n state_b = State.new(\"b\", 0)\n state_c = State.new(\"c\", 10)\n\n move_to_b = Action.new(\"move_to_b\", 1, state_b)\n\n move_to_c = Action.new(\"move_to_c\", 1, state_c)\n\n state_a.actions = [move_to_b, move_to_c]\n\n return state_a\n \nend"
] |
[
"0.6163163",
"0.6045976",
"0.5946146",
"0.591683",
"0.5890051",
"0.58349305",
"0.5776858",
"0.5703237",
"0.5703237",
"0.5652805",
"0.5621621",
"0.54210985",
"0.5411113",
"0.5411113",
"0.5411113",
"0.5391541",
"0.53794575",
"0.5357573",
"0.53402257",
"0.53394014",
"0.53321576",
"0.53124547",
"0.529654",
"0.5296262",
"0.52952296",
"0.52600986",
"0.52442724",
"0.52385926",
"0.52385926",
"0.52385926",
"0.52385926",
"0.52385926",
"0.5232394",
"0.523231",
"0.5227454",
"0.52226824",
"0.52201617",
"0.5212327",
"0.52079266",
"0.52050185",
"0.51754695",
"0.51726824",
"0.51710224",
"0.5166172",
"0.5159343",
"0.51578903",
"0.51522785",
"0.5152022",
"0.51518047",
"0.51456624",
"0.51398855",
"0.5133759",
"0.5112076",
"0.5111866",
"0.5111866",
"0.5110294",
"0.5106169",
"0.509231",
"0.50873137",
"0.5081088",
"0.508059",
"0.50677156",
"0.50562143",
"0.5050554",
"0.50474834",
"0.50474834",
"0.5036181",
"0.5026331",
"0.5022976",
"0.5015441",
"0.50121695",
"0.5000944",
"0.5000019",
"0.4996878",
"0.4989888",
"0.4989888",
"0.49864885",
"0.49797225",
"0.49785787",
"0.4976161",
"0.49683493",
"0.4965126",
"0.4958034",
"0.49559742",
"0.4954353",
"0.49535993",
"0.4952725",
"0.49467874",
"0.49423352",
"0.49325448",
"0.49282882",
"0.49269363",
"0.49269104",
"0.49252945",
"0.4923091",
"0.49194667",
"0.49174926",
"0.49173003",
"0.49171105",
"0.4915879",
"0.49155936"
] |
0.0
|
-1
|
Never trust parameters from the scary internet, only allow the white list through.
|
def sales_forecast_params
params.require(:sales_forecast).permit(:service_center_id, :segment_id, :status_id, :area_id, :company, :division, :customer_name, :instrument, :price, :budget_id, :certainty_id, :reporting_date, :fixed_date, :order_date, :agent, :condition_text, :person_in_charge, :visible)
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def strong_params\n params.require(:user).permit(param_whitelist)\n end",
"def strong_params\n params.require(:listing_member).permit(param_whitelist)\n end",
"def allow_params_authentication!; end",
"def allowed_params\n ALLOWED_PARAMS\n end",
"def default_param_whitelist\n [\"mode\"]\n end",
"def param_whitelist\n [:role, :title]\n end",
"def expected_permitted_parameter_names; end",
"def safe_params\n params.except(:host, :port, :protocol).permit!\n end",
"def strong_params\n params.require(:team_member).permit(param_whitelist)\n end",
"def permitir_parametros\n \t\tparams.permit!\n \tend",
"def strong_params\n params.require(:community).permit(param_whitelist)\n end",
"def permitted_strong_parameters\n :all #or an array of parameters, example: [:name, :email]\n end",
"def strong_params\n params.require(:education).permit(param_whitelist)\n end",
"def restricted_params\n #params.require(self.controller_name.classify.underscore.to_sym).permit([])\n raise(\"No strong params set, override restricted_params method in your controller. E.g. params.require(:model).permit(:attribute1, :attribute2)\")\n end",
"def allowed_params\n params.require(:user).permit(:username, :email, :password, :password_confirmation)\n end",
"def param_whitelist\n [:rating, :review]\n end",
"def param_whitelist\n whitelist = [\n :username, :name,\n :parent_id,\n :headline, :description, :video,\n :policy, :signup_mode, :category,\n :website, :facebook, :twitter, :linkedin,\n :founded_at,\n privacy: [\n :events,\n :resources\n ],\n permission: [\n :profile,\n :members,\n :children,\n :statistics,\n :posts,\n :listings,\n :resources,\n :events\n ],\n location: [\n :description,\n :street,\n :city,\n :state,\n :zip,\n :country,\n :latitude,\n :longitude\n ]\n ]\n \n if action_name === 'update'\n whitelist.delete(:parent_id)\n unless current_user.role_in(@community) === 'owner'\n whitelist.delete(:privacy)\n whitelist.delete(:permission)\n end\n end\n \n whitelist\n end",
"def param_whitelist\n if @user.present? && current_user != @user\n return [:followed]\n end\n \n whitelist = [\n :username, :email, :password,\n :first_name, :last_name,\n :birthday, :gender,\n :headline, :biography, :ask_about, :focus,\n :website, :facebook, :linkedin, :twitter, :github,\n roles: [],\n skills: [],\n interests: [],\n privacy: { contact: [] },\n location: [\n :description,\n :street,\n :city,\n :state,\n :zip,\n :country,\n :latitude,\n :longitude\n ]\n ]\n \n if action_name === 'update'\n whitelist.delete(:email)\n whitelist.delete(:password)\n end\n \n whitelist\n end",
"def user_params \n \tparams.require(:user).permit(:name, :email, :password, :password_confirmation)# preventing CSTR\n end",
"def user_params\n params.permit(:name, :phoneNumber, :address, :postalCode, :local, :link, :counter, :latitude, :longitude) \n end",
"def valid_params_request?; end",
"def strong_params\n params.require(:experience).permit(param_whitelist)\n end",
"def trim_whitelisted(params, whitelist)\n # remove any parameters that are not whitelisted\n params.each do |key, value|\n # if white listed\n if whitelist.include? key\n # strip the parameters of any extra spaces, save as string\n params[key] = value.to_s.strip\n else\n # delete any unauthorized parameters\n params.delete key\n end\n end\n params\n end",
"def whitelist_url_params\n params.require(:whitelist_url).permit(:domain)\n end",
"def allowed_params\n params.require(:allowed).permit(:email)\n end",
"def permitted_params\n []\n end",
"def trim_whitelisted(params, whitelist)\n # remove any parameters that are not whitelisted\n params.each do |key, value|\n # if white listed\n if whitelist.include? key\n # strip the parameters of any extra spaces, save as string\n params[key] = value.to_s.strip\n else\n # delete any unauthorized parameters\n params.delete key\n end\n end\n params\n end",
"def safe_params\n params.permit(:id, :name, :origin, :emails => []); #emails is an array\n end",
"def query_param\n\t\tparams.permit(:first_name, :last_name, :phone)\n\tend",
"def strong_params\n params.require(:success_metric).permit(param_whitelist)\n end",
"def devise_filter\r\n logger.debug(\"In devise_filter =>PARAMS: #{params.inspect}\")\r\n\r\n # White list for sign_up\r\n devise_parameter_sanitizer.for(:sign_up) { |u| u.permit(user_whitelist) }\r\n\r\n # White list for account update\r\n devise_parameter_sanitizer.for(:account_update) { |u| u.permit(user_whitelist, :current_password) }\r\n\r\n # White list for Invitation creation\r\n devise_parameter_sanitizer.for(:invite) { |u| u.permit(:account_type, :email, :invitation_token)}\r\n\r\n # White list for accept invitation\r\n devise_parameter_sanitizer.for(:accept_invitation) { |u| u.permit(user_whitelist, :invitation_token)}\r\n\r\n end",
"def whitelisted_user_params\n params.require(:user).\n permit( :first_name, :last_name, :email,:password,:password_confirmation,:birthday,:gender)\n end",
"def user_params\n ActionController::Parameters.permit_all_parameters = true\n params.require(:user) #.permit(:name, :surname, :phone, :password, :email, :time_zone)\n end",
"def strong_params\n params.require(:metric_change).permit(param_whitelist)\n end",
"def safe_params\n params.require(:user).permit(:name)\n end",
"def get_params\n\t\treturn ActionController::Parameters.new(self.attributes).permit(\"account_id\", \"title\", \"category\", \"introduction\", \"tags\", \"segment_type\", \"visible\", \"status\", \"main_image\")\n\tend",
"def grant_params\n @whitelisted = params.require(:grant).permit(:name, :description, :agency_id, :acronym)\n end",
"def check_params; true; end",
"def param_whitelist\n whitelist = [\n :description,\n :progress,\n :kpi_id\n ]\n \n unless action_name === 'create'\n whitelist.delete(:kpi_id)\n end\n \n whitelist\n end",
"def quote_params\n params.permit!\n end",
"def valid_params?; end",
"def paramunold_params\n params.require(:paramunold).permit!\n end",
"def user_params\n\t\tparams.permit(:nickname, :avatar, :description, :password, :gender, :birthday, :email, :phone, :qq_id, :wechat_id)\n\tend",
"def filtered_parameters; end",
"def user_params\n params.permit(\n \t:id,\n \t:email, \n \t:first_name, \n \t:last_name, \n \t:password, \n \t:confirm_token, \n \t:phone_number,\n \t:facebook_link,\n \t:car_model,\n \t:license_plate)\n end",
"def filtering_params\n params.permit(:email, :name)\n end",
"def check_params\n true\n end",
"def wx_public_params\n params.require(:wx_public).permit(:nickname, :manager, :alias)\n end",
"def allowed_params\n params.require(:user).permit(:email, :password, :role, :first_name, :last_name, :password_confirmation)\n end",
"def allowed_params\n params.require(:user).permit(:email, :password, :role, :first_name, :last_name, :password_confirmation)\n end",
"def listing_params\n\t\tparams.permit(:address, :transit_info, :rules, :other_info, :lat, :lng)\n\tend",
"def social_account_params\n\t\t\tparams.require(:social_account).permit!\n\t\tend",
"def safe_params\n resurce_name = self.class.resource_name\n params_method_name = \"#{resurce_name}_params\".to_sym\n if params[resurce_name]\n if respond_to?(params_method_name) || private_methods.include?(params_method_name)\n send(params_method_name)\n else\n raise ActiveModel::ForbiddenAttributesError, \"Please, define the '#{params_method_name}' method in #{self.class.name}\"\n end\n end\n end",
"def url_params\n params.require(:url).permit(:short_url, :original_url, :clicks, :ip_addresses)\n end",
"def user_params\n params.require(:user).permit(:uri, :username, :password, :realname, :email, :publicvisible)\n end",
"def model_params\n\t\tparams.require(:manager).permit(\n\t :user_name,\n :password,\n :email,\n \t\t\t)\n\tend",
"def article_params_whitelist\n params.require(:article).permit(:title, :description, category_ids: [])\n end",
"def college_whitelist_params\n params.require(:college_whitelist).permit(:status)\n end",
"def active_code_params\n params[:active_code].permit\n end",
"def filtering_params\n params.permit(:email)\n end",
"def valid_params(params)\n params.permit(:user_id, :photo_id, :originX, :originY, :width, :height)\n end",
"def ip_address_params\n\t\t\tparams.require(:ip_address).permit!\n end",
"def pull_request_params\n whitelist = [\n :url,\n :id,\n :html_url,\n :diff_url,\n :patch_url,\n :issue_url,\n :number,\n :state,\n :locked,\n :title\n ]\n params.require(:pull_request).permit(whitelist)\n end",
"def reserved_params\n params.require(:reserved).permit(:name, :email, :pax, :address, :KTP, :title)\n end",
"def post_params\n if current_user.admin? \n params.permit(:title, :body, :city, :country, :gps_location, :privacy, :visible, :latitude, :longitude, images: [], files: [])\n else \n params.permit(:title, :body, :city, :country, :gps_location, :privacy,:latitude, :longitude, images: [], files: [])\n end \n end",
"def list_params\n params.permit(:name)\n end",
"def filter_parameters; end",
"def filter_parameters; end",
"def vineyard_params\n params.permit(:vineyard_name, :email, :website_url, :phone, :address, :city, :region, :postcode, :country, :specialty, :description, :pet_friendly, :holiday, :tours, :events, :family_friendly, :cover_image, :image_one, :image_two, :image_three, :image_four, :user_id, :base64)\n end",
"def available_activity_params\n # params.require(:available_activity).permit(:type,:geometry,:properties)\n whitelisted = ActionController::Parameters.new({\n type: params.require(:available_activity)[:type],\n geometry: params.require(:available_activity)[:geometry].try(:permit!).to_h,\n properties: params.require(:available_activity)[:properties].try(:permit!).to_h\n }).try(:permit!)\n end",
"def user_params\n params.permit(:name, :username, :email, :password, :img_url, :bg_url, :coinbank)\n end",
"def user_params_pub\n\t \tparams[:user].permit(:hruid)\n\t end",
"def user_params\n params.permit(:id, :email, :password, :nickname, :status, :avatar, :flat_picture, :flatsharing_id, :member,\n :user, :color, :solde)\n end",
"def validate_search_inputs\n @whitelisted = params.fetch(:user, nil)\n if @whitelisted.blank?\n render_error(400, \"#{I18n.t('general_error.params_missing_key')}\": [I18n.t('general_error.params_missing_value', model: \"review\")])\n return\n else\n @whitelisted = @whitelisted.permit(:name, :uen, :description)\n end\n end",
"def param_whitelist\n [\n :title,\n :description,\n :organization,\n :team_id,\n :started_at,\n :finished_at,\n location: [\n :description,\n :street,\n :city,\n :state,\n :zip,\n :country,\n :latitude,\n :longitude\n ]\n ]\n end",
"def url_whitelist; end",
"def admin_social_network_params\n params.require(:social_network).permit!\n end",
"def filter_params\n params.require(:filters).permit(:letters)\n end",
"def origin_params\n params.permit(:country, :state, :city, :postal_code, :address, :description)\n end",
"def valid_params(params)\n params.permit(:login, :first_name, :last_name, \n :password, :password_confirmation)\n end",
"def sensitive_params=(params)\n @sensitive_params = params\n end",
"def permit_request_params\n params.permit(:address)\n end",
"def user_params\n # Ensure a user can't give themselves admin priveleges\n params.delete(:admin) if current_user.admin?\n params.require(:user).permit(:name, :email, :admin, :image)\n end",
"def secure_params\n params.require(:location).permit(:name)\n end",
"def strong_params\n params.require( :setting ).\n permit( :global_scan_limit, :per_user_scan_limit,\n :target_whitelist_patterns, :target_blacklist_patterns )\n end",
"def question_params\n params.require(:survey_question).permit(question_whitelist)\n end",
"def case_insensitive_params\n params.require(:case_insensitive).permit(:name)\n end",
"def empire_master_no_match_params\n params.require(:empire_master_no_match).permit(:uid, :last_name, :list, :search_date, :double, :source)\n end",
"def maintenance_request_params\n params[:maintenance_request].permit! #allow all parameters for now\n end",
"def unwanted_params\n params.require(:unwanted).permit(:title, :description, :image)\n end",
"def url_params\n params[:url].permit(:full)\n end",
"def backend_user_params\n params.permit!\n end",
"def filter_params\n\t\treturn params[:candidate].permit(:name_for_filter)\n\tend",
"def speed_measurement_params\n\n #fuckit, to lazy to deal with permit crap right now\n ActionController::Parameters.permit_all_parameters = true\n\n params[:speed_measurement]\n end",
"def user_params\n params.permit(:name, :age, :username, :display_photo, :password)\n end",
"def get_params\r\n #params.require(:article).permit(:title, :permalink, :content, :source_site, :introtext, :type_id, :order_by, :searchable, :created_by, :edited_by, :published_by, :published_on, :user_id)\r\n params.require(:article).permit!\r\n\r\n end",
"def pub_params\n params.require(:pub).permit(:name, :description, :phone, :email, :hidden, :city_id, :address)\n end",
"def pass_params\n params[:pass].permit(:name, :price, :description, :colour, :events)\n end",
"def droptraining_params\n params.permit(:training_id,:user_id, :utf8, :authenticity_token, :commit)\n end",
"def person_params\n # params whitelist does *not* include admin, sub, remember_token\n # TBD: share this whitelist with the list used by configuration_permitted_parameters\n # TBD: should current_password be on this list? -- for now, leaving off, since it seems to work without\n # NOTE: do not include 'admin' in this list!\n params.require(:person).permit(\n :name, \n :email, \n :description,\n :password, \n :password_confirmation\n )\n end",
"def parameter_params\n params.require(:parameter).permit(:name, :description, :param_code, :param_value, :active_from, :active_to)\n end"
] |
[
"0.69792545",
"0.6781151",
"0.67419964",
"0.674013",
"0.6734356",
"0.6591046",
"0.6502396",
"0.6496313",
"0.6480641",
"0.6477825",
"0.64565",
"0.6438387",
"0.63791263",
"0.63740575",
"0.6364131",
"0.63192815",
"0.62991166",
"0.62978333",
"0.6292148",
"0.6290449",
"0.6290076",
"0.62894756",
"0.6283177",
"0.6242471",
"0.62382483",
"0.6217549",
"0.6214457",
"0.6209053",
"0.6193042",
"0.6177802",
"0.6174604",
"0.61714715",
"0.6161512",
"0.6151757",
"0.6150663",
"0.61461",
"0.61213595",
"0.611406",
"0.6106206",
"0.6105114",
"0.6089039",
"0.6081015",
"0.6071004",
"0.60620916",
"0.6019971",
"0.601788",
"0.6011056",
"0.6010898",
"0.6005122",
"0.6005122",
"0.6001556",
"0.6001049",
"0.59943926",
"0.5992201",
"0.59909594",
"0.5990628",
"0.5980841",
"0.59669393",
"0.59589154",
"0.5958826",
"0.5957911",
"0.5957385",
"0.5953072",
"0.59526145",
"0.5943361",
"0.59386164",
"0.59375334",
"0.59375334",
"0.5933856",
"0.59292704",
"0.59254247",
"0.5924164",
"0.59167904",
"0.59088355",
"0.5907542",
"0.59064597",
"0.5906243",
"0.5898226",
"0.589687",
"0.5896091",
"0.5894501",
"0.5894289",
"0.5891739",
"0.58860534",
"0.5882406",
"0.587974",
"0.58738774",
"0.5869024",
"0.58679986",
"0.5867561",
"0.5865932",
"0.5864461",
"0.58639693",
"0.58617616",
"0.5861436",
"0.5860451",
"0.58602303",
"0.5854586",
"0.58537364",
"0.5850427",
"0.5850199"
] |
0.0
|
-1
|
For example if the input is 4, then your program should return 10 because 1+2+3+4 = 10.
|
def SimpleAdding(num)
sum = 0
(1..num).each do |number|
sum += number
end
sum
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def SimpleAdding(num\n total = 1.upto(num).reduce(&:+)\n total \nend",
"def sum(input)\n arr = input.to_s.chars.map { |x| x.to_i }\n arr.reduce { |sum, num| sum + num }\n # number.to_s.chars.map(&:to_i).reduce(:+) # read up on symbols and procs\nend",
"def addition(number)\n (1..number).reduce(:+)\nend",
"def summation(num)\r\n puts 1.upto(num).reduce(0, :+)\r\nend",
"def compute_sum(number)\n (1..number).reduce(:+)\nend",
"def sum(number)\n n = number\n array = []\n answer = []\n while n > 1 do\n array << (n -= 1)\n end\n num = array.reduce(:*)\n num.to_s.split(\"\").each do |i|\n answer << i.to_i\n end\n return answer.reduce(:+)\nend",
"def compute_sum(number)\n (1..number).inject(:+)\nend",
"def sum_difference_product\n puts \"Please enter 2 integers:\"\n input = gets.chomp\n nums = input.split(\" \").map(&:to_i)\n sum = nums.inject(:+)\n difference = nums.inject(:-)\n product = nums.inject(:*)\n puts \"#{sum}\\n#{difference}\\n#{product}\\n\"\nend",
"def SimpleAdding(num)\n answer = (1..num).to_a.inject(:+)\nend",
"def simple_adding(num)\n (1..num).inject(:+)\nend",
"def sum_difference_product\n input = gets.split\n numbers = input.map {|num| num.to_i}\n sum = numbers.inject(:+)\n difference = numbers.inject(:-)\n product = numbers.inject(:*)\n\n puts sum\n puts difference\n puts product\n \nend",
"def sum3(integer)\n if integer < 10\n integer % 10\n else\n integer % 10 + sum(integer / 10)\n end\nend",
"def sum(num)\n num.to_s.split('').reduce do |sum, num|\n sum.to_i + num.to_i\n end\nend",
"def GetSum(num, sum)\n if num == 0\n return sum\n else\n return GetSum((num - 1), (sum + num))\n end\nend",
"def sum_of_n_numbers()\n\tsum=0\n\ta=gets.chomp.to_i\n#\tputs a\n\tfor i in 1 .. a\n\t\tsum=sum+gets.chomp.to_i\n\tend\n\tputs sum\nend",
"def sum_difference_product\n\tinput = gets.chomp\n\ti = input.split.collect{|x| x.to_i}\n\tputs (i[0] + i[1])\n\tputs (i[0] - i[1])\n\tputs (i[0] * i[1])\nend",
"def sum_difference_product(input)\n puts \"give me an integer\"\n input= gets.chomp\narray_of_input = input.split(\"\").map {|numb| numb.to_i }\nsum = array_of_input[0]+ array_of_input[2]\nputs \"#{sum}\"\nputs \"#{array_of_input[0] - array_of_input[2]}\"\nputs array_of_input[0] * array_of_input[2]\n\nend",
"def summation(num)\n (1..num).reduce(:+)\nend",
"def sum_of_amicable_numbers(n)\r\n amicable_numbers(n).reduce(:+)\r\nend",
"def sum(num)\n num.to_s.chars.map(&:to_i).reduce(:+)\nend",
"def sum(num)\n num.to_s.chars.map(&:to_i).reduce(:+)\nend",
"def SimpleAdding(num)\n (1..num).inject(:+)\nend",
"def SimpleAdding(num)\n (1..num).inject(:+)\nend",
"def sum(n)\n end",
"def sum(num)\n digits = num.to_s.chars.map(&:to_i) # Get an array of ints\n digits.reduce(:+) # Return sum of digits\nend",
"def sum(number)\n number.to_s.chars.map(&:to_i).reduce(:+)\n \nend",
"def sum(int)\n int.to_s.split(\"\").map(&:to_i).reduce(&:+)\nend",
"def AdditivePersistence(num)\n\n new = num.to_s.split('').map {|x| x.to_i}\n \n pers = new.size == 1 ? 0 : 1\n sum = new.reduce(:+)\n new = sum.to_s.split('').map { |x| x.to_i }\n \n until sum < 10 \n p sum = new.reduce(:+)\n p pers += 1\n new = sum.to_s.split('').map { |x| x.to_i }\n end\n \n # code goes here\n return pers \n \nend",
"def get_next(n_str)\n sum = n_str.to_s.split('').map(&:to_i).reduce(:+)\n sum + n_str\nend",
"def NumberAddition(str)\n str.scan(/\\d+/).map {|i| i.to_i}.length == 0 ? 0 : str.scan(/\\d+/).map {|i|i.to_i}.inject(:+)\nend",
"def solution(n)\n n.to_s.split(//).inject(1) { |a,d| a + d.to_i }\nend",
"def numSum number\nsum = 0\ncount = 0\nwhile (count < number.to_i)\nputs \"enter num:\"\nnum = gets.chomp\nsum += num.to_i\ncount += 1\nend\nreturn sum\nend",
"def sum(numbers)\n string = numbers.digits.reverse\n string.reduce(:+)\nend",
"def SimpleAdding(num)\n\n # code goes here\n range_sum = *(1..num)\n return range_sum.inject(:+)\n \nend",
"def sum_recursive(num)\n # can also compute sum with symbol (1..5).inject(:+)\n (1..num).inject { |sum, n| sum + n }\nend",
"def sum_multiples_3_and_5\n return 3 * 333 * 334 / 2 + 5 * 199 * 200 / 2 - 15 * 66 * 67 / 2\nend",
"def sum(number)\n number.to_s.chars.map(&:to_i).reduce(:+)\nend",
"def sum(number)\n number.to_s.chars.map(&:to_i).reduce(:+)\nend",
"def sumdig_r(num, result = 0)\n if num < 10\n result += num\n else\n result += num % 10\n result = sumdig_r(num / 10, result)\n end\n result\nend",
"def SimpleAdding(num)\nsum=0\n 1.upto(num) do |x|\n\tsum+=x\n\tend\n\t\nreturn sum\nend",
"def sum(numbers)\r\n numbers.reduce(0, :+)\r\nend",
"def summation(num)\n sum = (0..num).inject(:+)\n return sum\nend",
"def add_up(num)\n return (1..num).inject(:+)\nend",
"def total(nums)\n nums.inject(:+)\nend",
"def consec_ints\n puts \">> Please enter an integer greater than 0:\"\n num = gets.chomp.to_i\n arr = Array.new(num) { |i| i + 1 }\n puts \">> Enter 's' to compute the sum, 'p' to compute the product.\"\n method = gets.chomp.downcase\n if method == 's'\n puts \"#{arr.reduce(&:+)}\"\n else\n puts \"#{arr.reduce(&:*)}\"\n end\nend",
"def sum(x)\n solution = 0\n x.each do |num|\n solution += num\n end\n solution\nend",
"def sum_nums(num)\n (1..num).inject(&:+)\nend",
"def SimpleAdding(num)\n\n # code goes here\n range_num = *(1..num)\n return range_num.inject(:+)\n \nend",
"def sum_or_product\n puts \">> Please enter an integer greater than 0:\"\n integer = gets.chomp.to_i\n\n puts \">> Enter 's' to compute the sum, 'p' to compute the product.\"\n compute = gets.chomp\n\n if compute == 's'\n \"The sum of the integers between 1 and #{integer} is #{(1..integer).reduce(:+)}\"\n else\n \"The product of the integers between 1 and #{integer} is #{(1..integer).reduce(:*)}\"\n end\nend",
"def sum_or_product()\n n = 0\n procedure = ''\n loop do\n puts \"Please enter a number greater than 0\"\n response = gets.chomp\n n = response.to_i if response.to_i.to_s == response && response.to_i > 0\n break if n > 0\n end\n loop do\n puts \"Please enter 's' if you would like to calculate the sum, or 'p' if you would like to calculate the product\"\n response = gets.chomp\n procedure = response if %w(s p).include?(response)\n break unless procedure.empty?\n end\n\n range = (1..n).to_a\n if procedure == 's'\n range.reduce(:+)\n else\n range.reduce(:*)\n end\nend",
"def digital_root(n)\n # Return number if it can't be reduced further\n return n if n.size == 1\n # Evaluate the sum of digits use recursion on the sum\n digital_root(n.to_s.chars.inject {|sum, n| sum.to_i + n.to_i })\nend",
"def SimpleAdding(num)\n\ti = 0\n\ttotal = 0\n\twhile i <= num\n\t\ttotal += i\n\t\ti += 1\n\tend\n\tprint total\nend",
"def sum(int)\n int.to_s.chars.map(&:to_i).reduce(:+)\nend",
"def sum(int)\n digits = int.to_s.split('')\n digits.map!(&:to_i)\n digits.reduce(:+)\nend",
"def sum(num)\n num.to_s.split('').inject(0) do |total, num|\n total + num.to_i\n end\nend",
"def sum(numbers)\n numbers.inject(0, :+)\nend",
"def simple_adding (num)\n the_sum = 0\n i = 0\n while i <= num do\n puts the_sum += i\n i += 1\n end\n return the_sum\nend",
"def sum(n)\n return 1 if n == 1\n\n n + sum(n - 1)\nend",
"def add_numbers(input_number)\n $number += input_number\n return $number\nend",
"def get_number(result)\n # This is not mathematical correct, because we cant subtract\n puts \"Well done, #{result} == 24\" if 1 / 3 + 4 * 6 == result\n # But, if we can, here is solution: 6.0 / (1.0 - 3.0 / 4.0)\nend",
"def sum(number)\n a = (1..number).to_a\n puts a.inject(:+)\nend",
"def calculate(num)\n return 0 if num.zero?\n num + calculate(num - 1)\nend",
"def sum_of_num(num)\n arry = num.to_s.split('').map {|num| num.to_i}\n arry.reduce {|sum, num| sum + num}\nend",
"def sum_nums(num)\n\t\n\tnumbers = []\n\tnum_sum = num\n\t\nuntil numbers.length == num_sum\n\tnumbers.push(num)\n\tnum = num - 1\nend\n return\tnumbers.inject(:+)\nend",
"def sum(number)\n numbers = number.to_s.chars\n\n numbers.reduce {|sum, i| sum.to_i + i.to_i }\nend",
"def sum(number)\n sum = number.to_s.chars\n total = sum.map {|num| num.to_i }\n total.reduce(:+)\nend",
"def sum(integer)\n digits = integer.to_s.chars.map {|x|x.to_i}\n total = digits.reduce(0) {|sum,x| sum += x}\n #integer.to_s.chars.map(&:to_i).reduce(&:+)\nend",
"def addition\nprint \"What is your first number? \"\nfirst_num = gets().to_i\n\nprint \"What is the second number? \"\nsecond_num = gets().to_i\n\nresult = first_num + second_num\nputs \"#{first_num} + #{second_num} = #{result}\"\nend",
"def sum(integer)\n sum = integer.to_s.chars.map do |num|\n num.to_i\n end\n\n result = sum.inject(:+)\n result\nend",
"def sum_terms(n)\n # your code here\n (1..n).inject(&:+)\nend",
"def addition\n print \"What is your first number? \"\n first_num = gets().to_i\n\n print \"What is your second number? \"\n second_num = gets().to_i\n\n result = first_num + second_num\n puts \"#{first_num} + #{second_num} = #{result}\"\nend",
"def sum(n, m)\n if n > m\n return 0\n else\n return sum(n+1, m) + n\n end\nend",
"def sum_difference_product\n int = gets.chomp\n int_arr = int.split\n int1 = int_arr[0].to_i\n int2 = int_arr[1].to_i\n int_sum = int1 + int2\n int_dif = int1 - int2\n int_prod = int1 * int2\n print \"#{int_sum}\\n#{int_dif}\\n#{int_prod}\\n\"\nend",
"def compute_sum(int)\n sum = 0\n 1.upto(int) { |i| sum += i }\n sum\nend",
"def sum(num)\na_num = []\nnum_split_string = num.to_s.split('')\n \n while a_num.count < num do\n a_num.push(num_split_string)\n a_num.count\n break\n end\n a_num.flatten!.map! do |e|\n e.to_i\n end\n a_num.sum\nend",
"def sum(num)\n p num.to_s.chars.tally.keys.map(&:to_i).reduce(:+)\nend",
"def sum(n)\n result = 0\n n.each do |number|\n result = result + number\n end\n return result\n end",
"def sum_nums(num)\n\n value = 0\n i = 0\n while i < num\n value = value + (num - i) \n i += 1\n end\n\n return value\nend",
"def sum(input_integer)\n digits = input_integer.to_s.chars\n sum_array = digits.map { |digit| digit.to_i }.reduce(:+)\nend",
"def sum(nums)\n nums.reduce(&:+)\nend",
"def adding(num)\n\tsum = 0\n\t(1..num).each do |x|\n\t\tsum += x\n\tend\n\treturn sum\nend",
"def add(n, p)\n return n + p\nend",
"def compute_sum(number)\n total = 0\n 1.upto(number) { |value| total += value }\n total\nend",
"def rec_sum(num)\n return num if num <= 1\n num + rec_sum(num - 1)\nend",
"def number_addition(str)\n nums = str.gsub(/\\D/, \" \")\n num_ary = nums.split(\"\")\n clean_ary = remove_extra_spaces(num_ary)\n result_ary = clean_ary.join(\"\").split(\" \").map { |e| e.to_i }\n ans = result_ary.reduce(:+)\n ans.nil? ? 0 : ans\nend",
"def add_function \n\tputs \"Enter the numbers\"\n\tn1 = gets.to_i\n\tn2 = gets.to_i\n\tputs \"The sum is #{n1+n2}\"\nend",
"def add\n# nums()\n puts \"\"\n puts \"What is your first number?\"\n first_num = gets.chomp.to_f\n\n puts \"What is your second number?\"\n second_num = gets.chomp.to_f\n\n result = first_num + second_num\n puts \"#{first_num} + #{second_num} = #{result}\"\nend",
"def summation(num)\n num * (num + 1) / 2\nend",
"def always_threes\n puts \"Give me a number\"\n number = gets.to_i\n puts \"Always \" + (((number + 5) * 2 - 4) / 2 - number).to_s + \"!\"\nend",
"def get_sum(a, b)\n [a,b].reduce(&:+)\nend",
"def sum(i)\n sum = i.digits.reduce(:+)\n sum\nend",
"def sum_func(num)\n return num if num <= 1\n\n divided = num / 10\n moded = num % 10\n\n moded + sum_func(divided)\nend",
"def sum(integer)\n integer.digits.reduce(:+)\nend",
"def sum_nums2(num)\n\ncount = 0\nresult = 0\n\n while num >= count\n result += count\n count += 1\n end\n #puts( result )\n return result\nend",
"def sumdig_r(n)\n\n # puts \"#{n} and n /10 is #{n/10} and n%10 is #{n%10}\"\n\n if (n<10) \n return n\n else\n return n%10 + sumdig_r(n/10)\n end\nend",
"def numbers(num1, num2)\n puts \"#{num1} + #{num2}\"\n return num1 + num2\nend",
"def add\n print \"What is the first number? \"\n first_number = gets.to_f\n\n print \"What is the second number? \"\n second_number = gets.to_f\n\n result = first_number + second_number\n\n puts \"The results is #{result}.\"\nend",
"def sum(int)\n int.to_s.split(//).reduce(0){|sum, item| sum += item.to_i}\nend",
"def sum(input, n)\n array = Array.new(input.split(\"\").length, 0)\n input.split(\" \").each_with_index{ |x, i| array[i] = x.to_i }\n sum = array.inject(0, :+).to_s.split(\"\")\n print sum.join(\"\").to_i\n return sum[0..n].join(\"\").to_i\nend",
"def add\n puts \"\"\n puts \"what is you forst number\"\n first_num = gets.chomp.to_f\n puts \"what is you forst number\"\n second_num = gets.chomp.to_f\n\n puts \"#{first_num} + #{second_num} = #{first_num+second_num}\"\nend",
"def sum_difference_product\n a , b, c = gets.split.map { |num| num.to_i }\n puts a + b + c\n puts a - b - c\n puts a * b * c\nend"
] |
[
"0.7081776",
"0.7055382",
"0.70188284",
"0.6957403",
"0.6889566",
"0.6862021",
"0.68231344",
"0.6817671",
"0.6789057",
"0.678087",
"0.67713755",
"0.6766516",
"0.67608327",
"0.6744585",
"0.67346686",
"0.6716996",
"0.67154837",
"0.67126733",
"0.6705608",
"0.66938645",
"0.66938645",
"0.6689058",
"0.6689058",
"0.66818196",
"0.66816723",
"0.6678679",
"0.6642075",
"0.664175",
"0.662279",
"0.662215",
"0.6618726",
"0.6618112",
"0.6599783",
"0.65981686",
"0.65892166",
"0.65814334",
"0.6574515",
"0.6574515",
"0.65700406",
"0.6564297",
"0.6557278",
"0.65522736",
"0.65515196",
"0.6546496",
"0.65435827",
"0.65386873",
"0.6536578",
"0.653494",
"0.65291256",
"0.6525413",
"0.65238154",
"0.65162224",
"0.6510706",
"0.65106267",
"0.6507414",
"0.6504123",
"0.6494102",
"0.64895123",
"0.64889216",
"0.6487698",
"0.6486252",
"0.6486199",
"0.64828074",
"0.6482261",
"0.64796287",
"0.64673734",
"0.64666325",
"0.6462686",
"0.6445045",
"0.64445",
"0.6443093",
"0.6442955",
"0.64347535",
"0.6434285",
"0.6429108",
"0.6428073",
"0.6427048",
"0.64256585",
"0.6425172",
"0.6420639",
"0.64163566",
"0.64107656",
"0.6408059",
"0.64048976",
"0.6404037",
"0.6403875",
"0.6400773",
"0.639468",
"0.6392691",
"0.63895977",
"0.6388387",
"0.63833517",
"0.63694376",
"0.6365491",
"0.63652736",
"0.63648796",
"0.6353102",
"0.6352744",
"0.6343559",
"0.6343413",
"0.6340693"
] |
0.0
|
-1
|
GET /product_showrooms def index
|
def destroy
if current_user.id == @product_showroom.showroom.owner_id
@product_showroom.destroy
else
render json: {errors: ['Unauthorized']}, status: :unauthorized
end
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def index\n @rooms = Room.all\n respond_to do | format |\n format.html\n format.json\n end\n \n end",
"def index\n @rooms = Room.all\n \n respond_to do |f|\n f.html\n end\n end",
"def index\n @restrooms = Restroom.all\n end",
"def index\n @rooms = Room.all\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @rooms }\n end\n end",
"def index\n @rooms = Room.all\n end",
"def index\n @rooms = Room.all\n end",
"def index\n @rooms = Room.all\n end",
"def index\n @rooms = Room.all\n end",
"def index\n @rooms = Room.all\n end",
"def index\n @rooms = Room.all\n end",
"def index\n @rooms = Room.all\n end",
"def index\n @rooms = Room.all\n end",
"def index\n @rooms = Room.all\n end",
"def index\n @rooms = Room.all\n end",
"def index\n @rooms = Room.all\n end",
"def index\n @rooms = Room.all\n\n respond_to do |format|\n format.html # index.html.erb\n format.xml { render :xml => @rooms }\n end\n end",
"def show\n @rooms = Room.all\n end",
"def index\n @hotel_rooms = HotelRoom.all\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @hotel_rooms }\n end\n end",
"def index\n @rooms = current_user.rooms.all\n end",
"def index\n @rooms = @rooms.search(description_cont: params[:search]).result.paginate(per_page: 15, page: params[:page])\n end",
"def show\n @rooms = @blueprint.rooms\n end",
"def index\n @search = @rooms.search(params[:q])\n @rooms = @search.result.paginate(:page => params[:page])\n end",
"def index\n\n # get the gauges to choose from\n @gauges = Product.gauges\n\n @device = detect_browser()\n\n # get the applications to choose from\n @applications = Product.applications\n\n @products = Product.all\n\n prod_params = params[:product]\n @selected_app = \"\"\n @selected_awg = \"\"\n\n if prod_params != nil\n if prod_params[\"selected_app\"] != nil\n @selected_app = prod_params[\"selected_app\"]\n @products = @products.single_application(@selected_app)\n logger.debug \"products single app: #{@products.inspect}\"\n end\n if prod_params[\"selected_awg\"] != nil\n @selected_awg = prod_params[\"selected_awg\"]\n @products = @products.single_gauge(@selected_awg)\n logger.debug \"products single gauge: #{@products.inspect}\"\n end\n\n #params[:page] = prod_params[\"page\"] if params[:page] == nil\n\n end\n\n @search_terms = params[\"search_terms\"] unless params[\"search_terms\"] == nil\n\n if params[:page] == nil\n @products = @products.search(params[\"search_terms\"], 1)\n @page = 1\n else\n @products = @products.search(params[\"search_terms\"], params[:page])\n @page = params[:page]\n end\n end",
"def index\r\n @product_ware_houses = ProductWareHouse.all\r\n end",
"def index\n @rooms=Room.all\n respond_to do |format|\n format.html\n end\n end",
"def index\n\t\t#this is the home page\n\t\t@rooms = Room.limit(5)\n\tend",
"def index\n\t\t@rooms = Room.order(updated_at: :desc)\n\t\trespond_to do |format|\n\t\t\tformat.html\n\t\t\tformat.json { render json: @rooms }\n\t\tend\n\tend",
"def index\n @hostel_rooms = HostelRoom.all\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @hostel_rooms }\n end\n end",
"def index\n b_admin = current_user.admin? rescue false\n @products = Product.filter_by_params(b_admin, params)\n #@products = Product.available\n \n @title = Product.page_description(params)\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @products }\n end\n end",
"def index\n @rooms = Room.paginate(:page => params[:page])\n end",
"def index\n @reservation_rooms = ReservationRoom.all\n end",
"def index\n @branch_rooms = Room.get_all_rooms(params[:branch_id])\n \n render json: @branch_rooms\n end",
"def list_products\n render \"products_view/all\"\n end",
"def index\n case current_type\n when 'owner'\n @rooms = owner_rooms\n when 'access'\n @rooms = access_rooms\n else\n @rooms = public_rooms\n end\n end",
"def index\n @reserved_rooms = ReservedRoom.all\n end",
"def index\n @rooms = Room.all\n @rooms = @rooms.search(params[:search]) if params[:search].present?\n end",
"def index\n @rooms = Room.all\n @rooms = @venue.rooms.all\n end",
"def index\n @product_sections = ProductSection.all\n\n respond_to do |format|\n format.html # index.html.erb\n format.xml { render :xml => @product_sections }\n end\n end",
"def index\n @request_rooms = RequestRoom.all\n end",
"def index\n @reservation2_rooms = Reservation2Room.all\n end",
"def index\n @products = Product.all\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @inventories }\n end\n end",
"def index\n @rooms = Room.find(:all, :conditions => [\"site_id = ?\", params[:site_id]])\n\n respond_to do |format|\n format.html # index.html.erb\n format.xml { render :xml => @rooms }\n end\n end",
"def index\n\t\t@products = Product.all\n\tend",
"def room_list\n @rooms = Room.where(active: true)\n render partial: \"room_list\"\n end",
"def index\n @occupied_rooms = OccupiedRoom.all\n end",
"def index\n limit = params[:limit]&.to_i || 10\n page = params[:page]&.to_i || 0\n if params[:available] == \"1\"\n @products = Product.paginate(page, limit).available\n else\n @products = Product.paginate(page, limit)\n end\n render json: @products\n end",
"def index\n @student_rooms = StudentRoom.all\n end",
"def index\n @orden_products = OrdenProduct.all\n end",
"def index\n # Get all the products\n @products = Product.all\n end",
"def show\n @room = Room.find(params[:id])\n\n render '/api/rooms/show'\n end",
"def index\n @class_rooms = ClassRoom.all\n end",
"def index\n @availability_products = AvailabilityProduct.all\n end",
"def index\n @floor_plans = @product.floor_plans\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @floor_plans }\n end\n end",
"def index\n if authorise(request)\n auth_token = request.headers['Authorization1'].split(' ').last\n req = GetRoomsRequest.new(auth_token)\n serv = RoomServices.new\n resp = serv.get_rooms(req)\n render json: { status: resp.success, message: resp.message, data: resp.data }, status: :ok\n else\n render json: { status: false, message: 'Unauthorized' }, status: 401\n end\n end",
"def index\n @about_rooms = AboutRoom.all\n end",
"def index\n rooms = Room.includes(:problems, :registrations, :users, :owner)\n if params[:name]\n render json: rooms.where(name: params[:name])\n else\n render json: rooms\n end\n end",
"def index\n @productors = Productor.all\n end",
"def products\n run(:get,\"/school_products\", [200])\n end",
"def index\n @show_rooms = ShowRoom.find(:all,:conditions => {:user_id => current_user.id},:order => \"created_at DESC\")\n\n respond_to do |format|\n format.html # index.html.erb\n format.xml { render :xml => @show_rooms }\n end\n end",
"def index\n @employee_rooms = EmployeeRoom.all\n end",
"def show\n @hostel = Hostel.find(params[:id])\n @rooms = @hostel.rooms\n respond_to do |format|\n format.html # show.html.erb\n format.xml { render :xml => @hostel }\n end\n end",
"def index\n \t\t# if params[:search].present? # check if somebody typed into searchbar\n # \t\t@products = Product.search(params[:search])\n \t\t# else # if not show all active products\n\n \t\t# Show all active products and suffle them to always show a different view\n\t\t\t@products = Product.where(active: true)\n\t\t\t@products.shuffle\n \t\t# end\n\t\tjson_response(@products)\n\tend",
"def index\n @rooms = Room.all.map do |rm|\n Room.cleanFetch(rm, current_user)\n end\n end",
"def index\n @classrooms = Classroom.all\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @classrooms }\n end\n end",
"def products\n request :public, :get, :products\n end",
"def index\n before_filter { |c| c.include_stylesheet 'admin/extensions/shop/products/products' }\n before_filter { |c| c.include_javascript 'admin/pagefactory' }\n @shop_products = ShopProduct.search(params[:query])\n \n unless @shop_products.empty?\n @radiant_layout = Radiant::Config['shop.category_layout']\n\n respond_to do |format|\n format.html { render }\n format.js { render :partial => '/shop/products/products', :collection => @shop_products }\n format.xml { render :xml => @shop_products.to_xml(attr_hash) }\n format.json { render :json => @shop_products.to_json(attr_hash) }\n end\n else\n render :template => 'site/not_found', :status => 404\n end\n end",
"def index\n @products = Product.all\n end",
"def index\n ota_rooms = Array.new\n if params[:room].nil? && params[:ota].nil?\n # 自分の施設情報だけを返す\n\n current_user.ota.select do |otum|\n otum.provider == 'default'\n end.each do |otum|\n ota_rooms.concat(otum.ota_rooms)\n end\n elsif params[:ota].nil?\n room = Room.find_by(id: params[:room].to_i)\n # 部屋IDが存在しない場合\n if room.nil?\n response = {\n status: 'NotFound',\n message: 'room not found'\n }\n return render json: response, status: 404\n end\n # 権限がない場合\n if current_user.id != room.user.id\n response = {\n status: 'NotFound',\n message: 'room not found'\n }\n return render json: response, status: 404\n end\n ota_rooms.concat(room.ota_rooms)\n elsif params[:room].nil?\n otum = Otum.find_by(id: params[:ota].to_i)\n # OTA IDが存在しない場合\n if otum.nil?\n response = {\n status: 'NotFound',\n message: 'ota not found'\n }\n return render json: response, status: 404\n end\n # 権限がない場合\n if current_user.id != otum.user.id\n response = {\n status: 'NotFound',\n message: 'ota not found'\n }\n return render json: response, status: 404\n end\n ota_rooms.concat(otum.ota_rooms)\n else\n # facility=[facility_id]\n facility = Facility.find_by(id: params[:facility].to_i)\n # 施設IDが存在しない場合\n if facility.nil?\n response = {\n status: 'NotFound',\n message: 'ota_rooms not found'\n }\n return render json: response, status: 404\n end\n # ユーザーの施設IDでない場合\n if facility.user_id != current_user.id\n response = {\n status: 'NotFound',\n message: 'ota_rooms not found'\n }\n return render json: response, status: 404\n end\n ota_rooms.concat(facility.ota_rooms)\n end\n\n render json: {\n status: 'Success',\n message: '',\n ota_rooms: ota_rooms.as_json(except: [\n :created_at,\n :updated_at\n ])\n }, status: 200\n end",
"def index\n @room_plants = RoomPlant.all\n end",
"def index\n @kind_rooms = KindRoom.all\n end",
"def index\n\t@products=products.all\nend",
"def index\n render :json => Producto.all\n end",
"def index\n @products = get_products\n \n respond_to do |format|\n format.html # index.html.haml\n format.js # index.js.rjs\n format.xml { render :xml => @products }\n end\n end",
"def index\n expose Product.page(params[:page])\n end",
"def home\n @products = Product.all\n end",
"def index\n \tproducts = Product.all\n \trender json: products\n \tend",
"def index\n @products = Product.all\n end",
"def index\n @products = Product.all\n end",
"def index\n @products = Product.all\n end",
"def index\n @products = Product.all\n end",
"def index\n @products = Product.all\n end",
"def index\n @products = Product.all\n end",
"def index\n @products = Product.all\n end",
"def index\n @products = Product.all\n end",
"def index\n @products = Product.all\n end",
"def index\n @products = Product.all\n end",
"def index\n @products = Product.all\n end",
"def index\n @products = Product.all\n end",
"def index\n @products = Product.all\n end",
"def index\n @products = Product.all\n end",
"def index\n @products = Product.all\n end",
"def index\n @products = Product.all\n end",
"def index\n @products = Product.all\n end",
"def index\n @products = Product.all\n end",
"def index\n @products = Product.all\n end",
"def index\n @products = Product.all\n end",
"def index\n @products = Product.all\n end",
"def index\n @products = Product.all\n end",
"def index\n @products = Product.all\n end",
"def index\n @products = Product.all\n end",
"def index\n @products = Product.all\n end"
] |
[
"0.7250432",
"0.721175",
"0.7173299",
"0.71025777",
"0.707864",
"0.707864",
"0.707864",
"0.707864",
"0.707864",
"0.707864",
"0.707864",
"0.707864",
"0.707864",
"0.707864",
"0.707864",
"0.7077684",
"0.69886446",
"0.6947703",
"0.6929497",
"0.69233197",
"0.6887779",
"0.6874277",
"0.6850722",
"0.6799419",
"0.67928123",
"0.678888",
"0.67883164",
"0.6771616",
"0.6751901",
"0.6751519",
"0.67310065",
"0.6723371",
"0.67123556",
"0.67097527",
"0.6686631",
"0.6681118",
"0.6680773",
"0.66456515",
"0.6640627",
"0.66393477",
"0.6635735",
"0.66253626",
"0.66108966",
"0.66101766",
"0.66073066",
"0.6600759",
"0.6592339",
"0.6581271",
"0.65769863",
"0.6572499",
"0.65712094",
"0.65641695",
"0.65636855",
"0.65549856",
"0.6554727",
"0.6551839",
"0.65453315",
"0.6538917",
"0.6538882",
"0.6538082",
"0.65373987",
"0.65365416",
"0.65097183",
"0.65071785",
"0.6506884",
"0.64968914",
"0.6489309",
"0.6486221",
"0.64808214",
"0.64695454",
"0.64639515",
"0.6463458",
"0.64595103",
"0.6448577",
"0.64471376",
"0.6439932",
"0.64342976",
"0.64342976",
"0.64342976",
"0.64342976",
"0.64342976",
"0.64342976",
"0.64342976",
"0.64342976",
"0.64342976",
"0.64342976",
"0.64342976",
"0.64342976",
"0.64342976",
"0.64342976",
"0.64342976",
"0.64342976",
"0.64342976",
"0.64342976",
"0.64342976",
"0.64342976",
"0.64342976",
"0.64342976",
"0.64342976",
"0.64342976",
"0.64342976"
] |
0.0
|
-1
|
Use callbacks to share common setup or constraints between actions.
|
def set_product_showroom
@product_showroom = ProductShowroom.find(params[:id])
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def set_required_actions\n # TODO: check what fields change to asign required fields\n end",
"def action_hook; end",
"def run_actions; end",
"def define_action_hook; end",
"def actions; end",
"def define_action_helpers\n if super && action == :save\n @instance_helper_module.class_eval do\n define_method(:valid?) do |*args|\n self.class.state_machines.fire_event_attributes(self, :save, false) { super(*args) }\n end\n end\n end\n end",
"def add_actions; end",
"def callbacks; end",
"def callbacks; end",
"def setup *actions, &proc\n (@setup_procs ||= []) << [proc, actions.size > 0 ? actions : [:*]]\n end",
"def define_action_helpers; end",
"def post_setup\n end",
"def action_methods; end",
"def action_methods; end",
"def action_methods; end",
"def before_setup; end",
"def action_run\n end",
"def execute(setup)\n @action.call(setup)\n end",
"def define_action_helpers?; end",
"def set_actions\n actions :all\n end",
"def action_done(action)\n dispatch = { :migrate => :done_migrating, :map => :done_mapping, :reduce =>\n :done_reducing, :finalize => :done_finalizing } \n self.send dispatch[action[:action]], action\n end",
"def dependencies action, &block\n @actions.each do |other|\n if action[:requires].include? other[:provide]\n block.call other\n end\n end\n end",
"def setup!\n return unless @setup_procs\n http_actions = actions\n @setup_procs.each do |setup_proc|\n proc, actions = setup_proc\n @setup__actions = actions.map do |action|\n\n action.is_a?(Regexp) ?\n http_actions.select { |a| a.to_s =~ action } :\n action.is_a?(String) && action =~ /\\A\\./ ?\n http_actions.map { |a| a.to_s << action if format?(a).include?(action) }.compact :\n action\n\n end.flatten\n self.class_exec &proc\n @setup__actions = nil\n end\n @setup_procs = nil\n end",
"def before_actions(*logic)\n self.before_actions = logic\n end",
"def setup_handler\n end",
"def set_action(opts)\n opts = check_params(opts,[:actions])\n super(opts)\n end",
"def setup(action)\n @targets.clear\n unless action.item.target_filters.empty?\n @targets = SES::TargetManager.make_targets(action)\n else\n item = action.item\n if item.for_opponent?\n @targets = $game_troop.alive_members\n elsif item.for_dead_friend?\n @targets = $game_party.battle_members.select { |actor| actor.dead? }\n else\n $game_party.battle_members.select { |actor| actor.alive? }\n end\n end\n @item_max = @targets.size\n create_contents\n refresh\n show\n activate\n end",
"def action; end",
"def action; end",
"def action; end",
"def action; end",
"def action; end",
"def workflow\n end",
"def revisable_shared_setup(args, block)\n class << self\n attr_accessor :revisable_options\n end\n options = args.extract_options!\n self.revisable_options = Options.new(options, &block)\n \n self.send(:include, Common)\n self.send(:extend, Validations) unless self.revisable_options.no_validation_scoping?\n self.send(:include, WithoutScope::QuotedColumnConditions)\n end",
"def setup\n @action = SampleActionAndroid.new(os_name: 'android',\n app_name: APP_PATH)\n end",
"def before(action)\n invoke_callbacks *self.class.send(action).before\n end",
"def process_action(...)\n send_action(...)\n end",
"def before_dispatch(env); end",
"def after_actions(*logic)\n self.after_actions = logic\n end",
"def setup\n # override and do something appropriate\n end",
"def setup(client)\n return unless @setup\n actions = @setup['setup'].select { |action| action['do'] }.map { |action| Action.new(action['do']) }\n actions.each do |action|\n action.execute(client)\n end\n self\n end",
"def setup(_context)\n end",
"def setup(resources) ; end",
"def validate_actions\n errors.add(:base, :should_give_at_least_one_action) if !manage? && !forecasting? && !read? && !api?\n end",
"def setup\n @resource_config = {\n :callbacks => {\n :before_create => nil,\n :after_create => nil,\n :before_update => nil,\n :after_update => nil,\n :before_destroy => nil,\n :after_destroy => nil,\n },\n :child_assoc => nil,\n :model => nil,\n :parent => nil,\n :path => nil,\n :permission => {},\n :properties => {},\n :relation => {\n :create => nil,\n :delete => nil,\n },\n :roles => nil,\n }\n end",
"def determine_valid_action\n\n end",
"def process_shared\n handle_taxes\n handle_shippings\n create_adjustments_from_params\n handle_status\n handle_inventory_refunds\n handle_payment_transactions\n order.updater.update\n end",
"def startcompany(action)\n @done = true\n action.setup\n end",
"def init_actions\n am = action_manager()\n am.add_action(Action.new(\"&Disable selection\") { @selection_mode = :none; unbind_key(32); bind_key(32, :scroll_forward); } )\n am.add_action(Action.new(\"&Edit Toggle\") { @edit_toggle = !@edit_toggle; $status_message.value = \"Edit toggle is #{@edit_toggle}\" })\n end",
"def event_callbacks(event, metadata={})\n case event\n when :reset, :review\n if confirmed\n update_attributes(confirmed: false)\n end\n when :confirm\n confirm\n # trigger :order for all applicable items\n # NOTE: :order event is common to both physical and digital items\n items.each do |i|\n if i.event_permitted(:order)\n user_id = last_transition.user_id\n i.trigger!(:order, { order_id: id, user_id: user_id })\n end\n end\n when :complete_work\n request = metadata[:request]\n work_complete_notification(request)\n when :close\n close\n end\n if event != :close && !open\n reopen\n end\n end",
"def setup_action\n return unless PONY::ERRNO::check_sequence(current_act)\n new_sequence = @action_sequence[@sequence_index+1...@action_sequence.size]\n @sequence_index = 0\n new_sequence = DND::SkillSequence::ACTS[@acts[1]] + new_sequence\n execute_sequence\n end",
"def define_tasks\n define_weave_task\n connect_common_tasks\n end",
"def setup(&block)\n define_method(:setup, &block)\n end",
"def setup\n transition_to(:setup)\n end",
"def setup\n transition_to(:setup)\n end",
"def action\n end",
"def setup( *args )\n\t\t\tself.class.setupBlocks.each {|sblock|\n\t\t\t\tdebugMsg \"Calling setup block method #{sblock}\"\n\t\t\t\tself.send( sblock )\n\t\t\t}\n\t\t\tsuper( *args )\n\t\tend",
"def config(action, *args); end",
"def setup\n @setup_proc.call(self) if @setup_proc\n end",
"def before_action \n end",
"def setup_callbacks\n defined_callbacks.each do |meth|\n unless respond_to?(\"call_#{meth}_callbacks\".to_sym)\n self.class.module_eval <<-EOE\n def call_#{meth}_callbacks(*args)\n plugin_store.each {|a| a.call_#{meth}_callbacks(*args) } if respond_to?(:plugin_store) && plugin_store\n self.send :#{meth}, *args if respond_to?(:#{meth})\n end\n EOE\n end\n end\n end",
"def action\n end",
"def matt_custom_action_begin(label); end",
"def setup\n # override this if needed\n end",
"def setup\n\t\t\t\t\t\t# Do nothing\n\t\t\t\tend",
"def setup\n\t\t\t\t\t\t# Do nothing\n\t\t\t\tend",
"def action(options,&callback)\n new_action = Action===options ? options : Action.new(options,&callback)\n # replace any with (shared name/alias or both default) + same arity\n @actions.delete_if do |existing_action|\n ((existing_action.names & new_action.names).size > 0 ||\n existing_action.default? && new_action.default?) &&\n existing_action.required.size == new_action.required.size &&\n existing_action.optional.size <= new_action.optional.size\n end\n @actions = (@actions + [new_action]).sort\n new_action\n end",
"def set_target_and_action target, action\n self.target = target\n self.action = 'sugarcube_handle_action:'\n @sugarcube_action = action\n end",
"def after(action)\n invoke_callbacks *options_for(action).after\n end",
"def pre_task\n end",
"def setup(server)\n server.on('beforeMethod', method(:before_method), 10)\n end",
"def add_actions\n attribute = machine.attribute\n name = self.name\n \n owner_class.class_eval do\n define_method(name) {self.class.state_machines[attribute].events[name].fire(self)}\n define_method(\"#{name}!\") {self.class.state_machines[attribute].events[name].fire!(self)}\n define_method(\"can_#{name}?\") {self.class.state_machines[attribute].events[name].can_fire?(self)}\n end\n end",
"def init_actions\n @select_action = SelectAction.new\n @endpoint_mouse_action = EndpointMouseAction.new\n @move_action = MoveAction.new\n end",
"def setup_signals; end",
"def after_created\r\n return unless compile_time\r\n Array(action).each do |action|\r\n run_action(action)\r\n end\r\nend",
"def after_created\r\n return unless compile_time\r\n Array(action).each do |action|\r\n run_action(action)\r\n end\r\nend",
"def set_target_and_action target, action\n self.target = target\n self.action = 'sugarcube_handle_action:'\n @sugarcube_action = action.respond_to?('weak!') ? action.weak! : action\n end",
"def initialize(*args)\n super\n @action = :set\nend",
"def after_set_callback; end",
"def setup\n #implement in subclass;\n end",
"def lookup_action; end",
"def setup &block\n if block_given?\n @setup = block\n else\n @setup.call\n end\n end",
"def setup_action\n return TSBS.error(@acts[0], 1, @used_sequence) if @acts.size < 2\n actions = TSBS::AnimLoop[@acts[1]]\n if actions.nil?\n show_action_error(@acts[1])\n end\n @sequence_stack.push(@acts[1])\n @used_sequence = @acts[1]\n actions.each do |acts|\n @acts = acts\n execute_sequence\n break if @break_action\n end\n @sequence_stack.pop\n @used_sequence = @sequence_stack[-1]\n end",
"def release_actions; end",
"def around_hooks; end",
"def save_action; end",
"def setup(easy)\n super\n easy.customrequest = @verb\n end",
"def action_target()\n \n end",
"def setup\n callback(:setup) do\n notify(:setup)\n migration_check.last_deployed_commit\n end\n end",
"def setup\n return unless @setup\n\n actions = @setup['setup'].select { |action| action['do'] }.map { |action| Action.new(action['do']) }\n run_actions_and_retry(actions)\n self\n end",
"def before_setup\n # do nothing by default\n end",
"def my_actions(options)\n @setup = false\n get_template_part(\"custom_used\",\"action_users\",true)\n end",
"def default_action; end",
"def setup(&blk)\n @setup_block = blk\n end",
"def callback_phase\n super\n end",
"def advice\n end",
"def _handle_action_missing(*args); end",
"def duas1(action)\n action.call\n action.call\nend",
"def shared_action(name, &block)\n @controller.shared_actions[name] = block\n end",
"def before_action action, &block\n @audience[:before][action] ||= Set.new\n @audience[:before][action] << block\n end",
"def setup_initial_state\n\n state_a = State.new(\"a\", 0)\n state_b = State.new(\"b\", 0)\n state_c = State.new(\"c\", 10)\n\n move_to_b = Action.new(\"move_to_b\", 1, state_b)\n\n move_to_c = Action.new(\"move_to_c\", 1, state_c)\n\n state_a.actions = [move_to_b, move_to_c]\n\n return state_a\n \nend"
] |
[
"0.6163163",
"0.6045976",
"0.5946146",
"0.591683",
"0.5890051",
"0.58349305",
"0.5776858",
"0.5703237",
"0.5703237",
"0.5652805",
"0.5621621",
"0.54210985",
"0.5411113",
"0.5411113",
"0.5411113",
"0.5391541",
"0.53794575",
"0.5357573",
"0.53402257",
"0.53394014",
"0.53321576",
"0.53124547",
"0.529654",
"0.5296262",
"0.52952296",
"0.52600986",
"0.52442724",
"0.52385926",
"0.52385926",
"0.52385926",
"0.52385926",
"0.52385926",
"0.5232394",
"0.523231",
"0.5227454",
"0.52226824",
"0.52201617",
"0.5212327",
"0.52079266",
"0.52050185",
"0.51754695",
"0.51726824",
"0.51710224",
"0.5166172",
"0.5159343",
"0.51578903",
"0.51522785",
"0.5152022",
"0.51518047",
"0.51456624",
"0.51398855",
"0.5133759",
"0.5112076",
"0.5111866",
"0.5111866",
"0.5110294",
"0.5106169",
"0.509231",
"0.50873137",
"0.5081088",
"0.508059",
"0.50677156",
"0.50562143",
"0.5050554",
"0.50474834",
"0.50474834",
"0.5036181",
"0.5026331",
"0.5022976",
"0.5015441",
"0.50121695",
"0.5000944",
"0.5000019",
"0.4996878",
"0.4989888",
"0.4989888",
"0.49864885",
"0.49797225",
"0.49785787",
"0.4976161",
"0.49683493",
"0.4965126",
"0.4958034",
"0.49559742",
"0.4954353",
"0.49535993",
"0.4952725",
"0.49467874",
"0.49423352",
"0.49325448",
"0.49282882",
"0.49269363",
"0.49269104",
"0.49252945",
"0.4923091",
"0.49194667",
"0.49174926",
"0.49173003",
"0.49171105",
"0.4915879",
"0.49155936"
] |
0.0
|
-1
|
Only allow a trusted parameter "white list" through.
|
def product_showroom_params
params.require(:product_showroom).permit(:product_id, :showroom_id, :status)
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def allowed_params\n ALLOWED_PARAMS\n end",
"def expected_permitted_parameter_names; end",
"def param_whitelist\n [:role, :title]\n end",
"def default_param_whitelist\n [\"mode\"]\n end",
"def permitir_parametros\n \t\tparams.permit!\n \tend",
"def permitted_params\n []\n end",
"def strong_params\n params.require(:user).permit(param_whitelist)\n end",
"def strong_params\n params.require(:listing_member).permit(param_whitelist)\n end",
"def filtered_parameters; end",
"def permitted_strong_parameters\n :all #or an array of parameters, example: [:name, :email]\n end",
"def parameters_list_params\n params.require(:parameters_list).permit(:name, :description, :is_user_specific)\n end",
"def parameter_params\n params.require(:parameter).permit(:name, :description, :param_code, :param_value, :active_from, :active_to)\n end",
"def param_whitelist\n whitelist = [\n :description,\n :progress,\n :kpi_id\n ]\n \n unless action_name === 'create'\n whitelist.delete(:kpi_id)\n end\n \n whitelist\n end",
"def param_whitelist\n whitelist = [\n :username, :name,\n :parent_id,\n :headline, :description, :video,\n :policy, :signup_mode, :category,\n :website, :facebook, :twitter, :linkedin,\n :founded_at,\n privacy: [\n :events,\n :resources\n ],\n permission: [\n :profile,\n :members,\n :children,\n :statistics,\n :posts,\n :listings,\n :resources,\n :events\n ],\n location: [\n :description,\n :street,\n :city,\n :state,\n :zip,\n :country,\n :latitude,\n :longitude\n ]\n ]\n \n if action_name === 'update'\n whitelist.delete(:parent_id)\n unless current_user.role_in(@community) === 'owner'\n whitelist.delete(:privacy)\n whitelist.delete(:permission)\n end\n end\n \n whitelist\n end",
"def param_whitelist\n [:rating, :review]\n end",
"def valid_params?; end",
"def permitted_params\n declared(params, include_missing: false)\n end",
"def permitted_params\n declared(params, include_missing: false)\n end",
"def get_params\n\t\treturn ActionController::Parameters.new(self.attributes).permit(\"account_id\", \"title\", \"category\", \"introduction\", \"tags\", \"segment_type\", \"visible\", \"status\", \"main_image\")\n\tend",
"def filter_parameters; end",
"def filter_parameters; end",
"def strong_params\n params.require(:team_member).permit(param_whitelist)\n end",
"def strong_params\n params.require(:community).permit(param_whitelist)\n end",
"def check_params; true; end",
"def valid_params_request?; end",
"def strong_params\n params.require(:experience).permit(param_whitelist)\n end",
"def allowed_params\n params.require(:user).permit(:username, :email, :password, :password_confirmation)\n end",
"def list_params\n params.permit(:name)\n end",
"def check_params\n true\n end",
"def grant_params\n @whitelisted = params.require(:grant).permit(:name, :description, :agency_id, :acronym)\n end",
"def safe_params\n resurce_name = self.class.resource_name\n params_method_name = \"#{resurce_name}_params\".to_sym\n if params[resurce_name]\n if respond_to?(params_method_name) || private_methods.include?(params_method_name)\n send(params_method_name)\n else\n raise ActiveModel::ForbiddenAttributesError, \"Please, define the '#{params_method_name}' method in #{self.class.name}\"\n end\n end\n end",
"def additional_permitted_params\n []\n end",
"def strong_params\n params.require(:education).permit(param_whitelist)\n end",
"def resource_params\n params[resource_singular_name].try(:permit, self.class.param_whitelist)\n end",
"def allow_params_authentication!; end",
"def param_whitelist\n [\n :title,\n :description,\n :organization,\n :team_id,\n :started_at,\n :finished_at,\n location: [\n :description,\n :street,\n :city,\n :state,\n :zip,\n :country,\n :latitude,\n :longitude\n ]\n ]\n end",
"def param_whitelist\n if @user.present? && current_user != @user\n return [:followed]\n end\n \n whitelist = [\n :username, :email, :password,\n :first_name, :last_name,\n :birthday, :gender,\n :headline, :biography, :ask_about, :focus,\n :website, :facebook, :linkedin, :twitter, :github,\n roles: [],\n skills: [],\n interests: [],\n privacy: { contact: [] },\n location: [\n :description,\n :street,\n :city,\n :state,\n :zip,\n :country,\n :latitude,\n :longitude\n ]\n ]\n \n if action_name === 'update'\n whitelist.delete(:email)\n whitelist.delete(:password)\n end\n \n whitelist\n end",
"def person_params\n # params whitelist does *not* include admin, sub, remember_token\n # TBD: share this whitelist with the list used by configuration_permitted_parameters\n # TBD: should current_password be on this list? -- for now, leaving off, since it seems to work without\n # NOTE: do not include 'admin' in this list!\n params.require(:person).permit(\n :name, \n :email, \n :description,\n :password, \n :password_confirmation\n )\n end",
"def paramunold_params\n params.require(:paramunold).permit!\n end",
"def param_params\n params.require(:param).permit(:param_category_id, :param_table_id, :name, :english_name, :weighting, :description)\n end",
"def quote_params\n params.permit!\n end",
"def list_params\n params.permit(:list_name)\n end",
"def allowed_params(parameters)\n parameters.select do |name, values|\n values.location != \"path\"\n end\n end",
"def all_params; end",
"def permitted_resource_params\n params[resource.object_name].present? ? params.require(resource.object_name).permit! : ActionController::Parameters.new\n end",
"def source_params\n params.require(:source).permit(all_allowed_params)\n end",
"def user_params\n end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def get_allowed_parameters\n return _get_specific_action_config(:allowed_action_parameters, :allowed_parameters)&.map(&:to_s)\n end",
"def permitted_params\n @wfd_edit_parameters\n end",
"def user_params\r\n end",
"def param_whitelist\n whitelist = [\n :comment,\n :old_progress, :new_progress,\n :metric_id\n ]\n \n unless action_name === 'create'\n whitelist.delete(:metric_id)\n end\n \n whitelist\n end",
"def query_param\n\t\tparams.permit(:first_name, :last_name, :phone)\n\tend",
"def whitelisted_user_params\n params.require(:user).\n permit( :first_name, :last_name, :email,:password,:password_confirmation,:birthday,:gender)\n end",
"def filter_params\n\t\treturn params[:candidate].permit(:name_for_filter)\n\tend",
"def user_params\n params.permit(:id, :email, :password, :nickname, :status, :avatar, :flat_picture, :flatsharing_id, :member,\n :user, :color, :solde)\n end",
"def get_params\n\t\t\n\t\treturn ActionController::Parameters.new(self.attributes).permit(:first_name, :last_name, :email, :provider)\n\n\tend",
"def devise_filter\r\n logger.debug(\"In devise_filter =>PARAMS: #{params.inspect}\")\r\n\r\n # White list for sign_up\r\n devise_parameter_sanitizer.for(:sign_up) { |u| u.permit(user_whitelist) }\r\n\r\n # White list for account update\r\n devise_parameter_sanitizer.for(:account_update) { |u| u.permit(user_whitelist, :current_password) }\r\n\r\n # White list for Invitation creation\r\n devise_parameter_sanitizer.for(:invite) { |u| u.permit(:account_type, :email, :invitation_token)}\r\n\r\n # White list for accept invitation\r\n devise_parameter_sanitizer.for(:accept_invitation) { |u| u.permit(user_whitelist, :invitation_token)}\r\n\r\n end",
"def valid_params(params)\n params.permit(:user_id, :photo_id, :originX, :originY, :width, :height)\n end",
"def valid_parameters\n sort_symbols(@interface.allowed_parameters)\n end",
"def params_permit\n params.permit(:id)\n end",
"def allowed_params\n params.require(:allowed).permit(:email)\n end",
"def allowed_params\n params.require(:user).permit(:email, :password, :role, :first_name, :last_name, :password_confirmation)\n end",
"def allowed_params\n params.require(:user).permit(:email, :password, :role, :first_name, :last_name, :password_confirmation)\n end",
"def filter_params\n params.permit(*resource_filter_permitted_params)\n end",
"def community_params\n params.permit(:profile_image, :name, :description, :privacy_type, :viewed_by, {tags: []}, {features: []}, {admins: []}, :members, :location, :beacon, :creator, :ambassadors, :current_events, :past_events, :feed, :category, :address, :allow_member_post_to_feed, :allow_member_post_to_events)\n end",
"def specialty_params\n\t\tparams.require(:specialty).permit(*Specialty::DEFAULT_ACCESSIBLE_ATTRIBUTES)\n\tend",
"def authorize_params\n super.tap do |params|\n %w[display scope auth_type].each do |v|\n if request.params[v]\n params[v.to_sym] = request.params[v]\n end\n end\n end\n end",
"def feature_params_filter\n params.require(:feature).permit(:name, :cat, :lower, :upper, :opts, :category, :description, :company, :active, :unit, :icon)\n end",
"def available_activity_params\n # params.require(:available_activity).permit(:type,:geometry,:properties)\n whitelisted = ActionController::Parameters.new({\n type: params.require(:available_activity)[:type],\n geometry: params.require(:available_activity)[:geometry].try(:permit!).to_h,\n properties: params.require(:available_activity)[:properties].try(:permit!).to_h\n }).try(:permit!)\n end",
"def argument_params\n params.require(:argument).permit(:name)\n end",
"def user_params_pub\n\t \tparams[:user].permit(:hruid)\n\t end",
"def strong_params\n params.require(:success_metric).permit(param_whitelist)\n end",
"def property_params\n params.permit(:name, :is_available, :is_approved, :owner_id)\n end",
"def restricted_params\n #params.require(self.controller_name.classify.underscore.to_sym).permit([])\n raise(\"No strong params set, override restricted_params method in your controller. E.g. params.require(:model).permit(:attribute1, :attribute2)\")\n end",
"def sponsor_params\n params.require(:sponsor).permit(WHITE_LIST)\n end",
"def whitelist_person_params\n params.require(:person).permit(:family, :pre_title, :given_name, :dates, :post_title, :epithet, :dates_of_office, same_as: [], related_authority: [], altlabel: [], note: []) # Note - arrays need to go at the end or an error occurs!\n end",
"def parameters\n nil\n end",
"def user_params \n \tparams.require(:user).permit(:name, :email, :password, :password_confirmation)# preventing CSTR\n end",
"def sequence_param_whitelist\n default_param_whitelist << \"show_index\"\n end",
"def resource_filter_permitted_params\n raise(NotImplementedError, 'resource_filter_permitted_params method not implemented')\n end",
"def normal_params\n reject{|param, val| param_definitions[param][:internal] }\n end",
"def validate_search_inputs\n @whitelisted = params.fetch(:user, nil)\n if @whitelisted.blank?\n render_error(400, \"#{I18n.t('general_error.params_missing_key')}\": [I18n.t('general_error.params_missing_value', model: \"review\")])\n return\n else\n @whitelisted = @whitelisted.permit(:name, :uen, :description)\n end\n end",
"def special_device_list_params\n params.require(:special_device_list).permit(:name)\n end",
"def pull_request_params\n whitelist = [\n :url,\n :id,\n :html_url,\n :diff_url,\n :patch_url,\n :issue_url,\n :number,\n :state,\n :locked,\n :title\n ]\n params.require(:pull_request).permit(whitelist)\n end"
] |
[
"0.7121987",
"0.70541996",
"0.69483954",
"0.6902367",
"0.6733912",
"0.6717838",
"0.6687021",
"0.6676254",
"0.66612333",
"0.6555296",
"0.6527056",
"0.6456324",
"0.6450841",
"0.6450127",
"0.6447226",
"0.6434961",
"0.64121825",
"0.64121825",
"0.63913447",
"0.63804525",
"0.63804525",
"0.6373396",
"0.6360051",
"0.6355191",
"0.62856233",
"0.627813",
"0.62451434",
"0.6228103",
"0.6224965",
"0.6222941",
"0.6210244",
"0.62077755",
"0.61762565",
"0.61711127",
"0.6168448",
"0.6160164",
"0.61446255",
"0.6134175",
"0.6120522",
"0.6106709",
"0.60981655",
"0.6076113",
"0.60534036",
"0.60410434",
"0.6034582",
"0.6029977",
"0.6019861",
"0.6019158",
"0.6019158",
"0.6019158",
"0.6019158",
"0.6019158",
"0.6019158",
"0.6019158",
"0.6019158",
"0.6019158",
"0.6019158",
"0.6019158",
"0.6019158",
"0.6019158",
"0.6019158",
"0.6019158",
"0.6019158",
"0.6019158",
"0.60184896",
"0.60157263",
"0.6005857",
"0.6003803",
"0.60012573",
"0.59955895",
"0.5994598",
"0.5993604",
"0.5983824",
"0.5983166",
"0.5977431",
"0.597591",
"0.5968824",
"0.5965953",
"0.59647584",
"0.59647584",
"0.59566855",
"0.59506303",
"0.5950375",
"0.59485626",
"0.59440875",
"0.5930872",
"0.5930206",
"0.5925668",
"0.59235454",
"0.5917905",
"0.59164816",
"0.5913821",
"0.59128743",
"0.5906617",
"0.59053683",
"0.59052664",
"0.5901591",
"0.58987755",
"0.5897456",
"0.58970183",
"0.58942604"
] |
0.0
|
-1
|
The opposite of Arrayinclude i.e. return true if self appears in the array splatted into +first+ and +rest+.
|
def in? first, *rest
(first.is_a?(Array) ? first : [first].concat(rest)).include? self
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def my_include?(array, target)\n return false if array.empty?\n return true if array.first == target\n my_include?(array.drop(1), target)\nend",
"def includes?(array, target)\n if array.empty?\n false\n elsif array.first == target\n true\n else\n includes?(array.drop(1),target)\n end\nend",
"def include_any?(arr, arr2)\n #good for large sets w/ few matches\n # Set.new(self).intersection(arr).empty?\n arr2.any? {|e| arr.include?(e) }\n end",
"def includes?(array, target)\n return false if array.empty?\n return true if array.first == target\n includes?(array.drop(1), target)\nend",
"def includes?(array, target)\n return true if array.first == target\n return false if array.empty?\n includes?(array.drop(1), target)\nend",
"def includes?(array, target)\n return false if array.empty?\n return true if array.first == target\n includes?(array.drop(1), target)\nend",
"def includes?(array, target)\n return false if array.empty?\n return true if array[0] == target\n includes?(array.drop(1),target)\nend",
"def includes?(array, target)\n return false if array.empty?\n return true if array.first == target\n includes?(array[1..-1])\nend",
"def includes?(array, target)\n return true if array[0] == target\n return false if array[0] != target && array.length <= 1\n\n includes?(array.drop(1), target)\nend",
"def arr_includes_all?(arr, *inc)\n inc.map { |i| arr.include? i }.reduce(&:&)\n end",
"def includes?(array, target)\n return false if array.empty?\n \n return true if array[0] == target\n\n includes?(array[1..-1], target)\nend",
"def includes?(array, target)\n return false if array.empty?\n dup = array.dup\n return true if dup.pop == target\n includes?(dup, target)\nend",
"def include?(*args)\n args.inject(true) {|val, x| val = self.single_include?(x)}\n end",
"def includes?(array, target)\n return false if array.length <= 0\n return true if array[0] == target\n includes?(array[1..-1], target)\nend",
"def array_include(a1, a2) \n return (a1 - a2).empty?\n end",
"def includes?(array, target)\n \n if array.empty?\n return false\n end\n \n if array[-1] == target\n return true\n else\n return includes?(array[0...-1], target)\n end\n\nend",
"def array_includes?(array, given_thing)\n answer = false\n\n array.each do |thing|\n if thing == given_thing\n answer = true\n end\n end\n\n return answer\nend",
"def includes?(array, target)\n return false if array.empty?\n return true if array.pop == target\n includes?(array, target)\nend",
"def includes?(array, target)\n return false if array.empty?\n\n if array.pop == target\n return true\n else\n includes?(array, target)\n end\nend",
"def include_array?(array)\n array.any? { |member| array?(member) }\n end",
"def include_all?(*items)\n items = items.first if items.length == 1 && items.first.kind_of?(Array)\n (items - self).empty?\n end",
"def include?(arr, target)\n arr.any?{ |ele| ele == target} ? true : false\nend",
"def include?(array, target)\n matches = array.select { |element| element == target }\n !matches.empty?\nend",
"def include? array, item\n array.include?(item)\nend",
"def include?(*args_)\n _offset_for_args(args_) ? true : false\n end",
"def include?(array, value)\n array.any?(value)\nend",
"def using_include(array, element)\n\tarray.include?(element)\nend",
"def includes?(array, target)\n #empty array does not contain object\n return false if array == []\n\n #start with array[3]\n value = array.shift #pluck off\n if value == target\n true #the recursive call is not made\n else\n includes?(array, target)\n end\n\nend",
"def include?(ary, target)\n !ary.select { |value| value == target }.empty?\nend",
"def include?(array, arg)\n boolean_return = false\n array.each {|value| return boolean_return = true if value == arg}\n boolean_return\nend",
"def include?(array, query)\n array.each do |el|\n return true if el == query\n end\n false\nend",
"def exclude_all?(arr, arr2)\n ! include_any?(arr, arr2)\n end",
"def include?(arr, include_item)\n arr.each { |item| return true if item == include_item }\n false\nend",
"def include?(arr, search)\n arr.any? { |i| i == search }\nend",
"def include?(arr, search)\n arr.any? { |elem| elem == search }\nend",
"def include?(array, value)\n array.each { |element| return true if value == element }\n false\nend",
"def include?(arr, val)\n arr.each { |el| return el == val if el == val }\n false\nend",
"def arr_includes(arr1, arr2)\n temp_arr2 = arr2.dup\n return false if arr2.length == 0\n arr1.each do |num|\n if temp_arr2.include? num\n index = temp_arr2.index(num)\n temp_arr2.delete_at(index)\n else\n return false\n end\n end\n return true\nend",
"def include?(array, search)\n array.each do |value|\n return true if value == search\n end\n false\nend",
"def includes_all? *args\n args.all? { |arg| include? arg }\n end",
"def does_list_include?(array, obj)\n array.count(obj) > 0\nend",
"def include?(array, value)\n array.each do |elem|\n return true if elem == value\n end\n false\nend",
"def include?(arr, search)\n arr.each do |element|\n if element == search\n return true\n end\n end\n false\nend",
"def include? (array, search)\n array.each { |item| return true if search == item }\n return false\nend",
"def include?(arr, val)\n arr.any? { |element| element == val }\nend",
"def include?(array, value)\n array.each do |integer|\n return true if integer == value\n end\n return false\nend",
"def check_array(array,target)\n if array.include?(target)\n return true\n else\n return false\n end\nend",
"def include?(array, value)\r\n array.count(value) > 0\r\nend",
"def includ?(array, value)\n\tarray.each do |element|\n\t\tif element == value\n\t\t\treturn true\n\t\tend\n\tend\n\tfalse\nend",
"def include?(arr, value)\n !(arr.select { |element| element == value }).empty?\nend",
"def include?(array, value)\n # method A\n array.any? { |element| element == value }\n \n # method B\n # for item in array\n # return true if item == value\n # end\n \n # false\nend",
"def include?(element)\n @ary.include? element\n end",
"def include?(ary, value)\r\n ary.select{ |obj| obj == value } != []\r\nend",
"def include?(arr, search)\n result = false\n arr.each { |num| result = true if num == search }\n result\nend",
"def include?(array, search_value)\n if array.count(search_value) == 1\n return true\n else\n return false\n end\nend",
"def include?(o)\n @mut.synchronize{@array.include?(o)}\n end",
"def include?(other)\n `return self.indexOf(other) == -1 ? Qfalse : Qtrue;`\n end",
"def include?(other)\n `return self.indexOf(other) == -1 ? Qfalse : Qtrue;`\n end",
"def in(*list)\n if list.length == 1\n return list[0].include?(self)\n else\n return list.include?(self)\n end\n end",
"def in_array(array1, array2)\r\n \r\nend",
"def include?(array, search_element)\n array.each do |array_element|\n if array_element == search_element\n return true\n end\n false\n end\nend",
"def include?(array, search_element)\n array.each do |array_element|\n if array_element == search_element\n return true\n end\n end\n false\nend",
"def include?(arr, value)\n includes = false\n arr.each { |n| includes = true if n == value }\n includes\nend",
"def non_include?(*args)\n !include? *args\n end",
"def include?(array, integer)\r\n array.each do |element|\r\n return true if element == integer\r\n end\r\n\r\nfalse\r\nend",
"def b_include?(obj) # assumes the array is sorted\n range = bsearch_range { |x| x <=> obj } # returned range excludes its last element\n (self[range.first] == obj or self[range.last - 1] == obj)\n end",
"def include?(arr, num)\n arr.each{ |element| return true if element == num}\n return false\nend",
"def includes_all?(other)\n leftovers = other.dup\n leftovers.delete_elements!(self)\n leftovers.empty?\n end",
"def fully_includes?(other)\n overlaps?(other) && other == intersection(other)\n end",
"def include?(arr, search_value)\n new_arr = arr.select { |element| element == search_value}\n !new_arr.empty?\nend",
"def include?(array, search_value)\n !array.select{|element| element == search_value}.empty?\nend",
"def include?(ary, value)\n !!ary.index(value)\nend",
"def array_42(z)\n if z.include? 42\n return true\n else\n return false\n end\n\nend",
"def include?(array, value)\n !!array.find_index(value)\nend",
"def include?(array, value)\n !!array.find_index(value)\nend",
"def include?(arr, number)\n arr.any? { |num| num == number }\nend",
"def item_included?(str, arr)\n arr.each do |item|\n return true if item == str\n end\n false\nend",
"def include?(array, search_value)\n array.any? { |element| element == search_value }\nend",
"def orinclude?(haystack, needles)\n if haystack.nil? or needles.nil?\n return false\n end\n\n return (haystack & needles).any?\n end",
"def include?(array, search_value)\n array.select { |value| value == search_value }.empty? ? false : true\nend",
"def includes?(array, target)\n array.each do |k, v|\n if k == key\n return true\n elsif v.class.to_s == \"Array\"\n v.each do |inner_array|\n return has_key(inner_array, key)\n end\n else\n return false\n end\n end\n\nend",
"def include?(arr, search_value)\n arr.each do |num|\n if num == search_value\n return true\n end\n end\n false\nend",
"def contains_all?(other)\n (other & self) == self\n end",
"def include?(list,tst)\n list.each {|itm| return true if itm == tst}\n false\nend",
"def include?(arr, search_val)\n # arr.each { |e| return true if e == search_val }\n # false\n arr.count(search_val) > 0\n # or:\n # arr.any? { |i| i == val }\nend",
"def include?(array, search_value)\n array.count(search_value) > 0\nend",
"def is_item_in_array(array,item)\n return array.include?(item)\nend",
"def include?(array, search_value)\n # array.each do |element|\n # return true if element == search_value\n # end\n # false\n array.count(search_value) > 0\nend",
"def in_arr? (arr, text, exact)\n\tif exact\n\t\treturn arr.include? text\n\telse\n\t\treturn arr.any?{|s| s.include? text}\n\tend\nend",
"def require_array_match(arr)\n\t\tfilter_two_arrays(self, arr, true)\n\tend",
"def assert_in(first, second, message='', node=nil)\n assert Array === second && second.include?(first), \"expected #{second} to include '#{first}' #{message}\", node\n end",
"def include?(p0) end",
"def subset_of?( other_collection )\n all? {|e| other_collection.include? e }\n end",
"def include?(p0) end",
"def include?(p0) end",
"def include?(p0) end",
"def include?(p0) end",
"def include?(p0) end",
"def include?(low, high = nil)\n return @data.include?(low) if high.nil?\n (low..high).each { |i| return false if get(i).nil? }\n return true\n end",
"def include?(ary, number)\n return false if ary.size == 0\n result = ary.detect do |item|\n item == number\n end\n !!(result == number)\nend"
] |
[
"0.7352883",
"0.7275701",
"0.725362",
"0.7145677",
"0.712652",
"0.7106624",
"0.7047529",
"0.69749135",
"0.69622946",
"0.6862535",
"0.68409693",
"0.68224657",
"0.68108904",
"0.6761371",
"0.6692932",
"0.6681192",
"0.6680476",
"0.6668103",
"0.6651931",
"0.6624857",
"0.6623291",
"0.6609062",
"0.65859765",
"0.65799236",
"0.65679705",
"0.656661",
"0.6561546",
"0.6511988",
"0.6454108",
"0.6452048",
"0.6449058",
"0.6431395",
"0.6422954",
"0.63651687",
"0.6347993",
"0.63382465",
"0.63255244",
"0.6324486",
"0.63244104",
"0.63211614",
"0.63144064",
"0.63009465",
"0.62948155",
"0.6286314",
"0.6274213",
"0.6257483",
"0.62536263",
"0.6251657",
"0.6218265",
"0.6217975",
"0.62139964",
"0.6209562",
"0.6204791",
"0.6199325",
"0.6182145",
"0.6160799",
"0.61564976",
"0.61564976",
"0.6152346",
"0.6121406",
"0.6114804",
"0.6114509",
"0.6110232",
"0.6085886",
"0.6060123",
"0.60595447",
"0.6008873",
"0.60028636",
"0.59990376",
"0.5983802",
"0.596797",
"0.5965195",
"0.593825",
"0.5918562",
"0.5918562",
"0.5903378",
"0.5886717",
"0.5884071",
"0.58745146",
"0.5870764",
"0.5870608",
"0.5865493",
"0.5861609",
"0.5854356",
"0.5849865",
"0.58413565",
"0.5815602",
"0.57997715",
"0.5799621",
"0.57825017",
"0.57790077",
"0.57774484",
"0.5777313",
"0.5775846",
"0.5775846",
"0.5775846",
"0.5775846",
"0.5775846",
"0.57640916",
"0.57494843"
] |
0.8168852
|
0
|
Return this object's metaclass; i.e. the value of self within a 'class << self' block.
|
def metaclass
class << self; self end
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def metaclass\n class << self;\n self\n end\n end",
"def metaclass\n class << self;\n self\n end\n end",
"def metaclass\n class << self; self; end;\n end",
"def metaclass\n if !self.metaclass?\n raise NoMetaClass, self\n else\n class << self; self; end; \n end\n end",
"def metaclass\n class << self\n self\n end\n end",
"def metaclass\n class << self\n self\n end\n end",
"def metaclass\n class << self\n self\n end \n end",
"def metametaclass; self.metaclass.metaclass; end",
"def selfclass\n if defined?(self.class.selfclass)\n self.class.selfclass\n else\n eign = (class << self; self ; end)\n eign.selfclass\n end\n end",
"def class\n (class << self; self; end).superclass\n end",
"def get_class()\n l = get_type()\n #puts \"Type #{l.class} in #{self.class} , #{self}\"\n l.object_class()\n end",
"def get_class()\n l = get_type()\n #puts \"Type #{l.class} in #{self.class} , #{self}\"\n l.object_class()\n end",
"def class\n (class << self; self; end).superclass\n end",
"def meta_eval &blk; metaclass.instance_eval &blk; end",
"def meta_eval &blk; metaclass.instance_eval &blk; end",
"def meta_eval(&block)\n metaclass.instance_eval(&block)\n end",
"def meta_eval\n metaclass.instance_eval\n end",
"def sclass\n class << @obj; self; end\n end",
"def class_eval(*args, &block)\n metaclass.class_eval(*args, &block)\n end",
"def class_eval(*args, &block)\r\n metaclass.class_eval(*args, &block)\r\n end",
"def singleton_class\n class << self; self; end\n end",
"def singleton_class\r\n class<<self; self end\r\n end",
"def to_class\n\t\tObject.const_get(self)\n\t\t#ret = eval(self)\n\t\t#deb \"this is a class? cls=#{ret.class}\"\n\t\t#raise \"Exception its not a class!\" if ret.class.to_s != 'Class'\n\t\t#return ret\n\tend",
"def cloaking_class\n class << self\n self\n end\n end",
"def _singleton_class\n class << self\n self\n end\n end",
"def klass\n self.class\n end",
"def eigenclass\n class << self; self; end\n end",
"def klass\n class_name = self[\"_type\"]\n class_name ? class_name.constantize : nil\n end",
"def klass\n reflection.klass\n end",
"def klass\n reflection.klass\n end",
"def singleton_class(&block)\n\t\tif block_given?\n\t\t(class << self; self; end).class_eval(&block)\n\t\tself\n\t\telse\n\t\t(class << self; self; end)\n\t\tend\n\tend",
"def eigenclass\n class << self\n self\n end\n end",
"def class\n `$q(this.__native__.className)`\n end",
"def class_obj\n self.class\n end",
"def _class\n special_attribute('@class'.freeze) || self.class.name.demodulize\n end",
"def klass\n @klass ||= if self.cached && self.cached.is_a?(AridCache::CacheProxy::Result)\n self.cached.klass\n else\n object_base_class\n end\n end",
"def typed_self\n @typed_self ||= Object::const_get(self.type).new(id: self.id)\n end",
"def wrap_type(klass, &block)\n Class.new do\n include Mixins::Object\n include Mixins::Wrapper\n include Mixins::Castable\n include Mixins::OptionsStorage\n class_eval(&block) if block_given?\n end\n end",
"def wrapped_class\n @wrapped_class ||= name.sub(/Wrapper$/, \"\").constantize\n end",
"def class\n __getobj__.class\n end",
"def singleton_class\n\t\tclass << self; self; end\n\tend",
"def class_for(cls)\n return nil unless cls # superclass will return nil when none\n fqn = cls.qualified_type_name\n return classes[fqn] if classes[fqn]\n classes[fqn] = JClass.parse(self, cls)\n end",
"def klass\n @klass ||= compute_class(class_name)\n end",
"def klass\n @klass ||= compute_class(class_name)\n end",
"def klass\n @klass ||= compute_class(class_name)\n end",
"def type\n self.class\n end",
"def getRubyClass\n @ruby_class\n end",
"def meta_eval(&block)\n singleton_class.instance_eval(&block)\n end",
"def class() end",
"def class\n object.class\n end",
"def obj_singleton_class\n @cache[:obj_singleton] ||= begin\n class << obj #:nodoc:\n self\n end\n rescue\n nil\n end\n end",
"def target_singleton_class\n class << @obj; self; end\n end",
"def eigenclass\n class << self\n return self\n end\n end",
"def cls\n unless @cls or (not target.cls?)\n\n # it is a method?\n if target.cls.respond_to? :defs\n method = target.cls.defs.find do\n |method|\n method.name == message_name\n end\n if method\n return nil\n # @cls = method.cls #TODO\n end\n end\n\n # it is an instance attribute?\n if @cls.nil? and target.cls.respond_to? :attributes\n attr = target.cls.attributes.find do\n |attr|\n attr.name == message_name\n end\n if attr\n @cls = attr.cls\n end\n end\n end\n\n super\n end",
"def meta\n self.class.instance_variable_get(:@__meta)\n end",
"def singleton_class_eval &block\n self.singleton_class.class_eval &block\n end",
"def meta_eval string = nil, &block\n return metaclass.class_eval(string) if string\n metaclass.class_eval &block\n end",
"def orig_cls\n\t\treturn self.class.orig_cls\n\tend",
"def class\n @object.class\n end",
"def klass\n info.klass\n end",
"def +@\n self.class[ self ]\n end",
"def object_class\n self.class.object_class\n end",
"def __mock_class\n (class << self; self; end).superclass\n end",
"def item_class\n @item_class || self.superclass.get_item_class\n end",
"def inherit_type(klass, &block)\n Class.new(klass) do\n class_eval(&block) if block_given?\n end\n end",
"def get_class(name)\n # 'self' -- the class from which get_class is called.\n self.const_get(name)\n end",
"def class_type\n # -> uncomment the next line to manually enable rule tracing\n # trace_in( __method__, 23 )\n return_value = ClassTypeReturnValue.new\n\n # $rule.start = the first token seen before matching\n return_value.start = @input.look\n\n begin\n # at line 113:15: ( KIND | MIXIN | ACTION )\n if @input.peek( 1 ).between?( KIND, ACTION )\n @input.consume\n @state.error_recovery = false\n else\n mse = MismatchedSet( nil )\n raise mse\n end\n\n\n # - - - - - - - rule clean up - - - - - - - -\n return_value.stop = @input.look( -1 )\n\n rescue ANTLR3::Error::RecognitionError => re\n report_error(re)\n recover(re)\n\n ensure\n # -> uncomment the next line to manually enable rule tracing\n # trace_out( __method__, 23 )\n\n end\n \n return return_value\n end",
"def type\n self.class::TYPE\n end",
"def klass\n @klass ||= class_name.constantize\n end",
"def klass\n @klass ||= class_name.constantize\n end",
"def master_class(after_class)\n klass = single_class\n begin\n up = klass.superclass\n return klass if up == after_class\n end while klass = up\n return self.klass\n end",
"def base_class!\n @base_class = self\n end",
"def type_klass; end",
"def self_type; end",
"def self_type; end",
"def self_type; end",
"def ar_class(type = nil)\n return ActiveNode::Base.node_class(type).ar_class if type\n @ar_class ||= superclass.try(:ar_class) rescue nil\n end",
"def to_const\n [Module, Class].any? { |base| self.class == base } ? self : self.class\n end",
"def class_type\n # -> uncomment the next line to manually enable rule tracing\n # trace_in( __method__, 23 )\n return_value = ClassTypeReturnValue.new\n\n # $rule.start = the first token seen before matching\n return_value.start = @input.look\n\n begin\n # at line 139:15: ( 'kind' | 'mixin' | 'action' )\n if @input.peek(1).between?(T__40, T__42)\n @input.consume\n @state.error_recovery = false\n else\n mse = MismatchedSet(nil)\n raise mse\n end\n\n\n # - - - - - - - rule clean up - - - - - - - -\n return_value.stop = @input.look(-1)\n\n rescue ANTLR3::Error::RecognitionError => re\n report_error(re)\n recover(re)\n\n ensure\n # -> uncomment the next line to manually enable rule tracing\n # trace_out( __method__, 23 )\n\n end\n\n return return_value\n end",
"def class_info?; \"#{self.class.name}\" end",
"def className\n\t\tself.class.to_s\n\tend",
"def className\n\t\tself.class.to_s\n\tend",
"def create_class\n attributes = content['attributes']\n mod.const_set class_name, Class.new(parent_class) {\n attr_accessor(*attributes)\n # include Concerns::Extendable\n }\n end",
"def class; end",
"def class; end",
"def klass\n @object.name if (defined? @object.class.name)\n end",
"def __class__\n Rubinius.primitive :object_class\n raise PrimitiveFailure, \"Kernel#class primitive failed.\"\n end",
"def object_class\n decorator_class&.object_class || super\n end",
"def classes\n [self]\n end",
"def class_family\r\n begin\r\n result, klass = '', self\r\n\r\n begin\r\n klass = klass.class unless klass.instance_of?(Class)\r\n result << klass.to_s\r\n klass = klass.superclass\r\n result << \"<\" if klass\r\n end while klass\r\n\r\n result\r\n end\r\n end",
"def open_metaclass\n <<-CODE\n t1 = stack_pop();\n stack_push(object_metaclass(state, t1));\n CODE\n end",
"def type\n self.class.type(self)\n end",
"def inspect\n \"\\#<#{self.class}>\"\n end",
"def type\n return self.class.type # all instances of a class have the same type\n end",
"def object_class\n @object_class ||= begin\n oc = @store.find_class_named('Object') || add_class(RDoc::NormalClass, 'Object')\n oc.record_location self\n oc\n end\n end",
"def repeatable_thing_class\n base_type = attr_json_registration.type.base_type\n if base_type.is_a?(AttrJson::Model)\n base_type\n else\n # Will return a symbol name, confusingly\n base_type.type\n end\n end",
"def meta_classes\n META_CLASSES\n end",
"def className\r\n\t\tself.class.to_s\r\n\tend",
"def base_type\n self\n end",
"def type\n\t\tself.class.type\n\tend"
] |
[
"0.8166059",
"0.8166059",
"0.8026234",
"0.7975332",
"0.7844113",
"0.7844113",
"0.7740912",
"0.71944124",
"0.6931881",
"0.6778121",
"0.67406315",
"0.67406315",
"0.6682594",
"0.66146445",
"0.66146445",
"0.64811945",
"0.63630056",
"0.6283211",
"0.6247136",
"0.62076557",
"0.61530477",
"0.61237794",
"0.6091191",
"0.6073583",
"0.6038739",
"0.6034376",
"0.6031831",
"0.6027243",
"0.6017061",
"0.6017061",
"0.6012234",
"0.5998271",
"0.5968878",
"0.5949438",
"0.59421015",
"0.5928028",
"0.5925888",
"0.5904138",
"0.58689034",
"0.58559364",
"0.5828248",
"0.5813853",
"0.5813591",
"0.5813591",
"0.5813591",
"0.5784647",
"0.5766556",
"0.5765214",
"0.5756203",
"0.57408756",
"0.5725069",
"0.5720514",
"0.5716544",
"0.569536",
"0.56929374",
"0.5685466",
"0.5680884",
"0.56698394",
"0.5659392",
"0.56390655",
"0.56226236",
"0.5603523",
"0.55981964",
"0.559816",
"0.55848145",
"0.55604964",
"0.5558022",
"0.555736",
"0.5550812",
"0.5550812",
"0.55244863",
"0.5522802",
"0.5521785",
"0.54843974",
"0.54843974",
"0.54843974",
"0.5482734",
"0.5471811",
"0.5468219",
"0.54635465",
"0.5460635",
"0.5460635",
"0.54532874",
"0.5449218",
"0.5449218",
"0.54448056",
"0.5444564",
"0.5441998",
"0.5438005",
"0.5434326",
"0.54306084",
"0.541654",
"0.5407318",
"0.5403966",
"0.53956634",
"0.5395172",
"0.53944623",
"0.5388693",
"0.5377654",
"0.5365021"
] |
0.78683776
|
4
|
Return true if this object is +nil?+, or +empty?+ if it accepts that method.
|
def blank?
nil? || (respond_to?(:empty?) && empty?)
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def nil_or_empty?(obj)\n obj.nil? || (obj.respond_to?(:empty?) && obj.empty?)\n end",
"def empty?\n @value.nil? || (@value.respond_to?(:empty?) && @value.empty?)\n end",
"def blank?\n respond_to?(:empty?) ? !!empty? : !self\n end",
"def nil?\n\t\t\tempty?\n\t\tend",
"def blank?\n respond_to?(:empty?) ? empty? : !self\n end",
"def blank?\r\n nil? || (respond_to?(:empty?) && empty?)\r\n end",
"def object_blank?(obj)\n obj.respond_to?(:empty?) ? !!obj.empty? : !obj\n end",
"def check_nil_empty(object)\n return false if object.nil? || object.empty?\n true\n end",
"def nil_or_empty?(value)\n return true if value.nil? || (value.respond_to?(:empty?) && value.empty?)\n\n false\nend",
"def empty?\n none? { true }\n end",
"def is_empty?\n end",
"def blank?(obj)\n obj.respond_to?(:empty?) ? obj.empty? : !obj\n end",
"def empty?\n raise NotImplementedError\n end",
"def empty?\n @is_empty\n end",
"def __empty?\n return false unless super\n __empty(__value)\n end",
"def non_empty?(obj)\n !(obj.nil? || obj.empty?)\n end",
"def empty?\n @type == EMPTY_TYPE\n end",
"def empty?\n\t\treturn self.first.nil? ? true : false\n\tend",
"def empty?\r\n return @value == nil\r\n end",
"def blank_object?(obj)\n return obj.blank? if obj.respond_to?(:blank?)\n case obj\n when NilClass, FalseClass\n true\n when Numeric, TrueClass\n false\n when String\n obj.strip.empty?\n else\n obj.respond_to?(:empty?) ? obj.empty? : false\n end\n end",
"def __empty?\n all?(&:__empty?)\n end",
"def nil_or_empty_string?(obj)\n obj.nil? || obj.empty?\n end",
"def empty?\n `self.length === 0`\n end",
"def truthy?\n if respond_to? :blank?\n not blank?\n else\n not nil?\n end\n end",
"def truthy?\n if respond_to? :blank?\n not blank?\n else\n not nil?\n end\n end",
"def empty?\n empty\n end",
"def empty?\n end",
"def empty?\n end",
"def empty?\n end",
"def blank?\n self.nil? || self.empty?\n end",
"def blank?( object )\n return object.nil? || object.empty?\n end",
"def isEmpty?\n end",
"def empty?\n raise NotImplementedError, \"Method not implemented yet...\"\n end",
"def null?\n IsNull.new(self)\n end",
"def empty?\n raise NotImplementedError\n end",
"def empty?\n return self.length == 0\n end",
"def empty?\n @value.nil? || @value.empty?\n end",
"def null?\n if( @length == 0 )\n return true\n else\n return false\n end\n end",
"def empty?\n self <=> EMPTY\n end",
"def empty?\n self.length == 0\n end",
"def empty?\n return false\n end",
"def empty?\n if !block_given?\n return @j_del.java_method(:isEmpty, []).call()\n end\n raise ArgumentError, \"Invalid arguments when calling empty?()\"\n end",
"def empty?\n end",
"def empty?\n end",
"def empty?\n end",
"def empty?\n end",
"def empty?\n @data[:object].nil?\n end",
"def respond_to?(method, *args)\n return true if self.links && self.links.respond_to?(method)\n return true if self.attributes && self.attributes.respond_to?(method)\n return true if self.objects && self.objects.respond_to?(method)\n super\n end",
"def empty?\n true\n end",
"def empty?\n true\n end",
"def empty?\n\t\t\t\t\tfalse\n\t\t\t\tend",
"def is_empty?\n self.size == 0\n end",
"def empty?() end",
"def empty?() end",
"def empty?() end",
"def valid? object\n self.call(object).empty?\n end",
"def any?\n `self.length > 0`\n end",
"def none_empty?\n !any_empty?\n end",
"def none_empty?\n !any_empty?\n end",
"def empty?\n @input.empty?\n end",
"def empty?()\n return @length == 0\n end",
"def present?(object)\n return false if object.nil?\n return false if object == false\n return false if object.respond_to?(:empty?) && object.empty?\n true\n end",
"def nil_or_empty?(*values)\n values.any? { |v| v.nil? || (v.respond_to?(:empty?) && v.empty?) }\n end",
"def empty?\n self.length == 0\n end",
"def empty?\n false\n end",
"def empty?\n return self.length == 0\n end",
"def empty?\n return item.nil?\n end",
"def empty?\n\t\t@length == 0\n\tend",
"def empty?()\n return @first.nil?\n end",
"def nil_or_empty?(value)\n value.nil? || value.empty? ? true : false\n end",
"def respond_to?(meth)\n super(meth) || @value.respond_to?(meth)\n end",
"def empty?\n false\n end",
"def empty?\n value.empty?\n end",
"def empty?\n\t\treturn self.fields.empty?\n\tend",
"def blank?\n return true if @target.nil?\n @target.respond_to?(:empty?) ? @target.empty? : !@target\n end",
"def empty?\n return false if @opts[:values]\n super\n end",
"def empty?\n @val==nil\n end",
"def empty?\n super\n end",
"def empty?\n super\n end",
"def empty?\n @length == 0\n end",
"def empty?\n length.zero?\n end",
"def any?\n ! empty?\n end",
"def empty?\n @length == 0\n end",
"def empty?\n true\n end",
"def empty?\n true\n end",
"def empty?\n return get.first.nil?\n end",
"def empty?\r\n @length == 0\r\n end",
"def empty?\r\n @length == 0\r\n end",
"def empty?\r\n @length == 0\r\n end",
"def empty?\r\n @length == 0\r\n end",
"def empty?\r\n @length == 0\r\n end",
"def empty?\n false\n end",
"def empty?\n false\n end",
"def empty?\n value.empty?\n end",
"def empty?\n # raise NotImplementedError, \"Not yet implemented\"\n if size > 0\n return false\n end\n return true\n end",
"def empty?\n length == 0\n end",
"def empty?\n length == 0\n end",
"def empty?\n length == 0\n end"
] |
[
"0.7699884",
"0.71426743",
"0.7137497",
"0.7127501",
"0.7082801",
"0.7003134",
"0.6990346",
"0.69731843",
"0.697255",
"0.6909335",
"0.68669826",
"0.6866432",
"0.68478334",
"0.6824168",
"0.68132603",
"0.6797004",
"0.6787903",
"0.6774785",
"0.67682517",
"0.66865164",
"0.66799414",
"0.667294",
"0.6672024",
"0.6661146",
"0.6661146",
"0.664401",
"0.66420287",
"0.66420287",
"0.66420287",
"0.6638088",
"0.66344553",
"0.6634232",
"0.66319233",
"0.66171837",
"0.660633",
"0.66039056",
"0.6595856",
"0.65918225",
"0.6578969",
"0.6569767",
"0.6558283",
"0.65486884",
"0.6545441",
"0.6545441",
"0.6545441",
"0.6545441",
"0.65398276",
"0.6538166",
"0.65380514",
"0.65380514",
"0.6531402",
"0.65244526",
"0.65203804",
"0.65203804",
"0.65203804",
"0.65062225",
"0.6480393",
"0.64732796",
"0.64732796",
"0.64732796",
"0.6469829",
"0.6467834",
"0.64552915",
"0.6433735",
"0.64327645",
"0.64146173",
"0.6410904",
"0.6406296",
"0.6403868",
"0.63974077",
"0.63960123",
"0.63945234",
"0.63820785",
"0.6380112",
"0.63758385",
"0.6373182",
"0.6372691",
"0.63650095",
"0.63650095",
"0.6359102",
"0.63450223",
"0.6340233",
"0.63395447",
"0.6338072",
"0.6338072",
"0.63377327",
"0.6337085",
"0.6334169",
"0.6334169",
"0.6334169",
"0.6334169",
"0.6324888",
"0.6324888",
"0.6315922",
"0.63079613",
"0.6305429",
"0.6305429",
"0.6305429"
] |
0.6962996
|
11
|
Return self unmodified after logging the output of inspect, along with the point at which +tapp+ was called.
|
def tapp
tap { puts "#{File.basename caller[4]}: #{self.inspect}" }
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def tapp\n Babushka::LogHelpers.deprecated! '2017-09-01'\n tap { STDOUT.puts \"#{File.basename caller[2]}: #{self.inspect}\" }\n end",
"def inspect!\n $stderr.puts(self.inspect)\n end",
"def inspect!\n $stderr.puts(self.inspect)\n end",
"def get\n super\n .tap { |result| __ext_debug(\"--> #{result.inspect}\") }\n end",
"def taph\n tap {\n puts \"<pre>\" +\n \"#{File.basename caller[2]}: #{self.inspect}\".gsub('&', '&').gsub('<', '<') +\n \"</pre>\"\n }\n end",
"def backtrace\n wrapper.application_trace\n end",
"def inspect\n redacted_string(:inspect)\n end",
"def inspect\n redacted_string(:inspect)\n end",
"def tracing()\n #This is a stub, used for indexing\n end",
"def inspect\n \"#<#{self.class}: #{@orig.inspect}>\"\n end",
"def inspect\n \"#<#{self.class}: #{@orig.inspect}>\"\n end",
"def inspect\n \"#<#{self.class}: #{@orig.inspect}>\"\n end",
"def run\n print_debug 'In #run'\n audit self.class.payloads, self.class.options\n end",
"def t\n self\n end",
"def underscorize\n dup.tap(&:underscorize!)\n end",
"def delegate\n @app.call(@env)\n end",
"def inspect\n reload\n @target.inspect\n end",
"def assertion\n super\n .tap { |result| __ext_debug(\"--> #{result.inspect}\") }\n end",
"def inspect!\n warn(inspect)\n end",
"def inspect!\n warn(inspect)\n end",
"def log_state\n super\n end",
"def trace!\n request! :trace\n end",
"def initialize( * )\n\t\tself.class.dump_application_stack\n\t\tsuper\n\tend",
"def initialize(app, &block)\n\n @app = app\n\n set Footprint::Log::Basic, STDOUT\n\n @app.class.send(:define_method, :logger,\n Proc.new do\n @env[:footprint_logger]\n end)\n\n self.instance_eval &block if block\n end",
"def trace\n DSL.new(@source.append(Trace.new))\n end",
"def inspect\n \"#<#{self.class} #{super}>\"\n end",
"def log\n self.class.log\n end",
"def inspect\n \"#<#{self.class} #{self}>\"\n end",
"def call env\n env[\"rack.logger\"] = logger\n trace_id = extract_trace_id(env)\n logger.add_trace_id trace_id\n\n begin\n @app.call env\n ensure\n logger.delete_trace_id\n end\n end",
"def app\n @app\n end",
"def backtrace\n if original.nil?\n super\n else\n original.backtrace\n end\n end",
"def inspect\n original_inspect = super\n original_inspect.split( ' ' ).first << '>'\n end",
"def inspect\n original_inspect = super\n original_inspect.split( ' ' ).first << '>'\n end",
"def inspect\n original_inspect = super\n original_inspect.split( ' ' ).first << '>'\n end",
"def __trace_impl(*args, **opt, &block)\n opt[:no_log] = true unless opt.key?(:no_log)\n __output_impl(*args, **opt, &block)\nend",
"def initialize(app, tracer: OpenTracing.global_tracer)\n @app = app\n @tracer = tracer\n end",
"def logme\n\t\tlogger.debug \"============Main ApplicationController get calls at #{Time.now}\"\n\tend",
"def inspect\n\t\t\told_to_s\n\t\tend",
"def entrapment\n # setup trap chain once.\n Base.trapped = true\n app = self\n\n # 'get' previous trap by replacing any existing trap with 'IGNORE'\n previous_trap = trap(:INT, 'IGNORE')\n\n # substitute our trap and chain it to previous by explicitly invoking\n # the previous trap. ruby makes this somewhat difficult and rack then\n # makes it even harder.\n trap(:INT) do\n begin\n # loggers may have closed file handles in a trap so disconnect any\n # loggers from multiplexer before continuing. even when they do not\n # raise exceptions they still appear to log nothing at this point\n # (not sure about syslog, definitely not file or console).\n if app.logger.respond_to?(:targets)\n # HACK: it is bad that Multiplexer#targets exposes its internal\n # array in a manner that allows us to clear it. it would be better\n # if to have a Multiplexer#reset method we could call instead.\n # to ensure that cleaning continues to work, check the result\n # afterward. note that we tried iterating targets and calling the\n # Multiplexer#remove method but that had no effect.\n app.logger.targets.clear\n fail 'Unexpected targets' unless app.logger.targets.empty?\n app.logger.warn('cannot log traps') # no exception raised\n end\n\n # interrupt any running app threads to resolve outstanding requests.\n #\n # note that Mutex#synchronize is not allowed inside a trap context.\n #\n # FIX: duplicating the set is slightly unsafe but not sure how else\n # to deal with data protected by critical section in a trap. we also\n # have logic in ensure block to avoid modifying set on interrupt.\n app.class.interrupted = true\n app_threads = app.class.app_threads.dup\n app_threads.each do |app_thread|\n if app_thread.alive?\n app_thread.raise(::Interrupt)\n app_thread.join\n end\n end\n\n # cleanup fixtures, if requested.\n app.cleanup\n if previous_trap && previous_trap.respond_to?(:call)\n previous_trap.call\n else\n exit\n end\n rescue ::Exception => e\n # loggers are unreliable so write any rescued error home.\n msg = ([e.class, e.message] + (e.backtrace || [])).join(\"\\n\")\n dir = ::ENV['HOME'] || ::Dir.pwd\n path = ::File.join(dir, 'might_api_rescued_error.txt')\n ::File.open(path, 'w') { |f| f.puts msg }\n exit 1\n end\n end\n true\n end",
"def inspect\n return \"#<#{self.class.name}: #{self}>\"\n end",
"def inspect\n \"<Application name=#{@name} id=#{@id}>\"\n end",
"def inspect\n # Concise to not dump too much information on the dev\n \"#<#{self.class.name}>\"\n end",
"def app; @app; end",
"def investigation\n self\n end",
"def app\n @app ||= begin\n lock!\n to_app\n end\n end",
"def run\n p self\n end",
"def inspect\n target.inspect\n end",
"def inspect() end",
"def inspect() end",
"def inspect() end",
"def inspect() end",
"def inspect() end",
"def inspect() end",
"def inspect() end",
"def inspect() end",
"def inspect() end",
"def inspect() end",
"def inspect() end",
"def inspect() end",
"def inspect() end",
"def inspect() end",
"def re_stdout\n @logdev.re_stdout\n self\n end",
"def initialize app, options = {}\n @app = app\n @except = Array(options[:except]).map { |e| regexpify e }\n @log = options[:log]\n @only = Array(options[:only]).map { |o| regexpify o }\n @path = options[:path] || \"/\"\n\n if FalseClass === @log\n @log = lambda { |m| }\n elsif !@log\n if defined? Rails\n @log = lambda { |m| Rails.logger.info \" [Onesie] #{m}\" }\n else\n @log = lambda { |m| puts \"[Onesie] #{m}\" }\n end\n end\n end",
"def run( result )\n\t\t\t$stderr.puts self.name if $VERBOSE || $DEBUG\n\n\t\t\t# Support debugging for individual tests\n\t\t\tolddb = nil\n\t\t\tif $DebugPattern && $DebugPattern =~ @method_name\n\t\t\t\tolddb = $DEBUG\n\t\t\t\t$DEBUG = true\n\t\t\tend\n\t\t\t\n\t\t\tsuper\n\n\t\t\t$DEBUG = olddb unless olddb.nil?\n\t\tend",
"def app\n @app\n end",
"def trace\n set_trace_func proc { |event, _file, _line, id, binding, classname|\n if event == watched && id != :log && classes.include?(classname.to_s)\n vars = variables(binding)\n\n if vars.empty?\n log_this(sprintf(\"%s %-25s #%-20s\",\n class_count, classname, id))\n\n else\n log_this(sprintf(\"%s %-25s #%-20s\\n%s\\n\",\n class_count, classname, id, vars))\n\n end\n end\n } if trace?\n end",
"def run(app)\n raise '@run already set' if @run\n\n @run = app\n end",
"def call(env)\n\t\tbefore = Time.now.to_i\n\t\t# Use parallel assignment\n\t\tstatus, headers, body = @app.call(env)\n\t\tafter = Time.now.to_i\n\t\t# The message to log\n\t\tlog_message = \"App took #{after - before} seconds.\"\n\t\t\n\t\t# Append the log message to the body and return the expected array of three elements\n\t\t[status, headers, body << log_message]\n\tend",
"def top()\n @view__.top\n end",
"def log\n\t\t\t@log_proxy ||= ClassNameProxy.new( self.class )\n\t\tend",
"def log\n\t\t\t@log_proxy ||= ClassNameProxy.new( self.class )\n\t\tend",
"def log\n\t\t\t@log_proxy ||= ClassNameProxy.new( self.class )\n\t\tend",
"def log\n\t\t\t@log_proxy ||= ClassNameProxy.new( self.class )\n\t\tend",
"def log\n\t\t\t@log_proxy ||= ClassNameProxy.new( self.class )\n\t\tend",
"def application\n @application || SemanticLogger.application\n end",
"def application_instance\n target\n end",
"def inspect\n \"#<#{self.class.inspect} @args=#{args.inspect}\"\n end",
"def application\n self\n end",
"def inspect\n \"#<#{self.class}>\"\n end",
"def inspect\n \"#<#{self.class}>\"\n end",
"def inspect\n \"#<#{self.class}>\"\n end",
"def insp\r\n self.inspect\r\n end",
"def call(env)\n env[:footprint_logger] = @logger\n @app.call env\n end",
"def trace(*args)\n Trace.(self, *args)\n end",
"def trace(*args)\n Trace.(self, *args)\n end",
"def inspect\n \"(#{self})\"\n end",
"def app\n @app || configure_middleware!\n end",
"def call(*args, &block)\n call_with_tracing(args[0]) { super }\n end",
"def log; end",
"def log; end",
"def log; end",
"def log; end",
"def log; end",
"def log; end",
"def log; end",
"def log; end",
"def inspect\n \"\\#<#{self.class}>\"\n end",
"def logger\n @app.logger\n end",
"def run\n start\n trace\n wrapup\n end",
"def inspect\n \"#{self.class}(#{to_s.inspect})\"\n end"
] |
[
"0.6483636",
"0.5369693",
"0.5369693",
"0.52905303",
"0.52770275",
"0.5134703",
"0.50808525",
"0.50808525",
"0.502376",
"0.49886099",
"0.49886099",
"0.49886099",
"0.4965967",
"0.49399868",
"0.49362803",
"0.48896167",
"0.4859325",
"0.48477146",
"0.4812033",
"0.4812033",
"0.47991428",
"0.47985652",
"0.4776972",
"0.47702518",
"0.4765173",
"0.4764623",
"0.47523513",
"0.4732488",
"0.47278988",
"0.47167528",
"0.47152638",
"0.47139648",
"0.47139648",
"0.47139648",
"0.47026584",
"0.47012523",
"0.46979982",
"0.46971637",
"0.46827707",
"0.46693316",
"0.46675217",
"0.4658326",
"0.46581537",
"0.46514142",
"0.46461824",
"0.46441442",
"0.46417987",
"0.4621384",
"0.4621384",
"0.4621384",
"0.4621384",
"0.4621384",
"0.4621384",
"0.4621384",
"0.4621384",
"0.4621384",
"0.4621384",
"0.4621384",
"0.4621384",
"0.4621384",
"0.4621384",
"0.46156934",
"0.45977986",
"0.45951924",
"0.45907867",
"0.45853993",
"0.45839742",
"0.45702258",
"0.45664257",
"0.456181",
"0.456181",
"0.456181",
"0.456181",
"0.456181",
"0.4552695",
"0.4550835",
"0.4537381",
"0.4531992",
"0.45303595",
"0.45303595",
"0.45303595",
"0.45205227",
"0.45190197",
"0.45172253",
"0.45172253",
"0.45163003",
"0.45161748",
"0.45150477",
"0.4512292",
"0.4512292",
"0.4512292",
"0.4512292",
"0.4512292",
"0.4512292",
"0.4512292",
"0.4512292",
"0.45089865",
"0.45084482",
"0.44956318",
"0.4490403"
] |
0.6342416
|
1
|
Log and return unmodified in the same manner as tapp, but escape the output to be HTML safe and easily readable. For example, becomes <Object:0x00000100bda208>
|
def taph
tap {
puts "<pre>" +
"#{File.basename caller[2]}: #{self.inspect}".gsub('&', '&').gsub('<', '<') +
"</pre>"
}
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def debug(object)\n puts \"<pre>\"\n puts object.pretty_inspect.gsub('<', '<').gsub('>', '>')\n puts \"</pre>\"\nend",
"def ru_debug(object)\n begin\n \"<pre class='debug_dump'>#{object.to_yaml}</pre> <p><em>Raw dump</em></p>\".html_safe\n rescue Exception => e # errors from Marshal or YAML\n # Object couldn't be dumped, perhaps because of singleton methods -- this is the fallback\n \"<code class='debug_dump'>#{object.inspect}</code>\".html_safe\n end\n end",
"def h obj; obj.to_s.html_escape end",
"def echo obj\n str = obj.to_s\n write str\n end",
"def dumpAsHTML\n\t\tvalue = dumpAsString\n\t\treturn CGI.escapeHTML(value)\n\tend",
"def escape_output(data)\n (data && defined?(ERB::Util.h) && Rabl.configuration.escape_all_output) ? ERB::Util.h(data) : data\n end",
"def log\n @log_output.string\n end",
"def to_s\n \"#{@text}: #{@context}\"\n end",
"def h(obj)\n CGI.escapeHTML(obj.to_s)\n end",
"def to_s\n render().join \"\\n\"\n end",
"def inspect\n redacted_string(:inspect)\n end",
"def inspect\n redacted_string(:inspect)\n end",
"def format_body_for_log_output(obj)\n output = ''\n if obj.content_type == 'application/json'\n if @log_pretty_print_body\n _body = obj.body\n output << \"\\n\"\n output << JSON.pretty_generate(JSON.parse(_body)) rescue _body\n return output\n else\n return obj.body\n end\n else\n return obj.body.inspect\n end\n end",
"def render_to_string(*args, &block); end",
"def format_body_for_log_output(obj)\n #obj.body.inspect\n output = ''\n if obj.content_type == 'application/json'\n if @log_pretty_print_body\n output << \"\\n\"\n output << JSON.pretty_generate(JSON.parse(obj.body))\n return output\n else\n return obj.body\n end\n else\n return obj.body.inspect\n end\n end",
"def inspect\n s = self.to_s\n s = s[0,s.length-1]\n display_hooks = hooks.dup\n display_hooks.each do |k,v|\n display_hooks.delete(k) if v.empty?\n end\n unless display_hooks.empty?\n s << \" hooks=#{display_hooks.inspect}\"\n end\n unless requirements.empty?\n s << \" requirements=#{requirements.inspect}\"\n end\n s << \" targets=#{targets.map(&:to_sym).inspect}\" if targets\n s << \" origins=#{origins.map(&:to_sym).inspect}\" if origins\n s << \">\"\n s\n end",
"def to_string\n render().join \"\\n\"\n end",
"def render_plain\n short_category(object.action) << \"\\n\" << instance_attributes\n end",
"def escape(str)\n str.dump\n end",
"def inspect\n original_inspect = super\n original_inspect.split( ' ' ).first << '>'\n end",
"def inspect\n original_inspect = super\n original_inspect.split( ' ' ).first << '>'\n end",
"def inspect\n original_inspect = super\n original_inspect.split( ' ' ).first << '>'\n end",
"def html_escape\n return to_s\n end",
"def html_escape\n return to_s\n end",
"def inspect\n id = format('%x', object_id << 1)\n \"#<#{self.class.name}:0x#{id.rjust(14, '0')} 💻=“#{to_esc(true)}███#{CLEAR_TERM}” ✎=“\\\\e[#{to_esc(false)}m” 🗔=“#{to_html}”>\"\n end",
"def out(*args); to_s(*args); end",
"def log(text)\n puts text\n end",
"def debug_out()\n return JSON.pretty_generate(self).gsub(\"\\\": \", \"\\\" => \").gsub(\" => null\", \" => nil\")\n end",
"def canonical_log logevent\n write(format(logevent))\n end",
"def h obj\n CGI.escapeHTML obj.to_s\n end",
"def format_body_for_log_output(obj)\n if obj.content_type == 'application/json'\n if @log_pretty_print_body\n _body = obj.body\n output = JSON.pretty_generate(JSON.parse(_body)) rescue _body\n return output\n else\n return obj.body\n end\n elsif obj.content_type == 'application/xml'\n return obj.body\n else\n return obj.body.inspect\n end\n end",
"def format_body_for_log_output(obj)\n if obj.content_type == 'application/json'\n if @log_pretty_print_body\n _body = obj.body\n output = JSON.pretty_generate(JSON.parse(_body)) rescue _body\n return output\n else\n return obj.body\n end\n elsif obj.content_type == 'application/xml'\n return obj.body\n else\n return obj.body.inspect\n end\n end",
"def to_s\n \"<%s name=\\\"%s\\\">\" % [self.class.name.sub(%r/^Logging::/, ''), self.name]\n end",
"def re_stdout\n @logdev.re_stdout\n self\n end",
"def dump_content_to_file_debug_text( content )\n\n return content unless logger.level == Logger::DEBUG\n\n outname = \"#{@name}.debug.text\"\n\n puts \" Dumping content before text-to-html conversion to #{outname}...\"\n\n File.open( outname, 'w' ) do |f|\n f.write( content )\t\n end\n\n content\n\nend",
"def to_s\n output || content || \"\"\n end",
"def inspect\n hex_address = (object_id << 1).to_s(16).rjust(14, \"0\")\n \"#<Lita::RackApp:0x#{hex_address}>\"\n end",
"def fancy_log\n\n @log.collect { |msg| fancy_print(msg) }\n end",
"def to_html_msg(str)\n str.gsub!(\"<stdin>:\", \"\") # Output from the compiler\n str.gsub!(\"<\", \"<\")\n str.gsub!(\">\", \">\")\n str.gsub!(/[\\n|\\r\\n]/, '<br\\>') # Replace newlines with line breaks \n str\n end",
"def log thing\n `console.orig_log(#{thing})`\n end",
"def result_to_string(object)\n if object.is_a?(String) then\n return object\n else\n if defined?(awesome_inspect)\n return object.awesome_inspect({:plain=>true, :index=>false})\n elsif defined?(pretty_inspect)\n return object.pretty_inspect.chomp # Remove new line that PrettyInspect adds at the end https://www.ruby-forum.com/topic/113429\n else\n return object.inspect\n end\n end\n end",
"def render(*args, &block)\n captured = helpers.capture do\n helpers.concat(helpers.render(*args, &block))\n helpers.output_buffer.to_s\n end\n rawtext(captured)\n end",
"def dumpAsString\n\t\tmarkup = \"\"\n\t\t@Value.each do |line|\n\t\t\tmarkup += \"#{line}<br />\\n\"\n\t\tend\n\t\treturn markup\n\tend",
"def inspect\n\t\t\treturn \"#<%p:%#0x %s>\" % [\n\t\t\t\tself.class,\n\t\t\t\tself.object_id * 2,\n\t\t\t\tself.inspect_details,\n\t\t\t]\n\t\tend",
"def render_to_string(context, properties = T.unsafe(nil)); end",
"def tapp\n Babushka::LogHelpers.deprecated! '2017-09-01'\n tap { STDOUT.puts \"#{File.basename caller[2]}: #{self.inspect}\" }\n end",
"def to_s\n return \"log(#{self.arg.to_s})\"\n end",
"def inspect\n\t\treturn \"#<%p:%#016x %s>\" % [\n\t\t\tself.class,\n\t\t\tself.object_id * 2,\n\t\t\tself.inspect_details,\n\t\t]\n\tend",
"def render\n inspect\n end",
"def raw_output(s)\n s\n end",
"def normal(string)\n to_console loggify(string)\n to_file loggify(string)\n end",
"def debug_output=(arg); end",
"def to_s\r\n pretty_output\r\n end",
"def to_s\n Utils::Escape.html(@content)\n end",
"def render_text(text)\n @out << text\n end",
"def write_tex_on(op, value_hash)\n esc = proc do |str|\n str.\n gsub(/</, '<').\n gsub(/>/, '>').\n gsub(/&/) { '\\\\&' }.\n gsub(/([$&%\\#{}_])/) { \"\\\\#$1\" }.\n gsub(/>/, '$>$').\n gsub(/</, '$<$')\n end\n str = \"\"\n \n str << write_common(value_hash, esc)\n $stderr.puts str.inspect\n op << str\n end",
"def render\n sb = StringIO.new\n append_to_string_builder(sb)\n sb.string\n end",
"def textLog _args\n \"textLog _args;\" \n end",
"def dprint(obj)\n self.log_puts(Php4r.print_r(obj, true))\n end",
"def echo(e)\n return \"#{e}\"\nend",
"def to_s\n redacted_string(:to_s)\n end",
"def to_s\n redacted_string(:to_s)\n end",
"def inspect\n\t\treturn \"#<%p:%#016x formatter: %s, level: %s, output: %s>\" % [\n\t\t\tself.class,\n\t\t\tself.object_id * 2,\n\t\t\tself.settings[:formatter] || '-',\n\t\t\tself.settings[:level] || '-',\n\t\t\tself.settings[:logdev] ? self.settings[:logdev].class : '-',\n\t\t]\n\tend",
"def logger_output; end",
"def inspect\n inspectables = self.class.inspectables\n if inspectables\n \"#<#{self.class}:0x#{object_id.to_s(16)} \" + inspectables.map {|i| \"@#{i}=\\\"#{send(i) rescue nil}\\\"\"}.join(' ') + \">\"\n else\n super\n end\n end",
"def inspect\n inspectables = self.class.inspectables\n if inspectables\n \"#<#{self.class}:0x#{object_id.to_s(16)} \" + inspectables.map {|i| \"@#{i}=\\\"#{send(i) rescue nil}\\\"\"}.join(' ') + \">\"\n else\n super\n end\n end",
"def to_s\n \"#<#{self.class.name}:#{object_id} #{info}>\"\n end",
"def to_html\n map { |x| x.output(\"\") }.join\n end",
"def output(o);printf o;end",
"def to_s\n \"[#{@type}]#{@content}\"\n end",
"def raw\n repr.raw\n end",
"def debug(thing, tag = 'pre')\n %Q{<#{tag}>#{thing.to_yaml}</#{tag}>}\n end",
"def <<(msg) debug(msg); end",
"def log_to_stdout\n log_to STDOUT\n end",
"def html_safe\n self\n end",
"def audit(data)\n stream << data.to_s.chomp + \"\\n\"\n stream.flush\n end",
"def output_log(log)\n puts \"\\e[32m#{log}\\e[m\\s\"\nend",
"def dump(object)\n internal = ::Riak::Serializers.serialize(content_type, object)\n encrypt(internal)\n end",
"def log(message)\n STDOUT.puts message\nend",
"def dump\r\n super + to_s\r\n end",
"def embed_action_as_string(options) #:doc:\n embedded_logging(options) do\n response_for_embeded_action = embedded_response(options, false)\n\n if redirected = response_for_embeded_action.redirected_to\n embed_action_as_string(redirected)\n else\n response_for_embeded_action.body\n end\n end\n end",
"def capture(*args)\n super.html_safe\n end",
"def to_s\r\n dump\r\n end",
"def log(s)\n #puts s\nend",
"def dump(obj, io)\n io.puts obj.send(inspect_method)\n end",
"def to_s()\n prettify()\n end",
"def _print_out data\n\t\tresponse << \"Someone said: #{data}\"\t\t\t\n\tend",
"def inspect\n \"#<#{self.class.name}:0x#{(self.__id__ * 2).to_s(16)} \" +\n \"@host=#{host.to_s}, \" +\n \"@log_level=#{log_level}, \" +\n \"@log_location=#{log_location.inspect}>\"\n end",
"def escape_html\n Rack::Utils.escape_html self\n end",
"def to_s\n \"#<Middleman::SourceWatcher:0x#{object_id} type=#{@type.inspect} directory=#{@directory.inspect}>\"\n end",
"def log str \n if $vips_debug\n puts str\n end\nend",
"def pretty(object)\n PP.pp(object, out)\n end",
"def log_sometext(str)\n log str\n end",
"def dumpAsHTML\n\t\tmarkup = \"\"\n\t\t@LoadString.each_line do |line|\n\t\t\tmarkup += \"#{CGI.escapeHTML(line.chomp)}<br />\\n\"\n\t\tend\n\t\treturn markup\n\tend",
"def to_log\n # Pretty - ain't it? :D\n \"\\n\" << voucher_rows.reduce(\"#{pretty_id} - #{title}\nDatum: #{I18n.l accounting_date.to_date}\nNämnd: #{organ}\nUnderlag från: #{material_from}\nBokfört av: #{bookkept_by}\nAttesterat av: #{authorized_by_to_s}\n -----\n \") do |acc, vr|\n acc << \"#{vr.to_log}\n \"\n end\n end",
"def text(text)\n text = text.gsub(/['\\\\]/, '\\\\\\\\\\&')\n \"@output_buffer.safe_concat('#{text}'); \"\n end",
"def to_s\n Array(render).map(&:to_s).join(\"\\n\")\n end",
"def p(obj)\n @contents << obj.inspect << \"\\n\"\n end",
"def to_s\n Alf::Renderer.text(self).execute(\"\")\n end",
"def to_s\n to_html\n end"
] |
[
"0.6483437",
"0.62061334",
"0.6087314",
"0.594068",
"0.58671564",
"0.5843327",
"0.5815329",
"0.57354784",
"0.5728294",
"0.5641781",
"0.563286",
"0.563286",
"0.56151956",
"0.56121206",
"0.5595415",
"0.55726415",
"0.55717295",
"0.5569038",
"0.5550239",
"0.5534598",
"0.5534598",
"0.5534598",
"0.5503541",
"0.5503541",
"0.5499175",
"0.54958135",
"0.54931664",
"0.5475328",
"0.5474885",
"0.5471879",
"0.5458912",
"0.5458808",
"0.54395",
"0.54308784",
"0.5429362",
"0.5410821",
"0.53943986",
"0.5390263",
"0.5382554",
"0.5354138",
"0.53319097",
"0.53201985",
"0.5320038",
"0.5319483",
"0.5313413",
"0.5306715",
"0.5300363",
"0.52920693",
"0.52677274",
"0.5262338",
"0.52616817",
"0.5259941",
"0.5259226",
"0.5258694",
"0.52550006",
"0.52536863",
"0.5243846",
"0.524362",
"0.5234482",
"0.5233488",
"0.5231333",
"0.5231333",
"0.5225445",
"0.5209505",
"0.5209267",
"0.5209267",
"0.5208625",
"0.52061176",
"0.52034473",
"0.52028126",
"0.51976967",
"0.5195907",
"0.5193556",
"0.51894945",
"0.51886916",
"0.51883656",
"0.5184913",
"0.5179941",
"0.51780015",
"0.5167992",
"0.51673836",
"0.51672864",
"0.5167248",
"0.51668006",
"0.5163358",
"0.5157675",
"0.5152377",
"0.51432",
"0.5139035",
"0.513516",
"0.5135152",
"0.5131645",
"0.5131622",
"0.51296306",
"0.51260024",
"0.51256937",
"0.5124195",
"0.512418",
"0.5120093",
"0.51181275"
] |
0.62805504
|
1
|
Executes the given block within the context of the receiver. In order to set the context, the variable self is set to this object while the block is executing, giving the code access to this object's instance variables. Arguments are passed as block parameters. This is a fallback implementation for older rubies that don't have a builtin instance_exec.
|
def instance_exec(*args, &block)
begin
old_critical, Thread.critical = Thread.critical, true
n = 0
n += 1 while respond_to?(mname="__instance_exec#{n}")
InstanceExecHelper.module_eval{ define_method(mname, &block) }
ensure
Thread.critical = old_critical
end
begin
ret = send(mname, *args)
ensure
InstanceExecHelper.module_eval{ remove_method(mname) } rescue nil
end
ret
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def _exec block, *args\n instance_exec(*args, &block)\n end",
"def run(context = nil, *args)\n if context\n context.instance_exec(*args, &@block)\n else\n @block.call(*args)\n end\n end",
"def evaluate(&block)\n if block.arity > 0\n yield self\n else\n instance_exec(&block)\n end\n end",
"def evaluate(block, *args)\n return if block.nil?\n if block.is_a? Symbol\n instance_exec(self, *args, &block)\n else\n instance_exec(*args, &block)\n end\n end",
"def __execute(&block)\n self.instance_eval(&block)\n end",
"def run(context, options = {})\n #@block.call(options)\n\n if @options[:defaults]\n options = @options[:defaults].merge(options)\n end\n\n context.instance_exec(options, &@block)\n end",
"def initialize(*args, &block)\n super(*args)\n instance_exec(_context, &block)\n end",
"def run\n @ctx.call(self,&@blk) if @blk\n end",
"def evaluate(&block)\n self.original_self = eval \"self\", block.binding\n instance_eval &block\n end",
"def _run(context)\n return unless matches_context?(context)\n\n if @run_block\n puts \"[#{context.host}] Executing \\\"#{self.name}\\\"...\"\n @run_block.call(context)\n end\n end",
"def initialize (&block)\n instance_exec(&block)\n end",
"def call\n if @block\n @block.arity < 1 ? self.instance_eval(&@block) : @block.call(self)\n end\n self\n end",
"def call\n @block.arity < 1 ? self.instance_eval(&@block) : @block.call(self) if @block\n self\n end",
"def run_stored_block\n self.run_in_context @stored_block if @stored_block\n end",
"def call(context, *args)\n context.instance_exec(*args, &callback)\n end",
"def call(&block)\n instance_eval(&block)\n end",
"def call(&block)\n instance_eval(&block)\n end",
"def call(&block)\n instance_eval(&block)\n end",
"def initialize(&block)\n instance_exec(&block) if block_given?\n end",
"def execute_block(block)\n case\n when block.is_a?(Symbol)\n send(block)\n when block.respond_to?(:call) && (block.arity == 1 || block.arity == -1)\n block.call(self)\n else\n block\n end \n end",
"def execute\n yield self if block_given?\n end",
"def call(*args)\n instance_exec(*correct_arg_arity(block.arity, args), &block)\n end",
"def parent &block\n unless @_iv.context\n raise 'Host scope is not available'\n end\n\n if block\n @_iv.context.instance_exec(&block)\n else\n @_iv.context\n end\n end",
"def exec (*args, &block)\n if @scope\n @scope.instance_variables.each do |name|\n instance_variable_set(name, @scope.instance_variable_get(name))\n end\n end\n\n @ivars.each do |name, value|\n instance_variable_set(name, value)\n end\n \n instance_exec(*args, &block)\n end",
"def execute\n if @block\n instance_eval(&@block)\n else\n raise MRubyTestSkip, \"(Not implemented)\"\n end\n end",
"def run\n if @block.arity >= 1\n @block.call self\n else\n @block.call\n end\n end",
"def execute( context )\n @instance_method.bind(context).call(*@args)\n end",
"def evaluate\n context = Context.new(self, node)\n context.instance_exec(&block)\n node\n end",
"def call(&block)\n block.call(self) if block_given?\n self\n end",
"def and_yield(*args)\n yield eval_context = Object.new if block_given?\n @plan = Proc.new do |&block|\n eval_context.instance_exec(*args, &block)\n end\n self\n end",
"def evaluate(&block)\n if block.arity.zero?\n instance_eval(&block)\n else\n yield(self)\n end\n end",
"def call(&block)\n execute(&block) if valid?\n execution_attempted!\n self\n end",
"def context(&block); end",
"def context(&block); end",
"def execute &block\n @executor = block\n end",
"def run_in_context(o={}, &block)\n proc = Proc.new do\n set_vars_from_options(:name => o[:name]) if o.has_key?(:name) # Name MUST be set first\n set_vars_from_options(o)\n instance_eval &block if block\n end\n super(&proc)\n end",
"def in_context(context_type, &block)\n context_type.new.instance_eval(&block)\n end",
"def call_block\n @_block.call(self) if @_block\n end",
"def with(&block)\n raise ArgumentError.new(\"Must pass a block\") unless block_given?\n case block.arity\n when 1\n yield self\n when 0\n instance_methods_eval(&block)\n else\n raise \"block arity must be 0 or 1\"\n end\n end",
"def call\n context.run(self) do\n #subject.run_setup(scope) if subject\n scope.instance_exec(*arguments, &procedure)\n #subject.run_teardown(scope) if subject\n end\n end",
"def in_the(instance, &block)\n instance.instance_eval(&block)\n end",
"def execution_context\n initialize_context!(Object.new)\n end",
"def call(*args)\n @ctx.instance_exec(args, &action)\n end",
"def run(*args,&ruby_block)\n # Gain access to the accesser as local variable.\n runner_proc = @runner_proc\n # Execute the code generating the accesser in context.\n HDLRuby::High.space_push(@namespace)\n HDLRuby::High.cur_block.open do\n instance_exec(ruby_block,*args,&runner_proc)\n end\n HDLRuby::High.space_pop\n end",
"def evaluate_sifter(scope, *args, &block)\n evaluate scope do |root|\n root.instance_exec(*args, &block)\n end\n end",
"def run(*args,&ruby_block)\n # Gain access to the runner as local variable.\n runner_proc = @runner_proc\n # Execute the code generating the accesser in context.\n HDLRuby::High.space_push(@namespace)\n HDLRuby::High.cur_block.open do\n instance_exec(ruby_block,*args,&runner_proc)\n end\n HDLRuby::High.space_pop\n end",
"def execute(&block)\n\tblock.call\nend",
"def execute(&block)\n\tblock.call\nend",
"def execute(&block)\n\tblock.call\nend",
"def execute(&block)\n\tblock.call\nend",
"def run(&blk)\n raise(\"A block is needed to run\") unless block_given?\n @run_logic = blk\n end",
"def execute(&block)\n block.call\n end",
"def execute(&block)\n block.call\n end",
"def use_block(obj, &method)\n\tmethod.call\nend",
"def eval\n @block.eval\n end",
"def resolve(&blk)\n @context = blk.binding.receiver\n instance_exec(&blk)\n end",
"def client(&block)\n eval('self', block.binding) if block_given?\n end",
"def client(&block)\n eval('self', block.binding)\n end",
"def instance_exec(*args, &block)\n begin\n old_critical, Thread.critical = Thread.critical, true\n n = 0\n n += 1 while respond_to?(mname=\"__instance_exec#{n}\")\n InstanceExecHelper.module_eval{ define_method(mname, &block) }\n ensure\n Thread.critical = old_critical\n end\n begin\n ret = send(mname, *args)\n ensure\n InstanceExecHelper.module_eval{ remove_method(mname) } rescue nil\n end\n ret\n end",
"def execute(block)\n model = @wrapper.object\n block.call model\n end",
"def instance_exec(*args, &block)\n begin\n old_critical, Thread.critical = Thread.critical, true\n n = 0\n n += 1 while respond_to?(method_name = \"__instance_exec#{n}\")\n InstanceExecMethods.module_eval { define_method(method_name, &block) }\n ensure\n Thread.critical = old_critical\n end\n\n begin\n send(method_name, *args)\n ensure\n InstanceExecMethods.module_eval { remove_method(method_name) } rescue nil\n end\n end",
"def instance_exec(*args, &block)\n begin\n old_critical, Thread.critical = Thread.critical, true\n n = 0\n n += 1 while respond_to?(method_name = \"__instance_exec#{n}\")\n InstanceExecMethods.module_eval { define_method(method_name, &block) }\n ensure\n Thread.critical = old_critical\n end\n\n begin\n send(method_name, *args)\n ensure\n InstanceExecMethods.module_eval { remove_method(method_name) } rescue nil\n end\n end",
"def context &block\n self.worker.context(&block) \n end",
"def execute(&block)\n\tblock\nend",
"def execute(&block)\n\tblock\nend",
"def with_context(&block)\n Fiber.current.hammer_context = self\n block.call\n Fiber.current.hammer_context = nil\n end",
"def execute(&block)\n define_method(:execute, &block)\n end",
"def instance_eval(*args, &block)\n file, line = __find_caller__(caller())\n $__gen_code__[[file, line]] = {\n :binding => self, \n :kind => :instance_eval, \n :code => args[0]\n }\n if block_given?\n $__instance_eval__.bind(self).call(&block)\n else\n $__instance_eval__.bind(self).call(*args)\n end\n end",
"def execute(block)\n block.call\nend",
"def execute(block)\n block.call\nend",
"def execute(block)\n block.call\nend",
"def execute(block)\n block.call\nend",
"def execute(block)\n block.call\nend",
"def execute(block)\n block.call\nend",
"def execute(block)\n block.call\nend",
"def call(*args, &block)\n @block.call *args, &block\n end",
"def execute(&block)\n block\n #puts block #This shows the object I think.\nend",
"def with( &blk )\n # blk[ self ]\n self.instance_eval &blk\n end",
"def execute(&block)\n # block\n block.call\nend",
"def execute(&block)\n block.call\nend",
"def call_this(&block)\n block.call\nend",
"def __execute__\n @result = @block[*@args]\n rescue Exception => e\n @result = e\n ensure\n # Prevent multiple executions\n def self.__execute__; nil; end\n end",
"def call(*args)\n block.call(*args)\n end",
"def safely(*args, &block)\n previous_state = @rye_safe\n enable_safe_mode\n ret = self.instance_exec *args, &block\n @rye_safe = previous_state\n ret\n end",
"def call(object, context = T.unsafe(nil), *args, &block); end",
"def execute(&block)\n block # needs .call method here\nend",
"def execute(&block)\r\n block\r\nend",
"def call(*args)\n in_context do\n self.execute(*args)\n end\n end",
"def run_block\n if @block\n _block = @block\n @block = nil\n instance_eval &_block\n true\n end\n end",
"def roby_execute(&block)\n execution_engine.execute(&block)\n end",
"def execute(&block)\n block.call \nend",
"def execute(&block)\n block.call\nend",
"def execute(&block)\n block.call\nend",
"def execute(&block)\n block.call\nend",
"def execute(&block)\n block.call\nend",
"def execute(&block)\n block.call\nend",
"def execute(&block)\n block.call\nend",
"def execute(&block)\n block.call\nend",
"def execute(&block)\n block.call\nend",
"def execute(&block)\n block.call\nend"
] |
[
"0.8038417",
"0.7608923",
"0.7133825",
"0.7016939",
"0.701285",
"0.699832",
"0.6978153",
"0.69703704",
"0.6927305",
"0.6897472",
"0.6874533",
"0.68489146",
"0.6776979",
"0.67618835",
"0.6744641",
"0.6741976",
"0.6741976",
"0.6741976",
"0.6737922",
"0.6699209",
"0.6692339",
"0.6662941",
"0.6650166",
"0.6616998",
"0.65756184",
"0.65437496",
"0.65082127",
"0.64799404",
"0.64712864",
"0.6456709",
"0.6437796",
"0.64261144",
"0.6407621",
"0.6407621",
"0.63942736",
"0.6392979",
"0.63801146",
"0.63642085",
"0.6329312",
"0.62999135",
"0.62759894",
"0.62595963",
"0.62369794",
"0.62260455",
"0.6225773",
"0.6223504",
"0.6203124",
"0.6203124",
"0.6203124",
"0.6202053",
"0.6190298",
"0.6185524",
"0.6185524",
"0.6184562",
"0.61751854",
"0.6172913",
"0.6169618",
"0.6159178",
"0.6140442",
"0.61251485",
"0.6118876",
"0.6118876",
"0.6114198",
"0.61011446",
"0.61011446",
"0.6093686",
"0.6088671",
"0.6077414",
"0.60701936",
"0.60701936",
"0.60701936",
"0.60701936",
"0.60701936",
"0.60701936",
"0.60701936",
"0.606233",
"0.6060643",
"0.60504663",
"0.60456634",
"0.6038858",
"0.6038692",
"0.603736",
"0.6016652",
"0.60149044",
"0.6003434",
"0.59839463",
"0.59772456",
"0.59711695",
"0.5955666",
"0.59511113",
"0.5943795",
"0.59389305",
"0.59389305",
"0.59389305",
"0.59389305",
"0.59389305",
"0.59389305",
"0.59389305",
"0.59389305",
"0.59389305"
] |
0.63130426
|
39
|
This method save data for better view and test (useful for having an overview of regexp matches)
|
def save( directory )
File.open( "parsed_data/#{directory}.yaml", 'w' ) do |f|
f.write( @pages.to_yaml )
end
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def save_additional_data\n end",
"def save_match(match_data)\n begin\n save_match_summary(match_data)\n save_player_stats(match_data)\n rescue\n end\n end",
"def record_data_for_later_saving(data, definition)\n word = data[:word]\n word_index = data[:index]\n\n # Add current definition's see_also list to map for processing later\n @relateds[definition] = data[:see_also]\n\n # Add definition to map of word obj. -> definition objects.\n # This mapping will be used to link all related definitions after they've all been created\n @word_definitions[word.downcase] ||= []\n @word_definitions[word.downcase][word_index] = definition\n end",
"def store_text(key, process, options)\n match_regex?(key, datas[key], options[:regex]) if(options.has_key?(:regex))\n store(key, process, options)\n end",
"def do_save_logic(record)\n # After validation,\n # I either scrape a record.parent.frgtxt or a record.rpt.uurl or record.inputurl\n case\n when record.valid? == false\n record.frgtxt = \"<b>The Data In This Fragment Is Invalid.</b>\"\n\n # 1st branch (the most common) is for when user specifies parent.frgtxt as input\n when record.parent != nil\n # parent is shorthand for record.parent.frgtxt\n # parent is easy for end-user to remember\n parent = record.parent.frgtxt\n\n # Now that I have the input, branch on the expression type\n case record.exprtype.name\n\n # peel-off-outer-tag returns parent with outer tag peeled off\n when \"peel-off-outer-tag\"\n record.frgtxt = peel_off_outer_tag(record)\n\n # glue-fragment-to-parent allows user to glue any fragment to end of parent fragment\n when \"glue-fragment-to-parent\"\n record.frgtxt = glue_fragment_to_parent(record)\n\n # string-sandwich allows user to sandwich parent between 2 strings\n when \"string-sandwich\"\n record.frgtxt = string_sandwich(record)\n\n # sub allows user to do this: parent.sub(/change this/,'to this')\n when \"sub\"\n record.frgtxt = mysub(record)\n\n # gsub allows user to do this: parent.gsub(/change this/,'to this')\n when \"gsub\"\n record.frgtxt = mygsub(record)\n\n # This allows: parent =~ /regexp/\n # Then returns data from matching-array\n when \"regexp-enumerable\"\n record.frgtxt = regexp_enumerable(record)\n\n # This branch works well on simple Hpricot scrape expressions like:\n # body or div#some-id or\n # table or a[@href*='yahoo.com'] or\n # span.some-class[text()*='Text I want to match']\n when \"hpricot\"\n record.frgtxt = Hpricot(record.parent.frgtxt).search(record.scrapeexpr).to_html\n\n # This branch might be the most useful.\n # It displays a set of enumerable objects.\n # Each is wrapped in a numbered-div-tag.\n # I want the user to enter a scrapeexpr like this:\n # table,[4,3] or\n # body div>table a,[0..999]\n when \"hpricot-enumerable\"\n record.frgtxt = get_my_hp_enum(record.parent.frgtxt, record.scrapeexpr)\n end # case record.exprtype.name\n\n\n # Branch here when user specifies EDGAR report as input\n when record.rpt != nil\n\n # Now that I have the input, branch on the expression type\n case record.exprtype.name\n when \"hpricot\"\n record.frgtxt = get_my_hp_elem(record.rpt.uurl).search(record.scrapeexpr).to_html\n when \"hpricot-enumerable\"\n record.frgtxt = get_my_hp_enum(get_my_html_from_open_uri(record.rpt.uurl), record.scrapeexpr)\n end # record.exprtype.name\n\n # Branch here when user specifies arbitrary URL as input\n # This is probably rare but is useful for testing purposes.\n when record.inputurl != nil\n\n # Now that I have the input, branch on the expression type\n case record.exprtype.name\n when \"hpricot\"\n record.frgtxt = get_my_hp_elem(record.inputurl).search(record.scrapeexpr).to_html\n when \"hpricot-enumerable\"\n record.frgtxt = get_my_hp_enum(get_my_html_from_open_uri(record.inputurl), record.scrapeexpr)\n end # record.exprtype.name\n\n end # The outer-case handles type of input: parent, rpt, inputurl\n\n end",
"def save_data(data)\n\tend",
"def saveListToFile()\n\t\t# f is going to equal data.txt with the 'write' capability:\n\t\tf = File.new('data.txt', 'w')\n\n\t\t# write searchSuggestionList to data.txt:\n\t\tf.write(\"#{@searchSuggestionList}\")\n\n\t\t# close data.txt/ end writing:\n\t\tf.close\n\tend",
"def save_files\n @law_h = [:law, :link, :section]\n @case_h = [:case, :link, :date]\n\n # Parse into hashes\n @single_matches = @single_matches-@double_matches\n case_hash = @double_matches.map{|i| {@case_h[0] => i[0], @case_h[1] => i[1], @case_h[2] => i[2]}}\n law_hash = @single_matches.map{|i| {@law_h[0] => i[0], @law_h[1] => i[1], @law_h[2] => i[2]}}\n \n # Write files\n write_files(case_hash, law_hash)\n end",
"def stored_data; end",
"def save_information(keyword, page)\n\t\t\tkeyword.save_adword_urls(page)\n\t\t\tkeyword.save_nonadword_urls(page)\n \tend",
"def save_cached_data\n # open file for writing\n @file_handle = File.open(file_path, 'w')\n \n # write data string into file\n @file_handle << (@file_data_to_write.respond_to?(:read) ? @file_data_to_write.read : @file_data_to_write)\n \n # close file\n close_file\n \n end",
"def save_cached_data\n # open file for writing\n @file_handle = File.open(file_path, 'w')\n \n # write data string into file\n @file_handle << (@file_data_to_write.respond_to?(:read) ? @file_data_to_write.read : @file_data_to_write)\n \n # close file\n close_file\n \n end",
"def store; end",
"def store; end",
"def store; end",
"def save_row_data(data)\n word = data[:word]\n\n # Attempt to retrieve Word obj & save to DB\n word_obj = @word_objs[word]\n\n # Create new word obj. if not found\n unless word_obj\n # TODO: Could use 'new' instead of 'create' here and batch save at end\n word_obj = Word.where(text: word).first_or_create\n @word_objs[word] = word_obj\n puts \"Error with word '#{word}': #{word_obj.errors.full_messages}\" if word_obj.errors.present?\n end\n\n # TODO: Could use 'new' instead of 'create' here and batch save at end\n # Create definition obj & save\n\n existing_defs = word_obj.definitions\n def_index = data[:index]\n definition = existing_defs[def_index] if existing_defs.size >= (def_index - 1)\n definition = existing_defs.create if definition.nil?\n definition.update(discussion: data[:discussion], text: data[:definition])\n\n # Report any errors with save\n puts \"Error with definition '#{word.downcase}': #{definition.errors.full_messages}\" if definition.errors.present?\n\n # Create alt_spellings\n data[:alt_spellings].each do |alt|\n alt_obj = definition.alt_spellings.where(text: alt).first_or_create\n puts \"Error with alternate spelling '#{alt}': #{alt_obj.errors.full_messages}\" if alt_obj.errors.present?\n end\n\n # Save source data\n @source_saver.save_sources(definition, data[:sources], @bibliography_data)\n\n # Some data can't be saved now since it relies on all definitions and words being saved first\n # Add this data to the relevant instance vars. for later\n record_data_for_later_saving data, definition\n end",
"def save(data)\n self.results[:processed] += 1\n\n # Create an observation from the data\n f = self.observation_from_data(data)\n\n new_record = f.new_record?\n changed = f.changed?\n\n if not changed\n self.results[:skipped] += 1\n return f\n else\n # changes = f.changes\n # changes.each do |i, data|\n # data[0] = data[0].to_s\n # data[1] = data[1].to_s\n # end\n # Rails.logger.debug changes\n end\n\n if not f.save\n Rails.logger.debug \"Could not save observation. Errors: #{f.errors.messages}\"\n self.results[:failed] += 1\n return f\n end\n\n if new_record\n self.results[:created] += 1\n elsif changed\n self.results[:updated] += 1\n end\n\n return f\n end",
"def save\n\t\titemtemp1 = Institutehandler.find(institutehandler_id)\n\t\titemtemp = itemtemp1.name+level+course+stream\n\t\titem = itemtemp.gsub(/[^0-9A-Za-z]/, '')\n \tself.uniquenesscheck = item.delete(' ')\n \tsuper # calls the rails save function to store our object to the database\n \tend",
"def get_data\n self.refer_count = self.refer_count + 1\n save\n return {:data => self.data, :code => self.shark_code, :similar => self.similar_data}\n end",
"def save_with_all_properties data\n @heart_rate_array = []\n self.laps = []\n self.distance_m = 0\n self.sport = data['Activities']['Activity']['Sport']\n self.activity_id = data['Activities']['Activity']['Id']\n self[:start_time] ||= DateTime.parse(data['Activities']['Activity']['Id'])\n if data['Activities']['Activity']['Lap'].class == Array\n data['Activities']['Activity']['Lap'].each do |lap|\n parse_lap lap\n end\n else\n parse_lap data['Activities']['Activity']['Lap']\n end\n self.avg_heart_rate = @heart_rate_array.sum / @heart_rate_array.count\n self.trimp = calculate_trimp\n end",
"def create_save\n @save_data = {:turns => @turns,:guesses => @guesses,:secret_word => @secret_word, :hidden_word => @hidden_word}\n save = File.new(\"./lib/save.txt\", \"w+\")\n save.puts JSON::dump(save_data)\n save.close\n end",
"def on_before_save\n\t\ts = \"#{self.make} #{self.model} #{self.year} #{self.trim} #{self.body} #{self.drive} #{self.transmission}\"\n\t\tputs s\n\t\tself.searchable = s\n\tend",
"def store_new_match(players, options, game_name)\r\n @info_match = {}\r\n pl_names = []\r\n players.each {|pl| pl_names << pl.name}\r\n @info_match[:players] = pl_names\r\n @info_match[:date] = Time.now\r\n @info_match[:giocate] = [ ] # items like { :deck => [], :first_plx => 0, :actions => [] }\r\n @info_match[:game] = {:name => \"#{game_name}\", :opt => {}\r\n }\r\n # set options of the match\r\n options.each do |k,v|\r\n @info_match[:game][:opt][k] = v\r\n end\r\n \r\n end",
"def configure_save\n save = load_save\n @word = save.word\n @hints = save.hints\n @incorrect = save.incorrect\n @lives = save.lives\n end",
"def save_processed_data\n attachment.update(processed_data: json_parser.final_hash)\n end",
"def save_guess\n if good_guess?\n matching_indexes(@code_word, @word_guesser.guess).each do |i|\n @good_guesses[i] = @word_guesser.guess\n end\n else\n @bad_guesses.push(@word_guesser.guess)\n end\n end",
"def make_save_contents\n\t\t\tcontents = local_make_save_contents\n\t\t\tcontents[:self_vars] = $game_selfVars\n\t\t\tcontents\n\t\tend",
"def persist(data)\n\t\tupdate_design_for self.shared_data_context\n\t\t\n\t\td = CouchRest::Document.new\n\t\td.database = proxy_database\n\n\t\t#copy data in to document\n\t\tdata.each do |key,value|\n\t\t\td[\"#{key}\"] = value\n\t\t\tputs \"added: #{key}\"\n\t\tend\n\t\td[:type] = self.shared_data_context.name\n\t\td[:date] = DateTime.now\n\n\t\tbegin\n\t\t\td.save false\n\t\trescue => e\n\t\t\tputs \"NOT SAVED: #{e.response}\"\n\t\t\t#TODO: properly handle this stuff\n\t\tend#\n\tend",
"def save(data)\n raise \"Implement this in subclass\"\n end",
"def save_direct_matches\n @@direct_matches.each do |i|\n keep_matches[i] = @@player_key[i]\n end\n end",
"def create_save\n @save_data = {:turns => @turns,:guesses => @guesses,:secret_word => @secret_word, :hidden_word => @hidden_word}\n save = File.new(\"save.txt\", \"w+\")\n save.puts JSON::dump(save_data)\n save.close\n end",
"def wiki_page_scrap_and_save_to_mongo\n page_infos = wiki_page_scrap\n\n #PUT THE COLLECTED TEXT INTO A MONGO_DB\n mongo = get_mongo\n collection = mongo[$mongo_wiki_scrap_collection_name]\n\n #REMOVE OLD CACHE DATA\n collection.remove\n #PUT NEW SCRAP DATA TO MONGO\n collection.insert page_infos\nend",
"def matchdata\n @matchdata\n end",
"def make_regexp\n @intent = self.intent\n regexp = self.pattern.dup.downcase\n words = regexp.split(\" \")\n words.each do |word|\n if word.include? '/'\n regexp = regexp.gsub(word,\"(#{word})\")\n\n end\n\n end\n regexp = regexp.gsub('/',\"|\")\n regexp = regexp.gsub('^ ','.{0,60}').gsub(' ^','.{0,60}').gsub(' *','.{1,60}').gsub('* ','.{1,60}').gsub('^','.{1,60}').gsub(' [','.{0,60}[')\n regexp = regexp.gsub(' .{0,60}','.{0,60}')\n regexp = regexp.gsub(' .{1,60}','.{1,60}')\n regexp = '.{0,60}' + regexp + '.{0,60}'\n self.regexp = regexp\n chunks = self.pattern.split(' ')\n chunks.each do |ch|\n result= Regexp.new(/\\[.{0,12}\\]/) =~ ch\n if(result==0)\n set = WordSet.find_by_keyword(ch[1..-2])\n str = '(' + set.words.join('|') + ')'\n regexp = self.regexp.gsub(ch,str)\n self.regexp = regexp\n end\n end\n self.save\n end",
"def save\n domain_info_location = @domain_location + FILENAME\n text = \"== %s\\n\\n\" % Time.now\n text << @record.map{|key, val| \"- %s: %s\" % [key,val]}.join(\"\\n\")\n text << \"\\n\\n\"\n if domain_info_location.exist?\n domain_info_location.append(text)\n else\n domain_info_location.create(text)\n end\n end",
"def save\n @data.map(&:save)\n end",
"def save!\n # Scrub some fields\n @data[\"installed\"].sort!\n @data[\"installed\"].uniq!\n\n # Save\n @path.open(\"w+\") do |f|\n f.write(JSON.dump(@data))\n end\n end",
"def save\n\t\tNg2::HashDb.add(@word, serialize)\n\tend",
"def save_unmatched_words # :nodoc:\n tokens = phrase_without_matches.split(' ')\n unmatched_db = Corpus.new(\"unmatched-#{program_name}.db\")\n tokens.each do |token|\n if !complex_token_matches?(token) # token was not transformed earlier\n @to_match << token\n unmatched_db[token] = @processor.original_text\n end\n end\n unmatched_db.close\n end",
"def save\n result = nil\n # iterate over each instance variable and insert create row to table\n\t\t\t\t obj = self.inst_strip_braces(self.object)\n self.instance_variables.each do |method|\n method = method.to_s.gsub(/@/,\"\")\n # Don't save objects with braces to database\n val = self.send(method.to_sym)\n # add rows excluding object, source_id and update_type\n unless self.method_name_reserved?(method) or val.nil?\n result = Rhom::RhomDbAdapter::insert_into_table(Rhom::TABLE_NAME,\n {\"source_id\"=>self.get_inst_source_id,\n \"object\"=>obj,\n \"attrib\"=>method,\n \"value\"=>val,\n \"update_type\"=>'create'})\n end\n end\n # Create a temporary query record to display in the list\n Rho::RhoConfig::sources[self.class.name.to_s]['attribs'].each do |attrib|\n result = Rhom::RhomDbAdapter::insert_into_table(Rhom::TABLE_NAME,\n {\"source_id\"=>self.get_inst_source_id,\n \"object\"=>obj,\n \"attrib\"=>attrib['attrib'],\n \"value\"=>self.send(attrib['attrib'].to_sym),\n \"update_type\"=>'query'})\n end\n result\n end",
"def store\n @store ||= JapaneseNames::Util::Kernel.deep_freeze(\n File.open(filepath, 'r:utf-8').each_with_object({}) do |line, hash|\n ary = line.chop.split('|')\n hash[ary[0]] ||= []\n hash[ary[0]] << ary\n end\n )\n end",
"def make_save_contents\n contents = rm_extender_make_save_contents\n contents[:self_vars] = $game_self_vars\n contents[:labels] = $game_labels\n contents[:self_labels] = $game_self_labels\n contents\n end",
"def pstore_save\n # Les métadonnées\n store_metadata\n # Scenes\n store_data :scene\n # Personnages\n store_data :personnage\n # Décors\n store_data :decor\n # Brins\n store_data :brin\n # Notes\n store_data :note\n end",
"def extract_data_by_pattern(regexp, message_pattern)\n if match = regexp.match(raw_message) then\n # Creates a hash maching :field_x => data\n new_data = Hash[match.names.zip(match.captures)].symbolize_keys\n\n new_data[:timestamp] = DateTime.strptime new_data[:timestamp], \n message_pattern.source.timestamp_definition\n new_data[:timestamp] += log.time_bias.seconds\n\n self.attributes = new_data\n self.message_pattern = message_pattern\n self.save\n end\n end",
"def save\n unless saved?\n # TODO: decide whether or not I should perform validations like this at\n # save-time, or only read-time. They can be quite slow.\n self.class.verify_type(hash, type)\n self.class.verify_length(hash, length, dump)\n self.class.verify_hash(hash, type, dump)\n \n contents = OUTPUT_FORMAT % [ type, length, dump ]\n \n self.class.write(hash, contents)\n end\n \n self\n end",
"def update_search_data\n self.search_data = [self.title, self.author, self.filename].join('|')\n .downcase\n .squeeze(' ')\n .gsub(/[^a-z\\s|]/, '')\n end",
"def save\n raise \"Unhashed entry during save\" unless @hash\n raise \"Entry without mime type during save\" unless @mimetype\n @data[:mimetype] = @mimetype\n @data[:ctime] = @ctime.utc.to_i\n @data[:mtime] = @mtime.utc.to_i\n @data[:name] = @name\n @data[:size] = @size\n @data[:tags] = @tags\n @data[:locations] ||= Array.new\n unless @data[:locations].include? @location\n @data[:locations] << @location\n end\n if ::Ordnung.database.read @hash\n ::Ordnung.database.update @hash, @data\n else\n ::Ordnung.database.create @hash, @data\n end\n @hash\n end",
"def validate_data(description, author, artist, type, status, genres, release, html_name, alternative_names, rank, rating, rating_max, cover_xpath)\n Utils_file::dir_create(@dir)\n Utils_connection::write_cover(@manga_data[:index_page], cover_xpath, @dir + 'cover.jpg', @params[:manga_path] + @manga_data[:website][:dir] + 'mangas/' + @manga_data[:name] + '.jpg')\n File.open(@dir + 'description.txt', 'w') do |txt|\n txt << Utils_file::data_concatenation(@manga_data, Utils_file::description_manipulation(description), author, artist, type, status, genres, release, html_name, alternative_names)\n end\n @aff.data_disp(@manga_data[:in_db])\n if @manga_data[:in_db]\n @db.update_manga(@manga_data[:name], description, author, artist, genres, html_name, alternative_names, rank, rating, rating_max)\n else\n @db.add_manga(@manga_data, description, author, artist, type, status, genres, release, html_name, alternative_names, rank, rating, rating_max)\n end\n end",
"def save()\n str = \"\"\n @base.each_value do |item|\n if item.is_a?(BasicFood)\n str += item.name + \",b,\" + item.calories\n elsif item.is_a?(Recipe)\n str += item.name + \",r\"\n item.foods.each do |food|\n str += \",\" + food.name \n end\n end\n str += \"\\n\"\n end\n File.open(\"FoodDB.txt\",\"w\").write(str)\n end",
"def store_fields(json=true)\n data = @data.map do |data|\n type = parse_store_renderer(data[\"renderer\"])\n hash = { :name => data[\"dataIndex\"] , :mapping => data[\"mapping\"] }\n hash.merge!(type) if type\n hash\n end\n json ? JSON.pretty_generate(data) : data\n end",
"def save_parsed_result(log_hash)\n log_hash.each do |visitor_id, visitor_hash|\n Visitor.create_entry(self.id, visitor_id, visitor_hash)\n end\n end",
"def post_match() end",
"def save; end",
"def save; end",
"def save; end",
"def save; end",
"def save; end",
"def save; end",
"def save; end",
"def save; end",
"def save_match_to_file(fname)\r\n #fname_old_loc = File.expand_path(File.join( File.dirname(__FILE__) + \"/../..\",fname))\r\n fname_old_loc = fname\r\n File.open( fname_old_loc, 'w' ) do |out|\r\n YAML.dump( @info_match, out )\r\n end\r\n end",
"def store_datas\n question = Question.find(self.question_id)\n self.exercise = question.exercise.as_json(include: {questions: {include: :test_cases }})\n self.lo = question.exercise.lo.as_json\n self.question = question.as_json(include: :test_cases)\n self.team = Team.find(self.team_id).as_json if self.team_id\n\ttrue\n end",
"def save(data)\n\t\t# NOTE: in a low-level implementation of this, you want to have two \"save slots\", and juggle pointers to figure out what data is \"this\" and what is \"prev\", rather moving the data down one slot. (ie, you alternate saving data into the left slot, and the right slot, rather than always saving in the right slot)\n\t\t@prev = @this\n\t\t@this = data\n\tend",
"def save_results\n\t\t# save search results\n\t\tCSV.open('results.csv', 'wb') {|csv|\n\t\t\t@checked_links.each {|link|\n\t\t\t\tcsv << [link[0], link[1][:res], link[1][:time]]\n\t\t\t}\n\t\t}\n\t\t# save list of external links\n\t\tCSV.open('external-links.csv', 'wb') {|csv|\n\t\t\t@external_links.each do |link|\n\t\t\t csv << [link[0], link[1][:res], link[1][:time]]\n\t\t\tend\n\t\t}\n\t\t# save list of invalid links\n\t\tCSV.open('invalid.csv', 'wb') {|csv|\n\t\t\t@error_links.each do |link|\n\t\t\t csv << link\n\t\t\tend\n\t\t}\n\tend",
"def save_state\n json_object = { :secret_word => @secret_word, :display_content => @display_content,\n \t :failed_attemps => @failed_attemps }.to_json\n File.open(\"saved_state.json\", \"w\") { |file| file.write(json_object) }\n end",
"def extract_save_contents(contents)\n\t\t\tlocal_extract_save_contents\n\t\t\t$game_selfVars = contents[:self_vars]\n\t\tend",
"def save_turn_data\n if SceneManager.scene_is?(Scene_Battle)\n $game_party.battle_members.each {|actor| fill_data(actor)}\n $game_troop.members.each {|enemy| fill_data(enemy)}\n fill_item_data($game_party.items, 0)\n fill_item_data($game_party.weapons, 1)\n fill_item_data($game_party.armors, 2)\n save_turn_without_rescue\n end\n end",
"def set_seo_data(hash)\n tt = Setting::get('Title for restaurant page')\n unless tt.blank?\n hash[:title] = self.process_seo_str tt\n end\n dt = Setting::get('Description for restaurant page')\n unless dt.blank?\n hash[:description] = self.process_seo_str dt\n end\n kt = Setting::get('Keywords for restaurant page')\n unless kt.blank?\n hash[:keywords] = self.process_seo_str kt\n end\n end",
"def save_candidate_data\n raise('Captcha found.') if captcha?\n puts \"Saving candidate's information.\"\n candidates ||= []\n candidate_results = select_elements(:result_tiles)\n candidate_results.each do |candidate_elem|\n candidates.push(candidate_info(candidate_elem)) \\\n unless company?(candidate_elem)\n end\n fail('No data found.') if candidates.empty?\n end_section\n candidates\n end",
"def save_all_to_files\n\t\tsave_hash_to_file($searched, $progDir+'searched.txt')\n\t\tsave_hash_to_file($downloaded, $progDir+'downloaded.txt')\n\t\tsave_array_to_file($toSearch, $progDir+'toSearch.txt')\n\t\tsave_array_to_file($toDownload, $progDir+'toDownload.txt')\n\t\t@pgCount=$toSearch.size\n\trescue => e\n\t\t$error.save_error_to_file(__method__, e, $specialErrorTxt)\n\tend",
"def save_values\n self.first_name = @first_name_field.text.strip.chomp\n self.last_name = @last_name_field.text.strip.chomp\n self.email = @email_field.text.strip.chomp\n self.github = @github_field.text.strip.chomp\n self.twitter = @twitter_field.text.strip.chomp\n self.fun_fact = @fun_fact_field.text.strip.chomp\n\n # TODO: 2. Finish the implementation to set the other fields. DONE\n end",
"def store_new_giocata(deck, first_player)\r\n info_giocata = { \r\n :deck => deck.dup, \r\n :first_plx => first_player, \r\n :actions => [] \r\n }\r\n @info_match[:giocate] << info_giocata\r\n end",
"def save_data(subject_list)\n File.open('data', 'w') do |file|\n subject_list.each do |subj|\n file.write(\"#{subj.code}|#{subj.name}\\n\")\n subj.plans.each do |plan|\n file.write(\"#{plan.name}\\n\")\n plan.classes.each do |cla|\n file.write(\"#{cla.day}\\n#{cla.start}\\n#{cla.finish}\\n\")\n end\n end\n end\n end\n File.open('data.json', 'w') do |file|\n file.write(create_subjects_json(subject_list))\n end\nend",
"def new_search_data\n # for now these are the same\n activity_search_data\n end",
"def save_data\n puts \"saving data\"\n\n File.open(generate_filename(self), \"w\") do |f|\n f.write(ostruct_to_hash(self.json).to_yaml)\n end\n end",
"def match_data\n Pdfh.verbose_print \"~~~~~~~~~~~~~~~~~~ Match Data RegEx\"\n Pdfh.verbose_print \" Using regex: #{@type.re_date}\"\n Pdfh.verbose_print \" named: #{@type.re_date.named_captures}\"\n matched = @type.re_date.match(@text)\n raise ReDateError unless matched\n\n Pdfh.verbose_print \" captured: #{matched.captures}\"\n\n return matched.captures.map(&:downcase) if @type.re_date.named_captures.empty?\n\n extra = matched.captures.size > 2 ? matched[:d] : nil\n [matched[:m].downcase, matched[:y], extra]\n end",
"def read_save_data(file)\r\n read_characters(file)\r\n read_frame(file)\r\n read_data(file)\r\n read_edit\r\n read_refresh\r\n end",
"def save_text\n @lexemes << @lexeme\n end",
"def record_save\n # Load mapinfo for map name\n mapinfos = load_data(\"Data/MapInfos.rxdata\")\n # Make the filename\n if Wep::Random_file_name and Wep::Modify_game_system\n filename = Wep::Game_version+\"-\"+$game_system.random_filename+\"-\"+Wep::Save_record_name\n else\n filename = Wep::Game_version+\"-\"+Wep::Save_record_name\n end\n estado = open(filename, \"a\")\n estado.write(\"------------------------------------------\\n\")\n estado.write(\"||||\"+Wep::Save_name+\" \"+($game_system.save_count+1).to_s+\"||||\")\n estado.write(\"\\n------------------------------------------\\n\") \n # Write time\n @total_sec = Graphics.frame_count / Graphics.frame_rate\n hour = @total_sec / 60 / 60\n min = @total_sec / 60 % 60\n sec = @total_sec % 60\n text = sprintf(\"%02d:%02d:%02d\", hour, min, sec)\n tiempo = \"\\nPlaytime: \"\n tiempo+= text\n estado.write(tiempo)\n # Write money\n money = \". Money: \"\n money += $game_party.gold.to_s\n estado.write(money)\n # Write actual map(and name)\n estado.write(\". Map: \"+$game_map.map_id.to_s+\" (\"+mapinfos[$game_map.map_id].name+\")\\n\")\n if Wep::Record_saves_actors\n if Wep::Record_saves_all_actors == false\n for actor in $game_party.actors\n estado.write(\"\\n\"+actor.name+\" (\"+$data_classes[actor.class_id].name+\") Nv\"+actor.level.to_s+\"\\n\")\n end\n else\n \n for i in 1...$data_actors.size\n estado.write(\"\\n\"+$game_actors[i].name+\" (\"+$data_classes[$game_actors[i].class_id].name+\") Nv\"+$game_actors[i].level.to_s+\"\\n\")\n end\n \n end\n end\n estado.write (\"\\n\")\n estado.close\n end",
"def add_data(vals)\n\t\tvals.each_index{|index|\n\t\t\tvalue = vals[index].downcase.strip\n\n unless value.eql?('')\n if value.include?(';')\n @data[@fields[index + 1]].concat(value.split(';'))\n else\n @data[@fields[index + 1]]<<value\n end\n\n @data[@fields[index + 1]].uniq!\n end\n\t\t}\n\tend",
"def prepare_data(data)\n data.gsub(/Do some fancy data manipulation here/, '')\n end",
"def search_data\n attributes.merge(\n brief_summary: brief_summary && brief_summary.description,\n detailed_description: detailed_description && detailed_description.description,\n browse_condition_mesh_terms: browse_conditions.map(&:mesh_term),\n conditions: conditions.map(&:downcase_name),\n browse_interventions_mesh_terms: browse_interventions.map(&:mesh_term),\n interventions_mesh_terms: interventions.map(&:name).reject(&:nil?),\n interventions: interventions.map(&:description).reject(&:nil?),\n design_outcome_measures: design_outcomes.map(&:measure),\n facility_names: facilities.map(&:name),\n facility_states: facilities.map(&:state),\n facility_cities: facilities.map(&:city),\n facility_countries: facilities.map(&:country),\n locations: facility_coords_hash,\n average_rating: average_rating,\n reviews_count: reviews.count,\n reviews: reviews && reviews.map(&:text),\n sponsors: sponsors && sponsors.map(&:name),\n rating_dimensions: rating_dimensions.keys,\n indexed_at: Time.now.utc,\n study_views_count: study_view_logs.count,\n wiki_page_edits: {\n email: wiki_page_edits.map(&:user).map(&:email),\n created_at: wiki_page_edits.map(&:created_at).map(&:to_time),\n },\n reactions:{\n email: reactions.order(:id).map(&:user).map(&:email),\n kind: reactions.order(:id).map(&:reaction_name)\n\n },\n ).merge(\n average_rating_dimensions,\n ).merge(\n wiki_search_data,\n ).except(\n # https://github.com/clinwiki-org/clinwiki/issues/111\n *NON_INDEX_FIELDS, *NON_INDEX_FIELDS.map(&:to_s)\n )\n end",
"def create\n @user_say = current_user.user_says.create(user_say_params)\n\n respond_to do |format|\n if @user_say.save\n if @user_say.extract_data == true\n Lesson.where(intent: @user_say.intent, bot_id: @user_say.bot_id).take.update(extract_data: true)\n #создаем регулярку с capture groups и сохраняем их в отдельное поле\n @user_say.update(regexp: @user_say.make_regexp(Regexp.new(Regexp.escape(@user_say.input), \"i\").to_s)) \n # создаем ключ-переменную для привязки примеров переменной\n if @user_say.input.match(/@[\\wа-я]+/i)\n @user_say.input.to_enum(:scan, /(@[\\wа-я]+)/i).map { Regexp.last_match }.each do |a|\n i = 0\n unless Key.where(bot_id: @user_say.bot_id, name: a[i]).take\n Key.create(name: a[i], bot_id: @user_say.bot_id, user_say_id: @user_say.id, lesson_id: @user_say.lesson_id)\n end\n i = i + 1 \n end\n end \n end\n format.html { redirect_to @user_say.bot, notice: 'User say was successfully created.' }\n format.json { render :show, status: :created, location: @user_say }\n else\n format.html { redirect_to @user_say.bot }\n format.json { render json: @user_say.errors, status: :unprocessable_entity }\n end\n end\n end",
"def characterize\n self.characterization.ng_xml = self.content.extract_metadata\n self.append_metadata\n self.filename = self.label\n save\n end",
"def save_unknown!\n\t\t @saved = []\n\t end",
"def write_save_data(file)\r\n write_characters(file)\r\n write_frame(file)\r\n write_setup(file)\r\n write_data(file)\r\n end",
"def make_info_mano_onlogs\r\n @log.debug \"make_info_mano_onlogs\"\r\n @mano_coll = []\r\n curr_ix = 0\r\n curr_mano_info = {:ix => curr_ix, :data => []}\r\n @data_logs.each do |line|\r\n #p line\r\n if line =~ /new_mano/\r\n @mano_coll << curr_mano_info if curr_mano_info.size > 0\r\n curr_ix += 1\r\n curr_mano_info = {:ix => curr_ix, :data => []}\r\n else\r\n curr_mano_info[:data] << line \r\n end\r\n end\r\n end",
"def save_meta_data(type)\n FileUtils.mkdir_p File.dirname(meta_file_path(type))\n File.open(meta_file_path(type), 'w') { |f| f.print self[type].to_yaml }\n if Mist.commit_meta_data\n Mist.repository.add meta_file_path(type)\n Mist.repository.commit '%s meta changes to %s' % [type, table_name]\n end\n\n # we must force meta to be reloaded because otherwise it could get out of sync with filesystem\n @meta = nil\n end",
"def saveOutputRecord(address, data)\r\n newOutput = Output.new\r\n newOutput.street = address.street\r\n newOutput.citystatezip = address.citystatezip\r\n newOutput.names = data[:metricsNames]\r\n newOutput.numbers = data[:metrics]\r\n newOutput.passes = data[:metricsPass]\r\n newOutput.urls = data[:urlsToHit]\r\n newOutput.reason = data[:reason]\r\n newOutput.comments = data[:metricsComments]\r\n newOutput.usage = data[:metricsUsage]\r\n newOutput.zpid = data[:zpid]\r\n newOutput.runid = data[:runID]\r\n #newOutput.time = (Time.now-@startTime-@sectionTimes.inject(:+)).round\r\n newOutput.date = Date.today \r\n newOutput.product = params[:product].to_s.upcase\r\n newOutput.save\r\n end",
"def save_all\n return if @loading_data\n\n json_data = JSON.dump(@model.to_h)\n\n LocalStorage['volt-store'] = json_data\n end",
"def extract_template_details(memo, mode, match)\n if mode == :localized\n memo[match[1]] ||= {}\n memo[match[1]][match[2]] ||= []\n memo[match[1]][match[2]] << match[3]\n else\n memo[match[1]] ||= []\n memo[match[1]] << match[2]\n end\n end",
"def save_values\n self.first_name = @first_name_field.text.strip.chomp\n self.last_name = @last_name_field.text.strip.chomp\n self.age = @age.text.strip.chomp\n self.email = @email.text.strip.chomp\n self.github = @github.text.strip.chomp\n self.twitter = @twitter.text.strip.chomp\n self.fun_fact = @fun_fact.text.strip.chomp\n puts self.first_name\n puts self.last_name\n puts self.age\n\n # TODO: 2. Finish the implementation to set the other fields.\n end",
"def save()\n File.write(@database_file, @data.to_json)\n end",
"def to_save\n hash = {}\n instance_variables.each do |var|\n ignored_fields = [:'@errors', :'@id']\n next if ignored_fields.include? var\n\n field = var.to_s.delete('@')\n\n case field\n when 'genero' then\n hash[:genero_id] = @genero.id\n when 'color1' then\n hash[:color1_id] = @color1 == '' ? nil : @color1\n when 'color2' then\n hash[:color2_id] = @color2 == '' ? nil : @color2\n when 'forma_vida1' then\n hash[:forma_vida1_id] = @forma_vida1 == '' ? nil : @forma_vida1\n when 'forma_vida2' then\n hash[:forma_vida2_id] = @forma_vida2 == '' ? nil : @forma_vida2\n else\n hash[field] = instance_variable_get(var)\n end\n end\n hash[:modified_date] = Time.now\n hash\n end",
"def atualizaTxt\n #abre o arquivo para leitura\n uploaded_file = params[:file]\n uploaded_file = uploaded_file.read();\n\n #declara a regexp a ser utilizada para o parse\n regexp = /(aspx\\?cod=([\\d]{6}))|(<div align=\"center\">([A-Z]{2})<\\/div>)/\n\n #faz o scan no arquivo\n result = uploaded_file.scan(regexp)\n\n #Transforma de matchData para um vetor de hash\n #i é uma flag auxiliar para orientar qual o indice a ser gravado no hash, visto\n #que o matchData armazena matchs diferentes em posicoes diferentes \n #a matricula está em [1] e a menção em [3].\n i = true\n array = Array.new()\n novoElem = Hash.new()\n result.each do |elem|\n if i\n sql = \"select id from disciplinas where codigo = \" + elem[1]\n records_array = ActiveRecord::Base.connection.execute(sql)\n if records_array[0] != nil\n records_array = records_array[0]\n novoElem[:id] = records_array[\"id\"]\n else\n novoElem[:id] = 1\n end\n i = false\n else\n novoElem[:nota] = elem[3]\n i = true\n array.push(novoElem)\n novoElem = Hash.new()\n end\n end\n\n #remove as disciplinas em que o aluno reprovou\n array.delete_if {|disc| (disc[:nota] != \"MM\" && disc[:nota] != \"MS\" && disc[:nota] != \"SS\")}\n \n #adiciona as disciplinas no Resultado da pessoa\n\n\n i = 0\n array.each do |disc|\n i = i + 1\n resultado = Resultado.new\n resultado.aluno_id = current_user.id\n resultado.mencao = disc[:nota]\n resultado.disciplina_id = disc[:id]\n resultado.comentario = \"Adicao Automatica\"\n resultado.semestre = \"2000/01\"\n resultado.save\n end\n\n #redirect_back(fallback_location: root_path)\n\n\n #puts array\n end",
"def add_info(model_data)\n @model_data = model_data\n ensure_file_end_with_empty_line\n if File.exists?(model_path)\n File.open(model_path, 'a') do |file|\n file.puts(ExtendedContent.from(@model_data.content))\n end\n end\n end",
"def to_db_hash\n super.merge(\n {\n 'text' => @text,\n 'url' => @url\n }\n )\n end",
"def save_values\n self.first_name = @first_name_field.text.strip.chomp\n self.last_name = @last_name_field.text.strip.chomp\n self.email = @email_field.text.strip.chomp\n self.github = @github_field.text.strip.chomp\n self.twitter = @twitter_field.text.strip.chomp\n self.fun_fact = @fun_fact_field.text.strip.chomp\n # self.preferred_text_editor = @preferred_text_editor.text.strip.chomp\n # self.teaching_experience = @teaching_experience.text.strip.chomp\n # TODO: 2. Finish the implementation to set the other fields.\n end",
"def save_input_data\n @inputs.each do |input, props|\n next unless props[:properties][\"save\"]\n next unless props[:properties][\"data\"]\n next unless props[:value]\n next if props[:dynamic]\n\n save_to_datastore(input)\n end\n end",
"def prepare_test_data(data)\n case data[:type][1]\n when 'file' then\n data[:file_type] = data[:file_type][1]\n data[:allowed_dirs] = data[:allowed_dirs][1]\n\n when 'relationship' then\n data[:channels] = data[:channels][0]\n data[:categories] = [] #data[:categories][1] #@todo swtich to names?\n data[:authors] = [] #data[:authors][1] #@todo swtich to names?\n data[:statuses] = [] #data[:statuses][1]\n data[:order_field] = data[:order_field][1]\n data[:order_dir] = data[:order_dir][1]\n\n when 'text' then\n data[:field_fmt] = data[:field_fmt][1]\n data[:field_content_type] = data[:field_content_type][1]\n data[:field_text_direction] = data[:field_text_direction][1]\n\n when 'textarea' then\n data[:field_fmt] = data[:field_fmt][1]\n data[:field_text_direction] = data[:field_text_direction][1]\n\n when 'rte' then\n data[:field_text_direction] = data[:field_text_direction][1]\n\n when 'checkboxes', 'multi_select', 'radio', 'select' then\n data[:field_fmt] = data[:field_fmt][1]\n end\n\n return data\n end",
"def cache_content_filter_data\n if name && name_id_changed?\n self.lifeform = name.lifeform\n self.text_name = name.text_name\n self.classification = name.classification\n end\n self.where = location.name if location && location_id_changed?\n end"
] |
[
"0.5834503",
"0.581124",
"0.5772738",
"0.57070047",
"0.56758887",
"0.5592121",
"0.54858935",
"0.54773647",
"0.54695505",
"0.53946084",
"0.5381834",
"0.5381834",
"0.52908367",
"0.52908367",
"0.52908367",
"0.5289626",
"0.5289544",
"0.5273991",
"0.52601457",
"0.5211825",
"0.51921237",
"0.5191284",
"0.5181915",
"0.5180077",
"0.5169888",
"0.516951",
"0.5168365",
"0.5168136",
"0.5165366",
"0.5151574",
"0.5147686",
"0.5143883",
"0.51404756",
"0.51268077",
"0.5081852",
"0.50813955",
"0.5077612",
"0.50758",
"0.5075791",
"0.5072424",
"0.5062681",
"0.50520104",
"0.5049999",
"0.504959",
"0.50350493",
"0.503017",
"0.5029867",
"0.5015334",
"0.5013128",
"0.50035405",
"0.49955276",
"0.4994511",
"0.4991864",
"0.4991864",
"0.4991864",
"0.4991864",
"0.4991864",
"0.4991864",
"0.4991864",
"0.4991864",
"0.49903867",
"0.49863464",
"0.49861333",
"0.49798274",
"0.49666032",
"0.49586144",
"0.49577042",
"0.49573404",
"0.49563324",
"0.49522275",
"0.49521846",
"0.49330786",
"0.49269095",
"0.4919263",
"0.49151155",
"0.49130094",
"0.4912581",
"0.4908402",
"0.4906534",
"0.4904157",
"0.48868912",
"0.4882504",
"0.4877972",
"0.4875004",
"0.48729885",
"0.4870237",
"0.48679143",
"0.486407",
"0.4863184",
"0.48608878",
"0.48590598",
"0.48577777",
"0.48473987",
"0.4846886",
"0.48456982",
"0.48423445",
"0.48411205",
"0.48396176",
"0.48353207",
"0.483489",
"0.48259956"
] |
0.0
|
-1
|
This method prepare the data for future VueJs requests
|
def prepare( directory )
path = '../vue_js/data/' + directory
FileUtils.mkpath( path )
@pages.each do |key, data|
File.open( "#{path}/#{key}.json", 'w' ) do |f|
f.puts( JSON.pretty_generate( data ) )
end
end
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def prepare_data\n\t\t\tsuper\n\t\t\t@data[:tuneable_data] = @tuneable_data\n\t\t\t@data[:@lowest_old] = @lowest_old\n\t\tend",
"def prepare_data\n # This method is just a stub.\n end",
"def prepare_data\n initial_data_eval \n initial_sorting if self.data_grid.sorting?\n filter\n global_summary_array\n pagination\n summary_array\n self.data_grid.row_styler\n prepare_data_for_filters\n end",
"def fill_with_data(data)\n if data.include? \"id\"\n @id = data[\"id\"]\n end\n if data.include? \"project\"\n @project = data[\"project\"]\n end\n if data.include? \"url\"\n @url = data[\"url\"]\n end\n if data.include? \"name\"\n @name = data[\"name\"]\n end\n if data.include? \"amount\"\n @amount = data[\"amount\"]\n end\n if data.include? \"currency\"\n @currency = data[\"currency\"]\n end\n if data.include? \"metadata\"\n @metadata = data[\"metadata\"]\n end\n if data.include? \"request_email\"\n @request_email = data[\"request_email\"]\n end\n if data.include? \"request_shipping\"\n @request_shipping = data[\"request_shipping\"]\n end\n if data.include? \"return_url\"\n @return_url = data[\"return_url\"]\n end\n if data.include? \"cancel_url\"\n @cancel_url = data[\"cancel_url\"]\n end\n if data.include? \"sandbox\"\n @sandbox = data[\"sandbox\"]\n end\n if data.include? \"created_at\"\n @created_at = data[\"created_at\"]\n end\n \n self\n end",
"def initialize_data\n end",
"def fill_with_data(data)\n if data.include? \"id\"\n @id = data[\"id\"]\n end\n if data.include? \"project\"\n @project = data[\"project\"]\n end\n if data.include? \"customer\"\n @customer = data[\"customer\"]\n end\n if data.include? \"token\"\n @token = data[\"token\"]\n end\n if data.include? \"url\"\n @url = data[\"url\"]\n end\n if data.include? \"authorized\"\n @authorized = data[\"authorized\"]\n end\n if data.include? \"name\"\n @name = data[\"name\"]\n end\n if data.include? \"currency\"\n @currency = data[\"currency\"]\n end\n if data.include? \"return_url\"\n @return_url = data[\"return_url\"]\n end\n if data.include? \"cancel_url\"\n @cancel_url = data[\"cancel_url\"]\n end\n if data.include? \"custom\"\n @custom = data[\"custom\"]\n end\n if data.include? \"sandbox\"\n @sandbox = data[\"sandbox\"]\n end\n if data.include? \"created_at\"\n @created_at = data[\"created_at\"]\n end\n \n self\n end",
"def set_data_in_details\n json_data.provision_details_hash!\n end",
"def prepare\n render json: prepare(params)\n end",
"def initialize(data)\n self.data = (self.respond_to? :prepare_data)? prepare_data(data): data\n self\n end",
"def fill_with_data(data)\n if data.include? \"id\"\n @id = data[\"id\"]\n end\n if data.include? \"project\"\n @project = data[\"project\"]\n end\n if data.include? \"gateway\"\n @gateway = data[\"gateway\"]\n end\n if data.include? \"enabled\"\n @enabled = data[\"enabled\"]\n end\n if data.include? \"public_keys\"\n @public_keys = data[\"public_keys\"]\n end\n \n self\n end",
"def initialize_data\n @data = parse_body || {}\n end",
"def init_data\n end",
"def fill_with_data(data)\n if data.include? \"id\"\n @id = data[\"id\"]\n end\n if data.include? \"project\"\n @project = data[\"project\"]\n end\n if data.include? \"brand\"\n @brand = data[\"brand\"]\n end\n if data.include? \"type\"\n @type = data[\"type\"]\n end\n if data.include? \"bank_name\"\n @bank_name = data[\"bank_name\"]\n end\n if data.include? \"level\"\n @level = data[\"level\"]\n end\n if data.include? \"iin\"\n @iin = data[\"iin\"]\n end\n if data.include? \"last_4_digits\"\n @last_4_digits = data[\"last_4_digits\"]\n end\n if data.include? \"exp_month\"\n @exp_month = data[\"exp_month\"]\n end\n if data.include? \"exp_year\"\n @exp_year = data[\"exp_year\"]\n end\n if data.include? \"metadata\"\n @metadata = data[\"metadata\"]\n end\n if data.include? \"sandbox\"\n @sandbox = data[\"sandbox\"]\n end\n if data.include? \"created_at\"\n @created_at = data[\"created_at\"]\n end\n \n self\n end",
"def prepare\n model.tap do |p|\n p.identifier = set_identifiers\n p.meta = set_meta\n p.text = set_text\n p.status = COMPLETED_STATUS\n p.authored = set_date\n p.author = set_author\n p.subject = set_subject\n p.questionnaire = set_questionnaire\n p.group = set_group\n end\n end",
"def prepare\n super\n end",
"def prepare_data\n @data = {\n :banned_points => @banned_points,\n :cost => @current_cost,\n :path => @path\n }\n if @current_cost > @epsilon\n @data[:failed] = true\n end\n end",
"def init(data)\n\t\t\traise InvalidDataError, 'No id found' unless data.has_key?('id')\n\t\t\t@id = data['id']\n\n\t\t\traise InvalidDataError, 'No name found' unless data.has_key?('name')\n\t\t\t@name = data['name']\n\n\t\t\traise InvalidDataError, 'No created_at found' unless data.has_key?('created_at')\n\t\t\t@created_at = DateTime.strptime(String(data['created_at']), '%s') unless data['created_at'].nil?\n\n\t\t\traise InvalidDataError, 'No status found' unless data.has_key?('status')\n\t\t\t@status = data['status']\n\n\t\t\traise InvalidDataError, 'No hash_type found' unless data.has_key?('hash_type')\n\t\t\t@hash_type = data['hash_type']\n\n\t\t\traise InvalidDataError, 'No hash found' unless data.has_key?('hash')\n\t\t\t@hash = data['hash']\n\n\t\t\traise InvalidDataError, 'No last_request found' unless data.has_key?('last_request')\n\t\t\t@last_request = DateTime.strptime(String(data['last_request']), '%s') unless data['last_request'].nil?\n\n\t\t\traise InvalidDataError, 'No last_success found' unless data.has_key?('last_success')\n\t\t\t@last_success = DateTime.strptime(String(data['last_success']), '%s') unless data['last_success'].nil?\n\n\t\t\traise InvalidDataError, 'No output_type found' unless data.has_key?('output_type')\n\t\t\t@output_type = data['output_type']\n\n\t\t\traise InvalidDataError, 'No output_params found' unless data.has_key?('output_params')\n\t\t\t@output_params = parseOutputParams(data['output_params'])\n\n\t\t\t@is_deleted = true if @status == STATUS_DELETED\n\t\tend",
"def prepare_data(params)\n build_ingest_form\n build_uploader(params[:upload], params[:upload_cache])\n build_asset(params[:id], params[:template_id])\n assign_form_attributes(params)\n find_unmapped_rdf\n end",
"def data\n options.merge(computed)\n end",
"def prepare_environment\n data = resource['data'] || {}\n data.each do |key, value|\n data[key] = nil if value == 'nil'\n data[key] = false if value == 'false'\n data[key] = true if value == 'true'\n data[key] = data[key].to_i if key == 'vlanId'\n end\n uri_validation(data)\n @data = data\n data_parse\nend",
"def set_api_response_data\n users_list = []\n @users.each do |u|\n ukd = @user_kyc_details[u.id]\n ukd_present = ukd.present?\n users_list << {\n user_id: u.id,\n case_id: ukd_present ? @user_kyc_details[u.id].id : 0,\n email: u.email,\n registration_timestamp: u.created_at.to_i,\n is_kyc_submitted: ukd_present.to_i,\n whitelist_status: ukd_present ? @user_kyc_details[u.id].whitelist_status : nil,\n action_to_perform: action_to_perform(ukd)\n }\n end\n\n meta = {\n page_number: @page_number,\n total_records: @total_filtered_users,\n page_payload: {\n },\n page_size: @page_size,\n filters: @allowed_filters,\n sortings: @sortings,\n }\n\n data = {\n meta: meta,\n result_set: 'users_list',\n users_list: users_list\n }\n\n @api_response_data = data\n\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize\n super()\n init_data()\n end",
"def prepare\n end",
"def prepare\n end",
"def prepared_data\n hash_list = Array.new\n devices.each { |device| hash_list << device.get_hash }\n\n data = Hash.new\n data['current_date'] = Time.now.strftime(\"%d.%m.%Y\")\n data['devices'] = hash_list\n return data\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def prepare!\n self.prepared_object = resolve_object\n end",
"def initialize data\n @data = data\n end",
"def prepare(property_data)\n build_room(property_data)\n property.instant_booking!\n\n set_base_info\n set_beds_count\n set_amenities\n set_deposit\n set_cleaning_service\n set_additional_info\n set_property_type\n set_images\n set_price_and_availabilities\n set_owner_info\n\n Result.new(property)\n end",
"def get_data\n @data = []\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def data\n @values.fetch('data') { \n @values['data'] = nil\n }\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def prepare\n model.tap do |p|\n p.name = name\n p.identifier = set_identifiers\n p.meta = set_meta\n end\n end",
"def prepare\n model.tap do |p|\n p.name = name\n p.identifier = set_identifiers\n p.meta = set_meta\n end\n end",
"def data\n update_data\n @data\n end",
"def collect_data( data )\n data\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def prepare\n @result = @content.dup\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def data\n {}\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end",
"def initialize()\n @additional_data = Hash.new\n end"
] |
[
"0.6801018",
"0.66666555",
"0.6521163",
"0.63313836",
"0.6284821",
"0.6186728",
"0.6174908",
"0.617153",
"0.6056353",
"0.6034315",
"0.59595585",
"0.5956695",
"0.5948949",
"0.5921524",
"0.5904306",
"0.5882984",
"0.585618",
"0.5830013",
"0.58214325",
"0.5817504",
"0.5799957",
"0.57968026",
"0.57935286",
"0.5786134",
"0.5786134",
"0.5784265",
"0.5768707",
"0.57582927",
"0.5756136",
"0.57466316",
"0.5745895",
"0.5739236",
"0.5739236",
"0.5721783",
"0.57215756",
"0.57215756",
"0.5713397",
"0.5713397",
"0.5704373",
"0.57005143",
"0.5692024",
"0.5692024",
"0.5692024",
"0.5692024",
"0.5692024",
"0.5692024",
"0.5692024",
"0.5692024",
"0.5683397",
"0.5683397",
"0.5683397",
"0.56468666",
"0.5643462",
"0.5643462",
"0.5643462",
"0.5643462",
"0.56282264",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938",
"0.5621938"
] |
0.0
|
-1
|
The absolute path to the artifact's release directory.
|
def release_path
::File.join(install_path, artifact_name, artifact_version)
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def release_path\n @release_path ||= Pathname.new(source_dir).join(data['release_path'] || './release').to_s\n end",
"def filepath\n groupId = @groupId.gsub('.', '/')\n artifactId = @artifactId.gsub('.', '/')\n return \"#{groupId}/#{artifactId}/#{@version}/#{artifactId}-#{@version}\"\n end",
"def path\n File.expand_path File.join(basepath, version_dir, basename)\n end",
"def relative_install_path\n a = artifact\n name = File.basename(a.to_s)\n a.respond_to?(:group) ? \"#{a.group.gsub('.', '/')}/#{name}\" : \"#{name}\"\n end",
"def destination\n @destination ||= File.join(tmpdir, File.basename(configuration[:release_path]))\n end",
"def artifact_directory_url\n [\n fetch(:maven_endpoint),\n maven_repository,\n *fetch(:maven_group_id).split('.'),\n fetch(:maven_artifact_name),\n fetch(:maven_artifact_version)\n ].join('/')\n end",
"def pathDist\n\t\"../../\" + distPackageName + \"_\" + version + \"/\"\nend",
"def current_path\n ::File.join(install_path, artifact_name, 'current')\n end",
"def relative_path\n File.join(@repo, @bundle)\n end",
"def abspath\n \"#{repo_base_path}/#{self.git_repo_path}\"\n end",
"def attachment_dir\n File.join(GIT_REPO, @basename + ATTACH_DIR_SUFFIX)\n end",
"def destination\n #configuration[:release_path]\n @destination ||= File.join(tmpdir, File.basename(configuration[:release_path]))\n\n end",
"def release_path\n @config[:remote_path]\n end",
"def tar_path\n File.expand_path(@env[\"package.output\"], FileUtils.pwd)\n end",
"def build_dist_path(item)\n sub_build_dir = File.join(@config.build_dir, item.name)\n return Dir[File.join(sub_build_dir, \"#{item.name}-[0-9.]*.tar.gz\")][0]\n end",
"def package_dir\r\n \"${0%/#{target_name}}\"\r\n end",
"def output_dir\n File.expand_path(OUTPUT_DIRECTORY, project.install_dir)\n end",
"def output_dir\n File.expand_path(OUTPUT_DIRECTORY, project.install_dir)\n end",
"def version_dir\n products_dir = Product.version_directory\n product_dir = File.join(products_dir, @product.name)\n File.join(product_dir, @version.to_s)\n end",
"def build_archive_dir\n out = config_source['build-archive-dir']\n out = nil if (out != nil && out.downcase == 'none')\n out = FilePath.new(out) unless out.nil?\n out\n end",
"def remote_path_for(package)\n File.join(\n Config.artifactory_base_path,\n package.metadata[:name],\n package.metadata[:version],\n publish_platform(package),\n publish_platform_version(package),\n package.metadata[:basename],\n )\n end",
"def asset_url\n URI(\"https://github.com/#{repo}/releases/download/#{release}/\" <<\n app_name)\n end",
"def target_path(target_name)\r\n File.join(package_dir, target_name)\r\n end",
"def remote_tmp_release_dir\n @remote_tmp_release_dir ||= File.join(remote_dir, File.basename(destination))\n end",
"def build_dir\n File.expand_path(\"#{Config.build_dir}/#{project.name}\")\n end",
"def remote_path_for(package)\n File.join(\n Config.artifactory_base_path,\n Config.artifactory_publish_pattern % package.metadata\n )\n end",
"def bundle_dir\n File.expand_path(File.join(Bixby.repo_path, self.relative_path))\n end",
"def absolute_repository_path\n File.join(\n OpenProject::Revisions::Git::GitoliteWrapper.gitolite_global_storage_path,\n git_path\n )\n end",
"def to_build_rel pn\n pn.relative_path_from(build_dir)\n end",
"def destination_rel_dir\n replace_build_path(super)\n end",
"def bundled_path\n File.dirname Wisp::Source.bundled_path\n end",
"def pkg_dir\n @pkg_dir ||= File.join( self.build_dir, \"#{name + ( version ? \"-#{version}\" : \"\" ) }\" )\n end",
"def maven_repository_path\n command = \"#{package_management_command} help:evaluate -Dexpression=settings.localRepository -q -DforceStdout\"\n command += \" #{@maven_options}\" unless @maven_options.nil?\n stdout, stderr, status = Dir.chdir(project_path) { Cmd.run(command) }\n raise \"Command '#{command}' failed to execute: #{stderr}\" unless status.success?\n\n Pathname(stdout)\n end",
"def downloaded_gem_path\n self.class.downloaded_gem_path @name, @version\n end",
"def repo_dir\n @repo_dir ||= File.join(staging_dir, \"publish\", \"repo\")\n end",
"def package_dir_path\n \"#{package_dir}/#{package_name}\"\n end",
"def build_path\n @build_path ||= \"#{@ant_path}/#{@config[:build_name]}\"\n end",
"def get_project_path\n return File.absolute_path File.join(root_dir, src)\n end",
"def expanded_path\n relative_path(dependency.berksfile.filepath)\n end",
"def repo_path\n @config.get_value('GIT_REPO_ROOT')\n end",
"def downloaded_gem_path\n self.class.downloaded_gem_path @name, @version\n end",
"def lib_out\n @mpc_project.recipe.get_relative_path(@mpc_project.lib_out)\n end",
"def get_build_dir\n if @build_dir\n File.join(@dirname, @build_dir)\n else\n @dirname\n end\n end",
"def get_build_dir\n if @build_dir\n File.join(@dirname, @build_dir)\n else\n @dirname\n end\n end",
"def destination_rel_dir\n @_dest_dir\n end",
"def gem_path\n @path || downloaded_gem_path\n end",
"def artifact_url(id = nil)\n \"#{artifact_directory_url}/#{remote_filename(id)}\"\n end",
"def get_versioned_destination\n original_destination / versioned_directory_name\n end",
"def build_path\n @build_path ||= Pathname.new(source_dir).join(data['build_path'] || './build').to_s\n end",
"def manifest_path\n path = @config[\"manifest\"][\"path\"] if @config[\"manifest\"]\n return Licensed::Git.repository_root.join(path) if path\n\n @config.cache_path.join(\"manifest.json\")\n end",
"def build_target_path\n File.join(destination, configuration[:build_target_path])\n end",
"def framework_path\n File.join(config.build_products_dir, name)\n end",
"def gem_root\n Pathname.new(__FILE__).dirname.parent.parent.expand_path\n end",
"def local_artifact_name\n if File.directory?(@resource[:path])\n b, f = bucket_and_file(@resource[:source])\n artifact = File.join(@resource[:path], f)\n else\n artifact = @resource[:path]\n end\n end",
"def gem_path\n @path || downloaded_gem_path\n end",
"def build_directory\n File.expand_path(config['build_directory'])\n end",
"def path\n @path ||= @project.dir.path\n end",
"def manifest_path\n build_output_dir.join('manifest.json')\n end",
"def target_asset_path(asset)\n File.join('public', asset.folder, asset.filename)\n end",
"def output_dir_path\n @output_dir_path ||= download_dir_path\n end",
"def path\n File.join(@base, @target)\n end",
"def output_path\n File.join(Musako.destination_path, @url)\n end",
"def downloaded_path\n @downloaded_path ||= File.join(install_path, file_name(url))\n return @downloaded_path\n end",
"def repo_path(node)\n node['delivery_builder']['repo'] || File.expand_path('..', __dir__)\n end",
"def assets_manifest_path\n build_output_dir.join('manifest-assets.json')\n end",
"def output_path\n \"build\"\n end",
"def repo_path\n @path\n end",
"def target_root\n APP_ROOT.join(\"deployments/#{target_id}\")\n end",
"def repo_path\n DeliveryGolang::Helpers.repo_path(node)\n end",
"def repository_root\n File.expand_path(@repo.path + \"/../\")\n end",
"def pathDistSource\n\tpathDist + \"source/\"\nend",
"def build_dir\n @build_dir ||= File.join(Crate.project.build_dir, name )\n end",
"def component_build_path(cmp)\n File.join WORK_PATH, 'build', cmp\nend",
"def component_build_path(cmp)\n File.join WORK_PATH, 'build', cmp\nend",
"def __path__\n File.join(root, 'tmp', 'build')\n end",
"def local_artifact_name\n @resource[:path]\n end",
"def go_artifact_uri\n server = ENV['ADMIN_SERVER']\n pipeline_counter = ENV['GO_PIPELINE_LABEL']\n stage_counter = ENV['GO_STAGE_COUNTER']\n \"http://#{server}:8153/go/files/MainBuild/#{pipeline_counter}/MainBuild/#{stage_counter}/MainBuild/dist/companyNews.war\"\n end",
"def project_path\n if(File.directory?(@library_path))\n # library is source dir\n File.join(project_lib, clean_name)\n else\n # library is a binary (like swc, jar, etc)\n File.join(project_lib, File.basename(@file_target.archive_path))\n end\n end",
"def get_svn_path( dir='.' )\n\troot = get_svn_repo_root( dir )\n\turl = get_svn_url( dir )\n\t\n\treturn url.sub( root + '/', '' )\nend",
"def build_root()\n \"#{project_root}/build\"\n end",
"def path\n @svn_name\n end",
"def dmg_dir_path\n \"#{package_dir}/#{dmg_name}\"\n end",
"def workdir\n @workdir ||= ::File.join(::File.dirname(@repo_dir), 'download')\n end",
"def source_root\n FilePath.new(build_module.root, name).canonicalize\n end",
"def log_path\n case Merb::Config[:log_file]\n when String then File.dirname(Merb::Config[:log_file])\n else Merb.root_path(\"log\")\n end\n end",
"def log_path\n case Merb::Config[:log_file]\n when String then File.dirname(Merb::Config[:log_file])\n else Merb.root_path(\"log\")\n end\n end",
"def asset_location()\n if fetch(:assets_path, \"\") != \"\" then\n return asset_location = \"#{assets_path}/#{assets_folder}\"\n end\n\n return asset_location = \"#{assets_folder}\"\n end",
"def package_dir\n config.package_dir\n end",
"def revision_file\n @root.join('REVISION')\n end",
"def asset_path_for( filename )\n\n \"#{Rails.application.config.assets[:path]}/#{versioned_filename_for( filename )}\"\n\n end",
"def cache_repository_path\n File.absolute_path(File.join(Rails.root, 'cache', cached_directory))\n end",
"def full_path\n container.root.join(path)\n end",
"def full_path\n container.root.join(path)\n end",
"def build_output_dir\n root.join(public_dir, public_output_dir)\n end",
"def path\n real_path = Pathname.new(root).realpath.to_s\n full_path.sub(%r{^#{real_path}/}, '')\n end",
"def project_name\n # using Dir.pwd will return something like: \n # /var/www/apps/staging.sweetspot.dm/releases/20061006155448\n # instead of\n # /var/www/apps/staging.sweetspot.dm/current\n pwd = ENV['PWD'] || Dir.pwd\n #another hack..ugh. If using standard capistrano setup, pwd will be the 'current' symlink.\n pwd = File.dirname(pwd) if File.symlink?(pwd)\n File.basename(pwd)\n end",
"def target_dir\n without_parent_dir ? \".\" : package_name\n end",
"def bin_folder configuration = 'Release'\n conf['bin'] || proj.output_path(configuration)\n end",
"def archive_path\n @archive_path ||= ''\n end",
"def path\n Rails.root.join(ROOT, type, name, executable).to_s\n end"
] |
[
"0.7805322",
"0.6771892",
"0.6750139",
"0.67314434",
"0.66772777",
"0.6619616",
"0.6613908",
"0.6569632",
"0.65583193",
"0.6538671",
"0.6538304",
"0.653421",
"0.64757997",
"0.6438014",
"0.64031345",
"0.63880557",
"0.63862365",
"0.63862365",
"0.63840806",
"0.63767666",
"0.6376554",
"0.6370008",
"0.6334297",
"0.63301414",
"0.63231295",
"0.6309706",
"0.62969434",
"0.6289892",
"0.6264804",
"0.62599605",
"0.624226",
"0.62417305",
"0.6234964",
"0.6216137",
"0.62080145",
"0.620362",
"0.61888534",
"0.61859715",
"0.61781204",
"0.61534196",
"0.61472505",
"0.613091",
"0.6109459",
"0.6109459",
"0.6097935",
"0.60783243",
"0.60772306",
"0.60730827",
"0.60633034",
"0.60407746",
"0.60402703",
"0.6038867",
"0.6034677",
"0.6033634",
"0.6032369",
"0.6031301",
"0.60284656",
"0.60193115",
"0.59979695",
"0.5988458",
"0.5986642",
"0.59822327",
"0.59819883",
"0.59693635",
"0.5961244",
"0.5958704",
"0.5955625",
"0.59547216",
"0.5945154",
"0.59442717",
"0.5937571",
"0.59279555",
"0.5919096",
"0.5919096",
"0.5914075",
"0.58912295",
"0.5883782",
"0.58678734",
"0.58625996",
"0.5860682",
"0.5855613",
"0.5853389",
"0.58462965",
"0.5842835",
"0.5838436",
"0.5838436",
"0.58377874",
"0.5825476",
"0.5821619",
"0.58157295",
"0.58131135",
"0.5801896",
"0.5801896",
"0.5801132",
"0.5795478",
"0.57877564",
"0.5787052",
"0.5782763",
"0.5774667",
"0.5770382"
] |
0.86605835
|
0
|
The absolute path to the current symlink for this artifact.
|
def current_path
::File.join(install_path, artifact_name, 'current')
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def linked_path\n File.readlink current_directory\n end",
"def path\n @link.TargetPath\n end",
"def symlink_current_dir\n @shell.symlink self.checkout_path, self.current_path\n end",
"def last_resurrect\n File.readlink(the_symlink)\n end",
"def working_directory\n @link.WorkingDirectory\n end",
"def link_target\n return if current_directory? && linked_path == current_path\n\n FileUtils.ln_s File.expand_path(current_path), current_directory\n end",
"def absolutepath\n if absolute?\n self\n elsif to_s == \".\"\n realpath\n else\n parent.absolutepath + self.basename\n end\n end",
"def current_path\n @current_path ||= @deploy_to + '/current'\n end",
"def rel_path(file)\n File.dirname(file)\n end",
"def abspath\n \"#{repo_base_path}/#{self.git_repo_path}\"\n end",
"def readlink() Path::Name.new(File.readlink(path)) end",
"def current_file_path\n clurl = AssetSettings[:local_assets][@file_id].last\n clurl.sub(/\\A#{AssetSettings[:local_assets].assets_url_prefix}/,\n '') if clurl\n end",
"def to_abs_path\n File.expand_path(self)\n end",
"def local_path\n src = if %i(direct repo).include?(new_resource.source)\n package_metadata[:url]\n else\n new_resource.source.to_s\n end\n ::File.join(Chef::Config[:file_cache_path], ::File.basename(src))\n end",
"def realpath\n if(self.page)\n return Rails.application.routes.url_helpers.page_path(self.page)\n end\n '/' + self.explicit_path\n end",
"def to_absolute_path\n File.join('', to.path(:default).to_s)\n end",
"def local_path\n fetch_path(DevTools.gem_root)\n end",
"def readlink_to_pathname\n rl = readlink\n rl.relative? ? parent + rl : rl\n end",
"def path\n real_path = Pathname.new(root).realpath.to_s\n full_path.sub(%r{^#{real_path}/}, '')\n end",
"def abs_path() \n if not @abs_path.nil? \n return @abs_path\n end\n\n if @parent.nil? or @parent.abs_path.nil?\n return @rel_path\n end\n\n @abs_path = \"#{@parent.abs_path}.#{rel_path}\"\n\n return @abs_path\n end",
"def getRealPath(path) Pathname.new(path).realpath.to_s; end",
"def getRealPath(path) Pathname.new(path).realpath.to_s; end",
"def relative_path\n @local_path.relative_path_from(@platform.local_path)\n end",
"def path\n File.join(@base, @target)\n end",
"def expanded_path\n relative_path(dependency.berksfile.filepath)\n end",
"def path\n application? ? application_path : local_path\n end",
"def path\n @unlinked ? nil : @tmpfile.path\n end",
"def relroot\n Pathname.new(File.expand_path(path)).\n relative_path_from(Pathname.new(File.expand_path(root))).to_s\n end",
"def symlink( *args )\n lnk, opts = self.destination_and_options( args )\n \n if opts[:absolute]\n lnk = lnk.fwf_filepath.expand\n else\n lnk = lnk.fwf_filepath\n end\n \n FileUtils.ln_s( self, lnk, ** Utils::Opts.narrow_file_utils_options( opts, :ln_s ) )\n lnk.fwf_filepath\n end",
"def current_file_path\n current_file.to_path\n end",
"def relative_install_path\n a = artifact\n name = File.basename(a.to_s)\n a.respond_to?(:group) ? \"#{a.group.gsub('.', '/')}/#{name}\" : \"#{name}\"\n end",
"def relative_path\n return self.avatar.match(/http[^|]*/)[0].to_s\n end",
"def right_link_dynamic_state_dir\n return pretty_path(File.join(Dir::COMMON_APPDATA, 'RightScale', 'right_link'))\n end",
"def manifest_path\n path = @config[\"manifest\"][\"path\"] if @config[\"manifest\"]\n return Licensed::Git.repository_root.join(path) if path\n\n @config.cache_path.join(\"manifest.json\")\n end",
"def remote_path_for(package)\n File.join(path, package.trigger)\n end",
"def absolute_repository_path\n File.join(\n OpenProject::Revisions::Git::GitoliteWrapper.gitolite_global_storage_path,\n git_path\n )\n end",
"def relative_path\n File.join(@repo, @bundle)\n end",
"def target_path\n Pathname.new(self.config[:target_path])\n end",
"def base_path\n Dir.pwd + \"/\"\n end",
"def full_rel_path()\n return nil if rel_path.nil?\n \n path = nil\n current_part = self\n while not current_part.nil? do\n if (not current_part.rel_path.nil?)\n if path.nil?\n path = current_part.rel_path\n else\n path = \"#{current_part.rel_path}.#{path}\"\n end\n end\n current_part = current_part.parent\n end\n \n return path\n end",
"def getRefPath()\n return @refPath\n end",
"def path\n File.join(self.drive.path, self.relative_path)\n end",
"def remote_path_for(package)\n File.join(\n Config.artifactory_base_path,\n package.metadata[:name],\n package.metadata[:version],\n publish_platform(package),\n publish_platform_version(package),\n package.metadata[:basename],\n )\n end",
"def private_bin_dir\n return pretty_path(File.join(right_link_home_dir, 'bin'))\n end",
"def determine_path\n if source_line.file == SourceLine::DEFAULT_FILE\n return source_line.file\n end\n\n full_path = File.expand_path(source_line.file)\n pwd = Dir.pwd\n\n if full_path.start_with?(pwd)\n from = Pathname.new(full_path)\n to = Pathname.new(pwd)\n\n return from.relative_path_from(to).to_s\n else\n return full_path\n end\n end",
"def relative_path\n @relative_path ||= absolute_path.sub(/^#{Bookshelf::remote_folder}\\/?/,'')\n end",
"def link_path\n File.join('/_attachment', @page_name, name)\n end",
"def path(rel)\n File.join(File.dirname(__FILE__), \"..\", rel)\nend",
"def path\n @path ||= filters.uri_escape(absolute_url) if absolute_url\n end",
"def path\n @path ||= filters.uri_escape(absolute_url) if absolute_url\n end",
"def relative_path\n name\n end",
"def relative_path_to(path, relative_to = nil)\n if relative_to\n path = File.expand_path(\n # symlink, e.g. \"../../../../grid5000/environments/etch-x64-base-1.0.json\"\n path,\n # e.g. : File.join(\"/\", File.dirname(\"grid5000/sites/rennes/environments/etch-x64-base-1.0\"))\n File.join('/', File.dirname(relative_to))\n ).gsub(%r{^/}, '')\n end\n path\n end",
"def target_path\n Pathname.new(config[:target_path])\n end",
"def path\n \"#{@parent.path}##{@aref}\"\n end",
"def path\n \"#{@parent.path}##{@aref}\"\n end",
"def realpath\n @site.tags[@tag].sort_by do |v|\n v.data[\"date\"]\n end.first.relative_path\n end",
"def past_scripts_path\n File.join(RIGHT_LINK_SPEC_HELPER_TEMP_PATH, '__past_scripts.js')\n end",
"def path()\n return ::File.join(@root, @name)\n end",
"def downloaded_path\n @downloaded_path ||= File.join(install_path, file_name(url))\n return @downloaded_path\n end",
"def local_path\n @io.local_path\n end",
"def right_link_static_state_dir\n return pretty_path(File.join(Dir::COMMON_APPDATA, 'RightScale', 'rightscale.d', 'right_link'))\n end",
"def current_path\n current_folder.path\n end",
"def gem_path\n @path || downloaded_gem_path\n end",
"def current\n\t\t\t\t\treturn Pathname.new(\".\")\n\t\t\t\tend",
"def download_path\n ::File.join(Chef::Config[:file_cache_path], ::File.basename(URL))\n end",
"def abs_filepath\n @epub.manifest.abs_path_from_id(@id)\n end",
"def gem_path\n @path || downloaded_gem_path\n end",
"def path\n (public_path + sitemaps_path + filename).expand_path.to_s\n end",
"def remote_path\n File.join(path, TRIGGER).sub(/^\\//, '')\n end",
"def target_of(path)\n # `path' may not be the symlink - perhaps an ancestor\n # directory points to the target.\n #\n # e.g., if the directory \"LAYER/dir\" points to \"/usr/lib/dir\",\n # and \"/usr/lib/dir/file\" exists, then the\n # target_of(\"dir/file\") is \"/usr/lib/dir/file\".\n current = self.path.dup\n segments = path.split(File::SEPARATOR)\n while (segment = segments.shift)\n current << File::SEPARATOR << segment\n if File.symlink?(current)\n segments.unshift File.readlink(current)\n return segments.join(File::SEPARATOR)\n elsif File.directory?(current)\n next\n else\n return nil\n end\n end\n nil\n end",
"def full_path_to_remote_dir\n (remote_dir[0] == ?/ ? remote_dir : \"$(pwd)/#{remote_dir}\").chomp('/')\n end",
"def release_path\n ::File.join(install_path, artifact_name, artifact_version)\n end",
"def absolute_path\n if is_url?\n # Use the last path component without the query string plus the name\n # of the resource in Base64. This should be both mildly readable and\n # also unique per invocation.\n url_part = URI(path).path.split(/\\//).last\n base64_name = Base64.strict_encode64(name).gsub(/\\=/, '')\n ::File.join(Chef::Config[:file_cache_path], \"#{base64_name}_#{url_part}\")\n else\n ::File.expand_path(path, Chef::Config[:file_cache_path])\n end\n end",
"def lib_out\n @mpc_project.recipe.get_relative_path(@mpc_project.lib_out)\n end",
"def path\n \"#{@parent.path}##{aref}\"\n end",
"def download_path\n ::File.join(Chef::Config[:file_cache_path],\n ::File.basename(source_path))\n end",
"def rel_path(path)\n Pathname(path).expand_path.relative_path_from(Pathname(Dir.pwd))\n end",
"def path\n @base\n end",
"def log_path\n File.join(RIGHT_LINK_SPEC_HELPER_TEMP_PATH, '__nanite.log')\n end",
"def target_file_path\n model.source_location(target_symbol)[:file_path]\n end",
"def local\n return Global.git_repository_directory + @address_digest\n end",
"def rel relative_path\r\n return File.dirname(__FILE__) + \"/../\" + relative_path\r\nend",
"def path\n File.expand_path File.join(basepath, version_dir, basename)\n end",
"def symlink(name)\n FileUtils.ln_s current_item, name\n ls\n end",
"def path\n File.join(RH_CONFIG[\"location\"], self.parent.pid.gsub(/:/,\"_\"), \"data\", self.name.first) unless self.parent.nil? or self.name.empty?\n end",
"def repo_path\n @path\n end",
"def relative_path; end",
"def relative_path; end",
"def relative_path; end",
"def relative_path; end",
"def relative_path; end",
"def abs_path_with(rel_path)\n path = abs_path\n return rel_path if path.nil?\n return \"#{path}.#{rel_path}\"\n end",
"def relative_path\n @relative_path ||= File.join(@dir, @target)\n end",
"def path\n Rails.root.join(ROOT, type, name, executable).to_s\n end",
"def full_defined_path()\n return nil if is_root?\n return name if parent.is_root?\n return \"#{parent.full_defined_path}.#{name}\"\n end",
"def full_path\n container.root.join(path)\n end",
"def full_path\n container.root.join(path)\n end",
"def relative_directory\n return '' unless @directory_root\n @path - @directory_root - name\n end",
"def current_path\n file.try(:path)\n end",
"def relative_path\n must_be File\n Pathname.new(self.full_path).relative_path_from(Pathname.new(Dir.pwd)).to_s\n end"
] |
[
"0.821351",
"0.73287606",
"0.7153724",
"0.7048193",
"0.6653185",
"0.6583951",
"0.6503875",
"0.64727736",
"0.647031",
"0.6444303",
"0.64361703",
"0.6422452",
"0.6421492",
"0.63861394",
"0.6382418",
"0.63717014",
"0.6364741",
"0.6350254",
"0.63377297",
"0.6308187",
"0.6257666",
"0.6257666",
"0.6256936",
"0.6254793",
"0.6254297",
"0.62270373",
"0.62260884",
"0.62209225",
"0.62193936",
"0.619377",
"0.6192097",
"0.6181091",
"0.6177643",
"0.6171603",
"0.6165289",
"0.6155148",
"0.6142763",
"0.6137964",
"0.6101035",
"0.60917616",
"0.60867226",
"0.6067223",
"0.6058203",
"0.6057239",
"0.60569274",
"0.605479",
"0.60526943",
"0.6045907",
"0.60365105",
"0.60365105",
"0.6032189",
"0.60262614",
"0.6006722",
"0.6002274",
"0.6002274",
"0.59730786",
"0.59678483",
"0.5956065",
"0.595334",
"0.59417045",
"0.5941615",
"0.59390324",
"0.5934843",
"0.59272045",
"0.59271055",
"0.59231794",
"0.591982",
"0.5919039",
"0.5910994",
"0.5903757",
"0.5902778",
"0.5900404",
"0.5898648",
"0.58970517",
"0.58947504",
"0.5879595",
"0.58756536",
"0.5874173",
"0.5873049",
"0.58716565",
"0.58707345",
"0.58655435",
"0.5864368",
"0.5860606",
"0.5856853",
"0.5856139",
"0.58534527",
"0.58534527",
"0.58534527",
"0.58534527",
"0.58534527",
"0.5846414",
"0.5844112",
"0.5840668",
"0.5836096",
"0.5828368",
"0.5828368",
"0.58101857",
"0.5810048",
"0.58011883"
] |
0.72428364
|
2
|
GET /image_paths/new GET /image_paths/new.xml
|
def new
@image_path = ImagePath.new
respond_to do |format|
format.html # new.html.erb
format.xml { render :xml => @image_path }
end
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def new\n @uploaded_image = UpdatedImage.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @uploaded_image }\n end\n end",
"def new\n @img = Img.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @img }\n end\n end",
"def new\n @image = Image.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @console_image }\n end\n end",
"def new\n @image = Image.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @image }\n end\n end",
"def new\n @image_set = ImageSet.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @image_set }\n end\n end",
"def new\n @dynamic_image = DynamicImage.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @dynamic_image }\n end\n end",
"def new\n @image = Image.new\n \n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @image }\n end\n end",
"def new\n @image = @owner.images.build\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @image }\n end\n end",
"def new\n @img_info = ImgInfo.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @img_info }\n end\n end",
"def new\n @images_index = ImagesIndex.new nil\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @images_index }\n end\n end",
"def new\n @map_image = MapImage.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @map_image }\n end\n end",
"def new\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @welcome_image }\n end\n end",
"def newFromSource\n\n @kite = Kite.new\n @images = []\n @isAutoAdd = true\n\n if uri_is_valid?(params[:path])\n site = URI.parse(URI.encode(params[:path]))\n\n #grab the reference page passed in as parameter\n doc = Nokogiri::HTML(open(site))\n\n #find all of the images\n doc.css('img').each do |image|\n #debugger\n path = image.attribute('src').value\n\n if path[0..0] == '/'\n path = URI.join(site, path).to_s\n end\n begin\n dimensions = FastImage.size(path, :timeout => 10.0)\n rescue => ex\n logger.error(\"Failed to retrieve image dimensions for image: #{ex}\")\n end\n if(!dimensions.nil? && dimensions.length > 1 && dimensions[0] > IMAGE_SQUARE_DIMENSION_LIMIT && dimensions[1] > IMAGE_SQUARE_DIMENSION_LIMIT )\n img = {:path => path,\n :source => doc.title,\n :alttext => image.attribute('alt').nil? ? \"\" : image.attribute('alt').value\n }\n\n @images << img\n else\n logger.debug \"The image #{path} was rejected as under size limit\"\n end\n end\n else\n logger.error \"The image was rejected as invalid path #{path}\"\n end\n @site = site\n if @images.count > 0\n @images[0][:first] = true\n end\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @kite }\n end\n end",
"def new\n @gallery_image = @project.gallery_images.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @gallery_image }\n end\n end",
"def new\n @jpeg_folder = JpegFolder.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @jpeg_folder }\n end\n end",
"def new\n @path = Path.new({:layer => @layer})\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @path }\n end\n end",
"def new\n @image_url = ImageUrl.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @image_url }\n end\n end",
"def new\n @image = Image.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @image }\n end\n end",
"def new\n @pic = Pic.new\n \n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @pic }\n end\n end",
"def new\n @image_gallery_image = @image_gallery_group.images.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @image_gallery_image }\n end\n end",
"def new\n @imagem = @evento.imagems.build\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @imagem }\n end\n end",
"def new\n @remote_image_content = RemoteImageContent.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @remote_image_content }\n end\n end",
"def new\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => new_vurl }\n end\n end",
"def new\n @photo_set = PhotoSet.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @photo_set }\n end\n end",
"def create\n @image_path = ImagePath.new(params[:image_path])\n\n respond_to do |format|\n if @image_path.save\n flash[:notice] = 'ImagePath was successfully created.'\n format.html { redirect_to(@image_path) }\n format.xml { render :xml => @image_path, :status => :created, :location => @image_path }\n else\n format.html { render :action => \"new\" }\n format.xml { render :xml => @image_path.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def new\n @project_image = ProjectImage.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @project_image }\n end\n end",
"def new\n @image_store = ImageStore.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @image_store }\n end\n end",
"def new\n @image = @user.images.build\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @image }\n end\n end",
"def new\n @photo = Photo.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @photo }\n end\n end",
"def new\n @picture = Picture.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @picture }\n end\n end",
"def new\n @picture = Picture.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @picture }\n end\n end",
"def new\n @photo = Photo.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @photo }\n end\n end",
"def new\n @photo = Photo.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @photo }\n end\n end",
"def new\n @photo = Photo.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @photo }\n end\n end",
"def new\n @photo = Photo.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @photo }\n end\n end",
"def new\n @image = Image.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render :json => @image }\n end\n end",
"def new\n @image = Image.new\n respond_to do |format|\n format.html # new.html.erb\n format.json { render :json => @image }\n end\n end",
"def new\n @image = Image.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @image }\n end\n end",
"def new\n @image = Image.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @image }\n end\n end",
"def new\n @image = Image.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @image }\n end\n end",
"def new\n @image = Image.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @image }\n end\n end",
"def new\n @image = Image.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @image }\n end\n end",
"def new\n @image = Image.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @image }\n end\n end",
"def new\n @gpath = Gpath.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @gpath }\n end\n end",
"def new\n @image = Image.where(\"id is not ?\",nil)\n\n respond_to do |format|\n format.html # new.html.haml\n format.xml { render :xml => @image }\n end\n end",
"def new\n @image_dataset = ImageDataset.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @image_dataset }\n end\n end",
"def new\n @image_dataset = ImageDataset.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @image_dataset }\n end\n end",
"def new\n @image = Image.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.js\n format.xml { render :xml => @image }\n end\n end",
"def new\n @image = Image.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.js\n format.xml { render :xml => @image }\n end\n end",
"def new\n @picture = Picture.new\n\n respond_to do |wants|\n wants.html # new.html.erb\n wants.xml { render :xml => @picture }\n end\n end",
"def new\n @media_path = MediaPath.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @media_path }\n end\n end",
"def register_new_image(image, xml_node, data_item)\n image.set_path(data_item)\n image.is_new = true\n image.id, image.target = @relationship_manager.new_relationship(:image, image.path)\n image.replace_new_image_id!(xml_node, data_item)\n @images << image\n image\n end",
"def element\n image_name_list = get_image_name_list ORIGINALS_PATH\n image_name_list.each { |image_name| post(image_name) }\n end",
"def new\n @old_point_tag = OldPointTag.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @old_point_tag }\n end\n end",
"def new\n @user_image = UserImage.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @user_image }\n end\n end",
"def new\n @gallery = Gallery.new\n @rights = {'Moderators' => 'moderators', 'Members' => 'members', 'Anybody' => 'all'}\n\n @gallery.parent_type = params[:parent_type]\n @gallery.parent_id = params[:parent_id]\n set_session_parent_pictures_root_path(@gallery.get_parent_object)\n\n @new_pictures = Array.new\n 1.upto(3) { @new_pictures << Picture.new }\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @gallery }\n end\n end",
"def new\n @imovel = Imovel.new\n @imovel.images.build\n \n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @imovel }\n end\n end",
"def new\n @graphic = Graphic.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @graphic }\n end\n end",
"def new\n @external_photo = ExternalPhoto.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @external_photo }\n end\n end",
"def new\n @image_datum = ImageDatum.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @image_datum }\n end\n end",
"def get_all_new\n uri = [@@base_uri, 'all', 'getAllNew'].join('/')\n return get(uri)\n end",
"def new\n @snap = Snap.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @snap }\n end\n end",
"def create_uri\n end",
"def new\n @especie_imagem = EspecieImagem.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @especie_imagem }\n end\n end",
"def new\n @action_graphic = ActionGraphic.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @action_graphic }\n end\n end",
"def new\n \n @page = Page.new\n @page.images.build\n \n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @page }\n end\n end",
"def new\n @exposure = Exposure.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @exposure }\n end\n end",
"def new\n @patient_photo = PatientPhoto.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @patient_photo }\n end\n end",
"def new\n @image = Image.new(:user => current_user)\n if params[:image_shell_id] && params[:request_id]\n @image.image_shell = ImageShell.find(params[:image_shell_id])\n @image.description = @image.image_shell.description\n @image.request = Request.find(params[:request_id]) \n end\n \n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @image }\n end\n end",
"def new\n @pix = Pix.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @pix }\n end\n end",
"def new\n @product_picture = ProductPicture.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @product_picture }\n end\n end",
"def create_url\n \"#{api_url}/gists\"\n end",
"def new\n @nspirefile = Nspirefile.new\n @categories = get_categories\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @nspirefile }\n end\n end",
"def index\n @images = @owner.images\n\n respond_to do |format|\n format.html # index.html.erb\n format.xml { render :xml => @images }\n end\n end",
"def new\n @image_position_template = ImagePositionTemplate.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @image_position_template }\n end\n end",
"def images\n end",
"def new\n @gallery = Gallery.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @gallery }\n end\n end",
"def new\n @gallery = Gallery.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @gallery }\n end\n end",
"def new\n @image_gallery_group = ImageGalleryGroup.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @image_gallery_group }\n end\n end",
"def new\n @bwimage = Bwimage.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @bwimage }\n end\n end",
"def get_svg\n\t expires_in(1.hours, :private => false, :public => true)\n\t source_fedora_object = Multiresimage.find(params[:id])\n\t authorize! :show, source_fedora_object\n\t @svg = source_fedora_object.DELIV_OPS.content()\n gon.url = DIL_CONFIG['dil_js_url']\n respond_to do |wants|\n wants.xml { render :xml => @svg }\n end\n end",
"def get_svg\n\t expires_in(1.hours, :private => false, :public => true)\n\t source_fedora_object = Multiresimage.find(params[:id])\n\t authorize! :show, source_fedora_object\n\t @svg = source_fedora_object.DELIV_OPS.content()\n gon.url = DIL_CONFIG['dil_js_url']\n respond_to do |wants|\n wants.xml { render :xml => @svg }\n end\n end",
"def new\n @item_photos = ItemPhoto.new\n @item_photos.item_id = params[:item_id]\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @item_photos }\n end\n end",
"def new\n @text2_image = Text2Image.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @text2_image }\n end\n end",
"def new\n @article = Article.find_by_id(params[:article_id])\n @photo = Photo.new\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @photo }\n end\n end",
"def new\n @event = Event.find(params[:event_id])\n @photo = Photo.new(:event => @event)\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @photo }\n end\n end",
"def new\n @svg_file = Cmtool::SvgFile.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @svg_file }\n end\n end",
"def new\n @news = News.new\n @news.build_image_reference\n @news.image_references.build\n @news.image_references.build\n# @news.image_references_attributes = [{:content_type=>\"news\"}]\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @news }\n end\n end",
"def new\n @photo_admin = PhotoAdmin.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @photo_admin }\n end\n end",
"def create_doc\n\t\theaders = {:content_type => 'application/xml'}\n\t\turi = CGI.escape(\"/\" + rand(2).to_s() + \"/\" + rand(10).to_s() + \"/\" + rand(1000000).to_s() + \".xml\")\n\t\tresponse = @resource[\"/documents/\" + uri].put('<stuff xmlns=\"stuff\">' + uri + '</stuff>', headers)\n\t\treturn uri\n\tend",
"def new\n @market_segment_image = MarketSegmentImage.new\n\n respond_to do |format|\n format.html # new.html.erb\n end\n end",
"def new\n @imaging_observation = ImagingObservation.new\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @imaging_observation }\n format.js # { render :layout => false } \n format.js { render :layout => false }\n end\n end",
"def show\n @uploaded_image = UpdatedImage.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.xml { render :xml => @uploaded_image }\n end\n end",
"def index\n @images = Image.order('created_at DESC')\n @image = Image.new\n @uploader = ImageUploader.new\n # @uploader.key = \"uploads/#{@image.id}-#{request.ip}/${filename}\"\n # @uploader.success_action_status = \"201\"\n @uploader.success_action_redirect = new_image_url\n end",
"def url\n #{}\"http://blah/\" + id.to_s + \"/rss\"\n url_for \"Delete Image\", { :action => \"show\", :id => id.to_s }\n end",
"def index\n @images = Image.all\n\n respond_to do |format|\n format.html # index.html.erb\n format.xml { render :xml => @images }\n end\n end",
"def new\n @event = Event.new\n NO_OF_IMAGES.times {@event.event_uploads.build}\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @event }\n end\n end",
"def clone_storage(request, params, action_xml)\n # --- Get the Image ---\n image = ImageOCCI.new(\n Image.build_xml(params[:id]),\n @client)\n\n rc = image.clone(action_xml[\"PARAMS/NAME\"])\n if OpenNebula.is_error?(rc)\n return rc, CloudServer::HTTP_ERROR_CODE[rc.errno]\n end\n\n new_image = ImageOCCI.new(\n Image.build_xml(rc),\n @client)\n\n new_image.info\n\n # --- Prepare XML Response ---\n return to_occi_xml(new_image, :code=>201)\n end",
"def new\n @icon = MavenJnlp::Icon.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @icon }\n end\n end",
"def new\n @slide_image = SlideImage.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @slide_image }\n end\n end"
] |
[
"0.6165053",
"0.6133824",
"0.5990479",
"0.5975609",
"0.5970181",
"0.59514606",
"0.593504",
"0.5910915",
"0.5905742",
"0.5843028",
"0.5842058",
"0.58256274",
"0.5822929",
"0.57385665",
"0.5719524",
"0.5719442",
"0.56936336",
"0.5641516",
"0.56367075",
"0.5625314",
"0.5615091",
"0.55960506",
"0.55807674",
"0.55661184",
"0.5559691",
"0.55578804",
"0.55542207",
"0.55525583",
"0.55499893",
"0.55410254",
"0.55410254",
"0.5526512",
"0.5526512",
"0.5526512",
"0.5526512",
"0.55179214",
"0.551288",
"0.5507239",
"0.5507239",
"0.5507239",
"0.5507239",
"0.5507239",
"0.5507239",
"0.55070597",
"0.54884565",
"0.54871887",
"0.54871887",
"0.54816055",
"0.54816055",
"0.54801345",
"0.5453583",
"0.54517996",
"0.5440476",
"0.5420225",
"0.5416524",
"0.53918606",
"0.5377479",
"0.53573084",
"0.5355933",
"0.53442806",
"0.53439695",
"0.5342612",
"0.5321583",
"0.5313169",
"0.531161",
"0.5306571",
"0.52945834",
"0.5283873",
"0.5280088",
"0.52659255",
"0.5259251",
"0.5256928",
"0.5244798",
"0.52398366",
"0.523303",
"0.52324665",
"0.52271587",
"0.52271587",
"0.5222096",
"0.5218119",
"0.52134573",
"0.52134573",
"0.5206385",
"0.5206263",
"0.52056265",
"0.51931304",
"0.51927364",
"0.51784295",
"0.51754785",
"0.5174332",
"0.51728684",
"0.51672184",
"0.5164607",
"0.5161868",
"0.5161743",
"0.5156099",
"0.5153642",
"0.51489925",
"0.5138401",
"0.5136069"
] |
0.69418824
|
0
|
POST /image_paths POST /image_paths.xml
|
def create
@image_path = ImagePath.new(params[:image_path])
respond_to do |format|
if @image_path.save
flash[:notice] = 'ImagePath was successfully created.'
format.html { redirect_to(@image_path) }
format.xml { render :xml => @image_path, :status => :created, :location => @image_path }
else
format.html { render :action => "new" }
format.xml { render :xml => @image_path.errors, :status => :unprocessable_entity }
end
end
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def element\n image_name_list = get_image_name_list ORIGINALS_PATH\n image_name_list.each { |image_name| post(image_name) }\n end",
"def make_request(img_paths)\n # Load contents of the imagefiles and instantiate Google::Apis::VisionV1::Image objects\n img_data = img_paths.map { |p| File.open(p).read }\n images = img_data.map { |d| Vision::Image.new(content: d)}\n\n # Maximum of 5 labels per image\n features = [ Vision::Feature.new(max_results: 5, type: 'LABEL_DETECTION') ]\n\n # Instantiate individual annotation API requests for each image\n requests = images.map { |i| Vision::AnnotateImageRequest.new(image: i, features: features) }\n\n # Combine requests into a single batch - max of 50\n batch_request = Vision::BatchAnnotateImagesRequest.new(requests: requests)\n end",
"def create\n @rent = Rent.new(rent_params)\n @rentpar = rent_params\n respond_to do |format|\n if @rent.save\n\n if params[:image]\n puts params[:image]\n params[:image].each { |image|\n @rent.rent_images.create(rent_id: @rent.id, image:image)\n }\n \n end\n\n format.html { redirect_to @rent, notice: 'Rent was successfully created.' }\n format.json { render :show, status: :created, location: @rent }\n else\n format.html { render :new }\n format.json { render json: @rent.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @post = Post.new(post_params)\n\n\n if @post.save && params[:images]\n params[:images].each { |image|\n @post.images.create(image: image)\n }\n end\n\n redirect_to @post\n end",
"def img_params\n\n end",
"def create\n\n post = Post.new(post_params)\n post.location = Location.create(name: params[:post][:location][:name], lat: params[:post][:location][:lat], long: params[:post][:location][:long])\n if post.save\n params[:post][:images].each do |i|\n img = Image.find(i[:id])\n img.update(active: 1, post_id: post.id)\n end\n\n render json: {\n status: \"success\",\n data: post.as_json(\n include: [\n {\n user:\n {\n only: [:id, :name, :avatar]\n }\n },\n :location,\n {\n images: {\n only: [:id, :src]\n }\n },\n :rates\n ])}, status: :ok\n\n else\n render json: post.errors, status: 404\n end\n end",
"def upload_image\n @image = Image.create(image_path: params[:upload][:image])\n p @image\n render json: @image\n end",
"def create\n\n if current_user.nil?\n redirect_to '/'\n end\n\n if(params[\"labels\"].nil?)\n respond_to do |format|\n format.html { redirect_to image_label_sets_url, error: 'Labels not present.' }\n end\n return\n end\n\n @image_label_set = ImageLabelSet.new\n @image_label_set.name = params[\"name\"]\n @image_label_set.user_id = current_user.id\n save_success = @image_label_set.save\n\n params[\"labels\"].split(\",\").each do |l|\n lb = Label.new\n lb.text = l\n lb.image_label_set_id = @image_label_set.id\n lb.save\n end\n\n images_folder_path = Rails.root.join('public', \"images/#{@image_label_set.id}\")\n FileUtils::mkdir_p images_folder_path\n\n accepted_formats = [\".jpg\", \".png\", \".bmp\"]\n\n params[\"upload\"].each do |uf|\n #Check if zipfile, raw images or URL textfile\n if (File.extname(uf.tempfile.path)==\".txt\")\n Image.transaction do\n File.readlines(uf.tempfile.path).each do |line|\n i = Image.new\n i.url = line.strip\n i.image_label_set_id = @image_label_set.id\n i.save\n end\n end\n end\n uf.tempfile.close\n uf.tempfile.unlink\n end\n\n respond_to do |format|\n if save_success\n format.html { redirect_to @image_label_set, notice: 'Image label set was successfully created.' }\n format.json { render :show, status: :created, location: @image_label_set }\n else\n format.html { render :new }\n format.json { render json: @image_label_set.errors, status: :unprocessable_entity }\n end\n end\n end",
"def add_images\n @parent = get_parent(params[:parent_id])\n\n #Images to Add\n if(!@parent.images.nil?)\n @parent.images.clear\n end\n if(!params[:Image_ids].nil?)\n for id in params[:Image_ids]\n @parent.images.push(Image.find(id))\n end\n end\n respond_to do |format|\n if @parent.save\n #Uses a session variable for reloading the current page assigned to the variable\n format.html { redirect_to session[:rq], notice: 'Image(s) successfully added.' }\n format.json { render json: @parent, status: :created, location: @parent }\n else\n format.html { render action: \"\" }\n format.json { render json: @parent.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @apartment = current_user.apartments.new(apartment_params)\n\n respond_to do |format|\n if @apartment.save\n if params[:images]\n # The magic is here ;)\n params[:images].each { |image|\n @apartment.pictures.create(image: image)\n }\n end\n format.html { redirect_to @apartment, notice: 'Propiedad creada correctamente.' }\n format.json { render :show, status: :created, location: @apartment }\n else\n format.html { render :new }\n format.json { render json: @apartment.errors, status: :unprocessable_entity }\n end\n end\n end",
"def postEntityImageBy_url( entity_id, image_url, image_name)\n params = Hash.new\n params['entity_id'] = entity_id\n params['image_url'] = image_url\n params['image_name'] = image_name\n return doCurl(\"post\",\"/entity/image/by_url\",params)\n end",
"def create\n @spot = Spot.new(spot_params)\n\n respond_to do |format|\n if @spot.save\n if params[:photos]\n params[:photos]['image'].each do |a|\n @photo = @spot.photos.create!(:image => a, :imageable_id => @spot.id)\n end\n end\n format.html { redirect_to [@country,@spot], notice: \"Spot was successfully created.\" }\n else\n format.html { render :new }\n end\n end\n end",
"def create\n @spot = Spot.new(spot_params)\n\n respond_to do |format|\n if @spot.save\n if params[:photos]\n params[:photos]['image'].each do |a|\n @photo = @spot.photos.create!(:image => a, :imageable_id => @spot.id)\n end\n end\n format.html { redirect_to [@country,@spot], notice: \"Spot was successfully created.\" }\n else\n format.html { render :new }\n end\n end\n end",
"def create\n respond_to do |format|\n @pix.images.attach(params[:images]) if params[:images]\n if @pix.save\n format.html do\n redirect_to go_to_pix(@pix),\n flash: { success: 'Pix was successfully saved.' }\n end\n format.json { render :show, status: :created, location: @pix }\n else\n format.html { render :new }\n format.json { render json: @pix.errors, status: :unprocessable_entity }\n end\n end\n end",
"def parse_sw_image_ids\n public_xml_doc.xpath('//resource[@type=\"page\" or @type=\"image\" or @type=\"thumb\"]').map do |node|\n node.xpath('./file[@mimetype=\"image/jp2\"]/@id').map do |x|\n \"#{@druid.gsub('druid:', '')}/\" + x\n end << node.xpath('./externalFile[@mimetype=\"image/jp2\"]').map do |y|\n \"#{y.attributes['objectId'].text.split(':').last}\" + '/' + \"#{y.attributes['fileId']}\"\n end\n end.flatten\n end",
"def upload_new_image_file(detection_flags, image = {})\n @client.post \"/service_json_ssl.svc/UploadNewImage_File\", {detection_flags:detection_flags,imagefile_data:image[:data],original_filename:image[:original_filename]}\n end",
"def upload_referenced_images(directory,filename,locale)\n version = Settings[:CURRENT_VERSION]\n begin\n doc_path = \"#{directory}/#{filename}\"\n relative_directory = File.dirname(doc_path) \n content_doc = Nokogiri::XML(open(doc_path)){|config| config.noent }\n content_doc.remove_namespaces!\n # Find each img element\n content_doc.xpath('//img').each do |img|\n # Steps for uploading content\n # 1. Create a hash of the file\n # 2. Get a unique path.\n # 3. Get the filename of the referenced document for the\n # attachment name\n # 4. Check to see if that document exists, if it does, compare\n # the hashes and only upload if it has changed.\n\n # If the image starts with a / assume the file will be in the public directory\n unless (img['src'].start_with?('/'))\n mime_type = get_mime_type(img['src'][/(?:.*)(\\..*$)/, 1])\n\n # Get the directory from the filename\n dir_match_re = /(.*)\\//\n file_dir = dir_match_re.match(filename)[1]\n\n # Fix relative paths here\n path_from_source = (Pathname.new img['src'])\n image_path = (Pathname.new(\"#{file_dir}/#{img['src']}\")).cleanpath.to_s\n id = \"#{Settings[:APP_NAME]}.#{image_path}.#{locale}.#{version}\"\n \n full_image_path = (Pathname.new(\"#{directory}/#{image_path}\")).cleanpath.to_s\n\n # Get the hash of the file on the filesystem\n np = Digest::MD5.file(full_image_path)\n attachment_hash = \"md5-#{Base64.encode64(np.digest)}\".strip\n \n # Look at the attachments on the document in the database\n # If there is an existing attachment with the same name, check the hash value.\n # If it's the same, don't upload it.\n # If it's different, upload it.\n\n #doc fields to create, if needed\n newDoc = {\n '_id' => id,\n :name => image_path,\n :locale => locale,\n :version => version,\n :updated => Time.now.to_date.iso8601,\n :content_hash => attachment_hash }\n \n #doc fields to update, if needed\n updatedDoc = {\n :updated => Time.now.to_date.iso8601,\n :content_hash => attachment_hash }\n \n #upsert the document\n upsert_doc(@db, newDoc, updatedDoc, :content_hash, @log)\n \n doc = @db.get(id)\n doc_attachments = JSON.parse(doc.to_json)[\"_attachments\"]\n\n # If there are no attachments, then doc_attachments will be Nil\n if (doc_attachments.is_a? Hash)\n # If there is already an attachment with the same name, check the hash.\n # If the hash is different, update it.\n unless (doc_attachments.has_key?(image_path) && doc_attachments[image_path][\"digest\"].eql?(attachment_hash))\n begin\n @db.put_attachment(doc, image_path, open(full_image_path, &:read), :content_type => mime_type)\n rescue RestClient::Conflict\n @log.warn \"Hit a conflict. Deleting the attachment and trying again.\"\n begin\n @db.delete_attachment(doc,image_path,true)\n begin\n # Have to get the document again, since the _rev has changed\n doc = @db.get(id)\n @db.put_attachment(doc, image_path, open(full_image_path, &:read), :content_type => mime_type)\n rescue => e\n @log.error\"The attachment was deleted, but could not be re-added.\"\n @log.error e.class\n @log.error e.message\n end\n rescue => e\n @log.warn \"Something went wrong when deleting the attachment. Unknown state.\"\n @log.error e.class\n @log.error e.message\n end\n rescue => e\n @log.error \"Something went wrong when adding an attachment - #{img['src']} on #{doc_path}\"\n @log.error e.message\n @log.error e.class\n end\n end\n else\n # There are no attachments on this document. Add this one.\n @db.put_attachment(doc, image_path, open(full_image_path, &:read), :content_type => mime_type)\n end\n end\n end\n rescue => e\n# @log.error \"Something went wrong when adding an attachment - #{img['src']} on #{doc_path}\"\n @log.error e.message\n @log.error e.class\n end\n end",
"def create\n @interest_point = InterestPoint.new(interest_point_params)\n\n respond_to do |format|\n if @interest_point.save\n @interest_point.image_url = rails_blob_path(@interest_point.point_image, only_path: true) if @interest_point.point_image.attached?\n @interest_point.save\n format.html { redirect_to interest_points_path, notice: 'Interest point was successfully created.' }\n format.json { render :index, status: :created, location: @interest_point }\n else\n format.html { render :new }\n format.json { render json: @interest_point.errors, status: :unprocessable_entity }\n end\n end\n end",
"def images\n end",
"def postEntityImage( entity_id, filedata, image_name)\n params = Hash.new\n params['entity_id'] = entity_id\n params['filedata'] = filedata\n params['image_name'] = image_name\n return doCurl(\"post\",\"/entity/image\",params)\n end",
"def create_image_features(request)\n http_request = request.to_http_info(@api_client.config)\n make_request(http_request, :POST, nil)\n nil\n end",
"def add_image(image_node)\n\t nodeset = self.find_by_terms(:vra)\n\t image_node.first.add_namespace_definition(\"vra\",\"http://www.vraweb.org/vracore4.htm\")\n\t nodeset.first.add_child(image_node)\n self.content = self.ng_xml.to_s\n return nodeset\n end",
"def create\n #debugger\n blog = blog_params[:name]\n @user = User.find(session[:user_id])\n\n params[:image][:attached_assets_attrs].each do |item|\n ImgGroup.post item[:asset], @user, blog\n end\n\n respond_to do |format|\n format.html { render action: 'done' }\n format.json { render text: 'All posted.', status: :created, location: @image }\n end\n end",
"def register_new_image(image, xml_node, data_item)\n image.set_path(data_item)\n image.is_new = true\n image.id, image.target = @relationship_manager.new_relationship(:image, image.path)\n image.replace_new_image_id!(xml_node, data_item)\n @images << image\n image\n end",
"def create\n @post = Post.new(params[:post])\n @post.user_id = session[:user_id]\n \n respond_to do |format|\n if @post.save\n flash[:notice] = 'Post was successfully created.'\n \n if params[:images]\n counter = 0\n for image in params[:images][\"file_data\"]\n if image != \"\"\n @image = Image.create(:file_data => image, :owner_id => @post.id, :owner_type => 'Post', :filename => image.original_filename, \n :body => params[:images][\"body\"][counter] )\n counter = counter + 1\n end\n end\n end\n \n format.html { redirect_to :controller => \"posts\" }\n format.xml { head :created, :location => post_url(:id => @post) }\n else\n format.html { render :action => \"new\" }\n format.xml { render :xml => @post.errors.to_xml }\n end\n end\n end",
"def postToolsImage( filedata, type)\n params = Hash.new\n params['filedata'] = filedata\n params['type'] = type\n return doCurl(\"post\",\"/tools/image\",params)\n end",
"def embed_images(image_path)\n @data['images'] ||= {}\n images = @data['notes'].map { |note| note['image'] }.uniq.sort\n images.each do |image|\n svg = Nokogiri::XML File.read(File.join(image_path, image + '.svg'))\n @data['images'][image] = process_note_svg svg, image\n end\n end",
"def image_params\n params.require(:image).permit(:url, :image_set_id)\n end",
"def image_params\n params.require(:image).permit(:image_url, :tagline, :created_date)\n end",
"def post_file_to_server id, content, size, page_count, extension = \".jpg\"\n\n @s.execute_file_post @s.url_for(\"system/pool/createfile.#{id}.page#{page_count}-#{size}\"), \"thumbnail\", \"thumbnail\", content, related_mimetype(extension)\n alt_url = @s.url_for(\"p/#{id}/page#{page_count}.#{size}\" + extension)\n @s.execute_post alt_url, {\"sakai:excludeSearch\" => true}\n log \"Uploaded image to curl #{alt_url}\"\nend",
"def create\n @results = []\n\n unless params[:files].nil?\n params[:files].each do |data|\n img = Image.new\n img.filename = data.original_filename\n img.data = data.read\n img.upload_id = params[:uuid]\n img.visitation_form_id = params[:formId]\n img.image_type = params[:imageType]\n img.content_type = data.content_type\n #img.temp_index = params[:birdIndex]\n img.bird_id = params[:birdId]\n\n if !img.save\n render :json => { :errors => img.errors.full_messages }, :status => 400 and return\n else\n @results << { name: img.filename, imageType: img.image_type, id: img.id }\n end\n end\n end\n\n render json: { files: @results }\n end",
"def image_params\n params.require(:inspection).permit(:image => [])\n end",
"def images_attributes=(image_attributes)\n image_attributes.each do |attributes| \n post_images.build(attributes) \n end\n end",
"def post(path = '/files/', params = {})\n request :post, path, params\n end",
"def create\n # @taxi_photo = TaxiPhoto.new(taxi_photo_params)\n\n if params[:image]\n puts params[:image]\n params[:image].each { |image|\n @tour_photo = TourPhoto.new(image:image)\n @tour_photo.save\n }\n \n end\n redirect_to \"/tour_photos/gallery\"\n end",
"def upload_images_and_documents\n if params[:project][:images]\n params[:project][:images].each do |project_image|\n @project.images.create(image: project_image, document: false)\n end\n end\n\n if params[:project][:legal_documents]\n params[:project][:legal_documents].each do |project_doc|\n @project.images.create(image: project_doc, document: true)\n end\n end\n end",
"def add_images\n\t\tif @current_user.present?\n\t\t\t@property = Property.find(params[:property_id])\n\t\t\tif @property.present?\n\t\t\t\t# if @property.images.present?\n\t\t\t\t# \t@property.images.destroy_all\n\t\t\t\t# end\n\t\t\t\tparams[:images].each { |image|\n\t i = @property.images.create(image: image)\n\t if i.save\n\t else\n\t \trender_json({\"status\" => \"Fail\", \"message\" => i.errors.full_messages.first}.to_json)\n\t \treturn\n\t end\n\t }\n\t @property.images.first.update_attributes(is_starred: true)\n\t render :file => 'api/v1/property/add_image'\n\t else\n\t \trender_json({\"status\" => \"Fail\", \"message\" => \"No property found.\"}.to_json)\n\t end\n\t\tend\n\tend",
"def saveImagesFromUrls(urls)\n urls.each_with_index do |image,idx|\n name = imageNameFromUrl(image.uri, idx)\n puts \"saving... \" + name\n ImageFinder.saveImage(name,image.uri)\n end \n end",
"def append_image(image)\n image_ele = @image_ele.clone\n image_ele.at(\".//v:shape\").attributes[\"style\"].value = \"width:#{image.width}pt;height:#{image.height}pt\"\n image_ele.at(\".//v:imagedata\").attributes[\"id\"].value = \"rId#{@rid_index}\"\n image_rel = @image_rel.clone\n image_rel.attributes[\"Id\"].value = \"rId#{@rid_index}\"\n image_rel.attributes[\"Target\"].value = \"media/image#{@rid_index}.#{image.file_type}\"\n FileUtils.cp(image.file_name, \"public/downloads/#{@filename}/blank_doc/word/media/image#{@rid_index}.#{image.file_type}\")\n @rel_doc.at('.//xmlns:Relationships').add_child(image_rel)\n @rid_index += 1\n\n last_para = @main_doc.xpath('.//w:p')[-1]\n last_para.add_child(image_ele)\n end",
"def create\n @image = Image.new(image_params)\n\n # fetch tags from google vision API\n helpers.fetch_tags(image_params)\n\n @image.image_file.attach(image_params[:image_file])\n\n respond_to do |format|\n if @image.save()\n format.html { redirect_to @image, notice: \"Image was successfully created.\" }\n format.json { render :show, status: :created, location: @image }\n else\n format.html { render :new, status: :unprocessable_entity }\n format.json { render json: @image.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @taxi_image = TaxiImage.new(taxi_image_params)\n\n respond_to do |format|\n if @taxi_image.save\n format.html { redirect_to @taxi_image, notice: 'Taxi image was successfully created.' }\n format.json { render :show, status: :created, location: @taxi_image }\n else\n format.html { render :new }\n format.json { render json: @taxi_image.errors, status: :unprocessable_entity }\n end\n end\n end",
"def upload_new_image_url(detection_flags, image = {})\n params = {detection_flags:detection_flags}\n params[:url] = image[:url] if image[:url].present?\n params[:original_filename] = image[:url].split(\"/\")[-1] if image[:url].present? and image[:original_filename].blank?\n @client.post \"/service_json_ssl.svc/UploadNewImage_Url\", params\n end",
"def post_storage(request)\n # --- Check OCCI XML from POST ---\n if request.params['occixml'] == nil\n error_msg = \"OCCI XML representation of Image\" +\n \" not present in the request\"\n error = OpenNebula::Error.new(error_msg)\n return error, 400\n end\n\n # --- Create and Add the new Image ---\n occixml = request.params['occixml']\n occixml = occixml[:tempfile].read if occixml.class == Hash\n\n image = ImageOCCI.new(\n Image.build_xml,\n @client,\n occixml,\n request.params['file'])\n\n # --- Generate the template and Allocate the new Instance ---\n template = image.to_one_template\n return template, 500 if OpenNebula.is_error?(template)\n\n rc = image.allocate(template, @config[:datastore_id]||1)\n if OpenNebula.is_error?(rc)\n return rc, CloudServer::HTTP_ERROR_CODE[rc.errno]\n end\n\n image.info\n #wait until image is ready to return\n while (image.state_str == 'LOCKED') && (image['RUNNING_VMS'] == '0') do\n sleep IMAGE_POLL_SLEEP_TIME\n image.info\n end\n\n # --- Prepare XML Response ---\n return to_occi_xml(image, :code=>201)\n end",
"def create\n request_image = params[:image]\n url_hash = SecureRandom.urlsafe_base64(6)\n name = url_hash + \"-\" + request_image.original_filename.downcase\n\n File.open(\"#{Rails.root}/public/theta/#{name}\", 'wb') do |f|\n f.write(request_image.read)\n end\n\n theta = Theta.create!(url_hash: url_hash, image_url: name)\n render json: {url: \"http://www.rakugaki.tk/h/#{theta.url_hash}\"}, status: 201\n end",
"def getimagesinfo\n trek = Trek.find_by_id(params[:id])\n send_data(trek.get_images_info.to_json,\n {:type => \"application/json\", :disposition => \"inline\"})\n end",
"def loc_image_params\n params.require(:loc_image).permit(:location_id, :image_url)\n end",
"def create\n @smartphone = Smartphone.new(smartphone_params)\n\n respond_to do |format|\n if @smartphone.save\n if params[:images]\n params[:images].each do |val|\n @smartphone.pictures.create(image: val)\n end\n end\n\n format.html { redirect_to @smartphone, notice: 'Smartphone was successfully created.' }\n format.json { render :show, status: :created, location: @smartphone }\n else\n format.html { render :new }\n format.json { render json: @smartphone.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n\n\n @img_info = ImgInfo.new \n @img_info.img_url=upload_file(params[:img_info]['imgdata']);\n \n respond_to do |format|\n if @img_info.save\n flash[:notice] = 'ImgInfo was successfully created.'\n format.html { redirect_to(@img_info) }\n format.xml { render :xml => @img_info, :status => :created, :location => @img_info }\n else\n format.html { render :action => \"new\" }\n format.xml { render :xml => @img_info.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @tour.update(tour_params)\n\n if params[:images]\n params[:images].each { |image|\n @tour.tour_days.create(images: image)\n }\n end\n\n format.html { redirect_to @tour, notice: 'Tour was successfully updated.' }\n format.json { render :show, status: :ok, location: @tour }\n else\n format.html { render :edit }\n format.json { render json: @tour.errors, status: :unprocessable_entity }\n end\n end\n end",
"def image_params\n params.require(:image).permit(:file, :description, :user_id, :location_id)\n end",
"def create\n @document = Document.new(document_params)\n # @document.images.attach(params[:document][:images])\n respond_to do |format|\n if @document.save\n @records = Document.all.with_attached_images\n\n format.html { redirect_to @document, notice: 'Document was successfully created.' }\n format.json { render :show, status: :created, location: @document }\n else\n format.html { render :new }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def generated_image_params\n params.require(:generated_image).permit(:image, :content_image_id, :style_image_id, :iterations)\n end",
"def save_image(data)\n\t\timages = []\n\n\t\tdata[\"photos\"].each do |image|\n\t\t\tset = {\n\t\t\t\t:sol => image['sol'],\n\t\t\t\t:img_src => image['img_src'],\n\t\t\t\t:rover => image['rover']['name'].downcase,\n\t\t\t\t:camera_name => image['camera']['name'],\n\t\t\t\t:camera_full_name => image['camera']['full_name'],\n\t\t\t\t:earth_date => image['earth_date'],\n\t\t\t}\n\n\t \t\t#save to database\n\t \t\timg = Image.create(set)\n\t \t\timg.save\n\n\t \t\timages.push(set)\n\t\tend\n\n\t\treturn images\n \tend",
"def add_image\n pass_query_params\n @observation = find_or_goto_index(Observation, params[:id].to_s)\n return unless @observation\n\n if !check_permission!(@observation)\n redirect_with_query(controller: \"observer\",\n action: \"show_observation\", id: @observation.id)\n elsif request.method != \"POST\"\n @image = Image.new\n @image.license = @user.license\n @image.copyright_holder = @user.legal_name\n @image.user = @user\n # Set the default date to the date of the observation\n # Don't know how to correctly test this.\n @image.when = @observation.when\n @licenses = License.current_names_and_ids(@image.license)\n init_project_vars_for_add_or_edit(@observation)\n elsif params[:upload].blank?\n flash_warning(:runtime_no_changes.t)\n redirect_with_query(controller: \"observer\",\n action: \"show_observation\", id: @observation.id)\n else\n args = params[:image]\n i = 1\n while i < 5 || params[:upload][\"image#{i}\"].present?\n process_image(args, params[:upload][\"image#{i}\"])\n i += 1\n end\n redirect_with_query(controller: \"observer\",\n action: \"show_observation\", id: @observation.id)\n end\n end",
"def pix_params\n params.require(:pix).permit(:character_id, :scenario_id, :msg, images: [])\n end",
"def post_paths\n %(/site/home\n /content/999/BOXID/TYPENAME\n /file/999/BOXID\n /shard/0000000000000000000000000000000000000000\n /schema/delete\n )\n end",
"def update\n respond_to do |format|\n if @rent.update(rent_params)\n\n if params[:image]\n puts params[:image]\n params[:image].each { |image|\n @rent.rent_images.create(rent_id: @rent.id, image:image)\n }\n \n end\n \n format.html { redirect_to @rent, notice: 'Rent was successfully updated.' }\n format.json { render :show, status: :ok, location: @rent }\n else\n format.html { render :edit }\n format.json { render json: @rent.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @position = Position.new(position_params)\n respond_to do |format|\n if @position.save\n @position.create_images(params[:images]) if params[:images]\n format.html { redirect_to @position, notice: 'Position was successfully created.' }\n format.json { render :show, status: :created, location: @position }\n else\n format.html { render :new }\n format.json { render json: @position.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create_image_files_where_needed()\n @file_info.data.each do |line|\n uri, filename = line\n process_file_info(uri, filename)\n end\n end",
"def process_images(pat, sp_name)\n sp = ServicePack.find_or_create_by_name(sp_name)\n images = ImagePath.find(:all, :conditions => \"path like '#{pat}'\")\n length = images.length\n images.each_with_index do |image, index|\n # puts \"processing image #{index} of #{length}\"\n image.package.filesets.each do |fileset|\n ServicePackFilesetMap.find_or_create_by_service_pack_id_and_fileset_id(sp.id, fileset.id)\n end\n end\nend",
"def drawPathWithPoints(points, image:image)\n screenSize = self.view.frame.size\n UIGraphicsBeginImageContext(screenSize)\n context = UIGraphicsGetCurrentContext()\n image.drawInRect(CGRectMake(0, 0, screenSize.width, screenSize.height))\n\n CGContextSetLineCap(context, KCGLineCapRound)\n CGContextSetLineWidth(context, LINE_WIDTH)\n CGContextSetRGBStrokeColor(context, 0, 0, 1, 1)\n CGContextBeginPath(context)\n\n count = points.count\n point = points[0].CGPointValue\n CGContextMoveToPoint(context, point.x, point.y)\n 1.upto(count - 1) do |i|\n point = points.objectAtIndex(i).CGPointValue\n CGContextAddLineToPoint(context, point.x, point.y)\n end\n CGContextStrokePath(context)\n\n ret = UIGraphicsGetImageFromCurrentImageContext()\n UIGraphicsEndImageContext()\n ret\n end",
"def image hash = {}\n hash = { :id => hash} if hash.is_a? String\n home hash.update :action => 'image', :trailing_slash => false\n end",
"def image_params\n params.require(:images).permit(:file)\n end",
"def image_params\n params.permit(:image, :build_id, :test_id, :approved, :test_image_ancestry)\n end",
"def new\n @image_path = ImagePath.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @image_path }\n end\n end",
"def image_params\n params.require(:image).permit(:file, :tag_list)\n end",
"def post_insert_image_with_http_info(name, page_number, llx, lly, urx, ury, opts = {})\n if @api_client.config.debugging\n @api_client.config.logger.debug \"Calling API: PdfApi.post_insert_image ...\"\n end\n # verify the required parameter 'name' is set\n if @api_client.config.client_side_validation && name.nil?\n fail ArgumentError, \"Missing the required parameter 'name' when calling PdfApi.post_insert_image\"\n end\n # verify the required parameter 'page_number' is set\n if @api_client.config.client_side_validation && page_number.nil?\n fail ArgumentError, \"Missing the required parameter 'page_number' when calling PdfApi.post_insert_image\"\n end\n # verify the required parameter 'llx' is set\n if @api_client.config.client_side_validation && llx.nil?\n fail ArgumentError, \"Missing the required parameter 'llx' when calling PdfApi.post_insert_image\"\n end\n # verify the required parameter 'lly' is set\n if @api_client.config.client_side_validation && lly.nil?\n fail ArgumentError, \"Missing the required parameter 'lly' when calling PdfApi.post_insert_image\"\n end\n # verify the required parameter 'urx' is set\n if @api_client.config.client_side_validation && urx.nil?\n fail ArgumentError, \"Missing the required parameter 'urx' when calling PdfApi.post_insert_image\"\n end\n # verify the required parameter 'ury' is set\n if @api_client.config.client_side_validation && ury.nil?\n fail ArgumentError, \"Missing the required parameter 'ury' when calling PdfApi.post_insert_image\"\n end\n # resource path\n local_var_path = \"/pdf/{name}/pages/{pageNumber}/images\".sub('{' + 'name' + '}', name.to_s).sub('{' + 'pageNumber' + '}', page_number.to_s)\n\n # query parameters\n query_params = {}\n query_params[:'llx'] = llx\n query_params[:'lly'] = lly\n query_params[:'urx'] = urx\n query_params[:'ury'] = ury\n query_params[:'imageFilePath'] = opts[:'image_file_path'] if !opts[:'image_file_path'].nil?\n query_params[:'storage'] = opts[:'storage'] if !opts[:'storage'].nil?\n query_params[:'folder'] = opts[:'folder'] if !opts[:'folder'].nil?\n\n # header parameters\n header_params = {}\n # HTTP header 'Accept' (if needed)\n header_params['Accept'] = @api_client.select_header_accept(['application/json'])\n # HTTP header 'Content-Type'\n header_params['Content-Type'] = @api_client.select_header_content_type(['multipart/form-data'])\n\n # form parameters\n form_params = {}\n # Fix header in file\n post_body = nil\n post_body = opts[:'image'] if !opts[:'image'].nil?\n header_params['Content-Type'] = @api_client.select_header_content_type(['application/octet-stream'])\n\n # http body (model)\n # Fix header in file\n # post_body = nil\n auth_names = ['JWT']\n data, status_code, headers = @api_client.call_api(:POST, local_var_path,\n :header_params => header_params,\n :query_params => query_params,\n :form_params => form_params,\n :body => post_body,\n :auth_names => auth_names,\n :return_type => 'AsposeResponse')\n if @api_client.config.debugging\n @api_client.config.logger.debug \"API called: PdfApi#post_insert_image\\nData: #{data.inspect}\\nStatus code: #{status_code}\\nHeaders: #{headers}\"\n end\n return data, status_code, headers\n end",
"def bulk_upload_images\n submissions = []\n params[:file].each do |key, image|\n registration_params = permitted_params.merge(image: image)\n \n @registration = Registration.new(registration_params)\n\n if @registration.save\n submissions << \"Image number #{key.to_i + 1}, upload sucessful\"\n else\n submissions << \"Image number #{key.to_i + 1}, upload unsucessful: #{@registration.errors.messages}\"\n end\n end\n submissions\n end",
"def create\n\n require 'RMagick'\n @images_added = []\n\n # params[:design_images] is set by a js library (Dropzone.js) and defined on dynamic_pages.js paramName\n params[:design_images].each do |array_file|\n\n # uploaded file example\n # {\"0\"=>#<ActionDispatch::Http::UploadedFile:0x007f88cf70b728 @tempfile=#<Tempfile:/var/folders/tx/fcpjq87x6b19sl3dvvm8fr040000gn/T/RackMultipart20150803-99318-18nmoiu.jpg>, @original_filename=\"10881656_10201963394846606_706439852029346946_n.jpg\", @content_type=\"image/jpeg\", @headers=\"Content-Disposition: form-data; name=\\\"singular_design[0]\\\"; filename=\\\"10881656_10201963394846606_706439852029346946_n.jpg\\\"\\r\\nContent-Type: image/jpeg\\r\\n\">, \"1\"=>#<ActionDispatch::Http::UploadedFile:0x007f88cf70b610 @tempfile=#<Tempfile:/var/folders/tx/fcpjq87x6b19sl3dvvm8fr040000gn/T/RackMultipart20150803-99318-1loctd5.jpg>, @original_filename=\"11062261_825163280904553_6934851262428256525_n.jpg\", @content_type=\"image/jpeg\", @headers=\"Content-Disposition: form-data; name=\\\"singular_design[1]\\\"; filename=\\\"11062261_825163280904553_6934851262428256525_n.jpg\\\"\\r\\nContent-Type: image/jpeg\\r\\n\">}\n\n uploaded_file = array_file.last\n\n # file checking (size and filetype)\n return false if !valid_img? uploaded_file\n\n # main image random filename\n random_filename = build_filename(uploaded_file)\n\n image = Magick::Image.from_blob(uploaded_file.read).first\n original_width = image.columns # columns are pixels\n original_height = image.rows # rows are pixels\n\n # just resize if image is larger in width\n image = image.resize_to_fit(2000, original_height) if original_width >= 2000\n image.write(Rails.root.join('public', 'assets', 'uploads','designs', random_filename))\n\n # thumbnail image random filename\n thumb_random_filename = build_filename(uploaded_file, true)\n thumbnail = image.resize_to_fill(250, 150)\n thumbnail.write(Rails.root.join('public', 'assets', 'uploads','designs', thumb_random_filename))\n\n # generating link for this new design\n random_link = SecureRandom.hex(4)\n\n while Design.exists?(link: random_link) == true do random_link = SecureRandom.hex(4) end\n\n @design = Design.new(user_id: current_user.id,\n link: random_link,\n image_path: random_filename,\n image_thumb_path: thumb_random_filename)\n\n @images_added << @design if @design.save\n end\n\n respond_to do |format|\n format.html { redirect_to :individuais }\n format.json { render json: @images_added }\n end\n end",
"def get_images_data\t\t\t\t\t\t\n\t\t{ payload: { 'makes' => makes_dropdown, 'images' => images_array('none').compact}, success: true}\n\tend",
"def zones_xml(image_url, zone_list=nil, url=nil)\n xml = Builder::XmlMarkup.new(:indent => 2)\n xml.dctl_ext_init{\n xml.img{\n xml.a(:r => self.id.to_s, :s => self.uri.to_s, :l => self.name, :u => image_url)\n }\n xml.xml{\n self.zones.each do |z|\n add_zone_to_xml(z, xml, self.uri.to_s, zone_list)\n end\n }\n #xml.cb(:u => \"/admin/images/update/\", :p => \"base64xml\")\n xml.cb(:u => url, :p => \"base64xml\")\n }\n base64 = Base64.encode64(xml.target!)\n # By default it splits up the base64 with \\n, strip them!\n base64.gsub(/\\s/, '')\n end",
"def upload_referenced_images(filename, mime_type, nokodoc, locale)\n nokodoc.xpath(\"//img/@src\").each do |image|\n # portal_images are in the public directory\n unless (image.text.include? '/portal_images')\n begin\n @original_filename = filename\n case\n when (image.text =~ /^[^\\/].*/)\n mydirectory = @original_filename.match(/([^\\/]*)\\/(.*)/)[1]\n mypath = \"#{DOCSRCDIR}#{mydirectory}/#{image.text}\"\n fullpath = File.expand_path(mypath)\n begin\n mime_type = get_mime_type(fullpath[/(?:.*)(\\..*$)/, 1])\n rescue\n @mime_type = \"\"\n end\n begin\n if File.exist?(fullpath)\n then\n upload_attachment(filename,locale,fullpath,mime_type,image.text)\n else\n STDERR.puts \"Failed to find the file\\n\\t #{fullpath}\\n referenced by \\n\\t #{@original_filename}\"\n end\n rescue\n STDERR.puts \"Failed to upload file\\n\\t #{fullpath}\\n referenced by \\n\\t #{@original_filename}\"\n end\n when (image.text =~ /^[\\/].*/)\n fullpath = \"#{APPSRCDIR}#{image.text}\"\n begin\n mime_type = get_mime_type(fullpath[/(?:.*)(\\..*$)/, 1])\n rescue\n @mime_type = \"\"\n end\n begin\n if File.exist?(fullpath)\n then\n upload_attachment('app_image_document',locale,fullpath,mime_type,image.text)\n else\n STDERR.puts \"APPSRCDIR -> #{APPSRCDIR}\"\n STDERR.puts \"image.text -> #{image.text}\"\n STDERR.puts \"Failed to find the file\\n\\t #{fullpath}\\n referenced by \\n\\t #{@original_filename}\"\n end\n rescue\n STDERR.puts \"Failed to upload linked file\\n\\t #{fullpath}\\n referenced by \\n\\t #{@original_filename}\"\n end\n else\n STDERR.puts \"Failed to upload file\\n\\t #{fullpath}\\n referenced by \\n\\t #{@original_filename}\"\n end\n rescue\n STDERR.puts \"File\\n\\t #{image.text}\\n referenced by \\n\\t #{@original_filename} is not formatted in a way I understand.\"\n end\n end\n end\nend",
"def upload_url_product_image(rid, add_params = nil)\n params = {\n uid: uid,\n rid: rid,\n }\n api_call('/stores/:uid/products/:rid/images/upload_url(.:format)',:post,params,add_params)\n end",
"def create\n @commission = Commission.new(commission_params)\n respond_to do |format|\n if @commission.save\n if params[:commission_images]\n params[:commission_images].each { |commission_image| @commission.commission_images.create(path: commission_image) }\n end\n format.html { redirect_to @commission, notice: 'Commission was successfully created.' }\n format.json { render :show, status: :created, location: @commission }\n else\n format.html { render :new }\n format.json { render json: @commission.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @periodic = Periodic.new(periodic_params)\n\n if @periodic.images.count>0\n @periodic.images.attach(params[:periodic][:images])\n end\n\n respond_to do |format|\n if @periodic.save\n format.html { redirect_to @periodic, notice: 'Periodic was successfully created.' }\n format.json { render :show, status: :created, location: @periodic }\n else\n format.html { render :new }\n format.json { render json: @periodic.errors, status: :unprocessable_entity }\n end\n end\n end",
"def image_params\n params.require(:image).permit(:name, :url, :image_type, :parent_id)\n end",
"def create\n @location = Location.new(location_params)\n\n respond_to do |format|\n if @location.save\n \n if params[:avatars]\n #===== The magic is here ;)\n params[:avatars].each { |avatar|\n @location.assets.create(avatar: avatar)\n }\n end\n\n format.html { redirect_to @location, notice: 'Gallery was successfully created.' }\n format.json { render json: @location, status: :created, location: @location }\n else\n format.html { render action: \"new\" }\n format.json { render json: @location.errors, status: :unprocessable_entity }\n end\n end\nend",
"def image_params\n params.require(:image).permit({files: []}, :file)\n end",
"def add_image\n obtain_product_image_params\n pi = ProductImage.new(picture: @image_params)\n @product.product_images << pi\n render json: @product.simple_info, status: :ok\n rescue => e\n render json: { error: e }, status: :bad_request\n end",
"def create\n @provider = Provider.new(provider_params)\n @provider.user_id = current_user.id\n\nif @provider.save\n\n if params[:images_p] \n params[:images_p].each do |image|\n @provider.provider_attachments.create(image2: image)\n end\n end\n\n @provider_attachments = @provider.provider_attachments\n redirect_to edit_provider_path(@provider), notice: \"Saved...\"\n else\n render :new\n end\n \n end",
"def post_params\n params.require(:post).permit(:title, :description, images: [], files: [], post_locations: [:lat, :lng, :location_id])\n end",
"def create\n @map_image = MapImage.new(params[:map_image])\n\n respond_to do |format|\n if @map_image.save\n format.html { redirect_to(@map_image, :notice => 'Map image was successfully created.') }\n format.xml { render :xml => @map_image, :status => :created, :location => @map_image }\n else\n format.html { render :action => \"new\" }\n format.xml { render :xml => @map_image.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def create_updated_image(request)\n http_request = request.to_http_info(@api_client.config)\n make_request(http_request, :POST, 'File')\n end",
"def image_params\n params.require(:image).permit(:file_path, :dataset_id, :user_id)\n end",
"def create\n @image_attrib = ImageAttrib.new(image_attrib_params)\n\n if @image_attrib.save\n render :show, status: :created, location: @image_attrib\n else\n render json: @image_attrib.errors, status: :unprocessable_entity\n end\n end",
"def create\n @driver = Driver.new(driver_params)\n\n if @driver.images.count>0\n @driver.images.attach(params[:driver][:images])\n end\n\n respond_to do |format|\n if @driver.save\n format.html { redirect_to @driver, notice: 'Driver was successfully created.' }\n format.json { render :show, status: :created, location: @driver }\n else\n format.html { render :new }\n format.json { render json: @driver.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @image = Image.new(image_params)\n respond_to do |format|\n if @image.save\n format.json { render :json => { url: @image.image.url} }\n else\n \n format.json { render json: @image.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @image_set = ImageSet.new(params[:image_set])\n\n respond_to do |format|\n if @image_set.save\n flash[:notice] = 'ImageSet was successfully created.'\n format.html { redirect_to(@image_set) }\n format.xml { render :xml => @image_set, :status => :created, :location => @image_set }\n else\n format.html { render :action => \"new\" }\n format.xml { render :xml => @image_set.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def create\n\n @event = Event.new(event_params)\n\n tag_array = event_params2['tags'].split(',')\n tag_array.each do |x|\n @tag = Tag.create(desc: x)\n @event.tags << @tag\n end\n EventTag.create(event_id: @event.id, tag_id: @tag.id)\n\n respond_to do |format|\n if @event.save\n params[:event]['images'].each do |a|\n @event_photo = @event.event_photos.create!(:image => a, :event_id => @event.id)\n end\n format.html { redirect_to @event, notice: 'Event was successfully created.' }\n format.json { render :show, status: :created, location: @event }\n else\n format.html { render :new }\n format.json { render json: @event.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @spot.update(spot_params)\n if params[:photos]\n params[:photos]['image'].each do |a|\n @photo = @spot.photos.create!(:image => a, :imageable_id => @spot.id)\n end\n end\n\n format.html { redirect_to [ @country, @spot ] , notice: 'Spot was successfully updated.' }\n else\n format.html { render :edit }\n end\n end\n end",
"def update\n respond_to do |format|\n if @spot.update(spot_params)\n if params[:photos]\n params[:photos]['image'].each do |a|\n @photo = @spot.photos.create!(:image => a, :imageable_id => @spot.id)\n end\n end\n\n format.html { redirect_to [ @country, @spot ] , notice: 'Spot was successfully updated.' }\n else\n format.html { render :edit }\n end\n end\n end",
"def create\n @image = Image.new(image_params)\n\n respond_to do |format|\n if @image.save\n format.json { render json: @image, status: :created, location: [:admin, @image] }\n else\n format.json { render json: @image.errors, status: :unprocessable_entity }\n end\n end\n end",
"def images; end",
"def create\r\n @family = Family.new(family_params)\r\n @image = @family.images.new\r\n @image.generate_filename\r\n @uploaded_io = params[:family][:uploaded_file]\r\n\r\n File.open(Rails.root.join('public', 'images', @image.filename), 'wb') do |file|\r\n file.write(@uploaded_io.read)\r\n end\r\n\r\n respond_to do |format|\r\n if @family.save\r\n format.html { redirect_to @family, notice: 'Family was successfully created.' }\r\n format.json { render :show, status: :created, location: @family }\r\n else\r\n format.html { render :new }\r\n format.json { render json: @family.errors, status: :unprocessable_entity }\r\n end\r\n end\r\n end",
"def update_images=(image_attributes) \n image_attributes.each do |attributes| \n post = post_images.detect {|n| n.id == attributes[:id].to_i}\n post.attributes = attributes\n end\n end",
"def create\n\n @drawing = Drawing.new(form_params)\n @drawing.picture = PictureAttachmentHelper.convert_data_uri_to_upload(params[\"picture\"])\n @drawing.save\n redirect_to root_path\n end",
"def export_measurements_image\n image_path(image_name: params[:export_measurements_image])\n end",
"def upload\n if File.exists?(file_path = \"/home/redroid/motion_detection_images/#{params[:img]}.jpg\")\n send_403_forbidden\n else\n File.open(file_path, 'wb') { |file| file.write(request.raw_post)}\n send_200_ok\n end\nend",
"def upload_receipt_pictures(body)\r\n # Prepare query url.\r\n _path_url = '/v2/pos/receipt_pictures'\r\n _query_builder = Configuration.get_base_uri\r\n _query_builder << _path_url\r\n _query_url = APIHelper.clean_url _query_builder\r\n # Prepare headers.\r\n _headers = {\r\n 'accept' => 'application/json',\r\n 'content-type' => 'application/json; charset=utf-8'\r\n }\r\n # Prepare and execute HttpRequest.\r\n _request = @http_client.post(\r\n _query_url,\r\n headers: _headers,\r\n parameters: body.to_json\r\n )\r\n OAuth2.apply(_request)\r\n _context = execute_request(_request)\r\n validate_response(_context)\r\n # Return appropriate response type.\r\n decoded = APIHelper.json_deserialize(_context.response.raw_body)\r\n ReceiptPictureResponse.from_hash(decoded)\r\n end",
"def create\n flash[:notice] = 'Image Created' if @owner.images << @image\n respond_with @owner, @image\n end"
] |
[
"0.6240876",
"0.5513908",
"0.5493714",
"0.5417255",
"0.5392265",
"0.5390093",
"0.53345364",
"0.5307892",
"0.5270956",
"0.5249065",
"0.5232015",
"0.523189",
"0.523189",
"0.5208456",
"0.5186416",
"0.5184379",
"0.5132494",
"0.5111202",
"0.51083183",
"0.510389",
"0.50965774",
"0.50938165",
"0.50932795",
"0.50928193",
"0.50908726",
"0.5086506",
"0.50806457",
"0.50772864",
"0.50760114",
"0.5075767",
"0.5074321",
"0.5071472",
"0.5067981",
"0.5066601",
"0.50651765",
"0.50631785",
"0.50577855",
"0.5055527",
"0.5035298",
"0.50345343",
"0.5028918",
"0.5027698",
"0.5025387",
"0.5021712",
"0.50208193",
"0.50185937",
"0.50036615",
"0.5000427",
"0.4989263",
"0.49746525",
"0.4965295",
"0.49622074",
"0.49593955",
"0.49482358",
"0.49462032",
"0.49273577",
"0.492368",
"0.49193403",
"0.49152952",
"0.4912629",
"0.49125925",
"0.4900126",
"0.4899247",
"0.48970887",
"0.48968226",
"0.48914465",
"0.48875153",
"0.48862126",
"0.4884386",
"0.4882672",
"0.48808026",
"0.4880618",
"0.48792958",
"0.4878458",
"0.48774388",
"0.4875394",
"0.48738265",
"0.4869176",
"0.48549378",
"0.48502663",
"0.48424408",
"0.48421723",
"0.4841919",
"0.48404098",
"0.48380995",
"0.48321655",
"0.48213384",
"0.481367",
"0.4812872",
"0.48077863",
"0.48077863",
"0.4805915",
"0.4803226",
"0.48016468",
"0.48008132",
"0.4798988",
"0.47989815",
"0.4797863",
"0.4796856",
"0.47945854"
] |
0.5596245
|
1
|
PUT /image_paths/1 PUT /image_paths/1.xml
|
def update
@image_path = ImagePath.find(params[:id])
respond_to do |format|
if @image_path.update_attributes(params[:image_path])
flash[:notice] = 'ImagePath was successfully updated.'
format.html { redirect_to(@image_path) }
format.xml { head :ok }
else
format.html { render :action => "edit" }
format.xml { render :xml => @image_path.errors, :status => :unprocessable_entity }
end
end
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def put(path = '/files/', params = {})\n request :put, path, params\n end",
"def update\n\n params = image_params\n params[\"tags\"] = params[\"tags\"].delete_suffix(',')\n\n respond_to do |format|\n if @image.update(params)\n format.html { redirect_to @image, notice: \"Image was successfully updated.\" }\n format.json { render :show, status: :ok, location: @image }\n else\n format.html { render :edit, status: :unprocessable_entity }\n format.json { render json: @image.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n new_images = params[:images]||[]\n\n @intention = Intention.find_by_uuid(params[:id])\n @intention.image_urls = IntentionsHelper.strip_duplicate_images(new_images)\n @intention.save\n\n points = VISBD_INTENTION_IMAGE_POINTS*new_images.count\n track_event(current_action_item(Intention::COMPONENT_TYPE)||current_enrollment, Intention::VISUALIZED_EVENT, target:@intention, points:points)\n\n render nothing: true\n end",
"def update\n respond_to do |format|\n if @interest_point.update(interest_point_params)\n @interest_point.image_url = rails_blob_path(@interest_point.point_image, only_path: true) if @interest_point.point_image.attached?\n @interest_point.save\n format.html { redirect_to interest_points_path, notice: 'Interest point was successfully updated.' }\n format.json { render :index, status: :ok, location: @interest_point }\n else\n format.html { render :edit }\n format.json { render json: @interest_point.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @spot.update(spot_params)\n if params[:photos]\n params[:photos]['image'].each do |a|\n @photo = @spot.photos.create!(:image => a, :imageable_id => @spot.id)\n end\n end\n\n format.html { redirect_to [ @country, @spot ] , notice: 'Spot was successfully updated.' }\n else\n format.html { render :edit }\n end\n end\n end",
"def update\n respond_to do |format|\n if @spot.update(spot_params)\n if params[:photos]\n params[:photos]['image'].each do |a|\n @photo = @spot.photos.create!(:image => a, :imageable_id => @spot.id)\n end\n end\n\n format.html { redirect_to [ @country, @spot ] , notice: 'Spot was successfully updated.' }\n else\n format.html { render :edit }\n end\n end\n end",
"def update\n @image.update(image_params)\n redirect_to '/images'\n end",
"def update\n params[:image].delete :created_at\n params[:image].delete :updated_at\n params[:image].delete :id\n @image = Image.find(params[:id])\n if @image.update_attributes(params[:image])\n render json: @image\n else\n render json: @image.errors, status: :unprocessable_entity\n end\n end",
"def update\n respond_to do |format|\n if @taxi_image.update(taxi_image_params)\n format.html { redirect_to @taxi_image, notice: 'Taxi image was successfully updated.' }\n format.json { render :show, status: :ok, location: @taxi_image }\n else\n format.html { render :edit }\n format.json { render json: @taxi_image.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @image_set = ImageSet.find(params[:id])\n\n respond_to do |format|\n if @image_set.update_attributes(params[:image_set])\n flash[:notice] = 'ImageSet was successfully updated.'\n format.html { redirect_to(@image_set) }\n format.xml { head :ok }\n else\n format.html { render :action => \"edit\" }\n format.xml { render :xml => @image_set.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def update\n @image = Image.find(params[:id])\n\n #we can allow updating an image name and description and unit, but not the image data. for that we need to create a new image\n update_params = image_params\n update_params.delete(\"image\")\n\n if @image.update(update_params)\n head :no_content\n else\n render json: @image.errors, status: :unprocessable_entity\n end\n end",
"def put(path, options={})\n request :put, path, options\n end",
"def update\n @image = Image.find(params[:id])\n unless params[:image][:tags].nil?\n @image.tag_with params[:image]['tags']\n params[:image].delete(\"tags\")\n end\n respond_to do |format|\n if @image.update_attributes(params[:image])\n flash[:notice] = 'Image was successfully updated.'\n format.html { redirect_to([:admin, @image]) }\n format.xml { head :ok }\n else\n format.html { render :action => \"edit\" }\n format.xml { render :xml => @image.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def update\n @img = Img.find(params[:id])\n\n respond_to do |format|\n if @img.update_attributes(params[:img])\n format.html { redirect_to(@img, :notice => 'Img was successfully updated.') }\n format.xml { head :ok }\n else\n format.html { render :action => \"edit\" }\n format.xml { render :xml => @img.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def element\n image_name_list = get_image_name_list ORIGINALS_PATH\n image_name_list.each { |image_name| post(image_name) }\n end",
"def update\n @image = Image.find(params[:id])\n\n respond_to do |format|\n if @image.update_attributes(params[:image])\n flash[:success] = 'Image was successfully updated.'\n format.html { redirect_to([:admin, @parents, :images]) }\n format.xml { head :ok }\n else\n format.html { render :action => \"edit\" }\n format.xml { render :xml => @image.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def update_radios_for_array(args = {}) \n id = args['id']\n temp_path = \"/radios.json/{arrayId}\"\n path = temp_path\nargs.keys.each do |key|\n if (key == \"radioId\")\n args.delete(key)\n path = temp_path.gsub(\"{#{key}}\", id)\n end\nend\n puts \" PATH : #{path}\"\n put(path, args)\nend",
"def update_photo_node(user,images_directory,previous_picture)\n Rails.logger.debug \"Call to photo.update_node\"\n if self.valid? #Validate if the Photo object is valid\n Rails.logger.debug \"The photo is valid!\"\n file = self.file #Set the photo file object\n if !file.blank? #Validate if a file was supplied by the user\n images_directory =images_directory.blank? ? @@images_directory : images_directory #Validate if an image_directory was supplied, otherwise we use the default one\n file_s3_path = Util.upload_image(images_directory,file) #Upload the new image\n if !previous_picture.blank?#Validate if there was a previous image file tied to the photo node\n Util.delete_image(previous_picture) #Delete the previous image file\n end\n else\n file_s3_path = self.file #If none was provided, keep the original file\n end\n #Create a raw photo object\n photo_req = { 'title'=>self.title,\n 'description'=>self.description,\n 'url'=>file_s3_path,\n 'date'=> Util.date_to_epoch(self.date), #Turn the date to epoch\n 'ownerId'=> self.owner_id \n } \n reqUrl = \"/api/photo/#{self.id}\" #Set the request url\n\n rest_response = MwHttpRequest.http_put_request(reqUrl,photo_req,user['email'],user['password']) #Make the PUT call to the server with the required parameters\n Rails.logger.debug \"Response from server: #{rest_response.code} #{rest_response.message}: #{rest_response.body}\"\n if rest_response.code == \"200\" #Validate if the response from the server is 200, which means OK\n photo = Photo.rest_to_photo(rest_response.body)\n return true, photo #Return success\n else\n return false, \"#{rest_response.code}\", \"#{rest_response.message}\" #Return error\n end\n else\n Rails.logger.debug self.errors.full_messages\n return false, self.errors.full_messages #Return invalid object error\n end\n end",
"def put(*args)\n prepare_request(:put, args)\n @@client.add(:put, @path, *args)\n end",
"def update(url, data)\n RestClient.put url, data, :content_type => :json\nend",
"def put(path, **args); end",
"def http_put(path, data, content_type = 'application/json')\n http_methods(path, :put, data, content_type)\n end",
"def update\n @image = Image.find(params[:id])\n\n respond_to do |format|\n if @image.update_attributes(params[:image])\n flash[:notice] = 'Image was successfully updated.'\n format.html { redirect_to(@image) }\n format.xml { head :ok }\n else\n format.html { render :action => \"edit\" }\n format.xml { render :xml => @image.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def update_attributes(attributes = {},user,images_directory)\n Rails.logger.debug \"Call to photo.update_attributes\"\n if self.valid? #Validate if the Photo object is valid\n Rails.logger.debug \"The photo is valid!\"\n file = attributes[:file]#Set the photo file object\n if !file.blank? #Validate if a file was supplied by the user\n images_directory =images_directory.blank? ? @@images_directory : images_directory #Validate if an image_directory was supplied, otherwise we use the default one\n file_s3_path = Util.upload_image(images_directory,file) #Upload the new image\n if !attributes[:previous_picture].blank? #Validate if there was a previous image file tied to the photo node\n Util.delete_image(attributes[:previous_picture]) #Delete the previous image file\n end\n else\n file_s3_path = self.file #If none was provided, keep the original file\n end\n #Create a raw photo object\n photo_req = { 'title'=>attributes[:title],\n 'description'=>attributes[:description],\n 'url'=>file_s3_path,\n 'date'=> Util.date_to_epoch(attributes[:date]), #Turn the date to epoch\n 'ownerId'=> self.owner_id \n } \n reqUrl = \"/api/photo/#{self.id}\" #Set the request url\n\n rest_response = MwHttpRequest.http_put_request(reqUrl,photo_req,user['email'],user['password']) #Make the PUT call to the server with the required parameters\n Rails.logger.debug \"Response from server: #{rest_response.code} #{rest_response.message}: #{rest_response.body}\"\n if rest_response.code == \"200\" #Validate if the response from the server is 200, which means OK\n photo = Photo.rest_to_photo(rest_response.body)\n return true, photo #Return success\n else\n return false, \"#{rest_response.code}\", \"#{rest_response.message}\" #Return error\n end\n else\n Rails.logger.debug self.errors.full_messages\n return false, self.errors.full_messages #Return invalid object error\n end\n end",
"def put_storage(request, params)\n xmldoc = XMLElement.build_xml(request.body, 'STORAGE')\n image_info = XMLElement.new(xmldoc) if xmldoc != nil\n\n image = ImageOCCI.new(\n Image.build_xml(params[:id]),\n @client)\n\n rc = nil\n if image_info['PERSISTENT'] && image_info['PUBLIC']\n error_msg = \"It is not allowed more than one change per request\"\n return OpenNebula::Error.new(error_msg), 400\n elsif image_info['PERSISTENT'] == 'YES'\n rc = image.persistent\n elsif image_info['PERSISTENT'] == 'NO'\n rc = image.nonpersistent\n elsif image_info['PUBLIC'] == 'YES'\n rc = image.publish\n elsif image_info['PUBLIC'] == 'NO'\n rc = image.unpublish\n end\n\n if OpenNebula.is_error?(rc)\n return rc, CloudServer::HTTP_ERROR_CODE[rc.errno]\n end\n\n # --- Prepare XML Response ---\n image.info\n return to_occi_xml(image, :code=>202)\n end",
"def update_image(request)\n http_request = request.to_http_info(@api_client.config)\n make_request(http_request, :GET, 'File')\n end",
"def update\n @path = Path.find(params[:id])\n\n respond_to do |format|\n if @path.update_attributes(params[:path])\n format.html { redirect_to([@layer, @path], :notice => 'Path was successfully updated.') }\n format.xml { head :ok }\n else\n format.html { render :action => \"edit\" }\n format.xml { render :xml => @path.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def update\n @image = Image.find(params[:id])\n\n respond_to do |format|\n if @image.update_attributes(params[:image])\n b = Bubo.from_heroku\n b.remove_image(@image.id.to_s)\n b.add_image(@image.id.to_s, @image.url)\n \n format.html { redirect_to(@image, :notice => 'Image was successfully updated.') }\n format.xml { head :ok }\n else\n format.html { render :action => \"edit\" }\n format.xml { render :xml => @image.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def put(path, options={})\n send_request 'put', path, options\n end",
"def put(path, params={})\n request(:put, path, params)\n end",
"def put(path, params={})\n request(:put, path, params)\n end",
"def put(path, params={})\n request(:put, path, params)\n end",
"def put(path, params={})\n request(:put, path, params)\n end",
"def put(path, params={})\n request(:put, path, params)\n end",
"def put(path, params={})\n request(:put, path, params)\n end",
"def put(path, params={})\n request(:put, path, params)\n end",
"def put(path, params={})\n request(:put, path, params)\n end",
"def update options={}\n client.put(\"/#{id}\", options)\n end",
"def put_path(base_path, payload)\n url = \"#{endpoint}/paths#{base_path}\"\n put_json(url, payload)\n end",
"def update\n @image = @user.images.find(params[:id])\n\n respond_to do |format|\n if @image.update_attributes(params[:image])\n flash[:notice] = 'Image was successfully updated.'\n format.html { redirect_to user_image_path(@user, @user) }\n format.xml { head :ok }\n else\n format.html { render :action => \"edit\" }\n format.xml { render :xml => @image.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def update_aos_version(args = {}) \n id = args['id']\n temp_path = \"/aosversions.json/{aosVersionId}\"\n path = temp_path\nargs.keys.each do |key|\n if (key == \"aosversionId\")\n args.delete(key)\n path = temp_path.gsub(\"{#{key}}\", id)\n end\nend\n puts \" PATH : #{path}\"\n put(path, args)\nend",
"def put(path, params = {})\n request(:put, path, params)\n end",
"def put(path, params = {})\n request(:put, path, params)\n end",
"def put(path, params = {})\n request(:put, path, params)\n end",
"def test_putpoi_update_valid\n nd = create(:node)\n cs_id = nd.changeset.id\n user = nd.changeset.user\n amf_content \"putpoi\", \"/1\", [\"#{user.email}:test\", cs_id, nd.version, nd.id, nd.lon, nd.lat, nd.tags, nd.visible]\n post :amf_write\n assert_response :success\n amf_parse_response\n result = amf_result(\"/1\")\n\n assert_equal 5, result.size\n assert_equal 0, result[0]\n assert_equal \"\", result[1]\n assert_equal nd.id, result[2]\n assert_equal nd.id, result[3]\n assert_equal nd.version + 1, result[4]\n\n # Now try to update again, with a different lat/lon, using the updated version number\n lat = nd.lat + 0.1\n lon = nd.lon - 0.1\n amf_content \"putpoi\", \"/2\", [\"#{user.email}:test\", cs_id, nd.version + 1, nd.id, lon, lat, nd.tags, nd.visible]\n post :amf_write\n assert_response :success\n amf_parse_response\n result = amf_result(\"/2\")\n\n assert_equal 5, result.size\n assert_equal 0, result[0]\n assert_equal \"\", result[1]\n assert_equal nd.id, result[2]\n assert_equal nd.id, result[3]\n assert_equal nd.version + 2, result[4]\n end",
"def put(uri, xml)\r\n req = Net::HTTP::Put.new(uri)\r\n req[\"content-type\"] = \"application/xml\"\r\n req.body = xml\r\n request(req)\r\n end",
"def update\n @request = Request.find(params[:id])\n\n if params[:images_attributes]\n if @request.update(request_params)\n if params[:images_attributes]\n params[:images_attributes]['image'].each do |a|\n @image_attachment = @request.images.create!(:picture => a, :name => 'RequestApp')\n end\n end\n\n redirect_to action: 'show'\n flash[:success] = \"Su solicitud ha sido editada con exito\"\n else\n render 'edit'\n flash.now[:danger] = \"Revise los siguentes errores\"\n end\n else\n flash.now[:danger] = \"La solicitud requiere de una imagen\"\n render 'edit'\n end\n end",
"def put(path, options = {})\n request(:put, path, options)\n end",
"def put(path, options = {})\n request(:put, path, options)\n end",
"def update\n @album = Album.find(params[:id])\n \n respond_to do |format|\n if @album.update_attributes(params[:album])\n @album.images.clear\n @album.images << Image.find([params[:images]].flatten)\n @album.save!\n format.html { redirect_to(albums_path, :notice => 'Album was successfully updated.') }\n format.xml { head :ok }\n else\n format.html { render :action => \"edit\" }\n format.xml { render :xml => @album.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def put(path, params={})\n request(:put, path, params)\n end",
"def perform_put(path, options = {})\n perform_request(:put, path, options)\n end",
"def update\n @map_image = MapImage.find(params[:id])\n\n respond_to do |format|\n if @map_image.update_attributes(params[:map_image])\n format.html { redirect_to(@map_image, :notice => 'Map image was successfully updated.') }\n format.xml { head :ok }\n else\n format.html { render :action => \"edit\" }\n format.xml { render :xml => @map_image.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def update\n @image = Image.find(params[:id])\n checkaccountobject(\"images\",@image)\n respond_to do |format|\n if @image.update_attributes(params[:image])\n format.html { redirect_to @image, notice: 'Image was successfully updated.' }\n format.json { head :ok }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @image.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @rent.update(rent_params)\n\n if params[:image]\n puts params[:image]\n params[:image].each { |image|\n @rent.rent_images.create(rent_id: @rent.id, image:image)\n }\n \n end\n \n format.html { redirect_to @rent, notice: 'Rent was successfully updated.' }\n format.json { render :show, status: :ok, location: @rent }\n else\n format.html { render :edit }\n format.json { render json: @rent.errors, status: :unprocessable_entity }\n end\n end\n end",
"def put(path, params)\n parse_response @client[path].put(params)\n end",
"def put(path, params)\n request(:put, path, params)\n end",
"def put(path, **options)\n execute :put, path, options\n end",
"def update\n par = img_properties\n logger.info(\"parameters: #{par.inspect}\")\n valid = true\n\n # find by user associated to app key, not by user from request parameter!\n @image = Image.editable_by(@user).find_by_guid(params[:id])\n\n return not_found if !@image\n\n # set these variables back to nil if they were in the request but blank\n if par[:ref]\n @image.ref = par[:ref].blank? ? nil : par[:ref]\n end\n if par[:name]\n @image.name = par[:name].blank? ? nil : par[:name]\n end\n if par[:page_url]\n @image.page_url = par[:page_url].blank? ? nil : par[:page_url]\n end\n @image.private = par[:private] if par[:private]\n\n # update calibration data if specified\n if !par[:calibrate_length].blank?\n @image.calibrate_length = par[:calibrate_length].to_f\n @image.calibrate_unit = par[:calibrate_unit].to_i if !par[:calibrate_unit].blank?\n @image.calibrate_coords = par[:calibrate_coords] if !par[:calibrate_coords].blank?\n @image.ppi = calculate_ppi(@image)\n end\n\n orig_url = par[:original_url] || par[:url]\n begin\n # this may fail\n if !orig_url.blank? && orig_url != @image.original_url\n # url was updated\n @image.remote_upload_url = orig_url\n @image.original_url = orig_url\n end\n rescue CarrierWave::DownloadError\n @image.errors.add(:remote_upload_url, \"^This url doesn't appear to be valid\")\n valid = false\n rescue CarrierWave::IntegrityError\n @image.errors.add(:remote_upload_url, \"^This url does not appear to point to a valid image\")\n valid = false\n rescue StandardError\n @image.errors.add(:remote_upload_url, \"There does not appear to be an image at this url\")\n valid = false\n end\n\n if valid && @image.save\n # update product if set\n @image.user_product = par[:product] if par[:product]\n\n image_data = @image.js_serialize\n # if the user hit the 'save and next' button, include the guid of the next image in the response.\n # The client side will redirect to the edit page for that image. \n if params[:commit] == 'save and next'\n image = Image.find_most_recent_uncalibrated(current_user.id)\n image_data['nextImage'] = image.guid if image\n end\n render :json => image_data, :callback => params[:callback]\n else\n render :json => { :error => 403, :messages => prepare_errors(@image), :callback => params[:callback] }, :status => 200\n end\n end",
"def update\n @img_info = ImgInfo.find(params[:id])\n\n respond_to do |format|\n if @img_info.update_attributes(params[:img_info])\n flash[:notice] = 'ImgInfo was successfully updated.'\n format.html { redirect_to(@img_info) }\n format.xml { head :ok }\n else\n format.html { render :action => \"edit\" }\n format.xml { render :xml => @img_info.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def put(path, params = {})\n request(:put, path, params)\n end",
"def put(path, params = {})\n request(:put, path, params)\n end",
"def image hash = {}\n hash = { :id => hash} if hash.is_a? String\n home hash.update :action => 'image', :trailing_slash => false\n end",
"def put(path, options = {})\n request(:put, path, options)\n end",
"def put(path, options = {})\n request(:put, path, options)\n end",
"def put(path, options = {}, &block)\n map(path, options, :put, &block)\n end",
"def put(path, params={})\n RestClient.put request_base+path, params\n end",
"def update_volumes(username, token, workset_name, volume_ids)\n\n #<?xml version=\"1.0\" encoding=\"UTF-8\" standalone=\"yes\"?>\n #<volumes xmlns=\"http://registry.htrc.i3.illinois.edu/entities/workset\">\n # <volume>\n # <id>9999999</id>\n # </volume>\n # <volume>\n # <id>3333333</id>\n # </volume>\n # </volumes>\n volumes_xml =\n \"<?xml version=\\\"1.0\\\" encoding=\\\"UTF-8\\\" standalone=\\\"yes\\\"?>\" +\n \"<volumes xmlns=\\\"http://registry.htrc.i3.illinois.edu/entities/workset\\\">\";\n\n for id in volume_ids\n volumes_xml += \"<volume><id>#{id}</id></volume>\"\n end\n volumes_xml += \"</volumes>\"\n\n\n # curl -v --data @new_volumes.xml -X PUT \\\n # -H \"Content-Type: application/vnd.htrc-volume+xml\" \\\n # -H \"Accept: application/vnd.htrc-volume+xml\" \\\n # http://localhost:9763/ExtensionAPI-0.1.0/services/worksets/workset1/volumes?user=fred\n\n url = URI.parse(\"#{APP_CONFIG['registry_url']}/worksets/#{workset_name}/volumes\")\n http = Net::HTTP.new(url.host, url.port)\n if Rails.env.development?\n http.set_debug_output($stdout)\n end\n http.use_ssl = true\n http.verify_mode = OpenSSL::SSL::VERIFY_NONE\n\n request = Net::HTTP::Put.new(url.request_uri)\n request[\"Content-Type\"] = \"application/vnd.htrc-volume+xml\"\n request.add_field(\"Authorization\", \"Bearer #{token}\")\n\n request.body = volumes_xml\n response = http.request(request)\n\n #xml = response.body\n\n case response\n when Net::HTTPUnauthorized then\n raise Exceptions::SessionExpiredError.new(\"Session expired. Please login again\")\n when Net::HTTPSuccess then\n # Do nothing\n else\n raise Exceptions::SystemError.new(\"Error retrieving worksets (HTTP #{response.code})\")\n end\n\n end",
"def upload_referenced_images(directory,filename,locale)\n version = Settings[:CURRENT_VERSION]\n begin\n doc_path = \"#{directory}/#{filename}\"\n relative_directory = File.dirname(doc_path) \n content_doc = Nokogiri::XML(open(doc_path)){|config| config.noent }\n content_doc.remove_namespaces!\n # Find each img element\n content_doc.xpath('//img').each do |img|\n # Steps for uploading content\n # 1. Create a hash of the file\n # 2. Get a unique path.\n # 3. Get the filename of the referenced document for the\n # attachment name\n # 4. Check to see if that document exists, if it does, compare\n # the hashes and only upload if it has changed.\n\n # If the image starts with a / assume the file will be in the public directory\n unless (img['src'].start_with?('/'))\n mime_type = get_mime_type(img['src'][/(?:.*)(\\..*$)/, 1])\n\n # Get the directory from the filename\n dir_match_re = /(.*)\\//\n file_dir = dir_match_re.match(filename)[1]\n\n # Fix relative paths here\n path_from_source = (Pathname.new img['src'])\n image_path = (Pathname.new(\"#{file_dir}/#{img['src']}\")).cleanpath.to_s\n id = \"#{Settings[:APP_NAME]}.#{image_path}.#{locale}.#{version}\"\n \n full_image_path = (Pathname.new(\"#{directory}/#{image_path}\")).cleanpath.to_s\n\n # Get the hash of the file on the filesystem\n np = Digest::MD5.file(full_image_path)\n attachment_hash = \"md5-#{Base64.encode64(np.digest)}\".strip\n \n # Look at the attachments on the document in the database\n # If there is an existing attachment with the same name, check the hash value.\n # If it's the same, don't upload it.\n # If it's different, upload it.\n\n #doc fields to create, if needed\n newDoc = {\n '_id' => id,\n :name => image_path,\n :locale => locale,\n :version => version,\n :updated => Time.now.to_date.iso8601,\n :content_hash => attachment_hash }\n \n #doc fields to update, if needed\n updatedDoc = {\n :updated => Time.now.to_date.iso8601,\n :content_hash => attachment_hash }\n \n #upsert the document\n upsert_doc(@db, newDoc, updatedDoc, :content_hash, @log)\n \n doc = @db.get(id)\n doc_attachments = JSON.parse(doc.to_json)[\"_attachments\"]\n\n # If there are no attachments, then doc_attachments will be Nil\n if (doc_attachments.is_a? Hash)\n # If there is already an attachment with the same name, check the hash.\n # If the hash is different, update it.\n unless (doc_attachments.has_key?(image_path) && doc_attachments[image_path][\"digest\"].eql?(attachment_hash))\n begin\n @db.put_attachment(doc, image_path, open(full_image_path, &:read), :content_type => mime_type)\n rescue RestClient::Conflict\n @log.warn \"Hit a conflict. Deleting the attachment and trying again.\"\n begin\n @db.delete_attachment(doc,image_path,true)\n begin\n # Have to get the document again, since the _rev has changed\n doc = @db.get(id)\n @db.put_attachment(doc, image_path, open(full_image_path, &:read), :content_type => mime_type)\n rescue => e\n @log.error\"The attachment was deleted, but could not be re-added.\"\n @log.error e.class\n @log.error e.message\n end\n rescue => e\n @log.warn \"Something went wrong when deleting the attachment. Unknown state.\"\n @log.error e.class\n @log.error e.message\n end\n rescue => e\n @log.error \"Something went wrong when adding an attachment - #{img['src']} on #{doc_path}\"\n @log.error e.message\n @log.error e.class\n end\n end\n else\n # There are no attachments on this document. Add this one.\n @db.put_attachment(doc, image_path, open(full_image_path, &:read), :content_type => mime_type)\n end\n end\n end\n rescue => e\n# @log.error \"Something went wrong when adding an attachment - #{img['src']} on #{doc_path}\"\n @log.error e.message\n @log.error e.class\n end\n end",
"def update(attrs, path=nil)\n resp = api_client.put(path || url, JSON.dump(attrs))\n refresh(JSON.load(resp.body))\n end",
"def update(id, name=\"Updated Name\", age=\"55\")\r\n xml_req =\r\n \"<?xml version='1.0' encoding='UTF-8'?>\r\n <person>\r\n <id type='integer'>#{id}</id>\r\n <name>#{name}</name>\r\n <age>#{age}</age> \r\n </person>\"\r\n request = Net::HTTP::Put.new(\"#{@url}/#{id}.xml\")\r\n request.add_field \"Content-Type\", \"application/xml\"\r\n request.body = xml_req\r\n http = Net::HTTP.new(@uri.host, @uri.port)\r\n response = http.request(request)\r\n # no response body will be returned\r\n case response\r\n when Net::HTTPSuccess\r\n return \"#{response.code} OK\"\r\n else\r\n return \"#{response.code} ERROR\"\r\n end\r\n end",
"def update!(params)\n res = @client.put(path, nil, params, \"Content-Type\" => \"application/json\")\n @attributes = res.json if res.status == 201\n res\n end",
"def put(*args)\n request :put, *args\n end",
"def update\n parse_path\n validate_path!\n\n if params.include?(:dir)\n @path.mkdir\n elsif params.include?(:file)\n @path.mv_from(params[:file].tempfile)\n elsif params.include?(:touch)\n @path.touch\n else\n content = request.body.read\n\n # forcing utf-8 because File.write seems to require it. request bodies are\n # in ASCII-8BIT and need to be re encoded otherwise errors are thrown.\n # see test cases for plain text, utf-8 text, images and binary files\n content.force_encoding('UTF-8')\n\n @path.write(content)\n end\n\n render json: {}\n rescue => e\n render json: { error_message: e.message }\n end",
"def api_put(path, data = {})\n api_request(:put, path, :data => data)\n end",
"def update\n @raw_image.update(raw_image_params)\n redirect_to admin_raw_images_path\n\n end",
"def update\n @accessory = Accessory.find(params[:id])\n\n respond_to do |format|\n if @accessory.update_attributes(params[:accessory])\n if @accessory.image_path.blank?\n @accessory.update_attributes(:image_path => (\"accessories/default.png\"))\n end\n format.html { redirect_to(@accessory, :notice => 'Accessory was successfully updated.') }\n format.xml { head :ok }\n else\n format.html { render :action => \"edit\" }\n format.xml { render :xml => @accessory.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @item.update(item_params)\n @item.images.delete_all\n unless params.require(:item)[:images_id].nil?\n params.require(:item)[:images_id].each do |id|\n image = Image.find_by_id(id)\n (@item.images << image) unless image.nil?\n end\n end\n format.html { redirect_to @item, notice: 'Item was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: 'edit' }\n format.json { render json: @item.errors, status: :unprocessable_entity }\n end\n end\n end",
"def put(path, options = {}, raw = false)\n request(:put, path, options, raw)\n end",
"def update opts = {}\n opts[:headers] ||= {}\n opts[:headers]['Content-Type'] ||= 'text/xml'\n post 'update', opts\n end",
"def _http_put resource, path\n uri = ::URI.parse(resource.auth_uri)\n path = _path uri, path\n request = Net::HTTP::Put.new(path)\n _build_request resource, request\nend",
"def update\n if(@check)\n @evento.update(event_params.except(:token))\n if @evento.save\n #first we delete all the current images if there are\n if params[:event_image_data]\n @evento.event_images.each do |image|\n image.destroy\n end\n #then we will create new ones\n params[:event_image_data].each do |file|\n @evento.event_images.create!(:image => file)\n end\n end\n render json: @evento, status: :ok, location: @evento\n else\n render json: @evento.errors, status: :unprocessable_entity\n end\n end\n end",
"def put(path, doc = nil, options = {})\n execute('PUT', path, options, doc)\n end",
"def _http_put resource, path\n uri = ::URI.parse(resource.auth_uri)\n path = _path uri, path\n request = Net::HTTP::Put.new(path)\n _build_request resource, request\nend",
"def update\n @image = Image.find(params[:id])\n \n respond_to do |format|\n if @image.update_attributes(params[:image])\n format.html { redirect_to([:admin,@image] ,:notice => 'Image was successfully updated.') }\n format.xml { head :ok }\n else\n format.html { render :action => \"edit\" }\n format.xml { render :xml => @image.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def destroy\n @image_path = ImagePath.find(params[:id])\n @image_path.destroy\n\n respond_to do |format|\n format.html { redirect_to(image_paths_url) }\n format.xml { head :ok }\n end\n end",
"def put(path, opts = {})\n request(:put, path, opts).body\n end",
"def update\n respond_to do |format|\n if @smartphone.update(smartphone_params)\n if params[:images]\n params[:images].each do |val|\n temp = @smartphone.pictures.find_by(image_file_name: val.original_filename)\n if temp\n temp.update_attributes(:image => val)\n else\n @smartphone.pictures.create(image: val)\n end\n end\n end\n format.html { redirect_to @smartphone, notice: 'Smartphone was successfully updated.' }\n format.json { render :show, status: :ok, location: @smartphone }\n else\n format.html { render :edit }\n format.json { render json: @smartphone.errors, status: :unprocessable_entity }\n end\n end\n end",
"def put(path, headers = {})\n process :put, path, headers\n end",
"def update_image_features(request)\n http_request = request.to_http_info(@api_client.config)\n make_request(http_request, :PUT, nil)\n nil\n end",
"def update\n @uploaded_image = UpdatedImage.find(params[:id])\n\n respond_to do |format|\n if @uploaded_image.update_attributes(params[:uploaded_image])\n format.html { redirect_to(@uploaded_image, :notice => 'UpdatedImage was successfully updated.') }\n format.xml { head :ok }\n else\n format.html { render :action => \"edit\" }\n format.xml { render :xml => @uploaded_image.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def update\n flash[:notice] = 'Image Successfully updated' if @image.update_attributes! params[:image]\n respond_with @owner, @image\n end",
"def update\n respond_to do |format|\n @pix.images.attach(params[:images]) if params[:images]\n if @pix.update(pix_params)\n format.html do\n redirect_to go_to_pix(@pix),\n flash: { success: 'Pix was successfully saved.' }\n end\n format.json { render :show, status: :ok, location: @pix }\n else\n format.html { render :edit }\n format.json { render json: @pix.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @place.update(place_params)\n if params[:images]\n params[:images].each do |image|\n @place.photos.create(image: image)\n end\n end\n format.html do\n redirect_to @place, notice: 'Place was successfully updated.'\n end\n format.json { render :show, status: :ok, location: @place }\n else\n format.html { render :edit }\n format.json { render json: @place.errors, status: :unprocessable_entity }\n end\n end\n end",
"def put(path, options={}, raw=false)\n request(:put, path, options, raw)\n end",
"def update\n @image = Image.find(params[:id])\n respond_to do |format|\n if @image.update_attributes(params[:image])\n format.html { redirect_to admins_image_path(@image, :notice => 'Image was successfully updated.') }\n format.xml { head :ok }\n else\n format.html { render :action => \"edit\" }\n format.xml { render :xml => @image.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def update\n @image = Image.find(params[:id])\n\n respond_to do |format|\n if @image.update_attributes(params[:image])\n format.html { redirect_to @image, :notice => 'Image was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render :action => \"edit\" }\n format.json { render :json => @image.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def create_update_volumes(username, token, workset_name, volume_ids)\n\n #<?xml version=\"1.0\" encoding=\"UTF-8\" standalone=\"yes\"?>\n #<volumes xmlns=\"http://registry.htrc.i3.illinois.edu/entities/workset\">\n # <volume>\n # <id>9999999</id>\n # </volume>\n # <volume>\n # <id>3333333</id>\n # </volume>\n # </volumes>\n volumes_xml =\n \"<?xml version=\\\"1.0\\\" encoding=\\\"UTF-8\\\" standalone=\\\"yes\\\"?>\" +\n \"<volumes xmlns=\\\"http://registry.htrc.i3.illinois.edu/entities/workset\\\">\";\n\n for id in volume_ids\n volumes_xml += \"<volume><id>#{id}</id></volume>\"\n end\n volumes_xml += \"</volumes>\"\n\n\n # curl -v --data @new_volumes.xml -X PUT \\\n # -H \"Content-Type: application/vnd.htrc-volume+xml\" \\\n # -H \"Accept: application/vnd.htrc-volume+xml\" \\\n # http://localhost:9763/ExtensionAPI-0.1.0/services/worksets/workset1/volumes?user=fred\n\n url = URI.parse(\"#{APP_CONFIG['registry_url']}/worksets/#{workset_name}\")\n http = Net::HTTP.new(url.host, url.port)\n if Rails.env.development?\n http.set_debug_output($stdout)\n end\n http.use_ssl = true\n http.verify_mode = OpenSSL::SSL::VERIFY_NONE\n\n request = Net::HTTP::Put.new(url.path)\n request[\"Content-Type\"] = \"application/vnd.htrc-volume+xml\"\n request.add_field(\"Authorization\", \"Bearer #{token}\")\n\n request.body = volumes_xml\n response = http.request(request)\n\n #xml = response.body\n\n case response\n when Net::HTTPUnauthorized then\n raise Exceptions::SessionExpiredError.new(\"Session expired. Please login again\")\n when Net::HTTPSuccess then\n # Do nothing\n else\n raise Exceptions::SystemError.new(\"Error retrieving worksets (HTTP #{response.code})\")\n end\n end",
"def update\n authorize @image_set\n if @image_set.update(params[:image_set].permit(:title, :description))\n # to handle multiple images upload on update when user add more pictures\n if params[:images]\n params[:images].each {|image|\n @image_set.images.create(image: image)\n }\n end\n\n flash[:notice] = \"Your image set has been updated!.\"\n redirect_to @image_set\n else\n flash[:alert] = \"Something went wrong.\"\n render 'edit'\n end\n end",
"def update\n @image = Image.find(params[:id])\n\n respond_to do |format|\n if @image.update_attributes(params[:image])\n format.html { redirect_to @image, notice: 'Image was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @image.errors, status: :unprocessable_entity }\n end\n end\n end"
] |
[
"0.6095052",
"0.5699765",
"0.56992865",
"0.5659542",
"0.56241524",
"0.56241524",
"0.56108135",
"0.5590362",
"0.5552536",
"0.55395144",
"0.5532609",
"0.5496175",
"0.5492888",
"0.5485352",
"0.5474991",
"0.54723066",
"0.54717857",
"0.5469874",
"0.5453326",
"0.5435665",
"0.5430836",
"0.5422867",
"0.54182315",
"0.54172885",
"0.5411479",
"0.53921735",
"0.5387833",
"0.53743213",
"0.53658456",
"0.53557676",
"0.53557676",
"0.53557676",
"0.53557676",
"0.53557676",
"0.53557676",
"0.53557676",
"0.53557676",
"0.5353188",
"0.53484994",
"0.5342825",
"0.53415954",
"0.5334688",
"0.5334688",
"0.5334688",
"0.5327907",
"0.5318498",
"0.53127676",
"0.53105754",
"0.53105754",
"0.53022563",
"0.5302122",
"0.5292475",
"0.52883327",
"0.5287283",
"0.5285751",
"0.52812093",
"0.5271193",
"0.5267143",
"0.52653754",
"0.5260491",
"0.5260414",
"0.5260414",
"0.5250416",
"0.5250413",
"0.5250413",
"0.52489233",
"0.52427405",
"0.52269214",
"0.5226437",
"0.5226333",
"0.522261",
"0.5222059",
"0.5219968",
"0.5214771",
"0.52115",
"0.52072346",
"0.5207153",
"0.5200632",
"0.52005196",
"0.5195859",
"0.5195275",
"0.5194072",
"0.5193825",
"0.51934737",
"0.51906663",
"0.5190641",
"0.5190583",
"0.5190008",
"0.51842856",
"0.5182614",
"0.5181946",
"0.5180441",
"0.51798695",
"0.5177431",
"0.51756275",
"0.5174848",
"0.5174231",
"0.5172066",
"0.51666594",
"0.51659703"
] |
0.6370364
|
0
|
DELETE /image_paths/1 DELETE /image_paths/1.xml
|
def destroy
@image_path = ImagePath.find(params[:id])
@image_path.destroy
respond_to do |format|
format.html { redirect_to(image_paths_url) }
format.xml { head :ok }
end
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def destroy\n @img = Img.find(params[:id])\n @img.destroy\n\n respond_to do |format|\n format.html { redirect_to(imgs_url) }\n format.xml { head :ok }\n end\n end",
"def destroy\n @image = Image.find(params[:id])\n \n imagen = @image.filename\n \n #function in manage_images.rb\n remove_image_file(imagen)\n \n @image.destroy\n\n respond_to do |format|\n format.html { redirect_to images_url }\n format.json { head :ok }\n end\n end",
"def delete_image(image_name)\n File.delete(image_name)\n end",
"def destroy\n @image = Image.find(params[:id])\n @image.destroy\n\n respond_to do |format|\n format.html { redirect_to(images_url) }\n format.xml { head :ok }\n end\n end",
"def destroy\n @image = Image.find(params[:id])\n @image.destroy\n\n respond_to do |format|\n format.html { redirect_to(images_url) }\n format.xml { head :ok }\n end\n end",
"def destroy\n @image = Image.find(params[:id])\n @image.destroy\n\n respond_to do |format|\n format.html { redirect_to(images_url) }\n format.xml { head :ok }\n end\n end",
"def destroy\n @image_attrib.destroy\n end",
"def destroy\n @image_set = ImageSet.find(params[:id])\n @image_set.destroy\n\n respond_to do |format|\n format.html { redirect_to(image_sets_url) }\n format.xml { head :ok }\n end\n end",
"def destroy\n @image = Image.find(params[:id])\n @image.destroy\n\n respond_to do |format|\n format.html { redirect_to(admin_images_url) }\n format.xml { head :ok }\n end\n end",
"def destroy\n @uploaded_image = UpdatedImage.find(params[:id])\n @uploaded_image.destroy\n\n respond_to do |format|\n format.html { redirect_to(template_files_url) }\n format.xml { head :ok }\n end\n end",
"def destroy\n ImagesIndex.delete params[:id]\n respond_to do |format|\n format.html { redirect_to(\"/images_indices\") }\n format.xml { head :ok }\n end\n end",
"def destroy\n @image = Image.find(params[:id])\n @image.destroy\n\n respond_to do |format|\n flash[:success] = 'Image was successfully deleted.'\n format.html { redirect_to([:admin, @parents, :images]) }\n format.xml { head :ok }\n end\n end",
"def delete_image tags\n all_images_by_tags( tags ).each { | image | image.deregister } \n end",
"def destroy\n @photo = Photo.find(params[:id])\n File.delete(Rails.root.join(\"app\",'assets','images',@photo.path))\n @photo.destroy\n\n respond_to do |format|\n format.html { redirect_to photos_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @image = Image.find(params[:id])\n @image.destroy\n\n respond_to do |format|\n format.html { redirect_to(admins_images_path) }\n format.xml { head :ok }\n end\n end",
"def delete_image(id)\n uri = URI.parse(\"http://\" + @location.host + \":\" + @location.port.to_s + \"/v2/images/\" + id)\n return delete_request(uri, @token)\n end",
"def delete_image(image_id)\n delete(\"cloud-instances/#{guid}/images/#{image_id}\")\n end",
"def delete_storage(request, params)\n # --- Get the Image ---\n image = ImageOCCI.new(\n Image.build_xml(params[:id]),\n @client)\n\n # --- Delete the Image ---\n rc = image.delete\n if OpenNebula.is_error?(rc)\n return rc, CloudServer::HTTP_ERROR_CODE[rc.errno]\n end\n\n return \"\", 204\n end",
"def destroy\n @image = Image.find(params[:id])\n @image.destroy\n\n\n redirect_to new_flat_image_path(@image.flat)\n\n end",
"def delete(*uris); end",
"def deleteEntityImage( entity_id, gen_id)\n params = Hash.new\n params['entity_id'] = entity_id\n params['gen_id'] = gen_id\n return doCurl(\"delete\",\"/entity/image\",params)\n end",
"def removeImage\n @groof = Greenroof.find(params[:id])\n if not (@groof.images.first.nil?)\n\n\n directory = \"/public/greenroofs/photos/\" + params[:id]\n photoFilename = @groof.images.first.photo\n thumbFilename = @groof.images.first.thumb\n photoPath = Dir.pwd + directory + \"/\" + photoFilename\n thumbPath = Dir.pwd + directory + \"/\" + thumbFilename\n File.delete(photoPath)\n File.delete(thumbPath)\n @groof.images.first.delete\n end\n redirect_to greenroof_path(@groof)\n end",
"def destroy\n @image = @user.images.find(params[:id])\n @image.destroy\n flash[:notice] = \"Image has been removed.\"\n\n respond_to do |format|\n format.html { redirect_to user_images_path(@user) }\n format.xml { head :ok }\n end\n end",
"def delete\n model.delete_attachment(@path)\n end",
"def destroy\n @img_info = ImgInfo.find(params[:id])\n @img_info.destroy\n\n respond_to do |format|\n format.html { redirect_to(img_infos_url) }\n format.xml { head :ok }\n end\n end",
"def delete_image_attachment\n @image = ActiveStorage::Attachment.find(params[:id])\n @image.purge\n redirect_back(fallback_location: items_path)\n end",
"def delete_image_files\n [ :original , :medium , :thumb ].each do |size|\n path = image_path(size, true)\n File.delete(path) if path\n end\n end",
"def remove_image( image_id )\n image_id = image_id.to_s.upcase\n upload_key = UPLOAD_PATH % image_id\n hires_key = HIRES_PATH % image_id\n lowres_key = LOWRES_PATH % image_id\n\n client.delete_objects(\n bucket: aws.bucket,\n delete: {\n objects: [\n {key: upload_key},\n {key: hires_key},\n {key: lowres_key}\n ]\n }\n )\n end",
"def destroy\n @raw_image.destroy\n redirect_to admin_raw_images_path\n end",
"def delete\n [filename, thumbnail_name].each do |name|\n image = \"#{DIRECTORY}/#{name}\"\n File.delete(image) if File.exists?(image)\n end\n end",
"def delete\n [filename, thumbnail_name].each do |name|\n image = \"#{DIRECTORY}/#{name}\"\n File.delete(image) if File.exists?(image)\n end\n end",
"def destroy\n @image = Image.find(params[:id])\n @image.destroy\n\n respond_to do |format|\n format.html { redirect_to(admin_images_url) }\n format.xml { head :ok }\n end\n end",
"def destroy\n request(:delete, \"/computing/image/#{uuid}\")\n true\n end",
"def delete(session, id)\n write_task('rvpe.image.delete', session) do\n err_msg = \"You don't have permission to delete the image.\"\n sanity_check(session, id, err_msg) do\n call_one_xmlrpc('one.image.delete', session, id)\n end\n end\n end",
"def destroy\n @image_store = ImageStore.find(params[:id])\n @image_store.destroy\n\n respond_to do |format|\n format.html { redirect_to(image_stores_url) }\n format.xml { head :ok }\n end\n end",
"def destroy\n file_url = @image.url\n @image.destroy\n\n File.delete(\"public/uploads/#{file_url}\")\n\n respond_to do |format|\n format.html { redirect_to images_url, notice: 'Image was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @image = Image.find(params[:id])\n @image.destroy\n redirect_to console_images_path\n end",
"def destroy\n @image_dataset = ImageDataset.find(params[:id])\n @image_dataset.destroy\n\n respond_to do |format|\n format.html { redirect_to(image_datasets_url) }\n format.xml { head :ok }\n end\n end",
"def destroy\n @map_image = MapImage.find(params[:id])\n @map_image.destroy\n\n respond_to do |format|\n format.html { redirect_to(map_images_url) }\n format.xml { head :ok }\n end\n end",
"def destroy\n @user_image = UserImage.find(params[:id])\n @user_image.destroy\n\n respond_to do |format|\n format.html { redirect_to(user_images_url) }\n format.xml { head :ok }\n end\n end",
"def destroy\n #Finds selected image\n @image = Image.find(params[:id])\n #destroy image\n @image.destroy\n respond_to do |format|\n format.html { redirect_to '/admin' }\n format.json { head :ok }\n end\n end",
"def destroy(paths)\n\t\tlogin_filter\n\t\tpaths = [paths].flatten\n\t\tpaths = paths.collect { |path| namespace_path(path) }\n\t\t@agent.post(\"/cmd/delete\", {\"files\"=> paths, \"t\" => @token }).code == \"200\"\n\tend",
"def destroy\n Image.find(params[:id]).destroy\n redirect_to :action => 'index' \n end",
"def destroy\n @image.destroy\n end",
"def destroy\n @image.destroy\n end",
"def destroy\n id = @taxi_image.taxi_sevice_id\n @taxi_image.destroy\n respond_to do |format|\n format.html { redirect_to \"/taxi_sevices/\" + id.to_s, notice: 'Taxi image was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @client = Client.find(params[:id])\n Client.transaction do\n FileUtils.rm Dir[\"#{Rails.root}/public/files/logo_files/\"+@client.logo.to_s]\n @client.destroy\n\n respond_to do |format|\n format.html { redirect_to clients_url, notice: 'Client was successfully deleted.' }\n format.json { head :no_content }\n end\n end\n end",
"def rm path\n end",
"def destroy\n #If image has parent, update children and vice versa\n if image.root_version? then\n newroot = image.child_versions.order(:created_at).last\n image.child_versions.delete(newroot)\n image.child_versions.each do |v| v.parent_image = newroot and v.save end\n else\n image.child_versions.each do |v| v.parent_image = image.parent_image and v.save end\n end\n\n image.destroy\n respond_to do |format|\n format.html { redirect_to images_url, notice: 'Image was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @imagem = @evento.imagems.find(params[:id])\n @imagem.destroy\n flash[:notice] = 'Imagem excluida com sucesso !'\n respond_to do |format|\n format.html { redirect_to(admin_evento_imagems_url) }\n format.xml { head :ok }\n end\n end",
"def destroy\r\n @location = Location.find(params[:id])\r\n RemovedLocation.create(server_id: Integer(params[:id]))\r\n directory = Rails.root.join('app','assets','locations');\r\n\r\n path = File.join(directory, @location.image)\r\n File.delete(path)\r\n @location.destroy\r\n mv = MapsVersion.first\r\n mv.version = mv.version+1\r\n mv.save\r\n respond_to do |format|\r\n format.html { redirect_to locations_url }\r\n format.json { head :no_content }\r\n end\r\n end",
"def destroy\n @image = Image.find(params[:id])\n if @image.destroy\n flash[:notice] = t(:image_deleted)\n else\n flash[:error] = t(:image_not_deleted)\n end\n\n respond_to do |format|\n format.html { redirect_back_or_default(admin_images_path) }\n format.xml { head :ok }\n end\n end",
"def destroy\n output = \"oneimage delete #{resource[:name]} \", self.class.login\n `#{output}`\n end",
"def destroy\n @image = Image.find(params[:id])\n @image.destroy\n\n head :no_content\n end",
"def delete\n conn.delete(escaped_path)\n true\n rescue StandardError => e\n puts \"carrierwave-upyun delete failed: #{e.inspect}\"\n nil\n end",
"def destroy\n @especie_imagem = EspecieImagem.find(params[:id])\n @especie_imagem.destroy\n\n respond_to do |format|\n format.html { redirect_to(especie_imgens_url) }\n format.xml { head :ok }\n end\n end",
"def destroy\n @article = Article.get(params[:id])\n \n # 업로드된 이미지 파일 삭제 =========================================================================\n file_name = @article.image_filename\n if File.exist?(IMAGE_PATH + file_name)\n \tFile.delete(IMAGE_PATH + file_name) #original image file\n \tFile.delete(IMAGE_PATH + \"t_\" + file_name) #thumbnail file\n end\n # 업로드된 이미지 파일 삭제 =========================================================================\n @article.destroy\n\n respond_to do |format|\n format.html { redirect_to(articles_url) }\n format.xml { head :ok }\n end\n end",
"def remove_uploaded_file\n remove_image!\n end",
"def delete(path)\n path = relativize_path path\n\n Precog.connect self do |http|\n uri = Addressable::URI.new\n uri.query_values = { :apiKey => api_key }\n\n http.delete \"/ingest/v#{VERSION}/fs/#{path}?#{uri.query}\"\n end\n end",
"def destroy\n @image.destroy\n\n respond_to do |format|\n format.html { redirect_to(images_url) }\n format.json { head :ok }\n end\n end",
"def destroy\n @exposure = Exposure.find(params[:id])\n @image = Image.find(:first, :conditions => [\"id = ?\",@exposure.image_id])\n @image.destroy unless @image.nil?\n @exposure.destroy\n\n respond_to do |format|\n format.html { redirect_to(exposures_url) }\n format.xml { head :ok }\n end\n end",
"def destroy\n @image_gallery_image = ImageGalleryImage.find(params[:id])\n @image_gallery_image.destroy\n\n respond_to do |format|\n format.html { redirect_to([:admin, @image_gallery_group]) }\n format.xml { head :ok }\n end\n end",
"def destroy\n @image = Image.find(params[:id])\n @image.destroy\n render json: {status: \"success\"}, status: :ok\n end",
"def delete_image\n @page.first('span[data-vh-id=\"remove\"]').click\n end",
"def destroy\n if @article.img_path != ''\n File.delete('app/assets/images/' + @article.img_path) if File.exist?('app/assets/images/' + @article.img_path)\n end\n @article.destroy\n respond_to do |format|\n format.html { redirect_to articles_url, notice: 'Article was successfully destroyed.' }\n format.json { head :no_content }\n end\n\n\n end",
"def destroy\n Rails.logger.debug {\"destroying gridfs file #{@id}\"}\n if persisted?\n Photo.mongo_client.database.fs.find(:_id=>BSON::ObjectId.from_string(@id)).delete_one\n end\n end",
"def destroy\n @remote_image_content = RemoteImageContent.find(params[:id])\n @remote_image_content.destroy\n\n respond_to do |format|\n format.html { redirect_to(remote_image_contents_url) }\n format.xml { head :ok }\n end\n end",
"def remove_photos(db, selected)\n\n photos = selected.split(\",\")\n\n photos.each do |photo|\n delete_s3_file(photo)\n delete_db_record(db, photo)\n end\n\nend",
"def destroy\n @album = Album.find(params[:id])\n @album.images.clear\n @album.save!\n @album.destroy\n Image.all(:conditions => {:album_id => params[:id]}).each do |x| \n x.album_id = nil\n x.save\n end\n respond_to do |format|\n format.html { redirect_to(albums_url) }\n format.xml { head :ok }\n end\n end",
"def delete_all_uploaded_images\n images_dir = FileUploader.file_dir('article', id)\n FileUtils.rm_rf(images_dir) if File.directory?(images_dir)\n end",
"def remove_image # :norobots:\n pass_query_params\n if @image = find_or_goto_index(Image, params[:image_id]) and\n @observation = find_or_goto_index(Observation, params[:observation_id])\n if !check_permission!(@observation)\n flash_error(:runtime_image_remove_denied.t(:id => @image.id))\n elsif !@observation.images.include?(@image)\n flash_error(:runtime_image_remove_missing.t(:id => @image.id))\n else\n @observation.remove_image(@image)\n @observation.log_remove_image(@image)\n Transaction.put_observation(\n :id => @observation,\n :del_image => @image\n )\n flash_notice(:runtime_image_remove_success.t(:id => @image.id))\n end\n redirect_to(:controller => 'observer', :action => 'show_observation',\n :id => @observation.id, :params => query_params)\n end\n end",
"def destroy\n @estate_agent_image = EstateAgentsImage.find(params[:id])\n @estate_agent_image.destroy\n\n respond_to do |format|\n format.html { redirect_to estate_agent_image_images_url }\n format.json { head :no_content }\n end\n end",
"def run_on_deletion(paths)\n end",
"def run_on_deletion(paths)\n end",
"def destroy\n @image.destroy\n respond_to do |format|\n format.html { redirect_to images_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @image.destroy\n respond_to do |format|\n format.html { redirect_to images_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @image.destroy\n respond_to do |format|\n format.html { redirect_to images_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @image.destroy\n respond_to do |format|\n format.html { redirect_to images_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @image.destroy\n respond_to do |format|\n format.html { redirect_to images_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @image.destroy\n respond_to do |format|\n format.html { redirect_to images_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @gallery_image = GalleryImage.find(params[:id])\n @gallery_image.destroy\n\n respond_to do |format|\n format.html { redirect_to(home_url) }\n format.xml { head :ok }\n end\n end",
"def destroy\n @jpeg_folder = JpegFolder.find(params[:id])\n @jpeg_folder.destroy\n\n respond_to do |format|\n format.html { redirect_to(jpeg_folders_url) }\n format.xml { head :ok }\n end\n end",
"def destroy\n @related_content = RelatedContent.find(params[:id])\n @related_content.destroy\n dirname = \"#{RelatedContent::UPLOAD_DIR}/#{@related_content.id}\"\n FileUtils.rm_rf dirname\t\n redirect_to @related_content.node\t \n\nend",
"def destroy\n @image_url = ImageUrl.find(params[:id])\n @image_url.destroy\n\n respond_to do |format|\n format.html { redirect_to image_urls_url }\n format.json { head :no_content }\n end\n end",
"def remove_images(object, name)\n FileUtils.rm Dir.glob(File.join(path_for(object.has_image_id), name + '*'))\n Dir.rmdir path_for(object.has_image_id)\n rescue SystemCallError \n end",
"def delete(container_name, file_name)\n validate_path_elements(container_name, file_name)\n\n client.request(\n method: :delete,\n path: \"#{container_name}/#{file_name}\",\n expected: 204\n )\n end",
"def destroy\n @photo = Photo.find(params[:id])\n @photo.destroy\n\n respond_to do |format|\n format.html { redirect_to(photos_url) }\n format.xml { head :ok }\n end\n end",
"def destroy\n @photo = Photo.find(params[:id])\n @photo.destroy\n\n respond_to do |format|\n format.html { redirect_to(photos_url) }\n format.xml { head :ok }\n end\n end",
"def destroy\n @photo = Photo.find(params[:id])\n @photo.destroy\n\n respond_to do |format|\n format.html { redirect_to(photos_url) }\n format.xml { head :ok }\n end\n end",
"def destroy\n @photo = Photo.find(params[:id])\n @photo.destroy\n\n respond_to do |format|\n format.html { redirect_to(photos_url) }\n format.xml { head :ok }\n end\n end",
"def destroy\n RestClient.delete \"#{REST_API_URI}/contents/#{id}.xml\" \n self\n end",
"def destroy\n @image_reference.destroy\n respond_to do |format|\n format.html { redirect_to image_references_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @image.destroy\n\n respond_to do |format|\n format.json { head :no_content }\n end\n end",
"def delete_files\n return if id.nil?\n return unless File.exist?(directory_path)\n\n FileUtils.rm_rf(directory_path)\n s3_delete_files\n end",
"def destroy\n @loc_image.destroy\n respond_to do |format|\n format.html { redirect_to loc_images_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n path = 'public/uploads/batale/text/image/' + @batale_text.id\n FileUtils.remove_dir(path) unless Dir.glob(path).empty? # Remove imagem associada ao texto, caso exista\n @batale_text.destroy\n respond_to do |format|\n format.html { redirect_to batale_texts_url, notice: 'Texto deletado com sucesso.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @image = Image.find(params[:id])\n @image.destroy\n\n respond_to do |format|\n format.html { redirect_to images_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @image = Image.find(params[:id])\n @image.destroy\n\n respond_to do |format|\n format.html { redirect_to images_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @image = Image.find(params[:id])\n @image.destroy\n\n respond_to do |format|\n format.html { redirect_to images_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @image = Image.find(params[:id])\n @image.destroy\n\n respond_to do |format|\n format.html { redirect_to images_url }\n format.json { head :no_content }\n end\n end"
] |
[
"0.6799447",
"0.67660975",
"0.6710873",
"0.66433656",
"0.66433656",
"0.66433656",
"0.66241515",
"0.6623799",
"0.6623578",
"0.65880513",
"0.6587223",
"0.6548239",
"0.65446794",
"0.6543704",
"0.65315104",
"0.65198755",
"0.6518764",
"0.65038085",
"0.65017927",
"0.64876753",
"0.6474715",
"0.64593446",
"0.6452045",
"0.6435428",
"0.64352214",
"0.6412728",
"0.64087474",
"0.63871217",
"0.63854337",
"0.63847816",
"0.63847816",
"0.63824785",
"0.63823736",
"0.6372444",
"0.6370153",
"0.63679355",
"0.6366305",
"0.63419455",
"0.6337835",
"0.63251674",
"0.6321925",
"0.63207364",
"0.63069147",
"0.6305747",
"0.6305747",
"0.6290661",
"0.6288964",
"0.6284363",
"0.6284011",
"0.62807065",
"0.6265379",
"0.626537",
"0.6263096",
"0.62624335",
"0.625566",
"0.6250511",
"0.62427145",
"0.6230584",
"0.62291086",
"0.62248015",
"0.6223318",
"0.6217038",
"0.6206948",
"0.61872536",
"0.618297",
"0.61779344",
"0.61764014",
"0.6176368",
"0.61761844",
"0.61755204",
"0.6171952",
"0.6166929",
"0.61665213",
"0.61665213",
"0.61660445",
"0.61660445",
"0.61660445",
"0.61660445",
"0.61660445",
"0.61660445",
"0.61647475",
"0.6160257",
"0.61559784",
"0.61539274",
"0.61472666",
"0.6146233",
"0.6141567",
"0.6141567",
"0.6141567",
"0.6141567",
"0.61387306",
"0.6136846",
"0.6136814",
"0.6133505",
"0.6132814",
"0.6132546",
"0.6131908",
"0.6131908",
"0.6131908",
"0.6131908"
] |
0.7535475
|
0
|
=begin def xyz(abc) azz=[] abc.each do |i| azz= i if i.to_s.include?("wa") break if azz.size>0 end print azz end =end =begin
|
def count_elements(abc)
z=0
i=1
xyz=Array.new()
xyz[0]=abc[0]
xyz[0][:count]=1
while i<abc.size do
if xyz[z][:name]==abc[i][:name]
xyz[z][:count]+=1
i+=1
else
z+=1
xyz[z]=abc[i]
xyz[z][:count]=1
i+=1
end
end
#puts xyz.size
print xyz
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def sch(emaillist,finalPrint,cmlInput,argm,search)\n\temaillist.each {|e| finalPrint.push(e) if yield(e)} if\n\tcmlInput[argm+1].eql? search\n\tfinalPrint\nend",
"def xyz2\n\n end",
"def bs; end",
"def start=(_arg0); end",
"def test;\"def test;;end;array = [attempt];puts array\";end",
"def get_letters(instr)\n MORSE.each { |c, re|\n if (instr =~ re) then\n yield c,$1\n end\n }\nend",
"def _print_ace(a,b,species)\n ace=String.new\n a.each{|_a|\n next if b.nil? || _a.nil? ||b.size<1\n ace << \"Gene : \\\"#{_a}\\\"\\n\"\n b.each{|_b|\n ace << \"Ortholog #{_b} \\\"#{species}\\\" Inferred_automatically OrthoMCL\\n\"\n }\n ace << \"\\n\"\n }\n return ace\nend",
"def _print_ace(a,b,species)\n ace=String.new\n a.each{|_a|\n next if b.nil? || _a.nil? ||b.size<1\n ace << \"Gene : \\\"#{_a}\\\"\\n\"\n b.each{|_b|\n ace << \"Ortholog #{_b} \\\"#{species}\\\" Inferred_automatically OrthoMCL\\n\"\n }\n ace << \"\\n\"\n }\n return ace\nend",
"def sea; end",
"def vowel(*argString)\n strArray = []\n argString.each {|x|if x.scan(/[aeiou]/).count>=2\n strArray << x\n end}\n print strArray\n puts\nend",
"def names\n yield \"ylva\"\n yield \"brighid\"\n yield \"shifra\"\n yield \"yesamin\"\nensure\n puts \"grim\"\nend",
"def start_re=(_); end",
"def outputcase \n for i in 0...@records.recordsArr.length do\n\n isfit = checkMustHaveNotLimit @records.recordsArr[i]\n \n\n for j in 0...@records.recordsArr[i].valuesArr.length do\n print parameters.paramsArr[j].elementsArr[@records.recordsArr[i].valuesArr[j]].value\n print ' '\n end\n puts ' '\n \n end\n end",
"def start_of_word; end",
"def lex_en_expr_beg=(_arg0); end",
"def lex_en_expr_beg=(_arg0); end",
"def lex_en_expr_beg=(_arg0); end",
"def e_words(str)\r\n \r\nend",
"def first_loop(letters)\n 0.upto(letters.length-1) do |idx1|\n (idx1+4).step(letters.length-1,2) do |idx2|\n return [idx1,idx2] if letters[idx1].casecmp(letters[idx2])==0\n end\n end\n nil\nend",
"def malts; end",
"def each(*) end",
"def yyerrok; end",
"def yyerrok; end",
"def kelp(a,b)\n for i in a..b \n puts i\n end\n \"done\"\n end",
"def str3; end",
"def str3; end",
"def testComprehension\n=begin\n yin = proc do |x|\n Transcript.cr\n x\n end.call(Continuation.current)\n\n yang = proc do |x|\n Transcript.nextPut('*')\n x\n end.call(Continuation.current)\n\n yin.call(yang)\n=end\n end",
"def find_cool(abc)\n xyz=[]\n\n abc.each do |i|\n if i[:temperature]==\"cool\"\n print i\n xyz=i\n end\n end\n\n #print xyz\nend",
"def can_make_word(string)\n blocks = [\n ['B','O'],\n ['X','K'],\n ['D','Q'],\n ['C','P'],\n ['N','A'],\n ['G','T'],\n ['R','E'],\n ['T','G'],\n ['Q','D'],\n ['F','S'],\n ['J','W'],\n ['H','U'],\n ['V','I'],\n ['A','N'],\n ['E','R'],\n ['F','S'],\n ['L','Y'],\n ['P','C'],\n ['Z','M']\n ]\n\n letters = string.upcase.chars\n\n letters.each do |letter|\n blocks.each_with_index do |block, i|\n if block.include?(letter)\n blocks.delete_at(i)\n binding.pry\n return true\n # else \n # return false \n end \n end\n end \n\nend",
"def fun_name(defn)\n # The ultimate masterpiece of illegible code.\n defn[3][1][1][3][1][1]\n end",
"def lex_en_expr_beg; end",
"def lex_en_expr_beg; end",
"def lex_en_expr_beg; end",
"def print_students(students)\n\tstudents.each_with_index do |student , index| \n\t\tputs \"#{index + 1} - #{student}\" if student.start_with? \"A\"\n\tend\n\nend",
"def vowels(string)\r\n i=0\r\n p vowels_test = string.downcase.scan(/\\w/)\r\n p vowels_test.keep_if { |a| a == \"a\" }\r\n # if string == vowels[0]\r\n # return vowels\r\n # end\r\n # end\r\n \r\nend",
"def begin() end",
"def lines_of_code; end",
"def lines_of_code; end",
"def outer; end",
"def stderrs; end",
"def str1; end",
"def str1; end",
"def each; end",
"def each; end",
"def each; end",
"def each; end",
"def each; end",
"def each; end",
"def each; end",
"def each; end",
"def each; end",
"def each; end",
"def each; end",
"def bbb\n puts \"bbb\"\n end",
"def ABcheck(str)\ncounter = 0 \nstr.each_char.do |x|\nif( x == 'a' && counter + 4 < str.length str.chars.to_a[counter + 4] =='b'\nreturn true\nend",
"def looper\r\n for i in 1..10\r\n puts i #wrong method puts.\r\nend #wrong position of end word.\r\nend",
"def sb(t, s, c)\n\n end",
"def scan_addicts(param)\r\n #array = []\r\n arr = []\r\n param.each do |i|\r\n \r\n if i >= 90\r\n arr.push('mal') \r\n else\r\n arr.push('bien')\r\n end \r\n end\r\n print arr\r\nend",
"def each_mid(&block)\n \n end",
"def nearby_az(string)\n\tspecial = []\n\tstring.split('').each_with_index do |x, index| \n\t\tif x == 'a'\n\t\t\tputs \"hey\"\n\t\t\treturn true if string.split('')[(index + 1)..(index+3)].include?('z')\n\t\tend\n\tend\n\tfalse\nend",
"def main_end ; end",
"def find_word(search, word, maxRow, maxCol)\n (0..maxRow).each do |i|\n (0..maxCol).each do |j|\n if starts_here(search, word, i, j)\n puts \"The word '#{word}' starts at (#{i},#{j}).\"\n return\n end\n end\n end\n puts \"The word '#{word}' was not found!\"\nend",
"def begin(p0) end",
"def lh(t, s, c)\n\n end",
"def get_sandwich(str) # needs some work\n (str.size - 4).times do |i|\n slice = str[i..(i + 4)]\n if slice != \"bread\"\n puts slice\n end\n end\n\nend",
"def find_it(seq)\r\n\r\nend",
"def my_method\n\tputs \"ruby\".concat(\" is awesome\")\n\n\tbinding.pry\n\tletters_array = (\"a\"..\"c\").to_a\n\tletters_array.each do |letter|\n\t\tbinding.pry\n\t\tputs letter\n\n\tend\nend",
"def lbu(t, s, c)\n\n end",
"def sokongan\n a='(staffgrade_id=? ' if Employgrade.find(:all, :conditions=>['group_id=? or group_id=?', 4,2]).map(&:id).uniq.count!=0\n 0.upto(Employgrade.find(:all, :conditions=>['group_id=? or group_id=?', 4,2]).map(&:id).uniq.count-2) do |l| \n a=a+'OR staffgrade_id=? '\n end\n a=a+')'\n return a if position==1 && position2==1 && position3==0\n \n b='(staffgrade_id=? ' if Employgrade.find(:all, :conditions=>['group_id=? or group_id=?', 4,1]).map(&:id).uniq.count!=0\n 0.upto(Employgrade.find(:all, :conditions=>['group_id=? or group_id=?', 4,1]).map(&:id).uniq.count-2) do |l| \n b=b+'OR staffgrade_id=? '\n end\n b=b+')'\n return b if position==1 && position2=0 && position3==1\n \n c='(staffgrade_id=? ' if Employgrade.find(:all, :conditions=>['group_id=? or group_id=?', 2,1]).map(&:id).uniq.count!=0\n 0.upto(Employgrade.find(:all, :conditions=>['group_id=? or group_id=?', 2,1]).map(&:id).uniq.count-2) do |l| \n c=c+'OR staffgrade_id=? '\n end\n c=c+')'\n return c if position==0 && position2=1 && position3==1\n end",
"def process_code_listing(lines)\n # extract caption\n caption = nil\n # no caption for listings in appendix\n if !lines.first.index(\",\").nil?\n raise \"error [#{lines.first.index(\",\")}] [#{lines.first}]\" if lines.first.split(\",\").size != 2\n raise \"error\" if lines.first.split(\",\").first.split(\"=\").size != 2\n caption = lines.first.split(\",\").first.split(\"=\")[1] \n end\n # remove start and end\n lines = lines[1..-1]\n # pretty print \n return final_pretty_code_listing(lines, caption)\nend",
"def finally(def_arry)\n\twhile def_arry != []\n\tout = def_arry.shift.transpose[0]\n\tp out\n\tend\nend",
"def cracklepop1\n def aux x\n /[A-Z]+/i.match(x.to_s).to_s\n end\n (1..100).map do |i|\n x = i%3==0 ? 'crackle' : i\n i%5==0 ? (aux(x)+'pop') : x\n end\nend",
"def translate(a)\n\tvoyel = [\"a\",\"e\",\"i\",\"o\",\"u\"]\ncheck = 0\nn = 0\nx = a \n words = a.split(/\\W+/)\n words.each do |a|\n\tok = voyel.include?(a[0])\n\tif ok == true \n\t\ta = a + \"ay\"\n\t\treturn a \n\tend\n while check <= 4\n\tb = a.slice(0..check)\n\n\tcheck = check + 1\n \n\tok = voyel.include?(x[check])\n\ttest1 = \"qu\".include?(x[check])\n\tif test1 == true \n\t\tif check == 1\n\t\t\tb = a.slice(0..check)\n\t\t\ta = a + b + \"ay\"\n\t\treturn a[2..-1]\n\t elsif check == 2\n\t \tb = a.slice(1..check)\n\t \t\ta = a + b + \"ay\"\n\t \treturn a[3..-1]\n\t elsif check == 3 \n\t \t\ta = a + b + \"ay\"\n\t \treturn a[4..-1]\n\t end\n\tend\n\n\tif ok == true \n\t\tif check == 1\n\t\t\ta = a + b + \"ay\"\n\t\treturn a[1..-1]\n\t elsif check == 2\n\t \t\ta = a + b + \"ay\"\n\t \treturn a[2..-1]\n\t elsif check == 3 \n\t \t\ta = a + b + \"ay\"\n\t \treturn a[3..-1]\n\t end\n\tend\nend\nend\nend",
"def begin_pos=(_); end",
"def bellini; end",
"def abc_three\n\t2.upto(1000) do |a|\n\t (a+1).upto(1000) do |b|\n\t \tc = Math.sqrt(a**2 + b**2)\n\t \tif (a+b+c==1000)\n\t \t\treturn a*b*c \n\t \tend\n\t end\n\tend\nend",
"def start_with_h(arr_with_h)\n #iterate into the array and when found the first string finish\n arr_with_h.find {|letter| letter[0] == \"h\"}\nend",
"def start_word_pattern; end",
"def scream_c(words)\n #words = words + \"!!!!!\"\n return\n #puts words\nend",
"def berlioz; end",
"def find_a(arry)\n\n arry.select do |word|\n \"#{word}\".start_with?(\"a\")\n\n end\n # newlist\n end",
"def ex2\n for i in (0..5)\n for j in (1..(4 - (4 - i).abs))\n print '#'\n end\n print \"\\n\"\n end\nend",
"def find_missing_letter(range)\nsplit_range = []\nsplit_range = range.split(//)\narray = []\n\nrange_two = (split_range[0]..split_range[-1]).to_a\narray << range_two - split_range\n\n if array.length > 0\n p array.join\n else\n p nil\n end\nend",
"def ex1\n for i in (0..5)\n for j in (1..(6 - i))\n print '#'\n end\n print \"\\n\"\n end\nend",
"def FizzBuzz\n\t\n1.upto(100) {|i| \n\tputs \"Fizz\" if i%3 == 0 && i%5 != 0\n\tputs \"Buzz\" if i%3 != 0 && i%5 == 0\n\tputs \"FizzBuzz\" if i%3 == 0 && i%5 == 0\n\tputs i if i%3 != 0 && i%5 != 0\n}\t\n\nend",
"def alpha_search(str)\r\n\r\nend",
"def tt; end",
"def test2\n i = 100\n j = 200\n k = 300\nreturn i, j, k\nend",
"def startstop minsize=30\n stopstop(minsize).find_all { | orf | \n codon1= orf.nt.seq[0..2].upcase\n ['ATG','TTG','CTG','AUG','UUG','CUG'].index(codon1) != nil\n }\n end",
"def fizzbuzz(start, finish)\n (start..finish).each do |x|\n output = case \n when x % 3 == 0 && x % 5 == 0\n 'FizzBuzz'\n when x % 3 == 0\n 'Fizz'\n when x % 5 == 0\n \"Buzz\"\n else x\n end\n \n \n puts output\n end\nend",
"def z\n end",
"def string_behavior (strings)\n\n strings.each do |str|\n if not str =~ /[A-Z]/ and str.length > 4\n puts \"long and lowercase\"\n elsif not str =~ /[A-Z]/\n puts \"lowercase\"\n elsif str.length > 4\n puts \"long\"\n\n else\n puts str\n end\n end\n end",
"def _reduce_101(val, _values, result)\n index = @block_parser.add_footnote val[1].rdoc\n result = \"{*#{index}}[rdoc-label:foottext-#{index}:footmark-#{index}]\"\n\n result\nend",
"def skips=(_arg0); end",
"def each(&block); end",
"def each(&block); end",
"def each(&block); end",
"def each(&block); end",
"def each(&block); end",
"def each(&block); end",
"def line_range=(_arg0); end"
] |
[
"0.57231677",
"0.5534685",
"0.5487405",
"0.54800254",
"0.54774016",
"0.53817517",
"0.5361834",
"0.5361834",
"0.53134114",
"0.5307534",
"0.5299324",
"0.52880657",
"0.5272075",
"0.5254862",
"0.5244559",
"0.5244559",
"0.5244559",
"0.5241725",
"0.5236698",
"0.52181625",
"0.5217735",
"0.52144414",
"0.52144414",
"0.520647",
"0.5205659",
"0.5205659",
"0.5195368",
"0.51948136",
"0.51888585",
"0.516894",
"0.51606333",
"0.51606333",
"0.51606333",
"0.5146415",
"0.51439613",
"0.5141644",
"0.5138522",
"0.5138522",
"0.5129676",
"0.5129348",
"0.51215094",
"0.51215094",
"0.51205075",
"0.51205075",
"0.51205075",
"0.51205075",
"0.51205075",
"0.51205075",
"0.51205075",
"0.51205075",
"0.51205075",
"0.51205075",
"0.51205075",
"0.51109004",
"0.5108604",
"0.51069015",
"0.5088054",
"0.5085912",
"0.5083391",
"0.5081462",
"0.50764006",
"0.50755596",
"0.5073618",
"0.5069832",
"0.50648916",
"0.5062747",
"0.5058474",
"0.50302505",
"0.50280803",
"0.50267047",
"0.50250953",
"0.5021904",
"0.5020638",
"0.5019497",
"0.5018739",
"0.5006408",
"0.49995795",
"0.49954838",
"0.4987809",
"0.4987285",
"0.4977702",
"0.49768198",
"0.49742618",
"0.4973824",
"0.49665678",
"0.4959971",
"0.4959806",
"0.4957351",
"0.49541423",
"0.49504796",
"0.49500132",
"0.49486914",
"0.49416208",
"0.4935094",
"0.4924217",
"0.4924217",
"0.4924217",
"0.4924217",
"0.4924217",
"0.4924217",
"0.49241707"
] |
0.0
|
-1
|
find_cool(ttt) puts ttt[0] puts ttt[1] find_cool(ttt) =begin
|
def find_cool(abc)
xyz=[]
abc.each do |i|
if i[:temperature]=="cool"
print i
xyz=i
end
end
#print xyz
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def find_it(seq)\r\n\r\nend",
"def beer_search_first(arg)\n beer_search(arg).find {|beer|beer}\nend",
"def find; end",
"def find(t)\n text = t\n text.downcase! unless @case_sensitive\n text.gsub!(/\\s+/,' ') # Get rid of multiple spaces.\n @state = 0\n index = 0\n text.each_char do |char|\n # Incrementing now so that I can announce index - @length.\n index += 1\n @state = step(@state,char)\n if @state == @length # Yay, we've got ourselves a match!\n puts \"Match found for #{@word[1,@length]} at #{index - @length}\"\n @state = 0\n end\n end\n end",
"def work!\n (@start..@stop).each {|i| return if $found; d = $data[i]; $found = true and puts \"Found #{d}\" if d == $search_key }\n end",
"def find_it(words, &block)\n words.sort(&block).first\nend",
"def next2(t) \n @start1 = t \n return @start1 \n end",
"def find(input)\n end",
"def slow_dance(target, array)\n\n array.each.with_index do |phrase, i|\n return i if array[i] == target \n end \n end",
"def tt_current\n tt_intersect()\n end",
"def scan_for_at(token); end",
"def find_needle(haystack)\n \"found the needle at position #{haystack.index('needle')}\"\nend",
"def find_item(word, array)\n iter = 0\n while iter < array.length\n # is \"this\" same as \"the\"?\n if word == array[iter]\n return iter + 1 #in this case, iter +1 will add 1 to the position number\n end\n\n iter += 1\n end\n puts \"#{word} doesn't exist\"\nend",
"def find_needle(haystack)\n position = 0\n haystack.each do |element|\n if element == \"needle\"\n puts \"found the needle at position #{position}\"\n else\n position += 1\n end\n end\nend",
"def match(tt)\n \ttt = \"#{tt}\"\n tt = tt.to_s.split(',')\n out = Cites.match(tt)\n puts out\n end",
"def first_wa(array)\n strings = array.join(\",\")\n # puts \"strings are #{strings}\"\n strings_in_array = strings.split(\",\")\n # puts \"strings in array #{strings_in_array}\"\n strings_in_array.find do |w| \n what_do_you_return = w.start_with?(\"wa\")\n # puts \"what do you return #{what_do_you_return}\"\n end\nend",
"def begin_pos=(_); end",
"def tt; end",
"def find_the_cheese(food) #method and argument\n cheese_types = [\"cheddar\", \"gouda\", \"camembert\"] #array\n food.find do |i|\n if cheese_types.include?(i)\n return i\n else\n end\n nil\n end\nend",
"def find_word(lexical_item)\n \n end",
"def first_index(arr, &blck)\n #arr.find_index { |el| blck.call(el) }\n arr.each_with_index do |el, i|\n if blck.call(el)\n return i\n end\n end\n return nil\nend",
"def first_wa(arrays)\narrays.find_index { |word| if (word.to_s.include? (\"wa\"))\n return word\n break\nend\n}\nend",
"def find_the_cheese(array)\n array.find do |cheese|\n cheese == \"cheddar\" || cheese == \"gouda\" || cheese == \"camembert\"\n \n end \n end",
"def find_the_cheese(array) \n array.find do |cheese|\n cheese == \"cheddar\" || cheese == \"gouda\" || cheese == \"camembert\"\n end \nend",
"def findARG(thing, index)\r\n\r\n aux = thing.strip.downcase\r\n result = aux.split\r\n\r\n first_word = result[0]\r\n second_word = result[1]\r\n\r\n len = first_word.length\r\n\r\n if index == 0 && first_word.start_with?(\"!\")\r\n target = first_word.slice(1..len)\r\n return target\r\n end\r\n\r\n if index == 1\r\n return second_word\r\n end\r\nend",
"def hello_t(array)\n i = 0\n names = []\n while i < array.length\n if array[i].start_with?(\"T\", \"t\")\n yield(array[i])\n names.push(array[i])\n end\n i = i + 1\n end\n names\nend",
"def bulk_tweet_shortener(array)\n array.each do |tweet|\n puts word_substituter(tweet)\n end\n #binding.pry\nend",
"def search(head=nil, tail=nil )\n head = 0 unless head # maybe backwards 200 changelists (or of yesterday's qa approved changelist.)\n # tail = changelist.size() - 1\n tail = 4 unless tail\n mid = nil\n # TODO: ensure the head is ok and tail is not ok, thus search is meaningful!\n while !( (mid == head) or (mid == tail) ) \n mid = ((head + tail ) / 2 ).round()\n puts \"-----------------\"\n puts \"ready to search, head:#{head} mid:#{mid} tail:#{tail} ...\"\n # ok = try_this_changelist(changelists[mid])\n ok = try_this_changelist(mid)\n puts \"deteck ok? .... #{ok}\"\n # log it\n if ok\n head = mid\n else\n tail = mid\n end\n mid = ((head + tail ) / 2 ).round()\n puts \"ranged changed to, head:#{head} mid:#{mid} tail:#{tail} ...\"\n end\n\n culprit = nil\n \n # TODO: just guess, not maths-proven\n if mid == head\n print \"error found: index is #{tail}\" \n culprit = tail\n elsif mid == tail\n print \"error found: index is #{head + 1}\" \n culprit = head + 1\n else\n raise \"Interrupted, not go down enough\"\n end\n \n return culprit\n end",
"def find_a(array)\n array.find_all do |word|\n word[0] == \"a\"\n end\nend",
"def linear_search(object, array)\n\t\tputs \"nil\" if array.empty? == true\n\t\tputs \"nil\" if array.include?(object) == false\n\t\n\t\t\ti = 0\n\t\tuntil array[i] == object\n\t\t\ti +=1\n\t\tend\n\t\nreturn i if array[i] = object\n\t\nend",
"def find_word(search, word, maxRow, maxCol)\n (0..maxRow).each do |i|\n (0..maxCol).each do |j|\n if starts_here(search, word, i, j)\n puts \"The word '#{word}' starts at (#{i},#{j}).\"\n return\n end\n end\n end\n puts \"The word '#{word}' was not found!\"\nend",
"def test_0150_find\n @@log.debug \"test_0150_find starts\" if @@log.debug?\n assert_respond_to(@list, :find, \"test_0150_find_respond\")\n # Object with .last == \"Dev\"\n persobj = @list.find {|obj| obj.last == \"Dev\" }\n assert_equal(@dad, persobj, \"test_0150_find_feq_01\")\n # Object with .last == \"Allard\"\n sorry = lambda { \"not found\" }\n persobj = @list.find(sorry) {|obj| obj.last == \"Allard\" } \n assert_equal(\"not found\", persobj, \"test_0150_find_feq_02\")\n # Check Enumerator or Enumerable::Enumerator return, no block given\n new_list = @list.find\nif RUBY_VERSION >= \"1.9\"\n result = new_list.is_a? Enumerator\n assert(result, \"test_0150_find_enumcheck\")\nelse\n result = new_list.is_a? Enumerable::Enumerator\n assert(result, \"test_0150_find_enumenumcheck\")\nend\n @@log.debug \"test_0150_find ends\" if @@log.debug?\n end",
"def test_upper_lower\n assert_equal(\"A\", @target.find_first_non_repeating(\"aaaAaaa\"))\n end",
"def at(index); end",
"def start_with_h(arr)\n arr.find {|word| word[0] == \"h\"}\nend",
"def element_selection(array)\n puts array[0]\n puts array[-1]\n puts array.first(2)\n return array.last(2)\nend",
"def find_cool(cool)\n ans = []\n cool.each do |data|\n ans << data if data[:temperature] == \"cool\"\n end\n ans\nend",
"def find_needle(haystack)\r\n index = haystack.find_index('needle')\r\n haystack.find_index('needle')? \"found the needle at position #{index}\" : nil\r\nend",
"def find(value)\n end",
"def find_needle(haystack)\n needle = haystack.index('needle')\n return \"found the needle at position #{needle}\"\nend",
"def find_a(array)\n new_a_array = []\n #binding.pry\n array.select do |item|\n if item[0]==\"a\"\n new_a_array<< item\n else\n puts \"nil\"\n end\n end\nnew_a_array\nend",
"def find_the_cheese(foodstuffs)\n cheese_types = [\"cheddar\", \"gouda\", \"camembert\"]\n foodstuffs.find do |cheese|\n if cheese_types.include?(cheese)\n return cheese\n end\nend\nend",
"def slow_dance(tile, array)\n array.each_with_index do |el, i|\n return i if el == tile\n end\nend",
"def second_great_low(arr)\r\n\r\nend",
"def find_a(array)\n array.find_all do |word|\n word[0] == \"a\"\n end\nend",
"def find_a(array)\n array.find_all do |word|\n word[0] == \"a\"\n end\nend",
"def bulk_tweet_shortener(tweet_array)\n tweet_array.each do |tweet|\n #binding.pry\n puts word_substituter(tweet)\n end\nend",
"def find_the_cheese(array)\n array.find do |cheese| \n cheese == \"cheddar\" || cheese == \"gouda\" || cheese == \"camembert\"\n end \n end",
"def start_with_h(arr_with_h)\n #iterate into the array and when found the first string finish\n arr_with_h.find {|letter| letter[0] == \"h\"}\nend",
"def search_insert_position(nums, target)\n\nend",
"def first() end",
"def find_index(a, target) # this is where returning directly from withing block is ok!\n a.each_with_index do |e, index|\n return index if e == target # returns to caller on line 19. This is exactly what we want\n end\n return nil # omitting return here is OK and doing so is more idiomatic\nend",
"def find_within(tags, span, pointer); end",
"def find_within(tags, span, pointer); end",
"def begin(p0) end",
"def scan_for_t(token); end",
"def find list, &block\n list.find &block\nend",
"def first_in_array(arr, el1, el2)\n\tarr.each_with_index do |letter, i|\n \tif (letter == el1)\n return el1\n end\n \tif (letter == el2)\n return el2\n end\n end\nend",
"def find_cool(array)\n array.select {|word| word[:temperature] && word[:temperature] == \"cool\"}\nend",
"def start_with_h(array)\n array.find {|word| word[0] == 'h'}\nend",
"def find_it(seq)\n seq.sort!\n res = seq[0]\n res = find_it(seq[2..seq.length]) if seq[1] == res\n res\nend",
"def my_array_finding_method(i_want_pets, thing_to_find)\n new_array = []\n i_want_pets.each do |word|\n range = word.to_s.downcase[0..-1]\n if range.include?(thing_to_find)\n new_array << word\n end\n end\n return new_array\nend",
"def find_the_cheese(no_cheese)\n cheese_types = [\"cheddar\", \"gouda\",\n \"camembert\"]\n no_cheese.find {|cheese| cheese_types.include?(cheese)}\nend",
"def find_it(seq)\n seq.find { |n| seq.count(n).odd? } # аналог метода find метод detect\nend",
"def return_t(hash)\n hash.each do |key, value|\n if value.to_s.start_with?(\"t\")\n puts value\n end\n end\nend",
"def find_a(array)\n array.select do |word|\n word[0] == \"a\"\n end\nend",
"def get_the_cheese(arr)\n cheeses = ['cheddar', 'gouda', 'cambert']\n arr.find { |x| cheeses.include?(x) }\n end",
"def first_element(my_first_element)\n my_first_element = @taylor_swift[1-1]\nend",
"def find_a (arr)\n arr.select do |i|\n i[0] == \"a\"\n end\nend",
"def first_wa(strings)\n strings.find do |string|\n string[0] ==\"w\" && string[1] == \"a\"\n end\nend",
"def begin_pos; end",
"def linear_search_simple(a, v)\n\ta.find_index(v)\nend",
"def first; end",
"def first; end",
"def find(name); end",
"def my_find(collection)\n i = 0 \n while i < collection.length \n return collection[i] if\n yield(collection[i])\n i = i += 1 #i can put += if its easier to remember\n end\nend",
"def find_index list, &block #were taking a list, and a block we want to find, find_index = func name\n list.index(&block) #look in list using .index(&block) to find the index of the block \nend",
"def find_code_insert_point (fileArray, utilHash)\n find_match_index(fileArray, utilHash[\"match_code\"])\nend",
"def find_cool(cool)\n cool.select { |data| data[:temperature] == \"cool\" }\nend",
"def my_array_finding_method(source, thing_to_find)\n p final_array = source.select {|test| test.to_s.include?(thing_to_find.to_s)}\nend",
"def my_array_finding_method(source, thing_to_find)\n p final_array = source.select {|test| test.to_s.include?(thing_to_find.to_s)}\nend",
"def first_offset; end",
"def first_offset; end",
"def b_find_number(n, number_to_find)\n n.bsearch {|x| puts x }\nend",
"def begin() end",
"def hint()\r\n #Get the first _ position in the shadow word\r\n returnValue=\"\"\r\n hintIndex = @shadowWord.index(\"_\");\r\n if(@count > 1)\r\n #Decrement the count\r\n @count-=1\r\n #Return the hind index\r\n returnValue = @word[hintIndex]\r\n end\r\n returnValue\r\n end",
"def found\n return @found\n end",
"def saved_for_special_occasion(array)\n puts array[2]\nend",
"def test_9_return_positions_of_letter_in_word\r\n $word = \"repeated\"\r\n letter = \"e\"\r\n results = find_locations(letter)\r\n assert_equal([1, 3, 6], results)\r\n end",
"def at(expr, &blk)\n search(expr, &blk).first\n end",
"def first_wa (array)\n array.find { |word| word[0..1] == \"wa\" }\nend",
"def find_a(array)\n array.select do |word|\n word[0] == \"a\"\n end\nend",
"def before(complete = T.unsafe(nil), index = T.unsafe(nil), &block); end",
"def nameri arr, query \n len = arr.length\n for idx in 0 .. len - 1 do\n if query == arr[idx]\n return idx \n else \n end\n end\nend",
"def look_and_say(array)\n\nend",
"def find(thing)\n case thing\n when Symbol\n find_by_tag(thing)\n when String\n find_by_string(thing)\n when Array\n find_by_string(thing.join(' '))\n when Node\n thing\n end\n end",
"def find(prepostion)\n\t\tpartial_search_kb = string_to_internal(preposition)\n\t\tpartial_search_kb.each do |sentence|\n\t\t\tind = @kb.index(sentence)\n\t\tend\n\t\treturn ind\n\tend",
"def search_array(array, integer)\r\n # Search array for given integer\r\n array.each_with_index do |item, index|\r\n\r\n if item.eql? integer\r\n p index\r\n elsif !item.eql? integer\r\n nil\r\n else puts \"nil\"\r\n end\r\n \r\n end\r\nend",
"def lookup(obj, pos); end",
"def slow_dance(tile, tiles_array)\n tiles_array.each_with_index do |direction, idx|\n return idx if direction == tile\n end\nend"
] |
[
"0.6378533",
"0.60212594",
"0.5975574",
"0.58719736",
"0.5759661",
"0.5624258",
"0.5575567",
"0.55651367",
"0.5559199",
"0.5477042",
"0.5470051",
"0.5444748",
"0.5424308",
"0.5420687",
"0.54182726",
"0.54171497",
"0.5416891",
"0.5416079",
"0.54078174",
"0.5387714",
"0.5374093",
"0.53633314",
"0.53313184",
"0.5316537",
"0.5314265",
"0.5310169",
"0.5292803",
"0.5284798",
"0.52817076",
"0.5269562",
"0.5259944",
"0.5251228",
"0.5242211",
"0.52381456",
"0.52255225",
"0.52249855",
"0.52229273",
"0.5221251",
"0.5218746",
"0.5216725",
"0.5209235",
"0.52074486",
"0.5204986",
"0.5196786",
"0.51952326",
"0.51952326",
"0.51950645",
"0.51900315",
"0.5180849",
"0.5179224",
"0.5175875",
"0.51730174",
"0.5172094",
"0.5172094",
"0.51678795",
"0.5163706",
"0.51563793",
"0.5150722",
"0.514959",
"0.5149579",
"0.51472795",
"0.51435244",
"0.5131241",
"0.51307046",
"0.51184744",
"0.51173556",
"0.51119864",
"0.5111101",
"0.51010185",
"0.5098249",
"0.5096313",
"0.50931805",
"0.509189",
"0.509189",
"0.5082889",
"0.50748384",
"0.5070371",
"0.50627315",
"0.5059656",
"0.5055699",
"0.5055699",
"0.505325",
"0.505325",
"0.50497043",
"0.5048836",
"0.50400275",
"0.50351393",
"0.50345355",
"0.5034101",
"0.5031851",
"0.5028849",
"0.502706",
"0.5026953",
"0.50266016",
"0.5025078",
"0.50238866",
"0.5018181",
"0.5007764",
"0.49971136",
"0.4994763"
] |
0.53779054
|
20
|
Delegate all method calls we don't know about to target object
|
def method_missing(sym, *args, &block)
proxy_target.__send__(sym, *args, &block)
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def delegating_method; end",
"def delegate_all\n # noop\n end",
"def method_missing(method_name, *args, &block)\n if target.respond_to?(method_name)\n target.public_send(method_name, *args, &block)\n else\n super\n end\n end",
"def method_missing(method, *args, &block)\n _target.send method, *args, &block\n end",
"def method_missing(m, *args, &block)\n target = self.__getobj__\n unless target.respond_to?(m)\n super(m, *args, &block)\n else\n target.__send__(m, *args, &block)\n end\n end",
"def method_missing(method, *args, &block)\n @delegate.send(method, *args, &block)\n rescue NoMethodError => e\n raise e, e.message.sub(/ for #<.*$/, \" via proxy for #{@delegate}\")\n end",
"def method_missing(id,*args,&block)\n return @delegated_to_object.send(id,*args,&block) if delegated_instance_methods.include?(id)\n super\n end",
"def method_missing(id,*args,&block)\n return @delegated_to_object.send(id,*args,&block) if delegated_instance_methods.include?(id)\n super\n end",
"def method_missing(method, *args, &block)\n send(target_method(method), method, *args, &block).tap do \n create_delegation_method!(method)\n end\n end",
"def delegate_method; end",
"def method_missing(method, *args, &block)\n super unless original_self\n original_self.send method, *args, &block\n end",
"def method_missing(method, *args, &block)\n generate_unsafe_delegation(method)\n send(method, *args, &block)\n end",
"def method_missing(method, *args, &block)\n send(target_method(method), method, *args, &block).tap do\n create_delegation_method!(method)\n end\n end",
"def method_missing(name, *args, &block)\n target.send(name, *args, &block)\n end",
"def method_missing(method_name, *args, &block)\n if respond_to_missing?(method_name)\n handle_known_exceptions do\n @target.send(method_name, *args, &block)\n end\n else\n super\n end\n end",
"def delegate_object; end",
"def doTarget _obj, _args\n \"_obj doTarget _args;\" \n end",
"def method_missing(sym, *args, &block)\n proxy_target.__send__(sym, *args, &block)\n end",
"def original_method(obj, meth, klass=Object)\n klass.instance_method(meth).bind(obj).call\n end",
"def original_method(obj, meth, klass=Object)\n klass.instance_method(meth).bind(obj).call\n end",
"def normal_dispatch\n emit_naked_receiver\n emit_selector_replacement\n mutate_receiver\n emit_argument_propagation\n mutate_arguments\n end",
"def delegate_public_instance_methods(options)\n from = options.fetch(:from)\n\n methods = (send(from).public_instance_methods - Object.methods).\n reject {|m| m =~ /(?:=|!|_changed\\?|change|_will_change|_was)$/}\n\n target = options.fetch(:to)\n\n methods.each do |method|\n define_method(method) do |*args, &block|\n\n instance = instance_variable_get(target) # @person or @account\n\n if instance.nil?\n raise Error,\n \"#{self.class}##{method} delegates to #{target} but it's nil\"\n end\n\n instance.public_send(method, *args, &block)\n end\n end\n end",
"def method_missing(name, *args, &block)\n target.send(name, *args, &block)\n end",
"def method_missing(m, *args, &block)\n # ignore\nend",
"def method_missing(method_name, *args, &block)\n # If `delegate_all_to` has been set up for the method name\n # then delegate to it, otherwise pass it up the food chain\n if !delegating_all_to? && delegate_all_responds_to?(method_name)\n delegate_all_to.send(method_name, *args, &block)\n else\n super\n end\n end",
"def __delegate\n@o\nend",
"def ignore_method_conflicts; end",
"def method_missing(method, *args, &block)\n self.class.define_proxy method\n send(method, *args, &block)\n end",
"def wrap_in_tracing_delegator(object)\n delegator = Object.new\n\n def delegator.method_missing(meth, *args, &blk)\n p :meth => meth, :args => args\n @delegate.send(meth, *args, &blk)\n end\n\n delegator.instance_variable_set(\"@delegate\", object)\n delegator\n end",
"def method_missing (method, *args, &block)\n __proxy_result__.send(method, *args, &block)\n end",
"def method_missing(name, *args, &block)\n @target.send(name, *args, &block)\n end",
"def with_target(method)\n target_context.push method\n yield\n target_context.pop\n nil # should not chain\n end",
"def ignore_method_conflicts=(_arg0); end",
"def method_missing(method, *args, &block)\n __proxy_result__.send(method, *args, &block)\n end",
"def method_missing(*args, &block)\nm = args.shift\n# First arg is the name of the method\nbegin\n# Trace the invocation of the method.\narglist = args.map {|a| a.inspect}.join(', ')\n@trace << \"Invoking: #{@n}.#{m}(#{arglist}) at #{caller[0]}\\n\"\n# Invoke the method on our delegate object and get the return value.\nr = @o.send m, *args, &block\n# Trace a normal return of the method.\n@trace << \"Returning: #{r.inspect} from #{@n}.#{m} to #{caller[0]}\\n\"\n# Return whatever value the delegate object returned.\nr\nrescue Exception => e\n# Trace an abnormal return from the method.\n@trace << \"Raising: #{e.class}:#{e} from #{@n}.#{m}\\n\"\n# And re-raise whatever exception the delegate object raised.\nraise\nend\nend",
"def original_method; end",
"def disable_unsafe_method(target, *names, &block)\n block ||= ->(*) {}\n names.each { |name| target.define_singleton_method name, &block }\n end",
"def method_missing(method, *args, &block)\n if delegate.respond_to?(method)\n delegate.public_send(method, *args, &block)\n else\n super\n end\n end",
"def method_missing(method, *args, &block)\n if respond_to_missing?(method)\n source_object.send(method, *args, &block)\n else\n super\n end\n end",
"def method_missing selector, *arguments, &block\n _log { \"#{selector} #{arguments.inspect}\" }\n if @target_list.empty?\n super\n else\n target = @mutex.synchronize do\n @target_list[@target_index = \n (@target_index + 1) % @target_list.size]\n end\n raise Error, \"No target\" unless target\n target.method_missing(selector, *arguments, &block)\n end\n end",
"def method_missing(meth, *args, &block)\n if respond_to?(meth)\n object.__send__(meth, *args, &block)\n else\n super\n end\n end",
"def to_proxy(*)\n raise NotImplementedError, \"#{__method__} must be implemented in a descendant class!\"\n end",
"def to_proxy(*)\n raise NotImplementedError, \"#{__method__} must be implemented in a descendant class!\"\n end",
"def ignore_unexpected_calls!\n return if @ignoring_unexpected_calls # do it once!\n\n @ignoring_unexpected_calls = true\n\n def self.method_missing(sym, *args)\n super if @expected_calls.has_key?(sym)\n end\n\n def self.respond_to?(sym)\n true\n end\n end",
"def add_method_object_options(opt)\n @method_target = target\n opt.on :M, \"instance-methods\", \"Operate on instance methods.\"\n opt.on :m, :methods, \"Operate on methods.\"\n opt.on :s, :super, \"Select the 'super' method. Can be repeated to traverse the ancestors.\"\n opt.on :c, :context, \"Select object context to run under.\", true do |context|\n @method_target = Pry.binding_for(target.eval(context))\n end\n end",
"def method_missing(method, *args, &block)\n @targets.each { |t| t.send(method, *args, &block) }\n end",
"def method_missing(meth, *args, &block)\n if !(meth.to_s =~ /=/) and (suffixed_method = \"#{meth.to_s}__c\") and respond_to?(suffixed_method)\n send(suffixed_method)\n else\n super\n end\n end",
"def method_missing(*args, &block)\n m = args.shift # Firt arg is the name of the method \n begin \n # Trace the invocation of the method.\n arglist = args.map {|a| a.inspect }.join(', ')\n @trace << \"Invoking: #{@n}.#{m}(#{arglist}) at #{caller[0]}\\n\"\n # Invoke the method on our delegate object and get the return value. \n r = @o.send m, *args, &block\n # Trace a normal return of the method. \n @trace << \"Returning: #{r.inspect} from #{@n}.#{m} to #{caller[0]}\\n\"\n # Return whaterver value the delegate object returned.\n r\n rescue Exception => e\n # Trace an abnormal return from the method.\n @trace << \"Rasing: #{e.class}:#{e} from #{@n}.#{m}\\n\"\n # And re-raise whaterver exception the delegate object raised.\n raise \n end \n end",
"def delegate(*methods)\n options = methods.pop\n unless options.is_a?(Hash) && to = options[:to]\n raise ArgumentError, \"Delegation needs a target. Supply an options hash with a :to key as the last argument (e.g. delegate :hello, :to => :greeter).\"\n end\n\n if options[:prefix] == true && options[:to].to_s =~ /^[^a-z_]/\n raise ArgumentError, \"Can only automatically set the delegation prefix when delegating to a method.\"\n end\n\n prefix = options[:prefix] && \"#{options[:prefix] == true ? to : options[:prefix]}_\" || ''\n\n file, line = caller.first.split(':', 2)\n line = line.to_i\n\n methods.each do |method|\n on_nil =\n if options[:allow_nil]\n 'return'\n else\n %(raise \"#{self}##{prefix}#{method} delegated to #{to}.#{method}, but #{to} is nil: \\#{self.inspect}\")\n end\n\n module_eval(<<-EOS, file, line - 5)\n if instance_methods(false).map(&:to_s).include?(\"#{prefix}#{method}\")\n remove_possible_method(\"#{prefix}#{method}\")\n end\n\n def #{prefix}#{method}(*args, &block) # def customer_name(*args, &block)\n #{to}.__send__(#{method.inspect}, *args, &block) # client.__send__(:name, *args, &block)\n rescue NoMethodError # rescue NoMethodError\n if #{to}.nil? # if client.nil?\n #{on_nil} # return # depends on :allow_nil\n else # else\n raise # raise\n end # end\n end # end\n EOS\n end\n end",
"def method_missing(method, *args, &block)\n return super unless object.respond_to?(method)\n\n object.public_send(method, *args, &block)\n end",
"def flexmock_invoke_original(object, method, *args, &block)\n object.instance_variable_get(:@flexmock_proxy).proxy.flexmock_invoke_original(method, args, &block)\n end",
"def flexmock_invoke_original(object, method, *args, &block)\n object.instance_variable_get(:@flexmock_proxy).proxy.flexmock_invoke_original(method, args, &block)\n end",
"def method_missing(method, *args)\n if reload\n unless @target.respond_to?(method)\n message = \"undefined method `#{method.to_s}' for \\\"#{@target}\\\":#{@target.class.to_s}\"\n raise NoMethodError, message\n end\n\n if block_given?\n @target.send(method, *args) { |*block_args| yield(*block_args) }\n else\n @target.send(method, *args)\n end\n end\n end",
"def method_missing(method, *args, &block)\n #$stderr.puts \"Method missing: #{method}\"\n if @_receivers.last.respond_to?(method)\n #$stderr.puts \"Proxy [#{method}] to receiver\"\n @_receivers.last.__send__(method, *args, &block)\n else\n #$stderr.puts \"Proxy [#{method}] to context\"\n @_context.__send__(method, *args, &block)\n end\n end",
"def aimedAtTarget _obj, _args\n \"_obj aimedAtTarget _args;\" \n end",
"def method_missing(method, *arguments, &block); end",
"def call(object); end",
"def underscorize\n dup.tap(&:underscorize!)\n end",
"def delegate_to klass_or_object\n @delegators.each do |data|\n unless data.empty?\n name = data[:name]\n args = data[:params]\n klass_or_object.send(name, *args)\n end\n end\n end",
"def method_missing(method, *args, &blk); end",
"def method_missing(method, *args, &block)\n Rails.logger.debug(\"#{self.class.name}##{__method__}() looking for: #{method.to_s}\")\n if proxy_u.respond_to?(method)\n block_given? ? proxy_u.send(method, *args, block) :\n (args.size == 0 ? proxy_u.send(method) : proxy_u.send(method, *args))\n else\n super\n end\n end",
"def method_missing(call)\n object.send(call)\n end",
"def stop_intercepting!\n @original_methods.each do |m|\n class_eval do\n alias_method m, \"intercepted_#{m}\"\n remove_method \"intercepted_#{m}\"\n end\n end\n\n class_eval do\n remove_method :synthesis_expectation\n remove_method :synthesis_expectation=\n end\n end",
"def method_missing(meth, *args, &block)\n @object.send(meth, *args, &block)\n end",
"def alias_class_method_chain(target, feature)\n # Strip out punctuation on predicates or bang methods since\n # e.g. target?_without_feature is not a valid method name.\n aliased_target, punctuation = target.to_s.sub(/([?!=])$/, ''), $1\n yield(aliased_target, punctuation) if block_given?\n alias_class_method \"#{aliased_target}_without_#{feature}#{punctuation}\", target\n alias_class_method target, \"#{aliased_target}_with_#{feature}#{punctuation}\"\n end",
"def method_object\n @method_object ||= get_method_or_raise(args.empty? ? nil : args.join(\" \"), @method_target,\n :super => opts[:super],\n :instance => opts.present?(:'instance-methods') && !opts.present?(:'methods'),\n :methods => opts.present?(:'methods') && !opts.present?(:'instance-methods')\n )\n end",
"def method_missing(meth, *args, &block); end",
"def method_missing(meth, *args, &block); end",
"def method_missing(meth, *args, &block); end",
"def alias_method_chain(target, feature)\n # Strip out punctuation on predicates or bang methods since\n # e.g. target?_without_feature is not a valid method name.\n aliased_target, punctuation = target.to_s.sub(/([?!=])$/, ''), $1\n yield(aliased_target, punctuation) if block_given?\n\n with_method, without_method = \"#{aliased_target}_with_#{feature}#{punctuation}\", \"#{aliased_target}_without_#{feature}#{punctuation}\"\n\n alias_method without_method, target\n alias_method target, with_method\n\n case\n when public_method_defined?(without_method)\n public target\n when protected_method_defined?(without_method)\n protected target\n when private_method_defined?(without_method)\n private target\n end\n end",
"def call_without_propagation(context) # :nodoc:\n super\n \trescue EventNotExecutable => e\n\t refine_call_exception(e)\n end",
"def call_without_propagation(context) # :nodoc:\n super\n \trescue EventNotExecutable => e\n\t refine_call_exception(e)\n end",
"def def_delegators_with_super (target, *methods)\n methods.each do |method|\n target_method = \"#{method}_without_super\".to_sym\n def_delegator(target, method, target_method)\n\n define_method(method) do |*args, &block|\n send(target_method, *args, &block)\n super(*args, &block)\n end\n end\n end",
"def method_missing(method, *args, &block)\n object.public_send(method, *args, &block)\n end",
"def chained_call(*methods)\n raise \"Not implemented yet.\"\n end",
"def test_Method_InstanceMethods_unbind\n\t\tpass\n\tend",
"def methods(*args)\n (super + analyser.delegatable_methods).uniq\n end",
"def decorated_object_behavior\n #code\n end",
"def method_missing(method, *args, &block); end",
"def do_invocation()\n obj = get_receiver_object\n if obj.nil?\n Undef\n else\n begin\n obj.send(@name)\n rescue NameError\n Undef\n end\n end\n end",
"def default_delegation_behavior\n\t\t@delegate_called = true\n\tend",
"def suppressFor _obj, _args\n \"_obj suppressFor _args;\" \n end",
"def method_missing method, *opt\n method.to_s =~ /^do_[A-Z]+$/ ? process_request(*opt) : super\n end",
"def send_to_target(name, *args, &block)\n raise \"Mutable methods not allowed\" unless allowed?(name)\n @target.__send__(name, *args, &block)\n end",
"def method_missing(m, *args, &_block)\n set_accessors(m)\n if valid_proxy_method?\n define_proxy_method(m)\n return send(m, args.first)\n end\n\n super\n end",
"def method_missing(method_name, *args, &block)\n if @object.respond_to? method_name.to_sym\n @object.__send__(method_name, *args, &block)\n else\n super\n end\n end",
"def delegate(*methods)\n options = methods.pop\n unless options.is_a?(Hash) && to = options[:to]\n raise ArgumentError, \"Delegation needs a target. Supply an options hash with a :to key as the last argument (e.g. delegate :hello, :to => :greeter).\"\n end\n\n if options[:prefix] == true && options[:to].to_s =~ /^[^a-z_]/\n raise ArgumentError, \"Can only automatically set the delegation prefix when delegating to a method.\"\n end\n\n prefix = options[:prefix] && \"#{options[:prefix] == true ? to : options[:prefix]}_\"\n\n file, line = caller.first.split(':', 2)\n line = line.to_i\n\n methods.each do |method|\n on_nil =\n if options[:allow_nil]\n 'return'\n else\n %(raise \"#{self}##{prefix}#{method} delegated to #{to}.#{method}, but #{to} is nil: \\#{self.inspect}\")\n end\n\n module_eval(<<-EOS, file, line)\n def #{prefix}#{method}(*args, &block) # def customer_name(*args, &block)\n #{to}.__send__(#{method.inspect}, *args, &block) # client.__send__(:name, *args, &block)\n rescue NoMethodError # rescue NoMethodError\n if #{to}.nil? # if client.nil?\n #{on_nil}\n else # else\n raise # raise\n end # end\n end # end\n EOS\n end\n end",
"def apply(target, method, options = nil)\n select_filters(target, options).each{|block| target.send(method, &block) }\n end",
"def process_method_object_options(args, opts)\n opts[:instance] = opts['instance-methods'] if opts.m?\n # TODO: de-hack when we upgrade Slop: https://github.com/injekt/slop/pull/30\n opts.options[:super].force_argument_value opts.options[:super].count if opts.super?\n method_obj = get_method_or_raise(args.empty? ? nil : args.join(\" \"), @method_target, opts.to_hash(true))\n opts.on(:method_object, :default => method_obj)\n end",
"def call\n __proxied_call\n end",
"def intercept(klass, meth_name, type, &block)\n orig_name = \"aop_orig_#{meth_name}\".to_sym\n meth_name = meth_name.to_sym\n @intercepted_methods ||= Hash.new do |h,k| \n # h[class_name] = hash\n h[k] = Hash.new do |h,k|\n # h[class_name][method_name] = hash\n h[k] = Hash.new do |h,k| \n # h[class_name][method_name][interception_type] = array\n h[k] = []\n end\n end\n end\n \n make_interception = !@intercepted_methods[klass].has_key?(meth_name)\n @intercepted_methods[klass][meth_name][type] << block\n method_chain = @intercepted_methods[klass][meth_name]\n \n if make_interception\n klass.class_eval do\n alias_method orig_name, meth_name\n define_method(meth_name) do |*args|\n method_chain[:before].each { |m| m.call(self, args) }\n # The result of the callcc block will either be the last line in the actual\n # ruby block, or it will be whatever is passed as arguments when calling the \n # +abort_continuation+ proc\n callcc do |abort_continuation|\n # First lambda in chain is the call to the original method\n call_lambda = lambda { send(orig_name, *args) }\n method_chain[:around].each do |m|\n # Make a chain of lambdas that calls the previouly defined\n # lambda, thus creating a chain of around blocks that will\n # all finally reach the original method block\n prev_call_lambda = call_lambda\n call_lambda = lambda {\n # If +prev_call_lambda+ is called, the next around block in\n # chain until the last one which corresponds to the original method call\n # if +abort_continuation+ is called, then this loop is aborted and the\n # callcc block returns whatever was passed as an argument to the proc call\n m.call(self, args, prev_call_lambda, abort_continuation)\n }\n end\n result = call_lambda.call\n method_chain[:after].each { |m| m.call(self, result, args) }\n result\n end\n end\n end\n end\n end",
"def method_missing(undefined_meth_yet,*args,&block)\n \"#{self.class.to_s} doesn't know how to perform :#{undefined_meth_yet}\"\n end",
"def method_missing(meth, *args, &block)\n\n end",
"def method_undefined(*) end",
"def method_missing(wh,*therest)\n # xxx internal methods must be protected at some point\n end",
"def method_missing(name, *args, &block)\n return owning_object.send(name, *args, &block) if owning_object.respond_to?(name, true)\n super\n end",
"def method_missing(method_name, *args, &block)\n objects.send method_name, *args, &block\n end",
"def method_missing(meth, *args)\n if @target.respond_to? meth, true # The second `true` argument checks private methods.\n # Create an instance of the job through the proxy and \n # configure it with the options given to the proxy.\n @block.call configure @target.build_job(meth, *args)\n else\n super\n end\n end",
"def method_missing(meth, *args, &block)\n Interrogate.is_interrogatory?(meth) ? Interrogate.interrogate(meth, *args, &block) : super\n end",
"def method_missing symbol, *args, &block\n @forwarder.call symbol, *args, &block\n end"
] |
[
"0.6739356",
"0.6451455",
"0.64021194",
"0.6251275",
"0.61972314",
"0.6168438",
"0.61553776",
"0.61553776",
"0.612947",
"0.6107595",
"0.6094749",
"0.604592",
"0.6037904",
"0.5988997",
"0.5972005",
"0.5959359",
"0.5889798",
"0.58865637",
"0.58574474",
"0.58574474",
"0.58418",
"0.58406276",
"0.5840495",
"0.5827716",
"0.58251524",
"0.5791537",
"0.57606274",
"0.57391435",
"0.572623",
"0.57130593",
"0.57055044",
"0.5701132",
"0.5692083",
"0.56910485",
"0.5690088",
"0.5689734",
"0.56847864",
"0.56746316",
"0.56600773",
"0.56531847",
"0.565215",
"0.5644683",
"0.5644683",
"0.5612001",
"0.5611309",
"0.559487",
"0.5590727",
"0.55809367",
"0.55727905",
"0.5548502",
"0.5543548",
"0.5543548",
"0.553826",
"0.5519606",
"0.55191004",
"0.55116874",
"0.5504696",
"0.5501507",
"0.5499234",
"0.54915714",
"0.5481318",
"0.54795367",
"0.54744846",
"0.5471428",
"0.5466438",
"0.5454024",
"0.54475963",
"0.54475963",
"0.54475963",
"0.5445272",
"0.54452693",
"0.54452693",
"0.54312426",
"0.5427617",
"0.5426419",
"0.542506",
"0.5419175",
"0.5415975",
"0.54123735",
"0.5410888",
"0.5406018",
"0.5403177",
"0.5400906",
"0.5396614",
"0.5384277",
"0.5381148",
"0.5381117",
"0.53804094",
"0.5378538",
"0.5373054",
"0.53638995",
"0.5361149",
"0.535731",
"0.53558975",
"0.5354706",
"0.5333804",
"0.5330462",
"0.53304183",
"0.5319831",
"0.53171957"
] |
0.5902514
|
16
|
table data insert process
|
def insert_cako
sql = "insert into CacheTako values (?, ?, ?)"
return sql
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def insert_into(table, data)\n\t\tkeys = \"(#{data.keys.join(', ')})\"\n\t\tvalues = \"(#{data.values.map{ |value| \"'#{value}'\" }.join(', ')})\"\n\t\texecute_with_retry \"INSERT INTO #{table} #{keys} VALUES #{values}; \"\n\tend",
"def insert_by_data data, table\n sql = \"insert into #{table} \"\n case data\n when Array\n data.each do |d|\n insert_by_data(d, table)\n end\n when Hash\n columns = data.keys.to_s.gsub('[','(').gsub(']',')').gsub('\"','')\n values = data.values.to_s.gsub('[','(').gsub(']',')').gsub('nil','NULL')\n sql = sql + columns + \" values \" + values\n query(sql)\n end\n end",
"def _insert\n return super if model.cti_tables.length == 1\n model.cti_models.each do |m|\n v = {}\n m.cti_table_columns.each{|c| v[c] = @values[c] if @values.include?(c)}\n ds = use_server(m.cti_instance_dataset)\n if ds.supports_insert_select? && (h = ds.insert_select(v))\n @values.merge!(h)\n else\n nid = ds.insert(v)\n @values[primary_key] ||= nid\n end\n end\n db.dataset.supports_insert_select? ? nil : @values[primary_key]\n end",
"def insert_record(table, values)\n execute table_insert_query(table, values)\n end",
"def insert_data(output, db, table)\n result = query(db, \"SELECT * FROM #{table}\", \"hash\")\n changes = false\n result.each do |row|\n to_insert(output, table, row)\n changes = true\n end\n output << \"\\n\" if changes\n end",
"def insert()\n\n end",
"def insert(table, id, attributes) # abstract\n end",
"def insert(data)\n query = \"INSERT INTO `#{@table_name}` \"+build_insert(data)\n\n begin\n queryresult = @mysql.query(query)\n rescue Exception => e\n @log.error(\"#{e}\")\n return nil\n end\n\n expire_table_cache(get_all_related_tables)\n\n if @auto_primary_key\n get_one({@primary_key => get_last_id})\n else\n get_one({@primary_key => data[@primary_key]})\n end\n end",
"def insert_table(table, identifier, data)\n if existing_data = get_table(table, identifier)\n if existing_data.empty? || !existing_data.has_key?('.members')\n data.each { |key, value| existing_data[key] = value }\n return data\n else\n error_report \"Error in insert_table: Redefinition of #{identifier}\"\n raise ParseError\n end\n end\n\n table['table_data'].push([identifier, data])\n table['quick_look'][identifier] = 1\n return data\nend",
"def insert_table(table, identifier, data)\n if existing_data = get_table(table, identifier)\n if existing_data.empty? || !existing_data.has_key?('.members')\n data.each { |key, value| existing_data[key] = value }\n return data\n else\n error_report \"Error in insert_table: Redefinition of #{identifier}\"\n raise ParseError\n end\n end\n\n table['table_data'].push([identifier, data])\n table['quick_look'][identifier] = 1\n return data\nend",
"def _insert_raw(ds)\n ds.insert(_insert_values)\n end",
"def _insert_raw(ds)\n ds.insert(_insert_values)\n end",
"def insert(object, table)\n sql = object.to_sql(table)\n execute(sql)\n end",
"def process_and_insert(dataset)\n begin\n dataset.each do |field|\n if field.nil? or field.empty?\n raise\n end\n end\n newdataset = parse_wifi_data(dataset)\n db_insert(newdataset)\n rescue Exception => e\n puts e.message\n puts e.backtrace.inspect\n ensure\n\n end\n\n end",
"def insert(sql, name = nil, pk = nil, id_value = nil) end",
"def db_insert table, fields= {}\n #client = Mysql2::Client.new(:host => \"localhost\", :username => \"root\", :password => \"toor\", :database => \"filesorter\")\n @query= \"INSERT INTO `#{table}` (`#{fields.keys.join('`, `')}`) VALUES ('\"+fields.values.join(\"', '\")+\"');\"\n do_query\n end",
"def insert(data) \n set data, 1\n end",
"def insert(data) \n set data, 1\n end",
"def insert_rows(rows, field, table_struct, dest_table_name = NEW_TABLE_NAME)\n fields = get_fields(table_struct)\n insert_tmplt = row_sql_insert(dest_table_name, table_struct)\n primary_keys = get_pkey_fields(table_struct) \n errs = []\n row_action_data = []\n del_keys = []\n \n if (rows) then\n rows.each_hash do | row |\n row_action_data << {\n :sql_insert => make_sql_insert_row(fields, insert_tmplt, row), \n :key => make_key_hash_for_row(primary_keys, row)\n }\n end\n end\n\n row_action_data.each { |row|\n begin\n dbres = do_sql_command(row[:sql_insert])\n if dbres.nil?\n del_keys << row[:key]\n end\n rescue Mysql::Error\n if !($! =~ /^Duplicate entry .* for key/).nil?\n # i'll consider a duplicate entry okay for a delete\n LOGGER.warn \"Database error! Duplicate key found on insert, marking for deletion anyway, moving on: #{$!}\"\n del_keys << row[:key]\n else\n #errs << \"Database error, moving on: #{$!}\"\n LOGGER.error \"Database error, not sure what, moving on: #{$!}\"\n end\n end\n }\n\n del_keys\nend",
"def insert()\n\n # Kräver att det finns ett \"set_table(\"Table\")\" i klassen\n @insertable_vars_full = self.instance_variables # Ta med namnen user.username osv\n @insertable_vars_full.shift(1) # Kinda frisky\n @insertable_vars = []\n @insertable_values = []\n @insertable_vars_full.each do |var|\n @insertable_vars << var[1..-1]\n @insertable_values << self.instance_variable_get(var)\n end\n\n\n @insertable_vars_str = @insertable_vars.join(\", \")\n\n @question_marks = \"\"\n @insertable_vars.each do |key|\n @question_marks.concat(\"?,\")\n end\n @question_marks = @question_marks[0..-2]\n\n DB.execute(\"INSERT INTO #{@table} (#{@insertable_vars_str})\n VALUES (#{@question_marks})\", @insertable_values)\n\n end",
"def exec_insert(sql, name, binds)\n exec_query(sql, name, binds)\n end",
"def insert_records file_id,records\n sql = \"INSERT INTO #{@table_records} (file_id,\" \n @db.connect do \n records.each do |name,hash|\n fields = hash[:fields]\n values = hash[:values]\n sql_ = sql + fields.keys.join(',') + \") VALUES \"\n sql_ += values.map do |rec| \n index_value fields,rec ## THIS LINE ;)\n row = [file_id] + rec.values_at(*fields.values)\n RubyUtil::sqlize(row)\n end.join(',')\n @db.query(sql_)\n end\n end\n end",
"def to_insert(output, table, row)\n columns = @columns[table].map {|i| i[0] }.join(',')\n values = map_values(row, @columns[table])\n output << \"INSERT INTO #{table} (#{columns}) VALUES (#{values});\\n\"\n end",
"def insert\n array = [[@name, @tagline, @github, @twitter, @blog_url, @image_url, @biography]]\n ins = DB[:conn].prepare(\"INSERT INTO students (name, tagline, github, twitter, blog_url, image_url, biography) VALUES (?, ?, ?, ?, ?, ?, ?);\")\n array.each { |s| ins.execute(s)}\n self.id = DB[:conn].execute(\"SELECT last_insert_rowid() FROM students;\")[0][0]\n #ask steven re. index figures\n #inserting data into an instance\n end",
"def insert(table_name)\n if table_name == \"authors\"\n DATABASE.execute(\"INSERT INTO authors (name) VALUES ('#{@name}')\")\n @id = DATABASE.last_insert_row_id\n \n elsif table_name == \"location_keys\"\n DATABASE.execute(\"INSERT INTO location_keys (location_name, street, city, \n state, country, address, latitude, longitude) \n \n VALUES ('#{@location_name}', '#{@street}', '#{@city}', '#{@state}', \n '#{@country}', '#{@address}', #{@latitude}, #{@longitude})\")\n \n @id = DATABASE.last_insert_row_id\n \n elsif table_name == \"articles\"\n DATABASE.execute(\"INSERT INTO articles (date, author, text, title) \n \n VALUES ('#{@date}', #{@author}, '#{@text}', '#{@title}')\")\n \n @id = DATABASE.last_insert_row_id\n \n elsif table_name == \"articles_with_locations\"\n DATABASE.execute(\"INSERT INTO articles_with_locations (article_id, location_id) \n \n VALUES (#{@article_id}, #{@location_id})\")\n \n end #if loop\n \n end",
"def insert(table,doc)\n connection.insert(path(table),doc)\n end",
"def build_insert(data)\n fields = \"\"\n values = \"\"\n data.each do |k,v|\n fields += \"`#{escape_str_field(k)}`, \"\n values += escape_value(v)+\", \"\n end\n \"(\"+fields.chomp(', ')+\") VALUES (\"+values.chomp(', ')+\")\"\n end",
"def make_insert(table, columns, fields, row)\n statement = \"INSERT INTO #{table['name']} (#{fields.join(',')}) VALUES (\"\n values = []\n fields.each do |field|\n values << make_val(row[field], columns[field])\n end\n statement << \"#{values.join(',')});\\n\"\n statement\n end",
"def db_insert(image_dataset_id)\n ####\"INSERT INTO raw_image_files\n ####(filename, header_reader, file_type, timestamp, source, rmr_number, series_description, \n ####gender, num_slices, slice_thickness, slice_spacing, reconstruction_diameter, \n ####acquisition_matrix_x, acquisition_matrix_y, rep_time, bold_reps, created_at, updated_at, image_dataset_id)\n ####VALUES ('#{@filename}', '#{@hdr_reader}', '#{@file_type}', '#{@timestamp.to_s}', '#{@source}', '#{@rmr_number}', \n ####'#{@series_description}', '#{@gender}', #{@num_slices}, #{@slice_thickness}, #{@slice_spacing}, \n #### #{@reconstruction_diameter}, #{@acquisition_matrix_x}, #{@acquisition_matrix_y}, #{@rep_time}, \n #### #{@bold_reps}, '#{DateTime.now}', '#{DateTime.now}', #{image_dataset_id})\" \n puts \"Old no raw_image_files table\"\n end",
"def insert(template, *data) # :nodoc:\n chk_conn\n conn = @hibernate_session.connection\n stmt = conn.prepare_statement(template)\n data.each do |d|\n d.each_with_index do |item, index|\n if item.kind_of?(Array)\n set_prepared_statement(stmt, item[0], index+1, item[1])\n else\n set_prepared_statement(stmt, item, index+1, nil)\n end\n end\n stmt.execute_update\n end\n conn.commit\n ensure\n stmt.close rescue nil\n end",
"def insert_data(test)\n is_singular = test['insert'].length == 1\n entity_or_entities = nil\n if is_singular\n entity_or_entities = 'entity'\n else\n entity_or_entities = 'entities'\n end\n puts \" Inserting #{test['insert'].length} #{entity_or_entities}\"\n\n insert_data = translate_column_names(test['insert'])\n if @verbose\n puts insert_data\n end\n\n @client.insert(insert_data)\n\n # Wait a few seconds so the data can be inserted by SlicingDice\n sleep @sleep_time\n end",
"def ins table, col, val\n pst = @con.prepare 'INSERT INTO ' + s(table) + '(' + s(col) + ') VALUES(?)'\n pst.execute s(val)\n #puts 'INSERT INTO ' + s(table) + '(' + s(col) + ') VALUES(' + s(val) + ')'\n end",
"def _insert_raw(ds)\n if use_prepared_statements_for?(:insert)\n _set_prepared_statement_server(model.send(:prepared_insert, @values.keys)).call(@values)\n else\n super\n end\n end",
"def single_insert(table_name, hash)\n status = true\n begin\n columns = []\n values = []\n hash.keys.each do |item|\n columns.push(item)\n values.push(\"'#{hash[item]}'\")\n end\n columns = columns.join(\",\")\n values = values.join(\",\")\n @mysql_client.query(\"INSERT INTO #{table_name} (#{columns}) VALUES (#{values})\")\n rescue\n status = false\n end\n return status\n end",
"def insert_table dataset_id, new_table_gapi\n execute { service.insert_table @project, dataset_id, new_table_gapi }\n end",
"def reaktor_insert(row)\n insert_id = 0\n unless row.idstore.nil?\n Log.write_log($import_log, \"reaktor_insert: Table: #{row.table_name} args: #{row.idstore.args * ', '}\")\n else\n Log.write_log($import_log, \"reaktor_insert: Table: #{row.table_name} No IdStore object\")\n end\n query = \"INSERT INTO #{row.table_name} (#{row.get_column_name_string})\\n VALUES (#{(['?']*row.size).join(', ')})\"\n sth = $dbh_ms.prepare(query)\n begin\n sth.execute(*row.get_column_values)\n rescue\n $stderr.puts \"### Error in #{__FILE__} on line #{__LINE__}. See errorlog\"\n Log.write_log('error', \"Could not insert data. Message: #{$!}. query: \\\"#{get_query_string(sth)}\\\"\")\n raise\n exit\n end\n begin\n insert_id = $dbh_ms.func(:insert_id) unless row.idstore.nil?\n rescue\n $stderr.puts \"### Error in #{__FILE__} on line #{__LINE__}. See errorlog\"\n Log.write_log('error', \"Could not get insert id. Message: #{$!}.\")\n raise\n exit\n end\n if insert_id > 0\n row.store_id(insert_id)\n Log.write_log($import_log, \"Insert id store to table: #{row.table_name} id_store parameters: (#{row.idstore.args * ', '}) id: #{insert_id}\")\n else\n unless row.idstore.nil?\n Log.write_log($import_log, \"No id stored for table: #{row.table_name} id_store parameters: (#{row.idstore.args * ', '})\")\n else\n Log.write_log($import_log, \"No id stored for table: #{row.table_name} No IdStore object\")\n end\n \n end\nend",
"def create_insert(headers, values, model, upsert_fields, ret_vals, ig_cols)\r\n\t\t\tputs \"Creating insert query:\"\r\n\t\t\tputs \"There are #{values.length} rows to insert.\"\r\n\r\n\t\t\tp \"HEADERS:\"\r\n\t\t\tp headers\r\n\r\n\t\t\tp \"IGNORED COLUMNS BEFORE HACK:\"\r\n\t\t\tp ig_cols\r\n\r\n\t\t\tig_cols = [] if ig_cols == nil\t# Weird hack because of an error ruby was throwing\r\n\t\t\treturn_results = []\r\n\r\n\t\t\tp \"IGNORED COLUMNS:\"\r\n\t\t\tp ig_cols\r\n\r\n\t\t\t# Loop through the array of arrays of values to insert\r\n\t\t\tvalues.each do |values_array|\r\n\t\t\t\tupsert_attributes = {}\r\n\t\t\t\tinner_array = []\r\n\t\t\t\t# Now loop through the single array of values\r\n\t\t\t\tp \"VALUES ARRAY:\"\r\n\t\t\t\tp values_array\r\n\r\n\t\t\t\tvalues_array.each_with_index do |val, index|\r\n\t\t\t\t\t# puts \"INDEX: #{index}\"\r\n\r\n\t\t\t\t\tnext if ig_cols.include?(index) # IMPORTANT: Need to ignore the indices of the columns in the CSV that the user specifies\r\n\t\t\t\t\tassociated_column_name = headers.at(index).to_sym\t# Get the header name for the row - need it to match in return values\r\n\r\n\t\t\t\t\t# Store the attributes we want to do the upsert on to pass into find_or_create_by method\r\n\t\t\t\t\tupsert_attributes[associated_column_name] = val if upsert_fields.include?(associated_column_name)\r\n\r\n\t\t\t\t\t# puts \"Line 282: #{upsert_attributes}\"\r\n\t\t\t\tend\r\n\r\n\t\t\t\t# Use ActiveRecord's method to return the updated or inserted row\r\n\t\t\t\t# Workaround - do a select and then insert since I can't figure out how to dynamically add the values to the class\r\n\t\t\t\t# select_result = model.find_by(upsert_attributes)\r\n\t\t\t\tinsert_attributes = {}\r\n\t\t\t\tvalues_array.each_with_index do |val, i|\r\n\t\t\t\t\tif !(upsert_attributes.has_key?(headers[i]))\r\n\t\t\t\t\t\t# puts \"VALUE: #{val}\"\r\n\t\t\t\t\t\tinsert_attributes[headers[i].to_sym] = val\r\n\t\t\t\t\tend\r\n\t\t\t\tend\r\n\t\t\t\tinsert_attributes = insert_attributes.merge upsert_attributes\r\n\r\n\t\t\t\t# if select_result == nil\r\n\t\t\t\t# \tinsert_result = model.create(insert_attributes)\r\n\t\t\t\t# else\r\n\t\t\t\t# \tinsert_result = model.update(insert_attributes)\r\n\t\t\t\t# end\r\n\r\n\t\t\t\t# upsert_result = model.find_or_create_by(upsert_attributes) do |klass|\r\n\t\t\t\t# \t# Check to see that we haven't already included the column and value in the upsert_attributes\r\n\t\t\t\t# \t# and if we haven't, include it as a field we need to add to the database along with the value\r\n\t\t\t\t# \t# puts \"#{klass.instance_variables}\"\r\n\t\t\t\t# \tvalues_array.each_with_index do |val, i|\r\n\t\t\t\t# \t\tif !(upsert_attributes.has_key?(headers[i]))\r\n\t\t\t\t# \t\t\t# puts \"VALUE: #{val}\"\r\n\t\t\t\t# \t\t\tklass.send :write_attribute, headers[i].to_sym, val\r\n\t\t\t\t# \t\tend\r\n\t\t\t\t# \tend\r\n\t\t\t\t# end\r\n\r\n\t\t\t\tp upsert_attributes\r\n\r\n\t\t\t\tupsert_result = model.find_or_initialize_by(upsert_attributes)\r\n\t\t\t\tupsert_result.update_attributes(insert_attributes)\r\n\r\n\t\t\t\t# Return what the user asked for\r\n\t\t\t\t#ret_vals.each { |val| inner_array.push(insert_result[val]) }\r\n\t\t\t\tret_vals.each { |val| inner_array.push(upsert_result[val]) }\r\n\r\n\t\t\t\t# Concatenate the arrays of information the user wants back\r\n\t\t\t\treturn_results.push(inner_array)\r\n\r\n\t\t\tend\r\n\t\t\t# p return_results\r\n\t\t\treturn return_results\r\n\t\tend",
"def insert values\n im = InsertManager.new @engine\n im.insert values\n @engine.connection.insert im.to_sql\n end",
"def insert_\n # new files to insert\n nfiles = @curr_schema.select_new_files\n count = nfiles.size\n Logger.<<(__FILE__,\"INFO\",\"Found #{nfiles.size} files to decode & insert for #{@curr_source.name}...\");\n return unless nfiles.size > 0\n\n SignalHandler.check\n ## Will decode them and insert their records, one by one (file)\n base_path = ::File.join(Conf::directories.store,@curr_source.name.to_s)\n ids_processed = []\n file_counter = 0\n iterate_over nfiles do |file|\n file_path = File.join(base_path,file[:folder],file[:file_name])\n begin\n file_ = CDR::File.new(file_path,search: true)\n rescue => e\n Logger.<<(__FILE__,\"WARNING\",\"File Error : #{e}\")\n raise e\n end\n records = @curr_source.decoder.decode file_\n if records.nil? \n Logger.<<(__FILE__,\"WARNING\",\"Found null output for file #{file}\")\n else\n @curr_schema.insert_records file[:file_id], records\n end\n @curr_schema.processed_files RubyUtil::arrayize(file[:file_id])\n backup_file file[:folder],file_\n str = log_file_summary file_,records\n Logger.<<(__FILE__,\"INFO\",\"(#{file_counter}/#{count}) #{str}\",inline: true)\n file_counter += 1\n end\n # so only one lookup for table cdr\n #mark_processed_decoded_files (ids_processed)\n Logger.<<(__FILE__,\"INFO\",\"Decoded & Inserted #{count} files ...\")\n Logger.<<(__FILE__,\"INFO\",\"Insert operation finished !\")\n end",
"def save\r\n # raise error if schema is not same between from and to.\r\n # so , create sql from record map\r\n #keys = get_schema.map{|v| v[0]}\r\n keys = @__stored_map__.keys\r\n sql = \"insert into #{self.class.table_name} (#{keys.join(',')}) values( \"\r\n\r\n keys.each_with_index do |key , index|\r\n v = @__stored_map__[key]\r\n sql << ',' if index != 0\r\n sql << (v.kind_of?(String) ? \"'#{v}'\" : v.to_s)\r\n end\r\n sql << \")\"\r\n puts sql if $DEBUG\r\n self.class.execute_sql(sql)\r\n end",
"def insert()\n\t\t@db.execute(\"insert into ausgaben (jahr, monat, name, betrag, gemeinsam, tags) values(:jahr, :monat, :name, :betrag, :gemeinsam, :tags)\", @options)\n\tend",
"def exec_insert(sql, name, binds)\n exec_query(sql, name, binds)\n end",
"def flush\n conn.transaction do\n buffer.flatten.each do |row|\n # check to see if this row's compound key constraint already exists\n # note that the compound key constraint may not utilize virtual fields\n next unless row_allowed?(row)\n\n # add any virtual fields\n add_virtuals!(row)\n \n names = []\n values = []\n order.each do |name|\n names << \"`#{name}`\"\n values << conn.quote(row[name]) # TODO: this is probably not database agnostic\n end\n q = \"INSERT INTO `#{table_name}` (#{names.join(',')}) VALUES (#{values.join(',')})\"\n ETL::Engine.logger.debug(\"Executing insert: #{q}\")\n conn.insert(q, \"Insert row #{current_row}\")\n @current_row += 1\n end\n buffer.clear\n end\n end",
"def insert_into_table(db, product, exp_date, prod_type)\n\tdb.execute(\"INSERT INTO products (name, expiration_date, type_id) VALUES (?, ?, ?)\",[product, exp_date, prod_type])\nend",
"def exec_insert(sql, name = nil, binds = []) #:nodoc:\r\n log(sql, \"insert\", binds) { query(sql, binds) }\r\n end",
"def insert_takoble\n\t\tsql = \"insert into TAKO_List values (?, ?)\"\n\t\treturn sql\n\tend",
"def dump_insert_multi(io, table_obj, rows)\n debug \"Inserting #{rows.length} into #{table_obj.name}.\"\n sqls = @export_db.insert_multi(\n table_obj.name,\n rows,\n replace_line_breaks: true,\n return_sql: true,\n keys: @keys\n )\n sqls.each do |sql|\n io.write(\"#{sql};\\n\")\n end\n\n rows.clear\n\n # Ensure garbage collection or we might start using A LOT of memory.\n GC.start\n end",
"def dump_insert_multi(io, table_obj, rows)\n print \"Inserting #{rows.length} into #{table_obj.name}.\\n\" if @debug\n sqls = @args[:db].insert_multi(table_obj.name, rows, :return_sql => true, :keys => @keys)\n sqls.each do |sql|\n io.write(\"#{sql};\\n\")\n end\n \n rows.clear\n \n #Ensure garbage collection or we might start using A LOT of memory.\n GC.start\n end",
"def create_threaded_insert(headers, values, model, upsert_fields, ret_vals)\r\n\t\t\tnum_rows = values.length\t\t# Cache the number of rows we have to insert\r\n\t\t\tnum_threads = (num_rows.to_f / 7000).ceil\t# Calculate the number of threads using an arbitrary way (for now...)\r\n\t\t\tthreads = []\t\t# Create an array to hold the threads we are going to spawn to handle the file\r\n\r\n\t\t\t# IMPORTANT: We create a new array using the values array and splitting it with chunks of 7000 rows\r\n\t\t\tnew_values = values.each_slice(7000).to_a\r\n\r\n\t\t\t# abort(\"#{new_values[0]}\")\r\n\r\n\t\t\tp \"There are #{num_rows} rows to insert.\"\r\n\t\t\tp \"We are creating #{num_threads} threads to handle the file.\"\r\n\r\n\t\t\t#headers.map! { |head_val| head_val.gsub(\"/\\\"/\", \"'\"); }\r\n\r\n\t\t\t#abort(\"#{headers}\")\r\n\r\n\t\t\t# Use the Parallel gem to handle parallel computation of the insertions\r\n\t\t\tresults = Parallel.map(new_values, :in_threads=>num_threads) do |val|\r\n\t\t\t\tp headers\r\n\t\t\t\tcreate_insert(headers, val, model, upsert_fields, ret_vals)\r\n\t\t\tend\r\n\r\n\t\t\t# We can try and maximize the threads to attain the best performance.\r\n\t\t\t# For now, we will just use an arbitrarily calculated number of threads.\r\n\t\t\t# num_threads.times do |i|\r\n\t\t\t# \tthreads[i] = Thread.new {\r\n\t\t\t# \t\tcreate_insert(headers, new_values[i], model, upsert_fields, ret_vals)\r\n\t\t\t# \t}\r\n\t\t\t# end\r\n\r\n\t\t\t# threads.each { |thread| thread.join }\r\n\t\t\t# ThreadsWait.all_waits(*threads)\r\n\t\t\tputs \"Finished inserting data successfully!\"\r\n\t\tend",
"def insert_data(pstmt, row)\n if @language == 'french'\n pstmt.execute(\n date_default_value(row[\"#{MOASHeader::HEADERS['GO_LIVE_DATE'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['APP_STATUS'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['SKU'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['SHORT_TITLE'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['LONG_TITLE'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['GENDER'].downcase}\"]),\n number_default_value(row[\"#{MOASHeader::HEADERS['AGE_FROM_MONTHS'].downcase}\"]),\n number_default_value(row[\"#{MOASHeader::HEADERS['AGE_TO_MONTHS'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['SKILL'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['CURRICULUM'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['LONG_DESCRIPTION'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['PLATFORM_COMPATIBILITY_FR'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['SPECIAL_MESSAGE'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['TEACHES_FR'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['LICENSE_LEGAL'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['LICENSE_NON_LICENSE'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['LICENSE'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['LANGUAGE'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['PRICE_TIER'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['CATEGORY'].downcase}\"]),\n '',\n '',\n '',\n '',\n '',\n '',\n convert_data(row[\"#{MOASHeader::HEADERS['FRANCE'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['FRENCH_CANADA'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['FRENCH_ROW'].downcase}\"])\n )\n elsif @language == 'english'\n pstmt.execute(\n date_default_value(row[\"#{MOASHeader::HEADERS['GO_LIVE_DATE'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['APP_STATUS'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['SKU'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['SHORT_TITLE'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['LONG_TITLE'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['GENDER'].downcase}\"]),\n number_default_value(row[\"#{MOASHeader::HEADERS['AGE_FROM_MONTHS'].downcase}\"]),\n number_default_value(row[\"#{MOASHeader::HEADERS['AGE_TO_MONTHS'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['SKILL'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['CURRICULUM'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['LONG_DESCRIPTION'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['PLATFORM_COMPATIBILITY'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['SPECIAL_MESSAGE'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['TEACHES'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['LICENSE_LEGAL'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['LICENSE_NON_LICENSE'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['LICENSE'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['LANGUAGE'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['PRICE_TIER'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['CATEGORY'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['US'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['CA'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['UK'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['IE'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['AU'].downcase}\"]),\n convert_data(row[\"#{MOASHeader::HEADERS['ROW'].downcase}\"]),\n '',\n '',\n ''\n )\n end\n end",
"def _merge_insert_sql(sql, data)\n sql << \" THEN INSERT \"\n columns, values = _parse_insert_sql_args(data[:values])\n _insert_columns_sql(sql, columns)\n if override = data[:override]\n sql << override\n end\n _insert_values_sql(sql, values)\n end",
"def insert_in_database\n Fetch.new(insertion_query).array\n end",
"def insertar (id_usu_den,id_tip_doc,num_doc_den,fec_exp_doc_den,fec_nac_den,pri_nom_den,seg_nom_den,pri_ape_den,seg_ape_den,tel_den,dir_den,push_den,ema_den)\n \tputs \"insertar: \"\n \tprint \"INSERT INTO public.denunciante(\n id_usu_den, id_tip_doc, num_doc_den, fec_exp_doc_den, fec_nac_den, \n pri_nom_den, seg_nom_den, pri_ape_den, seg_ape_den, tel_den, \n dir_den, push_den, ema_den)\n VALUES (?, ?, ?, ?, ?, \n ?, ?, ?, ?, ?, \n ?, ?, ?);\"\n\n end",
"def fast_insert(rows, base_cmd, end_cmd = '')\n RawDB.fast_insert(db, rows, base_cmd, end_cmd)\n end",
"def sqlite3_insert(table_name, values, mem_db)\n @assert.check_table_name(table_name, @dbh)\n\n table_ast = Hash.new\n table_ast = @dbm.get_table_ast(table_name)\n # TODO: Each value element type. All types are acceptable.\n # If hash, {key:value} needs to be {column_name:value}.\n\n # This hash holds a map of column position in the table\n index_to_column = Hash.new\n # Since the position of column in a table is static on\n # creation, we can keep a track of each column position\n # simple by assigning an index to each (A.K.A. indexing).\n tb_index = 0\n table_ast.each {|col_name, col_info|\n index_to_column[tb_index] = [col_name, col_info]\n tb_index += 1\n }\n\n @assert.check_class(values.class, Array, @dbh)\n vl_index = 0\n column_to_value = Hash.new\n values.each {|value|\n if value.class == Hash\n status = value.size == 1\n error_msg = \"Error: Size of hash for insert value cannot exeed 1.\\n\"\n error_msg += \"#{value} has size #{value.size}\"\n @assert.default_error_check(status, error_msg, @dbh)\n \n # Extract user input value\n col_name = value.keys[0]\n status = table_ast.has_key?(col_name.to_s)\n error_msg = \"Column #{col_name.to_s} does not exist in table #{table_name}.\\n\"\n error_msg += @dbm.get_table_schema(table_name)\n @assert.default_error_check(status, error_msg, @dbh)\n column_to_value[col_name.to_s] = value[col_name]\n else\n # Compare input value type with table column type\n cur_column_in_table = index_to_column[vl_index]\n @assert.check_column_value(cur_column_in_table, value, mem_db[table_name], @dbh)\n col_name = cur_column_in_table[0]\n column_to_value[col_name] = value\n end\n\n vl_index += 1\n }\n\n insert_query = \"INSERT INTO #{table_name} (\"\n column_to_value.each_key {|col_name|\n insert_query += \"#{col_name},\"\n }\n insert_query.chomp!(',')\n insert_query += \") VALUES (\"\n column_to_value.each_value {|value|\n if value.class == String\n insert_query += \"'#{value}',\"\n else\n insert_query += \"#{value},\"\n end\n }\n insert_query.chomp!(',')\n insert_query += ');'\n\n @dbh.execute(insert_query)\n return insert_query + \"\\n\"\n end",
"def exec_insert(sql, name = nil, binds = [], pk = nil, sequence_name = nil)\n exec_query(sql, name, binds)\n end",
"def insert\n \n attributes = []\n instance_variables.each do |i|\n attributes << i.to_s.delete(\"@\") if (i != :@id && i != :@table)\n end\n \n values = []\n attributes.each do |a|\n value = self.send(a)\n \n if value.is_a?(Integer)\n values << \"#{value}\"\n else values << \"'#{value}'\"\n end\n end\n \n DATABASE.execute(\"INSERT INTO students (#{attributes.join(\", \")}) \n VALUES (#{values.join(\", \")})\")\n @id = DATABASE.last_insert_row_id\n\n end",
"def insert_sql_each\n return enum_for(__method__) unless block_given?\n each_row do |row|\n yield table_dataset.insert_sql( row )\n end\n end",
"def process\n create, insert, table = extract_sql #Extract mysql create/insert statements from the dump file\n raise \"Couldn't extract create syntax from MySql Dump File\" if create.nil?\n create = escape_create_string(create)\n begin\n @connection.execute(\"DROP TABLE #{table}\") rescue ''#Drop existing table first\n @connection.execute(create) #Recreate the table \n if insert && @import_data\n values = row_values(insert) \n values.each do |val|\n sql = \"INSERT INTO #{table} VALUES #{val}\"\n begin\n @connection.execute(sql) #Insert rows\n rescue Exception => e\n puts e.message\n puts sql\n puts \"table #{table}\"\n end\n end\n else\n puts \"There's no records to be added\" if @import_data && !insert\n end\n rescue Exception => e\n puts e.message\n puts \"table #{table}\"\n end\n end",
"def handle_insert action, result; end",
"def row_sql_insert(table_name, table_struct)\n fields = get_fields(table_struct)\n\n sql = <<-EOF\n INSERT INTO `#{DBNAME}`.`#{table_name}` (\n #{fields.collect { |f| \"`#{f}`\" }.join(\", \")}\n )\n VALUES (\n #{fields.collect { |f| \"'%s'\" }.join(\", \")}\n );\n EOF\n\n sql\nend",
"def staff_insert(h)\n \t\trows = @database.execute <<-SQL\n\t\t\tINSERT INTO staff\n\t\t\t\t( person\n\t\t\t\t, first_name\n\t\t\t\t, last_name\n\t\t\t\t, staff_id\n\t\t\t\t, rate\n\t\t\t\t)\n\t\t\tVALUES\n\t\t\t\t('#{h['person']}'\n\t\t\t\t, '#{h['first_name']}'\n\t\t\t\t, '#{h['last_name']}'\n\t\t\t\t, '#{h['staff_id']}'\n\t\t\t\t, '#{h['rate']}'\n\t\t\t\t)\n\t\tSQL\n\tend",
"def sql_insert(record)\n flds, vals = parse_fldsvalues(record)\n ph = vals.map{|x| placeholder }\n\n sql = %Q|insert into #{quoted_table}\n ( #{flds.join ','} )\n output inserted.#{quote_field id_fld}\n values( #{ph.join ','} );|\n\n [sql, vals]\n end",
"def insert\n col_names = self.class.columns.join(\", \")\n question_marks = ([\"?\"] * self.class.columns.length).join(\", \")\n DBConnection.execute(<<-SQL, *attribute_values)\n INSERT INTO\n #{self.class.table_name} (#{col_names})\n VALUES\n (#{question_marks})\n SQL\n\n self.id = DBConnection.last_insert_row_id\n end",
"def insert(*args)\n dataset.insert(*args)\n self\n end",
"def copy_rows( field, \n table_struct, \n src_table_name = TABLE_NAME, \n dest_table_name = NEW_TABLE_NAME, \n num_rows = ROWS_PER_TRANSACTION)\n rows = grab_rows(field, src_table_name, num_rows)\n keys_for_delete = insert_rows(rows, field, table_struct, dest_table_name)\n keys_for_delete\nend",
"def insert_sequenced(row)\n sql = row.type.insert_sql_minus_key\n vals = row.field_values_minus_key\n#$stderr.puts sql\n#$stderr.puts vals.inspect\n\n db.do(sql, *vals)\n insert_id = db.select_one(row.type.get_insert_id_sql)[0]\n row.send(row.type.primary_key.setter_name, insert_id)\n row.reset_changed\n end",
"def _insert\n ds = _insert_dataset\n if _use_insert_select?(ds) && (h = _insert_select_raw(ds))\n _save_set_values(h)\n nil\n else\n iid = _insert_raw(ds)\n # if we have a regular primary key and it's not set in @values,\n # we assume it's the last inserted id\n if (pk = autoincrementing_primary_key) && pk.is_a?(Symbol) && !(vals = @values)[pk]\n vals[pk] = iid\n end\n pk\n end\n end",
"def _insert\n ds = _insert_dataset\n if _use_insert_select?(ds) && (h = _insert_select_raw(ds))\n _save_set_values(h)\n nil\n else\n iid = _insert_raw(ds)\n # if we have a regular primary key and it's not set in @values,\n # we assume it's the last inserted id\n if (pk = autoincrementing_primary_key) && pk.is_a?(Symbol) && !(vals = @values)[pk]\n vals[pk] = iid\n end\n pk\n end\n end",
"def flush\n conn.transaction do\n buffer.flatten.each do |row|\n # check to see if this row's compound key constraint already exists\n # note that the compound key constraint may not utilize virtual fields\n next unless row_allowed?(row)\n\n # add any virtual fields\n add_virtuals!(row)\n \n key_names = []\n key_values = []\n @key_columns.each do |name|\n key_names << \"#{name}\"\n key_values << conn.quote(row[name]) # TODO: this is probably not database agnostic\n end\n\n names = []\n values = []\n (order - @key_columns).each do |name|\n names << \"#{name}\"\n values << conn.quote(row[name]) # TODO: this is probably not database agnostic\n end\n\n all_name_values = (key_names+names).zip(key_values+values)\n\n q = <<EOF\nMERGE INTO #{table_name} d \nUSING (SELECT #{all_name_values.collect {|c,v| \"#{v} #{c}\"}.join(',')} FROM DUAL) s\nON (#{map_src_to_dest(key_names,'s','d').join(' AND ')})\nWHEN MATCHED THEN \nUPDATE SET #{[map_src_to_dest(names,'s','d'), \"d.#{@update_ts_column}=CURRENT_TIMESTAMP\"].flatten.join(',')}\nWHEN NOT MATCHED THEN\nINSERT (#{all_name_values.collect {|c,v| 'd.'+c}.join(',')},d.#{@insert_ts_column})\nVALUES (#{all_name_values.collect {|c,v| 's.'+c}.join(',')},CURRENT_TIMESTAMP)\nEOF\n #q = \"INSERT INTO `#{table_name}` (#{names.join(',')}) VALUES (#{values.join(',')})\"\n ETL::Engine.logger.debug(\"Executing upsert: #{q}\")\n conn.insert(q, \"Upsert row #{current_row}\")\n @current_row += 1\n end\n buffer.clear\n end\n end",
"def add_to (table_name,title,code)\r\n\t\"INSERT INTO \" + table_name + \r\n\t\"(title,code) VALUES \r\n\t('\" + title + \"','\" + code + \"')\"\r\nend",
"def execute_insert(sql, opts={})\n synchronize(opts[:server]) do |conn|\n conn.execute(sql)\n insert_result(conn, opts[:table], opts[:values])\n end\n end",
"def _bulkinsert (data,first)\n @length = 0\n cur = first\n while cur\n k = data[cur]\n # FIXME: Check for `Deleted`?\n if k\n if not_deleted?(k)\n # FIXME: Doing k == Deleted or k != Deleted here fails.\n # FIXME: Combining these on one line triggers bug.\n v = data[cur + 1]\n self[k] = v\n end\n end\n cur = data[cur + 2]\n end\n nil\n end",
"def insert(values)\n primary_key_value = nil\n\n if primary_key && Hash === values\n primary_key_value = values[values.keys.find { |k|\n k.name == primary_key\n }]\n\n if !primary_key_value && connection.prefetch_primary_key?(klass.table_name)\n primary_key_value = connection.next_sequence_value(klass.sequence_name)\n values[klass.arel_table[klass.primary_key]] = primary_key_value\n end\n end\n\n im = arel.create_insert\n\n # ****** BEGIN PARTITIONED PATCH ******\n actual_arel_table = @klass.dynamic_arel_table(Hash[*values.map{|k,v| [k.name,v]}.flatten]) if @klass.respond_to?(:dynamic_arel_table)\n actual_arel_table = @table unless actual_arel_table\n # Original line:\n # im.into @table\n im.into actual_arel_table\n # ****** END PARTITIONED PATCH ******\n\n conn = @klass.connection\n\n substitutes = values.sort_by { |arel_attr,_| arel_attr.name }\n binds = substitutes.map do |arel_attr, value|\n [@klass.columns_hash[arel_attr.name], value]\n end\n\n substitutes.each_with_index do |tuple, i|\n tuple[1] = conn.substitute_at(binds[i][0], i)\n end\n\n if values.empty? # empty insert\n im.values = Arel.sql(connection.empty_insert_statement_value)\n else\n im.insert substitutes\n end\n\n conn.insert(\n im,\n 'SQL',\n primary_key,\n primary_key_value,\n nil,\n binds)\n end",
"def generate_insert\n @binds = Array.new\n @insert_statement = \"insert into #{fully_qualified_table_name} (\"\n @insert_statement << column_details.keys.sort.map { |k| column_detail(k).column_name }.join(',')\n @insert_statement << ') values ('\n @insert_statement << column_details.keys.sort.map { |k|\n \":#{k}\"\n }.join(',')\n column_details.keys.sort.each { |k|\n if @column_values[k] == nil\n @binds.push [column_type_to_ruby_type(column_details[k]), nil]\n else\n @binds.push @column_values[k]\n end\n }\n @insert_statement << ')'\n @insert_statement\n end",
"def insert\n # the array of ::columns of the class joined with commas, drop id\n col_names = self.class.columns[1..-1].join(\", \") \n # an array of question marks\n question_marks = ([\"?\"] * col_names.split.size).join(\", \")\n\n DBConnection.execute(<<-SQL, *attribute_values[1..-1])\n INSERT INTO\n #{self.class.table_name} (#{col_names})\n VALUES\n (#{question_marks})\n SQL\n\n self.id = DBConnection.last_insert_row_id\n end",
"def insert!(*rows)\n Mao.sql(with_options(:insert => rows.flatten).sql) do |pg_result|\n if @options[:returning]\n pg_result.map {|result| Mao.normalize_result(result, @col_types)}\n else\n pg_result.cmd_tuples\n end\n end\n end",
"def batch_insert_time_serials_service_development_cost(data)\n connection = ActiveRecord::Base.connection\n ActiveRecord::Base.transaction {\n sql = \"INSERT INTO ts_service_development_cost_usage_patterns (user_id, service_development_cost_id, \\\"current_time\\\", year, month, date, quantity)\n VALUES #{data.join(\", \")}\"\n connection.execute sql\n }\n end",
"def time_entry_insert(h)\n \t\trows = @database.execute <<-SQL\n\t\t\tINSERT INTO time_entries\n\t\t\t\t(time_entry_id\n\t\t\t\t, staff_id\n\t\t\t\t, project_id\n\t\t\t\t, task_id\n\t\t\t\t, hours\n\t\t\t\t, date\n\t\t\t\t, notes\n\t\t\t\t, billed\n\t\t\t\t)\n\t\t\tVALUES\n\t\t\t\t('#{h['time_entry_id']}'\n\t\t\t\t, '#{h['staff_id']}'\n\t\t\t\t, '#{h['project_id']}'\n\t\t\t\t, '#{h['task_id']}'\n\t\t\t\t, '#{h['hours']}'\n\t\t\t\t, '#{h['date']}'\n\t\t\t\t, '#{h['notes']}'\n\t\t\t\t, '#{h['billed']}'\n\t\t\t\t)\n\t\tSQL\n\tend",
"def batch_insert(data)\n hash = Hash[data.collect { |v| [v, '1'] }]\n @redis.write_multi(hash)\n end",
"def generate_insert tab_name, argv\n\n\t\t\t\t# Attribute of table.\n\t\tattribute = \"\"\n\t\tattribute << \"id\"\n\t\ti = 0\n\t\targv.each do |k,v|\n\t\t\tif i == 0 and argv.length != 0\n\t\t\t \tattribute << \",\"\n\t\t\t end\n\t\t\tattribute << k\n\t\t\ti = i.to_i + 1\n\t\t\tif i != argv.length\n\t\t\t\tattribute << \",\"\n\t\t\tend\n\t\tend\n\t\t\t\t# Values of the Attributes.\n\t\tval = []\n\t\tval << \"NULL\"\n\t\ti = 0\n\t\tif argv.length != 0\n\t\t\tbegin\n\t\t\t\tval << \"\\\"#{argv[argv.keys[i]]}\\\"\"\n\t\t\t\ti = i.to_i + 1 \t\t\n\t\t\tend while i < argv.length.to_i\t\n\t\tend\t\t\n\n\t\tif argv.length == 0\n\t\t\tquery = \" INSERT INTO #{tab_name}(#{attribute}) VALUES (#{val * \"\"}) \"\n\t\t\treturn query\n\t\telse\n\t\t\tquery = \" INSERT INTO #{tab_name}(#{attribute}) VALUES (#{val * \",\"}) \"\n\t\t\treturn query\n\t\tend\n\tend",
"def insert(value)\n #YOUR WORK HERE\n end",
"def to_inserts(args={})\n args[:table] ||= Pathname.new(@filename).basename.to_s.downcase.gsub(/\\W/, '_')\n args[:before] ||= @@defaults[:before]\n args[:after] ||= @@defaults[:after]\n insert_sql = args[:ignore] ? 'insert ignore' : 'insert'\n if args[:bulk]\n args[:before] += \"#{insert_sql} into #{args[:table]} values\"\n args[:values_glue] ||= \", \"\n args[:row_format] ||= \" (%s)\"\n args[:row_glue] ||= \",\\n\"\n else\n args[:before] ||= \"\"\n args[:values_glue] ||= \", \"\n args[:row_format] ||= \"#{insert_sql} into #{args[:table]} values(%s)\"\n args[:row_glue] ||= \";\\n\"\n end\n to_any args\n end",
"def insert(value)\n # YOUR WORK HERE\n end",
"def generate_pg_insert_query(table_name, keys, rows)\n \"INSERT INTO #{table_name}(#{keys.map { |i| \"\\\"#{i}\\\"\" }.join(',')}) VALUES(#{keys.map { |i| rows[i] == nil ? 'NULL' : \"'\" + pg_conn.escape_string(rows[i]) + \"'\" }.join(',')});\\n\"\n end",
"def invoice_insert(h)\n\t\trows = @database.execute <<-SQL\n\t\tINSERT INTO invoices\n\t\t\t( number\n\t\t\t, organization\n\t\t\t, updated\n\t\t , amount\n\t\t\t, amount_outstanding\n\t\t\t, discount\n\t\t \t, invoice_id\n\t\t\t, matter\n\t\t\t, date\n\t\t\t, status)\n\t\tVALUES\n\t\t('#{h['number']}'\n\t\t, '#{h['organization']}'\n\t\t, '#{h['updated']}'\n\t\t, '#{h['amount']}'\n\t\t, '#{h['amount_outstanding']}'\n\t\t, '#{h['discount']}'\n\t\t, '#{h['invoice_id']}'\n\t\t, '#{h['matter']}'\n\t\t, '#{h['date']}'\n\t\t, '#{h['status']}'\n\t \t )\n\t\t SQL\n\tend",
"def save_data(uniq_keys, values_array, tbl_name)\n values_array = [values_array].flatten(1) # coerce to an array\n all_field_names = values_array.map(&:keys).flatten.uniq\n all_field_names_as_string = format_field_names_as_string(all_field_names)\n all_field_names_as_symbol_string = all_field_names.map{ |k| \":#{k}\" }.join(',') # need to appear as symbols\n begin\n values_array.each do |values_hash|\n # mustn't use nil value in unique value due to fact that SQLite considers NULL values to be different from\n # each other in UNIQUE indexes. See http://www.sqlite.org/lang_createindex.html\n raise DatabaseError.new(\"Data has nil value for unique key. Unique keys are #{uniq_keys}. Offending data: #{values_hash.inspect}\") unless uniq_keys.all?{ |k| values_hash[k] }\n sql_query = \"INSERT OR REPLACE INTO #{tbl_name} (#{all_field_names_as_string}) VALUES (#{all_field_names_as_symbol_string})\"\n database.execute(sql_query, values_hash)\n end\n rescue SQLite3::SQLException => e\n puts \"Exception (#{e.inspect}) raised\" if verbose?\n case e.message\n when /no such table/\n create_table(tbl_name, all_field_names, uniq_keys)\n retry\n when /has no column/\n add_columns(tbl_name, all_field_names)\n retry\n else\n raise e\n end\n end\n end",
"def insert(db, volume_id, page_number, access_date, ip_token)\n command = \"insert into results values (?, ?, datetime(?), ?);\"\n db.execute(command, volume_id, page_number, access_date, ip_token)\n end",
"def insert_sql(c, insert)\n\n\t\ttime = Time.now.to_s(:db)\n\n\t\tfirstname = remove_apostrophe(c.first_name)\n\t\tlastname = remove_apostrophe(c.last_name)\n\t\tcompany = remove_apostrophe(c.company)\n\t\tnotes = remove_apostrophe(c.notes)\n\t\temail = remove_apostrophe(c.email)\n\t\tdate_created = map_date(c.date_created)\n\t\tdate_modified = map_date(c.date_modified)\n\t\tphone = c.phone.gsub(/\\s+/, \"\")\n\n\t\tsql = cust = \"\"\n\n\t\tif insert == 1\n\n\t\t\tunallocated_staff_id = 34\n\n\t\t\tcust = \"('#{c.id}', '#{firstname}', '#{lastname}', '#{company}',\\\n\t\t\t'#{email}', '#{phone}', '#{c.store_credit}', '#{c.registration_ip_address}',\\\n\t\t\t'#{notes}', '#{date_created}', '#{date_modified}', '#{time}', '#{time}', '#{c.customer_group_id}', '#{unallocated_staff_id}')\"\n\n\t\t\tsql = \"INSERT INTO customers(id, firstname, lastname, company, email, phone,\\\n\t\t\tstore_credit, registration_ip_address, notes, date_created, date_modified,\\\n\t\t\tcreated_at, updated_at, cust_type_id, staff_id) VALUES #{cust}\"\n\t\telse\n\n\t\t\tsql = \"UPDATE customers SET firstname = '#{firstname}', lastname = '#{lastname}', company = '#{company}',\\\n\t\t\temail = '#{email}', phone = '#{phone}', store_credit = '#{c.store_credit}',\\\n\t\t\tregistration_ip_address = '#{c.registration_ip_address}', notes = '#{notes}', date_created = '#{date_created}',\\\n\t\t\tdate_modified = '#{date_modified}', updated_at = '#{time}', cust_type_id = '#{c.customer_group_id}' WHERE id = '#{c.id}'\"\n\n\n\t\tend\n\n ActiveRecord::Base.connection.execute(sql) \n\n\tend",
"def insert_into_sql(sql)\n sql << \" INTO \"\n if (f = @opts[:from]) && f.length == 1\n identifier_append(sql, server_version >= 90500 ? f.first : unaliased_identifier(f.first))\n else\n source_list_append(sql, f)\n end\n end",
"def insert_record(name, type, ttl, content)\n records_table.insert(\n :domain_id => domain.id,\n :name => name,\n :type => type, \n :ttl => ttl, \n :content => content,\n :change_date => Time.now.to_i\n )\n end",
"def post(data)\n jss = self.get_structure() # For referencing purposes\n\n input = self.query_to_hash(data)\n bad_request = false\n resp = nil\n jss.each do |key, value|\n # Check if we have it on the query too\n unless input.has_key? key\n resp = MIDB::Interface::Server.json_error(400, \"Bad Request - Not enough data for a new resource\")\n @engine.http_status = 400\n bad_request = true\n end\n end\n input.each do |key, value|\n # Check if we have it on the structure too\n unless jss.has_key? key\n resp = MIDB::Interface::Server.json_error(400, \"Bad Request - Wrong argument #{key}\")\n @engine.http_status = 400\n bad_request = true\n end\n end\n \n\n # Insert the values if we have a good request\n unless bad_request\n fields = Hash.new\n inserts = Hash.new\n main_table = self.get_structure.values[0].split('/')[0]\n input.each do |key, value|\n struct = jss[key]\n table = struct.split(\"/\")[0]\n inserts[table] ||= []\n fields[table] ||= []\n inserts[table].push \"\\\"\" + value + \"\\\"\"\n fields[table].push struct.split(\"/\")[1]\n if struct.split(\"/\").length > 2\n match = struct.split(\"/\")[2]\n matching_field = match.split(\"->\")[0]\n row_field = match.split(\"->\")[1]\n fields[table].push matching_field\n if @engine.config[\"dbengine\"] == :mysql\n inserts[table].push \"(SELECT #{row_field} FROM #{main_table} WHERE id=(SELECT LAST_INSERT_ID()))\"\n else\n inserts[table].push \"(SELECT #{row_field} FROM #{main_table} WHERE id=(last_insert_rowid()))\"\n end\n end\n end\n queries = [] \n inserts.each do |table, values|\n queries.push \"INSERT INTO #{table}(#{fields[table].join(',')}) VALUES (#{inserts[table].join(',')});\"\n end\n # Connect to the database\n dbe = MIDB::API::Dbengine.new(@engine.config, @db)\n dblink = dbe.connect()\n results = []\n rid = nil\n # Find the ID to return in the response (only for the first query)\n queries.each do |q|\n results.push dbe.query(dblink, q)\n if @engine.config[\"dbengine\"] == :mysql\n rid ||= dbe.extract(dbe.query(dblink, \"SELECT id FROM #{main_table} WHERE id=(SELECT LAST_INSERT_ID());\"), \"id\")\n else\n rid ||= dbe.extract(dbe.query(dblink, \"SELECT id FROM #{main_table} WHERE id=(last_insert_rowid());\"), \"id\")\n end\n end\n @engine.http_status = \"201 Created\"\n resp = {status: \"201 created\", id: rid}\n end\n return resp\n end",
"def pre_insert(sql, name, pk, id_value, sequence_name)\n @logger.unknown(\"ODBCAdapter#pre_insert>\") if @trace\n @logger.unknown(\"args=[#{sql}|#{name}|#{pk}|#{id_value}|#{sequence_name}]\") if @trace\n @iiTable = get_table_name(sql)\n @logger.unknown(\"@iiTable=#{@iiTable}\") if @trace\n @iiCol = get_autounique_column(@iiTable)\n @logger.unknown(\"@iiCol=#{@iiCol}\") if @trace\n\n if @iiCol != nil\n if query_contains_autounique_col(sql, @iiCol)\n begin\n remove_null_sequence_value_from_sql(sql, @iiCol)\n# rescue Exception => e\n# raise ActiveRecordError, \"IDENTITY_INSERT could not be turned on\"\n end\n end\n end\n end",
"def add_to_database\n hash = self.attr_hash\n columns = hash.keys\n values = hash.values\n if self.valid?\n CONNECTION.execute(\"INSERT INTO #{tablename} (#{columns.join \", \"}) VALUES (#{values.to_s[1...-1]});\")\n @id = CONNECTION.last_insert_row_id\n else\n false\n end\n end",
"def insert(idx, row)\n order_check(row)\n super\n end",
"def visit_Arel_Nodes_InsertStatement o, *a\n [\n \"INSERT INTO #{visit(o.relation).gsub(/\"/, '')}\",\n \"(#{o.columns.map { |x| x.name }.join ', '})\",\n \" VALUES (#{o.values.left.map { |value| value }.join ', '})\"\n ].compact.join ' '\n end",
"def prepared_insert(cols)\n cached_prepared_statement(:insert, prepared_columns(cols)){prepare_statement(dataset, :insert, prepared_statement_key_hash(cols))}\n end",
"def insert\n # Preparing for the query...\n cols = self.class.columns\n col_names = cols.map(&:to_s).join(\", \")\n question_marks = ([\"?\"] * cols.count).join(\", \")\n \n # The actual query\n DBConnection.execute(<<-SQL, *attribute_values)\n INSERT INTO\n #{ self.class.table_name } (#{ col_names })\n VALUES\n (#{ question_marks })\n SQL\n \n # Add an id number for the record\n self.id = DBConnection.last_insert_row_id\n end",
"def insertOperation\n\t\tputs \"Enter Product Id\"\n\t\tproid = gets.chomp.to_i\n\t\tputs \"Enter Product Name\"\n\t\tproname = gets.chomp\n\t\tproname = \"'\"+proname+\"'\"\n\t\tputs \"Enter Product Quantity\"\n\t\tproqty = gets.chomp.to_i\n\t\tputs \"Enter Product Cost\"\n\t\tprocost = gets.chomp.to_i\t\n\t\t\t\n\t\tbegin\n\t\t\tresult = @statement.executeUpdate(\"insert into products values(#{proid},#{proname},#{proqty},#{procost})\")\n\t\t\tif result\n\t\t\t\tputs \"Record Inserted Successfully...........\"\n\t\t\telse\n\t\t\t\tputs \"Error Occur while inserting Record\"\n\t\t\tend\n\t\t\t\t#Closing the Connections\n\t\t\t@connection.close\n\t\t\t@statement.close\n\t\trescue Exception => e\n\t\t\tputs e.message\n\t\tend\n\tend",
"def _insert_dataset\n if upsert_plugin_upserting\n if postgres?\n super.insert_conflict(update: values_to_update, target: self.class.upsert_plugin_identifying_columns)\n elsif mysql?\n columns_to_update = values_to_update.keys - self.class.upsert_plugin_identifying_columns\n super.on_duplicate_key_update(*columns_to_update)\n else\n super\n end\n else\n super\n end\n end",
"def _insert_dataset\n use_server(model.instance_insert_dataset)\n end"
] |
[
"0.74943197",
"0.72217417",
"0.69258296",
"0.6924332",
"0.69196045",
"0.68222225",
"0.67787504",
"0.6745788",
"0.67008656",
"0.67008656",
"0.66724616",
"0.66724616",
"0.65907043",
"0.65866613",
"0.65754974",
"0.65745676",
"0.6538777",
"0.6538777",
"0.65291214",
"0.64823574",
"0.6453642",
"0.6444754",
"0.64074934",
"0.64035606",
"0.63756895",
"0.6364254",
"0.6361346",
"0.6345638",
"0.6343343",
"0.63003707",
"0.62913656",
"0.62867",
"0.6270572",
"0.6266004",
"0.62608695",
"0.62395954",
"0.62340504",
"0.62173456",
"0.62139666",
"0.6213261",
"0.61929905",
"0.617928",
"0.61774856",
"0.6171273",
"0.61640394",
"0.6153086",
"0.61229664",
"0.61078644",
"0.61070055",
"0.610142",
"0.60894823",
"0.6066396",
"0.60624367",
"0.6057257",
"0.6048671",
"0.60346633",
"0.6022275",
"0.60215104",
"0.6021446",
"0.6016677",
"0.6015545",
"0.60135037",
"0.6010565",
"0.5996531",
"0.59729606",
"0.5966089",
"0.59642386",
"0.59522516",
"0.59522516",
"0.5949405",
"0.59373105",
"0.5936672",
"0.59337074",
"0.59282887",
"0.5925485",
"0.59244496",
"0.5924438",
"0.5910725",
"0.59007496",
"0.5899245",
"0.5890338",
"0.58902836",
"0.58901006",
"0.5889329",
"0.5883626",
"0.58721495",
"0.586089",
"0.5853921",
"0.584671",
"0.58436036",
"0.5824917",
"0.58140016",
"0.58130634",
"0.58116704",
"0.5808459",
"0.58084345",
"0.58041865",
"0.579097",
"0.5787408",
"0.5784351",
"0.5781604"
] |
0.0
|
-1
|
table data update process
|
def update_cako
sql = "update from CacheTako"
return sql
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def update(table,cond,op)\n connection.update(path(table),cond,op)\n end",
"def update data, table, condition = {}\n sql = \"update #{table} set\"\n data.each do |k,v|\n v = v.to_json if v.is_a?(Hash)\n if !!v == v\n sql = \"#{sql} #{k}=#{v},\"\n else\n sql = \"#{sql} #{k}='#{v}',\"\n end\n end\n sql = sql[0..-2] + \" where\"\n condition.each do |k,v|\n sql = \"#{sql} #{k} = '#{v}' and\"\n end\n query(sql[0..-4])\n end",
"def process_updates(table_name, field, rows)\n ActiveRecord::Base.transaction do\n mass_import(rows)\n mass_update(table_name, field)\n end\n end",
"def set(recs, data)\r\n # If updates are not in the form of a Proc, convert updates, which\r\n # could be an array, a hash, or a Struct into a common format (i.e.\r\n # hash).\r\n update_rec = convert_input_data(data) unless data.is_a?(Proc)\r\n\r\n updated_recs = []\r\n\r\n # For each one of the recs that matched the update query, apply the\r\n # updates to it and write it back to the database table.\r\n recs.each do |rec|\r\n temp_rec = rec.dup\r\n\r\n if data.is_a?(Proc)\r\n begin\r\n data.call(temp_rec)\r\n rescue NoMethodError\r\n raise 'Invalid field name in code block: %s' % $!\r\n end\r\n else\r\n @field_names.each { |fn| temp_rec[fn] = update_rec.fetch(fn,\r\n temp_rec.send(fn)) }\r\n end\r\n\r\n # Is the user trying to change something they shouldn't?\r\n raise 'Cannot update recno field!' unless \\\r\n rec.recno == temp_rec.recno\r\n raise 'Cannot update internal fpos field!' unless \\\r\n rec.fpos == temp_rec.fpos\r\n raise 'Cannot update internal line_length field!' unless \\\r\n rec.line_length == temp_rec.line_length\r\n\r\n # Are the data types of the updates correct?\r\n validate_input(temp_rec)\r\n\r\n check_required_fields(temp_rec)\r\n\r\n check_against_input_for_specials(temp_rec)\r\n\r\n # Apply updates to the record and add it to an array holding\r\n # updated records. We need the fpos and line_length because\r\n # the engine will use them to determine where to write the\r\n # update and whether the updated record will fit in the old\r\n # record's spot.\r\n updated_recs << { :rec => @field_names.zip(@field_types\r\n ).collect { |fn, ft| convert_to_encoded_string(ft,\r\n temp_rec.send(fn)) }, :fpos => rec.fpos,\r\n :line_length => rec.line_length }\r\n \r\n\r\n # Update any associated blob/memo fields.\r\n temp_rec.each { |r| r.write_to_file if r.is_a?(KBMemo) } if \\\r\n @field_types.include?(:Memo)\r\n temp_rec.each { |r| r.write_to_file if r.is_a?(KBBlob) } if \\\r\n @field_types.include?(:Blob)\r\n end\r\n\r\n # Take all of the update records and write them back out to the\r\n # table's file.\r\n @db.engine.update_records(self, updated_recs)\r\n\r\n # Return the number of records updated.\r\n return recs.size\r\n end",
"def update_record(table, values, org_key = nil)\n update table_update_query(table, values, org_key)\n end",
"def update(table, id, attributes) # abstract\n end",
"def update ; end",
"def update!(**args)\n @delete_all_data_from_table = args[:delete_all_data_from_table] if args.key?(:delete_all_data_from_table)\n @row_key_prefix = args[:row_key_prefix] if args.key?(:row_key_prefix)\n end",
"def update!(**args)\n @header = args[:header] if args.key?(:header)\n @is_table = args[:is_table] if args.key?(:is_table)\n @row = args[:row] if args.key?(:row)\n @total_rows = args[:total_rows] if args.key?(:total_rows)\n end",
"def update() end",
"def update\r\n end",
"def update!(**args)\n @column_headers = args[:column_headers] if args.key?(:column_headers)\n @contains_sampled_data = args[:contains_sampled_data] if args.key?(:contains_sampled_data)\n @data_last_refreshed = args[:data_last_refreshed] if args.key?(:data_last_refreshed)\n @data_table = args[:data_table] if args.key?(:data_table)\n @id = args[:id] if args.key?(:id)\n @items_per_page = args[:items_per_page] if args.key?(:items_per_page)\n @kind = args[:kind] if args.key?(:kind)\n @next_link = args[:next_link] if args.key?(:next_link)\n @previous_link = args[:previous_link] if args.key?(:previous_link)\n @profile_info = args[:profile_info] if args.key?(:profile_info)\n @query = args[:query] if args.key?(:query)\n @rows = args[:rows] if args.key?(:rows)\n @sample_size = args[:sample_size] if args.key?(:sample_size)\n @sample_space = args[:sample_space] if args.key?(:sample_space)\n @self_link = args[:self_link] if args.key?(:self_link)\n @total_results = args[:total_results] if args.key?(:total_results)\n @totals_for_all_results = args[:totals_for_all_results] if args.key?(:totals_for_all_results)\n end",
"def update_all(table, table_hash, values = {}, conditions = {})\n\t\t\t ctx = 0\n\t\t\t\tlen = table.length - 1\n\t\t\t\t\n\t\t\t\tfor i in 0..len\n\t\t\t\t\tupdate = true\n\t\t\t\t\t\n\t\t\t\t\tconditions.each do |key, val|\n\t\t\t\t\t\tif (table[i][table_hash[key]] != val)\n\t\t\t\t\t\t\tupdate = false\n\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\tend\n\t\t\t\t\tend\n\t\t\t\t\t\n\t\t\t\t\tif (update)\n\t\t\t\t\t\tvalues.each do |field, val|\n\t\t\t\t\t\t\ttable[i][table_hash[field]] = val\n\t\t\t\t\t\tend\n\t\t\t\t\t\tctx = ctx + 1\n\t\t\t\t\tend\n\t\t\t\tend\n\t\t\t\t\n\t\t\t\treturn ctx\n\t\t\tend",
"def update; end",
"def update; end",
"def update; end",
"def update; end",
"def update; end",
"def update; end",
"def update; end",
"def update; end",
"def updated_data; end",
"def update_tables\n @old_ethereum_address = @user_extended_details.ethereum_address\n\n @user_extended_details.ethereum_address = @encrypted_ethereum_address\n @user_extended_details.save!\n\n update_user_md5_extended_details\n\n @edit_kyc_request.status = GlobalConstant::EditKycRequest.processed_status\n @edit_kyc_request.save!\n\n @user_kyc_detail.last_acted_by = @admin_id\n @user_kyc_detail.last_acted_timestamp = Time.now.to_i\n @user_kyc_detail.save!\n\n log_activity\n\n success\n end",
"def update\r\n\r\n end",
"def updated_data\n\tend",
"def update\n\n end",
"def update\r\n end",
"def update\r\n end",
"def update\r\n end",
"def update\r\n end",
"def update(model_data)\n end",
"def update_by_primary_key(primary_key, data)\n query = \"UPDATE `#{@table_name}` SET \"+build_update(data)+\" WHERE \"+build_where({@primary_key => primary_key})\n\n begin\n queryresult = @mysql.query(query)\n rescue Exception => e\n @log.error(\"#{e}\")\n return false\n end\n\n expire_table_cache(get_all_related_tables)\n\n get_one({@primary_key => primary_key})\n end",
"def call(_obj, args, _ctx)\n\n update_table = Table.find(args[:table_id])\n return update_table if update_table.update_attributes(name: args[:name], quantity: args[:quantity])\n GraphQL::ExecutionError.new(\"invalid data\")\n end",
"def update!(**args)\n @results_table = args[:results_table] if args.key?(:results_table)\n end",
"def update!(**args)\n @results_table = args[:results_table] if args.key?(:results_table)\n end",
"def update\n \n end",
"def update!(**args)\n @data = args[:data] if args.key?(:data)\n @next_page_token = args[:next_page_token] if args.key?(:next_page_token)\n @column_header = args[:column_header] if args.key?(:column_header)\n end",
"def update!(**args)\n @initial_splits = args[:initial_splits] if args.key?(:initial_splits)\n @table = args[:table] if args.key?(:table)\n @table_id = args[:table_id] if args.key?(:table_id)\n end",
"def update\n check_fields\n sql = \"UPDATE #{table} SET #{to_update_record_str} WHERE id=#{@id}\"\n Database.transaction(sql)\n @log.debug \"Record[#{self}] is updated on Table[#{table}]\"\n end",
"def update \n end",
"def update(sql, name = nil) end",
"def update_data(table_name, column_values, criteria, config={})\n body = column_values.merge!({:ZOHO_CRITERIA => criteria})\n body = body.merge!(config) if config.any?\n\n options = {\n :query => {\n 'ZOHO_ACTION' => 'UPDATE',\n },\n :body => body\n }\n\n send_request get_uri(table_name), 'post', options\n end",
"def exec_update(sql, name, binds)\n exec_query(sql, name, binds)\n end",
"def flush\n conn.transaction do\n buffer.flatten.each do |row|\n # check to see if this row's compound key constraint already exists\n # note that the compound key constraint may not utilize virtual fields\n next unless row_allowed?(row)\n\n # add any virtual fields\n add_virtuals!(row)\n \n key_names = []\n key_values = []\n @key_columns.each do |name|\n key_names << \"#{name}\"\n key_values << conn.quote(row[name]) # TODO: this is probably not database agnostic\n end\n\n names = []\n values = []\n (order - @key_columns).each do |name|\n names << \"#{name}\"\n values << conn.quote(row[name]) # TODO: this is probably not database agnostic\n end\n\n all_name_values = (key_names+names).zip(key_values+values)\n\n q = <<EOF\nMERGE INTO #{table_name} d \nUSING (SELECT #{all_name_values.collect {|c,v| \"#{v} #{c}\"}.join(',')} FROM DUAL) s\nON (#{map_src_to_dest(key_names,'s','d').join(' AND ')})\nWHEN MATCHED THEN \nUPDATE SET #{[map_src_to_dest(names,'s','d'), \"d.#{@update_ts_column}=CURRENT_TIMESTAMP\"].flatten.join(',')}\nWHEN NOT MATCHED THEN\nINSERT (#{all_name_values.collect {|c,v| 'd.'+c}.join(',')},d.#{@insert_ts_column})\nVALUES (#{all_name_values.collect {|c,v| 's.'+c}.join(',')},CURRENT_TIMESTAMP)\nEOF\n #q = \"INSERT INTO `#{table_name}` (#{names.join(',')}) VALUES (#{values.join(',')})\"\n ETL::Engine.logger.debug(\"Executing upsert: #{q}\")\n conn.insert(q, \"Upsert row #{current_row}\")\n @current_row += 1\n end\n buffer.clear\n end\n end",
"def update;end",
"def update\n\t\t\n\t\tend",
"def update \n end",
"def update_table\n\t$table.each do |row|\n\t\tif unbroken?(row)\n\t\t\tclear(row)\n\t\t\t$cleared_lines += 1\n\t\t\t$score += 100\n\t\tend\n\tend\nend",
"def update\n\t\tend",
"def update\n\t\tend",
"def update!(**args)\n @oracle_columns = args[:oracle_columns] if args.key?(:oracle_columns)\n @table = args[:table] if args.key?(:table)\n end",
"def update()\n sql = \"UPDATE customers (name, funds) = ($1, $2) WHERE id = $3\"\n values = [@name, @funds, @id]\n SqlRunner.run(sql,values)\n# => Updating only, no need to return any array\n end",
"def update_values\n end",
"def update_values\n end",
"def update(db, tbl, updates, condition)\n # Build our SQL Query with info provided\n count=updates.size\n prep = \"UPDATE #{tbl} SET \"\n updates.each do |columnset|\n if count.to_i == 1\n prep += \"#{columnset} \"\n else\n prep += \"#{columnset}, \"\n end\n count = count.to_i - 1 \n end\n prep += \"WHERE #{condition};\"\n\n # User Confirmation & Execution of Update Query\n @db_connection.close if @db_connection\n begin\n db_connection = Mysql.connect(@host, @user, @pass, db, @port.to_i)\n query = db_connection.query(\"SELECT * FROM #{tbl} WHERE #{condition};\")\n print_caution(\"Before Update: \")\n query.each { |x| print_line(\"#{x[0]}\") }\n print_line(\"\")\n\n print_caution(\"Please confirm this UPDATE statement looks correct before we execute: \")\n print_caution(\"#{prep}\")\n answer = Readline.readline(\"(Does this look good (Y/N)?)> \", true)\n if \"#{answer.chomp.upcase}\" == \"YES\" or \"#{answer.chomp.upcase}\" == \"Y\"\n print_status(\"OK, sending UPDATE request...\")\n query = db_connection.query(\"#{prep}\")\n print_status(\"After Update: \")\n query = db_connection.query(\"SELECT * FROM #{tbl} WHERE #{condition};\")\n query.each { |x| print_line(\"#{x[0]}\") }\n db_connection.close\n print_line(\"\")\n print_status(\"Hope things worked, if not you can try custom SQL option from the Main Menu\")\n else\n print_error(\"OK, aborting Update request.....\")\n end\n print_status(\"Returning to Main Menu....\")\n rescue Mysql::Error => e\n print_error(\"Problem with Update for #{db}.#{tbl}!\")\n print_error(\"#{e}\")\n end\n @db_connection = Mysql.connect(@host, @user, @pass, @db, @port.to_i)\n end",
"def update(newdata)\n self.class.db.update(self.class.table, newdata, {:id => @id})\n self.should_reload\n self.class.ob.call(\"object\" => self, \"signal\" => \"update\")\n end",
"def update\n \n end",
"def update\n respond_to do |format|\n if @data_table.update(data_table_params)\n format.html { redirect_to @data_table, notice: 'Data table was successfully updated.' }\n format.json { render :show, status: :ok, location: @data_table }\n else\n format.html { render :edit }\n format.json { render json: @data_table.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n \t\n end",
"def update\n \t\n end",
"def update; end",
"def update()\n\n sql = \"UPDATE customers SET (\n name,\n funds\n ) =\n (\n $1, $2\n )\n WHERE id = $3\"\n values = [@title, @price, @id]\n tickets_update = SqlRunner.run(sql, values)\n\nend",
"def update\n update_all\n end",
"def update_records(table_name, old_records, new_records)\n raise \"implement in subclass\"\n end",
"def update\n\n @table = Table.find(params[:id])\n\n @table.number = params[:table][:number].to_i\n @table.size = params[:table][:size].to_i\n @table.status = params[:table][:status].to_i\n @table.name = params[:table][:name]\n\n begin\n Table.transaction do\n @table.save!\n flash[:notice] = '修改餐桌成功'\n\n\n end\n redirect_to :action => :show, :id => @table.id #, :port => PORT\n rescue\n error_message = ''\n @table.errors[:error_message].each do |error|\n if error == @table.errors[:error_message].last\n error_message += error.to_s\n else\n error_message += error.to_s + ','\n end\n end\n flash[:msg] = error_message\n @table.errors.clear\n render :edit and return\n\n end\n\n end",
"def update\n\n end",
"def update!(**args)\n @base_table = args[:base_table] if args.key?(:base_table)\n @sql_query = args[:sql_query] if args.key?(:sql_query)\n @view_type = args[:view_type] if args.key?(:view_type)\n end",
"def sql_update(sql)\n exec(sql)\n # return affected rows.\n end",
"def _update_without_checking(columns)\n ds = _update_dataset\n lc = model.lock_column\n rows = ds.clone(ds.send(:default_server_opts, :sql=>ds.output(nil, [Sequel[:inserted][lc]]).update_sql(columns))).all\n values[lc] = rows.first[lc] unless rows.empty?\n rows.length\n end",
"def update_remaining(table_type_instance)\n update_details table_type_instance.records_to_complete\n end",
"def update\n\n\tend",
"def update\n\n\tend",
"def update\n end",
"def update\n end",
"def update\n end",
"def update\n end",
"def update\n end",
"def update\n end",
"def update\n end",
"def update\n end",
"def update\n end",
"def update\n end",
"def update\n end",
"def update\n respond_to do |format|\n if @sub_data_table.update(sub_data_table_params)\n format.html { redirect_to @sub_data_table.data_table, notice: 'Se actualizó el set de datos correctamente.' }\n format.json { render :show, status: :ok, location: @sub_data_table }\n else\n format.html { render :edit }\n format.json { render json: @sub_data_table.errors, status: :unprocessable_entity }\n end\n end\n end",
"def exec_update(sql, name = nil, binds = []) #:nodoc:\r\n log(sql, \"update\", binds) { query(sql, binds) }\r\n end",
"def update\n respond_to do |format|\n if @table.update(table_params)\n\n # force related running match to refresh\n if @table.banner_id.blank?\n @table.matches.where(is_running: true).each do |match|\n match.touch\n end\n end\n\n format.html { redirect_to tables_path, notice: 'Table was successfully updated.' }\n format.json { render action: 'show', status: :ok, location: @table }\n else\n format.html { render action: 'edit' }\n format.json { render json: @table.errors, status: :unprocessable_entity }\n end\n end\n end",
"def rebuild(table); end",
"def update\n\tend",
"def update\n\tend",
"def update\n\tend",
"def update\n\tend",
"def update\n\tend",
"def update\n \n end",
"def update\n \n end",
"def update\n \n end",
"def update\n \n end",
"def update\n \n end",
"def update\n \n end",
"def update\n \n end",
"def update\n \n end",
"def update\n \n end"
] |
[
"0.6961447",
"0.69002265",
"0.6833356",
"0.6782169",
"0.66434956",
"0.6561265",
"0.6556899",
"0.6556851",
"0.6473217",
"0.645009",
"0.63937616",
"0.6382954",
"0.63729477",
"0.63588774",
"0.63588774",
"0.63588774",
"0.63588774",
"0.63588774",
"0.63588774",
"0.63588774",
"0.63588774",
"0.6350292",
"0.63120216",
"0.62939876",
"0.62920815",
"0.6285859",
"0.62479335",
"0.62479335",
"0.62479335",
"0.62479335",
"0.6241797",
"0.62406385",
"0.62231433",
"0.6214403",
"0.6214403",
"0.6213336",
"0.62113833",
"0.62063456",
"0.6183875",
"0.61735165",
"0.6163871",
"0.61599207",
"0.6148853",
"0.6148518",
"0.614411",
"0.61375564",
"0.6130122",
"0.61271197",
"0.60908383",
"0.60908383",
"0.60836315",
"0.608123",
"0.60802245",
"0.60802245",
"0.6079096",
"0.60779345",
"0.60750675",
"0.6039569",
"0.6032654",
"0.6032654",
"0.60295683",
"0.6028335",
"0.60042065",
"0.60030144",
"0.6000881",
"0.5995025",
"0.59933984",
"0.59866947",
"0.59684527",
"0.596732",
"0.5955772",
"0.5955772",
"0.5947527",
"0.5947527",
"0.5947527",
"0.5947527",
"0.5947527",
"0.5947527",
"0.5947527",
"0.5947527",
"0.5947527",
"0.5947527",
"0.5947527",
"0.59473115",
"0.59460676",
"0.59417576",
"0.59394395",
"0.5937024",
"0.5937024",
"0.5937024",
"0.5937024",
"0.5937024",
"0.5920771",
"0.5920771",
"0.5920771",
"0.5920771",
"0.5920771",
"0.5920771",
"0.5920771",
"0.5920771",
"0.5920771"
] |
0.0
|
-1
|
Return: id of the created object
|
def post_xml(url, ls_data)
uri = URI.parse(url)
request = Net::HTTP::Post.new(uri.request_uri, HEADER_XML)
request.body = ls_data
request.basic_auth(@nsx_user, @nsx_password)
response = Net::HTTP.start(uri.host, uri.port, :use_ssl => true,
:verify_mode => OpenSSL::SSL::VERIFY_NONE) do |https|
https.request(request)
end
return response.body if check_response(response, 201)
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def object_id() end",
"def id\n object.id.to_i\n end",
"def id\n object.id.to_s\n end",
"def id\n object.id\n end",
"def id\n object.id\n end",
"def id\n object.object_id\n end",
"def id\n object_id\n end",
"def return_id(obj)\r\n if (obj.new_record?)\r\n 'new_rec'\r\n else\r\n obj.id\r\n end\r\n end",
"def __object_unique_id__\n name\n end",
"def __object_unique_id__\n name\n end",
"def getId()\n @obj.getId()\n end",
"def __id; @id || @obj.id; end",
"def __object_unique_id__\n return 0 if self.deleted?\n return self.id\n end",
"def id\n __id\n end",
"def id() end",
"def id\n object[\"id\"]\n end",
"def object_id\n object.object_id\n end",
"def id\n object.to_param\n end",
"def id\n @obj['id']\n end",
"def id\n raise Errno::ENOENT, \"This object has been deleted.\" if self.deleted?\n raise \"No ID on object.\" if !@id\n return @id\n end",
"def __object_unique_id__\n return @args[:data][:Key_name]\n end",
"def id; 1; end",
"def get_object\n class_name.create_from_database(id)\n end",
"def sobject_id\n sobject.id\n end",
"def id\n @properties[self.class.primary_key].to_i unless new_record?\n end",
"def _id\n @id\n end",
"def get_id()\n return @id\n end",
"def get_object_id object\n object.respond_to?(:id) ? object.id : object\n end",
"def generate_id \n end",
"def id\n @instance.id\n end",
"def get_id()\n return @id\n end",
"def get_id()\n @id\n end",
"def get_id()\n @id\n end",
"def get_id\n\t\treturn id\n\tend",
"def objname\n id\n end",
"def getId()\n\t\t\treturn @_id\n\t\tend",
"def id()\n #This is a stub, used for indexing\n end",
"def id; 1; end",
"def id_for(obj)\n \"#{obj.class.name.underscore}-#{obj.id}\"\n end",
"def id\n GitlabSchema.id_from_object(object)\n end",
"def record_id(object)\n #eval \"object.#{object.class.primary_key}\"\n object.id\n end",
"def id\n @id ||= Time.now.utc.to_i\n end",
"def id; end",
"def id; end",
"def id; end",
"def id; end",
"def id; end",
"def id; end",
"def id; end",
"def id; end",
"def id; end",
"def id; end",
"def id; end",
"def id; end",
"def id\n return @id\n end",
"def id\n return @id\n end",
"def id\n return @id\n end",
"def id\n return @id\n end",
"def id\n return @id\n end",
"def id\n return @id\n end",
"def id\n return @id\n end",
"def id\n return @id\n end",
"def rails_id\n self.object.id\n end",
"def id\n @id\n end",
"def hash\n object_id\n end",
"def id\n\t\tbuild_id\n\tend",
"def id\n @id\n end",
"def id\n @id\n end",
"def id\n @id\n end",
"def id\n @id\n end",
"def unique_id\n object_id.abs.to_s(16)\n end",
"def object_identifier\n [\"#{self.class.name}\", (id.nil? ? nil : \"##{id}\"), \":0x#{self.object_id.to_s(16)}\"].join\n end",
"def real_id\n @id\n end",
"def real_id\n @id\n end",
"def getId\n return @id\n end",
"def oid\n id(get_oid())\n end",
"def creator_id\n self.id\n end",
"def get_id\n @id\n end",
"def get_id\n @id\n end",
"def get_id\n @id\n end",
"def get_id\n @id\n end",
"def id\n @model.guid\n end",
"def id\n self._id.to_s if self._id\n end",
"def id\n object.external_id\n end",
"def id\n @id\n end",
"def id\n model.id.to_s\n end",
"def id\n self[:identifier]\n end",
"def id\n _id rescue nil\n end",
"def id\n @id || self.class.name.underscore.split('/').last #gsub('/', '_')\n end",
"def identity\n :id\n end",
"def __id__() end",
"def to_id(obj)\n obj.nil? ? nil : obj.__id__\n end",
"def id\r\n @id\r\n end",
"def getId\r\n\t\t\t\t\treturn @id\r\n\t\t\t\tend",
"def getId\r\n\t\t\t\t\treturn @id\r\n\t\t\t\tend",
"def getId\r\n\t\t\t\t\treturn @id\r\n\t\t\t\tend",
"def getId\r\n\t\t\t\t\treturn @id\r\n\t\t\t\tend",
"def getId\r\n\t\t\t\t\treturn @id\r\n\t\t\t\tend",
"def getId\r\n\t\t\t\t\treturn @id\r\n\t\t\t\tend",
"def getId\r\n\t\t\t\t\treturn @id\r\n\t\t\t\tend",
"def getId\r\n\t\t\t\t\treturn @id\r\n\t\t\t\tend"
] |
[
"0.79416245",
"0.7843138",
"0.77955467",
"0.7787047",
"0.7787047",
"0.77575487",
"0.77150655",
"0.76802975",
"0.7676885",
"0.7676885",
"0.7663011",
"0.7623117",
"0.760327",
"0.7476846",
"0.73703384",
"0.7355571",
"0.7333011",
"0.7297001",
"0.7292987",
"0.7258343",
"0.7224329",
"0.7200097",
"0.7172586",
"0.7148814",
"0.71450865",
"0.7121336",
"0.71117085",
"0.7108882",
"0.7105927",
"0.71007067",
"0.7096556",
"0.7096454",
"0.7096454",
"0.70933926",
"0.70703775",
"0.70364726",
"0.70334226",
"0.6997357",
"0.69903725",
"0.6988533",
"0.6982253",
"0.6948005",
"0.69448787",
"0.69448787",
"0.69448787",
"0.69448787",
"0.69448787",
"0.69448787",
"0.69448787",
"0.69448787",
"0.69448787",
"0.69448787",
"0.69448787",
"0.69448787",
"0.69151485",
"0.69145614",
"0.69145614",
"0.69145614",
"0.69145614",
"0.69145614",
"0.69145614",
"0.69145614",
"0.6909493",
"0.69094354",
"0.6908913",
"0.69079477",
"0.69041574",
"0.69041574",
"0.69041574",
"0.69041574",
"0.68990976",
"0.68779075",
"0.6874728",
"0.6874728",
"0.6857764",
"0.685039",
"0.68424803",
"0.6833161",
"0.6833161",
"0.6833161",
"0.6833161",
"0.68207747",
"0.6807971",
"0.67958057",
"0.679101",
"0.67817545",
"0.6771855",
"0.6765914",
"0.67580354",
"0.67568785",
"0.67501336",
"0.6742469",
"0.67275447",
"0.6727374",
"0.6727374",
"0.6727374",
"0.6727374",
"0.6727374",
"0.6727374",
"0.6727374",
"0.6727374"
] |
0.0
|
-1
|
Return: id of the created object
|
def post_json(url, ls_data)
uri = URI.parse(url)
request = Net::HTTP::Post.new(uri.request_uri, HEADER_JSON)
request.body = ls_data
request.basic_auth(@nsx_user, @nsx_password)
response = Net::HTTP.start(uri.host, uri.port, :use_ssl => true,
:verify_mode => OpenSSL::SSL::VERIFY_NONE) do |https|
https.request(request)
end
return JSON.parse(response.body)['id'] \
if check_response(response, 201)
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def object_id() end",
"def id\n object.id.to_i\n end",
"def id\n object.id.to_s\n end",
"def id\n object.id\n end",
"def id\n object.id\n end",
"def id\n object.object_id\n end",
"def id\n object_id\n end",
"def return_id(obj)\r\n if (obj.new_record?)\r\n 'new_rec'\r\n else\r\n obj.id\r\n end\r\n end",
"def __object_unique_id__\n name\n end",
"def __object_unique_id__\n name\n end",
"def getId()\n @obj.getId()\n end",
"def __id; @id || @obj.id; end",
"def __object_unique_id__\n return 0 if self.deleted?\n return self.id\n end",
"def id\n __id\n end",
"def id() end",
"def id\n object[\"id\"]\n end",
"def object_id\n object.object_id\n end",
"def id\n object.to_param\n end",
"def id\n @obj['id']\n end",
"def id\n raise Errno::ENOENT, \"This object has been deleted.\" if self.deleted?\n raise \"No ID on object.\" if !@id\n return @id\n end",
"def __object_unique_id__\n return @args[:data][:Key_name]\n end",
"def id; 1; end",
"def get_object\n class_name.create_from_database(id)\n end",
"def sobject_id\n sobject.id\n end",
"def id\n @properties[self.class.primary_key].to_i unless new_record?\n end",
"def _id\n @id\n end",
"def get_id()\n return @id\n end",
"def get_object_id object\n object.respond_to?(:id) ? object.id : object\n end",
"def generate_id \n end",
"def id\n @instance.id\n end",
"def get_id()\n return @id\n end",
"def get_id()\n @id\n end",
"def get_id()\n @id\n end",
"def get_id\n\t\treturn id\n\tend",
"def objname\n id\n end",
"def getId()\n\t\t\treturn @_id\n\t\tend",
"def id()\n #This is a stub, used for indexing\n end",
"def id; 1; end",
"def id_for(obj)\n \"#{obj.class.name.underscore}-#{obj.id}\"\n end",
"def id\n GitlabSchema.id_from_object(object)\n end",
"def record_id(object)\n #eval \"object.#{object.class.primary_key}\"\n object.id\n end",
"def id\n @id ||= Time.now.utc.to_i\n end",
"def id; end",
"def id; end",
"def id; end",
"def id; end",
"def id; end",
"def id; end",
"def id; end",
"def id; end",
"def id; end",
"def id; end",
"def id; end",
"def id; end",
"def id\n return @id\n end",
"def id\n return @id\n end",
"def id\n return @id\n end",
"def id\n return @id\n end",
"def id\n return @id\n end",
"def id\n return @id\n end",
"def id\n return @id\n end",
"def id\n return @id\n end",
"def rails_id\n self.object.id\n end",
"def id\n @id\n end",
"def hash\n object_id\n end",
"def id\n\t\tbuild_id\n\tend",
"def id\n @id\n end",
"def id\n @id\n end",
"def id\n @id\n end",
"def id\n @id\n end",
"def unique_id\n object_id.abs.to_s(16)\n end",
"def object_identifier\n [\"#{self.class.name}\", (id.nil? ? nil : \"##{id}\"), \":0x#{self.object_id.to_s(16)}\"].join\n end",
"def real_id\n @id\n end",
"def real_id\n @id\n end",
"def getId\n return @id\n end",
"def oid\n id(get_oid())\n end",
"def creator_id\n self.id\n end",
"def get_id\n @id\n end",
"def get_id\n @id\n end",
"def get_id\n @id\n end",
"def get_id\n @id\n end",
"def id\n @model.guid\n end",
"def id\n self._id.to_s if self._id\n end",
"def id\n object.external_id\n end",
"def id\n @id\n end",
"def id\n model.id.to_s\n end",
"def id\n self[:identifier]\n end",
"def id\n _id rescue nil\n end",
"def id\n @id || self.class.name.underscore.split('/').last #gsub('/', '_')\n end",
"def identity\n :id\n end",
"def __id__() end",
"def to_id(obj)\n obj.nil? ? nil : obj.__id__\n end",
"def id\r\n @id\r\n end",
"def getId\r\n\t\t\t\t\treturn @id\r\n\t\t\t\tend",
"def getId\r\n\t\t\t\t\treturn @id\r\n\t\t\t\tend",
"def getId\r\n\t\t\t\t\treturn @id\r\n\t\t\t\tend",
"def getId\r\n\t\t\t\t\treturn @id\r\n\t\t\t\tend",
"def getId\r\n\t\t\t\t\treturn @id\r\n\t\t\t\tend",
"def getId\r\n\t\t\t\t\treturn @id\r\n\t\t\t\tend",
"def getId\r\n\t\t\t\t\treturn @id\r\n\t\t\t\tend",
"def getId\r\n\t\t\t\t\treturn @id\r\n\t\t\t\tend"
] |
[
"0.79416245",
"0.7843138",
"0.77955467",
"0.7787047",
"0.7787047",
"0.77575487",
"0.77150655",
"0.76802975",
"0.7676885",
"0.7676885",
"0.7663011",
"0.7623117",
"0.760327",
"0.7476846",
"0.73703384",
"0.7355571",
"0.7333011",
"0.7297001",
"0.7292987",
"0.7258343",
"0.7224329",
"0.7200097",
"0.7172586",
"0.7148814",
"0.71450865",
"0.7121336",
"0.71117085",
"0.7108882",
"0.7105927",
"0.71007067",
"0.7096556",
"0.7096454",
"0.7096454",
"0.70933926",
"0.70703775",
"0.70364726",
"0.70334226",
"0.6997357",
"0.69903725",
"0.6988533",
"0.6982253",
"0.6948005",
"0.69448787",
"0.69448787",
"0.69448787",
"0.69448787",
"0.69448787",
"0.69448787",
"0.69448787",
"0.69448787",
"0.69448787",
"0.69448787",
"0.69448787",
"0.69448787",
"0.69151485",
"0.69145614",
"0.69145614",
"0.69145614",
"0.69145614",
"0.69145614",
"0.69145614",
"0.69145614",
"0.6909493",
"0.69094354",
"0.6908913",
"0.69079477",
"0.69041574",
"0.69041574",
"0.69041574",
"0.69041574",
"0.68990976",
"0.68779075",
"0.6874728",
"0.6874728",
"0.6857764",
"0.685039",
"0.68424803",
"0.6833161",
"0.6833161",
"0.6833161",
"0.6833161",
"0.68207747",
"0.6807971",
"0.67958057",
"0.679101",
"0.67817545",
"0.6771855",
"0.6765914",
"0.67580354",
"0.67568785",
"0.67501336",
"0.6742469",
"0.67275447",
"0.6727374",
"0.6727374",
"0.6727374",
"0.6727374",
"0.6727374",
"0.6727374",
"0.6727374",
"0.6727374"
] |
0.0
|
-1
|
GET /therapists GET /therapists.json
|
def index
@therapists = Therapist.all
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def show\n @therapist = Therapist.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @therapist }\n end\n end",
"def artists\n if RESPONSE.code == 200\n # Return data to page\n JSON.parse(RESPONSE.to_s)['topartists']['artist']\n else\n # print error message\n \"Error Code #{RESPONSE.code}\"\n end\n end",
"def set_therapist\n @therapist = Therapist.includes(:specialties).find(params[:id])\n end",
"def index\n getProfile\n @consents = TherapistConsent.where('therapist_id = ? ', @therapist.id)\n respond_to do |format|\n format.html { render action: 'index' }\n format.json { render :status => 200, :json => { action: 'index', consents: @consents }}\n end\n end",
"def hotness artist\n url = \"http://developer.echonest.com/api/v4/artist/hotttnesss?api_key=#{ECHONEST_API_KEY}&name=#{artist}&format=json\"\n result = parseURL url\n result[\"response\"][\"artist\"][\"hotttnesss\"]\nend",
"def artists\n # TODO: add param for albums and songs\n a = Artist.all\n render json: a\n end",
"def gist_url\n \"#{api_url}/gists/%s\"\n end",
"def artists\n render :nothing => true and return if params[:id].nil?\n\n if ( params[:meta] == 'tattoos' )\n @artist = Artist.find(params[:id], :include => :tattoos)\n render :json => @artist.tattoos.to_json(:include => { :assets => { :only => [:id, :data_file_name] } })\n return\n end\n \n @artist = Artist.find(params[:id])\n render :json => @artist.to_json(:include => :assets)\n end",
"def gist(id)\n get \"/gists/#{id}\"\n end",
"def show\n @sleuths = HTTParty.get('https://webservice.wikipathways.org/findPathwaysByText?query=' + @sleuth.ext_gene + '&species=homo+sapiens&format=json',\n :headers =>{'Content-Type' => 'application/json'} )\n @pubs = HTTParty.get('https://eutils.ncbi.nlm.nih.gov/entrez/eutils/esearch.fcgi?db=pubmed&retmode=json&term='+ @sleuth.ext_gene,\n :headers =>{'Content-Type' => 'application/json'} )\n end",
"def list_tenants_for_circles(args = {}) \n get(\"/tenants.json/circles\", args)\nend",
"def show\n @artists = Artist.find(params[:id])\n puts @artists\n render json:@artists, status: :ok\n end",
"def index\n @uchronists = Uchronist.all\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @uchronists }\n end\n end",
"def set_therapist\n @therapist = Therapist.find(params[:id])\n end",
"def set_therapist\n @therapist = Therapist.find(params[:id])\n end",
"def set_therapist\n @therapist = Therapist.find(params[:id])\n end",
"def familiarity artist\n url = \"http://developer.echonest.com/api/v4/artist/familiarity?api_key=#{ECHONEST_API_KEY}&name=#{artist}&format=json\"\n result = parseURL url\n result[\"response\"][\"artist\"][\"familiarity\"]\nend",
"def new\n @therapist = Therapist.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @therapist }\n end\n end",
"def index \n artists = Artist.all \n render json: artists \n end",
"def similar artist\n url = \"http://developer.echonest.com/api/v4/artist/similar?api_key=#{ECHONEST_API_KEY}&bucket=years_active&name=#{artist}&format=json&start=0&results=#{RESULTS}\"\n parseURL url\n #also get their hotness?\nend",
"def terms artist\n url = URI.parse(\"http://developer.echonest.com/api/v4/artist/terms?api_key=#{ECHONEST_API_KEY}&name=#{artist}&format=json\")\nend",
"def create\n @therapist = Therapist.new(params[:therapist])\n\n respond_to do |format|\n if @therapist.save\n format.html { redirect_to @therapist, notice: 'Therapist was successfully created.' }\n format.json { render json: @therapist, status: :created, location: @therapist }\n else\n format.html { render action: \"new\" }\n format.json { render json: @therapist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def index # public\n if params[:shelter_id]\n set_shelter\n render json: @shelter.animals\n else\n @animals = Animal.includes(:shelter).all\n render 'index.json.jbuilder'\n end\n end",
"def index\n @pists = Pist.all\n end",
"def show\n @kennel_litter = KennelLitter.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @kennel_litter }\n end\n end",
"def artists(user, options={})\n get(:standard, {:method => \"library.getArtists\", :user => user}.merge(options))\n end",
"def get_gists(screen_name)\n result = self.class.get(\"/users/#{screen_name}/gists\", :headers => @headers)\n puts \"#{result.headers['x-ratelimit-remaining']} requests left!\"\n JSON.parse(result.body)\n result.each do |gist_hash|\n gist_hash['files'].each do |filename, file_hash|\n puts filename\n end\n end\n end",
"def create\n @therapist = Therapist.new(therapist_params)\n\n respond_to do |format|\n if @therapist.save\n format.html { redirect_to @therapist, notice: 'Therapist was successfully created.' }\n format.json { render :show, status: :created, location: @therapist }\n else\n format.html { render :new }\n format.json { render json: @therapist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @therapist = Therapist.new(therapist_params)\n\n respond_to do |format|\n if @therapist.save\n format.html { redirect_to @therapist, notice: 'Therapist was successfully created.' }\n format.json { render :show, status: :created, location: @therapist }\n else\n format.html { render :new }\n format.json { render json: @therapist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def show\n @twitter_user = TwitterUser.find(params[:id])\n @follower_ids = Twitter.follower_ids(@twitter_user.twitter_username).collection\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @twitter_user }\n end\n end",
"def index\n @followships = Follower.all\n render json: @followships\n end",
"def index\n @therapies = Therapy.all\n end",
"def coach\n fetch('football.coaches')\n end",
"def images artist\n url = \"http://developer.echonest.com/api/v4/artist/images?api_key=#{ECHONEST_API_KEY}&name=#{artist}&format=json&results=#{RESULTS}&start=0&license=unknown\"\n result = parseURL url\n result[\"response\"][\"images\"]\nend",
"def artists\n link :top_artists, :Artist, country\n end",
"def list_tenants_for_circle(args = {}) \n get(\"/tenantcircles.json/tenants\", args)\nend",
"def artists_hash\n parse(get_artists)\n end",
"def favourite_artists\n @favourite_artists = FavouriteArtist.all\n render :json => @favourite_artists.to_json\n end",
"def show\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @neuropsychologist }\n end\n end",
"def artists\n link :top_artists, :Artist, name\n end",
"def index\n weathers = Weather.all\n render json: weathers, status: 200\n end",
"def index\n @laboratorists = Laboratorist.all\n end",
"def show\n @uchronist = Uchronist.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @uchronist }\n end\n end",
"def follower\n @users = User.find(params[:id]).followers\n render json: @users\n end",
"def show\n @gists = Gist.where(\"matome_id = ?\", @matome.id)\n end",
"def index\n if params[:album_id]\n @artists = Album.resolve(params[:album_id]).artist\n else\n @artists = Artist.order(:name)\n end\n\n render json: @artists\n end",
"def artists(args={})\n query = \"/?client_id=#{@client_id}&format=#{format.to_s}&#{format_parameters(args)}\"\n path = __method__.to_s\n http_get(path, query)\n end",
"def show\n @party = Party.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @party }\n end\n end",
"def show\n render json: @shelter\n end",
"def show\n @litter = Litter.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @litter }\n end\n end",
"def show\n @artist = Artist.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @artist }\n end\n end",
"def index\n @botanists = Botanist.all\n end",
"def index\n @upcoming_harvests = Harvest.upcoming\n @past_harvests = Harvest.past\n\n if params[:person_id]\n @person = Person.find(params[:person_id])\n @upcoming_harvests = @person.upcoming_harvests\n @past_harvests = @person.past_harvests\n end\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @harvests }\n end\n end",
"def findFollowees\n @user = User.find(params[:id])\n @followees = @user.followees.all\n render json: @followees, status: :ok\n end",
"def people\n Harvest::Resources::Person\n end",
"def index\n ingredients = params[:ingredients]\n if params[:limit]\n searchlimit = params[:limit]\n else\n searchlimit = '10'\n end\n\n require 'nokogiri'\n require 'uri-handler'\n require 'open-uri'\n\n ingredients = ingredients.to_uri\n doc = Nokogiri::HTML(open(\"http://www.webtender.com/cgi-bin/search?name=&ingr=#{ingredients}&what=drink&show=#{searchlimit}&verbose=on\"))\n index = 0\n\n drinknames = []\n ings = []\n cocktails = Hash.new\n\n doc.css(\"td[width='90%']\").each do |node|\n # puts node.xpath('/a').text\n drinknames[index] = node.css('a').text\n ings = node.css(\"small\")[1].text\n\n ings.gsub!('Ingredients: ', '')\n ings = ings.split(', ')\n\n\n cocktails[drinknames[index]] = [ings]\n\n index += 1\n end\n\n respond_to do |format|\n format.json { render :json => cocktails }\n end\n end",
"def index\n @order_yurists = OrderYurist.all\n end",
"def route\n hitch = Hitch.find(params[:hitch_id])\n render json: hitch.geojson\n end",
"def destroy\n @therapist = Therapist.find(params[:id])\n @therapist.destroy\n\n respond_to do |format|\n format.html { redirect_to therapists_url }\n format.json { head :no_content }\n end\n end",
"def index\n if (@shelter.present?)\n @animals = @shelter.animals\n else\n @animals = Animal.all\n end\n end",
"def index\n @specialties = Specialty.all\n\n render json: @specialties\n end",
"def biography artist\n url = \"http://developer.echonest.com/api/v4/artist/biographies?api_key=#{ECHONEST_API_KEY}&name=#{artist}&format=json\"\nend",
"def update\n respond_to do |format|\n if @therapist.update(therapist_params)\n format.html { redirect_to @therapist, notice: 'Therapist was successfully updated.' }\n format.json { render :show, status: :ok, location: @therapist }\n else\n format.html { render :edit }\n format.json { render json: @therapist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @therapist.update(therapist_params)\n format.html { redirect_to @therapist, notice: 'Therapist was successfully updated.' }\n format.json { render :show, status: :ok, location: @therapist }\n else\n format.html { render :edit }\n format.json { render json: @therapist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @therapist.update(therapist_params)\n format.html { redirect_to @therapist, notice: 'Therapist was successfully updated.' }\n format.json { render :show, status: :ok, location: @therapist }\n else\n format.html { render :edit }\n format.json { render json: @therapist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def list_artists\n all_artists = artist_data[\"artists\"][\"items\"]\n all_artists.each do | artist |\n artist_name = artist[\"name\"]\n artists[artist_name] ||= {}\n artists[artist_name][:popularity] = artist[\"popularity\"]\n end\n artists\n end",
"def followers_info\n followers = []\n self.followers.each do |follower|\n followers << follower.attributes.merge(avatar: follower.avatar.url)\n end\n response = {\n message: \"Successfully fetch #{ self.username } followers\",\n relations: followers\n }\n end",
"def index\n manager = Manager.first(:conditions => \"login_id = #{current_login.id}\")\n @neuropsychologists = Neuropsychologist.all(:conditions => \"clinic_id = #{manager.clinic.id}\")\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @neuropsychologists }\n end\n end",
"def show\n hub = Hub.find(params[:id])\n respond_to do |format|\n format.json { render :json => hub.warehouses }\n end\n end",
"def get_json(state, city)\n HTTParty.get(\"http://api.wunderground.com/api/b0938627f87459c4/conditions/q/#{state}/#{city}.json\").parsed_response\nend",
"def index\n @species = Species.eager_load(:family, :species_locations, :images).order('families.name')\n respond_to do |format|\n format.html {\n not_found\n }\n format.xml { render :xml => @species }\n format.json {\n render :template => 'species/index.json'\n }\n end\n end",
"def friends\n #get friends page\n #get json from friends page\n #parse\n []\n end",
"def index\n me = david.new\n me.greets\n @capthurits = Capthurit.all\n end",
"def create_url\n \"#{api_url}/gists\"\n end",
"def show\n @slitter = Slitter.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @slitter }\n end\n end",
"def index\n @tenants = keystone.tenants\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render :json => @tenants }\n end\n end",
"def show\n @gist = Gist.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @gist }\n end\n end",
"def show\n @kitchen = Kitchen.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @kitchen }\n end\n end",
"def index\n @laws = Law.ordered_roots\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @laws }\n end\n end",
"def show\n @trainer = Trainer.find(params[:id])\n\n render json: @trainer.as_json(only: [:name], include: [:trainees])\n end",
"def people\n Birdman::Requester.get(\"movies/#{id}/people\")\n end",
"def index\n @kristine_toys = KristineToy.all\n end",
"def index\n @teathers = Teather.all\n end",
"def index\n @installation = Installation.find(params[:installation_id]) \n @eats = @installation.eats.find(:all, :order => 'name')\n \n @neighborhoods = @installation.neighborhoods\n respond_to do |format|\n format.html # index.html.erb\n format.xml { render :xml => @eats }\n end\n end",
"def destroy\n @therapist.destroy\n respond_to do |format|\n format.html { redirect_to therapists_url, notice: 'Therapist was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @therapist.destroy\n respond_to do |format|\n format.html { redirect_to therapists_url, notice: 'Therapist was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @therapist.destroy\n respond_to do |format|\n format.html { redirect_to therapists_url, notice: 'Therapist was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def get_related_artists(artist_code)\n\n params = {\n seed_artists: artist_code\n }\n JSON.parse(api_call(\"https://api.spotify.com/\",\"v1/recommendations\",params))\n\n end",
"def show\n coach = Coach.find(params[:id])\n json = coach.to_json({:image_url => ActionController::Base.helpers.asset_path(coach.image_url)})\n\n respond_to do |format|\n format.json {render :json => json}\n end\n end",
"def create\n @therapist = Therapist.new(therapist_params)\n @therapist.user_id = current_user.id\n @therapist.specialty_ids = params[:therapist][:specialty_ids]\n\n respond_to do |format|\n if @therapist.save\n format.html { redirect_to root_path, notice: \"Therapist was successfully created.\" }\n else\n format.html { render :new }\n end\n end\n end",
"def show\n @artist_tag = ArtistTag.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @artist_tag }\n end\n end",
"def getToolsSyndicateThomsonlocal( entity_id, destructive)\n params = Hash.new\n params['entity_id'] = entity_id\n params['destructive'] = destructive\n return doCurl(\"get\",\"/tools/syndicate/thomsonlocal\",params)\n end",
"def index\n @likes = target.likes.includes(:author => :profile)\n @people = @likes.map(&:author)\n\n respond_to do |format|\n format.all { render :layout => false }\n format.json { render :json => @likes.as_api_response(:backbone) }\n end\n end",
"def show\n @nutritionist = Nutritionist.includes(:testimonials).find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @nutritionist }\n end\n end",
"def artists\n paintings.map{|art| art.artist.name}\n\n end",
"def show\n user = User.find(params[:id])\n friendships = user.followers + user.followees\n render json: {user: user, friendships: friendships}\n end",
"def heros\n get('/hero/')\n end"
] |
[
"0.6562694",
"0.6169816",
"0.60333884",
"0.5904657",
"0.58211124",
"0.5700912",
"0.5698166",
"0.56811047",
"0.5671451",
"0.5653851",
"0.5626245",
"0.56250376",
"0.55675656",
"0.553357",
"0.553357",
"0.553357",
"0.5489247",
"0.54392534",
"0.54169357",
"0.54157764",
"0.5379153",
"0.534477",
"0.5327641",
"0.53268045",
"0.5308161",
"0.52905476",
"0.528855",
"0.5266234",
"0.5266234",
"0.524193",
"0.52309585",
"0.52273387",
"0.52091795",
"0.51931405",
"0.51816493",
"0.5177794",
"0.517179",
"0.51692903",
"0.5131191",
"0.5129141",
"0.51255316",
"0.51223224",
"0.5120789",
"0.51198953",
"0.51145107",
"0.5102494",
"0.5100297",
"0.5098383",
"0.50642496",
"0.50578636",
"0.505611",
"0.5049861",
"0.504954",
"0.5048536",
"0.5046339",
"0.5041995",
"0.5037188",
"0.5033818",
"0.5032496",
"0.50316733",
"0.5029847",
"0.5027344",
"0.5019966",
"0.5019966",
"0.50194454",
"0.50187266",
"0.50094455",
"0.50079226",
"0.50038135",
"0.49967694",
"0.49869725",
"0.4979084",
"0.49742225",
"0.49729678",
"0.4970908",
"0.49621016",
"0.4960119",
"0.49600467",
"0.49595523",
"0.4956112",
"0.4948961",
"0.49459887",
"0.4945199",
"0.4944154",
"0.4943315",
"0.4943315",
"0.4943315",
"0.49410966",
"0.49408773",
"0.49273515",
"0.4927319",
"0.49267194",
"0.49186194",
"0.4916442",
"0.49150708",
"0.4908951",
"0.4908309"
] |
0.67537636
|
2
|
GET /therapists/1 GET /therapists/1.json
|
def show
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def show\n @therapist = Therapist.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @therapist }\n end\n end",
"def index\n @therapists = Therapist.all\n end",
"def index\n @therapists = Therapist.all\n end",
"def index\n @therapists = Therapist.all\n end",
"def index\n @therapists = Therapist.all\n end",
"def set_therapist\n @therapist = Therapist.includes(:specialties).find(params[:id])\n end",
"def new\n @therapist = Therapist.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @therapist }\n end\n end",
"def index\n getProfile\n @consents = TherapistConsent.where('therapist_id = ? ', @therapist.id)\n respond_to do |format|\n format.html { render action: 'index' }\n format.json { render :status => 200, :json => { action: 'index', consents: @consents }}\n end\n end",
"def show\n @artists = Artist.find(params[:id])\n puts @artists\n render json:@artists, status: :ok\n end",
"def hotness artist\n url = \"http://developer.echonest.com/api/v4/artist/hotttnesss?api_key=#{ECHONEST_API_KEY}&name=#{artist}&format=json\"\n result = parseURL url\n result[\"response\"][\"artist\"][\"hotttnesss\"]\nend",
"def artists\n if RESPONSE.code == 200\n # Return data to page\n JSON.parse(RESPONSE.to_s)['topartists']['artist']\n else\n # print error message\n \"Error Code #{RESPONSE.code}\"\n end\n end",
"def index\n @uchronists = Uchronist.all\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @uchronists }\n end\n end",
"def set_therapist\n @therapist = Therapist.find(params[:id])\n end",
"def set_therapist\n @therapist = Therapist.find(params[:id])\n end",
"def set_therapist\n @therapist = Therapist.find(params[:id])\n end",
"def artists\n render :nothing => true and return if params[:id].nil?\n\n if ( params[:meta] == 'tattoos' )\n @artist = Artist.find(params[:id], :include => :tattoos)\n render :json => @artist.tattoos.to_json(:include => { :assets => { :only => [:id, :data_file_name] } })\n return\n end\n \n @artist = Artist.find(params[:id])\n render :json => @artist.to_json(:include => :assets)\n end",
"def show\n @uchronist = Uchronist.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @uchronist }\n end\n end",
"def gist_url\n \"#{api_url}/gists/%s\"\n end",
"def gist(id)\n get \"/gists/#{id}\"\n end",
"def familiarity artist\n url = \"http://developer.echonest.com/api/v4/artist/familiarity?api_key=#{ECHONEST_API_KEY}&name=#{artist}&format=json\"\n result = parseURL url\n result[\"response\"][\"artist\"][\"familiarity\"]\nend",
"def create\n @therapist = Therapist.new(params[:therapist])\n\n respond_to do |format|\n if @therapist.save\n format.html { redirect_to @therapist, notice: 'Therapist was successfully created.' }\n format.json { render json: @therapist, status: :created, location: @therapist }\n else\n format.html { render action: \"new\" }\n format.json { render json: @therapist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def index # public\n if params[:shelter_id]\n set_shelter\n render json: @shelter.animals\n else\n @animals = Animal.includes(:shelter).all\n render 'index.json.jbuilder'\n end\n end",
"def show\n @sleuths = HTTParty.get('https://webservice.wikipathways.org/findPathwaysByText?query=' + @sleuth.ext_gene + '&species=homo+sapiens&format=json',\n :headers =>{'Content-Type' => 'application/json'} )\n @pubs = HTTParty.get('https://eutils.ncbi.nlm.nih.gov/entrez/eutils/esearch.fcgi?db=pubmed&retmode=json&term='+ @sleuth.ext_gene,\n :headers =>{'Content-Type' => 'application/json'} )\n end",
"def artists\n # TODO: add param for albums and songs\n a = Artist.all\n render json: a\n end",
"def show\n @kennel_litter = KennelLitter.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @kennel_litter }\n end\n end",
"def index \n artists = Artist.all \n render json: artists \n end",
"def show\n @artist = Artist.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @artist }\n end\n end",
"def create\n @therapist = Therapist.new(therapist_params)\n\n respond_to do |format|\n if @therapist.save\n format.html { redirect_to @therapist, notice: 'Therapist was successfully created.' }\n format.json { render :show, status: :created, location: @therapist }\n else\n format.html { render :new }\n format.json { render json: @therapist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @therapist = Therapist.new(therapist_params)\n\n respond_to do |format|\n if @therapist.save\n format.html { redirect_to @therapist, notice: 'Therapist was successfully created.' }\n format.json { render :show, status: :created, location: @therapist }\n else\n format.html { render :new }\n format.json { render json: @therapist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def show\n @slitter = Slitter.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @slitter }\n end\n end",
"def show\n @gist = Gist.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @gist }\n end\n end",
"def show\n @litter = Litter.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @litter }\n end\n end",
"def show\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @neuropsychologist }\n end\n end",
"def show\n @forest = Forest.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @forest }\n end\n end",
"def show\n @party = Party.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @party }\n end\n end",
"def show\n @kitchen = Kitchen.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @kitchen }\n end\n end",
"def show\n @artist_tag = ArtistTag.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @artist_tag }\n end\n end",
"def show\n hub = Hub.find(params[:id])\n respond_to do |format|\n format.json { render :json => hub.warehouses }\n end\n end",
"def show\n @specialist = Specialist.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @specialist }\n end\n end",
"def index\n weathers = Weather.all\n render json: weathers, status: 200\n end",
"def destroy\n @therapist = Therapist.find(params[:id])\n @therapist.destroy\n\n respond_to do |format|\n format.html { redirect_to therapists_url }\n format.json { head :no_content }\n end\n end",
"def route\n hitch = Hitch.find(params[:hitch_id])\n render json: hitch.geojson\n end",
"def show\n @brother = Brother.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @brother }\n end\n end",
"def terms artist\n url = URI.parse(\"http://developer.echonest.com/api/v4/artist/terms?api_key=#{ECHONEST_API_KEY}&name=#{artist}&format=json\")\nend",
"def show\n render json: @shelter\n end",
"def show\n @herb = Herb.find_by_permalink(params[:id])\n \n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @herb }\n end\n end",
"def list_tenants_for_circles(args = {}) \n get(\"/tenants.json/circles\", args)\nend",
"def show\n @stundent = Stundent.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @stundent }\n end\n end",
"def show\r\n\r\n url = URI(\"https://spoonacular-recipe-food-nutrition-v1.p.rapidapi.com/recipes/#{params[:id]}/information\")\r\n\r\n http = Net::HTTP.new(url.host, url.port)\r\n http.use_ssl = true\r\n http.verify_mode = OpenSSL::SSL::VERIFY_NONE\r\n\r\n request = Net::HTTP::Get.new(url)\r\n request[\"x-rapidapi-key\"] = ENV[\"SPOONACULAR_API_KEY\"] # hidden API key\r\n request[\"x-rapidapi-host\"] = 'spoonacular-recipe-food-nutrition-v1.p.rapidapi.com'\r\n\r\n response = http.request(request)\r\n @recipe = JSON.parse response.read_body # gets the recipe\r\n\r\n p url_ingredients = URI(\"https://spoonacular-recipe-food-nutrition-v1.p.rapidapi.com/recipes/#{params[:id]}/ingredientWidget.json\")\r\n\r\n http_ingredients = Net::HTTP.new(url_ingredients.host, url_ingredients.port)\r\n http_ingredients.use_ssl = true\r\n http_ingredients.verify_mode = OpenSSL::SSL::VERIFY_NONE\r\n\r\n request_ingredients = Net::HTTP::Get.new(url_ingredients)\r\n request_ingredients[\"x-rapidapi-key\"] = ENV[\"SPOONACULAR_API_KEY\"]\r\n request[\"x-rapidapi-host\"] = 'spoonacular-recipe-food-nutrition-v1.p.rapidapi.com'\r\n\r\n response_ingredients = http.request(request_ingredients)\r\n # puts response_ingredients.read_body\r\n @ingredients = JSON.parse # data is a string (which looks like a hash -> convert to hash) response_ingredients.read_body\r\n p \"RECIPES\"\r\n # p @recipe\r\n p \"INGREDIENTS\"\r\n p @ingredients\r\n\r\n end",
"def show\n @ingredient = Ingredient.find_by_url_slug(params[:id])\n @ingredient = Ingredient.find(params[:id]) if @ingredient.nil?\n @recipes = @ingredient.recipes.order('created_at DESC')\n logger.debug @recipes.inspect\n respond_to do |format|\n format.html {render :layout => 'wall'}\n format.json { render json: @ingredient }\n end\n end",
"def index\n @specialties = Specialty.all\n\n render json: @specialties\n end",
"def index\n @therapies = Therapy.all\n end",
"def show\n @core_termo_vistorium = Core::TermoVistorium.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @core_termo_vistorium }\n end\n end",
"def show\n @neighborhood = Neighborhood.find(params[:id])\n\n render json: @neighborhood\n end",
"def update\n respond_to do |format|\n if @therapist.update(therapist_params)\n format.html { redirect_to @therapist, notice: 'Therapist was successfully updated.' }\n format.json { render :show, status: :ok, location: @therapist }\n else\n format.html { render :edit }\n format.json { render json: @therapist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @therapist.update(therapist_params)\n format.html { redirect_to @therapist, notice: 'Therapist was successfully updated.' }\n format.json { render :show, status: :ok, location: @therapist }\n else\n format.html { render :edit }\n format.json { render json: @therapist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @therapist.update(therapist_params)\n format.html { redirect_to @therapist, notice: 'Therapist was successfully updated.' }\n format.json { render :show, status: :ok, location: @therapist }\n else\n format.html { render :edit }\n format.json { render json: @therapist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def show\n @kitty = Kitty.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @kitty }\n end\n end",
"def show\n @trainer = Trainer.find(params[:id])\n\n render json: @trainer.as_json(only: [:name], include: [:trainees])\n end",
"def show\n @microplst = Microplst.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @microplst }\n end\n end",
"def show\n @nutritionist = Nutritionist.includes(:testimonials).find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @nutritionist }\n end\n end",
"def index\n if params[:album_id]\n @artists = Album.resolve(params[:album_id]).artist\n else\n @artists = Artist.order(:name)\n end\n\n render json: @artists\n end",
"def show\n @venue = get_venue\n @tour = Tour.find(params[:id])\n @attraction_tours = @tour.attraction_tours.order(\"tour_order ASC\")\n @attractions = @tour.venue.attractions\n respond_to do |format|\n format.html # show.html.erb\n format.json { render :json => @tour.to_json( :include => {:attraction_tours => { :include => :attraction} , :media => {} } ) }\n end\n end",
"def show\n @harvest_trello = HarvestTrello.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @harvest_trello }\n end\n end",
"def show\n coach = Coach.find(params[:id])\n json = coach.to_json({:image_url => ActionController::Base.helpers.asset_path(coach.image_url)})\n\n respond_to do |format|\n format.json {render :json => json}\n end\n end",
"def show\n @taxon = Taxon.find(params[:id])\n @layers = @taxon.layers.page(page).per(25)\n @geo_contexts = @taxon.geo_contexts.page(page).per(25)\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @taxon }\n end\n end",
"def show\n \n @artist = Artist.where(id: params[:id]).first\n if @artist \n render json: @artist, status: 200\n else\n #render error: {error: 'Artist not found'}, status: 404\n return_not_found\n # format.any { render :json => {:response => 'Artista not found' }, :status => 404 }\n \n end\n end",
"def show\n @town = Town.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.xml { render :xml => @town } # hash returned get_stats does not serialize to_xml\n format.json { render :json => @town.to_json({ :methods => [:current_stats, :url, :public_url]}) }\n end\n end",
"def show\n @personerium = Personerium.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @personerium }\n end\n end",
"def show\n recipe = Recipe.find(params[:id])\n # recipes = Recipe.find_by(params[:id])\n # render json: recipe\n render json: recipe\n end",
"def show \n bird = Bird.find_by(id: params[:id])\n if bird \n render bird.json: {id: bird.id, name: bird.name, species: bird.species}\n else \n render {\"Bird not found\"}\n end\nend",
"def show\n @harvest = Harvest.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @harvest }\n end\n end",
"def show\n @harvest = Harvest.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @harvest }\n end\n end",
"def index\n @species = Species.eager_load(:family, :species_locations, :images).order('families.name')\n respond_to do |format|\n format.html {\n not_found\n }\n format.xml { render :xml => @species }\n format.json {\n render :template => 'species/index.json'\n }\n end\n end",
"def show\n @behatweb = Behatweb.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @behatweb }\n end\n end",
"def show\n @father = Father.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @father }\n end\n end",
"def get_json(state, city)\n HTTParty.get(\"http://api.wunderground.com/api/b0938627f87459c4/conditions/q/#{state}/#{city}.json\").parsed_response\nend",
"def show\n @torneo = Torneo.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @torneo }\n end\n end",
"def show\n @neighborhood = Neighborhood.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @neighborhood }\n end\n end",
"def show\n @thermo_oil_distribution_year = ThermoOilDistributionYear.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @thermo_oil_distribution_year }\n end\n end",
"def index\n @upcoming_harvests = Harvest.upcoming\n @past_harvests = Harvest.past\n\n if params[:person_id]\n @person = Person.find(params[:person_id])\n @upcoming_harvests = @person.upcoming_harvests\n @past_harvests = @person.past_harvests\n end\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @harvests }\n end\n end",
"def destroy\n @therapist.destroy\n respond_to do |format|\n format.html { redirect_to therapists_url, notice: 'Therapist was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @therapist.destroy\n respond_to do |format|\n format.html { redirect_to therapists_url, notice: 'Therapist was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @therapist.destroy\n respond_to do |format|\n format.html { redirect_to therapists_url, notice: 'Therapist was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def view\n res = @client.get(path)\n @attributes = res.json if res.success?\n end",
"def index\n @pists = Pist.all\n end",
"def get_thing(query_hash)\n query = {\"type\" => \"/type/edition\"}.merge(query_hash)\n response = open(@api_url + \"/things?query=\" + CGI.escape(query.to_json) ).read\n JSON.parse(response)\n end",
"def show\n @ref_sanatorium = Ref::Sanatorium.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @ref_sanatorium }\n end\n end",
"def similar artist\n url = \"http://developer.echonest.com/api/v4/artist/similar?api_key=#{ECHONEST_API_KEY}&bucket=years_active&name=#{artist}&format=json&start=0&results=#{RESULTS}\"\n parseURL url\n #also get their hotness?\nend",
"def images artist\n url = \"http://developer.echonest.com/api/v4/artist/images?api_key=#{ECHONEST_API_KEY}&name=#{artist}&format=json&results=#{RESULTS}&start=0&license=unknown\"\n result = parseURL url\n result[\"response\"][\"images\"]\nend",
"def show\n @mosttinymobtrail = Mosttinymobtrail.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @mosttinymobtrail }\n end\n end",
"def show\n @galaxies_lenticular_galaxy = Galaxies::LenticularGalaxy.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render :json => @galaxies_lenticular_galaxy }\n end\n end",
"def index\n @followships = Follower.all\n render json: @followships\n end",
"def show\n @taxon = Taxon.find(params[:id])\n @page_title = @taxon.common_name\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: [@taxon.as_json(:include => { :sightings => { :include => [ :trip ] }, :locations => { }, :photos => { :include => [ :taxon, :trip, :location ], :methods => [ :photo_URL ] } } )]}\n end\n end",
"def show\n @interest = Interest.find(params[:id])\n creator = User.find(@interest.user_id)\n creatorName = creator.name\n\n respond_to do |format|\n format.html #show.html.erb\n format.json {render json: @interest}\n end\n end",
"def show\n @interest = Interest.find(params[:id])\n\n respond_to do |format|\n format.json { render json: @interest }\n end\n end",
"def index\n @foodhampers = Foodhamper.all\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @foodhampers }\n end\n end",
"def index\n @universes = Universe.all.page(params[:page]).per(25)\n respond_to do |format|\n format.html\n format.json { render json: @universes }\n end\n end",
"def getToolsSyndicateThomsonlocal( entity_id, destructive)\n params = Hash.new\n params['entity_id'] = entity_id\n params['destructive'] = destructive\n return doCurl(\"get\",\"/tools/syndicate/thomsonlocal\",params)\n end",
"def show\n @tupian = Tupian.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @tupian }\n end\n end",
"def show\n @person_interest = PersonInterest.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @person_interest }\n end\n end"
] |
[
"0.7174276",
"0.6771803",
"0.6771803",
"0.6771803",
"0.6771803",
"0.621438",
"0.613181",
"0.61102575",
"0.5975644",
"0.5957762",
"0.5954145",
"0.5898816",
"0.58721733",
"0.58721733",
"0.58721733",
"0.58653253",
"0.58572644",
"0.58281696",
"0.579335",
"0.57567984",
"0.57113117",
"0.56681275",
"0.56646156",
"0.56535774",
"0.5619497",
"0.5617526",
"0.56058645",
"0.56040835",
"0.56040835",
"0.55671716",
"0.5560353",
"0.5559481",
"0.55538553",
"0.55422425",
"0.5521439",
"0.5519133",
"0.548864",
"0.5485033",
"0.54770696",
"0.5459705",
"0.54297835",
"0.54224604",
"0.5418929",
"0.54166037",
"0.541273",
"0.54088193",
"0.54026634",
"0.53780043",
"0.5361198",
"0.53603584",
"0.53496015",
"0.5345876",
"0.5336389",
"0.53284514",
"0.5325221",
"0.5325221",
"0.53247637",
"0.5298435",
"0.5294167",
"0.5293845",
"0.5293285",
"0.5274089",
"0.5265795",
"0.52640146",
"0.5262574",
"0.526153",
"0.5261061",
"0.52602637",
"0.5243497",
"0.5241111",
"0.5238391",
"0.5230874",
"0.5230874",
"0.52266854",
"0.5219566",
"0.5219005",
"0.52188563",
"0.5218577",
"0.52156556",
"0.52122766",
"0.52085936",
"0.5206806",
"0.5206806",
"0.5206806",
"0.520065",
"0.51967126",
"0.5192728",
"0.51899904",
"0.5188362",
"0.518703",
"0.51859",
"0.5185517",
"0.5185098",
"0.51810044",
"0.5180213",
"0.51792896",
"0.51788545",
"0.5178302",
"0.5172346",
"0.51716495",
"0.5170448"
] |
0.0
|
-1
|
POST /therapists POST /therapists.json
|
def create
@therapist = Therapist.new(therapist_params)
respond_to do |format|
if @therapist.save
format.html { redirect_to @therapist, notice: 'Therapist was successfully created.' }
format.json { render :show, status: :created, location: @therapist }
else
format.html { render :new }
format.json { render json: @therapist.errors, status: :unprocessable_entity }
end
end
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def create\n @therapist = Therapist.new(params[:therapist])\n\n respond_to do |format|\n if @therapist.save\n format.html { redirect_to @therapist, notice: 'Therapist was successfully created.' }\n format.json { render json: @therapist, status: :created, location: @therapist }\n else\n format.html { render action: \"new\" }\n format.json { render json: @therapist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @therapist = Therapist.new(therapist_params)\n @therapist.user_id = current_user.id\n @therapist.specialty_ids = params[:therapist][:specialty_ids]\n\n respond_to do |format|\n if @therapist.save\n format.html { redirect_to root_path, notice: \"Therapist was successfully created.\" }\n else\n format.html { render :new }\n end\n end\n end",
"def create\n @therapist = Therapist.new(therapist_params.merge({code: create_code}))\n if @therapist.save\n render :json => {\n id: @therapist.id,\n email: @therapist.email,\n firstName: @therapist.first_name,\n lastName: @therapist.last_name,\n prefix: @therapist.prefix,\n profession: @therapist.profession,\n code: @therapist.code,\n token: get_token(@therapist)\n }\n else\n render json: @therapist.errors, status: :unprocessable_entity\n # render json: @therapist.errors, status: :unprocessable_entity\n end\n end",
"def follow_artist\n relationship = Relationship.create(params.require(:relationship).permit(:follower_id, :followed_id))\n render json: relationship\n end",
"def register_therapist(attributes)\n user = register_user(attributes)\n user.add_role(Role::Therapist)\n user.save\n user\n end",
"def create\n @artist = Artist.new(artist_params)\n @artist.artist_echo_info\n @artist.related_artists_echo\n @artist.twitter_echo\n\n respond_to do |format|\n if @artist.save\n format.html { redirect_to @artist }\n format.json { render :show, status: :created, location: @artist }\n else\n format.html { redirect_to root_path, notice: \"Musiak couldn't index that request. Maybe try a different spelling.\" }\n format.json { render json: @artist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @pist = Pist.new(pist_params)\n\n respond_to do |format|\n if @pist.save\n format.html { redirect_to @pist, notice: 'Pist was successfully created.' }\n format.json { render :show, status: :created, location: @pist }\n else\n format.html { render :new }\n format.json { render json: @pist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def set_therapist\n @therapist = Therapist.includes(:specialties).find(params[:id])\n end",
"def index\n @therapists = Therapist.all\n end",
"def index\n @therapists = Therapist.all\n end",
"def index\n @therapists = Therapist.all\n end",
"def index\n @therapists = Therapist.all\n end",
"def new\n @therapist = Therapist.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @therapist }\n end\n end",
"def follow\n if request.post?\n fo_ids = params[:follow] \n #fo_str = \"\"\n #fo_cnt = fo_ids.length - 1\n #for i in 0..fo_cnt\n # fo_str +=fo_ids[i].to_s\n # fo_str += \",\" unless fo_cnt == i\n #end\n \n fo_ids.each do |fid|\n hydra = Typhoeus::Hydra.new\n uri = \"http://api.twitter.com/1/friendships/create.json\"\n req = Typhoeus::Request.new(uri,\n :method =>\"post\",\n :params =>{:user_id=>fid, :include_entities=>\"true\"})\n \n sign_request(req,uri)\n hydra.queue(req)\n hydra.run\n #puts req.response.inspect\n end\n end\n redirect_to :action=>\"index\", :page=>\"1\" \n end",
"def therapist_params\n params.require(:therapist).permit(:name, :address, :specialty)\n end",
"def post\n Typhoeus.post(@url,\n body: @results_hash.to_json,\n headers: { 'Content-Type' => 'application/json' })\n end",
"def update\n respond_to do |format|\n if @therapist.update(therapist_params)\n format.html { redirect_to @therapist, notice: 'Therapist was successfully updated.' }\n format.json { render :show, status: :ok, location: @therapist }\n else\n format.html { render :edit }\n format.json { render json: @therapist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @therapist.update(therapist_params)\n format.html { redirect_to @therapist, notice: 'Therapist was successfully updated.' }\n format.json { render :show, status: :ok, location: @therapist }\n else\n format.html { render :edit }\n format.json { render json: @therapist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @therapist.update(therapist_params)\n format.html { redirect_to @therapist, notice: 'Therapist was successfully updated.' }\n format.json { render :show, status: :ok, location: @therapist }\n else\n format.html { render :edit }\n format.json { render json: @therapist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @nutritionist = Nutritionist.new(params[:nutritionist])\n\n respond_to do |format|\n if @nutritionist.save\n format.html { redirect_to @nutritionist, notice: 'Nutritionist was successfully created.' }\n format.json { render json: @nutritionist, status: :created, location: @nutritionist }\n else\n format.html { render action: \"new\" }\n format.json { render json: @nutritionist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n artist = Artist.new({\n name: params[:artist][:name],\n first_name: params[:artist][:first_name],\n age: params[:artist][:age],\n alive: params[:artist][:alive]\n })\n\n artist.save\n redirect_to '/artists'\n end",
"def therapist_params\n params.require(:therapist).permit(:about_me, :hourly_rate, :profile_image, :user_id, :specialty_ids)\n end",
"def set_therapist\n @therapist = Therapist.find(params[:id])\n end",
"def set_therapist\n @therapist = Therapist.find(params[:id])\n end",
"def set_therapist\n @therapist = Therapist.find(params[:id])\n end",
"def create_url\n \"#{api_url}/gists\"\n end",
"def create\n @botanist = Botanist.new(botanist_params)\n\n respond_to do |format|\n if @botanist.save\n format.html { redirect_to @botanist, notice: 'Botanist was successfully created.' }\n format.json { render :show, status: :created, location: @botanist }\n else\n format.html { render :new }\n format.json { render json: @botanist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def add_tenant_circle(args = {}) \n post(\"/tenantcircles.json/\", args)\nend",
"def create\n neo = Neography::Rest.new\n city = neo.create_node(params[:city])\n redirect_to cities_path\n end",
"def show\n @therapist = Therapist.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @therapist }\n end\n end",
"def create\n @gist = Gist.new(params[:gist])\n current_user.gists << @gist\n \n respond_to do |format|\n if @gist.save\n format.html { redirect_to @gist, notice: 'Gist was successfully created.' }\n format.json { render json: @gist, status: :created, location: @gist }\n format.js\n else\n format.html { render action: \"new\" }\n format.json { render json: @gist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @artist = Artist.new(params[:artist])\n\n respond_to do |format|\n if @artist.save\n format.html { redirect_to @artist, :notice => 'Artist was successfully created.' }\n format.json { render :json => @artist, :status => :created, :location => @artist }\n else\n format.html { render :action => \"new\" }\n format.json { render :json => @artist.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def create\n @artist = Artist.new(params[:artist])\n\n respond_to do |format|\n if @artist.save\n format.html { redirect_to @artist, :notice => 'Artist was successfully created.' }\n format.json { render :json => @artist, :status => :created, :location => @artist }\n else\n format.html { render :action => \"new\" }\n format.json { render :json => @artist.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def create\n @artist = Artist.new(artist_params)\n\n respond_to do |format|\n if @artist.save\n format.html { redirect_to @artist, notice: 'Artist was successfully created.' }\n format.json { render :show, status: :created, location: @artist }\n else\n format.html { render :new }\n format.json { render json: @artist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @artist = Artist.new(artist_params)\n\n respond_to do |format|\n if @artist.save\n format.html { redirect_to @artist, notice: 'Artist was successfully created.' }\n format.json { render :show, status: :created, location: @artist }\n else\n format.html { render :new }\n format.json { render json: @artist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @uchronist = Uchronist.new(params[:uchronist])\n\n respond_to do |format|\n if @uchronist.save\n format.html { redirect_to @uchronist, notice: 'Uchronist was successfully created.' }\n format.json { render json: @uchronist, status: :created, location: @uchronist }\n else\n format.html { render action: \"new\" }\n format.json { render json: @uchronist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def artists\n if RESPONSE.code == 200\n # Return data to page\n JSON.parse(RESPONSE.to_s)['topartists']['artist']\n else\n # print error message\n \"Error Code #{RESPONSE.code}\"\n end\n end",
"def terms artist\n url = URI.parse(\"http://developer.echonest.com/api/v4/artist/terms?api_key=#{ECHONEST_API_KEY}&name=#{artist}&format=json\")\nend",
"def create\n\t\t@owner = Prover.find(params[:owner])\n\t\t@owner.follow(Prover.find(params[:follows]))\n\n\t\trespond_to do |format|\n\t\t\tif @owner.follow(Prover.find(params[:follows]))\n\t\t\t\tformat.html { render :json => {prover_id: @owner.id, following_id: params[:follows]}.to_json }\n\t\t\telse\n\t\t\t\tformat.html { render :json => {it: \"BROKE\"}.to_json, :status => :unprocessable_entity }\n\t\t\tend\n\t\tend\n\tend",
"def create\n @animal = @shelter.animals.build(animal_params)\n respond_with(@shelter)\n end",
"def create\n @interest = Interest.new(params[:interest])\n \n respond_to do |format|\n if @interest.save\n format.json { render :json => @interest,\n :status => :created, :location => @interest }\n else\n format.json { render :json => @interest.errors,\n :status => :unprocessable_entity }\n end\n end\n end",
"def create\n @artist = Artist.new(artist_params)\n\n respond_to do |format|\n if @artist.save\n\n format.html { redirect_to @artist, notice: 'Artist was successfully created.' }\n\n format.json { render :show, status: :created, location: @artist }\n\n else\n\n format.html { render :new }\n\n format.json { render json: @artist.errors, status: :unprocessable_entity }\n\n end\n end\n end",
"def create\n @kristine_toy = KristineToy.new(kristine_toy_params)\n\n respond_to do |format|\n if @kristine_toy.save\n format.html { redirect_to @kristine_toy, notice: 'Kristine toy was successfully created.' }\n format.json { render :show, status: :created, location: @kristine_toy }\n else\n format.html { render :new }\n format.json { render json: @kristine_toy.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @order_yurist = OrderYurist.new(order_yurist_params)\n\n respond_to do |format|\n if @order_yurist.save\n format.html { redirect_to @order_yurist, notice: 'Order yurist was successfully created.' }\n format.json { render :show, status: :created, location: @order_yurist }\n else\n set_project_and_apartments\n format.html { render :new }\n format.json { render json: @order_yurist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @artist = Artist.new(artist_params)\n\n respond_to do |format|\n if @artist.save\n format.html { redirect_to @artist, notice: t(:create_notice) }\n format.json { render :show, status: :created, location: @artist }\n else\n format.html { render :new }\n format.json { render json: @artist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @therapy = Therapy.new(therapy_params)\n category = TherapyCategory.create(name: @therapy.name)\n category.therapies << @therapy\n\n respond_to do |format|\n if @therapy.save\n format.html { redirect_to @therapy }\n format.json { render :show, status: :created, location: @therapy }\n else\n format.html { render :new }\n format.json { render json: @therapy.errors, status: :unprocessable_entity }\n end\n end\n end",
"def destroy\n @therapist.destroy\n respond_to do |format|\n format.html { redirect_to therapists_url, notice: 'Therapist was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @therapist.destroy\n respond_to do |format|\n format.html { redirect_to therapists_url, notice: 'Therapist was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @therapist.destroy\n respond_to do |format|\n format.html { redirect_to therapists_url, notice: 'Therapist was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def create(data)\n data.each do |response|\n puts person = @person_repository.create_or_find(response)\n homeworld_response = StarwarsService.get_response(response[\"homeworld\"])\n planet = @planet_repository.find(homeworld_response[\"name\"]).first\n person.planet_id = planet.id\n\n if response[\"species\"].empty? == false\n species_response = StarwarsService.get_response(response[\"species\"].first)\n specie = @specie_repository.find(species_response[\"name\"]).first\n person.specie_id = specie.id\n end\n person.save\n end\n end",
"def create\n @user=User.find(params[:user_id])\n @follower = @user.followers.create(follower_params)\n if @follower.save\n render json: @follower, status: :created\n else\n render json: @follower.errors, status: :unprocessable_entity\n end\n end",
"def create_gists(file, options={})\n opts = {:files => {File.basename(file) => { 'content' => File.read(file) }}}\n options.merge!(opts)\n result = self.class.post(\"/gists\", :body => options.to_json, :headers => @headers)\n if result.code == 201\n puts \"You have successfully created a new gist!\"\n end\n end",
"def create\n @specialist = Specialist.new(specialist_params)\n\n respond_to do |format|\n if @specialist.save\n format.html { redirect_to [:admins, @specialist], notice: 'Specialist was successfully created.' }\n format.json { render :show, status: :created, location: [:admins, @specialist] }\n else\n format.html { render :new }\n format.json { render json: @specialist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @consists_of = ConsistsOf.new(consists_of_params)\n\n respond_to do |format|\n if @consists_of.save\n format.html { redirect_to @consists_of, notice: 'Consists of was successfully created.' }\n format.json { render :show, status: :created, location: @consists_of }\n else\n format.html { render :new }\n format.json { render json: @consists_of.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n\n @recipe = Recipe.new(recipe_params)\n @recipe.recipe_creator_id = current_user.id\n @recipe.allergies = get_allergies_from_params\n @recipe.ingredients = get_selected_ingredients\n\n #Shows an example on how to automatically check whether a recipe is suitable for a person with an intolerance\n laktoseintoleranz = Allergy.where(name: 'Laktoseintoleranz').first\n unless @recipe.allergies.include?(laktoseintoleranz)\n neo = Neography::Rest.new({:username => \"user\", :password => \"user\"})\n includes_laktose_ingredient = false\n\n @recipe.ingredients.each do |ingredient|\n node = neo.execute_query(\"MATCH (n)-[]->(i) WHERE n.name = 'Laktoseintoleranz' AND i.name = '#{ingredient.name}' RETURN i\")\n if node[\"data\"].present?\n includes_laktose_ingredient = true\n end\n end\n unless includes_laktose_ingredient\n @recipe.allergies << laktoseintoleranz\n end\n end\n\n respond_to do |format|\n if @recipe.save\n format.html { redirect_to @recipe, notice: 'Recipe was successfully created.' }\n format.json { render :show, status: :created, location: @recipe }\n else\n format.html { render :new }\n format.json { render json: @recipe.errors, status: :unprocessable_entity }\n end\n end\n end",
"def destroy\n @therapist = Therapist.find(params[:id])\n @therapist.destroy\n\n respond_to do |format|\n format.html { redirect_to therapists_url }\n format.json { head :no_content }\n end\n end",
"def create\n @artist_tag = ArtistTag.new(params[:artist_tag])\n\n respond_to do |format|\n if @artist_tag.save\n format.html { redirect_to artist_tags_path, notice: 'Artist tag was successfully created.' }\n format.json { render json: @artist_tag, status: :created, location: @artist_tag }\n else\n format.html { render action: \"new\" }\n format.json { render json: @artist_tag.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @gist = Gist.new(gist_params)\n\n respond_to do |format|\n if @gist.save\n format.html { redirect_to @gist, notice: 'Gist was successfully created.' }\n format.json { render action: 'show', status: :created, location: @gist }\n else\n format.html { render action: 'new' }\n format.json { render json: @gist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @party = Party.new(params[:party])\n\n respond_to do |format|\n if @party.save\n format.html { redirect_to @party, notice: 'Party was successfully created.' }\n format.json { render json: @party, status: :created, location: @party }\n else\n format.html { render action: \"new\" }\n format.json { render json: @party.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @forest = Forest.new(params[:forest])\n\n respond_to do |format|\n if @forest.save\n format.html { redirect_to @forest, notice: 'Forest was successfully created.' }\n format.json { render json: @forest, status: :created, location: @forest }\n else\n format.html { render action: \"new\" }\n format.json { render json: @forest.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @artist = Artist.new(params[:artist])\n @artist.user = @current_user\n\n respond_to do |format|\n if @artist.save\n format.html { redirect_to session[:redirect_origin].pop || artist_url(@artist), notice: 'Artist was successfully created.' }\n format.json { render json: @artist, status: :created, location: @artist }\n else\n format.html { render action: 'new' }\n format.json { render json: @artist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def createCharities\n\tcharity_list = [\"Direct Relief\", \"Catholic Medical Mission Board\", \"MAP International\", \"United Nations Foundation\", \"The Rotary Foundation of Rotary International\", \"Samaritan's Purse\", \"Institute of International Education\", \"International Rescue Committee\", \"Compassion International\", \"United States Fund for UNICEF\"]\n\tcharity_list.each do |charity|\n\t\tRestClient.post 'http://api.reimaginebanking.com/merchants?key=e0486a76005721ee6d86b140eaea2a40', { \"name\": \"#{charity}\"}.to_json, :content_type => :json, :accept => :json\n\tend\nend",
"def create\n megam_rest.post_node(to_hash)\n end",
"def create\n @laboratorist = Laboratorist.new(laboratorist_params)\n\n respond_to do |format|\n if @laboratorist.save\n format.html { redirect_to @laboratorist, notice: 'Laboratorist was successfully created.' }\n format.json { render :show, status: :created, location: @laboratorist }\n else\n format.html { render :new }\n format.json { render json: @laboratorist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create_item()\n\n request_body = {\n 'name' => 'Milkshake',\n 'variations' => [\n {\n 'name' => 'Small',\n 'pricing_type' => 'FIXED_PRICING',\n 'price_money' => {\n 'currency_code' => 'USD',\n 'amount' => 400\n }\n }\n ]\n }\n\n response = Unirest.post CONNECT_HOST + '/v1/' + LOCATION_ID + '/items',\n headers: REQUEST_HEADERS,\n parameters: request_body.to_json\n\n if response.code == 200\n puts 'Successfully created item:'\n puts JSON.pretty_generate(response.body)\n return response.body\n else\n puts 'Item creation failed'\n puts response.body\n return nil\n end\nend",
"def create_recipe_request(version, auth_headers, data = {})\n post \"/api/recipes\", params: data, headers: {'Content-Type' => \"application/json\", 'Accept' => \"application/vnd.ink.#{version}\" }.merge(auth_headers)\nend",
"def create\n @seeker = Seeker.new(seeker_params)\n @seeker.user_id = User.find(session[:user_id]).uid\n @seeker.skill_list.add(params[:seeker][:skill_list].to_s.downcase, parse: true)\n\n respond_to do |format|\n if @seeker.save\n format.html { redirect_to root_path, notice: 'Seeker was successfully created.' }\n format.json { render action: 'show', status: :created, location: @seeker }\n else\n format.html { render action: 'new' }\n format.json { render json: @seeker.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @similar_artist = SimilarArtist.new(similar_artist_params)\n\n respond_to do |format|\n if @similar_artist.save\n format.html { redirect_to @similar_artist, notice: 'Similar artist was successfully created.' }\n format.json { render :show, status: :created, location: @similar_artist }\n else\n format.html { render :new }\n format.json { render json: @similar_artist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @therapist.specialty_ids = params[:therapist][:specialty_ids]\n respond_to do |format|\n if @therapist.update(therapist_params)\n format.html { redirect_to root_path, notice: \"therapist was successfully updated.\" }\n else\n format.html { render :edit }\n end\n end\n end",
"def therapist_params\n params.require(:therapist).permit(:pt_id, :name, :summary, :phone)\n end",
"def create\n @party = Party.new(party_params)\n\n respond_to do |format|\n if @party.save\n format.html { redirect_to @party, notice: 'Party was successfully created.' }\n format.json { render :show, status: :created, location: @party }\n else\n format.html { render :new }\n format.json { render json: @party.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @forest = Forest.new(forest_params)\n\n respond_to do |format|\n if @forest.save\n format.html { redirect_to @forest, notice: 'Forest was successfully created.' }\n format.json { render action: 'show', status: :created, location: @forest }\n else\n format.html { render action: 'new' }\n format.json { render json: @forest.errors, status: :unprocessable_entity }\n end\n end\n end",
"def add_contributors\n @hash[:contributors].each do |json_contrib|\n my_hash = json_contrib.slice('contributor_name', 'contributor_type', 'created_at', 'updated_at', 'award_number')\n @ar_resource.contributors << StashDatacite::Contributor.create(my_hash)\n # btw affiliations for contributors are not really used in Dash data, so skipping that headache, though we have some test data for it\n end\n end",
"def gist_url\n \"#{api_url}/gists/%s\"\n end",
"def test_start_following_a_user\r\n post \"/follows.json?api_key=bobbyapikey&followee_id=3\" \r\n assert_response :created\r\n follow = JSON.parse(response.body) \r\n assert follow['follower_id'] == 6\r\n assert follow['followee_id'] == 3\r\n end",
"def create\n @user_artist = UserArtist.new\n\t@artist = Artist.find params[:artist_id]\n\t@user_artist.user_id = @current_user.id\n\t@user_artist.artist_id = @artist.id\n\t\n respond_to do |format|\n if @user_artist.save\n format.html { redirect_to @artist, notice: 'User artist was successfully created.' }\n format.json { render :show, status: :created, location: @user_artist }\n else\n format.html { render :new }\n format.json { render json: @user_artist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def post_spoonacular\n # %encode ingredients to url\n encoded_ingr = URI.escape(@translated_recipe[:ingredients_list])\n # post call block :\n url = URI(\"https://spoonacular-recipe-food-nutrition-v1.p.rapidapi.com/recipes/parseIngredients?includeNutrition=true\")\n http = Net::HTTP.new(url.host, url.port)\n http.use_ssl = true\n http.verify_mode = OpenSSL::SSL::VERIFY_NONE\n request = Net::HTTP::Post.new(url)\n request[\"content-type\"] = \"application/x-www-form-urlencoded\"\n request[\"x-rapidapi-key\"] = ENV[\"X_RAPIDAPI_KEY\"]\n request[\"x-rapidapi-host\"] = \"spoonacular-recipe-food-nutrition-v1.p.rapidapi.com\"\n # body of the call with ingredients and servings\n request.body = \"ingredientList=#{encoded_ingr}&#{@recipe_hash[:servings]}\"\n # response\n response = http.request(request)\n end",
"def create\n @shelter = Shelter.new(shelter_params)\n\n if @shelter.save\n render json: @shelter, status: :created, location: @shelter\n else\n render json: @shelter.errors, status: :unprocessable_entity\n end\n end",
"def create\n @odist = Odist.new(odist_params)\n\n respond_to do |format|\n if @odist.save\n format.html { redirect_to @odist, notice: 'Odist was successfully created.' }\n format.json { render :show, status: :created, location: @odist }\n else\n format.html { render :new }\n format.json { render json: @odist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n name_normalized = ActsAsTaggableOn::Tag.normalize(params['name'])\n\n if name_normalized.nil? || name_normalized.empty?\n render :nothing => true, :status => 403\n else\n @tag = ActsAsTaggableOn::Tag.find_or_create_by_name(name_normalized)\n @tag_following = current_user.tag_followings.new(:tag_id => @tag.id)\n\n if @tag_following.save\n render :json => @tag.to_json, :status => 201\n else\n render :nothing => true, :status => 403\n end\n end\n end",
"def create\n res = self.class.post('/', body: attrs)\n res.created?\n end",
"def create\n @turistum = Turistum.new(turistum_params)\n\n respond_to do |format|\n if @turistum.save\n format.html { redirect_to @turistum, notice: 'Turistum was successfully created.' }\n format.json { render :show, status: :created, location: @turistum }\n else\n format.html { render :new }\n format.json { render json: @turistum.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n dish = Dish.where(name: params[:name]).first\n dish ||= Dish.create(name: params[:name], party: current_party, price: params[:price])\n\n OrderItem.create(user: current_user, dish: dish)\n\n render json: { :name => dish.name, :price => dish.price, :participants => [current_user.login] }\n end",
"def index\n getProfile\n @consents = TherapistConsent.where('therapist_id = ? ', @therapist.id)\n respond_to do |format|\n format.html { render action: 'index' }\n format.json { render :status => 200, :json => { action: 'index', consents: @consents }}\n end\n end",
"def create\n @ministry = Ministry.find(params[:ministry_id])\n current_user.follow(@ministry)\n respond_to do |format|\n format.html { redirect_to @ministry, notice: 'Ministry Followed' }\n format.js\n end\n end",
"def api_request(name) \n\tresponse = HTTParty.post('https://api.talentiq.co/v2/person', \n :body => {:name => name }.to_json,\n :headers => { 'Content-Type' => 'application/json', \n \t\t\t'x-api-key:' => 'e058cdd334c69c31d45d87e44b405d128a9e8937939e7d0ab2496334',\n 'Accept' => 'application/json' })\n puts response.body\nend",
"def create\n @person_interest = PersonInterest.new(params[:person_interest])\n\n respond_to do |format|\n if @person_interest.save\n format.html { redirect_to @person_interest, notice: 'Person interest was successfully created.' }\n format.json { render json: @person_interest, status: :created, location: @person_interest }\n else\n format.html { render action: \"new\" }\n format.json { render json: @person_interest.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @hood = Hood.new(hood_params)\n\n respond_to do |format|\n if @hood.save\n format.html { redirect_to @hood, notice: 'Hood was successfully created.' }\n format.json { render action: 'show', status: :created, location: @hood }\n else\n format.html { render action: 'new' }\n format.json { render json: @hood.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n if @mission.torched_by? current_user\n return respond_to do |format|\n format.html { redirect_to mission_path(@mission) }\n format.json { head :no_content }\n end\n end\n @mission.torches.create torcher: current_user\n respond_to do |format|\n # TODO: notification and activity\n format.html { redirect_to mission_path(@mission), notice: t('missions.torch_awarded_successfully') }\n format.json { render json: { count: @mission.torches.count } }\n end\n end",
"def create\n @stundent = Stundent.new(params[:stundent])\n\n respond_to do |format|\n if @stundent.save\n format.html { redirect_to @stundent, notice: 'Stundent was successfully created.' }\n format.json { render json: @stundent, status: :created, location: @stundent }\n else\n format.html { render action: \"new\" }\n format.json { render json: @stundent.errors, status: :unprocessable_entity }\n end\n end\n end",
"def artists\n render :nothing => true and return if params[:id].nil?\n\n if ( params[:meta] == 'tattoos' )\n @artist = Artist.find(params[:id], :include => :tattoos)\n render :json => @artist.tattoos.to_json(:include => { :assets => { :only => [:id, :data_file_name] } })\n return\n end\n \n @artist = Artist.find(params[:id])\n render :json => @artist.to_json(:include => :assets)\n end",
"def create\n authorize! :index, @user, :message => 'Not authorized as an administrator.'\n\n @specialist = Specialist.new(params[:specialist])\n\n respond_to do |format|\n if @specialist.save\n format.html { redirect_to @specialist, notice: 'Specialist was successfully created.' }\n format.json { render json: @specialist, status: :created, location: @specialist }\n else\n format.html { render action: \"new\" }\n format.json { render json: @specialist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @specialty = Specialty.new(specialty_params)\n\n if @specialty.save\n render json: @specialty, status: :created, location: @specialty\n else\n render json: @specialty.errors, status: :unprocessable_entity\n end\n end",
"def perform(*args)\n # call petfinder\n response = HTTParty.get('http://api.petfinder.com/pet.find?key=49b6e87785e4e2811b2d5a9668eee5af&format=xml&location=ontario&count=1000')\n # parse xml\n xml_doc = Nokogiri::XML(response.body)\n\n # create hash of attributes\n xml_doc.xpath('//pet').each do |pet_xml|\n pet_attributes = {\n petfinder_id: pet_xml.xpath('id/text()').to_s.to_i,\n city: pet_xml.xpath('contact/city/text()').to_s,\n animal_type: pet_xml.xpath('animal/text()').to_s,\n breed: pet_xml.xpath('breeds/breed/text()')[0].to_s,\n name: pet_xml.xpath('name/text()').to_s,\n gender: pet_xml.xpath('sex/text()').to_s,\n size: pet_xml.xpath('size/text()').to_s,\n age: pet_xml.xpath('age/text()').to_s,\n status: pet_xml.xpath('status/text()').to_s,\n photo: pet_xml.xpath('media/photos/photo[3]/text()').to_s,\n description: pet_xml.xpath('description/text()').to_s,\n shelter_id: pet_xml.xpath('shelterId/text()').to_s\n }\n # save in DB\n #Pet.find_or_create_by!(petfinder_id: pet_id) do |pet|\n # pet.update(pet_attributes)\n #end\n Pet.create!(pet_attributes) unless (Pet.exists?(petfinder_id: pet_attributes[:petfinder_id]) || !Shelter.exists?(shelter_id: pet_attributes[:shelter_id]))\n end\n end",
"def artist_params\n params.permit(:city, :state, :spotify_id, :name, :followers, :popularity, :genres, :images)\n end",
"def create\n @plant = Plant.find(params[:plant])\n @friendship = @plant.friendships.build(:friend_id => params[:friend])\n if @friendship.save\n flash[:notice] = \"added friend\"\n redirect_to request.referrer\n else\n flash[:error] = \"Unable to add friend.\"\n redirect_to request.referer\n end\n end",
"def follow\r\n @relationship = Relationship.create(follower_id: current_user.id, followed_id: params[:followed_id])\r\n @relationship.create_activity key: 'relationship.follow', owner: current_user, recipient: User.find(params[:followed_id])\r\n\r\n if @relationship.save\r\n render json: @relationship\r\n else\r\n render json: { error: \"Relationship creating error\" }, status: :unprocessable_entity\r\n end\r\n end",
"def create\n passenger = Passenger.new(:name => params[:name], :contact_number => params[:contact_number], :nationality => params[:nationality], :meal_pref => params[:meal_pref])\n passenger.save\n render :json => passenger\n end",
"def create\n @trustee = Trustee.new(trustee_params)\n\n respond_to do |format|\n if @trustee.save\n format.html { redirect_to @trustee, notice: 'Trustee was successfully created.' }\n format.json { render action: 'show', status: :created, location: @trustee }\n else\n format.html { render action: 'new' }\n format.json { render json: @trustee.errors, status: :unprocessable_entity }\n end\n end\n end"
] |
[
"0.6404418",
"0.61935127",
"0.61634153",
"0.5756791",
"0.5664453",
"0.5594197",
"0.55832016",
"0.5535142",
"0.55215055",
"0.55215055",
"0.55215055",
"0.55215055",
"0.5423509",
"0.5402538",
"0.5380349",
"0.5351084",
"0.5314024",
"0.5314024",
"0.53133786",
"0.52829635",
"0.5273785",
"0.5254459",
"0.5253865",
"0.5253865",
"0.5253865",
"0.52493906",
"0.52077246",
"0.5180468",
"0.5144813",
"0.51397806",
"0.51341313",
"0.5099847",
"0.5099847",
"0.50804955",
"0.50804955",
"0.50750184",
"0.50662345",
"0.50660586",
"0.5051959",
"0.50441873",
"0.50343186",
"0.5029074",
"0.5023066",
"0.5021044",
"0.5020091",
"0.5014946",
"0.5003758",
"0.5003758",
"0.5003758",
"0.4986775",
"0.49859405",
"0.49855566",
"0.49828592",
"0.4972621",
"0.4948824",
"0.49308655",
"0.49238023",
"0.49216193",
"0.49191955",
"0.489859",
"0.4893734",
"0.48933136",
"0.48846135",
"0.4869391",
"0.48623377",
"0.48555806",
"0.48508",
"0.4849883",
"0.4848606",
"0.4847001",
"0.4846062",
"0.4845333",
"0.48398837",
"0.48392913",
"0.48297027",
"0.48266825",
"0.48153412",
"0.4809261",
"0.48086667",
"0.4801805",
"0.47931466",
"0.4788733",
"0.47846723",
"0.47780132",
"0.47760132",
"0.47727567",
"0.4767157",
"0.4766767",
"0.4766524",
"0.47647718",
"0.47631273",
"0.47604078",
"0.47576073",
"0.47556278",
"0.47541225",
"0.47473776",
"0.47450373",
"0.47396383",
"0.4735402"
] |
0.63292545
|
2
|
PATCH/PUT /therapists/1 PATCH/PUT /therapists/1.json
|
def update
respond_to do |format|
if @therapist.update(therapist_params)
format.html { redirect_to @therapist, notice: 'Therapist was successfully updated.' }
format.json { render :show, status: :ok, location: @therapist }
else
format.html { render :edit }
format.json { render json: @therapist.errors, status: :unprocessable_entity }
end
end
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def update\n @therapist = Therapist.find(params[:id])\n\n respond_to do |format|\n if @therapist.update_attributes(params[:therapist])\n format.html { redirect_to @therapist, notice: 'Therapist was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @therapist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @therapist.specialty_ids = params[:therapist][:specialty_ids]\n respond_to do |format|\n if @therapist.update(therapist_params)\n format.html { redirect_to root_path, notice: \"therapist was successfully updated.\" }\n else\n format.html { render :edit }\n end\n end\n end",
"def update!(params)\n res = @client.put(path, nil, params, \"Content-Type\" => \"application/json\")\n @attributes = res.json if res.status == 201\n res\n end",
"def update_tenant_circle(args = {}) \n put(\"/tenantcircles.json/#{args[:circleId]}\", args)\nend",
"def update\n respond_to do |format|\n if @pist.update(pist_params)\n format.html { redirect_to @pist, notice: 'Pist was successfully updated.' }\n format.json { render :show, status: :ok, location: @pist }\n else\n format.html { render :edit }\n format.json { render json: @pist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update # PATCH\n raise NotImplementedError\n end",
"def update(attrs, path=nil)\n resp = api_client.put(path || url, JSON.dump(attrs))\n refresh(JSON.load(resp.body))\n end",
"def put!\n request! :put\n end",
"def update\n @animal.update(animal_params)\n respond_with(@shelter)\n end",
"def update(url, data)\n RestClient.put url, data, :content_type => :json\nend",
"def update\n respond_to_update({thing: @author})\n end",
"def patch(type, info)\n path, info = type_info(type, :path), force_case(info)\n ida = type == :client ? 'client_id' : 'id'\n raise ArgumentError, \"info must include #{ida}\" unless id = info[ida]\n hdrs = headers\n if info && info['meta'] && (etag = info['meta']['version'])\n hdrs.merge!('if-match' => etag)\n end\n reply = json_parse_reply(@key_style,\n *json_patch(@target, \"#{path}/#{Addressable::URI.encode(id)}\", info, hdrs))\n\n # hide client endpoints that are not quite scim compatible\n type == :client && !reply ? get(type, info['client_id']): reply\n end",
"def update_aos_version(args = {}) \n id = args['id']\n temp_path = \"/aosversions.json/{aosVersionId}\"\n path = temp_path\nargs.keys.each do |key|\n if (key == \"aosversionId\")\n args.delete(key)\n path = temp_path.gsub(\"{#{key}}\", id)\n end\nend\n puts \" PATH : #{path}\"\n put(path, args)\nend",
"def update options={}\n client.put(\"/#{id}\", options)\n end",
"def update\n @artist = Artist.find(params[:id])\n\n respond_to do |format|\n if @artist.update_attributes(params[:artist])\n format.html { redirect_to @artist, notice: 'Artist was successfully updated.' }\n format.json { head :ok }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @artist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update attrs\n put '', :body => {self.class.api_name => self.class.opts_to_query_opts(attrs)}\n end",
"def update\n @supermarket = Supermarket.find(params[:id]) \n respond_to do |format|\n if @supermarket.update(supermarket_params)\n format.json { render json: @supermarket, status: :ok }\n end\n end\n end",
"def update_by_body\n @person = Person.find(person_update_params[:id])\n\n if @person.update_attributes(person_update_params)\n render json: { status: 'PUT Success' }, status: :ok\n else\n render json: { status: 'Error', message:'Error updating person', person: @person.errors }, status: :unprocessable_entity\n end\n end",
"def update\n @artist = Artist.find(params[:id])\n\n respond_to do |format|\n if @artist.update_attributes(params[:artist])\n format.html { redirect_to artist_url(@artist), notice: 'Artist was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: 'edit' }\n format.json { render json: @artist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def edit\n @therapist_consent = TherapistConsent.find(params[:id])\n respond_to do |format|\n format.html { render action: 'edit' }\n format.json { render :status => 200, :json => { action: 'edit', therapist_consent: @therapist_consent}}\n end\n end",
"def patch!\n request! :patch\n end",
"def update\n @uchronist = Uchronist.find(params[:id])\n\n respond_to do |format|\n if @uchronist.update_attributes(params[:uchronist])\n format.html { redirect_to @uchronist, notice: 'Uchronist was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @uchronist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @gist = Gist.find(params[:id])\n\n respond_to do |format|\n if @gist.update_attributes(params[:gist])\n format.html { redirect_to @gist, notice: 'Gist was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @gist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def jsonapi_update!(attributes)\n assign_jsonapi_attributes(attributes)\n save!\n end",
"def update\n @serving = Serving.find(params[:id])\n\n respond_to do |format|\n if @serving.update_attributes(params[:serving])\n format.html { redirect_to @serving, notice: 'Serving was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @serving.errors, status: :unprocessable_entity }\n end\n end\n end",
"def patch\n headers = {\"If-Match\" => @version}\n response = @context.request :patch, \"#{@path}/#{@id}\", @data.to_json, headers\n @version += 1\n response\n # 'X-HTTP-Method-Override' => 'PATCH'\n end",
"def update(attributes: {})\n attributes = attributes.with_indifferent_access\n clean_attributes(attributes)\n titleize(attributes)\n sync_tags(attributes)\n parse_ingredients(attributes)\n recipe.update(attributes)\n recipe_response\n rescue StandardError => e\n error_response(e)\n end",
"def update!(**args)\n @id = args[:id] if args.key?(:id)\n @kind = args[:kind] if args.key?(:kind)\n @name = args[:name] if args.key?(:name)\n @starred = args[:starred] if args.key?(:starred)\n @web_properties = args[:web_properties] if args.key?(:web_properties)\n end",
"def update!(**args)\n @attributes = args[:attributes] if args.key?(:attributes)\n @id = args[:id] if args.key?(:id)\n @kind = args[:kind] if args.key?(:kind)\n end",
"def update\n respond_to do |format|\n if @specialist.update(specialist_params)\n format.html { redirect_to '/mySpecialists' }\n format.json { render :show, status: :ok, location: @specialist }\n else\n format.html { render :edit }\n format.json { render json: @specialist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @slitter = Slitter.find(params[:id])\n\n respond_to do |format|\n if @slitter.update_attributes(params[:slitter])\n format.html { redirect_to @slitter, notice: 'Slitter was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @slitter.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @artist = Artist.find(params[:id])\n\n respond_to do |format|\n if @artist.update_attributes(params[:artist])\n format.html { redirect_to @artist, :notice => 'Artist was successfully updated.' }\n format.json { head :no_content }\n format.js\n else\n format.html { render :action => \"edit\" }\n format.json { render :json => @artist.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def update\n @artist = Artist.find(params[:id])\n\n respond_to do |format|\n if @artist.update_attributes(params[:artist])\n format.html { redirect_to @artist, :notice => 'Artist was successfully updated.' }\n format.json { head :no_content }\n format.js\n else\n format.html { render :action => \"edit\" }\n format.json { render :json => @artist.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def update\n @species = Species.find(params[:id])\n @species.update_attributes(params[:species])\n \n respond_with(@species, :location => admin_species_path(@species))\n end",
"def update\n ingredient.update(ingredient_params)\n render json: ingredient\n end",
"def put(*args)\n request :put, *args\n end",
"def update\n if request.content_type == \"application/json\"\n # .update is like a \"update people set ...\" in sql\n if @person.update(person_params)\n render json: @person\n else\n render json: @person.errors, status: :not_found\n end\n else\n render status: :bad_request\n end\n end",
"def update\n @todo = Todo.find(params[:id])\n @todo.update_attributes(params[:todo])\n render :json => @todo\n end",
"def update\n official = Official.find(params[:id])\n if official.update(official_params)\n render json: official, status: 200, location: [:api, official]\n else\n failed_to_update(official, \"official\")\n end\n end",
"def update\n @nutritionist = Nutritionist.find(params[:id])\n\n respond_to do |format|\n if @nutritionist.update_attributes(params[:nutritionist])\n format.html { redirect_to @nutritionist, notice: 'Nutritionist was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @nutritionist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update!(**args)\n @group_kinds = args[:group_kinds] if args.key?(:group_kinds)\n @json_path = args[:json_path] if args.key?(:json_path)\n @namespaces = args[:namespaces] if args.key?(:namespaces)\n end",
"def update!(**args)\n @description = args[:description] if args.key?(:description)\n @format = args[:format] if args.key?(:format)\n @json_path = args[:json_path] if args.key?(:json_path)\n @name = args[:name] if args.key?(:name)\n @priority = args[:priority] if args.key?(:priority)\n @type = args[:type] if args.key?(:type)\n end",
"def update\n put :update\n end",
"def update!(**args)\n @description = args[:description] if args.key?(:description)\n @etag = args[:etag] if args.key?(:etag)\n @multi_cluster_routing_use_any = args[:multi_cluster_routing_use_any] if args.key?(:multi_cluster_routing_use_any)\n @name = args[:name] if args.key?(:name)\n @single_cluster_routing = args[:single_cluster_routing] if args.key?(:single_cluster_routing)\n end",
"def update\n authorize! :index, @user, :message => 'Not authorized as an administrator.'\n\n @specialist = Specialist.find(params[:id])\n\n respond_to do |format|\n if @specialist.update_attributes(params[:specialist])\n format.html { redirect_to @specialist, notice: 'Specialist was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @specialist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @kristine_toy.update(kristine_toy_params)\n format.html { redirect_to @kristine_toy, notice: 'Kristine toy was successfully updated.' }\n format.json { render :show, status: :ok, location: @kristine_toy }\n else\n format.html { render :edit }\n format.json { render json: @kristine_toy.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update(*args)\n put(*args)\n end",
"def update(*args)\n put(*args)\n end",
"def create_method\n :put_json\n end",
"def update\n @person = Person.find(params[:id])\n\n populate_attributes(@person, params[:person])\n respond_to do |format|\n \n if @person.save && @person.identifiable_entries.each(&:save!)\n format.html { redirect_to @person, notice: 'Person was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @person.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @specialist.update(specialist_params)\n format.html { redirect_to [:admins, @specialist], notice: 'Specialist was successfully updated.' }\n format.json { render :show, status: :ok, location: [:admins, @specialist] }\n else\n format.html { render :edit }\n format.json { render json: @specialist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @interesting = Interesting.find(params[:id])\n\n respond_to do |format|\n if @interesting.update_attributes(params[:interesting])\n format.html { redirect_to @interesting, notice: 'Interesting was successfully updated.' }\n format.json { head :ok }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @interesting.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @artist.update(artist_params)\n format.html { redirect_to organization_artist_url(organization_id: @organization.id, id: @artist.id), notice: 'Artist was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: 'edit' }\n format.json { render json: @artist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @interest = Interest.find(params[:id])\n \n respond_to do |format|\n if @interest.update_attributes(params[:interest])\n format.json { head :ok }\n else\n format.json { render :json => @interest.errors,\n :status => :unprocessable_entity }\n end\n end\n end",
"def update_gist\n @gist[\"list.json\"] = JSON.pretty_generate @things\n @gists.save @gist\n end",
"def patch(path, **args); end",
"def update\n @fred = Fred.find(params[:id])\n\n respond_to do |format|\n if @fred.update_attributes(params[:fred])\n format.html { redirect_to @fred, notice: 'Fred was successfully updated.' }\n format.json { head :ok }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @fred.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @microplst = Microplst.find(params[:id])\n\n respond_to do |format|\n if @microplst.update_attributes(params[:microplst])\n format.html { redirect_to @microplst, notice: 'Microplst was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @microplst.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @artist.update(artist_params)\n format.html { redirect_to @artist, notice: \"Artist was successfully updated.\" }\n format.json { render :show, status: :ok, location: @artist }\n else\n format.html { render :edit, status: :unprocessable_entity }\n format.json { render json: @artist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @artist.update(artist_params)\n format.html { redirect_to @artist, notice: 'Artist was successfully updated.' }\n format.json { render :show, status: :ok, location: @artist }\n else\n format.html { render :edit }\n format.json { render json: @artist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @artist.update(artist_params)\n format.html { redirect_to @artist, notice: 'Artist was successfully updated.' }\n format.json { render :show, status: :ok, location: @artist }\n else\n format.html { render :edit }\n format.json { render json: @artist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @artist.update(artist_params)\n format.html { redirect_to @artist, notice: 'Artist was successfully updated.' }\n format.json { render :show, status: :ok, location: @artist }\n else\n format.html { render :edit }\n format.json { render json: @artist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @artist.update(artist_params)\n format.html { redirect_to @artist, notice: 'Artist was successfully updated.' }\n format.json { render :show, status: :ok, location: @artist }\n else\n format.html { render :edit }\n format.json { render json: @artist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @person = Person.find(params[:id]) \n respond_to do |format|\n if @person.update(person_params)\n format.json { render json: @person, status: :ok }\n else\n format.json { render json: @person.errors, status: :unprocessable_entity }\n end\n end\n end",
"def api_patch(path, data = {})\n api_request(:patch, path, :data => data)\n end",
"def update\n #Finding the specific chore where the id matches the one we pass in with the body\n @v1_chore = Chore.where(id: params[:id]).first\n #Here we're checking if we have user_id in our body, and if we do, we'll change the selected chore's properties\n #with the parameters of the body, we go through the specific group to our specific chore with the path\n if v1_chore_params[:user_id]\n @v1_chore.user_id = params[:user_id]\n @v1_chore.assigned = true\n if @v1_chore.save\n render :show, status: :ok\n end\n else\n render json: @v1_chore.errors, status: :unprocessable_entity\n end\n end",
"def update\n @shelter = Shelter.find(params[:id])\n\n if @shelter.update(shelter_params)\n head :no_content\n else\n render json: @shelter.errors, status: :unprocessable_entity\n end\n end",
"def update\n @artist_tag = ArtistTag.find(params[:id])\n\n respond_to do |format|\n if @artist_tag.update_attributes(params[:artist_tag])\n format.html { redirect_to @artist_tag, notice: 'Artist tag was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @artist_tag.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update!(**args)\n @description = args[:description] if args.key?(:description)\n @fingerprint = args[:fingerprint] if args.key?(:fingerprint)\n @id = args[:id] if args.key?(:id)\n @insert_time = args[:insert_time] if args.key?(:insert_time)\n @labels = args[:labels] if args.key?(:labels)\n @manifest = args[:manifest] if args.key?(:manifest)\n @name = args[:name] if args.key?(:name)\n @operation = args[:operation] if args.key?(:operation)\n @self_link = args[:self_link] if args.key?(:self_link)\n @target = args[:target] if args.key?(:target)\n @update = args[:update] if args.key?(:update)\n end",
"def update\n recipe.update(recipe_params)\n render json: recipe\n end",
"def update\n respond_to do |format|\n if @artist.update(artist_params)\n\n format.html { redirect_to @artist, notice: 'Artist was successfully updated.' }\n\n format.json { render :show, status: :ok, location: @artist }\n\n else\n\n format.html { render :edit }\n\n format.json { render json: @artist.errors, status: :unprocessable_entity }\n\n end\n end\n end",
"def update \n sneaker = find_sneaker\n # update! exceptions will be handled by the rescue_from ActiveRecord::RecordInvalid code\n sneaker.update(sneaker_params)\n render json: sneaker\n end",
"def update_specialist\n\t\t # update the speciality name ans sub speciality\n\t\tif @user.update_attributes(specialist_name:params[:specialist_name],sub_speciality_name:params[:sub_speciality_name])\n\t\t # response to the JSON\n\t\t\trender json: { success: true,message: \"Specialist Successfully Updated.\", response: {specialist_name: @user.specialist_name.as_json,sub_speciality_name:@user.sub_speciality_name.as_json } },:status=>200\n\t else\n\t render :json=> { success: false, message: @user.errors },:status=> 203\n\t end\t\n\tend",
"def update(id, attributes)\n # attributes = {name: 'chocolate and peanuts', calories: 10}\nend",
"def update_plant\n @plant.deleted = false\n\n respond_to do |format|\n if set_attributes_from_filemaker(@plant)\n format.json do\n render status: :created,\n json: {\n id: @plant.id,\n botanical_name: @plant.botanical_name,\n alternative_names: @plant.alternative_names,\n updated_at: @plant.updated_at,\n visible: (!@plant.deleted).to_s\n }\n end\n else\n format.json do\n render json: @plant.errors, status: :unprocessable_entity\n end\n end\n end\n end",
"def update!(**args)\n @id = args[:id] if args.key?(:id)\n @kind = args[:kind] if args.key?(:kind)\n @name = args[:name] if args.key?(:name)\n @starred = args[:starred] if args.key?(:starred)\n @type = args[:type] if args.key?(:type)\n end",
"def update\n @person.update_attributes(params[:person])\n respond_with(@person)\n end",
"def update!(**args)\n @metadata = args[:metadata] if args.key?(:metadata)\n @pomeroy_id = args[:pomeroy_id] if args.key?(:pomeroy_id)\n @trust_level = args[:trust_level] if args.key?(:trust_level)\n end",
"def update\n respond_to do |format|\n if @lunch.update(lunch_params)\n format.html { redirect_to @lunch, notice: 'Lunch was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: 'edit' }\n format.json { render json: @lunch.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @fridge = Fridge.find(params[:id])\n\n respond_to do |format|\n if @fridge.update_attributes(params[:fridge])\n format.html { redirect_to @fridge, notice: 'Fridge was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @fridge.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @artist.update(artist_params)\n format.html { redirect_to @artist, notice: t(:update_notice) }\n format.json { render :show, status: :ok, location: @artist }\n else\n format.html { render :edit }\n format.json { render json: @artist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def patch\n end",
"def update\n @forest = Forest.find(params[:id])\n\n respond_to do |format|\n if @forest.update_attributes(params[:forest])\n format.html { redirect_to @forest, notice: 'Forest was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @forest.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n if @person.seat\n render json: {errors: 'Cannot update a seated person'}, status: 422\n else\n @person.update person_params\n render json: @person\n end\n end",
"def update\n @artist = Artist.find(params[:id])\n\n respond_to do |format|\n if @artist.update_attributes(params[:artist])\n format.html { redirect_to(@artist, :notice => 'Artist was successfully updated.') }\n format.xml { head :ok }\n else\n format.html { render :action => \"edit\" }\n format.xml { render :xml => @artist.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @resist.update(resist_params)\n format.html { redirect_to @resist, notice: 'Resist was successfully updated.' }\n format.json { render :show, status: :ok, location: @resist }\n else\n format.html { render :edit }\n format.json { render json: @resist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @serving.update(serving_params)\n format.html { redirect_to @serving, notice: 'Serving was successfully updated.' }\n format.json { render :show, status: :ok, location: @serving }\n else\n format.html { render :edit }\n format.json { render json: @serving.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update!(**args)\n @allergen_absent = args[:allergen_absent] if args.key?(:allergen_absent)\n @allergen_present = args[:allergen_present] if args.key?(:allergen_present)\n @calories = args[:calories] if args.key?(:calories)\n @ingredients = args[:ingredients] if args.key?(:ingredients)\n @media = args[:media] if args.key?(:media)\n @name_info = args[:name_info] if args.key?(:name_info)\n @nutrition_facts = args[:nutrition_facts] if args.key?(:nutrition_facts)\n @portion_size = args[:portion_size] if args.key?(:portion_size)\n @preparation_methods = args[:preparation_methods] if args.key?(:preparation_methods)\n @price = args[:price] if args.key?(:price)\n @restriction = args[:restriction] if args.key?(:restriction)\n @serves_num_people = args[:serves_num_people] if args.key?(:serves_num_people)\n @spiciness = args[:spiciness] if args.key?(:spiciness)\n end",
"def update!(**args)\n @child_link = args[:child_link] if args.key?(:child_link)\n @created = args[:created] if args.key?(:created)\n @id = args[:id] if args.key?(:id)\n @kind = args[:kind] if args.key?(:kind)\n @name = args[:name] if args.key?(:name)\n @permissions = args[:permissions] if args.key?(:permissions)\n @self_link = args[:self_link] if args.key?(:self_link)\n @starred = args[:starred] if args.key?(:starred)\n @updated = args[:updated] if args.key?(:updated)\n end",
"def update\n @stable = Stable.find(params[:id])\n\n respond_to do |format|\n if @stable.update_attributes(params[:stable])\n format.html { redirect_to @stable, notice: 'Stable was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @stable.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n head :ok\n end",
"def update\n @my_ministry = MyMinistry.find(params[:id])\n\n respond_to do |format|\n if @my_ministry.update_attributes(params[:my_ministry])\n format.html { redirect_to @my_ministry, notice: 'My ministry was successfully updated.' }\n format.json { head :ok }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @my_ministry.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @gist.update(gist_params)\n format.html { redirect_to @gist, notice: 'Gist was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: 'edit' }\n format.json { render json: @gist.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update_user_for_tenant(args = {}) \n id = args['id']\n temp_path = \"/tenants.json/{tenantId}/users/{userId}\"\n path = temp_path\nargs.keys.each do |key|\n if (key == \"tenantId\")\n args.delete(key)\n path = temp_path.gsub(\"{#{key}}\", id)\n end\nend\n puts \" PATH : #{path}\"\n put(path, args)\nend",
"def update\n respond_to do |format|\n if @thirtyfife.update(thirtyfife_params)\n format.html { redirect_to @thirtyfife, notice: 'Thirtyfive was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: 'edit' }\n format.json { render json: @thirtyfife.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @shelter.update(shelter_params)\n format.html { redirect_to @shelter, notice: 'Shelter was successfully updated.' }\n format.json { render :show, status: :ok, location: @shelter }\n else\n format.html { render :edit }\n format.json { render json: @shelter.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update!(**args)\n @api_version = args[:api_version] if args.key?(:api_version)\n @kind = args[:kind] if args.key?(:kind)\n @metadata = args[:metadata] if args.key?(:metadata)\n @spec = args[:spec] if args.key?(:spec)\n end",
"def update\n @specialty = Specialty.find(params[:id])\n\n if @specialty.update(specialty_params)\n head :no_content\n else\n render json: @specialty.errors, status: :unprocessable_entity\n end\n end"
] |
[
"0.6600802",
"0.64254564",
"0.62123835",
"0.6177276",
"0.59769326",
"0.5938455",
"0.5935275",
"0.58813494",
"0.58594966",
"0.58484775",
"0.58376294",
"0.5820418",
"0.5815003",
"0.57989615",
"0.57927454",
"0.5791307",
"0.5763348",
"0.5754602",
"0.5751635",
"0.57411397",
"0.5734432",
"0.5726046",
"0.5715198",
"0.5714646",
"0.56793404",
"0.56694376",
"0.5666991",
"0.56636536",
"0.56380665",
"0.5620303",
"0.5612868",
"0.5600077",
"0.5600077",
"0.55945545",
"0.55898184",
"0.5585784",
"0.55843514",
"0.55788964",
"0.5575984",
"0.55679643",
"0.5567952",
"0.5565402",
"0.5565249",
"0.5565128",
"0.55646574",
"0.55629295",
"0.5561243",
"0.5561243",
"0.5561219",
"0.55583185",
"0.5556083",
"0.55500305",
"0.55370575",
"0.55369616",
"0.5525714",
"0.5517932",
"0.5516797",
"0.55146605",
"0.5510982",
"0.5510445",
"0.5510445",
"0.5510445",
"0.5510445",
"0.5506743",
"0.55054116",
"0.5504094",
"0.5496086",
"0.5494446",
"0.5491526",
"0.5490208",
"0.54897505",
"0.54882306",
"0.5486449",
"0.547914",
"0.54774797",
"0.54721373",
"0.54671013",
"0.5465714",
"0.5463492",
"0.54632986",
"0.5458031",
"0.5437964",
"0.543459",
"0.5432048",
"0.54271793",
"0.5426214",
"0.54254967",
"0.5425091",
"0.54249865",
"0.5423067",
"0.541888",
"0.54158175",
"0.5414226",
"0.54101413",
"0.54057497",
"0.5401122",
"0.53991187",
"0.53978455"
] |
0.67339206
|
1
|
DELETE /therapists/1 DELETE /therapists/1.json
|
def destroy
@therapist.destroy
respond_to do |format|
format.html { redirect_to therapists_url, notice: 'Therapist was successfully destroyed.' }
format.json { head :no_content }
end
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def destroy\n @therapist = Therapist.find(params[:id])\n @therapist.destroy\n\n respond_to do |format|\n format.html { redirect_to therapists_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @uchronist = Uchronist.find(params[:id])\n @uchronist.destroy\n\n respond_to do |format|\n format.html { redirect_to uchronists_url }\n format.json { head :no_content }\n end\n end",
"def delete_tenant_circle(args = {}) \n delete(\"/tenantcircles.json/#{args[:circleId]}\", args)\nend",
"def delete\n client.delete(\"/#{id}\")\n end",
"def destroy\n @pist.destroy\n respond_to do |format|\n format.html { redirect_to pists_url, notice: 'Pist was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def delete_aos_version(args = {}) \n delete(\"/aosversions.json/#{args[:aosVersionId]}\", args)\nend",
"def destroy\n @forest = Forest.find(params[:id])\n @forest.destroy\n\n respond_to do |format|\n format.html { redirect_to forests_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @forest.destroy\n respond_to do |format|\n format.html { redirect_to forests_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @odist.destroy\n respond_to do |format|\n format.html { redirect_to odists_url, notice: 'Odist was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @hydrant = Hydrant.find(params[:id])\n @hydrant.destroy\n\n respond_to do |format|\n format.html { redirect_to(hydrants_url) }\n format.xml { head :ok }\n end\n end",
"def destroy\n @gist.destroy\n respond_to do |format|\n format.html { redirect_to gists_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @stundent = Stundent.find(params[:id])\n @stundent.destroy\n\n respond_to do |format|\n format.html { redirect_to stundents_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @kota_stone.destroy\n respond_to do |format|\n format.html { redirect_to kota_stones_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @botanist.destroy\n respond_to do |format|\n format.html { redirect_to botanists_url, notice: 'Botanist was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def delete_json(path)\n url = [base_url, path].join\n resp = HTTParty.delete(url, headers: standard_headers)\n parse_json(url, resp)\n end",
"def destroy\n @orphan.destroy\n respond_to do |format|\n format.html { redirect_to orphans_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @asthenium.destroy\n respond_to do |format|\n format.html { redirect_to asthenia_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @herald.destroy\n respond_to do |format|\n format.html { redirect_to heralds_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @shelf.destroy\n\n respond_to do |format|\n format.html { redirect_to shelves_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @thirtythree.destroy\n respond_to do |format|\n format.html { redirect_to thirtythrees_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @shelf.destroy\n respond_to do |format|\n format.html { redirect_to shelves_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @hood.destroy\n respond_to do |format|\n format.html { redirect_to hoods_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n return if @name.nil?\n delete_rest \"extra/#{@name}\"\n end",
"def destroy\n @nutritionist = Nutritionist.find(params[:id])\n @nutritionist.destroy\n\n respond_to do |format|\n format.html { redirect_to nutritionists_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @father = Father.find(params[:id])\n @father.destroy\n\n respond_to do |format|\n format.html { redirect_to fathers_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @artist = Artist.find(params[:id])\n @artist.destroy\n\n respond_to do |format|\n format.html { redirect_to root_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @thirtyone.destroy\n respond_to do |format|\n format.html { redirect_to thirtyones_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @thirtyfife.destroy\n respond_to do |format|\n format.html { redirect_to thirtyfives_url }\n format.json { head :no_content }\n end\n end",
"def delete\n render json: Alien.delete(params[\"id\"])\n end",
"def destroy\n @flat.destroy\n\n respond_to do |format|\n format.html { redirect_to flats_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @thermostat.destroy\n respond_to do |format|\n format.html { redirect_to thermostats_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @attri = Attri.find(params[:id])\n @attri.destroy\n\n respond_to do |format|\n format.html { redirect_to attris_url }\n format.json { head :ok }\n end\n end",
"def destroy\n @artist = Artist.find(params[:id])\n @artist.destroy\n\n respond_to do |format|\n format.html { redirect_to artists_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @artist = Artist.find(params[:id])\n @artist.destroy\n\n respond_to do |format|\n format.html { redirect_to artists_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @kristine_toy.destroy\n respond_to do |format|\n format.html { redirect_to kristine_toys_url, notice: 'Kristine toy was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @three.destroy\n respond_to do |format|\n format.html { redirect_to threes_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @simple_chore.destroy\n respond_to do |format|\n format.html { redirect_to simple_chores_url, notice: 'Simple chore was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @dish_mixture.destroy\n respond_to do |format|\n format.html { redirect_to dish_mixtures_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @v1_chore = Chore.where(id: params[:id])\n if @v1_chore.destroy\n head(:ok)\n else\n head(:unprocessable_entity)\n end\n end",
"def destroy\n @capthurit.destroy\n respond_to do |format|\n format.html { redirect_to capthurits_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @stone = Stone.find(params[:id])\n @stone.destroy\n\n respond_to do |format|\n format.html { redirect_to stones_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n relationship = Relationships.find(params[:id])\n relationship.destroy\n head 204\n end",
"def destroy\n official = Official.find(params[:id])\n official.destroy\n head 204\n end",
"def destroy\n @shelf.destroy\n respond_to do |format|\n format.html { redirect_to house_shelves_path(params[:house_id]), notice: 'Shelf was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @homily.destroy\n respond_to do |format|\n format.html { redirect_to homilies_url, notice: 'Homily was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @stash = Stash.find(params[:id])\n @stash.destroy\n\n respond_to do |format|\n format.html { redirect_to stashes_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @specialist.destroy\n respond_to do |format|\n format.html { redirect_to admins_specialists_url, notice: 'Specialist was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @specialist.destroy\n respond_to do |format|\n format.html { redirect_to specialists_url, notice: 'Specialist was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @hit = Hit.find(params[:id])\n @hit.reactions.destroy_all\n @hit.destroy\n\n respond_to do |format|\n format.html { redirect_to hits_url }\n format.json { head :no_content }\n end\n end",
"def test_del\n header 'Content-Type', 'application/json'\n\n data = File.read 'sample-traces/0.json'\n post('/traces', data, 'CONTENT_TYPE': 'application/json')\n\n id = last_response.body\n\n delete \"/traces/#{id}\"\n assert last_response.ok?\n\n get \"/traces/#{id}\"\n\n contents = JSON.parse last_response.body\n assert_kind_of(Hash, contents, 'Response contents is not a hash')\n assert contents.key? 'description'\n assert(!last_response.ok?)\n end",
"def destroy\n @club_path = ClubPath.find(params[:id])\n @club_path.destroy\n\n respond_to do |format|\n format.html { redirect_to club_paths_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @teather.destroy\n respond_to do |format|\n format.html { redirect_to teathers_url, notice: 'Teather was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @tantosha.destroy\n # @tantosha.delete\n respond_to do |format|\n format.html { redirect_to tantoshas_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @shelf = Shelf.find(params[:id])\n @shelf.destroy\n\n respond_to do |format|\n format.html { redirect_to(shelves_url) }\n format.xml { head :ok }\n end\n end",
"def destroy\n @admin_artist = Artist.find(params[:id])\n @admin_artist.destroy\n\n respond_to do |format|\n format.html { redirect_to admin_artists_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @distro.destroy\n respond_to do |format|\n format.html { redirect_to distros_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @ingredients_nutrient.destroy\n respond_to do |format|\n format.html { redirect_to ingredients_nutrients_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @bunny.destroy\n respond_to do |format|\n format.html { redirect_to bunnies_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @tangent.destroy\n respond_to do |format|\n format.html { redirect_to tangents_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @ruby.destroy\n respond_to do |format|\n format.html { redirect_to rubies_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @gist = Gist.find(params[:id])\n @gist.destroy\n\n respond_to do |format|\n format.html { redirect_to gists_url }\n format.json { head :no_content }\n format.js\n end\n end",
"def unfollow_artist\n relationship_to_delete = Relationship.where(follower_id: params[:follower_id], followed_id: params[:followed_id])\n relationship_to_delete[0].destroy\n relationships = Relationship.where(followed_id: params[:followed_id])\n render json: relationships\n end",
"def destroy\n @animal.destroy\n respond_to do |format|\n format.html { redirect(person_animals_url, :delete, 'animal') }\n format.json { head :no_content }\n end\n end",
"def delete path\n make_request(path, \"delete\", {})\n end",
"def destroy\n animal = Animal.find(params[:id])\n animal.destroy\n head 204\n end",
"def destroy\n @fnss_simple_kary_tree.destroy\n respond_to do |format|\n format.html { redirect_to fnss_simple_kary_trees_url, notice: 'Fnss simple kary tree was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @stanza = Stanza.find_by_no(params[:id])\n @stanza.destroy\n\n respond_to do |format|\n format.html { redirect_to stanzas_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @order_yurist.destroy\n respond_to do |format|\n format.html { redirect_to order_yurists_url, notice: 'Order yurist was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @society.destroy\n respond_to do |format|\n format.html { redirect_to societies_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @dis_duplicate_therapy.destroy\n respond_to do |format|\n format.html { redirect_to dis_duplicate_therapies_url, notice: 'Dis duplicate therapy was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @brother = Brother.find(params[:id])\n @brother.destroy\n\n respond_to do |format|\n format.html { redirect_to brothers_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @click_thru = ClickThru.find(params[:id])\n @click_thru.destroy\n\n respond_to do |format|\n format.html { redirect_to click_thrus_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @bustour.destroy\n respond_to do |format|\n format.html { redirect_to bustours_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @json.destroy\n respond_to do |format|\n format.html { redirect_to jsons_url, notice: 'Json was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @json.destroy\n respond_to do |format|\n format.html { redirect_to jsons_url, notice: 'Json was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def deleteFlatpack( flatpack_id)\n params = Hash.new\n params['flatpack_id'] = flatpack_id\n return doCurl(\"delete\",\"/flatpack\",params)\n end",
"def destroy\n @thuctap.destroy\n respond_to do |format|\n format.html { redirect_to thuctaps_url, notice: 'Thuctap was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @distro = Distro.find(params[:id])\n @distro.destroy\n\n respond_to do |format|\n format.html { redirect_to distros_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @gather = Gather.find(params[:id])\n @gather.destroy\n\n respond_to do |format|\n format.html { redirect_to gathers_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @ref_sanatorium = Ref::Sanatorium.find(params[:id])\n @ref_sanatorium.destroy\n\n respond_to do |format|\n format.html { redirect_to ref_sanatoria_url }\n format.json { head :ok }\n end\n end",
"def destroy\n @consensu = Consensu.find(params[:id])\n @consensu.destroy\n\n respond_to do |format|\n format.html { redirect_to consensus_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @monkey = Monkey.find(params[:id])\n @monkey.destroy\n\n respond_to do |format|\n format.html { redirect_to monkeys_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @hoge = Hoge.find(params[:id])\n @hoge.destroy\n\n respond_to do |format|\n format.html { redirect_to hoges_url }\n format.json { head :ok }\n end\n end",
"def destroy\n @interest = Interest.find(params[:id])\n @interest.destroy\n\n respond_to do |format|\n format.json { head :ok }\n end \n end",
"def destroy\n @thirtytwo.destroy\n respond_to do |format|\n format.html { redirect_to thirtytwos_url }\n format.json { head :no_content }\n end\n end",
"def delete\n client.delete(url)\n @deleted = true\nend",
"def destroy\n @harvest_trello = HarvestTrello.find(params[:id])\n @harvest_trello.destroy\n\n respond_to do |format|\n format.html { redirect_to harvest_trellos_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @taxinomy = Taxinomy.find(params[:id])\n @taxinomy.destroy\n\n respond_to do |format|\n format.html { redirect_to taxinomies_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @subway.destroy\n respond_to do |format|\n format.html { redirect_to subways_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @herb = Herb.find_by_permalink(params[:id])\n @herb.destroy\n\n respond_to do |format|\n format.html { redirect_to herbs_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @harvest = Harvest.find(params[:id])\n @harvest.destroy\n\n respond_to do |format|\n format.html { redirect_to harvests_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @harvest = Harvest.find(params[:id])\n @harvest.destroy\n\n respond_to do |format|\n format.html { redirect_to harvests_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @foodhamper = Foodhamper.find(params[:id])\n @foodhamper.destroy\n\n respond_to do |format|\n format.html { redirect_to foodhampers_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @artist_medium.destroy\n respond_to do |format|\n format.html { redirect_to artist_mediums_url, notice: 'Artist was successfully deleted.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @persona = Persona.find(params[:id])\n @persona.destroy\n\n respond_to do |format|\n format.json { head :ok }\n end\n \n end",
"def destroy\n @thirtyfour.destroy\n respond_to do |format|\n format.html { redirect_to thirtyfours_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @krish.destroy\n respond_to do |format|\n format.html { redirect_to krishes_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @wash = Wash.find(params[:id])\n @wash.destroy\n\n respond_to do |format|\n format.html { redirect_to washes_url }\n format.json { head :no_content }\n end\n end"
] |
[
"0.7385645",
"0.6731119",
"0.66847366",
"0.6673562",
"0.6646744",
"0.65873194",
"0.658058",
"0.65756166",
"0.6543277",
"0.6516445",
"0.65016747",
"0.6499451",
"0.64946085",
"0.6491436",
"0.6486369",
"0.6471839",
"0.64355314",
"0.6434971",
"0.6432992",
"0.6419995",
"0.6416921",
"0.64132833",
"0.63995767",
"0.63924545",
"0.6388952",
"0.63806146",
"0.6380165",
"0.6373919",
"0.63659185",
"0.63564044",
"0.63557065",
"0.6351028",
"0.63442814",
"0.63442814",
"0.6334122",
"0.63310087",
"0.6326554",
"0.63201344",
"0.63198066",
"0.63197696",
"0.6318282",
"0.6317304",
"0.6314017",
"0.63085556",
"0.63039035",
"0.62993723",
"0.62976277",
"0.62962663",
"0.6291932",
"0.628933",
"0.6288818",
"0.628838",
"0.6284864",
"0.6282065",
"0.6281954",
"0.62816954",
"0.6279343",
"0.6271477",
"0.62654877",
"0.6265117",
"0.62618387",
"0.62601274",
"0.62579095",
"0.6257481",
"0.6252118",
"0.6251739",
"0.62515837",
"0.62506765",
"0.6249351",
"0.624914",
"0.6245702",
"0.62456447",
"0.62434006",
"0.6242961",
"0.6242961",
"0.6239907",
"0.62362206",
"0.62351835",
"0.6233852",
"0.6233578",
"0.62317866",
"0.623152",
"0.6230435",
"0.6223789",
"0.6222624",
"0.6221902",
"0.62194115",
"0.6213699",
"0.6211133",
"0.6210638",
"0.62097955",
"0.62097955",
"0.62083393",
"0.6207735",
"0.620686",
"0.6206044",
"0.6205803",
"0.6205474"
] |
0.7257804
|
3
|
Returns a pair where p[0] is an array containing the top three categories, and p[1] is an array containing therapist ids in descending order of relevance
|
def getCategoriesAndTherapists(query)
query = query.gsub("'", "")
f = open("|python ../nlp/FindTherapistsByQuery.py #{query}")
output = f.read().strip().gsub("\n", ", ").split(", ")
categories = Array.new()
output[0..2].each do |x|
x.sub!('[', '')
x.sub!(']', '')
categories << x
end
therapists = Array.new()
output[3..-1].each do |x|
x.sub!('[', '')
x.sub!(']', '')
therapists << x.to_i
end
return [categories,therapists]
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def top_three_recipes \n a = recipes.sort_by do |i| \n i.rating \n end \n a[-3..-1]\n end",
"def top_three_recipes\n recipes_sorted_by_rating.reverse[0..2]\n end",
"def top_three_recipes\n my_recipes = self.recipe_cards\n my_recipes.sort{|a, b| b.rating <=> a.rating }.slice(0,3)\n end",
"def top_three_recipes\n top_three_cards = recipe_cards.sort_by { |card| card.rating}.reverse.slice(0, 3)\n top_three_cards.map { |c| c.recipe }\n end",
"def top_matches(n=3)\n \n scores = Array.new\n \n @prefs.each_pair do |key,value|\n if key != @person\n scores << [@similarity.compute(key),key]\n end\n end\n \n (scores.sort!.reverse!)[0..n]\n \n end",
"def top_3\n freq_hash = each_with_object(Hash.new(0)) { |v, h| h[v] += 1 }\n freq_hash.sort_by { |k, v| -v }.first(3).map(&:first)\n end",
"def top_three_recipes\n self.recipes.sort_by {|info| info.rating}.pop(3)\n end",
"def top_matches(item_prefs, item, number=5)\n scores = []\n\n item_prefs.keys.each do |other|\n if other != item\n similarity = distance_similarity(item_prefs, item, other)\n\n if similarity > 0\n scores << [similarity, other]\n end\n end\n end\n\n scores = scores.sort_by { |score| -score[0] }\n\n return scores[0, number]\n end",
"def top_three_recipes\n\t\trecipes.select { |recipe| recipe.rating }.sort_by { |recipe| recipe.rating }.last(3)\n\tend",
"def top_three_recipes\n top_three = RecipeCard.all.select {|atr| atr.user == self}\n top_three.sort_by {|atr| atr.rating}.reverse\n top_three[0..2]\n end",
"def top_three_recipes\n self.find_user_recipe_cards.sort_by{|rcard| rcard.rating}.reverse![0..2]\n end",
"def scored_categories(doc, cutoff = 0.30, &block)\n icutoff = (@items.size * cutoff).round\n carry = proximity_array_for_content(doc, &block)\n carry = carry[0..icutoff - 1]\n votes = Hash.new(0.0)\n carry.each do |pair|\n @items[pair[0]].categories.each do |category|\n votes[category] += pair[1]\n end\n end\n\n votes.sort_by { |_, score| score }\n end",
"def top_three_recipes\n recipe_cards.max_by(3) {|recipe_cards| recipe_cards.rating}\n end",
"def calculate_similar_items(preferences, limit = 10)\n result = {} \n # Invert the preference matrix to be item-centric \n item_preferences = transform_preferences(preferences) \n item_preferences.keys.each do |item|\n # Find the most similar items to this one \n scores = top_matches(item_preferences, item, limit) \n result[item] = scores\n end\n return result\nend",
"def get_keywords(top_n)\n @weighted_keywords.sort_by {|_key, value| value}.reverse[1..top_n]\n end",
"def filter_articles(articles)\n include PopularitySearch\n #this isn't fair to more recent articles\n popularity_cutoff = 300\n articles.each do |article|\n article[:twitter_pop] = twitter_popularity(article[:url])\n end\n articles.select do |article|\n article[:twitter_pop] > popularity_cutoff\n end\n articles = articles.sort_by{|article| article[:twitter_pop]}.reverse\n return articles[0..2] #only pick the top 3 if there's more than 3\nend",
"def top_three_memes\n # Get top 3 rarity total scores\n top_rarities = Rarity.order('total_score DESC').limit(3)\n top_rarities.map do |rarity|\n Meme.find(rarity.meme_id)\n end\n end",
"def song_sorter(arr1, arr2, arr3)\n sorted_songs = {}\n\n (arr1 + arr2 + arr3).each do |classifier|\n song = classifier.song\n sorted_songs[song] ? sorted_songs[song] += 1 : sorted_songs[song] = 1\n end\n\n result = {\n first_recommendation: [],\n second_recommendation: [],\n third_recommendation: []\n }\n\n sorted_songs.each do |song, value|\n # TODO: Come back to this when you have async figured out\n # RecommendedSong.create(song_id: song.id,\n # recommendation_id: @recommendation.id)\n case value\n when 3\n result[:first_recommendation] << song\n when 2\n result[:second_recommendation] << song\n when 1\n result[:third_recommendation] << song\n end\n end\n\n result\n end",
"def top_three_recipes\n sorted_recipe_cards = self.recipe_cards.sort do |recipe_card|\n recipe_card.rating end\n if sorted_recipe_cards.length < 3\n sorted_recipe_cards\n else\n sorted_recipe_cards[-3,3]\n end\n end",
"def topMatches( prefs, person, n=5, scorefunc = :sim_pearson )\n scores = []\n for other in prefs.keys\n if scorefunc == :sim_pearson\n scores << [ sim_pearson(prefs,person,other), other] if other != person\n else\n scores << [sim_distance(prefs,person,other), other] if other != person\n end\n end\n return scores.sort.reverse.slice(0,n)\n end",
"def top_matches(preferences, person, limit = 5)\n scores = preferences.map {|pref| [sim_pearson(preferences, person, pref[0]), pref[0]] unless pref[0] == person}.compact\n scores.sort! {|a,b| b[0] <=> a[0]}\n return scores[0...limit]\nend",
"def topMatches( prefs, person, n=5, scorefunc = :sim_pearson )\n scores = []\n for other in prefs.keys\n if scorefunc == :sim_pearson\n scores << [ sim_pearson(prefs,person,other), other] if other != person\n else\n scores << [ sim_distance(prefs,person,other), other] if other != person\n end\n end\n return scores.sort.reverse.slice(0,n)\n end",
"def criterions\n objectives.collect(&:criterion).uniq.sort{|a,b| a.category <=> b.category}\n end",
"def match(categories)\n result = []\n for category in categories\n score = score(category)\n result << {value: score, category: category}\n end\n\n highest_value = result.map{|x| x[:value]}.sort.last\n selected_category = nil\n if highest_value > 0\n result.each do |hash|\n if hash[:value] == highest_value\n selected_category = hash[:category]\n end\n end\n end\n\n return selected_category\n\n end",
"def sorted_by_frequency\n @categories.sort { |a, b| b[1] <=> a[1] }\n end",
"def top_readers\n readers_with_duplicates = []\n self.articles.each { |article| readers_with_duplicates << article.readers }\n\n readers_with_duplicates.flatten!\n readers = readers_with_duplicates.uniq\n frequency = Hash.new(0)\n readers_with_duplicates.each { |r| frequency[r] += 1 }\n array = frequency.sort_by { |key, value| value }\n return [array[-1], array[-2], array[-3]]\n\n end",
"def sorted_results(unsorted)\n scores = []\n sorted = []\n\n unsorted.group_by(&:itself).each_pair do |result, results|\n relevance = result_relevance(result, results.length)\n new_index = scores.bsearch_index { |val| val < relevance }\n # if this result has the lowest score, add it to the end\n new_index ||= -1 \n\n scores.insert(new_index, relevance)\n sorted.insert(new_index, result.first)\n end\n sorted\n end",
"def winning_results_frequency_for_each_nominee\n category_id = self.id\n results = Result.where({\"category_id\" => category_id})\n nominee_hash = Hash.new(0)\n\n results.each do |winner| \n winner = winner.loved_id\n if winner != nil\n nominee_hash[winner] +=1\n end\n end\n\n tallied_results = nominee_hash.sort_by{ |key, value| value}.reverse.to_h\n return tallied_results\n end",
"def related_posts\n return [] unless docs.count > 1\n\n highest_freq = tag_freq.values.max\n related_scores = Hash.new(0)\n\n docs.each do |doc|\n doc.data[\"tags\"].each do |tag|\n if self.data[\"tags\"].include?(tag) && doc != self\n cat_freq = tag_freq[tag]\n related_scores[doc] += (1 + highest_freq - cat_freq)\n end\n end\n end\n\n sort_related_posts(related_scores)\n end",
"def display_stories stories\n #Double the score for each title with \"cat\" or \"dog\" in the title.\n cat_dog(stories)\n\n stories.sort_by! { |story| story[:score] }\n\n stories.reverse!\n\n stories.each do |story|\n # \"Title: title, Category: category, Upvotes: score\"\n puts \"Title: #{story[:title]}, Category: #{story[:category]}, Upvotes: (#{story[:score]})\"\n end\nend",
"def top_10 (matches)\n matches.sort_by! { |match| match['score'] }.reverse[:10]\nend",
"def ordered_by_qualifications(candiates)\n # Candidates with the most experience are at the top\n return candiates.sort_by { | obj | [obj[:years_of_experience], obj[:github_points]] }\n # return candiates.sort! { |a, b| b[:years_of_experience] <=> a[:years_of_experience] }\n\n # return candiates.sort_by {|:years_of_experience, :github_points|[ :github_points, :years_of_experience]}\n\n # return candiates.sort { | a, b | }\n # array.sort { |a,b| [ a[1], a[0] ] <=> [ b[1], b[0] ] }\n\n\n # For Candidates that have the same years of experience, \n #they are further sorted by their number of Github points (highest first)\nend",
"def categories_by_vote_count\n\t\tself.given_votes.group(:category_id).count.each_with_object({}) { |(k, v), h| ( h[v] ||= [] ) << self.given_votes.find_by(category_id: k) }\n\tend",
"def top_topics\n topics_with_views = Hash.new\n topics.each do |topic|\n topics_with_views[topic] = getViews(topic)\n end\n topics_with_views = topics_with_views.sort_by { |k, v| v }.reverse\n return topics_with_views\n end",
"def ordered_by_qualifications(candidates)\n myArray = candidates\n #first sort by github points\n myArray.sort! { |a, b| a[:github_points] <=> b[:github_points] }\n #sort by experience\n myArray.sort! { |a, b| a[:years_of_experience] <=> b[:years_of_experience] }\n\n return myArray\nend",
"def tfivetags\n tag_array = self.tags\n sort_array = tag_array.sort!{|a,b| a.reputation_for(:votes) <=> b.reputation_for(:votes)}\n sort_array.reverse!\n topfive = sort_array.take(5)\n topfive\n end",
"def top_ten_list(category)\n # Select all of category from work instances\n work_by_category = Work.all.select { |work| work.category.downcase == \"#{category}\" }\n # Return max by vote count for top 10\n return work_by_category.max_by(10) { |work| work.votes.length }\n end",
"def classify_with_all_result(doc)\n result = []\n return nil unless doc\n\n @categories_count.keys().each do |category|\n # log(P(doc|cat))\n document_category = calc_document_category(doc, category)\n\n # log(P(cat)) = log(@categories_count[cat]) - log( @all_category_num )\n category_probability = calc_category_probability(category)\n\n # log(P(cat|doc)) = log(P(doc|cat)) + log(P(cat))\n category_document_probability = document_category + category_probability\n\n result << [category, category_document_probability]\n end\n result\n end",
"def personal_top_three\n scores.max([scores.length, 3].min)\n end",
"def popularity_list\n @movies.values.sort { |movie1, movie2| popularity(movie2.movie_id) <=> popularity(movie1.movie_id) }.map(&:movie_id)\n end",
"def top_10_sorted\n\t\t\t\t\traw = notable_order_by_cvss_raw\n\t\t\t\t\tdata = Array.new\n\n\t\t\t\t\traw.each do |vuln|\n\t\t\t\t\t\trow = Array.new\n\t\t\t\t\t\tplugin_id = vuln[0]\n\t\t\t\t\t\tcount = vuln[1]\n\n\t\t\t\t\t\tname = scrub_plugin_name(Plugin.find_by_id(plugin_id).plugin_name)\n\n\t\t\t\t\t\trow.push(name)\n\t\t\t\t\t\trow.push(count)\n\t\t\t\t\t\tdata.push(row)\n\t\t\t\t\tend\n\n\t\t\t\t\tdata = data.sort do |a, b|\n\t\t\t\t\t\tb[1] <=> a[1]\n\t\t\t\t\tend\n\n\t\t\t\t\treturn data\n\t\t\t\tend",
"def results\n query = parse\n\n results = []\n positive = query[:positive].to_a\n results << Language.where(name: positive)\n results << Language.where(type: positive)\n results << Language.where(designers: positive)\n\n weights = weight_results results\n\n positive = results.flatten.uniq(&:name).index_by &:name\n\n results = []\n negative = query[:negative].to_a\n\n results << Language.where_not(name: negative).map(&:name)\n results << Language.where_not(type: negative).map(&:name)\n results << Language.where_not(designers: negative).map(&:name)\n\n negative = results.inject(results[0]) {|result, array| result & array }.uniq\n\n final_results = positive.slice(*negative).values\n sort_results set_hits(final_results, weights), weights\n end",
"def most_common_nationalities_list\n nationalities_hash = self.players.group_by{\n |player| player.nationality\n }\n count_hash={}\n nationalities_string = nationalities_hash.map do |nationality, players|\n # puts \"Number of players from \" + nationality + \": \" + players.count.to_s\n count_hash[nationality]=players.count\n end\n count_hash.sort_by {|nationality, num| num}.reverse\n end",
"def k_most_visited input, k\n \n hotels_to_visits = Hash.new\n \n input.each do |user|\n user.each do |hotel_id|\n hotels_to_visits[hotel_id] = (hotels_to_visits[hotel_id] || 0) + 1\n end\n end\n \n sorted_hotels_and_visits = hotels_to_visits.to_a.sort { |a,b| a[1] < b[1] }\n \n return sorted_hotels_and_visits[0..k]\n \nend",
"def sort_by_tag_count_with_random_buckets\n end",
"def most_common_number(array)\n #var\n numbers_and_count = Hash.new 0\n \n array.each do |num|\n \tnumbers_and_count[num] += 1\n end\n #var\n most = numbers_and_count.sort_by { |key, value| value }.reverse\n \n top = Hash[*most.first]\n top.select { |k,v| p k }\n\nend",
"def dominant(arr)\n merge_sort(arr)[-1]\nend",
"def prematch_top_markets\n {\n '6046' => top_markets(%w[1 2 7 59]), # 'Football',\n '54094' => top_markets(%w[52 202 166]), # 'Tennis',\n '687888' => top_markets(%w[]), # 'Horse Racing', *no matches\n '452674' => top_markets(%w[52 762 41 21]), # 'Cricket',\n '530129' => top_markets(%w[1 2 7 59]), # 'Hockey', *assumptions\n '274792' => top_markets(%w[1 2 52 202]), # 'Rugby League',\n '48242' => top_markets(%w[1 2 7 59]), # 'Basketball',\n '131506' => top_markets(%w[1 2 7 59]), # 'American Football', *assumptions\n '687889' => top_markets(%w[]), # 'Golf', *multiple outcomes\n '154919' => top_markets(%w[1 52]), # 'Boxing',\n #'687890' => top_markets(%w[]), # 'E-Games'\n '68789002' => top_markets(%w[52 202 21]), #\"CS:GO\"\n '68789004' => top_markets(%w[52]), #\"King Of Glory\" *no matches\n '68789005' => top_markets(%w[52 1175]), #\"League of Legends\"\n '68789006' => top_markets(%w[52 202 203 204]), #\"Overwatch\"\n '68789007' => top_markets(%w[52]), #\"Rainbow Six\"\n '68789008' => top_markets(%w[52 2]), #\"Rocket League\"\n '68789009' => top_markets(%w[52 2]), #\"StarCraft II\"\n '68789010' => top_markets(%w[52]), #\"Street Fighter V\" *no matches\n '68789011' => top_markets(%w[52]), #\"Warcraft III\"\n '68789001' => top_markets(%w[52 2 202 203 204]), #\"Call of Duty\"\n '68789003' => top_markets(%w[1 52]) #\"Dota 2\n }\n end",
"def select_top_five\n ed1_ed2_merged = @ed1.merge(@ed2)\n top5_words = ed1_ed2_merged.sort_by{|k,v| -v}.first 5\n top5_words = Hash[*top5_words.flatten]\n final_list = Hash.new{}\n # If the typed_word isnt in the dictionary, replace 5th suggestion by typed_word \n if !working_dictionary.has_key?( @typed_name )\n top5_words.each_with_index { |(k,v),index|\n if index == (top5_words.length-1)\n final_list[@typed_name] = v*0.5\n else\n final_list[k] = v\n end\n }\n top5_words = final_list\n end\n @suggestion = top5_words\n return top5_words #Hash of words and corresponding score \n end",
"def classify(tokens)\n max_prob, max_category = -1, -1\n\n if tokens.empty? \n # If the example is empty, find the category with the highest prior probability.\n (0..@num_categories - 1).each do |i|\n prior_prob = get_prior_category_probability(i)\n max_prob, max_category = prior_prob, i if prior_prob > max_prob\n end\n else\n # Otherwise, find the category with the highest posterior probability.\n get_posterior_category_probabilities(tokens).each_with_index do |prob, category|\n max_prob, max_category = prob, category if prob > max_prob\n end\n end\n \n return max_category\n end",
"def top_5\n count = @sentence.blips.group(:body).distinct.count\n percent = count.each {|k, v| count[k] = v / @sentence.blips_count.to_f }\n statistics = percent.sort_by { |k, v| v }.reverse[0..4].flatten.each { |k, v| puts \"#{k}: #{v}\" }\n end",
"def top_3_words(text)\n count = Hash.new { 0 }\n text.scan(/\\w+'*\\w*/) { |word| count[word.downcase] += 1 }\n count.map{|k,v| [-v,k]}.sort.first(3).map(&:last)\nend",
"def most_similar (user1)\n\t\tuser1 = user1.to_s\n\t\tsimilarity_list = Array.new\n\t\t@user_rating_index.each_key do |key|\n\t\t\tunless key == user1 \n\t\t\t\tpair = Array.new()\n\t\t\t\tpair = [similarity(user1, key), key]\n\t\t\t\tsimilarity_list << pair\n\t\t\tend \n\t\tend\n\t\tsimilarity_list= similarity_list.sort\n\t\treturn similarity_list\n\tend",
"def hits_overall\n @categories.inject(0) { |sum, (name, cat)| sum + cat[:hits] }\n end",
"def popularity_list\n @movie_list.values.sort{|a,b| b.popularity <=> a.popularity}\n end",
"def top_categories(options={})\n options = DEFAULT_CATEGORY_OPTIONS.merge(options)\n\n sql = <<-GO\n SELECT tags.name, COUNT(*) number\n FROM users\n INNER JOIN entries\n ON users.id = entries.user_id\n INNER JOIN taggings\n ON entries.id = taggings.taggable_id and 'Entry' = taggings.taggable_type\n INNER JOIN tags\n ON taggings.tag_id = tags.id\n WHERE users.id = #{self.id}\n #{' AND entries.is_private = 0 ' unless options[:include_private] == true}\n GROUP BY tags.name\n ORDER BY number DESC, tags.name ASC\n #{\"limit %d \" % options[:max_rows] unless options[:max_rows] == 0}\n GO\n\n result = connection.execute(sql.gsub(\"\\n\", ' ').squeeze(' '))\n tags = []\n result.each {|row| tags << [row[0], row[1].to_i]}\n tags\n end",
"def sort_searched_tweets(tweets)\n word_scores = []\n scores = tweets.map do |t|\n score = 0\n score, str = score_for_tweet(t)\n word_scores << str\n score\n end\n tweets = tweets.map.with_index.sort_by { |t, index| -scores[index] }.map(&:first)\n word_scores = word_scores.map.with_index.sort_by { |t, index| -scores[index] }.map(&:first)\n [tweets, word_scores]\n end",
"def ordered_by_qualifications(candidates)\n candidates.sort_by{ |candidate| [candidate[:years_of_experience], candidate[:github_points]] }\n .reverse\n\nend",
"def top_3_words (text)\n word_hash = Hash.new(0)\n len = text.length\n while (len > 0)\n\n nextIndex = text.index(' ')\n\n if (nextIndex == nil)\n nextIndex = text.length\n else\n nextIndex += 1\n end\n\n word = text.slice!(0, nextIndex).strip.downcase\n word_hash[word] += 1\n\n len = text.length\n\n end\n\n array = word_hash.sort_by{ |item, count| count}.last(3).reverse\n \n return array.map{|item| item[0]}\nend",
"def most_popular_tags(options = {})\n if all_tags = Tag.by_pseud(self).by_type(options[:categories]).canonical\n tags_with_count = {}\n all_tags.uniq.each do |tag|\n tags_with_count[tag] = all_tags.find_all{|t| t == tag}.size\n end\n all_tags = tags_with_count.to_a.sort {|x,y| y.last <=> x.last }\n options[:limit].blank? ? all_tags : all_tags[0..(options[:limit]-1)]\n end\n end",
"def ordered_by_qualifications(candidates)\n ordered_by_qualification = [] #creates an empty array for the candidates\n years_of_experience = candidates.sort { |a, b| [a[:years_of_experience], a[:github_points]] <=> [b[:years_of_experience], b[:github_points]] }\n ordered_by_qualification << years_of_experience.reverse\n return ordered_by_qualification\nend",
"def get_related_hashtags\n hash_matrix = []\n relateds.each do |r|\n #if r.related.numtweets > numtweets\n intersect = numtweets == 0 ? 0 : (r.intersection*100)/numtweets\n #else\n # intersect = r.related.numtweets == 0 ? 0 :(r.intersection*100)/r.related.numtweets\n #end\n hash_matrix << [r.related.name, r.intersection, r.related.numtweets, r.related, intersect]\n end\n hash_matrix.sort_by{|hash| hash[4]}.reverse\n end",
"def top_categories(options={})\n options = DEFAULT_CATEGORY_OPTIONS.merge(options)\n\n sql = <<-GO\n SELECT tags.name, COUNT(*) number \n FROM users \n INNER JOIN entries \n ON users.id = entries.user_id \n INNER JOIN taggings \n ON entries.id = taggings.taggable_id and 'Entry' = taggings.taggable_type \n INNER JOIN tags \n ON taggings.tag_id = tags.id \n WHERE users.id = #{self.id} \n #{' AND entries.is_private = 0 ' unless options[:include_private] == true}\n GROUP BY tags.name \n ORDER BY number DESC, tags.name ASC\n #{\"limit %d \" % options[:max_rows] unless options[:max_rows] == 0}\n GO\n\n result = connection.execute(sql.gsub(\"\\n\", ' ').squeeze(' '))\n tags = []\n result.each {|row| tags << [row[0], row[1].to_i]} \n tags\n end",
"def most_similar(u)\n\t\tmSimilar= {}\n\t\treviews_hash.each {|user, moveis|\tmSimilar[user] = similarity(u,user)} #find the similarity with every other user\n\t\tm = mSimilar.sort_by {|a,b| b}.reverse.transpose[0] #sort according to similarities and then choose the first ten (exclude itself)\n\t\treturn m[1..10]\n\tend",
"def get_recommendations\n result = []\n DIST[@pain_severity].each do |element|\n unique = false\n while !unique\n recommendation = weighted_sample(Recommendation.send(element).suggestable)\n if result.include? recommendation\n recommendation = weighted_sample(Recommendation.send(element).suggestable)\n else\n result << recommendation\n unique = true\n end\n end\n end\n result\n end",
"def findTopMovies(actor, top_number=100) \r\n movie_array = []\r\n\r\n actor.film_actor_hash.each_key {|key| movie_array.push(key)}\r\n\r\n return movie_array.take(top_number)\r\nend",
"def sort_kids(num_groups)\n kids = self.elemental_session_bookings.map(&:kid_id)\n total_kids = kids.count\n group_quorum = 10\n num_groups = 4\n categories = [\"age\"]\n match_points_arr = []\n\t\n\tif group_quorum < total_kids\n\t\t(0...kids.count).each do |i|\n\t\t\t(0...i).each do |j|\n\t\t\t\t match_points_arr.push({:kids[i]=> kids[j], :match_points => get_match_points(categories, kids[i], kids[j]) })\n\t\t\tend\n\t\tend\n\t\treturn get_best_groups(num_groups, match_points_arr.sort_by { |hash| hash[:match_points].reverse! })\n\tend\n end",
"def top_3_words(text)\n # 앞, 뒤에 있는 기호는 무시한다.\n # ' 는 몇개라도 봐준다.\n # 중간에 있는 기호는 모르겠는데 일단 그냥 가자.\n\n # words = {}\n # text.downcase.split.each do |w|\n # w = w.match(/\\w+(?:'*)\\w*/).to_s\n # next if w.empty?\n # words[w] ? words[w] += 1 : words[w] = 1\n # end\n # words.sort_by{|_, v| -v}[0..2].map{|x| x.first }\n\n text.downcase.scan(/\\w+(?:'*)\\w*/).inject(Hash.new(0)){|h, w| h[w] += 1; h}.sort_by{|_, v| -v}[0..2].map{|x| x.first }\n\n # text.downcase.scan(/[\\w']+/).select{|x| /\\w/ =~ x}.group_by{|x| x.downcase}.sort_by{|_, v| -v.count}.first(3).map(&:first)\nend",
"def top_categories\n Category\n .joins(:memes)\n .group(:id)\n .order('COUNT(memes.id) DESC')\n .limit(10)\n end",
"def top_10_sorted_raw\n\t\t\t\t\traw = notable_order_by_cvss_raw\n\n\t\t\t\t\tdata = Array.new\n\n\t\t\t\t\traw.each do |vuln|\n\t\t\t\t\t\trow = Array.new\n\t\t\t\t\t\tplugin_id = vuln[0]\n\t\t\t\t\t\tcount = vuln[1]\n\n\t\t\t\t\t\trow.push(plugin_id)\n\t\t\t\t\t\trow.push(count)\n\t\t\t\t\t\tdata.push(row)\n\t\t\t\t\tend\n\n\t\t\t\t\tdata = data.sort do |a, b|\n\t\t\t\t\t\tb[1] <=> a[1]\n\t\t\t\t\tend\n\n\t\t\t\t\treturn data\n\t\t\t\tend",
"def most_explored_tags(topic, tag_groups = [], limit = 5)\n # Take 5 mostly visited places\n most_visited = self.checkins.by_topic(topic.id).most.all(:limit => 5)\n most_visited_tags = sort_tags(tag_groups, most_visited.collect(&:restaurant)) \n\n # Take 5 mostly reviewed and loved places\n most_loved = self.reviews.by_topic(topic.id).most.all(:limit => 5) \n most_loved_tags = sort_tags(tag_groups, most_loved.collect(&:restaurant))\n \n # Take 5 recently explored places\n recent_explored = self.restaurants.by_topic(topic.id).recent.all(:limit => 5) \n recent_explored_tags = sort_tags(tag_groups, recent_explored)\n \n # Merge all tags (5 + 5 + 5) = 15 (At most)\n # Remove duplicates and pick 5 based on their frequencies\n (most_visited_tags + most_loved_tags + recent_explored_tags).sort_by{|a| a.last}.reverse.collect{|a| a.first}.uniq[0..4]\n end",
"def suggestions(limit=3)\n # todo: cache query results rails cache/redis,etc\n top_order_build_ids = self.related_order_items.group(['products.id', 'related_items.product_id']).order('related_items.product_id DESC').limit(limit).pluck(:id)\n return OrderItem.includes(products: :accessories).where(order_id: top_order_build_ids).map(&:products)\n end",
"def recommendations(url,num=5)\n url = url_host_path(url)\n uid = REDIS[\"id:#{url}\"]\n results = REDIS.zrevrangebyscore(\"recos:#{uid}\", '+inf', '-inf', {:withscores => true} )\n \n recos = results[0...num].map do |x| uid,score = x[0],x[1]\n { :title=>REDIS[\"title:#{uid}\"], :url=>REDIS[\"url:#{uid}\"], :score=>score }\n end\n \n return recos\nend",
"def dominant_octopus(fish_arr)\n merge_sort(fish_arr).last\nend",
"def ordered_by_qualifications(candidates)\n\n return @candidates.sort_by{ |candidate| [candidate[:years_of_experience], candidate[:github_points]] }\n\nend",
"def top_students(grade_hash, number_of_students)\n scores = averages(grade_hash)\n shortlist = scores.sort_by{|k,v| v}.reverse.take(number_of_students)\n names = shortlist.map{|k,v| k}\nend",
"def sorted_triples(array)\n\tnew_array = []\n\tarray.each_slice(3) {|grouped_words| new_array << grouped_words.sort}\n\tnew_array\nend",
"def extract\n # create hash of words with number of their instances in tokens excluding stopwords\n words_hash = Hash.new(0)\n @tokens.each { |w| \n unless w.empty? or stop_words_for(@language)[w]\n words_hash[w] += 1 \n end\n }\n\n idfs_hash = get_idfs(words_hash.keys)\n\n # calculate tf-idf for each word into keywords array\n keywords = []\n max_num = words_hash.values.max.to_f\n words_hash.each do |word, num|\n tf = num / max_num\n idf = idfs_hash[word]\n keywords << [word, (tf * idf).round(5), idf.round(5)]\n end\n\n # return keywords sorted by rank descending\n keywords.sort_by {|word, rank, idf| -rank}\n end",
"def ordered_by_qualifications(candidates)\n\n candidates.sort_by{ |candidate| [-candidate[:years_of_experience], -candidate[:github_points]] }\n\nend",
"def sort_categories_by_price(categorized_items)\n sorted_categories = Array.new\n categorized_items.each{ |sub_array|\n if sub_array.size > 1\n sorted_categories.push(sort_category_by_price(sub_array))\n else\n sorted_categories.push(sub_array)\n end\n }\n sorted_categories\n end",
"def ordered_by_qualifications(candidates)\n sortedCandidates = candidates.sort_by! { |candidate| [candidate[:years_of_experience], candidate[:github_points]] }\n return sortedCandidates.reverse! \nend",
"def most_used_hashtags\n # my_tags = self.hashtags.each do |hashtag|\n # hashtag.posts.select do |post|\n # post.user_id == self \n # end\n # end \n my_tags = []\n self.posts.each do |post|\n my_tags << post.hashtags\n end\n flattened = my_tags.flatten\n tallied = Hash[flattened.group_by{|x|x}.map{|x,y|[x,y.size]}]\n # sorted.each do |hashtag|\n # # need to associate hashtag w it's count\n \n # # gives number of times hashtag appears in sorted list\n # sorted.find_all[hashtag].count \n # end\n tallies_sorted = tallied.sort_by{|x,y| y}.reverse\n my_most = tallies_sorted[0..4]\n most_used = my_most.map do |array|\n array[0]\n end\n most_used\n end",
"def top_k_frequent_elements(list, k)\n return [] if list == []\n hash = Hash.new(0)\n counter = 0;\n list.each do |value|\n hash[value] = [] if hash[value].nil?\n hash[value]+=1\n end\n \n if k > hash.size()\n raise ArgumentError, \"Not Enough Values to Display\"\n end\n \n solution = hash.max(k){ |a, b| b<=>a} \n p solution\n numbers = []\n solution.each do |array|\n numbers << array[0]\n end\n return numbers\nend",
"def top_students(grade_hash, number_of_students)\nsorted_hash = averages(grade_hash).sort_by {|name, grade| grade}\nsorted_hash.reverse!\ntop_array = []\nsorted_hash.each do |k, v|\n top_array << k\nend\ntop_array.take(number_of_students)\nend",
"def top_rentals\n return @badfruit.parse_movies_array(JSON.parse(@badfruit.get_lists_action(\"top_rentals\")))\n end",
"def bestCand(voterOpinion, candList)\r\n\ttopCands = Array.new # []\r\n\ttopScore = nil\r\n\t\r\n\t# for each element in the candList array, calling the current element cand.\r\n\t#\tthe {/} are the same as the do/end shown in the Main section. \r\n\t#\tIncluded here to show the availabailty for different style choices\r\n\tcandList.each { |cand| \r\n\t\t\t\t\tcurrScore = getCandScore(voterOpinion,cand.drop(1))\r\n\t\t\t\t\tif topScore.nil? then\r\n\t\t\t\t\t\ttopScore = currScore\r\n\t\t\t\t\tend\r\n\t\t\t\t\tunless currScore < topScore # essentially if !(currScore < topScore)\r\n\t\t\t\t\t\t\r\n\t\t\t\t\t\tif currScore == topScore\r\n\t\t\t\t\t\t\ttopCands.push(cand[0])\r\n\t\t\t\t\t\telse # must be greater\r\n\t\t\t\t\t\t\ttopScore = currScore\r\n\t\t\t\t\t\t\ttopCands = [cand[0]]\r\n\t\t\t\t\t\tend\r\n\t\t\t\t\tend\r\n\t\t\t\t}\r\n\t\t\t\t\r\n\treturn topCands\r\nend",
"def top_three_rated_afterthoughts\n self.afterthoughts_by_rating.first(3)\n end",
"def extract_tags(frequently_reviewed_restaurants, tag_group)\n tags = frequently_reviewed_restaurants.collect{|r| r.first.tags_belongs_to(tag_group)}.flatten\n tag_usages_map = {}\n tags.each do |tag|\n tag_usages_map[tag] ||= 0\n tag_usages_map[tag] += 1\n end\n\n Tag.find_all_by_name(tag_usages_map.sort{|v1, v2| v2.last <=> v1.last}.collect{|t| t.first})\n end",
"def most_popular_tags(options = {})\n all_tags = []\n if options[:categories].blank?\n all_tags = self.tags + self.bookmark_tags\n else\n type_tags = []\n options[:categories].each do |type_name|\n type_tags << type_name.constantize.all\n end\n all_tags = [self.tags + self.bookmark_tags].flatten & type_tags.flatten\n end\n tags_with_count = {}\n all_tags.uniq.each do |tag|\n tags_with_count[tag] = all_tags.find_all{|t| t == tag}.size\n end\n all_tags = tags_with_count.to_a.sort {|x,y| y.last <=> x.last }\n popular_tags = options[:limit].blank? ? all_tags.collect {|pair| pair.first} : all_tags.collect {|pair| pair.first}[0..(options[:limit]-1)]\n end",
"def applicable_category_filters\n return {} if current_category_level > 3\n facet_name = ProductSearcher::CATEGORY_LEVEL_FACET_MAP[current_category_level + 1]\n facet = self.facets[facet_name.to_sym]\n facet_id_and_count_to_model_and_count(facet, Category).sort{|a,b| b[1] <=> a[1]}[0..19]\n end",
"def categories\n Category.position_sorted\n end",
"def suggestions\n suggest_response['suggest'].values.first[q]['suggestions']\n .uniq { |s| s['term'].downcase }\n .sort_by { |s| (-s['weight']) * 100 + s['term'].length }\n .take(5)\n end",
"def top_students(grade_hash, number_of_students)\n outArray = []\n grade_hash.each do |name, scores|\n sum, n = 0, 0\n scores.each do |x|\n n += 1\n sum += x\n end\n outArray.push([sum/n, name])\n end\n final_answer = []\n outArray.sort.reverse[0...number_of_students].each do |grade,name|\n final_answer.push(name)\n end\n return final_answer\nend",
"def related_posts(posts)\n return [] unless posts.size > 1\n highest_freq = category_freq(posts).values.max\n related_scores = Hash.new(0)\n posts.each do |post|\n post.categories.each do |category|\n if self.categories.include?(category) && post != self\n cat_freq = category_freq(posts)[category]\n related_scores[post] += (1+highest_freq-cat_freq)\n end\n end\n end\n\n sort_related_posts(related_scores)\n end",
"def extract_categories cats\n cats.inject Hash.new do |hash, tag|\n\n # iterate through groups if the tag belongs to multiple\n tag[\"groups\"].each do |group|\n name = group[\"name\"]\n hash[name] ||= []\n hash[name] << tag[\"name\"]\n end\n hash\n end\n end",
"def index\n @mobile_phones = MobilePhone._search(params)\n @suggestions = @mobile_phones.suggestions\n @categories = {}\n @mobile_phones.aggs.keys.sort.each do |category|\n @categories[category] = @mobile_phones.aggs[category][\"buckets\"].sort_by{ |e| e[\"key\"] }\n end\n end",
"def top_3_words(text)\n text.gsub(\"\\n\", \" \")\n .split(\" \")\n .map(&:downcase)\n .map(&sanitize)\n .reject(&empty)\n .reject(&no_letters)\n .reduce({}, &top_word)\n .sort_by(&word_count)\n .reverse\n .take(3)\n .map(&:first)\nend",
"def get_keywords\n tweets = self.get_tweets\n counts = Hash.new(0)\n tweets.each do |tweet|\n tweet.text.downcase.split(/\\s+/).each do |word|\n word.gsub!(/\\p{^Alnum}/,'')\n next if word.size < 1\n counts[word] += 1\n end\n end\n temp_nest_array = (counts.select{ |k, v| v.to_i > 1}).sort_by{ |k, v| -v } # sort by count (descending) on counts of more than one word\n count_hash = Hash.new(0)\n temp_nest_array.each do |k, v|\n count_hash[k] = v\n end\n count_hash\n end",
"def top_drink_descriptors(how_many)\n # create empty array to hold top descriptors list for beer being rated\n @this_beer_descriptors = Array.new\n # find all descriptors for this drink\n @this_beer_all_descriptors = self.descriptors\n # Rails.logger.debug(\"this beer's descriptors: #{@this_beer_all_descriptors.inspect}\")\n @this_beer_all_descriptors.each do |descriptor|\n @descriptor = descriptor[\"name\"]\n @this_beer_descriptors << @descriptor\n end\n \n # attach count to each descriptor type to find the drink's most common descriptors\n @this_beer_descriptor_count = @this_beer_descriptors.each_with_object(Hash.new(0)) { |word,counts| counts[word] += 1 }\n # put descriptors in descending order of importance\n @this_beer_descriptor_count = Hash[@this_beer_descriptor_count.sort_by{ |_, v| -v }]\n # grab top 5 of most common descriptors for this drink\n @this_beer_descriptor_count.first(how_many)\n end",
"def _get_keywords_set_int_v_1_55()\n \n result = []\n \n genre_int = Genre.find(:all, :conditions => {:code => 'int'})[0]\n \n categories = Category.find(:all, :conditions => {:genre_id => genre_int.id})\n \n #debug\n if categories != nil\n \n write_log(\n @log_path,\n \"categories.size => #{categories.size}\", \n # __FILE__,\n __FILE__.split(\"/\")[-1],\n __LINE__.to_s)\n\n else\n write_log(\n @log_path,\n \"categories => nil\", \n # __FILE__,\n __FILE__.split(\"/\")[-1],\n __LINE__.to_s) \n end\n \n if categories != nil and categories.size > 0\n \n kws1 = KeyWordSet.new\n kws1.category = 'China'\n kws1.keywords = '中国 日中'\n \n kws2 = KeyWordSet.new\n kws2.category = 'Europe'\n # kws2.keywords = '米国 アメリカ'\n kws2.keywords = '欧州 イギリス ドイツ フランス ロシア'\n \n result.push(kws1)\n result.push(kws2)\n \n else\n \n kws1 = KeyWordSet.new\n kws1.category = 'China'\n kws1.keywords = '中国 日中'\n \n kws2 = KeyWordSet.new\n kws2.category = 'US'\n # kws2.keywords = '米国 アメリカ'\n kws2.keywords = '米国 アメリカ オバマ 米選挙'\n \n result.push(kws1)\n result.push(kws2)\n \n end\n \n \n return result\n \n end"
] |
[
"0.63290876",
"0.63278717",
"0.6326671",
"0.6317078",
"0.6298532",
"0.6271997",
"0.6177567",
"0.61019856",
"0.60731256",
"0.6015085",
"0.6009022",
"0.58957064",
"0.585386",
"0.58149844",
"0.5773825",
"0.57319534",
"0.5720246",
"0.56945884",
"0.5692115",
"0.56269914",
"0.56250197",
"0.56212527",
"0.5569006",
"0.55687344",
"0.55676156",
"0.55533534",
"0.55456406",
"0.55350214",
"0.55262786",
"0.5524376",
"0.5515285",
"0.551234",
"0.5505779",
"0.5494244",
"0.54806936",
"0.5473151",
"0.54347914",
"0.5422246",
"0.5408288",
"0.53716457",
"0.53712183",
"0.5346302",
"0.53214973",
"0.5315654",
"0.5312916",
"0.5298486",
"0.5291574",
"0.5285758",
"0.5284547",
"0.5284264",
"0.527735",
"0.527635",
"0.5270444",
"0.526787",
"0.5266879",
"0.52657294",
"0.52616304",
"0.52567196",
"0.52497315",
"0.52449834",
"0.524152",
"0.52376014",
"0.52281976",
"0.52246326",
"0.52114624",
"0.5203967",
"0.5200828",
"0.51961136",
"0.51873845",
"0.51848185",
"0.51814955",
"0.51794016",
"0.5162439",
"0.5160334",
"0.51601875",
"0.5157529",
"0.51548773",
"0.5153257",
"0.5152536",
"0.51462924",
"0.5144659",
"0.51440144",
"0.5139153",
"0.5135065",
"0.51280814",
"0.512259",
"0.5118036",
"0.5116829",
"0.51079017",
"0.51027304",
"0.5100325",
"0.509818",
"0.5093592",
"0.5092057",
"0.50912577",
"0.509125",
"0.5088639",
"0.50739974",
"0.5068303",
"0.5067889"
] |
0.6130565
|
7
|
Use callbacks to share common setup or constraints between actions.
|
def set_therapist
@therapist = Therapist.find(params[:id])
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def set_required_actions\n # TODO: check what fields change to asign required fields\n end",
"def action_hook; end",
"def run_actions; end",
"def define_action_hook; end",
"def actions; end",
"def define_action_helpers\n if super && action == :save\n @instance_helper_module.class_eval do\n define_method(:valid?) do |*args|\n self.class.state_machines.fire_event_attributes(self, :save, false) { super(*args) }\n end\n end\n end\n end",
"def add_actions; end",
"def callbacks; end",
"def callbacks; end",
"def setup *actions, &proc\n (@setup_procs ||= []) << [proc, actions.size > 0 ? actions : [:*]]\n end",
"def define_action_helpers; end",
"def post_setup\n end",
"def action_methods; end",
"def action_methods; end",
"def action_methods; end",
"def before_setup; end",
"def action_run\n end",
"def execute(setup)\n @action.call(setup)\n end",
"def define_action_helpers?; end",
"def set_actions\n actions :all\n end",
"def action_done(action)\n dispatch = { :migrate => :done_migrating, :map => :done_mapping, :reduce =>\n :done_reducing, :finalize => :done_finalizing } \n self.send dispatch[action[:action]], action\n end",
"def dependencies action, &block\n @actions.each do |other|\n if action[:requires].include? other[:provide]\n block.call other\n end\n end\n end",
"def setup!\n return unless @setup_procs\n http_actions = actions\n @setup_procs.each do |setup_proc|\n proc, actions = setup_proc\n @setup__actions = actions.map do |action|\n\n action.is_a?(Regexp) ?\n http_actions.select { |a| a.to_s =~ action } :\n action.is_a?(String) && action =~ /\\A\\./ ?\n http_actions.map { |a| a.to_s << action if format?(a).include?(action) }.compact :\n action\n\n end.flatten\n self.class_exec &proc\n @setup__actions = nil\n end\n @setup_procs = nil\n end",
"def before_actions(*logic)\n self.before_actions = logic\n end",
"def setup_handler\n end",
"def set_action(opts)\n opts = check_params(opts,[:actions])\n super(opts)\n end",
"def setup(action)\n @targets.clear\n unless action.item.target_filters.empty?\n @targets = SES::TargetManager.make_targets(action)\n else\n item = action.item\n if item.for_opponent?\n @targets = $game_troop.alive_members\n elsif item.for_dead_friend?\n @targets = $game_party.battle_members.select { |actor| actor.dead? }\n else\n $game_party.battle_members.select { |actor| actor.alive? }\n end\n end\n @item_max = @targets.size\n create_contents\n refresh\n show\n activate\n end",
"def action; end",
"def action; end",
"def action; end",
"def action; end",
"def action; end",
"def workflow\n end",
"def revisable_shared_setup(args, block)\n class << self\n attr_accessor :revisable_options\n end\n options = args.extract_options!\n self.revisable_options = Options.new(options, &block)\n \n self.send(:include, Common)\n self.send(:extend, Validations) unless self.revisable_options.no_validation_scoping?\n self.send(:include, WithoutScope::QuotedColumnConditions)\n end",
"def setup\n @action = SampleActionAndroid.new(os_name: 'android',\n app_name: APP_PATH)\n end",
"def before(action)\n invoke_callbacks *self.class.send(action).before\n end",
"def process_action(...)\n send_action(...)\n end",
"def before_dispatch(env); end",
"def after_actions(*logic)\n self.after_actions = logic\n end",
"def setup\n # override and do something appropriate\n end",
"def setup(client)\n return unless @setup\n actions = @setup['setup'].select { |action| action['do'] }.map { |action| Action.new(action['do']) }\n actions.each do |action|\n action.execute(client)\n end\n self\n end",
"def setup(_context)\n end",
"def setup(resources) ; end",
"def validate_actions\n errors.add(:base, :should_give_at_least_one_action) if !manage? && !forecasting? && !read? && !api?\n end",
"def setup\n @resource_config = {\n :callbacks => {\n :before_create => nil,\n :after_create => nil,\n :before_update => nil,\n :after_update => nil,\n :before_destroy => nil,\n :after_destroy => nil,\n },\n :child_assoc => nil,\n :model => nil,\n :parent => nil,\n :path => nil,\n :permission => {},\n :properties => {},\n :relation => {\n :create => nil,\n :delete => nil,\n },\n :roles => nil,\n }\n end",
"def determine_valid_action\n\n end",
"def process_shared\n handle_taxes\n handle_shippings\n create_adjustments_from_params\n handle_status\n handle_inventory_refunds\n handle_payment_transactions\n order.updater.update\n end",
"def startcompany(action)\n @done = true\n action.setup\n end",
"def init_actions\n am = action_manager()\n am.add_action(Action.new(\"&Disable selection\") { @selection_mode = :none; unbind_key(32); bind_key(32, :scroll_forward); } )\n am.add_action(Action.new(\"&Edit Toggle\") { @edit_toggle = !@edit_toggle; $status_message.value = \"Edit toggle is #{@edit_toggle}\" })\n end",
"def event_callbacks(event, metadata={})\n case event\n when :reset, :review\n if confirmed\n update_attributes(confirmed: false)\n end\n when :confirm\n confirm\n # trigger :order for all applicable items\n # NOTE: :order event is common to both physical and digital items\n items.each do |i|\n if i.event_permitted(:order)\n user_id = last_transition.user_id\n i.trigger!(:order, { order_id: id, user_id: user_id })\n end\n end\n when :complete_work\n request = metadata[:request]\n work_complete_notification(request)\n when :close\n close\n end\n if event != :close && !open\n reopen\n end\n end",
"def setup_action\n return unless PONY::ERRNO::check_sequence(current_act)\n new_sequence = @action_sequence[@sequence_index+1...@action_sequence.size]\n @sequence_index = 0\n new_sequence = DND::SkillSequence::ACTS[@acts[1]] + new_sequence\n execute_sequence\n end",
"def define_tasks\n define_weave_task\n connect_common_tasks\n end",
"def setup(&block)\n define_method(:setup, &block)\n end",
"def setup\n transition_to(:setup)\n end",
"def setup\n transition_to(:setup)\n end",
"def action\n end",
"def setup( *args )\n\t\t\tself.class.setupBlocks.each {|sblock|\n\t\t\t\tdebugMsg \"Calling setup block method #{sblock}\"\n\t\t\t\tself.send( sblock )\n\t\t\t}\n\t\t\tsuper( *args )\n\t\tend",
"def config(action, *args); end",
"def setup\n @setup_proc.call(self) if @setup_proc\n end",
"def before_action \n end",
"def setup_callbacks\n defined_callbacks.each do |meth|\n unless respond_to?(\"call_#{meth}_callbacks\".to_sym)\n self.class.module_eval <<-EOE\n def call_#{meth}_callbacks(*args)\n plugin_store.each {|a| a.call_#{meth}_callbacks(*args) } if respond_to?(:plugin_store) && plugin_store\n self.send :#{meth}, *args if respond_to?(:#{meth})\n end\n EOE\n end\n end\n end",
"def action\n end",
"def matt_custom_action_begin(label); end",
"def setup\n # override this if needed\n end",
"def setup\n\t\t\t\t\t\t# Do nothing\n\t\t\t\tend",
"def setup\n\t\t\t\t\t\t# Do nothing\n\t\t\t\tend",
"def action(options,&callback)\n new_action = Action===options ? options : Action.new(options,&callback)\n # replace any with (shared name/alias or both default) + same arity\n @actions.delete_if do |existing_action|\n ((existing_action.names & new_action.names).size > 0 ||\n existing_action.default? && new_action.default?) &&\n existing_action.required.size == new_action.required.size &&\n existing_action.optional.size <= new_action.optional.size\n end\n @actions = (@actions + [new_action]).sort\n new_action\n end",
"def set_target_and_action target, action\n self.target = target\n self.action = 'sugarcube_handle_action:'\n @sugarcube_action = action\n end",
"def after(action)\n invoke_callbacks *options_for(action).after\n end",
"def pre_task\n end",
"def setup(server)\n server.on('beforeMethod', method(:before_method), 10)\n end",
"def add_actions\n attribute = machine.attribute\n name = self.name\n \n owner_class.class_eval do\n define_method(name) {self.class.state_machines[attribute].events[name].fire(self)}\n define_method(\"#{name}!\") {self.class.state_machines[attribute].events[name].fire!(self)}\n define_method(\"can_#{name}?\") {self.class.state_machines[attribute].events[name].can_fire?(self)}\n end\n end",
"def init_actions\n @select_action = SelectAction.new\n @endpoint_mouse_action = EndpointMouseAction.new\n @move_action = MoveAction.new\n end",
"def setup_signals; end",
"def after_created\r\n return unless compile_time\r\n Array(action).each do |action|\r\n run_action(action)\r\n end\r\nend",
"def after_created\r\n return unless compile_time\r\n Array(action).each do |action|\r\n run_action(action)\r\n end\r\nend",
"def set_target_and_action target, action\n self.target = target\n self.action = 'sugarcube_handle_action:'\n @sugarcube_action = action.respond_to?('weak!') ? action.weak! : action\n end",
"def initialize(*args)\n super\n @action = :set\nend",
"def after_set_callback; end",
"def setup\n #implement in subclass;\n end",
"def lookup_action; end",
"def setup &block\n if block_given?\n @setup = block\n else\n @setup.call\n end\n end",
"def setup_action\n return TSBS.error(@acts[0], 1, @used_sequence) if @acts.size < 2\n actions = TSBS::AnimLoop[@acts[1]]\n if actions.nil?\n show_action_error(@acts[1])\n end\n @sequence_stack.push(@acts[1])\n @used_sequence = @acts[1]\n actions.each do |acts|\n @acts = acts\n execute_sequence\n break if @break_action\n end\n @sequence_stack.pop\n @used_sequence = @sequence_stack[-1]\n end",
"def release_actions; end",
"def around_hooks; end",
"def save_action; end",
"def setup(easy)\n super\n easy.customrequest = @verb\n end",
"def action_target()\n \n end",
"def setup\n callback(:setup) do\n notify(:setup)\n migration_check.last_deployed_commit\n end\n end",
"def setup\n return unless @setup\n\n actions = @setup['setup'].select { |action| action['do'] }.map { |action| Action.new(action['do']) }\n run_actions_and_retry(actions)\n self\n end",
"def before_setup\n # do nothing by default\n end",
"def my_actions(options)\n @setup = false\n get_template_part(\"custom_used\",\"action_users\",true)\n end",
"def default_action; end",
"def setup(&blk)\n @setup_block = blk\n end",
"def callback_phase\n super\n end",
"def advice\n end",
"def _handle_action_missing(*args); end",
"def duas1(action)\n action.call\n action.call\nend",
"def shared_action(name, &block)\n @controller.shared_actions[name] = block\n end",
"def before_action action, &block\n @audience[:before][action] ||= Set.new\n @audience[:before][action] << block\n end",
"def setup_initial_state\n\n state_a = State.new(\"a\", 0)\n state_b = State.new(\"b\", 0)\n state_c = State.new(\"c\", 10)\n\n move_to_b = Action.new(\"move_to_b\", 1, state_b)\n\n move_to_c = Action.new(\"move_to_c\", 1, state_c)\n\n state_a.actions = [move_to_b, move_to_c]\n\n return state_a\n \nend"
] |
[
"0.6163163",
"0.6045976",
"0.5946146",
"0.591683",
"0.5890051",
"0.58349305",
"0.5776858",
"0.5703237",
"0.5703237",
"0.5652805",
"0.5621621",
"0.54210985",
"0.5411113",
"0.5411113",
"0.5411113",
"0.5391541",
"0.53794575",
"0.5357573",
"0.53402257",
"0.53394014",
"0.53321576",
"0.53124547",
"0.529654",
"0.5296262",
"0.52952296",
"0.52600986",
"0.52442724",
"0.52385926",
"0.52385926",
"0.52385926",
"0.52385926",
"0.52385926",
"0.5232394",
"0.523231",
"0.5227454",
"0.52226824",
"0.52201617",
"0.5212327",
"0.52079266",
"0.52050185",
"0.51754695",
"0.51726824",
"0.51710224",
"0.5166172",
"0.5159343",
"0.51578903",
"0.51522785",
"0.5152022",
"0.51518047",
"0.51456624",
"0.51398855",
"0.5133759",
"0.5112076",
"0.5111866",
"0.5111866",
"0.5110294",
"0.5106169",
"0.509231",
"0.50873137",
"0.5081088",
"0.508059",
"0.50677156",
"0.50562143",
"0.5050554",
"0.50474834",
"0.50474834",
"0.5036181",
"0.5026331",
"0.5022976",
"0.5015441",
"0.50121695",
"0.5000944",
"0.5000019",
"0.4996878",
"0.4989888",
"0.4989888",
"0.49864885",
"0.49797225",
"0.49785787",
"0.4976161",
"0.49683493",
"0.4965126",
"0.4958034",
"0.49559742",
"0.4954353",
"0.49535993",
"0.4952725",
"0.49467874",
"0.49423352",
"0.49325448",
"0.49282882",
"0.49269363",
"0.49269104",
"0.49252945",
"0.4923091",
"0.49194667",
"0.49174926",
"0.49173003",
"0.49171105",
"0.4915879",
"0.49155936"
] |
0.0
|
-1
|
Never trust parameters from the scary internet, only allow the white list through.
|
def therapist_params
params.require(:therapist).permit(:pt_id, :name, :summary, :phone)
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def strong_params\n params.require(:user).permit(param_whitelist)\n end",
"def strong_params\n params.require(:listing_member).permit(param_whitelist)\n end",
"def allow_params_authentication!; end",
"def allowed_params\n ALLOWED_PARAMS\n end",
"def default_param_whitelist\n [\"mode\"]\n end",
"def param_whitelist\n [:role, :title]\n end",
"def expected_permitted_parameter_names; end",
"def safe_params\n params.except(:host, :port, :protocol).permit!\n end",
"def strong_params\n params.require(:team_member).permit(param_whitelist)\n end",
"def permitir_parametros\n \t\tparams.permit!\n \tend",
"def strong_params\n params.require(:community).permit(param_whitelist)\n end",
"def permitted_strong_parameters\n :all #or an array of parameters, example: [:name, :email]\n end",
"def strong_params\n params.require(:education).permit(param_whitelist)\n end",
"def restricted_params\n #params.require(self.controller_name.classify.underscore.to_sym).permit([])\n raise(\"No strong params set, override restricted_params method in your controller. E.g. params.require(:model).permit(:attribute1, :attribute2)\")\n end",
"def allowed_params\n params.require(:user).permit(:username, :email, :password, :password_confirmation)\n end",
"def param_whitelist\n [:rating, :review]\n end",
"def param_whitelist\n whitelist = [\n :username, :name,\n :parent_id,\n :headline, :description, :video,\n :policy, :signup_mode, :category,\n :website, :facebook, :twitter, :linkedin,\n :founded_at,\n privacy: [\n :events,\n :resources\n ],\n permission: [\n :profile,\n :members,\n :children,\n :statistics,\n :posts,\n :listings,\n :resources,\n :events\n ],\n location: [\n :description,\n :street,\n :city,\n :state,\n :zip,\n :country,\n :latitude,\n :longitude\n ]\n ]\n \n if action_name === 'update'\n whitelist.delete(:parent_id)\n unless current_user.role_in(@community) === 'owner'\n whitelist.delete(:privacy)\n whitelist.delete(:permission)\n end\n end\n \n whitelist\n end",
"def param_whitelist\n if @user.present? && current_user != @user\n return [:followed]\n end\n \n whitelist = [\n :username, :email, :password,\n :first_name, :last_name,\n :birthday, :gender,\n :headline, :biography, :ask_about, :focus,\n :website, :facebook, :linkedin, :twitter, :github,\n roles: [],\n skills: [],\n interests: [],\n privacy: { contact: [] },\n location: [\n :description,\n :street,\n :city,\n :state,\n :zip,\n :country,\n :latitude,\n :longitude\n ]\n ]\n \n if action_name === 'update'\n whitelist.delete(:email)\n whitelist.delete(:password)\n end\n \n whitelist\n end",
"def user_params \n \tparams.require(:user).permit(:name, :email, :password, :password_confirmation)# preventing CSTR\n end",
"def user_params\n params.permit(:name, :phoneNumber, :address, :postalCode, :local, :link, :counter, :latitude, :longitude) \n end",
"def valid_params_request?; end",
"def strong_params\n params.require(:experience).permit(param_whitelist)\n end",
"def trim_whitelisted(params, whitelist)\n # remove any parameters that are not whitelisted\n params.each do |key, value|\n # if white listed\n if whitelist.include? key\n # strip the parameters of any extra spaces, save as string\n params[key] = value.to_s.strip\n else\n # delete any unauthorized parameters\n params.delete key\n end\n end\n params\n end",
"def whitelist_url_params\n params.require(:whitelist_url).permit(:domain)\n end",
"def allowed_params\n params.require(:allowed).permit(:email)\n end",
"def permitted_params\n []\n end",
"def trim_whitelisted(params, whitelist)\n # remove any parameters that are not whitelisted\n params.each do |key, value|\n # if white listed\n if whitelist.include? key\n # strip the parameters of any extra spaces, save as string\n params[key] = value.to_s.strip\n else\n # delete any unauthorized parameters\n params.delete key\n end\n end\n params\n end",
"def safe_params\n params.permit(:id, :name, :origin, :emails => []); #emails is an array\n end",
"def query_param\n\t\tparams.permit(:first_name, :last_name, :phone)\n\tend",
"def strong_params\n params.require(:success_metric).permit(param_whitelist)\n end",
"def devise_filter\r\n logger.debug(\"In devise_filter =>PARAMS: #{params.inspect}\")\r\n\r\n # White list for sign_up\r\n devise_parameter_sanitizer.for(:sign_up) { |u| u.permit(user_whitelist) }\r\n\r\n # White list for account update\r\n devise_parameter_sanitizer.for(:account_update) { |u| u.permit(user_whitelist, :current_password) }\r\n\r\n # White list for Invitation creation\r\n devise_parameter_sanitizer.for(:invite) { |u| u.permit(:account_type, :email, :invitation_token)}\r\n\r\n # White list for accept invitation\r\n devise_parameter_sanitizer.for(:accept_invitation) { |u| u.permit(user_whitelist, :invitation_token)}\r\n\r\n end",
"def whitelisted_user_params\n params.require(:user).\n permit( :first_name, :last_name, :email,:password,:password_confirmation,:birthday,:gender)\n end",
"def user_params\n ActionController::Parameters.permit_all_parameters = true\n params.require(:user) #.permit(:name, :surname, :phone, :password, :email, :time_zone)\n end",
"def strong_params\n params.require(:metric_change).permit(param_whitelist)\n end",
"def safe_params\n params.require(:user).permit(:name)\n end",
"def get_params\n\t\treturn ActionController::Parameters.new(self.attributes).permit(\"account_id\", \"title\", \"category\", \"introduction\", \"tags\", \"segment_type\", \"visible\", \"status\", \"main_image\")\n\tend",
"def grant_params\n @whitelisted = params.require(:grant).permit(:name, :description, :agency_id, :acronym)\n end",
"def check_params; true; end",
"def param_whitelist\n whitelist = [\n :description,\n :progress,\n :kpi_id\n ]\n \n unless action_name === 'create'\n whitelist.delete(:kpi_id)\n end\n \n whitelist\n end",
"def quote_params\n params.permit!\n end",
"def valid_params?; end",
"def paramunold_params\n params.require(:paramunold).permit!\n end",
"def user_params\n\t\tparams.permit(:nickname, :avatar, :description, :password, :gender, :birthday, :email, :phone, :qq_id, :wechat_id)\n\tend",
"def filtered_parameters; end",
"def user_params\n params.permit(\n \t:id,\n \t:email, \n \t:first_name, \n \t:last_name, \n \t:password, \n \t:confirm_token, \n \t:phone_number,\n \t:facebook_link,\n \t:car_model,\n \t:license_plate)\n end",
"def filtering_params\n params.permit(:email, :name)\n end",
"def check_params\n true\n end",
"def wx_public_params\n params.require(:wx_public).permit(:nickname, :manager, :alias)\n end",
"def allowed_params\n params.require(:user).permit(:email, :password, :role, :first_name, :last_name, :password_confirmation)\n end",
"def allowed_params\n params.require(:user).permit(:email, :password, :role, :first_name, :last_name, :password_confirmation)\n end",
"def listing_params\n\t\tparams.permit(:address, :transit_info, :rules, :other_info, :lat, :lng)\n\tend",
"def social_account_params\n\t\t\tparams.require(:social_account).permit!\n\t\tend",
"def safe_params\n resurce_name = self.class.resource_name\n params_method_name = \"#{resurce_name}_params\".to_sym\n if params[resurce_name]\n if respond_to?(params_method_name) || private_methods.include?(params_method_name)\n send(params_method_name)\n else\n raise ActiveModel::ForbiddenAttributesError, \"Please, define the '#{params_method_name}' method in #{self.class.name}\"\n end\n end\n end",
"def url_params\n params.require(:url).permit(:short_url, :original_url, :clicks, :ip_addresses)\n end",
"def user_params\n params.require(:user).permit(:uri, :username, :password, :realname, :email, :publicvisible)\n end",
"def model_params\n\t\tparams.require(:manager).permit(\n\t :user_name,\n :password,\n :email,\n \t\t\t)\n\tend",
"def article_params_whitelist\n params.require(:article).permit(:title, :description, category_ids: [])\n end",
"def college_whitelist_params\n params.require(:college_whitelist).permit(:status)\n end",
"def active_code_params\n params[:active_code].permit\n end",
"def filtering_params\n params.permit(:email)\n end",
"def valid_params(params)\n params.permit(:user_id, :photo_id, :originX, :originY, :width, :height)\n end",
"def ip_address_params\n\t\t\tparams.require(:ip_address).permit!\n end",
"def pull_request_params\n whitelist = [\n :url,\n :id,\n :html_url,\n :diff_url,\n :patch_url,\n :issue_url,\n :number,\n :state,\n :locked,\n :title\n ]\n params.require(:pull_request).permit(whitelist)\n end",
"def reserved_params\n params.require(:reserved).permit(:name, :email, :pax, :address, :KTP, :title)\n end",
"def post_params\n if current_user.admin? \n params.permit(:title, :body, :city, :country, :gps_location, :privacy, :visible, :latitude, :longitude, images: [], files: [])\n else \n params.permit(:title, :body, :city, :country, :gps_location, :privacy,:latitude, :longitude, images: [], files: [])\n end \n end",
"def list_params\n params.permit(:name)\n end",
"def filter_parameters; end",
"def filter_parameters; end",
"def vineyard_params\n params.permit(:vineyard_name, :email, :website_url, :phone, :address, :city, :region, :postcode, :country, :specialty, :description, :pet_friendly, :holiday, :tours, :events, :family_friendly, :cover_image, :image_one, :image_two, :image_three, :image_four, :user_id, :base64)\n end",
"def available_activity_params\n # params.require(:available_activity).permit(:type,:geometry,:properties)\n whitelisted = ActionController::Parameters.new({\n type: params.require(:available_activity)[:type],\n geometry: params.require(:available_activity)[:geometry].try(:permit!).to_h,\n properties: params.require(:available_activity)[:properties].try(:permit!).to_h\n }).try(:permit!)\n end",
"def user_params\n params.permit(:name, :username, :email, :password, :img_url, :bg_url, :coinbank)\n end",
"def user_params_pub\n\t \tparams[:user].permit(:hruid)\n\t end",
"def user_params\n params.permit(:id, :email, :password, :nickname, :status, :avatar, :flat_picture, :flatsharing_id, :member,\n :user, :color, :solde)\n end",
"def validate_search_inputs\n @whitelisted = params.fetch(:user, nil)\n if @whitelisted.blank?\n render_error(400, \"#{I18n.t('general_error.params_missing_key')}\": [I18n.t('general_error.params_missing_value', model: \"review\")])\n return\n else\n @whitelisted = @whitelisted.permit(:name, :uen, :description)\n end\n end",
"def param_whitelist\n [\n :title,\n :description,\n :organization,\n :team_id,\n :started_at,\n :finished_at,\n location: [\n :description,\n :street,\n :city,\n :state,\n :zip,\n :country,\n :latitude,\n :longitude\n ]\n ]\n end",
"def url_whitelist; end",
"def admin_social_network_params\n params.require(:social_network).permit!\n end",
"def filter_params\n params.require(:filters).permit(:letters)\n end",
"def origin_params\n params.permit(:country, :state, :city, :postal_code, :address, :description)\n end",
"def valid_params(params)\n params.permit(:login, :first_name, :last_name, \n :password, :password_confirmation)\n end",
"def sensitive_params=(params)\n @sensitive_params = params\n end",
"def permit_request_params\n params.permit(:address)\n end",
"def user_params\n # Ensure a user can't give themselves admin priveleges\n params.delete(:admin) if current_user.admin?\n params.require(:user).permit(:name, :email, :admin, :image)\n end",
"def secure_params\n params.require(:location).permit(:name)\n end",
"def strong_params\n params.require( :setting ).\n permit( :global_scan_limit, :per_user_scan_limit,\n :target_whitelist_patterns, :target_blacklist_patterns )\n end",
"def question_params\n params.require(:survey_question).permit(question_whitelist)\n end",
"def case_insensitive_params\n params.require(:case_insensitive).permit(:name)\n end",
"def empire_master_no_match_params\n params.require(:empire_master_no_match).permit(:uid, :last_name, :list, :search_date, :double, :source)\n end",
"def maintenance_request_params\n params[:maintenance_request].permit! #allow all parameters for now\n end",
"def unwanted_params\n params.require(:unwanted).permit(:title, :description, :image)\n end",
"def url_params\n params[:url].permit(:full)\n end",
"def backend_user_params\n params.permit!\n end",
"def filter_params\n\t\treturn params[:candidate].permit(:name_for_filter)\n\tend",
"def speed_measurement_params\n\n #fuckit, to lazy to deal with permit crap right now\n ActionController::Parameters.permit_all_parameters = true\n\n params[:speed_measurement]\n end",
"def user_params\n params.permit(:name, :age, :username, :display_photo, :password)\n end",
"def get_params\r\n #params.require(:article).permit(:title, :permalink, :content, :source_site, :introtext, :type_id, :order_by, :searchable, :created_by, :edited_by, :published_by, :published_on, :user_id)\r\n params.require(:article).permit!\r\n\r\n end",
"def pub_params\n params.require(:pub).permit(:name, :description, :phone, :email, :hidden, :city_id, :address)\n end",
"def pass_params\n params[:pass].permit(:name, :price, :description, :colour, :events)\n end",
"def droptraining_params\n params.permit(:training_id,:user_id, :utf8, :authenticity_token, :commit)\n end",
"def person_params\n # params whitelist does *not* include admin, sub, remember_token\n # TBD: share this whitelist with the list used by configuration_permitted_parameters\n # TBD: should current_password be on this list? -- for now, leaving off, since it seems to work without\n # NOTE: do not include 'admin' in this list!\n params.require(:person).permit(\n :name, \n :email, \n :description,\n :password, \n :password_confirmation\n )\n end",
"def parameter_params\n params.require(:parameter).permit(:name, :description, :param_code, :param_value, :active_from, :active_to)\n end"
] |
[
"0.69792545",
"0.6781151",
"0.67419964",
"0.674013",
"0.6734356",
"0.6591046",
"0.6502396",
"0.6496313",
"0.6480641",
"0.6477825",
"0.64565",
"0.6438387",
"0.63791263",
"0.63740575",
"0.6364131",
"0.63192815",
"0.62991166",
"0.62978333",
"0.6292148",
"0.6290449",
"0.6290076",
"0.62894756",
"0.6283177",
"0.6242471",
"0.62382483",
"0.6217549",
"0.6214457",
"0.6209053",
"0.6193042",
"0.6177802",
"0.6174604",
"0.61714715",
"0.6161512",
"0.6151757",
"0.6150663",
"0.61461",
"0.61213595",
"0.611406",
"0.6106206",
"0.6105114",
"0.6089039",
"0.6081015",
"0.6071004",
"0.60620916",
"0.6019971",
"0.601788",
"0.6011056",
"0.6010898",
"0.6005122",
"0.6005122",
"0.6001556",
"0.6001049",
"0.59943926",
"0.5992201",
"0.59909594",
"0.5990628",
"0.5980841",
"0.59669393",
"0.59589154",
"0.5958826",
"0.5957911",
"0.5957385",
"0.5953072",
"0.59526145",
"0.5943361",
"0.59386164",
"0.59375334",
"0.59375334",
"0.5933856",
"0.59292704",
"0.59254247",
"0.5924164",
"0.59167904",
"0.59088355",
"0.5907542",
"0.59064597",
"0.5906243",
"0.5898226",
"0.589687",
"0.5896091",
"0.5894501",
"0.5894289",
"0.5891739",
"0.58860534",
"0.5882406",
"0.587974",
"0.58738774",
"0.5869024",
"0.58679986",
"0.5867561",
"0.5865932",
"0.5864461",
"0.58639693",
"0.58617616",
"0.5861436",
"0.5860451",
"0.58602303",
"0.5854586",
"0.58537364",
"0.5850427",
"0.5850199"
] |
0.0
|
-1
|
Helper method to ensure wait time with additional buffer
|
def wait_until_with_buffer(args, &block)
original_timeout = args[:timeout] || ENV['WAIT_TIMEOUT'].to_i
args_buffered = args.dup
args_buffered[:timeout] = 60
start_time = Time.now
Frank::Cucumber::WaitHelper.wait_until(args_buffered) { block.call() }
end_time = Time.now
delta = end_time - start_time
puts("wait_until exceeded timeout #{original_timeout}. Took #{delta}. #{caller[0]}") if delta > original_timeout
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def wait_readable_or_timeout; end",
"def wait_until_not_full; end",
"def wait_writable_or_timeout; end",
"def wait; end",
"def wait; end",
"def wait; end",
"def wait_for_rx\n sleep DATA_REFRESH_RATE + @latency\n end",
"def wait\n sleep WAIT_TIME unless @skip_wait\n end",
"def wait\n sleep 0.0001\n end",
"def brute_wait(delay)\n sleep(delay)\n end",
"def wait\n 0\n end",
"def wait_for_pending_sends; end",
"def wait_for_tokens size\n available_time = @last_fetched + (size / @bandwidth)\n waiting_time = [0, available_time - time].max\n sleep waiting_time\n @last_fetched = time\n increase_bandwidth\n end",
"def abs_wait_short\n wait(15)\n end",
"def waiting; end",
"def waiting; end",
"def cheap_wait; end",
"def wait\n true\n end",
"def ev_do_calculate_rw_wait(wait); end",
"def implicit_wait; end",
"def wait\n loop do sleep 1 end\n end",
"def wait(arg0)\n end",
"def wait_interval(offset = T.unsafe(nil)); end",
"def wait!\n sleep(@sleep)\n end",
"def buffer_command(cmd)\n @cmd_time ||= Time.now.to_f\n @cmds << cmd\n flush_commands if Time.now.to_f - @cmd_time > BUFFER_TIME\n end",
"def wait\n @wait.synchronize do\n sleep 1 while @count >= THREAD_MAX\n @count += 1\n end\n end",
"def wait(frames)\n return if @wait > 0\n @wait = frames\n end",
"def wait_for_less_busy_worker(val = T.unsafe(nil)); end",
"def wait_for(pattern)\n $_api_exec_state = :wait_for\n if pattern.is_a?(Array)\n pattern = Regexp.new(pattern.join('|'))\n end\n\n (0..1000000).each do\n line = API::sync_read\n if line and line.match(pattern)\n $_api_exec_state = :none\n return\n end\n sleep 0.01\n end\nend",
"def wait(what = T.unsafe(nil)); end",
"def update_wait?\r\n # If waiting\r\n if @wait_count > 0\r\n # Reduce wait count\r\n @wait_count -= 1\r\n return true\r\n end\r\n return false\r\n end",
"def wait_until_available\n return unless @locked\n\n @mutex.lock\n @mutex.unlock\n end",
"def update_wait?\r\n # If waiting\r\n if @wait_count > 0\r\n # Decrease wait count\r\n @wait_count -= 1\r\n return true\r\n end\r\n return false\r\n end",
"def sync_wait\n if IO.select([sync_read], nil, nil, 20).nil?\n # timeout reading from the sync pipe.\n send_side_channel_error(\"Error syncing processes in run lock test (timeout)\")\n exit!(1)\n else\n sync_read.getc\n end\n end",
"def wait\n\tend",
"def wait\n\t\t\t\t@available.wait\n\t\t\tend",
"def wait\n #$_api_queue.clear\n wait_for(/>/)\nend",
"def wait(seconds)\n @waiting = seconds * 1000\n end",
"def wait_until_available\n return unless locked?\n\n @mutex.synchronize {}\n end",
"def implicit_wait=(seconds); end",
"def wait!\n now = Time.now.utc.to_i\n duration = (reset.to_i - now) + 1\n\n sleep duration if duration >= 0\n\n yield if block_given?\n\n duration\n end",
"def wait_for_launching\n sleep @delay\n end",
"def wait!\n wait\n raise @error if timeout?\n @messages\n end",
"def wait\r\n Ragweed::Wrap32::wait_for_single_object @h\r\n end",
"def remainder_as_buffer; end",
"def acquire_or_wait message_size\n waiter = nil\n while is_new_and_others_wait?(waiter) ||\n would_exceed_byte_limit?(message_size) ||\n would_exceed_message_limit?\n\n if waiter.nil?\n waiter = Concurrent::Event.new\n # This waiter gets added to the back of the line.\n @awaiting << waiter\n else\n waiter = Concurrent::Event.new\n # This waiter already in line stays at the head of the line.\n @awaiting[0] = waiter\n end\n @mutex.unlock\n waiter.wait\n @mutex.lock\n end\n @outstanding_messages += 1\n @outstanding_bytes += message_size\n\n @awaiting.shift if waiter # Remove the newly released waiter from the head of the queue.\n\n # There may be some surplus left; let the next message waiting try to acquire a permit.\n if !@awaiting.empty? && @outstanding_bytes < byte_limit && @outstanding_messages < message_limit\n @awaiting.first.set\n end\n end",
"def wait(duration, variable)\n for i in 0...duration\n @wait_time += 1 if variable\n @wait_time_thirst if variable == false \n break if i >= duration / 2\n end\n end",
"def wait_async(time)\n EM::Synchrony.sleep(time)\nend",
"def wait(frames)\n frames.times do\n wait_internal\n end\n end",
"def wait(timeout: nil)\n\t\t\t\tif timeout\n\t\t\t\t\tread_ready, _, _ = IO.select([@output], [], [], timeout)\n\t\t\t\t\t\n\t\t\t\t\treturn false unless read_ready and read_ready.any?\n\t\t\t\tend\n\t\t\t\t\n\t\t\t\t@signalled or @output.read(1)\n\t\t\t\t\n\t\t\t\t# Just in case that this was split across multiple processes.\n\t\t\t\t@signalled = true\n\t\t\tend",
"def lock_timeout; end",
"def wait\n @running.reset\n @waiting.set\n @running.wait\n end",
"def timed_write(buf, wait = def_write_timeout, opts = {})\n if (wait and wait > 0)\n Timeout.timeout(wait) {\n return write(buf, opts)\n }\n else\n return write(buf, opts)\n end\n end",
"def waitTillLimitReset\n timeTillReset = CLIENT.rate_limit.resets_in + 5\n @logger.info(\"API limit reached while fetching... Sleeping for #{timeTillReset} seconds 😴 brb\")\n sleep(timeTillReset)\nend",
"def pre_sleep; end",
"def work_pending?\n !buffer.empty?\n end",
"def wait(duration, value)\n for i in 0...duration\n @wait_time += 1 if value == false\n @wait_time2 += 1 if value\n break if i >= duration / 2\n end\n end",
"def wait(duration, value)\n for i in 0...duration\n @wait_time += 1 if value == false\n @wait_time2 += 1 if value\n break if i >= duration / 2\n end\n end",
"def num_waiting\n end",
"def num_waiting\n end",
"def wait(timeout = nil)\n @latch.wait(timeout)\n end",
"def wait_for_ready\n sleep 0.1 until ready?\n end",
"def wait_for_seconds\n\t\tsleep(1 * rand + 1)\n\tend",
"def queued_messages?\r\n @pointer < @buffer.length\r\n end",
"def wait_done\n sleep 0.01 until done?\n end",
"def wait_connection=(_arg0); end",
"def recvn(num_bytes, timeout: nil)\n @timer.countdown(timeout) do\n fillbuffer while @timer.active? && @buffer.size < num_bytes\n @buffer.size >= num_bytes ? @buffer.get(num_bytes) : ''\n end\n end",
"def check_buffer(conf)\n log.trace \"checkbuffer invoked\"\n if !conf.elements(name: 'buffer').empty?\n log.info \"buffer set; size=\", conf.elements(name: 'chunk_limit_records')\n end \n end",
"def waitUntil\n until yield\n sleep 0.5\n end\nend",
"def wait_objects_on_the_screen(expected_num_objects_on_screen, query_image_path, timeout = 20, take_res_if_expect_fail = true)\n logc(\"method: #{__method__}, params: #{expected_num_objects_on_screen}, #{query_image_path},\" + \n \" #{timeout}, #{take_res_if_expect_fail}\")\n\n time_start = Time.now\n time_end = time_start + timeout.to_i\n logc(\"Checking will be ended at '#{time_end}'\")\n\n res_image_path = File.join(@report_path,\n \"#{@scenario_name.to_s.gsub(\" \", \"_\").downcase}\" +\n \"_result_find_object_\" +\n \"#{File.basename(query_image_path)}\")\n\n #wait for time_end reached OR expectation reached\n res_of_finding = nil\n occurrences = nil\n attempt_counter = 1\n while true\n attempt_time = Time.now\n logc(\"Attempt: '#{attempt_counter}', time: #{attempt_time}\")\n res_of_finding = find_objects_on_the_screen(query_image_path, res_image_path)\n occurrences = res_of_finding[\"rectangle_centers\"].size.to_i\n\n if (occurrences == expected_num_objects_on_screen) || (attempt_time > time_end)\n break\n else\n attempt_counter += 1\n sleep 1\n end\n end\n\n assert_false_custom(occurrences.nil? || res_of_finding.nil?,\n \"Error: object occurrences or res_of_finding on image can't be nil. Check find method\")\n\n logc(\"Occurrences found: #{occurrences}\\n Time spent to wait expected result: #{Time.now - time_start}\")\n\n is_expectation_reached = (occurrences == expected_num_objects_on_screen)\n # remove_file_if_exist(res_image_path) if is_expectation_reached || !take_res_if_expect_fail\n\n return res_of_finding\nend",
"def send_and_wait(type, *args); end",
"def wait_available_slot\n while true\n index = available_slot\n return index if index\n sleep 0.01\n end\n end",
"def wait_for_sequence\n tsbs_wait_update\n tsbs_wait_update while @spriteset.busy?\n end",
"def fill\n begin\n @buffer << @io.read_nonblock(@buffer_size)\n rescue ::IO::WaitReadable\n IO.select([@io], nil, nil, @read_timeout) ? retry : (raise ::Timeout::Error)\n rescue ::IO::WaitWritable\n IO.select(nil, [@io], nil, @read_timeout) ? retry : (raise ::Timeout::Error)\n end\n end",
"def fill\n begin\n @buffer << @io.read_nonblock(@buffer_size)\n rescue ::IO::WaitReadable\n IO.select([@io], nil, nil, @read_timeout) ? retry : (raise ::Timeout::Error)\n rescue ::IO::WaitWritable\n IO.select(nil, [@io], nil, @read_timeout) ? retry : (raise ::Timeout::Error)\n end\n end",
"def wait_connection; end",
"def wait_connection; end",
"def wait_until(timeout=20, &block)\n time_to_stop = Time.now + timeout\n until yield do\n sleep(0.1) # much less cpu stress\n break if Time.now > time_to_stop\n end\nend",
"def exp_process\n sleep(@wait_sec.to_f)\n # handle input\n @exp_buffer << @exp_buffer_values.shift.to_s\n end",
"def process(wait = T.unsafe(nil), &block); end",
"def wait\n while source\n saved_source = source\n saved_source.wait\n break if saved_source.equal?(source)\n end\n end",
"def while_time_remaining; end",
"def io_sleep\n # sleep 2.5/1000.0\n end",
"def read_buffer; end",
"def wait(per = nil)\n @target += per || @period\n error = @target - Time.now\n sleep error if error > 0\n true\n end",
"def hard_test\n wait(10) # let some capacitor get up some charge.\n \n 5.times do\n wait(5)\n cmd(\"CFS CFS_WHE_OBS_START\")\n wait(5) \n cmd(\"CFS CFS_WHE_HTR_ON\")\n wait(5)\n cmd(\"CFS CFS_WHE_LOUVER_CLOSE\")\n wait(5)\n \n end\nend",
"def first_data_timeout(seconds); end",
"def wait(timeout = nil)\n synchronize do\n touch\n # TODO interruptions ?\n super timeout if incomplete?\n self\n end\n end",
"def queued_messages?\n @pointer < @buffer.length\n end",
"def send_blocking_command(command, _timeout, &block); end",
"def wait_until_ready\n # this method may be left unimplemented if that is applicable log\n end",
"def cwWatchExecute(watchCommand,resBuffer,resBufferSize,wantResult,timtimeout=10.0)\r\n begin\r\n Timeout::timeout(timtimeout) { \r\n res = eval(watchCommand).to_s + \"\\000\"\r\n cwSetWatchResult(resBuffer,resBufferSize,res.to_s) if(wantResult && resBuffer.to_i!=0)\r\n return true\r\n }\r\n rescue Exception => e\r\n res = e.class.to_s + \"\\000\"\r\n cwSetWatchResult(resBuffer,resBufferSize,res.to_s) if(resBuffer.to_i!=0)\r\n return false\r\n end\r\nend",
"def lock_timeout=(_arg0); end",
"def take_a_nap\n sleep(2)\nend",
"def wait(duration)\n for i in 0...duration\n update_basic(false, true, true)\n end\n end",
"def balloon_wait\r\r\n return 20\r\r\n end",
"def wait\n @timer_thread.join\n end",
"def buffered\n @buffer.length\n end",
"def buffered\n @buffer.length\n end",
"def ready?\n warn \"Hmm, too many records in queue: #{queue}\" if queue.size > delay+1\n queue.size > delay\n end"
] |
[
"0.697314",
"0.6893673",
"0.65395635",
"0.65077627",
"0.65077627",
"0.65077627",
"0.64972615",
"0.6451797",
"0.63891727",
"0.6375768",
"0.63728917",
"0.6355762",
"0.6327225",
"0.62768304",
"0.62280184",
"0.62280184",
"0.616895",
"0.6149683",
"0.6135843",
"0.61253667",
"0.6086527",
"0.60821784",
"0.60546505",
"0.6023302",
"0.6004895",
"0.5979746",
"0.5972244",
"0.59693706",
"0.59656715",
"0.59539545",
"0.59057206",
"0.58403915",
"0.5836461",
"0.5835514",
"0.58330756",
"0.5812222",
"0.5781701",
"0.5743904",
"0.57285553",
"0.5721087",
"0.5717715",
"0.56738836",
"0.56656516",
"0.5648275",
"0.56282943",
"0.5620363",
"0.5618905",
"0.5618895",
"0.5616237",
"0.5614189",
"0.55972916",
"0.55845195",
"0.55832666",
"0.556251",
"0.55475384",
"0.5546968",
"0.5529038",
"0.5529038",
"0.5523908",
"0.5523908",
"0.5513339",
"0.5499862",
"0.5495078",
"0.549329",
"0.5475454",
"0.5471685",
"0.5470346",
"0.5459639",
"0.5444097",
"0.5443977",
"0.5442505",
"0.54402936",
"0.5428341",
"0.5424155",
"0.5424155",
"0.54233944",
"0.54233944",
"0.54184705",
"0.54114676",
"0.5408559",
"0.54061073",
"0.5399458",
"0.53920376",
"0.5385313",
"0.5381946",
"0.53814906",
"0.53806055",
"0.53796655",
"0.5374567",
"0.5372288",
"0.5370389",
"0.53656095",
"0.5360086",
"0.53554845",
"0.5353857",
"0.5353501",
"0.53516066",
"0.53499436",
"0.53499436",
"0.5348316"
] |
0.6911476
|
1
|
Wait for the UI to finish processing an action
|
def wait_for_ui_to_update
sleep(UI_UPDATE_DELAY)
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def action_complete() self.finished = true; end",
"def wait_done\n sleep 0.01 until done?\n end",
"def wait; end",
"def wait; end",
"def wait; end",
"def wait\n\tend",
"def waiting; end",
"def waiting; end",
"def wait\n true\n end",
"def wait_for_elements\n super\n @wait_on << :continue_button\n end",
"def wait_until_ready\n # this method may be left unimplemented if that is applicable\n end",
"def wait\n\t\t\t\t@notification.wait\n\t\t\tend",
"def wait\n @future.value\n end",
"def wait\n sleep 0.0001\n end",
"def event_wait_for finishable\n ScriptActionHandler::HandlerResult::waitFor finishable\n end",
"def wait\n raise StateError.new('actor not running!') unless @running\n @future.value\n end",
"def wait\n\t\t\t\t@group.wait\n\t\t\tend",
"def wait\n\t\t\t\t@available.wait\n\t\t\tend",
"def wait_for_callback\n @turnstile.wait unless @paused\n end",
"def submit\r\n assert_exists\r\n submit_form \r\n @o.wait\r\n end",
"def finish\n @action = Action.find(params[:id])\n unless @action.finished\n @action.finished = true\n @action.finished_at = Time.now\n @action.save\n end\n render :partial => 'box'\n end",
"def wait\n 0\n end",
"def finish_turn\n waiting_to_confirm_done = true\n invalid_confirmation = nil\n\n while waiting_to_confirm_done\n DisplayManager.prepare_ingame_display\n show_state\n puts 'Done! Your turn is now over.'\n puts InputManager.input_options({ affirmative: 'Hide my Rack'}, invalid_confirmation)\n invalid_confirmation = nil\n \n response = InputManager.get\n\n if InputManager.affirmative?(response)\n waiting_to_confirm_done = false\n elsif InputManager.negative?(response)\n # do nothing, wait\n else\n invalid_confirmation = response\n end\n end \n\n DisplayManager.prepare_ingame_display\n end",
"def wait_for_callback\n turnstile.wait unless paused\n end",
"def wait\n\t\t\t\tif @count > 0\n\t\t\t\t\tFiber.yield\n\t\t\t\tend\n\t\t\t\t\n\t\t\t\tsucceeded?\n\t\t\tend",
"def wait_for_completion(id: nil, timeout: nil, cancel: false)\n id ||= @id\n\n if cancel\n puts \"Waiting for the action to cancel (hit Ctrl+C to skip)...\"\n else\n puts \"Waiting for the action to complete (hit Ctrl+C to skip)...\"\n end\n \n last_status = false\n can_cancel = false\n \n begin\n ret = HaveAPI::Client::Action.wait_for_completion(\n @client,\n id,\n timeout: timeout,\n ) do |state|\n last_status = state.status\n can_cancel = state.can_cancel?\n \n update_progress(state, cancel)\n end\n\n rescue Interrupt\n @pb && @pb.stop\n puts\n\n cancel_action(timeout: timeout) if can_cancel && !cancel && last_status\n\n puts\n print_help(id)\n exit(false)\n end\n\n if ret\n @pb && @pb.finish\n else\n @pb && @pb.stop\n end\n\n ret\n end",
"def wait_for\n subject = nil\n wait.until { (subject = yield).displayed? }\n sleep(1.2) # FIXME: wait animations to complete\n subject\n end",
"def finish\n pbSEPlay(\"EBDX/SE_Select3\")\n @finished = true\n Input.update\n end",
"def wait_until_ready\n # this method may be left unimplemented if that is applicable log\n end",
"def wait_for_elements\n super\n @wait_on << :return_button\n @wait_on << :circulation_desk_selector\n end",
"def wait_final_result()\n\t\t@queue.pop\n\t\t@appsrc.ready\n\t\treturn @result\n\tend",
"def run_action(msg)\n ui.info(\"#{msg}... \", :nonewline)\n begin\n result = yield\n ui.puts ui.color(\"complete!\", :green, :bold)\n if result\n ui.puts \"---> Results:\"\n case result\n when Hash\n result.each do |k, v|\n ui.puts \" \" << ui.color(\"#{k}: \", :bold) << v\n end\n else\n ui.puts result\n end\n end\n rescue\n ui.puts ui.color(\"error!\", :red, :bold)\n raise\n end\n true\n end",
"def wait_for_message\n message = gets.strip\n publish_message(@current_user, message)\n wait_for_message\n end",
"def completed()\n end",
"def wait(arg0)\n end",
"def wait_until_not_full; end",
"def run_action(msg)\n ui.info(\"#{msg}... \", :nonewline)\n begin\n result = yield\n ui.puts ui.color('complete!', :green, :bold)\n if(result)\n ui.puts '---> Results:'\n case result\n when Hash\n result.each do |k,v|\n ui.puts ' ' << ui.color(\"#{k}: \", :bold) << v\n end\n else\n ui.puts result\n end\n end\n rescue => e\n ui.puts ui.color('error!', :red, :bold)\n ui.error \"Reason - #{e}\"\n raise\n end\n true\n end",
"def wait_release_action(external_user, service_id, role_name, nodes, report)\n if report\n Log.info LOG_COMP, \"Waiting #{nodes} to report ready\"\n rc = wait_report_ready(nodes)\n else\n Log.info LOG_COMP, \"Waiting #{nodes} to be (ACTIVE, RUNNING)\"\n rc = wait(nodes, 'ACTIVE', 'RUNNING')\n end\n\n if rc[0]\n @lcm.trigger_action(:release_cb,\n service_id,\n external_user,\n service_id,\n role_name,\n rc[1])\n else\n @lcm.trigger_action(:deploy_failure_cb,\n service_id,\n external_user,\n service_id,\n role_name)\n end\n end",
"def wait_async\n @wait_async = true\n end",
"def wait\n @notifier.wait if @notifier\n end",
"def wait\n @notifier.wait if @notifier\n end",
"def completed\n end",
"def wait_for_launching\n sleep @delay\n end",
"def wait_for_message\n @message_window.update\n while $game_message.visible \n update_basic\n end\n end",
"def continue\n frm.div(:class=>\"highlightPanel\").span(:id=>\"submitnotifxxx\").wait_while_present\n frm.button(:value=>\"Continue\").click\n end",
"def click_on_end_tour_button\n\n click_button END_TOUR_BUTTON\n sleep(THREAD_SLEEP_1)\n\n end",
"def complete_tagging_task\n wait_for_css(input_elements[:change_tags_button])\n @page.find(input_elements[:change_tags_button]).click\n # wait_until_css_is_gone(input_elements[:tag_input_field])\n end",
"def ai_done_tb\n # print \"\\n\\n\\n~~~~~~~ AI DONE ~~~~~~~~~~ \\nstart_queue=#{TM.start_queue}\\n\\n\\n\"\n \n @wait_on_ai = false # nothing left in queue\n Era::AI.produce_units \n # do_turn_end_evs\n # start_player_turn_tb\n go_next_turn_tb\n end",
"def wait\n @running.reset\n @waiting.set\n @running.wait\n end",
"def waiting_for_response_tb?\n true # done waiting during scene_map otherwise need to wait\n end",
"def wait\n #$_api_queue.clear\n wait_for(/>/)\nend",
"def wait_for_state_transitions\n # don't wait for anything unless told otherwise\n end",
"def finish\r\n #\r\n end",
"def waitQuit()\n @device.waitQuit() ;\n end",
"def finish\n #\n end",
"def finish\n transition_to(:successful) unless transition_to(:waiting_funds)\n end",
"def busy( &block )\n self.connection.status( :dnd, \"Working...\" )\n yield\n self.connection.status( :chat, \"JabberListener waiting for instructions\" )\n end",
"def wait\n sleep WAIT_TIME unless @skip_wait\n end",
"def wait_until\n poll do\n transition! if yield\n end\n end",
"def wait_for_message\r\n Fiber.yield while $game_message.busy?\r\n end",
"def wait\n @thread.join\n end",
"def wait_deploy_action(external_user, service_id, role_name, nodes, report)\n if report\n Log.info LOG_COMP, \"Waiting #{nodes} to report ready\"\n rc = wait_report_ready(nodes)\n else\n Log.info LOG_COMP, \"Waiting #{nodes} to be (ACTIVE, RUNNING)\"\n rc = wait(nodes, 'ACTIVE', 'RUNNING')\n end\n\n if rc[0]\n @lcm.trigger_action(:deploy_cb,\n service_id,\n external_user,\n service_id,\n role_name,\n rc[1])\n else\n @lcm.trigger_action(:deploy_failure_cb,\n service_id,\n external_user,\n service_id,\n role_name)\n end\n end",
"def wait\n @t.each{|t| t.join}\n end",
"def wait_for_update_and_click(element)\n click_element(element, Utils.short_wait)\n end",
"def wait_for_message\n loop do\n message = gets.strip\n publish_message(@current_user, message)\n end\n end",
"def click_wait\n @webpage.click_wait(@locator)\n end",
"def wait!\n sleep(@sleep)\n end",
"def action_done(action)\n dispatch = { :migrate => :done_migrating, :map => :done_mapping, :reduce =>\n :done_reducing, :finalize => :done_finalizing } \n self.send dispatch[action[:action]], action\n end",
"def check_completed\n\t\tend",
"def grabber_done(playlist_id)\n @outfile.close\n @bar_progress.percent = 100\n @button_start.state = 'normal'\n \n filename = \"#{playlist_id}#{OUTFILE_EXT}\"\n show_msg('Done', \"Finished OK. Output to #{filename}\")\nend",
"def op_is_done\n return if @op_notifier.nil?\n @op_notifier.notify(self)\n end",
"def on_completion(actor); end",
"def finish\n #\n end",
"def wait\n self.listener.wait if self.listener\n end",
"def wait_for_player\n wait_character_move_completion 0\n end",
"def wait_for_browser\r\n # NOTE: no need any more\r\n end",
"def finish\n @finish || @break_action\n end",
"def wait_for_press\n wait_for_presses(1)\n end",
"def wait_for_input\n wait_for_input_core = lambda do\n text = %x{#{TM_DIALOG} -w #{@dialog_token} }\n raise WindowNotFound if $CHILD_STATUS == 54528 # -43\n raise \"Error (#{text})\" if $CHILD_STATUS != 0\n\n OSX::PropertyList::load(text)\n end\n\n if block_given? then\n loop do\n should_continue = yield(wait_for_input_core.call)\n break unless should_continue\n end\n else\n wait_for_input_core.call\n end\n end",
"def wait_until_finished\n @active_requests.dup.each do |n|\n n.join\n end\n end",
"def execute(action, data = {})\n data[:action] = action\n result = nil\n Thread.new do\n LOCK.synchronize do\n Thread.main.wakeup\n result = Net::HTTP.post_form(@uri, data).body\n Thread.main.wakeup\n end\n end\n sleep unless LOCK.locked? || result\n GTS.loading_screen&.process\n Graphics.update while LOCK.locked? # Security\n return result\n end",
"def request_ui_refresh\n @ui_refresh_executor.call\n end",
"def begin_analysis\n Thread.new do\n if @type == 'full'\n full_analysis\n elsif @type == 'refresh'\n refresh_analysis\n end\n end\n msg = { :success => true, :message => \"Please wait while we process the repository!\" }\n end",
"def complete!\n status \"Success\"\n self\n end",
"def wait\n waitUntilStarted\n\n @resultsSemaphore.wait\n end",
"def implicit_wait; end",
"def finish\n noth\n end",
"def wait_for_effect\n update_for_wait\n update_for_wait while @spriteset.effect?\n end",
"def enter_result(test_data)\n hide_notifications_bar\n wait_for_element_and_type(result_text_area, test_data[UseOfCollections::RESULT.name])\n end",
"def wait\n if defined? @result\n return @result\n else\n @waiters << Eventlet.current\n Eventlet.sleep\n end\n end",
"def pbWaitMessage\r\n return if !@briefMessage\r\n pbShowWindow(MESSAGE_BOX)\r\n cw = @sprites[\"messageWindow\"]\r\n MESSAGE_PAUSE_TIME.times do\r\n pbUpdate(cw)\r\n end\r\n cw.text = \"\"\r\n cw.visible = false\r\n @briefMessage = false\r\n end",
"def wait\r\n Ragweed::Wrap32::wait_for_single_object @h\r\n end",
"def run_completed\n end",
"def complete\n end",
"def run_actions; end",
"def finish\n end",
"def wait_for_progress\n sleep 2\n wait_for(timeout: 60) { element_does_not_exist \"* marked:'#{loading_screen}'\" }\n end",
"def wait_and_proceed\n Thread.new do\n sleep TIMEOUT\n proceed\n end\n end",
"def action_run\n end",
"def finish\n @executor.shutdown\n @executor.wait_for_termination(0.25)\n end"
] |
[
"0.7182441",
"0.6788431",
"0.66238284",
"0.66238284",
"0.66238284",
"0.66030705",
"0.64908904",
"0.64908904",
"0.6476253",
"0.6194249",
"0.6187958",
"0.6163954",
"0.6137924",
"0.61210084",
"0.61168545",
"0.6092502",
"0.60601664",
"0.6040198",
"0.60121256",
"0.5994139",
"0.59869677",
"0.5965594",
"0.5961089",
"0.59600556",
"0.5945012",
"0.593265",
"0.59181327",
"0.59071827",
"0.58846265",
"0.58838075",
"0.5875272",
"0.58566654",
"0.5852144",
"0.5830866",
"0.5825244",
"0.5809746",
"0.5804831",
"0.5778213",
"0.5776096",
"0.5760086",
"0.5760086",
"0.5755023",
"0.575312",
"0.57372725",
"0.57040435",
"0.5702265",
"0.5684871",
"0.5680687",
"0.5680186",
"0.5678853",
"0.56775665",
"0.5674776",
"0.56634146",
"0.56582886",
"0.5655388",
"0.5639234",
"0.56376004",
"0.56292653",
"0.562875",
"0.5623208",
"0.56227916",
"0.5621992",
"0.56192476",
"0.56067806",
"0.55950934",
"0.5583335",
"0.5572669",
"0.55722165",
"0.55680627",
"0.5564974",
"0.55549204",
"0.5552702",
"0.5548892",
"0.5546233",
"0.5539391",
"0.5537438",
"0.5536264",
"0.5533775",
"0.55222154",
"0.552167",
"0.5521379",
"0.55199444",
"0.55173635",
"0.5515314",
"0.55056244",
"0.5505525",
"0.5503349",
"0.55024076",
"0.5500065",
"0.5497344",
"0.5492651",
"0.5489559",
"0.54892516",
"0.5484952",
"0.54707634",
"0.54610085",
"0.5457873",
"0.54511017",
"0.5448325",
"0.5437095"
] |
0.6586998
|
6
|
Similar to wait_until but won't fail if the condition is never true. wait_until is prefered, but this can be useful if the condition might not occur.
|
def wait_until_without_failing(timeout=UI_UPDATE_DELAY)
start = Time.now
while ((Time.now - start) <= timeout)
break if yield
sleep(WaitHelpers::LOOP_DELAY)
end
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def wait_until times: 5, delay: 1, &condition\n times.times do\n return if condition.call\n sleep delay\n end\n raise \"Condition not met. Waited #{times} times with #{delay} sec delay\"\n end",
"def wait_until_true(*args, &block)\n Selenium::WebDriver::Wait.new(*args).until(&block)\n end",
"def wait_until_not_full; end",
"def wait_until(options = {}, &block)\n eventually(options, &block)\n end",
"def wait_for(condition_name, max_wait_time: Capybara.default_max_wait_time,\n polling_interval: 0.01)\n wait_until = Time.current + max_wait_time.seconds\n loop do\n break if yield\n\n raise \"Condition not met: #{condition_name}\" if Time.current > wait_until\n\n sleep(polling_interval)\n end\n end",
"def wait_until(timeout=10, &block)\n time = Time.now\n success = false\n until success\n if (Time.now - time) >= timeout\n raise \"Waited for #{timeout} seconds, but block never returned true\"\n end\n sleep 0.5\n success = yield\n end\n end",
"def wait_for_condition(timeout, step = 0.1, &block)\n defer{ wait_for_condition_sync(timeout, step, &block) }\n end",
"def wait_while(timeout: nil, message: nil, interval: nil, **opt, &blk)\n message ||= proc { |obj| \"waiting for false condition on #{obj.inspect}\" }\n\n # TODO: Consider throwing argument error for mixing block & options\n proc = create_proc(opt, &blk)\n\n Wait.while(timeout: timeout, message: message, interval: interval, object: self, &proc)\n\n self\n end",
"def wait_for timeout = 0, &condition\n if condition\n SeleniumAdapter.wait_for(timeout, &condition)\n return nil\n else\n @wait_for = timeout\n return self\n end\n end",
"def waitUntil\n until yield\n sleep 0.5\n end\nend",
"def wait_until\n until yield\n wait\n end\n end",
"def wait_until timeout=10, &block\n wait = Selenium::WebDriver::Wait.new(:timeout => timeout)\n wait.until &block\nend",
"def wait_until\n until yield\n\twait\n end\n end",
"def wait_until\n Timeout.timeout(Capybara.default_wait_time) do\n loop until yield\n end\n end",
"def call_until(caller_object, condition, body)\n until_lambda = lambda_or_raise(caller_object, :until)\n until_lambda.call(condition, body)\n end",
"def wait_until\n require 'timeout'\n Timeout.timeout(Capybara.default_wait_time) do\n sleep(0.1) until (value = yield)\n value\n end\n end",
"def wait_for(waiting_message = 'something to happen')\n 30.times do\n ret = yield\n return ret if ret\n sleep 0.3\n end\n assert false, \"waiting for #{waiting_message} was not successful\"\n end",
"def wait_until(ev)\n if inside_control?\n raise ThreadMismatch, \"cannot use #wait_until in execution threads\"\n end\n\n Roby.condition_variable(true) do |cv, mt|\n caller_thread = Thread.current\n # Note: no need to add the caller thread in waiting_threads,\n # since the event will become unreachable if the execution\n # thread quits\n\n mt.synchronize do\n done = false\n once do\n ev.if_unreachable(true) do |reason, event|\n mt.synchronize do\n done = true\n caller_thread.raise UnreachableEvent.new(event, reason)\n end\n end\n ev.on do |ev|\n mt.synchronize do\n done = true\n cv.broadcast\n end\n end\n yield if block_given?\n end\n\n while !done\n cv.wait(mt)\n end\n end\n end\n end",
"def wait_until\n until yield\n wait\n end\n end",
"def wait_until(timeout = 30, retry_interval = 0.1, &block)\n start = Time.now\n while (result = !block.call)\n break if (Time.now - start).to_i >= timeout\n sleep(retry_interval)\n end\n !result\nend",
"def wait_until\n require \"timeout\"\n Timeout.timeout(Capybara.default_wait_time + 5) do\n sleep(0.1) until value = yield\n value\n end\n end",
"def wait_for_condition script, timeout\r\n command 'waitForCondition', script, timeout\r\n end",
"def wait_for_condition script, timeout\r\n command 'waitForCondition', script, timeout\r\n end",
"def wait_until\n poll do\n transition! if yield\n end\n end",
"def wait_for_condition(timeout)\n deadline = Process.clock_gettime(Process::CLOCK_MONOTONIC) + timeout\n loop do\n break if yield ||\n Process.clock_gettime(Process::CLOCK_MONOTONIC) > deadline\n\n sleep 1\n end\n end",
"def wait_until(timeout=20, &block)\n time_to_stop = Time.now + timeout\n until yield do\n sleep(0.1) # much less cpu stress\n break if Time.now > time_to_stop\n end\nend",
"def wait_until iterations = 100\n count = 0\n loop do\n raise \"criteria not met\" if count >= iterations\n break if yield\n sleep 0.0001\n count += 1\n end\n end",
"def wait_until\n Timeout.timeout(Capybara.default_max_wait_time) do\n loop until yield\n end\nend",
"def wait_until_blocked(timeout=nil)\n @mutex.synchronize do\n return true unless @blocked == NOT_YET\n\n start = Time.now\n time_to_stop = timeout ? (start + timeout) : nil\n\n logger.debug { \"#{__method__} @blocked: #{@blocked.inspect} about to wait\" } \n @cond.wait(timeout)\n\n if (time_to_stop and (Time.now > time_to_stop)) and (@blocked == NOT_YET)\n return nil\n end\n\n (@blocked == NOT_YET) ? nil : true\n end\n end",
"def wait_until_blocked(timeout=nil)\n @mutex.synchronize do\n return true unless @blocked == NOT_YET\n\n start = Time.now\n time_to_stop = timeout ? (start + timeout) : nil\n\n logger.debug { \"#{__method__} @blocked: #{@blocked.inspect} about to wait\" }\n @cond.wait(timeout)\n\n if (time_to_stop and (Time.now > time_to_stop)) and (@blocked == NOT_YET)\n return nil\n end\n\n (@blocked == NOT_YET) ? nil : true\n end\n end",
"def wait_for_no_field_value(locator, expected_value, options={})\n builder = JavascriptExpressionBuilder.new\n builder.find_element(locator).element_value_is_not(expected_value)\n wait_for_condition builder.script, options[:timeout_in_seconds]\n end",
"def wait_for_no_element(locator, options={})\n builder = JavascriptExpressionBuilder.new\n builder.find_element(locator).append(\"element == null;\")\n wait_for_condition builder.script, options[:timeout_in_seconds]\n end",
"def wait_until(wait_time = Capybara.default_max_wait_time)\n Timeout.timeout(wait_time) do\n loop until yield\n end\n end",
"def wait_until(wait_time = Capybara.default_max_wait_time)\n Timeout.timeout(wait_time) do\n loop until yield\n end\n end",
"def unless_condition; end",
"def wait_until_true timeout=3, interval=0.1\n time_limit = Time.now + timeout\n loop do\n result = yield\n return if result || Time.now >= time_limit\n sleep interval\n end\n end",
"def wait_until_element_present(element)\n element.when_present.exists?\n end",
"def fail_script_unless(condition, message)\n fail_script message unless condition\n end",
"def fail_script_unless(condition, message)\n fail_script message unless condition\n end",
"def wait_until_available\n return unless @locked\n\n @mutex.lock\n @mutex.unlock\n end",
"def wait_while\n while yield\n\twait\n end\n end",
"def wait_for(timeout = DEFAULT_TIMEOUT)\n Selenium::WebDriver::Wait.new(:timeout => timeout).until {yield}\nend",
"def wait(timeout = nil)\n begin\n Timeout::timeout timeout do\n @mutex.synchronize do\n @conditional.wait @mutex if @count > 0\n end\n end\n true\n rescue Timeout::Error\n false\n end\n end",
"def wait(timeout = nil)\n return true if set?\n\n @mutex.synchronize { @waiters << Thread.current }\n return true if set? # if event was set while waiting for mutex\n\n if timeout.nil?\n slept = sleep\n else\n slept = sleep(timeout)\n end\n rescue\n # let it fail\n ensure\n @mutex.synchronize { @waiters.delete(Thread.current) }\n return set?\n end",
"def sleep_until(timeout = QUE_SLEEP_UNTIL_TIMEOUT)\n deadline = Time.now + timeout\n loop do\n break if yield\n if Time.now > deadline\n puts \"sleep_until timeout reached!\"\n raise \"sleep_until timeout reached!\"\n end\n sleep 0.01\n end\nend",
"def wait_condition(name, &block)\n resource = WaitCondition.new\n resource.evaluate &block\n\n if resource.handle.nil?\n handle_name = \"#{name}Handle\"\n\n if not resources.include? handle_name\n wait_condition_handle handle_name\n end\n\n resource.handle = handle_name\n end\n\n add_resource(name, resource)\n end",
"def wait(name)\n raise \"cannot wait for signals while exclusive\" if Celluloid.exclusive?\n\n @conditions[name] ||= Condition.new\n @conditions[name].wait\n end",
"def wait_until_available\n return unless locked?\n\n @mutex.synchronize {}\n end",
"def waitForElementPresent(how, what, waitTime = THE_TIMES)\n sleep 1\n puts waitTime\n if isElementPresent?(how, what)\n return true\n else\n raise Selenium::WebDriver::Error::NoSuchElementError\n end\n sleep 1\n end",
"def wait_while\n while yield\n wait\n end\n end",
"def wait_for(element, msg = \"Error waiting for element #{element}\")\n wait = Selenium::WebDriver::Wait.new(:timeout => WAIT_TIME)\n wait.until {\n begin\n send(element)\n true\n rescue Selenium::WebDriver::Error::NoSuchElementError, NoMethodError => e\n puts msg\n false\n end\n }\n end",
"def wait_true max_wait=30, interval=0.5, &block\n max_wait = 1 if max_wait <= 0\n result = nil\n timeout max_wait do\n until (result = begin; block.call; rescue; end)\n sleep interval\n end\n end\n result\n end",
"def wait_until_ready\n # this method may be left unimplemented if that is applicable\n end",
"def wait(what = T.unsafe(nil)); end",
"def wait_for_element(locator, options={})\n builder = JavascriptExpressionBuilder.new\n builder.find_element(locator).append(\"element != null;\")\n wait_for_condition builder.script, options[:timeout_in_seconds]\n end",
"def wait\n sleep WAIT_TIME unless @skip_wait\n end",
"def waiting?\n @status[:description] == :wait\n end",
"def wait_until(cycle)\n until yield\n\t sleep(cycle)\n\tend\n end",
"def wait timeout = nil\n @mutex.synchronize {\n if @cnt != 0\n if timeout\n @cond_var.wait @mutex, timeout\n else\n @cond_var.wait @mutex\n end\n end\n }\n end",
"def wait_for(seconds = timeout)\n Selenium::WebDriver::Wait.new(timeout: seconds).until { yield }\n end",
"def wait_for(wait_max: 3, step: 0.001, &block)\n stop_at = wait_max.seconds.from_now\n\n sleep step while !block.call && (@time = Time.now) < stop_at\n\n fail \"Timeout of #{wait_max} seconds exceeded!\" unless @time < stop_at\nend",
"def wait_till_not_ready(user, seconds)\n res = nil\n iterations = 0\n start_time = monotonic_seconds\n\n success = wait_for(seconds) {\n res = ready?(user: user, quiet: true)\n\n logger.info res[:command] if iterations == 0\n iterations = iterations + 1\n\n ! res[:success]\n }\n\n duration = monotonic_seconds - start_time\n logger.info \"After #{iterations} iterations and #{duration.to_i} \" <<\n \"seconds:\\n#{res[:response]}\"\n\n res[:success] = success\n return res\n end",
"def wait_while\n while yield\n wait\n end\n end",
"def wait_while_present\n container.wait_while_present\n end",
"def Wait_For_Element(locator_type, locator_value)\n for iSecond in 0..$config['Longwait']\n sleep 1\n if(@driver.find_element(\"#{locator_type}\", \"#{locator_value}\").displayed?)\n\tbreak\n else \t\n if (iSecond >= $config['Longwait']) \n return false\n @driver.find_element(\"#{locator_type}\", \"#{locator_value}\").displayed?\n break\n end\n end \n end\n return true\nend",
"def wait_for(seconds)\n Selenium::WebDriver::Wait.new(timeout: seconds).until { yield }\nend",
"def wait_until(type, *find_args)\n wait = Wait.new(timeout: Appom.max_wait_time)\n wait.until do\n case type\n # Function only return true if element enabled or raise an error if time out\n when 'element enable'\n _find(*find_args).enabled?\n # Function only return true if element disabled or raise an error if time out\n when 'element disable'\n !_find(*find_args).enabled?\n # Function only return true if we can find at leat one element (array is not empty) or raise error\n when 'at least one element exists'\n args, text, visible = deduce_element_args(find_args)\n !page.find_elements(*args).empty?\n # Function only return true if we can't find at leat one element (array is empty) or raise error\n when 'no element exists'\n args, text, visible = deduce_element_args(find_args)\n page.find_elements(*args).empty?\n end\n end\n end",
"def wait_until_has_no_selector(selector_expr)\n element_present = TRUE\n time_before = Time.now\n begin\n while element_present\n fail \"\\nERROR: Element can still be found even after #{Time.now - time_before} secs\\n\" if (Time.now - time_before) > PAGE_ELEMENT_TIMEOUT_SECS\n element_present = FALSE if !@page.has_selector?(selector_expr)\n log(\"\\n\\nExample #{$example_counter}: WAIT_UNTIL_HAS_NO_SELECTOR(#{selector_expr}): elapsed_time of CAPYBARA 'has_selector' call = #{Time.now - time_before} secs\\n\") if ENABLED_LOGGING\n sleep 0.5 if element_present\n end\n rescue Selenium::WebDriver::Error::StaleElementReferenceError => e\n puts \"Trapped Exception: #{e} : Retrying\"\n sleep 0.5\n retry\n rescue Selenium::WebDriver::Error::ObsoleteElementError => e\n puts \"Trapped Exception: #{e} : Retrying\"\n sleep 0.5\n retry\n end\n end",
"def signal_wait_until(pr, &block)\n #NOTE: busy waiting!!!\n while true do\n torrent = yield\n break if pr.call torrent\n end\n end",
"def wait(duration, variable)\n for i in 0...duration\n @wait_time += 1 if variable\n @wait_time_thirst if variable == false \n break if i >= duration / 2\n end\n end",
"def wait_until(msg = nil)\n if block_given?\n msg_thread, work_thread = nil, nil\n \n # prints incremental '...' message while waiting for the given block to finish successfully\n if msg\n print \"#{msg.to_s}...\"\n msg_thread = Thread.new do\n loop do\n print \".\"\n $stdout.flush\n sleep 1\n end\n end\n end\n \n # repeatedly yields to the given block until it returns true\n work_thread = Thread.new do\n result = false\n until result\n result = yield\n sleep 1\n end\n end\n \n work_thread.join if work_thread\n msg_thread.kill if msg_thread\n \n print \"\\n\" if msg\n end\n end",
"def wait_until_with_buffer(args, &block)\n original_timeout = args[:timeout] || ENV['WAIT_TIMEOUT'].to_i\n args_buffered = args.dup\n\n args_buffered[:timeout] = 60\n\n start_time = Time.now\n Frank::Cucumber::WaitHelper.wait_until(args_buffered) { block.call() }\n end_time = Time.now\n\n delta = end_time - start_time\n puts(\"wait_until exceeded timeout #{original_timeout}. Took #{delta}. #{caller[0]}\") if delta > original_timeout\nend",
"def wait timeout: 3, &block\n wait = Selenium::WebDriver::Wait.new timeout: timeout\n wait.until(&block)\nend",
"def wait_until(type, *find_args)\n wait = Wait.new(timeout: Appom.max_wait_time)\n wait.until do\n case type\n # Function only return true if element enabled or raise an error if time out\n when 'element enable'\n _find(*find_args).enabled?\n # Function only return true if element disabled or raise an error if time out\n when 'element disable'\n result = _find(*find_args)\n if result.enabled?\n raise StandardError, \"Still found an element enable with args = #{find_args}\"\n end\n return true\n # Function only return true if we can find at least one element (array is not empty) or raise error\n when 'at least one element exists'\n result = _all(*find_args)\n if result.empty?\n raise StandardError, \"Could not find any elements with args = #{find_args}\"\n end\n return true\n\n # Function only return true if we can't find at least one element (array is empty) or raise error\n when 'no element exists'\n result = _all(*find_args)\n if !result.empty?\n if result.size > 1\n raise StandardError, \"Still found #{result.size} elements with args = #{find_args}\"\n else\n raise StandardError, \"Still found #{result.size} element with args = #{find_args}\"\n end\n end\n return true\n end\n end\n end",
"def wait_until_ready\n # this method may be left unimplemented if that is applicable log\n end",
"def wait_for_less_busy_worker(val = T.unsafe(nil)); end",
"def wait_for(event)\n @mootex.synchronize do\n (@waiting[event] ||= ConditionVariable.new).wait(@mootex)\n end\n self\n end",
"def brute_wait(delay)\n sleep(delay)\n end",
"def wait_for_element(element)\n @wait.until {\n bool = false\n\n if(element.displayed?)\n bool = true\n @element = element\n break\n end \n\n bool == true\n }\n end",
"def waiting? #:nodoc:\n false|@waiting\n end",
"def wait(duration, value)\n for i in 0...duration\n @wait_time += 1 if value == false\n @wait_time2 += 1 if value\n break if i >= duration / 2\n end\n end",
"def wait(duration, value)\n for i in 0...duration\n @wait_time += 1 if value == false\n @wait_time2 += 1 if value\n break if i >= duration / 2\n end\n end",
"def wait_until_zero(js)\n wait_until { page.evaluate_script(js).zero? }\n end",
"def waiting?\n entered_at.nil?\n end",
"def wait\n # Here we use a loop-sleep combination instead of using\n # ThreadPoolExecutor's `wait_for_termination`. See issue #21 for more\n # information.\n loop do\n break if @executor.shutdown?\n sleep 0.1\n end\n end",
"def is_new_and_others_wait? waiter\n waiter.nil? && !@awaiting.empty?\n end",
"def wait_until(states)\n Timeout.timeout(timeout, Errors::Backend::EntityTimeoutError) do\n loop do\n sleep waiter_step\n refresher.call waitee\n early_fail!\n break if wanted?(states)\n end\n end\n\n yield waitee if block_given?\n end",
"def wait(timeout = nil)\n result = true\n c = nil\n @mutex.synchronize do\n @count -= 1\n if @count < 0\n @sleeping.push Thread.current\n @mutex.sleep(timeout)\n end\n end\n if timeout\n # If we had a timeout we may have woken due to it expiring rather than\n # due to signal being called. In that case we need to remove ourself from the sleepers.\n @mutex.synchronize do\n i = @sleeping.index(Thread.current)\n if i\n @count += 1\n @sleeping.delete_at(i)\n result = false\n end\n end\n end\n result\n end",
"def condition &block\n return unless block\n\n case @deferred_status\n when :succeeded, :failed\n SetCallbackFailed.new\n else\n @conditions ||= []\n @conditions.unshift block\n end\n end",
"def ignore_wait_alert_accept_if_exist\n begin\n wait_alert_accept_if_exist\n rescue\n # nothing\n end\n end",
"def wait\n @notifier.wait if @notifier\n end",
"def wait\n @notifier.wait if @notifier\n end",
"def condition\n expect :if\n self[1]\n end",
"def wait_for_not_visible(locator, options={})\n builder = JavascriptExpressionBuilder.new\n wait_for_condition builder.not_visible(locator).script, options[:timeout_in_seconds]\n end",
"def wait_until_get_not_empty(*find_args)\n wait = Wait.new(timeout: Appom.max_wait_time)\n wait.until do\n result = page.find_elements(*find_args)\n # If response is empty we will return false to make it not pass Wait condition\n if result.empty?\n raise StandardError, \"Can not found any elements with args = #{find_args}\"\n end\n # Return result\n return result\n end\n end",
"def wait_until_unhealthy!\n agent = consul.get(\"/agent/self\")[\"Member\"][\"Name\"]\n consul.get_while(\"/health/node/#{agent}\") do |data|\n status = data.detect {|x| x[\"CheckID\"] == \"service:#{name}\" }[\"Status\"]\n status == 'passing'\n end\n end",
"def check_until(created_at)\n return true if @opts[:until].nil?\n\n until_time = created_at + @opts[:until] * DAY\n\n until_time > Time.now\n end",
"def wait(arg0)\n end",
"def wait_for_element(element)\n wait = Selenium::WebDriver::Wait.new(timeout: 60)\n wait.until { find_element(element) }\n end",
"def refute(condition, message='')\n assert(!condition, message)\n end"
] |
[
"0.7180849",
"0.6953944",
"0.6926367",
"0.690409",
"0.69028425",
"0.681789",
"0.6814156",
"0.6801709",
"0.67515063",
"0.67208886",
"0.6720026",
"0.67182094",
"0.67060846",
"0.67034185",
"0.66877306",
"0.6651108",
"0.66334397",
"0.6567752",
"0.6567567",
"0.6545148",
"0.6474974",
"0.6447293",
"0.6447293",
"0.64336854",
"0.64268285",
"0.6413692",
"0.6313837",
"0.63061386",
"0.6248234",
"0.6207434",
"0.6193891",
"0.6188947",
"0.6184964",
"0.6184964",
"0.6152945",
"0.6129276",
"0.6091753",
"0.60902",
"0.60902",
"0.6081096",
"0.60795844",
"0.60754144",
"0.6070891",
"0.6056203",
"0.6024015",
"0.60236645",
"0.6017395",
"0.5986792",
"0.5966946",
"0.5965033",
"0.5956159",
"0.59549946",
"0.5952667",
"0.5937875",
"0.59336627",
"0.5928324",
"0.5919629",
"0.59157634",
"0.5888102",
"0.58816916",
"0.58628815",
"0.5861615",
"0.5858535",
"0.5822751",
"0.5818201",
"0.5816874",
"0.5810348",
"0.5796739",
"0.5792328",
"0.5790098",
"0.57746756",
"0.5770853",
"0.57650733",
"0.5762253",
"0.5750684",
"0.57373625",
"0.57341146",
"0.5733962",
"0.5731942",
"0.57281315",
"0.57259953",
"0.57259953",
"0.5724488",
"0.5723744",
"0.57225126",
"0.5705695",
"0.5681633",
"0.5643024",
"0.56404734",
"0.5631217",
"0.56238854",
"0.56238854",
"0.5619187",
"0.56175476",
"0.56122917",
"0.56076944",
"0.5603968",
"0.5598422",
"0.5585734",
"0.5584217"
] |
0.665238
|
15
|
Expected options: :output_dir: A folder to dump filtered sql files into. This folder will be emptied before running the dump. :mysql_options: Command line options to pass mysql (e.g. u root some_db_name). :db_connection: An active record connection to the database to dump. :column_filters: A hash of column types to filter mapped to an array of values to filter on. :table_filters: An array of tables names that should have all data filtered out.
|
def initialize(options = {})
self.output_dir = options[:output_dir]
self.mysql_options = options[:mysql_options]
self.db_connection = options[:db_connection]
self.column_filters = (options[:column_filters] || {}).stringify_keys!
self.table_filters = (options[:table_filters] || []).map(&:to_s)
self.post_dump_command = options[:post_dump_command]
self.file_extension = options[:file_extension] || ".sql"
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def _dump_tables(opts)\n if opts[:schema]\n _literal_table_sort(tables(opts.merge(:qualify=>true)))\n else\n tables(opts).sort\n end\n end",
"def dump(db = config[\"database\"], data_selection: {})\n file_namer = Postgres::FileNameGenerator.new(working_dir)\n exclude_tables_args = data_selection.keys.map { |table| %Q|--exclude-table-data=\"#{table}\"| }\n\n [\n cli.dump(file_namer.next(db, :schema), db, [exclude_tables_args]),\n *dump_partial_selected_data(db, file_namer, data_selection)\n ]\n end",
"def dump_table(table)\n conditions = conditions_for(table)\n\n cmd = \"mysqldump #{ mysql_options } --tables #{ table }\"\n cmd += \" \\\"--where=#{ conditions }\\\"\" if conditions.present?\n\n if post_dump_command\n cmd += \"| #{post_dump_command}\"\n end\n\n cmd += \" > #{ output_dir }/#{ table }#{file_extension}\"\n\n system(cmd)\n end",
"def extract(outfile_path = 'dbextractor_output.sql')\n @outfile = open(outfile_path, 'w')\n puts \"Loading config...\\n\"\n dbconfig = parse_db_config(@config_path)\n\n extract_tables(dbconfig[\"tables\"])\n @outfile.close\n end",
"def dump\n check_dependencies('mysqldump', 'bzip2', '/bin/sh')\n options = {}\n while arg = args.shift\n case arg\n when '-f', '--force'\n options[:force] = true\n when /^[^-]/\n raise CommandFailed, \"too many arguments passed\" if options[:filename]\n options[:filename] = arg\n else\n raise CommandFailed, \"unsupported option: #{arg}\"\n end\n end\n\n options[:filename] ||= \"#{app}-#{Time.now.strftime('%Y-%m-%d')}.sql.bz2\"\n options[:filename] += '.sql.bz2' unless options[:filename] =~ /\\.sql(\\.bz2)?$/\n options[:filename] += '.bz2' unless options[:filename] =~ /\\.bz2$/\n\n if File.exists?(options[:filename]) && !options[:force]\n raise CommandFailed, \"file already exists. use --force to override.\"\n end\n\n exec('/bin/sh', '-c',\n \"mysqldump --compress --single-transaction #{args_to_s(mysql_args(database_uri))}\" +\n pv_pipe +\n %{| bzip2 > '#{options[:filename]}'})\n end",
"def dump!\n MODEL_TO_OUTPUT_FILEMAPPING.each do |klass, filepath|\n write_table_to_file(klass, filepath, %w[created_at updated_at])\n end\n end",
"def dump_schema_migration(options=OPTS)\n options = options.dup\n if options[:indexes] == false && !options.has_key?(:foreign_keys)\n # Unless foreign_keys option is specifically set, disable if indexes\n # are disabled, as foreign keys that point to non-primary keys rely\n # on unique indexes being created first\n options[:foreign_keys] = false\n end\n\n ts = sort_dumped_tables(_dump_tables(options), options)\n skipped_fks = if sfk = options[:skipped_foreign_keys]\n # Handle skipped foreign keys by adding them at the end via\n # alter_table/add_foreign_key. Note that skipped foreign keys\n # probably result in a broken down migration.\n sfka = sfk.sort.map{|table, fks| dump_add_fk_constraints(table, fks.values)}\n sfka.join(\"\\n\\n\").gsub(/^/, ' ') unless sfka.empty?\n end\n\n <<END_MIG\nSequel.migration do\n change do\n#{ts.map{|t| dump_table_schema(t, options)}.join(\"\\n\\n\").gsub(/^/, ' ')}#{\"\\n \\n\" if skipped_fks}#{skipped_fks}\n end\nend\nEND_MIG\n end",
"def dump( out = Dumper.new )\n out.dump \"table #{out_name}\" do\n for column in columns\n column.dump( out )\n end\n for index in indexes\n index.dump( out )\n end\n for foreign_key in foreign_keys\n foreign_key.dump( out )\n end\n end\n end",
"def perform!\n super\n\n pipeline = Pipeline.new\n dump_ext = 'sql'\n\n pipeline << mysqldump\n\n model.compressor.compress_with do |command, ext|\n pipeline << command\n dump_ext << ext\n end if model.compressor\n\n pipeline << \"#{ utility(:cat) } > \" +\n \"'#{ File.join(dump_path, dump_filename) }.#{ dump_ext }'\"\n\n pipeline.run\n if pipeline.success?\n log!(:finished)\n else\n raise Error, \"Dump Failed!\\n\" + pipeline.error_messages\n end\n end",
"def psql_db_batch__cli_or_apply_dumps psql_db, db_dump_paths=[nil], options=\"\"\n batch = db_dump_paths.map { |db_dump_path|\n cli = psql_db_command__cli psql_db\n [cli, db_dump_path && \" #{options} -f #{quoted_shell_param db_dump_path}\"]\n }\n end",
"def make_mysql_backup\n if @all_databases\n options = {\n :name => \"--all-databases\",\n :dump_options => \"\",\n :append_name => \"\"\n }\n file_name = mysqldump(options)\n compress_file(file_name)\n end\n if @databases && !@databases.empty?\n @databases.each do |db|\n options = {\n :name => db[:name].to_s,\n :dump_options => db[:dump_options].to_s,\n :append_name => db[:append_name].to_s\n }\n file_name = mysqldump(options)\n compress_file(file_name)\n end\n end\nend",
"def perform!\n super\n\n dump = \"echo '.dump' | #{sqlitedump_utility} #{path}\"\n\n pipeline = Pipeline.new\n dump_ext = \"sql\".dup\n\n pipeline << dump\n if model.compressor\n model.compressor.compress_with do |command, ext|\n pipeline << command\n dump_ext << ext\n end\n end\n\n pipeline << \"cat > '#{File.join(dump_path, dump_filename)}.#{dump_ext}'\"\n\n pipeline.run\n\n if pipeline.success?\n log!(:finished)\n else\n raise Error,\n \"#{database_name} Dump Failed!\\n\" + pipeline.error_messages\n end\n end",
"def run\n tables.each do |table|\n dump_table(table)\n end\n end",
"def table_options_sql(options)\n\t sql = []\n\t sql << flag_option_sql(options, :parallel)\n\t sql << flag_option_sql(options, :logging)\n\t sql << flag_option_sql(options, :monitoring)\n\t sql << \"TABLESPACE #{quote_identifier(options[:tablespace])}\" if options[:tablespace]\n\t sql << compress_option_sql(options)\n\t sql << options[:options] if String === options[:options]\n\t sql.compact.join ' '\n\t end",
"def dump_schema_migration(options={})\n ts = tables(options)\n <<END_MIG\nSequel.migration do\n up do\n#{ts.sort_by{|t| t.to_s}.map{|t| dump_table_schema(t, options)}.join(\"\\n\\n\").gsub(/^/o, ' ')}\n end\n \n down do\n drop_table(#{ts.sort_by{|t| t.to_s}.inspect[1...-1]})\n end\nend\nEND_MIG\n end",
"def pg_dump_args\n ['-x', '-O', if schema_only\n '-s'\n end, conf['database']].compact.join(' ')\n end",
"def prep_command\n \"mysqldump #{dump_options} -u #{db_user} --password=#{db_password} #{database} > #{remote_backup_path}\"\n end",
"def filter_options options\n other_tables = options[:other_tables].dup || []\n temp_options = options.dup\n temp_options.keep_if do |k,v|\n ['name','page','search','order'].include? k.to_s\n end\n other_tables << temp_options\n end",
"def psql_db_batch__cli_or_generate_dumps psql_db, db_dump_paths=[nil], options=\"\"\n psql_db = array__from(psql_db)\n db_dump_paths = array__from(db_dump_paths)\n batch = db_dump_paths.map { |db_dump_path|\n cli = psql_db_command__cli psql_db\n pg_dump = psql_db_command__dump psql_db\n program = db_dump_path && pg_dump || cli\n [program, db_dump_path && \" #{options} -f #{quoted_shell_param db_dump_path}\"]\n }\n end",
"def sort_dumped_tables(tables, options=OPTS)\n if options[:foreign_keys] != false && supports_foreign_key_parsing?\n table_fks = {}\n tables.each{|t| table_fks[t] = foreign_key_list(t)}\n # Remove self referential foreign keys, not important when sorting.\n table_fks.each{|t, fks| fks.delete_if{|fk| fk[:table] == t}}\n tables, skipped_foreign_keys = sort_dumped_tables_topologically(table_fks, [])\n options[:skipped_foreign_keys] = skipped_foreign_keys\n tables\n else\n tables\n end\n end",
"def mysql_union_usage\n puts \"List of available commands and general description\".light_yellow + \": \".white\n puts \"back \".light_yellow + \" => \".white + \"Return to Main Menu\".light_red\n puts \"basic\".light_yellow + \" => \".white + \"Get Basic Info (User, Version, etc)\".light_red\n puts \"dbs\".light_yellow + \" => \".white + \"Get Available Database Names\".light_red\n puts \"tables\".light_yellow + \" => \".white + \"Get Tables in Current DB\".light_red\n puts \"dbtables\".light_yellow + \" => \".white + \"Get Tables in Another DB\".light_red\n puts \"tcolumns\".light_yellow + \" => \".white + \"Find Columns for Table in Current DB\".light_red\n puts \"dbcolumns\".light_yellow + \" => \".white + \"Find Columns for Table in Another DB\".light_red\n puts \"tdump\".light_yellow + \" => \".white + \"Dump Table from current DB\".light_red\n puts \"dbdump\".light_yellow + \" => \".white + \"Dump Table from another DB\".light_red\n puts \"passwords\".light_yellow + \" => \".white + \"Dump DBMS Usernames & Passwords (privileged)\".light_red\n # TBD => Add Search Option (By DB Name, Table Name, Column Name, Custom)\n # User can decide if it is run using LIKE statement, REGEXP or an EQUAL comparison\n # i.e. where column='password'; where username like %admin% or %super% or %moderator%;\n puts \"fuzz_tables\".light_yellow + \" => \".white + \"Fuzz Tables in DB\".light_red\n puts \"fuzz_cols\".light_yellow + \" => \".white + \"Fuzz Columns in Table\".light_red\n puts \"read\".light_yellow + \" => \".white + \"Read Files via load_file() (privileged)\".light_red\n puts \"write\".light_yellow + \" => \".white + \"Write Files via INTO OUTFILE (privileged)\".light_red\n puts \"save \".light_yellow + \" => \".white + \"Save Basic Injection Info to Results directory\".light_red\n print_line(\"\")\n end",
"def collect_args\n Optimist.options do\n opt(\n :work_dir,\n 'REQUIRED: Specify your work/build directory, where the SQLite and .csv files will be built',\n type: :string, required: true, short: '-w'\n )\n opt(\n :country,\n 'Specify the country code for processing, or all countries if not specified',\n type: :string, required: false, short: '-f'\n )\n opt(\n :generate_files,\n 'Generate CSV files: [counties.csv, states.csv, countries.csv, zipcodes.csv]',\n type: :boolean, required: false, short: '-g', default: false\n )\n opt(\n :country_tablename,\n 'Specify the name for the `countries` table',\n type: :string, required: false, default: 'countries'\n )\n opt(\n :state_tablename,\n 'Specify the name for the `states` table',\n type: :string, required: false, default: 'states'\n )\n opt(\n :county_tablename,\n 'Specify the name for the `counties` table',\n type: :string, required: false, default: 'counties'\n )\n opt(\n :zipcode_tablename,\n 'Specify the name for the `zipcodes` table',\n type: :string, required: false, default: 'zipcodes'\n )\n opt(\n :clobber,\n 'Overwrite existing files',\n type: :boolean, required: false, short: '-c', default: false\n )\n opt(\n :dry_run,\n 'Do not actually move or copy files',\n type: :boolean, required: false, short: '-d',\n default: false\n )\n opt(\n :verbose,\n 'Be verbose with output',\n type: :boolean, required: false, short: '-v',\n default: false\n )\n end\n end",
"def psql_db_dump_replacer__for_psql_db__sample_example\n [\n psql_db__sample_example,\n [\n \"/tmp/psql_db_original_dump\"\n ],\n [\n \"/tmp/database_dump\"\n ],\n \"ON_ERROR_STOP=off\",\n ]\n end",
"def dump_table_schema(table, options=OPTS)\n gen = dump_table_generator(table, options)\n commands = [gen.dump_columns, gen.dump_constraints, gen.dump_indexes].reject{|x| x == ''}.join(\"\\n\\n\")\n \"create_table(#{table.inspect}#{', :ignore_index_errors=>true' if !options[:same_db] && options[:indexes] != false && !gen.indexes.empty?}) do\\n#{commands.gsub(/^/, ' ')}\\nend\"\n end",
"def mysqldump(options)\n name = options[:name].to_s\n append_name = options[:append_name].to_s\n dump_options = options[:dump_options].to_s\n if @db_password.to_s.empty?\n db_password = \"\"\n else\n db_password = \"-p#{@db_password}\"\n end\n file_name = \"#{@data_dir}/#{append_name}#{name}_#{dump_options}_#{@filename}.sql\"\n puts \"Dumping #{options[:name]} into #{file_name}\\n\"\n command = \" nice -n #{@nice} mysqldump -u #{@db_username} #{dump_options} #{db_password} #{@extra_dump_options} #{name} > #{file_name}\"\n puts \"\\nEXECUTING:\\n #{command}\"\n system(command)\n return file_name\nend",
"def get_filter_sql\n return @schema_filter.filter_sql\n end",
"def initialize(options)\n raise 'Missing config username on output.mysql' unless options.key? :username\n raise 'Missing config database on output.mysql' unless options.key? :database\n raise 'Missing config table on output.mysql' unless options.key? :table\n conf = DEFAULT_MYSQL.merge options\n @table = conf.delete :table\n conf[:port] = conf[:port].to_i\n conf[:password] = conf[:password].to_s\n @encoding = conf[:encoding]\n conf[:ssl_mode] = :disabled if conf[:ssl_mode] != :enabled\n @mysql = Mysql2::Client.new conf\n @mysql_conf = conf\n create_table_if_not_exists\n @sqls = Hash.new { |hash, key| hash[key] = generate_batch_sql key }\n end",
"def dump( out = Dumper.new )\n for table in tables\n table.dump( out )\n end\n out\n end",
"def schema_dump(file_name = nil)\n schema = ActiveRecord::SchemaDumper.dump(ActiveRecord::Base.connection, StringIO.new).string\n return schema unless file_name\n\n File.open(file_name, 'w') { |f| f.puts(schema) }\n end",
"def schema_dump(file_name = nil)\n schema = ActiveRecord::SchemaDumper.dump(ActiveRecord::Base.connection, StringIO.new).string\n return schema unless file_name\n\n File.open(file_name, 'w') { |f| f.puts(schema) }\n end",
"def droby_dump(dest)\n\t droby = super\n droby.plan_predicates.concat(plan_predicates)\n droby.neg_plan_predicates.concat(neg_plan_predicates)\n droby\n\tend",
"def register_dump(name, clean: true, **options)\n called_from = caller_locations(1, 1).first.path\n watch = options.delete(:watch) || [called_from]\n cache_key = options.delete(:cache_key)\n skip = options.delete(:skip_if)\n\n id = \"sql/#{name}\"\n\n register_method = clean ? :register : :cached\n\n public_send(register_method, id) do\n dump = Dump.new(name, watch: watch, cache_key: cache_key)\n\n unless dump.force?\n next if skip&.call(dump: dump)\n\n next dump.within_prepared_env(import: true, **options) { dump.load } if dump.exists?\n end\n\n subscriber = ActiveSupport::Notifications.subscribe(\"sql.active_record\", dump.subscriber)\n res = dump.within_prepared_env(**options) { yield }\n\n dump.commit!\n\n res\n ensure\n ActiveSupport::Notifications.unsubscribe(subscriber) if subscriber\n end\n end",
"def schema_ds_filter(table_name, opts)\n if table_name\n [{:c__table_name=>table_name.to_s}]\n else\n [{:t__table_type=>'BASE TABLE'}]\n end\n end",
"def generate_query\n unless databases.nil?\n databases.each do |db|\n create_query[db] = [\"create table #{tablename} (\"]\n end\n csv_column_datatypes.each do |header, datatype|\n append_to_query = build_query_for_datatype(header, datatype)\n append_to_query.each do |key, value|\n create_query[key].push(value)\n end\n end\n prepare_sql_statements\n prepare_import_csv\n # Pass the prepared statements to options varaible.\n # Which gets passed on to print_metadata_analysis\n options[:create_query] = create_query\n options[:import_query] = import_query\n end\n print_metadata_analysis\n end",
"def write_sql model_name, model_attributes,output\n model_attributes.each do|key,query|\n sql= ActiveRecord::Base.connection();\n (sql.select_all query).each do |row|\n make_triples(row,model_name,\"\")\n end\n end\n end",
"def prune(options=nil)\n log.warn{ \"Pruning DB records\" }\n pruner(options || {}).go\n end",
"def saveme\n logdir = RESULTS + $config['INJECTOR']['MYSQL']['URL'].sub('http://', '').sub('https://', '').sub(/\\/$/, '').split(\"/\")[0]\n file = logdir + \"/union_info.txt\"\n if File.exists?(file)\n # backup old copy and move out the way\n end\n Dir.mkdir(logdir) unless File.exists?(logdir) and File.directory?(logdir)\n f=File.open(file, 'w+')\n f.puts \"Target: #{$config['INJECTOR']['MYSQL']['URL'].sub('http://', '').sub('https://', '').sub(/\\/$/, '').split(\"/\")[0]}\\n\"\n f.puts \"Injection Point: #{$config['INJECTOR']['MYSQL']['LOC']}\"\n f.puts \"Method: Union\"\n f.puts \"Column Count: #{$config['INJECTOR']['MYSQL']['UNION']['COLUMN_COUNT']}\"\n f.puts \"Vulnerable Column: #{$config['INJECTOR']['MYSQL']['UNION']['VULN_COLUMN']}\\n\\n\"\n f.puts \"GET: #{$config['INJECTOR']['MYSQL']['URL']}\" if $config['INJECTOR']['MYSQL']['DATA'].nil? \n f.puts \"POST: #{$config['INJECTOR']['MYSQL']['URL']}\" unless $config['INJECTOR']['MYSQL']['DATA'].nil? \n f.puts \"DATA: #{$config['INJECTOR']['MYSQL']['DATA']}\" unless $config['INJECTOR']['MYSQL']['DATA'].nil? \n case $config['INJECTOR']['MYSQL']['LOC']\n when 'UA'\n f.puts \"User-Agent: #{$config['INJECTOR']['MYSQL']['UA']}\"\n when 'REF'\n f.puts \"Referer: #{$config['INJECTOR']['MYSQL']['REF']}\"\n when 'HEADER'\n $config['INJECTOR']['MYSQL']['HEADERS'].each do |k, v|\n f.puts \"Header Name: #{k}\"\n f.puts \"Header Value: #{v}\"\n end\n when 'COOKIE'\n $config['INJECTOR']['MYSQL']['COOKIES'].each do |k, v|\n f.puts \"Cookie Name: #{k}\"\n f.puts \"Cookie Value: #{v}\"\n end\n end\n f.puts \"\\n\\nDBMS: MySQL\"\n f.puts \"Version: #{@target_config['VERSION']}\" unless @target_config['VERSION'].nil?\n f.puts \"Host: #{@target_config['HOST']}\" unless @target_config['HOST'].nil?\n f.puts \"User: #{@target_config['USER']}\" unless @target_config['USER'].nil?\n f.puts \"Basedir: #{@target_config['BASEDIR']}\" unless @target_config['BASEDIR'].nil?\n f.puts \"Datadir: #{@target_config['DATADIR']}\" unless @target_config['DATADIR'].nil?\n f.puts \"Tmpdir: #{@target_config['TMPDIR']}\" unless @target_config['TMPDIR'].nil?\n f.puts \"DB: #{@target_config['CURRENT_DB']}\" unless @target_config['CURRENT_DB'].nil?\n f.puts \"DBS: #{@target_config['DBS'].join(', ')}\\n\\n\" unless @target_config['DBS'].empty?\n f.puts \"\\n\\n\" if @target_config['DBS'].empty?\n if not @target_config['PASSWORDS'].nil? or @target_config['PASSWORDS'].size != 0\n f.puts \"MySQL Users & Passwords:\"\n f.puts @target_config['PASSWORDS']\n f.puts \"\\n\\n\"\n end\n f.puts \"Tables from #{@target_config['CURRENT_DB']}:\\n#{@target_config['CDB_TABLES'].join(', ')}\\n\\n\" unless @target_config['CDB_TABLES'].empty?\n if @target_config['DB_TABLES'].size > 1\n count=0\n @target_config['DB_TABLES'].each do |db, tables|\n if count.to_i > 1\n f.puts \"Tables from #{db}:\\n#{tables.join(', ')}\\n\\n\" unless tables.empty?\n else\n count = count.to_i + 1 #Skip the placeholder foofuck entry....\n end\n end\n end\n f.close\n print_good(\"Basic info succesfully saved!\")\n print_good(\"Saved to: #{file}\")\n end",
"def dump(schemas, output_filename)\n\n # Redirect output to output_filename (or use stdout)\n\n if output_filename then\n begin\n $stdout = File.open(output_filename, 'w');\n rescue => ex\n $stderr.puts \"Error: output file: #{ex.message}\"\n exit 1\n end\n end\n\n # Dump schemas\n\n schemas.each do |s|\n puts \"#---\"\n puts s.to_dump()\n puts\n end\n\n # Close output\n\n if output_filename then\n $stdout.close\n end\n\nend",
"def initialize(args)\n @options = args\n @create_query = {}\n @import_query = {}\n @csv_column_datatypes = args[:csv_column_datatypes]\n @nullable = args[:nullable]\n @sql_helper_options = {:tablename => tablename, :filename => @options[:filename], :delimiter => @options[:delimiter]}\n @mysql_helper.extend(CsvImportAnalyzer::MysqlQueryHelper)\n @pg_helper.extend(CsvImportAnalyzer::PgQueryHelper)\n end",
"def log_sql\n @opts[:log_sql]\n end",
"def dump( out )\n out << \"foreign_key #{out_columns}, #{out_table_name}#{out_opts}\"\n end",
"def union_data_dump(db_name, table_name, columns, start=0,stop=5)\n data=[] # WE will make an array of arrays to keep things trackable\n titlerow = []\n columns.each { |col| titlerow << col }\n count=0\n badcount=0 #Help us track things when no known count is available\n data << titlerow\n while count.to_i <= stop.to_i\n row_data = []\n columns.each do |col|\n if db_name =~ /CURRENT-DB|database()|schema()/\n results = union_basic_inject($config['INJECTOR']['MYSQL']['UNION']['VULN_COLUMN'].to_i, \"select #{col} from #{table_name} limit #{count},1\")\n else\n results = union_basic_inject($config['INJECTOR']['MYSQL']['UNION']['VULN_COLUMN'].to_i, \"select #{col} from #{db_name}.#{table_name} limit #{count},1\")\n end\n if results.nil? or results == ''\n row_data << 'NULL' unless count.to_i == stop.to_i\n badcount = badcount.to_i + 1\n else\n row_data << results\n if badcount > 0\n badcount = badcount.to_i - 1\n end\n end\n end\n pad = ' ' * (row_data.size + 25) unless row_data.empty?\n pad = ' ' * 50 if row_data.nil? if row_data.empty?\n print \"\\r(ROW##{count})> #{row_data.join(',')}#{pad}\".cyan unless row_data.empty?\n data << row_data unless row_data.empty?\n if count.to_i == stop.to_i\n break\n elsif badcount.to_i > 5\n print_caution(\"Noticing a High Number of Empty Values!\")\n print_caution(\"Do you want to continue dumping (Y/N)?\")\n answer=gets.chomp\n print_line(\"\")\n if answer.upcase == 'N' or answer.upcase == 'NO'\n print_status(\"OK, closing down dump session....\")\n break\n else\n badcount=0\n print_status(\"OK, continuing dump session....\")\n end\n end\n count = count.to_i + 1\n end\n puts\n if data.size == 1\n print_error(\"Unable to dump any data for #{db_name}.#{table_name}:#{columns.join(', ').sub(/, $/, '')}!\")\n print_error(\"Lack of privileges? IDK....\")\n print_status(\"Possible Solutions include:\")\n print_caution(\"A) Become HR's best friend by updating the code and sending him a copy\")\n print_caution(\"B) Tweak Settings and try things again\")\n print_caution(\"C) Be a bawz and do it manually\")\n print_line(\"\")\n else\t\n print_good(\"DB: #{db_name}\")\n print_good(\"Table: #{table_name}\")\n print_good(\"Columns: #{columns.join(', ').sub(/, $/, '')}\")\n pad = (data[0].size * 3) + data[0].size\n strsize = data[0].join().to_s.size\n breaker=\"\\#\" * (pad + strsize)\n print_good(\"#{breaker}\")\n table = data.to_table(:first_row_is_head => true)\n print table.to_s.white\n print_good(\"#{breaker}\")\n # Log Success for offline review\n logs = RESULTS + $config['INJECTOR']['MYSQL']['URL'].sub('http://', '').sub('https://', '').sub(/\\/$/, '').split(\"/\")[0]\n logdir = logs + '/dumps/'\n file = \"#{db_name}_#{table_name}\".gsub('/', '_').gsub('\\\\', '_').gsub(/[;:'\",.~`!@#$\\%^&*\\(\\)=\\[\\]]/, '_')\n csvfile = logdir + file + '.csv'\n txtfile = logdir + file + '.txt'\n Dir.mkdir(logs) unless File.exists?(logs) and File.directory?(logs)\n Dir.mkdir(logdir) unless File.exists?(logdir) and File.directory?(logdir)\n CSV.open(csvfile, \"wb\") do |csv|\n data.each { |row| csv << row }\n end\n t=File.open(txtfile, 'w+')\n t.puts \"DB: #{db_name}\"\n t.puts \"Table: #{table_name}\"\n t.puts \"Columns: #{columns.join(', ').sub(/, $/, '')}\"\n t.puts ''\n t.puts table.to_s\n t.close\n end\n end",
"def exec__psql_db_batch__cli_or_apply_dumps *args\n psql_db = psql_db__sample_example\n db_dumps = db_dumps__sample_example\n batch = psql_db_batch__cli_or_apply_dumps psql_db, db_dumps, \"ON_ERROR_STOP=off\"\n batch_commands batch\n end",
"def dump!\n run(mongodump)\n end",
"def dump!\n run(mongodump)\n end",
"def dump_directory\n \"--out='#{ dump_path }'\"\n end",
"def process\n create, insert, table = extract_sql #Extract mysql create/insert statements from the dump file\n raise \"Couldn't extract create syntax from MySql Dump File\" if create.nil?\n create = escape_create_string(create)\n begin\n @connection.execute(\"DROP TABLE #{table}\") rescue ''#Drop existing table first\n @connection.execute(create) #Recreate the table \n if insert && @import_data\n values = row_values(insert) \n values.each do |val|\n sql = \"INSERT INTO #{table} VALUES #{val}\"\n begin\n @connection.execute(sql) #Insert rows\n rescue Exception => e\n puts e.message\n puts sql\n puts \"table #{table}\"\n end\n end\n else\n puts \"There's no records to be added\" if @import_data && !insert\n end\n rescue Exception => e\n puts e.message\n puts \"table #{table}\"\n end\n end",
"def db_dumps__sample_example *args\n [\n \"install/10.0_to_10.1.sql\",\n \"/root/hack_10.0_to_10.1.sql\",\n ]\n\n end",
"def dump_to_file(options = {})\r\n options.reverse_merge!(:bulk => false)\r\n bulk_mode = options[:bulk]\r\n version = Time.now.utc.stamp\r\n file_name = \"#{version}_#{bulk_mode ? BULK_FILENAME_ID : INCREMENT_FILENAME_ID}.yml\"\r\n records = []\r\n\r\n # require 'ruby-debug'\r\n # debugger\r\n\r\n conditions = 'text <> \"dnr\"' # Don't dump out records marked Do Not Resuscitate\r\n conditions << \" AND to_dump\" unless bulk_mode\r\n records = model.find(:all, :conditions => conditions).map {|r| r.attributes}\r\n records.each do |r|\r\n %w(id type facet item_id namespace obsolete table_name from_bundle to_dump).each {|field| r.delete(field)}\r\n end\r\n\r\n if records.empty?\r\n @report << \"No records found, no sense in creating a file\\n\"\r\n return nil, 0\r\n end\r\n\r\n write_file(File.join(translation_base, file_name), records.ya2yaml)\r\n if bulk_mode\r\n model.update_all(['to_dump = ?', false])\r\n #no sence in loading our own bundle later.\r\n #but this is only for bulks because partials are downloaded from site and might not make it to SVN\r\n TranslationBundle.create!(:version => version, :bulk => true)\r\n end\r\n\r\n record_n = records.size \r\n puts \"\\tDumped #{record_n} records into file #{file_name}\"\r\n\r\n if options[:svn]\r\n a = `svn add #{File.join(translation_base, file_name)}`\r\n puts a =~ /^A/ ? \"Added #{file_name} to svn\" : \"Error adding #{file_name} to svn:\\n\\t#{a}\"\r\n end\r\n\r\n return file_name, record_n\r\n end",
"def dumpable\n set(variable_table: variable_table.dumpable, rule_table: rule_table.dumpable)\n end",
"def post_sql_statements( table_name, options ) # :nodoc:\n post_sql_statements = []\n if options[:on_duplicate_key_update]\n post_sql_statements << sql_for_on_duplicate_key_update( table_name, options[:on_duplicate_key_update] )\n end\n\n #custom user post_sql\n post_sql_statements << options[:post_sql] if options[:post_sql]\n\n #with rollup\n post_sql_statements << rollup_sql if options[:rollup]\n\n post_sql_statements\n end",
"def dump(db, filename, source)\n open(filename, \"w\") do |file|\n file.write(\"[\")\n source.call(db) do |record|\n file.write(record)\n end\n file.write(\"]\")\n end\nend",
"def dump_foreign_key_migration(options=OPTS)\n ts = _dump_tables(options)\n <<END_MIG\nSequel.migration do\n change do\n#{ts.map{|t| dump_table_foreign_keys(t)}.reject{|x| x == ''}.join(\"\\n\\n\").gsub(/^/, ' ')}\n end\nend\nEND_MIG\n end",
"def dbDumptoCsv(req,res)\n qs = req.query[\"select\"].split(\",\")\n @dbout = []\n qs.each do |q|\n @dbout = @dbout + @db.all(:url => q)\n end\n \n # Open the output csv file\n @oFile = File.new(\"#{$root_dir}/\" + $config.get_value('burpdotcsv'), 'w')\n \n @dbout.each do |entry|\n @oFile.syswrite(entry.ref + \",\" + entry.url + \"\\n\") if entry.ref.length > 0\n end\n \n @oFile.close\n end",
"def run_cmd!\n \"pg_dump #{pg_dump_args} | psql #{new_db_name}\"\n end",
"def obfuscate\n @mysqlcmds ||= ::Rds::S3::Backup::MySqlCmds.new(backup_server.endpoint['Address'],\n @opts['mysql_username'],\n @opts['mysql_password'],\n @opts['mysql_database'])\n\n @mysqlcmds.exec(@opts['obfuscate_sql'])\n end",
"def dump(database_name, dst)\n print \"Dummping #{database_name} ...\"\n command = @mysqldump\n command << \" -u #{@username}\"\n command << \" --skip-comments\"\n command << \" --skip-extended-insert\"\n command << \" --single-transaction\"\n command << \" --quick\"\n command << \" --password=#{@password}\" unless @password.blank?\n command << \" #{database_name}\"\n command << \" > #{dst}\"\n puts command\n output = `#{command}`\n if $?.success?\n puts 'Ok'\n else\n puts 'ERROR: Check your configuration file'\n end\n end",
"def droby_dump(dest)\n droby = super\n droby.arguments.merge!(arguments.droby_dump(dest))\n droby\n\tend",
"def dump_indexes(options={})\n is = indexes.map do |c|\n c = c.dup\n cols = c.delete(:columns)\n if table = options[:add_index] || options[:drop_index]\n \"#{options[:drop_index] ? 'drop' : 'add'}_index #{table.inspect}, #{cols.inspect}#{', :ignore_errors=>true' if options[:ignore_errors]}#{opts_inspect(c)}\"\n else\n \"index #{cols.inspect}#{opts_inspect(c)}\"\n end\n end\n is.join(\"\\n\")\n end",
"def cleanse(args)\n # Clean sql, need to add code to escape quotes and prevent deletions without conditions\n @sql\n end",
"def dump(table_name)\n dumping_tables << table_name\n end",
"def to_sql\n source.select(bin_sql).to_sql\n end",
"def prepare_from_yaml\n set(:db_file) { \"#{application}-dump.sql.bz2\" }\n set(:db_remote_file) { \"#{current_path}/backup/#{db_file}\" }\n set(:db_local_file) { \"tmp/#{db_file}\" }\n set(:db_user) { db_config[rails_env][\"username\"] }\n set(:db_pass) { db_config[rails_env][\"password\"] }\n set(:db_host) { db_config[rails_env][\"host\"] }\n set(:db_name) { db_config[rails_env][\"database\"] }\n end",
"def db_dump_cmd\n \"mongodump\"\n end",
"def dump(io)\n print \"Going through tables.\\n\" if @debug\n @rows_count = 0\n \n if @args[:tables]\n tables = @args[:tables]\n else\n tables = @args[:db].tables.list.values\n end\n \n if @on_status\n @on_status.call(:text => \"Preparing.\")\n \n @rows_count_total = 0\n tables.each do |table_obj|\n @rows_count_total += table_obj.rows_count\n end\n end\n \n tables.each do |table_obj|\n table_obj = @args[:db].tables[table_obj] if table_obj.is_a?(String) or table_obj.is_a?(Symbol)\n \n #Figure out keys.\n @keys = []\n table_obj.columns do |col|\n @keys << col.name\n end\n \n @table_obj = table_obj\n self.update_status\n print \"Dumping table: '#{table_obj.name}'.\\n\" if @debug\n self.dump_table(io, table_obj)\n end\n end",
"def climate_dump\r\n\ttableMetadata = LsrsClimate.where(:WarehouseName=>params[:table]).first\r\n\toutputCsvFile = File.open(\"/development/data/climate/#{params[:table]}.txt\", 'w')\r\n\toutputCsvFile.puts '--- #YAML'\r\n\toutputCsvFile.puts 'Title: \"' + tableMetadata.Title_en + '\"'\r\n\toutputCsvFile.puts 'Geography: \"Polygons\"'\r\n\toutputCsvFile.puts 'Framework: \"' + tableMetadata.PolygonTable + '\"'\r\n\toutputCsvFile.puts 'Timeframe: \"???\"'\r\n\toutputCsvFile.puts 'Origin: \"Observations???\"'\r\n\toutputCsvFile.puts 'Description: \"???\"'\r\n\toutputCsvFile.puts '--- #TSV'\r\n\toutputCsvFile.puts \"id\tlong\tlat\telev\ttmax01\ttmax02\ttmax03\ttmax04\ttmax05\ttmax06\ttmax07\ttmax08\ttmax09\ttmax10\ttmax11\ttmax12\ttmin01\ttmin02\ttmin03\ttmin04\ttmin05\ttmin06\ttmin07\ttmin08\ttmin09\ttmin10\ttmin11\ttmin12\tptot01\tptot02\tptot03\tptot04\tptot05\tptot06\tptot07\tptot08\tptot09\tptot10\tptot11\tptot12\"\r\n\trecords = eval(params[:table].capitalize).all\r\n\tfor r in records do\r\n\t\toutputCsvFile.puts \"#{r.id}\t#{r.long}\tlat\telev\ttmax01\ttmax02\ttmax03\ttmax04\ttmax05\ttmax06\ttmax07\ttmax08\ttmax09\ttmax10\ttmax11\ttmax12\ttmin01\ttmin02\ttmin03\ttmin04\ttmin05\ttmin06\ttmin07\ttmin08\ttmin09\ttmin10\ttmin11\ttmin12\tptot01\tptot02\tptot03\tptot04\tptot05\tptot06\tptot07\tptot08\tptot09\tptot10\tptot11\tptot12\"\r\n\tend\r\n\toutputCsvFile.close\r\n\t\t\r\n\t\r\nend",
"def backup(dest_dir)\n logger.info \"Dumping MySQL#{db.db_and_table_names}\"\n @db.backup dest_dir\n end",
"def apply_filters(query, options)\n if @model_ids && options[:mview]\n query = query.where(model_id: @model_ids)\n elsif @model_ids\n query = query.where('scenarios.model_id' => @model_ids)\n end\n query = query.where(scenario_id: @scenario_ids) if @scenario_ids\n query = query.where(indicator_id: @indicator_ids) if @indicator_ids\n query = query.where(location_id: @location_ids) if @location_ids\n query = query.joins(:indicator).where(indicators: {subcategory_id: apply_category_filter}) if @category_ids\n puts query.to_sql.inspect\n query\n end",
"def perform!\n super\n\n pipeline = Pipeline.new\n dump_ext = 'ldif'\n\n pipeline << slapcat\n\n model.compressor.compress_with do |command, ext|\n pipeline << command\n dump_ext << ext\n end if model.compressor\n\n pipeline << \"#{ utility(:cat) } > \" +\n \"'#{ File.join(dump_path, dump_filename) }.#{ dump_ext }'\"\n\n pipeline.run\n if pipeline.success?\n log!(:finished)\n else\n raise Error, \"Dump Failed!\\n\" + pipeline.error_messages\n end\n end",
"def prepare_from_yaml\n set(:db_file) { \"#{application}-dump.sql.bz2\" }\n set(:db_remote_file) { \"#{shared_path}/backup/#{db_file}\" }\n set(:db_local_file) { \"tmp/#{db_file}\" }\n set(:db_user) { db_config[rails_env][\"username\"] }\n set(:db_pass) { db_config[rails_env][\"password\"] }\n set(:db_host) { db_config[rails_env][\"host\"] }\n set(:db_name) { db_config[rails_env][\"database\"] }\n end",
"def dump_directory\n \"--out='#{ @dump_path }'\"\n end",
"def prepare_from_yaml\n set(:db_file) { \"#{application}-dump.sql.bz2\" }\n set(:db_remote_file) { \"#{shared_path}/backup/#{db_file}\" }\n set(:db_local_file) { \"tmp/#{db_file}\" }\n set(:db_user) { db_config[rails_env][\"username\"] }\n set(:db_password) { db_config[rails_env][\"password\"] }\n set(:db_host) { db_config[rails_env][\"host\"] }\n set(:db_name) { db_config[rails_env][\"database\"] }\n end",
"def sql(options={})\n get_location\n # TODO: validate options\n @params[:sql] = FEATURE_DEFAULTS[:sql].merge(options)\n @params[:sql][:generate] = true\n end",
"def mongodump\n \"#{ utility(:mongodump) } #{ database } #{ credential_options } \" +\n \"#{ connectivity_options } #{ ipv6 } #{ additional_options } #{ dump_directory }\"\n end",
"def filter(name, options = {})\n name = name.to_s\n options.each { |k,v| options[k] = v.to_s if Symbol === v }\n \n # Adds the new value to the sql_filters hash\n #\n self.sql_filters ||= {}\n self.sql_filters[name] = options\n self.sql_filters[name][:type] ||= :string\n\n if human_name = options.delete(:human_name)\n name.instance_variable_set('@human_name', human_name)\n def name.humanize; @human_name; end\n end\n \n columns << ActiveRecord::ConnectionAdapters::Column.new(\n name,\n sql_filters[name][:default],\n sql_filters[name][:type],\n options.include?(:null) ? options[:null] : true\n )\n \n # raise ArgumentError.new(\"unknown option(s) #{options.inspect}\") unless options.empty?\n end",
"def dump_indexes(options=OPTS)\n is = indexes.map do |c|\n c = c.dup\n cols = c.delete(:columns)\n if table = options[:add_index] || options[:drop_index]\n \"#{options[:drop_index] ? 'drop' : 'add'}_index #{table.inspect}, #{cols.inspect}#{', :ignore_errors=>true' if options[:ignore_errors]}#{opts_inspect(c)}\"\n else\n \"index #{cols.inspect}#{opts_inspect(c)}\"\n end\n end\n is = is.reverse if options[:drop_index]\n is.join(\"\\n\")\n end",
"def filters= options = {}\n prominent = [ :eid, :url ]\n narrowing = options.keys & Ydl::Videos.columns - prominent\n @filters = { prominent: prominent, narrowing: narrowing }\n end",
"def inner_dump( &encode_block )\n # could possibly overrride Dataset#paginate(page_no, page_size, record_count=nil)\n on_conditions = primary_keys.map{|f| [f,f]}.to_h\n (0..table_dataset.count).step(page_size).each do |offset|\n limit_dataset = table_dataset.select( *primary_keys ).limit( page_size, offset ).order( *primary_keys )\n page = table_dataset.join( limit_dataset, on_conditions ).order( *primary_keys ).qualify(table_name)\n logger.info \"#{__method__} #{table_name} #{offset}\"\n logger.debug page.sql\n page.each &encode_block\n end\n end",
"def schema_migrations_cleanup\n # Read all schema_migrations values from the dump.\n values = dump.scan(/^(\\(\\'\\d{14}\\'\\))[,;]\\n/).flatten.sort\n\n # Replace the schema_migrations values.\n dump.sub!(\n /(?<=INSERT INTO \"schema_migrations\" \\(version\\) VALUES).+;\\n*/m,\n \"\\n #{values.join(\"\\n,\")}\\n;\\n\\n\"\n )\n end",
"def schema_migrations_cleanup\n # Read all schema_migrations values from the dump.\n values = dump.scan(/^(\\(\\'\\d{14}\\'\\))[,;]\\n/).flatten.sort\n\n # Replace the schema_migrations values.\n dump.sub!(\n /(?<=INSERT INTO \"schema_migrations\" \\(version\\) VALUES).+;\\n*/m,\n \"\\n #{values.join(\"\\n,\")}\\n;\\n\\n\"\n )\n end",
"def to_sql\n return nil unless filters.present? and valid?\n filters.map { |n, f| sql_for(n) }.reject(&:blank?).join(' AND ')\n end",
"def mongo_dump_command\n cmd = \"\"\n cmd = \"mongodump -h #{config['host']}:#{config['port']} -d #{config['database']} -o . \"\n cmd += \"--username #{config['username']} --password #{config['password']}\" unless config[\"username\"].nil? and config[\"password\"].nil?\n cmd\n end",
"def delete_all_packet_filters\n super\n end",
"def csv_dump(addr, root_dir)\n filename = \"full.csv\"\n dump_dir = File.join(root_dir)\n filepath = File.join(dump_dir, filename)\n\n cmd = \"psql -c '\\\\copy (SELECT * FROM outdoor_env WHERE (apogee_w_m2 IS NOT NULL AND address=#{addr}) ORDER BY db_time DESC) To #{filepath} With CSV header\\'\"\n run_command(cmd)\nend",
"def prepare_import_csv\n databases.each do |db|\n if db == :mysql\n import_query[db] = mysql_helper.import_csv(tablename: tablename, filename: filename, delimiter: delimiter)\n elsif db == :pg\n import_query[db] = pg_helper.import_csv(tablename: tablename, filename: filename, delimiter: delimiter)\n end\n end\n end",
"def mysqldump(namespace, database_name = nil)\n puts 'Starting migration...'\n open_database_connection do\n database_uri = build_database_uri(namespace)\n database_name ||= database_uri.path.split('/')[1]\n output_file = \"#{database_name}-#{Time.now.strftime('%Y%m%d-%H%M%S')}.sql\"\n compress_cmd = compress_command(output_file)\n puts '-> Running the mysqldump'\n puts `mysqldump -h #{database_uri.host} -P #{database_uri.port} -u #{database_uri.user} --password=#{database_uri.password} #{database_name} | #{compress_cmd}`\n end\n end",
"def construct_finder_sql(options) #:nodoc:\n add_piggy_back!(options)\n old_construct_finder_sql(options)\n end",
"def to_dump(opts={})\n dump = \"\"\n dump << \"add_column_check_constraint #{table_name.to_s}, #{column_name.to_s}\" unless opts[:inline]\n dump << \", check: #{check.inspect}\"\n dump << \"\\n\"\n dump\n end",
"def dumpRulesets()\r\n 79.times {print \"=\"}\r\n puts\r\n puts \"RULESET DUMP\".center(80)\r\n 79.times {print \"=\"}\r\n puts\r\n\r\n if(@rulesets.length > 0)\r\n rulesets = @rulesets.sort\r\n rulesets.each do |key, ruleset|\r\n\r\n 40.times {print \"-\"}\r\n puts\r\n puts \"#{ruleset.name}\\t(#{ruleset.alias}) : #{ruleset.type} : #{ruleset.execType}\"\r\n ruleset.rules.each do |ruleAlias|\r\n puts \"\\t#{ruleAlias}\"\r\n end # rules.each\r\n puts \"\"\r\n\r\n end # rulesets.each\r\n\r\n else\r\n puts \"No rulesets to dump.\"\r\n end\r\n\r\n puts \"\"\r\n end",
"def copy_table_sql(table, opts)\n if table.is_a?(String)\n table\n else\n if opts[:options] || opts[:format]\n options = String.new\n options << \" (\"\n options << \"FORMAT #{opts[:format]}\" if opts[:format]\n options << \"#{', ' if opts[:format]}#{opts[:options]}\" if opts[:options]\n options << ')'\n end\n table = if table.is_a?(::Sequel::Dataset)\n \"(#{table.sql})\"\n else\n literal(table)\n end\n \"COPY #{table} TO STDOUT#{options}\"\n end\n end",
"def xDumpAllDatabase()\n puts \"Back up commencing...\"\n Dir.chdir('/Users/jeydurai')\n system('start_mongodump.bat')\n end",
"def dump(io)\n debug \"Going through tables.\"\n @rows_count = 0\n\n if @on_status\n @on_status.call(text: \"Preparing.\")\n\n @rows_count_total = 0\n tables.each do |table_obj|\n @rows_count_total += table_obj.rows_count\n end\n end\n\n each_table do |table_obj|\n # Figure out keys.\n @keys = []\n table_obj.columns do |col|\n @keys << col.name\n end\n\n @table_obj = table_obj\n update_status\n debug \"Dumping table: '#{table_obj.name}'.\"\n dump_table(io, table_obj)\n end\n\n dump_foreign_keys(io)\n end",
"def select_table_options_sql(sql)\n sql << \" WITH #{@opts[:table_options]}\" if @opts[:table_options]\n end",
"def get_filter_sql\n filter_sql = \"( \"\n self.foreach_filter do |filter|\n filter_sql += \"#{filter.get_filter_sql} #{filter.next_condition_as_sql} \"\n end\n filter_sql += \")\"\n return filter_sql\n end",
"def dump!(output_dir = MODEL_DIR)\n model_classes.each do |model_class|\n filename = \"#{ model_class.name.tableize }.#{ FILE_EXT }\".gsub(\"/\", \"-\")\n\n # Open the output file and replace its contents.\n File.open( File.join(output_dir.split(\"/\"), filename), \"w\" ) do |file|\n if model_class.any?\n records = []\n\n model_class.find_in_batches do |batch|\n # Only extract the attributes of each record since that's what we'll\n # use to recreate these records later.\n new_records = batch.map do |record|\n record.__send__(:attributes)\n end\n\n records.concat( new_records )\n end\n\n file << JSON.pretty_generate( records )\n end\n end\n end\n\n true\n end",
"def remove_dump\n %Q{Remove raw dump. To disable that use --remove-dump false}\n end",
"def wDumpAllDatabase()\n puts \"Back up commencing...\"\n Dir.chdir('/Users/jeydurai')\n system('start_mongodump.bat')\n end",
"def dump_database(app_name, instance_name)\n app_config = RailsPwnerer::Config[app_name, instance_name]\n db_name, db_user, db_pass = app_config[:db_name], app_config[:db_user], app_config[:db_pass]\n\n pwnerer_user = app_config[:pwnerer_user]\n pwnerer_uid = uid_for_username(pwnerer_user)\n pwnerer_gid = gid_for_username(pwnerer_user)\n \n timestamp = Time.now.strftime '%Y%m%d%H%M%S'\n dump_file = \"db/#{app_name}.#{instance_name}_#{timestamp}.sql\"\n Dir.chdir app_config[:backup_path] do\n system(\"mysqldump --add-drop-database --add-drop-table\" +\n \" --skip-extended-insert --single-transaction\" +\n \" -u#{db_user} -p#{db_pass} #{db_name} > #{dump_file}\")\n # lockdown the file\n File.chmod(0400, dump_file)\n File.chown(pwnerer_uid, pwnerer_gid, dump_file)\n end\n end",
"def dump_table_schema(table, options={})\n table = table.value.to_s if table.is_a?(SQL::Identifier)\n raise(Error, \"must provide table as a Symbol, String, or Sequel::SQL::Identifier\") unless [String, Symbol].any?{|c| table.is_a?(c)}\n s = schema(table).dup\n pks = s.find_all{|x| x.last[:primary_key] == true}.map{|x| x.first}\n options = options.merge(:single_pk=>true) if pks.length == 1\n m = method(:column_schema_to_generator_opts)\n im = method(:index_to_generator_opts)\n begin\n indexes = indexes(table).sort_by{|k,v| k.to_s} if options[:indexes] != false\n rescue Sequel::NotImplemented\n nil\n end\n gen = Schema::Generator.new(self) do\n s.each{|name, info| send(*m.call(name, info, options))}\n primary_key(pks) if !@primary_key && pks.length > 0\n indexes.each{|iname, iopts| send(:index, iopts[:columns], im.call(table, iname, iopts))} if indexes\n end\n commands = [gen.dump_columns, gen.dump_constraints, gen.dump_indexes].reject{|x| x == ''}.join(\"\\n\\n\")\n \"create_table(#{table.inspect}#{', :ignore_index_errors=>true' if !options[:same_db] && options[:indexes] != false && indexes && !indexes.empty?}) do\\n#{commands.gsub(/^/o, ' ')}\\nend\"\n end",
"def usage\n puts \"Usage examples:\n ./dump_from_db.rb --headers OUTDIR\n ^ Dumps headings only into the dir OUTDIR\n ./dump_from_db.rb OUTDIR -file INFILE\n ^ Dumps projects listed in INFILE \n (id only, one per line) to OUTDIR\n ./dump_from_db.rb OUTDIR -list 1000 1001 1002\n ^ Dumps projects 1000, 1001, etc to OUTDIR.\"\nend"
] |
[
"0.60283154",
"0.589884",
"0.5849439",
"0.57685965",
"0.5689041",
"0.56366765",
"0.5518343",
"0.5504224",
"0.5498455",
"0.54618114",
"0.54434943",
"0.5412365",
"0.53963107",
"0.53621936",
"0.5312326",
"0.52636266",
"0.52486044",
"0.52257913",
"0.5223067",
"0.5200668",
"0.5179168",
"0.5174994",
"0.5173375",
"0.514024",
"0.5137177",
"0.5133492",
"0.5125662",
"0.5101853",
"0.50895685",
"0.50895685",
"0.5035399",
"0.5035137",
"0.50100416",
"0.50053185",
"0.4966119",
"0.49603602",
"0.49263516",
"0.4923722",
"0.4908782",
"0.4905246",
"0.49021897",
"0.48937953",
"0.48888865",
"0.48886684",
"0.48886684",
"0.48836717",
"0.48742938",
"0.48618713",
"0.4856415",
"0.4840031",
"0.48386365",
"0.48379675",
"0.4831533",
"0.48162806",
"0.4815539",
"0.48128006",
"0.48115316",
"0.48075446",
"0.48047107",
"0.4803841",
"0.47770438",
"0.47686788",
"0.47587645",
"0.4756252",
"0.47541758",
"0.4742769",
"0.47423443",
"0.47296387",
"0.47223857",
"0.47179633",
"0.4714899",
"0.47085363",
"0.47041428",
"0.47033718",
"0.4701587",
"0.46896106",
"0.4689443",
"0.46851978",
"0.467996",
"0.467996",
"0.46788955",
"0.46787244",
"0.4677627",
"0.4674235",
"0.4660198",
"0.46597466",
"0.46592742",
"0.46487108",
"0.46317127",
"0.46306372",
"0.46305937",
"0.4626353",
"0.46176657",
"0.46169493",
"0.4607629",
"0.4599815",
"0.45925614",
"0.4589075",
"0.45862612",
"0.45842722"
] |
0.7232968
|
0
|
Run the dump. Call this method for a good time.
|
def run
tables.each do |table|
dump_table(table)
end
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def start_dump\n end",
"def start_dump\n end",
"def start_dumping\n %Q{Start dumping ...}\n end",
"def perform!\n super\n\n dump = \"echo '.dump' | #{sqlitedump_utility} #{path}\"\n\n pipeline = Pipeline.new\n dump_ext = \"sql\".dup\n\n pipeline << dump\n if model.compressor\n model.compressor.compress_with do |command, ext|\n pipeline << command\n dump_ext << ext\n end\n end\n\n pipeline << \"cat > '#{File.join(dump_path, dump_filename)}.#{dump_ext}'\"\n\n pipeline.run\n\n if pipeline.success?\n log!(:finished)\n else\n raise Error,\n \"#{database_name} Dump Failed!\\n\" + pipeline.error_messages\n end\n end",
"def perform!\n log!\n\n begin\n lock_database if @lock.eql?(true)\n if collections_to_dump.is_a?(Array) and not collections_to_dump.empty?\n specific_collection_dump!\n else\n dump!\n end\n unlock_database if @lock.eql?(true)\n rescue => exception\n unlock_database if @lock.eql?(true)\n raise exception\n end\n end",
"def run\n super() \n true_file_name = \"%s_%s_%s.%s\"%[file, platform_setting[:interface][interface], port, EXTENSION]\n #start tcpdump with the file name and trap process id\n command = \"(tcpdump -X -s 0 -i #{platform_setting[:interface][interface]} -w #{true_file_name} port #{port} >/dev/null 2>&1&); ps -elf | grep 'tcpdump' | egrep -v 'sh|grep|rb$'\" \n result = `#{command}`\n sleep(5) #need to wait for a bit before tcpdump is ready\n self.pidlist.push(true_file_name)\n end",
"def perform!\n super\n\n lock_database if @lock\n @only_collections.empty? ? dump! : specific_collection_dump!\n\n rescue => err\n raise Errors::Database::MongoDBError.wrap(err, 'Database Dump Failed!')\n ensure\n unlock_database if @lock\n package! unless err\n end",
"def run\n Kernel.system \"#{mongo_dump_command}\"\n self\n end",
"def dump!\n run(mongodump)\n end",
"def dump!\n run(mongodump)\n end",
"def run(instance = nil)\n report_start\n build_data_structure instance\n create_dump_file\n create_schema\n restore\n report_completion\n end",
"def dump_loop\n while (cycle = @dump_queue.pop)\n logfile.dump(cycle)\n logfile.flush if sync?\n end\n end",
"def perform!\n super\n\n pipeline = Pipeline.new\n dump_ext = 'sql'\n\n pipeline << mysqldump\n\n model.compressor.compress_with do |command, ext|\n pipeline << command\n dump_ext << ext\n end if model.compressor\n\n pipeline << \"#{ utility(:cat) } > \" +\n \"'#{ File.join(dump_path, dump_filename) }.#{ dump_ext }'\"\n\n pipeline.run\n if pipeline.success?\n log!(:finished)\n else\n raise Error, \"Dump Failed!\\n\" + pipeline.error_messages\n end\n end",
"def run!\n report_startup\n setup_stage\n stage_operations\n managed_copy\n remove_stage\n report_complete\n end",
"def dump!\n @dump = true\n end",
"def run\n init\n\n printer = config.flamegraph? ? Printers::Flamegraph : Printers::Simple\n\n at_exit do\n File.write(build_path(), JSON.dump(result)) if config.json?\n printer.dump(result)\n end\n\n start\n end",
"def perform!\n super\n\n pipeline = Pipeline.new\n dump_ext = 'ldif'\n\n pipeline << slapcat\n\n model.compressor.compress_with do |command, ext|\n pipeline << command\n dump_ext << ext\n end if model.compressor\n\n pipeline << \"#{ utility(:cat) } > \" +\n \"'#{ File.join(dump_path, dump_filename) }.#{ dump_ext }'\"\n\n pipeline.run\n if pipeline.success?\n log!(:finished)\n else\n raise Error, \"Dump Failed!\\n\" + pipeline.error_messages\n end\n end",
"def dump() end",
"def dump\n do_dump(0)\n end",
"def run\n\t\tself.rsync_to_temp\n\t\tself.convert_to_mp3\n\t\tself.rsync_to_usb\n\t\tself.delete_temp_dir\n\tend",
"def start_compressing\n %Q{Start compressing of dump ... }\n end",
"def _dump() end",
"def do_run\n\n @unit.logger.log_run_start(self)\n\n counter_next('runs')\n\n t0 = Time.now\n\n (@unit.conf['exe_max_messages'] || 77).times do |i|\n\n break if @shutdown\n\n m = @messages.shift\n break unless m\n\n m = (@messages << m).shift \\\n if m['point'] == 'terminated' && @messages.any?\n #\n # handle 'terminated' messages last\n\n ms = process(m)\n\n @consumed << m\n\n ims, oms = ms.partition { |mm| mm['exid'] == @exid }\n # qui est \"in\", qui est \"out\"?\n\n counter_add('omsgs', oms.size)\n # keep track of \"out\" messages, messages to other executions\n\n @messages.concat(ims)\n @unit.storage.put_messages(oms)\n end\n\n @alive = false\n\n @execution.merge!(\n closing_messages: @consumed.select { |m|\n CLOSING_POINTS.include?(m['point']) })\n\n @unit.storage.put_execution(@execution)\n @unit.storage.consume(@consumed)\n\n @unit.storage.put_messages(@messages)\n\n du = Time.now - t0\n t0 = Flor.tstamp(t0)\n\n @unit.logger.log_run_end(self, t0, du)\n @unit.hooker.notify(self, make_end_message(t0, du, @execution['size']))\n\n @consumed.clear\n\n rescue Exception => exc\n\n# TODO eventually, have a dump dir\n\n fn =\n [ 'flor', @unit.conf['env'], @unit.identifier, @exid,\n 'r' + counter('runs').to_s ].collect(&:to_s).join('_') + '.dump'\n\n @unit.logger.error(\n \"#{self.class}#do_run()\", exc, \"(dumping to #{fn} ...)\")\n\n File.open(fn, 'wb') do |f|\n f.puts(Flor.to_pretty_s({\n execution: @execution,\n messages: @messages,\n consumed: @consumed,\n traps: @traps.collect(&:to_h),\n exid: @exid,\n alive: @alive,\n shutdown: @shutdown,\n thread: [ @thread.object_id, @thread.to_s ]\n }))\n f.puts('-' * 80)\n f.puts(on_do_run_exc(exc))\n end\n\n @unit.logger.error(\n \"#{self.class}#do_run()\", exc, \"(dumped to #{fn})\")\n\n #puts on_do_run_exc(exc)\n # dump notification above\n end",
"def xDumpAllDatabase()\n puts \"Back up commencing...\"\n Dir.chdir('/Users/jeydurai')\n system('start_mongodump.bat')\n end",
"def main_program\n \n check_directories\n \n check_answer\n \n ### START MYSQL DUMP ###\n puts @lines\n puts \"Starting MYSQL Dump \\n\"\n sleep 1\n if @all_databases \n puts \"INFO: Going to dump all databases into\"\n puts \" '#{@data_dir}'\"\n check_answer\n else\n puts \"INFO: Going to dump '#{@databases.join(\", \")}' databases into\"\n puts \" '#{@data_dir}'\"\n end\n \n puts @lines\n puts \"Starting MYSQL dump...\"\n make_mysql_backup\n sleep 1\n \n puts @lines\n puts \"Stablishing Connection to S3 account.\"\n stablish_connection\n \n find_or_create_bucket\n \n puts @lines\n puts \"Now Going to copy Data to S3 bucket #{@bucket_name}.\"\n send_data\n \n puts @lines\n puts \"#{@time} -- DONE\"\n puts @lines\n\nend",
"def wDumpAllDatabase()\n puts \"Back up commencing...\"\n Dir.chdir('/Users/jeydurai')\n system('start_mongodump.bat')\n end",
"def _dump\n end",
"def wRestoreDump()\n puts \"Back up commencing...\"\n Dir.chdir('/Users/jeydurai')\n system('start_mongorestore.bat')\n end",
"def xRestoreDump()\n puts \"Back up commencing...\"\n Dir.chdir('/Users/jeydurai')\n system('start_mongorestore.bat')\n end",
"def run\n start\n trace\n wrapup\n end",
"def run\n end",
"def run\n end",
"def packet dump\n Capp.offline(dump).loop.first\n end",
"def run\n Laze.debug 'Starting source processing'\n target.reset\n store.each do |item|\n target.create item\n end\n target.save\n Laze.debug 'Source processing ready'\n end",
"def run\n create_file_appender\n # The first and ONLY place the results database is created and connected to.\n load_database\n # Create any application specific databases.\n load_application_database\n end",
"def run() end",
"def _dump()\n #This is a stub, used for indexing\n end",
"def run\n end",
"def begin_dump(name)\n # Create temp file & open\n @files[name] = temporary_file(name)\n @original_files[name] = temporary_file(name, ORIGINAL_FILE_EXTENSION) if @debug_mode\n @headers_file = temporary_file('', HEADERS_FILE_EXTENSION) if @debug_mode\n end",
"def perform\n tmp_mongo_dir = \"mongodump-#{Time.now.strftime(\"%Y%m%d%H%M%S\")}\"\n tmp_dump_dir = File.join(tmp_path, tmp_mongo_dir)\n\n case self.backup_method.to_sym\n when :mongodump\n #this is the default options \n # PROS:\n # * non-locking\n # * much smaller archive sizes\n # * can specifically target different databases or collections to dump\n # * de-fragements the datastore\n # * don't need to run under sudo\n # * simple logic\n # CONS:\n # * a bit longer to restore as you have to do an import\n # * does not include indexes or other meta data\n log system_messages[:mongo_dump]\n exit 1 unless run \"#{mongodump} #{mongodump_options} #{collections_to_include} -o #{tmp_dump_dir} #{additional_options} > /dev/null 2>&1\"\n when :disk_copy\n #this is a bit more complicated AND potentially a lot riskier: \n # PROS:\n # * byte level copy, so it includes all the indexes, meta data, etc\n # * fast recovery; you just copy the files into place and startup mongo\n # CONS:\n # * locks the database, so ONLY use against a slave instance\n # * copies everything; cannot specify a collection or a database\n # * will probably need to run under sudo as the mongodb db_path file is probably under a different owner. \n # If you do run under sudo, you will probably need to run rake RAILS_ENV=... if you aren't already\n # * the logic is a bit brittle... \n log system_messages[:mongo_copy]\n\n cmd = \"#{mongo} #{mongo_disk_copy_options} --quiet --eval 'printjson(db.isMaster());' admin\"\n output = JSON.parse(run(cmd, :exit_on_failure => true))\n if output['ismaster']\n puts \"You cannot run in disk_copy mode against a master instance. This mode will lock the database. Please use :mongodump instead.\"\n exit 1\n end\n \n begin\n cmd = \"#{mongo} #{mongo_disk_copy_options} --quiet --eval 'db.runCommand({fsync : 1, lock : 1}); printjson(db.runCommand({getCmdLineOpts:1}));' admin\"\n output = JSON.parse(run(cmd, :exit_on_failure => true))\n\n #lets go find the dbpath. it is either going to be in the argv just returned OR we are going to have to parse through the mongo config file\n cmd = \"mongo --quiet --eval 'printjson(db.runCommand({getCmdLineOpts:1}));' admin\"\n output = JSON.parse(run(cmd, :exit_on_failure => true))\n #see if --dbpath was passed in\n db_path = output['argv'][output['argv'].index('--dbpath') + 1] if output['argv'].index('--dbpath') \n #see if --config is passed in, and if so, lets parse it\n db_path ||= $1 if output['argv'].index('--config') && File.read(output['argv'][output['argv'].index('--config') + 1]) =~ /dbpath\\s*=\\s*([^\\s]*)/ \n db_path ||= \"/data/db/\" #mongo's default path\n run \"cp -rp #{db_path} #{tmp_dump_dir}\" \n ensure\n #attempting to unlock\n cmd = \"#{mongo} #{mongo_disk_copy_options} --quiet --eval 'printjson(db.currentOp());' admin\"\n output = JSON.parse(run(cmd, :exit_on_failure => true))\n (output['fsyncLock'] || 1).to_i.times do\n run \"#{mongo} #{mongo_disk_copy_options} --quiet --eval 'db.$cmd.sys.unlock.findOne();' admin\"\n end\n end\n else\n puts \"you did not enter a valid backup_method option. Your choices are: #{BACKUP_METHOD_OPTIONS.join(', ')}\"\n exit 1\n end \n \n log system_messages[:compressing]\n run \"tar -cz -C #{tmp_path} -f #{File.join(tmp_path, compressed_file)} #{tmp_mongo_dir}\"\n end",
"def run_cmd!\n \"pg_dump #{pg_dump_args} | psql #{new_db_name}\"\n end",
"def run\n create_log_folder\n \n in_tmp_dir do\n start_frame = tmp_path START_FRAME\n extract_start_transition_frame(start_frame) # 1.\n \n end_frame = tmp_path END_FRAME\n extract_end_transition_frame(end_frame) # 2.\n \n transitions = tmp_path TRANSITIONS\n Cmd::GenerateTransitionFrames.new(start_frame, end_frame, transitions, INNER_FRAMES_AMOUNT).run! *logs('3_generate_transition_frames') # 3.\n \n Queue.run *FORMATS.map{ |format| proc{ transition(format) } }, close_connection_before_execution: true # 4.\n end\n \n outputs\n end",
"def run!\n run\n die\n end",
"def run; end",
"def run; end",
"def run; end",
"def run; end",
"def run; end",
"def run; end",
"def run; end",
"def run; end",
"def run; end",
"def run\n end",
"def run\n end",
"def run\n end",
"def run\n end",
"def run\n end",
"def run\n end",
"def run\n end",
"def run\n sleep INTERVAL\n loop do\n start = Time.now\n @db.purge\n first_id = @db_id_seen+1\n @db_id_seen, records = @db.nodes_down(first_id)\n sleep INFLIGHT_WAIT\n records = records.all\n @graphite.add @db.id_range(first_id, @db_id_seen).all if @graphite\n @buffer.push records.map { |record| record.peer }\n @buffer.exceed_median? ? @alarm.set(@buffer) : @alarm.clear(@buffer)\n delay = INTERVAL-(Time.now-start)\n # in case delay happens to be too big\n if delay > INTERVAL\n delay = INTERVAL\n Log.warn \"delay became larger than #{INTERVAL}, capping it. (did ntp just sync?)\"\n end\n if delay > 0\n sleep delay\n else\n Log.error \"Analyzer loop took longer than #{INTERVAL}, wanted to sleep for #{delay}s\"\n end\n end\n end",
"def run\n end",
"def _dump(limit)\n code\n end",
"def _dump(*args)\n raise \"not implemented yet\"\n end",
"def dump\n @dump ||= Dump.new(dump_url)\n end",
"def dump contents;\r\n # p contents\r\n @dumps = Array.new unless @dumps \r\n @dumps << contents \r\n end",
"def run\n log \"Nothing to do.\"\n end",
"def run\n \n end",
"def run\n \n end",
"def run\n end",
"def run\n @hydra.run\n end",
"def run_yaml\r\n #puts @prefs.inspect\r\n cur_offset = @prefs['cur_pos']\r\n cur_timeout = @prefs['timeout']\r\n \r\n cur_offset.upto(99999) do |i|\r\n puts 'on ' + i.to_s\r\n self.find_and_save_zip(i)\r\n \r\n # update position.\r\n @prefs.store('cur_pos', i)\r\n #puts @prefs.inspect\r\n #puts @config_filename.to_s\r\n YamlUtil.write(@config_filename.to_s, @prefs)\r\n \r\n # sleep\r\n sleep cur_timeout\r\n end\r\n end",
"def dump\n puts \"Memory dump:\"\n @content.each_index do |index|\n puts \"%0.5d: %0.4o\" % [index, @content[index]] unless @content[index].nil?\n end\n end",
"def dump(*args)\n argv = to_pointer([\"dump\"] + args)\n rrd_dump(args.size+1, argv) == 0\n ensure\n free_pointers\n end",
"def finished_dumping(database)\n %Q{ -- Dumping of #{database} done ...}\n end",
"def run()\r\n rounds.times do |i|\r\n round(i+1)\r\n end\r\n #$stderr.puts \"#{name} Finished!\\n\"\r\n #below codes is used to output report\r\n output()\r\n end",
"def execute\n setup\n begin\n data = main\n rescue ArgumentError => e\n warn e.message\n exit 1\n end\n puts format_data(data)\n end",
"def live_gps_dump\n\t\tputs \"Reading...\\n\"\n\t\tbuffer = \"\"\n\t\tdata = {}\n\t\trows = 1\n\t\terrors = 0\n\n\t\twhile true do\n\t\t\tbegin\n\t\t\t\tread\n\n\t\t\t\t# Clear previous data\n\t\t\t\tif rows > 0\n\t\t\t\t\t$stdout.print \"\\e[#{rows}A\\e[E\\e[J\"\n\t\t\t\t\trows = 0\n\t\t\t\tend\n\t\t\t\terrors = 0\n\t\t\t\t\n\t\t\t\t# Get date\n\t\t\t\tdate = self.date_time \n\t\t\t\tunless date.nil?\n\t\t\t\t\tdate = date.strftime(\"%b %d %I:%M %p\")\n\t\t\t\tend\n\t\t\t\t\n\t\t\t\tnum_sat = @data[:num_sat] || 0\n\t\t\t\t$stdout.print \"Time: #{date}\t\tSatellites: #{num_sat}\t\tQuality:#{@data[:quality]}\\n\"\n\t\t\t\t$stdout.print \"Latitude: #{@data[:latitude]}#{@data[:lat_ref]}\"\n\t\t\t\t$stdout.print \"\\tLongitude: #{@data[:longitude]}#{@data[:long_ref]}\"\n\t\t\t\t$stdout.print \"\\tElevation: #{@data[:altitude]}#{@data[:alt_unit]}\\n\"\n\t\t\t\trows += 3\n\t\t\t\t\n\t\t\t\t# Satellites\n\t\t\t\tif @data.key?(:num_sat)\n\t\t\t\t\t$stdout.print \"-- Satellites --\\n\"\n\t\t\t\t\t\n\t\t\t\t\t@data[:num_sat].times do | i | \n\t\t\t\t\t\t\n\t\t\t\t\t\tif @data.key?(:satellites) && @data[:satellites].size > i\n\t\t\t\t\t\t\tsat = @data[:satellites][i]\n\t\t\t\t\t\t\trows += 1\n\t\t\t\t\t\t\t\n\t\t\t\t\t\t\t$stdout.print \"#{sat[:id]}: \"\n\t\t\t\t\t\t\t$stdout.print \"Elevation: #{sat[:elevation]}\"\n\t\t\t\t\t\t\t$stdout.print \"\\t\\tAzimuth: #{sat[:azimuth]}\\n\"\n\t\t\t\t\t\tend\n\t\t\t\t\tend\n\t\t\t\t\trows += 1\n\t\t\t\tend\n\t\t\t\n\t\t\trescue Exception => e\n\t\t\t\t# Clear previous error\n\t\t\t\tif errors > 0\n\t\t\t\t\t$stdout.print \"\\e[1A\\e[E\\e[J\"\n\t\t\t\t\terrors = 0\n\t\t\t\tend\n\n\t\t\t\t$stdout.print \"\\nERROR: #{e.message}\\n\"\n\t\t\t\tbreak\n\t\t\tend\n\t\t\t\n\t\t\t$stdout.flush\n\t\tend\n\tend",
"def run\n puts self.name + \" runs\"\n end",
"def run\n\t\t\tflush_database\n\t\t\tseed_countries\n\t\t\tseed_group_organizations\n\t\t\tseed_organizations\n\t\t\tseed_locations\n\t\t\tseed_admin\n\t\t\tseed_api_key\n\t\tend",
"def run\n puts \"Called with time of #{Time.now.strftime(\"%m/%d/%Y %H:%M:%S %P\")}\"\n update_accounts \n create_statements \n create_coverge\n create_freight_benefit_records \nend",
"def dump!\n MODEL_TO_OUTPUT_FILEMAPPING.each do |klass, filepath|\n write_table_to_file(klass, filepath, %w[created_at updated_at])\n end\n end",
"def main\n \n operations.each do |op|\n in_collection = op.input(INPUT).collection\n keep_plt = op.input(KEEP_PLT).val.to_s\n \n # export_filename(collection=in_collection, method='timeseries', timepoint=20) # timepoint is duration of timeseries plate reader\n filename = Item.find(in_collection.id).get('timeseries_filename')\n \n # Directs tech through biotek plate reader software in order to export time series measurements\n export_timeseries_data(filename)\n \n # Find measurements and upload\n show {\n title \"Locating Upload\"\n separator\n note \"The file you just exported should be in the <b>'_UWBIOFAB'</b> directory\"\n note \"It is called: #{filename}\"\n }\n # Show block upload button and retrieval of file uploaded\n up_show, up_sym = upload_show(filename)\n if (up_show[up_sym].nil?)\n show {warning \"No upload found for Plate Reader measurement. Try again!!!\"}\n up_show, up_sym = upload_show(filename)\n else\n upload = find_upload_from_show(up_show, up_sym)\n key = \"#{filename}\"\n associate_to_plans(key, upload)\n associate_to_item(in_collection, key, upload)\n end\n \n (keep_plt == 'No') ? (in_collection.mark_as_deleted) : (in_collection.location = 'Bench')\n in_collection.save\n if (keep_plt == 'No')\n show {\n title \"Cleaning Up...\"\n separator\n note \"Rinse out Plate <b>#{in_collection}</b> with DI H2O and bleach\"\n note \"Then rinse once more with EtOH\"\n }\n else\n show {\n title \"Cleaning Up...\"\n separator\n note \"Move <b>#{in_collection}</b> to <b>#{in_collection.location}</b>\"\n }\n end\n end\n \n end",
"def load\n flush\n replay\n end",
"def dump\n @formatters.each{|f| f.start_dump}\n dump_failures\n @formatters.each{|f| f.dump_summary(duration, @example_names.length, @failures.length)}\n @failures.length\n end",
"def dump(path)\n infobase.designer do\n dumpDBCfg path\n end.run.wait.result.verify!\n path\n end",
"def run\n print_debug 'In #run'\n audit self.class.payloads, self.class.options\n end",
"def run\n\t\t\tsummary\n\t\tend",
"def run!\n # Have to evalute/execute the block on the instance\n instance_eval &@block\n summary_at_exit\n end",
"def run(type, action, repetitions = Benchmarking::TEST_REPETITIONS)\n file_name = type.to_s << \"_bson.json\"\n GC.disable\n file_path = [Benchmarking::DATA_PATH, file_name].join('/')\n puts \"#{action} : #{send(action, file_path, repetitions)}\"\n end",
"def snapshot()\n puts \"TODO\"\n end",
"def dump\r\n @db_mon.synchronize {\r\n File.open(@path,'wt') { |file|\r\n @seek_db.each_value { |x|\r\n file.write(x.to_json)\r\n file.write(\"\\n\")\r\n }\r\n } rescue nil\r\n }\r\n end",
"def dump\n puts\n puts \"WorkFlow '#{ @label }' consists of the following tasks:\"\n @task_list.each{ |t| t.dump }\n puts\n end",
"def run\n ## self.download_events\n files = prepare_files\n events = parse_events(files)\n save_events(events)\n end",
"def dump\n super\n\n puts \"page header:\"\n pp page_header\n puts\n\n puts \"fseg header:\"\n pp fseg_header\n puts\n\n puts \"sizes:\"\n puts \" %-15s%5i\" % [\"header\", header_space]\n puts \" %-15s%5i\" % [\"trailer\", trailer_space]\n puts \" %-15s%5i\" % [\"directory\", directory_space]\n puts \" %-15s%5i\" % [\"free\", free_space]\n puts \" %-15s%5i\" % [\"used\", used_space]\n puts \" %-15s%5i\" % [\"record\", record_space]\n puts \" %-15s%5.2f\" % [\"per record\", space_per_record]\n puts\n\n puts \"page directory:\"\n pp directory\n puts\n\n puts \"system records:\"\n pp infimum.record\n pp supremum.record\n puts\n\n puts \"garbage records:\"\n each_garbage_record do |rec|\n pp rec.record\n puts\n end\n puts\n\n puts \"records:\"\n each_record do |rec|\n pp rec.record\n puts\n end\n puts\n end",
"def run\n @hydra.run\n status\n end",
"def run\n start_time = Time.now\n\n setup\n\n importer_run.update_attributes( started_at: start_time,\n source_model: source_model.name,\n destination_model: destination_model.name,\n importer_version: VERSION )\n\n # get a total count of records to process, bail out if none are found\n count = base_query(false).count\n\n logger.info \"\"\n if count == 0\n logger.info \"no #{source_model.name.pluralize} to import, exiting\"\n return\n end\n\n logger.info \"Starting import from #{source_model.table_name} into #{destination_model.name}...\"\n\n # step through the records in batches\n (0..count).step(batch_size) do |offset|\n thread_pool.schedule do\n with_connections do\n batch_elapsed_time = Benchmark.realtime do\n logger.info \"Importing from #{source_model.table_name} into #{destination_model.name} (#{offset / batch_size + 1}/#{count / batch_size + 1})\"\n\n # wipe the slate from the last batch\n prepare_for_new_batch\n\n benchmarks[:source_db] << Benchmark.realtime do\n # grab this batch of records from source\n fetch_records(offset)\n end\n\n # bail if there aren't any\n next if records.empty?\n\n logger.info \" #{records.count} source records fetched in #{benchmarks[:source_db].last}s\"\n if source_order_by\n logger.info \" #{source_order_by}: from #{records.first.read_attribute(source_order_by)} to #{records.last.read_attribute(source_order_by)}\"\n end\n\n # process this batch of records\n process_batch(records)\n\n logger.info \" #{records.count} records processed in #{benchmarks[:processing].last}s\"\n\n insert_and_update_batch\n end\n logger.info \" batch processed in #{batch_elapsed_time}s\"\n end\n end\n end\n thread_pool.shutdown\n \n print_validation_errors\n\n logger.info \"-------------------------------------------------\"\n logger.info \"Processing: #{benchmarks[:processing].sum}s total, #{benchmarks[:processing].sum / count}s per record\"\n logger.info \"source Database: #{benchmarks[:source_db].sum}s total, #{benchmarks[:source_db].sum / count}s per record\"\n logger.info \"dest Database: #{benchmarks[:destination_db].sum}s total, #{benchmarks[:destination_db].sum / count}s per record\"\n logger.info \"Total: #{Time.now - start_time}s elapsed\"\n importer_run.update_attributes( completed_at: Time.now )\n rescue Exception => e\n importer_run.update_attributes( error_trace: \"#{e.class} - #{e.message}\\n#{e.backtrace.join(\"\\n\")}\" )\n raise e\n ensure\n importer_run.update_attributes( records_created: records_created,\n records_updated: records_updated,\n duration: ((Time.now - start_time) * 1000).round,\n validation_errors: validation_errors )\n end",
"def run(limit, offset)\n timestamp = Time.now.utc.iso8601.gsub(/\\W/, '')\n @log = File.open(\"import_#{timestamp}.log\", 'w')\n create_indexes\n load_institutions\n import_objects(limit, offset)\n import_work_items(limit, offset)\n @log.close\n end",
"def run()\n end",
"def _load(dump)\n self.contents = dump\n end",
"def _dump(arg0)\n end"
] |
[
"0.75981015",
"0.75981015",
"0.6972107",
"0.66652936",
"0.66576517",
"0.6498528",
"0.6441552",
"0.6416067",
"0.63914543",
"0.63914543",
"0.63421273",
"0.6309646",
"0.6275292",
"0.61923075",
"0.61918044",
"0.61461705",
"0.6140818",
"0.6140184",
"0.6100221",
"0.602161",
"0.5991765",
"0.5990528",
"0.59049076",
"0.5899771",
"0.5881621",
"0.5853858",
"0.58526057",
"0.58521724",
"0.58288115",
"0.582306",
"0.5821065",
"0.5821065",
"0.57488334",
"0.5744669",
"0.57434297",
"0.5728481",
"0.57021713",
"0.5694764",
"0.56944335",
"0.56887144",
"0.5665231",
"0.5631402",
"0.56255734",
"0.5616304",
"0.5616304",
"0.5616304",
"0.5616304",
"0.5616304",
"0.5616304",
"0.5616304",
"0.5616304",
"0.5616304",
"0.56034905",
"0.56034905",
"0.56034905",
"0.56034905",
"0.56034905",
"0.56034905",
"0.56034905",
"0.5583805",
"0.55433255",
"0.55340177",
"0.5521186",
"0.5502144",
"0.5490468",
"0.54883796",
"0.54710126",
"0.54710126",
"0.5464757",
"0.54572105",
"0.5447092",
"0.5431661",
"0.5430828",
"0.5391463",
"0.5379595",
"0.53602785",
"0.53596646",
"0.5357929",
"0.5357494",
"0.53550553",
"0.5338919",
"0.5336292",
"0.53220487",
"0.53184325",
"0.5315642",
"0.53144675",
"0.52955455",
"0.5293828",
"0.529047",
"0.5285335",
"0.52832603",
"0.52810276",
"0.5280449",
"0.5278496",
"0.5273324",
"0.5258934",
"0.52578825",
"0.5252187",
"0.5250662",
"0.5248692"
] |
0.6629076
|
5
|
Execute the mysqldump command for the given table. Any column or table filters will be applied.
|
def dump_table(table)
conditions = conditions_for(table)
cmd = "mysqldump #{ mysql_options } --tables #{ table }"
cmd += " \"--where=#{ conditions }\"" if conditions.present?
if post_dump_command
cmd += "| #{post_dump_command}"
end
cmd += " > #{ output_dir }/#{ table }#{file_extension}"
system(cmd)
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def dump\n check_dependencies('mysqldump', 'bzip2', '/bin/sh')\n options = {}\n while arg = args.shift\n case arg\n when '-f', '--force'\n options[:force] = true\n when /^[^-]/\n raise CommandFailed, \"too many arguments passed\" if options[:filename]\n options[:filename] = arg\n else\n raise CommandFailed, \"unsupported option: #{arg}\"\n end\n end\n\n options[:filename] ||= \"#{app}-#{Time.now.strftime('%Y-%m-%d')}.sql.bz2\"\n options[:filename] += '.sql.bz2' unless options[:filename] =~ /\\.sql(\\.bz2)?$/\n options[:filename] += '.bz2' unless options[:filename] =~ /\\.bz2$/\n\n if File.exists?(options[:filename]) && !options[:force]\n raise CommandFailed, \"file already exists. use --force to override.\"\n end\n\n exec('/bin/sh', '-c',\n \"mysqldump --compress --single-transaction #{args_to_s(mysql_args(database_uri))}\" +\n pv_pipe +\n %{| bzip2 > '#{options[:filename]}'})\n end",
"def dump_table(db, tbl, num)\n t = Time.now\n timez = t.strftime(\"%m.%d.%Y\")\n logs = RESULTS + @host\n logdir = logs + '/dumps/'\n Dir.mkdir(logs) unless File.exists?(logs)\n Dir.mkdir(logdir) unless File.exists?(logdir)\n print_status(\"Attempting to dump #{db}.#{tbl}....\")\n if num.to_i == 1\n system(\"`which mysqldump` --host=#{@host} --user=#{@user} --password=#{@pass} #{db} #{tbl} --add-locks --create-options --disable-keys --extended-insert --lock-tables --quick -C --dump-date | gzip -c > #{logdir}#{db}_#{tbl}_#{timez}.sql.gz\")\n print_good(\"Table Dump Complete!\")\n print_good(\"You can view it here: #{logdir}#{db}_#{tbl}_#{timez}.sql.gz\")\n else\n system(\"`which mysqldump` --host=#{@host} --user=#{@user} --password=#{@pass} #{db} #{tbl} --add-locks --create-options --disable-keys --extended-insert --lock-tables --quick -C --dump-date > #{logdir}#{db}_#{tbl}_#{timez}.sql\")\n print_good(\"Table Dump Complete!\")\n print_good(\"You can view it here: #{logdir}#{db}_#{tbl}_#{timez}.sql\")\n end\n end",
"def perform!\n super\n\n pipeline = Pipeline.new\n dump_ext = 'sql'\n\n pipeline << mysqldump\n\n model.compressor.compress_with do |command, ext|\n pipeline << command\n dump_ext << ext\n end if model.compressor\n\n pipeline << \"#{ utility(:cat) } > \" +\n \"'#{ File.join(dump_path, dump_filename) }.#{ dump_ext }'\"\n\n pipeline.run\n if pipeline.success?\n log!(:finished)\n else\n raise Error, \"Dump Failed!\\n\" + pipeline.error_messages\n end\n end",
"def prep_command\n \"mysqldump #{dump_options} -u #{db_user} --password=#{db_password} #{database} > #{remote_backup_path}\"\n end",
"def mysqldump(options)\n name = options[:name].to_s\n append_name = options[:append_name].to_s\n dump_options = options[:dump_options].to_s\n if @db_password.to_s.empty?\n db_password = \"\"\n else\n db_password = \"-p#{@db_password}\"\n end\n file_name = \"#{@data_dir}/#{append_name}#{name}_#{dump_options}_#{@filename}.sql\"\n puts \"Dumping #{options[:name]} into #{file_name}\\n\"\n command = \" nice -n #{@nice} mysqldump -u #{@db_username} #{dump_options} #{db_password} #{@extra_dump_options} #{name} > #{file_name}\"\n puts \"\\nEXECUTING:\\n #{command}\"\n system(command)\n return file_name\nend",
"def dump(database_name, dst)\n print \"Dummping #{database_name} ...\"\n command = @mysqldump\n command << \" -u #{@username}\"\n command << \" --skip-comments\"\n command << \" --skip-extended-insert\"\n command << \" --single-transaction\"\n command << \" --quick\"\n command << \" --password=#{@password}\" unless @password.blank?\n command << \" #{database_name}\"\n command << \" > #{dst}\"\n puts command\n output = `#{command}`\n if $?.success?\n puts 'Ok'\n else\n puts 'ERROR: Check your configuration file'\n end\n end",
"def mysqldump(namespace, database_name = nil)\n puts 'Starting migration...'\n open_database_connection do\n database_uri = build_database_uri(namespace)\n database_name ||= database_uri.path.split('/')[1]\n output_file = \"#{database_name}-#{Time.now.strftime('%Y%m%d-%H%M%S')}.sql\"\n compress_cmd = compress_command(output_file)\n puts '-> Running the mysqldump'\n puts `mysqldump -h #{database_uri.host} -P #{database_uri.port} -u #{database_uri.user} --password=#{database_uri.password} #{database_name} | #{compress_cmd}`\n end\n end",
"def run\n tables.each do |table|\n dump_table(table)\n end\n end",
"def backup(dest_dir)\n logger.info \"Dumping MySQL#{db.db_and_table_names}\"\n @db.backup dest_dir\n end",
"def make_mysql_backup\n if @all_databases\n options = {\n :name => \"--all-databases\",\n :dump_options => \"\",\n :append_name => \"\"\n }\n file_name = mysqldump(options)\n compress_file(file_name)\n end\n if @databases && !@databases.empty?\n @databases.each do |db|\n options = {\n :name => db[:name].to_s,\n :dump_options => db[:dump_options].to_s,\n :append_name => db[:append_name].to_s\n }\n file_name = mysqldump(options)\n compress_file(file_name)\n end\n end\nend",
"def perform!\n super\n\n dump = \"echo '.dump' | #{sqlitedump_utility} #{path}\"\n\n pipeline = Pipeline.new\n dump_ext = \"sql\".dup\n\n pipeline << dump\n if model.compressor\n model.compressor.compress_with do |command, ext|\n pipeline << command\n dump_ext << ext\n end\n end\n\n pipeline << \"cat > '#{File.join(dump_path, dump_filename)}.#{dump_ext}'\"\n\n pipeline.run\n\n if pipeline.success?\n log!(:finished)\n else\n raise Error,\n \"#{database_name} Dump Failed!\\n\" + pipeline.error_messages\n end\n end",
"def run_cmd!\n \"pg_dump #{pg_dump_args} | psql #{new_db_name}\"\n end",
"def dump(table_name)\n dumping_tables << table_name\n end",
"def dump!\n MODEL_TO_OUTPUT_FILEMAPPING.each do |klass, filepath|\n write_table_to_file(klass, filepath, %w[created_at updated_at])\n end\n end",
"def copy_table_sql(table, opts)\n if table.is_a?(String)\n table\n else\n if opts[:options] || opts[:format]\n options = String.new\n options << \" (\"\n options << \"FORMAT #{opts[:format]}\" if opts[:format]\n options << \"#{', ' if opts[:format]}#{opts[:options]}\" if opts[:options]\n options << ')'\n end\n table = if table.is_a?(::Sequel::Dataset)\n \"(#{table.sql})\"\n else\n literal(table)\n end\n \"COPY #{table} TO STDOUT#{options}\"\n end\n end",
"def dump(table) # :nodoc:\n Marshal::dump(table)\n end",
"def psql_db_command__dump psql_db\n psql_db_command__program \"pg_dump\", psql_db\n end",
"def export_table_to_file(table)\n export_table.export_one_table(get_case_name(), table)\n end",
"def post_sql_statements( table_name, options ) # :nodoc:\n post_sql_statements = []\n if options[:on_duplicate_key_update]\n post_sql_statements << sql_for_on_duplicate_key_update( table_name, options[:on_duplicate_key_update] )\n end\n\n #custom user post_sql\n post_sql_statements << options[:post_sql] if options[:post_sql]\n\n #with rollup\n post_sql_statements << rollup_sql if options[:rollup]\n\n post_sql_statements\n end",
"def dump_table(io, table_obj)\n create_data = table_obj.data.clone\n create_data.delete(:name)\n create_data[:return_sql] = true\n\n # Get SQL for creating table and add it to IO.\n sqls = @export_db.tables.create(table_obj.name, **create_data)\n sqls.each do |sql|\n io.write(\"#{sql};\\n\")\n end\n\n\n # Try to find a primary column in the table.\n prim_col = nil\n table_obj.columns do |col|\n if col.primarykey?\n prim_col = col\n break\n end\n end\n\n\n debug \"Dumping data for table: #{table_obj.name}\"\n\n # Set up rows and way to fill rows.\n rows = []\n\n\n @db.select(table_obj.name, nil, unbuffered: true) do |row|\n rows << row\n @rows_count += 1\n\n if rows.length >= 1000\n update_status\n dump_insert_multi(io, table_obj, rows)\n end\n end\n\n\n # Dump the last rows if any.\n dump_insert_multi(io, table_obj, rows) unless rows.empty?\n end",
"def post_sql_statements( table_name, options ) # :nodoc:\n post_sql_statements = []\n\n if supports_on_duplicate_key_update? && options[:on_duplicate_key_update]\n post_sql_statements << sql_for_on_duplicate_key_update( table_name, options[:on_duplicate_key_update], options[:model], options[:primary_key], options[:locking_column] )\n elsif logger && options[:on_duplicate_key_update]\n logger.warn \"Ignoring on_duplicate_key_update because it is not supported by the database.\"\n end\n\n # custom user post_sql\n post_sql_statements << options[:post_sql] if options[:post_sql]\n\n # with rollup\n post_sql_statements << rollup_sql if options[:rollup]\n\n post_sql_statements\n end",
"def wDumpAllDatabase()\n puts \"Back up commencing...\"\n Dir.chdir('/Users/jeydurai')\n system('start_mongodump.bat')\n end",
"def copy_into_sql(table, opts)\n sql = String.new\n sql << \"COPY #{literal(table)}\"\n if cols = opts[:columns]\n sql << literal(Array(cols))\n end\n sql << \" FROM STDIN\"\n if opts[:options] || opts[:format]\n sql << \" (\"\n sql << \"FORMAT #{opts[:format]}\" if opts[:format]\n sql << \"#{', ' if opts[:format]}#{opts[:options]}\" if opts[:options]\n sql << ')'\n end\n sql\n end",
"def dump_table(io, table_obj)\n #Get SQL for creating table and add it to IO.\n sqls = @args[:db].tables.create(table_obj.name, table_obj.data, :return_sql => true)\n sqls.each do |sql|\n io.write(\"#{sql};\\n\")\n end\n \n \n #Try to find a primary column in the table.\n prim_col = nil\n table_obj.columns do |col|\n if col.primarykey?\n prim_col = col\n break\n end\n end\n \n \n #Set up rows and way to fill rows.\n rows = []\n block_data = proc do |row|\n rows << row\n @rows_count += 1\n \n if rows.length >= 1000\n self.update_status\n self.dump_insert_multi(io, table_obj, rows)\n end\n end\n \n \n #If a primary column is found then use IDQuery. Otherwise use cloned unbuffered query.\n args = {:idquery => prim_col.name.to_sym} if prim_col\n \n \n #Clone the connecting with array-results and execute query.\n @args[:db].clone_conn(:result => \"array\") do |db|\n db.select(table_obj.name, nil, args, &block_data)\n end\n \n \n #Dump the last rows if any.\n self.dump_insert_multi(io, table_obj, rows) if !rows.empty?\n end",
"def dump_table_schema(table, options=OPTS)\n gen = dump_table_generator(table, options)\n commands = [gen.dump_columns, gen.dump_constraints, gen.dump_indexes].reject{|x| x == ''}.join(\"\\n\\n\")\n \"create_table(#{table.inspect}#{', :ignore_index_errors=>true' if !options[:same_db] && options[:indexes] != false && !gen.indexes.empty?}) do\\n#{commands.gsub(/^/, ' ')}\\nend\"\n end",
"def xDumpAllDatabase()\n puts \"Back up commencing...\"\n Dir.chdir('/Users/jeydurai')\n system('start_mongodump.bat')\n end",
"def export\n Log.run(:job_type => \"export\", :name => schedule_name, \n :file => filename, :export_id => id) do |log|\n exporter = MysqlExporter.new\n exporter.export_table self # pass current object to exporter\n end\n end",
"def dump_to_file filename=nil\r\n raise \"Missing param host\" if @host.empty?\r\n raise \"Missing param name\" if @name.empty?\r\n raise \"Missing param user\" if @user.empty?\r\n raise \"Missing param pass\" if @pass.empty?\r\n\r\n filename = \"#{@name}.sql\" if filename.nil?\r\n system \"touch #{filename}\" unless Pathname(filename).exist?\r\n raise \"Invalid file \"+filename unless Pathname(filename).writable_real?\r\n\r\n %x(mysqldump --user=#{@user} --password='#{@pass}' --host=#{@host} --databases #{@name} > #{filename} 2>&1)\r\n end",
"def dump(backup_file_name)\n @mysqlcmds ||= ::Rds::S3::Backup::MySqlCmds.new(backup_server.endpoint['Address'],\n @opts['mysql_username'],\n @opts['mysql_password'],\n @opts['mysql_database'])\n\n\n\n\n @mysqlcmds.dump(backup_file_path(backup_file_name)) # returns the dump file path\n end",
"def dump_database(app_name, instance_name)\n app_config = RailsPwnerer::Config[app_name, instance_name]\n db_name, db_user, db_pass = app_config[:db_name], app_config[:db_user], app_config[:db_pass]\n\n pwnerer_user = app_config[:pwnerer_user]\n pwnerer_uid = uid_for_username(pwnerer_user)\n pwnerer_gid = gid_for_username(pwnerer_user)\n \n timestamp = Time.now.strftime '%Y%m%d%H%M%S'\n dump_file = \"db/#{app_name}.#{instance_name}_#{timestamp}.sql\"\n Dir.chdir app_config[:backup_path] do\n system(\"mysqldump --add-drop-database --add-drop-table\" +\n \" --skip-extended-insert --single-transaction\" +\n \" -u#{db_user} -p#{db_pass} #{db_name} > #{dump_file}\")\n # lockdown the file\n File.chmod(0400, dump_file)\n File.chown(pwnerer_uid, pwnerer_gid, dump_file)\n end\n end",
"def perform_backup\n \n add_memories_to_dropbox! if has_dropbox?\n\n # Clone the repo incase something is writing to it while we are backing up\n run \"cd #{@home} && git clone --bare #{@repo} #{@repo}.mirror\"\n output=run \"backup perform --trigger=daily_backup --log-path #{@dirs[:logs ]}\"\n run \"cd #{@home} && rm -fr #{@repo}.mirror\"\n \n get_timestamp(output)\n end",
"def backup_database\n #todo handle db prefix\n #todo proper error handling\n\n dbuser = @attributes[:dbuser]\n dbhost = @attributes[:dbhost]\n dbpass = @attributes[:dbpass]\n dbname = @attributes[:dbname]\n\n # see https://docs.moodle.org/20/en/Site_backup\n\n cmd = %Q{mysqldump -u #{dbuser} -h'#{dbhost}' -p'#{dbpass}' -C -Q -e --create-options '#{dbname}' | gzip -9 > '#{mk_backup_filename('database')}'}\n system cmd\n\n nil\n end",
"def copy_statement(table_name, options = {})\n format_options = replication.csv? ? \"CSV\" : \"GZIP DELIMITER ',' ESCAPE REMOVEQUOTES\"\n sql = <<-CS\n COPY #{table_name} from '#{import_file}' #{\"NOLOAD\" if options[:noload]}\n REGION '#{RailsRedshiftReplicator.s3_bucket_params[:region]}'\n CREDENTIALS 'aws_access_key_id=#{RailsRedshiftReplicator.aws_credentials[:key]};aws_secret_access_key=#{RailsRedshiftReplicator.aws_credentials[:secret]}'\n #{format_options}\n #{copy_options}\n CS\n sql.squish\n end",
"def exec__psql_db_batch__cli_or_apply_dumps *args\n psql_db = psql_db__sample_example\n db_dumps = db_dumps__sample_example\n batch = psql_db_batch__cli_or_apply_dumps psql_db, db_dumps, \"ON_ERROR_STOP=off\"\n batch_commands batch\n end",
"def db_export_cmd(server)\n raise \"No server given\" if !server\n env = \"RAILS_ENV=#{fetch(:rails_env, 'production')}\"\n import_cmd = \"cd #{release_path} && #{rake_cmd} alchemy:db:import #{env}\"\n ssh_cmd = \"#{ssh_command(server)} '#{import_cmd}'\"\n \"#{database_dump_command(database_config['adapter'])} | #{ssh_cmd}\"\n end",
"def psql_db_batch__cli_or_apply_dumps psql_db, db_dump_paths=[nil], options=\"\"\n batch = db_dump_paths.map { |db_dump_path|\n cli = psql_db_command__cli psql_db\n [cli, db_dump_path && \" #{options} -f #{quoted_shell_param db_dump_path}\"]\n }\n end",
"def csv_dump(addr, root_dir)\n filename = \"full.csv\"\n dump_dir = File.join(root_dir)\n filepath = File.join(dump_dir, filename)\n\n cmd = \"psql -c '\\\\copy (SELECT * FROM outdoor_env WHERE (apogee_w_m2 IS NOT NULL AND address=#{addr}) ORDER BY db_time DESC) To #{filepath} With CSV header\\'\"\n run_command(cmd)\nend",
"def process_table(table)\n \n $stderr.puts \"Processing table: %-15s to default table: #{table.__default_table__}\" % table.name\n Log.write_log(table.name, \"Processing table: #{table.name}\")\n table.each_column do |attr_name, maps_to|\n if maps_to.kind_of?(ReaktorColumn)\n maps_to.fill_in_defaults(table.__default_table__, attr_name)\n end\n end\n table.__set__.each do |rc|\n rc.fill_in_defaults(table.__default_table__, nil)\n end\n if table.instance_variable_defined?(:@__query__)\n query = table.__query__\n else\n query = build_query(table.name, table.__filter_rows__)\n end\n Log.write_log(table.name, \"Query: #{query}\")\n begin\n sth = $dbh_pg.execute(query)\n rescue\n $stderr.puts \"### Error in #{__FILE__} on line #{__LINE__}. See errorlog\"\n Log.write_log('error', \"Could not process query. Message: #{$!} query: #{get_query_string(sth)}.\")\n raise\n exit\n end\n while row = sth.fetch\n next unless preconditions_met?(row, table.__precondition__)\n\n process_row(table, row).each do |r|\n reaktor_insert(r) unless r.nil?\n end\n end\nend",
"def backup \n begin\n check_if_db_exists\n is_allowed = @conn.exec(\"select datallowconn from pg_catalog.pg_database where datname='#{@options[:database]}'\")[0]['datallowconn']\n if is_allowed == 'f'\n # unquiece temporarily\n set_dataallowcon_to true\n end\n\n # Check to see if the directory for backups exists and if not, create it with parents\n unless File.exist?(@options[:bkdir])\n FileUtils.mkdir_p @options[:bkdir]\n end\n filename = \"postgresbk_#{@options[:database]}_#{Time.new.strftime(\"%m%d%y%H%M%S\")}.dump\"\n\n # The below system call assumes you have passwordless access as the user passed into the executable tool\n # either due to ~/.pgpass or pg_hba.conf has your user as a 'trust' auth method\n `pg_dump -U #{@options[:user]} #{@options[:database]} -F c -f #{@options[:bkdir]}/#{filename}`\n\n rescue Exception => e\n raise e\n ensure\n if is_allowed == 'f'\n # re quiesce \n set_dataallowcon_to false\n end\n end\n end",
"def dump(db = config[\"database\"], data_selection: {})\n file_namer = Postgres::FileNameGenerator.new(working_dir)\n exclude_tables_args = data_selection.keys.map { |table| %Q|--exclude-table-data=\"#{table}\"| }\n\n [\n cli.dump(file_namer.next(db, :schema), db, [exclude_tables_args]),\n *dump_partial_selected_data(db, file_namer, data_selection)\n ]\n end",
"def dump_prod id\n target_path = File.expand_path(\"../../fixtures/trunk-#{id}.dump\", __FILE__)\n puts \"Dumping production database from Heroku (works only if you have access to the database)\"\n command = \"curl -o #{target_path} \\`heroku pg:backups public-url #{id} -a cocoapods-trunk-service\\`\"\n puts \"Executing command:\"\n puts command\n result = system command\n if result\n puts \"Production database snapshot #{id} dumped into #{target_path}\"\n else\n raise \"Could not dump #{id} from production database.\"\n end\n end",
"def mof_write_bin_file(file, exe_dest)\n data = \"0x\" + File.open(file, 'rb').read.unpack('H*').first\n begin\n @db_connection.query(\"SELECT #{data} INTO DUMPFILE '#{exe_dest}'\")\n print_good(\"Appears things were a success!\")\n return true\n rescue Mysql::Error => e\n print_error(\"Problem writing payload to file!\")\n puts \"\\t=> \".white + \"#{e}\".light_red\n if e =~ /MySQL server has gone away/\n print_error(\"This is likely due to payload which is too large in size.....\")\n print_error(\"Try compressing with UPX to shrink size down: upx 9 -qq #{file}\")\n puts \"\\t=> \".white + \"Then try again\".light_red + \".....\".white\n end\n return false\n end\n end",
"def dump_table(client, fusion_tables, fusion_table_id, backup_directory)\n backup_directory ||= \"backups\"\n FileUtils.mkdir_p backup_directory\n\n fusion_table = client.execute(\n :api_method => fusion_tables.table.get,\n :parameters => {'tableId' => \"#{fusion_table_id}\"}\n )\n fusion_table.data.to_hash\n filename = File.join(backup_directory ,\"#{fusion_table.data.to_hash['name']}-#{fusion_table_id}\")\n $stderr.puts filename\n\n File.open(\"#{filename}.json\",\"w\") do |f|\n f.write(JSON.pretty_generate(fusion_table.data.to_hash))\n end\n\n result = client.execute(\n :api_method => fusion_tables.query.sql_get,\n :parameters => {'sql' => \"SELECT * FROM #{fusion_table_id}\"}\n )\n fusion_table_data = result.data.to_hash\n\n if fusion_table_data['error']\n if fusion_table_data['error']['errors'][0]['reason'] == 'responseSizeTooLarge'\n # use Fusion Tables V2 media downloads API\n result = client.execute(\n :api_method => fusion_tables.query.sql_get,\n :parameters => {'sql' => \"SELECT * FROM #{fusion_table_id}\", 'alt' => 'media'}\n )\n File.open(\"#{filename}.csv\", 'w') { |file| file.write(result.response.body) }\n else\n $stderr.puts \"Unhandled Error:\"\n $stderr.puts fusion_table_data.inspect\n end\n else\n CSV.open(\"#{filename}.csv\", 'w') do |csv|\n if fusion_table_data['rows'] && (fusion_table_data['rows'].length > 0)\n csv << fusion_table_data['columns']\n fusion_table_data['rows'].each do |row|\n csv << row\n end\n end\n end\n end\nend",
"def perform!\n super\n\n pipeline = Pipeline.new\n\n pipeline << pg_probackup_cmd\n\n pipeline.run\n if pipeline.success?\n log!(:finished)\n else\n raise Error, \"Backup failed!\\n\" + pipeline.error_messages\n end\n end",
"def analyze_ftdump(table)\n returning Array.new do |wc|\n @myisam_ftpdump ||= `which myisam_ftdump`.strip\n @mysql_db_dir ||= File.join(@conn.instance_variable_get(\"@config\")[:db_dir], @conn.instance_variable_get(\"@config\")[:database])\n ftdump = `#{@myisam_ftpdump} -c #{@mysql_db_dir}/#{table} 0`\n ftdump.split(\"\\n\").each do |ws|\n parts = ws.split(/\\s+/)\n puts parts.inspect\n wc << [parts[3], parts[1].to_i]\n end\n end\n end",
"def db_dump_cmd\n \"mongodump\"\n end",
"def commit(on_kill = false)\n\n unless @destination.sql.empty?\n\n $log.debug(self) {'Launch last commit'} if on_kill\n\n if @use_cache\n cache_time = Time.now\n\n # And add to the cache\n @cache.commit\n\n Thread.current[:threshold_log].add 'Cache time', (Time.now - cache_time).round(3)\n end\n\n pg_time = Time.now\n # Launch SQL commands\n @destination.commit\n Thread.current[:threshold_log].add 'Replicate time', (Time.now - pg_time).round(3)\n\n end\n\n end",
"def table(table, stream)\n return if already_dumped?(table)\n\n new_stream = StringIO.new\n super(table, new_stream)\n string = new_stream.string\n\n if (parent_table = @connection.parent_table(table))\n table(parent_table, stream)\n string = inject_inherits_for_create_table(string, table, parent_table)\n string = remove_parent_table_columns(string, @connection.columns(parent_table))\n\n pindexes = Hash[@connection.indexes(parent_table).map { |index| [index.columns, index] }]\n cindexes = Hash[@connection.indexes(table).map { |index| [index.columns, index] }]\n\n string = remove_parent_table_indexes(string, (pindexes & cindexes).values)\n end\n\n # We've done this table\n dumped_tables << table\n\n stream.write string\n stream\n end",
"def open_bz2( table_name, &block )\n table_file =\n case table_name\n when Symbol\n container + \"#{table_name}.dbp.bz2\"\n when Pathname\n table_name\n else\n raise \"Don't know what to do with #{table_name.inspect}\"\n end\n\n IO.popen \"#{STREAM_DCMP} #{table_file}\", &block\n end",
"def pg_dump_args\n ['-x', '-O', if schema_only\n '-s'\n end, conf['database']].compact.join(' ')\n end",
"def process\n create, insert, table = extract_sql #Extract mysql create/insert statements from the dump file\n raise \"Couldn't extract create syntax from MySql Dump File\" if create.nil?\n create = escape_create_string(create)\n begin\n @connection.execute(\"DROP TABLE #{table}\") rescue ''#Drop existing table first\n @connection.execute(create) #Recreate the table \n if insert && @import_data\n values = row_values(insert) \n values.each do |val|\n sql = \"INSERT INTO #{table} VALUES #{val}\"\n begin\n @connection.execute(sql) #Insert rows\n rescue Exception => e\n puts e.message\n puts sql\n puts \"table #{table}\"\n end\n end\n else\n puts \"There's no records to be added\" if @import_data && !insert\n end\n rescue Exception => e\n puts e.message\n puts \"table #{table}\"\n end\n end",
"def dump( out = Dumper.new )\n out.dump \"table #{out_name}\" do\n for column in columns\n column.dump( out )\n end\n for index in indexes\n index.dump( out )\n end\n for foreign_key in foreign_keys\n foreign_key.dump( out )\n end\n end\n end",
"def run(direction, database, version)\n ensure_schema_migrations_table_exists(database)\n Database.process(database) do\n ActiveRecord::Migrator.run(direction, migration_paths(database), version)\n end\n dump_schema(database)\n end",
"def extract_table(table)\n puts \"Extracting from #{table['name']}\\n\"\n puts \">>> assoc: #{table['associations']}\"\n extract_associations(table['associations']) if table['associations']\n @outfile.puts create_table(table)\n columns = table_columns(table)\n ids = table_ids(table)\n query = table_query(table, columns, ids)\n results = db_client.query(query)\n results.each do |row|\n @outfile.puts make_insert(table, columns, results.fields, row)\n end\n end",
"def execute(sql, *args, &block)\n @db.execute(rewrite_table_names(sql), *args, &block)\n end",
"def copy_and_clear_table(params)\n table_name = params[:table]\n game = params[:game]\n team_table = params[:team_table]\n result_filter = params[:result_filter] # for Log only\n main_class = table_name.constantize\n archive_class = (\"Archive#{table_name}\").constantize\n\n main_class.where(game_id: game.id).each do |row|\n\n condition = result_filter.blank? || result_filter.include?(row.result_code) # check the result_filter\n condition &&= archive_class.find_by_id(row.id).blank? # check if this record is already archived\n if condition\n archive_instance = archive_class.new\n main_class.column_names.each do |column|\n if column == 'team_id'\n # if this team is not in game requests\n unless team_table[row.team_id]\n team = Team.find(row.team_id)\n archive_team = ArchiveTeam.create(name: team.name, alternative_name: team.alternative_name,\n image_url: team.image_url, team_id: team.id, game_id: game.id)\n team_table.merge!({team.id => archive_team.id})\n end\n\n archive_instance.send( \"#{column}=\", team_table[row.team_id] )\n else\n archive_instance.send(\"#{column}=\", row.send(column))\n end\n end\n archive_instance.save\n end\n end\n main_class.where(game_id: game.id).map(&:delete)\n end",
"def perform!\n super\n\n pipeline = Pipeline.new\n dump_ext = 'ldif'\n\n pipeline << slapcat\n\n model.compressor.compress_with do |command, ext|\n pipeline << command\n dump_ext << ext\n end if model.compressor\n\n pipeline << \"#{ utility(:cat) } > \" +\n \"'#{ File.join(dump_path, dump_filename) }.#{ dump_ext }'\"\n\n pipeline.run\n if pipeline.success?\n log!(:finished)\n else\n raise Error, \"Dump Failed!\\n\" + pipeline.error_messages\n end\n end",
"def obfuscate\n @mysqlcmds ||= ::Rds::S3::Backup::MySqlCmds.new(backup_server.endpoint['Address'],\n @opts['mysql_username'],\n @opts['mysql_password'],\n @opts['mysql_database'])\n\n @mysqlcmds.exec(@opts['obfuscate_sql'])\n end",
"def dump_table_schema(table, options={})\n table = table.value.to_s if table.is_a?(SQL::Identifier)\n raise(Error, \"must provide table as a Symbol, String, or Sequel::SQL::Identifier\") unless [String, Symbol].any?{|c| table.is_a?(c)}\n s = schema(table).dup\n pks = s.find_all{|x| x.last[:primary_key] == true}.map{|x| x.first}\n options = options.merge(:single_pk=>true) if pks.length == 1\n m = method(:column_schema_to_generator_opts)\n im = method(:index_to_generator_opts)\n begin\n indexes = indexes(table).sort_by{|k,v| k.to_s} if options[:indexes] != false\n rescue Sequel::NotImplemented\n nil\n end\n gen = Schema::Generator.new(self) do\n s.each{|name, info| send(*m.call(name, info, options))}\n primary_key(pks) if !@primary_key && pks.length > 0\n indexes.each{|iname, iopts| send(:index, iopts[:columns], im.call(table, iname, iopts))} if indexes\n end\n commands = [gen.dump_columns, gen.dump_constraints, gen.dump_indexes].reject{|x| x == ''}.join(\"\\n\\n\")\n \"create_table(#{table.inspect}#{', :ignore_index_errors=>true' if !options[:same_db] && options[:indexes] != false && indexes && !indexes.empty?}) do\\n#{commands.gsub(/^/o, ' ')}\\nend\"\n end",
"def dump!\n run(mongodump)\n end",
"def dump!\n run(mongodump)\n end",
"def dump( out = Dumper.new )\n for table in tables\n table.dump( out )\n end\n out\n end",
"def generate(table_name, statement)\n alter_argument = AlterArgument.new(statement)\n dsn = DSN.new(connection_details.database, table_name)\n\n \"#{command} #{all_options} #{dsn} #{alter_argument}\"\n end",
"def save\n hash = self.attr_hash\n sql_hash = hash.to_s.delete \"\\>\"\n\n CONNECTION.execute(\"UPDATE '#{tablename}' SET #{sql_hash[1...-1]} WHERE id = ?;\", @id)\n \"Saved.\"\n end",
"def restore_table( table_file )\n logger.info \"restoring from #{table_file}\"\n pump.table_name = table_file.basename.sub_ext('').sub_ext('').to_s.to_sym\n open_bz2 table_file do |io|\n pump.io = io\n pump.restore filename: table_file\n end\n end",
"def write_sql model_name, model_attributes,output\n model_attributes.each do|key,query|\n sql= ActiveRecord::Base.connection();\n (sql.select_all query).each do |row|\n make_triples(row,model_name,\"\")\n end\n end\n end",
"def dump_table_generator(table, options=OPTS)\n s = schema(table, options).dup\n pks = s.find_all{|x| x.last[:primary_key] == true}.map(&:first)\n options = options.merge(:single_pk=>true) if pks.length == 1\n m = method(:recreate_column)\n im = method(:index_to_generator_opts)\n\n if options[:indexes] != false && supports_index_parsing?\n indexes = indexes(table).sort\n end\n\n if options[:foreign_keys] != false && supports_foreign_key_parsing?\n fk_list = foreign_key_list(table)\n \n if (sfk = options[:skipped_foreign_keys]) && (sfkt = sfk[table])\n fk_list.delete_if{|fk| sfkt.has_key?(fk[:columns])}\n end\n\n composite_fks, single_fks = fk_list.partition{|h| h[:columns].length > 1}\n fk_hash = {}\n\n single_fks.each do |fk|\n column = fk.delete(:columns).first\n fk.delete(:name)\n fk_hash[column] = fk\n end\n\n s = s.map do |name, info|\n if fk_info = fk_hash[name]\n [name, fk_info.merge(info)]\n else\n [name, info]\n end\n end\n end\n\n create_table_generator do\n s.each{|name, info| m.call(name, info, self, options)}\n primary_key(pks) if !@primary_key && pks.length > 0\n indexes.each{|iname, iopts| send(:index, iopts[:columns], im.call(table, iname, iopts, options))} if indexes\n composite_fks.each{|fk| send(:foreign_key, fk[:columns], fk)} if composite_fks\n end\n end",
"def dumpable\n set(variable_table: variable_table.dumpable, rule_table: rule_table.dumpable)\n end",
"def execute_batch(sql, *args, &block)\n @db.execute_batch(rewrite_table_names(sql), *args, &block)\n end",
"def export(file = nil)\n Command::Batch::Export.new(\n *command_params(\n file: file\n )\n ).execute\n rescue => e\n catch_errors(e)\n end",
"def backup\n run_command 'pgbackups:capture', %W(--expire #{database})\n end",
"def exec_query_truncate_table\n execute_remote(query_truncate_table)\n end",
"def do_get_server_dump(address, table_name)\n return nil unless Module.const_get(table_name)\n # call the remote server sending the table that we want to import\n req = http_open_post(address, $SYNC_CONFIG[:dump_uri]) do |request|\n start_id = current_client.last_synced || -1\n form_data = { 'table_name' => table_name }\n request.set_form_data form_data\n end\n\n result ||= ActsAsReplica::Structs::SyncDumpResult.new\n result.status = case req\n when Net::HTTPSuccess\n fix_model_loading\n package = YAML.load(expand(req.body))\n result.last_synced = package.last_synced\n result.total = package.total.to_i\n \n # unpacks the received CSV package into a temporary file\n tmp_file = File.join(RAILS_ROOT, 'tmp', \"#{UUID.timestamp_create().to_s}.csv\")\n begin\n File.open(tmp_file, 'w') { |f| f.write package.payload } \n if Replica.compare_version(package.version)\n result.errors << \"Remote #{package.version} and Local #{Replica.version} versions are different.\"\n 500\n else\n if result.total > 0\n # loads the CSV file into the table\n model_class = Module.const_get(table_name)\n model_class.import_from_csv(tmp_file)\n result.result_message = \"Loaded #{model_class.count} rows from #{File.size(tmp_file)} csv file.\"\n end\n 200\n end\n ensure\n File.delete(tmp_file) if File.exists?(tmp_file)\n end\n else\n result.errors << \"Unknown error #{req.to_yaml}.\"\n 500\n end \n result\n end",
"def copy_into(table, opts=OPTS)\n data = opts[:data]\n data = Array(data) if data.is_a?(String)\n\n if block_given? && data\n raise Error, \"Cannot provide both a :data option and a block to copy_into\"\n elsif !block_given? && !data\n raise Error, \"Must provide either a :data option or a block to copy_into\"\n end\n\n synchronize(opts[:server]) do |conn|\n conn.execute(copy_into_sql(table, opts))\n begin\n if block_given?\n while buf = yield\n conn.put_copy_data(buf)\n end\n else\n data.each{|buff| conn.put_copy_data(buff)}\n end\n rescue Exception => e\n conn.put_copy_end(\"ruby exception occurred while copying data into PostgreSQL\")\n ensure\n conn.put_copy_end unless e\n while res = conn.get_result\n raise e if e\n check_database_errors{res.check}\n end\n end\n end \n end",
"def threaddump(gear, component)\n args = build_base_gear_args(gear)\n cart = component.cartridge_name\n args = build_base_component_args(component, args)\n\n run_cartridge_command(cart, gear, \"threaddump\", args)\n end",
"def dump\n _dump do |row|\n codec.encode( row.values, io ) unless dry_run?\n end\n ensure\n io.flush\n end",
"def dump(io)\n print \"Going through tables.\\n\" if @debug\n @rows_count = 0\n \n if @args[:tables]\n tables = @args[:tables]\n else\n tables = @args[:db].tables.list.values\n end\n \n if @on_status\n @on_status.call(:text => \"Preparing.\")\n \n @rows_count_total = 0\n tables.each do |table_obj|\n @rows_count_total += table_obj.rows_count\n end\n end\n \n tables.each do |table_obj|\n table_obj = @args[:db].tables[table_obj] if table_obj.is_a?(String) or table_obj.is_a?(Symbol)\n \n #Figure out keys.\n @keys = []\n table_obj.columns do |col|\n @keys << col.name\n end\n \n @table_obj = table_obj\n self.update_status\n print \"Dumping table: '#{table_obj.name}'.\\n\" if @debug\n self.dump_table(io, table_obj)\n end\n end",
"def export_database(server)\n puts \"Exporting the database. Please wait...\"\n system db_export_cmd(server)\n end",
"def psql_db_batch__cli_or_generate_dumps psql_db, db_dump_paths=[nil], options=\"\"\n psql_db = array__from(psql_db)\n db_dump_paths = array__from(db_dump_paths)\n batch = db_dump_paths.map { |db_dump_path|\n cli = psql_db_command__cli psql_db\n pg_dump = psql_db_command__dump psql_db\n program = db_dump_path && pg_dump || cli\n [program, db_dump_path && \" #{options} -f #{quoted_shell_param db_dump_path}\"]\n }\n end",
"def export_db\n result = `ssh -t #{USERNAME}@#{STAGING_SERVER} 'mysqldump -u#{MYSQL_USER} -p#{MYSQL_PASSWORD} #{SITECODE} > /tmp/#{SITECODE}.sql'`\n if $?.success?\n puts 'Database exported from staging.'\n else\n puts 'FAILED exporting database from staging'\n puts result\n exit(2)\n end\nend",
"def run_backups\n puts \"Starting backups...\" if options[:verbose]\n for database in database_list\n Dir.chdir(temp_directory)\n name, user, pass = database.split(\",\")\n password = pass.strip.empty? ? '' : \"-p#{pass.strip}\"\n tgz_filename = \"#{name}.#{DATE}.#{TIME}.tgz\"\n # stop the slave if necessary\n puts \"Stopping the slave...\" if options[:verbose] && options[:slave]\n `mysql -u #{user} #{password} --execute='stop slave;'` if options[:slave]\n \n # switch to the current database and backup each table\n tables = `echo 'show tables' | mysql -u #{user} #{password} #{name} | grep -v Tables_in_`\n for table in tables\n table.strip!\n puts \"Backing up table #{table}...\" if options[:verbose]\n filename = \"#{table}.#{DATE}.#{TIME}.sql\"\n `mysqldump --add-drop-table --allow-keywords -q -c -u #{user} #{password} #{name} #{table} > #{filename}`\n end\n \n # restart the slave if necessary\n puts \"Restarting the slave...\" if options[:verbose] && options[:slave]\n `mysql -u #{user} #{password} --execute='start slave;'` if options[:slave]\n \n # zip it up and move it to the backup directory\n puts \"Completed backups, zipping it up...\" if options[:verbose]\n `tar -zcvf #{backup_directory}/#{tgz_filename} *`\n puts \"Cleaning up...\" if options[:verbose]\n Dir.chdir(backup_directory)\n `rm -rf #{temp_directory}`\n \n # copy it to any remote hosts if needed\n scp_results(tgz_filename) unless scp_hosts.empty?\n puts \"And we're done!\" if options[:verbose]\n end\n end",
"def extract(outfile_path = 'dbextractor_output.sql')\n @outfile = open(outfile_path, 'w')\n puts \"Loading config...\\n\"\n dbconfig = parse_db_config(@config_path)\n\n extract_tables(dbconfig[\"tables\"])\n @outfile.close\n end",
"def print_table(table)\r\n if table_exists?(table)\r\n names = get_column_names(table)\r\n puts names.join(\"|\")\r\n table_info = @db.execute(\"SELECT * FROM #{table}\")\r\n print_execute(table_info)\r\n end\r\n end",
"def _dump\n unless @table.empty?\n key, value = @table.first\n suffix = _eval_join(\n *[\n key == \"_function_\" ? nil : key,\n !value.nil? ? value._dump : nil,\n ].compact\n )\n end\n if _fn_name\n args = _fn_args.map do |arg|\n if arg.respond_to?(:_dump)\n arg._dump\n elsif arg.is_a?(::Symbol)\n quote = __single_quote_strings ? \"'\" : '\"'\n \"#{quote}#{::Bogo::Utility.camel(arg.to_s, false)}#{quote}\"\n elsif arg.is_a?(::String) && __single_quote_strings\n \"'#{arg}'\"\n else\n arg.inspect\n end\n end.join(\", \")\n unless _fn_name.to_s.empty?\n function_name = args.empty? ? \"#{_fn_name}#{__empty_argument_list}\" : \"#{_fn_name}(#{args})\"\n end\n internal = _eval_join(\n *[\n function_name,\n suffix,\n ].compact\n )\n if root? || (!__single_anchor? && function_name)\n if !root? && __quote_nested_funcs?\n quote = __single_quote_strings ? \"'\" : '\"'\n end\n \"#{quote}#{__anchor_start}#{internal}#{__anchor_stop}#{quote}\"\n else\n internal\n end\n else\n suffix\n end\n end",
"def execute(sql)\n tmp = Digest::MD5.hexdigest(sql)\n tmp_path = \"#{TMP_DIR}/#{tmp}\"\n File.write tmp_path, sql \n scp_upload! tmp_path, tmp_path \n result = C.exec! \"psql -A -t -d #{TMP_DB} -f #{tmp_path}\"\n C.exec! \"rm #{tmp_path}\"\n File.delete tmp_path\n result\nend",
"def run\n connect\n src = @source_connection\n dest = @destination_connection\n\n log \"BEGINNING MIGRATION\\n\" +\n \"-------------------\"\n\n # Iterate through table mappings\n maps.each do |table|\n log \"MIGRATING TABLE #{table.source}\"\n\n # Fetch all rows from source table\n rows = src[table.source.to_sym].all\n \n # Iterate through the records and migrate each one\n rows.each do |row|\n new_record = {}\n\n # Identify the columns we need to grab data from\n table.maps.each do |column|\n new_record.merge!(\n column.destination.to_sym => row[column.source.to_sym]\n )\n end\n\n # Insert new record to destination table\n log \" -> Inserting record into #{table.destination}\"\n dest[table.destination.to_sym].insert(new_record)\n end\n end\n\n log \"-------------------\\n\" +\n \"MIGRATION COMPLETED\"\n end",
"def _dump( &encode_block )\n return enum_for(__method__) unless block_given?\n case\n when table_dataset.respond_to?( :stream )\n stream_dump &encode_block\n\n when primary_keys.empty?\n paginated_dump &encode_block\n\n when primary_keys.all?{|i| i == :id }\n min_max_dump &encode_block\n\n else\n inner_dump &encode_block\n end\n end",
"def snapshot(table, snapshot_name, *args)\n # Table name should be a string\n raise(ArgumentError, 'Table name must be of type String') unless table.is_a?(String)\n\n # Snapshot name should be a string\n raise(ArgumentError, 'Snapshot name must be of type String') unless\n snapshot_name.is_a?(String)\n\n table_name = TableName.valueOf(table)\n if args.empty?\n @admin.snapshot(snapshot_name, table_name)\n else\n args.each do |arg|\n ttl = arg[TTL]\n ttl = ttl ? ttl.to_java(:long) : -1\n snapshot_props = java.util.HashMap.new\n snapshot_props.put(\"TTL\", ttl)\n max_filesize = arg[MAX_FILESIZE]\n max_filesize = max_filesize ? max_filesize.to_java(:long) : -1\n snapshot_props.put(\"MAX_FILESIZE\", max_filesize)\n if arg[SKIP_FLUSH] == true\n @admin.snapshot(snapshot_name, table_name,\n org.apache.hadoop.hbase.client.SnapshotType::SKIPFLUSH, snapshot_props)\n else\n @admin.snapshot(snapshot_name, table_name, snapshot_props)\n end\n end\n end\n end",
"def cmd_backup argv\n setup argv\n command = @hash['command']\n name = @hash['name']\n response = @api.backup(command, name)\n msg response\n return response\n end",
"def inner_dump( &encode_block )\n # could possibly overrride Dataset#paginate(page_no, page_size, record_count=nil)\n on_conditions = primary_keys.map{|f| [f,f]}.to_h\n (0..table_dataset.count).step(page_size).each do |offset|\n limit_dataset = table_dataset.select( *primary_keys ).limit( page_size, offset ).order( *primary_keys )\n page = table_dataset.join( limit_dataset, on_conditions ).order( *primary_keys ).qualify(table_name)\n logger.info \"#{__method__} #{table_name} #{offset}\"\n logger.debug page.sql\n page.each &encode_block\n end\n end",
"def perform!\n log!\n\n begin\n lock_database if @lock.eql?(true)\n if collections_to_dump.is_a?(Array) and not collections_to_dump.empty?\n specific_collection_dump!\n else\n dump!\n end\n unlock_database if @lock.eql?(true)\n rescue => exception\n unlock_database if @lock.eql?(true)\n raise exception\n end\n end",
"def copy_to(db, args = {})\r\n data[\"tables\"].each do |table|\r\n table_args = nil\r\n table_args = args[\"tables\"][table[\"name\"].to_s] if args and args[\"tables\"] and args[\"tables\"][table[\"name\"].to_s]\r\n next if table_args and table_args[\"skip\"]\r\n table.delete(\"indexes\") if table.key?(\"indexes\") and args[\"skip_indexes\"]\r\n db.tables.create(table[\"name\"], table)\r\n \r\n limit_from = 0\r\n limit_incr = 1000\r\n \r\n loop do\r\n ins_arr = []\r\n q_rows = self.select(table[\"name\"], {}, {\"limit_from\" => limit_from, \"limit_to\" => limit_incr})\r\n while d_rows = q_rows.fetch\r\n col_args = nil\r\n \r\n if table_args and table_args[\"columns\"]\r\n d_rows.each do |col_name, col_data|\r\n col_args = table_args[\"columns\"][col_name.to_s] if table_args and table_args[\"columns\"]\r\n d_rows[col_name] = \"\" if col_args and col_args[\"empty\"]\r\n end\r\n end\r\n \r\n ins_arr << d_rows\r\n end\r\n \r\n break if ins_arr.empty?\r\n \r\n db.insert_multi(table[\"name\"], ins_arr)\r\n limit_from += limit_incr\r\n end\r\n end\r\n end",
"def marshal_dump\n @table\n end",
"def alter_table_sql(table, op)\n case op[:op]\n when :rename_column\n unless sch = op[:schema]\n raise(Error, \"can't find existing schema entry for #{op[:name]}\") unless sch = op[:schema] || schema(table).find{|c| c.first == op[:name]}\n sch = sch.last\n end\n [\n alter_table_sql(table, :op=>:add_column, :name=>op[:new_name], :default=>sch[:ruby_default], :type=>sch[:db_type], :null=>sch[:allow_null]),\n from(table).update_sql(op[:new_name]=>op[:name]),\n alter_table_sql(table, :op=>:drop_column, :name=>op[:name])\n ]\n when :set_column_null, :set_column_default\n raise(Error, \"can't find existing schema entry for #{op[:name]}\") unless sch = op[:schema] || schema(table).find{|c| c.first == op[:name]}\n sch = sch.last\n\n sch = if op[:op] == :set_column_null\n sch.merge(:allow_null=>op[:null])\n else\n sch.merge(:ruby_default=>op[:default])\n end\n\n [\n alter_table_sql(table, :op=>:rename_column, :name=>op[:name], :new_name=>:sequel_access_backup_column, :schema=>sch),\n alter_table_sql(table, :op=>:rename_column, :new_name=>op[:name], :name=>:sequel_access_backup_column, :schema=>sch)\n ]\n else\n super\n end\n end",
"def perform!\n super\n\n lock_database if @lock\n @only_collections.empty? ? dump! : specific_collection_dump!\n\n rescue => err\n raise Errors::Database::MongoDBError.wrap(err, 'Database Dump Failed!')\n ensure\n unlock_database if @lock\n package! unless err\n end",
"def execute_export cmd, metadata\n metadata.store!\n result = \"\"\n\n # Execute Export command on the source DB server\n result = metadata.ssh.exec!(cmd)\n\n return result\n end",
"def insert_data(output, db, table)\n result = query(db, \"SELECT * FROM #{table}\", \"hash\")\n changes = false\n result.each do |row|\n to_insert(output, table, row)\n changes = true\n end\n output << \"\\n\" if changes\n end",
"def get_copy_data(sql, file)\n query = 'COPY (' + sql + \") TO STDOUT WITH DELIMITER ',' CSV HEADER\"\n @db.copy_data query do\n until (row = @db.get_copy_data).nil?\n file.write(row)\n end\n end\n end",
"def run_bootstrap\n bootstrap = File.join(@manifest_path, UP_DIRNAME, BOOTSTRAP_FILENAME)\n execute_migration('bootstrap.sql', bootstrap)\n end",
"def set_pruned(table)\n if !@dry_run\n conn = connection\n res = conn.query(\"UPDATE pruned SET prune_time = now() WHERE table_name = '#{table}'\")\n if conn.affected_rows == 0\n conn.query(\"INSERT INTO pruned (table_name, prune_time) VALUES ('#{table}', now())\")\n end\n verbose \" Updated prune_time.\"\n end\n end"
] |
[
"0.58444566",
"0.57970226",
"0.5688865",
"0.56162184",
"0.5516886",
"0.5512034",
"0.5509626",
"0.5462216",
"0.5455626",
"0.54280394",
"0.5311821",
"0.52377546",
"0.5107581",
"0.50713354",
"0.50572675",
"0.5049411",
"0.5007488",
"0.49645013",
"0.4902564",
"0.48862857",
"0.48524424",
"0.48377573",
"0.4788537",
"0.478522",
"0.47664833",
"0.47437486",
"0.47211695",
"0.4689695",
"0.46891537",
"0.46755812",
"0.46609312",
"0.46468523",
"0.46399376",
"0.46228728",
"0.4604137",
"0.45751277",
"0.45610118",
"0.45500422",
"0.45289147",
"0.45242092",
"0.45169935",
"0.44768432",
"0.44684532",
"0.4468446",
"0.4467437",
"0.44527707",
"0.44310945",
"0.44152477",
"0.4414693",
"0.44117862",
"0.4386005",
"0.43424043",
"0.43393844",
"0.4327141",
"0.43262893",
"0.43176147",
"0.43113527",
"0.42900044",
"0.42894936",
"0.42782897",
"0.42782897",
"0.42726946",
"0.42614013",
"0.42612553",
"0.4259571",
"0.42590785",
"0.42273793",
"0.4222594",
"0.4217685",
"0.41990212",
"0.41985896",
"0.41947374",
"0.4193462",
"0.4192354",
"0.41886958",
"0.4188595",
"0.4180563",
"0.41716108",
"0.41715962",
"0.41708156",
"0.41550368",
"0.41522068",
"0.41492817",
"0.4144494",
"0.41431284",
"0.4141645",
"0.413942",
"0.41289318",
"0.41243088",
"0.41215384",
"0.4119298",
"0.4111791",
"0.41112104",
"0.40821928",
"0.4076603",
"0.40641397",
"0.4062908",
"0.4058623",
"0.40477318",
"0.40474597"
] |
0.78548235
|
0
|
Return a string suitable for use in the mysqldump command line where option.
|
def conditions_for(table)
conditions = []
column_filters.keys.each do |type|
conditions << column_filter_conditions_for(table, type)
end
# exclude all rows from table filtered tables
conditions << '1 = 0' if table_filters.include?(table)
conditions.compact.join(' or ')
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def pg_dump_args\n ['-x', '-O', if schema_only\n '-s'\n end, conf['database']].compact.join(' ')\n end",
"def dump_path\n dump = {:postgresql => 'pg_dump',:mysql => 'mysqldump'}\n # Note: the 'mysqldump_path' config option is DEPRECATED but keeping this in for legacy config file support\n @fu_conf[:mysqldump_path] || @fu_conf[:dump_path] || dump[@db_conf[:adapter].intern]\n end",
"def mysqldump(options)\n name = options[:name].to_s\n append_name = options[:append_name].to_s\n dump_options = options[:dump_options].to_s\n if @db_password.to_s.empty?\n db_password = \"\"\n else\n db_password = \"-p#{@db_password}\"\n end\n file_name = \"#{@data_dir}/#{append_name}#{name}_#{dump_options}_#{@filename}.sql\"\n puts \"Dumping #{options[:name]} into #{file_name}\\n\"\n command = \" nice -n #{@nice} mysqldump -u #{@db_username} #{dump_options} #{db_password} #{@extra_dump_options} #{name} > #{file_name}\"\n puts \"\\nEXECUTING:\\n #{command}\"\n system(command)\n return file_name\nend",
"def generate_where tab_name, argv #, att_argv\n\n\n\t\t\t\t# Parsing conditional hashing argument.\t\n\t\tcondition = \"\"\n\t\tcondition << \"( \"\n\t\ti = 0\n\t\tbegin\n\t\t\tif \"condition\" != argv.keys[i] \n\t\t\t\tcondition << argv.keys[i] \n\t\t\t\tcondition << \"=\"\n\t\t\t\tcondition << \"\\\"#{argv[argv.keys[i]]}\\\"\"\n\t\t\t\tcondition << \" \"\n\t\t\t\ti = i.to_i + 1\n\t\t\telse\n\t\t\t\tcondition << \"\\\"#{argv[argv.keys[i]]}\\\"\"\n\t\t\t\tcondition << \" \"\n\t\t\t\ti = i.to_i + 1\n\t\t\tend\t\n\t\tend while i < argv.length\n\t\tcondition << \")\"\n\n\n\t\t#if att_argv.length == 0\n\t\t\tquery = \" SELECT * FROM #{tab_name} WHERE #{condition} \"\n\t\t\tputs query\n\t\t\treturn query\n\t\t#else\n\t\t\t#query = \" SELECT #{att_argv * \",\"} FROM #{tab_name} WHERE #{condition} \"\t\n\t\t\t#puts query\n\t\t\t#return query\n\t\t#end\n\tend",
"def prep_command\n \"mysqldump #{dump_options} -u #{db_user} --password=#{db_password} #{database} > #{remote_backup_path}\"\n end",
"def db_dump_cmd\n \"mongodump\"\n end",
"def string\n @sql_string = \"SELECT #{@select} FROM #{@from}#{@join} WHERE #{@where};\"\n end",
"def sql_command_string(query, database, ctrl, grep_for = nil)\n raw_query = query.is_a?(String) ? query : query.join(\";\\n\")\n Chef::Log.debug(\"Control Hash: [#{ctrl.to_json}]\\n\")\n cmd = \"/usr/bin/mysql -B -e \\\"#{raw_query}\\\"\"\n cmd << \" --user=#{ctrl[:user]}\" if ctrl && ctrl.key?(:user) && !ctrl[:user].nil?\n cmd << \" -p#{ctrl[:password]}\" if ctrl && ctrl.key?(:password) && !ctrl[:password].nil?\n cmd << \" -h #{ctrl[:host]}\" if ctrl && ctrl.key?(:host) && !ctrl[:host].nil? && ctrl[:host] != 'localhost'\n cmd << \" -P #{ctrl[:port]}\" if ctrl && ctrl.key?(:port) && !ctrl[:port].nil? && ctrl[:host] != 'localhost'\n cmd << \" -S #{default_socket}\" if ctrl && ctrl.key?(:host) && !ctrl[:host].nil? && ctrl[:host] == 'localhost'\n cmd << \" #{database}\" unless database.nil?\n cmd << \" | grep #{grep_for}\" if grep_for\n Chef::Log.debug(\"Executing this command: [#{cmd}]\\n\")\n cmd\n end",
"def dump\n check_dependencies('mysqldump', 'bzip2', '/bin/sh')\n options = {}\n while arg = args.shift\n case arg\n when '-f', '--force'\n options[:force] = true\n when /^[^-]/\n raise CommandFailed, \"too many arguments passed\" if options[:filename]\n options[:filename] = arg\n else\n raise CommandFailed, \"unsupported option: #{arg}\"\n end\n end\n\n options[:filename] ||= \"#{app}-#{Time.now.strftime('%Y-%m-%d')}.sql.bz2\"\n options[:filename] += '.sql.bz2' unless options[:filename] =~ /\\.sql(\\.bz2)?$/\n options[:filename] += '.bz2' unless options[:filename] =~ /\\.bz2$/\n\n if File.exists?(options[:filename]) && !options[:force]\n raise CommandFailed, \"file already exists. use --force to override.\"\n end\n\n exec('/bin/sh', '-c',\n \"mysqldump --compress --single-transaction #{args_to_s(mysql_args(database_uri))}\" +\n pv_pipe +\n %{| bzip2 > '#{options[:filename]}'})\n end",
"def to_s()\n @sql ||= print_to(Printer.new()).to_s()\n end",
"def mongodump_options\n options = String.new\n options += \" --username='#{user}' \" unless user.blank?\n options += \" --password='#{password}' \" unless password.blank?\n options += \" --host='#{host}' \" unless host.blank?\n options += \" --port='#{port}' \" unless port.blank?\n options += \" --db='#{database}' \" unless database.blank?\n options\n end",
"def to_dump(opts={})\n dump = \"\"\n dump << \"add_column_check_constraint #{table_name.to_s}, #{column_name.to_s}\" unless opts[:inline]\n dump << \", check: #{check.inspect}\"\n dump << \"\\n\"\n dump\n end",
"def where_clause_from(options)\n return nil unless options\n query = String.new()\n query = sanitize_wql(options)\n puts \"NOW, we'll execute this: #{query}\"\n return query\n end",
"def short_commit() (commit || \"\")[0..6] end",
"def build_sql(value_str)\n \"COPY #{value_str} FROM STDIN WITH (FORMAT csv, DELIMITER E'\\x0B')\"\n end",
"def build_opts_string(*additional_opts)\n options = build_opts + additional_opts\n options = options << \"2>&1 | ocunit2junit\" if test_output == :junit\n return options.compact.join(\" \")\n end",
"def to_s\n \n options = \"\"\n options << \" -t\" << self.target_type unless self.target_type.nil?\n options << \" -m\" << self.min_size.to_s unless self.min_size.nil?\n options << \" -n\" << self.max_size.to_s unless self.max_size.nil?\n options << \" -s\" << self.min_supp.to_s unless self.min_supp.nil?\n options << \" -c\" << self.min_conf.to_s unless self.min_conf.nil?\n options << \" -e\" << self.add_eval.to_s unless self.add_eval.nil?\n options << \" -d\" << self.min_add_eval.to_s unless self.min_add_eval.nil?\n options << \" -v\" << \"\\\";\" << self.output_format << \"\\\"\" unless self.output_format.nil?\n \n options\n \n end",
"def dest\n if (@currentCommand.include?'=')\n result = \"\"\n if (@currentCommand.slice(0) == \"0\")\n result = \"null\"\n else\n endmarks = Array['=', ';']\n i = 0\n\n while (endmarks.include?(@currentCommand.slice(i)) != true)\n result += @currentCommand.slice(i)\n i += 1\n end\n end\n else\n result = 'null'\n end\n return result\n end",
"def to_sql\n @line.gsub(OUTPUT_REGEX, '')\n end",
"def generate_where_phrase\n opt_where = @opt[:where]\n\n return '' if opt_where.nil? or opt_where.empty?\n\n where = []\n opt_where.each_pair { |k, v|\n a = @entity.attr(k)\n r = @entity.rel(k)\n if a\n raise \"Field #{k.inspect} is not indexed\" unless a.indexed?\n wh = where_entry_single_column(k, v)\n elsif r\n if r.multi?\n @tables << \" LEFT JOIN `#{r.sql_table}` ON `#{@entity.name}`.`_id`=`#{r.sql_table}`.`#{r.sql_column1}`\"\n wh = where_entry_join_id(r, v)\n else\n wh = where_entry_single_column(k, v)\n end\n elsif k == '_id'\n # ID column is not an attribute, so it requires special case\n wh = where_entry_single_column(k, v)\n else\n raise \"Invalid field name: #{k.inspect}\" unless a\n end\n\n where << wh if wh\n }\n\n @where_phrase = where.empty? ? '' : \"WHERE #{where.join(' AND ')}\"\n end",
"def table_options_sql(options)\n\t sql = []\n\t sql << flag_option_sql(options, :parallel)\n\t sql << flag_option_sql(options, :logging)\n\t sql << flag_option_sql(options, :monitoring)\n\t sql << \"TABLESPACE #{quote_identifier(options[:tablespace])}\" if options[:tablespace]\n\t sql << compress_option_sql(options)\n\t sql << options[:options] if String === options[:options]\n\t sql.compact.join ' '\n\t end",
"def dump_directory\n \"--out='#{ @dump_path }'\"\n end",
"def dump(database_name, dst)\n print \"Dummping #{database_name} ...\"\n command = @mysqldump\n command << \" -u #{@username}\"\n command << \" --skip-comments\"\n command << \" --skip-extended-insert\"\n command << \" --single-transaction\"\n command << \" --quick\"\n command << \" --password=#{@password}\" unless @password.blank?\n command << \" #{database_name}\"\n command << \" > #{dst}\"\n puts command\n output = `#{command}`\n if $?.success?\n puts 'Ok'\n else\n puts 'ERROR: Check your configuration file'\n end\n end",
"def psql_db_command__dump psql_db\n psql_db_command__program \"pg_dump\", psql_db\n end",
"def dump_directory\n \"--out='#{ dump_path }'\"\n end",
"def to_s\n str = \"\"\n\n str << \":#{@prefix} \" unless @prefix.empty?\n str << @command\n\n if @params\n f = false\n @params.each do |param|\n f = !f && (param.empty? || param[0] == ?: || param.include?(\" \"))\n str << \" \"\n str << \":\" if f\n str << param\n end\n end\n\n str << \"\\x0D\\x0A\"\n\n str\n end",
"def cmdline(command, options = {})\n command = [command].flatten\n cmdline = [where]\n cmdline += @options.merge(options).map { |k, v| !v ? nil : \"#{option(k)} #{v == true ? '' : v.to_s}\".rstrip }.compact\n cmdline << option(command.shift)\n cmdline += command\n cmdline.flatten.join(' ').strip\n end",
"def db_dumps__sample_example *args\n [\n \"install/10.0_to_10.1.sql\",\n \"/root/hack_10.0_to_10.1.sql\",\n ]\n\n end",
"def command_str_for(*args)\n args.collect{|arg|\n case arg\n when Array\n arg.join(' ')\n when Hash\n arg.collect{|k,v| \"-#{k} #{v}\"}.join(' ')\n else\n arg.to_s\n end\n }.join(' ')\n end",
"def to_s\n s = \"#@method \"\n s << \"-db '#@db' \" if @db\n s << \"-query #@qfile \" if @qfile\n s << @options.to_s if @options\n s\n end",
"def to_option\n \"--#{@name}\"\n end",
"def psql_db_batch__cli_or_generate_dumps psql_db, db_dump_paths=[nil], options=\"\"\n psql_db = array__from(psql_db)\n db_dump_paths = array__from(db_dump_paths)\n batch = db_dump_paths.map { |db_dump_path|\n cli = psql_db_command__cli psql_db\n pg_dump = psql_db_command__dump psql_db\n program = db_dump_path && pg_dump || cli\n [program, db_dump_path && \" #{options} -f #{quoted_shell_param db_dump_path}\"]\n }\n end",
"def mongo_dump_command\n cmd = \"\"\n cmd = \"mongodump -h #{config['host']}:#{config['port']} -d #{config['database']} -o . \"\n cmd += \"--username #{config['username']} --password #{config['password']}\" unless config[\"username\"].nil? and config[\"password\"].nil?\n cmd\n end",
"def copy_table_sql(table, opts)\n if table.is_a?(String)\n table\n else\n if opts[:options] || opts[:format]\n options = String.new\n options << \" (\"\n options << \"FORMAT #{opts[:format]}\" if opts[:format]\n options << \"#{', ' if opts[:format]}#{opts[:options]}\" if opts[:options]\n options << ')'\n end\n table = if table.is_a?(::Sequel::Dataset)\n \"(#{table.sql})\"\n else\n literal(table)\n end\n \"COPY #{table} TO STDOUT#{options}\"\n end\n end",
"def to_s\n if path.to_s.include?(Berkshelf.berkshelf_path.to_s)\n \"#{self.class.location_key}\"\n else\n \"#{self.class.location_key}: '#{path}'\"\n end\n end",
"def to_s\n @argv.join(' ')\n end",
"def dump_table(table)\n conditions = conditions_for(table)\n\n cmd = \"mysqldump #{ mysql_options } --tables #{ table }\"\n cmd += \" \\\"--where=#{ conditions }\\\"\" if conditions.present?\n\n if post_dump_command\n cmd += \"| #{post_dump_command}\"\n end\n\n cmd += \" > #{ output_dir }/#{ table }#{file_extension}\"\n\n system(cmd)\n end",
"def to_sql\n @applied.reverse.map{ |e| \"#{e.type}(#{e.args.join(\", \")})\" }.join('.')\n end",
"def to_s\n #N Without this we won't see the command as a command and a list of arguments\n return command.join(\" \")\n end",
"def to_s\n \"executed `#{@command.to_s.ellipsisize}'\"\n end",
"def command_string\n\t\treturn datastore['CMD'] || ''\n\tend",
"def to_sql\n source.select(bin_sql).to_sql\n end",
"def to_s\n\t\t\tout = \"\"\n\t\t\tout << banner << \"\\n\"\n\t\t\t\n\t\t\t@order.each {|option|\n\t\t\t\tif (option.nil?)\n\t\t\t\t\tout << \"\\n\"\n\t\t\t\t\tnext\n\t\t\t\tend\n\t\t\t\t\n line = sprintf(\"-%-2s --%-#{longest+6}s \",\n option.short_name,\n option.long_name + (option.flag ? \"\" : \" [VAL]\"))\n\n out << line\n if (line.length + option.description.length <= 80)\n out << option.description\n else\n rem = 80 - line.length\n desc = option.description\n i = 0\n while (i < desc.length)\n out << \"\\n\" if i > 0\n j = [i + rem, desc.length].min\n while desc[j..j] =~ /[\\w\\d]/\n j -= 1\n end\n chunk = desc[i..j].strip\n out << \" \" * line.length if i > 0\n out << chunk\n i = j + 1\n end\n end\n\n if (!option.flag)\n out << \" (default: #{option.default_value})\"\n end\n\n out << \"\\n\"\n }\n return out\n\t\tend",
"def dump(db = config[\"database\"], data_selection: {})\n file_namer = Postgres::FileNameGenerator.new(working_dir)\n exclude_tables_args = data_selection.keys.map { |table| %Q|--exclude-table-data=\"#{table}\"| }\n\n [\n cli.dump(file_namer.next(db, :schema), db, [exclude_tables_args]),\n *dump_partial_selected_data(db, file_namer, data_selection)\n ]\n end",
"def to_s\n return @string if @string\n result = @type + \" \" + to_sWithAliasesIndexes( @join_sources )\n result += @exporter.separator\n result += \"ON \" + @join_on.val.to_s if @join_on\n @string = result\n end",
"def build_cmd(sql)\n # Only exists within the context of this script (not exported), so this\n # does not degrade security posture after the script has completed\n ENV['PGPASSWORD'] = conn_opts[:password] if conn_opts.has_key? :password\n\n cmd = \"psql\"\n cmd << \" -d #{conn_opts[:database]}\" if conn_opts.has_key? :database\n cmd << \" -h #{conn_opts[:host]}\" if conn_opts.has_key? :host\n cmd << \" -U #{conn_opts[:username]}\" if conn_opts.has_key? :username\n cmd << \" -p #{conn_opts[:port]}\" if conn_opts.has_key? :port\n cmd << \" -c \\\"#{sql}\\\"\"\n\n return cmd\n end",
"def describe_prune(options)\n statement = ['Pruning']\n\n if options[:remote]\n statement << 'remote'\n elsif options[:local]\n statement << 'local'\n end\n\n statement << 'branches'\n\n if options[:age]\n statement << \"older than #{time_ago_in_words(options[:age])}\"\n end\n\n if options[:merged]\n statement << \"that are fully merged into #{options[:merged]}\"\n end\n\n if options[:only]\n naming = \"with a name containing '#{options[:only]}'\"\n if options[:except]\n naming << \" (but not '#{options[:except]}')\"\n end\n statement << naming\n end\n\n statement.join(' ')\n end",
"def where_clause\n \"\"\n end",
"def csv_dump(addr, root_dir)\n filename = \"full.csv\"\n dump_dir = File.join(root_dir)\n filepath = File.join(dump_dir, filename)\n\n cmd = \"psql -c '\\\\copy (SELECT * FROM outdoor_env WHERE (apogee_w_m2 IS NOT NULL AND address=#{addr}) ORDER BY db_time DESC) To #{filepath} With CSV header\\'\"\n run_command(cmd)\nend",
"def dump( out )\n out << \"foreign_key #{out_columns}, #{out_table_name}#{out_opts}\"\n end",
"def get_backup_file_name(backup_options = nil)\n backup_options ||= @backup_options\n return get_backup_file_name_with_generation(backup_options,0)\n end",
"def make_mysql_backup\n if @all_databases\n options = {\n :name => \"--all-databases\",\n :dump_options => \"\",\n :append_name => \"\"\n }\n file_name = mysqldump(options)\n compress_file(file_name)\n end\n if @databases && !@databases.empty?\n @databases.each do |db|\n options = {\n :name => db[:name].to_s,\n :dump_options => db[:dump_options].to_s,\n :append_name => db[:append_name].to_s\n }\n file_name = mysqldump(options)\n compress_file(file_name)\n end\n end\nend",
"def psql_db_batch__cli_or_apply_dumps psql_db, db_dump_paths=[nil], options=\"\"\n batch = db_dump_paths.map { |db_dump_path|\n cli = psql_db_command__cli psql_db\n [cli, db_dump_path && \" #{options} -f #{quoted_shell_param db_dump_path}\"]\n }\n end",
"def sql! sql=nil\n require 'niceql'\n puts Niceql::Prettifier.prettify_sql sql || $last_sql_command\n end",
"def syncer_name\n ARGV.detect { |arg| arg !~ /^--/ }\n end",
"def to_s\n return \"#{@column} #{@comparison.upcase}\" if nil_comparison?\n\n \"#{@column} #{@comparison} ?\"\n end",
"def to_s\n option.to_s\n end",
"def to_dump\n opts = {column: self.column}.merge options_for_dump\n dump = \"add_foreign_key #{from_table.inspect}, #{to_table.inspect}, #{opts.to_s.sub(/^{(.*)}$/, '\\1')}\"\n end",
"def to_s\n \"if \" + @condition.to_s + \" then\\n\" + indent(@then_part.to_s) + (else_part.skip? ? \"\" : (\"\\nelse\\n\" + indent(@else_part.to_s))) + \"\\nend\"\n end",
"def form_flagstring(f, fall)\n\t\tflagSelectAll = (!fall.nil? && fall.to_s.downcase == \"all\")\n\t\tif(flagSelectAll || f.nil? || f.empty?)\n\t\t\tflagStr = \"all\"\n\t\telse\n\t\t\tflagStr = f.join(\"|\")\n\t\tend\n\n\t\treturn flagStr\n\tend",
"def cmdline(keys = [:username, :password, :host, :database, :port], string = nil)\n @utility ||= search_path_for(\"mysql\", \"mysql5\")\n\n cmd = [ @utility ]\n self.config.only(*keys).each do |k, v|\n cmd << case k\n when :username then \"--user='#{v}'\"\n when :password then v.blank? ? \"\" : \"--password='#{v}'\"\n when :port then \"--port='#{v}'\" unless v.blank?\n when :host then \"--host='#{v}'\" unless\n v.blank? or v.among?(\"localhost\")\n end\n end\n\n cmd << string if string\n\n return cmd.compact * ' '\n end",
"def to_s\n \"<#{options.inspect}>\"\n end",
"def literal_string_append(sql, v)\n sql << \"'\" << db.synchronize(@opts[:server]){|c| c.escape(v)} << \"'\"\n end",
"def out_database_name\n if options[:out]\n if options[:out].respond_to?(:keys) && (db = options[:out][:db])\n db\n else\n database.name\n end\n end\n end",
"def nmap_options_string\n h = targets.join(' ')\n opt = scan_option.options.each_with_object([h.to_s]) do |(opt_key, opt_value), memo|\n if opt_key == 'top_ports' && opt_value.present?\n memo << \"--top-ports #{opt_value}\"\n elsif opt_key == 'ports' && opt_value.present? && ports.empty?\n memo << \"-p #{ScanJob.normalize_ports_as_string(opt_value)}\"\n elsif opt_value == '1' && ScanOption::NMAP_OPT_MAP[opt_key.to_sym]\n memo << ScanOption::NMAP_OPT_MAP[opt_key.to_sym]\n end\n end\n opt << \"-p #{ScanJob.normalize_ports_as_string(ports)}\" if ports.present?\n opt.join(' ')\n end",
"def options\n db = (opts[\"-d\"] or opts[\"d\"] or opts[\"--database\"] or opts[\"database\"] or \"mysql\")\n rubypath = (opts[\"-r\"] or opts[\"r\"] or opts[\"--ruby\"] or opts[\"ruby\"] )\n rubypath = \"--ruby=#{rubypath}\" if rubypath\n nullary_opts = %w{ -f --freeze --force -s --skip -q --quiet -c --svn -g --git }.reject { |x| !opts.has_key?(x) and !opts.has_key?(x.gsub(/-/, \"\"))}\n \"-d #{db} #{rubypath}\" + nullary_opts.join(\" \")\n end",
"def get_database_option(value)\n return value.nil? ? nil : \"Initial Catalog=#{value}\"\n end",
"def opts_inspect(opts)\n if opts[:default]\n opts = opts.dup\n de = Sequel.eval_inspect(opts.delete(:default)) \n \", :default=>#{de}#{\", #{opts.inspect[1...-1]}\" if opts.length > 0}\"\n else\n \", #{opts.inspect[1...-1]}\" if opts.length > 0\n end\n end",
"def to_sql\n [@sql_insert, @sql_on_conflict, @sql_returning].reject(&:nil?).join(' ')\n end",
"def format_sql_comment(comment)\n \" -- #{comment.to_s.gsub(/\\s+/, ' ')}\\n\"\n end",
"def generate_string(options = nil)\n if comment\n suffix = \" ##{comment}\\n\"\n else\n suffix = \"\\n\"\n end\n\n [address, name, *aliases].join(' ') << suffix\n end",
"def literal_string_append(sql, v)\n sql << \"'\" << db.synchronize(@opts[:server]){|c| c.escape_string(v)} << \"'\"\n end",
"def binary_path_for(version)\n \"/usr/pgsql-#{version}/bin\"\n end",
"def literal_string_append(sql, v)\n sql << \"'\" << ::Mysql.quote(v) << \"'\"\n end",
"def literal_string_append(sql, v)\n db.synchronize(@opts[:server]){|c| sql << c.quote(v)}\n end",
"def to_s\n\t\treturn \"%s/%s\" % [ self.branch.dn, self.filter_string ]\n\tend",
"def mongodump\n \"#{ utility(:mongodump) } #{ database } #{ credential_options } \" +\n \"#{ connectivity_options } #{ ipv6 } #{ additional_options } #{ dump_directory }\"\n end",
"def custom_sql(q)\n query = q + ';' unless q =~ /;$/\n query = @db_connection.query(\"#{query}\")\n query.each { |x| print_line(\"#{x.join(',')}\") } unless query.empty?\n end",
"def make_flag(options)\n\tflagString=\" \"\n\tif(options.list != nil)\n\t\tflagString+=\" -l\"\n\tend\n\tif(options.all != nil)\n\t\tflagString+= \" -a\"\n\tend\n\treturn flagString\nend",
"def binary_path_for(version)\n \"/opt/#{ChefUtils::Dist::Org::LEGACY_CONF_DIR}/embedded/postgresql/#{version}/bin\"\n end",
"def to_s\n \"Option :#{@option_name} is not a recognised option.\"\n end",
"def to_s\n require \"shellwords\"\n\n escaped_args = @args.map do |arg|\n escaped = Shellwords.escape(arg)\n next escaped if escaped == arg\n next escaped if arg.include?(\"'\")\n \"'#{arg}'\"\n end\n escaped_args.join(\" \")\n end",
"def sql_string\n self\n end",
"def to_s_append( ds, s )\n s << '$' << name.to_s\n s << ':' << sql_field.to_s if sql_field\n s << '/*' << small_source << '*/'\n end",
"def argument_string(args)\n args.inject([]) do |opts, opt|\n opts << (opt =~ /^-/ ? opt : opt.inspect)\n end.join(' ') \n end",
"def compress_option_sql(attrs)\n\t case value=attrs[:compress]\n\t when Fixnum, Integer then \"COMPRESS(#{value})\"\n\t else flag_option_sql attrs, :compress\n\t end\n end",
"def generate_upsert_options\n if options.empty?\n ''\n else\n ' USING ' <<\n options.map do |key, value|\n serialized_value =\n case key\n when :timestamp then (value.to_f * 1_000_000).to_i\n else value\n end\n \"#{key.to_s.upcase} #{serialized_value}\"\n end.join(' AND ')\n end\n end",
"def mongodump\n \"#{ mongodump_utility } #{ database } #{ credential_options } \" +\n \"#{ connectivity_options } #{ ipv6_option } #{ user_options } #{ dump_directory }\"\n end",
"def psql_db_dump_replacer__for_psql_db__sample_example\n [\n psql_db__sample_example,\n [\n \"/tmp/psql_db_original_dump\"\n ],\n [\n \"/tmp/database_dump\"\n ],\n \"ON_ERROR_STOP=off\",\n ]\n end",
"def dumpFilename\n DumpDir + '/' + @name + '-' + dateMark\n end",
"def string_format\n (command_environment.map do |key, value|\n [key, value].join(\"=\")\n end + [command]).join(\" \")\n end",
"def to_s\n regex = REPLACEMENT\n\n if Gitlab::Database.mysql?\n regex = Regexp.union(regex, MYSQL_REPLACEMENTS)\n end\n\n sql = @sql.gsub(regex, '?').gsub(CONSECUTIVE) do |match|\n \"#{match.count(',') + 1} values\"\n end\n\n # InfluxDB escapes double quotes upon output, so lets get rid of them\n # whenever we can.\n if Gitlab::Database.postgresql?\n sql = sql.delete('\"')\n end\n\n sql.tr(\"\\n\", ' ')\n end",
"def create_cron_command\n command = @options.cron + \" sycbackup #{@options.backup_folder}\"\n command += ' -d ' + @options.database + \n ' -u' + @options.user + \n ' -p' + @options.password if @options.database\n command += ' -f ' + @options.files.join(',') if @options.files\n command += ' -m ' + @options.max_backups if @options.max_backups\n command += ' --no-compress' if @options.no_compress\n command += ' --override' if @options.override\n\n command\n end",
"def ip_cmd_from_file\n\n ip_cmd = ''\n ip_cmd += 'to ' + self.to + ' ' if ! self.to.nil? && ! self.to.empty?\n ip_cmd += 'via ' + self.via + ' ' if ! self.via.nil? && ! self.via.empty?\n ip_cmd += 'dev ' + self.device + ' ' if ! self.device.nil? && ! self.device.empty?\n ip_cmd += 'table ' + self.table + ' ' if ! self.table.nil? && ! self.table.empty?\n return ip_cmd\n\n end",
"def argument_string(args)\n args.inject([]) do |opts, opt|\n opts << (opt =~ /^-/ ? opt : opt.inspect)\n end.join(' ')\n end",
"def command\n return \"#{@command} #{options}\"\n end",
"def sql_test(name, value, **opt)\n sql_name_value(name, value, **opt).join(' ')\n end",
"def literal_string_append(sql, s)\n sql << \"'\" << db.synchronize(@opts[:server]){|c| c.escape(s)} << \"'\"\n end",
"def bin_path(_opts)\n '/bin'\n end",
"def bin_path(_opts)\n '/bin'\n end",
"def bin_path(_opts)\n '/bin'\n end"
] |
[
"0.56896085",
"0.5640027",
"0.56199324",
"0.5397596",
"0.53662616",
"0.52767104",
"0.52732885",
"0.52212965",
"0.5166593",
"0.5133538",
"0.51143086",
"0.5015769",
"0.5003394",
"0.49865937",
"0.4960623",
"0.4915719",
"0.49095353",
"0.48870468",
"0.48632118",
"0.48536316",
"0.48382688",
"0.48194334",
"0.48122874",
"0.48118055",
"0.48088524",
"0.48042884",
"0.48036656",
"0.47931647",
"0.47930065",
"0.47927693",
"0.47809532",
"0.4780438",
"0.47786772",
"0.47760922",
"0.47618097",
"0.47486275",
"0.4742915",
"0.47302946",
"0.4715283",
"0.47146228",
"0.47060812",
"0.47003773",
"0.4690964",
"0.46718067",
"0.46536586",
"0.46494362",
"0.46466964",
"0.46447808",
"0.4638004",
"0.46370533",
"0.463543",
"0.4626261",
"0.46224958",
"0.4621881",
"0.4618993",
"0.46142656",
"0.4602272",
"0.46006843",
"0.45978498",
"0.45957002",
"0.4595526",
"0.4591729",
"0.4591474",
"0.45885283",
"0.45797133",
"0.4567443",
"0.45648628",
"0.45608613",
"0.45607924",
"0.45583656",
"0.45519975",
"0.4549597",
"0.45481884",
"0.45478874",
"0.45421398",
"0.45414397",
"0.4540731",
"0.4534779",
"0.45127246",
"0.45112658",
"0.45108566",
"0.45089504",
"0.44947836",
"0.44947562",
"0.44927266",
"0.44899237",
"0.44891033",
"0.44883606",
"0.4484101",
"0.4478792",
"0.44759023",
"0.44751558",
"0.44747418",
"0.44712362",
"0.4460904",
"0.44579092",
"0.4455904",
"0.44463742",
"0.44419816",
"0.44419816",
"0.44419816"
] |
0.0
|
-1
|
Any conditions that should be used to filter table based on type
|
def column_filter_conditions_for(table, type)
values = column_filters[type].join(', ')
foreign_key = type.singularize.foreign_key
if table == type
"\\`id\\` in (#{ values })"
elsif columns_for(table).include?(foreign_key)
"\\`#{ foreign_key }\\` in (#{ values })"
end
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def get_conditions(type)\n sql = ' WHERE ' + (T_DELETABLE.include?(type) ? 'deleted_at IS NULL AND ' : '')\n\n if T_LISTS.include? type\n check_api_token\n\n unless @current_user\n return 'ERROR'\n end\n\n id = @current_user.theatre_id\n\n if type == 'theatres' && id != 0\n sql += 'id = ' + id.to_s\n\n elsif type == 'posters'\n sql += 't.theatre_id = ' + id.to_s\n\n elsif id != 0\n sql += 'theatre_id = ' + id.to_s\n end\n\n elsif type == 'u_perms'\n sql += \"perm NOT LIKE 'theatres%' AND perm NOT LIKE '%approve' AND perm NOT LIKE '%choose'\"\n\n elsif type == 'performances'\n check_api_token\n\n unless @current_user\n return 'ERROR'\n end\n\n id = @current_user.theatre_id\n\n sql += 'approved = 0 OR approved = ' + id.to_s\n\n end\n\n # Return\n if T_DELETABLE.include? type\n sql\n else\n sql == ' WHERE ' ? '' : sql\n end\n end",
"def dataset_to_select_tables(*tables)\n if @restricted_to_types && !@tables_restricting_type.values_at(*tables).any?\n super.filter(@qualified_type_column => @restricted_to_types)\n else\n super\n end\n end",
"def conditions_for(table)\n conditions = []\n\n column_filters.keys.each do |type|\n conditions << column_filter_conditions_for(table, type)\n end\n\n # exclude all rows from table filtered tables\n conditions << '1 = 0' if table_filters.include?(table)\n\n conditions.compact.join(' or ')\n end",
"def filter_group query_params, *types\n types = [:select, :scope, :date_range, :search] if types.empty?\n query_params.filters.each do |field, config|\n type = config[:type]\n concat render_filter(query_params, field, type) if types.include?(type || :search)\n end\n end",
"def table_filter(schemaName, tblName, tblType)\n [\"information_schema\", \"pg_catalog\"].include?(schemaName) || tblType !~ /TABLE/i\n end",
"def query_type_to_clause_params()\n where_clause = nil\n sql_parameters = []\n qextra = false\n all_query = false\n\n type = self.type\n query = self.query\n if type == 'All'\n all_query = true\n elsif type == 'None'\n where_clause = '1=0'\n elsif type == 'Date'\n where_clause = \"(Begin_Date REGEXP ? OR EndDate REGEXP ?)\"\n sql_parameters << query\n sql_parameters << query\n elsif type == 'after'\n where_clause = \"Begin_Date > ?\"\n sql_parameters << query\n elsif type == 'before'\n where_clause = \"EndDate < ?\"\n sql_parameters << query\n elsif type == 'parameter'\n where_clause = \"`bottle_dbs`.`parameters` LIKE ?\"\n sql_parameters << \"%#{query}%\"\n qextra = 'params'\n elsif type == 'basin'\n where_clause = \"spatial_groups.`#{query}` = 1\"\n qextra = 'spatial_group'\n elsif type == 'Chief_Scientist'\n where_clause = \"contacts.LastName LIKE ?\"\n sql_parameters << \"%#{query}%\"\n qextra = 'contacts'\n else\n if @@cruise_columns.include?(type)\n where_clause = \"`cruises`.`#{type}` REGEXP ?\"\n sql_parameters << query\n else\n Rails.logger.debug(\"Search: unrecognized type #{type}\")\n where_clause = nil\n end\n end\n [where_clause, sql_parameters, qextra, all_query]\n end",
"def filtered_index(type)\n return unless (term = params.dig(:filter, :term).strip)\n\n case type\n when :clade\n clade_filter(term)\n when :region\n region_filter(term)\n # when :user\n # user_filter(term)\n end\n end",
"def filter(query_type, options = {})\n bool.filter(query_type, options)\n self\n end",
"def condition_names_for_column\n [condition_type_name]\n end",
"def type_condition(table = arel_table)\n if using_multi_table_inheritance?\n nil\n else\n sti_column = table[inheritance_column]\n sti_names = ([self] + descendants).map { |model| model.sti_name }\n\n sti_column.in(sti_names)\n end\n end",
"def type_condition(table = arel_table)\n super unless mti_table\n end",
"def add_type_filter(criteria)\n unless params[:type_uri].blank?\n @type_filter = params[:type_uri]\n @type = RdfType.find(@type_filter) rescue nil\n criteria.where(\"?uri a <#{@type_filter}>\")\n end\n criteria\n end",
"def filter_item_by_type(item, field, value)\n item[field] == value\nend",
"def filter(type)\n print_items @items.select { |item| item.type.casecmp(type).zero? }\n end",
"def conditions; end",
"def conditions; end",
"def conditions_for_collection\n [\"activities.type IS NULL \"]\n end",
"def filter type\n if type_allowed?(type.downcase) # Downcase the filter type and check if it's allowed (exists in lists)\n list_type = @@list_types[type.downcase.to_sym]\n @items.select { |item| item.is_a? list_type }\n # Set the title equal to the filter and output the\n # Header and items (as long as there are items to put)\n @title = \"Filtered by: \" + type.capitalize\n puts header\n puts output_for items if items\n else\n raise InvalidFilter, \"Invalid Filter: #{filter}\"\n end\n end",
"def table_filtered?(schema_name, table_type)\n %w[information_schema pg_catalog].include?(schema_name) || table_type !~ /TABLE/i\n end",
"def search_type\n case type\n when \"opinions\"\n query.only_amendables\n when \"amendments\"\n query.only_visible_emendations_for(@current_user, @component)\n else # Assume 'all'\n query.amendables_and_visible_emendations_for(@current_user, @component)\n end\n end",
"def filter_collection(type, collection)\n # Ignore filtering if `where` is not in frontmatter.\n return collection unless cfg.dig(type, 'where')\n cfg.dig(type, 'where').each do |attr, value|\n # Resolve attribute if it is referencing object(s).\n attr = (attr.split('.').size > 1) ? attr.split('.') : attr\n # Resolve value if it is meant to be dynamic.\n value = page.data[value[1..-1]] if value.start_with?(':')\n # Filter the collection.\n collection = collection.select do |obj|\n # If the attribute is referencing a string we know the value is\n # either a string or an array. If it is a string, match exactly, but\n # if it is an array, look for inclusion.\n if attr.is_a?(String)\n v = obj.data[attr]\n v.is_a?(Array) ? v.include?(value) : v == value\n # If the attribute is an object or array of objects, dig into the\n # object(s) to find the appropriate value to match or be included,\n # as specified above.\n else\n v = obj.data[attr.first]\n v.is_a?(Array) ? v.collect{|x| x.dig(*attr[1..-1])}.include?(value) : v.dig(*attr[1..-1]) == value\n end\n end\n end\n collection\n end",
"def _filter_by_type(collection, owner_type)\n negate_owner_type = owner_type[0] == '!'\n tracker_owner_type = negate_owner_type ? owner_type.slice(1..-1) : owner_type\n type_constant_names = _tracker_constant_names_for_type(tracker_owner_type)\n collection.where(\"trackers.type #{ negate_owner_type ? 'NOT' : '' } IN (?)\", type_constant_names)\n end",
"def rules_for_type(type)\n FILTER_VALUES.each do |rule|\n return rule if rule[:type] == type\n end\n nil\n end",
"def filter\n @filter_params.each do |key, val|\n # Strip empty values from the array if the given value is an array.\n val.select!{ |val| !val.to_s.empty? } if val.is_a?(Array)\n\n # Convert key if it starts with a column name.\n key = key.slice(@model_method_name.length + 1, key.length) if key.start_with?(\"#{@model_method_name}_\")\n\n if @force_filter_for && @force_filter_for.include?(key)\n ret = @args[:filter].call(:key => key, :val => val, :query => @query)\n @query = ret if ret\n elsif @model.column_names.include?(key)\n if val.is_a?(Array) && val.empty?\n # Ignore.\n else\n @query = @query.where(key => val)\n end\n elsif match = key.to_s.match(/^(.+)_like$/) and @model.column_names.include?(match[1])\n next if val.blank?\n table = @model.arel_table\n \n val.to_s.strip.split(/\\s+/).each do |str|\n @query = @query.where(table[match[1].to_sym].matches(\"%#{escape(str)}%\"))\n end\n elsif @args[:filter]\n ret = @args[:filter].call(:key => key, :val => val, :query => @query)\n @query = ret if ret\n else\n raise \"Dont know what to do regarding filter with key: '#{key}'.\"\n end\n end\n end",
"def filters(type)\n case type\n when :before\n namespaces = self.ancestors + [self]\n all_filters = namespaces.map(&:befores).flatten\n when :after\n namespaces = [self] + self.ancestors.reverse\n all_filters = namespaces.map(&:afters).flatten\n else\n raise 'Invalid filter type. Use :before or :after'\n end\n all_filters.select{|f| f[:opts][:all] || self.send(\"#{type}s\").include?(f) }\n end",
"def init_contract_filters\n\t \n\t statusable = Category.statusable\n\t if statusable.length == 0\n\t @conditions << \"(true = false)\"\n\t else\n\t @conditions << \"(category_id in (#{Category.statusable.collect{|c|c.id}.join(',')}))\"\n end\n\tend",
"def products_by_type(type)\n all_products.select { |product| filter_item_by_type(product, TARGET_PRODUCT_FIELD, type) }\nend",
"def filter(param, type:, operation:, column: nil, required: false, format_value: nil)\n query_param = param.to_s.underscore\n column ||= param.to_s.underscore\n op = operations.fetch(operation.to_sym) { raise ArgumentError, \"unknown operation #{operation}\" }\n sql = op.call(column, query_param, type)\n raw_filter(param, sql:, required:, format_value:, query_param:, array: op.array)\n end",
"def types_filter\n types_filter = self[:types_filter] || []\n return types_filter if types_filter.any?\n options.fetch 'types_filter', []\n end",
"def build_conditions_for_datatable\n criteria = search_for.inject([]) do |crit, atom|\n search = AjaxDatatablesRails::Datatable::SimpleSearch.new(value: atom, regex: datatable.search.regexp?)\n disjunction = searchable_columns.map do |simple_column|\n simple_column.search = search\n simple_column.search_query\n end.reduce(:or)\n bag_of_text_class = get_bag_of_text_class\n if bag_of_text_class\n bag_of_text_query = bag_of_text_class.arel_table[:bag_of_text].matches(\"%#{atom}%\")\n if disjunction\n disjunction = disjunction.or(bag_of_text_query)\n else\n disjunction = bag_of_text_query\n end\n end\n crit << disjunction\n end.compact.reduce(:and)\n criteria\n end",
"def filters(type=nil)\n unless(type)\n @filters.dup\n else\n const = Splib.find_const(type)\n type = const unless const.nil?\n @filters[type] ? @filters[type].dup : nil\n end\n end",
"def conditions_for(fields=[])\n predicate = []\n values = []\n fields.each do |field|\n predicate << \"lower(#{field.to_s}) like ?\"\n values << \"'%' + @search_key.downcase + '%'\"\n end\n eval(\"[\\\"#{predicate.join(' OR ')}\\\", #{values.join(',')}]\")\n end",
"def conditions\n [[]]\n end",
"def conditions\n return ['1=0'] if !valid?\n\n conds = []\n values = []\n for col in klass.columns do\n col_str = col_name(col)\n val = @attrs[encode_name(col)]\n if !val.blank?\n case col.type\n when :string, :text\n if val =~ /\\*$/\n conds << col_str + ' like ?'\n values << val.gsub(/\\*$/, '%')\n else\n conds << col_str + '=?'\n values << val\n end\n when :datetime\n case is_dbms?\n when Dbms::MYSQL\n# conds << col_str + @_operator.exp(encode_name(col))\n conds << sprintf(\"date(%s)\",\n col_str) +\n @_operator.exp(encode_name(col))\n else\n conds << sprintf(\"date(timezone('%s',%s))\",\n DEFAULT_TIMEZONE_FOR_TIMESTAMP_DATE_COMPARISON,\n col_str) +\n @_operator.exp(encode_name(col))\n end\n values << val.to_s\n when :boolean\n case is_dbms?\n when Dbms::MYSQL\n conds << col_str + @_operator.exp(encode_name(col))\n values << (val=='true')\n else\n conds << col_str + @_operator.exp(encode_name(col))\n values << val\n end\n else\n conds << col_str + @_operator.exp(encode_name(col))\n values << val\n end\n end\n end\n return [conds.join(' and ')] + values\n end",
"def filter(filters = {})\n all.select do |rule|\n [:country, :type].each do |condition|\n break false if filters[condition] && filters[condition] != rule[condition]\n true\n end\n end\n end",
"def filters\n # Call model column on self (metadata since element in array is a string, not a variable hence we use send) <=> self.send(profile) == true\n %w[sight_seeing_adventurer art_lover serial_shopper nature_lover food_addict sport_lover history_passionate tech_fan relaxed city_wanderer].select! {|profile| send(profile) == true }\n end",
"def type_condition(table = arel_table)\n sti_column = table[ sti_config[:foreign_key].to_sym ]\n sti_enum = sti_config[:enum_class]\n\n unless sti_enum.include?( self.name )\n raise KeyError, \"#{sti_enum.name} should include #{self.name} but doesn't.\"\n end\n\n sti_ids = descendants.inject([]) do |subclass, sid|\n if sti_enum.include? subclass.name\n sid << sti_enum[subclass.name].id\n else\n raise KeyError, \"#{sti_enum.name} should include #{subclass.name} but doesn't.\"\n end\n sid\n end\n sti_ids << sti_enum[self.name].id\n\n sti_column.in(sti_ids)\n end",
"def process_filters(records,filter_params)\n return records unless filter_params\n filter_params.each do |field,filter_param|\n if filter_param.has_key?(\"value\")\n value = filter_param[\"value\"]\n next unless value.present?\n condition = filter_param[\"condition\"] || 'eq'\n case condition\n when \"eq\"\n value = true if value == 'true'\n value = [false, nil] if value == 'false'\n records = records.where(field.to_sym => value)\n when \"cont\"\n records = records.where(\"#{field} LIKE '%#{value}%'\")\n when \"ncont\"\n records = records.where(\"#{field} NOT LIKE '%#{value}%'\")\n when \"gt\"\n records = records.where(\"#{field} > ?\", value)\n when \"lt\"\n records = records.where(\"#{field} < ?\", value)\n end\n end\n end\n return records\n end",
"def filters; end",
"def filters; end",
"def filter_types(input)\n input.reject { |k,v|\n ( skipped_types + [:notype, :other] ).include?(v[:type])\n }\n end",
"def schema_ds_filter(table_name, opts)\n if table_name\n [{:c__table_name=>table_name.to_s}]\n else\n [{:t__table_type=>'BASE TABLE'}]\n end\n end",
"def apply_narrowing_filters\n @filters[:narrowing].each do |filter|\n @query = @query.where(filter => @options[filter])\n end\n @query\n end",
"def apply_filter(data, filter_model)\n filter_type = filter_model.filter_type\n variable_name = filter_model.variable_name;\n value1 = filter_model.value1\n value2 = filter_model.value2\n\n case filter_type\n when 'equals'\n filter_hash = Hash.new\n filter_hash[variable_name] = value1\n data = data.where(filter_hash)\n\n when 'not_equals'\n filter_hash = Hash.new\n filter_hash[variable_name] = value1\n data = data.where.not(filter_hash)\n\n when 'greater_than'\n data = data.where(\"#{variable_name} > ?\", value1)\n\n when 'greater_than_or_equal'\n data = data.where(\"#{variable_name} >= ?\", value1)\n\n when 'less_than'\n data = data.where(\"#{variable_name} < ?\", value1)\n\n when 'less_than_or_equal'\n data = data.where(\"#{variable_name} <= ?\", value1)\n\n when 'from_to'\n data = data.where(\"#{variable_name} >= ? AND #{variable_name} <= ?\", value1, value2)\n\n else\n data\n end\n end",
"def filter\n filter_type = params[:filter][:type]\n case filter_type\n when \"last_seven\", \"weekly\"\n @filter = \"Weekly\"\n @filtered_runs = current_user.runs.in_the_last_week\n when \"last_thirty\", \"monthly\"\n @filter = \"Monthly\"\n @filtered_runs = current_user.runs.in_the_last_thirty_days\n when \"year\", \"yearly\"\n @filter = \"Yearly\"\n @filtered_runs = current_user.runs.in_the_last_year\n when \"lifetime\"\n @filter = \"Lifetime\"\n @filtered_runs = current_user.runs.most_recent_by_date\n end\n\n respond_to do |format|\n format.js\n end\n\n end",
"def get_matches(query_type, filter, select_cond)\r\n result_struct = get_result_struct(query_type, filter)\r\n match_array = KBResultSet.new(self, filter, filter.collect { |f|\r\n @field_types[@field_names.index(f)] })\r\n\r\n tbl_rec = @table_class.new(self)\r\n\r\n # Loop through table.\r\n @db.engine.get_recs(self).each do |rec|\r\n tbl_rec.populate(rec)\r\n\r\n next if select_cond and not select_cond.call(tbl_rec)\r\n\r\n match_array << create_result_rec(query_type, filter,\r\n result_struct, tbl_rec, rec)\r\n end\r\n return match_array\r\n end",
"def index\n\n if params[:search] and params[:type]\n @questions = Question.where(\"category LIKE ? and (title LIKE ? or quest LIKE ?)\",\"%#{params[:type]}%\",\"%#{params[:search]}%\",\"%#{params[:search]}%\")\n elsif params[:type]\n @questions = Question.where(\"category LIKE ?\",\"%#{params[:type]}%\")\n elsif params[:search]\n @questions = Question.where(\"title LIKE ? or quest LIKE ?\",\"%#{params[:search]}%\",\"%#{params[:search]}%\")\n else\n @questions = Question.all\n end\n\n\n end",
"def typus_fields_for(filter); end",
"def to_scope\r\n\r\n table_name = @model.table_name\r\n\r\n @filtering_attributes.select { |attr|\r\n @filtering_values.key?(attr)\r\n }.reduce(@model.scoped) { |model_scope, attr|\r\n\r\n filtering_value = @filtering_values[attr]\r\n\r\n filtering_column_type = @model.attribute_type(attr)\r\n\r\n column_sql = %'\"#{ table_name }\".\"#{ attr }\"'\r\n\r\n case filtering_column_type\r\n when :string\r\n case filtering_value\r\n when Set\r\n model_scope.where(\"#{ column_sql } IN (?)\", filtering_value)\r\n else\r\n model_scope.where(\"#{ column_sql } LIKE ?\", filtering_value)\r\n end\r\n\r\n when :boolean\r\n model_scope.where(\"#{ column_sql } = ?\", filtering_value)\r\n\r\n when :integer\r\n case filtering_value\r\n when Hash\r\n new_model_scope = model_scope\r\n if filtering_value.key?(:min)\r\n unless filtering_value[:min] == -Float::INFINITY\r\n new_model_scope =\r\n model_scope.where(\"#{ column_sql } >= ?\", filtering_value[:min])\r\n end\r\n end\r\n if filtering_value.key?(:max)\r\n unless filtering_value[:max] == Float::INFINITY\r\n new_model_scope =\r\n model_scope.where(\"#{ column_sql } <= ?\", filtering_value[:max])\r\n end\r\n end\r\n new_model_scope\r\n when Set\r\n model_scope.where(\"#{ column_sql } IN (?)\", filtering_value)\r\n when Range\r\n new_model_scope = model_scope\r\n unless filtering_value.first == -Float::INFINITY\r\n new_model_scope =\r\n model_scope.where(\"#{ column_sql } >= ?\", filtering_value.first)\r\n end\r\n unless filtering_value.last == Float::INFINITY\r\n new_model_scope =\r\n if filtering_value.exclude_end?\r\n model_scope.where(\"#{ column_sql } < ?\", filtering_value.last)\r\n else\r\n model_scope.where(\"#{ column_sql } <= ?\", filtering_value.last)\r\n end\r\n end\r\n new_model_scope\r\n else\r\n model_scope.where(\"#{ column_sql } = ?\", filtering_value)\r\n end\r\n\r\n when :date\r\n case filtering_value\r\n when Hash\r\n new_model_scope = model_scope\r\n if filtering_value.key?(:from)\r\n new_model_scope =\r\n model_scope.where(\"#{ column_sql } >= ?\", filtering_value[:from])\r\n end\r\n if filtering_value.key?(:until)\r\n new_model_scope =\r\n model_scope.where(\"#{ column_sql } <= ?\", filtering_value[:until])\r\n end\r\n new_model_scope\r\n when Set\r\n model_scope.where(\"#{ column_sql } IN (?)\", filtering_value)\r\n when Range\r\n new_model_scope = model_scope\r\n unless filtering_value.first == -Float::INFINITY\r\n new_model_scope =\r\n model_scope.where(\"#{ column_sql } >= ?\", filtering_value.first)\r\n end\r\n unless filtering_value.last == Float::INFINITY\r\n new_model_scope =\r\n if filtering_value.exclude_end?\r\n model_scope.where(\"#{ column_sql } < ?\", filtering_value.last)\r\n else\r\n model_scope.where(\"#{ column_sql } <= ?\", filtering_value.last)\r\n end\r\n end\r\n new_model_scope\r\n else\r\n model_scope.where(\"#{ column_sql } = ?\", filtering_value)\r\n end\r\n else\r\n model_scope\r\n end\r\n }\r\n end",
"def add_conditions!(sql, conditions)\n sql << \"WHERE #{sanitize_conditions(conditions)} \" unless conditions.nil?\n sql << (conditions.nil? ? \"WHERE \" : \" AND \") + type_condition unless descends_from_active_record?\n end",
"def filtered_by_association_type(name, type)\n type.present? ? where(\"#{name}_key.begins_with\" => \"#{type}#{ActivityNotification.config.composite_key_delimiter}\") : none\n end",
"def typus_filters\n\n fields_with_type = ActiveSupport::OrderedHash.new\n\n if self.respond_to?('admin_filters')\n fields = self.admin_filters\n else\n return [] unless Typus::Configuration.config[self.name]['filters']\n fields = Typus::Configuration.config[self.name]['filters'].split(', ').collect { |i| i.to_sym }\n end\n\n fields.each do |field|\n attribute_type = self.model_fields[field.to_sym]\n if self.reflect_on_association(field.to_sym)\n attribute_type = self.reflect_on_association(field.to_sym).macro\n end\n fields_with_type[field.to_s] = attribute_type\n end\n\n return fields_with_type\n\n end",
"def filter\n if params[:modelsearch]\n @results = Carmodel.where('title LIKE ?', \"%#{params[:modelsearch]}%\")\n @model_type = 'modeltype'\n end\n if params[:subtypesearch]\n @results = Carmodelsubtype.where('title LIKE ?', \"%#{params[:subtypesearch]}%\")\n @model_type = 'subtype'\n end\n respond_to do |format|\n format.js\n end\n end",
"def where(*cond)\n return super if block_given?\n\n filters_per_role = self.cache_get(:filters_per_role) || init_filters([])\n\n Array.wrap(cond).each do |condition|\n filters = case condition\n when Hash\n hash_condition_to_filters(condition)\n when Sequel::SQL::BooleanExpression\n boolean_expression_to_filters(condition)\n else\n raise \"Unsupported condition type: #{condition.class}\"\n end\n\n append_filters(filters_per_role, filters)\n end\n\n dataset = self.from(RoleDataset.build(filters_per_role))\n dataset.cache_set(:filters_per_role, filters_per_role)\n dataset\n end",
"def extra_search_actions(items, extra_filters = [], kind = nil)\n (extra_filters || []).each do |filter|\n case filter\n when 'my_country'\n case kind || params[:type]\n when 'people', 'counselors'\n items = items.where(country: current_user.country)\n when 'churches', 'groups'\n items = items.joins(:user).where(users:{ country: current_user.country })\n when 'contents'\n items = items.joins(:user).where(users:{ country: current_user.country })\n when 'events'\n items = items.joins('inner join user_groups on user_groups.id = events.eventable_id and events.eventable_type = \\'UserGroup\\' inner join users on users.id = user_groups.user_id').where('users.country = ?', current_user.country)\n \n # TODO\n end\n when 'my_groups'\n case kind || params[:type]\n when 'people', 'counselors'\n items = items.joins(:user_groups).where(user_groups: {id: current_user.user_groups.pluck(:id)})\n when 'churches', 'groups'\n items = items.where(id: current_user.user_groups.select(:id))\n when 'contents'\n items = items.where(user_id: current_user.user_groups_members.select(:id))\n when 'events'\n items = items.where(id: current_user.user_groups_events.select(:id))\n end\n end\n end\n items\n end",
"def filters\n end",
"def conditions_from_params\n @conditions_from_params ||= begin\n conditions = [{}]\n params.except(:controller, :action, :page, :sort, :sort_direction, :format, :id).each do |key, value|\n distinct = true if key.match?(/!$/)\n column = active_scaffold_config._columns_hash[key.to_s[0..(distinct ? -2 : -1)]]\n next unless column\n key = column.name.to_sym\n not_string = %i[string text].exclude?(column.type)\n next if active_scaffold_constraints[key]\n next if nested? && nested.param_name == key\n\n range = %i[date datetime integer decimal float bigint].include?(column.type) && value.is_a?(String) && value.scan('..').size == 1\n value = value.split('..') if range\n value =\n if value.is_a?(Array)\n value.map { |v| v == '' && not_string ? nil : ActiveScaffold::Core.column_type_cast(v, column) }\n elsif value == '' && (not_string || column.null)\n ActiveScaffold::Core.column_type_cast(column.default, column)\n else\n ActiveScaffold::Core.column_type_cast(value, column)\n end\n value = Range.new(*value) if range\n if distinct\n conditions << active_scaffold_config.model.arel_table[key].not_eq(value)\n else\n conditions[0][key] = value\n end\n end\n conditions\n end\n end",
"def filter(item_type)\n case item_type\n when 'todo' then item_type = TodoItem\n when 'event' then item_type = EventItem\n when 'link' then item_type = LinkItem\n else raise UdaciListErrors::InvalidFilterType,\n \"#{type_desc} is not a valid filter type\"\n end\n output(item_type)\n end",
"def check_filter_options() #:nodoc:\r\n table_name = @tables.first[1]\r\n model = @tables.first[0]\r\n session[table_name] ||= {}\r\n# process page\r\n session[table_name][:page] = params[:page] if params[:page]\r\n# new filter is applied\r\n if params[:filter]\r\n set_session_filter(table_name)\r\n session[table_name][:page] = 1\r\n end\r\n# if data model has field dc_site_id ensure that only documents which belong to the site are selected.\r\n site_id = dc_get_site._id if dc_get_site\r\n# dont't filter site if no dc_site_id field or user is ADMIN\r\n site_id = nil if !model.method_defined?('dc_site_id') or dc_user_can(DcPermission::CAN_ADMIN)\r\n# \r\n if @records = DcFilter.get_filter(session[table_name][:filter])\r\n @records = @records.and(dc_site_id: site_id) if site_id\r\n else\r\n @records = if site_id\r\n model.where(dc_site_id: site_id)\r\n else\r\n model\r\n end\r\n end\r\n=begin \r\n# TODO Use only fields requested. Higly experimental but necessary in some scenarios\r\n if (columns = @form['result_set']['columns'])\r\n cols = []\r\n columns.each { |k,v| cols << v['name'] }\r\n p '*',cols,'*'\r\n @records = @records.only(cols)\r\n end\r\n=end \r\n# pagination if required\r\n per_page = (@form['result_set']['per_page'] || 30).to_i\r\n if per_page > 0\r\n @records = @records.page(session[table_name][:page]).per(per_page)\r\n end\r\nend",
"def filtered_by_association_type_and_id(name, type, id)\n type.present? && id.present? ? where(\"#{name}_key\" => \"#{type}#{ActivityNotification.config.composite_key_delimiter}#{id}\") : none\n end",
"def filter_available?\r\n # -1 = unused, 2 = Filter used\r\n self.query.ole_object.ortype == -1 ||\r\n self.query.ole_object.ortype == 2\r\n end",
"def applyable(table_class)\n for condition in conditions\n\treturn false unless condition.applyable(table_class)\n end\n return true\n end",
"def validate_query_filters\n super\n\n or_filters.each_key do |field|\n if values_for(field)\n case type_for(field)\n when :integer\n add_filter_error(field, :invalid, true) if or_values_for(field).detect {|v| v.present? && !v.match(/\\A[+-]?\\d+(,[+-]?\\d+)*\\z/) }\n when :float\n add_filter_error(field, :invalid, true) if or_values_for(field).detect {|v| v.present? && !v.match(/\\A[+-]?\\d+(\\.\\d*)?\\z/) }\n when :date, :date_past\n case or_operator_for(field)\n when \"=\", \">=\", \"<=\", \"><\"\n add_filter_error(field, :invalid, true) if or_values_for(field).detect {|v|\n v.present? && (!v.match(/\\A\\d{4}-\\d{2}-\\d{2}(T\\d{2}((:)?\\d{2}){0,2}(Z|\\d{2}:?\\d{2})?)?\\z/) || parse_date(v).nil?)\n }\n when \">t-\", \"<t-\", \"t-\", \">t+\", \"<t+\", \"t+\", \"><t+\", \"><t-\"\n add_filter_error(field, :invalid, true) if or_values_for(field).detect {|v| v.present? && !v.match(/^\\d+$/) }\n end\n end\n end\n\n add_filter_error(field, :blank, true) unless\n # filter requires one or more values\n (or_values_for(field) and !or_values_for(field).first.blank?) or\n # filter doesn't require any value\n [\"o\", \"c\", \"!*\", \"*\", \"t\", \"ld\", \"w\", \"lw\", \"l2w\", \"m\", \"lm\", \"y\", \"*o\", \"!o\"].include? or_operator_for(field)\n end if or_filters && or_filters.respond_to?(:each_key)\n end",
"def resource_type_filter\n return filter_for(:resource_type_id,\n objects_to_names_and_ids(current_user.company.resource_types),\n session[:resource_filters], ResourceType.model_name.human)\n end",
"def filter_listings(start_datetime, end_datetime, sale_lease, type)\n # Default to nil (which resolves to \"Sale\" if empty string provided\n filter_sale_lease = sale_lease == \"\" ? nil : sale_lease\n \n # Filter listings by required fields\n filter = Listing.where(municipality_id: params[:id])\n .where(sale_lease: filter_sale_lease || \"Sale\")\n .where(\n \"(sold_date > ? OR sold_date IS NULL) AND list_date < ?\",\n start_datetime,\n end_datetime\n )\n\n # Further filter by type, if provided\n filter = filter.where(listing_type: type) if type\n\n return filter\n end",
"def auth_records_condition(table_name = \"\")\n ret = \"\"\n table_pre = table_name.blank? ? '' : table_name + '.'\n class_name = params[:controller].titleize.sub(' ','').singularize.constantize\n if class_name.column_names.include?(\"employe_id\")\n ret += \" \" + table_pre + \"employe_id in (\" + @current_user.employe.id.to_s + \") \"\n end\n if class_name.column_names.include?(\"department_id\")\n str_managed_depart_id = auth_user_managed_departments(@current_user.id).join(',')\n ret += \" or \" + table_pre + \"department_id in (\" + str_managed_depart_id + \")\" if not str_managed_depart_id.blank?\n end\n ret = \"1=1\" if ret.blank?\n\n return \" (\" + ret + \") \"\n end",
"def do_params_filter scope\n filter_params.each do |k,value|\n if value.present?\n if self.class.custom_filter_fields[k].present?\n scope = self.class.custom_filter_fields[k].call scope, value\n elsif resource_class.column_names.include? k\n if resource_class.columns_hash[k].type == :boolean\n if value == '0'\n puts \"Should filter\"\n scope = scope.where(k => [false,nil])\n else\n scope = scope.where(k => true)\n end\n else\n scope = scope.where(k => value)\n end\n elsif resource_class.reflect_on_association(k.to_sym).present?\n klass = resource_class.reflect_on_association(k.to_sym).klass\n scope = do_inner_params_filter klass, value, scope\n else\n Rails.logger.warn(\"No filter is available for field #{k}\")\n end\n end\n end\n scope\n end",
"def time_filter(selection)\n\t\t\tif self.method_defined? :gc_created_at\n\t\t\t\ttime_column = 'gc_created_at'\n\t\t\telsif self.method_defined? :source_created_at\n\t\t\t\ttime_column = 'source_created_at'\n\t\t\telse\n\t\t\t\ttime_column = 'created_at'\n\t\t\tend\n\t\t\t\n\t\t\tif selection.is_a?(Hash)\n\t\t\t\tif selection[:filter] == 'between'\n\t\t\t\t\tself.where(name_of_class + '.' + time_column + '>= ?', Time.parse(selection[:from]))\n\t\t\t\t\t\t.where(name_of_class + '.' + time_column + '<= ?', Time.parse(selection[:to]))\n\t\t\t\telse\n\t\t\t\t\tself.all\n\t\t\t\tend\n\t\t\telse\n\t\t\t\tcase selection\n\t\t\t\twhen 'any'\n\t\t\t\t\tself.all\n\t\t\t\twhen 'today'\n\t\t\t\t\tself.where(name_of_class + '.' + time_column + '>= ?', Date.today)\n\t\t\t\twhen 'yesterday'\n\t\t\t\t\tself.where(name_of_class + '.' + time_column + '>= ?', Date.today - 1)\n\t\t\t\twhen 'this_week'\n\t\t\t\t\tself.where(name_of_class + '.' + time_column + '>= ?', Time.now.beginning_of_week)\n\t\t\t\twhen 'this_month'\n\t\t\t\t\tself.where(name_of_class + '.' + time_column + '>= ?', Time.now.beginning_of_month)\n\t\t\t\twhen 'last_month'\n\t\t\t\t\tself.where(name_of_class + '.' + time_column + '>= ?', Time.now.last_month.beginning_of_month)\n\t\t\t\t\t\t.where(name_of_class + '.' + time_column + '<= ?', Time.now.last_month.end_of_month)\n\t\t\t\twhen 'last_7_days'\n\t\t\t\t\tself.where(name_of_class + '.' + time_column + '>= ?', Date.today - 7)\n\t\t\t\twhen 'last_30_days'\n\t\t\t\t\tself.where(name_of_class + '.' + time_column + '>= ?', Date.today - 30)\n\t\t\t\telse\n\t\t\t\t\tself.all\n\t\t\t\tend\n\t\t\tend\n\t\tend",
"def filtertype_op\n\t\t\tFILTERTYPE_OP[ self.filtertype.to_sym ]\n\t\tend",
"def parse_where(remove_table_names=false)\n filters = {}\n done = false\n while !done && @tokenizer.more?\n name = @tokenizer.next_token\n raise \"sql parser can't handle nested stuff yet: #{@tokenizer.sql}\" if name == '('\n name.sub!(/.*\\./, '') if remove_table_names # Remove \"schema.table.\" from \"schema.table.col\"\n\n op = @tokenizer.next_token\n op += (' ' + @tokenizer.next_token) if op.downcase == 'not'\n op = op.downcase\n\n val = @tokenizer.next_token\n\n case op\n when \"=\"\n filters[name] = val\n when \"<\"\n filters[name] = { :$lt => val }\n when \"<=\"\n filters[name] = { :$lte => val }\n when \">\"\n filters[name] = { :$gt => val }\n when \">=\"\n filters[name] = { :$gte => val }\n when \"<>\", \"!=\"\n filters[name] = { :$ne => val }\n when \"like\"\n filters[name] = regexp_from_string(val)\n when \"in\"\n raise \"'in' must be followed by a list of values: #{@tokenizer.sql}\" unless val == '('\n filters[name] = { :$in => read_array }\n when \"between\"\n conjunction = @tokenizer.next_token.downcase\n raise \"syntax error: expected 'between X and Y', but saw '\" + conjunction + \"' instead of 'and'\" unless conjunction == 'and'\n val2 = @tokenizer.next_token\n val2, val = val, val2 if val > val2 # Make sure val <= val2\n filters[name] = { :$gte => val, :$lte => val2 }\n else\n raise \"can't handle sql operator [#{op}] yet: #{@tokenizer.sql}\"\n end\n\n break unless @tokenizer.more?\n\n tok = @tokenizer.next_token.downcase\n case tok\n when 'and'\n next\n when 'or'\n raise \"sql parser can't handle ors yet: #{@tokenizer.sql}\"\n when 'order', 'group', 'limit'\n @tokenizer.add_extra_token(tok)\n done = true\n else\n raise \"can't handle [#{tok}] yet\"\n end\n end\n filters\n end",
"def prepare_orm_filters\n filters = [[]]\n date_format = I18n.t(:\"date.formats.default\", {:locale => I18n.locale })\n self.data_grid.columns.each_with_index do |col, col_index|\n if col.filter and !col.filter_value.blank?\n case col.filter\n when :boolean\n filters[0] << \"#{col.filter_by} = ?\"\n filters << (col.filter_value == '1') ? true : false\n when :auto\n filters[0] << \"#{col.filter_by} = ?\"\n filters << col.filter_value\n when :text\n filters[0] << \"#{col.filter_by} #{ActiveRecord::Base.connection.adapter_name.downcase.to_sym == :postgresql ? 'ILIKE' : 'LIKE'} ?\"\n filters << \"%#{col.filter_value}%\"\n when :number\n filters[0] << \"#{col.filter_by} = ?\"\n filters << col.filter_value.to_i\n when :range\n range = col.filter_value.split(DataGrid.range_separator)\n\n if !range[0].blank? and !range[1].blank?\n begin\n range[0] < 2\n rescue\n range[0] = range[0].to_f\n range[1] = range[1].to_f\n end\n filters[0] << \"#{col.filter_by} >= ? AND #{col.filter_by} <= ?\"\n filters << range[0]\n filters << range[1]\n elsif range[0].blank? and !range[1].blank?\n begin\n range[1] < 2\n rescue\n range[1] = range[1].to_f\n end\n filters[0] << \"#{col.filter_by} <= ?\"\n filters << range[1]\n elsif range[1].blank? and !range[0].blank?\n begin\n range[0] < 2\n rescue\n range[0] = range[0].to_f\n end\n filters[0] << \"#{col.filter_by} >= ?\"\n filters << range[0]\n end\n\n when :date\n range = col.filter_value.split(DataGrid.range_separator)\n\n if !range[0].blank? and !range[1].blank?\n begin\n range[0] < 2\n rescue\n range[0] = DateTime.strptime(range[0], date_format)\n range[1] = DateTime.strptime(range[1], date_format)\n end\n filters[0] << \"#{col.filter_by} >= ? AND #{col.filter_by} <= ?\"\n filters << range[0]\n filters << range[1]\n elsif range[0].blank? and !range[1].blank?\n begin\n range[1] < 2\n rescue\n range[1] = DateTime.strptime(range[1], date_format)\n end\n filters[0] << \"#{col.filter_by} <= ?\"\n filters << range[1]\n elsif range[1].blank? and !range[0].blank?\n begin\n range[0] < 2\n rescue\n range[0] = DateTime.strptime(range[0], date_format)\n end\n filters[0] << \"#{col.filter_by} >= ?\"\n filters << range[0]\n end\n end\n end\n end\n \n filters[0] = filters[0].join(' AND ')\n filters\n end",
"def filters_sql\n if filters.blank?\n ''\n else\n conditions = JSON.parse filters\n sql_array = []\n conditions.each do |condition|\n if condition['values']\n values = condition['values'].map { |x| \"'#{x}'\" }.join(',')\n sql_array << \" #{condition['name']} in (#{values})\"\n else\n sql_array << \" #{condition['name']} between #{condition['from']} and #{condition['to']}\"\n end\n end\n sql_array.join(' AND ')\n end\n end",
"def relation_filter_for_relation_type(relation_type)\n self.class.relation_filter_for_relation_type(relation_type)\n end",
"def throttle_filter_class\n {:_type.nin => [self.class.name]}\n end",
"def get_form_filter_types\n [\n [\"From..To\", \"from_to\"],\n [\"Equals\", \"equals\"],\n [\"Not Equal To\", \"not_equals\"],\n [\"Is Greater Than\", \"greater_than\"],\n [\"Is Greater Than Or Equal To\", \"greater_than_or_equal\"],\n [\"Is Less Than\", \"less_than\"],\n [\"Is Less Than Or Equal To\", \"less_than_or_equal\"]\n\n ]\n end",
"def conditions_for_collection\n selector = case session[:travel_filter]\n when 'arrivals' then ['travels.arrival AND travels.date >= (?)', Date::today]\n when 'departures' then ['travels.arrival = ? AND travels.date >= (?)', false, Date::today]\n when 'all_dates' then TRUE\n when 'current' then ['travels.date >= (?)', Date::today]\n else ['travels.date >= (?)', Date::today] # This will include nil case where filter has not been set for session\n end\n return selector\n end",
"def sql(options = {}) \n table = options[:table_class]\t\n if table.respond_to?(:do_not_use_view_for_query) and table.do_not_use_view_for_query\n\tconditions.map {|c| c.sql(options) }\n else\n\tconditions.map {|c| c.sql(options[:table_class]) }\n end\n end",
"def resource_type_filter\n return filter_for(:resource_type_id, \n objects_to_names_and_ids(current_user.company.resource_types),\n session[:resource_filters], _(\"Resource Type\"))\n end",
"def build_conditions(params)\n\n conditions, joins = merge_conditions, []\n\n query_params = params.dup\n %w(action controller).each { |param| query_params.delete(param) }\n\n # Remove from params those with empty string.\n query_params.delete_if { |k, v| v.empty? }\n\n # If a search is performed.\n if query_params[:search]\n query = ActiveRecord::Base.connection.quote_string(query_params[:search].downcase)\n search = []\n typus_search_fields.each do |key, value|\n _query = case value\n when \"=\" then query\n when \"^\" then \"#{query}%\"\n when \"@\" then \"%#{query}%\"\n end\n search << \"#{key} LIKE '#{_query}'\"\n end\n conditions = merge_conditions(conditions, search.join(\" OR \"))\n end\n\n query_params.each do |key, value|\n\n filter_type = model_fields[key.to_sym] || model_relationships[key.to_sym]\n\n case filter_type\n when :boolean\n condition = { key => (value == 'true') ? true : false }\n conditions = merge_conditions(conditions, condition)\n when :datetime\n interval = case value\n when 'today' then Time.new.midnight..Time.new.midnight.tomorrow\n when 'last_few_days' then 3.days.ago.midnight..Time.new.midnight.tomorrow\n when 'last_7_days' then 6.days.ago.midnight..Time.new.midnight.tomorrow\n when 'last_30_days' then Time.new.midnight.prev_month..Time.new.midnight.tomorrow\n end\n condition = [\"#{key} BETWEEN ? AND ?\", interval.first.to_s(:db), interval.last.to_s(:db)]\n conditions = merge_conditions(conditions, condition)\n when :date\n if value.is_a?(Hash)\n date_format = Date::DATE_FORMATS[typus_date_format(key)]\n\n begin\n unless value[\"from\"].blank?\n date_from = Date.strptime(value[\"from\"], date_format)\n conditions = merge_conditions(conditions, [\"#{key} >= ?\", date_from])\n end\n\n unless value[\"to\"].blank?\n date_to = Date.strptime(value[\"to\"], date_format)\n conditions = merge_conditions(conditions, [\"#{key} <= ?\", date_to])\n end\n rescue\n end\n else\n # TODO: Improve and test filters.\n interval = case value\n when 'today' then nil\n when 'last_few_days' then 3.days.ago.to_date..Date.tomorrow\n when 'last_7_days' then 6.days.ago.midnight..Date.tomorrow\n when 'last_30_days' then (Date.today << 1)..Date.tomorrow\n end\n if interval\n condition = [\"#{key} BETWEEN ? AND ?\", interval.first, interval.last]\n elsif value == 'today'\n condition = [\"#{key} = ?\", Date.today]\n end\n conditions = merge_conditions(conditions, condition)\n end\n when :integer, :string\n condition = { key => value }\n conditions = merge_conditions(conditions, condition)\n when :has_and_belongs_to_many\n condition = { key => { :id => value } }\n conditions = merge_conditions(conditions, condition)\n joins << key.to_sym\n end\n\n end\n\n return conditions, joins\n\n end",
"def select_filter_drop_downs search_filter , value\n case search_filter\n when 'filter_type'\n select_from_drop_down 's:swfield', value\n when 'filter_category'\n select_from_drop_down 'rc:swfield', value\n when 'filter_sub_category'\n select_from_drop_down 'c1:swfield', value\n when 'bathroom minimum'\n select_from_drop_down 'bathrooms__gte:swfield', value\n when 'item sorting'\n select_from_drop_down 'id_sort_by', value\n end\n end",
"def by_type\n filter(:type) do |post, type|\n post.post_type.casecmp(type).zero?\n end\n end",
"def get_filters_sql\n if self.filters.blank?\n return ''\n else\n conditions = JSON.parse self.filters\n sql_array = []\n conditions.each do |condition|\n if condition['values']\n values = condition['values'].map{|x| \"'#{x}'\"}.join(',')\n sql_array << \" #{condition['name']} in (#{values})\"\n else\n sql_array << \" #{condition['name']} between #{condition['from']} and #{condition['to']}\"\n end\n end\n sql_array.join(' AND ')\n end\n end",
"def test_ut_fdw10a_t5_fic_009\n\t\tfilter_condition = FilterCondition.new(\"Action Status\", \"is\", 1)\n\t\tstring_return = filter_condition.create_query\n\t\tassert_equal \"risk_types.id = 1\", string_return\n end",
"def for_favoritable_type favoritable_type\n where favoritable_type: favoritable_type\n end",
"def filtered_by_type(notifiable_type)\n filtered_by_association_type(\"notifiable\", notifiable_type)\n end",
"def index\n @article_type_filter = (params[:article_type_filter] || ArticleType.all.each.collect{|article_type| article_type.id}).collect{|filter| filter.to_i}\n @articles = Article.where(article_type_id: @article_type_filter)\n end",
"def cast_records!(*types, **options)\n where!(regclass.cast(:varchar).in(types.map(&:table_name))) if options[:filter]\n self.select_extra_values += [regclass.as(_record_class_attribute.to_s)]\n self.cast_records_value = (types.present? ? types : model.casted_dependents.values)\n self\n end",
"def filter_options options\n other_tables = options[:other_tables].dup || []\n temp_options = options.dup\n temp_options.keep_if do |k,v|\n ['name','page','search','order'].include? k.to_s\n end\n other_tables << temp_options\n end",
"def index\n # if filter is blank, set up default values.\n # at last, we need to covert array contents all to integer.\n @filter = (params[:filter] || ProductType.all.collect{|type| type.id}).collect{|filter| filter.to_i}\n @products = Product.where(product_type_id: @filter)\n end",
"def select(table, type, where = {}, and_or = \"and\", equals_like = \"=\", order_by = \"\")\n order = order_by.empty? ? '' : \" ORDER BY #{order_by}\"\n sql = \"select * from #{table} #{where_clause(where, and_or, equals_like)}#{order};\"\n execute(sql).collect { |r| type.new.from_sql(r) }\n end",
"def filter(filtering_params)\n\t\t\tresults = self.where(nil)\n\t\t\tfiltering_params.each do |key, value|\n\t\t\t\tresults = results.public_send(key, value) if value.present?\n\t\t\tend\n\t\t\tresults\n\t\tend",
"def get_by_type(type)\r\n facts2typerule = {:methodcall => [:cant_access, :can_access],\r\n :inheritance => [:cant_extend, :can_extend],\r\n :objcreation => [:cant_create, :can_create]}\r\n\r\n find_all { |x| facts2typerule[type].index(x.type_interaction) != nil }\r\n end",
"def check_filter_options #:nodoc:\r\n table_name = CmsHelper.table_param(params).strip.split(';').first.underscore\r\n model = table_name.classify.constantize\r\n session[table_name] ||= {}\r\n # page is set\r\n session[table_name][:page] = params[:page] if params[:page]\r\n # if data model has field dc_site_id ensure that only documents which belong to the site are selected.\r\n site_id = dc_get_site._id if dc_get_site\r\n\r\n # don't filter site if no dc_site_id field or user is ADMIN\r\n site_id = nil if !model.method_defined?('dc_site_id') || dc_user_can(DcPermission::CAN_ADMIN)\r\n site_id = nil if session[table_name][:filter].to_s.match('dc_site_id')\r\n\r\n if @records = DcFilter.get_filter(session[table_name][:filter])\r\n @records = @records.and(dc_site_id: site_id) if site_id\r\n else\r\n @records = site_id ? model.where(dc_site_id: site_id) : model\r\n end\r\n process_select_and_deny_fields\r\n # pagination if required\r\n per_page = (@form['result_set']['per_page'] || 25).to_i\r\n @records = @records.page(session[table_name][:page]).per(per_page) if per_page > 0\r\nend",
"def filter\n end",
"def set_filters\n @filters = ''\n @filters.concat(\"status:'Available'\")\n unless @manufacturer_or_publisher.blank?\n @filters.concat(\" AND (manufacturer:'#{@manufacturer_or_publisher}'\")\n @filters.concat(\" OR publisher:'#{@manufacturer_or_publisher}')\")\n end\n @filters.concat(\" AND category:'#{@category}'\") unless @category.blank?\n @filters.concat(\" AND seller_name:'#{@seller_name}'\") unless @seller_name.blank?\n end",
"def build_selects_from_types(order); end",
"def filter_by_associations_add_conditions_dataset_filter(ds)\n k = filter_by_associations_conditions_associated_keys\n ds.select(*k).where(Sequel.negate(k.zip([])))\n end",
"def index\n @user_types_map = User.user_types.map{|k, v| [k, v.humanize.pluralize]}.to_h\n @filter = params[:filter] || 'Current'\n if @filter == 'All'\n @users = User.order(:email)\n elsif @user_types_map.values.include? @filter\n @users = User.where(locked_out: false, user_type: @user_types_map.key(@filter)).order(:email)\n else\n @users = User.where(locked_out: false).order(:email)\n end\n end",
"def filtering_params(params)\n params.slice(:id, :tipo)\n end",
"def set_filters\n @filters = []\n section_ids = Section.pluck(:id)\n\n [:ministers, :departments].each do |filter_type|\n if params[filter_type].present?\n id_list = params[filter_type].map(&:to_i)\n\n id_list.reject! do |item|\n !section_ids.include? item\n end\n\n @filters += Section.where(id: id_list)\n end\n end\n end"
] |
[
"0.70685226",
"0.7011513",
"0.6773623",
"0.67075807",
"0.65949965",
"0.6562494",
"0.65536016",
"0.65347576",
"0.64803755",
"0.6434349",
"0.64013046",
"0.6395308",
"0.6290136",
"0.6280743",
"0.6252328",
"0.6252328",
"0.62393916",
"0.6214773",
"0.6202415",
"0.6193779",
"0.61319244",
"0.61223334",
"0.6114015",
"0.6104669",
"0.60825694",
"0.60810965",
"0.6079737",
"0.6062547",
"0.60333246",
"0.6023865",
"0.60235196",
"0.6006511",
"0.5996981",
"0.5995883",
"0.5978894",
"0.59587103",
"0.5923492",
"0.5920643",
"0.59152603",
"0.59152603",
"0.59007955",
"0.5898893",
"0.58932614",
"0.5877978",
"0.5865823",
"0.5865004",
"0.58611566",
"0.58492535",
"0.5845786",
"0.583144",
"0.58279467",
"0.5827921",
"0.58277875",
"0.5819094",
"0.58146137",
"0.5810746",
"0.5776116",
"0.57503384",
"0.57480997",
"0.57407653",
"0.57199067",
"0.5716299",
"0.5716074",
"0.5712653",
"0.5710725",
"0.57101667",
"0.5703428",
"0.5694429",
"0.56942976",
"0.56926185",
"0.56874275",
"0.5681388",
"0.56673276",
"0.5662597",
"0.5659476",
"0.5652158",
"0.5635702",
"0.56325835",
"0.5629883",
"0.56295586",
"0.56294656",
"0.56162703",
"0.5616227",
"0.5610209",
"0.56012917",
"0.55996436",
"0.55992097",
"0.55980945",
"0.5584985",
"0.55846834",
"0.55840605",
"0.55827385",
"0.55707777",
"0.55704135",
"0.55672795",
"0.5564305",
"0.55611354",
"0.5559236",
"0.5558076",
"0.55568486"
] |
0.7718533
|
0
|
A list of all tables in the db
|
def tables
db_connection.select_values("show tables")
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def tables\n execute(\"SHOW TABLES FROM #{@opts[:database]}\")\n .map{ |i| i.values }\n .reduce(:+)\n .map{ |i| i.to_sym }\n end",
"def tables()\n tables = []\n @client.query(\"SHOW TABLES;\").each do |table|\n tables.push(table.values[0])\n end\n\n tables\n end",
"def get_tables\r\n table_names = []\r\n tables = @db.execute <<-SQL\r\n SELECT name FROM sqlite_master\r\n WHERE type='table'\r\n ORDER BY name;\r\n SQL\r\n tables.each do |row| \r\n table_names << row[\"name\"]\r\n end\r\n table_names\r\n end",
"def sqlite3_all_tables\n return @dbh.execute(\"select * from sqlite_master where type='table';\")\n end",
"def tables\n execute('select table_name from information_schema.tables where table_schema = ?', [schema]).map(&:first)\n end",
"def tables\n table_names.map { |tn| table(tn) }\n end",
"def tables\n sanity_check\n @handle.tables\n end",
"def tables\n connection.list_tables.select { |table| !exclude?(table) }\n end",
"def list_names_of_all_tables\n tables = @stargate.list_tables\n tables.collect { |t| t.name}\n end",
"def tables(name = nil)\n @connection.table_names\n end",
"def tables(name = nil) #:nodoc:\n select_all(\"select decode(table_name,upper(table_name),lower(table_name),table_name) name from all_tables where owner = sys_context('userenv','session_user')\").map {|t| t['name']}\n end",
"def find_tables\n sql = <<-eos\n SELECT table_name\n FROM information_schema.tables\n WHERE table_type = 'BASE TABLE'\n AND table_schema = 'public'\n eos\n @tables = @connection.exec(sql).values.flatten\n end",
"def tables\n []\n end",
"def list_tables(table=nil)\n q = table ? \"show tables like '#{quote table}'\" : \"show tables\"\n query(q).map(&:first)\n end",
"def tables\r\n return @engine.tables\r\n end",
"def get_all_tables\n query = @db_connection.query('SHOW DATABASES;')\n @db_connection.close if @db_connection\n query.each do |x|\n begin\n db_connection = Mysql.connect(@host, @user, @pass, x[0], @port.to_i)\n print_good(\"Tables for #{x[0]}\")\n query = db_connection.query('SHOW TABLES;')\n query.each { |y| print_line(\"#{y[0]}\") }\n puts\n db_connection.close if db_connection\n rescue Mysql::Error => e\n print_error(\"Problem getting tables from #{x[0]}!\")\n print_error(\"#{e}\")\n end\n end\n @db_connection = Mysql.connect(@host, @user, @pass, @db, @port.to_i)\n end",
"def sqlite3_list_tables\n tables = @dbm.sqlite3_all_tables\n if !tables.empty?\n tables.each {|table|\n printf \"#{table[1]} | #{table[4]}\\n\"\n }\n else\n printf \"Database [#{@db_name}] is empty.\\n\" \n end\n end",
"def tables(opts={})\n execute('SHOW TABLES').map{|i| i.values}.reduce(:+).map{|i| i.to_sym}\n end",
"def tables(_name = nil)\n stmt = @connection.tables\n result = stmt.fetch_all || []\n stmt.drop\n\n result.each_with_object([]) do |row, table_names|\n schema_name, table_name, table_type = row[1..3]\n next if respond_to?(:table_filtered?) && table_filtered?(schema_name, table_type)\n table_names << format_case(table_name)\n end\n end",
"def tables(_name = nil)\n stmt = @connection.tables\n result = stmt.fetch_all || []\n stmt.drop\n\n result.each_with_object([]) do |row, table_names|\n schema_name, table_name, table_type = row[1..3]\n next if respond_to?(:table_filtered?) && table_filtered?(schema_name, table_type)\n table_names << format_case(table_name)\n end\n end",
"def list\n Table.with_db do |db|\n return @db\n .execute(\"select name from sqlite_master where type='table' ORDER BY name;\")\n .flatten \n end\n end",
"def get_tables\n tables\n end",
"def tables_from(db=current_database)\n end",
"def tables\n return nil unless has_connection?\n self.connection.tables\n end",
"def show_tables\n execute('select table_name from information_schema.tables where table_schema = ?', [schema])\n end",
"def tables(name = nil)\n query(<<-SQL, 'SCHEMA').map { |row| row[0] }\n SELECT tablename\n FROM pg_tables\n WHERE schemaname = ANY (ARRAY['public'])\n SQL\n end",
"def tables\n @connection.tables.select {|row| row.to_s !~ /^system_/i }\n end",
"def tables(name = nil) #:nodoc:\n tables = @connection.tables\n tables.reject! { |t| /\\A_SYS_/ === t }\n tables\n end",
"def base_tables(name = nil)\n # this is untested\n select_values(\"SELECT table_name FROM information_schema.tables\", name)\n end",
"def tables(name = nil)\n query(<<-SQL, name).map { |row| row[0] }\n SELECT table_name\n FROM v_catalog.tables\n WHERE table_schema = 'public'\n SQL\n end",
"def table_names\n @tables.keys\n end",
"def tables\n self.probe_tables unless @tables\n @tables\n end",
"def tables(name = nil)\n select_rows(<<-SQL, name).map { |row| row[0] }\nSELECT name\nFROM sysobjects\nWHERE type = 'U'\nSQL\n end",
"def tables(name = nil, schema = nil, table = nil)\n schema = schema ? \"'#{quote_string(schema)}'\" : 'CURRENT_SCHEMA'\n select_rows(\n \"SELECT table_name \"+\n \"FROM information_schema.tables \"+\n \"WHERE table_type = 'TABLE' \"+\n \" AND table_schema = #{schema} \"+\n (table ? \"AND table_name = '#{quote_string(table)}'\" : \"\"),\n SCHEMA_LOG_NAME\n ).map { |row|\n row[0]\n }\n end",
"def list_tables\n data.keys\n end",
"def table_names\n with_admin { |admin| admin.list_tables.map(&:name_as_string) }\n end",
"def tables(name = nil)\n if name\n ActiveSupport::Deprecation.warn(<<-MSG.squish)\n Passing arguments to #tables is deprecated without replacement.\n MSG\n end\n\n select_values(\"SELECT tablename FROM pg_tables WHERE schemaname = ANY(current_schemas(false))\", 'SCHEMA')\n end",
"def tables(name = nil) #:nodoc:\n tables = @connection.tables\n tables.reject! { |t| /\\A_SYS_/ === t }\n end",
"def tables\n @tables ||= if @registration[:tables].present?\n @registration[:tables].call(@connection)\n else\n @connection.tables\n end\n end",
"def tables; ActiveRecord::Base.connection.tables; end",
"def tables(name = nil, database = nil, like = nil)\n database = database ? quote_table_name(database) : \"DATABASE()\"\n by_name = like ? \"AND table_name LIKE #{quote(like)}\" : \"\"\n\n sql = <<-SQL.squish\n SELECT table_name, table_type\n FROM information_schema.tables\n WHERE table_schema = #{database}\n AND table_type = 'BASE TABLE'\n #{by_name}\n SQL\n\n execute_and_free(sql, 'SCHEMA') do |result|\n rows_from(result).map(&:first)\n end\n end",
"def tables\n tables = []\n @a.keys.each{|e| if !e.match(/_base$/) then tables.push(e) end}\n tables\n end",
"def tables\n [\n ]\n end",
"def get_table_names\n raise \"No database file selected.\" unless $current_db\n\n $tables = get_data \"select name from sqlite_master\"\n $tables.collect!{|x| x[0] } ## 1.9 hack, but will it run on 1.8 ??\n $tables\nend",
"def tables(name = nil) #:nodoc:\r\n sql = \"EXECUTE PROCEDURE sp_GetTables( NULL, NULL, NULL, 'TABLE' );\"\r\n select(sql, name).map { |row| strip_or_self(row[\"TABLE_NAME\"]) }\r\n end",
"def get_table_list()\n resp = @dynamo_db.list_tables()\n puts \"Table list : \" + resp.table_names.to_s\nend",
"def tables\n Dir[File.join(dir, \"*.table\")].map { |file| Table.new(File.read(file)) }\n end",
"def get_tables\n get_schemas.keys\n end",
"def get_tables(db)\n if db == @db\n query = @db_connection.query('SHOW TABLES;')\n query.each { |x| puts \"#{x[0]}\".white }\n else\n @db_connection.close if @db_connection\n begin\n db_connection = Mysql.connect(@host, @user, @pass, db, @port.to_i)\n query = db_connection.query('SHOW TABLES;')\n query.each { |x| puts \"#{x[0]}\".white }\n db_connection.close if db_connection\n rescue Mysql::Error => e\n print_error(\"Problem getting tables from #{db}!\")\n print_error(\"#{e}\")\n end\n @db_connection = Mysql.connect(@host, @user, @pass, @db, @port.to_i)\n end\n end",
"def table_names\n @table_names ||= ActiveRecord::Base.connection.tables - excluded_table_names\n end",
"def application_tables\n exclusions = %w[schema_migrations ar_internal_metadata vat_rates fee_schemes]\n ActiveRecord::Base.connection.tables.uniq.sort - exclusions\n end",
"def tables\n Mongo::DocumentDefinition.defined_document_names\n #db.collection_names\n end",
"def compute_table_list\n if @configuration.download_tables\n # If the list is explicitly set then use that\n tables = @configuration.download_tables.to_set\n else\n # Otherwise guess via the tables actually in the database\n tables = @adapter.guess_tables.to_set\n\n if @configuration.allow_tables\n # Only allow tables tables that we specify\n tables = tables.intersection @configuration.allow_tables\n end\n if @configuration.disallow_tables\n # Remove any tables that we don't want included\n tables = tables.difference @configuration.disallow_tables\n end\n end\n return tables\n end",
"def tables_to_export_from_connection(conn)\n tables = []\n conn.tables.each do |table|\n next if table == 'schema_migrations'\n count_rows = conn.execute(\"SELECT 1 FROM `#{table}` LIMIT 1\").count\n tables << table if count_rows > 0\n end\n tables\n end",
"def get_tables(db_name)\n response = @glue_client.get_tables(database_name: db_name)\n response.table_list\n rescue Aws::Glue::Errors::GlueException => e\n @logger.error(\"Glue could not get tables #{db_name}: \\n#{e.message}\")\n raise\n end",
"def tables(opts={})\n m = output_identifier_meth\n metadata_dataset.from(:information_schema__tables___t).\n select(:table_name).\n filter(:table_type=>'BASE TABLE', :table_schema=>(opts[:schema]||default_schema||'dbo').to_s).\n map{|x| m.call(x[:table_name])}\n end",
"def table_names\n @table_names ||= entities.map{|entity|entity.model.table_name}.uniq\n end",
"def schema\n schema = Schema.new\n\n for name in db.tables\n next if IGNORED_TABLES.include? name\n table = schema.add_table( name )\n import_table( table )\n end\n\n schema\n end",
"def tables\n ActiveRecord::Base.connection.tables.sort!\nend",
"def print_table_names\r\n get_tables.each do |name| \r\n print \"#{name} \"\r\n end\r\n puts \"\"\r\n end",
"def show_tables\r\n tables = Document.new(self).get_tables\r\n puts \"There are #{tables.length} tables\"\r\n index = 1\r\n tables.each do |l|\r\n puts \"table: id: #{l.id}\"\r\n puts \" rows: #{l.row_count}\"\r\n puts \" columns: #{l.column_count}\"\r\n puts \" index: #{index}\"\r\n index += 1\r\n end\r\n end",
"def get_tables\n\t\t@@tables\n\tend",
"def tables(query)\n SqlAssess::Parsers::Tables.new(query).tables.map do |table|\n if table.key?(:join_type)\n table[:table][:table].remove('`')\n else\n table[:table].remove('`')\n end\n end\n end",
"def tables\r\n conn_exec do |driver|\r\n if !@tables\r\n require \"#{File.dirname(__FILE__)}/drivers/#{@opts[:type]}/knjdb_#{@opts[:type]}_tables\" if (!@opts.key?(:require) or @opts[:require])\r\n @tables = Kernel.const_get(\"KnjDB_#{@opts[:type]}\".to_sym).const_get(:Tables).new(\r\n :db => self\r\n )\r\n end\r\n \r\n return @tables\r\n end\r\n end",
"def tables(opts=OPTS, &block)\n pg_class_relname(['r', 'p'], opts, &block)\n end",
"def tables\n raise 'SevenZip#tables should never be called'\n end",
"def get_tables_db(db_name)\n if @target_config['VERSION'].scan(/./)[0].to_i >= 5\n # MySQL >= 5\n if not @target_config['DB_TABLES'].keys.include?(db_name)\n query = \"select count(table_name) from information_schema.tables where table_schema=#{db_name.mysqlhex}\"\n results = union_basic_inject($config['INJECTOR']['MYSQL']['UNION']['VULN_COLUMN'].to_i, query)\n if results.nil?\n print_error(\"Unable to determine number of tables in current database....\")\n else\n print_good(\"Fetching #{results} Tables from Current DB\")\n end\n count=0\n tables=[]\n while not results.nil?\n results = union_basic_inject($config['INJECTOR']['MYSQL']['UNION']['VULN_COLUMN'].to_i, \"select table_name from information_schema.tables where table_schema=#{db_name.mysqlhex} limit #{count},1\")\n pad = ' ' * (results.size + 25) unless results == '' or results.nil?\n pad = ' ' * 50 if results.nil?\n print \"\\r(#{count})> #{results}#{pad}\".cyan unless results == ''\n tables << results unless results == ''\n count = count.to_i + 1\n end\n print_line(\"\")\n if tables.empty?\n print_error(\"Unable to get any tables from the current database!\")\n print_error(\"Lack of privileges? IDK....\")\n print_status(\"Possible Solutions include:\")\n print_caution(\"A) Become HR's best friend by updating the code and sending him a copy\")\n print_caution(\"B) Tweak Settings and try things again\")\n print_caution(\"C) Be a bawz and do it manually\")\n print_line(\"\")\n return nil\n else\t\n @target_config['DB_TABLES'][\"#{db_name}\"] = tables\n print_good(\"DB: #{db_name}\")\n print_good(\"Tables: #{tables.join(', ').sub(/, $/, '')}\")\n return tables.join(' ')\n end\n else\n db = db_name\n tables = @target_config['DB_TABLES'][\"#{db_name}\"].join(', ').sub(/, $/, '')\n print_good(\"DB: #{db}\")\n print_good(\"Tables: #{tables}\")\n end\n else\n # MySQL < 5\n print_error(\"MySQL < 5: #{@target_config['VERSION']}\")\n print_error(\"There is no information_schema to query for tables as result.....\")\n if not @target_config['CURRENT_DB'].nil?\n print_error(\"Do you want to try Common Table Names (Y/N)?\")\n answer=gets.chomp\n print_line(\"\")\n if answer.upcase == 'Y' or answer.upcase == 'YES'\n tables = common_tables('CURRENT-DB', \"#{HOME}fuzz/common_tables.lst\")\n if tables.empty?\n print_error(\"OK, returning to menu...\")\n return nil\n else\n print_good(\"DB: #{db_name}\")\n print_good(\"Tables: #{tables.join(', ').sub(/, $/, '')}\")\n @target_config['DB_TABLES'][\"#{db_name}\"] = tables\n return tables.join(' ')\n end\n else\n print_error(\"OK, returning to menu...\")\n return nil\n end\n else\n print_error(\"OK, returning to menu...\")\n return nil\n end\n end\n end",
"def all\n self.as_objects(CONNECTION.execute(\"SELECT * FROM #{table_name};\"))\n end",
"def get_all_databases\n databases = ::MySQL::Database.find_all\n names = []\n databases.each do |database|\n names << database.name\n end\n names\n end",
"def all\n partitions.group_by { |row| row['table_name'] }.map(&method(:to_tablature_table))\n end",
"def schema\n execute(<<-eosql).collect { |row| row[0] }.collect { |t| table_schema(t) }\nSELECT rdb$relation_name FROM rdb$relations WHERE rdb$system_flag != 1\neosql\n end",
"def list_tables_by_state(isEnabled)\n @admin.listTableNamesByState(isEnabled).map(&:getNameAsString)\n end",
"def tables\n [\n ::GDO::Login::GDO_LoginHistory,\n ::GDO::Login::GDO_LoginAttempts,\n ]\n end",
"def list_tables instance_id, view: nil\n execute do\n tables.list_tables(\n instance_path(instance_id),\n view: view\n )\n end\n end",
"def tables(opts=OPTS, &block)\n tables_or_views('TABLE', opts, &block)\n end",
"def list(regex = '.*')\n @admin.listTableNames(Pattern.compile(regex)).map(&:getNameAsString)\n end",
"def get_tables(options={})\n return send_message(SkyDB::Message::GetTables.new(options))\n end",
"def tables\n options.fetch(:tables)\n end",
"def get_storage_names\n # This gets all the non view tables, but has to strip column 0 out of the two column response.\n select(\"SHOW FULL TABLES FROM #{options[:path][1..-1]} WHERE Table_type = 'BASE TABLE'\").map { |item| item.first }\n end",
"def table_list(our_join = join, our_tables = tables)\n flatten_joins([model.table_name] + our_join + our_tables, false).uniq\n end",
"def get_all_related_tables\n tables = [ @table_name ]\n return tables if @relations.nil?\n @relations.each do |n,r|\n tables = tables | get_relation_tables(r)\n end\n tables.sort\n end",
"def get_all_tables\n\t\ttc = new_sub(@discovery['tables'])\n\t\ttc.poll[:messages].map {|x| pp x; JSON.parse(x.content) rescue nil}.compact\n\tend",
"def get_all_tables\n\t\ttc = new_sub(@discovery['tables']['url'], @discovery['tables']['capability'])\n\t\ttc.listen.map {|x| JSON.parse(x) rescue nil}.compact\n\tend",
"def all_schemas\n query('SELECT schema_name FROM information_schema.schemata').flatten\n end",
"def load_tables\n tables = @table_ast.keys\n tables.each {|table|\n columns = Hash.new\n column_names = @table_ast[table].keys\n rows = @dbh.execute(\"SELECT * FROM #{table}\")\n @mem_db_row[table] = rows\n column_names.each {|col|\n col_data = Array.new\n rows.each {|data|\n col_data.push(data[col])\n }\n columns[col] = col_data\n }\n @mem_db_col[table] = columns\n }\n return @mem_db_col, @mem_db_row\n end",
"def all_tables\n log_warning('error identifying tables in AWS.. please check credentials') do\n glue_tables(\n glue_client,\n glue_databases(glue_client),\n get_deployment_id\n )\n end\n end",
"def revision_tables\n database.tables.select { |t| revision_table?(t) }.sort\n end",
"def index\n @schema_tables = SchemaTable.all\n end",
"def tables view: nil\n client.list_tables(\n instance_path,\n view: view\n )\n end",
"def list_tables instance_id, view: nil\n tables.list_tables parent: instance_path(instance_id), view: view\n end",
"def list\n @table.keys\n end",
"def get_all_data(table)\n return $db.execute('SELECT * FROM ' + table);\n end",
"def blob_tables\n execute('select table_name from information_schema.tables where table_schema = ?', ['blob']).map(&:first)\n end",
"def all_dbs\n @conn.query({url_path: \"_all_dbs\", method: :get})\n end",
"def get_tables_db(db_name)\n if @target_config['VERSION'].scan(/./)[0].to_i < 5\n # MySQL < 5\n print_error(\"MySQL < 5: #{@target_config['VERSION']}\")\n print_error(\"There is no information_schema to query for tables as result.....\")\n if not @target_config['CURRENT_DB'].nil?\n print_error(\"Do you want to try Common Table Names (Y/N)?\")\n answer = gets.chomp\n print_line(\"\")\n if answer.upcase == 'Y' or answer.upcase == 'YES'\n tables = common_tables('CURRENT-DB', \"#{HOME}/fuzz/common_tables.lst\")\n if tables.empty?\n print_error(\"OK, returning to menu...\")\n return nil\n else\n print_good(\"DB: #{db_name}\")\n print_good(\"Tables: #{tables.join(', ').sub(/, $/, '')}\")\n @target_config['DB_TABLES'][\"#{db_name}\"] = tables\n return tables.join(' ')\n end\n else\n print_error(\"OK, returning to menu...\")\n return nil\n end\n else\n print_error(\"OK, returning to menu...\")\n return nil\n end\n elsif @target_config['VERSION'].scan(/./)[0].to_i >= 5\n # MySQL >= 5\n if not @target_config['DB_TABLES'].keys.include?(db_name)\n query = \"select count(table_name) from information_schema.tables where table_schema=#{db_name.mysqlhex}\"\n results = error_basic_inject(query)\n if results.nil?\n print_error(\"Unable to determine number of tables in current database....\")\n else\n print_good(\"Fetching #{results} Tables from Current DB\")\n end\n count = 0\n tables=[]\n while not results.nil?\n results = error_basic_inject(\"select table_name from information_schema.tables where table_schema=#{db_name.mysqlhex} limit #{count},1\")\n pad = ' ' * (results.size + 25) unless results.nil? or results == ''\n pad = ' ' * 50 if results.nil? or results == ''\n print \"\\r(ROW##{count})> #{results}#{pad}\".cyan unless results == ''\n tables << results unless results == ''\n count = count.to_i + 1\n end\n print_line(\"\")\n if tables.empty?\n print_error(\"Unable to get any tables from the current database!\")\n print_error(\"Lack of privileges? IDK....\")\n print_status(\"Possible Solutions include:\")\n print_caution(\"A) Become HR's best friend by updating the code and sending him a copy\")\n print_caution(\"B) Tweak Settings and try things again\")\n print_caution(\"C) Be a bawz and do it manually\")\n print_line(\"\")\n return nil\n else\t\n @target_config['DB_TABLES'][\"#{db_name}\"] = tables\n print_good(\"DB: #{db_name}\")\n print_good(\"Tables: #{tables.join(', ').sub(/, $/, '')}\")\n return tables.join(' ')\n end\n else\n db = db_name\n tables = @target_config['DB_TABLES'][\"#{db_name}\"].join(', ').sub(/, $/, '')\n print_good(\"DB: #{db}\")\n print_good(\"Tables: #{tables}\")\n end\n end\n end",
"def tables(key = nil)\n key ||= properties.key1\n response = table_response(key, nil, \"Tables\")\n JSON.parse(response.body)['value'].map{ |t| Table.new(t) }\n end",
"def revision_tables\n Spontaneous::Publishing::Revision.tables(self)\n # database.tables.select { |t| revision_table?(t) }.sort\n end",
"def included_tables\n tables = table_names\n if group_name\n g = group(group_name)\n return [] if g['include_all']\n tables += (g['include_tables'] || [])\n end\n tables.uniq.sort\n end",
"def inherited_tables\n tables = query(<<-SQL, 'SCHEMA')\n SELECT child.relname AS table_name,\n array_agg(parent.relname) AS inheritances\n FROM pg_inherits\n JOIN pg_class parent ON pg_inherits.inhparent = parent.oid\n JOIN pg_class child ON pg_inherits.inhrelid = child.oid\n GROUP BY child.relname, pg_inherits.inhrelid\n ORDER BY pg_inherits.inhrelid\n SQL\n\n tables.map do |(table, refs)|\n [table, Coder.decode(refs)]\n end.to_h\n end",
"def get_tables\n if @target_config['CDB_TABLES'].empty?\n if not @target_config['VERSION'].nil?\n if @target_config['VERSION'].scan(/./)[0].to_i < 5\n # MySQL < 5\n print_error(\"MySQL < 5: #{@target_config['VERSION']}\")\n print_error(\"There is no information_schema to query for tables as result.....\")\n if not @target_config['CURRENT_DB'].nil?\n print_error(\"Do you want to try Common Table Names (Y/N)?\")\n answer=gets.chomp\n print_line(\"\")\n if answer.upcase == 'Y' or answer.upcase == 'YES'\n tables = common_tables('CURRENT-DB', \"#{HOME}fuzz/common_tables.lst\")\n if tables.empty?\n print_error(\"OK, returning to menu...\")\n return nil\n else\n print_good(\"DB: #{@target_config['CURRENT_DB']}\")\n print_good(\"Tables: #{tables.join(', ').sub(/, $/, '')}\")\n @target_config['CDB_TABLES'] = tables\n @target_config['DB_TABLES'][\"#{@target_config['CURRENT_DB']}\"] = @target_config['CDB_TABLES'] unless @target_config['CURRENT_DB'].nil?\n @target_config['DB_TABLES']['Current DB'] = @target_config['CDB_TABLES'] if @target_config['CURRENT_DB'].nil?\n return tables.join(' ')\n end\n else\n print_error(\"OK, returning to menu...\")\n return nil\n end\n else\n print_error(\"OK, returning to menu...\")\n return nil\n end\n else\n # MySQL >= 5\n if @target_config['CURRENT_DB'].nil?\n print_error(\"Current DB has not yet been discovered!\")\n print_error(\"Try BASIC command if you haven't already!\")\n print_error(\"You can also use DBTABLES command if you know the DB Name....\")\n print_error(\"FUZZ_TABLES can be used as a last resort as well if you know the DB Name...\")\n return nil\n else\n count=0\n tables=[]\n query = 'select count(table_name) from information_schema.tables where table_schema=database()'\n results = union_basic_inject($config['INJECTOR']['MYSQL']['UNION']['VULN_COLUMN'].to_i, query)\n if results.nil?\n query = 'select count(table_name) from information_schema.tables where table_schema=schema()'\n results = union_basic_inject($config['INJECTOR']['MYSQL']['UNION']['VULN_COLUMN'].to_i, query)\n if results.nil?\n query = \"select count(table_name) from information_schema.tables where table_schema=#{@target_config['CURRENT_DB'].mysqlhex}\"\n results = union_basic_inject($config['INJECTOR']['MYSQL']['UNION']['VULN_COLUMN'].to_i, query)\n if results.nil?\n print_error(\"Unable to determine number of tables in current database, sorry....\")\n end\n end\n end\n if not results.nil?\n print_good(\"Fetching #{results} Tables from Current DB\")\n case query\n when /database()/\n dbn='database()'\n when /schema()/\n dbn='schema()'\n when /0x[a-z0-9]{1,}/\n dbn=\"#{@target_config['CURRENT_DB'].mysqlhex}\"\n end\n while not results.nil?\n results = union_basic_inject($config['INJECTOR']['MYSQL']['UNION']['VULN_COLUMN'].to_i, \"select table_name from information_schema.tables where table_schema=#{dbn} limit #{count},1\")\n pad = ' ' * (results.size + 25) unless results == '' or results.nil?\n pad = ' ' * 50 if results.nil?\n print \"\\r(#{count})> #{results}#{pad}\".cyan unless results == ''\n tables << results unless results == ''\n count = count.to_i + 1\n end\n print_line(\"\")\n if tables.empty?\n print_line(\"\")\n print_error(\"Unable to get any tables from the current database!\")\n print_error(\"Lack of privileges? IDK....\")\n print_status(\"Possible Solutions include:\")\n print_caution(\"A) Become HR's best friend by updating the code and sending him a copy\")\n print_caution(\"B) Tweak Settings and try things again\")\n print_caution(\"C) Be a bawz and do it manually\")\n print_line(\"\")\n return nil\n else\t\n @target_config['CDB_TABLES'] = tables\n @target_config['DB_TABLES'][\"#{@target_config['CURRENT_DB']}\"] = @target_config['CDB_TABLES'] unless @target_config['CURRENT_DB'].nil?\n @target_config['DB_TABLES']['Current DB'] = @target_config['CDB_TABLES'] if @target_config['CURRENT_DB'].nil?\n print_good(\"Current DB: #{@target_config['CURRENT_DB']}\")\n print_good(\"Tables: #{tables.join(', ').sub(/, $/, '')}\")\n return tables.join(' ')\n end\n else\n print_error(\"Do you want to try Common Table Names (Y/N)?\")\n answer=gets.chomp\n print_line(\"\")\n if answer.upcase == 'Y' or answer.upcase == 'YES'\n tables = common_tables('CURRENT-DB', \"#{HOME}fuzz/common_tables.lst\")\n if tables.empty?\n print_error(\"OK, returning to menu...\")\n return nil\n else\n print_good(\"DB: #{@target_config['CURRENT_DB']}\")\n print_good(\"Tables: #{tables.join(', ').sub(/, $/, '')}\")\n @target_config['CDB_TABLES'] = tables\n @target_config['DB_TABLES'][\"#{@target_config['CURRENT_DB']}\"] = @target_config['CDB_TABLES']\n return tables.join(' ')\n end\n else\n print_error(\"OK, returning to menu...\")\n return nil\n end\n end\n end\n end\n else\n print_error(\"No version info collected yet!\")\n print_error(\"Try using the BASIC comamnd and then try again....\")\n print_error(\"You could also use the FUZZ TABLES option....\")\n end\n else\n print_good(\"DB: #{@target_config['CURRENT_DB']}\") unless @target_config['CURRENT_DB'].nil?\n print_good(\"Tables: #{@target_config['CDB_TABLES'].join(', ').sub(/, $/, '')}\")\n @target_config['DB_TABLES'][\"#{@target_config['CURRENT_DB']}\"] = @target_config['CDB_TABLES'] unless @target_config['CURRENT_DB'].nil?\n @target_config['DB_TABLES']['Current DB'] = @target_config['CDB_TABLES'] if @target_config['CURRENT_DB'].nil?\n end\n end"
] |
[
"0.8499479",
"0.83081716",
"0.81848574",
"0.81427336",
"0.80861485",
"0.80534613",
"0.7979202",
"0.7953699",
"0.79384804",
"0.79357654",
"0.7846949",
"0.7829539",
"0.7824308",
"0.7812821",
"0.78055245",
"0.77726704",
"0.7763799",
"0.7755373",
"0.7754824",
"0.77542955",
"0.76937014",
"0.7679",
"0.7668839",
"0.765747",
"0.76425403",
"0.75763243",
"0.75344557",
"0.75114274",
"0.7505624",
"0.7449119",
"0.73935956",
"0.73891276",
"0.7388535",
"0.7362489",
"0.7349033",
"0.73448133",
"0.73417693",
"0.7332874",
"0.73269117",
"0.73143935",
"0.7291985",
"0.72618556",
"0.72547203",
"0.7249728",
"0.72404605",
"0.7226649",
"0.7221901",
"0.7119809",
"0.7108405",
"0.7104909",
"0.7098598",
"0.7061676",
"0.7034794",
"0.6999538",
"0.69775844",
"0.6940123",
"0.6921449",
"0.68955785",
"0.6849039",
"0.6834285",
"0.68209624",
"0.67976993",
"0.6772682",
"0.6748659",
"0.6695718",
"0.6665534",
"0.66563445",
"0.66529214",
"0.6627077",
"0.65988314",
"0.6597701",
"0.6595438",
"0.65925455",
"0.6561537",
"0.65615076",
"0.65442216",
"0.6531632",
"0.65281606",
"0.6516305",
"0.65113163",
"0.65025127",
"0.6493889",
"0.648513",
"0.6483758",
"0.64827317",
"0.64545876",
"0.64390445",
"0.64345",
"0.643437",
"0.64338183",
"0.64293945",
"0.6423553",
"0.64198834",
"0.6352874",
"0.63399196",
"0.6333307",
"0.6320036",
"0.63167876",
"0.6312548",
"0.62858117"
] |
0.84634954
|
1
|
A list of all columns for table
|
def columns_for(table)
db_connection.select_values("show columns from `#{ table }`")
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def list_columns\n columns\n end",
"def table_columns\n klass.column_names\n end",
"def all\n __getobj__.column_names\n end",
"def get_table_column_names(table)\r\n @db[table].columns\r\n end",
"def columns(table_name)\n @db.table_info(table_name).map { |info| info[\"name\"].to_sym }\n end",
"def columns(table_name, name = nil) end",
"def columns\n @columns ||= []\n end",
"def columns; @columns; end",
"def columns\n @columns\n end",
"def columns\n @columns\n end",
"def columns( table )\n sanity_check\n @handle.columns( table ).collect {|col| ColumnInfo.new(col) }\n end",
"def columns()\n return @columns.keys\n end",
"def columns\n @columns ||= connection.columns(table_name, \"#{name} Columns\")\n end",
"def columns; end",
"def get_columns\n return @columns\n end",
"def column_names\n columns.map(&:name)\n end",
"def columns_for(table_name)\n self.connection.columns(table_name)\n end",
"def columns(table_name, *rest)\n @table.fields.map {|s| ::ActiveRecord::ConnectionAdapters::Column.new(s.to_s, nil, String)}\n end",
"def column_names(table_name)\n columns(table_name).map(&:name)\n end",
"def columns\n @columns\n end",
"def columns\n collect = []\n recursive_columns(@base, collect)\n end",
"def column_names\n @columns.keys\n end",
"def columns\n @columns ||= [];\n end",
"def columns\n _columns || columns!\n end",
"def columns \n []\n end",
"def columns\n @model.columns.map{|c| c.name}\n end",
"def table_columns\n ActiveRecord::Base.connection.columns(table_name)\n end",
"def columns\n @columns_hash.values\n end",
"def columns\n @columns\n end",
"def columns\n @columns\n end",
"def columns\n end",
"def columns\n @stmt.columns\n end",
"def all_columns(excluded_supertypes) #:nodoc:\n columns = []\n trace :columns, \"All Columns for #{name}\" do\n if is_table\n self_value_reference\n else\n columns << ActiveFacts::RMap::Column.new\n end\n references_from.each do |ref|\n trace :columns, \"Columns absorbed via #{ref}\" do\n columns += ref.columns({})\n end\n end\n end\n columns\n end",
"def columns\n self.class.instance_variable_get(:@columns)\n end",
"def columns \n @c\n end",
"def columns; @columns_hash.values; end",
"def columns\n @stmt.columns\n end",
"def columns\n raise NotImplementedError\n end",
"def columns(table = nil)\n if table\n @columns[table]\n else\n ActiveSupport::Deprecation.warn('call columns with a table name!')\n @columns.dup\n end\n end",
"def columns\n return @columns\n end",
"def columns\n if @column_names\n columns = []\n\n # First match against fully named columns, e.g. 'attribute:name'\n @column_names.each{|cn| columns << owner.columns_hash[cn] if owner.columns_hash.has_key?(cn)}\n\n # Now match against aliases if the number of columns found previously do not\n # match the expected @columns_names size, i.e. there's still some missing.\n if columns.size != @column_names.size\n columns_left = @column_names - columns.map{|column| column.name}\n owner.columns_hash.each { |name,column| columns << column if columns_left.include?(column.alias) }\n end\n\n columns\n else\n owner.columns\n end\n end",
"def columns\r\n unless @columns\r\n @columns = connection.columns(table_name, \"#{name} Columns\")\r\n @columns.each {|column| column.primary = primary_keys.include?(column.name.to_sym)}\r\n end\r\n @columns\r\n end",
"def columns\n @columns\n end",
"def columns\n\t\t\t@columns ||= Txtboard::Column.columns.sort_by(&:order)\n\t\tend",
"def column_names\n @column_names ||= columns.collect(&:name)\n end",
"def columns\n self.class.columns\n end",
"def columns\n self.class.columns\n end",
"def get_column_names(table_name)\r\n column_names = get_info(table_name, 'name')\r\n end",
"def columns\n @columns || self.class.columns\n end",
"def columns\n get_metadata unless @columns\n return @columns\n end",
"def columns\n self.class.const_get(:COLUMNS) rescue []\n end",
"def columns(table_name)\n column_definitions(table_name.to_s).map do |column_name, type, default, notnull, oid, fmod|\n default_value = extract_value_from_default(default)\n type_metadata = fetch_type_metadata(column_name, type, oid, fmod)\n default_function = extract_default_function(default_value, default)\n new_column(column_name, default_value, type_metadata, notnull == 'f', table_name, default_function)\n end\n end",
"def column_definitions(table_name)\r\n query <<-end_sql\r\n SELECT a.attname, format_type(a.atttypid, a.atttypmod), d.adsrc, a.attnotnull, c.consrc\r\n FROM pg_attribute a LEFT JOIN pg_attrdef d\r\n ON a.attrelid = d.adrelid AND a.attnum = d.adnum\r\n LEFT JOIN pg_constraint c ON a.attrelid = c.conrelid AND \r\n c.contype = 'c' AND c.conkey[1] = a.attnum\r\n WHERE a.attrelid = '#{table_name}'::regclass\r\n AND a.attnum > 0 AND NOT a.attisdropped\r\n ORDER BY a.attnum\r\n end_sql\r\n end",
"def columns\n self.columns = @core.columns._inheritable unless @columns # lazy evaluation\n @columns\n end",
"def columns\n self.columns = @core.columns._inheritable unless @columns # lazy evaluation\n @columns\n end",
"def columns(table_name, name = nil)#:nodoc:\n sql = \"SHOW FIELDS FROM #{quote_table_name(table_name)}\"\n columns = []\n result = execute(sql, name)\n result.each do |field|\n klass = field[1] =~ /geometry|point|linestring|polygon|multipoint|multilinestring|multipolygon|geometrycollection/i ? ActiveRecord::ConnectionAdapters::SpatialMysql2Column : ActiveRecord::ConnectionAdapters::Mysql2Column\n columns << klass.new(field[0], field[4], field[1], field[2] == \"YES\")\n end\n columns\n end",
"def columns!\n @columns = nil\n columns\n end",
"def columns!\n ds = clone(opts.merge(:sql => \"SELECT TOP 1 [T1].* FROM (#{opts[:sql]}) \\\"T1\\\"\"))\n ds.each{break}\n\n if cols = ds.cache[:_columns]\n self.columns = cols\n else\n []\n end\n end",
"def column_definitions(table_name)\n select_rows(<<-end_sql, 'SCHEMA')\n SELECT a.attname, format_type(a.atttypid, a.atttypmod),\n pg_get_expr(d.adbin, d.adrelid), a.attnotnull, a.atttypid, a.atttypmod\n FROM pg_attribute a LEFT JOIN pg_attrdef d\n ON a.attrelid = d.adrelid AND a.attnum = d.adnum\n WHERE a.attrelid = '#{quote_table_name(table_name)}'::regclass\n AND a.attnum > 0 AND NOT a.attisdropped\n ORDER BY a.attnum\n end_sql\n end",
"def column_names\n klass.properties.keys\n end",
"def columns\n single_record unless @columns\n @columns || []\n end",
"def columns\n @columns = columns_hash.values\n end",
"def column_names\n @column_names ||= @columns.keys\n end",
"def columns\n @columns ||= @stats.values.flatten.map { |a| a.column_name }\n end",
"def columns\n @header.to_a\n end",
"def columns\n [\"#{@base}.*\"]\n end",
"def column_all_sql(ca)\n \"#{quote_identifier(ca.table)}.*\"\n end",
"def columns\n @columns ||= dataset.columns\n end",
"def columns\n @columns ||= dataset.columns\n end",
"def column_definitions(table_name)\n select_rows(<<-end_sql, 'SCHEMA')\n SELECT a.attname, format_type(a.atttypid, a.atttypmod),\n pg_get_expr(d.adbin, d.adrelid), a.attnotnull, a.atttypid, a.atttypmod,\n (SELECT c.collname FROM pg_collation c, pg_type t\n WHERE c.oid = a.attcollation AND t.oid = a.atttypid\n AND a.attcollation <> t.typcollation),\n col_description(a.attrelid, a.attnum) AS comment\n FROM pg_attribute a\n LEFT JOIN pg_attrdef d ON a.attrelid = d.adrelid AND a.attnum = d.adnum\n WHERE a.attrelid = #{quote(quote_table_name(table_name))}::regclass\n AND a.attnum > 0 AND NOT a.attisdropped\n ORDER BY a.attnum\n end_sql\n end",
"def columns\n # we want to delay initializing to the @core.columns set for as long as possible. Too soon and .search_sql will not be available to .searchable?\n unless defined? @columns\n self.columns = @core.columns.collect { |c| c.name if @core.columns._inheritable.include?(c.name) && c.searchable? && c.association.nil? && c.text? }.compact\n end\n @columns\n end",
"def column_names(table)\n self.table_column_names ||= {}\n unless table_column_names.include? table\n table_column_names[table] = columns(table).map {|c| c.name}\n end\n table_column_names[table]\n end",
"def wanted_column_names\n nil # nil => select all columns\n end",
"def columns(tables)\n columns = execute <<-SQL\n SELECT\n s.nspname || '.' || t.relname AS table_name,\n a.attname AS column_name,\n pg_catalog.format_type(a.atttypid, a.atttypmod) AS data_type\n FROM pg_attribute a\n JOIN pg_class t on a.attrelid = t.oid\n JOIN pg_namespace s on t.relnamespace = s.oid\n WHERE a.attnum > 0\n AND NOT a.attisdropped\n AND s.nspname || '.' || t.relname IN (#{tables.map {|t| quote(t)}.join(\", \")})\n ORDER BY\n 1, 2\n SQL\n\n columns.map {|v| {table: v[\"table_name\"], column: v[\"column_name\"], type: v[\"data_type\"]}}\n end",
"def columns(sql)\n return self.dbh.execute(sql).column_names\n end",
"def column_definitions(table_name) #:nodoc:\n exec_query(<<-end_sql, 'SCHEMA')\n SELECT a.attname as column_name, format_type(a.atttypid, a.atttypmod) as column_type, d.adsrc as column_default, a.attnotnull as column_not_null\n FROM pg_attribute a LEFT JOIN pg_attrdef d\n ON a.attrelid = d.adrelid AND a.attnum = d.adnum\n WHERE a.attrelid = '#{quote_table_name(table_name)}'::regclass\n AND a.attnum > 0 AND NOT a.attisdropped\n ORDER BY a.attnum\n end_sql\n end",
"def columns\n unless @columns\n @columns = collection_info[table_name].columns.collect { |col_def|\n col = ActiveRecord::ConnectionAdapters::Column.new(col_def.name, col_def.default, col_def.sql_type, col_def.null)\n col.primary = col.name == primary_key\n col\n }\n end\n @columns\n end",
"def cols\n @cols\n end",
"def attr_columns\n @attr_columns\n end",
"def columns(table_name, name = nil)#:nodoc:\n # Each table always has a row key called 'ROW'\n columns = [\n Column.new('ROW', '')\n ]\n schema = describe_table(table_name)\n doc = REXML::Document.new(schema)\n column_families = doc.elements['Schema/AccessGroup[@name=\"default\"]'].elements.to_a\n\n @hypertable_column_names[table_name] ||= {}\n for cf in column_families\n column_name = cf.elements['Name'].text\n rubified_name = rubify_column_name(column_name)\n @hypertable_column_names[table_name][rubified_name] = column_name\n columns << new_column(rubified_name, '')\n end\n\n columns\n end",
"def column_names\n @entity_columns.map(&:name).map(&:to_s)\n end",
"def columns\n select.inject([]) do |columns, column|\n columns << {\n :type => all_columns[column][:type]\n }.merge({\n :id => column.to_s,\n :label => column_label(column)\n })\n end\n end",
"def column_definitions(table_name) #:nodoc:\n exec_query(<<-end_sql, 'SCHEMA').rows\n SELECT a.attname, format_type(a.atttypid, a.atttypmod),\n pg_get_expr(d.adbin, d.adrelid), a.attnotnull, a.atttypid, a.atttypmod\n FROM pg_attribute a LEFT JOIN pg_attrdef d\n ON a.attrelid = d.adrelid AND a.attnum = d.adnum\n WHERE a.attrelid = '#{quote_table_name(table_name)}'::regclass\n AND a.attnum > 0 AND NOT a.attisdropped\n ORDER BY a.attnum\n end_sql\n end",
"def column_names\n keys = COLUMN_NAMES\n keys.map{|key| [Person.field_name(key), key]}\n end",
"def columns(table_name, name = nil)\n # Limit, precision, and scale are all handled by the superclass.\n column_definitions(table_name).collect do |name, type, default, notnull|\n VerticaColumn.new(name, default, type, notnull == 'false')\n end\n end",
"def columns\n @columns.to_enum\n end",
"def columns\n @columns ||= begin\n column_indexes.map do |e|\n e.map do |e|\n @input[e]\n end\n end\n end\n end",
"def columns\n @columns ||= [].tap do |fields|\n @model.columns.reject { |a| a.name.match(/_id$/) }.each do |column|\n fields << resolve(column.type, column.name)\n end\n end\n end",
"def column_names\n klass.attributes\n end",
"def columns\n unless defined?(@columns) && @columns\n @columns = connection.columns(table_name, \"#{name} Columns\").select do |column| \n column.name =~ Regexp.new(\"^#{self.to_s.underscore}__\") || column.name == primary_key\n end\n @columns.each { |column| column.primary = column.name == primary_key }\n end\n @columns\n end",
"def column_definitions(table_name) # :nodoc:\n query(<<-end_sql, 'SCHEMA')\n SELECT a.attname, format_type(a.atttypid, a.atttypmod),\n pg_get_expr(d.adbin, d.adrelid), a.attnotnull, a.atttypid, a.atttypmod,\n format_encoding(a.attencodingtype::integer)\n FROM pg_attribute a LEFT JOIN pg_attrdef d\n ON a.attrelid = d.adrelid AND a.attnum = d.adnum\n WHERE a.attrelid = '#{quote_table_name(table_name)}'::regclass\n AND a.attnum > 0 AND NOT a.attisdropped\n ORDER BY a.attnum\n end_sql\n end",
"def columns(table_name, _name = nil)\n result = @connection.columns(table_name.to_s)\n\n result.each_with_object([]) do |field, cols|\n col_name = field.name\n col_sql_type = native_database_types.invert[name: field.type]\n col_nullable = (field.mode == 'NULLABLE')\n\n args = { sql_type: col_sql_type, type: col_sql_type, limit: nil }\n args[:scale] = nil\n args[:precision] = nil\n\n sql_type_metadata =\n ActiveRecord::ConnectionAdapters::SqlTypeMetadata.new(**args)\n\n cols << ActiveRecord::ConnectionAdapters::Column.new(\n col_name,\n nil,\n sql_type_metadata,\n col_nullable,\n table_name\n )\n end\n end",
"def column_definitions(table_name) # :nodoc:\n local_condition = 'AND a.attislocal IS TRUE' if @_dump_mode\n query(<<-SQL, 'SCHEMA')\n SELECT a.attname, format_type(a.atttypid, a.atttypmod),\n pg_get_expr(d.adbin, d.adrelid), a.attnotnull, a.atttypid, a.atttypmod,\n (SELECT c.collname FROM pg_collation c, pg_type t\n WHERE c.oid = a.attcollation AND t.oid = a.atttypid AND a.attcollation <> t.typcollation),\n col_description(a.attrelid, a.attnum) AS comment\n FROM pg_attribute a\n LEFT JOIN pg_attrdef d ON a.attrelid = d.adrelid AND a.attnum = d.adnum\n WHERE a.attrelid = '#{quote_table_name(table_name)}'::regclass\n AND a.attnum > 0\n AND a.attisdropped IS FALSE\n #{local_condition}\n ORDER BY a.attnum\n SQL\n end",
"def column_names\n klass.new.attributes.keys\n end",
"def columns; self.class.columns; end",
"def column_names\n @column_names = columns_hash.keys\n end",
"def columns(table_name, name = nil)\n type_part = @sql_layer_version >= 10903 ?\n \" COLUMN_TYPE_STRING(table_schema, table_name, column_name), \" :\n \" data_type||COALESCE('('||character_maximum_length||')', '('||numeric_precision||','||numeric_scale||')', ''), \"\n select_rows(\n \"SELECT column_name, \"+\n \" column_default, \"+\n type_part +\n \" is_nullable \"+\n \"FROM information_schema.columns \"+\n \"WHERE table_schema = CURRENT_SCHEMA \"+\n \" AND table_name = '#{quote_string(table_name.to_s)}' \"+\n \"ORDER BY ordinal_position\",\n name || SCHEMA_LOG_NAME\n ).map { |row|\n # Base Column depends on lower and no space (e.g. DECIMAL(10, 0) => decimal(10,0))\n type_str = row[2].gsub(' ', '').downcase\n FdbSqlColumn.new(row[0], row[1], type_str, row[3] == 'YES')\n }\n end",
"def query_columns\n explicit_columns\n end",
"def column_names\n @column_names || owner.column_names\n end",
"def columns\n @columns = @clazz.columns_hash.map do |name, column|\n [name, @columns[name] || Column.new(column, @dump)]\n end.to_h.with_indifferent_access\n end"
] |
[
"0.8558935",
"0.80598795",
"0.7951822",
"0.7878924",
"0.7862773",
"0.78619456",
"0.78464586",
"0.7842156",
"0.78395635",
"0.78395635",
"0.7834244",
"0.7816846",
"0.7771062",
"0.7770335",
"0.7751243",
"0.77316546",
"0.7706285",
"0.7704942",
"0.7699515",
"0.76968247",
"0.76959527",
"0.7689465",
"0.76713926",
"0.7670336",
"0.7670178",
"0.7669566",
"0.7649246",
"0.76311463",
"0.7622673",
"0.7622673",
"0.7614399",
"0.76104975",
"0.7608065",
"0.7587725",
"0.7572523",
"0.75140697",
"0.75030136",
"0.7492773",
"0.7476627",
"0.74766093",
"0.7458561",
"0.7451839",
"0.74457115",
"0.74347675",
"0.7421496",
"0.7413731",
"0.7413731",
"0.73925394",
"0.73856413",
"0.73834264",
"0.73770374",
"0.735634",
"0.7354736",
"0.73527455",
"0.73527455",
"0.73514897",
"0.7346251",
"0.73277897",
"0.730467",
"0.73032534",
"0.7299421",
"0.7294942",
"0.72839904",
"0.7278857",
"0.7273209",
"0.7253279",
"0.7246749",
"0.72385687",
"0.72385687",
"0.7227542",
"0.7224112",
"0.72208595",
"0.7218777",
"0.71997637",
"0.71977013",
"0.7195602",
"0.7194149",
"0.7189271",
"0.71842194",
"0.7182767",
"0.71742666",
"0.71723205",
"0.71712077",
"0.7169647",
"0.7159833",
"0.7157946",
"0.7153332",
"0.7151565",
"0.7149195",
"0.71187526",
"0.7116663",
"0.71130913",
"0.7107479",
"0.7099016",
"0.7097812",
"0.7095434",
"0.7095186",
"0.70899767",
"0.70898217",
"0.7084183"
] |
0.8322581
|
1
|
GET /married_statuses GET /married_statuses.json
|
def index
@married_statuses = MarriedStatus.all
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def statuses\n Sifter.\n get(\"/api/statuses\").\n parsed_response[\"statuses\"]\n end",
"def set_married_status\n @married_status = MarriedStatus.find(params[:id])\n end",
"def show_status(status_id)\n get \"statuses/show/#{status_id}\"\n end",
"def update\n respond_to do |format|\n if @married_status.update(married_status_params)\n format.html { redirect_to @married_status, notice: 'Married status was successfully updated.' }\n format.json { render :show, status: :ok, location: @married_status }\n else\n format.html { render :edit }\n format.json { render json: @married_status.errors, status: :unprocessable_entity }\n end\n end\n end",
"def marital_status\n fetch('demographic.marital_status')\n end",
"def married_status_params\n params.require(:married_status).permit(:name)\n end",
"def list_statuses(user, list)\n get(\"/#{user}/lists/#{list}/statuses.json\")\n end",
"def index\n @mstatuses = Mstatus.all\n end",
"def create\n @married_status = MarriedStatus.new(married_status_params)\n\n respond_to do |format|\n if @married_status.save\n format.html { redirect_to @married_status, notice: 'Married status was successfully created.' }\n format.json { render :show, status: :created, location: @married_status }\n else\n format.html { render :new }\n format.json { render json: @married_status.errors, status: :unprocessable_entity }\n end\n end\n end",
"def index\n @search = Status.search(params[:search], :order => \"id DESC\")\n @statuses = @search.paginate(:page => params[:page], :order => \"id DESC\")\n\n @statuses = @statuses.where(:uid => params[:uid]) unless params[:uid].nil?\n @statuses = @statuses.where(:screen_name => params[:screen_name]) unless params[:screen_name].nil?\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @statuses }\n end\n end",
"def get\n\t\t\t result = Status.find_by(windmillid: params[:windmillid]) \n \t\t\trender json: [result.as_json(only: [:status,:power,:gen,:frequency,:rotor,:wind,:pitch])]\n\tend",
"def mentions(params = {})\n get \"statuses/mentions\", params\n end",
"def index\n @statuses = Status.all\n end",
"def index\n @statuses = Status.all\n end",
"def index\n @search = MaritalStatus.search(params[:search])\n @marital_statuses = @search.page(params[:page]).per(10)\n\n respond_to do |format|\n format.html # index.html.erbb\n format.xml { render :xml => @marital_statuses }\n end\n end",
"def list_timeline(list_owner_username, slug, query = {})\n perform_get(\"/#{list_owner_username}/lists/#{slug}/statuses.json\", :query => query)\nend",
"def destroy\n @married_status.destroy\n respond_to do |format|\n format.html { redirect_to married_statuses_url, notice: 'Married status was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def show\n @roof = Roof.find(params[:roof_id])\n @status = @roof.statuses.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @status }\n end\n end",
"def current_statuses_three\n @current_statuses = CurrentStatus.where(status:true, id: [1,2,3]).order('name ASC')\n respond_to do |format|\n format.html # index.html.erb\n format.json { render :json => @current_statuses}\n end\n end",
"def statuses\n\t\trender :json => {:status => 1, :categories => {\"1\" => \"Available\", \"2\" => \"On Hold\", \"3\" => \"Borrowed\", \"4\" => \"Unavailable\"}}, :status => 200\n\t\treturn\n\tend",
"def show(project_token = @project_token, id = @id, user = @@default_user)\n @attributes = send_request(\"statuses/#{id}\", :get) do |req|\n req.params = {\n token: project_token,\n auth_token: user.auth_token\n }\n end\n end",
"def marital_status; end",
"def show_user_tweets\n @user_tweets = TwitterClient.user_timeline(params[:name])\n render json: @user_tweets\n end",
"def tweets(opts={})\n params = {\n :screen_name => NAME,\n :trim_user => true,\n :include_entities => true\n }.merge(opts)\n get(\"/statuses/user_timeline.json\",params)\n end",
"def status(status_id)\n response = @client.get(\"/api/v1/statuses/#{status_id}\")\n return JSON.parse(response.body)\n end",
"def statuses( params={} )\n statuses = get_connections(\"statuses\", params)\n return map_connections statuses, :to => Facebook::Graph::Statuse\n end",
"def index\n @statuses = Status.search.page params[:page]\n end",
"def index\n @swit_statuses = SwitStatus.all\n end",
"def statuses\n request(:get, \"applicant_tracking/statuses\")\n end",
"def tweets\n user = User.find(params[:id])\n render json: user.list_tweets, status: :ok\n end",
"def update\n respond_to do |format|\n if @current_statuses.update(current_statuses_params)\n format.html { redirect_to @current_statuses, notice: 'Current Statuses was successfully updated.' }\n format.json { render :show, status: :ok, location: @current_statuses }\n else\n format.html { render :edit }\n format.json { render json: @current_statuses.errors, status: :unprocessable_entity }\n end\n end\n end",
"def friends_timeline(params = {})\n get \"statuses/friends_timeline\", params\n end",
"def index\n @legalstatuses = Legalstatus.all\n end",
"def show\n user = user_from_token\n @tweet = user.tweets.where(:id => params[:id])\n render :json => @tweet\n end",
"def show\n\n respond_to do |format|\n format.html # show.html.erb\n format.xml { render :xml => @marital_status }\n end\n end",
"def statuses\n api.get('status')\n end",
"def statuses( params={} )\n statuses = get_connections(\"statuses\", params)\n return map_connections statuses, :to => Facebook::Graph::Generic\n end",
"def statuses; end",
"def list_timeline(list_owner_screen_name, slug, options={})\n perform_get(\"#{list_owner_screen_name}/lists/#{slug}/statuses.#{Twitter.format}\", options)\n end",
"def index\n @barrel_statuses = BarrelStatus.all\n end",
"def index\n @statuses = Status.find(:all)\n\n respond_to do |format|\n format.html # index.html.erb\n format.xml { render :xml => @statuses }\n end\n end",
"def get_redemption_request_statuses\r\n\r\n # prepare query url\r\n _query_builder = Configuration.base_uri.dup\r\n _query_builder << '/utilities/redemption-request-statuses'\r\n _query_url = APIHelper.clean_url _query_builder\r\n\r\n # prepare headers\r\n _headers = {\r\n 'accept' => 'application/json',\r\n 'Content-Type' => Configuration.content_type\r\n }\r\n\r\n # prepare and execute HttpRequest\r\n _request = @http_client.get _query_url, headers: _headers\r\n CustomHeaderAuth.apply(_request)\r\n _context = execute_request(_request)\r\n validate_response(_context)\r\n\r\n # return appropriate response type\r\n decoded = APIHelper.json_deserialize(_context.response.raw_body) if not (_context.response.raw_body.nil? or _context.response.raw_body.to_s.strip.empty?)\r\n return decoded\r\n end",
"def index\n @application_statuses = ApplicationStatus.all\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @application_statuses }\n end\n end",
"def parseResults( resp )\n data = JSON.parse(resp.body)\n\treturn data['statuses']\nend",
"def status\n sanitized = whitelist(params, :status)\n\n @travel = Travels::Travel.find(sanitized[:id])\n\n respond_to do |format|\n format.html # status.html.erb\n format.json { render json: { status: @travel.state.to_s } }\n end\n end",
"def index\n @request_statuses = RequestStatus.all\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @request_statuses }\n end\n end",
"def index\n @following_user_ids = @user.following.pluck(:id)\n\n @tweets = Tweet.where(user_id: @following_user_ids).order(created_at: :desc).includes(:user)\n json_response(@tweets)\n end",
"def show\n @mood = Mood.find(params[:id])\n respond_to do |format|\n format.any { render :json => @mood }\n end\n end",
"def index\n @reward_and_levels = RewardAndLevel.all\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @reward_and_levels }\n end\n end",
"def index\n @nvs_mig_statuses = NvsMigStatus.all\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @nvs_mig_statuses }\n end\n end",
"def get_twitter_status\n logger.info 'Getting twitter'\n begin\n c = Grackle::Client.new\n twitter = c.statuses.user_timeline?(:screen_name => 'paulcarlile', :count => 2)\n rescue Grackle::TwitterError\n twitter = Grackle::TwitterError\n end\n end",
"def index\n @tweets = @user.tweets.all\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @tweets }\n end\n end",
"def create_statuses\n end",
"def create_statuses\n end",
"def index\n @goal_rewards = user.goal_rewards\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @goal_rewards }\n end\n end",
"def url\n \"http://twitter.com/#{self.username}/statuses/#{self.twitter_id}\"\n end",
"def url\n \"http://twitter.com/#{attribute_get(:username)}/statuses/#{attribute_get(:id)}\"\n end",
"def print_timeline(response)\n \n tweets = response[\"statuses\"]\n \n for tweet in tweets\n puts tweet[\"user\"][\"screen_name\"]\n puts tweet[\"text\"]\n end\n \nend",
"def index\n timeline = Tweet.where(user_id: User.find(@current_user_id).follows).or(Tweet.where(user_id: @current_user_id))\n render json: {timeline: timeline}\n end",
"def list\r\n #page = params[:page].nil? ? 0 : params[:page]\r\n tweets = Tweet.all\r\n render json: tweets\r\n end",
"def update!\n response = Tessellator::Fetcher.new.call('get', 'https://howamidoing-duckinator.herokuapp.com/status.json')\n @@statuses = JSON.parse(response.body)['statuses']\n end",
"def show\n respond_to do |format|\n begin\n @tweet = TweetApp::ClientContext.status(:get, params[:id])\n format.html # show.rhtml\n format.json { render :json => @tweet.to_json }\n format.xml { render :xml => @tweet.to_xml }\n rescue Twitter::RESTError => re\n handle_rest_error(re, format)\n end\n end\n end",
"def show\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @reward_and_level }\n end\n end",
"def get_tweets(list)\n if list.authorized?(@user)\n list.tweets\n else\n []\n end\nend",
"def get_status(m, params)\n friends = params[:friends]\n show_reply = @bot.config['twitter.status_show_replies']\n if m.message.match(/^twitter status reply/)\n show_reply = true\n elsif m.message.match(/^twitter status noreply/)\n show_reply = false\n end\n\n if @registry.has_key?(m.sourcenick + \"_access_token\")\n @access_token = YAML::load(@registry[m.sourcenick + \"_access_token\"])\n nick = params[:nick] || @access_token.params[:screen_name]\n else\n if friends\n if @has_oauth\n m.reply \"You are not authorized with Twitter. Please use 'twitter authorize' first to use this feature.\"\n else\n report_oauth_missing(m, \"I cannot retrieve your friends status\")\n end\n return false\n end\n nick = params[:nick]\n end\n\n if not nick\n m.reply \"you should specify the username of the twitter to use, or identify using 'twitter authorize'\"\n return false\n end\n\n count = friends ? @bot.config['twitter.friends_status_count'] : @bot.config['twitter.status_count']\n if show_reply\n request_count = count\n else \n request_count = [ count*5,50 ].max\n end\n user = URI.escape(nick)\n # receive the public timeline per default (this works even without an access_token)\n uri = \"https://api.twitter.com/1/statuses/user_timeline.xml?screen_name=#{user}&count=#{request_count}&include_rts=true\"\n if @has_oauth and @registry.has_key?(m.sourcenick + \"_access_token\")\n if friends\n #no change to count variable\n uri = \"https://api.twitter.com/1/statuses/friends_timeline.xml?count=#{request_count}&include_rts=true\"\n end\n response = @access_token.get(uri).body\n else\n response = @bot.httputil.get(uri, :cache => false)\n end\n debug response\n\n texts = []\n\n if response\n begin\n rex = REXML::Document.new(response)\n rex.root.elements.each(\"status\") { |st|\n # month, day, hour, min, sec, year = st.elements['created_at'].text.match(/\\w+ (\\w+) (\\d+) (\\d+):(\\d+):(\\d+) \\S+ (\\d+)/)[1..6]\n # debug [year, month, day, hour, min, sec].inspect\n # time = Time.local(year.to_i, month, day.to_i, hour.to_i, min.to_i, sec.to_i)\n time = Time.parse(st.elements['created_at'].text)\n now = Time.now\n # Sometimes, time can be in the future; invert the relation in this case\n delta = ((time > now) ? time - now : now - time)\n msg = st.elements['text'].to_s + \" (#{Utils.secs_to_string(delta.to_i)} ago via #{st.elements['source'].to_s})\"\n author = \"\"\n if friends\n author = Utils.decode_html_entities(st.elements['user'].elements['name'].text) + \": \" rescue \"\"\n end\n texts << author+Utils.decode_html_entities(msg).ircify_html\n }\n if friends\n # friends always return the latest 20 updates, so we clip the count\n texts[count..-1]=nil\n end\n rescue\n error $!\n if friends\n m.reply \"could not parse status for #{nick}'s friends\"\n else\n m.reply \"could not parse status for #{nick}\"\n end\n return false\n end\n if !show_reply\n nonreplytexts = texts.grep(/^[^@]/)\n if nonreplytexts.length > 0\n texts = nonreplytexts\n end\n end\n # Make sure we have the right number\n texts[count..-1]=nil\n if texts.empty?\n m.reply \"No status updates!\"\n else\n m.reply texts.reverse.join(\"\\n\")\n end\n return true\n else\n if friends\n rep = \"could not get status for #{nick}'s friends\"\n rep << \", try asking in private\" unless m.private?\n else\n rep = \"could not get status for #{nick}\"\n end\n m.reply rep\n return false\n end\n end",
"def show\n @watcher = Watcher.includes(:tweets).friendly.find(params[:id])\n @statuses = Status.all\n respond_to do |format|\n format.html { render :show, offset: params[:offset]}\n format.json { render json: @watcher.tweets }\n end\n end",
"def index\n @tweets = Tweet.all\n\n render json: @tweets\n end",
"def parse_retweeted_status(rs)\n if rs.nil?\n nil\n else\n rs = { \n :created_at => rs.created_at,\n :id => rs.id,\n :text => rs.text,\n :source => rs.source, \n :truncated => rs[\"truncated\"],\n :in_reply_to_status_id => rs[\"in_reply_to_status_id\"],\n :in_reply_to_user_id => rs[\"in_reply_to_user_id\"],\n :in_reply_to_screen_name => rs[\"in_reply_to_screen_name\"],\n :user_id => rs[\"user\"][\"id\"] \n }\n rs\n end\n end",
"def update\n\n respond_to do |format|\n if @marital_status.update_attributes(params[:marital_status])\n format.html { redirect_to(@marital_status, :notice => 'Marital status was successfully updated.') }\n format.xml { head :ok }\n else\n format.html { render :action => \"edit\" }\n format.xml { render :xml => @marital_status.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def index\n @member_statuses = MemberStatus.all\n end",
"def show\n @annual_stat = AnnualStat.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @annual_stat }\n end\n end",
"def meals\n get(\"/user/#{@user_id}/meals.json\")\n end",
"def index\n @rh21_statuses = Rh21Status.all\n end",
"def index\n \tif params[:for_timeline]\n\t\t\t@bills = Bill.where(:initiator_id.in => params[:politicians]).where(:complete =>true).limit(10)\n\t\telse\n\t \t@bills = Bill.all.limit(10)\n \tend\n\n respond_to do |format|\n format.json { render json: @bills.to_json(:include => {:initiator => {:only =>[:name]}}) }\n end\n\n end",
"def index\n @idea_statuses = IdeaStatus.all\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @idea_statuses }\n end\n end",
"def get_timeline\n HTTParty.post(\"#{@api_path}/tweets/#{@handle}/#{@password}\")\n end",
"def show\n @twitter_list = TwitterList.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @twitter_list }\n end\n end",
"def show\r\n tweet = Tweet.find(params[:id])\r\n render json: tweet\r\n end",
"def twubbles\n @sad_tweets = WordTweet.recent_sad_tweets\n respond_to do |f|\n f.json { render :json => twubbles_json(@sad_tweets) }\n f.html\n end\n end",
"def index\n @registration_statuses = RegistrationStatus.all\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render :json => @registration_statuses }\n end\n end",
"def ineligible_statuses\n ['1 Day Loan',\n '1 Week No Renew',\n '1 Week No Renew Equip',\n '2 Day Loan',\n '24 Hour Loan',\n '3 Day Loan',\n 'Audio Recorder',\n 'Due at Closing',\n 'Journal Loan',\n 'Music CD/DVD',\n 'Pass',\n 'See Note Above',\n 'Two Week Loan']\n end",
"def statuses=(statuses)\n assert_unloaded\n @statuses = Array(statuses)\n end",
"def build_statuses(options = {})\n build_statuses_resource(options)\n end",
"def user_tweets(user, count=10, since_id=nil, max_id=nil)\n print \"Getting Last %d Statuses for User %s\" % [count, user.to_s]\n print \" since %s\" % since_id if since_id\n print \" until %s\" % max_id if max_id\n print \"\\n\"\n options = {:count => count, :trim_user => true, :include_rts => true, :include_entities => true}\n options[:since_id] = since_id if since_id\n options[:max_id] = max_id if max_id\n begin\n statuses = @MT.user_timeline(user, options)\n if statuses.size > 0\n status_data = statuses.map do |s|\n {\n :user_id => s.user.id,\n :created_at => s.created_at,\n :id => s.id,\n :text => s.text,\n :source => s.source,\n :truncated => s[\"truncated\"],\n :in_reply_to_user_id => s[\"in_reply_to_user_id\"],\n :in_reply_to_screen_name => s[\"in_reply_to_screen_name\"],\n :geo => s[\"geo\"],\n :coordinates => s[\"coordinates\"],\n :place => parse_place(s[\"place\"]),\n :contributors => s[\"contributors\"],\n :retweet_count => s[\"retweet_count\"],\n :entities => parse_entities(s.attrs[\"entities\"]),\n :retweeted_status => parse_retweeted_status(s[\"retweeted_status\"])\n }\n end\n status_data\n else\n []\n end\n rescue Twitter::Error::Unauthorized, Twitter::Error::Forbidden\n puts \"Failed for %s (Protected)\" % user.to_s\n []\n end\n end",
"def index\n params[:per_page] = perpage if params[:per_page].nil?\n params[:page] = 1 if params[:page].blank?\n conditions = {:status_type => 0 ,:relation_id => 0 ,:count_type => 0 }\n conditions[:status_type] = params[:status_type] unless params[:status_type].blank?\n conditions[:relation_id] = params[:relation_id] unless params[:relation_id].blank?\n conditions[:count_type] = params[:count_type] unless params[:count_type].blank?\n @kf_statuses = Kf::Status.order(\"date DESC\").where(conditions).paginate(:page => params[:page], :per_page => params[:per_page])\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @kf_statuses }\n end\n end",
"def show\n @twitter_user = TwitterUser.find(params[:id])\n @follower_ids = Twitter.follower_ids(@twitter_user.twitter_username).collection\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @twitter_user }\n end\n end",
"def show\n @received_military_award = ReceivedMilitaryAward.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @received_military_award }\n end\n end",
"def realm_status(options = {})\n options = options.dup\n perform_with_objects(:get, '/wow/realm/status', options, Armory::Realm, :realms)\n end",
"def show\n @status = Status.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @status }\n end\n end",
"def show\n @status = Status.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @status }\n end\n end",
"def index\n @lesson_statuses = LessonStatus.all\n end",
"def statusesReceived_forRequest(statuses,identifier)\n\t\tputs \"statusesReceived_forRequest\"\n\t\tself.statuses = statuses\n\tend",
"def index\n\t\tuser = User.find_by(id: params[:user_id])\n\t\tif user.present?\n\t\t\tfollower_ids = user.followers.pluck(:id)\n\t\t\ttweets = Tweet.where(\"user_id IN (?)\", follower_ids).order(\"updated_at DESC\")\n\t\t\trender json: {:status=>\"success\", :code=>200, :message=>\"List of tweets from the users you follow.\", data: tweets}\n\t\telse\n\t\t\trender json: {:status=>\"failure\", :message=>\"User is not present.\", data: tweets}\n\t\tend\n\tend",
"def index\n @reqstatuses = Reqstatus.all\n end",
"def get_statuses_with_http_info(tid, opts = {})\n if @api_client.config.debugging\n @api_client.config.logger.debug \"Calling API: DevicesManagementApi.get_statuses ...\"\n end\n # verify the required parameter 'tid' is set\n fail ArgumentError, \"Missing the required parameter 'tid' when calling DevicesManagementApi.get_statuses\" if tid.nil?\n # resource path\n local_var_path = \"/devicemgmt/tasks/{tid}/statuses\".sub('{format}','json').sub('{' + 'tid' + '}', tid.to_s)\n\n # query parameters\n query_params = {}\n query_params[:'count'] = opts[:'count'] if !opts[:'count'].nil?\n query_params[:'offset'] = opts[:'offset'] if !opts[:'offset'].nil?\n query_params[:'status'] = opts[:'status'] if !opts[:'status'].nil?\n query_params[:'dids'] = opts[:'dids'] if !opts[:'dids'].nil?\n\n # header parameters\n header_params = {}\n # HTTP header 'Accept' (if needed)\n header_params['Accept'] = @api_client.select_header_accept(['application/json'])\n\n # form parameters\n form_params = {}\n\n # http body (model)\n post_body = nil\n auth_names = ['artikcloud_oauth']\n data, status_code, headers = @api_client.call_api(:GET, local_var_path,\n :header_params => header_params,\n :query_params => query_params,\n :form_params => form_params,\n :body => post_body,\n :auth_names => auth_names,\n :return_type => 'TaskStatusesEnvelope')\n if @api_client.config.debugging\n @api_client.config.logger.debug \"API called: DevicesManagementApi#get_statuses\\nData: #{data.inspect}\\nStatus code: #{status_code}\\nHeaders: #{headers}\"\n end\n return data, status_code, headers\n end",
"def get_tweets(list)\n if list.authorized?(@user)\n list.tweets\n else\n [] # \"magic\" return value\n end\nend",
"def index\n authenticate_user! if params[:page]\n @statuses = Status.all.paginate(:page => params[:page], :per_page => 10).order(\"created_at DESC\")\n end",
"def index\n\t\t@leads = Lead.all\n\t\t\n\t\t\n\n\t\tlogger.info params[:status]\n\t\trespond_to do |format|\n\t\t format.html # index.html.erb\n\t\t format.json { render json: @leads }\n\t\tend\n\tend",
"def retweeters_of(id, options={})\n ids_only = !!(options.delete(:ids_only))\n perform_get(\"statuses/#{id}/retweeted_by#{\"/ids\" if ids_only}.#{Twitter.format}\", options)\n end",
"def validate_marital_status(val)\n unless @valid_marital_statuses.include?(val)\n raise \"Invalid value: #{val}\"\n end\n end"
] |
[
"0.64134794",
"0.63754165",
"0.59984237",
"0.595277",
"0.59112424",
"0.5827594",
"0.5805667",
"0.5798412",
"0.57732296",
"0.56787175",
"0.5674036",
"0.56538707",
"0.56222755",
"0.56222755",
"0.56143117",
"0.55412096",
"0.55409783",
"0.55313104",
"0.5514922",
"0.5490368",
"0.548507",
"0.5480444",
"0.542057",
"0.5409821",
"0.53926593",
"0.53912646",
"0.53907126",
"0.5372074",
"0.5365035",
"0.53596205",
"0.5346489",
"0.53457594",
"0.5317123",
"0.5316048",
"0.53089875",
"0.525744",
"0.52558446",
"0.5242158",
"0.52332866",
"0.52242136",
"0.5221123",
"0.52008873",
"0.51908594",
"0.51870525",
"0.516806",
"0.5161209",
"0.5158458",
"0.51538765",
"0.515246",
"0.51466626",
"0.51294047",
"0.51285964",
"0.51215065",
"0.51215065",
"0.51203036",
"0.51142776",
"0.51117295",
"0.51099265",
"0.51093197",
"0.5106341",
"0.5094474",
"0.50694555",
"0.5064608",
"0.50629216",
"0.50553876",
"0.5042387",
"0.50399363",
"0.5039582",
"0.5038437",
"0.5026367",
"0.5025776",
"0.5005801",
"0.50052947",
"0.49839824",
"0.4981292",
"0.498001",
"0.4970221",
"0.496911",
"0.49689204",
"0.49687597",
"0.49596876",
"0.49573126",
"0.4955752",
"0.4955637",
"0.4952649",
"0.4950409",
"0.49498025",
"0.49486607",
"0.49386036",
"0.49386036",
"0.49368763",
"0.493587",
"0.4931822",
"0.49275637",
"0.49263886",
"0.49235284",
"0.4918178",
"0.49169165",
"0.4915422",
"0.4914198"
] |
0.7358902
|
0
|
GET /married_statuses/1 GET /married_statuses/1.json
|
def show
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def index\n @married_statuses = MarriedStatus.all\n end",
"def show_status(status_id)\n get \"statuses/show/#{status_id}\"\n end",
"def statuses\n Sifter.\n get(\"/api/statuses\").\n parsed_response[\"statuses\"]\n end",
"def set_married_status\n @married_status = MarriedStatus.find(params[:id])\n end",
"def show\n @roof = Roof.find(params[:roof_id])\n @status = @roof.statuses.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @status }\n end\n end",
"def mentions(params = {})\n get \"statuses/mentions\", params\n end",
"def get\n\t\t\t result = Status.find_by(windmillid: params[:windmillid]) \n \t\t\trender json: [result.as_json(only: [:status,:power,:gen,:frequency,:rotor,:wind,:pitch])]\n\tend",
"def show(project_token = @project_token, id = @id, user = @@default_user)\n @attributes = send_request(\"statuses/#{id}\", :get) do |req|\n req.params = {\n token: project_token,\n auth_token: user.auth_token\n }\n end\n end",
"def status(status_id)\n response = @client.get(\"/api/v1/statuses/#{status_id}\")\n return JSON.parse(response.body)\n end",
"def index\n @mstatuses = Mstatus.all\n end",
"def show\n respond_to do |format|\n begin\n @tweet = TweetApp::ClientContext.status(:get, params[:id])\n format.html # show.rhtml\n format.json { render :json => @tweet.to_json }\n format.xml { render :xml => @tweet.to_xml }\n rescue Twitter::RESTError => re\n handle_rest_error(re, format)\n end\n end\n end",
"def update\n respond_to do |format|\n if @married_status.update(married_status_params)\n format.html { redirect_to @married_status, notice: 'Married status was successfully updated.' }\n format.json { render :show, status: :ok, location: @married_status }\n else\n format.html { render :edit }\n format.json { render json: @married_status.errors, status: :unprocessable_entity }\n end\n end\n end",
"def list_statuses(user, list)\n get(\"/#{user}/lists/#{list}/statuses.json\")\n end",
"def show\n user = user_from_token\n @tweet = user.tweets.where(:id => params[:id])\n render :json => @tweet\n end",
"def show_user_tweets\n @user_tweets = TwitterClient.user_timeline(params[:name])\n render json: @user_tweets\n end",
"def create\n @married_status = MarriedStatus.new(married_status_params)\n\n respond_to do |format|\n if @married_status.save\n format.html { redirect_to @married_status, notice: 'Married status was successfully created.' }\n format.json { render :show, status: :created, location: @married_status }\n else\n format.html { render :new }\n format.json { render json: @married_status.errors, status: :unprocessable_entity }\n end\n end\n end",
"def index\n @search = Status.search(params[:search], :order => \"id DESC\")\n @statuses = @search.paginate(:page => params[:page], :order => \"id DESC\")\n\n @statuses = @statuses.where(:uid => params[:uid]) unless params[:uid].nil?\n @statuses = @statuses.where(:screen_name => params[:screen_name]) unless params[:screen_name].nil?\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @statuses }\n end\n end",
"def marital_status\n fetch('demographic.marital_status')\n end",
"def index\n @statuses = Status.all\n end",
"def index\n @statuses = Status.all\n end",
"def married_status_params\n params.require(:married_status).permit(:name)\n end",
"def url\n \"http://twitter.com/#{self.username}/statuses/#{self.twitter_id}\"\n end",
"def show\n @mood = Mood.find(params[:id])\n respond_to do |format|\n format.any { render :json => @mood }\n end\n end",
"def show\r\n tweet = Tweet.find(params[:id])\r\n render json: tweet\r\n end",
"def index\n @search = MaritalStatus.search(params[:search])\n @marital_statuses = @search.page(params[:page]).per(10)\n\n respond_to do |format|\n format.html # index.html.erbb\n format.xml { render :xml => @marital_statuses }\n end\n end",
"def tweets(opts={})\n params = {\n :screen_name => NAME,\n :trim_user => true,\n :include_entities => true\n }.merge(opts)\n get(\"/statuses/user_timeline.json\",params)\n end",
"def url\n \"http://twitter.com/#{attribute_get(:username)}/statuses/#{attribute_get(:id)}\"\n end",
"def destroy\n @married_status.destroy\n respond_to do |format|\n format.html { redirect_to married_statuses_url, notice: 'Married status was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def current_statuses_three\n @current_statuses = CurrentStatus.where(status:true, id: [1,2,3]).order('name ASC')\n respond_to do |format|\n format.html # index.html.erb\n format.json { render :json => @current_statuses}\n end\n end",
"def tweets\n user = User.find(params[:id])\n render json: user.list_tweets, status: :ok\n end",
"def get_tweet_by_id(id)\n tweet = Twitter.status(id)\n {:message => tweet.text, :author => tweet.user.screen_name, :date => tweet.created_at} \n end",
"def friends_timeline(params = {})\n get \"statuses/friends_timeline\", params\n end",
"def get_twitter_status\n logger.info 'Getting twitter'\n begin\n c = Grackle::Client.new\n twitter = c.statuses.user_timeline?(:screen_name => 'paulcarlile', :count => 2)\n rescue Grackle::TwitterError\n twitter = Grackle::TwitterError\n end\n end",
"def show\n @status = Status.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @status }\n end\n end",
"def show\n @status = Status.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @status }\n end\n end",
"def print_timeline(response)\n \n tweets = response[\"statuses\"]\n \n for tweet in tweets\n puts tweet[\"user\"][\"screen_name\"]\n puts tweet[\"text\"]\n end\n \nend",
"def list_timeline(list_owner_username, slug, query = {})\n perform_get(\"/#{list_owner_username}/lists/#{slug}/statuses.json\", :query => query)\nend",
"def parseResults( resp )\n data = JSON.parse(resp.body)\n\treturn data['statuses']\nend",
"def show\n @twitter_id = TwitterId.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @twitter_id }\n end\n end",
"def show\n @watcher = Watcher.includes(:tweets).friendly.find(params[:id])\n @statuses = Status.all\n respond_to do |format|\n format.html { render :show, offset: params[:offset]}\n format.json { render json: @watcher.tweets }\n end\n end",
"def show\n\n respond_to do |format|\n format.html # show.html.erb\n format.xml { render :xml => @marital_status }\n end\n end",
"def show\n @retweet = Retweet.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @retweet }\n end\n end",
"def show\n @twitter_list = TwitterList.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @twitter_list }\n end\n end",
"def retweets_of_me(options = {})\n @req.get(\"/1.1/statuses/retweets_of_me.json\", options)\n end",
"def parse_retweeted_status(rs)\n if rs.nil?\n nil\n else\n rs = { \n :created_at => rs.created_at,\n :id => rs.id,\n :text => rs.text,\n :source => rs.source, \n :truncated => rs[\"truncated\"],\n :in_reply_to_status_id => rs[\"in_reply_to_status_id\"],\n :in_reply_to_user_id => rs[\"in_reply_to_user_id\"],\n :in_reply_to_screen_name => rs[\"in_reply_to_screen_name\"],\n :user_id => rs[\"user\"][\"id\"] \n }\n rs\n end\n end",
"def index\n @statuses = Status.search.page params[:page]\n end",
"def update\n respond_to do |format|\n if @current_statuses.update(current_statuses_params)\n format.html { redirect_to @current_statuses, notice: 'Current Statuses was successfully updated.' }\n format.json { render :show, status: :ok, location: @current_statuses }\n else\n format.html { render :edit }\n format.json { render json: @current_statuses.errors, status: :unprocessable_entity }\n end\n end\n end",
"def index\n @statuses = Status.find(:all)\n\n respond_to do |format|\n format.html # index.html.erb\n format.xml { render :xml => @statuses }\n end\n end",
"def update!\n response = Tessellator::Fetcher.new.call('get', 'https://howamidoing-duckinator.herokuapp.com/status.json')\n @@statuses = JSON.parse(response.body)['statuses']\n end",
"def index\n @tweets = @user.tweets.all\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @tweets }\n end\n end",
"def new\n @roof = Roof.find(params[:roof_id])\n @status = @roof.statuses.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @status }\n end\n end",
"def show\n @tweet = Tweet.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @tweet }\n end\n end",
"def show\n @tweet = Tweet.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @tweet }\n end\n end",
"def get_status(m, params)\n friends = params[:friends]\n show_reply = @bot.config['twitter.status_show_replies']\n if m.message.match(/^twitter status reply/)\n show_reply = true\n elsif m.message.match(/^twitter status noreply/)\n show_reply = false\n end\n\n if @registry.has_key?(m.sourcenick + \"_access_token\")\n @access_token = YAML::load(@registry[m.sourcenick + \"_access_token\"])\n nick = params[:nick] || @access_token.params[:screen_name]\n else\n if friends\n if @has_oauth\n m.reply \"You are not authorized with Twitter. Please use 'twitter authorize' first to use this feature.\"\n else\n report_oauth_missing(m, \"I cannot retrieve your friends status\")\n end\n return false\n end\n nick = params[:nick]\n end\n\n if not nick\n m.reply \"you should specify the username of the twitter to use, or identify using 'twitter authorize'\"\n return false\n end\n\n count = friends ? @bot.config['twitter.friends_status_count'] : @bot.config['twitter.status_count']\n if show_reply\n request_count = count\n else \n request_count = [ count*5,50 ].max\n end\n user = URI.escape(nick)\n # receive the public timeline per default (this works even without an access_token)\n uri = \"https://api.twitter.com/1/statuses/user_timeline.xml?screen_name=#{user}&count=#{request_count}&include_rts=true\"\n if @has_oauth and @registry.has_key?(m.sourcenick + \"_access_token\")\n if friends\n #no change to count variable\n uri = \"https://api.twitter.com/1/statuses/friends_timeline.xml?count=#{request_count}&include_rts=true\"\n end\n response = @access_token.get(uri).body\n else\n response = @bot.httputil.get(uri, :cache => false)\n end\n debug response\n\n texts = []\n\n if response\n begin\n rex = REXML::Document.new(response)\n rex.root.elements.each(\"status\") { |st|\n # month, day, hour, min, sec, year = st.elements['created_at'].text.match(/\\w+ (\\w+) (\\d+) (\\d+):(\\d+):(\\d+) \\S+ (\\d+)/)[1..6]\n # debug [year, month, day, hour, min, sec].inspect\n # time = Time.local(year.to_i, month, day.to_i, hour.to_i, min.to_i, sec.to_i)\n time = Time.parse(st.elements['created_at'].text)\n now = Time.now\n # Sometimes, time can be in the future; invert the relation in this case\n delta = ((time > now) ? time - now : now - time)\n msg = st.elements['text'].to_s + \" (#{Utils.secs_to_string(delta.to_i)} ago via #{st.elements['source'].to_s})\"\n author = \"\"\n if friends\n author = Utils.decode_html_entities(st.elements['user'].elements['name'].text) + \": \" rescue \"\"\n end\n texts << author+Utils.decode_html_entities(msg).ircify_html\n }\n if friends\n # friends always return the latest 20 updates, so we clip the count\n texts[count..-1]=nil\n end\n rescue\n error $!\n if friends\n m.reply \"could not parse status for #{nick}'s friends\"\n else\n m.reply \"could not parse status for #{nick}\"\n end\n return false\n end\n if !show_reply\n nonreplytexts = texts.grep(/^[^@]/)\n if nonreplytexts.length > 0\n texts = nonreplytexts\n end\n end\n # Make sure we have the right number\n texts[count..-1]=nil\n if texts.empty?\n m.reply \"No status updates!\"\n else\n m.reply texts.reverse.join(\"\\n\")\n end\n return true\n else\n if friends\n rep = \"could not get status for #{nick}'s friends\"\n rep << \", try asking in private\" unless m.private?\n else\n rep = \"could not get status for #{nick}\"\n end\n m.reply rep\n return false\n end\n end",
"def show\n client = Twitter::REST::Client.new do |config|\n config.consumer_key = ENV[\"TWITTER_CONSUMER_KEY\"]\n config.consumer_secret = ENV[\"TWITTER_CONSUMER_SECRET\"]\n config.access_token = ENV[\"TWITTER_ACCESS_TOKEN\"]\n config.access_token_secret = ENV[\"TWITTER_ACCESS_TOKEN_SECRET\"]\n end\n\n begin\n timelines = client.user_timeline(@club.twitter_id)\n @tweets = []\n rescue\n end\n\n if timelines.present?\n timelines.each do |timeline_tweet|\n tweet = client.status(timeline_tweet.id)\n @tweets << tweet.text\n end\n end\n end",
"def show\n @new_status = NewStatus.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @new_status }\n end\n end",
"def index\n @swit_statuses = SwitStatus.all\n end",
"def show\n @annual_stat = AnnualStat.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @annual_stat }\n end\n end",
"def retweets_of_me(options={})\n perform_get(\"statuses/retweets_of_me.#{Twitter.format}\", options)\n end",
"def index\n @request_statuses = RequestStatus.all\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @request_statuses }\n end\n end",
"def status(id)\n get(\"mailers/status/#{id.to_i}\")\n end",
"def update_status(status)\n post \"statuses/update\", :post => {:status => status}\n end",
"def status\n sanitized = whitelist(params, :status)\n\n @travel = Travels::Travel.find(sanitized[:id])\n\n respond_to do |format|\n format.html # status.html.erb\n format.json { render json: { status: @travel.state.to_s } }\n end\n end",
"def index\n @application_statuses = ApplicationStatus.all\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @application_statuses }\n end\n end",
"def index\n @tweets = Tweet.all\n\n render json: @tweets\n end",
"def show\n @story_status = StoryStatus.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render :json => @story_status }\n end\n end",
"def show\n @twitter_user = TwitterUser.find(params[:id])\n @follower_ids = Twitter.follower_ids(@twitter_user.twitter_username).collection\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @twitter_user }\n end\n end",
"def index\n @following_user_ids = @user.following.pluck(:id)\n\n @tweets = Tweet.where(user_id: @following_user_ids).order(created_at: :desc).includes(:user)\n json_response(@tweets)\n end",
"def show\n @tweeter = Tweeter.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @tweeter }\n end\n end",
"def list\r\n #page = params[:page].nil? ? 0 : params[:page]\r\n tweets = Tweet.all\r\n render json: tweets\r\n end",
"def marital_status; end",
"def twubbles\n @sad_tweets = WordTweet.recent_sad_tweets\n respond_to do |f|\n f.json { render :json => twubbles_json(@sad_tweets) }\n f.html\n end\n end",
"def twitter_url(json)\n \"http://twitter.com/#{json['from_user']}/status/#{json['id']}\"\n end",
"def get_timeline\n HTTParty.post(\"#{@api_path}/tweets/#{@handle}/#{@password}\")\n end",
"def sample(query_parameters = {})\n start('statuses/sample', query_parameters)\n end",
"def statuses\n\t\trender :json => {:status => 1, :categories => {\"1\" => \"Available\", \"2\" => \"On Hold\", \"3\" => \"Borrowed\", \"4\" => \"Unavailable\"}}, :status => 200\n\t\treturn\n\tend",
"def get_tweet\r\n begin\r\n\t return $client.status(tweet_id.to_i, tweet_mode: \"extended\")\r\n\t rescue Twitter::Error\r\n\t return nil\r\n\t end\r\n end",
"def show\n @received_military_award = ReceivedMilitaryAward.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @received_military_award }\n end\n end",
"def index\n @nvs_mig_statuses = NvsMigStatus.all\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @nvs_mig_statuses }\n end\n end",
"def get_tweets\n # TODO-JW: look into -- respond_to :json\n @card = Card.find(params[:id])\n @api = Twitter.user_timeline(@card.twitter_handle, options={count: 10})\n if @api\n tweets = []\n @api.each_with_index do |tweet,i|\n tweets[i] = {}\n tweets[i][:tweet_id] = String(tweet.id)\n tweets[i][:text] = auto_link(tweet.text)\n tweets[i][:created] = tweet.created_at\n tweets[i][:user_id] = tweet.user.screen_name\n end\n render json: tweets \n else\n [].to_json\n end\n end",
"def show\n @rainbow = Rainbow.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @rainbow }\n end\n end",
"def index\n chain = Tweet\n chain = chain.since_id(params[:since_id]) if params[:since_id]\n @tweets = chain.all(:order => 'msg_twid ASC')\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @tweets }\n end\n end",
"def statuses\n api.get('status')\n end",
"def print_timeline(tweets)\n \n puts tweets [0][\"user\"][\"screen_name\"]\n puts tweets [0][\"text\"]\n\n \n\nend",
"def show\n render json: @tweet\n end",
"def index\n timeline = Tweet.where(user_id: User.find(@current_user_id).follows).or(Tweet.where(user_id: @current_user_id))\n render json: {timeline: timeline}\n end",
"def index\n @barrel_statuses = BarrelStatus.all\n end",
"def get_tweet\n\t\tbegin\n\t return $client.status(tweet_id.to_i, tweet_mode: \"extended\")\n\t rescue Twitter::Error\n\t return nil\n\t end\n end",
"def statuses\n request(:get, \"applicant_tracking/statuses\")\n end",
"def statuses( params={} )\n statuses = get_connections(\"statuses\", params)\n return map_connections statuses, :to => Facebook::Graph::Statuse\n end",
"def statuses; end",
"def user_tweets(user, count=10, since_id=nil, max_id=nil)\n print \"Getting Last %d Statuses for User %s\" % [count, user.to_s]\n print \" since %s\" % since_id if since_id\n print \" until %s\" % max_id if max_id\n print \"\\n\"\n options = {:count => count, :trim_user => true, :include_rts => true, :include_entities => true}\n options[:since_id] = since_id if since_id\n options[:max_id] = max_id if max_id\n begin\n statuses = @MT.user_timeline(user, options)\n if statuses.size > 0\n status_data = statuses.map do |s|\n {\n :user_id => s.user.id,\n :created_at => s.created_at,\n :id => s.id,\n :text => s.text,\n :source => s.source,\n :truncated => s[\"truncated\"],\n :in_reply_to_user_id => s[\"in_reply_to_user_id\"],\n :in_reply_to_screen_name => s[\"in_reply_to_screen_name\"],\n :geo => s[\"geo\"],\n :coordinates => s[\"coordinates\"],\n :place => parse_place(s[\"place\"]),\n :contributors => s[\"contributors\"],\n :retweet_count => s[\"retweet_count\"],\n :entities => parse_entities(s.attrs[\"entities\"]),\n :retweeted_status => parse_retweeted_status(s[\"retweeted_status\"])\n }\n end\n status_data\n else\n []\n end\n rescue Twitter::Error::Unauthorized, Twitter::Error::Forbidden\n puts \"Failed for %s (Protected)\" % user.to_s\n []\n end\n end",
"def show\n @tweetsandwich = Tweetsandwich.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @tweetsandwich }\n end\n end",
"def retweeters_of(id, options={})\n ids_only = !!(options.delete(:ids_only))\n perform_get(\"statuses/#{id}/retweeted_by#{\"/ids\" if ids_only}.#{Twitter.format}\", options)\n end",
"def status\n get_json('status.json')\n end",
"def show\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @reward_and_level }\n end\n end",
"def add_retweet(status)\n\t@tweet_db[status.retweeted_status.id] ||= {\n\t\t:original => status.retweeted_status,\n\t\t:retweets => {}\n\t}\n\t@tweet_db[status.retweeted_status.id][:retweets][status.id] = status\nend",
"def index\n number_tweets = if params[\"count\"] then params[\"count\"].to_i else 10 end\n tweet_ids = []\n if @user.interests\n for i in 1..number_tweets\n interest = @user.interests.sample\n tweet = Rails.application.config.twitter_client.search(\"#{interest[:hashtag]}\", count: 1).take(1)\n tweet_ids.push(tweet.first.id.to_s)\n end\n end\n\n render json: tweet_ids, status: :ok\n end",
"def parse_twitter_api_response(tweets)\n tweets = tweets.map do |tweet|\n if tweet.attrs[:retweeted_status] then\n { created_at: tweet.created_at.getlocal.strftime(\"%l:%M%p - %b %e, %Y\"), name: tweet.user.name, body: \"Retweet: \" << CGI.unescapeHTML(tweet.attrs[:retweeted_status][:full_text]), avatar: tweet.user.profile_image_url_https, screen_name: \"@\" + tweet.user.screen_name}\n else\n { created_at: tweet.created_at.getlocal.strftime(\"%l:%M%p - %b %e, %Y\"), name: tweet.user.name, body: CGI.unescapeHTML(tweet.attrs[:full_text]), avatar: tweet.user.profile_image_url_https, screen_name: \"@\" + tweet.user.screen_name }\n end\n end\n\n return tweets\nend",
"def show \n \n @parsed_raw_tweet = ParsedRawTweet.find(params[:id])\n \n @title = \"Positive Sentiment | Parsed Raw Tweet ID: \"+params[:id]\n @keywords = \"\"\n \n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @parsed_raw_tweet }\n end\n end",
"def show\n @status_animal = StatusAnimal.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @status_animal }\n end\n end"
] |
[
"0.7028147",
"0.6499806",
"0.642677",
"0.6307701",
"0.61127007",
"0.59730375",
"0.59716785",
"0.59531933",
"0.59261584",
"0.5906872",
"0.58738357",
"0.585114",
"0.58212036",
"0.58149934",
"0.5803932",
"0.5800421",
"0.5783163",
"0.5762086",
"0.5734228",
"0.5734228",
"0.5723452",
"0.57042617",
"0.56889564",
"0.5676701",
"0.5646543",
"0.56337965",
"0.5623887",
"0.56186324",
"0.5615919",
"0.5610116",
"0.5585158",
"0.5575222",
"0.5572778",
"0.5567558",
"0.5567558",
"0.55659705",
"0.55601394",
"0.5550676",
"0.55417675",
"0.5533711",
"0.5531722",
"0.55233955",
"0.550043",
"0.5487252",
"0.5484525",
"0.54774725",
"0.54773885",
"0.5457422",
"0.5448854",
"0.54478145",
"0.5445602",
"0.5438813",
"0.5438813",
"0.5403051",
"0.53945893",
"0.5375579",
"0.5374211",
"0.53725326",
"0.5369581",
"0.53502077",
"0.5341518",
"0.53383213",
"0.5332931",
"0.5331027",
"0.53308487",
"0.5327849",
"0.53126514",
"0.5305401",
"0.5303047",
"0.52970076",
"0.52910095",
"0.5290908",
"0.52885014",
"0.52836686",
"0.5274716",
"0.5266018",
"0.5258465",
"0.52562416",
"0.52526075",
"0.5247974",
"0.52413887",
"0.5236509",
"0.5226623",
"0.5219374",
"0.52130353",
"0.521139",
"0.52104944",
"0.52022946",
"0.5190449",
"0.51894695",
"0.51837504",
"0.5182057",
"0.51806295",
"0.5180243",
"0.5172872",
"0.51606226",
"0.51588184",
"0.5158026",
"0.5147229",
"0.5140212",
"0.5134658"
] |
0.0
|
-1
|
POST /married_statuses POST /married_statuses.json
|
def create
@married_status = MarriedStatus.new(married_status_params)
respond_to do |format|
if @married_status.save
format.html { redirect_to @married_status, notice: 'Married status was successfully created.' }
format.json { render :show, status: :created, location: @married_status }
else
format.html { render :new }
format.json { render json: @married_status.errors, status: :unprocessable_entity }
end
end
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def married_status_params\n params.require(:married_status).permit(:name)\n end",
"def create_statuses\n end",
"def create_statuses\n end",
"def index\n @married_statuses = MarriedStatus.all\n end",
"def set_married_status\n @married_status = MarriedStatus.find(params[:id])\n end",
"def update\n respond_to do |format|\n if @married_status.update(married_status_params)\n format.html { redirect_to @married_status, notice: 'Married status was successfully updated.' }\n format.json { render :show, status: :ok, location: @married_status }\n else\n format.html { render :edit }\n format.json { render json: @married_status.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @marital_status = MaritalStatus.new(params[:marital_status])\n\n respond_to do |format|\n if @marital_status.save\n format.html { redirect_to(@marital_status, :notice => 'Marital status was successfully created.') }\n format.xml { render :xml => @marital_status, :status => :created, :location => @marital_status }\n else\n format.html { render :action => \"new\" }\n format.xml { render :xml => @marital_status.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def postTweet(status)\n\t\t\t@client.update(status)\n\t\tend",
"def create\n @status = current_user.statuses.new(status_params)\n\n if current_user.state_facebook? && @status.content?\n current_user.facebook(current_user).put_wall_post(@status.content)\n end\n\n if current_user.state_twitter? && @status.content?\n current_user.twitter(current_user).update(@status.content)\n end\n \n respond_to do |format|\n if @status.save\n format.html { redirect_to statuses_url, notice: 'Post was successfully created.' }\n format.json { render :show, status: :created, location: @status }\n else\n format.html { redirect_to statuses_url }\n format.json { render json: @status.errors, status: :unprocessable_entity }\n end\n end\n end",
"def destroy\n @married_status.destroy\n respond_to do |format|\n format.html { redirect_to married_statuses_url, notice: 'Married status was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def create\n @status = current_user.statuses.new(params[:status])\n\n respond_to do |format|\n if @status.save\n current_user.create_activity(@status, 'created')\n format.html { redirect_to profile_path(current_user), notice: 'Status was successfully created.' }\n format.json { render json: @status, status: :created, location: @status }\n else\n format.html { render action: \"new\" }\n format.json { render json: @status.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @current_statuses = CurrentStatus.new(current_statuses_params)\n\n respond_to do |format|\n if @current_statuses.save\n format.html { redirect_to @current_statuses, notice: 'Current Statuses was successfully created.' }\n format.json { render :show, status: :created, location: @current_statuses }\n else\n format.html { render :new }\n format.json { render json: @current_statuses.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n user = user_from_token\n @tweet = user.tweets.new()\n @tweet.tweet = params[:tweet]\n if @tweet.save\n render :json => @tweet, :status => :created\n else\n render :json => @tweet.errors, :status => :unprocessable_entity\n end\n end",
"def create(attrs, user = @@default_user)\n attrs = { project_token: @project_token }.merge(attrs)\n @attributes = send_request('statuses', :post) do |req|\n req.body = {\n status_object: attrs.slice(:name),\n token: attrs[:project_token],\n auth_token: user.auth_token\n }\n end\n end",
"def create\n @tweet = current_user.tweets.create(params[:tweet])\n respond_with(@tweet, :location => tweet_url(@tweet))\n end",
"def create\n @user = current_user\n @status = @user.statuses.build(status_params)\n\n #@status = Status.new(status_params)\n respond_to do |format|\n if @status.save\n format.html { redirect_to @status, notice: 'Status was successfully created.' }\n format.json { render :show, status: :created, location: @status }\n else\n format.html { render :new }\n format.json { render json: @status.errors, status: :unprocessable_entity }\n end\n end\n end",
"def add_retweet(status)\n\t@tweet_db[status.retweeted_status.id] ||= {\n\t\t:original => status.retweeted_status,\n\t\t:retweets => {}\n\t}\n\t@tweet_db[status.retweeted_status.id][:retweets][status.id] = status\nend",
"def update_status(status)\n post \"statuses/update\", :post => {:status => status}\n end",
"def create\n respond_to do |format|\n begin\n @tweet = TweetApp::ClientContext.status(:post, params[:text])\n flash[:notice] = 'Tweet was successfully created.'\n format.html { redirect_to tweet_url(@tweet) }\n format.json { head :created, :location => tweet_url(@tweet) }\n format.xml { head :created, :location => tweet_url(@tweet) }\n rescue Twitter::RESTError => re\n handle_rest_error(re, format, 'new')\n end\n end\n end",
"def tweet_params\n params.require(:tweet).permit(:status, :zombie_id)\n end",
"def create\n params[:kf_status][:status_type] = params[:status_type] unless params[:status_type].blank?\n params[:kf_status][:relation_id] = params[:relation_id] unless params[:relation_id].blank?\n params[:kf_status][:count_type] = params[:count_type] unless params[:count_type].blank?\n @kf_status = Kf::Status.new(params[:kf_status])\n\n respond_to do |format|\n if @kf_status.save\n format.html { redirect_to \"/kf/statuses?page=#{params[:page]}&relation_id=#{params[:relation_id]}&status_type=#{params[:status_type]}&count_type=#{params[:count_type]}\",notice: 'Status was successfully created.' }\n format.json { render json: @kf_status, status: :created, location: @kf_status }\n else\n format.html { render action: \"new\" }\n format.json { render json: @kf_status.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @tweet = @user.tweets.build(params[:tweet])\n\n respond_to do |format|\n if @tweet.save\n format.html { redirect_to user_tweet_path(@user, @tweet), notice: 'Tweet was successfully created.' }\n format.json { render json: @tweet, status: :created, location: user_tweet_path(@user, @tweet) }\n else\n format.html { render action: \"new\" }\n format.json { render json: @tweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @tweet = current_user.tweets.build(tweet_params)\n\n respond_to do |format|\n if @tweet.save\n format.html { redirect_to @tweet, notice: 'Tweet was successfully created.' }\n format.json { render :show, status: :created, location: @tweet }\n else\n format.html { render :new }\n format.json { render json: @tweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n params['user_id'] = current_user.id if current_user\n @mood = Mood.new(mood_params)\n\n if @mood.save\n render json: { location: format_mood(@mood) }, status: :created\n else\n render json: @mood.errors, status: :unprocessable_entity\n end\n end",
"def create\n begin \n new_tweet = Tweet.new(tweet_params)\n if new_tweet.save()\n render(status: 201, json: {tweet: new_tweet})\n else # if validation fails\n render(status: 422, json: {error: new_tweet.errors})\n end\n rescue => error\n render(status: 500, json: {error: \"Internal Server Error: #{error.message}\"})\n end\n end",
"def create\n # Use current_user below, so that new tweets are only created by the logged in user #MDM\n #@tweet = Tweet.new(tweet_params)\n @tweet = current_user.tweets.new(tweet_params)\n\n respond_to do |format|\n if @tweet.save\n format.html { redirect_to @tweet, notice: 'Tweet was successfully created.' }\n format.json { render :show, status: :created, location: @tweet }\n else\n format.html { render :new }\n format.json { render json: @tweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @status = Status.new(status_params)\n\n if @status.save\n render json: @status\n else\n render json: @status.errors, status: :unprocessable_entity\n end\n end",
"def create\n @tweet = Tweet.create(tweet_params)\n\n @tweet = get_tagged(@tweet)\n\n respond_to do |format|\n if @tweet.save\n format.html { redirect_to home_path, notice: 'Tweet was successfully created.' }\n format.json { render :show, status: :created, location: @tweet }\n else\n format.html { render :new }\n format.json { render json: @tweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @roof = Roof.find(params[:roof_id])\n @status = @roof.statuses.new(params[:roof_status])\n\n respond_to do |format|\n if @status.save\n format.html { redirect_to @roof, notice: 'Спасибо что помогаете нам! Заявка отправлена на модерацию :)' }\n format.json { render json: @status, status: :created, location: @status }\n else\n format.html { render action: \"new\" }\n format.json { render json: @status.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @mstatus = Mstatus.new(mstatus_params)\n @mstatus.user_id = current_user.id\n\n respond_to do |format|\n if @mstatus.save\n format.html { redirect_to @mstatus, notice: 'Status was successfully created.' }\n format.json { render :show, status: :created, location: @mstatus }\n else\n format.html { render :new }\n format.json { render json: @mstatus.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @tweet = Tweet.new(params[:tweet])\n\n respond_to do |format|\n if @tweet.save\n format.html { redirect_to @tweet, notice: 'Tweet was successfully created.' }\n format.json { render json: @tweet, status: :created, location: @tweet }\n else\n format.html { render action: \"new\" }\n format.json { render json: @tweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @tweet = Tweet.new(params[:tweet])\n\n respond_to do |format|\n if @tweet.save\n format.html { redirect_to @tweet, notice: 'Tweet was successfully created.' }\n format.json { render json: @tweet, status: :created, location: @tweet }\n else\n format.html { render action: \"new\" }\n format.json { render json: @tweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @tw_stat = TwStat.new(tw_stat_params)\n\n respond_to do |format|\n if @tw_stat.save\n format.html { redirect_to @tw_stat, notice: 'Tw stat was successfully created.' }\n format.json { render :show, status: :created, location: @tw_stat }\n else\n format.html { render :new }\n format.json { render json: @tw_stat.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @tweet = Tweet.new(tweet_params)\n if @tweet.username == nil\n # This is for the current user posting tweets\n @tweet.username = current_user.name\n @tweet.user_id = current_user.id\n # Updates to Twitter\n current_user.twitter.update(@tweet.tweetbody)\n else \n # Incoming tweets from the daemon script\n @tweet.save\n end\n respond_with(@tweet)\n end",
"def create\n @new_status = NewStatus.new(params[:new_status])\n\n respond_to do |format|\n if @new_status.save\n format.html { redirect_to @new_status, notice: 'New status was successfully created.' }\n format.json { render json: @new_status, status: :created, location: @new_status }\n else\n format.html { render action: \"new\" }\n format.json { render json: @new_status.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @tweet = Tweet.new(tweet_params)\n \n respond_to do |format|\n if @tweet.save\n format.html { redirect_to tweets_path, notice: 'Tweet was successfully created.' }\n format.json { render :show, status: :created, location: @tweet }\n else\n format.html { render :new }\n format.json { render json: @tweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @retweet = Retweet.new(params[:retweet])\n\n respond_to do |format|\n if @retweet.save\n format.html { redirect_to @retweet, notice: 'Retweet was successfully created.' }\n format.json { render json: @retweet, status: :created, location: @retweet }\n else\n format.html { render action: \"new\" }\n format.json { render json: @retweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def mention(status, *names)\n tweet(\"#{names.join(' ')} #{status}\")\nend",
"def mention(status, *names)\n tweet(\"#{names.join(' ')} #{status}\")\nend",
"def create\r\n @tweet = Tweet.new(tweet_params)\r\n @tweet.user = current_user\r\n \r\n respond_to do |format|\r\n if @tweet.save\r\n format.html { redirect_to @tweet, notice: \"Tweet was successfully created.\" }\r\n format.json { render :show, status: :created, location: @tweet }\r\n else\r\n format.html { render :new, status: :unprocessable_entity }\r\n format.json { render json: @tweet.errors, status: :unprocessable_entity }\r\n end\r\n end\r\n end",
"def create\n username = params[:username]\n tweets = timeline_query username\n tweet_sentiment_list = get_sentiment(tweets)\n redirect_to :action => \"index\", :tweets => tweet_sentiment_list\n end",
"def update_status(payload, status)\n sha = payload.after\n repo = payload.repository.full_name\n state, description = status.first\n\n # setup http post\n uri = URI.parse(\"#{GITHUB_ROOT}/repos/#{repo}/statuses/#{sha}\")\n http = Net::HTTP.new(uri.host, uri.port)\n http.use_ssl = true\n\n # post to GitHub\n params = {:state => state, :description => description, :context => CONTEXT}\n http.post(uri.path, params.to_json, HEADERS)\nend",
"def create\n @story_status = StoryStatus.new(params[:story_status])\n\n respond_to do |format|\n if @story_status.save\n format.html { redirect_to @story_status, :notice => 'Story status was successfully created.' }\n format.json { render :json => @story_status, :status => :created, :location => @story_status }\n else\n format.html { render :action => \"new\" }\n format.json { render :json => @story_status.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def update\n\n respond_to do |format|\n if @marital_status.update_attributes(params[:marital_status])\n format.html { redirect_to(@marital_status, :notice => 'Marital status was successfully updated.') }\n format.xml { head :ok }\n else\n format.html { render :action => \"edit\" }\n format.xml { render :xml => @marital_status.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def create\n @anniversary = Anniversary.new\n form = params[:anniversary]\n @anniversary.anniversary = form[:anniversary]\n @anniversary.description = form[:description]\n\n @anniversary.user_id = current_user.id\n last_anniversary = Anniversary.find(:all, :conditions => ['user_id', '1']).first\n if last_anniversary\n @anniversary.date = last_anniversary.date + 1.days\n else\n @anniversary.date = Date.today + 1.days\n end\n @anniversary.save\n # FIXME 3rd parameter\n tweet_body = \"@#{@anniversary.user.twitter_id} は最高の幸せを噛み締めた。そう、一日は終わったのだ。http://cutend.me:3000/anniversaries/#{@anniversary.id} #cutend\"\n tweet_result = current_user.tweet({:tweet_body => tweet_body, :user => current_user})\n @anniversary.tweet_id = tweet_result[:id].to_i\n\n respond_to do |format|\n if @anniversary.save\n format.html { redirect_to '/anniversaries', notice: 'Anniversary was successfully created.' }\n format.json { render json: @anniversary, status: :created, location: @anniversary }\n else\n format.html { render action: \"new\" }\n format.json { render json: @anniversary.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @tweet = current_user.tweets.new(tweet_params)\n\n respond_to do |format|\n if @tweet.save\n format.html { redirect_back fallback_location: root_path, notice: 'ツーイトの投稿が完了しました。' }\n format.json { render :show, status: :created, location: @tweet }\n else\n format.html { render :new }\n format.json { render json: @tweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @status = Status.new(params[:status])\n @status.user = current_user\n \n respond_to do |format|\n if @status.save\n flash[:notice] = 'Status was successfully created.'\n format.html { redirect_to(statuses_path) }\n format.xml { render :xml => @status, :status => :created, :location => @status }\n else\n format.html { render :action => \"new\" }\n format.xml { render :xml => @status.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def create\n @mood = Mood.new(mood_params)\n @mood.user = current_user\n respond_to do |format|\n if @mood.save\n format.html { redirect_to moods_path, notice: \"Mood was successfully created.\" }\n format.json { render :show, status: :created, location: @mood }\n else\n format.html { render :new, status: :unprocessable_entity }\n format.json { render json: @mood.errors, status: :unprocessable_entity }\n end\n end\n end",
"def generate_tweet\n raw_tweet = raw['status']\n JsonTweet.new(raw_tweet, twitter_user_id).generate_tweet if raw_tweet\n end",
"def create\n @tweet = current_user.tweets.build(tweet_params)\n\n respond_to do |format|\n if @tweet.save\n format.html { redirect_to member_url(current_user), notice: '投稿しました' }\n format.json { render :show, status: :created, location: @tweet }\n else\n format.html { render :new }\n format.json { render json: @tweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @status = Status.new(status_params)\n\n respond_to do |format|\n if @status.save\n format.html { redirect_to @status, notice: 'Status was successfully created.' }\n format.json { render :show, status: :created, location: @status }\n else\n format.html { render :new }\n format.json { render json: @status.errors, status: :unprocessable_entity }\n end\n end\n end",
"def mood_params\n params.require(:mood).permit(:status, :time, :reason, :user_id)\n end",
"def mention(status, *name)\n tweet(\"#{name.join('')} #{status}\")\nend",
"def update\n respond_to do |format|\n if @current_statuses.update(current_statuses_params)\n format.html { redirect_to @current_statuses, notice: 'Current Statuses was successfully updated.' }\n format.json { render :show, status: :ok, location: @current_statuses }\n else\n format.html { render :edit }\n format.json { render json: @current_statuses.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @tweet = current_user.tweets.new(tweet_params)\n\n # respond_to do |format|\n if @tweet.save\n # format.html { \n # puts \"html\"\n # render json: TweetBlueprint.render(@tweet), status: :created \n # }\n # format.json { \n puts \"json\"\n render json: TweetBlueprint.render(@tweet), status: :created \n # }\n else\n # format.html { \n # puts \"error html\"\n # render :new \n # }\n # format.json { \n puts \"error json\"\n render json: { result: 'error', error: @tweet.errors }, status: :unprocessable_entity \n # }\n end\n # end\n end",
"def create\n tweet.user = current_user\n\n respond_to do |format|\n if tweet.save\n format.html { redirect_to tweet, notice: 'Tweet was successfully created.' }\n format.json { render :show, status: :created, location: tweet }\n else\n format.html { render :new }\n format.json { render json: tweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @status = Status.new(status_params)\n\n respond_to do |format|\n if @status.save\n format.html { redirect_to @status, notice: 'Status was successfully created.' }\n format.json { render action: 'show', status: :created, location: @status }\n else\n format.html { render action: 'new' }\n format.json { render json: @status.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @tweet = Tweet.new(tweet_params.merge(user: current_user))\n\n respond_to do |format|\n if @tweet.save\n format.html { redirect_to root_path, notice: \"Tweet was successfully created.\" }\n format.json { render :show, status: :created, location: @tweet }\n else\n format.html { render :new, status: :unprocessable_entity }\n format.json { render json: @tweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def new\n @roof = Roof.find(params[:roof_id])\n @status = @roof.statuses.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @status }\n end\n end",
"def create\n @tweet = Tweet.new(tweet_params)\n\n if @tweet.save\n render :show, status: :created, location: { tweet: @tweet }\n else\n render json: @tweet.errors, status: :unprocessable_entity\n end\n end",
"def statuses=(statuses)\n assert_unloaded\n @statuses = Array(statuses)\n end",
"def create\n @tweet = current_user.tweets.build(tweet_params)\n\n #respond_to do |format|\n if @tweet.save\n flash[:success] = 'ツイートを投稿しました。'\n redirect_to tweets_path\n #format.html { redirect_to @tweet }\n #format.json { render :show, status: :created, location: @tweet }\n else\n @tweets = current_user.feed_tweets.order(id: :desc).page(params[:page])\n flash.now[:danger] = 'ツイートを投稿できませんでした。'\n render 'tweets/index'\n #format.html { render :new }\n #format.json { render json: @tweet.errors, status: :unprocessable_entity }\n end\n end",
"def create\n @tweet = Tweet.new(tweet_params)\n @tweet.user_vote = 0\n\n respond_to do |format|\n if @tweet.save\n format.html { redirect_to @tweet, notice: 'Tweet was successfully created.' }\n format.json { render :show, status: :created, location: @tweet }\n else\n format.html { render :new }\n format.json { render json: @tweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @relationship_status = RelationshipStatus.new(relationship_status_params)\n\n respond_to do |format|\n if @relationship_status.save\n format.html { redirect_to @relationship_status, notice: 'Relationship status was successfully created.' }\n format.json { render action: 'show', status: :created, location: @relationship_status }\n else\n format.html { render action: 'new' }\n format.json { render json: @relationship_status.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @pony = Pony.new(params[:pony])\n\n respond_to do |format|\n if @pony.save\n t = Tweet.create(:status =>\"Ya llegue al mundo magico de ponies\", :pony_id => @pony[:id])\n\n format.html { redirect_to @pony, notice: 'Pony was successfully created.' }\n format.json { render json: @pony, status: :created, location: @pony }\n else\n format.html { render action: \"new\" }\n format.json { render json: @pony.errors, status: :unprocessable_entity }\n end\n \n\n end\n end",
"def create_status(name, description, level, image)\n response = @client.post(\"/api/v1/statuses\", { \"name\" => name, \"description\" => description, \"level\" => level, \"image\" => image })\n return JSON.parse(response.body)\n end",
"def create\n @received_military_award = ReceivedMilitaryAward.new(params[:received_military_award])\n\n respond_to do |format|\n if @received_military_award.save\n format.html { redirect_to @received_military_award.military_history.personal_detail, notice: 'Received military award was successfully created.' }\n format.json { render json: @received_military_award, status: :created, location: @received_military_award }\n else\n format.html { render action: \"new\" }\n format.json { render json: @received_military_award.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @tweet = Tweet.create(:user_id => current_user.id, :text => tweet_params[:text])\n\n respond_to do |format|\n if @tweet.save\n format.html { redirect_to @tweet, notice: 'Tweet was successfully created.' }\n format.json { render :show, status: :created, location: @tweet }\n else\n format.html { render :new }\n format.json { render json: @tweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create_status(*args)\n arguments(args, required: [:user, :repo, :id]) do\n assert_required %w[ state ]\n end\n params = arguments.params\n params['accept'] ||= PREVIEW_MEDIA\n\n post_request(\"repos/#{arguments.user}/#{arguments.repo}/deployments/#{arguments.id}/statuses\", params)\n end",
"def add_status( thing, statusArray )\n data = {\n token: @applicationToken,\n thing_id: thing.id,\n id: 'null',\n status_array: statusArray\n }\n @httpHelper.post_data( 'status', data )\n end",
"def create\n @mood = current_user.moods.new(mood_params)\n\n respond_to do |format|\n if @mood.save\n format.html { redirect_to moods_path, notice: 'Mood was successfully created.' }\n format.json { render json: @mood, status: :created, location: @mood }\n else\n format.html { render action: \"new\" }\n format.json { render json: @mood.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n require 'twitter'\t \n @muscle_diary = MuscleDiary.new(muscle_diary_params)\n @muscle_diary.user_id = current_user.id\n\n # tweet記述\n auth = request.env[\"omniauth.auth\"]\n \n client = Twitter::REST::Client.new do |config|\n # developer\n\n config.consumer_key = \"p1kXruMcvR9qzaANAphm3sgav\"\n config.consumer_secret = \"rLLYwiOdTqq7HJNoSQLiqhMEQ55IKGriax9mlYKIj3AKGKgXCl\"\n\n #config.consumer_key = Rails.application.secrets.twitter_consumer_key\n #config.consumer_secret = Rails.application.secrets.twitter_consumer_secret\n \n # user\n config.access_token = \"hogehoge\" \n config.access_token_secret = \"hogehoge\"\n\n #config.access_token = auth.credentials.token\n #config.access_token_secret = auth.credentials.secret\n end\n \n # Twitter投稿\n client.update(\"test\")\n\n respond_to do |format|\n if @muscle_diary.save\n format.html { redirect_to '/', notice: '日記をつけました。ナイスバルク!' }\n format.json { render action: 'show', status: :created, location: @muscle_diary }\n else\n format.html { render action: 'new' }\n format.json { render json: @muscle_diary.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @rented_unit = appointment.rented_units.new(rented_unit_params)\n\n respond_to do |format|\n if @rented_unit.save\n format.html { redirect_to [building, listing, appointment, @rented_unit], notice: 'Rented unit was successfully created.' }\n format.json { render :show, status: :created, location: @rented_unit }\n else\n format.html { render :new }\n format.json { render json: @rented_unit.errors, status: :unprocessable_entity }\n end\n end\n end",
"def statuses\n Sifter.\n get(\"/api/statuses\").\n parsed_response[\"statuses\"]\n end",
"def create\n @mood = Mood.new(mood_params)\n\n respond_to do |format|\n if @mood.save\n format.html { redirect_to @mood, notice: 'Mood was successfully created.' }\n format.json { render action: 'show', status: :created, location: @mood }\n else\n format.html { render action: 'new' }\n format.json { render json: @mood.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create_status(*args)\n arguments(args, required: [:user, :repo, :id]) do\n assert_required %w[ state ]\n permit VALID_STATUS_OPTIONS\n end\n params = arguments.params\n params['accept'] ||= PREVIEW_MEDIA\n\n post_request(\"repos/#{arguments.user}/#{arguments.repo}/deployments/#{arguments.id}/statuses\", params)\n end",
"def create\n @tweet = Tweet.new(tweet_params)\n @tweet.user = current_user\n respond_to do |format|\n if @tweet.save\n format.html { redirect_to root_path, notice: \"Tweet creado felizmente.\" }\n format.json { redirect_to root_path, status: :created, location: @tweet }\n else\n format.html { redirect_to root_path, notice: \"No se puede tuitear nada.\" }\n format.json { render json: @tweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def build_statuses(options = {})\n build_statuses_resource(options)\n end",
"def create\n @new_tweet = NewTweet.new(new_tweet_params)\n\n respond_to do |format|\n if @new_tweet.save\n format.html { redirect_to @new_tweet, notice: 'New tweet was successfully created.' }\n format.json { render :show, status: :created, location: @new_tweet }\n else\n format.html { render :new }\n format.json { render json: @new_tweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n \n # API Status Update - Basic Authentication\n if !session[\"user_id\"]\n authenticate_or_request_with_http_basic do |username, password|\n if request.url.index('localhost')\n user = User.find(:first, :conditions => ['username LIKE ?', username.strip])\n else\n user = User.find(:first, :conditions => ['username ILIKE ?', username.strip])\n end\n\n if user && user.authenticate(password)\n session[\"user_id\"] = user.id\n end\n end\n end\n \n\n respond_to do |format|\n \n if params[:status] and params[:status][:message] \n postmessage = params[:status][:message]\n elsif params[:message] \n postmessage = params[:message]\n end\n \n if session[\"user_id\"] and postmessage\n\n @status = Status.new \n @status.user_id = session[\"user_id\"]\n @status.message = postmessage \n @status.save\n \n @activity = Activity.new\n @activity.user_id = session[\"user_id\"]\n @activity.status_id = @status.id\n @activity.council_id = @council.id\n @activity.save\n \n # Check if message includes a mention \n txtcall = @status.message.gsub(\"@\", \"\")\n txtarray = txtcall.split(\" \")\n if request.url.index('localhost')\n @mention = User.find(:first, :conditions => ['username LIKE ?', txtarray[0]])\n else\n @mention = User.find(:first, :conditions => ['username ILIKE ?', txtarray[0]])\n end\n \n if @mention and @mention.email and @mention.allowemail != false\n begin\n MentionMailer.alerter(session[:username], @mention.username, session[:image], @mention.email, @activity.id.to_s).deliver\n rescue\n end\n end \n \n # Tweet\n if session['access_token'] and params[:status][:twitter] == \"1\"\n @client.update(@status.message)\n end \n \n # FB\n if session['fbtoken'] and params[:status][:facebook] == \"1\"\n RestClient.post 'https://graph.facebook.com/' + session['fbid'] + '/feed', :access_token => session['fbtoken'], :message => @status.message\n end\n \n \n format.html { redirect_to @status, notice: 'Status was successfully created.' }\n # format.json { render json: @status, status: :created, location: @status }\n format.js { render :action => 'create.js.coffee', :content_type => 'text/javascript'}\n format.json { render :json => @status}\n else\n format.html { render action: \"new\" }\n format.json { }\n end\n end\n end",
"def maritalinfo_params\n params.require(:maritalinfo).permit(:marital_status, :wedding_anniversary_date, :spouse_type, :spouse_firstt_name, :spouse_middle_name, :spouse_sur_name, :no_of_family_members, :kids, :parents_dependent_on_you, :member_id)\n end",
"def create\n @status = current_user.statuses.new(status_params)\n if params[:borrowed_playlist].empty?\n @status.image = RSpotify::Playlist.find(current_user.uid, @status.playlist).images.first[\"url\"]\n # @status.image = RSpotify::Playlist.find(\"wizzler\", @status.playlist).images.first[\"url\"]\n else\n playlist_attrs = URI(params[:borrowed_playlist]).path.split(\"/\")\n @status.spotify_uid, @status.playlist = playlist_attrs[2], playlist_attrs[4]\n @status.image = RSpotify::Playlist.find(@status.spotify_uid, @status.playlist).images.first[\"url\"]\n end\n\n respond_to do |format|\n if @status.save\n format.html { redirect_to @status, notice: 'Status was successfully created.' }\n format.json { render :show, status: :created, location: @status }\n else\n format.html { render :new }\n format.json { render json: @status.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @tweett = Tweett.new(tweett_params)\n\n respond_to do |format|\n if @tweett.save\n format.html { redirect_to @tweett, notice: 'Tweett was successfully created.' }\n format.json { render :show, status: :created, location: @tweett }\n else\n format.html { render :new }\n format.json { render json: @tweett.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @tweet = current_user.tweets.new tweet_params\n\n if current_user\n @tweet.user_id = current_user.id\n end\n\n respond_to do |format|\n if @tweet.save\n format.html { redirect_to @tweet, notice: \"Tweet was successfully created.\" }\n format.json { render :show, status: :created, location: @tweet }\n else\n format.html { render :new, status: :unprocessable_entity }\n format.json { render json: @tweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @tweet = FuKing::Twitter.update(params[:tweet][:text])\n\n respond_to do |format|\n if @tweet\n flash[:notice] = 'Tweet was successfully created.'\n format.html { redirect_to(:action => :index) }\n format.mobile { redirect_to(:action => :index) }\n format.xml { render :xml => @tweet, :status => :created, :tweet => @tweet }\n else\n format.html { render :action => \"new\" }\n format.mobile { render :action => \"new\" }\n format.xml { render :xml => @tweet.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def create\n @status_animal = StatusAnimal.new(params[:status_animal])\n\n respond_to do |format|\n if @status_animal.save\n format.html { redirect_to @status_animal, notice: 'Status animal was successfully created.' }\n format.json { render json: @status_animal, status: :created, location: @status_animal }\n else\n format.html { render action: \"new\" }\n format.json { render json: @status_animal.errors, status: :unprocessable_entity }\n end\n end\n end",
"def tweet_params\n params.require(:tweet).permit(:status, :message, :location, :user_id)\n end",
"def create\n rental = sporting_goods.rentals.new(rental_params)\n rental.user_id = current_user.id\n if rental.save\n render json: rental, status: 200\n else\n render json: rental, status: 400\n end\n end",
"def create\n @tweeter = Tweeter.new(params[:tweeter])\n\n respond_to do |format|\n if @tweeter.save\n format.html { redirect_to @tweeter, notice: 'Tweeter was successfully created.' }\n format.json { render json: @tweeter, status: :created, location: @tweeter }\n else\n format.html { render action: \"new\" }\n format.json { render json: @tweeter.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @status = Status.new(params[:status])\n\n respond_to do |format|\n if @status.save \n @user = @status.user\n @user.view_status = @status.view_status\n @user.save\n \n if @user.view_status == 'live' && @user.teacher?\n Innsights.report(\"Maestro aceptado\", :user => current_user, :group => current_network).run\n elsif @user.teacher?\n Innsights.report(\"Maestro suspendido\", :user => current_user, :group => current_network).run\n end\n\n format.html { redirect_to :back, notice: 'El status ha sido agregado correctamente al usuario' }\n format.json { render json: :back , status: :created, location: @status }\n else\n format.html { render action: \"new\" }\n format.json { render json: @status.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @annual_stat = AnnualStat.new(params[:annual_stat])\n\n respond_to do |format|\n if @annual_stat.save\n format.html { redirect_to @annual_stat, notice: 'Annual stat was successfully created.' }\n format.json { render json: @annual_stat, status: :created, location: @annual_stat }\n else\n format.html { render action: \"new\" }\n format.json { render json: @annual_stat.errors, status: :unprocessable_entity }\n end\n end\n end",
"def statusesReceived_forRequest(statuses,identifier)\n\t\tputs \"statusesReceived_forRequest\"\n\t\tself.statuses = statuses\n\tend",
"def create\n @barrel = Barrel.find(params[:barrel_id])\n if @barrel\n @barrel_status = @barrel.barrel_statuses.build(barrel_status_params)\n if @barrel_status.save\n render :show, :status => :created\n else\n render :json => { :errors => @barrel_status.errors }, :status => :unprocessable_entity\n end\n else\n render :json => { :message => \"No barrel found\" }, :status => :not_found\n end\n\n end",
"def create\n @user_mood = UserMood.new(user_mood_params)\n\n respond_to do |format|\n if @user_mood.save\n format.html { redirect_to @user_mood, notice: 'User mood was successfully created.' }\n format.json { render :show, status: :created, location: @user_mood }\n else\n format.html { render :new }\n format.json { render json: @user_mood.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @learning_status = LearningStatus.new(learning_status_params)\n @learning_status.user = current_user\n\n respond_to do |format|\n if @learning_status.save\n format.html { redirect_to @learning_status, notice: 'Learning status was successfully created.' }\n format.json { render :show, status: :created, location: @learning_status }\n else\n format.html { render :new }\n format.json { render json: @learning_status.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n\n #player = Player.all.sample\n #game = Game.all.sample\n #tr = TweetRecord.new\n #tr.user_screen_name=\"rebeccag_dev\"\n #tr.user_twitter_id=1234567890123\n #tr.status_text=\"@c2sb #g#{game.id}p#{player.id}sFGM\"\n #Rails.logger.info \"Sending tweet #{tr.inspect}\"\n #TweetCollector.add_tweet(tr)\n ##StatisticsCollector.add_tweet(68,\"#g17p#{player.id}sFGM\")\n #Rails.logger.info \"Submitted tweet for player #{player.id} - #{player.name}\"\n #Rails.logger.info(\"Tweet log #{StatisticsCollector.get_tweet_log.last.inspect}\")\n #has_error = false\n\n\n @user_reported_statistic = UserReportedStatistic.new()\n stat_params=params[:user_reported_statistic]\n @tweet = params[:tweet]\n user_id = stat_params[:user]\n tr = TweetRecord.new\n tr.status_text=@tweet\n tr.user_id= user_id\n TweetCollector.add_tweet(tr)\n @user_reported_statistic = UserReportedStatistic.new\n\n if tr.has_error?\n @statistic_types = StatisticType.all\n @games = Game.all\n @teams = Team.all\n @players = Player.all\n @users = User.all\n tr.error_msgs.each do | x|\n @user_reported_statistic.errors.add(:tweet,x)\n end\n end\n logger.info(\"logger update_stat\")\n\n respond_to do |format|\n if tr.has_error?\n format.html { render action: \"new\" }\n format.json { render json: @user_reported_statistic.errors, status: :unprocessable_entity }\n else\n format.html { redirect_to user_reported_statistics_url, notice: 'User reported statistic was successfully created.' }\n format.json { render json: @user_reported_statistic, status: :created, location: @user_reported_statistic }\n end\n\n end\n end",
"def create\n @tweet = Tweet.new(tweet_params)\n @tweet.user_id = current_user.id\n respond_to do |format|\n if @tweet.save\n format.html { redirect_to root_path, notice: \"Tu Tweet se ha publicado con exito!\" }\n format.json { render :show, status: :created, location: @tweet }\n else\n format.html { render :new, status: :unprocessable_entity }\n format.json { render json: @tweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def tweet_params\n params.require(:tweet).permit(:content, :total_tweet, :retweet)\n end",
"def parse_retweeted_status(rs)\n if rs.nil?\n nil\n else\n rs = { \n :created_at => rs.created_at,\n :id => rs.id,\n :text => rs.text,\n :source => rs.source, \n :truncated => rs[\"truncated\"],\n :in_reply_to_status_id => rs[\"in_reply_to_status_id\"],\n :in_reply_to_user_id => rs[\"in_reply_to_user_id\"],\n :in_reply_to_screen_name => rs[\"in_reply_to_screen_name\"],\n :user_id => rs[\"user\"][\"id\"] \n }\n rs\n end\n end",
"def create\n @status = Status.new(params[:status])\n @status.save\n\n redirect_to @status, notice: 'Status was successfully created.'\n end"
] |
[
"0.647825",
"0.6294594",
"0.6294594",
"0.6277677",
"0.6146193",
"0.59509134",
"0.58110416",
"0.5783689",
"0.57458234",
"0.56536686",
"0.56203765",
"0.55974174",
"0.5560564",
"0.55356175",
"0.551086",
"0.55105096",
"0.5473909",
"0.54488045",
"0.54235417",
"0.53783315",
"0.53690994",
"0.5294834",
"0.52481496",
"0.5245258",
"0.5205642",
"0.5204757",
"0.52029604",
"0.5197876",
"0.51934487",
"0.5184146",
"0.517981",
"0.517981",
"0.514653",
"0.5141974",
"0.5140182",
"0.51169497",
"0.51095027",
"0.5098448",
"0.5098448",
"0.5078373",
"0.5077172",
"0.50635725",
"0.5059467",
"0.5053378",
"0.50528485",
"0.5050878",
"0.50416",
"0.50411975",
"0.5038852",
"0.5038655",
"0.5035547",
"0.5034964",
"0.5032144",
"0.50287867",
"0.50231564",
"0.50229913",
"0.50125533",
"0.5010521",
"0.49827635",
"0.4977633",
"0.49609604",
"0.4960936",
"0.49606577",
"0.4953037",
"0.49426788",
"0.49385706",
"0.49298373",
"0.49276626",
"0.49264878",
"0.49250263",
"0.49156296",
"0.4911248",
"0.49071845",
"0.49071696",
"0.49068943",
"0.49068713",
"0.49027073",
"0.49023595",
"0.48864233",
"0.48859736",
"0.4884415",
"0.48838297",
"0.48686668",
"0.48678818",
"0.4865208",
"0.48636594",
"0.48609605",
"0.48597717",
"0.4857612",
"0.48562142",
"0.48561478",
"0.4853979",
"0.4851007",
"0.4848537",
"0.4847304",
"0.48470187",
"0.48450744",
"0.4844191",
"0.4842855",
"0.4840956"
] |
0.6917205
|
0
|
PATCH/PUT /married_statuses/1 PATCH/PUT /married_statuses/1.json
|
def update
respond_to do |format|
if @married_status.update(married_status_params)
format.html { redirect_to @married_status, notice: 'Married status was successfully updated.' }
format.json { render :show, status: :ok, location: @married_status }
else
format.html { render :edit }
format.json { render json: @married_status.errors, status: :unprocessable_entity }
end
end
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def update\n respond_to do |format|\n if @current_statuses.update(current_statuses_params)\n format.html { redirect_to @current_statuses, notice: 'Current Statuses was successfully updated.' }\n format.json { render :show, status: :ok, location: @current_statuses }\n else\n format.html { render :edit }\n format.json { render json: @current_statuses.errors, status: :unprocessable_entity }\n end\n end\n end",
"def set_married_status\n @married_status = MarriedStatus.find(params[:id])\n end",
"def update\n\n respond_to do |format|\n if @marital_status.update_attributes(params[:marital_status])\n format.html { redirect_to(@marital_status, :notice => 'Marital status was successfully updated.') }\n format.xml { head :ok }\n else\n format.html { render :action => \"edit\" }\n format.xml { render :xml => @marital_status.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def update\n @roof = Roof.find(params[:roof_id])\n @status = @roof.statuses.find(params[:id])\n\n respond_to do |format|\n if @status.update_attributes(params[:roof_status])\n format.html { redirect_to @roof, notice: 'Status was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @status.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n\n #LA\n #@status = current_user.statuses.find(status_params)\n\n #if (status_params) && (status_params).has_key?(:user_id)\n # (status_params).delete(:user_id) \n #end\n\n respond_to do |format|\n if @status.update(status_params)\n format.html { redirect_to statuses_url, notice: 'Status was successfully updated.' }\n format.json { render :show, status: :ok, location: @status }\n else\n format.html { render action: 'edit' }\n format.json { render json: @status.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update(attrs, user = @@default_user)\n attrs = { id: @id, project_token: @project_token }.merge(attrs)\n @attributes = send_request(\"statuses/#{attrs[:id]}\", :put) do |req|\n req.body = {\n status_object: attrs.slice(:name),\n token: attrs[:project_token],\n auth_token: user.auth_token\n }\n end\n end",
"def update\n respond_to do |format|\n if @status.update(isshared: params[:isshared])\n format.json { head :no_content }\n else\n format.json {\n render json: @status.errors,\n status: :unprocessable_entity\n }\n end\n end\n end",
"def update\n @new_status = NewStatus.find(params[:id])\n\n respond_to do |format|\n if @new_status.update_attributes(params[:new_status])\n format.html { redirect_to @new_status, notice: 'New status was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @new_status.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n params[:kf_status][:status_type] = params[:status_type] unless params[:status_type].blank?\n params[:kf_status][:relation_id] = params[:relation_id] unless params[:relation_id].blank?\n params[:kf_status][:count_type] = params[:count_type] unless params[:count_type].blank?\n @kf_status = Kf::Status.find(params[:id])\n\n respond_to do |format|\n if @kf_status.update_attributes(params[:kf_status])\n format.html { redirect_to \"/kf/statuses?page=#{params[:page]}&relation_id=#{params[:relation_id]}&status_type=#{params[:status_type]}&count_type=#{params[:count_type]}\", notice: 'Status was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @kf_status.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n # Add in this #MDM\n @tweet = Tweet.find(params[:id]) \n \n respond_to do |format|\n if @tweet.update(tweet_params)\n format.html { redirect_to @tweet, notice: 'Tweet was successfully updated.' }\n format.json { render :show, status: :ok, location: @tweet }\n else\n format.html { render :edit }\n format.json { render json: @tweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n ticket = Ticket.find(params[:id])\n status = params[:status]\n\n if status and ticket.update_attributes(status: status)\n render json: {\"notice\"=>\"status updated successfully to '#{status}'\"}\n else\n render json: {\"alert\"=>\"status not updated. check params.\"}\n end\n end",
"def update\n @status = Status.find(params[:id])\n\n respond_to do |format|\n if @status.update_attributes(params[:status])\n format.html { redirect_to @status, notice: 'Status was successfully updated.' }\n format.json { head :ok }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @status.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @status = Status.find(params[:id])\n\n respond_to do |format|\n if @status.update_attributes(params[:status])\n format.html { redirect_to @status, notice: 'Status was successfully updated.' }\n format.json { head :ok }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @status.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update_status(status)\n post \"statuses/update\", :post => {:status => status}\n end",
"def update\n @mood = Mood.find(params[:id])\n @user = User.find(params[:user_id])\n\n respond_to do |format|\n if @mood.update_attributes(params[:mood])\n format.json { head :ok }\n else\n format.json { render :xml => @mood.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def married_status_params\n params.require(:married_status).permit(:name)\n end",
"def update\n respond_to do |format|\n if @status.update(status_params)\n format.html { redirect_to @status, notice: 'Status was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: 'edit' }\n format.json { render json: @status.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update_status(new_stat)\n\n attrs = ActionController::Parameters.new({status: new_stat, req_to_del_at: nil})\n self.update_attributes(attrs.permit(Team::PERMIT_BASE))\n end",
"def update\n respond_to do |format|\n if @status.update(status_params)\n format.html { redirect_to @status, notice: 'Status was successfully updated.' }\n format.json { render :show, status: :ok, location: @status }\n else\n format.html { render :edit }\n format.json { render json: @status.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @status.update(status_params)\n format.html { redirect_to @status, notice: 'Status was successfully updated.' }\n format.json { render :show, status: :ok, location: @status }\n else\n format.html { render :edit }\n format.json { render json: @status.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @retweet = Retweet.find(params[:id])\n\n respond_to do |format|\n if @retweet.update_attributes(params[:retweet])\n format.html { redirect_to @retweet, notice: 'Retweet was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @retweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @tweet.update(tweet_params)\n respond_with(@tweet)\n end",
"def update\n respond_to do |format|\n if @mstatus.update(mstatus_params)\n format.html { redirect_to @mstatus, notice: 'Status was successfully updated.' }\n format.json { render :show, status: :ok, location: @mstatus }\n else\n format.html { render :edit }\n format.json { render json: @mstatus.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update_status\n update_params = {status: params[:status]}\n update_params[:password] = \"\" if params[:object] == \"User\"\n\n model = params[:object].constantize\n object = model.find(params[:status_id])\n if object.update_attributes(update_params)\n render json: {success:{msg: \"Updated #{params[:object]}\", id: object.id.to_s}}\n else \n render json: {failure:{msg: object.errors.full_messages.first}}\n end\n end",
"def update_status(payload, status)\n sha = payload.after\n repo = payload.repository.full_name\n state, description = status.first\n\n # setup http post\n uri = URI.parse(\"#{GITHUB_ROOT}/repos/#{repo}/statuses/#{sha}\")\n http = Net::HTTP.new(uri.host, uri.port)\n http.use_ssl = true\n\n # post to GitHub\n params = {:state => state, :description => description, :context => CONTEXT}\n http.post(uri.path, params.to_json, HEADERS)\nend",
"def update\n if can? :update, @status\n respond_to do |format|\n\n if @status.update(status_params)\n\n format.html { redirect_to @status, notice: 'Status was successfully updated.' }\n format.json { render :show, status: :ok, location: @status }\n else\n format.html { render :edit }\n format.json { render json: @status.errors, status: :unprocessable_entity }\n end\n\n end\n end\n end",
"def update\n respond_to do |format|\n if @tw_stat.update(tw_stat_params)\n format.html { redirect_to @tw_stat, notice: 'Tw stat was successfully updated.' }\n format.json { render :show, status: :ok, location: @tw_stat }\n else\n format.html { render :edit }\n format.json { render json: @tw_stat.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @tweet = Tweet.find(params[:id])\n\n respond_to do |format|\n if @tweet.update_attributes(params[:tweet])\n format.html { redirect_to @tweet, notice: 'Tweet was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @tweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @tweet = Tweet.find(params[:id])\n\n respond_to do |format|\n if @tweet.update_attributes(params[:tweet])\n format.html { redirect_to @tweet, notice: 'Tweet was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @tweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @tweet = Tweet.find(params[:id])\n\n respond_to do |format|\n if @tweet.update_attributes(params[:tweet])\n format.html { redirect_to @tweet, notice: 'Tweet was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @tweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @tweet.update_attributes(params[:tweet])\n format.html { redirect_to user_tweet_path(@user, @tweet), notice: 'Tweet was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @tweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if tweet.save\n format.html { redirect_to tweets_path, notice: 'Tweet was successfully updated.' }\n format.json { render :show, status: :ok, location: tweet }\n else\n format.html { render :edit }\n format.json { render json: tweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update_status\n \t update_user_book = UsersBook.find(params[:id])\n \t update_user_book.status = params[:status]\n\n \t render :json => {:status => \"success\"}\n end",
"def change_status\n @user = User.find(params[:data])\n\n respond_to do |format|\n if @user.update_attributes(status: params[:status])\n format.json { render json: @user }\n else\n format.json { render json: @user.errors.messages, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @male200mts_head.update(male200mts_head_params)\n format.html { redirect_to @male200mts_head, notice: 'Male200mts head was successfully updated.' }\n format.json { render :show, status: :ok, location: @male200mts_head }\n else\n format.html { render :edit }\n format.json { render json: @male200mts_head.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update!\n response = Tessellator::Fetcher.new.call('get', 'https://howamidoing-duckinator.herokuapp.com/status.json')\n @@statuses = JSON.parse(response.body)['statuses']\n end",
"def update\n respond_to do |format|\n if @swit_status.update(swit_status_params)\n format.html { redirect_to @swit_status, notice: 'Swit status was successfully updated.' }\n format.json { render :show, status: :ok, location: @swit_status }\n else\n format.html { render :edit }\n format.json { render json: @swit_status.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @relationship_status.update(relationship_status_params)\n format.html { redirect_to @relationship_status, notice: 'Relationship status was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: 'edit' }\n format.json { render json: @relationship_status.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @male_shot_put_head.update(male_shot_put_head_params)\n format.html { redirect_to @male_shot_put_head, notice: 'Male shot put head was successfully updated.' }\n format.json { render :show, status: :ok, location: @male_shot_put_head }\n else\n format.html { render :edit }\n format.json { render json: @male_shot_put_head.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @request_status = RequestStatus.find(params[:id])\n\n respond_to do |format|\n if @request_status.update_attributes(params[:request_status])\n format.html { redirect_to @request_status, notice: 'Request status was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @request_status.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @status_animal = StatusAnimal.find(params[:id])\n\n respond_to do |format|\n if @status_animal.update_attributes(params[:status_animal])\n format.html { redirect_to @status_animal, notice: 'Status animal was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @status_animal.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @mood.update_attributes(mood_params)\n format.html { redirect_to moods_path, notice: 'Mood was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @mood.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @reqstatus.update(reqstatus_params)\n format.html { redirect_to @reqstatus, notice: 'Reqstatus was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: 'edit' }\n format.json { render json: @reqstatus.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @retweet.update(tweet_params)\n format.html { redirect_to retweets_path, notice: 'Retweet was successfully updated.' }\n format.json { render :show, status: :ok, location: @retweet }\n else\n format.html { render :edit }\n format.json { render json: @retweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @rainbow = Rainbow.find(params[:id])\n\n respond_to do |format|\n if @rainbow.update_attributes(params[:rainbow])\n format.html { redirect_to @rainbow, notice: 'Rainbow was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @rainbow.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @remind.update(remind_params)\n format.html { redirect_to @remind, notice: 'Reminder was successfully updated.' }\n format.json { render :show, status: :ok, location: @remind }\n else\n format.html { render :edit }\n format.json { render json: @remind.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @annual_stat = AnnualStat.find(params[:id])\n\n respond_to do |format|\n if @annual_stat.update_attributes(params[:annual_stat])\n format.html { redirect_to @annual_stat, notice: 'Annual stat was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @annual_stat.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update_status!( status , in_reply_to_status_id = nil )\n\t\tif in_reply_to_status_id\n\t\t\tresponse = access_token.post('/statuses/update.json', { :status => status, :in_reply_to_status_id => in_reply_to_status_id })\n\t\telse\n\t\t\tresponse = access_token.post('/statuses/update.json', { :status => status })\n\t\tend\n\t\tcase response\n\t\twhen Net::HTTPSuccess\n\t\t\tmessage=JSON.parse(response.body)\n\t\t\traise TwitterOauth::UnexpectedResponse unless message.is_a? Hash\n\t\t\tmessage\n\t\telse\n\t\t\traise TwitterOauth::APIError\n\t\tend\n\trescue => err\n\t\tputs \"Exception in update_status!: #{err}\"\n\t\traise err\n\tend",
"def update\r\n respond_to do |format|\r\n if @tweet.update(tweet_params)\r\n format.html { redirect_to @tweet}\r\n format.json { head :no_content }\r\n else\r\n format.html { render action: 'edit' }\r\n format.json { render json: @tweet.errors, status: :unprocessable_entity }\r\n end\r\n end\r\n end",
"def update\n @verb = Verb.find(params[:id])\n\n if @verb.update(verb_params)\n head :no_content\n else\n render json: @verb.errors, status: :unprocessable_entity\n end\n end",
"def update\n @story_status = StoryStatus.find(params[:id])\n\n respond_to do |format|\n if @story_status.update_attributes(params[:story_status])\n format.html { redirect_to @story_status, :notice => 'Story status was successfully updated.' }\n format.json { head :ok }\n else\n format.html { render :action => \"edit\" }\n format.json { render :json => @story_status.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def update\n @fred = Fred.find(params[:id])\n\n respond_to do |format|\n if @fred.update_attributes(params[:fred])\n format.html { redirect_to @fred, notice: 'Fred was successfully updated.' }\n format.json { head :ok }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @fred.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @mood.update(mood_params)\n format.html { redirect_to moods_path, notice: \"Mood was successfully updated.\" }\n format.json { render :show, status: :ok, location: @mood }\n else\n format.html { render :edit, status: :unprocessable_entity }\n format.json { render json: @mood.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n @tweet.assign_attributes(tweet_params)\n @tweet.uuid = session[:uid]\n if @tweet.update(tweet_params)\n format.html { redirect_to @tweet, notice: 'Tweet was successfully updated.' }\n format.json { render :show, status: :ok, location: @tweet }\n else\n format.html { render :edit }\n format.json { render json: @tweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @appointment_status.update(appointment_status_params)\n format.html { redirect_to @appointment_status, notice: 'Appointment status was successfully updated.' }\n format.json { render :show, status: :ok, location: @appointment_status }\n else\n format.html { render :edit }\n format.json { render json: @appointment_status.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update_by_body\n @person = Person.find(person_update_params[:id])\n\n if @person.update_attributes(person_update_params)\n render json: { status: 'PUT Success' }, status: :ok\n else\n render json: { status: 'Error', message:'Error updating person', person: @person.errors }, status: :unprocessable_entity\n end\n end",
"def postTweet(status)\n\t\t\t@client.update(status)\n\t\tend",
"def create\n @married_status = MarriedStatus.new(married_status_params)\n\n respond_to do |format|\n if @married_status.save\n format.html { redirect_to @married_status, notice: 'Married status was successfully created.' }\n format.json { render :show, status: :created, location: @married_status }\n else\n format.html { render :new }\n format.json { render json: @married_status.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @master_user_status.update(master_user_status_params)\n format.html { redirect_to @master_user_status}\n format.json { head :no_content }\n else\n format.html { render action: 'edit' }\n format.json { render json: @master_user_status.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n # binding.pry\n @ringtone = Ringtone.find(params[:id])\n\n respond_to do |format|\n if @ringtone.update_attributes(params[:ringtone])\n format.html { redirect_to @ringtone, notice: 'Ringtone was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @ringtone.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @stat = Stat.find(params[:id])\n\n if @stat.update(stat_params)\n head :no_content\n else\n render json: @stat.errors, status: :unprocessable_entity\n end\n end",
"def update\n respond_to do |format|\n if @mood.update(mood_params)\n format.html { redirect_to @mood, notice: 'Mood was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: 'edit' }\n format.json { render json: @mood.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @tweet.update(tweet_params)\n format.html { redirect_to @tweet, notice: \"Tweet was successfully updated.\" }\n format.json { render :show, status: :ok, location: @tweet }\n else\n format.html { render :edit, status: :unprocessable_entity }\n format.json { render json: @tweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @tweet.update(tweet_params)\n format.html { redirect_to @tweet, notice: \"Tweet was successfully updated.\" }\n format.json { render :show, status: :ok, location: @tweet }\n else\n format.html { render :edit, status: :unprocessable_entity }\n format.json { render json: @tweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @tweet.update(tweet_params)\n format.html { redirect_to @tweet, notice: \"Tweet was successfully updated.\" }\n format.json { render :show, status: :ok, location: @tweet }\n else\n format.html { render :edit, status: :unprocessable_entity }\n format.json { render json: @tweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @barrel_status.update(barrel_status_params)\n format.html { redirect_to @barrel_status, notice: 'Barrel status was successfully updated.' }\n format.json { render :show, status: :ok, location: @barrel_status }\n else\n format.html { render :edit }\n format.json { render json: @barrel_status.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @bookmark_stat.update_attributes(params[:bookmark_stat])\n format.html { redirect_to @bookmark_stat, :notice => t('controller.successfully_updated', :model => t('activerecord.models.bookmark_stat')) }\n format.json { head :no_content }\n else\n format.html { render :action => \"edit\" }\n format.json { render :json => @bookmark_stat.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @member_status.update(member_status_params)\n format.html { redirect_to @member_status, notice: 'Member status was successfully updated.' }\n format.json { render :show, status: :ok, location: @member_status }\n else\n format.html { render :edit }\n format.json { render json: @member_status.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @tweet.update(tweet_params)\n format.html { redirect_to @tweet }\n format.json { render :show, status: :ok, location: @tweet }\n else\n format.html { render :edit, status: :unprocessable_entity }\n format.json { render json: @tweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @twitter_user = TwitterUser.find(params[:id])\n\n respond_to do |format|\n if @twitter_user.update_attributes(params[:twitter_user])\n format.html { redirect_to @twitter_user, notice: 'Twitter user was successfully updated.' }\n format.json { head :ok }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @twitter_user.errors, status: :unprocessable_entity }\n end\n end\n end",
"def index\n @married_statuses = MarriedStatus.all\n end",
"def update\n respond_to do |format|\n if @tweet.update(tweet_params)\n format.html { redirect_to @tweet }\n format.json { render :show, status: :ok, location: @tweet }\n else\n format.html { render :edit }\n format.json { render json: @tweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\r\n respond_to do |format|\r\n if @tweet.update(tweet_params)\r\n format.html { redirect_to @tweet, notice: \"Tweet was successfully updated.\" }\r\n format.json { render :show, status: :ok, location: @tweet }\r\n else\r\n format.html { render :edit, status: :unprocessable_entity }\r\n format.json { render json: @tweet.errors, status: :unprocessable_entity }\r\n end\r\n end\r\n end",
"def update\n @medication = Medication.find_by_id(params[:id])\n @user_id = @medication.user_id\n\n if @medication.nil?\n send_error_json(nil, \"Param 'medication' missing\", 400)\n return\n end\n\n if !check_owner()\n send_error_json(@medication.id, \"Unauthorized\", 403)\n return\n end\n\n fav = true\n if params['medication'].nil? || params['medication']['favourite'].nil? || params['medication']['favourite']=='false'\n fav = false\n end\n update_hash = {:favourite => fav}\n\n if params['medication'] && params['medication']['amount']\n update_hash[:amount] = params['medication']['amount'].to_i\n end\n if params['medication'] && params['medication']['medication_type_id']\n mt = MedicationType.find_by_id(params['medication']['medication_type_id'].to_i)\n if !mt.nil?\n update_hash[:medication_type_id] = mt.id\n else\n send_error_json(nil, \"Invalid medication_type_id\", 400)\n return\n end\n end\n\n if @medication.update_attributes(update_hash)\n send_success_json(@medication.id, {:msg => \"Updated successfully\"})\n else\n send_error_json(@medication.id, \"Update error\", 400)\n end\n\n end",
"def update\n respond_to do |format|\n if @tweet.update(tweet_params)\n format.html { redirect_to @tweet, notice: 'Tweet was successfully updated.' }\n format.json { render :show, status: :ok, location: @tweet }\n else\n format.html { render :edit }\n format.json { render json: @tweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @tweet.update(tweet_params)\n format.html { redirect_to @tweet, notice: 'Tweet was successfully updated.' }\n format.json { render :show, status: :ok, location: @tweet }\n else\n format.html { render :edit }\n format.json { render json: @tweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @tweet.update(tweet_params)\n format.html { redirect_to @tweet, notice: 'Tweet was successfully updated.' }\n format.json { render :show, status: :ok, location: @tweet }\n else\n format.html { render :edit }\n format.json { render json: @tweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @tweet.update(tweet_params)\n format.html { redirect_to @tweet, notice: 'Tweet was successfully updated.' }\n format.json { render :show, status: :ok, location: @tweet }\n else\n format.html { render :edit }\n format.json { render json: @tweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @custom_vital.update(custom_vital_params)\n format.html { redirect_to @custom_vital, notice: 'Custom vital was successfully updated.' }\n format.json { render :show, status: :ok, location: @custom_vital }\n else\n format.html { render :edit }\n format.json { render json: @custom_vital.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @status = Status.find(params[:id])\n\n respond_to do |format|\n if @status.update_attributes(params[:status])\n flash[:notice] = 'Status Type was successfully updated.'\n format.html { redirect_to(@status) }\n format.xml { head :ok }\n else\n format.html { render :action => \"edit\" }\n format.xml { render :xml => @status.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def update\n @red_flag = RedFlag.find(params[:id])\n\n respond_to do |format|\n if @red_flag.update_attributes(params[:red_flag])\n format.html { redirect_to @red_flag, notice: 'Red flag was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @red_flag.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @customized_concept.update(customized_concept_params)\n format.html { redirect_to @customized_concept, notice: 'Customized concept was successfully updated.' }\n format.json { render :show, status: :ok, location: @customized_concept }\n else\n format.html { render :edit }\n format.json { render json: @customized_concept.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @maritalinfo.update(maritalinfo_params)\n format.html { redirect_to @maritalinfo, notice: 'Maritalinfo was successfully updated.' }\n format.json { render :show, status: :ok, location: @maritalinfo }\n else\n format.html { render :edit }\n format.json { render json: @maritalinfo.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @status = Status.find(params[:id])\n\n respond_to do |format|\n if @status.update_attributes(params[:status])\n flash[:notice] = 'Status was successfully updated.'\n format.html { redirect_to(@status) }\n format.xml { head :ok }\n else\n format.html { render :action => \"edit\" }\n format.xml { render :xml => @status.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def update\n @irregular_verb = IrregularVerb.find(params[:id])\n\n respond_to do |format|\n if @irregular_verb.update_attributes(params[:irregular_verb])\n format.html { redirect_to @irregular_verb, notice: 'Irregular verb was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @irregular_verb.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @verb.update(verb_params)\n format.html { redirect_to @verb, notice: 'Verb was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: 'edit' }\n format.json { render json: @verb.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @status = @object\n\n respond_to do |format|\n if @status.update_attributes(params[:status])\n flash[:notice] = 'Status was successfully updated.'\n format.html { redirect_to status_url(@status) }\n format.xml { head :ok }\n else\n format.html { render :action => \"edit\" }\n format.xml { render :xml => @status.errors.to_xml, :status => :unprocessable_entity }\n end\n end\n end",
"def update_status(user_id:, status:)\n path = '/users/{userId}/status'\n .gsub('{userId}', user_id)\n\n if user_id.nil?\n raise Appwrite::Exception.new('Missing required parameter: \"userId\"')\n end\n\n if status.nil?\n raise Appwrite::Exception.new('Missing required parameter: \"status\"')\n end\n\n params = {\n status: status,\n }\n \n headers = {\n \"content-type\": 'application/json',\n }\n\n @client.call(\n method: 'PATCH',\n path: path,\n headers: headers,\n params: params,\n response_type: Models::User\n )\n end",
"def update\r\n @reminder = Reminder.find(params[:id])\r\n\r\n respond_to do |format|\r\n if @reminder.update_attributes(params[:reminder])\r\n format.html { redirect_to @reminder, notice: 'Reminder was successfully updated.' }\r\n format.json { head :no_content }\r\n else\r\n format.html { render action: \"edit\" }\r\n format.json { render json: @reminder.errors, status: :unprocessable_entity }\r\n end\r\n end\r\n end",
"def update\n @lost = Lost.find(params[:id])\n\n respond_to do |format|\n if @lost.update_attributes(params[:lost])\n format.html { redirect_to @lost, notice: 'Lost was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @lost.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n begin\n Twitter.user_timeline(params[:foodtruck][:handle])\n foodtruck = Foodtruck.find(params[:id])\n foodtruck.update_attributes(params[:foodtruck])\n redirect_to '/foodtrucks'\n rescue\n redirect_to \"/foodtrucks/#{params[:id]}\"\n end\n end",
"def update\n respond_to do |format|\n if @new_tweet.update(new_tweet_params)\n format.html { redirect_to @new_tweet, notice: 'New tweet was successfully updated.' }\n format.json { render :show, status: :ok, location: @new_tweet }\n else\n format.html { render :edit }\n format.json { render json: @new_tweet.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @health_status.update(health_status_params)\n format.html { redirect_to @health_status, notice: 'Health status was successfully updated.' }\n format.json { render :show, status: :ok, location: @health_status }\n else\n format.html { render :edit }\n format.json { render json: @health_status.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @todo = Todo.find(params[:id])\n if @todo.update_attributes(todo_params)\n render json: @todo, status: :ok\n else\n render json: @todo.errors, status: 422\n end\n end",
"def update\n if :opr == 'edit'\n update\n else\n @team = Team.find_by_id(params[:id])\n @team.update_attributes({:id => params[:id], :name => params[:name], :status => params[:status]})\n\n if request.xhr?\n render :json => @team\n end\n end\n end",
"def update\n respond_to do |format|\n if @status_request.update(status_request_params)\n format.html { redirect_to @status_request, notice: 'Status request was successfully updated.' }\n format.json { render :show, status: :ok, location: @status_request }\n else\n format.html { render :edit }\n format.json { render json: @status_request.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @invite_status = InviteStatus.find(params[:id])\n\n respond_to do |format|\n if @invite_status.update_attributes(params[:invite_status])\n format.html { redirect_to @invite_status, notice: 'Invite status was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @invite_status.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\r\n @status_task = StatusTask.find(params[:id])\r\n\r\n respond_to do |format|\r\n if @status_task.update_attributes(params[:status_task])\r\n format.html { redirect_to @status_task, notice: 'Status task was successfully updated.' }\r\n format.json { head :no_content }\r\n else\r\n format.html { render action: \"edit\" }\r\n format.json { render json: @status_task.errors, status: :unprocessable_entity }\r\n end\r\n end\r\n end",
"def update\n @stat = Stat.find(params[:id])\n\n respond_to do |format|\n if @stat.update_attributes(params[:stat])\n format.html { redirect_to @stat, :notice => 'Stat was successfully updated.' }\n format.json { head :ok }\n else\n format.html { render :action => \"edit\" }\n format.json { render :json => @stat.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @entity_status.update(entity_status_params)\n format.html { redirect_to entity_statuses_path, notice: 'Entity status was successfully updated.' }\n format.json { render :show, status: :ok, location: @entity_status }\n else\n format.html { render :edit }\n format.js { render :edit }\n format.json { render json: @entity_status.errors, status: :unprocessable_entity }\n end\n end\n end"
] |
[
"0.6699571",
"0.6523629",
"0.65057445",
"0.6272926",
"0.62283516",
"0.6162448",
"0.6126403",
"0.6071017",
"0.6068237",
"0.60560423",
"0.604338",
"0.60276586",
"0.60276586",
"0.6022956",
"0.6012158",
"0.5965091",
"0.59637815",
"0.593372",
"0.5921117",
"0.5921117",
"0.59066504",
"0.590588",
"0.58869326",
"0.5882511",
"0.5881693",
"0.5866379",
"0.5832283",
"0.5824587",
"0.5824587",
"0.5824587",
"0.58115524",
"0.5799857",
"0.5793853",
"0.5791157",
"0.5780189",
"0.57799286",
"0.57766134",
"0.57744104",
"0.5761605",
"0.5747767",
"0.574203",
"0.5737717",
"0.5733381",
"0.5723069",
"0.5721382",
"0.571122",
"0.5704356",
"0.5700698",
"0.5678707",
"0.5676911",
"0.56583303",
"0.5645392",
"0.56389225",
"0.5628723",
"0.56279135",
"0.562574",
"0.5623688",
"0.5620388",
"0.5611689",
"0.56105167",
"0.56073415",
"0.56055856",
"0.56026447",
"0.56026447",
"0.56026447",
"0.5602129",
"0.5596328",
"0.5595925",
"0.5594933",
"0.55784595",
"0.5576719",
"0.5576103",
"0.55757606",
"0.5566764",
"0.5563395",
"0.5563395",
"0.5563395",
"0.5563395",
"0.5562994",
"0.55622214",
"0.55564433",
"0.5556383",
"0.55532646",
"0.5551514",
"0.5550804",
"0.55500245",
"0.5544605",
"0.55441004",
"0.55424607",
"0.5540787",
"0.55404615",
"0.5536253",
"0.5533776",
"0.55302495",
"0.5527806",
"0.55257505",
"0.55185866",
"0.55145395",
"0.55135643",
"0.55135083"
] |
0.7180156
|
0
|
DELETE /married_statuses/1 DELETE /married_statuses/1.json
|
def destroy
@married_status.destroy
respond_to do |format|
format.html { redirect_to married_statuses_url, notice: 'Married status was successfully destroyed.' }
format.json { head :no_content }
end
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def destroy_status(status_id)\n delete \"statuses/destroy/#{status_id}\"\n end",
"def delete_status(id)\n delete(\"/statuses/#{id}\")\n end",
"def destroy\n @status.destroy\n respond_to do |format|\n format.html { redirect_to statuses_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @roof = Roof.find(params[:roof_id])\n @status = @roof.statuses.find(params[:id])\n @status.destroy\n\n respond_to do |format|\n format.html { redirect_to statuseses_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @current_statuses.destroy\n respond_to do |format|\n format.html { redirect_to current_statuses_url, notice: 'Current Statuses was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @status = Status.find(params[:id])\n @status.destroy\n\n respond_to do |format|\n format.html { redirect_to(statuses_url) }\n format.xml { head :ok }\n end\n end",
"def destroy\n @new_status = NewStatus.find(params[:id])\n @new_status.destroy\n\n respond_to do |format|\n format.html { redirect_to new_statuses_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @reqstatus.destroy\n respond_to do |format|\n format.html { redirect_to reqstatuses_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @kf_status = Kf::Status.find(params[:id])\n @kf_status.destroy\n\n respond_to do |format|\n format.html { redirect_to \"/kf/statuses?page=#{params[:page]}&relation_id=#{params[:relation_id]}&status_type=#{params[:status_type]}&count_type=#{params[:count_type]}\" }\n format.json { head :no_content }\n end\n end",
"def destroy\n @status = Status.find(params[:id])\n @status.destroy\n flash[:notice] = 'Status deleted'\n \n respond_to do |format|\n format.html { redirect_to(statuses_url) }\n format.xml { head :ok }\n end\n end",
"def destroy\n @mstatus.destroy\n respond_to do |format|\n format.html { redirect_to mstatuses_url, notice: 'Status was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @request_status = RequestStatus.find(params[:id])\n @request_status.destroy\n\n respond_to do |format|\n format.html { redirect_to request_statuses_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @status = Status.find(params[:id])\n @status.destroy\n\n redirect_to statuses_url\n end",
"def destroy\n @status.destroy\n respond_to do |format|\n format.html { redirect_to statuses_url, notice: 'Status was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @status.destroy\n respond_to do |format|\n format.html { redirect_to statuses_url, notice: 'Status was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @master_user_status.destroy\n respond_to do |format|\n format.html { redirect_to master_user_statuses_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @story_status = StoryStatus.find(params[:id])\n @story_status.destroy\n\n respond_to do |format|\n format.html { redirect_to story_statuses_url }\n format.json { head :ok }\n end\n end",
"def destroy\n @swit_status.destroy\n respond_to do |format|\n format.html { redirect_to swit_statuses_url, notice: 'Swit status was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @status.destroy\n respond_to do |format|\n format.html { redirect_to statuses_url, notice: 'Status was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @relationship_status.destroy\n respond_to do |format|\n format.html { redirect_to relationship_statuses_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @nvs_mig_status = NvsMigStatus.find(params[:id])\n @nvs_mig_status.destroy\n\n respond_to do |format|\n format.html { redirect_to nvs_mig_statuses_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @rh21_status.destroy\n respond_to do |format|\n format.html { redirect_to rh21_statuses_url, notice: 'Rh21 status was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @status = Status.find(params[:id])\n @status.destroy\n\n respond_to do |format|\n format.html { redirect_to(statuss_url) }\n format.xml { head :ok }\n end\n end",
"def destroy\n @status = Status.find(params[:id])\n @status.destroy\n\n respond_to do |format|\n format.html { redirect_to profile_path(current_user) }\n format.json { head :no_content }\n end\n end",
"def destroy\n @barrel_status.destroy\n respond_to do |format|\n format.html { redirect_to barrel_statuses_url, notice: 'Barrel status was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @status_animal = StatusAnimal.find(params[:id])\n @status_animal.destroy\n\n respond_to do |format|\n format.html { redirect_to status_animais_url }\n format.json { head :no_content }\n end\n end",
"def delete(project_token = @project_token, id = @id, user = @@default_user)\n @attributes = send_request(\"statuses/#{id}\", :delete) do |req|\n req.body = {\n token: project_token,\n auth_token: user.auth_token\n }\n end\n end",
"def destroy\n @twitter_id = TwitterId.find(params[:id])\n @twitter_id.destroy\n\n respond_to do |format|\n format.html { redirect_to twitter_ids_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n if can? :delete, @status\n @status.destroy\n respond_to do |format|\n\n format.html { redirect_to statuses_url, notice: 'Status was successfully destroyed.' }\n format.json { head :no_content }\n end\n end\n end",
"def destroy\n @contaminationstatus.destroy\n respond_to do |format|\n format.html { redirect_to contaminationstatuses_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @sa_request_status.destroy\n respond_to do |format|\n format.html { redirect_to sa_request_statuses_url, notice: 'Sa request status was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n respond_to do |format|\n begin\n @tweet = TweetApp::ClientContext.status(:delete, params[:id])\n flash[:message] = \"Tweet with id #{params[:id]} was deleted from Twitter\"\n format.html { redirect_to tweets_url }\n format.json { head :ok }\n format.xml { head :ok }\n rescue Twitter::RESTError => re\n handle_rest_error(re, format)\n end\n end\n end",
"def destroy\n @reqdevstatus.destroy\n respond_to do |format|\n format.html { redirect_to reqdevstatuses_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @itemstatus = Itemstatus.find(params[:id])\n @itemstatus.destroy\n\n respond_to do |format|\n format.html { redirect_to itemstatuses_url }\n format.json { head :ok }\n end\n end",
"def destroy\n @retweet = Retweet.find(params[:id])\n @retweet.destroy\n\n respond_to do |format|\n format.html { redirect_to retweets_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @tw_stat.destroy\n respond_to do |format|\n format.html { redirect_to tw_stats_url, notice: 'Tw stat was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n user = user_from_token\n user.tweets.destroy(params[:id])\n head :no_content\n end",
"def destroy\n @status_record = StatusRecord.find(params[:id])\n @status_record.destroy\n\n respond_to do |format|\n format.html { redirect_to status_records_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @mood.destroy\n\n respond_to do |format|\n format.html { redirect_to moods_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @mood.destroy\n respond_to do |format|\n format.html { redirect_to moods_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @typeofstatus.destroy\n respond_to do |format|\n format.html { redirect_to typeofstatuses_url, notice: 'Typeofstatus was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @posts_status.destroy\n respond_to do |format|\n format.html { redirect_to posts_statuses_url, notice: 'Status was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @searchstatus = Searchstatus.find(params[:id])\n @searchstatus.destroy\n\n respond_to do |format|\n format.html { redirect_to searchstatuses_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @member_status.destroy\n respond_to do |format|\n format.html { redirect_to member_statuses_url, notice: 'Member status was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @record_status = RecordStatus.find(params[:id])\n @record_status.destroy\n\n respond_to do |format|\n format.html { redirect_to record_statuses_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n begin\n @marital_status.destroy\n flash[:success] = \"successfully destroyed.\"\n rescue ActiveRecord::DeleteRestrictionError => e\n @marital_status.errors.add(:base, e)\n flash[:error] = \"#{e}\"\n redirect_to marital_status_url\n else\n redirect_to marital_statuses_url\n end\n end",
"def destroy\r\n death = []\r\n death << Retweet.where(tweet_id: @tweet.id)\r\n death.flatten!\r\n death.each do |die|\r\n die.destroy\r\n end\r\n\r\n @tweet.destroy\r\n\r\n respond_to do |format|\r\n format.html { redirect_to tweets_url }\r\n format.json { head :no_content }\r\n end\r\n end",
"def destroy\n @agent_status = AgentStatus.find(params[:id])\n @agent_status.destroy\n\n respond_to do |format|\n format.html { redirect_to agent_statuses_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @active_status = ActiveStatus.find(params[:id])\n @active_status.destroy\n\n respond_to do |format|\n format.html { redirect_to(active_statuses_url) }\n format.xml { head :ok }\n end\n end",
"def destroy\n @status_ativ = StatusAtiv.find(params[:id])\n @status_ativ.destroy\n\n respond_to do |format|\n format.html { redirect_to status_ativs_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @invite_status = InviteStatus.find(params[:id])\n @invite_status.destroy\n\n respond_to do |format|\n format.html { redirect_to invite_statuses_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @payumoney_status.destroy\n respond_to do |format|\n format.html { redirect_to payumoney_statuses_url, notice: 'Payumoney status was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @status_request.destroy\n respond_to do |format|\n format.html { redirect_to status_requests_url, notice: 'Status request was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @tweet.destroy\n\n respond_to do |format|\n format.html { redirect_to tweets_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @tweet.destroy\n respond_to do |format|\n format.html { redirect_to tweets_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @service_status.destroy\n respond_to do |format|\n format.html { redirect_to service_statuses_url, notice: 'Service status was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @tweet = Tweet.find(params[:id])\n @tweet.destroy\n\n respond_to do |format|\n format.html { redirect_to tweets_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @tweet = Tweet.find(params[:id])\n @tweet.destroy\n\n respond_to do |format|\n format.html { redirect_to tweets_url }\n format.json { head :no_content }\n end\n end",
"def delete; update(:status => 'DELETED'); end",
"def destroy\n @status.destroy\n respond_to do |format|\n format.html { redirect_to statuses_url, notice: 'Status was successfully destroyed.' }\n format.json { head :no_content }\n format.js\n end\n end",
"def delete()\n\n client.delete(\"/project_statuses/#{gid}\") && true\n end",
"def destroy\n @medium_status_mod = MediumStatusMod.find(params[:id])\n @medium_status_mod.destroy\n\n respond_to do |format|\n format.html { redirect_to medium_status_mods_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @status = @object\n @status.destroy\n flash[:error] = @status.errors.on_base unless @status.errors.empty?\n\n respond_to do |format|\n format.html { redirect_to statuses_url }\n format.xml { head :ok }\n end\n end",
"def destroy\n @health_status.destroy\n respond_to do |format|\n format.html { redirect_to health_statuses_url, notice: 'Health status was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @technician_status.destroy\n respond_to do |format|\n format.html { redirect_to technician_statuses_url, notice: 'Technician status was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @store_status.destroy\n respond_to do |format|\n format.html { redirect_to store_statuses_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @dish_mood.destroy\n respond_to do |format|\n format.html { redirect_to dish_moods_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @reward_and_level.destroy\n respond_to do |format|\n format.html { redirect_to reward_and_levels_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @application_status = ApplicationStatus.find(params[:id])\n @application_status.destroy\n\n respond_to do |format|\n format.html { redirect_to application_statuses_url }\n format.json { head :no_content }\n end\n end",
"def delete_status(status_id)\n\t\t\tdata = oauth_request(\"/user_status/destroy/#{status_id}\", {}, \"post\")\n\t\tend",
"def destroy\n @rainbow = Rainbow.find(params[:id])\n @rainbow.destroy\n\n respond_to do |format|\n format.html { redirect_to rainbows_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @moresmalltrial = Moresmalltrial.find(params[:id])\n @moresmalltrial.destroy\n\n respond_to do |format|\n format.html { redirect_to moresmalltrials_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @red.destroy\n respond_to do |format|\n format.html { redirect_to reds_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @invoice_status = InvoiceStatus.find(params[:id])\n @invoice_status.destroy\n\n respond_to do |format|\n format.html { redirect_to invoice_statuses_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n tweet.destroy\n respond_to do |format|\n format.html { redirect_to tweets_path, notice: 'Tweet was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @appointment_status.destroy\n respond_to do |format|\n format.html { redirect_to appointment_statuses_url, notice: 'Appointment status was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @male_shot_put_head.destroy\n respond_to do |format|\n format.html { redirect_to male_shot_put_heads_url, notice: 'Male shot put head was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @observance.destroy\n respond_to do |format|\n format.html { redirect_to observances_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @status2.destroy\n respond_to do |format|\n format.html { redirect_to status2s_url, notice: 'Status2 was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def delete_dm(settings, dmid)\n Net::HTTP.start('twitter.com') {|http|\n req = Net::HTTP::Post.new(\"/direct_messages/destroy/#{dmid}.json\")\n req.basic_auth settings[:user], settings[:pass]\n response = http.request(req)\n response.body\n }\nend",
"def destroy\n @moodboard.destroy\n respond_to do |format|\n format.html { redirect_to moodboards_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @annual_stat = AnnualStat.find(params[:id])\n @annual_stat.destroy\n\n respond_to do |format|\n format.html { redirect_to annual_stats_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @subcontractor_status.destroy\n respond_to do |format|\n format.html { redirect_to subcontractor_statuses_url, notice: 'Subcontractor status was successfully deleted.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @makrana_marble.destroy\n respond_to do |format|\n format.html { redirect_to makrana_marbles_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @male200mts_head.destroy\n respond_to do |format|\n format.html { redirect_to male200mts_heads_url, notice: 'Male200mts head was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\r\n @customerstatus.destroy\r\n respond_to do |format|\r\n format.html { redirect_to customerstatuses_url, notice: 'Customerstatus was successfully destroyed.' }\r\n format.json { head :no_content }\r\n end\r\n end",
"def destroy\n @status_item.destroy\n respond_to do |format|\n format.html { redirect_to status_items_url, notice: 'El estatus del articulo fue eliminado.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @status_akun.destroy\n respond_to do |format|\n format.html { redirect_to status_akun_index_url, notice: 'Status akun was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @learning_status.destroy\n respond_to do |format|\n format.html { redirect_to learning_statuses_url, notice: 'Learning status was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy_multiple\n Remind.destroy(params[:reminds])\n respond_to do |format|\n format.html { redirect_to :back, notice: 'Reminder(s) was deleted.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @civil_status = CivilStatus.find(params[:id])\n @civil_status.destroy\n\n respond_to do |format|\n format.html { redirect_to civil_statuses_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @stat = Stat.find(params[:id])\n @stat.destroy\n\n respond_to do |format|\n format.html { redirect_to stats_url }\n format.json { head :ok }\n end\n end",
"def destroy\n @actor_tweet.destroy\n respond_to do |format|\n format.html { redirect_to @tweet, notice: 'Actor tweet eliminado correctamente.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @purchase_item_status = PurchaseItemStatus.find(params[:id])\n @purchase_item_status.destroy\n\n respond_to do |format|\n format.html { redirect_to purchase_item_statuses_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @stat = Stat.find(params[:id])\n @stat.destroy\n\n respond_to do |format|\n format.html { redirect_to stats_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @entity_status.destroy\n respond_to do |format|\n format.html { redirect_to entity_statuses_url, notice: 'Entity status was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @status = Status.find(params[:id])\n if session[\"user_id\"] == @status.user_id\n @status.destroy\n end\n\n respond_to do |format|\n format.html { redirect_to statuses_url }\n # format.json { head :ok }\n format.js { render :action => 'destroy.js.coffee', :content_type => 'text/javascript'}\n end\n end",
"def destroy\n @mission_status.destroy\n respond_to do |format|\n format.html { redirect_to admin_mission_statuses_url }\n format.json { head :no_content }\n end\n end",
"def delete\n client.delete(\"/#{id}\")\n end",
"def destroy\n @interesting_tweet.destroy\n respond_to do |format|\n format.html { redirect_to interesting_tweets_url }\n format.json { head :no_content }\n end\n end"
] |
[
"0.7316664",
"0.7198055",
"0.7114044",
"0.7109347",
"0.7086772",
"0.707513",
"0.7041838",
"0.7009426",
"0.6996029",
"0.6927694",
"0.6922116",
"0.6827844",
"0.68117946",
"0.6797788",
"0.6797788",
"0.67199755",
"0.6717694",
"0.6704238",
"0.6693786",
"0.6670758",
"0.6669584",
"0.66482425",
"0.66475236",
"0.66372323",
"0.6631518",
"0.66273105",
"0.66105056",
"0.6606596",
"0.66049826",
"0.6598547",
"0.65985215",
"0.6595572",
"0.6581637",
"0.65743846",
"0.6568389",
"0.65654016",
"0.65645117",
"0.65416896",
"0.6532369",
"0.65221006",
"0.6509492",
"0.6505266",
"0.649463",
"0.6494522",
"0.6491024",
"0.6485117",
"0.64840925",
"0.64788204",
"0.6468163",
"0.6464778",
"0.64620644",
"0.64609957",
"0.64587915",
"0.6441092",
"0.64284235",
"0.6426855",
"0.6426079",
"0.6426079",
"0.6426019",
"0.6420104",
"0.6414151",
"0.6409154",
"0.6408896",
"0.64059234",
"0.6399474",
"0.6393944",
"0.63932085",
"0.63926226",
"0.6391141",
"0.6382847",
"0.63746184",
"0.6374336",
"0.6374124",
"0.63736445",
"0.6372938",
"0.63725823",
"0.6360551",
"0.6359926",
"0.6358732",
"0.63569677",
"0.6353644",
"0.6351755",
"0.6348745",
"0.63412464",
"0.6340157",
"0.6339785",
"0.6336609",
"0.63342595",
"0.6332051",
"0.63299114",
"0.6328375",
"0.6325505",
"0.63234824",
"0.6322872",
"0.6319304",
"0.63188344",
"0.63179046",
"0.63081056",
"0.63071954",
"0.6305145"
] |
0.7700158
|
0
|
Use callbacks to share common setup or constraints between actions.
|
def set_married_status
@married_status = MarriedStatus.find(params[:id])
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def set_required_actions\n # TODO: check what fields change to asign required fields\n end",
"def action_hook; end",
"def run_actions; end",
"def define_action_hook; end",
"def actions; end",
"def define_action_helpers\n if super && action == :save\n @instance_helper_module.class_eval do\n define_method(:valid?) do |*args|\n self.class.state_machines.fire_event_attributes(self, :save, false) { super(*args) }\n end\n end\n end\n end",
"def add_actions; end",
"def callbacks; end",
"def callbacks; end",
"def setup *actions, &proc\n (@setup_procs ||= []) << [proc, actions.size > 0 ? actions : [:*]]\n end",
"def define_action_helpers; end",
"def post_setup\n end",
"def action_methods; end",
"def action_methods; end",
"def action_methods; end",
"def before_setup; end",
"def action_run\n end",
"def execute(setup)\n @action.call(setup)\n end",
"def define_action_helpers?; end",
"def set_actions\n actions :all\n end",
"def action_done(action)\n dispatch = { :migrate => :done_migrating, :map => :done_mapping, :reduce =>\n :done_reducing, :finalize => :done_finalizing } \n self.send dispatch[action[:action]], action\n end",
"def dependencies action, &block\n @actions.each do |other|\n if action[:requires].include? other[:provide]\n block.call other\n end\n end\n end",
"def setup!\n return unless @setup_procs\n http_actions = actions\n @setup_procs.each do |setup_proc|\n proc, actions = setup_proc\n @setup__actions = actions.map do |action|\n\n action.is_a?(Regexp) ?\n http_actions.select { |a| a.to_s =~ action } :\n action.is_a?(String) && action =~ /\\A\\./ ?\n http_actions.map { |a| a.to_s << action if format?(a).include?(action) }.compact :\n action\n\n end.flatten\n self.class_exec &proc\n @setup__actions = nil\n end\n @setup_procs = nil\n end",
"def before_actions(*logic)\n self.before_actions = logic\n end",
"def setup_handler\n end",
"def set_action(opts)\n opts = check_params(opts,[:actions])\n super(opts)\n end",
"def setup(action)\n @targets.clear\n unless action.item.target_filters.empty?\n @targets = SES::TargetManager.make_targets(action)\n else\n item = action.item\n if item.for_opponent?\n @targets = $game_troop.alive_members\n elsif item.for_dead_friend?\n @targets = $game_party.battle_members.select { |actor| actor.dead? }\n else\n $game_party.battle_members.select { |actor| actor.alive? }\n end\n end\n @item_max = @targets.size\n create_contents\n refresh\n show\n activate\n end",
"def action; end",
"def action; end",
"def action; end",
"def action; end",
"def action; end",
"def workflow\n end",
"def revisable_shared_setup(args, block)\n class << self\n attr_accessor :revisable_options\n end\n options = args.extract_options!\n self.revisable_options = Options.new(options, &block)\n \n self.send(:include, Common)\n self.send(:extend, Validations) unless self.revisable_options.no_validation_scoping?\n self.send(:include, WithoutScope::QuotedColumnConditions)\n end",
"def setup\n @action = SampleActionAndroid.new(os_name: 'android',\n app_name: APP_PATH)\n end",
"def before(action)\n invoke_callbacks *self.class.send(action).before\n end",
"def process_action(...)\n send_action(...)\n end",
"def before_dispatch(env); end",
"def after_actions(*logic)\n self.after_actions = logic\n end",
"def setup\n # override and do something appropriate\n end",
"def setup(client)\n return unless @setup\n actions = @setup['setup'].select { |action| action['do'] }.map { |action| Action.new(action['do']) }\n actions.each do |action|\n action.execute(client)\n end\n self\n end",
"def setup(_context)\n end",
"def setup(resources) ; end",
"def validate_actions\n errors.add(:base, :should_give_at_least_one_action) if !manage? && !forecasting? && !read? && !api?\n end",
"def setup\n @resource_config = {\n :callbacks => {\n :before_create => nil,\n :after_create => nil,\n :before_update => nil,\n :after_update => nil,\n :before_destroy => nil,\n :after_destroy => nil,\n },\n :child_assoc => nil,\n :model => nil,\n :parent => nil,\n :path => nil,\n :permission => {},\n :properties => {},\n :relation => {\n :create => nil,\n :delete => nil,\n },\n :roles => nil,\n }\n end",
"def determine_valid_action\n\n end",
"def process_shared\n handle_taxes\n handle_shippings\n create_adjustments_from_params\n handle_status\n handle_inventory_refunds\n handle_payment_transactions\n order.updater.update\n end",
"def startcompany(action)\n @done = true\n action.setup\n end",
"def init_actions\n am = action_manager()\n am.add_action(Action.new(\"&Disable selection\") { @selection_mode = :none; unbind_key(32); bind_key(32, :scroll_forward); } )\n am.add_action(Action.new(\"&Edit Toggle\") { @edit_toggle = !@edit_toggle; $status_message.value = \"Edit toggle is #{@edit_toggle}\" })\n end",
"def event_callbacks(event, metadata={})\n case event\n when :reset, :review\n if confirmed\n update_attributes(confirmed: false)\n end\n when :confirm\n confirm\n # trigger :order for all applicable items\n # NOTE: :order event is common to both physical and digital items\n items.each do |i|\n if i.event_permitted(:order)\n user_id = last_transition.user_id\n i.trigger!(:order, { order_id: id, user_id: user_id })\n end\n end\n when :complete_work\n request = metadata[:request]\n work_complete_notification(request)\n when :close\n close\n end\n if event != :close && !open\n reopen\n end\n end",
"def setup_action\n return unless PONY::ERRNO::check_sequence(current_act)\n new_sequence = @action_sequence[@sequence_index+1...@action_sequence.size]\n @sequence_index = 0\n new_sequence = DND::SkillSequence::ACTS[@acts[1]] + new_sequence\n execute_sequence\n end",
"def define_tasks\n define_weave_task\n connect_common_tasks\n end",
"def setup(&block)\n define_method(:setup, &block)\n end",
"def setup\n transition_to(:setup)\n end",
"def setup\n transition_to(:setup)\n end",
"def action\n end",
"def setup( *args )\n\t\t\tself.class.setupBlocks.each {|sblock|\n\t\t\t\tdebugMsg \"Calling setup block method #{sblock}\"\n\t\t\t\tself.send( sblock )\n\t\t\t}\n\t\t\tsuper( *args )\n\t\tend",
"def config(action, *args); end",
"def setup\n @setup_proc.call(self) if @setup_proc\n end",
"def before_action \n end",
"def setup_callbacks\n defined_callbacks.each do |meth|\n unless respond_to?(\"call_#{meth}_callbacks\".to_sym)\n self.class.module_eval <<-EOE\n def call_#{meth}_callbacks(*args)\n plugin_store.each {|a| a.call_#{meth}_callbacks(*args) } if respond_to?(:plugin_store) && plugin_store\n self.send :#{meth}, *args if respond_to?(:#{meth})\n end\n EOE\n end\n end\n end",
"def action\n end",
"def matt_custom_action_begin(label); end",
"def setup\n # override this if needed\n end",
"def setup\n\t\t\t\t\t\t# Do nothing\n\t\t\t\tend",
"def setup\n\t\t\t\t\t\t# Do nothing\n\t\t\t\tend",
"def action(options,&callback)\n new_action = Action===options ? options : Action.new(options,&callback)\n # replace any with (shared name/alias or both default) + same arity\n @actions.delete_if do |existing_action|\n ((existing_action.names & new_action.names).size > 0 ||\n existing_action.default? && new_action.default?) &&\n existing_action.required.size == new_action.required.size &&\n existing_action.optional.size <= new_action.optional.size\n end\n @actions = (@actions + [new_action]).sort\n new_action\n end",
"def set_target_and_action target, action\n self.target = target\n self.action = 'sugarcube_handle_action:'\n @sugarcube_action = action\n end",
"def after(action)\n invoke_callbacks *options_for(action).after\n end",
"def pre_task\n end",
"def setup(server)\n server.on('beforeMethod', method(:before_method), 10)\n end",
"def add_actions\n attribute = machine.attribute\n name = self.name\n \n owner_class.class_eval do\n define_method(name) {self.class.state_machines[attribute].events[name].fire(self)}\n define_method(\"#{name}!\") {self.class.state_machines[attribute].events[name].fire!(self)}\n define_method(\"can_#{name}?\") {self.class.state_machines[attribute].events[name].can_fire?(self)}\n end\n end",
"def init_actions\n @select_action = SelectAction.new\n @endpoint_mouse_action = EndpointMouseAction.new\n @move_action = MoveAction.new\n end",
"def setup_signals; end",
"def after_created\r\n return unless compile_time\r\n Array(action).each do |action|\r\n run_action(action)\r\n end\r\nend",
"def after_created\r\n return unless compile_time\r\n Array(action).each do |action|\r\n run_action(action)\r\n end\r\nend",
"def set_target_and_action target, action\n self.target = target\n self.action = 'sugarcube_handle_action:'\n @sugarcube_action = action.respond_to?('weak!') ? action.weak! : action\n end",
"def initialize(*args)\n super\n @action = :set\nend",
"def after_set_callback; end",
"def setup\n #implement in subclass;\n end",
"def lookup_action; end",
"def setup &block\n if block_given?\n @setup = block\n else\n @setup.call\n end\n end",
"def setup_action\n return TSBS.error(@acts[0], 1, @used_sequence) if @acts.size < 2\n actions = TSBS::AnimLoop[@acts[1]]\n if actions.nil?\n show_action_error(@acts[1])\n end\n @sequence_stack.push(@acts[1])\n @used_sequence = @acts[1]\n actions.each do |acts|\n @acts = acts\n execute_sequence\n break if @break_action\n end\n @sequence_stack.pop\n @used_sequence = @sequence_stack[-1]\n end",
"def release_actions; end",
"def around_hooks; end",
"def save_action; end",
"def setup(easy)\n super\n easy.customrequest = @verb\n end",
"def action_target()\n \n end",
"def setup\n callback(:setup) do\n notify(:setup)\n migration_check.last_deployed_commit\n end\n end",
"def setup\n return unless @setup\n\n actions = @setup['setup'].select { |action| action['do'] }.map { |action| Action.new(action['do']) }\n run_actions_and_retry(actions)\n self\n end",
"def before_setup\n # do nothing by default\n end",
"def my_actions(options)\n @setup = false\n get_template_part(\"custom_used\",\"action_users\",true)\n end",
"def default_action; end",
"def setup(&blk)\n @setup_block = blk\n end",
"def callback_phase\n super\n end",
"def advice\n end",
"def _handle_action_missing(*args); end",
"def duas1(action)\n action.call\n action.call\nend",
"def shared_action(name, &block)\n @controller.shared_actions[name] = block\n end",
"def before_action action, &block\n @audience[:before][action] ||= Set.new\n @audience[:before][action] << block\n end",
"def setup_initial_state\n\n state_a = State.new(\"a\", 0)\n state_b = State.new(\"b\", 0)\n state_c = State.new(\"c\", 10)\n\n move_to_b = Action.new(\"move_to_b\", 1, state_b)\n\n move_to_c = Action.new(\"move_to_c\", 1, state_c)\n\n state_a.actions = [move_to_b, move_to_c]\n\n return state_a\n \nend"
] |
[
"0.6163163",
"0.6045976",
"0.5946146",
"0.591683",
"0.5890051",
"0.58349305",
"0.5776858",
"0.5703237",
"0.5703237",
"0.5652805",
"0.5621621",
"0.54210985",
"0.5411113",
"0.5411113",
"0.5411113",
"0.5391541",
"0.53794575",
"0.5357573",
"0.53402257",
"0.53394014",
"0.53321576",
"0.53124547",
"0.529654",
"0.5296262",
"0.52952296",
"0.52600986",
"0.52442724",
"0.52385926",
"0.52385926",
"0.52385926",
"0.52385926",
"0.52385926",
"0.5232394",
"0.523231",
"0.5227454",
"0.52226824",
"0.52201617",
"0.5212327",
"0.52079266",
"0.52050185",
"0.51754695",
"0.51726824",
"0.51710224",
"0.5166172",
"0.5159343",
"0.51578903",
"0.51522785",
"0.5152022",
"0.51518047",
"0.51456624",
"0.51398855",
"0.5133759",
"0.5112076",
"0.5111866",
"0.5111866",
"0.5110294",
"0.5106169",
"0.509231",
"0.50873137",
"0.5081088",
"0.508059",
"0.50677156",
"0.50562143",
"0.5050554",
"0.50474834",
"0.50474834",
"0.5036181",
"0.5026331",
"0.5022976",
"0.5015441",
"0.50121695",
"0.5000944",
"0.5000019",
"0.4996878",
"0.4989888",
"0.4989888",
"0.49864885",
"0.49797225",
"0.49785787",
"0.4976161",
"0.49683493",
"0.4965126",
"0.4958034",
"0.49559742",
"0.4954353",
"0.49535993",
"0.4952725",
"0.49467874",
"0.49423352",
"0.49325448",
"0.49282882",
"0.49269363",
"0.49269104",
"0.49252945",
"0.4923091",
"0.49194667",
"0.49174926",
"0.49173003",
"0.49171105",
"0.4915879",
"0.49155936"
] |
0.0
|
-1
|
Never trust parameters from the scary internet, only allow the white list through.
|
def married_status_params
params.require(:married_status).permit(:name)
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def strong_params\n params.require(:user).permit(param_whitelist)\n end",
"def strong_params\n params.require(:listing_member).permit(param_whitelist)\n end",
"def allow_params_authentication!; end",
"def allowed_params\n ALLOWED_PARAMS\n end",
"def default_param_whitelist\n [\"mode\"]\n end",
"def param_whitelist\n [:role, :title]\n end",
"def expected_permitted_parameter_names; end",
"def safe_params\n params.except(:host, :port, :protocol).permit!\n end",
"def strong_params\n params.require(:team_member).permit(param_whitelist)\n end",
"def permitir_parametros\n \t\tparams.permit!\n \tend",
"def strong_params\n params.require(:community).permit(param_whitelist)\n end",
"def permitted_strong_parameters\n :all #or an array of parameters, example: [:name, :email]\n end",
"def strong_params\n params.require(:education).permit(param_whitelist)\n end",
"def restricted_params\n #params.require(self.controller_name.classify.underscore.to_sym).permit([])\n raise(\"No strong params set, override restricted_params method in your controller. E.g. params.require(:model).permit(:attribute1, :attribute2)\")\n end",
"def allowed_params\n params.require(:user).permit(:username, :email, :password, :password_confirmation)\n end",
"def param_whitelist\n [:rating, :review]\n end",
"def param_whitelist\n whitelist = [\n :username, :name,\n :parent_id,\n :headline, :description, :video,\n :policy, :signup_mode, :category,\n :website, :facebook, :twitter, :linkedin,\n :founded_at,\n privacy: [\n :events,\n :resources\n ],\n permission: [\n :profile,\n :members,\n :children,\n :statistics,\n :posts,\n :listings,\n :resources,\n :events\n ],\n location: [\n :description,\n :street,\n :city,\n :state,\n :zip,\n :country,\n :latitude,\n :longitude\n ]\n ]\n \n if action_name === 'update'\n whitelist.delete(:parent_id)\n unless current_user.role_in(@community) === 'owner'\n whitelist.delete(:privacy)\n whitelist.delete(:permission)\n end\n end\n \n whitelist\n end",
"def param_whitelist\n if @user.present? && current_user != @user\n return [:followed]\n end\n \n whitelist = [\n :username, :email, :password,\n :first_name, :last_name,\n :birthday, :gender,\n :headline, :biography, :ask_about, :focus,\n :website, :facebook, :linkedin, :twitter, :github,\n roles: [],\n skills: [],\n interests: [],\n privacy: { contact: [] },\n location: [\n :description,\n :street,\n :city,\n :state,\n :zip,\n :country,\n :latitude,\n :longitude\n ]\n ]\n \n if action_name === 'update'\n whitelist.delete(:email)\n whitelist.delete(:password)\n end\n \n whitelist\n end",
"def valid_params_request?; end",
"def user_params \n \tparams.require(:user).permit(:name, :email, :password, :password_confirmation)# preventing CSTR\n end",
"def strong_params\n params.require(:experience).permit(param_whitelist)\n end",
"def user_params\n params.permit(:name, :phoneNumber, :address, :postalCode, :local, :link, :counter, :latitude, :longitude) \n end",
"def trim_whitelisted(params, whitelist)\n # remove any parameters that are not whitelisted\n params.each do |key, value|\n # if white listed\n if whitelist.include? key\n # strip the parameters of any extra spaces, save as string\n params[key] = value.to_s.strip\n else\n # delete any unauthorized parameters\n params.delete key\n end\n end\n params\n end",
"def whitelist_url_params\n params.require(:whitelist_url).permit(:domain)\n end",
"def allowed_params\n params.require(:allowed).permit(:email)\n end",
"def permitted_params\n []\n end",
"def trim_whitelisted(params, whitelist)\n # remove any parameters that are not whitelisted\n params.each do |key, value|\n # if white listed\n if whitelist.include? key\n # strip the parameters of any extra spaces, save as string\n params[key] = value.to_s.strip\n else\n # delete any unauthorized parameters\n params.delete key\n end\n end\n params\n end",
"def safe_params\n params.permit(:id, :name, :origin, :emails => []); #emails is an array\n end",
"def query_param\n\t\tparams.permit(:first_name, :last_name, :phone)\n\tend",
"def strong_params\n params.require(:success_metric).permit(param_whitelist)\n end",
"def devise_filter\r\n logger.debug(\"In devise_filter =>PARAMS: #{params.inspect}\")\r\n\r\n # White list for sign_up\r\n devise_parameter_sanitizer.for(:sign_up) { |u| u.permit(user_whitelist) }\r\n\r\n # White list for account update\r\n devise_parameter_sanitizer.for(:account_update) { |u| u.permit(user_whitelist, :current_password) }\r\n\r\n # White list for Invitation creation\r\n devise_parameter_sanitizer.for(:invite) { |u| u.permit(:account_type, :email, :invitation_token)}\r\n\r\n # White list for accept invitation\r\n devise_parameter_sanitizer.for(:accept_invitation) { |u| u.permit(user_whitelist, :invitation_token)}\r\n\r\n end",
"def whitelisted_user_params\n params.require(:user).\n permit( :first_name, :last_name, :email,:password,:password_confirmation,:birthday,:gender)\n end",
"def user_params\n ActionController::Parameters.permit_all_parameters = true\n params.require(:user) #.permit(:name, :surname, :phone, :password, :email, :time_zone)\n end",
"def strong_params\n params.require(:metric_change).permit(param_whitelist)\n end",
"def safe_params\n params.require(:user).permit(:name)\n end",
"def get_params\n\t\treturn ActionController::Parameters.new(self.attributes).permit(\"account_id\", \"title\", \"category\", \"introduction\", \"tags\", \"segment_type\", \"visible\", \"status\", \"main_image\")\n\tend",
"def grant_params\n @whitelisted = params.require(:grant).permit(:name, :description, :agency_id, :acronym)\n end",
"def check_params; true; end",
"def param_whitelist\n whitelist = [\n :description,\n :progress,\n :kpi_id\n ]\n \n unless action_name === 'create'\n whitelist.delete(:kpi_id)\n end\n \n whitelist\n end",
"def quote_params\n params.permit!\n end",
"def valid_params?; end",
"def paramunold_params\n params.require(:paramunold).permit!\n end",
"def user_params\n\t\tparams.permit(:nickname, :avatar, :description, :password, :gender, :birthday, :email, :phone, :qq_id, :wechat_id)\n\tend",
"def filtered_parameters; end",
"def user_params\n params.permit(\n \t:id,\n \t:email, \n \t:first_name, \n \t:last_name, \n \t:password, \n \t:confirm_token, \n \t:phone_number,\n \t:facebook_link,\n \t:car_model,\n \t:license_plate)\n end",
"def filtering_params\n params.permit(:email, :name)\n end",
"def check_params\n true\n end",
"def wx_public_params\n params.require(:wx_public).permit(:nickname, :manager, :alias)\n end",
"def allowed_params\n params.require(:user).permit(:email, :password, :role, :first_name, :last_name, :password_confirmation)\n end",
"def allowed_params\n params.require(:user).permit(:email, :password, :role, :first_name, :last_name, :password_confirmation)\n end",
"def listing_params\n\t\tparams.permit(:address, :transit_info, :rules, :other_info, :lat, :lng)\n\tend",
"def social_account_params\n\t\t\tparams.require(:social_account).permit!\n\t\tend",
"def safe_params\n resurce_name = self.class.resource_name\n params_method_name = \"#{resurce_name}_params\".to_sym\n if params[resurce_name]\n if respond_to?(params_method_name) || private_methods.include?(params_method_name)\n send(params_method_name)\n else\n raise ActiveModel::ForbiddenAttributesError, \"Please, define the '#{params_method_name}' method in #{self.class.name}\"\n end\n end\n end",
"def user_params\n params.require(:user).permit(:uri, :username, :password, :realname, :email, :publicvisible)\n end",
"def model_params\n\t\tparams.require(:manager).permit(\n\t :user_name,\n :password,\n :email,\n \t\t\t)\n\tend",
"def url_params\n params.require(:url).permit(:short_url, :original_url, :clicks, :ip_addresses)\n end",
"def article_params_whitelist\n params.require(:article).permit(:title, :description, category_ids: [])\n end",
"def college_whitelist_params\n params.require(:college_whitelist).permit(:status)\n end",
"def active_code_params\n params[:active_code].permit\n end",
"def filtering_params\n params.permit(:email)\n end",
"def valid_params(params)\n params.permit(:user_id, :photo_id, :originX, :originY, :width, :height)\n end",
"def ip_address_params\n\t\t\tparams.require(:ip_address).permit!\n end",
"def pull_request_params\n whitelist = [\n :url,\n :id,\n :html_url,\n :diff_url,\n :patch_url,\n :issue_url,\n :number,\n :state,\n :locked,\n :title\n ]\n params.require(:pull_request).permit(whitelist)\n end",
"def reserved_params\n params.require(:reserved).permit(:name, :email, :pax, :address, :KTP, :title)\n end",
"def post_params\n if current_user.admin? \n params.permit(:title, :body, :city, :country, :gps_location, :privacy, :visible, :latitude, :longitude, images: [], files: [])\n else \n params.permit(:title, :body, :city, :country, :gps_location, :privacy,:latitude, :longitude, images: [], files: [])\n end \n end",
"def list_params\n params.permit(:name)\n end",
"def filter_parameters; end",
"def filter_parameters; end",
"def vineyard_params\n params.permit(:vineyard_name, :email, :website_url, :phone, :address, :city, :region, :postcode, :country, :specialty, :description, :pet_friendly, :holiday, :tours, :events, :family_friendly, :cover_image, :image_one, :image_two, :image_three, :image_four, :user_id, :base64)\n end",
"def available_activity_params\n # params.require(:available_activity).permit(:type,:geometry,:properties)\n whitelisted = ActionController::Parameters.new({\n type: params.require(:available_activity)[:type],\n geometry: params.require(:available_activity)[:geometry].try(:permit!).to_h,\n properties: params.require(:available_activity)[:properties].try(:permit!).to_h\n }).try(:permit!)\n end",
"def user_params\n params.permit(:name, :username, :email, :password, :img_url, :bg_url, :coinbank)\n end",
"def user_params_pub\n\t \tparams[:user].permit(:hruid)\n\t end",
"def user_params\n params.permit(:id, :email, :password, :nickname, :status, :avatar, :flat_picture, :flatsharing_id, :member,\n :user, :color, :solde)\n end",
"def url_whitelist; end",
"def validate_search_inputs\n @whitelisted = params.fetch(:user, nil)\n if @whitelisted.blank?\n render_error(400, \"#{I18n.t('general_error.params_missing_key')}\": [I18n.t('general_error.params_missing_value', model: \"review\")])\n return\n else\n @whitelisted = @whitelisted.permit(:name, :uen, :description)\n end\n end",
"def param_whitelist\n [\n :title,\n :description,\n :organization,\n :team_id,\n :started_at,\n :finished_at,\n location: [\n :description,\n :street,\n :city,\n :state,\n :zip,\n :country,\n :latitude,\n :longitude\n ]\n ]\n end",
"def admin_social_network_params\n params.require(:social_network).permit!\n end",
"def filter_params\n params.require(:filters).permit(:letters)\n end",
"def valid_params(params)\n params.permit(:login, :first_name, :last_name, \n :password, :password_confirmation)\n end",
"def sensitive_params=(params)\n @sensitive_params = params\n end",
"def origin_params\n params.permit(:country, :state, :city, :postal_code, :address, :description)\n end",
"def permit_request_params\n params.permit(:address)\n end",
"def user_params\n # Ensure a user can't give themselves admin priveleges\n params.delete(:admin) if current_user.admin?\n params.require(:user).permit(:name, :email, :admin, :image)\n end",
"def secure_params\n params.require(:location).permit(:name)\n end",
"def strong_params\n params.require( :setting ).\n permit( :global_scan_limit, :per_user_scan_limit,\n :target_whitelist_patterns, :target_blacklist_patterns )\n end",
"def question_params\n params.require(:survey_question).permit(question_whitelist)\n end",
"def case_insensitive_params\n params.require(:case_insensitive).permit(:name)\n end",
"def maintenance_request_params\n params[:maintenance_request].permit! #allow all parameters for now\n end",
"def empire_master_no_match_params\n params.require(:empire_master_no_match).permit(:uid, :last_name, :list, :search_date, :double, :source)\n end",
"def backend_user_params\n params.permit!\n end",
"def url_params\n params[:url].permit(:full)\n end",
"def unwanted_params\n params.require(:unwanted).permit(:title, :description, :image)\n end",
"def filter_params\n\t\treturn params[:candidate].permit(:name_for_filter)\n\tend",
"def speed_measurement_params\n\n #fuckit, to lazy to deal with permit crap right now\n ActionController::Parameters.permit_all_parameters = true\n\n params[:speed_measurement]\n end",
"def user_params\n params.permit(:name, :age, :username, :display_photo, :password)\n end",
"def get_params\r\n #params.require(:article).permit(:title, :permalink, :content, :source_site, :introtext, :type_id, :order_by, :searchable, :created_by, :edited_by, :published_by, :published_on, :user_id)\r\n params.require(:article).permit!\r\n\r\n end",
"def pub_params\n params.require(:pub).permit(:name, :description, :phone, :email, :hidden, :city_id, :address)\n end",
"def pass_params\n params[:pass].permit(:name, :price, :description, :colour, :events)\n end",
"def droptraining_params\n params.permit(:training_id,:user_id, :utf8, :authenticity_token, :commit)\n end",
"def person_params\n # params whitelist does *not* include admin, sub, remember_token\n # TBD: share this whitelist with the list used by configuration_permitted_parameters\n # TBD: should current_password be on this list? -- for now, leaving off, since it seems to work without\n # NOTE: do not include 'admin' in this list!\n params.require(:person).permit(\n :name, \n :email, \n :description,\n :password, \n :password_confirmation\n )\n end",
"def parameter_params\n params.require(:parameter).permit(:name, :description, :param_code, :param_value, :active_from, :active_to)\n end"
] |
[
"0.6981537",
"0.67835593",
"0.6748275",
"0.67436063",
"0.6736311",
"0.65937173",
"0.6503359",
"0.6498499",
"0.6482832",
"0.6478776",
"0.645703",
"0.6439998",
"0.63802195",
"0.6377008",
"0.6366287",
"0.632018",
"0.63016284",
"0.63011277",
"0.62932974",
"0.62919617",
"0.62905645",
"0.6289235",
"0.6283876",
"0.62425834",
"0.62410337",
"0.6218672",
"0.62151134",
"0.62096137",
"0.6192354",
"0.6178057",
"0.6177618",
"0.61727077",
"0.6162073",
"0.6152049",
"0.61515594",
"0.61458135",
"0.6122875",
"0.61165285",
"0.6107696",
"0.6104097",
"0.6091097",
"0.6080201",
"0.60699946",
"0.6063739",
"0.60206395",
"0.60169303",
"0.60134894",
"0.601003",
"0.6007347",
"0.6007347",
"0.6001054",
"0.59997267",
"0.5997844",
"0.5991826",
"0.5991213",
"0.59911627",
"0.5980111",
"0.5967009",
"0.59597385",
"0.5958542",
"0.595787",
"0.5957425",
"0.59522784",
"0.5951228",
"0.59423685",
"0.5939385",
"0.5939122",
"0.5939122",
"0.59325653",
"0.5930178",
"0.59248054",
"0.59243476",
"0.59164625",
"0.59106",
"0.59101933",
"0.59084356",
"0.5905666",
"0.58975077",
"0.58974737",
"0.5895128",
"0.58946574",
"0.589308",
"0.58916",
"0.5885987",
"0.58838505",
"0.58792",
"0.58723736",
"0.58684355",
"0.58677715",
"0.5865701",
"0.5865538",
"0.5865288",
"0.586385",
"0.5862139",
"0.58614355",
"0.58593005",
"0.5857459",
"0.58541363",
"0.58536613",
"0.58520085",
"0.585011"
] |
0.0
|
-1
|
Build Apartments using their description and features. Get HTML, then parse it and collect parameters.
|
def build(apartment_url)
@apartment = Nokogiri::HTML(Net::HTTP.get(URI(apartment_url)))
description = apartment_description
features = apartment_features
description['apartment_features'] = features
Apartments.new(description)
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def create\n\n if params[:sample]\n analyse = Ca::Analyse.new(HTMLReader.instance.page(params[:sample][:address]));\n end\n if params[:text]\n analyse = Ca::Analyse.new(params[:text][:content]);\n end\n descript = analyse.description\n @problems = descript.problems\n @text = descript.text\n @best_phrases = Hash[analyse.description.first_n]\n @nr_of_chars = descript.text_number_of_chars\n @nr_of_words = descript.text_number_of_words\n @nr_of_nodes = descript.nr_of_nodes\n @score = descript.score\n @plagiarism = descript.plagiarism\n @html = descript.text.to_s.force_encoding(\"UTF-8\")\n @tags_problem = descript.tag_problem_flag\n end",
"def create_digest(html)\n apartments = []\n items = html.css('p.row')\n items.each do |item|\n details = item.css('.pl').text\n new_digest = {\n :description => item.css('a').text,\n :price => parse_price(item.css('.price').text),\n :bedrooms => parse_bedrooms(item.css('.pnr').text)\n }\n apartments << new_digest\n end\n apartments\nend",
"def html_description\n if long_description.nil?\n paragraphs = []\n else\n paragraphs = long_description.delete(\"\\r\").gsub(\"\\n\\n\", \"\\n\").split(\"\\n\")\n end\n\n str = paragraphs.map { |x| \"<p>#{x}</p>\" }.join\n\n str += \"\\n<!--product-details-->\\n\"\n bullets = []\n\n (1..10).each do |x|\n point = get_property(\"Bullet #{x}\")\n bullets << point unless point.blank?\n end\n\n if bullets.length > 0\n str += \"<ul>\"\n str += bullets.map { |x| \"<li>#{x}</li>\" }.join\n str += \"</ul>\"\n end\n\n str += \"\\n<!--sizing-->\\n\"\n str += get_property(\"Sizing Info\")\n\n str += \"\\n<!--care-instructions-->\\n\"\n str += \"<p>\" + get_property(\"Care Instructions\") + \"</p>\"\n end",
"def parse_description_for_tasks(description)\n html = Nokogiri::HTML description\n leaves = []\n requirements = []\n\n recurse = lambda { |contents|\n contents.each do |content|\n if content.name == 'b' || content.name == 'li'\n leaves << content\n else\n recurse.call content.children\n end\n end\n }\n\n recurse.call html.css('body').children\n\n requirements_leaf = leaves.find { |leaf|\n leaf.name == 'b' && leaf.text.match(/requirements/i)\n }\n\n i = leaves.index(requirements_leaf) + 1\n\n loop do\n break if i == leaves.length || leaves[i].name == 'b'\n\n requirements << leaves[i].text\n\n i += 1\n end\n\n requirements\n end",
"def build_item_from_html(html, index)\n\n new_item_attributes = {\n xiv_index: index,\n name: html.css('.content-page-title').text,\n }\n\n item = Item.new(new_item_attributes)\n\n additional_info = get_additional_info(index)\n\n job_data = html.css('#page-crafted span[style=\"display:inline-block;width:300px;\"]').text\n matches = /.*Attempt Level:\\s*(\\d+)\\s*(\\w+)/.match(job_data)\n\n item.save\n\n if (matches)\n job_requirements = {\n level: matches[1],\n }\n jr = JobRequirement.new(job_requirements)\n jr.job = Job.find_by({ name: matches[2] })\n jr.item = item\n jr.save\n end\n\n item\nend",
"def data_results(data)\n #data search p row. Creates a tree pattern that mirrors the html.\n results = data.search('p.row')\n\n apartments = []\n results.each do |listing| #listing is a bunch of data for each p row.\n apartments.push(create_apartment(listing))\n end\n create_csv(apartments) #list of apartments here.\nend",
"def process_descriptions(n)\n descriptions = self.feature.descriptions\n delete_descriptions = self.fields.delete('descriptions.delete')\n descriptions.clear if !delete_descriptions.blank? && delete_descriptions.downcase == 'yes'\n 0.upto(n) do |i|\n prefix = i>0 ? \"#{i}.descriptions\" : 'descriptions'\n description_content = self.fields.delete(\"#{prefix}.content\")\n if !description_content.blank?\n description_content = \"<p>#{description_content}</p>\"\n author_name = self.fields.delete(\"#{prefix}.author.fullname\")\n description_title = self.fields.delete(\"#{prefix}.title\")\n author = author_name.blank? ? nil : AuthenticatedSystem::Person.find_by(fullname: author_name)\n description = description_title.blank? ? descriptions.find_by(content: description_content) : descriptions.find_by(title: description_title) # : descriptions.find_by(['LEFT(content, 200) = ?', description_content[0...200]])\n language = Language.get_by_code_or_name(self.fields.delete(\"#{prefix}.languages.code\"), self.fields.delete(\"#{prefix}.languages.name\"))\n attributes = {:content => description_content, :title => description_title}\n attributes[:language_id] = language.id if !language.nil?\n if description.nil?\n if language.nil?\n self.say \"Language needed to create description for feature #{self.feature.pid}.\"\n description = nil\n else\n description = descriptions.create(attributes)\n end\n else\n description.update_attributes(attributes)\n end\n if !description.nil?\n self.spreadsheet.imports.create(:item => description) if description.imports.find_by(spreadsheet_id: self.spreadsheet.id).nil?\n description.authors << author if !author.nil? && !description.author_ids.include?(author.id)\n end\n end\n end \n end",
"def offers_parser\n SpinningCursor.run { message 'Done!' }\n\n parsed_offers = PageScrapper.page_scrapper.css('div#tab_specials_content a') # 40 deals\n SpinningCursor.stop\n parsed_offers\n end",
"def create_guide(title, description)\n result = \"\"\n result << \"<div class='explanation-unit'>\"\n result << \"<h1>#{title}</h1>\"\n result << \"<p>#{description}</p>\"\n result << \"</div>\"\n end",
"def create_guide(title, description)\n result = \"\"\n result << \"<div class='explanation-unit'>\"\n result << \"<h1>#{title}</h1>\"\n result << \"<p>#{description}</p>\"\n result << \"</div>\"\n end",
"def create_guide(title, description)\n result = \"\"\n result << \"<div class='explanation-unit'>\"\n result << \"<h1>#{title}</h1>\"\n result << \"<p>#{description}</p>\"\n result << \"</div>\"\n end",
"def create_html(options)\n if(options[:html_template] && options[:html_template] == :scientific) then\n\n if(options[:language] && options[:language] == :english) then\n html = '<h2>Academic Interests<\\/h2>\\r\\n' +\n '<p>Add information about academic fields of interest.<\\/p>\\r\\n' +\n '<h2>Teaching<\\/h2>\\r\\n' +\n '<ul>\\r\\n' +\n ' <li><Link to programme of study/course><\\/li>\\r\\n' +\n ' <li><Link to programme of study/course><\\/li>\\r\\n' +\n ' <li>...<\\/li>\\r\\n' +\n '<\\/ul>\\r\\n' +\n '<h2>Higher education and employment history<\\/h2>\\r\\n' +\n '<p>Brief introduction to previous education and employment.<\\/p>\\r\\n' +\n '<h2>Honoraria<\\/h2>\\r\\n' +\n '<ul>\\r\\n' +\n ' <li><Name of prize and (if applicable) link 1><\\/li>\\r\\n' +\n ' <li><Name of prize and (if applicable) link 2><\\/li>\\r\\n' +\n ' <li>...<\\/li>\\r\\n' +\n '<\\/ul>\\r\\n' +\n '<h2>Appointments<\\/h2>\\r\\n' +\n '<ul>\\r\\n' +\n ' <li><Title and (if applicable) link 1><\\/li>\\r\\n' +\n ' <li><Title and (if applicable) link 2><\\/li>\\r\\n' +\n ' <li>...<\\/li>\\r\\n' +\n '<\\/ul>\\r\\n' +\n '<h2>Cooperation<\\/h2>\\r\\n' +\n '<p> <\\/p>'\n else\n html = '<h2>Faglige interesser<\\/h2>\\r\\n' +\n '<p>Her kan du skrive om faglige interesser.<\\/p>\\r\\n' +\n '<h2>Undervisning<\\/h2>\\r\\n<p>' +\n '<Lenke til studieprogram/emne> <br />\\r\\n' +\n '<Lenke til studieprogram/emne> <br />\\r\\n...<\\/p>\\r\\n' +\n '<h2>Bakgrunn<\\/h2>\\r\\n' +\n '<p>Kort om tidligere arbeidserfaring og utdanning<\\/p>\\r\\n' +\n '<h2>Priser<\\/h2>\\r\\n' +\n '<p><Navn og eventuelt lenke til pris 1> <br />\\r\\n' +\n '<Navn og eventuelt lenke til pris 2> <br />\\r\\n' +\n '...<\\/p>\\r\\n' +\n '<h2>Verv<\\/h2>\\r\\n<p>' +\n '<Navn og eventuelt lenke til verv 1> <br />\\r\\n' +\n '<Navn og eventuelt lenke til verv 2> <br />\\r\\n...' +\n '<\\/p>\\r\\n' +\n '<h2>Samarbeid<\\/h2>\\r\\n' +\n '<p> <\\/p>'\n end\n else\n\n if(options[:language] && options[:language] == :english) then\n html = '<h2>Tasks performed<\\/h2>\\r\\n' +\n '<p>Add information about job duties, as a short text or a bulleted list:<\\/p>' +\n '\\r\\n<ul>\\r\\n' +\n ' <li><Task 1><\\/li>\\r\\n' +\n ' <li><Task 1><\\/li>\\r\\n' +\n ' <li>...<\\/li>\\r\\n' +\n '<\\/ul>\\r\\n' +\n '<h2>Background<\\/h2>\\r\\n' +\n '<p>Add information about previous education and employment.<\\/p>'\n else\n html = '<h2>Arbeidsområder<\\/h2>\\r\\n' +\n '<p>Her kan du skrive om arbeidsområder, ' +\n 'enten som kort tekst eller som listepunkter:</p>' +\n '\\r\\n' +\n '<ul>\\r\\n' +\n ' <li><Arbeidsområde 1></li>\\r\\n' +\n ' <li><Arbeidsområde 1></li>\\r\\n' +\n ' <li>...</li>\\r\\n' +\n '</ul>' +\n '\\r\\n' +\n '<h2>Bakgrunn</h2>\\r\\n' +\n '<p>Eventuelt kort om tidligere arbeidserfaring og utdanning.</p>'\n end\n end\n return html\n end",
"def parse(max_apts = 100)\n @mech.get(@url) do |page|\n i = 0\n page.links.each do |link|\n next if link.href == nil\n next if i >= max_apts\n next unless links_to_apt?(link)\n\t\t\t\tnext if Apartment.find_by_url(link.href)\n @apts << AptDetailsPage.new(link.href, link.to_s).parse\n sleep 0.1 # Don't hit CL too quickly\n i += 1\n end\n end\n self\n end",
"def add_courses(url, counter, parsed_page)\n all_paragraphs = parsed_page.xpath '//p' # all <p> on the page\n paragraph_number = 8 # The description paragraph for most pages\n\n # get the course's description\n course_description = \"\"\n while !all_paragraphs[paragraph_number].text.eql? \"Qualification\" do\n course_description += all_paragraphs[paragraph_number].text.strip\n course_description += \"\\n\\n\"\n paragraph_number += 1\n end\n # some pages are set up differently\n if course_description.empty?\n course_description = all_paragraphs[7].text\n end\n course_description = course_description.strip\n\n # if it exists, get the provider's url for the course\n provider_url = \"\"\n if !parsed_page.at_css('[id=\"ProviderCourseUrl\"]').nil?\n provider_url = parsed_page.at_css('[id=\"ProviderCourseUrl\"]').attributes[\"href\"].value\n end\n\n department = \"\"\n if !parsed_page.css('span').css('[id=\"contact_Title\"]')[0].nil?\n department = parsed_page.css('span').css('[id=\"contact_Title\"]')[0].text\n end\n\n email = \"\"\n if !parsed_page.at_css('.contact-email').nil?\n #email = parsed_page.at_css('.contact-email').attributes[\"href\"].value\n email = parsed_page.css('[class=\"contact-email\"]')[0].text.strip\n end\n\n # if a contact exists then\n contact = \"\"\n if !parsed_page.at_css('[id=\"contact_Phone\"]').nil?\n contact = parsed_page.at_css('[id=\"contact_Phone\"]').text\n end\n\n # Entry requirements [[Exam type, Grade/Mark, Info]]\n requirements = Array.new\n requirements_section = parsed_page.at_css('[id=\"entry-requirements-section\"]')\n requirements_rows = requirements_section.css('tr').drop(1)\n requirements_rows.each do |row|\n row_info = Array.new\n row_info << row.css('th').text.strip\n row.css('td').each do |col|\n row_info << col.text.strip\n end\n if row_info.count == 2\n row_info << \"\"\n end\n requirements << row_info\n end\n\n # extract additional entry requirements info\n requirements_info = \"\"\n requirements_section.css('p').each do |p|\n requirements_info += p.text.strip\n requirements_info += \"\\n\"\n end\n requirements_info += requirements_section.css('a').text.strip\n requirements_info = requirements_info.strip\n\n # entry points for the course\n entry_points = Array.new\n parsed_page.at_css('[id=\"howToApply\"]').parent.css('li').each do |row|\n entry_points << row.text.strip\n end\n\n # fees info [[student type, fee, fee period]]\n fees = Array.new\n empty_fees_message = \"No fee information has been provided for this course\"\n empty_fees_scraped_message = parsed_page.at_css('[id=\"feesAndFunding\"]').parent.children[3].children[2].text.strip\n if !(empty_fees_scraped_message.eql? empty_fees_message)\n fees_table = parsed_page.css('[class=\"table-responsive table-responsive--list table-borderless table-col1-bold\"]')[-1].children[1]\n fees_rows = fees_table.css('tr').count\n fees = Array.new(fees_rows){Array.new(3)}\n (0...fees_rows).each do |row_number|\n row_info = fees_table.css('tr')[row_number]\n fees[row_number][0] = row_info.css('td')[0].text\n fees[row_number][1] = row_info.css('td')[1].text.split(\"\\u00A3\")[1].split[0].tr(',','')\n fees[row_number][2] = row_info.css('td')[2].text\n end\n end\n\n # extract additional fees info\n fees_sections = parsed_page.at_css('[id=\"feesAndFunding\"]').parent.css('section').drop(1)\n fees_info = \"\"\n fees_sections.each do |section|\n paragraph = section.css('div').text.strip\n if paragraph.empty?\n paragraph = section.css('p').text.strip\n end\n fees_info += paragraph\n end\n fees_info = fees_info.strip\n\n delivery = \"\"\n if parsed_page.css('p').text.eql? \"lectures\"\n delivery = parsed_page.css('p').text.strip\n end\n\n notes = \"\"\n if !parsed_page.css('[id=\"courseDetails\"]').empty?\n notes_sections = parsed_page.at_css('[id=\"courseDetails\"]').parent.css('section')\n notes_sections.each do |section|\n #notes += \"[h3]\"+section.css('h3').text.strip + \"[/h3]\\n\\n\"\n notes += section.text.strip + \"\\n\\n\"\n end\n end\n\n\n # course object with all of the scraped info\n course = {\n name: parsed_page.css('h1.search-result__result-provider').children[0].text.strip,\n qualification: all_paragraphs[paragraph_number+1].text,\n provider: parsed_page.css('h1.search-result__result-provider').children[1].text.strip,\n provider_url: provider_url,\n ucas_url: url,\n description: course_description,\n study_mode: all_paragraphs[paragraph_number+3].text.strip,\n location: all_paragraphs[paragraph_number+5].text.strip,\n start_date: all_paragraphs[paragraph_number+7].text.strip,\n duration: all_paragraphs[paragraph_number+9].text.strip,\n entry_points: entry_points,\n department: department,\n contact_number: contact,\n email: email,\n requirements: requirements,\n requirements_info: requirements_info,\n fees: fees,\n fees_info: fees_info,\n institution: parsed_page.css('td[id=\"institution-code\"]').text,\n course_code: parsed_page.css('td[id=\"application-code\"]').text,\n delivery: delivery,\n notes: notes\n }\n\n puts \"Course #{counter}: #{course[:name]} #{course[:provider]}\"#, delivery: #{course[:delivery]}\"\n @all_courses << course\n end",
"def build( special_contexts=[] )\r\n\r\n self.ehtmls, self.ecss, self.ejs = self.theme.original_page_layout_root.build_content(special_contexts)\r\n self.ehtml = self.ehtmls.first\r\n\r\n end",
"def create\n html = HTMLEntities.new\n params[:administration_offering][:description] = html.encode(params[:administration_offering][:description])\n @administration_offering = Administration::Offering.new(params[:administration_offering])\n\n respond_to do |format|\n if @administration_offering.save\n @administration_offering.add_offering_merchants(params[:administration_merchant_offering])\n format.html { redirect_to(@administration_offering, :notice => 'Offering was successfully created.') }\n format.xml { render :xml => @administration_offering, :status => :created, :location => @administration_offering }\n else\n load_merchants\n @merchant_offerings = Administration::MerchantOffering.find(:all)\n format.html { render :action => \"new\" }\n format.xml { render :xml => @administration_offering.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def make_apartment(some)\n puts \"How much does the apartment cost?\"\n @price = \"$\" + gets.chomp\n @is_occupied = \"is occupied\"\n \n puts \"How much sq footage does it have?\"\n @sqft = gets.chomp + \" sqft\"\n \n puts \"How many bedrooms?\"\n @num_bedrooms = gets.chomp + \" bedrooms\"\n \n puts \"How many bathrooms?\"\n @num_bathrooms = gets.chomp + \" bathrooms\"\n #pushes attr into hash \n $listing.push(some,@price,@is_occupied,@sqft,@num_bedrooms,@num_bathrooms)\n $building.add_apartment($listing[0], $listing[1..5])\n end",
"def build_prosecution(attrs, page)\n doc = Nokogiri::HTML(page.body) {|c| c.noblanks}\n elements = doc.search('div.wysiwyg').first.children\n header = elements.search(\"//a[@id='#{attrs['id']}']\").first.parent\n\n els = [header]\n current = header.next\n until current.nil? || current.name == 'h3' do\n els << current\n current = current.next\n end\n\n if more_attrs = extract_attrs(els)\n puts \"Extracting #{more_attrs['address']}\"\n attrs.merge(more_attrs)\n else\n nil\n end\nend",
"def build\n \n puts \"[Cutlist.build] Building cutlist...\" if $cutlister_debug\n \n data = heading().to_s\n \n data += page_title().to_s\n \n # Show sheet goods section.\n if @options[\"show_sheets\"]\n \n if @parts.sheets != nil\n\n data += section_heading(\"Sheet Goods\").to_s\n data += rows(@parts.sheets).to_s\n data += section_footer(@parts.sheets).to_s\n \n end\n \n end\n \n # Show solid stock section.\n if @options[\"show_solids\"]\n \n if @parts.solids != nil\n \n data += section_heading(\"Solid Stock\").to_s\n data += rows(@parts.solids).to_s\n data += section_footer(@parts.solids).to_s\n \n end\n \n end\n \n # Show hardware section.\n if @options[\"show_hardware\"]\n \n if @parts.hardware != nil\n \n data += section_heading(\"Hardware\").to_s\n data += rows(@parts.hardware).to_s\n data += section_footer(@parts.hardware).to_s\n \n end\n \n end\n \n data += footer().to_s\n \n # Return the results.\n data\n \n end",
"def build_park_data(ple, with_details = true)\n add = { }\n reservation_uri = nil\n\n if with_details\n res = get_park_details_page(ple['park_id'])\n if res.is_a?(Net::HTTPOK)\n doc = Nokogiri::HTML(res.body)\n\n pfl = extract_park_features(doc, res.uri)\n ACTIVITY_MAP.each do |ak, al|\n a = filter_activities(pfl, al)\n add[ak] = a.join(', ') if a.count > 0\n end\n\n reservation_uri = extract_park_reservation_uri(doc, res, ple)\n end\n end\n\n name = ple['park_name']\n\n cpd = {\n signature: \"state/oregon/#{name.downcase}/#{name.downcase}\",\n organization: ORGANIZATION_NAME,\n name: name,\n uri: park_details_url(ple['park_id']),\n region: REGION_NAME,\n area: '',\n location: {\n lat: ple['park_latitude'],\n lon: ple['park_longitude']\n },\n additional_info: add\n }\n\n cpd[:reservation_uri] = reservation_uri unless reservation_uri.nil?\n\n self.logger.info { \"extracted park data for (#{cpd[:region]}) (#{cpd[:area]}) (#{cpd[:name]})\" }\n cpd\n end",
"def part_A(f)\n panel(f, \"Part A Services\") do\n a = content_tag :p, \"Complete this section \" \\\n \"to authorize payment \" \\\n \"to a service provider \" \\\n \"who is providing autism intervention \" \\\n \"for the child.\"\n\n a += form_row do\n f.text_field(:service_provider_name, column_width: 8, lstrip: \"\") +\n f.form_group(:payment,\n label: { text: \"Payment to be provided to:\" },\n column_width: 4) do\n f.radio_button(:payment, \"provider\", label: \"Service Provider\") +\n f.radio_button(:payment, \"agency\", label: \"Agency\")\n end\n end\n a += form_row do\n f.text_field(:agency_name,\n column_width: 8,\n label: \"Agency Name (if applicable)\")\n end\n a += form_row do\n f.text_field(:service_provider_address,\n column_width: 5,\n label: \"Address\") +\n f.text_field(:service_provider_city,\n column_width: 3,\n label: \"City/Town\") +\n f.text_field(:service_provider_postal_code,\n column_width: 2,\n label: \"Postal Code\") +\n f.phone_field(:service_provider_phone,\n column_width: 2,\n label: \"Phone Number\")\n end\n a += form_row do\n f.text_field(:service_provider_service_1,\n column_width: 6,\n placeholder: \"Service 1\",\n label: \"Type of Service(s)\") +\n f.date_field(:service_provider_service_start,\n column_width: 3,\n label: \"Start Date\") +\n f.date_field(:service_provider_service_end,\n column_width: 3,\n label: \"End Date\")\n end\n a += form_row do\n f.text_field(:service_provider_service_2,\n column_width: 6,\n placeholder: \"Service 2\",\n hide_label: true) +\n f.currency_field(:service_provider_service_fee,\n column_width: 2,\n label: \"Fee (incl PST)\") +\n f.select(:service_provider_service_hour,\n %w(Hour Day),\n column_width: 2,\n label: \"Per\") +\n f.currency_field(:service_provider_service_amount,\n column_width: 2,\n label: \"Total Amount\")\n end\n a + form_row do\n f.text_field(:service_provider_service_3,\n column_width: 6,\n placeholder: \"Service 3\",\n hide_label: true)\n end\n end\n end",
"def perform\n url = build_request_url\n req = open(url).read\n response_body = JSON.parse(req)\n build_museum_objects_from_response(response_body['features'])\n end",
"def BuildExcerpts(docs, index, words, opts = {})\n assert { docs.instance_of? Array }\n assert { index.instance_of? String }\n assert { words.instance_of? String }\n assert { opts.instance_of? Hash }\n\n # fixup options\n opts['before_match'] ||= '<b>';\n opts['after_match'] ||= '</b>';\n opts['chunk_separator'] ||= ' ... ';\n\t opts['html_strip_mode'] ||= 'index';\n opts['limit'] ||= 256;\n\t opts['limit_passages'] ||= 0;\n\t opts['limit_words'] ||= 0;\n opts['around'] ||= 5;\n\t opts['start_passage_id'] ||= 1;\n opts['exact_phrase'] ||= false\n opts['single_passage'] ||= false\n opts['use_boundaries'] ||= false\n opts['weight_order'] ||= false\n\t opts['load_files'] ||= false\n\t opts['allow_empty'] ||= false\n \n # build request\n \n # v.1.0 req\n flags = 1\n flags |= 2 if opts['exact_phrase']\n flags |= 4 if opts['single_passage']\n flags |= 8 if opts['use_boundaries']\n flags |= 16 if opts['weight_order']\n\t flags |= 32 if opts['query_mode']\n\t flags |= 64 if opts['force_all_words']\n\t flags |= 128 if opts['load_files']\n\t flags |= 256 if opts['allow_empty']\n \n request = Request.new\n request.put_int 0, flags # mode=0, flags=1 (remove spaces)\n # req index\n request.put_string index\n # req words\n request.put_string words\n \n # options\n request.put_string opts['before_match']\n request.put_string opts['after_match']\n request.put_string opts['chunk_separator']\n request.put_int opts['limit'].to_i, opts['around'].to_i\n\t \n\t # options v1.2\n\t request.put_int opts['limit_passages'].to_i\n\t request.put_int opts['limit_words'].to_i\n\t request.put_int opts['start_passage_id'].to_i\n\t request.put_string opts['html_strip_mode']\n \n # documents\n request.put_int docs.size\n docs.each do |doc|\n assert { doc.instance_of? String }\n\n request.put_string doc\n end\n \n response = PerformRequest(:excerpt, request)\n \n # parse response\n begin\n res = []\n docs.each do |doc|\n res << response.get_string\n end\n rescue EOFError\n @error = 'incomplete reply'\n raise SphinxResponseError, @error\n end\n return res\n end",
"def scrape_individual_offer(offer_object, url)\n puts \"Pulling details from posting #{url}\"\n domain = 'https://www.indeed.com.mx/'\n html_content = open(domain + url).read\n doc = Nokogiri::HTML(html_content)\n description = doc.search('.jobsearch-jobDescriptionText')\n offer_object[:description] = description # unrefined but workable\n collect_posting_date(doc, offer_object)\n collect_job_type(description.text.downcase, offer_object)\n collect_tags_indeed(description.text.downcase, offer_object)\n sleep 0.7\n end",
"def build\n \n data = heading(:css_location => \"css/html-cutlist.css\").to_s\n \n data += page_title().to_s\n \n # Get all the parts in an array of part hashes.\n all_parts = @parts.grouped\n\n # # Sort parts.\n # all_parts = all_parts.sort { |a, b|\n # a['material'] <=> b['material']\n # # a['thickness'] <=> b['thickness']\n # # a['width'] <=> b['width']\n # # a['length'] <=> b['length']\n # }\n\n # Create an empty array of materials.\n materials = []\n\n # Put all the materials in the list\n all_parts.each { |p|\n materials.push(p['material']) \n }\n\n # Make sure there is only one of each material.\n materials = materials.uniq\n\n # Create a blank hash to put sorted parts into.\n grouped_parts = {}\n\n # Create a new list of hashes that represent the material and then the parts \n # that are of that material (a blank hash).\n materials.each { |m| \n grouped_parts[m] = {}\n }\n\n # Loop through each part, adding it to the right key in the sorted_list.\n all_parts.each { |p| \n\n # Go through the list of materials.\n materials.each { |m|\n\n if p['material'] == m\n \n # If the \"round_dimensions\" setting is on, round the dimension to \n # three decimal places so we don't have a lot of entries in the \n # grouped_parts list like \"3.750000001\", \"3.750000002\", etc... but \n # instead have one entry like \"3.750\". This way grouping behaves \n # as expected.\n rounded_thickness = @options['round_dimensions'] ? format(\"%0.3f\", p['thickness']).to_f : p['thickness']\n \n # Check to see if there is a key for this thickness and if there is \n # append the part to the array of parts.\n if grouped_parts[m][rounded_thickness]\n grouped_parts[m][rounded_thickness] += [p]\n # If there isn't a key for this thickness, create it now and add the \n # part array.\n else\n grouped_parts[m][rounded_thickness] = [p]\n end\n\n end\n\n }\n\n }\n \n # Sort by materials.\n parts_by_material = grouped_parts.sort { |a,b|\n a[0] <=> b[0]\n }\n\n # List all the parts, grouped.\n parts_by_material.each { |t| # t for thickness.\n\n # Sort thicknesses.\n parts_by_thickness = t[1].sort { |a,b|\n a[0] <=> b[0]\n }\n parts_by_thickness.reverse!\n\n # Go through each thickness key.\n parts_by_thickness.each { |p| # p for parts\n\n # Sort parts by width, then length.\n parts = p[1].sort { |a,b|\n a['width'] <=> b['width']\n # TODO: Get sub-sorting by length.\n # a['length'] <=> b['length']\n }\n parts.reverse!\n \n # Create a parts array to store the parts in.\n parts_array = []\n \n # Go through the parts that are of a specific thickness.\n parts.each { |part| \n \n # Check if part is a sheet good.\n if @options[\"show_sheets\"] && part['is_sheet']\n parts_array.push(part)\n end\n\n # Check if part is solid stock.\n if @options[\"show_solids\"] && part['is_solid']\n parts_array.push(part)\n end\n\n # Check if part is hardware.\n if @options[\"show_hardware\"] && part['is_hardware']\n parts_array.push(part)\n end\n # puts \"#{p['sub_assembly']} -- #{p['part_name']} -- #{p['quantity']} -- #{p['material']} -- Sheet? #{p['is_sheet']} -- Solid? #{p['is_solid']} -- Hardware? #{p['is_hardware']} -- #{p['width']} x #{p['length']} x #{p['thickness']}\"\n \n }\n \n # If there are items in the parts_array, then add the section heading \n # and the array of parts. If there are no parts, show nothing.\n if parts_array.length != 0\n data += section_heading(\"#{p[0].to_fraction(true)} #{t[0]}\")\n data += rows(parts_array)\n end\n \n # TODO: Put section footer here, if needed...\n \n }\n }\n \n data += footer().to_s\n \n # Return the results.\n data\n \n end",
"def build\r\n self.ehtml, self.ecss, self.ejs = self.theme.page_layout.build_content() \r\n return self.ehtml, self.ecss, self.ejs\r\n end",
"def initialize(doc, dept_abbrevs)\n @valid = true\n @attributes = {}\n \n # split up the page title to extract information, or fail if not the expected format\n title = doc.search(\"title\").inner_text\n # => \"AERONAUTICS & ASTRONAUTICS A A 101 A ?? Joseph M Blakely ??Other?? AU07\"\n title = title.split(\"??\")\n # => [\"AERONAUTICS & ASTRONAUTICS A A 101 A \", \" Joseph M Blakely \", \"Other\", \" AU07\"]\n return @valid = false if title.size < 4\n \n @attributes[:instructor_name] = title[1].strip\n @attributes[:instructor_type] = title[2].strip\n \n # get quarter ('AU07') and convert it to the format used in the time schedule ('AUT2007')\n quarter = title[3].strip\n if (quarter =~ /97|98|99/)\n year_prefix = '19'\n else\n year_prefix = '20'\n end\n quarter.sub!(/AU/, \"AUT#{year_prefix}\"); quarter.sub!(/WI/, \"WIN#{year_prefix}\"); quarter.sub!(/SP/, \"SPR#{year_prefix}\"); quarter.sub!(/SU/, \"SUM#{year_prefix}\")\n @attributes[:quarter] = quarter\n \n # get the course number, or fail\n match = title[0].match(/\\d+/)\n return @valid = false unless match\n @attributes[:number] = match[0].strip\n @attributes[:section] = match.post_match.strip if match.post_match\n \n # match the dept abbrev using the list passed to us (guaranteed to be sorted long to short).\n # this is the only way to separate department name from department abbrev. fail if no match\n title_and_abbrev = match.pre_match.strip\n dept = nil\n dept_abbrevs.each do |abbrev|\n match = title_and_abbrev.match(\" #{Regexp::escape(abbrev)}$\") \n break dept = match[0].strip if match\n end\n unless dept\n #puts \"FAILED on \" + title_and_abbrev\n return @valid = false\n end\n @attributes[:dept] = dept\n \n # parse the eval scores table\n table = doc.search(\"table\")\n caption = table.search(\"caption\").inner_text\n # => \"Form B: Large Lecture?????\\\"39\\\" surveyed ?? \\\"56\\\" enrolled\"\n caption = caption.split(/\\?+/)\n # => [\"Form B: Large Lecture\", \"\\\"39\\\" surveyed \", \" \\\"56\\\" enrolled\"]\n @attributes[:course_type] = caption[0].strip\n @attributes[:surveyed] = caption[1].match(/\\d+/)[0].to_i\n @attributes[:enrolled] = caption[2].match(/\\d+/)[0].to_i\n @attributes[:stats] = {}\n rows = table.search(\"tr\")\n rows.each do |row|\n row = row.inner_text\n # => \"The course as a whole: 28%38%28%3%3%0% 3.93\"\n key = nil\n key = :whole if row =~ /whole:/\n key = :content if row =~ /content:|taught:/\n key = :contribution if row =~ /contribution:/\n key = :effectiveness if row =~ /effectiveness:/\n key = :interest if row =~ /interest:|enthusiastic:/\n key = :learned if row =~ /learned:/\n key = :grading if row =~ /techniques:/\n next unless key\n row = row.split(/ |:/).reverse # sometimes there is no space after \"whole:\"\n # [\"3.93\", \"28%38%28%3%3%0%\", \"whole:\", \"a\", \"as\", \"course\", \"The\"]\n median = row[0].to_f\n scores = row[1].split('%').reverse\n # => [\"0\", \"3\", \"3\", \"28\", \"38\", \"28\"]\n scores.map! { |x| x = x.to_i }\n @attributes[:stats][key] = {:median => median, :scores => scores}\n end\n end",
"def process_and_build_components \n end",
"def format_description(text)\n # Look for signs of structure, otherwise just treat as unstructured.\n case text\n when /\"\";/ then double_quotes_to_sections(text)\n when /\\.--v\\. */ then double_dash_to_sections(text)\n when /; *PART */i then # Seen in some IA records.\n when /:;/ then # Observed in one unusual case.\n when /[[:punct:]] *--.* +-- +/ then # Blurbs/quotes with attribution.\n when / +-- +.* +-- +/ then # Table-of-contents title list.\n when /(;[^;]+){4,}/ then # Many sections indicated.\n else return format_multiline(text)\n end\n q_section = nil\n text.split(/ *; */).flat_map { |part|\n next if (part = part.strip).blank?\n case part\n when /^\"\"(.*)\"\"$/\n # === Rare type of table-of-contents listing entry\n line = $1.to_s\n if line.match(SECTION_TITLE_RE)\n gap = (\"\\n\" unless q_section)\n q_section = $1.to_s\n [gap, \"#{q_section} #{$2}\", \"\\n\"].compact\n else\n q_section = nil\n line.match?(/^\\d+ +/) ? line : \"#{BLACK_CIRCLE}#{EN_SPACE}#{line}\"\n end\n\n when / +-- +.* +-- +/\n # === Table-of-contents listing\n section = nil\n part.split(/ +-- +/).flat_map { |line|\n if line.match(SECTION_TITLE_RE)\n gap = (\"\\n\" unless section)\n section = $1.to_s.delete_suffix('.')\n [gap, \"#{section}. #{$2}\", \"\\n\"].compact\n else\n section = nil\n \"#{BLACK_CIRCLE}#{EN_SPACE}#{line}\"\n end\n }.tap { |toc| toc << \"\\n\" unless toc.last == \"\\n\" }\n\n when /[[:punct:]] *--/\n # === Blurbs/quotes with attribution\n part.scan(BLURB_RE).flat_map do |paragraph, attribution|\n attribution.remove!(/[.\\s]+$/)\n [\"#{paragraph} #{EM_DASH}#{attribution}.\", \"\\n\"]\n end\n\n when /^v[^.]*\\. *\\d/\n # === Apparent table-of-contents volume title\n [part]\n\n else\n # === Plain text section\n part = \"#{part}.\" unless part.match?(/[[:punct:]]$/)\n [part, \"\\n\"]\n end\n }.compact.map { |line|\n line.gsub(/---/, EM_DASH).gsub(/--/, EN_DASH)\n }\n end",
"def extract_ingredients_structured\n # NOTE manish 9/17/12 This is for food52 http://www.food52.com/recipes/4175_gin_spritz\n # Interestingly it has to be evaluated before li[contains(@itemprop]!)]\n ingredients = @doc.xpath(\"//p[contains(@itemprop, 'ingredients')]\").collect { |s| clean_text(s.text)}\n\n ingredients = @doc.xpath(\"//div[contains(@class, 'ingredients')]//li\").collect { |s| clean_text(s.text)\n } if ingredients.empty?\n ingredients = @doc.xpath(\"//div[contains(@id, 'ingredients')]//li\").collect { |s| clean_text(s.text)} if ingredients.empty?\n\n # READ READ READ READ: a lot of these patterns have no test. You can remove two or three of these\n # patterns and the test will still pass. In the future if you add a pattern, document the website\n # or page for which you are adding the pattern\n\n ingredients = @doc.xpath(\"//li[contains(@class,'ingredient')]\").collect { |s| clean_text(s.text)} if ingredients.empty?\n ingredients = @doc.xpath(\"//li[contains(@itemprop,'ingredient')]\").collect { |s| clean_text(s.text)} if ingredients.empty?\n ingredients = @doc.xpath(\"//span[contains(@class, 'ingredient')]\").collect { |s| clean_text(s.text)} if ingredients.empty?\n ingredients = @doc.css(\"div.ingredients-section li\").collect { |s| clean_text(s.text)} if ingredients.empty?\n ingredients\n end",
"def parse_from_website\n log 'Parsing from Website:'\n offer_links = HomepageParser.new(get_body_from('/leistungen/alle-leistungen-2016/index.html')).offer_links\n offer_links.each do |offer_link|\n create_and_parse_offer(offer_link)\n end\n log\n end",
"def initialize(parsed_element = nil)\n @name = ''\n @description = []\n @description_text = ''\n\n build_feature_element(parsed_element) if parsed_element\n end",
"def parse_sections\n AIPP::Downloader.new(storage: AIPP.options.storage, source: sources_file) do |downloader|\n @dependencies.tsort(AIPP.options.section).each do |section|\n info(\"parsing #{section.sectionize}\")\n class_for(section).new(\n downloader: downloader,\n aixm: aixm\n ).attach_patches.tap(&:parse).detach_patches\n end\n end\n end",
"def create_proposal(title, oversight_department, admin_org, opts={})\n \n lo_cat = \"Scientific method\"\n lo_cat_text = \"LO Cat Text\"\n \n defaults = {\n :propose_person => '%%_username%%', #user the dynvar from users.csv\n :mode => 'blank',\n :nav_homepage => true,\n :submit => true,\n :append_unique_id => false, #tell tsung to append unique id on title\n :instructor => 'fred', #BUG - HARDCODED - can't use dynvar though because of ajax search\n :collaborator => @request.config.directory[\"users\"][\"collaborator\"][\"username\"],\n :first_expected_offering => @request.config.directory[\"atp\"][\"name\"],\n :subject_area => \"BSCI\",\n :oversight_dept_number => \"65\", #BUG - right now hardcoded to BSCI, search that returned this was removed\n :course_suffix => \"123\",\n :course_short_title => \"Perf Course\",\n :course_title => title,\n :course_description => \"My fake description.\",\n :course_rationale => \"My fake rationale.\",\n :lo_create => FALSE,\n :lo_category => lo_cat,\n :lo_cat_text => lo_cat_text,\n :lo_name => @request.config.directory[\"lo\"][\"name\"],\n :admin_dep_var_name => \"admin_dep_org_id\",\n :admin_dep_var_regexp => 'org.resultColumn.orgId\\\"\\,\\\"\\([^\\\"]+\\)',\n :proposal_dyn_var_name => \"proposal_id\",\n :proposal_dyn_var_regexp => '\\\"proposal\\\"\\,\\\"workflowNode\\\"\\,\\\"PreRoute\\\"\\,\\\"\\([^\\\"]+\\)',\n :proposal_doc_id_var_name => \"proposal_doc_id\",\n :proposal_doc_id_var_regexp => 'workflowId\\\"\\,\\\"\\([^\\\"]+\\)\\\"',\n :clu_ref_dyn_var_name => \"clu_ref_id\",\n :clu_ref_dyn_var_regexp => '\\\"id\\\"\\,\\\"\\([^\\\"]+\\)',\n :result_com_var_name => \"result_component_id\",\n :result_com_var_regexp => '\\\"ResultComponent 1\\\"\\,\\\"\\([^\\\"]+\\)',\n :enroll_est_var_name => \"default_enrollment_estimate_id\",\n :enroll_est_var_regexp => 'defaultEnrollmentEstimate\\\"\\,\\\"kuali.atp.duration.Week\\\"\\,\\\"Week\\\"\\,\\\"\\([^\\\"]+\\)',\n :lab_var_name => \"lab_id\",\n :lab_var_regexp => 'draft\\\"\\,\\\"unitsContentOwner\\\"\\,\\\"Lab\\\"\\,\\\"\\([^\\\"]+\\)',\n :lab_fee_id_name => 'cp_lab_fee_id',\n :lab_fee_id_regexp => 'kuali.enum.type.feeTypes.labFee\\\"\\,\\\"\\([^\\\"]+\\)',\n :revenues_id_name => 'cp_revenues_id',\n :revenues_id_regexp => 'revenues\\\"\\,\\\"\\([^\\\"]+\\)',\n :revenue_id_name => 'cp_revenue_id',\n :revenue_id_regexp => 'REVENUE\\\"\\,\\\"\\([^\\\"]+\\)',\n :joints_var_name => \"joints_num\",\n :joints_var_regexp => 'joints\\\"\\,\\\"[^\\\"]+\\\"\\,\\\"\\([^\\\"]+\\)',\n :fee_info_id_dyn_var_name => 'fee_info_id',\n :fee_info_id_dyn_var_regexp => '\\\"fees\\\"\\,\\\"\\([^\\\"]+\\)',\n :fee_info_dyn_var_name => 'fee_info',\n :fee_info_dyn_var_regexp => 'org.kuali.student.lum.lu.dto.CluFeeInfo\\\"\\,\\\"\\([^\\\"]+\\)',\n :clu_info_dyn_var_name => 'clu_info',\n :clu_info_dyn_var_regexp => 'org.kuali.student.lum.lu.dto.CluInfo\\\"\\,\\\"\\([^\\\"]+\\)',\n :lu_dto_clu_format_dyn_var_name => \"lu_dto_clu_format\",\n :lu_dto_clu_format_dyn_var_regexp => 'org.kuali.student.lum.lu.dto.CluInfo\\\"\\,\\\"Credit Course\\\"\\,\\\"[^\\\"]+\\\"\\,\\\"formats\\\"\\,\\\"\\([^\\\"]+\\)',\n :lu_dto_clu_activities_dyn_var_name => \"lu_dto_clu_activites\",\n :lu_dto_clu_activities_dyn_var_regexp => 'org.kuali.student.lum.lu.dto.CluInfo\\\"\\,\\\"Credit Course\\\"\\,\\\"[^\\\"]+\\\"\\,\\\"formats\\\"\\,\\\"[^\\\"]+\\\"\\,\\\"0\\\"\\,\\\"activities\\\"\\,\\\"\\([^\\\"]+\\)',\n :outcome_id_var_name => \"outcome_id\",\n :outcome_id_var_regexp => 'outcomeId\\\"\\,\\\"\\([^\\\"]+\\)',\n :lo_category_var_name => \"lo_category\",\n #:lo_category_var_regexp => lo_cat_text + '\\\"\\,\\\"plain\\\"\\,\\\"id\\\"\\,\\\"\\([^\\\"]+\\)',\n :lo_category_var_regexp => lo_cat_text + '\\\"\\,\\\"plain\\\"\\,\\\"\\([^\\\"]+\\)',\n :lo_category_id_var_name => \"lo_category_id\",\n :lo_category_id_var_regexp => 'lo.resultColumn.categoryId\"\\,\\\"\\([^\\\"]+\\)',\n :lo_child_id_dyn_var_name => \"lo_child_id\",\n :lo_child_id_dyn_var_regexp => 'childLo\\\"\\,\\\"\\([^\\\"]+\\)',\n :single_use_lo_dyn_var_name => \"single_use_lo\",\n :single_use_lo_dyn_var_regexp => 'includedSingleUseLo\\\"\\,\\\"\\([^\\\"]+\\)',\n :atp_duration_week_var_name => 'atp_duration_week',\n :atp_duration_week_var_regexp => 'kuali.atp.duration.Week\\\"\\,\\\"Week\\\"\\,\\\"\\([^\\\"]+\\)',\n :version_ind_id_name => 'cp_version_ind_id',\n :version_ind_id_regexp => 'versionIndId\\\"\\,\\\"\\([^\\\"]+\\)',\n :affliated_orgs_id_name => 'cp_affiliated_orgs_id',\n :affliated_orgs_id_regexp => 'affiliatedOrgs\\\"\\,\\\"\\([^\\\"]+\\)',\n :action_request_id_name => 'cp_action_request_id',\n :action_request_id_regexp => 'actionRequestId\\\"\\,\\\"\\([^\\\"]+\\)'\n }\n \n # Version for the doc at each step. We'll increment on each usage\n # So first usage should eval to 0\n version_indicator = -1\n \n opts = defaults.merge(opts)\n \n title << '_%%ts_user_server:get_unique_id%%' if(opts[:append_unique_id])\n \n if(opts[:mode] != \"blank\")\n # select template or copy course...\n end\n \n # Navigate to Curriculum Mgmt\n self.homepage() unless(!opts[:nav_homepage])\n \n puts \"creating proposal as: #{opts[:propose_person]}\"\n \n # Create a course\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SecurityRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|7|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|13BFCB3640903B473D12816447D1469D|org.kuali.student.common.ui.client.service.SecurityRpcService|checkAdminPermission|java.lang.String/2004016611|#{opts[:propose_person]}|useCurriculumReview|1|2|3|4|2|5|5|6|7|\"\n }, {'subst' => 'true'}\n )\n \n \n # Click Start blank proposal\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/CreditCourseProposalRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|15|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|D60D3C6E0D395C18A0F44A2D9D2A7348|org.kuali.student.lum.lu.ui.course.client.service.CreditCourseProposalRpcService|getMetadata|java.lang.String/2004016611|java.util.Map||java.util.HashMap/962170901|documentTypeName|kuali.proposal.type.course.create|DtoState|Draft|DtoNextState|DtoWorkflowNode|PreRoute|1|2|3|4|2|5|6|7|8|4|5|9|5|10|5|11|5|12|5|13|5|7|5|14|5|15|\"\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/statementRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|2543249A07E7952228E0E500F14F1B17|org.kuali.student.lum.program.client.rpc.StatementRpcService|getStatementTypesForStatementTypeForCourse|java.lang.String/2004016611|kuali.statement.type.course|1|2|3|4|1|5|6|\"\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|11|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|enumeration.queryParam.enumerationType|kuali.lu.campusLocation|enumeration.management.search|enumeration.resultColumn.sortKey|1|2|3|4|1|5|5|0|0|6|1|7|8|0|9|10|11|0|0|\"\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|8|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|atp.search.atpSeasonTypes|atp.resultColumn.atpSeasonTypeName|1|2|3|4|1|5|5|0|0|6|0|7|8|0|0|\"\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|8|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|atp.search.atpDurationTypes|atp.resultColumn.atpDurTypeName|1|2|3|4|1|5|5|0|0|6|0|7|8|0|0|\"\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|18|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|lrc.queryParam.resultComponent.type|kuali.resultComponentType.grade.finalGrade|lrc.queryParam.resultComponent.idRestrictionList|java.lang.String/2004016611|kuali.resultComponent.grade.letter|kuali.resultComponent.grade.passFail|kuali.resultComponent.grade.satisfactory|kuali.resultComponent.grade.completedNotation|kuali.resultComponent.grade.percentage|lrc.search.resultComponent|lrc.resultColumn.resultComponent.id|1|2|3|4|1|5|5|0|0|6|2|7|8|0|9|7|10|6|5|11|12|11|13|11|14|11|15|11|16|0|17|18|0|0|\"\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|11|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|enumeration.queryParam.enumerationType|kuali.lu.finalExam.status|enumeration.management.search|enumeration.resultColumn.sortKey|1|2|3|4|1|5|5|0|0|6|1|7|8|0|9|10|11|0|0|\"\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|14|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|lrc.queryParam.resultComponent.idRestrictionList|java.lang.String/2004016611|kuali.resultComponentType.credit.degree.fixed|kuali.resultComponentType.credit.degree.range|kuali.resultComponentType.credit.degree.multiple|lrc.search.resultComponentType|lrc.resultColumn.resultComponent.id|1|2|3|4|1|5|5|0|0|6|1|7|8|6|3|9|10|9|11|9|12|0|13|14|0|0|\"\n }\n )\n \n # DUPE\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|14|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|lrc.queryParam.resultComponent.idRestrictionList|java.lang.String/2004016611|kuali.resultComponentType.credit.degree.fixed|kuali.resultComponentType.credit.degree.range|kuali.resultComponentType.credit.degree.multiple|lrc.search.resultComponentType|lrc.resultColumn.resultComponent.id|1|2|3|4|1|5|5|0|0|6|1|7|8|6|3|9|10|9|11|9|12|0|13|14|0|0|\"\n }\n )\n \n # DUPE\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|14|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|lrc.queryParam.resultComponent.idRestrictionList|java.lang.String/2004016611|kuali.resultComponentType.credit.degree.fixed|kuali.resultComponentType.credit.degree.range|kuali.resultComponentType.credit.degree.multiple|lrc.search.resultComponentType|lrc.resultColumn.resultComponent.id|1|2|3|4|1|5|5|0|0|6|1|7|8|6|3|9|10|9|11|9|12|0|13|14|0|0|\"\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|18|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|atp.advancedAtpSearchParam.atpType|java.lang.String/2004016611|kuali.atp.type.Spring|kuali.atp.type.Summer|kuali.atp.type.Fall|kuali.atp.type.Session1|kuali.atp.type.Session2|kuali.atp.type.Mini-mester1A|kuali.atp.type.Mini-mester1B|atp.search.advancedAtpSearch|atp.resultColumn.atpStartDate|1|2|3|4|1|5|5|0|0|6|1|7|8|6|7|9|10|9|11|9|12|9|13|9|14|9|15|9|16|0|17|18|0|0|\"\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|11|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|enumeration.queryParam.enumerationType|kuali.lu.fee.rateType|enumeration.management.search|enumeration.resultColumn.sortKey|1|2|3|4|1|5|5|0|0|6|1|7|8|0|9|10|11|0|0|\"\n }\n )\n \n # DUPE\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|11|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|enumeration.queryParam.enumerationType|kuali.lu.fee.rateType|enumeration.management.search|enumeration.resultColumn.sortKey|1|2|3|4|1|5|5|0|0|6|1|7|8|0|9|10|11|0|0|\"\n }\n )\n \n # DUPE\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|11|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|enumeration.queryParam.enumerationType|kuali.lu.fee.rateType|enumeration.management.search|enumeration.resultColumn.sortKey|1|2|3|4|1|5|5|0|0|6|1|7|8|0|9|10|11|0|0|\"\n }\n )\n \n # DUPE\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|11|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|enumeration.queryParam.enumerationType|kuali.lu.fee.rateType|enumeration.management.search|enumeration.resultColumn.sortKey|1|2|3|4|1|5|5|0|0|6|1|7|8|0|9|10|11|0|0|\"\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|11|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|enumeration.queryParam.enumerationType|kuali.lu.campusLocation|enumeration.management.search|enumeration.resultColumn.sortKey|1|2|3|4|1|5|5|0|0|6|1|7|8|0|9|10|11|0|0|\"\n }\n )\n \n # DUPE\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|11|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|enumeration.queryParam.enumerationType|kuali.lu.campusLocation|enumeration.management.search|enumeration.resultColumn.sortKey|1|2|3|4|1|5|5|0|0|6|1|7|8|0|9|10|11|0|0|\"\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|8|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|atp.search.atpSeasonTypes|atp.resultColumn.atpSeasonTypeName|1|2|3|4|1|5|5|0|0|6|0|7|8|0|0|\"\n }\n )\n \n # DUPE\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|8|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|atp.search.atpSeasonTypes|atp.resultColumn.atpSeasonTypeName|1|2|3|4|1|5|5|0|0|6|0|7|8|0|0|\"\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|18|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|lrc.queryParam.resultComponent.type|kuali.resultComponentType.grade.finalGrade|lrc.queryParam.resultComponent.idRestrictionList|java.lang.String/2004016611|kuali.resultComponent.grade.letter|kuali.resultComponent.grade.passFail|kuali.resultComponent.grade.satisfactory|kuali.resultComponent.grade.completedNotation|kuali.resultComponent.grade.percentage|lrc.search.resultComponent|lrc.resultColumn.resultComponent.id|1|2|3|4|1|5|5|0|0|6|2|7|8|0|9|7|10|6|5|11|12|11|13|11|14|11|15|11|16|0|17|18|0|0|\"\n }\n )\n \n # DUPE\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|18|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|lrc.queryParam.resultComponent.type|kuali.resultComponentType.grade.finalGrade|lrc.queryParam.resultComponent.idRestrictionList|java.lang.String/2004016611|kuali.resultComponent.grade.letter|kuali.resultComponent.grade.passFail|kuali.resultComponent.grade.satisfactory|kuali.resultComponent.grade.completedNotation|kuali.resultComponent.grade.percentage|lrc.search.resultComponent|lrc.resultColumn.resultComponent.id|1|2|3|4|1|5|5|0|0|6|2|7|8|0|9|7|10|6|5|11|12|11|13|11|14|11|15|11|16|0|17|18|0|0|\"\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/CommentRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|58FCBA6F511FF769D9DF8CAE72C3C369|org.kuali.student.core.comments.ui.client.service.CommentRpcService|getUserRealName|java.lang.String/2004016611|#{opts[:propose_person]}|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|11|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|subjectCode.queryParam.code||subjectCode.search.orgsForSubjectCode|subjectCode.resultColumn.orgLongName|1|2|3|4|1|5|5|0|0|6|1|7|8|0|9|10|11|0|0|\"\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|19|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|atp.advancedAtpSearchParam.atpType|java.lang.String/2004016611|kuali.atp.type.Spring|kuali.atp.type.Summer|kuali.atp.type.Fall|kuali.atp.type.Session1|kuali.atp.type.Session2|kuali.atp.type.Mini-mester1A|kuali.atp.type.Mini-mester1B|atp.advancedAtpSearchParam.atpStartDateAtpConstraintId|atp.search.advancedAtpSearch|atp.resultColumn.atpStartDate|1|2|3|4|1|5|5|0|0|6|2|7|8|6|7|9|10|9|11|9|12|9|13|9|14|9|15|9|16|0|7|17|0|0|18|19|0|0|\"\n }\n )\n \n \n #\n # Pg1 - Course Information\n #\n\n @request.add_thinktime(5)\n\n # Course Subject Area\n # AJAX popup while typing in subject area\n for i in 1..opts[:subject_area].length\n itr = i-1\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|12|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|search|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.lang.Boolean/476441737|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|subjectCode.queryParam.code|#{opts[:subject_area][0..itr]}|subjectCode.search.subjectCodeGeneric|subjectCode.resultColumn.code|1|2|3|4|1|5|5|0|6|0|7|1|8|9|0|10|11|12|0|0|\"\n } \n ) \n end\n\n @request.add_thinktime(3)\n\n # Instructor\n for i in 1..opts[:instructor].length\n itr = i-1\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|12|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|search|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.lang.Boolean/476441737|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|person.queryParam.personGivenName|#{opts[:instructor][0..itr]}|person.search.personQuickViewByGivenName|person.resultColumn.DisplayName|1|2|3|4|1|5|5|0|6|0|7|1|8|9|0|10|11|12|0|0|\"\n } \n ) \n end\n\n\n @request.add_thinktime(22)\n\n # Save & Continue\n contents1 = \"5|0|41|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|D60D3C6E0D395C18A0F44A2D9D2A7348|org.kuali.student.lum.lu.ui.course.client.service.CreditCourseProposalRpcService|saveData|org.kuali.student.common.assembly.data.Data/3184510345|org.kuali.student.common.assembly.data.Data|java.util.LinkedHashMap/1551059846|org.kuali.student.common.assembly.data.Data$StringKey/758802082|proposal|org.kuali.student.common.assembly.data.Data$DataValue/1692468409|type|org.kuali.student.common.assembly.data.Data$StringValue/3151113388|kuali.proposal.type.course.create|workflowNode|PreRoute|name|#{opts[:course_title]}|_runtimeData|dirty|org.kuali.student.common.assembly.data.Data$BooleanValue/4261226833|java.lang.Boolean/476441737|rationale|#{opts[:course_rationale]}|courseTitle|#{opts[:course_rationale]}|transcriptTitle|subjectArea|courseNumberSuffix|instructors|#{opts[:course_short_title]}|#{opts[:subject_area]}|#{opts[:course_suffix]}|org.kuali.student.common.assembly.data.Data$IntegerKey/134469241|java.lang.Integer/3438268394|personId|#{opts[:instructor]}|id-translation|#{opts[:instructor]}, #{opts[:instructor]}(#{opts[:instructor]})|descr|plain|#{opts[:course_description]}\"\n contents2 = \"|1|2|3|4|1|5|5|6|7|0|8|8|9|10|5|6|7|0|5|8|11|12|13|8|14|12|15|8|16|12|17|8|18|10|5|6|7|0|1|8|19|10|5|6|7|0|2|-11|20|21|1|8|22|20|-22|-15|-17|-5|-13|-23|12|23|-1|-3|8|24|12|25|-13|10|5|6|7|0|1|-17|10|5|6|7|0|5|-26|20|-22|8|26|20|-22|8|27|20|-22|8|28|20|-22|8|29|20|-22|-29|-17|-1|-13|-35|12|30|-37|12|31|-39|12|32|-41|10|5|6|7|0|1|33|34|0|10|5|6|7|0|2|8|35|12|36|8|18|10|5|6|7|0|1|8|35|10|5|6|7|0|1|8|37|12|38|-58|-60|-52|-56|-47|-49|-1|8|29|8|39|10|5|6|7|0|2|8|40|12|41|-13|10|5|6|7|0|1|-17|10|5|6|7|0|1|-71|20|-22|-74|-17|-69|-13|-1|-67|0|0|\"\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/CreditCourseProposalRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"#{contents1}#{contents2}\"\n },\n {\n :dyn_variables => [\n {\"name\" => opts[:proposal_dyn_var_name], \"regexp\" => opts[:proposal_dyn_var_regexp]},\n {\"name\" => opts[:clu_ref_dyn_var_name], \"regexp\" => opts[:clu_ref_dyn_var_regexp]},\n {\"name\" => opts[:proposal_doc_id_var_name], \"regexp\" => opts[:proposal_doc_id_var_regexp]},\n {\"name\" => opts[:version_ind_id_name], \"regexp\" => opts[:version_ind_id_regexp]}\n ]\n }\n )\n\n #@request.add(\"DEBUG/proposal_dyn_var_name/%%_#{opts[:proposal_dyn_var_name]}%%\", {}, {'subst' => 'true'})\n #@request.add(\"DEBUG/clu_ref_dyn_var_name/%%_#{opts[:clu_ref_dyn_var_name]}%%\", {}, {'subst' => 'true'})\n #@request.add(\"DEBUG/proposal_doc_id_var_name/%%_#{opts[:proposal_doc_id_var_name]}%%\", {}, {'subst' => 'true'})\n #@request.add(\"DEBUG/version_ind_id_name/%%_#{opts[:version_ind_id_name]}%%\", {}, {'subst' => 'true'})\n \n \n \n #\n # Pg 2 - Governance\n # Campus Locations: All\n #\n\n @request.add_thinktime(5)\n\n # COC Org\n # Biology Dept\n\n @request.add_thinktime(3)\n\n # Admin Org\n # Botany\n for i in 1..admin_org.length\n if(i == admin_org.length)\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|16|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|search|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.lang.Boolean/476441737|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|org.queryParam.orgOptionalLongName|#{admin_org[0..itr]}|org.queryParam.orgOptionalType|java.lang.String/2004016611|kuali.org.Department|kuali.org.College|org.search.generic||1|2|3|4|1|5|5|0|6|0|7|2|8|9|0|10|8|11|7|2|12|13|12|14|0|15|16|0|0|\"\n },\n {\n :dyn_variables => [\n {\"name\" => opts[:admin_dep_var_name], \"regexp\" => opts[:admin_dep_var_regexp]}\n ]\n }\n )\n\n #@request.add(\"DEBUG/admin_dep_var_name/%%_#{opts[:admin_dep_var_name]}%%\", {}, {'subst' => 'true'})\n else\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|16|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|search|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.lang.Boolean/476441737|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|org.queryParam.orgOptionalLongName|#{admin_org[0..itr]}|org.queryParam.orgOptionalType|java.lang.String/2004016611|kuali.org.Department|kuali.org.College|org.search.generic||1|2|3|4|1|5|5|0|6|0|7|2|8|9|0|10|8|11|7|2|12|13|12|14|0|15|16|0|0|\"\n } \n )\n end \n end\n\n @request.add_thinktime(15)\n\n # Save & Continue\n contents1 = \"5|0|101|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|D60D3C6E0D395C18A0F44A2D9D2A7348|org.kuali.student.lum.lu.ui.course.client.service.CreditCourseProposalRpcService|saveData|org.kuali.student.common.assembly.data.Data/3184510345|org.kuali.student.common.assembly.data.Data|java.util.LinkedHashMap/1551059846|org.kuali.student.common.assembly.data.Data$StringKey/758802082|campusLocations|org.kuali.student.common.assembly.data.Data$DataValue/1692468409|org.kuali.student.common.assembly.data.Data$IntegerKey/134469241|java.lang.Integer/3438268394|org.kuali.student.common.assembly.data.Data$StringValue/3151113388|AL|code|#{opts[:subject_area]}#{opts[:course_suffix]}|courseNumberSuffix|#{opts[:course_suffix]}|courseSpecificLOs|courseTitle|#{opts[:course_title]}|creditOptions|crossListings|descr|plain|#{opts[:course_description]}|expenditure|affiliatedOrgs|fees|formats|gradingOptions|id|%%_#{opts[:clu_ref_dyn_var_name]}%%|instructors|personId|#{opts[:instructor]}|_runtimeData|id-translation|#{opts[:instructor]}, #{opts[:instructor]}(#{opts[:instructor]})|joints|level|100|metaInfo|createId|#{opts[:propose_person]}|createTime|org.kuali.student.common.assembly.data.Data$DateValue/2929953165|java.util.Date/1659716317|updateId|updateTime|versionInd|#{version_indicator+=1}|pilotCourse|org.kuali.student.common.assembly.data.Data$BooleanValue/4261226833|java.lang.Boolean/476441737|revenues|specialTopicsCourse|state|draft|subjectArea|#{opts[:subject_area]}|termsOffered|transcriptTitle|#{opts[:course_short_title]}|type|kuali.lu.type.CreditCourse|unitsContentOwner|#{opts[:oversight_dept_number]}|#{oversight_department}|unitsDeployment|%%_#{opts[:admin_dep_var_name]}%%|#{admin_org}|variations|versionInfo|currentVersionStart|sequenceNumber|org.kuali.student.common.assembly.data.Data$LongValue/3784756947|java.lang.Long/4227064769|versionIndId|%%_#{opts[:version_ind_id_name]}%%|dirty|proposal|workflowNode|PreRoute|%%_#{opts[:proposal_dyn_var_name]}%%|2|name|#{opts[:course_title]}|proposalReference|proposalReferenceType|kuali.proposal.referenceType.clu|proposerOrg|proposerPerson|rationale|#{opts[:course_rationale]}|Saved|kuali.proposal.type.course.create|workflowId|%%_#{opts[:proposal_doc_id_var_name]}%%|collaboratorInfo|collaborators\"\n contents2 = \"|1|2|3|4|1|5|5|6|7|0|32|8|9|10|5|6|7|0|1|11|12|0|13|14|-1|8|9|8|15|13|16|8|17|13|18|8|19|10|5|6|7|0|0|-1|-15|8|20|13|21|8|22|10|5|6|7|0|0|-1|-21|8|23|10|5|6|7|0|0|-1|-25|8|24|10|5|6|7|0|1|8|25|13|26|-1|-29|8|27|10|5|6|7|0|1|8|28|10|5|6|7|0|0|-37|-39|-1|-35|8|29|10|5|6|7|0|0|-1|-43|8|30|10|5|6|7|0|0|-1|-47|8|31|10|5|6|7|0|0|-1|-51|8|32|13|33|8|34|10|5|6|7|0|1|11|-8|10|5|6|7|0|2|8|35|13|36|8|37|10|5|6|7|0|1|8|35|10|5|6|7|0|1|8|38|13|39|-69|-71|-63|-67|-59|-61|-1|-57|8|40|10|5|6|7|0|0|-1|-77|8|41|13|42|8|43|10|5|6|7|0|5|8|44|13|45|8|46|47|48|3467218721|1309965025280|8|49|13|45|8|50|47|48|3467218721|1309965025280|8|51|13|52|-1|-83|8|53|54|55|0|8|56|10|5|6|7|0|0|-1|-102|8|57|54|-101|8|58|13|59|8|60|13|61|8|62|10|5|6|7|0|0|-1|-112|8|63|13|64|8|65|13|66|8|67|10|5|6|7|0|2|11|-8|13|68|8|37|10|5|6|7|0|1|11|-8|10|5|6|7|0|1|8|38|13|69|-128|-130|-122|-126|-1|8|67|8|70|10|5|6|7|0|2|11|-8|13|71|8|37|10|5|6|7|0|1|11|-8|10|5|6|7|0|1|8|38|13|72|-145|-147|-139|-143|-1|8|70|8|73|10|5|6|7|0|0|-1|-154|8|74|10|5|6|7|0|3|8|75|47|48|3467218721|1309965025280|8|76|77|78|1|0|8|79|13|80|-1|-158|8|37|10|5|6|7|0|2|8|60|10|5|6|7|0|1|8|38|13|61|-172|-174|8|81|10|5|6|7|0|3|8|9|54|55|1|8|67|54|-186|8|70|54|-186|-172|-180|-1|-170|8|82|10|5|6|7|0|12|8|83|13|84|8|32|13|85|8|43|10|5|6|7|0|5|8|44|13|45|8|46|47|48|3467218994|1309965025280|8|49|13|45|8|50|47|48|3467219879|1309965025280|8|51|13|86|-193|-199|8|87|13|88|8|89|10|5|6|7|0|1|11|-8|13|33|-193|-217|8|90|13|91|8|92|10|5|6|7|0|0|-193|-225|8|93|10|5|6|7|0|0|-193|-229|8|94|13|95|8|58|13|96|8|65|13|97|8|98|13|99|-1|-191|8|100|10|5|6|7|0|1|8|101|10|5|6|7|0|0|-243|-245|-1|-241|0|0|\"\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/CreditCourseProposalRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"#{contents1}#{contents2}\"\n },\n {\n 'subst' => 'true'\n }\n )\n\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|11|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|lu.queryParam.luOptionalLuTypeStartsWith|kuali.lu.type.activity.|lu.search.all.lu.Types|lu.resultColumn.luTypeName|1|2|3|4|1|5|5|0|0|6|1|7|8|0|9|10|11|0|0|\"\n }\n )\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|11|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|enumeration.queryParam.enumerationType|kuali.atptype.duration|enumeration.management.search|enumeration.resultColumn.sortKey|1|2|3|4|1|5|5|0|0|6|1|7|8|0|9|10|11|0|0|\"\n }\n )\n \n \n \n #\n # Course Logistics \n # Term: Any\n # Duration Type: Semester\n # Duration Count: 2\n # Assessment Scale: Letter\n # Standard Final Exam\n # Outcome: 10 credits\n # Course Format\n # => Activity Type: Lab, Contact Hours: 5, Frequency: weekly\n # => Duration Type: Weekly\n # => Duration: 13\n # => Anticipated class size: 100\n\n @request.add_thinktime(30)\n\n # Save & Continue\n contents1 = \"5|0|126|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|D60D3C6E0D395C18A0F44A2D9D2A7348|org.kuali.student.lum.lu.ui.course.client.service.CreditCourseProposalRpcService|saveData|org.kuali.student.common.assembly.data.Data/3184510345|org.kuali.student.common.assembly.data.Data|java.util.LinkedHashMap/1551059846|org.kuali.student.common.assembly.data.Data$StringKey/758802082|campusLocations|org.kuali.student.common.assembly.data.Data$DataValue/1692468409|org.kuali.student.common.assembly.data.Data$IntegerKey/134469241|java.lang.Integer/3438268394|org.kuali.student.common.assembly.data.Data$StringValue/3151113388|AL|_runtimeData|id-translation|All|code|#{opts[:subject_area]}#{opts[:course_suffix]}|courseNumberSuffix|#{opts[:course_suffix]}|courseSpecificLOs|courseTitle|#{opts[:course_title]}|creditOptions|dirty|type|org.kuali.student.common.assembly.data.Data$BooleanValue/4261226833|java.lang.Boolean/476441737|fixedCreditValue|created|kuali.resultComponentType.credit.degree.fixed|10|crossListings|descr|plain|#{opts[:course_description]}|expenditure|affiliatedOrgs|fees|formats|activities|activityType|defaultEnrollmentEstimate|kuali.lu.type.activity.Lab|contactHours|unitQuantity|5|unitType|kuali.atp.duration.week|duration|atpDurationTypeKey|timeQuantity|kuali.atp.duration.Week|org.kuali.student.common.assembly.data.Data$IntegerValue/3605481012|gradingOptions|kuali.resultComponent.grade.letter|id|%%_#{opts[:clu_ref_dyn_var_name]}%%|instructors|personId|#{opts[:instructor]}|#{opts[:instructor]}, #{opts[:instructor]}(#{opts[:instructor]})|joints|level|100|metaInfo|createId|#{opts[:propose_person]}|createTime|org.kuali.student.common.assembly.data.Data$DateValue/2929953165|java.sql.Timestamp/1769758459|updateId|updateTime|versionInd|#{version_indicator+=1}|pilotCourse|revenues|specialTopicsCourse|state|draft|subjectArea|#{opts[:subject_area]}|termsOffered|kuali.atp.season.Any|transcriptTitle|#{opts[:course_short_title]}|kuali.lu.type.CreditCourse|unitsContentOwner|#{opts[:oversight_dept_number]}|#{oversight_department}|unitsDeployment|%%_#{opts[:admin_dep_var_name]}%%|#{admin_org}|variations|versionInfo|currentVersionStart|sequenceNumber|org.kuali.student.common.assembly.data.Data$LongValue/3784756947|java.lang.Long/4227064769|versionIndId|%%_#{opts[:version_ind_id_name]}%%|finalExamStatus|audit|passFail|proposal|workflowNode|PreRoute|%%_#{opts[:proposal_dyn_var_name]}%%|3|name|proposalReference|proposalReferenceType|kuali.proposal.referenceType.clu|proposerOrg|proposerPerson|rationale|#{opts[:course_rationale]}|Saved|kuali.proposal.type.course.create|workflowId|%%_#{opts[:proposal_doc_id_var_name]}%%|collaboratorInfo|collaborators|kuali.atp.duration.Semester|STD\"\n contents2 = \"|1|2|3|4|1|5|5|6|7|0|36|8|9|10|5|6|7|0|2|11|12|0|13|14|8|15|10|5|6|7|0|1|11|-8|10|5|6|7|0|1|8|16|13|17|-12|-14|-5|-10|-1|-3|8|18|13|19|8|20|13|21|8|22|10|5|6|7|0|0|-1|-24|8|23|13|24|8|25|10|5|6|7|0|1|11|-8|10|5|6|7|0|3|8|15|10|5|6|7|0|3|8|26|10|5|6|7|0|2|8|27|28|29|1|8|30|28|-48|-40|-42|8|31|28|-48|-38|10|5|6|7|0|1|-42|10|5|6|7|0|1|-51|28|-48|-54|-42|-40|-38|-36|-38|-46|13|32|-49|13|33|-32|-34|-1|-30|8|34|10|5|6|7|0|0|-1|-62|8|35|10|5|6|7|0|1|8|36|13|37|-1|-66|8|38|10|5|6|7|0|1|8|39|10|5|6|7|0|0|-74|-76|-1|-72|8|40|10|5|6|7|0|0|-1|-80|8|41|10|5|6|7|0|1|11|-8|10|5|6|7|0|2|8|42|10|5|6|7|0|1|11|-8|10|5|6|7|0|5|-38|10|5|6|7|0|3|-42|10|5|6|7|0|2|8|43|28|-48|8|44|28|-48|-101|-42|8|31|28|-48|-38|10|5|6|7|0|1|-42|10|5|6|7|0|1|-110|28|-48|-113|-42|-101|-38|-98|-38|-106|13|45|8|46|10|5|6|7|0|3|8|47|13|48|-38|10|5|6|7|0|1|-42|10|5|6|7|0|2|-124|28|-48|8|49|28|-48|-127|-42|-122|-38|-133|13|50|-98|-120|8|51|10|5|6|7|0|3|-38|10|5|6|7|0|1|-42|10|5|6|7|0|2|8|52|28|-48|8|53|28|-48|-141|-42|-138|-38|-146|13|54|-148|55|12|13|-98|-136|-108|55|12|100|-94|-96|-90|-92|8|15|10|5|6|7|0|2|8|31|28|-48|-38|10|5|6|7|0|1|-42|10|5|6|7|0|1|-159|28|-48|-162|-42|-157|-38|-90|-155|-86|-88|-1|-84|8|56|10|5|6|7|0|1|11|-8|13|57|-1|8|56|8|58|13|59|8|60|10|5|6|7|0|1|11|-8|10|5|6|7|0|2|8|61|13|62|8|15|10|5|6|7|0|1|8|61|10|5|6|7|0|1|8|16|13|63|-189|-191|-183|-187|-179|-181|-1|-177|8|64|10|5|6|7|0|0|-1|-197|8|65|13|66|8|67|10|5|6|7|0|5|8|68|13|69|8|70|71|72|3469526397|1309965025280|677000000|8|73|13|69|8|74|71|72|3469529329|1309965025280|609000000|8|75|13|76|-1|-203|8|77|28|29|0|8|78|10|5|6|7|0|0|-1|-222|8|79|28|-221|8|80|13|81|8|82|13|83|8|84|10|5|6|7|0|1|11|-8|13|85|-1|8|84|8|86|13|87|8|27|13|88|8|89|10|5|6|7|0|2|11|-8|13|90|8|15|10|5|6|7|0|1|11|-8|10|5|6|7|0|1|8|16|13|91|-251|-253|-245|-249|-1|-243|8|92|10|5|6|7|0|2|11|-8|13|93|8|15|10|5|6|7|0|1|11|-8|10|5|6|7|0|1|8|16|13|94|-267|-269|-261|-265|-1|-259|8|95|10|5|6|7|0|0|-1|-275|8|96|10|5|6|7|0|3|8|97|71|72|3469526397|1309965025280|677000000|8|98|99|100|1|0|8|101|13|102|-1|-279|8|15|10|5|6|7|0|2|8|82|10|5|6|7|0|1|8|16|13|83|-293|-295|-42|10|5|6|7|0|5|8|84|28|-48|8|56|28|-48|8|103|28|-48|8|104|28|-48|8|105|28|-48|-293|-42|-1|-291|8|106|10|5|6|7|0|12|8|107|13|108|8|58|13|109|8|67|10|5|6|7|0|5|8|68|13|69|8|70|71|72|3469526678|1309965025280|958000000|8|73|13|69|8|74|71|72|3469530172|1309965025280|452000000|8|75|13|110|-316|-322|8|111|13|24|8|112|10|5|6|7|0|1|11|-8|13|59|-316|-340|8|113|13|114|8|115|10|5|6|7|0|0|-316|-348|8|116|10|5|6|7|0|0|-316|-352|8|117|13|118|8|80|13|119|8|27|13|120|8|121|13|122|-1|-314|8|123|10|5|6|7|0|1|8|124|10|5|6|7|0|0|-366|-368|-1|-364|8|51|10|5|6|7|0|3|-38|10|5|6|7|0|1|-42|10|5|6|7|0|2|8|52|28|-48|8|53|28|-48|-377|-42|-374|-38|-382|13|125|-384|55|12|2|-1|-372|-308|13|126|-310|28|-221|-312|28|-221|0|0|\"\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/CreditCourseProposalRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"#{contents1}#{contents2}\"\n },\n {\n 'subst' => 'true',\n :dyn_variables => [\n {\"name\" => opts[:enroll_est_var_name], \"regexp\" => opts[:enroll_est_var_regexp]},\n {\"name\" => opts[:lab_var_name], \"regexp\" => opts[:lab_var_regexp]},\n {\"name\" => opts[:atp_duration_week_var_name], \"regexp\" => opts[:atp_duration_week_var_regexp]}\n ]\n }\n )\n\n #@request.add(\"DEBUG/enroll_est_var_name/%%_#{opts[:enroll_est_var_name]}%%\", {}, {'subst' => 'true'})\n #@request.add(\"DEBUG/lab_var_name/%%_#{opts[:lab_var_name]}%%\", {}, {'subst' => 'true'})\n #@request.add(\"DEBUG/atp_duration_week_var_name/%%_#{opts[:atp_duration_week_var_name]}%%\", {}, {'subst' => 'true'})\n\n @request.add_thinktime(2)\n \n \n \n #\n # Learning Objectives\n #\n\n @request.add_thinktime(5)\n\n # Category\n for i in 1..opts[:lo_category].length\n if(i == opts[:lo_category].length)\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|11|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|search|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.lang.Boolean/476441737|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|lo.queryParam.loOptionalCategoryName|#{opts[:lo_category][0..itr]}|lo.search.loCategories|1|2|3|4|1|5|5|0|6|0|7|1|8|9|0|10|11|0|0|0|\"\n },\n {\n :dyn_variables => [\n {\"name\" => opts[:lo_category_id_var_name], \"regexp\" => opts[:lo_category_id_var_regexp]}\n ]\n }\n )\n\n #@request.add(\"DEBUG/lo_category_id_var_name/%%_#{opts[:lo_category_id_var_name]}%%\", {}, {'subst' => 'true'})\n\n else\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|11|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|search|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.lang.Boolean/476441737|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|lo.queryParam.loOptionalCategoryName|#{opts[:lo_category][0..itr]}|lo.search.loCategories|1|2|3|4|1|5|5|0|6|0|7|1|8|9|0|10|11|0|0|0|\"\n } \n )\n end \n end\n\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/LoCategoryRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|EDE1DECBD54F5894284944CD8AC3661C|org.kuali.student.lum.common.client.lo.rpc.LoCategoryRpcService|getData|java.lang.String/2004016611|%%_#{opts[:lo_category_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {'subst' => 'true'}\n )\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/LoCategoryRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|EDE1DECBD54F5894284944CD8AC3661C|org.kuali.student.lum.common.client.lo.rpc.LoCategoryRpcService|getLoCategoryType|java.lang.String/2004016611|loCategoryType.subject|1|2|3|4|1|5|6|\"\n }\n )\n\n @request.add_thinktime(25)\n\n # Save & Continue\n contents1 = \"5|0|154|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|D60D3C6E0D395C18A0F44A2D9D2A7348|org.kuali.student.lum.lu.ui.course.client.service.CreditCourseProposalRpcService|saveData|org.kuali.student.common.assembly.data.Data/3184510345|org.kuali.student.common.assembly.data.Data|java.util.LinkedHashMap/1551059846|org.kuali.student.common.assembly.data.Data$StringKey/758802082|passFail|org.kuali.student.common.assembly.data.Data$BooleanValue/4261226833|java.lang.Boolean/476441737|audit|finalExamStatus|org.kuali.student.common.assembly.data.Data$StringValue/3151113388|STD|campusLocations|org.kuali.student.common.assembly.data.Data$DataValue/1692468409|org.kuali.student.common.assembly.data.Data$IntegerKey/134469241|java.lang.Integer/3438268394|AL|_runtimeData|id-translation|All|code|#{opts[:subject_area]}#{opts[:course_suffix]}|courseNumberSuffix|#{opts[:course_suffix]}|courseSpecificLOs|loInfo|id|desc|formatted|#{opts[:lo_cat_text]}|plain|name|SINGLE USE LO|sequence|0|metaInfo|loCategoryInfoList|%%_#{opts[:lo_category_id_var_name]}%%|#{opts[:lo_category]}|loRepository|kuali.loRepository.key.singleUse|effectiveDate|org.kuali.student.common.assembly.data.Data$DateValue/2929953165|expirationDate|state|active|type|loCategoryType.subject|createId|admin|createTime|java.sql.Timestamp/1769758459|updateId|updateTime|versionInd|courseTitle|#{opts[:course_title]}|creditOptions|fixedCreditValue|10.0|kuali.creditType.credit.degree.10.0|#{opts[:propose_person]}|resultValues|draft|kuali.resultComponentType.credit.degree.fixed|Credits, Fixed|crossListings|descr|#{opts[:course_description]}|duration|atpDurationTypeKey|kuali.atp.duration.Semester|timeQuantity|org.kuali.student.common.assembly.data.Data$IntegerValue/3605481012|Semester|expenditure|affiliatedOrgs|fees|formats|activities|activityType|kuali.lu.type.activity.Lab|contactHours|unitQuantity|5|unitType|kuali.atp.duration.week|per week|defaultEnrollmentEstimate|kuali.atp.duration.Week|Week|%%_#{opts[:atp_duration_week_var_name]}%%|unitsContentOwner|Lab|%%_#{opts[:lab_var_name]}%%|termsOffered|kuali.lu.type.CreditCourseFormatShell|gradingOptions|kuali.resultComponent.grade.letter|Letter|%%_#{opts[:clu_ref_dyn_var_name]}%%|instructors|personId|#{opts[:instructor]}|#{opts[:instructor]}, #{opts[:instructor]}(#{opts[:instructor]})|joints|level|100|2|pilotCourse|revenues|specialTopicsCourse|subjectArea|#{opts[:subject_area]}|kuali.atp.season.Any|Any|transcriptTitle|#{opts[:course_short_title]}|kuali.lu.type.CreditCourse|#{opts[:oversight_dept_number]}|#{oversight_department}|unitsDeployment|%%_#{opts[:admin_dep_var_name]}%%|#{admin_org}|variations|versionInfo|currentVersionStart|sequenceNumber|org.kuali.student.common.assembly.data.Data$LongValue/3784756947|java.lang.Long/4227064769|versionIndId|%%_#{opts[:version_ind_id_name]}%%|Standard final Exam|proposal|workflowNode|PreRoute|%%_#{opts[:proposal_dyn_var_name]}%%|4|proposalReference|proposalReferenceType|kuali.proposal.referenceType.clu|proposerOrg|proposerPerson|rationale|#{opts[:course_rationale]}|Saved|kuali.proposal.type.course.create|workflowId|%%_#{opts[:proposal_doc_id_var_name]}%%|collaboratorInfo|collaborators\"\n contents2 = \"|1|2|3|4|1|5|5|6|7|0|36|8|9|10|11|0|8|12|10|-5|8|13|14|15|8|16|17|5|6|7|0|2|18|19|0|14|20|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|23|-19|-21|-12|-17|-1|-10|8|24|14|25|8|26|14|27|8|28|17|5|6|7|0|1|18|-15|17|5|6|7|0|2|8|29|17|5|6|7|0|5|8|30|14|0|8|31|17|5|6|7|0|2|8|32|14|33|8|34|14|33|-41|-45|8|35|14|36|8|37|14|38|8|39|17|0|-37|-39|8|40|17|5|6|7|0|1|18|-15|17|5|6|7|0|9|8|30|14|41|8|35|14|42|8|31|17|5|6|7|0|2|8|32|14|0|8|34|14|0|-65|-71|8|43|14|44|8|45|46|0|8|47|46|0|8|48|14|49|8|50|14|51|8|39|17|5|6|7|0|5|8|52|14|53|8|54|46|55|3759152200|1288490188800|0|8|56|14|53|8|57|46|55|3759152200|1288490188800|0|8|58|14|38|-65|-89|-61|-63|-37|-59|-33|-35|-1|8|28|8|59|14|60|8|61|17|5|6|7|0|1|18|-15|17|5|6|7|0|7|8|62|14|63|8|30|14|64|8|39|17|5|6|7|0|5|8|52|14|65|8|54|46|55|3479039543|1309965025280|823000000|8|56|14|65|8|57|46|55|3479039543|1309965025280|823000000|8|58|14|38|-114|-120|8|66|17|5|6|7|0|1|18|-15|14|63|-114|-136|8|48|14|67|8|50|14|68|8|21|17|5|6|7|0|1|8|50|17|5|6|7|0|1|8|22|14|69|-148|-150|-114|-146|-110|-112|-1|-108|8|70|17|5|6|7|0|0|-1|-156|8|71|17|5|6|7|0|1|8|34|14|72|-1|-160|8|73|17|5|6|7|0|3|8|74|14|75|8|76|77|19|2|8|21|17|5|6|7|0|1|8|74|17|5|6|7|0|1|8|22|14|78|-177|-179|-168|-175|-1|-166|8|79|17|5|6|7|0|1|8|80|17|5|6|7|0|0|-187|-189|-1|-185|8|81|17|5|6|7|0|0|-1|-193|8|82|17|5|6|7|0|1|18|-15|17|5|6|7|0|6|8|83|17|5|6|7|0|1|18|-15|17|5|6|7|0|9|8|84|14|85|8|86|17|5|6|7|0|3|8|87|14|88|8|89|14|90|8|21|17|5|6|7|0|1|8|89|17|5|6|7|0|1|8|22|14|91|-225|-227|-217|-223|-211|-215|8|92|77|19|100|8|73|17|5|6|7|0|3|8|74|14|93|8|76|77|19|13|8|21|17|5|6|7|0|1|8|74|17|5|6|7|0|1|8|22|14|94|-247|-249|-238|-245|-211|-236|8|30|14|95|8|39|17|5|6|7|0|5|8|52|14|65|8|54|46|55|3479549402|1309965025280|682000000|8|56|14|65|8|57|46|55|3479549402|1309965025280|682000000|8|58|14|38|-211|-257|8|48|14|67|8|96|17|5|6|7|0|0|-211|-275|8|21|17|5|6|7|0|1|8|84|17|5|6|7|0|1|8|22|14|97|-281|-283|-211|-279|-207|-209|-203|-205|8|30|14|98|8|39|17|5|6|7|0|5|8|52|14|65|8|54|46|55|3479549392|1309965025280|672000000|8|56|14|65|8|57|46|55|3479549392|1309965025280|672000000|8|58|14|38|-203|-291|8|48|14|67|8|99|17|5|6|7|0|0|-203|-309|8|50|14|100|-199|-201|-1|-197|8|101|17|5|6|7|0|2|18|-15|14|102|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|103|-323|-325|-317|-321|-1|-315|8|30|14|104|8|105|17|5|6|7|0|1|18|-15|17|5|6|7|0|2|8|106|14|107|8|21|17|5|6|7|0|1|8|106|17|5|6|7|0|1|8|22|14|108|-345|-347|-339|-343|-335|-337|-1|-333|8|109|17|5|6|7|0|0|-1|-353|8|110|14|111|8|39|17|5|6|7|0|5|8|52|14|65|8|54|46|55|3479545033|1309965025280|313000000|8|56|14|65|8|57|46|55|3479549354|1309965025280|634000000|8|58|14|112|-1|-359|8|113|10|-5|8|114|17|5|6|7|0|0|-1|-377|8|115|10|-5|8|48|14|67|8|116|14|117|8|99|17|5|6|7|0|2|18|-15|14|118|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|119|-395|-397|-389|-393|-1|-387|8|120|14|121|8|50|14|122|8|96|17|5|6|7|0|2|18|-15|14|123|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|124|-415|-417|-409|-413|-1|-407|8|125|17|5|6|7|0|2|18|-15|14|126|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|127|-431|-433|-425|-429|-1|-423|8|128|17|5|6|7|0|0|-1|-439|8|129|17|5|6|7|0|3|8|130|46|55|3479545033|1309965025280|313000000|8|131|132|133|1|0|8|134|14|135|-1|-443|8|21|17|5|6|7|0|2|8|116|17|5|6|7|0|1|8|22|14|117|-457|-459|8|13|17|5|6|7|0|1|8|22|14|136|-457|-465|-1|-455|8|137|17|5|6|7|0|12|8|138|14|139|8|30|14|140|8|39|17|5|6|7|0|5|8|52|14|65|8|54|46|55|3479545302|1309965025280|582000000|8|56|14|65|8|57|46|55|3479550248|1309965025280|528000000|8|58|14|141|-473|-479|8|35|14|60|8|142|17|5|6|7|0|1|18|-15|14|104|-473|-497|8|143|14|144|8|145|17|5|6|7|0|0|-473|-505|8|146|17|5|6|7|0|0|-473|-509|8|147|14|148|8|48|14|149|8|50|14|150|8|151|14|152|-1|-471|8|153|17|5|6|7|0|1|8|154|17|5|6|7|0|0|-523|-525|-1|-521|0|0|\"\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/CreditCourseProposalRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"#{contents1}#{contents2}\"\n },\n {\n 'subst' => 'true',\n :dyn_variables => [\n {\"name\" => opts[:lo_category_var_name], \"regexp\" => opts[:lo_category_var_regexp]}\n ]\n }\n )\n\n #@request.add(\"DEBUG/lo_category_var_name/%%_#{opts[:lo_category_var_name]}%%\", {}, {'subst' => 'true'})\n\n @request.add_thinktime(2)\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/LoCategoryRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|4|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|EDE1DECBD54F5894284944CD8AC3661C|org.kuali.student.lum.common.client.lo.rpc.LoCategoryRpcService|getLoCategoryTypes|1|2|3|4|0|\"\n }\n )\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/statementRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|2543249A07E7952228E0E500F14F1B17|org.kuali.student.lum.program.client.rpc.StatementRpcService|getStatementTypesForStatementTypeForCourse|java.lang.String/2004016611|kuali.statement.type.course|1|2|3|4|1|5|6|\"\n }\n )\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/CourseRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|8|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|1ED48DA6F48F82765FE7B58378EA94E0|org.kuali.student.lum.lu.ui.course.client.service.CourseRpcService|getCourseStatements|java.lang.String/2004016611|%%_#{opts[:clu_ref_dyn_var_name]}%%|KUALI.RULE|en|1|2|3|4|3|5|5|5|6|7|8|\"\n },\n {\n 'subst' => 'true'\n }\n )\n \n \n \n #\n # Course Requisites\n # Save without editing anything\n\n @request.add_thinktime(5)\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/CourseRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|10|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|1ED48DA6F48F82765FE7B58378EA94E0|org.kuali.student.lum.lu.ui.course.client.service.CourseRpcService|storeCourseStatements|java.lang.String/2004016611|java.util.Map|%%_#{opts[:clu_ref_dyn_var_name]}%%|draft|java.util.HashMap/962170901|java.util.LinkedHashMap/1551059846|1|2|3|4|4|5|5|6|6|7|8|9|0|10|0|0|\"\n },\n {\n 'subst' => 'true'\n }\n )\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/statementRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|2543249A07E7952228E0E500F14F1B17|org.kuali.student.lum.program.client.rpc.StatementRpcService|getStatementTypesForStatementTypeForCourse|java.lang.String/2004016611|kuali.statement.type.course|1|2|3|4|1|5|6|\"\n }\n )\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/CourseRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|8|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|1ED48DA6F48F82765FE7B58378EA94E0|org.kuali.student.lum.lu.ui.course.client.service.CourseRpcService|getCourseStatements|java.lang.String/2004016611|%%_#{opts[:clu_ref_dyn_var_name]}%%|KUALI.RULE|en|1|2|3|4|3|5|5|5|6|7|8|\"\n },\n {\n 'subst' => 'true'\n }\n )\n \n \n \n #\n # Active Dates\n # Start = Fall Sem 2008\n #\n\n @request.add_thinktime(10)\n\n # Save & Continue\n contents1 = \"5|0|160|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|D60D3C6E0D395C18A0F44A2D9D2A7348|org.kuali.student.lum.lu.ui.course.client.service.CreditCourseProposalRpcService|saveData|org.kuali.student.common.assembly.data.Data/3184510345|org.kuali.student.common.assembly.data.Data|java.util.LinkedHashMap/1551059846|org.kuali.student.common.assembly.data.Data$StringKey/758802082|passFail|org.kuali.student.common.assembly.data.Data$BooleanValue/4261226833|java.lang.Boolean/476441737|audit|finalExamStatus|org.kuali.student.common.assembly.data.Data$StringValue/3151113388|STD|campusLocations|org.kuali.student.common.assembly.data.Data$DataValue/1692468409|org.kuali.student.common.assembly.data.Data$IntegerKey/134469241|java.lang.Integer/3438268394|AL|_runtimeData|id-translation|All|code|#{opts[:subject_area]}#{opts[:course_suffix]}|courseNumberSuffix|#{opts[:course_suffix]}|courseSpecificLOs|loCategoryInfoList|id|%%_#{opts[:lo_category_id_var_name]}%%|loRepository|kuali.loRepository.key.singleUse|metaInfo|createId|admin|createTime|org.kuali.student.common.assembly.data.Data$DateValue/2929953165|java.sql.Timestamp/1769758459|updateId|updateTime|versionInd|0|name|#{opts[:lo_category]}|state|active|type|loCategoryType.subject|loDisplayInfoList|loInfo|sequence|desc|formatted|#{opts[:lo_cat_text]}|plain|%%_#{opts[:lo_category_var_name]}%%|loRepositoryKey|#{opts[:propose_person]}|SINGLE USE LO|draft|kuali.lo.type.singleUse|courseTitle|#{opts[:course_title]}|creditOptions|fixedCreditValue|10.0|kuali.creditType.credit.degree.10.0|resultValues|kuali.resultComponentType.credit.degree.fixed|Credits, Fixed|crossListings|descr|#{opts[:course_description]}|duration|atpDurationTypeKey|kuali.atp.duration.Semester|timeQuantity|org.kuali.student.common.assembly.data.Data$IntegerValue/3605481012|Semester|expenditure|affiliatedOrgs|fees|formats|activities|activityType|kuali.lu.type.activity.Lab|contactHours|unitQuantity|5|unitType|kuali.atp.duration.week|per week|defaultEnrollmentEstimate|kuali.atp.duration.Week|Week|%%_#{opts[:atp_duration_week_var_name]}%%|1|unitsContentOwner|Lab|%%_#{opts[:lab_var_name]}%%|termsOffered|kuali.lu.type.CreditCourseFormatShell|gradingOptions|kuali.resultComponent.grade.letter|Letter|%%_#{opts[:clu_ref_dyn_var_name]}%%|instructors|personId|#{opts[:instructor]}|#{opts[:instructor]}, #{opts[:instructor]}(#{opts[:instructor]})|joints|level|100|3|pilotCourse|revenues|specialTopicsCourse|subjectArea|#{opts[:subject_area]}|kuali.atp.season.Any|Any|transcriptTitle|#{opts[:course_short_title]}|kuali.lu.type.CreditCourse|#{opts[:oversight_dept_number]}|#{oversight_department}|unitsDeployment|%%_#{opts[:admin_dep_var_name]}%%|#{admin_org}|variations|versionInfo|currentVersionStart|sequenceNumber|org.kuali.student.common.assembly.data.Data$LongValue/3784756947|java.lang.Long/4227064769|versionIndId|%%_#{opts[:version_ind_id_name]}%%|Standard final Exam|dirty|startTerm|endTerm|proposal|workflowNode|PreRoute|%%_#{opts[:proposal_dyn_var_name]}%%|proposalReference|proposalReferenceType|kuali.proposal.referenceType.clu|proposerOrg|proposerPerson|rationale|#{opts[:course_rationale]}|Saved|kuali.proposal.type.course.create|workflowId|%%_#{opts[:proposal_doc_id_var_name]}%%|collaboratorInfo|collaborators|kuali.atp.FA2008-2009\"\n contents2 = \"|1|2|3|4|1|5|5|6|7|0|38|8|9|10|11|0|8|12|10|-5|8|13|14|15|8|16|17|5|6|7|0|2|18|19|0|14|20|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|23|-19|-21|-12|-17|-1|-10|8|24|14|25|8|26|14|27|8|28|17|5|6|7|0|1|18|-15|17|5|6|7|0|3|8|29|17|5|6|7|0|1|18|-15|17|5|6|7|0|6|8|30|14|31|8|32|14|33|8|34|17|5|6|7|0|5|8|35|14|36|8|37|38|39|3759152200|1288490188800|0|8|40|14|36|8|41|38|39|3759152200|1288490188800|0|8|42|14|43|-45|-51|8|44|14|45|8|46|14|47|8|48|14|49|-41|-43|-37|-39|8|50|17|5|6|7|0|0|-37|-73|8|51|17|5|6|7|0|8|8|52|14|43|8|53|17|5|6|7|0|2|8|54|14|55|8|56|14|55|-79|-83|8|30|14|57|8|58|14|33|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3487349987|1309965025280|267000000|8|40|14|59|8|41|38|39|3487349987|1309965025280|267000000|8|42|14|43|-79|-95|8|44|14|60|8|46|14|61|8|48|14|62|-37|-77|-33|-35|-1|-31|8|63|14|64|8|65|17|5|6|7|0|1|18|-15|17|5|6|7|0|7|8|66|14|67|8|30|14|68|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3479039543|1309965025280|823000000|8|40|14|59|8|41|38|39|3479039543|1309965025280|823000000|8|42|14|43|-125|-131|8|69|17|5|6|7|0|1|18|-15|14|67|-125|-147|8|46|14|61|8|48|14|70|8|21|17|5|6|7|0|1|8|48|17|5|6|7|0|1|8|22|14|71|-159|-161|-125|-157|-121|-123|-1|-119|8|72|17|5|6|7|0|0|-1|-167|8|73|17|5|6|7|0|1|8|56|14|74|-1|-171|8|75|17|5|6|7|0|3|8|76|14|77|8|78|79|19|2|8|21|17|5|6|7|0|1|8|76|17|5|6|7|0|1|8|22|14|80|-188|-190|-179|-186|-1|-177|8|81|17|5|6|7|0|1|8|82|17|5|6|7|0|0|-198|-200|-1|-196|8|83|17|5|6|7|0|0|-1|-204|8|84|17|5|6|7|0|1|18|-15|17|5|6|7|0|6|8|85|17|5|6|7|0|1|18|-15|17|5|6|7|0|9|8|86|14|87|8|88|17|5|6|7|0|3|8|89|14|90|8|91|14|92|8|21|17|5|6|7|0|1|8|91|17|5|6|7|0|1|8|22|14|93|-236|-238|-228|-234|-222|-226|8|94|79|19|100|8|75|17|5|6|7|0|3|8|76|14|95|8|78|79|19|13|8|21|17|5|6|7|0|1|8|76|17|5|6|7|0|1|8|22|14|96|-258|-260|-249|-256|-222|-247|8|30|14|97|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3487345884|1309965025280|164000000|8|40|14|59|8|41|38|39|3487349962|1309965025280|242000000|8|42|14|98|-222|-268|8|46|14|61|8|99|17|5|6|7|0|0|-222|-286|8|21|17|5|6|7|0|1|8|86|17|5|6|7|0|1|8|22|14|100|-292|-294|-222|-290|-218|-220|-214|-216|8|30|14|101|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3487345875|1309965025280|155000000|8|40|14|59|8|41|38|39|3487349936|1309965025280|216000000|8|42|14|98|-214|-302|8|46|14|61|8|102|17|5|6|7|0|0|-214|-320|8|48|14|103|-210|-212|-1|-208|8|104|17|5|6|7|0|2|18|-15|14|105|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|106|-334|-336|-328|-332|-1|-326|8|30|14|107|8|108|17|5|6|7|0|1|18|-15|17|5|6|7|0|2|8|109|14|110|8|21|17|5|6|7|0|1|8|109|17|5|6|7|0|1|8|22|14|111|-356|-358|-350|-354|-346|-348|-1|-344|8|112|17|5|6|7|0|0|-1|-364|8|113|14|114|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3487341491|1309965025280|771000000|8|40|14|59|8|41|38|39|3487349901|1309965025280|181000000|8|42|14|115|-1|-370|8|116|10|-5|8|117|17|5|6|7|0|0|-1|-388|8|118|10|-5|8|46|14|61|8|119|14|120|8|102|17|5|6|7|0|2|18|-15|14|121|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|122|-406|-408|-400|-404|-1|-398|8|123|14|124|8|48|14|125|8|99|17|5|6|7|0|2|18|-15|14|126|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|127|-426|-428|-420|-424|-1|-418|8|128|17|5|6|7|0|2|18|-15|14|129|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|130|-442|-444|-436|-440|-1|-434|8|131|17|5|6|7|0|0|-1|-450|8|132|17|5|6|7|0|3|8|133|38|39|3487341491|1309965025280|771000000|8|134|135|136|1|0|8|137|14|138|-1|-454|8|21|17|5|6|7|0|3|8|119|17|5|6|7|0|1|8|22|14|120|-468|-470|8|13|17|5|6|7|0|1|8|22|14|139|-468|-476|8|140|17|5|6|7|0|2|8|141|10|11|1|8|142|10|-488|-468|-482|-1|-466|8|143|17|5|6|7|0|12|8|144|14|145|8|30|14|146|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3487341710|1309965025280|990000000|8|40|14|59|8|41|38|39|3487350824|1309965025280|104000000|8|42|14|90|-493|-499|8|44|14|64|8|147|17|5|6|7|0|1|18|-15|14|107|-493|-517|8|148|14|149|8|150|17|5|6|7|0|0|-493|-525|8|151|17|5|6|7|0|0|-493|-529|8|152|14|153|8|46|14|154|8|48|14|155|8|156|14|157|-1|-491|8|158|17|5|6|7|0|1|8|159|17|5|6|7|0|0|-543|-545|-1|-541|-486|14|160|-489|14|0|0|0|\"\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/CreditCourseProposalRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"#{contents1}#{contents2}\"\n },\n {\n 'subst' => 'true'\n }\n )\n \n \n \n #\n # Financials\n # $100 lab fee\n # Admin org 100% rev and exp\n #\n\n @request.add_thinktime(5)\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|11|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|enumeration.queryParam.enumerationType|kuali.lu.fee.feeType|enumeration.management.search|enumeration.resultColumn.sortKey|1|2|3|4|1|5|5|0|0|6|1|7|8|0|9|10|11|0|0|\"\n }\n )\n\n # Revenue - Admin Org\n for i in 1..admin_org.length\n itr = i-1\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|16|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|search|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.lang.Boolean/476441737|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|org.queryParam.orgOptionalLongName|#{admin_org[0..itr]}|org.queryParam.orgOptionalType|java.lang.String/2004016611|kuali.org.Department|kuali.org.College|org.search.generic||1|2|3|4|1|5|5|0|6|0|7|2|8|9|0|10|8|11|7|2|12|13|12|14|0|15|16|0|0|\"\n } \n ) \n end\n\n @request.add_thinktime(5)\n\n # Expense - Admin Org\n for i in 1..admin_org.length\n if(i == admin_org.length)\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|16|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|search|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.lang.Boolean/476441737|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|org.queryParam.orgOptionalLongName|#{admin_org[0..itr]}|org.queryParam.orgOptionalType|java.lang.String/2004016611|kuali.org.Department|kuali.org.College|org.search.generic||1|2|3|4|1|5|5|0|6|0|7|2|8|9|0|10|8|11|7|2|12|13|12|14|0|15|16|0|0|\"\n },\n {\n :dyn_variables => [\n {\"name\" => opts[:admin_dep_var_name], \"regexp\" => opts[:admin_dep_var_regexp]}\n ]\n } \n )\n else\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|16|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|search|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.lang.Boolean/476441737|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|org.queryParam.orgOptionalLongName|#{admin_org[0..itr]}|org.queryParam.orgOptionalType|java.lang.String/2004016611|kuali.org.Department|kuali.org.College|org.search.generic||1|2|3|4|1|5|5|0|6|0|7|2|8|9|0|10|8|11|7|2|12|13|12|14|0|15|16|0|0|\"\n } \n )\n end \n end\n\n #@request.add(\"DEBUG/admin_dep_var_name/%%_#{opts[:admin_dep_var_name]}%%\", {}, {'subst' => 'true'})\n\n @request.add_thinktime(20)\n\n # Save & Continue\n contents1 = \"5|0|172|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|D60D3C6E0D395C18A0F44A2D9D2A7348|org.kuali.student.lum.lu.ui.course.client.service.CreditCourseProposalRpcService|saveData|org.kuali.student.common.assembly.data.Data/3184510345|org.kuali.student.common.assembly.data.Data|java.util.LinkedHashMap/1551059846|org.kuali.student.common.assembly.data.Data$StringKey/758802082|passFail|org.kuali.student.common.assembly.data.Data$BooleanValue/4261226833|java.lang.Boolean/476441737|audit|finalExamStatus|org.kuali.student.common.assembly.data.Data$StringValue/3151113388|STD|campusLocations|org.kuali.student.common.assembly.data.Data$DataValue/1692468409|org.kuali.student.common.assembly.data.Data$IntegerKey/134469241|java.lang.Integer/3438268394|AL|_runtimeData|id-translation|All|code|#{opts[:subject_area]}#{opts[:course_suffix]}|courseNumberSuffix|#{opts[:course_suffix]}|courseSpecificLOs|loCategoryInfoList|id|%%_#{opts[:lo_category_id_var_name]}%%|loRepository|kuali.loRepository.key.singleUse|metaInfo|createId|admin|createTime|org.kuali.student.common.assembly.data.Data$DateValue/2929953165|java.sql.Timestamp/1769758459|updateId|updateTime|versionInd|0|name|#{opts[:lo_category]}|state|active|type|loCategoryType.subject|loDisplayInfoList|loInfo|sequence|desc|formatted|#{opts[:lo_cat_text]}|plain|%%_#{opts[:lo_category_var_name]}%%|loRepositoryKey|#{opts[:propose_person]}|1|SINGLE USE LO|draft|kuali.lo.type.singleUse|courseTitle|#{opts[:course_title]}|creditOptions|fixedCreditValue|10.0|kuali.creditType.credit.degree.10.0|resultValues|kuali.resultComponentType.credit.degree.fixed|Credits, Fixed|crossListings|descr|#{opts[:course_description]}|duration|atpDurationTypeKey|kuali.atp.duration.Semester|timeQuantity|org.kuali.student.common.assembly.data.Data$IntegerValue/3605481012|Semester|effectiveDate|expenditure|affiliatedOrgs|dirty|orgId|percentage|created|%%_#{opts[:admin_dep_var_name]}%%|org.kuali.student.common.assembly.data.Data$LongValue/3784756947|java.lang.Long/4227064769|fees|feeType|rateType|kuali.enum.type.feeTypes.labFee|fixedRateFee|feeAmounts|currencyQuantity|formats|activities|activityType|kuali.lu.type.activity.Lab|contactHours|unitQuantity|5|unitType|kuali.atp.duration.week|per week|defaultEnrollmentEstimate|kuali.atp.duration.Week|Week|%%_#{opts[:atp_duration_week_var_name]}%%|2|unitsContentOwner|Lab|%%_#{opts[:lab_var_name]}%%|termsOffered|kuali.lu.type.CreditCourseFormatShell|gradingOptions|kuali.resultComponent.grade.letter|Letter|%%_#{opts[:clu_ref_dyn_var_name]}%%|instructors|personId|#{opts[:instructor]}|#{opts[:instructor]}, #{opts[:instructor]}(#{opts[:instructor]})|joints|level|100|4|pilotCourse|revenues|specialTopicsCourse|startTerm|kuali.atp.FA2008-2009|subjectArea|#{opts[:subject_area]}|kuali.atp.season.Any|Any|transcriptTitle|#{opts[:course_short_title]}|kuali.lu.type.CreditCourse|#{opts[:oversight_dept_number]}|#{oversight_department}|unitsDeployment|#{admin_org}|variations|versionInfo|currentVersionStart|sequenceNumber|versionIndId|%%_#{opts[:version_ind_id_name]}%%|Fall Semester of 2008|Standard final Exam|proposal|workflowNode|PreRoute|%%_#{opts[:proposal_dyn_var_name]}%%|6|proposalReference|proposalReferenceType|kuali.proposal.referenceType.clu|proposerOrg|proposerPerson|rationale|#{opts[:course_rationale]}|Saved|kuali.proposal.type.course.create|workflowId|%%_#{opts[:proposal_doc_id_var_name]}%%|collaboratorInfo|collaborators\"\n contents2 = \"|1|2|3|4|1|5|5|6|7|0|38|8|9|10|11|0|8|12|10|-5|8|13|14|15|8|16|17|5|6|7|0|2|18|19|0|14|20|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|23|-19|-21|-12|-17|-1|-10|8|24|14|25|8|26|14|27|8|28|17|5|6|7|0|1|18|-15|17|5|6|7|0|3|8|29|17|5|6|7|0|1|18|-15|17|5|6|7|0|6|8|30|14|31|8|32|14|33|8|34|17|5|6|7|0|5|8|35|14|36|8|37|38|39|3759152200|1288490188800|0|8|40|14|36|8|41|38|39|3759152200|1288490188800|0|8|42|14|43|-45|-51|8|44|14|45|8|46|14|47|8|48|14|49|-41|-43|-37|-39|8|50|17|5|6|7|0|0|-37|-73|8|51|17|5|6|7|0|8|8|52|14|43|8|53|17|5|6|7|0|2|8|54|14|55|8|56|14|55|-79|-83|8|30|14|57|8|58|14|33|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3488193507|1309965025280|787000000|8|40|14|59|8|41|38|39|3488196257|1309965025280|537000000|8|42|14|60|-79|-95|8|44|14|61|8|46|14|62|8|48|14|63|-37|-77|-33|-35|-1|-31|8|64|14|65|8|66|17|5|6|7|0|1|18|-15|17|5|6|7|0|7|8|67|14|68|8|30|14|69|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3479039543|1309965025280|823000000|8|40|14|59|8|41|38|39|3479039543|1309965025280|823000000|8|42|14|43|-125|-131|8|70|17|5|6|7|0|1|18|-15|14|68|-125|-147|8|46|14|62|8|48|14|71|8|21|17|5|6|7|0|1|8|48|17|5|6|7|0|1|8|22|14|72|-159|-161|-125|-157|-121|-123|-1|-119|8|73|17|5|6|7|0|0|-1|-167|8|74|17|5|6|7|0|1|8|56|14|75|-1|-171|8|76|17|5|6|7|0|3|8|77|14|78|8|79|80|19|2|8|21|17|5|6|7|0|1|8|77|17|5|6|7|0|1|8|22|14|81|-188|-190|-179|-186|-1|-177|8|82|38|39|470887936|1219770712064|0|8|83|17|5|6|7|0|1|8|84|17|5|6|7|0|1|18|-15|17|5|6|7|0|3|8|21|17|5|6|7|0|3|8|85|17|5|6|7|0|2|8|86|10|11|1|8|87|10|-221|-213|-215|8|88|10|-221|-211|17|5|6|7|0|1|-215|17|5|6|7|0|1|-224|10|-221|-227|-215|-213|-211|-209|-211|-219|14|89|-222|90|91|100|0|-205|-207|-201|-203|-1|-199|8|92|17|5|6|7|0|1|18|-15|17|5|6|7|0|4|-211|17|5|6|7|0|3|-215|17|5|6|7|0|2|8|93|10|-221|8|94|10|-221|-245|-215|8|88|10|-221|-211|17|5|6|7|0|1|-215|17|5|6|7|0|1|-254|10|-221|-257|-215|-245|-211|-242|-211|-250|14|95|-252|14|96|8|97|17|5|6|7|0|1|18|-15|17|5|6|7|0|2|8|98|80|19|100|-211|17|5|6|7|0|1|-215|17|5|6|7|0|1|-273|10|-221|-277|-215|-271|-211|-267|-269|-242|-265|-238|-240|-1|-236|8|99|17|5|6|7|0|1|18|-15|17|5|6|7|0|6|8|100|17|5|6|7|0|1|18|-15|17|5|6|7|0|9|8|101|14|102|8|103|17|5|6|7|0|3|8|104|14|105|8|106|14|107|8|21|17|5|6|7|0|1|8|106|17|5|6|7|0|1|8|22|14|108|-311|-313|-303|-309|-297|-301|8|109|80|-275|8|76|17|5|6|7|0|3|8|77|14|110|8|79|80|19|13|8|21|17|5|6|7|0|1|8|77|17|5|6|7|0\"\n contents3 = \"|1|8|22|14|111|-332|-334|-323|-330|-297|-321|8|30|14|112|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3488189822|1309965025280|102000000|8|40|14|59|8|41|38|39|3488196198|1309965025280|478000000|8|42|14|113|-297|-342|8|46|14|62|8|114|17|5|6|7|0|0|-297|-360|8|21|17|5|6|7|0|1|8|101|17|5|6|7|0|1|8|22|14|115|-366|-368|-297|-364|-293|-295|-289|-291|8|30|14|116|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3488189788|1309965025280|68000000|8|40|14|59|8|41|38|39|3488196188|1309965025280|468000000|8|42|14|113|-289|-376|8|46|14|62|8|117|17|5|6|7|0|0|-289|-394|8|48|14|118|-285|-287|-1|-283|8|119|17|5|6|7|0|2|18|-15|14|120|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|121|-408|-410|-402|-406|-1|-400|8|30|14|122|8|123|17|5|6|7|0|1|18|-15|17|5|6|7|0|2|8|124|14|125|8|21|17|5|6|7|0|1|8|124|17|5|6|7|0|1|8|22|14|126|-430|-432|-424|-428|-420|-422|-1|-418|8|127|17|5|6|7|0|0|-1|-438|8|128|14|129|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3488185173|1309965025280|453000000|8|40|14|59|8|41|38|39|3488196156|1309965025280|436000000|8|42|14|130|-1|-444|8|131|10|-5|8|132|17|5|6|7|0|1|18|-15|17|5|6|7|0|2|8|84|17|5|6|7|0|1|18|-15|17|5|6|7|0|3|-211|17|5|6|7|0|1|-215|17|5|6|7|0|2|8|86|10|-221|8|87|10|-221|-479|-215|-476|-211|-484|14|89|-486|90|-235|-472|-474|-468|-470|8|21|17|5|6|7|0|2|8|88|10|-221|-211|17|5|6|7|0|1|-215|17|5|6|7|0|1|-494|10|-221|-497|-215|-492|-211|-468|-490|-464|-466|-1|-462|8|133|10|-5|8|134|14|135|8|46|14|62|8|136|14|137|8|117|17|5|6|7|0|2|18|-15|14|138|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|139|-519|-521|-513|-517|-1|-511|8|140|14|141|8|48|14|142|8|114|17|5|6|7|0|2|18|-15|14|143|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|144|-539|-541|-533|-537|-1|-531|8|145|17|5|6|7|0|2|18|-15|14|89|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|146|-555|-557|-549|-553|-1|-547|8|147|17|5|6|7|0|0|-1|-563|8|148|17|5|6|7|0|3|8|149|38|39|3488185173|1309965025280|453000000|8|150|90|91|1|0|8|151|14|152|-1|-567|8|21|17|5|6|7|0|3|8|134|17|5|6|7|0|1|8|22|14|153|-581|-583|8|136|17|5|6|7|0|1|8|22|14|137|-581|-589|8|13|17|5|6|7|0|1|8|22|14|154|-581|-595|-1|-579|8|155|17|5|6|7|0|12|8|156|14|157|8|30|14|158|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3488185428|1309965025280|708000000|8|40|14|59|8|41|38|39|3488197141|1309965025280|421000000|8|42|14|159|-603|-609|8|44|14|65|8|160|17|5|6|7|0|1|18|-15|14|122|-603|-627|8|161|14|162|8|163|17|5|6|7|0|0|-603|-635|8|164|17|5|6|7|0|0|-603|-639|8|165|14|166|8|46|14|167|8|48|14|168|8|169|14|170|-1|-601|8|171|17|5|6|7|0|1|8|172|17|5|6|7|0|0|-653|-655|-1|-651|0|0|\"\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/CreditCourseProposalRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"#{contents1}#{contents2}#{contents3}\"\n },\n {\n 'subst' => 'true',\n :dyn_variables => [\n {\"name\" => opts[:affliated_orgs_id_name], \"regexp\" => opts[:affliated_orgs_id_regexp]},\n {\"name\" => opts[:lab_fee_id_name], \"regexp\" => opts[:lab_fee_id_regexp]},\n {\"name\" => opts[:revenues_id_name], \"regexp\" => opts[:revenues_id_regexp]},\n {\"name\" => opts[:revenue_id_name], \"regexp\" => opts[:revenue_id_regexp]}\n ]\n }\n )\n\n #@request.add(\"DEBUG/affliated_orgs_id_name/%%_#{opts[:affliated_orgs_id_name]}%%\", {}, {'subst' => 'true'})\n #@request.add(\"DEBUG/lab_fee_id_name/%%_#{opts[:lab_fee_id_name]}%%\", {}, {'subst' => 'true'})\n #@request.add(\"DEBUG/revenues_id_name/%%_#{opts[:revenues_id_name]}%%\", {}, {'subst' => 'true'})\n #@request.add(\"DEBUG/revenue_id_name/%%_#{opts[:revenue_id_name]}%%\", {}, {'subst' => 'true'})\n\n @request.add_thinktime(2)\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/WorkflowRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|71417C94A72A0CF76A43A2B36B8E3E1B|org.kuali.student.core.workflow.ui.client.service.WorkflowRpcService|getDocumentStatus|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/WorkflowRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|71417C94A72A0CF76A43A2B36B8E3E1B|org.kuali.student.core.workflow.ui.client.service.WorkflowRpcService|isAuthorizedAddReviewer|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/WorkflowRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|71417C94A72A0CF76A43A2B36B8E3E1B|org.kuali.student.core.workflow.ui.client.service.WorkflowRpcService|isAuthorizedRemoveReviewers|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n \n \n \n \n #\n # Authors and Collaborators\n #\n\n @request.add_thinktime(5)\n\n if(!opts[:collaborator].nil?)\n # Collaborator seach\n for i in 1..opts[:collaborator].length\n itr = i-1\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|14|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|search|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.lang.Boolean/476441737|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|person.queryParam.personGivenName|#{opts[:collaborator][0..itr]}|person.queryParam.excludedUserId|psycho1|person.search.personQuickViewByGivenName|person.resultColumn.DisplayName|1|2|3|4|1|5|5|0|6|0|7|2|8|9|0|10|8|11|0|12|13|14|0|0|\"\n } \n ) \n end\n\n @request.add_thinktime(5)\n\n\n contents1 = \"5|0|190|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|D60D3C6E0D395C18A0F44A2D9D2A7348|org.kuali.student.lum.lu.ui.course.client.service.CreditCourseProposalRpcService|saveData|org.kuali.student.common.assembly.data.Data/3184510345|org.kuali.student.common.assembly.data.Data|java.util.LinkedHashMap/1551059846|org.kuali.student.common.assembly.data.Data$StringKey/758802082|passFail|org.kuali.student.common.assembly.data.Data$BooleanValue/4261226833|java.lang.Boolean/476441737|audit|finalExamStatus|org.kuali.student.common.assembly.data.Data$StringValue/3151113388|STD|campusLocations|org.kuali.student.common.assembly.data.Data$DataValue/1692468409|org.kuali.student.common.assembly.data.Data$IntegerKey/134469241|java.lang.Integer/3438268394|AL|_runtimeData|id-translation|All|code|#{opts[:subject_area]}#{opts[:course_suffix]}|courseNumberSuffix|#{opts[:course_suffix]}|courseSpecificLOs|loCategoryInfoList|id|%%_#{opts[:lo_category_id_var_name]}%%|loRepository|kuali.loRepository.key.singleUse|metaInfo|createId|admin|createTime|org.kuali.student.common.assembly.data.Data$DateValue/2929953165|java.sql.Timestamp/1769758459|updateId|updateTime|versionInd|0|name|#{opts[:lo_category]}|state|active|type|loCategoryType.subject|loDisplayInfoList|loInfo|sequence|desc|formatted|#{opts[:lo_cat_text]}|plain|%%_#{opts[:lo_category_var_name]}%%|loRepositoryKey|#{opts[:propose_person]}|2|SINGLE USE LO|draft|kuali.lo.type.singleUse|courseTitle|#{opts[:course_title]}|creditOptions|fixedCreditValue|10.0|kuali.creditType.credit.degree.10.0|resultValues|kuali.resultComponentType.credit.degree.fixed|Credits, Fixed|crossListings|descr|#{opts[:course_description]}|duration|atpDurationTypeKey|kuali.atp.duration.Semester|timeQuantity|org.kuali.student.common.assembly.data.Data$IntegerValue/3605481012|Semester|effectiveDate|expenditure|affiliatedOrgs|%%_#{opts[:affliated_orgs_id_name]}%%|orgId|%%_#{opts[:admin_dep_var_name]}%%|percentage|org.kuali.student.common.assembly.data.Data$LongValue/3784756947|java.lang.Long/4227064769|#{admin_org}|fees|feeAmounts|currencyQuantity|currencyTypeKey|kuali.currency.type.usdollars.cents|feeType|kuali.enum.type.feeTypes.labFee|%%_#{opts[:lab_fee_id_name]}%%|1|rateType|fixedRateFee|Fixed Rate Fee|Laboratory Fee|formats|activities|activityType|kuali.lu.type.activity.Lab|contactHours|unitQuantity|5|unitType|kuali.atp.duration.week|per week|defaultEnrollmentEstimate|kuali.atp.duration.Week|Week|%%_#{opts[:atp_duration_week_var_name]}%%|3|unitsContentOwner|Lab|%%_#{opts[:lab_var_name]}%%|termsOffered|kuali.lu.type.CreditCourseFormatShell|gradingOptions|kuali.resultComponent.grade.letter|Letter|%%_#{opts[:clu_ref_dyn_var_name]}%%|instructors|personId|#{opts[:instructor]}|#{opts[:instructor]}, #{opts[:instructor]}(#{opts[:instructor]})|joints|level|100|pilotCourse|revenues|%%_#{opts[:revenues_id_name]}%%|REVENUE|%%_#{opts[:revenue_id_name]}%%|specialTopicsCourse|startTerm|kuali.atp.FA2008-2009|subjectArea|#{opts[:subject_area]}|kuali.atp.season.Any|Any|transcriptTitle|#{opts[:course_short_title]}|kuali.lu.type.CreditCourse|#{opts[:oversight_dept_number]}|#{oversight_department}|unitsDeployment|variations|versionInfo|currentVersionStart|sequenceNumber|versionIndId|%%_#{opts[:version_ind_id_name]}%%|Fall Semester of 2008|Standard final Exam|proposal|workflowNode|PreRoute|%%_#{opts[:proposal_dyn_var_name]}%%|7|proposalReference|proposalReferenceType|kuali.proposal.referenceType.clu|proposerOrg|proposerPerson|rationale|#{opts[:course_rationale]}|Saved|kuali.proposal.type.course.create|workflowId|%%_#{opts[:proposal_doc_id_var_name]}%%|collaboratorInfo|collaborators|principalId|#{opts[:collaborator]}|permission|KS-SYS~Edit Document|action|F|firstName|lastName|actionRequestStatus|New|author\"\n contents2 = \"|1|2|3|4|1|5|5|6|7|0|38|8|9|10|11|0|8|12|10|-5|8|13|14|15|8|16|17|5|6|7|0|2|18|19|0|14|20|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|23|-19|-21|-12|-17|-1|-10|8|24|14|25|8|26|14|27|8|28|17|5|6|7|0|1|18|-15|17|5|6|7|0|3|8|29|17|5|6|7|0|1|18|-15|17|5|6|7|0|6|8|30|14|31|8|32|14|33|8|34|17|5|6|7|0|5|8|35|14|36|8|37|38|39|3759152200|1288490188800|0|8|40|14|36|8|41|38|39|3759152200|1288490188800|0|8|42|14|43|-45|-51|8|44|14|45|8|46|14|47|8|48|14|49|-41|-43|-37|-39|8|50|17|5|6|7|0|0|-37|-73|8|51|17|5|6|7|0|8|8|52|14|43|8|53|17|5|6|7|0|2|8|54|14|55|8|56|14|55|-79|-83|8|30|14|57|8|58|14|33|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3489516951|1309965025280|231000000|8|40|14|59|8|41|38|39|3489523415|1309965025280|695000000|8|42|14|60|-79|-95|8|44|14|61|8|46|14|62|8|48|14|63|-37|-77|-33|-35|-1|-31|8|64|14|65|8|66|17|5|6|7|0|1|18|-15|17|5|6|7|0|7|8|67|14|68|8|30|14|69|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3479039543|1309965025280|823000000|8|40|14|59|8|41|38|39|3479039543|1309965025280|823000000|8|42|14|43|-125|-131|8|70|17|5|6|7|0|1|18|-15|14|68|-125|-147|8|46|14|62|8|48|14|71|8|21|17|5|6|7|0|1|8|48|17|5|6|7|0|1|8|22|14|72|-159|-161|-125|-157|-121|-123|-1|-119|8|73|17|5|6|7|0|0|-1|-167|8|74|17|5|6|7|0|1|8|56|14|75|-1|-171|8|76|17|5|6|7|0|3|8|77|14|78|8|79|80|19|2|8|21|17|5|6|7|0|1|8|77|17|5|6|7|0|1|8|22|14|81|-188|-190|-179|-186|-1|-177|8|82|38|39|470887936|1219770712064|0|8|83|17|5|6|7|0|1|8|84|17|5|6|7|0|1|18|-15|17|5|6|7|0|4|8|30|14|85|8|86|14|87|8|88|89|90|100|0|8|21|17|5|6|7|0|1|8|86|17|5|6|7|0|1|8|22|14|91|-220|-222|-209|-218|-205|-207|-201|-203|-1|-199|8|92|17|5|6|7|0|1|18|-15|17|5|6|7|0|6|8|93|17|5|6|7|0|1|18|-15|17|5|6|7|0|2|8|94|80|19|100|8|95|14|96|-238|-240|-234|-236|8|97|14|98|8|30|14|99|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3489523331|1309965025280|611000000|8|40|14|59|8|41|38|39|3489523333|1309965025280|613000000|8|42|14|100|-234|-253|8|101|14|102|8|21|17|5|6|7|0|2|8|101|17|5|6|7|0|1|8|22|14|103|-273|-275|8|97|17|5|6|7|0|1|8|22|14|104|-273|-281|-234|-271|-230|-232|-1|-228|8|105|17|5|6|7|0|1|18|-15|17|5|6|7|0|6|8|106|17|5|6|7|0|1|18|-15|17|5|6|7|0|9|8|107|14|108|8|109|17|5|6|7|0|3|8|110|14|111|8|112|14|113|8|21|17|5|6|7|0|1|8|112|17|5|6|7|0|1|8|22|14|114|-315|-317|-307|-313|-301|-305|8|115|80|-246|8|76|17|5|6|7|0|3|8|77|14|116|8|79|80|19|13|8|21|17|5|6|7|0|1|8|77|17|5|6|7|0|1|8|22|14|117|-336|-338|-327|-334|-301|-325\"\n contents3 = \"|8|30|14|118|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3489513388|1309965025280|668000000|8|40|14|59|8|41|38|39|3489523378|1309965025280|658000000|8|42|14|119|-301|-346|8|46|14|62|8|120|17|5|6|7|0|0|-301|-364|8|21|17|5|6|7|0|1|8|107|17|5|6|7|0|1|8|22|14|121|-370|-372|-301|-368|-297|-299|-293|-295|8|30|14|122|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3489513379|1309965025280|659000000|8|40|14|59|8|41|38|39|3489523370|1309965025280|650000000|8|42|14|119|-293|-380|8|46|14|62|8|123|17|5|6|7|0|0|-293|-398|8|48|14|124|-289|-291|-1|-287|8|125|17|5|6|7|0|2|18|-15|14|126|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|127|-412|-414|-406|-410|-1|-404|8|30|14|128|8|129|17|5|6|7|0|1|18|-15|17|5|6|7|0|2|8|130|14|131|8|21|17|5|6|7|0|1|8|130|17|5|6|7|0|1|8|22|14|132|-434|-436|-428|-432|-424|-426|-1|-422|8|133|17|5|6|7|0|0|-1|-442|8|134|14|135|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3489508904|1309965025280|184000000|8|40|14|59|8|41|38|39|3489523333|1309965025280|613000000|8|42|14|111|-1|-448|8|136|10|-5|8|137|17|5|6|7|0|1|18|-15|17|5|6|7|0|4|8|84|17|5|6|7|0|1|18|-15|17|5|6|7|0|4|8|30|14|138|8|86|14|87|8|88|89|90|100|0|8|21|17|5|6|7|0|1|8|86|17|5|6|7|0|1|8|22|14|91|-491|-493|-480|-489|-476|-478|-472|-474|8|97|14|139|8|30|14|140|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3489523331|1309965025280|611000000|8|40|14|59|8|41|38|39|3489523333|1309965025280|613000000|8|42|14|100|-472|-503|-468|-470|-1|-466|8|141|10|-5|8|142|14|143|8|46|14|62|8|144|14|145|8|123|17|5|6|7|0|2|18|-15|14|146|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|147|-535|-537|-529|-533|-1|-527|8|148|14|149|8|48|14|150|8|120|17|5|6|7|0|2|18|-15|14|151|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|152|-555|-557|-549|-553|-1|-547|8|153|17|5|6|7|0|2|18|-15|14|87|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|91|-571|-573|-565|-569|-1|-563|8|154|17|5|6|7|0|0|-1|-579|8|155|17|5|6|7|0|3|8|156|38|39|3489508904|1309965025280|184000000|8|157|89|90|1|0|8|158|14|159|-1|-583|8|21|17|5|6|7|0|3|8|142|17|5|6|7|0|1|8|22|14|160|-597|-599|8|144|17|5|6|7|0|1|8|22|14|145|-597|-605|8|13|17|5|6|7|0|1|8|22|14|161|-597|-611|-1|-595|8|162|17|5|6|7|0|12|8|163|14|164|8|30|14|165|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3489509135|1309965025280|415000000|8|40|14|59|8|41|38|39|3489524194|1309965025280|474000000|8|42|14|166|-619|-625|8|44|14|65|8|167|17|5|6|7|0|1|18|-15|14|128|-619|-643|8|168|14|169|8|170|17|5|6|7|0|0|-619|-651|8|171|17|5|6|7|0|0|-619|-655|8|172|14|173|8|46|14|174|8|48|14|175|8|176|14|177|-1|-617|8|178|17|5|6|7|0|1|8|179|17|5|6|7|0|1|18|-15|17|5|6|7|0|7|8|180|14|181|8|182|14|183|8|184|14|185|8|186|14|181|8|187|14|181|8|188|14|189|8|190|10|-5|-673|-675|-669|-671|-1|-667|0|0|\"\n else\n contents1 = \"5|0|179|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|D60D3C6E0D395C18A0F44A2D9D2A7348|org.kuali.student.lum.lu.ui.course.client.service.CreditCourseProposalRpcService|saveData|org.kuali.student.common.assembly.data.Data/3184510345|org.kuali.student.common.assembly.data.Data|java.util.LinkedHashMap/1551059846|org.kuali.student.common.assembly.data.Data$StringKey/758802082|passFail|org.kuali.student.common.assembly.data.Data$BooleanValue/4261226833|java.lang.Boolean/476441737|audit|finalExamStatus|org.kuali.student.common.assembly.data.Data$StringValue/3151113388|STD|campusLocations|org.kuali.student.common.assembly.data.Data$DataValue/1692468409|org.kuali.student.common.assembly.data.Data$IntegerKey/134469241|java.lang.Integer/3438268394|AL|_runtimeData|id-translation|All|code|#{opts[:subject_area]}#{opts[:course_suffix]}|courseNumberSuffix|#{opts[:course_suffix]}|courseSpecificLOs|loCategoryInfoList|id|%%_#{opts[:lo_category_id_var_name]}%%|loRepository|kuali.loRepository.key.singleUse|metaInfo|createId|admin|createTime|org.kuali.student.common.assembly.data.Data$DateValue/2929953165|java.sql.Timestamp/1769758459|updateId|updateTime|versionInd|0|name|#{opts[:lo_category]}|state|active|type|loCategoryType.subject|loDisplayInfoList|loInfo|sequence|desc|formatted|#{opts[:lo_cat_text]}|plain|%%_#{opts[:lo_category_var_name]}%%|loRepositoryKey|#{opts[:propose_person]}|2|SINGLE USE LO|draft|kuali.lo.type.singleUse|courseTitle|#{opts[:course_title]}|creditOptions|fixedCreditValue|10.0|kuali.creditType.credit.degree.10.0|resultValues|kuali.resultComponentType.credit.degree.fixed|Credits, Fixed|crossListings|descr|#{opts[:course_description]}|duration|atpDurationTypeKey|kuali.atp.duration.Semester|timeQuantity|org.kuali.student.common.assembly.data.Data$IntegerValue/3605481012|Semester|effectiveDate|expenditure|affiliatedOrgs|%%_#{opts[:affliated_orgs_id_name]}%%|orgId|%%_#{opts[:admin_dep_var_name]}%%|percentage|org.kuali.student.common.assembly.data.Data$LongValue/3784756947|java.lang.Long/4227064769|#{admin_org}|fees|feeAmounts|currencyQuantity|currencyTypeKey|kuali.currency.type.usdollars.cents|feeType|kuali.enum.type.feeTypes.labFee|%%_#{opts[:lab_fee_id_name]}%%|1|rateType|fixedRateFee|Fixed Rate Fee|Laboratory Fee|formats|activities|activityType|kuali.lu.type.activity.Lab|contactHours|unitQuantity|5|unitType|kuali.atp.duration.week|per week|defaultEnrollmentEstimate|kuali.atp.duration.Week|Week|%%_#{opts[:atp_duration_week_var_name]}%%|3|unitsContentOwner|Lab|%%_#{opts[:lab_var_name]}%%|termsOffered|kuali.lu.type.CreditCourseFormatShell|gradingOptions|kuali.resultComponent.grade.letter|Letter|%%_#{opts[:clu_ref_dyn_var_name]}%%|instructors|personId|#{opts[:instructor]}|#{opts[:instructor]}, #{opts[:instructor]}(#{opts[:instructor]})|joints|level|100|pilotCourse|revenues|%%_#{opts[:revenues_id_name]}%%|REVENUE|%%_#{opts[:revenue_id_name]}%%|specialTopicsCourse|startTerm|kuali.atp.FA2008-2009|subjectArea|#{opts[:subject_area]}|kuali.atp.season.Any|Any|transcriptTitle|#{opts[:course_short_title]}|kuali.lu.type.CreditCourse|#{opts[:oversight_dept_number]}|#{oversight_department}|unitsDeployment|variations|versionInfo|currentVersionStart|sequenceNumber|versionIndId|%%_#{opts[:version_ind_id_name]}%%|Fall Semester of 2008|Standard final Exam|proposal|workflowNode|PreRoute|%%_#{opts[:proposal_dyn_var_name]}%%|7|proposalReference|proposalReferenceType|kuali.proposal.referenceType.clu|proposerOrg|proposerPerson|rationale|#{opts[:course_rationale]}|Saved|kuali.proposal.type.course.create|workflowId|%%_#{opts[:proposal_doc_id_var_name]}%%|collaboratorInfo|collaborators\"\n contents2 = \"|1|2|3|4|1|5|5|6|7|0|38|8|9|10|11|0|8|12|10|-5|8|13|14|15|8|16|17|5|6|7|0|2|18|19|0|14|20|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|23|-19|-21|-12|-17|-1|-10|8|24|14|25|8|26|14|27|8|28|17|5|6|7|0|1|18|-15|17|5|6|7|0|3|8|29|17|5|6|7|0|1|18|-15|17|5|6|7|0|6|8|30|14|31|8|32|14|33|8|34|17|5|6|7|0|5|8|35|14|36|8|37|38|39|3759152200|1288490188800|0|8|40|14|36|8|41|38|39|3759152200|1288490188800|0|8|42|14|43|-45|-51|8|44|14|45|8|46|14|47|8|48|14|49|-41|-43|-37|-39|8|50|17|5|6|7|0|0|-37|-73|8|51|17|5|6|7|0|8|8|52|14|43|8|53|17|5|6|7|0|2|8|54|14|55|8|56|14|55|-79|-83|8|30|14|57|8|58|14|33|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3491266071|1309965025280|351000000|8|40|14|59|8|41|38|39|3491272352|1309965025280|632000000|8|42|14|60|-79|-95|8|44|14|61|8|46|14|62|8|48|14|63|-37|-77|-33|-35|-1|-31|8|64|14|65|8|66|17|5|6|7|0|1|18|-15|17|5|6|7|0|7|8|67|14|68|8|30|14|69|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3479039543|1309965025280|823000000|8|40|14|59|8|41|38|39|3479039543|1309965025280|823000000|8|42|14|43|-125|-131|8|70|17|5|6|7|0|1|18|-15|14|68|-125|-147|8|46|14|62|8|48|14|71|8|21|17|5|6|7|0|1|8|48|17|5|6|7|0|1|8|22|14|72|-159|-161|-125|-157|-121|-123|-1|-119|8|73|17|5|6|7|0|0|-1|-167|8|74|17|5|6|7|0|1|8|56|14|75|-1|-171|8|76|17|5|6|7|0|3|8|77|14|78|8|79|80|19|2|8|21|17|5|6|7|0|1|8|77|17|5|6|7|0|1|8|22|14|81|-188|-190|-179|-186|-1|-177|8|82|38|39|470887936|1219770712064|0|8|83|17|5|6|7|0|1|8|84|17|5|6|7|0|1|18|-15|17|5|6|7|0|4|8|30|14|85|8|86|14|87|8|88|89|90|100|0|8|21|17|5|6|7|0|1|8|86|17|5|6|7|0|1|8|22|14|91|-220|-222|-209|-218|-205|-207|-201|-203|-1|-199|8|92|17|5|6|7|0|1|18|-15|17|5|6|7|0|6|8|93|17|5|6|7|0|1|18|-15|17|5|6|7|0|2|8|94|80|19|100|8|95|14|96|-238|-240|-234|-236|8|97|14|98|8|30|14|99|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3491272235|1309965025280|515000000|8|40|14|59|8|41|38|39|3491272236|1309965025280|516000000|8|42|14|100|-234|-253|8|101|14|102|8|21|17|5|6|7|0|2|8|101|17|5|6|7|0|1|8|22|14|103|-273|-275|8|97|17|5|6|7|0|1|8|22|14|104|-273|-281|-234|-271|-230|-232|-1|-228|8|105|17|5|6|7|0|1|18|-15|17|5|6|7|0|6|8|106|17|5|6|7|0|1|18|-15|17|5|6|7|0|9|8|107|14|108|8|109|17|5|6|7|0|3|8|110|14|111|8|112|14|113|8|21|17|5|6|7|0|1|8|112|17|5|6|7|0|1|8|22|14|114|-315|-317|-307|-313|-301|-305|8|115|80|-246|8|76|17|5|6|7|0|3|8|77|14|116|8|79|80|19|13|8|21|17|5|6|7|0|1|8|77|17|5|6|7|0|1|8|22|14|117|-336|-338|-327|-334|-301|-325|8|30|14|118|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3491262219|1309965025280|499000000|8|40|14|59|8|41|38|39|3491272295|1309965025280|575000000|8|42|14|119|-301|-346|8|46|14|62|8|120|17|5|6|7|0|0|-301|-364|8|21|17|5|6|7|0|1|8|107|17|5|6|7|0|1|8|22|14|121|-370|-372|-301|-368|-297|-299|-293|-295|8|30|14|122|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3491262210|1309965025280|490000000|8|40|14|59|8|41|38|39|3491272289|1309965025280|569000000|8|42|14|119|-293|-380|8|46|14|62|8|123|17|5|6|7|0|0|-293|-398|8|48|14|124|-289|-291|-1|-287|8|125|17|5|6|7|0|2|18|-15|14|126|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|127|-412|-414|-406|-410|-1|-404|8|30|14|128|8|129|17|5|6|7|0|1|18|-15|17|5|6|7|0|2|8|130|14|131|8|21|17|5|6|7|0|1|8|130|17|5|6|7|0|1|8|22|14|132|-434|-436|-428\"\n contents3 = \"|-432|-424|-426|-1|-422|8|133|17|5|6|7|0|0|-1|-442|8|134|14|135|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3491257291|1309965025280|571000000|8|40|14|59|8|41|38|39|3491272236|1309965025280|516000000|8|42|14|111|-1|-448|8|136|10|-5|8|137|17|5|6|7|0|1|18|-15|17|5|6|7|0|4|8|84|17|5|6|7|0|1|18|-15|17|5|6|7|0|4|8|30|14|138|8|86|14|87|8|88|89|90|100|0|8|21|17|5|6|7|0|1|8|86|17|5|6|7|0|1|8|22|14|91|-491|-493|-480|-489|-476|-478|-472|-474|8|97|14|139|8|30|14|140|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3491272233|1309965025280|513000000|8|40|14|59|8|41|38|39|3491272236|1309965025280|516000000|8|42|14|100|-472|-503|-468|-470|-1|-466|8|141|10|-5|8|142|14|143|8|46|14|62|8|144|14|145|8|123|17|5|6|7|0|2|18|-15|14|146|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|147|-535|-537|-529|-533|-1|-527|8|148|14|149|8|48|14|150|8|120|17|5|6|7|0|2|18|-15|14|151|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|152|-555|-557|-549|-553|-1|-547|8|153|17|5|6|7|0|2|18|-15|14|87|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|91|-571|-573|-565|-569|-1|-563|8|154|17|5|6|7|0|0|-1|-579|8|155|17|5|6|7|0|3|8|156|38|39|3491257291|1309965025280|571000000|8|157|89|90|1|0|8|158|14|159|-1|-583|8|21|17|5|6|7|0|3|8|142|17|5|6|7|0|1|8|22|14|160|-597|-599|8|144|17|5|6|7|0|1|8|22|14|145|-597|-605|8|13|17|5|6|7|0|1|8|22|14|161|-597|-611|-1|-595|8|162|17|5|6|7|0|12|8|163|14|164|8|30|14|165|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3491257553|1309965025280|833000000|8|40|14|59|8|41|38|39|3491273130|1309965025280|410000000|8|42|14|166|-619|-625|8|44|14|65|8|167|17|5|6|7|0|1|18|-15|14|128|-619|-643|8|168|14|169|8|170|17|5|6|7|0|0|-619|-651|8|171|17|5|6|7|0|0|-619|-655|8|172|14|173|8|46|14|174|8|48|14|175|8|176|14|177|-1|-617|8|178|17|5|6|7|0|1|8|179|17|5|6|7|0|0|-669|-671|-1|-667|0|0|\"\n end\n\n # Save & Continue\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/CreditCourseProposalRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"#{contents1}#{contents2}#{contents3}\"\n },\n {\n 'subst' => 'true',\n :dyn_variables => [\n {\"name\" => opts[:action_request_id_name], \"regexp\" => opts[:action_request_id_regexp]}\n ]\n }\n )\n \n #@request.add(\"DEBUG/affliated_orgs_id_name/%%_#{opts[:action_request_id_name]}%%\", {}, {'subst' => 'true'})\n\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/DocumentRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|7|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|F189A2C8AED6D8071F8F9362674CF016|org.kuali.student.core.document.ui.client.service.DocumentRpcService|isAuthorizedUploadDocuments|java.lang.String/2004016611|%%_#{opts[:proposal_dyn_var_name]}%%|referenceType.clu.proposal|1|2|3|4|2|5|5|6|7|\"\n },\n {\n 'subst' => 'true'\n }\n )\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/DocumentRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|7|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|F189A2C8AED6D8071F8F9362674CF016|org.kuali.student.core.document.ui.client.service.DocumentRpcService|getRefDocIdsForRef|java.lang.String/2004016611|kuali.org.RefObjectType.ProposalInfo|%%_#{opts[:proposal_dyn_var_name]}%%|1|2|3|4|2|5|5|6|7|\"\n },\n {\n 'subst' => 'true'\n }\n )\n \n \n \n #\n # Support Documents\n # Nothing uploaded\n #\n\n @request.add_thinktime(5)\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/DocumentRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|7|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|F189A2C8AED6D8071F8F9362674CF016|org.kuali.student.core.document.ui.client.service.DocumentRpcService|getRefDocIdsForRef|java.lang.String/2004016611|kuali.org.RefObjectType.ProposalInfo|%%_#{opts[:proposal_dyn_var_name]}%%|1|2|3|4|2|5|5|6|7|\"\n },\n {\n 'subst' => 'true'\n }\n )\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/WorkflowRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|71417C94A72A0CF76A43A2B36B8E3E1B|org.kuali.student.core.workflow.ui.client.service.WorkflowRpcService|getActionsRequested|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/WorkflowRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|71417C94A72A0CF76A43A2B36B8E3E1B|org.kuali.student.core.workflow.ui.client.service.WorkflowRpcService|getDocumentStatus|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n\n\n if(!opts[:collaborator].nil?)\n contents1 = \"5|0|194|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|1ED48DA6F48F82765FE7B58378EA94E0|org.kuali.student.lum.lu.ui.course.client.service.CourseRpcService|validate|org.kuali.student.common.assembly.data.Data/3184510345|org.kuali.student.common.assembly.data.Data|java.util.LinkedHashMap/1551059846|org.kuali.student.common.assembly.data.Data$StringKey/758802082|passFail|org.kuali.student.common.assembly.data.Data$BooleanValue/4261226833|java.lang.Boolean/476441737|audit|finalExamStatus|org.kuali.student.common.assembly.data.Data$StringValue/3151113388|STD|campusLocations|org.kuali.student.common.assembly.data.Data$DataValue/1692468409|org.kuali.student.common.assembly.data.Data$IntegerKey/134469241|java.lang.Integer/3438268394|AL|_runtimeData|id-translation|All|code|#{opts[:subject_area]}#{opts[:course_suffix]}|courseNumberSuffix|#{opts[:course_suffix]}|courseSpecificLOs|loCategoryInfoList|id|%%_#{opts[:lo_category_id_var_name]}%%|loRepository|kuali.loRepository.key.singleUse|metaInfo|createId|admin|createTime|org.kuali.student.common.assembly.data.Data$DateValue/2929953165|java.sql.Timestamp/1769758459|updateId|updateTime|versionInd|0|name|#{opts[:lo_category]}|state|active|type|loCategoryType.subject|loDisplayInfoList|loInfo|sequence|desc|formatted|#{opts[:lo_cat_text]}|plain|%%_#{opts[:lo_category_var_name]}%%|loRepositoryKey|#{opts[:propose_person]}|3|SINGLE USE LO|draft|kuali.lo.type.singleUse|courseTitle|#{opts[:course_title]}|creditOptions|fixedCreditValue|10.0|kuali.creditType.credit.degree.10.0|resultValues|kuali.resultComponentType.credit.degree.fixed|Credits, Fixed|crossListings|descr|#{opts[:course_description]}|duration|atpDurationTypeKey|kuali.atp.duration.Semester|timeQuantity|org.kuali.student.common.assembly.data.Data$IntegerValue/3605481012|Semester|effectiveDate|expenditure|affiliatedOrgs|%%_#{opts[:affliated_orgs_id_name]}%%|orgId|%%_#{opts[:admin_dep_var_name]}%%|percentage|org.kuali.student.common.assembly.data.Data$LongValue/3784756947|java.lang.Long/4227064769|#{admin_org}|fees|feeAmounts|currencyQuantity|currencyTypeKey|kuali.currency.type.usdollars.cents|feeType|kuali.enum.type.feeTypes.labFee|%%_#{opts[:lab_fee_id_name]}%%|2|rateType|fixedRateFee|Fixed Rate Fee|Laboratory Fee|formats|activities|activityType|kuali.lu.type.activity.Lab|contactHours|unitQuantity|5|unitType|kuali.atp.duration.week|per week|defaultEnrollmentEstimate|kuali.atp.duration.Week|Week|%%_#{opts[:atp_duration_week_var_name]}%%|4|unitsContentOwner|Lab|%%_#{opts[:lab_var_name]}%%|termsOffered|kuali.lu.type.CreditCourseFormatShell|gradingOptions|kuali.resultComponent.grade.letter|Letter|%%_#{opts[:clu_ref_dyn_var_name]}%%|instructors|personId|#{opts[:instructor]}|#{opts[:instructor]}, #{opts[:instructor]}(#{opts[:instructor]})|joints|level|100|6|pilotCourse|revenues|%%_#{opts[:revenues_id_name]}%%|REVENUE|%%_#{opts[:revenue_id_name]}%%|specialTopicsCourse|startTerm|kuali.atp.FA2008-2009|subjectArea|#{opts[:subject_area]}|kuali.atp.season.Any|Any|transcriptTitle|#{opts[:course_short_title]}|kuali.lu.type.CreditCourse|#{opts[:oversight_dept_number]}|#{oversight_department}|unitsDeployment|variations|versionInfo|currentVersionStart|sequenceNumber|versionIndId|%%_#{opts[:version_ind_id_name]}%%|Fall Semester of 2008|Standard final Exam|proposal|workflowNode|PreRoute|%%_#{opts[:proposal_dyn_var_name]}%%|8|proposalReference|proposalReferenceType|kuali.proposal.referenceType.clu|proposerOrg|proposerPerson|rationale|#{opts[:course_rationale]}|Saved|kuali.proposal.type.course.create|workflowId|%%_#{opts[:proposal_doc_id_var_name]}%%|collaboratorInfo|collaborators|action|F|actionRequestId|%%_#{opts[:action_request_id_name]}%%|actionRequestStatus|Active|author|canRevokeRequest|firstName|#{opts[:collaborator]}|lastName|permission|KS-SYS~Edit Document|principalId\"\n contents2 = \"|1|2|3|4|1|5|5|6|7|0|38|8|9|10|11|0|8|12|10|-5|8|13|14|15|8|16|17|5|6|7|0|2|18|19|0|14|20|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|23|-19|-21|-12|-17|-1|-10|8|24|14|25|8|26|14|27|8|28|17|5|6|7|0|1|18|-15|17|5|6|7|0|3|8|29|17|5|6|7|0|1|18|-15|17|5|6|7|0|6|8|30|14|31|8|32|14|33|8|34|17|5|6|7|0|5|8|35|14|36|8|37|38|39|3759152200|1288490188800|0|8|40|14|36|8|41|38|39|3759152200|1288490188800|0|8|42|14|43|-45|-51|8|44|14|45|8|46|14|47|8|48|14|49|-41|-43|-37|-39|8|50|17|5|6|7|0|0|-37|-73|8|51|17|5|6|7|0|8|8|52|14|43|8|53|17|5|6|7|0|2|8|54|14|55|8|56|14|55|-79|-83|8|30|14|57|8|58|14|33|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3491070679|1309965025280|959000000|8|40|14|59|8|41|38|39|3491081941|1309965025280|221000000|8|42|14|60|-79|-95|8|44|14|61|8|46|14|62|8|48|14|63|-37|-77|-33|-35|-1|-31|8|64|14|65|8|66|17|5|6|7|0|1|18|-15|17|5|6|7|0|7|8|67|14|68|8|30|14|69|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3479039543|1309965025280|823000000|8|40|14|59|8|41|38|39|3479039543|1309965025280|823000000|8|42|14|43|-125|-131|8|70|17|5|6|7|0|1|18|-15|14|68|-125|-147|8|46|14|62|8|48|14|71|8|21|17|5|6|7|0|1|8|48|17|5|6|7|0|1|8|22|14|72|-159|-161|-125|-157|-121|-123|-1|-119|8|73|17|5|6|7|0|0|-1|-167|8|74|17|5|6|7|0|1|8|56|14|75|-1|-171|8|76|17|5|6|7|0|3|8|77|14|78|8|79|80|19|2|8|21|17|5|6|7|0|1|8|77|17|5|6|7|0|1|8|22|14|81|-188|-190|-179|-186|-1|-177|8|82|38|39|470887936|1219770712064|0|8|83|17|5|6|7|0|1|8|84|17|5|6|7|0|1|18|-15|17|5|6|7|0|4|8|30|14|85|8|86|14|87|8|88|89|90|100|0|8|21|17|5|6|7|0|1|8|86|17|5|6|7|0|1|8|22|14|91|-220|-222|-209|-218|-205|-207|-201|-203|-1|-199|8|92|17|5|6|7|0|1|18|-15|17|5|6|7|0|6|8|93|17|5|6|7|0|1|18|-15|17|5|6|7|0|2|8|94|80|19|100|8|95|14|96|-238|-240|-234|-236|8|97|14|98|8|30|14|99|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3491076835|1309965025280|115000000|8|40|14|59|8|41|38|39|3491081856|1309965025280|136000000|8|42|14|100|-234|-253|8|101|14|102|8|21|17|5|6|7|0|2|8|101|17|5|6|7|0|1|8|22|14|103|-273|-275|8|97|17|5|6|7|0|1|8|22|14|104|-273|-281|-234|-271|-230|-232|-1|-228|8|105|17|5|6|7|0|1|18|-15|17|5|6|7|0|6|8|106|17|5|6|7|0|1|18|-15|17|5|6|7|0|9|8|107|14|108|8|109|17|5|6|7|0|3|8|110|14|111|8|112|14|113|8|21|17|5|6|7|0|1|8|112|17|5|6|7|0|1|8|22|14|114|-315|-317|-307|-313|-301|-305|8|115|80|-246|8|76|17|5|6|7|0|3|8|77|14|116|8|79|80|19|13|8|21|17|5|6|7|0|1|8|77|17|5|6|7|0|1|8|22|14|117|-336|-338|-327|-334|-301|-325|8|30|14|118|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3491067152|1309965025280|432000000|8|40|14|59|8|41|38|39|3491081913|1309965025280|193000000|8|42|14|119|-301|-346|8|46|14|62|8|120|17|5|6|7|0|0|-301|-364|8|21|17|5|6|7|0|1|8|107|17|5|6|7|0|1|8|22|14|121|-370|-372|-301|-368|-297|-299|-293|-295|8|30|14|122|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3491067143|1309965025280|423000000|8|40|14|59|8|41|38|39|3491081907|1309965025280|187000000|8\"\n contents3 = \"|42|14|119|-293|-380|8|46|14|62|8|123|17|5|6|7|0|0|-293|-398|8|48|14|124|-289|-291|-1|-287|8|125|17|5|6|7|0|2|18|-15|14|126|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|127|-412|-414|-406|-410|-1|-404|8|30|14|128|8|129|17|5|6|7|0|1|18|-15|17|5|6|7|0|2|8|130|14|131|8|21|17|5|6|7|0|1|8|130|17|5|6|7|0|1|8|22|14|132|-434|-436|-428|-432|-424|-426|-1|-422|8|133|17|5|6|7|0|0|-1|-442|8|134|14|135|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3491062523|1309965025280|803000000|8|40|14|59|8|41|38|39|3491081856|1309965025280|136000000|8|42|14|136|-1|-448|8|137|10|-5|8|138|17|5|6|7|0|1|18|-15|17|5|6|7|0|4|8|84|17|5|6|7|0|1|18|-15|17|5|6|7|0|4|8|30|14|139|8|86|14|87|8|88|89|90|100|0|8|21|17|5|6|7|0|1|8|86|17|5|6|7|0|1|8|22|14|91|-491|-493|-480|-489|-476|-478|-472|-474|8|97|14|140|8|30|14|141|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3491076834|1309965025280|114000000|8|40|14|59|8|41|38|39|3491081856|1309965025280|136000000|8|42|14|100|-472|-503|-468|-470|-1|-466|8|142|10|-5|8|143|14|144|8|46|14|62|8|145|14|146|8|123|17|5|6|7|0|2|18|-15|14|147|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|148|-535|-537|-529|-533|-1|-527|8|149|14|150|8|48|14|151|8|120|17|5|6|7|0|2|18|-15|14|152|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|153|-555|-557|-549|-553|-1|-547|8|154|17|5|6|7|0|2|18|-15|14|87|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|91|-571|-573|-565|-569|-1|-563|8|155|17|5|6|7|0|0|-1|-579|8|156|17|5|6|7|0|3|8|157|38|39|3491062523|1309965025280|803000000|8|158|89|90|1|0|8|159|14|160|-1|-583|8|21|17|5|6|7|0|3|8|143|17|5|6|7|0|1|8|22|14|161|-597|-599|8|145|17|5|6|7|0|1|8|22|14|146|-597|-605|8|13|17|5|6|7|0|1|8|22|14|162|-597|-611|-1|-595|8|163|17|5|6|7|0|12|8|164|14|165|8|30|14|166|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3491062758|1309965025280|38000000|8|40|14|59|8|41|38|39|3491082765|1309965025280|45000000|8|42|14|167|-619|-625|8|44|14|65|8|168|17|5|6|7|0|1|18|-15|14|128|-619|-643|8|169|14|170|8|171|17|5|6|7|0|0|-619|-651|8|172|17|5|6|7|0|0|-619|-655|8|173|14|174|8|46|14|175|8|48|14|176|8|177|14|178|-1|-617|8|179|17|5|6|7|0|1|8|180|17|5|6|7|0|1|18|-15|17|5|6|7|0|9|8|181|14|182|8|183|14|184|8|185|14|186|8|187|10|-5|8|188|10|11|1|8|189|14|190|8|191|14|190|8|192|14|193|8|194|14|190|-673|-675|-669|-671|-1|-667|0|0|\"\n else\n contents1 = \"5|0|180|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|1ED48DA6F48F82765FE7B58378EA94E0|org.kuali.student.lum.lu.ui.course.client.service.CourseRpcService|validate|org.kuali.student.common.assembly.data.Data/3184510345|org.kuali.student.common.assembly.data.Data|java.util.LinkedHashMap/1551059846|org.kuali.student.common.assembly.data.Data$StringKey/758802082|passFail|org.kuali.student.common.assembly.data.Data$BooleanValue/4261226833|java.lang.Boolean/476441737|audit|finalExamStatus|org.kuali.student.common.assembly.data.Data$StringValue/3151113388|STD|campusLocations|org.kuali.student.common.assembly.data.Data$DataValue/1692468409|org.kuali.student.common.assembly.data.Data$IntegerKey/134469241|java.lang.Integer/3438268394|AL|_runtimeData|id-translation|All|code|#{opts[:subject_area]}#{opts[:course_suffix]}|courseNumberSuffix|#{opts[:course_suffix]}|courseSpecificLOs|loCategoryInfoList|id|%%_#{opts[:lo_category_id_var_name]}%%|loRepository|kuali.loRepository.key.singleUse|metaInfo|createId|admin|createTime|org.kuali.student.common.assembly.data.Data$DateValue/2929953165|java.sql.Timestamp/1769758459|updateId|updateTime|versionInd|0|name|#{opts[:lo_category]}|state|active|type|loCategoryType.subject|loDisplayInfoList|loInfo|sequence|desc|formatted|#{opts[:lo_cat_text]}|plain|%%_#{opts[:lo_category_var_name]}%%|loRepositoryKey|#{opts[:propose_person]}|3|SINGLE USE LO|draft|kuali.lo.type.singleUse|courseTitle|#{opts[:course_title]}|creditOptions|fixedCreditValue|10.0|kuali.creditType.credit.degree.10.0|resultValues|kuali.resultComponentType.credit.degree.fixed|Credits, Fixed|crossListings|descr|#{opts[:course_description]}|duration|atpDurationTypeKey|kuali.atp.duration.Semester|timeQuantity|org.kuali.student.common.assembly.data.Data$IntegerValue/3605481012|Semester|effectiveDate|expenditure|affiliatedOrgs|%%_#{opts[:affliated_orgs_id_name]}%%|orgId|%%_#{opts[:admin_dep_var_name]}%%|percentage|org.kuali.student.common.assembly.data.Data$LongValue/3784756947|java.lang.Long/4227064769|#{admin_org}|fees|feeAmounts|currencyQuantity|currencyTypeKey|kuali.currency.type.usdollars.cents|feeType|kuali.enum.type.feeTypes.labFee|%%_#{opts[:lab_fee_id_name]}%%|2|rateType|fixedRateFee|Fixed Rate Fee|Laboratory Fee|formats|activities|activityType|kuali.lu.type.activity.Lab|contactHours|unitQuantity|5|unitType|kuali.atp.duration.week|per week|defaultEnrollmentEstimate|kuali.atp.duration.Week|Week|%%_#{opts[:atp_duration_week_var_name]}%%|4|unitsContentOwner|Lab|%%_#{opts[:lab_var_name]}%%|termsOffered|kuali.lu.type.CreditCourseFormatShell|gradingOptions|kuali.resultComponent.grade.letter|Letter|%%_#{opts[:clu_ref_dyn_var_name]}%%|instructors|personId|#{opts[:instructor]}|#{opts[:instructor]}, #{opts[:instructor]}(#{opts[:instructor]})|joints|level|100|6|pilotCourse|revenues|%%_#{opts[:revenues_id_name]}%%|REVENUE|%%_#{opts[:revenue_id_name]}%%|specialTopicsCourse|startTerm|kuali.atp.FA2008-2009|subjectArea|#{opts[:subject_area]}|kuali.atp.season.Any|Any|transcriptTitle|#{opts[:course_short_title]}|kuali.lu.type.CreditCourse|#{opts[:oversight_dept_number]}|#{oversight_department}|unitsDeployment|variations|versionInfo|currentVersionStart|sequenceNumber|versionIndId|%%_#{opts[:version_ind_id_name]}%%|Fall Semester of 2008|Standard final Exam|proposal|workflowNode|PreRoute|%%_#{opts[:proposal_dyn_var_name]}%%|8|proposalReference|proposalReferenceType|kuali.proposal.referenceType.clu|proposerOrg|proposerPerson|rationale|#{opts[:course_rationale]}|Saved|kuali.proposal.type.course.create|workflowId|%%_#{opts[:proposal_doc_id_var_name]}%%|collaboratorInfo|collaborators\"\n contents2 = \"|1|2|3|4|1|5|5|6|7|0|38|8|9|10|11|0|8|12|10|-5|8|13|14|15|8|16|17|5|6|7|0|2|18|19|0|14|20|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|23|-19|-21|-12|-17|-1|-10|8|24|14|25|8|26|14|27|8|28|17|5|6|7|0|1|18|-15|17|5|6|7|0|3|8|29|17|5|6|7|0|1|18|-15|17|5|6|7|0|6|8|30|14|31|8|32|14|33|8|34|17|5|6|7|0|5|8|35|14|36|8|37|38|39|3759152200|1288490188800|0|8|40|14|36|8|41|38|39|3759152200|1288490188800|0|8|42|14|43|-45|-51|8|44|14|45|8|46|14|47|8|48|14|49|-41|-43|-37|-39|8|50|17|5|6|7|0|0|-37|-73|8|51|17|5|6|7|0|8|8|52|14|43|8|53|17|5|6|7|0|2|8|54|14|55|8|56|14|55|-79|-83|8|30|14|57|8|58|14|33|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3498783321|1309965025280|601000000|8|40|14|59|8|41|38|39|3498791687|1309965025280|967000000|8|42|14|60|-79|-95|8|44|14|61|8|46|14|62|8|48|14|63|-37|-77|-33|-35|-1|-31|8|64|14|65|8|66|17|5|6|7|0|1|18|-15|17|5|6|7|0|7|8|67|14|68|8|30|14|69|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3479039543|1309965025280|823000000|8|40|14|59|8|41|38|39|3479039543|1309965025280|823000000|8|42|14|43|-125|-131|8|70|17|5|6|7|0|1|18|-15|14|68|-125|-147|8|46|14|62|8|48|14|71|8|21|17|5|6|7|0|1|8|48|17|5|6|7|0|1|8|22|14|72|-159|-161|-125|-157|-121|-123|-1|-119|8|73|17|5|6|7|0|0|-1|-167|8|74|17|5|6|7|0|1|8|56|14|75|-1|-171|8|76|17|5|6|7|0|3|8|77|14|78|8|79|80|19|2|8|21|17|5|6|7|0|1|8|77|17|5|6|7|0|1|8|22|14|81|-188|-190|-179|-186|-1|-177|8|82|38|39|470887936|1219770712064|0|8|83|17|5|6|7|0|1|8|84|17|5|6|7|0|1|18|-15|17|5|6|7|0|4|8|30|14|85|8|86|14|87|8|88|89|90|100|0|8|21|17|5|6|7|0|1|8|86|17|5|6|7|0|1|8|22|14|91|-220|-222|-209|-218|-205|-207|-201|-203|-1|-199|8|92|17|5|6|7|0|1|18|-15|17|5|6|7|0|6|8|93|17|5|6|7|0|1|18|-15|17|5|6|7|0|2|8|94|80|19|100|8|95|14|96|-238|-240|-234|-236|8|97|14|98|8|30|14|99|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3498789104|1309965025280|384000000|8|40|14|59|8|41|38|39|3498791612|1309965025280|892000000|8|42|14|100|-234|-253|8|101|14|102|8|21|17|5|6|7|0|2|8|101|17|5|6|7|0|1|8|22|14|103|-273|-275|8|97|17|5|6|7|0|1|8|22|14|104|-273|-281|-234|-271|-230|-232|-1|-228|8|105|17|5|6|7|0|1|18|-15|17|5|6|7|0|6|8|106|17|5|6|7|0|1|18|-15|17|5|6|7|0|9|8|107|14|108|8|109|17|5|6|7|0|3|8|110|14|111|8|112|14|113|8|21|17|5|6|7|0|1|8|112|17|5|6|7|0|1|8|22|14|114|-315|-317|-307|-313|-301|-305|8|115|80|-246|8|76|17|5|6|7|0|3|8|77|14|116|8|79|80|19|13|8|21|17|5|6|7|0|1|8|77|17|5|6|7|0|1|8|22|14|117|-336|-338|-327|-334|-301|-325|8|30|14|118|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3498779786|1309965025280|66000000|8|40|14|59|8|41|38|39|3498791652|1309965025280|932000000|8|42|14|119|-301|-346|8|46|14|62|8|120|17|5|6|7|0|0|-301|-364|8|21|17|5|6|7|0|1|8|107|17|5|6|7|0|1|8|22|14|121|-370|-372|-301|-368|-297|-299|-293|-295|8|30|14|122|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3498779778|1309965025280|58000000|8|40|14|59|8|41|38|39|3498791646|1309965025280|926000000|8|42|14|119|-293|-380|8|46|14|62|8|123|17|5|6|7|0|0|-293|-398|8|48|14|124|-289|-291|-1|-287|8|125|17|5|6|7|0|2|18|-15|14|126|8|21|17|5|6\"\n contents3 = \"|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|127|-412|-414|-406|-410|-1|-404|8|30|14|128|8|129|17|5|6|7|0|1|18|-15|17|5|6|7|0|2|8|130|14|131|8|21|17|5|6|7|0|1|8|130|17|5|6|7|0|1|8|22|14|132|-434|-436|-428|-432|-424|-426|-1|-422|8|133|17|5|6|7|0|0|-1|-442|8|134|14|135|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3498775348|1309965025280|628000000|8|40|14|59|8|41|38|39|3498791612|1309965025280|892000000|8|42|14|136|-1|-448|8|137|10|-5|8|138|17|5|6|7|0|1|18|-15|17|5|6|7|0|4|8|84|17|5|6|7|0|1|18|-15|17|5|6|7|0|4|8|30|14|139|8|86|14|87|8|88|89|90|100|0|8|21|17|5|6|7|0|1|8|86|17|5|6|7|0|1|8|22|14|91|-491|-493|-480|-489|-476|-478|-472|-474|8|97|14|140|8|30|14|141|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3498789104|1309965025280|384000000|8|40|14|59|8|41|38|39|3498791612|1309965025280|892000000|8|42|14|100|-472|-503|-468|-470|-1|-466|8|142|10|-5|8|143|14|144|8|46|14|62|8|145|14|146|8|123|17|5|6|7|0|2|18|-15|14|147|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|148|-535|-537|-529|-533|-1|-527|8|149|14|150|8|48|14|151|8|120|17|5|6|7|0|2|18|-15|14|152|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|153|-555|-557|-549|-553|-1|-547|8|154|17|5|6|7|0|2|18|-15|14|87|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|91|-571|-573|-565|-569|-1|-563|8|155|17|5|6|7|0|0|-1|-579|8|156|17|5|6|7|0|3|8|157|38|39|3498775348|1309965025280|628000000|8|158|89|90|1|0|8|159|14|160|-1|-583|8|21|17|5|6|7|0|3|8|143|17|5|6|7|0|1|8|22|14|161|-597|-599|8|145|17|5|6|7|0|1|8|22|14|146|-597|-605|8|13|17|5|6|7|0|1|8|22|14|162|-597|-611|-1|-595|8|163|17|5|6|7|0|12|8|164|14|165|8|30|14|166|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3498775596|1309965025280|876000000|8|40|14|59|8|41|38|39|3498792455|1309965025280|735000000|8|42|14|167|-619|-625|8|44|14|65|8|168|17|5|6|7|0|1|18|-15|14|128|-619|-643|8|169|14|170|8|171|17|5|6|7|0|0|-619|-651|8|172|17|5|6|7|0|0|-619|-655|8|173|14|174|8|46|14|175|8|48|14|176|8|177|14|178|-1|-617|8|179|17|5|6|7|0|1|8|180|17|5|6|7|0|0|-669|-671|-1|-667|0|0|\"\n end\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/CreditCourseProposalRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"#{contents1}#{contents2}#{contents3}\"\n },\n {\n 'subst' => 'true'\n }\n )\n\n @request.add_thinktime(5)\n\n\n # Submit to worflow\n if(opts[:submit])\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/WorkflowRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|71417C94A72A0CF76A43A2B36B8E3E1B|org.kuali.student.core.workflow.ui.client.service.WorkflowRpcService|submitDocumentWithId|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/WorkflowRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|71417C94A72A0CF76A43A2B36B8E3E1B|org.kuali.student.core.workflow.ui.client.service.WorkflowRpcService|getActionsRequested|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/WorkflowRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|71417C94A72A0CF76A43A2B36B8E3E1B|org.kuali.student.core.workflow.ui.client.service.WorkflowRpcService|getDocumentStatus|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n end\n \n \n end",
"def call\n html.gsub(/\\[\\[([^\\]|]*)(\\|([^\\]]*))?\\]\\]/) do\n link = $1\n desc = $3 ? $3 : $1\n\n \"<a href=\\\"#{to_link link}\\\">#{to_description desc}</a>\"\n end\n end",
"def parse_facts(elem)\n para = elem.following_siblings()[0]\n values = para.search(\"strong\")\n launchdate = nil\n launchvehicle = nil\n launchsite = nil\n values.each { |value| \n field_name = value.inner_html\n field_value = value.next.to_plain_text.gsub /^\\s/, \"\"\n # substrings are a hack for presence of at the start of some values\n if field_name == \"Launch Date:\"\n launchdate = field_value \n elsif field_name == \"Launch Vehicle:\"\n launchvehicle = field_value\n elsif field_name == \"Launch Site:\" \n launchsite = field_value\n elsif field_name == \"Mass:\"\n @mass = field_value.gsub /\\skg/, \"\"\n elsif field_name == \"Nominal Power:\"\n #\"Nominal Power: 375.0 W\", 1966-073A.html \n @nominalPower = clean_ws( field_value.gsub(/\\sW/, \"\") ).gsub /\\s/, \"\"\n else\n puts \"Unknown field: #{field_name}\"\n end\n \n }\n @launch = Launch.new(launchdate, launchsite, @id, launchvehicle)\n end",
"def get_pdp_info\n wait_for_product_detail_div(TimeOut::WAIT_BIG_CONST * 2)\n\n # Get information on pdp page\n long_name = long_name_txt.text\n age = RspecEncode.remove_nbsp(age_txt.text.strip)\n description = description_txt.text\n moreinfo_lb = get_more_info[:moreinfo_lb].gsub(\"\\n\", '').gsub(/\\s+/, ' ') # Remove '\\n' and double space characters\n moreinfo_text = get_more_info[:moreinfo_txt].gsub(\"\\n\", '').gsub(/\\s+/, ' ')\n special_message = (has_special_message_txt?) ? special_message_txt.text : '' # If special message exist => return special text. Else, return ''\n legal_top = (has_legal_top_txt?) ? legal_top_txt.text : ''\n legal_bottom = (has_legal_bottom_txt?) ? legal_bottom_txt.text : ''\n learning_difference = (has_learning_difference_txt?(wait: TimeOut::WAIT_SMALL_CONST)) ? learning_difference_txt.text : ''\n review = has_review_box? # If Review box exist ->'true' else 'false'\n more_like_this = has_more_like_this? # If More Like this box exist ->'true' else 'false'\n write_a_review = has_write_a_review_div?(wait: TimeOut::WAIT_SMALL_CONST) # If 'Write a Review' box exist ->'true' else 'false'\n add_to_wishlist = has_add_to_wishlist_lnk? # If 'Add to Wishlist' button exist -> 'true' else 'false'\n\n # Get price:\n if has_strike_price_txt?(wait: TimeOut::WAIT_SMALL_CONST)\n price = strike_price_txt.text\n else\n price = (has_price_txt?(wait: TimeOut::WAIT_SMALL_CONST)) ? price_txt.text : ''\n end\n\n # Get attributes info\n content_type = ''\n curriculum = ''\n notable = ''\n work_with = ''\n publisher = ''\n size = ''\n attributes_div.each do |a|\n attr = a.text.split(':')\n content_type = attr[1].strip if attr[0].include?('Type')\n curriculum = attr[1].strip if attr[0].include?('Curriculum')\n notable = attr[1..-1].join(':').strip if attr[0].include?('Notable')\n work_with = attr[1].gsub(', ', ',').strip if attr[0].include?('Works With')\n publisher = attr[1].strip if attr[0].include?('Publisher')\n size = attr[1].strip if attr[0].include?('Size')\n end\n\n # Get trailer\n has_trailer = trailer?\n trailer_link = ''\n trailer_link = find('.video')['data-largeimage'].to_s.gsub('\"', '\\\"') if has_trailer_box?\n\n # Get teaches (Skills list)\n teaches = []\n teaches_txt.each do |teach|\n teaches.push(teach.text)\n end\n\n # Get product detail\n details = get_detail\n\n # get value of Add to Cart button: Add to Cart\n add_to_cart_val = ''\n\n if has_add_to_cart_btn?(wait: 0)\n add_to_cart_val = find(:xpath, \"(//input[contains(@value,'Add to Cart')])[1]\")[:value] if has_xpath?(\"//input[contains(@value,'Add to Cart')]\", wait: 0)\n add_to_cart_val = find(:xpath, \"(//*[contains(text(),'Add to Cart')])[1]\").text if has_xpath?(\"(//*[contains(text(),'Add to Cart')])[1]\", wait: 0)\n else\n add_to_cart_val = 'Not Available'\n end\n\n # get value of buy now button: Buy Now\n buy_now_btn = ''\n execute_script(\"$('#sub-nav-grnbar-btn').css('display', 'block');\")\n buy_now_btn = find('#sub-nav-grnbar-btn')[:value] if has_css?('#sub-nav-grnbar-btn', wait: 0)\n\n has_credits_link = has_credits_lnk?\n\n # Put all info into array\n { long_name: long_name,\n age: age,\n description: description,\n content_type: content_type,\n curriculum: curriculum,\n notable: notable,\n work_with: work_with,\n publisher: publisher,\n size: size,\n moreinfo_lb: moreinfo_lb,\n moreinfo_txt: moreinfo_text,\n special_message: special_message,\n legal_top: legal_top,\n price: price,\n details: details,\n learning_difference: learning_difference,\n legal_bottom: legal_bottom,\n teaches: teaches,\n has_trailer: has_trailer,\n trailer_link: trailer_link,\n has_credits_link: has_credits_link,\n review: review,\n more_like_this: more_like_this,\n write_a_review: write_a_review,\n add_to_wishlist: add_to_wishlist,\n add_to_cart_btn: add_to_cart_val,\n buy_now_btn: buy_now_btn }\n end",
"def gather_offers_per_page(doc, indeed_offers)\n # Gathers all the cards on the page and collects info from each owne of them\n puts \"Pulling information from each card offer per page\"\n doc.search('.jobsearch-SerpJobCard').each do |job_card|\n # If offer already exists none will be created\n unless Offer.where(external_id: job_card['data-jk']).present?\n\n puts \"Getting information from offer #{job_card['data-jk']}, #{job_card.search('h2').text}\"\n new_offer_hash = {\n external_id: job_card['data-jk'],\n company: job_card.search('.company').text,\n title: job_card.search('h2').text,\n salary: \"\", # unrefined text, not suited for ranges yet\n category: \"Software Development\",\n position: '',\n job_type: \"\",\n tags: [],\n location: job_card.search('.location').text,\n listing_url: \"https://www.indeed.com.mx/ver-empleo?jk=#{job_card['data-jk']}\",\n candidate_required_location: \"Mexico\",\n source: 'indeed'\n }\n collect_salary(job_card, new_offer_hash)\n scrape_individual_offer(new_offer_hash, \"ver-empleo?jk=#{job_card['data-jk']}\")\n # Save each job with complete information into a list of all the offers from indeed\n\n puts 'create a new offer test'\n\n new_offer = Offer.where(external_id: new_offer_hash['id'].to_s, source: 'indeed').first_or_initialize\n copy_offer_variables(new_offer, new_offer_hash)\n new_offer.source = 'indeed'\n new_offer_hash[:tags].each do |tag_name|\n new_offer.tags << Tag.find_by(name: tag_name)\n end\n new_offer.save!\n end\n end\n end",
"def generate_description\n text, _ = parse_description\n parts = []\n parts << text if text.length > 0\n parts << JSON.generate(config) if config && config.length > 0\n parts.join(\"\\n\")\n end",
"def goals\n assessment_type = params[:assessment_type]\n country_name = params[:country_name]\n technical_area_ids = params[:areas].to_s.split(\"-\")\n @disease_ids = params[:diseases]\n country = Country.find_by_name country_name\n @publication = AssessmentPublication.find_by_named_id assessment_type\n if country.present? && @publication.present?\n @assessment =\n Assessment.deep_load(country.try(:alpha3), @publication.try(:id))\n end\n if @assessment.blank?\n render \"assessment_not_found\"\n return\n end\n @plan =\n Plan.new_from_assessment(\n assessment: @assessment,\n technical_area_ids: technical_area_ids,\n is_5_year_plan: params[:plan_term]&.start_with?(\"5\")\n )\n end",
"def create_proposal(title, oversight_department, admin_org, opts={})\n \n lo_cat = \"Scientific method\"\n lo_cat_text = \"LO Cat Text\"\n \n defaults = {\n :propose_person => '%%_username%%', #user is the dynvar from users.csv\n :mode => 'blank',\n :nav_homepage => true,\n :submit => true,\n :append_unique_id => false, #tell tsung to append unique id on title\n :instructor => 'fred', #BUG - HARDCODED - can't use dynvar though because of ajax search\n :collaborator => @request.config.directory[\"users\"][\"collaborator\"][\"username\"],\n :first_expected_offering => @request.config.directory[\"atp\"][\"name\"],\n :subject_area => \"BSCI\",\n :oversight_dept_number => \"65\", #BUG - right now hardcoded to BSCI, search that returned this was removed\n :course_suffix => \"123\",\n :course_short_title => \"Perf Course\",\n :course_title => title,\n :proposal_title => title,\n :course_description => \"My fake description.\",\n :course_rationale => \"My fake rationale.\",\n :lo_create => FALSE,\n :lo_category => lo_cat,\n :lo_cat_text => lo_cat_text,\n :lo_name => @request.config.directory[\"lo\"][\"name\"],\n :admin_dep_var_name => \"admin_dep_org_id\",\n :admin_dep_var_regexp => 'org.resultColumn.orgId\\\"\\,\\\"\\([^\\\"]+\\)',\n :proposal_dyn_var_name => \"proposal_id\",\n :proposal_dyn_var_regexp => '\\\"proposal\\\"\\,\\\"workflowNode\\\"\\,\\\"PreRoute\\\"\\,\\\"\\([^\\\"]+\\)',\n :proposal_doc_id_var_name => \"proposal_doc_id\",\n :proposal_doc_id_var_regexp => 'workflowId\\\"\\,\\\"\\([^\\\"]+\\)\\\"',\n :clu_ref_dyn_var_name => \"clu_ref_id\",\n :clu_ref_dyn_var_regexp => '\\\"id\\\"\\,\\\"\\([^\\\"]+\\)',\n :result_com_var_name => \"result_component_id\",\n :result_com_var_regexp => '\\\"ResultComponent 1\\\"\\,\\\"\\([^\\\"]+\\)',\n :enroll_est_var_name => \"default_enrollment_estimate_id\",\n :enroll_est_var_regexp => 'defaultEnrollmentEstimate\\\"\\,\\\"kuali.atp.duration.Week\\\"\\,\\\"Week\\\"\\,\\\"\\([^\\\"]+\\)',\n :lab_var_name => \"lab_id\",\n :lab_var_regexp => 'draft\\\"\\,\\\"unitsContentOwner\\\"\\,\\\"Lab\\\"\\,\\\"\\([^\\\"]+\\)',\n :lab_fee_id_name => 'cp_lab_fee_id',\n :lab_fee_id_regexp => 'kuali.enum.type.feeTypes.labFee\\\"\\,\\\"\\([^\\\"]+\\)',\n :revenues_id_name => 'cp_revenues_id',\n :revenues_id_regexp => 'revenues\\\"\\,\\\"\\([^\\\"]+\\)',\n :revenue_id_name => 'cp_revenue_id',\n :revenue_id_regexp => 'REVENUE\\\"\\,\\\"\\([^\\\"]+\\)',\n :joints_var_name => \"joints_num\",\n :joints_var_regexp => 'joints\\\"\\,\\\"[^\\\"]+\\\"\\,\\\"\\([^\\\"]+\\)',\n :fee_info_id_dyn_var_name => 'fee_info_id',\n :fee_info_id_dyn_var_regexp => '\\\"fees\\\"\\,\\\"\\([^\\\"]+\\)',\n :fee_info_dyn_var_name => 'fee_info',\n :fee_info_dyn_var_regexp => 'org.kuali.student.lum.lu.dto.CluFeeInfo\\\"\\,\\\"\\([^\\\"]+\\)',\n :clu_info_dyn_var_name => 'clu_info',\n :clu_info_dyn_var_regexp => 'org.kuali.student.lum.lu.dto.CluInfo\\\"\\,\\\"\\([^\\\"]+\\)',\n :lu_dto_clu_format_dyn_var_name => \"lu_dto_clu_format\",\n :lu_dto_clu_format_dyn_var_regexp => 'org.kuali.student.lum.lu.dto.CluInfo\\\"\\,\\\"Credit Course\\\"\\,\\\"[^\\\"]+\\\"\\,\\\"formats\\\"\\,\\\"\\([^\\\"]+\\)',\n :lu_dto_clu_activities_dyn_var_name => \"lu_dto_clu_activites\",\n :lu_dto_clu_activities_dyn_var_regexp => 'org.kuali.student.lum.lu.dto.CluInfo\\\"\\,\\\"Credit Course\\\"\\,\\\"[^\\\"]+\\\"\\,\\\"formats\\\"\\,\\\"[^\\\"]+\\\"\\,\\\"0\\\"\\,\\\"activities\\\"\\,\\\"\\([^\\\"]+\\)',\n :outcome_id_var_name => \"outcome_id\",\n :outcome_id_var_regexp => 'outcomeId\\\"\\,\\\"\\([^\\\"]+\\)',\n :lo_category_var_name => \"lo_category\",\n #:lo_category_var_regexp => lo_cat_text + '\\\"\\,\\\"plain\\\"\\,\\\"id\\\"\\,\\\"\\([^\\\"]+\\)',\n :lo_category_var_regexp => lo_cat_text + '\\\"\\,\\\"plain\\\"\\,\\\"\\([^\\\"]+\\)',\n :lo_category_id_var_name => \"lo_category_id\",\n :lo_category_id_var_regexp => 'lo.resultColumn.categoryId\"\\,\\\"\\([^\\\"]+\\)',\n :lo_child_id_dyn_var_name => \"lo_child_id\",\n :lo_child_id_dyn_var_regexp => 'childLo\\\"\\,\\\"\\([^\\\"]+\\)',\n :single_use_lo_dyn_var_name => \"single_use_lo\",\n :single_use_lo_dyn_var_regexp => 'includedSingleUseLo\\\"\\,\\\"\\([^\\\"]+\\)',\n :atp_duration_week_var_name => 'atp_duration_week',\n :atp_duration_week_var_regexp => 'kuali.atp.duration.Week\\\"\\,\\\"Week\\\"\\,\\\"\\([^\\\"]+\\)',\n :version_ind_id_name => 'cp_version_ind_id',\n :version_ind_id_regexp => 'versionIndId\\\"\\,\\\"\\([^\\\"]+\\)',\n :affliated_orgs_id_name => 'cp_affiliated_orgs_id',\n :affliated_orgs_id_regexp => 'affiliatedOrgs\\\"\\,\\\"\\([^\\\"]+\\)',\n :action_request_id_name => 'cp_action_request_id',\n :action_request_id_regexp => 'actionRequestId\\\"\\,\\\"\\([^\\\"]+\\)'\n }\n \n # Version for the doc at each step. We'll increment on each usage\n # So first usage should eval to 0\n version_indicator = -1\n \n opts = defaults.merge(opts)\n \n title << '_%%ts_user_server:get_unique_id%%' if(opts[:append_unique_id])\n \n if(opts[:mode] != \"blank\")\n # select template or copy course...\n end\n \n # Navigate to Curriculum Mgmt\n self.homepage() unless(!opts[:nav_homepage])\n \n puts \"creating proposal as: #{opts[:propose_person]}\"\n \n # Create a course\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SecurityRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|7|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|13BFCB3640903B473D12816447D1469D|org.kuali.student.common.ui.client.service.SecurityRpcService|checkAdminPermission|java.lang.String/2004016611|#{opts[:propose_person]}|useCurriculumReview|1|2|3|4|2|5|5|6|7|\"\n }, {'subst' => 'true'}\n )\n \n \n # Click Start blank proposal\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/CreditCourseProposalRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|15|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|A239E8C5A2EDCD8BCE6061BF191A8095|org.kuali.student.lum.lu.ui.course.client.service.CreditCourseProposalRpcService|getMetadata|java.lang.String/2004016611|java.util.Map||java.util.HashMap/962170901|documentTypeName|kuali.proposal.type.course.create|DtoState|Draft|DtoNextState|DtoWorkflowNode|PreRoute|1|2|3|4|2|5|6|7|8|4|5|9|5|10|5|11|5|12|5|13|5|7|5|14|5|15|\"\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/statementRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|335FF062A700107AB2A642B325C6C5C5|org.kuali.student.lum.program.client.rpc.StatementRpcService|getStatementTypesForStatementTypeForCourse|java.lang.String/2004016611|kuali.statement.type.course|1|2|3|4|1|5|6|\"\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|11|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|enumeration.queryParam.enumerationType|kuali.lu.campusLocation|enumeration.management.search|enumeration.resultColumn.sortKey|1|2|3|4|1|5|5|0|0|6|1|7|8|0|9|10|11|0|0|\"\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|8|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|atp.search.atpSeasonTypes|atp.resultColumn.atpSeasonTypeName|1|2|3|4|1|5|5|0|0|6|0|7|8|0|0|\"\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|8|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|atp.search.atpDurationTypes|atp.resultColumn.atpDurTypeName|1|2|3|4|1|5|5|0|0|6|0|7|8|0|0|\"\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|18|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|lrc.queryParam.resultComponent.type|kuali.resultComponentType.grade.finalGrade|lrc.queryParam.resultComponent.idRestrictionList|java.lang.String/2004016611|kuali.resultComponent.grade.letter|kuali.resultComponent.grade.passFail|kuali.resultComponent.grade.satisfactory|kuali.resultComponent.grade.completedNotation|kuali.resultComponent.grade.percentage|lrc.search.resultComponent|lrc.resultColumn.resultComponent.id|1|2|3|4|1|5|5|0|0|6|2|7|8|0|9|7|10|6|5|11|12|11|13|11|14|11|15|11|16|0|17|18|0|0|\"\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|11|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|enumeration.queryParam.enumerationType|kuali.lu.finalExam.status|enumeration.management.search|enumeration.resultColumn.sortKey|1|2|3|4|1|5|5|0|0|6|1|7|8|0|9|10|11|0|0|\"\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|14|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|lrc.queryParam.resultComponent.idRestrictionList|java.lang.String/2004016611|kuali.resultComponentType.credit.degree.fixed|kuali.resultComponentType.credit.degree.range|kuali.resultComponentType.credit.degree.multiple|lrc.search.resultComponentType|lrc.resultColumn.resultComponent.id|1|2|3|4|1|5|5|0|0|6|1|7|8|6|3|9|10|9|11|9|12|0|13|14|0|0|\"\n }\n )\n \n # DUPE\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|14|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|lrc.queryParam.resultComponent.idRestrictionList|java.lang.String/2004016611|kuali.resultComponentType.credit.degree.fixed|kuali.resultComponentType.credit.degree.range|kuali.resultComponentType.credit.degree.multiple|lrc.search.resultComponentType|lrc.resultColumn.resultComponent.id|1|2|3|4|1|5|5|0|0|6|1|7|8|6|3|9|10|9|11|9|12|0|13|14|0|0|\"\n }\n )\n \n # DUPE\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|14|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|lrc.queryParam.resultComponent.idRestrictionList|java.lang.String/2004016611|kuali.resultComponentType.credit.degree.fixed|kuali.resultComponentType.credit.degree.range|kuali.resultComponentType.credit.degree.multiple|lrc.search.resultComponentType|lrc.resultColumn.resultComponent.id|1|2|3|4|1|5|5|0|0|6|1|7|8|6|3|9|10|9|11|9|12|0|13|14|0|0|\"\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|11|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|enumeration.queryParam.enumerationType|kuali.lu.fee.rateType|enumeration.management.search|enumeration.resultColumn.sortKey|1|2|3|4|1|5|5|0|0|6|1|7|8|0|9|10|11|0|0|\"\n }\n )\n \n # DUPE\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|11|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|enumeration.queryParam.enumerationType|kuali.lu.fee.rateType|enumeration.management.search|enumeration.resultColumn.sortKey|1|2|3|4|1|5|5|0|0|6|1|7|8|0|9|10|11|0|0|\"\n }\n )\n \n # DUPE\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|11|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|enumeration.queryParam.enumerationType|kuali.lu.fee.rateType|enumeration.management.search|enumeration.resultColumn.sortKey|1|2|3|4|1|5|5|0|0|6|1|7|8|0|9|10|11|0|0|\"\n }\n )\n \n # DUPE\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|11|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|enumeration.queryParam.enumerationType|kuali.lu.fee.rateType|enumeration.management.search|enumeration.resultColumn.sortKey|1|2|3|4|1|5|5|0|0|6|1|7|8|0|9|10|11|0|0|\"\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|11|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|enumeration.queryParam.enumerationType|kuali.lu.campusLocation|enumeration.management.search|enumeration.resultColumn.sortKey|1|2|3|4|1|5|5|0|0|6|1|7|8|0|9|10|11|0|0|\"\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|11|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|enumeration.queryParam.enumerationType|kuali.lu.campusLocation|enumeration.management.search|enumeration.resultColumn.sortKey|1|2|3|4|1|5|5|0|0|6|1|7|8|0|9|10|11|0|0|\"\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|8|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|atp.search.atpSeasonTypes|atp.resultColumn.atpSeasonTypeName|1|2|3|4|1|5|5|0|0|6|0|7|8|0|0|\"\n }\n )\n \n # DUPE\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|8|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|atp.search.atpSeasonTypes|atp.resultColumn.atpSeasonTypeName|1|2|3|4|1|5|5|0|0|6|0|7|8|0|0|\"\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|18|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|lrc.queryParam.resultComponent.type|kuali.resultComponentType.grade.finalGrade|lrc.queryParam.resultComponent.idRestrictionList|java.lang.String/2004016611|kuali.resultComponent.grade.letter|kuali.resultComponent.grade.passFail|kuali.resultComponent.grade.satisfactory|kuali.resultComponent.grade.completedNotation|kuali.resultComponent.grade.percentage|lrc.search.resultComponent|lrc.resultColumn.resultComponent.id|1|2|3|4|1|5|5|0|0|6|2|7|8|0|9|7|10|6|5|11|12|11|13|11|14|11|15|11|16|0|17|18|0|0|\"\n }\n )\n \n # DUPE\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|18|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|lrc.queryParam.resultComponent.type|kuali.resultComponentType.grade.finalGrade|lrc.queryParam.resultComponent.idRestrictionList|java.lang.String/2004016611|kuali.resultComponent.grade.letter|kuali.resultComponent.grade.passFail|kuali.resultComponent.grade.satisfactory|kuali.resultComponent.grade.completedNotation|kuali.resultComponent.grade.percentage|lrc.search.resultComponent|lrc.resultColumn.resultComponent.id|1|2|3|4|1|5|5|0|0|6|2|7|8|0|9|7|10|6|5|11|12|11|13|11|14|11|15|11|16|0|17|18|0|0|\"\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/CommentRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|62D53D0C5087061126A72510E98E7E9A|org.kuali.student.core.comments.ui.client.service.CommentRpcService|getUserRealName|java.lang.String/2004016611|#{opts[:propose_person]}|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|11|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|subjectCode.queryParam.code||subjectCode.search.orgsForSubjectCode|subjectCode.resultColumn.orgLongName|1|2|3|4|1|5|5|0|0|6|1|7|8|0|9|10|11|0|0|\"\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|20|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|atp.advancedAtpSearchParam.atpType|java.lang.String/2004016611|kuali.atp.type.Spring|kuali.atp.type.Summer|kuali.atp.type.Fall|kuali.atp.type.Session1|kuali.atp.type.Session2|kuali.atp.type.Mini-mester1A|kuali.atp.type.Mini-mester1B|atp.advancedAtpSearchParam.atpStartDateAtpConstraintIdExclusive||atp.search.advancedAtpSearch|atp.resultColumn.atpStartDate|1|2|3|4|1|5|5|0|0|6|2|7|8|6|7|9|10|9|11|9|12|9|13|9|14|9|15|9|16|0|7|17|0|18|19|20|0|0|\"\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|19|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|atp.advancedAtpSearchParam.atpType|java.lang.String/2004016611|kuali.atp.type.Spring|kuali.atp.type.Summer|kuali.atp.type.Fall|kuali.atp.type.Session1|kuali.atp.type.Session2|kuali.atp.type.Mini-mester1A|kuali.atp.type.Mini-mester1B|atp.advancedAtpSearchParam.atpStartDateAtpConstraintId|atp.search.advancedAtpSearch|atp.resultColumn.atpStartDate|1|2|3|4|1|5|5|0|0|6|2|7|8|6|7|9|10|9|11|9|12|9|13|9|14|9|15|9|16|0|7|17|0|0|18|19|0|0|\"\n }\n )\n \n \n \n #\n # Pg1 - Course Information\n #\n\n @request.add_thinktime(5)\n\n # Course Subject Code\n # AJAX popup while typing in subject area\n for i in 1..opts[:subject_area].length\n itr = i-1\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|12|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|search|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.lang.Boolean/476441737|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|subjectCode.queryParam.code|#{opts[:subject_area][0..itr]}|subjectCode.search.subjectCodeGeneric|subjectCode.resultColumn.code|1|2|3|4|1|5|5|0|6|0|7|1|8|9|0|10|11|12|0|0|\"\n } \n ) \n end\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|11|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|subjectCode.queryParam.code|#{opts[:subject_area]}|subjectCode.search.orgsForSubjectCode|subjectCode.resultColumn.orgLongName|1|2|3|4|1|5|5|0|0|6|1|7|8|0|9|10|11|0|0|\"\n } \n )\n\n @request.add_thinktime(3)\n\n # Instructor\n for i in 1..opts[:instructor].length\n itr = i-1\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|12|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|search|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.lang.Boolean/476441737|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|person.queryParam.personGivenName|#{opts[:instructor][0..itr]}|person.search.personQuickViewByGivenName|person.resultColumn.DisplayName|1|2|3|4|1|5|5|0|6|0|7|1|8|9|0|10|11|12|0|0|\"\n } \n ) \n end\n\n\n @request.add_thinktime(22)\n\n # Save & Continue\n contents1 = \"5|0|41|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|A239E8C5A2EDCD8BCE6061BF191A8095|org.kuali.student.lum.lu.ui.course.client.service.CreditCourseProposalRpcService|saveData|org.kuali.student.common.assembly.data.Data/3184510345|org.kuali.student.common.assembly.data.Data|java.util.LinkedHashMap/1551059846|org.kuali.student.common.assembly.data.Data$StringKey/758802082|proposal|org.kuali.student.common.assembly.data.Data$DataValue/1692468409|type|org.kuali.student.common.assembly.data.Data$StringValue/3151113388|kuali.proposal.type.course.create|workflowNode|PreRoute|name|#{opts[:proposal_title]}|_runtimeData|dirty|org.kuali.student.common.assembly.data.Data$BooleanValue/4261226833|java.lang.Boolean/476441737|rationale|#{opts[:course_rationale]}|courseTitle|#{opts[:course_title]}|transcriptTitle|subjectArea|courseNumberSuffix|instructors|#{opts[:course_short_title]}|#{opts[:subject_area]}|#{opts[:course_suffix]}|org.kuali.student.common.assembly.data.Data$IntegerKey/134469241|java.lang.Integer/3438268394|personId|#{opts[:instructor]}|id-translation|#{opts[:instructor]}, #{opts[:instructor]}(#{opts[:instructor]})|descr|plain|#{opts[:course_description]}\"\n contents2 = \"|1|2|3|4|1|5|5|6|7|0|8|8|9|10|5|6|7|0|5|8|11|12|13|8|14|12|15|8|16|12|17|8|18|10|5|6|7|0|1|8|19|10|5|6|7|0|2|-11|20|21|1|8|22|20|-22|-15|-17|-5|-13|-23|12|23|-1|-3|8|24|12|25|-13|10|5|6|7|0|1|-17|10|5|6|7|0|5|-26|20|-22|8|26|20|-22|8|27|20|-22|8|28|20|-22|8|29|20|-22|-29|-17|-1|-13|-35|12|30|-37|12|31|-39|12|32|-41|10|5|6|7|0|1|33|34|0|10|5|6|7|0|2|8|35|12|36|8|18|10|5|6|7|0|1|8|35|10|5|6|7|0|1|8|37|12|38|-58|-60|-52|-56|-47|-49|-1|8|29|8|39|10|5|6|7|0|2|8|40|12|41|-13|10|5|6|7|0|1|-17|10|5|6|7|0|1|-71|20|-22|-74|-17|-69|-13|-1|-67|0|0|\"\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/CreditCourseProposalRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"#{contents1}#{contents2}\"\n },\n {\n :dyn_variables => [\n {\"name\" => opts[:proposal_dyn_var_name], \"re\" => opts[:proposal_dyn_var_regexp]},\n {\"name\" => opts[:clu_ref_dyn_var_name], \"re\" => opts[:clu_ref_dyn_var_regexp]},\n {\"name\" => opts[:proposal_doc_id_var_name], \"re\" => opts[:proposal_doc_id_var_regexp]},\n {\"name\" => opts[:version_ind_id_name], \"re\" => opts[:version_ind_id_regexp]}\n ]\n }\n )\n\n #@request.add(\"DEBUG/proposal_dyn_var_name/%%_#{opts[:proposal_dyn_var_name]}%%\", {}, {'subst' => 'true'})\n #@request.add(\"DEBUG/clu_ref_dyn_var_name/%%_#{opts[:clu_ref_dyn_var_name]}%%\", {}, {'subst' => 'true'})\n #@request.add(\"DEBUG/proposal_doc_id_var_name/%%_#{opts[:proposal_doc_id_var_name]}%%\", {}, {'subst' => 'true'})\n #@request.add(\"DEBUG/version_ind_id_name/%%_#{opts[:version_ind_id_name]}%%\", {}, {'subst' => 'true'})\n \n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/WorkflowRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|71417C94A72A0CF76A43A2B36B8E3E1B|org.kuali.student.core.workflow.ui.client.service.WorkflowRpcService|getActionsRequested|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/WorkflowRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|71417C94A72A0CF76A43A2B36B8E3E1B|org.kuali.student.core.workflow.ui.client.service.WorkflowRpcService|getDocumentStatus|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/ProposalRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|12BDE6C2DA6A7CF74BE0FBF074E806E1|org.kuali.student.core.proposal.ui.client.service.ProposalRpcService|getProposalByWorkflowId|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n \n \n \n \n #\n # Pg 2 - Governance\n # Campus Locations: All\n #\n\n # COC Org\n # Biology Dept\n\n @request.add_thinktime(8)\n\n # Admin Org\n # Botany\n for i in 1..admin_org.length\n itr = i-1\n if(i == admin_org.length)\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|16|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|search|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.lang.Boolean/476441737|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|org.queryParam.orgOptionalLongName|#{admin_org[0..itr]}|org.queryParam.orgOptionalType|java.lang.String/2004016611|kuali.org.Department|kuali.org.College|org.search.generic||1|2|3|4|1|5|5|0|6|0|7|2|8|9|0|10|8|11|7|2|12|13|12|14|0|15|16|0|0|\"\n },\n {\n :dyn_variables => [\n {\"name\" => opts[:admin_dep_var_name], \"re\" => opts[:admin_dep_var_regexp]}\n ]\n }\n )\n\n #@request.add(\"DEBUG/admin_dep_var_name/%%_#{opts[:admin_dep_var_name]}%%\", {}, {'subst' => 'true'})\n else\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|16|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|search|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.lang.Boolean/476441737|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|org.queryParam.orgOptionalLongName|#{admin_org[0..itr]}|org.queryParam.orgOptionalType|java.lang.String/2004016611|kuali.org.Department|kuali.org.College|org.search.generic||1|2|3|4|1|5|5|0|6|0|7|2|8|9|0|10|8|11|7|2|12|13|12|14|0|15|16|0|0|\"\n } \n )\n end \n end\n\n @request.add_thinktime(15)\n\n # Save & Continue\n contents1 = \"5|0|101|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|A239E8C5A2EDCD8BCE6061BF191A8095|org.kuali.student.lum.lu.ui.course.client.service.CreditCourseProposalRpcService|saveData|org.kuali.student.common.assembly.data.Data/3184510345|org.kuali.student.common.assembly.data.Data|java.util.LinkedHashMap/1551059846|org.kuali.student.common.assembly.data.Data$StringKey/758802082|campusLocations|org.kuali.student.common.assembly.data.Data$DataValue/1692468409|org.kuali.student.common.assembly.data.Data$IntegerKey/134469241|java.lang.Integer/3438268394|org.kuali.student.common.assembly.data.Data$StringValue/3151113388|AL|code|#{opts[:subject_area]}#{opts[:course_suffix]}|courseNumberSuffix|#{opts[:course_suffix]}|courseSpecificLOs|courseTitle|#{opts[:course_title]}|creditOptions|crossListings|descr|plain|#{opts[:course_description]}|expenditure|affiliatedOrgs|fees|formats|gradingOptions|id|%%_#{opts[:clu_ref_dyn_var_name]}%%|instructors|personId|#{opts[:instructor]}|_runtimeData|id-translation|#{opts[:instructor]}, #{opts[:instructor]}(#{opts[:instructor]})|joints|level|100|metaInfo|createId|#{opts[:propose_person]}|createTime|org.kuali.student.common.assembly.data.Data$DateValue/2929953165|java.sql.Timestamp/1769758459|updateId|updateTime|versionInd|#{version_indicator+=1}|pilotCourse|org.kuali.student.common.assembly.data.Data$BooleanValue/4261226833|java.lang.Boolean/476441737|revenues|specialTopicsCourse|state|draft|subjectArea|#{opts[:subject_area]}|termsOffered|transcriptTitle|#{opts[:course_short_title]}|type|kuali.lu.type.CreditCourse|unitsContentOwner|#{opts[:oversight_dept_number]}|#{oversight_department}|unitsDeployment|%%_#{opts[:admin_dep_var_name]}%%|#{admin_org}|variations|versionInfo|currentVersionStart|sequenceNumber|org.kuali.student.common.assembly.data.Data$LongValue/3784756947|java.lang.Long/4227064769|versionIndId|%%_#{opts[:version_ind_id_name]}%%|dirty|proposal|workflowNode|PreRoute|%%_#{opts[:proposal_dyn_var_name]}%%|2|name|#{opts[:proposal_title]}|proposalReference|proposalReferenceType|kuali.proposal.referenceType.clu|proposerOrg|proposerPerson|rationale|#{opts[:course_rationale]}|Saved|kuali.proposal.type.course.create|workflowId|%%_#{opts[:proposal_doc_id_var_name]}%%|collaboratorInfo|collaborators\"\n contents2 = \"|1|2|3|4|1|5|5|6|7|0|32|8|9|10|5|6|7|0|1|11|12|0|13|14|-1|8|9|8|15|13|16|8|17|13|18|8|19|10|5|6|7|0|0|-1|-15|8|20|13|21|8|22|10|5|6|7|0|0|-1|-21|8|23|10|5|6|7|0|0|-1|-25|8|24|10|5|6|7|0|1|8|25|13|26|-1|-29|8|27|10|5|6|7|0|1|8|28|10|5|6|7|0|0|-37|-39|-1|-35|8|29|10|5|6|7|0|0|-1|-43|8|30|10|5|6|7|0|0|-1|-47|8|31|10|5|6|7|0|0|-1|-51|8|32|13|33|8|34|10|5|6|7|0|1|11|-8|10|5|6|7|0|2|8|35|13|36|8|37|10|5|6|7|0|1|8|35|10|5|6|7|0|1|8|38|13|39|-69|-71|-63|-67|-59|-61|-1|-57|8|40|10|5|6|7|0|0|-1|-77|8|41|13|42|8|43|10|5|6|7|0|5|8|44|13|45|8|46|47|48|1854995943|1314259992576|519000000|8|49|13|45|8|50|47|48|1854995943|1314259992576|519000000|8|51|13|52|-1|-83|8|53|54|55|0|8|56|10|5|6|7|0|0|-1|-102|8|57|54|-101|8|58|13|59|8|60|13|61|8|62|10|5|6|7|0|0|-1|-112|8|63|13|64|8|65|13|66|8|67|10|5|6|7|0|2|11|-8|13|68|8|37|10|5|6|7|0|1|11|-8|10|5|6|7|0|1|8|38|13|69|-128|-130|-122|-126|-1|8|67|8|70|10|5|6|7|0|2|11|-8|13|71|8|37|10|5|6|7|0|1|11|-8|10|5|6|7|0|1|8|38|13|72|-145|-147|-139|-143|-1|8|70|8|73|10|5|6|7|0|0|-1|-154|8|74|10|5|6|7|0|3|8|75|47|48|1854995943|1314259992576|519000000|8|76|77|78|1|0|8|79|13|80|-1|-158|8|37|10|5|6|7|0|2|8|60|10|5|6|7|0|1|8|38|13|61|-172|-174|8|81|10|5|6|7|0|3|8|9|54|55|1|8|67|54|-186|8|70|54|-186|-172|-180|-1|-170|8|82|10|5|6|7|0|12|8|83|13|84|8|32|13|85|8|43|10|5|6|7|0|5|8|44|13|45|8|46|47|48|1854996146|1314259992576|722000000|8|49|13|45|8|50|47|48|1854997119|1314259992576|695000000|8|51|13|86|-193|-199|8|87|13|88|8|89|10|5|6|7|0|1|11|-8|13|33|-193|-217|8|90|13|91|8|92|10|5|6|7|0|0|-193|-225|8|93|10|5|6|7|0|0|-193|-229|8|94|13|95|8|58|13|96|8|65|13|97|8|98|13|99|-1|-191|8|100|10|5|6|7|0|1|8|101|10|5|6|7|0|0|-243|-245|-1|-241|0|0|\"\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/CreditCourseProposalRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"#{contents1}#{contents2}\"\n },\n {\n 'subst' => 'true'\n }\n )\n \n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/WorkflowRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|71417C94A72A0CF76A43A2B36B8E3E1B|org.kuali.student.core.workflow.ui.client.service.WorkflowRpcService|getActionsRequested|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/WorkflowRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|71417C94A72A0CF76A43A2B36B8E3E1B|org.kuali.student.core.workflow.ui.client.service.WorkflowRpcService|getDocumentStatus|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/ProposalRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|12BDE6C2DA6A7CF74BE0FBF074E806E1|org.kuali.student.core.proposal.ui.client.service.ProposalRpcService|getProposalByWorkflowId|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|11|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|lu.queryParam.luOptionalLuTypeStartsWith|kuali.lu.type.activity.|lu.search.all.lu.Types|lu.resultColumn.luTypeName|1|2|3|4|1|5|5|0|0|6|1|7|8|0|9|10|11|0|0|\"\n }\n )\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|11|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|enumeration.queryParam.enumerationType|kuali.atptype.duration|enumeration.management.search|enumeration.resultColumn.sortKey|1|2|3|4|1|5|5|0|0|6|1|7|8|0|9|10|11|0|0|\"\n }\n )\n \n \n \n #\n # Course Logistics \n # Term: Any\n # Duration Type: Semester\n # Duration Count: 2\n # Assessment Scale: Letter\n # Standard Final Exam\n # Outcome: 10 credits\n # Course Format\n # => Activity Type: Lab, Contact Hours: 5, Frequency: weekly\n # => Duration Type: Weekly\n # => Duration: 13\n # => Anticipated class size: 100\n\n @request.add_thinktime(30)\n\n # Save & Continue\n contents1 = \"5|0|126|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|A239E8C5A2EDCD8BCE6061BF191A8095|org.kuali.student.lum.lu.ui.course.client.service.CreditCourseProposalRpcService|saveData|org.kuali.student.common.assembly.data.Data/3184510345|org.kuali.student.common.assembly.data.Data|java.util.LinkedHashMap/1551059846|org.kuali.student.common.assembly.data.Data$StringKey/758802082|campusLocations|org.kuali.student.common.assembly.data.Data$DataValue/1692468409|org.kuali.student.common.assembly.data.Data$IntegerKey/134469241|java.lang.Integer/3438268394|org.kuali.student.common.assembly.data.Data$StringValue/3151113388|AL|_runtimeData|id-translation|All|code|#{opts[:subject_area]}#{opts[:course_suffix]}|courseNumberSuffix|#{opts[:course_suffix]}|courseSpecificLOs|courseTitle|#{opts[:course_title]}|creditOptions|dirty|type|org.kuali.student.common.assembly.data.Data$BooleanValue/4261226833|java.lang.Boolean/476441737|fixedCreditValue|created|kuali.resultComponentType.credit.degree.fixed|10|crossListings|descr|plain|#{opts[:course_description]}|expenditure|affiliatedOrgs|fees|formats|activities|activityType|defaultEnrollmentEstimate|kuali.lu.type.activity.Lab|contactHours|unitQuantity|unitType|kuali.atp.duration.week|duration|atpDurationTypeKey|timeQuantity|kuali.atp.duration.Week|org.kuali.student.common.assembly.data.Data$IntegerValue/3605481012|gradingOptions|kuali.resultComponent.grade.letter|id|%%_#{opts[:clu_ref_dyn_var_name]}%%|instructors|personId|#{opts[:instructor]}|#{opts[:instructor]}, #{opts[:instructor]}(#{opts[:instructor]})|joints|level|100|metaInfo|createId|#{opts[:propose_person]}|createTime|org.kuali.student.common.assembly.data.Data$DateValue/2929953165|java.sql.Timestamp/1769758459|updateId|updateTime|versionInd|#{version_indicator+=1}|pilotCourse|revenues|specialTopicsCourse|state|draft|subjectArea|#{opts[:subject_area]}|termsOffered|kuali.atp.season.Any|transcriptTitle|#{opts[:course_short_title]}|kuali.lu.type.CreditCourse|unitsContentOwner|#{opts[:oversight_dept_number]}|#{oversight_department}|unitsDeployment|%%_#{opts[:admin_dep_var_name]}%%|#{admin_org}|variations|versionInfo|currentVersionStart|sequenceNumber|org.kuali.student.common.assembly.data.Data$LongValue/3784756947|java.lang.Long/4227064769|versionIndId|%%_#{opts[:version_ind_id_name]}%%|finalExamStatus|audit|passFail|proposal|workflowNode|PreRoute|%%_#{opts[:proposal_dyn_var_name]}%%|3|name|#{opts[:proposal_title]}|proposalReference|proposalReferenceType|kuali.proposal.referenceType.clu|proposerOrg|proposerPerson|rationale|#{opts[:course_rationale]}|Saved|kuali.proposal.type.course.create|workflowId|%%_#{opts[:proposal_doc_id_var_name]}%%|collaboratorInfo|collaborators|kuali.atp.duration.Semester|STD\"\n contents2 = \"|1|2|3|4|1|5|5|6|7|0|36|8|9|10|5|6|7|0|2|11|12|0|13|14|8|15|10|5|6|7|0|1|11|-8|10|5|6|7|0|1|8|16|13|17|-12|-14|-5|-10|-1|-3|8|18|13|19|8|20|13|21|8|22|10|5|6|7|0|0|-1|-24|8|23|13|24|8|25|10|5|6|7|0|1|11|-8|10|5|6|7|0|3|8|15|10|5|6|7|0|3|8|26|10|5|6|7|0|2|8|27|28|29|1|8|30|28|-48|-40|-42|8|31|28|-48|-38|10|5|6|7|0|1|-42|10|5|6|7|0|1|-51|28|-48|-54|-42|-40|-38|-36|-38|-46|13|32|-49|13|33|-32|-34|-1|-30|8|34|10|5|6|7|0|0|-1|-62|8|35|10|5|6|7|0|1|8|36|13|37|-1|-66|8|38|10|5|6|7|0|1|8|39|10|5|6|7|0|0|-74|-76|-1|-72|8|40|10|5|6|7|0|0|-1|-80|8|41|10|5|6|7|0|1|11|-8|10|5|6|7|0|2|8|42|10|5|6|7|0|1|11|-8|10|5|6|7|0|5|-38|10|5|6|7|0|3|-42|10|5|6|7|0|2|8|43|28|-48|8|44|28|-48|-101|-42|8|31|28|-48|-38|10|5|6|7|0|1|-42|10|5|6|7|0|1|-110|28|-48|-113|-42|-101|-38|-98|-38|-106|13|45|8|46|10|5|6|7|0|3|8|47|13|33|-38|10|5|6|7|0|1|-42|10|5|6|7|0|2|-124|28|-48|8|48|28|-48|-127|-42|-122|-38|-133|13|49|-98|-120|8|50|10|5|6|7|0|3|-38|10|5|6|7|0|1|-42|10|5|6|7|0|2|8|51|28|-48|8|52|28|-48|-141|-42|-138|-38|-146|13|53|-148|54|12|13|-98|-136|-108|54|12|100|-94|-96|-90|-92|8|15|10|5|6|7|0|2|8|31|28|-48|-38|10|5|6|7|0|1|-42|10|5|6|7|0|1|-159|28|-48|-162|-42|-157|-38|-90|-155|-86|-88|-1|-84|8|55|10|5|6|7|0|1|11|-8|13|56|-1|8|55|8|57|13|58|8|59|10|5|6|7|0|1|11|-8|10|5|6|7|0|2|8|60|13|61|8|15|10|5|6|7|0|1|8|60|10|5|6|7|0|1|8|16|13|62|-189|-191|-183|-187|-179|-181|-1|-177|8|63|10|5|6|7|0|0|-1|-197|8|64|13|65|8|66|10|5|6|7|0|5|8|67|13|68|8|69|70|71|1854995943|1314259992576|519000000|8|72|13|68|8|73|70|71|1856244060|1314259992576|636000000|8|74|13|75|-1|-203|8|76|28|29|0|8|77|10|5|6|7|0|0|-1|-222|8|78|28|-221|8|79|13|80|8|81|13|82|8|83|10|5|6|7|0|1|11|-8|13|84|-1|8|83|8|85|13|86|8|27|13|87|8|88|10|5|6|7|0|2|11|-8|13|89|8|15|10|5|6|7|0|1|11|-8|10|5|6|7|0|1|8|16|13|90|-251|-253|-245|-249|-1|-243|8|91|10|5|6|7|0|2|11|-8|13|92|8|15|10|5|6|7|0|1|11|-8|10|5|6|7|0|1|8|16|13|93|-267|-269|-261|-265|-1|-259|8|94|10|5|6|7|0|0|-1|-275|8|95|10|5|6|7|0|3|8|96|70|71|1854995943|1314259992576|519000000|8|97|98|99|1|0|8|100|13|101|-1|-279|8|15|10|5|6|7|0|2|8|81|10|5|6|7|0|1|8|16|13|82|-293|-295|-42|10|5|6|7|0|5|8|83|28|-48|8|55|28|-48|8|102|28|-48|8|103|28|-48|8|104|28|-48|-293|-42|-1|-291|8|105|10|5|6|7|0|12|8|106|13|107|8|57|13|108|8|66|10|5|6|7|0|5|8|67|13|68|8|69|70|71|1854996146|1314259992576|722000000|8|72|13|68|8|73|70|71|1856244989|1314259992576|565000000|8|74|13|109|-316|-322|8|110|13|111|8|112|10|5|6|7|0|1|11|-8|13|58|-316|-340|8|113|13|114|8|115|10|5|6|7|0|0|-316|-348|8|116|10|5|6|7|0|0|-316|-352|8|117|13|118|8|79|13|119|8|27|13|120|8|121|13|122|-1|-314|8|123|10|5|6|7|0|1|8|124|10|5|6|7|0|0|-366|-368|-1|-364|8|50|10|5|6|7|0|3|-38|10|5|6|7|0|1|-42|10|5|6|7|0|2|8|51|28|-48|8|52|28|-48|-377|-42|-374|-38|-382|13|125|-384|54|12|2|-1|-372|-308|13|126|-310|28|-221|-312|28|-221|0|0|\"\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/CreditCourseProposalRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"#{contents1}#{contents2}\"\n },\n {\n 'subst' => 'true',\n :dyn_variables => [\n {\"name\" => opts[:enroll_est_var_name], \"re\" => opts[:enroll_est_var_regexp]},\n {\"name\" => opts[:lab_var_name], \"re\" => opts[:lab_var_regexp]},\n {\"name\" => opts[:atp_duration_week_var_name], \"re\" => opts[:atp_duration_week_var_regexp]}\n ]\n }\n )\n\n #@request.add(\"DEBUG/enroll_est_var_name/%%_#{opts[:enroll_est_var_name]}%%\", {}, {'subst' => 'true'})\n #@request.add(\"DEBUG/lab_var_name/%%_#{opts[:lab_var_name]}%%\", {}, {'subst' => 'true'})\n #@request.add(\"DEBUG/atp_duration_week_var_name/%%_#{opts[:atp_duration_week_var_name]}%%\", {}, {'subst' => 'true'})\n\n @request.add_thinktime(2)\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/WorkflowRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|71417C94A72A0CF76A43A2B36B8E3E1B|org.kuali.student.core.workflow.ui.client.service.WorkflowRpcService|getActionsRequested|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/WorkflowRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|71417C94A72A0CF76A43A2B36B8E3E1B|org.kuali.student.core.workflow.ui.client.service.WorkflowRpcService|getDocumentStatus|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/ProposalRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|12BDE6C2DA6A7CF74BE0FBF074E806E1|org.kuali.student.core.proposal.ui.client.service.ProposalRpcService|getProposalByWorkflowId|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n \n \n \n #\n # Learning Objectives\n #\n\n @request.add_thinktime(5)\n\n # Category\n for i in 1..opts[:lo_category].length\n itr = i-1\n if(i == opts[:lo_category].length)\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|11|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|search|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.lang.Boolean/476441737|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|lo.queryParam.loOptionalCategoryName|#{opts[:lo_category][0..itr]}|lo.search.loCategories|1|2|3|4|1|5|5|0|6|0|7|1|8|9|0|10|11|0|0|0|\"\n },\n {\n :dyn_variables => [\n {\"name\" => opts[:lo_category_id_var_name], \"re\" => opts[:lo_category_id_var_regexp]}\n ]\n }\n )\n\n #@request.add(\"DEBUG/lo_category_id_var_name/%%_#{opts[:lo_category_id_var_name]}%%\", {}, {'subst' => 'true'})\n\n else\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|11|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|search|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.lang.Boolean/476441737|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|lo.queryParam.loOptionalCategoryName|#{opts[:lo_category][0..itr]}|lo.search.loCategories|1|2|3|4|1|5|5|0|6|0|7|1|8|9|0|10|11|0|0|0|\"\n } \n )\n end \n end\n\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/LoCategoryRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|AC3B9DCF992DD862E331BCB0704203E2|org.kuali.student.lum.common.client.lo.rpc.LoCategoryRpcService|getData|java.lang.String/2004016611|%%_#{opts[:lo_category_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {'subst' => 'true'}\n )\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/LoCategoryRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|AC3B9DCF992DD862E331BCB0704203E2|org.kuali.student.lum.common.client.lo.rpc.LoCategoryRpcService|getLoCategoryType|java.lang.String/2004016611|loCategoryType.subject|1|2|3|4|1|5|6|\"\n }\n )\n\n @request.add_thinktime(25)\n\n # Save & Continue\n contents1 = \"5|0|155|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|A239E8C5A2EDCD8BCE6061BF191A8095|org.kuali.student.lum.lu.ui.course.client.service.CreditCourseProposalRpcService|saveData|org.kuali.student.common.assembly.data.Data/3184510345|org.kuali.student.common.assembly.data.Data|java.util.LinkedHashMap/1551059846|org.kuali.student.common.assembly.data.Data$StringKey/758802082|passFail|org.kuali.student.common.assembly.data.Data$BooleanValue/4261226833|java.lang.Boolean/476441737|audit|finalExamStatus|org.kuali.student.common.assembly.data.Data$StringValue/3151113388|STD|campusLocations|org.kuali.student.common.assembly.data.Data$DataValue/1692468409|org.kuali.student.common.assembly.data.Data$IntegerKey/134469241|java.lang.Integer/3438268394|AL|_runtimeData|id-translation|All|code|#{opts[:subject_area]}#{opts[:course_suffix]}|courseNumberSuffix|#{opts[:course_suffix]}|courseSpecificLOs|loInfo|id|desc|formatted|#{opts[:lo_cat_text]}|plain|name|SINGLE USE LO|sequence|0|metaInfo|loCategoryInfoList|%%_#{opts[:lo_category_id_var_name]}%%|#{opts[:lo_category]}|loRepository|kuali.loRepository.key.singleUse|effectiveDate|org.kuali.student.common.assembly.data.Data$DateValue/2929953165|expirationDate|state|active|type|loCategoryType.subject|createId|admin|createTime|java.sql.Timestamp/1769758459|updateId|updateTime|versionInd|courseTitle|#{opts[:course_title]}|creditOptions|fixedCreditValue|10.0|kuali.creditType.credit.degree.10.0|#{opts[:propose_person]}|resultValues|draft|kuali.resultComponentType.credit.degree.fixed|Credits, Fixed|crossListings|descr|#{opts[:course_description]}|duration|atpDurationTypeKey|kuali.atp.duration.Semester|timeQuantity|org.kuali.student.common.assembly.data.Data$IntegerValue/3605481012|Semester|expenditure|affiliatedOrgs|fees|formats|activities|activityType|kuali.lu.type.activity.Lab|contactHours|unitQuantity|10|unitType|kuali.atp.duration.week|per week|defaultEnrollmentEstimate|kuali.atp.duration.Week|Week|%%_#{opts[:atp_duration_week_var_name]}%%|unitsContentOwner|Lab|%%_#{opts[:lab_var_name]}%%|termsOffered|kuali.lu.type.CreditCourseFormatShell|gradingOptions|kuali.resultComponent.grade.letter|Letter|%%_#{opts[:clu_ref_dyn_var_name]}%%|instructors|personId|#{opts[:instructor]}|#{opts[:instructor]}, #{opts[:instructor]}(#{opts[:instructor]})|joints|level|100|2|pilotCourse|revenues|specialTopicsCourse|subjectArea|#{opts[:subject_area]}|kuali.atp.season.Any|Any|transcriptTitle|#{opts[:course_short_title]}|kuali.lu.type.CreditCourse|#{opts[:oversight_dept_number]}|#{oversight_department}|unitsDeployment|%%_#{opts[:admin_dep_var_name]}%%|#{admin_org}|variations|versionInfo|currentVersionStart|sequenceNumber|org.kuali.student.common.assembly.data.Data$LongValue/3784756947|java.lang.Long/4227064769|versionIndId|%%_#{opts[:version_ind_id_name]}%%|Standard final Exam|proposal|workflowNode|PreRoute|%%_#{opts[:proposal_dyn_var_name]}%%|4|#{opts[:proposal_title]}|proposalReference|proposalReferenceType|kuali.proposal.referenceType.clu|proposerOrg|proposerPerson|rationale|#{opts[:course_rationale]}|Saved|kuali.proposal.type.course.create|workflowId|%%_#{opts[:proposal_doc_id_var_name]}%%|collaboratorInfo|collaborators\"\n contents2 = \"|1|2|3|4|1|5|5|6|7|0|36|8|9|10|11|0|8|12|10|-5|8|13|14|15|8|16|17|5|6|7|0|2|18|19|0|14|20|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|23|-19|-21|-12|-17|-1|-10|8|24|14|25|8|26|14|27|8|28|17|5|6|7|0|1|18|-15|17|5|6|7|0|2|8|29|17|5|6|7|0|5|8|30|14|0|8|31|17|5|6|7|0|2|8|32|14|33|8|34|14|33|-41|-45|8|35|14|36|8|37|14|38|8|39|17|0|-37|-39|8|40|17|5|6|7|0|1|18|-15|17|5|6|7|0|9|8|30|14|41|8|35|14|42|8|31|17|5|6|7|0|2|8|32|14|0|8|34|14|0|-65|-71|8|43|14|44|8|45|46|0|8|47|46|0|8|48|14|49|8|50|14|51|8|39|17|5|6|7|0|5|8|52|14|53|8|54|46|55|3759152200|1288490188800|0|8|56|14|53|8|57|46|55|3759152200|1288490188800|0|8|58|14|38|-65|-89|-61|-63|-37|-59|-33|-35|-1|8|28|8|59|14|60|8|61|17|5|6|7|0|1|18|-15|17|5|6|7|0|7|8|62|14|63|8|30|14|64|8|39|17|5|6|7|0|5|8|52|14|65|8|54|46|55|1861297882|1314259992576|458000000|8|56|14|65|8|57|46|55|1861297882|1314259992576|458000000|8|58|14|38|-114|-120|8|66|17|5|6|7|0|1|18|-15|14|63|-114|-136|8|48|14|67|8|50|14|68|8|21|17|5|6|7|0|1|8|50|17|5|6|7|0|1|8|22|14|69|-148|-150|-114|-146|-110|-112|-1|-108|8|70|17|5|6|7|0|0|-1|-156|8|71|17|5|6|7|0|1|8|34|14|72|-1|-160|8|73|17|5|6|7|0|3|8|74|14|75|8|76|77|19|2|8|21|17|5|6|7|0|1|8|74|17|5|6|7|0|1|8|22|14|78|-177|-179|-168|-175|-1|-166|8|79|17|5|6|7|0|1|8|80|17|5|6|7|0|0|-187|-189|-1|-185|8|81|17|5|6|7|0|0|-1|-193|8|82|17|5|6|7|0|1|18|-15|17|5|6|7|0|6|8|83|17|5|6|7|0|1|18|-15|17|5|6|7|0|9|8|84|14|85|8|86|17|5|6|7|0|3|8|87|14|88|8|89|14|90|8|21|17|5|6|7|0|1|8|89|17|5|6|7|0|1|8|22|14|91|-225|-227|-217|-223|-211|-215|8|92|77|19|100|8|73|17|5|6|7|0|3|8|74|14|93|8|76|77|19|13|8|21|17|5|6|7|0|1|8|74|17|5|6|7|0|1|8|22|14|94|-247|-249|-238|-245|-211|-236|8|30|14|95|8|39|17|5|6|7|0|5|8|52|14|65|8|54|46|55|1861297868|1314259992576|444000000|8|56|14|65|8|57|46|55|1861297868|1314259992576|444000000|8|58|14|38|-211|-257|8|48|14|67|8|96|17|5|6|7|0|0|-211|-275|8|21|17|5|6|7|0|1|8|84|17|5|6|7|0|1|8|22|14|97|-281|-283|-211|-279|-207|-209|-203|-205|8|30|14|98|8|39|17|5|6|7|0|5|8|52|14|65|8|54|46|55|1861297859|1314259992576|435000000|8|56|14|65|8|57|46|55|1861297859|1314259992576|435000000|8|58|14|38|-203|-291|8|48|14|67|8|99|17|5|6|7|0|0|-203|-309|8|50|14|100|-199|-201|-1|-197|8|101|17|5|6|7|0|2|18|-15|14|102|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|103|-323|-325|-317|-321|-1|-315|8|30|14|104|8|105|17|5|6|7|0|1|18|-15|17|5|6|7|0|2|8|106|14|107|8|21|17|5|6|7|0|1|8|106|17|5|6|7|0|1|8|22|14|108|-345|-347|-339|-343|-335|-337|-1|-333|8|109|17|5|6|7|0|0|-1|-353|8|110|14|111|8|39|17|5|6|7|0|5|8|52|14|65|8|54|46|55|1854995943|1314259992576|519000000|8|56|14|65|8|57|46|55|1861297799|1314259992576|375000000|8|58|14|112|-1|-359|8|113|10|-5|8|114|17|5|6|7|0|0|-1|-377|8|115|10|-5|8|48|14|67|8|116|14|117|8|99|17|5|6|7|0|2|18|-15|14|118|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|119|-395|-397|-389|-393|-1|-387|8|120|14|121|8|50|14|122|8|96|17|5|6|7|0|2|18|-15|14|123|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|124|-415|-417|-409|-413|-1|-407|8|125|17|5|6|7|0|2|18|-15|14|126|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|127|-431|-433|-425|-429|-1|-423|8|128|17|5|6|7|0|0|-1|-439|8|129|17|5|6|7|0|3|8|130|46|55|1854995943|1314259992576|519000000|8|131|132|133|1|0|8|134|14|135|-1|-443|8|21|17|5|6|7|0|2|8|116|17|5|6|7|0|1|8|22|14|117|-457|-459|8|13|17|5|6|7|0|1|8|22|14|136|-457|-465|-1|-455|8|137|17|5|6|7|0|12|8|138|14|139|8|30|14|140|8|39|17|5|6|7|0|5|8|52|14|65|8|54|46|55|1854996146|1314259992576|722000000|8|56|14|65|8|57|46|55|1861299120|1314259992576|696000000|8|58|14|141|-473|-479|8|35|14|142|8|143|17|5|6|7|0|1|18|-15|14|104|-473|-497|8|144|14|145|8|146|17|5|6|7|0|0|-473|-505|8|147|17|5|6|7|0|0|-473|-509|8|148|14|149|8|48|14|150|8|50|14|151|8|152|14|153|-1|-471|8|154|17|5|6|7|0|1|8|155|17|5|6|7|0|0|-523|-525|-1|-521|0|0|\"\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/CreditCourseProposalRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"#{contents1}#{contents2}\"\n },\n {\n 'subst' => 'true',\n :dyn_variables => [\n {\"name\" => opts[:lo_category_var_name], \"re\" => opts[:lo_category_var_regexp]}\n ]\n }\n )\n\n #@request.add(\"DEBUG/lo_category_var_name/%%_#{opts[:lo_category_var_name]}%%\", {}, {'subst' => 'true'})\n\n @request.add_thinktime(2)\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/LoCategoryRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|4|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|AC3B9DCF992DD862E331BCB0704203E2|org.kuali.student.lum.common.client.lo.rpc.LoCategoryRpcService|getLoCategoryTypes|1|2|3|4|0|\"\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/WorkflowRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|71417C94A72A0CF76A43A2B36B8E3E1B|org.kuali.student.core.workflow.ui.client.service.WorkflowRpcService|getActionsRequested|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/WorkflowRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|71417C94A72A0CF76A43A2B36B8E3E1B|org.kuali.student.core.workflow.ui.client.service.WorkflowRpcService|getDocumentStatus|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/ProposalRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|12BDE6C2DA6A7CF74BE0FBF074E806E1|org.kuali.student.core.proposal.ui.client.service.ProposalRpcService|getProposalByWorkflowId|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/statementRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|335FF062A700107AB2A642B325C6C5C5|org.kuali.student.lum.program.client.rpc.StatementRpcService|getStatementTypesForStatementTypeForCourse|java.lang.String/2004016611|kuali.statement.type.course|1|2|3|4|1|5|6|\"\n }\n )\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/CourseRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|8|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|3C9BBAD14113E13A72476EEE8100687B|org.kuali.student.lum.lu.ui.course.client.service.CourseRpcService|getCourseStatements|java.lang.String/2004016611|%%_#{opts[:clu_ref_dyn_var_name]}%%|KUALI.RULE|en|1|2|3|4|3|5|5|5|6|7|8|\"\n },\n {\n 'subst' => 'true'\n }\n )\n \n \n #\n # Course Requisites\n # Save without editing anything\n\n @request.add_thinktime(5)\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/CourseRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|10|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|3C9BBAD14113E13A72476EEE8100687B|org.kuali.student.lum.lu.ui.course.client.service.CourseRpcService|storeCourseStatements|java.lang.String/2004016611|java.util.Map|%%_#{opts[:clu_ref_dyn_var_name]}%%|draft|java.util.HashMap/962170901|java.util.LinkedHashMap/1551059846|1|2|3|4|4|5|5|6|6|7|8|9|0|10|0|0|\"\n },\n {\n 'subst' => 'true'\n }\n )\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/statementRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|335FF062A700107AB2A642B325C6C5C5|org.kuali.student.lum.program.client.rpc.StatementRpcService|getStatementTypesForStatementTypeForCourse|java.lang.String/2004016611|kuali.statement.type.course|1|2|3|4|1|5|6|\"\n }\n )\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/CourseRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|8|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|3C9BBAD14113E13A72476EEE8100687B|org.kuali.student.lum.lu.ui.course.client.service.CourseRpcService|getCourseStatements|java.lang.String/2004016611|%%_#{opts[:clu_ref_dyn_var_name]}%%|KUALI.RULE|en|1|2|3|4|3|5|5|5|6|7|8|\"\n },\n {\n 'subst' => 'true'\n }\n )\n \n \n #\n # Active Dates\n # Start = Fall Sem 2008\n #\n\n @request.add_thinktime(10)\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|20|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|atp.advancedAtpSearchParam.atpType|java.lang.String/2004016611|kuali.atp.type.Spring|kuali.atp.type.Summer|kuali.atp.type.Fall|kuali.atp.type.Session1|kuali.atp.type.Session2|kuali.atp.type.Mini-mester1A|kuali.atp.type.Mini-mester1B|atp.advancedAtpSearchParam.atpStartDateAtpConstraintId|kuali.atp.FA2008-2009|atp.search.advancedAtpSearch|atp.resultColumn.atpStartDate|1|2|3|4|1|5|5|0|0|6|2|7|8|6|7|9|10|9|11|9|12|9|13|9|14|9|15|9|16|0|7|17|0|18|19|20|0|0|\"\n }\n )\n\n # Save & Continue\n contents1 = \"5|0|162|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|A239E8C5A2EDCD8BCE6061BF191A8095|org.kuali.student.lum.lu.ui.course.client.service.CreditCourseProposalRpcService|saveData|org.kuali.student.common.assembly.data.Data/3184510345|org.kuali.student.common.assembly.data.Data|java.util.LinkedHashMap/1551059846|org.kuali.student.common.assembly.data.Data$StringKey/758802082|passFail|org.kuali.student.common.assembly.data.Data$BooleanValue/4261226833|java.lang.Boolean/476441737|audit|finalExamStatus|org.kuali.student.common.assembly.data.Data$StringValue/3151113388|STD|campusLocations|org.kuali.student.common.assembly.data.Data$DataValue/1692468409|org.kuali.student.common.assembly.data.Data$IntegerKey/134469241|java.lang.Integer/3438268394|AL|_runtimeData|id-translation|All|code|#{opts[:subject_area]}#{opts[:course_suffix]}|courseNumberSuffix|#{opts[:course_suffix]}|courseSpecificLOs|loCategoryInfoList|id|%%_#{opts[:lo_category_id_var_name]}%%|loRepository|kuali.loRepository.key.singleUse|metaInfo|createId|admin|createTime|org.kuali.student.common.assembly.data.Data$DateValue/2929953165|java.sql.Timestamp/1769758459|updateId|updateTime|versionInd|0|name|#{opts[:lo_category]}|state|active|type|loCategoryType.subject|loDisplayInfoList|loInfo|sequence|desc|formatted|#{opts[:lo_cat_text]}|plain|%%_#{opts[:lo_category_var_name]}%%|loRepositoryKey|#{opts[:propose_person]}|SINGLE USE LO|draft|kuali.lo.type.singleUse|courseTitle|#{opts[:course_title]}|creditOptions|fixedCreditValue|10.0|kuali.creditType.credit.degree.10.0|resultValues|kuali.resultComponentType.credit.degree.fixed|Credits, Fixed|crossListings|descr|#{opts[:course_description]}|duration|atpDurationTypeKey|kuali.atp.duration.Semester|timeQuantity|org.kuali.student.common.assembly.data.Data$IntegerValue/3605481012|Semester|expenditure|affiliatedOrgs|fees|formats|activities|activityType|kuali.lu.type.activity.Lab|contactHours|unitQuantity|10|unitType|kuali.atp.duration.week|per week|defaultEnrollmentEstimate|kuali.atp.duration.Week|Week|%%_#{opts[:atp_duration_week_var_name]}%%|1|unitsContentOwner|Lab|%%_#{opts[:lab_var_name]}%%|termsOffered|kuali.lu.type.CreditCourseFormatShell|gradingOptions|kuali.resultComponent.grade.letter|Letter|%%_#{opts[:clu_ref_dyn_var_name]}%%|instructors|personId|#{opts[:instructor]}|#{opts[:instructor]}, #{opts[:instructor]}(#{opts[:instructor]})|joints|level|100|3|pilotCourse|revenues|specialTopicsCourse|subjectArea|#{opts[:subject_area]}|kuali.atp.season.Any|Any|transcriptTitle|#{opts[:course_short_title]}|kuali.lu.type.CreditCourse|#{opts[:oversight_dept_number]}|#{oversight_department}|unitsDeployment|%%_#{opts[:admin_dep_var_name]}%%|#{admin_org}|variations|versionInfo|currentVersionStart|sequenceNumber|org.kuali.student.common.assembly.data.Data$LongValue/3784756947|java.lang.Long/4227064769|versionIndId|%%_#{opts[:version_ind_id_name]}%%|Standard final Exam|dirty|startTerm|endTerm|proposal|workflowNode|PreRoute|%%_#{opts[:proposal_dyn_var_name]}%%|5|#{opts[:proposal_title]}|proposalReference|proposalReferenceType|kuali.proposal.referenceType.clu|proposerOrg|proposerPerson|rationale|#{opts[:course_rationale]}|Saved|kuali.proposal.type.course.create|workflowId|%%_#{opts[:proposal_doc_id_var_name]}%%|collaboratorInfo|collaborators|kuali.atp.FA2008-2009\"\n contents2 = \"|1|2|3|4|1|5|5|6|7|0|38|8|9|10|11|0|8|12|10|-5|8|13|14|15|8|16|17|5|6|7|0|2|18|19|0|14|20|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|23|-19|-21|-12|-17|-1|-10|8|24|14|25|8|26|14|27|8|28|17|5|6|7|0|1|18|-15|17|5|6|7|0|3|8|29|17|5|6|7|0|1|18|-15|17|5|6|7|0|6|8|30|14|31|8|32|14|33|8|34|17|5|6|7|0|5|8|35|14|36|8|37|38|39|3759152200|1288490188800|0|8|40|14|36|8|41|38|39|3759152200|1288490188800|0|8|42|14|43|-45|-51|8|44|14|45|8|46|14|47|8|48|14|49|-41|-43|-37|-39|8|50|17|5|6|7|0|0|-37|-73|8|51|17|5|6|7|0|8|8|52|14|43|8|53|17|5|6|7|0|2|8|54|14|55|8|56|14|55|-79|-83|8|30|14|57|8|58|14|33|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|1864356667|1314259992576|243000000|8|40|14|59|8|41|38|39|1864356667|1314259992576|243000000|8|42|14|43|-79|-95|8|44|14|60|8|46|14|61|8|48|14|62|-37|-77|-33|-35|-1|-31|8|63|14|64|8|65|17|5|6|7|0|1|18|-15|17|5|6|7|0|7|8|66|14|67|8|30|14|68|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|1861297882|1314259992576|458000000|8|40|14|59|8|41|38|39|1861297882|1314259992576|458000000|8|42|14|43|-125|-131|8|69|17|5|6|7|0|1|18|-15|14|67|-125|-147|8|46|14|61|8|48|14|70|8|21|17|5|6|7|0|1|8|48|17|5|6|7|0|1|8|22|14|71|-159|-161|-125|-157|-121|-123|-1|-119|8|72|17|5|6|7|0|0|-1|-167|8|73|17|5|6|7|0|1|8|56|14|74|-1|-171|8|75|17|5|6|7|0|3|8|76|14|77|8|78|79|19|2|8|21|17|5|6|7|0|1|8|76|17|5|6|7|0|1|8|22|14|80|-188|-190|-179|-186|-1|-177|8|81|17|5|6|7|0|1|8|82|17|5|6|7|0|0|-198|-200|-1|-196|8|83|17|5|6|7|0|0|-1|-204|8|84|17|5|6|7|0|1|18|-15|17|5|6|7|0|6|8|85|17|5|6|7|0|1|18|-15|17|5|6|7|0|9|8|86|14|87|8|88|17|5|6|7|0|3|8|89|14|90|8|91|14|92|8|21|17|5|6|7|0|1|8|91|17|5|6|7|0|1|8|22|14|93|-236|-238|-228|-234|-222|-226|8|94|79|19|100|8|75|17|5|6|7|0|3|8|76|14|95|8|78|79|19|13|8|21|17|5|6|7|0|1|8|76|17|5|6|7|0|1|8|22|14|96|-258|-260|-249|-256|-222|-247|8|30|14|97|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|1861297868|1314259992576|444000000|8|40|14|59|8|41|38|39|1864356631|1314259992576|207000000|8|42|14|98|-222|-268|8|46|14|61|8|99|17|5|6|7|0|0|-222|-286|8|21|17|5|6|7|0|1|8|86|17|5|6|7|0|1|8|22|14|100|-292|-294|-222|-290|-218|-220|-214|-216|8|30|14|101|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|1861297859|1314259992576|435000000|8|40|14|59|8|41|38|39|1864356620|1314259992576|196000000|8|42|14|98|-214|-302|8|46|14|61|8|102|17|5|6|7|0|0|-214|-320|8|48|14|103|-210|-212|-1|-208|8|104|17|5|6|7|0|2|18|-15|14|105|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|106|-334|-336|-328|-332|-1|-326|8|30|14|107|8|108|17|5|6|7|0|1|18|-15|17|5|6|7|0|2|8|109|14|110|8|21|17|5|6|7|0|1|8|109|17|5|6|7|0|1|8|22|14|111|-356|-358|-350|-354|-346|-348|-1|-344|8|112|17|5|6|7|0|0|-1|-364|8|113|14|114|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|1854995943|1314259992576|519000000|8|40|14|59|8|41|38|39|1864356597|1314259992576|173000000|8|42|14|115|-1|-370|8|116|10|-5|8|117|17|5|6|7|0|0|-1|-388|8|118|10|-5|8|46|14|61|8|119|14|120|8|102|17|5|6|7|0|2|18|-15|14|121|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|122|-406|-408|-400|-404|-1|-398|8|123|14|124|8|48|14|125|8|99|17|5|6|7|0|2|18|-15|14|126|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|127|-426|-428|-420|-424|-1|-418|8|128|17|5|6|7|0|2|18|-15|14|129|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|130|-442|-444|-436|-440|-1|-434|8|131|17|5|6|7|0|0|-1|-450|8|132|17|5|6|7|0|3|8|133|38|39|1854995943|1314259992576|519000000|8|134|135|136|1|0|8|137|14|138|-1|-454|8|21|17|5|6|7|0|3|8|119|17|5|6|7|0|1|8|22|14|120|-468|-470|8|13|17|5|6|7|0|1|8|22|14|139|-468|-476|8|140|17|5|6|7|0|2|8|141|10|11|1|8|142|10|-488|-468|-482|-1|-466|8|143|17|5|6|7|0|12|8|144|14|145|8|30|14|146|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|1854996146|1314259992576|722000000|8|40|14|59|8|41|38|39|1864357928|1314259992576|504000000|8|42|14|147|-493|-499|8|44|14|148|8|149|17|5|6|7|0|1|18|-15|14|107|-493|-517|8|150|14|151|8|152|17|5|6|7|0|0|-493|-525|8|153|17|5|6|7|0|0|-493|-529|8|154|14|155|8|46|14|156|8|48|14|157|8|158|14|159|-1|-491|8|160|17|5|6|7|0|1|8|161|17|5|6|7|0|0|-543|-545|-1|-541|-486|14|162|-489|14|0|0|0|\"\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/CreditCourseProposalRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"#{contents1}#{contents2}\"\n },\n {\n 'subst' => 'true'\n }\n )\n \n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/WorkflowRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|71417C94A72A0CF76A43A2B36B8E3E1B|org.kuali.student.core.workflow.ui.client.service.WorkflowRpcService|getActionsRequested|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/WorkflowRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|71417C94A72A0CF76A43A2B36B8E3E1B|org.kuali.student.core.workflow.ui.client.service.WorkflowRpcService|getDocumentStatus|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/ProposalRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|12BDE6C2DA6A7CF74BE0FBF074E806E1|org.kuali.student.core.proposal.ui.client.service.ProposalRpcService|getProposalByWorkflowId|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n \n \n \n #\n # Financials\n # $100 lab fee\n # Admin org 100% rev and exp\n #\n\n @request.add_thinktime(5)\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|11|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|cachingSearch|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|enumeration.queryParam.enumerationType|kuali.lu.fee.feeType|enumeration.management.search|enumeration.resultColumn.sortKey|1|2|3|4|1|5|5|0|0|6|1|7|8|0|9|10|11|0|0|\"\n }\n )\n\n # Revenue - Admin Org\n for i in 1..admin_org.length\n itr = i-1\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|16|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|search|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.lang.Boolean/476441737|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|org.queryParam.orgOptionalLongName|#{admin_org[0..itr]}|org.queryParam.orgOptionalType|java.lang.String/2004016611|kuali.org.Department|kuali.org.College|org.search.generic||1|2|3|4|1|5|5|0|6|0|7|2|8|9|0|10|8|11|7|2|12|13|12|14|0|15|16|0|0|\"\n } \n ) \n end\n\n @request.add_thinktime(5)\n\n # Expense - Admin Org\n for i in 1..admin_org.length\n itr = i-1\n if(i == admin_org.length)\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|16|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|search|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.lang.Boolean/476441737|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|org.queryParam.orgOptionalLongName|#{admin_org[0..itr]}|org.queryParam.orgOptionalType|java.lang.String/2004016611|kuali.org.Department|kuali.org.College|org.search.generic||1|2|3|4|1|5|5|0|6|0|7|2|8|9|0|10|8|11|7|2|12|13|12|14|0|15|16|0|0|\"\n },\n {\n :dyn_variables => [\n {\"name\" => opts[:admin_dep_var_name], \"re\" => opts[:admin_dep_var_regexp]}\n ]\n } \n )\n else\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|16|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|search|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.lang.Boolean/476441737|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|org.queryParam.orgOptionalLongName|#{admin_org[0..itr]}|org.queryParam.orgOptionalType|java.lang.String/2004016611|kuali.org.Department|kuali.org.College|org.search.generic||1|2|3|4|1|5|5|0|6|0|7|2|8|9|0|10|8|11|7|2|12|13|12|14|0|15|16|0|0|\"\n } \n )\n end \n end\n\n #@request.add(\"DEBUG/admin_dep_var_name/%%_#{opts[:admin_dep_var_name]}%%\", {}, {'subst' => 'true'})\n\n @request.add_thinktime(20)\n\n # Save & Continue\n contents1 = \"5|0|173|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|A239E8C5A2EDCD8BCE6061BF191A8095|org.kuali.student.lum.lu.ui.course.client.service.CreditCourseProposalRpcService|saveData|org.kuali.student.common.assembly.data.Data/3184510345|org.kuali.student.common.assembly.data.Data|java.util.LinkedHashMap/1551059846|org.kuali.student.common.assembly.data.Data$StringKey/758802082|passFail|org.kuali.student.common.assembly.data.Data$BooleanValue/4261226833|java.lang.Boolean/476441737|audit|finalExamStatus|org.kuali.student.common.assembly.data.Data$StringValue/3151113388|STD|campusLocations|org.kuali.student.common.assembly.data.Data$DataValue/1692468409|org.kuali.student.common.assembly.data.Data$IntegerKey/134469241|java.lang.Integer/3438268394|AL|_runtimeData|id-translation|All|code|#{opts[:subject_area]}#{opts[:course_suffix]}|courseNumberSuffix|#{opts[:course_suffix]}|courseSpecificLOs|loCategoryInfoList|id|%%_#{opts[:lo_category_id_var_name]}%%|loRepository|kuali.loRepository.key.singleUse|metaInfo|createId|admin|createTime|org.kuali.student.common.assembly.data.Data$DateValue/2929953165|java.sql.Timestamp/1769758459|updateId|updateTime|versionInd|0|name|#{opts[:lo_category]}|state|active|type|loCategoryType.subject|loDisplayInfoList|loInfo|sequence|desc|formatted|#{opts[:lo_cat_text]}|plain|%%_#{opts[:lo_category_var_name]}%%|loRepositoryKey|#{opts[:propose_person]}|1|SINGLE USE LO|draft|kuali.lo.type.singleUse|courseTitle|#{opts[:course_title]}|creditOptions|fixedCreditValue|10.0|kuali.creditType.credit.degree.10.0|resultValues|kuali.resultComponentType.credit.degree.fixed|Credits, Fixed|crossListings|descr|#{opts[:course_description]}|duration|atpDurationTypeKey|kuali.atp.duration.Semester|timeQuantity|org.kuali.student.common.assembly.data.Data$IntegerValue/3605481012|Semester|effectiveDate|expenditure|affiliatedOrgs|dirty|orgId|percentage|created|%%_#{opts[:admin_dep_var_name]}%%|org.kuali.student.common.assembly.data.Data$LongValue/3784756947|java.lang.Long/4227064769|fees|feeType|rateType|kuali.enum.type.feeTypes.labFee|fixedRateFee|feeAmounts|currencyQuantity|formats|activities|activityType|kuali.lu.type.activity.Lab|contactHours|unitQuantity|10|unitType|kuali.atp.duration.week|per week|defaultEnrollmentEstimate|kuali.atp.duration.Week|Week|%%_#{opts[:atp_duration_week_var_name]}%%|2|unitsContentOwner|Lab|%%_#{opts[:lab_var_name]}%%|termsOffered|kuali.lu.type.CreditCourseFormatShell|gradingOptions|kuali.resultComponent.grade.letter|Letter|%%_#{opts[:clu_ref_dyn_var_name]}%%|instructors|personId|#{opts[:instructor]}|#{opts[:instructor]}, #{opts[:instructor]}(#{opts[:instructor]})|joints|level|100|4|pilotCourse|revenues|specialTopicsCourse|startTerm|kuali.atp.FA2008-2009|subjectArea|#{opts[:subject_area]}|kuali.atp.season.Any|Any|transcriptTitle|#{opts[:course_short_title]}|kuali.lu.type.CreditCourse|#{opts[:oversight_dept_number]}|#{oversight_department}|unitsDeployment|#{admin_org}|variations|versionInfo|currentVersionStart|sequenceNumber|versionIndId|%%_#{opts[:version_ind_id_name]}%%|Fall Semester of 2008|Standard final Exam|proposal|workflowNode|PreRoute|%%_#{opts[:proposal_dyn_var_name]}%%|6|#{opts[:proposal_title]}|proposalReference|proposalReferenceType|kuali.proposal.referenceType.clu|proposerOrg|proposerPerson|rationale|#{opts[:course_rationale]}|Saved|kuali.proposal.type.course.create|workflowId|%%_#{opts[:proposal_doc_id_var_name]}%%|collaboratorInfo|collaborators\"\n contents2 = \"|1|2|3|4|1|5|5|6|7|0|38|8|9|10|11|0|8|12|10|-5|8|13|14|15|8|16|17|5|6|7|0|2|18|19|0|14|20|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|23|-19|-21|-12|-17|-1|-10|8|24|14|25|8|26|14|27|8|28|17|5|6|7|0|1|18|-15|17|5|6|7|0|3|8|29|17|5|6|7|0|1|18|-15|17|5|6|7|0|6|8|30|14|31|8|32|14|33|8|34|17|5|6|7|0|5|8|35|14|36|8|37|38|39|3759152200|1288490188800|0|8|40|14|36|8|41|38|39|3759152200|1288490188800|0|8|42|14|43|-45|-51|8|44|14|45|8|46|14|47|8|48|14|49|-41|-43|-37|-39|8|50|17|5|6|7|0|0|-37|-73|8|51|17|5|6|7|0|8|8|52|14|43|8|53|17|5|6|7|0|2|8|54|14|55|8|56|14|55|-79|-83|8|30|14|57|8|58|14|33|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|1864356667|1314259992576|243000000|8|40|14|59|8|41|38|39|1866434943|1314259992576|519000000|8|42|14|60|-79|-95|8|44|14|61|8|46|14|62|8|48|14|63|-37|-77|-33|-35|-1|-31|8|64|14|65|8|66|17|5|6|7|0|1|18|-15|17|5|6|7|0|7|8|67|14|68|8|30|14|69|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|1861297882|1314259992576|458000000|8|40|14|59|8|41|38|39|1861297882|1314259992576|458000000|8|42|14|43|-125|-131|8|70|17|5|6|7|0|1|18|-15|14|68|-125|-147|8|46|14|62|8|48|14|71|8|21|17|5|6|7|0|1|8|48|17|5|6|7|0|1|8|22|14|72|-159|-161|-125|-157|-121|-123|-1|-119|8|73|17|5|6|7|0|0|-1|-167|8|74|17|5|6|7|0|1|8|56|14|75|-1|-171|8|76|17|5|6|7|0|3|8|77|14|78|8|79|80|19|2|8|21|17|5|6|7|0|1|8|77|17|5|6|7|0|1|8|22|14|81|-188|-190|-179|-186|-1|-177|8|82|38|39|470887936|1219770712064|0|8|83|17|5|6|7|0|1|8|84|17|5|6|7|0|1|18|-15|17|5|6|7|0|3|8|21|17|5|6|7|0|3|8|85|17|5|6|7|0|2|8|86|10|11|1|8|87|10|-221|-213|-215|8|88|10|-221|-211|17|5|6|7|0|1|-215|17|5|6|7|0|1|-224|10|-221|-227|-215|-213|-211|-209|-211|-219|14|89|-222|90|91|100|0|-205|-207|-201|-203|-1|-199|8|92|17|5|6|7|0|1|18|-15|17|5|6|7|0|4|-211|17|5|6|7|0|3|-215|17|5|6|7|0|2|8|93|10|-221|8|94|10|-221|-245|-215|8|88|10|-221|-211|17|5|6|7|0|1|-215|17|5|6|7|0|1|-254|10|-221|-257|-215|-245|-211|-242|-211|-250|14|95|-252|14|96|8|97|17|5|6|7|0|1|18|-15|17|5|6|7|0|2|8|98|80|19|100|-211|17|5|6|7|0|1|-215|17|5|6|7|0|1|-273|10|-221|-277|-215|-271|-211|-267|-269|-242|-265|-238|-240|-1|-236|8|99|17|5|6|7|0|1|18|-15|17|5|6|7|0|6|8|100|17|5|6|7|0|1|18|-15|17|5|6|7|0|9|8|101|14|102|8|103|17|5|6|7|0|3|8|104|14|105|8|106|14|107|8|21|17|5|6|7|0|1|8|106|17|5|6|7|0|1|8|22|14|108|-311|-313|-303|-309|-297|-301|8|109|80|-275|8|76|17|5|6|7|0|3|8|77|14|110|8|79|80|19|13|8|21|17|5|6|7|0|1|8|77|17|5|6|7|0|1|8|22|14|111|-332|-334|-323|-330|-297|-321|8|30|14|112|8\"\n contents3 = \"|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|1861297868|1314259992576|444000000|8|40|14|59|8|41|38|39|1866434894|1314259992576|470000000|8|42|14|113|-297|-342|8|46|14|62|8|114|17|5|6|7|0|0|-297|-360|8|21|17|5|6|7|0|1|8|101|17|5|6|7|0|1|8|22|14|115|-366|-368|-297|-364|-293|-295|-289|-291|8|30|14|116|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|1861297859|1314259992576|435000000|8|40|14|59|8|41|38|39|1866434884|1314259992576|460000000|8|42|14|113|-289|-376|8|46|14|62|8|117|17|5|6|7|0|0|-289|-394|8|48|14|118|-285|-287|-1|-283|8|119|17|5|6|7|0|2|18|-15|14|120|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|121|-408|-410|-402|-406|-1|-400|8|30|14|122|8|123|17|5|6|7|0|1|18|-15|17|5|6|7|0|2|8|124|14|125|8|21|17|5|6|7|0|1|8|124|17|5|6|7|0|1|8|22|14|126|-430|-432|-424|-428|-420|-422|-1|-418|8|127|17|5|6|7|0|0|-1|-438|8|128|14|129|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|1854995943|1314259992576|519000000|8|40|14|59|8|41|38|39|1866434860|1314259992576|436000000|8|42|14|130|-1|-444|8|131|10|-5|8|132|17|5|6|7|0|1|18|-15|17|5|6|7|0|2|8|84|17|5|6|7|0|1|18|-15|17|5|6|7|0|3|-211|17|5|6|7|0|1|-215|17|5|6|7|0|2|8|86|10|-221|8|87|10|-221|-479|-215|-476|-211|-484|14|89|-486|90|-235|-472|-474|-468|-470|8|21|17|5|6|7|0|2|8|88|10|-221|-211|17|5|6|7|0|1|-215|17|5|6|7|0|1|-494|10|-221|-497|-215|-492|-211|-468|-490|-464|-466|-1|-462|8|133|10|-5|8|134|14|135|8|46|14|62|8|136|14|137|8|117|17|5|6|7|0|2|18|-15|14|138|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|139|-519|-521|-513|-517|-1|-511|8|140|14|141|8|48|14|142|8|114|17|5|6|7|0|2|18|-15|14|143|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|144|-539|-541|-533|-537|-1|-531|8|145|17|5|6|7|0|2|18|-15|14|89|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|146|-555|-557|-549|-553|-1|-547|8|147|17|5|6|7|0|0|-1|-563|8|148|17|5|6|7|0|3|8|149|38|39|1854995943|1314259992576|519000000|8|150|90|91|1|0|8|151|14|152|-1|-567|8|21|17|5|6|7|0|3|8|134|17|5|6|7|0|1|8|22|14|153|-581|-583|8|136|17|5|6|7|0|1|8|22|14|137|-581|-589|8|13|17|5|6|7|0|1|8|22|14|154|-581|-595|-1|-579|8|155|17|5|6|7|0|12|8|156|14|157|8|30|14|158|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|1854996146|1314259992576|722000000|8|40|14|59|8|41|38|39|1866435719|1314259992576|295000000|8|42|14|159|-603|-609|8|44|14|160|8|161|17|5|6|7|0|1|18|-15|14|122|-603|-627|8|162|14|163|8|164|17|5|6|7|0|0|-603|-635|8|165|17|5|6|7|0|0|-603|-639|8|166|14|167|8|46|14|168|8|48|14|169|8|170|14|171|-1|-601|8|172|17|5|6|7|0|1|8|173|17|5|6|7|0|0|-653|-655|-1|-651|0|0|\"\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/CreditCourseProposalRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"#{contents1}#{contents2}#{contents3}\"\n },\n {\n 'subst' => 'true',\n :dyn_variables => [\n {\"name\" => opts[:affliated_orgs_id_name], \"re\" => opts[:affliated_orgs_id_regexp]},\n {\"name\" => opts[:lab_fee_id_name], \"re\" => opts[:lab_fee_id_regexp]},\n {\"name\" => opts[:revenues_id_name], \"re\" => opts[:revenues_id_regexp]},\n {\"name\" => opts[:revenue_id_name], \"re\" => opts[:revenue_id_regexp]}\n ]\n }\n )\n\n #@request.add(\"DEBUG/affliated_orgs_id_name/%%_#{opts[:affliated_orgs_id_name]}%%\", {}, {'subst' => 'true'})\n #@request.add(\"DEBUG/lab_fee_id_name/%%_#{opts[:lab_fee_id_name]}%%\", {}, {'subst' => 'true'})\n #@request.add(\"DEBUG/revenues_id_name/%%_#{opts[:revenues_id_name]}%%\", {}, {'subst' => 'true'})\n #@request.add(\"DEBUG/revenue_id_name/%%_#{opts[:revenue_id_name]}%%\", {}, {'subst' => 'true'})\n\n @request.add_thinktime(2)\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/WorkflowRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|71417C94A72A0CF76A43A2B36B8E3E1B|org.kuali.student.core.workflow.ui.client.service.WorkflowRpcService|getActionsRequested|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/WorkflowRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|71417C94A72A0CF76A43A2B36B8E3E1B|org.kuali.student.core.workflow.ui.client.service.WorkflowRpcService|getDocumentStatus|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/ProposalRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|12BDE6C2DA6A7CF74BE0FBF074E806E1|org.kuali.student.core.proposal.ui.client.service.ProposalRpcService|getProposalByWorkflowId|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/WorkflowRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|71417C94A72A0CF76A43A2B36B8E3E1B|org.kuali.student.core.workflow.ui.client.service.WorkflowRpcService|getDocumentStatus|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/WorkflowRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|71417C94A72A0CF76A43A2B36B8E3E1B|org.kuali.student.core.workflow.ui.client.service.WorkflowRpcService|isAuthorizedAddReviewer|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/WorkflowRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|71417C94A72A0CF76A43A2B36B8E3E1B|org.kuali.student.core.workflow.ui.client.service.WorkflowRpcService|isAuthorizedRemoveReviewers|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n \n \n \n #\n # Authors and Collaborators\n #\n\n @request.add_thinktime(5)\n\n if(!opts[:collaborator].nil?)\n # Collaborator seach\n for i in 1..opts[:collaborator].length\n itr = i-1\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/SearchRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|14|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|DB85114A8D2B33860498043707FB831D|org.kuali.student.common.ui.client.service.SearchRpcService|search|org.kuali.student.common.search.dto.SearchRequest/2597477947|java.lang.Boolean/476441737|java.util.ArrayList/3821976829|org.kuali.student.common.search.dto.SearchParam/1222427352|person.queryParam.personGivenName|#{opts[:collaborator][0..itr]}|person.queryParam.excludedUserId|psycho1|person.search.personQuickViewByGivenName|person.resultColumn.DisplayName|1|2|3|4|1|5|5|0|6|0|7|2|8|9|0|10|8|11|0|12|13|14|0|0|\"\n } \n ) \n end\n\n @request.add_thinktime(5)\n\n\n contents1 = \"5|0|192|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|A239E8C5A2EDCD8BCE6061BF191A8095|org.kuali.student.lum.lu.ui.course.client.service.CreditCourseProposalRpcService|saveData|org.kuali.student.common.assembly.data.Data/3184510345|org.kuali.student.common.assembly.data.Data|java.util.LinkedHashMap/1551059846|org.kuali.student.common.assembly.data.Data$StringKey/758802082|passFail|org.kuali.student.common.assembly.data.Data$BooleanValue/4261226833|java.lang.Boolean/476441737|audit|finalExamStatus|org.kuali.student.common.assembly.data.Data$StringValue/3151113388|STD|campusLocations|org.kuali.student.common.assembly.data.Data$DataValue/1692468409|org.kuali.student.common.assembly.data.Data$IntegerKey/134469241|java.lang.Integer/3438268394|AL|_runtimeData|id-translation|All|code|#{opts[:subject_area]}#{opts[:course_suffix]}|courseNumberSuffix|#{opts[:course_suffix]}|courseSpecificLOs|loCategoryInfoList|id|%%_#{opts[:lo_category_id_var_name]}%%|loRepository|kuali.loRepository.key.singleUse|metaInfo|createId|admin|createTime|org.kuali.student.common.assembly.data.Data$DateValue/2929953165|java.sql.Timestamp/1769758459|updateId|updateTime|versionInd|0|name|#{opts[:lo_category]}|state|active|type|loCategoryType.subject|loDisplayInfoList|loInfo|sequence|desc|formatted|#{opts[:lo_cat_text]}|plain|%%_#{opts[:lo_category_var_name]}%%|loRepositoryKey|#{opts[:propose_person]}|2|SINGLE USE LO|draft|kuali.lo.type.singleUse|courseTitle|#{opts[:course_title]}|creditOptions|fixedCreditValue|10.0|kuali.creditType.credit.degree.10.0|resultValues|kuali.resultComponentType.credit.degree.fixed|Credits, Fixed|crossListings|descr|#{opts[:course_description]}|duration|atpDurationTypeKey|kuali.atp.duration.Semester|timeQuantity|org.kuali.student.common.assembly.data.Data$IntegerValue/3605481012|Semester|effectiveDate|expenditure|affiliatedOrgs|%%_#{opts[:affliated_orgs_id_name]}%%|orgId|%%_#{opts[:admin_dep_var_name]}%%|percentage|org.kuali.student.common.assembly.data.Data$LongValue/3784756947|java.lang.Long/4227064769|#{admin_org}|fees|feeAmounts|currencyQuantity|currencyTypeKey|kuali.currency.type.usdollars.cents|feeType|kuali.enum.type.feeTypes.labFee|%%_#{opts[:lab_fee_id_name]}%%|1|rateType|fixedRateFee|Fixed Rate Fee|Laboratory Fee|formats|activities|activityType|kuali.lu.type.activity.Lab|contactHours|unitQuantity|10|unitType|kuali.atp.duration.week|per week|defaultEnrollmentEstimate|kuali.atp.duration.Week|Week|%%_#{opts[:atp_duration_week_var_name]}%%|3|unitsContentOwner|Lab|%%_#{opts[:lab_var_name]}%%|termsOffered|kuali.lu.type.CreditCourseFormatShell|gradingOptions|kuali.resultComponent.grade.letter|Letter|%%_#{opts[:clu_ref_dyn_var_name]}%%|instructors|personId|#{opts[:instructor]}|#{opts[:instructor]}, #{opts[:instructor]}(#{opts[:instructor]})|joints|level|100|5|pilotCourse|revenues|%%_#{opts[:revenues_id_name]}%%|REVENUE|%%_#{opts[:revenue_id_name]}%%|specialTopicsCourse|startTerm|kuali.atp.FA2008-2009|subjectArea|#{opts[:subject_area]}|kuali.atp.season.Any|Any|transcriptTitle|#{opts[:course_short_title]}|kuali.lu.type.CreditCourse|#{opts[:oversight_dept_number]}|#{oversight_department}|unitsDeployment|variations|versionInfo|currentVersionStart|sequenceNumber|versionIndId|%%_#{opts[:version_ind_id_name]}%%|Fall Semester of 2008|Standard final Exam|proposal|workflowNode|PreRoute|%%_#{opts[:proposal_dyn_var_name]}%%|7|#{opts[:proposal_title]}|proposalReference|proposalReferenceType|kuali.proposal.referenceType.clu|proposerOrg|proposerPerson|rationale|#{opts[:course_rationale]}|Saved|kuali.proposal.type.course.create|workflowId|%%_#{opts[:proposal_doc_id_var_name]}%%|collaboratorInfo|collaborators|principalId|#{opts[:collaborator]}|permission|KS-SYS~Edit Document|action|F|firstName|lastName|actionRequestStatus|New|author\"\n contents2 = \"|1|2|3|4|1|5|5|6|7|0|38|8|9|10|11|0|8|12|10|-5|8|13|14|15|8|16|17|5|6|7|0|2|18|19|0|14|20|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|23|-19|-21|-12|-17|-1|-10|8|24|14|25|8|26|14|27|8|28|17|5|6|7|0|1|18|-15|17|5|6|7|0|3|8|29|17|5|6|7|0|1|18|-15|17|5|6|7|0|6|8|30|14|31|8|32|14|33|8|34|17|5|6|7|0|5|8|35|14|36|8|37|38|39|3759152200|1288490188800|0|8|40|14|36|8|41|38|39|3759152200|1288490188800|0|8|42|14|43|-45|-51|8|44|14|45|8|46|14|47|8|48|14|49|-41|-43|-37|-39|8|50|17|5|6|7|0|0|-37|-73|8|51|17|5|6|7|0|8|8|52|14|43|8|53|17|5|6|7|0|2|8|54|14|55|8|56|14|55|-79|-83|8|30|14|57|8|58|14|33|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|1864356667|1314259992576|243000000|8|40|14|59|8|41|38|39|1867163616|1314259992576|192000000|8|42|14|60|-79|-95|8|44|14|61|8|46|14|62|8|48|14|63|-37|-77|-33|-35|-1|-31|8|64|14|65|8|66|17|5|6|7|0|1|18|-15|17|5|6|7|0|7|8|67|14|68|8|30|14|69|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|1861297882|1314259992576|458000000|8|40|14|59|8|41|38|39|1861297882|1314259992576|458000000|8|42|14|43|-125|-131|8|70|17|5|6|7|0|1|18|-15|14|68|-125|-147|8|46|14|62|8|48|14|71|8|21|17|5|6|7|0|1|8|48|17|5|6|7|0|1|8|22|14|72|-159|-161|-125|-157|-121|-123|-1|-119|8|73|17|5|6|7|0|0|-1|-167|8|74|17|5|6|7|0|1|8|56|14|75|-1|-171|8|76|17|5|6|7|0|3|8|77|14|78|8|79|80|19|2|8|21|17|5|6|7|0|1|8|77|17|5|6|7|0|1|8|22|14|81|-188|-190|-179|-186|-1|-177|8|82|38|39|470887936|1219770712064|0|8|83|17|5|6|7|0|1|8|84|17|5|6|7|0|1|18|-15|17|5|6|7|0|4|8|30|14|85|8|86|14|87|8|88|89|90|100|0|8|21|17|5|6|7|0|1|8|86|17|5|6|7|0|1|8|22|14|91|-220|-222|-209|-218|-205|-207|-201|-203|-1|-199|8|92|17|5|6|7|0|1|18|-15|17|5|6|7|0|6|8|93|17|5|6|7|0|1|18|-15|17|5|6|7|0|2|8|94|80|19|100|8|95|14|96|-238|-240|-234|-236|8|97|14|98|8|30|14|99|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|1867163501|1314259992576|77000000|8|40|14|59|8|41|38|39|1867163502|1314259992576|78000000|8|42|14|100|-234|-253|8|101|14|102|8|21|17|5|6|7|0|2|8|101|17|5|6|7|0|1|8|22|14|103|-273|-275|8|97|17|5|6|7|0|1|8|22|14|104|-273|-281|-234|-271|-230|-232|-1|-228|8|105|17|5|6|7|0|1|18|-15|17|5|6|7|0|6|8|106|17|5|6|7|0|1|18|-15|17|5|6|7|0|9|8|107|14|108|8|109|17|5|6|7|0|3|8|110|14|111|8|112|14|113|8|21|17|5|6|7|0|1|8|112|17|5|6|7|0|1|8|22|14|114|-315|-317|-307|-313|-301|-305|8|115|80|-246|8|76|17|5|6|7|0|3|8|77|14|116|8|79|80|19|13|8|21|17|5|6|7|0|1|8|77|17|5|6|7|0|1|8|22|14|117|-336|-338|-327|-334|-301|-325|8|30|14|118|8|34|17|5|6|7|0|5|8|35|14|59|8|37\"\n contents3 = \"|38|39|1861297868|1314259992576|444000000|8|40|14|59|8|41|38|39|1867163581|1314259992576|157000000|8|42|14|119|-301|-346|8|46|14|62|8|120|17|5|6|7|0|0|-301|-364|8|21|17|5|6|7|0|1|8|107|17|5|6|7|0|1|8|22|14|121|-370|-372|-301|-368|-297|-299|-293|-295|8|30|14|122|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|1861297859|1314259992576|435000000|8|40|14|59|8|41|38|39|1867163562|1314259992576|138000000|8|42|14|119|-293|-380|8|46|14|62|8|123|17|5|6|7|0|0|-293|-398|8|48|14|124|-289|-291|-1|-287|8|125|17|5|6|7|0|2|18|-15|14|126|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|127|-412|-414|-406|-410|-1|-404|8|30|14|128|8|129|17|5|6|7|0|1|18|-15|17|5|6|7|0|2|8|130|14|131|8|21|17|5|6|7|0|1|8|130|17|5|6|7|0|1|8|22|14|132|-434|-436|-428|-432|-424|-426|-1|-422|8|133|17|5|6|7|0|0|-1|-442|8|134|14|135|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|1854995943|1314259992576|519000000|8|40|14|59|8|41|38|39|1867163502|1314259992576|78000000|8|42|14|136|-1|-448|8|137|10|-5|8|138|17|5|6|7|0|1|18|-15|17|5|6|7|0|4|8|84|17|5|6|7|0|1|18|-15|17|5|6|7|0|4|8|30|14|139|8|86|14|87|8|88|89|90|100|0|8|21|17|5|6|7|0|1|8|86|17|5|6|7|0|1|8|22|14|91|-491|-493|-480|-489|-476|-478|-472|-474|8|97|14|140|8|30|14|141|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|1867163501|1314259992576|77000000|8|40|14|59|8|41|38|39|1867163502|1314259992576|78000000|8|42|14|100|-472|-503|-468|-470|-1|-466|8|142|10|-5|8|143|14|144|8|46|14|62|8|145|14|146|8|123|17|5|6|7|0|2|18|-15|14|147|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|148|-535|-537|-529|-533|-1|-527|8|149|14|150|8|48|14|151|8|120|17|5|6|7|0|2|18|-15|14|152|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|153|-555|-557|-549|-553|-1|-547|8|154|17|5|6|7|0|2|18|-15|14|87|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|91|-571|-573|-565|-569|-1|-563|8|155|17|5|6|7|0|0|-1|-579|8|156|17|5|6|7|0|3|8|157|38|39|1854995943|1314259992576|519000000|8|158|89|90|1|0|8|159|14|160|-1|-583|8|21|17|5|6|7|0|3|8|143|17|5|6|7|0|1|8|22|14|161|-597|-599|8|145|17|5|6|7|0|1|8|22|14|146|-597|-605|8|13|17|5|6|7|0|1|8|22|14|162|-597|-611|-1|-595|8|163|17|5|6|7|0|12|8|164|14|165|8|30|14|166|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|1854996146|1314259992576|722000000|8|40|14|59|8|41|38|39|1867164349|1314259992576|925000000|8|42|14|167|-619|-625|8|44|14|168|8|169|17|5|6|7|0|1|18|-15|14|128|-619|-643|8|170|14|171|8|172|17|5|6|7|0|0|-619|-651|8|173|17|5|6|7|0|0|-619|-655|8|174|14|175|8|46|14|176|8|48|14|177|8|178|14|179|-1|-617|8|180|17|5|6|7|0|1|8|181|17|5|6|7|0|1|18|-15|17|5|6|7|0|7|8|182|14|183|8|184|14|185|8|186|14|187|8|188|14|183|8|189|14|183|8|190|14|191|8|192|10|-5|-673|-675|-669|-671|-1|-667|0|0|\"\n else\n # UPDATE\n contents1 = \"5|0|179|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|D60D3C6E0D395C18A0F44A2D9D2A7348|org.kuali.student.lum.lu.ui.course.client.service.CreditCourseProposalRpcService|saveData|org.kuali.student.common.assembly.data.Data/3184510345|org.kuali.student.common.assembly.data.Data|java.util.LinkedHashMap/1551059846|org.kuali.student.common.assembly.data.Data$StringKey/758802082|passFail|org.kuali.student.common.assembly.data.Data$BooleanValue/4261226833|java.lang.Boolean/476441737|audit|finalExamStatus|org.kuali.student.common.assembly.data.Data$StringValue/3151113388|STD|campusLocations|org.kuali.student.common.assembly.data.Data$DataValue/1692468409|org.kuali.student.common.assembly.data.Data$IntegerKey/134469241|java.lang.Integer/3438268394|AL|_runtimeData|id-translation|All|code|#{opts[:subject_area]}#{opts[:course_suffix]}|courseNumberSuffix|#{opts[:course_suffix]}|courseSpecificLOs|loCategoryInfoList|id|%%_#{opts[:lo_category_id_var_name]}%%|loRepository|kuali.loRepository.key.singleUse|metaInfo|createId|admin|createTime|org.kuali.student.common.assembly.data.Data$DateValue/2929953165|java.sql.Timestamp/1769758459|updateId|updateTime|versionInd|0|name|#{opts[:lo_category]}|state|active|type|loCategoryType.subject|loDisplayInfoList|loInfo|sequence|desc|formatted|#{opts[:lo_cat_text]}|plain|%%_#{opts[:lo_category_var_name]}%%|loRepositoryKey|#{opts[:propose_person]}|2|SINGLE USE LO|draft|kuali.lo.type.singleUse|courseTitle|#{opts[:course_title]}|creditOptions|fixedCreditValue|10.0|kuali.creditType.credit.degree.10.0|resultValues|kuali.resultComponentType.credit.degree.fixed|Credits, Fixed|crossListings|descr|#{opts[:course_description]}|duration|atpDurationTypeKey|kuali.atp.duration.Semester|timeQuantity|org.kuali.student.common.assembly.data.Data$IntegerValue/3605481012|Semester|effectiveDate|expenditure|affiliatedOrgs|%%_#{opts[:affliated_orgs_id_name]}%%|orgId|%%_#{opts[:admin_dep_var_name]}%%|percentage|org.kuali.student.common.assembly.data.Data$LongValue/3784756947|java.lang.Long/4227064769|#{admin_org}|fees|feeAmounts|currencyQuantity|currencyTypeKey|kuali.currency.type.usdollars.cents|feeType|kuali.enum.type.feeTypes.labFee|%%_#{opts[:lab_fee_id_name]}%%|1|rateType|fixedRateFee|Fixed Rate Fee|Laboratory Fee|formats|activities|activityType|kuali.lu.type.activity.Lab|contactHours|unitQuantity|5|unitType|kuali.atp.duration.week|per week|defaultEnrollmentEstimate|kuali.atp.duration.Week|Week|%%_#{opts[:atp_duration_week_var_name]}%%|3|unitsContentOwner|Lab|%%_#{opts[:lab_var_name]}%%|termsOffered|kuali.lu.type.CreditCourseFormatShell|gradingOptions|kuali.resultComponent.grade.letter|Letter|%%_#{opts[:clu_ref_dyn_var_name]}%%|instructors|personId|#{opts[:instructor]}|#{opts[:instructor]}, #{opts[:instructor]}(#{opts[:instructor]})|joints|level|100|pilotCourse|revenues|%%_#{opts[:revenues_id_name]}%%|REVENUE|%%_#{opts[:revenue_id_name]}%%|specialTopicsCourse|startTerm|kuali.atp.FA2008-2009|subjectArea|#{opts[:subject_area]}|kuali.atp.season.Any|Any|transcriptTitle|#{opts[:course_short_title]}|kuali.lu.type.CreditCourse|#{opts[:oversight_dept_number]}|#{oversight_department}|unitsDeployment|variations|versionInfo|currentVersionStart|sequenceNumber|versionIndId|%%_#{opts[:version_ind_id_name]}%%|Fall Semester of 2008|Standard final Exam|proposal|workflowNode|PreRoute|%%_#{opts[:proposal_dyn_var_name]}%%|7|proposalReference|proposalReferenceType|kuali.proposal.referenceType.clu|proposerOrg|proposerPerson|rationale|#{opts[:course_rationale]}|Saved|kuali.proposal.type.course.create|workflowId|%%_#{opts[:proposal_doc_id_var_name]}%%|collaboratorInfo|collaborators\"\n contents2 = \"|1|2|3|4|1|5|5|6|7|0|38|8|9|10|11|0|8|12|10|-5|8|13|14|15|8|16|17|5|6|7|0|2|18|19|0|14|20|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|23|-19|-21|-12|-17|-1|-10|8|24|14|25|8|26|14|27|8|28|17|5|6|7|0|1|18|-15|17|5|6|7|0|3|8|29|17|5|6|7|0|1|18|-15|17|5|6|7|0|6|8|30|14|31|8|32|14|33|8|34|17|5|6|7|0|5|8|35|14|36|8|37|38|39|3759152200|1288490188800|0|8|40|14|36|8|41|38|39|3759152200|1288490188800|0|8|42|14|43|-45|-51|8|44|14|45|8|46|14|47|8|48|14|49|-41|-43|-37|-39|8|50|17|5|6|7|0|0|-37|-73|8|51|17|5|6|7|0|8|8|52|14|43|8|53|17|5|6|7|0|2|8|54|14|55|8|56|14|55|-79|-83|8|30|14|57|8|58|14|33|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3491266071|1309965025280|351000000|8|40|14|59|8|41|38|39|3491272352|1309965025280|632000000|8|42|14|60|-79|-95|8|44|14|61|8|46|14|62|8|48|14|63|-37|-77|-33|-35|-1|-31|8|64|14|65|8|66|17|5|6|7|0|1|18|-15|17|5|6|7|0|7|8|67|14|68|8|30|14|69|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3479039543|1309965025280|823000000|8|40|14|59|8|41|38|39|3479039543|1309965025280|823000000|8|42|14|43|-125|-131|8|70|17|5|6|7|0|1|18|-15|14|68|-125|-147|8|46|14|62|8|48|14|71|8|21|17|5|6|7|0|1|8|48|17|5|6|7|0|1|8|22|14|72|-159|-161|-125|-157|-121|-123|-1|-119|8|73|17|5|6|7|0|0|-1|-167|8|74|17|5|6|7|0|1|8|56|14|75|-1|-171|8|76|17|5|6|7|0|3|8|77|14|78|8|79|80|19|2|8|21|17|5|6|7|0|1|8|77|17|5|6|7|0|1|8|22|14|81|-188|-190|-179|-186|-1|-177|8|82|38|39|470887936|1219770712064|0|8|83|17|5|6|7|0|1|8|84|17|5|6|7|0|1|18|-15|17|5|6|7|0|4|8|30|14|85|8|86|14|87|8|88|89|90|100|0|8|21|17|5|6|7|0|1|8|86|17|5|6|7|0|1|8|22|14|91|-220|-222|-209|-218|-205|-207|-201|-203|-1|-199|8|92|17|5|6|7|0|1|18|-15|17|5|6|7|0|6|8|93|17|5|6|7|0|1|18|-15|17|5|6|7|0|2|8|94|80|19|100|8|95|14|96|-238|-240|-234|-236|8|97|14|98|8|30|14|99|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3491272235|1309965025280|515000000|8|40|14|59|8|41|38|39|3491272236|1309965025280|516000000|8|42|14|100|-234|-253|8|101|14|102|8|21|17|5|6|7|0|2|8|101|17|5|6|7|0|1|8|22|14|103|-273|-275|8|97|17|5|6|7|0|1|8|22|14|104|-273|-281|-234|-271|-230|-232|-1|-228|8|105|17|5|6|7|0|1|18|-15|17|5|6|7|0|6|8|106|17|5|6|7|0|1|18|-15|17|5|6|7|0|9|8|107|14|108|8|109|17|5|6|7|0|3|8|110|14|111|8|112|14|113|8|21|17|5|6|7|0|1|8|112|17|5|6|7|0|1|8|22|14|114|-315|-317|-307|-313|-301|-305|8|115|80|-246|8|76|17|5|6|7|0|3|8|77|14|116|8|79|80|19|13|8|21|17|5|6|7|0|1|8|77|17|5|6|7|0|1|8|22|14|117|-336|-338|-327|-334|-301|-325|8|30|14|118|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3491262219|1309965025280|499000000|8|40|14|59|8|41|38|39|3491272295|1309965025280|575000000|8|42|14|119|-301|-346|8|46|14|62|8|120|17|5|6|7|0|0|-301|-364|8|21|17|5|6|7|0|1|8|107|17|5|6|7|0|1|8|22|14|121|-370|-372|-301|-368|-297|-299|-293|-295|8|30|14|122|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3491262210|1309965025280|490000000|8|40|14|59|8|41|38|39|3491272289|1309965025280|569000000|8|42|14|119|-293|-380|8|46|14|62|8|123|17|5|6|7|0|0|-293|-398|8|48|14|124|-289|-291|-1|-287|8|125|17|5|6|7|0|2|18|-15|14|126|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|127|-412|-414|-406|-410|-1|-404|8|30|14|128|8|129|17|5|6|7|0|1|18|-15|17|5|6|7|0|2|8|130|14|131|8|21|17|5|6|7|0|1|8|130|17|5|6|7|0|1|8|22|14|132|-434|-436|-428\"\n contents3 = \"|-432|-424|-426|-1|-422|8|133|17|5|6|7|0|0|-1|-442|8|134|14|135|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3491257291|1309965025280|571000000|8|40|14|59|8|41|38|39|3491272236|1309965025280|516000000|8|42|14|111|-1|-448|8|136|10|-5|8|137|17|5|6|7|0|1|18|-15|17|5|6|7|0|4|8|84|17|5|6|7|0|1|18|-15|17|5|6|7|0|4|8|30|14|138|8|86|14|87|8|88|89|90|100|0|8|21|17|5|6|7|0|1|8|86|17|5|6|7|0|1|8|22|14|91|-491|-493|-480|-489|-476|-478|-472|-474|8|97|14|139|8|30|14|140|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3491272233|1309965025280|513000000|8|40|14|59|8|41|38|39|3491272236|1309965025280|516000000|8|42|14|100|-472|-503|-468|-470|-1|-466|8|141|10|-5|8|142|14|143|8|46|14|62|8|144|14|145|8|123|17|5|6|7|0|2|18|-15|14|146|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|147|-535|-537|-529|-533|-1|-527|8|148|14|149|8|48|14|150|8|120|17|5|6|7|0|2|18|-15|14|151|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|152|-555|-557|-549|-553|-1|-547|8|153|17|5|6|7|0|2|18|-15|14|87|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|91|-571|-573|-565|-569|-1|-563|8|154|17|5|6|7|0|0|-1|-579|8|155|17|5|6|7|0|3|8|156|38|39|3491257291|1309965025280|571000000|8|157|89|90|1|0|8|158|14|159|-1|-583|8|21|17|5|6|7|0|3|8|142|17|5|6|7|0|1|8|22|14|160|-597|-599|8|144|17|5|6|7|0|1|8|22|14|145|-597|-605|8|13|17|5|6|7|0|1|8|22|14|161|-597|-611|-1|-595|8|162|17|5|6|7|0|12|8|163|14|164|8|30|14|165|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3491257553|1309965025280|833000000|8|40|14|59|8|41|38|39|3491273130|1309965025280|410000000|8|42|14|166|-619|-625|8|44|14|65|8|167|17|5|6|7|0|1|18|-15|14|128|-619|-643|8|168|14|169|8|170|17|5|6|7|0|0|-619|-651|8|171|17|5|6|7|0|0|-619|-655|8|172|14|173|8|46|14|174|8|48|14|175|8|176|14|177|-1|-617|8|178|17|5|6|7|0|1|8|179|17|5|6|7|0|0|-669|-671|-1|-667|0|0|\"\n end\n\n # Save & Continue\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/CreditCourseProposalRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"#{contents1}#{contents2}#{contents3}\"\n },\n {\n 'subst' => 'true',\n :dyn_variables => [\n {\"name\" => opts[:action_request_id_name], \"re\" => opts[:action_request_id_regexp]}\n ]\n }\n )\n \n #@request.add(\"DEBUG/affliated_orgs_id_name/%%_#{opts[:action_request_id_name]}%%\", {}, {'subst' => 'true'})\n \n \n @request.add_thinktime(2)\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/WorkflowRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|71417C94A72A0CF76A43A2B36B8E3E1B|org.kuali.student.core.workflow.ui.client.service.WorkflowRpcService|getActionsRequested|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/WorkflowRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|71417C94A72A0CF76A43A2B36B8E3E1B|org.kuali.student.core.workflow.ui.client.service.WorkflowRpcService|getDocumentStatus|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/ProposalRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|12BDE6C2DA6A7CF74BE0FBF074E806E1|org.kuali.student.core.proposal.ui.client.service.ProposalRpcService|getProposalByWorkflowId|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/DocumentRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|7|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|5771428875B68D3E8EC7527EC8D18D40|org.kuali.student.core.document.ui.client.service.DocumentRpcService|isAuthorizedUploadDocuments|java.lang.String/2004016611|%%_#{opts[:proposal_dyn_var_name]}%%|referenceType.clu.proposal|1|2|3|4|2|5|5|6|7|\"\n },\n {\n 'subst' => 'true'\n }\n )\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/DocumentRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|7|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|5771428875B68D3E8EC7527EC8D18D40|org.kuali.student.core.document.ui.client.service.DocumentRpcService|getRefDocIdsForRef|java.lang.String/2004016611|kuali.org.RefObjectType.ProposalInfo|%%_#{opts[:proposal_dyn_var_name]}%%|1|2|3|4|2|5|5|6|7|\"\n },\n {\n 'subst' => 'true'\n }\n )\n \n \n \n #\n # Support Documents\n # Nothing uploaded\n #\n\n @request.add_thinktime(5)\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/DocumentRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|7|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|5771428875B68D3E8EC7527EC8D18D40|org.kuali.student.core.document.ui.client.service.DocumentRpcService|getRefDocIdsForRef|java.lang.String/2004016611|kuali.org.RefObjectType.ProposalInfo|%%_#{opts[:proposal_dyn_var_name]}%%|1|2|3|4|2|5|5|6|7|\"\n },\n {\n 'subst' => 'true'\n }\n )\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/WorkflowRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|71417C94A72A0CF76A43A2B36B8E3E1B|org.kuali.student.core.workflow.ui.client.service.WorkflowRpcService|getActionsRequested|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/WorkflowRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|71417C94A72A0CF76A43A2B36B8E3E1B|org.kuali.student.core.workflow.ui.client.service.WorkflowRpcService|getDocumentStatus|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/ProposalRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|12BDE6C2DA6A7CF74BE0FBF074E806E1|org.kuali.student.core.proposal.ui.client.service.ProposalRpcService|getProposalByWorkflowId|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n\n\n if(!opts[:collaborator].nil?)\n contents1 = \"5|0|195|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|3C9BBAD14113E13A72476EEE8100687B|org.kuali.student.lum.lu.ui.course.client.service.CourseRpcService|validate|org.kuali.student.common.assembly.data.Data/3184510345|org.kuali.student.common.assembly.data.Data|java.util.LinkedHashMap/1551059846|org.kuali.student.common.assembly.data.Data$StringKey/758802082|passFail|org.kuali.student.common.assembly.data.Data$BooleanValue/4261226833|java.lang.Boolean/476441737|audit|finalExamStatus|org.kuali.student.common.assembly.data.Data$StringValue/3151113388|STD|campusLocations|org.kuali.student.common.assembly.data.Data$DataValue/1692468409|org.kuali.student.common.assembly.data.Data$IntegerKey/134469241|java.lang.Integer/3438268394|AL|_runtimeData|id-translation|All|code|#{opts[:subject_area]}#{opts[:course_suffix]}|courseNumberSuffix|#{opts[:course_suffix]}|courseSpecificLOs|loCategoryInfoList|id|%%_#{opts[:lo_category_id_var_name]}%%|loRepository|kuali.loRepository.key.singleUse|metaInfo|createId|admin|createTime|org.kuali.student.common.assembly.data.Data$DateValue/2929953165|java.sql.Timestamp/1769758459|updateId|updateTime|versionInd|0|name|#{opts[:lo_category]}|state|active|type|loCategoryType.subject|loDisplayInfoList|loInfo|sequence|desc|formatted|#{opts[:lo_cat_text]}|plain|%%_#{opts[:lo_category_var_name]}%%|loRepositoryKey|#{opts[:propose_person]}|3|SINGLE USE LO|draft|kuali.lo.type.singleUse|courseTitle|#{opts[:course_title]}|creditOptions|fixedCreditValue|10.0|kuali.creditType.credit.degree.10.0|resultValues|kuali.resultComponentType.credit.degree.fixed|Credits, Fixed|crossListings|descr|#{opts[:course_description]}|duration|atpDurationTypeKey|kuali.atp.duration.Semester|timeQuantity|org.kuali.student.common.assembly.data.Data$IntegerValue/3605481012|Semester|effectiveDate|expenditure|affiliatedOrgs|%%_#{opts[:affliated_orgs_id_name]}%%|orgId|%%_#{opts[:admin_dep_var_name]}%%|percentage|org.kuali.student.common.assembly.data.Data$LongValue/3784756947|java.lang.Long/4227064769|#{admin_org}|fees|feeAmounts|currencyQuantity|currencyTypeKey|kuali.currency.type.usdollars.cents|feeType|kuali.enum.type.feeTypes.labFee|%%_#{opts[:lab_fee_id_name]}%%|2|rateType|fixedRateFee|Fixed Rate Fee|Laboratory Fee|formats|activities|activityType|kuali.lu.type.activity.Lab|contactHours|unitQuantity|10|unitType|kuali.atp.duration.week|per week|defaultEnrollmentEstimate|kuali.atp.duration.Week|Week|%%_#{opts[:atp_duration_week_var_name]}%%|4|unitsContentOwner|Lab|%%_#{opts[:lab_var_name]}%%|termsOffered|kuali.lu.type.CreditCourseFormatShell|gradingOptions|kuali.resultComponent.grade.letter|Letter|%%_#{opts[:clu_ref_dyn_var_name]}%%|instructors|personId|#{opts[:instructor]}|#{opts[:instructor]}, #{opts[:instructor]}(#{opts[:instructor]})|joints|level|100|6|pilotCourse|revenues|%%_#{opts[:revenues_id_name]}%%|REVENUE|%%_#{opts[:revenue_id_name]}%%|specialTopicsCourse|startTerm|kuali.atp.FA2008-2009|subjectArea|#{opts[:subject_area]}|kuali.atp.season.Any|Any|transcriptTitle|#{opts[:course_short_title]}|kuali.lu.type.CreditCourse|#{opts[:oversight_dept_number]}|#{oversight_department}|unitsDeployment|variations|versionInfo|currentVersionStart|sequenceNumber|versionIndId|%%_#{opts[:version_ind_id_name]}%%|Fall Semester of 2008|Standard final Exam|proposal|workflowNode|PreRoute|%%_#{opts[:proposal_dyn_var_name]}%%|8|#{opts[:proposal_title]}|proposalReference|proposalReferenceType|kuali.proposal.referenceType.clu|proposerOrg|proposerPerson|rationale|#{opts[:course_rationale]}|Saved|kuali.proposal.type.course.create|workflowId|%%_#{opts[:proposal_doc_id_var_name]}%%|collaboratorInfo|collaborators|action|F|actionRequestId|%%_#{opts[:action_request_id_name]}%%|actionRequestStatus|Active|author|canRevokeRequest|firstName|#{opts[:collaborator]}|lastName|permission|KS-SYS~Edit Document|principalId\"\n contents2 = \"|1|2|3|4|1|5|5|6|7|0|38|8|9|10|11|0|8|12|10|-5|8|13|14|15|8|16|17|5|6|7|0|2|18|19|0|14|20|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|23|-19|-21|-12|-17|-1|-10|8|24|14|25|8|26|14|27|8|28|17|5|6|7|0|1|18|-15|17|5|6|7|0|3|8|29|17|5|6|7|0|1|18|-15|17|5|6|7|0|6|8|30|14|31|8|32|14|33|8|34|17|5|6|7|0|5|8|35|14|36|8|37|38|39|3759152200|1288490188800|0|8|40|14|36|8|41|38|39|3759152200|1288490188800|0|8|42|14|43|-45|-51|8|44|14|45|8|46|14|47|8|48|14|49|-41|-43|-37|-39|8|50|17|5|6|7|0|0|-37|-73|8|51|17|5|6|7|0|8|8|52|14|43|8|53|17|5|6|7|0|2|8|54|14|55|8|56|14|55|-79|-83|8|30|14|57|8|58|14|33|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|1864356667|1314259992576|243000000|8|40|14|59|8|41|38|39|1868678977|1314259992576|553000000|8|42|14|60|-79|-95|8|44|14|61|8|46|14|62|8|48|14|63|-37|-77|-33|-35|-1|-31|8|64|14|65|8|66|17|5|6|7|0|1|18|-15|17|5|6|7|0|7|8|67|14|68|8|30|14|69|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|1861297882|1314259992576|458000000|8|40|14|59|8|41|38|39|1861297882|1314259992576|458000000|8|42|14|43|-125|-131|8|70|17|5|6|7|0|1|18|-15|14|68|-125|-147|8|46|14|62|8|48|14|71|8|21|17|5|6|7|0|1|8|48|17|5|6|7|0|1|8|22|14|72|-159|-161|-125|-157|-121|-123|-1|-119|8|73|17|5|6|7|0|0|-1|-167|8|74|17|5|6|7|0|1|8|56|14|75|-1|-171|8|76|17|5|6|7|0|3|8|77|14|78|8|79|80|19|2|8|21|17|5|6|7|0|1|8|77|17|5|6|7|0|1|8|22|14|81|-188|-190|-179|-186|-1|-177|8|82|38|39|470887936|1219770712064|0|8|83|17|5|6|7|0|1|8|84|17|5|6|7|0|1|18|-15|17|5|6|7|0|4|8|30|14|85|8|86|14|87|8|88|89|90|100|0|8|21|17|5|6|7|0|1|8|86|17|5|6|7|0|1|8|22|14|91|-220|-222|-209|-218|-205|-207|-201|-203|-1|-199|8|92|17|5|6|7|0|1|18|-15|17|5|6|7|0|6|8|93|17|5|6|7|0|1|18|-15|17|5|6|7|0|2|8|94|80|19|100|8|95|14|96|-238|-240|-234|-236|8|97|14|98|8|30|14|99|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|1867163501|1314259992576|77000000|8|40|14|59|8|41|38|39|1868678891|1314259992576|467000000|8|42|14|100|-234|-253|8|101|14|102|8|21|17|5|6|7|0|2|8|101|17|5|6|7|0|1|8|22|14|103|-273|-275|8|97|17|5|6|7|0|1|8|22|14|104|-273|-281|-234|-271|-230|-232|-1|-228|8|105|17|5|6|7|0|1|18|-15|17|5|6|7|0|6|8|106|17|5|6|7|0|1|18|-15|17|5|6|7|0|9|8|107|14|108|8|109|17|5|6|7|0|3|8|110|14|111|8|112|14|113|8|21|17|5|6|7|0|1|8|112|17|5|6|7|0|1|8|22|14|114|-315|-317|-307|-313|-301|-305|8|115|80|-246|8|76|17|5|6|7|0|3|8|77|14|116|8|79|80|19|13|8|21|17|5|6|7|0|1|8|77|17|5|6|7|0|1|8|22|14|117|-336|-338|-327|-334|-301|-325|8|30|14|118|8|34|17|5|6|7|0|5|8|35|14|59|8\"\n contents3 = \"|37|38|39|1861297868|1314259992576|444000000|8|40|14|59|8|41|38|39|1868678938|1314259992576|514000000|8|42|14|119|-301|-346|8|46|14|62|8|120|17|5|6|7|0|0|-301|-364|8|21|17|5|6|7|0|1|8|107|17|5|6|7|0|1|8|22|14|121|-370|-372|-301|-368|-297|-299|-293|-295|8|30|14|122|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|1861297859|1314259992576|435000000|8|40|14|59|8|41|38|39|1868678932|1314259992576|508000000|8|42|14|119|-293|-380|8|46|14|62|8|123|17|5|6|7|0|0|-293|-398|8|48|14|124|-289|-291|-1|-287|8|125|17|5|6|7|0|2|18|-15|14|126|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|127|-412|-414|-406|-410|-1|-404|8|30|14|128|8|129|17|5|6|7|0|1|18|-15|17|5|6|7|0|2|8|130|14|131|8|21|17|5|6|7|0|1|8|130|17|5|6|7|0|1|8|22|14|132|-434|-436|-428|-432|-424|-426|-1|-422|8|133|17|5|6|7|0|0|-1|-442|8|134|14|135|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|1854995943|1314259992576|519000000|8|40|14|59|8|41|38|39|1868678891|1314259992576|467000000|8|42|14|136|-1|-448|8|137|10|-5|8|138|17|5|6|7|0|1|18|-15|17|5|6|7|0|4|8|84|17|5|6|7|0|1|18|-15|17|5|6|7|0|4|8|30|14|139|8|86|14|87|8|88|89|90|100|0|8|21|17|5|6|7|0|1|8|86|17|5|6|7|0|1|8|22|14|91|-491|-493|-480|-489|-476|-478|-472|-474|8|97|14|140|8|30|14|141|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|1867163501|1314259992576|77000000|8|40|14|59|8|41|38|39|1868678891|1314259992576|467000000|8|42|14|100|-472|-503|-468|-470|-1|-466|8|142|10|-5|8|143|14|144|8|46|14|62|8|145|14|146|8|123|17|5|6|7|0|2|18|-15|14|147|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|148|-535|-537|-529|-533|-1|-527|8|149|14|150|8|48|14|151|8|120|17|5|6|7|0|2|18|-15|14|152|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|153|-555|-557|-549|-553|-1|-547|8|154|17|5|6|7|0|2|18|-15|14|87|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|91|-571|-573|-565|-569|-1|-563|8|155|17|5|6|7|0|0|-1|-579|8|156|17|5|6|7|0|3|8|157|38|39|1854995943|1314259992576|519000000|8|158|89|90|1|0|8|159|14|160|-1|-583|8|21|17|5|6|7|0|3|8|143|17|5|6|7|0|1|8|22|14|161|-597|-599|8|145|17|5|6|7|0|1|8|22|14|146|-597|-605|8|13|17|5|6|7|0|1|8|22|14|162|-597|-611|-1|-595|8|163|17|5|6|7|0|12|8|164|14|165|8|30|14|166|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|1854996146|1314259992576|722000000|8|40|14|59|8|41|38|39|1868679626|1314259992576|202000000|8|42|14|167|-619|-625|8|44|14|168|8|169|17|5|6|7|0|1|18|-15|14|128|-619|-643|8|170|14|171|8|172|17|5|6|7|0|0|-619|-651|8|173|17|5|6|7|0|0|-619|-655|8|174|14|175|8|46|14|176|8|48|14|177|8|178|14|179|-1|-617|8|180|17|5|6|7|0|1|8|181|17|5|6|7|0|1|18|-15|17|5|6|7|0|9|8|182|14|183|8|184|14|185|8|186|14|187|8|188|10|-5|8|189|10|11|1|8|190|14|191|8|192|14|191|8|193|14|194|8|195|14|191|-673|-675|-669|-671|-1|-667|0|0|\"\n else\n #UPDATE\n contents1 = \"5|0|180|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|1ED48DA6F48F82765FE7B58378EA94E0|org.kuali.student.lum.lu.ui.course.client.service.CourseRpcService|validate|org.kuali.student.common.assembly.data.Data/3184510345|org.kuali.student.common.assembly.data.Data|java.util.LinkedHashMap/1551059846|org.kuali.student.common.assembly.data.Data$StringKey/758802082|passFail|org.kuali.student.common.assembly.data.Data$BooleanValue/4261226833|java.lang.Boolean/476441737|audit|finalExamStatus|org.kuali.student.common.assembly.data.Data$StringValue/3151113388|STD|campusLocations|org.kuali.student.common.assembly.data.Data$DataValue/1692468409|org.kuali.student.common.assembly.data.Data$IntegerKey/134469241|java.lang.Integer/3438268394|AL|_runtimeData|id-translation|All|code|#{opts[:subject_area]}#{opts[:course_suffix]}|courseNumberSuffix|#{opts[:course_suffix]}|courseSpecificLOs|loCategoryInfoList|id|%%_#{opts[:lo_category_id_var_name]}%%|loRepository|kuali.loRepository.key.singleUse|metaInfo|createId|admin|createTime|org.kuali.student.common.assembly.data.Data$DateValue/2929953165|java.sql.Timestamp/1769758459|updateId|updateTime|versionInd|0|name|#{opts[:lo_category]}|state|active|type|loCategoryType.subject|loDisplayInfoList|loInfo|sequence|desc|formatted|#{opts[:lo_cat_text]}|plain|%%_#{opts[:lo_category_var_name]}%%|loRepositoryKey|#{opts[:propose_person]}|3|SINGLE USE LO|draft|kuali.lo.type.singleUse|courseTitle|#{opts[:course_title]}|creditOptions|fixedCreditValue|10.0|kuali.creditType.credit.degree.10.0|resultValues|kuali.resultComponentType.credit.degree.fixed|Credits, Fixed|crossListings|descr|#{opts[:course_description]}|duration|atpDurationTypeKey|kuali.atp.duration.Semester|timeQuantity|org.kuali.student.common.assembly.data.Data$IntegerValue/3605481012|Semester|effectiveDate|expenditure|affiliatedOrgs|%%_#{opts[:affliated_orgs_id_name]}%%|orgId|%%_#{opts[:admin_dep_var_name]}%%|percentage|org.kuali.student.common.assembly.data.Data$LongValue/3784756947|java.lang.Long/4227064769|#{admin_org}|fees|feeAmounts|currencyQuantity|currencyTypeKey|kuali.currency.type.usdollars.cents|feeType|kuali.enum.type.feeTypes.labFee|%%_#{opts[:lab_fee_id_name]}%%|2|rateType|fixedRateFee|Fixed Rate Fee|Laboratory Fee|formats|activities|activityType|kuali.lu.type.activity.Lab|contactHours|unitQuantity|5|unitType|kuali.atp.duration.week|per week|defaultEnrollmentEstimate|kuali.atp.duration.Week|Week|%%_#{opts[:atp_duration_week_var_name]}%%|4|unitsContentOwner|Lab|%%_#{opts[:lab_var_name]}%%|termsOffered|kuali.lu.type.CreditCourseFormatShell|gradingOptions|kuali.resultComponent.grade.letter|Letter|%%_#{opts[:clu_ref_dyn_var_name]}%%|instructors|personId|#{opts[:instructor]}|#{opts[:instructor]}, #{opts[:instructor]}(#{opts[:instructor]})|joints|level|100|6|pilotCourse|revenues|%%_#{opts[:revenues_id_name]}%%|REVENUE|%%_#{opts[:revenue_id_name]}%%|specialTopicsCourse|startTerm|kuali.atp.FA2008-2009|subjectArea|#{opts[:subject_area]}|kuali.atp.season.Any|Any|transcriptTitle|#{opts[:course_short_title]}|kuali.lu.type.CreditCourse|#{opts[:oversight_dept_number]}|#{oversight_department}|unitsDeployment|variations|versionInfo|currentVersionStart|sequenceNumber|versionIndId|%%_#{opts[:version_ind_id_name]}%%|Fall Semester of 2008|Standard final Exam|proposal|workflowNode|PreRoute|%%_#{opts[:proposal_dyn_var_name]}%%|8|proposalReference|proposalReferenceType|kuali.proposal.referenceType.clu|proposerOrg|proposerPerson|rationale|#{opts[:course_rationale]}|Saved|kuali.proposal.type.course.create|workflowId|%%_#{opts[:proposal_doc_id_var_name]}%%|collaboratorInfo|collaborators\"\n contents2 = \"|1|2|3|4|1|5|5|6|7|0|38|8|9|10|11|0|8|12|10|-5|8|13|14|15|8|16|17|5|6|7|0|2|18|19|0|14|20|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|23|-19|-21|-12|-17|-1|-10|8|24|14|25|8|26|14|27|8|28|17|5|6|7|0|1|18|-15|17|5|6|7|0|3|8|29|17|5|6|7|0|1|18|-15|17|5|6|7|0|6|8|30|14|31|8|32|14|33|8|34|17|5|6|7|0|5|8|35|14|36|8|37|38|39|3759152200|1288490188800|0|8|40|14|36|8|41|38|39|3759152200|1288490188800|0|8|42|14|43|-45|-51|8|44|14|45|8|46|14|47|8|48|14|49|-41|-43|-37|-39|8|50|17|5|6|7|0|0|-37|-73|8|51|17|5|6|7|0|8|8|52|14|43|8|53|17|5|6|7|0|2|8|54|14|55|8|56|14|55|-79|-83|8|30|14|57|8|58|14|33|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3498783321|1309965025280|601000000|8|40|14|59|8|41|38|39|3498791687|1309965025280|967000000|8|42|14|60|-79|-95|8|44|14|61|8|46|14|62|8|48|14|63|-37|-77|-33|-35|-1|-31|8|64|14|65|8|66|17|5|6|7|0|1|18|-15|17|5|6|7|0|7|8|67|14|68|8|30|14|69|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3479039543|1309965025280|823000000|8|40|14|59|8|41|38|39|3479039543|1309965025280|823000000|8|42|14|43|-125|-131|8|70|17|5|6|7|0|1|18|-15|14|68|-125|-147|8|46|14|62|8|48|14|71|8|21|17|5|6|7|0|1|8|48|17|5|6|7|0|1|8|22|14|72|-159|-161|-125|-157|-121|-123|-1|-119|8|73|17|5|6|7|0|0|-1|-167|8|74|17|5|6|7|0|1|8|56|14|75|-1|-171|8|76|17|5|6|7|0|3|8|77|14|78|8|79|80|19|2|8|21|17|5|6|7|0|1|8|77|17|5|6|7|0|1|8|22|14|81|-188|-190|-179|-186|-1|-177|8|82|38|39|470887936|1219770712064|0|8|83|17|5|6|7|0|1|8|84|17|5|6|7|0|1|18|-15|17|5|6|7|0|4|8|30|14|85|8|86|14|87|8|88|89|90|100|0|8|21|17|5|6|7|0|1|8|86|17|5|6|7|0|1|8|22|14|91|-220|-222|-209|-218|-205|-207|-201|-203|-1|-199|8|92|17|5|6|7|0|1|18|-15|17|5|6|7|0|6|8|93|17|5|6|7|0|1|18|-15|17|5|6|7|0|2|8|94|80|19|100|8|95|14|96|-238|-240|-234|-236|8|97|14|98|8|30|14|99|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3498789104|1309965025280|384000000|8|40|14|59|8|41|38|39|3498791612|1309965025280|892000000|8|42|14|100|-234|-253|8|101|14|102|8|21|17|5|6|7|0|2|8|101|17|5|6|7|0|1|8|22|14|103|-273|-275|8|97|17|5|6|7|0|1|8|22|14|104|-273|-281|-234|-271|-230|-232|-1|-228|8|105|17|5|6|7|0|1|18|-15|17|5|6|7|0|6|8|106|17|5|6|7|0|1|18|-15|17|5|6|7|0|9|8|107|14|108|8|109|17|5|6|7|0|3|8|110|14|111|8|112|14|113|8|21|17|5|6|7|0|1|8|112|17|5|6|7|0|1|8|22|14|114|-315|-317|-307|-313|-301|-305|8|115|80|-246|8|76|17|5|6|7|0|3|8|77|14|116|8|79|80|19|13|8|21|17|5|6|7|0|1|8|77|17|5|6|7|0|1|8|22|14|117|-336|-338|-327|-334|-301|-325|8|30|14|118|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3498779786|1309965025280|66000000|8|40|14|59|8|41|38|39|3498791652|1309965025280|932000000|8|42|14|119|-301|-346|8|46|14|62|8|120|17|5|6|7|0|0|-301|-364|8|21|17|5|6|7|0|1|8|107|17|5|6|7|0|1|8|22|14|121|-370|-372|-301|-368|-297|-299|-293|-295|8|30|14|122|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3498779778|1309965025280|58000000|8|40|14|59|8|41|38|39|3498791646|1309965025280|926000000|8|42|14|119|-293|-380|8|46|14|62|8|123|17|5|6|7|0|0|-293|-398|8|48|14|124|-289|-291|-1|-287|8|125|17|5|6|7|0|2|18|-15|14|126|8|21|17|5|6\"\n contents3 = \"|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|127|-412|-414|-406|-410|-1|-404|8|30|14|128|8|129|17|5|6|7|0|1|18|-15|17|5|6|7|0|2|8|130|14|131|8|21|17|5|6|7|0|1|8|130|17|5|6|7|0|1|8|22|14|132|-434|-436|-428|-432|-424|-426|-1|-422|8|133|17|5|6|7|0|0|-1|-442|8|134|14|135|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3498775348|1309965025280|628000000|8|40|14|59|8|41|38|39|3498791612|1309965025280|892000000|8|42|14|136|-1|-448|8|137|10|-5|8|138|17|5|6|7|0|1|18|-15|17|5|6|7|0|4|8|84|17|5|6|7|0|1|18|-15|17|5|6|7|0|4|8|30|14|139|8|86|14|87|8|88|89|90|100|0|8|21|17|5|6|7|0|1|8|86|17|5|6|7|0|1|8|22|14|91|-491|-493|-480|-489|-476|-478|-472|-474|8|97|14|140|8|30|14|141|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3498789104|1309965025280|384000000|8|40|14|59|8|41|38|39|3498791612|1309965025280|892000000|8|42|14|100|-472|-503|-468|-470|-1|-466|8|142|10|-5|8|143|14|144|8|46|14|62|8|145|14|146|8|123|17|5|6|7|0|2|18|-15|14|147|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|148|-535|-537|-529|-533|-1|-527|8|149|14|150|8|48|14|151|8|120|17|5|6|7|0|2|18|-15|14|152|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|153|-555|-557|-549|-553|-1|-547|8|154|17|5|6|7|0|2|18|-15|14|87|8|21|17|5|6|7|0|1|18|-15|17|5|6|7|0|1|8|22|14|91|-571|-573|-565|-569|-1|-563|8|155|17|5|6|7|0|0|-1|-579|8|156|17|5|6|7|0|3|8|157|38|39|3498775348|1309965025280|628000000|8|158|89|90|1|0|8|159|14|160|-1|-583|8|21|17|5|6|7|0|3|8|143|17|5|6|7|0|1|8|22|14|161|-597|-599|8|145|17|5|6|7|0|1|8|22|14|146|-597|-605|8|13|17|5|6|7|0|1|8|22|14|162|-597|-611|-1|-595|8|163|17|5|6|7|0|12|8|164|14|165|8|30|14|166|8|34|17|5|6|7|0|5|8|35|14|59|8|37|38|39|3498775596|1309965025280|876000000|8|40|14|59|8|41|38|39|3498792455|1309965025280|735000000|8|42|14|167|-619|-625|8|44|14|65|8|168|17|5|6|7|0|1|18|-15|14|128|-619|-643|8|169|14|170|8|171|17|5|6|7|0|0|-619|-651|8|172|17|5|6|7|0|0|-619|-655|8|173|14|174|8|46|14|175|8|48|14|176|8|177|14|178|-1|-617|8|179|17|5|6|7|0|1|8|180|17|5|6|7|0|0|-669|-671|-1|-667|0|0|\"\n end\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/CreditCourseProposalRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"#{contents1}#{contents2}#{contents3}\"\n },\n {\n 'subst' => 'true'\n }\n )\n\n @request.add_thinktime(5)\n\n\n # Submit to worflow\n if(opts[:submit])\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/WorkflowRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|71417C94A72A0CF76A43A2B36B8E3E1B|org.kuali.student.core.workflow.ui.client.service.WorkflowRpcService|submitDocumentWithId|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/WorkflowRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|71417C94A72A0CF76A43A2B36B8E3E1B|org.kuali.student.core.workflow.ui.client.service.WorkflowRpcService|getActionsRequested|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n\n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/WorkflowRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|71417C94A72A0CF76A43A2B36B8E3E1B|org.kuali.student.core.workflow.ui.client.service.WorkflowRpcService|getDocumentStatus|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n \n @request.add('/org.kuali.student.lum.lu.ui.main.LUMMain/rpcservices/ProposalRpcService',\n {\n 'method' => 'POST',\n 'content_type' => 'text/x-gwt-rpc; charset=utf-8',\n 'contents' => \"5|0|6|#{@request.url}/org.kuali.student.lum.lu.ui.main.LUMMain/|12BDE6C2DA6A7CF74BE0FBF074E806E1|org.kuali.student.core.proposal.ui.client.service.ProposalRpcService|getProposalByWorkflowId|java.lang.String/2004016611|%%_#{opts[:proposal_doc_id_var_name]}%%|1|2|3|4|1|5|6|\"\n },\n {\n 'subst' => 'true'\n }\n )\n \n end\n \n \n end",
"def initialize(html, name, section, tagline, manual=nil, version=nil, date=nil)\n @buf = []\n title_heading name, section, tagline, manual, version, date\n doc = Hpricot(html)\n remove_extraneous_elements! doc\n normalize_whitespace! doc\n block_filter doc\n write \"\\n\"\n end",
"def to_html(datums)\n #, :id, :description, :conditions, :explanation, :format\n require 'markaby'\n mab = Markaby::Builder.new\n mab.html do\n head { title \"Loonheffing stuff\" }\n body do \n datums.each do |d|\n h1 d.name\n table do\n tr { td \"Identifier\"; td d.id }\n tr { td \"Description\"; td d.description }\n cs = d.conditions.split(/\\[[0-9]+\\]/)\n cs = cs[1..-1] # get strange first thing out of the way\n tr do \n td \"Conditions\"; td do\n ul do \n cs.each do |c|\n li c\n end\n end\n end\n end\n tr { td \"Explanation\"; td d.explanation }\n tr { td \"Format\"; td d.format }\n end\n end\n end\n end\n puts mab.to_s\nend",
"def perform\n page_ref.ensure_attributes :content\n # The recipe_page will assert path markers and clear our content\n # if changes during page parsing were significant\n if content_needed?\n if page_ref.recipe_page_ready? # Ready to build\n # reload if persisted? # Possibly the recipe_page changed us\n recipe_page.ensure_attributes :content # Parse the page into one or more recipes\n end\n content_to_parse =\n (recipe_page&.selected_content(anchor_path, focus_path) if anchor_path.present? && focus_path.present?) ||\n page_ref.trimmed_content\n return unless content_to_parse.present?\n ps = ParserServices.parse entity: self, content: content_to_parse\n new_content = ps.annotate\n nc1 = ParsingServices.new(self).parse_and_annotate content_to_parse\n return unless new_content.present? # Parsing was a success\n accept_attribute :content, new_content, true # Force the new content\n end\n # super if defined?(super)\n end",
"def part_B(f)\n panel(f,\n \"Part B Additional Expenses: \".html_safe +\n content_tag(:small, \"Travel, Training, Equipment, and Supplies\")) do\n a = content_tag :p, \"Complete this section \" \\\n \"to authorize payment \" \\\n \"to a supplier \" \\\n \"for expenses related to travel, \" \\\n \"training, \" \\\n \"equipment \" \\\n \"or materials \" \\\n \"directly on behalf of a parent or guardian.\"\n\n a += form_row do\n # puts \"RENDERING: #{f.object.funded_person.valid_fiscal_years.map(&:to_s)}\"\n # puts \"SHOULD BE SELECTED: #{f.object.part_b_fiscal_year.to_s}\"\n # puts options_for_select(f.object.funded_person.valid_fiscal_years.map(&:to_s), f.object.part_b_fiscal_year.to_s)\n # puts options_for_select(f\n # .object\n # .funded_person\n # .valid_fiscal_years\n # .collect(&:to_s),\n # f.object.part_b_fiscal_year)\n f.select(:part_b_fiscal_year,\n options_for_select(f\n .object\n .funded_person\n .valid_fiscal_years\n .collect(&:to_s),\n f.object.part_b_fiscal_year),\n column_width: 3,\n label: \"Fiscal Year\")\n end\n\n a += form_row do\n f.text_field(:supplier_name,\n column_width: 4,\n label: \"Supplier Name\") +\n f.text_field(:supplier_contact_person,\n column_width: 5,\n label: \"Contact Person\") +\n f.phone_field(:supplier_phone,\n column_width: 3,\n label: \"Phone Number\")\n end\n a += form_row do\n f.text_field(:supplier_address, column_width: 6, lstrip: \"Supplier\") +\n f.text_field(:supplier_city, column_width: 4, label: \"City/Town\") +\n f.text_field(:supplier_postal_code, column_width: 2, lstrip: \"Supplier\")\n end\n a += form_row do\n f.text_field(:item_desp_1, column_width: 6, label: \"Item\", placeholder: \"Item 1\") +\n f.currency_field(:item_cost_1, column_width: 2, label: \"Cost\", placeholder: \"Item Cost 1\")\n end\n a += form_row do\n f.text_field(:item_desp_2, column_width: 6, hide_label: true, placeholder: \"Item 2\") +\n f.currency_field(:item_cost_2, column_width: 2, hide_label: true, lstrip: \"Supplier\")\n end\n a + form_row do\n f.text_field(:item_desp_3, column_width: 6, hide_label: true, placeholder: \"Item 3\") +\n f.currency_field(:item_cost_3, column_width: 2, hide_label: true, lstrip: \"Supplier\") +\n f.currency_field(:item_total, column_width: 4, label: \"Total\", disabled: \"disabled\")\n end\n end\n end",
"def new\n @section_piece_param = SectionPieceParam.new\n @html_attributes = HtmlAttribute.all\n respond_to do |format|\n format.html # new.html.erb\n format.xml { render :xml => @section_piece_param }\n end\n end",
"def pullingListingAddress(buildingURL, listing)\n buildingDocContact = Nokogiri::HTML(open(URI.parse(buildingURL + \"/contact\")))\n buildingDocDetails = Nokogiri::HTML(open(URI.parse(buildingURL + \"/features\")))\n buildingDocImages = Nokogiri::HTML(open(URI.parse(buildingURL + \"/media\")))\n #REMEMBER TO ADD BROOKLYN INTO THIS!!!!!\n contactText = buildingDocContact.css(\"#community-contact-text\")\n #contactText.css(\"p\")[1].text\n #puts buildingDocImages.css(\".slides\").css(\"img\").length\n #puts buildingDocImages.css(\".slides\").css(\"img\")\n listing[:images] = []\n countInt = 0 #used for image count\n while (countInt < (buildingDocImages.css(\".slides\").css(\"img\").length)/2) do #too many photos so i cut in half\n listing[:images] << { origin_url: buildingDocImages.css(\".slides\").css(\"img\")[countInt][\"src\"] }\n countInt = countInt + 1\n end\n #retrieve_images(buildingDocImages, listing)\n #puts contactText.css(\"p\")[2].text\n #puts buildingDocContact.css(\".phone-number\")\n if listing[:description] = buildingDocDetails.css(\".row.feature\").css(\"p\").css(\"span\")[0] != nil\n listing[:description] = buildingDocDetails.css(\".row.feature\").css(\"p\").css(\"span\")[0].text\n end\n listing[:amenities] = []\n #listing[:amenities] <<\n #puts (buildingDocDetails.css(\".row.feature\").css(\"#li_cont2\")).length\n listing[:contact_name] = \"Bozzuto Management\"\n\n if contactText.css(\"p\")[1].text.include?(\"New York\")\n matchText = (/New York, /).match(contactText.css(\"p\")[1].text)\n #Save one match text as string to enable string manipulations\n matchTextString = matchText.to_s\n listing[:title] = (matchText.pre_match).to_s #Street Address, but the matching part is like \"New York, \"\n listing[:city_name] = ((/, /).match(matchTextString).pre_match).to_s\n listing[:state_name] = (((/ /).match((matchText.post_match).to_s)).pre_match).to_s\n listing[:zipcode] = (((/ /).match((matchText.post_match).to_s)).post_match).to_s\n end\n if contactText.css(\"p\")[1].text.include?(\"Hoboken\")\n #((/Hoboken, /) === (contactText.css(\"p\")[1].text))\n matchText = (/Hoboken, /).match(contactText.css(\"p\")[1].text)\n #Save one match text as string to enable string manipulations\n matchTextString = (matchText.pre_match).to_s\n listing[:title] = matchText.pre_match\n listing[:raw_neighborhood] = \"Hoboken\"\n listing[:city_name] = \"Jersey City\"\n listing[:state_name] = (((/ /).match((matchText.post_match).to_s)).pre_match).to_s\n listing[:zipcode] = (((/ /).match((matchText.post_match).to_s)).post_match).to_s\n end\n if contactText.css(\"p\")[1].text.include?(\"Brooklyn\")\n matchText = (/Brooklyn, /).match(contactText.css(\"p\")[1].text)\n #Save one match text as string to enable string manipulations\n matchTextString = (matchText.pre_match).to_s\n listing[:title] = matchText.pre_match\n listing[:city_name] = \"Brooklyn\"\n listing[:state_name] = (((/ /).match((matchText.post_match).to_s)).pre_match).to_s\n listing[:zipcode] = (((/ /).match((matchText.post_match).to_s)).post_match).to_s\n end\n listing\n end",
"def builder\n\n add_breadcrumb \"SOGR Capital Project Analyzer\"\n\n # Select the asset seed that they are allowed to build\n\n @asset_seed = []\n asset_class_name = Rails.application.config.asset_base_class_name == 'TransamAsset' ? 'TransitAsset' : Rails.application.config.asset_base_class_name\n\n asset_class_name.constantize.asset_seed_class_name.constantize.active.each do |seed|\n assets = asset_class_name.constantize.where(seed.class.to_s.underscore => seed)\n if assets.where(organization: @organization_list).count > 0\n @asset_seed << {id: seed.id, name: seed.to_s, orgs: @organization_list.select{|o| assets.where(organization_id: o).count > 0}}\n else\n @asset_seed << {id: seed.id, name: seed.to_s, orgs: []}\n end\n end\n\n @fiscal_years = get_fiscal_years(Date.today)\n @range_fiscal_years = ((1..14).to_a + (3..10).to_a.map{|x| x * 5}).map{|x| [\"#{x} years\", x-1]}\n @scenarios = Scenario.where(organization: current_user.viewable_organizations)\n @builder_proxy = BuilderProxy.new\n\n @has_locked_sogr_this_fiscal_year = CapitalPlanModule.joins(:capital_plan_module_type, :capital_plan).where(capital_plan_module_types: {name: ['Unconstrained Plan', 'Constrained Plan']}, capital_plans: {organization_id: @organization_list, fy_year: current_planning_year_year}).where('capital_plan_modules.completed_at IS NOT NULL').pluck('capital_plans.organization_id')\n\n if @organization_list.count == 1\n if @has_locked_sogr_this_fiscal_year && (@has_locked_sogr_this_fiscal_year.include? @organization_list.first)\n @fiscal_years = @fiscal_years[(@fiscal_years.index{|x| x[1]==current_planning_year_year}+1)..-1]\n end\n @builder_proxy.start_fy = current_planning_year_year\n else\n @has_sogr_project_org_list = CapitalProject.joins(:organization).where(organization_id: @organization_list).sogr.group(:organization_id).count\n end\n\n @message = \"Creating SOGR capital projects. This process might take a while.\"\n\n end",
"def parsed_description\n @parsed_description ||= Nokogiri::HTML(description) \n end",
"def run\n url = p params[:my_url]\n if url.include? \"damndelicious.net\"\n doc = Nokogiri::HTML(open(url))\n recipe = Recipe.create do |recipe|\n #TODO: I see a lot of recipe manipulation. Could you extract lines 15 to 19 and similar codebase below into a function? That would make this a lot smaller and easier to read.\n recipe.user_id = current_user.id\n recipe.title = doc.at_css(\"h1\").text\n recipe.description = doc.at_css(\"em\").text\n recipe.url = url\n recipe.image_url = doc.at('//img[@class=\"photo nopin pib-hover-img\"]/@src').to_s\n # recipe end\n end\n doc.css(\".ingredient\").each do |classes|\n Ingredient.create!(name: classes.text, recipe_id: Recipe.last.id)\n # instruction end\n end\n doc.css(\".instructions li\").each do |classes|\n Direction.create!(step: classes.text, recipe_id: Recipe.last.id)\n # direction end\n end\n redirect_to recipe, notice: \"Scraped Recipe, please check to verify everything looks correct\"\n elsif url.include? \"thepioneerwoman.com\"\n doc = Nokogiri::HTML(open(url))\n recipe = Recipe.create do |recipe|\n recipe.user_id = current_user.id\n recipe.title = doc.at_css(\".recipe-title\").text\n # recipe.description = doc.at_css(\".col-xs-7 p\").text\n recipe.image_url = doc.at('//img[@class=\"alignnone size-full wp-image-91195\"]/@src').to_s\n recipe.url = url\n # recipe end\n end\n doc.css(\".list-ingredients li\").each do |classes|\n Ingredient.create!(name: classes.text, recipe_id: Recipe.last.id)\n # instruction end\n end\n doc.css(\".panel+ .panel .panel-body\").each do |classes|\n Direction.create!(step: classes.text, recipe_id: Recipe.last.id)\n # direction end\n end\n redirect_to recipe, notice: \"Scraped Recipe, please check to verify everything looks correct\"\n elsif url.include? \"allrecipes.com\"\n doc = Nokogiri::HTML(open(url))\n recipe = Recipe.create do |recipe|\n recipe.user_id = current_user.id\n recipe.title = doc.at_css(\".recipe-summary__h1\").text\n recipe.description = doc.at_css(\".submitter__description\").text\n recipe.url = url\n recipe.image_url = doc.at('//img[@class=\"rec-photo\"]/@src').to_s\n\n # recipe end\n end\n doc.css(\".added\").each do |classes|\n Ingredient.create!(name: classes.text, recipe_id: Recipe.last.id)\n # instruction end\n end\n doc.css(\".step\").each do |classes|\n Direction.create!(step: classes.text, recipe_id: Recipe.last.id)\n # direction end\n end\n redirect_to recipe, notice: \"Scraped Recipe, please check to verify everything looks correct\"\n else\n redirect_to new_recipe_url, alert: \"the domain your trying to access is not yet supported\"\n # if statement end\n end\n # Def run end\n end",
"def build_provisions(run_directory, data_directory, whitelist, blacklist)\n \n Scarcity::Submission.new do\n \n # declare where runs happen and where data comes from\n # and indicate that we gather default provisions accordingly \n # This will not need to change in almost all cases.\n runs_in run_directory\n pulls_from data_directory, :only => whitelist, :except => blacklist\n gathers_provisions :zip_data => true\n \n # Declare other goods that will be provided to each dataset at run time.\n provides :from => 'app/executables', :to => :each_dataset do\n file '<%= base_name %>.py', :chmod => 0755\n end\n \n provides :from => 'app/scripts', :to => :each_dataset do\n file 'prejob.py', :chmod => 0755\n file 'postjob.py', :chmod => 0755\n end\n \n provides :from => 'app/submits', :to => :each_dataset do\n file '<%= base_name %>.submit'\n end\n \n provides :from => 'lib/submits', :to => :each_dataset do\n file 'null.submit'\n end\n \n # You might also declare goods to be provided to the segment as a whole\n # provides :from => 'lib/submits', :to => :segment do\n # file 'null.submit'\n # end\n \n end\n \nend",
"def setup\n @agent = Mechanize.new\n uri = URI.join('file:///', File.absolute_path(\"test/data/view.html\"))\n @page = @agent.get uri.to_s\n @view = PortfolioView.new uri.to_s, @page, \"test mameber\", \"test portfolio\", \"View1\", \"testgroup\"\n\n @analyzer = PortfolioStatistics::DescriptiveViewAnalyzer.new\n end",
"def build_source(feature)\n # clean up redmine text formatting\n description = feature.description.gsub(/<\\/?code[^>]*>/ui,'').gsub(/<\\/?pre[^>]*>/ui,'\"\"\"').gsub(/^$\\n/, '')\n subject = \"Feature: #{feature.subject}\\n\"\n \"#{@tags.join(' ') + \"\\n\" if @tags}#{subject}\\n #{description}\\n\"\n end",
"def process_markup()\n [title,content].each do |text|\n next if !text\n clear_webpage(text).scan(PTN_MARKUP).map{|e|e[0].split(PTN_ELEMENT_SEP)}.flatten.each do |element|\n #debug \"[process_markup] element: \"+element.inspect\n next if !element\n ptn = element.scan(PTN_METADATA)\n #debug \"[process_markup] ptn: \"+ptn.inspect\n if ptn.size > 0\n m[ptn[0][0].to_sym] = parse_value(ptn[0][1])\n else\n create_and_link(element, 'tag' , \"t\")\n #add_tags(element, \"m\")\n end\n end#scan\n end#each\n end",
"def parse\n parts = text.split(/^\\-\\-\\-/)\n\n if parts.size == 1\n data = {}\n #@parts << Part.new(sect[0]) #, *[@stencil, @default].compact.flatten)\n else\n parts.shift if parts.first.strip.empty?\n data = YAML::load(parts.first)\n if Hash === data\n parts.shift\n else\n data = {}\n end\n end\n\n parse_metadata(data)\n\n @parts = parts.map{ |part| Part.parse(self, part) }\n end",
"def scrape_medium_story_create_inspirations\n url = \"https://medium.com/topic/editors-picks\"\n html_file = open(url).read\n html_doc = Nokogiri::HTML(html_file)\n\n # browser = Watir::Browser.new\n # browser.goto 'https://medium.com/topic/editors-picks'\n\n # html_doc = Nokogiri::HTML.parse(browser.html)\n\n\n html_doc.xpath(\"//div[contains(concat(' ',normalize-space(@class)), 'l er q s es bz et eu ev v')]\").each do |card|\n #ai y cl bj cm bk dq ez fa ak an ds cr gk gl\n article_name = card.search('.ai.y.ce.at.cf.au.di.ex.ey.ak.an.dk.ck.cl.am').first.children.first.children.text unless card.search('.ai.y.ce.at.cf.au.di.ex.ey.ak.an.dk.ck.cl.am').empty?\n article_link = card.xpath(\".//div[contains(concat(' ',normalize-space(@class)), 'ew d ev v cc')]/div/a/@href\").first.value\n duration_text = card.xpath(\".//div[contains(concat(' ',normalize-space(@class)), 'ec s ed')]\").first.text\n article_duration = duration_text.gsub(/(\\D+\\s)(\\d+.)(\\d+)(.*)/, '\\3').to_i\n Inspiration.create(\n inspiration_type: 'article',\n source: 'Medium',\n duration: article_duration + 1, # because Medium round it below ?\n name: article_name,\n url: article_link\n )\n end\n\n end",
"def parse\n products = []\n agent = WWW::Mechanize.new\n\n self.categories.each do |category|\n page = agent.get category.url\n page.search(\"//table[@width=175]\").each do |table|\n p = Product.new\n p.name = table.children[1].children[0].text\n p.price = table.children[3].children[2].text\n p.model = table.children[4].children[0].text\n str = table.children[5].text.gsub(/\\n|\\t|'/, '').strip.squeeze(\" \") #this position can hold the warranty info or description\n if str.include? 'Warranty'\n p.warranty = str\n else\n p.description = str\n p.warranty = table.children[6].text.gsub(/\\n|\\t|'/, '').strip.squeeze(\" \")\n end\n p.tag = category.name\n products << p\n end\n end\n\n products\n end",
"def to_html\n out = \"<p>\\n\"\n self.techniques.each do |t| \n tmp = t.is_a?(Ingredient) ? t.to_html : \"#{CGI.escapeHTML(t.chomp)}\"\n out << tmp\n end\n out << \"\\n</p>\\n\"\n out\n end",
"def setup\n # create a page_part containing a part storing the gallery\n new_page = Page.new(:title => 'Test Gallery', :slug => 'gallery-test', :breadcrumb => 'gallery', :behavior_id => 'Gallery')\n new_page.save\n @page_part_gallery= PagePart.new(:name => 'gallery', \n :content => 'test-gallery',\n :page_id => new_page.id)\n @page_part_gallery.save\n @page_part_body = PagePart.new(:name => 'body', \n :content => '<r:gallery/>',\n :page_id => new_page.id)\n @page_part_body.save\n @page = Page.find_by_id(new_page.id)\n end",
"def get_html_template(passed, failed, env, view_build_link, view_allure_link)\n file = File.read('features/data/email_template.html')\n email_template = Nokogiri::HTML.fragment(file)\n email_template.at_css('#passed').content += passed.to_s\n email_template.at_css('#failed').content += failed.to_s\n email_template.at_css('#env').content += env\n email_template.at_css(\"#circleci\")['href'] = view_build_link\n email_template.at_css(\"#allure\")['href'] = view_allure_link\n email_template.to_html\nend",
"def process_specification_page(response_body, short_dataset_only=true)\n hash = {}\n\n if !response_body.nil?\n doc = Nokogiri.HTML(response_body)\n\n # msrp\n hash['msrp'] = nil\n x = doc.css('#style-price a').first\n hash['msrp'] = x.text.gsub('$', '').gsub(',', '').to_i if !x.nil? && x.text.length > 0\n\n get_specifications_value(doc, hash, 'style', 'Style Name')\n get_specifications_value(doc, hash, 'body_style', 'Body Style')\n get_specifications_value(doc, hash, 'passenger_doors', 'Passenger Doors', integer: true)\n get_specifications_value(doc, hash, 'transmission', 'Transmission')\n\n # DIMENSIONS\n get_specifications_value(doc, hash, 'epa_classification', 'EPA Classification')\n get_specifications_value(doc, hash, 'passenger_capacity', 'Passenger Capacity', integer: true)\n get_specifications_value(doc, hash, 'front_head_room', 'Front Head Room (in)', float: true)\n get_specifications_value(doc, hash, 'front_shoulder_room', 'Front Shoulder Room (in)', float: true)\n get_specifications_value(doc, hash, 'front_hip_room', 'Front Hip Room (in)', float: true)\n get_specifications_value(doc, hash, 'front_leg_room', 'Front Leg Room (in)', float: true)\n get_specifications_value(doc, hash, 'second_head_room', 'Second Head Room (in)', float: true)\n get_specifications_value(doc, hash, 'second_shoulder_room', 'Second Shoulder Room (in)', float: true)\n get_specifications_value(doc, hash, 'second_hip_room', 'Second Hip Room (in)', float: true)\n get_specifications_value(doc, hash, 'second_leg_room', 'Second Leg Room (in)', float: true)\n get_specifications_value(doc, hash, 'third_head_room', 'Third Head Room (in)', float: true)\n get_specifications_value(doc, hash, 'third_shoulder_room', 'Third Shoulder Room (in)', float: true)\n get_specifications_value(doc, hash, 'third_hip_room', 'Third Hip Room (in)', float: true)\n get_specifications_value(doc, hash, 'third_leg_room', 'Third Leg Room (in)', float: true)\n\n get_specifications_value(doc, hash, 'width', 'Width, Max w/o mirrors (in)', float: true)\n get_specifications_value(doc, hash, 'height', 'Height, Overall (in)', float: true)\n get_specifications_value(doc, hash, 'length', 'Length, Overall (in)', float: true)\n get_specifications_value(doc, hash, 'ground_clearance', 'Min Ground Clearance (in)', float: true)\n\n get_specifications_value(doc, hash, 'trunk_volume', 'Trunk Volume (ft³)', float: true)\n get_specifications_value(doc, hash, 'cargo_area_width', 'Cargo Box Width @ Wheelhousings (in)', float: true)\n get_specifications_value(doc, hash, 'cargo_area_height', 'Cargo Box (Area) Height (in)', float: true)\n get_specifications_value(doc, hash, 'cargo_area_length_floor_to_seat2', 'Cargo Area Length @ Floor to Seat 2 (in)', float: true)\n get_specifications_value(doc, hash, 'cargo_area_volume_to_seat1', 'Cargo Volume to Seat 1 (ft³)', float: true)\n get_specifications_value(doc, hash, 'cargo_area_volume_to_seat2', 'Cargo Volume to Seat 2 (ft³)', float: true)\n get_specifications_value(doc, hash, 'cargo_are_volume_to_seat3', 'Cargo Volume to Seat 3 (ft³)', float: true)\n\n\n # FUEL ECONOMY\n get_specifications_value(doc, hash, 'fuel_tank_capacity', 'Fuel Tank Capacity, Approx (gal)', integer: true)\n get_specifications_value(doc, hash, 'mpg_city', 'EPA Fuel Economy Est - City (MPG)', integer: true)\n get_specifications_value(doc, hash, 'mpg_highway', 'EPA Fuel Economy Est - Hwy (MPG)', integer: true)\n get_specifications_value(doc, hash, 'mpg_combined', 'Fuel Economy Est-Combined (MPG)', integer: true)\n get_specifications_value(doc, hash, 'battery_range', 'Battery Range (mi)')\n get_specifications_value(doc, hash, 'mpg_equivalent_city', 'EPA MPG Equivalent - City')\n get_specifications_value(doc, hash, 'mpg_equivalent_hwy', 'EPA MPG Equivalent - Hwy')\n get_specifications_value(doc, hash, 'mpg_equivalent_combined', 'EPA MPG Equivalent - Combined')\n\n\n # PERFORMANCE SPECS\n get_specifications_value(doc, hash, 'engine_type', 'Engine Type')\n get_specifications_value(doc, hash, 'engine_displacement', 'Displacement')\n get_specifications_value(doc, hash, 'horsepower', 'SAE Net Horsepower @ RPM')\n get_specifications_value(doc, hash, 'fuel_system', 'Fuel System')\n\n get_specifications_value(doc, hash, 'brakes_abs', 'Brake ABS System')\n get_specifications_value(doc, hash, 'brakes_disc_front', 'Disc - Front (Yes or )')\n get_specifications_value(doc, hash, 'brakes_disc_rear', 'Disc - Rear (Yes or )')\n get_specifications_value(doc, hash, 'brakes_drum_rear', 'Drum - Rear (Yes or )')\n\n get_specifications_value(doc, hash, 'steering_type', 'Steering Type')\n get_specifications_value(doc, hash, 'turning_diameter', 'Turning Diameter - Curb to Curb (ft)', float: true)\n\n\n # SAFETY FEATURES\n get_specifications_value(doc, hash, 'air_bag_front_driver', 'Air Bag-Frontal-Driver')\n get_specifications_value(doc, hash, 'air_bag_front_passenger', 'Air Bag-Frontal-Passenger')\n get_specifications_value(doc, hash, 'air_bag_front_passenger_switch', 'Air Bag-Passenger Switch (On/Off)')\n get_specifications_value(doc, hash, 'air_bag_side_head_front', 'Air Bag-Side Head-Front')\n get_specifications_value(doc, hash, 'air_bag_side_body_front', 'Air Bag-Side Body-Front')\n get_specifications_value(doc, hash, 'air_bag_side_head_rear', 'Air Bag-Side Head-Rear')\n get_specifications_value(doc, hash, 'air_bag_side_body_rear', 'Air Bag-Side Body-Rear')\n get_specifications_value(doc, hash, 'brakes_abs', 'Brakes-ABS')\n get_specifications_value(doc, hash, 'child_door_locks', 'Child Safety Rear Door Locks')\n get_specifications_value(doc, hash, 'other_features', 'Other Features')\n\n\n\n # if the entire dataset is desired, continue\n if !short_dataset_only\n\n # DIMENSIONS\n get_specifications_value(doc, hash, 'base_curb_weight', 'Base Curb Weight (lbs)', integer: true)\n get_specifications_value(doc, hash, 'passenger_volume', 'Passenger Volume (ft³)', float: true)\n get_specifications_value(doc, hash, '', 'Gross Combined Wt Rating (lbs)')\n get_specifications_value(doc, hash, '', 'Curb Weight - Front (lbs)')\n get_specifications_value(doc, hash, '', 'Curb Weight - Rear (lbs)')\n get_specifications_value(doc, hash, '', 'Gross Axle Wt Rating - Front (lbs)')\n get_specifications_value(doc, hash, '', 'Gross Axle Wt Rating - Rear (lbs)')\n get_specifications_value(doc, hash, '', 'Gross Vehicle Weight Rating Cap (lbs)')\n\n get_specifications_value(doc, hash, 'wheelbase', 'Wheelbase (in)', float: true)\n get_specifications_value(doc, hash, 'track_width_front', 'Track Width, Front (in)', float: true)\n get_specifications_value(doc, hash, 'track_width_rear', 'Track Width, Rear (in)', float: true)\n get_specifications_value(doc, hash, 'liftover_height', 'Liftover Height (in)', float: true)\n get_specifications_value(doc, hash, '', 'Ground Clearance, Front (in)')\n get_specifications_value(doc, hash, '', 'Overhang, Rear w/o bumper (in)')\n get_specifications_value(doc, hash, '', 'Ground to Top of Frame (in)')\n get_specifications_value(doc, hash, '', 'Cab to End of Frame (in)')\n get_specifications_value(doc, hash, '', 'Ground Clearance, Rear (in)')\n get_specifications_value(doc, hash, '', 'Length, Overall w/o rear bumper (in)')\n get_specifications_value(doc, hash, '', 'Front Bumper to Back of Cab (in)')\n get_specifications_value(doc, hash, '', 'Frame Width, Rear (in)')\n get_specifications_value(doc, hash, '', 'Overhang, Front (in)')\n get_specifications_value(doc, hash, '', 'Ground to Top of Load Floor (in)')\n get_specifications_value(doc, hash, '', 'Cab to Axle (in)')\n get_specifications_value(doc, hash, '', 'Rear Door Type')\n get_specifications_value(doc, hash, '', 'Rear Door Opening Height (in)')\n get_specifications_value(doc, hash, '', 'Step Up Height - Side (in)')\n get_specifications_value(doc, hash, '', 'Side Door Opening Width (in)')\n get_specifications_value(doc, hash, '', 'Overhang, Rear w/bumper (in)')\n get_specifications_value(doc, hash, '', 'Rear Door Opening Width (in)')\n get_specifications_value(doc, hash, '', 'Step Up Height - Front (in)')\n get_specifications_value(doc, hash, '', 'Length, Overall w/rear bumper (in)')\n get_specifications_value(doc, hash, '', 'Side Door Opening Height (in)')\n\n\n get_specifications_value(doc, hash, '', 'Cargo Box Length @ Floor (in)')\n get_specifications_value(doc, hash, '', 'Cargo Box Width @ Floor (in)')\n get_specifications_value(doc, hash, '', 'Cargo Box Width @ Top, Rear (in)')\n get_specifications_value(doc, hash, '', 'Cargo Volume (ft³)')\n get_specifications_value(doc, hash, '', 'Ext\\'d Cab Cargo Volume (ft³)')\n get_specifications_value(doc, hash, '', 'Cargo Area Width @ Beltline (in)')\n get_specifications_value(doc, hash, '', 'Cargo Area Length @ Floor to Seat 1 (in)')\n get_specifications_value(doc, hash, '', 'Tailgate Width (in)')\n get_specifications_value(doc, hash, '', 'Cargo Area Length @ Floor to Seat 4 (in)')\n get_specifications_value(doc, hash, '', 'Cargo Area Length @ Floor to Console (in)')\n get_specifications_value(doc, hash, '', 'Cargo Area Length @ Floor to Seat 3 (in)')\n get_specifications_value(doc, hash, '', 'Cargo Volume to Seat 4 (ft³)')\n get_specifications_value(doc, hash, '', 'Cargo Volume with Rear Seat Up (ft³)')\n get_specifications_value(doc, hash, '', 'Cargo Volume with Rear Seat Down (ft³)')\n\n # PERFORMANCE SPECS\n get_specifications_value(doc, hash, 'torque', 'SAE Net Torque @ RPM')\n get_specifications_value(doc, hash, 'engine_order_code', 'Engine Order Code')\n\n get_specifications_value(doc, hash, '', 'Aux Fuel Tank Location')\n get_specifications_value(doc, hash, '', 'Aux Fuel Tank Capacity, Approx (gal)')\n get_specifications_value(doc, hash, '', 'Fuel Tank Location')\n\n get_specifications_value(doc, hash, '', 'Engine Oil Cooler')\n\n get_specifications_value(doc, hash, 'drivetrain', 'Drivetrain')\n get_specifications_value(doc, hash, 'first_gear_ratio', 'First Gear Ratio (:1)')\n get_specifications_value(doc, hash, '', 'Second Gear Ratio (:1)')\n get_specifications_value(doc, hash, '', 'Third Gear Ratio (:1)')\n get_specifications_value(doc, hash, '', 'Fourth Gear Ratio (:1)')\n get_specifications_value(doc, hash, '', 'Fifth Gear Ratio (:1)')\n get_specifications_value(doc, hash, '', 'Sixth Gear Ratio (:1)')\n get_specifications_value(doc, hash, '', 'Seventh Gear Ratio (:1)')\n get_specifications_value(doc, hash, '', 'Eighth Gear Ratio (:1)')\n get_specifications_value(doc, hash, '', 'Clutch Size (in)')\n get_specifications_value(doc, hash, '', 'Transfer Case Gear Ratio (:1), Low')\n get_specifications_value(doc, hash, '', 'Transfer Case Model')\n get_specifications_value(doc, hash, '', 'Trans Power Take Off')\n get_specifications_value(doc, hash, '', 'Transfer Case Power Take Off')\n get_specifications_value(doc, hash, '', 'Transfer Case Gear Ratio (:1), High')\n get_specifications_value(doc, hash, '', 'Trans PTO Access')\n\n get_specifications_value(doc, hash, '', 'Axle Type - Front')\n get_specifications_value(doc, hash, '', 'Axle Capacity - Front (lbs)')\n get_specifications_value(doc, hash, '', 'Axle Ratio (:1) - Front')\n get_specifications_value(doc, hash, '', 'Axle Type - Rear')\n get_specifications_value(doc, hash, '', 'Axle Ratio (:1) - Rear')\n get_specifications_value(doc, hash, '', 'Axle Capacity - Rear (lbs)')\n get_specifications_value(doc, hash, '', 'Spring Capacity - Front (lbs)')\n get_specifications_value(doc, hash, '', 'Shock Absorber Diameter - Front (mm)')\n get_specifications_value(doc, hash, '', 'Stabilizer Bar Diameter - Front (in)')\n get_specifications_value(doc, hash, '', 'Spring Capacity - Rear (lbs)')\n get_specifications_value(doc, hash, '', 'Shock Absorber Diameter - Rear (mm)')\n get_specifications_value(doc, hash, '', 'Stabilizer Bar Diameter - Rear (in)')\n\n get_specifications_value(doc, hash, 'reverse_ratio', 'Reverse Ratio (:1)', float: true)\n get_specifications_value(doc, hash, 'final_drive_axle_ratio', 'Final Drive Axle Ratio (:1)', float: true)\n get_specifications_value(doc, hash, 'trans_type', 'Trans Type')\n get_specifications_value(doc, hash, 'trans_desc_cont', 'Trans Description Cont.')\n get_specifications_value(doc, hash, 'trans_desc_cont2', 'Trans Description Cont. Again')\n get_specifications_value(doc, hash, 'trans_order_code', 'Trans Order Code')\n\n get_specifications_value(doc, hash, 'brakes_front_drum_thickness', 'Front Brake Rotor Diam x Thickness (in)', float: true)\n get_specifications_value(doc, hash, 'brakes_rear_drum_thickness', 'Rear Brake Rotor Diam x Thickness (in)', float: true)\n get_specifications_value(doc, hash, 'brakes_rear_drum_width', 'Rear Drum Diam x Width (in)', float: true)\n get_specifications_value(doc, hash, '', 'Brake Type')\n get_specifications_value(doc, hash, '', 'Brake ABS System (Second Line)')\n\n get_specifications_value(doc, hash, '', 'Steering Ratio (:1), On Center')\n get_specifications_value(doc, hash, '', 'Turning Diameter - Wall to Wall (ft)')\n get_specifications_value(doc, hash, '', 'Steering Ratio (:1), At Lock')\n\n get_specifications_value(doc, hash, '', 'Revolutions/Mile @ 45 mph - Rear')\n get_specifications_value(doc, hash, '', 'Spare Tire Capacity (lbs)')\n get_specifications_value(doc, hash, '', 'Front Tire Capacity (lbs)')\n get_specifications_value(doc, hash, '', 'Revolutions/Mile @ 45 mph - Spare')\n get_specifications_value(doc, hash, '', 'Revolutions/Mile @ 45 mph - Front')\n get_specifications_value(doc, hash, '', 'Rear Tire Capacity (lbs)')\n\n get_specifications_value(doc, hash, 'tire_front_size', 'Front Tire Size')\n get_specifications_value(doc, hash, 'tire_front_code', 'Front Tire Order Code')\n get_specifications_value(doc, hash, 'tire_rear_size', 'Rear Tire Size')\n get_specifications_value(doc, hash, 'tire_rear_code', 'Rear Tire Order Code')\n get_specifications_value(doc, hash, 'tire_spare_size', 'Spare Tire Size')\n get_specifications_value(doc, hash, 'tire_spare_code', 'Spare Tire Order Code')\n\n get_specifications_value(doc, hash, 'wheel_front_size', 'Front Wheel Size (in)')\n get_specifications_value(doc, hash, 'wheel_front_material', 'Front Wheel Material')\n get_specifications_value(doc, hash, 'wheel_rear_size', 'Rear Wheel Size (in)')\n get_specifications_value(doc, hash, 'wheel_rear_material', 'Rear Wheel Material')\n get_specifications_value(doc, hash, 'wheel_spare_size', 'Spare Wheel Size (in)')\n get_specifications_value(doc, hash, 'wheel_spare_material', 'Spare Wheel Material')\n\n get_specifications_value(doc, hash, 'suspension_type_front', 'Suspension Type - Front')\n get_specifications_value(doc, hash, 'suspension_type_front2', 'Suspension Type - Front (Cont.)')\n get_specifications_value(doc, hash, 'suspension_type_rear', 'Suspension Type - Rear')\n get_specifications_value(doc, hash, 'suspension_type_rear2', 'Suspension Type - Rear (Cont.)')\n\n\n # SAFETY FEATURES\n get_specifications_value(doc, hash, 'daytime_lights', 'Daytime Running Lights')\n get_specifications_value(doc, hash, 'fog_lamps', 'Fog Lamps')\n get_specifications_value(doc, hash, 'night_vision', 'Night Vision')\n get_specifications_value(doc, hash, 'backup_camera', 'Back-Up Camera')\n get_specifications_value(doc, hash, 'parking_aid', 'Parking Aid')\n get_specifications_value(doc, hash, 'traction_control', 'Traction Control')\n get_specifications_value(doc, hash, 'tire_pressure_monitor', 'Tire Pressure Monitor')\n get_specifications_value(doc, hash, 'stability_control', 'Stability Control')\n get_specifications_value(doc, hash, 'rollover_protection_bars', 'Rollover Protection Bars')\n\n\n # WARRANTY\n get_specifications_value(doc, hash, 'warranty_years', 'Basic Years', integer: true)\n get_specifications_value(doc, hash, 'warranty_miles', 'Basic Miles/km')\n get_specifications_value(doc, hash, 'warranty_drivetrain_year', 'Drivetrain Years', integer: true)\n get_specifications_value(doc, hash, 'warranty_drivetrain_mils', 'Drivetrain Miles/km')\n get_specifications_value(doc, hash, 'warranty_corrosion_years', 'Corrosion Years', integer: true)\n get_specifications_value(doc, hash, 'warranty_corrosion_miles', 'Corrosion Miles/km')\n get_specifications_value(doc, hash, 'warranty_roadside_years', 'Roadside Assistance Years', integer: true)\n get_specifications_value(doc, hash, 'warranty_roadside_miles', 'Roadside Assistance Miles/km')\n get_specifications_value(doc, hash, '', 'Hybrid/Electric Components Miles/km')\n get_specifications_value(doc, hash, '', 'Hybrid/Electric Components Years')\n get_specifications_value(doc, hash, '', 'Maintenance Miles/km')\n get_specifications_value(doc, hash, '', 'Maintenance Years')\n get_specifications_value(doc, hash, '', 'Drivetrain Note')\n get_specifications_value(doc, hash, '', 'Maintenance Note')\n get_specifications_value(doc, hash, '', 'Roadside Assistance Note')\n get_specifications_value(doc, hash, '', 'Emissions Miles/km')\n get_specifications_value(doc, hash, '', 'Emissions Years')\n\n\n\n # OTHER SPECS\n get_specifications_value(doc, hash, 'cold_cranking_amps', 'Cold Cranking Amps @ 0° F (Primary)', integer: true)\n get_specifications_value(doc, hash, '', 'Total Cooling System Capacity (qts)')\n get_specifications_value(doc, hash, '', 'Maximum Alternator Watts')\n get_specifications_value(doc, hash, '', 'Cold Cranking Amps @ 0° F (2nd)')\n get_specifications_value(doc, hash, 'max_alternator_capacity', 'Maximum Alternator Capacity (amps)', integer: true)\n get_specifications_value(doc, hash, 'max_trailering_capacity', 'Maximum Trailering Capacity (lbs)', integer: true)\n get_specifications_value(doc, hash, 'max_trailer_weight_distributing_hitch', 'Wt Distributing Hitch - Max Trailer Wt. (lbs)', integer: true)\n get_specifications_value(doc, hash, 'max_tongue_weight_distributing_hitch', 'Wt Distributing Hitch - Max Tongue Wt. (lbs)', integer: true)\n get_specifications_value(doc, hash, 'max_trailer_weight_dead_weight_hitch', 'Dead Weight Hitch - Max Trailer Wt. (lbs)', integer: true)\n get_specifications_value(doc, hash, 'max_tongue_weight_dead_weight_hitch', 'Dead Weight Hitch - Max Tongue Wt. (lbs)', integer: true)\n get_specifications_value(doc, hash, '', 'Fifth Wheel Hitch - Max Tongue Wt. (lbs)')\n get_specifications_value(doc, hash, '', 'Fifth Wheel Hitch - Max Trailer Wt. (lbs)')\n get_specifications_value(doc, hash, '', 'Wt Distributing Hitch - Max Trailer Wt. (lbs)')\n\n end\n\n end\n\n return hash\nend",
"def parse_data(page)\n if $debug\n puts 'Opening local file.'\n begin\n f = File.open(\"html/research-it.html\")\n doc = Nokogiri::HTML(f)\n f.close\n rescue\n puts \"Failed to open research-it.html file.\"\n end\n else\n puts \"Opening: #{page}\"\n doc = Nokogiri::HTML(open(page))\n end\n\n doc.xpath('//h3/a').each do |record|\n url = record['href']\n name = record.content\n $lessons[url] = {}\n $lessons[url]['name'] = name\n page = Nokogiri::HTML(open(url))\n desc = page.xpath('//p')[1].content\n if desc == ''\n $lessons[url]['short_description'] = 'No description available at this time.'\n else\n $lessons[url]['short_description'] = desc\n end\n end\n\nend",
"def setup_recipe url\n # In practice, grammar mods will get bound to the site\n # The selector will get associated with the recipe's site (as a 'Content' finder)\n # The trimmers will kept on the site as well, to remove extraneous elements\n # The grammar_mods will get applied to the parser's grammar for site-specific modification\n @recipe = load_recipe url, @selector, @trimmers, @grammar_mods\n @page_ref = @recipe.page_ref\n @recipe_page = @page_ref.recipe_page\nend",
"def build #title, form_name, form_action, script='', append_body='')\n \n @disable = []\n \n xhtml = %Q{ \n <html> \n <head>\n <title>#{self.meta['title']}</title>\n }\n if self.meta['stylesheets']\n self.meta['stylesheets'].each { |css|\n xhtml += %Q{ <link type=\"text/css\" src=\"#{css}\" /> \\n }\n }\n end\n if self.meta['javascripts']\n self.meta['javascripts'].each { |js|\n xhtml += %Q{ <script langauge=\"javascript\" src=\"#{jc}\" /> \\n }\n }\n end\n #xhtml += %Q{\n # <style type=\"text/css\">\n # #{build_css(self.style)}\n # </style>\n #}\n xhtml += %Q{ </head> }\n xhtml += %Q{\n <body onload=\"javascript: loadScript();\">\n #{self.meta['prepend']}\n #{self.main.build}\n #{self.meta['append']}\n </body>\n </html>\n }\n return CGI::pretty(xhtml)\n #return PrettyXML.pretty(html, 2)\n end",
"def new\n @thng = Thng.new\n @frqtags_from_db= Frqtag.find :all\n @frqtags3= @frqtags_from_db.map{ |t| \"<option>#{t.tgnm}</option>\" }.to_s\n @frqtags1= \"<option>My thing</option> \" + @frqtags3\n @frqtags2= \"<option>No 2nd tag needed</option> \" + @frqtags3\n\n # Fill objects with blanks so I can share code among haml files.\n @tagnms = [1,2,3,4].map{ |i| \"\"}\n @hhrefs = [1,2,3,4].map{ |i| \"\"}\n @imgsrcs = [1,2,3,4].map{ |i| \"\"}\n @videohtmls= [1,2,3,4].map{ |i| \"\"}\n end",
"def build_listing_array(all_listings)\n parsed_listings_array = []\n all_listings.map do |val, info|\n parsed_listings_array <<\n {\n \"listing_class\" => val.css('div.details-title a')[0]['data-gtm-listing-type'],\n \"address\" => val.css('.details-title>a').text.split('#')[0].gsub(/\\s+$/,''),\n \"unit\" => val.css('.details-title>a')[0].text.partition('#').last,\n \"url\" => \"www.streeteasy.com/\" + val.css('.details-title>a')[0]['href'],\n \"price\" => val.css('span.price').text.delete('$ ,').to_i\n }\n end\n parsed_listings_array\n end",
"def xml_builder\n lambda do |builder|\n builder[:api].amend do\n builder[:api].requests do\n Array.wrap(amendments).each do |amendment|\n build_object builder, :amendments, amendment, :obj\n end\n build_object builder, :amend_options, amend_options, :api\n build_object builder, :preview_options, preview_options, :api\n end\n end\n end\n end",
"def apartment_description\n apartment_description = {}\n apartment_description.merge!(APARTMENT_DESCRIPTION)\n apartment_description.keys.each { |key| apartment_description[key] = send(apartment_description[key]) }\n apartment_description\n end",
"def pull_emilys_event_data(event_website) #event_website must be a string\n\n event_title = ScrapeEventURLs.create(event_website, \".bsd-contribForm-aboveContent/h1\")[0][7..-8]\n\n description_raw, date_location_raw = ScrapeEventURLs.create(event_website, \".bsd-contribForm-aboveContent/p\")[0..1]\n description = description_raw.gsub(\"<p>\", \"\").gsub(\"</p>\", \"\").gsub(\"\\r\", \"\").gsub(\"\\n\", \"\")\n\n date_times_raw, location_raw = date_location_raw.split(\"<br><br>\")\n stripped_date_times_raw = date_times_raw.gsub(\"\\n\", \"\").gsub(\"\\r\", \"\").gsub(\"<p>\", \"\")\n date_raw, both_times = stripped_date_times_raw.split(\"<br>\")\n start_time_raw, end_time_raw = both_times.split(\" - \")\n event_location = location_raw.gsub(\"\\r\\n\", \"\").gsub(\"<br>\", \" \").gsub(\"</p>\", \"\")\n\n\n #This block transforms intermediate date, times, and location into start_time and end_time strings that include date, 24time, and timezone\n date_string = FormatDateTime.date_string(date_raw)\n start_time, end_time = FormatDateTime.time_string(start_time_raw, end_time_raw)\n timezone = \"+0000\" #using UTC as default for now because DetermineTimezone.zone(event_location) is not yet working\n start_time = \"#{date_string} #{start_time} #{timezone}\"\n end_time = \"#{date_string} #{end_time} #{timezone}\"\n\n\n #These 2 categories can't yet be determined from url data, so they are given assumptions based on knowledge of Emilys List events in general\n free = false\n cta_type = \"onsite\"\n\n parsed_event = CreateJsonObject.create_json_object(event_title, description, free, start_time, end_time, cta_type, event_website)\n [parsed_event, event_location]\nend",
"def populate_offers_living_social\n testing = \"true\"\n if testing != \"true\"\n division_request = \"http://www.livingsocial.com/services/city/v2/cities\"\n @division_response = `curl \"#{division_request}\"`\n @division_response = JSON.parse(@division_response)\n else\n @division_response = {\"divisions\"=>[{\"id\"=>\"26\"}, {\"id\"=>\"864\"}]}\n end\n\n @division_response['divisions'].each do |division|\n @division_deals_request = \"http://monocle.livingsocial.com/v2/deals?city=#{URI::escape(division['id'])}&api-key=2574AD58578A419596D95D4D0549A9CF&full=1\"\n @division_deals_response = `curl \"#{@division_deals_request}\"`\n @division_deals_response = JSON.parse(@division_deals_response)\n\n @division_deals_response['deals'].each do |deal|\n if deal['sold_out'].to_s.match(/false/i)\n @offer = Offer.find_by_deal_url deal['url']\n @offer.destroy if @offer.present?\n @offer = Offer.new\n @offer.deal_id = deal['id']\n @offer.deal_end = deal['offer_ends_at']\n @offer.deal_start = deal['offer_starts_at']\n @offer.deal_source = \"livingsocial\"\n @offer.merchant_type = ''\n @offer.deal_header = deal['long_title']\n @offer.merchant_name = deal['merchant_name']\n\n @offer.division_id = division['id']\n @offer.large_image_url = deal['images'][0]['size220']\n @offer.status = \"open\"\n @offer.deal_url = deal['url']\n @offer.redemption_location = deal['locations'][0].present? ? (deal['locations'][0]['city'] ) : (\"Online Deal\")\n\n deal['options'].each do |option|\n @offer_option = @offer.offer_options.build\n \n @offer_option.value_currency = deal['currency_code']\n @offer_option.value_amount = option['original_price']\n @offer_option.price_currency = deal['currency_code']\n @offer_option.price_amount = option['price']\n\n if option['savings'].blank?\n @offer_option.discount_amount = option['original_price'].to_f - option['price'].to_f\n else\n @offer_option.discount_amount = option['savings'].to_f\n end\n\n if option['discount'].blank?\n @offer_option.discount_percent = ((100.0 * (@offer_option.value_amount.to_f - @offer_option.price_amount.to_f))/\n @offer_option.value_amount.to_f).to_i\n else\n @offer_option.discount_percent = option['discount'].to_i\n end\n \n @offer_option.price_amount = \"$\" + option['price'].to_s\n @offer_option.discount_amount = \"$\" + sprintf(\"%0.02f\", @offer_option.discount_amount).to_s\n @offer_option.discount_currency = deal['currency_code']\n @offer_option.offer_id = @offer.id\n @offer.update_trend_score(deal['orders_count'])\n @offer.save\n end\n \n deal['locations'].each do |location|\n @offer_redemption_location = @offer.offer_redemption_locations.build\n @offer_redemption_location.redemption_neighborhood = location['city']\n \n @offer_redemption_location.redemption_street_address1 = location['address1']\n @offer_redemption_location.redemption_street_address2 = location['address2']\n @offer_redemption_location.redemption_city = location['city']\n @offer_redemption_location.redemption_state = location['state']\n \n if Offer::States::List.include?(@offer_redemption_location.redemption_state)\n @offer_redemption_location.redemption_country = deal['county_code']\n else\n @offer_redemption_location.redemption_country = \"NONUS\"\n end\n\n @offer_redemption_location.redemption_zip_code = location['postal_code']\n @offer_redemption_location.redemption_lat = location['latlng'][0]\n @offer_redemption_location.redemption_lng = location['latlng'][1]\n @offer_redemption_location.redemption_phone_number = location['phone']\n @offer_redemption_location.update_woeid\n @offer_redemption_location.offer_option_id = @offer_option.id\n @offer.save\n end\n \n @offer.category_id = deal['categories'][0].to_s + '-' + deal['categories'][1].to_s\n p @offer.category_id\n @offer.get_snapgadget_category_id_living_social\n @offer.save\n end\n end\n end\n\n end",
"def processSection2(section)\n string = '';\n text = section.inner_html\n text = text.delete('<p>').strip().delete('\\/').delete(','); #clean up string\n string += ',';\n string += text;\n return string;\nend",
"def prepare_data(params)\n build_ingest_form\n build_uploader(params[:upload], params[:upload_cache])\n build_asset(params[:id], params[:template_id])\n assign_form_attributes(params)\n find_unmapped_rdf\n end",
"def featured_offers_and_deals_list\n $tracer.trace(__method__)\n return GameStopProductContainerList.new(ToolTag.new(div.className(create_ats_regex_string(\"featuredoff\")).a.id(\"/aImageUrl/\"), __method__), self)\n end",
"def adopt_dependencies\n super if defined? super\n # After everything has settled down, we can extract our attributes\n accept_attributes gleaning.ready_attribute_values\n # Note that if we got an attribute from the Gleaning, we no longer need it from MercuryResult\n accept_attributes mercury_result.ready_attribute_values\n if recipe_page_needed?\n recipe_page || build_recipe_page\n accept_attribute :recipe_page, recipe_page\n # Could do this to get the RecipePage parsing done sooner\n # recipe_page.request_attributes :content\n end\n end",
"def description_html\n convert_html description\n end",
"def optionnalProtect(xml,release)\n i = 1\n xml.xpath(\"//Short_description\").each do |node|\n release[i].add_description(node.content) rescue nil\n i = i + 1\n end\n i = 1\n xml.xpath(\"//FeaturingArtists\").each do |node|\n release[i].add_featArtists(node.content) rescue nil\n i = i + 1\n end\n build_xml(release,xml)\nend",
"def generate_tags\n uri = URI.parse(\"https://api.thomsonreuters.com/permid/calais\")\n http = Net::HTTP.new(uri.host, uri.port)\n http.use_ssl = true\n post_body = []\n post_body << \"<Document><Body>\"\n # stip html\n post_body << ActionView::Base.full_sanitizer.sanitize(params[:desc])\n # no strip\n # post_body << params[:desc]\n post_body << \"</Body></Document>\"\n request = Net::HTTP::Post.new(uri.request_uri)\n request.add_field(\"Content-Type\",\"text/xml\")\n request.add_field(\"outputFormat\",\"application/json\")\n #request.add_field(\"outputFormat\",\"text/n3\") \n request.add_field(\"x-ag-access-token\",\"fY7WUM3GGCXHm9ATOhtzhrvlWX8oPo5X\")\n request.body = post_body.join\n # request[\"Content-Type\"] = \"multipart/form-data, boundary=#{BOUNDARY}\"\n\n render :json => http.request(request).body\n end",
"def parse\n @parts = []\n\n sect = text.split(/^\\-\\-\\-/)\n\n if sect.size == 1\n @header = {}\n @parts << Part.new(sect[0]) #, *[@stencil, @default].compact.flatten)\n else\n sect.shift if sect.first.strip.empty?\n\n head = sect.shift\n head = YAML::load(head)\n parse_header(head)\n\n sect.each do |body|\n index = body.index(\"\\n\")\n formats = body[0...index].strip\n formats = formats.split(/\\s+/) if String===formats\n #formats = @default if formats.empty?\n #formats.unshift(@stencil) if @stencil\n text = body[index+1..-1]\n @parts << Part.new(text, *formats)\n end\n end\n end",
"def setup\n @html_parser = Hatemile::Util::Html::NokogiriLib::NokogiriHTMLDOMParser.new(\n \"<!DOCTYPE html>\n <html>\n <head>\n <title>HaTeMiLe Tests</title>\n <meta charset=\\\"UTF-8\\\" />\n </head>\n <body>\n <section>\n <header></header>\n <article>\n \\n\n </article>\n <footer><!-- Footer --></footer>\n </section>\n <span attribute=\\\"value\\\" data-attribute=\\\"custom_value\\\">\n <!-- Comment -->\n Text node\n <strong>Strong text</strong>\n <hr />\n </span>\n <div></div>\n <p>\n <del>Deleted text</del>\n </p>\n <table>\n <thead><tr>\n <th>Table header</th>\n </tr></thead>\n <tbody>\n <tr>\n <td>Table <ins>cell</ins></td>\n </tr>\n </tbody>\n <tfoot><!-- Table footer --></tfoot>\n </table>\n <ul>\n <li id=\\\"li-1\\\">1</li>\n <li id=\\\"li-3\\\">3</li>\n </ul>\n <ol>\n <li>1</li>\n <li>2</li>\n <li>3</li>\n <li>4</li>\n <li>5</li>\n </ol>\n #{ASIDE_CONTENT}\n <form>\n <label>\n Text:\n <input type=\\\"text\\\" name=\\\"number\\\" />\n </label>\n </form>\n <h1></h1>\n <h2></h2>\n </body>\n </html>\"\n )\n end",
"def info_content\n\n info = []\n info << \"<b>\" + @place[:title].escape_html + \"</b>\"\n unless $params.election_description.empty?\n info << \"<i>\" + $params.election_description.escape_html + \"</i>\"\n info << \"\"\n end\n\n a = \"#{@place[:name]}, #{@place[:street]}, #{@place[:city]}, #{@place[:state]} #{@place[:zip]}\"\n info << \"<a href=\\\"http://maps.google.com/?daddr=#{a.escape_uri}\\\" target=\\\"_blank\\\">#{@place[:name].escape_html}</a>\"\n info << @place[:street].escape_html\n info << \"#{@place[:city]}, #{@place[:state]} #{@place[:zip]}\".escape_html\n info << \"\"\n info << \"Hours of operation:\"\n info += @place[:schedule_formatted].escape_html.split(\"\\n\").map {|s| \"\\u2022 \" + s}\n unless @place[:notes].empty?\n info << \"\"\n info << @place[:notes].escape_html\n end\n unless $params.election_info.empty?\n info << \"\"\n info << $params.election_info\n end\n info.join(\"\\n\")\n end",
"def g_goal_escenarios\n timeInit = Time.now\n empresa = Enterprise.find(params[:idEmp].to_i)\n log = []\n goalIds = params[:goalIds].split(\"|\")\n\n log.push(\"Creating goal assessment scenarios content...\")\n\n # ::ESCENARIOS:: ES: Comentar si el engine no esta activo! - EN: Comment if the engine is not active!\n log2 = view_context.escenariosGenerateGoalsHTML(empresa, goalIds, log)\n log.concat(log2) \n \n timeFin = Time.now\n timeExpend = (timeFin - timeInit).to_f\n log.push(\"Goal assessment scenarios content creation finished - Time required: \" << timeExpend.to_s << ' (seconds)')\n\n respond_to do |format|\n format.json {render json: log}\n end\n\n\n end",
"def eds_abstract(options = nil)\n values, opt = extract_config_value(options)\n separator = opt[:separator] || \"<br/>\\n\"\n result = Array.wrap(values).reject(&:blank?).join(separator)\n\n # === Insert breaks before bullets\n result.gsub!(EBSCO_BREAK_BEFORE_REGEX, '<br/>\\1 ')\n\n # === Make implied sections explicit\n if result.gsub!(%r{<br\\s*/?>([^:\\s]+:)\\s*}) { abstract_subsection($1) }\n # (1) For implied sections that follow a <br>, the first implied section\n # will be at the start of the abstract without a <br>.\n result.sub!(/\\A([^:\\s]+:)\\s*/) { abstract_subsection($1) }\n elsif result.gsub!(%r{(\\*?\\s*)([A-Z]{3,}\\s*:)}) { abstract_subsection($2) }\n # (2)\n else\n # (3) Handle implied sections within <b>.\n result.gsub!(/<b>\\s*([^:<]*:)\\s*<\\/b>/) { abstract_subsection($1) }\n end\n\n # === Eliminate leading and trailing breaks\n result.sub!(%r{^\\s*(</?\\s*br\\s*/?\\s*>\\s*)+}i, '')\n result.sub!(%r{(</?\\s*br\\s*/?\\s*>\\s*)+\\s*$}i, '')\n\n if rendering_html?(opt)\n result.html_safe.presence\n else\n result = result.split(separator)\n (values.is_a?(Array) || (result.size > 1)) ? result : result.first\n end\n end",
"def process_html\n benchmark \"Process HTML for #{self.url}\" do\n doc = Readability::Document.new(self.html)\n html = doc.html\n self.title = content_for_open_graph_tag('og:title', html) || doc.title\n self.description =\n content_for_open_graph_tag('og:description', html) ||\n content_for_meta_tag('name=\"description\"', html) ||\n html.xpath('//head/meta/@description', html).first.try(:content)\n image_url = content_for_open_graph_tag('og:image', html) || doc.images.first\n self.image_url = image_url if image_url =~ URI.regexp\n self.site_name = content_for_open_graph_tag('og:site_name', html) || get_url_domain.try(:humanize)\n self.content_html = doc.content.encode_from_charset!(doc.html.encoding)\n self.content = Nokogiri::HTML(self.content_html).text\n end\n self\n end",
"def explanation\n explanation_intro_html + explanations_array_html_list\n end",
"def create\n visit(CentralAdmin).create_institutional_proposal\n on ProposalLogLookup do |look|\n look.proposal_number.set @proposal_number\n look.search\n look.select_item @proposal_number\n end\n on InstitutionalProposal do |create|\n create.expand_all\n @document_id=create.document_id\n @doc_header=create.doc_title\n @proposal_number=create.institutional_proposal_number\n fill_out create, :proposal_type, :award_id, :activity_type, :project_title, :description\n set_sponsor_code\n create.save\n end\n if @proposal_log && $current_page.errors.size==0\n pi = make ProjectPersonnelObject, principal_name: @proposal_log.principal_investigator,\n full_name: @proposal_log.pi_full_name,\n document_id: @document_id,\n lookup_class: @lookup_class,\n search_key: @search_key,\n doc_header: @doc_header\n @project_personnel << pi\n view :contacts\n @project_personnel.principal_investigator.set_up_units\n end\n end",
"def prepare_html(content , page_type = 'N')\n #header\n 1.upto 5 do |no| content.gsub! /^(={#{no}}) (.*) (={#{no}})/ ,\"\\nh#{no+1}. \\\\2\\n\" end\n 1.upto 5 do |no| content.gsub! /^(={#{no}}) (.*)/ ,\"\\nh#{no+1}. \\\\2\\n\" end\n\n #list\n 1.upto 5 do |no| content.gsub! /^([ ]{#{no}})(\\*) ?(.*)/ ,\"#{'*'*no} \\\\3\" end\n 1.upto 5 do |no| content.gsub! /^([ ]{#{no}})(#) ?(.*)/ ,\"#{'#'*no} \\\\3\" end\n #content.gsub! /(\\*) v (.*)/ , \"\\\\1 -\\\\2-\"\n \n #block\n content.gsub! /^\\{\\{\\{/ , \"<pre>\" ; content.gsub! /^\\}\\}\\}/ , \"</pre>\"\n content.gsub! /^\\{\\{\\\"/ , \"<blockquote>\" ; content.gsub! /^\\\"\\}\\}/ , \"</blockquote>\"\n content.gsub! /^\\{\\{\\[/ , \"<math>\" ; content.gsub! /^\\]\\}\\}/ , \"</math>\"\n \n #concept & property\n content.gsub! /\\[\\[(.*?):=(.*?)\\]\\]/ , '\\1(\\2)'\n #content.gsub! /\\[\\[(.*?)[<>=].*?\\]\\]/ , \\\"\\\\1\\\":#{APP_ROOT}/page/\\\\1\" \n content.gsub! /\\[\\[(.*?)\\]\\]/ , \"\\\"\\\\1\\\":#{APP_ROOT}/entry/\\\\1\" if defined?(APP_ROOT)\n\n #comment\n content.gsub! PTN_COMMENT , \"\\\\1\"\n content.gsub! PTN_COMMENT_MULTILINE , \"\"\n if defined? SystemConfig\n SystemConfig.site_info.each do |e|\n content.gsub! /(\\s)#{e[1]}:/ , \"\\\\1#{e[2]}\"\n end\n content.gsub! SystemConfig.ptn_url_unnamed , \"\\\\1\\\"\\\\2\\\":\\\\2\"\n content.gsub! \"%ROOT%\" , APP_ROOT\n end\n \n #Process by page_type\n case page_type\n when 'N'\n math_list = content.scan( PTN_MATH ) ; math_list.each do |m|\n #content.gsub! \"$#{m[0]}$\" , latex_render(m[0])\n content.gsub! \"$#{m[0]}$\" , get_math_img(m[0])\n end\n math_block_list = content.scan( PTN_MATH_BLOCK ) ; math_block_list.each do |m|\n #content.gsub! \"#{m[0]}\" , latex_render(m[0])\n content.gsub! \"#{m[0]}\" , get_math_img(m[0])\n end\n when 'S'\n menu_list = content.scan( PTN_MENU ) ; menu_list.each do |m|\n menu_title = m[0] ; menu_target = m[1] ; menu_str = \"M{{#{menu_title}|#{menu_target}}}\"\n #$lgr.info \"#{menu_title} / #{menu_target}\"\n result = link_to_remote(menu_title , :url => { :action => 'menu' , :query => CGI.escape(menu_target) })\n content.gsub! menu_str , result\n end\n end\n #$lgr.info \"[prepare_html] \"+content\n query_list = content.scan( PTN_QUERY ) ; query_list.each do |q|\n query_type = q[0] ; query_content = q[1] ; query_str = \"#{query_type}{{#{query_content}}}\"\n case query_type\n when 'P'\n result = eval(\"find_page :display=>'|@title|@tags|@created_at|' ,\" + query_content )\n result = result.join(\"\\n\") if result.class == Array\n result = \"|_.Title|_.Tag|_.CreatedAt|\\n\"+result if query_content.scan(/:display/).size == 0\n #$lgr.info \"[prepare_html] Query : #{query_str} , #{result}\"\n content.gsub! query_str , result\n end\n end\n #content.gsub! SystemConfig.ptn_url , \"\\\"\\\\0\\\":\\\\0\"\n #???content.gsub!(SystemConfig.ptn_site) \"\\\"#{ApplicationController.SystemConfig(\\\\0)}\\\":\\\\0\"\n content\n end",
"def fetch_from_g750 url\n if g750_host? url\n page = Nokogiri::HTML(open(url).read)\n @title = page.css('h1.c-article__title').text\n\n # get times\n @preptime = sanitize(page.css('ul.c-recipe-summary > li.c-recipe-summary__rating[title=\"Temps de préparation\"]').text).to_i\n @cooktime = sanitize(page.css('ul.c-recipe-summary > li.c-recipe-summary__rating[title=\"Temps de cuisson\"]').text).to_i\n\n @steps = []\n css_step = \"div[itemprop=recipeInstructions] p\"\n @steps = page.css(css_step).text.split /[( ),(<br>)]/\n\n @ingredients = []\n css_ingredient = \"div.c-recipe-ingredients ul.c-recipe-ingredients__list li.ingredient\"\n page.css(css_ingredient).each { |ing_node|\n @ingredients << sanitize(ing_node.text)\n }\n\n # get image\n css_image = 'div.swiper-wrapper img.photo'\n begin\n @image = page.css(css_image).attr('src').to_s\n rescue NoMethodError => e\n end\n\n else\n raise ArgumentError, \"Instantiation cancelled (ulr not from #{G750_HOST}).\"\n end\n end",
"def guides_posts_list\n $tracer.trace(__method__)\n return ToolTag.new(div.id(\"/fragment-3805/\").ul.className(\"content-list\").li.className(\"/content-item/\"), __method__)\n end",
"def create\n if text.match(/\\_QUOTE/)\n require 'organismo/element/quote'\n Organismo::Element::Quote.new(text, location)\n elsif text.match(/\\_SRC/)\n require 'organismo/element/code'\n Organismo::Element::Code.new(text, location)\n elsif text.match(/\\_EXAMPLE/)\n require 'organismo/element/example'\n Organismo::Element::Example.new(text, location)\n elsif text.match(/\\*/)\n require 'organismo/element/header'\n Organismo::Element::Header.new(text, location)\n elsif text.match(/\\[\\[\\S*(\\.png)|(\\jpg)|(\\.jpeg)\\]\\]/)\n require 'organismo/element/image'\n Organismo::Element::Image.new(text, location)\n elsif text.match(/\\[\\[\\S*\\]\\]/)\n require 'organismo/element/link'\n Organismo::Element::Link.new(text, location) \n elsif text.match(/\\-/)\n require 'organismo/element/plain_list'\n Organismo::Element::PlainList.new(text, location)\n else\n require 'organismo/element/text'\n Organismo::Element::Text.new(text, location)\n end\n end",
"def get_features(frag, model)\n w1 = (frag.cleaned.last or '')\n w2 = (frag.next or '')\n\n frag.features = [\"w1_#{w1}\", \"w2_#{w2}\", \"both_#{w1}_#{w2}\"]\n\n if not w2.empty?\n if w1.chop.is_alphabetic? \n frag.features.push \"w1length_#{[10, w1.length].min}\", \"w1abbr_#{model.non_abbrs[w1.chop]}\"\n end\n\n if w2.chop.is_alphabetic?\n frag.features.push \"w2cap_#{w2[0,1].is_upper_case?}\", \"w2lower_#{model.lower_words[w2.downcase]}\"\n end\n end\n end",
"def scrape_course_prereqs(parsed_html)\n prereqs = Array.new\n\n parsed_html.css('a.sc-courselink').each do |link|\n co = Course.new\n co.name = link.children.text\n co.url = 'http://pdx.smartcatalogiq.com' + link.attributes['href'].value\n prereqs << co\n end\n\n prereqs\n end",
"def initialize(opts={})\n @url = opts[:url].strip\n\n s = RecipeDocument.read_document(opts)\n\n @options = DEFAULT_OPTIONS.merge(opts)\n # remove hardspaces   with a simple space.\n s.gsub!(' ', ' ')\n @doc = Nokogiri::HTML(s)\n\n\n @trimmed_doc = Nokogiri::HTML(s)\n\n @trimmed_doc.css(\"object, embed\").each do |elem|\n elem.remove\n end\n\n remove_unlikely_candidates!\n remove_divs_with_high_link_density!\n\n @title = @doc.xpath(\"//title\").text.lstrip.rstrip.gsub(/[\\n]+/, \" \")\n end",
"def extract_data( page )\n strip_tags page[:content]\n\n #extract_concepts( page[:title] , page , LOC_TITLE ) if page[:title] && page[:title] !=~ PTN_EMPTY\n if page[:tag] && page[:tag] !=~ PTN_EMPTY\n page[:tag] = page[:tag].split(SEP_TAG).map{|e|\"[[\"+e+\"]]\"}.join(\";\")\n extract_concepts( page[:tag] , page , LOC_TAG )\n end\n\n extract_concepts( page[:title] , page , LOC_TITLE)\n extract_concepts( page[:content] , page )\n extract_resources( page[:content] , page )\n end",
"def site_description\n headings = @doc.xpath(\"//h3[@class='clearl']\")\n content_sections = @doc.xpath(\"//h3[@class='clearl']/following-sibling::p[1]\")\n content = \"\"\n headings.zip(content_sections).each do |h, c| \n unless (c.to_s().squeeze().empty?)\n content << \"<h3>#{sanitize(h.to_s)}</h3>\" \n content << \"<p>#{sanitize(c.to_s)}\"\n end\n end\n rhtml = IO.read(File.expand_path(\"site_description.rhtml\", File.dirname(__FILE__)))\n content_html = Erubis::Eruby.new(rhtml)\n content_html.result(:content => content)\n end",
"def process_holdings_data\n extract_call_numbers\n extract_notes\n extract_textual_summary_holdings\n build_summary_holdings if @summary_holding.empty?\n end",
"def initialize(description, url)\n @url = url\n\n @type = description.at('serviceType').text.strip\n @id = description.at('serviceId').text.strip\n @control_url = @url + description.at('controlURL').text.strip\n @event_sub_url = @url + description.at('eventSubURL').text.strip\n @scpd_url = @url + description.at('SCPDURL').text.strip\n\n create_driver\n end",
"def get_articles(sub_topics)\n FileUtils::mkdir_p \"#{output_dir}/guides\"\n sub_topics.map do |sub_topic|\n link = sub_topic.css(\"a\")[0]\n title = link.inner_text\n href = \"http://emberjs.com#{link[:href]}\"\n $stderr.puts \"- #{title}\"\n\n path = href[/(guides\\/[^\\/]*)\\/?.*/, 0] + '.html'\n dirpath = href[/(guides\\/[^\\/]*)\\/?.*/, 1]\n FileUtils::mkdir_p \"#{output_dir}/#{dirpath}\"\n\n html = run_shell_command \"curl -s #{href}/\"\n chapter_doc = Nokogiri::HTML html\n\n File.open(\"#{output_dir}/#{path}\", 'w') do |f|\n f.puts(chapter_doc.at('.chapter').inner_html)\n end\n\n {\n title: title,\n path: path\n }\n end\n end",
"def part(part_label, part_title, part_intro, opts={})\n return if hide_content?(opts)\n\n concat_output <<EOT\n <part label=\"#{part_label}\">\n <title>#{part_title}</title>\n <partintro>\n <para>#{part_intro}</para>\n </partintro>\nEOT\n\n yield\n\n concat_output <<EOT\n </part>\nEOT\n end",
"def acquire(links, agent)\n puts \"Acquiring data...\".foreground(:cyan)\n\n links.each do |link|\n name = link.inner_html.to_s\n code = link[\"href\"].to_s.sub course_url, \"\"\n\n if name == \"\" || code == \"\"\n raise \"Incorrect parsing. Please contact the developer.\".foreground(:red)\n end\n \n grade = Grade.new name, code, agent\n course = Course.new name, code, grade\n\n @learn_courses << course\n @learn_grades << grade\n end \n end",
"def call_all_apartments(ruby_apartments)\n ruby_apartments.each do |apartment|\n puts apartment.show_all_apartments\n end\nend"
] |
[
"0.5359373",
"0.53573185",
"0.52481955",
"0.52323633",
"0.5228217",
"0.5209506",
"0.5098252",
"0.50677514",
"0.505511",
"0.505511",
"0.505511",
"0.5024265",
"0.49952686",
"0.49942705",
"0.49534038",
"0.49341407",
"0.49321362",
"0.49249387",
"0.4918939",
"0.49103743",
"0.48976934",
"0.4887873",
"0.48861572",
"0.48839667",
"0.4875752",
"0.48695752",
"0.48641443",
"0.4861048",
"0.48569828",
"0.4837246",
"0.4834528",
"0.48012275",
"0.47950798",
"0.47910616",
"0.47894177",
"0.47849295",
"0.47828895",
"0.47750524",
"0.47600895",
"0.47575334",
"0.47556037",
"0.47555283",
"0.47266805",
"0.47211713",
"0.47082603",
"0.47007528",
"0.467792",
"0.46695387",
"0.46317944",
"0.46312526",
"0.46265787",
"0.46173477",
"0.46091703",
"0.4608599",
"0.46059933",
"0.45930365",
"0.45882022",
"0.45856854",
"0.4585168",
"0.4583274",
"0.45807758",
"0.45763975",
"0.45628554",
"0.45604718",
"0.45587915",
"0.4555707",
"0.45546213",
"0.45481092",
"0.4544624",
"0.45421946",
"0.45378876",
"0.4537881",
"0.4537254",
"0.45327312",
"0.4531931",
"0.4531446",
"0.45313528",
"0.45298746",
"0.45288548",
"0.4524628",
"0.45204532",
"0.4516884",
"0.4512082",
"0.45117173",
"0.4505753",
"0.45045748",
"0.45011434",
"0.45008412",
"0.44994783",
"0.44955763",
"0.4490737",
"0.44890437",
"0.44880772",
"0.44870406",
"0.44862935",
"0.44833186",
"0.4482437",
"0.44775927",
"0.44716695",
"0.4466803"
] |
0.7131718
|
0
|
Makes common apartments description. Hash, based on constant APARTMENT_DESCRIPTION, first merge it, then find unique description
|
def apartment_description
apartment_description = {}
apartment_description.merge!(APARTMENT_DESCRIPTION)
apartment_description.keys.each { |key| apartment_description[key] = send(apartment_description[key]) }
apartment_description
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def merge_info_description(info, val)\n\t\tmerge_info_string(info, 'Description', val)\n\tend",
"def merge_description_notes\n src_notes = @src.all_notes\n dest_notes = @dest.all_notes\n src_notes.each_key do |f|\n if dest_notes[f].blank?\n dest_notes[f] = src_notes[f]\n elsif src_notes[f].present?\n dest_notes[f] += \"\\n\\n--------------------------------------\\n\\n\"\n dest_notes[f] += src_notes[f].to_s\n end\n end\n @dest.all_notes = dest_notes\n end",
"def description_for_google\n google_description.present? ? google_description : description\n end",
"def best_description\n if description?\n description\n else\n additional_description\n end\n end",
"def reason_codes_and_descriptions\n rc_and_desc = []\n reason_code_record = nil\n if @crosswalk_records && @crosswalk_records.length > 0\n rcc_log.debug \"Obtaining All Reason codes and descriptions.\"\n @crosswalk_records.each do |reason_code_record|\n if reason_code_record.present?\n reason_code = reason_code_record.reason_code\n reason_code_description = reason_code_record.reason_code_description.to_s.upcase\n reason_code = reason_code.to_s.upcase\n if reason_code.present? && reason_code_description.present?\n rc_and_desc << [reason_code, reason_code_description]\n end\n end\n end\n end\n rc_and_desc = rc_and_desc.compact.uniq\n rcc_log.debug \"Reason Codes and descriptions are : #{rc_and_desc.join(', ')}\"\n rc_and_desc\n end",
"def merge_description_notes(src, dest)\n src_notes = src.all_notes\n dest_notes = dest.all_notes\n for f in src_notes.keys\n if dest_notes[f].blank?\n dest_notes[f] = src_notes[f]\n elsif !src_notes[f].blank?\n dest_notes[f] += \"\\n\\n--------------------------------------\\n\\n\"\n dest_notes[f] += src_notes[f].to_s\n end\n end\n dest.all_notes = dest_notes\n end",
"def description\n text = ['A']\n text << [duration, 'minute'] if duration.present?\n text << [format, 'publication']\n text << [ 'from', pretty_date(published_on) ] if published_on.present?\n text << ['by', presentations.first.speaker_names] if presentations.present?\n text.join(' ')\n end",
"def description_lookup\n @@description_lookup ||= {}\n end",
"def get_all_reason_codes_and_descriptions\n rc_and_desc = []\n if @crosswalk_records && @crosswalk_records.length > 0\n rcc_log.debug \"Obtaining All Reason codes and descriptions for Output Segments and tags.\"\n @crosswalk_records.each do |reason_code_record|\n if reason_code_record.present?\n code = reason_code_record.reason_code.to_s.upcase\n if @facility.details[:rc_crosswalk_done_by_client] && code == 'NORC'\n @code_norc_found = true\n end\n is_hipaa_crosswalk_present = reason_code_record.hipaa_code_active_indicator && reason_code_record.hipaa_adjustment_code.present?\n if (@facility.details[:rc_crosswalk_done_by_client].blank? ||\n (@facility.details[:rc_crosswalk_done_by_client] &&\n !is_hipaa_crosswalk_present && code != 'NORC'))\n if @fetch_footnote_code\n reason_code = footnote_code(reason_code_record)\n else\n reason_code = reason_code_record.reason_code\n end\n reason_code_description = reason_code_record.reason_code_description.to_s.upcase\n reason_code = reason_code.to_s.upcase\n notify = reason_code_record.notify\n if reason_code.present? && reason_code_description.present?\n rc_and_desc << [reason_code, reason_code_description, notify, is_hipaa_crosswalk_present]\n end\n end\n end\n end\n end\n rc_and_desc = rc_and_desc.compact.uniq\n rcc_log.debug \"Reason Codes and descriptions are : #{rc_and_desc.join(', ')}\"\n rc_and_desc\n end",
"def get_economic_categories_from_budget_items_list(items)\n def count_different_descriptions(items)\n items.map{|i| i[:description]}.uniq.count\n end\n\n # First, group items by economic concept\n buckets = {}\n items.each do |item|\n concept = item[:economic_concept]\n next if concept.nil? or concept.empty?\n next if concept.length > 4 # Budget item\n # Note: We don't need economic categories for budget items (concept length==5), they are \n # just items belonging to a heading. At one point the obstacle to this was distinguishing \n # heading subtotals from the items themselves in the output files, but we've sorted that \n # out through a new 'budget item' column in the output (see below).\n\n buckets[concept] = [] if buckets[concept].nil? \n buckets[concept].push item\n end\n\n # Then, for each bucket, decide whether we need to tag the economic concept\n categories = {}\n buckets.each do |concept, items|\n if count_different_descriptions(items) > 1 # We need to tag the concept\n # We expect this to happen only for headings\n if concept.length < 3\n puts \"Warning: inconsistent descriptions for article or chapter #{concept}!\"\n end\n\n # Create a category for each item, and modify the items to point to them\n items.each do |item|\n tagged_concept = \"#{concept}/#{get_entity_id(item[:section], item[:service])}\"\n item[:economic_concept] = tagged_concept\n categories[tagged_concept] = item[:description]\n end\n else\n categories[concept] = items.first[:description] # Pick the first, they're all the same\n end\n end\n categories\nend",
"def add_info_on_tag_ids\n desc_hash = Notion.where(locale: @lecture.locale || I18n.default_locale)\n .pluck(:title, :tag_id, :aliased_tag_id)\n .map { |x| [x.first.downcase, x.second || x.third] }\n .select { |x| x.first.in?(@content_descriptions.map(&:downcase)) }\n .to_h\n @content.each do |c|\n c['tag_id'] = desc_hash[c['description'].downcase]\n end\n end",
"def full_description(ulterior_desc='', debug = false)\n return (description rescue '(no desc.)' ) unless debug\n arr_descr = [ \n \"Full Description:\",\n\t\t\tkv('description'), \n\t\t\tkv('location'), \n kv('repeat_frequency'), \n kv('repeats'), \n kv('url'), \n kv('price') ,\n kv('active') ,\n kv('repetition_magic_string'),\n \"#RICCAL=#{self.calendar.abbrev rescue \"No Abbrev available\"}\",\n 'Ulteriore descrizione: ',\n ulterior_desc ,\n ]\n if self.venue\n arr_descr << self.venue\n #arr_descr << self.venue.lat.to_s\n end\n return arr_descr[debug].join(\"\\n--\\n\") # rescue \"Exception w/ full_description: #{$!}\"\n end",
"def description\n [basic_description, extra_description].compact.join(\" \")\n end",
"def kase_description_in_words(kase)\n result = []\n result << followers_count_in_words(kase).to_s.capitalize\n\t result << replies_count_in_words(kase)\n\t result << kase_type_and_time_in_words(kase)\n\t result.compact.map {|m| m.to_s.strip}.reject {|i| i.empty?}.join(', ')\n end",
"def name_merge hsh\n str = \"\"\n @log.debug \"name merge for #{hsh}\"\n str += hsh[:name].strip.force_encoding(\"UTF-8\") if hsh[:name]\n str += \"\\n#{hsh[:description].strip.force_encoding(\"UTF-8\")}\" if hsh[:description]\n str\n end",
"def _search_text\n [_concatenated_brand,\n _concatenated_description,\n _concatenated_sell_unit,\n classic_mbid\n ].compact.map { |w| w.hanize.split(' ') }.flatten.uniq.reject { |w| w.size < 3 || self.class.stop_words.include?(w) }.join(' ')\nend",
"def aai_strings one_way_aai, two_way_aai\n aai_strings = {}\n keys = (one_way_aai.keys + two_way_aai.keys).\n map { |key| key.sort }.uniq\n\n keys.each do |key|\n a_to_b_aai = one_way_aai[key] || \"NA\"\n b_to_a_aai = one_way_aai[key.reverse] || \"NA\"\n two_way = two_way_aai[key] || \"NA\"\n\n aai_strings[key] = [a_to_b_aai,\n b_to_a_aai,\n two_way]\n end\n\n aai_strings.map do |genome_pair, aais|\n [genome_pair, aais].flatten.join \"\\t\"\n end\n end",
"def unique_tag\n\n\t\t##collect transforms given array into the new array abbr\n\t\tabbr = self.author.split(\" \").collect do |sub_string|\n\t\t\tsub_string[0] #puts first letter of each substring into intials array\n\t\tend\n\n\t\treturn abbr.join + '#' + self.id.to_s ##join concatenates an array into a string\n\tend",
"def explain_application_info(org_text)\n retval = {} # Default\n return retval unless org_text\n\n if org_text.match('Application = ')\n appl = SyspApplication.get_cached_instance(org_text.split(' ')[2].to_i)\n if appl\n retval[:short_info] = appl.name\n retval[:long_info] = \"#{appl.description} >> Team: #{appl.developmentteam.name}\"\n else\n retval[:short_info] = \"Application not found for #{org_text}\"\n end\n end\n\n if org_text.match('ID_WSMethod = ')\n ws = Wsmethod.get_cached_instance(org_text.split(' ')[2].to_i)\n if ws\n retval[:short_info] = ws.name\n retval[:long_info] = \"#{ws.name}\"\n else\n retval[:short_info] = \"WSMethod not found for #{org_text}\"\n end\n end\n\n if org_text.match('ID_OFMsgType = ')\n mt = Ofmessagetype.get_cached_instance(org_text.split(' ')[2].to_i, session[:database].hash)\n if mt\n retval[:short_info] = mt.name\n retval[:long_info] = \"#{mt.description} >> Domain: #{mt.domain.name}\"\n else\n retval[:short_info] = \"OFMessagetype not found for #{org_text}\"\n end\n end\n\n\n\n retval\n end",
"def format_description(text)\n # Look for signs of structure, otherwise just treat as unstructured.\n case text\n when /\"\";/ then double_quotes_to_sections(text)\n when /\\.--v\\. */ then double_dash_to_sections(text)\n when /; *PART */i then # Seen in some IA records.\n when /:;/ then # Observed in one unusual case.\n when /[[:punct:]] *--.* +-- +/ then # Blurbs/quotes with attribution.\n when / +-- +.* +-- +/ then # Table-of-contents title list.\n when /(;[^;]+){4,}/ then # Many sections indicated.\n else return format_multiline(text)\n end\n q_section = nil\n text.split(/ *; */).flat_map { |part|\n next if (part = part.strip).blank?\n case part\n when /^\"\"(.*)\"\"$/\n # === Rare type of table-of-contents listing entry\n line = $1.to_s\n if line.match(SECTION_TITLE_RE)\n gap = (\"\\n\" unless q_section)\n q_section = $1.to_s\n [gap, \"#{q_section} #{$2}\", \"\\n\"].compact\n else\n q_section = nil\n line.match?(/^\\d+ +/) ? line : \"#{BLACK_CIRCLE}#{EN_SPACE}#{line}\"\n end\n\n when / +-- +.* +-- +/\n # === Table-of-contents listing\n section = nil\n part.split(/ +-- +/).flat_map { |line|\n if line.match(SECTION_TITLE_RE)\n gap = (\"\\n\" unless section)\n section = $1.to_s.delete_suffix('.')\n [gap, \"#{section}. #{$2}\", \"\\n\"].compact\n else\n section = nil\n \"#{BLACK_CIRCLE}#{EN_SPACE}#{line}\"\n end\n }.tap { |toc| toc << \"\\n\" unless toc.last == \"\\n\" }\n\n when /[[:punct:]] *--/\n # === Blurbs/quotes with attribution\n part.scan(BLURB_RE).flat_map do |paragraph, attribution|\n attribution.remove!(/[.\\s]+$/)\n [\"#{paragraph} #{EM_DASH}#{attribution}.\", \"\\n\"]\n end\n\n when /^v[^.]*\\. *\\d/\n # === Apparent table-of-contents volume title\n [part]\n\n else\n # === Plain text section\n part = \"#{part}.\" unless part.match?(/[[:punct:]]$/)\n [part, \"\\n\"]\n end\n }.compact.map { |line|\n line.gsub(/---/, EM_DASH).gsub(/--/, EN_DASH)\n }\n end",
"def get_licence_step1_per_item\n a = @dc[:description] + @dc[:rights]\n a.each{|desc|\n # It is important to test for abbreviations in the correct order\n # (as per the LICENCE_KEYS array).\n LICENCE_KEYS.each{|k| return k.to_s.upcase.gsub(\"_\", \"-\") if\n desc.match(LICENCE_ABBR_REGEX_LIST[k]) || desc.match(LICENCE_URL_REGEX_LIST[k])}\n }\n nil\n end",
"def create_shared_alt\r\n shared_alts = Array.new\r\n \r\n @specimen_alts.values.each do |a|\r\n if(!shared_alts.include?(a) && !a.eql?(@ref))\r\n shared_alts.push(a)\r\n end\r\n end\r\n\r\n alt_str = \"[\" + @ref + \"/\"\r\n alts_without_nc = shared_alts.reject{ |entry| entry.eql?(\"nc\") }\r\n alts_without_nc.each_with_index do |sa, i|\r\n if (i < alts_without_nc.size-1)\r\n alt_str += sa + \"/\"\r\n else\r\n alt_str += sa\r\n end\r\n end\r\n alt_str += \"]\"\r\n\r\n @shared_alt = alt_str\r\n end",
"def explain_application_info(org_text)\n return {}\n\n # Example:\n # retval = {}\n # retval[:short_info] = ws.name\n # retval[:long_info] = \"#{ws.name}\"\n # return retval\n end",
"def grouped_attributes_description(attributes)\n make_sentence(\n attributes.sort_by{ |key, value| key.to_s }.map do |key, value|\n attribute_description(key, value)\n end\n )\n end",
"def get_original_combination\n unless GENUS_AND_SPECIES_RANK_NAMES.include?(self.rank_string) && self.class == Protonym\n cached_html = nil\n else\n relationships = self.original_combination_relationships\n relationships = relationships.sort_by{|r| r.type_class.order_index }\n genus = ''\n subgenus = ''\n superspecies = ''\n species = ''\n gender = nil\n relationships.each do |i|\n case i.type_class.object_relationship_name\n when 'original genus'\n genus = '<em>' + i.subject_taxon_name.name_with_misspelling(nil) + '</em> '\n gender = i.subject_taxon_name.gender_name\n when 'original subgenus' \n subgenus += '<em>' + i.subject_taxon_name.name_with_misspelling(nil) + '</em> '\n when 'original section' \n subgenus += 'sect. <em>' + i.subject_taxon_name.name_with_misspelling(nil) + '</em> '\n when 'original subsection' \n subgenus += 'subsect. <em>' + i.subject_taxon_name.name_with_misspelling(nil) + '</em> '\n when 'original series' \n subgenus += 'ser. <em>' + i.subject_taxon_name.name_with_misspelling(nil) + '</em> '\n when 'original subseries' \n subgenus += 'subser. <em>' + i.subject_taxon_name.name_with_misspelling(nil) + '</em> '\n when 'original species' \n species += '<em>' + i.subject_taxon_name.name_with_misspelling(gender) + '</em> '\n when 'original subspecies' \n species += '<em>' + i.subject_taxon_name.name_with_misspelling(gender) + '</em> '\n when 'original variety' \n species += 'var. <em>' + i.subject_taxon_name.name_with_misspelling(gender) + '</em> '\n when 'original subvariety' \n species += 'subvar. <em>' + i.subject_taxon_name.name_with_misspelling(gender) + '</em> '\n when 'original form' \n species += 'f. <em>' + i.subject_taxon_name.name_with_misspelling(gender) + '</em> '\n end\n end\n if self.rank_string =~ /Genus/\n if genus.blank?\n genus += '<em>' + self.name_with_misspelling(nil) + '</em> '\n else\n subgenus += '<em>' + self.name_with_misspelling(nil) + '</em> '\n end\n elsif self.rank_string =~ /Species/\n species += '<em>' + self.name_with_misspelling(nil) + '</em> '\n genus = '<em>' + self.ancestor_at_rank('genus').name_with_misspelling(nil) + '</em> ' if genus.empty? && !self.ancestor_at_rank('genus').nil?\n end\n subgenus = '(' + subgenus.squish + ') ' unless subgenus.empty?\n cached_html = (genus + subgenus + superspecies + species).squish.gsub('</em> <em>', ' ')\n cached_html.blank? ? nil : cached_html\n end\n end",
"def get_description(room)\n lo = room.learning_objects.shuffle.first\n description = \"Zastúpené témy: \"\n list = Array.new\n\n list = lo.concepts.take(Room::NUMBER_OF_CONCEPTS_IN_DESCRIPTION).map(&:name)\n\n \"#{list.join(\", \")}.\"\n end",
"def log_description_merged(old_desc)\n @description.parent.log(:log_object_merged_by_user,\n user: @user.login, touch: true,\n from: old_desc.unique_partial_format_name,\n to: @description.unique_partial_format_name)\n end",
"def content_description(element, essence_hash)\n essence_hash.stringify_keys!\n # No name given. We build the content from essence type.\n if essence_hash['name'].blank? && essence_hash['essence_type'].present?\n content_description_from_essence_type(element, essence_hash['essence_type'])\n else\n content_description_from_element(element, essence_hash['name'])\n end\n end",
"def get_description(animal)\n animals = [\n {name: \"Dwarf Mongoose\", description: \"You are the Dwarf Mongoose, Africa's smallest carnivorous mammal. You aren't afraid of a fight, even though you are only 300 grams, and you would never abandon your fellow mongeese.\"},\n {name: \"Capybara\", description: \"You are the Capybara, a beloved member of the rodent family from South America. You are renown for your ability to get along with others regardless of differences and you make friends easily.\"},\n {name: \"Leopard\", description: \"You are the Leopard, an agile hunter of the cat family, who prefers to hunt alone. You are determined in going after the things you want, no matter the opinion of others.\"},\n {name: \"Koala\", description: \"You are the Koala, the tree-loving Australian marsupial. While others waste energy running around and worrying, you just prefer to sit back by yourself and relax.\"}\n ]\n\n for hash in animals\n if hash[:name] == animal\n description = hash[:description]\n end\n end\n\n return description\nend",
"def calculate_unique_hash\n unique = ''\n unique += self.content if self.content.present?\n unique += self.summary if self.summary.present?\n unique += self.title if self.title.present?\n self.unique_hash = Digest::MD5.hexdigest unique\n end",
"def attributes_description\n grouped_attributes = @attributes.group_by { |key, value| !!value }\n\n make_sentence(\n # Yes, this is functionally equivalent to grouped_attributes.map, except this forces the\n # keys to be evalutated in the order [true, false]. This is necessary to maintain\n # compatibility with Ruby 1.8.7, because hashes in 1.8.7 aren't ordered.\n [true, false].reduce([]) do |memo, group_key|\n attributes = grouped_attributes[group_key]\n\n if attributes\n memo << grouped_attributes_prefix(group_key, attributes.count > 1) +\n grouped_attributes_description(attributes)\n end\n\n memo\n end\n )\n end",
"def map_description_part(description, query)\n start_of_find = description.gsub(/_/, \" \").downcase.index(query)\n substring_start = if start_of_find - 17 < 0 then 0 else start_of_find - 17 end\n substring_end = if start_of_find + 20 > description.size then description.size else start_of_find + 20 end\n\n matching_description = description[substring_start..substring_end].gsub(\"/~/\", \",\")\n\n if substring_end != description.size\n matching_description << \"...\"\n end\n\n if substring_start != 0\n matching_description = \"...\" + matching_description\n end\n\n matching_description\n end",
"def get_title_description()\n\t\tif !self.building_number.nil?\n\t\t\tif self.title.include? self.building_number\n\t\t\t\treturn \"Building \" + self.building_number\n\t\t\telse\n\t\t\t\treturn \"Building \" + self.building_number + \"- \" + self.title\n\t\t\tend\n\t\telse\n\t\t\treturn self.title\n\t\tend\n\tend",
"def extract_work_description_display\n description_display_array = {}\n self.find_by_terms(:vra_work,:descriptionSet,:descriptionSet_display).each do |description_display| \n ::Solrizer::Extractor.insert_solr_field_value(description_display_array, \"description_display_tesim\", description_display.text) \n end\n return description_display_array\n end",
"def clear_desc_helper(description)\n return \"N/A\" if description.nil? || description.empty?\n\n tmp = description.dup\n tmp = tmp.delete \"#{1.chr}-#{31.chr}\".split.join ' '\n\n return tmp\n end",
"def replace_keyword_in_requirement\n reqs = Requirement.where(project: project)\n reqs.each do |req|\n next unless req.description.include?(key)\n abbr = \"<abbr title='#{definition}'>#{key}</abbr>\"\n newdesc = if req.description[%r{<abbr title='[A-Za-z0-9 ,.\\-]*'>#{key}</abbr>}]\n req.description.gsub(%r{<abbr title='[A-Za-z0-9 ,.\\-]*'>#{key}</abbr>}, abbr)\n else\n req.description.gsub(key, abbr)\n end\n req.update(description: newdesc)\n req.save\n end\n end",
"def generate_description\n text, _ = parse_description\n parts = []\n parts << text if text.length > 0\n parts << JSON.generate(config) if config && config.length > 0\n parts.join(\"\\n\")\n end",
"def test_format_reason_code_description\n valid_rc_desc = \"The impact of prior payer(s) adjudication including payments and/or adjustments\"\n assert_equal(valid_rc_desc, @output_xml_obj.format_reason_code_description(@reason_code1))\n assert_equal(valid_rc_desc, @output_xml_obj.format_reason_code_description(@reason_code2))\n assert_equal(valid_rc_desc, @output_xml_obj.format_reason_code_description(@reason_code3))\n end",
"def hash\n [sequence_number, corporate_number, process, correct, update_date, change_date, name, name_image_id, kind, prefecture_name, city_name, street_number, address_image_id, prefecture_code, city_code, post_code, address_outside, address_outside_image_id, close_date, close_cause, successor_corporate_number, change_cause, assignment_date, latest, en_name, en_prefecture_name, en_city_name, en_address_outside, furigana, hihyoji].hash\n end",
"def get_full_name\n return nil unless GENUS_AND_SPECIES_RANK_NAMES.include?(self.rank_string) || self.type == 'Combination'\n d = full_name_hash\n elements = []\n eo = '<em>'\n ec = '</em>'\n d.merge!('genus' => [nil, '[GENUS NOT PROVIDED]']) if !d['genus']\n\n elements.push(\"#{eo}#{d['genus'][1]}#{ec}\")\n elements.push ['(', %w{subgenus section subsection series subseries}.collect { |r| d[r] ? [d[r][0], \"#{eo}#{d[r][1]}#{ec}\"] : nil }, ')']\n elements.push ['(', eo, d['superspecies'], ec, ')'] if d['superspecies']\n\n %w{species subspecies variety subvariety form subform}.each do |r|\n elements.push(d[r][0], \"#{eo}#{d[r][1]}#{ec}\") if d[r]\n end\n\n elements.flatten.compact.join(\" \").gsub(/\\(\\s*\\)/, '').gsub(/\\(\\s/, '(').gsub(/\\s\\)/, ')').squish.gsub('</em> <em>', ' ')\n end",
"def map_oai_qdc\n { :title => [:title],\n :alternative => [:alternative],\n :description => [:description, :descriptionOfManifestation, :culturalContext, :stylePeriod, :designerLabel, :motif],\n :abstract => [:abstract],\n :identifier => [:accessURL],\n :date => [:date, :viewDate, :earliestDate, :created, :awardDate, :issued],\n :created => [:created],\n :issued => [:issued],\n :creator => [:creator],\n :contributor => [:arranger, :artist, :author, :cartographer, :collector, :composer, :contributor, :editor, :illustrator,\n :interviewee, :interviewer, :lyricist, :patron, :photographer, :printMaker, :scribe, :transcriber, :translator, \n :dedicatee, :recipient, :owner, :designer, :landscapeArchitect],\n :subject => [:lcsubject, :subject, :phylum, :taxonClass, :order, :family, :genus, :species, :commonNames, :ethnographicTerm, :event, :militaryBranch, :sportsTeam,\n :tribalClasses, :tribalTerms, :award, :legalName],\n :rights => [:rights, :useRestrictions, :accessRestrictions],\n :rightsHolder => [:rightsHolder],\n :license => [:license],\n :publisher => [:publisher, :od_repository],\n :provenance => [:provenance],\n :spatial => [:location, :tgn, :waterBasin, :rangerDistrict, :streetAddress, :geobox, :latitude, :longitude],\n :type => [:type],\n :language => [:language],\n :isPartOf => [:set, :containedInJournal, :localCollectionName, :isPartOf, :largerWork],\n :tableOfContents => [:tableOfContents],\n :temporal => [:temporal],\n :bibliographicCitation => [:citation],\n :relation => [:relation, :artSeries],\n :isReferencedBy => [:findingAid],\n :hasPart => [:hasPart],\n :isVersionOf => [:isVersionOf],\n :extent => [:extent, :physicalExtent, :measurements],\n :format => [:format]\n }\n end",
"def comparisons_ids_with_descriptions(comparisons)\n selected_ids = comparisons.select { |comparison| comparison.description.present? }.collect(&:id)\n\n { comparisonsIdsWithDescription: selected_ids }\n end",
"def get_original_combination_html\n nil\n end",
"def get_default_policies_and_programmes\n {\n \"0\" => { description: \"Transferencias internas\" },\n \"1\" => { description: \"Servicios públicos básicos\" },\n \"2\" => { description: \"Protección y promoción social\" },\n \"3\" => { description: \"Bienes públicos de carácter preferente\" },\n \"4\" => { description: \"Actuaciones de carácter económico\" },\n \"9\" => { description: \"Actuaciones de carácter general\" },\n \"00\" => { description: \"Transferencias internas\" },\n \"11\" => { description: \"Justicia\" },\n \"12\" => { description: \"Defensa\" },\n \"13\" => { description: \"Seguridad ciudadana e instituciones penitenciarias\" },\n \"14\" => { description: \"Política exterior\" },\n \"21\" => { description: \"Pensiones\" },\n \"22\" => { description: \"Otras prestaciones económicas\" },\n \"23\" => { description: \"Servicios sociales y promoción social\" },\n \"24\" => { description: \"Fomento del empleo\" },\n \"25\" => { description: \"Desempleo\" },\n \"26\" => { description: \"Acceso a la vivienda y fomento de la edificación\" },\n \"29\" => { description: \"Gestión y administración de la Seguridad Social\" },\n \"31\" => { description: \"Sanidad\" },\n \"32\" => { description: \"Educación\" },\n \"33\" => { description: \"Cultura\" },\n \"41\" => { description: \"Agricultura, pesca y alimentación\" },\n \"42\" => { description: \"Industria y energía\" },\n \"43\" => { description: \"Comercio, turismo y PYMES\" },\n \"44\" => { description: \"Subvenciones al transporte\" },\n \"45\" => { description: \"Infraestructuras\" },\n \"46\" => { description: \"Investigación, desarrollo e innovación\" },\n \"49\" => { description: \"Otras actuaciones de carácter económico\" },\n \"91\" => { description: \"Alta dirección\" },\n \"92\" => { description: \"Servicios de carácter general\" },\n \"93\" => { description: \"Administración financiera y tributaria\" },\n \"94\" => { description: \"Transferencias a otras admones. públicas\" },\n \"95\" => { description: \"Deuda pública\" } \n }\nend",
"def get_study_group_title\n temp = sg_title.text.delete('()').gsub('client_division_uuid','').gsub('uuid','').split(':').map(&:strip)\n name_and_uuid = {}\n name_and_uuid[:name] = temp.first\n name_and_uuid[:uuid] = temp.last\n name_and_uuid\n end",
"def social_description\n @social_description ||= social_description_card&.format(:text)&.text_description\n end",
"def description\n [@group.description,@description].join(' ')\n end",
"def set_title_and_description(merge_request)\n if match = merge_request.source_branch.match(/\\A(\\d+)-/)\n iid = match[1]\n end\n\n commits = merge_request.compare_commits\n if commits && commits.count == 1\n commit = commits.first\n merge_request.title = commit.title\n merge_request.description ||= commit.description.try(:strip)\n elsif iid && (issue = merge_request.target_project.get_issue(iid)) && !issue.try(:confidential?)\n case issue\n when Issue\n merge_request.title = \"Resolve \\\"#{issue.title}\\\"\"\n when ExternalIssue\n merge_request.title = \"Resolve #{issue.title}\"\n end\n else\n merge_request.title = merge_request.source_branch.titleize.humanize\n end\n\n if iid\n closes_issue = \"Closes ##{iid}\"\n\n if merge_request.description.present?\n merge_request.description += closes_issue.prepend(\"\\n\\n\")\n else\n merge_request.description = closes_issue\n end\n end\n\n merge_request\n end",
"def merge_duplicates_based_on_case\n POSSIBLY_WRONGLY_CASED_ATTRIBUTES.each do |attr|\n merged_runners = 0\n find_runners_only_differing_in(attr, [\"f_unaccent(lower(#{attr})) as low\"], ['low']).each do |entries|\n # We prefer the version with capital first letter and more lowercase\n # characters. E. g. for\n # Reichenbach I. K.\n # reichenbach i. K.\n # Reichenbach i. K.\n # the version at the bottom is preferred.\n merged_runners += reduce_to_one_runner_by_condition(entries) do |runner|\n [runner[attr][0] == runner[attr][0].upcase ? 1 : 0,\n runner[attr].scan(/[[:lower:]]/).size]\n end\n end\n puts \"Merged #{merged_runners} entries based on case of #{attr}.\" unless Rails.env.test?\n end\n end",
"def compute_description(scopes)\n scopes.reduce do |desc, element|\n desc = if element =~ /^(#|\\.|:)/ && desc.split(\" \").length == 1\n desc+element\n else\n \"#{desc} #{element}\"\n end\n end\n end",
"def normalized_reason_code_description(crosswalk_record = nil)\n reason_code_record = crosswalk_record || reason_code_object\n if reason_code_record.present?\n rcc_log.debug \"Obtaining Reason Code description having ID : #{reason_code_record.id} \"\n reason_code_description = reason_code_record.reason_code_description\n end\n rcc_log.debug \"Reason Code Description : #{reason_code_description}\"\n reason_code_description.to_s.upcase\n end",
"def general_explanation\n # The .html_safe is a bit lame, but we need to be able to include\n # HTML in this setting.\n Settings.send(\"#{slug}_build_pair_explanation\").html_safe\n end",
"def merge_descriptions\n return unless check_dest_permission!\n\n # Try merge.\n if perform_merge\n # Merged successfully.\n log_the_merge_flash_and_redirect\n else\n # If conflict: render edit form.\n warn_and_render_edit_description_form\n end\n end",
"def common_tags_to_add\n common_tags = []\n common_tags << self.merger\n common_tags << self if self.canonical\n common_tags << self.parents\n common_tags.flatten.uniq.compact\n end",
"def alter_description(description, additions, subtractions)\n new_description = description\n additions.each do |addition|\n stylized_weights = addition.weights&.join('->')\n if stylized_weights == '' || stylized_weights.nil?\n stylized_weights = 'body weight'\n end\n\n # This will look off for things I do 1 set or rep of,\n # but I'm ok with that - it makes eventually parsing\n # a dump of my data from Strava in the future easier.\n new_description << <<~EOF\n #{addition.exercise}: #{addition.num_sets} sets of #{addition.num_reps} reps at #{stylized_weights} lbs\n EOF\n end\n subtractions.each do |subtraction|\n new_description_array = new_description.split(\"\\n\").reject do |line|\n line.index(subtraction) == 0\n end\n new_description = new_description_array.join(\"\\n\")\n end\n\n new_description\nend",
"def merge_info\n num_obs = observations.count\n num_namings = namings.count\n \"#{:NAME.l} ##{id}: #{real_search_name} [o=#{num_obs}, n=#{num_namings}]\"\n end",
"def get_full_name\n # [Steve, 20140725] Too long/repetitive: \"#{description} #{header_year} - #{get_federation_type}\"\n description\n end",
"def format_acknowledgements()\n s = \"<ul>\"\n if @book.style.attribute?\n s << \"<li>Layout and style '#{@book.style.name}' provided by \"\n if @book.style.author_url\n s << \"<a href='#{@book.style.author_url}'>#{@book.style.author}</a>\"\n else\n s << @book.style.author\n end\n s << \".\"\n s << \"</ul><ul>\"\n end\n @book.style.assets.each do |a|\n next unless a.attribute?\n next unless a.name\n s << \"<li>\"\n if a.collection_url\n s << \"<a href='#{a.collection_url}'>#{a.name}</a>\"\n elsif a.name\n s << a.name\n end\n s << \" is provided by \"\n if a.author_url\n s << \"<a href='#{a.author_url}'>#{a.author}</a>\"\n elsif a.author\n s << a.author\n else\n s << \"anonymous\"\n end\n if a.collection\n s << \" as part of \"\n s << \"#{a.collection}\"\n end\n s << \".</li>\"\n end\n s << \"</ul>\"\n end",
"def og_desc(desc)\n\t\tbase_desc = \"Informing and entertaining the global internet generation.\"\n\t\tif desc.empty?\n\t\t\tbase_desc\n\t\telse\n\t\t\tdesc.to_s\n\t\tend\n\tend",
"def product_description\n return product_presentation.description if product_presentation\n return \"\"\n end",
"def composite_med_proc_id\n elem = []\n if ((!cpt_code.blank? and !revenue_code.blank?)|| revenue_code.blank?)\n proc_code = \"HC:#{cpt_code}\"\n else\n proc_code = \"NU:#{revenue_code}\"\n end\n\n # proc_code = ((!cpt_code.blank? and !revenue_code.blank?) || revenue_code.blank?) ? \"HC:#{cpt_code}\" : \"NU:#{revenue_code}\"\n elem = [proc_code, service.service_modifier1 , service.service_modifier2 ,\n service.service_modifier3 , service.service_modifier4]\n elem = Output835.trim_segment(elem)\n elem.join(':')\n end",
"def get_meta_description(include_color, include_defaults)\n \"Soletron, which is owned by Superbowl MVP Santonio Holmes, offers the finest #{self.name_from_taxon} Products\"\n end",
"def one_of_description\n\t\treturn nil if self.one_of.empty?\n\t\treturn \" with at least one of: %s\" % [ self.one_of.map(&:name).join(', ') ]\n\tend",
"def meta_description\n description = [\"Organize your U.S. coin collection with free software. \"]\n if @coin and !@coin.new_record?\n description << [@coin.meta_description]\n end\n description << page_no\n raw description.join(\" \")\n end",
"def meta_description\n description = [\"Description here.\"] # Change this for your site\n description << @category.title if @category\n description << [@post.title, @post.cat_title ] if @post\n description.join(\". \")\n end",
"def uppercase_text\n if self.desc.present?\n self.desc.upcase! \n self.desc.strip!\n end\n self.part_code.upcase! if self.part_code.present?\n self.tarif_code.upcase! if self.tarif_code.present?\n self.origin.upcase! if self.origin.present?\n self.brand.upcase! if self.brand.present?\n self.grade.upcase! if self.grade.present?\n self.packing_method.upcase! if self.packing_method.present?\n self.container_no.upcase! if self.container_no.present?\n self.seal_no.upcase! if self.seal_no.present?\n self.despatch_advise.upcase! if self.despatch_advise.present?\n self.size.upcase! if self.size.present?\n self.model_no.upcase! if self.model_no.present?\n self.reorder_no.upcase! if self.reorder_no.present?\n self.art_no.upcase! if self.art_no.present?\n self.ref_no.upcase! if self.ref_no.present?\n self.work_order_no.upcase! if self.work_order_no.present?\n self.film_no.upcase! if self.film_no.present?\n self.bar_code.upcase! if self.bar_code.present?\n self.lot_no.upcase! if self.lot_no.present?\n self.item_code.upcase! if self.item_code.present?\n self.pantone_code.upcase! if self.pantone_code.present?\n self.color.upcase! if self.color.present?\n self.serial_no.upcase! if self.serial_no.present?\n self.warranty.upcase! if self.warranty.present?\n self.capacity.upcase! if self.capacity.present?\n self.consignee.upcase! if self.consignee.present?\n self.destination.upcase! if self.destination.present?\n self.process.upcase! if self.process.present?\n self.mould_no.upcase! if self.mould_no.present?\n self.category.upcase! if self.category.present?\n self.revision.upcase! if self.revision.present?\n self.window_code.upcase! if self.window_code.present?\n end",
"def long_appearance\n # How Med does it:\n # description\n # glance\n # eq\n # inventory\n \"{!{FG#{description}\\n\\n{FY{#{short_name} is in excellent condition.\"\n end",
"def content_description_from_essence_type(element, essence_type)\n {\n 'type' => essence_type,\n 'name' => content_name_from_element_and_essence_type(element, essence_type)\n }\n end",
"def description\n values = super\n values = Deepblue::MetadataHelper.ordered( ordered_values: description_ordered, values: values )\n return values\n end",
"def get_organization\n return \"NSBE\" if affiliation == 1\n \"SHPE\"\n end",
"def common_word\n words = {}\n record.records.each do |record|\n record.title.split(' ').each do |word|\n words[word] = 0 if words[word].nil?\n words[word] += 1\n end\n end\n return if words.empty?\n words.sort.first[0]\n end",
"def get_description( forte_set_name )\n dictionary_entry = @dictionary[forte_set_name]\n dictionary_entry.nil? ? nil : dictionary_entry[2].clone\n end",
"def map_oai_dc\n { :title => [:title, :alternative],\n :description => [:description, :abstract, :descriptionOfManifestation, :culturalContext, :stylePeriod, :awardDate],\n :date => [:date, :created, :viewDate, :earliestDate],\n :identifier => [:accessURL],\n :creator => [:creator],\n :contributor => [:arranger, :artist, :author, :cartographer, :collector, :composer, :contributor, :editor, :illustrator,\n :interviewee, :interviewer, :lyricist, :patron, :photographer, :printMaker, :scribe, :transcriber, :translator,\n :dedicatee, :recipient, :owner, :designer, :landscapeArchitect],\n :subject => [:lcsubject, :subject, :phylum, :taxonClass, :order, :family, :genus, :species, :commonNames, :ethnographicTerm, :event, :militaryBranch, :sportsTeam,\n :tribalClasses, :tribalTerms, :award, :legalName],\n :coverage => [:location, :tgn, :waterBasin, :rangerDistrict, :streetAddress, :temporal, :geobox, :latitude, :longitude],\n :publisher => [:publisher, :od_repository],\n :type => [:type],\n :language => [:language],\n :relation => [:relation, :set, :localCollectionName, :artSeries, :findingAid],\n :rights => [:rights, :rightsHolder, :license, :useRestrictions, :accessRestrictions],\n :format => [:format]\n }\n end",
"def upc_a_with_composite_symbology; end",
"def descriptions\n return attributes[:descriptions] if attributes[:descriptions]\n attributes[:descriptions] = {}\n attributes[:descriptions]\n end",
"def default_value_for_description\n warn(\"Using README as description\")\n # RubyGems refuses to build a gem if the description contains `FIXME` or `TODO`,\n # which are perfectly valid words to use in a description, but alas.\n @chunked_source.readme.gsub(/FIXME/i, \"FIZZIX-ME\").gsub(/TODO/i, \"TOODLES\")\n end",
"def description\n return @description if @description\n \n parts = []\n \n tp = trial_period(false)\n parts << \"#{tp}-day trial\" if tp && tp > 0\n \n sa = setup_amount(false)\n parts << \"#{number_to_currency(sa)} setup fee\" if sa && sa > 0\n \n am = amount(false)\n parts << \"#{number_to_currency(am)}/mo\"\n \n if prepaid_message_count >= UNLIMITED\n texts = \"unlimited\"\n else\n texts = \"#{prepaid_message_count}/mo prepaid\"\n end\n \n @description = parts.join(', ') + \" and #{texts} texts after that.\"\n\n return @description\n end",
"def concat_description\n id = Readline.readline(\"ID of task to concatenate to: \").to_i\n str = Readline.readline(\"Information to concatenate: \").chomp\n ok = ConcatDescription.new(id, str).execute\n puts \"No such issue\" if !ok\n end",
"def description\n description_section.join(\" \").tr(\"\\n\", ' ').gsub(/[{}]/,'').gsub(/\\[[^\\]]+\\]/,'') # strip rdoc\n end",
"def collect_places_information\n ### Places information\n mark_data_places_information = Array.new\n places_information = @case_details[:case][:system][:places]\n\n mark_data_places_information << 'Answer the following about each places present in the evidence on the system'\n places_information.each do | place_category, place_information |\n mark_data_places_information << place_category.capitalize\n place_information.each { | key, value | mark_data_places_information << \"-#{key.capitalize}::#{value}\" }\n mark_data_places_information << ''\n end\n\n return mark_data_places_information\n end",
"def generate_description\n\t\t\tself.log.debug \"Auto-generating description for %p\" % [ self ]\n\t\t\tdesc = self.name.to_s.\n\t\t\t\tgsub( /.*\\[(\\w+)\\]/, \"\\\\1\" ).\n\t\t\t\tgsub( /_(.)/ ) {|m| \" \" + m[1,1].upcase }.\n\t\t\t\tgsub( /^(.)/ ) {|m| m.upcase }\n\t\t\tself.log.debug \" generated: %p\" % [ desc ]\n\t\t\treturn desc\n\t\tend",
"def existing_kases_list_header_in_words(kind)\n case kind\n when :idea then \"Existing Ideas in the Community\".t\n when :question then \"Existing Questions in the Community\".t\n when :problem then \"Existing Problems in the Community\".t\n when :praise then \"Existing Praise in the Community\".t\n else \"Existing Cases in the Community\".t\n end\n end",
"def extra_description\n attributes_description + content_description + count_description\n end",
"def test_overriding_description\r\n\t\t# first, a feature with no description assigned\r\n\t\tfeatures = Feature.for_static_page(static_pages(:aboutus).id)\r\n\r\n\t\tassert_equal(\"Short description\", features.first[:short_description])\r\n\r\n\t\t# now, a feature with a description to override\r\n\t\tfeatures = Feature.for_category(categories(:livestock).id)\r\n\r\n\t\tassert_equal(\"Different description\", features.first[:short_description])\r\n\tend",
"def full_ipa_phrase(delimiter = '')\n @full_ipa_phrase ||= {}\n @full_ipa_phrase[delimiter] ||= entry_chain.map(&:match).compact.join(delimiter)\n end",
"def main_description; end",
"def descr_short\n descr = self[:descr].to_s.gsub(\"\\n\", \" \").gsub(/\\s{2,}/, \" \")\n descr = Knj::Strings.shorten(descr, 20)\n #descr = \"[#{_(\"no description\")}]\" if descr.to_s.strip.length <= 0\n return descr\n end",
"def text\n #[@organism, @symbol, @description, @aka, @protnames, @summary,@comentaries.join(\". \")].join(\". \") \n [@symbol, @description, @aka, @protnames, @summary].flatten.join(\". \") \n end",
"def description\n meta_description.nil? ? secondary_description : meta_description\n end",
"def description\n meta_description.nil? ? secondary_description : meta_description\n end",
"def combine\n other_survivor_supplies = [ \"warm clothes\", \"rations\", \"compass\", \"camp stove\",\n \"solar battery\", \"flashlight\"]\n other_survivor_supplies.each do |others|\n @zombie_apocalypse_supplies.push(others)\n end\n @zombie_apocalypse_supplies = @zombie_apocalypse_supplies.uniq\n\nend",
"def description\n description = {}\n for name in @hash.keys\n text = @hash[name].last.description\n if @hash[name].last.depends and @hash[name].last.depends.length > 0\n depends = ' [' + @hash[name].last.depends.join(', ') + ']'\n else\n depends = nil\n end\n description[name] = (text ? text : '') + (depends ? depends : '')\n end\n return description\n end",
"def html_description2\n ::Util.simple_format(self.description2)\n end",
"def description_bonita\n \t\"#{description} #{product.name}\"\n end",
"def nyc_pigeon_organizer_two(data)\n# MAP will return a changed value, EACH will return original\n\n # coerce each name as the key in the hash\n #access all names\n #names - ['']\n # unique list of names\n names = data.values[0].values.flatten.uniq\n\n # use name as key in hash - fill in with scaffold hash\n initial_structure = names.each_with_object ({}) do |name, hash|\n hash[name] = {color: [], gender: [], lives:[]}\n end\n\n names.each_with_object(initial_structure) do |pigeon, my_initial_structure|\n attributes = data.keys\n attributes.each do |attribute|\n data[attribute].each do |value, names|\n if names.include?(pigeon)\n my_initial_structure[pigeon][attribute] << value.to_s\n end\n end\n end\n end\n\n\nend",
"def program_long_desc(desc)\n abstract!\n end",
"def upc_a_with_composite_symbology\n \"#{upc_a}|#{bothify(parse('barcode.composite_symbol'))}\"\n end",
"def description_complexity(pr)\n pull_req = pull_req_entry(pr[:id])\n (pull_req['title'] + ' ' + pull_req['body']).gsub(/[\\n\\r]\\s+/, ' ').split(/\\s+/).size\n end",
"def combine_anagrams(words) \n # for each word, sort the word to create the \n # signature for the anagrams\n anagram_groups = {}\n words.each do |w|\n key = w.downcase.chars.sort.join\n if anagram_groups[key].nil?\n anagram_groups[key] = [w]\n else\n anagram_groups[key].push(w)\n end\n end\n \n return anagram_groups.values\n\nend",
"def metadata_cities(structure)\n place = []\n courses = structure.courses\n\n place << 'A domicile' if courses.where(type: 'Course::Private').any?\n if (public_courses = courses.where.not(type: 'Course::Private')).any?\n place += public_courses.flat_map(&:places).flat_map(&:city).uniq.map(&:name).sort\n end\n\n place.first(3).to_sentence\n end"
] |
[
"0.58629906",
"0.5643776",
"0.5629066",
"0.55231047",
"0.5436626",
"0.5427282",
"0.53373724",
"0.5335062",
"0.5315231",
"0.53100395",
"0.5269072",
"0.5264237",
"0.52463824",
"0.5243661",
"0.523188",
"0.5208591",
"0.52035874",
"0.51994884",
"0.5180714",
"0.5176603",
"0.5171239",
"0.51616615",
"0.51414436",
"0.5138139",
"0.51128423",
"0.5093141",
"0.5083794",
"0.5076672",
"0.50635654",
"0.50619024",
"0.5059321",
"0.5056084",
"0.5050118",
"0.5038892",
"0.50283426",
"0.5019171",
"0.5014994",
"0.5014866",
"0.5000855",
"0.498694",
"0.49742624",
"0.49722838",
"0.4968065",
"0.49668467",
"0.4962879",
"0.49618158",
"0.49552223",
"0.49536714",
"0.4953343",
"0.49504998",
"0.4941418",
"0.49368665",
"0.4936762",
"0.49169126",
"0.4915642",
"0.49107936",
"0.49040005",
"0.49024835",
"0.4900465",
"0.48968175",
"0.48853254",
"0.4869834",
"0.4869609",
"0.4865939",
"0.48569727",
"0.48546368",
"0.48542434",
"0.485242",
"0.48453262",
"0.4838064",
"0.4838",
"0.48343414",
"0.48317978",
"0.48275805",
"0.4824853",
"0.4820332",
"0.4819803",
"0.48104456",
"0.480942",
"0.47968635",
"0.47954732",
"0.4791679",
"0.47898233",
"0.47887626",
"0.47873583",
"0.4786913",
"0.4785099",
"0.47846094",
"0.47766516",
"0.47766516",
"0.47724584",
"0.47701123",
"0.47686347",
"0.47650295",
"0.4758717",
"0.47410947",
"0.47399282",
"0.47359434",
"0.47332817",
"0.47332418"
] |
0.6304396
|
0
|
Makes common apartments features. Hash, based on constant APARTMENT_FEATURES, first merge it, then find unique features
|
def apartment_features
apartment_features = {}
apartment_features.merge!(APARTMENT_FEATURES)
features_unavailable.each { |feature| apartment_features[feature] = false }
apartment_features
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def get_required_features(features)\n features.merge(@features) unless @features.nil?\n end",
"def get_required_features(features)\n features.merge(@features) unless @features.nil?\n end",
"def selected_features(activating_contexts, active_contexts)\n concerned = concerned_features(activating_contexts)\n selected = []\n # activated_contexts = activating_contexts + active_contexts\n \n concerned.each {\n |feature|\n concerned_contexts = @feature_to_contexts[feature]\n # TODO remove already active features from active contexts\n # then create relevant_contexts and include feature in\n # selected the number of times its mapping appears in\n # relevant_contexts\n }\n end",
"def get_features(frag, model)\n w1 = (frag.cleaned.last or '')\n w2 = (frag.next or '')\n\n frag.features = [\"w1_#{w1}\", \"w2_#{w2}\", \"both_#{w1}_#{w2}\"]\n\n if not w2.empty?\n if w1.chop.is_alphabetic? \n frag.features.push \"w1length_#{[10, w1.length].min}\", \"w1abbr_#{model.non_abbrs[w1.chop]}\"\n end\n\n if w2.chop.is_alphabetic?\n frag.features.push \"w2cap_#{w2[0,1].is_upper_case?}\", \"w2lower_#{model.lower_words[w2.downcase]}\"\n end\n end\n end",
"def hash_genetics (my_features={}, sig_other={})\n \n\t\tif my_features.respond_to? :merge and sig_other.respond_to? :merge\n \n\t\t\tchild = sig_other.merge(my_features)\n\t\telse\n\t\t\treturn \"error\"\n\t\tend\n\tend",
"def get_features(frag, model)\n w1 = (frag.cleaned.last or '')\n w2 = (frag.next or '')\n\n frag.features = [\"w1_#{w1}\", \"w2_#{w2}\", \"both_#{w1}_#{w2}\"]\n\n if not w2.empty?\n if w1.chop.is_alphabetic? \n frag.features.push \"w1length_#{[10, w1.length].min}\"\n frag.features.push \"w1abbr_#{model.non_abbrs[w1.chop]}\"\n end\n\n if w2.chop.is_alphabetic?\n frag.features.push \"w2cap_#{w2[0].is_upper_case?}\"\n frag.features.push \"w2lower_#{model.lower_words[w2.downcase]}\"\n end\n end\n end",
"def get_required_features(features)\n end",
"def features\n hash = {}\n hash['price_range'] = [(our_price_range_low / 100).floor, \n (our_price_range_high / 100).floor]\n hash['locality'] = venue.locality\n hash['category'] = categories\n return hash\n end",
"def parse_features\n check = Hash.new(0)\n @features.each do |feat|\n feature = feat.feature\n position = feat.position\n\n # try to link gene-related features (CDS, mRNA etc.) by matching /locus_tag or /gene qualifier values\n qual = feat.to_hash\n gene_name = locus_tag = gene = nil\n if qual[\"locus_tag\"]\n if locus_tag = qual[\"locus_tag\"].first\n gene_name = locus_tag\n end\n elsif qual[\"gene\"]\n if gene = qual[\"gene\"].first\n gene_name = gene\n end\n end\n\n @feature_count[feature] += 1\n locations = Bio::Locations.new(position)\n min, max = locations.span\n strand = locations.first.strand\n feature_id = new_feature_uri(feature, min, max, strand, @feature_count[feature])\n if feature == \"gene\"\n @gene[gene_name] = feature_id\n check[gene_name] += 1\n end\n gene_id = @gene[gene_name] # expect that \"gene\" feature appears before \"CDS\", \"mRNA\", etc. features derived from a gene (otherwize use parse_genes() for fail-safe)\n\n # add type by Sequence Ontology\n so_id = \"SO:0000001\"\n so_obo_id = \"obo:SO_0000001\"\n so_term = \"region\"\n ft_id = \"Feature\"\n if so_id = @ft_so.so_id(feature)\n if so_id != \"undefined\"\n so_obo_id = @ft_so.obo_id(so_id)\n so_term = @ft_so.so_term(feature)\n ft_id = @ft_so.ft_id(feature)\n end\n end\n\n # feature types and labels\n puts triple(feature_id, \"rdf:type\", \"insdc:#{ft_id}\")\n puts triple(feature_id, \"rdfs:subClassOf\", so_obo_id) + \" # SO:#{so_term}\"\n # to make compatible with Ensembl RDF\n puts triple(feature_id, \"obo:RO_0002162\", \"<http://identifiers.org/taxonomy/#{@taxonomy_id}>\") + \" # RO:in taxon\"\n puts triple(feature_id, \"rdfs:label\", quote(locus_tag || gene || feature))\n puts triple(feature_id, \"dc:identifier\", quote(locus_tag || gene)) if locus_tag || gene\n if locus_tag || gene\n puts triple(feature_id, \"skos:prefLabel\", quote(locus_tag || gene))\n if qual[\"gene_synonym\"]\n qual[\"gene_synonym\"].first.split(/;\\s+/).each do |synonym|\n puts triple(feature_id, \"skos:altLabel\", quote(synonym))\n end\n end\n end\n\n # feature qualifiers\n parse_qualifiers(feature_id, qual)\n\n # parent-child relationship (gene -> mRNA|CDS|misc_RNA etc.)\n parent_uri = @sequence_uri\n if gene_id and gene_id != feature_id\n parent_uri = gene_id\n puts triple(feature_id, \"sio:SIO_010081\", gene_id) + \" # sio:is-transcribed-from\"\n # to make compatible with Ensembl RDF\n puts triple(feature_id, \"rdfs:subClassOf\", \"obo:SO_0000673\") + \" # SO:transcript\"\n if encoding = @ft_so.so_encoding(feature)\n if feature == \"ncRNA\"\n if ncrna_class = qual[\"ncRNA_class\"]\n if @ft_so.so_encoding(ncrna_class.first)\n encoding = @ft_so.so_encoding(ncrna_class.first)\n end\n end\n end\n so_encoding_id = encoding[\"so_id\"]\n so_encoding_term = encoding[\"so_term\"]\n puts triple(gene_id, \"rdfs:subClassOf\", @ft_so.obo_id(so_encoding_id)) + \" # SO:#{so_encoding_term}\"\n end\n end\n puts triple(feature_id, \"obo:so_part_of\", parent_uri)\n\n # add FALDO location and subparts (exons etc.)\n region_id, locations = new_location(feature_id, position)\n # Uncomment to eliminate single exon genes.\n #if locations.count > 1\n if gene_id\n # link to exons in join(exon1, exon2, ...)\n feature_type = { :id => \"obo:SO_0000147\", :term => \"exon\", :ft => \"Exon\" }\n else\n # [TODO] need to confirm that if there are any features having subparts other than exons\n feature_type = { :id => \"obo:SO_0000001\", :term => \"region\", :ft => \"Feature\" }\n end\n sub_parts, sub_ordered_parts = add_subparts(locations, feature_type)\n #puts triple(feature_id, \"obo:so_has_part\", \"(#{sub_parts.join(' ')})\") # rdf:List\n # exon URIs\n puts triple(feature_id, \"obo:so_has_part\", sub_parts.join(', '))\n # part URIs\n puts triple(feature_id, \"sio:SIO_000974\", sub_ordered_parts.join(', ')) + \" # sio:has-ordered-part\"\n #end\n end\n $stderr.puts \"Features: #{@feature_count.to_json}\"\n check.each do |k, v|\n $stderr.puts \"Warning: gene ID #{k} occured #{v} times\" if v > 1\n end\n end",
"def configured_feature_flags\n feature_flag_options.sort_by(&:name).map(&:to_h).reduce({}, :merge).with_indifferent_access\n end",
"def addFeaturesAndLabel(earliest_date, latest_date, examples, labels)\n all_games = Game.where(\"game_date > ? AND game_date < ?\", earliest_date, latest_date)\n\n all_games.each do |game|\n feature = Feature.find_by_game_id(game.id)\n if feature == nil\n feature = Feature.new\n feature.game_id = game.id\n feature.home_team_won = game.home_team_won\n feature.save\n end\n\n feature_set = []\n\n # Add in individual features\n feature_set << feature.h2h_diff_1\n feature_set << feature.h2h_diff_2\n feature_set << feature.h2h_diff_3\n \n#=begin\n feature_set << feature.run_differentials_1\n feature_set << feature.opp_differentials_1\n feature_set << feature.run_differentials_2\n feature_set << feature.opp_differentials_2\n feature_set << feature.run_differentials_5\n feature_set << feature.opp_differentials_5\n feature_set << feature.run_differentials_10\n feature_set << feature.opp_differentials_10\n feature_set << feature.run_differentials_20\n feature_set << feature.opp_differentials_20\n\n feature_set << feature.win_differentials_1\n feature_set << feature.opp_win_differentials_1\n feature_set << feature.win_differentials_2\n feature_set << feature.opp_win_differentials_2\n feature_set << feature.win_differentials_5\n feature_set << feature.opp_win_differentials_5\n feature_set << feature.win_differentials_10\n feature_set << feature.opp_win_differentials_10\n feature_set << feature.win_differentials_20\n feature_set << feature.opp_win_differentials_20\n#=end\n\n=begin\n # Add in the differences between features. Could be preferable?\n feature_set << feature.run_differentials_1 - feature.opp_differentials_1\n feature_set << feature.run_differentials_2 - feature.opp_differentials_2\n feature_set << feature.run_differentials_5 - feature.opp_differentials_5\n feature_set << feature.run_differentials_10 - feature.opp_differentials_10\n feature_set << feature.run_differentials_20 - feature.opp_differentials_20\n \n feature_set << feature.win_differentials_1 - feature.opp_win_differentials_1\n feature_set << feature.win_differentials_2 - feature.opp_win_differentials_2\n feature_set << feature.win_differentials_5 - feature.opp_win_differentials_5\n feature_set << feature.win_differentials_10 - feature.opp_win_differentials_10\n feature_set << feature.win_differentials_20 - feature.opp_win_differentials_20\n=end\n\n#=begin\n feature_set << feature.home_batting_spot_1_walks_last_1_game\n feature_set << feature.home_batting_spot_2_walks_last_1_game\n feature_set << feature.home_batting_spot_3_walks_last_1_game\n feature_set << feature.home_batting_spot_4_walks_last_1_game\n feature_set << feature.home_batting_spot_5_walks_last_1_game\n feature_set << feature.home_batting_spot_6_walks_last_1_game\n feature_set << feature.home_batting_spot_7_walks_last_1_game\n feature_set << feature.home_batting_spot_8_walks_last_1_game\n feature_set << feature.home_batting_spot_9_walks_last_1_game\n\n feature_set << feature.away_batting_spot_1_walks_last_1_game\n feature_set << feature.away_batting_spot_2_walks_last_1_game\n feature_set << feature.away_batting_spot_3_walks_last_1_game\n feature_set << feature.away_batting_spot_4_walks_last_1_game\n feature_set << feature.away_batting_spot_5_walks_last_1_game\n feature_set << feature.away_batting_spot_6_walks_last_1_game\n feature_set << feature.away_batting_spot_7_walks_last_1_game\n feature_set << feature.away_batting_spot_8_walks_last_1_game\n feature_set << feature.away_batting_spot_9_walks_last_1_game\n\n feature_set << feature.home_batting_spot_1_walks_last_2_games\n feature_set << feature.home_batting_spot_2_walks_last_2_games\n feature_set << feature.home_batting_spot_3_walks_last_2_games\n feature_set << feature.home_batting_spot_4_walks_last_2_games\n feature_set << feature.home_batting_spot_5_walks_last_2_games\n feature_set << feature.home_batting_spot_6_walks_last_2_games\n feature_set << feature.home_batting_spot_7_walks_last_2_games\n feature_set << feature.home_batting_spot_8_walks_last_2_games\n feature_set << feature.home_batting_spot_9_walks_last_2_games\n\n feature_set << feature.away_batting_spot_1_walks_last_2_games\n feature_set << feature.away_batting_spot_2_walks_last_2_games\n feature_set << feature.away_batting_spot_3_walks_last_2_games\n feature_set << feature.away_batting_spot_4_walks_last_2_games\n feature_set << feature.away_batting_spot_5_walks_last_2_games\n feature_set << feature.away_batting_spot_6_walks_last_2_games\n feature_set << feature.away_batting_spot_7_walks_last_2_games\n feature_set << feature.away_batting_spot_8_walks_last_2_games\n feature_set << feature.away_batting_spot_9_walks_last_2_games\n\n feature_set << feature.home_batting_spot_1_walks_last_5_games\n feature_set << feature.home_batting_spot_2_walks_last_5_games\n feature_set << feature.home_batting_spot_3_walks_last_5_games\n feature_set << feature.home_batting_spot_4_walks_last_5_games\n feature_set << feature.home_batting_spot_5_walks_last_5_games\n feature_set << feature.home_batting_spot_6_walks_last_5_games\n feature_set << feature.home_batting_spot_7_walks_last_5_games\n feature_set << feature.home_batting_spot_8_walks_last_5_games\n feature_set << feature.home_batting_spot_9_walks_last_5_games\n\n feature_set << feature.away_batting_spot_1_walks_last_5_games\n feature_set << feature.away_batting_spot_2_walks_last_5_games\n feature_set << feature.away_batting_spot_3_walks_last_5_games\n feature_set << feature.away_batting_spot_4_walks_last_5_games\n feature_set << feature.away_batting_spot_5_walks_last_5_games\n feature_set << feature.away_batting_spot_6_walks_last_5_games\n feature_set << feature.away_batting_spot_7_walks_last_5_games\n feature_set << feature.away_batting_spot_8_walks_last_5_games\n feature_set << feature.away_batting_spot_9_walks_last_5_games\n\n feature_set << feature.home_batting_spot_1_walks_last_10_games\n feature_set << feature.home_batting_spot_2_walks_last_10_games\n feature_set << feature.home_batting_spot_3_walks_last_10_games\n feature_set << feature.home_batting_spot_4_walks_last_10_games\n feature_set << feature.home_batting_spot_5_walks_last_10_games\n feature_set << feature.home_batting_spot_6_walks_last_10_games\n feature_set << feature.home_batting_spot_7_walks_last_10_games\n feature_set << feature.home_batting_spot_8_walks_last_10_games\n feature_set << feature.home_batting_spot_9_walks_last_10_games\n\n feature_set << feature.away_batting_spot_1_walks_last_10_games\n feature_set << feature.away_batting_spot_2_walks_last_10_games\n feature_set << feature.away_batting_spot_3_walks_last_10_games\n feature_set << feature.away_batting_spot_4_walks_last_10_games\n feature_set << feature.away_batting_spot_5_walks_last_10_games\n feature_set << feature.away_batting_spot_6_walks_last_10_games\n feature_set << feature.away_batting_spot_7_walks_last_10_games\n feature_set << feature.away_batting_spot_8_walks_last_10_games\n feature_set << feature.away_batting_spot_9_walks_last_10_games\n\n feature_set << feature.home_batting_spot_1_walks_last_20_games\n feature_set << feature.home_batting_spot_2_walks_last_20_games\n feature_set << feature.home_batting_spot_3_walks_last_20_games\n feature_set << feature.home_batting_spot_4_walks_last_20_games\n feature_set << feature.home_batting_spot_5_walks_last_20_games\n feature_set << feature.home_batting_spot_6_walks_last_20_games\n feature_set << feature.home_batting_spot_7_walks_last_20_games\n feature_set << feature.home_batting_spot_8_walks_last_20_games\n feature_set << feature.home_batting_spot_9_walks_last_20_games\n\n feature_set << feature.away_batting_spot_1_walks_last_20_games\n feature_set << feature.away_batting_spot_2_walks_last_20_games\n feature_set << feature.away_batting_spot_3_walks_last_20_games\n feature_set << feature.away_batting_spot_4_walks_last_20_games\n feature_set << feature.away_batting_spot_5_walks_last_20_games\n feature_set << feature.away_batting_spot_6_walks_last_20_games\n feature_set << feature.away_batting_spot_7_walks_last_20_games\n feature_set << feature.away_batting_spot_8_walks_last_20_games\n feature_set << feature.away_batting_spot_9_walks_last_20_games\n\n feature_set << feature.home_batting_spot_1_batting_percentage_last_1_game\n feature_set << feature.home_batting_spot_2_batting_percentage_last_1_game\n feature_set << feature.home_batting_spot_3_batting_percentage_last_1_game\n feature_set << feature.home_batting_spot_4_batting_percentage_last_1_game\n feature_set << feature.home_batting_spot_5_batting_percentage_last_1_game\n feature_set << feature.home_batting_spot_6_batting_percentage_last_1_game\n feature_set << feature.home_batting_spot_7_batting_percentage_last_1_game\n feature_set << feature.home_batting_spot_8_batting_percentage_last_1_game\n feature_set << feature.home_batting_spot_9_batting_percentage_last_1_game\n\n feature_set << feature.away_batting_spot_1_batting_percentage_last_1_game\n feature_set << feature.away_batting_spot_2_batting_percentage_last_1_game\n feature_set << feature.away_batting_spot_3_batting_percentage_last_1_game\n feature_set << feature.away_batting_spot_4_batting_percentage_last_1_game\n feature_set << feature.away_batting_spot_5_batting_percentage_last_1_game\n feature_set << feature.away_batting_spot_6_batting_percentage_last_1_game\n feature_set << feature.away_batting_spot_7_batting_percentage_last_1_game\n feature_set << feature.away_batting_spot_8_batting_percentage_last_1_game\n feature_set << feature.away_batting_spot_9_batting_percentage_last_1_game\n\n feature_set << feature.home_batting_spot_1_batting_percentage_last_2_games\n feature_set << feature.home_batting_spot_2_batting_percentage_last_2_games\n feature_set << feature.home_batting_spot_3_batting_percentage_last_2_games\n feature_set << feature.home_batting_spot_4_batting_percentage_last_2_games\n feature_set << feature.home_batting_spot_5_batting_percentage_last_2_games\n feature_set << feature.home_batting_spot_6_batting_percentage_last_2_games\n feature_set << feature.home_batting_spot_7_batting_percentage_last_2_games\n feature_set << feature.home_batting_spot_8_batting_percentage_last_2_games\n feature_set << feature.home_batting_spot_9_batting_percentage_last_2_games\n\n feature_set << feature.away_batting_spot_1_batting_percentage_last_2_games\n feature_set << feature.away_batting_spot_2_batting_percentage_last_2_games\n feature_set << feature.away_batting_spot_3_batting_percentage_last_2_games\n feature_set << feature.away_batting_spot_4_batting_percentage_last_2_games\n feature_set << feature.away_batting_spot_5_batting_percentage_last_2_games\n feature_set << feature.away_batting_spot_6_batting_percentage_last_2_games\n feature_set << feature.away_batting_spot_7_batting_percentage_last_2_games\n feature_set << feature.away_batting_spot_8_batting_percentage_last_2_games\n feature_set << feature.away_batting_spot_9_batting_percentage_last_2_games\n\n feature_set << feature.home_batting_spot_1_batting_percentage_last_5_games\n feature_set << feature.home_batting_spot_2_batting_percentage_last_5_games\n feature_set << feature.home_batting_spot_3_batting_percentage_last_5_games\n feature_set << feature.home_batting_spot_4_batting_percentage_last_5_games\n feature_set << feature.home_batting_spot_5_batting_percentage_last_5_games\n feature_set << feature.home_batting_spot_6_batting_percentage_last_5_games\n feature_set << feature.home_batting_spot_7_batting_percentage_last_5_games\n feature_set << feature.home_batting_spot_8_batting_percentage_last_5_games\n feature_set << feature.home_batting_spot_9_batting_percentage_last_5_games\n\n feature_set << feature.away_batting_spot_1_batting_percentage_last_5_games\n feature_set << feature.away_batting_spot_2_batting_percentage_last_5_games\n feature_set << feature.away_batting_spot_3_batting_percentage_last_5_games\n feature_set << feature.away_batting_spot_4_batting_percentage_last_5_games\n feature_set << feature.away_batting_spot_5_batting_percentage_last_5_games\n feature_set << feature.away_batting_spot_6_batting_percentage_last_5_games\n feature_set << feature.away_batting_spot_7_batting_percentage_last_5_games\n feature_set << feature.away_batting_spot_8_batting_percentage_last_5_games\n feature_set << feature.away_batting_spot_9_batting_percentage_last_5_games\n\n feature_set << feature.home_batting_spot_1_batting_percentage_last_10_games\n feature_set << feature.home_batting_spot_2_batting_percentage_last_10_games\n feature_set << feature.home_batting_spot_3_batting_percentage_last_10_games\n feature_set << feature.home_batting_spot_4_batting_percentage_last_10_games\n feature_set << feature.home_batting_spot_5_batting_percentage_last_10_games\n feature_set << feature.home_batting_spot_6_batting_percentage_last_10_games\n feature_set << feature.home_batting_spot_7_batting_percentage_last_10_games\n feature_set << feature.home_batting_spot_8_batting_percentage_last_10_games\n feature_set << feature.home_batting_spot_9_batting_percentage_last_10_games\n\n feature_set << feature.away_batting_spot_1_batting_percentage_last_10_games\n feature_set << feature.away_batting_spot_2_batting_percentage_last_10_games\n feature_set << feature.away_batting_spot_3_batting_percentage_last_10_games\n feature_set << feature.away_batting_spot_4_batting_percentage_last_10_games\n feature_set << feature.away_batting_spot_5_batting_percentage_last_10_games\n feature_set << feature.away_batting_spot_6_batting_percentage_last_10_games\n feature_set << feature.away_batting_spot_7_batting_percentage_last_10_games\n feature_set << feature.away_batting_spot_8_batting_percentage_last_10_games\n feature_set << feature.away_batting_spot_9_batting_percentage_last_10_games\n\n feature_set << feature.home_batting_spot_1_batting_percentage_last_20_games\n feature_set << feature.home_batting_spot_2_batting_percentage_last_20_games\n feature_set << feature.home_batting_spot_3_batting_percentage_last_20_games\n feature_set << feature.home_batting_spot_4_batting_percentage_last_20_games\n feature_set << feature.home_batting_spot_5_batting_percentage_last_20_games\n feature_set << feature.home_batting_spot_6_batting_percentage_last_20_games\n feature_set << feature.home_batting_spot_7_batting_percentage_last_20_games\n feature_set << feature.home_batting_spot_8_batting_percentage_last_20_games\n feature_set << feature.home_batting_spot_9_batting_percentage_last_20_games\n\n feature_set << feature.away_batting_spot_1_batting_percentage_last_20_games\n feature_set << feature.away_batting_spot_2_batting_percentage_last_20_games\n feature_set << feature.away_batting_spot_3_batting_percentage_last_20_games\n feature_set << feature.away_batting_spot_4_batting_percentage_last_20_games\n feature_set << feature.away_batting_spot_5_batting_percentage_last_20_games\n feature_set << feature.away_batting_spot_6_batting_percentage_last_20_games\n feature_set << feature.away_batting_spot_7_batting_percentage_last_20_games\n feature_set << feature.away_batting_spot_8_batting_percentage_last_20_games\n feature_set << feature.away_batting_spot_9_batting_percentage_last_20_games\n\n feature_set << feature.home_batting_spot_1_OPS_last_1_game\n feature_set << feature.home_batting_spot_2_OPS_last_1_game\n feature_set << feature.home_batting_spot_3_OPS_last_1_game\n feature_set << feature.home_batting_spot_4_OPS_last_1_game\n feature_set << feature.home_batting_spot_5_OPS_last_1_game\n feature_set << feature.home_batting_spot_6_OPS_last_1_game\n feature_set << feature.home_batting_spot_7_OPS_last_1_game\n feature_set << feature.home_batting_spot_8_OPS_last_1_game\n feature_set << feature.home_batting_spot_9_OPS_last_1_game\n\n feature_set << feature.away_batting_spot_1_OPS_last_1_game\n feature_set << feature.away_batting_spot_2_OPS_last_1_game\n feature_set << feature.away_batting_spot_3_OPS_last_1_game\n feature_set << feature.away_batting_spot_4_OPS_last_1_game\n feature_set << feature.away_batting_spot_5_OPS_last_1_game\n feature_set << feature.away_batting_spot_6_OPS_last_1_game\n feature_set << feature.away_batting_spot_7_OPS_last_1_game\n feature_set << feature.away_batting_spot_8_OPS_last_1_game\n feature_set << feature.away_batting_spot_9_OPS_last_1_game\n\n feature_set << feature.home_batting_spot_1_OPS_last_2_games\n feature_set << feature.home_batting_spot_2_OPS_last_2_games\n feature_set << feature.home_batting_spot_3_OPS_last_2_games\n feature_set << feature.home_batting_spot_4_OPS_last_2_games\n feature_set << feature.home_batting_spot_5_OPS_last_2_games\n feature_set << feature.home_batting_spot_6_OPS_last_2_games\n feature_set << feature.home_batting_spot_7_OPS_last_2_games\n feature_set << feature.home_batting_spot_8_OPS_last_2_games\n feature_set << feature.home_batting_spot_9_OPS_last_2_games\n\n feature_set << feature.away_batting_spot_1_OPS_last_2_games\n feature_set << feature.away_batting_spot_2_OPS_last_2_games\n feature_set << feature.away_batting_spot_3_OPS_last_2_games\n feature_set << feature.away_batting_spot_4_OPS_last_2_games\n feature_set << feature.away_batting_spot_5_OPS_last_2_games\n feature_set << feature.away_batting_spot_6_OPS_last_2_games\n feature_set << feature.away_batting_spot_7_OPS_last_2_games\n feature_set << feature.away_batting_spot_8_OPS_last_2_games\n feature_set << feature.away_batting_spot_9_OPS_last_2_games\n\n feature_set << feature.home_batting_spot_1_OPS_last_5_games\n feature_set << feature.home_batting_spot_2_OPS_last_5_games\n feature_set << feature.home_batting_spot_3_OPS_last_5_games\n feature_set << feature.home_batting_spot_4_OPS_last_5_games\n feature_set << feature.home_batting_spot_5_OPS_last_5_games\n feature_set << feature.home_batting_spot_6_OPS_last_5_games\n feature_set << feature.home_batting_spot_7_OPS_last_5_games\n feature_set << feature.home_batting_spot_8_OPS_last_5_games\n feature_set << feature.home_batting_spot_9_OPS_last_5_games\n\n feature_set << feature.away_batting_spot_1_OPS_last_5_games\n feature_set << feature.away_batting_spot_2_OPS_last_5_games\n feature_set << feature.away_batting_spot_3_OPS_last_5_games\n feature_set << feature.away_batting_spot_4_OPS_last_5_games\n feature_set << feature.away_batting_spot_5_OPS_last_5_games\n feature_set << feature.away_batting_spot_6_OPS_last_5_games\n feature_set << feature.away_batting_spot_7_OPS_last_5_games\n feature_set << feature.away_batting_spot_8_OPS_last_5_games\n feature_set << feature.away_batting_spot_9_OPS_last_5_games\n\n feature_set << feature.home_batting_spot_1_OPS_last_10_games\n feature_set << feature.home_batting_spot_2_OPS_last_10_games\n feature_set << feature.home_batting_spot_3_OPS_last_10_games\n feature_set << feature.home_batting_spot_4_OPS_last_10_games\n feature_set << feature.home_batting_spot_5_OPS_last_10_games\n feature_set << feature.home_batting_spot_6_OPS_last_10_games\n feature_set << feature.home_batting_spot_7_OPS_last_10_games\n feature_set << feature.home_batting_spot_8_OPS_last_10_games\n feature_set << feature.home_batting_spot_9_OPS_last_10_games\n\n feature_set << feature.away_batting_spot_1_OPS_last_10_games\n feature_set << feature.away_batting_spot_2_OPS_last_10_games\n feature_set << feature.away_batting_spot_3_OPS_last_10_games\n feature_set << feature.away_batting_spot_4_OPS_last_10_games\n feature_set << feature.away_batting_spot_5_OPS_last_10_games\n feature_set << feature.away_batting_spot_6_OPS_last_10_games\n feature_set << feature.away_batting_spot_7_OPS_last_10_games\n feature_set << feature.away_batting_spot_8_OPS_last_10_games\n feature_set << feature.away_batting_spot_9_OPS_last_10_games\n\n feature_set << feature.home_batting_spot_1_OPS_last_20_games\n feature_set << feature.home_batting_spot_2_OPS_last_20_games\n feature_set << feature.home_batting_spot_3_OPS_last_20_games\n feature_set << feature.home_batting_spot_4_OPS_last_20_games\n feature_set << feature.home_batting_spot_5_OPS_last_20_games\n feature_set << feature.home_batting_spot_6_OPS_last_20_games\n feature_set << feature.home_batting_spot_7_OPS_last_20_games\n feature_set << feature.home_batting_spot_8_OPS_last_20_games\n feature_set << feature.home_batting_spot_9_OPS_last_20_games\n\n feature_set << feature.away_batting_spot_1_OPS_last_20_games\n feature_set << feature.away_batting_spot_2_OPS_last_20_games\n feature_set << feature.away_batting_spot_3_OPS_last_20_games\n feature_set << feature.away_batting_spot_4_OPS_last_20_games\n feature_set << feature.away_batting_spot_5_OPS_last_20_games\n feature_set << feature.away_batting_spot_6_OPS_last_20_games\n feature_set << feature.away_batting_spot_7_OPS_last_20_games\n feature_set << feature.away_batting_spot_8_OPS_last_20_games\n feature_set << feature.away_batting_spot_9_OPS_last_20_games\n\n feature_set << feature.home_batting_spot_1_strikeout_rate_last_1_game\n feature_set << feature.home_batting_spot_2_strikeout_rate_last_1_game\n feature_set << feature.home_batting_spot_3_strikeout_rate_last_1_game\n feature_set << feature.home_batting_spot_4_strikeout_rate_last_1_game\n feature_set << feature.home_batting_spot_5_strikeout_rate_last_1_game\n feature_set << feature.home_batting_spot_6_strikeout_rate_last_1_game\n feature_set << feature.home_batting_spot_7_strikeout_rate_last_1_game\n feature_set << feature.home_batting_spot_8_strikeout_rate_last_1_game\n feature_set << feature.home_batting_spot_9_strikeout_rate_last_1_game\n\n feature_set << feature.away_batting_spot_1_strikeout_rate_last_1_game\n feature_set << feature.away_batting_spot_2_strikeout_rate_last_1_game\n feature_set << feature.away_batting_spot_3_strikeout_rate_last_1_game\n feature_set << feature.away_batting_spot_4_strikeout_rate_last_1_game\n feature_set << feature.away_batting_spot_5_strikeout_rate_last_1_game\n feature_set << feature.away_batting_spot_6_strikeout_rate_last_1_game\n feature_set << feature.away_batting_spot_7_strikeout_rate_last_1_game\n feature_set << feature.away_batting_spot_8_strikeout_rate_last_1_game\n feature_set << feature.away_batting_spot_9_strikeout_rate_last_1_game\n\n feature_set << feature.home_batting_spot_1_strikeout_rate_last_2_games\n feature_set << feature.home_batting_spot_2_strikeout_rate_last_2_games\n feature_set << feature.home_batting_spot_3_strikeout_rate_last_2_games\n feature_set << feature.home_batting_spot_4_strikeout_rate_last_2_games\n feature_set << feature.home_batting_spot_5_strikeout_rate_last_2_games\n feature_set << feature.home_batting_spot_6_strikeout_rate_last_2_games\n feature_set << feature.home_batting_spot_7_strikeout_rate_last_2_games\n feature_set << feature.home_batting_spot_8_strikeout_rate_last_2_games\n feature_set << feature.home_batting_spot_9_strikeout_rate_last_2_games\n\n feature_set << feature.away_batting_spot_1_strikeout_rate_last_2_games\n feature_set << feature.away_batting_spot_2_strikeout_rate_last_2_games\n feature_set << feature.away_batting_spot_3_strikeout_rate_last_2_games\n feature_set << feature.away_batting_spot_4_strikeout_rate_last_2_games\n feature_set << feature.away_batting_spot_5_strikeout_rate_last_2_games\n feature_set << feature.away_batting_spot_6_strikeout_rate_last_2_games\n feature_set << feature.away_batting_spot_7_strikeout_rate_last_2_games\n feature_set << feature.away_batting_spot_8_strikeout_rate_last_2_games\n feature_set << feature.away_batting_spot_9_strikeout_rate_last_2_games\n\n feature_set << feature.home_batting_spot_1_strikeout_rate_last_5_games\n feature_set << feature.home_batting_spot_2_strikeout_rate_last_5_games\n feature_set << feature.home_batting_spot_3_strikeout_rate_last_5_games\n feature_set << feature.home_batting_spot_4_strikeout_rate_last_5_games\n feature_set << feature.home_batting_spot_5_strikeout_rate_last_5_games\n feature_set << feature.home_batting_spot_6_strikeout_rate_last_5_games\n feature_set << feature.home_batting_spot_7_strikeout_rate_last_5_games\n feature_set << feature.home_batting_spot_8_strikeout_rate_last_5_games\n feature_set << feature.home_batting_spot_9_strikeout_rate_last_5_games\n\n feature_set << feature.away_batting_spot_1_strikeout_rate_last_5_games\n feature_set << feature.away_batting_spot_2_strikeout_rate_last_5_games\n feature_set << feature.away_batting_spot_3_strikeout_rate_last_5_games\n feature_set << feature.away_batting_spot_4_strikeout_rate_last_5_games\n feature_set << feature.away_batting_spot_5_strikeout_rate_last_5_games\n feature_set << feature.away_batting_spot_6_strikeout_rate_last_5_games\n feature_set << feature.away_batting_spot_7_strikeout_rate_last_5_games\n feature_set << feature.away_batting_spot_8_strikeout_rate_last_5_games\n feature_set << feature.away_batting_spot_9_strikeout_rate_last_5_games\n\n feature_set << feature.home_batting_spot_1_strikeout_rate_last_10_games\n feature_set << feature.home_batting_spot_2_strikeout_rate_last_10_games\n feature_set << feature.home_batting_spot_3_strikeout_rate_last_10_games\n feature_set << feature.home_batting_spot_4_strikeout_rate_last_10_games\n feature_set << feature.home_batting_spot_5_strikeout_rate_last_10_games\n feature_set << feature.home_batting_spot_6_strikeout_rate_last_10_games\n feature_set << feature.home_batting_spot_7_strikeout_rate_last_10_games\n feature_set << feature.home_batting_spot_8_strikeout_rate_last_10_games\n feature_set << feature.home_batting_spot_9_strikeout_rate_last_10_games\n\n feature_set << feature.away_batting_spot_1_strikeout_rate_last_10_games\n feature_set << feature.away_batting_spot_2_strikeout_rate_last_10_games\n feature_set << feature.away_batting_spot_3_strikeout_rate_last_10_games\n feature_set << feature.away_batting_spot_4_strikeout_rate_last_10_games\n feature_set << feature.away_batting_spot_5_strikeout_rate_last_10_games\n feature_set << feature.away_batting_spot_6_strikeout_rate_last_10_games\n feature_set << feature.away_batting_spot_7_strikeout_rate_last_10_games\n feature_set << feature.away_batting_spot_8_strikeout_rate_last_10_games\n feature_set << feature.away_batting_spot_9_strikeout_rate_last_10_games\n\n feature_set << feature.home_batting_spot_1_strikeout_rate_last_20_games\n feature_set << feature.home_batting_spot_2_strikeout_rate_last_20_games\n feature_set << feature.home_batting_spot_3_strikeout_rate_last_20_games\n feature_set << feature.home_batting_spot_4_strikeout_rate_last_20_games\n feature_set << feature.home_batting_spot_5_strikeout_rate_last_20_games\n feature_set << feature.home_batting_spot_6_strikeout_rate_last_20_games\n feature_set << feature.home_batting_spot_7_strikeout_rate_last_20_games\n feature_set << feature.home_batting_spot_8_strikeout_rate_last_20_games\n feature_set << feature.home_batting_spot_9_strikeout_rate_last_20_games\n\n feature_set << feature.away_batting_spot_1_strikeout_rate_last_20_games\n feature_set << feature.away_batting_spot_2_strikeout_rate_last_20_games\n feature_set << feature.away_batting_spot_3_strikeout_rate_last_20_games\n feature_set << feature.away_batting_spot_4_strikeout_rate_last_20_games\n feature_set << feature.away_batting_spot_5_strikeout_rate_last_20_games\n feature_set << feature.away_batting_spot_6_strikeout_rate_last_20_games\n feature_set << feature.away_batting_spot_7_strikeout_rate_last_20_games\n feature_set << feature.away_batting_spot_8_strikeout_rate_last_20_games\n feature_set << feature.away_batting_spot_9_strikeout_rate_last_20_games\n \n feature_set << feature.home_batting_spot_1_walks_per_game_career\n feature_set << feature.home_batting_spot_2_walks_per_game_career\n feature_set << feature.home_batting_spot_3_walks_per_game_career\n feature_set << feature.home_batting_spot_4_walks_per_game_career\n feature_set << feature.home_batting_spot_5_walks_per_game_career\n feature_set << feature.home_batting_spot_6_walks_per_game_career\n feature_set << feature.home_batting_spot_7_walks_per_game_career\n feature_set << feature.home_batting_spot_8_walks_per_game_career\n feature_set << feature.home_batting_spot_9_walks_per_game_career\n \n feature_set << feature.away_batting_spot_1_walks_per_game_career\n feature_set << feature.away_batting_spot_2_walks_per_game_career\n feature_set << feature.away_batting_spot_3_walks_per_game_career\n feature_set << feature.away_batting_spot_4_walks_per_game_career\n feature_set << feature.away_batting_spot_5_walks_per_game_career\n feature_set << feature.away_batting_spot_6_walks_per_game_career\n feature_set << feature.away_batting_spot_7_walks_per_game_career\n feature_set << feature.away_batting_spot_8_walks_per_game_career\n feature_set << feature.away_batting_spot_9_walks_per_game_career\n \n feature_set << feature.home_batting_spot_1_batting_percentage_career\n feature_set << feature.home_batting_spot_2_batting_percentage_career\n feature_set << feature.home_batting_spot_3_batting_percentage_career\n feature_set << feature.home_batting_spot_4_batting_percentage_career\n feature_set << feature.home_batting_spot_5_batting_percentage_career\n feature_set << feature.home_batting_spot_6_batting_percentage_career\n feature_set << feature.home_batting_spot_7_batting_percentage_career\n feature_set << feature.home_batting_spot_8_batting_percentage_career\n feature_set << feature.home_batting_spot_9_batting_percentage_career\n \n feature_set << feature.away_batting_spot_1_batting_percentage_career\n feature_set << feature.away_batting_spot_2_batting_percentage_career\n feature_set << feature.away_batting_spot_3_batting_percentage_career\n feature_set << feature.away_batting_spot_4_batting_percentage_career\n feature_set << feature.away_batting_spot_5_batting_percentage_career\n feature_set << feature.away_batting_spot_6_batting_percentage_career\n feature_set << feature.away_batting_spot_7_batting_percentage_career\n feature_set << feature.away_batting_spot_8_batting_percentage_career\n feature_set << feature.away_batting_spot_9_batting_percentage_career\n \n feature_set << feature.home_batting_spot_1_OPS_career\n feature_set << feature.home_batting_spot_2_OPS_career\n feature_set << feature.home_batting_spot_3_OPS_career\n feature_set << feature.home_batting_spot_4_OPS_career\n feature_set << feature.home_batting_spot_5_OPS_career\n feature_set << feature.home_batting_spot_6_OPS_career\n feature_set << feature.home_batting_spot_7_OPS_career\n feature_set << feature.home_batting_spot_8_OPS_career\n feature_set << feature.home_batting_spot_9_OPS_career\n \n feature_set << feature.away_batting_spot_1_OPS_career\n feature_set << feature.away_batting_spot_2_OPS_career\n feature_set << feature.away_batting_spot_3_OPS_career\n feature_set << feature.away_batting_spot_4_OPS_career\n feature_set << feature.away_batting_spot_5_OPS_career\n feature_set << feature.away_batting_spot_6_OPS_career\n feature_set << feature.away_batting_spot_7_OPS_career\n feature_set << feature.away_batting_spot_8_OPS_career\n feature_set << feature.away_batting_spot_9_OPS_career\n \n feature_set << feature.home_batting_spot_1_strikeout_rate_career\n feature_set << feature.home_batting_spot_2_strikeout_rate_career\n feature_set << feature.home_batting_spot_3_strikeout_rate_career\n feature_set << feature.home_batting_spot_4_strikeout_rate_career\n feature_set << feature.home_batting_spot_5_strikeout_rate_career\n feature_set << feature.home_batting_spot_6_strikeout_rate_career\n feature_set << feature.home_batting_spot_7_strikeout_rate_career\n feature_set << feature.home_batting_spot_8_strikeout_rate_career\n feature_set << feature.home_batting_spot_9_strikeout_rate_career\n \n feature_set << feature.away_batting_spot_1_strikeout_rate_career\n feature_set << feature.away_batting_spot_2_strikeout_rate_career\n feature_set << feature.away_batting_spot_3_strikeout_rate_career\n feature_set << feature.away_batting_spot_4_strikeout_rate_career\n feature_set << feature.away_batting_spot_5_strikeout_rate_career\n feature_set << feature.away_batting_spot_6_strikeout_rate_career\n feature_set << feature.away_batting_spot_7_strikeout_rate_career\n feature_set << feature.away_batting_spot_8_strikeout_rate_career\n feature_set << feature.away_batting_spot_9_strikeout_rate_career \n#=end\n\n=begin\n feature_set << feature.home_batting_spot_1_walks_last_1_game - feature.away_batting_spot_1_walks_last_1_game\n feature_set << feature.home_batting_spot_2_walks_last_1_game - feature.away_batting_spot_2_walks_last_1_game\n feature_set << feature.home_batting_spot_3_walks_last_1_game - feature.away_batting_spot_3_walks_last_1_game\n feature_set << feature.home_batting_spot_4_walks_last_1_game - feature.away_batting_spot_4_walks_last_1_game\n feature_set << feature.home_batting_spot_5_walks_last_1_game - feature.away_batting_spot_5_walks_last_1_game\n feature_set << feature.home_batting_spot_6_walks_last_1_game - feature.away_batting_spot_6_walks_last_1_game\n feature_set << feature.home_batting_spot_7_walks_last_1_game - feature.away_batting_spot_7_walks_last_1_game\n feature_set << feature.home_batting_spot_8_walks_last_1_game - feature.away_batting_spot_8_walks_last_1_game\n feature_set << feature.home_batting_spot_9_walks_last_1_game - feature.away_batting_spot_9_walks_last_1_game\n=end\n\n=begin\n walk_diff = 0\n walk_diff += feature.home_batting_spot_1_walks_last_1_game - feature.away_batting_spot_1_walks_last_1_game\n walk_diff += feature.home_batting_spot_2_walks_last_1_game - feature.away_batting_spot_2_walks_last_1_game\n walk_diff += feature.home_batting_spot_3_walks_last_1_game - feature.away_batting_spot_3_walks_last_1_game\n walk_diff += feature.home_batting_spot_4_walks_last_1_game - feature.away_batting_spot_4_walks_last_1_game\n walk_diff += feature.home_batting_spot_5_walks_last_1_game - feature.away_batting_spot_5_walks_last_1_game\n walk_diff += feature.home_batting_spot_6_walks_last_1_game - feature.away_batting_spot_6_walks_last_1_game\n walk_diff += feature.home_batting_spot_7_walks_last_1_game - feature.away_batting_spot_7_walks_last_1_game\n walk_diff += feature.home_batting_spot_8_walks_last_1_game - feature.away_batting_spot_8_walks_last_1_game\n walk_diff += feature.home_batting_spot_9_walks_last_1_game - feature.away_batting_spot_9_walks_last_1_game\n feature_set << walk_diff\n=end\n \n\n #feature_set << (feature.home_team_won ? 1 : -1)\n\n examples << feature_set\n labels << (feature.home_team_won ? 1 : 0)\n end\nend",
"def full_classification_list\n \n vs = Hash[Vineyard.all.map { |v| [v.id, {:name => v.name, :up_id => v.subregion_id}] }]\n ss = Hash[Subregion.all.map { |s| [s.id, {:name => s.name, :up_id => s.region_id}] }]\n rs = Hash[Region.all.map { |r| [r.id, {:name => r.name, :up_id => r.country_id}] }]\n cs = Hash[Country.all.map { |c| [c.id, {:name => c.name, :up_id => nil}] }]\n\n full_list = []\n \n vs.reject { |k,v| v[:name].nil? || v[:name] == \"\" }.each do |k, v|\n s = ss[v[:up_id]]\n r = rs[s[:up_id]]\n c = cs[r[:up_id]]\n full_list = full_list.append([k, v[:name], v[:up_id], s[:name], s[:up_id], r[:name], r[:up_id], c[:name]])\n end\n \n ss.reject { |k,s| s[:name].nil? || s[:name] == \"\" }.each do |k, s|\n r = rs[s[:up_id]]\n c = cs[r[:up_id]]\n full_list = full_list.append([nil, nil, k, s[:name], s[:up_id], r[:name], r[:up_id], c[:name]])\n end\n \n rs.reject { |k,r| r[:name].nil? || r[:name] == \"\" }.each do |k, r|\n c = cs[r[:up_id]]\n full_list = full_list.append([nil, nil, nil, nil, k, r[:name], r[:up_id], c[:name]])\n end\n \n cs.reject { |k,c| c[:name].nil? || c[:name] == \"\" }.each do |k, c|\n full_list = full_list.append([nil, nil, nil, nil, nil, nil, k, c[:name]])\n end\n\n return full_list\n end",
"def features\n features = Hash[api.get_features(app).body.map{|feature| [feature['name'], feature['enabled']]}]\n actual_features = Hash[api.get_features(target_app).body.map{|feature| [feature['name'], feature['enabled']]}]\n\n features_to_enable = features.select{|feature, enabled| enabled && !actual_features[feature]}\n features_to_disable = actual_features.select{|feature, enabled| enabled && !features[feature]}\n\n action(\"Copying labs features from #{app} and restarting #{target_app}\") do\n features_to_enable.each do |feature|\n puts \"Adding #{feature} to #{target_app}\"\n api.post_feature(feature, target_app)\n end\n\n features_to_disable.each do |feature|\n puts \"Deleting #{feature} from #{target_app}\"\n api.delete_feature(feature, target_app)\n end\n end\n end",
"def find_all_scenarios_by_feature(project_name, feature_name)\n if @scenario_look_up_map.has_key? project_name\n features = @scenario_look_up_map[project_name]\n if features.has_key? feature_name\n features[feature_name].dup\n end\n end\n end",
"def features\n @features ||= {}\n @features.keys\n end",
"def without_any(*features)\n features = setify(*features)\n self.class.new Hash[@sets.select {|key, val| key.intersection(features).empty?}]\n end",
"def grouped_by_access_feature!\n groups = {'audio_described_performance' => [], 'captioned_performance' => [], 'signed_performance' => [], 'touch_tour' => [], 'relaxed_performance' => [], 'talk_back' => []}\n\n @instances.each do |instance|\n instance_types = instance.meta_attributes.select{|attr_key, attr_value| groups.keys.include?(attr_key) && attr_value==\"true\"}.keys\n\n if instance_types.any?\n instance_types.each do |type|\n groups[type].push(instance)\n end\n end\n end\n\n groups.each do |type, instances|\n groups[type] = slice_instances_by_date(instances.reverse)\n end\n\n groups\n end",
"def feature_objects(*args, &block)\n maesb_sets.compact + (super(*args, &block))\n end",
"def combine\n other_survivor_supplies = [ \"warm clothes\", \"rations\", \"compass\", \"camp stove\",\n \"solar battery\", \"flashlight\"]\n other_survivor_supplies.each do |others|\n @zombie_apocalypse_supplies.push(others)\n end\n @zombie_apocalypse_supplies = @zombie_apocalypse_supplies.uniq\n\nend",
"def update!(**args)\n @anchors_common_feature_set = args[:anchors_common_feature_set] if args.key?(:anchors_common_feature_set)\n @caption_entity_anchor_set_features = args[:caption_entity_anchor_set_features] if args.key?(:caption_entity_anchor_set_features)\n @caption_span_anchor_set_features = args[:caption_span_anchor_set_features] if args.key?(:caption_span_anchor_set_features)\n @comment_anchor_set_features = args[:comment_anchor_set_features] if args.key?(:comment_anchor_set_features)\n @description_anchor_set_features = args[:description_anchor_set_features] if args.key?(:description_anchor_set_features)\n @filtered = args[:filtered] if args.key?(:filtered)\n @list_anchor_set_features = args[:list_anchor_set_features] if args.key?(:list_anchor_set_features)\n @list_training_data_set_features = args[:list_training_data_set_features] if args.key?(:list_training_data_set_features)\n @ocr_anchor_cluster_feature = args[:ocr_anchor_cluster_feature] if args.key?(:ocr_anchor_cluster_feature)\n @ocr_description_training_data_set_features = args[:ocr_description_training_data_set_features] if args.key?(:ocr_description_training_data_set_features)\n @qna_anchor_set_features = args[:qna_anchor_set_features] if args.key?(:qna_anchor_set_features)\n @rating_score = args[:rating_score] if args.key?(:rating_score)\n @sports_key_moments_anchor_set_features = args[:sports_key_moments_anchor_set_features] if args.key?(:sports_key_moments_anchor_set_features)\n end",
"def active_features(options={})\n options = Util.normalized_options options, controller\n ret={}\n features.each {|feature_name,feature| ret[feature_name]=feature if active_internal(feature_name,options)}\n ret\n end",
"def merge_components(c1, c2)\n\nend",
"def find_features\n @features = if params[ :only ]\n params[ :only ].to_s.split( \",\" ).collect( &:to_sym ).uniq\n\n elsif skip = params[ :skip ] || params[ :exclude ]\n monitored_features.keys - skip.to_s.split( \",\" ).collect( &:to_sym )\n\n else\n monitored_features.keys\n end\n end",
"def all_features\r\n feature_objects.inject([]) {|r, obj| r + obj.features }\r\n end",
"def _init_global_features()\n @_global_features = {}\n end",
"def with(*features)\n pos, neg = mangle_args(*features)\n self.class.new(Hash[@sets.select do |key, val|\n !key.intersection(pos).empty?\n end]).without_any(neg)\n end",
"def addFeaturesAndLabel(team, earliest_date, latest_date, examples, labels)\n home_faceoffs = Game.where(\"home_team = ? and game_date > ? and game_date <= ?\", team, earliest_date, latest_date).order(\"game_date desc\")\n away_faceoffs = Game.where(\"away_team = ? and game_date > ? and game_date <= ?\", team, earliest_date, latest_date).order(\"game_date desc\")\n past_games = home_faceoffs.concat(away_faceoffs)\n # games sort in ascending order. The first game of the 2001 season is located first\n past_games = past_games.sort {|game1, game2| game1.game_date <=> game2.game_date }\n \n player_performances = Hash.new\n\n past_games.each_with_index do |past_game, index|\n feature_set = Feature.find_by_game_id(past_game.id)\n performances = Performance.where(\"game_id = ?\", past_game.id)\n\n performances.each do |perf|\n player = Player.find_by_id(perf.player_id)\n if !player_performances.has_key?(player.retrosheet_id)\n addBlankPlayer(player.retrosheet_id, player_performances)\n end\n end\n\n if past_game.home_team == team.to_s\n feature_set.home_batting_spot_1_walks_last_1_game = player_performances[past_game.home_batting_spot_1][\"walks_last_1_game\"].reduce(:+)\n feature_set.home_batting_spot_2_walks_last_1_game = player_performances[past_game.home_batting_spot_2][\"walks_last_1_game\"].reduce(:+)\n feature_set.home_batting_spot_3_walks_last_1_game = player_performances[past_game.home_batting_spot_3][\"walks_last_1_game\"].reduce(:+)\n feature_set.home_batting_spot_4_walks_last_1_game = player_performances[past_game.home_batting_spot_4][\"walks_last_1_game\"].reduce(:+)\n feature_set.home_batting_spot_5_walks_last_1_game = player_performances[past_game.home_batting_spot_5][\"walks_last_1_game\"].reduce(:+)\n feature_set.home_batting_spot_6_walks_last_1_game = player_performances[past_game.home_batting_spot_6][\"walks_last_1_game\"].reduce(:+)\n feature_set.home_batting_spot_7_walks_last_1_game = player_performances[past_game.home_batting_spot_7][\"walks_last_1_game\"].reduce(:+)\n feature_set.home_batting_spot_8_walks_last_1_game = player_performances[past_game.home_batting_spot_8][\"walks_last_1_game\"].reduce(:+)\n feature_set.home_batting_spot_9_walks_last_1_game = player_performances[past_game.home_batting_spot_9][\"walks_last_1_game\"].reduce(:+)\n \n feature_set.home_batting_spot_1_walks_last_2_games = player_performances[past_game.home_batting_spot_1][\"walks_last_2_games\"].reduce(:+)\n feature_set.home_batting_spot_2_walks_last_2_games = player_performances[past_game.home_batting_spot_2][\"walks_last_2_games\"].reduce(:+)\n feature_set.home_batting_spot_3_walks_last_2_games = player_performances[past_game.home_batting_spot_3][\"walks_last_2_games\"].reduce(:+)\n feature_set.home_batting_spot_4_walks_last_2_games = player_performances[past_game.home_batting_spot_4][\"walks_last_2_games\"].reduce(:+)\n feature_set.home_batting_spot_5_walks_last_2_games = player_performances[past_game.home_batting_spot_5][\"walks_last_2_games\"].reduce(:+)\n feature_set.home_batting_spot_6_walks_last_2_games = player_performances[past_game.home_batting_spot_6][\"walks_last_2_games\"].reduce(:+)\n feature_set.home_batting_spot_7_walks_last_2_games = player_performances[past_game.home_batting_spot_7][\"walks_last_2_games\"].reduce(:+)\n feature_set.home_batting_spot_8_walks_last_2_games = player_performances[past_game.home_batting_spot_8][\"walks_last_2_games\"].reduce(:+)\n feature_set.home_batting_spot_9_walks_last_2_games = player_performances[past_game.home_batting_spot_9][\"walks_last_2_games\"].reduce(:+)\n\n feature_set.home_batting_spot_1_walks_last_5_games = player_performances[past_game.home_batting_spot_1][\"walks_last_5_games\"].reduce(:+)\n feature_set.home_batting_spot_2_walks_last_5_games = player_performances[past_game.home_batting_spot_2][\"walks_last_5_games\"].reduce(:+)\n feature_set.home_batting_spot_3_walks_last_5_games = player_performances[past_game.home_batting_spot_3][\"walks_last_5_games\"].reduce(:+)\n feature_set.home_batting_spot_4_walks_last_5_games = player_performances[past_game.home_batting_spot_4][\"walks_last_5_games\"].reduce(:+)\n feature_set.home_batting_spot_5_walks_last_5_games = player_performances[past_game.home_batting_spot_5][\"walks_last_5_games\"].reduce(:+)\n feature_set.home_batting_spot_6_walks_last_5_games = player_performances[past_game.home_batting_spot_6][\"walks_last_5_games\"].reduce(:+)\n feature_set.home_batting_spot_7_walks_last_5_games = player_performances[past_game.home_batting_spot_7][\"walks_last_5_games\"].reduce(:+)\n feature_set.home_batting_spot_8_walks_last_5_games = player_performances[past_game.home_batting_spot_8][\"walks_last_5_games\"].reduce(:+)\n feature_set.home_batting_spot_9_walks_last_5_games = player_performances[past_game.home_batting_spot_9][\"walks_last_5_games\"].reduce(:+)\n\n feature_set.home_batting_spot_1_walks_last_10_games = player_performances[past_game.home_batting_spot_1][\"walks_last_10_games\"].reduce(:+)\n feature_set.home_batting_spot_2_walks_last_10_games = player_performances[past_game.home_batting_spot_2][\"walks_last_10_games\"].reduce(:+)\n feature_set.home_batting_spot_3_walks_last_10_games = player_performances[past_game.home_batting_spot_3][\"walks_last_10_games\"].reduce(:+)\n feature_set.home_batting_spot_4_walks_last_10_games = player_performances[past_game.home_batting_spot_4][\"walks_last_10_games\"].reduce(:+)\n feature_set.home_batting_spot_5_walks_last_10_games = player_performances[past_game.home_batting_spot_5][\"walks_last_10_games\"].reduce(:+)\n feature_set.home_batting_spot_6_walks_last_10_games = player_performances[past_game.home_batting_spot_6][\"walks_last_10_games\"].reduce(:+)\n feature_set.home_batting_spot_7_walks_last_10_games = player_performances[past_game.home_batting_spot_7][\"walks_last_10_games\"].reduce(:+)\n feature_set.home_batting_spot_8_walks_last_10_games = player_performances[past_game.home_batting_spot_8][\"walks_last_10_games\"].reduce(:+)\n feature_set.home_batting_spot_9_walks_last_10_games = player_performances[past_game.home_batting_spot_9][\"walks_last_10_games\"].reduce(:+)\n\n feature_set.home_batting_spot_1_walks_last_20_games = player_performances[past_game.home_batting_spot_1][\"walks_last_20_games\"].reduce(:+)\n feature_set.home_batting_spot_2_walks_last_20_games = player_performances[past_game.home_batting_spot_2][\"walks_last_20_games\"].reduce(:+)\n feature_set.home_batting_spot_3_walks_last_20_games = player_performances[past_game.home_batting_spot_3][\"walks_last_20_games\"].reduce(:+)\n feature_set.home_batting_spot_4_walks_last_20_games = player_performances[past_game.home_batting_spot_4][\"walks_last_20_games\"].reduce(:+)\n feature_set.home_batting_spot_5_walks_last_20_games = player_performances[past_game.home_batting_spot_5][\"walks_last_20_games\"].reduce(:+)\n feature_set.home_batting_spot_6_walks_last_20_games = player_performances[past_game.home_batting_spot_6][\"walks_last_20_games\"].reduce(:+)\n feature_set.home_batting_spot_7_walks_last_20_games = player_performances[past_game.home_batting_spot_7][\"walks_last_20_games\"].reduce(:+)\n feature_set.home_batting_spot_8_walks_last_20_games = player_performances[past_game.home_batting_spot_8][\"walks_last_20_games\"].reduce(:+)\n feature_set.home_batting_spot_9_walks_last_20_games = player_performances[past_game.home_batting_spot_9][\"walks_last_20_games\"].reduce(:+)\n\n \n feature_set.home_batting_spot_1_walks_per_game_career = player_performances[past_game.home_batting_spot_1][\"career_games\"] == 0 ? 0 : player_performances[past_game.home_batting_spot_1][\"career_walks\"] / player_performances[past_game.home_batting_spot_1][\"career_games\"]\n feature_set.home_batting_spot_2_walks_per_game_career = player_performances[past_game.home_batting_spot_2][\"career_games\"] == 0 ? 0 : player_performances[past_game.home_batting_spot_2][\"career_walks\"] / player_performances[past_game.home_batting_spot_2][\"career_games\"]\n feature_set.home_batting_spot_3_walks_per_game_career = player_performances[past_game.home_batting_spot_3][\"career_games\"] == 0 ? 0 : player_performances[past_game.home_batting_spot_3][\"career_walks\"] / player_performances[past_game.home_batting_spot_3][\"career_games\"]\n feature_set.home_batting_spot_4_walks_per_game_career = player_performances[past_game.home_batting_spot_4][\"career_games\"] == 0 ? 0 : player_performances[past_game.home_batting_spot_4][\"career_walks\"] / player_performances[past_game.home_batting_spot_4][\"career_games\"]\n feature_set.home_batting_spot_5_walks_per_game_career = player_performances[past_game.home_batting_spot_5][\"career_games\"] == 0 ? 0 : player_performances[past_game.home_batting_spot_5][\"career_walks\"] / player_performances[past_game.home_batting_spot_5][\"career_games\"]\n feature_set.home_batting_spot_6_walks_per_game_career = player_performances[past_game.home_batting_spot_6][\"career_games\"] == 0 ? 0 : player_performances[past_game.home_batting_spot_6][\"career_walks\"] / player_performances[past_game.home_batting_spot_6][\"career_games\"]\n feature_set.home_batting_spot_7_walks_per_game_career = player_performances[past_game.home_batting_spot_7][\"career_games\"] == 0 ? 0 : player_performances[past_game.home_batting_spot_7][\"career_walks\"] / player_performances[past_game.home_batting_spot_7][\"career_games\"]\n feature_set.home_batting_spot_8_walks_per_game_career = player_performances[past_game.home_batting_spot_8][\"career_games\"] == 0 ? 0 : player_performances[past_game.home_batting_spot_8][\"career_walks\"] / player_performances[past_game.home_batting_spot_8][\"career_games\"]\n feature_set.home_batting_spot_9_walks_per_game_career = player_performances[past_game.home_batting_spot_9][\"career_games\"] == 0 ? 0 : player_performances[past_game.home_batting_spot_9][\"career_walks\"] / player_performances[past_game.home_batting_spot_9][\"career_games\"]\n \n\n feature_set.home_batting_spot_1_batting_percentage_last_1_game = player_performances[past_game.home_batting_spot_1][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_1][\"hits_last_1_game\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_1][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.home_batting_spot_2_batting_percentage_last_1_game = player_performances[past_game.home_batting_spot_2][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_2][\"hits_last_1_game\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_2][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.home_batting_spot_3_batting_percentage_last_1_game = player_performances[past_game.home_batting_spot_3][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_3][\"hits_last_1_game\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_3][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.home_batting_spot_4_batting_percentage_last_1_game = player_performances[past_game.home_batting_spot_4][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_4][\"hits_last_1_game\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_4][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.home_batting_spot_5_batting_percentage_last_1_game = player_performances[past_game.home_batting_spot_5][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_5][\"hits_last_1_game\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_5][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.home_batting_spot_6_batting_percentage_last_1_game = player_performances[past_game.home_batting_spot_6][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_6][\"hits_last_1_game\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_6][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.home_batting_spot_7_batting_percentage_last_1_game = player_performances[past_game.home_batting_spot_7][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_7][\"hits_last_1_game\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_7][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.home_batting_spot_8_batting_percentage_last_1_game = player_performances[past_game.home_batting_spot_8][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_8][\"hits_last_1_game\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_8][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.home_batting_spot_9_batting_percentage_last_1_game = player_performances[past_game.home_batting_spot_9][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_9][\"hits_last_1_game\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_9][\"at_bats_last_1_game\"].reduce(:+)\n\n feature_set.home_batting_spot_1_batting_percentage_last_2_games = player_performances[past_game.home_batting_spot_1][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_1][\"hits_last_2_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_1][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.home_batting_spot_2_batting_percentage_last_2_games = player_performances[past_game.home_batting_spot_2][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_2][\"hits_last_2_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_2][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.home_batting_spot_3_batting_percentage_last_2_games = player_performances[past_game.home_batting_spot_3][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_3][\"hits_last_2_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_3][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.home_batting_spot_4_batting_percentage_last_2_games = player_performances[past_game.home_batting_spot_4][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_4][\"hits_last_2_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_4][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.home_batting_spot_5_batting_percentage_last_2_games = player_performances[past_game.home_batting_spot_5][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_5][\"hits_last_2_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_5][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.home_batting_spot_6_batting_percentage_last_2_games = player_performances[past_game.home_batting_spot_6][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_6][\"hits_last_2_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_6][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.home_batting_spot_7_batting_percentage_last_2_games = player_performances[past_game.home_batting_spot_7][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_7][\"hits_last_2_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_7][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.home_batting_spot_8_batting_percentage_last_2_games = player_performances[past_game.home_batting_spot_8][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_8][\"hits_last_2_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_8][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.home_batting_spot_9_batting_percentage_last_2_games = player_performances[past_game.home_batting_spot_9][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_9][\"hits_last_2_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_9][\"at_bats_last_2_games\"].reduce(:+)\n\n feature_set.home_batting_spot_1_batting_percentage_last_5_games = player_performances[past_game.home_batting_spot_1][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_1][\"hits_last_5_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_1][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.home_batting_spot_2_batting_percentage_last_5_games = player_performances[past_game.home_batting_spot_2][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_2][\"hits_last_5_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_2][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.home_batting_spot_3_batting_percentage_last_5_games = player_performances[past_game.home_batting_spot_3][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_3][\"hits_last_5_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_3][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.home_batting_spot_4_batting_percentage_last_5_games = player_performances[past_game.home_batting_spot_4][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_4][\"hits_last_5_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_4][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.home_batting_spot_5_batting_percentage_last_5_games = player_performances[past_game.home_batting_spot_5][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_5][\"hits_last_5_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_5][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.home_batting_spot_6_batting_percentage_last_5_games = player_performances[past_game.home_batting_spot_6][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_6][\"hits_last_5_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_6][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.home_batting_spot_7_batting_percentage_last_5_games = player_performances[past_game.home_batting_spot_7][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_7][\"hits_last_5_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_7][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.home_batting_spot_8_batting_percentage_last_5_games = player_performances[past_game.home_batting_spot_8][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_8][\"hits_last_5_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_8][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.home_batting_spot_9_batting_percentage_last_5_games = player_performances[past_game.home_batting_spot_9][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_9][\"hits_last_5_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_9][\"at_bats_last_5_games\"].reduce(:+)\n\n feature_set.home_batting_spot_1_batting_percentage_last_10_games = player_performances[past_game.home_batting_spot_1][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_1][\"hits_last_10_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_1][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.home_batting_spot_2_batting_percentage_last_10_games = player_performances[past_game.home_batting_spot_2][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_2][\"hits_last_10_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_2][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.home_batting_spot_3_batting_percentage_last_10_games = player_performances[past_game.home_batting_spot_3][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_3][\"hits_last_10_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_3][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.home_batting_spot_4_batting_percentage_last_10_games = player_performances[past_game.home_batting_spot_4][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_4][\"hits_last_10_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_4][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.home_batting_spot_5_batting_percentage_last_10_games = player_performances[past_game.home_batting_spot_5][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_5][\"hits_last_10_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_5][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.home_batting_spot_6_batting_percentage_last_10_games = player_performances[past_game.home_batting_spot_6][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_6][\"hits_last_10_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_6][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.home_batting_spot_7_batting_percentage_last_10_games = player_performances[past_game.home_batting_spot_7][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_7][\"hits_last_10_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_7][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.home_batting_spot_8_batting_percentage_last_10_games = player_performances[past_game.home_batting_spot_8][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_8][\"hits_last_10_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_8][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.home_batting_spot_9_batting_percentage_last_10_games = player_performances[past_game.home_batting_spot_9][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_9][\"hits_last_10_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_9][\"at_bats_last_10_games\"].reduce(:+)\n\n feature_set.home_batting_spot_1_batting_percentage_last_20_games = player_performances[past_game.home_batting_spot_1][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_1][\"hits_last_20_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_1][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.home_batting_spot_2_batting_percentage_last_20_games = player_performances[past_game.home_batting_spot_2][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_2][\"hits_last_20_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_2][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.home_batting_spot_3_batting_percentage_last_20_games = player_performances[past_game.home_batting_spot_3][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_3][\"hits_last_20_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_3][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.home_batting_spot_4_batting_percentage_last_20_games = player_performances[past_game.home_batting_spot_4][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_4][\"hits_last_20_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_4][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.home_batting_spot_5_batting_percentage_last_20_games = player_performances[past_game.home_batting_spot_5][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_5][\"hits_last_20_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_5][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.home_batting_spot_6_batting_percentage_last_20_games = player_performances[past_game.home_batting_spot_6][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_6][\"hits_last_20_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_6][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.home_batting_spot_7_batting_percentage_last_20_games = player_performances[past_game.home_batting_spot_7][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_7][\"hits_last_20_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_7][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.home_batting_spot_8_batting_percentage_last_20_games = player_performances[past_game.home_batting_spot_8][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_8][\"hits_last_20_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_8][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.home_batting_spot_9_batting_percentage_last_20_games = player_performances[past_game.home_batting_spot_9][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_9][\"hits_last_20_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_9][\"at_bats_last_20_games\"].reduce(:+)\n\n feature_set.home_batting_spot_1_batting_percentage_career = player_performances[past_game.home_batting_spot_1][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.home_batting_spot_1][\"career_hits\"] / player_performances[past_game.home_batting_spot_1][\"career_at_bats\"]\n feature_set.home_batting_spot_2_batting_percentage_career = player_performances[past_game.home_batting_spot_2][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.home_batting_spot_2][\"career_hits\"] / player_performances[past_game.home_batting_spot_2][\"career_at_bats\"]\n feature_set.home_batting_spot_3_batting_percentage_career = player_performances[past_game.home_batting_spot_3][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.home_batting_spot_3][\"career_hits\"] / player_performances[past_game.home_batting_spot_3][\"career_at_bats\"]\n feature_set.home_batting_spot_4_batting_percentage_career = player_performances[past_game.home_batting_spot_4][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.home_batting_spot_4][\"career_hits\"] / player_performances[past_game.home_batting_spot_4][\"career_at_bats\"]\n feature_set.home_batting_spot_5_batting_percentage_career = player_performances[past_game.home_batting_spot_5][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.home_batting_spot_5][\"career_hits\"] / player_performances[past_game.home_batting_spot_5][\"career_at_bats\"]\n feature_set.home_batting_spot_6_batting_percentage_career = player_performances[past_game.home_batting_spot_6][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.home_batting_spot_6][\"career_hits\"] / player_performances[past_game.home_batting_spot_6][\"career_at_bats\"]\n feature_set.home_batting_spot_7_batting_percentage_career = player_performances[past_game.home_batting_spot_7][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.home_batting_spot_7][\"career_hits\"] / player_performances[past_game.home_batting_spot_7][\"career_at_bats\"]\n feature_set.home_batting_spot_8_batting_percentage_career = player_performances[past_game.home_batting_spot_8][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.home_batting_spot_8][\"career_hits\"] / player_performances[past_game.home_batting_spot_8][\"career_at_bats\"]\n feature_set.home_batting_spot_9_batting_percentage_career = player_performances[past_game.home_batting_spot_9][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.home_batting_spot_9][\"career_hits\"] / player_performances[past_game.home_batting_spot_9][\"career_at_bats\"]\n\n feature_set.home_batting_spot_1_OPS_last_1_game = player_performances[past_game.home_batting_spot_1][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_1][\"total_bases_last_1_game\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_1][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.home_batting_spot_2_OPS_last_1_game = player_performances[past_game.home_batting_spot_2][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_2][\"total_bases_last_1_game\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_2][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.home_batting_spot_3_OPS_last_1_game = player_performances[past_game.home_batting_spot_3][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_3][\"total_bases_last_1_game\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_3][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.home_batting_spot_4_OPS_last_1_game = player_performances[past_game.home_batting_spot_4][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_4][\"total_bases_last_1_game\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_4][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.home_batting_spot_5_OPS_last_1_game = player_performances[past_game.home_batting_spot_5][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_5][\"total_bases_last_1_game\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_5][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.home_batting_spot_6_OPS_last_1_game = player_performances[past_game.home_batting_spot_6][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_6][\"total_bases_last_1_game\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_6][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.home_batting_spot_7_OPS_last_1_game = player_performances[past_game.home_batting_spot_7][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_7][\"total_bases_last_1_game\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_7][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.home_batting_spot_8_OPS_last_1_game = player_performances[past_game.home_batting_spot_8][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_8][\"total_bases_last_1_game\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_8][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.home_batting_spot_9_OPS_last_1_game = player_performances[past_game.home_batting_spot_9][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_9][\"total_bases_last_1_game\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_9][\"at_bats_last_1_game\"].reduce(:+)\n\n feature_set.home_batting_spot_1_OPS_last_2_games = player_performances[past_game.home_batting_spot_1][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_1][\"total_bases_last_2_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_1][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.home_batting_spot_2_OPS_last_2_games = player_performances[past_game.home_batting_spot_2][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_2][\"total_bases_last_2_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_2][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.home_batting_spot_3_OPS_last_2_games = player_performances[past_game.home_batting_spot_3][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_3][\"total_bases_last_2_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_3][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.home_batting_spot_4_OPS_last_2_games = player_performances[past_game.home_batting_spot_4][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_4][\"total_bases_last_2_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_4][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.home_batting_spot_5_OPS_last_2_games = player_performances[past_game.home_batting_spot_5][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_5][\"total_bases_last_2_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_5][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.home_batting_spot_6_OPS_last_2_games = player_performances[past_game.home_batting_spot_6][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_6][\"total_bases_last_2_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_6][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.home_batting_spot_7_OPS_last_2_games = player_performances[past_game.home_batting_spot_7][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_7][\"total_bases_last_2_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_7][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.home_batting_spot_8_OPS_last_2_games = player_performances[past_game.home_batting_spot_8][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_8][\"total_bases_last_2_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_8][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.home_batting_spot_9_OPS_last_2_games = player_performances[past_game.home_batting_spot_9][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_9][\"total_bases_last_2_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_9][\"at_bats_last_2_games\"].reduce(:+)\n\n feature_set.home_batting_spot_1_OPS_last_5_games = player_performances[past_game.home_batting_spot_1][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_1][\"total_bases_last_5_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_1][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.home_batting_spot_2_OPS_last_5_games = player_performances[past_game.home_batting_spot_2][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_2][\"total_bases_last_5_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_2][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.home_batting_spot_3_OPS_last_5_games = player_performances[past_game.home_batting_spot_3][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_3][\"total_bases_last_5_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_3][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.home_batting_spot_4_OPS_last_5_games = player_performances[past_game.home_batting_spot_4][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_4][\"total_bases_last_5_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_4][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.home_batting_spot_5_OPS_last_5_games = player_performances[past_game.home_batting_spot_5][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_5][\"total_bases_last_5_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_5][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.home_batting_spot_6_OPS_last_5_games = player_performances[past_game.home_batting_spot_6][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_6][\"total_bases_last_5_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_6][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.home_batting_spot_7_OPS_last_5_games = player_performances[past_game.home_batting_spot_7][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_7][\"total_bases_last_5_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_7][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.home_batting_spot_8_OPS_last_5_games = player_performances[past_game.home_batting_spot_8][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_8][\"total_bases_last_5_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_8][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.home_batting_spot_9_OPS_last_5_games = player_performances[past_game.home_batting_spot_9][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_9][\"total_bases_last_5_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_9][\"at_bats_last_5_games\"].reduce(:+)\n\n feature_set.home_batting_spot_1_OPS_last_10_games = player_performances[past_game.home_batting_spot_1][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_1][\"total_bases_last_10_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_1][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.home_batting_spot_2_OPS_last_10_games = player_performances[past_game.home_batting_spot_2][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_2][\"total_bases_last_10_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_2][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.home_batting_spot_3_OPS_last_10_games = player_performances[past_game.home_batting_spot_3][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_3][\"total_bases_last_10_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_3][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.home_batting_spot_4_OPS_last_10_games = player_performances[past_game.home_batting_spot_4][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_4][\"total_bases_last_10_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_4][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.home_batting_spot_5_OPS_last_10_games = player_performances[past_game.home_batting_spot_5][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_5][\"total_bases_last_10_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_5][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.home_batting_spot_6_OPS_last_10_games = player_performances[past_game.home_batting_spot_6][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_6][\"total_bases_last_10_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_6][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.home_batting_spot_7_OPS_last_10_games = player_performances[past_game.home_batting_spot_7][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_7][\"total_bases_last_10_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_7][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.home_batting_spot_8_OPS_last_10_games = player_performances[past_game.home_batting_spot_8][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_8][\"total_bases_last_10_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_8][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.home_batting_spot_9_OPS_last_10_games = player_performances[past_game.home_batting_spot_9][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_9][\"total_bases_last_10_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_9][\"at_bats_last_10_games\"].reduce(:+)\n\n feature_set.home_batting_spot_1_OPS_last_20_games = player_performances[past_game.home_batting_spot_1][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_1][\"total_bases_last_20_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_1][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.home_batting_spot_2_OPS_last_20_games = player_performances[past_game.home_batting_spot_2][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_2][\"total_bases_last_20_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_2][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.home_batting_spot_3_OPS_last_20_games = player_performances[past_game.home_batting_spot_3][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_3][\"total_bases_last_20_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_3][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.home_batting_spot_4_OPS_last_20_games = player_performances[past_game.home_batting_spot_4][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_4][\"total_bases_last_20_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_4][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.home_batting_spot_5_OPS_last_20_games = player_performances[past_game.home_batting_spot_5][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_5][\"total_bases_last_20_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_5][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.home_batting_spot_6_OPS_last_20_games = player_performances[past_game.home_batting_spot_6][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_6][\"total_bases_last_20_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_6][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.home_batting_spot_7_OPS_last_20_games = player_performances[past_game.home_batting_spot_7][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_7][\"total_bases_last_20_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_7][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.home_batting_spot_8_OPS_last_20_games = player_performances[past_game.home_batting_spot_8][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_8][\"total_bases_last_20_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_8][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.home_batting_spot_9_OPS_last_20_games = player_performances[past_game.home_batting_spot_9][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_9][\"total_bases_last_20_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_9][\"at_bats_last_20_games\"].reduce(:+)\n\n feature_set.home_batting_spot_1_OPS_career = player_performances[past_game.home_batting_spot_1][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.home_batting_spot_1][\"career_total_bases\"] / player_performances[past_game.home_batting_spot_1][\"career_at_bats\"]\n feature_set.home_batting_spot_2_OPS_career = player_performances[past_game.home_batting_spot_2][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.home_batting_spot_2][\"career_total_bases\"] / player_performances[past_game.home_batting_spot_2][\"career_at_bats\"]\n feature_set.home_batting_spot_3_OPS_career = player_performances[past_game.home_batting_spot_3][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.home_batting_spot_3][\"career_total_bases\"] / player_performances[past_game.home_batting_spot_3][\"career_at_bats\"]\n feature_set.home_batting_spot_4_OPS_career = player_performances[past_game.home_batting_spot_4][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.home_batting_spot_4][\"career_total_bases\"] / player_performances[past_game.home_batting_spot_4][\"career_at_bats\"]\n feature_set.home_batting_spot_5_OPS_career = player_performances[past_game.home_batting_spot_5][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.home_batting_spot_5][\"career_total_bases\"] / player_performances[past_game.home_batting_spot_5][\"career_at_bats\"]\n feature_set.home_batting_spot_6_OPS_career = player_performances[past_game.home_batting_spot_6][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.home_batting_spot_6][\"career_total_bases\"] / player_performances[past_game.home_batting_spot_6][\"career_at_bats\"]\n feature_set.home_batting_spot_7_OPS_career = player_performances[past_game.home_batting_spot_7][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.home_batting_spot_7][\"career_total_bases\"] / player_performances[past_game.home_batting_spot_7][\"career_at_bats\"]\n feature_set.home_batting_spot_8_OPS_career = player_performances[past_game.home_batting_spot_8][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.home_batting_spot_8][\"career_total_bases\"] / player_performances[past_game.home_batting_spot_8][\"career_at_bats\"]\n feature_set.home_batting_spot_9_OPS_career = player_performances[past_game.home_batting_spot_9][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.home_batting_spot_9][\"career_total_bases\"] / player_performances[past_game.home_batting_spot_9][\"career_at_bats\"]\n\n feature_set.home_batting_spot_1_strikeout_rate_last_1_game = player_performances[past_game.home_batting_spot_1][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_1][\"strikeouts_last_1_game\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_1][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.home_batting_spot_2_strikeout_rate_last_1_game = player_performances[past_game.home_batting_spot_2][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_2][\"strikeouts_last_1_game\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_2][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.home_batting_spot_3_strikeout_rate_last_1_game = player_performances[past_game.home_batting_spot_3][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_3][\"strikeouts_last_1_game\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_3][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.home_batting_spot_4_strikeout_rate_last_1_game = player_performances[past_game.home_batting_spot_4][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_4][\"strikeouts_last_1_game\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_4][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.home_batting_spot_5_strikeout_rate_last_1_game = player_performances[past_game.home_batting_spot_5][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_5][\"strikeouts_last_1_game\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_5][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.home_batting_spot_6_strikeout_rate_last_1_game = player_performances[past_game.home_batting_spot_6][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_6][\"strikeouts_last_1_game\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_6][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.home_batting_spot_7_strikeout_rate_last_1_game = player_performances[past_game.home_batting_spot_7][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_7][\"strikeouts_last_1_game\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_7][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.home_batting_spot_8_strikeout_rate_last_1_game = player_performances[past_game.home_batting_spot_8][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_8][\"strikeouts_last_1_game\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_8][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.home_batting_spot_9_strikeout_rate_last_1_game = player_performances[past_game.home_batting_spot_9][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_9][\"strikeouts_last_1_game\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_9][\"at_bats_last_1_game\"].reduce(:+)\n\n feature_set.home_batting_spot_1_strikeout_rate_last_2_games = player_performances[past_game.home_batting_spot_1][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_1][\"strikeouts_last_2_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_1][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.home_batting_spot_2_strikeout_rate_last_2_games = player_performances[past_game.home_batting_spot_2][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_2][\"strikeouts_last_2_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_2][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.home_batting_spot_3_strikeout_rate_last_2_games = player_performances[past_game.home_batting_spot_3][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_3][\"strikeouts_last_2_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_3][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.home_batting_spot_4_strikeout_rate_last_2_games = player_performances[past_game.home_batting_spot_4][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_4][\"strikeouts_last_2_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_4][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.home_batting_spot_5_strikeout_rate_last_2_games = player_performances[past_game.home_batting_spot_5][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_5][\"strikeouts_last_2_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_5][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.home_batting_spot_6_strikeout_rate_last_2_games = player_performances[past_game.home_batting_spot_6][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_6][\"strikeouts_last_2_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_6][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.home_batting_spot_7_strikeout_rate_last_2_games = player_performances[past_game.home_batting_spot_7][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_7][\"strikeouts_last_2_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_7][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.home_batting_spot_8_strikeout_rate_last_2_games = player_performances[past_game.home_batting_spot_8][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_8][\"strikeouts_last_2_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_8][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.home_batting_spot_9_strikeout_rate_last_2_games = player_performances[past_game.home_batting_spot_9][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_9][\"strikeouts_last_2_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_9][\"at_bats_last_2_games\"].reduce(:+)\n\n feature_set.home_batting_spot_1_strikeout_rate_last_5_games = player_performances[past_game.home_batting_spot_1][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_1][\"strikeouts_last_5_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_1][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.home_batting_spot_2_strikeout_rate_last_5_games = player_performances[past_game.home_batting_spot_2][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_2][\"strikeouts_last_5_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_2][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.home_batting_spot_3_strikeout_rate_last_5_games = player_performances[past_game.home_batting_spot_3][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_3][\"strikeouts_last_5_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_3][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.home_batting_spot_4_strikeout_rate_last_5_games = player_performances[past_game.home_batting_spot_4][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_4][\"strikeouts_last_5_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_4][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.home_batting_spot_5_strikeout_rate_last_5_games = player_performances[past_game.home_batting_spot_5][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_5][\"strikeouts_last_5_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_5][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.home_batting_spot_6_strikeout_rate_last_5_games = player_performances[past_game.home_batting_spot_6][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_6][\"strikeouts_last_5_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_6][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.home_batting_spot_7_strikeout_rate_last_5_games = player_performances[past_game.home_batting_spot_7][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_7][\"strikeouts_last_5_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_7][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.home_batting_spot_8_strikeout_rate_last_5_games = player_performances[past_game.home_batting_spot_8][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_8][\"strikeouts_last_5_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_8][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.home_batting_spot_9_strikeout_rate_last_5_games = player_performances[past_game.home_batting_spot_9][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_9][\"strikeouts_last_5_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_9][\"at_bats_last_5_games\"].reduce(:+)\n\n feature_set.home_batting_spot_1_strikeout_rate_last_10_games = player_performances[past_game.home_batting_spot_1][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_1][\"strikeouts_last_10_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_1][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.home_batting_spot_2_strikeout_rate_last_10_games = player_performances[past_game.home_batting_spot_2][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_2][\"strikeouts_last_10_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_2][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.home_batting_spot_3_strikeout_rate_last_10_games = player_performances[past_game.home_batting_spot_3][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_3][\"strikeouts_last_10_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_3][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.home_batting_spot_4_strikeout_rate_last_10_games = player_performances[past_game.home_batting_spot_4][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_4][\"strikeouts_last_10_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_4][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.home_batting_spot_5_strikeout_rate_last_10_games = player_performances[past_game.home_batting_spot_5][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_5][\"strikeouts_last_10_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_5][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.home_batting_spot_6_strikeout_rate_last_10_games = player_performances[past_game.home_batting_spot_6][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_6][\"strikeouts_last_10_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_6][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.home_batting_spot_7_strikeout_rate_last_10_games = player_performances[past_game.home_batting_spot_7][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_7][\"strikeouts_last_10_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_7][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.home_batting_spot_8_strikeout_rate_last_10_games = player_performances[past_game.home_batting_spot_8][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_8][\"strikeouts_last_10_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_8][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.home_batting_spot_9_strikeout_rate_last_10_games = player_performances[past_game.home_batting_spot_9][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_9][\"strikeouts_last_10_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_9][\"at_bats_last_10_games\"].reduce(:+)\n\n feature_set.home_batting_spot_1_strikeout_rate_last_20_games = player_performances[past_game.home_batting_spot_1][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_1][\"strikeouts_last_20_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_1][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.home_batting_spot_2_strikeout_rate_last_20_games = player_performances[past_game.home_batting_spot_2][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_2][\"strikeouts_last_20_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_2][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.home_batting_spot_3_strikeout_rate_last_20_games = player_performances[past_game.home_batting_spot_3][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_3][\"strikeouts_last_20_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_3][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.home_batting_spot_4_strikeout_rate_last_20_games = player_performances[past_game.home_batting_spot_4][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_4][\"strikeouts_last_20_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_4][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.home_batting_spot_5_strikeout_rate_last_20_games = player_performances[past_game.home_batting_spot_5][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_5][\"strikeouts_last_20_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_5][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.home_batting_spot_6_strikeout_rate_last_20_games = player_performances[past_game.home_batting_spot_6][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_6][\"strikeouts_last_20_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_6][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.home_batting_spot_7_strikeout_rate_last_20_games = player_performances[past_game.home_batting_spot_7][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_7][\"strikeouts_last_20_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_7][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.home_batting_spot_8_strikeout_rate_last_20_games = player_performances[past_game.home_batting_spot_8][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_8][\"strikeouts_last_20_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_8][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.home_batting_spot_9_strikeout_rate_last_20_games = player_performances[past_game.home_batting_spot_9][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.home_batting_spot_9][\"strikeouts_last_20_games\"].reduce(:+).to_f / player_performances[past_game.home_batting_spot_9][\"at_bats_last_20_games\"].reduce(:+)\n \n feature_set.home_batting_spot_1_strikeout_rate_career = player_performances[past_game.home_batting_spot_1][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.home_batting_spot_1][\"career_strikeouts\"] / player_performances[past_game.home_batting_spot_1][\"career_at_bats\"]\n feature_set.home_batting_spot_2_strikeout_rate_career = player_performances[past_game.home_batting_spot_2][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.home_batting_spot_2][\"career_strikeouts\"] / player_performances[past_game.home_batting_spot_2][\"career_at_bats\"]\n feature_set.home_batting_spot_3_strikeout_rate_career = player_performances[past_game.home_batting_spot_3][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.home_batting_spot_3][\"career_strikeouts\"] / player_performances[past_game.home_batting_spot_3][\"career_at_bats\"]\n feature_set.home_batting_spot_4_strikeout_rate_career = player_performances[past_game.home_batting_spot_4][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.home_batting_spot_4][\"career_strikeouts\"] / player_performances[past_game.home_batting_spot_4][\"career_at_bats\"]\n feature_set.home_batting_spot_5_strikeout_rate_career = player_performances[past_game.home_batting_spot_5][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.home_batting_spot_5][\"career_strikeouts\"] / player_performances[past_game.home_batting_spot_5][\"career_at_bats\"]\n feature_set.home_batting_spot_6_strikeout_rate_career = player_performances[past_game.home_batting_spot_6][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.home_batting_spot_6][\"career_strikeouts\"] / player_performances[past_game.home_batting_spot_6][\"career_at_bats\"]\n feature_set.home_batting_spot_7_strikeout_rate_career = player_performances[past_game.home_batting_spot_7][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.home_batting_spot_7][\"career_strikeouts\"] / player_performances[past_game.home_batting_spot_7][\"career_at_bats\"]\n feature_set.home_batting_spot_8_strikeout_rate_career = player_performances[past_game.home_batting_spot_8][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.home_batting_spot_8][\"career_strikeouts\"] / player_performances[past_game.home_batting_spot_8][\"career_at_bats\"]\n feature_set.home_batting_spot_9_strikeout_rate_career = player_performances[past_game.home_batting_spot_9][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.home_batting_spot_9][\"career_strikeouts\"] / player_performances[past_game.home_batting_spot_9][\"career_at_bats\"]\n else\n \n feature_set.away_batting_spot_1_walks_last_1_game = player_performances[past_game.away_batting_spot_1][\"walks_last_1_game\"].reduce(:+)\n feature_set.away_batting_spot_2_walks_last_1_game = player_performances[past_game.away_batting_spot_2][\"walks_last_1_game\"].reduce(:+)\n feature_set.away_batting_spot_3_walks_last_1_game = player_performances[past_game.away_batting_spot_3][\"walks_last_1_game\"].reduce(:+)\n feature_set.away_batting_spot_4_walks_last_1_game = player_performances[past_game.away_batting_spot_4][\"walks_last_1_game\"].reduce(:+)\n feature_set.away_batting_spot_5_walks_last_1_game = player_performances[past_game.away_batting_spot_5][\"walks_last_1_game\"].reduce(:+)\n feature_set.away_batting_spot_6_walks_last_1_game = player_performances[past_game.away_batting_spot_6][\"walks_last_1_game\"].reduce(:+)\n feature_set.away_batting_spot_7_walks_last_1_game = player_performances[past_game.away_batting_spot_7][\"walks_last_1_game\"].reduce(:+)\n feature_set.away_batting_spot_8_walks_last_1_game = player_performances[past_game.away_batting_spot_8][\"walks_last_1_game\"].reduce(:+)\n feature_set.away_batting_spot_9_walks_last_1_game = player_performances[past_game.away_batting_spot_9][\"walks_last_1_game\"].reduce(:+)\n\n feature_set.away_batting_spot_1_walks_last_2_games = player_performances[past_game.away_batting_spot_1][\"walks_last_2_games\"].reduce(:+)\n feature_set.away_batting_spot_2_walks_last_2_games = player_performances[past_game.away_batting_spot_2][\"walks_last_2_games\"].reduce(:+)\n feature_set.away_batting_spot_3_walks_last_2_games = player_performances[past_game.away_batting_spot_3][\"walks_last_2_games\"].reduce(:+)\n feature_set.away_batting_spot_4_walks_last_2_games = player_performances[past_game.away_batting_spot_4][\"walks_last_2_games\"].reduce(:+)\n feature_set.away_batting_spot_5_walks_last_2_games = player_performances[past_game.away_batting_spot_5][\"walks_last_2_games\"].reduce(:+)\n feature_set.away_batting_spot_6_walks_last_2_games = player_performances[past_game.away_batting_spot_6][\"walks_last_2_games\"].reduce(:+)\n feature_set.away_batting_spot_7_walks_last_2_games = player_performances[past_game.away_batting_spot_7][\"walks_last_2_games\"].reduce(:+)\n feature_set.away_batting_spot_8_walks_last_2_games = player_performances[past_game.away_batting_spot_8][\"walks_last_2_games\"].reduce(:+)\n feature_set.away_batting_spot_9_walks_last_2_games = player_performances[past_game.away_batting_spot_9][\"walks_last_2_games\"].reduce(:+)\n\n feature_set.away_batting_spot_1_walks_last_5_games = player_performances[past_game.away_batting_spot_1][\"walks_last_5_games\"].reduce(:+)\n feature_set.away_batting_spot_2_walks_last_5_games = player_performances[past_game.away_batting_spot_2][\"walks_last_5_games\"].reduce(:+)\n feature_set.away_batting_spot_3_walks_last_5_games = player_performances[past_game.away_batting_spot_3][\"walks_last_5_games\"].reduce(:+)\n feature_set.away_batting_spot_4_walks_last_5_games = player_performances[past_game.away_batting_spot_4][\"walks_last_5_games\"].reduce(:+)\n feature_set.away_batting_spot_5_walks_last_5_games = player_performances[past_game.away_batting_spot_5][\"walks_last_5_games\"].reduce(:+)\n feature_set.away_batting_spot_6_walks_last_5_games = player_performances[past_game.away_batting_spot_6][\"walks_last_5_games\"].reduce(:+)\n feature_set.away_batting_spot_7_walks_last_5_games = player_performances[past_game.away_batting_spot_7][\"walks_last_5_games\"].reduce(:+)\n feature_set.away_batting_spot_8_walks_last_5_games = player_performances[past_game.away_batting_spot_8][\"walks_last_5_games\"].reduce(:+)\n feature_set.away_batting_spot_9_walks_last_5_games = player_performances[past_game.away_batting_spot_9][\"walks_last_5_games\"].reduce(:+)\n\n feature_set.away_batting_spot_1_walks_last_10_games = player_performances[past_game.away_batting_spot_1][\"walks_last_10_games\"].reduce(:+)\n feature_set.away_batting_spot_2_walks_last_10_games = player_performances[past_game.away_batting_spot_2][\"walks_last_10_games\"].reduce(:+)\n feature_set.away_batting_spot_3_walks_last_10_games = player_performances[past_game.away_batting_spot_3][\"walks_last_10_games\"].reduce(:+)\n feature_set.away_batting_spot_4_walks_last_10_games = player_performances[past_game.away_batting_spot_4][\"walks_last_10_games\"].reduce(:+)\n feature_set.away_batting_spot_5_walks_last_10_games = player_performances[past_game.away_batting_spot_5][\"walks_last_10_games\"].reduce(:+)\n feature_set.away_batting_spot_6_walks_last_10_games = player_performances[past_game.away_batting_spot_6][\"walks_last_10_games\"].reduce(:+)\n feature_set.away_batting_spot_7_walks_last_10_games = player_performances[past_game.away_batting_spot_7][\"walks_last_10_games\"].reduce(:+)\n feature_set.away_batting_spot_8_walks_last_10_games = player_performances[past_game.away_batting_spot_8][\"walks_last_10_games\"].reduce(:+)\n feature_set.away_batting_spot_9_walks_last_10_games = player_performances[past_game.away_batting_spot_9][\"walks_last_10_games\"].reduce(:+)\n\n feature_set.away_batting_spot_1_walks_last_20_games = player_performances[past_game.away_batting_spot_1][\"walks_last_20_games\"].reduce(:+)\n feature_set.away_batting_spot_2_walks_last_20_games = player_performances[past_game.away_batting_spot_2][\"walks_last_20_games\"].reduce(:+)\n feature_set.away_batting_spot_3_walks_last_20_games = player_performances[past_game.away_batting_spot_3][\"walks_last_20_games\"].reduce(:+)\n feature_set.away_batting_spot_4_walks_last_20_games = player_performances[past_game.away_batting_spot_4][\"walks_last_20_games\"].reduce(:+)\n feature_set.away_batting_spot_5_walks_last_20_games = player_performances[past_game.away_batting_spot_5][\"walks_last_20_games\"].reduce(:+)\n feature_set.away_batting_spot_6_walks_last_20_games = player_performances[past_game.away_batting_spot_6][\"walks_last_20_games\"].reduce(:+)\n feature_set.away_batting_spot_7_walks_last_20_games = player_performances[past_game.away_batting_spot_7][\"walks_last_20_games\"].reduce(:+)\n feature_set.away_batting_spot_8_walks_last_20_games = player_performances[past_game.away_batting_spot_8][\"walks_last_20_games\"].reduce(:+)\n feature_set.away_batting_spot_9_walks_last_20_games = player_performances[past_game.away_batting_spot_9][\"walks_last_20_games\"].reduce(:+)\n\n feature_set.away_batting_spot_1_walks_per_game_career = player_performances[past_game.away_batting_spot_1][\"career_games\"] == 0 ? 0 : player_performances[past_game.away_batting_spot_1][\"career_walks\"] / player_performances[past_game.away_batting_spot_1][\"career_games\"]\n feature_set.away_batting_spot_2_walks_per_game_career = player_performances[past_game.away_batting_spot_2][\"career_games\"] == 0 ? 0 : player_performances[past_game.away_batting_spot_2][\"career_walks\"] / player_performances[past_game.away_batting_spot_2][\"career_games\"]\n feature_set.away_batting_spot_3_walks_per_game_career = player_performances[past_game.away_batting_spot_3][\"career_games\"] == 0 ? 0 : player_performances[past_game.away_batting_spot_3][\"career_walks\"] / player_performances[past_game.away_batting_spot_3][\"career_games\"]\n feature_set.away_batting_spot_4_walks_per_game_career = player_performances[past_game.away_batting_spot_4][\"career_games\"] == 0 ? 0 : player_performances[past_game.away_batting_spot_4][\"career_walks\"] / player_performances[past_game.away_batting_spot_4][\"career_games\"]\n feature_set.away_batting_spot_5_walks_per_game_career = player_performances[past_game.away_batting_spot_5][\"career_games\"] == 0 ? 0 : player_performances[past_game.away_batting_spot_5][\"career_walks\"] / player_performances[past_game.away_batting_spot_5][\"career_games\"]\n feature_set.away_batting_spot_6_walks_per_game_career = player_performances[past_game.away_batting_spot_6][\"career_games\"] == 0 ? 0 : player_performances[past_game.away_batting_spot_6][\"career_walks\"] / player_performances[past_game.away_batting_spot_6][\"career_games\"]\n feature_set.away_batting_spot_7_walks_per_game_career = player_performances[past_game.away_batting_spot_7][\"career_games\"] == 0 ? 0 : player_performances[past_game.away_batting_spot_7][\"career_walks\"] / player_performances[past_game.away_batting_spot_7][\"career_games\"]\n feature_set.away_batting_spot_8_walks_per_game_career = player_performances[past_game.away_batting_spot_8][\"career_games\"] == 0 ? 0 : player_performances[past_game.away_batting_spot_8][\"career_walks\"] / player_performances[past_game.away_batting_spot_8][\"career_games\"]\n feature_set.away_batting_spot_9_walks_per_game_career = player_performances[past_game.away_batting_spot_9][\"career_games\"] == 0 ? 0 : player_performances[past_game.away_batting_spot_9][\"career_walks\"] / player_performances[past_game.away_batting_spot_9][\"career_games\"]\n \n\n feature_set.away_batting_spot_1_batting_percentage_last_1_game = player_performances[past_game.away_batting_spot_1][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_1][\"hits_last_1_game\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_1][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.away_batting_spot_2_batting_percentage_last_1_game = player_performances[past_game.away_batting_spot_2][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_2][\"hits_last_1_game\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_2][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.away_batting_spot_3_batting_percentage_last_1_game = player_performances[past_game.away_batting_spot_3][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_3][\"hits_last_1_game\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_3][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.away_batting_spot_4_batting_percentage_last_1_game = player_performances[past_game.away_batting_spot_4][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_4][\"hits_last_1_game\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_4][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.away_batting_spot_5_batting_percentage_last_1_game = player_performances[past_game.away_batting_spot_5][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_5][\"hits_last_1_game\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_5][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.away_batting_spot_6_batting_percentage_last_1_game = player_performances[past_game.away_batting_spot_6][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_6][\"hits_last_1_game\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_6][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.away_batting_spot_7_batting_percentage_last_1_game = player_performances[past_game.away_batting_spot_7][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_7][\"hits_last_1_game\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_7][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.away_batting_spot_8_batting_percentage_last_1_game = player_performances[past_game.away_batting_spot_8][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_8][\"hits_last_1_game\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_8][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.away_batting_spot_9_batting_percentage_last_1_game = player_performances[past_game.away_batting_spot_9][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_9][\"hits_last_1_game\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_9][\"at_bats_last_1_game\"].reduce(:+)\n\n feature_set.away_batting_spot_1_batting_percentage_last_2_games = player_performances[past_game.away_batting_spot_1][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_1][\"hits_last_2_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_1][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.away_batting_spot_2_batting_percentage_last_2_games = player_performances[past_game.away_batting_spot_2][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_2][\"hits_last_2_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_2][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.away_batting_spot_3_batting_percentage_last_2_games = player_performances[past_game.away_batting_spot_3][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_3][\"hits_last_2_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_3][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.away_batting_spot_4_batting_percentage_last_2_games = player_performances[past_game.away_batting_spot_4][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_4][\"hits_last_2_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_4][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.away_batting_spot_5_batting_percentage_last_2_games = player_performances[past_game.away_batting_spot_5][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_5][\"hits_last_2_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_5][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.away_batting_spot_6_batting_percentage_last_2_games = player_performances[past_game.away_batting_spot_6][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_6][\"hits_last_2_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_6][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.away_batting_spot_7_batting_percentage_last_2_games = player_performances[past_game.away_batting_spot_7][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_7][\"hits_last_2_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_7][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.away_batting_spot_8_batting_percentage_last_2_games = player_performances[past_game.away_batting_spot_8][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_8][\"hits_last_2_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_8][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.away_batting_spot_9_batting_percentage_last_2_games = player_performances[past_game.away_batting_spot_9][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_9][\"hits_last_2_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_9][\"at_bats_last_2_games\"].reduce(:+)\n\n feature_set.away_batting_spot_1_batting_percentage_last_5_games = player_performances[past_game.away_batting_spot_1][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_1][\"hits_last_5_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_1][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.away_batting_spot_2_batting_percentage_last_5_games = player_performances[past_game.away_batting_spot_2][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_2][\"hits_last_5_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_2][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.away_batting_spot_3_batting_percentage_last_5_games = player_performances[past_game.away_batting_spot_3][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_3][\"hits_last_5_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_3][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.away_batting_spot_4_batting_percentage_last_5_games = player_performances[past_game.away_batting_spot_4][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_4][\"hits_last_5_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_4][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.away_batting_spot_5_batting_percentage_last_5_games = player_performances[past_game.away_batting_spot_5][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_5][\"hits_last_5_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_5][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.away_batting_spot_6_batting_percentage_last_5_games = player_performances[past_game.away_batting_spot_6][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_6][\"hits_last_5_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_6][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.away_batting_spot_7_batting_percentage_last_5_games = player_performances[past_game.away_batting_spot_7][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_7][\"hits_last_5_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_7][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.away_batting_spot_8_batting_percentage_last_5_games = player_performances[past_game.away_batting_spot_8][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_8][\"hits_last_5_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_8][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.away_batting_spot_9_batting_percentage_last_5_games = player_performances[past_game.away_batting_spot_9][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_9][\"hits_last_5_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_9][\"at_bats_last_5_games\"].reduce(:+)\n\n feature_set.away_batting_spot_1_batting_percentage_last_10_games = player_performances[past_game.away_batting_spot_1][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_1][\"hits_last_10_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_1][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.away_batting_spot_2_batting_percentage_last_10_games = player_performances[past_game.away_batting_spot_2][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_2][\"hits_last_10_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_2][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.away_batting_spot_3_batting_percentage_last_10_games = player_performances[past_game.away_batting_spot_3][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_3][\"hits_last_10_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_3][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.away_batting_spot_4_batting_percentage_last_10_games = player_performances[past_game.away_batting_spot_4][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_4][\"hits_last_10_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_4][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.away_batting_spot_5_batting_percentage_last_10_games = player_performances[past_game.away_batting_spot_5][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_5][\"hits_last_10_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_5][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.away_batting_spot_6_batting_percentage_last_10_games = player_performances[past_game.away_batting_spot_6][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_6][\"hits_last_10_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_6][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.away_batting_spot_7_batting_percentage_last_10_games = player_performances[past_game.away_batting_spot_7][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_7][\"hits_last_10_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_7][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.away_batting_spot_8_batting_percentage_last_10_games = player_performances[past_game.away_batting_spot_8][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_8][\"hits_last_10_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_8][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.away_batting_spot_9_batting_percentage_last_10_games = player_performances[past_game.away_batting_spot_9][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_9][\"hits_last_10_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_9][\"at_bats_last_10_games\"].reduce(:+)\n\n feature_set.away_batting_spot_1_batting_percentage_last_20_games = player_performances[past_game.away_batting_spot_1][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_1][\"hits_last_20_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_1][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.away_batting_spot_2_batting_percentage_last_20_games = player_performances[past_game.away_batting_spot_2][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_2][\"hits_last_20_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_2][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.away_batting_spot_3_batting_percentage_last_20_games = player_performances[past_game.away_batting_spot_3][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_3][\"hits_last_20_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_3][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.away_batting_spot_4_batting_percentage_last_20_games = player_performances[past_game.away_batting_spot_4][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_4][\"hits_last_20_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_4][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.away_batting_spot_5_batting_percentage_last_20_games = player_performances[past_game.away_batting_spot_5][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_5][\"hits_last_20_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_5][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.away_batting_spot_6_batting_percentage_last_20_games = player_performances[past_game.away_batting_spot_6][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_6][\"hits_last_20_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_6][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.away_batting_spot_7_batting_percentage_last_20_games = player_performances[past_game.away_batting_spot_7][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_7][\"hits_last_20_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_7][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.away_batting_spot_8_batting_percentage_last_20_games = player_performances[past_game.away_batting_spot_8][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_8][\"hits_last_20_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_8][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.away_batting_spot_9_batting_percentage_last_20_games = player_performances[past_game.away_batting_spot_9][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_9][\"hits_last_20_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_9][\"at_bats_last_20_games\"].reduce(:+)\n \n feature_set.away_batting_spot_1_batting_percentage_career = player_performances[past_game.away_batting_spot_1][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.away_batting_spot_1][\"career_hits\"] / player_performances[past_game.away_batting_spot_1][\"career_at_bats\"]\n feature_set.away_batting_spot_2_batting_percentage_career = player_performances[past_game.away_batting_spot_2][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.away_batting_spot_2][\"career_hits\"] / player_performances[past_game.away_batting_spot_2][\"career_at_bats\"]\n feature_set.away_batting_spot_3_batting_percentage_career = player_performances[past_game.away_batting_spot_3][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.away_batting_spot_3][\"career_hits\"] / player_performances[past_game.away_batting_spot_3][\"career_at_bats\"]\n feature_set.away_batting_spot_4_batting_percentage_career = player_performances[past_game.away_batting_spot_4][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.away_batting_spot_4][\"career_hits\"] / player_performances[past_game.away_batting_spot_4][\"career_at_bats\"]\n feature_set.away_batting_spot_5_batting_percentage_career = player_performances[past_game.away_batting_spot_5][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.away_batting_spot_5][\"career_hits\"] / player_performances[past_game.away_batting_spot_5][\"career_at_bats\"]\n feature_set.away_batting_spot_6_batting_percentage_career = player_performances[past_game.away_batting_spot_6][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.away_batting_spot_6][\"career_hits\"] / player_performances[past_game.away_batting_spot_6][\"career_at_bats\"]\n feature_set.away_batting_spot_7_batting_percentage_career = player_performances[past_game.away_batting_spot_7][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.away_batting_spot_7][\"career_hits\"] / player_performances[past_game.away_batting_spot_7][\"career_at_bats\"]\n feature_set.away_batting_spot_8_batting_percentage_career = player_performances[past_game.away_batting_spot_8][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.away_batting_spot_8][\"career_hits\"] / player_performances[past_game.away_batting_spot_8][\"career_at_bats\"]\n feature_set.away_batting_spot_9_batting_percentage_career = player_performances[past_game.away_batting_spot_9][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.away_batting_spot_9][\"career_hits\"] / player_performances[past_game.away_batting_spot_9][\"career_at_bats\"]\n\n feature_set.away_batting_spot_1_OPS_last_1_game = player_performances[past_game.away_batting_spot_1][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_1][\"total_bases_last_1_game\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_1][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.away_batting_spot_2_OPS_last_1_game = player_performances[past_game.away_batting_spot_2][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_2][\"total_bases_last_1_game\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_2][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.away_batting_spot_3_OPS_last_1_game = player_performances[past_game.away_batting_spot_3][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_3][\"total_bases_last_1_game\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_3][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.away_batting_spot_4_OPS_last_1_game = player_performances[past_game.away_batting_spot_4][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_4][\"total_bases_last_1_game\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_4][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.away_batting_spot_5_OPS_last_1_game = player_performances[past_game.away_batting_spot_5][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_5][\"total_bases_last_1_game\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_5][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.away_batting_spot_6_OPS_last_1_game = player_performances[past_game.away_batting_spot_6][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_6][\"total_bases_last_1_game\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_6][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.away_batting_spot_7_OPS_last_1_game = player_performances[past_game.away_batting_spot_7][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_7][\"total_bases_last_1_game\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_7][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.away_batting_spot_8_OPS_last_1_game = player_performances[past_game.away_batting_spot_8][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_8][\"total_bases_last_1_game\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_8][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.away_batting_spot_9_OPS_last_1_game = player_performances[past_game.away_batting_spot_9][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_9][\"total_bases_last_1_game\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_9][\"at_bats_last_1_game\"].reduce(:+)\n\n feature_set.away_batting_spot_1_OPS_last_2_games = player_performances[past_game.away_batting_spot_1][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_1][\"total_bases_last_2_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_1][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.away_batting_spot_2_OPS_last_2_games = player_performances[past_game.away_batting_spot_2][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_2][\"total_bases_last_2_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_2][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.away_batting_spot_3_OPS_last_2_games = player_performances[past_game.away_batting_spot_3][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_3][\"total_bases_last_2_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_3][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.away_batting_spot_4_OPS_last_2_games = player_performances[past_game.away_batting_spot_4][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_4][\"total_bases_last_2_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_4][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.away_batting_spot_5_OPS_last_2_games = player_performances[past_game.away_batting_spot_5][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_5][\"total_bases_last_2_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_5][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.away_batting_spot_6_OPS_last_2_games = player_performances[past_game.away_batting_spot_6][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_6][\"total_bases_last_2_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_6][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.away_batting_spot_7_OPS_last_2_games = player_performances[past_game.away_batting_spot_7][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_7][\"total_bases_last_2_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_7][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.away_batting_spot_8_OPS_last_2_games = player_performances[past_game.away_batting_spot_8][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_8][\"total_bases_last_2_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_8][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.away_batting_spot_9_OPS_last_2_games = player_performances[past_game.away_batting_spot_9][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_9][\"total_bases_last_2_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_9][\"at_bats_last_2_games\"].reduce(:+)\n\n feature_set.away_batting_spot_1_OPS_last_5_games = player_performances[past_game.away_batting_spot_1][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_1][\"total_bases_last_5_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_1][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.away_batting_spot_2_OPS_last_5_games = player_performances[past_game.away_batting_spot_2][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_2][\"total_bases_last_5_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_2][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.away_batting_spot_3_OPS_last_5_games = player_performances[past_game.away_batting_spot_3][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_3][\"total_bases_last_5_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_3][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.away_batting_spot_4_OPS_last_5_games = player_performances[past_game.away_batting_spot_4][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_4][\"total_bases_last_5_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_4][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.away_batting_spot_5_OPS_last_5_games = player_performances[past_game.away_batting_spot_5][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_5][\"total_bases_last_5_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_5][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.away_batting_spot_6_OPS_last_5_games = player_performances[past_game.away_batting_spot_6][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_6][\"total_bases_last_5_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_6][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.away_batting_spot_7_OPS_last_5_games = player_performances[past_game.away_batting_spot_7][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_7][\"total_bases_last_5_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_7][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.away_batting_spot_8_OPS_last_5_games = player_performances[past_game.away_batting_spot_8][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_8][\"total_bases_last_5_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_8][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.away_batting_spot_9_OPS_last_5_games = player_performances[past_game.away_batting_spot_9][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_9][\"total_bases_last_5_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_9][\"at_bats_last_5_games\"].reduce(:+)\n\n feature_set.away_batting_spot_1_OPS_last_10_games = player_performances[past_game.away_batting_spot_1][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_1][\"total_bases_last_10_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_1][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.away_batting_spot_2_OPS_last_10_games = player_performances[past_game.away_batting_spot_2][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_2][\"total_bases_last_10_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_2][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.away_batting_spot_3_OPS_last_10_games = player_performances[past_game.away_batting_spot_3][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_3][\"total_bases_last_10_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_3][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.away_batting_spot_4_OPS_last_10_games = player_performances[past_game.away_batting_spot_4][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_4][\"total_bases_last_10_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_4][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.away_batting_spot_5_OPS_last_10_games = player_performances[past_game.away_batting_spot_5][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_5][\"total_bases_last_10_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_5][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.away_batting_spot_6_OPS_last_10_games = player_performances[past_game.away_batting_spot_6][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_6][\"total_bases_last_10_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_6][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.away_batting_spot_7_OPS_last_10_games = player_performances[past_game.away_batting_spot_7][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_7][\"total_bases_last_10_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_7][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.away_batting_spot_8_OPS_last_10_games = player_performances[past_game.away_batting_spot_8][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_8][\"total_bases_last_10_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_8][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.away_batting_spot_9_OPS_last_10_games = player_performances[past_game.away_batting_spot_9][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_9][\"total_bases_last_10_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_9][\"at_bats_last_10_games\"].reduce(:+)\n\n feature_set.away_batting_spot_1_OPS_last_20_games = player_performances[past_game.away_batting_spot_1][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_1][\"total_bases_last_20_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_1][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.away_batting_spot_2_OPS_last_20_games = player_performances[past_game.away_batting_spot_2][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_2][\"total_bases_last_20_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_2][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.away_batting_spot_3_OPS_last_20_games = player_performances[past_game.away_batting_spot_3][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_3][\"total_bases_last_20_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_3][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.away_batting_spot_4_OPS_last_20_games = player_performances[past_game.away_batting_spot_4][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_4][\"total_bases_last_20_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_4][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.away_batting_spot_5_OPS_last_20_games = player_performances[past_game.away_batting_spot_5][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_5][\"total_bases_last_20_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_5][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.away_batting_spot_6_OPS_last_20_games = player_performances[past_game.away_batting_spot_6][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_6][\"total_bases_last_20_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_6][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.away_batting_spot_7_OPS_last_20_games = player_performances[past_game.away_batting_spot_7][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_7][\"total_bases_last_20_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_7][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.away_batting_spot_8_OPS_last_20_games = player_performances[past_game.away_batting_spot_8][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_8][\"total_bases_last_20_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_8][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.away_batting_spot_9_OPS_last_20_games = player_performances[past_game.away_batting_spot_9][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_9][\"total_bases_last_20_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_9][\"at_bats_last_20_games\"].reduce(:+)\n\n \n feature_set.away_batting_spot_1_OPS_career = player_performances[past_game.away_batting_spot_1][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.away_batting_spot_1][\"career_total_bases\"] / player_performances[past_game.away_batting_spot_1][\"career_at_bats\"]\n feature_set.away_batting_spot_2_OPS_career = player_performances[past_game.away_batting_spot_2][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.away_batting_spot_2][\"career_total_bases\"] / player_performances[past_game.away_batting_spot_2][\"career_at_bats\"]\n feature_set.away_batting_spot_3_OPS_career = player_performances[past_game.away_batting_spot_3][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.away_batting_spot_3][\"career_total_bases\"] / player_performances[past_game.away_batting_spot_3][\"career_at_bats\"]\n feature_set.away_batting_spot_4_OPS_career = player_performances[past_game.away_batting_spot_4][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.away_batting_spot_4][\"career_total_bases\"] / player_performances[past_game.away_batting_spot_4][\"career_at_bats\"]\n feature_set.away_batting_spot_5_OPS_career = player_performances[past_game.away_batting_spot_5][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.away_batting_spot_5][\"career_total_bases\"] / player_performances[past_game.away_batting_spot_5][\"career_at_bats\"]\n feature_set.away_batting_spot_6_OPS_career = player_performances[past_game.away_batting_spot_6][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.away_batting_spot_6][\"career_total_bases\"] / player_performances[past_game.away_batting_spot_6][\"career_at_bats\"]\n feature_set.away_batting_spot_7_OPS_career = player_performances[past_game.away_batting_spot_7][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.away_batting_spot_7][\"career_total_bases\"] / player_performances[past_game.away_batting_spot_7][\"career_at_bats\"]\n feature_set.away_batting_spot_8_OPS_career = player_performances[past_game.away_batting_spot_8][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.away_batting_spot_8][\"career_total_bases\"] / player_performances[past_game.away_batting_spot_8][\"career_at_bats\"]\n feature_set.away_batting_spot_9_OPS_career = player_performances[past_game.away_batting_spot_9][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.away_batting_spot_9][\"career_total_bases\"] / player_performances[past_game.away_batting_spot_9][\"career_at_bats\"]\n\n feature_set.away_batting_spot_1_strikeout_rate_last_1_game = player_performances[past_game.away_batting_spot_1][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_1][\"strikeouts_last_1_game\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_1][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.away_batting_spot_2_strikeout_rate_last_1_game = player_performances[past_game.away_batting_spot_2][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_2][\"strikeouts_last_1_game\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_2][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.away_batting_spot_3_strikeout_rate_last_1_game = player_performances[past_game.away_batting_spot_3][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_3][\"strikeouts_last_1_game\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_3][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.away_batting_spot_4_strikeout_rate_last_1_game = player_performances[past_game.away_batting_spot_4][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_4][\"strikeouts_last_1_game\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_4][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.away_batting_spot_5_strikeout_rate_last_1_game = player_performances[past_game.away_batting_spot_5][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_5][\"strikeouts_last_1_game\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_5][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.away_batting_spot_6_strikeout_rate_last_1_game = player_performances[past_game.away_batting_spot_6][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_6][\"strikeouts_last_1_game\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_6][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.away_batting_spot_7_strikeout_rate_last_1_game = player_performances[past_game.away_batting_spot_7][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_7][\"strikeouts_last_1_game\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_7][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.away_batting_spot_8_strikeout_rate_last_1_game = player_performances[past_game.away_batting_spot_8][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_8][\"strikeouts_last_1_game\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_8][\"at_bats_last_1_game\"].reduce(:+)\n feature_set.away_batting_spot_9_strikeout_rate_last_1_game = player_performances[past_game.away_batting_spot_9][\"at_bats_last_1_game\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_9][\"strikeouts_last_1_game\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_9][\"at_bats_last_1_game\"].reduce(:+)\n\n feature_set.away_batting_spot_1_strikeout_rate_last_2_games = player_performances[past_game.away_batting_spot_1][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_1][\"strikeouts_last_2_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_1][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.away_batting_spot_2_strikeout_rate_last_2_games = player_performances[past_game.away_batting_spot_2][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_2][\"strikeouts_last_2_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_2][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.away_batting_spot_3_strikeout_rate_last_2_games = player_performances[past_game.away_batting_spot_3][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_3][\"strikeouts_last_2_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_3][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.away_batting_spot_4_strikeout_rate_last_2_games = player_performances[past_game.away_batting_spot_4][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_4][\"strikeouts_last_2_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_4][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.away_batting_spot_5_strikeout_rate_last_2_games = player_performances[past_game.away_batting_spot_5][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_5][\"strikeouts_last_2_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_5][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.away_batting_spot_6_strikeout_rate_last_2_games = player_performances[past_game.away_batting_spot_6][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_6][\"strikeouts_last_2_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_6][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.away_batting_spot_7_strikeout_rate_last_2_games = player_performances[past_game.away_batting_spot_7][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_7][\"strikeouts_last_2_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_7][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.away_batting_spot_8_strikeout_rate_last_2_games = player_performances[past_game.away_batting_spot_8][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_8][\"strikeouts_last_2_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_8][\"at_bats_last_2_games\"].reduce(:+)\n feature_set.away_batting_spot_9_strikeout_rate_last_2_games = player_performances[past_game.away_batting_spot_9][\"at_bats_last_2_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_9][\"strikeouts_last_2_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_9][\"at_bats_last_2_games\"].reduce(:+)\n\n feature_set.away_batting_spot_1_strikeout_rate_last_5_games = player_performances[past_game.away_batting_spot_1][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_1][\"strikeouts_last_5_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_1][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.away_batting_spot_2_strikeout_rate_last_5_games = player_performances[past_game.away_batting_spot_2][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_2][\"strikeouts_last_5_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_2][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.away_batting_spot_3_strikeout_rate_last_5_games = player_performances[past_game.away_batting_spot_3][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_3][\"strikeouts_last_5_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_3][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.away_batting_spot_4_strikeout_rate_last_5_games = player_performances[past_game.away_batting_spot_4][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_4][\"strikeouts_last_5_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_4][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.away_batting_spot_5_strikeout_rate_last_5_games = player_performances[past_game.away_batting_spot_5][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_5][\"strikeouts_last_5_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_5][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.away_batting_spot_6_strikeout_rate_last_5_games = player_performances[past_game.away_batting_spot_6][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_6][\"strikeouts_last_5_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_6][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.away_batting_spot_7_strikeout_rate_last_5_games = player_performances[past_game.away_batting_spot_7][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_7][\"strikeouts_last_5_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_7][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.away_batting_spot_8_strikeout_rate_last_5_games = player_performances[past_game.away_batting_spot_8][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_8][\"strikeouts_last_5_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_8][\"at_bats_last_5_games\"].reduce(:+)\n feature_set.away_batting_spot_9_strikeout_rate_last_5_games = player_performances[past_game.away_batting_spot_9][\"at_bats_last_5_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_9][\"strikeouts_last_5_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_9][\"at_bats_last_5_games\"].reduce(:+)\n\n feature_set.away_batting_spot_1_strikeout_rate_last_10_games = player_performances[past_game.away_batting_spot_1][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_1][\"strikeouts_last_10_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_1][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.away_batting_spot_2_strikeout_rate_last_10_games = player_performances[past_game.away_batting_spot_2][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_2][\"strikeouts_last_10_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_2][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.away_batting_spot_3_strikeout_rate_last_10_games = player_performances[past_game.away_batting_spot_3][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_3][\"strikeouts_last_10_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_3][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.away_batting_spot_4_strikeout_rate_last_10_games = player_performances[past_game.away_batting_spot_4][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_4][\"strikeouts_last_10_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_4][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.away_batting_spot_5_strikeout_rate_last_10_games = player_performances[past_game.away_batting_spot_5][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_5][\"strikeouts_last_10_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_5][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.away_batting_spot_6_strikeout_rate_last_10_games = player_performances[past_game.away_batting_spot_6][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_6][\"strikeouts_last_10_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_6][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.away_batting_spot_7_strikeout_rate_last_10_games = player_performances[past_game.away_batting_spot_7][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_7][\"strikeouts_last_10_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_7][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.away_batting_spot_8_strikeout_rate_last_10_games = player_performances[past_game.away_batting_spot_8][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_8][\"strikeouts_last_10_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_8][\"at_bats_last_10_games\"].reduce(:+)\n feature_set.away_batting_spot_9_strikeout_rate_last_10_games = player_performances[past_game.away_batting_spot_9][\"at_bats_last_10_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_9][\"strikeouts_last_10_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_9][\"at_bats_last_10_games\"].reduce(:+)\n\n feature_set.away_batting_spot_1_strikeout_rate_last_20_games = player_performances[past_game.away_batting_spot_1][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_1][\"strikeouts_last_20_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_1][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.away_batting_spot_2_strikeout_rate_last_20_games = player_performances[past_game.away_batting_spot_2][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_2][\"strikeouts_last_20_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_2][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.away_batting_spot_3_strikeout_rate_last_20_games = player_performances[past_game.away_batting_spot_3][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_3][\"strikeouts_last_20_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_3][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.away_batting_spot_4_strikeout_rate_last_20_games = player_performances[past_game.away_batting_spot_4][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_4][\"strikeouts_last_20_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_4][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.away_batting_spot_5_strikeout_rate_last_20_games = player_performances[past_game.away_batting_spot_5][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_5][\"strikeouts_last_20_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_5][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.away_batting_spot_6_strikeout_rate_last_20_games = player_performances[past_game.away_batting_spot_6][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_6][\"strikeouts_last_20_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_6][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.away_batting_spot_7_strikeout_rate_last_20_games = player_performances[past_game.away_batting_spot_7][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_7][\"strikeouts_last_20_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_7][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.away_batting_spot_8_strikeout_rate_last_20_games = player_performances[past_game.away_batting_spot_8][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_8][\"strikeouts_last_20_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_8][\"at_bats_last_20_games\"].reduce(:+)\n feature_set.away_batting_spot_9_strikeout_rate_last_20_games = player_performances[past_game.away_batting_spot_9][\"at_bats_last_20_games\"].reduce(:+) == 0 ? 0 : player_performances[past_game.away_batting_spot_9][\"strikeouts_last_20_games\"].reduce(:+).to_f / player_performances[past_game.away_batting_spot_9][\"at_bats_last_20_games\"].reduce(:+)\n \n feature_set.away_batting_spot_1_strikeout_rate_career = player_performances[past_game.away_batting_spot_1][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.away_batting_spot_1][\"career_strikeouts\"] / player_performances[past_game.away_batting_spot_1][\"career_at_bats\"]\n feature_set.away_batting_spot_2_strikeout_rate_career = player_performances[past_game.away_batting_spot_2][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.away_batting_spot_2][\"career_strikeouts\"] / player_performances[past_game.away_batting_spot_2][\"career_at_bats\"]\n feature_set.away_batting_spot_3_strikeout_rate_career = player_performances[past_game.away_batting_spot_3][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.away_batting_spot_3][\"career_strikeouts\"] / player_performances[past_game.away_batting_spot_3][\"career_at_bats\"]\n feature_set.away_batting_spot_4_strikeout_rate_career = player_performances[past_game.away_batting_spot_4][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.away_batting_spot_4][\"career_strikeouts\"] / player_performances[past_game.away_batting_spot_4][\"career_at_bats\"]\n feature_set.away_batting_spot_5_strikeout_rate_career = player_performances[past_game.away_batting_spot_5][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.away_batting_spot_5][\"career_strikeouts\"] / player_performances[past_game.away_batting_spot_5][\"career_at_bats\"]\n feature_set.away_batting_spot_6_strikeout_rate_career = player_performances[past_game.away_batting_spot_6][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.away_batting_spot_6][\"career_strikeouts\"] / player_performances[past_game.away_batting_spot_6][\"career_at_bats\"]\n feature_set.away_batting_spot_7_strikeout_rate_career = player_performances[past_game.away_batting_spot_7][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.away_batting_spot_7][\"career_strikeouts\"] / player_performances[past_game.away_batting_spot_7][\"career_at_bats\"]\n feature_set.away_batting_spot_8_strikeout_rate_career = player_performances[past_game.away_batting_spot_8][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.away_batting_spot_8][\"career_strikeouts\"] / player_performances[past_game.away_batting_spot_8][\"career_at_bats\"]\n feature_set.away_batting_spot_9_strikeout_rate_career = player_performances[past_game.away_batting_spot_9][\"career_at_bats\"] == 0 ? 0 : player_performances[past_game.away_batting_spot_9][\"career_strikeouts\"] / player_performances[past_game.away_batting_spot_9][\"career_at_bats\"]\n end\n\n feature_set.save\n\n performances.each do |perf|\n player = Player.find_by_id(perf.player_id)\n # career updates\n player_performances[player.retrosheet_id][\"career_games\"] += 1\n player_performances[player.retrosheet_id][\"career_at_bats\"] += perf.at_bats\n player_performances[player.retrosheet_id][\"career_walks\"] += perf.walks\n player_performances[player.retrosheet_id][\"career_hits\"] += perf.hits\n player_performances[player.retrosheet_id][\"career_strikeouts\"] += perf.strikeouts \n player_performances[player.retrosheet_id][\"career_total_bases\"] += perf.total_bases \n \n player_performances[player.retrosheet_id][\"at_bats_last_1_game\"] << perf.at_bats\n player_performances[player.retrosheet_id][\"at_bats_last_2_games\"] << perf.at_bats\n player_performances[player.retrosheet_id][\"at_bats_last_5_games\"] << perf.at_bats\n player_performances[player.retrosheet_id][\"at_bats_last_10_games\"] << perf.at_bats\n player_performances[player.retrosheet_id][\"at_bats_last_20_games\"] << perf.at_bats\n \n player_performances[player.retrosheet_id][\"at_bats_last_1_game\"].shift\n player_performances[player.retrosheet_id][\"at_bats_last_2_games\"].shift\n player_performances[player.retrosheet_id][\"at_bats_last_5_games\"].shift\n player_performances[player.retrosheet_id][\"at_bats_last_10_games\"].shift\n player_performances[player.retrosheet_id][\"at_bats_last_20_games\"].shift\n\n player_performances[player.retrosheet_id][\"walks_last_1_game\"] << perf.walks\n player_performances[player.retrosheet_id][\"walks_last_2_games\"] << perf.walks\n player_performances[player.retrosheet_id][\"walks_last_5_games\"] << perf.walks\n player_performances[player.retrosheet_id][\"walks_last_10_games\"] << perf.walks\n player_performances[player.retrosheet_id][\"walks_last_20_games\"] << perf.walks\n \n player_performances[player.retrosheet_id][\"walks_last_1_game\"].shift\n player_performances[player.retrosheet_id][\"walks_last_2_games\"].shift\n player_performances[player.retrosheet_id][\"walks_last_5_games\"].shift\n player_performances[player.retrosheet_id][\"walks_last_10_games\"].shift\n player_performances[player.retrosheet_id][\"walks_last_20_games\"].shift\n\n player_performances[player.retrosheet_id][\"hits_last_1_game\"] << perf.hits\n player_performances[player.retrosheet_id][\"hits_last_2_games\"] << perf.hits\n player_performances[player.retrosheet_id][\"hits_last_5_games\"] << perf.hits\n player_performances[player.retrosheet_id][\"hits_last_10_games\"] << perf.hits\n player_performances[player.retrosheet_id][\"hits_last_20_games\"] << perf.hits\n \n player_performances[player.retrosheet_id][\"hits_last_1_game\"].shift\n player_performances[player.retrosheet_id][\"hits_last_2_games\"].shift\n player_performances[player.retrosheet_id][\"hits_last_5_games\"].shift\n player_performances[player.retrosheet_id][\"hits_last_10_games\"].shift\n player_performances[player.retrosheet_id][\"hits_last_20_games\"].shift\n\n player_performances[player.retrosheet_id][\"strikeouts_last_1_game\"] << perf.strikeouts\n player_performances[player.retrosheet_id][\"strikeouts_last_2_games\"] << perf.strikeouts\n player_performances[player.retrosheet_id][\"strikeouts_last_5_games\"] << perf.strikeouts\n player_performances[player.retrosheet_id][\"strikeouts_last_10_games\"] << perf.strikeouts\n player_performances[player.retrosheet_id][\"strikeouts_last_20_games\"] << perf.strikeouts\n \n player_performances[player.retrosheet_id][\"strikeouts_last_1_game\"].shift\n player_performances[player.retrosheet_id][\"strikeouts_last_2_games\"].shift\n player_performances[player.retrosheet_id][\"strikeouts_last_5_games\"].shift\n player_performances[player.retrosheet_id][\"strikeouts_last_10_games\"].shift\n player_performances[player.retrosheet_id][\"strikeouts_last_20_games\"].shift\n\n player_performances[player.retrosheet_id][\"total_bases_last_1_game\"] << perf.total_bases\n player_performances[player.retrosheet_id][\"total_bases_last_2_games\"] << perf.total_bases\n player_performances[player.retrosheet_id][\"total_bases_last_5_games\"] << perf.total_bases\n player_performances[player.retrosheet_id][\"total_bases_last_10_games\"] << perf.total_bases\n player_performances[player.retrosheet_id][\"total_bases_last_20_games\"] << perf.total_bases\n \n player_performances[player.retrosheet_id][\"total_bases_last_1_game\"].shift\n player_performances[player.retrosheet_id][\"total_bases_last_2_games\"].shift\n player_performances[player.retrosheet_id][\"total_bases_last_5_games\"].shift\n player_performances[player.retrosheet_id][\"total_bases_last_10_games\"].shift\n player_performances[player.retrosheet_id][\"total_bases_last_20_games\"].shift\n end\n end\nend",
"def merge_counties\n all_counties = County.distinct(:chapman_code)\n all_counties\n end",
"def to_h\n features.group_by(&:postcode).map do |postcode, objects|\n Hash[\n postcode => objects.map do |object|\n {\n 'name' => object.name\n }\n end\n ]\n end.reduce(:merge)\n end",
"def misc_features(code)\n \tanswer = Array.new\n if code.nil?\n self.seq_region.misc_features.each do |mf|\n if mf.seq_region_start > self.start and mf.seq_region_end < self.stop\n answer.push(mf)\n end\n end\n else\n self.seq_region.misc_features.each do |mf|\n if mf.misc_sets[0].code == code\n if mf.seq_region_start > self.start and mf.seq_region_end < self.stop\n answer.push(mf)\n end\n end\n end\n end\n \treturn answer\n end",
"def append_features(mod) end",
"def merge_duplicates_based_on_case\n POSSIBLY_WRONGLY_CASED_ATTRIBUTES.each do |attr|\n merged_runners = 0\n find_runners_only_differing_in(attr, [\"f_unaccent(lower(#{attr})) as low\"], ['low']).each do |entries|\n # We prefer the version with capital first letter and more lowercase\n # characters. E. g. for\n # Reichenbach I. K.\n # reichenbach i. K.\n # Reichenbach i. K.\n # the version at the bottom is preferred.\n merged_runners += reduce_to_one_runner_by_condition(entries) do |runner|\n [runner[attr][0] == runner[attr][0].upcase ? 1 : 0,\n runner[attr].scan(/[[:lower:]]/).size]\n end\n end\n puts \"Merged #{merged_runners} entries based on case of #{attr}.\" unless Rails.env.test?\n end\n end",
"def overlapping_token_support\n attributes.fetch(:overlappingTokenSupport)\n end",
"def assigned_features=(array_hash)\n # Find new features (but no duplicates)\n self.added_features = []\n array_hash.each do |h|\n unless features.detect { |c| c.id.to_s == h[:id] } || self.added_features.detect { |f| f.id.to_s == h[:id] }\n c = !h[:id].blank? ? Feature.find(h[:id]) : Feature.new({:page_object => self})\n c.attributes = h.reject { |k,v| k == :id } # input values, but don't try to overwrite the id\n self.added_features << c unless c.nil?\n end\n end\n # Delete removed features\n features.each do |c|\n if h = array_hash.detect { |h| h[:id] == c.id.to_s }\n c.attributes = h.reject { |k,v| k == :id }\n else\n c.destroy_association = 1\n end\n end\n end",
"def combine_regions_for_championship(params_logic)\n ff_headers = Team.region_names_final_four\n champ_header = Team.region_names_championship.first\n params_logic[champ_header] = {\"0\" => params_logic[ff_headers.first][\"0\"].merge(params_logic[ff_headers.last][\"0\"])}\n ff_headers.each{|header| params_logic.delete(header)}\n params_logic\n end",
"def discretize_by_ChiMerge!(alpha=0.10)\n # degree of freedom equals one less than number of classes\n df = get_classes.size-1\n chisq = pval2chisq(alpha, df)\n \n # for intialization\n hzero = {}\n each_class do |k|\n hzero[k] = 0.0\n end\n \n # determine the final boundaries for each feature\n f2bs = {}\n each_feature do |f|\n #f = :\"sepal-length\"\n # 1a. initialize boundaries\n bs, cs, qs = [], [], []\n fvs = get_feature_values(f).uniq.sort\n fvs.each do |v|\n bs << v\n cs << hzero.dup\n end\n \n # 1b. initialize counts for each interval\n each_sample do |k, s|\n next if not s.has_key? f\n i = bs.rindex { |x| s[f] >= x }\n cs[i][k] += 1.0\n end\n \n # 1c. initialize chi-squared values between two adjacent intervals\n cs.each_with_index do |c, i|\n if i+1 < cs.size\n qs << chisq_calc(c, cs[i+1])\n end\n end\n \n # 2. iteratively merge intervals\n until qs.empty? or qs.min > chisq\n qs.each_with_index do |q, i|\n next if q != qs.min\n \n # update cs for merged two intervals\n cm = {}\n each_class do |k|\n cm[k] = cs[i][k]+cs[i+1][k]\n end\n \n # update qs if necessary\n # before merged intervals\n if i-1 >= 0\n qs[i-1] = chisq_calc(cs[i-1], cm)\n end\n # after merged intervals\n if i+1 < qs.size\n qs[i+1] = chisq_calc(cm, cs[i+2])\n end\n \n # merge up\n bs.delete_at(i+1)\n cs.delete_at(i);cs.delete_at(i);cs.insert(i, cm)\n qs.delete_at(i)\n \n # break out\n break\n end\n end\n \n # 3. record the final boundaries\n f2bs[f] = bs\n end\n \n # discretize according to each feature's boundaries\n discretize_at_cutpoints!(f2bs)\n end",
"def common_tags_to_add\n common_tags = []\n common_tags << self.merger\n common_tags << self if self.canonical\n common_tags << self.parents\n common_tags.flatten.uniq.compact\n end",
"def create_features_ensembl_seq_obj(bioseq,targets)\r\n targets.each do |key,value|\r\n f1 = Bio::Feature.new(\"target_CTTCTT\",\"#{key[0]}..#{key[1]}\")\r\n #im no sure if it is a interior coding exon \r\n f1.append(Bio::Feature::Qualifier.new('interior coding exon', \"#{value[0]}\"))\r\n f1.append(Bio::Feature::Qualifier.new('strand', \"#{value[1]}\"))\r\n bioseq.features << f1\r\n end\r\nend",
"def service_learners\n service_learning_positions.collect(&:placements).flatten.collect(&:person).compact.uniq\n end",
"def merge; end",
"def misc_features(code)\r\n \tanswer = Array.new\r\n if code.nil?\r\n self.seq_region.misc_features.each do |mf|\r\n if mf.seq_region_start > self.start and mf.seq_region_end < self.stop\r\n answer.push(mf)\r\n end\r\n end\r\n else\r\n self.seq_region.misc_features.each do |mf|\r\n if mf.misc_sets[0].code == code\r\n if mf.seq_region_start > self.start and mf.seq_region_end < self.stop\r\n answer.push(mf)\r\n end\r\n end\r\n end\r\n end\r\n \treturn answer\r\n end",
"def with_all(*features)\n pos, neg = mangle_args(*features)\n self.class.new(Hash[@sets.select do |key, val|\n pos.subset?(key)\n end]).without_any(neg)\n end",
"def get_features()#:nodoc:\n # These features are more or less accurate. Many values couldn't be verified because\n # they are used in context of Java programming language a sometimes Nokogiri just\n # dont't provide any information on the feature\n\n features = {\n \"external-general-entities\" => false,\n \"external-parameter-entities\" => false,\n \"is-standalone\" => false, #Zjistit jak ziskat z dokumentu standalone\n \"lexical-handler/parameter-entities\" => false,\n \"namespaces\" => true,\n \"namespace-prefixes\" => true,\n \"resolve-dtd-uris\" => false,\n \"string-interning\" => false,\n \"unicode-normalization-checking\" => true, #Nokogiri probably supports this, needs to be verified\n \"use-attributes2\" => false,\n \"use-locator2\" => false,\n \"use-entity-resolver2\" => false,\n \"validation\" => true, #Nokogiri is probably reporting erros\n \"xmlns-uris\" => true, #Nokogiri is probably treating xmlns declarations as part of xmlns namespace\n \"xml-1.1\" => false #Don't know\n }\n\n return features\n end",
"def cluster(cluster)\n features = {\n 'vsan' => true,\n 'enableDrs' => true,\n 'enableHA' => true,\n }\n\n return cluster.merge(features)\nend",
"def intersect(array1, array2)\n hash_table = {}\n mergedArray = []\n\n array1.each do |a1|\n hash_table[a1] = true\n end\n\n array2.each do |a2|\n mergedArray << a2 if hash_table[a2]\n end\n\n return mergedArray\nend",
"def assignable_services\n\t\t(Service.predefined + services).sort_by(&:lower_case_name).uniq\n\tend",
"def addFeaturesAndLabel(home_team, away_team, earliest_date, latest_date)\n home_faceoffs = Game.where(\"home_team = ? and away_team = ? and game_date > ? and game_date <= ?\", home_team, away_team, earliest_date, latest_date).order(\"game_date desc\")\n away_faceoffs = Game.where(\"home_team = ? and away_team = ? and game_date > ? and game_date <= ?\", away_team, home_team, earliest_date, latest_date).order(\"game_date desc\")\n past_games = home_faceoffs.concat(away_faceoffs)\n past_games = past_games.sort {|game1, game2| game1.game_date <=> game2.game_date }\n\n if past_games.size < 3\n return\n end\n\n run_differentials = [0, 0, 0]\n\n past_games.each_with_index do |past_game, index|\n feature_set = Feature.new\n\n feature_set.game_id = past_game.id\n feature_set.home_team_won = past_game.home_team_won\n\n if past_game.home_team == home_team.to_s\n feature_set.h2h_diff_1 = run_differentials[2]\n feature_set.h2h_diff_2 = run_differentials[1]\n feature_set.h2h_diff_3 = run_differentials[0]\n\n run_differentials << past_game.home_team_runs - past_game.away_team_runs\n else \n feature_set.h2h_diff_1 = -1*run_differentials[2]\n feature_set.h2h_diff_2 = -1*run_differentials[1]\n feature_set.h2h_diff_3 = -1*run_differentials[0]\n\n run_differentials << past_game.away_team_runs - past_game.home_team_runs \n end\n\n feature_set.save\n\n if run_differentials.size > 3\n run_differentials.shift\n end\n end\nend",
"def cleanup\n datas[:model_feature_values].each_pair { |model_key, feature_values| \n model = Model.find(model_key)\n feature_keys = []\n model.features_list { |feature| feature_keys << feature.key }\n values(model).delete_if { |feature_key, value| !feature_keys.include?(feature_key) }\n }\n end",
"def copy(cleanse = true)\n a = dup\n a.cleanse if cleanse\n vehicle_features.each do |x|\n a.vehicle_features << x\n end\n fta_service_types.each do |x|\n a.fta_service_types << x\n end\n fta_mode_types.each do |x|\n a.fta_mode_types << x\n end\n a\n end",
"def copy(cleanse = true)\n a = dup\n a.cleanse if cleanse\n vehicle_features.each do |x|\n a.vehicle_features << x\n end\n fta_service_types.each do |x|\n a.fta_service_types << x\n end\n fta_mode_types.each do |x|\n a.fta_mode_types << x\n end\n a\n end",
"def test_get_feature_vect()\n rs = ClinicalTCGA::RetrieveSamples.new([\"TCGA-A6-2671\",\"TCGA-A6-2672\"], \n [\"vital_status\", \"death_days_to\",\"percent_tumor_nuclei\"], \n false)\n followup = \"#{ENV['TCGA_CLINICAL_TEST_DATA']}/Biotab/nationwidechildrens.org_clinical_follow_up_v1.0_coad.txt\"\n tumor_sample = \"#{ENV['TCGA_CLINICAL_TEST_DATA']}/Biotab/nationwidechildrens.org_biospecimen_slide_coad.txt\"\n rs.add_tcga_source(followup)\n rs.add_tcga_source(tumor_sample)\n \n h = Hash.new\n rs.get_feature_vector do |sample,fV|\n h[sample] = fV\n end\n\n assert_equal(h[\"TCGA-A6-2671\"], [\"Dead\", \"1331\", 35.0], \"returned unexpected result for TCGA-A6-2671\")\n assert_equal(h[\"TCGA-A6-2672\"], [\"Alive\", \"[Not Available]\", 40.0], \"returned unexpected result for TCGA-A6-2672\")\n end",
"def update!(**args)\n @concierge_features = args[:concierge_features] if args.key?(:concierge_features)\n end",
"def copy_features(options={})\r\n hash = {}\r\n @current_object.class.ecore.eAllStructuralFeatures.each do |f|\r\n next if f.derived\r\n next if options[:except] && options[:except].include?(f.name.to_sym)\r\n hash[f.name.to_sym] = trans(@current_object.send(f.name))\r\n end\r\n hash.merge!(yield) if block_given?\r\n hash\r\n end",
"def merge_attributes\n attrs = self.attributes.dup.reject{ |k,v| ignored_merge_attributes.include?(k) }\n attrs.merge!(address_attributes) # we want addresses to be shown in the UI\n sorted = attrs.sort do |a,b|\n (ordered_merge_attributes.index(a.first) || 1000) <=> (ordered_merge_attributes.index(b.first) || 1000)\n end\n sorted.inject({}) do |h, item|\n h[item.first] = item.second\n h\n end\n end",
"def features_set(code)\r\n features(code).inject([]) {|r, ft| r |= [ft.data_id] }\r\n end",
"def build_accordions_and_trees_only\n # Build the Explorer screen from scratch\n allowed_features = ApplicationController::Feature.allowed_features(features)\n @trees = allowed_features.collect { |feature| feature.build_tree(@sb) }\n @accords = allowed_features.map(&:accord_hash)\n\n allowed_features\n end",
"def feature_flags_with_defaults\n flag_names = FeatureFlag.pluck(:name).sort\n FeatureFlag.default_flag_hash.merge(feature_flags_for(*flag_names)).with_indifferent_access\n end",
"def build!\n set1.each do | target , options |\n candidate =\n Candidate.new \\\n target,\n *( options.first.is_a?( Array ) ? options : [ options ] )\n\n candidates.push candidate\n candidate_set1[ target ] = candidate\n end\n\n set2.each do | target , options |\n candidate =\n Candidate.new \\\n target,\n *( options.first.is_a?( Array ) ? options : [ options ] )\n\n candidates.push candidate\n candidate_set2[ target ] = candidate\n end\n\n candidate_set1.each do | target , candidate |\n candidate.preferences =\n candidate.raw_preferences.map { | preference_target | candidate_set2[ preference_target ] }\n end\n\n candidate_set2.each do | target , candidate |\n candidate.preferences =\n candidate.raw_preferences.map { | preference_target | candidate_set1[ preference_target ] }\n end\n\n # We've built the candidates\n self.built = true\n end",
"def update!(**args)\n @anchor_common_feature_set = args[:anchor_common_feature_set] if args.key?(:anchor_common_feature_set)\n @attachments = args[:attachments] if args.key?(:attachments)\n @babel_checkpoint_path = args[:babel_checkpoint_path] if args.key?(:babel_checkpoint_path)\n @caption_entity_anchor_features = args[:caption_entity_anchor_features] if args.key?(:caption_entity_anchor_features)\n @caption_span_anchor_features = args[:caption_span_anchor_features] if args.key?(:caption_span_anchor_features)\n @description_anchor_features = args[:description_anchor_features] if args.key?(:description_anchor_features)\n @filter_reason = args[:filter_reason] if args.key?(:filter_reason)\n @filtered = args[:filtered] if args.key?(:filtered)\n @generative_features = args[:generative_features] if args.key?(:generative_features)\n @instruction_anchor_features = args[:instruction_anchor_features] if args.key?(:instruction_anchor_features)\n @instruction_training_data_anchor_features = args[:instruction_training_data_anchor_features] if args.key?(:instruction_training_data_anchor_features)\n @label_language = args[:label_language] if args.key?(:label_language)\n @label_transformation = args[:label_transformation] if args.key?(:label_transformation)\n @list_anchor_features = args[:list_anchor_features] if args.key?(:list_anchor_features)\n @list_training_data_anchor_features = args[:list_training_data_anchor_features] if args.key?(:list_training_data_anchor_features)\n @multimodal_topic_features = args[:multimodal_topic_features] if args.key?(:multimodal_topic_features)\n @multimodal_topic_training_features = args[:multimodal_topic_training_features] if args.key?(:multimodal_topic_training_features)\n @normalized_babel_embedding = args[:normalized_babel_embedding] if args.key?(:normalized_babel_embedding)\n @ocr_anchor_feature = args[:ocr_anchor_feature] if args.key?(:ocr_anchor_feature)\n @ocr_description_training_data_anchor_features = args[:ocr_description_training_data_anchor_features] if args.key?(:ocr_description_training_data_anchor_features)\n @opinions_anchor_features = args[:opinions_anchor_features] if args.key?(:opinions_anchor_features)\n @qna_anchor_features = args[:qna_anchor_features] if args.key?(:qna_anchor_features)\n @rating_score = args[:rating_score] if args.key?(:rating_score)\n @safe_search_classifier_output = args[:safe_search_classifier_output] if args.key?(:safe_search_classifier_output)\n @text_similarity_features = args[:text_similarity_features] if args.key?(:text_similarity_features)\n @thumbnail_info = args[:thumbnail_info] if args.key?(:thumbnail_info)\n end",
"def features_set(code)\n features(code).inject([]) {|r, ft| r |= [ft.data_id] }\n end",
"def features\n s = Set.new\n @plugins.each { |x| s += x.features }\n s\n end",
"def test_get_larger_feature_vect()\n feature_v = [\"percent_normal_cells\", \n \"percent_stromal_cells\", \n \"percent_tumor_cells\",\n \"percent_lymphocyte_infiltration\",\n \"vital_status\",\n \"death_days_to\",\n \"last_contact_days_to\", \n \"tumor_status\",\n \"ajcc_tumor_pathologic_pt\",\n \"ajcc_nodes_pathologic_pn\", \n \"ajcc_metastasis_pathologic_pm\", \n \"ajcc_pathologic_tumor_stage\"\n ]\n \n rs = ClinicalTCGA::RetrieveSamples.new([\"TCGA-QG-A5YW\", \"TCGA-A6-2676\"],\n feature_v,\n false)\n rs.add_all_sources(\"#{ENV['TCGA_CLINICAL_TEST_DATA']}/Biotab/\", false) # suppress progress bar for unit test\n \n h = Hash.new\n rs.get_feature_vector{|sample,fV| h[sample] = fV}\n \n assert_equal(h[\"TCGA-QG-A5YW\"][0], 10.0, \"not expected value at index 0\")\n assert_equal(h[\"TCGA-A6-2676\"][0], 2.5, \"not expected value at index 0\")\n assert_equal(h[\"TCGA-QG-A5YW\"][-2],nil, \"not expected value at index -2\")\n assert_equal(h[\"TCGA-A6-2676\"][-2],\"M0\", \"not expected value at index -2\")\n end",
"def without(*features)\n features = setify(*features)\n self.class.new Hash[@sets.select {|key, val| !features.subset?(key)}]\n end",
"def discretize_by_TID!\n # cut points for each feature\n f2cp = {}\n \n each_feature do |f|\n cv = get_class_labels\n fv = get_feature_values(f)\n \n n = cv.size\n abort \"[#{__FILE__}@#{__LINE__}]: \\n\"+\n \" missing feature value is not allowed!\" if n != fv.size\n \n # sort cv and fv according to ascending order of fv\n sis = (0...n).to_a.sort { |i,j| fv[i] <=> fv[j] }\n cv = cv.values_at(*sis)\n fv = fv.values_at(*sis)\n \n # get initial boundaries\n bs = []\n fv_u = fv.uniq\n fv_u.each_with_index do |v, i|\n # cut points are the mean of two adjacent data points\n if i < fv_u.size-1\n bs << (v+fv_u[i+1])/2.0\n end\n end\n \n # test each pair cut point\n s_best, h1_best, h2_best = nil, nil, nil\n \n bs.each_with_index do |h1, i| \n bs.each_with_index do |h2, j|\n next if j <= i\n \n n_h1 = (0...n).to_a.select { |x| fv[x] < h1 }.size.to_f\n n_h1_h2 = (0...n).to_a.select { |x| fv[x] > h1 and fv[x] < h2 }.size.to_f\n n_h2 = (0...n).to_a.select { |x| fv[x] > h2 }.size.to_f\n \n s = 0.0\n \n each_class do |k|\n n_h1_k = (0...n).to_a.select { |x| fv[x] < h1 and cv[x] == k }.size.to_f\n n_h1_h2_k = (0...n).to_a.select { |x| fv[x] > h1 and fv[x] < h2 and cv[x] == k }.size.to_f\n n_h2_k = (0...n).to_a.select { |x| fv[x] > h2 and cv[x] == k }.size.to_f\n \n s += n_h1_k * Math.log2(n_h1_k/n_h1) if not n_h1_k.zero?\n s += n_h1_h2_k * Math.log2(n_h1_h2_k/n_h1_h2) if not n_h1_h2_k.zero?\n s += n_h2_k * Math.log2(n_h2_k/n_h2) if not n_h2_k.zero?\n \n #pp [s_best, s, h1, h2] + [n_h1, n_h1_k] + [n_h1_h2, n_h1_h2_k] + [n_h2, n_h2_k]\n end\n \n if not s_best or s > s_best\n s_best, h1_best, h2_best = s, h1, h2\n #pp [s_best, h1_best, h2_best]\n end\n \n break if s_best.zero? # allow early temination at maximum value 0.0\n end\n \n break if s_best.zero? # allow early temination at maximum value 0.0\n end\n \n #pp [s_best, h1_best, h2_best]\n f2cp[f] = [h1_best, h2_best]\n end\n \n # discretize based on cut points\n discretize_at_cutpoints!(f2cp, true)\n end",
"def define_features\n @fvs.each do |vector, label|\n vector.each do |term, value|\n @features.add(term)\n end\n end\n end",
"def merge_power_distribition\n # method to be developed for any attributes to be aggregated or merged\n end",
"def _flag_sets_of_automation(automation_config)\n my_flags = automation_config.fetch(:flags, {})\n target_sets = automation_config.fetch(FLAG_SETS_KEY, [])\n if target_sets.is_a?(Array) && !target_sets.empty?\n # create hash of set_name => merge of that set INTO the explicit flags for this automation\n target_sets.map! { |set_name| [set_name, my_flags.merge(@config[FLAG_SETS_KEY][set_name.to_sym])] }.to_h\n else\n # no flag sets, so we just have one set of flags -- the explicit flags configured for this automation\n target_sets = { flags: my_flags }\n end\n target_sets\n end",
"def active_tags\n active_tags = [:global]\n active_tags << feature_tag.to_sym if feature_tag\n active_tags + super\n end",
"def update!(**args)\n @concierge_product_features = args[:concierge_product_features] if args.key?(:concierge_product_features)\n end",
"def supported_features\n\t\treturn self.supported_feature_oids.collect {|oid| FEATURE_NAMES[oid] || oid }\n\tend",
"def mastered_weapon_proficiencies\n self.weapon_categories.map(&:common_weapons).flatten.map(&:id) + self.common_weapons.map(&:id)\n end",
"def features\n @_features ||= Hash.new do |hash, key|\n class_feature = self.class.features[key]\n\n if class_feature\n hash[key] = class_feature.bind(self)\n else\n hash[key] = Feature.new(key, {}).bind(self).fake!\n end\n end\n end",
"def consolidate_classes(original_line, list_of_classes)\n record = {\n :original_ocr => original_line,\n :attributes_parsed => {\n :subject =>\n [\n #{:value => \"Curran Sarah\", :type => \"primary\", :occupation => \"widowed\"},\n #{:value => \"Richard\", :type => \"widower of primary\"}\n ],\n :location =>\n [\n #{:value => \"7 Sixth\", :position => \"rear\", :type => \"home\"}\n ]\n }\n }\n\n list_of_classes.each_with_index do |classed_token, index|\n parsed_class = classed_token[1][0]\n value = classed_token[0]\n if index == 0 && parsed_class == :name_component\n record[:attributes_parsed][:subject] << {:value => value, :type => 'primary'}\n end\n if index > 0\n case parsed_class\n when :job_component\n unless record[:attributes_parsed][:subject].count < 1\n record[:attributes_parsed][:subject][0][:occupation] = value\n end\n when :predicate\n case value\n when \"wid\"\n unless record[:attributes_parsed][:subject].count < 1\n record[:attributes_parsed][:subject][0][:occupation] = 'widow'\n end\n deceased_name = look_for_name_of_deceased(list_of_classes,index)\n unless deceased_name.nil?\n record[:attributes_parsed][:subject] << {:value => deceased_name, :type => 'deceased spouse of primary'}\n end\n #attach_to_next(list_of_classes, index, :name_component, [{:type => 'deceased spouse of primary'}])\n when \"h\"\n attach_to_next(list_of_classes, index, :address_component, [{:type => 'home'}])\n when \"r\"\n attach_to_next(list_of_classes, index, :address_component, [{:position => 'rear'}])\n else\n end\n ## inner case\n when :address_component\n loc = {:value => value}\n classed_token[2..-1].each do |xtra_attr| ## add in any additional attributes from predicates\n xtra_attr.each do |k, v|\n loc[k] = v\n end\n end\n unless merge_if_directly_subsequent_is_alike(list_of_classes, index, classed_token)\n record[:attributes_parsed][:location] << loc\n end\n else\n end\n end ## indices after 0\n end ## loop of classes\n\n return record\nend",
"def merge_duplicates_based_on_msm_prefix\n merged_runners = 0\n suffix_length = 4\n attribute = :club_or_hometown\n find_runners_only_differing_in(attribute, [\"lower(replace(#{attribute},'MSM - ', '')) as no_msm_prefix\"], ['no_msm_prefix']).each do |entries|\n # Pick the version including 'MSM' (is always the longer one).\n merged_runners += reduce_to_one_runner_by_condition(entries) do |runner|\n runner[:club_or_hometown].length\n end\n end\n puts \"Merged #{merged_runners} entries based on prefix MSM prefix\" unless Rails.env.test?\n end",
"def new_classifiers\n CLASSIFIERS.map { |classifier_name| [classifier_name, {}] }.to_h\n end",
"def feature_module\n unless defined?(@feature_module)\n @features ||= {}\n @feature_module = ::Module.new\n const_set(\"FeatureModule\", @feature_module)\n features = @features\n # Create a feature? method that can be passed a feature name and\n # determine if the feature is present.\n @feature_module.send(:define_method, :feature?) do |name|\n method = name.to_s + \"?\"\n return !!(respond_to?(method) and send(method))\n end\n\n # Create a method that will list all functional features.\n @feature_module.send(:define_method, :features) do\n return false unless defined?(features)\n features.keys.find_all { |n| feature?(n) }.sort { |a,b|\n a.to_s <=> b.to_s\n }\n end\n\n # Create a method that will determine if a provided list of\n # features are satisfied by the curred provider.\n @feature_module.send(:define_method, :satisfies?) do |*needed|\n ret = true\n needed.flatten.each do |feature|\n unless feature?(feature)\n ret = false\n break\n end\n end\n ret\n end\n\n # Create a boolean method for each feature so you can test them\n # individually as you might need.\n @features.each do |name, feature|\n method = name.to_s + \"?\"\n @feature_module.send(:define_method, method) do\n (is_a?(Class) ? declared_feature?(name) : self.class.declared_feature?(name)) or feature.available?(self)\n end\n end\n\n # Allow the provider to declare that it has a given feature.\n @feature_module.send(:define_method, :has_features) do |*names|\n @declared_features ||= []\n names.each do |name|\n @declared_features << name.intern\n end\n end\n # Aaah, grammatical correctness\n @feature_module.send(:alias_method, :has_feature, :has_features)\n end\n @feature_module\n end",
"def features_by_postcode(features)\n return unless features\n\n @features_by_postcode ||= features.group_by do |feature|\n feature['context'].find { |ctx| postcode?(ctx) }['text']\n end\n end",
"def autofill_kit\n return if kit || components.empty?\n kits = components.map {|c| c.kit }\n kit = kits.uniq\n end",
"def combine_regions_for_final_four(params_logic)\n ff_headers = Team.region_names_final_four\n Team.final_four_region_pairings.each.with_index do |region_pair, i|\n params_logic[ff_headers[i]] = {\"0\" => params_logic[region_pair.first][\"0\"].merge(params_logic[region_pair.last][\"0\"])}\n region_pair.each{|region_name| params_logic.delete(region_name)}\n end\n params_logic\n end",
"def insert_gaps_between( features )\n features_with_gaps = []\n gap_feature = AlleleImage::Feature.new( Bio::Feature.new( \"misc_feature\", \"1..1\" ).append( Bio::Feature::Qualifier.new( \"note\", \"gap\" ) ) )\n\n return features_with_gaps if features.nil?\n\n features.each_index do |current_index|\n features_with_gaps.push( features[current_index] )\n next_index = current_index + 1\n unless features[next_index].nil?\n consecutive_names = [ features[current_index].feature_name, features[next_index].feature_name ]\n consecutive_types = [ features[current_index].feature_type, features[next_index].feature_type ]\n if consecutive_names.include?(\"loxP\") ||\n consecutive_names.include?(\"FRT\") ||\n consecutive_names.include?(\"Rox\") ||\n consecutive_names.include?(\"F3\") ||\n consecutive_names.include?(\"AttP\") ||\n consecutive_names.include?(\"intervening sequence\") ||\n consecutive_types.include?(\"exon\")\n features_with_gaps.push( gap_feature )\n end\n end\n end\n\n return features_with_gaps\n end",
"def qualified_candidates (collection)\n match=[]\n \n collection.each do |x|\n if years_of_experience(x[:years_of_experience]) && github_points(x[:github_points]) && knowledge(x[:languages]) && applied_time(x[:date_applied]) && old_enough(x[:age])\n match << (x)\n end\n end\n\n match\nend",
"def features\n\t\tcollection = []\n\t\tclasses.each do |_class|\n\t\t\t_class.features.each do |feature|\n\t\t\t\tcollection << feature\n\t\t\tend\n\t\tend\n\t\trace.features.each do |feature|\n\t\t\tcollection << feature\n\t\tend\n\t\tfeats.each do |feat|\n\t\t\tcollection << feat\n\t\tend\n\t\tcollection\n\tend",
"def test_scenario1\n data = [\n {\"filename\" => File.dirname(__FILE__)+\"/data/predictions_c.json\", \n \"method\" => 0,\n \"prediction\" => \"a\",\n\t \"confidence\" => 0.450471270879},\n {\"filename\" => File.dirname(__FILE__)+\"/data/predictions_c.json\",\n \"method\" => 1,\n \"prediction\" => \"a\",\n \"confidence\" => 0.552021302649},\n {\"filename\" => File.dirname(__FILE__)+\"/data/predictions_c.json\",\n \"method\" => 2,\n \"prediction\" => \"a\",\n \"confidence\" => 0.403632421178},\n {\"filename\" => File.dirname(__FILE__)+\"/data/predictions_r.json\",\n \"method\" => 0,\n \"prediction\" => 1.55555556667, \n \"confidence\" => 0.400079152063},\n {\"filename\" => File.dirname(__FILE__)+\"/data/predictions_r.json\",\n \"method\" => 1,\n \"prediction\" => 1.59376845074,\n \"confidence\" => 0.248366474212},\n {\"filename\" => File.dirname(__FILE__)+\"/data/predictions_r.json\",\n \"method\" => 2,\n \"prediction\" => 1.55555556667,\n \"confidence\" => 0.400079152063}\n ]\n\n puts \"Scenario: Successfully computing predictions combinations\"\n data.each do |item|\n puts\n\n puts \"Given I create a MultiVote for the set of predictions in file <%s>\" % item[\"filename\"]\n multivote = BigML::MultiVote.new(JSON.parse(File.open(item[\"filename\"], \"rb\").read))\n\n puts \"When I compute the prediction with confidence using method <%s>\" % item[\"method\"]\n combined_results = multivote.combine(item[\"method\"], nil, true)\n\n puts \"And I compute the prediction without confidence using method <%s>\" % item[\"method\"] \n combined_results_no_confidence = multivote.combine(item[\"method\"])\n\n if multivote.is_regression() \n puts \"Then the combined prediction is <%s>\" % item[\"prediction\"]\n assert_equal(combined_results[\"prediction\"].round(6), item[\"prediction\"].round(6))\n puts \"And the combined prediction without confidence is <%s>\" % item[\"prediction\"]\n assert_equal(combined_results_no_confidence.round(6), item[\"prediction\"].round(6))\n else\n puts \"Then the combined prediction is <%s>\" % item[\"prediction\"]\n assert_equal(combined_results[\"prediction\"], item[\"prediction\"])\n puts \"And the combined prediction without confidence is <%s>\" % item[\"prediction\"]\n assert_equal(combined_results_no_confidence,item[\"prediction\"])\n end\n puts \"And the confidence for the combined prediction is %s \" % item[\"confidence\"]\n assert_equal(combined_results[\"confidence\"].round(5), item[\"confidence\"].round(5)) \n end\n\n end",
"def environment_features(environment)\n @environment_features ||= {}\n @environment_features[environment] ||= load_environment_features(environment)\n end",
"def move_features(to)\n cartridges.delete_if do |c|\n if c.tags.include?(:ci_builder) and not c.tags.include?(:web_framework)\n to.cartridges.select{ |d| d.tags.include?(:web_framework) }.each{ |d| d.builds_with(c, self) }.present?\n end\n end\n cartridges.delete_if{ |c| cartridges.any?{ |other| other != c && other.scales_with == c.name } }\n if self != to && cartridges.empty?\n to.gears.concat(gears)\n gears.clear\n end\n end",
"def expected_results_with_all_supplemental_codes\n # Since this is a CMS IG requirement, only do this for CVU+ or C3 tests\n return expected_results unless product.cvuplus? || product.c3_test?\n\n required_codes = { 'PAYER' => %w[1 2 6 349], 'SEX' => %w[M F], 'RACE' => %w[2106-3 2076-8 2054-5 2028-9 1002-5 2131-1],\n 'ETHNICITY' => %w[2135-2 2186-5] }.freeze\n new_hash = expected_results\n new_hash.each do |_measure_id, pop_set_hash|\n pop_set_hash.each do |_pop_set_id, pop_set|\n sup_data = pop_set['supplemental_data']\n %w[IPP DENOM NUMER NUMEX DENEX DENEXCEP MSRPOPL MSRPOPLEX].each do |pop_key|\n next unless pop_set[pop_key]\n\n sup_data[pop_key] = { 'RACE' => {}, 'ETHNICITY' => {}, 'SEX' => {}, 'PAYER' => {} } unless sup_data[pop_key]\n required_codes.each do |sup_data_type, codes|\n codes.each do |code|\n sup_data[pop_key][sup_data_type][code] = 0 unless sup_data[pop_key][sup_data_type][code]\n end\n end\n end\n end\n end\n new_hash\n end",
"def customization_pairs(product_customizations)\n pairs = product_customizations.map(&:customized_product_options).flatten.map do |m|\n [m.customizable_product_option.id, m.value.present? ? m.value : m.customization_image.to_s ]\n end\n\n Set.new pairs\n end",
"def single_features\r\n\tcase self.feature_groups.where(singles: true).first\r\n\twhen nil\r\n\t\tself.feature_groups.create(singles: true)\r\n\telse \r\n\t\tself.feature_groups.where(singles: true).first\r\n\tend\r\n end",
"def requires(include_pending=false)\n features = component_instances.map {|ci| ci.cartridge_name} #get_feature(ci.cartridge_name, ci.component_name)}\n\n if include_pending\n self.pending_op_groups.each do |op_group|\n case op_group.op_type\n when :add_features\n features += op_group[:args][\"features\"]\n when :remove_features\n features -= op_group[:args][\"features\"]\n end\n end\n end\n\n features || []\n end",
"def add_features(features, group_overrides=[], init_git_url=nil, user_env_vars=nil)\n ssl_endpoint = Rails.application.config.openshift[:ssl_endpoint]\n cart_name_map = {}\n\n features.each do |feature_name|\n cart = CartridgeCache.find_cartridge(feature_name, self)\n\n # Make sure this is a valid cartridge\n if cart.nil?\n raise OpenShift::UserException.new(\"Invalid cartridge '#{feature_name}' specified.\", 109)\n end\n\n # ensure that the user isn't trying to add multiple versions of the same cartridge\n if cart_name_map.has_key?(cart.original_name)\n raise OpenShift::UserException.new(\"#{cart.name} cannot co-exist with #{cart_name_map[cart.original_name]} in the same application\", 109)\n else\n cart_name_map[cart.original_name] = cart.name\n end\n\n if cart.is_web_framework?\n component_instances.each do |ci|\n if ci.is_web_framework?\n raise OpenShift::UserException.new(\"You can only have one framework cartridge in your application '#{name}'.\", 109)\n end\n end\n end\n\n # check if the requested feature is provided by any existing/embedded application cartridge\n component_instances.each do |ci|\n ci_cart = ci.get_cartridge\n if ci_cart.original_name == cart.original_name\n raise OpenShift::UserException.new(\"#{feature_name} cannot co-exist with cartridge #{ci.cartridge_name} in your application\", 109)\n end\n end\n\n if cart.is_web_framework? and defined?(cart.endpoints) and cart.endpoints.respond_to?(:each)\n cart_req_ssl_endpoint = false\n cart.endpoints.each do |endpoint|\n if endpoint.options and endpoint.options[\"ssl_to_gear\"]\n cart_req_ssl_endpoint = true\n end\n end\n if (((ssl_endpoint == \"deny\") and cart_req_ssl_endpoint ) or\n ((ssl_endpoint == \"force\") and not cart_req_ssl_endpoint))\n raise OpenShift::UserException.new(\"Invalid cartridge '#{feature_name}' conflicts with platform SSL_ENDPOINT setting.\", 109, \"cartridge\")\n end\n end\n\n # Validate that the features support scalable if necessary\n if self.scalable && !(cart.is_plugin? || cart.is_service?)\n if cart.is_web_framework?\n raise OpenShift::UserException.new(\"Scalable app cannot be of type '#{feature_name}'.\", 109)\n else\n raise OpenShift::UserException.new(\"#{feature_name} cannot be embedded in scalable app '#{name}'.\", 109)\n end\n end\n\n # prevent a proxy from being added to a non-scalable (single-gear) application\n if cart.is_web_proxy? and !self.scalable\n raise OpenShift::UserException.new(\"#{feature_name} cannot be added to existing applications. It is automatically added when you create a scaling application.\", 137)\n end\n\n if self.scalable and cart.is_web_framework?\n prof = cart.profile_for_feature(feature_name)\n cart_scalable = false\n prof.components.each do |component|\n next if component.scaling.min==1 and component.scaling.max==1\n cart_scalable = true\n end\n if !cart_scalable\n raise OpenShift::UserException.new(\"Scalable app cannot be of type '#{feature_name}'.\", 109)\n end\n end\n\n # Validate that this feature either does not have the domain_scope category\n # or if it does, then no other application within the domain has this feature already\n if cart.is_domain_scoped?\n begin\n if Application.where(domain_id: self.domain._id, \"component_instances.cartridge_name\" => cart.name).count() > 0\n raise OpenShift::UserException.new(\"An application with #{feature_name} already exists within the domain. You can only have a single application with #{feature_name} within a domain.\")\n end\n rescue Mongoid::Errors::DocumentNotFound\n #ignore\n end\n end\n end\n\n result_io = ResultIO.new\n Application.run_in_application_lock(self) do\n self.pending_op_groups.push PendingAppOpGroup.new(op_type: :add_features, args: {\"features\" => features, \"group_overrides\" => group_overrides, \"init_git_url\" => init_git_url,\n \"user_env_vars\" => user_env_vars}, user_agent: self.user_agent)\n self.run_jobs(result_io)\n end\n\n # adding this feature may have caused pending_ops to be created on the domain\n # for adding env vars and ssh keys\n # execute run_jobs on the domain to take care of those\n domain.reload\n domain.run_jobs\n result_io\n end",
"def append_features(other)\n if other.__check_include(self)\n other.__include_module(self)\n end\n self\n end",
"def merge_duplicates_based_on_hometown_suffix\n merged_runners = 0\n suffix_length = 4\n attribute = :club_or_hometown\n find_runners_only_differing_in(attribute, [\"lower(substring(#{attribute} from length(#{attribute}) - #{suffix_length})) as suffix_only_attr\"], ['suffix_only_attr']).each do |entries|\n # The longer club_or_hometown entry is assumed to be correct/contains more information.\n # If there is a version with all uppercase, it is disprioritized.\n merged_runners += reduce_to_one_runner_by_condition(entries) do |runner|\n if runner[:club_or_hometown].upcase == runner[:club_or_hometown]\n -1\n else\n runner[:club_or_hometown].length\n end\n end\n end\n puts \"Merged #{merged_runners} entries based on prefix of club or hometown\" unless Rails.env.test?\n end",
"def techniques_by_tactic(only_platform: /.*/)\n techniques_by_tactic = Hash.new {|h, k| h[k] = []}\n techniques.each do |technique|\n next unless !technique['x_mitre_platforms'].nil?\n next unless technique['x_mitre_platforms'].any? { |platform| platform.downcase.sub(\" \", \"-\") =~ only_platform }\n\n technique.fetch('kill_chain_phases', []).select { |phase| phase['kill_chain_name'] == 'mitre-attack' }.each do |tactic|\n techniques_by_tactic[tactic.fetch('phase_name')] << technique\n end\n end\n techniques_by_tactic\n end",
"def build_associations\n candidate_sheet || build_candidate_sheet\n baptismal_certificate || build_baptismal_certificate\n sponsor_covenant || build_sponsor_covenant\n sponsor_eligibility || build_sponsor_eligibility\n pick_confirmation_name || build_pick_confirmation_name\n christian_ministry || build_christian_ministry\n retreat_verification || build_retreat_verification\n true\n end",
"def compose_algorithm_list(supported, option, append_all_supported_algorithms = T.unsafe(nil)); end",
"def action_plus_set\r\n features(FEATURE_ACTION_PLUS).collect {|ft| ft.value }\r\n end",
"def intersection_hash(array1, array2)\n hash = Hash.new()\n intersection = []\n\n array1.each do |el|\n hash[el] = true\n end\n\n array2.uniq.each do |el|\n if hash[el]\n intersection << el\n end\n end\n\n intersection\nend",
"def use(*features); end",
"def random_service_hash\n\t\tbelt \t\t= ['white', 'green', 'blue', 'red', 'black'].sample\n\t\tlocation \t= ['Los Angeles', 'USC', 'Southern California', 'Silicon Valley', 'San Francisco', 'Boston', 'United States'].sample\n\t\tprice \t\t= [*10..200].sample\n\t\tdesc \t\t= [ { title: 'Cheap Guitar Lessons', category: 'Music', tag: 'Guitar', headline: 'Experience professional teaching guitar'},\n\t\t\t\t\t\t{ title: 'Piano Lessons!', category: 'Music', tag: 'Piano', headline: 'Can teach beginners to experienced'},\n\t\t\t\t\t\t{ title: 'Rock and Roll!', category: 'Music', tag: 'Guitar', headline: 'Teaching rock and rock style guitar'},\n\t\t\t\t\t\t{ title: 'Dog Walking All Day', category: 'Errands', tag: 'Dog Walking', headline: 'I love spending time with dogs!'},\n\t\t\t\t\t\t{ title: 'How to Ride A Bike', category: 'Sports/Fitness', tag:'Biking', headline: 'Teaching you how to ride a bike!'}, \n\t\t\t\t\t\t{ title: 'BUAD Tutor Session', category: 'Education', tag: 'Tutoring', headline: 'Helping students with BUAD 425' },\n\t\t\t\t\t\t{ title: 'Car Washing Service', category: 'Errands', tag: 'Car Wash', headline: \"We've been doing this for a long time!\"} ].sample\n\t\thash \t\t= { belt: belt, location: location, price: price, desc: desc }\n\n\t\thash\n\tend",
"def concat_pro_ser(product_service_map,pro2,ser2)\n temp_map = product_service_map\n all = Array.new\n size = pro2.size\n 0.upto(size-1){ |i| \n p = pro2[i]\n s = ser2[i]\n if temp_map.has_key?(p)\n temparray = temp_map[p]\n else\n temparray = Array.new\n end\n temparray.push(ser2[i])\n temp_map[p]=temparray.uniq\n }\n #puts \"\\n\"\n show_info(\"exclude_p/s:\\n#{temp_map}\")\n return temp_map\n end"
] |
[
"0.5694013",
"0.5694013",
"0.54719406",
"0.5449497",
"0.5445721",
"0.54392433",
"0.5367435",
"0.5349743",
"0.53452206",
"0.5340121",
"0.531569",
"0.5293815",
"0.52930105",
"0.5287015",
"0.52672917",
"0.52443814",
"0.52270395",
"0.52179515",
"0.5188723",
"0.5186548",
"0.5169435",
"0.5149645",
"0.5126853",
"0.5094202",
"0.50928116",
"0.5062883",
"0.5054197",
"0.50532424",
"0.5022456",
"0.50163746",
"0.5010977",
"0.4989347",
"0.49765316",
"0.49563885",
"0.49524447",
"0.49519593",
"0.49411488",
"0.4940748",
"0.49275893",
"0.49203107",
"0.49163",
"0.49109006",
"0.4909614",
"0.48891774",
"0.48888794",
"0.48794305",
"0.4871068",
"0.48631296",
"0.48610222",
"0.48610222",
"0.48474434",
"0.48305205",
"0.48251933",
"0.48195937",
"0.47924852",
"0.47828567",
"0.47798938",
"0.4771578",
"0.4765288",
"0.47632188",
"0.47582862",
"0.47553885",
"0.47518364",
"0.47415152",
"0.47412527",
"0.47388735",
"0.47331926",
"0.47281766",
"0.4728099",
"0.47228524",
"0.47176605",
"0.47148547",
"0.470638",
"0.47038206",
"0.4687733",
"0.46856695",
"0.46799698",
"0.46773696",
"0.46772608",
"0.46736565",
"0.4671386",
"0.4670086",
"0.4665352",
"0.4655488",
"0.4642317",
"0.46416274",
"0.46363673",
"0.46357626",
"0.46344298",
"0.46318072",
"0.46309087",
"0.4625881",
"0.46246123",
"0.46191013",
"0.4613214",
"0.46124342",
"0.46101868",
"0.46013474",
"0.46009862",
"0.46009478"
] |
0.6113703
|
0
|
method that takes 2 args and prints the sum
|
def sum_these_numbers(a,b)
puts a + b
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def print_sum(num1, num2)\n p num1 + num2\nend",
"def calc_sum_two(first, second)\n\n puts \"the sum of #{first} and #{second} is #{first + second}\"\n\nend",
"def sum_two(num1, num2)\n p \"#{num1} + #{num2} = #{num1 + num2}\"\nend",
"def sum(a, b)\n puts \"Sum total: #{a} + #{b} = #{10 + 20}\"\n\nend",
"def print_sum num1=0,num2=0\r\n\tputs num1 + num2\t\r\nend",
"def sum(num1, num2)\n puts \"Really? #{num1 + num2} foxes?!\"\nend",
"def total(num1, num2)\n puts num1 + num2\nend",
"def sum(a, b)\n puts \"The sum of #{a} + #{b} = \" + \"#{(a + b)}\"\n #return a + b not needed for this\nend",
"def sum_numbers(num_1,num_2)\n puts \"The sum of #{num_1} and #{num_2} is #{num_1 + num_2}\"\nend",
"def sum(x, y)\n puts \"The sum of #{x} + #{y} is:\"\n puts \"-> #{x+y}\"\nend",
"def add(num_one,num_two)\n puts \"#{num_one} + #{num_two} = #{num_one + num_two}\"\nend",
"def sum_method(a, b)\n puts \"Adding #{a} + #{b}\"\n puts a + b\nend",
"def sum_these_numbers(num1,num2)\n p num1 + num2\nend",
"def add_sum(a, b)\n puts \"ADDING #{a} + #{b}\"\n return a + b\nend",
"def get_sum(num1, num2)\n puts num1 + num2\nend",
"def sum (a, b)\n puts a + b\nend",
"def sum(num1, num2)\n puts num1 + num2\nend",
"def sum(number_A, number_B)\n p number_A + number_B\nend",
"def sum_two_num(a, b)\n puts a + b\nend",
"def sum(a,b)\n p a + b\nend",
"def sum2(n1, n2)\n puts n1+n2\nend",
"def sum(a, b)\n p a + b\nend",
"def sum(a, b)\n p a + b\nend",
"def sum_these_numbers (a, b)\n puts a + b\nend",
"def sum_nums(num1, num2)\n puts num1 + num2\nend",
"def sum x,y \n puts x+y \nend",
"def sum_these_numbers(a, b)\n puts a + b\nend",
"def add(a, b)\n puts \"SUM #{a} + #{b}\"\n return a + b\nend",
"def sum_finder(num_1, num_2)\n p num_1 + num_2\nend",
"def find_sum(num1, num2)\n puts num1 + num2\nend",
"def sum(x, y)\n puts x + y\nend",
"def sum(x, y) # 2 arguments\n p \"The sum of #{x} and #{y} is #{x + y}\"\n # p math.sum(sum_numbers) # This will not work because the arguments are not in an array\nend",
"def print_sums(nums1, nums2)\n nums1.each do |num1|\n nums2.each do |num2|\n p num1 + num2\n end\n end\nend",
"def add(num1, num2)\n total = num1 + num2\n puts \"The sum of #{num1} + #{num2} = \" + total.to_s\nend",
"def numbers(num1, num2)\n puts \"#{num1} + #{num2}\"\n return num1 + num2\nend",
"def total(num1, num2)\n num1 + num2\nend",
"def summator(x, y)\n p x + y\nend",
"def maths2(firstNum, secondNum)\n total = firstNum + secondNum\n puts \"firstNum is #{firstNum}\"\n puts \"the secondnum is #{secondNum}\"\n puts \"and the total is #{total} \"\n puts firstNum\n puts secondNum\n puts total\n puts \"firstNum is firstNum\"\n puts \"the secondnum is secondNum\"\n puts \"and the total is total\"\nend",
"def add_two_numbers (x,y)\n puts x + y\nend",
"def add(a,b)\n puts \"Adding these numbers #{a}, #{b}\"\n a + b\n end",
"def two_sum(nums)\nend",
"def add_two_numbers(x,y)\n puts x + y\nend",
"def add_two_numbers( x, y )\n puts x + y\nend",
"def sum_these_numbers(x,y)\n sum = x + y\n puts sum\nend",
"def add(num_1, num_2)\n sum = num_1 + num_2\n puts sum\n return sum\nend",
"def sum num1, num2\n\ttotal = num1 + num2\n\treturn total\nend",
"def two_numbers(one, two)\n p \"This is the sum of #{one} and #{two}\"\nend",
"def sum(num1, num2)\n num1 * num2\nend",
"def sum_num(a, b)\n ab = a + b\n puts \"#{ab}\"\nend",
"def addition(value1, value2)\n result = value1 + value2\n puts \"The sum of #{ value1 } + #{ value2 } = #{ result }\"\nend",
"def add(x, y)\n\tputs \"#{ x } + #{ y } = #{ x + y }\"\nend",
"def simp_add(arg1, arg2)\n \n ans=arg1+arg2\n \n puts \"the answer you are looking for is: #{ans}\"\n \nend",
"def numbers(num1, num2)\n p num1 + num2\nend",
"def add(x, y)\n puts \"The sum of x & y is: #{x} + #{y}\"\n return x + y\nend",
"def sum(n1, n2)\n n1 + n2\nend",
"def add(a,b)\n puts \"Adding #{a} and #{b}:\"\n return a + b\nend",
"def add(a, b)\n puts \"addng #{a} + #{b}\"\n return a + b\nend",
"def add(num_1, num_2)\n p num_1 + num_2\nend",
"def add(num1,num2)\n\tputs \"num1 has the value : #{ num1 }\"\n\tputs \"num2 has the value : #{ num2 }\"\n\tresults = num1 + num2\n\tputs \"results is then: #{ results}\"\nend",
"def sum(x,y)\r\n\t x + y\r\n\tend",
"def add(x,y)\n total = x + y\n puts total\n calculate(total)\nend",
"def sum(number1, number2)\ntotal = number1 + number2 \nend",
"def numbers(a, b)\n p \"#{a} + #{b}\"\n return a + b\nend",
"def add(a, b)\n\t# a method that adds two numbers together\n\tputs \"ADDING #{a} + #{b}\"\n\treturn a + b\nend",
"def total(a,b)\r\n return (a+b)\r\nend",
"def add_nums(num1, num2)\n p \"Addition of #{num1} + #{num2} = #{num1 + num2}\"\nend",
"def add(a, b)\n puts \"adding #{a} and #{b}:\"\n return a + b\nend",
"def sum (a, b)\n a + b\nend",
"def addition(a,b)\n puts a + b\nend",
"def sum(num1, num2)\n return num1 + num2\nend",
"def addition(a,b)\n puts c = a + b\n end",
"def sum a, b\n a + b\nend",
"def add(num_1, num_2)\n puts num_1 + num_2\nend",
"def add(a,b)\n puts \"Adding #{a} + #{b}\"\n a + b\nend",
"def add(a,b)\n p (a + b)\nend",
"def add_nums(num1,num2)\n p num1 + num2\nend",
"def addition(a, b)\n puts a + b\nend",
"def add_two(num1, num2)\n\tnumber = num1 + num2\n\tputs number\nend",
"def add (arg1, arg2)\n puts arg1 + arg2\nend",
"def add(number_1, number_2)\n p number_1 + number_2\nend",
"def sum a, b\n return a + b\nend",
"def sum(a,b)\n a + b\nend",
"def add(a, b)\n puts \"Add #{a} + #{b}\"\n return a + b\nend",
"def sum_two_numbers(a, b)\n a + b\nend",
"def addition num1,num2 #num1 and num2 are known as parameters\nputs num1 + num2 \nend",
"def add (a, b)\n puts \"Adding #{a} + #{b}\"\n return a+b\nend",
"def add_twos(num1, num2)\n number = (num1 + num2)\n puts number\nend",
"def somma(a, b)\n puts \"SOMMANDO #{a} + #{b}\"\n a + b\nend",
"def add(a, b)\n puts \"ADDING #{a} + #{b}\"\n a + b\nend",
"def addition(x, y)\n puts \"Addition: #{x} + #{y}\"\n return x + y\nend",
"def sum(a,b)\n a + b \nend",
"def add(a, b)\n puts \"#{a} + #{b}\"\n return a + b\nend",
"def add(a, b)\n puts \"#{a} + #{b}\"\n return a + b\nend",
"def add(a, b)\n puts \"#{a} + #{b}\"\n return a + b\nend",
"def sum(a, b)\n a + b\nend",
"def sum(a, b)\n a + b\nend",
"def addition(num1, num2)\n p num1 + num2\nend",
"def addition_method(num1, num2)\n\tputs num1 + num2\nend",
"def add_numbers(num1, num2)\n p num1 + num2\nend",
"def add(x, y)\n puts \"Adding #{x}, #{y}\"\n return x + y\nend"
] |
[
"0.8242617",
"0.8135315",
"0.81145835",
"0.7973638",
"0.79689616",
"0.79239273",
"0.7917015",
"0.7899012",
"0.78814495",
"0.78180134",
"0.7816875",
"0.7777289",
"0.7758747",
"0.7752694",
"0.77397704",
"0.77377224",
"0.77055234",
"0.7689922",
"0.7674478",
"0.7660498",
"0.7638418",
"0.7606267",
"0.7606267",
"0.75983644",
"0.7585424",
"0.7546624",
"0.75439906",
"0.752594",
"0.7519623",
"0.75078213",
"0.74812984",
"0.747192",
"0.74385023",
"0.74383795",
"0.74351645",
"0.73968136",
"0.739678",
"0.7387185",
"0.73870254",
"0.7374685",
"0.7358254",
"0.7342969",
"0.73354375",
"0.73193264",
"0.7313369",
"0.7310191",
"0.73094094",
"0.7288084",
"0.7271018",
"0.7266416",
"0.7266071",
"0.7245424",
"0.72398627",
"0.7237568",
"0.72346246",
"0.72328717",
"0.7217567",
"0.7212101",
"0.72063375",
"0.7192321",
"0.7191647",
"0.7177966",
"0.71776855",
"0.7170745",
"0.71682096",
"0.71655846",
"0.71588016",
"0.7157398",
"0.715163",
"0.71493924",
"0.7147888",
"0.71425724",
"0.71418333",
"0.7132982",
"0.7132778",
"0.7129291",
"0.7124438",
"0.71228623",
"0.71124446",
"0.71075296",
"0.7096816",
"0.7081319",
"0.70733035",
"0.70676804",
"0.70615757",
"0.7052957",
"0.7051506",
"0.704847",
"0.7044685",
"0.7040819",
"0.7036627",
"0.7036526",
"0.7036526",
"0.7036526",
"0.70337415",
"0.70337415",
"0.70329154",
"0.7026249",
"0.7024486",
"0.7015"
] |
0.75571847
|
25
|
Writes a method for division and one for addition. Calls upon the return value in the second method.
|
def divide(a,b)
a/b
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def post_divide; end",
"def division (first_number, second_number)\nfirst_number / second_number\nend",
"def divide(number_one, number_two)\nnumber_one / number_two\nend",
"def divide\n match '/'\n factor\n emit_ln 'MOVE (SP)+,D1'\n emit_ln 'DIVS D1,D0'\nend",
"def numerator() end",
"def divide(num1, num2)\n p num1/num2\nend",
"def division(value1, value2)\n result = value1 / value2\n puts \"The sum of #{ value1 } / #{ value2 } = #{ result }\"\nend",
"def divide(numerator, denominator)\n puts \"#{numerator} / #{denominator} = #{numerator / denominator}\\n\"\nend",
"def divide(a, b)\n puts \"DIVIDING #{a} / #{b}\"\n return a / b\n a + b # This is ignored as the output due to the return\nend",
"def div(x, op, y)\n x.send(op, y)\nend",
"def calculate(operator, num1, num2)\n # Use case when to invoke each method\n case operator\n when \"add\", \"+\"\n puts \"#{num1} + #{num2} = #{add(num1, num2)}\"\n # Do I need a return here\n when \"subtract\", \"-\"\n puts \"#{num1} - #{num2} = #{subtract(num1, num2)}\"\n when \"multiply\", \"*\"\n puts \"#{num1} * #{num2} = #{multiply(num1, num2)}\"\n when \"divide\", \"/\"\n puts \"#{num1} / #{num2} = #{divide(num1, num2)}\"\n when \"modulo\", \"%\"\n puts \"#{num1} % #{num2} = #{mod(num1, num2)}\"\n when \"exponify\", \"**\"\n puts \"#{num1}^#{num2} = #{exponify(num1, num2)}\"\n end\nend",
"def /(int)\n\t\tself + int\n\tend",
"def divide(num1, num2)\n return num1 / num2\nend",
"def addition_method(num1, num2)\n\tputs num1 + num2\nend",
"def division num1, num2\n total = num1.to_i / num2.to_i\n puts \"--> The total is... #{total}\"\nend",
"def denominator() end",
"def divide(a, b)\n puts \"DIVIDING #{a} / #{b}\"\n a / b\nend",
"def divisione(a, b)\n puts \"DIVIDENDO #{a} / #{b}\"\n a / b\nend",
"def pre_divide; end",
"def divide(number_1, number_2)\n return number_1 / number_2\nend",
"def divide(number_1, number_2)\n return number_1 / number_2\nend",
"def divide(number_1, number_2)\n return number_1 / number_2\nend",
"def divisionm(num_1,num_2)\n return num_1 / num_2\nend",
"def my_math_method(num1, num2)\n\tnumber = num1 + num2\n\tputs number\nend",
"def my_math_method(num1, num2)\n\tnumber = num1 + num2\n\tputs number\nend",
"def operation(operator, num1, num2)\n case operator \n when 1 \n puts \"The operational result is: #{num1.to_f + num2.to_f}\"\n puts \"==> #{num1.to_s} + #{num2.to_s} = #{num1.to_f + num2.to_f}\"\n when 2 \n puts \"The operational result is: #{num1.to_f - num2.to_f}\"\n puts \"==> #{num1.to_s} - #{num2.to_s} = #{num1.to_f - num2.to_f}\"\n when 3\n puts \"The operational result is: #{num1.to_f * num2.to_f}\"\n puts \"==> #{num1.to_s} * #{num2.to_s} = #{num1.to_f * num2.to_f}\"\n else\n puts \"The operational result is: #{num1.to_f / num2.to_f}\"\n puts \"==> #{num1.to_s} / #{num2.to_s} = #{num1.to_f / num2.to_f}\"\n end\nend",
"def divide(other)\n Rubinius.primitive :float_div\n redo_coerced :/, other\n end",
"def divideThem(numOne, numTwo)\n numOne / numTwo\nend",
"def addition (a, b)\n a + b\nend",
"def divide(first_number, second_number)\n first_number / second_number\nend",
"def operations\n puts \"(a) - addition (+)\"\n puts \"(s) - subtraction (-)\"\n puts \"(m) - multiplication (*)\"\n puts \"(d) - division (/)\"\nend",
"def calculate(op, x, y)\n case op\n when 'plus'\n x.to_f + y.to_f\n when 'minus'\n x.to_f - y.to_f\n when 'divide'\n x.to_f / y.to_f\n when 'multiple'\n x.to_f * y.to_f\n end\nend",
"def div(x, y)\n x / y\nend",
"def add one, two\r\n\tone + two\r\nend",
"def /(other)\n\tself.num / other.num\nend",
"def calculator_method(operator, number_1, number_2)\n case operator\n when \"add\"\n total = number_1 + number_2\n when \"+\"\n total = number_1 + number_2\n when \"subtract\"\n total = number_1 - number_2\n when \"-\"\n total = number_1 - number_2\n when \"multiply\"\n total = number_1 * number_2\n when \"x\"\n total = number_1 * number_2\n when \"*\"\n total = number_1 * number_2\n when \"divide\"\n total = number_1.to_f / number_2.to_f\n when \"/\"\n total = number_1.to_f / number_2.to_f\n end\n return total\nend",
"def addition(a, b)\n\treturn a + b\nend",
"def math(operator, first_num, second_num)\n if operator ==\"*\"\n first_num*second_num\n elsif operator == \"/\"\n first_num/second_num\n elsif operator == \"+\"\n first_num+second_num\n else\n first_num-second_num\n end\nend",
"def calculate(operation, n1, n2)\n if operation == \"add\" || operation == \"+\"\n return \"#{n1} + #{n2} = #{n1+n2}\"\n elsif operation == \"subtract\" || operation == \"-\"\n return \"#{n1} - #{n2} = #{n1-n2}\"\n elsif operation == \"multiply\" || operation == \"*\"\n return \"#{n1} * #{n2} = #{n1*n2}\"\n elsif operation == \"divide\" || operation == \"/\"\n if n2 == 0\n return \"undefined\"\n else\n return \"#{n1} / #{n2} = #{n1/n2}\"\n end\n elsif operation == \"exponent\" || operation == \"^\"\n return \"#{n1} ^ #{n2} = #{n1**n2}\"\n elsif operation == \"modulo\" || operation == \"%\"\n return \"#{n1} % #{n2} = #{n1%n2}\"\n end\nend",
"def addition(a, b)\n\ta + b\nend",
"def calculator_method(number_one, operator, number_two)\n if operator == \"+\"\n return result = number_one + number_two\n elsif operator == \"-\"\n return result = number_one - number_two\n elsif operator == \"*\"\n return result = number_one * number_two\n else operator == \"/\"\n return result = number_one / number_two\n end\nend",
"def div_stringed(num1, num2)\r\n\r\nend",
"def my_math_method(num1, num2)\n sum = num1 + num2\n\tputs \"The sum of 2 + 2 is #{sum}\"\nend",
"def divide(first_number, second_number)\n return first_number / second_number\nend",
"def my_math_method(num1, num2)\n number = num1 + num2\n puts number\nend",
"def divide\n\n\t\t# make sure when can perform the operation\n\t\tempty?\n\n\t\t# pop the last 2 numbers and store but need to be converted to a float\n\t\tnum_1 = @calculator.pop.to_f\n\t\tnum_2 = @calculator.pop.to_f\n\n\t\t# division is performed and pushed to @calc\n\t\t@calculator << num_2 / num_1\n\n\tend",
"def addition(a, b) \n return a + b\nend",
"def division(input_number1, input_number2)\n\tquotient = input_number1 / input_number2\n\tputs \"The quotient of #{input_number1} and #{input_number2}th is #{quotient}.\"\n\tputs \"Thank you for using the calculator. Goodbye!\"\nend",
"def my_math_method(num1, num2)\n number = num1 + num2\n puts number\nend",
"def calculator(num1, num2)\n\treturn num1 + num2, num1 - num2, num1 * num2, num1 / num2\nend",
"def writeArithmetic(cmd,ln)\n\tcase cmd[:c]\n\twhen 'add','sub','and','or' then l = arith(cmd[:s])\n\twhen 'neg','not' then l = neginv(cmd[:s])\n\twhen 'gt','lt','eq' then l = glte(cmd[:s],ln)\n\telse return -1 end \n\treturn l\nend",
"def divide\n\t\tif @operands.size>= 2\n\t\t\tnew_n = (@operands[-2].to_f / @operands[-1].to_f)\n\t\t\t@operands.pop(2)\n\t\t\t@operands.push new_n\n\t\telse\n\t\t\traise \"calculator is empty\"\n\t\tend\n\tend",
"def add(a, b)\n\t# a method that adds two numbers together\n\tputs \"ADDING #{a} + #{b}\"\n\treturn a + b\nend",
"def divide(a, b)\n divide = a / b\n return divide\nend",
"def basic_operation(a, b, type)\n case type\n when 1\n puts \"Answer: #{a + b}\"\n when 2\n puts \"Answer: #{a - b}\"\n when 3\n puts \"Answer: #{a * b}\"\n when 4\n puts \"Answer: #{a / b}\"\n end\nend",
"def add (a, b) \n return a + b; \nend",
"def addition(a,b)\n\ta + b\nend",
"def divide(left, right, result) #method\n left = get_dir(left)\n right = get_dir(right)\n @current_context[result] = get_value(left).to_f / get_value(right).to_f\n end",
"def my_math_method(num1, num2)\n\tanswer = num1 + num2\n\tputs \"The sum of #{num1} and #{num2} is #{answer}.\"\nend",
"def mathy(n1, n2, operation)\n answer = n1.send(operation, n2).round(4)\n return answer\nend",
"def addition(num)\n\treturn num+1\nend",
"def /(value)\n mul(value) ;\n end",
"def divide (a,b)\n f = a.to_i / b.to_i\n puts \"Division is: #{f}\"\n end",
"def plus(number)\n @operation = '+' \n @number = number\nend",
"def addition (num1, num2)\n\tnum1 + num2\nend",
"def arithmetic2(a, b)\nend",
"def add\n\t number_1 + number_2\n end",
"def fraction_calculator fraction_one, fraction_two, operator\n num_one = Rational(fraction_one)\n num_two = Rational(fraction_two)\n \n final_result = case operator\n when '/' then num_one / num_two\n when '*' then num_one * num_two\n when '+' then num_one + num_two\n when '-' then num_one - num_two\n end\n\n String(final_result)\n\n print String(final_result)\n\nend",
"def add(first_num, second_num)\n @result = (first_num + second_num)\nend",
"def addition(a, b)\n puts a + b\nend",
"def addition(num_one, num_two)\n return num_one + num_two\nend",
"def fdiv(arg0)\n end",
"def divide (a,b)\n f = a.to_i / b.to_i\n puts \"Division of both number is: #{f}\"\n end",
"def arithmetic(a, b, operator)\n if operator == \"add\"\n a + b\n elsif operator == \"subtract\"\n a - b\n elsif operator == \"multiply\"\n a * b\n elsif operator == \"divide\"\n a/b\n else\n print \"hahaa\"\n end\nend",
"def basic_op(operator, value1, value2)\n case operator\n when \"+\"\n value1 + value2\n when \"-\"\n value1 - value2\n when \"*\"\n value1 * value2\n when \"/\"\n value1 / value2\n end\nend",
"def add num1, num2\n num1 + num2\nend",
"def basic_op(operator, value1, value2)\n case operator\n when '+'\n value1 + value2\n when '-'\n value1 - value2\n when '*'\n value1 * value2\n when '/'\n value1 / value2\n end\nend",
"def add(num_one, num_two)\n return num_one + num_two\nend",
"def add(num_one, num_two)\n return num_one + num_two\nend",
"def add(num_one, num_two)\n return num_one + num_two\nend",
"def add(num_one, num_two)\n return num_one + num_two\nend",
"def add(one, two)\n\tone + two\nend",
"def on_div(ast_node, context)\n left, right = *ast_node\n\n return on_call_number(context, left) / on_call_number(context, right)\n end",
"def /(op)\n CAS::Div.new self, op\n end",
"def sum (operator, a, b)\n if operator == \"add\"\n a + b\n elsif operator == \"subtract\"\n a - b\n elsif operator == \"multiply\"\n a * b\n elsif operator == \"divide\"\n a / b\n end\nend",
"def add(a, b)\n puts \"#{a} + #{b}\"\n return a + b\nend",
"def add(a, b)\n puts \"#{a} + #{b}\"\n return a + b\nend",
"def add(a, b)\n puts \"#{a} + #{b}\"\n return a + b\nend",
"def total(a,b)\r\n return (a+b)\r\nend",
"def add(a, b)\n return a + b\nend",
"def plus(a, b)\n a + b\nend",
"def plus(a, b)\n a + b\nend",
"def divide(one,two)\n\treturn \"I don't think so\" if two ==0\n\tone / two\nend",
"def do_math(num1, num2, operation)\n case operation\n when '+'\n num1.to_i + num2.to_i\n when '-'\n num1.to_i - num2.to_i\n when '*'\n num1.to_i * num2.to_i\n when '/'\n num1.to_f / num2.to_f\n end\nend",
"def operation_of(a, b, op)\n if op == \"add\"\n return a + b\n elsif op == \"subtract\"\n return a - b \n elsif op == \"multiply\"\n return a * b \n elsif op == \"divide\"\n if b == 0\n return \"undefined\"\n else\n return a / b\n end\n end\nend",
"def two_method(x,y,z)\n\t\tx + y + z\nend",
"def add (number1, number2)\n number1 +number2\nend",
"def add(a, b)\n return a + b\nend",
"def add(a, b)\n return a + b\nend",
"def add(a, b)\n return a + b\nend"
] |
[
"0.6588389",
"0.64899135",
"0.63537985",
"0.6314706",
"0.63010716",
"0.6295089",
"0.6290365",
"0.62898433",
"0.6272394",
"0.6244718",
"0.6205226",
"0.6169197",
"0.61394906",
"0.6136974",
"0.6125868",
"0.61156917",
"0.6095771",
"0.6093122",
"0.607734",
"0.6053234",
"0.6053234",
"0.6053234",
"0.60387385",
"0.60343814",
"0.60343814",
"0.6008222",
"0.5988967",
"0.5988259",
"0.5978238",
"0.5976249",
"0.5971578",
"0.59715027",
"0.5968943",
"0.596278",
"0.59517235",
"0.59433055",
"0.5936092",
"0.59301245",
"0.59296256",
"0.58988565",
"0.5896206",
"0.58959335",
"0.5894294",
"0.5894081",
"0.58837444",
"0.58792514",
"0.58730507",
"0.58629155",
"0.5862851",
"0.58562475",
"0.5853205",
"0.5846966",
"0.58409995",
"0.583341",
"0.5828945",
"0.5815451",
"0.5812831",
"0.5801287",
"0.5801057",
"0.5798074",
"0.57931376",
"0.5791826",
"0.57863724",
"0.57783175",
"0.57581097",
"0.5755215",
"0.57544655",
"0.5753887",
"0.5739164",
"0.57382095",
"0.5732889",
"0.5732721",
"0.5724572",
"0.57203674",
"0.5716728",
"0.5707087",
"0.5698212",
"0.5686546",
"0.5686546",
"0.5686546",
"0.5686546",
"0.5685443",
"0.5683317",
"0.5679106",
"0.56772435",
"0.56643295",
"0.56643295",
"0.56643295",
"0.56634974",
"0.56609213",
"0.56542933",
"0.56542933",
"0.5653952",
"0.5653777",
"0.56489193",
"0.56479925",
"0.5645892",
"0.56394196",
"0.56394196",
"0.56394196"
] |
0.629748
|
5
|
Return VDOE Cluster contact. The contact will respond to name, email, and phone, returning Strings.
|
def contact
@contact ||= OpenStruct.new(get_attr(:contact))
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def contact\n response[\"contact\"]\n end",
"def contact\n ::HubEdos::Common::Reference::Descriptor.new(@data['contact']) if @data['contact']\n end",
"def contact\n Zapi::Models::Contact.new\n end",
"def contact\n @contact ||= get_contact(@data_object.user_contact)\n end",
"def contact\n doc \"The identifier of this contact.\"\n id\n contact_attributes\n record_timestamps\n end",
"def contact\n find('Contact', contact_id)\n end",
"def get_contact(id)\n\t\t@nutshell.get_contact(id)\n\tend",
"def contact\n @contact ||= ActsAsIcontact::Contact.find(contactId.to_i) if contactId.to_i > 0\n end",
"def contact\n\t\t@contact\n\tend",
"def contact\n\t\t@contact\n\tend",
"def get_contact(contact_name, project_name, optional={})\n\t\targs = self.class.new_params\n\t\targs[:method] = 'GET'\n\t\targs[:path]['ContactName'] = contact_name\n\t\targs[:path]['ProjectName'] = project_name\n\t\targs[:pattern] = '/projects/[ProjectName]/contacts/[ContactName]'\n\t\targs[:query]['Action'] = 'GetContact'\n\t\targs[:region] = optional[:_region] if (optional.key? :_region)\n\t\targs[:scheme] = 'http'\n\t\tself.run(args)\n\tend",
"def contact(contact, options = {})\n get(\"contacts/#{contact}\", options).pop\n end",
"def get_contact\n if self.contact_id\n return self.contact\n end\n end",
"def contact; Contact.get(self.contact_id); end",
"def contact; Contact.get(self.contact_id); end",
"def contact; Contact.get(self.contact_id); end",
"def contact; Contact.get(self.contact_id); end",
"def contact; Contact.get(self.contact_id); end",
"def get_contact(contact)\n ensure_cache_up_to_date\n namespace = \"bluster:objects:contact:#{contact}\"\n keys = self.redis.keys(\"#{namespace}:*\")\n data = {}\n keys.each { |key|\n short_key = key.split(\":\")[4] \n data[short_key] = self.redis.get(\"#{namespace}:#{short_key}\")\n }\n data\n end",
"def contact; end",
"def contact(options={})\n if self.padma_contact.nil? || options[:force_service_call]\n self.padma_contact = PadmaContact.find(contact_id, {select: :all, account_name: self.account_name})\n end\n ret = padma_contact\n if options[:decorated] && padma_contact\n ret = PadmaContactDecorator.decorate(padma_contact)\n end\n ret\n end",
"def get_cluster(name)\n cl_obj = Com::Vmware::Vcenter::Cluster.new(vapi_config)\n\n # @todo: Use Cluster::FilterSpec to only get the cluster which was asked\n # filter = Com::Vmware::Vcenter::Cluster::FilterSpec.new(clusters: Set.new(['...']))\n clusters = cl_obj.list.select { |cluster| cluster.name == name }\n raise format(\"Unable to find Cluster: %s\", name) if clusters.empty?\n\n cluster_id = clusters[0].cluster\n cl_obj.get(cluster_id)\n end",
"def ntd_contact\n usrs = users_with_role :ntd_contact\n if usrs.empty?\n nil\n else\n usrs.first\n end\n end",
"def get(request_configuration=nil)\n request_info = self.to_get_request_information(\n request_configuration\n )\n error_mapping = Hash.new\n error_mapping[\"4XX\"] = lambda {|pn| MicrosoftGraph::Models::ODataErrorsODataError.create_from_discriminator_value(pn) }\n error_mapping[\"5XX\"] = lambda {|pn| MicrosoftGraph::Models::ODataErrorsODataError.create_from_discriminator_value(pn) }\n return @request_adapter.send_async(request_info, lambda {|pn| MicrosoftGraph::Models::OrgContact.create_from_discriminator_value(pn) }, error_mapping)\n end",
"def retrieve_contact_by_email\n unless @rolodex.is_empty?\n print \"Please, provide the email of the contact in question: \"\n contact = @rolodex.search_contact(gets.chomp)\n if contact\n @rolodex.display_particular(contact)\n else\n puts \"Error: contact not found.\"\n end\n else\n contact = false\n puts \"Error: the Rolodex is empty.\"\n end\n return contact\n end",
"def find(name, config, datacenter_name)\n cluster_mob = cluster_mobs[name]\n raise \"Can't find cluster '#{name}'\" if cluster_mob.nil?\n\n cluster_properties = @client.cloud_searcher.get_properties(\n cluster_mob, VimSdk::Vim::ClusterComputeResource,\n Cluster::PROPERTIES, :ensure_all => true\n )\n raise \"Can't find properties for cluster '#{name}'\" if cluster_properties.nil?\n\n Cluster.new(\n config,\n cluster_properties,\n @client,\n datacenter_name,\n )\n end",
"def contact\n end",
"def contact\n end",
"def contact\n end",
"def contact\n end",
"def contact\n end",
"def contact\n end",
"def contact\n end",
"def contact\n end",
"def contact\n end",
"def contact\n end",
"def contact\n end",
"def contact\n end",
"def contact\n end",
"def contact\n end",
"def get_contact\r\n @contact = @customer.contact_people.find_by_name params[:id]\r\n end",
"def contact\n\n end",
"def contact\n\n end",
"def get_contact_by_number(contact_number)\n get_contact(nil, contact_number)\n end",
"def get_contact_by_number(contact_number)\n get_contact(nil, contact_number)\n end",
"def contact\n query = self\n # only one:\n if email.blank? || phone.blank?\n # if phone, do a constant_score on the match clause\n if email.blank?\n filters = [Clauses::Phone.exact(DSL::Filter.new, query.phone)]\n build_with_constant_score(filters)\n # if email, just do the match clause\n else\n DSL::Search.new { query Clauses::Email.exact(DSL::Query.new, query.email) }\n end\n # both?\n else\n # use phone number as a filter\n filters = [Clauses::Phone.exact(DSL::Filter.new, phone)]\n # use email as the filter function, build with function_score\n build_with_function_score(filters)\n end\n end",
"def get_cluster(name)\n cluster_id = get_cluster_id(name)\n\n host_api = VSphereAutomation::VCenter::HostApi.new(api_client)\n raise_if_unauthenticated host_api, \"checking for cluster `#{name}`\"\n\n hosts = host_api.list({ filter_clusters: cluster_id, connection_states: \"CONNECTED\" }).value\n filter_maintenance!(hosts)\n raise_if_missing hosts, format(\"Unable to find active hosts in cluster `%s`\", name)\n\n cluster_api = VSphereAutomation::VCenter::ClusterApi.new(api_client)\n cluster_api.get(cluster_id).value\n end",
"def contacts\n contact_client.contacts\n end",
"def display_one_contact(contact)\n if contact\n puts \"Name: #{contact.name}\"\n puts \"E-mail: #{contact.email}\"\n puts \"Phone Number: #{contact.phone_number}\"\n else\n puts \"Not found!\"\n end\n end",
"def query_contacts(options={}) path = \"/api/v2/contacts\"\n get(path, options, AvaTax::VERSION) end",
"def fetch_contact_info\n lambda {\n user.vet360_contact_info\n }\n end",
"def build_contact(element, type)\n Record::Contact.new(\n :type => type,\n :id => node(\"#{element} ID\"),\n :name => node(\"#{element} Name\"),\n :organization => node(\"#{element} Organization\"),\n :address => node(\"#{element} Address\"),\n :city => node(\"#{element} City\"),\n :zip => node(\"#{element} Postal Code\"),\n :state => node(\"#{element} State/Province\"),\n :country_code => node(\"#{element} Country\"),\n :phone => node(\"#{element} Phone Number\"),\n :fax => node(\"#{element} Fax Number\"),\n :email => node(\"#{element} Email\")\n )\n end",
"def display_one_contact\n contact = retrieve_contact_by_email\n end",
"def contacts()\n return MicrosoftGraph::Contacts::ContactsRequestBuilder.new(@path_parameters, @request_adapter)\n end",
"def my_contacts(opts = {})\n client.get_my_contacts(opts)\n end",
"def get_contacts(params={})\n @obj.get('get-contacts', @auth.merge(params))\n end",
"def cluster( name )\n\n return ::Module::Cluster.instance_controller( self ).cluster( name )\n \n end",
"def contact\n @voucher.customer.nil? ? nil : ContactDrop.new(@voucher.customer.main_contact)\n end",
"def find_or_create_xero_contact(name)\n existing = @api.get('Contacts', where: \"Name=\\\"#{name}\\\"\")['Contacts']&.first\n if existing\n logger.debug \"Found existing contact with name: #{name}\"\n logger.debug \"ID: #{existing['ContactID']}\"\n return existing['ContactID']\n end\n\n logger.debug \"Creating new contact with name: #{name}\"\n response = @api.post('Contacts', 'Name' => name)\n id = response['Contacts'].first['ContactID']\n logger.debug \"Contact created with ID: #{id}\"\n id\n end",
"def list_contact(project_name, optional={})\n\t\targs = self.class.new_params\n\t\targs[:method] = 'GET'\n\t\targs[:path]['ProjectName'] = project_name\n\t\targs[:pattern] = '/projects/[ProjectName]/contacts'\n\t\targs[:query]['Action'] = 'ListContact'\n\t\targs[:region] = optional[:_region] if (optional.key? :_region)\n\t\targs[:scheme] = 'http'\n\t\tif optional.key? :contact_name\n\t\t\targs[:query]['ContactName'] = optional[:contact_name]\n\t\tend\n\t\tif optional.key? :page\n\t\t\targs[:query]['Page'] = optional[:page]\n\t\tend\n\t\tif optional.key? :page_size\n\t\t\targs[:query]['PageSize'] = optional[:page_size]\n\t\tend\n\t\tself.run(args)\n\tend",
"def contact?\r\n infoxml = get_info\r\n \r\n if contact = infoxml['contact']\r\n return contact == '1'\r\n end\r\n \r\n return false\r\n end",
"def cluster(name = context&.cluster)\n named_cluster(name)&.cluster\n end",
"def contacts\n @contacts ||= get_attr(:contacts).collect { |c| OpenStruct.new(c) }\n end",
"def contactDataGet(options={})\n assert_valid_keys(options, :customerId)\n assert_keys_exists(options, :customerId)\n execute(:contactDataGet, options)\n end",
"def contactDataGet(options={})\n assert_valid_keys(options, :customerId)\n assert_keys_exists(options, :customerId)\n execute(:contactDataGet, options)\n end",
"def cluster() node[:cluster_name] ; end",
"def contacts\n\t\t@contact = Contact.first();\n\tend",
"def contact\n \t@contact = Contact.new\n end",
"def contacts\n Easybill::Api::Contacts\n end",
"def find_one(_entity_type, uuid, _params)\n Contact.new(fake_contact_data.merge('contactid' => uuid))\n end",
"def contact_attributes(options = {})\n with_options options do\n # Specify the 'name' attribute\n doc <<-EOS.strip_heredoc\n The name of the contact.\n\n First name and last name are separated by white space.\n EOS\n string \"name\"\n # Specify the 'age' attribute.\n doc \"How old is the contact.\"\n integer \"age\"\n # Specify the 'homepage' attribute.\n uri \"homepage\", doc: \"The URL of the contact's homepage\"\n end\n end",
"def contact \n\n end",
"def where(options = {})\n _, _, root = @client.get(\"/contacts\", options)\n\n root[:items].map{ |item| Contact.new(item[:data]) }\n end",
"def contacts\n @contacts ||= Harvest::API::Contacts.new(credentials)\n end",
"def new_contact\n @contact = Spree::Address.new\n end",
"def contacts(params = {})\n # contacts in this group\n @contacts ||= get_contacts({\"group\" => self.id}.merge(params))\n end",
"def get_agent_contacts(params)\n message = {\n serviceGroupID: params[:service_group_id],\n serviceID: params[:service_id],\n teamID: params[:team_name],\n agentID: params[:agent_id],\n startDate: params[:start_date],\n endDate: params[:end_date],\n contactTypes: params[:contact_type],\n useServiceTime: false\n }\n\n reply = @client.call(:get_contacts, message: message)\n data = reply.body.dig(:get_contacts_response,\n :get_contacts_result,\n :array_of_string)\n\n data = check_if_data_exists(data)\n data = map_contacts_data(data)\n delete_contact_headers(data)\n data\n end",
"def contact_name\n contact['name']\n end",
"def contact_email\n self.contact[:contact_email]\n end",
"def new\n\t\t\t\t# we are going to make a new contact yall\n\t\t\t\t# comes in like post\n\t\t\t\t# {'api_token': ..., 'contact': {}}\n\t\t\t\tcontact_params = params[:contact] # be sure to clean all the values\n\t\t\t\t# clean them up\n\t\t\t\tcontact_params = sanitize_obj(contact_params);\n\t\t\t\t# lets allow rails to build this for us automagically\n\t\t\t\tc = Contact.new\n\t\t\t\tc.from_json(contact_params.to_json) # generate from our cleaned params\n\t\t\t\t# should be it for that, as long as the keys match, rails should set it\n\t\t\t\t\n\t\t\t\t# now we can save the contact\n\t\t\t\tc.save\n\t\t\t\t@user.accounts.first.contacts << c\n\t\t\t\t@user.accounts.first.save\n\t\t\t\t\n\t\t\t\t# now let's this new contact to the client\n\t\t\t\trender json: {:status => \"success\", :contact => c}\n\t\t\tend",
"def contact(id)\n self.contacts.detect { |c| c.id == id }\n end",
"def select_contact(options={})\n contact_list_id = options.delete(:list_id)\n email = options.delete(:email)\n request_body = String.new\n xml = Builder::XmlMarkup.new(:target => request_body, :indent => 1)\n\n xml.instruct!\n xml.Envelope do\n xml.Body do\n xml.SelectRecipientData do\n xml.LIST_ID contact_list_id\n xml.EMAIL email\n end\n end\n end\n\n doc = send_xml_api_request(request_body)\n result_dom(doc)\n end",
"def contact_mix\n $testCaseID = \"VT229-0019\"\n con_remove\n createContact 20\n @contact = Rho::RhoContact.find(:all, :per_page => 5, :offset => 5, :select => [\"id\", \"last_name\"], :conditions => {:phone => 'not_nil', :email => 'not_nil'})\n puts @contact\n redirect :action => :index\n end",
"def recipient_contact\n @recipient_contact ||= SimpleShipping::Contact.new(\n :person_name => \"John Recipient Smith\",\n :phone_number => \"1234567890\"\n )\n end",
"def create_contact()\n Contact.new(id: rand(2000)).tap do |c|\n contacts << c\n end\n end",
"def user\n User.get(contact, client: @client)\n end",
"def get_contacts(options = {})\n request_params = {}\n request_params[:type] = options[:type] if options[:type]\n request_params[:sortBy] = options[:sort] if options[:sort] \n request_params[:direction] = options[:direction] if options[:direction] \n \n response_xml = http_get(\"#{@xero_url}/contacts\", request_params)\n \n parse_response(response_xml, :request_params => request_params)\n end",
"def create(name, email)\n # TODO: Instantiate a Contact, add its data to the 'contacts.csv' file, and return it.\n CSV.foreach('contacts.csv') do |row|\n end\n id=$.\n new_contact = Contact.new(name, email, id)\n full_contact = new_contact.id, new_contact.name, new_contact.email\n CSV.open('contacts.csv', 'a+') do |csv|\n csv << full_contact\n end\n return new_contact\n end",
"def create_contact(options = {})\n post(:contacts, contacts: [options]).pop\n end",
"def show_contact(id)\n get(\"contacts/#{id}\")\n end",
"def campaign_for_contact entity\n entity.ensure_salesforce_id\n return if entity.salesforce_id.nil?\n \n campaign_member = campaign_member_for_contact(entity)\n return nil if campaign_member.nil?\n \n client.materialize('Campaign')\n SFDC_Models::Campaign.find(campaign_member.CampaignId)\n end",
"def contact_email\n return @contact_email\n end",
"def contactDataGet(options = {})\n assert_valid_keys(options, :accessKey, :testMode, :customerId)\n assert_keys_exists(options, :customerId)\n execute(:contactDataGet, options)\n end",
"def get_contact_by_id(contact_id)\n get_contact(contact_id)\n end",
"def get_contact_by_id(contact_id)\n get_contact(contact_id)\n end",
"def create_contact(project_name, optional={})\n\t\targs = self.class.new_params\n\t\targs[:method] = 'POST'\n\t\targs[:path]['ProjectName'] = project_name\n\t\targs[:pattern] = '/projects/[ProjectName]/contacts'\n\t\targs[:query]['Action'] = 'CreateContact'\n\t\targs[:region] = optional[:_region] if (optional.key? :_region)\n\t\targs[:scheme] = 'http'\n\t\tif optional.key? :contact\n\t\t\targs[:body]['Contact'] = optional[:contact]\n\t\tend\n\t\tself.run(args)\n\tend",
"def get_contact(contact_id, list_id, opts = {})\n data, _status_code, _headers = get_contact_with_http_info(contact_id, list_id, opts)\n data\n end",
"def contact_to(subject)\n sent_contacts.received_by(subject).first\n end",
"def create(name, email)\n # TODO: Instantiate a Contact, add its data to the 'contacts.csv' file, and return it.\n results = connection.exec(\"SELECT * FROM contacts ORDER BY id DESC LIMIT 1;\")\n results.each do |row|\n each_info = Contact.new(row['name'], row['email'], row['id'])\n puts \"ID #{row['id']}: #{row['name']} #{row['email']}\".blue\n end\n puts \"The contact was created successfully.\".green\n end",
"def get_contact_detail(business, category)\n contact_details.find_by(business: business, category: category)\n end"
] |
[
"0.66833985",
"0.66090137",
"0.65540034",
"0.64098597",
"0.62773436",
"0.6145131",
"0.609668",
"0.60651255",
"0.60434",
"0.60434",
"0.59495527",
"0.59093964",
"0.5896344",
"0.5817725",
"0.5817725",
"0.5817725",
"0.5817725",
"0.5817725",
"0.57928205",
"0.5744455",
"0.56779766",
"0.56750816",
"0.56317717",
"0.5619594",
"0.5594274",
"0.55569243",
"0.55180377",
"0.55180377",
"0.55180377",
"0.55180377",
"0.55180377",
"0.55180377",
"0.55180377",
"0.55180377",
"0.55180377",
"0.55180377",
"0.55180377",
"0.55180377",
"0.55180377",
"0.55180377",
"0.5504318",
"0.5492562",
"0.5492562",
"0.5450975",
"0.5450975",
"0.54458755",
"0.5436668",
"0.54334944",
"0.5432021",
"0.54282296",
"0.54027396",
"0.53924155",
"0.539158",
"0.5390824",
"0.5385284",
"0.53827745",
"0.5381669",
"0.53611815",
"0.5356826",
"0.53376853",
"0.532946",
"0.5327468",
"0.5308559",
"0.53013253",
"0.53013253",
"0.53006434",
"0.52990323",
"0.52973104",
"0.52937263",
"0.5280699",
"0.52787006",
"0.5259074",
"0.52539736",
"0.5230242",
"0.5225923",
"0.52226126",
"0.5220394",
"0.52175796",
"0.5216729",
"0.52115315",
"0.52102846",
"0.5208956",
"0.5201804",
"0.5195266",
"0.5179164",
"0.5170714",
"0.5168884",
"0.5167123",
"0.51595724",
"0.51514554",
"0.5138476",
"0.5136244",
"0.5128729",
"0.5116795",
"0.5116795",
"0.5112418",
"0.51103204",
"0.5109598",
"0.5107859",
"0.5098026"
] |
0.654526
|
3
|
Courses related to the Cluster
|
def courses
@courses ||= CourseList.new(:cluster => slug.gsub('-', ' ')).
sort_by { |c| c.title + c.edition }.
uniq { |c| c.code + c.edition }
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def courses\n [self]\n end",
"def courses\n course_list = []\n \n semesters.collect do |semester|\n course_list << semester.cis_courses\n end\n \n course_list << course_bin.cis_courses\n \n return course_list\n end",
"def courses\n Content::Courses.new(token: @token)\n end",
"def section_courses\n all_sections = sections.to_a.concat(sections_as_student).uniq\n\n # In the future we may want to make it so that if assigned a script, but that\n # script has a default course, it shows up as a course here\n all_sections.map(&:course).compact.uniq\n end",
"def courses\n request(COURSES_URL, {}, 'GET').map do |course|\n Course.new(self, course)\n end\n end",
"def cluster_list\n super\n end",
"def courses\n @courses ||= Stellar::Courses.new self\n end",
"def list_courses(courses_collection)\n courses_collection.each do |course|\n\n end\n end",
"def courses\n bins.map(&:courses).flatten\n end",
"def make_courses\n course_array = Scraper.scrape_index_page(MODIFIED_BASE_PATH)\n Courses.create_from_collection(course_array)\n end",
"def courses\n @learn_courses\n end",
"def addCourses\n\t\tself.ccounter = 1 \t\n\tend",
"def courses\n Course.all.select { |course_inst| course_inst.student == self }\n end",
"def courses\n @courses = Course.where(\"teacher_id=?\",session[:user].teacher_id)\n end",
"def all_courses\n [\n {\n id: 1,\n title: 'The Complete Node.js Developer Course',\n author: 'Andrew Mead, Rob Percival',\n description: 'Learn Node.js by building real-world applications with Node, Express, MongoDB, Mocha, and more!',\n topic: 'Node.js',\n url: 'https://codingthesmartway.com/courses/nodejs/'\n },\n {\n id: 2,\n title: 'Node.js, Express & MongoDB Dev to Deployment',\n author: 'Brad Traversy',\n description: 'Learn by example building & deploying real-world Node.js applications from absolute scratch',\n topic: 'Node.js',\n url: 'https://codingthesmartway.com/courses/nodejs-express-mongodb/'\n },\n {\n id: 3,\n title: 'JavaScript: Understanding The Weird Parts',\n author: 'Anthony Alicea',\n description: 'An advanced JavaScript course for everyone! Scope, closures, prototypes, this, build your own framework, and more.',\n topic: 'JavaScript',\n url: 'https://codingthesmartway.com/courses/understand-javascript/'\n }\n ]\n end",
"def create_courses(ed_org_id)\n courses = Hash.new\n course_counter = 0\n GradeLevelType::get_ordered_grades.each do |grade|\n current_grade_courses = Array.new\n if !@scenarioYAML[grade.to_s + \"_COURSES\"].nil?\n @scenarioYAML[grade.to_s + \"_COURSES\"].each do |course|\n course_counter += 1\n current_grade_courses << {\"id\" => course_counter, \"title\" => course, \"ed_org_id\" => ed_org_id}\n end\n else\n course_counter += 1\n current_grade_courses << {\"id\" => course_counter, \"title\" => GradeLevelType.to_string(grade), \"ed_org_id\" => ed_org_id}\n end\n courses[grade] = current_grade_courses\n end\n courses\n end",
"def cluster_by\n return @cluster_by\n end",
"def index \n @courses = Course.all\n puts 'Courses-Index -mashal is super helpful'\n end",
"def parallel_drug_courses\n ConceptName.where(name: ['CPT', 'INH', 'Rifapentine', 'Isoniazid/Rifapentine'])\n end",
"def set_courses_category\n @courses_categories = Course.order(:name).pluck(:name, :id)\n end",
"def check_multiple_courses\n \n end",
"def current_course\n\n @user = current_user\n @courses = []\n \n if @user.parent or @user.student\n @student = Student.find(:all,\n #:select => \"\",\n :joins => \"inner join batches b on students.batch_id = b.id \",\n :conditions => \"b.is_active = true and b.is_deleted = false \"+\n \"and students.is_active = true and students.is_deleted = false \"+\n \"and students.user_id = \"+@user.id.to_s+\"\"\n ).first\n\n if !@student.nil?\n @courses = Course.find(:all,\n :conditions => \"is_deleted = false \"+\n \"and id in ( \"+ @student['batch_id'].to_s + \")\"\n )\n end\n end\n @courses\n end",
"def available_courses\n MAX_COURSES - self.course_selections_count\n end",
"def index\n @courses = Courses.all\n end",
"def show_courses_on_waitlist\n Waitlist.where(student: self).order(:course_id).map(&:student)\n end",
"def manage_cluster\n @nodes = Node.find(:all)\n end",
"def cluster( name )\n\n return ::Module::Cluster.instance_controller( self ).cluster( name )\n \n end",
"def course_array\n self.current_active_courses.all.map { |course| [course.name, course.id] }\n end",
"def courses\n unless user_signed_in? && current_user.instructor\n render :nothing => true, :status => :unauthorized\n end\n \n $selected_course = nil\n\n @courses = current_user.instructor.courses.collect{ |course|\n {\n name: course.name,\n id: course.id\n }\n }\n\n render :template => \"home/index\"\n end",
"def index\n @clusters = current_account.clusters\n end",
"def course\n unless user_signed_in? && current_user.instructor\n render :nothing => true, :status => :unauthorized\n end\n \n $selected_course = Course.where({id: params[:id]}).first\n\n @course_projects = $selected_course.projects.collect{ |project|\n {\n project_name: project.name,\n team_names: project.teams.reduce(''){|names, team| names + team.name + ', '}[0..-3],\n due: project.due.in_time_zone('Eastern Time (US & Canada)').strftime('%Y-%m-%d %I:%M:%S %p')\n }\n }\n\n render :template => \"home/index\"\n end",
"def index\n\t\t@courses = Course.all\n\tend",
"def create(params)\n response = self.class.post(\n '/webservice/rest/server.php',\n {\n :query => query_hash('core_course_create_courses', token),\n :body => {\n :courses => {\n '0' => {\n :fullname => params[:full_name],\n :shortname => params[:short_name],\n :categoryid => params[:parent_category],\n :idnumber => params[:idnumber]\n }\n }\n }\n }.merge(query_options)\n )\n check_for_errors(response)\n response.parsed_response.first\n end",
"def courses_all\n call_path = \"courses/all\"\n data = build_post_data(\"\")\n perform_post(build_url(call_path), data)\n end",
"def process_courses(user)\n if !user.global_role.can_edit_system_configuration? &&\n !user.global_role.can_manage_all_courses?\n\n # Everyone can manage their own course enrollments\n can :manage, CourseEnrollment, user_id: user.id\n\n can :enroll, CourseOffering, self_enrollment_allowed: true\n\n can :unenroll, CourseOffering\n\n # A user can manage a CourseOffering if they are enrolled in that\n # offering and have a CourseRole where can_manage_course? is true.\n can [:edit, :update], CourseOffering,\n CourseOffering.managed_by_user(user) do |co|\n co.is_manager? user\n end\n\n # A user can grade a CourseOffering if they are enrolled in that\n # offering and have a CourseRole where can_grade_submissions? is true.\n can :generate_gradebook, CourseOffering do |co|\n co.is_staff? user\n end\n\n # Likewise, a user can only manage enrollments in a CourseOffering\n # that they have can_manage_courses? permission in.\n can :manage, CourseEnrollment do |enrollment|\n enrollment.course_offering.is_manager? user\n end\n end\n end",
"def index\n if current_user.is_admin\n @courses = Course.all\n else\n @courses = current_user.courses\n end\n end",
"def index\n #to see tenant info\n #@tenant = Tenant.current_tenant\n @courses = Course.all\n end",
"def course_map\n @course_map ||= all_courses.each_with_object({}) do |course, map|\n if map[course.comp_key]\n map[course.comp_key] << course\n else\n map[course.comp_key] = [course]\n end\n end\n end",
"def clusters\n @clusters ||= parse_clusters(ENV['OOD_CLUSTERS'])\n end",
"def cluster() node[:cluster_name] ; end",
"def courses\n item_courses = items.flat_map do |item|\n item.fetch('courses', []).map do |course|\n {\n course_name: course['name'],\n course_id: course['courseNumber'],\n instructors: Array(course['instructorNames']), # NOTE: we've seen cases where instructorNames is nil.\n reserve_desk: course['locationCode']\n }\n end\n end\n\n item_courses.uniq { |c| c[:course_id] }\n end",
"def process_courses(user)\n # A user can manage a CourseOffering if they are enrolled in that\n # offering and have a CourseRole where can_manage_course? is true.\n\n can :read, CourseOffering do |offering|\n user.course_offerings.include?(offering)\n end\n\n can :manage, CourseOffering do |offering|\n user.managing_course_offerings.include?(offering)\n end\n\n # Likewise, a user can only manage enrollments in a CourseOffering\n # that they have can_manage_courses? permission in.\n can :manage, CourseEnrollment do |enrollment|\n user_enrollment = CourseEnrollment.where(\n user_id: user.id,\n course_offering_id: enrollment.course_offering.id).first\n\n user_enrollment && user_enrollment.course_role.can_manage_course?\n end\n end",
"def index\n @secondary_courses = SecondaryCourse.all\n end",
"def index\n @courses = current_teacher.courses.all.to_a + Course.find(current_teacher.jobs.pluck(:course_id).uniq)\n @course = current_teacher.courses.new\n end",
"def course_sections\n ['1', '2']\n end",
"def conversation_courses(conversation, structure)\n courses = structure.courses.with_deleted.find(conversation.mailboxer_course_ids.split(',')) if conversation.mailboxer_course_ids.present?\n end",
"def cluster_roles\n iterate :clusterroles do |r|\n setup_role role_kind: :ClusterRole, role: r\n end\n\n # For cluster roles with aggregation rules create an edge betweeen those roles\n @aggregable_roles.each do |aggregating_role, composite_roles|\n composite_roles.each do |composite_role|\n edge :aggregate, {\n aggregating_role_name: aggregating_role, \n composite_role_name: composite_role\n }\n edge :composite, {\n aggregating_role_name: aggregating_role,\n composite_role_name: composite_role\n }\n end\n end\n end",
"def cluster(name = context&.cluster)\n named_cluster(name)&.cluster\n end",
"def index\n reservations = Reservation.where(:user_id => current_user.id)\n for reservation in reservations\n c = Course.find(reservation.course_id)\n @courses.append(c)\n end\n\n render layout: \"application\"\n \n end",
"def to_json_d3v3_cola(required_courses, elective_courses, elective_category)\n\t\tvert_to_i_table = self.vertices.each_with_index.to_a.to_h\n\t\t\n\t\t\n\t\tout = Hash.new\n\t\t\n\t\tout['nodes'] = \n\t\t\tself.vertices.each_with_index.collect do |v, i|\n\t\t\t\ttype = node_type(v, required_courses, elective_courses)\n\t\t\t\tclass_string = type.to_s.tr('_', '-')\n\t\t\t\t\n\t\t\t\telective_type =\n\t\t\t\t\tif type == :elective\n\t\t\t\t\t\telective_category[v]*2\n\t\t\t\t\telse\n\t\t\t\t\t\t-1\n\t\t\t\t\tend\n\t\t\t\t\n\t\t\t\t\n\t\t\t\t{\n\t\t\t\t\t'name' => v,\n\t\t\t\t\t'number' => i,\n\t\t\t\t\t'chain_deps' => ancestors(v).to_a.collect{ |x| vert_to_i_table[x] },\n\t\t\t\t\t'descendants' => descendants(v).to_a.collect{ |x| vert_to_i_table[x] },\n\t\t\t\t\t'class' => class_string,\n\t\t\t\t\t'elective_type' => elective_type\n\t\t\t\t}\n\t\t\tend\n\t\t\n\t\t\n\t\tout['links'] = \n\t\t\tself.each_edge.collect do |u,v|\n\t\t\t\t{\n\t\t\t\t\t'source' => vert_to_i_table[u],\n\t\t\t\t\t'target' => vert_to_i_table[v]\n\t\t\t\t}\n\t\t\tend\n\t\t\n\t\tout['constraints'] = constraints_foo()\n\t\t\n\t\tJSON.generate(out)\n\tend",
"def cluster\n ElasticSearchHelpers.memoized_cluster\n end",
"def initialize\n @mech = mech\n \n @courses = nil\n end",
"def index\n @sections = Section.all\n @courses_for_sections = {}\n @courses_for_links = {}\n @all_courses = Course.all\n @all_courses.each do |course|\n @courses_for_links[course.id] = course\n course_title_raw = course.course_title.split(\" \")\n course_title_fixed = \"\"\n course_title_raw.each do |word|\n if word != \"of\" && word != \"the\"\n course_title_fixed = course_title_fixed + word[0] \n end\n end\n @courses_for_sections[course.id] = course_title_fixed + \" \" + course.course_description \n end\n\n puts @all_courses.inspect\n puts @courses_for_sections.inspect \n\n end",
"def course; end",
"def set_course\n @courses = Course.find(params[:course_id])\n end",
"def list_courses\n if current_user.is_admin?\n @user = User.find(params[:id])\n @courses = @user.courses\n respond_to do |format|\n format.xml { render :xml => @courses }\n end\n else\n respond_to do |format|\n format.xml { render :text => \"error\" }\n end\n end\n end",
"def course\n return direct_course if direct_course.present?\n group && group.course\n end",
"def listCluster()\n @crosssiteadmin.listClusters()\n end",
"def index\n @path_courses = PathCourse.all\n end",
"def index\n @classcourses = Classcourse.all\n end",
"def index\n @courses = Array.new\n Rails.configuration.x.displayOptions.courseCategories.each do |cat|\n @courses.push(Course.where(\"college = ?\", cat).order(\"title ASC\"))\n end\n end",
"def course\n if validateurl({\"name\": params[:name], \"semester\": params[:semester], \"coursename\": params[:course]})\n @semester = params[:semester]\n @coursename = params[:course]\n courseid = Course.select(\"id\").where(\"name = '\" + @coursename + \"' AND semester = '\" + @semester + \"'\" ).ids[0].to_s\n @students = ActiveRecord::Base.connection.execute(\"SELECT grades.grade, students.name FROM grades, students WHERE '\" + courseid.to_s + \"' = grades.course_id AND '\" + @semester.to_s + \"' = grades.semester AND students.id = grades.student_id\")\n end\n end",
"def index\n @subject_courses_relations = SubjectCoursesRelation.all\n end",
"def list_abstract_courses\n AbstractCourse.find_each do |abstract_course|\n puts \" \" + abstract_course_to_s( abstract_course )\n end\n end",
"def build_courses_object(course_arr)\n \tcourses = {}\n\tcourse_arr.each {|course|\n\t\tif course != nil\n\t\t\tif course.course_name != \"\"\n\t\t\t\tcourses[course.id] = course.course_name + \" \" + course.CourseTitle\n\t\t\tend\n\t\tend\n\t} \t\n\treturn courses\n end",
"def active_assignments \n courses.each do |course|\n course.active_assignments\n end\n end",
"def update_courses\n unless course_ids.nil?\n self.enrollments.each do |e|\n e.destroy unless course_ids.include?(e.course_id.to_s)\n course_ids.delete(e.course_id.to_s)\n end \n course_ids.each do |c|\n self.enrollments.create(:course_id => c) unless c.blank?\n @cse = Course.find(c)\n @cse.assignments.each do |assignment|\n Gradation.create(:assignment_id => assignment.id, :student_id => self.id, :course_id => c)\n end\n reload\n self.course_ids = nil\n end\n end\nend",
"def create\n\t\tcourse = Course.new(course_params)\n\t\tif course.save\n\n\t\t params[:course][\"major_id\"].each do |major_id|\n\t if !major_id.empty?\n\t course.major << Major.find(major_id)\n\t end\n\t end\n\n\t params[:course][\"minor_id\"].each do |minor_id|\n\t if !minor_id.empty?\n\t course.minor << Minor.find(minor_id)\n\t end\n\t end\n\n\t params[:course][\"concentration_id\"].each do |concentration_id|\n\t if !concentration_id.empty?\n\t course.concentration << Concentration.find(concentration_id)\n\t end\n\t end\n\n\t params[:course][\"distribution_id\"].each do |distribution_id|\n\t if !distribution_id.empty?\n\t course.distribution << Distribution.find(distribution_id)\n\t end\n\t end\n\n\t\t\tredirect_to '/courses'\n\t\telse\n\t\t\tflash[:danger] = \"The form you submitted is invalid.\"\n\t\t\tredirect_to '/courses/new'\n\t\tend\n\tend",
"def each_course(category_url, &block)\n visit category_url\n\n courses = []\n\n all('.columnas-familiafp ul:nth-of-type(1) li').each do |course|\n courses << [course.text, :medium]\n end\n\n all('.columnas-familiafp ul:nth-of-type(2) li').each do |course|\n courses << [course.text, :high]\n end\n\n courses.each do |course|\n block.call *course\n end\n end",
"def index\r\n @courses = Course.all\r\n end",
"def index\n @courts = Court.all\n end",
"def set_course\n @course = Courses.find(params[:id])\n end",
"def index\n @admin_courses = Course.all\n end",
"def star_cluster; end",
"def schedules\n\n\n\n main_courses = Course.where(department_id: self.department_id)\n .where(course_num: self.course_num)\n .where(term_id: self.search.term_id)\n .where(parent_course_id: nil)\n .includes(:child_courses)\n\n # Loop thru the main courses (the LEC )\n sched_list = main_courses.collect do |course|\n # This is a single section course, so just bail out\n return [course] if course.child_courses.empty?\n \n types = {}\n self.course_types.each do |type|\n types[type] = [] unless type == course.type\n end\n\n course.child_courses.each do |crs|\n types[crs.type] << crs\n end\n\n [course].product(*types.values)\n end\n \n sched_list.flatten(1)\n end",
"def courses_taked\n\t\tself.normal_scores.includes(:course_detail, :semester, :course_teachership, :course, :course_field)\n\tend",
"def addcourse\n @courses = Course.all\n @me.courses.each do |mine|\n @courses.delete(mine)\n end\n end",
"def index\n begin\n case\n when params[:semester_id]\n @courses = Semester.find(params[:semester_id]).courses\n when params[:teacher_id]\n check_token(params[:teacher_id], params[:token], true)\n @courses = User.find(params[:teacher_id]).real_teaching_courses\n when params[:student_id]\n check_token(params[:student_id])\n @courses = Course.none\n User.find(params[:student_id]).participations.each do |participation|\n @courses <<= participation.course if ROLE_STUDENT == participation.role\n end\n when params[:assistant_id]\n check_token(params[:assistant_id])\n @courses = Course.none\n User.find(params[:assistant_id]).participations.each do |participation|\n @courses <<= participation.course if ROLE_ASSISTANT == participation.role\n end\n else\n json_failed\n end\n rescue ActiveRecord::RecordNotFound\n json_failed(REASON_RESOURCE_NOT_FOUND)\n end\n end",
"def index\n @courses = Course.all\n @titre = \"All courses\"\n @description = \"Here, you will find the list of all the Courses available in our High school :\"\n end",
"def make_courses\n self.get_courses.each do |post|\n course = Course.new\n course.title = post.css(\"h2\").text\n course.schedule = post.css(\".date\").text\n course.description = post.css(\"p\").text\n end\n end",
"def index\n @courses = Course.all\n @current_semester = Semester.find_by_active(1)\n if @courses.present?\n @current_semester_courses = @current_semester.courses\n end\n @check_user = current_user.admin\n @semesters = Semester.all\n end",
"def get_cluster_by_key\n return @cluster_by\n end",
"def courses(institution, pageIndex=0, options={})\n options.merge!({:query => {:pageIndex => pageIndex}})\n self.class.get(\"/Institution/#{institution}/Courses.json\", options)\n end",
"def index\n @standard_categories = StandardCategory.where course:@course\n end",
"def course\n\t\treturn self.section.course if self.section\n\t\treturn nil\n\tend",
"def group_by_used!(feed)\n # prepare details of existing course site\n course_term_year = feed[:canvas_course][:term][:term_yr]\n course_term_code = feed[:canvas_course][:term][:term_cd]\n course_ccns = []\n feed[:canvas_course][:officialSections].each do |official_section|\n section_term_match = (official_section[:term_cd] == course_term_code) && (official_section[:term_yr] == course_term_year)\n raise RuntimeError, \"Invalid term specified for official section with CCN '#{official_section[:ccn]}'\" unless section_term_match\n course_ccns << official_section[:ccn]\n end\n\n associatedCourses = []\n unassociatedCourses = []\n\n feed[:teachingSemesters].each do |semester|\n course_semester_match = (semester[:termCode] == course_term_code) && (semester[:termYear] == course_term_year)\n if course_semester_match\n semester[:classes].each do |course|\n # either iterate and count the matches\n # or loop through and return the matches, then count that\n course[:containsCourseSections] = false\n course[:sections].each do |section|\n if course_ccns.include?(section[:ccn])\n course[:containsCourseSections] = true\n section[:isCourseSection] = true\n else\n section[:isCourseSection] = false\n end\n end\n if course[:containsCourseSections]\n associatedCourses << course\n else\n unassociatedCourses << course\n end\n end\n semester[:classes] = associatedCourses + unassociatedCourses\n else\n semester[:classes].each do |course|\n course[:containsCourseSections] = false\n end\n end\n end\n feed\n end",
"def index\n @courses = Course.all\n\n end",
"def index\n @course=Coursecontent.all\t \n end",
"def new_assessment_course\n @assessments = current_user.assessments.where(\"no_of_questions > 0\")\n @courses = current_user.courses.where(\"size > 0\")\n @current_package = current_user.packages.where(\"id = #{params[:id]}\")[0]\n end",
"def index\n @courses = Course.all\n end",
"def index\n @courses = Course.all\n end",
"def index\n @courses = Course.all\n end",
"def index\n @courses = Course.all\n end",
"def index\n @courses = Course.all\n end",
"def index\n @courses = Course.all\n end",
"def index\n @courses = Course.all\n end",
"def index\n @courses = Course.all\n end",
"def index\n @courses = Course.all\n end",
"def index\n @courses = Course.all\n end",
"def index\n @courses = Course.all\n end"
] |
[
"0.6577266",
"0.64513713",
"0.64507455",
"0.6379219",
"0.63550264",
"0.6263102",
"0.62602764",
"0.6251284",
"0.6235173",
"0.6113608",
"0.60380566",
"0.5993471",
"0.5989406",
"0.5977029",
"0.5928846",
"0.5905281",
"0.5895975",
"0.58901745",
"0.5882325",
"0.5856688",
"0.5842029",
"0.5830289",
"0.5828971",
"0.5818052",
"0.58112264",
"0.57976145",
"0.57819045",
"0.57553816",
"0.5732665",
"0.57195836",
"0.5710453",
"0.5692805",
"0.56726223",
"0.56704557",
"0.5668289",
"0.56615835",
"0.56587493",
"0.56447923",
"0.56405866",
"0.5637685",
"0.56362545",
"0.5620334",
"0.561657",
"0.56142616",
"0.56140023",
"0.56111366",
"0.56097376",
"0.55983984",
"0.55876464",
"0.55800974",
"0.55758",
"0.5558121",
"0.55509776",
"0.55480033",
"0.5538731",
"0.55372685",
"0.5535687",
"0.55349505",
"0.55339915",
"0.5532156",
"0.5520587",
"0.5517927",
"0.5515271",
"0.55135",
"0.5505866",
"0.5504998",
"0.55020714",
"0.5490454",
"0.5489604",
"0.5484922",
"0.54829997",
"0.5477665",
"0.5473961",
"0.5473231",
"0.54725945",
"0.5466676",
"0.54630494",
"0.5460841",
"0.545977",
"0.5459075",
"0.54566205",
"0.54563",
"0.54556286",
"0.54551953",
"0.5445626",
"0.54426056",
"0.5440946",
"0.54395974",
"0.54381746",
"0.54334944",
"0.54334944",
"0.54334944",
"0.54334944",
"0.54334944",
"0.54334944",
"0.54334944",
"0.54334944",
"0.54334944",
"0.54334944",
"0.54334944"
] |
0.6826941
|
0
|
Store a global content data object as a part of content data keeping service. If type is full flush then content data is saved as it. Otherwise incremental diff of comparison with the latest saved file is calculated and saved. In the later case content data objects containing data regarding added files and removed files is stored in a separate files. If there is no added or/and removed files, then appropreate diff files are not created. NOTE global content data object should be locked for writing while adding it to this service.
|
def add(is_full_flush)
content_data = $local_content_data
return if content_data.empty?
# for the latest content data added to the DB
latest_snapshot = nil
# for instances that were added regarding latest_snapshot
added_cd = nil
# for insrances that were removed regarding latest_snapshot
removed_cd = nil
latest_index_time = nil
content_data.each_instance do |_,_,_,_,_,_,index_time|
if latest_index_time.nil? || latest_index_time < index_time.to_i
latest_index_time = index_time.to_i
end
end
# Checking time consistency
content_data_timestamp = DateTime.from_epoch(latest_index_time)
if !@latest_timestamp.nil? && (content_data_timestamp <= @latest_timestamp)
# It is possible when instances added at @latest_timestamp
# were removed and any new instances addded,
# then latest indexed time in the new content data
# is earlier then @latest_timestamp
# Example:
# ContentData for Date1 (noted as latest_snapshot):
# Content1
# location1
# location2
# Between Date1 and Date2 location2 was removed
# and no other file operations were done.
# ContentData for Date2 (noted as content_data):
# Content1
# location1
# Then:
# content_data.remove_instances(latest_snapshot) is empty.
latest_snapshot = get(@latest_timestamp)
added_cd = content_data.remove_instances(latest_snapshot)
msg = "latest index time of the content data: #{content_data_timestamp}" +
"must be later then latest timestamp: #{latest_timestamp}"
if added_cd.empty?
# In this case we do not know exactly when the indexation was
# then the timestamp is fake
# TODO better solution?
latest_timestamp_epoch = @latest_timestamp.strftime('%s')
content_data_timestamp = DateTime.from_epoch(latest_timestamp_epoch.to_i + 1)
Log.warning msg
else
fail msg
end
end
if is_full_flush
save(content_data,
DiffFile::SNAPSHOT_TYPE,
nil, # 'from' param is not relevant for full flush
content_data_timestamp)
end
# If it is not the first content data that we store,
# i.e there are already stored content data,
# then a diff files are relevant.
# NOTE we save diff (added/removed) files even in the case of full flush
# cause of data consistency. It is crucial for the diff operation.
# Example (simple, without removed):
# When:
# date1-date2.added
# date2-date3.added (added along with a snapshot)
# date3.snapshot
# date3-date4.added
# Then:
# ContentDataDb.diff(date2, date4) = date2-date3.added + date3-date4.added
if @latest_timestamp.nil?
earliest_index_time = nil
content_data.each_instance do |_,_,_,_,_,_,index_time|
if earliest_index_time.nil? || earliest_index_time > index_time
earliest_index_time = index_time
end
end
content_data_from = DateTime.from_epoch(earliest_index_time)
save(content_data,
DiffFile::ADDED_TYPE,
content_data_from,
content_data_timestamp)
else
latest_snapshot ||= get(@latest_timestamp)
added_cd ||= content_data.remove_instances(latest_snapshot)
unless added_cd.empty?
save(added_cd,
DiffFile::ADDED_TYPE,
@latest_timestamp,
content_data_timestamp)
end
removed_cd = latest_snapshot.remove_instances(content_data)
unless removed_cd.empty?
save(removed_cd,
DiffFile::REMOVED_TYPE,
@latest_timestamp,
content_data_timestamp)
end
end
@latest_timestamp = content_data_timestamp
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def global_data\n @global_data ||= DataStore.new(File.expand_path(\"global_data.json\", home_path))\n end",
"def save(content_data, type, from, till)\n filename = DiffFile.compose_filename(type, from, till)\n path = case type\n when DiffFile::SNAPSHOT_TYPE\n File.join(@snapshots_path, till.year.to_s, filename)\n when DiffFile::ADDED_TYPE, DiffFile::REMOVED_TYPE\n File.join(@diffs_path, till.year.to_s, filename)\n else\n fail ArgumentError, \"Unrecognized type: #{type}\"\n end\n dirname = File.dirname(path)\n unless Dir.exist?(dirname)\n FileUtils.mkdir_p dirname\n end\n\n content_data.to_file(path)\n filename\n end",
"def save_object(type: nil, object_id: nil, title: nil, content: nil, visible: true, data: {})\n base_object = { objectID: object_id, visible: visible, type: type }\n object = { title: title, content: content, data: data }.merge(base_object)\n\n index_objects(object_index(type), [data.merge(base_object)])\n index_objects(global_index, [object])\n end",
"def store_local(ltype=nil, ctype=nil, data=nil, filename=nil)\n if ! ::File.directory?(Msf::Config.local_directory)\n FileUtils.mkdir_p(Msf::Config.local_directory)\n end\n\n # Split by fname an extension\n if filename and not filename.empty?\n if filename =~ /(.*)\\.(.*)/\n ext = $2\n fname = $1\n else\n fname = filename\n end\n else\n fname = ctype || \"local_#{Time.now.utc.to_i}\"\n end\n\n # Split by path separator\n fname = ::File.split(fname).last\n\n case ctype # Probably could use more cases\n when \"text/plain\"\n ext ||= \"txt\"\n when \"text/xml\"\n ext ||= \"xml\"\n when \"text/html\"\n ext ||= \"html\"\n when \"application/pdf\"\n ext ||= \"pdf\"\n else\n ext ||= \"bin\"\n end\n\n fname.gsub!(/[^a-z0-9\\.\\_\\-]+/i, '')\n fname << \".#{ext}\"\n\n ltype.gsub!(/[^a-z0-9\\.\\_\\-]+/i, '')\n\n path = File.join(Msf::Config.local_directory, fname)\n full_path = ::File.expand_path(path)\n File.open(full_path, \"wb\") { |fd| fd.write(data) }\n\n # This will probably evolve into a new database table\n report_note(\n :data => full_path.dup,\n :type => \"#{ltype}.localpath\"\n )\n\n return full_path.dup\n end",
"def store obj, msg\n Dir.chdir(working_dir) do\n dirname = File.dirname(obj.path)\n FileUtils.mkdir_p(dirname) unless File.directory?(dirname)\n File.open(obj.path, 'w'){ |f| f.write(obj.content) }\n add(obj.path)\n end\n commit_index(msg)\n end",
"def put_raw_object(content, type)\n size = content.length.to_s\n LooseStorage.verify_header(type, size)\n \n header = \"#{type} #{size}\\0\"\n store = header + content\n \n sha1 = Digest::SHA1.hexdigest(store)\n path = @directory+'/'+sha1[0...2]+'/'+sha1[2..40]\n \n if !File.exists?(path)\n content = Zlib::Deflate.deflate(store)\n \n FileUtils.mkdir_p(@directory+'/'+sha1[0...2])\n File.open(path, 'w') do |f|\n f.write content\n end\n end\n return sha1\n end",
"def persist(data)\n\t\tupdate_design_for self.shared_data_context\n\t\t\n\t\td = CouchRest::Document.new\n\t\td.database = proxy_database\n\n\t\t#copy data in to document\n\t\tdata.each do |key,value|\n\t\t\td[\"#{key}\"] = value\n\t\t\tputs \"added: #{key}\"\n\t\tend\n\t\td[:type] = self.shared_data_context.name\n\t\td[:date] = DateTime.now\n\n\t\tbegin\n\t\t\td.save false\n\t\trescue => e\n\t\t\tputs \"NOT SAVED: #{e.response}\"\n\t\t\t#TODO: properly handle this stuff\n\t\tend#\n\tend",
"def save_meta_data(type)\n FileUtils.mkdir_p File.dirname(meta_file_path(type))\n File.open(meta_file_path(type), 'w') { |f| f.print self[type].to_yaml }\n if Mist.commit_meta_data\n Mist.repository.add meta_file_path(type)\n Mist.repository.commit '%s meta changes to %s' % [type, table_name]\n end\n\n # we must force meta to be reloaded because otherwise it could get out of sync with filesystem\n @meta = nil\n end",
"def store(type, file)\n require_writeable_layers!\n hash = store_immediately!(type, file)\n if layers.delayed.writeable.any?\n Dis::Jobs::Store.perform_later(type, hash)\n end\n hash\n end",
"def save_data(resource_type, name, data, action_handler)\n _chef_server = self.chef_server\n Chef::Provisioning.inline_resource(action_handler) do\n if resource_type == :machine\n chef_node name do\n chef_server _chef_server\n raw_json data\n end\n else\n chef_data_bag resource_type.to_s do\n chef_server _chef_server\n end\n chef_data_bag_item name do\n chef_server _chef_server\n data_bag resource_type.to_s\n raw_data data\n end\n end\n end\n\n backcompat_type = ChefManagedEntryStore.type_names_for_backcompat[resource_type]\n if backcompat_type && backcompat_type != resource_type\n delete_data(backcompat_type, name, action_handler)\n end\n end",
"def save_cached_data\n # open file for writing\n @file_handle = File.open(file_path, 'w')\n \n # write data string into file\n @file_handle << (@file_data_to_write.respond_to?(:read) ? @file_data_to_write.read : @file_data_to_write)\n \n # close file\n close_file\n \n end",
"def save_cached_data\n # open file for writing\n @file_handle = File.open(file_path, 'w')\n \n # write data string into file\n @file_handle << (@file_data_to_write.respond_to?(:read) ? @file_data_to_write.read : @file_data_to_write)\n \n # close file\n close_file\n \n end",
"def save\n raise \"Unhashed entry during save\" unless @hash\n raise \"Entry without mime type during save\" unless @mimetype\n @data[:mimetype] = @mimetype\n @data[:ctime] = @ctime.utc.to_i\n @data[:mtime] = @mtime.utc.to_i\n @data[:name] = @name\n @data[:size] = @size\n @data[:tags] = @tags\n @data[:locations] ||= Array.new\n unless @data[:locations].include? @location\n @data[:locations] << @location\n end\n if ::Ordnung.database.read @hash\n ::Ordnung.database.update @hash, @data\n else\n ::Ordnung.database.create @hash, @data\n end\n @hash\n end",
"def store_content(key, klass, content)\n raise MogileFS::ReadOnlyError if readonly?\n\n new_file key, klass do |mfp|\n if content.is_a?(MogileFS::Util::StoreContent)\n mfp.streaming_io = content\n else\n mfp << content\n end\n end\n\n content.length\n end",
"def sync\n @data_file.fsync\n end",
"def commit\n @read_lock.synchronize do\n @write_lock.synchronize do\n unless @saved\n storage.store(self.class, @id, @data.dup)\n end\n @saved = true\n end\n end\n end",
"def store(path, content, mode = 0o644)\n put_at(parse_path(path), content && repo.data_sha(content), mode)\n end",
"def refresh\n @write_lock.synchronize do\n @data = storage.retrieve(self.class, @id)\n @saved = true\n end\n end",
"def update_foreign_file_level_data(\n foreign_content_at_file,\n sync_commit,\n sts_that_require_review\n )\n existing_data = foreign_content_at_file.read_file_level_data\n already_flagged_sts = existing_data['st_sync_subtitles_to_review'] || {}\n\n new_flagged_sts = already_flagged_sts.dup\n # Convert existing data in json file from scalar String to Array of Strings.\n new_flagged_sts.each { |stid, ops_type_or_types|\n new_flagged_sts[stid] = [*ops_type_or_types]\n }\n # Add new sts that require review, or add more ops types to existing ones.\n sts_that_require_review.each { |stid, ops_types|\n new_flagged_sts[stid] ||= []\n new_flagged_sts[stid] += ops_types\n new_flagged_sts[stid].uniq!\n new_flagged_sts[stid].sort!\n }\n # TODO: Check if this clashes with #update_file_level_data! in merge\n foreign_content_at_file.update_file_level_data!(\n existing_data.merge({\n 'st_sync_commit' => sync_commit,\n 'st_sync_subtitles_to_review' => new_flagged_sts,\n })\n )\n end",
"def save_data(metadata_entity, local_path)\n $log.info 'Saving data to the BDS.'\n local_path = pack_data(local_path)\n metadata_entity.store_runtime_param('source_filename', local_path)\n metadata_entity.store_runtime_param('date_from', Time.now)\n metadata_entity.store_runtime_param('date_to', Time.now)\n @metadata.save_data(metadata_entity)\n GoodData::Connectors::Metadata::Runtime.reset_now\n end",
"def store_loot(ltype, ctype, host, data, filename=nil, info=nil, service=nil)\n if ! ::File.directory?(Msf::Config.loot_directory)\n FileUtils.mkdir_p(Msf::Config.loot_directory)\n end\n\n ext = 'bin'\n if filename\n parts = filename.to_s.split('.')\n if parts.length > 1 and parts[-1].length < 4\n ext = parts[-1]\n end\n end\n\n case ctype\n when /^text\\/[\\w\\.]+$/\n ext = \"txt\"\n end\n # This method is available even if there is no database, don't bother checking\n host = Msf::Util::Host.normalize_host(host)\n\n ws = (db ? myworkspace.name[0,16] : 'default')\n name =\n Time.now.strftime(\"%Y%m%d%H%M%S\") + \"_\" + ws + \"_\" +\n (host || 'unknown') + '_' + ltype[0,16] + '_' +\n Rex::Text.rand_text_numeric(6) + '.' + ext\n\n name.gsub!(/[^a-z0-9\\.\\_]+/i, '')\n\n path = File.join(Msf::Config.loot_directory, name)\n full_path = ::File.expand_path(path)\n File.open(full_path, \"wb\") do |fd|\n fd.write(data)\n end\n\n if (db)\n # If we have a database we need to store it with all the available\n # metadata.\n conf = {}\n conf[:host] = host if host\n conf[:type] = ltype\n conf[:content_type] = ctype\n conf[:path] = full_path\n conf[:workspace] = myworkspace\n conf[:name] = filename if filename\n conf[:info] = info if info\n conf[:data] = data if data\n\n if service and service.kind_of?(::Mdm::Service)\n conf[:service] = service if service\n end\n\n framework.db.report_loot(conf)\n end\n\n return full_path.dup\n end",
"def cache_content(type, data)\n return nil unless type.present?\n\n Rails.cache.write(type, data, expires_in: 60.minutes)\n rescue StandardError => e\n logger.error(\"Unable to add #{type} to the Rails cache: #{e}.\")\n end",
"def flush\n File.open(\"data/#{Time.now.to_i.to_s}\",'w') { |f| Marshal.dump(@cache, f) }\n File.open(\"data2/#{Time.now.to_i.to_s}\",'w') { |f| Marshal.dump(@cache, f) }\n puts \"=\"\n end",
"def save\n return if @content.nil?\n put_rest \"extra/#{@name}\", @content, :content_type => \"application/octet-stream\"\n end",
"def put_object(type, content)\n data = \"#{type} #{content.length}\\0#{content}\"\n id = sha(data)\n path = object_path(id)\n\n unless File.exists?(path)\n FileUtils.mkpath(File.dirname(path))\n open(path, 'wb') do |f|\n f.write Zlib::Deflate.deflate(data)\n end\n end\n\n id\n end",
"def update_stored_data(t = Time.now.utc)\n stored_data.delete(stored_data_last_key) unless stored_data.keys.length == 1\n stored_data[t] = showback_event.data\n save\n end",
"def stored_data; end",
"def store_content\n if @content_changed\n @s3_object = pool.default_file_store.put(bucket, storage_location_id, @content)\n set_metadata\n end\n @content_changed=false\n end",
"def save_to_storage\n if save_attachment?\n (db_file || build_db_file).data = temp_data\n db_file.save!\n self.class.update_all ['db_file_id = ?', self.db_file_id = db_file.id], ['id = ?', id]\n end\n true\n end",
"def save_sync_data(local_path, sync_data)\n\n # save our sync data\n File.open(get_sync_data_file(local_path), 'w') do |out|\n\n YAML::dump(sync_data, out)\n end\n end",
"def sync\n #debug 'content sync'\n # We're safe not testing for the 'source' if there's no 'should'\n # because we wouldn't have gotten this far if there weren't at least\n # one valid value somewhere.\n @resource.write(:content)\n end",
"def update_object_cache\n in_object = false\n objects = {}\n data = {}\n type = \"\"\n File.open(self.objects_path, \"r\").readlines.each { |line|\n line = line.strip\n if line =~ %r{^define (\\w+) .*}\n type = $1\n in_object = true\n data = {}\n if objects[type].nil?\n objects[type] = []\n end\n else\n if in_object == true\n if line == \"}\"\n in_object = false\n objects[type] << data\n else\n chunks = line.squeeze(' ').split(' ')\n data[chunks.first] = chunks[1..-1].join(' ')\n end\n end\n end\n }\n \n objects[\"contact\"].each { |contact|\n namespace = \"bluster:objects:contact:#{contact['contact_name']}\"\n contact.keys.each { |key|\n self.redis.set(\"#{namespace}:#{key}\", contact[key]) if key != \"contact_name\"\n }\n }\n \n objects[\"command\"].each { |command|\n namespace = \"bluster:objects:command:#{command['command_name']}\"\n command.keys.each { |key|\n self.redis.set(\"#{namespace}:#{key}\", command[key]) if key != \"command_name\"\n }\n }\n \n self.redis.set(\"bluster:last_update_timestamp\", File.new(self.objects_path).mtime.to_i)\n end",
"def flush\n data = {}\n end",
"def store_object(robject, options = {})\n raw_data = begin; robject.raw_data.dup; rescue TypeError; robject.raw_data; end\n\n bucket_data(robject.bucket, options[:type])[:keys][robject.key] = {\n :value => raw_data,\n :content_type => robject.content_type.dup,\n :links => robject.links.dup,\n :indexes => robject.indexes.dup,\n :meta => robject.meta.dup,\n :etag => Digest::MD5.hexdigest(raw_data.to_s),\n :last_modified => Time.now.gmtime,\n :vclock => Base64.encode64(Time.now.to_f.to_s).chomp\n }\n end",
"def save\n if !persisted? \n if @contents\n gps = EXIFR::JPEG.new(@contents).gps\n @location = Point.new(:lng => gps.longitude, :lat => gps.latitude)\n \n @contents.rewind # Reposition read location to beggining of file\n \n grid_file = Mongo::Grid::File.new(@contents.read, get_description) # Unsaved gridfs file\n\n # Store file to mongo db\n id = self.class.mongo_client.database.fs.insert_one(grid_file)\n\n @id = id.to_s\n end\n else\n file = self.class.mongo_client.database.fs.find(_id: BSON::ObjectId.from_string(@id))\n file.update_one(get_description)\n end\n end",
"def store(path ,data)\n @sync.synchronize(Sync::EX){\n if data.kind_of? Hash\n data.each do |key, value|\n if key.kind_of? Array\n store( path + key, value)\n next\n end\n key = key.to_sym\n if key == VALUE_KEY\n store( path, value )\n else\n store( path + [key], value)\n end\n end\n else\n @source[path] = data\n end\n }\n end",
"def save\n return if @blob.data == content\n repository.store(self, commit_message)\n end",
"def store(key, data, opts={})\n if data.respond_to?(:read)\n multipart_store(key, data, opts)\n else\n singlepart_store(key, data, opts)\n end\n end",
"def set(type, content)\n grit.git.put_raw_object(content, type.to_s)\n end",
"def save\n if((ActiveRDF::ConnectionPool.read_adapters.size == 1) &&\n (ActiveRDF::ConnectionPool.write_adapter == ActiveRDF::ConnectionPool.read_adapters.first))\n save_default_types # Only write the \"default\" types to the store\n else\n full_save # Do the full save operation\n end\n end",
"def store(id, data)\n storage[id] = data\n end",
"def delayed_store(type, hash)\n file = get(type, hash)\n layers.delayed.writeable.each do |layer|\n layer.store(type, hash, file)\n end\n end",
"def updateTransactions(file, type, newdata=nil, store_amount=5)\n File.open(type.to_s + \"_lock\", File::CREAT|File::RDONLY) do |lock|\n begin\n lock.flock(File::LOCK_EX)\n File.open(file, File::RDONLY|File::CREAT) do |read|\n session[type] = Array.new\n\n read.each do |line|\n obj = JSON.load(line)\n session[type].push(obj) \n end\n\n if newdata\n session[type].push newdata unless session[type].include? newdata\n end\n end\n\n #only store x amount of transactions \n session[type].delete_at 0 if session[type].length > store_amount\n\n File.open(file, 'w') do |out|\n session[type].each do |obj|\n out.puts(obj.to_json)\n end\n end\n\n ensure\n lock.flock(File::LOCK_UN)\n end\n end\nend",
"def save\n # first, get a lock for this process\n db = File.open(@fume_db, \"r+\")\n db.flock(File::LOCK_EX)\n\n # now make sure this thread is also locked\n @thread_lock.synchronize do \n \n # then, check if any changes occurred and merge them if necessary\n modified = modified?\n\n if modified\n old_entries = @entries\n \n # reload to minimize chance of overwriting anything\n load_files\n\n # add changes; additions are accepted, but conflicts have to be resolved manually\n @entries.merge! old_entries do |id, old_e, new_e|\n old_e.merge(new_e) do |attr, old_v, new_v|\n if old_v != new_v\n error_db = \"fume_db_error_#{Time.now.strftime(\"%s\")}.yaml\"\n File.open(File.join(Fume::Config[\"fume_dir\"], error_db), \"w\") do |f|\n YAML.dump(old_entries, f)\n end\n raise \"conflict for #{id}: #{attr} '#{old_v}' != '#{new_v}'\"\n else\n new_v\n end\n end\n end\n end\n\n # write entries to file\n YAML.dump(@entries, db)\n end\n\n # minimize the necessity of reloads\n @last_modified = File.ctime(@fume_db)\n\n # let go of lock\n db.flock(File::LOCK_UN)\n \n # update caches again (always necessary)\n update_caches\n end",
"def save_to_storage\n if save_attachment?\n object = self.class.bucket.objects.build(full_filename)\n\n object.content_type = content_type\n object.acl = attachment_options[:acl]\n object.content = temp_path ? File.open(temp_path) : temp_data\n object.save\n end\n true\n end",
"def save_git_content\n MergeRequest\n .where('id = ? AND COALESCE(latest_merge_request_diff_id, 0) < ?', self.merge_request_id, self.id)\n .update_all(latest_merge_request_diff_id: self.id)\n\n ensure_commit_shas\n save_commits\n save_diffs\n save\n keep_around_commits\n end",
"def data\n @data ||= file? ?\n repos.file(fs_path, revision) :\n repos.dir(fs_path, revision)\n end",
"def store_wffile\n # verify if there is actually a file to be saved\n if @file_data\n # create the WORKFLOW_STORE Folder if it does not exist\n FileUtils.mkdir_p File.join WORKFLOW_STORE, \"#{id}\"\n # create the file and write the data to the file system\n File.open(workflow_filename, 'wb') do |f|\n f.write(@file_data.read)\n end\n # ensure that the data is only save once by clearing the cache after savig\n @file_data = nil\n end\n end",
"def save_data(resource_type, name, data, action_handler)\n raise NotImplementedError, :save_data\n end",
"def put key, value\n # append value to open file\n # write new keydict value\n @keydict[key] = write_data_entry key, value\n\n # sync\n\n if @data_file.size > MAX_FILE_SIZE\n fname = @data_file.path\n @data_file.close\n @data_file = open_new_data_file\n @old_data_files[fname] = File.open(fname, \"rb\")\n end\n end",
"def store_oauth_data\n not_modified = [:access_token, :refresh_token, :expires_in, :issued_at].map do |key|\n oauth_data[key] == (oauth_data[key] = client.authorization.send(key))\n end.all?\n File.open(oauth_data_file,'w') { |f| f.puts oauth_data.to_yaml } unless not_modified\n end",
"def save\n if modified? and @entries and !@entries.empty?\n save!\n end\n end",
"def _synced\n @_synced ||= {}\n end",
"def add_replace(storage_obj)\n cache_has_key = @cache.key? storage_obj.key\n store_data = (!cache_has_key && storage_obj.command_name == ADD_CMD_NAME)\n store_data ||= (cache_has_key && storage_obj.command_name == REPLACE_CMD_NAME)\n\n if store_data\n store_new_item storage_obj.key, storage_obj.flags, storage_obj.expdate, storage_obj.length, storage_obj.data_block\n else\n NOT_STORED_MSG\n end\n end",
"def store\n # Calculate rule memory\n (@reps.to_a + @site.layouts.to_a).each do |obj|\n rule_memory_store[obj] = action_provider.memory_for(obj).serialize\n end\n\n # Calculate checksums\n objects_to_checksum =\n site.items.to_a + site.layouts.to_a + site.code_snippets + [site.config]\n objects_to_checksum.each { |obj| checksum_store.add(obj) }\n\n # Store\n checksum_store.store\n rule_memory_store.store\n end",
"def track_save save_type = :quicksave, args = {}\n files = save_type.eql?(:both) ? ['persistent.sfs', 'quicksave.sfs'] : [(save_type.eql?(:persistent) ? 'persistent' : 'quicksave') << '.sfs']\n r = self.repo\n\n within_dir(self.path) do \n files.each do |file|\n changed_file = r.changed.include?(file)\n if File.exists?(file) && (changed_file || r.untracked.include?(file))\n message = \"added #{file}\" \n message = \"updated #{file}\" if changed_file\n message = args[:message] unless args[:message].blank? \n puts message unless Rails.env.eql?(\"test\")\n r.add(file)\n r.commit(message)\n end\n end\n end\n end",
"def save\n Chef::FileCache.store(\"remote_file/#{sanitized_cache_file_basename}\", json_data)\n end",
"def save_all\n return if @loading_data\n\n json_data = JSON.dump(@model.to_h)\n\n LocalStorage['volt-store'] = json_data\n end",
"def _store(io, context)\n if (hash = io).is_a?(Hash)\n raise Error, \":location is not applicable to versions\" if context.key?(:location)\n raise Error, \"detected multiple versions that point to the same IO object: given versions: #{hash.keys}, unique versions: #{hash.invert.invert.keys}\" if hash.invert.invert != hash\n\n hash.inject({}) do |result, (name, version)|\n result.update(name => _store(version, version: name, **context))\n end\n else\n super\n end\n end",
"def save!\n # Scrub some fields\n @data[\"installed\"].sort!\n @data[\"installed\"].uniq!\n\n # Save\n @path.open(\"w+\") do |f|\n f.write(JSON.dump(@data))\n end\n end",
"def transfer_applicable_st_ops_to_foreign_file!(foreign_content_at_file, applicable_st_ops_for_file)\n if applicable_st_ops_for_file.any?\n # An st_ops_for_file exists. That means the file is being synced.\n # NOTE: Not sure why we're passing applicable_st_ops_for_file\n # as an array as there should really be only one.\n\n # Iterate over st_ops and incrementally update both content and data\n found_st_ops = false\n applicable_st_ops_for_file.each do |st_ops_for_file|\n # Detect if there are st_ops for file, or if it's time slice\n # changes only.\n found_st_ops ||= st_ops_for_file.operations.any?\n transfer_st_ops_to_foreign_file!(\n foreign_content_at_file,\n st_ops_for_file\n )\n # We have to reload the file contents as they were changed on\n # disk by #transfer_st_ops_to_foreign_file!\n foreign_content_at_file.reload_contents!\n end\n # We need to manually write the @to_git_commit to st_sync_commit.\n # We can't rely on transfer_st_ops_to_foreign_file! alone since\n # it will only write sync commits that actually contained st_ops\n # for the current file. However we want to record on the file\n # that it has been synced to the current primary st_sync_commit.\n update_foreign_file_level_data(\n foreign_content_at_file,\n @to_git_commit,\n {} # Don't touch sts that require review\n )\n if found_st_ops\n # Actual st ops\n print \" - Synced\".color(:green)\n else\n # Time slice changes only\n print \" - Synced (Time slice changes only)\".color(:green)\n end\n else\n # No applicable st ops, just update file level st_sync data\n update_foreign_file_level_data(\n foreign_content_at_file,\n @to_git_commit,\n {} # Don't touch sts that require review\n )\n print \" - No applicable st_ops\"\n end\n true\n end",
"def share_datastore(ds)\n\t\tself.datastore = ds\n\t\tself.datastore.import_options(self.options)\n\tend",
"def _flush_cache(*types)\n if types.size == 0 or types.include?(:hash)\n @cache_hash = { }\n @hash_times = Hash.new(0)\n end\n\n if types.size == 0 or types.include?(:file)\n @file_times = Hash.new(0)\n @file_cache = { }\n end\n self\n end",
"def save_stored_data(dir)\n return unless exist?\n file = File.join(dir, \"#{name}.#{Gem::Version.new(ole.Version)}.cfe\")\n ole.GetData.Write(real_win_path(file))\n file\n end",
"def store(io, context = {})\n _store(io, context)\n end",
"def files_to_final_location\n if @temp_file && (@temp_file.size > 0)\n logger.info(\"Saving attachment '#{self.filename}' (#{@temp_file.size} bytes) to database\")\n md5 = Digest::MD5.new\n buffer = \"\"\n self.data = \"\"\n while (buffer = @temp_file.read(8192))\n md5.update(buffer)\n self.data << buffer\n end\n self.digest = md5.hexdigest\n end\n @temp_file = nil\n # Don't save the content type if it's longer than the authorized length\n if self.content_type && self.content_type.length > 255\n self.content_type = nil\n end\n end",
"def sync_data(data)\n begin\n operation = create_operation(data)\n sync_categories(operation) if operation.kind\n\n rescue Exception => e\n log_data_invalid_error(data, e)\n end\n end",
"def save_processed_data\n attachment.update(processed_data: json_parser.final_hash)\n end",
"def save(data=cached_data)\n lock do\n File.safe_write(path, JSON.pretty_generate(data))\n end\n end",
"def add_object(name, data, content_type='binary/octet-stream', *args)\n data.rewind\n additional_headers = {\n 'Content-MD5' => Base64.encode64(Digest::MD5.digest(data.read)).strip,\n 'Content-Type' => content_type\n }\n additional_headers.merge!(args.pop) if args.last.is_a? Hash\n connection.put(\"/#{URI.escape(name)}\", @name, data, additional_headers)\n end",
"def save_fabrication(type, time)\n return unless Thread.current.key?(type)\n return unless top_level_fabrication?\n\n Thread.current[type] += time\n end",
"def global_merge_data\n @global_merge_data\n end",
"def store\n # Calculate rule memory\n (@reps.to_a + @site.layouts.to_a).each do |obj|\n rule_memory_store[obj] = action_provider.memory_for(obj).serialize\n end\n\n # Calculate checksums\n objects_to_checksum =\n site.items.to_a + site.layouts.to_a + site.code_snippets + [site.config]\n objects_to_checksum.each { |obj| checksum_store.add(obj) }\n\n # Store\n stores.each(&:store)\n end",
"def store\n object = bucket.objects.build(final_file)\n object.content = open(File.join(tmp_path, final_file))\n object.save\n end",
"def write_file_after_save(file_data_to_write=nil)\n # check if there are data to write\n return unless(@file_data_to_write)\n \n begin\n self.class.benchmark(\"\\033[36m\\033[1m\\033[4mFileStore\\033[0m Saving file for #{self.id}\") do\n # create data directory path\n FileUtils.mkdir_p(data_directory)\n \n if(@file_data_to_write.is_a?(DataPath))\n copy_data_file\n else\n save_cached_data\n end\n \n @file_data_to_write = nil\n end\n rescue Exception => e\n assit_fail(\"Exception on writing file #{self.location}: #{e}\")\n end\n\n end",
"def add(path, data, ctype=DEFAULT_CTYPE)\n # FIXME: determine if ADD or UPDATE EVENT\n # evt = File.exist? @content_tree.node_path(path)\n # FIXME: should this always be create-or-update? what about replace=false?\n n = @content_tree.add(path, data, ctype)\n notify(EVENT_ADD, path, ctype)\n n\n end",
"def save\n MiGA.DEBUG \"Metadata.save #{path}\"\n self[:updated] = Time.now.to_s\n json = JSON.pretty_generate(data)\n sleeper = 0.0\n while File.exist?(lock_file)\n sleeper += 0.1 if sleeper <= 10.0\n sleep(sleeper.to_i)\n end\n FileUtils.touch lock_file\n ofh = File.open(\"#{path}.tmp\", \"w\")\n ofh.puts json\n ofh.close\n raise \"Lock-racing detected for #{path}.\" unless\n File.exist?(\"#{path}.tmp\") and File.exist?(lock_file)\n File.rename(\"#{path}.tmp\", path)\n File.unlink(lock_file)\n end",
"def store\n @files.each do |file, lines|\n text = \"\"\n dirty = false\n lines.each do |l|\n if l.is_a?(Section)\n dirty ||= l.dirty?\n text << l.format\n l.mark_clean\n else\n text << l\n end\n end\n if dirty\n Puppet::Util::FileType.filetype(:flat).new(file).write(text)\n return file\n end\n end\n end",
"def store_file(data, filename)\r\n ltype = \"exploit.fileformat.#{self.shortname}\"\r\n\r\n if ! ::File.directory?(Msf::Config.local_directory)\r\n FileUtils.mkdir_p(Msf::Config.local_directory)\r\n end\r\n\r\n if filename and not filename.empty?\r\n if filename =~ /(.*)\\.(.*)/\r\n ext = $2\r\n fname = $1\r\n else\r\n fname = filename\r\n end\r\n else\r\n fname = \"local_#{Time.now.utc.to_i}\"\r\n end\r\n\r\n fname = ::File.split(fname).last\r\n\r\n fname.gsub!(/[^a-z0-9\\.\\_\\-]+/i, '')\r\n fname << \".#{ext}\"\r\n\r\n path = File.join(\"#{Msf::Config.local_directory}/\", fname)\r\n full_path = ::File.expand_path(path)\r\n File.open(full_path, \"wb\") { |fd| fd.write(data) }\r\n\r\n full_path.dup\r\n end",
"def record_data_for_later_saving(data, definition)\n word = data[:word]\n word_index = data[:index]\n\n # Add current definition's see_also list to map for processing later\n @relateds[definition] = data[:see_also]\n\n # Add definition to map of word obj. -> definition objects.\n # This mapping will be used to link all related definitions after they've all been created\n @word_definitions[word.downcase] ||= []\n @word_definitions[word.downcase][word_index] = definition\n end",
"def save_additional_data\n end",
"def edit_data(&block)\n lock do\n @cached_data = load_data_from_file\n block.call(self)\n File.safe_write(path, JSON.pretty_generate(@cached_data))\n @cached_data\n end\n end",
"def merge(content_data)\n content_data.contents.values.each { |content|\n add_content(content)\n }\n content_data.instances.values.each { |instance|\n add_instance(instance)\n }\n end",
"def marshal_save\n donnee_totale.merge!(updated_at: Time.now.to_i)\n # log \"DATA FILM ENREGISTRÉES (film#donnee_totale) :#{RC}#{donnee_totale.inspect}\"\n File.open(marshal_file,'wb'){|f| f.write Marshal.dump(donnee_totale)}\n end",
"def getContentTree\n #N Without this we won't have timestamp and the map of file hashes used to efficiently determine the hash of a file which hasn't been modified after the timestamp\n cachedTimeAndMapOfHashes = getCachedContentTreeMapOfHashes\n #N Without this we won't have the timestamp to compare against file modification times\n cachedTime = cachedTimeAndMapOfHashes[0]\n #N Without this we won't have the map of file hashes\n cachedMapOfHashes = cachedTimeAndMapOfHashes[1]\n #N Without this we won't have an empty content tree which can be populated with data describing the files and directories within the base directory\n contentTree = ContentTree.new()\n #N Without this we won't have a record of a time which precedes the recording of directories, files and hashes (which can be used when this content tree is used as a cached for data when constructing some future content tree)\n contentTree.time = Time.now.utc\n #N Without this, we won't record information about all sub-directories within this content tree\n for subDir in @baseDirectory.subDirs\n #N Without this, this sub-directory won't be recorded in the content tree\n contentTree.addDir(subDir.relativePath)\n end\n #N Without this, we won't record information about the names and contents of all files within this content tree\n for file in @baseDirectory.allFiles\n #N Without this, we won't know the digest of this file (if we happen to have it) from the cached content tree\n cachedDigest = cachedMapOfHashes[file.relativePath]\n #N Without this check, we would assume that the cached digest applies to the current file, even if one wasn't available, or if the file has been modified since the time when the cached value was determined.\n # (Extra note: just checking the file's mtime is not a perfect check, because a file can \"change\" when actually it or one of it's enclosing sub-directories has been renamed, which might not reset the mtime value for the file itself.)\n if cachedTime and cachedDigest and File.stat(file.fullPath).mtime < cachedTime\n #N Without this, the digest won't be recorded from the cached digest in those cases where we know the file hasn't changed\n digest = cachedDigest\n else\n #N Without this, a new digest won't be determined from the calculated hash of the file's actual contents\n digest = hashClass.file(file.fullPath).hexdigest\n end\n #N Without this, information about this file won't be added to the content tree\n contentTree.addFile(file.relativePath, digest)\n end\n #N Without this, the files and directories in the content tree might be listed in some indeterminate order\n contentTree.sort!\n #N Without this check, a new version of the cached content file will attempt to be written, even when no name has been specified for the cached content file\n if cachedContentFile != nil\n #N Without this, a new version of the cached content file (ready to be used next time) won't be created\n contentTree.writeToFile(cachedContentFile)\n end\n return contentTree\n end",
"def update_content_entry(content_type, content_entry)\n locale = Locomotive::Mounter.locale\n\n # log before\n self.output_resource_op content_entry\n\n # get the params\n params = self.buffer_log { self.content_entry_to_params(content_entry) }\n\n # send the request\n response = self.put \"content_types/#{content_type}/entries\", content_entry._id, params, locale\n\n status = self.response_to_status(response)\n\n # log after\n self.output_resource_op_status content_entry, status\n self.flush_log_buffer\n end",
"def storage; end",
"def update_global_statistic(new_data)\n path = \"#{@results_dir_path}/global_statistic.json\"\n old_data = ensure_load_json(path, {}, symbolize_names: true)\n File.write(path, JSON.pretty_generate(old_data.merge(new_data)))\n end",
"def write_if_empty\n return if cached_content.present?\n\n @diff_collection.diff_files.each do |diff_file|\n next unless cacheable?(diff_file)\n\n diff_file_id = diff_file.file_identifier\n\n cached_content[diff_file_id] = diff_file.highlighted_diff_lines.map(&:to_hash)\n end\n\n cache.write(key, cached_content, expires_in: 1.week)\n end",
"def set_content\n unless compare_content\n description = []\n description << \"update content in file #{@new_resource.path} from #{short_cksum(@current_resource.checksum)} to #{short_cksum(new_resource_content_checksum)}\"\n description << diff_current_from_content(@new_resource.content) \n converge_by(description) do\n backup @new_resource.path if ::File.exists?(@new_resource.path)\n ::File.open(@new_resource.path, \"w\") {|f| f.write @new_resource.content }\n Chef::Log.info(\"#{@new_resource} contents updated\")\n end\n end\n end",
"def full_save\n types.each do |t|\n ActiveRDF::FederationManager.add(self, N::RDF::type, t)\n end\n\n ActiveRDF::Query.new(N::URI).distinct(:p,:o).where(self, :p, :o).execute do |p, o|\n ActiveRDF::FederationManager.add(self, p, o)\n end\n end",
"def save_content_type_and_size\n model.content_type = file.content_type if model.respond_to?(:content_type) && file.content_type\n model.file_size = file.size if model.respond_to?(:file_size)\n end",
"def save_data(file, data)\n if ENV['RACK_ENV'] == 'test'\n local = File.expand_path(\"../test/data/#{file}\", __FILE__)\n File.open(local, 'w') { |open_file| open_file.write(YAML.dump(data)) }\n else\n local = File.expand_path(\"../data/#{file}\", __FILE__)\n File.open(local, 'w') { |open_file| open_file.write(YAML.dump(data)) }\n\n if USE_GOOGLE_DRIVE\n remote = google_session.file_by_title(file.to_s)\n remote.update_from_file(local)\n end\n end\nend",
"def _store(io, context)\n _enforce_io(io)\n context[:location] ||= get_location(io, context)\n context[:metadata] ||= get_metadata(io, context)\n\n put(io, context)\n\n self.class::UploadedFile.new(\n \"id\" => context[:location],\n \"storage\" => storage_key.to_s,\n \"metadata\" => context[:metadata],\n )\n end",
"def log_content_type(type)\n File.open(Path + \"/content.log\", \"a\") {|f| f << type + \"\\n\"}\nend",
"def save(data)\n File.open(@local_file_path, \"wb\") {|file| file.write(data) }\n end",
"def save!; File.write @path, @data end",
"def persist_info\n file = VER.loadpath.first / 'buffer_info.json'\n l \"Persisting Buffer info into: #{file}\"\n\n JSON::Store.new(file.to_s, true).transaction do |buffer_info|\n syntax_name = @syntax.name if @syntax\n\n buffer_info[uri.to_s] = {\n 'insert' => index('insert').to_s,\n 'syntax' => syntax_name\n }\n end\n end",
"def append_content(path, content)\n \n puts \"Sending path and content via MCollective Files client\"\n @mc.append(:path => path, :content => content)\n printrpcstats\n \n end",
"def add storage\n Storage.new (storage.data+@data), @length\n end"
] |
[
"0.5787639",
"0.57433087",
"0.5566787",
"0.55094874",
"0.54253364",
"0.5327555",
"0.53268236",
"0.5261412",
"0.5210791",
"0.51719934",
"0.51689905",
"0.51689905",
"0.51615727",
"0.5160665",
"0.5109377",
"0.5072805",
"0.505124",
"0.50323355",
"0.5016275",
"0.49901885",
"0.4990059",
"0.49844044",
"0.49805334",
"0.4943894",
"0.49039742",
"0.49014518",
"0.49012583",
"0.4895492",
"0.48808238",
"0.48625642",
"0.48609123",
"0.485693",
"0.48532134",
"0.48488855",
"0.48435852",
"0.48361817",
"0.48328945",
"0.48150736",
"0.4813205",
"0.48057505",
"0.48024228",
"0.47955862",
"0.47899204",
"0.47813892",
"0.47723764",
"0.47684607",
"0.4746923",
"0.47412208",
"0.4737515",
"0.47289607",
"0.47195807",
"0.47028875",
"0.4700071",
"0.46979663",
"0.46976882",
"0.46970382",
"0.46962923",
"0.46943286",
"0.46914592",
"0.46907744",
"0.4689572",
"0.4682031",
"0.4681566",
"0.46771023",
"0.4658409",
"0.46577716",
"0.46576574",
"0.46554098",
"0.4653556",
"0.46533614",
"0.46490914",
"0.46490797",
"0.46485275",
"0.4644507",
"0.46442813",
"0.46372068",
"0.4635209",
"0.46311387",
"0.46244556",
"0.46232846",
"0.46229404",
"0.46186668",
"0.46158952",
"0.46055198",
"0.46004242",
"0.46001527",
"0.45982784",
"0.45922294",
"0.458909",
"0.45875338",
"0.4582591",
"0.45824993",
"0.4579722",
"0.4571467",
"0.4571463",
"0.45606032",
"0.45592228",
"0.45587072",
"0.45544285",
"0.4549472"
] |
0.5880467
|
0
|
Save content data file in the DB. ContentData DB directories created in the lazy manner, so if they are still absent, they will be created during add content data operation.
|
def save(content_data, type, from, till)
filename = DiffFile.compose_filename(type, from, till)
path = case type
when DiffFile::SNAPSHOT_TYPE
File.join(@snapshots_path, till.year.to_s, filename)
when DiffFile::ADDED_TYPE, DiffFile::REMOVED_TYPE
File.join(@diffs_path, till.year.to_s, filename)
else
fail ArgumentError, "Unrecognized type: #{type}"
end
dirname = File.dirname(path)
unless Dir.exist?(dirname)
FileUtils.mkdir_p dirname
end
content_data.to_file(path)
filename
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def save()\n File.write(@database_file, @data.to_json)\n end",
"def save_to_storage\n if save_attachment?\n (db_file || build_db_file).data = temp_data\n db_file.save!\n self.class.update_all ['db_file_id = ?', self.db_file_id = db_file.id], ['id = ?', id]\n end\n true\n end",
"def save\n \n transaction do |transaction|\n \n check_content_type! if self.content_type # Update the content type\n check_categories! if self.categories and not self.categories.empty? # Update the categories\n \n begin \n super # Invokes the super class \n rescue DataMapper::SaveFailureError => error\n p \"Error saving content #{error} #{self.errors.inspect}\"\n raise error \n end\n\n transaction.commit\n\n end\n\n end",
"def write_content\n File.open(absolute_path,'w') do |file|\n file << content if content\n end\n # TODO git functionality\n end",
"def store(path, content, mode = 0o644)\n put_at(parse_path(path), content && repo.data_sha(content), mode)\n end",
"def save_file(path, content)\n FileUtils.mkdir_p(File.expand_path(\"..\", path))\n File.open(path, \"w\") do |f|\n f.write(content)\n end\n end",
"def save_cached_data\n # open file for writing\n @file_handle = File.open(file_path, 'w')\n \n # write data string into file\n @file_handle << (@file_data_to_write.respond_to?(:read) ? @file_data_to_write.read : @file_data_to_write)\n \n # close file\n close_file\n \n end",
"def save_cached_data\n # open file for writing\n @file_handle = File.open(file_path, 'w')\n \n # write data string into file\n @file_handle << (@file_data_to_write.respond_to?(:read) ? @file_data_to_write.read : @file_data_to_write)\n \n # close file\n close_file\n \n end",
"def save\n repository.create_contents(path, \"Upload #{path}\", content)\n end",
"def save_file\n if platinum_user_and_above?\n uid = current_user.id\n data_dir = Rails.root.join('shared','data')\n @file = data_dir.join('domains')\n file = File.open(@file, 'r')\n @restore = ''\n file.each_line { |line| @restore += line }\n file.close\n file = File.open(@file, 'w+')\n file.write(params[:file_content])\n file.close\n domain_table_reload(uid,data_dir.to_s)\n render json: { message: 'Saving successful.' }\n else\n render json: { message: 'Saving failed, please check your file again.' }\n end\n rescue Psych::SyntaxError\n file = File.open(@file, 'w+')\n file.write(@restore)\n file.close\n render json: { message: 'Saving failed, please check your file again.' }\n end",
"def save_to(path, content) \n begin \n dir = File.dirname(path)\n # Create a new directory (does nothing if directory exists or is a file)\n FileUtils.mkdir_p dir #unless File.dirname(path) == \".\"\n \n open(path, \"w\") do |f|\n f.write(content)\n end \n\n rescue StandardError => e\n puts e\n end\n end",
"def save\n return if @content.nil?\n put_rest \"extra/#{@name}\", @content, :content_type => \"application/octet-stream\"\n end",
"def save!; File.write @path, @data end",
"def save\n ApplicationDatabase.new.save_compressed_to_file(@file.path)\n end",
"def save_as_attachment(content)\n file_name = \"#{self.uid}.json\"\n\n # Use Tempfile - so we can handle large amounts of data.\n tmp = File.new(\"tmp/\" + file_name, \"w\")\n tmp << content\n tmp.flush\n\n self.attachment_file_name = file_name\n self.update(:attachment => tmp)\n\n # Make sure the tmp file is deleted.\n File.delete(tmp)\n end",
"def write_file_after_save(file_data_to_write=nil)\n # check if there are data to write\n return unless(@file_data_to_write)\n \n begin\n self.class.benchmark(\"\\033[36m\\033[1m\\033[4mFileStore\\033[0m Saving file for #{self.id}\") do\n # create data directory path\n FileUtils.mkdir_p(data_directory)\n \n if(@file_data_to_write.is_a?(DataPath))\n copy_data_file\n else\n save_cached_data\n end\n \n @file_data_to_write = nil\n end\n rescue Exception => e\n assit_fail(\"Exception on writing file #{self.location}: #{e}\")\n end\n\n end",
"def save_to_file(path = nil)\n content\n file = Store::File.find_by(id: store_file_id)\n if !file\n raise \"No such file #{store_file_id}!\"\n end\n\n if !path\n path = Rails.root.join('tmp', filename)\n end\n ::File.open(path, 'wb') do |handle|\n handle.write file.content\n end\n path\n end",
"def content=(content)\n if GalleryConfig.storage.database_notebooks\n notebookFile = NotebookFile.find_or_initialize_by(save_type: \"stage\", uuid: uuid)\n notebookFile.stage_id = id\n notebookFile.content = content\n notebookFile.save\n else\n File.write(filename, content)\n end\n end",
"def store_wffile\n # verify if there is actually a file to be saved\n if @file_data\n # create the WORKFLOW_STORE Folder if it does not exist\n FileUtils.mkdir_p File.join WORKFLOW_STORE, \"#{id}\"\n # create the file and write the data to the file system\n File.open(workflow_filename, 'wb') do |f|\n f.write(@file_data.read)\n end\n # ensure that the data is only save once by clearing the cache after savig\n @file_data = nil\n end\n end",
"def update_db(cont)\n db = File.open('./database', 'w')\n db.write(Marshal.dump(cont))\n db.close\n end",
"def save_file(path, content)\n File.open(path, 'w') do |f|\n f.write content\n end\n end",
"def save(name = @name)\n raise OneCfg::Config::Exception::NoContent if @content.nil?\n\n file_operation(name, 'w') {|file| file.write(to_s) }\n end",
"def store_content(key, klass, content)\n raise MogileFS::ReadOnlyError if readonly?\n\n new_file key, klass do |mfp|\n if content.is_a?(MogileFS::Util::StoreContent)\n mfp.streaming_io = content\n else\n mfp << content\n end\n end\n\n content.length\n end",
"def save\n pathname.open('w') { |file| file.write(data) }\n end",
"def write(key, content)\n temp_file = File.join(root, ['tmp', $$, Thread.current.unique_id].join('-'))\n File.open(temp_file, 'wb') do |dest|\n if content.respond_to? :to_str\n dest.write(content.to_str)\n else\n content.each {|s| dest.write(s) }\n end\n end\n\n path = cache_path(key)\n if File.exist?(path)\n File.unlink temp_file\n else\n FileUtils.mkdir_p File.dirname(path), :mode => 0755\n FileUtils.mv temp_file, path\n end\n true\n rescue\n File.unlink temp_file rescue false\n false\n end",
"def persist!\n ::File.write(self.path, Marshal.dump(self))\n rescue => e\n puts e.message\n exit\n end",
"def persist\n if current_file.new_record?\n file_set.save\n else\n current_file.save\n end\n end",
"def write content, *opts\n raise \"Will not overwrite: #{self}\" if File.exist? self and not opts.include? :force\n FileUtils.mkdir_p(self.dirname)\n File.open(self, 'w'){ |f| f.write(content) }\n end",
"def write_to_file(path, content)\n directory = File.dirname(path)\n FileUtils.mkdir_p(directory)\n File.write(path, content)\n after_rendering_run(\"rm -rf #{path}\")\n path\n end",
"def persistData!\n File.open(getWorkingDir+'/data.host', 'w') do |f|\n f.write(@plataforms.to_yaml)\n end\n end",
"def store_content\n if @content_changed\n @s3_object = pool.default_file_store.put(bucket, storage_location_id, @content)\n set_metadata\n end\n @content_changed=false\n end",
"def save\n return if @filename.nil?\n FileUtils.mkdir_p File.dirname(@filename)\n Utils.atomic_write(@filename) { |f| f.write(JSON.generate(@data)) }\n end",
"def store_content(world, scenario)\n # Make sure the content directory exists.\n basedir = File.join(Dir.pwd, 'content')\n FileUtils.mkdir_p(basedir)\n\n # Store content. Note that not all drivers may support this.\n filename = File.join(basedir, base_filename(scenario))\n filename += '.txt'\n\n File.open(filename, 'w') do |file|\n file.write(world.driver.page_source)\n end\n end",
"def save_stored_data(dir)\n return unless exist?\n file = File.join(dir, \"#{name}.#{Gem::Version.new(ole.Version)}.cfe\")\n ole.GetData.Write(real_win_path(file))\n file\n end",
"def save\n # first, get a lock for this process\n db = File.open(@fume_db, \"r+\")\n db.flock(File::LOCK_EX)\n\n # now make sure this thread is also locked\n @thread_lock.synchronize do \n \n # then, check if any changes occurred and merge them if necessary\n modified = modified?\n\n if modified\n old_entries = @entries\n \n # reload to minimize chance of overwriting anything\n load_files\n\n # add changes; additions are accepted, but conflicts have to be resolved manually\n @entries.merge! old_entries do |id, old_e, new_e|\n old_e.merge(new_e) do |attr, old_v, new_v|\n if old_v != new_v\n error_db = \"fume_db_error_#{Time.now.strftime(\"%s\")}.yaml\"\n File.open(File.join(Fume::Config[\"fume_dir\"], error_db), \"w\") do |f|\n YAML.dump(old_entries, f)\n end\n raise \"conflict for #{id}: #{attr} '#{old_v}' != '#{new_v}'\"\n else\n new_v\n end\n end\n end\n end\n\n # write entries to file\n YAML.dump(@entries, db)\n end\n\n # minimize the necessity of reloads\n @last_modified = File.ctime(@fume_db)\n\n # let go of lock\n db.flock(File::LOCK_UN)\n \n # update caches again (always necessary)\n update_caches\n end",
"def save_to_file\n f = File.open(\"#{IMAGE_DATA_DIR}/#{self.id}\",\"w\")\n f.write(self.data)\n f.close\n end",
"def save\n create_ok = exists? ? true : create\n upload_ok = @unwritten_contents ? upload_new_content : true\n\n create_ok && upload_ok\n end",
"def save\n if !persisted? \n if @contents\n gps = EXIFR::JPEG.new(@contents).gps\n @location = Point.new(:lng => gps.longitude, :lat => gps.latitude)\n \n @contents.rewind # Reposition read location to beggining of file\n \n grid_file = Mongo::Grid::File.new(@contents.read, get_description) # Unsaved gridfs file\n\n # Store file to mongo db\n id = self.class.mongo_client.database.fs.insert_one(grid_file)\n\n @id = id.to_s\n end\n else\n file = self.class.mongo_client.database.fs.find(_id: BSON::ObjectId.from_string(@id))\n file.update_one(get_description)\n end\n end",
"def sync\n #debug 'content sync'\n # We're safe not testing for the 'source' if there's no 'should'\n # because we wouldn't have gotten this far if there weren't at least\n # one valid value somewhere.\n @resource.write(:content)\n end",
"def save(filename, content)\n\t\t# FIXME - if the file exists, this should bail out\n\t\t\n\t\t# write the contents into the file\n\t\tfile = dir + '/' + filename + '.md'\n\t\tf = File.new(file, \"w\")\n\t\tf.write(content)\n\t\tf.close\n\t\t\n\t\t# return the new file\n\t\treturn page(filename)\n\tend",
"def save(filename, content)\n\t\t# FIXME - if the file exists, this should bail out\n\t\t\n\t\t# write the contents into the file\n\t\tfile = dir + '/' + filename + '.md'\n\t\tf = File.new(file, \"w\")\n\t\tf.write(content)\n\t\tf.close\n\t\t\n\t\t# return the new file\n\t\treturn page(filename)\n\tend",
"def add_content content\n return unless content\n self.transaction do\n content = content.versions.latest if Content === content\n raise ArgumentError, \"Expected Content or Content::Version\" unless Content::Version === content\n\n save! unless self.id\n\n # Dont add if it's already been added.\n return if content_versions.find(:first, :conditions => [ 'content_id = ?', content ])\n\n connection.\n execute(\"DELETE FROM revision_list_contents \n WHERE revision_list_id = #{self.id}\n AND content_version_id IN\n (SELECT id FROM content_versions WHERE content_id = #{content.content.id})\"\n )\n revision_list_contents.create!(:content_version => content)\n\n # Invalidate association caches.\n revision_list_contents.reset\n content_versions.reset\n end\n self\n end",
"def write(persistable)\n if persistable.persistence_data\n return directory.files.create(:key => persistable.persistence_key, :body => persistable.persistence_data)\n end\n return false\n end",
"def save_content(record)\n\n\tid = record[\"id\"]\n\ttitle = record[\"title\"]\n\tdesc = record[\"desc\"]\n\turl = record[\"url\"]\n\tpath = \"\"\n\t\n\tbegin\n\t\tdb = Mysql.real_connect('localhost', 'ashish', 'ashish', 'content')\n#\t puts \"Server version: \" + db.get_server_info\n\t db.query(\"INSERT INTO data_image ( `id`, `title`,`url`,`path`,`desc` ) VALUES (#{id},'#{db.escape_string(title)}','#{db.escape_string(url)}','#{db.escape_string(path)}','#{db.escape_string(desc)}') \")\n\trescue Mysql::Error => e\n\t puts \"Error code: #{e.errno}\"\n\t puts \"Error message: #{e.error}\"\n\t puts \"Error SQLSTATE: #{e.sqlstate}\" if e.respond_to?(\"sqlstate\")\n\tend\t\nend",
"def commit\n @read_lock.synchronize do\n @write_lock.synchronize do\n unless @saved\n storage.store(self.class, @id, @data.dup)\n end\n @saved = true\n end\n end\n end",
"def update_file_content(new_content)\n return false if new_content.nil?\n self.file_content = new_content\n return self.save\n end",
"def create_document(name, content = \"\")\n File.open(File.join(data_path, name), \"w\") do |file|\n file.write(content)\n end\n end",
"def create_document(name, content = \"\")\n File.open(File.join(data_path, name), \"w\") do |file|\n file.write(content)\n end\n end",
"def persist(data)\n\t\tupdate_design_for self.shared_data_context\n\t\t\n\t\td = CouchRest::Document.new\n\t\td.database = proxy_database\n\n\t\t#copy data in to document\n\t\tdata.each do |key,value|\n\t\t\td[\"#{key}\"] = value\n\t\t\tputs \"added: #{key}\"\n\t\tend\n\t\td[:type] = self.shared_data_context.name\n\t\td[:date] = DateTime.now\n\n\t\tbegin\n\t\t\td.save false\n\t\trescue => e\n\t\t\tputs \"NOT SAVED: #{e.response}\"\n\t\t\t#TODO: properly handle this stuff\n\t\tend#\n\tend",
"def save(message)\n Log.debug(\"Database save: #{message}\")\n db_add_entry(message)\n\n if message.is_a? Evesync::IPC::Data::File\n Log.debug(\"Database save file action: #{message.action}\")\n unless message.action ==\n IPC::Data::File::Action::DELETE\n save_file(message)\n end\n end\n true\n end",
"def save_to(path)\n unless File.exists?(File.dirname(path))\n raise ArgumentError, \"Output path does not exist!\"\n end\n if File.directory?(path)\n raise ArgumentError, \"Output path should be a file!\"\n end\n File.open(path, 'w') { |f| f.write(@content) }\n end",
"def content_data=(value)\n @content_data = value\n end",
"def content_data=(value)\n @content_data = value\n end",
"def save(data)\n File.open(@local_file_path, \"wb\") {|file| file.write(data) }\n end",
"def write_file(path, content)\n dir = File.dirname(path)\n if !File.exist?(dir)\n FileUtils.mkdir_p(dir)\n end\n File.open(path, 'wb') do |f|\n f.write(content)\n end\n end",
"def store\n File.open(@file_name, 'w') do |file|\n file.write YAML::dump(@data)\n end\n end",
"def save\n return if @content.nil?\n put_rest \"rules/#{@name}\", @content, :content_type => \"application/octet-stream\"\n end",
"def create\n File.open(@db_file, \"w\" ) do |file|\n end\n end",
"def update_file_content(new_content)\n return false if new_content.nil?\n self.file_content = new_content\n self.save\n end",
"def update_file_content(new_content)\n return false if new_content.nil?\n self.file_content = new_content\n self.save\n end",
"def write_content(file)\n (content = property(:content)) && content.write(file)\n end",
"def save\n generate_archive(generate_content_types, generate_rels, generate_document)\n end",
"def persistUserContentToFile(dataDir)\n threadForEachUser do |account|\n #account.contentMap.keys.each do |tag|\n #userFile = @dataDir+\"other/#{account.user}.yml\"\n # if tag.match(/programming/)\n # userFile = @dataDir+\"programming/#{account.user}.yml\"\n # elsif tag.match(/travel/)\n # userFile = @dataDir+\"travel/#{account.user}.yml\"\n # end\n userFile = dataDir+\"/#{account.user}.yml\"\n # Remove user file if already present.\n File.delete(userFile) if File.exists?(userFile)\n File.open(userFile,\"a\") do |outputFile|\n puts \"Persisting content for '#{account.user}'\"\n #outputFile.puts account.contentMap[tag].to_yaml \n outputFile.puts account.contentArr.to_yaml\n end\n #end\n end\n end",
"def save\n return if File.exists?(file)\n\n # Create parent directories\n FileUtils.mkdir_p(File.dirname(file))\n\n File.open(file, \"w\") do |f|\n f.write(compressed_contents)\n end\n\n puts \"Wrote blob #{file}\"\n end",
"def create\n @content = Content.new(params[:content])\n file_size = params[:content][:content_file_file_size]\n\n respond_to do |format|\n if @content.save\n @content_saved = true\n \n @content.user = current_user\n current_user.contents << @content\n \n format.html { redirect_to @content}\n format.json { render json: @content, status: :created, location: @content }\n else\n @content_saved = false\n \n if file_size == nil\n @file_nil = true\n elsif file_size > 1000000\n @file_too_big = true \n end \n \n format.html { redirect_to root_path, notice: \"Content could not be saved - please pick a jpg or png file less than 1 MB\" }\n format.json { render json: @content.errors, status: :unprocessable_entity }\n end\n end\n end",
"def save\n return if @blob.data == content\n repository.store(self, commit_message)\n end",
"def save_data(metadata_entity, local_path)\n $log.info 'Saving data to the BDS.'\n local_path = pack_data(local_path)\n metadata_entity.store_runtime_param('source_filename', local_path)\n metadata_entity.store_runtime_param('date_from', Time.now)\n metadata_entity.store_runtime_param('date_to', Time.now)\n @metadata.save_data(metadata_entity)\n GoodData::Connectors::Metadata::Runtime.reset_now\n end",
"def write_file(content, path)\n File.delete path if File.exists? path\n\n file = File.new path, 'w+'\n file.write content\n\n file.close\n end",
"def add_content_to_file(file_path, content)\n names = file_path.split('/')\n file_name = names.pop\n directory = self.mkdir(names.join('/'))\n directory.add_content_to_file(file_name, content)\n end",
"def save\n File.open(file_path, 'w') do |file|\n YAML.dump(data, file)\n end\n end",
"def prepare_db_data( f, cl, tbd, df )\n s = File::Stat.new( f )\n if ! s\n puts \"couldn't stat #{f}\\n\"\n next\n end\n\n # grab extension\n m = /(\\.\\w+)$/.match( f )\n if m && m[1]\n # yes it's redundant, but this way, if the file is outside of it's directory i have half a chance of knowing what it is\n new_pathfile = s.mtime.strftime( \"%m/%d/%m%d%H%M\" ) + m[1]\n else \n puts \"couldn't find file extension for #{f}\\n\"\n next\n end\n\n md5_checksum = Digest::MD5.hexdigest( File.read( f ) )\n\n # make directories if needed\n testfile = tbd + s.mtime.strftime( \"/%m\" )\n if ! File.exists?( testfile )\n Dir.mkdir( testfile )\n end\n\n testfile += s.mtime.strftime( \"/%d\" )\n if ! File.exists?( testfile )\n Dir.mkdir( testfile )\n end\n\n # copy file to new location\n FileUtils.copy( f, \"#{tbd}/\" + new_pathfile )\n\n # save data for db push\n df.push( { :class => cl, :created_text_date => s.mtime, :created_epoch_seconds => s.mtime.to_i, :pathfile => new_pathfile, :md5_checksum => md5_checksum } )\nend",
"def save_to_sites\n if Site.where(:id => self.site_id).count == 0\n p self\n return\n end\n STDERR.puts \"save_to_sites: Site: #{site}\"\n site_dir = Rails.root.join(\"sites\", site.name)\n unless Dir.exists? site_dir\n Dir.mkdir site_dir\n end\n payload_dir = site_dir.join(self.name)\n unless Dir.exists? payload_dir\n Dir.mkdir payload_dir\n end\n STDERR.puts \"Writing: #{payload_dir}\"\n File.open(payload_dir.join(\"client_script.rb\"), \"wb\") do |f|\n f.write self.client_script\n end \n File.open(payload_dir.join(\"data_generator.rb\"), \"wb\") do |f|\n f.write self.data_generator\n end \n end",
"def save(host, uri, content)\n end",
"def files_to_final_location\n if @temp_file && (@temp_file.size > 0)\n logger.info(\"Saving attachment '#{self.filename}' (#{@temp_file.size} bytes) to database\")\n md5 = Digest::MD5.new\n buffer = \"\"\n self.data = \"\"\n while (buffer = @temp_file.read(8192))\n md5.update(buffer)\n self.data << buffer\n end\n self.digest = md5.hexdigest\n end\n @temp_file = nil\n # Don't save the content type if it's longer than the authorized length\n if self.content_type && self.content_type.length > 255\n self.content_type = nil\n end\n end",
"def save(data: content, name: subject + '.html')\n filename = 'tmp/' + name\n\n File.open(filename, 'w+') do |f|\n f.write(data)\n end\n end",
"def content=(value)\n self.temp_files = [value].flatten.select do |f|\n FILE_CLASSES.member?(f.class.name)\n end\n # correctly triggering dirty\n if temp_files.present?\n write_attribute(:content, nil)\n content_will_change!\n else\n write_attribute(:content, value)\n end\n end",
"def data\n read_attribute(:data) || write_attribute(:data, (db_file_id ? db_file.data : nil))\n end",
"def save\n if id\n update(title: title, description: description, file: file)\n else\n new_instance = self.class.create(sys[:space].id, fields: instance_variable_get(:@fields))\n refresh_data(new_instance)\n end\n end",
"def save(path, mode = ::File::WRONLY|::File::CREAT|::File::TRUNC, **options)\n\t\t\t\t\tif @body\n\t\t\t\t\t\t::File.open(path, mode, **options) do |file|\n\t\t\t\t\t\t\tself.each do |chunk|\n\t\t\t\t\t\t\t\tfile.write(chunk)\n\t\t\t\t\t\t\tend\n\t\t\t\t\t\tend\n\t\t\t\t\tend\n\t\t\t\tend",
"def []=(key, content)\n temp_file = File.join(root, ['tmp', $$, Thread.current.object_id].join('-'))\n File.open(temp_file, 'wb') do |dest|\n if content.respond_to? :to_str\n dest.write(content.to_str)\n else\n content.each {|s| dest.write(s) }\n end\n end\n\n path = cache_path(key)\n File.unlink path if File.exist?(path)\n FileUtils.mkdir_p File.dirname(path), :mode => 0755\n FileUtils.mv temp_file, path\n rescue\n File.unlink temp_file rescue nil\n ensure\n content\n end",
"def save\n r = self.class.put(\"/domains/#{@@parent_id}/records/#{self.id}\", :query => { \"record[name]\" => self.name,\n \"record[ttl]\" => self.ttl,\n \"record[content]\" => self.content,\n \"record[prio]\" => self.prio,\n \"record[type]\" => self.type })\n r[\"errors\"] and raise StandardError, r[\"errors\"][\"error\"].to_a.join(\", \")\n if r.code == 200\n self.class.find(self.id)\n else\n raise StandardError, 'Could not update the record'\n end\n end",
"def []=(file_name, content)\n path = absolute_path_to_file(file_name)\n directory = File.dirname(path)\n FileUtils.mkdir_p(directory) unless File.exist?(directory)\n File.open(path, 'w') { |f| f.write(content) }\n end",
"def write_cache(resource, content)\n expanded_path = cache_path_for(resource)\n return false unless expanded_path\n FileUtils.mkdir_p(File.dirname(expanded_path))\n @logger.info(self.class) { \"Caching #{content.length} B for #{resource}\" }\n File.write(expanded_path, content)\n end",
"def write_file(filename, content)\n FileUtils.mkdir_p File.dirname(filename)\n IO.binwrite(filename, content)\n end",
"def save(dir, data, filename)\n FileUtils.mkdir_p dir\n fn = dir+\"/\"+filename \n puts fn\n open(fn, 'w') { |f|\n f.puts data\n f.close\n }\n end",
"def create\n @content = Content.new(params[:content])\n\n respond_to do |format|\n if @content.save\n Searcher.open(Const.get('searcher_db')) do |db|\n db.regist(@content)\n end\n flash[:notice] = \"'#{@content.title}'が作成されました。: #{@content.path}\"\n \n @parent = Content.find(params[:parent_id]) if params[:parent_id]\n if @parent != nil\n @parent.children << @content\n format.html { redirect_to(@parent) }\n else\n fid = session[:folder_id]\n fid = params[:content][:folder_id] if fid == nil\n @folder = Folder.find(fid)\n format.html { redirect_to(@folder) }\n end\n format.xml { render :xml => @content, :status => :created, :location => @content }\n else\n format.html { render :action => \"new\" }\n format.xml { render :xml => @content.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def save\n @@data_store.insert(self)\n self\n end",
"def create(path, content)\n begin\n (@cmd.model[:public] + path).write(content)\n return true\n rescue => e\n return false\n end\n end",
"def write\n file_name = \"course_#{@file_term}.json\"\n # Write the JSON to the course_work_content folder where the app will pick it up later\n Rails.root.join(\"lib/course_work_content\", file_name).write(@courses.to_json)\n end",
"def save_to(path); end",
"def save\n File.open(path, 'w') do |out|\n YAML.dump(data, out)\n end\n end",
"def content_path\n path = \"esm/#{self.esm.name}/#{self.name}/content\"\n file_path = \"public/#{path}\"\n p = file_path.split('/')\n i=0\n unless FileTest.exist?(file_path)\n while(s = \"public/esm/#{p[2...2+i].join('/')}\" and s!=file_path)\n Dir.mkdir(s) unless FileTest.exist?(s)\n i+=1\n end\n Dir.mkdir(s) unless FileTest.exist?(s)\n end\n return path\n end",
"def save_as(path)\n FileUtils.mkdir_p(path)\n File.open(File.join(path, 'contents.xcworkspacedata'), 'w') do |out|\n out << to_s\n end\n end",
"def store obj, msg\n Dir.chdir(working_dir) do\n dirname = File.dirname(obj.path)\n FileUtils.mkdir_p(dirname) unless File.directory?(dirname)\n File.open(obj.path, 'w'){ |f| f.write(obj.content) }\n add(obj.path)\n end\n commit_index(msg)\n end",
"def save\n CSV.open('./db/gossip.csv', 'ab') do |csv|\n csv << [@content, @author]\n end\n end",
"def create\n @content = Content.new(permitted_params.content)\n\n @content.project = @project\n @content.content_type = @content_type\n\n respond_to do |format|\n if @content.save\n\n\n @content.ctbs.each do |ctb|\n if params[:link]\n ctb.link = (params[:link] == ctb.branch_id.to_s) ? true : false\n end\n\n if params[:caption]\n ctb.caption = (params[:caption][ctb.branch_id.to_s]) ? params[:caption][ctb.branch_id.to_s] : ''\n end\n\n ctb.save\n end\n\n\n if params[:content_elements][:add]\n params[:content_elements][:add].each do |content_element_type, value|\n begin\n @content.content_elements.build(:content_element_type_id => content_element_type, :language => @locale, :value => value).save\n rescue\n flash[:error] = t('fehler.ascii')\n @content.content_elements.build(:content_element_type_id => content_element_type, :language => @locale, :value => cleanup(value)).save\n end\n end\n end\n @content.proof_bracketcommands\n @content.setFileNames\n\n format.html { redirect_to edit_project_content_type_content_path(@project, @content_type, @content, :locale => @locale), notice: 'Content was successfully created.' }\n format.json { render action: 'show', status: :created, location: [@project, @content_type, @content] }\n else\n format.html { render action: 'new' }\n format.json { render json: @content.errors, status: :unprocessable_entity }\n end\n end\n end",
"def write\n if self.site.persisted?\n self.check_locales! unless self.force? # requirements\n\n if self.force?\n self.update_site\n end\n else\n self.create_site\n end\n end",
"def migrate_content_datastreams\n save\n target.attached_files.keys.each do |ds|\n mover = FedoraMigrate::DatastreamMover.new(source.datastreams[ds.to_s], target.attached_files[ds.to_s], options)\n report.content_datastreams << ContentDatastreamReport.new(ds, mover.migrate)\n end\n end",
"def save_data(file, data)\n if ENV['RACK_ENV'] == 'test'\n local = File.expand_path(\"../test/data/#{file}\", __FILE__)\n File.open(local, 'w') { |open_file| open_file.write(YAML.dump(data)) }\n else\n local = File.expand_path(\"../data/#{file}\", __FILE__)\n File.open(local, 'w') { |open_file| open_file.write(YAML.dump(data)) }\n\n if USE_GOOGLE_DRIVE\n remote = google_session.file_by_title(file.to_s)\n remote.update_from_file(local)\n end\n end\nend",
"def write_file_at(file, content)\n\t\tFileUtils.mkdir_p(File.dirname(file))\n\t\tFile.open(file, 'w') do |file|\n\t\t\tfile.write(content)\n\t\tend\n\tend"
] |
[
"0.6420183",
"0.6406604",
"0.6371505",
"0.6282888",
"0.6026979",
"0.59837586",
"0.5962015",
"0.5962015",
"0.5951333",
"0.5940295",
"0.58668655",
"0.58661467",
"0.58489925",
"0.58402663",
"0.5838136",
"0.582204",
"0.58186257",
"0.5813595",
"0.580001",
"0.5793781",
"0.5789255",
"0.57285416",
"0.5718909",
"0.5701694",
"0.5685606",
"0.5677654",
"0.56763303",
"0.5654316",
"0.56531084",
"0.5650732",
"0.56457925",
"0.5628021",
"0.5623283",
"0.5608877",
"0.5607243",
"0.558375",
"0.55636066",
"0.5547832",
"0.5546184",
"0.5538801",
"0.5538801",
"0.5532239",
"0.55239975",
"0.5521001",
"0.55006367",
"0.5494177",
"0.5493804",
"0.5493804",
"0.5482371",
"0.5462572",
"0.5458341",
"0.5455733",
"0.5455733",
"0.54480475",
"0.54472965",
"0.54398113",
"0.54275453",
"0.5417155",
"0.5408824",
"0.5408824",
"0.54056543",
"0.5399373",
"0.539117",
"0.5389493",
"0.5383987",
"0.5383364",
"0.53796965",
"0.5376065",
"0.5372036",
"0.5369816",
"0.5368064",
"0.5365242",
"0.53598815",
"0.5353771",
"0.53348583",
"0.5334851",
"0.5332657",
"0.5332029",
"0.5305091",
"0.52861655",
"0.5277375",
"0.52679664",
"0.525369",
"0.5246832",
"0.5242487",
"0.5240195",
"0.52382433",
"0.5220395",
"0.52189696",
"0.5209812",
"0.52097034",
"0.5191565",
"0.51902324",
"0.51843286",
"0.5183065",
"0.5178445",
"0.51748955",
"0.5174035",
"0.5169173",
"0.51645416"
] |
0.57324785
|
21
|
Returns a snapshot of instances indexed before (including) provided timestamp.
|
def get(till = DateTime.now)
# looking for the latest base file that is earlier than till argument
base = snapshot_files.inject(nil) do |cur_base, f|
if (cur_base.nil? || f.same_time_as?(till) ||
(f.earlier_than?(till) && f.later_than?(cur_base.till)))
cur_base = f
end
cur_base
end
base_cd = ContentData::ContentData.new
base_cd.from_file(base.filename)
# applying diff files between base timestamp and till argument
diff_from_base = diff(base.till, till)
added_content_data = diff_from_base[DiffFile::ADDED_TYPE]
removed_content_data = diff_from_base[DiffFile::REMOVED_TYPE]
result = base_cd.merge(added_content_data)
result.remove_instances!(removed_content_data)
result
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def preceding(obj, timestamp_arg = false)\n if timestamp_arg != true && primary_key_is_int?\n return where(arel_table[primary_key].lt(obj.id)).order(arel_table[primary_key].desc)\n end\n\n obj = obj.send(PaperTrail.timestamp_field) if obj.is_a?(self)\n where(arel_table[PaperTrail.timestamp_field].lt(obj)).\n order(timestamp_sort_order(\"desc\"))\n end",
"def newer_than_timestamp(timestamp)\n non_future_partitions.select do |p|\n timestamp <= p.timestamp\n end\n end",
"def version_at(timestamp, reify_options = {})\n # Because a version stores how its object looked *before* the change,\n # we need to look for the first version created *after* the timestamp.\n v = versions.subsequent(timestamp, true).first\n return v.reify(reify_options) if v\n @record unless @record.destroyed?\n end",
"def version_at(timestamp, reify_options = {})\n # Because a version stores how its object looked *before* the change,\n # we need to look for the first version created *after* the timestamp.\n v = versions.subsequent(timestamp, true).first\n return v.reify(reify_options) if v\n @record unless @record.destroyed?\n end",
"def initialize(timestamp)\n @timestamp = timestamp\n @records = []\n end",
"def history_timestamps\n self.class.history.timestamps do |query|\n query.where(:id => self)\n end\n end",
"def records_modified_since(time)\n []\n end",
"def version_at(timestamp, reify_options={})\n # Because a version stores how its object looked *before* the change,\n # we need to look for the first version created *after* the timestamp.\n version = versions.after(timestamp).first\n version ? version.reify(reify_options) : self\n end",
"def touch_index_timestamp\n self.index_timestamp = Time.now\n end",
"def capped_records(database, collection, conditions, last = Time.now)\n db = @connection.db(database)\n coll = db.collection(collection)\n conditions = conditions.merge({\"timestamp\" => {\"$lt\" => last}})\n coll.find(conditions).sort([[\"$natural\", \"-1\"]]).limit(50)\n end",
"def subsequent(obj, timestamp_arg = false)\n if timestamp_arg != true && primary_key_is_int?\n return where(arel_table[primary_key].gt(obj.id)).order(arel_table[primary_key].asc)\n end\n\n obj = obj.send(PaperTrail.timestamp_field) if obj.is_a?(self)\n where(arel_table[PaperTrail.timestamp_field].gt(obj)).order(timestamp_sort_order)\n end",
"def records_modified_since(time)\n condition = []\n %w(updated_at created_at).each do |col|\n condition << \"#{col} >= ?\" if column_names.include? col\n end\n if condition.empty?\n logger.warn \"#{self.name}: Override records_modified_since(time) to keep the index up to date with records changed during rebuild.\"\n []\n else\n find :all, :conditions => [ condition.join(' AND '), *([time]*condition.size) ]\n end\n end",
"def timestamp\n first(:timestamp)\n end",
"def previous(count = 1)\n dupe = self.dup\n dupe.assign_attributes start_time: start_time + push(count), end_time: end_time + push(count)\n dupe\n end",
"def timestamp # :nodoc:\n @timestamp.dup\n end",
"def find_snapshot_timestamp\n last_snap = behavior(:find_snapshots).last\n last_snap.tags.detect { |t| t[\"name\"] =~ /timestamp=(\\d+)$/ }\n timestamp = $1\n end",
"def starting(time_since:, time_until:)\n table.where(\"start >= ? AND start <= ?\", time_since, time_until).entries\n end",
"def getSnapshot(month, day, hour)\n time = getUnixTime(month, day, hour)\n difference = false\n index = 0 \n for i in @@snapshots\n diff = ((Time.parse(i[4])).to_i - time).abs\n if (not difference or diff < difference)\n difference = diff\n index = i \n end \n end \n index\n end",
"def earlier?(timestamp)\n starts_at && timestamp < starts_at\n end",
"def earlier_than(created_time)\n where('created_at.lt': created_time)\n end",
"def earliest_order\n # order(created_at: :asc)\n scan_index_forward(true)\n end",
"def get_hits(timestamp)\n \n end",
"def where_history()\r\n @history_offset\r\n end",
"def get_oplog_cursor(timestamp)\n Mongo::Cursor.new(@oplog_coll,\n { :tailable => true,\n :selector => { \"op\" => { \"$ne\" => \"n\" },\n \"ts\" => { \"$gte\" => timestamp } },\n :order => [\"$natural\", :asc]\n })\n end",
"def current_partition(current_timestamp)\n non_future_partitions.select do |p|\n p.timestamp > current_timestamp\n end.min_by { |p| p.timestamp }\n end",
"def stamp_new_rows\n db.query(\"UPDATE #{audit} SET `_copied_at` = NOW() WHERE `_copied_at` IS NULL\")\n end",
"def past\n @matches = get_query(:matches).where('ends_at < now()')\n render(:index)\n end",
"def hit(timestamp)\n \n end",
"def range_records(database, collection, conditions, first = Time.now, last = Time.now, limit = 2000)\n db = @connection.db(database)\n coll = db.collection(collection)\n conditions = conditions.merge({\"timestamp\" => {\"$lte\" => last, \"$gte\" => first}})\n\n baseQuery = coll.find(conditions).sort([[\"$natural\", \"-1\"]]).limit(limit)\n end",
"def at(time)\n from, to = quoted_history_fields\n unscoped.\n select(\"#{quoted_table_name}.*, '#{time}' AS as_of_time\").\n where(\"'#{time}' >= #{from} AND '#{time}' < #{to}\")\n end",
"def timestamps\n assocs = reflect_on_all_associations.select {|a|\n [:belongs_to, :has_one, :has_many].include?(a.macro) && a.klass.chrono?\n }\n\n models = [self].concat(assocs.map {|a| a.klass.history})\n fields = models.inject([]) {|a,m| a.concat m.quoted_history_fields}\n\n relation = self.\n joins(*assocs.map(&:name)).\n select(\"DISTINCT UNNEST(ARRAY[#{fields.join(',')}]) AS ts\").\n order('ts')\n\n relation = yield relation if block_given?\n\n sql = \"SELECT ts FROM ( #{relation.to_sql} ) foo WHERE ts IS NOT NULL AND ts < NOW()\"\n sql.gsub! 'INNER JOIN', 'LEFT OUTER JOIN'\n\n connection.on_schema(Adapter::HISTORY_SCHEMA) do\n connection.select_values(sql, \"#{self.name} periods\").map! do |ts|\n Conversions.string_to_utc_time ts\n end\n end\n end",
"def sorted_snapshot_list()\n snapshot_list().sort{|a,b| b.time <=> a.time }\n end",
"def get_since(timestamp, limit=20)\n uri = '/api/v1/pulses/events'\n params = {limit: limit, since: timestamp}\n events = []\n begin\n json_data = get(uri, params)\n page = json_data['next']\n\n params = URI::decode_www_form(URI(page).query).to_h unless page.nil?\n\n events += json_data['results']\n end while !page.nil?\n\n results = []\n events.each do |event|\n results << OTX::Event.new(event)\n end\n\n return results\n end",
"def pred_timestamp(options = {})\n if historical?\n options[:before] ||= as_of_time\n timeline(options.merge(:limit => 1, :reverse => true)).first\n else\n timeline(options.merge(:limit => 2, :reverse => true)).second\n end\n end",
"def pred_timestamp(options = {})\n if historical?\n options[:before] ||= as_of_time\n timeline(options.merge(limit: 1, reverse: true)).first\n else\n timeline(options.merge(limit: 2, reverse: true)).second\n end\n end",
"def past\n index\n end",
"def get_volumes(timestamp, duration = 31.minutes)\n volume_class = \"::#{exchange_name.capitalize}Volume\".constantize\n execute_query(\n volume_class.where('timestamp > ?', (timestamp - duration.to_i)).to_sql\n )\n end",
"def index\n @recent_objects = Serial.order(updated_at: :desc).limit(10)\n end",
"def [](time)\n prev_time = prev_time time\n fetch prev_time if prev_time\n end",
"def hits_for_day(timestamp)\n properties = Advert.where(user_id: self.id).pluck(:property_id)\n beg_time = Time.zone.at(timestamp).beginning_of_day\n end_time = Time.zone.at(timestamp).end_of_day\n Hit.where(\"property_id IN (?)\", properties).where(created_at: beg_time..end_time).all\n end",
"def timestamp\n self[:timestamp]\n end",
"def history\n generic_index(true)\n end",
"def diff_from(time: nil, version: nil)\n all.map { |record| record.diff_from(time: time, version: version) }\n end",
"def timestamp\n TimeStamp.new\n end",
"def timestamp\n @timestamp ||= Time.now.xs_datetime\n end",
"def get_earliest_timestamp\n @timestamp = (@raw_image_files.sort_by { |i| i.timestamp }).first.timestamp\n end",
"def past\n Event.all.where('time_event < ?', Time.now)\n end",
"def get_oplog_cursor_w_check(timestamp)\n ret = nil\n\n unless timestamp.nil? then\n cursor = get_oplog_cursor(timestamp)\n doc = cursor.next_document\n\n if doc.nil? or not cursor.has_next? then\n # This does not necessarily mean that the cursor is too stale, since the\n # timestamp passed can be a timestamp of a no-op entry. So double check\n # if that said entry still exists.\n entry = @oplog_coll.find_one({ \"ts\" => timestamp })\n\n if entry.nil? then\n less_than_doc = @oplog_coll.find_one({ \"ts\" => { \"$lt\" => timestamp }})\n\n unless less_than_doc.nil? then\n ret = get_oplog_cursor_w_check(rollback)\n end\n else\n ret = cursor\n end\n elsif timestamp == doc[\"ts\"] then\n ret = cursor\n else\n @logger.warn(\"#{@name}: (#{timestamp.inspect}) is too old and not in the oplog\")\n end\n end\n\n return ret\n end",
"def beginning\n event.timestamp\n end",
"def snapshot\n if snapshot_key\n snapshots.find_by(key: snapshot_key)\n else\n snapshots.order('created_at DESC').limit(1).first\n end\n end",
"def search_updated_guests(timestamp)\n ldap_timestamp = timestamp.to_time.utc.strftime(TIMESTAMP_FORMAT)\n modified_timestamp_filter = Net::LDAP::Filter.ge('modifytimestamp', ldap_timestamp)\n client.search(base: GUEST_DN, filter: modified_timestamp_filter)\n end",
"def in_hot_list\n where(:added_to_hot_list.gte => 4.days.ago.utc)\n end",
"def sort_timestamp\n self.created_at\n end",
"def focussed_index\n @current_index # 2009-01-07 14:35 \n end",
"def prune\n if Event.count > 60\n cutoff = Event.limit(60).last.created_at\n Event.where([\"created_at < ?\", cutoff]).delete_all\n end\n end",
"def get_revision_by_timestamp(at_or_earlier_than, path = nil, later_than = nil)\n unless at_or_earlier_than.is_a?(Time)\n raise 'Was expecting a timestamp of type Time'\n end\n\n (@revision_history + [@current_revision]).reverse_each do |revision|\n return nil if !later_than.nil? && revision.server_timestamp <= later_than\n return revision if revision.server_timestamp <= at_or_earlier_than &&\n (path.nil? || revision.changes_at_path?(path))\n end\n nil\n end",
"def recent(n = 5)\n return @history[-n..-1] if n < @history.size\n return @history[0..n]\n end",
"def recents_get(since_timestamp, opts = {})\n recents_get_with_http_info(since_timestamp, opts)\n end",
"def findLogs (objectId, objectType, timestamp)\n Log.where(object_id: objectId, object_type: objectType)\n .where(\"log_timestamp <= ?\", Time.at(timestamp.to_i))\n .order(\"log_timestamp\")\n end",
"def by_created_at_before(events)\n return events unless params[:before]\n\n events.where('events.created_at < ?', params[:before].beginning_of_day)\n end",
"def recent from\n where ['created_at > ?', (from || 2.weeks.ago).to_s(:db)]\n end",
"def all_previous\n self.class.base_class.where(\"#{scoped_position} < ?\", position)\n end",
"def get_scans_for_last(seconds_ago)\n Scan.find(:all, :conditions => [\"complete > ? and complete <= ?\", complete - seconds_ago, complete])\n end",
"def find_latest_backup(lineage, timestamp, from_master = nil)\n filter = [\n \"latest_before==#{timestamp.utc.strftime('%Y/%m/%d %H:%M:%S %z')}\",\n 'committed==true',\n 'completed==true'\n ]\n filter << \"from_master==#{from_master}\" if from_master\n\n # Add cloud href to the list of filters so we only get backups in the current cloud as we can't restore backups\n # from a different cloud.\n cloud_href = @api_client.get_instance.cloud.href\n filter << \"cloud_href==#{cloud_href}\"\n\n backup = @api_client.backups.index(lineage: lineage, filter: filter)\n backup.first\n end",
"def bt_current(instant=Time.zone.now)\n vt_intersect(instant).tt_intersect(instant)\n end",
"def hits_prev\n # see PsiBlastJob for comment\n end",
"def get_revision_by_timestamp(at_or_earlier_than, path = nil, later_than = nil)\n raise NotImplementedError\n end",
"def load_previous_version\n Version.all(:limit => 1, :order => [ :processed.desc ]).first\n end",
"def search_result\n klass.where(\"object_id = ? AND object_type = ? AND timestamp <= ?\", object_id, object_type, timestamp).order(:timestamp)\n end",
"def previous_visits\n user.visits.where('entry_date <= ? and id <> ?', entry_date, id)\n end",
"def future_events\n instances = []\n Event.where(place: self).each do |e|\n instances << e.occurrences.where(date: Date.current..2.weeks.from_now)\n end\n instances.flatten!.sort_by(&:date) unless instances.empty?\n end",
"def find_all_at_ts(ts)\n repository.find_all_at_ts(ts_bucket(ts))\n end",
"def history\n from = @from.to_i\n limit = @to.to_i - from\n # Excecute the find of comparations\n histories = @to ? History.offset(from).limit(limit) : History.all\n histories\n end",
"def index\n @snaps = Snap.order(created_at: :desc).limit(20).all\n end",
"def rl_beginning_of_history(count, key)\r\n rl_get_previous_history(1 + where_history(), key)\r\n end",
"def get_new_petitions\n @new_petitions = Petition.where(created_at: time_range).order('created_at ASC')\n end",
"def request_history(timestamp=1.day.ago)\n @ws.send({request: \"history\", timestamp: timestamp}.to_json)\n end",
"def oldest(constraints = {})\n constraints.merge!(order: :created_at.asc)\n _q = query(constraints)\n _q.define_singleton_method(:method_missing) { |m, *args, &block| self.results.send(m, *args, &block) }\n _q\n end",
"def new_timestamp # :nodoc:\n @properties['timestamp'].dup\n end",
"def previous_events(event)\n # Initialise the hash to use an empty array as default value\n @dups ||= Hash.new { |h, k| h[k] = [] }\n one_hour_earlier = event.start.advance(:hours => -1)\n @dups[one_hour_earlier]\n end",
"def timestamp\n Time.at((attributes[:timestamp] || Time.now).to_i)\n end",
"def find_since(time_since:)\n @time_now = Time.now\n #puts \"find_since, Recieve: #{time_since}, time_now: #{time_now}\"\n return [] if (time_now - time_since) < shorter_time_window_span\n open_window = get_open_window(time_since, time_now)\n [open_window] + find_in_range(time_from: time_since, time_to: open_window.fetch(:window_starts), time_windows: time_span_windows.clone)\n end",
"def timestamp; end",
"def timestamp; end",
"def timestamp; end",
"def timestamp; end",
"def timestamp; end",
"def timestamp; end",
"def past_events\n events.where(\"date < ?\", Time.now)\n end",
"def works_index_timestamp\n REDIS_GENERAL.get(redis_works_index_key) || update_works_index_timestamp!\n end",
"def since_fetch_at(fetch_time)\n # Default to 0 to remain compatible with old clients\n last_fetched_at = Time.at(@params.fetch(:last_fetched_at, 0).to_i)\n\n @notes.where('updated_at > ?', last_fetched_at - FETCH_OVERLAP).fresh\n end",
"def active_bucket_times(starting_at=Time.now)\n keys = []\n starting_at = starting_at.utc\n @active_bucket_count.times do \n keys << quantize_time_to_slize_size(starting_at)\n starting_at -= @slice_size\n end\n keys\n end",
"def current_snapshot\n @snapshots[0][1]\n end",
"def starting(starts_at)\n merge(starts: starts_at)\n end",
"def [](index)\n warn \"Timestamps are no longer deserialized as arrays. If you're working \" +\n \"with BSON Timestamp objects, see the BSON::Timestamp class. This usage will \" +\n \"be deprecated in Ruby Driver v2.0.\"\n if index == 0\n self.increment\n elsif index == 1\n self.seconds\n else\n nil\n end\n end",
"def earliest_backtrack_time\n data[:earliest_backtrack_time]\n end",
"def recent_uniques\n visits.where(created_at: 10.minutes.ago..Time.now).distinct\n end",
"def previous_snapshot(snapname)\n prev = nil\n tags.each do |tag|\n if tag == snapname\n return prev\n else\n prev = tag\n end\n end\n raise 'Snapshot not found'\n end",
"def collect_movements\n Measurement.collect_most_recent_measurements_for self\n index_movements\n end",
"def bt_history(vtparams=AllTime, ttparams=nil)\n if ttparams\n bt_versions.vt_intersect(vtparams).tt_intersect(ttparams).order(:vtstart_at)\n else\n bt_versions.vt_intersect(vtparams).tt_forever.order(:vtstart_at)\n end\n end",
"def check_timestamp\n @recent.save! if Time.now - @recent.updated_at >= 600\n end"
] |
[
"0.60803396",
"0.5694255",
"0.55965793",
"0.55965793",
"0.5550635",
"0.55179596",
"0.54913265",
"0.5463175",
"0.5459604",
"0.5356755",
"0.5335492",
"0.53346974",
"0.533032",
"0.5284458",
"0.5280035",
"0.5266561",
"0.52608395",
"0.5223791",
"0.5218488",
"0.52169335",
"0.5212088",
"0.51819944",
"0.51746744",
"0.51449805",
"0.51401716",
"0.5133076",
"0.5110332",
"0.5103277",
"0.5101246",
"0.5096228",
"0.5072019",
"0.50413847",
"0.50216943",
"0.49796867",
"0.49788275",
"0.4968867",
"0.49380666",
"0.49129716",
"0.4892195",
"0.48778322",
"0.48489222",
"0.48453143",
"0.48445034",
"0.48372748",
"0.48246524",
"0.48207328",
"0.48198432",
"0.48154187",
"0.48142025",
"0.48012048",
"0.47972035",
"0.4796583",
"0.47953877",
"0.478815",
"0.4787755",
"0.47851312",
"0.47628984",
"0.47615367",
"0.4755874",
"0.47473168",
"0.47416148",
"0.4729034",
"0.47286773",
"0.47264993",
"0.47197083",
"0.4712323",
"0.47092447",
"0.47053364",
"0.4704388",
"0.46921209",
"0.468903",
"0.46864012",
"0.4683842",
"0.46832252",
"0.46798846",
"0.46789265",
"0.46708986",
"0.46702814",
"0.46697554",
"0.46556628",
"0.46548903",
"0.46514165",
"0.46364766",
"0.46364766",
"0.46364766",
"0.46364766",
"0.46364766",
"0.46364766",
"0.46354154",
"0.4625697",
"0.4623771",
"0.46220928",
"0.4609635",
"0.4607419",
"0.46050522",
"0.45909512",
"0.45891336",
"0.45840192",
"0.45838496",
"0.4581365",
"0.45702678"
] |
0.0
|
-1
|
Diff between two timestamps.
|
def diff(from, till = DateTime.now)
if from.nil?
err_msg = 'from parameter should be defined. Did you mean a get method?'
fail ArgumentError, err_msg
end
diff_files_in_range = diff_files.select do |df|
df.later_than?(from) &&
(df.earlier_than?(till) || df.same_time_as?(till))
end
added_content_data = ContentData::ContentData.new
removed_content_data = ContentData::ContentData.new
# diff files sorted by date to enable correct processing of files
# that were changed few times between the timestamps,
# thus resulted revision is indeed latest revision.
diff_files_in_range.sort!.each do |df|
cd = ContentData::ContentData.new
cd.from_file(df.filename)
if df.type == DiffFile::ADDED_TYPE
added_content_data.merge!(cd)
elsif df.type == DiffFile::REMOVED_TYPE
removed_content_data.merge!(cd)
end
end
{ DiffFile::REMOVED_TYPE => removed_content_data,
DiffFile::ADDED_TYPE => added_content_data }
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def time_diff\n diff = (@start_at and @end_at) ? @end_at.to_time - @start_at.to_time : 0\n diff.round.to_i\n end",
"def time_diff\n return ((time_2 - time_1) / 3600).round\n end",
"def timeDifference\n receive_timestamp-@startTime\n end",
"def timestamp_delta\n @timestamp_delta ||= @records.empty? ? 0.0 : (@records.last.timestamp - @records.first.timestamp)\n end",
"def calc_difference\n @difference = @date_time.to_f - @now_date_time.to_f\n end",
"def time_diff(start_time, end_time)\n #calculate time elapsed\n elapsed_seconds = ((self.end_time - self.start_time)).to_i\n mins, secs = elapsed_seconds.divmod(60)\n hours, mins = mins.divmod(60)\n days, hours = hours.divmod(24)\n return days, hours, mins, secs\nend",
"def realDifference\n timeDifference-(latency/2)\n end",
"def triptime(time1, time2)\r\n\r\n\t #For precision let the answer be in minutes\r\n\t t = ((Time.parse(time2) - Time.parse(time1))/60)\r\n\t t\r\n\tend",
"def time_diff(time1, time2)\n time1_a = time1.split(':')\n time2_a = time2.split(':')\n \n hours_output = 0\n hours1 = time1_a[0].to_i\n hours2 = time2_a[0].to_i\n hours = hours2 - hours1\n \n mins = time2_a[1].to_i - time1_a[1].to_i\n \n while hours > 0\n hours_output += 60\n hours -= 1\n end\n return hours_output + mins \nend",
"def difference(dt1,dt2)\n f = (dt1-dt2)\n (f*f.denominator).abs\nend",
"def time_diff_milli(start, finish)\n (finish - start) * 1000.0\n end",
"def time_between(start_point, end_point)\n return 0 unless start_point.time && end_point.time\n end_point.time - start_point.time\n end",
"def elapsed_time\n seconds = (self.end_at.to_i - self.start_at.to_i)\n (seconds / 60)\n end",
"def diff_since_last\n last.blank? ? [0,0,0,0] : distance_of_time_as_array( timestamp_server, last.timestamp_server)\n end",
"def time_diff_milli( start, finish )\n\n\t( finish - start )\n\nend",
"def time_diff(start_time, end_time)\n seconds_diff = (start_time - end_time).to_i.abs\n\n days = seconds_diff / 86400\n seconds_diff -= days * 86400\n\n hours = seconds_diff / 3600\n seconds_diff -= hours * 3600\n\n minutes = seconds_diff / 60\n seconds_diff -= minutes * 60\n\n seconds = seconds_diff\n\n \"#{days.to_s.rjust(3,' ')}d #{hours.to_s.rjust(2, '0')}:#{minutes.to_s.rjust(2, '0')}:#{seconds.to_s.rjust(2, '0')}\"\nend",
"def time_diff(start_time, end_time)\n seconds_diff = (start_time - end_time).to_i.abs\n\n hours = seconds_diff / 3600\n seconds_diff -= hours * 3600\n\n minutes = seconds_diff / 60\n seconds_diff -= minutes * 60\n\n seconds = seconds_diff\n\n \"#{hours.to_s.rjust(2, '0')}:#{minutes.to_s.rjust(2, '0')}:#{seconds.to_s.rjust(2, '0')}\"\nend",
"def precise_diff(to, from)\n years = to.year - from.year\n months = to.month - from.month\n days = to.day - from.day\n hours = to.hour - from.hour\n minutes = to.minute - from.minute\n\n if minutes < 0\n minutes += 60\n hours -= 1\n end\n\n if hours < 0\n hours += 24\n days -= 1\n end\n\n if days < 0\n days += last_day_in_month(months-1)\n end\n\n if months < 0\n months += 12\n years -= 1\n end\n\n\n [years, months, days, hours, minutes]\n end",
"def time_distance_in_ms(time1, time2)\n ((time1 - time2) * 1_000).round\n end",
"def getDiff ( x , y )\n\tdiff = x - y\n\treturn diff\nend",
"def get_deltatime\n return @deltatime\n end",
"def delta(v1, v2)\n if block_given?\n v1 = yield(v1)\n v2 = yield(v2)\n end\n return (v1 - v2).abs\n end",
"def get_time_delta\n return(@rundate - @histodate)\n end",
"def elapsed_seconds(start_time, end_time)\n end_time.to_i - start_time.to_i\nend",
"def time_diff(start_time, end_time)\n seconds_diff = (start_time - end_time).to_i.abs\n hours = seconds_diff / 3600\n return hours\n end",
"def delta\n @end_point - @start_point\n rescue\n 0\n end",
"def travel_time\n return 0 if points.count < 2\n\n points.last.record_time - points.first.record_time\n end",
"def duration\n if start_time && end_time\n end_time.to_i - start_time.to_i\n else\n 0.0\n end\n end",
"def time_duration\n t1 = Time.now.to_f\n Time.now.to_f - t1\nend",
"def difference_in_minutes time_one, time_two\n time_one_with_resetted_date = reset_date_for_time time_one\n time_two_with_resetted_date = reset_date_for_time time_two\n (time_one_with_resetted_date - time_two_with_resetted_date) / 60\n end",
"def diff\n @start_position - @position\n end",
"def get_time_diff(time)\n return (time*60*60*24).to_i\n end",
"def diff(other)\n Distance.diff_total(self, other, true)\n end",
"def elapsed_seconds(start_time, end_time)\r\n end_time - start_time\r\nend",
"def time\n end_time - start_time\n end",
"def timestamp\n ((Time.now.to_f - StartTime)*1000).round\n end",
"def time_diff_milli(start, finish=Time.now)\n rc=(finish - start) * 1000.0\n end",
"def time_delta\n device_time = time\n delta = device_time - Time.now\n end",
"def duration\n self.updated_at - self.created_at\n end",
"def calculate_time\n if @event.present? && @event.ends != nil\n event_time = @event.ends - @event.starts\n \n if event_time < 3600\n difference = ((event_time / 60) % 60)\n @event_difference = \"#{difference.round(0)} minutes\"\n else\n difference = event_time / (60 * 60)\n if difference == 1 \n @event_difference = \"#{difference.round(0)} hour\"\n else\n @event_difference = \"#{difference.round(1)} hours\"\n end\n end\n end\n end",
"def elapsed_seconds(start_time, end_time)\n return (end_time - start_time)\nend",
"def date_diff()\n start_time = Time.current\n @d = distance_of_time_in_words(start_time, @due_date, false)\n end",
"def elapsed_time\n if end_time && start_time\n return ((end_time - start_time)/60).round\n else\n return 0\n end\n end",
"def check_timestamps(block1, block2)\n timestamp_one_string = block1.timestamp_string.split('.')\n timestamp_two_string = block2.timestamp_string.split('.')\n\n timestamp_one_partition = timestamp_one_string.map(&:to_i)\n timestamp_two_partition = timestamp_two_string.map(&:to_i)\n\n if timestamp_two_partition[0] < timestamp_one_partition[0]\n puts \"Line #{block2.block_number}: Previous timestamp #{block1.timestamp_string}\n \t>= new timestamp #{block2.timestamp_string}\"\n return false\n end\n\n if timestamp_two_partition[0] == timestamp_one_partition[0]\n if timestamp_two_partition[1] <= timestamp_one_partition[1]\n puts \"Line #{block2.block_number}: Previous timestamp #{block1.timestamp_string}\n \t >= new timestamp #{block2.timestamp_string}\"\n return false\n end\n end\n true\n end",
"def difference_in_days(date1, date2)\n return ((date1.utc.at_midnight-date2.utc.at_midnight)/SECONDS_PER_DAY).to_i\n end",
"def subtract(t1,t2) ##revisar name\n\t\treturn (t1-t2).to_i\n\tend",
"def calculate_elapsed_time(result)\n Time.parse(result.finish_time).to_i - Time.parse(result.start_time).to_i\n end",
"def time_elapsed_secs\n ((winned_at || updated_at || Time.zone.now) - created_at).to_i.seconds\n end",
"def -(other)\n ((@time - other.time) / SECONDS_IN_A_DAY).to_i\n end",
"def calculate_time_left(difference)\n calculated_time = {}\n \n #calculating the days, hours, minutes and seconds out of the initial large seconds value generated by 'difference', then dropping the values after the decimal\n calculated_time[:days] = (difference / (60*60*24)).to_i\n days_remainder = (difference % (60*60*24)).to_i\n calculated_time[:hours] = (days_remainder / (60*60)).to_i\n hours_remainder = (days_remainder % (60 * 60)).to_i\n calculated_time[:minutes] = (hours_remainder / 60).to_i\n calculated_time[:seconds] = (hours_remainder % 60).to_i\n calculated_time\nend",
"def timestamp(now)\n (now - @start_time) * 1000\n end",
"def compare time_1, time_2\n return 1 if time_1.hour > time_2.hour\n return -1 if time_1.hour < time_2.hour\n return 1 if time_1.minute > time_2.minute\n return -1 if time_1.minute < time_2.minute\n\n 0\n end",
"def diff(other, callbacks = nil, &block)\n Diff::LCS.diff(self, other, callbacks, &block)\n end",
"def difference\n end",
"def time_days() (@time_end.jd - @time_start.jd) + 1; end",
"def duration\n finished? ? finished_at.to_f - started_at.to_f : -1\n end",
"def days_diff\n if self.end_at.nil? || self.start_at.nil?\n return 0\n end\n n = Integer(self.end_at - self.start_at) + 1\n return n\n end",
"def get_price_difference\n if( (Time.current - self.updated_at) >= 2 * Security.PRICE_UPDATE_CONSTANT )\n 0\n else\n self.price - self.past_price\n end\n end",
"def calculate_seconds_passed?(entry)\n incident_created_at = entry['incident']['created_at']\n most_recent_update = entry['created_at']\n start_time = Time.parse(incident_created_at)\n recent_time = Time.parse(most_recent_update)\n result = recent_time - start_time\n result >= 0 ? (return result) : (return 0)\n end",
"def subtractUTF(t1,t2)\n\t\tt3=subtract(t1,t2)\n\t\tdays = (t3 / (60*60*24)).floor\n\t\trest = (t3 % (60*60*24))\n\t\thours = (rest / (60*60)).floor\n\t\trest = (rest % (60*60))\n\t\tminutes = (rest / 60).floor\n\t\trest = (rest % 60)\n\t\tresult=[days,hours,minutes,rest]\n\t\treturn result\n\tend",
"def duration\n TingYun::Helper.time_to_millis(@exit_timestamp - @entry_timestamp)\n end",
"def age\n ((Time.now - self.started_at)/(3600*24)).to_i\n end",
"def time_length\n (end_time.hour - start_time.hour) + (end_time.min - start_time.min) / 60.0\n end",
"def duration; ((endtime()- starttime()) / 60).to_i; end",
"def lead_time\n (self.date_start.to_date - self.booking_date.to_date).to_i\n end",
"def to_seconds(timestamp)\n hours = timestamp[0].to_i * 60 * 60\n minutes = timestamp[1].to_i * 60\n hours + minutes + timestamp[2].to_i\nend",
"def diff(other)\n\t\tif other.is_a? Hash\n\t\t\treturn int_diff(other)\n\t\tend\n\t\t\n\t\tif other.is_a? RHash\n\t\t\treturn int_diff(other)\n\t\tend\t\n\tend",
"def days_till_old\n if modified_at.nil? || old?\n 0\n else\n (modified_at - self.class.old_in_days.days.ago.to_i)/(1.day.to_i)\n end\n end",
"def seconds_since_last\n last.blank? ? 0 : (timestamp_server - last.timestamp_server)\n end",
"def ago(other)\n since(-other)\n end",
"def elapsed_days t1\n ((Time.now - t1) / 86400).to_i + 1\n end",
"def elapsed_time\n if @start_time && @end_time\n @end_time - @start_time\n else\n nil\n end\n end",
"def cycle_time\n if completed_at && started_at\n completed_at - started_at\n else\n 0.0\n end\n end",
"def duration\n @duration ||= timestamp_delta / 256.0\n end",
"def dwell_time\n return nil if created_at.nil? || down_at.nil?\n\n # Convert the created at Time to date.\n created_at_date = Date.new(\n created_at.year,\n created_at.month,\n created_at.day\n )\n\n (down_at - created_at_date).to_i.abs\n end",
"def lead_time\n if completed_at\n completed_at - created_at\n else\n 0.0\n end\n end",
"def elapsed_time\n return nil if !started_at.present? || aborted_at.present?\n\n (finished_at.present? ? finished_at : Time.now) - started_at\n end",
"def duration\n (self.end_date.to_i/60/60/24) - (self.start_date.to_i/60/60/24)\n end",
"def duration\n self.end_time - self.start_time\n end",
"def diff((x1, y1), (x2, y2))\n [x1 - x2, y1 - y2]\nend",
"def sdiff(other, callbacks = nil, &block)\n Diff::LCS.sdiff(self, other, callbacks, &block)\n end",
"def duration\n ran? ? (completed_at || failed_at) - started_at : 0\n end",
"def diff_to_compare; end",
"def duration\n (finish - start)/3600\n end",
"def time_elapsed\n if !self.finished.blank?\n ((self.finished - self.started) / 60).to_i\n end\n end",
"def distance_of_time_as_array( dt_1, dt_2)\n difference = ((dt_1 > dt_2) ? (dt_1 - dt_2) : (dt_2 - dt_1))\n\n seconds = difference % 60\n difference = (difference - seconds) / 60\n minutes = difference % 60\n difference = (difference - minutes) / 60\n hours = difference % 24\n difference = (difference - hours) / 24\n days = difference % 7\n \n return [days, hours, minutes, seconds]\n end",
"def total_seconds\n (ends_at - Time.current).round\n end",
"def difference(rt1, rt2)\n return template_tags(rt1) - template_tags(rt2)\n end",
"def to_i\n return @tv_sec\n end",
"def -( other )\n TimeDelta.new( @msecs - other.msecs )\n end",
"def _get_timestamp_or_duration\n return [nil, nil] unless @buffer\n\n now, now_dup = _init_now_then\n @_mode = nil\n @buffer.each do |time_unit, time_value|\n now_dup = _apply(time_unit, time_value, now, now_dup)\n end\n return [nil, diff_in_seconds(now, now_dup)] if @_mode.nil?\n\n [now_dup.to_i, nil]\n end",
"def diff(other)\n HQMF::Measure::LogicExtractor.get_measure_logic_diff(self,other,true)\n end",
"def format_time_diff(tstart)\n diff = Time.now - tstart\n format_time(diff)\nend",
"def differences_between_arrays(first_points, second_points)\n YquotesSignalTools.truncate_to_shortest!(first_points, second_points)\n differences = []\n first_points.each_with_index { |fp, index| differences << fp - second_points[index] }\n differences\n end",
"def timestamps\n @timestamps ||= bytes[T_BYTES].each_slice(4).map do |t_bytes|\n ByteArray.to_i(t_bytes)\n end.reject{ |t| t == 0 }\n end",
"def diff(other, options={})\n Differences::Entry.compute(self, other, diff_criteria(options), options)\n end",
"def zone_to_diff(zone)\n if m = /\\A([+-](?:\\d{4}|\\d\\d:\\d\\d))\\z/.match(zone)\n x = m[1].gsub(':','')\n x[0..2].to_i*3600 + x[3..4].to_i*60\n else\n 0\n end\n end",
"def get_elapse_time\n @start_time ||= @time_now\n return @time_now - @start_time\n end",
"def pill_taking_time_diff(record)\n ptime = record.pill_time_at\n ctime = record.actual_pill_time_at\n\tif ctime == nil\n\t\treturn \"No submission\"\n\tend\n range = (ptime - 10.minutes)..(ptime + 10.minutes)\n\n if range.cover? ctime\n \"Taken on time\"\n elsif record.created_at < record.pill_time_at\n \"Taken early by #{distance_of_time_in_words(ctime, ptime)}\"\n else\n \"Taken late by #{distance_of_time_in_words(ctime, ptime)}\"\n end\n end",
"def angle_diff(angle1, angle2)\n end",
"def elapsed\n ms = duration\n s = ms.to_i\n ms = ((ms - s) * 1000).to_i\n h = s / 3600\n s = s % 3600\n d = h / 24\n h = h % 24\n m = s / 60\n s = s % 60\n return d, h, m, s, ms\n end"
] |
[
"0.73357546",
"0.72436374",
"0.711391",
"0.7014129",
"0.6795968",
"0.65606296",
"0.655958",
"0.64303714",
"0.64100695",
"0.6349768",
"0.6255267",
"0.62479687",
"0.6245177",
"0.62373",
"0.6197899",
"0.61782825",
"0.61694765",
"0.61478966",
"0.61191404",
"0.60989463",
"0.6080314",
"0.6071505",
"0.60489815",
"0.6027253",
"0.601895",
"0.59756565",
"0.5963319",
"0.594715",
"0.5943241",
"0.59202194",
"0.5906005",
"0.59056944",
"0.59014857",
"0.5874619",
"0.58674765",
"0.5850608",
"0.58398277",
"0.58347356",
"0.58211905",
"0.5815622",
"0.5791391",
"0.57864374",
"0.5770771",
"0.57399166",
"0.5736651",
"0.5718382",
"0.57039243",
"0.568607",
"0.5677467",
"0.5662563",
"0.5625894",
"0.5616545",
"0.5601565",
"0.5599643",
"0.5596708",
"0.5590031",
"0.5589656",
"0.55882233",
"0.5580019",
"0.5576563",
"0.5574444",
"0.55683404",
"0.5561181",
"0.55516684",
"0.5544146",
"0.5539855",
"0.55306596",
"0.5513299",
"0.5511695",
"0.54973567",
"0.5490671",
"0.54885906",
"0.547862",
"0.54765815",
"0.5464758",
"0.54633826",
"0.5460335",
"0.5456795",
"0.5456787",
"0.5451768",
"0.5446517",
"0.5441855",
"0.5440416",
"0.5426101",
"0.5425672",
"0.5419353",
"0.54188246",
"0.5401114",
"0.5396249",
"0.53960633",
"0.53952324",
"0.5394747",
"0.53887266",
"0.5385741",
"0.53808534",
"0.5379813",
"0.5379546",
"0.5377604",
"0.536797",
"0.5367658",
"0.53592944"
] |
0.0
|
-1
|
Use this to call any builtin object methods that are masked by ImmutableProxy's builtin object methods.
|
def send_to_target(name, *args, &block)
raise "Mutable methods not allowed" unless allowed?(name)
@target.__send__(name, *args, &block)
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def unboring_instance_methods\n if [::BasicObject,::Object,::Kernel].include? self\n self.instance_methods\n # elsif [Class,Module].include? self\n # self.instance_methods - Object.instance_methods\n else\n self.instance_methods - (::Object.instance_methods - self.local_instance_methods)\n end\n end",
"def getObjectProxy _obj, _args\n \"_obj getObjectProxy _args;\" \n end",
"def setObjectProxy _obj, _args\n \"_obj setObjectProxy _args;\" \n end",
"def method_missing(m, *args, &_block)\n set_accessors(m)\n if valid_proxy_method?\n define_proxy_method(m)\n return send(m, args.first)\n end\n\n super\n end",
"def method_missing(method, *args, &block)\n return super unless object.respond_to?(method)\n\n object.public_send(method, *args, &block)\n end",
"def method_missing(method_name, *args, &block)\n objects.send method_name, *args, &block\n end",
"def public_methods(all=true)\n __getobj__.public_methods(all) | super\n end",
"def methods(all=true)\n __getobj__.methods(all) | super\n end",
"def run\n methods = self.class.instance_methods - Object.instance_methods\n methods.grep(%r!^set_!).each do |v|\n send(v)\n end\n end",
"def method_missing(call)\n object.send(call)\n end",
"def method_missing(method, *args, &block)\n object.public_send(method, *args, &block)\n end",
"def method_missing(sym, *args)\n # Extend this object only when needed and immediately redefine\n # #method_missing so that the new version is used on all future calls.\n extensions.each {|e| extend(e) } if @extensions\n redefine_method_missing!\n __send__(sym, *args)\n end",
"def decorated_object_behavior\n #code\n end",
"def all(_obj)\n raise NotImplementedError\n end",
"def not_sandboxed_methods(include_superclasses = false, allowed_mixins=[], *disallowed_methods)\n\n __the_methods_to_check = public_instance_methods(false)\n puts \"#{self.name}: direct: #{__the_methods_to_check.inspect}\" if $DEBUG\n if include_superclasses\n clz = self.superclass\n while !clz.nil?\n unless clz == Object || (defined? BasicObject && clz == BasicObject)\n puts \"#{self.name}: #{clz.name}: #{clz.public_instance_methods(false).inspect}\" if $DEBUG\n __the_methods_to_check += clz.public_instance_methods(false)\n end\n clz = clz.superclass\n end\n \n if allowed_mixins.length > 0\n #we include any mixins\n for m in self.included_modules\n if allowed_mixins.include?(m)\n puts \"#{self.name}: #{m.name}: #{m.public_instance_methods(false).inspect}\" if $DEBUG\n __the_methods_to_check += m.public_instance_methods(false)\n end\n end\n end\n end\n \n __the_methods_to_check << \"nil?\".intern\n \n __the_methods_to_check.uniq!\n \n unless disallowed_methods.nil? || disallowed_methods.length == 0\n not_bang = false\n if disallowed_methods.include?(:bang_methods) #just remove all xxx! methods that modify in place\n __the_methods_to_check.reject! { |meth| meth.to_s[-1, 1] == \"!\"}\n not_bang = true\n end\n unless not_bang || disallowed_methods.length > 1\n __the_methods_to_check.reject! { |meth| disallowed_methods.include?(meth)}\n end\n end\n \n puts \"#{self.name}: #{__the_methods_to_check.inspect}\" if $DEBUG\n \n sandboxed_methods(*__the_methods_to_check)\n \n \n \n end",
"def method_missing(method, *args, &block)\n self.class.define_proxy method\n send(method, *args, &block)\n end",
"def method_missing(meth, *args, &block)\n if respond_to?(meth)\n object.__send__(meth, *args, &block)\n else\n super\n end\n end",
"def call(object); end",
"def __setobj__(obj)\n __raise__ ::NotImplementedError, \"need to define `__setobj__'\"\n end",
"def safe_send(obj, method, *args, &block)\n (Module === obj ? Module : Object).instance_method(method).bind(obj).call(*args, &block)\n end",
"def method_missing(m, *args, &block)\n target = self.__getobj__\n unless target.respond_to?(m)\n super(m, *args, &block)\n else\n target.__send__(m, *args, &block)\n end\n end",
"def all_convenience_methods\n @mutex.synchronize do\n @methods.keys\n end\n end",
"def methods_to_try(obj)\n ret = obj.methods.map(&:intern)\n blacklist = obj.is_a?(Module) ? CLASS_METHOD_BLACKLIST : INSTANCE_METHOD_BLACKLIST\n klass = obj.is_a?(Module) ? obj : obj.class\n\n klass.ancestors.each { |ancestor| ret -= blacklist[ancestor.to_s.intern] }\n\n # 1.8.7 lacks Symbol#<=>\n ret.sort_by(&:to_s)\n end",
"def safe_send(obj, method, *args, &block)\n (Module === obj ? Module : Object).instance_method(method).bind(obj).call(*args, &block)\n end",
"def safe_send(obj, method, *args, &block)\n (Module === obj ? Module : Object).instance_method(method).bind(obj).call(*args, &block)\n end",
"def delegate_object_reader_method; end",
"def setOvercast _obj, _args\n \"_obj setOvercast _args;\" \n end",
"def methods(inherited_too=true)\n ensure_apply_object_class\n target_names = @attr_methods.keys + @attr_aliases.keys\n target_names -= ['objectClass', Inflector.underscore('objectClass')]\n super + target_names.uniq.collect do |x|\n [x, \"#{x}=\", \"#{x}?\", \"#{x}_before_type_cast\"]\n end.flatten\n end",
"def method_missing(method_id, object)\n return object(object)\n end",
"def [](object)\n if running?\n load_features\n MethodProxy.new object, @proxy\n else\n raise NotRunning\n end\n end",
"def method_missing(method_name, *args, &block)\n if @object.respond_to? method_name.to_sym\n @object.__send__(method_name, *args, &block)\n else\n super\n end\n end",
"def method_missing(method, *args, &block)\n __proxy_result__.send(method, *args, &block)\n end",
"def flexmock_invoke_original(object, method, *args, &block)\n object.instance_variable_get(:@flexmock_proxy).proxy.flexmock_invoke_original(method, args, &block)\n end",
"def flexmock_invoke_original(object, method, *args, &block)\n object.instance_variable_get(:@flexmock_proxy).proxy.flexmock_invoke_original(method, args, &block)\n end",
"def method_missing(method, *args, &block)\n if @object.key? method\n @object[method]\n elsif @object.respond_to? method\n @object.send(method, *args, &block)\n end\n end",
"def method_missing (method, *args, &block)\n __proxy_result__.send(method, *args, &block)\n end",
"def original_method(obj, meth, klass=Object)\n klass.instance_method(meth).bind(obj).call\n end",
"def original_method(obj, meth, klass=Object)\n klass.instance_method(meth).bind(obj).call\n end",
"def *(obj)\n raise 'Not Implemented'\n end",
"def unhook_object!(object, methods = :all, hooks = :all)\n methods = (methods == :all) ? SINGLETON_METHOD_HOOKS[object].keys : (methods.is_a?(Symbol) ? [methods] : methods)\n raise ArgumentError(\"Valid arguments: :before, :after or :all\") unless [:before, :after, :all].include?(hooks)\n\n # Validate all methods exist before doing anything\n methods.each do |method|\n raise UndefinedHookException.new(\"No hook defined for class method #{method})\") unless SINGLETON_METHOD_HOOKS[object][method]\n end\n\n methods.each do |method|\n if hooks == :all\n object.metaclass.class_eval <<-REDEF_OLD_METHOD\n alias :\"#{method}\" :\"#{SINGLETON_HOOK_METHOD_PREFIX}#{method}\"\n undef :\"#{SINGLETON_HOOK_METHOD_PREFIX}#{method}\"\n REDEF_OLD_METHOD\n SINGLETON_METHOD_HOOKS[object].delete(method)\n else\n unless SINGLETON_METHOD_HOOKS[object][method][hooks]\n raise UndefinedHookException.new(\"No hook defined for singleton method #{method}) at #{hooks.inspect}\")\n end\n SINGLETON_METHOD_HOOKS[object][method].delete(hooks)\n SINGLETON_METHOD_HOOKS[object].delete(method) if SINGLETON_METHOD_HOOKS[object][method].empty?\n end\n end\n\n SINGLETON_METHOD_HOOKS.delete(object) if methods == :all or (SINGLETON_METHOD_HOOKS[object] || {}).empty?\n end",
"def method_missing(sym, *args, &block)\n proxy_target.__send__(sym, *args, &block)\n end",
"def get_methods(obj)\n meths = obj.methods.grep(/impl/)\n meths.map { |m| m.to_s.gsub('impl_', '') + '!' }\n end",
"def unboring_methods\n if [::Class,::Module].include? self\n # Only those instance methods that we have not by virtue of being an instance of ourself\n self.methods - (self.instance_methods - self.singleton_methods)\n elsif self.is_a? ::Class\n # Only those instance methods that we have not by virtue of being a Class, unless we have overridden them\n self.methods - (::Class.instance_methods - self.singleton_methods)\n else\n # Only those instance methods that we have not by virtue of being a Module, unless we have overridden them\n self.methods - (::Module.instance_methods - self.singleton_methods)\n end\n end",
"def method_missing(method, *args, &body)\n\n if in_set?([\"<<\",\"clear\",\"delete\"],method.to_s)\n\n # DEPRICATED CODE:\n if !call_if_exists(@parent, \"gr_#{@target.proxy_reflection.name.to_s}_w?\")\n guard_rails_error(\"Not authorized to use #{method.to_s} on read-only object\")\n end\n # END OF DEPRICATED CODE\n\n failed = false\n \n # If the parent, the proxy, the Rails proxy, or the array itself\n # fail to allow edits, then the edit methods will be prohibited\n\n if @parent.respond_to?(\"gr_can_edit?\")\n if !@parent.gr_can_edit?\n failed = true\n end\n end\n if @target.respond_to?(\"gr_can_edit?\")\n if !@target.gr_can_edit?\n failed = true\n end\n end\n if self.respond_to?(\"gr_can_edit?\")\n if !self.gr_can_edit?\n failed = true\n end\n end\n if @target.respond_to?(\"proxy_reflection\")\n if !@target.target.gr_can_edit?\n failed = true\n end\n end\n\n # << counts as an append method, so it can be performed, even without\n # write access, as long as one of the objects in quesiton actively \n # allows append access (note that not specifying append access at all)\n # does not count the same as directly saying append access is allowed,\n # as it would with any other type of access control policy\n\n if failed and method.to_s==\"<<\" \n if @parent.respond_to?(\"gr_can_append?\")\n if @parent.gr_can_append?\n failed = false\n end\n end\n if @target.respond_to?(\"gr_can_append?\")\n if @target.gr_can_append?\n failed = false\n end\n end\n if self.respond_to?(\"gr_can_append?\")\n if self.gr_can_append?\n failed = false\n end\n end\n if @target.respond_to?(\"proxy_reflection\")\n if @target.target.gr_can_append?\n failed = false\n end\n end\n # If the method in question is << and there is no\n # append access, the violation policy for append access\n # will be run.\n\n # TODO: check to make sure that this line cannot be \n # reached if << was called but no append access annotation\n # was defined\n return eval_violation(:append_access) if failed\n end\n \n # If the method is blocked for write access reasons, call\n # the appropriate violation method\n eval_violation(:write_access) if failed\n end\n\n # DEPRICATED CODE: Allows specific methods to be singled out as allowed\n # or not based on gr_ methods in the object that \"owns\" this reflection.\n # This feature is currently unused, but might be useful\n special_function = \"gr_#{@target.proxy_reflection.name.to_s}_#{method.to_s}\"\n if @parent.respond_to?(special_function)\n if !@parent.send(special_function)\n guard_rails_error(\"Not authorized to use #{method.to_s} on this object\")\n end\n end\n # END OF DEPRICATED CODE\n\n # If 'find' is used on the plural association, make sure that none of its\n # parameters contain unsafe SQL (aka call the appropriate 'transform' method\n # on that string).\n if method.to_s == \"find\"\n args = clean_args(args)\n end\n\n # As with all objects potentially pulled from the database, policies need\n # to be set up immediatly\n target.each do |obj|\n obj.populate_policies\n end\n \n # If no errors have been raised up to this point, pass the method on to the\n # target, along with any arguments or block that may have been provided\n if block_given?\n @target.target.send(method,*args,&body)\n else\n @target.send(method,*args,&body)\n end\n end",
"def method_missing(sym, *args, &block)\n proxy_target.__send__(sym, *args, &block)\n end",
"def method_missing(meth, *args, &block)\n @object.send(meth, *args, &block)\n end",
"def try_helper(method, obj)\n if obj.respond_to?(method)\n obj.send(method)\n end\n end",
"def proxy meth\n Proxy.new self, meth\n end",
"def process_method_object_options(args, opts)\n opts[:instance] = opts['instance-methods'] if opts.m?\n # TODO: de-hack when we upgrade Slop: https://github.com/injekt/slop/pull/30\n opts.options[:super].force_argument_value opts.options[:super].count if opts.super?\n method_obj = get_method_or_raise(args.empty? ? nil : args.join(\" \"), @method_target, opts.to_hash(true))\n opts.on(:method_object, :default => method_obj)\n end",
"def __setobj__\n raise \"ObjectProxy does not support changing referenced object\"\n end",
"def !\n !__getobj__\n end",
"def method_missing(symbol, args)\n\t@obj.send(symbol, args)\nend",
"def __getobj__\n __raise__ ::NotImplementedError, \"need to define `__getobj__'\"\n end",
"def proxy; end",
"def proxy; end",
"def proxy; end",
"def mod _obj, _args\n \"_obj mod _args;\" \n end",
"def api\n methods - Object.public_methods\n end",
"def method_conflicts\n (@module.instance_methods + @module.private_instance_methods) &\n (Boson.main_object.methods + Boson.main_object.private_methods)\n end",
"def underscorize\n dup.tap(&:underscorize!)\n end",
"def proxy; self end",
"def __send_for_obj__ obj,sym,*args\n sym = sym.to_sym\n voff = Class.method_to_voff[sym]\n if !voff\n # FIXME: This needs to change once we handle \"define_method\"\n return obj.method_missing(sym, *args)\n else\n # We can't inline this in the call, as our updated callm\n # doesn't allow method/function calls in the method slot\n # for simplicity, for now anyway.\n %s(assign raw (callm voff __get_raw))\n %s(callm obj (index self raw) ((splat args)))\n end\n end",
"def freeze\n __getobj__.freeze\n super()\n end",
"def make_public(obj)\n obj.private_methods.each do |method_name|\n obj.singleton_class.class_eval { public method_name }\n end\n obj\nend",
"def extend_object(obj) end",
"def method_missing(method, *args, &blk); end",
"def object_methods(obj)\n @object_methods[obj] ||= {}\n end",
"def refute_respond_to(obj, meth, msg = T.unsafe(nil)); end",
"def invoke(object, method_name)\n object.send(method_name)\nend",
"def method_missing(m, *args, &block)\n # Attempt to call a method on the object directly\n if(method_names.index(m.to_s))\n $DEBUG and puts(\"Attempting to invoke_method '#{m}'\")\n invoke_method(get_method(m.to_s), args)\n else\n super.method_missing(m, *args, &block)\n end\n end",
"def method_missing(method_name, *args) self end",
"def method_missing(method_name, *args) self end",
"def __setobj__(obj); end",
"def hooked_singleton_methods_for(object)\n SINGLETON_METHOD_HOOKS[object] || {}\n end",
"def initialize(obj, mock, safe_mode)\n @obj = obj\n @mock = mock\n @method_definitions = {}\n @methods_proxied = []\n unless safe_mode\n add_mock_method(@obj, :should_receive)\n MOCK_METHODS.each do |sym|\n unless @obj.respond_to?(sym)\n add_mock_method(@obj, sym)\n end\n end\n end\n end",
"def unproxify(*methods)\n methods = proxified_methods.keys if methods.empty?\n\n self.proxified_methods = proxified_methods.except(*methods)\n\n methods.each { |method| remove_proxy_method(method) }\n end",
"def initialize(obj, mock, safe_mode)\n @obj = obj\n @mock = mock\n @method_definitions = {}\n @methods_proxied = []\n unless safe_mode\n add_mock_method(:should_receive)\n MOCK_METHODS.each do |sym|\n unless @obj.respond_to?(sym)\n add_mock_method(sym)\n end\n end\n end\n end",
"def safe_methods\n SafeClass.safe_methods_for(self)\n end",
"def protected_methods(all=true)\n __getobj__.protected_methods(all) | super\n end",
"def method_missing(method, *args)\n api_obj.send(method, *args)\n end",
"def method_missing(method, *arguments, &block); end",
"def do_target_object_map\r\n (to = self[:obj]) && to.map_foorth_exclusive(@symbol)\r\n end",
"def real_object\n result = self\n while result.respond_to?(:__getobj__)\n result = result.__getobj__\n end\n result\n end",
"def patched_methods\n @patched_methods ||= {}.with_indifferent_access\n end",
"def allowGetIn _obj, _args\n \"_obj allowGetIn _args;\" \n end",
"def inherited_meths(opts = T.unsafe(nil)); end",
"def global_proxy_exclusion\n super\n end",
"def unmemoized_instance_method(method_name); end",
"def public_instance_methods(include_super=true) end",
"def method_missing(name, *args)\n if wrapped_object.respond_to? name\n wrapped_object.send name, *args\n else\n super\n end\n end",
"def method_missing(method, *args, &block); end",
"def protected_instance_methods(arg0, arg1, *rest)\n end",
"def method_missing(method, *args, &block)\n super unless original_self\n original_self.send method, *args, &block\n end",
"def instance_method_list\n warn '#instance_method_list is obsoleted, please use #instance_methods'\n @instance_methods ||= method_list.reject { |a| a.singleton }\n end",
"def method_missing(name, *args, &block)\n proxy = if main.respond_to?(name)\n main\n elsif builtin.respond_to?(name)\n builtin\n else\n super(name, *args)\n end\n result = if proxy.is_real_method?(name)\n proxy.__send__(name, *args)\n else\n proxy.__send__(:method_missing, name, *args)\n end\n\n if block\n block.call(result)\n else\n result\n end\n end",
"def impersonate_methods(actual_object, *methods)\n raise Impersonator::Errors::ConfigurationError, 'You must start a recording to impersonate objects. Use Impersonator.recording {}' unless @current_recording\n\n ::Impersonator::Proxy.new(actual_object, recording: current_recording, impersonated_methods: methods)\n end",
"def unop!(meth); map!{|i| meth.call(i)}; end",
"def own_methods\n (self.methods - Object.methods)\n end",
"def plain_old_ruby_object\n\t\treturn false\n\tend",
"def callExtension _obj, _args\n \"_obj callExtension _args;\" \n end",
"def w(obj)\n obj.public_methods.sort - Object.methods\nend"
] |
[
"0.61126834",
"0.5994068",
"0.5979825",
"0.59736365",
"0.59670377",
"0.5911224",
"0.5899633",
"0.58934534",
"0.58731234",
"0.58501977",
"0.58149904",
"0.57679796",
"0.57244265",
"0.57203865",
"0.5716093",
"0.57036877",
"0.5686787",
"0.5656572",
"0.5639709",
"0.56335324",
"0.5612448",
"0.5609231",
"0.5608805",
"0.5594926",
"0.5594926",
"0.5589412",
"0.5584085",
"0.5578528",
"0.5551143",
"0.5529556",
"0.55276734",
"0.5517517",
"0.55022025",
"0.55022025",
"0.5494603",
"0.5494467",
"0.54824734",
"0.54824734",
"0.54769975",
"0.54662985",
"0.5466111",
"0.54593915",
"0.5437509",
"0.5431466",
"0.542577",
"0.5421884",
"0.54204845",
"0.54174477",
"0.5409148",
"0.53948885",
"0.53791744",
"0.53766423",
"0.5372614",
"0.53723294",
"0.53723294",
"0.53723294",
"0.5367798",
"0.5366366",
"0.5365271",
"0.5346338",
"0.5301788",
"0.5297507",
"0.5280802",
"0.5269",
"0.5264382",
"0.5263442",
"0.5261792",
"0.5261029",
"0.5259088",
"0.52570873",
"0.5254012",
"0.5254012",
"0.5248063",
"0.52439845",
"0.5243201",
"0.5242098",
"0.5241986",
"0.5240192",
"0.5236426",
"0.52303576",
"0.52223843",
"0.5218504",
"0.52168286",
"0.5212429",
"0.52013874",
"0.51983184",
"0.5193267",
"0.51930785",
"0.5189205",
"0.517389",
"0.5173775",
"0.517177",
"0.51691914",
"0.51663744",
"0.5163927",
"0.51594436",
"0.5151162",
"0.5140565",
"0.5137466",
"0.51339906",
"0.5131565"
] |
0.0
|
-1
|
and returns true if the string includes all the letters in the alphabet and false otherwise "the quick brown fox jumps over the lazy dog"
|
def string_has_all_letters(str)
result = {}
str.each_char do |c|
return true if result.length == 26
if c == " "
next
elsif !result.include?(c)
result[c] = c
end
end
result.length == 26 ? true : false
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def all_letters? str\r\n str[/[a-zA-z]+/] == str\r\nend",
"def letters?(word)\n\t\t# Split word and check if each letter is within the range a-z\n\t\tword.split('').each do |letter| # Use each loop as it is slightly better in performance, 'letter' iterator is encapsulated in a way that it cannot be accessed beyond each loop\n\t\t\treturn false unless(START_LETTER..END_LETTER).include? letter\n\t\tend\n\t\ttrue # All characters in word are letters from a-z\n\tend",
"def valid_word?(word, inner_letter, outer_letters)\n word.include?(inner_letter) && word.chars.uniq.all? {|l| @all_letters.chars.include?(l) }\n end",
"def include?(word, letters)\n word.upcase.chars.all? { |letter| word.count(letter) <= letters.count(letter) }\n end",
"def letter?(string)\n string =~ /[A-Za-z]/\n end",
"def is_all_this_letter?(word, character)\n a = is_all_as?(word)\n\n end",
"def letter?(s)\n\ts =~ /[A-Za-z]/\nend",
"def is_letter?(l)\n return l.match?(/[A-Za-z]/)\nend",
"def all_letters(str)\r\n # Use 'str[/[a-zA-Z]*/] == str' to let all_letters\r\n # yield true for the empty string\r\n str[/[a-zA-Z]+/] == str\r\nend",
"def alphabetical?(str)\r\n ('A'..'Z').to_a.include? str.upcase\r\nend",
"def letter_in_word?(letter)\n end",
"def letter?(letter)\n letter =~ /[[:alpha:]]/\n end",
"def is_letter?(character)\n character =~ /[[:alpha:]]/ ? true : false\nend",
"def has_letter(c, word)\n word.each_char do |i|\n if i == c\n return true\n end\n end\n false\nend",
"def letter?(char)\n char[/[a-zA-Z]/] == char\nend",
"def word_contained_in_letters?(word, letters)\n word.chars.all? do |char|\n if pos = letters.index(char)\n letters.delete_at(pos)\n true\n else\n false\n end\n end\nend",
"def pangram?(str)\n letters = \"abcdefghijklmnopqrstuvwxyz\".split(\"\")\n\n str.downcase!\n\n letters.all? { |letter|\n str.include? (letter)\n }\nend",
"def is_letter?(letter)\r\n (Alphabet.include? letter.downcase) && (letter.length == 1)\r\n end",
"def is_alpha(c)\n c =~ /[a-z]/i\nend",
"def is_alpha(c)\n c =~ /[a-z]/i\nend",
"def is_alpha(c)\n c =~ /[a-z]/i\nend",
"def string_check(elem)\n (('a'..'z').to_a + ('A'..'Z').to_a).each { |k| return true if elem.include?(k) }\n false\n end",
"def is_alpha(char)\n char.match?(/[a-zA-Z]/)\n end",
"def panagram2?(string)\n ('a'..'z').all? { |letter| string.downcase.include?(letter) }\nend",
"def is_isogram(string)\n return false if string.class != String\n string = string.downcase\n used_letters = []\n string.split(\"\").each do |e|\n if used_letters.include?(e)\n return false\n else\n used_letters << e\n end\n end\n return true\nend",
"def not_in_alphabet?(letter)\n !UPPERCASE.include?(letter) && !LOWERCASE.include?(letter)\nend",
"def repeating_letters?(str)\n dStr = str.downcase.chars\n checkStr = \"\"\n\n dStr.each do |char|\n if checkStr.index(char) != nil\n return true\n else\n checkStr += char\n end\n end\n false\nend",
"def letter?(lookAhead)\n if lookAhead =~ /[[:alpha:]]/\n return true\n else\n return false\n end\nend",
"def isLetter(c) \n /[A-Za-z]/ =~ c\n #return ((?a <= c and c <= ?z) or (?A <= c and c <= ?Z))\n end",
"def starts_and_ends_with_same_letter?(word)\nend",
"def valid_letter?(input)\n input.length == 1 && input.match?(/^[[:alpha:]]+$/)\n end",
"def contains?(char)\r\n any_letter_found = false\r\n each do |letter|\r\n if char.upcase == letter.content\r\n letter.state = :found\r\n any_letter_found = true\r\n end\r\n end\r\n any_letter_found\r\n end",
"def panagram?(string)\n alphabet = \"abcdefghijklmnopqrstuvwxyz\"\n string = string.gsub(/[^A-Za-z]/, \"\")\n string.each_char do |chr|\n alphabet.gsub!(chr, \"\")\n end\n alphabet == \"\"\nend",
"def only_chars?(text)\n text.match(/^\\w[\\w| ]+$/i)\n end",
"def isAlpha(char)\n alph = \"abcdefghijklmnopqrstuvwxyz\"\n for ch in (0..alph.length - 1)\n\tif ((char.to_s.downcase <=> alph[ch]) == 0)\n\t return true\n\tend\n end\n return false\nend",
"def check_a_string_for_special_characters(string)\n\t# match() is used to check if there is any match as in the regexp. /\\W/ means non word characters.\n\tstring.chars.select {|char| char.match(/\\W/) }.any?\nend",
"def check_a_string_for_special_characters(string)\n true_or_false = false\n # collect all non-special characters\n chars = []\n chars = ('a'..'z').to_a + ('A'..'Z').to_a + (0..9).to_a + (0..9).to_a + [\"0\",\"1\",\"2\",\"3\",\"4\",\"5\",\"6\",\"7\",\"8\",\"9\"]\n chars << \" \"\n # check each part of string to see if any special characters\n string.split(\"\").each {|character|\n if !chars.include?(character)\n true_or_false = true\n else\n next\n end\n }\n true_or_false\nend",
"def repeating_letters?(str)\r\n # your code goes here\r\n str = str.downcase.split(\"\")\r\n temp = \"\"\r\n i = 0\r\n while i < str.length\r\n if temp.include?(str[i])\r\n return true\r\n else\r\n temp << str[i]\r\n end\r\n i += 1\r\n end\r\n false\r\nend",
"def letter?(lookAhead)\n lookAhead =~ /[A-Za-z]/\n end",
"def is_word?(word)\r\n word = word.downcase\r\n word.each_char { |c| return false if not is_letter?(c) }\r\n !word.empty?\r\n end",
"def is_alpha?(c)\n # Match alpha character and make bool (via double negation !!)\n c && !!c.match(/^[[:alpha:]]$/)\n\n # !!c.match(/[a-zA-Z0-9]/)\n end",
"def is_a_letter?(guess)\n return true if @word.include?(guess)\n return false\n end",
"def is_letter_space(text)\n alphabet = \"abcdefghijklmnopqrstuvwxyz \"\n text_array = []\n text.chars.each do |char|\n if alphabet.index(char) == nil\n text_array << false\n else\n text_array << true\n end\n end\n\n if !text_array.include?(false)\n return true\n end\nend",
"def isAlpha(c)\n ('A' .. 'Z').include?(c.capitalize)\nend",
"def is_lowercase(string)\n lowercase = ('a'..'z').to_a\n lowercase.include? string\nend",
"def alphabetic(s)\n s.chars.sort.join == s\nend",
"def filter_invalid_word(word)\n # Define a string which includes all valid letter\n letters = \"abcdefghijklmnopqrstuvwxyz\"\n # Define return variable and give a default value\n valid = true\n # transfer the word to lowercase and take out off \\r\\n\n word = word.chomp.downcase\n # set return value as false if the length of word not exactly equal 5\n if (word.split(//).size != 5)\n valid = false\n end\n # loop each word\n word.split(//).each do |letter|\n # If the letter occurs more than once in the word, set return value to false\n if (word.count(letter.to_s)) > 1 \n valid = false\n end\n # If the letter does not included in valid letter, set return value to false\n if (letters.include?(letter) == false) \n valid = false\n end\n end\n # return a bool value to method\n return valid\n end",
"def panagram?(string)\n alphabet_arr = ('a'..'z').to_a\n \n alphabet_arr.each do |element|\n return false if string.downcase.include?(element) == false\n end\n\n true\nend",
"def alphanumeric?(string)\n string.match /\\A[a-zA-Z\\d]+\\z/\nend",
"def repeating_letters?(str)\n # your code goes here\n str.each_char do |ch|\n if str.downcase.count(ch) > 1\n return true\n end\n end\n false\nend",
"def alphanumeric?(str)\n return false if str.empty? || (str.length == 1 && str[0] == ' ')\n match = (str.match /^[a-zA-Z\\d]+$/)\n match != nil\nend",
"def block_word?(string)\n hash = { 'B' => 'O', 'G' => 'T', 'V' => 'I', \n 'X' => 'K', 'R' => 'E', 'L' => 'Y',\n 'D' => 'Q', 'F' => 'S', 'Z' => 'M',\n 'C' => 'P', 'J' => 'W', 'N' => 'A', 'H' => 'U'}\n\n banned_letters = []\n bool = true\n letters = string.upcase.chars\n letters.each do |letter|\n if hash.has_key?(letter)\n banned_letters << hash[letter]\n elsif hash.has_value?(letter)\n banned_letters << hash.key(letter)\n end\n end\n banned_letters.each do |letter|\n if letters.include?(letter)\n bool = false\n break\n else\n next\n end\n end\n bool\nend",
"def is_isogram(word)\n new_word = word.downcase.split('')\n letter_list = []\n for letter in new_word\n if letter_list.include?(letter)\n return false\n else \n letter_list << letter\n end\n end\n return true\nend",
"def check_a_string_for_special_characters(string)\n string =~ /\\W/\nend",
"def valid_word?(word, letters)\n # take the word from user and split into Array of each letter\n input_word = word.upcase.split('')\n # iterate over the array of letters then\n # .all? checks each condition, returns T/F >>>\n # count of characters in input and the array @letters\n input_word.all? { |letter| input_word.count(letter) <= letters.count(letter) }\n end",
"def all_letters_uniq?\n @my_word.chars.each do |letter|\n return false if (@my_word.count letter) > 1\n end\n true\nend",
"def letter?(lookAhead)\n\t\tlookAhead =~ /[A-Za-z]/\n\tend",
"def isIdentifier(str)\n digitsOfAlphabet = getAlphabet[0] # if its a digit\n charsOfAlphabet = getAlphabet[1] # if a letter\n\n # first character in name cannot be digit\n if digitsOfAlphabet.include?(str[0])\n return false\n end\n\n # if keyword includes str\n if getKeywords.include?(str)\n return false\n end\n\n # check is in the alphabet\n for i in 1..str.size-1\n character = str[i]\n\n unless digitsOfAlphabet.include?(character) or charsOfAlphabet.include?(character)\n return false\n end\n end\n\n return true\nend",
"def is_lower?\n r = 'a'..'z'\n self.split('').keep_if{ |c| r.include?(c) }.count == self.length\n end",
"def word_has?(letter)\n @word.include? letter\n end",
"def check_a_string_for_special_characters(string)\n string.scan(/\\W+/).length > 0\nend",
"def check_a_string_for_special_characters(string)\n string.include?('@') ||\n string.include?('!') ||\n string.include?('£') ||\n string.include?('$') ||\n string.include?('%')\nend",
"def charIsAlpha\n c = getChar\n return ((c >= 'A') && (c <= 'Z')) || ((c >= 'a') && (c <= 'z'))\n end",
"def unique_letters(str)\n # store a key/value pair of each letter in a word where\n # key = letter and value = nil or 0 or 1\n letter_count = {}\n has_unique_letters = true\n\n # iterate through each letter of a word\n str.chars.each do |letter|\n # check if the count of the letter has already been incremented\n # if no, add 1\n # if yes, return false\n if letter_count[letter].nil? || (letter_count[letter]).zero?\n letter_count[letter] = 1\n else\n has_unique_letters = false\n end\n end\n\n # return status of a\n has_unique_letters\n end",
"def alnum?(str)\n str =~ /[[:alnum:]]/\nend",
"def word_char?(char)\n char =~ /\\w+/\n end",
"def alnum?(str)\n\t\t\tstr.match /^[[:alnum:]]$/\n\t\tend",
"def pangram?(s)\n y = s.downcase.split(//).sort.uniq\n y.keep_if { |i| i =~ /[a-z]/ }\n y.length == 26 ? true : false\nend",
"def letter?(lookAhead)\r\n\t\tlookAhead =~ /^[a-z]|[A-Z]$/\r\n\tend",
"def letters(arg)\n return tuc(arg.sub(/[eiou]/, 'a')) != \"\"\n end",
"def missing_letters(string)\n missing_letters = []\n str = string.downcase\n ('a'..'z').each do |c|\n missing_letters << c if !str.include?(c)\n end\n missing_letters\nend",
"def only_alpha(string)\r\n\ts = string.split(\"\")\r\n\tstring1=\"\"\r\n\ts.each do |letter|\r\n\t\tif (letter >= \"a\" && letter <= \"z\") || (letter >= \"A\" && letter <= \"Z\")\r\n\t\t\tstring1 += letter\r\n\t\tend\r\n\tend\r\n\tputs string1\r\nend",
"def danglers?\n !valid_word?(@str[@first_letter..@last_letter]) && @last_letter == @str.length && @words.join.length < @str.length\n end",
"def repeating_letters?(str)\n str.downcase.chars.uniq.length != str.length\nend",
"def word_has?(letter)\n @word.include?(letter)\n end",
"def alpha(x)\n c = x.downcase\n is_alpha?(c) ? c : false\n end",
"def word_has?(letter)\r\n @word.include?(letter)\r\n end",
"def missing_letters(string)\n return ('a'..'z').to_a.select {|e| !string.include?(e)}\nend",
"def missing_letters(string)\n alphabet='a'..'z'\n alphabet.reject {|el|string.downcase.include?(el)}\nend",
"def test_6_accepts_includes_non_alphanumeric\n result = is_alphanumeric?(\"1Abjils&\")\n assert(result, \"1ABjils& has non alphanumeric, should be valid\")\n end",
"def repeating_letters?(str)\n i = 0\n while i < str.length - 1\n return true if str[i].downcase == str[i + 1].downcase\n i += 1\n end\n false\nend",
"def panagram?(str)\n alph = ('a'..'z').to_a\n\n str.downcase.chars.each do |char|\n if alph.include?(char)\n alph.delete(char)\n end\n end\n\n alph.empty?\nend",
"def panagram?(string)\n return false if string.length < 26\n\n letters = string.downcase.gsub(/[^a-z]/, '')\n letters.chars.uniq.count >= 26\nend",
"def clean?(string)\n !all.any? { |s| string.downcase.include? s }\n end",
"def panagram?(string)\n return false if string.gsub(/[^a-zA-Z]/, '').length < 26\n seen = string.split(\"\").reduce([]) do |seen, char|\n char = char.downcase\n seen.push(char) if !seen.include?(char) && char =~ /[a-z]/\n seen()\n end\n !(seen.length < 26)\nend",
"def missing_letters(string)\n alphabet = (\"a\"..\"z\").to_a\n alphabet.reject { |x| string.downcase.include?(x) }\nend",
"def is_latin_only?(input)\n !!(input.match(LATIN_ALPHABET_PLUS_SYMBOLS))\n end",
"def repeating_letters?(str)\n # your code goes here\n str.downcase.chars.uniq.length != str.downcase.chars.length\n\n end",
"def letters_only(case_matters=false)\n just_letters = split('').find_all do |char| \n DUAL_CASE_ALPHABET.include?(char) \n end.join('')\n return just_letters if (case_matters)\n return just_letters.downcase\n end",
"def is_good?(letter)\n @letters.include?(letter)\n end",
"def alphabet_only(str)\n return str.gsub(/[^abcdefghijklmnopqrstuvwxyz]/, '')\n end",
"def starts_with_consonant?(s)\n !!(s =~ /^[A-Z&&[^AEIOU]]/i)\nend",
"def alphanumeric_words\n /([a-zA-Z0-9]*\\w)/\n end",
"def repeating_letters?(str)\n str = str.downcase\n i = 0\n while i < str.length - 1\n return true if str[i] == str[i + 1]\n i += 1\n end\n false\nend",
"def letters\n the_letters = []\n letter_regex = /[a-z]/i\n chars.each do |character|\n the_letters << character if character.match(letter_regex)\n end\n the_letters.join\n end",
"def block_word?(string)\nblocks = [['B','O'], ['X','K'], ['D','Q'], ['C','P'], ['N','A'], ['G','T'], ['R','E'], ['F','S'], ['J','W'], ['H','U'],\n['V','I'], ['L','Y'], ['Z','M']]\nresult = []\nchars_array = string.chars\nblocks.each do |sub_array|\n chars_array.each do |letter|\n if sub_array.include?(letter.upcase)\n result << sub_array\n end\n end\nend\nresult == result.uniq ? true : false\nend",
"def alpha_search(str)\r\n\r\nend",
"def verify_lowercase(password)\n lowercase_letters = ('a'..'z')\n lowercase = false\n password.split(\"\").each do |letter|\n if lowercase_letters.include?(letter)\n lowercase = true\n end\n end\n return lowercase\nend",
"def is_special?\n # Only consonants\n return true if self =~ /^[bcdfghjklmnpqrstvwxz]+$/i\n\n # Not a word\n return false if self =~ /[^\\s]\\s[^\\s]/;\n return false if self.length < 3;\n # Alphanumeric\n return true if self =~ /[0-9]/ && self =~ /[a-z]/i\n # All Caps\n return true if self =~ /[A-Z]{2,}/;\n # Caps Mix\n return true if self =~ /[a-z][A-Z]/;\n # All consonants\n return true if self =~ /^[a-z]$/i && self !~ /[aeiou]/i\n # Dashed word\n return true if self =~ /(^\\w-|-\\w$)/\n # To many consonants (very heuristic)\n if self =~ /([^aeiouy]{3,})/i && !CONSONANTS.include?($1.downcase)\n return true\n end\n\n return false\n end",
"def letter_match?(letter)\n if @secret.include?(letter)\n return true; end\n return false\n end"
] |
[
"0.8151747",
"0.76416415",
"0.7584866",
"0.7579621",
"0.75392896",
"0.7463846",
"0.7459924",
"0.74596894",
"0.73800147",
"0.7341007",
"0.7337289",
"0.7306161",
"0.73020756",
"0.7297379",
"0.7256316",
"0.7241348",
"0.7197539",
"0.7149373",
"0.7136688",
"0.7136688",
"0.7136688",
"0.7122292",
"0.7100352",
"0.7075267",
"0.7059148",
"0.7036504",
"0.7017185",
"0.70148915",
"0.7013449",
"0.69789916",
"0.6963597",
"0.6945628",
"0.6939851",
"0.6927727",
"0.69213",
"0.69174933",
"0.69114006",
"0.690545",
"0.6883798",
"0.68779325",
"0.68758565",
"0.68650526",
"0.6859735",
"0.6852817",
"0.684996",
"0.68348235",
"0.68336606",
"0.68206406",
"0.68076634",
"0.67943066",
"0.6788828",
"0.6788089",
"0.6787767",
"0.67649335",
"0.6756429",
"0.6742392",
"0.67312413",
"0.6729712",
"0.6723413",
"0.671435",
"0.66870654",
"0.6681933",
"0.6671728",
"0.66665524",
"0.66623497",
"0.66596854",
"0.6656004",
"0.6645647",
"0.6626191",
"0.66178375",
"0.65925115",
"0.6573626",
"0.6572104",
"0.6565846",
"0.65597373",
"0.6556255",
"0.6555865",
"0.6553534",
"0.65497553",
"0.65478086",
"0.65382904",
"0.6535811",
"0.65278816",
"0.6518401",
"0.651422",
"0.6513871",
"0.65113515",
"0.6503357",
"0.65010613",
"0.6495034",
"0.6484393",
"0.64838237",
"0.6480523",
"0.6477094",
"0.6475062",
"0.6474822",
"0.64726907",
"0.6468879",
"0.64672375",
"0.64512026"
] |
0.7732584
|
1
|
Retourne le premier mot
|
def first_word (a)
return a.split[0]
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def motto; end",
"def Pot\n exp1, msg = self.Primario\n if exp1 then\n if @token.get_tipo == \"^\" then\n op, msg = le(\"^\")\n if op\n exp2, msg = self.Fator\n if exp2 then\n return Hash[\"tag\" => \"pot\", \"1\" => exp1, \"2\" => exp2]\n else\n return nil, msg\n end\n else\n return nil, msg\n end \n end\n return exp1\n end \n return nil, msg \n end",
"def povuci_mrezu\n return @mreza\n end",
"def word\n @word ||= Word.current_word('a-zA-Z0-9_.', :left).split('.').last\nend",
"def find_next_tunnel\n @position.mark(:minotaurus) \n @position.tunnels.reverse.each { |tunnel|\n if not tunnel.marks.include?(:minotaurus) # timto tunelem minotaur jeste nesel \n tunnel.mark(:minotaurus)\n return tunnel \n end\n }\n \n return @position.tunnels[rand(@position.tunnels.size)] \n end",
"def first_word (phrase)\n morceaux = phrase.split\n return morceaux[0]\nend",
"def term\n return unless lecture.present?\n lecture.term\n end",
"def preamp\n @ole.Preamp\n end",
"def getM()\n\t\t@conexion.gets.chomp\n\tend",
"def get_mission\n if $game_system.worldtip.nil?\n missiontext = MISSIONS[$game_variables[STORY_VAR_ID]]\n else\n missiontext = $game_system.worldtip\n end\n return Vocab.no_info if missiontext == nil\n missiontext\n end",
"def tu_turno\n Jugador.tu_turno\n end",
"def unit\n parts[3]\n end",
"def morphword\n @morphword\n end",
"def juega_primero mano\n\t\tcarta_a_jugar = nil\n\t\tcadena = (Util.captura_entrada \"Escoge una carta por su posicion en la mano (0=inicial)=>\").chomp\n\t\tcarta_a_jugar = mano[cadena.to_i]\n\t\treturn carta_a_jugar\t\n\tend",
"def winner\n @words.first\n end",
"def jugar\n ganador = case self.maquina_tirada\n when self.humano_tirada\n\t\"empatan\"\n when Ppt.jugadas_posibles[self.humano_tirada]\n\t\"humano\"\n else \n\t\"maquina\"\n end\n return ganador\n end",
"def verb\n\t\tif link = self.links.find {|link| link.llabel =~ /^(O([DFNTX]?)|P|BI|K|LI|MV|Q)[a-z\\*]*/ }\n\t\t\t# $deferr.puts \"Picking %s: LL of %p is %s\" % [ link.lword, link, link.llabel ]\n\t\t\treturn link.lword.sub( /\\.v$/, '' )\n\t\telsif link = self.links.find {|link| link.rlabel =~ /^(SI|S|AF)[a-z\\*]*/ }\n\t\t\t# $deferr.puts \"Picking %s: RL of %p is %s\" % [ link.rword, link, link.rlabel ]\n\t\t\treturn link.rword.sub( /\\.v$/, '' )\n\t\telse\n\t\t\treturn nil\n\t\tend\n\tend",
"def getMove()\n\t\t# Null warning\n\t\tputs(\"NO getMove() FUNCTION FOUND!!\")\n\t\treturn nil\n\tend",
"def phrase\n @phrase\n end",
"def first\n @poss[0]\n end",
"def preco\n @preco\n end",
"def gram_schmidtR\n gram_schmidt[1]\n end",
"def primary_mentor\n primaries = mentors.select{ |m| m.primary? }\n primaries.empty? ? mentors.first : primaries.first\n end",
"def get_move\n print \"#{@name}, please choose r, p, or s: \" \n @move = gets.chomp.downcase\n until is_move_valid(@move) do\n puts \"Sorry, that's not a valid move!\"\n @move = gets.chomp.downcase\n end\n @move\n end",
"def start\n @parts.first.start\n end",
"def result\n @s.first[1] || t.default\n end",
"def get_word\r\n return_word\r\n end",
"def set_mot new_mot\n\n # Différence de longueur entre le mot précédent et le nouveau mot.\n diff_len = new_mot.length - length\n\n old_mot_base = \"#{self.mot_base}\".freeze\n new_mot_base = Texte::Mot.get_mot_base(new_mot).freeze\n Tests::Log << <<-EOT\nold mot base : #{old_mot_base.inspect}\nnew mot base : #{new_mot_base.inspect}\n EOT\n\n # Il faut forcément le faire avant de mettre les nouvelles valeurs,\n # car l'occurence se sert de mot_base, par exemple.\n new_mot_base != old_mot_base && begin\n Tests::Log << <<-EOT\nOccurences[old_mot_base].mot = Occurences[#{old_mot_base}].mot = #{Occurences[old_mot_base] ? Occurences[old_mot_base].mot : '---'}\n EOT\n Occurences[old_mot_base].retire_mot(self)\n end\n\n # On peut véritablement modifier le mot.\n init\n @mot = @real_mot = new_mot\n @mot_base = new_mot_base\n @offset_correction = diff_len\n\n # Il faut ajouter le mot à sa nouvelle instance d'occurences\n # On ajoute le mot à cette occurence, comme les autres. La méthode +add+\n # se charge de tout, notamment de créer l'instance s'il le faut.\n Occurences.add(self)\n\n end",
"def parrot (phrase = \"Squawk!\")\n puts phrase\n\n #returns the phrase element\n return phrase\nend",
"def previous_word after\n RetortApi.get_word_before(after, identity: { medium: 'bible' })\nend",
"def machine\n\t\tif @argv.size < 1\n\t\t\treturn nil\n\t\tend\n\n\t\tif @argv[0] == '_'\n\t\t\treturn nil\n\t\tend\n\n\t\treturn @argv[0]\n\tend",
"def last_move; @last_move ||= match.moves.last; end",
"def word()\r\n\t\treturn @word\r\n\tend",
"def get_move\n @players[0].get_move(@board)\n end",
"def just_first\n first.split(' ')[0]\n end",
"def object\n\t\tlink = self.links.find {|link| link.rlabel[0] == ?O } or return nil\n\t\treturn link.rword.sub( /\\.[np]$/, '' )\n\tend",
"def part_of_speech\n\t\treturn self.class.postype_table[ self.pos.to_sym ]\n\tend",
"def get_Method(initial)\n\tinitial.split(' ',3)[0]\nend",
"def Terreno\n\t\treturn @terreno\n\tend",
"def get_kaduki_pose_direction\n #read pose information from module\n #using pose as key, access direction info that relates to current _suffix image\n \n #..It may be that I want to actually store this in the battler class and just\n #read it, as it will have already had to do this to get the image _suffix name\n #for the corresponding pose. Slightly more effective use of CPU time????\n msgbox \"Kaduki Type battlers are not currently supported!\"\n return get_direction; \n end",
"def get_turn\n @pos, @action = @player.prompt\n end",
"def get_swimmer_modifier\n modifier = 0.0\n age = @swimmer.get_swimmer_age(@goggle_cup.get_end_date)\n if age < @goggle_cup.age_for_negative_modifier\n modifier = @goggle_cup.negative_modifier\n else\n modifier = @goggle_cup.positive_modifier if age > @goggle_cup.age_for_positive_modifier\n end\n modifier\n end",
"def reflex\r\n Pronouns.get_pronoun_set(@object)[3]\r\n end",
"def sektion\n chapter\n end",
"def magic\n room = Room.random\n return self.magic if room.name == \"Entrance\" || room.name == \"Exit\"\n @game.current_room = room.name\n\n \"You moved to another room...\\n\"\n end",
"def get_nominee\n @data.search(\"div.award p strong a\").collect { |nominee| nominee.text }\n end",
"def telegram\n self.matched_line(\"TELEGRAM\")&.strip\n end",
"def get_music_destination\n dest = Readline::readline(\"Destination [G/c]: \")\n return dest.downcase == 'c' ? 'Classical' : 'General'\nend",
"def motion_name=(n)\n self.motion = meeting_section.meeting.committee.motions.find_by_name(\n n.slice( /^(?:R\\. \\d+\\: )?(.*)/, 1 ) )\n n\n end",
"def get_move\n print \"#{@name}, what is your letter? \"\n move = gets.chomp\n move\n end",
"def name\n self.text =~ /^(.*)$/\n return $1\n end",
"def citeste_parola\n Kernel.puts 'Tasteaza parola, aka tokenul tau Campfire:'\n @parola = Kernel.gets.strip\n end",
"def getTilte\r\n\t\t\t\t\treturn @title\r\n\t\t\t\tend",
"def mission\n \"To seek the Holy Grail\"\nend",
"def target\n @parts.last.target\n end",
"def current_step\n\t\t@current_step || steps.first\t\n\tend",
"def Fator\n if @token.get_tipo == \"-\" then\n op, msg = le(\"-\")\n if op\n exp2, msg = self.Fator\n if exp2 \n return Hash[\"tag\" => \"neg\", \"1\" => exp2]\n else \n return nil, msg\n end\n else \n return nil, msg\n end\n else\n pot, msg = self.Pot\n if pot\n return pot\n else\n return nil, msg\n end \n end\n end",
"def last_move\n @last_move\n end",
"def get_start_line\n puts \"\\nHello commuter. What line would you like to travel on?\"\n @start_line = gets.strip.chomp.downcase\n end",
"def pos(part_of_speech)\n case part_of_speech\n when 'noun'\n 'n'\n when 'adjective'\n 'adj'\n when 'verb'\n 'v'\n when 'adverb'\n 'adv'\n when 'interjection'\n 'interj'\n when 'pronoun'\n 'pron'\n when 'preposition'\n 'prep'\n when 'abbreviation'\n 'abbr'\n when 'auxiliary-verb'\n 'aux v'\n when 'conjunction'\n 'conj'\n when 'definite-article'\n 'definite article'\n when 'family-name'\n 'fam'\n when 'idiom'\n 'idiom'\n when 'imperative'\n 'imper'\n when 'noun-plural'\n 'n pl'\n when 'noun-posessive'\n 'n posessive'\n when 'past-participle'\n 'past part'\n when 'phrasal-prefix'\n 'phrasal prefix'\n when 'proper-noun'\n 'proper n'\n when 'proper-noun-plural'\n 'proper n pl'\n when 'proper-noun-posessive'\n 'proper n posessive'\n when 'verb-intransitive'\n 'vi'\n when 'verb-transitive'\n 'vt'\n else\n part_of_speech.nil? ? nil : part_of_speech.gsub('-',' ')\n end\n end",
"def getCurrentMonster() # : Monster\n @currentMonster\n end",
"def character\n ang_moment.character\n end",
"def secondo\n self[1]\n end",
"def get_meaning\n unless self.verbs.empty?\n self.verbs.first.meanings.first\n end\n end",
"def first_title\n title.first\n end",
"def prox(contrincante=nil) # :nodoc:\n if contrincante.nil?\n return @mano \n end\n jugada_oponente = contrincante.to_s # Copio la jugada del oponente\n jugada_oponente = $posibles_jugadas[jugada_oponente] \n @mano = Jugada.new(jugada_oponente) # La guardo como siguiente movimiento\n\n # Manejo si la jugada del oponente fue invalida \n if not @mano.jugada_valida?\n @mano = $sin_jugada\n end\n return @mano\n end",
"def last_word\n self.split(' ').last\n end",
"def lemma\n # TODO Return @string.stem\n @word.downcase\n end",
"def code\n\t self.video.try(:split, '/').try(:last) || ''\n\tend",
"def meme_phrase; end",
"def current_step\n fail 'Conversation without messages' if conversation.messages.count == 0\n\n # If the all messages in a conversation doesn't have a step,\n # it's the first interaction\n conversation.messages.where.not(step: nil).order(step: :desc).try(:first).try(:step)\n end",
"def get_smoking_pre(smoking_pre)\n SMOKING_PREFERRNCES[smoking_pre].to_s\n end",
"def prox() \n print \"Introduzca una jugada: \"\n entrada = gets\n entrada.downcase! # Lleva todo el input a minusculas \n entrada.capitalize! # Convierte a la primera letra en mayusculas\n entrada = entrada.slice(0,entrada.length - 1) # Quita el salto de linea\n @mano = Jugada.new(entrada) # Se crea una instancia de Jugada\n return @mano\n \n # Manejador si ocurre que el usuario ingresa una opcion no valida \n if not @mano.jugada_valida? \n @mano = $sin_jugada\n return \"Esta jugada no es válida, introduzca una opción que lo sea\"\n end \n end",
"def next_of current_player\n \t\tif current_player == self.player \n self.robot\n else\n self.player\t\n end\n \tend",
"def get_cur_mth\n\treturn $cur_mth.to_i\nend",
"def display_motto\n puts \"A LA LUNA!!!\"\n end",
"def which_token_next\n moves.count.even? ? \"X\" : \"O\" \n # if moves.count.even?\n # return \"x\"\n # else\n # return \"o\"\n # end\n end",
"def current_player()\n # Assume player X goes first\n return turn_count() % 2 == 0 ? \"X\" : \"O\"\n end",
"def mentor_title\n alternate_mentor_title.blank? ? \"Mentor\" : alternate_mentor_title\n end",
"def juega_primero mano\n\t\tcartas_palo_menor = obtiene_cartas_palo_menor mano\n\t\tif es_palo_corto cartas_palo_menor\n\t\t\tthen \n\t\t\tcarta_a_jugar = obtiene_carta_menor cartas_palo_menor\n\t\t\tUtil.muestra_msg \"Palo Corto=>\" + carta_a_jugar.carta.split(\"+\")[1] + \"\\n\", @verbose\n\t\telse\n\t\t\tcarta_a_jugar = obtiene_carta_menor mano\n\t\tend\n\t\treturn carta_a_jugar\n\tend",
"def get_paragraph\n wo = self.words_from_markov_data\n\n self.paragraph_from_words(wo).strip\n end",
"def get_winner_against(move)\n\t\tif move == :r\n\t\t\t:p\n\t\telsif move == :p\n\t\t\t:s\n\t\telse\n\t\t\t:r\n\t\tend\n\tend",
"def subj\r\n Pronouns.get_pronoun_set(@object)[0]\r\n end",
"def get_swimmer_name\n swimmer ? swimmer.get_full_name : '?'\n end",
"def get_swimmer_name\n swimmer ? swimmer.get_full_name : '?'\n end",
"def current_player_move\n @on_move == HUMAN_PLAYER ? human.move : computer.move\n end",
"def last_word\n @last_word_pronunciation.first.word.downcase\n end",
"def get_pokemon(pokemon)\n if pokemon.position >= 0\n return @actors[pokemon.position]\n end\n return @enemies[-pokemon.position-1]\n end",
"def current_player\n\tif $turn == 0\n\t\tplayerone\n\telse\n\t\tplayertwo\n\tend\nend",
"def get_move\n while true\n print \"#{@name}, make your move (e.g. e2 e4): \"\n move = gets.chomp.downcase.split\n break if valid_input?(move)\n puts \"Invalid input\"\n end\n return parse_move(move)\n end",
"def serial_noun\n @serial_noun ||= RandomWord.nouns.next\n end",
"def citeste_mesajul\n Kernel.puts('Scrie un mesaj:')\n return Kernel.gets.strip\n end",
"def presentacion\n \"La marca del Ventilador es #{@marca}\"\n end",
"def unit\n trigger.last\n end",
"def get_move\n puts \"Give me Order to move\"\n self.move = gets.chomp\n steps = move.upcase.split(//)\n puts \"get_move: steps = #{steps}\"\n return steps\n end",
"def next_objective\n next_obj = turn.listen.first\n return unless next_obj\n next_direction = turn.direction_of next_obj\n if turn.feel(next_direction).stairs?\n empty_directions = DIRECTIONS.select do |dir|\n turn.feel(dir).empty?\n end\n empty_directions.first\n else\n next_direction\n end\n end",
"def come\r\n 'rico'\r\n end",
"def distance\n @distance ||= mot_apres.offset - mot_avant.offset\n end",
"def jeuTermine\n\t\tlancementAventure(@tailleGrille+1)\n\tend",
"def recolectar_una\n\t\test = \"\"\n\t\tif @estado == ESTADO::MUERTE\n\t\t\test = \"El árbol está muerto\"\n\t\telse\n\t\t\tif @contador == 0\n\t\t\t\test = \"No hay más naranjas\"\n\t\t\telse\n\t\t\t\test = \"La naranja estaba deliciosa\"\n\t\t\t\t@contador -= 1\n\t\t\tend\n\t\tend\n\t\test\n\tend",
"def last_mover\n Player.find(self.last_player_id) if last_player_id.present?\n end"
] |
[
"0.63641864",
"0.6079449",
"0.58775055",
"0.5490274",
"0.54848814",
"0.5482",
"0.54228705",
"0.5412627",
"0.54090995",
"0.5408171",
"0.5401835",
"0.53941286",
"0.538468",
"0.5331964",
"0.53057104",
"0.53020376",
"0.527282",
"0.52647406",
"0.52609843",
"0.5253508",
"0.5251829",
"0.5233153",
"0.5228228",
"0.52274674",
"0.52166104",
"0.5215669",
"0.5209553",
"0.5201712",
"0.5181496",
"0.51751107",
"0.5174449",
"0.5161234",
"0.51581675",
"0.5157756",
"0.51533383",
"0.5146454",
"0.51440424",
"0.51399446",
"0.51389885",
"0.5115506",
"0.5110272",
"0.510439",
"0.5094286",
"0.5093537",
"0.5087124",
"0.5081602",
"0.50801176",
"0.5075202",
"0.50729465",
"0.50708055",
"0.5069064",
"0.5057697",
"0.5057325",
"0.50526255",
"0.50488585",
"0.50475234",
"0.50382674",
"0.50353944",
"0.5034772",
"0.50341904",
"0.5032434",
"0.50248337",
"0.5024829",
"0.50231695",
"0.5021878",
"0.50160366",
"0.50142974",
"0.50115377",
"0.5008315",
"0.500732",
"0.5003615",
"0.50014645",
"0.50009954",
"0.49967462",
"0.49956864",
"0.49915665",
"0.4987844",
"0.49857396",
"0.4982778",
"0.49791268",
"0.49772587",
"0.49748275",
"0.49746013",
"0.49724934",
"0.49724934",
"0.49692053",
"0.49674776",
"0.49629477",
"0.495765",
"0.49571067",
"0.49570647",
"0.4955639",
"0.4954408",
"0.4952866",
"0.49479082",
"0.49458194",
"0.49435353",
"0.49433672",
"0.49418655",
"0.49388683",
"0.49381793"
] |
0.0
|
-1
|
Mettre une majuscule // Bug //
|
def titleize(a)
word = a.split
phrase = ""
if word[1].empty?
return a.split[0]
else
word.each do |element|
phrase = phrase + element.capitalize + ""
return phrase
end
else
end
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def marquer!\n fail \"Doit etre defini dans la sous-classe\"\n end",
"def mambo_no_5; end",
"def mi_carrera\n\n\tend",
"def schumann; end",
"def suivre; end",
"def minor; end",
"def minor; end",
"def minor; end",
"def celebrity; end",
"def celebrity; end",
"def celebration; end",
"def malts; end",
"def somme\n fail \"Doit etre defini dans la sous-classe\"\n end",
"def verdi; end",
"def zuruecksetzen()\n end",
"def test_min_dispersa\n\t\tassert_equal(Fraccion.new(66,5), @h3.min, \"Resultado Incorrecto\" )\n\tend",
"def povuci_mrezu\n return @mreza\n end",
"def cardinal; end",
"def minor=(_arg0); end",
"def test_mcm\n assert_equal 6,Fraccion.new(1,1).mcm(@a.denom,@b.denom)\n end",
"def moon; end",
"def imc\n\t\tnum = (@peso/(@talla*@talla)).round(2)\n\t\tif num < 18.5\n\t\t\tnum #- Bajo peso\"\n\t\telsif num > 18.5 and num < 24.9\n\t\t\tnum #- Adecuado\"\n\t\telsif num > 25.0 and num < 29.9\n\t\t\tnum #- Sobrepeso\"\n\t\telsif num > 30.0 and num < 34.9\n\t\t\tnum #Obesidad grado 1\"\n\t\telsif num > 35.0 and num < 39.9\n\t\t\tnum #- Obesidad grado 2\"\n\t\telsif num > 40\n\t\t\tnum #- Obesidad grado 2\"\n\t\tend\t\t\t\n\tend",
"def bug\n end",
"def jeuTermine\n\t\tlancementAventure(@tailleGrille+1)\n\tend",
"def romeo_and_juliet; end",
"def test_minimizar\n assert_equal 2,Fraccion.new(1,1).minimizar(@c.num,@c.denom).num\n assert_equal 1,Fraccion.new(1,1).minimizar(@c.num,@c.denom).denom\n end",
"def test_max_dispersa\n\t\tassert_equal(Fraccion.new(77,12), @h3.max, \"Resultado Incorrecto\" )\n\tend",
"def primer\n end",
"def ismn; end",
"def major; end",
"def major; end",
"def major; end",
"def sichtbar_machen()\n end",
"def fml (m)\n\tend",
"def bellini; end",
"def m3; 3 end",
"def schubert; end",
"def reset_marque\n fail \"Doit etre defini dans la sous-classe\"\n end",
"def cuadrada\n @n==@m\n end",
"def changerEnSuivant #-1=>1=>0>-1\n\t\t@couleur=@couleur%3-1\n\tend",
"def unsichtbar_machen()\n end",
"def mascot; end",
"def chondromyxoma(buckshee, uncongenially_chiquitan)\n end",
"def llamar() #imprimira marca y color este metodo, ese metodo usa las variables de instancia\n puts @marca + \"-\"+ @color + \" esta llamando...\"\n end",
"def m3\n\t\t\t\t\tsuper * 2\t\t\t\t\t\t\t\t\t\t\t\t# first override to #m3\n\t\t\t\tend",
"def m3\n\t\t\t\t\tsuper * 2\t\t\t\t\t\t\t\t\t\t\t\t# first override to #m3\n\t\t\t\tend",
"def m3\n\t\t\t\t\tsuper * 2\t\t\t\t\t\t\t\t\t\t\t\t# first override to #m3\n\t\t\t\tend",
"def no_cm\n self.gsub(/[mm]/,\"\")\n end",
"def monica\n end",
"def sichtbar_machen()\n # TODO\n end",
"def changerEnVide\n\t\t@couleur=-1\n\tend",
"def major=(_arg0); end",
"def slogan\n # 'A maneira mais fácil de pré-qualificar ao Atlas.'\n ''\n end",
"def marcoReal\n\t\t@marcoReal\n\tend",
"def matricula\n end",
"def no_mm\n self.gsub(/[cm]/,\"\")\n end",
"def jeuTermine\n\t\t@grille.score.recupererTemps(self.timer)\n\t\tscoreFinal = @grille.score.calculerScoreFinal\n\t\t@@joueur.score = scoreFinal > 0 ? scoreFinal : 0\n\t\tself.lancementFinDeJeu\n\t\tself\n\tend",
"def changerEnRouge\n\t\t@couleur=1\n\tend",
"def mayor\n max = 0\n for i in 0...@f do\n for j in 0...@c do\n\tif((!@matriz[i].nil?) && (!@matriz[i][j].nil?))\n\t if (max < @matriz[i][j])\n\t max = @matriz[i][j]\n\t end\n\tend\n end\n end\n end",
"def motto; end",
"def musician; end",
"def musician; end",
"def seuil()\n\t\treturn 0\n\tend",
"def test_resta\n \t\tassert_equal(@ma, (@md-@mz)) #densa\n \t\tassert_equal(@a, (@c-@b)) #fraccionales\n\t\tassert_equal(@mp, (@mr-@mq)) #dispersa\n \tend",
"def test_min_densa\n\t\tassert_equal(3, @h6.min, \"Resultado Incorrecto\" )\n\tend",
"def genclase\n \"M\"\n end",
"def exercise_1113 (matrix)\n end",
"def jugada_max\r\n movimiento.max_by{|x| x.nivel} #Devuelve el objeto de enumeración que da el valor máximo\r\n end",
"def test_soma_100_200\n\t\tassert_equal(300, Calculadora.new(100,200).soma)\n\tend",
"def m\nend",
"def m\nend",
"def blg; end",
"def mitch_hedberg; end",
"def manufacture; end",
"def test_max_densa\n\t\tassert_equal(8, @h6.max, \"Resultado Incorrecto\" )\n\tend",
"def mille\n self > 9999 || (return self)\n case\n when self < 10000\n return self.to_s\n when self < 999999\n s = self.to_s.rjust(6,'0')\n return \"#{s[0..2].to_i} #{s[3..5]}\"\n when self < 999999999\n s = self.to_s.rjust(9,'0')\n return \"#{s[0..2].to_i} #{s[3..5]} #{s[6..8]}\"\n else\n return self.to_s\n end\n end",
"def warnings; end",
"def warnings; end",
"def warnings; end",
"def warnings; end",
"def warnings; end",
"def warnings; end",
"def warnings; end",
"def warnings; end",
"def terpene; end",
"def rebalance_ming_crisis!\n # First, make sure crisis lasts at least 10 years, no cheating\n patch_mod_file!(\"common/disasters/ming_crisis.txt\") do |node|\n crisis = node[\"ming_crisis\"]\n modify_node! crisis,\n [\"modifier\", \"global_unrest\", 5, 10],\n [\"modifier\", \"land_morale\", -0.15, -0.30]\n crisis[\"can_end\"][\"custom_trigger_tooltip\"] = PropertyList[\n \"tooltip\", \"disaster_active_for_10_years_tooltip\",\n \"had_country_flag\", PropertyList[\n \"flag\", \"had_ming_crisis\",\n \"days\", 3650,\n ]\n ]\n end\n\n # Set flag to make sure it last 10 years (flag only for event above)\n patch_mod_file!(\"events/disaster_ming_crisis.txt\") do |node|\n event = node.to_a.find{|prop| prop.key == \"country_event\" and prop.val[\"id\"] == \"ming_crisis.1\"}\n event.val[\"immediate\"].add! Property[\"set_country_flag\", \"had_ming_crisis\"]\n end\n\n # Then make low mandate actually painful\n # And losing mandate basically a death sentence\n # -100% manpower sounds like BS, but there are enormous\n # positive modifiers in 1.30 (especially estates) so it's de facto closer to -50%\n patch_mod_file!(\"common/static_modifiers/00_static_modifiers.txt\") do |node|\n modify_node!(node,\n [\"negative_mandate\", \"global_unrest\", 5, 15],\n [\"negative_mandate\", \"manpower_recovery_speed\", nil, -0.5],\n [\"negative_mandate\", \"mercenary_manpower\", -0.5, -2.0],\n [\"negative_mandate\", \"fire_damage_received\", 0.5, 1.0],\n [\"negative_mandate\", \"shock_damage_received\", 0.5, 1.0],\n [\"negative_mandate\", \"mercenary_discipline\", nil, -1.0],\n # [\"negative_mandate\", \"reduced_liberty_desire\", nil, -50],\n [\"negative_mandate\", \"diplomatic_reputation\", nil, -5],\n [\"negative_mandate\", \"liberty_desire_from_subject_development\", nil, 1.0],\n [\"lost_mandate_of_heaven\", \"global_unrest\", 10, 15],\n [\"lost_mandate_of_heaven\", \"manpower_recovery_speed\", nil, -0.5],\n [\"lost_mandate_of_heaven\", \"mercenary_manpower\", -0.5, -2.0],\n [\"lost_mandate_of_heaven\", \"fire_damage_received\", 0.5, 1.0],\n [\"lost_mandate_of_heaven\", \"shock_damage_received\", 0.5, 1.0],\n [\"lost_mandate_of_heaven\", \"mercenary_discipline\", nil, -1.0],\n # [\"lost_mandate_of_heaven\", \"reduced_liberty_desire\", -50, -50]\n [\"lost_mandate_of_heaven\", \"diplomatic_reputation\", nil, -5],\n [\"lost_mandate_of_heaven\", \"liberty_desire_from_subject_development\", nil, 1.0],\n )\n end\n\n # Make winning/losing mandate last 50 years\n # from 20 (lost) / 25 (gained)\n patch_mod_file!(\"common/on_actions/00_on_actions.txt\") do |node|\n gain_mods = node[\"on_mandate_of_heaven_gained\"].find_all(\"add_country_modifier\")\n loss_mods = node[\"on_mandate_of_heaven_lost\"].find_all(\"add_country_modifier\")\n\n raise unless loss_mods == [PropertyList[\"name\", \"lost_mandate_of_heaven\", \"duration\", 7300]]\n raise unless gain_mods == [PropertyList[\"name\", \"new_mandate_holder\", \"duration\", 9125]]\n\n gain_mods[0][\"duration\"] = 365*50\n loss_mods[0][\"duration\"] = 365*50\n end\n\n # Prevent kicking new Empire down too early\n # It makes flavor sense, but kicking AI before it can establish itself\n # just destroys any chance of new China emerging\n patch_mod_file!(\"common/disasters/empire_of_china_culture.txt\") do |node|\n node[\"empire_of_china_in_name\"][\"potential\"].add! Property::NOT[\"has_country_modifier\", \"new_mandate_holder\"]\n # Actually just disable it for AI for now\n end\n\n # Waging war for China helps more\n patch_mod_files!(\"common/triggered_modifiers/00_triggered_modifiers.txt\") do |node|\n modify_node! node,\n [\"waging_war_for_china\", \"imperial_mandate\", 0.05, 0.10]\n end\n\n # Buff mandate winners harder\n # +0.10 barely counters not having 2/3 capitals,\n # and new holder will often lack all 3\n patch_mod_file!(\"common/event_modifiers/00_event_modifiers.txt\") do |node|\n modify_node! node,\n [\"new_mandate_holder\", \"imperial_mandate\", 0.05, 0.10],\n [\"new_mandate_holder\", \"global_unrest\", nil, -5]\n end\n\n # Fix bug that makes lost mandate holder not have proper rebels\n # Code obviously tries to make them have nationalists, but code bug breaks it\n # as country which lost mandate will also lose celestial_empire reform\n patch_mod_files!(\"common/rebel_types/nationalist.txt\") do |node|\n spawn_chance = node[\"nationalist_rebels\"][\"spawn_chance\"]\n lost_mandate = spawn_chance.find_all(\"modifier\").find{|m| m[\"owner\"] and m[\"owner\"][\"has_reform\"] == \"celestial_empire\" and m[\"factor\"] == 100 }\n lost_mandate[\"owner\"] = PropertyList[\n \"NOT\", PropertyList[\"primary_culture\", \"ROOT\"],\n \"OR\", PropertyList[\n \"has_country_modifier\", \"the_mandate_of_heaven_lost\",\n \"has_country_modifier\", \"lost_mandate_of_heaven\",\n \"AND\", PropertyList[\n \"has_reform\", \"celestial_empire\",\n \"has_dlc\", \"Mandate of Heaven\",\n \"NOT\", PropertyList[\"imperial_mandate\", 50],\n ],\n ],\n ]\n end\n end",
"def konversiMenit(menit) \n # =================================\n # Area Kode Kode di sini\n \n \n \n # =================================\n end",
"def test_biggest_metro\n info = @sao_paulo\n assert_equal(info, biggest_metro(@south_america))\n end",
"def mon() end",
"def medical_use; end",
"def mozart; end",
"def hambriento?\n # Los nombres de los metodos pueden terminar en \"?\".\n # Generalmente, hacemos esto si el método debe\n # devolver verdadero o falso, como esto:\n @panzaLlena <= 2\n end",
"def hambriento?\n # Los nombres de los metodos pueden terminar en \"?\".\n # Generalmente, hacemos esto si el método debe\n # devolver verdadero o falso, como esto:\n @panzaLlena <= 2\n end",
"def rebalance_ming_crisis!\n warn \"Disabled code #{__method__}. Do not enable in release. #{__FILE__}:#{__LINE__}\"\n # First, make sure crisis lasts at least 10 years, no cheating\n patch_mod_file!(\"common/disasters/ming_crisis.txt\") do |node|\n crisis = node[\"ming_crisis\"]\n modify_node! crisis,\n [\"modifier\", \"global_unrest\", 15, 15], # used to be 5->10 before 1.32\n [\"modifier\", \"land_morale\", -0.15, -0.30]\n crisis[\"can_end\"][\"custom_trigger_tooltip\"] = PropertyList[\n \"tooltip\", \"disaster_active_for_10_years_tooltip\",\n \"had_country_flag\", PropertyList[\n \"flag\", \"had_ming_crisis\",\n \"days\", 3650,\n ]\n ]\n end\n\n # Set flag to make sure it last 10 years (flag only for event above)\n patch_mod_file!(\"events/disaster_ming_crisis.txt\") do |node|\n event = node.to_a.find{|prop| prop.key == \"country_event\" and prop.val[\"id\"] == \"ming_crisis.1\"}\n event.val[\"immediate\"].add! Property[\"set_country_flag\", \"had_ming_crisis\"]\n end\n\n # Then make low mandate actually painful\n # And losing mandate basically a death sentence\n # -100% manpower sounds like BS, but there are enormous\n # positive modifiers in 1.30 (especially estates) so it's de facto closer to -50%\n patch_mod_file!(\"common/static_modifiers/00_static_modifiers.txt\") do |node|\n modify_node!(node,\n [\"negative_mandate\", \"global_unrest\", 5, 10],\n [\"negative_mandate\", \"manpower_recovery_speed\", nil, -0.5],\n [\"negative_mandate\", \"mercenary_manpower\", -0.5, -2.0],\n # [\"negative_mandate\", \"fire_damage_received\", 0.5, 1.0],\n # [\"negative_mandate\", \"shock_damage_received\", 0.5, 1.0],\n [\"negative_mandate\", \"mercenary_discipline\", nil, -1.0],\n # [\"negative_mandate\", \"reduced_liberty_desire\", nil, -50],\n [\"negative_mandate\", \"diplomatic_reputation\", nil, -5],\n [\"negative_mandate\", \"liberty_desire_from_subject_development\", nil, 1.0],\n # [\"lost_mandate_of_heaven\", \"global_unrest\", 10, 10],\n [\"lost_mandate_of_heaven\", \"manpower_recovery_speed\", nil, -0.5],\n [\"lost_mandate_of_heaven\", \"mercenary_manpower\", -0.5, -2.0],\n # [\"lost_mandate_of_heaven\", \"fire_damage_received\", 0.5, 1.0],\n # [\"lost_mandate_of_heaven\", \"shock_damage_received\", 0.5, 1.0],\n [\"lost_mandate_of_heaven\", \"mercenary_discipline\", nil, -1.0],\n # [\"lost_mandate_of_heaven\", \"reduced_liberty_desire\", -50, -50]\n [\"lost_mandate_of_heaven\", \"diplomatic_reputation\", nil, -5],\n [\"lost_mandate_of_heaven\", \"liberty_desire_from_subject_development\", nil, 1.0],\n )\n end\n\n # Make winning/losing mandate last 50 years\n # from 20 (lost) / 25 (gained)\n patch_mod_file!(\"common/on_actions/00_on_actions.txt\") do |node|\n gain_mods = node[\"on_mandate_of_heaven_gained\"].find_all(\"add_country_modifier\")\n loss_mods = node[\"on_mandate_of_heaven_lost\"].find_all(\"add_country_modifier\")\n\n raise unless loss_mods == [PropertyList[\"name\", \"lost_mandate_of_heaven\", \"duration\", 7300]]\n raise unless gain_mods == [PropertyList[\"name\", \"new_mandate_holder\", \"duration\", 9125]]\n\n gain_mods[0][\"duration\"] = 365*50\n loss_mods[0][\"duration\"] = 365*50\n end\n\n # Prevent kicking new Empire down too early\n # It makes flavor sense, but kicking AI before it can establish itself\n # just destroys any chance of new China emerging\n patch_mod_file!(\"common/disasters/empire_of_china_culture.txt\") do |node|\n node[\"empire_of_china_in_name\"][\"potential\"].add! Property::NOT[\"has_country_modifier\", \"new_mandate_holder\"]\n # Actually just disable it for AI for now\n end\n\n # Waging war for China helps more\n patch_mod_files!(\"common/triggered_modifiers/00_triggered_modifiers.txt\") do |node|\n modify_node! node,\n [\"waging_war_for_china\", \"imperial_mandate\", 0.05, 0.10]\n end\n\n # Buff mandate winners harder\n # +0.10 barely counters not having 2/3 capitals,\n # and new holder will often lack all 3\n patch_mod_file!(\"common/event_modifiers/00_event_modifiers.txt\") do |node|\n modify_node! node,\n [\"new_mandate_holder\", \"imperial_mandate\", 0.05, 0.10],\n [\"new_mandate_holder\", \"global_unrest\", nil, -5]\n end\n\n # Fix bug that makes lost mandate holder not have proper rebels\n # Code obviously tries to make them have nationalists, but code bug breaks it\n # as country which lost mandate will also lose celestial_empire reform\n patch_mod_files!(\"common/rebel_types/nationalist.txt\") do |node|\n spawn_chance = node[\"nationalist_rebels\"][\"spawn_chance\"]\n lost_mandate = spawn_chance.find_all(\"modifier\").find{|m| m[\"owner\"] and m[\"owner\"][\"has_reform\"] == \"celestial_empire\" and m[\"factor\"] == 100 }\n lost_mandate[\"owner\"] = PropertyList[\n \"NOT\", PropertyList[\"primary_culture\", \"ROOT\"],\n \"OR\", PropertyList[\n \"has_country_modifier\", \"the_mandate_of_heaven_lost\",\n \"has_country_modifier\", \"lost_mandate_of_heaven\",\n \"AND\", PropertyList[\n \"has_reform\", \"celestial_empire\",\n \"has_dlc\", \"Mandate of Heaven\",\n \"NOT\", PropertyList[\"imperial_mandate\", 50],\n ],\n ],\n ]\n end\n end",
"def calcula_imc\n \n if @peso/@altura*@altura < 18\n puts \"vc esta magro\"\n elsif @peso/@altura*@altura <= 25\n puts \"vc esta no peso ideal\"\n elsif @peso/@altura*@altura > 25\n puts \"vc esta acima do peso\"\n end\n \n end",
"def imei; end",
"def monster; end",
"def monster; end",
"def recolectar_una\n\t\test = \"\"\n\t\tif @estado == ESTADO::MUERTE\n\t\t\test = \"El árbol está muerto\"\n\t\telse\n\t\t\tif @contador == 0\n\t\t\t\test = \"No hay más naranjas\"\n\t\t\telse\n\t\t\t\test = \"La naranja estaba deliciosa\"\n\t\t\t\t@contador -= 1\n\t\t\tend\n\t\tend\n\t\test\n\tend",
"def blam!\n berzerk? ? w00t! : super\n end",
"def jack_handey; end"
] |
[
"0.6316823",
"0.61429083",
"0.60477185",
"0.5976003",
"0.5945212",
"0.59011716",
"0.59011716",
"0.59011716",
"0.5892193",
"0.5892193",
"0.58853245",
"0.5699722",
"0.5673872",
"0.5654073",
"0.56159574",
"0.5498585",
"0.5498479",
"0.54802686",
"0.54697686",
"0.54601276",
"0.54397404",
"0.5431828",
"0.54214555",
"0.5420095",
"0.540989",
"0.54021066",
"0.53698283",
"0.5368631",
"0.5365544",
"0.53571695",
"0.53571695",
"0.53571695",
"0.5353654",
"0.534272",
"0.5325098",
"0.5310217",
"0.5308462",
"0.53069246",
"0.5304058",
"0.5298648",
"0.5292684",
"0.52826244",
"0.52647686",
"0.52588886",
"0.52492374",
"0.52492374",
"0.52492374",
"0.52463067",
"0.52325225",
"0.52299917",
"0.5229765",
"0.5206534",
"0.52026534",
"0.5190262",
"0.51732445",
"0.51577175",
"0.5154569",
"0.5152871",
"0.51330847",
"0.51237494",
"0.51132196",
"0.51132196",
"0.51062506",
"0.5092782",
"0.5085188",
"0.5084342",
"0.50711715",
"0.5054705",
"0.50535077",
"0.5044258",
"0.5044258",
"0.50314426",
"0.5023816",
"0.5018129",
"0.50159556",
"0.5008053",
"0.50022227",
"0.50022227",
"0.50022227",
"0.50022227",
"0.50022227",
"0.50022227",
"0.50022227",
"0.50022227",
"0.50006366",
"0.49973166",
"0.4997042",
"0.4993276",
"0.49932253",
"0.49925238",
"0.4991972",
"0.4988858",
"0.4988858",
"0.49845216",
"0.49789116",
"0.49727693",
"0.49721858",
"0.49721858",
"0.49672782",
"0.49671915",
"0.49669608"
] |
0.0
|
-1
|
Copied over code and then added w/ extra current user only condition.
|
def collection
if params[:deleted] == 'on'
base_variant_scope ||= super.with_deleted
else
base_variant_scope ||= super
end
search = Spree::Config.variant_search_class.new(params[:variant_search_term], scope: base_variant_scope)
@collection = search.results.includes(variant_includes).page(params[:page]).per(Spree::Config[:admin_variants_per_page])
@collection = @collection.where(user_id: spree_current_user.try(:id) ) if spree_current_user && !spree_current_user.admin?
@collection
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def applies_to?(user); false end",
"def current_user\n # super: don't change anything, i just want the exact same behavior \n # as in the method that we are overriding\n\n # what this line means is that if the user is logged in, super is true,\n # then call super treat everything normal, and ignore right hand side\n # if super == false, call right hand side\n\n # super comes from devise class\n # meanwhile the r.h.s comes from open struct class\n super || guest_user\n end",
"def current_user_required\n\t\t# Have to add \".filter(self)\" when not in before_filter line.\n\t\tCASClient::Frameworks::Rails::Filter.filter(self)\n\tend",
"def check_if_current_user\r\n User.current_user && User.current_user != self\r\n end",
"def check_if_user_access\n if current_user && current_user.admin?\n @all_orders = Order.all\n @orders = @all_orders.paginate(page: params[:page])\n elsif current_user && current_user.role == \"Partner\"\n redirect_to root_path\n flash[:danger] = I18n.t 'flash.your_company_not_verified'\n elsif current_user && current_user.role == \"Client\"\n @all_orders = @user_company.orders\n @orders = @all_orders.paginate(page: params[:page])\n else\n end\n end",
"def current_user?\n \n end",
"def current_user\n@current_user ||= (login_from_session || login_from_basic_auth || login_from_cookie) unless @current_user == false\nend",
"def user_check(resource)\n return true unless self[:name] == \"user\"\n return true unless self[:unless_system_user]\n\n resource[:audit] = :uid\n\n return false if system_users.include?(resource[:name])\n\n current_values = resource.retrieve_resource\n current_values[resource.property(:uid)] > self[:unless_system_user]\n end",
"def current_user\n @current_user ||= User.find(session[:user_id]) if session[:user_id]\n #if current_user exist or equil to this user then keep going\n end",
"def before_filter\n if current_user\n true\n end\n end",
"def current_user?(call_node)\n call_node.message == :current_user || (call_node.subject.message == :current_user && call_node.message == :id)\n end",
"def current_user\n current_admin || current_consumer || current_employee\n end",
"def current_user? user\n if user == nil\n false\n elsif self.username == user.username\n true\n elsif\n false\n end\n end",
"def check_user_background\n self.status = Status::PENDING_ADMIN_ACTION if offender && offender.new_user?\n end",
"def current_user?(user)\n \t\tuser == current_user\n \tend",
"def current_user?(user)\n \tuser == current_user\n \tend",
"def current_user?(user)\n \tuser == current_user\n \tend",
"def current_user?(user)\n\t\tuser == self.current_user\n\t\t\n\tend",
"def effective_user\r\n self.current_proxy || self.current_user\r\n end",
"def effective_user\n current_user || demo_user\n end",
"def current_user? user\n \tuser == self.current_user\n end",
"def user_permission(instance,current_user)\n\t\tcase instance\n\t\twhen User #IF THAT THING IS A USER, ONLY A USER OR AN ADMIN CAN CHANGE\n\t\t\ttrue if self.admin || instance == current_user\n\t\telse\n\t\t\t#IF THAT THING IS SOMETHING BESIDES A USER, ONLY AN ADMIN OR THE USER THAT THE THING BELONGS TO CAN CHANGE IT\n\t\t\ttrue if self.admin || instance.user == current_user\n\t\tend\n\tend",
"def current_user\n @current_user||= User.find(session[:user_id]) if session[:user_id]\n\n #here User.find(session[:user_id]) this will fetch the user based on the active session's user_id \n #as the above line will hit the database numerous time. We don't want that so to restrict it we created a local variable to store the user once find\n #|| this pipeline symbol is for checking \nend",
"def current_user?(user)\n user && user == current_user\n end",
"def check_current_user_owns\n head :not_found unless @user == current_user\n end",
"def current_user?(user)\n \t\tuser == current_user\n \tend",
"def current_user?(user)\r\n user == current_user\r\n end",
"def current_user?(user)\n user == current_person.user\n end",
"def owner_required\n ## before filter for owner of channel. \n if logged_in? && current_user.login == THUMBWEBS_AUTHORIZED_USER\n return true\n else\n flash[:error] = \"Unauthorized Access-Must be logged-in as owner.\"\n redirect_to thumbwebs_root_path\n end\nend",
"def check_if_user\n render text: \"Fuck you user\", status: 403 unless current_user.id == @book.user_id || current_user.rank_id == 2\n end",
"def check_current_user \n if params[:user_id]\n redirect_to '/' and return unless (current_user == @user) || current_user.is_admin?\n elsif params[:mband_id]\n redirect_to '/' and return unless @mband.members.include?(current_user) || current_user.is_admin?\n end\n end",
"def current_user?(user)\n (user == self.current_user) || admin?\n end",
"def is_user?\ncurrent_user.username == @userl.username\n end",
"def current_user?(user) \n user == current_user\n end",
"def current_user?(user)\n\tuser == current_user\n end",
"def is_user_specific?\n false\n end",
"def is_this_user\n\t\t@user.id == current_user.id\n\tend",
"def current_user\n\n #if $apiflag==1\n #@current_user ||= Fbuser.find(session[:user_id]) if session[:user_id]\n #else\n @current_user ||= User.find(session[:user_id]) if session[:user_id]\n #end\n end",
"def current_user?(user)\n user == current_user\n end",
"def is_current\n \treturn unless self.controller_name == \"users\"\n \tif current_user == nil\n \t\tflash[:notice] = \"You cannot access this page without being logged in to the proper account.\"\n \t\tredirect_to root_path\n \telsif !current_user.is_admin\n \t\tif params[:id] != current_user.id && params[:user] != current_user\n \t\t\tflash[:notice] = \"You cannot access this page without being logged in to the proper account.\"\n \t\t\tredirect_to root_path\n \t\telse\n \t\t\treturn\n \t\tend\n \telse\n \t\treturn\n \tend\n end",
"def current_user?(user)\n\t\tuser == current_user\n\tend",
"def current_user?(user)\n\t\tuser == current_user\n\tend",
"def current_user?(user)\n\t\tuser == current_user\n\tend",
"def current_user?(user)\n\t\tuser == current_user\n\tend",
"def current_user?(user)\n\t\tuser == current_user\n\tend",
"def current_user?(user)\n\t\tuser == current_user\n\tend",
"def current_user?(user)\n\t\tuser == current_user\n\tend",
"def current_user?(user)\n\t\tuser == current_user\n\tend",
"def current_user?(user)\n\t\tuser == current_user\n\tend",
"def current_user?(user)\n\t\tuser == current_user\n\tend",
"def current_user?(user)\n\t\tuser == current_user\n\tend",
"def user_is_current\n fetch_user\n\n unless current_user?(@user)\n flash[:error] = t(:not_authorized)\n redirect_to root_path\n end\n end",
"def current_user\n @current_user ||= session[:cas_user] #User.where(liu_id: session[:cas_user]) \n end",
"def check_correct_user(user)\n \tcurrent_user?(user)\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end",
"def current_user?(user)\n user == current_user\n end"
] |
[
"0.6458769",
"0.63881606",
"0.6372016",
"0.6280778",
"0.62759703",
"0.6185129",
"0.6178637",
"0.612089",
"0.61104846",
"0.6082802",
"0.60506845",
"0.6046031",
"0.60411066",
"0.6038733",
"0.6010903",
"0.6007372",
"0.6007372",
"0.5996541",
"0.5981947",
"0.5977512",
"0.59721637",
"0.5960242",
"0.59576255",
"0.59524614",
"0.5943943",
"0.5935516",
"0.5932836",
"0.59324104",
"0.5927181",
"0.5920405",
"0.5909946",
"0.59085417",
"0.5898686",
"0.5884263",
"0.5879769",
"0.58791345",
"0.5875043",
"0.58716464",
"0.58706224",
"0.58686227",
"0.5863001",
"0.5861164",
"0.5861164",
"0.5861164",
"0.5861164",
"0.5861164",
"0.5861164",
"0.5861164",
"0.5861164",
"0.5861164",
"0.5861164",
"0.58583516",
"0.58561754",
"0.5853485",
"0.58490586",
"0.58490586",
"0.58490586",
"0.58490586",
"0.58490586",
"0.58490586",
"0.58490586",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396",
"0.5848396"
] |
0.0
|
-1
|
Check a relation tuple To learn how relation tuples and the check works, head over to [the documentation](../concepts/relationtuples.mdx).
|
def get_check(namespace, object, relation, opts = {})
data, _status_code, _headers = get_check_with_http_info(namespace, object, relation, opts)
data
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def valid_relation_literal!(heading, tuples, raise_on_error = true)\n if Array === tuples\n invalid = tuples.find{|t| !valid_tuple_literal?(heading, t)}\n if invalid\n raise_on_error ? raise(::Veritas::RelationMismatchError, \"Invalid tuple #{invalid.inspect} for heading #{heading}\", caller) : false\n else\n tuples\n end\n else\n raise_on_error ? raise(::Veritas::RelationMismatchError, \"Invalid relation literal #{tuples.inspect}\", caller) : false \n end\n end",
"def check_for_tuple(tuple)\n ltuple = @left_pattern.map {|i| tuple[i]}\n if @left.check_for_tuple(ltuple)\n rtuple = @right_pattern.map {|i| tuple[i]}\n if @right.check_for_tuple(rtuple)\n return true\n end\n end\n return false\n end",
"def contains?(tuple)\n case tuple\n when Hash\n eql_hash = {}\n relvar_def.primary_key.attributes.each{|a| eql_hash[a.name] = tuple[a.name]}\n inside_tuple = underlying_table.filter(eql_hash).first\n return inside_tuple === tuple\n else\n raise ArgumentError, \"Unable to check inclusion of #{tuple} inside a relation variable\"\n end\n end",
"def valid_tuple_literal!(heading, tuple, raise_on_error = true)\n if Hash===tuple and tuple.size == heading.to_ary.size\n ok = heading.all?{|a| tuple.key?(a.name) and a.valid_value?(tuple[a.name])}\n return tuple if ok\n end\n raise_on_error ? raise(ArgumentError, \"Invalid tuple #{tuple} for #{heading}\", caller) : false\n end",
"def _tuple2tuple(tuple)\n end",
"def tuple_plan()\n raise NotTupleValuedException.new(self) unless tuple_valued?\n @definition.type.tuple.plan\n end",
"def create_relation_tuple(opts = {})\n data, _status_code, _headers = create_relation_tuple_with_http_info(opts)\n data\n end",
"def check_relation(relation)\n (relation - reflection.map{|r| r[:rel] }).empty?\n end",
"def relation\n # -> uncomment the next line to manually enable rule tracing\n # trace_in( __method__, 28 )\n\n\n value = nil\n\n\n type = nil\n a = nil\n b = nil\n\n\n begin\n # at line 178:5: a= add (type= ( '==' | '!=' | '<' | '<=' | '>=' | '>' ) b= relation )?\n @state.following.push( TOKENS_FOLLOWING_add_IN_relation_1322 )\n a = add\n @state.following.pop\n\n # syntactic predicate action gate test\n if @state.backtracking == 0\n # --> action\n value = a \n # <-- action\n end\n\n # at line 179:3: (type= ( '==' | '!=' | '<' | '<=' | '>=' | '>' ) b= relation )?\n alt_44 = 2\n look_44_0 = @input.peek( 1 )\n\n if ( look_44_0 == T__31 || look_44_0.between?( T__47, T__48 ) || look_44_0.between?( T__50, T__52 ) )\n alt_44 = 1\n end\n case alt_44\n when 1\n # at line 179:6: type= ( '==' | '!=' | '<' | '<=' | '>=' | '>' ) b= relation\n type = @input.look\n\n if @input.peek(1) == T__31 || @input.peek( 1 ).between?( T__47, T__48 ) || @input.peek( 1 ).between?( T__50, T__52 )\n @input.consume\n @state.error_recovery = false\n\n else\n @state.backtracking > 0 and raise( ANTLR3::Error::BacktrackingFailed )\n\n\n mse = MismatchedSet( nil )\n raise mse\n\n end\n\n\n @state.following.push( TOKENS_FOLLOWING_relation_IN_relation_1359 )\n b = relation\n @state.following.pop\n\n # syntactic predicate action gate test\n if @state.backtracking == 0\n # --> action\n value = DoubleOperandExpressionEval.new(type.text, a, b) \n # <-- action\n end\n\n\n end\n\n rescue ANTLR3::Error::RecognitionError => re\n report_error(re)\n recover(re)\n\n ensure\n # -> uncomment the next line to manually enable rule tracing\n # trace_out( __method__, 28 )\n\n\n end\n\n return value\n end",
"def eval_tuple(node, env, stack_trace)\n\n\tevaluated_nodes = []\n\tnode.values.each{ |sub_node|\n\t\tsub_node_value, _ = eval_node_under(sub_node, env, stack_trace) \n\t\tevaluated_nodes.append(sub_node_value)\n\t}\n\tif evaluated_nodes.length == 1\n\t\treturn evaluated_nodes[0], env\n\telse\n\t\treturn Tuple.new(node.line, node.col, evaluated_nodes)\n\tend\nend",
"def triangle?((x1, y1), (x2, y2), (x3, y3))\n return false if (x1, y1) == (x2, y2) || (x1, y1) == (x3, y3) || (x2, y2) == (x3, y3)\n return false \n true\nend",
"def patch_relation_tuples(opts = {})\n patch_relation_tuples_with_http_info(opts)\n nil\n end",
"def is_meta_tuple? tuple\n if meta_subspace\n meta_subspace === tuple\n else\n # meta_subspace hasn't arrived yet, so use approximation\n tuple.kind_of? Hash and tuple.key? client.tupelo_meta_key and\n tuple[client.tupelo_meta_key] == \"subspace\"\n end\n end",
"def t\n BERT::Tuple\nend",
"def activity_relations?(*args)\n activity_relations(*args).any?\n end",
"def tuples; end",
"def tuples; end",
"def assert_working_relations( model, relations )\n failed = []\n relations.each do | name |\n begin\n model.send( name )\n rescue Exception => e\n failed << \"#{name} - #{e}\"\n end\n end\n if failed.empty?\n assert true\n else\n assert false, failed.join('; ')\n end\n end",
"def check_relation(x)\n h, qx = compute_q(x)\n return if qx == 0\n es, l = exponent_bitvector(qx)\n\n # discard this x if the residue L is too big\n return if l > @pmax ** @t\n\n if l == 1\n # complete relation found:\n # Q(x) = p0^e0 * p1^e1 * ... * pk^ek (pi in the factor base)\n qx_vec = Hash.new(0)\n PrimeFactorization.prime_factorization(qx) {|p, e| qx_vec[p] += e }\n collect_relation(es, h, qx_vec)\n\n elsif @incomplete_relations[l]\n # large prime procedure:\n # make a complete relation by multiplying two incomplete relations\n es2, h2, qx2 = @incomplete_relations[l]\n\n # XXX: use FactoredInteger\n qx_vec = Hash.new(0)\n PrimeFactorization.prime_factorization(qx) {|p, e| qx_vec[p] += e }\n PrimeFactorization.prime_factorization(qx2) {|p, e| qx_vec[p] += e }\n\n collect_relation(es ^ es2, h * h2 % @kn, qx_vec)\n\n else\n @incomplete_relations[l] = [es, h, qx]\n end\n end",
"def verify_uniquness!(tuple)\n stream = tuple[:stream]\n attrs = %i[position event_id]\n attrs.delete(:position) if Stream.new(stream).global?\n\n attrs.each do |key|\n next if key == :position && tuple[key].nil?\n next if restrict(:stream => stream, key => tuple.fetch(key)).none?\n\n raise TupleUniquenessError.public_send(:\"for_stream_and_#{key}\", stream, tuple.fetch(key))\n end\n end",
"def create_relation_tuple_with_http_info(opts = {})\n if @api_client.config.debugging\n @api_client.config.logger.debug 'Calling API: WriteApi.create_relation_tuple ...'\n end\n # resource path\n local_var_path = '/relation-tuples'\n\n # query parameters\n query_params = opts[:query_params] || {}\n\n # header parameters\n header_params = opts[:header_params] || {}\n # HTTP header 'Accept' (if needed)\n header_params['Accept'] = @api_client.select_header_accept(['application/json'])\n # HTTP header 'Content-Type'\n header_params['Content-Type'] = @api_client.select_header_content_type(['application/json'])\n\n # form parameters\n form_params = opts[:form_params] || {}\n\n # http body (model)\n post_body = opts[:debug_body] || @api_client.object_to_http_body(opts[:'payload'])\n\n # return_type\n return_type = opts[:debug_return_type] || 'RelationQuery'\n\n # auth_names\n auth_names = opts[:debug_auth_names] || []\n\n new_options = opts.merge(\n :operation => :\"WriteApi.create_relation_tuple\",\n :header_params => header_params,\n :query_params => query_params,\n :form_params => form_params,\n :body => post_body,\n :auth_names => auth_names,\n :return_type => return_type\n )\n\n data, status_code, headers = @api_client.call_api(:PUT, local_var_path, new_options)\n if @api_client.config.debugging\n @api_client.config.logger.debug \"API called: WriteApi#create_relation_tuple\\nData: #{data.inspect}\\nStatus code: #{status_code}\\nHeaders: #{headers}\"\n end\n return data, status_code, headers\n end",
"def delete_relation_tuple_with_http_info(namespace, object, relation, opts = {})\n if @api_client.config.debugging\n @api_client.config.logger.debug 'Calling API: WriteApi.delete_relation_tuple ...'\n end\n # verify the required parameter 'namespace' is set\n if @api_client.config.client_side_validation && namespace.nil?\n fail ArgumentError, \"Missing the required parameter 'namespace' when calling WriteApi.delete_relation_tuple\"\n end\n # verify the required parameter 'object' is set\n if @api_client.config.client_side_validation && object.nil?\n fail ArgumentError, \"Missing the required parameter 'object' when calling WriteApi.delete_relation_tuple\"\n end\n # verify the required parameter 'relation' is set\n if @api_client.config.client_side_validation && relation.nil?\n fail ArgumentError, \"Missing the required parameter 'relation' when calling WriteApi.delete_relation_tuple\"\n end\n # resource path\n local_var_path = '/relation-tuples'\n\n # query parameters\n query_params = opts[:query_params] || {}\n query_params[:'namespace'] = namespace\n query_params[:'object'] = object\n query_params[:'relation'] = relation\n query_params[:'subject_id'] = opts[:'subject_id'] if !opts[:'subject_id'].nil?\n query_params[:'subject_set.namespace'] = opts[:'subject_set_namespace'] if !opts[:'subject_set_namespace'].nil?\n query_params[:'subject_set.object'] = opts[:'subject_set_object'] if !opts[:'subject_set_object'].nil?\n query_params[:'subject_set.relation'] = opts[:'subject_set_relation'] if !opts[:'subject_set_relation'].nil?\n\n # header parameters\n header_params = opts[:header_params] || {}\n # HTTP header 'Accept' (if needed)\n header_params['Accept'] = @api_client.select_header_accept(['application/json'])\n\n # form parameters\n form_params = opts[:form_params] || {}\n\n # http body (model)\n post_body = opts[:debug_body]\n\n # return_type\n return_type = opts[:debug_return_type]\n\n # auth_names\n auth_names = opts[:debug_auth_names] || []\n\n new_options = opts.merge(\n :operation => :\"WriteApi.delete_relation_tuple\",\n :header_params => header_params,\n :query_params => query_params,\n :form_params => form_params,\n :body => post_body,\n :auth_names => auth_names,\n :return_type => return_type\n )\n\n data, status_code, headers = @api_client.call_api(:DELETE, local_var_path, new_options)\n if @api_client.config.debugging\n @api_client.config.logger.debug \"API called: WriteApi#delete_relation_tuple\\nData: #{data.inspect}\\nStatus code: #{status_code}\\nHeaders: #{headers}\"\n end\n return data, status_code, headers\n end",
"def tuple_extract(*args, &bl)\n alf_connection.tuple_extract(*args, &bl)\n end",
"def relation(relation)\n true\n end",
"def tuple_valued?()\n @definition.type.named_type?\n end",
"def delete_relation_tuple(namespace, object, relation, opts = {})\n delete_relation_tuple_with_http_info(namespace, object, relation, opts)\n nil\n end",
"def check_relation_with friend_id\n response = {\n message: \"Successfully fecth user relation\",\n following: self.active_relationships.where(follower_id: friend_id).present?,\n follower: self.passive_relationships.where(follower_id: friend_id).present?\n }\n end",
"def is_valid_relation?(parent, child, require_both = self.require_both)\n # depending upon the\n require_both ? is_valid_down_relation?(parent, child) && is_valid_up_relation?(parent, child)\n : is_valid_down_relation?(parent, child) || is_valid_up_relation?(parent, child)\n end",
"def parse_tuple cur_tok\n\t\tcur_ast = Tuple.new(cur_tok.line, cur_tok.col, splitter( nil, \")\", \",\")) # start=nil because \"(\" token has already been read.\n\n\t\t# Use look ahead to see if the tuple needs to be a child node of an operation.\n\t\tif is_operation \n\t\t\treturn parse_operation(cur_ast, true)\n\t\telse\n\t\t\treturn cur_ast \n\t\tend\n\tend",
"def deduce\n each_vertex do |v1|\n each_vertex do |v2|\n each_vertex do |v3|\n \n if self[v1,v2] == :provable_true and self[v2,v3] == :provable_true\n add_assertion(v1, v3, :provable_true)\n end\n \n if self[v2,v1] == :provable_false and self[v2,v3] == :provable_true\n add_assertion(v3, v1, :provable_false)\n end\n \n if self[v1,v2] == :provable_true and self[v3,v2] == :provable_false\n add_assertion(v3, v1, :provable_false)\n end\n \n break if @contradiction\n end\n end\n end\n end",
"def cast_tuple\n <<-CODE\n t1 = stack_pop();\n if(REFERENCE_P(t1) && object_kind_of_p(state, t1, global->array)) {\n j = N2I(array_get_total(t1));\n t2 = tuple_new(state, j);\n\n for(k = 0; k < j; k++) {\n tuple_put(state, t2, k, array_get(state, t1, k));\n }\n t1 = t2;\n } else if(!REFERENCE_P(t1) || !object_kind_of_p(state, t1, global->tuple)) {\n t2 = tuple_new(state, 1);\n tuple_put(state, t2, 0, t1);\n t1 = t2;\n }\n stack_push(t1);\n CODE\n end",
"def assert_relations_are_equal(a, b)\n assert_not_nil a, \"first relation is not allowed to be nil\"\n assert_not_nil b, \"second relation #{a.id} is not allowed to be nil\"\n assert_equal a.id, b.id, \"relation IDs\"\n assert_equal a.changeset_id, b.changeset_id, \"changeset ID on relation #{a.id}\"\n assert_equal a.visible, b.visible, \"visible on relation #{a.id}, #{a.visible.inspect} != #{b.visible.inspect}\"\n assert_equal a.version, b.version, \"version on relation #{a.id}\"\n assert_equal a.tags, b.tags, \"tags on relation #{a.id}\"\n assert_equal a.members, b.members, \"member references on relation #{a.id}\"\n end",
"def tuple(*elements)\n Tuple.new(*elements)\n end",
"def cmp_tuples(left, right)\n reduce(0) do |cmp, direction|\n break cmp if cmp.nonzero?\n direction.call(left, right)\n end\n end",
"def get_relation_tuples(namespace, opts = {})\n data, _status_code, _headers = get_relation_tuples_with_http_info(namespace, opts)\n data\n end",
"def func_le(args)\n p1 = _eval(car(args))\n p2 = _eval(car(cdr(args)))\n\n if p1.type != LObject::OBJ_INTEGER and p2.type != LObject::OBJ_INTEGER\n if @lint\n Error.warn(\"warning: comparison with a non integer operand\")\n end\n return @o_man.nil\n end\n\n if p1.value.i <= p2.value.i\n return @o_man.t\n end\n return @o_man.nil\n end",
"def tuple(*args)\n \"( #{args.map { |b| if b.nil? then 'NULL' else b end }.join(', ')} )\"\nend",
"def relation\n # -> uncomment the next line to manually enable rule tracing\n # trace_in( __method__, 55 )\n return_value = RelationReturnValue.new\n\n # $rule.start = the first token seen before matching\n return_value.start = @input.look\n\n root_0 = nil\n shift249 = nil\n relation_op250 = nil\n shift251 = nil\n\n\n begin\n root_0 = @adaptor.create_flat_list\n\n\n # at line 609:5: shift ( relation_op shift )*\n @state.following.push( TOKENS_FOLLOWING_shift_IN_relation_4033 )\n shift249 = shift\n @state.following.pop\n if @state.backtracking == 0\n @adaptor.add_child( root_0, shift249.tree )\n end\n # at line 609:11: ( relation_op shift )*\n while true # decision 59\n alt_59 = 2\n look_59_0 = @input.peek( 1 )\n\n if ( look_59_0 == GEQ || look_59_0 == GREATER || look_59_0 == INSTANCEOF || look_59_0 == LEQ || look_59_0 == LESS )\n alt_59 = 1\n elsif ( look_59_0 == IN ) and ( ( !( @InFor_stack.last && @InFor_stack.last.active ) ) )\n alt_59 = 1\n\n end\n case alt_59\n when 1\n # at line 609:14: relation_op shift\n @state.following.push( TOKENS_FOLLOWING_relation_op_IN_relation_4038 )\n relation_op250 = relation_op\n @state.following.pop\n if @state.backtracking == 0\n root_0 = @adaptor.become_root( relation_op250.tree, root_0 )\n end\n @state.following.push( TOKENS_FOLLOWING_shift_IN_relation_4042 )\n shift251 = shift\n @state.following.pop\n if @state.backtracking == 0\n @adaptor.add_child( root_0, shift251.tree )\n end\n\n else\n break # out of loop for decision 59\n end\n end # loop for decision 59\n # - - - - - - - rule clean up - - - - - - - -\n return_value.stop = @input.look( -1 )\n\n if @state.backtracking == 0\n\n return_value.tree = @adaptor.rule_post_processing( root_0 )\n @adaptor.set_token_boundaries( return_value.tree, return_value.start, return_value.stop )\n\n end\n rescue ANTLR3::Error::RecognitionError => re\n report_error(re)\n recover(re)\n return_value.tree = @adaptor.create_error_node( @input, return_value.start, @input.look(-1), re )\n\n ensure\n # -> uncomment the next line to manually enable rule tracing\n # trace_out( __method__, 55 )\n\n end\n \n return return_value\n end",
"def get_relation_tuples_with_http_info(namespace, opts = {})\n if @api_client.config.debugging\n @api_client.config.logger.debug 'Calling API: ReadApi.get_relation_tuples ...'\n end\n # verify the required parameter 'namespace' is set\n if @api_client.config.client_side_validation && namespace.nil?\n fail ArgumentError, \"Missing the required parameter 'namespace' when calling ReadApi.get_relation_tuples\"\n end\n # resource path\n local_var_path = '/relation-tuples'\n\n # query parameters\n query_params = opts[:query_params] || {}\n query_params[:'namespace'] = namespace\n query_params[:'page_token'] = opts[:'page_token'] if !opts[:'page_token'].nil?\n query_params[:'page_size'] = opts[:'page_size'] if !opts[:'page_size'].nil?\n query_params[:'object'] = opts[:'object'] if !opts[:'object'].nil?\n query_params[:'relation'] = opts[:'relation'] if !opts[:'relation'].nil?\n query_params[:'subject_id'] = opts[:'subject_id'] if !opts[:'subject_id'].nil?\n query_params[:'subject_set.namespace'] = opts[:'subject_set_namespace'] if !opts[:'subject_set_namespace'].nil?\n query_params[:'subject_set.object'] = opts[:'subject_set_object'] if !opts[:'subject_set_object'].nil?\n query_params[:'subject_set.relation'] = opts[:'subject_set_relation'] if !opts[:'subject_set_relation'].nil?\n\n # header parameters\n header_params = opts[:header_params] || {}\n # HTTP header 'Accept' (if needed)\n header_params['Accept'] = @api_client.select_header_accept(['application/json'])\n\n # form parameters\n form_params = opts[:form_params] || {}\n\n # http body (model)\n post_body = opts[:debug_body]\n\n # return_type\n return_type = opts[:debug_return_type] || 'GetRelationTuplesResponse'\n\n # auth_names\n auth_names = opts[:debug_auth_names] || []\n\n new_options = opts.merge(\n :operation => :\"ReadApi.get_relation_tuples\",\n :header_params => header_params,\n :query_params => query_params,\n :form_params => form_params,\n :body => post_body,\n :auth_names => auth_names,\n :return_type => return_type\n )\n\n data, status_code, headers = @api_client.call_api(:GET, local_var_path, new_options)\n if @api_client.config.debugging\n @api_client.config.logger.debug \"API called: ReadApi#get_relation_tuples\\nData: #{data.inspect}\\nStatus code: #{status_code}\\nHeaders: #{headers}\"\n end\n return data, status_code, headers\n end",
"def valid_triangle? triple\n a, b, c = triple\n return false if a + b <= c\n return false if a + c <= b\n return false if b + c <= a\n\n true\nend",
"def tuples\n relation = self.relation\n return relation if materialized?\n\n adapter.read(relation)\n end",
"def create_tuple!(attributes)\n # split the attrs (needs DRYing)\n rel_attrs = {}\n attributes.each do |key, value|\n if key.to_s[0..3] == 'rel_'\n rel_attrs[key] = attributes.delete(key)\n end\n end\n\n # find or create the one node\n if target_node = @target_class.find(attributes[:id])\n target_node.update_attributes!(attributes)\n else\n target_node = @target_class.create!(attributes)\n end\n\n self.relate(target_node, rel_attrs)\n\n # finally, make and return the tuple\n #Tuple.new(@node, target_node)\n #Tuple.new(@node, target_node)\n\n end",
"def grid_operation_checker(candidate)\n candidate.node_type == :call and\n candidate.sons.first.node_type == :lval and\n (candidate.grid_objects.map {|g_obj| g_obj.name}).include? candidate.sons.first.name\n end",
"def tuple_print(results)\n results.each do |tuple|\n puts tuple\n end\nend",
"def check_permission_or_error_with_http_info(opts = {})\n if @api_client.config.debugging\n @api_client.config.logger.debug 'Calling API: PermissionApi.check_permission_or_error ...'\n end\n # resource path\n local_var_path = '/relation-tuples/check'\n\n # query parameters\n query_params = opts[:query_params] || {}\n query_params[:'namespace'] = opts[:'namespace'] if !opts[:'namespace'].nil?\n query_params[:'object'] = opts[:'object'] if !opts[:'object'].nil?\n query_params[:'relation'] = opts[:'relation'] if !opts[:'relation'].nil?\n query_params[:'subject_id'] = opts[:'subject_id'] if !opts[:'subject_id'].nil?\n query_params[:'subject_set.namespace'] = opts[:'subject_set_namespace'] if !opts[:'subject_set_namespace'].nil?\n query_params[:'subject_set.object'] = opts[:'subject_set_object'] if !opts[:'subject_set_object'].nil?\n query_params[:'subject_set.relation'] = opts[:'subject_set_relation'] if !opts[:'subject_set_relation'].nil?\n query_params[:'max-depth'] = opts[:'max_depth'] if !opts[:'max_depth'].nil?\n\n # header parameters\n header_params = opts[:header_params] || {}\n # HTTP header 'Accept' (if needed)\n header_params['Accept'] = @api_client.select_header_accept(['application/json'])\n\n # form parameters\n form_params = opts[:form_params] || {}\n\n # http body (model)\n post_body = opts[:debug_body]\n\n # return_type\n return_type = opts[:debug_return_type] || 'CheckPermissionResult'\n\n # auth_names\n auth_names = opts[:debug_auth_names] || ['oryAccessToken']\n\n new_options = opts.merge(\n :operation => :\"PermissionApi.check_permission_or_error\",\n :header_params => header_params,\n :query_params => query_params,\n :form_params => form_params,\n :body => post_body,\n :auth_names => auth_names,\n :return_type => return_type\n )\n\n data, status_code, headers = @api_client.call_api(:GET, local_var_path, new_options)\n if @api_client.config.debugging\n @api_client.config.logger.debug \"API called: PermissionApi#check_permission_or_error\\nData: #{data.inspect}\\nStatus code: #{status_code}\\nHeaders: #{headers}\"\n end\n return data, status_code, headers\n end",
"def is_triplet(n1,n2,n3)\n n1**2 + n2**2 == n3**2 or n1**2 + n3**2 == n2**2 or n2**2 + n3**2 == n1**2\nend",
"def is_pythag_triplet?(a,b,c)\n a < b && b < c\nend",
"def create\n @tuple = Tuple.new(params[:tuple])\n\n respond_to do |format|\n if @tuple.save\n format.html { redirect_to(@tuple, :notice => 'Tuple was successfully created.') }\n format.xml { render :xml => @tuple, :status => :created, :location => @tuple }\n else\n format.html { render :action => \"new\" }\n format.xml { render :xml => @tuple.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def _add_tuple(objdata_, tuple_) # :nodoc:\n return false if @tuples.has_key?(tuple_)\n @tuples[tuple_] = objdata_\n tupcats_ = []\n @categories.each do |category_, catdata_|\n if Utils.matches?(catdata_[0], tuple_)\n index_ = catdata_[1].map{ |i_| tuple_[i_] }\n ((@catdata[category_] ||= {})[index_] ||= {})[tuple_] = objdata_\n tupcats_ << category_\n end\n end\n objdata_[2][tuple_] = tupcats_\n true\n end",
"def has_edge?(vertex1, vertex2)\n\tend",
"def isTriangle(a,b,c)\n a + b > c && b + c > a && a + c > b\nend",
"def isTriangle(a,b,c)\n a + b > c && b + c > a && a + c > b\nend",
"def lift_tuple(tuple)\n if Pione::TupleSpace[tuple.first]\n if pos = Pione::TupleSpace[tuple.first].location_position\n if new_location = lift_location(tuple[pos])\n tuple = tuple.clone\n tuple[pos] = new_location\n end\n end\n end\n return tuple\n end",
"def _remove_tuple(objdata_, tuple_) # :nodoc:\n tupcats_ = objdata_[2][tuple_]\n return false unless tupcats_\n @tuples.delete(tuple_)\n tupcats_.each do |cat_|\n index_ = @categories[cat_][1].map{ |i_| tuple_[i_] }\n @catdata[cat_][index_].delete(tuple_)\n end\n objdata_[2].delete(tuple_)\n true\n end",
"def validate_relation(klass, name, options = {})\n [name, \"#{name}?\".to_sym, \"#{name}=\".to_sym].each do |n|\n if Mongoid.destructive_fields.include?(n)\n raise Errors::InvalidRelation.new(klass, n)\n end\n end\n end",
"def call(relation)\n relation\n end",
"def valid?\n # Check tx = <lx, rx>\n unless FIELD.mod(lx.zip(rx).map { |a, b| FIELD.mod(a * b) }.sum) == tx\n return false\n end\n\n return false unless valid_poly_t?\n\n p =\n vec_g\n .zip(lx)\n .map { |a, b| a * b }\n .zip(vec_h2.zip(rx).map { |a, b| a * b })\n .map { |a, b| a + b }\n .sum(GROUP.infinity)\n\n p1 == p2 && p2 == p\n end",
"def op_rel!\n # -> uncomment the next line to manually enable rule tracing\n # trace_in( __method__, 17 )\n\n\n\n type = OP_REL\n channel = ANTLR3::DEFAULT_CHANNEL\n # - - - - label initialization - - - -\n\n\n # - - - - main rule block - - - -\n # at line 321:4: ( '>=' | '<=' | '==' | '!=' | '<' | '>' )\n # at line 321:4: ( '>=' | '<=' | '==' | '!=' | '<' | '>' )\n alt_1 = 6\n case look_1 = @input.peek( 1 )\n when 0x3e then look_1_1 = @input.peek( 2 )\n\n if ( look_1_1 == 0x3d )\n alt_1 = 1\n else\n alt_1 = 6\n\n end\n when 0x3c then look_1_2 = @input.peek( 2 )\n\n if ( look_1_2 == 0x3d )\n alt_1 = 2\n else\n alt_1 = 5\n\n end\n when 0x3d then alt_1 = 3\n when 0x21 then alt_1 = 4\n else\n raise NoViableAlternative( \"\", 1, 0 )\n\n end\n case alt_1\n when 1\n # at line 321:5: '>='\n match( \">=\" )\n\n\n when 2\n # at line 321:10: '<='\n match( \"<=\" )\n\n\n when 3\n # at line 321:15: '=='\n match( \"==\" )\n\n\n when 4\n # at line 321:20: '!='\n match( \"!=\" )\n\n\n when 5\n # at line 321:25: '<'\n match( 0x3c )\n\n when 6\n # at line 321:29: '>'\n match( 0x3e )\n\n end\n\n\n @state.type = type\n @state.channel = channel\n ensure\n # -> uncomment the next line to manually enable rule tracing\n # trace_out( __method__, 17 )\n\n\n end",
"def test_two_proper_args\n assert check_args([1, 1])\n end",
"def test_check_let\n RPN rpn = RPN.new(1)\n let = 'let'\n assert_equal let, 'let'\n assert rpn.check_let(let)\n end",
"def valid_triangle?(a, b, c)\n # Your code goes here!\na+b > c && a+c > b && b+c >a\nend",
"def tuples_for(arg_)\n objdata_ = _get_objdata(arg_)\n objdata_ ? objdata_[2].keys : nil\n end",
"def relation(rel, &relation_definition)\n raise NotImplementedError, 'relation is not supported yet'\n end",
"def tuple_count\n target.count\n end",
"def get_relational(expr)\n\t\t[\"==\", \"!=\", \">\", \"<\", \">=\", \"<=\"].each do |ope|\n\t\t\tif expr.scan(ope).length == 1\n\t\t\t\treturn ope\n\t\t\tend\n\t\tend\n\tend",
"def one_pair?\n n_of_a_kind? 2\n end",
"def test_history_equals_versions\n way = create(:way, :with_history)\n used_way = create(:way, :with_history)\n create(:relation_member, :member => used_way)\n way_with_versions = create(:way, :with_history, :version => 4)\n\n check_history_equals_versions(way.id)\n check_history_equals_versions(used_way.id)\n check_history_equals_versions(way_with_versions.id)\n end",
"def <= other\n call_enum \"relational\", other, :lesseq\n end",
"def to_generate_tuple(&block_)\n if block_\n @gen_tuple = block_\n else\n @gen_tuple\n end\n end",
"def check_rule_ast(pt)\n # :defn format: node tag, block name, args, body_0, ..., body_n\n tag, name, args, *body = pt\n return pt if tag != :defn\n\n body.each_with_index do |n,i|\n next if i == 0 and n == s(:nil) # a block got rewritten to an empty block\n\n # Check for a common case\n if n.sexp_type == :lasgn\n return [n, \"illegal operator: '='\"]\n end\n return pt unless n.sexp_type == :call and n.length == 4\n\n # Rule format: call tag, lhs, op, rhs\n _, lhs, op, rhs = n\n\n # Check that LHS references a named collection\n lhs_name = get_qual_name(lhs)\n return [n, \"unexpected lhs format: #{lhs}\"] if lhs_name.nil?\n unless @bud_instance.tables.has_key? lhs_name.to_sym or\n @bud_instance.lattices.has_key? lhs_name.to_sym\n return [n, \"Collection does not exist: '#{lhs_name}'\"]\n end\n\n return [n, \"illegal operator: '#{op}'\"] unless [:<, :<=].include? op\n\n # Check superator invocation. A superator that begins with \"<\" is parsed\n # as a call to the binary :< operator. The right operand to :< is a :call\n # node; the LHS of the :call is the actual rule body, the :call's oper is\n # the rest of the superator (unary ~, -, +), and the RHS is empty. Note\n # that ParseTree encodes unary \"-\" and \"+\" as :-@ and :-+, respectively.\n # XXX: We don't check for illegal superators (e.g., \"<--\"). That would be\n # tricky, because they are encoded as a nested unary op in the rule body.\n if op == :<\n return n unless rhs.sexp_type == :call\n op_tail = rhs[2]\n return n unless [:~, :-@, :+@].include? op_tail\n rhs_args = rhs[3..-1]\n return n unless rhs_args.empty?\n end\n end\n\n return nil # No errors found\n end",
"def valid_token_tuple?(type, *args)\n args.flatten!\n case type\n when :amazon then valid_amazon_tokens?(*args)\n when :bitlyv1 then valid_bitlyv1_tokens?(*args)\n when :bitlyv2 then valid_bitlyv2_tokens?(*args)\n when :facebook then valid_facebook_tokens?(*args)\n when :flickr then valid_flickr_tokens?(*args)\n when :foursquare then valid_foursquare_tokens?(*args)\n when :google then valid_google_tokens?(*args)\n when :google_oauth then valid_google_oauth_tokens?(*args)\n when :linkedin then valid_linkedin_tokens?(*args)\n when :twitter then valid_twitter_tokens?(*args)\n when :yelpv1 then valid_yelpv1_tokens?(*args)\n when :yelpv2 then valid_yelpv2_tokens?(*args)\n else\n # Rails.logger.error \"Unknown token type: #{type}\"\n end\nend",
"def check_error(val_a, val_b, val_c)\n raise TriangleError if (val_b + val_c <= val_a) || (val_c + val_a <= val_b)\n raise TriangleError if val_a + val_b <= val_c\nend",
"def valid_triangle?(a, b, c)\n ((a+b>c) && (b+c>a) && (a+c>b))? true:false\nend",
"def test_friendp_returns_true_if_users_are_friends\n assert users(:active).friend?(users(:passive))\n assert users(:passive).friend?(users(:active))\n end",
"def to_ruby_literal\n \"Alf::Relation[\" +\n tuples.collect{|t| Tools.to_ruby_literal(t)}.join(', ') + \"]\"\n end",
"def good_tuples(nums)\n count = 0\n (0..nums.size - 3).each do |i|\n first, second, third = nums[i], nums[i + 1], nums[i + 2]\n next if 3 * first == first + second + third\n count += 1 if [first, second, third].include?(first ^ second ^ third)\n\n # next if nums[i] == nums[i + 1] && nums[i + 1] == nums[i + 2]\n # next if nums[i] != nums[i + 1] && nums[i + 1] != nums[i + 2] && nums[i] != nums[i + 2]\n end\n\n count\nend",
"def valid_triangle?(a, b, c)\n a + b > c && a + c > b && b + c > a\nend",
"def valid_triangle?(a, b, c)\n a + b > c && a + c > b && b + c > a\nend",
"def valid_triangle?(a, b, c)\n a + b > c && a + c > b && b + c > a\nend",
"def triplet_check(a,b,c)\n\tif a**2+b**2 == c**2\n\t\treturn true\n\telse\n\t\treturn false\n\tend\nend",
"def visit_axiom_relation_operation_binary(binary)\n generator = self.class.visit(binary)\n @name = generator.name\n @from = aliased_subquery(generator)\n generator\n end",
"def valid_triangle?(a, b, c)\n a + b > c && b + c > a && c + a > b\nend",
"def nested?\n self.resource_tuple.length > 1\n end",
"def right_triangle?(a, b, c)\n a**2 + b**2 == c**2\nend",
"def relationship(*args)\n options = args.extract_options!\n if options[:raise_on_multiple]\n rels = relationships.take(2)\n raise _(\"Multiple relationships found\") if rels.length > 1\n rels.first\n else\n relationships.first\n end\n end",
"def isTriangle(a,b,c)\n (a+b>c && a+c>b && c+b>a)? true : false\nend",
"def >= other\n call_enum \"relational\", other, :moreeq\n end",
"def isTriangle(a, b, c)\n a, b, c = [a, b, c].sort\n a + b > c\nend",
"def test_get_location_bad_input\n assert_raises 'Improper typing of location input argument in method get_location' do\n locations = LocationRelation.new\n locations.get_location 55\n end\n end",
"def insert(tuples)\n pks = tuples.map { |tuple| relation.insert(tuple) }\n relation.where(relation.primary_key => pks).to_a\n end",
"def to_ruby_literal\n \"Alf::Relation([\" + tuples.map{|t| Support.to_ruby_literal(t) }.join(', ') + \"])\"\n end",
"def evaluate\n distinct? ? distinct_relation : relation\n end",
"def check_permission_with_http_info(opts = {})\n if @api_client.config.debugging\n @api_client.config.logger.debug 'Calling API: PermissionApi.check_permission ...'\n end\n # resource path\n local_var_path = '/relation-tuples/check/openapi'\n\n # query parameters\n query_params = opts[:query_params] || {}\n query_params[:'namespace'] = opts[:'namespace'] if !opts[:'namespace'].nil?\n query_params[:'object'] = opts[:'object'] if !opts[:'object'].nil?\n query_params[:'relation'] = opts[:'relation'] if !opts[:'relation'].nil?\n query_params[:'subject_id'] = opts[:'subject_id'] if !opts[:'subject_id'].nil?\n query_params[:'subject_set.namespace'] = opts[:'subject_set_namespace'] if !opts[:'subject_set_namespace'].nil?\n query_params[:'subject_set.object'] = opts[:'subject_set_object'] if !opts[:'subject_set_object'].nil?\n query_params[:'subject_set.relation'] = opts[:'subject_set_relation'] if !opts[:'subject_set_relation'].nil?\n query_params[:'max-depth'] = opts[:'max_depth'] if !opts[:'max_depth'].nil?\n\n # header parameters\n header_params = opts[:header_params] || {}\n # HTTP header 'Accept' (if needed)\n header_params['Accept'] = @api_client.select_header_accept(['application/json'])\n\n # form parameters\n form_params = opts[:form_params] || {}\n\n # http body (model)\n post_body = opts[:debug_body]\n\n # return_type\n return_type = opts[:debug_return_type] || 'CheckPermissionResult'\n\n # auth_names\n auth_names = opts[:debug_auth_names] || ['oryAccessToken']\n\n new_options = opts.merge(\n :operation => :\"PermissionApi.check_permission\",\n :header_params => header_params,\n :query_params => query_params,\n :form_params => form_params,\n :body => post_body,\n :auth_names => auth_names,\n :return_type => return_type\n )\n\n data, status_code, headers = @api_client.call_api(:GET, local_var_path, new_options)\n if @api_client.config.debugging\n @api_client.config.logger.debug \"API called: PermissionApi#check_permission\\nData: #{data.inspect}\\nStatus code: #{status_code}\\nHeaders: #{headers}\"\n end\n return data, status_code, headers\n end",
"def with_input_tuples(tuples, &block)\n input_tuples = Array([tuples]).flatten(1).map\n return input_tuples unless block_given?\n\n input_tuples.each(&block)\n end",
"def test_two_valid_arguments\n\t\tc = Check.new\n\t\trefute c.check_arguments([1,1])\n\tend",
"def test_triangle\n if @side1.eql?(@side2) && @side2.eql?(@side3)\n \"It's an equilateral\"\n elsif @side1.eql?(@side2) || @side2.eql?(@side3)\n \"It's an isosceles\"\n elsif @side1.eql?(@side2) == false && @side2.eql?(@side3) == false\n \"It's a scalene triangle\"\n else\n \"Not a triangle\"\n end\n end",
"def parse_relational\n add = parse_add\n if consume(:lt)\n Node.new(:binary, op: :lt, lhs: add, rhs: parse_relational)\n elsif consume(:lteq)\n Node.new(:binary, op: :lteq, lhs: add, rhs: parse_relational)\n elsif consume(:gt)\n Node.new(:binary, op: :gt, lhs: add, rhs: parse_relational)\n elsif consume(:gteq)\n Node.new(:binary, op: :gteq, lhs: add, rhs: parse_relational)\n else\n add\n end\n end",
"def pitagoric_triplet?(a,b,c)\n (a**2 + b**2) == c**2\nend",
"def selecionErr(method)\n\n tnTableCreation('tuple_node_test_result') if @is_new\n\n if @unwanted_tuple_count + @missing_tuple_count == 0\n p 'no failed rows found. There is no selection error'\n return\n end\n\n if @unwanted_tuple_count > 0\n # p \"Unwanted Pk count #{unWantedPK.count()}\"\n # create unwanted_tuple_branch table\n # binding.pry\n\n # whereErrList = whereCondTest(@unWantedPK, 'U')\n where_cond_test('U')\n # joinErrList = jointypeErr(query,'U')\n end\n\n if @missing_tuple_count > 0\n # p \"Missing PK count #{missinPK.count()}\"\n # binding.pry\n # whereErrList = whereCondTest(@missingPK, 'M')\n where_cond_test('M')\n # joinErrList = jointypeErr(query,'M')\n end\n # create aggregated tuple_suspicious_nodes\n pk = @pkFullList.map { |pk| pk['alias'] }.join(',')\n query = 'create materialized view tuple_node_test_result_aggr as '\\\n \"select #{pk}, string_agg(branch_name||'-'||node_name, ',' order by node_name,branch_name) as suspicious_nodes from tuple_node_test_result group by #{pk}\"\n pp query\n DBConn.exec(query)\n\n unless @predicateTree.nil?\n suspicious_score_upd(@predicateTree.branches)\n end\n # exnorate algorithm\n # binding.pry\n @column_combinations = method.start_with?('o') ? Columns_Combination.new(@all_columns) : @all_columns\n case method\n # when 'o'\n # puts 'old exonerate algorithm'\n\n # true_query_PT_construct\n # constraint_query = constraint_predicate_construct\n # # column_combinations_construct\n # tuple_mutation_test(@missingPK, 'M', constraint_query, false)\n # tuple_mutation_test(@unWantedPK, 'U', constraint_query, false)\n when 'or'\n puts 'old exonerate algorithm with duplicate removal'\n # reset suspicious score\n query = \"update node_query_mapping set suspicious_score = 0 where type = 'f'\"\n res = DBConn.exec(query)\n\n true_query_PT_construct\n constraint_query = constraint_predicate_construct\n # allcolumns_construct()\n puts 'Missing starts'\n puts Time.now()\n tuple_mutation_test_with_dup_removal('M', constraint_query)\n puts 'Unwanted starts'\n puts Time.now()\n\n tuple_mutation_test_with_dup_removal('U', constraint_query)\n # when 'n'\n # puts 'new exonerate algorithm'\n # true_query_PT_construct\n # constraint_query = constraint_predicate_construct\n # tuple_mutation_test_reverse(@missingPK, 'M', constraint_query, false)\n # tuple_mutation_test_reverse(@unWantedPK, 'U', constraint_query, false)\n\n when 'b'\n puts 'baseline'\n else\n puts 'Unknown method'\n end\n # remove constraint_nodes in node_query_mapping\n query = \"delete from node_query_mapping where test_id = #{@test_id} and type = 't'\"\n DBConn.exec(query)\n\n # j = Hash.new\n # # j['JoinErr'] = joinErrList if joinErrList.count >0\n # j['WhereErr'] = whereErrList\n # j\n # whereErrList\n end",
"def isTriangle(a,b,c)\n a+b>c && a + c > b && b + c > a ? true : false\nend",
"def arel_bt_intersect(*bt_tuple)\n arel_vt_intersect(bt_tuple.at(0), bt_tuple.at(1)).and(arel_tt_intersect(bt_tuple.at(2), bt_tuple.at(3)))\n end"
] |
[
"0.7015953",
"0.6605728",
"0.6361857",
"0.6198903",
"0.60191625",
"0.579029",
"0.5698508",
"0.5596643",
"0.5500909",
"0.53615284",
"0.5320704",
"0.52337974",
"0.51784104",
"0.51652044",
"0.5137385",
"0.5097203",
"0.5097203",
"0.5095431",
"0.50804377",
"0.5077863",
"0.5056793",
"0.5047188",
"0.50432706",
"0.50223356",
"0.49960107",
"0.49366918",
"0.49293387",
"0.4916955",
"0.4907835",
"0.48614576",
"0.48245752",
"0.4788206",
"0.4757173",
"0.47517347",
"0.4732002",
"0.46899974",
"0.46806657",
"0.46671054",
"0.46462044",
"0.46259937",
"0.46217015",
"0.46073496",
"0.4602068",
"0.45971423",
"0.4588765",
"0.45760465",
"0.4574195",
"0.45620072",
"0.45608",
"0.45552087",
"0.4553977",
"0.45391572",
"0.45349598",
"0.45325094",
"0.45259938",
"0.45203832",
"0.4519309",
"0.45055634",
"0.4499442",
"0.44971564",
"0.44606513",
"0.44566202",
"0.44505966",
"0.4447705",
"0.4445908",
"0.4441292",
"0.4428633",
"0.44250175",
"0.44135484",
"0.44133824",
"0.4411091",
"0.44052505",
"0.44050676",
"0.44036365",
"0.4403217",
"0.44016722",
"0.44010103",
"0.44010103",
"0.44010103",
"0.43980426",
"0.43945882",
"0.4391476",
"0.43906316",
"0.43885761",
"0.43848813",
"0.43774366",
"0.4370648",
"0.43687898",
"0.43655345",
"0.4363082",
"0.43569046",
"0.43532142",
"0.43471527",
"0.43461266",
"0.4344165",
"0.43420166",
"0.4339705",
"0.43395022",
"0.43379533",
"0.4334619",
"0.43341088"
] |
0.0
|
-1
|
Check a relation tuple To learn how relation tuples and the check works, head over to [the documentation](../concepts/relationtuples.mdx).
|
def get_check_with_http_info(namespace, object, relation, opts = {})
if @api_client.config.debugging
@api_client.config.logger.debug 'Calling API: ReadApi.get_check ...'
end
# verify the required parameter 'namespace' is set
if @api_client.config.client_side_validation && namespace.nil?
fail ArgumentError, "Missing the required parameter 'namespace' when calling ReadApi.get_check"
end
# verify the required parameter 'object' is set
if @api_client.config.client_side_validation && object.nil?
fail ArgumentError, "Missing the required parameter 'object' when calling ReadApi.get_check"
end
# verify the required parameter 'relation' is set
if @api_client.config.client_side_validation && relation.nil?
fail ArgumentError, "Missing the required parameter 'relation' when calling ReadApi.get_check"
end
# resource path
local_var_path = '/check'
# query parameters
query_params = opts[:query_params] || {}
query_params[:'namespace'] = namespace
query_params[:'object'] = object
query_params[:'relation'] = relation
query_params[:'subject_id'] = opts[:'subject_id'] if !opts[:'subject_id'].nil?
query_params[:'subject_set.namespace'] = opts[:'subject_set_namespace'] if !opts[:'subject_set_namespace'].nil?
query_params[:'subject_set.object'] = opts[:'subject_set_object'] if !opts[:'subject_set_object'].nil?
query_params[:'subject_set.relation'] = opts[:'subject_set_relation'] if !opts[:'subject_set_relation'].nil?
# header parameters
header_params = opts[:header_params] || {}
# HTTP header 'Accept' (if needed)
header_params['Accept'] = @api_client.select_header_accept(['application/json'])
# form parameters
form_params = opts[:form_params] || {}
# http body (model)
post_body = opts[:debug_body]
# return_type
return_type = opts[:debug_return_type] || 'GetCheckResponse'
# auth_names
auth_names = opts[:debug_auth_names] || []
new_options = opts.merge(
:operation => :"ReadApi.get_check",
:header_params => header_params,
:query_params => query_params,
:form_params => form_params,
:body => post_body,
:auth_names => auth_names,
:return_type => return_type
)
data, status_code, headers = @api_client.call_api(:GET, local_var_path, new_options)
if @api_client.config.debugging
@api_client.config.logger.debug "API called: ReadApi#get_check\nData: #{data.inspect}\nStatus code: #{status_code}\nHeaders: #{headers}"
end
return data, status_code, headers
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def valid_relation_literal!(heading, tuples, raise_on_error = true)\n if Array === tuples\n invalid = tuples.find{|t| !valid_tuple_literal?(heading, t)}\n if invalid\n raise_on_error ? raise(::Veritas::RelationMismatchError, \"Invalid tuple #{invalid.inspect} for heading #{heading}\", caller) : false\n else\n tuples\n end\n else\n raise_on_error ? raise(::Veritas::RelationMismatchError, \"Invalid relation literal #{tuples.inspect}\", caller) : false \n end\n end",
"def check_for_tuple(tuple)\n ltuple = @left_pattern.map {|i| tuple[i]}\n if @left.check_for_tuple(ltuple)\n rtuple = @right_pattern.map {|i| tuple[i]}\n if @right.check_for_tuple(rtuple)\n return true\n end\n end\n return false\n end",
"def contains?(tuple)\n case tuple\n when Hash\n eql_hash = {}\n relvar_def.primary_key.attributes.each{|a| eql_hash[a.name] = tuple[a.name]}\n inside_tuple = underlying_table.filter(eql_hash).first\n return inside_tuple === tuple\n else\n raise ArgumentError, \"Unable to check inclusion of #{tuple} inside a relation variable\"\n end\n end",
"def valid_tuple_literal!(heading, tuple, raise_on_error = true)\n if Hash===tuple and tuple.size == heading.to_ary.size\n ok = heading.all?{|a| tuple.key?(a.name) and a.valid_value?(tuple[a.name])}\n return tuple if ok\n end\n raise_on_error ? raise(ArgumentError, \"Invalid tuple #{tuple} for #{heading}\", caller) : false\n end",
"def _tuple2tuple(tuple)\n end",
"def tuple_plan()\n raise NotTupleValuedException.new(self) unless tuple_valued?\n @definition.type.tuple.plan\n end",
"def create_relation_tuple(opts = {})\n data, _status_code, _headers = create_relation_tuple_with_http_info(opts)\n data\n end",
"def check_relation(relation)\n (relation - reflection.map{|r| r[:rel] }).empty?\n end",
"def relation\n # -> uncomment the next line to manually enable rule tracing\n # trace_in( __method__, 28 )\n\n\n value = nil\n\n\n type = nil\n a = nil\n b = nil\n\n\n begin\n # at line 178:5: a= add (type= ( '==' | '!=' | '<' | '<=' | '>=' | '>' ) b= relation )?\n @state.following.push( TOKENS_FOLLOWING_add_IN_relation_1322 )\n a = add\n @state.following.pop\n\n # syntactic predicate action gate test\n if @state.backtracking == 0\n # --> action\n value = a \n # <-- action\n end\n\n # at line 179:3: (type= ( '==' | '!=' | '<' | '<=' | '>=' | '>' ) b= relation )?\n alt_44 = 2\n look_44_0 = @input.peek( 1 )\n\n if ( look_44_0 == T__31 || look_44_0.between?( T__47, T__48 ) || look_44_0.between?( T__50, T__52 ) )\n alt_44 = 1\n end\n case alt_44\n when 1\n # at line 179:6: type= ( '==' | '!=' | '<' | '<=' | '>=' | '>' ) b= relation\n type = @input.look\n\n if @input.peek(1) == T__31 || @input.peek( 1 ).between?( T__47, T__48 ) || @input.peek( 1 ).between?( T__50, T__52 )\n @input.consume\n @state.error_recovery = false\n\n else\n @state.backtracking > 0 and raise( ANTLR3::Error::BacktrackingFailed )\n\n\n mse = MismatchedSet( nil )\n raise mse\n\n end\n\n\n @state.following.push( TOKENS_FOLLOWING_relation_IN_relation_1359 )\n b = relation\n @state.following.pop\n\n # syntactic predicate action gate test\n if @state.backtracking == 0\n # --> action\n value = DoubleOperandExpressionEval.new(type.text, a, b) \n # <-- action\n end\n\n\n end\n\n rescue ANTLR3::Error::RecognitionError => re\n report_error(re)\n recover(re)\n\n ensure\n # -> uncomment the next line to manually enable rule tracing\n # trace_out( __method__, 28 )\n\n\n end\n\n return value\n end",
"def eval_tuple(node, env, stack_trace)\n\n\tevaluated_nodes = []\n\tnode.values.each{ |sub_node|\n\t\tsub_node_value, _ = eval_node_under(sub_node, env, stack_trace) \n\t\tevaluated_nodes.append(sub_node_value)\n\t}\n\tif evaluated_nodes.length == 1\n\t\treturn evaluated_nodes[0], env\n\telse\n\t\treturn Tuple.new(node.line, node.col, evaluated_nodes)\n\tend\nend",
"def triangle?((x1, y1), (x2, y2), (x3, y3))\n return false if (x1, y1) == (x2, y2) || (x1, y1) == (x3, y3) || (x2, y2) == (x3, y3)\n return false \n true\nend",
"def patch_relation_tuples(opts = {})\n patch_relation_tuples_with_http_info(opts)\n nil\n end",
"def is_meta_tuple? tuple\n if meta_subspace\n meta_subspace === tuple\n else\n # meta_subspace hasn't arrived yet, so use approximation\n tuple.kind_of? Hash and tuple.key? client.tupelo_meta_key and\n tuple[client.tupelo_meta_key] == \"subspace\"\n end\n end",
"def t\n BERT::Tuple\nend",
"def activity_relations?(*args)\n activity_relations(*args).any?\n end",
"def tuples; end",
"def tuples; end",
"def assert_working_relations( model, relations )\n failed = []\n relations.each do | name |\n begin\n model.send( name )\n rescue Exception => e\n failed << \"#{name} - #{e}\"\n end\n end\n if failed.empty?\n assert true\n else\n assert false, failed.join('; ')\n end\n end",
"def check_relation(x)\n h, qx = compute_q(x)\n return if qx == 0\n es, l = exponent_bitvector(qx)\n\n # discard this x if the residue L is too big\n return if l > @pmax ** @t\n\n if l == 1\n # complete relation found:\n # Q(x) = p0^e0 * p1^e1 * ... * pk^ek (pi in the factor base)\n qx_vec = Hash.new(0)\n PrimeFactorization.prime_factorization(qx) {|p, e| qx_vec[p] += e }\n collect_relation(es, h, qx_vec)\n\n elsif @incomplete_relations[l]\n # large prime procedure:\n # make a complete relation by multiplying two incomplete relations\n es2, h2, qx2 = @incomplete_relations[l]\n\n # XXX: use FactoredInteger\n qx_vec = Hash.new(0)\n PrimeFactorization.prime_factorization(qx) {|p, e| qx_vec[p] += e }\n PrimeFactorization.prime_factorization(qx2) {|p, e| qx_vec[p] += e }\n\n collect_relation(es ^ es2, h * h2 % @kn, qx_vec)\n\n else\n @incomplete_relations[l] = [es, h, qx]\n end\n end",
"def verify_uniquness!(tuple)\n stream = tuple[:stream]\n attrs = %i[position event_id]\n attrs.delete(:position) if Stream.new(stream).global?\n\n attrs.each do |key|\n next if key == :position && tuple[key].nil?\n next if restrict(:stream => stream, key => tuple.fetch(key)).none?\n\n raise TupleUniquenessError.public_send(:\"for_stream_and_#{key}\", stream, tuple.fetch(key))\n end\n end",
"def create_relation_tuple_with_http_info(opts = {})\n if @api_client.config.debugging\n @api_client.config.logger.debug 'Calling API: WriteApi.create_relation_tuple ...'\n end\n # resource path\n local_var_path = '/relation-tuples'\n\n # query parameters\n query_params = opts[:query_params] || {}\n\n # header parameters\n header_params = opts[:header_params] || {}\n # HTTP header 'Accept' (if needed)\n header_params['Accept'] = @api_client.select_header_accept(['application/json'])\n # HTTP header 'Content-Type'\n header_params['Content-Type'] = @api_client.select_header_content_type(['application/json'])\n\n # form parameters\n form_params = opts[:form_params] || {}\n\n # http body (model)\n post_body = opts[:debug_body] || @api_client.object_to_http_body(opts[:'payload'])\n\n # return_type\n return_type = opts[:debug_return_type] || 'RelationQuery'\n\n # auth_names\n auth_names = opts[:debug_auth_names] || []\n\n new_options = opts.merge(\n :operation => :\"WriteApi.create_relation_tuple\",\n :header_params => header_params,\n :query_params => query_params,\n :form_params => form_params,\n :body => post_body,\n :auth_names => auth_names,\n :return_type => return_type\n )\n\n data, status_code, headers = @api_client.call_api(:PUT, local_var_path, new_options)\n if @api_client.config.debugging\n @api_client.config.logger.debug \"API called: WriteApi#create_relation_tuple\\nData: #{data.inspect}\\nStatus code: #{status_code}\\nHeaders: #{headers}\"\n end\n return data, status_code, headers\n end",
"def delete_relation_tuple_with_http_info(namespace, object, relation, opts = {})\n if @api_client.config.debugging\n @api_client.config.logger.debug 'Calling API: WriteApi.delete_relation_tuple ...'\n end\n # verify the required parameter 'namespace' is set\n if @api_client.config.client_side_validation && namespace.nil?\n fail ArgumentError, \"Missing the required parameter 'namespace' when calling WriteApi.delete_relation_tuple\"\n end\n # verify the required parameter 'object' is set\n if @api_client.config.client_side_validation && object.nil?\n fail ArgumentError, \"Missing the required parameter 'object' when calling WriteApi.delete_relation_tuple\"\n end\n # verify the required parameter 'relation' is set\n if @api_client.config.client_side_validation && relation.nil?\n fail ArgumentError, \"Missing the required parameter 'relation' when calling WriteApi.delete_relation_tuple\"\n end\n # resource path\n local_var_path = '/relation-tuples'\n\n # query parameters\n query_params = opts[:query_params] || {}\n query_params[:'namespace'] = namespace\n query_params[:'object'] = object\n query_params[:'relation'] = relation\n query_params[:'subject_id'] = opts[:'subject_id'] if !opts[:'subject_id'].nil?\n query_params[:'subject_set.namespace'] = opts[:'subject_set_namespace'] if !opts[:'subject_set_namespace'].nil?\n query_params[:'subject_set.object'] = opts[:'subject_set_object'] if !opts[:'subject_set_object'].nil?\n query_params[:'subject_set.relation'] = opts[:'subject_set_relation'] if !opts[:'subject_set_relation'].nil?\n\n # header parameters\n header_params = opts[:header_params] || {}\n # HTTP header 'Accept' (if needed)\n header_params['Accept'] = @api_client.select_header_accept(['application/json'])\n\n # form parameters\n form_params = opts[:form_params] || {}\n\n # http body (model)\n post_body = opts[:debug_body]\n\n # return_type\n return_type = opts[:debug_return_type]\n\n # auth_names\n auth_names = opts[:debug_auth_names] || []\n\n new_options = opts.merge(\n :operation => :\"WriteApi.delete_relation_tuple\",\n :header_params => header_params,\n :query_params => query_params,\n :form_params => form_params,\n :body => post_body,\n :auth_names => auth_names,\n :return_type => return_type\n )\n\n data, status_code, headers = @api_client.call_api(:DELETE, local_var_path, new_options)\n if @api_client.config.debugging\n @api_client.config.logger.debug \"API called: WriteApi#delete_relation_tuple\\nData: #{data.inspect}\\nStatus code: #{status_code}\\nHeaders: #{headers}\"\n end\n return data, status_code, headers\n end",
"def tuple_extract(*args, &bl)\n alf_connection.tuple_extract(*args, &bl)\n end",
"def relation(relation)\n true\n end",
"def tuple_valued?()\n @definition.type.named_type?\n end",
"def delete_relation_tuple(namespace, object, relation, opts = {})\n delete_relation_tuple_with_http_info(namespace, object, relation, opts)\n nil\n end",
"def check_relation_with friend_id\n response = {\n message: \"Successfully fecth user relation\",\n following: self.active_relationships.where(follower_id: friend_id).present?,\n follower: self.passive_relationships.where(follower_id: friend_id).present?\n }\n end",
"def is_valid_relation?(parent, child, require_both = self.require_both)\n # depending upon the\n require_both ? is_valid_down_relation?(parent, child) && is_valid_up_relation?(parent, child)\n : is_valid_down_relation?(parent, child) || is_valid_up_relation?(parent, child)\n end",
"def parse_tuple cur_tok\n\t\tcur_ast = Tuple.new(cur_tok.line, cur_tok.col, splitter( nil, \")\", \",\")) # start=nil because \"(\" token has already been read.\n\n\t\t# Use look ahead to see if the tuple needs to be a child node of an operation.\n\t\tif is_operation \n\t\t\treturn parse_operation(cur_ast, true)\n\t\telse\n\t\t\treturn cur_ast \n\t\tend\n\tend",
"def deduce\n each_vertex do |v1|\n each_vertex do |v2|\n each_vertex do |v3|\n \n if self[v1,v2] == :provable_true and self[v2,v3] == :provable_true\n add_assertion(v1, v3, :provable_true)\n end\n \n if self[v2,v1] == :provable_false and self[v2,v3] == :provable_true\n add_assertion(v3, v1, :provable_false)\n end\n \n if self[v1,v2] == :provable_true and self[v3,v2] == :provable_false\n add_assertion(v3, v1, :provable_false)\n end\n \n break if @contradiction\n end\n end\n end\n end",
"def cast_tuple\n <<-CODE\n t1 = stack_pop();\n if(REFERENCE_P(t1) && object_kind_of_p(state, t1, global->array)) {\n j = N2I(array_get_total(t1));\n t2 = tuple_new(state, j);\n\n for(k = 0; k < j; k++) {\n tuple_put(state, t2, k, array_get(state, t1, k));\n }\n t1 = t2;\n } else if(!REFERENCE_P(t1) || !object_kind_of_p(state, t1, global->tuple)) {\n t2 = tuple_new(state, 1);\n tuple_put(state, t2, 0, t1);\n t1 = t2;\n }\n stack_push(t1);\n CODE\n end",
"def assert_relations_are_equal(a, b)\n assert_not_nil a, \"first relation is not allowed to be nil\"\n assert_not_nil b, \"second relation #{a.id} is not allowed to be nil\"\n assert_equal a.id, b.id, \"relation IDs\"\n assert_equal a.changeset_id, b.changeset_id, \"changeset ID on relation #{a.id}\"\n assert_equal a.visible, b.visible, \"visible on relation #{a.id}, #{a.visible.inspect} != #{b.visible.inspect}\"\n assert_equal a.version, b.version, \"version on relation #{a.id}\"\n assert_equal a.tags, b.tags, \"tags on relation #{a.id}\"\n assert_equal a.members, b.members, \"member references on relation #{a.id}\"\n end",
"def tuple(*elements)\n Tuple.new(*elements)\n end",
"def cmp_tuples(left, right)\n reduce(0) do |cmp, direction|\n break cmp if cmp.nonzero?\n direction.call(left, right)\n end\n end",
"def get_relation_tuples(namespace, opts = {})\n data, _status_code, _headers = get_relation_tuples_with_http_info(namespace, opts)\n data\n end",
"def func_le(args)\n p1 = _eval(car(args))\n p2 = _eval(car(cdr(args)))\n\n if p1.type != LObject::OBJ_INTEGER and p2.type != LObject::OBJ_INTEGER\n if @lint\n Error.warn(\"warning: comparison with a non integer operand\")\n end\n return @o_man.nil\n end\n\n if p1.value.i <= p2.value.i\n return @o_man.t\n end\n return @o_man.nil\n end",
"def tuple(*args)\n \"( #{args.map { |b| if b.nil? then 'NULL' else b end }.join(', ')} )\"\nend",
"def relation\n # -> uncomment the next line to manually enable rule tracing\n # trace_in( __method__, 55 )\n return_value = RelationReturnValue.new\n\n # $rule.start = the first token seen before matching\n return_value.start = @input.look\n\n root_0 = nil\n shift249 = nil\n relation_op250 = nil\n shift251 = nil\n\n\n begin\n root_0 = @adaptor.create_flat_list\n\n\n # at line 609:5: shift ( relation_op shift )*\n @state.following.push( TOKENS_FOLLOWING_shift_IN_relation_4033 )\n shift249 = shift\n @state.following.pop\n if @state.backtracking == 0\n @adaptor.add_child( root_0, shift249.tree )\n end\n # at line 609:11: ( relation_op shift )*\n while true # decision 59\n alt_59 = 2\n look_59_0 = @input.peek( 1 )\n\n if ( look_59_0 == GEQ || look_59_0 == GREATER || look_59_0 == INSTANCEOF || look_59_0 == LEQ || look_59_0 == LESS )\n alt_59 = 1\n elsif ( look_59_0 == IN ) and ( ( !( @InFor_stack.last && @InFor_stack.last.active ) ) )\n alt_59 = 1\n\n end\n case alt_59\n when 1\n # at line 609:14: relation_op shift\n @state.following.push( TOKENS_FOLLOWING_relation_op_IN_relation_4038 )\n relation_op250 = relation_op\n @state.following.pop\n if @state.backtracking == 0\n root_0 = @adaptor.become_root( relation_op250.tree, root_0 )\n end\n @state.following.push( TOKENS_FOLLOWING_shift_IN_relation_4042 )\n shift251 = shift\n @state.following.pop\n if @state.backtracking == 0\n @adaptor.add_child( root_0, shift251.tree )\n end\n\n else\n break # out of loop for decision 59\n end\n end # loop for decision 59\n # - - - - - - - rule clean up - - - - - - - -\n return_value.stop = @input.look( -1 )\n\n if @state.backtracking == 0\n\n return_value.tree = @adaptor.rule_post_processing( root_0 )\n @adaptor.set_token_boundaries( return_value.tree, return_value.start, return_value.stop )\n\n end\n rescue ANTLR3::Error::RecognitionError => re\n report_error(re)\n recover(re)\n return_value.tree = @adaptor.create_error_node( @input, return_value.start, @input.look(-1), re )\n\n ensure\n # -> uncomment the next line to manually enable rule tracing\n # trace_out( __method__, 55 )\n\n end\n \n return return_value\n end",
"def get_relation_tuples_with_http_info(namespace, opts = {})\n if @api_client.config.debugging\n @api_client.config.logger.debug 'Calling API: ReadApi.get_relation_tuples ...'\n end\n # verify the required parameter 'namespace' is set\n if @api_client.config.client_side_validation && namespace.nil?\n fail ArgumentError, \"Missing the required parameter 'namespace' when calling ReadApi.get_relation_tuples\"\n end\n # resource path\n local_var_path = '/relation-tuples'\n\n # query parameters\n query_params = opts[:query_params] || {}\n query_params[:'namespace'] = namespace\n query_params[:'page_token'] = opts[:'page_token'] if !opts[:'page_token'].nil?\n query_params[:'page_size'] = opts[:'page_size'] if !opts[:'page_size'].nil?\n query_params[:'object'] = opts[:'object'] if !opts[:'object'].nil?\n query_params[:'relation'] = opts[:'relation'] if !opts[:'relation'].nil?\n query_params[:'subject_id'] = opts[:'subject_id'] if !opts[:'subject_id'].nil?\n query_params[:'subject_set.namespace'] = opts[:'subject_set_namespace'] if !opts[:'subject_set_namespace'].nil?\n query_params[:'subject_set.object'] = opts[:'subject_set_object'] if !opts[:'subject_set_object'].nil?\n query_params[:'subject_set.relation'] = opts[:'subject_set_relation'] if !opts[:'subject_set_relation'].nil?\n\n # header parameters\n header_params = opts[:header_params] || {}\n # HTTP header 'Accept' (if needed)\n header_params['Accept'] = @api_client.select_header_accept(['application/json'])\n\n # form parameters\n form_params = opts[:form_params] || {}\n\n # http body (model)\n post_body = opts[:debug_body]\n\n # return_type\n return_type = opts[:debug_return_type] || 'GetRelationTuplesResponse'\n\n # auth_names\n auth_names = opts[:debug_auth_names] || []\n\n new_options = opts.merge(\n :operation => :\"ReadApi.get_relation_tuples\",\n :header_params => header_params,\n :query_params => query_params,\n :form_params => form_params,\n :body => post_body,\n :auth_names => auth_names,\n :return_type => return_type\n )\n\n data, status_code, headers = @api_client.call_api(:GET, local_var_path, new_options)\n if @api_client.config.debugging\n @api_client.config.logger.debug \"API called: ReadApi#get_relation_tuples\\nData: #{data.inspect}\\nStatus code: #{status_code}\\nHeaders: #{headers}\"\n end\n return data, status_code, headers\n end",
"def valid_triangle? triple\n a, b, c = triple\n return false if a + b <= c\n return false if a + c <= b\n return false if b + c <= a\n\n true\nend",
"def tuples\n relation = self.relation\n return relation if materialized?\n\n adapter.read(relation)\n end",
"def create_tuple!(attributes)\n # split the attrs (needs DRYing)\n rel_attrs = {}\n attributes.each do |key, value|\n if key.to_s[0..3] == 'rel_'\n rel_attrs[key] = attributes.delete(key)\n end\n end\n\n # find or create the one node\n if target_node = @target_class.find(attributes[:id])\n target_node.update_attributes!(attributes)\n else\n target_node = @target_class.create!(attributes)\n end\n\n self.relate(target_node, rel_attrs)\n\n # finally, make and return the tuple\n #Tuple.new(@node, target_node)\n #Tuple.new(@node, target_node)\n\n end",
"def grid_operation_checker(candidate)\n candidate.node_type == :call and\n candidate.sons.first.node_type == :lval and\n (candidate.grid_objects.map {|g_obj| g_obj.name}).include? candidate.sons.first.name\n end",
"def tuple_print(results)\n results.each do |tuple|\n puts tuple\n end\nend",
"def check_permission_or_error_with_http_info(opts = {})\n if @api_client.config.debugging\n @api_client.config.logger.debug 'Calling API: PermissionApi.check_permission_or_error ...'\n end\n # resource path\n local_var_path = '/relation-tuples/check'\n\n # query parameters\n query_params = opts[:query_params] || {}\n query_params[:'namespace'] = opts[:'namespace'] if !opts[:'namespace'].nil?\n query_params[:'object'] = opts[:'object'] if !opts[:'object'].nil?\n query_params[:'relation'] = opts[:'relation'] if !opts[:'relation'].nil?\n query_params[:'subject_id'] = opts[:'subject_id'] if !opts[:'subject_id'].nil?\n query_params[:'subject_set.namespace'] = opts[:'subject_set_namespace'] if !opts[:'subject_set_namespace'].nil?\n query_params[:'subject_set.object'] = opts[:'subject_set_object'] if !opts[:'subject_set_object'].nil?\n query_params[:'subject_set.relation'] = opts[:'subject_set_relation'] if !opts[:'subject_set_relation'].nil?\n query_params[:'max-depth'] = opts[:'max_depth'] if !opts[:'max_depth'].nil?\n\n # header parameters\n header_params = opts[:header_params] || {}\n # HTTP header 'Accept' (if needed)\n header_params['Accept'] = @api_client.select_header_accept(['application/json'])\n\n # form parameters\n form_params = opts[:form_params] || {}\n\n # http body (model)\n post_body = opts[:debug_body]\n\n # return_type\n return_type = opts[:debug_return_type] || 'CheckPermissionResult'\n\n # auth_names\n auth_names = opts[:debug_auth_names] || ['oryAccessToken']\n\n new_options = opts.merge(\n :operation => :\"PermissionApi.check_permission_or_error\",\n :header_params => header_params,\n :query_params => query_params,\n :form_params => form_params,\n :body => post_body,\n :auth_names => auth_names,\n :return_type => return_type\n )\n\n data, status_code, headers = @api_client.call_api(:GET, local_var_path, new_options)\n if @api_client.config.debugging\n @api_client.config.logger.debug \"API called: PermissionApi#check_permission_or_error\\nData: #{data.inspect}\\nStatus code: #{status_code}\\nHeaders: #{headers}\"\n end\n return data, status_code, headers\n end",
"def is_triplet(n1,n2,n3)\n n1**2 + n2**2 == n3**2 or n1**2 + n3**2 == n2**2 or n2**2 + n3**2 == n1**2\nend",
"def is_pythag_triplet?(a,b,c)\n a < b && b < c\nend",
"def create\n @tuple = Tuple.new(params[:tuple])\n\n respond_to do |format|\n if @tuple.save\n format.html { redirect_to(@tuple, :notice => 'Tuple was successfully created.') }\n format.xml { render :xml => @tuple, :status => :created, :location => @tuple }\n else\n format.html { render :action => \"new\" }\n format.xml { render :xml => @tuple.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def _add_tuple(objdata_, tuple_) # :nodoc:\n return false if @tuples.has_key?(tuple_)\n @tuples[tuple_] = objdata_\n tupcats_ = []\n @categories.each do |category_, catdata_|\n if Utils.matches?(catdata_[0], tuple_)\n index_ = catdata_[1].map{ |i_| tuple_[i_] }\n ((@catdata[category_] ||= {})[index_] ||= {})[tuple_] = objdata_\n tupcats_ << category_\n end\n end\n objdata_[2][tuple_] = tupcats_\n true\n end",
"def has_edge?(vertex1, vertex2)\n\tend",
"def isTriangle(a,b,c)\n a + b > c && b + c > a && a + c > b\nend",
"def isTriangle(a,b,c)\n a + b > c && b + c > a && a + c > b\nend",
"def lift_tuple(tuple)\n if Pione::TupleSpace[tuple.first]\n if pos = Pione::TupleSpace[tuple.first].location_position\n if new_location = lift_location(tuple[pos])\n tuple = tuple.clone\n tuple[pos] = new_location\n end\n end\n end\n return tuple\n end",
"def _remove_tuple(objdata_, tuple_) # :nodoc:\n tupcats_ = objdata_[2][tuple_]\n return false unless tupcats_\n @tuples.delete(tuple_)\n tupcats_.each do |cat_|\n index_ = @categories[cat_][1].map{ |i_| tuple_[i_] }\n @catdata[cat_][index_].delete(tuple_)\n end\n objdata_[2].delete(tuple_)\n true\n end",
"def validate_relation(klass, name, options = {})\n [name, \"#{name}?\".to_sym, \"#{name}=\".to_sym].each do |n|\n if Mongoid.destructive_fields.include?(n)\n raise Errors::InvalidRelation.new(klass, n)\n end\n end\n end",
"def call(relation)\n relation\n end",
"def valid?\n # Check tx = <lx, rx>\n unless FIELD.mod(lx.zip(rx).map { |a, b| FIELD.mod(a * b) }.sum) == tx\n return false\n end\n\n return false unless valid_poly_t?\n\n p =\n vec_g\n .zip(lx)\n .map { |a, b| a * b }\n .zip(vec_h2.zip(rx).map { |a, b| a * b })\n .map { |a, b| a + b }\n .sum(GROUP.infinity)\n\n p1 == p2 && p2 == p\n end",
"def op_rel!\n # -> uncomment the next line to manually enable rule tracing\n # trace_in( __method__, 17 )\n\n\n\n type = OP_REL\n channel = ANTLR3::DEFAULT_CHANNEL\n # - - - - label initialization - - - -\n\n\n # - - - - main rule block - - - -\n # at line 321:4: ( '>=' | '<=' | '==' | '!=' | '<' | '>' )\n # at line 321:4: ( '>=' | '<=' | '==' | '!=' | '<' | '>' )\n alt_1 = 6\n case look_1 = @input.peek( 1 )\n when 0x3e then look_1_1 = @input.peek( 2 )\n\n if ( look_1_1 == 0x3d )\n alt_1 = 1\n else\n alt_1 = 6\n\n end\n when 0x3c then look_1_2 = @input.peek( 2 )\n\n if ( look_1_2 == 0x3d )\n alt_1 = 2\n else\n alt_1 = 5\n\n end\n when 0x3d then alt_1 = 3\n when 0x21 then alt_1 = 4\n else\n raise NoViableAlternative( \"\", 1, 0 )\n\n end\n case alt_1\n when 1\n # at line 321:5: '>='\n match( \">=\" )\n\n\n when 2\n # at line 321:10: '<='\n match( \"<=\" )\n\n\n when 3\n # at line 321:15: '=='\n match( \"==\" )\n\n\n when 4\n # at line 321:20: '!='\n match( \"!=\" )\n\n\n when 5\n # at line 321:25: '<'\n match( 0x3c )\n\n when 6\n # at line 321:29: '>'\n match( 0x3e )\n\n end\n\n\n @state.type = type\n @state.channel = channel\n ensure\n # -> uncomment the next line to manually enable rule tracing\n # trace_out( __method__, 17 )\n\n\n end",
"def test_two_proper_args\n assert check_args([1, 1])\n end",
"def test_check_let\n RPN rpn = RPN.new(1)\n let = 'let'\n assert_equal let, 'let'\n assert rpn.check_let(let)\n end",
"def valid_triangle?(a, b, c)\n # Your code goes here!\na+b > c && a+c > b && b+c >a\nend",
"def tuples_for(arg_)\n objdata_ = _get_objdata(arg_)\n objdata_ ? objdata_[2].keys : nil\n end",
"def relation(rel, &relation_definition)\n raise NotImplementedError, 'relation is not supported yet'\n end",
"def tuple_count\n target.count\n end",
"def get_relational(expr)\n\t\t[\"==\", \"!=\", \">\", \"<\", \">=\", \"<=\"].each do |ope|\n\t\t\tif expr.scan(ope).length == 1\n\t\t\t\treturn ope\n\t\t\tend\n\t\tend\n\tend",
"def one_pair?\n n_of_a_kind? 2\n end",
"def test_history_equals_versions\n way = create(:way, :with_history)\n used_way = create(:way, :with_history)\n create(:relation_member, :member => used_way)\n way_with_versions = create(:way, :with_history, :version => 4)\n\n check_history_equals_versions(way.id)\n check_history_equals_versions(used_way.id)\n check_history_equals_versions(way_with_versions.id)\n end",
"def <= other\n call_enum \"relational\", other, :lesseq\n end",
"def to_generate_tuple(&block_)\n if block_\n @gen_tuple = block_\n else\n @gen_tuple\n end\n end",
"def check_rule_ast(pt)\n # :defn format: node tag, block name, args, body_0, ..., body_n\n tag, name, args, *body = pt\n return pt if tag != :defn\n\n body.each_with_index do |n,i|\n next if i == 0 and n == s(:nil) # a block got rewritten to an empty block\n\n # Check for a common case\n if n.sexp_type == :lasgn\n return [n, \"illegal operator: '='\"]\n end\n return pt unless n.sexp_type == :call and n.length == 4\n\n # Rule format: call tag, lhs, op, rhs\n _, lhs, op, rhs = n\n\n # Check that LHS references a named collection\n lhs_name = get_qual_name(lhs)\n return [n, \"unexpected lhs format: #{lhs}\"] if lhs_name.nil?\n unless @bud_instance.tables.has_key? lhs_name.to_sym or\n @bud_instance.lattices.has_key? lhs_name.to_sym\n return [n, \"Collection does not exist: '#{lhs_name}'\"]\n end\n\n return [n, \"illegal operator: '#{op}'\"] unless [:<, :<=].include? op\n\n # Check superator invocation. A superator that begins with \"<\" is parsed\n # as a call to the binary :< operator. The right operand to :< is a :call\n # node; the LHS of the :call is the actual rule body, the :call's oper is\n # the rest of the superator (unary ~, -, +), and the RHS is empty. Note\n # that ParseTree encodes unary \"-\" and \"+\" as :-@ and :-+, respectively.\n # XXX: We don't check for illegal superators (e.g., \"<--\"). That would be\n # tricky, because they are encoded as a nested unary op in the rule body.\n if op == :<\n return n unless rhs.sexp_type == :call\n op_tail = rhs[2]\n return n unless [:~, :-@, :+@].include? op_tail\n rhs_args = rhs[3..-1]\n return n unless rhs_args.empty?\n end\n end\n\n return nil # No errors found\n end",
"def valid_token_tuple?(type, *args)\n args.flatten!\n case type\n when :amazon then valid_amazon_tokens?(*args)\n when :bitlyv1 then valid_bitlyv1_tokens?(*args)\n when :bitlyv2 then valid_bitlyv2_tokens?(*args)\n when :facebook then valid_facebook_tokens?(*args)\n when :flickr then valid_flickr_tokens?(*args)\n when :foursquare then valid_foursquare_tokens?(*args)\n when :google then valid_google_tokens?(*args)\n when :google_oauth then valid_google_oauth_tokens?(*args)\n when :linkedin then valid_linkedin_tokens?(*args)\n when :twitter then valid_twitter_tokens?(*args)\n when :yelpv1 then valid_yelpv1_tokens?(*args)\n when :yelpv2 then valid_yelpv2_tokens?(*args)\n else\n # Rails.logger.error \"Unknown token type: #{type}\"\n end\nend",
"def to_ruby_literal\n \"Alf::Relation[\" +\n tuples.collect{|t| Tools.to_ruby_literal(t)}.join(', ') + \"]\"\n end",
"def check_error(val_a, val_b, val_c)\n raise TriangleError if (val_b + val_c <= val_a) || (val_c + val_a <= val_b)\n raise TriangleError if val_a + val_b <= val_c\nend",
"def test_friendp_returns_true_if_users_are_friends\n assert users(:active).friend?(users(:passive))\n assert users(:passive).friend?(users(:active))\n end",
"def valid_triangle?(a, b, c)\n ((a+b>c) && (b+c>a) && (a+c>b))? true:false\nend",
"def good_tuples(nums)\n count = 0\n (0..nums.size - 3).each do |i|\n first, second, third = nums[i], nums[i + 1], nums[i + 2]\n next if 3 * first == first + second + third\n count += 1 if [first, second, third].include?(first ^ second ^ third)\n\n # next if nums[i] == nums[i + 1] && nums[i + 1] == nums[i + 2]\n # next if nums[i] != nums[i + 1] && nums[i + 1] != nums[i + 2] && nums[i] != nums[i + 2]\n end\n\n count\nend",
"def valid_triangle?(a, b, c)\n a + b > c && a + c > b && b + c > a\nend",
"def valid_triangle?(a, b, c)\n a + b > c && a + c > b && b + c > a\nend",
"def valid_triangle?(a, b, c)\n a + b > c && a + c > b && b + c > a\nend",
"def triplet_check(a,b,c)\n\tif a**2+b**2 == c**2\n\t\treturn true\n\telse\n\t\treturn false\n\tend\nend",
"def visit_axiom_relation_operation_binary(binary)\n generator = self.class.visit(binary)\n @name = generator.name\n @from = aliased_subquery(generator)\n generator\n end",
"def nested?\n self.resource_tuple.length > 1\n end",
"def valid_triangle?(a, b, c)\n a + b > c && b + c > a && c + a > b\nend",
"def right_triangle?(a, b, c)\n a**2 + b**2 == c**2\nend",
"def relationship(*args)\n options = args.extract_options!\n if options[:raise_on_multiple]\n rels = relationships.take(2)\n raise _(\"Multiple relationships found\") if rels.length > 1\n rels.first\n else\n relationships.first\n end\n end",
"def isTriangle(a,b,c)\n (a+b>c && a+c>b && c+b>a)? true : false\nend",
"def >= other\n call_enum \"relational\", other, :moreeq\n end",
"def isTriangle(a, b, c)\n a, b, c = [a, b, c].sort\n a + b > c\nend",
"def test_get_location_bad_input\n assert_raises 'Improper typing of location input argument in method get_location' do\n locations = LocationRelation.new\n locations.get_location 55\n end\n end",
"def insert(tuples)\n pks = tuples.map { |tuple| relation.insert(tuple) }\n relation.where(relation.primary_key => pks).to_a\n end",
"def to_ruby_literal\n \"Alf::Relation([\" + tuples.map{|t| Support.to_ruby_literal(t) }.join(', ') + \"])\"\n end",
"def evaluate\n distinct? ? distinct_relation : relation\n end",
"def with_input_tuples(tuples, &block)\n input_tuples = Array([tuples]).flatten(1).map\n return input_tuples unless block_given?\n\n input_tuples.each(&block)\n end",
"def check_permission_with_http_info(opts = {})\n if @api_client.config.debugging\n @api_client.config.logger.debug 'Calling API: PermissionApi.check_permission ...'\n end\n # resource path\n local_var_path = '/relation-tuples/check/openapi'\n\n # query parameters\n query_params = opts[:query_params] || {}\n query_params[:'namespace'] = opts[:'namespace'] if !opts[:'namespace'].nil?\n query_params[:'object'] = opts[:'object'] if !opts[:'object'].nil?\n query_params[:'relation'] = opts[:'relation'] if !opts[:'relation'].nil?\n query_params[:'subject_id'] = opts[:'subject_id'] if !opts[:'subject_id'].nil?\n query_params[:'subject_set.namespace'] = opts[:'subject_set_namespace'] if !opts[:'subject_set_namespace'].nil?\n query_params[:'subject_set.object'] = opts[:'subject_set_object'] if !opts[:'subject_set_object'].nil?\n query_params[:'subject_set.relation'] = opts[:'subject_set_relation'] if !opts[:'subject_set_relation'].nil?\n query_params[:'max-depth'] = opts[:'max_depth'] if !opts[:'max_depth'].nil?\n\n # header parameters\n header_params = opts[:header_params] || {}\n # HTTP header 'Accept' (if needed)\n header_params['Accept'] = @api_client.select_header_accept(['application/json'])\n\n # form parameters\n form_params = opts[:form_params] || {}\n\n # http body (model)\n post_body = opts[:debug_body]\n\n # return_type\n return_type = opts[:debug_return_type] || 'CheckPermissionResult'\n\n # auth_names\n auth_names = opts[:debug_auth_names] || ['oryAccessToken']\n\n new_options = opts.merge(\n :operation => :\"PermissionApi.check_permission\",\n :header_params => header_params,\n :query_params => query_params,\n :form_params => form_params,\n :body => post_body,\n :auth_names => auth_names,\n :return_type => return_type\n )\n\n data, status_code, headers = @api_client.call_api(:GET, local_var_path, new_options)\n if @api_client.config.debugging\n @api_client.config.logger.debug \"API called: PermissionApi#check_permission\\nData: #{data.inspect}\\nStatus code: #{status_code}\\nHeaders: #{headers}\"\n end\n return data, status_code, headers\n end",
"def test_two_valid_arguments\n\t\tc = Check.new\n\t\trefute c.check_arguments([1,1])\n\tend",
"def parse_relational\n add = parse_add\n if consume(:lt)\n Node.new(:binary, op: :lt, lhs: add, rhs: parse_relational)\n elsif consume(:lteq)\n Node.new(:binary, op: :lteq, lhs: add, rhs: parse_relational)\n elsif consume(:gt)\n Node.new(:binary, op: :gt, lhs: add, rhs: parse_relational)\n elsif consume(:gteq)\n Node.new(:binary, op: :gteq, lhs: add, rhs: parse_relational)\n else\n add\n end\n end",
"def test_triangle\n if @side1.eql?(@side2) && @side2.eql?(@side3)\n \"It's an equilateral\"\n elsif @side1.eql?(@side2) || @side2.eql?(@side3)\n \"It's an isosceles\"\n elsif @side1.eql?(@side2) == false && @side2.eql?(@side3) == false\n \"It's a scalene triangle\"\n else\n \"Not a triangle\"\n end\n end",
"def selecionErr(method)\n\n tnTableCreation('tuple_node_test_result') if @is_new\n\n if @unwanted_tuple_count + @missing_tuple_count == 0\n p 'no failed rows found. There is no selection error'\n return\n end\n\n if @unwanted_tuple_count > 0\n # p \"Unwanted Pk count #{unWantedPK.count()}\"\n # create unwanted_tuple_branch table\n # binding.pry\n\n # whereErrList = whereCondTest(@unWantedPK, 'U')\n where_cond_test('U')\n # joinErrList = jointypeErr(query,'U')\n end\n\n if @missing_tuple_count > 0\n # p \"Missing PK count #{missinPK.count()}\"\n # binding.pry\n # whereErrList = whereCondTest(@missingPK, 'M')\n where_cond_test('M')\n # joinErrList = jointypeErr(query,'M')\n end\n # create aggregated tuple_suspicious_nodes\n pk = @pkFullList.map { |pk| pk['alias'] }.join(',')\n query = 'create materialized view tuple_node_test_result_aggr as '\\\n \"select #{pk}, string_agg(branch_name||'-'||node_name, ',' order by node_name,branch_name) as suspicious_nodes from tuple_node_test_result group by #{pk}\"\n pp query\n DBConn.exec(query)\n\n unless @predicateTree.nil?\n suspicious_score_upd(@predicateTree.branches)\n end\n # exnorate algorithm\n # binding.pry\n @column_combinations = method.start_with?('o') ? Columns_Combination.new(@all_columns) : @all_columns\n case method\n # when 'o'\n # puts 'old exonerate algorithm'\n\n # true_query_PT_construct\n # constraint_query = constraint_predicate_construct\n # # column_combinations_construct\n # tuple_mutation_test(@missingPK, 'M', constraint_query, false)\n # tuple_mutation_test(@unWantedPK, 'U', constraint_query, false)\n when 'or'\n puts 'old exonerate algorithm with duplicate removal'\n # reset suspicious score\n query = \"update node_query_mapping set suspicious_score = 0 where type = 'f'\"\n res = DBConn.exec(query)\n\n true_query_PT_construct\n constraint_query = constraint_predicate_construct\n # allcolumns_construct()\n puts 'Missing starts'\n puts Time.now()\n tuple_mutation_test_with_dup_removal('M', constraint_query)\n puts 'Unwanted starts'\n puts Time.now()\n\n tuple_mutation_test_with_dup_removal('U', constraint_query)\n # when 'n'\n # puts 'new exonerate algorithm'\n # true_query_PT_construct\n # constraint_query = constraint_predicate_construct\n # tuple_mutation_test_reverse(@missingPK, 'M', constraint_query, false)\n # tuple_mutation_test_reverse(@unWantedPK, 'U', constraint_query, false)\n\n when 'b'\n puts 'baseline'\n else\n puts 'Unknown method'\n end\n # remove constraint_nodes in node_query_mapping\n query = \"delete from node_query_mapping where test_id = #{@test_id} and type = 't'\"\n DBConn.exec(query)\n\n # j = Hash.new\n # # j['JoinErr'] = joinErrList if joinErrList.count >0\n # j['WhereErr'] = whereErrList\n # j\n # whereErrList\n end",
"def pitagoric_triplet?(a,b,c)\n (a**2 + b**2) == c**2\nend",
"def arel_bt_intersect(*bt_tuple)\n arel_vt_intersect(bt_tuple.at(0), bt_tuple.at(1)).and(arel_tt_intersect(bt_tuple.at(2), bt_tuple.at(3)))\n end",
"def isTriangle(a,b,c)\n a+b>c && a + c > b && b + c > a ? true : false\nend"
] |
[
"0.70170575",
"0.6604974",
"0.6361968",
"0.6199828",
"0.601983",
"0.5791429",
"0.56987613",
"0.5596251",
"0.5501244",
"0.5362164",
"0.53187114",
"0.52338684",
"0.5177114",
"0.5165867",
"0.5137209",
"0.5097905",
"0.5097905",
"0.5094715",
"0.5081012",
"0.5077448",
"0.50573194",
"0.5047579",
"0.50442135",
"0.50223505",
"0.49971905",
"0.4937294",
"0.4928462",
"0.49154636",
"0.49083236",
"0.486156",
"0.4825083",
"0.47883636",
"0.47573668",
"0.47520363",
"0.4731951",
"0.4689812",
"0.46817315",
"0.46669012",
"0.46462342",
"0.4623728",
"0.4622148",
"0.4607798",
"0.4602458",
"0.4598743",
"0.4588246",
"0.45754355",
"0.45736194",
"0.45631528",
"0.45610237",
"0.45553726",
"0.4551607",
"0.45367697",
"0.45349476",
"0.45326313",
"0.45260912",
"0.4520512",
"0.4519029",
"0.45056605",
"0.4499548",
"0.44970715",
"0.4458524",
"0.44575095",
"0.445076",
"0.44476923",
"0.44457793",
"0.44411728",
"0.44283494",
"0.4425244",
"0.44153854",
"0.44136786",
"0.441124",
"0.44048145",
"0.4403664",
"0.44030887",
"0.4402576",
"0.44009736",
"0.43985027",
"0.43985027",
"0.43985027",
"0.4397282",
"0.4395862",
"0.4389849",
"0.43888873",
"0.4387113",
"0.43837047",
"0.43751907",
"0.4370491",
"0.4366688",
"0.43657264",
"0.43639547",
"0.4358688",
"0.4352702",
"0.4347642",
"0.4346974",
"0.4343862",
"0.43400222",
"0.4339471",
"0.43389374",
"0.43388397",
"0.43324643",
"0.43324327"
] |
0.0
|
-1
|
Expand a Relation Tuple Use this endpoint to expand a relation tuple.
|
def get_expand(namespace, object, relation, max_depth, opts = {})
data, _status_code, _headers = get_expand_with_http_info(namespace, object, relation, max_depth, opts)
data
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def patch_relation_tuples(opts = {})\n patch_relation_tuples_with_http_info(opts)\n nil\n end",
"def create_relation_tuple_with_http_info(opts = {})\n if @api_client.config.debugging\n @api_client.config.logger.debug 'Calling API: WriteApi.create_relation_tuple ...'\n end\n # resource path\n local_var_path = '/relation-tuples'\n\n # query parameters\n query_params = opts[:query_params] || {}\n\n # header parameters\n header_params = opts[:header_params] || {}\n # HTTP header 'Accept' (if needed)\n header_params['Accept'] = @api_client.select_header_accept(['application/json'])\n # HTTP header 'Content-Type'\n header_params['Content-Type'] = @api_client.select_header_content_type(['application/json'])\n\n # form parameters\n form_params = opts[:form_params] || {}\n\n # http body (model)\n post_body = opts[:debug_body] || @api_client.object_to_http_body(opts[:'payload'])\n\n # return_type\n return_type = opts[:debug_return_type] || 'RelationQuery'\n\n # auth_names\n auth_names = opts[:debug_auth_names] || []\n\n new_options = opts.merge(\n :operation => :\"WriteApi.create_relation_tuple\",\n :header_params => header_params,\n :query_params => query_params,\n :form_params => form_params,\n :body => post_body,\n :auth_names => auth_names,\n :return_type => return_type\n )\n\n data, status_code, headers = @api_client.call_api(:PUT, local_var_path, new_options)\n if @api_client.config.debugging\n @api_client.config.logger.debug \"API called: WriteApi#create_relation_tuple\\nData: #{data.inspect}\\nStatus code: #{status_code}\\nHeaders: #{headers}\"\n end\n return data, status_code, headers\n end",
"def create_relation_tuple(opts = {})\n data, _status_code, _headers = create_relation_tuple_with_http_info(opts)\n data\n end",
"def expand_permissions_with_http_info(namespace, object, relation, opts = {})\n if @api_client.config.debugging\n @api_client.config.logger.debug 'Calling API: PermissionApi.expand_permissions ...'\n end\n # verify the required parameter 'namespace' is set\n if @api_client.config.client_side_validation && namespace.nil?\n fail ArgumentError, \"Missing the required parameter 'namespace' when calling PermissionApi.expand_permissions\"\n end\n # verify the required parameter 'object' is set\n if @api_client.config.client_side_validation && object.nil?\n fail ArgumentError, \"Missing the required parameter 'object' when calling PermissionApi.expand_permissions\"\n end\n # verify the required parameter 'relation' is set\n if @api_client.config.client_side_validation && relation.nil?\n fail ArgumentError, \"Missing the required parameter 'relation' when calling PermissionApi.expand_permissions\"\n end\n # resource path\n local_var_path = '/relation-tuples/expand'\n\n # query parameters\n query_params = opts[:query_params] || {}\n query_params[:'namespace'] = namespace\n query_params[:'object'] = object\n query_params[:'relation'] = relation\n query_params[:'max-depth'] = opts[:'max_depth'] if !opts[:'max_depth'].nil?\n\n # header parameters\n header_params = opts[:header_params] || {}\n # HTTP header 'Accept' (if needed)\n header_params['Accept'] = @api_client.select_header_accept(['application/json'])\n\n # form parameters\n form_params = opts[:form_params] || {}\n\n # http body (model)\n post_body = opts[:debug_body]\n\n # return_type\n return_type = opts[:debug_return_type] || 'ExpandedPermissionTree'\n\n # auth_names\n auth_names = opts[:debug_auth_names] || ['oryAccessToken']\n\n new_options = opts.merge(\n :operation => :\"PermissionApi.expand_permissions\",\n :header_params => header_params,\n :query_params => query_params,\n :form_params => form_params,\n :body => post_body,\n :auth_names => auth_names,\n :return_type => return_type\n )\n\n data, status_code, headers = @api_client.call_api(:GET, local_var_path, new_options)\n if @api_client.config.debugging\n @api_client.config.logger.debug \"API called: PermissionApi#expand_permissions\\nData: #{data.inspect}\\nStatus code: #{status_code}\\nHeaders: #{headers}\"\n end\n return data, status_code, headers\n end",
"def delete_relation_tuple_with_http_info(namespace, object, relation, opts = {})\n if @api_client.config.debugging\n @api_client.config.logger.debug 'Calling API: WriteApi.delete_relation_tuple ...'\n end\n # verify the required parameter 'namespace' is set\n if @api_client.config.client_side_validation && namespace.nil?\n fail ArgumentError, \"Missing the required parameter 'namespace' when calling WriteApi.delete_relation_tuple\"\n end\n # verify the required parameter 'object' is set\n if @api_client.config.client_side_validation && object.nil?\n fail ArgumentError, \"Missing the required parameter 'object' when calling WriteApi.delete_relation_tuple\"\n end\n # verify the required parameter 'relation' is set\n if @api_client.config.client_side_validation && relation.nil?\n fail ArgumentError, \"Missing the required parameter 'relation' when calling WriteApi.delete_relation_tuple\"\n end\n # resource path\n local_var_path = '/relation-tuples'\n\n # query parameters\n query_params = opts[:query_params] || {}\n query_params[:'namespace'] = namespace\n query_params[:'object'] = object\n query_params[:'relation'] = relation\n query_params[:'subject_id'] = opts[:'subject_id'] if !opts[:'subject_id'].nil?\n query_params[:'subject_set.namespace'] = opts[:'subject_set_namespace'] if !opts[:'subject_set_namespace'].nil?\n query_params[:'subject_set.object'] = opts[:'subject_set_object'] if !opts[:'subject_set_object'].nil?\n query_params[:'subject_set.relation'] = opts[:'subject_set_relation'] if !opts[:'subject_set_relation'].nil?\n\n # header parameters\n header_params = opts[:header_params] || {}\n # HTTP header 'Accept' (if needed)\n header_params['Accept'] = @api_client.select_header_accept(['application/json'])\n\n # form parameters\n form_params = opts[:form_params] || {}\n\n # http body (model)\n post_body = opts[:debug_body]\n\n # return_type\n return_type = opts[:debug_return_type]\n\n # auth_names\n auth_names = opts[:debug_auth_names] || []\n\n new_options = opts.merge(\n :operation => :\"WriteApi.delete_relation_tuple\",\n :header_params => header_params,\n :query_params => query_params,\n :form_params => form_params,\n :body => post_body,\n :auth_names => auth_names,\n :return_type => return_type\n )\n\n data, status_code, headers = @api_client.call_api(:DELETE, local_var_path, new_options)\n if @api_client.config.debugging\n @api_client.config.logger.debug \"API called: WriteApi#delete_relation_tuple\\nData: #{data.inspect}\\nStatus code: #{status_code}\\nHeaders: #{headers}\"\n end\n return data, status_code, headers\n end",
"def get_relation_tuples_with_http_info(namespace, opts = {})\n if @api_client.config.debugging\n @api_client.config.logger.debug 'Calling API: ReadApi.get_relation_tuples ...'\n end\n # verify the required parameter 'namespace' is set\n if @api_client.config.client_side_validation && namespace.nil?\n fail ArgumentError, \"Missing the required parameter 'namespace' when calling ReadApi.get_relation_tuples\"\n end\n # resource path\n local_var_path = '/relation-tuples'\n\n # query parameters\n query_params = opts[:query_params] || {}\n query_params[:'namespace'] = namespace\n query_params[:'page_token'] = opts[:'page_token'] if !opts[:'page_token'].nil?\n query_params[:'page_size'] = opts[:'page_size'] if !opts[:'page_size'].nil?\n query_params[:'object'] = opts[:'object'] if !opts[:'object'].nil?\n query_params[:'relation'] = opts[:'relation'] if !opts[:'relation'].nil?\n query_params[:'subject_id'] = opts[:'subject_id'] if !opts[:'subject_id'].nil?\n query_params[:'subject_set.namespace'] = opts[:'subject_set_namespace'] if !opts[:'subject_set_namespace'].nil?\n query_params[:'subject_set.object'] = opts[:'subject_set_object'] if !opts[:'subject_set_object'].nil?\n query_params[:'subject_set.relation'] = opts[:'subject_set_relation'] if !opts[:'subject_set_relation'].nil?\n\n # header parameters\n header_params = opts[:header_params] || {}\n # HTTP header 'Accept' (if needed)\n header_params['Accept'] = @api_client.select_header_accept(['application/json'])\n\n # form parameters\n form_params = opts[:form_params] || {}\n\n # http body (model)\n post_body = opts[:debug_body]\n\n # return_type\n return_type = opts[:debug_return_type] || 'GetRelationTuplesResponse'\n\n # auth_names\n auth_names = opts[:debug_auth_names] || []\n\n new_options = opts.merge(\n :operation => :\"ReadApi.get_relation_tuples\",\n :header_params => header_params,\n :query_params => query_params,\n :form_params => form_params,\n :body => post_body,\n :auth_names => auth_names,\n :return_type => return_type\n )\n\n data, status_code, headers = @api_client.call_api(:GET, local_var_path, new_options)\n if @api_client.config.debugging\n @api_client.config.logger.debug \"API called: ReadApi#get_relation_tuples\\nData: #{data.inspect}\\nStatus code: #{status_code}\\nHeaders: #{headers}\"\n end\n return data, status_code, headers\n end",
"def extend_relation_instance(relation)\n relation\n end",
"def patch_relation_tuples_with_http_info(opts = {})\n if @api_client.config.debugging\n @api_client.config.logger.debug 'Calling API: WriteApi.patch_relation_tuples ...'\n end\n # resource path\n local_var_path = '/relation-tuples'\n\n # query parameters\n query_params = opts[:query_params] || {}\n\n # header parameters\n header_params = opts[:header_params] || {}\n # HTTP header 'Accept' (if needed)\n header_params['Accept'] = @api_client.select_header_accept(['application/json'])\n # HTTP header 'Content-Type'\n header_params['Content-Type'] = @api_client.select_header_content_type(['application/json'])\n\n # form parameters\n form_params = opts[:form_params] || {}\n\n # http body (model)\n post_body = opts[:debug_body] || @api_client.object_to_http_body(opts[:'payload'])\n\n # return_type\n return_type = opts[:debug_return_type]\n\n # auth_names\n auth_names = opts[:debug_auth_names] || []\n\n new_options = opts.merge(\n :operation => :\"WriteApi.patch_relation_tuples\",\n :header_params => header_params,\n :query_params => query_params,\n :form_params => form_params,\n :body => post_body,\n :auth_names => auth_names,\n :return_type => return_type\n )\n\n data, status_code, headers = @api_client.call_api(:PATCH, local_var_path, new_options)\n if @api_client.config.debugging\n @api_client.config.logger.debug \"API called: WriteApi#patch_relation_tuples\\nData: #{data.inspect}\\nStatus code: #{status_code}\\nHeaders: #{headers}\"\n end\n return data, status_code, headers\n end",
"def call(relation)\n relation\n end",
"def update_relation(field_name, *obj)\n # Should be an Array of RDF::Term objects\n return unless obj\n\n obj.map! { |item| item.is_a?(RDF::URI) ? Ladder::Resource.from_uri(item) : item }\n relation = send(field_name)\n\n if Mongoid::Relations::Targets::Enumerable == relation.class\n obj.map { |item| relation.send(:push, item) unless relation.include? item }\n else\n send(\"#{field_name}=\", obj.size > 1 ? obj : obj.first)\n end\n end",
"def relation(rel, &relation_definition)\n raise NotImplementedError, 'relation is not supported yet'\n end",
"def expand(rel)\n return rel unless rel && templated?\n\n href&.gsub('{rel}', rel)\n end",
"def convert_path_relation(relation)\n relation.map{|r| \"<\" + Xplain::Namespace.expand_uri(r.id) + \">\"}.join(\"/\")\n end",
"def tuples\n relation = self.relation\n return relation if materialized?\n\n adapter.read(relation)\n end",
"def relation(rel, &relation_definition)\n RelationContext.new(@resource, rel, &relation_definition)\n end",
"def new_word_relation(sax, author_id)\n WordRelation.new.tap do |relation|\n relation.author_id = author_id\n relation.word1_id = sax.word1_id\n relation.word2_id = sax.word2_id\n end\nend",
"def expand(*props)\n @query[:expand] = props.join(',')\n self\n end",
"def delete_relation_tuple(namespace, object, relation, opts = {})\n delete_relation_tuple_with_http_info(namespace, object, relation, opts)\n nil\n end",
"def expand_permissions(namespace, object, relation, opts = {})\n data, _status_code, _headers = expand_permissions_with_http_info(namespace, object, relation, opts)\n data\n end",
"def relation\n relation = nodes.reduce(root) do |a, e|\n a.associations[e.name.key].join(:join, a, e)\n end\n schema.(relation)\n end",
"def relation(*fields)\n inline_keys = fields.extract_options!\n if api_response?\n fields.each do |column|\n raise ArgumentError, \"You specify relation with `_id` on the end.\" if column.end_with?(\"_id\")\n\n property column, inline_keys do\n key :type, \"object\"\n key :readOnly, true\n property \"id\", type: \"integer\"\n property \"name\", type: \"string\"\n end\n end\n else\n fields.each do |column|\n property \"#{column}_id\", inline_keys do\n key :type, \"integer\"\n end\n end\n end\n end",
"def relation(*args, &block)\n boot.relation(*args, &block)\n end",
"def forward(method, args, &block)\n result = @relation.send(method, *args, &block)\n\n unless result.kind_of?(::Veritas::Relation)\n return result\n end\n\n self.class.new(@mapper, result)\n end",
"def new_word_relation(sax, author_id)\n AntonomyRelation.new.tap do |relation|\n relation.author_id = author_id\n relation.synset1_id = sax.synset1_id\n relation.synset2_id = sax.synset2_id\n relation.word1_id = sax.word1_id\n relation.word2_id = sax.word2_id\n end\nend",
"def rel(rels)\n # rel must be an array.\n data[:rel] = Array(rels)\n end",
"def update!(**args)\n @relation = args[:relation] if args.key?(:relation)\n end",
"def op_add_relation!(field, objects = [])\n objects = [objects] unless objects.is_a?(Array)\n return false if objects.empty?\n relation_action = Parse::RelationAction.new(field, polarity: true, objects: objects)\n operate_field! field, relation_action\n end",
"def relation(related, arguments=[])\n self.send(related.to_sym)\n end",
"def relation(*args)\n source_node.join(target_node, *args)\n end",
"def related_to(*args)\n@relations = args.map(&:to_s)\nend",
"def visit_axiom_relation_base(base_relation)\n @name = base_relation.name.to_s.freeze\n @from = visit_identifier(@name)\n @header = base_relation.header\n @columns = columns_for(base_relation)\n self\n end",
"def get_relation_tuples(namespace, opts = {})\n data, _status_code, _headers = get_relation_tuples_with_http_info(namespace, opts)\n data\n end",
"def visit_axiom_relation_operation_binary(binary)\n generator = self.class.visit(binary)\n @name = generator.name\n @from = aliased_subquery(generator)\n generator\n end",
"def visit_ActiveRecord_Relation(o, parent)\n o.arel\n end",
"def read_join_relations(obj, res_row, row, join_relations)\n offset = obj.class.serializable_attributes.size\n\n for rel in join_relations\n rel_obj = rel[:target_class].og_allocate(res_row, row)\n rel_obj.og_read(res_row, row, offset)\n offset += rel_obj.class.serializable_attributes.size\n obj.instance_variable_set(\"@#{rel[:name]}\", rel_obj)\n end\n end",
"def new_relation(rel_name, internal_relation)\n relations_info[rel_name.to_sym][:relation].new(internal_relation) # internal_relation is a java neo object\n end",
"def resolve!\n # default to base relation if not specified\n @relation_type ||= RelationTypes::BASE\n @attributes.map(&:resolve!)\n end",
"def add_relationship(rel_attr); end",
"def call(*args)\n ((combine? || composite?) ? relation : (relation >> mapper)).call(*args)\n end",
"def read_relation_list(node)\n target_type = Utils.add_namespace(node.attributes['target-type'], Model::NS_REL_1)\n node.elements.each('relation') {|child|\n yield create_relation(child, target_type)\n }\n end",
"def to_ruby_literal\n \"Alf::Relation[\" +\n tuples.collect{|t| Tools.to_ruby_literal(t)}.join(', ') + \"]\"\n end",
"def set_relation\n @relation = Relation.find(params[:id])\n end",
"def set_relation\n @relation = Relation.find(params[:id])\n end",
"def gateway(relation)\n Axiom::Relation::Gateway.new(self, relation)\n end",
"def each_relation\n\t for rel in relations\n\t\tyield(rel)\n\t end\n\tend",
"def relation_method\n :product\n end",
"def relation\n raise NotImplementedError\n end",
"def _expand(uri_variables = {})\n self.class.new(@key, @link, @entry_point, uri_variables)\n end",
"def get_relation(args)\n\tapi_url = \"#{@base_url}/#{args[:collection]}/#{args[:key]}/relations/#{args[:relation]}\"\n\tdo_the_get_call( url: api_url, user: @user )\nend",
"def set_rel_props(rel, hash)\n @batch_inserter.set_relationship_properties(rel, hash)\n end",
"def <<(values)\n values = prepare_relation(values) if values.is_a?(Relation)\n self.set(objects.to_a | Array.wrap(values))\n end",
"def relations\n @relations ||= process_rels\n end",
"def set_relation\n @relation = Relation.find(params[:id])\n end",
"def relation\n @relation ||= begin\n additions = false\n # TODO: merge! has unexpected results.\n frag = include_params.reduce(model.all) do |result, inclusion|\n config = config_for(inclusion)\n query = config.respond_to?(:keys) ? config.values.first : nil\n unless query.nil?\n additions = true\n result = result.merge(query)\n end\n result\n end\n additions ? frag : nil\n end\n end",
"def expand_url(pairs)\n url.expand Hash[pairs]\n end",
"def create_tuple!(attributes)\n # split the attrs (needs DRYing)\n rel_attrs = {}\n attributes.each do |key, value|\n if key.to_s[0..3] == 'rel_'\n rel_attrs[key] = attributes.delete(key)\n end\n end\n\n # find or create the one node\n if target_node = @target_class.find(attributes[:id])\n target_node.update_attributes!(attributes)\n else\n target_node = @target_class.create!(attributes)\n end\n\n self.relate(target_node, rel_attrs)\n\n # finally, make and return the tuple\n #Tuple.new(@node, target_node)\n #Tuple.new(@node, target_node)\n\n end",
"def coerce_to_relation(summarize_with)\n if summarize_with.kind_of?(Relation)\n summarize_with\n else\n project(Relation::Header.coerce(summarize_with))\n end\n end",
"def expand(*args)\n\t\t\t#\tjust the default value, not the Response\n\t\t\t(super *args).to_s\n\t\tend",
"def adding_plan_relation(transaction, parent, child, relations, info); end",
"def expand(binding)\n begin\n context.push(binding)\n render\n ensure\n context.pop\n end\n end",
"def generate_person_and_task_relations(relations)\n relations.each_pair { |person_id, values|\n values.each { |task_id|\n @db_base.map_task_to_person(person_id, task_id)\n }\n }\n end",
"def expand\n map { |p| p&.flatten || p }.flatten\n end",
"def expand_more(more)\n response = client.get(\n \"/api/morechildren\",\n children: more.join(\",\"),\n link_id: fullname\n )\n\n client.object_from_body(\n kind: \"Listing\",\n data: {\n before: \"\", after: \"\",\n children: response.body[:json][:data][:things]\n }\n )\n end",
"def relationship_aux(action, opts)\n\n if action != \"destroy\"\n\n data = LibXML::XML::Parser.string(request.raw_post).parse\n\n subject = parse_element(data, :resource, '/relationship/subject')\n predicate = parse_element(data, :resource, '/relationship/predicate')\n objekt = parse_element(data, :resource, '/relationship/object')\n context = parse_element(data, :resource, '/relationship/context')\n end\n\n # Obtain object\n\n case action\n when 'create';\n return rest_response(401, :reason => \"Not authorised to create a relationship\") unless Authorization.check('create', Relationship, opts[:user], context)\n ob = Relationship.new(:user => opts[:user])\n when 'view', 'edit', 'destroy';\n ob, error = obtain_rest_resource('Relationship', opts[:query]['id'], opts[:query]['version'], opts[:user], action)\n else\n raise \"Invalid action '#{action}'\"\n end\n\n return error if ob.nil? # appropriate rest response already given\n\n if action == \"destroy\"\n\n ob.destroy\n\n else\n\n # build it\n\n ob.subject = subject if subject\n ob.predicate = predicate if predicate\n ob.objekt = objekt if objekt\n ob.context = context if context\n\n if not ob.save\n return rest_response(400, :object => ob)\n end\n end\n\n rest_get_request(ob, opts[:user], { \"id\" => ob.id.to_s })\nend",
"def <<(relation)\n new_node(relation.name, relation)\n end",
"def to_ruby_literal\n \"Alf::Relation([\" + tuples.map{|t| Support.to_ruby_literal(t) }.join(', ') + \"])\"\n end",
"def relation_params\n _relation_params\n end",
"def update!(**args)\n @relation_descriptor = args[:relation_descriptor] if args.key?(:relation_descriptor)\n end",
"def method_missing(method, *args, &block)\n relation.send(method, *args, &block)\n end",
"def show\n\t\t@relation = Relation.find(params[:id])\n\t\trespond_to do |format|\n\t\t\tformat.html # show.html.erb\n\t\t\tformat.xml { render :xml => @relation }\n\t\tend\n\tend",
"def add_rel(predicate, object, required = false)\n object = check_objects(object)\n if(!object)\n raise(ArgumentError, \"Relation with empty object on #{predicate} (#{@current.attributes['uri']}).\") if(required)\n return\n end\n if(object.kind_of?(Array))\n object.each do |obj| \n raise(ArgumentError, \"Cannot add relation on database field <#{predicate}> - <#{object.inspect}>\") if(ActiveSource.db_attr?(predicate))\n set_element(predicate, \"<#{irify(obj)}>\", required) \n end\n else\n raise(ArgumentError, \"Cannot add relation on database field\") if(ActiveSource.db_attr?(predicate))\n set_element(predicate, \"<#{irify(object)}>\", required)\n end\n end",
"def relation\n super.extending!(TranslatedAttributesQuery)\n end",
"def rdfa_relation(name,tag,relation,options={})\n property_value = relation\n property_uri = name\n property_datatype = options[:datatype]\n if(name.instance_of?(Symbol))\n key = self.class.resource_mapping.keys.detect do |k|\n k == name\n end\n raise Exception.new(\"Unknown rdfa property #{name} for object #{self}\") if key.nil?\n property_uri = \"#{SemanticResource::Configuration.namespaces_map[self.class.resource_mapping[key].first]}#{self.class.resource_mapping[key].last}\"\n end\n options.delete(:datatype)\n\n rdfa = StringIO.new\n\n rdfa << \"<#{tag} rel='#{property_uri}'\"\n rdfa << \" href='#{property_value}'\"\n\n options.each_pair do |k,v|\n rdfa << \"#{k}='#{v}' \"\n end\n if(block_given?)\n rdfa << \">\"\n rdfa << yield if block_given?\n rdfa << \"</#{tag}>\"\n else\n rdfa << \" />\"\n end\n\n rdfa.string\n end",
"def show\n\t\tshow_\n\t\trespond_to do |format|\n\t\t\tformat.html # show.html.erb\n\t\t\tformat.xml { render :xml => @relation }\n\t\tend\n\tend",
"def update(tuple)\n hash = relvar_def.__prepare_for_update(self, tuple)\n underlying_table.update(relvar_def.__physical_tuple_encode(self, hash))\n end",
"def expand_routes(*slice)\n expanded = parent_resource.expand_routes(route_variables)\n\n unless slice.empty?\n expanded = expanded.send(:slice, *slice)\n end\n\n expanded.transform_keys do |key|\n \"#{ key }_url\"\n end\n end",
"def lift_tuple(tuple)\n if Pione::TupleSpace[tuple.first]\n if pos = Pione::TupleSpace[tuple.first].location_position\n if new_location = lift_location(tuple[pos])\n tuple = tuple.clone\n tuple[pos] = new_location\n end\n end\n end\n return tuple\n end",
"def expand!\n if attributes.nil?\n @attributes = {}\n if @parent\n @parent.expand!\n\n @parent.send(:\"_#{@path.last}=\", self)\n end\n end\n end",
"def expand\n inst = model.new(api)\n inst.data[:id] = self.id\n inst.reload\n end",
"def add_relation source_label, relation, destination_label, direction = '<->'\n # More popular nodes have more weigth\n @node_weights[source_label] += 1\n @node_weights[destination_label] += 1\n\n @edge_buffer << Edge.new(\n source_label: source_label,\n relation: relation,\n destination_label: destination_label,\n direction: direction\n )\n end",
"def related_attrs\n relationship = flex_options[:relationship_name]\n send relationship\n end",
"def decorate_relation(obj)\n if obj.respond_to?(:records)\n # Rails 5.0\n obj.extend ActiveDecorator::RelationDecorator unless ActiveDecorator::RelationDecorator === obj\n else\n # Rails 3.x and 4.x\n obj.extend ActiveDecorator::RelationDecoratorLegacy unless ActiveDecorator::RelationDecoratorLegacy === obj\n end\n obj\n end",
"def populate_relation(args)\n\targs[:keys_B].each do |key_B|\n\t\tapi_url = \"#{@base_url}/#{args[:coll_A]}/#{args[:key_A]}/relation/\" +\n\t\t \"#{args[:relation]}/#{args[:coll_B]}/#{key_B}\"\n\t\tputs do_the_put_call({ url: api_url, user: @user, json: '{}' })\n\tend\nend",
"def rels(*rel_types)\n if rel_types.empty?\n AllRelsDsl.new(@_relationships, _java_node)\n else\n storage = _create_or_get_storage(rel_types.first)\n RelsDSL.new(storage)\n end\n end",
"def set_relation(name, relation)\n instance_variable_set(\"@_#{name}\", relation)\n end",
"def relationship(rel_class)\n @relationship = rel_class\n self\n end",
"def relation(relation)\n true\n end",
"def relation=(relation)\n if relation.nil?\n fail ArgumentError, 'invalid value for \"relation\", relation cannot be nil.'\n end\n @relation = relation\n end",
"def relation\n Proxy\n end",
"def relation\n Proxy\n end",
"def relation\n Proxy\n end",
"def resolve!\n @relations.map(&:resolve!)\n end",
"def relations\n Relations::RelationTraverser.new(@internal_node)\n end",
"def relation_params\n params.require(:relation).permit(:user_id, :friend_id)\n end",
"def relationship(rel_class)\n @relationship = rel_class\n self\n end",
"def get_relation(id)\n get_object('relation', id)\n end",
"def get_expand_with_http_info(namespace, object, relation, max_depth, opts = {})\n if @api_client.config.debugging\n @api_client.config.logger.debug 'Calling API: ReadApi.get_expand ...'\n end\n # verify the required parameter 'namespace' is set\n if @api_client.config.client_side_validation && namespace.nil?\n fail ArgumentError, \"Missing the required parameter 'namespace' when calling ReadApi.get_expand\"\n end\n # verify the required parameter 'object' is set\n if @api_client.config.client_side_validation && object.nil?\n fail ArgumentError, \"Missing the required parameter 'object' when calling ReadApi.get_expand\"\n end\n # verify the required parameter 'relation' is set\n if @api_client.config.client_side_validation && relation.nil?\n fail ArgumentError, \"Missing the required parameter 'relation' when calling ReadApi.get_expand\"\n end\n # verify the required parameter 'max_depth' is set\n if @api_client.config.client_side_validation && max_depth.nil?\n fail ArgumentError, \"Missing the required parameter 'max_depth' when calling ReadApi.get_expand\"\n end\n # resource path\n local_var_path = '/expand'\n\n # query parameters\n query_params = opts[:query_params] || {}\n query_params[:'namespace'] = namespace\n query_params[:'object'] = object\n query_params[:'relation'] = relation\n query_params[:'max-depth'] = max_depth\n\n # header parameters\n header_params = opts[:header_params] || {}\n # HTTP header 'Accept' (if needed)\n header_params['Accept'] = @api_client.select_header_accept(['application/json'])\n\n # form parameters\n form_params = opts[:form_params] || {}\n\n # http body (model)\n post_body = opts[:debug_body]\n\n # return_type\n return_type = opts[:debug_return_type] || 'ExpandTree'\n\n # auth_names\n auth_names = opts[:debug_auth_names] || []\n\n new_options = opts.merge(\n :operation => :\"ReadApi.get_expand\",\n :header_params => header_params,\n :query_params => query_params,\n :form_params => form_params,\n :body => post_body,\n :auth_names => auth_names,\n :return_type => return_type\n )\n\n data, status_code, headers = @api_client.call_api(:GET, local_var_path, new_options)\n if @api_client.config.debugging\n @api_client.config.logger.debug \"API called: ReadApi#get_expand\\nData: #{data.inspect}\\nStatus code: #{status_code}\\nHeaders: #{headers}\"\n end\n return data, status_code, headers\n end",
"def get_relation(name, metadata, object, reload = false)\n if !reload && (value = ivar(name)) != false\n value\n else\n _building do\n _loading do\n if object && needs_no_database_query?(object, metadata)\n __build__(name, object, metadata)\n else\n __build__(name, attributes[metadata.key], metadata)\n end\n end\n end\n end\n end",
"def relations; enum_relations.to_a end",
"def relation(relation_name)\n state_depth_must_be(States::SCHEMA)\n r = relation_lookup(relation_name)\n raise \"relation #{relation_name} already exists\" if r\n n = Relation.new(@current_schema, relation_name)\n @current_schema.relations << n\n @current_relation = n\n clear_state_below(States::RELATION)\n end"
] |
[
"0.61490965",
"0.59804386",
"0.58907497",
"0.5821089",
"0.5736419",
"0.56363285",
"0.5592303",
"0.54958576",
"0.5446701",
"0.539018",
"0.5334504",
"0.5320094",
"0.5302275",
"0.5155383",
"0.51421624",
"0.51375866",
"0.50912446",
"0.50813067",
"0.5055887",
"0.5038976",
"0.50217754",
"0.49961802",
"0.49806532",
"0.49726167",
"0.49679613",
"0.49421933",
"0.48941278",
"0.4887914",
"0.4884876",
"0.4874631",
"0.48691714",
"0.4837445",
"0.4815686",
"0.47924477",
"0.47823134",
"0.47798088",
"0.47472924",
"0.47453338",
"0.47446445",
"0.47147676",
"0.47086906",
"0.47057733",
"0.47057733",
"0.46897528",
"0.46852958",
"0.46655217",
"0.4651653",
"0.46256477",
"0.46078",
"0.46064228",
"0.4606086",
"0.4601006",
"0.45932102",
"0.45912397",
"0.45909864",
"0.45722052",
"0.4564711",
"0.45508045",
"0.45469016",
"0.45461497",
"0.45417917",
"0.454026",
"0.4539158",
"0.45363066",
"0.45305914",
"0.4528904",
"0.45128122",
"0.44858193",
"0.44804946",
"0.4469743",
"0.4463853",
"0.4463781",
"0.44557557",
"0.44544983",
"0.44494262",
"0.44423935",
"0.44272506",
"0.4417288",
"0.44001055",
"0.4396538",
"0.43860424",
"0.4384188",
"0.4378816",
"0.437295",
"0.43682742",
"0.43669412",
"0.435734",
"0.43559918",
"0.43553385",
"0.43553385",
"0.43553385",
"0.43454716",
"0.43437704",
"0.4341211",
"0.43387687",
"0.4337756",
"0.43312055",
"0.4331147",
"0.4326733",
"0.43159848"
] |
0.45721352
|
56
|
Expand a Relation Tuple Use this endpoint to expand a relation tuple.
|
def get_expand_with_http_info(namespace, object, relation, max_depth, opts = {})
if @api_client.config.debugging
@api_client.config.logger.debug 'Calling API: ReadApi.get_expand ...'
end
# verify the required parameter 'namespace' is set
if @api_client.config.client_side_validation && namespace.nil?
fail ArgumentError, "Missing the required parameter 'namespace' when calling ReadApi.get_expand"
end
# verify the required parameter 'object' is set
if @api_client.config.client_side_validation && object.nil?
fail ArgumentError, "Missing the required parameter 'object' when calling ReadApi.get_expand"
end
# verify the required parameter 'relation' is set
if @api_client.config.client_side_validation && relation.nil?
fail ArgumentError, "Missing the required parameter 'relation' when calling ReadApi.get_expand"
end
# verify the required parameter 'max_depth' is set
if @api_client.config.client_side_validation && max_depth.nil?
fail ArgumentError, "Missing the required parameter 'max_depth' when calling ReadApi.get_expand"
end
# resource path
local_var_path = '/expand'
# query parameters
query_params = opts[:query_params] || {}
query_params[:'namespace'] = namespace
query_params[:'object'] = object
query_params[:'relation'] = relation
query_params[:'max-depth'] = max_depth
# header parameters
header_params = opts[:header_params] || {}
# HTTP header 'Accept' (if needed)
header_params['Accept'] = @api_client.select_header_accept(['application/json'])
# form parameters
form_params = opts[:form_params] || {}
# http body (model)
post_body = opts[:debug_body]
# return_type
return_type = opts[:debug_return_type] || 'ExpandTree'
# auth_names
auth_names = opts[:debug_auth_names] || []
new_options = opts.merge(
:operation => :"ReadApi.get_expand",
:header_params => header_params,
:query_params => query_params,
:form_params => form_params,
:body => post_body,
:auth_names => auth_names,
:return_type => return_type
)
data, status_code, headers = @api_client.call_api(:GET, local_var_path, new_options)
if @api_client.config.debugging
@api_client.config.logger.debug "API called: ReadApi#get_expand\nData: #{data.inspect}\nStatus code: #{status_code}\nHeaders: #{headers}"
end
return data, status_code, headers
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def patch_relation_tuples(opts = {})\n patch_relation_tuples_with_http_info(opts)\n nil\n end",
"def create_relation_tuple_with_http_info(opts = {})\n if @api_client.config.debugging\n @api_client.config.logger.debug 'Calling API: WriteApi.create_relation_tuple ...'\n end\n # resource path\n local_var_path = '/relation-tuples'\n\n # query parameters\n query_params = opts[:query_params] || {}\n\n # header parameters\n header_params = opts[:header_params] || {}\n # HTTP header 'Accept' (if needed)\n header_params['Accept'] = @api_client.select_header_accept(['application/json'])\n # HTTP header 'Content-Type'\n header_params['Content-Type'] = @api_client.select_header_content_type(['application/json'])\n\n # form parameters\n form_params = opts[:form_params] || {}\n\n # http body (model)\n post_body = opts[:debug_body] || @api_client.object_to_http_body(opts[:'payload'])\n\n # return_type\n return_type = opts[:debug_return_type] || 'RelationQuery'\n\n # auth_names\n auth_names = opts[:debug_auth_names] || []\n\n new_options = opts.merge(\n :operation => :\"WriteApi.create_relation_tuple\",\n :header_params => header_params,\n :query_params => query_params,\n :form_params => form_params,\n :body => post_body,\n :auth_names => auth_names,\n :return_type => return_type\n )\n\n data, status_code, headers = @api_client.call_api(:PUT, local_var_path, new_options)\n if @api_client.config.debugging\n @api_client.config.logger.debug \"API called: WriteApi#create_relation_tuple\\nData: #{data.inspect}\\nStatus code: #{status_code}\\nHeaders: #{headers}\"\n end\n return data, status_code, headers\n end",
"def create_relation_tuple(opts = {})\n data, _status_code, _headers = create_relation_tuple_with_http_info(opts)\n data\n end",
"def expand_permissions_with_http_info(namespace, object, relation, opts = {})\n if @api_client.config.debugging\n @api_client.config.logger.debug 'Calling API: PermissionApi.expand_permissions ...'\n end\n # verify the required parameter 'namespace' is set\n if @api_client.config.client_side_validation && namespace.nil?\n fail ArgumentError, \"Missing the required parameter 'namespace' when calling PermissionApi.expand_permissions\"\n end\n # verify the required parameter 'object' is set\n if @api_client.config.client_side_validation && object.nil?\n fail ArgumentError, \"Missing the required parameter 'object' when calling PermissionApi.expand_permissions\"\n end\n # verify the required parameter 'relation' is set\n if @api_client.config.client_side_validation && relation.nil?\n fail ArgumentError, \"Missing the required parameter 'relation' when calling PermissionApi.expand_permissions\"\n end\n # resource path\n local_var_path = '/relation-tuples/expand'\n\n # query parameters\n query_params = opts[:query_params] || {}\n query_params[:'namespace'] = namespace\n query_params[:'object'] = object\n query_params[:'relation'] = relation\n query_params[:'max-depth'] = opts[:'max_depth'] if !opts[:'max_depth'].nil?\n\n # header parameters\n header_params = opts[:header_params] || {}\n # HTTP header 'Accept' (if needed)\n header_params['Accept'] = @api_client.select_header_accept(['application/json'])\n\n # form parameters\n form_params = opts[:form_params] || {}\n\n # http body (model)\n post_body = opts[:debug_body]\n\n # return_type\n return_type = opts[:debug_return_type] || 'ExpandedPermissionTree'\n\n # auth_names\n auth_names = opts[:debug_auth_names] || ['oryAccessToken']\n\n new_options = opts.merge(\n :operation => :\"PermissionApi.expand_permissions\",\n :header_params => header_params,\n :query_params => query_params,\n :form_params => form_params,\n :body => post_body,\n :auth_names => auth_names,\n :return_type => return_type\n )\n\n data, status_code, headers = @api_client.call_api(:GET, local_var_path, new_options)\n if @api_client.config.debugging\n @api_client.config.logger.debug \"API called: PermissionApi#expand_permissions\\nData: #{data.inspect}\\nStatus code: #{status_code}\\nHeaders: #{headers}\"\n end\n return data, status_code, headers\n end",
"def delete_relation_tuple_with_http_info(namespace, object, relation, opts = {})\n if @api_client.config.debugging\n @api_client.config.logger.debug 'Calling API: WriteApi.delete_relation_tuple ...'\n end\n # verify the required parameter 'namespace' is set\n if @api_client.config.client_side_validation && namespace.nil?\n fail ArgumentError, \"Missing the required parameter 'namespace' when calling WriteApi.delete_relation_tuple\"\n end\n # verify the required parameter 'object' is set\n if @api_client.config.client_side_validation && object.nil?\n fail ArgumentError, \"Missing the required parameter 'object' when calling WriteApi.delete_relation_tuple\"\n end\n # verify the required parameter 'relation' is set\n if @api_client.config.client_side_validation && relation.nil?\n fail ArgumentError, \"Missing the required parameter 'relation' when calling WriteApi.delete_relation_tuple\"\n end\n # resource path\n local_var_path = '/relation-tuples'\n\n # query parameters\n query_params = opts[:query_params] || {}\n query_params[:'namespace'] = namespace\n query_params[:'object'] = object\n query_params[:'relation'] = relation\n query_params[:'subject_id'] = opts[:'subject_id'] if !opts[:'subject_id'].nil?\n query_params[:'subject_set.namespace'] = opts[:'subject_set_namespace'] if !opts[:'subject_set_namespace'].nil?\n query_params[:'subject_set.object'] = opts[:'subject_set_object'] if !opts[:'subject_set_object'].nil?\n query_params[:'subject_set.relation'] = opts[:'subject_set_relation'] if !opts[:'subject_set_relation'].nil?\n\n # header parameters\n header_params = opts[:header_params] || {}\n # HTTP header 'Accept' (if needed)\n header_params['Accept'] = @api_client.select_header_accept(['application/json'])\n\n # form parameters\n form_params = opts[:form_params] || {}\n\n # http body (model)\n post_body = opts[:debug_body]\n\n # return_type\n return_type = opts[:debug_return_type]\n\n # auth_names\n auth_names = opts[:debug_auth_names] || []\n\n new_options = opts.merge(\n :operation => :\"WriteApi.delete_relation_tuple\",\n :header_params => header_params,\n :query_params => query_params,\n :form_params => form_params,\n :body => post_body,\n :auth_names => auth_names,\n :return_type => return_type\n )\n\n data, status_code, headers = @api_client.call_api(:DELETE, local_var_path, new_options)\n if @api_client.config.debugging\n @api_client.config.logger.debug \"API called: WriteApi#delete_relation_tuple\\nData: #{data.inspect}\\nStatus code: #{status_code}\\nHeaders: #{headers}\"\n end\n return data, status_code, headers\n end",
"def get_relation_tuples_with_http_info(namespace, opts = {})\n if @api_client.config.debugging\n @api_client.config.logger.debug 'Calling API: ReadApi.get_relation_tuples ...'\n end\n # verify the required parameter 'namespace' is set\n if @api_client.config.client_side_validation && namespace.nil?\n fail ArgumentError, \"Missing the required parameter 'namespace' when calling ReadApi.get_relation_tuples\"\n end\n # resource path\n local_var_path = '/relation-tuples'\n\n # query parameters\n query_params = opts[:query_params] || {}\n query_params[:'namespace'] = namespace\n query_params[:'page_token'] = opts[:'page_token'] if !opts[:'page_token'].nil?\n query_params[:'page_size'] = opts[:'page_size'] if !opts[:'page_size'].nil?\n query_params[:'object'] = opts[:'object'] if !opts[:'object'].nil?\n query_params[:'relation'] = opts[:'relation'] if !opts[:'relation'].nil?\n query_params[:'subject_id'] = opts[:'subject_id'] if !opts[:'subject_id'].nil?\n query_params[:'subject_set.namespace'] = opts[:'subject_set_namespace'] if !opts[:'subject_set_namespace'].nil?\n query_params[:'subject_set.object'] = opts[:'subject_set_object'] if !opts[:'subject_set_object'].nil?\n query_params[:'subject_set.relation'] = opts[:'subject_set_relation'] if !opts[:'subject_set_relation'].nil?\n\n # header parameters\n header_params = opts[:header_params] || {}\n # HTTP header 'Accept' (if needed)\n header_params['Accept'] = @api_client.select_header_accept(['application/json'])\n\n # form parameters\n form_params = opts[:form_params] || {}\n\n # http body (model)\n post_body = opts[:debug_body]\n\n # return_type\n return_type = opts[:debug_return_type] || 'GetRelationTuplesResponse'\n\n # auth_names\n auth_names = opts[:debug_auth_names] || []\n\n new_options = opts.merge(\n :operation => :\"ReadApi.get_relation_tuples\",\n :header_params => header_params,\n :query_params => query_params,\n :form_params => form_params,\n :body => post_body,\n :auth_names => auth_names,\n :return_type => return_type\n )\n\n data, status_code, headers = @api_client.call_api(:GET, local_var_path, new_options)\n if @api_client.config.debugging\n @api_client.config.logger.debug \"API called: ReadApi#get_relation_tuples\\nData: #{data.inspect}\\nStatus code: #{status_code}\\nHeaders: #{headers}\"\n end\n return data, status_code, headers\n end",
"def extend_relation_instance(relation)\n relation\n end",
"def patch_relation_tuples_with_http_info(opts = {})\n if @api_client.config.debugging\n @api_client.config.logger.debug 'Calling API: WriteApi.patch_relation_tuples ...'\n end\n # resource path\n local_var_path = '/relation-tuples'\n\n # query parameters\n query_params = opts[:query_params] || {}\n\n # header parameters\n header_params = opts[:header_params] || {}\n # HTTP header 'Accept' (if needed)\n header_params['Accept'] = @api_client.select_header_accept(['application/json'])\n # HTTP header 'Content-Type'\n header_params['Content-Type'] = @api_client.select_header_content_type(['application/json'])\n\n # form parameters\n form_params = opts[:form_params] || {}\n\n # http body (model)\n post_body = opts[:debug_body] || @api_client.object_to_http_body(opts[:'payload'])\n\n # return_type\n return_type = opts[:debug_return_type]\n\n # auth_names\n auth_names = opts[:debug_auth_names] || []\n\n new_options = opts.merge(\n :operation => :\"WriteApi.patch_relation_tuples\",\n :header_params => header_params,\n :query_params => query_params,\n :form_params => form_params,\n :body => post_body,\n :auth_names => auth_names,\n :return_type => return_type\n )\n\n data, status_code, headers = @api_client.call_api(:PATCH, local_var_path, new_options)\n if @api_client.config.debugging\n @api_client.config.logger.debug \"API called: WriteApi#patch_relation_tuples\\nData: #{data.inspect}\\nStatus code: #{status_code}\\nHeaders: #{headers}\"\n end\n return data, status_code, headers\n end",
"def call(relation)\n relation\n end",
"def update_relation(field_name, *obj)\n # Should be an Array of RDF::Term objects\n return unless obj\n\n obj.map! { |item| item.is_a?(RDF::URI) ? Ladder::Resource.from_uri(item) : item }\n relation = send(field_name)\n\n if Mongoid::Relations::Targets::Enumerable == relation.class\n obj.map { |item| relation.send(:push, item) unless relation.include? item }\n else\n send(\"#{field_name}=\", obj.size > 1 ? obj : obj.first)\n end\n end",
"def relation(rel, &relation_definition)\n raise NotImplementedError, 'relation is not supported yet'\n end",
"def expand(rel)\n return rel unless rel && templated?\n\n href&.gsub('{rel}', rel)\n end",
"def convert_path_relation(relation)\n relation.map{|r| \"<\" + Xplain::Namespace.expand_uri(r.id) + \">\"}.join(\"/\")\n end",
"def tuples\n relation = self.relation\n return relation if materialized?\n\n adapter.read(relation)\n end",
"def relation(rel, &relation_definition)\n RelationContext.new(@resource, rel, &relation_definition)\n end",
"def new_word_relation(sax, author_id)\n WordRelation.new.tap do |relation|\n relation.author_id = author_id\n relation.word1_id = sax.word1_id\n relation.word2_id = sax.word2_id\n end\nend",
"def expand(*props)\n @query[:expand] = props.join(',')\n self\n end",
"def delete_relation_tuple(namespace, object, relation, opts = {})\n delete_relation_tuple_with_http_info(namespace, object, relation, opts)\n nil\n end",
"def expand_permissions(namespace, object, relation, opts = {})\n data, _status_code, _headers = expand_permissions_with_http_info(namespace, object, relation, opts)\n data\n end",
"def relation\n relation = nodes.reduce(root) do |a, e|\n a.associations[e.name.key].join(:join, a, e)\n end\n schema.(relation)\n end",
"def relation(*fields)\n inline_keys = fields.extract_options!\n if api_response?\n fields.each do |column|\n raise ArgumentError, \"You specify relation with `_id` on the end.\" if column.end_with?(\"_id\")\n\n property column, inline_keys do\n key :type, \"object\"\n key :readOnly, true\n property \"id\", type: \"integer\"\n property \"name\", type: \"string\"\n end\n end\n else\n fields.each do |column|\n property \"#{column}_id\", inline_keys do\n key :type, \"integer\"\n end\n end\n end\n end",
"def relation(*args, &block)\n boot.relation(*args, &block)\n end",
"def forward(method, args, &block)\n result = @relation.send(method, *args, &block)\n\n unless result.kind_of?(::Veritas::Relation)\n return result\n end\n\n self.class.new(@mapper, result)\n end",
"def new_word_relation(sax, author_id)\n AntonomyRelation.new.tap do |relation|\n relation.author_id = author_id\n relation.synset1_id = sax.synset1_id\n relation.synset2_id = sax.synset2_id\n relation.word1_id = sax.word1_id\n relation.word2_id = sax.word2_id\n end\nend",
"def rel(rels)\n # rel must be an array.\n data[:rel] = Array(rels)\n end",
"def update!(**args)\n @relation = args[:relation] if args.key?(:relation)\n end",
"def op_add_relation!(field, objects = [])\n objects = [objects] unless objects.is_a?(Array)\n return false if objects.empty?\n relation_action = Parse::RelationAction.new(field, polarity: true, objects: objects)\n operate_field! field, relation_action\n end",
"def relation(related, arguments=[])\n self.send(related.to_sym)\n end",
"def relation(*args)\n source_node.join(target_node, *args)\n end",
"def related_to(*args)\n@relations = args.map(&:to_s)\nend",
"def visit_axiom_relation_base(base_relation)\n @name = base_relation.name.to_s.freeze\n @from = visit_identifier(@name)\n @header = base_relation.header\n @columns = columns_for(base_relation)\n self\n end",
"def get_relation_tuples(namespace, opts = {})\n data, _status_code, _headers = get_relation_tuples_with_http_info(namespace, opts)\n data\n end",
"def visit_axiom_relation_operation_binary(binary)\n generator = self.class.visit(binary)\n @name = generator.name\n @from = aliased_subquery(generator)\n generator\n end",
"def visit_ActiveRecord_Relation(o, parent)\n o.arel\n end",
"def read_join_relations(obj, res_row, row, join_relations)\n offset = obj.class.serializable_attributes.size\n\n for rel in join_relations\n rel_obj = rel[:target_class].og_allocate(res_row, row)\n rel_obj.og_read(res_row, row, offset)\n offset += rel_obj.class.serializable_attributes.size\n obj.instance_variable_set(\"@#{rel[:name]}\", rel_obj)\n end\n end",
"def new_relation(rel_name, internal_relation)\n relations_info[rel_name.to_sym][:relation].new(internal_relation) # internal_relation is a java neo object\n end",
"def resolve!\n # default to base relation if not specified\n @relation_type ||= RelationTypes::BASE\n @attributes.map(&:resolve!)\n end",
"def add_relationship(rel_attr); end",
"def call(*args)\n ((combine? || composite?) ? relation : (relation >> mapper)).call(*args)\n end",
"def read_relation_list(node)\n target_type = Utils.add_namespace(node.attributes['target-type'], Model::NS_REL_1)\n node.elements.each('relation') {|child|\n yield create_relation(child, target_type)\n }\n end",
"def to_ruby_literal\n \"Alf::Relation[\" +\n tuples.collect{|t| Tools.to_ruby_literal(t)}.join(', ') + \"]\"\n end",
"def set_relation\n @relation = Relation.find(params[:id])\n end",
"def set_relation\n @relation = Relation.find(params[:id])\n end",
"def gateway(relation)\n Axiom::Relation::Gateway.new(self, relation)\n end",
"def each_relation\n\t for rel in relations\n\t\tyield(rel)\n\t end\n\tend",
"def relation_method\n :product\n end",
"def relation\n raise NotImplementedError\n end",
"def _expand(uri_variables = {})\n self.class.new(@key, @link, @entry_point, uri_variables)\n end",
"def get_relation(args)\n\tapi_url = \"#{@base_url}/#{args[:collection]}/#{args[:key]}/relations/#{args[:relation]}\"\n\tdo_the_get_call( url: api_url, user: @user )\nend",
"def set_rel_props(rel, hash)\n @batch_inserter.set_relationship_properties(rel, hash)\n end",
"def <<(values)\n values = prepare_relation(values) if values.is_a?(Relation)\n self.set(objects.to_a | Array.wrap(values))\n end",
"def relations\n @relations ||= process_rels\n end",
"def set_relation\n @relation = Relation.find(params[:id])\n end",
"def relation\n @relation ||= begin\n additions = false\n # TODO: merge! has unexpected results.\n frag = include_params.reduce(model.all) do |result, inclusion|\n config = config_for(inclusion)\n query = config.respond_to?(:keys) ? config.values.first : nil\n unless query.nil?\n additions = true\n result = result.merge(query)\n end\n result\n end\n additions ? frag : nil\n end\n end",
"def expand_url(pairs)\n url.expand Hash[pairs]\n end",
"def get_expand(namespace, object, relation, max_depth, opts = {})\n data, _status_code, _headers = get_expand_with_http_info(namespace, object, relation, max_depth, opts)\n data\n end",
"def create_tuple!(attributes)\n # split the attrs (needs DRYing)\n rel_attrs = {}\n attributes.each do |key, value|\n if key.to_s[0..3] == 'rel_'\n rel_attrs[key] = attributes.delete(key)\n end\n end\n\n # find or create the one node\n if target_node = @target_class.find(attributes[:id])\n target_node.update_attributes!(attributes)\n else\n target_node = @target_class.create!(attributes)\n end\n\n self.relate(target_node, rel_attrs)\n\n # finally, make and return the tuple\n #Tuple.new(@node, target_node)\n #Tuple.new(@node, target_node)\n\n end",
"def coerce_to_relation(summarize_with)\n if summarize_with.kind_of?(Relation)\n summarize_with\n else\n project(Relation::Header.coerce(summarize_with))\n end\n end",
"def expand(*args)\n\t\t\t#\tjust the default value, not the Response\n\t\t\t(super *args).to_s\n\t\tend",
"def adding_plan_relation(transaction, parent, child, relations, info); end",
"def expand(binding)\n begin\n context.push(binding)\n render\n ensure\n context.pop\n end\n end",
"def generate_person_and_task_relations(relations)\n relations.each_pair { |person_id, values|\n values.each { |task_id|\n @db_base.map_task_to_person(person_id, task_id)\n }\n }\n end",
"def expand\n map { |p| p&.flatten || p }.flatten\n end",
"def expand_more(more)\n response = client.get(\n \"/api/morechildren\",\n children: more.join(\",\"),\n link_id: fullname\n )\n\n client.object_from_body(\n kind: \"Listing\",\n data: {\n before: \"\", after: \"\",\n children: response.body[:json][:data][:things]\n }\n )\n end",
"def relationship_aux(action, opts)\n\n if action != \"destroy\"\n\n data = LibXML::XML::Parser.string(request.raw_post).parse\n\n subject = parse_element(data, :resource, '/relationship/subject')\n predicate = parse_element(data, :resource, '/relationship/predicate')\n objekt = parse_element(data, :resource, '/relationship/object')\n context = parse_element(data, :resource, '/relationship/context')\n end\n\n # Obtain object\n\n case action\n when 'create';\n return rest_response(401, :reason => \"Not authorised to create a relationship\") unless Authorization.check('create', Relationship, opts[:user], context)\n ob = Relationship.new(:user => opts[:user])\n when 'view', 'edit', 'destroy';\n ob, error = obtain_rest_resource('Relationship', opts[:query]['id'], opts[:query]['version'], opts[:user], action)\n else\n raise \"Invalid action '#{action}'\"\n end\n\n return error if ob.nil? # appropriate rest response already given\n\n if action == \"destroy\"\n\n ob.destroy\n\n else\n\n # build it\n\n ob.subject = subject if subject\n ob.predicate = predicate if predicate\n ob.objekt = objekt if objekt\n ob.context = context if context\n\n if not ob.save\n return rest_response(400, :object => ob)\n end\n end\n\n rest_get_request(ob, opts[:user], { \"id\" => ob.id.to_s })\nend",
"def <<(relation)\n new_node(relation.name, relation)\n end",
"def to_ruby_literal\n \"Alf::Relation([\" + tuples.map{|t| Support.to_ruby_literal(t) }.join(', ') + \"])\"\n end",
"def relation_params\n _relation_params\n end",
"def update!(**args)\n @relation_descriptor = args[:relation_descriptor] if args.key?(:relation_descriptor)\n end",
"def method_missing(method, *args, &block)\n relation.send(method, *args, &block)\n end",
"def show\n\t\t@relation = Relation.find(params[:id])\n\t\trespond_to do |format|\n\t\t\tformat.html # show.html.erb\n\t\t\tformat.xml { render :xml => @relation }\n\t\tend\n\tend",
"def add_rel(predicate, object, required = false)\n object = check_objects(object)\n if(!object)\n raise(ArgumentError, \"Relation with empty object on #{predicate} (#{@current.attributes['uri']}).\") if(required)\n return\n end\n if(object.kind_of?(Array))\n object.each do |obj| \n raise(ArgumentError, \"Cannot add relation on database field <#{predicate}> - <#{object.inspect}>\") if(ActiveSource.db_attr?(predicate))\n set_element(predicate, \"<#{irify(obj)}>\", required) \n end\n else\n raise(ArgumentError, \"Cannot add relation on database field\") if(ActiveSource.db_attr?(predicate))\n set_element(predicate, \"<#{irify(object)}>\", required)\n end\n end",
"def relation\n super.extending!(TranslatedAttributesQuery)\n end",
"def rdfa_relation(name,tag,relation,options={})\n property_value = relation\n property_uri = name\n property_datatype = options[:datatype]\n if(name.instance_of?(Symbol))\n key = self.class.resource_mapping.keys.detect do |k|\n k == name\n end\n raise Exception.new(\"Unknown rdfa property #{name} for object #{self}\") if key.nil?\n property_uri = \"#{SemanticResource::Configuration.namespaces_map[self.class.resource_mapping[key].first]}#{self.class.resource_mapping[key].last}\"\n end\n options.delete(:datatype)\n\n rdfa = StringIO.new\n\n rdfa << \"<#{tag} rel='#{property_uri}'\"\n rdfa << \" href='#{property_value}'\"\n\n options.each_pair do |k,v|\n rdfa << \"#{k}='#{v}' \"\n end\n if(block_given?)\n rdfa << \">\"\n rdfa << yield if block_given?\n rdfa << \"</#{tag}>\"\n else\n rdfa << \" />\"\n end\n\n rdfa.string\n end",
"def show\n\t\tshow_\n\t\trespond_to do |format|\n\t\t\tformat.html # show.html.erb\n\t\t\tformat.xml { render :xml => @relation }\n\t\tend\n\tend",
"def update(tuple)\n hash = relvar_def.__prepare_for_update(self, tuple)\n underlying_table.update(relvar_def.__physical_tuple_encode(self, hash))\n end",
"def expand_routes(*slice)\n expanded = parent_resource.expand_routes(route_variables)\n\n unless slice.empty?\n expanded = expanded.send(:slice, *slice)\n end\n\n expanded.transform_keys do |key|\n \"#{ key }_url\"\n end\n end",
"def lift_tuple(tuple)\n if Pione::TupleSpace[tuple.first]\n if pos = Pione::TupleSpace[tuple.first].location_position\n if new_location = lift_location(tuple[pos])\n tuple = tuple.clone\n tuple[pos] = new_location\n end\n end\n end\n return tuple\n end",
"def expand!\n if attributes.nil?\n @attributes = {}\n if @parent\n @parent.expand!\n\n @parent.send(:\"_#{@path.last}=\", self)\n end\n end\n end",
"def expand\n inst = model.new(api)\n inst.data[:id] = self.id\n inst.reload\n end",
"def add_relation source_label, relation, destination_label, direction = '<->'\n # More popular nodes have more weigth\n @node_weights[source_label] += 1\n @node_weights[destination_label] += 1\n\n @edge_buffer << Edge.new(\n source_label: source_label,\n relation: relation,\n destination_label: destination_label,\n direction: direction\n )\n end",
"def related_attrs\n relationship = flex_options[:relationship_name]\n send relationship\n end",
"def decorate_relation(obj)\n if obj.respond_to?(:records)\n # Rails 5.0\n obj.extend ActiveDecorator::RelationDecorator unless ActiveDecorator::RelationDecorator === obj\n else\n # Rails 3.x and 4.x\n obj.extend ActiveDecorator::RelationDecoratorLegacy unless ActiveDecorator::RelationDecoratorLegacy === obj\n end\n obj\n end",
"def populate_relation(args)\n\targs[:keys_B].each do |key_B|\n\t\tapi_url = \"#{@base_url}/#{args[:coll_A]}/#{args[:key_A]}/relation/\" +\n\t\t \"#{args[:relation]}/#{args[:coll_B]}/#{key_B}\"\n\t\tputs do_the_put_call({ url: api_url, user: @user, json: '{}' })\n\tend\nend",
"def rels(*rel_types)\n if rel_types.empty?\n AllRelsDsl.new(@_relationships, _java_node)\n else\n storage = _create_or_get_storage(rel_types.first)\n RelsDSL.new(storage)\n end\n end",
"def set_relation(name, relation)\n instance_variable_set(\"@_#{name}\", relation)\n end",
"def relationship(rel_class)\n @relationship = rel_class\n self\n end",
"def relation(relation)\n true\n end",
"def relation=(relation)\n if relation.nil?\n fail ArgumentError, 'invalid value for \"relation\", relation cannot be nil.'\n end\n @relation = relation\n end",
"def relation\n Proxy\n end",
"def relation\n Proxy\n end",
"def relation\n Proxy\n end",
"def resolve!\n @relations.map(&:resolve!)\n end",
"def relations\n Relations::RelationTraverser.new(@internal_node)\n end",
"def relation_params\n params.require(:relation).permit(:user_id, :friend_id)\n end",
"def relationship(rel_class)\n @relationship = rel_class\n self\n end",
"def get_relation(id)\n get_object('relation', id)\n end",
"def get_relation(name, metadata, object, reload = false)\n if !reload && (value = ivar(name)) != false\n value\n else\n _building do\n _loading do\n if object && needs_no_database_query?(object, metadata)\n __build__(name, object, metadata)\n else\n __build__(name, attributes[metadata.key], metadata)\n end\n end\n end\n end\n end",
"def relations; enum_relations.to_a end",
"def relation(relation_name)\n state_depth_must_be(States::SCHEMA)\n r = relation_lookup(relation_name)\n raise \"relation #{relation_name} already exists\" if r\n n = Relation.new(@current_schema, relation_name)\n @current_schema.relations << n\n @current_relation = n\n clear_state_below(States::RELATION)\n end"
] |
[
"0.6148216",
"0.597972",
"0.58912784",
"0.58220965",
"0.5735235",
"0.5635795",
"0.5595332",
"0.54948616",
"0.54483986",
"0.53932273",
"0.5336877",
"0.5321562",
"0.5305559",
"0.5154556",
"0.51443547",
"0.514021",
"0.50917137",
"0.50810313",
"0.5056913",
"0.50432634",
"0.50231636",
"0.4998036",
"0.49817613",
"0.49755242",
"0.49703172",
"0.49444935",
"0.48969886",
"0.48900455",
"0.48868823",
"0.48776072",
"0.4871543",
"0.48372898",
"0.48177755",
"0.47938666",
"0.47838473",
"0.47827587",
"0.47493902",
"0.47479904",
"0.4745684",
"0.47168532",
"0.4710221",
"0.47089288",
"0.47089288",
"0.46924704",
"0.46875665",
"0.4666554",
"0.46535966",
"0.46261346",
"0.46099108",
"0.46079123",
"0.46066675",
"0.4604007",
"0.45964155",
"0.45953074",
"0.45900917",
"0.4574557",
"0.45720065",
"0.4567852",
"0.4550259",
"0.45497584",
"0.454591",
"0.45442685",
"0.45420626",
"0.4540911",
"0.45372537",
"0.45325226",
"0.45304716",
"0.45146233",
"0.4487139",
"0.44810325",
"0.44711804",
"0.44668144",
"0.44655257",
"0.44576895",
"0.44552317",
"0.44481727",
"0.44406423",
"0.44258112",
"0.44187057",
"0.44010544",
"0.4399764",
"0.43879187",
"0.43860164",
"0.43804097",
"0.43746665",
"0.43713412",
"0.43694067",
"0.43597174",
"0.4359455",
"0.43577155",
"0.43577155",
"0.43577155",
"0.43481386",
"0.4346786",
"0.43430996",
"0.43412992",
"0.43406665",
"0.4333126",
"0.43295112",
"0.4319514"
] |
0.4332848
|
98
|
Query relation tuples Get all relation tuples that match the query. Only the namespace field is required.
|
def get_relation_tuples(namespace, opts = {})
data, _status_code, _headers = get_relation_tuples_with_http_info(namespace, opts)
data
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def get_relation_tuples_with_http_info(namespace, opts = {})\n if @api_client.config.debugging\n @api_client.config.logger.debug 'Calling API: ReadApi.get_relation_tuples ...'\n end\n # verify the required parameter 'namespace' is set\n if @api_client.config.client_side_validation && namespace.nil?\n fail ArgumentError, \"Missing the required parameter 'namespace' when calling ReadApi.get_relation_tuples\"\n end\n # resource path\n local_var_path = '/relation-tuples'\n\n # query parameters\n query_params = opts[:query_params] || {}\n query_params[:'namespace'] = namespace\n query_params[:'page_token'] = opts[:'page_token'] if !opts[:'page_token'].nil?\n query_params[:'page_size'] = opts[:'page_size'] if !opts[:'page_size'].nil?\n query_params[:'object'] = opts[:'object'] if !opts[:'object'].nil?\n query_params[:'relation'] = opts[:'relation'] if !opts[:'relation'].nil?\n query_params[:'subject_id'] = opts[:'subject_id'] if !opts[:'subject_id'].nil?\n query_params[:'subject_set.namespace'] = opts[:'subject_set_namespace'] if !opts[:'subject_set_namespace'].nil?\n query_params[:'subject_set.object'] = opts[:'subject_set_object'] if !opts[:'subject_set_object'].nil?\n query_params[:'subject_set.relation'] = opts[:'subject_set_relation'] if !opts[:'subject_set_relation'].nil?\n\n # header parameters\n header_params = opts[:header_params] || {}\n # HTTP header 'Accept' (if needed)\n header_params['Accept'] = @api_client.select_header_accept(['application/json'])\n\n # form parameters\n form_params = opts[:form_params] || {}\n\n # http body (model)\n post_body = opts[:debug_body]\n\n # return_type\n return_type = opts[:debug_return_type] || 'GetRelationTuplesResponse'\n\n # auth_names\n auth_names = opts[:debug_auth_names] || []\n\n new_options = opts.merge(\n :operation => :\"ReadApi.get_relation_tuples\",\n :header_params => header_params,\n :query_params => query_params,\n :form_params => form_params,\n :body => post_body,\n :auth_names => auth_names,\n :return_type => return_type\n )\n\n data, status_code, headers = @api_client.call_api(:GET, local_var_path, new_options)\n if @api_client.config.debugging\n @api_client.config.logger.debug \"API called: ReadApi#get_relation_tuples\\nData: #{data.inspect}\\nStatus code: #{status_code}\\nHeaders: #{headers}\"\n end\n return data, status_code, headers\n end",
"def tuples\n results = run_tuples(@last_query_context.to_sparql_select)\n results[\"results\"][\"bindings\"].map do |h|\n h.keys.each do |k|\n h[k.to_sym] = QL.from_tuple_binding(h[k])\n h.delete(k)\n end\n h\n end\n end",
"def tuples\n relation = self.relation\n return relation if materialized?\n\n adapter.read(relation)\n end",
"def relations\n @relations ||= process_rels\n end",
"def relations\n Relations::RelationTraverser.new(@internal_node)\n end",
"def relations\n _relations\n end",
"def relations\n _relations\n end",
"def get_relations\n # Descobrir as relacoes:\n # Orientados\n # Projetos de Pesquisa\n # Producoes Bibliograficas\n # Consultar o RP de cada uma\n # Somar os RPs e retornar\n orientados = self.get_relations_of_orientados\n projetos = self.get_relations_of_projetos_pesquisa\n producoes = self.get_relations_of_producoes_bibliograficas\n\n relations = (orientados | projetos | producoes).delete self.id\n return relations.to_a\n end",
"def traverse(start_elements, relation_filter, stop_filter, from_path = [])\n return [] if from_path.size > 100\n start_elements.each_with_object([]) do |el, relations|\n concrete_rels = concrete_relationships(el, relation_filter, from_path)\n relations.concat(\n concrete_rels,\n *derived_relationship_paths(concrete_rels, relation_filter, stop_filter, from_path)\n )\n end\n end",
"def relation_get_all(user,kind)\n results = []\n Note_Relation.all( :note_id => self.id, :user_id => user.id, :kind => kind ).each do |r|\n results.push(r.value)\n end\n return results\n end",
"def find_relations\n puts '2nd pass: find relations'\n find :relations\n self\n end",
"def relations_for_relation_type(relation_type)\n # Find all the relations that belong to us for this RelationType, ordered by position\n related_ids = relations.where(relation_type_id: relation_type.id).order(:position).select(:related_to_id)\n\n # Construct a query for all these records\n result = relation_type.applies_to.constantize.where(id: related_ids)\n\n # Merge in the relation_filter if it's available\n result = result.merge(relation_filter_for_relation_type(relation_type)) if relation_filter_for_relation_type(relation_type)\n\n # make sure results are in same order as related_ids array (position order)\n if result.present?\n result.where(id: related_ids).order(:position)\n end\n\n result\n end",
"def each_relation\n\t for rel in relations\n\t\tyield(rel)\n\t end\n\tend",
"def relation_all(model)\n validate_model(model)\n model.all\n end",
"def relation_subjects(record)\n query = ActiveFedora::SolrQueryBuilder.construct_query_for_rel(\n [[:has_model, proxy_class.to_rdf_representation], [:proxyFor, record.id]]\n )\n rows = ActiveFedora::SolrService::MAX_ROWS\n ActiveFedora::SolrService.query(query, fl: 'id', rows: rows).map(&:rdf_uri)\n end",
"def relationships\n return [] if empty?\n\n map { |pivot_table| Relationship.new(pivot_table, PIVOT_TABLE_R, \"../#{pivot_table.pn}\") }\n end",
"def roots(relation)\n\t @result_set = plan.query_roots(result_set, relation)\n\t self\n\tend",
"def distinct_relations(from_cache=false)\n # @@neo = Neography::Rest.new\n begin\n if from_cache\n self_node = self.get_node\n @neo.get_node_properties(node1)\n else\n self_node_id = self.get_node_id\n qur = \"start n=node(\"+self_node_id.to_s+\") match n-[r]->() return distinct(type(r)), count(r), r.magnitude;\"\n response = @@neo.execute_query(qur)\n distinct_relations = response[\"data\"]\n return distinct_relations\n end\n rescue Exception\n return nil\n end\n end",
"def relationships\n return [] if empty?\n\n map { |table| Relationship.new(table, TABLE_R, \"../#{table.pn}\") }\n end",
"def read_relation_list(node)\n target_type = Utils.add_namespace(node.attributes['target-type'], Model::NS_REL_1)\n node.elements.each('relation') {|child|\n yield create_relation(child, target_type)\n }\n end",
"def relations_for(node) #:nodoc:\n Enumerator.new do |yielder|\n @description.relations.each do |relation_accessor|\n begin\n relations = relation_accessor[node]\n enumerable = relations.is_a?(Enumerable) ? relations : [relations].compact\n\n enumerable.each { |e| yielder << e unless visited?(e) }\n rescue NoMethodError\n # ignore errors on relation_accessor[node]\n end\n end\n end\n end",
"def map_in_included_relations!(result, query)\n dat = get_relation_data_as_hash(query)\n result.each do |res|\n dat.each do |name, lookup|\n res[name] = lookup[res[@relations[name][:this_key]]]\n if @relations[name][:type] == :has_one\n res[name] = res[name][0] unless res[name].nil?\n else \n res[name] = [] if res[name].nil?\n end\n end\n end\n end",
"def relations\n @relations ||= {}\n end",
"def related_objects(relation = nil, result = nil)\n\t result ||= ValueSet.new\n\t if relation\n\t\tresult.merge(parent_objects(relation).to_value_set)\n\t\tresult.merge(child_objects(relation).to_value_set)\n\t else\n\t\teach_relation { |rel| related_objects(rel, result) }\n\t end\n\t result\n\tend",
"def derived_relations(start_elements, relationship_filter, target_filter, stop_filter = FAIL_ALL)\n traverse(start_elements, relationship_filter, stop_filter)\n .reject(&single_relation_paths) # See #2 above\n .select(&target_relations(target_filter))\n .map(&create_relationship_for_path)\n .uniq { |rel| [rel.type, rel.source, rel.target] }\n end",
"def get_relations_from_api(api=Rosemary::API.new)\n api.get_relations_referring_to_object(type, self.id.to_i)\n end",
"def get_relation(args)\n\tapi_url = \"#{@base_url}/#{args[:collection]}/#{args[:key]}/relations/#{args[:relation]}\"\n\tdo_the_get_call( url: api_url, user: @user )\nend",
"def find_relations\n sql = <<-eos\n SELECT\n tc.constraint_name, tc.table_name, kcu.column_name,\n ccu.table_name AS foreign_table_name,\n ccu.column_name AS foreign_column_name\n FROM\n information_schema.table_constraints AS tc\n JOIN information_schema.key_column_usage AS kcu ON tc.constraint_name = kcu.constraint_name\n JOIN information_schema.constraint_column_usage AS ccu ON ccu.constraint_name = tc.constraint_name\n WHERE constraint_type = 'FOREIGN KEY'\n eos\n @relations = @connection.exec(sql).values\n end",
"def relations; enum_relations.to_a end",
"def get_relations_for_actor(search)\n # Self\n q = \"SELECT *\n WHERE { ?x <http://www.semanticweb.org/ontologies/2011/10/moviesandtv.owl#hasName> \\\"#{search}\\\" }\"\n results = query(q)\n relations = [{:person_uri => results.last[:x], :name => search}]\n # Shows\n q = \"SELECT *\n WHERE { ?x <http://www.semanticweb.org/ontologies/2011/10/moviesandtv.owl#hasName> \\\"#{search}\\\" .\n ?x <http://www.semanticweb.org/ontologies/2011/10/moviesandtv.owl#isActorIn> ?show .\n ?show <http://www.semanticweb.org/ontologies/2011/10/moviesandtv.owl#hasTitle> ?name }\"\n results = query(q)\n results.each do |result|\n relations << {:show_uri => result[:show], :name => result[:name]}\n end\n return relations\n end",
"def printout_relations\n @relations.each do |shape, relations|\n relations.each do |relation|\n relation.each do |key, value|\n puts \"\\\"#{shape}\\\" #{value} \\\"#{key}\\\"\"\n end\n end\n end\n end",
"def relations\n return @artist_data[\"relations\"]\n end",
"def relations\n self.class.relations\n end",
"def relation_types\n Spree::RelationType.where(applies_to: to_relation_name).order(:name)\n end",
"def relation\n relation = nodes.reduce(root) do |a, e|\n a.associations[e.name.key].join(:join, a, e)\n end\n schema.(relation)\n end",
"def all_objects\n objects = []\n each_namespace{|n| objects << n}\n each_namespace{|n| \n n.each_relvar{|r| objects << r}\n }\n each_namespace{|n| \n n.each_relvar{|r| r.each_candidate_key{|k| objects << k}}\n }\n each_namespace{|n| \n n.each_relvar{|r| r.each_foreign_key{|k| objects << k}}\n }\n objects\n end",
"def get_relations(tokens, base_noun_chunks)\n graph = Graph.new\n tokens.each do |t|\n if t[:args]\n t[:args].each do |type, arg|\n graph.add_edge(t[:idx], arg, 1) if arg >= 0\n end\n end\n end\n \n rels = []\n heads = base_noun_chunks.collect{|c| c[:head]}\n base_noun_chunks.combination(2) do |c|\n path = graph.shortest_path(c[0][:head], c[1][:head])\n s = path.shift\n o = path.pop\n rels << [s, path, o] if (path & heads).empty?\n end\n return rels\n end",
"def get_relations(service)\n\t\treturn @transport.get_path(\"relations\",service)\n\tend",
"def get_relations(type, options)\n configure_twitter\n result = case type\n when :follows\n Twitter.friend_ids(options)\n when :followers\n Twitter.follower_ids(options)\n end\n # next_cursor = result.next_cursor\n ids = result.ids\n options[:page] = 0 if options[:page].nil?\n offset = options[:page].to_i * USERS_PER_PAGE\n users = Twitter.users(result.collection[offset, USERS_PER_PAGE+1]).index_by(&:id)\n data = []\n ids.each {|id|\n data << users[id] unless users[id].nil?\n }\n data\n end",
"def related_to(*args)\n@relations = args.map(&:to_s)\nend",
"def list(relation = nil)\n relation ||= apply_filter(self.relation)\n\n apply_paging(\n apply_sorting(relation)\n ).to_a\n end",
"def containing_relation_members\n []\n end",
"def findRelation(query)\n curNode = @head\n while curNode != nil\n if curNode.data.relation == query\n curNode.data.print\n end\n curNode = curNode.next\n end\n end",
"def get_relation(*items, **opt)\n terms = []\n meth = opt.delete(:meth) || \"#{self_class}.#{__method__}\"\n id_opt = opt.extract!(:id_key, :sid_key).transform_values!(&:to_sym)\n id_key = id_opt[:id_key] ||= id_column\n sid_key = id_opt[:sid_key] ||= sid_column\n\n # === Record specifiers\n ids = id_key ? Array.wrap(opt.delete(id_key)) : []\n sids = sid_key ? Array.wrap(opt.delete(sid_key)) : []\n if items.present?\n recs = expand_ids(*items).map { |term| id_term(term, **id_opt) }\n ids = recs.map { |rec| rec[id_key] } + ids if id_key\n sids = recs.map { |rec| rec[sid_key] } + sids if sid_key\n end\n ids = ids.compact_blank!.uniq.presence\n sids = sids.compact_blank!.uniq.presence\n if ids && sids\n terms << sql_terms(id_key => ids, sid_key => sids, join: :or)\n elsif ids\n opt[id_key] = ids\n elsif sids\n opt[sid_key] = sids\n end\n\n # === Sort order\n # Avoid applying a sort order if identifiers were specified or if\n # opt[:sort] was explicitly *nil* or *false*. Permit :asc as shorthand for\n # the default sort order ascending; :desc as shorthand for the default sort\n # order descending.\n if (sort = opt.key?(:sort) ? opt.delete(:sort) : (ids || sids).blank?)\n case sort\n when Hash then col, dir = sort.first\n when TrueClass then col, dir = [nil, nil]\n when /^ASC$/i, /^DESC$/i then col, dir = [nil, sort]\n else col, dir = [sort, nil]\n end\n col ||= implicit_order_column || pagination_column\n dir &&= dir.to_s.upcase\n sort = col && \"#{col} #{dir}\".squish\n Log.info { \"#{meth}: no default sort\" } unless sort\n end\n\n # === Filter by user\n user_opt = opt.extract!(:user, :user_id)\n if user_column && user_opt.present?\n users = user_opt.values.flatten.map { |u| User.id_value(u) }.uniq\n users = users.first if users.size == 1\n terms << sql_terms(user_column => users, join: :or)\n end\n\n # === Filter by state\n state_opt = state_column && opt.extract!(state_column)\n terms << sql_terms(state_opt, join: :or) if state_opt.present?\n\n # === Update time lower bound\n exclusive, inclusive = [opt.delete(:after), opt.delete(:start_date)]\n lower = exclusive || inclusive\n day, month, year = day_string(lower)\n lower = day if day\n if (lower &&= (lower.to_datetime rescue nil))\n lower += 1.month if exclusive && month\n lower += 1.year if exclusive && year\n on_or_after = (exclusive && !month && !year) ? '>' : '>='\n terms << \"updated_at #{on_or_after} '#{lower}'::date\"\n end\n\n # === Update time upper bound\n exclusive, inclusive = [opt.delete(:before), opt.delete(:end_date)]\n upper = exclusive || inclusive\n day, month, year = day_string(upper)\n upper = day if day\n if (upper &&= (upper.to_datetime rescue nil))\n upper += 1.month - 1.day if inclusive && month\n upper += 1.year - 1.day if inclusive && year\n on_or_before = exclusive ? '<' : '<='\n terms << \"updated_at #{on_or_before} '#{upper}'::date\"\n end\n\n # === Record limit/offset\n limit = positive(opt.delete(:limit))\n offset = positive(opt.delete(:offset))\n if offset && pagination_column\n terms << \"#{pagination_column} > #{offset}\"\n elsif offset\n Log.warn { \"#{meth}: pagination not supported\" }\n end\n\n # === Filter by association\n assoc = opt.keys.map(&:to_s).select { |k| k.include?('.') }\n assoc.map! { |k| k.split('.').first.singularize.to_sym }\n assoc = assoc.presence\n\n # === Generate the relation\n query = sql_terms(opt, *terms, join: :and)\n result = assoc ? joins(*assoc).where(query) : where(query)\n result.order!(sort) if sort.present?\n result.limit!(limit) if limit.present?\n result\n end",
"def all\n @data_adapter.relations\n end",
"def xpath_all(pdoc, path, namespace = '')\n if namespace!=\"\"\n return REXML::XPath.match(pdoc, path, namespace)\n else\n return REXML::XPath.match(pdoc, path);\n end\n return []\n end",
"def get_relations(project_id = nil, base_ids = nil, sort = false)\n\t\t_relations = self.subcatrels.in_project(project_id).among_denotations(base_ids)\n\t\tif sort\n\t\t\t_relations.sort{|r1, r2| r1.id <=> r2.id}\n\t\telse\n\t\t\t_relations\n\t\tend\n\tend",
"def get_relations_referring_to_object(type, id)\n api_call_with_type(type, id, \"#{type}/#{id}/relations\")\n end",
"def direct_predicates\n ActiveRDF::Query.new(N::Predicate).distinct(:p).where(self, :p, :o).execute\n end",
"def query_all_years(attribute, options)\n query = super(attribute, options)\n query = query.unscope(where: :context_id)\n where_conds = []\n where_vars = []\n @node_indices_by_context.each do |context_id, node_idx|\n where_conds << \"flows.context_id = ? AND ? = path[?]\"\n where_vars += [context_id, @node.id, node_idx]\n end\n query.where(\n where_conds.join(\" OR \"), *where_vars\n )\n end",
"def queryns \n e = first_element('query')\n if e\n return e.namespace\n else\n return nil\n end\n end",
"def xpath_all(pdoc, path, namespace = '')\n begin\n if namespace != \"\"\n return pdoc.find(path, namespace) if pdoc.find(path, namespace)\n else\n return pdoc.find(path) if pdoc.find(path)\n end\n rescue\n return []\n end\n end",
"def rels\n @traversal_result = :rels\n self\n end",
"def rels(node)\n @batch_inserter.getRelationships(node)\n end",
"def query(*triples)\n bindings = nil\n triples.each do |triple|\n binding_position = {}\n query = []\n triple.each_with_index do |e, i|\n if query_variable?(e)\n binding_position[e] = i\n query << nil\n else\n query << e \n end\n end\n rows = find(*query)\n if bindings.nil?\n bindings = rows.inject([]) do |list, row|\n binding = {}\n binding_position.each do |var, pos|\n binding[var] = row[pos]\n end\n list << binding\n end\n else\n new_binding = []\n bindings.each do |binding|\n rows.each do |row|\n valid_match = true\n temp_binding = binding.dup\n binding_position.each do |var, pos|\n if temp_binding.include?(var)\n valid_match = false if temp_binding[var] != row[pos]\n else\n temp_binding[var] = row[pos]\n end\n end\n new_binding << temp_binding if valid_match\n end\n end\n bindings = new_binding.dup\n end\n bindings\n end\n return bindings\n end",
"def relationships_with_name(input, name)\n input.relations(name)\n end",
"def containing_relation_members \n return [] \n end",
"def tuples\n @storage.enum_for(:each_tuple, @dir)\n end",
"def get_all_related_tables\n tables = [ @table_name ]\n return tables if @relations.nil?\n @relations.each do |n,r|\n tables = tables | get_relation_tables(r)\n end\n tables.sort\n end",
"def all(*vars)\n result = Query.get self, *vars\n result.get_docs\n end",
"def find_related_items(pAuthToken, p_args)\r\n relation_type = p_args[:relation_type]\r\n item_id = p_args[:item_id]\r\n offset = p_args[:offset]\r\n max = p_args[:max]\r\n\r\n unless relation_type.nil?\r\n service_uri = \"#{@@m_service_url.path}#{@@m_service_find_related_items}#{relation_type}&#{item_id}.#{@@m_content_format}?authToken=#{pAuthToken}\"\r\n params = nil\r\n unless offset.nil?\r\n params = \"&offset=#{offset}\"\r\n end\r\n unless max.nil?\r\n if params.nil?\r\n params = \"?max=#{max}\"\r\n else\r\n params += \"&max=#{max}\"\r\n end\r\n end\r\n service_uri += params\r\n end\r\n search_result = SearchResult.new\r\n state_true?(nil, post_request(service_uri, nil).body) {|root|\r\n build_search_result(search_result, root)\r\n }\r\n search_result\r\n end",
"def all\n results = query '*::*::*::*'\n (results.each {|r| yield r}) if block_given?\n results\n end",
"def child_rels(*args)\n options = args.extract_options!\n rels = relationships.flat_map(&:children).uniq\n Relationship.filter_by_resource_type(rels, options)\n end",
"def path_rels(*args)\n options = args.extract_options!\n rel = relationship(:raise_on_multiple => true) # TODO: Handle multiple nodes with a way to detect which node you want\n rels = rel.nil? ? [relationship_for_isolated_root] : rel.path\n Relationship.filter_by_resource_type(rels, options)\n end",
"def sibling_rels(*args)\n options = args.extract_options!\n rels = relationships.flat_map(&:siblings).uniq\n Relationship.filter_by_resource_type(rels, options)\n end",
"def target_relations\n query = \"at:link[@rel = '#{Rel[repository.cmis_version][:relationships]}']/@href\"\n link = data.xpath(query, NS::COMBINED)\n if link.length == 1\n link = Internal::Utils.append_parameters(link.text, \"relationshipDirection\" => \"target\", \"includeSubRelationshipTypes\" => true)\n Collection.new(repository, link)\n else\n raise \"Expected exactly 1 relationships link for #{key}, got #{link.length}, are you sure this is a document/folder?\"\n end\n end",
"def search_graph(query)\n (search_following(query) + search_followers(query)).uniq\n end",
"def relation\n context = Order\n context = context.where(from_group_id: from_group_id) if from_group_id\n context = context.where(to_producer_id: to_producer_id) if to_producer_id\n context = context.where(confirm_before: confirm_before) if confirm_before\n context = context.order(:updated_at)\n context\n end",
"def get_relations(tokens, base_noun_chunks)\n graph = Graph.new\n tokens.each do |t|\n if t[:args]\n t[:args].each do |type, arg|\n graph.add_edge(t[:idx], arg, 1) if arg >= 0\n end\n end\n end\n\n rels = []\n heads = base_noun_chunks.collect{|c| c[:head]}\n base_noun_chunks.combination(2) do |c|\n path = graph.shortest_path(c[0][:head], c[1][:head])\n s = path.shift\n o = path.pop\n rels << [s, path, o] if (path & heads).empty?\n end\n rels\n end",
"def includes(*_args)\n relations.map { |rel| rel[1] }\n end",
"def subtree_rels(*args)\n options = args.extract_options!\n # TODO: make this a single query (vs 3)\n # thus making filter_by_resource_type into a query\n rels = relationships.flat_map(&:subtree).uniq\n rels = [relationship_for_isolated_root] if rels.empty?\n Relationship.filter_by_resource_type(rels, options)\n end",
"def rels\n @rels ||= root.data._rels\n end",
"def query(scope = :all)\n # execute query\n self.semantic_relations.find(scope, :conditions => 'rel_order IS NOT NULL', :order => :rel_order)\n end",
"def each\n enum = @relation.to_enum(:in_batches, of: @of, start: @start, finish: @finish, load: false)\n return enum.each { |relation| yield relation } if block_given?\n enum\n end",
"def relationships\n r = Relationships.new\n r << Relationship.new(cache_definition, PIVOT_TABLE_CACHE_DEFINITION_R, \"../#{cache_definition.pn}\")\n r\n end",
"def items\n @items ||= begin\n offset = @per_page * (@page - 1)\n @relation.limit(@per_page).offset(offset)\n end\n end",
"def relationships\n model.relationships(repository.name)\n end",
"def fulltree_ids(*args)\n Relationship.resource_pairs(fulltree_rels(*args))\n end",
"def source_relations\n query = \"at:link[@rel = '#{Rel[repository.cmis_version][:relationships]}']/@href\"\n link = data.xpath(query, NS::COMBINED)\n if link.length == 1\n link = Internal::Utils.append_parameters(link.text, \"relationshipDirection\" => \"source\", \"includeSubRelationshipTypes\" => true)\n Collection.new(repository, link)\n else\n raise \"Expected exactly 1 relationships link for #{key}, got #{link.length}, are you sure this is a document/folder?\"\n end\n end",
"def getRestaurantTuples()\n\t@cheesies = IDTuple.new('@CheesiesChicago', 'cheesies-pub-and-grub-chicago')\n\t@niusushi = IDTuple.new('@SushiNiu','niu-japanese-fusion-lounge-chicago')\n\t@bob = IDTuple.new('@thebaronbuena','bar-on-buena-chicago')\n\t#@michaels = IDTuple.new('','michaels-pizzeria-and-tavern-chicago-2') #Now here's a business thats not on twitter yet. Womp womp.\n\t@polkstreet = IDTuple.new('@PolkPub', 'polk-street-pub-chicago')\n\t@fatcat = IDTuple.new('@FatCatChicago', 'fat-cat-chicago')\n\treturn [@cheesies, @niusushi, @bob, @polkstreet, @fatcat]\nend",
"def connections(relation_name, limit = 0)\n return Relation.get_connections(relation_name, self.node, limit)\n end",
"def edges\n @pathway.relations.collect do |rel|\n [ rel.node[0], rel.node[1], rel.relation ]\n end\n end",
"def neo4j_query(*args)\n ActiveGraph::Base.query(*args)\n end",
"def fetch_related_oids( oid )\n\t\toid = normalize_oid( oid )\n\t\tself.log.debug \"Fetching OIDs of resources related to %s\" % [ oid ]\n\t\treturn self.search( :criteria => {:relation => oid}, :include_related => true )\n\tend",
"def visit_axiom_relation_operation_binary(binary)\n generator = self.class.visit(binary)\n @name = generator.name\n @from = aliased_subquery(generator)\n generator\n end",
"def nodes\n relation.respond_to?(:nodes) ? relation.nodes : []\n end",
"def queries\n qrs = []\n self.each_query {|qr| qrs << qr }\n qrs\n end",
"def path_ids(*args)\n return call_ancestry_method(:path_ids) if use_ancestry?\n\n Relationship.resource_pairs(path_rels(*args))\n end",
"def relation_complements\n @relation_complements ||= [ self.class ].freeze\n end",
"def query_block(rel, &block)\n rel = yield(rel) if block_given?\n rel.all\n end",
"def relations_by_type\n relations = {}\n for i in 0..(self.relations.count-1)\n type = self.relations(i).type.first\n relations[type] ||= []\n relations[type] += self.relations(i).id\n end\n relations\n end",
"def pairs_for(week)\n pairs = []\n relationships = Relationship.relationships_for(week)\n relationships.each do |relationship|\n pairs << [Person.find_by(id: relationship.partner1_id), Person.find_by(id: relationship.partner2_id)]\n end\n pairs\n end",
"def relationmodule\n @itemtype = params[:itemtype]\n @itemid = params[:itemid]\n \n # Get all current relations\n relatedItems = ElementsAssoc.where('(element1_type = :elm1type AND element1_id = :elm1id) OR (element2_type = :elm1type AND element2_id = :elm1id)', { :elm1type => @itemtype, :elm1id => @itemid })\n @associated = Array.new\n relatedItems.each do |item|\n if item.element1.class.to_s == @itemtype and item.element1[:id].to_s == @itemid\n @associated.push(item.element2.getCat + '_' + item.element2[:id].to_s)\n elsif item.element2.class.to_s == @itemtype and item.element2[:id].to_s == @itemid\n @associated.push(item.element1.getCat + '_' + item.element1[:id].to_s)\n end\n end\n \n self.do_search\n render layout: false\n end",
"def infer_relations_relations\n datasets.each do |gateway, schema|\n schema.each do |name|\n if infer_relation?(gateway, name)\n klass = Relation.build_class(name, adapter: adapter_for(gateway))\n klass.gateway(gateway)\n klass.dataset(name)\n else\n next\n end\n end\n end\n end",
"def get_tuples(params = {})\r\n quantity = calculate_quantity(params[:quantity])\r\n read_tuples(params)\r\n @log.info(\"Found #{@tuples.size} tuples.\")\r\n if @tuples.size > -1 then\r\n @tuples[0..quantity].each do |tuple|\r\n if params[:hostnames]\r\n filter_tuple_by_hostname(tuple, params)\r\n else\r\n add_tuple_to_browsers(tuple)\r\n take_tuple(tuple) if params[:take_all] == true\r\n end\r\n end\r\n end\r\n end",
"def group_relations\n relations = EntityGroupRelationship.where(entity_id: self.id)\n relations.sort_by { |r| r[:position] }\n end",
"def query_all_years(attribute, options)\n query = super(attribute, options)\n\n query.where(\"? = path[?]\", @node.id, @node_index)\n end",
"def tuples_for(arg_)\n objdata_ = _get_objdata(arg_)\n objdata_ ? objdata_[2].keys : nil\n end",
"def and_relation(relation)\n q = all\n raise \"incompatible FROM clauses: #{q.to_sql}; #{relation.to_sql}\" if !q.from_clause.empty? && q.from_clause != relation.from_clause\n raise \"incompatible GROUP BY clauses: #{q.to_sql}; #{relation.to_sql}\" if !q.group_values.empty? && q.group_values != relation.group_values\n\n q = q.select(q.select_values + relation.select_values) if !relation.select_values.empty?\n q = q.from(relation.from_clause.value) if !relation.from_clause.empty?\n q = q.joins(relation.joins_values + q.joins_values) if relation.joins_values.present?\n q = q.where(relation.where_clause.ast) if relation.where_clause.present?\n q = q.group(relation.group_values) if relation.group_values.present?\n q = q.order(relation.order_values) if relation.order_values.present? && !relation.reordering_value\n q = q.reorder(relation.order_values) if relation.order_values.present? && relation.reordering_value\n q\n end",
"def descendant_rels(*args)\n options = args.extract_options!\n rels = relationships.flat_map(&:descendants).uniq\n Relationship.filter_by_resource_type(rels, options)\n end"
] |
[
"0.64533687",
"0.5897142",
"0.58628005",
"0.5677368",
"0.55983114",
"0.555004",
"0.555004",
"0.5404969",
"0.5391372",
"0.53506917",
"0.5291705",
"0.52539784",
"0.52229613",
"0.51843756",
"0.51743656",
"0.51691085",
"0.5137401",
"0.51104915",
"0.50819343",
"0.50815237",
"0.50468194",
"0.50306576",
"0.50285023",
"0.5013627",
"0.5004177",
"0.49946725",
"0.49825084",
"0.49808937",
"0.49674985",
"0.49573094",
"0.49417803",
"0.49349216",
"0.4922973",
"0.4882387",
"0.4878357",
"0.4849209",
"0.48345685",
"0.48326066",
"0.48237476",
"0.4813844",
"0.48093227",
"0.48059556",
"0.48026505",
"0.47901073",
"0.47899348",
"0.47816685",
"0.47796416",
"0.47759828",
"0.47700748",
"0.4756369",
"0.475492",
"0.47524664",
"0.47450423",
"0.4741355",
"0.47399312",
"0.4734702",
"0.47322893",
"0.47315213",
"0.47280338",
"0.4725268",
"0.47233558",
"0.47195435",
"0.4714183",
"0.47114235",
"0.47042465",
"0.4693779",
"0.4693388",
"0.46854144",
"0.46844938",
"0.46832478",
"0.4682374",
"0.46688017",
"0.46659043",
"0.46407655",
"0.46298903",
"0.46040392",
"0.46039885",
"0.4581035",
"0.45788062",
"0.45730612",
"0.4536463",
"0.453561",
"0.4534388",
"0.45340356",
"0.45287737",
"0.4521206",
"0.4519688",
"0.45185846",
"0.45183858",
"0.45047548",
"0.45023477",
"0.4500682",
"0.4492731",
"0.44884148",
"0.4485797",
"0.44815236",
"0.44803587",
"0.4464631",
"0.4453385",
"0.44514224"
] |
0.70719904
|
0
|
Query relation tuples Get all relation tuples that match the query. Only the namespace field is required.
|
def get_relation_tuples_with_http_info(namespace, opts = {})
if @api_client.config.debugging
@api_client.config.logger.debug 'Calling API: ReadApi.get_relation_tuples ...'
end
# verify the required parameter 'namespace' is set
if @api_client.config.client_side_validation && namespace.nil?
fail ArgumentError, "Missing the required parameter 'namespace' when calling ReadApi.get_relation_tuples"
end
# resource path
local_var_path = '/relation-tuples'
# query parameters
query_params = opts[:query_params] || {}
query_params[:'namespace'] = namespace
query_params[:'page_token'] = opts[:'page_token'] if !opts[:'page_token'].nil?
query_params[:'page_size'] = opts[:'page_size'] if !opts[:'page_size'].nil?
query_params[:'object'] = opts[:'object'] if !opts[:'object'].nil?
query_params[:'relation'] = opts[:'relation'] if !opts[:'relation'].nil?
query_params[:'subject_id'] = opts[:'subject_id'] if !opts[:'subject_id'].nil?
query_params[:'subject_set.namespace'] = opts[:'subject_set_namespace'] if !opts[:'subject_set_namespace'].nil?
query_params[:'subject_set.object'] = opts[:'subject_set_object'] if !opts[:'subject_set_object'].nil?
query_params[:'subject_set.relation'] = opts[:'subject_set_relation'] if !opts[:'subject_set_relation'].nil?
# header parameters
header_params = opts[:header_params] || {}
# HTTP header 'Accept' (if needed)
header_params['Accept'] = @api_client.select_header_accept(['application/json'])
# form parameters
form_params = opts[:form_params] || {}
# http body (model)
post_body = opts[:debug_body]
# return_type
return_type = opts[:debug_return_type] || 'GetRelationTuplesResponse'
# auth_names
auth_names = opts[:debug_auth_names] || []
new_options = opts.merge(
:operation => :"ReadApi.get_relation_tuples",
:header_params => header_params,
:query_params => query_params,
:form_params => form_params,
:body => post_body,
:auth_names => auth_names,
:return_type => return_type
)
data, status_code, headers = @api_client.call_api(:GET, local_var_path, new_options)
if @api_client.config.debugging
@api_client.config.logger.debug "API called: ReadApi#get_relation_tuples\nData: #{data.inspect}\nStatus code: #{status_code}\nHeaders: #{headers}"
end
return data, status_code, headers
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def get_relation_tuples(namespace, opts = {})\n data, _status_code, _headers = get_relation_tuples_with_http_info(namespace, opts)\n data\n end",
"def tuples\n results = run_tuples(@last_query_context.to_sparql_select)\n results[\"results\"][\"bindings\"].map do |h|\n h.keys.each do |k|\n h[k.to_sym] = QL.from_tuple_binding(h[k])\n h.delete(k)\n end\n h\n end\n end",
"def tuples\n relation = self.relation\n return relation if materialized?\n\n adapter.read(relation)\n end",
"def relations\n @relations ||= process_rels\n end",
"def relations\n Relations::RelationTraverser.new(@internal_node)\n end",
"def relations\n _relations\n end",
"def relations\n _relations\n end",
"def get_relations\n # Descobrir as relacoes:\n # Orientados\n # Projetos de Pesquisa\n # Producoes Bibliograficas\n # Consultar o RP de cada uma\n # Somar os RPs e retornar\n orientados = self.get_relations_of_orientados\n projetos = self.get_relations_of_projetos_pesquisa\n producoes = self.get_relations_of_producoes_bibliograficas\n\n relations = (orientados | projetos | producoes).delete self.id\n return relations.to_a\n end",
"def traverse(start_elements, relation_filter, stop_filter, from_path = [])\n return [] if from_path.size > 100\n start_elements.each_with_object([]) do |el, relations|\n concrete_rels = concrete_relationships(el, relation_filter, from_path)\n relations.concat(\n concrete_rels,\n *derived_relationship_paths(concrete_rels, relation_filter, stop_filter, from_path)\n )\n end\n end",
"def relation_get_all(user,kind)\n results = []\n Note_Relation.all( :note_id => self.id, :user_id => user.id, :kind => kind ).each do |r|\n results.push(r.value)\n end\n return results\n end",
"def find_relations\n puts '2nd pass: find relations'\n find :relations\n self\n end",
"def relations_for_relation_type(relation_type)\n # Find all the relations that belong to us for this RelationType, ordered by position\n related_ids = relations.where(relation_type_id: relation_type.id).order(:position).select(:related_to_id)\n\n # Construct a query for all these records\n result = relation_type.applies_to.constantize.where(id: related_ids)\n\n # Merge in the relation_filter if it's available\n result = result.merge(relation_filter_for_relation_type(relation_type)) if relation_filter_for_relation_type(relation_type)\n\n # make sure results are in same order as related_ids array (position order)\n if result.present?\n result.where(id: related_ids).order(:position)\n end\n\n result\n end",
"def each_relation\n\t for rel in relations\n\t\tyield(rel)\n\t end\n\tend",
"def relation_all(model)\n validate_model(model)\n model.all\n end",
"def relation_subjects(record)\n query = ActiveFedora::SolrQueryBuilder.construct_query_for_rel(\n [[:has_model, proxy_class.to_rdf_representation], [:proxyFor, record.id]]\n )\n rows = ActiveFedora::SolrService::MAX_ROWS\n ActiveFedora::SolrService.query(query, fl: 'id', rows: rows).map(&:rdf_uri)\n end",
"def relationships\n return [] if empty?\n\n map { |pivot_table| Relationship.new(pivot_table, PIVOT_TABLE_R, \"../#{pivot_table.pn}\") }\n end",
"def roots(relation)\n\t @result_set = plan.query_roots(result_set, relation)\n\t self\n\tend",
"def distinct_relations(from_cache=false)\n # @@neo = Neography::Rest.new\n begin\n if from_cache\n self_node = self.get_node\n @neo.get_node_properties(node1)\n else\n self_node_id = self.get_node_id\n qur = \"start n=node(\"+self_node_id.to_s+\") match n-[r]->() return distinct(type(r)), count(r), r.magnitude;\"\n response = @@neo.execute_query(qur)\n distinct_relations = response[\"data\"]\n return distinct_relations\n end\n rescue Exception\n return nil\n end\n end",
"def relationships\n return [] if empty?\n\n map { |table| Relationship.new(table, TABLE_R, \"../#{table.pn}\") }\n end",
"def read_relation_list(node)\n target_type = Utils.add_namespace(node.attributes['target-type'], Model::NS_REL_1)\n node.elements.each('relation') {|child|\n yield create_relation(child, target_type)\n }\n end",
"def relations_for(node) #:nodoc:\n Enumerator.new do |yielder|\n @description.relations.each do |relation_accessor|\n begin\n relations = relation_accessor[node]\n enumerable = relations.is_a?(Enumerable) ? relations : [relations].compact\n\n enumerable.each { |e| yielder << e unless visited?(e) }\n rescue NoMethodError\n # ignore errors on relation_accessor[node]\n end\n end\n end\n end",
"def map_in_included_relations!(result, query)\n dat = get_relation_data_as_hash(query)\n result.each do |res|\n dat.each do |name, lookup|\n res[name] = lookup[res[@relations[name][:this_key]]]\n if @relations[name][:type] == :has_one\n res[name] = res[name][0] unless res[name].nil?\n else \n res[name] = [] if res[name].nil?\n end\n end\n end\n end",
"def relations\n @relations ||= {}\n end",
"def related_objects(relation = nil, result = nil)\n\t result ||= ValueSet.new\n\t if relation\n\t\tresult.merge(parent_objects(relation).to_value_set)\n\t\tresult.merge(child_objects(relation).to_value_set)\n\t else\n\t\teach_relation { |rel| related_objects(rel, result) }\n\t end\n\t result\n\tend",
"def derived_relations(start_elements, relationship_filter, target_filter, stop_filter = FAIL_ALL)\n traverse(start_elements, relationship_filter, stop_filter)\n .reject(&single_relation_paths) # See #2 above\n .select(&target_relations(target_filter))\n .map(&create_relationship_for_path)\n .uniq { |rel| [rel.type, rel.source, rel.target] }\n end",
"def get_relations_from_api(api=Rosemary::API.new)\n api.get_relations_referring_to_object(type, self.id.to_i)\n end",
"def get_relation(args)\n\tapi_url = \"#{@base_url}/#{args[:collection]}/#{args[:key]}/relations/#{args[:relation]}\"\n\tdo_the_get_call( url: api_url, user: @user )\nend",
"def find_relations\n sql = <<-eos\n SELECT\n tc.constraint_name, tc.table_name, kcu.column_name,\n ccu.table_name AS foreign_table_name,\n ccu.column_name AS foreign_column_name\n FROM\n information_schema.table_constraints AS tc\n JOIN information_schema.key_column_usage AS kcu ON tc.constraint_name = kcu.constraint_name\n JOIN information_schema.constraint_column_usage AS ccu ON ccu.constraint_name = tc.constraint_name\n WHERE constraint_type = 'FOREIGN KEY'\n eos\n @relations = @connection.exec(sql).values\n end",
"def relations; enum_relations.to_a end",
"def get_relations_for_actor(search)\n # Self\n q = \"SELECT *\n WHERE { ?x <http://www.semanticweb.org/ontologies/2011/10/moviesandtv.owl#hasName> \\\"#{search}\\\" }\"\n results = query(q)\n relations = [{:person_uri => results.last[:x], :name => search}]\n # Shows\n q = \"SELECT *\n WHERE { ?x <http://www.semanticweb.org/ontologies/2011/10/moviesandtv.owl#hasName> \\\"#{search}\\\" .\n ?x <http://www.semanticweb.org/ontologies/2011/10/moviesandtv.owl#isActorIn> ?show .\n ?show <http://www.semanticweb.org/ontologies/2011/10/moviesandtv.owl#hasTitle> ?name }\"\n results = query(q)\n results.each do |result|\n relations << {:show_uri => result[:show], :name => result[:name]}\n end\n return relations\n end",
"def printout_relations\n @relations.each do |shape, relations|\n relations.each do |relation|\n relation.each do |key, value|\n puts \"\\\"#{shape}\\\" #{value} \\\"#{key}\\\"\"\n end\n end\n end\n end",
"def relations\n return @artist_data[\"relations\"]\n end",
"def relations\n self.class.relations\n end",
"def relation_types\n Spree::RelationType.where(applies_to: to_relation_name).order(:name)\n end",
"def relation\n relation = nodes.reduce(root) do |a, e|\n a.associations[e.name.key].join(:join, a, e)\n end\n schema.(relation)\n end",
"def all_objects\n objects = []\n each_namespace{|n| objects << n}\n each_namespace{|n| \n n.each_relvar{|r| objects << r}\n }\n each_namespace{|n| \n n.each_relvar{|r| r.each_candidate_key{|k| objects << k}}\n }\n each_namespace{|n| \n n.each_relvar{|r| r.each_foreign_key{|k| objects << k}}\n }\n objects\n end",
"def get_relations(tokens, base_noun_chunks)\n graph = Graph.new\n tokens.each do |t|\n if t[:args]\n t[:args].each do |type, arg|\n graph.add_edge(t[:idx], arg, 1) if arg >= 0\n end\n end\n end\n \n rels = []\n heads = base_noun_chunks.collect{|c| c[:head]}\n base_noun_chunks.combination(2) do |c|\n path = graph.shortest_path(c[0][:head], c[1][:head])\n s = path.shift\n o = path.pop\n rels << [s, path, o] if (path & heads).empty?\n end\n return rels\n end",
"def get_relations(service)\n\t\treturn @transport.get_path(\"relations\",service)\n\tend",
"def get_relations(type, options)\n configure_twitter\n result = case type\n when :follows\n Twitter.friend_ids(options)\n when :followers\n Twitter.follower_ids(options)\n end\n # next_cursor = result.next_cursor\n ids = result.ids\n options[:page] = 0 if options[:page].nil?\n offset = options[:page].to_i * USERS_PER_PAGE\n users = Twitter.users(result.collection[offset, USERS_PER_PAGE+1]).index_by(&:id)\n data = []\n ids.each {|id|\n data << users[id] unless users[id].nil?\n }\n data\n end",
"def related_to(*args)\n@relations = args.map(&:to_s)\nend",
"def list(relation = nil)\n relation ||= apply_filter(self.relation)\n\n apply_paging(\n apply_sorting(relation)\n ).to_a\n end",
"def containing_relation_members\n []\n end",
"def findRelation(query)\n curNode = @head\n while curNode != nil\n if curNode.data.relation == query\n curNode.data.print\n end\n curNode = curNode.next\n end\n end",
"def get_relation(*items, **opt)\n terms = []\n meth = opt.delete(:meth) || \"#{self_class}.#{__method__}\"\n id_opt = opt.extract!(:id_key, :sid_key).transform_values!(&:to_sym)\n id_key = id_opt[:id_key] ||= id_column\n sid_key = id_opt[:sid_key] ||= sid_column\n\n # === Record specifiers\n ids = id_key ? Array.wrap(opt.delete(id_key)) : []\n sids = sid_key ? Array.wrap(opt.delete(sid_key)) : []\n if items.present?\n recs = expand_ids(*items).map { |term| id_term(term, **id_opt) }\n ids = recs.map { |rec| rec[id_key] } + ids if id_key\n sids = recs.map { |rec| rec[sid_key] } + sids if sid_key\n end\n ids = ids.compact_blank!.uniq.presence\n sids = sids.compact_blank!.uniq.presence\n if ids && sids\n terms << sql_terms(id_key => ids, sid_key => sids, join: :or)\n elsif ids\n opt[id_key] = ids\n elsif sids\n opt[sid_key] = sids\n end\n\n # === Sort order\n # Avoid applying a sort order if identifiers were specified or if\n # opt[:sort] was explicitly *nil* or *false*. Permit :asc as shorthand for\n # the default sort order ascending; :desc as shorthand for the default sort\n # order descending.\n if (sort = opt.key?(:sort) ? opt.delete(:sort) : (ids || sids).blank?)\n case sort\n when Hash then col, dir = sort.first\n when TrueClass then col, dir = [nil, nil]\n when /^ASC$/i, /^DESC$/i then col, dir = [nil, sort]\n else col, dir = [sort, nil]\n end\n col ||= implicit_order_column || pagination_column\n dir &&= dir.to_s.upcase\n sort = col && \"#{col} #{dir}\".squish\n Log.info { \"#{meth}: no default sort\" } unless sort\n end\n\n # === Filter by user\n user_opt = opt.extract!(:user, :user_id)\n if user_column && user_opt.present?\n users = user_opt.values.flatten.map { |u| User.id_value(u) }.uniq\n users = users.first if users.size == 1\n terms << sql_terms(user_column => users, join: :or)\n end\n\n # === Filter by state\n state_opt = state_column && opt.extract!(state_column)\n terms << sql_terms(state_opt, join: :or) if state_opt.present?\n\n # === Update time lower bound\n exclusive, inclusive = [opt.delete(:after), opt.delete(:start_date)]\n lower = exclusive || inclusive\n day, month, year = day_string(lower)\n lower = day if day\n if (lower &&= (lower.to_datetime rescue nil))\n lower += 1.month if exclusive && month\n lower += 1.year if exclusive && year\n on_or_after = (exclusive && !month && !year) ? '>' : '>='\n terms << \"updated_at #{on_or_after} '#{lower}'::date\"\n end\n\n # === Update time upper bound\n exclusive, inclusive = [opt.delete(:before), opt.delete(:end_date)]\n upper = exclusive || inclusive\n day, month, year = day_string(upper)\n upper = day if day\n if (upper &&= (upper.to_datetime rescue nil))\n upper += 1.month - 1.day if inclusive && month\n upper += 1.year - 1.day if inclusive && year\n on_or_before = exclusive ? '<' : '<='\n terms << \"updated_at #{on_or_before} '#{upper}'::date\"\n end\n\n # === Record limit/offset\n limit = positive(opt.delete(:limit))\n offset = positive(opt.delete(:offset))\n if offset && pagination_column\n terms << \"#{pagination_column} > #{offset}\"\n elsif offset\n Log.warn { \"#{meth}: pagination not supported\" }\n end\n\n # === Filter by association\n assoc = opt.keys.map(&:to_s).select { |k| k.include?('.') }\n assoc.map! { |k| k.split('.').first.singularize.to_sym }\n assoc = assoc.presence\n\n # === Generate the relation\n query = sql_terms(opt, *terms, join: :and)\n result = assoc ? joins(*assoc).where(query) : where(query)\n result.order!(sort) if sort.present?\n result.limit!(limit) if limit.present?\n result\n end",
"def all\n @data_adapter.relations\n end",
"def xpath_all(pdoc, path, namespace = '')\n if namespace!=\"\"\n return REXML::XPath.match(pdoc, path, namespace)\n else\n return REXML::XPath.match(pdoc, path);\n end\n return []\n end",
"def get_relations(project_id = nil, base_ids = nil, sort = false)\n\t\t_relations = self.subcatrels.in_project(project_id).among_denotations(base_ids)\n\t\tif sort\n\t\t\t_relations.sort{|r1, r2| r1.id <=> r2.id}\n\t\telse\n\t\t\t_relations\n\t\tend\n\tend",
"def get_relations_referring_to_object(type, id)\n api_call_with_type(type, id, \"#{type}/#{id}/relations\")\n end",
"def direct_predicates\n ActiveRDF::Query.new(N::Predicate).distinct(:p).where(self, :p, :o).execute\n end",
"def query_all_years(attribute, options)\n query = super(attribute, options)\n query = query.unscope(where: :context_id)\n where_conds = []\n where_vars = []\n @node_indices_by_context.each do |context_id, node_idx|\n where_conds << \"flows.context_id = ? AND ? = path[?]\"\n where_vars += [context_id, @node.id, node_idx]\n end\n query.where(\n where_conds.join(\" OR \"), *where_vars\n )\n end",
"def queryns \n e = first_element('query')\n if e\n return e.namespace\n else\n return nil\n end\n end",
"def xpath_all(pdoc, path, namespace = '')\n begin\n if namespace != \"\"\n return pdoc.find(path, namespace) if pdoc.find(path, namespace)\n else\n return pdoc.find(path) if pdoc.find(path)\n end\n rescue\n return []\n end\n end",
"def rels\n @traversal_result = :rels\n self\n end",
"def rels(node)\n @batch_inserter.getRelationships(node)\n end",
"def query(*triples)\n bindings = nil\n triples.each do |triple|\n binding_position = {}\n query = []\n triple.each_with_index do |e, i|\n if query_variable?(e)\n binding_position[e] = i\n query << nil\n else\n query << e \n end\n end\n rows = find(*query)\n if bindings.nil?\n bindings = rows.inject([]) do |list, row|\n binding = {}\n binding_position.each do |var, pos|\n binding[var] = row[pos]\n end\n list << binding\n end\n else\n new_binding = []\n bindings.each do |binding|\n rows.each do |row|\n valid_match = true\n temp_binding = binding.dup\n binding_position.each do |var, pos|\n if temp_binding.include?(var)\n valid_match = false if temp_binding[var] != row[pos]\n else\n temp_binding[var] = row[pos]\n end\n end\n new_binding << temp_binding if valid_match\n end\n end\n bindings = new_binding.dup\n end\n bindings\n end\n return bindings\n end",
"def relationships_with_name(input, name)\n input.relations(name)\n end",
"def containing_relation_members \n return [] \n end",
"def tuples\n @storage.enum_for(:each_tuple, @dir)\n end",
"def get_all_related_tables\n tables = [ @table_name ]\n return tables if @relations.nil?\n @relations.each do |n,r|\n tables = tables | get_relation_tables(r)\n end\n tables.sort\n end",
"def all(*vars)\n result = Query.get self, *vars\n result.get_docs\n end",
"def find_related_items(pAuthToken, p_args)\r\n relation_type = p_args[:relation_type]\r\n item_id = p_args[:item_id]\r\n offset = p_args[:offset]\r\n max = p_args[:max]\r\n\r\n unless relation_type.nil?\r\n service_uri = \"#{@@m_service_url.path}#{@@m_service_find_related_items}#{relation_type}&#{item_id}.#{@@m_content_format}?authToken=#{pAuthToken}\"\r\n params = nil\r\n unless offset.nil?\r\n params = \"&offset=#{offset}\"\r\n end\r\n unless max.nil?\r\n if params.nil?\r\n params = \"?max=#{max}\"\r\n else\r\n params += \"&max=#{max}\"\r\n end\r\n end\r\n service_uri += params\r\n end\r\n search_result = SearchResult.new\r\n state_true?(nil, post_request(service_uri, nil).body) {|root|\r\n build_search_result(search_result, root)\r\n }\r\n search_result\r\n end",
"def all\n results = query '*::*::*::*'\n (results.each {|r| yield r}) if block_given?\n results\n end",
"def child_rels(*args)\n options = args.extract_options!\n rels = relationships.flat_map(&:children).uniq\n Relationship.filter_by_resource_type(rels, options)\n end",
"def path_rels(*args)\n options = args.extract_options!\n rel = relationship(:raise_on_multiple => true) # TODO: Handle multiple nodes with a way to detect which node you want\n rels = rel.nil? ? [relationship_for_isolated_root] : rel.path\n Relationship.filter_by_resource_type(rels, options)\n end",
"def sibling_rels(*args)\n options = args.extract_options!\n rels = relationships.flat_map(&:siblings).uniq\n Relationship.filter_by_resource_type(rels, options)\n end",
"def target_relations\n query = \"at:link[@rel = '#{Rel[repository.cmis_version][:relationships]}']/@href\"\n link = data.xpath(query, NS::COMBINED)\n if link.length == 1\n link = Internal::Utils.append_parameters(link.text, \"relationshipDirection\" => \"target\", \"includeSubRelationshipTypes\" => true)\n Collection.new(repository, link)\n else\n raise \"Expected exactly 1 relationships link for #{key}, got #{link.length}, are you sure this is a document/folder?\"\n end\n end",
"def search_graph(query)\n (search_following(query) + search_followers(query)).uniq\n end",
"def relation\n context = Order\n context = context.where(from_group_id: from_group_id) if from_group_id\n context = context.where(to_producer_id: to_producer_id) if to_producer_id\n context = context.where(confirm_before: confirm_before) if confirm_before\n context = context.order(:updated_at)\n context\n end",
"def get_relations(tokens, base_noun_chunks)\n graph = Graph.new\n tokens.each do |t|\n if t[:args]\n t[:args].each do |type, arg|\n graph.add_edge(t[:idx], arg, 1) if arg >= 0\n end\n end\n end\n\n rels = []\n heads = base_noun_chunks.collect{|c| c[:head]}\n base_noun_chunks.combination(2) do |c|\n path = graph.shortest_path(c[0][:head], c[1][:head])\n s = path.shift\n o = path.pop\n rels << [s, path, o] if (path & heads).empty?\n end\n rels\n end",
"def includes(*_args)\n relations.map { |rel| rel[1] }\n end",
"def subtree_rels(*args)\n options = args.extract_options!\n # TODO: make this a single query (vs 3)\n # thus making filter_by_resource_type into a query\n rels = relationships.flat_map(&:subtree).uniq\n rels = [relationship_for_isolated_root] if rels.empty?\n Relationship.filter_by_resource_type(rels, options)\n end",
"def rels\n @rels ||= root.data._rels\n end",
"def query(scope = :all)\n # execute query\n self.semantic_relations.find(scope, :conditions => 'rel_order IS NOT NULL', :order => :rel_order)\n end",
"def each\n enum = @relation.to_enum(:in_batches, of: @of, start: @start, finish: @finish, load: false)\n return enum.each { |relation| yield relation } if block_given?\n enum\n end",
"def relationships\n r = Relationships.new\n r << Relationship.new(cache_definition, PIVOT_TABLE_CACHE_DEFINITION_R, \"../#{cache_definition.pn}\")\n r\n end",
"def items\n @items ||= begin\n offset = @per_page * (@page - 1)\n @relation.limit(@per_page).offset(offset)\n end\n end",
"def relationships\n model.relationships(repository.name)\n end",
"def fulltree_ids(*args)\n Relationship.resource_pairs(fulltree_rels(*args))\n end",
"def source_relations\n query = \"at:link[@rel = '#{Rel[repository.cmis_version][:relationships]}']/@href\"\n link = data.xpath(query, NS::COMBINED)\n if link.length == 1\n link = Internal::Utils.append_parameters(link.text, \"relationshipDirection\" => \"source\", \"includeSubRelationshipTypes\" => true)\n Collection.new(repository, link)\n else\n raise \"Expected exactly 1 relationships link for #{key}, got #{link.length}, are you sure this is a document/folder?\"\n end\n end",
"def getRestaurantTuples()\n\t@cheesies = IDTuple.new('@CheesiesChicago', 'cheesies-pub-and-grub-chicago')\n\t@niusushi = IDTuple.new('@SushiNiu','niu-japanese-fusion-lounge-chicago')\n\t@bob = IDTuple.new('@thebaronbuena','bar-on-buena-chicago')\n\t#@michaels = IDTuple.new('','michaels-pizzeria-and-tavern-chicago-2') #Now here's a business thats not on twitter yet. Womp womp.\n\t@polkstreet = IDTuple.new('@PolkPub', 'polk-street-pub-chicago')\n\t@fatcat = IDTuple.new('@FatCatChicago', 'fat-cat-chicago')\n\treturn [@cheesies, @niusushi, @bob, @polkstreet, @fatcat]\nend",
"def connections(relation_name, limit = 0)\n return Relation.get_connections(relation_name, self.node, limit)\n end",
"def edges\n @pathway.relations.collect do |rel|\n [ rel.node[0], rel.node[1], rel.relation ]\n end\n end",
"def neo4j_query(*args)\n ActiveGraph::Base.query(*args)\n end",
"def fetch_related_oids( oid )\n\t\toid = normalize_oid( oid )\n\t\tself.log.debug \"Fetching OIDs of resources related to %s\" % [ oid ]\n\t\treturn self.search( :criteria => {:relation => oid}, :include_related => true )\n\tend",
"def visit_axiom_relation_operation_binary(binary)\n generator = self.class.visit(binary)\n @name = generator.name\n @from = aliased_subquery(generator)\n generator\n end",
"def nodes\n relation.respond_to?(:nodes) ? relation.nodes : []\n end",
"def queries\n qrs = []\n self.each_query {|qr| qrs << qr }\n qrs\n end",
"def path_ids(*args)\n return call_ancestry_method(:path_ids) if use_ancestry?\n\n Relationship.resource_pairs(path_rels(*args))\n end",
"def relation_complements\n @relation_complements ||= [ self.class ].freeze\n end",
"def query_block(rel, &block)\n rel = yield(rel) if block_given?\n rel.all\n end",
"def relations_by_type\n relations = {}\n for i in 0..(self.relations.count-1)\n type = self.relations(i).type.first\n relations[type] ||= []\n relations[type] += self.relations(i).id\n end\n relations\n end",
"def pairs_for(week)\n pairs = []\n relationships = Relationship.relationships_for(week)\n relationships.each do |relationship|\n pairs << [Person.find_by(id: relationship.partner1_id), Person.find_by(id: relationship.partner2_id)]\n end\n pairs\n end",
"def relationmodule\n @itemtype = params[:itemtype]\n @itemid = params[:itemid]\n \n # Get all current relations\n relatedItems = ElementsAssoc.where('(element1_type = :elm1type AND element1_id = :elm1id) OR (element2_type = :elm1type AND element2_id = :elm1id)', { :elm1type => @itemtype, :elm1id => @itemid })\n @associated = Array.new\n relatedItems.each do |item|\n if item.element1.class.to_s == @itemtype and item.element1[:id].to_s == @itemid\n @associated.push(item.element2.getCat + '_' + item.element2[:id].to_s)\n elsif item.element2.class.to_s == @itemtype and item.element2[:id].to_s == @itemid\n @associated.push(item.element1.getCat + '_' + item.element1[:id].to_s)\n end\n end\n \n self.do_search\n render layout: false\n end",
"def infer_relations_relations\n datasets.each do |gateway, schema|\n schema.each do |name|\n if infer_relation?(gateway, name)\n klass = Relation.build_class(name, adapter: adapter_for(gateway))\n klass.gateway(gateway)\n klass.dataset(name)\n else\n next\n end\n end\n end\n end",
"def get_tuples(params = {})\r\n quantity = calculate_quantity(params[:quantity])\r\n read_tuples(params)\r\n @log.info(\"Found #{@tuples.size} tuples.\")\r\n if @tuples.size > -1 then\r\n @tuples[0..quantity].each do |tuple|\r\n if params[:hostnames]\r\n filter_tuple_by_hostname(tuple, params)\r\n else\r\n add_tuple_to_browsers(tuple)\r\n take_tuple(tuple) if params[:take_all] == true\r\n end\r\n end\r\n end\r\n end",
"def group_relations\n relations = EntityGroupRelationship.where(entity_id: self.id)\n relations.sort_by { |r| r[:position] }\n end",
"def query_all_years(attribute, options)\n query = super(attribute, options)\n\n query.where(\"? = path[?]\", @node.id, @node_index)\n end",
"def tuples_for(arg_)\n objdata_ = _get_objdata(arg_)\n objdata_ ? objdata_[2].keys : nil\n end",
"def and_relation(relation)\n q = all\n raise \"incompatible FROM clauses: #{q.to_sql}; #{relation.to_sql}\" if !q.from_clause.empty? && q.from_clause != relation.from_clause\n raise \"incompatible GROUP BY clauses: #{q.to_sql}; #{relation.to_sql}\" if !q.group_values.empty? && q.group_values != relation.group_values\n\n q = q.select(q.select_values + relation.select_values) if !relation.select_values.empty?\n q = q.from(relation.from_clause.value) if !relation.from_clause.empty?\n q = q.joins(relation.joins_values + q.joins_values) if relation.joins_values.present?\n q = q.where(relation.where_clause.ast) if relation.where_clause.present?\n q = q.group(relation.group_values) if relation.group_values.present?\n q = q.order(relation.order_values) if relation.order_values.present? && !relation.reordering_value\n q = q.reorder(relation.order_values) if relation.order_values.present? && relation.reordering_value\n q\n end",
"def descendant_rels(*args)\n options = args.extract_options!\n rels = relationships.flat_map(&:descendants).uniq\n Relationship.filter_by_resource_type(rels, options)\n end"
] |
[
"0.70719904",
"0.5897142",
"0.58628005",
"0.5677368",
"0.55983114",
"0.555004",
"0.555004",
"0.5404969",
"0.5391372",
"0.53506917",
"0.5291705",
"0.52539784",
"0.52229613",
"0.51843756",
"0.51743656",
"0.51691085",
"0.5137401",
"0.51104915",
"0.50819343",
"0.50815237",
"0.50468194",
"0.50306576",
"0.50285023",
"0.5013627",
"0.5004177",
"0.49946725",
"0.49825084",
"0.49808937",
"0.49674985",
"0.49573094",
"0.49417803",
"0.49349216",
"0.4922973",
"0.4882387",
"0.4878357",
"0.4849209",
"0.48345685",
"0.48326066",
"0.48237476",
"0.4813844",
"0.48093227",
"0.48059556",
"0.48026505",
"0.47901073",
"0.47899348",
"0.47816685",
"0.47796416",
"0.47759828",
"0.47700748",
"0.4756369",
"0.475492",
"0.47524664",
"0.47450423",
"0.4741355",
"0.47399312",
"0.4734702",
"0.47322893",
"0.47315213",
"0.47280338",
"0.4725268",
"0.47233558",
"0.47195435",
"0.4714183",
"0.47114235",
"0.47042465",
"0.4693779",
"0.4693388",
"0.46854144",
"0.46844938",
"0.46832478",
"0.4682374",
"0.46688017",
"0.46659043",
"0.46407655",
"0.46298903",
"0.46040392",
"0.46039885",
"0.4581035",
"0.45788062",
"0.45730612",
"0.4536463",
"0.453561",
"0.4534388",
"0.45340356",
"0.45287737",
"0.4521206",
"0.4519688",
"0.45185846",
"0.45183858",
"0.45047548",
"0.45023477",
"0.4500682",
"0.4492731",
"0.44884148",
"0.4485797",
"0.44815236",
"0.44803587",
"0.4464631",
"0.4453385",
"0.44514224"
] |
0.64533687
|
1
|
Check a relation tuple To learn how relation tuples and the check works, head over to [the documentation](../concepts/relationtuples.mdx).
|
def post_check(opts = {})
data, _status_code, _headers = post_check_with_http_info(opts)
data
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def valid_relation_literal!(heading, tuples, raise_on_error = true)\n if Array === tuples\n invalid = tuples.find{|t| !valid_tuple_literal?(heading, t)}\n if invalid\n raise_on_error ? raise(::Veritas::RelationMismatchError, \"Invalid tuple #{invalid.inspect} for heading #{heading}\", caller) : false\n else\n tuples\n end\n else\n raise_on_error ? raise(::Veritas::RelationMismatchError, \"Invalid relation literal #{tuples.inspect}\", caller) : false \n end\n end",
"def check_for_tuple(tuple)\n ltuple = @left_pattern.map {|i| tuple[i]}\n if @left.check_for_tuple(ltuple)\n rtuple = @right_pattern.map {|i| tuple[i]}\n if @right.check_for_tuple(rtuple)\n return true\n end\n end\n return false\n end",
"def contains?(tuple)\n case tuple\n when Hash\n eql_hash = {}\n relvar_def.primary_key.attributes.each{|a| eql_hash[a.name] = tuple[a.name]}\n inside_tuple = underlying_table.filter(eql_hash).first\n return inside_tuple === tuple\n else\n raise ArgumentError, \"Unable to check inclusion of #{tuple} inside a relation variable\"\n end\n end",
"def valid_tuple_literal!(heading, tuple, raise_on_error = true)\n if Hash===tuple and tuple.size == heading.to_ary.size\n ok = heading.all?{|a| tuple.key?(a.name) and a.valid_value?(tuple[a.name])}\n return tuple if ok\n end\n raise_on_error ? raise(ArgumentError, \"Invalid tuple #{tuple} for #{heading}\", caller) : false\n end",
"def _tuple2tuple(tuple)\n end",
"def tuple_plan()\n raise NotTupleValuedException.new(self) unless tuple_valued?\n @definition.type.tuple.plan\n end",
"def create_relation_tuple(opts = {})\n data, _status_code, _headers = create_relation_tuple_with_http_info(opts)\n data\n end",
"def check_relation(relation)\n (relation - reflection.map{|r| r[:rel] }).empty?\n end",
"def relation\n # -> uncomment the next line to manually enable rule tracing\n # trace_in( __method__, 28 )\n\n\n value = nil\n\n\n type = nil\n a = nil\n b = nil\n\n\n begin\n # at line 178:5: a= add (type= ( '==' | '!=' | '<' | '<=' | '>=' | '>' ) b= relation )?\n @state.following.push( TOKENS_FOLLOWING_add_IN_relation_1322 )\n a = add\n @state.following.pop\n\n # syntactic predicate action gate test\n if @state.backtracking == 0\n # --> action\n value = a \n # <-- action\n end\n\n # at line 179:3: (type= ( '==' | '!=' | '<' | '<=' | '>=' | '>' ) b= relation )?\n alt_44 = 2\n look_44_0 = @input.peek( 1 )\n\n if ( look_44_0 == T__31 || look_44_0.between?( T__47, T__48 ) || look_44_0.between?( T__50, T__52 ) )\n alt_44 = 1\n end\n case alt_44\n when 1\n # at line 179:6: type= ( '==' | '!=' | '<' | '<=' | '>=' | '>' ) b= relation\n type = @input.look\n\n if @input.peek(1) == T__31 || @input.peek( 1 ).between?( T__47, T__48 ) || @input.peek( 1 ).between?( T__50, T__52 )\n @input.consume\n @state.error_recovery = false\n\n else\n @state.backtracking > 0 and raise( ANTLR3::Error::BacktrackingFailed )\n\n\n mse = MismatchedSet( nil )\n raise mse\n\n end\n\n\n @state.following.push( TOKENS_FOLLOWING_relation_IN_relation_1359 )\n b = relation\n @state.following.pop\n\n # syntactic predicate action gate test\n if @state.backtracking == 0\n # --> action\n value = DoubleOperandExpressionEval.new(type.text, a, b) \n # <-- action\n end\n\n\n end\n\n rescue ANTLR3::Error::RecognitionError => re\n report_error(re)\n recover(re)\n\n ensure\n # -> uncomment the next line to manually enable rule tracing\n # trace_out( __method__, 28 )\n\n\n end\n\n return value\n end",
"def eval_tuple(node, env, stack_trace)\n\n\tevaluated_nodes = []\n\tnode.values.each{ |sub_node|\n\t\tsub_node_value, _ = eval_node_under(sub_node, env, stack_trace) \n\t\tevaluated_nodes.append(sub_node_value)\n\t}\n\tif evaluated_nodes.length == 1\n\t\treturn evaluated_nodes[0], env\n\telse\n\t\treturn Tuple.new(node.line, node.col, evaluated_nodes)\n\tend\nend",
"def triangle?((x1, y1), (x2, y2), (x3, y3))\n return false if (x1, y1) == (x2, y2) || (x1, y1) == (x3, y3) || (x2, y2) == (x3, y3)\n return false \n true\nend",
"def patch_relation_tuples(opts = {})\n patch_relation_tuples_with_http_info(opts)\n nil\n end",
"def is_meta_tuple? tuple\n if meta_subspace\n meta_subspace === tuple\n else\n # meta_subspace hasn't arrived yet, so use approximation\n tuple.kind_of? Hash and tuple.key? client.tupelo_meta_key and\n tuple[client.tupelo_meta_key] == \"subspace\"\n end\n end",
"def t\n BERT::Tuple\nend",
"def activity_relations?(*args)\n activity_relations(*args).any?\n end",
"def tuples; end",
"def tuples; end",
"def assert_working_relations( model, relations )\n failed = []\n relations.each do | name |\n begin\n model.send( name )\n rescue Exception => e\n failed << \"#{name} - #{e}\"\n end\n end\n if failed.empty?\n assert true\n else\n assert false, failed.join('; ')\n end\n end",
"def check_relation(x)\n h, qx = compute_q(x)\n return if qx == 0\n es, l = exponent_bitvector(qx)\n\n # discard this x if the residue L is too big\n return if l > @pmax ** @t\n\n if l == 1\n # complete relation found:\n # Q(x) = p0^e0 * p1^e1 * ... * pk^ek (pi in the factor base)\n qx_vec = Hash.new(0)\n PrimeFactorization.prime_factorization(qx) {|p, e| qx_vec[p] += e }\n collect_relation(es, h, qx_vec)\n\n elsif @incomplete_relations[l]\n # large prime procedure:\n # make a complete relation by multiplying two incomplete relations\n es2, h2, qx2 = @incomplete_relations[l]\n\n # XXX: use FactoredInteger\n qx_vec = Hash.new(0)\n PrimeFactorization.prime_factorization(qx) {|p, e| qx_vec[p] += e }\n PrimeFactorization.prime_factorization(qx2) {|p, e| qx_vec[p] += e }\n\n collect_relation(es ^ es2, h * h2 % @kn, qx_vec)\n\n else\n @incomplete_relations[l] = [es, h, qx]\n end\n end",
"def verify_uniquness!(tuple)\n stream = tuple[:stream]\n attrs = %i[position event_id]\n attrs.delete(:position) if Stream.new(stream).global?\n\n attrs.each do |key|\n next if key == :position && tuple[key].nil?\n next if restrict(:stream => stream, key => tuple.fetch(key)).none?\n\n raise TupleUniquenessError.public_send(:\"for_stream_and_#{key}\", stream, tuple.fetch(key))\n end\n end",
"def create_relation_tuple_with_http_info(opts = {})\n if @api_client.config.debugging\n @api_client.config.logger.debug 'Calling API: WriteApi.create_relation_tuple ...'\n end\n # resource path\n local_var_path = '/relation-tuples'\n\n # query parameters\n query_params = opts[:query_params] || {}\n\n # header parameters\n header_params = opts[:header_params] || {}\n # HTTP header 'Accept' (if needed)\n header_params['Accept'] = @api_client.select_header_accept(['application/json'])\n # HTTP header 'Content-Type'\n header_params['Content-Type'] = @api_client.select_header_content_type(['application/json'])\n\n # form parameters\n form_params = opts[:form_params] || {}\n\n # http body (model)\n post_body = opts[:debug_body] || @api_client.object_to_http_body(opts[:'payload'])\n\n # return_type\n return_type = opts[:debug_return_type] || 'RelationQuery'\n\n # auth_names\n auth_names = opts[:debug_auth_names] || []\n\n new_options = opts.merge(\n :operation => :\"WriteApi.create_relation_tuple\",\n :header_params => header_params,\n :query_params => query_params,\n :form_params => form_params,\n :body => post_body,\n :auth_names => auth_names,\n :return_type => return_type\n )\n\n data, status_code, headers = @api_client.call_api(:PUT, local_var_path, new_options)\n if @api_client.config.debugging\n @api_client.config.logger.debug \"API called: WriteApi#create_relation_tuple\\nData: #{data.inspect}\\nStatus code: #{status_code}\\nHeaders: #{headers}\"\n end\n return data, status_code, headers\n end",
"def delete_relation_tuple_with_http_info(namespace, object, relation, opts = {})\n if @api_client.config.debugging\n @api_client.config.logger.debug 'Calling API: WriteApi.delete_relation_tuple ...'\n end\n # verify the required parameter 'namespace' is set\n if @api_client.config.client_side_validation && namespace.nil?\n fail ArgumentError, \"Missing the required parameter 'namespace' when calling WriteApi.delete_relation_tuple\"\n end\n # verify the required parameter 'object' is set\n if @api_client.config.client_side_validation && object.nil?\n fail ArgumentError, \"Missing the required parameter 'object' when calling WriteApi.delete_relation_tuple\"\n end\n # verify the required parameter 'relation' is set\n if @api_client.config.client_side_validation && relation.nil?\n fail ArgumentError, \"Missing the required parameter 'relation' when calling WriteApi.delete_relation_tuple\"\n end\n # resource path\n local_var_path = '/relation-tuples'\n\n # query parameters\n query_params = opts[:query_params] || {}\n query_params[:'namespace'] = namespace\n query_params[:'object'] = object\n query_params[:'relation'] = relation\n query_params[:'subject_id'] = opts[:'subject_id'] if !opts[:'subject_id'].nil?\n query_params[:'subject_set.namespace'] = opts[:'subject_set_namespace'] if !opts[:'subject_set_namespace'].nil?\n query_params[:'subject_set.object'] = opts[:'subject_set_object'] if !opts[:'subject_set_object'].nil?\n query_params[:'subject_set.relation'] = opts[:'subject_set_relation'] if !opts[:'subject_set_relation'].nil?\n\n # header parameters\n header_params = opts[:header_params] || {}\n # HTTP header 'Accept' (if needed)\n header_params['Accept'] = @api_client.select_header_accept(['application/json'])\n\n # form parameters\n form_params = opts[:form_params] || {}\n\n # http body (model)\n post_body = opts[:debug_body]\n\n # return_type\n return_type = opts[:debug_return_type]\n\n # auth_names\n auth_names = opts[:debug_auth_names] || []\n\n new_options = opts.merge(\n :operation => :\"WriteApi.delete_relation_tuple\",\n :header_params => header_params,\n :query_params => query_params,\n :form_params => form_params,\n :body => post_body,\n :auth_names => auth_names,\n :return_type => return_type\n )\n\n data, status_code, headers = @api_client.call_api(:DELETE, local_var_path, new_options)\n if @api_client.config.debugging\n @api_client.config.logger.debug \"API called: WriteApi#delete_relation_tuple\\nData: #{data.inspect}\\nStatus code: #{status_code}\\nHeaders: #{headers}\"\n end\n return data, status_code, headers\n end",
"def tuple_extract(*args, &bl)\n alf_connection.tuple_extract(*args, &bl)\n end",
"def relation(relation)\n true\n end",
"def tuple_valued?()\n @definition.type.named_type?\n end",
"def delete_relation_tuple(namespace, object, relation, opts = {})\n delete_relation_tuple_with_http_info(namespace, object, relation, opts)\n nil\n end",
"def check_relation_with friend_id\n response = {\n message: \"Successfully fecth user relation\",\n following: self.active_relationships.where(follower_id: friend_id).present?,\n follower: self.passive_relationships.where(follower_id: friend_id).present?\n }\n end",
"def is_valid_relation?(parent, child, require_both = self.require_both)\n # depending upon the\n require_both ? is_valid_down_relation?(parent, child) && is_valid_up_relation?(parent, child)\n : is_valid_down_relation?(parent, child) || is_valid_up_relation?(parent, child)\n end",
"def parse_tuple cur_tok\n\t\tcur_ast = Tuple.new(cur_tok.line, cur_tok.col, splitter( nil, \")\", \",\")) # start=nil because \"(\" token has already been read.\n\n\t\t# Use look ahead to see if the tuple needs to be a child node of an operation.\n\t\tif is_operation \n\t\t\treturn parse_operation(cur_ast, true)\n\t\telse\n\t\t\treturn cur_ast \n\t\tend\n\tend",
"def deduce\n each_vertex do |v1|\n each_vertex do |v2|\n each_vertex do |v3|\n \n if self[v1,v2] == :provable_true and self[v2,v3] == :provable_true\n add_assertion(v1, v3, :provable_true)\n end\n \n if self[v2,v1] == :provable_false and self[v2,v3] == :provable_true\n add_assertion(v3, v1, :provable_false)\n end\n \n if self[v1,v2] == :provable_true and self[v3,v2] == :provable_false\n add_assertion(v3, v1, :provable_false)\n end\n \n break if @contradiction\n end\n end\n end\n end",
"def cast_tuple\n <<-CODE\n t1 = stack_pop();\n if(REFERENCE_P(t1) && object_kind_of_p(state, t1, global->array)) {\n j = N2I(array_get_total(t1));\n t2 = tuple_new(state, j);\n\n for(k = 0; k < j; k++) {\n tuple_put(state, t2, k, array_get(state, t1, k));\n }\n t1 = t2;\n } else if(!REFERENCE_P(t1) || !object_kind_of_p(state, t1, global->tuple)) {\n t2 = tuple_new(state, 1);\n tuple_put(state, t2, 0, t1);\n t1 = t2;\n }\n stack_push(t1);\n CODE\n end",
"def assert_relations_are_equal(a, b)\n assert_not_nil a, \"first relation is not allowed to be nil\"\n assert_not_nil b, \"second relation #{a.id} is not allowed to be nil\"\n assert_equal a.id, b.id, \"relation IDs\"\n assert_equal a.changeset_id, b.changeset_id, \"changeset ID on relation #{a.id}\"\n assert_equal a.visible, b.visible, \"visible on relation #{a.id}, #{a.visible.inspect} != #{b.visible.inspect}\"\n assert_equal a.version, b.version, \"version on relation #{a.id}\"\n assert_equal a.tags, b.tags, \"tags on relation #{a.id}\"\n assert_equal a.members, b.members, \"member references on relation #{a.id}\"\n end",
"def tuple(*elements)\n Tuple.new(*elements)\n end",
"def cmp_tuples(left, right)\n reduce(0) do |cmp, direction|\n break cmp if cmp.nonzero?\n direction.call(left, right)\n end\n end",
"def get_relation_tuples(namespace, opts = {})\n data, _status_code, _headers = get_relation_tuples_with_http_info(namespace, opts)\n data\n end",
"def func_le(args)\n p1 = _eval(car(args))\n p2 = _eval(car(cdr(args)))\n\n if p1.type != LObject::OBJ_INTEGER and p2.type != LObject::OBJ_INTEGER\n if @lint\n Error.warn(\"warning: comparison with a non integer operand\")\n end\n return @o_man.nil\n end\n\n if p1.value.i <= p2.value.i\n return @o_man.t\n end\n return @o_man.nil\n end",
"def tuple(*args)\n \"( #{args.map { |b| if b.nil? then 'NULL' else b end }.join(', ')} )\"\nend",
"def relation\n # -> uncomment the next line to manually enable rule tracing\n # trace_in( __method__, 55 )\n return_value = RelationReturnValue.new\n\n # $rule.start = the first token seen before matching\n return_value.start = @input.look\n\n root_0 = nil\n shift249 = nil\n relation_op250 = nil\n shift251 = nil\n\n\n begin\n root_0 = @adaptor.create_flat_list\n\n\n # at line 609:5: shift ( relation_op shift )*\n @state.following.push( TOKENS_FOLLOWING_shift_IN_relation_4033 )\n shift249 = shift\n @state.following.pop\n if @state.backtracking == 0\n @adaptor.add_child( root_0, shift249.tree )\n end\n # at line 609:11: ( relation_op shift )*\n while true # decision 59\n alt_59 = 2\n look_59_0 = @input.peek( 1 )\n\n if ( look_59_0 == GEQ || look_59_0 == GREATER || look_59_0 == INSTANCEOF || look_59_0 == LEQ || look_59_0 == LESS )\n alt_59 = 1\n elsif ( look_59_0 == IN ) and ( ( !( @InFor_stack.last && @InFor_stack.last.active ) ) )\n alt_59 = 1\n\n end\n case alt_59\n when 1\n # at line 609:14: relation_op shift\n @state.following.push( TOKENS_FOLLOWING_relation_op_IN_relation_4038 )\n relation_op250 = relation_op\n @state.following.pop\n if @state.backtracking == 0\n root_0 = @adaptor.become_root( relation_op250.tree, root_0 )\n end\n @state.following.push( TOKENS_FOLLOWING_shift_IN_relation_4042 )\n shift251 = shift\n @state.following.pop\n if @state.backtracking == 0\n @adaptor.add_child( root_0, shift251.tree )\n end\n\n else\n break # out of loop for decision 59\n end\n end # loop for decision 59\n # - - - - - - - rule clean up - - - - - - - -\n return_value.stop = @input.look( -1 )\n\n if @state.backtracking == 0\n\n return_value.tree = @adaptor.rule_post_processing( root_0 )\n @adaptor.set_token_boundaries( return_value.tree, return_value.start, return_value.stop )\n\n end\n rescue ANTLR3::Error::RecognitionError => re\n report_error(re)\n recover(re)\n return_value.tree = @adaptor.create_error_node( @input, return_value.start, @input.look(-1), re )\n\n ensure\n # -> uncomment the next line to manually enable rule tracing\n # trace_out( __method__, 55 )\n\n end\n \n return return_value\n end",
"def get_relation_tuples_with_http_info(namespace, opts = {})\n if @api_client.config.debugging\n @api_client.config.logger.debug 'Calling API: ReadApi.get_relation_tuples ...'\n end\n # verify the required parameter 'namespace' is set\n if @api_client.config.client_side_validation && namespace.nil?\n fail ArgumentError, \"Missing the required parameter 'namespace' when calling ReadApi.get_relation_tuples\"\n end\n # resource path\n local_var_path = '/relation-tuples'\n\n # query parameters\n query_params = opts[:query_params] || {}\n query_params[:'namespace'] = namespace\n query_params[:'page_token'] = opts[:'page_token'] if !opts[:'page_token'].nil?\n query_params[:'page_size'] = opts[:'page_size'] if !opts[:'page_size'].nil?\n query_params[:'object'] = opts[:'object'] if !opts[:'object'].nil?\n query_params[:'relation'] = opts[:'relation'] if !opts[:'relation'].nil?\n query_params[:'subject_id'] = opts[:'subject_id'] if !opts[:'subject_id'].nil?\n query_params[:'subject_set.namespace'] = opts[:'subject_set_namespace'] if !opts[:'subject_set_namespace'].nil?\n query_params[:'subject_set.object'] = opts[:'subject_set_object'] if !opts[:'subject_set_object'].nil?\n query_params[:'subject_set.relation'] = opts[:'subject_set_relation'] if !opts[:'subject_set_relation'].nil?\n\n # header parameters\n header_params = opts[:header_params] || {}\n # HTTP header 'Accept' (if needed)\n header_params['Accept'] = @api_client.select_header_accept(['application/json'])\n\n # form parameters\n form_params = opts[:form_params] || {}\n\n # http body (model)\n post_body = opts[:debug_body]\n\n # return_type\n return_type = opts[:debug_return_type] || 'GetRelationTuplesResponse'\n\n # auth_names\n auth_names = opts[:debug_auth_names] || []\n\n new_options = opts.merge(\n :operation => :\"ReadApi.get_relation_tuples\",\n :header_params => header_params,\n :query_params => query_params,\n :form_params => form_params,\n :body => post_body,\n :auth_names => auth_names,\n :return_type => return_type\n )\n\n data, status_code, headers = @api_client.call_api(:GET, local_var_path, new_options)\n if @api_client.config.debugging\n @api_client.config.logger.debug \"API called: ReadApi#get_relation_tuples\\nData: #{data.inspect}\\nStatus code: #{status_code}\\nHeaders: #{headers}\"\n end\n return data, status_code, headers\n end",
"def valid_triangle? triple\n a, b, c = triple\n return false if a + b <= c\n return false if a + c <= b\n return false if b + c <= a\n\n true\nend",
"def tuples\n relation = self.relation\n return relation if materialized?\n\n adapter.read(relation)\n end",
"def create_tuple!(attributes)\n # split the attrs (needs DRYing)\n rel_attrs = {}\n attributes.each do |key, value|\n if key.to_s[0..3] == 'rel_'\n rel_attrs[key] = attributes.delete(key)\n end\n end\n\n # find or create the one node\n if target_node = @target_class.find(attributes[:id])\n target_node.update_attributes!(attributes)\n else\n target_node = @target_class.create!(attributes)\n end\n\n self.relate(target_node, rel_attrs)\n\n # finally, make and return the tuple\n #Tuple.new(@node, target_node)\n #Tuple.new(@node, target_node)\n\n end",
"def grid_operation_checker(candidate)\n candidate.node_type == :call and\n candidate.sons.first.node_type == :lval and\n (candidate.grid_objects.map {|g_obj| g_obj.name}).include? candidate.sons.first.name\n end",
"def tuple_print(results)\n results.each do |tuple|\n puts tuple\n end\nend",
"def check_permission_or_error_with_http_info(opts = {})\n if @api_client.config.debugging\n @api_client.config.logger.debug 'Calling API: PermissionApi.check_permission_or_error ...'\n end\n # resource path\n local_var_path = '/relation-tuples/check'\n\n # query parameters\n query_params = opts[:query_params] || {}\n query_params[:'namespace'] = opts[:'namespace'] if !opts[:'namespace'].nil?\n query_params[:'object'] = opts[:'object'] if !opts[:'object'].nil?\n query_params[:'relation'] = opts[:'relation'] if !opts[:'relation'].nil?\n query_params[:'subject_id'] = opts[:'subject_id'] if !opts[:'subject_id'].nil?\n query_params[:'subject_set.namespace'] = opts[:'subject_set_namespace'] if !opts[:'subject_set_namespace'].nil?\n query_params[:'subject_set.object'] = opts[:'subject_set_object'] if !opts[:'subject_set_object'].nil?\n query_params[:'subject_set.relation'] = opts[:'subject_set_relation'] if !opts[:'subject_set_relation'].nil?\n query_params[:'max-depth'] = opts[:'max_depth'] if !opts[:'max_depth'].nil?\n\n # header parameters\n header_params = opts[:header_params] || {}\n # HTTP header 'Accept' (if needed)\n header_params['Accept'] = @api_client.select_header_accept(['application/json'])\n\n # form parameters\n form_params = opts[:form_params] || {}\n\n # http body (model)\n post_body = opts[:debug_body]\n\n # return_type\n return_type = opts[:debug_return_type] || 'CheckPermissionResult'\n\n # auth_names\n auth_names = opts[:debug_auth_names] || ['oryAccessToken']\n\n new_options = opts.merge(\n :operation => :\"PermissionApi.check_permission_or_error\",\n :header_params => header_params,\n :query_params => query_params,\n :form_params => form_params,\n :body => post_body,\n :auth_names => auth_names,\n :return_type => return_type\n )\n\n data, status_code, headers = @api_client.call_api(:GET, local_var_path, new_options)\n if @api_client.config.debugging\n @api_client.config.logger.debug \"API called: PermissionApi#check_permission_or_error\\nData: #{data.inspect}\\nStatus code: #{status_code}\\nHeaders: #{headers}\"\n end\n return data, status_code, headers\n end",
"def is_triplet(n1,n2,n3)\n n1**2 + n2**2 == n3**2 or n1**2 + n3**2 == n2**2 or n2**2 + n3**2 == n1**2\nend",
"def is_pythag_triplet?(a,b,c)\n a < b && b < c\nend",
"def create\n @tuple = Tuple.new(params[:tuple])\n\n respond_to do |format|\n if @tuple.save\n format.html { redirect_to(@tuple, :notice => 'Tuple was successfully created.') }\n format.xml { render :xml => @tuple, :status => :created, :location => @tuple }\n else\n format.html { render :action => \"new\" }\n format.xml { render :xml => @tuple.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def _add_tuple(objdata_, tuple_) # :nodoc:\n return false if @tuples.has_key?(tuple_)\n @tuples[tuple_] = objdata_\n tupcats_ = []\n @categories.each do |category_, catdata_|\n if Utils.matches?(catdata_[0], tuple_)\n index_ = catdata_[1].map{ |i_| tuple_[i_] }\n ((@catdata[category_] ||= {})[index_] ||= {})[tuple_] = objdata_\n tupcats_ << category_\n end\n end\n objdata_[2][tuple_] = tupcats_\n true\n end",
"def has_edge?(vertex1, vertex2)\n\tend",
"def isTriangle(a,b,c)\n a + b > c && b + c > a && a + c > b\nend",
"def isTriangle(a,b,c)\n a + b > c && b + c > a && a + c > b\nend",
"def lift_tuple(tuple)\n if Pione::TupleSpace[tuple.first]\n if pos = Pione::TupleSpace[tuple.first].location_position\n if new_location = lift_location(tuple[pos])\n tuple = tuple.clone\n tuple[pos] = new_location\n end\n end\n end\n return tuple\n end",
"def _remove_tuple(objdata_, tuple_) # :nodoc:\n tupcats_ = objdata_[2][tuple_]\n return false unless tupcats_\n @tuples.delete(tuple_)\n tupcats_.each do |cat_|\n index_ = @categories[cat_][1].map{ |i_| tuple_[i_] }\n @catdata[cat_][index_].delete(tuple_)\n end\n objdata_[2].delete(tuple_)\n true\n end",
"def validate_relation(klass, name, options = {})\n [name, \"#{name}?\".to_sym, \"#{name}=\".to_sym].each do |n|\n if Mongoid.destructive_fields.include?(n)\n raise Errors::InvalidRelation.new(klass, n)\n end\n end\n end",
"def call(relation)\n relation\n end",
"def valid?\n # Check tx = <lx, rx>\n unless FIELD.mod(lx.zip(rx).map { |a, b| FIELD.mod(a * b) }.sum) == tx\n return false\n end\n\n return false unless valid_poly_t?\n\n p =\n vec_g\n .zip(lx)\n .map { |a, b| a * b }\n .zip(vec_h2.zip(rx).map { |a, b| a * b })\n .map { |a, b| a + b }\n .sum(GROUP.infinity)\n\n p1 == p2 && p2 == p\n end",
"def op_rel!\n # -> uncomment the next line to manually enable rule tracing\n # trace_in( __method__, 17 )\n\n\n\n type = OP_REL\n channel = ANTLR3::DEFAULT_CHANNEL\n # - - - - label initialization - - - -\n\n\n # - - - - main rule block - - - -\n # at line 321:4: ( '>=' | '<=' | '==' | '!=' | '<' | '>' )\n # at line 321:4: ( '>=' | '<=' | '==' | '!=' | '<' | '>' )\n alt_1 = 6\n case look_1 = @input.peek( 1 )\n when 0x3e then look_1_1 = @input.peek( 2 )\n\n if ( look_1_1 == 0x3d )\n alt_1 = 1\n else\n alt_1 = 6\n\n end\n when 0x3c then look_1_2 = @input.peek( 2 )\n\n if ( look_1_2 == 0x3d )\n alt_1 = 2\n else\n alt_1 = 5\n\n end\n when 0x3d then alt_1 = 3\n when 0x21 then alt_1 = 4\n else\n raise NoViableAlternative( \"\", 1, 0 )\n\n end\n case alt_1\n when 1\n # at line 321:5: '>='\n match( \">=\" )\n\n\n when 2\n # at line 321:10: '<='\n match( \"<=\" )\n\n\n when 3\n # at line 321:15: '=='\n match( \"==\" )\n\n\n when 4\n # at line 321:20: '!='\n match( \"!=\" )\n\n\n when 5\n # at line 321:25: '<'\n match( 0x3c )\n\n when 6\n # at line 321:29: '>'\n match( 0x3e )\n\n end\n\n\n @state.type = type\n @state.channel = channel\n ensure\n # -> uncomment the next line to manually enable rule tracing\n # trace_out( __method__, 17 )\n\n\n end",
"def test_two_proper_args\n assert check_args([1, 1])\n end",
"def test_check_let\n RPN rpn = RPN.new(1)\n let = 'let'\n assert_equal let, 'let'\n assert rpn.check_let(let)\n end",
"def valid_triangle?(a, b, c)\n # Your code goes here!\na+b > c && a+c > b && b+c >a\nend",
"def tuples_for(arg_)\n objdata_ = _get_objdata(arg_)\n objdata_ ? objdata_[2].keys : nil\n end",
"def relation(rel, &relation_definition)\n raise NotImplementedError, 'relation is not supported yet'\n end",
"def tuple_count\n target.count\n end",
"def get_relational(expr)\n\t\t[\"==\", \"!=\", \">\", \"<\", \">=\", \"<=\"].each do |ope|\n\t\t\tif expr.scan(ope).length == 1\n\t\t\t\treturn ope\n\t\t\tend\n\t\tend\n\tend",
"def one_pair?\n n_of_a_kind? 2\n end",
"def test_history_equals_versions\n way = create(:way, :with_history)\n used_way = create(:way, :with_history)\n create(:relation_member, :member => used_way)\n way_with_versions = create(:way, :with_history, :version => 4)\n\n check_history_equals_versions(way.id)\n check_history_equals_versions(used_way.id)\n check_history_equals_versions(way_with_versions.id)\n end",
"def <= other\n call_enum \"relational\", other, :lesseq\n end",
"def to_generate_tuple(&block_)\n if block_\n @gen_tuple = block_\n else\n @gen_tuple\n end\n end",
"def check_rule_ast(pt)\n # :defn format: node tag, block name, args, body_0, ..., body_n\n tag, name, args, *body = pt\n return pt if tag != :defn\n\n body.each_with_index do |n,i|\n next if i == 0 and n == s(:nil) # a block got rewritten to an empty block\n\n # Check for a common case\n if n.sexp_type == :lasgn\n return [n, \"illegal operator: '='\"]\n end\n return pt unless n.sexp_type == :call and n.length == 4\n\n # Rule format: call tag, lhs, op, rhs\n _, lhs, op, rhs = n\n\n # Check that LHS references a named collection\n lhs_name = get_qual_name(lhs)\n return [n, \"unexpected lhs format: #{lhs}\"] if lhs_name.nil?\n unless @bud_instance.tables.has_key? lhs_name.to_sym or\n @bud_instance.lattices.has_key? lhs_name.to_sym\n return [n, \"Collection does not exist: '#{lhs_name}'\"]\n end\n\n return [n, \"illegal operator: '#{op}'\"] unless [:<, :<=].include? op\n\n # Check superator invocation. A superator that begins with \"<\" is parsed\n # as a call to the binary :< operator. The right operand to :< is a :call\n # node; the LHS of the :call is the actual rule body, the :call's oper is\n # the rest of the superator (unary ~, -, +), and the RHS is empty. Note\n # that ParseTree encodes unary \"-\" and \"+\" as :-@ and :-+, respectively.\n # XXX: We don't check for illegal superators (e.g., \"<--\"). That would be\n # tricky, because they are encoded as a nested unary op in the rule body.\n if op == :<\n return n unless rhs.sexp_type == :call\n op_tail = rhs[2]\n return n unless [:~, :-@, :+@].include? op_tail\n rhs_args = rhs[3..-1]\n return n unless rhs_args.empty?\n end\n end\n\n return nil # No errors found\n end",
"def valid_token_tuple?(type, *args)\n args.flatten!\n case type\n when :amazon then valid_amazon_tokens?(*args)\n when :bitlyv1 then valid_bitlyv1_tokens?(*args)\n when :bitlyv2 then valid_bitlyv2_tokens?(*args)\n when :facebook then valid_facebook_tokens?(*args)\n when :flickr then valid_flickr_tokens?(*args)\n when :foursquare then valid_foursquare_tokens?(*args)\n when :google then valid_google_tokens?(*args)\n when :google_oauth then valid_google_oauth_tokens?(*args)\n when :linkedin then valid_linkedin_tokens?(*args)\n when :twitter then valid_twitter_tokens?(*args)\n when :yelpv1 then valid_yelpv1_tokens?(*args)\n when :yelpv2 then valid_yelpv2_tokens?(*args)\n else\n # Rails.logger.error \"Unknown token type: #{type}\"\n end\nend",
"def check_error(val_a, val_b, val_c)\n raise TriangleError if (val_b + val_c <= val_a) || (val_c + val_a <= val_b)\n raise TriangleError if val_a + val_b <= val_c\nend",
"def valid_triangle?(a, b, c)\n ((a+b>c) && (b+c>a) && (a+c>b))? true:false\nend",
"def test_friendp_returns_true_if_users_are_friends\n assert users(:active).friend?(users(:passive))\n assert users(:passive).friend?(users(:active))\n end",
"def to_ruby_literal\n \"Alf::Relation[\" +\n tuples.collect{|t| Tools.to_ruby_literal(t)}.join(', ') + \"]\"\n end",
"def good_tuples(nums)\n count = 0\n (0..nums.size - 3).each do |i|\n first, second, third = nums[i], nums[i + 1], nums[i + 2]\n next if 3 * first == first + second + third\n count += 1 if [first, second, third].include?(first ^ second ^ third)\n\n # next if nums[i] == nums[i + 1] && nums[i + 1] == nums[i + 2]\n # next if nums[i] != nums[i + 1] && nums[i + 1] != nums[i + 2] && nums[i] != nums[i + 2]\n end\n\n count\nend",
"def valid_triangle?(a, b, c)\n a + b > c && a + c > b && b + c > a\nend",
"def valid_triangle?(a, b, c)\n a + b > c && a + c > b && b + c > a\nend",
"def valid_triangle?(a, b, c)\n a + b > c && a + c > b && b + c > a\nend",
"def triplet_check(a,b,c)\n\tif a**2+b**2 == c**2\n\t\treturn true\n\telse\n\t\treturn false\n\tend\nend",
"def visit_axiom_relation_operation_binary(binary)\n generator = self.class.visit(binary)\n @name = generator.name\n @from = aliased_subquery(generator)\n generator\n end",
"def valid_triangle?(a, b, c)\n a + b > c && b + c > a && c + a > b\nend",
"def nested?\n self.resource_tuple.length > 1\n end",
"def right_triangle?(a, b, c)\n a**2 + b**2 == c**2\nend",
"def relationship(*args)\n options = args.extract_options!\n if options[:raise_on_multiple]\n rels = relationships.take(2)\n raise _(\"Multiple relationships found\") if rels.length > 1\n rels.first\n else\n relationships.first\n end\n end",
"def isTriangle(a,b,c)\n (a+b>c && a+c>b && c+b>a)? true : false\nend",
"def >= other\n call_enum \"relational\", other, :moreeq\n end",
"def isTriangle(a, b, c)\n a, b, c = [a, b, c].sort\n a + b > c\nend",
"def test_get_location_bad_input\n assert_raises 'Improper typing of location input argument in method get_location' do\n locations = LocationRelation.new\n locations.get_location 55\n end\n end",
"def insert(tuples)\n pks = tuples.map { |tuple| relation.insert(tuple) }\n relation.where(relation.primary_key => pks).to_a\n end",
"def to_ruby_literal\n \"Alf::Relation([\" + tuples.map{|t| Support.to_ruby_literal(t) }.join(', ') + \"])\"\n end",
"def evaluate\n distinct? ? distinct_relation : relation\n end",
"def check_permission_with_http_info(opts = {})\n if @api_client.config.debugging\n @api_client.config.logger.debug 'Calling API: PermissionApi.check_permission ...'\n end\n # resource path\n local_var_path = '/relation-tuples/check/openapi'\n\n # query parameters\n query_params = opts[:query_params] || {}\n query_params[:'namespace'] = opts[:'namespace'] if !opts[:'namespace'].nil?\n query_params[:'object'] = opts[:'object'] if !opts[:'object'].nil?\n query_params[:'relation'] = opts[:'relation'] if !opts[:'relation'].nil?\n query_params[:'subject_id'] = opts[:'subject_id'] if !opts[:'subject_id'].nil?\n query_params[:'subject_set.namespace'] = opts[:'subject_set_namespace'] if !opts[:'subject_set_namespace'].nil?\n query_params[:'subject_set.object'] = opts[:'subject_set_object'] if !opts[:'subject_set_object'].nil?\n query_params[:'subject_set.relation'] = opts[:'subject_set_relation'] if !opts[:'subject_set_relation'].nil?\n query_params[:'max-depth'] = opts[:'max_depth'] if !opts[:'max_depth'].nil?\n\n # header parameters\n header_params = opts[:header_params] || {}\n # HTTP header 'Accept' (if needed)\n header_params['Accept'] = @api_client.select_header_accept(['application/json'])\n\n # form parameters\n form_params = opts[:form_params] || {}\n\n # http body (model)\n post_body = opts[:debug_body]\n\n # return_type\n return_type = opts[:debug_return_type] || 'CheckPermissionResult'\n\n # auth_names\n auth_names = opts[:debug_auth_names] || ['oryAccessToken']\n\n new_options = opts.merge(\n :operation => :\"PermissionApi.check_permission\",\n :header_params => header_params,\n :query_params => query_params,\n :form_params => form_params,\n :body => post_body,\n :auth_names => auth_names,\n :return_type => return_type\n )\n\n data, status_code, headers = @api_client.call_api(:GET, local_var_path, new_options)\n if @api_client.config.debugging\n @api_client.config.logger.debug \"API called: PermissionApi#check_permission\\nData: #{data.inspect}\\nStatus code: #{status_code}\\nHeaders: #{headers}\"\n end\n return data, status_code, headers\n end",
"def with_input_tuples(tuples, &block)\n input_tuples = Array([tuples]).flatten(1).map\n return input_tuples unless block_given?\n\n input_tuples.each(&block)\n end",
"def test_two_valid_arguments\n\t\tc = Check.new\n\t\trefute c.check_arguments([1,1])\n\tend",
"def test_triangle\n if @side1.eql?(@side2) && @side2.eql?(@side3)\n \"It's an equilateral\"\n elsif @side1.eql?(@side2) || @side2.eql?(@side3)\n \"It's an isosceles\"\n elsif @side1.eql?(@side2) == false && @side2.eql?(@side3) == false\n \"It's a scalene triangle\"\n else\n \"Not a triangle\"\n end\n end",
"def parse_relational\n add = parse_add\n if consume(:lt)\n Node.new(:binary, op: :lt, lhs: add, rhs: parse_relational)\n elsif consume(:lteq)\n Node.new(:binary, op: :lteq, lhs: add, rhs: parse_relational)\n elsif consume(:gt)\n Node.new(:binary, op: :gt, lhs: add, rhs: parse_relational)\n elsif consume(:gteq)\n Node.new(:binary, op: :gteq, lhs: add, rhs: parse_relational)\n else\n add\n end\n end",
"def pitagoric_triplet?(a,b,c)\n (a**2 + b**2) == c**2\nend",
"def selecionErr(method)\n\n tnTableCreation('tuple_node_test_result') if @is_new\n\n if @unwanted_tuple_count + @missing_tuple_count == 0\n p 'no failed rows found. There is no selection error'\n return\n end\n\n if @unwanted_tuple_count > 0\n # p \"Unwanted Pk count #{unWantedPK.count()}\"\n # create unwanted_tuple_branch table\n # binding.pry\n\n # whereErrList = whereCondTest(@unWantedPK, 'U')\n where_cond_test('U')\n # joinErrList = jointypeErr(query,'U')\n end\n\n if @missing_tuple_count > 0\n # p \"Missing PK count #{missinPK.count()}\"\n # binding.pry\n # whereErrList = whereCondTest(@missingPK, 'M')\n where_cond_test('M')\n # joinErrList = jointypeErr(query,'M')\n end\n # create aggregated tuple_suspicious_nodes\n pk = @pkFullList.map { |pk| pk['alias'] }.join(',')\n query = 'create materialized view tuple_node_test_result_aggr as '\\\n \"select #{pk}, string_agg(branch_name||'-'||node_name, ',' order by node_name,branch_name) as suspicious_nodes from tuple_node_test_result group by #{pk}\"\n pp query\n DBConn.exec(query)\n\n unless @predicateTree.nil?\n suspicious_score_upd(@predicateTree.branches)\n end\n # exnorate algorithm\n # binding.pry\n @column_combinations = method.start_with?('o') ? Columns_Combination.new(@all_columns) : @all_columns\n case method\n # when 'o'\n # puts 'old exonerate algorithm'\n\n # true_query_PT_construct\n # constraint_query = constraint_predicate_construct\n # # column_combinations_construct\n # tuple_mutation_test(@missingPK, 'M', constraint_query, false)\n # tuple_mutation_test(@unWantedPK, 'U', constraint_query, false)\n when 'or'\n puts 'old exonerate algorithm with duplicate removal'\n # reset suspicious score\n query = \"update node_query_mapping set suspicious_score = 0 where type = 'f'\"\n res = DBConn.exec(query)\n\n true_query_PT_construct\n constraint_query = constraint_predicate_construct\n # allcolumns_construct()\n puts 'Missing starts'\n puts Time.now()\n tuple_mutation_test_with_dup_removal('M', constraint_query)\n puts 'Unwanted starts'\n puts Time.now()\n\n tuple_mutation_test_with_dup_removal('U', constraint_query)\n # when 'n'\n # puts 'new exonerate algorithm'\n # true_query_PT_construct\n # constraint_query = constraint_predicate_construct\n # tuple_mutation_test_reverse(@missingPK, 'M', constraint_query, false)\n # tuple_mutation_test_reverse(@unWantedPK, 'U', constraint_query, false)\n\n when 'b'\n puts 'baseline'\n else\n puts 'Unknown method'\n end\n # remove constraint_nodes in node_query_mapping\n query = \"delete from node_query_mapping where test_id = #{@test_id} and type = 't'\"\n DBConn.exec(query)\n\n # j = Hash.new\n # # j['JoinErr'] = joinErrList if joinErrList.count >0\n # j['WhereErr'] = whereErrList\n # j\n # whereErrList\n end",
"def isTriangle(a,b,c)\n a+b>c && a + c > b && b + c > a ? true : false\nend",
"def arel_bt_intersect(*bt_tuple)\n arel_vt_intersect(bt_tuple.at(0), bt_tuple.at(1)).and(arel_tt_intersect(bt_tuple.at(2), bt_tuple.at(3)))\n end"
] |
[
"0.7015953",
"0.6605728",
"0.6361857",
"0.6198903",
"0.60191625",
"0.579029",
"0.5698508",
"0.5596643",
"0.5500909",
"0.53615284",
"0.5320704",
"0.52337974",
"0.51784104",
"0.51652044",
"0.5137385",
"0.5097203",
"0.5097203",
"0.5095431",
"0.50804377",
"0.5077863",
"0.5056793",
"0.5047188",
"0.50432706",
"0.50223356",
"0.49960107",
"0.49366918",
"0.49293387",
"0.4916955",
"0.4907835",
"0.48614576",
"0.48245752",
"0.4788206",
"0.4757173",
"0.47517347",
"0.4732002",
"0.46899974",
"0.46806657",
"0.46671054",
"0.46462044",
"0.46259937",
"0.46217015",
"0.46073496",
"0.4602068",
"0.45971423",
"0.4588765",
"0.45760465",
"0.4574195",
"0.45620072",
"0.45608",
"0.45552087",
"0.4553977",
"0.45391572",
"0.45349598",
"0.45325094",
"0.45259938",
"0.45203832",
"0.4519309",
"0.45055634",
"0.4499442",
"0.44971564",
"0.44606513",
"0.44566202",
"0.44505966",
"0.4447705",
"0.4445908",
"0.4441292",
"0.4428633",
"0.44250175",
"0.44135484",
"0.44133824",
"0.4411091",
"0.44052505",
"0.44050676",
"0.44036365",
"0.4403217",
"0.44016722",
"0.44010103",
"0.44010103",
"0.44010103",
"0.43980426",
"0.43945882",
"0.4391476",
"0.43906316",
"0.43885761",
"0.43848813",
"0.43774366",
"0.4370648",
"0.43687898",
"0.43655345",
"0.4363082",
"0.43569046",
"0.43532142",
"0.43471527",
"0.43461266",
"0.4344165",
"0.43420166",
"0.4339705",
"0.43395022",
"0.43379533",
"0.4334619",
"0.43341088"
] |
0.0
|
-1
|
Check a relation tuple To learn how relation tuples and the check works, head over to [the documentation](../concepts/relationtuples.mdx).
|
def post_check_with_http_info(opts = {})
if @api_client.config.debugging
@api_client.config.logger.debug 'Calling API: ReadApi.post_check ...'
end
# resource path
local_var_path = '/check'
# query parameters
query_params = opts[:query_params] || {}
# header parameters
header_params = opts[:header_params] || {}
# HTTP header 'Accept' (if needed)
header_params['Accept'] = @api_client.select_header_accept(['application/json'])
# HTTP header 'Content-Type'
header_params['Content-Type'] = @api_client.select_header_content_type(['application/json'])
# form parameters
form_params = opts[:form_params] || {}
# http body (model)
post_body = opts[:debug_body] || @api_client.object_to_http_body(opts[:'payload'])
# return_type
return_type = opts[:debug_return_type] || 'GetCheckResponse'
# auth_names
auth_names = opts[:debug_auth_names] || []
new_options = opts.merge(
:operation => :"ReadApi.post_check",
:header_params => header_params,
:query_params => query_params,
:form_params => form_params,
:body => post_body,
:auth_names => auth_names,
:return_type => return_type
)
data, status_code, headers = @api_client.call_api(:POST, local_var_path, new_options)
if @api_client.config.debugging
@api_client.config.logger.debug "API called: ReadApi#post_check\nData: #{data.inspect}\nStatus code: #{status_code}\nHeaders: #{headers}"
end
return data, status_code, headers
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def valid_relation_literal!(heading, tuples, raise_on_error = true)\n if Array === tuples\n invalid = tuples.find{|t| !valid_tuple_literal?(heading, t)}\n if invalid\n raise_on_error ? raise(::Veritas::RelationMismatchError, \"Invalid tuple #{invalid.inspect} for heading #{heading}\", caller) : false\n else\n tuples\n end\n else\n raise_on_error ? raise(::Veritas::RelationMismatchError, \"Invalid relation literal #{tuples.inspect}\", caller) : false \n end\n end",
"def check_for_tuple(tuple)\n ltuple = @left_pattern.map {|i| tuple[i]}\n if @left.check_for_tuple(ltuple)\n rtuple = @right_pattern.map {|i| tuple[i]}\n if @right.check_for_tuple(rtuple)\n return true\n end\n end\n return false\n end",
"def contains?(tuple)\n case tuple\n when Hash\n eql_hash = {}\n relvar_def.primary_key.attributes.each{|a| eql_hash[a.name] = tuple[a.name]}\n inside_tuple = underlying_table.filter(eql_hash).first\n return inside_tuple === tuple\n else\n raise ArgumentError, \"Unable to check inclusion of #{tuple} inside a relation variable\"\n end\n end",
"def valid_tuple_literal!(heading, tuple, raise_on_error = true)\n if Hash===tuple and tuple.size == heading.to_ary.size\n ok = heading.all?{|a| tuple.key?(a.name) and a.valid_value?(tuple[a.name])}\n return tuple if ok\n end\n raise_on_error ? raise(ArgumentError, \"Invalid tuple #{tuple} for #{heading}\", caller) : false\n end",
"def _tuple2tuple(tuple)\n end",
"def tuple_plan()\n raise NotTupleValuedException.new(self) unless tuple_valued?\n @definition.type.tuple.plan\n end",
"def create_relation_tuple(opts = {})\n data, _status_code, _headers = create_relation_tuple_with_http_info(opts)\n data\n end",
"def check_relation(relation)\n (relation - reflection.map{|r| r[:rel] }).empty?\n end",
"def relation\n # -> uncomment the next line to manually enable rule tracing\n # trace_in( __method__, 28 )\n\n\n value = nil\n\n\n type = nil\n a = nil\n b = nil\n\n\n begin\n # at line 178:5: a= add (type= ( '==' | '!=' | '<' | '<=' | '>=' | '>' ) b= relation )?\n @state.following.push( TOKENS_FOLLOWING_add_IN_relation_1322 )\n a = add\n @state.following.pop\n\n # syntactic predicate action gate test\n if @state.backtracking == 0\n # --> action\n value = a \n # <-- action\n end\n\n # at line 179:3: (type= ( '==' | '!=' | '<' | '<=' | '>=' | '>' ) b= relation )?\n alt_44 = 2\n look_44_0 = @input.peek( 1 )\n\n if ( look_44_0 == T__31 || look_44_0.between?( T__47, T__48 ) || look_44_0.between?( T__50, T__52 ) )\n alt_44 = 1\n end\n case alt_44\n when 1\n # at line 179:6: type= ( '==' | '!=' | '<' | '<=' | '>=' | '>' ) b= relation\n type = @input.look\n\n if @input.peek(1) == T__31 || @input.peek( 1 ).between?( T__47, T__48 ) || @input.peek( 1 ).between?( T__50, T__52 )\n @input.consume\n @state.error_recovery = false\n\n else\n @state.backtracking > 0 and raise( ANTLR3::Error::BacktrackingFailed )\n\n\n mse = MismatchedSet( nil )\n raise mse\n\n end\n\n\n @state.following.push( TOKENS_FOLLOWING_relation_IN_relation_1359 )\n b = relation\n @state.following.pop\n\n # syntactic predicate action gate test\n if @state.backtracking == 0\n # --> action\n value = DoubleOperandExpressionEval.new(type.text, a, b) \n # <-- action\n end\n\n\n end\n\n rescue ANTLR3::Error::RecognitionError => re\n report_error(re)\n recover(re)\n\n ensure\n # -> uncomment the next line to manually enable rule tracing\n # trace_out( __method__, 28 )\n\n\n end\n\n return value\n end",
"def eval_tuple(node, env, stack_trace)\n\n\tevaluated_nodes = []\n\tnode.values.each{ |sub_node|\n\t\tsub_node_value, _ = eval_node_under(sub_node, env, stack_trace) \n\t\tevaluated_nodes.append(sub_node_value)\n\t}\n\tif evaluated_nodes.length == 1\n\t\treturn evaluated_nodes[0], env\n\telse\n\t\treturn Tuple.new(node.line, node.col, evaluated_nodes)\n\tend\nend",
"def triangle?((x1, y1), (x2, y2), (x3, y3))\n return false if (x1, y1) == (x2, y2) || (x1, y1) == (x3, y3) || (x2, y2) == (x3, y3)\n return false \n true\nend",
"def patch_relation_tuples(opts = {})\n patch_relation_tuples_with_http_info(opts)\n nil\n end",
"def is_meta_tuple? tuple\n if meta_subspace\n meta_subspace === tuple\n else\n # meta_subspace hasn't arrived yet, so use approximation\n tuple.kind_of? Hash and tuple.key? client.tupelo_meta_key and\n tuple[client.tupelo_meta_key] == \"subspace\"\n end\n end",
"def t\n BERT::Tuple\nend",
"def activity_relations?(*args)\n activity_relations(*args).any?\n end",
"def tuples; end",
"def tuples; end",
"def assert_working_relations( model, relations )\n failed = []\n relations.each do | name |\n begin\n model.send( name )\n rescue Exception => e\n failed << \"#{name} - #{e}\"\n end\n end\n if failed.empty?\n assert true\n else\n assert false, failed.join('; ')\n end\n end",
"def check_relation(x)\n h, qx = compute_q(x)\n return if qx == 0\n es, l = exponent_bitvector(qx)\n\n # discard this x if the residue L is too big\n return if l > @pmax ** @t\n\n if l == 1\n # complete relation found:\n # Q(x) = p0^e0 * p1^e1 * ... * pk^ek (pi in the factor base)\n qx_vec = Hash.new(0)\n PrimeFactorization.prime_factorization(qx) {|p, e| qx_vec[p] += e }\n collect_relation(es, h, qx_vec)\n\n elsif @incomplete_relations[l]\n # large prime procedure:\n # make a complete relation by multiplying two incomplete relations\n es2, h2, qx2 = @incomplete_relations[l]\n\n # XXX: use FactoredInteger\n qx_vec = Hash.new(0)\n PrimeFactorization.prime_factorization(qx) {|p, e| qx_vec[p] += e }\n PrimeFactorization.prime_factorization(qx2) {|p, e| qx_vec[p] += e }\n\n collect_relation(es ^ es2, h * h2 % @kn, qx_vec)\n\n else\n @incomplete_relations[l] = [es, h, qx]\n end\n end",
"def verify_uniquness!(tuple)\n stream = tuple[:stream]\n attrs = %i[position event_id]\n attrs.delete(:position) if Stream.new(stream).global?\n\n attrs.each do |key|\n next if key == :position && tuple[key].nil?\n next if restrict(:stream => stream, key => tuple.fetch(key)).none?\n\n raise TupleUniquenessError.public_send(:\"for_stream_and_#{key}\", stream, tuple.fetch(key))\n end\n end",
"def create_relation_tuple_with_http_info(opts = {})\n if @api_client.config.debugging\n @api_client.config.logger.debug 'Calling API: WriteApi.create_relation_tuple ...'\n end\n # resource path\n local_var_path = '/relation-tuples'\n\n # query parameters\n query_params = opts[:query_params] || {}\n\n # header parameters\n header_params = opts[:header_params] || {}\n # HTTP header 'Accept' (if needed)\n header_params['Accept'] = @api_client.select_header_accept(['application/json'])\n # HTTP header 'Content-Type'\n header_params['Content-Type'] = @api_client.select_header_content_type(['application/json'])\n\n # form parameters\n form_params = opts[:form_params] || {}\n\n # http body (model)\n post_body = opts[:debug_body] || @api_client.object_to_http_body(opts[:'payload'])\n\n # return_type\n return_type = opts[:debug_return_type] || 'RelationQuery'\n\n # auth_names\n auth_names = opts[:debug_auth_names] || []\n\n new_options = opts.merge(\n :operation => :\"WriteApi.create_relation_tuple\",\n :header_params => header_params,\n :query_params => query_params,\n :form_params => form_params,\n :body => post_body,\n :auth_names => auth_names,\n :return_type => return_type\n )\n\n data, status_code, headers = @api_client.call_api(:PUT, local_var_path, new_options)\n if @api_client.config.debugging\n @api_client.config.logger.debug \"API called: WriteApi#create_relation_tuple\\nData: #{data.inspect}\\nStatus code: #{status_code}\\nHeaders: #{headers}\"\n end\n return data, status_code, headers\n end",
"def delete_relation_tuple_with_http_info(namespace, object, relation, opts = {})\n if @api_client.config.debugging\n @api_client.config.logger.debug 'Calling API: WriteApi.delete_relation_tuple ...'\n end\n # verify the required parameter 'namespace' is set\n if @api_client.config.client_side_validation && namespace.nil?\n fail ArgumentError, \"Missing the required parameter 'namespace' when calling WriteApi.delete_relation_tuple\"\n end\n # verify the required parameter 'object' is set\n if @api_client.config.client_side_validation && object.nil?\n fail ArgumentError, \"Missing the required parameter 'object' when calling WriteApi.delete_relation_tuple\"\n end\n # verify the required parameter 'relation' is set\n if @api_client.config.client_side_validation && relation.nil?\n fail ArgumentError, \"Missing the required parameter 'relation' when calling WriteApi.delete_relation_tuple\"\n end\n # resource path\n local_var_path = '/relation-tuples'\n\n # query parameters\n query_params = opts[:query_params] || {}\n query_params[:'namespace'] = namespace\n query_params[:'object'] = object\n query_params[:'relation'] = relation\n query_params[:'subject_id'] = opts[:'subject_id'] if !opts[:'subject_id'].nil?\n query_params[:'subject_set.namespace'] = opts[:'subject_set_namespace'] if !opts[:'subject_set_namespace'].nil?\n query_params[:'subject_set.object'] = opts[:'subject_set_object'] if !opts[:'subject_set_object'].nil?\n query_params[:'subject_set.relation'] = opts[:'subject_set_relation'] if !opts[:'subject_set_relation'].nil?\n\n # header parameters\n header_params = opts[:header_params] || {}\n # HTTP header 'Accept' (if needed)\n header_params['Accept'] = @api_client.select_header_accept(['application/json'])\n\n # form parameters\n form_params = opts[:form_params] || {}\n\n # http body (model)\n post_body = opts[:debug_body]\n\n # return_type\n return_type = opts[:debug_return_type]\n\n # auth_names\n auth_names = opts[:debug_auth_names] || []\n\n new_options = opts.merge(\n :operation => :\"WriteApi.delete_relation_tuple\",\n :header_params => header_params,\n :query_params => query_params,\n :form_params => form_params,\n :body => post_body,\n :auth_names => auth_names,\n :return_type => return_type\n )\n\n data, status_code, headers = @api_client.call_api(:DELETE, local_var_path, new_options)\n if @api_client.config.debugging\n @api_client.config.logger.debug \"API called: WriteApi#delete_relation_tuple\\nData: #{data.inspect}\\nStatus code: #{status_code}\\nHeaders: #{headers}\"\n end\n return data, status_code, headers\n end",
"def tuple_extract(*args, &bl)\n alf_connection.tuple_extract(*args, &bl)\n end",
"def relation(relation)\n true\n end",
"def tuple_valued?()\n @definition.type.named_type?\n end",
"def delete_relation_tuple(namespace, object, relation, opts = {})\n delete_relation_tuple_with_http_info(namespace, object, relation, opts)\n nil\n end",
"def check_relation_with friend_id\n response = {\n message: \"Successfully fecth user relation\",\n following: self.active_relationships.where(follower_id: friend_id).present?,\n follower: self.passive_relationships.where(follower_id: friend_id).present?\n }\n end",
"def is_valid_relation?(parent, child, require_both = self.require_both)\n # depending upon the\n require_both ? is_valid_down_relation?(parent, child) && is_valid_up_relation?(parent, child)\n : is_valid_down_relation?(parent, child) || is_valid_up_relation?(parent, child)\n end",
"def parse_tuple cur_tok\n\t\tcur_ast = Tuple.new(cur_tok.line, cur_tok.col, splitter( nil, \")\", \",\")) # start=nil because \"(\" token has already been read.\n\n\t\t# Use look ahead to see if the tuple needs to be a child node of an operation.\n\t\tif is_operation \n\t\t\treturn parse_operation(cur_ast, true)\n\t\telse\n\t\t\treturn cur_ast \n\t\tend\n\tend",
"def deduce\n each_vertex do |v1|\n each_vertex do |v2|\n each_vertex do |v3|\n \n if self[v1,v2] == :provable_true and self[v2,v3] == :provable_true\n add_assertion(v1, v3, :provable_true)\n end\n \n if self[v2,v1] == :provable_false and self[v2,v3] == :provable_true\n add_assertion(v3, v1, :provable_false)\n end\n \n if self[v1,v2] == :provable_true and self[v3,v2] == :provable_false\n add_assertion(v3, v1, :provable_false)\n end\n \n break if @contradiction\n end\n end\n end\n end",
"def cast_tuple\n <<-CODE\n t1 = stack_pop();\n if(REFERENCE_P(t1) && object_kind_of_p(state, t1, global->array)) {\n j = N2I(array_get_total(t1));\n t2 = tuple_new(state, j);\n\n for(k = 0; k < j; k++) {\n tuple_put(state, t2, k, array_get(state, t1, k));\n }\n t1 = t2;\n } else if(!REFERENCE_P(t1) || !object_kind_of_p(state, t1, global->tuple)) {\n t2 = tuple_new(state, 1);\n tuple_put(state, t2, 0, t1);\n t1 = t2;\n }\n stack_push(t1);\n CODE\n end",
"def assert_relations_are_equal(a, b)\n assert_not_nil a, \"first relation is not allowed to be nil\"\n assert_not_nil b, \"second relation #{a.id} is not allowed to be nil\"\n assert_equal a.id, b.id, \"relation IDs\"\n assert_equal a.changeset_id, b.changeset_id, \"changeset ID on relation #{a.id}\"\n assert_equal a.visible, b.visible, \"visible on relation #{a.id}, #{a.visible.inspect} != #{b.visible.inspect}\"\n assert_equal a.version, b.version, \"version on relation #{a.id}\"\n assert_equal a.tags, b.tags, \"tags on relation #{a.id}\"\n assert_equal a.members, b.members, \"member references on relation #{a.id}\"\n end",
"def tuple(*elements)\n Tuple.new(*elements)\n end",
"def cmp_tuples(left, right)\n reduce(0) do |cmp, direction|\n break cmp if cmp.nonzero?\n direction.call(left, right)\n end\n end",
"def get_relation_tuples(namespace, opts = {})\n data, _status_code, _headers = get_relation_tuples_with_http_info(namespace, opts)\n data\n end",
"def func_le(args)\n p1 = _eval(car(args))\n p2 = _eval(car(cdr(args)))\n\n if p1.type != LObject::OBJ_INTEGER and p2.type != LObject::OBJ_INTEGER\n if @lint\n Error.warn(\"warning: comparison with a non integer operand\")\n end\n return @o_man.nil\n end\n\n if p1.value.i <= p2.value.i\n return @o_man.t\n end\n return @o_man.nil\n end",
"def tuple(*args)\n \"( #{args.map { |b| if b.nil? then 'NULL' else b end }.join(', ')} )\"\nend",
"def relation\n # -> uncomment the next line to manually enable rule tracing\n # trace_in( __method__, 55 )\n return_value = RelationReturnValue.new\n\n # $rule.start = the first token seen before matching\n return_value.start = @input.look\n\n root_0 = nil\n shift249 = nil\n relation_op250 = nil\n shift251 = nil\n\n\n begin\n root_0 = @adaptor.create_flat_list\n\n\n # at line 609:5: shift ( relation_op shift )*\n @state.following.push( TOKENS_FOLLOWING_shift_IN_relation_4033 )\n shift249 = shift\n @state.following.pop\n if @state.backtracking == 0\n @adaptor.add_child( root_0, shift249.tree )\n end\n # at line 609:11: ( relation_op shift )*\n while true # decision 59\n alt_59 = 2\n look_59_0 = @input.peek( 1 )\n\n if ( look_59_0 == GEQ || look_59_0 == GREATER || look_59_0 == INSTANCEOF || look_59_0 == LEQ || look_59_0 == LESS )\n alt_59 = 1\n elsif ( look_59_0 == IN ) and ( ( !( @InFor_stack.last && @InFor_stack.last.active ) ) )\n alt_59 = 1\n\n end\n case alt_59\n when 1\n # at line 609:14: relation_op shift\n @state.following.push( TOKENS_FOLLOWING_relation_op_IN_relation_4038 )\n relation_op250 = relation_op\n @state.following.pop\n if @state.backtracking == 0\n root_0 = @adaptor.become_root( relation_op250.tree, root_0 )\n end\n @state.following.push( TOKENS_FOLLOWING_shift_IN_relation_4042 )\n shift251 = shift\n @state.following.pop\n if @state.backtracking == 0\n @adaptor.add_child( root_0, shift251.tree )\n end\n\n else\n break # out of loop for decision 59\n end\n end # loop for decision 59\n # - - - - - - - rule clean up - - - - - - - -\n return_value.stop = @input.look( -1 )\n\n if @state.backtracking == 0\n\n return_value.tree = @adaptor.rule_post_processing( root_0 )\n @adaptor.set_token_boundaries( return_value.tree, return_value.start, return_value.stop )\n\n end\n rescue ANTLR3::Error::RecognitionError => re\n report_error(re)\n recover(re)\n return_value.tree = @adaptor.create_error_node( @input, return_value.start, @input.look(-1), re )\n\n ensure\n # -> uncomment the next line to manually enable rule tracing\n # trace_out( __method__, 55 )\n\n end\n \n return return_value\n end",
"def get_relation_tuples_with_http_info(namespace, opts = {})\n if @api_client.config.debugging\n @api_client.config.logger.debug 'Calling API: ReadApi.get_relation_tuples ...'\n end\n # verify the required parameter 'namespace' is set\n if @api_client.config.client_side_validation && namespace.nil?\n fail ArgumentError, \"Missing the required parameter 'namespace' when calling ReadApi.get_relation_tuples\"\n end\n # resource path\n local_var_path = '/relation-tuples'\n\n # query parameters\n query_params = opts[:query_params] || {}\n query_params[:'namespace'] = namespace\n query_params[:'page_token'] = opts[:'page_token'] if !opts[:'page_token'].nil?\n query_params[:'page_size'] = opts[:'page_size'] if !opts[:'page_size'].nil?\n query_params[:'object'] = opts[:'object'] if !opts[:'object'].nil?\n query_params[:'relation'] = opts[:'relation'] if !opts[:'relation'].nil?\n query_params[:'subject_id'] = opts[:'subject_id'] if !opts[:'subject_id'].nil?\n query_params[:'subject_set.namespace'] = opts[:'subject_set_namespace'] if !opts[:'subject_set_namespace'].nil?\n query_params[:'subject_set.object'] = opts[:'subject_set_object'] if !opts[:'subject_set_object'].nil?\n query_params[:'subject_set.relation'] = opts[:'subject_set_relation'] if !opts[:'subject_set_relation'].nil?\n\n # header parameters\n header_params = opts[:header_params] || {}\n # HTTP header 'Accept' (if needed)\n header_params['Accept'] = @api_client.select_header_accept(['application/json'])\n\n # form parameters\n form_params = opts[:form_params] || {}\n\n # http body (model)\n post_body = opts[:debug_body]\n\n # return_type\n return_type = opts[:debug_return_type] || 'GetRelationTuplesResponse'\n\n # auth_names\n auth_names = opts[:debug_auth_names] || []\n\n new_options = opts.merge(\n :operation => :\"ReadApi.get_relation_tuples\",\n :header_params => header_params,\n :query_params => query_params,\n :form_params => form_params,\n :body => post_body,\n :auth_names => auth_names,\n :return_type => return_type\n )\n\n data, status_code, headers = @api_client.call_api(:GET, local_var_path, new_options)\n if @api_client.config.debugging\n @api_client.config.logger.debug \"API called: ReadApi#get_relation_tuples\\nData: #{data.inspect}\\nStatus code: #{status_code}\\nHeaders: #{headers}\"\n end\n return data, status_code, headers\n end",
"def valid_triangle? triple\n a, b, c = triple\n return false if a + b <= c\n return false if a + c <= b\n return false if b + c <= a\n\n true\nend",
"def tuples\n relation = self.relation\n return relation if materialized?\n\n adapter.read(relation)\n end",
"def create_tuple!(attributes)\n # split the attrs (needs DRYing)\n rel_attrs = {}\n attributes.each do |key, value|\n if key.to_s[0..3] == 'rel_'\n rel_attrs[key] = attributes.delete(key)\n end\n end\n\n # find or create the one node\n if target_node = @target_class.find(attributes[:id])\n target_node.update_attributes!(attributes)\n else\n target_node = @target_class.create!(attributes)\n end\n\n self.relate(target_node, rel_attrs)\n\n # finally, make and return the tuple\n #Tuple.new(@node, target_node)\n #Tuple.new(@node, target_node)\n\n end",
"def grid_operation_checker(candidate)\n candidate.node_type == :call and\n candidate.sons.first.node_type == :lval and\n (candidate.grid_objects.map {|g_obj| g_obj.name}).include? candidate.sons.first.name\n end",
"def tuple_print(results)\n results.each do |tuple|\n puts tuple\n end\nend",
"def check_permission_or_error_with_http_info(opts = {})\n if @api_client.config.debugging\n @api_client.config.logger.debug 'Calling API: PermissionApi.check_permission_or_error ...'\n end\n # resource path\n local_var_path = '/relation-tuples/check'\n\n # query parameters\n query_params = opts[:query_params] || {}\n query_params[:'namespace'] = opts[:'namespace'] if !opts[:'namespace'].nil?\n query_params[:'object'] = opts[:'object'] if !opts[:'object'].nil?\n query_params[:'relation'] = opts[:'relation'] if !opts[:'relation'].nil?\n query_params[:'subject_id'] = opts[:'subject_id'] if !opts[:'subject_id'].nil?\n query_params[:'subject_set.namespace'] = opts[:'subject_set_namespace'] if !opts[:'subject_set_namespace'].nil?\n query_params[:'subject_set.object'] = opts[:'subject_set_object'] if !opts[:'subject_set_object'].nil?\n query_params[:'subject_set.relation'] = opts[:'subject_set_relation'] if !opts[:'subject_set_relation'].nil?\n query_params[:'max-depth'] = opts[:'max_depth'] if !opts[:'max_depth'].nil?\n\n # header parameters\n header_params = opts[:header_params] || {}\n # HTTP header 'Accept' (if needed)\n header_params['Accept'] = @api_client.select_header_accept(['application/json'])\n\n # form parameters\n form_params = opts[:form_params] || {}\n\n # http body (model)\n post_body = opts[:debug_body]\n\n # return_type\n return_type = opts[:debug_return_type] || 'CheckPermissionResult'\n\n # auth_names\n auth_names = opts[:debug_auth_names] || ['oryAccessToken']\n\n new_options = opts.merge(\n :operation => :\"PermissionApi.check_permission_or_error\",\n :header_params => header_params,\n :query_params => query_params,\n :form_params => form_params,\n :body => post_body,\n :auth_names => auth_names,\n :return_type => return_type\n )\n\n data, status_code, headers = @api_client.call_api(:GET, local_var_path, new_options)\n if @api_client.config.debugging\n @api_client.config.logger.debug \"API called: PermissionApi#check_permission_or_error\\nData: #{data.inspect}\\nStatus code: #{status_code}\\nHeaders: #{headers}\"\n end\n return data, status_code, headers\n end",
"def is_triplet(n1,n2,n3)\n n1**2 + n2**2 == n3**2 or n1**2 + n3**2 == n2**2 or n2**2 + n3**2 == n1**2\nend",
"def is_pythag_triplet?(a,b,c)\n a < b && b < c\nend",
"def create\n @tuple = Tuple.new(params[:tuple])\n\n respond_to do |format|\n if @tuple.save\n format.html { redirect_to(@tuple, :notice => 'Tuple was successfully created.') }\n format.xml { render :xml => @tuple, :status => :created, :location => @tuple }\n else\n format.html { render :action => \"new\" }\n format.xml { render :xml => @tuple.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def _add_tuple(objdata_, tuple_) # :nodoc:\n return false if @tuples.has_key?(tuple_)\n @tuples[tuple_] = objdata_\n tupcats_ = []\n @categories.each do |category_, catdata_|\n if Utils.matches?(catdata_[0], tuple_)\n index_ = catdata_[1].map{ |i_| tuple_[i_] }\n ((@catdata[category_] ||= {})[index_] ||= {})[tuple_] = objdata_\n tupcats_ << category_\n end\n end\n objdata_[2][tuple_] = tupcats_\n true\n end",
"def has_edge?(vertex1, vertex2)\n\tend",
"def isTriangle(a,b,c)\n a + b > c && b + c > a && a + c > b\nend",
"def isTriangle(a,b,c)\n a + b > c && b + c > a && a + c > b\nend",
"def lift_tuple(tuple)\n if Pione::TupleSpace[tuple.first]\n if pos = Pione::TupleSpace[tuple.first].location_position\n if new_location = lift_location(tuple[pos])\n tuple = tuple.clone\n tuple[pos] = new_location\n end\n end\n end\n return tuple\n end",
"def _remove_tuple(objdata_, tuple_) # :nodoc:\n tupcats_ = objdata_[2][tuple_]\n return false unless tupcats_\n @tuples.delete(tuple_)\n tupcats_.each do |cat_|\n index_ = @categories[cat_][1].map{ |i_| tuple_[i_] }\n @catdata[cat_][index_].delete(tuple_)\n end\n objdata_[2].delete(tuple_)\n true\n end",
"def validate_relation(klass, name, options = {})\n [name, \"#{name}?\".to_sym, \"#{name}=\".to_sym].each do |n|\n if Mongoid.destructive_fields.include?(n)\n raise Errors::InvalidRelation.new(klass, n)\n end\n end\n end",
"def call(relation)\n relation\n end",
"def valid?\n # Check tx = <lx, rx>\n unless FIELD.mod(lx.zip(rx).map { |a, b| FIELD.mod(a * b) }.sum) == tx\n return false\n end\n\n return false unless valid_poly_t?\n\n p =\n vec_g\n .zip(lx)\n .map { |a, b| a * b }\n .zip(vec_h2.zip(rx).map { |a, b| a * b })\n .map { |a, b| a + b }\n .sum(GROUP.infinity)\n\n p1 == p2 && p2 == p\n end",
"def op_rel!\n # -> uncomment the next line to manually enable rule tracing\n # trace_in( __method__, 17 )\n\n\n\n type = OP_REL\n channel = ANTLR3::DEFAULT_CHANNEL\n # - - - - label initialization - - - -\n\n\n # - - - - main rule block - - - -\n # at line 321:4: ( '>=' | '<=' | '==' | '!=' | '<' | '>' )\n # at line 321:4: ( '>=' | '<=' | '==' | '!=' | '<' | '>' )\n alt_1 = 6\n case look_1 = @input.peek( 1 )\n when 0x3e then look_1_1 = @input.peek( 2 )\n\n if ( look_1_1 == 0x3d )\n alt_1 = 1\n else\n alt_1 = 6\n\n end\n when 0x3c then look_1_2 = @input.peek( 2 )\n\n if ( look_1_2 == 0x3d )\n alt_1 = 2\n else\n alt_1 = 5\n\n end\n when 0x3d then alt_1 = 3\n when 0x21 then alt_1 = 4\n else\n raise NoViableAlternative( \"\", 1, 0 )\n\n end\n case alt_1\n when 1\n # at line 321:5: '>='\n match( \">=\" )\n\n\n when 2\n # at line 321:10: '<='\n match( \"<=\" )\n\n\n when 3\n # at line 321:15: '=='\n match( \"==\" )\n\n\n when 4\n # at line 321:20: '!='\n match( \"!=\" )\n\n\n when 5\n # at line 321:25: '<'\n match( 0x3c )\n\n when 6\n # at line 321:29: '>'\n match( 0x3e )\n\n end\n\n\n @state.type = type\n @state.channel = channel\n ensure\n # -> uncomment the next line to manually enable rule tracing\n # trace_out( __method__, 17 )\n\n\n end",
"def test_two_proper_args\n assert check_args([1, 1])\n end",
"def test_check_let\n RPN rpn = RPN.new(1)\n let = 'let'\n assert_equal let, 'let'\n assert rpn.check_let(let)\n end",
"def valid_triangle?(a, b, c)\n # Your code goes here!\na+b > c && a+c > b && b+c >a\nend",
"def tuples_for(arg_)\n objdata_ = _get_objdata(arg_)\n objdata_ ? objdata_[2].keys : nil\n end",
"def relation(rel, &relation_definition)\n raise NotImplementedError, 'relation is not supported yet'\n end",
"def get_relational(expr)\n\t\t[\"==\", \"!=\", \">\", \"<\", \">=\", \"<=\"].each do |ope|\n\t\t\tif expr.scan(ope).length == 1\n\t\t\t\treturn ope\n\t\t\tend\n\t\tend\n\tend",
"def tuple_count\n target.count\n end",
"def one_pair?\n n_of_a_kind? 2\n end",
"def test_history_equals_versions\n way = create(:way, :with_history)\n used_way = create(:way, :with_history)\n create(:relation_member, :member => used_way)\n way_with_versions = create(:way, :with_history, :version => 4)\n\n check_history_equals_versions(way.id)\n check_history_equals_versions(used_way.id)\n check_history_equals_versions(way_with_versions.id)\n end",
"def <= other\n call_enum \"relational\", other, :lesseq\n end",
"def check_rule_ast(pt)\n # :defn format: node tag, block name, args, body_0, ..., body_n\n tag, name, args, *body = pt\n return pt if tag != :defn\n\n body.each_with_index do |n,i|\n next if i == 0 and n == s(:nil) # a block got rewritten to an empty block\n\n # Check for a common case\n if n.sexp_type == :lasgn\n return [n, \"illegal operator: '='\"]\n end\n return pt unless n.sexp_type == :call and n.length == 4\n\n # Rule format: call tag, lhs, op, rhs\n _, lhs, op, rhs = n\n\n # Check that LHS references a named collection\n lhs_name = get_qual_name(lhs)\n return [n, \"unexpected lhs format: #{lhs}\"] if lhs_name.nil?\n unless @bud_instance.tables.has_key? lhs_name.to_sym or\n @bud_instance.lattices.has_key? lhs_name.to_sym\n return [n, \"Collection does not exist: '#{lhs_name}'\"]\n end\n\n return [n, \"illegal operator: '#{op}'\"] unless [:<, :<=].include? op\n\n # Check superator invocation. A superator that begins with \"<\" is parsed\n # as a call to the binary :< operator. The right operand to :< is a :call\n # node; the LHS of the :call is the actual rule body, the :call's oper is\n # the rest of the superator (unary ~, -, +), and the RHS is empty. Note\n # that ParseTree encodes unary \"-\" and \"+\" as :-@ and :-+, respectively.\n # XXX: We don't check for illegal superators (e.g., \"<--\"). That would be\n # tricky, because they are encoded as a nested unary op in the rule body.\n if op == :<\n return n unless rhs.sexp_type == :call\n op_tail = rhs[2]\n return n unless [:~, :-@, :+@].include? op_tail\n rhs_args = rhs[3..-1]\n return n unless rhs_args.empty?\n end\n end\n\n return nil # No errors found\n end",
"def to_generate_tuple(&block_)\n if block_\n @gen_tuple = block_\n else\n @gen_tuple\n end\n end",
"def valid_token_tuple?(type, *args)\n args.flatten!\n case type\n when :amazon then valid_amazon_tokens?(*args)\n when :bitlyv1 then valid_bitlyv1_tokens?(*args)\n when :bitlyv2 then valid_bitlyv2_tokens?(*args)\n when :facebook then valid_facebook_tokens?(*args)\n when :flickr then valid_flickr_tokens?(*args)\n when :foursquare then valid_foursquare_tokens?(*args)\n when :google then valid_google_tokens?(*args)\n when :google_oauth then valid_google_oauth_tokens?(*args)\n when :linkedin then valid_linkedin_tokens?(*args)\n when :twitter then valid_twitter_tokens?(*args)\n when :yelpv1 then valid_yelpv1_tokens?(*args)\n when :yelpv2 then valid_yelpv2_tokens?(*args)\n else\n # Rails.logger.error \"Unknown token type: #{type}\"\n end\nend",
"def check_error(val_a, val_b, val_c)\n raise TriangleError if (val_b + val_c <= val_a) || (val_c + val_a <= val_b)\n raise TriangleError if val_a + val_b <= val_c\nend",
"def test_friendp_returns_true_if_users_are_friends\n assert users(:active).friend?(users(:passive))\n assert users(:passive).friend?(users(:active))\n end",
"def valid_triangle?(a, b, c)\n ((a+b>c) && (b+c>a) && (a+c>b))? true:false\nend",
"def to_ruby_literal\n \"Alf::Relation[\" +\n tuples.collect{|t| Tools.to_ruby_literal(t)}.join(', ') + \"]\"\n end",
"def good_tuples(nums)\n count = 0\n (0..nums.size - 3).each do |i|\n first, second, third = nums[i], nums[i + 1], nums[i + 2]\n next if 3 * first == first + second + third\n count += 1 if [first, second, third].include?(first ^ second ^ third)\n\n # next if nums[i] == nums[i + 1] && nums[i + 1] == nums[i + 2]\n # next if nums[i] != nums[i + 1] && nums[i + 1] != nums[i + 2] && nums[i] != nums[i + 2]\n end\n\n count\nend",
"def valid_triangle?(a, b, c)\n a + b > c && a + c > b && b + c > a\nend",
"def valid_triangle?(a, b, c)\n a + b > c && a + c > b && b + c > a\nend",
"def valid_triangle?(a, b, c)\n a + b > c && a + c > b && b + c > a\nend",
"def triplet_check(a,b,c)\n\tif a**2+b**2 == c**2\n\t\treturn true\n\telse\n\t\treturn false\n\tend\nend",
"def visit_axiom_relation_operation_binary(binary)\n generator = self.class.visit(binary)\n @name = generator.name\n @from = aliased_subquery(generator)\n generator\n end",
"def valid_triangle?(a, b, c)\n a + b > c && b + c > a && c + a > b\nend",
"def nested?\n self.resource_tuple.length > 1\n end",
"def right_triangle?(a, b, c)\n a**2 + b**2 == c**2\nend",
"def relationship(*args)\n options = args.extract_options!\n if options[:raise_on_multiple]\n rels = relationships.take(2)\n raise _(\"Multiple relationships found\") if rels.length > 1\n rels.first\n else\n relationships.first\n end\n end",
"def isTriangle(a,b,c)\n (a+b>c && a+c>b && c+b>a)? true : false\nend",
"def >= other\n call_enum \"relational\", other, :moreeq\n end",
"def isTriangle(a, b, c)\n a, b, c = [a, b, c].sort\n a + b > c\nend",
"def test_get_location_bad_input\n assert_raises 'Improper typing of location input argument in method get_location' do\n locations = LocationRelation.new\n locations.get_location 55\n end\n end",
"def insert(tuples)\n pks = tuples.map { |tuple| relation.insert(tuple) }\n relation.where(relation.primary_key => pks).to_a\n end",
"def to_ruby_literal\n \"Alf::Relation([\" + tuples.map{|t| Support.to_ruby_literal(t) }.join(', ') + \"])\"\n end",
"def evaluate\n distinct? ? distinct_relation : relation\n end",
"def check_permission_with_http_info(opts = {})\n if @api_client.config.debugging\n @api_client.config.logger.debug 'Calling API: PermissionApi.check_permission ...'\n end\n # resource path\n local_var_path = '/relation-tuples/check/openapi'\n\n # query parameters\n query_params = opts[:query_params] || {}\n query_params[:'namespace'] = opts[:'namespace'] if !opts[:'namespace'].nil?\n query_params[:'object'] = opts[:'object'] if !opts[:'object'].nil?\n query_params[:'relation'] = opts[:'relation'] if !opts[:'relation'].nil?\n query_params[:'subject_id'] = opts[:'subject_id'] if !opts[:'subject_id'].nil?\n query_params[:'subject_set.namespace'] = opts[:'subject_set_namespace'] if !opts[:'subject_set_namespace'].nil?\n query_params[:'subject_set.object'] = opts[:'subject_set_object'] if !opts[:'subject_set_object'].nil?\n query_params[:'subject_set.relation'] = opts[:'subject_set_relation'] if !opts[:'subject_set_relation'].nil?\n query_params[:'max-depth'] = opts[:'max_depth'] if !opts[:'max_depth'].nil?\n\n # header parameters\n header_params = opts[:header_params] || {}\n # HTTP header 'Accept' (if needed)\n header_params['Accept'] = @api_client.select_header_accept(['application/json'])\n\n # form parameters\n form_params = opts[:form_params] || {}\n\n # http body (model)\n post_body = opts[:debug_body]\n\n # return_type\n return_type = opts[:debug_return_type] || 'CheckPermissionResult'\n\n # auth_names\n auth_names = opts[:debug_auth_names] || ['oryAccessToken']\n\n new_options = opts.merge(\n :operation => :\"PermissionApi.check_permission\",\n :header_params => header_params,\n :query_params => query_params,\n :form_params => form_params,\n :body => post_body,\n :auth_names => auth_names,\n :return_type => return_type\n )\n\n data, status_code, headers = @api_client.call_api(:GET, local_var_path, new_options)\n if @api_client.config.debugging\n @api_client.config.logger.debug \"API called: PermissionApi#check_permission\\nData: #{data.inspect}\\nStatus code: #{status_code}\\nHeaders: #{headers}\"\n end\n return data, status_code, headers\n end",
"def with_input_tuples(tuples, &block)\n input_tuples = Array([tuples]).flatten(1).map\n return input_tuples unless block_given?\n\n input_tuples.each(&block)\n end",
"def test_two_valid_arguments\n\t\tc = Check.new\n\t\trefute c.check_arguments([1,1])\n\tend",
"def test_triangle\n if @side1.eql?(@side2) && @side2.eql?(@side3)\n \"It's an equilateral\"\n elsif @side1.eql?(@side2) || @side2.eql?(@side3)\n \"It's an isosceles\"\n elsif @side1.eql?(@side2) == false && @side2.eql?(@side3) == false\n \"It's a scalene triangle\"\n else\n \"Not a triangle\"\n end\n end",
"def parse_relational\n add = parse_add\n if consume(:lt)\n Node.new(:binary, op: :lt, lhs: add, rhs: parse_relational)\n elsif consume(:lteq)\n Node.new(:binary, op: :lteq, lhs: add, rhs: parse_relational)\n elsif consume(:gt)\n Node.new(:binary, op: :gt, lhs: add, rhs: parse_relational)\n elsif consume(:gteq)\n Node.new(:binary, op: :gteq, lhs: add, rhs: parse_relational)\n else\n add\n end\n end",
"def pitagoric_triplet?(a,b,c)\n (a**2 + b**2) == c**2\nend",
"def selecionErr(method)\n\n tnTableCreation('tuple_node_test_result') if @is_new\n\n if @unwanted_tuple_count + @missing_tuple_count == 0\n p 'no failed rows found. There is no selection error'\n return\n end\n\n if @unwanted_tuple_count > 0\n # p \"Unwanted Pk count #{unWantedPK.count()}\"\n # create unwanted_tuple_branch table\n # binding.pry\n\n # whereErrList = whereCondTest(@unWantedPK, 'U')\n where_cond_test('U')\n # joinErrList = jointypeErr(query,'U')\n end\n\n if @missing_tuple_count > 0\n # p \"Missing PK count #{missinPK.count()}\"\n # binding.pry\n # whereErrList = whereCondTest(@missingPK, 'M')\n where_cond_test('M')\n # joinErrList = jointypeErr(query,'M')\n end\n # create aggregated tuple_suspicious_nodes\n pk = @pkFullList.map { |pk| pk['alias'] }.join(',')\n query = 'create materialized view tuple_node_test_result_aggr as '\\\n \"select #{pk}, string_agg(branch_name||'-'||node_name, ',' order by node_name,branch_name) as suspicious_nodes from tuple_node_test_result group by #{pk}\"\n pp query\n DBConn.exec(query)\n\n unless @predicateTree.nil?\n suspicious_score_upd(@predicateTree.branches)\n end\n # exnorate algorithm\n # binding.pry\n @column_combinations = method.start_with?('o') ? Columns_Combination.new(@all_columns) : @all_columns\n case method\n # when 'o'\n # puts 'old exonerate algorithm'\n\n # true_query_PT_construct\n # constraint_query = constraint_predicate_construct\n # # column_combinations_construct\n # tuple_mutation_test(@missingPK, 'M', constraint_query, false)\n # tuple_mutation_test(@unWantedPK, 'U', constraint_query, false)\n when 'or'\n puts 'old exonerate algorithm with duplicate removal'\n # reset suspicious score\n query = \"update node_query_mapping set suspicious_score = 0 where type = 'f'\"\n res = DBConn.exec(query)\n\n true_query_PT_construct\n constraint_query = constraint_predicate_construct\n # allcolumns_construct()\n puts 'Missing starts'\n puts Time.now()\n tuple_mutation_test_with_dup_removal('M', constraint_query)\n puts 'Unwanted starts'\n puts Time.now()\n\n tuple_mutation_test_with_dup_removal('U', constraint_query)\n # when 'n'\n # puts 'new exonerate algorithm'\n # true_query_PT_construct\n # constraint_query = constraint_predicate_construct\n # tuple_mutation_test_reverse(@missingPK, 'M', constraint_query, false)\n # tuple_mutation_test_reverse(@unWantedPK, 'U', constraint_query, false)\n\n when 'b'\n puts 'baseline'\n else\n puts 'Unknown method'\n end\n # remove constraint_nodes in node_query_mapping\n query = \"delete from node_query_mapping where test_id = #{@test_id} and type = 't'\"\n DBConn.exec(query)\n\n # j = Hash.new\n # # j['JoinErr'] = joinErrList if joinErrList.count >0\n # j['WhereErr'] = whereErrList\n # j\n # whereErrList\n end",
"def isTriangle(a,b,c)\n a+b>c && a + c > b && b + c > a ? true : false\nend",
"def valid_triangle?(side1, side2, side3)\n\n triangle_sides=[side1, side2, side3].sort\n\n p (triangle_sides[0] + triangle_sides[1]) > triangle_sides[2]\n\nend"
] |
[
"0.70167816",
"0.66043293",
"0.6362414",
"0.6198588",
"0.6017634",
"0.57891464",
"0.56978273",
"0.5597392",
"0.55021274",
"0.5360187",
"0.53194505",
"0.5232657",
"0.51772624",
"0.5164011",
"0.51385814",
"0.50963265",
"0.50963265",
"0.5095716",
"0.50808156",
"0.5077837",
"0.50561583",
"0.50467813",
"0.5042676",
"0.5023282",
"0.4995738",
"0.4936109",
"0.49294013",
"0.49169928",
"0.4906028",
"0.48622432",
"0.48228234",
"0.4789072",
"0.4755289",
"0.47503346",
"0.47308597",
"0.46896136",
"0.46785426",
"0.46676442",
"0.46456912",
"0.46250802",
"0.46217775",
"0.46068227",
"0.46021903",
"0.45963222",
"0.45890284",
"0.45746356",
"0.4573108",
"0.45624387",
"0.45601568",
"0.45557684",
"0.45524535",
"0.45376578",
"0.45328152",
"0.45319912",
"0.45270094",
"0.4520923",
"0.4518909",
"0.45067638",
"0.44996852",
"0.44976512",
"0.44595206",
"0.44561327",
"0.44512",
"0.4446724",
"0.44466165",
"0.4440338",
"0.4429359",
"0.44253746",
"0.4413978",
"0.44123283",
"0.44101584",
"0.4403991",
"0.44039664",
"0.44037217",
"0.44030228",
"0.4400129",
"0.43997043",
"0.43997043",
"0.43997043",
"0.43969977",
"0.43949974",
"0.4390133",
"0.43899447",
"0.43873096",
"0.43844968",
"0.43760037",
"0.43710077",
"0.43677223",
"0.43654543",
"0.43632632",
"0.43567508",
"0.43544686",
"0.4347883",
"0.43457636",
"0.43446526",
"0.43410885",
"0.43397266",
"0.43383962",
"0.43377355",
"0.4333266",
"0.4332013"
] |
0.0
|
-1
|
Attempts to return the schema from table_name and table_name_prefix
|
def schema_name
return self.table_name.split('.')[0] if self.table_name && self.table_name =~ /\./
return self.table_name_prefix.to_s.gsub(/\./,'') if self.table_name_prefix && self.table_name_prefix =~ /\./
return self.connection.config[:database] if self.connection.mysql?
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def schema_and_table_name\n if qualified_table_name.include? '.'\n schema_name, table_name = qualified_table_name.split('.', 2)\n else\n table_name = qualified_table_name\n schema_name = self.class.default_schema_name\n end\n [schema_name, table_name]\n end",
"def extract_schema_and_table(name)\n schema, table = name.split('.', 2)\n\n unless table # A table was provided without a schema\n table = schema\n schema = nil\n end\n\n if name =~ /^\"/ # Handle quoted table names\n table = name\n schema = nil\n end\n [schema, table]\n end",
"def schema_and_table(table_name)\n schema_utility_dataset.schema_and_table(table_name)\n end",
"def get_table_schema(table)\n if (@options[:redshift_schema] == nil)\n raise 'Must pass :redshift_schema as option when creating RedTrack client'\n end\n\n schema = @options[:redshift_schema]\n\n if schema[table.to_sym]\n result = schema[table.to_sym]\n elsif schema[\"#{table}\"]\n result = schema[\"#{table}\"]\n end\n\n return result\n end",
"def roomer_full_table_name_prefix(schema_name)\n \"#{schema_name.to_s}#{Roomer.schema_seperator}\"\n end",
"def schema(table_name = nil, opts={})\n if opts[:reload] && @schemas\n if table_name\n @schemas.delete(table_name)\n else\n @schemas = nil\n end\n end\n\n if table_name\n return @schemas[table_name] if @schemas && @schemas[table_name]\n else\n return @schemas if @schemas\n end\n\n if table_name\n @schemas ||= {}\n @schemas[table_name] ||= schema_parse_table(table_name, opts)\n else\n @schemas = schema_parse_tables(opts)\n end\n end",
"def schema(table_name = nil, opts={})\n table_name = table_name.to_sym if table_name\n if opts[:reload] && @schemas\n if table_name\n @schemas.delete(table_name)\n else\n @schemas = nil\n end\n end\n\n if @schemas\n if table_name\n return @schemas[table_name] if @schemas[table_name]\n else\n return @schemas\n end\n end\n\n if table_name\n @schemas ||= {}\n if respond_to?(:schema_parse_table, true)\n @schemas[table_name] ||= schema_parse_table(table_name, opts)\n else\n raise Error, 'schema parsing is not implemented on this database'\n end\n else\n if respond_to?(:schema_parse_tables, true)\n @schemas = schema_parse_tables(opts)\n elsif respond_to?(:schema_parse_table, true) and respond_to?(:tables, true)\n tables.each{|t| schema(t, opts)}\n @schemas\n else\n raise Error, 'schema parsing is not implemented on this database'\n end\n end\n end",
"def table_name_prefix; end",
"def tables(name = nil, schema = nil, table = nil)\n schema = schema ? \"'#{quote_string(schema)}'\" : 'CURRENT_SCHEMA'\n select_rows(\n \"SELECT table_name \"+\n \"FROM information_schema.tables \"+\n \"WHERE table_type = 'TABLE' \"+\n \" AND table_schema = #{schema} \"+\n (table ? \"AND table_name = '#{quote_string(table)}'\" : \"\"),\n SCHEMA_LOG_NAME\n ).map { |row|\n row[0]\n }\n end",
"def schema_prefix\n ''\n end",
"def schema_and_table(result)\n :\"#{result.schema}__#{self.implicit_table_name}\"\n end",
"def table_name_prefix(model)\n return model::Base.table_name_prefix rescue \"\"\n end",
"def table_name\n @table_name ||= begin\n schema = self.class.schema_name\n table = self.sqltablename.blank? ? \"PS_#{self.recname}\" : self.sqltablename\n [schema, table].compact.join(\".\")\n end\n end",
"def schema_names\n select_value(<<-SQL, 'SCHEMA')\n SELECT nspname\n FROM pg_namespace\n WHERE nspname !~ '^pg_.*'\n AND nspname NOT IN ('information_schema')\n ORDER by nspname;\n SQL\n end",
"def roomer_set_table_name_prefix\n self.table_name_prefix = begin\n case @roomer_scope\n when :shared\n roomer_full_table_name_prefix(Roomer.shared_schema_name)\n when :tenanted\n roomer_full_table_name_prefix(Roomer.current_tenant.try(Roomer.tenant_schema_name_column))\n else\n \"\"\n end\n end\n end",
"def schema_names\n select_values(\n \"SELECT nspname FROM pg_namespace\" <<\n \" WHERE nspname !~ '^pg_.*' AND nspname NOT IN ('information_schema')\" <<\n \" ORDER by nspname;\",\n 'SCHEMA')\n end",
"def namespace_equivalent_for_db_server\n # HACK!\n return connection.respond_to?(:postgresql_connection) ? 'schema' : 'database'\n end",
"def tables\n execute('select table_name from information_schema.tables where table_schema = ?', [schema]).map(&:first)\n end",
"def schema_lookup(schema_name)\n s = nil\n @current_domain.schemas.each do |schema|\n s = schema if schema.name == schema_name\n end\n s\n end",
"def parse_schema_and_table_name table_name\n return [nil, table_name] if table_name.count('.') == 0\n\n if table_name.count('.') > 1\n raise \"Invalid relation reference #{table_name} (only one '.' is allowed)\"\n end\n\n schema_name, table_name = table_name.split('.')\n [schema_name, table_name]\n end",
"def schema_ds_from(table_name, opts)\n [:information_schema__tables___t]\n end",
"def namespaced_table_name\n [self.class.table_namespace, table_name].reject {|s| s.nil? || s.empty?}.join('-')\n end",
"def schema\n schema = Schema.new\n\n for name in db.tables\n next if IGNORED_TABLES.include? name\n table = schema.add_table( name )\n import_table( table )\n end\n\n schema\n end",
"def schema_name\n unless @schema_name\n @schema_name = collect_first(&:schema_name)\n end\n return @schema_name\n end",
"def schema_name\n @schema_name ||= owner_type.eql?('Root') ? owner.tenant.schema_name : schema_name_from_access_key_id\n end",
"def table_prefix component, use_table_name = nil\n case use_table_name\n when false, nil\n ''\n when true\n safe_table_name(component)+'.'\n else\n use_table_name+'.'\n end\n end",
"def schema(name)\n get(\"schemas/#{name}/\", \"schema\")\n end",
"def schemas\n sql = \"SELECT nspname FROM pg_namespace WHERE nspname !~ '^pg_.*'\"\n ActiveRecord::Base.connection.query(sql).flatten\n end",
"def schema_name=schema_name\n self.table_name_prefix = \"#{schema_name}.\" if schema_name && !schema_name.blank?\n self.table_name = \"#{self.table_name_prefix}#{self.table_name}\" unless self.abstract_class?\n end",
"def table_structure(table_name)\n execute('select * from information_schema.columns where table_schema = ?' \\\n 'AND table_name = ?', [schema, table_name])\n end",
"def schema_name_for(statement_name)\n unless statement_name =~ STATEMENT_NAME_REGEXP\n raise ArgumentError, \"statement '#{statement_name}' name must match #{STATEMENT_NAME_REGEXP.inspect}\"\n end\n const_name = statement_name.to_s\n if const_name == \"schema\"\n \"#{self.name}::Schema\"\n else\n \"#{self.name}::#{const_name.camelize}Schema\"\n end\n end",
"def unqualified(table_name)\n table_name.split('.').last\n end",
"def tables(name = nil)\n if name\n ActiveSupport::Deprecation.warn(<<-MSG.squish)\n Passing arguments to #tables is deprecated without replacement.\n MSG\n end\n\n select_values(\"SELECT tablename FROM pg_tables WHERE schemaname = ANY(current_schemas(false))\", 'SCHEMA')\n end",
"def schema\n execute(<<-eosql).collect { |row| row[0] }.collect { |t| table_schema(t) }\nSELECT rdb$relation_name FROM rdb$relations WHERE rdb$system_flag != 1\neosql\n end",
"def schema_search_path\n 'dbo'\n end",
"def schema\n @schema ||= (default_schema || ETL::Schema::Table.new)\n end",
"def table_name_prefix \n @table_name_prefix ||= 'gricer_'\n end",
"def base_tables(name = nil)\n # this is untested\n select_values(\"SELECT table_name FROM information_schema.tables\", name)\n end",
"def clean_table_name(table_name)\n class_from_table_name(table_name)&.table_name\n end",
"def get_tables\n get_schemas.keys\n end",
"def schema_exists?(name)\n select_value(\"SELECT COUNT(*) FROM pg_namespace WHERE nspname = '#{name}'\", 'SCHEMA').to_i > 0\n end",
"def schema_exists?(name)\n select_value(\"SELECT COUNT(*) FROM pg_namespace WHERE nspname = '#{name}'\", 'SCHEMA').to_i > 0\n end",
"def schema\n absolutize(@schema)\n end",
"def schema\n absolutize(@schema)\n end",
"def show_tables\n execute('select table_name from information_schema.tables where table_schema = ?', [schema])\n end",
"def prepare_schema\n tables = [];ActiveRecord::Base.connection.execute(\"show tables\").each{|t| tables << t[0].strip}\n \n ActiveRecord::Schema.define do\n App.log.info(\"preparing schema\")\n \n unless tables.include?(\"services\")\n # a service entry\n begin\n create_table :services do |t|\n t.string :name\n t.string :status, :null => false, :default => \"active\"\n end\n add_index :services, :name\n rescue\n App.log_exception\n end\n end\n \n unless tables.include?(\"queries\")\n begin\n # queries\n create_table :queries do |t|\n t.string :query \n t.column :last_twid, :bigint, :null => false, :default => 0\n t.timestamp :last_run\n t.integer :last_result_count\n t.string :status, :default => 'active', :null=> false\n end\n add_index :queries, :query\n rescue\n App.log_exception\n end\n end\n \n unless tables.include?(\"tweets\")\n begin\n # cache of tweets\n create_table :tweets do |t|\n t.column :twid, :bigint, :null => false\n t.string :from_user\n t.string :to_user\n t.integer :from_user_id\n t.integer :to_user_id\n t.string :text\n t.string :profile_image_url\n t.timestamp :created_at\n end\n add_index :tweets, :twid\n rescue\n App.log_exception\n end\n end\n \n end # define schema\n end",
"def prefix_table\n update unless @prefix_table\n @prefix_table\n end",
"def qualified(table_name, context:)\n return table_name if table_name['.'].present?\n return table_name if context['.'].blank?\n\n \"#{context.sub(%r{[^.]*$}, '')}#{table_name}\"\n end",
"def tables(_name = nil)\n stmt = @connection.tables\n result = stmt.fetch_all || []\n stmt.drop\n\n result.each_with_object([]) do |row, table_names|\n schema_name, table_name, table_type = row[1..3]\n next if respond_to?(:table_filtered?) && table_filtered?(schema_name, table_type)\n table_names << format_case(table_name)\n end\n end",
"def tables(_name = nil)\n stmt = @connection.tables\n result = stmt.fetch_all || []\n stmt.drop\n\n result.each_with_object([]) do |row, table_names|\n schema_name, table_name, table_type = row[1..3]\n next if respond_to?(:table_filtered?) && table_filtered?(schema_name, table_type)\n table_names << format_case(table_name)\n end\n end",
"def singular_table_name_from x\n x.to_s.tableize.singularize\n end",
"def current_schema\n exec_query('SELECT current_schema', 'SCHEMA')[0][\"current_schema\"]\n end",
"def real_table_name\n packageid.gsub('.', '_')\n end",
"def shared_schema_name\n Roomer.shared_schema_name ||= options[:shared_schema_name].to_s\n end",
"def current_schema\n select_value('SELECT current_schema', 'SCHEMA')\n end",
"def schema_parse_table(table_name, opts)\n m = output_identifier_meth(opts[:dataset])\n\n _schema_ds.where_all(Sequel[:pg_class][:oid]=>regclass_oid(table_name, opts)).map do |row|\n row[:default] = nil if blank_object?(row[:default])\n if row[:base_oid]\n row[:domain_oid] = row[:oid]\n row[:oid] = row.delete(:base_oid)\n row[:db_domain_type] = row[:db_type]\n row[:db_type] = row.delete(:db_base_type)\n else\n row.delete(:base_oid)\n row.delete(:db_base_type)\n end\n\n db_type = row[:db_type]\n row[:type] = if row.delete(:is_array)\n schema_array_type(db_type)\n else\n send(TYPTYPE_METHOD_MAP[row.delete(:typtype)], db_type)\n end\n identity = row.delete(:attidentity)\n if row[:primary_key]\n row[:auto_increment] = !!(row[:default] =~ /\\A(?:nextval)/i) || identity == 'a' || identity == 'd'\n end\n\n # :nocov:\n if server_version >= 90600\n # :nocov:\n case row[:oid]\n when 1082\n row[:min_value] = MIN_DATE\n row[:max_value] = MAX_DATE\n when 1184, 1114\n if Sequel.datetime_class == Time\n row[:min_value] = MIN_TIMESTAMP\n row[:max_value] = MAX_TIMESTAMP\n end\n end\n end\n\n [m.call(row.delete(:name)), row]\n end\n end",
"def table_name\n if qualified? && meta[:qualified].is_a?(Symbol)\n meta[:qualified]\n else\n source.dataset\n end\n end",
"def quote_table_name(name)\n schema, name_part = extract_vertica_identifier_from_name(name.to_s)\n\n unless name_part\n quote_column_name(schema)\n else\n table_name, name_part = extract_vertica_identifier_from_name(name_part)\n \"#{quote_column_name(schema)}.#{quote_column_name(table_name)}\"\n end\n end",
"def table_name\n self.name.split('::').last\n end",
"def parent_table_schema_name(*partition_key_values)\n return collect_first(*partition_key_values, &:parent_table_schema_name)\n end",
"def create_schema schema_name\n execute \"CREATE SCHEMA #{quote_schema_name(schema_name)}\"\n end",
"def schema_parse_table(table_name, opts)\n m = output_identifier_meth\n m2 = input_identifier_meth\n ds = metadata_dataset.from(:information_schema__tables___t).\n join(:information_schema__columns___c, :table_catalog=>:table_catalog,\n :table_schema => :table_schema, :table_name => :table_name).\n select(:column_name___column, :data_type___db_type, :character_maximum_length___max_chars, :column_default___default, :is_nullable___allow_null).\n filter(:c__table_name=>m2.call(table_name.to_s))\n if schema = opts[:schema] || default_schema\n ds.filter!(:table_schema=>schema)\n end\n ds.map do |row|\n row[:allow_null] = row[:allow_null] == 'YES' ? true : false\n row[:default] = nil if blank_object?(row[:default])\n row[:type] = schema_column_type(row[:db_type])\n [m.call(row.delete(:column)), row]\n end\n end",
"def get_table_name\n self.class.table_name\n end",
"def create_table_prefix_sql(name, options)\n \"CREATE #{temporary_table_sql if options[:temp]}TABLE#{' IF NOT EXISTS' if options[:if_not_exists]} #{options[:temp] ? quote_identifier(name) : quote_schema_table(name)}\"\n end",
"def chooseSchema\n @metadata.chooseSchema\n end",
"def quote_schema_table(table)\n schema_utility_dataset.quote_schema_table(table)\n end",
"def quote_schema_table(table)\n schema_utility_dataset.quote_schema_table(table)\n end",
"def current_schema\n select_value('SELECT current_schema()', 'SCHEMA')\n end",
"def compute_table_name\n \"#{full_table_name_prefix}#{undecorated_table_name(name)}#{table_name_suffix}\"\n end",
"def parent_tables(table_name)\n result = exec_query(<<-SQL, 'SCHEMA')\n SELECT pg_namespace.nspname, pg_class.relname\n FROM pg_catalog.pg_inherits\n INNER JOIN pg_catalog.pg_class ON (pg_inherits.inhparent = pg_class.oid)\n INNER JOIN pg_catalog.pg_namespace ON (pg_class.relnamespace = pg_namespace.oid)\n WHERE inhrelid = '#{table_name}'::regclass\n SQL\n result.map { |a| a['relname'] }\n end",
"def current_schema\n select_value(<<~SQL.squish, \"SCHEMA\")\n SELECT SYS_CONTEXT('userenv', 'current_schema') FROM dual\n SQL\n end",
"def tables(name = nil)\n query(<<-SQL, 'SCHEMA').map { |row| row[0] }\n SELECT tablename\n FROM pg_tables\n WHERE schemaname = ANY (ARRAY['public'])\n SQL\n end",
"def schema\n @schema ||= metadata.ancestors('Schema').first\n end",
"def pgt_mangled_table_name(table)\n quote_schema_table(table).gsub('\"', '').gsub(/[^A-Za-z0-9]/, '_').gsub(/_+/, '_')\n end",
"def tables(name = nil, database = nil, like = nil)\n database = database ? quote_table_name(database) : \"DATABASE()\"\n by_name = like ? \"AND table_name LIKE #{quote(like)}\" : \"\"\n\n sql = <<-SQL.squish\n SELECT table_name, table_type\n FROM information_schema.tables\n WHERE table_schema = #{database}\n AND table_type = 'BASE TABLE'\n #{by_name}\n SQL\n\n execute_and_free(sql, 'SCHEMA') do |result|\n rows_from(result).map(&:first)\n end\n end",
"def create_schema(name)\n sql = %{CREATE SCHEMA \"#{name}\"}\n ActiveRecord::Base.connection.execute sql\n end",
"def pg_table_name\n table_name\n end",
"def get_schema_struct(table_name)\n dbres = do_sql_command(\"DESC #{table_name};\")\n\n dbstruct = []\n\n if(dbres) then\n dbres.each_hash do | row |\n dbstruct_hash = {}\n row.each {|key, val|\n dbstruct_hash[key.downcase.to_sym] = val \n }\n dbstruct << dbstruct_hash\n end \n end\n\n dbstruct\nend",
"def create_schema(schema)\n ActiveRecord::Base.connection.execute(\"CREATE SCHEMA #{schema}\")\n end",
"def schema(work_class_name:, context: nil)\n context ||= AllinsonFlex::Context.where(name: 'default')\n AllinsonFlex::DynamicSchema.where(\n allinson_flex_class: work_class_name,\n context: context\n ).order('created_at').last.schema\n rescue StandardError\n {}\n end",
"def prefix_for_tables\n migration_version = ActiveRecord::Migrator.current_version\n end",
"def fdw_adjusted_table_name(name)\n name[0...PG_MAX_TABLE_NAME_LENGTH]\n end",
"def sqlite3_schema(table_name)\n table_schema = @dbm.sqlite3_pragma(table_name)\n\n # First, find the max lengths of each column's title strings\n # for nice print out in a fixed length of table format\n max_column_name_length = 0\n max_type_name_col_length = 22\n max_nullable_col_length = 10\n max_pk_col_length = 5\n\n table_schema.each {|schema|\n if max_column_name_length < schema[1].length\n max_column_name_length = schema[1].length\n end\n \n # Check null status and convert the stored binary value\n # into string \"YES\" or \"NO\"\n schema[3] = (schema[3] == 1 ? \"YES\":\"NO\")\n # Check primary key status and convert the stored\n # binary value into string \"YES\" or \"NO\"\n schema[5] = (schema[5] == 1 ? \"YES\":\"NO\")\n }\n\n # Print out the schema in a table format\n puts table_name\n printf \"%-5s | %-#{max_column_name_length}s | %-22s | %-5s | %-5s\\n\",\\\n \"CID\", \"NAME\", \"TYPE\", \"NULL?\", \"PK?\"\n dash_line_splitter = '-' * (50 + max_column_name_length)\n puts dash_line_splitter\n table_schema.each {|schema|\n printf \"%-5s | %-#{max_column_name_length}s | %-22s | %-5s | %-5s\\n\",\\\n schema[0].to_s, schema[1], schema[2], schema[3], schema[5]\n }\n end",
"def schema_tables\n @schema_tables ||= %i[\n projects\n amazon_clouds\n data_centers\n jmeter_plans\n load_agents\n target_hosts\n clusters\n execution_cycles\n client_stats\n page_stats\n target_stats\n jtl_files\n ]\n end",
"def identifier(schema, name = nil)\n name ||= @name\n \"`#{schema}`.`#{name}`\"\n end",
"def table_name_prefix\n \"ext_#{self.registered_name.to_s.underscore}_\"\n end",
"def _table_name\n self.class.table_name\n end",
"def schema_parse_table(table_name, opts)\n m = output_identifier_meth(opts[:dataset])\n\n table = if opts[:schema]\n Sequel.qualify(opts[:schema], table_name)\n else\n Sequel.identifier(table_name)\n end\n\n describe(table, opts).map do |row|\n row[:db_type] = row[:type]\n row[:type] = schema_column_type(row[:db_type])\n row[:default] = nil\n row[:primary_key] = false\n [m.call(row.delete(:name)), row]\n end\n end",
"def all_schemas\n query('SELECT schema_name FROM information_schema.schemata').flatten\n end",
"def get_schema_id()\n\t\tend",
"def table_name(table_name = nil)\n if table_name\n @table_name = table_name\n else\n @table_name\n end\n end",
"def get_schema schema_name, view, options = {}\n schema_view = Google::Cloud::PubSub::V1::SchemaView.const_get view.to_s.upcase\n schemas.get_schema name: schema_path(schema_name, options),\n view: schema_view\n end",
"def get_schema_sql(table_struct, table_name = NEW_TABLE_NAME)\n dbstruct = []\n pkeys = []\n\n table_struct.each do | row |\n dbstruct << \"`#{row[:field]}` #{row[:type]} #{(!row[:default].nil? && row[:default] != '' ) ? \"default '#{row[:default]}'\" : ''} #{row[:null] == 'NO' ? 'NOT NULL' : ''}\"\n pkeys << \"`#{row[:field]}`\" if row[:key] == 'PRI'\n end \n\n dbstruct << \"PRIMARY KEY (%s)\" % [pkeys.join(', ')]\n dbstring = \"CREATE TABLE `%s` (\\n\\t%s\\n)\" % [table_name, dbstruct.join(\",\\n\\t\")]\n\n dbstring\nend",
"def classify(table_name)\n # strip out any leading schema name\n camelize(singularize(table_name.to_s.sub(/.*\\./, '')))\n end",
"def create_schema_sql(name, opts=OPTS)\n \"CREATE SCHEMA #{'IF NOT EXISTS ' if opts[:if_not_exists]}#{quote_identifier(name)}#{\" AUTHORIZATION #{literal(opts[:owner])}\" if opts[:owner]}\"\n end",
"def tables(name = nil)\n select_rows(<<-SQL, name).map { |row| row[0] }\nSELECT name\nFROM sysobjects\nWHERE type = 'U'\nSQL\n end",
"def schema_parse_table(table, opts={})\n jdbc_schema_parse_table(table, opts)\n end",
"def tables(name = nil) #:nodoc:\r\n sql = \"EXECUTE PROCEDURE sp_GetTables( NULL, NULL, NULL, 'TABLE' );\"\r\n select(sql, name).map { |row| strip_or_self(row[\"TABLE_NAME\"]) }\r\n end",
"def schema\n if @old_schema.nil?\n @old_schema = Schema.new(get(link('schema')))\n end\n return @old_schema \n end",
"def create_schema(schema)\n execute \"CREATE SCHEMA #{schema}\", 'Create Schema'\n end"
] |
[
"0.7890468",
"0.766686",
"0.7419078",
"0.73171866",
"0.71402097",
"0.70742774",
"0.6926123",
"0.6907323",
"0.68817",
"0.68221694",
"0.67872965",
"0.6696679",
"0.6693215",
"0.6641224",
"0.662283",
"0.6589934",
"0.6589023",
"0.65569425",
"0.652769",
"0.6499903",
"0.6474521",
"0.6466262",
"0.6459383",
"0.6455617",
"0.64454263",
"0.6389734",
"0.63774127",
"0.6371248",
"0.63639987",
"0.6276949",
"0.6256685",
"0.6217615",
"0.6213644",
"0.61660355",
"0.614267",
"0.61216044",
"0.6121386",
"0.60651386",
"0.6030288",
"0.60280055",
"0.60016406",
"0.5970438",
"0.59377146",
"0.59377146",
"0.5923414",
"0.5910985",
"0.5905375",
"0.58916116",
"0.588726",
"0.5884871",
"0.5872151",
"0.5868284",
"0.5862812",
"0.58606875",
"0.58538604",
"0.58217835",
"0.5815061",
"0.5809723",
"0.5808724",
"0.58083063",
"0.5803218",
"0.5794697",
"0.5755989",
"0.5754436",
"0.57415456",
"0.5734379",
"0.5734379",
"0.5724696",
"0.57144475",
"0.5693247",
"0.56795293",
"0.5675579",
"0.56701845",
"0.5664668",
"0.5664569",
"0.56466347",
"0.56465864",
"0.5645544",
"0.56392354",
"0.5636541",
"0.56343603",
"0.5633819",
"0.56293625",
"0.56095856",
"0.5605394",
"0.56019306",
"0.5600614",
"0.5590795",
"0.55892",
"0.55844706",
"0.55624044",
"0.5544506",
"0.5539551",
"0.5532411",
"0.5531705",
"0.55296713",
"0.55241144",
"0.55236953",
"0.55235505",
"0.55210876"
] |
0.76656115
|
2
|
Set the unformatted schema name for the given model / connection class EX: class User 'users_db' User.table_name_prefix => 'users_db.' User.table_name => 'users_db.users'
|
def schema_name=schema_name
self.table_name_prefix = "#{schema_name}." if schema_name && !schema_name.blank?
self.table_name = "#{self.table_name_prefix}#{self.table_name}" unless self.abstract_class?
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def set_schema(schema)\n define_class_method(:schema) {schema.to_s.to_sym}\n end",
"def schema_name\n return self.table_name.split('.')[0] if self.table_name && self.table_name =~ /\\./\n return self.table_name_prefix.to_s.gsub(/\\./,'') if self.table_name_prefix && self.table_name_prefix =~ /\\./\n return self.connection.config[:database] if self.connection.mysql?\n end",
"def roomer_full_table_name_prefix(schema_name)\n \"#{schema_name.to_s}#{Roomer.schema_seperator}\"\n end",
"def __create_meta_schema_class klass\n binder = self.name\n self.const_set(klass,\n Class.new(ActiveRecord::Base) do # class MetaSchema < ActiveRecord::Base\n singleton_class.send(:define_method, :table_name_prefix) do # def self.table_name_prefix\n \"#{binder.underscore}_\" # \"foo_binder\" # If we have a class FooBinder < Binder::AR\n end # end\n end) # end\n end",
"def roomer_set_table_name_prefix\n self.table_name_prefix = begin\n case @roomer_scope\n when :shared\n roomer_full_table_name_prefix(Roomer.shared_schema_name)\n when :tenanted\n roomer_full_table_name_prefix(Roomer.current_tenant.try(Roomer.tenant_schema_name_column))\n else\n \"\"\n end\n end\n end",
"def table_name_prefix(model)\n return model::Base.table_name_prefix rescue \"\"\n end",
"def undecorated_table_name(class_name = base_class.name)\n table_name = Inflector.underscore(Inflector.demodulize(class_name))\n table_name = Inflector.pluralize(table_name) if pluralize_table_names\n table_name\n end",
"def undecorated_table_name(class_name = base_class.name)\n table_name = class_name.to_s.demodulize.underscore\n table_name = table_name.pluralize if pluralize_table_names\n table_name\n end",
"def model_table_name\n model_name = model_param.name\n model_class = model_name.demodulize\n model_class.tableize\n end",
"def schema_name_for(statement_name)\n unless statement_name =~ STATEMENT_NAME_REGEXP\n raise ArgumentError, \"statement '#{statement_name}' name must match #{STATEMENT_NAME_REGEXP.inspect}\"\n end\n const_name = statement_name.to_s\n if const_name == \"schema\"\n \"#{self.name}::Schema\"\n else\n \"#{self.name}::#{const_name.camelize}Schema\"\n end\n end",
"def schema_prefix\n ''\n end",
"def table_name_prefix; end",
"def table_name\n self.class.name.demodulize.tableize\n end",
"def table_name\n self.to_s.pluralize.underscore\n end",
"def schema(s)\n @config[:schema] = s\n end",
"def table_name\n @table_name ||= self.class.const_get(class_name).table_name\n end",
"def format_model_schema(klass)\n hsh = klass.columns_hash.map { |column_name,column| [column_name, column.type] }.to_h\n format_schema_helper(hsh)\n end",
"def table_name_prefix\n \"ext_#{self.registered_name.to_s.underscore}_\"\n end",
"def tableize(class_name)\n Geode::FastInflector.tableize(class_name)\n end",
"def undecorated_table_name(class_name = class_name_of_active_record_descendant(self))\n table_name = Inflector.underscore(Inflector.demodulize(class_name))\n table_name = Inflector.pluralize(table_name) if pluralize_table_names\n return table_name\n end",
"def namespaced_table_name\n [self.class.table_namespace, table_name].reject {|s| s.nil? || s.empty?}.join('-')\n end",
"def table_name_for_insert\n self.class.to_s.downcase.pluralize\n end",
"def migration_class_name\n migration_name.camelize\n end",
"def readonly_class_name(db_name)\n \"#{db_name.camelize}\"\n end",
"def table_name=(value)\n @original_table_name = @table_name if defined?(@table_name)\n @table_name = value && value.to_s\n @quoted_table_name = nil\n end",
"def class_name(table_name = table_name) # :nodoc:\n # remove any prefix and/or suffix from the table name\n class_name = Inflector.camelize(table_name[table_name_prefix.length..-(table_name_suffix.length + 1)])\n class_name = Inflector.singularize(class_name) if pluralize_table_names\n return class_name\n end",
"def _table_name\n self.class.table_name\n end",
"def schema_name\n unless @schema_name\n @schema_name = collect_first(&:schema_name)\n end\n return @schema_name\n end",
"def class_name\n table.model.name\n end",
"def model_class_name\n implementation_model_name.ns_camelize\n end",
"def implicit_table_name\n pluralize(underscore(demodulize(name))).to_sym\n end",
"def implicit_table_name\n pluralize(underscore(demodulize(name))).to_sym\n end",
"def shared_schema_name\n Roomer.shared_schema_name ||= options[:shared_schema_name].to_s\n end",
"def table_name\n raise \"You must override `table_name' in your class\"\n end",
"def table_name\n model_class.table_name\n end",
"def table_name\n model_class.table_name\n end",
"def set_table(table)\n define_class_method(:table) {table.to_s.to_sym}\n end",
"def real_table_name\n packageid.gsub('.', '_')\n end",
"def table_name_prefix \n @table_name_prefix ||= 'gricer_'\n end",
"def identifier(schema, name = nil)\n name ||= @name\n \"`#{schema}`.`#{name}`\"\n end",
"def table_name\n self.name.split('::').last\n end",
"def table_name\n self.class.table_name\n end",
"def tableize(class_name)\n pluralize(underscore(class_name))\n end",
"def tableize(class_name)\n pluralize(underscore(class_name))\n end",
"def singular_table_name_from x\n x.to_s.tableize.singularize\n end",
"def new_table_name=(name)\n @new_table_name = \"layer_table_#{name}\"\n end",
"def table_name(class_name = nil)\n if class_name.nil?\n class_name = class_name_of_active_record_descendant(self)\n table_name_prefix + undecorated_table_name(class_name) + table_name_suffix\n else\n table_name_prefix + undecorated_table_name(class_name) + table_name_suffix\n end\n end",
"def model_class\n table_name.classify.constantize\n end",
"def roomer_reset_table_name\n if @roomer_original_table_name\n self.table_name = \"#{table_name_prefix}#{@roomer_original_table_name}\"\n else\n reset_table_name\n end\n roomer_ensure_table_name_prefix\n end",
"def table_name=(name)\n @table_name = name\n end",
"def model_name\n configuration.model_name\n end",
"def classify(table_name)\n # strip out any leading schema name\n camelize(singularize(table_name.to_s.sub(/.*\\./, '')))\n end",
"def tableize(class_name)\n pluralize(underscore(class_name))\n end",
"def set_schema(schema)\n @current_schema = nil\n execute \"SET SCHEMA #{schema}\", 'SCHEMA'\n end",
"def model_name(model)\n model.class.to_s.downcase\n end",
"def clean_table_name(table_name)\n class_from_table_name(table_name)&.table_name\n end",
"def default_database_name\n self.name.tableize\n end",
"def csharp_model_fqn( model )\n [ sanitize_model_namespace(model), sanitize_model_class(model) ].compact.join(\".\")\n end",
"def set_model_name(name)\n class_eval <<-RUBY\n def self.model_name\n @model_name ||= ActiveModel::Name.new(self, nil, #{name.inspect})\n end\n RUBY\n end",
"def model_class_name(name=nil)\n @model_class_name = name if name\n @model_class_name ||= guess_model_class_name\n end",
"def model_name\n model_name = options.fetch(:model_name) do\n ActiveSupport::Inflector.singularize(name.to_s)\n end\n\n Name.new(model_name.to_s)\n end",
"def quote_type_name(string, schema = nil)\n name_schema, table = string.to_s.scan(/[^\".\\s]+|\"[^\"]*\"/)\n if table.nil?\n table = name_schema\n name_schema = nil\n end\n\n schema = schema || name_schema || 'public'\n Name.new(schema, table).quoted\n end",
"def sanitize_model_namespace( model )\n [ models_namespace, model.name.split(\"::\")[0...-1] ].flatten.compact.join(\".\")\n end",
"def quote_table_or_view(name, options)\n schema = options[:schema]\n if schema\n \"\\\"#{schema}\\\".\\\"#{name}\\\"\"\n else\n \"\\\"#{name}\\\"\"\n end\n end",
"def get_table_name\n self.class.table_name\n end",
"def pgt_mangled_table_name(table)\n quote_schema_table(table).gsub('\"', '').gsub(/[^A-Za-z0-9]/, '_').gsub(/_+/, '_')\n end",
"def set_naming_convention\n repository(:default).adapter.\n resource_naming_convention = lambda do |value|\n underscore(class_basename(value))\n end\n end",
"def table_name\n self.class.table_name\n end",
"def schema_key\n @schema_key ||= self.class.name.demodulize.gsub('Validator', '').underscore\n end",
"def model_name\n self.class.name.underscore.split('_').first.singularize\n end",
"def full_name\n \"#{@table_alias}#{name}\"\n end",
"def classify(table_name)\n # strip out any leading schema name\n camelize(singularize(table_name.to_s.sub(/.*\\./, '')))\n end",
"def classify(table_name)\n # strip out any leading schema name\n camelize(singularize(table_name.to_s.sub(/.*\\./, '')))\n end",
"def schema_name\n @schema_name ||= owner_type.eql?('Root') ? owner.tenant.schema_name : schema_name_from_access_key_id\n end",
"def _write_class_code\n cname = @table.to_s.split('_').map(&:capitalize).join\n klass = Taupe::Model.const_set cname, Class.new(Taupe::Model::Table)\n klass._cname = cname\n klass._table = @table\n klass._columns = @_column_stack\n end",
"def define_class(config, table_name = nil)\n Class.new ActiveRecord::Base do\n establish_connection config\n self.abstract_class = true\n cattr_accessor :model_name\n \n if table_name\n self.table_name = table_name\n self.model_name = ActiveModel::Name.new(self, nil, table_name.classify)\n end \n \n end\n end",
"def namespace\n \"#{database.name}.#{name}\"\n end",
"def namespace\n \"#{database.name}.#{name}\"\n end",
"def sanitize_model_class( model )\n model.name.split(\"::\").last\n end",
"def tableize(class_name)\n words = class_name.to_const_path.tr('/', '_').split('_')\n words[-1] = pluralize(words[-1])\n words.join('_')\n end",
"def set_table_name(name) # rubocop:disable Naming/AccessorMethodName\n @table_name = name\n end",
"def model_class_name\n options[:model] ? options[:model].classify : \"User\"\n end",
"def table_name\n @table_name ||= begin\n schema = self.class.schema_name\n table = self.sqltablename.blank? ? \"PS_#{self.recname}\" : self.sqltablename\n [schema, table].compact.join(\".\")\n end\n end",
"def set_schema(klass, hash)\n if use_class_var?(klass)\n klass.class_variable_set(CLASS_VARIABLE, hash)\n else\n klass.instance_variable_set(INSTANCE_VARIABLE, hash)\n end\n end",
"def set_schema( schema )\n @schema = schema if @schema == self\n schema\n end",
"def set_schema(klass, hash)\n if use_class_var?(klass)\n klass.class_variable_set(CLASS_VARIABLE, hash)\n else\n klass.instance_variable_set(INSTANCE_VARIABLE, hash)\n end\n end",
"def default_join_table\n ([self[:class_name].demodulize, self[:model].name.to_s.demodulize]. \\\n map{|i| i.pluralize.underscore}.sort.join('_')).to_sym\n end",
"def class_name\n @class_name ||= active_record.name\n end",
"def set_table_name(value = nil, &block)\n define_attr_method :table_name_without_shard, value, &block\n end",
"def hoovered_table_name(model)\n table_name_from_model_version(model, 0)\n end",
"def model_name; end",
"def model_name; end",
"def model_name\n @_model_name ||= ActiveModel::Naming.param_key(model.new)\n end",
"def table_name\n @table_name ||= (superclass == SmallRecord::Base) ? default_table_name : superclass.table_name\n end",
"def quoted_table_name\n @_quoted_table_name ||= acts_as_list_class.quoted_table_name\n end",
"def models_namespace\n options[:model_namespace]\n end",
"def schema_and_table_name\n if qualified_table_name.include? '.'\n schema_name, table_name = qualified_table_name.split('.', 2)\n else\n table_name = qualified_table_name\n schema_name = self.class.default_schema_name\n end\n [schema_name, table_name]\n end",
"def table_name\n\t\tself.table_name\n\tend",
"def model_name\n root.singularize.camelize if root.present?\n end",
"def record_class_name=(value)\n doc['record_class_name'] = value.nil? ? nil : value.to_s\n end"
] |
[
"0.7063645",
"0.665204",
"0.6544634",
"0.65237355",
"0.63980806",
"0.63883513",
"0.6381655",
"0.6375005",
"0.6353489",
"0.6264394",
"0.62488264",
"0.62342995",
"0.6219615",
"0.61584246",
"0.60872304",
"0.60706234",
"0.60491985",
"0.60254496",
"0.6018481",
"0.60173607",
"0.60087085",
"0.5994582",
"0.5983509",
"0.59763414",
"0.5962579",
"0.59516597",
"0.5928619",
"0.5914654",
"0.5907648",
"0.590239",
"0.5902006",
"0.5902006",
"0.5891429",
"0.5885758",
"0.58855677",
"0.58855677",
"0.5874797",
"0.58584666",
"0.58530027",
"0.58409494",
"0.5823418",
"0.5811703",
"0.57994384",
"0.57994384",
"0.5798287",
"0.57889694",
"0.57875836",
"0.57738554",
"0.57717115",
"0.5754152",
"0.5737112",
"0.5726864",
"0.57156265",
"0.5711016",
"0.57086694",
"0.5702074",
"0.56988055",
"0.56964064",
"0.56940484",
"0.5690276",
"0.56891966",
"0.56792176",
"0.5669163",
"0.5652795",
"0.5650312",
"0.5646791",
"0.5645299",
"0.56261367",
"0.56251615",
"0.5621864",
"0.56188476",
"0.5601398",
"0.5601398",
"0.55985695",
"0.5595843",
"0.5586884",
"0.5584547",
"0.5584547",
"0.55752635",
"0.55744636",
"0.55542797",
"0.5547424",
"0.55463123",
"0.5543864",
"0.55437016",
"0.5543067",
"0.55415654",
"0.55393773",
"0.5531623",
"0.55270445",
"0.55231893",
"0.55231893",
"0.55202836",
"0.54936403",
"0.5488308",
"0.54870534",
"0.548593",
"0.54768586",
"0.5460715",
"0.5444062"
] |
0.72890466
|
0
|
A place to store managed connections
|
def managed_connections
@@managed_connections
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def connections\n @connections ||= {}\n end",
"def connections\n @_connections ||= {}\n end",
"def connections\n @connections ||= {}\n end",
"def connections\n @connections ||= []\n end",
"def grab_connections\n connections = []\n connections << @host\n end",
"def connections\n @connection_pool\n end",
"def remote_connections; end",
"def all_active_connections\n super\n end",
"def connections\n return @connectors.keys\n end",
"def setup_shared_connection_pool\n handler = ActiveRecord::Base.connection_handler\n\n handler.connection_pool_names.each do |name|\n pool_manager = handler.send(:connection_name_to_pool_manager)[name]\n pool_manager.shard_names.each do |shard_name|\n writing_pool_config = pool_manager.get_pool_config(ActiveRecord.writing_role, shard_name)\n @saved_pool_configs[name][shard_name] ||= {}\n pool_manager.role_names.each do |role|\n next unless pool_config = pool_manager.get_pool_config(role, shard_name)\n next if pool_config == writing_pool_config\n\n @saved_pool_configs[name][shard_name][role] = pool_config\n pool_manager.set_pool_config(role, shard_name, writing_pool_config)\n end\n end\n end\n end",
"def redis_connections\n @redis_connections ||= Stockpile::RedisConnectionsFactory.build_connections\n end",
"def connection_identifiers; end",
"def create_connection_manager(ds)\n end",
"def connections\n attributes[:connections]\n end",
"def get_connections\n @connected_to.keys\n end",
"def store\n connection.store\n end",
"def connections\n @mutex.synchronize do\n @connections.values.inject(0, &:+)\n end\n end",
"def establish_connections\n klass = Struct.new(:ip,:port)\n if t_servers = BackgrounDRb::BDRB_CONFIG[:client]\n connections = t_servers.split(',')\n connections.each do |conn_string|\n ip = conn_string.split(':')[0]\n port = conn_string.split(':')[1].to_i\n @bdrb_servers << klass.new(ip,port)\n end\n end\n @bdrb_servers << klass.new(BackgrounDRb::BDRB_CONFIG[:backgroundrb][:ip],BackgrounDRb::BDRB_CONFIG[:backgroundrb][:port].to_i)\n @bdrb_servers.each_with_index do |connection_info,index|\n next if @backend_connections.detect { |x| x.server_info == \"#{connection_info.ip}:#{connection_info.port}\" }\n @backend_connections << Connection.new(connection_info.ip,connection_info.port,self)\n end\n end",
"def new_connection; end",
"def nicks_connected_with\n @connection_cache.keys\n end",
"def checkin_connection(conn)\n @available_connections << conn\n conn\n end",
"def keyspaces\n @@connections.keys\n end",
"def keyspaces\n @@connections.keys\n end",
"def available_connections(server=:default)\n @available_connections[server]\n end",
"def all_connections\n hold do |c|\n sync do\n yield c\n @available_connections.each{|conn| yield conn}\n end\n end\n end",
"def all_connections\n hold do |c|\n sync do\n yield c\n @available_connections.each{|conn| yield conn}\n end\n end\n end",
"def connect\n @connection_pool.get_connection\n end",
"def connections\n @connections.size\n end",
"def connections\n @connections.size\n end",
"def redis_connection_pool\n @redis_connection_pool ||= Stockpile::RedisConnection.connection_pool\n end",
"def connection(options={})\n raise NoToken if @token.to_s.empty?\n\n @connections ||= {}\n\n cached_connection? && !protocol_changed? ? cached_connection : new_connection\n end",
"def connections\n connections = @mutex.synchronize { connections = @connections.dup }\n connections.delete_if { |c| c.kind_of?(Thread) }\n connections.freeze\n connections\n end",
"def initialize\n bootstrap_connections unless File.exist?(connections_file)\n load_connections\n end",
"def all\n repo.connections.each do |name, connection|\n write \"#{name} : #{connection}\"\n end\n end",
"def pools\n @pools ||= {}\n end",
"def pools\n @pools ||= {}\n end",
"def inject_conn_into_models\n BaseModel.connection = self\n end",
"def get_connected_uri\n @connections.map {|x| x.target_uri}\n end",
"def connection\n if conn = @reserved_connections[current_connection_id]\n conn\n else\n @reserved_connections[current_connection_id] = checkout\n end\n end",
"def conndefaults\n\t\treturn self.class.conndefaults\n\tend",
"def conndefaults\n\t\treturn self.class.conndefaults\n\tend",
"def conndefaults_hash\n\t\treturn self.class.conndefaults_hash\n\tend",
"def conndefaults_hash\n\t\treturn self.class.conndefaults_hash\n\tend",
"def bootstrap_connections\n FileUtils.touch(connections_file)\n end",
"def close_connections\n @connections ||= {}\n @connections.values.each do |connection|\n begin\n connection.disconnect!\n rescue\n end\n end\n\n @connections = {}\n end",
"def connection\n\n @hash_connection = {\n :connection => {\n :domain => @domain,\n :store_id => @profile,\n :service_user => @seller_profile,\n :service_password => @password\n }\n }\n\n end",
"def connection\n begin\n @connections.hold { |dbh| yield(dbh) }\n rescue Mysql::Error => me\n \n @configuration.log.fatal(me)\n \n @connections.available_connections.each do |sock|\n begin\n sock.close\n rescue => se\n @configuration.log.error(se)\n end\n end\n \n @connections.available_connections.clear\n raise me\n end\n end",
"def load_connections\n @connections = YAML::load_file(connections_file)\n if(@connections == false)\n @connections = {}\n end\n end",
"def get_connections\n @followers_buffer = []\n @friends_buffer = []\n\n batch_connections\n\n friends_data = { 'context' => TwitterContext , 'connections' => @friends_buffer }\n followers_data = { 'context' => TwitterContext , 'connections' => @followers_buffer }\n\n if Rails.env != 'production'\n File.open( 'tmp/twitter_friends.json' , 'w'){ |f| f.write( friends_data.to_json ) }\n File.open( 'tmp/twitter_followers.json' , 'w'){ |f| f.write( followers_data.to_json ) }\n File.open( 'tmp/twitter_credentials.json', 'w'){ |f| f.write( @credentials ) }\n end\n \n ConnectionHandler.build_user_user_connections( friends_data , @usertp)\n ConnectionHandler.build_user_user_connections( followers_data , @usertp)\n end",
"def connection_manager\n @manager ||= (connection_base? ? Redis::ConnectionManager.new(@connection_specification) : superclass.connection_manager)\n end",
"def redis_pool; end",
"def connection\n @reserved_connections[current_connection_id] ||= checkout\n end",
"def conn(server=:default)\n @conns[@servers[server]]\n end",
"def connection\n @db = Connection.client\nend",
"def post_init\n $connection_list << self\n @my_connection_index = $connection_list.length\n puts \"Connection Initialized #{@my_connection_index}) #{@my_address} - #{@signature}\"\n #\n # TODO: get access to the HostIP and HostPort associated with this connection\n # if it is the local control connection, do the following\n # * do not repeat other connection traffic to this connection\n # * consider any input a command to be processed\n end",
"def connect\n @connection_manager.connect\n end",
"def conn= x\n @conn_owned = false\n @conn = x\n end",
"def connection_class; end",
"def conn\n\t\treturn @conn ||= self.connect\n\tend",
"def mysql_global_vip_connection_info(category='root')\n host_info = {\n host: node[:bcpc][:management][:vip],\n port: 3306\n }\n mysql_local_connection_info.merge(host_info)\nend",
"def connection\n @connection ||= make_connection\n end",
"def connection\n @connection_pool\n end",
"def connection(key)\n @connections[key]\n end",
"def connection(key)\n @connections[key]\n end",
"def create_connection_manager\n ThreadSafeClientConnManager.new( @mparams, @scheme_registry )\n end",
"def all_connections\n yield @conn.first if @conn\n end",
"def save_ssh_connection(conn)\n conn.exec! 'cd ~'\n @lock.synchronize do\n @connection_pool << conn\n end\n rescue\n output \"Discarding nonfunctional SSH connection\"\n end",
"def connect\n @connection.create\n end",
"def setup\n return unless connection_pool_enabled?\n\n @connection_pool = ::Queue.new\n\n start_pool_size.times { @connection_pool.push(new_client) }\n\n @created_connections = start_pool_size\n end",
"def active_local_sockets; end",
"def connection; end",
"def connection; end",
"def connection; end",
"def connection; end",
"def connection; end",
"def connection; end",
"def connection; end",
"def connection; end",
"def empty_connection_pools!; end",
"def remote_connections\n @remote_connections || @mutex.synchronize { @remote_connections ||= RemoteConnections.new(self) }\n end",
"def connection_information\n super\n end",
"def connection_identifiers\n config.connection_class.call.identifiers\n end",
"def connection_pool_default_options\n o = super.dup\n # Default to only a single connection if a memory database is used,\n # because otherwise each connection will get a separate database\n o[:max_connections] = 1 if @opts[:database] == ':memory:' || blank_object?(@opts[:database])\n o\n end",
"def connection_identifier; end",
"def jmx_connection\n if @@jmx_connections[name].nil?\n url=JMXServiceURL.new uri\n map = java.util.HashMap.new\n creds = [username, password].to_java :String if !username.nil? #TODO: Encrypt pass?\n jmxc = JMXConnectorFactory.connect url, {\"jmx.remote.credentials\" => creds} \n @@jmx_connections[name] = jmxc.get_mbean_server_connection\n end\n @@jmx_connections[name]\n end",
"def connection\n if defined?(@active_connection_name) && (conn = active_connections[@active_connection_name])\n conn\n else\n # retrieve_connection sets the cache key.\n conn = retrieve_connection\n active_connections[@active_connection_name] = conn\n end\n end",
"def make_conn(conn)\n conn.h_conn = @dbm.make_conn(conn.spec)\n end",
"def connection\n @connection ||= build_connection\n end",
"def connection\n CryptKeeper::Connection.http_instance\n end",
"def conn()\n @@connections ||= {}\n\n # TODO remove hidden dependency on ENV['RACK_ENV'] (maybe require config= to accept a sub hash?)\n config = @@config[ENV[\"RACK_ENV\"]]\n raise MissingConfiguration, \"CassandraModel config is missing a '#{ENV['RACK_ENV']}' section\" unless config\n\n thrift_client_options = {:timeout => RightSupport::DB::CassandraModel::DEFAULT_TIMEOUT}\n thrift_client_options.merge!({:protocol => Thrift::BinaryProtocolAccelerated})\\\n if defined? Thrift::BinaryProtocolAccelerated\n\n @@connections[self.keyspace] ||= Cassandra.new(self.keyspace, config[\"server\"], thrift_client_options)\n @@connections[self.keyspace].disable_node_auto_discovery!\n @@connections[self.keyspace]\n end",
"def mysql_remote_connection_info(category='root', remote)\n host_info = {\n host: remote,\n port: 3306\n }\n mysql_local_connection_info.merge(host_info)\nend",
"def supported_connections\n ConnectionProvider.providers.select { |connection| connection.supported_by?(@handle) }\n end",
"def connection\n self.server[self.database][self.layout]\n end",
"def connection_for nick\n @connection_cache[nick]\n end",
"def all_connections\n hold do |conn|\n yield conn\n\n # Use a hash to record all connections already seen. As soon as we\n # come across a connection we've already seen, we stop the loop.\n conns = {}\n conns.compare_by_identity\n while true\n conn = nil\n begin\n break unless (conn = @queue.pop(timeout: 0)) && !conns[conn]\n conns[conn] = true\n yield conn\n ensure\n @queue.push(conn) if conn\n end\n end\n end\n end",
"def delete_all_active_connections\n super\n end",
"def connect(*) end",
"def db_conn\n @db_conn_pool ||= ConnectionPool.new(size: 4, timeout: 2) do\n Mysql2::Client.new(self.db_settings)\n end\n end",
"def connection_pool_maximum_reuse\n super\n end",
"def propagate_connections(klass)\n connections.each { |n, o| klass.connection(n, o) }\n end"
] |
[
"0.78580385",
"0.7808314",
"0.7747195",
"0.7455886",
"0.6746446",
"0.67325324",
"0.6642948",
"0.66416085",
"0.642973",
"0.6412615",
"0.63878965",
"0.633877",
"0.63363284",
"0.6313833",
"0.62448925",
"0.6227242",
"0.62070274",
"0.61570615",
"0.6132078",
"0.6107775",
"0.6078849",
"0.6075809",
"0.6075809",
"0.60491586",
"0.59862673",
"0.59862673",
"0.5944043",
"0.59122205",
"0.59122205",
"0.59015995",
"0.58982414",
"0.58765167",
"0.5870956",
"0.5866229",
"0.58512384",
"0.58512384",
"0.5837681",
"0.58375925",
"0.5831611",
"0.5801228",
"0.5801228",
"0.5797129",
"0.5797129",
"0.5794645",
"0.57919204",
"0.57795155",
"0.5776277",
"0.5766863",
"0.57607335",
"0.57594544",
"0.57292175",
"0.57258844",
"0.57195103",
"0.57161194",
"0.57011425",
"0.56867933",
"0.5672197",
"0.5671785",
"0.5670187",
"0.5656171",
"0.565397",
"0.564775",
"0.5639841",
"0.5639841",
"0.56378084",
"0.5629574",
"0.56289124",
"0.5625398",
"0.5621714",
"0.56110895",
"0.56090295",
"0.56090295",
"0.56090295",
"0.56090295",
"0.56090295",
"0.56090295",
"0.56090295",
"0.56090295",
"0.5598701",
"0.55971014",
"0.5591618",
"0.55850136",
"0.558138",
"0.5574464",
"0.5562846",
"0.5554191",
"0.55522496",
"0.55491143",
"0.5548387",
"0.55476075",
"0.5544058",
"0.5543945",
"0.55427706",
"0.5541346",
"0.55403525",
"0.5531101",
"0.5526811",
"0.5524216",
"0.55227524",
"0.5512465"
] |
0.77632385
|
2
|
Makes a request to the regional api for the URL to a specific server's auction database.
|
def refreshRealmAPI
begin
uri = "https://" + @regionURL + "/wow/auction/data/" + @realm + "?locale=#{@locale}" + "&apikey=#{@apikey}"
puts uri
jsontemp = Yajl::Parser.parse(open(uri)) # Parse JSON to ruby object.
@dataURL = jsontemp["files"][0]["url"]
@lastModified = Time.at(jsontemp["files"][0]["lastModified"]/1000)
puts "Successfully retrived data URL for #{uri}\nURL: #{@dataURL}\nLatest data is from #{@lastModified}"
@log.info "Successfully retrived data URL for #{uri}\nURL: #{@dataURL}\nLatest data is from #{@lastModified}"
return true
rescue => e
puts "Failed to get the Auction data URL."
@log.error "Failed to get the Auction data URL."
puts "Error message from the server:\n\n #{jsontemp}\n\n"
@log.error "Error message from the server:\n\n #{jsontemp}\n\n"
puts e
@log.error e
return false
end
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def response_db_master(method, database, master, uuid)\n \n #5.times { puts }\n #puts \"#{method} requested [database] /#{database}/#{master};#{uuid}\"\n \n if database == XDCR_BUCKET \n return [200]\n else\n return [404]\n end\nend",
"def edubase_client\n @uri = API_PATH + \"RunQuery/?\"\n end",
"def query_couchdb(urn, verb, host = '127.0.0.1', options = @@default_options)\n secure = options[:secure] || @@default_options[:secure]\n body = options[:body] || @@default_options[:body]\n\n ## Set scheme\n scheme = secure ? 'https' : 'http'\n ## build uri\n url = \"#{scheme}://#{host}\"\n uri = URI.join(url, urn)\n Chef::Log.debug(\"query_couchdb built uri: #{uri}\")\n\n http = Net::HTTP.new(uri.host, options[:port])\n http.use_ssl = secure\n unless options[:verify_ssl] || @@default_options[:verify_ssl]\n Chef::Log.debug('verify_ssl is false setting verify to none')\n http.verify_mode = OpenSSL::SSL::VERIFY_NONE\n end\n\n if body.empty?\n retry_request(http, verb.upcase, uri.request_uri)\n else\n if body.is_a? Hash\n headers = { 'Content-Type' => 'application/json' }\n body = JSON.generate(body)\n else\n body = \"\\\"#{body}\\\"\"\n end\n retry_request(http, verb.upcase, uri.request_uri, body, headers)\n end\n end",
"def doDB\n \t\tbegin\n \t\t# Query the inventory\n \t\tinv = InventoryService.getInv(@@config)\n \t\tyield(inv)\n \t\trescue Exception => ex\n \t\terror = \"Inventory - Error connecting to the Inventory Database - '#{ex}''\"\n \t\traise HTTPStatus::InternalServerError, error\n \t\tend\n \t\tend",
"def response_db_vbucket(method, database, vbucket_number, uuid)\n \n #5.times { puts }\n #puts \"#{method} requested [database] /#{database}/#{vbucket_number};#{uuid}\"\n \n if database == XDCR_BUCKET \n return [200]\n else\n return [404]\n end\nend",
"def get url\n RestClient::Request.execute(:method => :get, :url => url, :headers => lbaas_headers, :timeout => @timeout, :open_timeout => @open_timeout)\n end",
"def request_uri\n uri = URI.parse(Europeana::URL + \"/record\" + \"#{@id}.json\")\n uri.query = params_with_authentication.to_query\n uri\n end",
"def get_api_url\n\t\t\t\"https://#{club_name}.tidyhq.com/api/v1/\"\n\t\tend",
"def couchdb_get(urn, host = '127.0.0.1', options = @@default_options)\n query_couchdb(urn, 'GET', host, options)\n end",
"def database! url\n parsed = parse url\n cr = CouchRest.new(parsed[:host])\n cr.database!(parsed[:database])\n end",
"def request(params)\n\n # Add auth header\n headers = params[:headers] || {}\n headers['x-vcloud-authorization'] = @auth_key if !@auth_key.nil? || !@auth_key.equal?('')\n\n # set connection options\n options = {:url => params[:url],\n :body => params[:body] || '',\n :expects => params[:expects] || 200,\n :headers => headers || {},\n :method => params[:method] || 'GET'\n }\n\n # connect\n res = RestClient::Request.execute options\n\n raise res if (res.code!=params[:expects] && res.code!=200)\n\n res\n\n\n end",
"def query(base, address, param_hash) # :nodoc:\n http = Net::HTTP.new(@servers[@server_index], 443)\n http.use_ssl = true\n if @ca_file\n http.verify_mode = OpenSSL::SSL::VERIFY_PEER\n http.ca_file = @ca_file\n end\n if param_hash.empty?\n request_path = [base, escape_param(address)].join('/')\n else\n param_str = param_hash.map{|k,v| [k.to_s, escape_param(v)].join('=')}.join('&')\n request_path = [[base, escape_param(address)].join('/'), param_str].join('?')\n end\n http.start do \n request = Net::HTTP::Get.new(request_path)\n request.basic_auth @username, @password\n response = http.request(request)\n case response.code.to_i \n when 200\n return Leadspend::Result.new(response.body)\n when 202\n return Leadspend::Result.unknown(address)\n when 400\n raise Leadspend::Exceptions::BadRequestException, response.body\n when 401\n raise Leadspend::Exceptions::UnauthorizedRequestException, response.body\n when 500\n raise Leadspend::Exceptions::ServerException, response.body\n when 503\n raise Leadspend::Exceptions::ServerBusyException, response.body\n else\n raise Leadspend::Exceptions::UnknownResponseException, response.body\n end\n end\n end",
"def db_info\n @conn.query({url_path: \"#{database}\", method: :get})\n end",
"def init_api(aws_final_url)\n\tbegin\n\t\tconn = Faraday.new(aws_final_url, :ssl => {:verify => false})\n\t\tresponse = conn.get do |aws|\n\t\t\taws.headers['Content-Type'] = 'application/json' \n\t\tend\n\t\tif response.status == 200\n\t\t\tregions(aws_final_url)\n\t\t\toffercode(aws_final_url)\n\t\t\t@json_res = JSON.parse(response.body)\n\t\t\tsave_pricing\n\t\telse\n\t\t\traise \"Could not get the response\"\n\t\tend\n\trescue Exception => e\n\t\tputs e\n\tend\nend",
"def database_info(dbname, options = {})\n version = options.fetch(:t, '-')\n get db_url(dbname, version) + \"/\", :Accept => 'application/edn'\n end",
"def hit_api_direct\n # CollegiateLink API needs some data to be hashed and sent for auth purposes\n time = (Time.now.to_f * 1000).to_i.to_s\n ipaddress = ENV['cl_ipaddress']\n apikey = ENV['cl_apikey']\n privatekey = ENV['cl_privatekey']\n random = SecureRandom.hex\n hash = Digest::SHA256.hexdigest(apikey + ipaddress + time + random + privatekey)\n\n url = ENV['cl_apiurl'] + @resource + \"?time=\" + time + \"&apikey=\" + apikey + \"&random=\" + random + \"&hash=\" + hash + @url_options\n return send_request(url, nil)\n end",
"def make_api_call (api_path)\n uri = URI.parse(GPS_CLIENT_SETTINGS[\"url\"] + api_path)\n http = Net::HTTP.new(uri.host, uri.port)\n if uri.scheme==\"https\"\n http.use_ssl = true\n http.verify_mode = OpenSSL::SSL::VERIFY_NONE\n end\n request = Net::HTTP::Get.new(uri.request_uri)\n request['API-SECRET'] = client_secret\n response = http.request(request) \n make_adapter_response(response)\n end",
"def execute(params = nil)\n params ||= @params\n request.uri = \"#{@db_name}\"\n request.params = params\n Couchdbtools.execute(request)\n end",
"def make_api_request(query)\n response = timeout(configuration.timeout) do\n uri = URI.parse(query_url(query))\n http_client.start(uri.host, uri.port, :use_ssl => true) do |client|\n req = Net::HTTP::Post.new(uri.request_uri, configuration.http_headers)\n req.basic_auth(uri.user, uri.password) if uri.user and uri.password\n req.body = sensis_query_json(query)\n req.content_type = 'application/json'\n req['X-Auth-Token'] = configuration.api_key[0]\n req['X-Auth-Password'] = configuration.api_key[1]\n client.request(req)\n end\n end\n case response.code.to_i\n when 200\n return response\n when 400\n raise_error ::Geocoder::InvalidRequest, \"Bad Request: #{response.body}\"\n else\n raise_error ::Geocoder::Error, \"Unable to access Sensis API: #{response.code}. Body:\\n#{response.body}\"\n end\n response\n end",
"def request(base_url, params)\n RestClient.get base_url, { params: params }\n end",
"def api_url\n ENV.fetch('GLIMR_API_URL',\n 'https://glimr-api.taxtribunals.dsd.io/Live_API/api/tdsapi')\n end",
"def server\n uri = \"#{options[:use_ssl] ? \"https\" : \"http\"}://#{Blupee.config.api_server}\"\n end",
"def api_request(params)\n authenticate\n response = cap_client.get \"#{API_PATH}#{params}\"\n JSON.parse(response.body)\n rescue Faraday::TimeoutError => e\n NotificationManager.error(e, 'Timeout error during CAP-API request', self)\n raise\n rescue StandardError => e\n NotificationManager.error(e, \"#{e.class.name} during CAP-API request\", self)\n raise\n end",
"def base_uri\n if mode == 'sandbox'\n @base_uri = 'https://sandbox.fundamerica.com/api/'\n else\n @base_uri = 'https://apps.fundamerica.com/api/'\n end\n end",
"def request_url\n base_url = \"http://#{api_host}#{PATH}?\"\n base_url << \"appid=#{app_id}\"\n base_url << \"&callname=#{call_name}\"\n base_url << \"&siteid=#{site_id}\"\n base_url << \"&version=#{API_VERSION}\"\n base_url\n end",
"def set_api\n @url = 'https://api.coinmarketcap.com/v1/ticker/'\n @uri = URI(@url)\n @response = Net::HTTP.get(@uri)\n # converts response to a Ruby hash \n @lookup_crypto = JSON.parse(@response)\n @gain_loss = 0\n end",
"def hit_api_local\n # Authentication info\n pass = ENV['stugov_api_user']\n priv = ENV['stugov_api_pass']\n # Our base URL hosted on stugov's server\n base_url = ENV['stugov_api_base_url']\n\n # We make a sha256 hash of this in binary format, then base64 encode that\n digest = Base64.encode64(OpenSSL::HMAC.digest(OpenSSL::Digest.new(\"sha256\"), priv, pass)).chomp\n\n url = base_url + \"?resource=\" + @resource + @url_options\n return send_request(url, digest)\n end",
"def request_over_http(options, endpoint, uri)\n uri = URI.parse([endpoint, uri].join('/') + '?' + options.collect{|k,v| \"%s=%s\" % [k,v]}.join('&'))\n\n http = Net::HTTP.new(uri.host, 443)\n http.use_ssl = true\n if TIES::DEBUG\n http.set_debug_output $stderr\n end\n request = Net::HTTP::Get.new(uri.request_uri)\n request.initialize_http_header({\n 'Accept' => 'application/json',\n 'Content-Type' => 'application/json',\n 'User-Agent' => 'TIEScloud ruby %s' % TIES::VERSION,\n 'ties-date' => (time = Time.now).utc.to_s(:rfc822),\n 'Authorization' => authentication('GET', time, uri.request_uri),\n 'DistrictNumber' => self.district_number.to_s\n })\n http.request(request)\n end",
"def connection\n @db = Connection.client\nend",
"def show\n @database = Database.find(params[:id])\n @database.refresh_database_state\n \n # If the rds was restored from a snapshot, a few fields have to be modify after becomes available: security_groups, password, size\n @database_client = @database.sync_agi_fields_to_rds unless @database.snapshot_id.blank?\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @database }\n end\n end",
"def connection(uri:, params:)\n Faraday.new(\n url: \"#{@endpoint}/#{@api_version}/#{uri}\",\n params: params,\n headers: {\n 'User-Agent' => \"RubyGem-zombie_battleground-api_#{VERSION}\",\n 'Content-Type' => 'application/json'\n }\n )\n end",
"def api_url(endpoint)\n \"https://www.bloc.io/api/v1/#{endpoint}\"\n end",
"def all_dbs\n @conn.query({url_path: \"_all_dbs\", method: :get})\n end",
"def create_db(database)\n @conn.query({url_path: database, method: :put})\n end",
"def query\n begin\n response = resource[\"/query/#{app}\"].post(:days => options[:days], :url => options[:url], :mode => options[:mode])\n rescue RestClient::InternalServerError\n display \"An error has occurred.\"\n end\n display response.to_s\n end",
"def api_url\n \"https://api.readmill.com/v2/\"\n end",
"def initialize\n @url = 'http://api.preachingcentral.com/'\n end",
"def drugbank_get(route, params)\n url = $drugbank_api + route\n res = HTTParty.get(url, :query => params, :headers => $drugbank_headers)\n return res\nend",
"def endpoint\n \"https://#{region.sub(/-\\d$/, '')}.power-iaas.cloud.ibm.com/pcloud/v1\"\n end",
"def call_api\n url = \"#{@base_uri}/stock\"\n params = {symbol: @symbol, api_token: @api_key}\n\n if @exchange\n params[:stock_exchange] = @exchange\n end\n\n Lita.logger.debug \"call_api: #{url} #{params.inspect}\"\n\n @response = RestClient.get url, {params: params}\n\n Lita.logger.debug \"response: #{@response}\"\n end",
"def server(api_method=nil)\n server = BigbluebuttonRails.configuration.select_server.call(self, api_method)\n end",
"def request (url_requested, api_key)\n url_builded = url_requested + \"&api_key=\" + api_key\n\n url = URI(url_builded)\n\n https = Net::HTTP.new(url.host, url.port)\n https.use_ssl = true\n\n request = Net::HTTP::Get.new(url)\n\n response = https.request(request)\n\n JSON.parse(response.read_body)\nend",
"def get_database(name)\n response = @glue_client.get_database(name: name)\n response.database\nrescue Aws::Glue::Errors::GlueException => e\n @logger.error(\"Glue could not get database #{name}: \\n#{e.message}\")\n raise\n end",
"def request(url, api_key)\n rover = get_data(\"#{url}#{api_key}\")\n\n return rover\nend",
"def api_request_url(object)\n api_uri = 'http://timezonedb.wellfounded.ca/api/v1'\n api_request_url = \"#{api_uri}/timezones?\"\n api_request_url << get_filters(object).join('&')\n log \"Making request to #{api_request_url} for timezone.\"\n api_request_url\n end",
"def simple_request(method, url)\n uri = URI.parse(url)\n request = Net::HTTP::Get.new(uri)\n request.content_type = \"application/json\"\n request[\"Accept\"] = \"application/json\"\n request[\"App-Id\"] = ENV[\"salt_edge_app_id\"]\n request[\"Secret\"] = ENV[\"salt_edge_secret\"]\n\n req_options = {\n use_ssl: uri.scheme == \"https\",\n } \n\n response = Net::HTTP.start(uri.hostname, uri.port, req_options) do |http|\n http.request(request)\n end\n\n #puts response.body\n return JSON.parse(response.body)\n end",
"def execute\n puts DB_MSG\n logger.info(DB_MSG)\n # connect to the database and query on a Site\n CaTissue::Database.current.open { find_in_transit_site }\n end",
"def call_api\n @client.build_url\n @client.get\n assign_data\n end",
"def iod_request (api, params)\r\n uri = URI(\"http://api.idolondemand.com/1/api/async/#{api}/v1\")\r\n uri.query = URI.encode_www_form(params)\r\n res = Net::HTTP.get_response(uri, p_addr = $proxy_host, p_port = $proxy_port)\r\n jobid = JSON.parse(res.body)['jobID']\r\n puts \"Post request jobid [#{jobid}]\"\r\n return job_results(jobid)\r\nend",
"def http_request(host)\n begin\n response = Nokogiri::HTML(open(\"http:\\/\\/bleed-1161785939.us-east-1.elb.amazonaws.com\\/bleed\\/#{host}\"))\n display_result(response, host)\n rescue\n puts \"[-] #{host}: Issues connecting to site\"\n end\nend",
"def database! url\n parsed = parse url\n cr = Sova.new(parsed[:host])\n cr.database!(parsed[:database])\n end",
"def fetch_server(uri)\n @server[uri]\n end",
"def make_request(method, url, params = {})\n path = \"#{EbanqApi.base_url}/#{url}\"\n response = case method\n when :get then get(path, headers, params)\n when :post then post(path, headers, params)\n when :put then put(path, headers, params)\n when :delete then delete(path, headers)\n else raise 'Error'\n end\n process_raw_response(response)\n rescue RestClient::ResourceNotFound, SocketError, Errno::ECONNREFUSED => e\n raise e\n end",
"def find(catalog_id)\n setup_request \"#{@@resource_url}/#{catalog_id}\"\n end",
"def index\n @init_url = \"https://crest-tq.eveonline.com/market/\" +\n \"10000002\" +\n \"/orders/\"+ \"sell\" + \"/?type=https://crest-tq.eveonline.com/types/\" +\n \"34\" +\n \"/\"\n end",
"def get(database_id:)\n path = '/databases/{databaseId}'\n .gsub('{databaseId}', database_id)\n\n if database_id.nil?\n raise Appwrite::Exception.new('Missing required parameter: \"databaseId\"')\n end\n\n params = {\n }\n \n headers = {\n \"content-type\": 'application/json',\n }\n\n @client.call(\n method: 'GET',\n path: path,\n headers: headers,\n params: params,\n response_type: Models::Database\n )\n end",
"def call_with_config_params\n disciplina = Iesde::Model::Disciplina.buscar(\n config: {\n user: ENV['IESDE_USER'],\n password: ENV['IESDE_PWD'],\n ead_api_key: ENV['EAD_API_KEY']\n }).select {|d| d.disciplina_id.to_i == 123 }.first\n\n ap disciplina.aulas(12345)\nend",
"def base_url\n \"https://api.beezup.com/\"\n end",
"def perform_api_request(query) \n request = Net::HTTP::Post.new( \n $url.path + ($url.query != nil ? (\"?\" + $url.query) : \"\"), \n initheader = {\"Content-Type\" => \"application/json\", \n \"X-Replicon-Security-Context\" => \"User\"}) \n request.basic_auth($companyKey + \"\\\\\" + $loginname, $password) \n request.body = JSON.generate(query) \n \n server = Net::HTTP.new($url.host, $url.port) \n server.use_ssl = $url.scheme == \"https\" \n response = server.start {|http| http.request(request)} \n \n http_code = response.code.to_i \n if http_code != 200 \n puts response.body \n raise \"Expected success code 200, but was #{http_code}\" \n end \n \n return JSON.parse(response.body) \nend",
"def query\n start do |connection|\n request = HTTP::Get.new \"#{ @uri.path }?#{ @uri.query }\"\n @headers.each { |header, value| request[header] = value }\n connection.request request\n end\n end",
"def get_api endpoint\n\turi = URI.parse(\"http://#{@HOST}:#{@HTTP}#{endpoint}\")\n\trequest = Net::HTTP::Get.new(uri)\n\trequest[\"Accept\"] = \"application/json\"\n\toptions = { use_ssl: uri.scheme == \"https\" }\n\tresponse = Net::HTTP.start(uri.hostname, uri.port, options) do |http|\n\t\thttp.request(request)\n\tend\nend",
"def query(options = {})\n options.each_pair do |key, value|\n self.send(\"#{key}=\", value)\n end\n \n raise 'No API base URL provided.' unless @@base_url\n raise 'No API key provided.' unless @@api_key\n raise 'No resource specified.' unless @resource\n raise 'No format specified.' unless @format\n \n return do_get\n end",
"def make_request_get_response_trend_availible\n @path_trend_availible = '/1.1/trends/available.json'\n @address_trend_availible = URI(\"#{@baseurl}#{@path_trend_availible}\")\n # Set up HTTP. Need ssL to make the connection\n @request_trend_availible = Net::HTTP::Get.new @address_trend_availible.request_uri\n @http = Net::HTTP.new @address_trend_availible.host, @address_trend_availible.port\n @http.use_ssl = true\n @http.verify_mode = OpenSSL::SSL::VERIFY_PEER\n # Issue the request.\n @request_trend_availible.oauth! @http, @consumer_key_country, @access_token_country\n @http.start\n @response_trend_availible = @http.request @request_trend_availible\n @response_trend_availible\n end",
"def book_server(user)\n query = {\n key: SSC_API_KEY,\n user: user,\n hours: MAX_MATCH_LENGTH,\n }\n result = Net::HTTP.post_form(URI.parse(SSC_ENDPOINT + \"/bookings/?#{query.to_query}\"), {})\n\n json = JSON.parse result.body\n if result.code != '200'\n puts \"Error: Failed booking for #{user}, #{json['statusMessage']}\"\n return nil\n end\n json['server']['connect-string']\nend",
"def request(*args, &blk)\n (@client ||= connect).request(*args, &blk)\n end",
"def item(opts={})\r\n opts[:output] = 'json'\r\n opts[:callback] = 'callback'\r\n Yahoo::Request.get(\"http://auctions.yahooapis.jp/AuctionWebService/V2/auctionItem\", Yahoo::Api.merge(opts))\r\n end",
"def db\n return @client if @client\n @client = Mysql2::Client.new(\n host: 'localhost',\n port: nil,\n username: 'root',\n password: nil,\n database: 'isucon5q',\n reconnect: true,\n )\n @client.query_options.merge!(symbolize_keys: true)\n @client\nend",
"def do_get\n Net::HTTP.get(URI.parse(api_url))\n end",
"def database(options = {})\n throw ArgumentError.new('Required arguments :database_id missing') if options[:database_id].nil?\n get(\"databases/#{options[:database_id]}\")\n end",
"def send\n uri = URI(@api_url)\n http = Net::HTTP.new(uri.host, uri.port)\n if uri.scheme == \"https\"\n http.use_ssl = true\n http.verify_mode = OpenSSL::SSL::VERIFY_PEER\n end\n http.read_timeout = GlobalConstant::CompanyApi.read_timeout\n http.open_timeout = GlobalConstant::CompanyApi.open_timeout\n req_obj = get_request_obj(uri.request_uri)\n\n http_response, e = nil, nil\n begin\n http_response = http.request(req_obj)\n set_api_response_cookie(http_response)\n parse_api_response(http_response)\n rescue Net::ReadTimeout, Net::OpenTimeout => e\n # Timeouts\n exception_with_internal_code(\n e,\n 'company_api_timeout',\n 'company api timeout',\n GlobalConstant::ErrorCode.internal_server_error,\n debug_data\n )\n rescue Exception => e\n # Exceptions\n exception_with_internal_code(e, 'company_api_exception', 'company api exception', GlobalConstant::ErrorCode.internal_server_error, debug_data)\n end\n\n end",
"def getToolsSyndicateLocaldatabase( entity_id, destructive)\n params = Hash.new\n params['entity_id'] = entity_id\n params['destructive'] = destructive\n return doCurl(\"get\",\"/tools/syndicate/localdatabase\",params)\n end",
"def construct_url\n @@API_URL + city + @@API_KEY\n end",
"def get_report(api_key, client_api_id, interval, query = \"\")\n uri = URI(API_ENDPOINT) + URI.escape(\"?api_key=#{api_key}&client_api_id=#{client_api_id}&interval=#{interval}&query=#{query}\")\n http = Net::HTTP.new(uri.host, uri.port)\n http.use_ssl = true\n request = Net::HTTP::Get.new(uri.request_uri)\n response = http.request(request)\n raise \"Server returned error #{response.code} processing your API request\" if response.code != \"200\"\n JSON.parse(response.body)\nend",
"def getApiEndpoint(endpoint)\n \n apiRegion = $drugbank_region\n\n if $drugbank_region != \"\"\n apiRegion = apiRegion + \"/\"\n end\n\n return apiRegion + endpoint\n\nend",
"def query(query, apikey, endpoint, accept = \"application/json\")\n\tjson = open(\"#{endpoint}?query=#{CGI.escape(query)}&apikey=#{apikey}\", \"Accept\" => accept).read\n\tJSON.parse(json)\n\nend",
"def get\n appid = ENV['TRIMET_APP_ID']\n response = Unirest.get( \"http://developer.trimet.org/ws/v2/vehicles?appid=#{appid}\" )\n response.body\nend",
"def base_uri\n\t\t\t\"#{server}/api/v4\"\n\t\tend",
"def make_api_request(request_url)\n url = URI.parse(request_url)\n request = Net::HTTP::Get.new(url.path)\n connection = Net::HTTP.new(url.host, url.port)\n connection.use_ssl = true\n connection.request(request)\n end",
"def seed_db\n\n # Grab login page\n @agent.get(@@server + @@standings) do |login_page|\n \n # Login and fetch the standings page\n standings_page = do_login(login_page)\n \n # Get a link to a random bracket (the first one)\n bracket_link = \"\"\n standings_page.links.each do |link|\n next unless /(#{@@server})\\/(brackets)\\/\\d+\\/\\d+/.match(link.href)\n bracket_link = link.href\n break\n end\n\n # Pull in the regions, teams, and games from the bracket page and store them in the DB\n scrape_data_and_create_records(bracket_link)\n\n end \n\n end",
"def index\n @all_products = Product.all.as_json\n @all_currencies = ExchangeRateService.valid_currencies\n @base_url = Rails.env.development? ? 'http://localhost:3000' : 'http://cabifycheckout.com'\n end",
"def api_url\n \"#{protocol}://api:#{api_key}@#{host}/#{api_version}\"\n end",
"def connection\n Faraday.new(api_url) do |conn|\n conn.use :breakers\n conn.response :snakecase\n conn.response :json, content_type: /\\bjson$/\n conn.adapter Faraday.default_adapter\n end\n end",
"def connection\n Faraday.new(api_url) do |conn|\n conn.use :breakers\n conn.response :snakecase\n conn.response :json, content_type: /\\bjson$/\n conn.adapter Faraday.default_adapter\n end\n end",
"def connect_to_server\n if server_url.present?\n @client = FHIR::Client.new(server_url)\n @client.use_r4\n @client.additional_headers = { 'Accept-Encoding' => 'identity' } # \n @client.set_basic_auth(\"fhiruser\",\"change-password\")\n cookies[:server_url] = server_url\n session[:server_url] = server_url \n end\n rescue => exception\n err = \"Connection failed: Ensure provided url points to a valid FHIR server\"\n redirect_to root_path, flash: { error: err }\n \n end",
"def link_to_database \n \"/connect/#{url_uuid}\"\n end",
"def request\n data = [\n exact_queries(%w[name user type]),\n switch_queries(%w[stable reviewed]),\n array_queries(%w[tags]),\n range_condition_queries(%w[downloads rating version]),\n sort_query,\n range_query\n ].reduce({}, :merge)\n\n url = \"/items/#{data.empty? ? '' : '?'}#{URI.encode_www_form(data)}\"\n @data = @api.request(url).map do |hash|\n hash['id'] = @api.normalize_id(hash['id'])\n hash\n end\n end",
"def use_database(subdomain)\n ActiveRecord::Base.establish_connection(website_connection(subdomain))\n end",
"def bc_api_url\n ENV['BC_API_ENDPOINT'] || 'https://api.bigcommerce.com'\nend",
"def elections(key = api_key)\n uri = \"#{base_path}/elections\"\n handle_timeouts do\n self.class.get(uri, query: { key: key })\n end\n end",
"def connection\n options = { url: api_url, ssl: { verify: false } }\n\n connection = Faraday.new(options) do |conn|\n conn.response :readmill_errors\n conn.response :mashify\n conn.response :json\n\n conn.adapter adapter\n end\n\n connection\n end",
"def database_sync_api\n snacks = Suggestion.all\n\n snacks.each do |snack|\n if !api_snack_checker?(snack.name)\n data_to_api(snack.name, snack.location, snack.optional)\n end\n end\n end",
"def api_url\n \"http://svcs.ebay.com/services/search/FindingService/v1?OPERATION-NAME=findItemsAdvanced\" +\n \"&SERVICE-VERSION=1.9.0\" +\n \"&SECURITY-APPNAME=#{Settings.ebay.app_id}\" +\n \"&RESPONSE-DATA-FORMAT=JSON\" +\n \"&REST-PAYLOAD\" +\n \"&keywords=#{query}\"\n end",
"def find_by_name(catalog_name)\n @request = setup_request \"#{@@resource_url}s\"\n @request.query = { catalogName: catalog_name }\n @request\n end",
"def set_auction\n @auction = Auction.find(params[:id])\n end",
"def set_auction\n @auction = Auction.find(params[:id])\n end",
"def set_auction\n @auction = Auction.find(params[:id])\n end",
"def set_auction\n @auction = Auction.find(params[:id])\n end",
"def set_auction\n @auction = Auction.find(params[:id])\n end",
"def perform_record_request\n Faraday.send(:get, request_url, nil, headers)\n end",
"def connection()\r\n\r\n begin\r\n db_params = {\r\n host: ENV['host'], # AWS link\r\n port: ENV['port'], # AWS port, always 5432\r\n username: ENV['username'],\r\n password: ENV['password'],\r\n database: ENV['database']\r\n }\r\n\r\n client = Mysql2::Client.new(db_params)\r\n\r\n rescue Mysql2::Error => e\r\n puts 'Exception occurred'\r\n puts e.message\r\n end\r\n\r\nend",
"def test_start_auction_to_one_server_should_replicate\n reset(REP_SERVERS)\n server1, server2, server3 = REP_SERVERS\n name = \"gau2\"\n \n do_rest(server1, :start_auction, {name: name, end_time: Time.now.to_i + 3})\n do_rest(server1, :bid, {name: name, client:1, bid:100})\n res = do_rest(server1, :status, {name: name})\n assert_equal(\"1\", res.strip)\n\n res = do_rest(server2, :status, {name: name})\n assert_equal(\"1\", res.strip)\n res = do_rest(server3, :status, {name: name})\n assert_equal(\"1\", res.strip)\n end"
] |
[
"0.5609157",
"0.5608949",
"0.54337823",
"0.5408552",
"0.5359365",
"0.5301148",
"0.52393365",
"0.52278036",
"0.5223498",
"0.5199291",
"0.517414",
"0.51336545",
"0.5120033",
"0.51006985",
"0.5059749",
"0.5048924",
"0.5047501",
"0.5018101",
"0.49888328",
"0.4981478",
"0.4977666",
"0.49497685",
"0.49464294",
"0.49362522",
"0.49320072",
"0.49289283",
"0.4928638",
"0.49281275",
"0.49174035",
"0.49168926",
"0.49109262",
"0.49094266",
"0.4903515",
"0.48967364",
"0.48917165",
"0.4888609",
"0.4885231",
"0.48825428",
"0.48820567",
"0.48810875",
"0.48806214",
"0.48749316",
"0.48668686",
"0.48506403",
"0.48480397",
"0.4844991",
"0.48449373",
"0.4842204",
"0.48417193",
"0.48397607",
"0.48389357",
"0.4832822",
"0.48291877",
"0.48273194",
"0.4823354",
"0.48173025",
"0.48170838",
"0.48141566",
"0.4807606",
"0.48012558",
"0.47955796",
"0.47939166",
"0.4792121",
"0.47894263",
"0.47855142",
"0.4785408",
"0.4780467",
"0.477988",
"0.4779211",
"0.47783336",
"0.47768912",
"0.47662652",
"0.4765411",
"0.4762112",
"0.47530928",
"0.4750687",
"0.47467682",
"0.4744976",
"0.47425744",
"0.47343242",
"0.47301638",
"0.47296143",
"0.47296143",
"0.47244197",
"0.47219703",
"0.47216606",
"0.4720258",
"0.47158408",
"0.4715799",
"0.47157842",
"0.4714983",
"0.47126666",
"0.47099754",
"0.4709923",
"0.4709923",
"0.4709923",
"0.4709923",
"0.4709923",
"0.47089896",
"0.47089788",
"0.47055975"
] |
0.0
|
-1
|
Downloads the actual auction database file from a specific server. The fileformat is JSON.
|
def getAuctionJSON
begin
json = open(@dataURL).read
if !json.include? "ownerRealm"
raise "Recieved something unexpected: \n #{json} \n of class: #{json.class}"
end
return json
rescue => e
puts "Failed to download the Auction JSON data.\n #{e}"
@log.error "Failed to download the Auction JSON data.\n #{e}"
return false
end
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def download_file url\n check_session\n result = api_client.execute(:uri => url)\n if result.status == 200\n return result.body\n else\n puts \"An error occurred: #{result.data['error']['message']}\"\n return nil\n end\n end",
"def download_internal_best_bets\n\tunless ENV['FUNNELBACK_PASSWORD']\n\t\tabort('The FUNNELBACK_PASSWORD enviornment variable is not set')\n\tend\n\tbegin\n\thttp = Net::HTTP.new(ENV['FUNNELBACK_DOMAIN'],ENV['FUNNELBACK_PORT'])\n\treq = Net::HTTP::Get.new(\"/search/admin/download-conf.cgi?collection=website&f=best_bets.cfg&dir=profile-folder-internal-preview\")\n\thttp.use_ssl = true\n\thttp.verify_mode = OpenSSL::SSL::VERIFY_NONE\n\treq.basic_auth ENV['FUNNELBACK_USERNAME'], ENV['FUNNELBACK_PASSWORD']\n\tresponse = http.request(req)\nrescue\n \terb :loaderror\n else\n\t\n\topen(\"data/best-bets-internal.txt\", \"wb\") { |file|\n\t\tfile.write(response.body)\n }\nend\nend",
"def download_external_best_bets\n\tunless ENV['FUNNELBACK_PASSWORD']\n\t\tabort('The FUNNELBACK_PASSWORD enviornment variable is not set')\n\tend\n\tbegin\n\t\thttp = Net::HTTP.new(ENV['FUNNELBACK_DOMAIN'],ENV['FUNNELBACK_PORT'])\n\t\treq = Net::HTTP::Get.new(\"/search/admin/download-conf.cgi?collection=website&f=best_bets.cfg&dir=profile-folder-_default_preview\")\n\t\thttp.use_ssl = true\n\t\thttp.verify_mode = OpenSSL::SSL::VERIFY_NONE\n\t\treq.basic_auth ENV['FUNNELBACK_USERNAME'], ENV['FUNNELBACK_PASSWORD']\n\t\tresponse = http.request(req)\n\trescue\n \terb :loaderror\n else\n\t \topen(\"data/best-bets-external.txt\", \"wb\") { |file|\n \tfile.write(response.body)\n \t}\n end\nend",
"def download_prices() \n\t\t \t \t \t\n\t\t\t\t@dropbox_token = DropboxSession.deserialize(Linkdropbox.first.dropbox_token)\n\t\t\t\tclient = DropboxClient.new(@dropbox_token)\n\t\t\t\tcontents, metadata = client.get_file_and_metadata('Grupo8/DBPrecios.accdb')\n\t\t\t\tbegin\n\t\t\t\t open('public/jars/DBPrecios.accdb', 'wb') {|f| f.puts contents }\n\t\t\t\t \n\t\t\t\t \n\t\t\t\trescue\n\t\t\t\t flash[:success] = \"Exception occured while downloading...\"\t\t\n\n\t \t\tend \n\n\t \t\tLinkdropbox.import_prices_to_csv\t\n\tend",
"def download(county)\n puts \" F95BA #{county}\"\n results = @client.find :all, {\n class: '1', # 1 Residential\n query: \"(246=|A),(61=|#{county})\", #246 ListingStatus\n #A ActiveAvailable\n #61 County\n select: '157,881,10,922,924,137,261,129,246,80,61,25,1424,102,214,314,96,1,131,1329,sysid', \n search_type: 'Property'\n }\n puts \"F95BA #{results.size} listings\"\n #puts \"F95BA saving\"\n pg_save(results)\n results\n end",
"def download_driver_http\n File.write(\n @target_path,\n open(@params[:jdbc_driver_jar]).read\n )\n end",
"def load_bitbucket()\n JSON.parse(IO.read('db-1.0.json'))\nend",
"def meeting_recordings_download_file(download_url)\n raise \"You must use JWT client\" unless self.class == Zoom::Clients::JWT\n file=Tempfile.create\n file.binmode\n response = HTTParty.get(\"#{download_url}?access_token=#{access_token}\",\n stream_body: true,\n follow_redirects: true\n ) do |fragment|\n if fragment.code == 200\n file.write(fragment)\n elsif fragment.code != 302\n raise StandardError, \"Non-success status code while streaming #{fragment.code}\"\n end\n end\n file\n end",
"def download\n api_url = build_url\n RestClient::Request.execute(method: 'GET', url: api_url, open_timeout: 20)\n end",
"def download(path)\n RestClient::Request.execute(:method => :get, :url => (@host + '/api/v1/files' + escape(path)), :user => @user, :password => @password, :raw_response => true, :headers => {:accept_encoding => ''}).file\n end",
"def get_file(options, path, filename)\n\n result = {}\n\n http = Net::HTTP.new(ENV['NESSUS_HOST'], options[:port])\n http.use_ssl = @use_ssl\n http.verify_mode = OpenSSL::SSL::VERIFY_NONE\n\n http.start do |http|\n req = Net::HTTP::Get.new(path)\n\n req['X-ApiKeys'] = \"accessKey=#{ENV['NESSUS_ACCESS_KEY']}; secretKey=#{ENV['NESSUS_SECRET_KEY']}\"\n \n resp, data = http.request(req)\n \n open(\"#{options[:directory]}/#{filename}\", \"wb\") do |file|\n file.write(resp.body)\n end\n \n if resp.code.eql? '200'\n puts \" downloaded #{filename} to #{options[:directory]}\"\n elsif resp.code.eql? '409'\n # the server isn't ready yet, it is\n # still exporting the file\n sleep(1)\n print '.'\n get_file(options, path, filename)\n else\n puts \"Error: \" + resp.code.to_s + \": \" + resp.body\n end\n end\n\n return result\n end",
"def getDBData(urlInput)\n url = URI.parse(urlInput)\n server = Couch::Server.new(url.host, url.port)\n res = server.GET(urlInput)\n puts \"PARSING JSON FILE--------------\"\n json = JSON.parse(res.body)\n return json\n end",
"def opendb\n db = Rho::Database.new(Rho::Application.databaseFilePath('local'), 'local')\n render :json => db.to_s\n return db.to_s\n end",
"def download\n #require 'debugger'; debugger\n generator = Generator.where(id: params[:id], user_id: current_user.id).first\n send_file TerrainLib::Component.generate JSON.parse(generator.generator_hash)\n end",
"def download_backup(host,port,user,password,name,path,format)\n connect_to_host(host,port,user,password,true)\n if format == 'binary'\n local_file = path+name+\".backup\"\n remote_file = name+\".backup\"\n elsif format== 'script'\n local_file = path+name+\".rsc\"\n remote_file = name+\".rsc\"\n end\n download_file(remote_file,local_file)\n @ssh_connect.close(@ssh_connect)\n end",
"def database_file(id)\n if repo\n if id == :not_specified || !id || id == ''\n f = 'store.json'\n else\n f = \"store_#{id.to_s.downcase}.json\"\n end\n \"#{git_database_dir}/#{f}\"\n end\n end",
"def download(server, id)\n make_request(\n :get,\n \"/download/#{server}/#{id}\",\n headers: { 'Accept' => '*/*' },\n base: @media_uri\n )\n end",
"def download_schema!(uri)\n file_name = \"#{TMP_DIR}/#{uri.split('/').last}\"\n File.write(file_name, HTTParty.get(uri).body)\n file_name\n end",
"def show\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @inventory_file }\n format.download {\n send_file @inventory_file.inventory.download,\n filename: @inventory_file.inventory_filename,\n type: 'application/octet-stream'\n }\n end\n end",
"def database_info(dbname, options = {})\n version = options.fetch(:t, '-')\n get db_url(dbname, version) + \"/\", :Accept => 'application/edn'\n end",
"def get_db\n manif = Project.miga_online_manif\n @downloadable = manif[:databases].map do |name, i|\n local = Project.find_by(path: name)\n local_v = (local.miga.metadata[:release] || 'unknown') if local && local.miga\n latest = i[:versions][i[:latest].to_sym]\n file = File.join(Settings.miga_projects, latest[:path])\n downloading = File.size(file) * 100 / latest[:size] if File.exist?(file)\n i.merge(name: name, local: local_v, downloading: downloading)\n end\n end",
"def download_link\n download_params = { :sub => 'download', :fileid => @fileid, :filename => @remote_filename, :cookie => @api.cookie }\n DOWNLOAD_URL % [ @server_id, @short_host, download_params.to_query ]\n end",
"def download\n create_agent\n login\n fetch_feed\n create_catalog\n download_catalog\n end",
"def get_response url, format\r\n begin\r\n response = Net::HTTP.get_response(URI.parse(url))\r\n if format.to_sym == :json\r\n res = JSON.parse response.body\r\n else\r\n res = response.body\r\n end\r\n rescue Exception => e\r\n res = \"ERROR: There is a problem while fetching data, please check whether OpenTSDB is running or not.\"\r\n end\r\n res\r\n end",
"def download\n URI.extract(json, ['http', 'https']).each do |url|\n get_asset(url)\n end\n\n json\n end",
"def download_data\n # Custom downloader functionality implementation, this is just a simplified example template\n $log.info 'Starting downloading data from the Dummy API.'\n entities = @metadata.list_entities\n results = []\n entities.each do |entity|\n entity_custom = entity.custom\n load_metadata entity, entity_custom['fields']\n start_date, end_date = get_date_interval(entity)\n results << [entity, download_data_range(entity, start_date, end_date)]\n end\n save_to_s3 results\n end",
"def download(path)\n downloader.get do |req|\n req.url path\n end.body\n end",
"def download\n respond_to do |format|\n if @package.present?\n format.html do\n send_file Munki::Application::PACKAGE_DIR + @package.installer_item_location, :filename => @package.to_s(:download_filename)\n fresh_when :etag => @package, :last_modified => @package.created_at.utc, :public => true\n end\n \n format.json { render :text => @package.to_json(:methods => [:name, :display_name]) }\n else\n render page_not_found\n end\n end\n end",
"def download\n respond_to do |format|\n if @package.present?\n format.html do\n send_file Munki::Application::PACKAGE_DIR + @package.installer_item_location, :filename => @package.to_s(:download_filename)\n fresh_when :etag => @package, :last_modified => @package.created_at.utc, :public => true\n end\n \n format.json { render :text => @package.to_json(:methods => [:name, :display_name]) }\n else\n render page_not_found\n end\n end\n end",
"def file_data\n @client.get_file @file_url\n end",
"def download(limit=100, download_to=File.expand_path(\".\")+File::SEPARATOR+\"edgar_data\")\n items=@content[\"channel\"][0][\"item\"]\n items.each_with_index do |item, index|\n break if index==limit\n files=get_xbrl_files(item)\n download_to += File::SEPARATOR unless download_to.end_with?(File::SEPARATOR)\n data_dir=download_to\n data_dir=data_dir+File::SEPARATOR+item[\"xbrlFiling\"][0][\"cikNumber\"][0][\"content\"]\n data_dir=data_dir+File::SEPARATOR+item[\"xbrlFiling\"][0][\"accessionNumber\"][0][\"content\"]\n mkdir(data_dir)\n files.each do |file|\n file_content=open(file[\"edgar:url\"]).read\n dump_to_file(data_dir+File::SEPARATOR+file[\"edgar:file\"], file_content)\n end\n end\n end",
"def download\n response = Clickmeetings.with_client(client_options) do\n Clickmeetings.client.connect.get remote_url(__method__, id: id) do |req|\n req.headers.merge! default_headers\n end\n end\n response.body\n end",
"def download\n open(download_url, \"rb\")\n end",
"def b2_download_file_by_name(file, *folder)\n\n if folder[0] != nil\n file_url = b2_generate_file_url(file, folder[0])\n else\n file_url = b2_generate_file_url(file)\n end\n\n uri = URI(file_url)\n req = Net::HTTP::Get.new(uri)\n http = Net::HTTP.new(req.uri.host, req.uri.port)\n http.use_ssl = true\n res = http.start {|http| http.request(req)}\n\n case res\n when Net::HTTPSuccess then\n res.body\n swapfile = File.new(\"./public/swap/#{file}\", 'wb')\n swapfile.puts(res.body)\n swapfile.close\n when Net::HTTPRedirection then\n fetch(res['location'], limit - 1)\n else\n res.error!\n end\n\nend",
"def fetch_movie_data_and_cache_in_json_files\n query = Imdb::Top250.new\n movies = query.movies.map do |m|\n HTTParty.get(\"http://www.omdbapi.com/?i=tt#{m.id}&plot=short&r=json\")\n end\n movies = fetch_movie_data_from_list\n File.open('db/data/movie_data.json', 'w') { |file| file.write(movies.to_json) }\nend",
"def download\n file = BruseFile.find_by(:download_hash => params[:download_hash])\n if file.identity.user == current_user\n # send the file to the user\n send_data file.identity.get_file(file.foreign_ref), filename: file.name, type: file.filetype\n end\n end",
"def saveJSON\n File.open(\"./db/larves.json\", 'w') do |file|\n file.write(@url)\n end\n puts \"Ecriture effectuer avec success!\"\n end",
"def download_file(source, filename)\n response = HTTParty.get(\"#{@host}/download_file\", query: {\n source_file: \"#{source}:#{filename}\",\n api_key: @api_key\n })\n \n return response.body\n end",
"def download_loot(p)\n\t\t# These are the files we wanna grab for the directory for future decryption\n\t\tfiles = ['signons.sqlite', 'key3.db', 'cert8.db']\n\n\t\tfiles.each do |item|\n\t\t\tloot = ''\n\n\t\t\t# Downaload the file\n\t\t\tif session.type =~ /meterpreter/\n\t\t\t\tvprint_status(\"Downloading: #{p + item}\")\n\t\t\t\tbegin\n\t\t\t\t\tf = session.fs.file.new(p + item, 'rb')\n\t\t\t\t\tuntil f.eof?\n\t\t\t\t\t\tloot << f.read\n\t\t\t\t\tend\n\t\t\t\trescue ::Exception => e\n\t\t\t\tensure\n\t\t\t\t\tf.close\n\t\t\t\tend\n\t\t\telsif session.type =~ /shell/\n\t\t\t\tcmd_show = (session.platform =~ /win/) ? 'type' : 'cat'\n\t\t\t\t# The type command will add a 0x0a character in the file? Pff.\n\t\t\t\t# Gotta lstrip that.\n\t\t\t\tloot = cmd_exec(cmd_show, \"\\\"#{p+item}\\\"\").lstrip\n\t\t\t\tnext if loot =~ /system cannot find the file specified|No such file/\n\t\t\tend\n\n\t\t\t# Save it\n\t\t\text = ::File.extname(item)\n\t\t\text = ext[1,ext.length]\n\n\t\t\tpath = store_loot(\n\t\t\t\t\"tb.#{item}\",\n\t\t\t\t\"binary/#{ext}\",\n\t\t\t\tsession,\n\t\t\t\tloot,\n\t\t\t\t\"thunderbird_raw_#{item}\",\n\t\t\t\t\"Thunderbird Raw File #{item}\")\n\n\t\t\tprint_status(\"#{item} saved in #{path}\")\n\n\t\t\t# Parse signons.sqlite\n\t\t\tif item =~ /signons\\.sqlite/ and datastore['PARSE']\n\t\t\t\tprint_status(\"Parsing signons.sqlite...\")\n\t\t\t\tdata_tbl = parse(path)\n\t\t\t\tif data_tbl.nil? or data_tbl.rows.empty?\n\t\t\t\t\tprint_status(\"No data parsed\")\n\t\t\t\telse\n\t\t\t\t\tpath = store_loot(\n\t\t\t\t\t\t\"tb.parsed.#{item}\",\n\t\t\t\t\t\t\"text/plain\",\n\t\t\t\t\t\tsession,\n\t\t\t\t\t\tdata_tbl.to_csv,\n\t\t\t\t\t\t\"thunderbird_parsed_#{item}\",\n\t\t\t\t\t\t\"Thunderbird Parsed File #{item}\")\n\t\t\t\t\tprint_status(\"Parsed signons.sqlite saved in: #{path}\")\n\t\t\t\tend\n\t\t\tend\n\t\tend\n\tend",
"def url\n @client.get_download_link(@path)\n end",
"def download(url)\n filedl = url.split(\"/\").last\n uri = URI.parse(url)\n http = Net::HTTP.new(uri.host, uri.port)\n request = Net::HTTP::Get.new(uri.request_uri)\n response = http.request(request)\n open(filedl, \"w\") do |file|\n file.write(response.body)\n end\n return filedl\nend",
"def download\n file = UserFile.exist_refresh_state(@context, params[:uid])\n\n if file.state != UserFile::STATE_CLOSED\n raise ApiError, \"Files can only be downloaded if they are in the 'closed' state\"\n end\n\n if file.license.present? && !file.licensed_by?(@context)\n raise ApiError, \"You must accept the license before you can download this\"\n end\n\n file_url = file.file_url(@context, params[:inline])\n\n respond_to do |format|\n format.html do\n redirect_to URI.parse(file_url).to_s\n end\n\n format.json do\n render json: {\n file_url: file_url,\n file_size: file.file_size,\n }, adapter: :json\n end\n end\n end",
"def download_article_data\n print \"Getting list of articles...\\n\"\n ArticleGetter.new([ARTICLE_SOURCE]).result\n end",
"def download_file\n source_file = Item.new(Path.new(params[:source_file]))\n response = {}\n response[:source_file] = source_file\n\n if !source_file.path.exist?\n response[:msg] = \"Fail\"\n render json: response, status: 404\n return\n end\n\n respond_to do |format|\n format.json { render json: response }\n format.file { send_file source_file.path.to_path }\n end\n \n end",
"def downloadTBD\n\n #just do a collection!\n\n @players = Player.all\n @players.each do |player|\n @playerHash = downloadStat(player.yahoo_pid)\n end \n\n end",
"def download\n @orders = Order.find(params[:ids])\n\n case params[:format]\n when 'xml'\n content = Order.get_xml_for_orders(@orders)\n when 'csv'\n content = Order.get_csv_for_orders(@orders)\n end\n\n directory = File.join(RAILS_ROOT, \"public/system/order_files\")\n file_name = Time.now.strftime(\"%m_%d_%Y_%H-%M\")\n file = \"#{file_name}.#{params[:format]}\"\n save_to = \"#{directory}/#{file}\"\n\n # make sure we have the directory to write these files to\n if Dir[directory].empty?\n FileUtils.mkdir_p(directory)\n end \n\n # write the file\n File.open(save_to, \"w\") { |f| f.write(content) }\n\n send_file(save_to, :type => \"text/#{params[:format]}\")\n end",
"def get_file(url, options = {})\n\t\t\t\n\t\t\t# better way of doing this?\n\t\t\t# Map custom keys to the HTTP request values\n\t\t\t# TODO add handles for searching based upon stats\n\t\t\treqs = {\n\t\t\t\t:character_name => 'n',\n\t\t\t\t:source => \"fl[source]\", # dungeon, badges, arena, etc\n\t\t\t\t:dungeon => \"fl[dungeon]\", # seems it needs the dungeons id rather than name\n\t\t\t\t:difficulty => \"fl[difficulty]\", # normal, heroic, etc\n\t\t\t\t:item_type => \"fl[type]\", # weapon, armor, trinket, etc\n\t\t\t\t:item_slot => \"fl[slot]\", # head, shoulders, etc\n\t\t\t\t:item_sub_type => \"fl[subTp]\", # leather, mail, etc\n\t\t\t\t:realm => 'r',\n\t\t\t\t:search => 'searchQuery',\n\t\t\t\t:type => 'searchType',\n\t\t\t\t:guild_name => 'gn',\n\t\t\t\t:item_id => 'i',\n\t\t\t\t:team_size => 'ts',\n\t\t\t\t:team_name => 't',\n\t\t\t\t:group => 'group',\n\t\t\t\t:callback => 'callback',\n\t\t\t\t:calendar_type => 'type',\n\t\t\t\t:month => 'month',\n\t\t\t\t:year => 'year',\n\t\t\t\t:event => 'e',\n :now => 'now',\n\t\t\t\t:achievement_category => 'c'\n\t\t\t} \n\t\t\t\n\t\t\tparams = []\n\t\t\toptions.each do |key, value|\n\t\t\t\tparams << \"#{reqs[key]}=#{u(value)}\" if reqs[key]\n\t\t\tend\n\t\t\t\n\t\t\tquery = ''\n\t\t\tquery = query + '?' + params.join('&') if params.size > 0\n\t\t\t#query = '?' + params.join('&') if params.size > 0\n\t\t\t\n\t\t\tbase = self.base_url(options[:locale], options)\n\t\t\tfull_query = base + url + query\n\n\t\t\tif options[:caching]\n\t\t\t\tresponse = get_cache(full_query, options)\n\t\t\telse\n\t\t\t\tresponse = http_request(full_query, options)\n\t\t\tend\n\t\tend",
"def server_get_file(server, path)\n request(\n :path => \"containers/#{server.id}/files\",\n :params => {\n :path => path\n },\n :disable_body_extraction => true\n ).get(:body)\n end",
"def _fetch\n curl @url, '-o', @tarball_path\n end",
"def download(id, filename)\n open(filename, \"w\").write(@res[\"/download?id=#{id}\"].get)\n return filename\n rescue\n puts $!\n return nil\n end",
"def download\n @config.each do |config|\n download_backup(config[:host],config[:port],config[:user],config[:password],config[:name],config[:path],config[:format])\n end\n return true\n end",
"def download_prepared_dump id\n name = \"trunk-#{id}.dump\"\n target_path = File.expand_path(\"../../fixtures/#{name}\", __FILE__)\n \n puts \"Accessing prepared DB test snapshot #{id} from S3.\"\n \n require 's3'\n service = S3::Service.new(:access_key_id => access_key_id, :secret_access_key => secret_access_key)\n bucket = service.buckets.find(\"cocoapods-org-testing-dumps\")\n \n # Due to a bug in the s3 gem we are searching for the object via iterating.\n bucket.objects.each do |obj|\n if obj.key == name\n puts \"Downloading prepared DB test snapshot #{id} from S3.\"\n File.open(target_path, 'w') do |file|\n file.write(obj.content)\n end\n break\n end\n end\n \n puts \"Prepared DB test snapshot #{id} downloaded to #{target_path}\"\n end",
"def download(save_path=\"\")\n url = prefix + \"download\"\n r = response(url) \n if r.class == String #success\n open(File.join(save_path,@filename), \"wb\").write(open(r).read)\n return r\n else #failed\n return r\n end\n end",
"def download(url)\n response = client.send(:get, url, {}, 'Content-Type': 'text/html')\n raise ::Errors::BadRequest unless response.success?\n\n response.body\n end",
"def save2blob()\n url = URI.parse(getUrl)\n h = Net::HTTP.start(url.host, url.port) do |http|\n resp, data = http.get(url.path + '?' + url.query)\n return data\n end\n end",
"def fetch_file(file_path)\n client.get_file(file_path)\n end",
"def download_iso\n IsoRepo.get(iso_url)\n end",
"def get_all()\n return JSON.parse(File.read(@database_file))\n \n # error handling\n rescue Errno::ENOENT\n File.open(@database_file, 'w+')\n File.write(@database_file, [])\n retry\n end",
"def download(local, item)\n if item[:bundled] then\n warning \"Not downloading into bundled item #{synckey(item)}\"\n return\n end\n local.dirname.mkpath\n id = item[@itemkey.to_sym]\n if id.nil? then\n debug \"!!! Missing '#{@itemkey}', using :id instead!\"\n debug \":id => #{item[:id]}\"\n id = item[:id]\n raise \"Both #{@itemkey} and id in item are nil!\" if id.nil?\n end\n local.open('wb') do |io|\n fetch(id) do |chunk|\n io.write chunk\n end\n end\n end",
"def download!(file)\n login\n warn \"DEBUG: downloading #{file}\" if debug\n if dry_run\n warn \"DEBUG: download skipped for dry run\" if dry_run\n filename = file\n body = \"no body\"\n else\n page = agent.get(file)\n filename = page.filename\n body = page.body\n end\n [ filename, body ]\n end",
"def AxeDownload(download)\n uri = URI('https://axeweb.intel.com/axe/api/testlist/295/latest/combined')\n puts uri\n req = Net::HTTP::Get.new(uri)\n req.basic_auth 'autoclient', 'gr@ph1c$'\n \n if download\n print \"#{Time.now.strftime(\"%l:%M:%S %p\")} - Start download\\n\"\n res = Net::HTTP.start(uri.hostname, uri.port, :use_ssl => true, :verify_mode => OpenSSL::SSL::VERIFY_NONE) {|http|\n http.request(req)\n }\n #puts res.body\n \n unless res.kind_of? Net::HTTPSuccess\n puts \"Error downloading results from Axe\"\n exit(9)\n end\n open('result.json', 'wb') do |fileAXE|\n fileAXE << res.body\n end\n print \"#{Time.now.strftime(\"%l:%M:%S %p\")} - End download\\n\"\n end\nend",
"def download_package\n path = download_path\n remote_file path do\n source URL\n action :create\n only_if { !::File.exist?(PATH) }\n end\n end",
"def call\n @response = connection.get(url)\n if status == 200\n context.file = save!\n else\n context.fail! message: \"Download failed\"\n end\n end",
"def download_b2_file(image_info, *sightings_count)\n\n folder = image_info[0]\n filename = image_info[1]\n b2_file_path = \"#{folder}/#{filename}\"\n\n if sightings_count\n sightings_count = sightings_count[0]\n subdirectory = folder.split(\"/\")[1]\n sightings_dir = create_directory(subdirectory, sightings_count)\n swap_file = \"#{sightings_dir}/#{filename}\"\n else\n create_folder()\n swap_file = \"./public/swap/#{filename}\" # use when running via app.rb\n # swap_file = \"../public/swap/#{file}\" # use when running locally from /lib/b2_bucket.rb\n end\n\n file_url = b2_generate_file_url(filename, folder)\n\n uri = URI(file_url)\n req = Net::HTTP::Get.new(uri)\n http = Net::HTTP.new(req.uri.host, req.uri.port)\n http.use_ssl = true\n res = http.start {|http| http.request(req)}\n\n case res\n when Net::HTTPSuccess then\n res.body\n swapfile = File.new(swap_file, 'wb')\n swapfile.puts(res.body)\n swapfile.close\n when Net::HTTPRedirection then\n fetch(res['location'], limit - 1)\n else\n res.error!\n end\n\nend",
"def download(url, download_to=File.expand_path(\".\")+File::SEPARATOR)\n $LOG.info \" Starting download of fillings from SEC url [\"+url+\"]\"\n files=[]\n content = open(url).read\n @links = Set.new\n uri=URI(url)\n @base_path=\"\"\n @base_path=(uri.scheme+\"://\"+uri.host+((uri.port==80 && \"\") || \":\"+uri.port.to_s)) unless uri.host.nil?\n parse(content)\n download_to += File::SEPARATOR unless download_to.end_with?(File::SEPARATOR)\n mkdir(download_to)\n @links.each do |link|\n file=download_to + link.split(\"/\")[-1]\n dump_to_file(file, open(link).read)\n files << file\n end unless uri.host.nil?\n files\n end",
"def database! url\n parsed = parse url\n cr = CouchRest.new(parsed[:host])\n cr.database!(parsed[:database])\n end",
"def download(save_path=\"\")\n url = prefix + \"download\"\n r = response(url)\n if r.class == String #success\n open(::File.join(save_path,@filename), \"wb\").write(open(r).read)\n return r\n else #failed\n return r\n end\n end",
"def dump_table(client, fusion_tables, fusion_table_id, backup_directory)\n backup_directory ||= \"backups\"\n FileUtils.mkdir_p backup_directory\n\n fusion_table = client.execute(\n :api_method => fusion_tables.table.get,\n :parameters => {'tableId' => \"#{fusion_table_id}\"}\n )\n fusion_table.data.to_hash\n filename = File.join(backup_directory ,\"#{fusion_table.data.to_hash['name']}-#{fusion_table_id}\")\n $stderr.puts filename\n\n File.open(\"#{filename}.json\",\"w\") do |f|\n f.write(JSON.pretty_generate(fusion_table.data.to_hash))\n end\n\n result = client.execute(\n :api_method => fusion_tables.query.sql_get,\n :parameters => {'sql' => \"SELECT * FROM #{fusion_table_id}\"}\n )\n fusion_table_data = result.data.to_hash\n\n if fusion_table_data['error']\n if fusion_table_data['error']['errors'][0]['reason'] == 'responseSizeTooLarge'\n # use Fusion Tables V2 media downloads API\n result = client.execute(\n :api_method => fusion_tables.query.sql_get,\n :parameters => {'sql' => \"SELECT * FROM #{fusion_table_id}\", 'alt' => 'media'}\n )\n File.open(\"#{filename}.csv\", 'w') { |file| file.write(result.response.body) }\n else\n $stderr.puts \"Unhandled Error:\"\n $stderr.puts fusion_table_data.inspect\n end\n else\n CSV.open(\"#{filename}.csv\", 'w') do |csv|\n if fusion_table_data['rows'] && (fusion_table_data['rows'].length > 0)\n csv << fusion_table_data['columns']\n fusion_table_data['rows'].each do |row|\n csv << row\n end\n end\n end\n end\nend",
"def fetch_attachment(document_id, attachment_name)\n server.get(\"#{name}/#{CGI.escape(document_id)}/#{CGI.escape(attachment_name)}\", :no_json => true)\n end",
"def download\n @package = Package.find(params[:id])\n if @package.present?\n send_file Munki::Application::PACKAGE_DIR + @package.installer_item_location, :filename => @package.to_s(:download_filename)\n fresh_when :etag => @package, :last_modified => @package.created_at.utc, :public => true\n else\n render page_not_found\n end\n end",
"def refreshRealmAPI\n\t\tbegin\n\t\t\turi = \"https://\" + @regionURL + \"/wow/auction/data/\" + @realm + \"?locale=#{@locale}\" + \"&apikey=#{@apikey}\"\n\t\t\tputs uri\n\t\t\tjsontemp = Yajl::Parser.parse(open(uri)) # Parse JSON to ruby object.\n\n\t\t\t@dataURL = jsontemp[\"files\"][0][\"url\"]\n\t\t\t@lastModified = Time.at(jsontemp[\"files\"][0][\"lastModified\"]/1000)\n\n\t\t\tputs \"Successfully retrived data URL for #{uri}\\nURL: #{@dataURL}\\nLatest data is from #{@lastModified}\"\n\t\t\t@log.info \"Successfully retrived data URL for #{uri}\\nURL: #{@dataURL}\\nLatest data is from #{@lastModified}\"\n\n\t\t\treturn true\n\n\t\trescue => e\n\t\t\t\n\t\t\tputs \"Failed to get the Auction data URL.\"\n\t\t\t@log.error \"Failed to get the Auction data URL.\"\n\t\t\tputs \"Error message from the server:\\n\\n #{jsontemp}\\n\\n\"\n\t\t\t@log.error \"Error message from the server:\\n\\n #{jsontemp}\\n\\n\"\n\t\t\tputs e\n\t\t\t@log.error e\n\n\t\t\treturn false\n\n\t\tend\n\t\t\n\t\t\n\tend",
"def download(url, filename)\n uri = URI.parse(url)\n f = open(filename,'wb')\n begin\n http = Net::HTTP.start(uri.host) {|http|\n http.request_get(uri.path) {|resp|\n resp.read_body {|segment|\n f.write(segment)\n }\n }\n }\n ensure\n f.close()\n end\nend",
"def run!\n db_uri = URI.parse(@configuration.remote_db)\n @adapter = (Fixtural.adapter_for_uri db_uri).new(self, db_uri)\n # Actually connect to the database and figure out which tables we need\n # to download\n @adapter.connect()\n\n # Figure out the tables we need to download\n tables = compute_table_list()\n\n total = tables.length\n puts \"Downloading #{total.to_s} tables:\"\n\n # Setup the output store and the writer for the chosen output format\n output_store = @configuration.output_store\n output_writer_klass = Fixtural.output_writer_for_name @configuration.output_format\n\n tables.each_with_index do |table, index|\n progressbar = ProgressBar.create(\n format: \"- #{table} (#{(index+1).to_s}/#{total.to_s}) (%j%%)\"\n )\n extension = output_writer_klass.extension\n name = \"#{table}.#{extension}\"\n\n columns = get_columns table\n\n output_store.open(name) do |fd|\n writer = output_writer_klass.new fd, table, columns\n\n download_table table, writer, progressbar\n\n writer.done\n end\n end\n end",
"def download\n @data = HeyDan::Helper.get_data_from_url(HeyDan.cdn + '/' + dataset_file_name)\n end",
"def get_data_download(project_id, data_download_id)\n get \"projects/#{project_id}/datadownload/#{data_download_id}\"\n end",
"def query_filedata(uid)\n json = {}\n id = uid.to_i\n dir = id%100\n dir = \"#{fs_root}/#{dir.to_s}/#{id}_lastread\"\n FileUtils.makedirs(dir)\n fname = \"#{dir}/jsondata\" \n #p \"query_filedata:filename #{fname}\"\n\n begin\n if FileTest::exists?(fname) \n data= nil \n open(fname, \"r\") {|f|\n data = f.read\n # f.seek(0)\n # f.write(\"\") \n # f.truncate(0)\n }\n # p \"data=#{data.inspect}\"\n json = JSON.parse(data) if data\n end\n rescue Exception=>e\n # logger.error e\n p e.inspect\n pe(e)\n \n end\n\n return json\n\n end",
"def download\n\n# Taking from the query parameter that was passed in from the HTML after'?' (Query string parameters):\n url = params[:url]\n\n# Get the filename unless last character is a /\n filename = url[-1,1] == '/' ? \"\" : url.split('/').last\n\n# Set the url without the filename so that we can encode any special characters of the filename later\n url.slice!(filename)\n\n# Querying the database for documents where the URL matches the url ^ and convert to array:\n document = Cache.where(url: url+CGI::escape(filename)).to_a\n\n# Send binary data of content to be downloaded as HTML document:\n send_data(document.first.content, :filename => filename, :type => \"text/html; charset=utf-8\", :disposition => 'attachment')\n end",
"def download(api_params)\n File.open(File.basename(api_params[:download_location]), 'wb') do |file|\n file.write(api_params[\"data\"])\n end\n end",
"def download(path, item)\n Download.download item.link, path\n Logger.instance.info \"Downloaded file #{item.title}\"\n @history.save item.title\n end",
"def get_file(file_id)\n\tputs \"Getting file: \" + file_id\n\tresponse = request_get('/api/partner/file/' + file_id)\n\tputs response.body\nend",
"def download(path, item)\n Download.download item.link, path\n Logger.instance.info \"Downloaded file #{item.title}\"\n @history.save item\n end",
"def download_url(is_serverUrl=true)\n DocumentHelper.get_download_url(@file_name, is_serverUrl)\n end",
"def download\n \n @tmpfile = fetch_remote(location)\n \n @fetched_at = Time.new\n \n return @tmpfile\n \n end",
"def get_data(class_name)\n return JSON.parse(File.read(@database_file))[class_name]\n \n # error handling\n rescue Errno::ENOENT\n File.open(@database_file, 'w+')\n File.write(@database_file, [])\n retry\n end",
"def fetch_all\n downloaded = 0\n megabyte = 1024 * 1024\n mb_down = 0\n File.open(@file, 'wb+') do |file|\n @downloader.parts do |part|\n begin\n @server.request_get(part, @headers) do |res|\n res.read_body do |body|\n file.write body\n end\n end # /@server\n rescue Timeout::Error, EOFError, Errno::ECONNRESET => exception\n yield -1\n @server = Net::HTTP.start(@base_url.host, @base_url.port)\n STDERR.puts \"Connection error...\"\n retry\n end\n\n yield part\n end # /parts\n end # /File\n end",
"def download(url, filename)\n uri = URI.parse(url)\n f = open(filename,'wb')\n begin\n http = Net::HTTP.start(uri.host) {|http|\n http.request_get(uri.request_uri) {|resp|\n resp.read_body {|segment|\n f.write(segment)\n }\n }\n }\n ensure\n f.close()\n end\nend",
"def download(url)\n base.get(url, @file_path)\n end",
"def show\n @download = Download.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @download }\n end\n end",
"def download(url)\n STDERR.print \"loading #{url}\"\n response = Faraday.get(url)\n STDERR.puts ' (done)'\n response.body\n end",
"def cve_2019_9960_pre25(cookie, ip)\n vprint_status('Attempting to retrieve file')\n traversal = '../' * datastore['DEPTH']\n res = send_request_cgi({\n 'method' => 'GET',\n 'uri' => normalize_uri(uri, 'index.php'),\n 'cookie' => cookie,\n 'vars_get' => {\n 'sZip' => \"#{traversal}#{datastore['FILE']}\",\n 'r' => 'admin/export/sa/downloadZip'\n }\n })\n if res && res.code == 200 && !res.body.empty?\n loot = store_loot('', 'text/plain', ip, res.body, datastore['FILE'], 'LimeSurvey Path Traversal')\n print_good(\"File stored to: #{loot}\")\n else\n print_bad('File not found or server not vulnerable')\n end\n end",
"def download_install_file(action, args={})\n company = @company\n username = @user\n password = @password\n url = \"https://#{company}.logicmonitor.com/santaba/do/#{action}?\"\n args.each_pair do |key, value|\n url << \"#{key}=#{value}&\"\n end\n url << \"c=#{company}&u=#{username}&p=#{password}\"\n uri = URI(url)\n begin\n http = Net::HTTP.new(uri.host, 443)\n http.use_ssl = true\n http.verify_mode = OpenSSL::SSL::VERIFY_NONE\n req = Net::HTTP::Get.new(uri.request_uri)\n response = http.request(req)\n return response.body\n rescue SocketError => se\n puts \"There was an issue communicating with #{url}. Please make sure everything is correct and try again.\"\n puts se.message\n rescue Error => e\n puts \"There was an issue.\"\n puts e.message\n end\n return nil\nend",
"def file\n TestIds.database_file(id)\n end",
"def http_test_3\n File.open('http_test_3.mp4', 'wb') do |f|\n f.write RestClient.get(URL)\n end\nend",
"def download(url, file)\n # First, open the remote file. (FD == 'File Descriptor', a C term)\n #\n # If we first open the remote file, if there are any exceptions in\n # attempting to open it, we won't lose the contents of the local file.\n open(url) do |remoteFD|\n # Open the local file, purging the contents.\n File.open(file, \"w\") do |genericFD|\n remoteFD.each_line do |line|\n # Take each line, stick it in the file.\n genericFD.puts line\n end\n end\n end\n @repo.add(file) # add the file to the list to be committed\nend",
"def fetch\n open(to_url) { |io| io.read }\n end",
"def download_file!\n retries = 3\n\n begin\n options = {\n :read_timeout => 300,\n }\n\n open(from_url, options) do |f|\n save_to_cache(f)\n end\n rescue SocketError,\n Errno::ECONNREFUSED,\n Errno::ECONNRESET,\n Errno::ENETUNREACH,\n Timeout::Error,\n OpenURI::HTTPError => e\n if retries != 0\n retries -= 1\n retry\n else\n raise Exceptions::NetworkError.new(from_url, e)\n end\n end\n end",
"def download_game(game)\n page = agent.get(game.fetch(:page_url))\n download_url = find_download_url(page)\n\n unless download_url\n puts \"Cannot download #{ game.fetch(:name) }\"\n return\n end\n\n Helpers.wait\n throw :interrupted if interrupted?\n print \"#{ download_url } -> #{ game.fetch(:file_path) }\"\n\n File.open(game.fetch(:file_path), 'wb') do |file|\n file << agent.get(download_url).body\n end\n\n puts\n end",
"def save_as_JSON\n @result_pretty = JSON.pretty_generate(@result_scrap)\n File.open(\"./db/#{@name}.json\",\"w\") do |f|\n f.write(@result_pretty)\n end\n end",
"def dbpedia_json\n return @dbpedia_json if defined? @dbpedia_json\n\n @dbpedia_json = nil\n return @dbpedia_json if dbpedia_uri.blank?\n\n begin\n api = \"#{dbpedia_uri.gsub('resource', 'data')}.json\"\n response = URI.parse(api).open\n resp = response.read\n @dbpedia_json = JSON.parse(resp)\n rescue\n end\n end",
"def dump(data_path=File.join(DATA_DIR, \"#{Time.now.to_i}.json\"))\n\n if(File.exists?(data_path))\n raise ArgumentError.new(\"Refusing to overwrite existing file #{data_path}\")\n else\n dir, base = File.split(data_path)\n cmd = \"mkdir -p #{dir}\"\n pid = status = stderr = nil\n #Open3.popen3([env,] cmd... [, opts]) {|stdin, stdout, stderr, wait_thr|\n Open3.popen3(cmd) {|stdin, stdout, stderr, wait_thr|\n pid = wait_thr.pid # pid of the started process.\n status = wait_thr.value # Process::Status object returned.\n }\n if(status.success?)\n else\n raise ArgumentError.new(\"Could not create directory #{dir}; stderr of #{cmd} : #{stderr}\")\n end\n end\n\n sql = \"select * from #{TABLE_KEY}\"\n query = [\n \"key=#{PROJECT_KEY}\",\n \"sql=#{CGI.escape(sql)}\"\n ]\n url = \"#{TABLE_URL}?#{query.join(\"&\")}\"\n response = get(url)\n \n File.open(data_path, \"w\"){|ff|\n size = ff.write(response.body)\n }\nend"
] |
[
"0.57935965",
"0.5785309",
"0.5725889",
"0.57110524",
"0.5689905",
"0.55466974",
"0.5521662",
"0.55145407",
"0.5514303",
"0.55055064",
"0.5498958",
"0.54974514",
"0.5487188",
"0.54548717",
"0.5402494",
"0.5394242",
"0.5391018",
"0.53770953",
"0.53466976",
"0.5340965",
"0.5339588",
"0.53380495",
"0.5323183",
"0.5303047",
"0.52783453",
"0.52655774",
"0.5258098",
"0.5247964",
"0.5247964",
"0.5225899",
"0.51975715",
"0.5193637",
"0.5192165",
"0.5191833",
"0.5187772",
"0.5187741",
"0.5170931",
"0.5162975",
"0.5150888",
"0.5141426",
"0.51357836",
"0.5134779",
"0.5116119",
"0.5110357",
"0.5109003",
"0.5108459",
"0.51018864",
"0.5098945",
"0.509698",
"0.5079738",
"0.5078649",
"0.5072928",
"0.50671554",
"0.5064373",
"0.50638103",
"0.50605506",
"0.50582653",
"0.5045978",
"0.50357383",
"0.5019809",
"0.50159955",
"0.501439",
"0.50054306",
"0.5001134",
"0.5000083",
"0.4992305",
"0.49901417",
"0.49841282",
"0.49705875",
"0.49668747",
"0.49657026",
"0.4955402",
"0.494735",
"0.49450618",
"0.4936846",
"0.49334845",
"0.49314106",
"0.49248293",
"0.49229482",
"0.49145624",
"0.49127278",
"0.49112353",
"0.49095422",
"0.49093774",
"0.48986164",
"0.48957676",
"0.489461",
"0.48916683",
"0.48912746",
"0.4890599",
"0.48896474",
"0.4888812",
"0.48865518",
"0.4886375",
"0.48816445",
"0.48803777",
"0.4875682",
"0.4874733",
"0.48730016",
"0.48683837"
] |
0.5675986
|
5
|
Collects all contained fields.
|
def all_fields
found_fields = Array.new
self.fields.each do |field|
found_fields << field
found_fields = found_fields + field.all_fields if field.type == 'ObjectField'
end
found_fields
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def all_fields\n fields.values\n end",
"def all_fields\n @fields.values\n end",
"def fields\n @fields ||= self.class.fields.collect { |f| f.clone_for(self) }\n end",
"def fetch_fields\n @fields\n end",
"def fields\n @_fields ||= {}\n end",
"def all_fields\n superclass.all_fields.merge(own_fields)\n rescue NoMethodError\n own_fields\n end",
"def fields\n return @fields if defined?(@fields)\n\n @fields = array_of_items_for(Fields, :fields)\n end",
"def fields\n @fields ||= []\n end",
"def fields\n self.class.fields(true)\n end",
"def fields\n all_fields\n end",
"def fields\n @fields ||= []\n end",
"def fields\n @fields ||= {}\n end",
"def fields\n @fields ||= {}\n end",
"def fields\n @fields ||= {}\n end",
"def fields\n @fields ||= {}\n end",
"def fields\n @fields ||= {}\n end",
"def all_fields\n return @all_fields if defined? @all_fields\n\n @all_fields ||= fields.includes(field_set: [:catalog]).each_with_object([]) do |field, all|\n all << field\n next unless field.is_a?(Field::ChoiceSet)\n\n field.choices.includes(:category).each do |choice|\n category = choice.category\n next unless category&.not_deleted?\n\n additional_fields = category.fields.includes(field_set: [:catalog]).map do |f|\n f.category_choice = choice\n f.category_choice_set = field.choice_set\n f\n end\n all.concat(additional_fields)\n end\n end\n end",
"def all_fields\n interface_fields.merge(self.fields).values\n end",
"def fields\n []\n end",
"def fields\n injectors = @injectors || @propagators\n injectors.flat_map(&fields).uniq\n end",
"def load_fields!\n attributes = {}\n data_hash.each do |entry|\n next unless entry.is_a? Hash\n entry.keys.each { |key| attributes[key] ||= nil }\n end\n @fields = attributes.keys\n end",
"def fields\n @fields\n end",
"def fields\n @fields\n end",
"def all_fields\n Java::CascadingTuple::Fields::ALL\n end",
"def fields\n @fields ||= FieldList.new\n end",
"def fetch_fields\n @result.fetch_fields\n end",
"def fields\n @fields.keys\n end",
"def fields\n @fields.keys\n end",
"def fields\n @fields\n end",
"def fields; end",
"def fields; end",
"def fields; end",
"def all_fields\n fields = Set.new(@fields)\n @times.each {|entry| fields += entry.keys} unless fields.size > 0\n return fields.to_a.sort\n end",
"def fields!\n @schema.fields!\n end",
"def fields(field_hash = {})\n @fields.merge!(field_hash) unless field_hash.empty?\n @fields\n end",
"def fields\n @fields ||= begin\n default_fields\n end\n end",
"def fields #:nodoc:\n f = @properties.to_a\n last = f.pop\n f.push @elements\n f.push last\n end",
"def to_a; @fields end",
"def result_fields\n @fields ||= result_meta.collect { |m| m.name }\n end",
"def fields\n raw['fields']\n end",
"def fields\n self.class.fields\n end",
"def fields\n self.class.fields\n end",
"def fields\n self.class.fields\n end",
"def fields\n self.class.fields\n end",
"def fields\n self.find('ns:field', :ns => self.class.registered_ns).map do |field|\n Field.new(field)\n end\n end",
"def fields\n self.class.fields\n end",
"def each(&block)\n @fields.each(&block)\n end",
"def fields; h.fields; end",
"def all_attributes\n @@all_attributes ||= \n begin\n to_return = {}\n @fields.each do |key, field|\n to_return[key] = field.human_name\n end\n to_return\n end\n end",
"def fields()\n if !@custom_fields || @fields.length == 0 then\n @elements.each { |field|\n if field.respond_to?(:is_form_field) then\n @fields << field.name\n @element_map[field.name.to_s] = field\n elsif field.is_a?(Fieldset) then\n @fields << { field.name => field.fields }\n @element_map.update(field.element_map)\n end\n }\n end\n @fields.uniq!\n return @fields\n end",
"def fields\n nil\n end",
"def staff_request_all_fields\n StaffRequest.fields\n end",
"def attributes\n self.class.fields.inject({}) do |attribute_hash, field|\n attribute_hash[field] = send(field)\n attribute_hash\n end\n end",
"def contractor_request_all_fields\n ContractorRequest.fields\n end",
"def fields\n FIELDS\n end",
"def display_all_fields\n @@all_fields\n end",
"def fields_for_query\n self.class.fields_coercions.keys.each_with_object({}) do |field_name, results|\n results[field_name] = @fields.each_with_object({}) do |(locale, fields), field_results|\n field_results[locale] = get_value_from(fields, field_name)\n end\n end\n end",
"def fields_for_query\n self.class.fields_coercions.keys.each_with_object({}) do |field_name, results|\n results[field_name] = @fields.each_with_object({}) do |(locale, fields), field_results|\n field_results[locale] = get_value_from(fields, field_name)\n end\n end\n end",
"def fields\n fields = orig_fields\n\n fields.instance_exec(self) do |content_type|\n fields.define_singleton_method(:add) do |field|\n content_type.update(fields: content_type.merged_fields(field))\n end\n\n fields.define_singleton_method(:create) do |params|\n field = Contentful::Management::Field.new\n Field.property_coercions.each do |key, _value|\n snakify_key = Support.snakify(key)\n param = params[snakify_key.to_sym]\n field.send(\"#{snakify_key}=\", param) if param\n end\n content_type.update(fields: content_type.merged_fields(field))\n end\n\n fields.define_singleton_method(:destroy) do |id|\n fields = content_type.fields.reject { |field| field.id == id }\n content_type.update(fields: fields)\n end\n end\n\n fields\n end",
"def fields\r\n @hash.keys\r\n end",
"def fields\n @fields ||= @stats.keys.select{ |k| k != RECORDS_LISTING_KEY}.flatten.uniq\n end",
"def fields\n if frozen?\n Array(@gapi.fields).map { |f| Field.from_gapi(f).freeze }.freeze\n else\n Array(@gapi.fields).map { |f| Field.from_gapi f }\n end\n end",
"def fields(attributes = nil)\n clone.tap { |crit| crit.options[:fields] = attributes || {} }\n end",
"def fields\n FIELDS\n end",
"def labor_request_all_fields\n LaborRequest.fields\n end",
"def fields\n FIELDS\n end",
"def fields\n missing_method :fields\n end",
"def reflect_fields\n send_and_receive('admin/luke', params: { fl: '*', 'json.nl' => 'map' })['fields']\n end",
"def fields\n schema.fields\n end",
"def fields\n [*lookup]\n end",
"def fields\n @fields ||= form.fields\n end",
"def fields\n self.class.fields\n end",
"def fields\n all_fields = super\n interfaces.each do |int|\n if int.is_a?(GraphQL::InterfaceType)\n int_f = {}\n int.fields.each do |name, legacy_field|\n int_f[name] = field_class.from_options(name, field: legacy_field)\n end\n all_fields = int_f.merge(all_fields)\n end\n end\n all_fields\n end",
"def fields()\n @@defined_subclass_field_lists[self.class]\n end",
"def fields\n schema.fields\n end",
"def fields\n schema.fields\n end",
"def fields\n schema.fields\n end",
"def fields\n return @fields if @fields\n @file.rewind\n @fields = @file.gets.split(@separator).map &:strip\n end",
"def all\n @all ||= begin\n path = '../../vendor/choosealicense.com/_data/fields.yml'\n path = File.expand_path path, __dir__\n fields = YAML.safe_load File.read(path)\n fields.map { |field| LicenseField.from_hash(field) }\n end\n end",
"def each( &block )\n return self.fields.each( &block ) if block\n self.fields.each\n end",
"def fields\n java_import org.dspace.content.MetadataField\n MetadataField.findAllInSchema(DSpace.context, @obj.getSchemaID)\n end",
"def info_hash\n @fields\n end",
"def index\n @fields = all_fields\n end",
"def fields\n a = attributes.keys.to_a\n a.delete_if { | e | e =~ /^id$/ }\n end",
"def fields_from_record(options = {})\n field_attributes = _filter_field_attributes(options)\n field_attributes += [ options.fetch(:include, []) ]\n field_attributes.flatten!\n field_attributes.compact!\n field_attributes.uniq!\n\n field_attributes = field_attributes.inject({}) do |hash, field|\n if _protobuf_field_transformers.has_key?(field)\n hash[field] = _protobuf_field_transformers[field].call(self)\n else\n value = respond_to?(field) ? __send__(field) : nil\n hash[field] = _protobuf_convert_attributes_to_fields(field, value)\n end\n hash\n end\n\n field_attributes\n end",
"def _field_assignments\n result = []\n @fields.each do |field|\n field_type = field.type.type_sym # Record, Union, Enum, Array or Map\n schema_base_type = _schema_base_class(field.type)\n field_base_type = _field_type(schema_base_type)\n method_argument = %i(array map).include?(field_type) ? 'values' : 'value'\n is_schema_class = %i(record enum).include?(schema_base_type.type_sym)\n\n field_initialization = method_argument\n\n if is_schema_class\n field_initialization = \"#{field_base_type}.initialize_from_value(value)\"\n end\n\n result << {\n field: field,\n field_type: field_type,\n is_schema_class: is_schema_class,\n method_argument: method_argument,\n deimos_type: deimos_field_type(field),\n field_initialization: field_initialization\n }\n end\n\n result\n end",
"def fields\n klass.members.map(&:to_sym)\n end",
"def fields\n call(\"Webs\", \"get_columns\").xpath(\"//sp:Field\", NS).map do |field|\n attributes = clean_attributes(field.attributes)\n supersite && supersite.field(attributes[\"ID\"].downcase) || Field.new(self, attributes[\"ID\"].downcase, attributes[\"StaticName\"], attributes[\"Type\"], nil, attributes) if attributes[\"ID\"] && attributes[\"StaticName\"]\n end.compact\n end",
"def fields\n class_name.constantize.fields\n end",
"def flattened_fields\n metadata_string = String.from_java_bytes @reader.getMetadata\n description_text = self.class.description_from_metadata_xml(metadata_string)\n begin\n metadata = JSON(description_text)\n flattened_fields = metadata.key?('jpdfer_flattened_fields') ? metadata['jpdfer_flattened_fields'] : {}\n rescue JSON::ParserError\n flattened_fields = {}\n end\n flattened_fields.each_with_object({}) do |(name, value), fields|\n fields[name.to_sym] = value\n end\n end",
"def get_field_list\n return make_request(\"#{self.endpoint}/list/fields\")\n end",
"def fields\n iterator = @form_fields.keySet.iterator\n map = {}\n while iterator.hasNext\n key = iterator.next.toString\n map[key.to_sym] = field(key)\n end\n map\n end",
"def added_fields\n self.class.added_fields\n end",
"def export_all\n @exported_fields = nil\n end",
"def fields\n @fields = headers.map.with_index do |header, index|\n Field.new(header, row_data[index])\n end\n end",
"def fields\n @locations.map(&:field).uniq\n end",
"def computed_fields; end",
"def each_field &block\n @fields.each(&block)\n end",
"def get_entered_fields\n @entered_fields = get_used_fields_only(@contact_form_field)\n @entered_fields\n end",
"def fields\n self.class::FIELDS\n end"
] |
[
"0.7575919",
"0.75224596",
"0.7179399",
"0.707724",
"0.70342255",
"0.6961201",
"0.69537",
"0.69453555",
"0.6915724",
"0.69142556",
"0.69055235",
"0.68843734",
"0.68843734",
"0.68843734",
"0.68843734",
"0.68843734",
"0.68798184",
"0.6860387",
"0.68035775",
"0.6751277",
"0.6727666",
"0.6689157",
"0.6689157",
"0.6687476",
"0.6654465",
"0.6642568",
"0.6615831",
"0.6615831",
"0.65934974",
"0.6579184",
"0.6579184",
"0.6579184",
"0.6514417",
"0.64734447",
"0.64616406",
"0.6449627",
"0.64438486",
"0.64240825",
"0.6416781",
"0.64159864",
"0.64117223",
"0.64117223",
"0.64117223",
"0.64117223",
"0.64038086",
"0.6385881",
"0.6379048",
"0.63722885",
"0.6333687",
"0.63326293",
"0.633194",
"0.6318105",
"0.63131523",
"0.63059396",
"0.6263296",
"0.62539625",
"0.62387705",
"0.62387705",
"0.6221205",
"0.6214081",
"0.62119347",
"0.62092066",
"0.6197517",
"0.61920506",
"0.61856675",
"0.61684704",
"0.616455",
"0.6152786",
"0.6147549",
"0.6135215",
"0.6123469",
"0.6110051",
"0.6101494",
"0.60743964",
"0.60715455",
"0.60715455",
"0.60715455",
"0.60641414",
"0.6062628",
"0.6056406",
"0.6049908",
"0.6022634",
"0.6022477",
"0.6008353",
"0.59980035",
"0.5985786",
"0.59829515",
"0.5982896",
"0.5967941",
"0.5961105",
"0.59544915",
"0.59487665",
"0.59206825",
"0.59151566",
"0.59101444",
"0.59090793",
"0.5906708",
"0.58932334",
"0.58891296",
"0.58855504"
] |
0.7798035
|
0
|
Use callbacks to share common setup or constraints between actions.
|
def set_pet
@pet = Pet.find(params[:id])
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def set_required_actions\n # TODO: check what fields change to asign required fields\n end",
"def action_hook; end",
"def run_actions; end",
"def define_action_hook; end",
"def actions; end",
"def define_action_helpers\n if super && action == :save\n @instance_helper_module.class_eval do\n define_method(:valid?) do |*args|\n self.class.state_machines.fire_event_attributes(self, :save, false) { super(*args) }\n end\n end\n end\n end",
"def add_actions; end",
"def callbacks; end",
"def callbacks; end",
"def setup *actions, &proc\n (@setup_procs ||= []) << [proc, actions.size > 0 ? actions : [:*]]\n end",
"def define_action_helpers; end",
"def post_setup\n end",
"def action_methods; end",
"def action_methods; end",
"def action_methods; end",
"def before_setup; end",
"def action_run\n end",
"def execute(setup)\n @action.call(setup)\n end",
"def define_action_helpers?; end",
"def set_actions\n actions :all\n end",
"def action_done(action)\n dispatch = { :migrate => :done_migrating, :map => :done_mapping, :reduce =>\n :done_reducing, :finalize => :done_finalizing } \n self.send dispatch[action[:action]], action\n end",
"def dependencies action, &block\n @actions.each do |other|\n if action[:requires].include? other[:provide]\n block.call other\n end\n end\n end",
"def setup!\n return unless @setup_procs\n http_actions = actions\n @setup_procs.each do |setup_proc|\n proc, actions = setup_proc\n @setup__actions = actions.map do |action|\n\n action.is_a?(Regexp) ?\n http_actions.select { |a| a.to_s =~ action } :\n action.is_a?(String) && action =~ /\\A\\./ ?\n http_actions.map { |a| a.to_s << action if format?(a).include?(action) }.compact :\n action\n\n end.flatten\n self.class_exec &proc\n @setup__actions = nil\n end\n @setup_procs = nil\n end",
"def before_actions(*logic)\n self.before_actions = logic\n end",
"def setup_handler\n end",
"def set_action(opts)\n opts = check_params(opts,[:actions])\n super(opts)\n end",
"def setup(action)\n @targets.clear\n unless action.item.target_filters.empty?\n @targets = SES::TargetManager.make_targets(action)\n else\n item = action.item\n if item.for_opponent?\n @targets = $game_troop.alive_members\n elsif item.for_dead_friend?\n @targets = $game_party.battle_members.select { |actor| actor.dead? }\n else\n $game_party.battle_members.select { |actor| actor.alive? }\n end\n end\n @item_max = @targets.size\n create_contents\n refresh\n show\n activate\n end",
"def action; end",
"def action; end",
"def action; end",
"def action; end",
"def action; end",
"def workflow\n end",
"def revisable_shared_setup(args, block)\n class << self\n attr_accessor :revisable_options\n end\n options = args.extract_options!\n self.revisable_options = Options.new(options, &block)\n \n self.send(:include, Common)\n self.send(:extend, Validations) unless self.revisable_options.no_validation_scoping?\n self.send(:include, WithoutScope::QuotedColumnConditions)\n end",
"def setup\n @action = SampleActionAndroid.new(os_name: 'android',\n app_name: APP_PATH)\n end",
"def before(action)\n invoke_callbacks *self.class.send(action).before\n end",
"def process_action(...)\n send_action(...)\n end",
"def before_dispatch(env); end",
"def after_actions(*logic)\n self.after_actions = logic\n end",
"def setup\n # override and do something appropriate\n end",
"def setup(client)\n return unless @setup\n actions = @setup['setup'].select { |action| action['do'] }.map { |action| Action.new(action['do']) }\n actions.each do |action|\n action.execute(client)\n end\n self\n end",
"def setup(_context)\n end",
"def setup(resources) ; end",
"def validate_actions\n errors.add(:base, :should_give_at_least_one_action) if !manage? && !forecasting? && !read? && !api?\n end",
"def setup\n @resource_config = {\n :callbacks => {\n :before_create => nil,\n :after_create => nil,\n :before_update => nil,\n :after_update => nil,\n :before_destroy => nil,\n :after_destroy => nil,\n },\n :child_assoc => nil,\n :model => nil,\n :parent => nil,\n :path => nil,\n :permission => {},\n :properties => {},\n :relation => {\n :create => nil,\n :delete => nil,\n },\n :roles => nil,\n }\n end",
"def determine_valid_action\n\n end",
"def process_shared\n handle_taxes\n handle_shippings\n create_adjustments_from_params\n handle_status\n handle_inventory_refunds\n handle_payment_transactions\n order.updater.update\n end",
"def startcompany(action)\n @done = true\n action.setup\n end",
"def init_actions\n am = action_manager()\n am.add_action(Action.new(\"&Disable selection\") { @selection_mode = :none; unbind_key(32); bind_key(32, :scroll_forward); } )\n am.add_action(Action.new(\"&Edit Toggle\") { @edit_toggle = !@edit_toggle; $status_message.value = \"Edit toggle is #{@edit_toggle}\" })\n end",
"def event_callbacks(event, metadata={})\n case event\n when :reset, :review\n if confirmed\n update_attributes(confirmed: false)\n end\n when :confirm\n confirm\n # trigger :order for all applicable items\n # NOTE: :order event is common to both physical and digital items\n items.each do |i|\n if i.event_permitted(:order)\n user_id = last_transition.user_id\n i.trigger!(:order, { order_id: id, user_id: user_id })\n end\n end\n when :complete_work\n request = metadata[:request]\n work_complete_notification(request)\n when :close\n close\n end\n if event != :close && !open\n reopen\n end\n end",
"def setup_action\n return unless PONY::ERRNO::check_sequence(current_act)\n new_sequence = @action_sequence[@sequence_index+1...@action_sequence.size]\n @sequence_index = 0\n new_sequence = DND::SkillSequence::ACTS[@acts[1]] + new_sequence\n execute_sequence\n end",
"def define_tasks\n define_weave_task\n connect_common_tasks\n end",
"def setup(&block)\n define_method(:setup, &block)\n end",
"def setup\n transition_to(:setup)\n end",
"def setup\n transition_to(:setup)\n end",
"def action\n end",
"def setup( *args )\n\t\t\tself.class.setupBlocks.each {|sblock|\n\t\t\t\tdebugMsg \"Calling setup block method #{sblock}\"\n\t\t\t\tself.send( sblock )\n\t\t\t}\n\t\t\tsuper( *args )\n\t\tend",
"def config(action, *args); end",
"def setup\n @setup_proc.call(self) if @setup_proc\n end",
"def before_action \n end",
"def setup_callbacks\n defined_callbacks.each do |meth|\n unless respond_to?(\"call_#{meth}_callbacks\".to_sym)\n self.class.module_eval <<-EOE\n def call_#{meth}_callbacks(*args)\n plugin_store.each {|a| a.call_#{meth}_callbacks(*args) } if respond_to?(:plugin_store) && plugin_store\n self.send :#{meth}, *args if respond_to?(:#{meth})\n end\n EOE\n end\n end\n end",
"def action\n end",
"def matt_custom_action_begin(label); end",
"def setup\n # override this if needed\n end",
"def setup\n\t\t\t\t\t\t# Do nothing\n\t\t\t\tend",
"def setup\n\t\t\t\t\t\t# Do nothing\n\t\t\t\tend",
"def action(options,&callback)\n new_action = Action===options ? options : Action.new(options,&callback)\n # replace any with (shared name/alias or both default) + same arity\n @actions.delete_if do |existing_action|\n ((existing_action.names & new_action.names).size > 0 ||\n existing_action.default? && new_action.default?) &&\n existing_action.required.size == new_action.required.size &&\n existing_action.optional.size <= new_action.optional.size\n end\n @actions = (@actions + [new_action]).sort\n new_action\n end",
"def set_target_and_action target, action\n self.target = target\n self.action = 'sugarcube_handle_action:'\n @sugarcube_action = action\n end",
"def after(action)\n invoke_callbacks *options_for(action).after\n end",
"def pre_task\n end",
"def setup(server)\n server.on('beforeMethod', method(:before_method), 10)\n end",
"def add_actions\n attribute = machine.attribute\n name = self.name\n \n owner_class.class_eval do\n define_method(name) {self.class.state_machines[attribute].events[name].fire(self)}\n define_method(\"#{name}!\") {self.class.state_machines[attribute].events[name].fire!(self)}\n define_method(\"can_#{name}?\") {self.class.state_machines[attribute].events[name].can_fire?(self)}\n end\n end",
"def init_actions\n @select_action = SelectAction.new\n @endpoint_mouse_action = EndpointMouseAction.new\n @move_action = MoveAction.new\n end",
"def setup_signals; end",
"def after_created\r\n return unless compile_time\r\n Array(action).each do |action|\r\n run_action(action)\r\n end\r\nend",
"def after_created\r\n return unless compile_time\r\n Array(action).each do |action|\r\n run_action(action)\r\n end\r\nend",
"def set_target_and_action target, action\n self.target = target\n self.action = 'sugarcube_handle_action:'\n @sugarcube_action = action.respond_to?('weak!') ? action.weak! : action\n end",
"def initialize(*args)\n super\n @action = :set\nend",
"def after_set_callback; end",
"def setup\n #implement in subclass;\n end",
"def lookup_action; end",
"def setup &block\n if block_given?\n @setup = block\n else\n @setup.call\n end\n end",
"def setup_action\n return TSBS.error(@acts[0], 1, @used_sequence) if @acts.size < 2\n actions = TSBS::AnimLoop[@acts[1]]\n if actions.nil?\n show_action_error(@acts[1])\n end\n @sequence_stack.push(@acts[1])\n @used_sequence = @acts[1]\n actions.each do |acts|\n @acts = acts\n execute_sequence\n break if @break_action\n end\n @sequence_stack.pop\n @used_sequence = @sequence_stack[-1]\n end",
"def release_actions; end",
"def around_hooks; end",
"def save_action; end",
"def setup(easy)\n super\n easy.customrequest = @verb\n end",
"def action_target()\n \n end",
"def setup\n callback(:setup) do\n notify(:setup)\n migration_check.last_deployed_commit\n end\n end",
"def setup\n return unless @setup\n\n actions = @setup['setup'].select { |action| action['do'] }.map { |action| Action.new(action['do']) }\n run_actions_and_retry(actions)\n self\n end",
"def before_setup\n # do nothing by default\n end",
"def my_actions(options)\n @setup = false\n get_template_part(\"custom_used\",\"action_users\",true)\n end",
"def default_action; end",
"def setup(&blk)\n @setup_block = blk\n end",
"def callback_phase\n super\n end",
"def advice\n end",
"def _handle_action_missing(*args); end",
"def duas1(action)\n action.call\n action.call\nend",
"def shared_action(name, &block)\n @controller.shared_actions[name] = block\n end",
"def before_action action, &block\n @audience[:before][action] ||= Set.new\n @audience[:before][action] << block\n end",
"def setup_initial_state\n\n state_a = State.new(\"a\", 0)\n state_b = State.new(\"b\", 0)\n state_c = State.new(\"c\", 10)\n\n move_to_b = Action.new(\"move_to_b\", 1, state_b)\n\n move_to_c = Action.new(\"move_to_c\", 1, state_c)\n\n state_a.actions = [move_to_b, move_to_c]\n\n return state_a\n \nend"
] |
[
"0.6163163",
"0.6045976",
"0.5946146",
"0.591683",
"0.5890051",
"0.58349305",
"0.5776858",
"0.5703237",
"0.5703237",
"0.5652805",
"0.5621621",
"0.54210985",
"0.5411113",
"0.5411113",
"0.5411113",
"0.5391541",
"0.53794575",
"0.5357573",
"0.53402257",
"0.53394014",
"0.53321576",
"0.53124547",
"0.529654",
"0.5296262",
"0.52952296",
"0.52600986",
"0.52442724",
"0.52385926",
"0.52385926",
"0.52385926",
"0.52385926",
"0.52385926",
"0.5232394",
"0.523231",
"0.5227454",
"0.52226824",
"0.52201617",
"0.5212327",
"0.52079266",
"0.52050185",
"0.51754695",
"0.51726824",
"0.51710224",
"0.5166172",
"0.5159343",
"0.51578903",
"0.51522785",
"0.5152022",
"0.51518047",
"0.51456624",
"0.51398855",
"0.5133759",
"0.5112076",
"0.5111866",
"0.5111866",
"0.5110294",
"0.5106169",
"0.509231",
"0.50873137",
"0.5081088",
"0.508059",
"0.50677156",
"0.50562143",
"0.5050554",
"0.50474834",
"0.50474834",
"0.5036181",
"0.5026331",
"0.5022976",
"0.5015441",
"0.50121695",
"0.5000944",
"0.5000019",
"0.4996878",
"0.4989888",
"0.4989888",
"0.49864885",
"0.49797225",
"0.49785787",
"0.4976161",
"0.49683493",
"0.4965126",
"0.4958034",
"0.49559742",
"0.4954353",
"0.49535993",
"0.4952725",
"0.49467874",
"0.49423352",
"0.49325448",
"0.49282882",
"0.49269363",
"0.49269104",
"0.49252945",
"0.4923091",
"0.49194667",
"0.49174926",
"0.49173003",
"0.49171105",
"0.4915879",
"0.49155936"
] |
0.0
|
-1
|
Only allow a trusted parameter "white list" through.
|
def pet_params
params.require(:pet).permit(:name, :age)
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def allowed_params\n ALLOWED_PARAMS\n end",
"def expected_permitted_parameter_names; end",
"def param_whitelist\n [:role, :title]\n end",
"def default_param_whitelist\n [\"mode\"]\n end",
"def permitir_parametros\n \t\tparams.permit!\n \tend",
"def permitted_params\n []\n end",
"def strong_params\n params.require(:user).permit(param_whitelist)\n end",
"def strong_params\n params.require(:listing_member).permit(param_whitelist)\n end",
"def filtered_parameters; end",
"def permitted_strong_parameters\n :all #or an array of parameters, example: [:name, :email]\n end",
"def parameters_list_params\n params.require(:parameters_list).permit(:name, :description, :is_user_specific)\n end",
"def parameter_params\n params.require(:parameter).permit(:name, :description, :param_code, :param_value, :active_from, :active_to)\n end",
"def param_whitelist\n whitelist = [\n :description,\n :progress,\n :kpi_id\n ]\n \n unless action_name === 'create'\n whitelist.delete(:kpi_id)\n end\n \n whitelist\n end",
"def param_whitelist\n whitelist = [\n :username, :name,\n :parent_id,\n :headline, :description, :video,\n :policy, :signup_mode, :category,\n :website, :facebook, :twitter, :linkedin,\n :founded_at,\n privacy: [\n :events,\n :resources\n ],\n permission: [\n :profile,\n :members,\n :children,\n :statistics,\n :posts,\n :listings,\n :resources,\n :events\n ],\n location: [\n :description,\n :street,\n :city,\n :state,\n :zip,\n :country,\n :latitude,\n :longitude\n ]\n ]\n \n if action_name === 'update'\n whitelist.delete(:parent_id)\n unless current_user.role_in(@community) === 'owner'\n whitelist.delete(:privacy)\n whitelist.delete(:permission)\n end\n end\n \n whitelist\n end",
"def param_whitelist\n [:rating, :review]\n end",
"def valid_params?; end",
"def permitted_params\n declared(params, include_missing: false)\n end",
"def permitted_params\n declared(params, include_missing: false)\n end",
"def get_params\n\t\treturn ActionController::Parameters.new(self.attributes).permit(\"account_id\", \"title\", \"category\", \"introduction\", \"tags\", \"segment_type\", \"visible\", \"status\", \"main_image\")\n\tend",
"def filter_parameters; end",
"def filter_parameters; end",
"def strong_params\n params.require(:team_member).permit(param_whitelist)\n end",
"def strong_params\n params.require(:community).permit(param_whitelist)\n end",
"def check_params; true; end",
"def valid_params_request?; end",
"def strong_params\n params.require(:experience).permit(param_whitelist)\n end",
"def allowed_params\n params.require(:user).permit(:username, :email, :password, :password_confirmation)\n end",
"def list_params\n params.permit(:name)\n end",
"def grant_params\n @whitelisted = params.require(:grant).permit(:name, :description, :agency_id, :acronym)\n end",
"def check_params\n true\n end",
"def safe_params\n resurce_name = self.class.resource_name\n params_method_name = \"#{resurce_name}_params\".to_sym\n if params[resurce_name]\n if respond_to?(params_method_name) || private_methods.include?(params_method_name)\n send(params_method_name)\n else\n raise ActiveModel::ForbiddenAttributesError, \"Please, define the '#{params_method_name}' method in #{self.class.name}\"\n end\n end\n end",
"def additional_permitted_params\n []\n end",
"def strong_params\n params.require(:education).permit(param_whitelist)\n end",
"def resource_params\n params[resource_singular_name].try(:permit, self.class.param_whitelist)\n end",
"def allow_params_authentication!; end",
"def param_whitelist\n [\n :title,\n :description,\n :organization,\n :team_id,\n :started_at,\n :finished_at,\n location: [\n :description,\n :street,\n :city,\n :state,\n :zip,\n :country,\n :latitude,\n :longitude\n ]\n ]\n end",
"def param_whitelist\n if @user.present? && current_user != @user\n return [:followed]\n end\n \n whitelist = [\n :username, :email, :password,\n :first_name, :last_name,\n :birthday, :gender,\n :headline, :biography, :ask_about, :focus,\n :website, :facebook, :linkedin, :twitter, :github,\n roles: [],\n skills: [],\n interests: [],\n privacy: { contact: [] },\n location: [\n :description,\n :street,\n :city,\n :state,\n :zip,\n :country,\n :latitude,\n :longitude\n ]\n ]\n \n if action_name === 'update'\n whitelist.delete(:email)\n whitelist.delete(:password)\n end\n \n whitelist\n end",
"def person_params\n # params whitelist does *not* include admin, sub, remember_token\n # TBD: share this whitelist with the list used by configuration_permitted_parameters\n # TBD: should current_password be on this list? -- for now, leaving off, since it seems to work without\n # NOTE: do not include 'admin' in this list!\n params.require(:person).permit(\n :name, \n :email, \n :description,\n :password, \n :password_confirmation\n )\n end",
"def paramunold_params\n params.require(:paramunold).permit!\n end",
"def param_params\n params.require(:param).permit(:param_category_id, :param_table_id, :name, :english_name, :weighting, :description)\n end",
"def quote_params\n params.permit!\n end",
"def list_params\n params.permit(:list_name)\n end",
"def allowed_params(parameters)\n parameters.select do |name, values|\n values.location != \"path\"\n end\n end",
"def all_params; end",
"def permitted_resource_params\n params[resource.object_name].present? ? params.require(resource.object_name).permit! : ActionController::Parameters.new\n end",
"def source_params\n params.require(:source).permit(all_allowed_params)\n end",
"def get_allowed_parameters\n return _get_specific_action_config(:allowed_action_parameters, :allowed_parameters)&.map(&:to_s)\n end",
"def user_params\n end",
"def permitted_params\n @wfd_edit_parameters\n end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def param_whitelist\n whitelist = [\n :comment,\n :old_progress, :new_progress,\n :metric_id\n ]\n \n unless action_name === 'create'\n whitelist.delete(:metric_id)\n end\n \n whitelist\n end",
"def user_params\r\n end",
"def query_param\n\t\tparams.permit(:first_name, :last_name, :phone)\n\tend",
"def whitelisted_user_params\n params.require(:user).\n permit( :first_name, :last_name, :email,:password,:password_confirmation,:birthday,:gender)\n end",
"def user_params\n params.permit(:id, :email, :password, :nickname, :status, :avatar, :flat_picture, :flatsharing_id, :member,\n :user, :color, :solde)\n end",
"def filter_params\n\t\treturn params[:candidate].permit(:name_for_filter)\n\tend",
"def devise_filter\r\n logger.debug(\"In devise_filter =>PARAMS: #{params.inspect}\")\r\n\r\n # White list for sign_up\r\n devise_parameter_sanitizer.for(:sign_up) { |u| u.permit(user_whitelist) }\r\n\r\n # White list for account update\r\n devise_parameter_sanitizer.for(:account_update) { |u| u.permit(user_whitelist, :current_password) }\r\n\r\n # White list for Invitation creation\r\n devise_parameter_sanitizer.for(:invite) { |u| u.permit(:account_type, :email, :invitation_token)}\r\n\r\n # White list for accept invitation\r\n devise_parameter_sanitizer.for(:accept_invitation) { |u| u.permit(user_whitelist, :invitation_token)}\r\n\r\n end",
"def get_params\n\t\t\n\t\treturn ActionController::Parameters.new(self.attributes).permit(:first_name, :last_name, :email, :provider)\n\n\tend",
"def valid_params(params)\n params.permit(:user_id, :photo_id, :originX, :originY, :width, :height)\n end",
"def valid_parameters\n sort_symbols(@interface.allowed_parameters)\n end",
"def params_permit\n params.permit(:id)\n end",
"def allowed_params\n params.require(:allowed).permit(:email)\n end",
"def allowed_params\n params.require(:user).permit(:email, :password, :role, :first_name, :last_name, :password_confirmation)\n end",
"def allowed_params\n params.require(:user).permit(:email, :password, :role, :first_name, :last_name, :password_confirmation)\n end",
"def filter_params\n params.permit(*resource_filter_permitted_params)\n end",
"def specialty_params\n\t\tparams.require(:specialty).permit(*Specialty::DEFAULT_ACCESSIBLE_ATTRIBUTES)\n\tend",
"def community_params\n params.permit(:profile_image, :name, :description, :privacy_type, :viewed_by, {tags: []}, {features: []}, {admins: []}, :members, :location, :beacon, :creator, :ambassadors, :current_events, :past_events, :feed, :category, :address, :allow_member_post_to_feed, :allow_member_post_to_events)\n end",
"def authorize_params\n super.tap do |params|\n %w[display scope auth_type].each do |v|\n if request.params[v]\n params[v.to_sym] = request.params[v]\n end\n end\n end\n end",
"def feature_params_filter\n params.require(:feature).permit(:name, :cat, :lower, :upper, :opts, :category, :description, :company, :active, :unit, :icon)\n end",
"def available_activity_params\n # params.require(:available_activity).permit(:type,:geometry,:properties)\n whitelisted = ActionController::Parameters.new({\n type: params.require(:available_activity)[:type],\n geometry: params.require(:available_activity)[:geometry].try(:permit!).to_h,\n properties: params.require(:available_activity)[:properties].try(:permit!).to_h\n }).try(:permit!)\n end",
"def argument_params\n params.require(:argument).permit(:name)\n end",
"def user_params_pub\n\t \tparams[:user].permit(:hruid)\n\t end",
"def strong_params\n params.require(:success_metric).permit(param_whitelist)\n end",
"def property_params\n params.permit(:name, :is_available, :is_approved, :owner_id)\n end",
"def restricted_params\n #params.require(self.controller_name.classify.underscore.to_sym).permit([])\n raise(\"No strong params set, override restricted_params method in your controller. E.g. params.require(:model).permit(:attribute1, :attribute2)\")\n end",
"def sponsor_params\n params.require(:sponsor).permit(WHITE_LIST)\n end",
"def whitelist_person_params\n params.require(:person).permit(:family, :pre_title, :given_name, :dates, :post_title, :epithet, :dates_of_office, same_as: [], related_authority: [], altlabel: [], note: []) # Note - arrays need to go at the end or an error occurs!\n end",
"def user_params \n \tparams.require(:user).permit(:name, :email, :password, :password_confirmation)# preventing CSTR\n end",
"def sequence_param_whitelist\n default_param_whitelist << \"show_index\"\n end",
"def parameters\n nil\n end",
"def resource_filter_permitted_params\n raise(NotImplementedError, 'resource_filter_permitted_params method not implemented')\n end",
"def normal_params\n reject{|param, val| param_definitions[param][:internal] }\n end",
"def special_device_list_params\n params.require(:special_device_list).permit(:name)\n end",
"def validate_search_inputs\n @whitelisted = params.fetch(:user, nil)\n if @whitelisted.blank?\n render_error(400, \"#{I18n.t('general_error.params_missing_key')}\": [I18n.t('general_error.params_missing_value', model: \"review\")])\n return\n else\n @whitelisted = @whitelisted.permit(:name, :uen, :description)\n end\n end",
"def pull_request_params\n whitelist = [\n :url,\n :id,\n :html_url,\n :diff_url,\n :patch_url,\n :issue_url,\n :number,\n :state,\n :locked,\n :title\n ]\n params.require(:pull_request).permit(whitelist)\n end"
] |
[
"0.71230334",
"0.70530915",
"0.69479465",
"0.6902122",
"0.67367256",
"0.67172784",
"0.6689043",
"0.66784793",
"0.6660117",
"0.6555213",
"0.6528485",
"0.6458438",
"0.6452378",
"0.6451654",
"0.64478326",
"0.6433326",
"0.6413599",
"0.6413599",
"0.63907677",
"0.63787645",
"0.63787645",
"0.6375229",
"0.63608277",
"0.635366",
"0.6283652",
"0.62798274",
"0.6245606",
"0.62283605",
"0.6224614",
"0.6223649",
"0.62118477",
"0.6207179",
"0.61780804",
"0.6173056",
"0.61674094",
"0.615996",
"0.6145132",
"0.613597",
"0.612235",
"0.6108622",
"0.6098955",
"0.60767287",
"0.6055062",
"0.60391796",
"0.60363555",
"0.6030472",
"0.6018476",
"0.60174584",
"0.60163116",
"0.60159874",
"0.60159874",
"0.60159874",
"0.60159874",
"0.60159874",
"0.60159874",
"0.60159874",
"0.60159874",
"0.60159874",
"0.60159874",
"0.60159874",
"0.60159874",
"0.60159874",
"0.60159874",
"0.60159874",
"0.60159874",
"0.60159874",
"0.60052663",
"0.6003681",
"0.6001089",
"0.5996807",
"0.5994288",
"0.59942675",
"0.5984987",
"0.59827954",
"0.59777087",
"0.5975369",
"0.59706473",
"0.5966046",
"0.5965166",
"0.5965166",
"0.59577847",
"0.5952617",
"0.59503365",
"0.59480196",
"0.5943258",
"0.5931462",
"0.59299",
"0.5927073",
"0.5924737",
"0.5919184",
"0.5918459",
"0.591457",
"0.59142643",
"0.59062785",
"0.59054136",
"0.59047925",
"0.5902357",
"0.5900476",
"0.5898475",
"0.5898218",
"0.5895328"
] |
0.0
|
-1
|
GET /documents GET /documents.json
|
def index
@attachments = Attachment.all
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def documents(params={})\n server.get(\"#{name}/_all_docs\", params)\n end",
"def index\n @documents = Document.all\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @documents }\n end\n end",
"def documents\n return bad_request unless params[:id] and request.format.json? || request.format.js? || request.format.text?\n return not_found unless current_document\n opts = {:access => true, :sections => true, :annotations => true, :data => true}\n if current_account\n opts[:account] = current_account\n opts[:allowed_to_edit] = current_account.allowed_to_edit?(current_document)\n opts[:allowed_to_review] = current_account.reviews?(current_document)\n end\n @response = {'document' => current_document.canonical(opts)}\n respond_to do |format|\n format.text do\n direct = [PRIVATE, ORGANIZATION, EXCLUSIVE].include? current_document.access\n redirect_to(current_document.full_text_url(direct: direct))\n end\n format.json { render_cross_origin_json }\n format.js { render_cross_origin_json }\n end\n end",
"def show\n \n @document = Document.find(params[:id])\n \n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @document }\n end\n \n end",
"def show\n @document = Document.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @document }\n end\n end",
"def show\n @document = Document.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @document }\n end\n end",
"def show\n @document = Document.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @document }\n end\n end",
"def show\n @document = Document.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @document }\n end\n end",
"def show\n @document = Document.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @document }\n end\n end",
"def show\n @document = Document.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @document }\n end\n end",
"def show\n @document = Document.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @document }\n end\n end",
"def index\n @documents = Document.all\n @document = Document.new\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @documents }\n end\n end",
"def show\n @document = Document.find(params[:id])\n\t\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @document }\n end\n end",
"def show\n @document = Document.find(params[:id])\n\n respond_to do |format|\n format.html # _resource.html.erb\n format.json { render json: @document }\n end\n end",
"def show\r\n @document = Document.find(params[:id])\r\n\r\n respond_to do |format|\r\n format.html # show.html.erb\r\n format.json { render json: @document }\r\n end\r\n end",
"def index\n @documents = current_user.documents.order(\"id DESC\").all\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @documents }\n end\n end",
"def index_documents\n @params = {}\n @action = 'index_documents'\n \n send_auth_request\n end",
"def show\n @knowledge = Knowledge.find(params[:id])\n @documents = @knowledge.documents\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @knowledge }\n end\n end",
"def show\n\t@document = Document.find(params[:id])\n\trender json: {status: 'SUCCESS', message:'Loaded document', data:@document}, status: :ok\n end",
"def show\n @doucment = Document.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @document }\n end\n end",
"def documents\n @documents ||=\n Array.wrap(options[:documents]).compact.presence ||\n (response['docs'] || []).map do |doc|\n document_factory.build(doc, self, options)\n end\n end",
"def index\n\n @documentable = find_resource\n @documents = @documentable.documents\n\n end",
"def index\n @documents = Document.for_user(current_user).order('created_at desc')\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @documents }\n end\n end",
"def show\n @doc = Doc.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @doc }\n end\n end",
"def show\n respond_to do |format|\n format.html {\n redirect_to collection_documents_path(@collection)\n }\n format.json\n end\n end",
"def index\n @documents = Document.all\n\n respond_to do |format|\n format.html{\n }\n format.json {\n ret = @documents.as_json\n render json:{\n status: 0,\n total: ret.count,\n result: ret\n }\n }\n end\n end",
"def show\n @documento = Documento.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @documento }\n end\n end",
"def index\n \t@documents = Document.paginate(:page => params[:page], per_page: 2)\n \trender json: {status: 'SUCCESS', message:'Loaded documents', data:@documents}, status: :ok\n end",
"def document\n json = Net::HTTP.get_response URI.parse(query_string(@query))\n json.body\n end",
"def info\n @document = Document.find(params[:id])\n respond_to do |format|\n format.json { render :json => @document, serializer: Api::Mobile::V2::DynamicDocumentSerializer, root: 'document' }\n end\n end",
"def get(id)\n Gini::Api::Document.new(self, \"/documents/#{id}\")\n end",
"def index\n if (params[:public])\n @documents = Document.where('public = ?', params[:public])\n else\n @documents = Document.all\n end\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render jbuilder: @documents }\n end\n end",
"def documents\n Hancock::Request.send_get_request(\"/envelopes/#{envelope_id}/documents\")[\"envelopeDocuments\"]\n end",
"def show\n @document = Document.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render jbuilder: @document }\n end\n end",
"def index\n @special_documents = ModifiedDocument.all\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @special_documents }\n end\n end",
"def documents; end",
"def documents; end",
"def service_document\n response = get(@url.to_s)\n response.body\n end",
"def show\n @document = Document.where(:id => params[:id])\n render :json => @document, :include => [:versions]\n end",
"def show\n @document = Document.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.xml { render :xml => @document }\n format.json { render :json => @document }\n end\n end",
"def show\n @document_page = DocumentPage.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @document_page }\n end\n end",
"def index\n @documents = Document.all.delete_if { |document| cannot? :read, document }\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @documents }\n format.xml { render xml: @documents }\n end\n end",
"def get_document index, id\n uri = URI(\"http://#{@host}:#{@port_s}/#{index}/_doc/#{id}\")\n req = Net::HTTP::Get.new(uri)\n run(uri, req)\n end",
"def show\n @documents = current_user.documents.find(params[:id])\n end",
"def show\n\n \t\n @companydocument = Companydocument.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @companydocument }\n end\n\n end",
"def index\n @documents = current_user.documents.all\n end",
"def index\n @documents = Document.all\n end",
"def index\n @documents = Document.all\n end",
"def index\n @documents = Document.all\n end",
"def index\n @documents = Document.all\n end",
"def index\n @documents = Document.all\n end",
"def index\n @documents = Document.all\n end",
"def index\n @documents = Document.all\n end",
"def show\n @document = Document.find(params[:id])\n \n not_found unless current_user == @document.user\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @document }\n end\n end",
"def show\n @oa_sent_document = Oa::SentDocument.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @oa_sent_document }\n end\n end",
"def get_document_info\n\n begin\n\n str_uri = $product_uri + '/words/' + @filename\n signed_str_uri = Aspose::Cloud::Common::Utils.sign(str_uri)\n\n response_stream = RestClient.get(signed_str_uri, {:accept => 'application/json'})\n\n stream_hash = JSON.parse(response_stream)\n stream_hash['Code'] == 200 ? stream_hash['Document'] : false\n\n rescue Exception => e\n print e\n end\n\n end",
"def show\n respond_with( @document = Document.find(params[:id]) )\n end",
"def index\n authorize Document\n\n @documents = apply_scopes(Document).valid.all\n\n if current_user.client?\n @documents = @documents.by_belongings([current_user.id])\n end\n\n respond_with do |format|\n format.html\n format.json { render json: DocumentsDatatable.new(view_context, @documents) }\n format.js\n end\n\n end",
"def getDocuments(projectId, queryMap)\t\t\r\n\t\t\t\turl = getBaseURL+\"projects/\"+String(projectId)+\"/documents/\"\t\t\r\n\t\t\t\tresponse = ZohoHTTPClient.get(url, getQueryMap(queryMap))\t\t\r\n\t\t\t\treturn $documentParser.getDocuments(response)\r\n\t\t\tend",
"def get_document_all_using_get_with_http_info(opts = {})\n if @api_client.config.debugging\n @api_client.config.logger.debug 'Calling API: DocumentApi.get_document_all_using_get ...'\n end\n # resource path\n local_var_path = '/nucleus/v1/document'\n\n # query parameters\n query_params = {}\n query_params[:'ascending'] = opts[:'ascending'] if !opts[:'ascending'].nil?\n query_params[:'filter'] = opts[:'filter'] if !opts[:'filter'].nil?\n query_params[:'order_by'] = opts[:'order_by'] if !opts[:'order_by'].nil?\n query_params[:'page'] = opts[:'page'] if !opts[:'page'].nil?\n query_params[:'size'] = opts[:'size'] if !opts[:'size'].nil?\n\n # header parameters\n header_params = {}\n # HTTP header 'Accept' (if needed)\n header_params['Accept'] = @api_client.select_header_accept(['*/*'])\n\n # form parameters\n form_params = {}\n\n # http body (model)\n post_body = nil\n auth_names = ['oauth2']\n data, status_code, headers = @api_client.call_api(:GET, local_var_path,\n :header_params => header_params,\n :query_params => query_params,\n :form_params => form_params,\n :body => post_body,\n :auth_names => auth_names,\n :return_type => 'PageDocument')\n if @api_client.config.debugging\n @api_client.config.logger.debug \"API called: DocumentApi#get_document_all_using_get\\nData: #{data.inspect}\\nStatus code: #{status_code}\\nHeaders: #{headers}\"\n end\n return data, status_code, headers\n end",
"def documents(type: nil) # \"path\", \"id\", \"key\" # TESTED\n body = {\n \"collection\" => @collection,\n \"type\" => type\n }.delete_if{|k,v| v.nil?}.to_json\n request = @@request.merge({ :body => body })\n result = self.class.put(\"/_db/#{@database}/_api/simple/all-keys\", request)\n return result.headers[\"x-arango-async-id\"] if @@async == \"store\"\n return true if @@async\n result = result.parsed_response\n if type.nil?\n @@verbose ? result : result[\"error\"] ? result[\"errorMessage\"] : result[\"result\"].map{|x| value = self.class.get(x).parsed_response; ArangoDocument.new(key: value[\"_key\"], collection: @collection, body: value)}\n else\n @@verbose ? result : result[\"error\"] ? result[\"errorMessage\"] : result[\"result\"]\n end\n end",
"def index\n if !@db\n render json: [], status: 200\n else\n docs = @db.all_docs(params={\"include_docs\":\"true\"})[\"rows\"]\n names = docs.map { |d| d[\"doc\"][\"name\"] }\n render json: JSON.dump(names), status: 200\n end\n end",
"def index\n @docs = Oa::SentDocument.all#.paginate(:page => params[:page], :order => 'created_at desc')\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @docs }\n end\n end",
"def show\n respond_to do |format|\n format.html {\n redirect_to project_documents_path(@project)\n }\n format.json\n end\n end",
"def show\n @documentotipo = Documentotipo.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @documentotipo }\n end\n end",
"def show\n @document_type = DocumentType.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @document_type }\n end\n end",
"def index\n @space = Space.where(:wiki_name => params[:space_id])[0]\n @documents = @space.documents.all\n\n respond_to do |format|\n format.html # index.html.erb\n format.xml { render :xml => @documents }\n format.json { render :json => @documents.to_json }\n end\n end",
"def document(document_id)\n Hancock::Request.send_get_request(\"/envelopes/#{envelope_id}/documents/#{document_id}\")\n end",
"def documents\n @assets = filtered_assets.where(type: \"Document\").order(\"created_at DESC\").paginate(params)\n @assets = @assets.search(params[:q], fuzzy: true) if params[:q].present?\n respond_to do |format|\n format.html do\n render template: \"/dash/chooser/documents\"\n end\n format.js do\n render template: \"/dash/chooser/documents_search\"\n end\n end\n end",
"def show \n @collection = Collection.find(params[:id])\n\n #@documents = Document.where(:collection_id => @collection.id).paginate(:per_page => 5, :page => params[:page])\n @documents_all = Document.where(:collection_id => @collection.id)\n\n @documents = []\n @documents_all.each do |doc|\n if doc_is_viewable(doc, current_user)\n @documents << doc\n end\n end\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @collection }\n end\n end",
"def list_documents(database_id:, collection_id:, queries: nil)\n path = '/databases/{databaseId}/collections/{collectionId}/documents'\n .gsub('{databaseId}', database_id)\n .gsub('{collectionId}', collection_id)\n\n if database_id.nil?\n raise Appwrite::Exception.new('Missing required parameter: \"databaseId\"')\n end\n\n if collection_id.nil?\n raise Appwrite::Exception.new('Missing required parameter: \"collectionId\"')\n end\n\n params = {\n queries: queries,\n }\n \n headers = {\n \"content-type\": 'application/json',\n }\n\n @client.call(\n method: 'GET',\n path: path,\n headers: headers,\n params: params,\n response_type: Models::DocumentList\n )\n end",
"def show\n @document = Document.find(params[:id])\n end",
"def get_document_info\n \n begin\n \n if @filename == \"\"\n raise \"Base file not specified.\"\n end\n \n str_uri = $productURI + \"/words/\" + @filename\n signed_str_uri = Common::Utils.sign(str_uri)\n \n response_stream = RestClient.get(signed_str_uri,{:accept=>\"application/json\"})\n \n stream_hash = JSON.parse(response_stream)\n \n if(stream_hash[\"Code\"] == 200)\n return stream_hash[\"Document\"]\n else\n return false\n end\n \n rescue Exception=>e\n print e\n end\n \n end",
"def show\n @paper = Paper.find(params[:id])\n @document = Document.where(paper_id: @paper.id)\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @paper }\n end\n end",
"def show\n @special_document = ModifiedDocument.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @special_document }\n end\n end",
"def show\n @modified_document = ModifiedDocument.find(params[:id])\n render json: @modified_document, status: :ok\n end",
"def show\n respond_to do |format|\n format.html # show.html.erb\n format.json { render :json => @shared_document }\n end\n end",
"def show\n @project = Project.find(params[:id])\n db = MongoMapper.database\n @documents = db[\"documents\"].find( { \"_id\" => { \"$in\" => @project.document_ids} } )\n\n respond_to do |format|\n format.html # show.html.erb\n format.xml { render :xml => @project }\n format.json { render :json => @project }\n end\n end",
"def show\r\n @customer_document = CustomerDocument.find(params[:id])\r\n\r\n respond_to do |format|\r\n format.html # show.html.erb\r\n format.json { render json: @customer_document }\r\n end\r\n end",
"def index\n #@documents = Document.all\n @documents = current_user.documents\n end",
"def show\n @document_history = DocumentHistory.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @document_history }\n end\n end",
"def document_details(guid)\n get \"/api/documents/#{guid}.xml\"\n end",
"def index\n params.delete :utf8\n @documents = Document.filter(params)\n #@documents = Document.all\n\t\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @documents }\n end\n end",
"def get_document( doc_id:, version:CURRENT, filename: )\n params = {}\n params[:backtrace] = @backtrace if @backtrace\n send_request :get, path_for(doc_id, filename, version), {}, :binary\n end",
"def documents\n Easybill::Api::Documents\n end",
"def get_document_info( doc_id: )\n params = {}\n params[:backtrace] = @backtrace if @backtrace\n send_request :get, url_for_base(doc_id), params, :json\n end",
"def get_document(database_id:, collection_id:, document_id:, queries: nil)\n path = '/databases/{databaseId}/collections/{collectionId}/documents/{documentId}'\n .gsub('{databaseId}', database_id)\n .gsub('{collectionId}', collection_id)\n .gsub('{documentId}', document_id)\n\n if database_id.nil?\n raise Appwrite::Exception.new('Missing required parameter: \"databaseId\"')\n end\n\n if collection_id.nil?\n raise Appwrite::Exception.new('Missing required parameter: \"collectionId\"')\n end\n\n if document_id.nil?\n raise Appwrite::Exception.new('Missing required parameter: \"documentId\"')\n end\n\n params = {\n queries: queries,\n }\n \n headers = {\n \"content-type\": 'application/json',\n }\n\n @client.call(\n method: 'GET',\n path: path,\n headers: headers,\n params: params,\n response_type: Models::Document\n )\n end",
"def get(projectId, documentId, queryMap)\r\n\t\t\t\turl = getBaseURL+\"projects/\"+String(projectId)+\"/documents/\"+String(documentId)+\"/\"\r\n\t\t\t\tresponse = ZohoHTTPClient.get(url, getQueryMap(queryMap))\t\t\r\n\t\t\t\treturn $documentParser.getDocument(response)\r\n\t\t\tend",
"def new\n @document = Document.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @document }\n end\n end",
"def new\n @document = Document.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @document }\n end\n end",
"def new\n @document = Document.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @document }\n end\n end",
"def new\n @document = Document.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @document }\n end\n end",
"def new\n @document = Document.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @document }\n end\n end",
"def new\n @document = Document.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @document }\n end\n end",
"def new\n @document = Document.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @document }\n end\n end",
"def new\n @document = Document.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @document }\n end\n end",
"def new\n @document = Document.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @document }\n end\n end",
"def document\n params.has_key?('_json') ? params['_json'] : params\n end",
"def eds_documents(data)\n docs = data&.dig('response', 'docs')\n docs = Array.wrap(docs).compact\n factory = blacklight_config.document_factory\n model_opt = { lens: blacklight_config.lens_key }\n docs.map { |doc| factory.build(doc, data, model_opt) }\n end",
"def new\r\n @document = Document.new\r\n\r\n respond_to do |format|\r\n format.html # new.html.erb\r\n format.json { render json: @document }\r\n end\r\n end",
"def index\n @person = Person.find((params[:person_id]))\n @person_identification_docs = PersonIdentificationDoc.where(\"person_id = ?\", params[:person_id])\n\n\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render :json => @person_identification_docs }\n end\n end"
] |
[
"0.7656662",
"0.7626486",
"0.75327045",
"0.7379236",
"0.73293954",
"0.73293954",
"0.73293954",
"0.73293954",
"0.73293954",
"0.73293954",
"0.73293954",
"0.7327912",
"0.73046535",
"0.7289063",
"0.72620124",
"0.7253349",
"0.7240533",
"0.7154175",
"0.7143549",
"0.7138256",
"0.7117629",
"0.7110345",
"0.7072527",
"0.70057225",
"0.70028484",
"0.6999755",
"0.6996226",
"0.6968826",
"0.6911097",
"0.6897962",
"0.6891248",
"0.6864924",
"0.6850267",
"0.6838529",
"0.6828052",
"0.6786018",
"0.6786018",
"0.67633694",
"0.6762873",
"0.6754818",
"0.67340624",
"0.671717",
"0.6715148",
"0.670757",
"0.6684289",
"0.66812176",
"0.6671473",
"0.6671473",
"0.6671473",
"0.6671473",
"0.6671473",
"0.6671473",
"0.6671473",
"0.6643592",
"0.6594412",
"0.6593195",
"0.6590831",
"0.6584786",
"0.65707844",
"0.6566993",
"0.6563477",
"0.656041",
"0.6560237",
"0.65491116",
"0.65403116",
"0.6533303",
"0.6532149",
"0.6493753",
"0.64918077",
"0.6468988",
"0.64674485",
"0.64645875",
"0.6462418",
"0.64599407",
"0.64593697",
"0.64567673",
"0.64554703",
"0.6454668",
"0.6446817",
"0.64405835",
"0.64392465",
"0.6435274",
"0.64334327",
"0.6428762",
"0.64278346",
"0.64268124",
"0.64167386",
"0.6415706",
"0.6411099",
"0.6411099",
"0.6411099",
"0.6411099",
"0.6411099",
"0.6411099",
"0.6411099",
"0.6411099",
"0.6411099",
"0.6397768",
"0.6394564",
"0.6390629",
"0.63814247"
] |
0.0
|
-1
|
GET /documents/1 GET /documents/1.json
|
def show
send_data(@attachment.file_contents,
type: @attachment.content_type,
filename: @attachment.filename)
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def show\n \n @document = Document.find(params[:id])\n \n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @document }\n end\n \n end",
"def index\n @documents = Document.all\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @documents }\n end\n end",
"def show\n @document = Document.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @document }\n end\n end",
"def show\n @document = Document.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @document }\n end\n end",
"def show\n @document = Document.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @document }\n end\n end",
"def show\n @document = Document.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @document }\n end\n end",
"def show\n @document = Document.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @document }\n end\n end",
"def show\n @document = Document.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @document }\n end\n end",
"def show\n @document = Document.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @document }\n end\n end",
"def show\n @document = Document.find(params[:id])\n\t\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @document }\n end\n end",
"def get_document index, id\n uri = URI(\"http://#{@host}:#{@port_s}/#{index}/_doc/#{id}\")\n req = Net::HTTP::Get.new(uri)\n run(uri, req)\n end",
"def show\n @document = Document.find(params[:id])\n\n respond_to do |format|\n format.html # _resource.html.erb\n format.json { render json: @document }\n end\n end",
"def show\r\n @document = Document.find(params[:id])\r\n\r\n respond_to do |format|\r\n format.html # show.html.erb\r\n format.json { render json: @document }\r\n end\r\n end",
"def index\n @documents = Document.all\n @document = Document.new\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @documents }\n end\n end",
"def get(id)\n Gini::Api::Document.new(self, \"/documents/#{id}\")\n end",
"def show\n\t@document = Document.find(params[:id])\n\trender json: {status: 'SUCCESS', message:'Loaded document', data:@document}, status: :ok\n end",
"def show\n @doc = Doc.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @doc }\n end\n end",
"def show\n @doucment = Document.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @document }\n end\n end",
"def documents\n return bad_request unless params[:id] and request.format.json? || request.format.js? || request.format.text?\n return not_found unless current_document\n opts = {:access => true, :sections => true, :annotations => true, :data => true}\n if current_account\n opts[:account] = current_account\n opts[:allowed_to_edit] = current_account.allowed_to_edit?(current_document)\n opts[:allowed_to_review] = current_account.reviews?(current_document)\n end\n @response = {'document' => current_document.canonical(opts)}\n respond_to do |format|\n format.text do\n direct = [PRIVATE, ORGANIZATION, EXCLUSIVE].include? current_document.access\n redirect_to(current_document.full_text_url(direct: direct))\n end\n format.json { render_cross_origin_json }\n format.js { render_cross_origin_json }\n end\n end",
"def documents(params={})\n server.get(\"#{name}/_all_docs\", params)\n end",
"def index_documents\n @params = {}\n @action = 'index_documents'\n \n send_auth_request\n end",
"def index\n @documents = current_user.documents.order(\"id DESC\").all\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @documents }\n end\n end",
"def show\n @documento = Documento.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @documento }\n end\n end",
"def show\n @knowledge = Knowledge.find(params[:id])\n @documents = @knowledge.documents\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @knowledge }\n end\n end",
"def show\n @document = Document.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render jbuilder: @document }\n end\n end",
"def show\n respond_to do |format|\n format.html {\n redirect_to collection_documents_path(@collection)\n }\n format.json\n end\n end",
"def get(id)\n @documents[id.to_i]\n end",
"def index\n\n @documentable = find_resource\n @documents = @documentable.documents\n\n end",
"def show\n @document = Document.where(:id => params[:id])\n render :json => @document, :include => [:versions]\n end",
"def info\n @document = Document.find(params[:id])\n respond_to do |format|\n format.json { render :json => @document, serializer: Api::Mobile::V2::DynamicDocumentSerializer, root: 'document' }\n end\n end",
"def index\n \t@documents = Document.paginate(:page => params[:page], per_page: 2)\n \trender json: {status: 'SUCCESS', message:'Loaded documents', data:@documents}, status: :ok\n end",
"def index\n @documents = Document.all\n\n respond_to do |format|\n format.html{\n }\n format.json {\n ret = @documents.as_json\n render json:{\n status: 0,\n total: ret.count,\n result: ret\n }\n }\n end\n end",
"def index\n @documents = Document.for_user(current_user).order('created_at desc')\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @documents }\n end\n end",
"def show\n @document = Document.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.xml { render :xml => @document }\n format.json { render :json => @document }\n end\n end",
"def document\n documents.first\n end",
"def get_document( doc_id:, version:CURRENT, filename: )\n params = {}\n params[:backtrace] = @backtrace if @backtrace\n send_request :get, path_for(doc_id, filename, version), {}, :binary\n end",
"def show\n @document_page = DocumentPage.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @document_page }\n end\n end",
"def index\n if (params[:public])\n @documents = Document.where('public = ?', params[:public])\n else\n @documents = Document.all\n end\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render jbuilder: @documents }\n end\n end",
"def index\n @special_documents = ModifiedDocument.all\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @special_documents }\n end\n end",
"def show\n @document = Document.find(params[:id])\n end",
"def show\n respond_with( @document = Document.find(params[:id]) )\n end",
"def get(id)\n client.get(\n index: name,\n type: document_type.name,\n id: id\n )\n end",
"def show\n @document_type = DocumentType.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @document_type }\n end\n end",
"def show\n @special_document = ModifiedDocument.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @special_document }\n end\n end",
"def show\n\n \t\n @companydocument = Companydocument.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @companydocument }\n end\n\n end",
"def show\n @documentotipo = Documentotipo.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @documentotipo }\n end\n end",
"def show\n @modified_document = ModifiedDocument.find(params[:id])\n render json: @modified_document, status: :ok\n end",
"def document\n json = Net::HTTP.get_response URI.parse(query_string(@query))\n json.body\n end",
"def get_doc(id,*opts)\n q = \"#{database}/#{id}\"\n q << build_query_string(opts.first,\"doc\") if opts && opts.any? && opts.first.is_a?(Hash)\n\n @conn.query({url_path: q, method: :get})\n end",
"def show\n @document = Document.find(params[:id])\n \n not_found unless current_user == @document.user\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @document }\n end\n end",
"def show\n @oa_sent_document = Oa::SentDocument.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @oa_sent_document }\n end\n end",
"def show\n @paper = Paper.find(params[:id])\n @document = Document.where(paper_id: @paper.id)\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @paper }\n end\n end",
"def index\n @space = Space.where(:wiki_name => params[:space_id])[0]\n @documents = @space.documents.all\n\n respond_to do |format|\n format.html # index.html.erb\n format.xml { render :xml => @documents }\n format.json { render :json => @documents.to_json }\n end\n end",
"def document\n params.has_key?('_json') ? params['_json'] : params\n end",
"def index\n @documents = Document.all\n end",
"def index\n @documents = Document.all\n end",
"def index\n @documents = Document.all\n end",
"def index\n @documents = Document.all\n end",
"def index\n @documents = Document.all\n end",
"def index\n @documents = Document.all\n end",
"def index\n @documents = Document.all\n end",
"def index\n @documents = Document.all.delete_if { |document| cannot? :read, document }\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @documents }\n format.xml { render xml: @documents }\n end\n end",
"def show\n @documents = current_user.documents.find(params[:id])\n end",
"def show\n @tdoc = Tdoc.find(params[:id])\n \n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @tdoc }\n end\n end",
"def show\n @project = Project.find(params[:id])\n db = MongoMapper.database\n @documents = db[\"documents\"].find( { \"_id\" => { \"$in\" => @project.document_ids} } )\n\n respond_to do |format|\n format.html # show.html.erb\n format.xml { render :xml => @project }\n format.json { render :json => @project }\n end\n end",
"def new\n \n @document = Document.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @document }\n end\n \n end",
"def show\n @document_history = DocumentHistory.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @document_history }\n end\n end",
"def show\n respond_to do |format|\n format.html {\n redirect_to project_documents_path(@project)\n }\n format.json\n end\n end",
"def new\n @document = Document.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @document }\n end\n end",
"def new\n @document = Document.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @document }\n end\n end",
"def new\n @document = Document.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @document }\n end\n end",
"def new\n @document = Document.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @document }\n end\n end",
"def new\n @document = Document.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @document }\n end\n end",
"def new\n @document = Document.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @document }\n end\n end",
"def new\n @document = Document.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @document }\n end\n end",
"def new\n @document = Document.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @document }\n end\n end",
"def new\n @document = Document.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @document }\n end\n end",
"def documents(type: nil) # \"path\", \"id\", \"key\" # TESTED\n body = {\n \"collection\" => @collection,\n \"type\" => type\n }.delete_if{|k,v| v.nil?}.to_json\n request = @@request.merge({ :body => body })\n result = self.class.put(\"/_db/#{@database}/_api/simple/all-keys\", request)\n return result.headers[\"x-arango-async-id\"] if @@async == \"store\"\n return true if @@async\n result = result.parsed_response\n if type.nil?\n @@verbose ? result : result[\"error\"] ? result[\"errorMessage\"] : result[\"result\"].map{|x| value = self.class.get(x).parsed_response; ArangoDocument.new(key: value[\"_key\"], collection: @collection, body: value)}\n else\n @@verbose ? result : result[\"error\"] ? result[\"errorMessage\"] : result[\"result\"]\n end\n end",
"def new\r\n @document = Document.new\r\n\r\n respond_to do |format|\r\n format.html # new.html.erb\r\n format.json { render json: @document }\r\n end\r\n end",
"def index\n @person = Person.find((params[:person_id]))\n @person_identification_docs = PersonIdentificationDoc.where(\"person_id = ?\", params[:person_id])\n\n\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render :json => @person_identification_docs }\n end\n end",
"def show\n transform_id if /^nyu_\\d{4}_\\d{5}$/.match(params[:id])\n @response, @document = fetch_document(params[:id])\n\n respond_to do |format|\n format.html { setup_next_and_previous_documents }\n format.json { render json: { response: { document: @document } } }\n additional_export_formats(@document, format)\n end\n end",
"def index\n if !@db\n render json: [], status: 200\n else\n docs = @db.all_docs(params={\"include_docs\":\"true\"})[\"rows\"]\n names = docs.map { |d| d[\"doc\"][\"name\"] }\n render json: JSON.dump(names), status: 200\n end\n end",
"def document\n params[:document]\n end",
"def show\n respond_to do |format|\n format.html # show.html.erb\n format.json { render :json => @shared_document }\n end\n end",
"def documents; end",
"def documents; end",
"def get_document_info( doc_id: )\n params = {}\n params[:backtrace] = @backtrace if @backtrace\n send_request :get, url_for_base(doc_id), params, :json\n end",
"def document_details(guid)\n get \"/api/documents/#{guid}.xml\"\n end",
"def show\n if request.put?\n logger.debug params.inspect\n else\n respond_to do |format|\n format.html\n format.json { render :json => @document.to_obj }\n end\n end\n end",
"def show\n @document_format = DocumentFormat.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @document_format }\n end\n end",
"def show\n @squadron_document = SquadronDocument.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render :json => @squadron_document }\n end\n end",
"def index\n @docs = Oa::SentDocument.all#.paginate(:page => params[:page], :order => 'created_at desc')\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @docs }\n end\n end",
"def show\r\n @customer_document = CustomerDocument.find(params[:id])\r\n\r\n respond_to do |format|\r\n format.html # show.html.erb\r\n format.json { render json: @customer_document }\r\n end\r\n end",
"def service_document\n response = get(@url.to_s)\n response.body\n end",
"def index\n params.delete :utf8\n @documents = Document.filter(params)\n #@documents = Document.all\n\t\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @documents }\n end\n end",
"def document(document_id)\n Hancock::Request.send_get_request(\"/envelopes/#{envelope_id}/documents/#{document_id}\")\n end",
"def index\n @documents = current_user.documents.all\n end",
"def show\n @document = Document.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.xml { render :xml => @document }\n end\n end",
"def show\n @document = Document.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.xml { render :xml => @document }\n end\n end",
"def get_document(id)\n @session.getDocument id\n end",
"def index\n @documents = Document.all\n [:country_code, :subject_type, :subject_id, :category, :year].each do |key|\n @documents = @documents.where(key => params[key]) unless params[key].nil?\n end\n render :json => @documents, :include => [:versions]\n end"
] |
[
"0.7391596",
"0.73689586",
"0.73263764",
"0.73263764",
"0.73263764",
"0.73263764",
"0.73263764",
"0.73263764",
"0.73263764",
"0.73170596",
"0.7306216",
"0.7303724",
"0.7286854",
"0.72555476",
"0.71793205",
"0.71743375",
"0.71661717",
"0.714282",
"0.71288323",
"0.7051852",
"0.69956195",
"0.6973213",
"0.6970693",
"0.69446284",
"0.6934252",
"0.692459",
"0.69116575",
"0.688864",
"0.6869882",
"0.6843845",
"0.68378395",
"0.68169975",
"0.6811576",
"0.6809019",
"0.6805765",
"0.6792871",
"0.67714137",
"0.6765573",
"0.6759939",
"0.67392385",
"0.6713349",
"0.6704634",
"0.6702019",
"0.66802585",
"0.66768974",
"0.6650748",
"0.66179174",
"0.6616438",
"0.66001326",
"0.6581267",
"0.65783656",
"0.65714663",
"0.6570603",
"0.65407324",
"0.6532111",
"0.6532111",
"0.6532111",
"0.6532111",
"0.6532111",
"0.6532111",
"0.6532111",
"0.6530772",
"0.65298676",
"0.6513702",
"0.6504066",
"0.6498369",
"0.6495558",
"0.6495557",
"0.6490711",
"0.6490711",
"0.6490711",
"0.6490711",
"0.6490711",
"0.6490711",
"0.6490711",
"0.6490711",
"0.6490711",
"0.64893234",
"0.64841187",
"0.6476979",
"0.6454224",
"0.6447528",
"0.6442026",
"0.64404166",
"0.64335865",
"0.64335865",
"0.64072186",
"0.64061415",
"0.6399104",
"0.63888896",
"0.63854843",
"0.63847864",
"0.63521683",
"0.63512355",
"0.63298464",
"0.63271624",
"0.6308484",
"0.6307677",
"0.6307677",
"0.62963086",
"0.62953496"
] |
0.0
|
-1
|
POST /documents POST /documents.json
|
def create
@attachment = Attachment.new(attachment_params)
if file_data.respond_to?(:read)
@@world = file_data.read
xml_contents = file_data.read
elsif file_data.respond_to?(:path)
xml_contents = File.read(file_data.path)
else
logger.error "Bad file_data: #{file_data.class.name}: #{file_data.inspect}"
end
respond_to do |format|
if @attachment.save
format.html { redirect_to attachments_path, notice: 'Document was successfully created.' }
format.json { render action: 'show', status: :created, location: @attachment }
else
format.html { render action: 'new' }
format.json { render json: @dattachment.errors, status: :unprocessable_entity }
end
end
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def create\n @document = Document.new(params[:document])\n\n respond_to do |format|\n if @document.save\n format.html { redirect_to @document, :notice => 'Document was successfully created.' }\n format.json { render :json => @document, :status => :created, :location => @document }\n else\n format.html { render :action => \"new\" }\n format.json { render :json => @document.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def create\n @document = Document.new(document_params)\n\n if @document.save\n render json: {status: 'SUCCESS', message:'Document created', data:@document}, status: :ok\n else\n render json: {status: 'ERROR', message:'Document not created', data:@document.errors}, status: :unprocessable_entity\n end\n\n end",
"def create\n @document = Document.new(params[:document])\n\n respond_to do |format|\n if @document.save\n format.html { redirect_to @document, notice: 'Document was successfully created.' }\n format.json { render json: @document, status: :created, location: @document }\n else\n format.html { render action: \"new\" }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @document = Document.new(params[:document])\n\n respond_to do |format|\n if @document.save\n format.html { redirect_to @document, notice: 'Document was successfully created.' }\n format.json { render json: @document, status: :created, location: @document }\n else\n format.html { render action: \"new\" }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create \n @document = Document.new(document_params)\n\n respond_to do |format|\n if @document.save\n format.html { redirect_to documents_path, notice: 'Document was successfully created.' }\n format.json { render :show, status: :created, location: document_path }\n else\n format.html { render :new }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n\n end\n end",
"def create\n @document = Document.new(params[:document])\n\n respond_to do |format|\n if @document.save\n format.html { redirect_to documents_path, notice: 'Document was successfully uploaded.' }\n format.json { render json: @document, status: :created, location: @document }\n else\n format.html { render new_document_path }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @document = Document.new(document_params)\n\n if @document.save\n render :show, status: :created, location: @document\n else\n render json: @document.errors, status: :unprocessable_entity\n end\n end",
"def create\n @document = Document.new(params[:document])\n\n respond_to do |format|\n if @document.save\n format.html { redirect_to @document, notice: 'Document was successfully created.' }\n format.json { render jbuilder: @document, status: :created, location: @document }\n else\n format.html { render action: \"new\" }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @document = Document.new(document_params)\n\n respond_to do |format|\n if @document.save\n format.html { redirect_to @document, notice: 'Document was successfully created.' }\n format.json { render :show, status: :created, location: @document }\n else\n format.html { render :new }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n\n \n @document = current_user.documents.create(params[:document])\n\n respond_to do |format|\n if @document.save\n format.html {\n render :json => [@document.to_jq_upload].to_json,\n :content_type => 'text/html',\n :layout => false\n }\n format.json { render json: [@document.to_jq_upload].to_json, status: :created, location: @document }\n else\n format.html { render action: \"new\" }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @document = Document.new(document_params)\n\n respond_to do |format|\n if @document.save\n format.html{\n flash[:success] = \"Document has saved successfully!\"\n redirect_to @document\n }\n format.json {\n url = url_for @document\n render json: {\n status_code: 0,\n response:{\n url: url\n }\n }\n }\n else\n format.html{ render 'new' }\n format.json {\n render json: {\n status_code: 1,\n errors: @document.errors\n }\n }\n end\n end\n end",
"def create\n @document = current_user.documents.new(params[:document])\n\n respond_to do |format|\n if @document.save\n format.html { redirect_to :action => \"index\", notice: 'Document was successfully created.' }\n format.json { render json: @document, status: :created, location: @document }\n else\n flash[:alert] = error_messages(@document.errors)\n\n format.html { render action: \"new\" }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @document = Document.new(document_params)\n @document.user = current_user\n\n respond_to do |format|\n if @document.save\n format.html { redirect_to @document, notice: 'Document was successfully created.' }\n format.json { render :show, status: :created, location: @document }\n else\n format.html { render :new }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @document = Document.new(resource_params)\n\n respond_to do |format|\n if @document.save\n format.html { redirect_to @document, notice: 'Primary document was successfully created.' }\n format.json { render json: @document, status: :created, location: @document }\n else\n format.html { render action: \"new\" }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @has_document = HasDocument.new(has_document_params)\n \n respond_to do |format|\n if @has_document.save\n format.html { redirect_to @has_document, notice: 'Has document was successfully created.' }\n format.json { render :show, status: :created, location: @has_document }\n else\n format.html { render :new }\n format.json { render json: @has_document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @document = Document.new(params[:document].merge(:owner => current_user))\n\n respond_to do |format|\n if @document.save\n format.html { redirect_to \"/view_pad/#{@document.id}\", notice: 'Document was successfully created.' }\n format.json { render json: @document, status: :created, location: @document }\n else\n format.html { render action: \"new\" }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @document = Document.new(document_params)\n\n respond_to do |format|\n if @document.save\n format.html { redirect_to @document, success: 'Documento creado con éxito.' }\n format.json { render :show, status: :created, location: @document }\n else\n format.html { render :new }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @document = Document.new(params[:document])\n\n respond_to do |format|\n if @document.save\n format.html { redirect_to edit_admin_document_path(@document), notice: 'Document was successfully created.' }\n format.json { render json: @document, status: :created, location: @document }\n format.js { head :no_content }\n else\n format.html { render action: \"new\" }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @documentable = find_documentable\n @document = @documentable.documents.build(document_params)\n\n respond_to do |format|\n if @document.save\n format.html { redirect_to @document, notice: t('.notice', document: @document) }\n format.json\n else\n format.html { render :new }\n format.json {\n render json: @document.errors[:attachment], status: :unprocessable_entity\n }\n end\n end\n end",
"def upload\n doc = DocumentService::Document.create!({\n original_filename: params[\"original_filename\"],\n content_type: params[\"file\"].content_type,\n document: params['file'].tempfile,\n public: true\n })\n render json: {id: doc.id}, status: :created\n end",
"def create\n @document = Document.new(params[:document])\n @document.stuffing_data = []\n @document.user = current_user\n\n #Hack for now - add all column keys to primary keys for search\n @document.stuffing_primary_keys = get_data_colnames(@document.stuffing_data)\n\n respond_to do |format|\n if @document.save\n format.html { redirect_to @document, notice: 'Document was successfully created.' }\n format.json { render json: @document, status: :created, location: @document }\n else\n format.html { render action: \"new\" }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def new\n @document = Document.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @documents.map{|document| document.to_jq_upload } }\n \n end\n end",
"def create\n @modified_document = ModifiedDocument.new(modified_document_params)\n \n if @modified_document.save\n render json: @modified_document, status: :ok\n else\n render json: @modified_document.errors, status: :unprocessable_entity\n end\n end",
"def create\n authorize! :create, Document.new, @project\n main = @project.document_mains.create\n rev = main.revisions.create\n document = rev.versions.new(document_params(true))\n if document.save\n send_emails_helper(document)\n render json: document.attributes_for_edit\n else\n rev.destroy\n main.destroy\n render json: document.errors, status: :unprocessable_entity\n end\n end",
"def create\n\t \n\tdata = unpack_document(params[:document][:datafile]) \n\tdoc_params = {:title => document_params[\"title\"], :date => get_date(document_params,\"date\")}\n\tif !data.nil? then\n\t\tdoc_params[:content] = data[:content]\n\t\tdoc_params[:styles] = data[:styles]\n\tend\n @document = Document.new(doc_params)\n\n respond_to do |format|\n if @document.save\n format.html { redirect_to @document, notice: 'Document was successfully created.' }\n format.json { render action: 'show', status: :created, location: @document }\n else\n format.html { render action: 'new' }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @document = Document.new()\n @document.process params[:document][:file]\n \n respond_to do |format|\n if @document.save\n format.html { redirect_to :action => :index , notice: 'Document was successfully created.' }\n format.json { render json: @document, status: :created, location: @document }\n else\n format.html { render action: \"new\" }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @document = Document.new(document_params)\n\n respond_to do |format|\n if @document.save\n if params[:attachments]\n params[:attachments].each do |a|\n @attachment = @document.attachments.create!(:attachment => a)\n end\n end\n format.html { redirect_to @document, notice: 'Document was successfully created.' }\n format.json { render :show, status: :created, location: @document }\n else\n format.html { render :new }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @document = Document.new(params[:document])\n\n respond_to do |format|\n if @document.save\n format.html { redirect_to(@document, :notice => 'Document was successfully created.') }\n format.xml { render :xml => @document, :status => :created, :location => @document }\n else\n format.html { render :action => \"new\" }\n format.xml { render :xml => @document.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def create\n @document = @department.documents.new(document_params)\n\n respond_to do |format|\n if @document.save\n format.html { redirect_to @document, notice: 'Document was successfully created.' }\n format.json { render action: 'show', status: :created, location: @document }\n else\n format.html { render action: 'new' }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @docfile = Docfile.new(docfile_params)\n\n respond_to do |format|\n if @docfile.save\n format.html { redirect_to @docfile, notice: \"Docfile was successfully created.\" }\n format.json { render :show, status: :created, location: @docfile }\n else\n format.html { render :new, status: :unprocessable_entity }\n format.json { render json: @docfile.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @current_user = User.find_by_id(session[:user_id])\n @document = Document.new(params[:document])\n @document.user = @current_user\n if @document.file && !document.file.empty?\n @document.name = @document.file.original_filename\n @document.status = 'pending'\n #@document.dt_reference = DocTrackrEnterprise.secure_document(@document.file, \"https://doctrackr-salesforce.herokuapp.com/documents/callback\")\n end\n respond_to do |format|\n if @document.save\n format.html { redirect_to @document, notice: 'Document was successfully created.' }\n format.json { render json: @document, status: :created, location: @document }\n else\n format.html { render action: \"new\" }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @document = @course.documents.build(document_params)\n\n respond_to do |format|\n if @document.save\n format.html { redirect_to @course, notice: \"Document was successfully created.\" }\n format.json { render :show, status: :created, location: @document }\n else\n format.html { render :new, status: :unprocessable_entity }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @squadron_document = SquadronDocument.new(params[:squadron_document])\n\n respond_to do |format|\n if @squadron_document.save\n format.html { redirect_to @squadron_document, :notice => 'Squadron document was successfully created.' }\n format.json { render :json => @squadron_document, :status => :created, :location => @squadron_document }\n else\n format.html { render :action => \"new\" }\n format.json { render :json => @squadron_document.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def create\n @document_file = DocumentFile.new(document_file_params)\n @document_file.document = @document\n respond_to do |format|\n if @document_file.save\n format.html { redirect_to edit_document_path(@document), notice: 'Document file was successfully created.' }\n format.json { render :show, status: :created, location: @document_file }\n else\n format.html { render :new }\n format.json { render json: @document_file.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n if signed_in?\n if current_user.admin?\n @document = Document.new(document_params)\n @document.user = current_user\n respond_to do |format|\n if @document.save\n format.html { redirect_to @document, notice: I18n.t('documents.messages.create_success') }\n format.json { render :show, status: :created, location: @document }\n else\n format.html { render :new }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n else\n admin_user\n end\n else\n redirect_to signin_path\n end\n end",
"def create\n @documents_person = DocumentsPerson.new(documents_person_params)\n\n respond_to do |format|\n if @documents_person.save\n format.html { redirect_to @documents_person, notice: 'Documents person was successfully created.' }\n format.json { render :show, status: :created, location: @documents_person }\n else\n format.html { render :new }\n format.json { render json: @documents_person.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @document = current_user.documents.build(document_params)\n get_identification_type\n respond_to do |format|\n if @document.save\n format.html { redirect_to @document, notice: 'Document was successfully created.' }\n format.json { render :show, status: :created, location: @document }\n else\n format.html { render :new }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def index\n @documents = Document.all\n @document = Document.new\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @documents }\n end\n end",
"def create\n @documentotipo = Documentotipo.new(params[:documentotipo])\n\n respond_to do |format|\n if @documentotipo.save\n format.html { redirect_to @documentotipo, notice: 'Documentotipo was successfully created.' }\n format.json { render json: @documentotipo, status: :created, location: @documentotipo }\n else\n format.html { render action: \"new\" }\n format.json { render json: @documentotipo.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n # puts \"*|*|\"*20\n # puts params[:document][:data].original_filename\n # puts params[:document][:data].content_type\n # puts document_params[:data].original_filename\n # puts document_params[:data].content_type\n # puts \"*^*^\"*20\n @document = Document.new\n @document.name = document_params[:name]\n @document.lab_id = document_params[:lab_id]\n @document.data = document_params[:data].read\n @document.filename = sanitize_filename(document_params[:data].original_filename)\n @document.content_type = document_params[:data].content_type\n\n\n respond_to do |format|\n if @document.save\n format.html { redirect_to documents_url, notice: 'Document was successfully created.' }\n format.json { render :show, status: :created, location: documents_url }\n else\n format.html { render :new }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @shared_document = SharedDocument.new(shared_document_params)\n\n respond_to do |format|\n if @shared_document.save\n # format.html { redirect_to @shared_document, notice: 'Shared document was successfully created.' }\n format.json { render json: @shared_document.to_json({}) }\n else\n format.html { render action: 'new' }\n format.json { render json: @shared_document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def add_document(birthdate:, first_name:, last_name:, street:, postal_code:, country_code:, document_type:, document_value:)\n data = {\n doc: {\n birth_day: birthdate.day,\n birth_month: birthdate.month,\n birth_year: birthdate.year,\n name_first: first_name,\n name_last: last_name,\n address_street1: street,\n address_postal_code: postal_code,\n address_country_code: country_code,\n document_type: document_type,\n document_value: document_value\n }\n }\n\n @client.patch(path: \"/users/#{@user_id}\", oauth_key: @oauth_key, fingerprint: @fingerprint, json: data)\n end",
"def create\r\n @customer_document = CustomerDocument.new(params[:customer_document])\r\n\r\n respond_to do |format|\r\n if @customer_document.save\r\n format.html { redirect_to @customer_document, notice: 'Customer document was successfully created.' }\r\n format.json { render json: @customer_document, status: :created, location: @customer_document }\r\n else\r\n format.html { render action: \"new\" }\r\n format.json { render json: @customer_document.errors, status: :unprocessable_entity }\r\n end\r\n end\r\n end",
"def create\n @order_document = OrderDocument.new(order_document_params)\n\n respond_to do |format|\n if @order_document.save\n format.html { redirect_to @order_document, notice: 'Order document was successfully created.' }\n format.json { render :show, status: :created, location: @order_document }\n else\n format.html { render :new }\n format.json { render json: @order_document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @document = Document.new(document_params)\n\n authorize @document\n\n respond_to do |format|\n if @document.save\n flash_label = \"Document was successfully created.\"\n flash.now[:notice] = flash_label\n\n format.html { redirect_to @document, notice: flash_label }\n format.json { render :show, status: :created, location: @document }\n format.js\n else\n format.html { render :new }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n format.js\n end\n end\n end",
"def new\n \n @document = Document.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @document }\n end\n \n end",
"def create\n @document = @instruction.documents.build(document_params)\n authorize @document\n disable_primary if @document.primary\n respond_to do |format|\n if @document.save\n format.html { redirect_to @instruction, notice: t('documents.create.success') }\n format.json { render :show, status: :created, location: @document }\n else\n format.html { render :new }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @document = Document.new(params[:document])\n\n respond_to do |format|\n if @document.save\n flash[:notice] = 'Document was successfully created.'\n #format.html { redirect_to(@document) }\n format.xml { render :xml => @document, :status => :created, :location => @document }\n else\n format.html { render :action => \"new\" }\n format.xml { render :xml => @document.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def add_documents\n\t\tif @current_user.present?\n\t\t\t@property = Property.find(params[:property_id])\n\t\t\tif @property.present?\n\t\t\t\tif @property.documents.present?\n\t\t\t\t\t@property.documents.destroy_all\n\t\t\t\tend\n\t\t\t\tparams[:files].each { |file|\n\t i = @property.documents.create(file: file)\n\t if i.save\n\t else\n\t \trender_json({\"status\" => \"Fail\", \"message\" => i.errors.full_messages.first}.to_json)\n\t \treturn\n\t end\n\t }\n\t render :file => 'api/v1/property/add_document'\n\t #render_json({\"status\" => \"Success\", \"message\" => \"Documents has been saved successfully.\"}.to_json)\n\t else\n\t \trender_json({\"status\" => \"Fail\", \"message\" => \"No property found.\"}.to_json)\n\t end\n\t\tend\n\tend",
"def document_params\n params.require(:document).permit(:title, :date, :content)\n end",
"def document_params\n params.require(:document).permit(:title, :doc)\n end",
"def create\n @document_page = DocumentPage.new(params[:document_page])\n\n respond_to do |format|\n if @document_page.save\n format.html { redirect_to @document_page, notice: 'Document page was successfully created.' }\n format.json { render json: @document_page, status: :created, location: @document_page }\n else\n format.html { render action: \"new\" }\n format.json { render json: @document_page.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n\n \t\n @companydocument = Companydocument.new(params[:companydocument])\n\n respond_to do |format|\n if @companydocument.save\n format.html { redirect_to @companydocument, notice: 'Companydocument was successfully created.' }\n format.json { render json: @companydocument, status: :created, location: @companydocument }\n else\n format.html { render action: \"new\" }\n format.json { render json: @companydocument.errors, status: :unprocessable_entity }\n end\n end\n\n end",
"def create\n @document = Document.new(document_params)\n respond_to do |format|\n ap tame_name = params[:document][:team]\n if @document.save\n format.html { redirect_to documents_url(:team_id => @document.team_id, :team => tame_name), notice: 'Document was successfully created.' }\n format.json { render :show, status: :created, location: @document }\n else\n format.html { redirect_to new_document_path(:team_id => @document.team_id), notice: 'Document not created.' }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n if(@document = Document.new(params[:document])).save\n flash['notice'] = 'Document was successfully created.'\n respond_with @document, \n :location => site_document_url(@document.site.slug , @document.id.to_s)\n else\n\n if doc = Document.where(uri: params[:document][:uri]).first\n params[:id] = doc.id\n update\n else\n respond_with @document\n end\n end\n \n end",
"def create\n @documentos_simposio = DocumentosSimposio.new(documentos_simposio_params)\n\n respond_to do |format|\n if @documentos_simposio.save\n format.html { redirect_to @documentos_simposio, notice: 'Documentos simposio was successfully created.' }\n format.json { render :show, status: :created, location: @documentos_simposio }\n else\n format.html { render :new }\n format.json { render json: @documentos_simposio.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @document = Document.new(document_params)\n\n respond_to do |format|\n if @document.save\n format.html { redirect_to event_documents_path(@document.event.id), notice: 'Document was successfully created .' }\n format.json { render action: 'show', status: :created, location: @document }\n format.js\n else\n format.html { render action: 'new' }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @special_document = ModifiedDocument.new(params[:special_document])\n\n respond_to do |format|\n if @special_document.save\n format.html { redirect_to @special_document, notice: 'Special document was successfully created.' }\n format.json { render json: @special_document, status: :created, location: @special_document }\n else\n format.html { render action: \"new\" }\n format.json { render json: @special_document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def indexDocument(index, type, document, id)\n self.class.post(\"#{index}/#{type}/#{id}\", :body => document.to_json)\n end",
"def create\n @client_doc = ClientDoc.new(client_doc_params)\n\n respond_to do |format|\n if @client_doc.save\n format.html { redirect_to @client_doc, notice: 'Client doc was successfully created.' }\n format.json { render :show, status: :created, location: @client_doc }\n else\n format.html { render :new }\n format.json { render json: @client_doc.errors, status: :unprocessable_entity }\n end\n end\n end",
"def add_document(opts={})\n raise ArgumentError, \":title is required\" if opts[:title].nil?\n raise ArgumentError, \":body is required\" if opts[:body].nil?\n raise ArgumentError, \":url is required\" if opts[:url].nil?\n \n if opts[:tags].is_a?(Array)\n opts[:tags] = opts[:tags].join(\",\")\n end\n \n doc = opts.to_json\n \n Srchio::Response.new(self.class.post(\"/searchers/#{searcher_id}/documents\", :body => doc))\n end",
"def create\n @document = Document.new(params[:document])\n\n respond_to do |wants|\n if @document.save\n flash[:notice] = 'Document was successfully created.'\n wants.html { redirect_to(@document) }\n wants.xml { render :xml => @document, :status => :created, :location => @document }\n else\n wants.html { render :action => \"new\" }\n wants.xml { render :xml => @document.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def create\n @department_document = DepartmentDocument.new(department_document_params)\n\n respond_to do |format|\n if @department_document.save\n format.html { redirect_to @department_document, notice: 'Department document was successfully created.' }\n format.json { render :show, status: :created, location: @department_document }\n else\n format.html { render :new }\n format.json { render json: @department_document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def document_params\n params.fetch(:document).permit(:title, :name, :content)\n end",
"def new\n @paper = Paper.new\n @document = @paper.documents.new\n\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @paper }\n end\n end",
"def create\n \n @doc = Doc.new(params[:doc])\n\n respond_to do |format|\n if @doc.save\n save_object_relationship\n format.html { redirect_to(@doc, :notice => 'Doc was successfully created.') }\n format.xml { render :xml => @doc, :status => :created, :location => @doc }\n else\n format.html { render :action => \"new\" }\n format.xml { render :xml => @doc.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def save doc\n if doc['_attachments']\n doc['_attachments'] = encode_attachments(doc['_attachments'])\n end\n if doc['_id']\n slug = CGI.escape(doc['_id'])\n CouchRest.put \"#{@root}/#{slug}\", doc\n else\n CouchRest.post \"#{@root}\", doc\n end\n end",
"def create\n @doc = Doc.new(params[:doc])\n\n respond_to do |format|\n if @doc.save\n format.html { redirect_to(@doc, :notice => 'Doc was successfully created.') }\n format.xml { render :xml => @doc, :status => :created, :location => @doc }\n else\n format.html { render :action => \"new\" }\n format.xml { render :xml => @doc.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def document_params\n params.require(:document).permit(:title, :body)\n end",
"def create\n @documentation = current_user.documentations.build(documentation_params)\n\n respond_to do |format|\n if @documentation.save\n format.html { redirect_to @documentation, notice: 'Documentation was successfully created.' }\n format.json { render :show, status: :created, location: @documentation }\n else\n format.html { render :new }\n format.json { render json: @documentation.errors, status: :unprocessable_entity }\n end\n end\n end",
"def save document\n ensure_connection!\n resp = connection.create_doc index_id, document.to_hash\n if resp.success?\n raw = document.instance_variable_get \"@raw\"\n raw.merge! JSON.parse(resp.body)\n return document\n end\n fail ApiError.from_response(resp)\n rescue JSON::ParserError\n raise ApiError.from_response(resp)\n end",
"def new\r\n @document = Document.new\r\n\r\n respond_to do |format|\r\n format.html # new.html.erb\r\n format.json { render json: @document }\r\n end\r\n end",
"def create\n @doc = Doc.new(params[:doc])\n\n respond_to do |format|\n if @doc.save\n format.html { redirect_to @doc, notice: \"\\\"#{@doc.component}\\\" was successfully created.\" }\n format.json { render json: @doc, status: :created, location: @doc }\n else\n format.html { render action: \"new\" }\n format.json { render json: @doc.errors, status: :unprocessable_entity }\n end\n end\n end",
"def post_document_save_as(request)\n data, _status_code, _headers = post_document_save_as_with_http_info(request)\n request_token if _status_code == 401\n data\n end",
"def add(document)\n @rest.post(\"/#{@db_name}\", [document])\n end",
"def create\n @document = Document.new(params[:document])\n @document.generate_token\n @document.archived = false\n if params[:group_id]\n @document.owner = Group.find(params[:group_id]) \n else\n @document.owner ||= default_company\n end\n @document.author = current_user\n\n respond_to do |format|\n if @document.save\n format.html { redirect_to document_path(@document), :flash => { :success => 'Document was successfully created.'} }\n format.json { render json: @document, status: :created, location: @document }\n else\n format.html { render action: \"new\" }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def document_params\n params.require(:document).permit(:name, :document_type, :document_url, :intake_code)\n end",
"def new\n @document = Document.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @document }\n end\n end",
"def new\n @document = Document.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @document }\n end\n end",
"def new\n @document = Document.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @document }\n end\n end",
"def new\n @document = Document.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @document }\n end\n end",
"def new\n @document = Document.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @document }\n end\n end",
"def new\n @document = Document.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @document }\n end\n end",
"def new\n @document = Document.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @document }\n end\n end",
"def new\n @document = Document.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @document }\n end\n end",
"def new\n @document = Document.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @document }\n end\n end",
"def index_documents\n @params = {}\n @action = 'index_documents'\n \n send_auth_request\n end",
"def document_params\n params.require(:document).permit(:name, :collection_id, :file_dir)\n end",
"def create_document_using_post_with_http_info(document_request, opts = {})\n if @api_client.config.debugging\n @api_client.config.logger.debug 'Calling API: DocumentApi.create_document_using_post ...'\n end\n # verify the required parameter 'document_request' is set\n if @api_client.config.client_side_validation && document_request.nil?\n fail ArgumentError, \"Missing the required parameter 'document_request' when calling DocumentApi.create_document_using_post\"\n end\n # resource path\n local_var_path = '/nucleus/v1/document'\n\n # query parameters\n query_params = {}\n\n # header parameters\n header_params = {}\n # HTTP header 'Accept' (if needed)\n header_params['Accept'] = @api_client.select_header_accept(['*/*'])\n # HTTP header 'Content-Type'\n header_params['Content-Type'] = @api_client.select_header_content_type(['application/json'])\n\n # form parameters\n form_params = {}\n\n # http body (model)\n post_body = @api_client.object_to_http_body(document_request)\n auth_names = ['oauth2']\n data, status_code, headers = @api_client.call_api(:POST, local_var_path,\n :header_params => header_params,\n :query_params => query_params,\n :form_params => form_params,\n :body => post_body,\n :auth_names => auth_names,\n :return_type => 'Document')\n if @api_client.config.debugging\n @api_client.config.logger.debug \"API called: DocumentApi#create_document_using_post\\nData: #{data.inspect}\\nStatus code: #{status_code}\\nHeaders: #{headers}\"\n end\n return data, status_code, headers\n end",
"def document_params\n params.require(:document).permit!\n end",
"def documents\n return bad_request unless params[:id] and request.format.json? || request.format.js? || request.format.text?\n return not_found unless current_document\n opts = {:access => true, :sections => true, :annotations => true, :data => true}\n if current_account\n opts[:account] = current_account\n opts[:allowed_to_edit] = current_account.allowed_to_edit?(current_document)\n opts[:allowed_to_review] = current_account.reviews?(current_document)\n end\n @response = {'document' => current_document.canonical(opts)}\n respond_to do |format|\n format.text do\n direct = [PRIVATE, ORGANIZATION, EXCLUSIVE].include? current_document.access\n redirect_to(current_document.full_text_url(direct: direct))\n end\n format.json { render_cross_origin_json }\n format.js { render_cross_origin_json }\n end\n end",
"def new\n #@current_user = User.find(session[:user_id])\n #@document = @current_user.documents.new\n @document = Document.new\n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @document }\n end\n end",
"def create\n @document = @memo.new_document(params[:document])\n\n respond_to do |format|\n if @document.save\n format.html { redirect_to memo_url(@memo), notice: 'Memo was successfully created.' }\n format.json { render json: @document, status: :created }\n else\n format.html { render memo_url(@memo) }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @document_type = DocumentType.new(params[:document_type])\n\n respond_to do |format|\n if @document_type.save\n format.html { redirect_to @document_type, notice: 'Document type was successfully created.' }\n format.json { render json: @document_type, status: :created, location: @document_type }\n else\n format.html { render action: \"new\" }\n format.json { render json: @document_type.errors, status: :unprocessable_entity }\n end\n end\n end",
"def document_params\n params.require(:document).permit(:name, :data, :lab_id)\n end",
"def create\n @documento = Documento.new(params[:documento])\n\n respond_to do |format|\n if @documento.save\n format.html { \n\t\tredirect_to @documento , notice: 'Documento creado con éxito.', flash: {estado: 0}\n }\n format.json { render json: @documento, status: :created, location: @documento }\n else\n format.html { render action: \"new\" }\n format.json { render json: @documento.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @tipo_documento = TipoDocumento.new(params[:tipo_documento])\n\n respond_to do |format|\n if @tipo_documento.save\n format.html { redirect_to @tipo_documento, notice: 'Tipo documento was successfully created.' }\n format.json { render json: @tipo_documento, status: :created, location: @tipo_documento }\n else\n format.html { render action: \"new\" }\n format.json { render json: @tipo_documento.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create_documents(type_identifier, events); end",
"def create\n @cdocument = Cdocument.new(cdocument_params)\n\n respond_to do |format|\n if @cdocument.save\n format.html { redirect_to @cdocument, notice: 'Cdocument was successfully created.' }\n format.json { render :show, status: :created, location: @cdocument }\n else\n format.html { render :new }\n format.json { render json: @cdocument.errors, status: :unprocessable_entity }\n end\n end\n end",
"def create\n @document = Document.new(params[:document])\n\n respond_to do |format|\n if @document.save\n format.html { redirect_to(@document, :notice => 'Project was successfully created.') }\n format.xml { render :xml => @document, :status => :created, :location => @document }\n else\n format.html { render :action => \"new\" }\n format.xml { render :xml => @document.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def document_params\n params.require(:document).permit(:title, :description, :url)\n end"
] |
[
"0.72675",
"0.72270125",
"0.7202067",
"0.7202067",
"0.71927583",
"0.71516854",
"0.7137915",
"0.71379024",
"0.712435",
"0.7049302",
"0.68928266",
"0.6876889",
"0.6837602",
"0.6818319",
"0.68161947",
"0.6751123",
"0.67454964",
"0.6734003",
"0.6723543",
"0.66879636",
"0.6685973",
"0.668132",
"0.6680458",
"0.6663098",
"0.6660174",
"0.66162896",
"0.65943253",
"0.6590682",
"0.6570422",
"0.65512085",
"0.6546826",
"0.6535173",
"0.6533918",
"0.65321255",
"0.652205",
"0.6515982",
"0.6513196",
"0.6507297",
"0.6495766",
"0.64839077",
"0.6479422",
"0.6472698",
"0.6470567",
"0.64556056",
"0.6440168",
"0.643289",
"0.64205045",
"0.6417518",
"0.6401603",
"0.63947314",
"0.639158",
"0.63845897",
"0.63733524",
"0.63690007",
"0.63634974",
"0.636292",
"0.6350792",
"0.6344701",
"0.63445014",
"0.6342723",
"0.6342058",
"0.63352257",
"0.63282084",
"0.63280314",
"0.6328017",
"0.6327471",
"0.63192314",
"0.6310641",
"0.63057226",
"0.62854344",
"0.6283745",
"0.62812275",
"0.6278156",
"0.6276319",
"0.6274944",
"0.6261227",
"0.6260864",
"0.62410796",
"0.62410796",
"0.62410796",
"0.62410796",
"0.62410796",
"0.62410796",
"0.62410796",
"0.62410796",
"0.62410796",
"0.6240315",
"0.6235294",
"0.62191933",
"0.62174124",
"0.6215783",
"0.6214476",
"0.6210557",
"0.6207363",
"0.6201699",
"0.6197934",
"0.61968803",
"0.6196743",
"0.61902386",
"0.61808026",
"0.61764014"
] |
0.0
|
-1
|
PATCH/PUT /documents/1 PATCH/PUT /documents/1.json
|
def update
respond_to do |format|
if @attachment.update(attachment_params)
format.html { redirect_to @attachment, notice: 'Document was successfully updated.' }
format.json { head :no_content }
else
format.html { render action: 'edit' }
format.json { render json: @attachment.errors, status: :unprocessable_entity }
end
end
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def update\n document = Document.find(params[:id])\n document.update!(update_params)\n render json: {}\n end",
"def update\n document = Document.find(params[:id])\n if document.update(params_document)\n render json: document, status: 200\n else\n render json: document.errors, status: 422\n end\n\n end",
"def update\n @document = Document.find(params[:id])\n respond_to do |format|\n if @document.update_attributes(params[:document])\n format.json { render :json => nil, :status => :ok }\n format.html { redirect_to(@document, :notice => 'Document was successfully updated.') }\n else\n format.json { render :json => @document.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def update\n @document = Document.find(params[:id])\n\n respond_to do |format|\n if @document.update_attributes(params[:document])\n format.html { redirect_to @document, notice: 'Document was successfully updated.' }\n # format.json { head :no_content }\n format.json { render jbuilder: @document }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @document = Document.find(params[:id])\n\n respond_to do |format|\n if @document.update_attributes(params[:document])\n format.html { redirect_to documents_path, notice: 'Document was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render :edit }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @document = Document.find(params[:id])\n\n respond_to do |format|\n if @document.update_attributes(params[:document])\n format.html { redirect_to @document, :notice => 'Document was successfully updated.' }\n format.json { head :ok }\n else\n format.html { render :action => \"edit\" }\n format.json { render :json => @document.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def update\n @document = Document.find(params[:id])\n\n respond_to do |format|\n if @document.update_attributes(resource_params)\n format.html { redirect_to @document, notice: 'Primary document was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @document = Document.find(params[:id])\n\n respond_to do |format|\n if @document.update_attributes(params[:document])\n format.html { redirect_to @document, notice: 'Document was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @document = Document.find(params[:id])\n\n respond_to do |format|\n if @document.update_attributes(params[:document])\n format.html { redirect_to @document, notice: 'Document was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @document = Document.find(params[:id])\n\n respond_to do |format|\n if @document.update_attributes(params[:document])\n format.html { redirect_to @document, notice: 'Document was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @document = Document.find(params[:id])\n\n respond_to do |format|\n if @document.update_attributes(params[:document])\n format.html { redirect_to root_url , notice: 'Document was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @document = Document.find(params[:id])\n\n respond_to do |format|\n if @document.update_attributes(params[:document])\n format.html { redirect_to @document, notice: 'Document was successfully updated.' }\n format.json { head :ok }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n \t@document = Document.find(params[:id])\n if @document.update_attributes(document_params)\n render json: {status: 'SUCCESS', message:'Document updated', data:@document}, status: :ok\n else\n\t\trender json: {status: 'ERROR', message:'Document not updated', data:@document.errors}, status: :unprocessable_entity\n end\n end",
"def update_document index, id, document\n uri = URI(\"http://#{@host}:#{@port_s}/#{index}/_doc/#{id}/_update\")\n req = Net::HTTP::Post.new(uri)\n req.body = { \"doc\": document }.to_json\n run(uri, req)\n end",
"def update\n @document = Document.find(params[:id])\n\n respond_to do |format|\n if @document.update_attributes(params[:document])\n format.html { redirect_to :action => \"index\", notice: 'Document was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update!(**args)\n @documents = args[:documents] if args.key?(:documents)\n end",
"def update\n @document = Document.find(params[:id])\n\n respond_to do |format|\n if @document.update_attributes(params[:document])\n format.html { redirect_to @document, :flash => { :success => 'Document was successfully updated.' } }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @document.update(document_params)\n format.html { redirect_to @document, notice: 'Document was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: 'edit' }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @document.update(document_params)\n format.html { redirect_to @document, notice: 'Document was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: 'edit' }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @document.update(document_params)\n format.html { redirect_to @document, notice: 'Document was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: 'edit' }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n Rails.logger.info \"Before update\"\n respond_to do |format|\n if @document.update(document_params)\n format.html { redirect_to @document, notice: 'Document was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: 'edit' }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n Rails.logger.info \"After update\"\n end",
"def update\n @modified_document = ModifiedDocument.find(params[:id])\n \n if @modified_document.update(modified_document_params)\n render json: @modified_document, status: :ok \n else\n render json: @modified_document.errors, status: :unprocessable_entity\n end\n end",
"def update\n\n @documentable = @document.documentable\n\n respond_to do |format|\n if @document.update(document_params)\n notify_user(:notice, 'Document was successfully updated.')\n format.html { redirect_to get_resource_url(@documentable) }\n format.json { head :no_content }\n else\n format.html { render action: 'edit' }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n\t\traise RuntimeError, \"Not authorized\" unless current_user && current_user.root? == true\n\n\t\tparams = doc_params\n\t\tparams[:body].gsub!(/\\r\\n/, \"\\n\")\n\t\t@doc = Doc.find(params[:id])\n\n\t\trespond_to do |format|\n\t\t\tif @doc.update(params)\n\t\t\t\tformat.html { redirect_to @doc, notice: t('controllers.shared.successfully_updated', :model => t('activerecord.models.doc')) }\n\t\t\t\tformat.json { head :no_content }\n\t\t\telse\n\t\t\t\tformat.html { render action: \"edit\" }\n\t\t\t\tformat.json { render json: @doc.errors, status: :unprocessable_entity }\n\t\t\tend\n\t\tend\n\tend",
"def update\n respond_to do |format|\n if @document.update_attributes(name: document_params[:name], lab_id: document_params[:lab_id])\n format.html { redirect_to documents_url, notice: 'Document was successfully updated.' }\n format.json { render :show, status: :ok, location: documents_url }\n else\n format.html { render :edit }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n \n @document = Document.find(params[:id])\n\n respond_to do |format|\n if @document.update_attributes(params[:document])\n format.html { redirect_to edit_document_path(@document), notice: 'Documento gardado correctamente' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n \n end",
"def update\n respond_to do |format|\n if @document.update(document_params)\n format.html { redirect_to documents_path, notice: 'Document was successfully updated.' }\n format.json { render :show, status: :ok, location: documents_path }\n else\n format.html { render :edit }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @document.update(document_params)\n format.html { redirect_to @document, notice: 'Document was successfully updated.' }\n format.json {\n render json: {\n status_code: 0\n }\n }\n else\n format.html { render action: 'edit' }\n format.json {\n render json: {\n status_code: 1,\n errors: @document.errors\n }\n }\n end\n end\n end",
"def update\n\tdata = unpack_document(params[:document][:datafile]) \n\tdoc_params = {:title => document_params[\"title\"], :date => get_date(document_params,\"date\")}\n\tif !data.nil? then\n\t\tdoc_params[:content] = data[:content]\n\t\tdoc_params[:styles] = data[:styles]\n\tend\n respond_to do |format|\n if @document.update(doc_params)\n format.html { redirect_to @document, notice: \"Document updated successfully\"}\n format.json { head :no_content }\n else\n format.html { render action: 'edit' }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n authorize @document\n disable_primary if document_params[:primary]\n respond_to do |format|\n if @document.update(document_params)\n format.html { redirect_to @document.instruction, notice: 'Document was successfully updated.' }\n format.json { render :show, status: :ok, location: @document }\n else\n format.html { render :edit }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @document.update(document_params)\n format.html { redirect_to documents_url, notice: 'Document was successfully updated.' }\n format.json { render :show, status: :ok, location: @document }\n else\n format.html { render :edit }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @document.update(document_params)\n format.html { redirect_to @document, notice: 'Document was successfully updated.' }\n format.json { render :show, status: :ok, location: @document }\n else\n format.html { render :edit }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @document.update(document_params)\n format.html { redirect_to @document, notice: 'Document was successfully updated.' }\n format.json { render :show, status: :ok, location: @document }\n else\n format.html { render :edit }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @document.update(document_params)\n format.html { redirect_to @document, notice: 'Document was successfully updated.' }\n format.json { render :show, status: :ok, location: @document }\n else\n format.html { render :edit }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @document.update(document_params)\n format.html { redirect_to @document, notice: 'Document was successfully updated.' }\n format.json { render :show, status: :ok, location: @document }\n else\n format.html { render :edit }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @document.update(document_params)\n format.html { redirect_to @document, notice: 'Document was successfully updated.' }\n format.json { render :show, status: :ok, location: @document }\n else\n format.html { render :edit }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @document.update(document_params)\n format.html { redirect_to @document, notice: 'Document was successfully updated.' }\n format.json { render :show, status: :ok, location: @document }\n else\n format.html { render :edit }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @document.update(document_params)\n format.html { redirect_to @document, notice: 'Document was successfully updated.' }\n format.json { render :show, status: :ok, location: @document }\n else\n format.html { render :edit }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @document.update(document_params)\n format.html { redirect_to @document, notice: 'Document was successfully updated.' }\n format.json { render :show, status: :ok, location: @document }\n else\n format.html { render :edit }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\r\n\r\n respond_to do |format|\r\n if @document.update(document_params)\r\n format.html { redirect_to root_url, notice: 'Document was successfully updated.' }\r\n format.json { render :show, status: :ok, location: @document }\r\n else\r\n format.html { render :edit }\r\n format.json { render json: @document.errors, status: :unprocessable_entity }\r\n end\r\n end\r\n end",
"def update\n @document = Document.find(params[:id])\n\n respond_to do |format|\n if @document.update_attributes(params[:document])\n format.html { redirect_to edit_admin_document_path(@document), notice: 'Document was successfully updated.' }\n format.js { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @collection = @document.collection\n respond_to do |format|\n if @document.update(document_params)\n format.html { redirect_back fallback_location: collection_documents_path(@collection), notice: 'The document was successfully updated.' }\n format.json { render :show, status: :ok, location: @document }\n else\n format.html { render :edit }\n format.json { render json: {error: @document.errors}, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @document.update(document_params)\n format.html { redirect_to @document, notice: \"Document was successfully updated.\" }\n format.json { render :show, status: :ok, location: @document }\n else\n format.html { render :edit, status: :unprocessable_entity }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def patch!\n request! :patch\n end",
"def update\r\n params[:document][:version] = ENV[\"VERSION\"]\r\n params[:document][:username] = current_user.username\r\n @document = Document.find(params[:id])\r\n\r\n respond_to do |format|\r\n if @document.update_attributes(document_params)\r\n format.html { redirect_to @document, notice: 'Document was successfully updated.' }\r\n format.json { head :no_content }\r\n else\r\n format.html { render action: \"edit\" }\r\n format.json { render json: @document.errors, status: :unprocessable_entity }\r\n end\r\n end\r\n end",
"def update\n respond_to do |format|\n if @document.update(document_params)\n format.html { redirect_to @document, notice: 'Document was successfully updated.' }\n format.json { render :show, status: :ok, location: @document }\n else\n format.html { render :edit }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n document = @document.revision.versions.new(document_params(true))\n if document.save\n send_emails_helper(document)\n render json: document.attributes_for_edit\n else\n render json: document.errors, status: :unprocessable_entity\n end\n end",
"def update\n add_breadcrumb @document.name, library_category_library_document_path(@category, @document)\n add_breadcrumb \"Modify\"\n\n respond_to do |format|\n if @document.update(document_params)\n notify_user(:notice, 'Document was successfully updated.')\n format.html { redirect_to library_category_library_document_path(@category, @document) }\n format.json { head :no_content }\n else\n format.html { render action: 'edit' }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @document = current_user.documents.find(params[:id])\n respond_to do |format|\n if @document.update(document_params)\n format.html { redirect_to @document, notice: 'Document was successfully updated.' }\n format.json { render :show, status: :ok, location: @document }\n else\n format.html { render :edit }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @special_document = ModifiedDocument.find(params[:id])\n\n respond_to do |format|\n if @special_document.update_attributes(params[:special_document])\n format.html { redirect_to @special_document, notice: 'Special document was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @special_document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @document = Document.find(params[:id])\n\n respond_to do |format|\n if @document.update_attributes(params[:document])\n format.html { redirect_to(@document, :notice => 'Document was successfully updated.') }\n format.xml { head :ok }\n else\n format.html { render :action => \"edit\" }\n format.xml { render :xml => @document.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def update!(**args)\n @document = args[:document] if args.key?(:document)\n end",
"def update!(**args)\n @document = args[:document] if args.key?(:document)\n end",
"def update\n respond_to do |format|\n if @doc_file.update(doc_file_params)\n format.html { redirect_to @doc_file, notice: 'Doc file was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: 'edit' }\n format.json { render json: @doc_file.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @has_document.update(has_document_params)\n format.html { redirect_to @has_document, notice: 'Has document was successfully updated.' }\n format.json { render :show, status: :ok, location: @has_document }\n else\n format.html { render :edit }\n format.json { render json: @has_document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n if params[:resource][:document].present?\n @resource.document.purge\n @resource.document.attach(params[:resource][:document])\n end\n if params[:resource][:sample].present?\n @resource.sample.purge\n @resource.sample.attach(params[:resource][:sample])\n end\n respond_to do |format|\n if @resource.update(resource_params)\n format.html { redirect_to @resource, notice: 'Resource was successfully updated.' }\n format.json { render :show, status: :ok, location: @resource }\n else\n format.html { render :edit }\n format.json { render json: @resource.errors, status: :unprocessable_entity }\n end\n end\n end",
"def put_document index, id, document\n uri = URI(\"http://#{@host}:#{@port_s}/#{index}/_doc/#{id}\")\n req = Net::HTTP::Put.new(uri)\n req.body = document.to_json\n run(uri, req)\n end",
"def patch\n headers = {\"If-Match\" => @version}\n response = @context.request :patch, \"#{@path}/#{@id}\", @data.to_json, headers\n @version += 1\n response\n # 'X-HTTP-Method-Override' => 'PATCH'\n end",
"def update\n @document = Document.find(params[:id])\n\n respond_to do |format|\n if @document.update_attributes(params[:document].merge({:updated_by => current_user}))\n format.html { redirect_to edit_admin_document_path(@document), notice: '<strong>Success!</strong> The document was successfully updated.' }\n format.json { head :ok }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @client_doc.update(client_doc_params)\n format.html { redirect_to @client_doc, notice: 'Client doc was successfully updated.' }\n format.json { render :show, status: :ok, location: @client_doc }\n else\n format.html { render :edit }\n format.json { render json: @client_doc.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update!(**args)\n @cloud_ai_document_option = args[:cloud_ai_document_option] if args.key?(:cloud_ai_document_option)\n @document = args[:document] if args.key?(:document)\n @request_metadata = args[:request_metadata] if args.key?(:request_metadata)\n @update_options = args[:update_options] if args.key?(:update_options)\n end",
"def update\n @documento = Documento.find(params[:id])\n\n respond_to do |format|\n if @documento.update_attributes(params[:documento])\n format.html { redirect_to @documento, notice: 'Documento was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @documento.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @document = Document.find(:first, :conditions => [\"id = ?\", params[:id]])\n\n respond_to do |format|\n if @document.update_attributes(params[:document])\n flash[:notice] = 'Document was successfully updated.'\n format.html { redirect_to(@document) }\n format.xml { head :ok }\n else\n format.html { render :action => \"edit\" }\n format.xml { render :xml => @document.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @document.update_attributes(params[:document])\n save_log({ :action => 'EDIT', :document => { :before => @old_document, :after => @document}})\n flash[:notice] = 'Document was successfully updated.'\n format.html { redirect_to(@document) }\n format.xml { head :ok }\n else\n format.html { render :action => \"edit\" }\n format.xml { render :xml => @document.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def update!(**args)\n @docs = args[:docs] if args.key?(:docs)\n end",
"def api_patch(path, data = {})\n api_request(:patch, path, :data => data)\n end",
"def update\n @doc = Doc.find(params[:id])\n \n respond_to do |format|\n if @doc.update_attributes(params[:doc])\n save_object_relationship\n format.html { redirect_to(@doc, :notice => 'Doc was successfully updated.') }\n format.xml { head :ok }\n else\n format.html { render :action => \"edit\" }\n format.xml { render :xml => @doc.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @document.update(document_params)\n @document.create_activity :update, owner: current_user\n format.html { redirect_to @document, notice: 'Document was successfully updated.' }\n format.json { render :show, status: :ok, location: @document }\n else\n format.html { render :edit }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @documentotipo = Documentotipo.find(params[:id])\n\n respond_to do |format|\n if @documentotipo.update_attributes(params[:documentotipo])\n format.html { redirect_to @documentotipo, notice: 'Documentotipo was successfully updated.' }\n format.json { head :ok }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @documentotipo.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update_document!(model, id)\n document = model.update(id, params)\n unless document.valid?\n error 400, convert(body_for(:invalid_document, document))\n end\n document\n end",
"def update\n if @doc.update_from_params(params[:course_document])\n render json: @doc, status: :ok\n else\n render json: doc.errors, status: :unprocessable_entity\n end\n end",
"def update!(**args)\n @document = args[:document] if args.key?(:document)\n @id = args[:id] if args.key?(:id)\n @metadata = args[:metadata] if args.key?(:metadata)\n end",
"def update\n @doc = Doc.find(params[:id])\n\n respond_to do |format|\n if @doc.update_attributes(params[:doc])\n format.html { redirect_to(@doc, :notice => 'Doc was successfully updated.') }\n format.xml { head :ok }\n else\n format.html { render :action => \"edit\" }\n format.xml { render :xml => @doc.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @document.update(document_params)\n format.html { redirect_to back_url, notice: 'Document was successfully updated.' }\n else\n format.html { 'documents/edit' }\n end\n end\n end",
"def update\n respond_to do |format|\n if @document.update(document_params)\n format.html { redirect_to @document, success: 'Documento actualizado con éxito.' }\n format.json { render :show, status: :ok, location: @document }\n else\n format.html { render :edit }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def operations\n { \"$set\" => { path => documents } }\n end",
"def update\n respond_to do |format|\n if @document.update(document_params)\n format.html { redirect_to back_index_case_url, notice: 'Document was successfully updated.' }\n else\n format.html { render :edit }\n end\n end\n end",
"def update\n respond_to do |wants|\n if @document.update_attributes(params[:document])\n flash[:notice] = 'Document was successfully updated.'\n wants.html { redirect_to(@document) }\n wants.xml { head :ok }\n else\n wants.html { render :action => \"edit\" }\n wants.xml { render :xml => @document.errors, :status => :unprocessable_entity }\n end\n end\n end",
"def update\n @document = Document.find(params[:id])\n suc_msg = 'Document was successfully updated. '\n\n #if we're in document text edit mode, or notes edit mode\n if (params.include?(\"document\")) and (params[\"document\"].include?(\"post\")) and (params[\"document\"][\"post\"] == \"edit_text\")\n @document.stuffing_text = params[\"document\"][\"stuffing_text\"]\n update_suc = @document.save\n elsif (params.include?(\"document\")) and (params[\"document\"].include?(\"post\")) and (params[\"document\"][\"post\"] == \"edit_notes\")\n @document.stuffing_notes = params[\"document\"][\"stuffing_notes\"]\n update_suc = @document.save\n else\n #Add doc to project\n if params.include?(\"proj\") and params[:proj].include?(\"id\") and params[:proj][:id] != \"\"\n project = Project.find(params[:proj][:id])\n if (project != nil)\n add_project_doc(project, @document) #call to document helper, adds doc to project\n end\n end\n \n #Add selected upload as a note to the document\n if (params.include?(\"note\") and params[\"note\"].include?(\"upload_id\") and (!params[\"note\"][\"upload_id\"].blank?) )\n add_note(params[\"note\"][\"upload_id\"])\n end\n \n #Removed selected notes from a documet\n if (params.include?(\"remove_ids\") and (!params[\"remove_ids\"].blank?) )\n remove_notes(params[\"remove_ids\"]) #Remove notes\n end\n\n user = User.where(:id => params[:new_user_id]).first\n\n #Update other attributes\n if (@document.user_id == current_user.id)\n update_suc = @document.update_attributes(params[:document])\n else\n #needed becuase update will drop collection id if an editor tries to use a filter\n coll_id = @document.collection_id\n update_suc = @document.update_attributes(params[:document])\n update_suc = @document.update_attributes(:collection_id => coll_id)\n end\n\n #Filter / Validate\n if ( params.include?(\"post\") and params[:post].include?(\"ifilter_id\") and params[:post][:ifilter_id] != \"\" )\n \n f = get_ifilter(params[:post][:ifilter_id].to_i)\n status, msg = validate_document(@document, f)\n suc_msg += msg\n end\n end #end if in text edit mode, else...\n\n respond_to do |format|\n if update_suc\n format.html { redirect_to edit_document_path(@document), notice: suc_msg }\n format.json { head :ok }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @shared_document.update(shared_document_params)\n format.html { redirect_to @shared_document, notice: 'Shared document was successfully updated.' }\n format.json { head :no_content }\n else\n format.html { render action: 'edit' }\n format.json { render json: @shared_document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def put!\n request! :put\n end",
"def update!(**args)\n @document_id = args[:document_id] if args.key?(:document_id)\n end",
"def update\n respond_to do |format|\n if @document.update(document_params)\n if params[:attachments]\n params[:attachments].each do |a|\n @attachment = @document.attachments.create!(:attachment => a)\n end\n end\n format.html { redirect_to @document, notice: 'Document was successfully updated.' }\n format.json { render :show, status: :ok, location: @document }\n else\n format.html { render :edit }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @tdoc = Tdoc.find(params[:id])\n \n respond_to do |format|\n if @tdoc.update_attributes(params[:tdoc])\n format.html { redirect_to @tdoc, notice: 'Tdoc was successfully updated.' }\n format.json { head :ok }\n else\n format.html { render action: \"edit\" }\n format.json { render json: @tdoc.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update # PATCH\n raise NotImplementedError\n end",
"def update\n authorize @document\n\n respond_to do |format|\n if @document.update(document_params)\n flash_label = \"Document was successfully updated.\"\n flash.now[:notice] = flash_label\n\n format.html { redirect_to @document, notice: flash_label }\n format.json { render :show, status: :ok, location: @document }\n format.js\n else\n format.html { render :edit }\n format.json { render json: @document.errors, status: :unprocessable_entity }\n format.js\n end\n end\n end",
"def rest_edit(path, options={}, &blk)\n callback = Proc.new { |*args|\n @object = yield(*args) or pass\n rest_params.each { |k, v| @object.send :\"#{k}=\", v unless k == 'id' }\n\n return 400, @object.errors.to_json unless @object.valid?\n\n @object.save\n rest_respond @object\n }\n\n # Make it work with `Backbone.emulateHTTP` on.\n put path, &callback\n post path, &callback\n end",
"def rest_edit(path, options={}, &blk)\n callback = Proc.new { |*args|\n @object = yield(*args) or pass\n rest_params.each { |k, v| @object.send :\"#{k}=\", v unless k == 'id' }\n\n return 400, @object.errors.to_json unless @object.valid?\n\n @object.save\n rest_respond @object\n }\n\n # Make it work with `Backbone.emulateHTTP` on.\n put path, &callback\n post path, &callback\n end",
"def update!(**args)\n @document_id = args[:document_id] if args.key?(:document_id)\n @status = args[:status] if args.key?(:status)\n end",
"def update\n respond_to do |format|\n if @collection.update(collection_params)\n format.html { redirect_to collection_documents_path(@collection), notice: 'The collection was successfully updated.' }\n format.json { render :show, status: :ok, location: @collection }\n else\n format.html { render :edit }\n format.json { render json: @collection.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @documents_person.update(documents_person_params)\n format.html { redirect_to @documents_person, notice: 'Documents person was successfully updated.' }\n format.json { render :show, status: :ok, location: @documents_person }\n else\n format.html { render :edit }\n format.json { render json: @documents_person.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update(url, data)\n RestClient.put url, data, :content_type => :json\nend",
"def update\n respond_to do |format|\n if @docfile.update(docfile_params)\n format.html { redirect_to @docfile, notice: \"Docfile was successfully updated.\" }\n format.json { render :show, status: :ok, location: @docfile }\n else\n format.html { render :edit, status: :unprocessable_entity }\n format.json { render json: @docfile.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @document_file.update(document_file_params)\n format.html { redirect_to edit_document_path(@document), notice: 'Document file was successfully updated.' }\n format.json { render json: 'ok', status: :ok, location: @document_file }\n else\n format.html { render :edit }\n format.json { render json: @folder.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @event_subscription.update(event_subscription_params)\n @event_subscription.save\n\n file_params.each do |requirement|\n if(requirement[\"doc\"])\n requirement.symbolize_keys\n requirement[:doc].symbolize_keys\n path = \"data:#{requirement[:doc][:filetype]};base64, #{requirement[:doc][:base64]}\"\n Document.update(id: requirement[:doc][:id],\n user_id: @event_subscription.user_id,\n requirement_id: requirement[:id],\n state: \"pending_review\",\n path: path\n )\n end\n end\n render json: @event_subscription, status: :updated\n end",
"def update\n respond_to do |format|\n if @admin_documento.update(admin_documento_params)\n format.html { redirect_to admin_documentos_url }\n format.json { render :show, status: :ok, location: @admin_documento }\n else\n format.html { render :edit }\n format.json { render json: @admin_documento.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n @documento = Documento.find(params[:id])\n respond_to do |format|\n if @documento.update_attributes(params[:documento])\n format.html { redirect_to @documento, notice: 'Documento actualizado exitosamente.' }\n format.json { head :no_content }\n else\n format.html { render action: \"edit\", alert: 'Documento no pudo ser actualizado.' }\n format.json { render json: @documento.errors, status: :unprocessable_entity }\n end\n end\n end",
"def show\n if request.put?\n logger.debug params.inspect\n else\n respond_to do |format|\n format.html\n format.json { render :json => @document.to_obj }\n end\n end\n end",
"def update!(**args)\n @documents = args[:documents] if args.key?(:documents)\n @processor_info = args[:processor_info] if args.key?(:processor_info)\n end",
"def update\n respond_to do |format|\n if @couch.update(couch_params)\n format.html { redirect_to @couch, notice: 'Couch was successfully updated.' }\n format.json { render :show, status: :ok, location: @couch }\n else\n format.html { render :edit }\n format.json { render json: @couch.errors, status: :unprocessable_entity }\n end\n end\n end",
"def update\n respond_to do |format|\n if @docs_ponto.update(docs_ponto_params)\n format.html { redirect_to :back }\n format.json { head :no_content }\n else\n format.html { render action: 'edit' }\n format.json { render json: @docs_ponto.errors, status: :unprocessable_entity }\n end\n end\n end"
] |
[
"0.70435137",
"0.6930837",
"0.6855188",
"0.6815346",
"0.68135005",
"0.67967653",
"0.6793707",
"0.67770743",
"0.67770743",
"0.67770743",
"0.6773916",
"0.67491776",
"0.6741166",
"0.6725393",
"0.67247325",
"0.6706539",
"0.67052364",
"0.66906583",
"0.66906583",
"0.66906583",
"0.66876215",
"0.6673912",
"0.6660242",
"0.6630856",
"0.6600182",
"0.6587207",
"0.656486",
"0.65352994",
"0.6522962",
"0.65210396",
"0.6512598",
"0.6500489",
"0.6500489",
"0.6500489",
"0.6500489",
"0.6500489",
"0.6500489",
"0.6500489",
"0.6498879",
"0.6490631",
"0.6476599",
"0.64736444",
"0.64561206",
"0.6431792",
"0.64239377",
"0.641092",
"0.6389844",
"0.63637245",
"0.63614476",
"0.6358351",
"0.6347044",
"0.6310242",
"0.6310242",
"0.6306436",
"0.63019973",
"0.6285808",
"0.62857354",
"0.6284708",
"0.62782955",
"0.62607545",
"0.6257117",
"0.62570024",
"0.62486964",
"0.624217",
"0.6238632",
"0.6230852",
"0.62306976",
"0.6204322",
"0.6203307",
"0.6194676",
"0.61945933",
"0.6175026",
"0.6165756",
"0.6162653",
"0.6161446",
"0.61510146",
"0.61439466",
"0.61396897",
"0.61217946",
"0.6106124",
"0.6102206",
"0.6093257",
"0.60845214",
"0.6072432",
"0.6068512",
"0.60651517",
"0.606361",
"0.606361",
"0.6061975",
"0.60580695",
"0.60567707",
"0.605002",
"0.6045748",
"0.6036312",
"0.6031565",
"0.6030904",
"0.60225654",
"0.60213816",
"0.60149306",
"0.6012468",
"0.6009832"
] |
0.0
|
-1
|
DELETE /documents/1 DELETE /documents/1.json
|
def destroy
@attachment.destroy
respond_to do |format|
format.html { redirect_to attachments_url }
format.json { head :no_content }
end
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def delete\n @client.delete_document(@path)\n end",
"def delete_document index, id\n uri = URI(\"http://#{@host}:#{@port_s}/#{index}/_doc/#{id}\")\n req = Net::HTTP::Delete.new(uri)\n run(uri, req)\n end",
"def destroy\n @doc = Doc.find(params[:id])\n @doc.destroy\n\n respond_to do |format|\n format.html { redirect_to docs_url }\n format.json { head :ok }\n end\n end",
"def destroy\n @document = Document.find(params[:id])\n @document.destroy\n\n respond_to do |format|\n format.html { redirect_to root_url }\n format.json { head :ok }\n end\n end",
"def destroy\n @document.destroy\n respond_to do |format|\n format.html { redirect_to documents_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document.destroy\n respond_to do |format|\n format.html { redirect_to documents_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document.destroy\n respond_to do |format|\n format.html { redirect_to documents_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document.destroy\n respond_to do |format|\n format.html { redirect_to documents_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document.destroy\n respond_to do |format|\n format.html { redirect_to documents_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document = Document.find(params[:id])\n @document.destroy\n\n respond_to do |format|\n format.html { redirect_to documents_url }\n format.json { head :ok }\n end\n end",
"def destroy\n @document = Document.find(params[:id])\n @document.destroy\n\n respond_to do |format|\n format.html { redirect_to documents_url }\n format.json { head :ok }\n end\n end",
"def destroy\n @document = Document.find(params[:id])\n @document.destroy\n\n respond_to do |format|\n format.html { redirect_to documents_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document = Document.find(params[:id])\n @document.destroy\n\n respond_to do |format|\n format.html { redirect_to documents_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document = Document.find(params[:id])\n @document.destroy\n\n respond_to do |format|\n format.html { redirect_to documents_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document = Document.find(params[:id])\n @document.destroy\n\n respond_to do |format|\n format.html { redirect_to documents_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document = Document.find(params[:id])\n @document.destroy\n\n respond_to do |format|\n format.html { redirect_to documents_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document = Document.find(params[:id])\n @document.destroy\n\n respond_to do |format|\n format.html { redirect_to documents_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document = Document.find(params[:id])\n @document.destroy\n\n respond_to do |format|\n format.html { redirect_to documents_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document = Document.find(params[:id])\n @document.destroy\n\n respond_to do |format|\n format.html { redirect_to documents_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document = Document.find(params[:id])\n @document.destroy\n\n respond_to do |format|\n format.html { redirect_to documents_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document = Document.find(params[:id])\n @document.destroy\n respond_to do |format|\n format.html { redirect_to documents_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document = Document.find(params[:id])\n @document.destroy\n\n respond_to do |format|\n format.html { redirect_to admin_documents_url }\n format.json { head :ok }\n end\n end",
"def destroy\r\n @document = Document.find(params[:id])\r\n @document.destroy\r\n\r\n respond_to do |format|\r\n format.html { redirect_to documents_url }\r\n format.json { head :no_content }\r\n end\r\n end",
"def destroy\n @doc.destroy\n respond_to do |format|\n format.html { redirect_to docs_url, notice: 'Документ был удалён' }\n format.json { head :no_content }\n end\n end",
"def destroy\n user = User.find(params[:user_id])\n user.documents.delete(Document.find(params[:document_id]))\n\n respond_to do |format|\n format.html { redirect_to :back }\n format.json { head :ok }\n end\n end",
"def delete_document( doc_id: )\n params = {}\n params[:backtrace] = @backtrace if @backtrace\n send_request :delete, url_for_base(doc_id), params, :json\n end",
"def destroy\n \n @document = Document.find(params[:id])\n @document.destroy\n\n respond_to do |format|\n format.html { redirect_to documents_url, notice: 'Documento eliminado correctamente' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document.destroy\n respond_to do |format|\n format.html { redirect_to back_index_case_url, notice: 'Document was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document.destroy\n respond_to do |format|\n format.html { redirect_to back_index_case_url, notice: 'Document was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document = Document.find(params[:id])\n @document.destroy\n\n respond_to do |format|\n format.html { redirect_to documents_url }\n # format.json { head :no_content }\n format.json { render json: @document }\n end\n end",
"def destroy\n \t@document = Document.find(params[:id])\n unless @document.blank?\n \tif @document.destroy\n \t\trender json: {status: 'SUCCESS', message:'Document deleted', data:@document}, status: :ok\n \telse \n \t\trender json: {status: 'ERROR', message:'Document was not deleted', data:@document}, status: :unprocessable_entity\n \tend\n else \n \t\trender json: {status: 'ERROR', message:'No such document', data:@document}, status: :unprocessable_entity\n end\n end",
"def destroy\n @documento = Documento.find(params[:id])\n @documento.destroy\n\n respond_to do |format|\n format.html { redirect_to documentos_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document = Document.find(params[:id])\n if @document.destroy\n render json: @document, status: :accepted\n else\n render json: {errors: @document.errors.full_messages}, status: :unprocessable_entity\n end\n end",
"def destroy\n @document.destroy\n respond_to do |format|\n format.html { redirect_to documents_url, notice: 'Document was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document.destroy\n respond_to do |format|\n format.html { redirect_to documents_url, notice: 'Document was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document.destroy\n respond_to do |format|\n format.html { redirect_to documents_url, notice: 'Document was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document.destroy\n respond_to do |format|\n format.html { redirect_to documents_url, notice: 'Document was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document.destroy\n respond_to do |format|\n format.html { redirect_to documents_url, notice: 'Document was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document.destroy\n respond_to do |format|\n format.html { redirect_to documents_url, notice: 'Document was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document.destroy\n respond_to do |format|\n format.html { redirect_to documents_url, notice: 'Document was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document.destroy\n respond_to do |format|\n format.html { redirect_to documents_url, notice: 'Document was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document.destroy\n respond_to do |format|\n format.html { redirect_to documents_url, notice: 'Document was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document.destroy\n respond_to do |format|\n format.html { redirect_to documents_url, notice: 'Document was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document.destroy\n respond_to do |format|\n format.html { redirect_to documents_url, notice: 'Document was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document.destroy\n respond_to do |format|\n format.html { redirect_to documents_url, notice: 'Document was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document.destroy\n respond_to do |format|\n format.html { redirect_to documents_url, notice: 'Document was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document.destroy\n respond_to do |format|\n format.html { redirect_to documents_url, notice: 'Document was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document.destroy\n respond_to do |format|\n format.html { redirect_to documents_url, notice: \"Document was successfully destroyed.\" }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document.destroy\n respond_to do |format|\n format.html { redirect_to documents_url, notice: \"Document was successfully destroyed.\" }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document = Document.find(:first, :conditions => [\"id = ?\", params[:id]])\n @document.destroy\n\n respond_to do |format|\n format.html { redirect_to(documents_url) }\n format.xml { head :ok }\n end\n end",
"def destroy\r\n @document.destroy\r\n respond_to do |format|\r\n format.html { redirect_to root_url, notice: 'Document was successfully destroyed.' }\r\n format.json { head :no_content }\r\n end\r\n end",
"def destroy\n @document.file = nil\n @document.save\n @document.destroy\n respond_to do |format|\n format.html { redirect_to :back, notice: 'Document was successfully deleted.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @client_doc.destroy\n respond_to do |format|\n format.html { redirect_to client_docs_url, notice: 'Client doc was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n authorize @document\n instruction = @document.instruction\n @document.destroy\n respond_to do |format|\n format.html { redirect_to instruction, notice: t('documents.destroy.success') }\n format.json { head :no_content }\n end\n end",
"def destroy\n @squadron_document = SquadronDocument.find(params[:id])\n @squadron_document.destroy\n\n respond_to do |format|\n format.html { redirect_to squadron_documents_url }\n format.json { head :ok }\n end\n end",
"def destroy\n @document = Document.find(params[:id])\n @document.destroy\n\n respond_to do |format|\n format.html { redirect_to(documents_url) }\n format.xml { head :ok }\n end\n end",
"def destroy\n @document = Document.find(params[:id])\n @document.destroy\n\n respond_to do |format|\n format.html { redirect_to(documents_url) }\n format.xml { head :ok }\n end\n end",
"def destroy\n @document = Document.find(params[:id])\n @document.destroy\n\n respond_to do |format|\n format.html { redirect_to(documents_url) }\n format.xml { head :ok }\n end\n end",
"def destroy\n\n @documentable = @document.documentable\n @document.destroy\n\n notify_user(:notice, 'Document was successfully removed.')\n respond_to do |format|\n format.html { redirect_back(fallback_location: root_path) }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document.destroy\n save_log({ :action => 'DELETE', :document => { :before => @old_document }})\n #TODO delete the file from disk\n respond_to do |format|\n format.html { redirect_to(documents_url) }\n format.xml { head :ok }\n end\n end",
"def destroy\n @document = Document.find(params[:id])\n collection = Collection.find(@document.collection_id)\n upload_remove(@document) #Removes upload record if file is deleted\n @document.destroy\n\n respond_to do |format|\n #format.html { redirect_to collections_path }\n format.html { redirect_to collection }\n format.json { head :ok }\n end\n end",
"def destroy\n @document = Document.find(params[:id])\n @document.destroy\n\n respond_to do |format|\n format.html { redirect_to admin_documents_path }\n format.json { head :no_content }\n format.js { head :no_content }\n end\n end",
"def destroy\n @admin_documento.destroy\n respond_to do |format|\n format.html { redirect_to admin_documentos_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @has_document.destroy\n respond_to do |format|\n format.html { redirect_to has_documents_url, notice: 'Has document was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @modified_document = ModifiedDocument.find(params[:id])\n \n if @modified_document.destroy\n render json: @modified_document, status: :ok \n else\n render json: @modified_document.errors, status: :unprocessable_entity\n end\n end",
"def destroy\n @doc = Doc.find(params[:id])\n @doc.destroy\n\n respond_to do |format|\n format.html { redirect_to(docs_url) }\n format.xml { head :ok }\n end\n end",
"def destroy\n @doc = Doc.find(params[:id])\n @doc.destroy\n\n respond_to do |format|\n format.html { redirect_to(docs_url) }\n format.xml { head :ok }\n end\n end",
"def deleteEntityDocument( entity_id, gen_id)\n params = Hash.new\n params['entity_id'] = entity_id\n params['gen_id'] = gen_id\n return doCurl(\"delete\",\"/entity/document\",params)\n end",
"def destroy\n \t@doc = Doc.find params[:id]\n @doc.destroy\nend",
"def destroy\n @document = Document.find(params[:id])\n @document.destroy\n\n respond_to do |format|\n format.html { redirect_to documents_url, notice: t('.notice', document: @document) }\n format.json { head :no_content }\n format.js\n end\n end",
"def destroy\n @document = Document.find(params[:id])\n if not @document.destroy\n flash[:error] = 'There was an error deleting the document'\n end\n\n respond_to do |format|\n format.html { redirect_to documents_path }\n format.json { head :no_content }\n end\n end",
"def destroy\n @oa_sent_document = Oa::SentDocument.find(params[:id])\n @oa_sent_document.destroy\n\n respond_to do |format|\n format.html { redirect_to oa_sent_documents_url }\n format.json { head :ok }\n end\n end",
"def destroy\n @docfile.destroy\n respond_to do |format|\n format.html { redirect_to docfiles_url, notice: \"Docfile was successfully destroyed.\" }\n format.json { head :no_content }\n end\n end",
"def destroy_document(opts = {})\n es_client.delete document_path(opts)\n end",
"def destroy\n @tdoc = Tdoc.find(params[:id])\n @tdoc.destroy\n \n respond_to do |format|\n format.html { redirect_to tdocs_url }\n format.json { head :ok }\n end\n end",
"def destroy\n @document.create_activity :destroy, owner: current_user\n @document.destroy\n respond_to do |format|\n format.html { redirect_to documents_url, notice: 'Document was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def delete(opts = {})\n response = Crocodoc.connection.post 'document/delete', :uuid => @uuid\n response.body.chomp.downcase == 'true'\n end",
"def destroy\n begin\n @collection = baseCollections.find(params[:collection_id]) \n rescue ActiveRecord::RecordNotFound\n redirect_to collections_url, :alert => \"Not exist collection or you are not allowed to see.\"\n return\n end\n @document = @collection.documents.find_by_id(params[:id])\n @document.destroy\n\n respond_to do |format|\n format.html { redirect_to @collection, notice: 'Document was successfully removed.'}\n format.json { head :no_content }\n end\n end",
"def destroy\n @special_document = ModifiedDocument.find(params[:id])\n @special_document.destroy\n\n respond_to do |format|\n format.html { redirect_to special_documents_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n authorize @document\n @document.destroy\n respond_to do |format|\n format.html { redirect_to documents_url, notice: t('document.destroy.confirmation') }\n format.json { head :no_content }\n end\n end",
"def destroy\n @tipo_documento = TipoDocumento.find(params[:id])\n @tipo_documento.destroy\n\n respond_to do |format|\n format.html { redirect_to tipo_documentos_url }\n format.json { head :no_content }\n end\n end",
"def delete\n client.delete(\"/#{id}\")\n end",
"def destroy\n @document.destroy\n respond_to do |format|\n format.html { redirect_to documents_url(document_filter: {kind: @document.kind}), success: 'Documento eliminado con éxito.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @document.destroy\n end",
"def delete\n request(:delete)\n end",
"def delete!\n request! :delete\n end",
"def destroy\n @document.destroy\n respond_to do |format|\n format.html { redirect_to department_documents_path @department }\n format.json { head :no_content }\n end\n end",
"def destroy\n @documento.destroy\n respond_to do |format|\n format.html { redirect_to documentos_url, notice: 'Documento was successfully destroyed.' }\n format.json { head :no_content }\n end\n end",
"def destroy\n @shared_document.destroy\n respond_to do |format|\n format.html { redirect_to shared_documents_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @docs_ponto.destroy\n respond_to do |format|\n format.html { redirect_to docs_pontos_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n @documento = Documento.find(params[:id])\n @documento.destroy\n\n respond_to do |format|\n format.html { redirect_to(documentos_url) }\n format.xml { head :ok }\n end\n end",
"def destroy\n @documentotipo = Documentotipo.find(params[:id])\n @documentotipo.destroy\n\n respond_to do |format|\n format.html { redirect_to documentotipos_url }\n format.json { head :ok }\n end\n end",
"def destroy\n @gdoc.destroy\n respond_to do |format|\n format.html { redirect_to gdocs_url }\n format.json { head :no_content }\n end\n end",
"def destroy\r\n @customer_document = CustomerDocument.find(params[:id])\r\n @customer_document.destroy\r\n\r\n respond_to do |format|\r\n format.html { redirect_to customer_documents_url }\r\n format.json { head :no_content }\r\n end\r\n end",
"def delete_document\n @document = Document.find(params[:id])\n @document.destroy\n flash[:notice] = 'Document successfully deleted. !'\n redirect_to :action => :list\n end",
"def destroy\n \t\n @companydocument = Companydocument.find(params[:id])\n @companydocument.destroy\n\n respond_to do |format|\n format.html { redirect_to companydocuments_url }\n format.json { head :no_content }\n end\n end",
"def destroy\n (@document = Document.find(params[:id])).destroy\n flash['notice'] = 'Destroyed document.'\n respond_with @document\n end",
"def destroy\n @doc.destroy\n redirect_to docs_path\n end",
"def delete(document)\n delete_path(document.path)\n end",
"def destroy\n @id = params[:id]\n @user = current_user\n @profile = @user.profile\n @user_document = @user.documents.find(@id)\n @user_document.document.destroy\n\n if @user_document.destroy\n respond_to do |format|\n format.json { render :json => [:id=>@id], :status => :ok }\n end\n\n else\n respond_to do |format|\n format.json { render :json => @profile.errors, :status => :unprocessable_entity }\n end\n\n end\n\n end",
"def destroy\n @document.destroy\n\n respond_to do |wants|\n wants.html { redirect_to(documents_url) }\n wants.xml { head :ok }\n end\n end"
] |
[
"0.78232354",
"0.761401",
"0.7518423",
"0.75177574",
"0.7488822",
"0.7488822",
"0.7488822",
"0.7488822",
"0.7488822",
"0.7483412",
"0.7483412",
"0.7474384",
"0.7474384",
"0.7474384",
"0.7474384",
"0.7474384",
"0.7474384",
"0.7474384",
"0.7474384",
"0.7474384",
"0.74585146",
"0.7431029",
"0.74234265",
"0.7416177",
"0.74146056",
"0.7329395",
"0.7328653",
"0.7327659",
"0.7327659",
"0.73240066",
"0.7313219",
"0.7313006",
"0.72855675",
"0.7229978",
"0.7229978",
"0.7229978",
"0.7229978",
"0.7229978",
"0.7229978",
"0.7229978",
"0.7229978",
"0.7229978",
"0.7229978",
"0.7229978",
"0.7229978",
"0.7229978",
"0.7229978",
"0.7224111",
"0.7224111",
"0.71999544",
"0.7199324",
"0.71628296",
"0.715822",
"0.71423656",
"0.71368116",
"0.71339154",
"0.71339154",
"0.71339154",
"0.71338505",
"0.7109069",
"0.7102997",
"0.71011657",
"0.7089564",
"0.70857257",
"0.70840544",
"0.7077002",
"0.7077002",
"0.70740056",
"0.7071343",
"0.7067764",
"0.70549285",
"0.7048996",
"0.70428175",
"0.70384675",
"0.7031552",
"0.7029917",
"0.7028915",
"0.702583",
"0.7012551",
"0.7009529",
"0.70047367",
"0.70042396",
"0.70019513",
"0.69984424",
"0.69947416",
"0.6981742",
"0.6967921",
"0.69560415",
"0.69552565",
"0.69497764",
"0.6945253",
"0.6942498",
"0.69332695",
"0.6914753",
"0.6909102",
"0.6903946",
"0.69033724",
"0.6901315",
"0.69004166",
"0.6892682",
"0.6890542"
] |
0.0
|
-1
|
Use callbacks to share common setup or constraints between actions.
|
def set_attachment
@attachment = Attachment.find(params[:id])
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def set_required_actions\n # TODO: check what fields change to asign required fields\n end",
"def action_hook; end",
"def run_actions; end",
"def define_action_hook; end",
"def actions; end",
"def define_action_helpers\n if super && action == :save\n @instance_helper_module.class_eval do\n define_method(:valid?) do |*args|\n self.class.state_machines.fire_event_attributes(self, :save, false) { super(*args) }\n end\n end\n end\n end",
"def add_actions; end",
"def callbacks; end",
"def callbacks; end",
"def setup *actions, &proc\n (@setup_procs ||= []) << [proc, actions.size > 0 ? actions : [:*]]\n end",
"def define_action_helpers; end",
"def post_setup\n end",
"def action_methods; end",
"def action_methods; end",
"def action_methods; end",
"def before_setup; end",
"def action_run\n end",
"def execute(setup)\n @action.call(setup)\n end",
"def define_action_helpers?; end",
"def set_actions\n actions :all\n end",
"def action_done(action)\n dispatch = { :migrate => :done_migrating, :map => :done_mapping, :reduce =>\n :done_reducing, :finalize => :done_finalizing } \n self.send dispatch[action[:action]], action\n end",
"def dependencies action, &block\n @actions.each do |other|\n if action[:requires].include? other[:provide]\n block.call other\n end\n end\n end",
"def setup!\n return unless @setup_procs\n http_actions = actions\n @setup_procs.each do |setup_proc|\n proc, actions = setup_proc\n @setup__actions = actions.map do |action|\n\n action.is_a?(Regexp) ?\n http_actions.select { |a| a.to_s =~ action } :\n action.is_a?(String) && action =~ /\\A\\./ ?\n http_actions.map { |a| a.to_s << action if format?(a).include?(action) }.compact :\n action\n\n end.flatten\n self.class_exec &proc\n @setup__actions = nil\n end\n @setup_procs = nil\n end",
"def before_actions(*logic)\n self.before_actions = logic\n end",
"def setup_handler\n end",
"def set_action(opts)\n opts = check_params(opts,[:actions])\n super(opts)\n end",
"def setup(action)\n @targets.clear\n unless action.item.target_filters.empty?\n @targets = SES::TargetManager.make_targets(action)\n else\n item = action.item\n if item.for_opponent?\n @targets = $game_troop.alive_members\n elsif item.for_dead_friend?\n @targets = $game_party.battle_members.select { |actor| actor.dead? }\n else\n $game_party.battle_members.select { |actor| actor.alive? }\n end\n end\n @item_max = @targets.size\n create_contents\n refresh\n show\n activate\n end",
"def action; end",
"def action; end",
"def action; end",
"def action; end",
"def action; end",
"def workflow\n end",
"def revisable_shared_setup(args, block)\n class << self\n attr_accessor :revisable_options\n end\n options = args.extract_options!\n self.revisable_options = Options.new(options, &block)\n \n self.send(:include, Common)\n self.send(:extend, Validations) unless self.revisable_options.no_validation_scoping?\n self.send(:include, WithoutScope::QuotedColumnConditions)\n end",
"def setup\n @action = SampleActionAndroid.new(os_name: 'android',\n app_name: APP_PATH)\n end",
"def before(action)\n invoke_callbacks *self.class.send(action).before\n end",
"def process_action(...)\n send_action(...)\n end",
"def before_dispatch(env); end",
"def after_actions(*logic)\n self.after_actions = logic\n end",
"def setup\n # override and do something appropriate\n end",
"def setup(client)\n return unless @setup\n actions = @setup['setup'].select { |action| action['do'] }.map { |action| Action.new(action['do']) }\n actions.each do |action|\n action.execute(client)\n end\n self\n end",
"def setup(_context)\n end",
"def setup(resources) ; end",
"def validate_actions\n errors.add(:base, :should_give_at_least_one_action) if !manage? && !forecasting? && !read? && !api?\n end",
"def setup\n @resource_config = {\n :callbacks => {\n :before_create => nil,\n :after_create => nil,\n :before_update => nil,\n :after_update => nil,\n :before_destroy => nil,\n :after_destroy => nil,\n },\n :child_assoc => nil,\n :model => nil,\n :parent => nil,\n :path => nil,\n :permission => {},\n :properties => {},\n :relation => {\n :create => nil,\n :delete => nil,\n },\n :roles => nil,\n }\n end",
"def determine_valid_action\n\n end",
"def process_shared\n handle_taxes\n handle_shippings\n create_adjustments_from_params\n handle_status\n handle_inventory_refunds\n handle_payment_transactions\n order.updater.update\n end",
"def startcompany(action)\n @done = true\n action.setup\n end",
"def init_actions\n am = action_manager()\n am.add_action(Action.new(\"&Disable selection\") { @selection_mode = :none; unbind_key(32); bind_key(32, :scroll_forward); } )\n am.add_action(Action.new(\"&Edit Toggle\") { @edit_toggle = !@edit_toggle; $status_message.value = \"Edit toggle is #{@edit_toggle}\" })\n end",
"def event_callbacks(event, metadata={})\n case event\n when :reset, :review\n if confirmed\n update_attributes(confirmed: false)\n end\n when :confirm\n confirm\n # trigger :order for all applicable items\n # NOTE: :order event is common to both physical and digital items\n items.each do |i|\n if i.event_permitted(:order)\n user_id = last_transition.user_id\n i.trigger!(:order, { order_id: id, user_id: user_id })\n end\n end\n when :complete_work\n request = metadata[:request]\n work_complete_notification(request)\n when :close\n close\n end\n if event != :close && !open\n reopen\n end\n end",
"def setup_action\n return unless PONY::ERRNO::check_sequence(current_act)\n new_sequence = @action_sequence[@sequence_index+1...@action_sequence.size]\n @sequence_index = 0\n new_sequence = DND::SkillSequence::ACTS[@acts[1]] + new_sequence\n execute_sequence\n end",
"def define_tasks\n define_weave_task\n connect_common_tasks\n end",
"def setup(&block)\n define_method(:setup, &block)\n end",
"def setup\n transition_to(:setup)\n end",
"def setup\n transition_to(:setup)\n end",
"def action\n end",
"def setup( *args )\n\t\t\tself.class.setupBlocks.each {|sblock|\n\t\t\t\tdebugMsg \"Calling setup block method #{sblock}\"\n\t\t\t\tself.send( sblock )\n\t\t\t}\n\t\t\tsuper( *args )\n\t\tend",
"def config(action, *args); end",
"def setup\n @setup_proc.call(self) if @setup_proc\n end",
"def before_action \n end",
"def setup_callbacks\n defined_callbacks.each do |meth|\n unless respond_to?(\"call_#{meth}_callbacks\".to_sym)\n self.class.module_eval <<-EOE\n def call_#{meth}_callbacks(*args)\n plugin_store.each {|a| a.call_#{meth}_callbacks(*args) } if respond_to?(:plugin_store) && plugin_store\n self.send :#{meth}, *args if respond_to?(:#{meth})\n end\n EOE\n end\n end\n end",
"def action\n end",
"def matt_custom_action_begin(label); end",
"def setup\n # override this if needed\n end",
"def setup\n\t\t\t\t\t\t# Do nothing\n\t\t\t\tend",
"def setup\n\t\t\t\t\t\t# Do nothing\n\t\t\t\tend",
"def action(options,&callback)\n new_action = Action===options ? options : Action.new(options,&callback)\n # replace any with (shared name/alias or both default) + same arity\n @actions.delete_if do |existing_action|\n ((existing_action.names & new_action.names).size > 0 ||\n existing_action.default? && new_action.default?) &&\n existing_action.required.size == new_action.required.size &&\n existing_action.optional.size <= new_action.optional.size\n end\n @actions = (@actions + [new_action]).sort\n new_action\n end",
"def set_target_and_action target, action\n self.target = target\n self.action = 'sugarcube_handle_action:'\n @sugarcube_action = action\n end",
"def after(action)\n invoke_callbacks *options_for(action).after\n end",
"def pre_task\n end",
"def setup(server)\n server.on('beforeMethod', method(:before_method), 10)\n end",
"def add_actions\n attribute = machine.attribute\n name = self.name\n \n owner_class.class_eval do\n define_method(name) {self.class.state_machines[attribute].events[name].fire(self)}\n define_method(\"#{name}!\") {self.class.state_machines[attribute].events[name].fire!(self)}\n define_method(\"can_#{name}?\") {self.class.state_machines[attribute].events[name].can_fire?(self)}\n end\n end",
"def init_actions\n @select_action = SelectAction.new\n @endpoint_mouse_action = EndpointMouseAction.new\n @move_action = MoveAction.new\n end",
"def setup_signals; end",
"def after_created\r\n return unless compile_time\r\n Array(action).each do |action|\r\n run_action(action)\r\n end\r\nend",
"def after_created\r\n return unless compile_time\r\n Array(action).each do |action|\r\n run_action(action)\r\n end\r\nend",
"def set_target_and_action target, action\n self.target = target\n self.action = 'sugarcube_handle_action:'\n @sugarcube_action = action.respond_to?('weak!') ? action.weak! : action\n end",
"def initialize(*args)\n super\n @action = :set\nend",
"def after_set_callback; end",
"def setup\n #implement in subclass;\n end",
"def lookup_action; end",
"def setup &block\n if block_given?\n @setup = block\n else\n @setup.call\n end\n end",
"def setup_action\n return TSBS.error(@acts[0], 1, @used_sequence) if @acts.size < 2\n actions = TSBS::AnimLoop[@acts[1]]\n if actions.nil?\n show_action_error(@acts[1])\n end\n @sequence_stack.push(@acts[1])\n @used_sequence = @acts[1]\n actions.each do |acts|\n @acts = acts\n execute_sequence\n break if @break_action\n end\n @sequence_stack.pop\n @used_sequence = @sequence_stack[-1]\n end",
"def release_actions; end",
"def around_hooks; end",
"def save_action; end",
"def setup(easy)\n super\n easy.customrequest = @verb\n end",
"def action_target()\n \n end",
"def setup\n callback(:setup) do\n notify(:setup)\n migration_check.last_deployed_commit\n end\n end",
"def setup\n return unless @setup\n\n actions = @setup['setup'].select { |action| action['do'] }.map { |action| Action.new(action['do']) }\n run_actions_and_retry(actions)\n self\n end",
"def before_setup\n # do nothing by default\n end",
"def my_actions(options)\n @setup = false\n get_template_part(\"custom_used\",\"action_users\",true)\n end",
"def default_action; end",
"def setup(&blk)\n @setup_block = blk\n end",
"def callback_phase\n super\n end",
"def advice\n end",
"def _handle_action_missing(*args); end",
"def duas1(action)\n action.call\n action.call\nend",
"def shared_action(name, &block)\n @controller.shared_actions[name] = block\n end",
"def before_action action, &block\n @audience[:before][action] ||= Set.new\n @audience[:before][action] << block\n end",
"def setup_initial_state\n\n state_a = State.new(\"a\", 0)\n state_b = State.new(\"b\", 0)\n state_c = State.new(\"c\", 10)\n\n move_to_b = Action.new(\"move_to_b\", 1, state_b)\n\n move_to_c = Action.new(\"move_to_c\", 1, state_c)\n\n state_a.actions = [move_to_b, move_to_c]\n\n return state_a\n \nend"
] |
[
"0.6163163",
"0.6045976",
"0.5946146",
"0.591683",
"0.5890051",
"0.58349305",
"0.5776858",
"0.5703237",
"0.5703237",
"0.5652805",
"0.5621621",
"0.54210985",
"0.5411113",
"0.5411113",
"0.5411113",
"0.5391541",
"0.53794575",
"0.5357573",
"0.53402257",
"0.53394014",
"0.53321576",
"0.53124547",
"0.529654",
"0.5296262",
"0.52952296",
"0.52600986",
"0.52442724",
"0.52385926",
"0.52385926",
"0.52385926",
"0.52385926",
"0.52385926",
"0.5232394",
"0.523231",
"0.5227454",
"0.52226824",
"0.52201617",
"0.5212327",
"0.52079266",
"0.52050185",
"0.51754695",
"0.51726824",
"0.51710224",
"0.5166172",
"0.5159343",
"0.51578903",
"0.51522785",
"0.5152022",
"0.51518047",
"0.51456624",
"0.51398855",
"0.5133759",
"0.5112076",
"0.5111866",
"0.5111866",
"0.5110294",
"0.5106169",
"0.509231",
"0.50873137",
"0.5081088",
"0.508059",
"0.50677156",
"0.50562143",
"0.5050554",
"0.50474834",
"0.50474834",
"0.5036181",
"0.5026331",
"0.5022976",
"0.5015441",
"0.50121695",
"0.5000944",
"0.5000019",
"0.4996878",
"0.4989888",
"0.4989888",
"0.49864885",
"0.49797225",
"0.49785787",
"0.4976161",
"0.49683493",
"0.4965126",
"0.4958034",
"0.49559742",
"0.4954353",
"0.49535993",
"0.4952725",
"0.49467874",
"0.49423352",
"0.49325448",
"0.49282882",
"0.49269363",
"0.49269104",
"0.49252945",
"0.4923091",
"0.49194667",
"0.49174926",
"0.49173003",
"0.49171105",
"0.4915879",
"0.49155936"
] |
0.0
|
-1
|
Never trust parameters from the scary internet, only allow the white list through.
|
def attachment_params
params.require(:attachment).permit(:file)
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def strong_params\n params.require(:user).permit(param_whitelist)\n end",
"def strong_params\n params.require(:listing_member).permit(param_whitelist)\n end",
"def allow_params_authentication!; end",
"def allowed_params\n ALLOWED_PARAMS\n end",
"def default_param_whitelist\n [\"mode\"]\n end",
"def param_whitelist\n [:role, :title]\n end",
"def expected_permitted_parameter_names; end",
"def safe_params\n params.except(:host, :port, :protocol).permit!\n end",
"def strong_params\n params.require(:team_member).permit(param_whitelist)\n end",
"def permitir_parametros\n \t\tparams.permit!\n \tend",
"def strong_params\n params.require(:community).permit(param_whitelist)\n end",
"def permitted_strong_parameters\n :all #or an array of parameters, example: [:name, :email]\n end",
"def strong_params\n params.require(:education).permit(param_whitelist)\n end",
"def restricted_params\n #params.require(self.controller_name.classify.underscore.to_sym).permit([])\n raise(\"No strong params set, override restricted_params method in your controller. E.g. params.require(:model).permit(:attribute1, :attribute2)\")\n end",
"def allowed_params\n params.require(:user).permit(:username, :email, :password, :password_confirmation)\n end",
"def param_whitelist\n [:rating, :review]\n end",
"def param_whitelist\n whitelist = [\n :username, :name,\n :parent_id,\n :headline, :description, :video,\n :policy, :signup_mode, :category,\n :website, :facebook, :twitter, :linkedin,\n :founded_at,\n privacy: [\n :events,\n :resources\n ],\n permission: [\n :profile,\n :members,\n :children,\n :statistics,\n :posts,\n :listings,\n :resources,\n :events\n ],\n location: [\n :description,\n :street,\n :city,\n :state,\n :zip,\n :country,\n :latitude,\n :longitude\n ]\n ]\n \n if action_name === 'update'\n whitelist.delete(:parent_id)\n unless current_user.role_in(@community) === 'owner'\n whitelist.delete(:privacy)\n whitelist.delete(:permission)\n end\n end\n \n whitelist\n end",
"def param_whitelist\n if @user.present? && current_user != @user\n return [:followed]\n end\n \n whitelist = [\n :username, :email, :password,\n :first_name, :last_name,\n :birthday, :gender,\n :headline, :biography, :ask_about, :focus,\n :website, :facebook, :linkedin, :twitter, :github,\n roles: [],\n skills: [],\n interests: [],\n privacy: { contact: [] },\n location: [\n :description,\n :street,\n :city,\n :state,\n :zip,\n :country,\n :latitude,\n :longitude\n ]\n ]\n \n if action_name === 'update'\n whitelist.delete(:email)\n whitelist.delete(:password)\n end\n \n whitelist\n end",
"def user_params \n \tparams.require(:user).permit(:name, :email, :password, :password_confirmation)# preventing CSTR\n end",
"def user_params\n params.permit(:name, :phoneNumber, :address, :postalCode, :local, :link, :counter, :latitude, :longitude) \n end",
"def valid_params_request?; end",
"def strong_params\n params.require(:experience).permit(param_whitelist)\n end",
"def trim_whitelisted(params, whitelist)\n # remove any parameters that are not whitelisted\n params.each do |key, value|\n # if white listed\n if whitelist.include? key\n # strip the parameters of any extra spaces, save as string\n params[key] = value.to_s.strip\n else\n # delete any unauthorized parameters\n params.delete key\n end\n end\n params\n end",
"def whitelist_url_params\n params.require(:whitelist_url).permit(:domain)\n end",
"def allowed_params\n params.require(:allowed).permit(:email)\n end",
"def permitted_params\n []\n end",
"def trim_whitelisted(params, whitelist)\n # remove any parameters that are not whitelisted\n params.each do |key, value|\n # if white listed\n if whitelist.include? key\n # strip the parameters of any extra spaces, save as string\n params[key] = value.to_s.strip\n else\n # delete any unauthorized parameters\n params.delete key\n end\n end\n params\n end",
"def safe_params\n params.permit(:id, :name, :origin, :emails => []); #emails is an array\n end",
"def query_param\n\t\tparams.permit(:first_name, :last_name, :phone)\n\tend",
"def strong_params\n params.require(:success_metric).permit(param_whitelist)\n end",
"def devise_filter\r\n logger.debug(\"In devise_filter =>PARAMS: #{params.inspect}\")\r\n\r\n # White list for sign_up\r\n devise_parameter_sanitizer.for(:sign_up) { |u| u.permit(user_whitelist) }\r\n\r\n # White list for account update\r\n devise_parameter_sanitizer.for(:account_update) { |u| u.permit(user_whitelist, :current_password) }\r\n\r\n # White list for Invitation creation\r\n devise_parameter_sanitizer.for(:invite) { |u| u.permit(:account_type, :email, :invitation_token)}\r\n\r\n # White list for accept invitation\r\n devise_parameter_sanitizer.for(:accept_invitation) { |u| u.permit(user_whitelist, :invitation_token)}\r\n\r\n end",
"def whitelisted_user_params\n params.require(:user).\n permit( :first_name, :last_name, :email,:password,:password_confirmation,:birthday,:gender)\n end",
"def user_params\n ActionController::Parameters.permit_all_parameters = true\n params.require(:user) #.permit(:name, :surname, :phone, :password, :email, :time_zone)\n end",
"def strong_params\n params.require(:metric_change).permit(param_whitelist)\n end",
"def safe_params\n params.require(:user).permit(:name)\n end",
"def get_params\n\t\treturn ActionController::Parameters.new(self.attributes).permit(\"account_id\", \"title\", \"category\", \"introduction\", \"tags\", \"segment_type\", \"visible\", \"status\", \"main_image\")\n\tend",
"def grant_params\n @whitelisted = params.require(:grant).permit(:name, :description, :agency_id, :acronym)\n end",
"def check_params; true; end",
"def param_whitelist\n whitelist = [\n :description,\n :progress,\n :kpi_id\n ]\n \n unless action_name === 'create'\n whitelist.delete(:kpi_id)\n end\n \n whitelist\n end",
"def quote_params\n params.permit!\n end",
"def valid_params?; end",
"def paramunold_params\n params.require(:paramunold).permit!\n end",
"def user_params\n\t\tparams.permit(:nickname, :avatar, :description, :password, :gender, :birthday, :email, :phone, :qq_id, :wechat_id)\n\tend",
"def filtered_parameters; end",
"def user_params\n params.permit(\n \t:id,\n \t:email, \n \t:first_name, \n \t:last_name, \n \t:password, \n \t:confirm_token, \n \t:phone_number,\n \t:facebook_link,\n \t:car_model,\n \t:license_plate)\n end",
"def filtering_params\n params.permit(:email, :name)\n end",
"def check_params\n true\n end",
"def wx_public_params\n params.require(:wx_public).permit(:nickname, :manager, :alias)\n end",
"def allowed_params\n params.require(:user).permit(:email, :password, :role, :first_name, :last_name, :password_confirmation)\n end",
"def allowed_params\n params.require(:user).permit(:email, :password, :role, :first_name, :last_name, :password_confirmation)\n end",
"def listing_params\n\t\tparams.permit(:address, :transit_info, :rules, :other_info, :lat, :lng)\n\tend",
"def social_account_params\n\t\t\tparams.require(:social_account).permit!\n\t\tend",
"def safe_params\n resurce_name = self.class.resource_name\n params_method_name = \"#{resurce_name}_params\".to_sym\n if params[resurce_name]\n if respond_to?(params_method_name) || private_methods.include?(params_method_name)\n send(params_method_name)\n else\n raise ActiveModel::ForbiddenAttributesError, \"Please, define the '#{params_method_name}' method in #{self.class.name}\"\n end\n end\n end",
"def url_params\n params.require(:url).permit(:short_url, :original_url, :clicks, :ip_addresses)\n end",
"def user_params\n params.require(:user).permit(:uri, :username, :password, :realname, :email, :publicvisible)\n end",
"def model_params\n\t\tparams.require(:manager).permit(\n\t :user_name,\n :password,\n :email,\n \t\t\t)\n\tend",
"def article_params_whitelist\n params.require(:article).permit(:title, :description, category_ids: [])\n end",
"def college_whitelist_params\n params.require(:college_whitelist).permit(:status)\n end",
"def active_code_params\n params[:active_code].permit\n end",
"def filtering_params\n params.permit(:email)\n end",
"def valid_params(params)\n params.permit(:user_id, :photo_id, :originX, :originY, :width, :height)\n end",
"def ip_address_params\n\t\t\tparams.require(:ip_address).permit!\n end",
"def pull_request_params\n whitelist = [\n :url,\n :id,\n :html_url,\n :diff_url,\n :patch_url,\n :issue_url,\n :number,\n :state,\n :locked,\n :title\n ]\n params.require(:pull_request).permit(whitelist)\n end",
"def reserved_params\n params.require(:reserved).permit(:name, :email, :pax, :address, :KTP, :title)\n end",
"def post_params\n if current_user.admin? \n params.permit(:title, :body, :city, :country, :gps_location, :privacy, :visible, :latitude, :longitude, images: [], files: [])\n else \n params.permit(:title, :body, :city, :country, :gps_location, :privacy,:latitude, :longitude, images: [], files: [])\n end \n end",
"def list_params\n params.permit(:name)\n end",
"def filter_parameters; end",
"def filter_parameters; end",
"def vineyard_params\n params.permit(:vineyard_name, :email, :website_url, :phone, :address, :city, :region, :postcode, :country, :specialty, :description, :pet_friendly, :holiday, :tours, :events, :family_friendly, :cover_image, :image_one, :image_two, :image_three, :image_four, :user_id, :base64)\n end",
"def available_activity_params\n # params.require(:available_activity).permit(:type,:geometry,:properties)\n whitelisted = ActionController::Parameters.new({\n type: params.require(:available_activity)[:type],\n geometry: params.require(:available_activity)[:geometry].try(:permit!).to_h,\n properties: params.require(:available_activity)[:properties].try(:permit!).to_h\n }).try(:permit!)\n end",
"def user_params\n params.permit(:name, :username, :email, :password, :img_url, :bg_url, :coinbank)\n end",
"def user_params_pub\n\t \tparams[:user].permit(:hruid)\n\t end",
"def user_params\n params.permit(:id, :email, :password, :nickname, :status, :avatar, :flat_picture, :flatsharing_id, :member,\n :user, :color, :solde)\n end",
"def validate_search_inputs\n @whitelisted = params.fetch(:user, nil)\n if @whitelisted.blank?\n render_error(400, \"#{I18n.t('general_error.params_missing_key')}\": [I18n.t('general_error.params_missing_value', model: \"review\")])\n return\n else\n @whitelisted = @whitelisted.permit(:name, :uen, :description)\n end\n end",
"def param_whitelist\n [\n :title,\n :description,\n :organization,\n :team_id,\n :started_at,\n :finished_at,\n location: [\n :description,\n :street,\n :city,\n :state,\n :zip,\n :country,\n :latitude,\n :longitude\n ]\n ]\n end",
"def url_whitelist; end",
"def admin_social_network_params\n params.require(:social_network).permit!\n end",
"def filter_params\n params.require(:filters).permit(:letters)\n end",
"def origin_params\n params.permit(:country, :state, :city, :postal_code, :address, :description)\n end",
"def valid_params(params)\n params.permit(:login, :first_name, :last_name, \n :password, :password_confirmation)\n end",
"def sensitive_params=(params)\n @sensitive_params = params\n end",
"def permit_request_params\n params.permit(:address)\n end",
"def user_params\n # Ensure a user can't give themselves admin priveleges\n params.delete(:admin) if current_user.admin?\n params.require(:user).permit(:name, :email, :admin, :image)\n end",
"def secure_params\n params.require(:location).permit(:name)\n end",
"def strong_params\n params.require( :setting ).\n permit( :global_scan_limit, :per_user_scan_limit,\n :target_whitelist_patterns, :target_blacklist_patterns )\n end",
"def question_params\n params.require(:survey_question).permit(question_whitelist)\n end",
"def case_insensitive_params\n params.require(:case_insensitive).permit(:name)\n end",
"def empire_master_no_match_params\n params.require(:empire_master_no_match).permit(:uid, :last_name, :list, :search_date, :double, :source)\n end",
"def maintenance_request_params\n params[:maintenance_request].permit! #allow all parameters for now\n end",
"def unwanted_params\n params.require(:unwanted).permit(:title, :description, :image)\n end",
"def url_params\n params[:url].permit(:full)\n end",
"def backend_user_params\n params.permit!\n end",
"def filter_params\n\t\treturn params[:candidate].permit(:name_for_filter)\n\tend",
"def speed_measurement_params\n\n #fuckit, to lazy to deal with permit crap right now\n ActionController::Parameters.permit_all_parameters = true\n\n params[:speed_measurement]\n end",
"def user_params\n params.permit(:name, :age, :username, :display_photo, :password)\n end",
"def get_params\r\n #params.require(:article).permit(:title, :permalink, :content, :source_site, :introtext, :type_id, :order_by, :searchable, :created_by, :edited_by, :published_by, :published_on, :user_id)\r\n params.require(:article).permit!\r\n\r\n end",
"def pub_params\n params.require(:pub).permit(:name, :description, :phone, :email, :hidden, :city_id, :address)\n end",
"def pass_params\n params[:pass].permit(:name, :price, :description, :colour, :events)\n end",
"def droptraining_params\n params.permit(:training_id,:user_id, :utf8, :authenticity_token, :commit)\n end",
"def person_params\n # params whitelist does *not* include admin, sub, remember_token\n # TBD: share this whitelist with the list used by configuration_permitted_parameters\n # TBD: should current_password be on this list? -- for now, leaving off, since it seems to work without\n # NOTE: do not include 'admin' in this list!\n params.require(:person).permit(\n :name, \n :email, \n :description,\n :password, \n :password_confirmation\n )\n end",
"def parameter_params\n params.require(:parameter).permit(:name, :description, :param_code, :param_value, :active_from, :active_to)\n end"
] |
[
"0.69792545",
"0.6781151",
"0.67419964",
"0.674013",
"0.6734356",
"0.6591046",
"0.6502396",
"0.6496313",
"0.6480641",
"0.6477825",
"0.64565",
"0.6438387",
"0.63791263",
"0.63740575",
"0.6364131",
"0.63192815",
"0.62991166",
"0.62978333",
"0.6292148",
"0.6290449",
"0.6290076",
"0.62894756",
"0.6283177",
"0.6242471",
"0.62382483",
"0.6217549",
"0.6214457",
"0.6209053",
"0.6193042",
"0.6177802",
"0.6174604",
"0.61714715",
"0.6161512",
"0.6151757",
"0.6150663",
"0.61461",
"0.61213595",
"0.611406",
"0.6106206",
"0.6105114",
"0.6089039",
"0.6081015",
"0.6071004",
"0.60620916",
"0.6019971",
"0.601788",
"0.6011056",
"0.6010898",
"0.6005122",
"0.6005122",
"0.6001556",
"0.6001049",
"0.59943926",
"0.5992201",
"0.59909594",
"0.5990628",
"0.5980841",
"0.59669393",
"0.59589154",
"0.5958826",
"0.5957911",
"0.5957385",
"0.5953072",
"0.59526145",
"0.5943361",
"0.59386164",
"0.59375334",
"0.59375334",
"0.5933856",
"0.59292704",
"0.59254247",
"0.5924164",
"0.59167904",
"0.59088355",
"0.5907542",
"0.59064597",
"0.5906243",
"0.5898226",
"0.589687",
"0.5896091",
"0.5894501",
"0.5894289",
"0.5891739",
"0.58860534",
"0.5882406",
"0.587974",
"0.58738774",
"0.5869024",
"0.58679986",
"0.5867561",
"0.5865932",
"0.5864461",
"0.58639693",
"0.58617616",
"0.5861436",
"0.5860451",
"0.58602303",
"0.5854586",
"0.58537364",
"0.5850427",
"0.5850199"
] |
0.0
|
-1
|
after dividing the larger number by the smaller number. Division by zero should return NaN. Arguments will both be integers. escribe 'Remainder Function' do it 'Should handle arguments and math as defined in specificaitons' do Test.assert_equals(remainder(17,5), 2, 'Returned value should be the value left over after dividing as much as possible.') Test.assert_equals(remainder(13, 72), remainder(72, 13), 'The order the arguments are passed should not matter.') Test.expect(remainder(1, 0).nil?, 'Divide by zero should return nil') Test.expect(remainder(0, 0).nil?, 'Divide by zero should return nil') end end
|
def my_remainder(a, b)
return nil if a == 0 || b == 0
return a.abs % b.abs if a.abs >= b.abs
return b.abs % a.abs if b.abs > a.abs
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def remainder(arg0)\n end",
"def remainder(val); end",
"def divmod(arg0)\n end",
"def divmod(arg0)\n end",
"def get_remainder(num1,num2)\n num1%num2\nend",
"def remainder(p0) end",
"def dec_remainder_of_two_integers(i_dividend, i_divisor)\n # your code goes here\n dec_remainder_of_two_floats(i_dividend.to_f, i_divisor.to_f)\nend",
"def dec_remainder_of_two_integers(i_dividend, i_divisor)\n # your code goes here\n return dec_remainder_of_two_floats(i_dividend.to_f, i_divisor.to_f)\nend",
"def dec_remainder_of_two_floats(f_dividend, f_divisor)\r\n # your code goes here\r\n f_dividend / f_divisor - (f_dividend / f_divisor).to_i\r\nend",
"def mod(a, b)\n a % b\n c = a % b\n puts \" the remainder of #{a} divided by #{b} is #{c}\"\nreturn c\nend",
"def dec_remainder_of_two_integers(i_dividend, i_divisor)\r\n # your code goes here\r\n f_dividend = i_dividend.to_f\r\n f_divisor = i_divisor.to_f\r\n dec_remainder_of_two_floats(f_dividend, f_divisor)\r\nend",
"def modulo(x, y)\n if y == 0\n return \"Psst. You can't divide by zero. Please try again.\"\n else\n return x % y\n end\nend",
"def divmod(val); end",
"def dec_remainder_of_two_floats(f_dividend, f_divisor)\n # your code goes here\n puts (f_divisor/f_dividend).round(1)\nend",
"def modulo(arg0)\n end",
"def modulo(arg0)\n end",
"def dec_remainder_of_two_floats(f_dividend, f_divisor)\n # your code goes here\n remainer = (f_dividend / f_divisor) % 10\n return remainer - (f_dividend.to_i / f_divisor.to_i)\nend",
"def divisor_plus_remainder(big_int, small_int)\n big_int / small_int + big_int % small_int\nend",
"def int_remainder_without_modulo(i_dividend, i_divisor)\r\n # take the remainder of the integers => .6\r\n # multiply that remainder by i_divisor => .6*5 = 3.0\r\n # convert the float to an integer\r\n x = dec_remainder_of_two_integers(i_dividend, i_divisor)\r\n y = x * (i_divisor)\r\n y.to_i\r\nend",
"def div(arg0)\n end",
"def mod(first_number, second_number)\n first_number % second_number\nend",
"def my_modulo(dividend, divisor)\r\n # your code goes here\r\n # use of modulo operator\r\n\r\n dividend % divisor\r\nend",
"def divmod(p0) end",
"def mod(num1, num2)\n num2.to_f % num1.to_f;\nend",
"def fdiv(arg0)\n end",
"def modulo(dividend, divisor)\n puts \"#{dividend} % #{divisor} = #{dividend.modulo(divisor)}\\n\"\nend",
"def divide(dividend, divisor)\n\n denom = divisor\n current = 0\n\n return 0 if denom > dividend\n return 1 if denom == dividend\n\n # 5 45\n while denom <= dividend do \n \tdenom += divisor\n \tcurrent += 1 \n end\n remainder = dividend - (denom - divisor)\n current\nend",
"def can_be_divided_by(n1, n2, n3)\n number_one_remainder = n1 % n2\n number_two_remainder = n1 % n3\n if number_one_remainder == 0\n puts \"Yes, you can divide by #{n2}\"\n else\n puts \"No, you can't divide by #{n2}\"\n end\n if number_two_remainder == 0\n puts \"Yes, you can divide by #{n3}\"\n else\n puts \"No, you can't divide by #{n3}\"\n end\nend",
"def divideme(first_number,second_number) # divides the numbers\r\n if second_number.to_i == 0\r\n puts \"Sorry cannot divide by zero\"\r\n else\r\n if first_number.to_i % second_number.to_i == 0\r\n puts (\"#{first_number} / #{second_number} is :\" + (first_number.to_i / second_number.to_i).to_s)\r\n else\r\n puts (\"#{first_number} / #{second_number} is :\" + (first_number.to_f / second_number.to_f).to_s)\r\n end\r\n end\r\nend",
"def rdiv(arg0)\n end",
"def divides?(a,b)\n return b%a == 0\nend",
"def divide_by_subtraction(dividend, divisor)\n #puts \"#{dividend}/#{divisor} is :\"\n quotient = 0 \n remainder = 0\n if dividend < divisor\n return [ 0, dividend]\n elsif divisor == 0 #can't do division by zero\n return nil\n else\n while dividend - divisor >= divisor\n dividend = dividend - divisor\n quotient += 1\n end\n \n if dividend != 0\n remainder = dividend - divisor\n quotient += 1\n end\n \n return [ quotient , remainder]\n end\n\n\nend",
"def test_10_div_0_is_bad\n\t\tassert_equal(\"can't do that, divide by something else.\", division(10,0))\n\tend",
"def modular_function\n num_1 % num_2\n return num_1 % num_2\nend",
"def divide(first_number, second_number)\n return first_number / second_number\nend",
"def divide(dividend, divisor)\r\n ((dividend < 0 && divisor >= 0) || (dividend >= 0 && divisor < 0)) ? negative = true : negative = false\r\n dividend = dividend.abs\r\n divisor = divisor.abs\r\n count = 0\r\n until dividend < divisor\r\n dividend -= divisor\r\n count += 1\r\n return negative ? 0 - (2**31) : 2**31 - 1 if count > 9999999\r\n end\r\n negative ? 0 - count : count\r\nend",
"def divide(dividend:, divisor:)\n dividend / divisor\nend",
"def divide(first_number, second_number)\n first_number / second_number\nend",
"def no_modulo(i_dividend, i_divisor)\n a = i_dividend / i_divisor\n b = a * i_divisor\n c = i_dividend - b\nend",
"def divisible?(number, divisor)\n number % divisor == 0\nend",
"def modulo_of(fraction); end",
"def is_divisor?(number, i) \n number % i == 0\nend",
"def modulus(d, e)\n puts \"MODULUS #{d} % #{e}\"\n return d % e\nend",
"def divisible_by(number, div_by)\n number % div_by == 0\nend",
"def self_dividing_numbers(left, right)\n (left..right).select{|x| x.digits.all?{|y| y!=0 && x%y == 0}}\nend",
"def divisible (first_number, second_number)\n if ((first_number % second_number) == 0 )\n puts \"#{first_number} is divisible by #{second_number}\"\n else\n puts \"#{first_number} is not divisible by #{second_number}\"\n end\nend",
"def division (first_number, second_number)\nfirst_number / second_number\nend",
"def divisible_by?(num, denominator)\n num % denominator == 0\nend",
"def func_mod(args)\n p1 = _eval(car(args))\n p2 = _eval(car(cdr(args)))\n\n if p1.type != LObject::OBJ_INTEGER or p2.type != LObject::OBJ_INTEGER\n if @lint\n Error.warn(\"warning: modulo with a non integer operand\")\n end\n return @o_man.nil\n end\n\n p3 = @o_man.new_object(LObject::OBJ_INTEGER)\n if p2.value.i == 0\n if @lint\n Error.warn(\"warning: modulo by zero\")\n end\n p3.value.i = 0\n else\n p3.value.i = p1.value.i % p2.value.i\n end\n\n return p3\n end",
"def div(x, y)\n x / y\nend",
"def divisible?(num, divisor)\n num % divisor == 0\nend",
"def modulus(a, b)\n return \"If you divide #{a} and #{b} you will get the remainer #{a%b}.\"\nend",
"def divide(number_1, number_2)\n return number_1 / number_2\nend",
"def divide(number_1, number_2)\n return number_1 / number_2\nend",
"def divide(number_1, number_2)\n return number_1 / number_2\nend",
"def divide(dividend, divisor)\n # new_dividend = dividend > 0 : -divident : dividend\n # new_divisor = divisor > 0 : -divisor : divisor\n\n new_dividend = dividend\n new_divisor = divisor\n\n quotient = 0\n while new_dividend >= new_divisor\n multiplier = 0\n while (new_divisor << multiplier + 1) < new_dividend\n multiplier += 1\n end\n\n quotient += (1 << multiplier)\n new_dividend -= new_divisor << multiplier\n end\n\n quotient\nend",
"def division_plus_remainder(big_number, small_number)\n (big_number / small_number) + (big_number % small_number)\n end",
"def divideThem(numOne, numTwo)\n numOne / numTwo\nend",
"def divide(number, divisor)\n begin\n answer = number / divisor\n rescue ZeroDivisionError => e\n puts e.message\n end\nend",
"def divide(number, divisor)\n begin\n answer = number / divisor\n rescue ZeroDivisionError => e\n puts e.message\n end\nend",
"def divide(number, divisor)\n begin\n answer = number / divisor\n rescue ZeroDivisionError => e\n puts e.message\n end\nend",
"def is_odd_using_remainder(int)\n int.remainder(2).abs == 1\nend",
"def modulus\n\tputs (5%3)\n\tputs ( -5 % 3)\n\tputs (5 % 3)\n\tputs (-5 % -3)\n\nend",
"def divisable_by_given?(number, given)\n number % given == 0\n end",
"def divide(dividend, divisor)\n negative_sign = (dividend < 0 && divisor >= 0) || (dividend >= 0 && divisor < 0)\n\n remaining = dividend.abs\n abs_divisor = divisor.abs\n quotient = 0\n while remaining >= abs_divisor\n remaining -= abs_divisor\n quotient += 1\n end\n\n negative_sign ? \"-#{quotient.to_s}\".to_i : quotient\nend",
"def divide(number, divisor)\n begin\n answer = number / divisor\n rescue ZeroDivisionError => e \n puts e.message\n end\nend",
"def divide(x, y)\n if x == 0 || y == 0\n return \"We dont do math on zeros\"\n end\n\n x / y\nend",
"def safeDivide( top, bottom )\n if( (top == nil) or ( bottom == nil ) or ( bottom == 0.0 )) then\n return 0.0;\n end\n return top/bottom;\nend",
"def divide(x, y)\n if y == 0\n return \"Psst. You can't divide by zero. Please try again.\"\n elsif x.is_a?(Integer) && y.is_a?(Integer) && x % y != 0\n return x.to_f / y\n else\n return x / y\n end\nend",
"def divide(number_one, number_two)\nnumber_one / number_two\nend",
"def p003(total_remaining = 600_851_475_143)\n divisor = 2\n until divisor >= total_remaining\n if (total_remaining % divisor).zero?\n total_remaining /= divisor\n else\n divisor += 1\n end\n end\n divisor\nend",
"def multiple?(number, divisor)\n number % divisor == 0\nend",
"def multiple?(number, divisor)\n number % divisor == 0\nend",
"def multiple?(number, divisor)\n number % divisor == 0\nend",
"def multiple?(number, divisor)\n number % divisor == 0\nend",
"def div_by(x,y)\n begin\n x / y\n rescue\n puts \"oops, what happened?\"\n retry\n end\nend",
"def modulo(p0) end",
"def divide(a,b,num_of_sig_figs=10)\n #output will be our answer\n output = 0\n #we loop until our dividend is smaller than our divisor\n while a >= b\n # we subtract the divisor from the dividend and the difference is our new dividend\n a -= b\n #we increase the output by 1\n output += 1\n end\n if a == 0\n #if there is no remainder we return the output\n output\n else\n #if there is a remainder we need to change the output to a string\n output = output.to_s\n #sig_figs keeps track of the significant figures we've made\n sig_figs = output.length\n #remainder_args takes an array of three arguments - the dividend, divisor and the final answer string which ends with a . as we will be adding decimals\n remainder_args = [a,b,output + '.']\n #the default number of significant figures is 10. The loop runs until we hit the number of significant figures or if there is no remainder left\n while sig_figs < num_of_sig_figs && remainder_args[0] != 0\n #we have to create a new function which we will call recursively to add any new digits after the decimal point. We must call it on an array because otherwise we cannot change the variables due to Ruby's scope\n def remainder_divide(arr)\n #we multiply the dividend by 10 to make it a non decimal number\n arr[0] = multiply(arr[0], 10)\n #this loop is the same as our original division\n output = 0\n while arr[0] >= arr[1]\n arr[0] -= arr[1]\n output += 1\n end\n #here we change the output to a string and attach it to our original answer.\n output = output.to_s\n arr[2] += output\n #We change the remainder_args to our new array\n remainder_args = [arr[0],arr[1],arr[2]]\n end\n #now that we've defined our function, we will call it on our array of arguments until we hit the desired amount of significant figures or we have the answer\n remainder_divide(remainder_args)\n #we add one to our significant figures so that we don't loop infinitely\n sig_figs += 1\n end\n #we return our answer converted back to a number\n remainder_args[2].to_f\n end\nend",
"def remainder(val)\n a, b = self, val\n b = b.exchange_to(a.currency) if b.is_a?(Money) and a.currency != b.currency\n\n a_sign, b_sign = :pos, :pos\n a_sign = :neg if a.cents < 0\n b_sign = :neg if (b.is_a?(Money) and b.cents < 0) or (b < 0)\n\n return a.modulo(b) if a_sign == b_sign\n a.modulo(b) - (b.is_a?(Money) ? b : Money.new(b, a.currency))\n end",
"def divide(a,b)\n raise Exception.new(\"Undefined. Divisor is 0\") if b == 0\n sign = 1\n result = -1\n\n # Handle corner cases\n if a < 0\n a *= -1\n sign *= -1\n end\n\n if b < 0\n b *= -1\n sign *= -1\n end\n\n while a >= 0\n a -= b\n result += 1\n end\n result * sign\nend",
"def gcd( a, b )\n return a if b.zero?\n gcd b, a % b\nend",
"def divisible_by number, divisor\n number % divisor == 0\n end",
"def divide(number, divisor)\n begin\n answer = number / divisor\n rescue ZeroDivisionError => e\n puts \"--error--\", e.message, \"-----------\"\n end\nend",
"def divide(num1, num2)\n return num1 / num2\nend",
"def test_divide_by_zero\n calc = RpnCalc.new\n calc.enter(1)\n calc.enter(0)\n assert_raise(ZeroDivisionError) do\n calc.enter('/')\n end\n end",
"def dec_remainder_two_ints(int1, int2)\n f_dividend = int1.to_f\n f_divisor = int2.to_f\n return dec_remainder_two_floats(f_dividend, f_divisor)\nend",
"def test_13_multiple_integers_and_floats\r\n\r\n results = divide(284.182, 3.2, 19, 0.01928, 39.2944)\r\n\r\n assert_equal(6.16958, results)\r\n\r\n end",
"def divide(a, b)\n puts \"DIVIDING #{a} / #{b}\"\n return a / b\n a + b # This is ignored as the output due to the return\nend",
"def divide_integers(dividend, divisor)\n return 1 if dividend == divisor\n \n quotient = 0\n counter = divisor.abs\n until dividend.abs < counter\n quotient += 1\n counter += divisor.abs\n end\n\n if dividend > 0 && divisor > 0 || dividend < 0 && divisor < 0\n quotient\n else\n quotient * -1\n end\nend",
"def number_is_divisible_by?(number, divisor)\n number % divisor == 0\nend",
"def calculate_remainder(amount:, total:)\n 100 - calculate_percentage(amount: amount, total: total)\n end",
"def divide(a, b)\n divide = a / b\n return divide\nend",
"def divide(first_number, second_number)\n return first_number / second_number\n end",
"def divide(a,b)\n a/b\nend",
"def divisor_game(n)\n n%2==0\nend",
"def div(a, b) #=> funkcja wykonujaca dzielenie\n\n\tif b == 0\n\n\t\tputs \"Nie dzielimy przez 0!\"\n\t\treturn\n\n\telse\n\n\t\tputs \"Wynik: #{a/b}\"\n\n\tend\n\n\tputs \"Dzialanie wykonane prawidlowo\" #=> ta informacja nie pojawi sie w przypadku dzielenia przez 0\n\nend",
"def is_odd_remainder?(num)\n num.abs.remainder(2) == 1\n # x.remainder(y) means x-y*(x/y).truncate.\nend",
"def divisione(a, b)\n puts \"DIVIDENDO #{a} / #{b}\"\n a / b\nend",
"def divide(num1, num2)\n\tresults = \"\"\n\tif num2 == 0\n\t\t\"False\"\n\telse\n\tresults = num1/num2\nend\n\n\t\t\t\n\t# if number == 0\n\t# \treturn \"Cant't divide a number by 0\"\n\t# else numbers.inject do |divedend, number|\n\t# quo = (divedend /= number).round(2)\n\n\n\t# end\n\n\t# if number.zero?\n # \treturn \"Cannot divide by 0!\"\n\t# else\n # \tquo\n\t# end\n\n\t\n\t# if number != 0 \n\t\t# \treturn quo\n\t# else \n\t\t# \treturn \"error\"\n\t# end\n\n\t# end\n\nend",
"def div_stringed(num1, num2)\r\n\r\nend"
] |
[
"0.76481676",
"0.7472466",
"0.74267524",
"0.74267524",
"0.7393483",
"0.71505165",
"0.71389467",
"0.7118643",
"0.7096791",
"0.698439",
"0.69758356",
"0.696419",
"0.69162863",
"0.69068396",
"0.6879262",
"0.6879262",
"0.6877722",
"0.6762794",
"0.6698192",
"0.668713",
"0.6686745",
"0.66530424",
"0.65963626",
"0.65818125",
"0.6555842",
"0.6534977",
"0.6529534",
"0.6523108",
"0.6503091",
"0.65021694",
"0.6482304",
"0.64776397",
"0.6457942",
"0.64358294",
"0.64104986",
"0.6388235",
"0.6386065",
"0.6353669",
"0.633478",
"0.6333406",
"0.632471",
"0.63203347",
"0.6303696",
"0.6284857",
"0.6278005",
"0.62746453",
"0.62698615",
"0.62546176",
"0.62526083",
"0.62509304",
"0.62423754",
"0.6240761",
"0.623979",
"0.623979",
"0.623979",
"0.62397546",
"0.62270194",
"0.62169796",
"0.62110186",
"0.62110186",
"0.62110186",
"0.6204677",
"0.6196755",
"0.6183917",
"0.6180897",
"0.61748254",
"0.61733186",
"0.6157737",
"0.61549973",
"0.61528516",
"0.61527455",
"0.6151076",
"0.6151076",
"0.6151076",
"0.6151076",
"0.61433643",
"0.6140798",
"0.61401755",
"0.61263895",
"0.61089987",
"0.61053365",
"0.61018795",
"0.6098991",
"0.60823995",
"0.6081453",
"0.60691833",
"0.60536003",
"0.60514164",
"0.603459",
"0.60216683",
"0.60162437",
"0.60121536",
"0.6006578",
"0.60041845",
"0.5993433",
"0.5989487",
"0.5986125",
"0.59822696",
"0.59806603",
"0.59755206"
] |
0.73172474
|
5
|
Use callbacks to share common setup or constraints between actions.
|
def set_pagamento
@pagamento = Pagamento.find(params[:id])
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def set_required_actions\n # TODO: check what fields change to asign required fields\n end",
"def action_hook; end",
"def run_actions; end",
"def define_action_hook; end",
"def actions; end",
"def define_action_helpers\n if super && action == :save\n @instance_helper_module.class_eval do\n define_method(:valid?) do |*args|\n self.class.state_machines.fire_event_attributes(self, :save, false) { super(*args) }\n end\n end\n end\n end",
"def add_actions; end",
"def callbacks; end",
"def callbacks; end",
"def setup *actions, &proc\n (@setup_procs ||= []) << [proc, actions.size > 0 ? actions : [:*]]\n end",
"def define_action_helpers; end",
"def post_setup\n end",
"def action_methods; end",
"def action_methods; end",
"def action_methods; end",
"def before_setup; end",
"def action_run\n end",
"def execute(setup)\n @action.call(setup)\n end",
"def define_action_helpers?; end",
"def set_actions\n actions :all\n end",
"def action_done(action)\n dispatch = { :migrate => :done_migrating, :map => :done_mapping, :reduce =>\n :done_reducing, :finalize => :done_finalizing } \n self.send dispatch[action[:action]], action\n end",
"def dependencies action, &block\n @actions.each do |other|\n if action[:requires].include? other[:provide]\n block.call other\n end\n end\n end",
"def setup!\n return unless @setup_procs\n http_actions = actions\n @setup_procs.each do |setup_proc|\n proc, actions = setup_proc\n @setup__actions = actions.map do |action|\n\n action.is_a?(Regexp) ?\n http_actions.select { |a| a.to_s =~ action } :\n action.is_a?(String) && action =~ /\\A\\./ ?\n http_actions.map { |a| a.to_s << action if format?(a).include?(action) }.compact :\n action\n\n end.flatten\n self.class_exec &proc\n @setup__actions = nil\n end\n @setup_procs = nil\n end",
"def before_actions(*logic)\n self.before_actions = logic\n end",
"def setup_handler\n end",
"def set_action(opts)\n opts = check_params(opts,[:actions])\n super(opts)\n end",
"def setup(action)\n @targets.clear\n unless action.item.target_filters.empty?\n @targets = SES::TargetManager.make_targets(action)\n else\n item = action.item\n if item.for_opponent?\n @targets = $game_troop.alive_members\n elsif item.for_dead_friend?\n @targets = $game_party.battle_members.select { |actor| actor.dead? }\n else\n $game_party.battle_members.select { |actor| actor.alive? }\n end\n end\n @item_max = @targets.size\n create_contents\n refresh\n show\n activate\n end",
"def action; end",
"def action; end",
"def action; end",
"def action; end",
"def action; end",
"def workflow\n end",
"def revisable_shared_setup(args, block)\n class << self\n attr_accessor :revisable_options\n end\n options = args.extract_options!\n self.revisable_options = Options.new(options, &block)\n \n self.send(:include, Common)\n self.send(:extend, Validations) unless self.revisable_options.no_validation_scoping?\n self.send(:include, WithoutScope::QuotedColumnConditions)\n end",
"def setup\n @action = SampleActionAndroid.new(os_name: 'android',\n app_name: APP_PATH)\n end",
"def before(action)\n invoke_callbacks *self.class.send(action).before\n end",
"def process_action(...)\n send_action(...)\n end",
"def before_dispatch(env); end",
"def after_actions(*logic)\n self.after_actions = logic\n end",
"def setup\n # override and do something appropriate\n end",
"def setup(client)\n return unless @setup\n actions = @setup['setup'].select { |action| action['do'] }.map { |action| Action.new(action['do']) }\n actions.each do |action|\n action.execute(client)\n end\n self\n end",
"def setup(_context)\n end",
"def setup(resources) ; end",
"def validate_actions\n errors.add(:base, :should_give_at_least_one_action) if !manage? && !forecasting? && !read? && !api?\n end",
"def setup\n @resource_config = {\n :callbacks => {\n :before_create => nil,\n :after_create => nil,\n :before_update => nil,\n :after_update => nil,\n :before_destroy => nil,\n :after_destroy => nil,\n },\n :child_assoc => nil,\n :model => nil,\n :parent => nil,\n :path => nil,\n :permission => {},\n :properties => {},\n :relation => {\n :create => nil,\n :delete => nil,\n },\n :roles => nil,\n }\n end",
"def determine_valid_action\n\n end",
"def process_shared\n handle_taxes\n handle_shippings\n create_adjustments_from_params\n handle_status\n handle_inventory_refunds\n handle_payment_transactions\n order.updater.update\n end",
"def startcompany(action)\n @done = true\n action.setup\n end",
"def init_actions\n am = action_manager()\n am.add_action(Action.new(\"&Disable selection\") { @selection_mode = :none; unbind_key(32); bind_key(32, :scroll_forward); } )\n am.add_action(Action.new(\"&Edit Toggle\") { @edit_toggle = !@edit_toggle; $status_message.value = \"Edit toggle is #{@edit_toggle}\" })\n end",
"def event_callbacks(event, metadata={})\n case event\n when :reset, :review\n if confirmed\n update_attributes(confirmed: false)\n end\n when :confirm\n confirm\n # trigger :order for all applicable items\n # NOTE: :order event is common to both physical and digital items\n items.each do |i|\n if i.event_permitted(:order)\n user_id = last_transition.user_id\n i.trigger!(:order, { order_id: id, user_id: user_id })\n end\n end\n when :complete_work\n request = metadata[:request]\n work_complete_notification(request)\n when :close\n close\n end\n if event != :close && !open\n reopen\n end\n end",
"def setup_action\n return unless PONY::ERRNO::check_sequence(current_act)\n new_sequence = @action_sequence[@sequence_index+1...@action_sequence.size]\n @sequence_index = 0\n new_sequence = DND::SkillSequence::ACTS[@acts[1]] + new_sequence\n execute_sequence\n end",
"def define_tasks\n define_weave_task\n connect_common_tasks\n end",
"def setup(&block)\n define_method(:setup, &block)\n end",
"def setup\n transition_to(:setup)\n end",
"def setup\n transition_to(:setup)\n end",
"def action\n end",
"def setup( *args )\n\t\t\tself.class.setupBlocks.each {|sblock|\n\t\t\t\tdebugMsg \"Calling setup block method #{sblock}\"\n\t\t\t\tself.send( sblock )\n\t\t\t}\n\t\t\tsuper( *args )\n\t\tend",
"def config(action, *args); end",
"def setup\n @setup_proc.call(self) if @setup_proc\n end",
"def before_action \n end",
"def setup_callbacks\n defined_callbacks.each do |meth|\n unless respond_to?(\"call_#{meth}_callbacks\".to_sym)\n self.class.module_eval <<-EOE\n def call_#{meth}_callbacks(*args)\n plugin_store.each {|a| a.call_#{meth}_callbacks(*args) } if respond_to?(:plugin_store) && plugin_store\n self.send :#{meth}, *args if respond_to?(:#{meth})\n end\n EOE\n end\n end\n end",
"def action\n end",
"def matt_custom_action_begin(label); end",
"def setup\n # override this if needed\n end",
"def setup\n\t\t\t\t\t\t# Do nothing\n\t\t\t\tend",
"def setup\n\t\t\t\t\t\t# Do nothing\n\t\t\t\tend",
"def action(options,&callback)\n new_action = Action===options ? options : Action.new(options,&callback)\n # replace any with (shared name/alias or both default) + same arity\n @actions.delete_if do |existing_action|\n ((existing_action.names & new_action.names).size > 0 ||\n existing_action.default? && new_action.default?) &&\n existing_action.required.size == new_action.required.size &&\n existing_action.optional.size <= new_action.optional.size\n end\n @actions = (@actions + [new_action]).sort\n new_action\n end",
"def set_target_and_action target, action\n self.target = target\n self.action = 'sugarcube_handle_action:'\n @sugarcube_action = action\n end",
"def after(action)\n invoke_callbacks *options_for(action).after\n end",
"def pre_task\n end",
"def setup(server)\n server.on('beforeMethod', method(:before_method), 10)\n end",
"def add_actions\n attribute = machine.attribute\n name = self.name\n \n owner_class.class_eval do\n define_method(name) {self.class.state_machines[attribute].events[name].fire(self)}\n define_method(\"#{name}!\") {self.class.state_machines[attribute].events[name].fire!(self)}\n define_method(\"can_#{name}?\") {self.class.state_machines[attribute].events[name].can_fire?(self)}\n end\n end",
"def init_actions\n @select_action = SelectAction.new\n @endpoint_mouse_action = EndpointMouseAction.new\n @move_action = MoveAction.new\n end",
"def setup_signals; end",
"def after_created\r\n return unless compile_time\r\n Array(action).each do |action|\r\n run_action(action)\r\n end\r\nend",
"def after_created\r\n return unless compile_time\r\n Array(action).each do |action|\r\n run_action(action)\r\n end\r\nend",
"def set_target_and_action target, action\n self.target = target\n self.action = 'sugarcube_handle_action:'\n @sugarcube_action = action.respond_to?('weak!') ? action.weak! : action\n end",
"def initialize(*args)\n super\n @action = :set\nend",
"def after_set_callback; end",
"def setup\n #implement in subclass;\n end",
"def lookup_action; end",
"def setup &block\n if block_given?\n @setup = block\n else\n @setup.call\n end\n end",
"def setup_action\n return TSBS.error(@acts[0], 1, @used_sequence) if @acts.size < 2\n actions = TSBS::AnimLoop[@acts[1]]\n if actions.nil?\n show_action_error(@acts[1])\n end\n @sequence_stack.push(@acts[1])\n @used_sequence = @acts[1]\n actions.each do |acts|\n @acts = acts\n execute_sequence\n break if @break_action\n end\n @sequence_stack.pop\n @used_sequence = @sequence_stack[-1]\n end",
"def release_actions; end",
"def around_hooks; end",
"def save_action; end",
"def setup(easy)\n super\n easy.customrequest = @verb\n end",
"def action_target()\n \n end",
"def setup\n callback(:setup) do\n notify(:setup)\n migration_check.last_deployed_commit\n end\n end",
"def setup\n return unless @setup\n\n actions = @setup['setup'].select { |action| action['do'] }.map { |action| Action.new(action['do']) }\n run_actions_and_retry(actions)\n self\n end",
"def before_setup\n # do nothing by default\n end",
"def my_actions(options)\n @setup = false\n get_template_part(\"custom_used\",\"action_users\",true)\n end",
"def default_action; end",
"def setup(&blk)\n @setup_block = blk\n end",
"def callback_phase\n super\n end",
"def advice\n end",
"def _handle_action_missing(*args); end",
"def duas1(action)\n action.call\n action.call\nend",
"def shared_action(name, &block)\n @controller.shared_actions[name] = block\n end",
"def before_action action, &block\n @audience[:before][action] ||= Set.new\n @audience[:before][action] << block\n end",
"def setup_initial_state\n\n state_a = State.new(\"a\", 0)\n state_b = State.new(\"b\", 0)\n state_c = State.new(\"c\", 10)\n\n move_to_b = Action.new(\"move_to_b\", 1, state_b)\n\n move_to_c = Action.new(\"move_to_c\", 1, state_c)\n\n state_a.actions = [move_to_b, move_to_c]\n\n return state_a\n \nend"
] |
[
"0.6163163",
"0.6045976",
"0.5946146",
"0.591683",
"0.5890051",
"0.58349305",
"0.5776858",
"0.5703237",
"0.5703237",
"0.5652805",
"0.5621621",
"0.54210985",
"0.5411113",
"0.5411113",
"0.5411113",
"0.5391541",
"0.53794575",
"0.5357573",
"0.53402257",
"0.53394014",
"0.53321576",
"0.53124547",
"0.529654",
"0.5296262",
"0.52952296",
"0.52600986",
"0.52442724",
"0.52385926",
"0.52385926",
"0.52385926",
"0.52385926",
"0.52385926",
"0.5232394",
"0.523231",
"0.5227454",
"0.52226824",
"0.52201617",
"0.5212327",
"0.52079266",
"0.52050185",
"0.51754695",
"0.51726824",
"0.51710224",
"0.5166172",
"0.5159343",
"0.51578903",
"0.51522785",
"0.5152022",
"0.51518047",
"0.51456624",
"0.51398855",
"0.5133759",
"0.5112076",
"0.5111866",
"0.5111866",
"0.5110294",
"0.5106169",
"0.509231",
"0.50873137",
"0.5081088",
"0.508059",
"0.50677156",
"0.50562143",
"0.5050554",
"0.50474834",
"0.50474834",
"0.5036181",
"0.5026331",
"0.5022976",
"0.5015441",
"0.50121695",
"0.5000944",
"0.5000019",
"0.4996878",
"0.4989888",
"0.4989888",
"0.49864885",
"0.49797225",
"0.49785787",
"0.4976161",
"0.49683493",
"0.4965126",
"0.4958034",
"0.49559742",
"0.4954353",
"0.49535993",
"0.4952725",
"0.49467874",
"0.49423352",
"0.49325448",
"0.49282882",
"0.49269363",
"0.49269104",
"0.49252945",
"0.4923091",
"0.49194667",
"0.49174926",
"0.49173003",
"0.49171105",
"0.4915879",
"0.49155936"
] |
0.0
|
-1
|
Only allow a trusted parameter "white list" through.
|
def pagamento_params
params.require(:pagamento).permit(:id, :valor, :state, :data_vencimento, :data_pagamento, :evento_financeiro_id)
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def allowed_params\n ALLOWED_PARAMS\n end",
"def expected_permitted_parameter_names; end",
"def param_whitelist\n [:role, :title]\n end",
"def default_param_whitelist\n [\"mode\"]\n end",
"def permitir_parametros\n \t\tparams.permit!\n \tend",
"def permitted_params\n []\n end",
"def strong_params\n params.require(:user).permit(param_whitelist)\n end",
"def strong_params\n params.require(:listing_member).permit(param_whitelist)\n end",
"def filtered_parameters; end",
"def permitted_strong_parameters\n :all #or an array of parameters, example: [:name, :email]\n end",
"def parameters_list_params\n params.require(:parameters_list).permit(:name, :description, :is_user_specific)\n end",
"def parameter_params\n params.require(:parameter).permit(:name, :description, :param_code, :param_value, :active_from, :active_to)\n end",
"def param_whitelist\n whitelist = [\n :description,\n :progress,\n :kpi_id\n ]\n \n unless action_name === 'create'\n whitelist.delete(:kpi_id)\n end\n \n whitelist\n end",
"def param_whitelist\n whitelist = [\n :username, :name,\n :parent_id,\n :headline, :description, :video,\n :policy, :signup_mode, :category,\n :website, :facebook, :twitter, :linkedin,\n :founded_at,\n privacy: [\n :events,\n :resources\n ],\n permission: [\n :profile,\n :members,\n :children,\n :statistics,\n :posts,\n :listings,\n :resources,\n :events\n ],\n location: [\n :description,\n :street,\n :city,\n :state,\n :zip,\n :country,\n :latitude,\n :longitude\n ]\n ]\n \n if action_name === 'update'\n whitelist.delete(:parent_id)\n unless current_user.role_in(@community) === 'owner'\n whitelist.delete(:privacy)\n whitelist.delete(:permission)\n end\n end\n \n whitelist\n end",
"def param_whitelist\n [:rating, :review]\n end",
"def valid_params?; end",
"def permitted_params\n declared(params, include_missing: false)\n end",
"def permitted_params\n declared(params, include_missing: false)\n end",
"def get_params\n\t\treturn ActionController::Parameters.new(self.attributes).permit(\"account_id\", \"title\", \"category\", \"introduction\", \"tags\", \"segment_type\", \"visible\", \"status\", \"main_image\")\n\tend",
"def filter_parameters; end",
"def filter_parameters; end",
"def strong_params\n params.require(:team_member).permit(param_whitelist)\n end",
"def strong_params\n params.require(:community).permit(param_whitelist)\n end",
"def check_params; true; end",
"def valid_params_request?; end",
"def strong_params\n params.require(:experience).permit(param_whitelist)\n end",
"def allowed_params\n params.require(:user).permit(:username, :email, :password, :password_confirmation)\n end",
"def list_params\n params.permit(:name)\n end",
"def check_params\n true\n end",
"def grant_params\n @whitelisted = params.require(:grant).permit(:name, :description, :agency_id, :acronym)\n end",
"def safe_params\n resurce_name = self.class.resource_name\n params_method_name = \"#{resurce_name}_params\".to_sym\n if params[resurce_name]\n if respond_to?(params_method_name) || private_methods.include?(params_method_name)\n send(params_method_name)\n else\n raise ActiveModel::ForbiddenAttributesError, \"Please, define the '#{params_method_name}' method in #{self.class.name}\"\n end\n end\n end",
"def additional_permitted_params\n []\n end",
"def strong_params\n params.require(:education).permit(param_whitelist)\n end",
"def resource_params\n params[resource_singular_name].try(:permit, self.class.param_whitelist)\n end",
"def allow_params_authentication!; end",
"def param_whitelist\n [\n :title,\n :description,\n :organization,\n :team_id,\n :started_at,\n :finished_at,\n location: [\n :description,\n :street,\n :city,\n :state,\n :zip,\n :country,\n :latitude,\n :longitude\n ]\n ]\n end",
"def param_whitelist\n if @user.present? && current_user != @user\n return [:followed]\n end\n \n whitelist = [\n :username, :email, :password,\n :first_name, :last_name,\n :birthday, :gender,\n :headline, :biography, :ask_about, :focus,\n :website, :facebook, :linkedin, :twitter, :github,\n roles: [],\n skills: [],\n interests: [],\n privacy: { contact: [] },\n location: [\n :description,\n :street,\n :city,\n :state,\n :zip,\n :country,\n :latitude,\n :longitude\n ]\n ]\n \n if action_name === 'update'\n whitelist.delete(:email)\n whitelist.delete(:password)\n end\n \n whitelist\n end",
"def person_params\n # params whitelist does *not* include admin, sub, remember_token\n # TBD: share this whitelist with the list used by configuration_permitted_parameters\n # TBD: should current_password be on this list? -- for now, leaving off, since it seems to work without\n # NOTE: do not include 'admin' in this list!\n params.require(:person).permit(\n :name, \n :email, \n :description,\n :password, \n :password_confirmation\n )\n end",
"def paramunold_params\n params.require(:paramunold).permit!\n end",
"def param_params\n params.require(:param).permit(:param_category_id, :param_table_id, :name, :english_name, :weighting, :description)\n end",
"def quote_params\n params.permit!\n end",
"def list_params\n params.permit(:list_name)\n end",
"def allowed_params(parameters)\n parameters.select do |name, values|\n values.location != \"path\"\n end\n end",
"def all_params; end",
"def permitted_resource_params\n params[resource.object_name].present? ? params.require(resource.object_name).permit! : ActionController::Parameters.new\n end",
"def source_params\n params.require(:source).permit(all_allowed_params)\n end",
"def user_params\n end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def params; end",
"def get_allowed_parameters\n return _get_specific_action_config(:allowed_action_parameters, :allowed_parameters)&.map(&:to_s)\n end",
"def permitted_params\n @wfd_edit_parameters\n end",
"def user_params\r\n end",
"def param_whitelist\n whitelist = [\n :comment,\n :old_progress, :new_progress,\n :metric_id\n ]\n \n unless action_name === 'create'\n whitelist.delete(:metric_id)\n end\n \n whitelist\n end",
"def query_param\n\t\tparams.permit(:first_name, :last_name, :phone)\n\tend",
"def whitelisted_user_params\n params.require(:user).\n permit( :first_name, :last_name, :email,:password,:password_confirmation,:birthday,:gender)\n end",
"def filter_params\n\t\treturn params[:candidate].permit(:name_for_filter)\n\tend",
"def user_params\n params.permit(:id, :email, :password, :nickname, :status, :avatar, :flat_picture, :flatsharing_id, :member,\n :user, :color, :solde)\n end",
"def get_params\n\t\t\n\t\treturn ActionController::Parameters.new(self.attributes).permit(:first_name, :last_name, :email, :provider)\n\n\tend",
"def devise_filter\r\n logger.debug(\"In devise_filter =>PARAMS: #{params.inspect}\")\r\n\r\n # White list for sign_up\r\n devise_parameter_sanitizer.for(:sign_up) { |u| u.permit(user_whitelist) }\r\n\r\n # White list for account update\r\n devise_parameter_sanitizer.for(:account_update) { |u| u.permit(user_whitelist, :current_password) }\r\n\r\n # White list for Invitation creation\r\n devise_parameter_sanitizer.for(:invite) { |u| u.permit(:account_type, :email, :invitation_token)}\r\n\r\n # White list for accept invitation\r\n devise_parameter_sanitizer.for(:accept_invitation) { |u| u.permit(user_whitelist, :invitation_token)}\r\n\r\n end",
"def valid_params(params)\n params.permit(:user_id, :photo_id, :originX, :originY, :width, :height)\n end",
"def valid_parameters\n sort_symbols(@interface.allowed_parameters)\n end",
"def params_permit\n params.permit(:id)\n end",
"def allowed_params\n params.require(:allowed).permit(:email)\n end",
"def allowed_params\n params.require(:user).permit(:email, :password, :role, :first_name, :last_name, :password_confirmation)\n end",
"def allowed_params\n params.require(:user).permit(:email, :password, :role, :first_name, :last_name, :password_confirmation)\n end",
"def filter_params\n params.permit(*resource_filter_permitted_params)\n end",
"def community_params\n params.permit(:profile_image, :name, :description, :privacy_type, :viewed_by, {tags: []}, {features: []}, {admins: []}, :members, :location, :beacon, :creator, :ambassadors, :current_events, :past_events, :feed, :category, :address, :allow_member_post_to_feed, :allow_member_post_to_events)\n end",
"def specialty_params\n\t\tparams.require(:specialty).permit(*Specialty::DEFAULT_ACCESSIBLE_ATTRIBUTES)\n\tend",
"def authorize_params\n super.tap do |params|\n %w[display scope auth_type].each do |v|\n if request.params[v]\n params[v.to_sym] = request.params[v]\n end\n end\n end\n end",
"def feature_params_filter\n params.require(:feature).permit(:name, :cat, :lower, :upper, :opts, :category, :description, :company, :active, :unit, :icon)\n end",
"def available_activity_params\n # params.require(:available_activity).permit(:type,:geometry,:properties)\n whitelisted = ActionController::Parameters.new({\n type: params.require(:available_activity)[:type],\n geometry: params.require(:available_activity)[:geometry].try(:permit!).to_h,\n properties: params.require(:available_activity)[:properties].try(:permit!).to_h\n }).try(:permit!)\n end",
"def argument_params\n params.require(:argument).permit(:name)\n end",
"def user_params_pub\n\t \tparams[:user].permit(:hruid)\n\t end",
"def strong_params\n params.require(:success_metric).permit(param_whitelist)\n end",
"def property_params\n params.permit(:name, :is_available, :is_approved, :owner_id)\n end",
"def restricted_params\n #params.require(self.controller_name.classify.underscore.to_sym).permit([])\n raise(\"No strong params set, override restricted_params method in your controller. E.g. params.require(:model).permit(:attribute1, :attribute2)\")\n end",
"def sponsor_params\n params.require(:sponsor).permit(WHITE_LIST)\n end",
"def whitelist_person_params\n params.require(:person).permit(:family, :pre_title, :given_name, :dates, :post_title, :epithet, :dates_of_office, same_as: [], related_authority: [], altlabel: [], note: []) # Note - arrays need to go at the end or an error occurs!\n end",
"def parameters\n nil\n end",
"def user_params \n \tparams.require(:user).permit(:name, :email, :password, :password_confirmation)# preventing CSTR\n end",
"def sequence_param_whitelist\n default_param_whitelist << \"show_index\"\n end",
"def resource_filter_permitted_params\n raise(NotImplementedError, 'resource_filter_permitted_params method not implemented')\n end",
"def normal_params\n reject{|param, val| param_definitions[param][:internal] }\n end",
"def validate_search_inputs\n @whitelisted = params.fetch(:user, nil)\n if @whitelisted.blank?\n render_error(400, \"#{I18n.t('general_error.params_missing_key')}\": [I18n.t('general_error.params_missing_value', model: \"review\")])\n return\n else\n @whitelisted = @whitelisted.permit(:name, :uen, :description)\n end\n end",
"def special_device_list_params\n params.require(:special_device_list).permit(:name)\n end",
"def pull_request_params\n whitelist = [\n :url,\n :id,\n :html_url,\n :diff_url,\n :patch_url,\n :issue_url,\n :number,\n :state,\n :locked,\n :title\n ]\n params.require(:pull_request).permit(whitelist)\n end"
] |
[
"0.7121987",
"0.70541996",
"0.69483954",
"0.6902367",
"0.6733912",
"0.6717838",
"0.6687021",
"0.6676254",
"0.66612333",
"0.6555296",
"0.6527056",
"0.6456324",
"0.6450841",
"0.6450127",
"0.6447226",
"0.6434961",
"0.64121825",
"0.64121825",
"0.63913447",
"0.63804525",
"0.63804525",
"0.6373396",
"0.6360051",
"0.6355191",
"0.62856233",
"0.627813",
"0.62451434",
"0.6228103",
"0.6224965",
"0.6222941",
"0.6210244",
"0.62077755",
"0.61762565",
"0.61711127",
"0.6168448",
"0.6160164",
"0.61446255",
"0.6134175",
"0.6120522",
"0.6106709",
"0.60981655",
"0.6076113",
"0.60534036",
"0.60410434",
"0.6034582",
"0.6029977",
"0.6019861",
"0.6019158",
"0.6019158",
"0.6019158",
"0.6019158",
"0.6019158",
"0.6019158",
"0.6019158",
"0.6019158",
"0.6019158",
"0.6019158",
"0.6019158",
"0.6019158",
"0.6019158",
"0.6019158",
"0.6019158",
"0.6019158",
"0.6019158",
"0.60184896",
"0.60157263",
"0.6005857",
"0.6003803",
"0.60012573",
"0.59955895",
"0.5994598",
"0.5993604",
"0.5983824",
"0.5983166",
"0.5977431",
"0.597591",
"0.5968824",
"0.5965953",
"0.59647584",
"0.59647584",
"0.59566855",
"0.59506303",
"0.5950375",
"0.59485626",
"0.59440875",
"0.5930872",
"0.5930206",
"0.5925668",
"0.59235454",
"0.5917905",
"0.59164816",
"0.5913821",
"0.59128743",
"0.5906617",
"0.59053683",
"0.59052664",
"0.5901591",
"0.58987755",
"0.5897456",
"0.58970183",
"0.58942604"
] |
0.0
|
-1
|
Default method, subclasses must override this
|
def run
super
# first, ensure we're fingerprinted
require_enrichment
opt_max_rate = _get_option("max_rate")
results = _masscan_netblock(@entity,[4786],[],opt_max_rate)
_log_error "Invalid params" unless results
results.each do |r|
_log "Result: #{r}"
# check to see if it's a smart install enabled device
ip_entity = _create_entity "IpAddress", {"name" => r["ip_address"]}
_create_network_service_entity(ip_entity,r["port"],r["protocol"],{})
end
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def overrides; end",
"def custom; end",
"def custom; end",
"def default; end",
"def default; end",
"def private; end",
"def special\n override\n end",
"def defaults\n super\n end",
"def implementation; end",
"def implementation; end",
"def default\n end",
"def defaults; end",
"def defaults; end",
"def defaults; end",
"def defaults; end",
"def defaults; end",
"def defaults; end",
"def defaults; end",
"def defaults; end",
"def defaults; end",
"def defaults; end",
"def method_missing(*args)\n default\n end",
"def specie; end",
"def specie; end",
"def specie; end",
"def specie; end",
"def tag; raise 'Override this method'; end",
"def extended(*) end",
"def set_default\n end",
"def default_proc() end",
"def initialize(*)\n super\n apply_defaults\n end",
"def standard\n end",
"def initialize\n super\n end",
"def initialize\n super\n end",
"def overrides=(_arg0); end",
"def defaults!; end",
"def defaults!; end",
"def initialize\n super\n end",
"def initialize\n super\n end",
"def initialize\n super\n end",
"def internal; end",
"def set_defaults\n super\n end",
"def set_defaults\n super\n end",
"def main\n super\n return self\n end",
"def call\n # implement in subclasses\n end",
"def initialize\n super\n end",
"def initialize\n super\n end",
"def initialize\n super\n end",
"def initialize()\n # override parent\n end",
"def original; end",
"def wrapper; end",
"def initialize\n super \n end",
"def ignores; end",
"def type; super; end",
"def overload; end",
"def invoke\r\n # TODO: rename to more appropriate one 2007/05/10 by shino\r\n raise 'must be implemented in subclasses'\r\n end",
"def initialize\n super(true)\n end",
"def choose\n raise NotImplementedError.new('Must override')\n end",
"def initialize\n super()\n end",
"def initialize\n super()\n end",
"def initialize\n super()\n end",
"def initialize\n super()\n end",
"def initialize\n super()\n end",
"def initialize\n super()\n end",
"def initialize\n super()\n end",
"def normal\n end",
"def normal\n end",
"def default_content; end",
"def virtual; end",
"def extra; end",
"def specialty; end",
"def initialize()\n super\n end",
"def initialize()\n super\n end",
"def initialize()\n super\n end",
"def initialize()\n super\n end",
"def initialize()\n super\n end",
"def initialize()\n super\n end",
"def initialize()\n super\n end",
"def initialize()\n super\n end",
"def initialize()\n super\n end",
"def initialize()\n super\n end",
"def initialize()\n super\n end",
"def initialize()\n super\n end",
"def initialize()\n super\n end",
"def initialize()\n super\n end",
"def initialize()\n super\n end",
"def initialize()\n super\n end",
"def initialize()\n super\n end",
"def initialize()\n super\n end",
"def initialize()\n super\n end",
"def initialize()\n super\n end",
"def initialize()\n super\n end",
"def initialize()\n super\n end",
"def initialize()\n super\n end",
"def initialize()\n super\n end",
"def initialize()\n super\n end",
"def initialize()\n super\n end",
"def initialize()\n super\n end",
"def initialize()\n super\n end",
"def initialize()\n super\n end",
"def initialize()\n super\n end"
] |
[
"0.7415348",
"0.73554313",
"0.73554313",
"0.6991004",
"0.6991004",
"0.69613177",
"0.69572926",
"0.6824949",
"0.6801543",
"0.6801543",
"0.67361915",
"0.66934764",
"0.66934764",
"0.66934764",
"0.66934764",
"0.66934764",
"0.66934764",
"0.66934764",
"0.66934764",
"0.66934764",
"0.66934764",
"0.65642005",
"0.65501094",
"0.65501094",
"0.65501094",
"0.65501094",
"0.6528684",
"0.6490566",
"0.64695704",
"0.64676064",
"0.64597005",
"0.6455708",
"0.64551467",
"0.64551467",
"0.64329106",
"0.64307904",
"0.64307904",
"0.6429658",
"0.6429658",
"0.6429658",
"0.63993496",
"0.6390495",
"0.6390495",
"0.6376105",
"0.63759387",
"0.6340433",
"0.6340433",
"0.6340433",
"0.63253945",
"0.6310201",
"0.62980604",
"0.62755877",
"0.62675124",
"0.6240711",
"0.6220586",
"0.6218145",
"0.62124836",
"0.6204448",
"0.62003416",
"0.62003416",
"0.62003416",
"0.62003416",
"0.62003416",
"0.62003416",
"0.62003416",
"0.6188848",
"0.6188848",
"0.6188653",
"0.6170095",
"0.6169804",
"0.61623544",
"0.61536336",
"0.61536336",
"0.61536336",
"0.61536336",
"0.61536336",
"0.61536336",
"0.61536336",
"0.61536336",
"0.61536336",
"0.61536336",
"0.61536336",
"0.61536336",
"0.61536336",
"0.61536336",
"0.61536336",
"0.61536336",
"0.61536336",
"0.61536336",
"0.61536336",
"0.61536336",
"0.61536336",
"0.61536336",
"0.61536336",
"0.61536336",
"0.61536336",
"0.61536336",
"0.61536336",
"0.61536336",
"0.61536336",
"0.61536336"
] |
0.0
|
-1
|
recursively build the tree display
|
def to_a list = @list
lines = []
return lines if list.length < 2
if list[0].is_a? Array
left = to_a(list[0]).flatten
right = to_a(list[1]).flatten
indent = (left + right).collect { |i| i.length }.max
all = left[0..(left.length / 2)]
left[(left.length / 2 + 1)..-1].each do |i|
all << ((i + (" "*indent))[0..(indent-1)] + "|")
end
all << (" "*indent) + "|---"
right[0..(right.length / 2 - 1)].each do |i|
all << ((i + (" "*indent))[0..(indent-1)] + "|")
end
all << right[(right.length / 2)..-1]
else
return [list[0].to_s, "---", " |---", "---", list[1].to_s]
end
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def show_tree\n htmltree = \"\"\n self.each { |node| \n htmltree += \"<li><a href='#{normalize(node.name)}'>\"\n htmltree += ' ' * node.node_depth * 3\n htmltree += \"#{node.name}</a></li>\\n\"\n }\n htmltree\n end",
"def show_tree\n\tend",
"def display_tree() \n list = []\n yield @tree.value\n left_child = @tree.left_child\n right_child = @tree.right_child\n\n list << left_child if left_child != nil\n list << right_child if right_child != nil\n\n loop do\n break if list.empty?\n node = list.shift\n yield node.value\n list << node.left_child if node.left_child != nil\n list << node.right_child if node.right_child != nil\n end\n end",
"def render\n\t\t\ttree.flatten.map(&:to_s).join\n\t\tend",
"def _display_tree(max_length=20, tabs='')\n\t\tif(@type != '')\n\t\t\tprint(tabs + \"[\" + @type + ((@param != '')? '(' + @param.to_s + ')': '') + \"]\\n\")\n\t\telse\n\t\t\tprint(tabs + \"[TEMPLATE:\" + @template.template_file + \"]\\n\")\n\t\tend\n\n\t\tfor content in @contents\n\t\t\tif(content.is_a?(SifterElement))\n\t\t\t\tcontent._display_tree(max_length, tabs + \"\\t\")\n\t\t\telsif(content.is_a?(SifterTemplate))\n\t\t\t\tcontent._display_tree(max_length, tabs + \"\\t\")\n\t\t\telse\n\t\t\t\tcontent.gsub!(/[\\r\\n]/, ' ')\n\t\t\t\tprint(tabs + \"\\t[TEXT:\" + content[0, max_length] + \"]\\n\")\n\t\t\tend\n\t\tend\n\tend",
"def print_tree\n ''\n end",
"def render_tree(elements, symbols, parents = [])\n i = 0\n x = elements.map do |li|\n last = elements.length == i+1\n\n current = indentation(parents, last, symbols) + li[:value]\n\n children = \"\"\n if li[:children].length > 0\n children = \"\\n\" + render_tree(li[:children], symbols, parents + [last])\n end\n\n i += 1\n current + children\n end\n\n x.join(\"\\n\")\n end",
"def print_tree\n if root.children\n puts \" - root : #{root.children.length} - \"\n root.children.each(&:print_node)\n puts ''\n end\n end",
"def draw_tree\n result = ''\n max_cols = @n + 1\n calc_tree.each do |row|\n print = ' ' * ((max_cols - row.length) / 2) * max_cols\n print << row.join(' ')\n result << \"#{print}\\n\"\n end\n puts result\n end",
"def build_tree(arr)\n\tend",
"def _display_tree(max_length=20, tabs='')\n\t\treturn @contents._display_tree(max_length, tabs)\n\tend",
"def printTree()\n def pTree(node, i)\n puts node.toString i\n node.children.each do|child|\n pTree(child, i+1)\n end\n end\n pTree(@root, 0)\n end",
"def show_tree\n\t\t@root_node.show\n\tend",
"def to_tree() = puts(TTY::Tree.new({ '.' => as_tree }).render)",
"def show_node(tree, node)\n print \"ID:#{node.id} Parent:#{node.parent} Keys:\"\n node.keys.each { |key| print \"[#{key}]\"}\n print \" Sub trees:\"\n node.sub_trees.each { |sub_tree| print \"-#{sub_tree}-\"}\n print \"\\n\"\n node.sub_trees.compact.each { |sub_tree| show_node(tree, tree.nodes[sub_tree])}\nend",
"def tree\n @roots = @cards.roots\n # TODO: remove jbuilder?\n render :tree, status: :ok\n end",
"def cmd_tree\n print_tree(Editor, 0)\n end",
"def produce_tree(ary); end",
"def print_tree(tree)\n\t\t\tif tree.is_leaf? and tree.depth > 0\n\t\t\t\tprint_line((\"|\\t\"*(tree.depth-1))+\"+------\"+tree.name)\n\t\t\telse\n\t\t\t\tprint_line((\"|\\t\"*tree.depth)+tree.name)\n\t\t\tend\n\t\t\ttree.children.each_pair do |name,child|\n\t\t\t\t\tprint_tree(child)\n\t\t\tend\n\t\tend",
"def render\n path = [[root,0]]\n current_height = 0\n\n while (current, height = path.shift)\n next if nil_node_proc.call(current)\n # don't print empty leafs\n next if height >= max_height\n\n # height increased: we print the / \\ separator\n if height > current_height\n current_height += 1\n print_height_separator(current_height)\n end\n\n current.render(padding(height))\n\n # navigate left\n if !nil_node_proc.call(current.l)\n path.push([current.l, height + 1])\n elsif height < max_height\n path.push([EmptyNode.from_node(current), height + 1])\n end\n\n # navigate right\n if !nil_node_proc.call(current.r)\n path.push([current.r, height + 1])\n elsif height < max_height\n path.push([EmptyNode.from_node(current), height + 1])\n end\n end\n puts \"\\n\"\n end",
"def show_binary_tree\n text = tree_levels_text_array\n puts \"+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++\"\n puts \"\\nBinary Tree Display:\\n \"\n text.each { |row| puts \"\\n\" + row }\n puts \"\\n+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++\"\n end",
"def render(node = nil)\n elt = node | tree\n puts \"\\nNodes: #{elt.body.count}\"\n puts \"\\nAttributes:\"\n elt.attrs.each{ |k, v| puts(\"#{k}: #{v}\") }\n children = elt.body.select{ |i| i.is_a?(Node) }\n puts \"\\nChildren: #{children.each{ |c| render(c) }}\"\n\n\n end",
"def print_tree\n current = nil\n stack = Stack.new\n stack.push(@root)\n until stack.empty?\n current = stack.pop\n if current.parent.nil?\n puts current.tag\n else\n puts \"#{current.tag}\\t<=\\t#{current.parent.tag}\"\n end\n unless current.childs.size.zero?\n current.childs.reverse_each { |child| stack.push(child) }\n end\n end\n end",
"def tree(object, locals = {})\n locals = {\n depth: [],\n }.merge(locals)\n\n if locals[:depth].size > @options[:drop]\n if object == object.parent.children.last\n prefix_char = @options[:edge_char]\n else\n prefix_char = @options[:connect_char]\n end\n else\n prefix_char = \"\"\n end\n\n indents = locals[:depth].each.with_index.collect {|e, i|\n if i > @options[:drop]\n tab = e ? @options[:tab_visible_char] : \"\"\n tab.toeuc.ljust(@options[:tab_space]).toutf8\n end\n }.join\n\n if @block\n label = @block.call(object, locals)\n else\n if locals[:depth].empty? && @options[:root_label] # Change if there is root and alternative label\n label = @options[:root_label]\n else\n label = TreeSupport.node_name(object)\n end\n end\n\n buffer = \"\"\n branch_char = nil\n\n if locals[:depth].size > @options[:drop]\n branch_char = @options[:branch_char]\n end\n if locals[:depth].size < @options[:take]\n if locals[:depth].size >= @options[:drop]\n buffer = \"#{indents}#{prefix_char}#{branch_char}#{label}#{@options[:debug] ? locals[:depth].inspect : \"\"}\\n\"\n end\n end\n\n flag = false\n if object.parent\n flag = (object != object.parent.children.last)\n end\n\n locals[:depth].push(flag)\n if locals[:depth].size < @options[:take]\n buffer << object.children.collect {|node| tree(node, locals)}.join\n end\n locals[:depth].pop\n\n buffer\n end",
"def nested_array_to_html(nodes)\n classes = \"depth-#{ nodes.first[:depth] }\"\n classes += \" folder-tree\" if nodes.first[:depth].zero?\n\n html = \"<ul class=\\\"#{ classes }\\\">\"\n nodes.each do |node|\n html += \"<li>\"\n\n if node[:children]\n html += '<i class=\"icon-folder-close\"></i>'\n else\n html += '<i class=\"icon-file-alt\"></i>'\n end\n\n html += \"<span class=\\\"name\\\">#{ node[:name].strip }</span>\"\n unless (node[:children].nil? || node[:children].empty?)\n html += nested_array_to_html(node[:children]) \n end\n html += '</li>'\n end\n\n html += '</ul>'\n\n html\n end",
"def tree_print(tree, folder)\n str = \"\";\n tree.each do |key, value|\n if not value.is_a?(Hash)\n # page\n if value.name != \"Home\"\n str += \"<li><a href=\\\"/#{@wiki.base_path}#{value.url_path}\\\">#{value.name}</a></li>\"\n end\n else\n # folder\n subfolder = folder + \"/\" + key\n str += \"<li><a href=\\\"#{@wiki.base_path}#{subfolder}/\\\">#{key}</a>\"\n str += \"<ul>\"\n str += tree_print(value, subfolder)\n str += \"</ul>\"\n str += \"</li>\"\n end\n end\n return str\n end",
"def print_tree(level = 0)\n if is_root?\n print \"\\n*\"\n else\n print \"|\" unless parent.is_root?\n print(' ' * (level - 1) * 4)\n print(is_root? ? \"+\" : \"|\")\n print \"---\"\n print(has_children? ? \"+\" : \">\")\n end\n\n if content\n content_hash = content.split(\"[\").first\n else\n content_hash = nil\n end\n\n puts \" #{content}\" + \" <Type: \" + (@node_type || \"no_type\") + \">\"\n\n children { |child| child.print_tree(level + 1)}\n end",
"def node_tree(nodes, &block)\n \n nodes = nodes.dup\n printed_nodes = []\n \n result = \"<ul>\"\n \n # top level nodes first, then others\n for node in nodes\n next if node.instance_of?(Center) || node.instance_of?(Team)\n next unless node.parent == nil\n printed_nodes << node\n result += \"<li>\"\n\n if block_given?\n result += yield node\n else\n result += node.title\n end\n\n children = node.children.dup\n children.delete_if { |r| not nodes.include?(r) }\n if not children.empty?\n result += node_tree_help(children, nodes, printed_nodes, &block)\n end\n \n result += \"</li>\"\n end\n \n # TODO: Add depth counting here to get a minimum of trees\n for node in nodes\n next if printed_nodes.include? node\n printed_nodes << node\n \n result += \"<li>\"\n\n if block_given?\n result += yield node\n else\n result += node.title\n end\n\n children = node.children #.dup\n children.delete_if { |r| not nodes.include?(r) }\n\n if not children.empty?\n result += node_tree_help(children, nodes, printed_nodes, &block)\n end\n \n result += \"</li>\"\n end\n\n result += '</ul>'\n\n return result\n end",
"def display\n urls = SiteUrl.where(:site => params[:site]).map(&:url) - [nil,\"\"]\n #render plain: urls\n @tree = urls.count>0? urls_2_tree(params[:site],urls) : []\n @site = params[:site]\n #@tree=[{\"id\"=>\"https://www.google.com/\", \"text\"=>\"https://www.google.com/\", \"children\"=>[{\"id\"=>\"images\", \"text\"=>\"images\", \"children\"=>[{\"id\"=>\"icon.png\", \"text\"=>\"icon.png\", \"children\"=>[], \"icon\"=>\"glyphicon glyphicon-list\", \"state\"=>{\"opened\"=>true}}], \"icon\"=>\"glyphicon glyphicon-list\", \"state\"=>{\"opened\"=>true}}], \"icon\"=>\"glyphicon glyphicon-list\", \"state\"=>{\"opened\"=>true}}]\n logger.debug(\"tree: #{@tree}\")\n respond_to do |format|\n format.html\n format.json { render json: @tree }\n end\n end",
"def tree\n @tree ||= build_tree\n end",
"def printTree(root = @root, indentation = \"\\n\")\r\n #print root the first time running mehtod, the rest will be printed as children\r\n puts \"#{indentation}#{root.value}\" if root == @root\r\n root.children.each do | child |\r\n puts \"#{indentation}\\t#{child.value}\\n\"\r\n #if there are children, again\r\n printTree(child, indentation + \"\\t\") if child.children.any?\r\n end\r\n end",
"def print_tree(tree)\n return \"-\" if tree.nil?\n puts \"#{tree.value}: \"\n print \"Left: \"\n puts \"#{print_tree(tree.children[0])}\"\n print \"Right: \"\n puts \"#{print_tree(tree.children[1])}\"\nend",
"def walk(elements, max_depth, *args)\n output = ''\n\n # invalid parameter or nothing to walk\n return output if max_depth < -1 || elements.blank?\n\n parent_field = @db_fields[:parent]\n\n # flat display\n if -1 == max_depth\n empty_array = {}\n elements.each do |e|\n display_element(e, empty_array, 1, 0, args, output)\n end\n return output\n end\n\n # Need to display in hierarchical order.\n # Separate elements into two buckets: top level and children elements.\n # Children_elements is two dimensional array, eg.\n # Children_elements[10][] contains all sub-elements whose parent is 10.\n top_level_elements = []\n children_elements = {}\n elements.each do |e|\n if e.send(parent_field).blank? || e.send(parent_field) == 0\n top_level_elements << e\n else\n children_elements[e.send(parent_field)] ||= []\n children_elements[e.send(parent_field)] << e\n end\n end\n\n # When none of the elements is top level.\n # Assume the first one must be root of the sub elements.\n if top_level_elements.blank?\n root = elements.first\n\n top_level_elements = []\n children_elements = {}\n elements.each do |e|\n if root.send(parent_field) == e.send(parent_field)\n top_level_elements << e\n else\n children_elements[e.send(parent_field).to_i] ||= []\n children_elements[e.send(parent_field).to_i] << e\n end\n end\n end\n\n top_level_elements.each do |e|\n display_element(e, children_elements, max_depth, 0, args, output)\n end\n\n # If we are displaying all levels, and remaining children_elements is not empty,\n # then we got orphans, which should be displayed regardless.\n if max_depth == 0 && !children_elements.empty?\n empty_array = {}\n children_elements.values.each do |orphans|\n orphans.each do |op|\n display_element(op, empty_array, 1, 0, args, output)\n end\n end\n end\n output\n end",
"def printf\n output = []\n children = []\n output.push(\"#{@root.title}: #{@root.rating}\")\n if @root.left != nil\n children.push(@root.left)\n end\n if @root.right != nil\n children.push(@root.right)\n end\n children.each do |i|\n output.push(\"#{i.title}: #{i.rating}\")\n if i.left != nil\n children.push(i.left)\n end\n if i.right != nil\n children.push(i.right)\n end\n end\n puts output\n end",
"def trees_to_html(trees)\nend",
"def printTree(options = {})\n # Set defaults\n options[:name] ||= true\n options[:content] ||= false\n \n result = \"\"\n \n options[:output] = result \n # Traverse tree and modify result by tacking on child names.\n printTraversal(options)\n \n puts result\n end",
"def display\n puts @root.to_s + \" \" + \"=\" * 10\n traverse(@root,0) {|n, l| puts \" \"*l + \"node: #{n}\"}\n self\n end",
"def print_tree(d)\n \n print_tree(d['l']) if d.key?('l')\n print \"#{d['v']},\" if d.key?('v')\n print_tree(d['r']) if d.key?('r')\n \nend",
"def print_tree(n, i = 0)\n i.times { print \" \" }\n puts \"+-#{n}\"\n $classtree[n].sort.each { |c| print_tree(c, i+2) }\nend",
"def tree(data_path, options = {})\n\t\t\t\tresult = \"\"\n\n\t\t\t\t# Unique hash\n\t\t\t\t@hash = Digest::SHA1.hexdigest(data_path.to_s)\n\n\t\t\t\t# Options\n\t\t\t\t@options = options.nil? ? {} : options\n\n\t\t\t\t# Clipboard\n\t\t\t\tif @options[:clipboard_attrs]\n\t\t\t\t\tclipboard = true\n\t\t\t\t\t@options[:clipboard_attrs] = [@options[:clipboard_attrs]] if !@options[:clipboard_attrs].is_a?(Array)\n\t\t\t\t\tclipboard_attrs_js = \"[\" + @options[:clipboard_attrs].map { |item| \"'#{item}'\" }.join(\",\") + \"]\"\n\t\t\t\telse\n\t\t\t\t\tclipboard = false\n\t\t\t\t\tclipboard_attrs_js = \"[]\"\n\t\t\t\tend\n\n\t\t\t\t# Actions\n\t\t\t\tif @options[:actions]\n\t\t\t\t\tactions_js = \"[\"\n\t\t\t\t\toptions[:actions].each do |key, action|\n\t\t\t\t\t\tactions_js += %{\n\t\t\t\t\t\t\t{\n\t\t\t\t\t\t\t\turl: '#{action[:path] ? @path_resolver.resolve(action[:path], \":id\") : \"\"}',\n\t\t\t\t\t\t\t\ticon: '#{action[:icon]}',\n\t\t\t\t\t\t\t\tlabel: '#{action[:label]}',\n\t\t\t\t\t\t\t\tcollapsed: #{action[:collapsed] == true ? \"true\" : \"false\"},\n\t\t\t\t\t\t\t\tstyle: '#{action[:style] ? action[:style] : \"default\"}',\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t}\n\t\t\t\t\tend\n\t\t\t\t\tactions_js += \"]\"\n\t\t\t\telse\n\t\t\t\t\tactions_js = \"[]\"\n\t\t\t\tend\n\n\t\t\t\t# Parent\n\t\t\t\tparent = (options[:parent] ? options[:parent] : nil)\n\n\t\t\t\t# Save state\n\t\t\t\tsave_state = (options[:save_state] ? options[:save_state] : :simple)\n\n\t\t\t\t# Application JS\n\t\t\t\tresult += @template.javascript_tag(%{\n\t\t\t\t\tvar rug_tree_#{@hash} = null;\n\t\t\t\t\t$(document).ready(function() {\n\t\t\t\t\t\trug_tree_#{@hash} = new RugTree('#{@hash}', {\n\t\t\t\t\t\t\t\n\t\t\t\t\t\t\t// Model\n\t\t\t\t\t\t\tmodel: 'node',\n\n\t\t\t\t\t\t\t// State\n\t\t\t\t\t\t\tsaveState: '#{save_state.to_s}',\n\n\t\t\t\t\t\t\t// Parent element\n\t\t\t\t\t\t\tparent: '#{parent.to_s}',\n\n\t\t\t\t\t\t\t// Icons\n\t\t\t\t\t\t\tclosedIcon: '#{@icon_builder.render(@options[:closed_icon] ? @options[:closed_icon] : \"chevron-right\")}',\n\t\t\t\t\t\t\topenedIcon: '#{@icon_builder.render(@options[:opened_icon] ? @options[:opened_icon] : \"chevron-down\")}',\n\n\t\t\t\t\t\t\t// Show\n\t\t\t\t\t\t\tshow: #{check_show(@options) ? 'true' : 'false'},\n\t\t\t\t\t\t\tshowEvent: '#{@options[:show_event] && @options[:show_event].to_sym == :double_click ? \"dblclick\" : \"click\"}',\n\t\t\t\t\t\t\tshowUrl: '#{@path_resolver.resolve(@options[:paths][:show], \":id\")}',\n\n\t\t\t\t\t\t\t// Create\n\t\t\t\t\t\t\tcreate: #{check_create(@options) ? 'true' : 'false'}, \n\t\t\t\t\t\t\tcreateUrl: '#{@path_resolver.resolve(@options[:paths][:create])}',\n\t\t\t\t\t\t\tcreateIcon: '#{@icon_builder.render(@options[:update_icon] ? @options[:update_icon] : \"plus\")}',\n\t\t\t\t\t\t\tcreateLabel: '#{I18n.t(\"general.action.create_child\").upcase_first}',\n\t\t\t\t\t\t\tcreateActionCollapsed: #{@options[:create_action_collapsed] == true ? 'true' : 'false'}, \n\t\t\t\t\t\t\tcreateSuccessMessage: '#{I18n.t(\"general.messages.create.success\")}',\n\n\t\t\t\t\t\t\t// Update\n\t\t\t\t\t\t\tupdate: #{check_update(@options) ? 'true' : 'false'}, \n\t\t\t\t\t\t\tupdateUrl: '#{@path_resolver.resolve(@options[:paths][:update], \":id\")}', \n\t\t\t\t\t\t\tupdateIcon: '#{@icon_builder.render(@options[:update_icon] ? @options[:update_icon] : \"pencil\")}',\n\t\t\t\t\t\t\tupdateLabel: '#{I18n.t(\"general.action.update\").upcase_first}',\n\t\t\t\t\t\t\tupdateActionCollapsed: #{@options[:update_action_collapsed] == true ? 'true' : 'false'}, \n\t\t\t\t\t\t\tupdateSuccessMessage: '#{I18n.t(\"general.messages.create.success\")}',\n\n\t\t\t\t\t\t\t// Destroy\n\t\t\t\t\t\t\tdestroy: #{check_destroy(@options) ? 'true' : 'false'}, \n\t\t\t\t\t\t\tdestroyUrl: '#{@path_resolver.resolve(@options[:paths][:destroy], \":id\")}', \n\t\t\t\t\t\t\tdestroyIcon: '#{@icon_builder.render(@options[:update_icon] ? @options[:update_icon] : \"trash\")}',\n\t\t\t\t\t\t\tdestroyLabel: '#{I18n.t(\"general.action.destroy\").upcase_first}',\n\t\t\t\t\t\t\tdestroyActionCollapsed: #{@options[:destroy_action_collapsed] == true ? 'true' : 'false'}, \n\t\t\t\t\t\t\tdestroyConfirmMessage: '#{I18n.t(\"general.are_you_sure\")}',\n\t\t\t\t\t\t\tdestroySuccessMessage: '#{I18n.t(\"general.messages.destroy.success\")}',\n\n\t\t\t\t\t\t\t// Moving\n\t\t\t\t\t\t\tmoving: #{check_moving(@options) ? 'true' : 'false'},\n\t\t\t\t\t\t\tmovingUrl: '#{@path_resolver.resolve(@options[:paths][:move], \":id\", \":relation\", \":destination_id\")}',\n\t\t\t\t\t\t\n\t\t\t\t\t\t\t// Type\n\t\t\t\t\t\t\ttypeIconTemplate: '#{@icon_builder.render(\":icon\", class: \"jqtree-icon\")}',\n\t\t\t\t\t\t\ttypeIconAttr: '#{@options[:type_icon_attr]}',\n\n\t\t\t\t\t\t\t// Actions\n\t\t\t\t\t\t\tactions: #{actions_js},\n\t\t\t\t\t\t\tactionsIconTemplate: '#{@icon_builder.render(\":icon\")}',\n\n\t\t\t\t\t\t\t// Clipboard\n\t\t\t\t\t\t\tclipboard: #{clipboard ? 'true' : 'false'},\n\t\t\t\t\t\t\tclipboardIcon: '#{@icon_builder.render(@options[:clipboard_icon] ? @options[:clipboard_icon] : \"clipboard\")}',\n\t\t\t\t\t\t\tclipboardTemplate: \"#{clipboard ? (@options[:clipboard_template] ? @options[:clipboard_template].gsub('\"', \"'\") : \":\" + @options[:clipboard_attrs].first) : \"\"}\",\n\t\t\t\t\t\t\tclipboardAttrs: #{clipboard_attrs_js},\n\t\t\t\t\t\t\tclipboardActionCollapsed: #{@options[:clipboard_action_collapsed] == true ? 'true' : 'false'}, \n\t\t\t\t\t\t\t\n\t\t\t\t\t\t\t// Select\n\t\t\t\t\t\t\tselectByDefault: #{@options[:select_by_default] ? @options[:select_by_default].to_i : \"null\"},\n\n\t\t\t\t\t\t\t// Reload\n\t\t\t\t\t\t\treloadIcon: '#{@icon_builder.render(@options[:update_icon] ? @options[:update_icon] : \"refresh\")}',\n\t\t\t\t\t\t\treloadLabel: '#{I18n.t(\"general.action.reload\").upcase_first}',\n\t\t\t\t\t\t});\n\t\t\t\t\t\trug_tree_#{@hash}.ready();\n\t\t\t\t\t});\n\t\t\t\t\t$(document).on('turbolinks:load', function() {\n\t\t\t\t\t\trug_tree_#{@hash}.repair();\n\t\t\t\t\t});\n\t\t\t\t})\n\n\t\t\t\tresult += %{\n\t\t\t\t\t<div id=\"tree-#{@hash}\" data-url=\"#{data_path.to_s}\"></div>\n\t\t\t\t}\n\n\t\t\t\treturn result.html_safe\n\t\t\tend",
"def build_up_frbr_tree(int_digital_edition_id, bln_primary)\n \t@arr_frbr_tree = Array.new\n \t\n \t\n \t@items = ItemsHasDigitalEditions.where(\"digital_edition_id = ?\", int_digital_edition_id).order(\"item_has_digital_edition_primary ASC\")\n \t@items.each do |items_item|\n \t\tint_assoc_man = 0\n \t\t@item = Item.where(\"id = ?\", items_item.item_id)\n \t\t@item.each do |item_item|\n \t\t\t\tint_item_id = item_item.id\n\t\t\t\tstr_item_name = \"I_\" + item_item.id + \": \" + item_item.item_siglum\n\t\t\t\tint_assoc_man = item_item.manifestation_id\n \t\tend\n \t\t@arr_manifestation_master = Array.new\n \t\t@manifestations = Manifestation.where(\"id = ?\", int_assoc_man).order(\"manifestation_name ASC\")\n \t\t@manifestations.each do |man_item|\n \t\t\t@arr_expressions_master = Array.new\n \t\t\t@exp_man_joins = ExpressionsHasManifestations.where(\"manifestation_id = ?\", man_item.id)\n \t\t\t@exp_man_joins.each do |exp_join_item|\n \t\t\t\n \t\t\t\t# here is where I have to put in the logic whether to display all expressions and works\n \t\t\t\t# or just the primary ones\n \t\t\t\tif (bln_primary)\n \t\t\t\t\t@expressions = Expression.joins(:works).where(\"expressions.id = ?\", exp_join_item.expression_id).order(\"expression_name ASC\")\n \t\t\t\telse\n \t\t\t\t\t@expressions = Expression.where(\"id = ?\", exp_join_item.expression_id).order(\"expression_siglum ASC\")\n \t\t\t\tend\n \t\t\t\t@expressions.each do |expression_item|\n \t\t\t\t\n \t\t\t\t\n \t\t\t\t\t@arr_works_master = Array.new\n \t\t\t\t\tif (bln_primary)\n \t\t\t\t\t\t@works = Work.where(\"work_frbr = 1 AND id = ?\", expression_item.work_id).order(\"work_name ASC\")\n \t\t\t\t\telse\n \t\t\t\t\t\t@works = Work.where(\"id = ?\", expression_item.work_id).order(\"work_siglum ASC\")\n \t\t\t\t\tend\n \t\t\t\t\t@works.each do |work_item|\n \t\t\t\t\t\t@arr_work_item = Array.new\n \t\t\t\t\t\tif bln_primary\n \t\t\t\t\t\t\tstr_work_item = \"W_\" + work_item.id.to_s + \": \" + work_item.work_name + \" (\" + work_item.work_siglum + \")\"\n \t\t\t\t\t\telse\n \t\t\t\t\t\t\tstr_work_item = \"(\" + work_item.work_siglum + \") \" + work_item.work_name + \": W_\" + work_item.id.to_s\n \t\t\t\t\t\tend\n \t\t\t\t\t\t@arr_work_item.push(work_item.id)\n \t\t\t\t\t\t@arr_work_item.push(str_work_item)\n \t\t\t\t\t\t@arr_works_master.push(@arr_work_item)\n \t\t\t\t\tend\n\n \t\t\t\t\t@arr_expression_item = Array.new\n \t\t\t\t\tstr_expression_item = \"E_\" + expression_item.id.to_s + \": \" + expression_item.expression_name + \" (\" + expression_item.expression_siglum + \")\"\n \t\t\t\t\t@arr_expression_item.push(expression_item.id)\n \t\t\t\t\t@arr_expression_item.push(str_expression_item)\n \t\t\t\t\t@arr_expression_item.push(@arr_works_master)\n \t\t\t\t\t@arr_expressions_master.push(@arr_expression_item)\n \t\t\t\t\t\n \t\t\t\tend\n\n \t\t\tend\n \t\t\t\n \t\t\t\n \t\t\t\n \t\t\t\n \t\t\t@arr_manifestation_item = Array.new\n \t\t\tstr_manifestation_name = \"M_\" + man_item.id.to_s + \": \" + man_item.manifestation_name + \" (\" + man_item.manifestation_siglum + \")\"\n \t\t\t@arr_manifestation_item.push(man_item.id)\n \t\t\t@arr_manifestation_item.push(str_manifestation_name)\n \t\t\t@arr_manifestation_item.push(@arr_expressions_master)\n\n \t\tend\n\n \t\t#fill the items array and push to tree master\n\t\tif (items_item.item_has_digital_edition_primary)\n\t\t\tint_item_primary = 1\n\t\telse\n\t\t\tint_item_primary = 0\n\t\tend\n\t\tstr_item_name = \"I_\" + item_item.id.to_s + \": \" + item_item.item_siglum\n\t\t@arr_item_item = Array.new\n\t\t@arr_item_item.push(int_item_id)\n\t\t@arr_item_item.push(str_item_name)\n\t\t@arr_item_item.push(int_item_primary)\n\t\t@arr_item_item.push(@arr_manifestation_master)\n \t\tarr_frbr_tree.push(@arr_item_item)\n \tend\n \t\n \treturn @arr_frbr_tree\n end",
"def generate_class_tree_level(parent='')\n $all.map { |klass|\n if parent == klass['parentname']\n [\n klass['name'],\n \"classes/#{klass['fullname']}.html\", # klass.path, \n '',\n generate_class_tree_level(klass['fullname'])\n ]\n else\n nil\n end\n }.compact\nend",
"def tree\n # Caution: use only for small projects, don't use in root.\n @title = 'Full Tree'\n # @files = `zsh -c 'print -rl -- **/*(#{@sorto}#{@hidden}M)'`.split(\"\\n\")\n @files = Dir['**/*']\n message \"#{@files.size} files.\"\nend",
"def print_tree(array , item, level)\n items = array[item]\n unless items == nil\n indent = level > 0 ? sprintf(\"%#{level * 2}s\", \" \") : \"\"\n items.each do |e|\n puts \"#{indent}-#{e[:title]}\"\n print_tree(e, level + 1)\n end\n end\nend",
"def generate_tree\n root =\tTreeNode.new(3)\n root.left =\tTreeNode.new(9)\n right = \t\tTreeNode.new(20)\n right.left = \tTreeNode.new(15)\n right.right = TreeNode.new(7)\n root.right = \tright\n root\nend",
"def see_tree(tree)\n tree.each do |i|\n puts \"---------------\"\n puts \"Node: #{i.value}\"\n puts \"Parent: \" + i.find_parent[0].value.to_s if i.find_parent.count > 0\n puts \"Left Child: \" + i.find_left_child[0].value.to_s if i.find_left_child.count > 0\n puts \"Right Child: \" + i.find_right_child[0].value.to_s if i.find_right_child.count > 0\n end\nend",
"def build_tree\n c1 = ComponentNode.new(110)\n c2 = ComponentNode.new(20)\n c3 = ComponentNode.new(20)\n c4 = ComponentNode.new(150)\n c5 = ComponentNode.new(80)\n c6 = ComponentNode.new(120, [c1, c2, c3])\n c7 = ComponentNode.new(180, [c4, c5])\n return(ComponentNode.new(200, [c6, c7]))\n end",
"def test_print_tree\n setup_test_tree\n #puts\n #@root.print_tree\n end",
"def build_tree(data)\n @root = Node.new(data[0])\n data.shift\n data.each { |value| @root.insert(value) }\n end",
"def build_tree( n , d )\n \n if d.key?('v')\n n < d['v'] ? build_tree(n , d['l']) : build_tree(n, d['r'])\n else\n d['l'] = {}\n d['v'] = n\n d['r'] = {}\n end\n \nend",
"def display_tree(an_array)\r\n an_array.length\r\n count = 1\r\n (count - 1).upto(count) do\r\n end\r\nend",
"def build_tree(arr)\n #arr.shuffle!\n arr.each do |x|\n if @root == nil\n @root = Node.new(x)\n else\n current_node = @root\n until current_node == nil\n if x < current_node.value\n parent = current_node\n direction = \"left\"\n current_node = current_node.left_child\n elsif x > current_node.value\n parent = current_node\n direction = \"right\"\n current_node = current_node.right_child\n end\n end\n if direction == \"left\"\n parent.left_child = Node.new(x)\n elsif direction == \"right\"\n parent.right_child = Node.new(x)\n end\n end\n end\n end",
"def render_recursively\n @obj = Node.find(params[:id])\n max_depth = params[:max_depth].to_f if params[:max_depth]\n\n respond_to do |format|\n format.html {render inline: NodeRepToHtml::convert(NodeRep.new(@obj, max_depth))}\n format.json {render json: NodeRep.new(@obj, max_depth)}\n end\n\n end",
"def printf()\n #node object array\n children = []\n\n #info array\n arr = []\n arr.push(\"#{@root.title}: #{@root.rating}\")\n if @root.left != nil\n children.push(@root.left)\n end\n if @root.right != nil\n children.push(@root.right)\n end\n\n # adds nodes to children array AND pushes data to arr\n children.each do |node|\n if node.left != nil\n children.push(node.left)\n end\n if node.right != nil\n children.push(node.right)\n end\n arr.push(\"#{node.title}: #{node.rating}\")\n end\n\n puts arr\n end",
"def new_tree(group, options = {}, &block)\n options[:root_ol] = true if options[:root_ol].nil?\n options[:vehicles] = true if options[:vehicles].nil?\n options[:users] = false if options[:users].nil?\n options[:close_level] ||= 99\n options[:stop_level] ||= 99\n \n html = [[]]\n if options[:root_ol]\n pending = [:ol, :li, group, :nli, :nol]\n level = -1\n else\n pending = [:li, group, :nli]\n level = 0\n end\n \n while node = pending.shift\n case node\n when :ol\n html << []\n level += 1\n when :li\n html << []\n when :nli\n content = content_tag(:li, html.pop.join)\n html.last << content\n when :nol\n content = content_tag(:ol, html.pop.join)\n html.last << content\n level -= 1\n else\n html.last << capture(node, level, &block)\n end\n \n if !node.is_a?(Symbol) && !(node.is_a?(Device) || node.is_a?(User))\n if options[:vehicles]\n children = [:ol, (node.children + node.devices).map {|c| [:li, c, :nli]}, :nol]\n elsif options[:users]\n children = [:ol, (node.children + node.users).map {|c| [:li, c, :nli]}, :nol]\n else\n children = [:ol, node.children.map {|c| [:li, c, :nli]}, :nol]\n end\n \n pending.unshift *(children.flatten)\n end\n end\n \n concat(html.to_s)\n end",
"def print_recursive_menu(menu_entry)\n menu_children = MenuEntry.find_all_by_parent_id(menu_entry.id)\n current_entry = \"<li>\" + menu_link(menu_entry) + \"</li>\"\n return current_entry if menu_children.size == 0\n current_entry = \"<li>\" + menu_link(menu_entry) + \"<ul>\"\n for menu_child in menu_children do\n current_entry += print_recursive_menu(menu_child)\n end\n current_entry += \"</ul></li>\"\n return current_entry\n end",
"def inorder_display\n puts \"ROOT: #{@root.value}\\n\"\n inorder_display_helper @root\n end",
"def show()\n printed = \"IPv4 Tree\\n---------\\n\"\n list4 = dump_children(@v4_root)\n list6 = dump_children(@v6_root)\n\n list4.each do |entry|\n cidr = entry[:CIDR]\n depth = entry[:Depth]\n\n if (depth == 0)\n indent = \"\"\n else\n indent = \" \" * (depth*3)\n end\n\n printed << \"#{indent}#{cidr.desc}\\n\"\n end\n\n printed << \"\\n\\nIPv6 Tree\\n---------\\n\" if (list6.length != 0)\n\n list6.each do |entry|\n cidr = entry[:CIDR]\n depth = entry[:Depth]\n\n if (depth == 0)\n indent = \"\"\n else\n indent = \" \" * (depth*3)\n end\n\n printed << \"#{indent}#{cidr.desc(:Short => true)}\\n\"\n end\n\n return(printed)\n end",
"def render_tag_tree(tree_node=@tree.root)\n string = \" \" * 2 * tree_node.depth\n string += \"<#{tree_node.info.type}\"\n string += \" class=#{tree_node.info.classes}\" unless tree_node.info.classes.nil?\n string += \" id=#{tree_node.info.id}\" unless tree_node.info.id.nil?\n string += \" name=#{tree_node.info.name}\" unless tree_node.info.name.nil?\n string += \">\"\n string += \" #{tree_node.info.text}\"\n puts string\n\n tree_node.children.each do |child|\n render_tag_tree(child)\n end\n end",
"def build_tree(unit, node, level = 0)\r\n return nil if level > @max_depth\r\n \t\r\n unit.next_move(node.current_case).each do |next_case|\r\n next if next_case[0] < 0 || next_case[0] > 7 ||\r\n next_case[1] < 0 || next_case[1] > 7 \r\n \r\n next_node = Node.new(next_case, node)\r\n node.children << next_node\r\n\r\n build_tree(unit, next_node, level + 1)\r\n end \r\n end",
"def tree\n # Caution: use only for small projects, don't use in root.\n $title = \"Full Tree\"\n $files = `zsh -c 'print -rl -- **/*(#{$sorto}#{$hidden}M)'`.split(\"\\n\")\nend",
"def build_tree(array)\n\t\t@root = Node.new(array.shift)\n\t\tarray.each { |value| add_node(value, @root)}\n\tend",
"def link_tree\n ''.html_safe.tap do |content|\n content << toggle_link\n\n unless leaf?\n content << h.content_tag(:ul) do\n h.content_tag_for(:li, children) do |c|\n c.decorate.link_tree\n end\n end\n end\n end\n end",
"def class_tree(root, show_methods = true, colorize = true)\n # get children of root\n children = Hash.new()\n maxlength = root.to_s.length\n ObjectSpace.each_object(Class) do |aClass|\n if (root != aClass && aClass.ancestors.include?(root))\n children[aClass.superclass] = Array.new() if children[aClass.superclass] == nil\n children[aClass.superclass].push(aClass)\n maxlength = aClass.to_s.length if aClass.to_s.length > maxlength\n end\n end\n maxlength += 3\n\n # print nice ascii class inheritance tree\n indentation = \" \"*4\n c = Hash.new(\"\")\n if colorize\n c[:lines] = term_color :blue\n c[:dots] = term_color :red\n c[:classNames] = term_color :yellow\n c[:moduleNames] = term_color :green\n c[:methodNames] = term_color :default\n end\n\n recursePrint = proc do |current_root,prefixString|\n if show_methods # show methods (but don't show mixed in modules)\n puts(prefixString.tr('`','|'))\n methods = (current_root.instance_methods - (begin current_root.superclass.instance_methods; rescue NameError; []; end))\n strings = methods.sort.collect {|m|\n prefixString.tr('`',' ') +\n ( children[current_root] == nil ? \" \"*maxlength : c[:lines]+indentation+\"|\"+\" \"*(maxlength-indentation.length-1)) + \n c[:dots]+\":.. \" +\n c[:methodNames]+m.to_s\n }\n strings[0] = prefixString + c[:lines]+\"- \"+c[:classNames]+current_root.to_s\n strings[0] += \" \" + c[:dots]+\".\"*(maxlength-current_root.to_s.length) + \" \"+c[:methodNames]+methods[0].to_s if methods[0] != nil\n strings.each {|aString| puts(aString) }\n else\n string = \"#{prefixString}#{c[:lines]}-#{c[:classNames]}#{current_root.to_s}\"\n modules = current_root.included_modules - [Kernel]\n if modules.size > 0\n string += \" \"*(maxlength-current_root.to_s.length)+c[:lines]+\"[ \"+c[:moduleNames]+\n modules.join( c[:lines]+\", \"+c[:moduleNames]) +\n c[:lines]+\" ]\"\n end\n puts(string)\n end\n if children[current_root] != nil\n children[current_root].sort! {|a, b| a.to_s <=> b.to_s}\n children[current_root].each do |child|\n recursePrint.call(\n child,\n prefixString.tr('`',' ') + indentation + c[:lines]+(child == children[current_root].last ? \"`\":\"|\"))\n end\n end\n end\n\n recursePrint.call(root,\"\")\nend",
"def printf(list=[@root], output='')\n # remove front list item & add it to the output\n first = list.shift()\n print_string = first.title + ': ' + first.rating.to_s + \"\\n\"\n output << print_string\n # add its children to the back of the list\n if first.left\n list << first.left\n end\n if first.right\n list << first.right\n end\n # repeat until all nodes have been visited, then print output\n if list.empty?\n puts output\n return\n else\n printf(list, output)\n end\n end",
"def render\n return ro_standard if @readonly \n set_initial_value('html','value')\n require 'sort_alphabetical' \n \n record = record_text_for(@yaml['name'])\n p '******', \"<div id=\\\"#{@yaml['name']}\\\" class=\\\"tree-select\\\" #{set_style()} >\"\n @html << \"<div id=\\\"#{@yaml['name']}\\\" class=\\\"tree-select\\\" #{set_style()} >\"\n# Fill @choices hash. The key is parent object id\n @choices = {}\n do_eval(@yaml['eval']).each {|data| @choices[ data[2].to_s ] ||= []; @choices[ data[2].to_s ] << (data << false)}\n# put current values hash with. To speed up selection when there is a lot of categories\n current_values = {}\n current = @record[@yaml['name']] || []\n current.each {|e| current_values[e.to_s] = true}\n# set third element of @choices when selected\n @choices.keys.each do |key|\n 0.upto( @choices[key].size - 1 ) do |i|\n choice = @choices[key][i]\n choice[3] = true if current_values[ choice[1].to_s ]\n end\n end\n make_tree(nil)\n @html << '</ul></div>'\n# add hidden communication field \n @html << @parent.hidden_field(record, @yaml['name'], value: current.join(','))\n# javascript to update hidden record field when tree looses focus\n @js =<<EOJS\n$(function(){\n $(\"##{@yaml['name']}\").jstree( {\n \"checkbox\" : {\"three_state\" : false}, \n \"core\" : { \"themes\" : { \"icons\": false } },\n \"plugins\" : [\"checkbox\"]\n });\n});\n \n$(document).ready(function() {\n $('##{@yaml['name']}').on('focusout', function(e) {\n var checked_ids = [];\n var checked = $('##{@yaml['name']}').jstree(\"get_checked\", true);\n $.each(checked, function() {\n checked_ids.push( this.data.id );\n });\n $('#record_#{@yaml['name']}').val( checked_ids.join(\",\") );\n });\n});\nEOJS\n self\nend",
"def printout\n\t\t\n\t\tindex = 0\n\t\t\n\t\tdef small_loop (node, index)\n\t\t\t\n\t\t\tfor i in 0...index\n\t\t\t\tprint \" \"\n\t\t\tend\n\t\t\t\n\t\t\tputs node.name\n\t\t\tindex += 1\n\t\t\t\n\t\t\tif node.children.length > 0\n\t\t\t\tindex += 1\n\t\t\t\tnode.children.cycle(1) { |child| small_loop(child, index) }\n\t\t\tend\n\t\tend\n\t\t\n\t\tsmall_loop(@root, index)\n\tend",
"def tree(dir='.')\n output = []\n visit_tree(output, '.', '', '', '', dir)\n output.join(\"\\n\")\n end",
"def render_recursively(output_path, title, obj, parent = '')\n renderer.render(output_path, parent, title, obj.get_thumbs, obj.filter)\n\n # For Doc object, returns a list of makes. For the NodeSet object, returns a list of models.\n obj.filter.each do |key|\n # Only recurse if not the same as the parent to avoid an infinite loop\n if title != key\n nodes = obj.search('work:contains(\"' + key + '\")')\n render_recursively(output_path, key, nodes, title)\n end\n end\n end",
"def build_tree(arr)\n @root = insert_node(nil, arr.shift)\n arr.each { |value| insert_node(@root, value) }\n end",
"def build_tree(arr)\n @root = Node.new(arr.shift)\n arr.each { |data| insert_data(data, @root) }\n end",
"def print_tree(out = $stdout)\n out.puts \"## Class: #{name}\"\n out.puts \" Visible Name: #{visible_name}\"\n out.puts \" Description : #{description}\"\n tables.each do |table|\n table.print_tree(out)\n end\n end",
"def printf(root=nil)# def printf(children=nil)\n if root == nil\n root = @root\n end\n node_array = []\n node_array.push(root)\n printf_breadth(node_array)\n end",
"def level_order_print(tree)\n unless tree\n return\n end\n\n nodes = [tree]\n current_line_count = 1\n next_line_count = 0\n\n while nodes.length != 0\n current_node = nodes.shift\n current_line_count -= 1\n print current_node.key.to_s + ' '\n if current_node.left_child\n nodes.push(current_node.left_child)\n next_line_count += 1\n end\n if current_node.right_child\n nodes.push(current_node.right_child)\n next_line_count += 1\n end\n if current_line_count == 0\n # finished printing current level\n puts ''\n current_line_count = next_line_count\n next_line_count = current_line_count\n end\n end\nend",
"def tree_results(word_data)\n #get the data ready for d3 view\n tree_data = {\"name\"=> (@topic.name), \"info\" => \"tst\", \"children\" => []}\n \n word_data.each do |text, v|\n tree_data[\"children\"].push({\"name\" => text.to_s, \"children\" => []})\n end\n \n tree_data[\"children\"][0][\"children\"] << Hash[\"name\", word_data[:word]]\n \n word_data[:definitions].each do |text|\n tree_data[\"children\"][1][\"children\"] << Hash[\"name\", text[\"text\"]]\n end\n \n word_data[:word_associations].each do |text|\n tree_data[\"children\"][2][\"children\"] << Hash[\"name\", text[\"relationshipType\"], \"children\", []]\n end \n if word_data[:reverse_definitions][\"results\"].nil? \n tree_data[\"children\"][3][\"children\"] << nil\n else\n word_data[:reverse_definitions][\"results\"].each do |result| \n tree_data[\"children\"][3][\"children\"] << Hash[\"name\", result[\"text\"]]\n end\n end\n i = 0\n word_data[:word_associations].each do |text|\n text[\"words\"].each do |word|\n tree_data[\"children\"][2][\"children\"][i][\"children\"] << Hash[\"name\", word]\n end\n i+=1\n end\n #reduce duplicates in word_association hash\n tree_data[\"children\"][3][\"children\"].uniq!\n tree_data[\"children\"][2][\"children\"].uniq!\n return tree_data\n end",
"def build_tree array\n\t\t@root = Node.new array[0]\n\t\t@nodes += 1\n\t\tarray[1..-1].each do |var|\n\t\t\tinsert(@root,var)\n\t\tend\n\tend",
"def build_tree(model)\n # inflate the node id to test id wrap around edge cases\n ENV[\"NODES\"].to_i.times { model.create!.destroy } if ENV[\"NODES\"]\n\n n1 = model.create!\n n2 = model.create!(:parent => n1)\n n3 = model.create!(:parent => n2)\n n4 = model.create!(:parent => n2)\n n5 = model.create!(:parent => n1)\n n6 = model.create!(:parent => n5)\n\n puts \"create: #{n1.id}..#{n6.id}\" if ENV[\"NODES\"]\n [n1, n2, n3, n4, n5, n6]\n end",
"def build_tree(nodes_fragement)\n nodes_element = LonelyPlanet::Node.new nodes_fragement\n node = LonelyPlanet::TreeNode.new(nodes_element.name, nodes_element.id)\n if nodes_element.has_child?\n nodes_element.children.all? { |child_frag|\n node << build_tree(child_frag)\n }\n end\n node\n end",
"def printf(children=nil)\n text = \"\"\n if (self.root)\n queue = Queue.new\n queue.push(self.root)\n while (!queue.empty?)\n temp = queue.pop\n text << temp.title << \":\" << temp.rating.to_s\n if (temp.left)\n queue.push(temp.left)\n end\n if (temp.right)\n queue.push(temp.right)\n end\n text << \"\\n\"\n end\n puts text\n return\n end\n end",
"def tree_for(stack, level, tree)\n stack.each do |debug_item|\n task = debug_item[0][0]\n\n if debug_item.size == 2 # flat\n introspect = debug_item[0].last\n\n name = (node = introspect[task]) ? node[:id] : task\n\n tree << [ level, name ]\n else # nesting\n tree << [ level, task ]\n\n tree_for(debug_item[1..-2], level + 1, tree)\n\n tree << [ level+1, debug_item[-1][0] ]\n end\n\n tree\n end\n end",
"def print_tree(id)\n puts \"--------------------- Tree Record: #{NewYorkTrees::Tree.all[id].name} ---------------------\"\n puts \"\"\n puts \"Common Name:\"\n puts \"#{NewYorkTrees::Tree.all[id].name.ljust(25)}\"\n puts \"\"\n puts \"Scientific Name:\"\n puts \"#{NewYorkTrees::Tree.all[id].scientific_name.ljust(25)}\"\n puts \"\"\n puts \"Bark:\"\n puts \"#{NewYorkTrees::Tree.all[id].bark.ljust(25)}\"\n puts \"\"\n puts \"Twigs:\"\n puts \"#{NewYorkTrees::Tree.all[id].twigs.ljust(25)}\"\n puts \"\"\n puts \"Winter Buds:\"\n puts \"#{NewYorkTrees::Tree.all[id].winter_buds.ljust(25)}\"\n puts \"\"\n puts \"Leaves:\"\n puts \"#{NewYorkTrees::Tree.all[id].leaves.ljust(25)}\"\n puts \"\"\n puts \"Fruit:\"\n puts \"#{NewYorkTrees::Tree.all[id].fruit.ljust(25)}\"\n puts \"\"\n puts \"Distinguishing Features:\"\n puts \"#{NewYorkTrees::Tree.all[id].distinguishing_features.ljust(25)}\"\n puts \"\"\n puts \"---------------------Description---------------------\"\n\n puts \"#{NewYorkTrees::Tree.all[id].description}\"\n puts \"\"\n end",
"def tree_gallery(gallery)\n str = \"\"\n if @gallery == gallery || @gallery.ancestors.include?(gallery)\n str = \"<ul>\"\n gallery.children.each do |gallery_child|\n str += \"<li>\"\n str += link_gallery_with_number(gallery_child)\n str += tree_gallery(gallery_child)\n str += \"</li>\"\n end\n str += \"</ul>\"\n end\n str\n end",
"def file_structure\n #@root_file = \"C:/App_factory/app_factory\"\n #@root_file = \"C:/Documents and Settings/Luxolo Matoti/My Documents/Exercises\"\n @root_file = \"C:/reports_yml\"\n #@root_file = \"C:/current projects/Kromco mes\"\n #@root_file = \"C:/Documents and Settings/Luxolo Matoti/My Documents/My Pictures\"\n\n#_______________________________________________________________\n tree_builder = ReportTreeBuilder.new\n @tree = tree_builder.build_tree(@root_file) # Store in session state to rebuild location of selected file\n session[:tree] = @tree\n#________________________________________________________________\n\n render :inline => %{\n <% @content_header_caption = \"'#{@root_file}'\" %>\n <% @tree_script = build_file_structure_form(@tree,@tree[0].values[0]) %>\n }, :layout => 'tree'\n end",
"def build_tree(arr)\n #take array, turn into bt with node objs\n return nil if arr.empty?\n\n mid = (arr.size - 1)/2\n current_node = Node.new(arr[mid])\n\n current_node.left = build_tree(arr[0...mid])\n current_node.right = build_tree(arr[(mid+1)..-1])\n \n current_node\n end",
"def build_tree(array)\n\t\t@root = Node.new(array[0])\n\t\ttemp_root = @root\n\n\t\tarray[1..-1].each do |node_value|\n\t\t\tinsert_node(node_value, temp_root)\n\t\tend\n\tend",
"def display\n # tree-font is where we start with the fonts\n # min-tree-font is how far we go down... translates into\n # tree-depth, one depth increment corresponds to 2 font increments\n # absolute-min-tree-font is what you think it is\n init_session\n\t@level = 1\n @maxlevel = session[:'max-level']\t\n @font = session[:'tree-font']\n @individual = Individual.by_uid( params[:uid] ) \n\tif !@individual\n\t redirect_to root_path\n\telsif session[:display] == \"graph\"\n\t is_user = @current_user and @current_user.user?\n\t is_editor = @current_user and @current_user.user?\t \n\t @nodes = [] \n\t @edges = [] \n\t @nodes << @individual.graph_focus_node( is_user, (is_editor and session[:edit] == 'on') )\t\t \n\t @nodes, @edges = @individual.graph_up( \n\t 0, @maxlevel, @nodes, @edges, is_user, (is_editor and session[:edit] == 'on') )\t\t \n\t @nodes, @edges = @individual.graph_down( \n\t 0, @maxlevel, @nodes, @edges, is_user, (is_editor and session[:edit] == 'on') )\t \n\t render :graph\t\n\telse\n\t render\n\tend\n end",
"def gen_tree\n new_tree = {}\n node_list = {}\n @json_tree.each do |k, v|\n if v['child_of'].nil?\n # top\n new_tree[k] = v\n node_list[k] = new_tree[k]\n else\n parent = v['child_of']\n if v['condition'] == 'and'\n node_list[parent]['and'] ||= {}\n node_list[parent]['and'][k] = v\n node_list[k] = node_list[parent]['and'][k]\n elsif v['condition'] == 'or'\n node_list[parent]['or'] ||= {}\n node_list[parent]['or'][k] = v\n node_list[k] = node_list[parent]['or'][k]\n else\n # TODO: sink?\n node_list[parent]['or'] ||= {}\n node_list[parent]['or'][k] = v\n node_list[k] = node_list[parent]['or'][k]\n end\n end\n end\n\n @json_tree_type = 'tree'\n @json_tree = new_tree\n end",
"def build\n reset\n visit(root, @root)\n end",
"def print_tree(indent = 0, lines = [])\n lines << (\" \" * indent) + self.to_s\n @nodes.keys.sort.each do |reference|\n node = @nodes[reference]\n if node.is_a? APISpec::Namespace\n node.print_tree(indent + 1, lines)\n else\n lines << (\" \" * (indent + 1)) + \"#{reference} => #{node.to_s}\"\n end\n end\n lines.join(\"\\n\")\n end",
"def navtree_build(ret, item, el, el_type = \"\")\n module_name = item[:classnames][2] ? item[:classnames][2] : item[:classnames][1];\n page_name = item[:crumb]\n\n if !ret[module_name]\n ret[module_name] = {:id => module_name, :children => Hash.new}\n #if item[:deprecated] ret[module_name][:deprecated] = true\n end\n\n if !ret[module_name][:children][page_name]\n ret[module_name][:children][page_name] = {\n :id => page_name,\n :path => item.identifier,\n :children => Hash.new\n }\n #if item[:deprecated] ret[module_name][:deprecated] = true\n end\n\n el_name = el.respond_to?(:name) ? el.name : el\n ret[module_name][:children][page_name][:children][el_name] = {\n :signature => el.respond_to?(:sass_signature) ? el.sass_signature(:none) : el,\n :el_type => el_type\n }\n\n ret\nend",
"def draw_tree(x, y, length, dir, dx, dy)\n if dir < 0\n dir = dir + 8\n elsif dir >= 8\n dir = dir - 8\n end\n\n xnew = x + (length * dx[dir])\n ynew = y - (length * dy[dir])\n color = branch_color(length)\n\n draw_line(x, y, color, xnew, ynew, color, 1.0)\n\n # recurse\n if length > 0\n draw_tree(xnew, ynew, length - 1, dir - 1, dx, dy)\n draw_tree(xnew, ynew, length - 1, dir + 1, dx, dy)\n end\n end",
"def create_tree(father,tree)\n tree.each do |name|\n n = Meta::create_class(father, name[0], name[1])\n create_tree(n, name[2])\n end\nend",
"def print_paths\n path, counter = [], 0\n puts '************************ Tree Paths ***********************************'\n puts\n print_tree_paths(self.root, path, counter)\n puts\n puts '***********************************************************************'\n end",
"def tree\n return nil if messages.size == 0\n build_tree unless @tree\n @tree\n end",
"def display_segment( node )\n html = \"<li>\".html_safe\n node_class = \"folder\"\n if node.class == Venue\n node_class = \"file\"\n end\n if node.class == Organisation\n node_class = \"web-browser\"\n end\n if node == nil then\n html << \"<span class=\\\"#{node_class}\\\"> <a href=\\\"/location_tree/NIL\\\">Found NIL</a> </span>\".html_safe\n html << \"<ul id=\\\"children_of_NIL\\\">\".html_safe\n\n else\n html << \"<span class=\\\"#{node_class}\\\"> <a href=\\\"/location_tree/#{node.id}\\\">#{node.to_s}</a> </span>\".html_safe\n html << \"<ul id=\\\"children_of_#{node.id}\\\">\".html_safe\n node.children.each{|child_node|\n html << display_segment( child_node )}\n end\n html << \"</ul></li>\".html_safe\n end",
"def printf(children=nil)\n h = height(@root) + 1\n root = @root\n h.times do |i|\n self.print_level(root, i)\n end\n end",
"def printf(children=nil)\n children = [@root] if children.nil?\n grand_children = []\n children.each do |node|\n puts \"#{node.title}: #{node.rating}\\n\"\n grand_children.push(node.left) if node.left\n grand_children.push(node.right) if node.right\n end\n printf(grand_children) unless grand_children.empty?\n end",
"def show\n @dossier = Dossier.find(params[:id])\n @acteurs = @dossier.acteurs\n @contact_acteurs = []\n tree = []\n @acteurs.each do |acteur|\n content = {:expanded => true, :cls => \"folder\", :id => acteur.id, :text => acteur.description, :qualite_procedurale => '', :institution=>'', :email => '', :telephone => ''}\n contact_acteurs = []\n acteur.contact_acteurs.each do |conact|\n @contact_acteurs.push(conact)\n if conact.contact\n contact_content = {:id => conact.id, :text => conact.contact.full_name_inc_civilite, :qualite_procedurale => conact.qualite_procedurale.try(:description), :institution=>conact.contact.institution.try(:nom), :email => conact.contact.try(:email), :telephone => conact.contact.try(:telephone), :leaf => true}\n contact_acteurs.push(contact_content)\n end\n end\n content[:children] = contact_acteurs\n tree.push(content)\n end\n \n\n \n respond_to do |format|\n format.html # show.html.erb\n format.xml { render :xml => @dossier }\n format.json { render :json => {:dossier => @dossier.attributes.merge(:institution_nom => @dossier.institution.try(:nom), :type_etat_dossier_description => @dossier.type_etat_dossier.try(:description), :juge_mission_id => @dossier.juge_mission.try(:contact_id), :juge_controlleur_id => @dossier.juge_controlleur.try(:contact_id)), :activites => @dossier.activites, :expenses => @dossier.expenses.map {|p| p.attributes.merge(:total_ht => p.total, :total_ttc => p.total_ttc, :activite_name => p.activite.try(:description))}, :reminders => @dossier.reminders, :documents => @dossier.documents.map {|p| p.attributes.merge(:short_link => p.generate_link,:long_link => p.generate_long_link )}, :acteurs =>@acteurs, :communications => @dossier.communications, :contact_acteurs=>@contact_acteurs, :tree => tree, :consignations => @dossier.consignations}}\n format.pdf {\n html = render_to_string( :action => \"show\")\n kit = PDFKit.new(html, :page_size => 'A4')\n kit = kit.to_pdf\n send_data(kit, :filename => \"labels.pdf\", :type => 'application/pdf', :disposition => 'inline')\n return\n }\n end\n end",
"def tree\r\n @rootNode\r\n end",
"def list_children(taxons_collection)\n taxons_collection.sort_by {|p| p.hierarchy}\n html_var = \"\"\n taxons_collection.each do |t|\n if not t.children.empty?\n html_var << \"<li><i class='icon-plus'> </i>\" << link_to(t.name, t) << \"<ul>\" << list_children(t.children) << \"</ul>\"\n else\n html_var << \"<li><i class='icon-white' style='visibility: hidden;'> </i>\" << link_to(t.name, t)\n end\n html_var << \"</li>\\n\"\n end\n return html_var.html_safe\nend",
"def build_tree\n t = RDTree.new\n t.add_node(0, DummyRoot)\n root = root_node_of\n t.add_edge(0, root.attributes[\"ID\"])\n do_build_tree(root, 1, t) \n t\n end"
] |
[
"0.7314121",
"0.71735543",
"0.7171269",
"0.6967468",
"0.6928821",
"0.69102544",
"0.6896934",
"0.68819135",
"0.6851404",
"0.6805692",
"0.67988634",
"0.6783851",
"0.675101",
"0.67384154",
"0.6703568",
"0.6641606",
"0.66398305",
"0.66274816",
"0.66012764",
"0.6561899",
"0.6561658",
"0.64810663",
"0.6477569",
"0.64580107",
"0.6452229",
"0.6451525",
"0.64345056",
"0.6433858",
"0.6429137",
"0.6417011",
"0.6389864",
"0.63881785",
"0.6386149",
"0.63236004",
"0.6319486",
"0.6312141",
"0.6307179",
"0.62950957",
"0.62912405",
"0.62902707",
"0.62765443",
"0.6276176",
"0.627297",
"0.62675613",
"0.6262332",
"0.62174714",
"0.6213479",
"0.62033755",
"0.619603",
"0.619562",
"0.6189922",
"0.6189308",
"0.6172751",
"0.6164013",
"0.6151649",
"0.61390686",
"0.613372",
"0.6109374",
"0.61061734",
"0.61022586",
"0.6098675",
"0.6094072",
"0.60749197",
"0.6073439",
"0.6056709",
"0.6049839",
"0.6040697",
"0.6031884",
"0.6029835",
"0.60039973",
"0.59960365",
"0.59896845",
"0.598261",
"0.5957131",
"0.5947528",
"0.59445703",
"0.5943909",
"0.5932283",
"0.5929581",
"0.5920075",
"0.59159046",
"0.59113705",
"0.5909831",
"0.5900472",
"0.5897222",
"0.58962446",
"0.589578",
"0.58794415",
"0.58757496",
"0.587509",
"0.58627737",
"0.5861799",
"0.5860949",
"0.58583635",
"0.5847092",
"0.5844454",
"0.5840745",
"0.5836136",
"0.58300734",
"0.58169234",
"0.5807275"
] |
0.0
|
-1
|
i = 1 j = 0 while i = 0 list.insert(current1, list[current]) list.delete_at(current+1) current = 1 end end j+=1 end i+=1 end list end
|
def insertion_sort(arr, &prc)
dir = (prc.call(1,2) == -1) ? :ASC : :DESC
while true
i = 0
j = 1
inserted = false
while j < arr.length
if dir == :ASC
if arr[i] > arr[j]
num = arr.delete_at(i)
arr.insert(j, num)
inserted = true
end
else
if arr[i] < arr[j]
num = arr.delete_at(i)
arr.insert(j, num)
inserted = true
end
end
i += 1
j += 1
end
break if !inserted
end
arr
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def insertion_sort(list)\n for i in 1..list.size - 1\n j = i - 1\n while j >=0 and list[j+1] < list[j] # if current < previous\n list[j], list[j+1] = list[j+1], list[j] # swap previous and current values. faster.\n\n# list[j+1] = list[j] --- typical book way. slower.\n j -= 1\n end\n# list[j+1] = current --- typical book way. slower.\n end\n list\nend",
"def insertion_sort(list)\n return list if list.size < 2\n\n (1...list.length).each do |i|\n j = i - 1\n e_next = list[i]\n while j >= 0 and list[j] > e_next\n list[j + 1] = list[j]\n j -= 1\n end\n list[j + 1] = e_next\n end\n list\n end",
"def insert(list, x)\n list_size = list.size\n if x < list[1]\n if x < list[0]\n return [x] + list\n elsif list_size == 2\n return [list[0]] + [x] + [list[1]]\n else\n return [list[0]] + [x] + list[1..2]\n end\n elsif list_size == 2 || x > list[2]\n return list + [x]\n else\n return list[0..1] + [x] + [list[2]]\n end\n end",
"def insertion_sort(list)\n (1...list.length).each do |i|\n k = i\n while k > 0 && list[k] < list[k-1]\n list[k], list[k-1] = list[k-1], list[k]\n k -= 1\n end\n end\n list\nend",
"def reverse!(list)\n count = 0\n\n list.size.times do\n list.insert(count, list.pop)\n count += 1\n end\n\n list\nend",
"def add_to_list(a)\n i = 0\n while i < @list.length\n if a == a.compareto(list[i])\n @list.insert(i, a)\n #puts \"inserted\"\n return list\n end\n i = i+1\n end\n @list.push(a) \n #puts \"at end\"\n end",
"def insertion2 arr\n for i in 1..(arr.size-1)\n cur_el = arr[i]\n j = i -1\n while j >= 0 and arr[j] > cur_el\n arr[j+1] = arr[j]\n j = j-1\n end\n arr[j+1] = cur_el\n puts arr.join(' ')\n end\n \nend",
"def remove_dupes list\n cur = list.head\n i = 0\n while cur != nil\n list.remove i\n if !contains(list, cur.node)\n list.insert cur.node, i\n i += 1\n end\n cur = cur.nxt\n end\n\n return list\nend",
"def insertion_sort(a)\n a.each_with_index do |el,i|\n # Insert el into the sorted sublist\n j = i - 1\n while j >= 0\n break if a[j] <= el\n a[j + 1] = a[j]\n j -= 1\n end\n a[j + 1] = el\n\n end\nend",
"def swap_with_next i\n #if first_item >= first_item.next_list_item\n\n end",
"def shift_left(list)\n list_new = list\n i = 0\n first = list[0]\n\n list.size.times do \n list_new[i] = list[i +1]\n i += 1\n end \n list_new[list.size - 1] = first\n\n return list_new\n \nend",
"def shift_left2(list)\n list_new = list\n i = 0\n first = list[0]\n\n list.size.times do |i|\n list_new[i] = list[i + 1]\n i += 1\n end\n list_new[list.size - 1] = first\n\n return list_new\n\nend",
"def insertion(a)\n\tlen = a.length\t\n\tfor i in 0..(len-1) #going through each element\n\t\tfor j in i..0 #elements go from right to left <-\n\t\t\tif a[j-1] > a[j]\n\t\t\t\ttemp = a[j-1]\n\t\t\t\ta[j - 1] = a[j]\n\t\t\t\ta[j] = temp\n\t\t\tend\n\t\tend\n\tend\n\treturn a\n\nend",
"def solution(a, k)\n # write your code in Ruby 2.2\n \n unless a.empty?\n for i in 1..k\n last = a.pop\n a.insert(0, last)\n end\n end\n \n return a\nend",
"def merged(list1, list2)\n total = list1.size + list2.size\n merged = [list1 + list2]\n fihish = []\n merged.each do |i|\n if i < (i + 1)\n finish.push(i)\n elsif i == (i+1) \n finish.push(i, i+1)\n i += 1 #is this how I skip the next interval?\n end\n end\n print finish \nend",
"def insertion_sort_two(arr)\n 1.upto(arr.length - 1) do |i|\n value = arr.delete_at(i)\n j = i - 1\n j -= 1 while j >= 0 && value < arr[j]\n arr.insert(j + 1, value)\n end\n arr\nend",
"def insertion_sort!(array)\n array.each_with_index do |el, i|\n j = i\n j -= 1 while j > 0 && el < array[j-1]\n array[i] = array[i-1] and i -= 1 while i > j\n array[j] = el\n end\n\n array\nend",
"def shoppingList\n a = []\n x = 1\n y = 1\n puts \"insert an item\"\n 5.times do\n a << gets.strip.upcase\n \n end\n b = a.sort.uniq\n\n 5.times do \n \n b.insert(x,y)\n x += 2\n y += 1\n # binding.pry\n end\n h=Hash[*b.flatten(1)]\n puts h\n end",
"def next_list\n @list_index += 1\n\n if @lists.size < @list_index\n @lists << []\n end\n\n @co_index = 1\n\n nil\n end",
"def increment_new_list(new_list, old_list, if_origin)\n list = Hash.new\n new_list.map { |k, v| v.to_i == 1 ? list[k+'s'] = v.to_i : list[k] = v.to_i }\n old_list.each do |k,v|\n if v.to_i == 1\n list[k+'s'] ? list[k+'s'] += v.to_i : list[k] = v.to_i\n else\n list[k] = list[k] ? list[k] + v.to_i : v.to_i\n end\n end\n if_origin ? old_list = list : list\n end",
"def switch_pairs(list)\n new_list = []\n (list.length / 2).times do\n new_list << list[1]\n new_list << list[0]\n list.delete_at(0)\n list.delete_at(0)\n end\n if list[0]\n new_list << list[0]\n end\n return new_list\nend",
"def slide(i, l)\n \"[2, 2, 3, 2, 2] => [2, 2, 2, 3, 2]\"\n if (i + 1) < l.length\n l[i + 1] = 3\n l[i] = 2\n end\n l\nend",
"def add_to_list(interval)\n # edge case \n if (interval.begin_t < @list[0].begin_t)\n @list.insert(0, IntervalCounter.new(interval.begin_t, @list[0].begin_t, interval.items))\n end\n # other cases\n i = 0\n while @list[i]\n if (interval.begin_t < @list[i].begin_t && interval.end_t > @list[i].end_t)\n @list[i].items = increment_new_list(interval.items, @list[i].items, true)\n \n elsif (interval.end_t > @list[i].begin_t && interval.end_t < @list[i].end_t)\n items_list = increment_new_list(interval.items, @list[i].items, false)\n time_slot = IntervalCounter.new(@list[i].begin_t, interval.end_t, items_list)\n @list[i].begin_t = interval.end_t\n @list.insert(i, time_slot)\n \n elsif (interval.begin_t > @list[i].begin_t && interval.begin_t < @list[i].end_t)\n items_list = increment_new_list(interval.items, @list[i].items, false)\n time_slot = IntervalCounter.new(interval.begin_t, @list[i].end_t, items_list)\n @list[i].end_t = interval.begin_t\n @list.insert(i + 1, time_slot)\n \n elsif (interval.begin_t > @list[i].begin_t && interval.end_t < @list[i].end_t)\n time_slot = IntervalCounter.new(interval.end_t, @list[i].end_t, list[i].items)\n @list[i].end_t = interval.begin_t\n items_list = increment_new_list(interval.items, @list[i].items, false)\n @list.insert(i + 1, IntervalCounter.new(interval.begin_t, interval.end_t, items_list))\n @list.insert(i + 1, time_slot)\n end\n if (interval.end_t > @list[@list.length-1].end_t) \n @list.insert(@list.length, IntervalCounter.new(@list[@list.length-1].end_t, interval.end_t, interval.items))\n end\n i += 1\n end\n @list\n end",
"def merge(list1,list2)\n x1 = list1.length\n x2 = list2.length\n out = []\n if x1 == 0 || x2 == 0 ; return out = list1 + list2 else\n x = x1 + x2 # add both lengths for # of steps\n x1 -= 1 # -1 to equal index.last\n x2 -= 1 # ^\n while x > 0 \n if list1[x1] > list2[x2] \n if x1 > -1; out.unshift(list1[x1])\n else out.unshift(list2[x2]); x2 -= 1 end\n x1 -= 1 \n elsif list1[x1] < list2[x2]\n if x2 > -1; out.unshift(list2[x2])\n else out.unshift(list1[x1]); x1 -= 1 end\n x2 -= 1\n else \n out.unshift(list2[x2],list1[x1])\n x1 -= 1\n x2 -= 1\n x -= 1 # extra count down cuz pushing 2 equal elements\n end\n x -= 1\n end\n out\n end\nend",
"def printList(list)\n\tputs \"____CURRENT LIST____\"\n\t\tlist.each do |x|\n\t\t\tputs (list.index(x)+1).to_s + \". \" + x\n\t\tend\n\tputs \"____CURRENT LIST____\"\nend",
"def insertion_sort(arr)\n step = arr.size - 1\n \n 1.upto(step).each do |i|\n current = arr[i]\n index = i\n \n (i-1).downto(0).each do |j|\n break if arr[j] < current\n \n arr[j], arr[j+1] = arr[j+1], arr[j]\n index = j\n end\n \n arr[index] = current\n end\n \n arr\nend",
"def insertion_sort(array)\n array.each_with_index do |item, index|\n new_card = array[index]\n @index_position = index\n array[0..index].each_with_index do |_item, index_2|\n next if index == 0\n compare_index = index - index_2\n if new_card < array[index - index_2]\n array.insert(compare_index, array.delete_at(@index_position))\n p \"Intermediate array step: #{array}\"\n @index_position = compare_index\n end\n end\n p array\n end\nend",
"def insert_sort(start_arr=[])\n index = 1\n (start_arr.length-1).times do \n current = start_arr[index]\n sub_arr_index = index - 1 \n while sub_arr_index >= 0 && start_arr[sub_arr_index]>current do\n start_arr[index] == start_arr[sub_arr_index]\n sub_arr_index += 1 \n end\n index +=1\n end\n return start_arr\nend",
"def insertion_sort(array) \r\narray.each_with_index do |el,i| \r\n j = i - 1 \r\n while j >= 0 \r\n break if array[j] <= el \r\n array[j + 1] = array[j] \r\n j -= 1 \r\n end \r\n array[j + 1] = el \r\nend \r\nend",
"def create_list arr\n list = SinglyLinkedList.new\n arr.each do |i|\n list.insert_at_end i\n end\n list\nend",
"def create_list arr\n list = SinglyLinkedList.new\n arr.each do |i|\n list.insert_at_end i\n end\n list\nend",
"def create_list arr\n list = SinglyLinkedList.new\n arr.each do |i|\n list.insert_at_end i\n end\n list\nend",
"def insertionSort1(n, arr)\n l = arr.length - 1\n last = arr[l]\n\n while l-1 >= 0 && arr[l-1] > last\n if last < arr[l-1] \n arr[l] = arr[l-1] \n arr.each{|v| print \"#{v} \"}\n puts\n end\n l -= 1\n end\n arr[l] = last\n arr.each{|v| print \"#{v} \"} \nend",
"def insertionSort(input)\n for j in 1..input.length - 1\n temp = input[j]\n i = j - 1\n\n while i >= 0 && input[i] > temp\n input[i + 1] = input[i]\n i -= 1\n end\n input[i + 1] = temp\n end\n return input\nend",
"def insert_final(*val)\n \n val.each do |nuevo_nodo|\n \n if @tail != nil\n @tail.next = nuevo_nodo\n nuevo_nodo.previous = @tail\n @tail = nuevo_nodo\n else\n @head = @tail = nuevo_nodo\n end\n @num_nodos += 1\n end\n end",
"def insertion_sort(array)\n # setting an array called final to [whatever is at the first position of the array]; an array of length 1\n final = [array[0]]\n # deleting the first index of the array so that it doesn't get considered in the while loop, since it is already in final array\n array.delete_at(0)\n # iterating through the array\n array.each do |i|\n # setting a value of 0 to a variable called final_index\n final_index = 0\n # while final_index value is less than the length of final array\n while final_index < final.length\n # if i is less than or equal to final[0]\n if i <= final[final_index]\n # then insert i into final array at position indicated by final_index value (i.e., i swaps positions with the value at the index indicated by final_index)\n final.insert(final_index, i)\n # breaks the loop when if returns true\n break\n # else if final_index equals the length of final - 1\n elsif final_index == final.length - 1\n # then insert i into final array at position indicated by final_index value plus 1 (i.e, i gets inserted behind the value at final_index plus 1)\n final.insert(final_index + 1, i)\n # breaks the loop when elsif returns true\n break\n end\n # add 1 to final_index each time it runs through the loop\n final_index += 1\n end\n end\n # puts final array\n final\nend",
"def insertionsort! arr\n (1..arr.size - 1).each do |i|\n j = i\n while j > 0 && arr[j] < arr[j - 1] do\n arr[j], arr[j - 1] = arr[j - 1], arr[j]\n j -= 1\n end\n end\n arr\nend",
"def insertion_sort(array)\t\t\t\t\t\t\t\t\t\t\t\t#This is not my code. I do not understand this code. I tried very hard to understand it but could not.\r\n final = [array[0]]\r\n array.delete_at(0)\r\n # main code\r\n for i in array\r\n final_index = 0\r\n while final_index < final.length\r\n if i <= final[final_index]\r\n final.insert(final_index,i)\r\n break\r\n elsif final_index == final.length-1\r\n final.insert(final_index+1,i)\r\n break\r\n end\r\n final_index+=1\r\n end\r\n end\r\n # output\r\n final\r\nend",
"def insertion_sort_one(arr)\n 1.upto(arr.length - 1) do |i|\n value = arr[i]\n j = i - 1\n while j >= 0 and arr[j] > value\n arr[j + 1] = arr[j]\n j -= 1\n end\n arr[j + 1] = value\n end\n arr\nend",
"def reverse!(list)\n index = list.length\n loop do\n break if index == 0\n list << list.delete_at(index -= 1)\n end\n list\nend",
"def shift_left(list)\n a = list\n a.push(list[0])\n a.shift\n a\nend",
"def sort2(collection)\n # p \"sort 2\"\n # p collection\n collection[1..-1].each_with_index do |el, i|\n k = i + 1\n # p \"******/***** i is: #{i} ************\"\n until collection[k] > collection[k-1] or k == 0\n # p \"k is: #{k}\"\n # p \"swapping #{collection[k]} with #{collection[k-1]}\"\n collection[k], collection[k-1] = collection[k-1], collection[k]\n # p collection\n # p \"subtracting 1 from k\"\n k -= 1\n end\n end\n collection\nend",
"def destutter2(seq)\n result = [] #result will be a new array\n last = nil #keep track of last thing\n\n seq.each do |item|\n if item != last\n result.push(item)\n #result << item\n end\n last = item\n end\n result\nend",
"def organize_list\n\t\t@list.each_with_index do |node, index|\n\t\t\tnode.next_node = @list[index + 1] if index < (@list.length - 1)\n\t\tend\n\t\t@list[-1].next_node = nil\n\tend",
"def insertion_sort(arr)\n (arr.length - 1).times do |index|\n index += 1\n position = index\n temp_value = arr[index]\n\n while position > 0 && arr[position - 1] > temp_value\n arr[position] = arr[position - 1]\n position -= 1\n end\n arr[position] = temp_value\n end\n arr\nend",
"def increment_position\n return unless in_list?\n set_list_position(current_position + 1)\n end",
"def ordenar_for\n\t @lista = self.map{ |a| a }\n\t \tfor i in (0..@lista.count-1)\n\t\t\tfor j in (0..@lista.count-1)\n\t\t\t\tif j+1 != @lista.count\n if @lista[j+1] < @lista[j]\n\t\t\t\t\t\t@lista[j],@lista[j+1] = @lista[j+1],@lista[j]\n \t\t\t\tend\n\n\t\t\t\tend\n\t\t\tend\n\t\tend\n\t\t@lista\n end",
"def rotate(nums, k)\n for i in 0...k\n nums.insert(0,nums.last)\n nums.pop\n end\nend",
"def insert(x,a)\r\n c = a.dup\r\n c.push(x)\r\n c = c.sort\r\nend",
"def shell_sort(list)\n gap = list.size / 2\n while gap > 0 do\n for i in gap..list.size - 1 \n j = i \n while j >= gap and list[j] < list[j-gap] # if current < previous. previous = gap values to the left\n list[j], list[j-gap] = list[j-gap], list[j] # swap current and previous\n j -= gap # keep swapping current and previous until previous > current\n end\n end\n gap = gap / 2\n end\n list\nend",
"def partition list, x\n # TODO implement size property in list class\n cur = list.head\n size = 0\n while cur != nil\n size += 1\n cur = cur.nxt\n end\n\n cur = list.head\n i = 0\n while i < size\n mv = cur\n list.remove(i)\n if mv.node < x\n list.insert(mv.node, 0)\n else\n list.insert(mv.node, size - 1)\n end\n cur = cur.nxt\n i += 1\n end\n\n return list\nend",
"def prep_sorted_lists\n 10.times do |i|\n list1.add(3*i)\n list2.add(5*i)\n end \n list1.print\n list2.print \n end",
"def insert_at(v, list, n)\n list.dup.insert(n, v)\nend",
"def remove_duplicates(list)\n if list.length == 0\n return []\n end\n \n length = list.length\n i = 0\n j = 0\n while i < length-1\n if list[i] != list[i+1]\n list[j] = list[i]\n j += 1\n end\n i += 1\n end\n list[j] = list[length - 1]\n \n ((j+1)..(length-1)).each do |index|\n list.pop()\n end\n \n # list.each_with_index do |item,index|\n # if item == nil\n # list.delete_at(index)\n # end\n # end\n \n # list = list[0..j]\n return list\nend",
"def delmid(linked_list)\n curr = linked_list\n count = 0\n\n while curr != null\n count += 1\n curr = curr.next\n end\n\n return linked_list if count <= 2\n mid = (count/2).round()\n curr = linked_list\n\n mid.each do |num|\n curr = curr.next\n end\n\n return curr\nend",
"def remove_duplicates(list)\n list.sort!\n first_element = list.first\n previous_element = list.first\n list[1..-1].map do |element|\n element = nil if element == previous_element\n previous_element = element\n end.compact.unshift(list.first)\nend",
"def shift_left(list)\n overflow = 0\n overflow = list[0]\n point = 0\n list.size.times do\n list[point] = list[point + 1]\n point += 1\n end\n list[list.size - 1] = overflow\n return list\nend",
"def swap_elements(array)\n array.insert(1,array.delete_at(2))\nend",
"def removeDup(list)\n hash = {}\n count = list.count\n i = 0\n while i < count\n key = list[i].key\n if hash[key]\n list.remove_pos(i)\n i = i-1\n count = count-1\n else\n hash[key] = true\n end\n i = i+1\n end\nend",
"def insertionSort(array)\n temp = 0\n l = array.length - 1\n (1..l).each do |i|\n j = i\n j.downto(1) do |j|\n if array[j] < array[j-1]\n temp = array[j]\n array[j] = array[j-1]\n array[j-1] = temp\n end\n end\n end\n puts \"Sorted array:\"\n puts array\nend",
"def merge(list,list2)\n i = 0\n merged = []\n finalmerged = []\n usmerged = []\n while i < list.length - 1\n if list[i] > list2[i]\n merged.push(list2[i] , list[i])\n else\n merged.push(list[i] , list2[i])\n end\n i += 1\n sort = true\n end\n (merged.length - 1).times do |i|\n if merged[i] <= merged[i+1]\n usmerged.push(merged[i],merged[i+1])\n else\n usmerged.push(merged[i+1],merged[i])\n end\n end\n return usmerged\nend",
"def insert(element)\n insertindex=@store.length\n\n @store[@store.length]=element\n return if insertindex==1\n\n #For even inserts parent index is insertedindex/2-1 otherwise its juts insertedindex/2\n\n parentindex=insertindex/2\n\n #Incase it does not fit in(violating Heap Property)\n while insertindex!=1 && @store[parentindex] > @store[insertindex]\n swap(parentindex,insertindex)\n insertindex=parentindex\n parentindex=insertindex/2\n end\n\n end",
"def merge(list_1, list_2)\n com_list = list_1 + list_2\n new_list = list_1 + list_2\n i_2 = 0\n new_list.each do |a|\n i = 0\n new_list.each do |b|\n if a >= b\n new_list[i_2] = com_list[i]\n end\n i += 1\n end\n i_2 += 1\n end\n return new_list\nend",
"def shift_left(list) # make the last element the first elem\n return list.unshift(list.pop)\nend",
"def insert_at(index, insertion)\n prev = get_index(index)\n post = get_index(index+1)\n\n #if index of insertion is bigger than list size it will insert at the end\n if prev.nil?\n @tail.next_node = insertion\n @tail = insertion #problem alert\n @tail.next_node = nil\n else\n prev.next_node = insertion\n insertion.next_node = post\n end\n end",
"def remove_dups(list)\n current = list.head\n until current.next.nil?\n compare = current.next\n until compare.next.nil?\n if compare.value == current.value\n # remove duplicate\n befr = compare.prev\n aftr = compare.next\n befr.next = aftr\n aftr.prev = befr\n end\n compare = compare.next\n end\n current = current.next\n end\nend",
"def mirror(list)\n list_copy = list.dup\n index_tracker = list.length - 1\n until index_tracker == 0\n list_copy << list[index_tracker]\n index_tracker -= 1\n end\n return list_copy\nend",
"def myNextFunc(list, last)\n list << last\nend",
"def find_and_replace(list,user,score,k)\n\t\tto_insert=OpenStruct.new(:user=>user,:score=>score)\n\t\tif list.length<k\n\t\t\tlist.push(to_insert)\n\t\telse\n\t\t\tminium=list[0]\n\t\t\tlist.each_index do |i|\n\t\t\t\tif list[i].score<to_insert.score #swap the insert one with current element\n\t\t\t\t\ttemp=list[i]\n\t\t\t\t\tlist[i]=to_insert\n\t\t\t\t\tto_insert=temp\n\t\t\t\tend\n\t\t\tend\n\t\tend\n\tend",
"def remove_duplicates!\n \t\t# current will move forward 1 each step\n \t\tcurrent = @head\n\n \t\t# prev will hold on to the last non-duplicate node\n \t\tprev = @head\n\n \t\twhile (current.next_node)\n \t\t\t# if we find a new value...\n \t\t\tif prev.val != current.val\n \t\t\t\t# add this current node to our non-duplicate list\n \t\t\t\tprev.next_node = current\n \t\t\t\t# update prev to store this most recent non-duplciate node\n \t\t\t\tprev = current\n \t\t\tend\n \t\t\t\n \t\t\t# always advance current\n \t\t\tcurrent = current.next_node\n \t\tend\n \t\t# finally, add tail value if it's unique\n \t\tif prev.val != current.val\n\t\t\t# add this current node to our non-duplicate list\n\t\t\tprev.next_node = current\n\t\t\t@tail = current\n\t\telse\n\t\t\tprev.next_node = nil\n\t\t\t@tail = prev\n\t\tend\n \tend",
"def insertion_sort(n,a)\n puts \"Insertion Sort\"\n for i in 1..n-1\n k = a[i]\n j = i-1\n while j>=0 && k < a[j] do\n a[j+1] = a[j]\n j = j-1\n end\n a[j+1] = k\n end\n for i in 0..n-1\n print a[i]\n end\n print \"\\n\"\nend",
"def remove_duplicates(list)\n if list.length < 2\n return list\n else\n i = 0\n list.each do |num|\n if num == list[(i+1)]\n list.delete_at(i)\n end\n i += 1\n end\n return list\n end\nend",
"def insertion_sort(array)\n # Takes n passes here.\n (1...array.size).each{ |i|\n j = i\n # and this also takes n, but the distance of the count decreases\n # each time.\n while j > 0 and array[j - 1] > array[j]\n temp = array[j]\n array[j] = array[j-1]\n array[j-1] = temp\n\n j -= 1\n end\n }\n array\nend",
"def swap_elements(array)\n array.insert(1, array.delete_at(2))\nend",
"def shift_left(list)\n i = 0\n shifted = []\n save = []\n save[0] = list[0]\n list.each do |num|\n if i == 0\n\n else\n shifted[i - 1] = num\n end\n i += 1\n end\n shifted[list.size - 1] = save[0]\n return shifted\nend",
"def insert_at(data, index)\n current = @head\n index.times do\n current = current.next\n end\n save = current.next.clone\n current.next = nil\n append(data)\n current.next.next = save\n end",
"def remove_duplicates(list)\n length = list.length\n i = 0\n while i < length - 1\n if list[i] == list[i + 1]\n list.delete_at(i + 1)\n end \n i += 1\n end \n return list \nend",
"def create_list\n list = SinglyLinkedList.new\n list.insert_at_start 1\n list.insert_at_start 2\n list.insert_at_start 3\n list.insert_at_end 4\n list.insert_at_start 4\n list.insert_at_end 5\n list\nend",
"def insertion_sort(a)\n for i in 1...(a.length)\n\n while i>0\n if a[i] < a[i-1]\n temp = a[i-1]\n a[i-1] = a[i]\n a[i] = temp\n else\n break\n end\n i = i-1\n end\n end\n return a\nend",
"def bubble(list)\n return nil if list.empty?\n sorted = false \n \n until sorted\n sorted = true \n \n (list.length - 1).times do |idx|\n \n if list[idx] > list[idx + 1]\n list[idx], list[idx + 1] = list[idx + 1], list[idx]\n sorted = false\n end \n end \n end\n \n list[0]\nend",
"def insertion_sort(array)\n binding.pry\n (1..array.length - 1).each do |n|\n number_to_insert = array[n]\n j = n - 1\n while array[j] > number_to_insert && j >= 0\n array[j + 1] = array[j]\n j -= 1\n end\n array[j + 1] = number_to_insert\n end\n array\nend",
"def iterative_insertion_sort(arr)\r\n 1.upto(arr.length - 1) do |index|\r\n key = arr[index]\r\n pos = index - 1\r\n while pos >= 0 && key <= arr[pos]\r\n # this could also have been written by swapping repeatedly\r\n # instead of copying the bigger element to the empty slot\r\n arr[pos + 1] = arr[pos]\r\n pos = pos - 1\r\n end\r\n arr[pos + 1] = key\r\n end\r\n arr\r\nend",
"def insert(index, element)\n raise OutOfBoundsException unless valid_index?(index)\n\n temp_list = array_list\n @array_list = FixedArray.new(length + 1)\n (index -1).times do |below_index|\n array_list.set(below_index, temp_list.get(below_index))\n end\n\n array_list.set(index,element)\n\n (index + 1).upto(length - 1) do |above_index|\n @array_list.set(above_index, temp_list.get(index))\n end\n\n element\n end",
"def create_list\n list = SinglyLinkedList.new\n list.insert_at_end 1\n list.insert_at_end 2\n list.insert_at_end 3\n list.insert_at_end 4\n list.insert_at_end 5\n list.insert_at_end 6\n list.insert_at_end 7\n list\nend",
"def remove_duplicate_in_place(head)\n head = LinkedListNode.from_array(head.to_array.sort)\n current = head\n while (current.next_item != nil) do\n if current.next_item.data != current.data\n current = current.next_item\n else\n current.next_item = find_different_next(current)\n end\n end\n\n head\nend",
"def insertion_sort(array)\n\tfinal = [array[0]]\n\tarray.delete_at(0)\n\t#main code\n\tfor i in array\n\t\tfinal_index = 0\n\t\twhile final_index < final.length\n\t\t\tif i <= final[final_index]\n\t\t\t\tfinal.insert(final_index, i)\n\t\t\t\tbreak\n\t\t\telsif \n\t\t\t\tfinal_index == final.length-1\n\t\t\t\tfinal.insert(final_index+1, i)\n\t\t\t\tbreak \n\t\t\tend\n\t\t\tfinal_index +=1\n\t\tend\n\tend\n\t#output\n\tfinal\nend",
"def bubble_up(i)\n while i > 0\n parent = (i+1) / 2 - 1\n if @contents[parent].key >= @contents[i].key then\n @contents[parent],@contents[i] = @contents[i],@contents[parent]\n i = parent\n else return\n end\n end\n end",
"def insertionSort(ia)\n #initialize value and j\n value = 0\n j = 0\n\n for i in 1..(ia.length-1) #for i in indexes of the array...\n value = ia[i]; #value is the current position\n j = i - 1 #j is the index of one less than the current position\n\n while (j >= 0 && ia[j] > value) #while j is not 0 and there is still something bigger than value...\n #move value down\n ia[j + 1] = ia[j]\n j -= 1\n end\n\n ia[j + 1] = value #finally set value where it should be.\n end\nend",
"def list_remove_test\n\tputs \"Testing the Remove index method\"\n\tlist = LinkedList.new()\n\tcount = 9\n\twhile (count >= 0)\n\t\tlist.push(count)\n\t\tcount -= 1\n\tend\n\tputs \"After initialization:\"\n\tlist.toString()\n\tlist.remove_index(9)\n\tlist.remove_index(0)\n\tlist.remove_index(3)\n\tputs \"After removing 10th,1st, and 4th elements (9, 0, 4 should be gone)\"\n\tlist.toString()\n\tcount = 10\n\twhile (count > 0)\n\t\tlist.remove_index(0)\n\t\tcount -= 1\n\tend\n\tputs \"After removing the 0th element a ton:\"\n\tlist.toString\nend",
"def bubble_sort(numbers)\n permutation = nil\n\n while permutation != 0 do # N times\n permutation = 0\n\n numbers.each_with_index do |number, index| # N times\n next if numbers[index + 1].nil?\n\n if number > numbers[index + 1]\n numbers.insert(index + 1, numbers.delete_at(index))\n permutation += 1\n end\n end\n end\n\n numbers\nend",
"def remove(index)\n @size -= 1\n if index == 0\n @first_item = @first_item.next_list_item\n else\n previous_item = get_item( index -1)\n next_list_item = previous_item.next_list_item.next_list_item\n previous_item.next_list_item = next_list_item\n #get_item( index - 1).next_list_item = nil\n # return @first_item = @first_item.next_list_item if index == 0\n # item = @first_item\n # (index - 1).times do\n # item = item.next_list_item\n # # puts \"(((((((()))))))))))))((((((((((((((@item))))))))))))))\"\n # # puts @first_item.payload\n # # puts\"((((((((((((((item.next_list_Item)))))))))))))\"\n # # puts item.payload\n # end\n # raise IndexError if item.nil?\n # item.next_list_item = item.next_list_item.next_list_item\n end\nend",
"def insertion_sort(a)\n\tfor i in 1...(a.length)\n\t\tj=i\n\t\twhile j>0\n\t\t\tif a[j-1]>a[j]\n\t\t\t\ttemp=a[j]\n\t\t\t\ta[j]=a[j-1]\n\t\t\t\ta[j-1]=temp\n\t\t\telse\n\t\t\t\tbreak\n\t\t\tend \n\t\t\tj=j-1\n\t\tend\n end \n return a \nend",
"def insertion_sort(array)\n (1...array.size).each do |i|\n value = array[i]\n j = i - 1\n\n while j >= 0 && array[j] > value\n array[j + 1] = array[j] # shift the value to the right\n j -= 1\n end\n\n array[j + 1] = value # `j + 1` is equal to `i` if the `while` wasn't entered\n end\n\n array\nend",
"def insertion_sort(arr)\n i = 1\n while i < arr.length\n k = i - 1\n num = arr[ i]\n while k >=0 && num < arr[k]\n arr[k], arr [k + 1] = num, arr[ k]\n k -= 1\n end\n i += 1\n end\n arr\nend",
"def insertion_sort(collection)\n sorted_collection = [collection.delete_at(0)]\n\n for val in collection\n sorted_collection_index = 0\n while sorted_collection_index < sorted_collection.length\n if val <= sorted_collection[sorted_collection_index]\n sorted_collection.insert(sorted_collection_index, val)\n break\n elsif sorted_collection_index == sorted_collection.length - 1\n sorted_collection.insert(sorted_collection_index + 1, val)\n break\n end\n\n sorted_collection_index += 1\n end\n end\n\n sorted_collection\n end",
"def reverse!(list)\n list2 = []\n loop do\n break if list.empty?\n\n element = list.pop\n list2 << element\n end\n list.replace(list2)\nend",
"def rollYourOwnSort(list)\n for i in 1..(list.length - 1)\n val = list[i]\n j = i - 1\n\n while j >= 0 && list[j] > val\n list[j + 1] = list[j]\n j -= 1\n end\n\n list[j + 1] = val\n end\nend",
"def update_current(next_idx)\n @current = @list[next_idx]\n end",
"def refresh\n return if first.nil?\n\n i = first.index\n while at(i).exists?\n i += 1\n end\n i -= 1\n\n @last = at(i)\n end",
"def sort3(collection)\n # p \"sort 3\"\n # p collection\n collection[1..-1].each_with_index do |el, i|\n k = i + 1\n # p \"*********** i is: #{i} ************\"\n while collection[k] < collection[k-1]\n # p \"k is: #{k}\"\n # p \"swapping #{collection[k]} with #{collection[k-1]}\"\n collection[k], collection[k-1] = collection[k-1], collection[k]\n # p collection\n # p \"subtracting 1 from k\"\n k -= 1\n break if k == 0\n end\n end\n collection\nend",
"def merge_k_lists(arr)\n new_head, cur = null, null\n\n until arr.empty?\n next_val, idx = nil, 0\n\n arr.each_with_index do |node, i|\n next if node == nil\n if next_val == nil || node.val < next_val\n next_val = node.val\n idx = i\n end\n end\n\n if arr[idx] == nil\n arr.delete_at(idx)\n next\n elsif new_head == nil\n new_head = ListNode.new(next_val)\n cur = new_head\n else\n cur.next = ListNode.new(next_val)\n cur = cur.next\n end\n arr[idx] = arr[idx].next\n end\n new_head\nend"
] |
[
"0.71921366",
"0.69812787",
"0.681147",
"0.6775793",
"0.66426057",
"0.6620658",
"0.64684635",
"0.6422278",
"0.6330906",
"0.63046795",
"0.6223973",
"0.6183446",
"0.6178061",
"0.615965",
"0.6151347",
"0.613927",
"0.61380535",
"0.6102281",
"0.6057625",
"0.60569406",
"0.60525376",
"0.60482186",
"0.6040985",
"0.60356563",
"0.60323256",
"0.60123795",
"0.6004187",
"0.59975123",
"0.59854376",
"0.59760565",
"0.59760565",
"0.59760565",
"0.5968385",
"0.59561574",
"0.5955886",
"0.59540987",
"0.59495634",
"0.5944179",
"0.5943313",
"0.5920835",
"0.5915767",
"0.58972913",
"0.5885736",
"0.5885633",
"0.58846796",
"0.58787084",
"0.5863852",
"0.58554405",
"0.58276194",
"0.58213073",
"0.58198017",
"0.58195573",
"0.5818411",
"0.5811635",
"0.5805792",
"0.58036286",
"0.58034873",
"0.57957643",
"0.5792987",
"0.5776737",
"0.5769444",
"0.5767351",
"0.57525325",
"0.5749297",
"0.5734988",
"0.57245713",
"0.57143605",
"0.57096905",
"0.568863",
"0.5684682",
"0.5683635",
"0.5681619",
"0.5670667",
"0.56705445",
"0.56610787",
"0.5660282",
"0.56589156",
"0.5653808",
"0.5653504",
"0.5652395",
"0.56439155",
"0.5641514",
"0.56398726",
"0.56389904",
"0.56372035",
"0.56348467",
"0.5633684",
"0.5632157",
"0.56186503",
"0.56140536",
"0.561394",
"0.5612369",
"0.5605454",
"0.5593589",
"0.5592441",
"0.55913967",
"0.55874914",
"0.5580688",
"0.5572234",
"0.55673957",
"0.556681"
] |
0.0
|
-1
|
We need to define it because of openurl method received
|
def open
object.open
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def open?(uri); end",
"def URI(url); end",
"def consume_url; end",
"def openurl(url)\n runcmd 'openurl', url\n end",
"def open(url)\n Net::HTTP.get(URI.parse(url))\nend",
"def open(url)\n Net::HTTP.get(URI.parse(url))\nend",
"def open_url(url)\t\n\t\tputs \"Open url #{url} by creating an open_uri object. Return the reference upon success.\" if @verbose\n\t\t#url_object = nil\n begin\n\t\t\tif url =~ /http\\:/i\n\t\t\t\t# patch for allow the 'un-safe' URL redirection i.e. https://www.example.com -> http://www.example.com\n\t\t\t\turl_object = open(url, :allow_redirections=>:safe, :read_timeout=>Max_http_timeout/1000)\n\t\t\t\t#url_object = open(url)\n\t\t\telsif url =~ /https\\:/i\n\t\t\t\turl_object = open(url,:ssl_verify_mode => 0, :allow_redirections =>:safe, :read_timeout=>Max_http_timeout/1000)\n\t\t\t\t#url_object = open(url,:ssl_verify_mode => 0)\n\t\t\telse\n\t\t\t\traise \"Invalid URL format - please specify the protocol prefix http(s) in the URL: #{url}\"\n\t\t\tend\n\t\t\treturn url_object\n rescue => ee\n puts \"Exception on method #{__method__} for #{url}: #{ee}\" if @verbose\n return nil\n end \n end",
"def url\n raise\n end",
"def ensure_http \n url.ensure_http\n end",
"def url=(_arg0); end",
"def url=(_arg0); end",
"def url=(_arg0); end",
"def url=(_arg0); end",
"def url=(_arg0); end",
"def url=(_arg0); end",
"def url=(_arg0); end",
"def url; end",
"def url; end",
"def url; end",
"def url; end",
"def url; end",
"def url; end",
"def url; end",
"def url; end",
"def url; end",
"def url; end",
"def url; end",
"def url; end",
"def url; end",
"def url; end",
"def url; end",
"def url; end",
"def url; end",
"def url; end",
"def url; end",
"def url\n raise ::NotImplementedError\n end",
"def url_options; end",
"def original_url; end",
"def uri(the_url)\n case @url_method\n when :graphical\n self.class.uri_graphical(the_url)\n when :text\n case the_url.to_s\n when /^http/\n puts \"USER ACTION: please enter this url in a browser:\\n\"+the_url.to_s.red()+\"\\n\"\n else\n puts \"USER ACTION: open this:\\n\"+the_url.to_s.red()+\"\\n\"\n end\n else\n raise StandardError,\"unsupported url open method: #{@url_method}\"\n end\n end",
"def get(url); end",
"def consume_bad_url; end",
"def url\n end",
"def url\n raise NotImplementedError, 'this should be overridden by subclass'\n end",
"def url=(_); end",
"def url\n raise \"Implement in child class\"\n end",
"def options(url); end",
"def link() url; end",
"def link() url; end",
"def open_url\n opts = [config.oui_full_url]\n opts << { 'User-Agent' => config.user_agent } if config.user_agent\n open(*opts).read\n rescue OpenURI::HTTPError\n ''\n end",
"def url\n end",
"def url_accessibility(url)\n open(url).status.last == 'OK'\n rescue ::SocketError, Timeout::Error, Errno::ECONNREFUSED, OpenURI::HTTPError\n false\n end",
"def connection\n open(url)\n end",
"def get_url\n Nokogiri::HTML(open(@url))\n end",
"def file_uri_scheme=(_arg0); end",
"def url_exist?\n\t\tbegin\n\t\t\turi = URI.parse(valid_url?)\n\t\t\tresponse = Net::HTTP.get_response(uri)\n\t\trescue \n\t\t\terrors.add(:long_url,\"is invalid url\")\n\t\t\t# in AR, error is a class by itself already \n\t\t\t# go to static.rb to check the errors\n\t\tend\n\tend",
"def check_uri_build\n\n end",
"def file_uri_scheme; end",
"def status_open_url(url)\n open(url) do |f|\n return f.status\n end\nend",
"def file_url\n end",
"def error(url); end",
"def get_file(url); end",
"def fetchURL(url)\n open(url) {|response| response.read}\nend",
"def open_url\n u = URI.parse(@host_url)\n u.open { |file| @html_doc = Nokogiri::HTML(file) }\n\n rescue OpenURI::HTTPError => excp\n raise \"#{excp}, could not open #{@host_url} \"\n ensure\n @html_doc\n end",
"def stream_url\n\t\t\"http://www.google.com\"\n\tend",
"def resolved_uri; end",
"def openurl(url)\n data = nil\n uri = URI.parse(url)\n proxy_uri = self.getproxyuri\n http = (proxy_uri != nil) ? Net::HTTP.new(uri.host, uri.port, proxy_uri.host, proxy_uri.port) : Net::HTTP.new(uri.host, uri.port)\n if uri.scheme == \"https\"\n http.use_ssl = true\n http.verify_mode = OpenSSL::SSL::VERIFY_NONE\n end\n http.start {\n http.request_get(uri.path) {|res|\n data = res.body\n }\n }\n return data\n end",
"def new_http(uri); end",
"def raw_url(file_name)\n fail('Requires implementation')\n end",
"def open(uri_s, options = T.unsafe(nil), &error_block); end",
"def open_url(url)\n url = url.is_a?(NSURL) ? url : NSURL.URLWithString(url)\n request = NSURLRequest.requestWithURL(url, cachePolicy: NSURLRequestUseProtocolCachePolicy, timeoutInterval: 20)\n web.loadRequest request\n end",
"def http; end",
"def robots_error(url); end",
"def error!(url); end",
"def request_uri; end",
"def url(url)\n @url = url\n end",
"def uri=(_arg0); end",
"def uri=(_arg0); end",
"def uri=(_arg0); end",
"def uri=(_arg0); end",
"def proxy_uri; end",
"def proxy_uri; end",
"def initialize(url)\n @url = url\n end",
"def url\n super\n end",
"def server_url=(_arg0); end",
"def open_url(url)\n url = url.to_s\n puts \"Opening URL: #{url}\" if verbose\n\n if offline\n if config[:offline_url_mapping].key?(url)\n puts \"... using local file: #{config[:offline_url_mapping][url]}\" if verbose\n File.open(config[:offline_url_mapping][url])\n elsif (cache_path = cache_file_path(Digest::SHA1.hexdigest(url))) && File.exists?(cache_path)\n puts \"... using cache: #{cache_path}\" if verbose\n File.open(cache_path)\n else\n puts \"... skipping! No offline file or cache entry found\" if verbose\n end\n else\n puts \"... from remote location\" if verbose\n options = {\n redirect: false, # We're doing redirects manually below, since open-uri can't handle http -> https redirection\n read_timeout: 5\n }\n options[:ssl_verify_mode] = config[:ssl_verify_mode] if config.key?(:ssl_verify_mode)\n redirect_attempts = 5\n begin\n URI.open(url, options).tap do |f|\n cache_file(url, f) if cache\n f.rewind\n end\n rescue OpenURI::HTTPRedirect => e\n url = e.uri.to_s\n retry if (redirect_attempts -= 1) > 0\n raise e\n rescue OpenURI::HTTPError => e\n puts(\"Error for URL #{url}: #{e}\")\n end\n end\n end",
"def initialize\n raise NotImplementedError, 'need to implement #intialize and set @url'\n end",
"def url?\n !url.nil?\n end",
"def get_url\n @url\n end",
"def open(uri, options = T.unsafe(nil)); end",
"def request_url(requirelogin=false, requireuserinfo=false, expirationdays=30, notifyonupload=false)\n end",
"def robots_error!(url); end",
"def parse_link; end",
"def request_url(requirelogin=false, requireuserinfo=false, expirationdays=30, notifyonupload=false)\n end",
"def uri; end",
"def uri; end",
"def uri; end",
"def uri; end",
"def uri; end",
"def uri; end",
"def uri; end",
"def uri; end"
] |
[
"0.70443976",
"0.6838619",
"0.6798097",
"0.6742751",
"0.6653469",
"0.6653469",
"0.66116345",
"0.65972686",
"0.6596559",
"0.65895385",
"0.65895385",
"0.65895385",
"0.65895385",
"0.65895385",
"0.65895385",
"0.65895385",
"0.65362185",
"0.65362185",
"0.65362185",
"0.65362185",
"0.65362185",
"0.65362185",
"0.65362185",
"0.65362185",
"0.65362185",
"0.65362185",
"0.65362185",
"0.65362185",
"0.65362185",
"0.65362185",
"0.65362185",
"0.65362185",
"0.65362185",
"0.65362185",
"0.65362185",
"0.6519413",
"0.65193737",
"0.6511591",
"0.6502447",
"0.6490718",
"0.64525634",
"0.6444608",
"0.6439874",
"0.64234495",
"0.6378042",
"0.6373384",
"0.63658077",
"0.63658077",
"0.6353787",
"0.63507146",
"0.6350647",
"0.63399047",
"0.6313107",
"0.62364024",
"0.62346876",
"0.6221422",
"0.62172437",
"0.6211544",
"0.62079585",
"0.6179098",
"0.6175039",
"0.61621976",
"0.6161625",
"0.6149392",
"0.6134422",
"0.6128989",
"0.6119057",
"0.61170745",
"0.6101427",
"0.60949916",
"0.6089895",
"0.6088714",
"0.60781115",
"0.60735834",
"0.60491526",
"0.6046299",
"0.6046299",
"0.6046299",
"0.6046299",
"0.6038121",
"0.6038121",
"0.6019333",
"0.60023403",
"0.59964865",
"0.5994008",
"0.5987614",
"0.5973705",
"0.59697664",
"0.5966615",
"0.5966477",
"0.59646875",
"0.59621274",
"0.59593177",
"0.59510463",
"0.59510463",
"0.59510463",
"0.59510463",
"0.59510463",
"0.59510463",
"0.59510463",
"0.59510463"
] |
0.0
|
-1
|
before_action :check_review_options, only: [:show] GET /show_search
|
def show_search
@tour = Tour.new
@tour.tour_locations.build
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def show\n @search_mode = false #true if we came here from a tire search\n @default_tab = 1\n # DG 7/16/15 - commented this out since it no longer applies\n #if (!session[:diameter].blank?)\n # @default_tab = 2\n #else\n # @default_tab = 1\n #end\n\n # we don't currently have an in-house review system...\n @th_reviews_available = false\n\n if !params[:inline_search].blank?\n if params[:inline_search].downcase == \"true\"\n session[:inline_search] = \"true\"\n else\n session[:inline_search] = nil \n end\n end\n\n @diameters = TireSize.all_diameters\n @ratios = TireSize.all_ratios(session[:diameter])\n @wheeldiameters = TireSize.all_wheeldiameters(session[:ratio])\n\n Rack::MiniProfiler.step(\"load store and get params\") do\n @tire_store = TireStore.find(params[:id]) unless @tire_store\n\n @tire_store.tire_size_id = params[:tire_size_id]\n\n impressionist(@tire_store)\n # DG 7/16/15 - commented this out since it no longer applies\n #get_search_params_and_store_defaults() # if params[:auto_search] || params[:size_search] || params[:tire_size_id]\n @reservations = @tire_store.reservations.paginate(page: params[:page])\n if params[:ajax_mode]\n @ajax_mode = params[:ajax_mode].to_s.to_bool\n else\n @ajax_mode = true\n end\n end\n\n # find nearby stores with reviews.\n ### @nearby_google_stores = TireStore.near([@tire_store.latitude, @tire_store.longitude], 20).limit(2).where(\"(EXIST(google_properties, 'google_place_id')=TRUE)'))\")\n \n # DG 7/16/15 - For now, ignore any search params unless they came with a tire_search id.\n #We might use the id to provide a link back to the tire search results page.\n if !params[:tire_search].blank?\n #We've arrived on the tire store page from a tire search.\n #Rack::MiniProfiler.step(\"get tire search params\") do\n @search_mode = true\n @tire_search_id = params[:tire_search].to_i\n get_search_params_and_query_string()\n #end\n end\n\n Rack::MiniProfiler.step(\"load tirelistings from db\") do\n if params[:tab] && params[:tab].downcase == \"used\" && @tire_store.has_new_and_used?\n @tire_listings = @tire_store.used_tirelistings\n elsif params[:tab] && params[:tab].downcase == \"new\" && @tire_store.has_new_and_used?\n @tire_listings = @tire_store.new_tirelistings\n else\n @tire_listings = @tire_store.tire_listings\n end\n end\n\n # DG 7/16/15 - commented this out since it no longer applies\n #@refine_tire_search = TireSearch.new\n #@auto_manufacturers = AutoManufacturer.order(\"name\")\n #@models = AutoModel.where(:auto_manufacturer_id => session[:manufacturer_id])\n #@years = AutoYear.where(:auto_model_id => session[:auto_model_id])\n #@options = AutoOption.where(:auto_year_id => session[:auto_year_id]) \n\n Rack::MiniProfiler.step(\"mapping stuff a\") do\n @tire_stores, @quantities, @tire_manufacturers, @categories, @sellers, @pictures, @conditions, @wheelsizes = create_mappings_turbo(@tire_listings)\n #@tire_stores, @quantities, @tire_manufacturers, @categories, @sellers, @pictures, @conditions, @wheelsizes = create_mappings(@tire_listings)\n @can_edit = edit_access\n end\n Rack::MiniProfiler.step(\"mapping stuff b\") do\n if @tire_store.storefront_assets.count > 0\n @photos = @tire_store.storefront_assets.map{ |x| {:img => x.image.to_s, :caption => x.caption, :url => x.url}}\n end\n end\n Rack::MiniProfiler.step(\"mapping stuff c\") do\n if @tire_store.promotion_assets.count > 0\n @promotions = @tire_store.promotion_assets.map{ |x| {:img => x.image.to_s, :caption => x.caption, :url => x.url}}\n end\n end\n Rack::MiniProfiler.step(\"phone image\") do\n if @tire_store.private_seller?\n t = TextToGIF.new()\n @phone_image = t.create_phone_gif_for_tire_store(@tire_store)\n end\n end\n\n Rack::MiniProfiler.step(\"loading search data (not anymore) and branding\") do\n # DG 7/16/15 - commented this out since it no longer applies\n #load_default_search_data\n \n @branding = Branding.find_or_create_by_tire_store_id(@tire_store.id)\n\n if @branding.slogan.to_s.size > 0 && @branding.slogan_description.to_s.size > 0\n @slogan = @branding.slogan\n @slogan_description = @branding.slogan_description\n else\n @slogan = \"Featured Tires\"\n @slogan_description = \"Here are some of the tires we feature at #{@tire_store.name}.\"\n end\n end\n\n @contact_seller = ContactSeller.new(:id => 1,\n :email => signed_in? ? current_user.email : '',\n :sender_name => signed_in? ? current_user.first_name + ' ' + current_user.last_name : \"\",\n :tire_store_id => @tire_store.id) \n\n if @tire_store.private_seller?\n if signed_in? && (super_user? or current_user.account_id == @tire_store.account_id) then\n @submenu = Hash.new\n @submenu[:menu] = \"Edit\"\n @submenu[:items] = []\n @submenu[:items] << {href: \"/tire_listings/new?tire_store_id=\" + @tire_store.id.to_s, link: \"Create a new listing\"}\n @submenu[:items] << {href: \"/tire_stores/#{@tire_store.id}/edit\", link: \"Edit Seller Information\"}\n\n\n render :show_private\n return\n else\n redirect_to '/', :notice => \"You do not have permission to view this page.\" \n return\n end\n end\n\n if signed_in? && (super_user? or current_user.account_id == @tire_store.account_id) then\n @submenu = Hash.new\n @submenu[:menu] = \"Create Listings\"\n @submenu[:items] = []\n @submenu[:items] << {href: \"/tire_listings/new?tire_store_id=\" + @tire_store.id.to_s, link: \"Create a new listing\"}\n if !@tire_store.private_seller?\n @submenu[:items] << {href: \"/new_multiple?tire_store_id=\" + @tire_store.id.to_s, link: \"Create Multiple New Tire Listings\"}\n @submenu[:items] << {href: \"/generic_tire_listings/new?tire_store_id=\" + @tire_store.id.to_s, link: \"Create Bulk Used Tire Listings\"}\n @submenu[:items] << {href: edit_tire_store_path(@tire_store), link: \"Edit Store Information\"}\n end\n end\n \n # DG 8/6/15 - Setup appointment object for appointment modal\n @appointment = Appointment.new\n if !current_user.nil?\n @appointment.user_id = current_user.id if @appointment.user_id.blank?\n @appointment.buyer_name = current_user.name if @appointment.buyer_name.blank?\n @appointment.buyer_email = current_user.email if @appointment.buyer_email.blank?\n @appointment.buyer_phone = current_user.phone if @appointment.buyer_phone.nil?\n @appointment.buyer_address = \"\" # current_user.address if @appointment.buyer_address.nil?\n @appointment.buyer_city = \"\" # current_user.city if @appointment.buyer_city.nil?\n @appointment.buyer_state = \"\" # current_user.state if @appointment.buyer_state.nil?\n @appointment.buyer_zip = \"\" # current_user.zipcode if @appointment.buyer_zip.nil?\n end\n if !params[:auto_options_id].blank?\n option = AutoOption.find(params[:auto_options_id])\n if option\n @appointment_auto = option.auto_year.modelyear + ' ' + \n option.auto_year.auto_model.auto_manufacturer.name + ' ' +\n option.auto_year.auto_model.name + ' ' +\n option.name\n @appointment.auto_option_id = option.id\n @appointment.auto_year_id = option.auto_year_id\n @appointment.auto_model_id = option.auto_year.auto_model_id\n @appointment.auto_manufacturer_id = option.auto_year.auto_model.auto_manufacturer_id\n end\n end\n @appointment.request_hour_primary = '12:00' #default to 12 noon (if available)\n \n @appointments = {}\n @appt_counts = Appointment.store_appointments_by_day(@tire_store.id, Date.today, Date.today + 30.days)\n @primary_hours = @tire_store.hours_array_for_date(@appointment.request_date_primary)\n @secondary_hours = @tire_store.hours_array_for_date(@appointment.request_date_secondary)\n\n\n @ar_months = Date::MONTHNAMES.each_with_index.collect{|m, i| [m, i.to_s.rjust(2, '0')]}\n @ar_years = [*Date.today.year..Date.today.year + 8] \n \n \n\n # 3/6/13 ksi Clean this mess up...\n if in_storefront?\n if @tire_search && @branding.template_number == 4\n ourtires()\n respond_to do |format|\n format.html {redirect_to params.merge!(:action => :ourtires) }\n end\n else \n respond_to do |format|\n case @branding.template_number\n when 1\n format.html\n when 2\n format.html { render :show_storefront } #, :layout => 'storefront' }\n when 3\n format.html { render :show_storefront_slidetabs } #, :layout => 'storefront' }\n else\n format.html { render :show_storefront_traditional } #, :layout => 'storefront' }\n end\n end\n end\n else\n @turbo_mode = true # (params[:turbo].to_s > \"\")\n if @turbo_mode && !@ajax_mode\n Rack::MiniProfiler.step(\"JSONify\") do\n @json = @tire_listings.to_json(:root => true,\n #:except => [:price, :teaser, :description],\n :only => [:id, :is_new, :tire_manufacturer_id, :treadlife, :quantity],\n #:include => [:tire_manufacturer], \n :methods => [:cost_per_tire, :tire_category_id, :wheelsize]\n )\n end\n else\n @json = ''\n end\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @tire_store, :methods => [:sunday_open, :sunday_close, :monday_open, :monday_close, :tuesday_open, :tuesday_close, :wednesday_open, :wednesday_close, :thursday_open, :thursday_close, :friday_open, :friday_close, :saturday_open, :saturday_close, :closed_all_day_array, :open_all_day_array, :realtime_quote_distributors, :logo_url, :th_customer] }\n end\n end\n end",
"def show\n @fullwidth = true\n if @user.is_artist\n @artist = @user\n if user_signed_in?\n @review = Review.find_by(receiving_user_id: @artist.id, leaving_user_id: current_user.id)\n @review.nil? ? @review = Review.new : @review\n end\n @reviews = @artist.received_reviews.page(params[:page]).order('updated_at DESC').per(25)\n @artist.view_count.present? ? @artist.view_count += 1 : @artist.view_count = 0\n @artist.save\n else\n @reviews = @user.left_reviews.page(params[:page]).order('updated_at DESC').per(25)\n end\n if params[:review].present?\n top_review = Review.find(params[:review])\n if params[:response_link].present?\n render 'show', locals: {top_review: top_review, response_link: true}\n else\n render 'show', locals: {top_review: top_review}\n end\n else\n respond_to do |format|\n format.html { render 'show' }\n format.js { render action: 'paginate_reviews' }\n end\n end\n end",
"def show\n @search_object = Search.find(params[:id])\n @review_options=params\n @review_options[:gender]=@search_object[:gender] unless @search_object[:gender].blank?\n @review_options[:age]=@search_object[:age] unless @search_object[:age].blank?\n @review_options[:location]=@search_object[:location] unless @search_object[:location].blank?\n @review_options[:ethnicity]=@search_object[:ethnicity] unless @search_object[:ethnicity].blank?\n @review_options[:keyword]=@search_object[:keyword] unless @search_object[:keyword].blank?\n @review_options[:smoking_status]=@search_object[:smoking_status] unless @search_object[:smoking_status].blank?\n @review_options[:weight]=@search_object[:weight] unless @search_object[:weight].blank?\n\n if params[:drug_name] != 'all'\n @drug=Drug.find_by_brand_name(@search_object.drug_name)\n @review_options[:for_drug_id]=@drug.id\n @type=@drug.brand_name\n else\n @type=\"All Drugs\"\n end\n @total_results=@search_object.get_all_reviews(@review_options).count\n @reviews = @search_object.get_all_reviews(@review_options).order(\"created_at DESC\").page(params[:page]).per(5)\n\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @search }\n end\n end",
"def review\n end",
"def review\n filters = []\n OpportunityGenerator.filter_conditions.keys.each do |field|\n if params[field].present?\n filters << OpportunityGenerator.filter_conditions[field].gsub(\"?\", params[field])\n end\n end\n condition = filters.present? ? filters.join(\" AND \") : \"1=1\"\n @opportunity_generators = OpportunityGenerator.review_search(condition)\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @opportunity_generators }\n end\n end",
"def show\n @search = Review.where(book_id: @book.id).search(params[:q])\n @reviews = @search.result.paginate(:page => params[:page]).order(rating: :desc)\n @search.build_condition\n end",
"def show\n @review = find_review\n end",
"def show\n find_review\n find_show\n respond_to do |f|\n f.html {render :show}\n f.json {render json: @review}\n end\n end",
"def show\n @is_adv_search=params[:advance_search]\n end",
"def index\n if current_user.type == \"Student\"\n @reviews = Review.all.search(params[:search])\n \n else\n @reviews = current_user.reviews\n end\n @review = Review.new\n end",
"def show\n\t\t#show now run by find_book and before_action\n\t\t#keeps code dry\n\tend",
"def show_review_submissions\n @review_permitted_params = params.permit(Search::RESULTS_PARAMS)\n @review_start = @review_permitted_params[:start].to_i\n @review_window = window_size\n @review_items = review_items(@review_start, @review_window)\n @review_count = @review_items.count\n @review_current_page = @review_items.page\n render partial: \"collections/show_review_submissions_tab\"\n end",
"def index\n # this is our list page for our review\n # variable is @.\n # creating a filter variable for price\n @price = params[:price]\n #creating a filter variable for cuisine\n @cuisine = params[:cuisine]\n # adding location filter using geocoder\n @location = params[:location]\n\n #filtering properly by get all the reviews \"Review\" model from the database\n #creating new review variable as ruby list[]\n #@reviews = [\"The Smile\", \"Baby Bo's\", \"Chipotle\", \"nandos\"]\n @reviews = Review.all\n\n # filtering by price. this will toggle on/off depend when it has filter\n if @price.present?\n #take all of the review we have and replace the original review with filtered ones\n # find the value of the price in db that matches the param above\n @reviews = @reviews.where(price: @price)\n end\n\n #filter by cuisine\n if @cuisine.present?\n @reviews = @reviews.where(cuisine: @cuisine)\n end\n #search near the location\n if @location.present?\n # .near is what geo lcation given to us - see docs\n @reviews = @reviews.near(@location)\n end\n\n end",
"def index\n if has_search_parameters? || advanced_controller?\n super\n else\n respond_to do |format|\n format.html { store_preferred_view }\n end\n end\n end",
"def enforce_viewing_context_for_show_requests \n end",
"def on_search_page?\n params['controller'] == 'catalog' && params['action'] == 'index'\n end",
"def review; end",
"def show\n admin_last_url\n search_params = YAML::load(cookies[:product_search])\n search_params[:page] = params[:page]\n do_search search_params\n end",
"def index\n if related_product_label = params[:search] and related_product_label = params[:search][:related_product]\n @related_product = Product.first(:label => related_product_label)\n end\n @max_nb_reviews = params[:max_nb_reviews] || 100\n @date_oldest = if date_oldest = params[:date_oldest]\n Date.new(Integer(date_oldest[\"year\"]), Integer(date_oldest[\"month\"]), Integer(date_oldest[\"day\"]))\n else\n Date.today - 1000\n end\n\n @output_mode = params[:output_mode] || \"standard\"\n @source_categories = params[:source_categories]\n @source_categories ||= Review.categories.collect { |category_name, weight| category_name }\n @state_names = params[:state_names]\n @state_names ||= Review.list_states.collect(&:first)\n\n\n select_options = { :category => @source_categories,\n :state => @state_names,\n :limit => @max_nb_reviews,\n :written_at => { '$gt' => @date_oldest.to_time },\n :order => \"written_at DESC\" }\n select_options[\"product_ids\"] = @related_product.id if @related_product\n\n # puts \"selection options=#{select_options.inspect}\"\n @reviews = Review.all(select_options)\n @nb_reviews = @reviews.size\n\n if @output_mode == \"xml\"\n render(:xml => Rcollection.new(@current_user.rpx_username, \"xml output #{Time.now}\", @reviews) )\n else\n # index.html.erb\n end\n\n\n end",
"def show\n if user_signed_in? && current_user.admin?\n @new_review = NewReview.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @new_review }\n end\n else\n redirect_to root_path\n end\n end",
"def show\n #Suficiente con el set_training en el before_action\n end",
"def before_action \n end",
"def review_points\n unless params[:filter]\n week_start = now_unless_test.beginning_of_week\n params[:filter] = {\n start: week_start,\n :end => week_start.end_of_week,\n restaurants: [ 'all' ],\n riders: [ 'all' ],\n status: [ 'all' ]\n } \n end\n load_filter_wrapper\n load_filter_path_params\n load_shifts\n @base_path = '/shifts/review_points/'\n @filter_submit_path = @base_path\n load_table review_points: true\n render 'index'\n end",
"def my_reviews\n @owner = current_user\n @reviews = Review.find(:all, :conditions => [\"user_id = ?\", @owner.id])\n render :action => \"index\"\n end",
"def show\n @company = Company.find(params[:id])\n @reviewable = @company\n\n respond_to do |format|\n format.html {render :layout => 'with_search_header'}\n format.xml { render :xml => @company }\n end\n end",
"def show_action(url_params)\n # get list of recipes for specified chef.\n chef_id = url_params[:chef_id] || url_params[:id] \n if url_params.has_key?(:sort_by) && url_params[:sort_by] == \"popular\"\n # list of my recipes sorted by the most up votes first.\n # Chef.all.joins(recipes: :likes).where(\"likes.like = ?\", \"t\")\n # .select(\"chefs.*, count(likes.like) as num_votes\").group(\"chefs.id\")\n # .reorder(\"num_votes DESC\").order(\"chefs.name ASC\")\n # .paginate(page: url_params[:page], per_page: 3)\n Recipe.joins(:likes).where(\"recipes.chef_id = ? and likes.like = ?\", chef_id, \"t\")\n .select(\"recipes.*, count(likes.like) as num_votes\").group(\"recipes.id\")\n .reorder(\"num_votes DESC\").order(\"recipes.name ASC\")\n .paginate(page: url_params[:page], per_page: 2)\n else\n Recipe.where(\"recipes.chef_id = ?\", chef_id).reorder(\"recipes.name ASC\").paginate(page: url_params[:page], per_page: 2)\n end\n end",
"def show\n if user_signed_in?\n @trigger = TriggerAction.pending_action(\"review_required\", current_user, @assignment)\n @my_answers = Answer.where(user: current_user, assignment: @assignment, active: true)\n @my_reviews = Review.where(answer_id: @my_answers, active: true, assignment_id: @assignment.id)\n\n if current_user.get_and_store_experimental_condition!(@assignment.course) == \"batched_email\"\n @my_reviews = Review.where(answer_id: @my_answers, active: true, assignment_id: @assignment.id).where('created_at < ?', 1.day.ago)\n end\n @reviews_by_me = Review.where(active: true, assignment_id: @assignment.id).where(\"user_id = ? or copilot_email = ?\", current_user.id,current_user.email)\n @out_of_box_answers_with_count = Review.where(assignment_id: @assignment.id, out_of_box_answer: true).group(:answer_id).count\n\n unless @out_of_box_answers_with_count.blank?\n @out_of_box_answers = @out_of_box_answers_with_count.reject {|k,v| v < 2 }\n end\n if @out_of_box_answers.blank?\n @out_of_box_answers = {}\n end\n end\n @all_answers = @assignment.answers.reviewable.limit(10)\n @starred_answers = @assignment.answers.reviewable.where(starred: true)\n render layout: \"one_column\"\n end",
"def load_review\n @review = Spree::Review.find(params[:review_id])\n end",
"def show\n admin_store_url\n if params[:id] == \"search\"\n # This happens when javascript is disabled and the user clicks\n # on the will_paginate links.\n redirect_to :action => \"index\"\n else\n @product = Product.find(params[:id])\n respond_to do |format|\n format.html # show.html.erb\n format.xml { render :xml => @product }\n end\n end\n end",
"def show\n\n respond_to do |format|\n format.html #show\n format.json { render json: @review }\n end\n end",
"def set_admin_review\n @admin_review = Review.find(params[:id])\n end",
"def show_review_action(movie, status)\n\tcheck = user_already_reviewed(movie)\n\tif check == -1\n\t @show_review_action = true \n\telsif (check == 0 && status == true) || (check == 1 && status == false) || check == 2\n\t @show_review_action = false \n\telse\n\t @show_review_action = true\n\tend\n\nend",
"def show\n @review = Review.find(params[:id])\n\n end",
"def adv_search\r\n\r\n redirect_to :action => :search\r\n end",
"def review\n expose Challenge.review(@oauth_token,\n params[:technology] ||= nil, \n params[:platform] ||= nil, \n params[:category] ||= nil,\n params[:limit] ||= 25, \n params[:offset] ||= 0)\n end",
"def show\n @review = Review.find(params[:id])\n end",
"def show\n @review = Review.find(params[:id])\n end",
"def show\n @review = Review.find(params[:id])\n end",
"def show\n # return redirect_to finalize_search_path(Search.create_unfinalized(current_user, @shop)) unless can_show?\n @cart = Cart.new_with_bookings(current_search)\n I18n.locale = admin_preview_language_abbr if !pretending_to_be_customer?\n end",
"def show\n #this is going to be the individual review page using no. in db row using params array of :id. make sure we now create view page\n @review = Review.find(params[:id])\n end",
"def find_review\n @review = Review.find(params[:id])\n end",
"def show\n authorize! :show, @gf_travel_request\n if hf_transition_to_in_review?(@gf_travel_request, @user) # defined in StateEvents\n @gf_travel_request.approval_state.review!\n end\n end",
"def quick_search_show\n @exp_token = nil\n @quick_search_active = true\n @qs_exp_table = exp_build_table(@edit[:adv_search_applied][:exp], true)\n @edit[:qs_tokens] = create_tokens(@qs_exp_table, @edit[:adv_search_applied][:exp])\n\n render :update do |page|\n page << javascript_prologue\n page.replace(:user_input_filter, :partial => \"layouts/user_input_filter\")\n page << \"$('#advsearchModal').modal('hide');\"\n page << \"$('#quicksearchbox').modal('show');\"\n page << \"miqSparkle(false);\"\n end\n end",
"def index\n @admin_reviews = Review.all\n end",
"def review\n # unapproved post viewer\n # MAKE ADMIN ONLY\n @posts = Post.order(\"approved, created_at desc\")\n .paginate(per_page: 7, page: params[:page])\n end",
"def set_review\n @agency_review = Review.find(params[:review_id])\n end",
"def show\n authorize! :show, @travel_request\n if hf_transition_to_in_review?(@travel_request, @user) # defined in StateEvents\n @travel_request.approval_state.review!\n end\n end",
"def set_review\n @review = Review.find_by_id(params[:id])\n if @review == nil\n head :not_found\n end\n end",
"def show_search_bar_options\n zone = controller.class.name.downcase\n zone =~ /homepage/ || zone =~ /records/\n end",
"def show_search_box?\n (controller.controller_name.eql? \"catalog\")\n end",
"def review\n @t = T\n @t = @t.paginate :page => params[:page], :per_page => params[:per_page]\n @data = @t.meanings\n \n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @t }\n end\n end",
"def admin_review_params\n params.fetch(:review, {}).permit(whitelisted_params)\n end",
"def show\n @review = Review.new\n end",
"def show\n @review = Review.new\n end",
"def show\n @review = Review.new\n end",
"def show\n @review = Review.new\n end",
"def index\n @reviews = Review.where(shop_id: params[:shop_id])\n @shop = Shop.find(params[:shop_id])\n end",
"def filter\n setup_instance_variables\n render 'index'\n end",
"def set_review\n @review = Review.find(params[:id])\n end",
"def set_review\n @review = Review.find(params[:id])\n end",
"def set_review\n @review = Review.find(params[:id])\n end",
"def set_review\n @review = Review.find(params[:id])\n end",
"def set_review\n @review = Review.find(params[:id])\n end",
"def set_review\n @review = Review.find(params[:id])\n end",
"def set_review\n @review = Review.find(params[:id])\n end",
"def index\n logger.ap params\n if params.has_key? 'search_term' and params[:search_term] and !params[:search_term].blank?\n @reviews = Review.basic_search(params[:search_term]).page(params[:page]).per(params[:per_page])\n else\n @reviews = Review.page(params[:page]).per(params[:per_page])\n end\n logger.ap @reviews.length\n render json: @reviews\n end",
"def enforce_viewing_context_for_show_requests \n if params[:viewing_context] == \"browse\"\n session[:viewing_context] = params[:viewing_context]\n elsif session[:viewing_context] == \"edit\"\n if editor?\n redirect_to :action=>:edit\n else\n session[:viewing_context] = \"browse\"\n end\n end\n end",
"def show\n if search_params = params[:search] && params[:search][:query]\n @search = Search.new(query: search_params)\n @search.perform if @search.valid?\n else\n @search = Search.new\n end\nend",
"def view_reviews\n @submission = Submission.find(params[:id])\n @questions = @submission.assignment.questions.sort_by {|obj| obj.created_at }\n evaluation = @evaluations.where(:user_id => current_user.id)[0]\n @responses = @evaluations[0].responses.sort_by {|obj| obj.created_at }\n\n respond_to do |format|\n format.html { render view, :layout => 'no_sidebar' } # show.html.erb\n format.json { render json: @submission }\n end\n end",
"def redirect_old_review_filters\r\n old_params = [:min_rating, :N, :Ne, :Nf, :Nrc, :Ns, :page, :sort]\r\n if old_params.inject(false) { |memo, key| memo |= params.has_key?(key) }\r\n permanent_redirect_to :profile_mode => @profile_mode,\r\n :screen_name => params[:screen_name],\r\n :controller => 'profile_reviews', :action => 'index', :N => params[:N]\r\n return false\r\n end\r\n end",
"def show_only\n @options = {:show_all => true, :show_only => true, :action => 'show_answer', :validation => true}\n survey_id = params[:id].to_i\n @survey = Survey.and_questions.find(params[:id])\n @color = @survey.color\n @page_title = @survey.get_title\n # flash[:notice] = \"Denne side viser ikke et brugbart spørgeskema. Du har tilgang til besvarelser gennem journaler.\"\n render :template => 'surveys/show', :layout => \"layouts/jsurvey\"\n end",
"def find_review\n\t\t\t@review = Review.find(params[:id])\n\t\tend",
"def set_review\n\t\t@review = Review.find(params[:id])\n\tend",
"def show\n if query_exist\n else\n @cocktail = Cocktail.find(params[:id])\n @dose = Dose.new\n # added review to show page\n @review = Review.new\n end\n end",
"def index\n\t\tparams[:search] ? @reviews=Review.search(params[:search]) : @reviews= Review.where(\"user_id = ?\", current_user.id)\n\t\t@reviews = @reviews.order(:heading).paginate(page: params[:page], per_page: 18)\n\t\trespond_to do |format|\n\t\t\tformat.html # index.html.erb\n\t\t\tformat.json { render json: @reviews }\n\t\t\tformat.js\n\t\tend \n\tend",
"def start_new_search_session?\n params[:action] == 'show'\n end",
"def set_review\n @review=Review.find(params[:id])\n end",
"def show\n @review = Review.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @review }\n end\n end",
"def show\n @review = Review.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @review }\n end\n end",
"def show\n @review = Review.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @review }\n end\n end",
"def show\n @review = Review.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @review }\n end\n end",
"def show\n @review = Review.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @review }\n end\n end",
"def action_enable\n end",
"def review\n @page_title = 'Review Item: ' + @item.display_title\n @menu_title = 'Review Item'\n set_referer_as_redirect_back\n render 'review', layout: 'markup'\n end",
"def show\n authorize_action_for(@revise)\n end",
"def index\n @reviews = Review.all\n \n end",
"def set_review\n @review = @order.review\n\n if @review.nil?\n render body: nil, status: :not_found\n end\n end",
"def search_form\n #this is finsihed!!\n end",
"def search_action_url options = {}\n search_catalog_url(options.except(:controller, :action))\n end",
"def set_review\n @review = Review.find(params[:id])\n end",
"def available_actions\n super.merge({ \"search\" => \"You search the grimy pool\"\n\n })\n end",
"def analyse_search\n if params[:search].nil? && params['action'] == 'index'\n redirect_to root_path\n elsif params[:search].nil? && params['action'] == 'show'\n nil\n elsif params[:search][:age].present?\n @deputies = select_by_age(params[:search][:age])\n elsif params[:search][:grp].present?\n @deputies = select_by_group(params[:search][:grp])\n elsif params[:search][:ini].present?\n @deputies = select_by_initial(params[:search][:ini])\n elsif params[:search][:dep].present?\n @deputies = select_by_department(params[:search][:dep])\n elsif params[:search][:pro].present?\n @deputies = select_by_profession(params[:search][:pro])\n elsif params['action'] == 'index'\n redirect_to root_path\n end\n end",
"def find_review\n Review.find(params[\"id\"])\n end",
"def review_params\n\t\tparams.require(:review).permit(:heading, :review, :book_id, :user_id, :status, :cover_page)\n\tend",
"def show\n render json: @review\n end",
"def show\n render json: @review\n end",
"def show\n render json: @review\n end",
"def set_review_status # :norobots:\n pass_query_params\n id = params[:id].to_s\n desc = NameDescription.find(id)\n if is_reviewer?\n desc.update_review_status(params[:value])\n end\n redirect_to(:action => 'show_name', :id => desc.name_id,\n :params => query_params)\n end",
"def action\n if action_name == 'advanced_search'\n :post\n else\n :get\n end\n end",
"def show\n @review = reviewable.reviews.find(params[:id])\n\n respond_to do |format|\n format.html do\n if request.xhr?\n render \"show-xhr\", :layout => nil\n else\n render\n end\n end\n format.json { render json: @review }\n end\n end",
"def index\n\t\t@store_reviews = @store.store_reviews;\t\t\n\tend"
] |
[
"0.6621513",
"0.638247",
"0.63462746",
"0.63395184",
"0.6327548",
"0.63262004",
"0.630805",
"0.61938184",
"0.617426",
"0.6169452",
"0.6143272",
"0.6129062",
"0.6127854",
"0.6090672",
"0.6058821",
"0.60522443",
"0.6029703",
"0.6023195",
"0.60205024",
"0.60189414",
"0.60008156",
"0.59650815",
"0.59645844",
"0.59199214",
"0.59139585",
"0.58934546",
"0.5892432",
"0.588832",
"0.5880183",
"0.587397",
"0.58405817",
"0.5837448",
"0.5836223",
"0.583369",
"0.5811079",
"0.58095616",
"0.58095616",
"0.58095616",
"0.58046705",
"0.57916623",
"0.5784889",
"0.5765982",
"0.5764044",
"0.5755787",
"0.5741931",
"0.57204205",
"0.5718945",
"0.5714225",
"0.57097864",
"0.5708072",
"0.570236",
"0.5701209",
"0.5691366",
"0.5691366",
"0.5691366",
"0.5691366",
"0.56902623",
"0.5687849",
"0.5686064",
"0.5686064",
"0.5686064",
"0.5686064",
"0.5686064",
"0.5686064",
"0.5686064",
"0.5683725",
"0.56825465",
"0.5676109",
"0.566552",
"0.56653184",
"0.56629956",
"0.5659663",
"0.5659193",
"0.56582326",
"0.56435466",
"0.5633711",
"0.5632539",
"0.5632066",
"0.5632066",
"0.5632066",
"0.5632066",
"0.5632066",
"0.5630042",
"0.5620401",
"0.56160927",
"0.561597",
"0.56095225",
"0.5608177",
"0.560502",
"0.56044763",
"0.5603931",
"0.559693",
"0.5594342",
"0.5591461",
"0.5586683",
"0.5586683",
"0.5586683",
"0.5578975",
"0.55780864",
"0.5578019",
"0.55767363"
] |
0.0
|
-1
|
GET /tours GET /tours.json
|
def index
if params[:search]
@page_title = "List of Tours from filter"
# function located above...
search_results = search()
@tours = search_results[0]
@filter = search_results[1]
else
personalize = params[:my_tours]
bookmarked_tours = params[:bookmarked_tours]
waitlisted_tours = params[:waitlisted_tours]
if personalize
if %w[admin agent].include?(current_user.role)
@tours = current_user.tours
@page_title = "My Tours"
elsif current_user.role.eql? 'customer'
booked_user_tours = current_user.user_tours.select {|x| x.booked? || x.wait_listed?}
@tours = booked_user_tours.map {|ut| Tour.find(ut[:tour_id])}
@page_title = "My Tours"
end
elsif bookmarked_tours
bookmarked_user_tours = current_user.user_tours.select {|x| x.bookmarked?}
@tours = bookmarked_user_tours.map {|ut| Tour.find(ut[:tour_id])}
@page_title = "My Bookmarked Tours"
elsif waitlisted_tours
waitlisted_user_tours = current_user.user_tours.select {|x| x.wait_listed?}
@tours = waitlisted_user_tours.map {|ut| Tour.find(ut[:tour_id])}
@page_title = "My Waitlisted Tours"
else
@tours = Tour.all
@page_title = "All Tours"
end
end
end
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def index\n @tours = Tour.all\n end",
"def index\n @tours = Tour.all\n end",
"def index\n #@tours = Tour.all\n end",
"def show\n @tours = Tour.find(params[:id])\n end",
"def index\n @tours = Tour.all\n if params[:search]\n @search = TourSearch.new(params[:search])\n @tours = @search.scope\n @tours = {} if @tours.nil?\n end\n end",
"def my_tours\n @tours = Tour.where(user_id: current_user.id).order(created_at: :desc)\n end",
"def index\n return render json: {message: 'Public users cannot list tours'}, status: 403 if current_user.public?\n @tours = TourServices::TourFilterApi.new(user: current_user,\n status: params[:status],\n type: params[:type],\n vehicle_type: params[:vehicle_type],\n latitude: params[:latitude],\n longitude: params[:longitude],\n distance: params[:distance],\n page: params[:page],\n per: per).tours\n render json: @tours, status: 200, each_serializer: ::V1::TourSerializer, scope: {user: current_user}\n end",
"def index\n @tours = Tour.find(1,2,3,4)\n end",
"def index\n @tours = Tour.all\n @category = Category.new\n @place = Place.new\n \n end",
"def index\n @tourneys = Tourney.all\n end",
"def index\n if(params[:search])\n @title = \"Tours matching #{params[:search]}\"\n else\n @title = \"Search for tours\"\n end\n search_sort_and_paginate_tours(params)\n\n respond_to do |format|\n format.html # index.html.erb\n format.xml { render :xml => @tours }\n format.json { render :json => @tours }\n end\n end",
"def index\n @tours = Tour.all\n if params[:search]\n @tours = Tour.search(params[:search]).order(\"created_at DESC\")\n else\n @tours = Tour.all.order('created_at DESC')\n end\n end",
"def show\n render json: @tour, status: 200, serializer: ::V1::TourSerializer, scope: {user: current_user}\n end",
"def index\n @tours = Tour.all\n @cities = City.all\n @places = Place.all\n @user = User.first\n end",
"def index\n @tutor = Tutor.all\n\n render json: @tutor, status: :ok\n end",
"def index\n redirect_to root_path unless session_logged_in?\n @tours = Tour.accessible_by(current_ability, :update)\n end",
"def index\n @tour_requests = TourRequest.all\n end",
"def index\n @mod_tours = ModTour.all\n end",
"def index\n\n # Get tours\n @tours = Tour.get_tours(params)\n\n # Support filtering tours according to user desires\n # https://www.justinweiss.com/articles/search-and-filter-rails-models-without-bloating-your-controller/\n flash[:filters] = {}\n filtering_params(params).each do |key, value|\n # Filter the tours by this criteria IF a \"real\" value was provided for the filter\n if value.length.positive? && (!is_number?(value) || value.to_f.positive?)\n @tours = @tours.public_send(key, value) if value.present?\n end\n # Persist this filter information for one request\n # so that we can still show the user what they filtered by\n flash[:filters][key] = value\n end\n\n # Set page title\n set_page_title\n\n\n end",
"def index\n @tour_dates = TourDate.all\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @tour_dates }\n end\n end",
"def index\n @tour_details = TourDetail.all\n end",
"def tournaments\n get('sports/en/tournaments.xml')\n end",
"def index\n @tours = Tour.all.order(created_at: :desc)\n \n #Search functionality\n if params[:search]\n @tours = Tour.search(params[:search]).order(created_at: :desc)\n else\n @tours = Tour.all.order(created_at: :desc)\n end\n end",
"def index\n p params[:tour_id]\n if params[:tour_id].nil?\n @points = Point.all\n render json: @points, include: :tour, status: :ok\n else\n @points = Point.where(tour_id: params[:tour_id])\n render json: @points, status: :ok\n end\n end",
"def show\n @tour = Tour.find_by_url(params[:id])\n @stops = @tour.stops\n @ratings = @tour.ratings\n @title = @tour.name\n\n respond_to do |format|\n format.html # show.html.erb\n format.xml { render :xml => @tour }\n format.json { render :json => @tour }\n end\n end",
"def taken\n @travels = Travels::Travel.taken.where(performer: current_user)\n\n respond_to do |format|\n format.html { render \"travels/index\" }\n format.json { as_json @travels }\n end\n end",
"def index\n @tutorados = Tutorado.all\n\n render json: @tutorados, status: :ok\n end",
"def index\n @tourpubs = Tourpub.all\n end",
"def index\n @tourist_places = TouristPlace.all\n end",
"def index\n @locationtours = Locationtour.all\n end",
"def participatingtournaments_feed\n tournaments= getpartipatingtournaments\n tournamentslist = tournaments.map do |t|\n { :tournament_id=>t.id, :title=> t.name, :start => t.start_date }\n end\n\n render :json=> tournamentslist.to_json\n end",
"def show\n travels = Travel.find(params[:id])\n render json: travels, status: :ok\n end",
"def index\n @uchronists = Uchronist.all\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @uchronists }\n end\n end",
"def index\n @tour_interest_lists = TourInterestList.all\n end",
"def index\n @tourism_articles = @tourism_articles.published.recent.page(params[:page]).per(10)\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @tourism_articles }\n end\n end",
"def index\n @tournaments = Tournament.all\n end",
"def index\n @tournaments = Tournament.all\n end",
"def index\n @tournaments = Tournament.all\n end",
"def index\n @tournaments = Tournament.all\n end",
"def index\n @tournaments = Tournament.all\n end",
"def index\n render json: TeachingActivity.all\n end",
"def index\n @trajets = Trajet.all\n end",
"def index *_\n @trades = owned_trades_view\n\n render json: @trades\n end",
"def get_travellers\n if current_user != nil\n @trip = Trip.find(params[:id])\n render json: @trip.trip_json\n else\n render json: session[:trip].trip_json\n end\n\n end",
"def index\n @tenures = Tenure.all.order(\"updated_at DESC\").order(\"created_at DESC\")\n\n render json: @tenures\n end",
"def show\n @journey = Journey.find(params[:id])\n render json: @journey\n end",
"def index\n @tutorials = Tutorial.all\n\n respond_to do |format|\n format.html\n format.json do\n render json: @tutorials\n end\n end\nend",
"def index\n @tournament = Tournament.new\n @tournaments = Tournament.all\n respond_to do |format|\n format.html \n format.json { render json: @tournaments }\n format.js\n end\n end",
"def index\n @teaches = Teach.all\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @teaches }\n end\n end",
"def index\n @tour = Tour.find_by_id params[:tour_id]\n @high_lights = @tour.high_lights\n render layout: \"shijiebang\"\n end",
"def index\n @tutors = Tutor.all\n end",
"def index\n @treks = Trek.all\n @title = \"Trekking routes and destinations\"\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @treks }\n end\n end",
"def index\n @teaches = Teach.all\n\t\trespond_to do |format|\n\t\t\tformat.html { render :index }\n\t\t\tformat.json { render json: Oj.dump(@teaches) }\n\t\tend\n\n end",
"def show\n @tour_date = TourDate.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @tour_date }\n end\n end",
"def show\n @timeline = Timeline.find(params[:id])\n @orbituarysite = current_user.orbituarysites.new \n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @timeline }\n end\n end",
"def index\n @tutorials = Tutorial.all\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @tutorials }\n end\n end",
"def index\n @travels = Travel.all\n end",
"def index\n @travels = Travel.all\n end",
"def timeline\n Chirp.timeline(current_user)\n render json: @chirp\n end",
"def index\n puts \"<><><><><><><><><><><><><><><><><><><><><>\"\n p params\n puts \"<><><><><><><><><><><><><><><><><><><><><>\"\n # @journeys = current_user.journeys\n @journeys = User.find_by_id(params[:user_id]).journeys\n render json: @journeys.to_a\n end",
"def index\n @users = current_city.coaches.joins(:schedule).order(\"schedules.updated_at desc\").paginate(page: params[:page], per_page: 10)\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @users }\n end\n end",
"def show\n render json: @tutor, status: :ok\n end",
"def index\n @turmas = Turma.all\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @turmas }\n end\n end",
"def index\n past_histories = PastHistory.where(user_id: params[:user_id])\n render json: past_histories.as_json(include: :recipe, only: :created_at)\n end",
"def index\n @tutors = Tutor.find(:all)\n\n respond_to do |format|\n format.html # index.html.erb\n format.xml { render :xml => @tutors }\n end\n end",
"def search\n @search = TourSearch.new(params[:search])\n @tours = @search.scope\n end",
"def index\n @turns = Turn.all\n\t\trespond_with @turns\n end",
"def index\n render json: Story.all\n end",
"def index\n @clues = Clue.all\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @clues }\n end\n end",
"def index\n @to_dos = ToDo.all\n\n render json: @to_dos\n end",
"def index\n @goals = @todo.goals.all\n render json: @goals\n end",
"def created\n @travels = Travels::Travel.of(current_user)\n\n respond_to do |format|\n format.html { render \"travels/index\" }\n format.json { as_json @travels }\n end\n end",
"def show\n @venue = get_venue\n @tour = Tour.find(params[:id])\n @attraction_tours = @tour.attraction_tours.order(\"tour_order ASC\")\n @attractions = @tour.venue.attractions\n respond_to do |format|\n format.html # show.html.erb\n format.json { render :json => @tour.to_json( :include => {:attraction_tours => { :include => :attraction} , :media => {} } ) }\n end\n end",
"def index\n @swit_sours = SwitSour.all\n end",
"def index\n @stories = Story.order(\"created_at DESC\").all\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render :json => @stories }\n end\n end",
"def index\n @sponsors = @event.sponsors\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @sponsors }\n end\n end",
"def index_5\n @joueurs = @parties.map(&:joueurs).as_json\n\n respond_to do |format|\n format.html { render :html => @joueurs.to_s }\n format.json { render :json => @joueurs.to_json }\n format.xml { render :xml => @joueurs.to_xml }\n end\n end",
"def index\n trail_api_request\n @trails = Trail.order(:city).order(:name)\n end",
"def active\n @travels = Travels::Travel.actual\n\n respond_to do |format|\n format.html { render \"travels/index\" }\n format.json { as_json @travels }\n end\n end",
"def index\n @towns = Town.all\n end",
"def index\n @settlement_histories = Settlement::History.all\n\n respond_to do |format|\n format.html # index.html.erb\n format.json { render json: @settlement_histories }\n end\n end",
"def index\n trips = Trip.all\n render json: trips\n end",
"def index\n @turmas = current_colegio.turmas\n end",
"def index\n @schedule_tournaments = ScheduleTournament.all\n end",
"def index\n @stories = Story.all\n render json: @stories, root: false\n end",
"def index\n @turista = Turistum.all\n end",
"def index\n @observaciontutors = Observaciontutor.all\n end",
"def show\n if signed_in?\n @tourtable = Tourtable.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @tourtable }\n end\n end\nend",
"def index\n @lessons = Lesson.all\n\n render 'index.json'\n end",
"def show\n @tournament = Tournament.find(params[:id])\n\n respond_to do |format|\n format.html # show.html.erb\n format.json { render json: @tournament }\n end\n end",
"def index\n @trails = Trail.all\n end",
"def index\n @tutors = Tutor.all\n\n respond_to do |format|\n format.html # index.html.erb\n end\n end",
"def index\n @tournaments = Tournament.all.order(:id)\n end",
"def index\n @trips = Trip.all\n\n render json: @trips\n end",
"def index\n @trips = Trip.all\n\n render json: @trips\n end",
"def new\n @tln = Orbituarysite.find(params[:id]) \n @timeline = @tln.timelines.new\n @orbituarysite = current_user.orbituarysites.new \n\n respond_to do |format|\n format.html # new.html.erb\n format.json { render json: @timeline }\n end\n end",
"def index\n @leagues = League.all\n render json: @leagues, status: :ok\n end",
"def sport_tournaments(sport_id)\n get(\"sports/en/sports/sr:sport:#{sport_id}/tournaments.xml\")\n end",
"def index\n @trades = Trade\n .only(:created_at, :is_fair, :ash_pokemons, :brock_pokemons)\n respond_to do |format|\n format.json { render json: @trades }\n end\n end",
"def index\n @tracks = Track.all\n render json: {tracks: @tracks}\n end"
] |
[
"0.7661255",
"0.7661255",
"0.7493558",
"0.7369424",
"0.7208394",
"0.71341306",
"0.7035222",
"0.69646",
"0.6800399",
"0.67208934",
"0.66991645",
"0.6634698",
"0.6579901",
"0.6518906",
"0.65050703",
"0.6435392",
"0.640606",
"0.6384809",
"0.63697237",
"0.63596994",
"0.63517576",
"0.6314923",
"0.6314329",
"0.6314192",
"0.6291391",
"0.62338567",
"0.62315017",
"0.61881924",
"0.6174923",
"0.6124474",
"0.6105693",
"0.60983586",
"0.6090506",
"0.60756284",
"0.6065553",
"0.6016695",
"0.6016695",
"0.6016695",
"0.6016695",
"0.6016695",
"0.60144454",
"0.6006299",
"0.5999879",
"0.5984719",
"0.5971552",
"0.5959039",
"0.59550405",
"0.59509116",
"0.59498143",
"0.5944164",
"0.5936847",
"0.59218484",
"0.59148484",
"0.5902487",
"0.5898822",
"0.589797",
"0.5895201",
"0.5895201",
"0.5892585",
"0.58915335",
"0.5886458",
"0.5882378",
"0.5871116",
"0.5870872",
"0.58701",
"0.5847173",
"0.5845641",
"0.58440983",
"0.5842396",
"0.5835886",
"0.5833752",
"0.58311784",
"0.582407",
"0.5822388",
"0.58168215",
"0.58063686",
"0.5800164",
"0.57906824",
"0.5789871",
"0.5781955",
"0.57777685",
"0.57755244",
"0.5772493",
"0.57714194",
"0.57581365",
"0.5756116",
"0.57561034",
"0.5755643",
"0.5754203",
"0.57486284",
"0.5741371",
"0.5731198",
"0.572448",
"0.5722363",
"0.5722363",
"0.57183456",
"0.57174295",
"0.57091534",
"0.56983846",
"0.5692333"
] |
0.59414524
|
50
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.