repo_full_name
stringlengths
6
93
repo_url
stringlengths
25
112
repo_api_url
stringclasses
28 values
owner
stringclasses
28 values
repo_name
stringclasses
28 values
description
stringclasses
28 values
stars
int64
617
98.8k
forks
int64
31
355
watchers
int64
990
999
license
stringclasses
2 values
default_branch
stringclasses
2 values
repo_created_at
timestamp[s]date
2012-07-24 23:12:50
2025-06-16 08:07:28
repo_updated_at
timestamp[s]date
2026-02-23 15:23:15
2026-05-03 18:52:12
repo_topics
listlengths
0
13
repo_languages
unknown
is_fork
bool
1 class
open_issues
int64
3
104
file_path
stringlengths
3
208
file_name
stringclasses
509 values
file_extension
stringclasses
1 value
file_size_bytes
int64
101
84k
file_url
stringclasses
627 values
file_raw_url
stringclasses
627 values
file_sha
stringclasses
624 values
language
stringclasses
8 values
parsed_at
stringdate
2026-05-04 01:12:36
2026-05-04 19:41:55
text
stringlengths
100
102k
LeCAR-Lab/ASAP
https://github.com/LeCAR-Lab/ASAP
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
humanoidverse/simulator/genesis/genesis_mjdebug.py
null
null
null
null
null
null
Python
2026-05-04T01:39:52.432810
import sys import os from loguru import logger # from isaacgym import gymtorch, gymapi, gymutil import torch import genesis as gs from genesis.engine.solvers.rigid.rigid_solver_decomp import RigidSolver from humanoidverse.simulator.genesis.tmp_gs_utils import * from humanoidverse.simulator.genesis.genesis_viewer import...
LeCAR-Lab/ASAP
https://github.com/LeCAR-Lab/ASAP
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
humanoidverse/agents/callbacks/base_callback.py
null
null
null
null
null
null
Python
2026-05-04T01:39:53.668465
import torch from torch.nn import Module class RL_EvalCallback(Module): def __init__(self, config, training_loop): super().__init__() self.config = config self.training_loop = training_loop self.device = self.training_loop.device def on_pre_evaluate_policy(self): pass ...
LeCAR-Lab/ASAP
https://github.com/LeCAR-Lab/ASAP
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
humanoidverse/agents/dagger/dagger.py
null
null
null
null
null
null
Python
2026-05-04T01:39:53.698042
import torch import torch.nn as nn import torch.optim as optim from humanoidverse.agents.modules.ppo_modules import PPOActorFixSigma from humanoidverse.agents.modules.models import BaseModels from humanoidverse.agents.modules.data_utils import RolloutStorage from humanoidverse.envs.base_task.base_task import BaseEnv f...
LeCAR-Lab/ASAP
https://github.com/LeCAR-Lab/ASAP
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
humanoidverse/agents/ppo/ppo.py
null
null
null
null
null
null
Python
2026-05-04T01:39:54.695508
import torch import torch.nn as nn import torch.optim as optim from humanoidverse.agents.modules.ppo_modules import PPOActor, PPOCritic from humanoidverse.agents.modules.data_utils import RolloutStorage from humanoidverse.envs.base_task.base_task import BaseTask from humanoidverse.agents.base_algo.base_algo import Bas...
LeCAR-Lab/ASAP
https://github.com/LeCAR-Lab/ASAP
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
humanoidverse/agents/mppi/mppi.py
null
null
null
null
null
null
Python
2026-05-04T01:39:54.948587
from typing import Dict, Any, Optional, Tuple from multiprocessing import Process, Manager,shared_memory, Event, set_start_method, Queue import struct import numpy as np import pickle as pkl import torch import torch.nn as nn from torch.utils.tensorboard import SummaryWriter as TensorboardSummaryWriter import wandb fr...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/cp.py
null
null
null
null
null
null
Python
2026-05-04T01:39:57.359066
# -*- coding: utf-8 -*- """Copy a file or directory. Multiple source files may be specified if the destination is an existing directory. """ from __future__ import print_function import argparse import os import shutil import sys def pprint(path): if path.startswith(os.environ["HOME"]): return "~" + pat...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/cowsay.py
null
null
null
null
null
null
Python
2026-05-04T01:39:57.370378
# -*- coding: utf-8 -*- from __future__ import print_function # by Siddharth Duahantha # 28 July 2017 import sys import argparse COW = r""" \ ^__^ \ (oo)\_______ (__)\ )\/\\ ||----w | || || """ def get_cow(text): """create a strin...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/clear.py
null
null
null
null
null
null
Python
2026-05-04T01:39:57.372390
# -*- coding: utf-8 -*- """Clear the stash console output window""" from __future__ import print_function print("\u009bc", end="")
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/crypt.py
null
null
null
null
null
null
Python
2026-05-04T01:39:57.373983
# -*- coding: utf-8 -*- """ File encryption for stash Uses AES in CBC mode. usage: crypt.py [-h] [-k KEY] [-d] infile [outfile] positional arguments: infile File to encrypt/decrypt. outfile Output file. optional arguments: -h, --help show this help message and exit -k KEY, --ke...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/cd.py
null
null
null
null
null
null
Python
2026-05-04T01:39:57.375040
#!/usr/bin/env python # -*- coding: utf-8 -*- """Change the current working directory.""" from __future__ import print_function import argparse import os import sys def main(args): p = argparse.ArgumentParser(description=__doc__) p.add_argument( "dir", action="store", nargs="?", ...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/cat.py
null
null
null
null
null
null
Python
2026-05-04T01:39:57.383773
#!/usr/bin/env python # -*- coding: utf-8 -*- """Print the contents of the given files.""" from __future__ import print_function import argparse import string import sys import fileinput def filter_non_printable(s): return "".join( [c if c.isalnum() or c.isspace() or c in string.punctuation else " " for...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/alias.py
null
null
null
null
null
null
Python
2026-05-04T01:39:57.387532
# -*- coding: utf-8 -*- # Example of accessing the shell object from script # This ability completely removes the need of plugins """List or define shell aliases.""" from __future__ import print_function import sys import argparse def main(args): ap = argparse.ArgumentParser() ap.add_argument("expr", nargs=...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/curl.py
null
null
null
null
null
null
Python
2026-05-04T01:39:57.388921
# coding=utf-8 """Transfer a URL""" from __future__ import print_function import sys import argparse import requests from six.moves.urllib.parse import urlparse try: import clipboard except ImportError: clipboard = None def main(args): ap = argparse.ArgumentParser() ap.add_argument("url", nargs="?"...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/cut.py
null
null
null
null
null
null
Python
2026-05-04T01:39:57.412792
# -*- coding: utf-8 -*- """Print selected parts of lines from each FILE to standard output.""" from __future__ import print_function import sys import argparse _stash = globals()["_stash"] def construct_indices_from_list_spec(list_spec): # Note unlike python, cut's indices start from 1 indices = [] for ...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/diff.py
null
null
null
null
null
null
Python
2026-05-04T01:39:57.962844
#!/usr/bin/env python import sys import os import re import json import argparse import time import pytz import console from datetime import datetime, timedelta from difflib import unified_diff, ndiff # _____________________________________________________ def argue(): parser = argparse.ArgumentParser() par...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/du.py
null
null
null
null
null
null
Python
2026-05-04T01:39:57.991049
# -*- coding: utf-8 -*- """Summarize disk usage of the set of FILEs, recursively for directories.""" from __future__ import print_function import os import sys from argparse import ArgumentParser from fnmatch import fnmatch def is_excluded(path, pattern): if pattern: while path != "": prefix,...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/dropbox_setup.py
null
null
null
null
null
null
Python
2026-05-04T01:39:58.011627
#! python2 # -*- coding: utf-8 -*- # StaSh utility """manage your dropbox configuration.""" import cmd import keychain import sys from stashutils import dbutils _stash = globals()["_stash"] class DropboxSetupCmd(cmd.Cmd): """The command loop for managing the dropbox""" intro = _stash.text_color( "...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/echo.py
null
null
null
null
null
null
Python
2026-05-04T01:39:58.013254
#!/usr/bin/env python # -*- coding: utf-8 -*- """Print all arguments to stdout, separated by spaces.""" from __future__ import print_function import sys def main(args): # Not using argparse here, because echo should echo anything that is not a # valid and usable flag. end = "\n" escapes = False # N...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/find.py
null
null
null
null
null
null
Python
2026-05-04T01:39:58.015358
# -*- coding: utf-8 -*- """Find files in specified paths""" from __future__ import print_function import os import sys import argparse import time import fnmatch from functools import partial class FilePredicate(object): def __init__(self): self.funclist = [] def add_filter(self, func): sel...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/fg.py
null
null
null
null
null
null
Python
2026-05-04T01:39:58.018519
# -*- coding: utf-8 -*- """ Bring a background job to foreground. """ from __future__ import print_function import sys import argparse import threading def main(args): ap = argparse.ArgumentParser() ap.add_argument( "job_id", nargs="?", type=int, help="ID of a running background job" ) ns = a...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/exit.py
null
null
null
null
null
null
Python
2026-05-04T01:39:58.019910
#!/usr/bin/env python # -*- coding: utf-8 -*- """Exit the current subshell, optionally with a specific status. If no status is given, the default of 0 is used, indicating successful execution with no errors. """ import argparse import os import sys def main(args): p = argparse.ArgumentParser(description=__doc__)...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/edit.py
null
null
null
null
null
null
Python
2026-05-04T01:39:58.030205
# -*- coding: utf-8 -*- """ Used to create/open and edit files. [-t --temp] - Opens the file as a temporary file. Allowing editing and renaming. Previous script in the pythonista editor will be restored unless a new tab is edited. [-o --old_tab] - Open file in an old editor tab (default is new tab, which is possible fr...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/easy_config.py
null
null
null
null
null
null
Python
2026-05-04T01:39:58.060875
# -*- coding: utf-8 -*- """a graphical config manager for StaSh""" import ast import os import threading from six import string_types import console import ui import dialogs import pythonista_add_action as paa from stash.system.shcommon import _STASH_CONFIG_FILES, _STASH_HISTORY_FILE _stash = globals()["_stash"] ...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/ftpserver.py
null
null
null
null
null
null
Python
2026-05-04T01:39:58.095170
# -*- coding: utf-8 -*- """Simple FTP Server""" from __future__ import print_function import argparse import os import sys import threading import time import logging _stash = globals()["_stash"] try: import pyftpdlib except ImportError: print("Installing pyftpdlib...") _stash("pip install pyftpdlib") ...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/gci.py
null
null
null
null
null
null
Python
2026-05-04T01:39:58.567390
# coding: utf-8 """Interface to pythons built-in garbage collector Warning: this command may crash StaSh! Only use it if you know what you are doing!""" import gc import argparse import sys _stash = globals()["_stash"] def main(): parser = argparse.ArgumentParser( description="access to pythons built-i...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/grep.py
null
null
null
null
null
null
Python
2026-05-04T01:39:58.608431
# -*- coding: utf-8 -*- """Search a regular expression pattern in one or more files""" from __future__ import print_function import argparse import collections import fileinput import os import re import sys def main(args): global _stash ap = argparse.ArgumentParser() ap.add_argument("pattern", help="th...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/gh.py
null
null
null
null
null
null
Python
2026-05-04T01:39:58.609418
# coding: utf-8 """ Usage: gh <command> [<args>...] gh <command> (-h|--help) supported commands are: gh fork <repo> forks user/repo gh create <repo> creates a new repo gh pull <repo> <base> <head> create a pull request gh list_keys list user keys gh create_key <title> ...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/git.py
null
null
null
null
null
null
Python
2026-05-04T01:39:58.632473
# -*- coding: utf-8 -*- """ Distributed version control system Commands: init: git init <directory> - initialize a new Git repository add: git add <file1> .. [file2] .. - stage one or more files rm: git rm <file1> .. [file2] .. - unstage one or more files commit: git commit <message> <name> <email> - ...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/httpserver.py
null
null
null
null
null
null
Python
2026-05-04T01:39:58.634753
#!/usr/bin/env python # -*- coding: utf-8 -*- """Simple HTTP Server With Upload (https://gist.github.com/UniIsland/3346170) (StaSh) Recommend to run this command as background job, i.e. "httpserver [port] &" This module builds on BaseHTTPServer by implementing the standard GET and HEAD requests in a fairly straightfo...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/head.py
null
null
null
null
null
null
Python
2026-05-04T01:39:58.655067
# -*- coding: utf-8 -*- """Print the first 10 lines of the given files.""" from __future__ import print_function import argparse import string import sys import fileinput def filter_non_printable(s): return "".join( [c if c.isalnum() or c.isspace() or c in string.punctuation else " " for c in s] ) ...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/jobs.py
null
null
null
null
null
null
Python
2026-05-04T01:39:58.656607
# -*- coding: utf-8 -*- """ List all jobs that are currently running. """ from __future__ import print_function import sys import argparse import threading def main(args): ap = argparse.ArgumentParser() ap.parse_args(args) current_worker = threading.currentThread() _stash = globals()["_stash"] ...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/latte.py
null
null
null
null
null
null
Python
2026-05-04T01:39:58.657933
# -*- coding: utf-8 -*- # A package manager meant for Pythonista, built on StaSh. from __future__ import print_function import argparse import sys from os import getcwd, mkdir, remove, rename from shutil import rmtree import requests cwd = getcwd() documentsIndex = cwd.index("Documents") documentsIndex += len("Docum...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/ln.py
null
null
null
null
null
null
Python
2026-05-04T01:39:58.659399
#!/usr/bin/env python import sys import os import re import json import argparse import time import pytz import console from datetime import datetime, timedelta from difflib import unified_diff, ndiff def argue(): parser = argparse.ArgumentParser() parser.add_argument("-v", "--verbose", action="store_true") ...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/kill.py
null
null
null
null
null
null
Python
2026-05-04T01:39:58.678972
# -*- coding: utf-8 -*- """ Terminate a running job. """ from __future__ import print_function import sys import argparse import time def main(args): ap = argparse.ArgumentParser() ap.add_argument("job_ids", nargs="+", type=int, help="ID of a running job") ns = ap.parse_args(args) _stash = globals(...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/ls.py
null
null
null
null
null
null
Python
2026-05-04T01:39:59.704194
#!/usr/bin/env python # -*- coding: utf-8 -*- """List information about files (the current directory by default)""" from __future__ import print_function import os import sys import time import tarfile import zipfile import imghdr from argparse import ArgumentParser from stashutils.mount_ctrl import get_manager def...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/man.py
null
null
null
null
null
null
Python
2026-05-04T01:39:59.708506
# -*- coding: utf-8 -*- """Display help for a command in $STASH_ROOT/bin/ or a topic, or list all commands if no name is given.""" from __future__ import print_function import argparse import ast import os import sys from stash.system.shcommon import _STASH_EXTENSION_BIN_PATH, _STASH_EXTENSION_MAN_PATH try: raw...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/logout.py
null
null
null
null
null
null
Python
2026-05-04T01:39:59.727515
""" Quit a shell. """ import argparse _stash = globals()["_stash"] def logout(n): """ Quit StaSh :param n: exitcode for the shell (not implemented) :type n: int """ import threading t = threading.Thread(target=_stash.close, name="close thread") t.daemon = True t.start() if __...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/mc.py
null
null
null
null
null
null
Python
2026-05-04T01:39:59.729372
# coding: utf-8 """easily work with multiple filesystems (e.g. local and FTP) synchronously""" # the name refers to midnight-commander, but this will probably # never be a true counterpart import os import shutil import cmd import sys import tempfile import shlex from stashutils.fsi.errors import OperationFailure, Is...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/mail.py
null
null
null
null
null
null
Python
2026-05-04T01:39:59.759286
# -*- coding: utf-8 -*- """ Send messages/files from stash. Calling mail without params will prompt for input. Passing '-' in the message param will read from stdin pipe. usage: mail.py [-h] [-s SUBJECT] [-v] [-f FILE] [-e] [sendto] [message] positional arguments: sendto Send email to. message ...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/mount.py
null
null
null
null
null
null
Python
2026-05-04T01:40:00.323905
# -*- coding: utf-8 -*- """mount a filesystem.""" from __future__ import print_function import argparse import sys from six import string_types from six.moves import input from stashutils import mount_ctrl, mount_manager from stashutils.fsi.interfaces import FILESYSTEM_TYPES _stash = globals()["_stash"] def list...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/more.py
null
null
null
null
null
null
Python
2026-05-04T01:40:00.525467
#! python2 # -*- coding: utf-8 -*- # StaSh utility - Dutcho, 16-17 Apr 2017 """Display output one screen page at a time (optionally with numbered lines)""" from __future__ import print_function import argparse import console import fileinput import sys def msi(chars): """Set terminal screen by ANSI code "MSI", ...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/monkeylord.py
null
null
null
null
null
null
Python
2026-05-04T01:40:00.526717
# -*- coding: utf-8 -*- """easiliy manage monkey-patches. See 'man monkeypatching' for more help.""" from __future__ import print_function import argparse import sys import json from mlpatches import base _stash = globals()["_stash"] from mlpatches import patches def patch_is_compatible(patch): """Return True ...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/pip.py
null
null
null
null
null
null
Python
2026-05-04T01:40:00.966109
# -*- coding: utf-8 -*- """ Install and manage python packages usage: pip.py [-h] [--verbose] sub-command ... optional arguments: -h, --help show this help message and exit --verbose be more chatty List of sub-commands: sub-command "pip sub-command -h" for more help on a sub-command list ...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/printenv.py
null
null
null
null
null
null
Python
2026-05-04T01:40:01.111881
#!/usr/bin/env python # -*- coding: utf-8 -*- """List current environment variables and values.""" from __future__ import division, print_function, unicode_literals import argparse import os import sys def main(args): p = argparse.ArgumentParser(description=__doc__) p.add_argument( "variables", acti...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/printhex.py
null
null
null
null
null
null
Python
2026-05-04T01:40:01.141583
#!/usr/bin/env python # -*- coding: utf-8 -*- """Print the given files' content and hexadecimal byte values.""" from __future__ import print_function import argparse import sys INVISIBLE = range(0x20) + [0x81, 0x8D, 0x8F, 0x90, 0x9D] def main(args): p = argparse.ArgumentParser(description=__doc__) p.add_ar...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/pwd.py
null
null
null
null
null
null
Python
2026-05-04T01:40:01.693774
#!/usr/bin/env python # -*- coding: utf-8 -*- """Print the current working directory.""" from __future__ import print_function import argparse import os import sys _stash = globals()["_stash"] collapseuser = _stash.libcore.collapseuser def main(args): p = argparse.ArgumentParser(description=__doc__) p.add_...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/python.py
null
null
null
null
null
null
Python
2026-05-04T01:40:03.174736
# -*- coding: utf-8 -*- """ Simulates a console call to python [-m module][-c cmd] [file] [args] Used for running standard library python modules such as: SimpleHTTPServer, unittest and .py files. Can also be used to run a script in the background, such as a server, with the bash character & at the end. usage: py...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/mkdir.py
null
null
null
null
null
null
Python
2026-05-04T01:40:04.321528
#!/usr/bin/env python # -*- coding: utf-8 -*- """Create a new directory. The parent directory must already exist, unless -p is specified. """ from __future__ import print_function import argparse import os import sys def main(args): p = argparse.ArgumentParser(description=__doc__) p.add_argument( "-...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/md5sum.py
null
null
null
null
null
null
Python
2026-05-04T01:40:04.743963
# -*- coding: utf-8 -*- """ Get md5 hash of a file or string. usage: md5sum.py [-h] [-c] [file [file ...]] positional arguments: file String or file to hash. optional arguments: -h, --help show this help message and exit -c, --check Check a file with md5 hashes and file names for a match. ...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/pbcopy.py
null
null
null
null
null
null
Python
2026-05-04T01:40:05.344506
# -*- coding: utf-8 -*- """Copy one or more files to the system clipboard""" from __future__ import print_function import argparse import fileinput import os import sys _stash = globals()["_stash"] def main(args): """ The main function. """ ap = argparse.ArgumentParser() ap.add_argument("file"...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/openin.py
null
null
null
null
null
null
Python
2026-05-04T01:40:05.345014
#!/usr/bin/env python # -*- coding: utf-8 -*- """Open file in an external app.""" from __future__ import print_function import argparse import sys _stash = globals()["_stash"] class ConsoleOpenin(object): def __init__(self, args): p = argparse.ArgumentParser(description=__doc__) p.add_argument...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/mv.py
null
null
null
null
null
null
Python
2026-05-04T01:40:05.374723
#!/usr/bin/env python # -*- coding: utf-8 -*- """Move (rename) a file or directory to a new name, or into a new directory. Multiple source files may be specified if the destination is an existing directory. """ from __future__ import print_function import argparse import os import shutil import sys def main(args): ...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/ping.py
null
null
null
null
null
null
Python
2026-05-04T01:40:05.406402
#!/usr/bin/env python2 # -*- coding: utf-8 -*- """ Usage: ping [-c <count>] [-i <interval>] [-W <timeout>] <destination> Options: -c <count>, --count=<count> [default: 5] -i <interval>, --interval=<interval> [default: 1.0] Wait interval seconds between sending each packet. The default is to wait ...
ywangd/stash
https://github.com/ywangd/stash
null
null
null
null
1,999
null
null
mit
null
null
null
null
null
null
null
bin/pbpaste.py
null
null
null
null
null
null
Python
2026-05-04T01:40:05.478679
# -*- coding: utf-8 -*- """Writes the contents of the system clipboard to a file.""" from __future__ import print_function import argparse import os import sys import io import six _stash = globals()["_stash"] def main(args): ap = argparse.ArgumentParser() ap.add_argument("file", nargs="?", help="the fil...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
crawler/spiders/__init__.py
null
null
null
null
null
null
Python
2026-05-04T01:40:07.504274
# This package will contain the spiders of your Scrapy project # # Please refer to the documentation for information on how to create and manage # your spiders.
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
crawler/pipelines.py
null
null
null
null
null
null
Python
2026-05-04T01:40:07.505725
# -*- coding: utf-8 -*- # Define your item pipelines here # # Don't forget to add your pipeline to the ITEM_PIPELINES setting # See: http://doc.scrapy.org/en/latest/topics/item-pipeline.html class CrawlerPipeline(object): def process_item(self, item, spider): return item
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
crawler/items.py
null
null
null
null
null
null
Python
2026-05-04T01:40:07.507276
# -*- coding: utf-8 -*- # Define here the models for your scraped items # # See documentation in: # http://doc.scrapy.org/en/latest/topics/items.html import scrapy class CrawlerItem(scrapy.Item): # define the fields for your item here like: # name = scrapy.Field() pass
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
crawler/settings.py
null
null
null
null
null
null
Python
2026-05-04T01:40:07.508345
# -*- coding: utf-8 -*- # Scrapy settings for crawler project # # For simplicity, this file contains only settings considered important or # commonly used. You can find more settings consulting the documentation: # # http://doc.scrapy.org/en/latest/topics/settings.html # http://scrapy.readthedocs.org/en/latest...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
crawler/spiders/proxy/basespider.py
null
null
null
null
null
null
Python
2026-05-04T01:40:07.512625
# coding=utf-8 import sys import config import utils import datetime from scrapy.spiders import Spider from scrapy.http import Request from sql import SqlManager class BaseSpider(Spider): name = 'basespider' def __init__(self, *a, **kw): super(BaseSpider, self).__init__(*a, **kw) self.urls...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
config.py
null
null
null
null
null
null
Python
2026-05-04T01:40:07.513139
# coding=utf-8 DB_config = { # 'db_type': 'mongodb', 'db_type': 'mysql', 'mysql': { 'host': 'localhost', 'port': 3306, 'user': 'root', 'password': '123456', 'charset': 'utf8', }, 'redis': { 'host': 'localhost', 'port': 6379, 'password...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
crawler/spiders/proxy/data5u.py
null
null
null
null
null
null
Python
2026-05-04T01:40:07.597417
# -*- coding: utf-8 -*- from scrapy import Selector from .basespider import BaseSpider from proxy import Proxy class Data5uSpider(BaseSpider): name = 'data5u' def __init__(self, *a, **kw): # 在类的继承中,如果重定义某个方法,该方法会覆盖父类的同名方法 # 但有时,我们希望能同时实现父类的功能,这时,我们就需要调用父类的方法了,可通过使用 super 来实现,比如: supe...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
crawler/spiders/proxy/proxydb.py
null
null
null
null
null
null
Python
2026-05-04T01:40:08.492211
# coding=utf-8 from proxy import Proxy from .basespider import BaseSpider from scrapy.selector import Selector import re from base64 import b64decode class ProxyDBSpider(BaseSpider): name = 'proxydb' def __init__(self, *a, **kwargs): super(ProxyDBSpider, self).__init__(*a, **kwargs) self.urls...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
crawler/spiders/proxy/kuaidaili.py
null
null
null
null
null
null
Python
2026-05-04T01:40:08.494398
#-*- coding: utf-8 -*- import re from proxy import Proxy from .basespider import BaseSpider class KuaiDaiLiSpider(BaseSpider): name = 'kuaidaili' def __init__(self, *a, **kwargs): super(KuaiDaiLiSpider, self).__init__(*a, **kwargs) self.urls = ['https://www.kuaidaili.com/free/inha/%s/' % i...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
crawler/spiders/proxy/sixsixip.py
null
null
null
null
null
null
Python
2026-05-04T01:40:08.495590
# coding=utf-8 import re from proxy import Proxy from .basespider import BaseSpider class SixSixIpSpider(BaseSpider): name = 'sixsixip' def __init__(self, *a, **kwargs): super(SixSixIpSpider, self).__init__(*a, **kwargs) self.urls = ['http://m.66ip.cn/%s.html' % n for n in range(1, 10)] ...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
crawler/spiders/proxy/gatherproxy.py
null
null
null
null
null
null
Python
2026-05-04T01:40:08.497185
# coding=utf-8 import json import random import re import requests from proxy import Proxy from .basespider import BaseSpider class GatherproxySpider(BaseSpider): name = 'gatherproxy' def __init__(self, *a, **kwargs): super(GatherproxySpider, self).__init__(*a, **kwargs) self.urls = [ ...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
crawler/spiders/proxy/ip181.py
null
null
null
null
null
null
Python
2026-05-04T01:40:08.498721
#-*- coding: utf-8 -*- from scrapy import Selector from .basespider import BaseSpider from proxy import Proxy class IpOneEightOneSpider(BaseSpider): name = 'ip181' def __init__(self, *a, **kw): super(IpOneEightOneSpider, self).__init__(*a, **kw) self.urls = ['http://www.ip181.com/'] ...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
crawler/spiders/proxy/freeproxylists.py
null
null
null
null
null
null
Python
2026-05-04T01:40:09.063349
# coding=utf-8 import urllib import re from proxy import Proxy from .basespider import BaseSpider from bs4 import BeautifulSoup class FreeProxyListsSpider(BaseSpider): name = 'freeproxylists' def __init__(self, *a, **kwargs): super(FreeProxyListsSpider, self).__init__(*a, **kwargs) self.url...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
crawler/spiders/proxy/usproxy.py
null
null
null
null
null
null
Python
2026-05-04T01:40:09.539519
# coding=utf-8 import re from proxy import Proxy from .basespider import BaseSpider class UsProxySpider(BaseSpider): name = 'usproxy' def __init__(self, *a, **kwargs): super(UsProxySpider, self).__init__(*a, **kwargs) self.urls = [ 'http://www.sslproxies.org/', 'htt...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
crawler/spiders/proxy/hidemy.py
null
null
null
null
null
null
Python
2026-05-04T01:40:09.730696
#-*- coding: utf-8 -*- import utils from scrapy import Selector from .basespider import BaseSpider from proxy import Proxy class HidemySpider(BaseSpider): name = 'hidemy' def __init__(self, *a, **kw): super(HidemySpider, self).__init__(*a, **kw) self.urls = ['https://hidemy.name/en/proxy-l...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
crawler/spiders/validator/baidu.py
null
null
null
null
null
null
Python
2026-05-04T01:40:10.288207
#-*- coding: utf-8 -*- from .validator import Validator class BaiduSpider(Validator): name = 'baidu' def __init__(self, name = None, **kwargs): super(BaiduSpider, self).__init__(name, **kwargs) self.urls = [ 'https://www.baidu.com/' ] self.headers = { ...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
crawler/spiders/validator/amazoncn.py
null
null
null
null
null
null
Python
2026-05-04T01:40:10.289631
#-*- coding: utf-8 -*- from .validator import Validator class AmazonCnSpider(Validator): name = 'amazoncn' def __init__(self, name = None, **kwargs): super(AmazonCnSpider, self).__init__(name, **kwargs) self.timeout = 5 self.urls = [ 'https://www.amazon.cn/dp/B00ID363S4...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
crawler/spiders/validator/__init__.py
null
null
null
null
null
null
Python
2026-05-04T01:40:10.290680
# This package will contain the spiders of your Scrapy project # # Please refer to the documentation for information on how to create and manage # your spiders.
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
crawler/spiders/validator/assetstore.py
null
null
null
null
null
null
Python
2026-05-04T01:40:10.291550
#-*- coding: utf-8 -*- import json import time import config from scrapy.http import Request from .validator import Validator class AssetStoreSpider(Validator): name = 'assetstore' def __init__(self, *a, **kwargs): super(AssetStoreSpider, self).__init__(*a, **kwargs) self.timeout = 10 ...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
crawler/spiders/validator/bbs.py
null
null
null
null
null
null
Python
2026-05-04T01:40:10.454304
# -*- coding: utf-8 -*- from .validator import Validator class BBSSpider(Validator): name = 'bbs' concurrent_requests = 8 def __init__(self, name = None, **kwargs): super(BBSSpider, self).__init__(name, **kwargs) self.urls = [ 'http://www.autohome.com.cn/beijing/', ...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
crawler/spiders/validator/douban.py
null
null
null
null
null
null
Python
2026-05-04T01:40:10.994879
#-*- coding: utf-8 -*- from .validator import Validator class DoubanSpider(Validator): name = 'douban' def __init__(self, name = None, **kwargs): super(DoubanSpider, self).__init__(name, **kwargs) self.timeout = 5 self.urls = [ 'https://movie.douban.com/subject/3434070/...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
crawler/spiders/validator/httpbin.py
null
null
null
null
null
null
Python
2026-05-04T01:40:11.158830
# -*- coding: utf-8 -*- import json import time import requests import config from scrapy import Request from .validator import Validator class HttpBinSpider(Validator): name = 'httpbin' concurrent_requests = 16 def __init__(self, name=None, **kwargs): super(HttpBinSpider, self).__init__(name, ...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
crawler/spiders/validator/gather.py
null
null
null
null
null
null
Python
2026-05-04T01:40:11.165499
#-*- coding: utf-8 -*- from .validator import Validator class GatherSpider(Validator): name = 'gather' def __init__(self, name = None, **kwargs): super(GatherSpider, self).__init__(name, **kwargs) self.timeout = 10 self.urls = [ 'http://gatherproxy.com/proxylist/anonymit...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
crawler/spiders/validator/jd.py
null
null
null
null
null
null
Python
2026-05-04T01:40:11.183140
#-*- coding: utf-8 -*- import random import time import re import config from scrapy import Request from .validator import Validator class JDSpider(Validator): name = 'jd' def __init__(self, name = None, **kwargs): super(JDSpider, self).__init__(name, **kwargs) self.urls = [ 'h...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
crawler/spiders/validator/lagou.py
null
null
null
null
null
null
Python
2026-05-04T01:40:11.188301
#-*- coding: utf-8 -*- import time import config import utils from .validator import Validator from scrapy.http import FormRequest class LagouSpider(Validator): name = 'lagou' concurrent_requests = 8 def __init__(self, name = None, **kwargs): super(LagouSpider, self).__init__(name, **kwargs) ...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
crawler/spiders/validator/anjuke.py
null
null
null
null
null
null
Python
2026-05-04T01:40:11.214410
# -*- coding=utf-8 -*- import datetime import json import random import re import time import config from scrapy import Request from scrapy.selector import Selector from crawler.spiders.validator.validator import Validator class AJKSpider(Validator): name = 'ajk' concurrent_requests = 16 def __init__(...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
crawler/spiders/validator/boss.py
null
null
null
null
null
null
Python
2026-05-04T01:40:11.215575
#-*- coding: utf-8 -*- from .validator import Validator class BossSpider(Validator): name = 'boss' concurrent_requests = 8 def __init__(self, name = None, **kwargs): super(BossSpider, self).__init__(name, **kwargs) self.urls = [ 'https://www.zhipin.com/c101010100/h_101010100...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
crawler/spiders/validator/liepin.py
null
null
null
null
null
null
Python
2026-05-04T01:40:11.271124
#-*- coding: utf-8 -*- from .validator import Validator class LiepinSpider(Validator): name = 'liepin' concurrent_requests = 8 def __init__(self, name = None, **kwargs): super(LiepinSpider, self).__init__(name, **kwargs) self.urls = [ 'https://www.liepin.com/zhaopin/?pubTime...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
crawler/spiders/validator/steam.py
null
null
null
null
null
null
Python
2026-05-04T01:40:12.010827
#-*- coding: utf-8 -*- from .validator import Validator class SteamSpider(Validator): name = 'steam' def __init__(self, name = None, **kwargs): super(SteamSpider, self).__init__(name, **kwargs) self.timeout = 10 self.urls = [ 'http://store.steampowered.com/app/602580/' ...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
crawler/spiders/validator/zhilian.py
null
null
null
null
null
null
Python
2026-05-04T01:40:12.062394
#-*- coding: utf-8 -*- from .validator import Validator class ZhiLianSpider(Validator): name = 'zhilian' concurrent_requests = 8 def __init__(self, name = None, **kwargs): super(ZhiLianSpider, self).__init__(name, **kwargs) self.urls = [ 'http://www.zhaopin.com/' ] ...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
ipproxytool.py
null
null
null
null
null
null
Python
2026-05-04T01:40:12.064099
# coding=utf-8 import logging import os import sys import subprocess import run_validator import run_validator_async if __name__ == '__main__': # 进入当前项目目录 os.chdir(sys.path[0]) if not os.path.exists('log'): os.makedirs('log') logging.basicConfig( filename = 'log/ipproxy.log', ...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
crawler/spiders/validator/validator.py
null
null
null
null
null
null
Python
2026-05-04T01:40:12.091544
# -*- coding: utf-8 -*- import random import time import datetime import utils import config from scrapy import Request from scrapy.spiders import Spider from sql import SqlManager class Validator(Spider): name = 'base' concurrent_requests = 16 retry_enabled = False def __init__(self, name = None, *...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
run_validator.py
null
null
null
null
null
null
Python
2026-05-04T01:40:12.648893
# -*- coding: utf-8 -*- import logging import os import subprocess import sys import time import scrapydo import utils from importlib import import_module VALIDATORS = { 'HttpBinSpider': 'crawler.spiders.validator.httpbin', # 'DoubanSpider':'ipproxytool.spiders.validator.douban', # 'AssetStoreSpider':'ipp...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
run_validator_async.py
null
null
null
null
null
null
Python
2026-05-04T01:40:12.675469
# -*- coding: utf-8 -*- import logging import os import sys import time import utils import aiohttp from aiohttp import ClientSession from sql.sql_manager import SqlManager import config import asyncio TEST_URL='http://httpbin.org/ip' async def test_connect(proxy,operator,mode=None): conn = aiohttp.TCPConnector(...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
server/dataserver.py
null
null
null
null
null
null
Python
2026-05-04T01:40:12.684320
# -*- coding: utf-8 -*- import json import logging import sys import config from proxy import Proxy from sql import SqlManager from flask import Flask from flask import request app = Flask(__name__) @app.route('/') def index(): return 'Hello, World!' @app.route('/insert') def insert(): sql = SqlManager()...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
proxy.py
null
null
null
null
null
null
Python
2026-05-04T01:40:12.804617
# -*- coding: utf-8 -*- class Proxy(object): def __init__(self): self.id = 1 self.ip = '' self.port = '' self.country = '' self.anonymity = '' self.https = '' self.speed = '' self.source = '' self.vali_count = 0 def __str__(self): ...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
run_server.py
null
null
null
null
null
null
Python
2026-05-04T01:40:12.852732
# -*- coding: utf-8 -*- import os import logging import config import utils from server import dataserver if __name__ == '__main__': if not os.path.exists('log'): os.makedirs('log') logging.basicConfig( filename='log/server.log', format='%(levelname)s %(asctime)s: %(message)s', ...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
run_spider.py
null
null
null
null
null
null
Python
2026-05-04T01:40:12.898209
# -*- coding: utf-8 -*- import os import logging import sys from scrapy.crawler import CrawlerProcess from scrapy.utils.log import configure_logging from scrapy.utils.project import get_project_settings def runspider(name): configure_logging(install_root_handler=False) logging.basicConfig( filename=...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
run_crawl_proxy.py
null
null
null
null
null
null
Python
2026-05-04T01:40:12.916373
# -*- coding: utf-8 -*- import logging import os import sys import scrapydo import time import utils import config from sql import SqlManager from crawler.spiders.proxy.xicidaili import XiCiDaiLiSpider from crawler.spiders.proxy.sixsixip import SixSixIpSpider from crawler.spiders.proxy.ip181 import IpOneEightOneSpide...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
sql/__init__.py
null
null
null
null
null
null
Python
2026-05-04T01:40:13.871055
#-*- coding: utf-8 -*- import sql.sql_base from sql.mysql import MySql from sql.mongodb import Mongodb from sql.sql_manager import SqlManager
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
sql/sql_base.py
null
null
null
null
null
null
Python
2026-05-04T01:40:13.871573
#-*- coding: utf-8 -*- class SqlBase(object): def __init__(self, **kwargs): pass def init_database(self, database_name): pass def init_proxy_table(self, table_name): pass def insert_proxy(self, table_name, proxy): pass def select_proxy(self, table_name, **kwargs)...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
crawler/spiders/proxy/proxylistplus.py
null
null
null
null
null
null
Python
2026-05-04T01:40:14.056127
#-*- coding: utf-8 -*- from scrapy import Selector from .basespider import BaseSpider from proxy import Proxy class ProxylistplusSpider(BaseSpider): name = 'proxylistplus' def __init__(self, *a, **kw): super(ProxylistplusSpider, self).__init__(*a, **kw) self.urls = ['https://list.proxylistp...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
sql/mongodb.py
null
null
null
null
null
null
Python
2026-05-04T01:40:14.156949
# -*- coding: utf-8 -*- import logging import pymongo import config import json import datetime from proxy import Proxy from sql.sql_base import SqlBase class Mongodb(SqlBase): def __init__(self, **kwargs): super(Mongodb, self).__init__(**kwargs) self.client = pymongo.MongoClient(**kwargs) ...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
sql/sql_manager.py
null
null
null
null
null
null
Python
2026-05-04T01:40:14.425820
# -*- coding: utf-8 -*- import config from sql.sql_base import SqlBase class SqlManager(object): def __init__(self): db_type = config.DB_config.get('db_type', 'mysql') db_config = config.DB_config.get(db_type) if db_type == 'mysql': from sql.mysql import MySql se...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
sql/mysql.py
null
null
null
null
null
null
Python
2026-05-04T01:40:14.556634
# -*- coding: utf-8 -*- import logging import utils import config import pymysql from proxy import Proxy from sql.sql_base import SqlBase class MySql(SqlBase): def __init__(self, **kwargs): super(MySql, self).__init__(**kwargs) self.conn = pymysql.connect(**kwargs) self.cursor = self.co...
awolfly9/IPProxyTool
https://github.com/awolfly9/IPProxyTool
null
null
null
null
1,998
null
null
mit
null
null
null
null
null
null
null
crawler/spiders/proxy/xicidaili.py
null
null
null
null
null
null
Python
2026-05-04T01:40:18.243295
#-*- coding: utf-8 -*- from proxy import Proxy from .basespider import BaseSpider from scrapy.selector import Selector class XiCiDaiLiSpider(BaseSpider): name = 'xici' def __init__(self, *a, **kw): super(XiCiDaiLiSpider, self).__init__(*a, **kw) self.urls = ['http://www.xicidaili.com/nn/%s'...