code stringlengths 22 1.05M | apis listlengths 1 3.31k | extract_api stringlengths 75 3.25M |
|---|---|---|
from Dialogflow_Api import rispondimi
from collegamentoSito import inserisci_utente
from Nutrition import get_food, traduzione
import re
tipo_cibo = ["frutta", "carne", "verdure", "ortaggi", "primi_piatti", "legumi"]
"""
controllo_intent(query_result, utente)--> text_respose
prende il risultato della query e lo confr... | [
"re.split",
"collegamentoSito.inserisci_utente",
"Nutrition.traduzione",
"Nutrition.get_food",
"Dialogflow_Api.rispondimi"
] | [((2545, 2569), 'collegamentoSito.inserisci_utente', 'inserisci_utente', (['utente'], {}), '(utente)\n', (2561, 2569), False, 'from collegamentoSito import inserisci_utente\n'), ((4345, 4369), 'collegamentoSito.inserisci_utente', 'inserisci_utente', (['utente'], {}), '(utente)\n', (4361, 4369), False, 'from collegament... |
# Copyright (c) 2019, NVIDIA CORPORATION. All rights reserved.
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
# http://www.apache.org/licenses/LICENSE-2.0
# Unless required by applicabl... | [
"architectures.Generator",
"argparse.ArgumentParser",
"kaolin.rep.voxel.max_connected",
"kaolin.rep.QuadMesh.from_tensors",
"kaolin.conversion.voxel.to_mesh_quad",
"torch.zeros",
"torch.ones"
] | [((914, 939), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (937, 939), False, 'import argparse\n'), ((1542, 1581), 'kaolin.rep.voxel.max_connected', 'kal.rep.voxel.max_connected', (['model', '(0.5)'], {}), '(model, 0.5)\n', (1569, 1581), True, 'import kaolin as kal\n'), ((1597, 1637), 'kaolin... |
"""Defines procedures for training, and evaluation automatic camfi annotation models,
and for using them for making automatic annotations (inference). Depends on camfi.util,
camfi.datamodel.autoannotation, camfi.datamodel.geometry, camfi.datamode.via, as well
as ._torchutils and ._models."""
from datetime import datet... | [
"numpy.count_nonzero",
"numpy.array",
"camfi.datamodel.via.ViaRegionAttributes",
"numpy.arange",
"camfi.datamodel.geometry.CircleShapeAttributes",
"pathlib.Path",
"numpy.where",
"camfi.datamodel.autoannotation.Prediction.from_tensor_dict",
"torch.hub.load_state_dict_from_url",
"scipy.sparse.csgrap... | [((3085, 3091), 'pathlib.Path', 'Path', ([], {}), '()\n', (3089, 3091), False, 'from pathlib import Path\n'), ((4578, 4598), 'torch.device', 'torch.device', (['device'], {}), '(device)\n', (4590, 4598), False, 'import torch\n'), ((4729, 4838), 'torch.utils.data.DataLoader', 'DataLoader', (['dataset'], {'batch_size': 'b... |
"""
Wrapper for running amplicon_analysis_inputs report for LAAgc (with an extra
input file).
"""
import logging
import sys
from pbcommand.models import FileTypes
from pbcommand.cli import pbparser_runner
from pbcommand.utils import setup_log
from pbreports.report.amplicon_analysis_input import _get_parser, make_rep... | [
"logging.getLogger",
"pbreports.report.amplicon_analysis_input._get_parser",
"pbreports.report.amplicon_analysis_input.make_report"
] | [((332, 359), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (349, 359), False, 'import logging\n'), ((551, 642), 'pbreports.report.amplicon_analysis_input.make_report', 'make_report', (['args.report_csv', 'args.report_json', 'args.locus_csv', 'args.barcoded_subreads'], {}), '(args.report... |
# The MIT License (MIT)
# Copyright © 2021 Opentensor.ai
# Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated
# documentation files (the “Software”), to deal in the Software without restriction, including without limitation
# the rights to use, copy, modify,... | [
"os.path.exists",
"loguru.logger.info",
"argparse.ArgumentParser",
"bittensor.crypto.decrypt_data",
"bittensor.crypto.is_encrypted",
"os.access",
"re.match",
"os.chmod",
"os.path.isfile",
"bittensor.crypto.keyfiles.load_keypair_from_data",
"loguru.logger.error",
"bittensor.utils.Cli.ask_passwo... | [((2475, 2499), 'os.path.expanduser', 'os.path.expanduser', (['path'], {}), '(path)\n', (2493, 2499), False, 'import os\n'), ((2597, 2644), 'loguru.logger.info', 'logger.info', (['"""Loaded coldkey: {}"""', 'self.coldkey'], {}), "('Loaded coldkey: {}', self.coldkey)\n", (2608, 2644), False, 'from loguru import logger\n... |
import unittest
from datetime import datetime, timezone
from pandas import DatetimeTZDtype
from parameterized import parameterized
import pandas as pd
from aiokraken.utils.timeindexeddataframe import TimeindexedDataframe
"""
Test module.
This is intended for extensive testing, using parameterized, hypothesis or simi... | [
"datetime.datetime",
"pandas.Series",
"datetime.datetime.fromtimestamp",
"pandas.api.types.is_datetime64_ns_dtype",
"pandas.api.types.is_numeric_dtype",
"aiokraken.utils.timeindexeddataframe.TimeindexedDataframe",
"pandas.api.types.is_datetime64tz_dtype",
"unittest.main",
"pandas.DatetimeTZDtype",
... | [((20171, 20186), 'unittest.main', 'unittest.main', ([], {}), '()\n', (20184, 20186), False, 'import unittest\n'), ((3430, 3477), 'aiokraken.utils.timeindexeddataframe.TimeindexedDataframe', 'TimeindexedDataframe', ([], {'data': 'df', 'index': '"""datetime"""'}), "(data=df, index='datetime')\n", (3450, 3477), False, 'f... |
#!/usr/bin/env python
# coding: utf-8
# In[1]:
from bs4 import BeautifulSoup
import re
import json
import numpy as np
import sys
sys.setrecursionlimit(10000)
# Write a function that parses HTML pages into BeautifulSoup objects
# In[2]:
def soupify(html):
soup = BeautifulSoup(html, 'lxml')
return soup
... | [
"sys.setrecursionlimit",
"json.loads",
"re.compile",
"bs4.BeautifulSoup",
"re.findall"
] | [((132, 160), 'sys.setrecursionlimit', 'sys.setrecursionlimit', (['(10000)'], {}), '(10000)\n', (153, 160), False, 'import sys\n'), ((275, 302), 'bs4.BeautifulSoup', 'BeautifulSoup', (['html', '"""lxml"""'], {}), "(html, 'lxml')\n", (288, 302), False, 'from bs4 import BeautifulSoup\n'), ((717, 730), 'json.loads', 'json... |
# coding: utf-8
"""
Engine api
Engine APIs # noqa: E501
The version of the OpenAPI document: 1.0.4
Generated by: https://openapi-generator.tech
"""
import pprint
import re # noqa: F401
import six
class EngineTask(object):
"""NOTE: This class is auto generated by OpenAPI Generator.
Ref:... | [
"six.iteritems"
] | [((12047, 12080), 'six.iteritems', 'six.iteritems', (['self.openapi_types'], {}), '(self.openapi_types)\n', (12060, 12080), False, 'import six\n')] |
"""
Analyses skewness for continuous features
Options:
A. Log
B. Yeo-Johnson
C. QuantileTransformer
"""
import json
import pandas as pd
import numpy as np
from sklearn.pipeline import make_pipeline
from sklearn.preprocessing import power_transform, quantile_transform
from pathlib import Path
p = Path(__f... | [
"numpy.reshape",
"pandas.concat",
"src.logger.LOGGER.info",
"pathlib.Path"
] | [((464, 488), 'src.logger.LOGGER.info', 'LOGGER.info', (['"""Load data"""'], {}), "('Load data')\n", (475, 488), False, 'from src.logger import LOGGER\n'), ((576, 615), 'src.logger.LOGGER.info', 'LOGGER.info', (['"""Process data - Logarithm"""'], {}), "('Process data - Logarithm')\n", (587, 615), False, 'from src.logge... |
import pygame
pygame.init()
FPS = 60
WIDTH, HEIGHT = 800, 700
ROWS, COLS = 7, 8
SQUARE_SIZE = HEIGHT//ROWS
# GRAPHICAL USER INTERFACE
ICON_PATH = './src/murus_gallicus/assets/noun_checkers_1684698.png'
WINDOW = pygame.display.set_mode((WIDTH, HEIGHT))
PADDING = 20
OUTLINE = 2
# RGB COLORS
RED = (255, 0, 0)
WHITE = ... | [
"pygame.display.set_mode",
"pygame.init"
] | [((14, 27), 'pygame.init', 'pygame.init', ([], {}), '()\n', (25, 27), False, 'import pygame\n'), ((213, 253), 'pygame.display.set_mode', 'pygame.display.set_mode', (['(WIDTH, HEIGHT)'], {}), '((WIDTH, HEIGHT))\n', (236, 253), False, 'import pygame\n')] |
import deepchem as dc
import numpy as np
import os
def test_numpy_dataset_get_shape():
"""Test that get_shape works for numpy datasets."""
num_datapoints = 100
num_features = 10
num_tasks = 10
# Generate data
X = np.random.rand(num_datapoints, num_features)
y = np.random.randint(2, size=(num_datapoints,... | [
"deepchem.data.DiskDataset.from_numpy",
"deepchem.data.CSVLoader",
"numpy.random.rand",
"os.path.join",
"numpy.array",
"numpy.random.randint",
"deepchem.feat.DummyFeaturizer",
"deepchem.data.NumpyDataset",
"os.path.abspath",
"deepchem.data.DiskDataset"
] | [((227, 271), 'numpy.random.rand', 'np.random.rand', (['num_datapoints', 'num_features'], {}), '(num_datapoints, num_features)\n', (241, 271), True, 'import numpy as np\n'), ((278, 332), 'numpy.random.randint', 'np.random.randint', (['(2)'], {'size': '(num_datapoints, num_tasks)'}), '(2, size=(num_datapoints, num_tasks... |
# -*- coding: utf-8 -*- #
# Copyright 2015 Google LLC. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless requir... | [
"collections.OrderedDict"
] | [((1176, 1278), 'collections.OrderedDict', 'collections.OrderedDict', (["[('second', 1), ('minute', 60), ('hour', 60 * 60), ('day', 60 * 60 * 24)]"], {}), "([('second', 1), ('minute', 60), ('hour', 60 * 60),\n ('day', 60 * 60 * 24)])\n", (1199, 1278), False, 'import collections\n')] |
import time
a = 3215.35127
b = 3.
start = time.time()
for i in range(100000000):
c = a / b
end = time.time()
time_elapsed = end - start
print('Time elapsed (div ver) = %.5f' % time_elapsed)
a = 3215.35127
b = 1./3.
start = time.time()
for i in range(100000000):
c = a * b
end = time.time()
time_elapsed = end... | [
"time.time"
] | [((44, 55), 'time.time', 'time.time', ([], {}), '()\n', (53, 55), False, 'import time\n'), ((103, 114), 'time.time', 'time.time', ([], {}), '()\n', (112, 114), False, 'import time\n'), ((231, 242), 'time.time', 'time.time', ([], {}), '()\n', (240, 242), False, 'import time\n'), ((290, 301), 'time.time', 'time.time', ([... |
# coding: utf-8
# Copyright (c) 2016, 2021, Oracle and/or its affiliates. All rights reserved.
# This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may c... | [
"oci.util.formatted_flat_dict",
"oci.util.value_allowed_none_or_none_sentinel"
] | [((5155, 5180), 'oci.util.formatted_flat_dict', 'formatted_flat_dict', (['self'], {}), '(self)\n', (5174, 5180), False, 'from oci.util import formatted_flat_dict, NONE_SENTINEL, value_allowed_none_or_none_sentinel\n'), ((4970, 5032), 'oci.util.value_allowed_none_or_none_sentinel', 'value_allowed_none_or_none_sentinel',... |
from django.contrib import admin
from .models import Search
# Register your models here.
admin.site.register(Search) | [
"django.contrib.admin.site.register"
] | [((90, 117), 'django.contrib.admin.site.register', 'admin.site.register', (['Search'], {}), '(Search)\n', (109, 117), False, 'from django.contrib import admin\n')] |
# <NAME>
# Institute of Mechatronic Systems
# Leibniz Universität Hannover, Germany
# 2019
# Code From https://github.com/mlaves/bayesian-temperature-scaling
import torch
__all__ = ['accuracy', 'kl_loss', 'nentr', 'xavier_normal_init']
def accuracy(input, target):
_, max_indices = torch.max(input.data, 1)
ac... | [
"torch.tensor",
"torch.max",
"torch.nn.functional.log_softmax",
"torch.nn.init.xavier_normal_"
] | [((289, 313), 'torch.max', 'torch.max', (['input.data', '(1)'], {}), '(input.data, 1)\n', (298, 313), False, 'import torch\n'), ((756, 794), 'torch.tensor', 'torch.tensor', (['[1e-16]'], {'device': 'p.device'}), '([1e-16], device=p.device)\n', (768, 794), False, 'import torch\n'), ((823, 881), 'torch.tensor', 'torch.te... |
import unittest
from core.drive import copy
from core.aesthetics import *
class TestingDrivepy(unittest.TestCase):
def test_copy_function(self):
source_tests = [
r"D:\Alexzander__\programming\python\Python2Executable",
r"D:\Alexzander__\programming\python\byzantion",
... | [
"unittest.main",
"core.drive.copy"
] | [((1875, 1890), 'unittest.main', 'unittest.main', ([], {}), '()\n', (1888, 1890), False, 'import unittest\n'), ((1266, 1340), 'core.drive.copy', 'copy', (['source', 'destination'], {'open_destination_when_done': '(False)', '__print': '(False)'}), '(source, destination, open_destination_when_done=False, __print=False)\n... |
from app import app
from flask_restful import Api
from app.resources.auth import TokenResource
api = Api(app)
# Token resource
api.add_resource(TokenResource, '/authservice/token', endpoint='auth_token')
| [
"flask_restful.Api"
] | [((103, 111), 'flask_restful.Api', 'Api', (['app'], {}), '(app)\n', (106, 111), False, 'from flask_restful import Api\n')] |
#! /usr/bin/env python
import os
import rospy
import rospkg
from readbag import restore
from qt_gui.plugin import Plugin
from python_qt_binding.QtCore import Qt
from python_qt_binding import loadUi
from python_qt_binding.QtGui import QFileDialog, QGraphicsView, QIcon, QWidget
from PyQt4 import QtGui, QtCore
from exam... | [
"PyQt4.QtGui.QApplication",
"PyQt4.QtGui.QWidget.__init__"
] | [((812, 840), 'PyQt4.QtGui.QApplication', 'QtGui.QApplication', (['sys.argv'], {}), '(sys.argv)\n', (830, 840), False, 'from PyQt4 import QtGui\n'), ((470, 506), 'PyQt4.QtGui.QWidget.__init__', 'QtGui.QWidget.__init__', (['self', 'parent'], {}), '(self, parent)\n', (492, 506), False, 'from PyQt4 import QtGui\n')] |
import models
import os
import copy
import torch
import torch.nn as nn
from lifelines import KaplanMeierFitter as KMFitter
import pycox
import numpy as np
# local
import catdist
import data_utils
import _concordance
import _nll
import _saver
def str_to_bool(arg):
"""Convert an argument string into its boolean val... | [
"numpy.unique",
"catdist.CatDist",
"torch.zeros_like",
"torch.isnan",
"torch.zeros"
] | [((1717, 1778), 'catdist.CatDist', 'catdist.CatDist', ([], {'logits': 'None', 'args': 'args', 'probs': 'Gprobs', 'k': 'None'}), '(logits=None, args=args, probs=Gprobs, k=None)\n', (1732, 1778), False, 'import catdist\n'), ((699, 713), 'torch.isnan', 'torch.isnan', (['x'], {}), '(x)\n', (710, 713), False, 'import torch\... |
from distutils.core import setup
from pathlib import Path
# TODO(joschnei): Add typing info
setup(
name="active",
version="0.1",
packages=["active",],
install_requires=[
"scipy",
"numpy",
"driver @ git+https://github.com/jordan-schneider/driver-env.git#egg=driver",
],
p... | [
"distutils.core.setup"
] | [((94, 318), 'distutils.core.setup', 'setup', ([], {'name': '"""active"""', 'version': '"""0.1"""', 'packages': "['active']", 'install_requires': "['scipy', 'numpy',\n 'driver @ git+https://github.com/jordan-schneider/driver-env.git#egg=driver'\n ]", 'package_data': "{'active': ['py.typed']}"}), "(name='active', ... |
import tableprint
from collections import Counter
from .args import *
from .util import *
if __name__ == "__main__":
args = get_args()
output_classes = Counter()
question_types = Counter()
with tableprint.TableContext(headers=["Type", "Question", "Answer"], width=[40,50,15]) as t:
for i in read_gqa(args):... | [
"collections.Counter",
"tableprint.TableContext"
] | [((160, 169), 'collections.Counter', 'Counter', ([], {}), '()\n', (167, 169), False, 'from collections import Counter\n'), ((188, 197), 'collections.Counter', 'Counter', ([], {}), '()\n', (195, 197), False, 'from collections import Counter\n'), ((206, 294), 'tableprint.TableContext', 'tableprint.TableContext', ([], {'h... |
#!/usr/bin/env python3
# Copyright (2021-) <NAME> <<EMAIL>>
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicab... | [
"tensorflow.nn.conv2d",
"tensorflow.shape",
"tensorflow.keras.activations.get",
"tensorflow.keras.initializers.GlorotUniform",
"kaldi_tflite.lib.layers.tdnn.utils.reshapeKaldiTdnnWeights",
"tensorflow.range",
"tensorflow.gather",
"tensorflow.constant",
"tensorflow.clip_by_value",
"tensorflow.expan... | [((1765, 1780), 'tensorflow.keras.initializers.GlorotUniform', 'GlorotUniform', ([], {}), '()\n', (1778, 1780), False, 'from tensorflow.keras.initializers import Initializer, GlorotUniform\n'), ((1831, 1846), 'tensorflow.keras.initializers.GlorotUniform', 'GlorotUniform', ([], {}), '()\n', (1844, 1846), False, 'from te... |
from django.shortcuts import render
from rest_framework.response import Response
from rest_framework.views import APIView
from rest_framework import status, permissions
import pyotp
import base64
from rest_framework_simplejwt.tokens import RefreshToken
from .serializers import CustomUserSerializer
from .models import ... | [
"rest_framework_simplejwt.tokens.RefreshToken.for_user",
"rest_framework.response.Response",
"pyotp.HOTP"
] | [((1059, 1111), 'rest_framework.response.Response', 'Response', (["{'status': 201, 'userdata': user.username}"], {}), "({'status': 201, 'userdata': user.username})\n", (1067, 1111), False, 'from rest_framework.response import Response\n'), ((1623, 1667), 'rest_framework.response.Response', 'Response', ([], {'status': '... |
import argparse
parser = argparse.ArgumentParser(
formatter_class=argparse.ArgumentDefaultsHelpFormatter)
parser.add_argument('--d_model', type=int, default=0, help='d_model')
parser.add_argument('--d_head', type=int, default=2, help='head')
parser.add_argument('--d_inner', type=bool, default=True, help='inner lay... | [
"argparse.ArgumentParser"
] | [((25, 104), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'formatter_class': 'argparse.ArgumentDefaultsHelpFormatter'}), '(formatter_class=argparse.ArgumentDefaultsHelpFormatter)\n', (48, 104), False, 'import argparse\n')] |
import logging
import sys
import os
from logging.handlers import RotatingFileHandler
from multiprocessing.pool import ThreadPool
from optparse import OptionParser
import requests
from requests.packages import urllib3
urllib3.disable_warnings()
# Workers configurations
ASYNC_WORKERS_COUNT = 100 # How many threads wi... | [
"logging.basicConfig",
"logging.getLogger",
"requests.session",
"os.path.exists",
"requests.packages.urllib3.disable_warnings",
"os.listdir",
"os.linesep.join",
"logging.Formatter",
"os.path.join",
"optparse.OptionParser",
"multiprocessing.pool.ThreadPool",
"os.path.isfile",
"os.path.isdir",... | [((219, 245), 'requests.packages.urllib3.disable_warnings', 'urllib3.disable_warnings', ([], {}), '()\n', (243, 245), False, 'from requests.packages import urllib3\n'), ((1927, 2005), 'logging.basicConfig', 'logging.basicConfig', ([], {'stream': 'sys.stdout', 'level': 'logging_level', 'format': 'LOG_FORMAT'}), '(stream... |
#Template of the Purkinje cell model, Zang et al. 2018
#Templating by Lungsi 2019 based on ~/PC2018Zang/purkinje.hoc
#purkinje.hoc has been converted from original purkinje_demo and using readme.html as a guide
from neuron import h
#from pdb import set_trace as breakpoint
from random import randint
class Purkinje(obje... | [
"neuron.h.xopen"
] | [((395, 418), 'neuron.h.xopen', 'h.xopen', (['"""purkinje.hoc"""'], {}), "('purkinje.hoc')\n", (402, 418), False, 'from neuron import h\n')] |
"""
@brief Class to store data from a specific event
@date Created July 2, 2018
@author <NAME>
@bug No known bugs
"""
from fprime.common.models.serialize import time_type
from fprime_gds.common.data_types import sys_data
class EventData(sys_data.SysData):
"""
The EventData class stores a specific event mes... | [
"fprime.common.models.serialize.time_type.TimeType"
] | [((1702, 1722), 'fprime.common.models.serialize.time_type.TimeType', 'time_type.TimeType', ([], {}), '()\n', (1720, 1722), False, 'from fprime.common.models.serialize import time_type\n')] |
from os.path import exists, join
import pytest
from flask_dj import startproject
from tests.basic_project_creator import ProjectCreate
class TestBaseSettingProjectConstructor(ProjectCreate):
def setup(self, need_static=False, need_templates=False):
super().setup()
def test_project_folder_exist(self):... | [
"os.path.exists",
"flask_dj.startproject",
"os.path.join",
"pytest.raises"
] | [((336, 361), 'os.path.exists', 'exists', (['self.project_path'], {}), '(self.project_path)\n', (342, 361), False, 'from os.path import exists, join\n'), ((423, 465), 'os.path.join', 'join', (['self.project_path', 'self.project_name'], {}), '(self.project_path, self.project_name)\n', (427, 465), False, 'from os.path im... |
from django.test import TestCase
from rest_framework.test import APIRequestFactory
from .views import LoginView, SignupView
from common.tests import USER_DATA, prepare_dummy_user_data
# Create your tests here.
class ApiViewTests(TestCase):
factory = APIRequestFactory()
def test_signup(self):
user_da... | [
"rest_framework.test.APIRequestFactory",
"common.tests.prepare_dummy_user_data"
] | [((257, 276), 'rest_framework.test.APIRequestFactory', 'APIRequestFactory', ([], {}), '()\n', (274, 276), False, 'from rest_framework.test import APIRequestFactory\n'), ((949, 974), 'common.tests.prepare_dummy_user_data', 'prepare_dummy_user_data', ([], {}), '()\n', (972, 974), False, 'from common.tests import USER_DAT... |
# Copyright 2015 The Chromium Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
'''
Utility functions for all things related to manipulating google play services
related files.
'''
import argparse
import filecmp
import json
import loggi... | [
"logging.debug",
"re.compile",
"devil.utils.cmd_helper.Call",
"json.dumps",
"os.path.isfile",
"os.path.dirname",
"json.load",
"filecmp.cmp"
] | [((490, 567), 're.compile', 're.compile', (['"""<integer name="google_play_services_version">(\\\\d+)<\\\\/integer>"""'], {}), '(\'<integer name="google_play_services_version">(\\\\d+)<\\\\/integer>\')\n', (500, 567), False, 'import re\n'), ((3565, 3612), 'json.dumps', 'json.dumps', (['json_data'], {'sort_keys': '(True... |
# -*-coding: utf-8
import sys
import argparse
from tagger import FileTagger
def tagger_add(args):
tg = FileTagger()
res = tg.add_tags(args.path, *args.tags)
if not res:
print("[-] Fail to add tags.")
def tagger_rm(args):
tg = FileTagger()
res = tg.rm_tags(args.path, *args.tags)
if n... | [
"argparse.ArgumentParser",
"tagger.FileTagger"
] | [((110, 122), 'tagger.FileTagger', 'FileTagger', ([], {}), '()\n', (120, 122), False, 'from tagger import FileTagger\n'), ((255, 267), 'tagger.FileTagger', 'FileTagger', ([], {}), '()\n', (265, 267), False, 'from tagger import FileTagger\n'), ((404, 416), 'tagger.FileTagger', 'FileTagger', ([], {}), '()\n', (414, 416),... |
import os
FILES_TO_REMOVE = ["setup.cfg", "pyproject.toml", "setup.py"]
if "{{ cookiecutter.as_package }}" == "no":
for file in FILES_TO_REMOVE:
os.remove(file)
if "{{ cookiecutter.remove_test_script }}" == "yes":
os.remove("test_log.py")
| [
"os.remove"
] | [((233, 257), 'os.remove', 'os.remove', (['"""test_log.py"""'], {}), "('test_log.py')\n", (242, 257), False, 'import os\n'), ((159, 174), 'os.remove', 'os.remove', (['file'], {}), '(file)\n', (168, 174), False, 'import os\n')] |
from django import forms
from crispy_forms.helper import FormHelper
class MiscForm(forms.Form):
notes = forms.CharField(
label='Notes for the registration team',
help_text='Anything else you need to describe. '
'The registration team will see this. '
'The bursa... | [
"django.forms.Textarea",
"crispy_forms.helper.FormHelper"
] | [((528, 540), 'crispy_forms.helper.FormHelper', 'FormHelper', ([], {}), '()\n', (538, 540), False, 'from crispy_forms.helper import FormHelper\n'), ((357, 390), 'django.forms.Textarea', 'forms.Textarea', ([], {'attrs': "{'rows': 3}"}), "(attrs={'rows': 3})\n", (371, 390), False, 'from django import forms\n')] |
#Import libraries
import cv2
#Read image file into n dimensional numpy array
#Read as greyscale image -> 0, Read as color image w/o transparency -> 1, Read as color image w transparency -> -1
img = cv2.imread("Practice/images/galaxy.jpg", 0)
print(type(img))
print(img)
#To know the number of rows and columns
print(i... | [
"cv2.imwrite",
"cv2.imshow",
"cv2.destroyAllWindows",
"cv2.waitKey",
"cv2.imread"
] | [((199, 242), 'cv2.imread', 'cv2.imread', (['"""Practice/images/galaxy.jpg"""', '(0)'], {}), "('Practice/images/galaxy.jpg', 0)\n", (209, 242), False, 'import cv2\n'), ((667, 700), 'cv2.imshow', 'cv2.imshow', (['"""Galaxy"""', 'resized_img'], {}), "('Galaxy', resized_img)\n", (677, 700), False, 'import cv2\n'), ((797, ... |
import asyncio
from .main import create_app
loop = asyncio.get_event_loop()
app = create_app(loop)
| [
"asyncio.get_event_loop"
] | [((53, 77), 'asyncio.get_event_loop', 'asyncio.get_event_loop', ([], {}), '()\n', (75, 77), False, 'import asyncio\n')] |
import io, re, codecs, sys, os.path
import pprint
REG = 0
SHFT = 1
CTRL = 2
ALT = 4
ALTGR = 6
COMPRESSED_OUTPUT=1
def get_kbd_layout(base_filename, load_patch = False):
filename_klc = base_filename
filename_changes = base_filename + 'patch'
f = io.open(filename_klc, mode="r", encoding="utf-8")
lines = f.r... | [
"re.split",
"pprint.pformat",
"io.open",
"sys.exit"
] | [((258, 307), 'io.open', 'io.open', (['filename_klc'], {'mode': '"""r"""', 'encoding': '"""utf-8"""'}), "(filename_klc, mode='r', encoding='utf-8')\n", (265, 307), False, 'import io, re, codecs, sys, os.path\n'), ((14988, 15050), 'sys.exit', 'sys.exit', (["('unknown locale format: ' + kbd_layout['localename'])"], {}), ... |
import pandas as pd
from gooey import Gooey, GooeyParser
import numpy as np
import xlsxwriter
import xlrd
@Gooey(program_name="FEC FILE FOR FRANCE", required_cols= 4,default_size=(710, 700),navigation='TABBED', header_bg_color = '#48a7fa')
def parse_args():
parser = GooeyParser()
FilesGL = par... | [
"pandas.ExcelWriter",
"gooey.Gooey",
"gooey.GooeyParser",
"pandas.read_excel"
] | [((118, 254), 'gooey.Gooey', 'Gooey', ([], {'program_name': '"""FEC FILE FOR FRANCE"""', 'required_cols': '(4)', 'default_size': '(710, 700)', 'navigation': '"""TABBED"""', 'header_bg_color': '"""#48a7fa"""'}), "(program_name='FEC FILE FOR FRANCE', required_cols=4, default_size=(\n 710, 700), navigation='TABBED', he... |
import boto3
def handler(event, _):
boto3.client('codebuild').start_build(
projectName=event['Records'][0]['customData'])
| [
"boto3.client"
] | [((38, 63), 'boto3.client', 'boto3.client', (['"""codebuild"""'], {}), "('codebuild')\n", (50, 63), False, 'import boto3\n')] |
# menu.py
# 维护暂停界面
import pygame
from pygame.locals import *
import sys
from utility import globe
from process.scene import menu_confirm
from PIL import Image, ImageFilter
class Pause_Menu(object):
# 暂停页面
def __init__(self):
self.button_rect = []
self.rs = globe.destiny.rsManager.image
... | [
"utility.globe.destiny.msManager.play_SE",
"pygame.image.tostring",
"pygame.quit",
"pygame.event.get",
"pygame.image.fromstring",
"utility.globe.destiny.back",
"utility.globe.destiny.call",
"utility.globe.destiny.msManager.unpause",
"sys.exit",
"utility.globe.destiny.screen.get_size"
] | [((1175, 1193), 'pygame.event.get', 'pygame.event.get', ([], {}), '()\n', (1191, 1193), False, 'import pygame\n'), ((5274, 5305), 'utility.globe.destiny.screen.get_size', 'globe.destiny.screen.get_size', ([], {}), '()\n', (5303, 5305), False, 'from utility import globe\n'), ((5502, 5551), 'pygame.image.tostring', 'pyga... |
from setuptools import setup
# get the version here
pkg_vars = {}
with open("version.py") as fp:
exec(fp.read(), pkg_vars)
setup(
name='sn_simulation',
version= pkg_vars['__version__'],
description='Simulations for supernovae',
url='http://github.com/lsstdesc/sn_simulation',
author='<NAME>',... | [
"setuptools.setup"
] | [((131, 514), 'setuptools.setup', 'setup', ([], {'name': '"""sn_simulation"""', 'version': "pkg_vars['__version__']", 'description': '"""Simulations for supernovae"""', 'url': '"""http://github.com/lsstdesc/sn_simulation"""', 'author': '"""<NAME>"""', 'author_email': '"""<EMAIL>"""', 'license': '"""BSD"""', 'packages':... |
import contextlib
import datetime
import json
import StringIO
from django.contrib.auth import logout as logout_user
from django.contrib import messages
from django.core.urlresolvers import reverse
from django.core.exceptions import ObjectDoesNotExist
from django.http import HttpResponseRedirect, JsonResponse, Http404,... | [
"simpleoncall.forms.schedule.TeamScheduleForm",
"django.http.HttpResponseBadRequest",
"simpleoncall.models.Alert.objects.get",
"django.db.models.Count",
"django.contrib.messages.warning",
"django.utils.http.urlencode",
"django.core.urlresolvers.reverse",
"django.utils.http.urlquote",
"simpleoncall.f... | [((1151, 1175), 'simpleoncall.decorators.require_authentication', 'require_authentication', ([], {}), '()\n', (1173, 1175), False, 'from simpleoncall.decorators import require_authentication, require_selected_team\n'), ((1177, 1200), 'simpleoncall.decorators.require_selected_team', 'require_selected_team', ([], {}), '(... |
from bs4 import BeautifulSoup
import requests
import os
""" This script download all songs with given genre from midiworld.com
"""
genre_name = input(
"type in genre name (lowercase, no space, no special characters): ")
# Just in case someone don't respect the rules.
genre_name = genre_name.lower()
genre_name = g... | [
"os.path.join",
"requests.get",
"bs4.BeautifulSoup",
"os.path.isdir",
"os.mkdir"
] | [((392, 446), 'os.path.join', 'os.path.join', (['"""genresDataset"""', 'genre_name', '"""midiworld"""'], {}), "('genresDataset', genre_name, 'midiworld')\n", (404, 446), False, 'import os\n'), ((454, 475), 'os.path.isdir', 'os.path.isdir', (['folder'], {}), '(folder)\n', (467, 475), False, 'import os\n'), ((481, 497), ... |
#!/usr/bin/env python
# coding: utf-8
__author__ = '<NAME>'
__copyright__ = 'Copyright 2017-2020, <NAME>'
__license__ = 'MIT'
__version__ = '0.5'
__email__ = '<EMAIL>'
__status__ = 'Development'
__description__ = 'Tkinter based GUI, visualizing PASCAL VOC object detection annotation'
"""
Changelog:
- 2020-06-16 11:3... | [
"logging.getLogger",
"Tkinter.Entry",
"cv2.rectangle",
"logging.StreamHandler",
"colorsys.hsv_to_rgb",
"numpy.array",
"PIL.ImageDraw.Draw",
"Tkinter.Scrollbar",
"Tkinter.LabelFrame",
"Tkinter.Label",
"os.path.exists",
"os.listdir",
"Tkinter.Tk.__init__",
"Tkinter.Listbox",
"Tkinter.Strin... | [((2452, 2471), 'PIL.Image.fromarray', 'Image.fromarray', (['im'], {}), '(im)\n', (2467, 2471), False, 'from PIL import Image, ImageTk, ImageFont, ImageDraw\n'), ((2483, 2505), 'PIL.ImageDraw.Draw', 'ImageDraw.Draw', (['im_pil'], {}), '(im_pil)\n', (2497, 2505), False, 'from PIL import Image, ImageTk, ImageFont, ImageD... |
__author__ = 'dimd'
from twisted.python import log
from zope.interface import implementer
from NetCatKS.Logger.api.interfaces import ILogger
GLOBAL_DEBUG = True
@implementer(ILogger)
class Logger(object):
def __init__(self):
pass
def debug(self, msg):
if GLOBAL_DEBUG is True:
... | [
"zope.interface.implementer"
] | [((167, 187), 'zope.interface.implementer', 'implementer', (['ILogger'], {}), '(ILogger)\n', (178, 187), False, 'from zope.interface import implementer\n')] |
from plenum.test.bls.helper import check_bls_multi_sig_after_send
from plenum.test.pool_transactions.conftest import looper, clientAndWallet1, \
client1, wallet1, client1Connected
nodeCount = 4
nodes_wth_bls = 0
def test_each_node_has_bls(txnPoolNodeSet):
for node in txnPoolNodeSet:
assert node.bls_b... | [
"plenum.test.bls.helper.check_bls_multi_sig_after_send"
] | [((491, 593), 'plenum.test.bls.helper.check_bls_multi_sig_after_send', 'check_bls_multi_sig_after_send', (['looper', 'txnPoolNodeSet', 'client1', 'wallet1'], {'saved_multi_sigs_count': '(0)'}), '(looper, txnPoolNodeSet, client1, wallet1,\n saved_multi_sigs_count=0)\n', (521, 593), False, 'from plenum.test.bls.helper... |
from __future__ import division, print_function
from .. import __version__
from ._global_imports import *
try:
import h5py
except ImportError:
print('Install h5py to enable signal caching.')
raise
class _Cache(object):
""" Cache numerical model objects computed during likelihood evaluation.
:pa... | [
"datetime.datetime.now",
"h5py.File"
] | [((1493, 1520), 'h5py.File', 'h5py.File', (['self._path', 'mode'], {}), '(self._path, mode)\n', (1502, 1520), False, 'import h5py\n'), ((6214, 6228), 'datetime.datetime.now', 'datetime.now', ([], {}), '()\n', (6226, 6228), False, 'from datetime import datetime\n')] |
from multiprocessing import Process, Manager
''' Custom Module Imports '''
from calculator.add import addition
from calculator.subtract import subtraction
from calculator.multiply import multiplication
from calculator.divide import division
class Main:
def __init__(self) -> None:
pass
d... | [
"calculator.subtract.subtraction",
"calculator.multiply.multiplication",
"multiprocessing.Process",
"calculator.divide.division",
"calculator.add.addition",
"multiprocessing.Manager"
] | [((397, 406), 'multiprocessing.Manager', 'Manager', ([], {}), '()\n', (404, 406), False, 'from multiprocessing import Process, Manager\n'), ((506, 534), 'calculator.add.addition', 'addition', (['i[0]', 'i[1]', 'result'], {}), '(i[0], i[1], result)\n', (514, 534), False, 'from calculator.add import addition\n'), ((558, ... |
"""minor fixes
Revision ID: <KEY>
Revises: 0<PASSWORD>
Create Date: 2020-09-18 07:56:14.159782
"""
from alembic import op
import geoalchemy2
import sqlalchemy as sa
import backend
# revision identifiers, used by Alembic.
revision = '<KEY>'
down_revision = '0fed690a57ce'
branch_labels = ()
depends_on = None
def up... | [
"alembic.op.drop_table",
"sqlalchemy.VARCHAR",
"alembic.op.f",
"sqlalchemy.PrimaryKeyConstraint",
"sqlalchemy.INTEGER",
"sqlalchemy.CheckConstraint",
"sqlalchemy.BigInteger"
] | [((1714, 1746), 'alembic.op.drop_table', 'op.drop_table', (['"""task_harvesting"""'], {}), "('task_harvesting')\n", (1727, 1746), False, 'from alembic import op\n'), ((1549, 1642), 'sqlalchemy.CheckConstraint', 'sa.CheckConstraint', (['"""(srid > 0) AND (srid <= 998999)"""'], {'name': '"""spatial_ref_sys_srid_check"""'... |
"""
Tests for Day 22
"""
from day22.module import part_1, part_2, \
FULL_INPUT_FILE, TEST_INPUT_FILE_1, TEST_INPUT_FILE_2, TEST_INPUT_FILE_3
def test_part_1_1():
result = part_1(TEST_INPUT_FILE_1)
assert result == 39
def test_part_1_2():
result = part_1(TEST_INPUT_FILE_2)
assert result == 59078... | [
"day22.module.part_2",
"day22.module.part_1"
] | [((182, 207), 'day22.module.part_1', 'part_1', (['TEST_INPUT_FILE_1'], {}), '(TEST_INPUT_FILE_1)\n', (188, 207), False, 'from day22.module import part_1, part_2, FULL_INPUT_FILE, TEST_INPUT_FILE_1, TEST_INPUT_FILE_2, TEST_INPUT_FILE_3\n'), ((268, 293), 'day22.module.part_1', 'part_1', (['TEST_INPUT_FILE_2'], {}), '(TES... |
# -*- coding: utf-8 -*-
# DO NOT EDIT THIS FILE!
# This file has been autogenerated by dephell <3
# https://github.com/dephell/dephell
try:
from setuptools import setup
except ImportError:
from distutils.core import setup
import os.path
readme = ""
here = os.path.abspath(os.path.dirname(__file__))
readme_p... | [
"distutils.core.setup"
] | [((483, 2453), 'distutils.core.setup', 'setup', ([], {'long_description': 'readme', 'name': '"""lightbus"""', 'version': '"""1.1.0"""', 'description': '"""RPC & event framework for Python 3"""', 'python_requires': '""">=3.7"""', 'project_urls': "{'documentation': 'https://lightbus.org', 'homepage':\n 'https://lightb... |
import ENVIRONMENT
from src.database.database_creation import createPlayerTrueSkillDictionary
from src.skill_algorithms.algorithms import trueSkillMatchWithRawNums, trueSkillTipWinProb
from src.skill_algorithms.common_data_processing import beforeMatchPredictions, runAlgoForSeason, runAlgoForAllSeasons
# backlogtodo... | [
"src.skill_algorithms.common_data_processing.beforeMatchPredictions",
"src.database.database_creation.createPlayerTrueSkillDictionary",
"src.skill_algorithms.algorithms.trueSkillMatchWithRawNums",
"src.skill_algorithms.common_data_processing.runAlgoForSeason",
"src.skill_algorithms.common_data_processing.ru... | [((568, 1049), 'src.skill_algorithms.common_data_processing.runAlgoForSeason', 'runAlgoForSeason', (['seasonCsv', 'ENVIRONMENT.PLAYER_TRUESKILL_DICT_PATH', 'ENVIRONMENT.TS_PREDICTION_SUMMARIES_PATH', 'trueSkillBeforeMatchPredictions', 'trueSkillUpdateDataSingleTipoff', 'winningBetThreshold'], {'columnAdds': "['Home TS ... |
__author__ = 'multiangle'
# 这是实现 霍夫曼树相关的文件, 主要用于 针对层次softmax进行 word2vec 优化方案的一种
'''
至于 为什么要进行层次softmax 可以简单理解 因为词表很大 针对上完个类别单词进行softmax 计算量大 更新参数过多 无法训练,而采用softmax 层次化 只需要 计算几个有限单词的sigmod 就可以 更新参数也非常少
提高训练速度
什么是霍夫曼树 简单理解就是 将训练文本 进行词频统计 通过构建加权最短路径来构造二叉树 这样 词频高的 位置在前 词频低的位置在后 每一个 霍夫曼编码代表一个词 路径 并且是唯一 不是其他词的前缀
'''
impo... | [
"numpy.zeros"
] | [((4947, 4974), 'numpy.zeros', 'np.zeros', (['[1, self.vec_len]'], {}), '([1, self.vec_len])\n', (4955, 4974), True, 'import numpy as np\n')] |
__all__ = [
"fit_gp",
"ft_gp",
"fit_lm",
"ft_lm",
"fit_rf",
"ft_rf",
"fit_kmeans",
"ft_kmeans",
]
## Fitting via sklearn package
try:
from sklearn.base import clone
from sklearn.linear_model import LinearRegression
from sklearn.gaussian_process import GaussianProcessRegresso... | [
"grama.add_pipe",
"sklearn.cluster.KMeans",
"sklearn.ensemble.RandomForestRegressor",
"grama.Model",
"sklearn.linear_model.LinearRegression",
"grama.cp_vec_function",
"copy.deepcopy",
"pandas.DataFrame",
"warnings.filterwarnings"
] | [((7062, 7078), 'grama.add_pipe', 'add_pipe', (['fit_gp'], {}), '(fit_gp)\n', (7070, 7078), False, 'from grama import add_pipe, pipe\n'), ((9811, 9827), 'grama.add_pipe', 'add_pipe', (['fit_rf'], {}), '(fit_rf)\n', (9819, 9827), False, 'from grama import add_pipe, pipe\n'), ((12030, 12046), 'grama.add_pipe', 'add_pipe'... |
from django.conf import settings
from django.test import TestCase
# Create your tests here.
class SettingsTest(TestCase):
def test_settings(self):
self.assertEqual(settings.SERVER, 'prodserver.com')
self.assertEqual(
settings.STATIC_URL, '/changed/in/settings.toml/by/dynaconf/')
... | [
"django.conf.settings.get",
"django.conf.settings.using_env"
] | [((466, 490), 'django.conf.settings.get', 'settings.get', (['"""PASSWORD"""'], {}), "('PASSWORD')\n", (478, 490), False, 'from django.conf import settings\n'), ((591, 624), 'django.conf.settings.using_env', 'settings.using_env', (['"""development"""'], {}), "('development')\n", (609, 624), False, 'from django.conf impo... |
# coding=utf-8
# Copyright 2020 The Google Research Authors.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicab... | [
"wget.download",
"shutil.copyfileobj",
"pandas.read_csv",
"gzip.open",
"sklearn.model_selection.train_test_split",
"os.mkdir"
] | [((880, 898), 'os.mkdir', 'os.mkdir', (['"""./data"""'], {}), "('./data')\n", (888, 898), False, 'import os\n'), ((910, 928), 'wget.download', 'wget.download', (['url'], {}), '(url)\n', (923, 928), False, 'import wget\n'), ((1059, 1090), 'pandas.read_csv', 'pd.read_csv', (['"""data/covtype.csv"""'], {}), "('data/covtyp... |
import pytest
import datetime
from django.test import TestCase
from django.utils import timezone
from gamer_registration_system.con.models import Convention, Event, EventSchedule
# Create your tests here.
class EventScheduleModelTests(TestCase):
new_con = Convention(convention_name='Test Future Con')
new_eve... | [
"gamer_registration_system.con.models.Event",
"gamer_registration_system.con.models.EventSchedule",
"django.utils.timezone.now",
"gamer_registration_system.con.models.Convention",
"datetime.timedelta"
] | [((263, 308), 'gamer_registration_system.con.models.Convention', 'Convention', ([], {'convention_name': '"""Test Future Con"""'}), "(convention_name='Test Future Con')\n", (273, 308), False, 'from gamer_registration_system.con.models import Convention, Event, EventSchedule\n'), ((325, 377), 'gamer_registration_system.c... |
#!/usr/bin/env python
import rospy
import tf
from jsk_recognition_msgs.msg import BoundingBox
class BoundingBoxToTf(object):
def __init__(self):
self.tf_frame = rospy.get_param('~tf_frame', 'bounding_box')
self.broadcaster = tf.TransformBroadcaster()
self.sub = rospy.Subscriber('~inpu... | [
"tf.TransformBroadcaster",
"rospy.init_node",
"rospy.get_param",
"rospy.Time.now",
"rospy.spin",
"rospy.Subscriber"
] | [((783, 821), 'rospy.init_node', 'rospy.init_node', (['""" bounding_box_to_tf"""'], {}), "(' bounding_box_to_tf')\n", (798, 821), False, 'import rospy\n'), ((854, 866), 'rospy.spin', 'rospy.spin', ([], {}), '()\n', (864, 866), False, 'import rospy\n'), ((178, 222), 'rospy.get_param', 'rospy.get_param', (['"""~tf_frame"... |
from defs_utils import *
import pdf2image
import os
def transforma_pdf_em_img_por_materia(materia, pdf_path=None):
searched = materia
if pdf_path:
list_files = list_dir(complete_name(searched, pre=pdf_path), True)
else:
list_files = list_dir(complete_name(searched), True)
volta = os.ge... | [
"os.getcwd",
"os.chdir",
"os.path.realpath",
"os.mkdir",
"pdf2image.convert_from_path"
] | [((315, 326), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (324, 326), False, 'import os\n'), ((372, 405), 'pdf2image.convert_from_path', 'pdf2image.convert_from_path', (['file'], {}), '(file)\n', (399, 405), False, 'import pdf2image\n'), ((434, 449), 'os.chdir', 'os.chdir', (['volta'], {}), '(volta)\n', (442, 449), Fal... |
# Copyright (c) 2017, Carnegie Mellon University. All rights reserved.
#
# Use of the K-NRM package is subject to the terms of the software license set
# forth in the LICENSE file included with this software, and also available at
# https://github.com/AdeDZY/K-NRM/blob/master/LICENSE
from setuptools import setup
from ... | [
"setuptools.find_packages"
] | [((528, 543), 'setuptools.find_packages', 'find_packages', ([], {}), '()\n', (541, 543), False, 'from setuptools import find_packages\n')] |
#Imports
import config
import telebot
from db_manager import SQL
import os
#Globals
restaurants = {"12345":"abc"}
current_restaurant_id = ""
current_restaurant_key = ""
couriers = {"12345":"abc"}
current_courier_id = ""
current_courier_key = ""
current_courier_altitude = 0
current_courier_longitude = 0
#DB + BOT CO... | [
"os.path.join",
"db_manager.SQL",
"telebot.types.ReplyKeyboardMarkup",
"os.path.abspath",
"telebot.TeleBot"
] | [((358, 380), 'telebot.TeleBot', 'telebot.TeleBot', (['token'], {}), '(token)\n', (373, 380), False, 'import telebot\n'), ((445, 482), 'os.path.join', 'os.path.join', (['BASE_DIR', '"""database.db"""'], {}), "(BASE_DIR, 'database.db')\n", (457, 482), False, 'import os\n'), ((488, 500), 'db_manager.SQL', 'SQL', (['db_pa... |
# -*- coding: utf-8 -*-
# Generated by Django 1.11.5 on 2017-09-10 20:53
from __future__ import unicode_literals
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
dependencies = [
("bdiadmin", "0013_auto_20170319_1415"),
("stock", "0... | [
"django.db.models.FloatField",
"django.db.models.ForeignKey",
"django.db.models.AutoField",
"django.db.models.PositiveIntegerField",
"django.db.models.CharField"
] | [((523, 616), 'django.db.models.AutoField', 'models.AutoField', ([], {'auto_created': '(True)', 'primary_key': '(True)', 'serialize': '(False)', 'verbose_name': '"""ID"""'}), "(auto_created=True, primary_key=True, serialize=False,\n verbose_name='ID')\n", (539, 616), False, 'from django.db import migrations, models\... |
# -*- coding: UTF-8 -*-
"""
Source processing routines
"""
from __future__ import absolute_import
from __future__ import unicode_literals
import warnings
from collections import OrderedDict
from astropy.cosmology import default_cosmology
import numpy as np
import os
import pysynphot as S
import astropy.table as at
fro... | [
"os.path.exists",
"collections.OrderedDict",
"numpy.isscalar",
"os.path.join",
"pysynphot.ArraySpectrum",
"numpy.array",
"astropy.cosmology.default_cosmology.get",
"warnings.warn",
"astropy.table.Table.read"
] | [((732, 745), 'collections.OrderedDict', 'OrderedDict', ([], {}), '()\n', (743, 745), False, 'from collections import OrderedDict\n'), ((753, 777), 'numpy.isscalar', 'np.isscalar', (['sourcenames'], {}), '(sourcenames)\n', (764, 777), True, 'import numpy as np\n'), ((2274, 2316), 'os.path.join', 'os.path.join', (['"""s... |
import pandas as pd
import numpy as np
import pickle
from sklearn.metrics.pairwise import linear_kernel
from sklearn.feature_extraction.text import TfidfVectorizer
import csv
from kivy.app import App
from kivy.uix.gridlayout import GridLayout
from kivy.uix.label import Label
from kivy.uix.textinput import TextInput
... | [
"kivy.uix.label.Label",
"numpy.ndarray.argsort",
"kivy.uix.textinput.TextInput",
"csv.reader"
] | [((1260, 1286), 'kivy.uix.textinput.TextInput', 'TextInput', ([], {'multiline': '(False)'}), '(multiline=False)\n', (1269, 1286), False, 'from kivy.uix.textinput import TextInput\n'), ((1398, 1439), 'kivy.uix.textinput.TextInput', 'TextInput', ([], {'password': '(True)', 'multiline': '(False)'}), '(password=True, multi... |
from django import template
register = template.Library()
@register.inclusion_tag('quiz/correct_answer.html', takes_context=True)
def correct_answer_for_all(context, question):
"""
processes the correct answer based on a given question object
if the answer is incorrect, informs the user
"""
answe... | [
"django.template.Library"
] | [((40, 58), 'django.template.Library', 'template.Library', ([], {}), '()\n', (56, 58), False, 'from django import template\n')] |
# -*- coding: utf-8 -*-
"""
Created on Mon Jun 29 14:14:45 2020
@author: Nikki
"""
import numpy as np
import cv2
import transform as tform
import sys
import math
import scipy.spatial
import markers
###---------------------------------------------------------------------------
# Allows video to be initialized usin... | [
"math.sqrt",
"cv2.imshow",
"math.cos",
"numpy.array",
"cv2.ellipse",
"sys.exc_info",
"cv2.destroyAllWindows",
"cv2.setMouseCallback",
"markers.aot_1_markers",
"numpy.asarray",
"markers.mrb3_markers",
"cv2.addWeighted",
"markers.aot_3_markers",
"cv2.waitKey",
"transform.get_best_transform... | [((1885, 1915), 'transform.get_best_transform', 'tform.get_best_transform', (['x', 'y'], {}), '(x, y)\n', (1909, 1915), True, 'import transform as tform\n'), ((1930, 1960), 'transform.get_best_transform', 'tform.get_best_transform', (['y', 'x'], {}), '(y, x)\n', (1954, 1960), True, 'import transform as tform\n'), ((300... |
# mp4.py
# ---------------
# Licensing Information: You are free to use or extend this projects for
# educational purposes provided that (1) you do not distribute or publish
# solutions, (2) you retain this notice, and (3) you provide clear
# attribution to the University of Illinois at Urbana-Champaign
#
# Created Fa... | [
"collections.Counter"
] | [((1845, 1866), 'collections.Counter', 'collections.Counter', ([], {}), '()\n', (1864, 1866), False, 'import collections\n'), ((2090, 2111), 'collections.Counter', 'collections.Counter', ([], {}), '()\n', (2109, 2111), False, 'import collections\n')] |
from ex2.sigmoid import sigmoid
def sigmoidGradient(z):
"""computes the gradient of the sigmoid function
evaluated at z. This should work regardless if z is a matrix or a
vector. In particular, if z is a vector or matrix, you should return
the gradient for each element."""
# =====================... | [
"ex2.sigmoid.sigmoid"
] | [((521, 531), 'ex2.sigmoid.sigmoid', 'sigmoid', (['z'], {}), '(z)\n', (528, 531), False, 'from ex2.sigmoid import sigmoid\n'), ((539, 549), 'ex2.sigmoid.sigmoid', 'sigmoid', (['z'], {}), '(z)\n', (546, 549), False, 'from ex2.sigmoid import sigmoid\n')] |
# -*- coding: utf-8 -*-
from __future__ import (print_function, unicode_literals, absolute_import,
division)
from pusher.config import Config
from pusher.request import Request
from pusher.sync import SynchronousBackend
from pusher.util import GET, POST, text, validate_channel
import collectio... | [
"json.dumps",
"pusher.util.validate_channel",
"pusher.sync.SynchronousBackend",
"six.text_type",
"pusher.request.Request"
] | [((3131, 3205), 'pusher.request.Request', 'Request', (['self.config', 'POST', "('/apps/%s/events' % self.config.app_id)", 'params'], {}), "(self.config, POST, '/apps/%s/events' % self.config.app_id, params)\n", (3138, 3205), False, 'from pusher.request import Request\n'), ((3623, 3698), 'pusher.request.Request', 'Reque... |
"""
To run this module directly
"""
# pylint: disable=no-else-return
import argparse
import os
try:
from . import prefetch_from_file
except ImportError:
import prefetch_from_file
try:
from . import prefetch_from_url
except ImportError:
import prefetch_from_url
def validate_filepath_or_url(filepath_o... | [
"argparse.ArgumentParser",
"os.access",
"os.path.isfile",
"prefetch_from_url.url_to_prefetch",
"prefetch_from_file.file_to_prefetch"
] | [((678, 781), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '__doc__', 'formatter_class': 'argparse.RawDescriptionHelpFormatter'}), '(description=__doc__, formatter_class=argparse.\n RawDescriptionHelpFormatter)\n', (701, 781), False, 'import argparse\n'), ((1642, 1718), 'prefetch_from_f... |
import numpy as np
from src import const
#TODO: should be imported from aguirregabiria_simple.py
def period_profit(p: np.ndarray, lambdas: np.ndarray, betas_transition=const.betas_transition):
"""
Correct expected period return profit. See ReadMe for derivation
"""
constant_part = (p-const.c) * np.e *... | [
"numpy.log",
"numpy.array",
"numpy.allclose"
] | [((518, 538), 'numpy.array', 'np.array', (['[1.4, 1.2]'], {}), '([1.4, 1.2])\n', (526, 538), True, 'import numpy as np\n'), ((553, 578), 'numpy.array', 'np.array', (['[0.5, 0.4, 0.1]'], {}), '([0.5, 0.4, 0.1])\n', (561, 578), True, 'import numpy as np\n'), ((598, 791), 'numpy.array', 'np.array', (['[[np.e ** (-3.0 * 0.... |
import mock
from pyramid.request import Request
from .support import DummyRequest, unittest
from kinto.core import authentication
from kinto.core.authorization import RouteFactory, AuthorizationPolicy
from kinto.core.storage import exceptions as storage_exceptions
class RouteFactoryTest(unittest.TestCase):
def... | [
"mock.patch",
"mock.Mock",
"pyramid.request.Request.blank",
"kinto.core.authorization.RouteFactory",
"kinto.core.authorization.AuthorizationPolicy",
"mock.MagicMock"
] | [((3505, 3528), 'pyramid.request.Request.blank', 'Request.blank', ([], {'path': '"""/"""'}), "(path='/')\n", (3518, 3528), False, 'from pyramid.request import Request\n'), ((3556, 3578), 'mock.Mock', 'mock.Mock', ([], {'settings': '{}'}), '(settings={})\n', (3565, 3578), False, 'import mock\n'), ((3707, 3728), 'kinto.c... |
import copy
#from enum import IntFlag
from time import sleep
# I tried to use enum here, but I was having a problem with packages in the image, so I gave up as I just want to get it done
class FieldValue:
Empty = 0
Wall = 1
Player = 2
Box = 4
Goal = 8
class SenseHATColour:
Red = (204, 4, 4)
... | [
"sense_hat.SenseHat",
"time.sleep",
"copy.deepcopy"
] | [((1862, 1872), 'sense_hat.SenseHat', 'SenseHat', ([], {}), '()\n', (1870, 1872), False, 'from sense_hat import SenseHat\n'), ((4169, 4189), 'copy.deepcopy', 'copy.deepcopy', (['level'], {}), '(level)\n', (4182, 4189), False, 'import copy\n'), ((6258, 6269), 'time.sleep', 'sleep', (['(0.25)'], {}), '(0.25)\n', (6263, 6... |
# coding: utf-8
"""
NiFi Rest Api
The Rest Api provides programmatic access to command and control a NiFi instance in real time. Start and stop processors, monitor queues, query provenance data, and more. Each endpoint below includes a description, ... | [
"six.iteritems"
] | [((3606, 3633), 'six.iteritems', 'iteritems', (["params['kwargs']"], {}), "(params['kwargs'])\n", (3615, 3633), False, 'from six import iteritems\n'), ((8990, 9017), 'six.iteritems', 'iteritems', (["params['kwargs']"], {}), "(params['kwargs'])\n", (8999, 9017), False, 'from six import iteritems\n'), ((15208, 15235), 's... |
#!/usr/bin/env python
"""
Usage: python vtk_reindent_code.py [--test] <file1> [<file2> ...]
This script takes old-style "Whitesmiths" indented VTK source files as
input, and re-indents the braces according to the new VTK style.
Only the brace indentation is modified.
If called with the --test option, then it ... | [
"re.compile",
"os.path.split",
"sys.stderr.write",
"sys.exc_info",
"sys.exit"
] | [((882, 902), 're.compile', 're.compile', (['"""[/"\']"""'], {}), '(\'[/"\\\']\')\n', (892, 902), False, 'import re\n'), ((958, 1002), 're.compile', 're.compile', (['"""\\\\/\\\\*(\\\\*(?!\\\\/)|[^*])*\\\\*\\\\/"""'], {}), "('\\\\/\\\\*(\\\\*(?!\\\\/)|[^*])*\\\\*\\\\/')\n", (968, 1002), False, 'import re\n'), ((1021, 1... |
import os.path
from scipy.optimize import fsolve
import math
import numpy as np
from matplotlib import pyplot as plt
import pandas as pd
import utils_Florian as utils
def equations(p, t_peak, t_half):
x, y = p
return (0.5 * (math.exp(-x * t_peak) - math.exp(-y * t_peak)) - (math.exp(-x * t_half) - math.exp... | [
"scipy.optimize.fsolve",
"numpy.abs",
"numpy.argmax",
"numpy.exp",
"numpy.linspace",
"numpy.argmin",
"pandas.DataFrame",
"math.exp"
] | [((404, 418), 'pandas.DataFrame', 'pd.DataFrame', ([], {}), '()\n', (416, 418), True, 'import pandas as pd\n'), ((1088, 1155), 'scipy.optimize.fsolve', 'fsolve', (['equations', 'initial_conditions[alpha]'], {'args': '(t_peak, t_half)'}), '(equations, initial_conditions[alpha], args=(t_peak, t_half))\n', (1094, 1155), F... |
import re
import string
import numpy as np
from tqdm import tqdm
from typing import List
from docqa.triviaqa.read_data import TriviaQaQuestion
from docqa.triviaqa.trivia_qa_eval import normalize_answer, f1_score
from docqa.utils import flatten_iterable, split
"""
Tools for turning the aliases and answer strings from... | [
"docqa.triviaqa.trivia_qa_eval.normalize_answer",
"docqa.utils.flatten_iterable",
"re.compile",
"tqdm.tqdm",
"trivia_qa.build_span_corpus.TriviaQaWebDataset",
"data_processing.text_utils.NltkAndPunctTokenizer",
"numpy.array",
"numpy.zeros",
"multiprocessing.Pool",
"docqa.triviaqa.trivia_qa_eval.f1... | [((10899, 10919), 'trivia_qa.build_span_corpus.TriviaQaWebDataset', 'TriviaQaWebDataset', ([], {}), '()\n', (10917, 10919), False, 'from trivia_qa.build_span_corpus import TriviaQaWebDataset\n'), ((5407, 5429), 'docqa.utils.flatten_iterable', 'flatten_iterable', (['para'], {}), '(para)\n', (5423, 5429), False, 'from do... |
# Generated by Django 3.2.9 on 2021-11-13 14:43
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('base', '0005_authquery_password'),
]
operations = [
migrations.AddField(
model_name='profile',
name='h... | [
"django.db.models.CharField"
] | [((348, 395), 'django.db.models.CharField', 'models.CharField', ([], {'default': '"""[]"""', 'max_length': '(1000)'}), "(default='[]', max_length=1000)\n", (364, 395), False, 'from django.db import migrations, models\n')] |
import nifty.tools as nt
import numpy as np
import z5py
from elf.label_multiset import deserialize_multiset
from tqdm import trange
def check_serialization(mset1, mset2):
if len(mset1) != len(mset2):
print("Serialization sizes disagree:", len(mset1), len(mset2))
return False
if not np.array_... | [
"numpy.unique",
"elf.label_multiset.deserialize_multiset",
"nifty.tools.blocking",
"z5py.File",
"numpy.array_equal",
"tqdm.trange"
] | [((1798, 1840), 'nifty.tools.blocking', 'nt.blocking', (['[0, 0, 0]', 'roi_end', '[1, 1, 1]'], {}), '([0, 0, 0], roi_end, [1, 1, 1])\n', (1809, 1840), True, 'import nifty.tools as nt\n'), ((1861, 1892), 'tqdm.trange', 'trange', (['blocking.numberOfBlocks'], {}), '(blocking.numberOfBlocks)\n', (1867, 1892), False, 'from... |
import torch
import torch.nn as nn
import torch.optim as optim
from torch.utils.data import DataLoader
from torchvision.transforms import ToTensor
from torchvision.datasets import MNIST
from vis_util import visual_mnist
##### settings
x_dim = 28 * 28 # size of mnist digit
z_dim = 100 # random noise
h_dim = 128 ... | [
"torch.nn.Sigmoid",
"torch.nn.ReLU",
"torch.randn",
"torch.nn.BCELoss",
"torch.nn.Linear",
"torch.utils.data.DataLoader",
"torchvision.transforms.ToTensor",
"torch.zeros",
"torch.ones"
] | [((548, 605), 'torch.utils.data.DataLoader', 'DataLoader', (['trainset'], {'batch_size': 'batch_size', 'shuffle': '(True)'}), '(trainset, batch_size=batch_size, shuffle=True)\n', (558, 605), False, 'from torch.utils.data import DataLoader\n'), ((686, 711), 'torch.ones', 'torch.ones', (['batch_size', '(1)'], {}), '(batc... |
from django.shortcuts import render, redirect, reverse
from django.core.paginator import Paginator, PageNotAnInteger, EmptyPage
from django.db.models import Q
from django.views.generic import View
from .models import OrgInfo, CityInfo, TeacherInfo
from operations.models import UserLove
# Create your views here.
clas... | [
"django.shortcuts.render",
"django.db.models.Q",
"django.core.paginator.Paginator"
] | [((1336, 1358), 'django.core.paginator.Paginator', 'Paginator', (['all_orgs', '(2)'], {}), '(all_orgs, 2)\n', (1345, 1358), False, 'from django.core.paginator import Paginator, PageNotAnInteger, EmptyPage\n'), ((1554, 1763), 'django.shortcuts.render', 'render', (['request', '"""orgs/org-list.html"""', "{'all_orgs': all... |
from keras.callbacks import TensorBoard,EarlyStopping,TerminateOnNaN,ReduceLROnPlateau,ModelCheckpoint
import os
import sys
import tensorflow as tf
import keras.backend.tensorflow_backend as KTF
file_abspath = os.path.abspath(sys.argv[0]) # exe所在文件地址
location = os.path.dirname(file_abspath) # exe所在文件夹目录地址
tbCallB... | [
"keras.callbacks.ModelCheckpoint",
"keras.callbacks.ReduceLROnPlateau",
"keras.callbacks.TerminateOnNaN",
"keras.callbacks.TensorBoard",
"os.path.dirname",
"keras.callbacks.EarlyStopping",
"os.path.abspath"
] | [((213, 241), 'os.path.abspath', 'os.path.abspath', (['sys.argv[0]'], {}), '(sys.argv[0])\n', (228, 241), False, 'import os\n'), ((266, 295), 'os.path.dirname', 'os.path.dirname', (['file_abspath'], {}), '(file_abspath)\n', (281, 295), False, 'import os\n'), ((327, 575), 'keras.callbacks.TensorBoard', 'TensorBoard', ([... |
# Generated by Django 3.1.1 on 2020-12-26 20:50
from django.db import migrations
class Migration(migrations.Migration):
dependencies = [
('dashboard', '0033_auto_20201226_2148'),
]
operations = [
migrations.RenameField(
model_name='sleepnight',
old_name='data',
... | [
"django.db.migrations.RenameField"
] | [((228, 319), 'django.db.migrations.RenameField', 'migrations.RenameField', ([], {'model_name': '"""sleepnight"""', 'old_name': '"""data"""', 'new_name': '"""diary_day"""'}), "(model_name='sleepnight', old_name='data', new_name=\n 'diary_day')\n", (250, 319), False, 'from django.db import migrations\n')] |
"""Fear and Greed Index Class"""
__docformat__ = "numpy"
from matplotlib import pyplot as plt
from fear_greed_index import scrape_cnn
from fear_greed_index.FearAndGreedIndicator import FearAndGreedIndicator
class CNNFearAndGreedIndex:
"""CNN Fear and Greed Index
Attributes
----------
junk_bond_deman... | [
"matplotlib.pyplot.imshow",
"fear_greed_index.FearAndGreedIndicator.FearAndGreedIndicator",
"fear_greed_index.scrape_cnn._get_chart",
"matplotlib.pyplot.tight_layout",
"fear_greed_index.scrape_cnn._get_fear_greed_index"
] | [((1459, 1500), 'fear_greed_index.FearAndGreedIndicator.FearAndGreedIndicator', 'FearAndGreedIndicator', (['"""Junk Bond Demand"""'], {}), "('Junk Bond Demand')\n", (1480, 1500), False, 'from fear_greed_index.FearAndGreedIndicator import FearAndGreedIndicator\n'), ((1534, 1576), 'fear_greed_index.FearAndGreedIndicator.... |
# -*- coding: utf-8 -*-
"""
idfy_rest_client.models.person_person_information
This file was automatically generated for Idfy by APIMATIC v2.0 ( https://apimatic.io )
"""
from idfy_rest_client.api_helper import APIHelper
class PersonPersonInformation(object):
"""Implementation of the 'Person.P... | [
"idfy_rest_client.api_helper.APIHelper.RFC3339DateTime"
] | [((2794, 2825), 'idfy_rest_client.api_helper.APIHelper.RFC3339DateTime', 'APIHelper.RFC3339DateTime', (['dead'], {}), '(dead)\n', (2819, 2825), False, 'from idfy_rest_client.api_helper import APIHelper\n')] |
import logging
import subprocess
# create logger
logger = logging.getLogger('simple_example')
logger.setLevel(logging.DEBUG)
# create console handler and set level to INFO
console_logger = logging.StreamHandler()
console_logger.setLevel(logging.INFO)
# create formatter
formatter = logging.Formatter('%(asctime)s - %(... | [
"logging.getLogger",
"logging.Formatter",
"logging.StreamHandler",
"subprocess.call"
] | [((59, 94), 'logging.getLogger', 'logging.getLogger', (['"""simple_example"""'], {}), "('simple_example')\n", (76, 94), False, 'import logging\n'), ((191, 214), 'logging.StreamHandler', 'logging.StreamHandler', ([], {}), '()\n', (212, 214), False, 'import logging\n'), ((285, 347), 'logging.Formatter', 'logging.Formatte... |
import numpy as np
from pandas import read_csv
import pandas as pd
from sklearn.model_selection import train_test_split
import matplotlib.pyplot as plt
from config import *
from lib.preprocess.read_data import DataReader
from lib.scaler.preprocessing_data.data_normalizer import DataNormalizer
class DataPreprocessor:... | [
"lib.scaler.preprocessing_data.data_normalizer.DataNormalizer",
"lib.preprocess.read_data.DataReader",
"numpy.column_stack",
"numpy.array",
"numpy.concatenate"
] | [((717, 729), 'lib.preprocess.read_data.DataReader', 'DataReader', ([], {}), '()\n', (727, 729), False, 'from lib.preprocess.read_data import DataReader\n'), ((2432, 2461), 'lib.scaler.preprocessing_data.data_normalizer.DataNormalizer', 'DataNormalizer', (['scaler_method'], {}), '(scaler_method)\n', (2446, 2461), False... |
# pylint: disable=arguments-differ,unused-argument,line-too-long
"""R2Plus1D, implemented in Gluon. https://arxiv.org/abs/1711.11248.
Code partially borrowed from https://github.com/pytorch/vision/blob/master/torchvision/models/video/resnet.py."""
__all__ = ['R2Plus1D', 'r2plus1d_resnet18_kinetics400',
'r2... | [
"mxnet.gluon.nn.GlobalAvgPool3D",
"mxnet.gluon.nn.HybridSequential",
"mxnet.init.Normal",
"mxnet.gluon.nn.Dropout",
"mxnet.init.MSRAPrelu",
"mxnet.context.cpu",
"mxnet.gluon.nn.Conv3D",
"mxnet.gluon.nn.Activation"
] | [((762, 967), 'mxnet.gluon.nn.Conv3D', 'nn.Conv3D', ([], {'in_channels': 'in_planes', 'channels': 'out_planes', 'kernel_size': '(3, 1, 1)', 'strides': '(temporal_stride, spatial_stride, spatial_stride)', 'padding': '(dilation, 0, 0)', 'dilation': 'dilation', 'use_bias': '(False)'}), '(in_channels=in_planes, channels=ou... |
#!/usr/bin/env python3
import retrieve_author_ppn as autppn
import retrieve_references as refs
import zot_helpers as pyzot
from itertools import islice
researchers = autppn.constructOutput('test.csv')
autppn.writeCsv('out.csv', researchers)
for researcher in researchers:
ppn = researcher['ppn']
creator_name... | [
"itertools.islice",
"retrieve_author_ppn.constructOutput",
"retrieve_references.getRefsByRole",
"zot_helpers.create_collection",
"retrieve_author_ppn.writeCsv",
"retrieve_references.getReferences"
] | [((168, 202), 'retrieve_author_ppn.constructOutput', 'autppn.constructOutput', (['"""test.csv"""'], {}), "('test.csv')\n", (190, 202), True, 'import retrieve_author_ppn as autppn\n'), ((204, 243), 'retrieve_author_ppn.writeCsv', 'autppn.writeCsv', (['"""out.csv"""', 'researchers'], {}), "('out.csv', researchers)\n", (2... |
#!/usr/bin/env python
from pkg_resources import get_distribution
from setuptools import setup, find_packages
with open("README.md", "r") as f:
long_description = f.read()
version = get_distribution("autolabel").version
setup(
packages=find_packages(),
install_requires=[
'click',
'more-ite... | [
"setuptools.find_packages",
"pkg_resources.get_distribution"
] | [((187, 216), 'pkg_resources.get_distribution', 'get_distribution', (['"""autolabel"""'], {}), "('autolabel')\n", (203, 216), False, 'from pkg_resources import get_distribution\n'), ((246, 261), 'setuptools.find_packages', 'find_packages', ([], {}), '()\n', (259, 261), False, 'from setuptools import setup, find_package... |
import os
from docker.models.containers import Container
import docker
import pytest
from tox_docker.config import runas_name
def find_container(instance_name: str) -> Container:
# TODO: refactor this as a pytest fixture
# this is running in a child-process of the tox instance which
# spawned the conta... | [
"pytest.fail",
"os.getppid",
"docker.from_env"
] | [((472, 503), 'docker.from_env', 'docker.from_env', ([], {'version': '"""auto"""'}), "(version='auto')\n", (487, 503), False, 'import docker\n'), ((680, 752), 'pytest.fail', 'pytest.fail', (['f"""No running container with instance name {running_name!r}"""'], {}), "(f'No running container with instance name {running_nam... |
#!usr/bin/env python3
# -*- coding: utf-8 -*-
#
# @author <NAME>
#
# get posts of a fb page, group or account
# returns a json lines files with a line for each post
#
from argparse import ArgumentParser
from datetime import date, datetime
from facebook_scraper import get_posts
from json import dumps, JSONEncoder
clas... | [
"datetime.date.today",
"json.dumps",
"facebook_scraper.get_posts",
"argparse.ArgumentParser"
] | [((1456, 1472), 'argparse.ArgumentParser', 'ArgumentParser', ([], {}), '()\n', (1470, 1472), False, 'from argparse import ArgumentParser\n'), ((1101, 1231), 'facebook_scraper.get_posts', 'get_posts', ([], {'group': 'account', 'cookies': 'cookies', 'pages': 'pages', 'extra_info': '(True)', 'options': "{'comments': comme... |
import gi
gi.require_version('Gtk', '3.0')
from gi.repository import Gtk as gtk
class Ked:
def __init__(self, app_path, ufile=None):
glade_layout = f'{app_path}/data/ked_layout.glade'
self.builder = gtk.Builder()
self.builder.add_from_file(glade_layout)
win = self.builder.get_obje... | [
"gi.repository.Gtk.Builder",
"gi.repository.Gtk.main",
"gi.require_version"
] | [((11, 43), 'gi.require_version', 'gi.require_version', (['"""Gtk"""', '"""3.0"""'], {}), "('Gtk', '3.0')\n", (29, 43), False, 'import gi\n'), ((536, 546), 'gi.repository.Gtk.main', 'gtk.main', ([], {}), '()\n', (544, 546), True, 'from gi.repository import Gtk as gtk\n'), ((221, 234), 'gi.repository.Gtk.Builder', 'gtk.... |
#!/usr/bin/env python
# vim: set fileencoding=utf-8 ts=4 sts=4 sw=4 et tw=80 :
#
# Compare an image file and its associated uncertainty image.
#
# <NAME>
# Created: 2021-06-03
# Last modified: 2021-06-03
#--------------------------------------------------------------------------
#*********************************... | [
"logging.basicConfig",
"logging.getLogger",
"numpy.__version__.split",
"numpy.abs",
"os.getenv",
"matplotlib.pyplot.gcf",
"theil_sen.linefit",
"imp.reload",
"sys.stderr.write",
"astropy.io.fits.getdata",
"numpy.linspace",
"matplotlib.pyplot.figure",
"numpy.isnan",
"sys.exit",
"matplotlib... | [((514, 553), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.INFO'}), '(level=logging.INFO)\n', (533, 553), False, 'import logging\n'), ((563, 590), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (580, 590), False, 'import logging\n'), ((8752, 8782), 'astropy.io.fit... |
import logging
import os
try:
import simplejson as json
except ImportError:
import json
from flask import Flask, request, make_response, Response
from cStringIO import StringIO
import zipfile
def get_mocked_server(binary_directory):
mocked_cb_server = Flask('cb')
files = os.listdir(binary_directory... | [
"flask.request.args.get",
"os.path.exists",
"cStringIO.StringIO",
"os.listdir",
"json.loads",
"zipfile.ZipFile",
"flask.Flask",
"json.dumps",
"os.path.join",
"flask.Response",
"os.path.abspath",
"flask.make_response"
] | [((268, 279), 'flask.Flask', 'Flask', (['"""cb"""'], {}), "('cb')\n", (273, 279), False, 'from flask import Flask, request, make_response, Response\n'), ((293, 321), 'os.listdir', 'os.listdir', (['binary_directory'], {}), '(binary_directory)\n', (303, 321), False, 'import os\n'), ((3167, 3215), 'os.path.join', 'os.path... |
from common.parse_tex import (
BeginDocumentExtractor,
BibitemExtractor,
DocumentclassExtractor,
EquationExtractor,
MacroExtractor,
PlaintextExtractor,
)
from common.types import MacroDefinition
from entities.sentences.extractor import SentenceExtractor
def test_extract_plaintext_with_newlines... | [
"common.parse_tex.BeginDocumentExtractor",
"common.parse_tex.DocumentclassExtractor",
"entities.sentences.extractor.SentenceExtractor",
"common.types.MacroDefinition",
"common.parse_tex.EquationExtractor",
"common.parse_tex.BibitemExtractor",
"common.parse_tex.PlaintextExtractor",
"common.parse_tex.Ma... | [((340, 360), 'common.parse_tex.PlaintextExtractor', 'PlaintextExtractor', ([], {}), '()\n', (358, 360), False, 'from common.parse_tex import BeginDocumentExtractor, BibitemExtractor, DocumentclassExtractor, EquationExtractor, MacroExtractor, PlaintextExtractor\n'), ((1023, 1042), 'entities.sentences.extractor.Sentence... |
from webdriver_manager.driver import EdgeDriver, IEDriver
from webdriver_manager.manager import DriverManager
from webdriver_manager import utils
class EdgeDriverManager(DriverManager):
def __init__(self, version=None,
os_type=utils.os_name()):
super(EdgeDriverManager, self).__init__()
... | [
"webdriver_manager.utils.os_type",
"webdriver_manager.utils.os_name",
"webdriver_manager.driver.IEDriver",
"webdriver_manager.driver.EdgeDriver"
] | [((250, 265), 'webdriver_manager.utils.os_name', 'utils.os_name', ([], {}), '()\n', (263, 265), False, 'from webdriver_manager import utils\n'), ((340, 384), 'webdriver_manager.driver.EdgeDriver', 'EdgeDriver', ([], {'version': 'version', 'os_type': 'os_type'}), '(version=version, os_type=os_type)\n', (350, 384), False... |
"""Parent class DataN."""
import os
import os.path
from warnings import warn
from typing import Union, NoReturn
from pycifstar import Data, to_data
from cryspy.A_functions_base.function_1_markdown import md_to_html
from cryspy.A_functions_base.function_1_objects import \
get_functions_of_objet, get_table_html_for_... | [
"pycifstar.to_data",
"os.path.isfile",
"warnings.warn",
"cryspy.A_functions_base.function_1_objects.get_table_html_for_variables",
"cryspy.A_functions_base.function_1_objects.get_functions_of_objet",
"pycifstar.Data"
] | [((13869, 13875), 'pycifstar.Data', 'Data', ([], {}), '()\n', (13873, 13875), False, 'from pycifstar import Data, to_data\n'), ((1332, 1366), 'cryspy.A_functions_base.function_1_objects.get_table_html_for_variables', 'get_table_html_for_variables', (['self'], {}), '(self)\n', (1360, 1366), False, 'from cryspy.A_functio... |
import os
import unittest
def require(f):
def skipit(*args, **kwargs):
raise unittest.SkipTest('VIVBINS env var...')
if os.getenv('VIVBINS') == None:
return skipit
return f
| [
"unittest.SkipTest",
"os.getenv"
] | [((90, 129), 'unittest.SkipTest', 'unittest.SkipTest', (['"""VIVBINS env var..."""'], {}), "('VIVBINS env var...')\n", (107, 129), False, 'import unittest\n'), ((138, 158), 'os.getenv', 'os.getenv', (['"""VIVBINS"""'], {}), "('VIVBINS')\n", (147, 158), False, 'import os\n')] |
from setuptools import setup
from setuptools.command.install import install as _install
class Install(_install):
def run(self):
_install.do_egg_install(self)
import nltk
nltk.download("popular")
setup(
cmdclass={'install': Install},
install_requires=['nltk'],
setup_requires=[... | [
"setuptools.setup",
"nltk.download",
"setuptools.command.install.install.do_egg_install"
] | [((227, 319), 'setuptools.setup', 'setup', ([], {'cmdclass': "{'install': Install}", 'install_requires': "['nltk']", 'setup_requires': "['nltk']"}), "(cmdclass={'install': Install}, install_requires=['nltk'],\n setup_requires=['nltk'])\n", (232, 319), False, 'from setuptools import setup\n'), ((142, 171), 'setuptool... |