code
stringlengths
22
1.05M
apis
listlengths
1
3.31k
extract_api
stringlengths
75
3.25M
from Dialogflow_Api import rispondimi from collegamentoSito import inserisci_utente from Nutrition import get_food, traduzione import re tipo_cibo = ["frutta", "carne", "verdure", "ortaggi", "primi_piatti", "legumi"] """ controllo_intent(query_result, utente)--> text_respose prende il risultato della query e lo confr...
[ "re.split", "collegamentoSito.inserisci_utente", "Nutrition.traduzione", "Nutrition.get_food", "Dialogflow_Api.rispondimi" ]
[((2545, 2569), 'collegamentoSito.inserisci_utente', 'inserisci_utente', (['utente'], {}), '(utente)\n', (2561, 2569), False, 'from collegamentoSito import inserisci_utente\n'), ((4345, 4369), 'collegamentoSito.inserisci_utente', 'inserisci_utente', (['utente'], {}), '(utente)\n', (4361, 4369), False, 'from collegament...
# Copyright (c) 2019, NVIDIA CORPORATION. All rights reserved. # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # http://www.apache.org/licenses/LICENSE-2.0 # Unless required by applicabl...
[ "architectures.Generator", "argparse.ArgumentParser", "kaolin.rep.voxel.max_connected", "kaolin.rep.QuadMesh.from_tensors", "kaolin.conversion.voxel.to_mesh_quad", "torch.zeros", "torch.ones" ]
[((914, 939), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (937, 939), False, 'import argparse\n'), ((1542, 1581), 'kaolin.rep.voxel.max_connected', 'kal.rep.voxel.max_connected', (['model', '(0.5)'], {}), '(model, 0.5)\n', (1569, 1581), True, 'import kaolin as kal\n'), ((1597, 1637), 'kaolin...
"""Defines procedures for training, and evaluation automatic camfi annotation models, and for using them for making automatic annotations (inference). Depends on camfi.util, camfi.datamodel.autoannotation, camfi.datamodel.geometry, camfi.datamode.via, as well as ._torchutils and ._models.""" from datetime import datet...
[ "numpy.count_nonzero", "numpy.array", "camfi.datamodel.via.ViaRegionAttributes", "numpy.arange", "camfi.datamodel.geometry.CircleShapeAttributes", "pathlib.Path", "numpy.where", "camfi.datamodel.autoannotation.Prediction.from_tensor_dict", "torch.hub.load_state_dict_from_url", "scipy.sparse.csgrap...
[((3085, 3091), 'pathlib.Path', 'Path', ([], {}), '()\n', (3089, 3091), False, 'from pathlib import Path\n'), ((4578, 4598), 'torch.device', 'torch.device', (['device'], {}), '(device)\n', (4590, 4598), False, 'import torch\n'), ((4729, 4838), 'torch.utils.data.DataLoader', 'DataLoader', (['dataset'], {'batch_size': 'b...
""" Wrapper for running amplicon_analysis_inputs report for LAAgc (with an extra input file). """ import logging import sys from pbcommand.models import FileTypes from pbcommand.cli import pbparser_runner from pbcommand.utils import setup_log from pbreports.report.amplicon_analysis_input import _get_parser, make_rep...
[ "logging.getLogger", "pbreports.report.amplicon_analysis_input._get_parser", "pbreports.report.amplicon_analysis_input.make_report" ]
[((332, 359), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (349, 359), False, 'import logging\n'), ((551, 642), 'pbreports.report.amplicon_analysis_input.make_report', 'make_report', (['args.report_csv', 'args.report_json', 'args.locus_csv', 'args.barcoded_subreads'], {}), '(args.report...
# The MIT License (MIT) # Copyright © 2021 Opentensor.ai # Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated # documentation files (the “Software”), to deal in the Software without restriction, including without limitation # the rights to use, copy, modify,...
[ "os.path.exists", "loguru.logger.info", "argparse.ArgumentParser", "bittensor.crypto.decrypt_data", "bittensor.crypto.is_encrypted", "os.access", "re.match", "os.chmod", "os.path.isfile", "bittensor.crypto.keyfiles.load_keypair_from_data", "loguru.logger.error", "bittensor.utils.Cli.ask_passwo...
[((2475, 2499), 'os.path.expanduser', 'os.path.expanduser', (['path'], {}), '(path)\n', (2493, 2499), False, 'import os\n'), ((2597, 2644), 'loguru.logger.info', 'logger.info', (['"""Loaded coldkey: {}"""', 'self.coldkey'], {}), "('Loaded coldkey: {}', self.coldkey)\n", (2608, 2644), False, 'from loguru import logger\n...
import unittest from datetime import datetime, timezone from pandas import DatetimeTZDtype from parameterized import parameterized import pandas as pd from aiokraken.utils.timeindexeddataframe import TimeindexedDataframe """ Test module. This is intended for extensive testing, using parameterized, hypothesis or simi...
[ "datetime.datetime", "pandas.Series", "datetime.datetime.fromtimestamp", "pandas.api.types.is_datetime64_ns_dtype", "pandas.api.types.is_numeric_dtype", "aiokraken.utils.timeindexeddataframe.TimeindexedDataframe", "pandas.api.types.is_datetime64tz_dtype", "unittest.main", "pandas.DatetimeTZDtype", ...
[((20171, 20186), 'unittest.main', 'unittest.main', ([], {}), '()\n', (20184, 20186), False, 'import unittest\n'), ((3430, 3477), 'aiokraken.utils.timeindexeddataframe.TimeindexedDataframe', 'TimeindexedDataframe', ([], {'data': 'df', 'index': '"""datetime"""'}), "(data=df, index='datetime')\n", (3450, 3477), False, 'f...
#!/usr/bin/env python # coding: utf-8 # In[1]: from bs4 import BeautifulSoup import re import json import numpy as np import sys sys.setrecursionlimit(10000) # Write a function that parses HTML pages into BeautifulSoup objects # In[2]: def soupify(html): soup = BeautifulSoup(html, 'lxml') return soup ...
[ "sys.setrecursionlimit", "json.loads", "re.compile", "bs4.BeautifulSoup", "re.findall" ]
[((132, 160), 'sys.setrecursionlimit', 'sys.setrecursionlimit', (['(10000)'], {}), '(10000)\n', (153, 160), False, 'import sys\n'), ((275, 302), 'bs4.BeautifulSoup', 'BeautifulSoup', (['html', '"""lxml"""'], {}), "(html, 'lxml')\n", (288, 302), False, 'from bs4 import BeautifulSoup\n'), ((717, 730), 'json.loads', 'json...
# coding: utf-8 """ Engine api Engine APIs # noqa: E501 The version of the OpenAPI document: 1.0.4 Generated by: https://openapi-generator.tech """ import pprint import re # noqa: F401 import six class EngineTask(object): """NOTE: This class is auto generated by OpenAPI Generator. Ref:...
[ "six.iteritems" ]
[((12047, 12080), 'six.iteritems', 'six.iteritems', (['self.openapi_types'], {}), '(self.openapi_types)\n', (12060, 12080), False, 'import six\n')]
""" Analyses skewness for continuous features Options: A. Log B. Yeo-Johnson C. QuantileTransformer """ import json import pandas as pd import numpy as np from sklearn.pipeline import make_pipeline from sklearn.preprocessing import power_transform, quantile_transform from pathlib import Path p = Path(__f...
[ "numpy.reshape", "pandas.concat", "src.logger.LOGGER.info", "pathlib.Path" ]
[((464, 488), 'src.logger.LOGGER.info', 'LOGGER.info', (['"""Load data"""'], {}), "('Load data')\n", (475, 488), False, 'from src.logger import LOGGER\n'), ((576, 615), 'src.logger.LOGGER.info', 'LOGGER.info', (['"""Process data - Logarithm"""'], {}), "('Process data - Logarithm')\n", (587, 615), False, 'from src.logge...
import pygame pygame.init() FPS = 60 WIDTH, HEIGHT = 800, 700 ROWS, COLS = 7, 8 SQUARE_SIZE = HEIGHT//ROWS # GRAPHICAL USER INTERFACE ICON_PATH = './src/murus_gallicus/assets/noun_checkers_1684698.png' WINDOW = pygame.display.set_mode((WIDTH, HEIGHT)) PADDING = 20 OUTLINE = 2 # RGB COLORS RED = (255, 0, 0) WHITE = ...
[ "pygame.display.set_mode", "pygame.init" ]
[((14, 27), 'pygame.init', 'pygame.init', ([], {}), '()\n', (25, 27), False, 'import pygame\n'), ((213, 253), 'pygame.display.set_mode', 'pygame.display.set_mode', (['(WIDTH, HEIGHT)'], {}), '((WIDTH, HEIGHT))\n', (236, 253), False, 'import pygame\n')]
import deepchem as dc import numpy as np import os def test_numpy_dataset_get_shape(): """Test that get_shape works for numpy datasets.""" num_datapoints = 100 num_features = 10 num_tasks = 10 # Generate data X = np.random.rand(num_datapoints, num_features) y = np.random.randint(2, size=(num_datapoints,...
[ "deepchem.data.DiskDataset.from_numpy", "deepchem.data.CSVLoader", "numpy.random.rand", "os.path.join", "numpy.array", "numpy.random.randint", "deepchem.feat.DummyFeaturizer", "deepchem.data.NumpyDataset", "os.path.abspath", "deepchem.data.DiskDataset" ]
[((227, 271), 'numpy.random.rand', 'np.random.rand', (['num_datapoints', 'num_features'], {}), '(num_datapoints, num_features)\n', (241, 271), True, 'import numpy as np\n'), ((278, 332), 'numpy.random.randint', 'np.random.randint', (['(2)'], {'size': '(num_datapoints, num_tasks)'}), '(2, size=(num_datapoints, num_tasks...
# -*- coding: utf-8 -*- # # Copyright 2015 Google LLC. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless requir...
[ "collections.OrderedDict" ]
[((1176, 1278), 'collections.OrderedDict', 'collections.OrderedDict', (["[('second', 1), ('minute', 60), ('hour', 60 * 60), ('day', 60 * 60 * 24)]"], {}), "([('second', 1), ('minute', 60), ('hour', 60 * 60),\n ('day', 60 * 60 * 24)])\n", (1199, 1278), False, 'import collections\n')]
import time a = 3215.35127 b = 3. start = time.time() for i in range(100000000): c = a / b end = time.time() time_elapsed = end - start print('Time elapsed (div ver) = %.5f' % time_elapsed) a = 3215.35127 b = 1./3. start = time.time() for i in range(100000000): c = a * b end = time.time() time_elapsed = end...
[ "time.time" ]
[((44, 55), 'time.time', 'time.time', ([], {}), '()\n', (53, 55), False, 'import time\n'), ((103, 114), 'time.time', 'time.time', ([], {}), '()\n', (112, 114), False, 'import time\n'), ((231, 242), 'time.time', 'time.time', ([], {}), '()\n', (240, 242), False, 'import time\n'), ((290, 301), 'time.time', 'time.time', ([...
# coding: utf-8 # Copyright (c) 2016, 2021, Oracle and/or its affiliates. All rights reserved. # This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may c...
[ "oci.util.formatted_flat_dict", "oci.util.value_allowed_none_or_none_sentinel" ]
[((5155, 5180), 'oci.util.formatted_flat_dict', 'formatted_flat_dict', (['self'], {}), '(self)\n', (5174, 5180), False, 'from oci.util import formatted_flat_dict, NONE_SENTINEL, value_allowed_none_or_none_sentinel\n'), ((4970, 5032), 'oci.util.value_allowed_none_or_none_sentinel', 'value_allowed_none_or_none_sentinel',...
from django.contrib import admin from .models import Search # Register your models here. admin.site.register(Search)
[ "django.contrib.admin.site.register" ]
[((90, 117), 'django.contrib.admin.site.register', 'admin.site.register', (['Search'], {}), '(Search)\n', (109, 117), False, 'from django.contrib import admin\n')]
# <NAME> # Institute of Mechatronic Systems # Leibniz Universität Hannover, Germany # 2019 # Code From https://github.com/mlaves/bayesian-temperature-scaling import torch __all__ = ['accuracy', 'kl_loss', 'nentr', 'xavier_normal_init'] def accuracy(input, target): _, max_indices = torch.max(input.data, 1) ac...
[ "torch.tensor", "torch.max", "torch.nn.functional.log_softmax", "torch.nn.init.xavier_normal_" ]
[((289, 313), 'torch.max', 'torch.max', (['input.data', '(1)'], {}), '(input.data, 1)\n', (298, 313), False, 'import torch\n'), ((756, 794), 'torch.tensor', 'torch.tensor', (['[1e-16]'], {'device': 'p.device'}), '([1e-16], device=p.device)\n', (768, 794), False, 'import torch\n'), ((823, 881), 'torch.tensor', 'torch.te...
import unittest from core.drive import copy from core.aesthetics import * class TestingDrivepy(unittest.TestCase): def test_copy_function(self): source_tests = [ r"D:\Alexzander__\programming\python\Python2Executable", r"D:\Alexzander__\programming\python\byzantion", ...
[ "unittest.main", "core.drive.copy" ]
[((1875, 1890), 'unittest.main', 'unittest.main', ([], {}), '()\n', (1888, 1890), False, 'import unittest\n'), ((1266, 1340), 'core.drive.copy', 'copy', (['source', 'destination'], {'open_destination_when_done': '(False)', '__print': '(False)'}), '(source, destination, open_destination_when_done=False, __print=False)\n...
from app import app from flask_restful import Api from app.resources.auth import TokenResource api = Api(app) # Token resource api.add_resource(TokenResource, '/authservice/token', endpoint='auth_token')
[ "flask_restful.Api" ]
[((103, 111), 'flask_restful.Api', 'Api', (['app'], {}), '(app)\n', (106, 111), False, 'from flask_restful import Api\n')]
#! /usr/bin/env python import os import rospy import rospkg from readbag import restore from qt_gui.plugin import Plugin from python_qt_binding.QtCore import Qt from python_qt_binding import loadUi from python_qt_binding.QtGui import QFileDialog, QGraphicsView, QIcon, QWidget from PyQt4 import QtGui, QtCore from exam...
[ "PyQt4.QtGui.QApplication", "PyQt4.QtGui.QWidget.__init__" ]
[((812, 840), 'PyQt4.QtGui.QApplication', 'QtGui.QApplication', (['sys.argv'], {}), '(sys.argv)\n', (830, 840), False, 'from PyQt4 import QtGui\n'), ((470, 506), 'PyQt4.QtGui.QWidget.__init__', 'QtGui.QWidget.__init__', (['self', 'parent'], {}), '(self, parent)\n', (492, 506), False, 'from PyQt4 import QtGui\n')]
import models import os import copy import torch import torch.nn as nn from lifelines import KaplanMeierFitter as KMFitter import pycox import numpy as np # local import catdist import data_utils import _concordance import _nll import _saver def str_to_bool(arg): """Convert an argument string into its boolean val...
[ "numpy.unique", "catdist.CatDist", "torch.zeros_like", "torch.isnan", "torch.zeros" ]
[((1717, 1778), 'catdist.CatDist', 'catdist.CatDist', ([], {'logits': 'None', 'args': 'args', 'probs': 'Gprobs', 'k': 'None'}), '(logits=None, args=args, probs=Gprobs, k=None)\n', (1732, 1778), False, 'import catdist\n'), ((699, 713), 'torch.isnan', 'torch.isnan', (['x'], {}), '(x)\n', (710, 713), False, 'import torch\...
from distutils.core import setup from pathlib import Path # TODO(joschnei): Add typing info setup( name="active", version="0.1", packages=["active",], install_requires=[ "scipy", "numpy", "driver @ git+https://github.com/jordan-schneider/driver-env.git#egg=driver", ], p...
[ "distutils.core.setup" ]
[((94, 318), 'distutils.core.setup', 'setup', ([], {'name': '"""active"""', 'version': '"""0.1"""', 'packages': "['active']", 'install_requires': "['scipy', 'numpy',\n 'driver @ git+https://github.com/jordan-schneider/driver-env.git#egg=driver'\n ]", 'package_data': "{'active': ['py.typed']}"}), "(name='active', ...
import tableprint from collections import Counter from .args import * from .util import * if __name__ == "__main__": args = get_args() output_classes = Counter() question_types = Counter() with tableprint.TableContext(headers=["Type", "Question", "Answer"], width=[40,50,15]) as t: for i in read_gqa(args):...
[ "collections.Counter", "tableprint.TableContext" ]
[((160, 169), 'collections.Counter', 'Counter', ([], {}), '()\n', (167, 169), False, 'from collections import Counter\n'), ((188, 197), 'collections.Counter', 'Counter', ([], {}), '()\n', (195, 197), False, 'from collections import Counter\n'), ((206, 294), 'tableprint.TableContext', 'tableprint.TableContext', ([], {'h...
#!/usr/bin/env python3 # Copyright (2021-) <NAME> <<EMAIL>> # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicab...
[ "tensorflow.nn.conv2d", "tensorflow.shape", "tensorflow.keras.activations.get", "tensorflow.keras.initializers.GlorotUniform", "kaldi_tflite.lib.layers.tdnn.utils.reshapeKaldiTdnnWeights", "tensorflow.range", "tensorflow.gather", "tensorflow.constant", "tensorflow.clip_by_value", "tensorflow.expan...
[((1765, 1780), 'tensorflow.keras.initializers.GlorotUniform', 'GlorotUniform', ([], {}), '()\n', (1778, 1780), False, 'from tensorflow.keras.initializers import Initializer, GlorotUniform\n'), ((1831, 1846), 'tensorflow.keras.initializers.GlorotUniform', 'GlorotUniform', ([], {}), '()\n', (1844, 1846), False, 'from te...
from django.shortcuts import render from rest_framework.response import Response from rest_framework.views import APIView from rest_framework import status, permissions import pyotp import base64 from rest_framework_simplejwt.tokens import RefreshToken from .serializers import CustomUserSerializer from .models import ...
[ "rest_framework_simplejwt.tokens.RefreshToken.for_user", "rest_framework.response.Response", "pyotp.HOTP" ]
[((1059, 1111), 'rest_framework.response.Response', 'Response', (["{'status': 201, 'userdata': user.username}"], {}), "({'status': 201, 'userdata': user.username})\n", (1067, 1111), False, 'from rest_framework.response import Response\n'), ((1623, 1667), 'rest_framework.response.Response', 'Response', ([], {'status': '...
import argparse parser = argparse.ArgumentParser( formatter_class=argparse.ArgumentDefaultsHelpFormatter) parser.add_argument('--d_model', type=int, default=0, help='d_model') parser.add_argument('--d_head', type=int, default=2, help='head') parser.add_argument('--d_inner', type=bool, default=True, help='inner lay...
[ "argparse.ArgumentParser" ]
[((25, 104), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'formatter_class': 'argparse.ArgumentDefaultsHelpFormatter'}), '(formatter_class=argparse.ArgumentDefaultsHelpFormatter)\n', (48, 104), False, 'import argparse\n')]
import logging import sys import os from logging.handlers import RotatingFileHandler from multiprocessing.pool import ThreadPool from optparse import OptionParser import requests from requests.packages import urllib3 urllib3.disable_warnings() # Workers configurations ASYNC_WORKERS_COUNT = 100 # How many threads wi...
[ "logging.basicConfig", "logging.getLogger", "requests.session", "os.path.exists", "requests.packages.urllib3.disable_warnings", "os.listdir", "os.linesep.join", "logging.Formatter", "os.path.join", "optparse.OptionParser", "multiprocessing.pool.ThreadPool", "os.path.isfile", "os.path.isdir",...
[((219, 245), 'requests.packages.urllib3.disable_warnings', 'urllib3.disable_warnings', ([], {}), '()\n', (243, 245), False, 'from requests.packages import urllib3\n'), ((1927, 2005), 'logging.basicConfig', 'logging.basicConfig', ([], {'stream': 'sys.stdout', 'level': 'logging_level', 'format': 'LOG_FORMAT'}), '(stream...
#Template of the Purkinje cell model, Zang et al. 2018 #Templating by Lungsi 2019 based on ~/PC2018Zang/purkinje.hoc #purkinje.hoc has been converted from original purkinje_demo and using readme.html as a guide from neuron import h #from pdb import set_trace as breakpoint from random import randint class Purkinje(obje...
[ "neuron.h.xopen" ]
[((395, 418), 'neuron.h.xopen', 'h.xopen', (['"""purkinje.hoc"""'], {}), "('purkinje.hoc')\n", (402, 418), False, 'from neuron import h\n')]
""" @brief Class to store data from a specific event @date Created July 2, 2018 @author <NAME> @bug No known bugs """ from fprime.common.models.serialize import time_type from fprime_gds.common.data_types import sys_data class EventData(sys_data.SysData): """ The EventData class stores a specific event mes...
[ "fprime.common.models.serialize.time_type.TimeType" ]
[((1702, 1722), 'fprime.common.models.serialize.time_type.TimeType', 'time_type.TimeType', ([], {}), '()\n', (1720, 1722), False, 'from fprime.common.models.serialize import time_type\n')]
from os.path import exists, join import pytest from flask_dj import startproject from tests.basic_project_creator import ProjectCreate class TestBaseSettingProjectConstructor(ProjectCreate): def setup(self, need_static=False, need_templates=False): super().setup() def test_project_folder_exist(self):...
[ "os.path.exists", "flask_dj.startproject", "os.path.join", "pytest.raises" ]
[((336, 361), 'os.path.exists', 'exists', (['self.project_path'], {}), '(self.project_path)\n', (342, 361), False, 'from os.path import exists, join\n'), ((423, 465), 'os.path.join', 'join', (['self.project_path', 'self.project_name'], {}), '(self.project_path, self.project_name)\n', (427, 465), False, 'from os.path im...
from django.test import TestCase from rest_framework.test import APIRequestFactory from .views import LoginView, SignupView from common.tests import USER_DATA, prepare_dummy_user_data # Create your tests here. class ApiViewTests(TestCase): factory = APIRequestFactory() def test_signup(self): user_da...
[ "rest_framework.test.APIRequestFactory", "common.tests.prepare_dummy_user_data" ]
[((257, 276), 'rest_framework.test.APIRequestFactory', 'APIRequestFactory', ([], {}), '()\n', (274, 276), False, 'from rest_framework.test import APIRequestFactory\n'), ((949, 974), 'common.tests.prepare_dummy_user_data', 'prepare_dummy_user_data', ([], {}), '()\n', (972, 974), False, 'from common.tests import USER_DAT...
# Copyright 2015 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. ''' Utility functions for all things related to manipulating google play services related files. ''' import argparse import filecmp import json import loggi...
[ "logging.debug", "re.compile", "devil.utils.cmd_helper.Call", "json.dumps", "os.path.isfile", "os.path.dirname", "json.load", "filecmp.cmp" ]
[((490, 567), 're.compile', 're.compile', (['"""<integer name="google_play_services_version">(\\\\d+)<\\\\/integer>"""'], {}), '(\'<integer name="google_play_services_version">(\\\\d+)<\\\\/integer>\')\n', (500, 567), False, 'import re\n'), ((3565, 3612), 'json.dumps', 'json.dumps', (['json_data'], {'sort_keys': '(True...
# -*-coding: utf-8 import sys import argparse from tagger import FileTagger def tagger_add(args): tg = FileTagger() res = tg.add_tags(args.path, *args.tags) if not res: print("[-] Fail to add tags.") def tagger_rm(args): tg = FileTagger() res = tg.rm_tags(args.path, *args.tags) if n...
[ "argparse.ArgumentParser", "tagger.FileTagger" ]
[((110, 122), 'tagger.FileTagger', 'FileTagger', ([], {}), '()\n', (120, 122), False, 'from tagger import FileTagger\n'), ((255, 267), 'tagger.FileTagger', 'FileTagger', ([], {}), '()\n', (265, 267), False, 'from tagger import FileTagger\n'), ((404, 416), 'tagger.FileTagger', 'FileTagger', ([], {}), '()\n', (414, 416),...
import os FILES_TO_REMOVE = ["setup.cfg", "pyproject.toml", "setup.py"] if "{{ cookiecutter.as_package }}" == "no": for file in FILES_TO_REMOVE: os.remove(file) if "{{ cookiecutter.remove_test_script }}" == "yes": os.remove("test_log.py")
[ "os.remove" ]
[((233, 257), 'os.remove', 'os.remove', (['"""test_log.py"""'], {}), "('test_log.py')\n", (242, 257), False, 'import os\n'), ((159, 174), 'os.remove', 'os.remove', (['file'], {}), '(file)\n', (168, 174), False, 'import os\n')]
from django import forms from crispy_forms.helper import FormHelper class MiscForm(forms.Form): notes = forms.CharField( label='Notes for the registration team', help_text='Anything else you need to describe. ' 'The registration team will see this. ' 'The bursa...
[ "django.forms.Textarea", "crispy_forms.helper.FormHelper" ]
[((528, 540), 'crispy_forms.helper.FormHelper', 'FormHelper', ([], {}), '()\n', (538, 540), False, 'from crispy_forms.helper import FormHelper\n'), ((357, 390), 'django.forms.Textarea', 'forms.Textarea', ([], {'attrs': "{'rows': 3}"}), "(attrs={'rows': 3})\n", (371, 390), False, 'from django import forms\n')]
#Import libraries import cv2 #Read image file into n dimensional numpy array #Read as greyscale image -> 0, Read as color image w/o transparency -> 1, Read as color image w transparency -> -1 img = cv2.imread("Practice/images/galaxy.jpg", 0) print(type(img)) print(img) #To know the number of rows and columns print(i...
[ "cv2.imwrite", "cv2.imshow", "cv2.destroyAllWindows", "cv2.waitKey", "cv2.imread" ]
[((199, 242), 'cv2.imread', 'cv2.imread', (['"""Practice/images/galaxy.jpg"""', '(0)'], {}), "('Practice/images/galaxy.jpg', 0)\n", (209, 242), False, 'import cv2\n'), ((667, 700), 'cv2.imshow', 'cv2.imshow', (['"""Galaxy"""', 'resized_img'], {}), "('Galaxy', resized_img)\n", (677, 700), False, 'import cv2\n'), ((797, ...
import asyncio from .main import create_app loop = asyncio.get_event_loop() app = create_app(loop)
[ "asyncio.get_event_loop" ]
[((53, 77), 'asyncio.get_event_loop', 'asyncio.get_event_loop', ([], {}), '()\n', (75, 77), False, 'import asyncio\n')]
import io, re, codecs, sys, os.path import pprint REG = 0 SHFT = 1 CTRL = 2 ALT = 4 ALTGR = 6 COMPRESSED_OUTPUT=1 def get_kbd_layout(base_filename, load_patch = False): filename_klc = base_filename filename_changes = base_filename + 'patch' f = io.open(filename_klc, mode="r", encoding="utf-8") lines = f.r...
[ "re.split", "pprint.pformat", "io.open", "sys.exit" ]
[((258, 307), 'io.open', 'io.open', (['filename_klc'], {'mode': '"""r"""', 'encoding': '"""utf-8"""'}), "(filename_klc, mode='r', encoding='utf-8')\n", (265, 307), False, 'import io, re, codecs, sys, os.path\n'), ((14988, 15050), 'sys.exit', 'sys.exit', (["('unknown locale format: ' + kbd_layout['localename'])"], {}), ...
import pandas as pd from gooey import Gooey, GooeyParser import numpy as np import xlsxwriter import xlrd @Gooey(program_name="FEC FILE FOR FRANCE", required_cols= 4,default_size=(710, 700),navigation='TABBED', header_bg_color = '#48a7fa') def parse_args(): parser = GooeyParser() FilesGL = par...
[ "pandas.ExcelWriter", "gooey.Gooey", "gooey.GooeyParser", "pandas.read_excel" ]
[((118, 254), 'gooey.Gooey', 'Gooey', ([], {'program_name': '"""FEC FILE FOR FRANCE"""', 'required_cols': '(4)', 'default_size': '(710, 700)', 'navigation': '"""TABBED"""', 'header_bg_color': '"""#48a7fa"""'}), "(program_name='FEC FILE FOR FRANCE', required_cols=4, default_size=(\n 710, 700), navigation='TABBED', he...
import boto3 def handler(event, _): boto3.client('codebuild').start_build( projectName=event['Records'][0]['customData'])
[ "boto3.client" ]
[((38, 63), 'boto3.client', 'boto3.client', (['"""codebuild"""'], {}), "('codebuild')\n", (50, 63), False, 'import boto3\n')]
# menu.py # 维护暂停界面 import pygame from pygame.locals import * import sys from utility import globe from process.scene import menu_confirm from PIL import Image, ImageFilter class Pause_Menu(object): # 暂停页面 def __init__(self): self.button_rect = [] self.rs = globe.destiny.rsManager.image ...
[ "utility.globe.destiny.msManager.play_SE", "pygame.image.tostring", "pygame.quit", "pygame.event.get", "pygame.image.fromstring", "utility.globe.destiny.back", "utility.globe.destiny.call", "utility.globe.destiny.msManager.unpause", "sys.exit", "utility.globe.destiny.screen.get_size" ]
[((1175, 1193), 'pygame.event.get', 'pygame.event.get', ([], {}), '()\n', (1191, 1193), False, 'import pygame\n'), ((5274, 5305), 'utility.globe.destiny.screen.get_size', 'globe.destiny.screen.get_size', ([], {}), '()\n', (5303, 5305), False, 'from utility import globe\n'), ((5502, 5551), 'pygame.image.tostring', 'pyga...
from setuptools import setup # get the version here pkg_vars = {} with open("version.py") as fp: exec(fp.read(), pkg_vars) setup( name='sn_simulation', version= pkg_vars['__version__'], description='Simulations for supernovae', url='http://github.com/lsstdesc/sn_simulation', author='<NAME>',...
[ "setuptools.setup" ]
[((131, 514), 'setuptools.setup', 'setup', ([], {'name': '"""sn_simulation"""', 'version': "pkg_vars['__version__']", 'description': '"""Simulations for supernovae"""', 'url': '"""http://github.com/lsstdesc/sn_simulation"""', 'author': '"""<NAME>"""', 'author_email': '"""<EMAIL>"""', 'license': '"""BSD"""', 'packages':...
import contextlib import datetime import json import StringIO from django.contrib.auth import logout as logout_user from django.contrib import messages from django.core.urlresolvers import reverse from django.core.exceptions import ObjectDoesNotExist from django.http import HttpResponseRedirect, JsonResponse, Http404,...
[ "simpleoncall.forms.schedule.TeamScheduleForm", "django.http.HttpResponseBadRequest", "simpleoncall.models.Alert.objects.get", "django.db.models.Count", "django.contrib.messages.warning", "django.utils.http.urlencode", "django.core.urlresolvers.reverse", "django.utils.http.urlquote", "simpleoncall.f...
[((1151, 1175), 'simpleoncall.decorators.require_authentication', 'require_authentication', ([], {}), '()\n', (1173, 1175), False, 'from simpleoncall.decorators import require_authentication, require_selected_team\n'), ((1177, 1200), 'simpleoncall.decorators.require_selected_team', 'require_selected_team', ([], {}), '(...
from bs4 import BeautifulSoup import requests import os """ This script download all songs with given genre from midiworld.com """ genre_name = input( "type in genre name (lowercase, no space, no special characters): ") # Just in case someone don't respect the rules. genre_name = genre_name.lower() genre_name = g...
[ "os.path.join", "requests.get", "bs4.BeautifulSoup", "os.path.isdir", "os.mkdir" ]
[((392, 446), 'os.path.join', 'os.path.join', (['"""genresDataset"""', 'genre_name', '"""midiworld"""'], {}), "('genresDataset', genre_name, 'midiworld')\n", (404, 446), False, 'import os\n'), ((454, 475), 'os.path.isdir', 'os.path.isdir', (['folder'], {}), '(folder)\n', (467, 475), False, 'import os\n'), ((481, 497), ...
#!/usr/bin/env python # coding: utf-8 __author__ = '<NAME>' __copyright__ = 'Copyright 2017-2020, <NAME>' __license__ = 'MIT' __version__ = '0.5' __email__ = '<EMAIL>' __status__ = 'Development' __description__ = 'Tkinter based GUI, visualizing PASCAL VOC object detection annotation' """ Changelog: - 2020-06-16 11:3...
[ "logging.getLogger", "Tkinter.Entry", "cv2.rectangle", "logging.StreamHandler", "colorsys.hsv_to_rgb", "numpy.array", "PIL.ImageDraw.Draw", "Tkinter.Scrollbar", "Tkinter.LabelFrame", "Tkinter.Label", "os.path.exists", "os.listdir", "Tkinter.Tk.__init__", "Tkinter.Listbox", "Tkinter.Strin...
[((2452, 2471), 'PIL.Image.fromarray', 'Image.fromarray', (['im'], {}), '(im)\n', (2467, 2471), False, 'from PIL import Image, ImageTk, ImageFont, ImageDraw\n'), ((2483, 2505), 'PIL.ImageDraw.Draw', 'ImageDraw.Draw', (['im_pil'], {}), '(im_pil)\n', (2497, 2505), False, 'from PIL import Image, ImageTk, ImageFont, ImageD...
__author__ = 'dimd' from twisted.python import log from zope.interface import implementer from NetCatKS.Logger.api.interfaces import ILogger GLOBAL_DEBUG = True @implementer(ILogger) class Logger(object): def __init__(self): pass def debug(self, msg): if GLOBAL_DEBUG is True: ...
[ "zope.interface.implementer" ]
[((167, 187), 'zope.interface.implementer', 'implementer', (['ILogger'], {}), '(ILogger)\n', (178, 187), False, 'from zope.interface import implementer\n')]
from plenum.test.bls.helper import check_bls_multi_sig_after_send from plenum.test.pool_transactions.conftest import looper, clientAndWallet1, \ client1, wallet1, client1Connected nodeCount = 4 nodes_wth_bls = 0 def test_each_node_has_bls(txnPoolNodeSet): for node in txnPoolNodeSet: assert node.bls_b...
[ "plenum.test.bls.helper.check_bls_multi_sig_after_send" ]
[((491, 593), 'plenum.test.bls.helper.check_bls_multi_sig_after_send', 'check_bls_multi_sig_after_send', (['looper', 'txnPoolNodeSet', 'client1', 'wallet1'], {'saved_multi_sigs_count': '(0)'}), '(looper, txnPoolNodeSet, client1, wallet1,\n saved_multi_sigs_count=0)\n', (521, 593), False, 'from plenum.test.bls.helper...
from __future__ import division, print_function from .. import __version__ from ._global_imports import * try: import h5py except ImportError: print('Install h5py to enable signal caching.') raise class _Cache(object): """ Cache numerical model objects computed during likelihood evaluation. :pa...
[ "datetime.datetime.now", "h5py.File" ]
[((1493, 1520), 'h5py.File', 'h5py.File', (['self._path', 'mode'], {}), '(self._path, mode)\n', (1502, 1520), False, 'import h5py\n'), ((6214, 6228), 'datetime.datetime.now', 'datetime.now', ([], {}), '()\n', (6226, 6228), False, 'from datetime import datetime\n')]
from multiprocessing import Process, Manager ''' Custom Module Imports ''' from calculator.add import addition from calculator.subtract import subtraction from calculator.multiply import multiplication from calculator.divide import division class Main: def __init__(self) -> None: pass d...
[ "calculator.subtract.subtraction", "calculator.multiply.multiplication", "multiprocessing.Process", "calculator.divide.division", "calculator.add.addition", "multiprocessing.Manager" ]
[((397, 406), 'multiprocessing.Manager', 'Manager', ([], {}), '()\n', (404, 406), False, 'from multiprocessing import Process, Manager\n'), ((506, 534), 'calculator.add.addition', 'addition', (['i[0]', 'i[1]', 'result'], {}), '(i[0], i[1], result)\n', (514, 534), False, 'from calculator.add import addition\n'), ((558, ...
"""minor fixes Revision ID: <KEY> Revises: 0<PASSWORD> Create Date: 2020-09-18 07:56:14.159782 """ from alembic import op import geoalchemy2 import sqlalchemy as sa import backend # revision identifiers, used by Alembic. revision = '<KEY>' down_revision = '0fed690a57ce' branch_labels = () depends_on = None def up...
[ "alembic.op.drop_table", "sqlalchemy.VARCHAR", "alembic.op.f", "sqlalchemy.PrimaryKeyConstraint", "sqlalchemy.INTEGER", "sqlalchemy.CheckConstraint", "sqlalchemy.BigInteger" ]
[((1714, 1746), 'alembic.op.drop_table', 'op.drop_table', (['"""task_harvesting"""'], {}), "('task_harvesting')\n", (1727, 1746), False, 'from alembic import op\n'), ((1549, 1642), 'sqlalchemy.CheckConstraint', 'sa.CheckConstraint', (['"""(srid > 0) AND (srid <= 998999)"""'], {'name': '"""spatial_ref_sys_srid_check"""'...
""" Tests for Day 22 """ from day22.module import part_1, part_2, \ FULL_INPUT_FILE, TEST_INPUT_FILE_1, TEST_INPUT_FILE_2, TEST_INPUT_FILE_3 def test_part_1_1(): result = part_1(TEST_INPUT_FILE_1) assert result == 39 def test_part_1_2(): result = part_1(TEST_INPUT_FILE_2) assert result == 59078...
[ "day22.module.part_2", "day22.module.part_1" ]
[((182, 207), 'day22.module.part_1', 'part_1', (['TEST_INPUT_FILE_1'], {}), '(TEST_INPUT_FILE_1)\n', (188, 207), False, 'from day22.module import part_1, part_2, FULL_INPUT_FILE, TEST_INPUT_FILE_1, TEST_INPUT_FILE_2, TEST_INPUT_FILE_3\n'), ((268, 293), 'day22.module.part_1', 'part_1', (['TEST_INPUT_FILE_2'], {}), '(TES...
# -*- coding: utf-8 -*- # DO NOT EDIT THIS FILE! # This file has been autogenerated by dephell <3 # https://github.com/dephell/dephell try: from setuptools import setup except ImportError: from distutils.core import setup import os.path readme = "" here = os.path.abspath(os.path.dirname(__file__)) readme_p...
[ "distutils.core.setup" ]
[((483, 2453), 'distutils.core.setup', 'setup', ([], {'long_description': 'readme', 'name': '"""lightbus"""', 'version': '"""1.1.0"""', 'description': '"""RPC & event framework for Python 3"""', 'python_requires': '""">=3.7"""', 'project_urls': "{'documentation': 'https://lightbus.org', 'homepage':\n 'https://lightb...
import ENVIRONMENT from src.database.database_creation import createPlayerTrueSkillDictionary from src.skill_algorithms.algorithms import trueSkillMatchWithRawNums, trueSkillTipWinProb from src.skill_algorithms.common_data_processing import beforeMatchPredictions, runAlgoForSeason, runAlgoForAllSeasons # backlogtodo...
[ "src.skill_algorithms.common_data_processing.beforeMatchPredictions", "src.database.database_creation.createPlayerTrueSkillDictionary", "src.skill_algorithms.algorithms.trueSkillMatchWithRawNums", "src.skill_algorithms.common_data_processing.runAlgoForSeason", "src.skill_algorithms.common_data_processing.ru...
[((568, 1049), 'src.skill_algorithms.common_data_processing.runAlgoForSeason', 'runAlgoForSeason', (['seasonCsv', 'ENVIRONMENT.PLAYER_TRUESKILL_DICT_PATH', 'ENVIRONMENT.TS_PREDICTION_SUMMARIES_PATH', 'trueSkillBeforeMatchPredictions', 'trueSkillUpdateDataSingleTipoff', 'winningBetThreshold'], {'columnAdds': "['Home TS ...
__author__ = 'multiangle' # 这是实现 霍夫曼树相关的文件, 主要用于 针对层次softmax进行 word2vec 优化方案的一种 ''' 至于 为什么要进行层次softmax 可以简单理解 因为词表很大 针对上完个类别单词进行softmax 计算量大 更新参数过多 无法训练,而采用softmax 层次化 只需要 计算几个有限单词的sigmod 就可以 更新参数也非常少 提高训练速度 什么是霍夫曼树 简单理解就是 将训练文本 进行词频统计 通过构建加权最短路径来构造二叉树 这样 词频高的 位置在前 词频低的位置在后 每一个 霍夫曼编码代表一个词 路径 并且是唯一 不是其他词的前缀 ''' impo...
[ "numpy.zeros" ]
[((4947, 4974), 'numpy.zeros', 'np.zeros', (['[1, self.vec_len]'], {}), '([1, self.vec_len])\n', (4955, 4974), True, 'import numpy as np\n')]
__all__ = [ "fit_gp", "ft_gp", "fit_lm", "ft_lm", "fit_rf", "ft_rf", "fit_kmeans", "ft_kmeans", ] ## Fitting via sklearn package try: from sklearn.base import clone from sklearn.linear_model import LinearRegression from sklearn.gaussian_process import GaussianProcessRegresso...
[ "grama.add_pipe", "sklearn.cluster.KMeans", "sklearn.ensemble.RandomForestRegressor", "grama.Model", "sklearn.linear_model.LinearRegression", "grama.cp_vec_function", "copy.deepcopy", "pandas.DataFrame", "warnings.filterwarnings" ]
[((7062, 7078), 'grama.add_pipe', 'add_pipe', (['fit_gp'], {}), '(fit_gp)\n', (7070, 7078), False, 'from grama import add_pipe, pipe\n'), ((9811, 9827), 'grama.add_pipe', 'add_pipe', (['fit_rf'], {}), '(fit_rf)\n', (9819, 9827), False, 'from grama import add_pipe, pipe\n'), ((12030, 12046), 'grama.add_pipe', 'add_pipe'...
from django.conf import settings from django.test import TestCase # Create your tests here. class SettingsTest(TestCase): def test_settings(self): self.assertEqual(settings.SERVER, 'prodserver.com') self.assertEqual( settings.STATIC_URL, '/changed/in/settings.toml/by/dynaconf/') ...
[ "django.conf.settings.get", "django.conf.settings.using_env" ]
[((466, 490), 'django.conf.settings.get', 'settings.get', (['"""PASSWORD"""'], {}), "('PASSWORD')\n", (478, 490), False, 'from django.conf import settings\n'), ((591, 624), 'django.conf.settings.using_env', 'settings.using_env', (['"""development"""'], {}), "('development')\n", (609, 624), False, 'from django.conf impo...
# coding=utf-8 # Copyright 2020 The Google Research Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicab...
[ "wget.download", "shutil.copyfileobj", "pandas.read_csv", "gzip.open", "sklearn.model_selection.train_test_split", "os.mkdir" ]
[((880, 898), 'os.mkdir', 'os.mkdir', (['"""./data"""'], {}), "('./data')\n", (888, 898), False, 'import os\n'), ((910, 928), 'wget.download', 'wget.download', (['url'], {}), '(url)\n', (923, 928), False, 'import wget\n'), ((1059, 1090), 'pandas.read_csv', 'pd.read_csv', (['"""data/covtype.csv"""'], {}), "('data/covtyp...
import pytest import datetime from django.test import TestCase from django.utils import timezone from gamer_registration_system.con.models import Convention, Event, EventSchedule # Create your tests here. class EventScheduleModelTests(TestCase): new_con = Convention(convention_name='Test Future Con') new_eve...
[ "gamer_registration_system.con.models.Event", "gamer_registration_system.con.models.EventSchedule", "django.utils.timezone.now", "gamer_registration_system.con.models.Convention", "datetime.timedelta" ]
[((263, 308), 'gamer_registration_system.con.models.Convention', 'Convention', ([], {'convention_name': '"""Test Future Con"""'}), "(convention_name='Test Future Con')\n", (273, 308), False, 'from gamer_registration_system.con.models import Convention, Event, EventSchedule\n'), ((325, 377), 'gamer_registration_system.c...
#!/usr/bin/env python import rospy import tf from jsk_recognition_msgs.msg import BoundingBox class BoundingBoxToTf(object): def __init__(self): self.tf_frame = rospy.get_param('~tf_frame', 'bounding_box') self.broadcaster = tf.TransformBroadcaster() self.sub = rospy.Subscriber('~inpu...
[ "tf.TransformBroadcaster", "rospy.init_node", "rospy.get_param", "rospy.Time.now", "rospy.spin", "rospy.Subscriber" ]
[((783, 821), 'rospy.init_node', 'rospy.init_node', (['""" bounding_box_to_tf"""'], {}), "(' bounding_box_to_tf')\n", (798, 821), False, 'import rospy\n'), ((854, 866), 'rospy.spin', 'rospy.spin', ([], {}), '()\n', (864, 866), False, 'import rospy\n'), ((178, 222), 'rospy.get_param', 'rospy.get_param', (['"""~tf_frame"...
from defs_utils import * import pdf2image import os def transforma_pdf_em_img_por_materia(materia, pdf_path=None): searched = materia if pdf_path: list_files = list_dir(complete_name(searched, pre=pdf_path), True) else: list_files = list_dir(complete_name(searched), True) volta = os.ge...
[ "os.getcwd", "os.chdir", "os.path.realpath", "os.mkdir", "pdf2image.convert_from_path" ]
[((315, 326), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (324, 326), False, 'import os\n'), ((372, 405), 'pdf2image.convert_from_path', 'pdf2image.convert_from_path', (['file'], {}), '(file)\n', (399, 405), False, 'import pdf2image\n'), ((434, 449), 'os.chdir', 'os.chdir', (['volta'], {}), '(volta)\n', (442, 449), Fal...
# Copyright (c) 2017, Carnegie Mellon University. All rights reserved. # # Use of the K-NRM package is subject to the terms of the software license set # forth in the LICENSE file included with this software, and also available at # https://github.com/AdeDZY/K-NRM/blob/master/LICENSE from setuptools import setup from ...
[ "setuptools.find_packages" ]
[((528, 543), 'setuptools.find_packages', 'find_packages', ([], {}), '()\n', (541, 543), False, 'from setuptools import find_packages\n')]
#Imports import config import telebot from db_manager import SQL import os #Globals restaurants = {"12345":"abc"} current_restaurant_id = "" current_restaurant_key = "" couriers = {"12345":"abc"} current_courier_id = "" current_courier_key = "" current_courier_altitude = 0 current_courier_longitude = 0 #DB + BOT CO...
[ "os.path.join", "db_manager.SQL", "telebot.types.ReplyKeyboardMarkup", "os.path.abspath", "telebot.TeleBot" ]
[((358, 380), 'telebot.TeleBot', 'telebot.TeleBot', (['token'], {}), '(token)\n', (373, 380), False, 'import telebot\n'), ((445, 482), 'os.path.join', 'os.path.join', (['BASE_DIR', '"""database.db"""'], {}), "(BASE_DIR, 'database.db')\n", (457, 482), False, 'import os\n'), ((488, 500), 'db_manager.SQL', 'SQL', (['db_pa...
# -*- coding: utf-8 -*- # Generated by Django 1.11.5 on 2017-09-10 20:53 from __future__ import unicode_literals from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ("bdiadmin", "0013_auto_20170319_1415"), ("stock", "0...
[ "django.db.models.FloatField", "django.db.models.ForeignKey", "django.db.models.AutoField", "django.db.models.PositiveIntegerField", "django.db.models.CharField" ]
[((523, 616), 'django.db.models.AutoField', 'models.AutoField', ([], {'auto_created': '(True)', 'primary_key': '(True)', 'serialize': '(False)', 'verbose_name': '"""ID"""'}), "(auto_created=True, primary_key=True, serialize=False,\n verbose_name='ID')\n", (539, 616), False, 'from django.db import migrations, models\...
# -*- coding: UTF-8 -*- """ Source processing routines """ from __future__ import absolute_import from __future__ import unicode_literals import warnings from collections import OrderedDict from astropy.cosmology import default_cosmology import numpy as np import os import pysynphot as S import astropy.table as at fro...
[ "os.path.exists", "collections.OrderedDict", "numpy.isscalar", "os.path.join", "pysynphot.ArraySpectrum", "numpy.array", "astropy.cosmology.default_cosmology.get", "warnings.warn", "astropy.table.Table.read" ]
[((732, 745), 'collections.OrderedDict', 'OrderedDict', ([], {}), '()\n', (743, 745), False, 'from collections import OrderedDict\n'), ((753, 777), 'numpy.isscalar', 'np.isscalar', (['sourcenames'], {}), '(sourcenames)\n', (764, 777), True, 'import numpy as np\n'), ((2274, 2316), 'os.path.join', 'os.path.join', (['"""s...
import pandas as pd import numpy as np import pickle from sklearn.metrics.pairwise import linear_kernel from sklearn.feature_extraction.text import TfidfVectorizer import csv from kivy.app import App from kivy.uix.gridlayout import GridLayout from kivy.uix.label import Label from kivy.uix.textinput import TextInput ...
[ "kivy.uix.label.Label", "numpy.ndarray.argsort", "kivy.uix.textinput.TextInput", "csv.reader" ]
[((1260, 1286), 'kivy.uix.textinput.TextInput', 'TextInput', ([], {'multiline': '(False)'}), '(multiline=False)\n', (1269, 1286), False, 'from kivy.uix.textinput import TextInput\n'), ((1398, 1439), 'kivy.uix.textinput.TextInput', 'TextInput', ([], {'password': '(True)', 'multiline': '(False)'}), '(password=True, multi...
from django import template register = template.Library() @register.inclusion_tag('quiz/correct_answer.html', takes_context=True) def correct_answer_for_all(context, question): """ processes the correct answer based on a given question object if the answer is incorrect, informs the user """ answe...
[ "django.template.Library" ]
[((40, 58), 'django.template.Library', 'template.Library', ([], {}), '()\n', (56, 58), False, 'from django import template\n')]
# -*- coding: utf-8 -*- """ Created on Mon Jun 29 14:14:45 2020 @author: Nikki """ import numpy as np import cv2 import transform as tform import sys import math import scipy.spatial import markers ###--------------------------------------------------------------------------- # Allows video to be initialized usin...
[ "math.sqrt", "cv2.imshow", "math.cos", "numpy.array", "cv2.ellipse", "sys.exc_info", "cv2.destroyAllWindows", "cv2.setMouseCallback", "markers.aot_1_markers", "numpy.asarray", "markers.mrb3_markers", "cv2.addWeighted", "markers.aot_3_markers", "cv2.waitKey", "transform.get_best_transform...
[((1885, 1915), 'transform.get_best_transform', 'tform.get_best_transform', (['x', 'y'], {}), '(x, y)\n', (1909, 1915), True, 'import transform as tform\n'), ((1930, 1960), 'transform.get_best_transform', 'tform.get_best_transform', (['y', 'x'], {}), '(y, x)\n', (1954, 1960), True, 'import transform as tform\n'), ((300...
# mp4.py # --------------- # Licensing Information: You are free to use or extend this projects for # educational purposes provided that (1) you do not distribute or publish # solutions, (2) you retain this notice, and (3) you provide clear # attribution to the University of Illinois at Urbana-Champaign # # Created Fa...
[ "collections.Counter" ]
[((1845, 1866), 'collections.Counter', 'collections.Counter', ([], {}), '()\n', (1864, 1866), False, 'import collections\n'), ((2090, 2111), 'collections.Counter', 'collections.Counter', ([], {}), '()\n', (2109, 2111), False, 'import collections\n')]
from ex2.sigmoid import sigmoid def sigmoidGradient(z): """computes the gradient of the sigmoid function evaluated at z. This should work regardless if z is a matrix or a vector. In particular, if z is a vector or matrix, you should return the gradient for each element.""" # =====================...
[ "ex2.sigmoid.sigmoid" ]
[((521, 531), 'ex2.sigmoid.sigmoid', 'sigmoid', (['z'], {}), '(z)\n', (528, 531), False, 'from ex2.sigmoid import sigmoid\n'), ((539, 549), 'ex2.sigmoid.sigmoid', 'sigmoid', (['z'], {}), '(z)\n', (546, 549), False, 'from ex2.sigmoid import sigmoid\n')]
# -*- coding: utf-8 -*- from __future__ import (print_function, unicode_literals, absolute_import, division) from pusher.config import Config from pusher.request import Request from pusher.sync import SynchronousBackend from pusher.util import GET, POST, text, validate_channel import collectio...
[ "json.dumps", "pusher.util.validate_channel", "pusher.sync.SynchronousBackend", "six.text_type", "pusher.request.Request" ]
[((3131, 3205), 'pusher.request.Request', 'Request', (['self.config', 'POST', "('/apps/%s/events' % self.config.app_id)", 'params'], {}), "(self.config, POST, '/apps/%s/events' % self.config.app_id, params)\n", (3138, 3205), False, 'from pusher.request import Request\n'), ((3623, 3698), 'pusher.request.Request', 'Reque...
""" To run this module directly """ # pylint: disable=no-else-return import argparse import os try: from . import prefetch_from_file except ImportError: import prefetch_from_file try: from . import prefetch_from_url except ImportError: import prefetch_from_url def validate_filepath_or_url(filepath_o...
[ "argparse.ArgumentParser", "os.access", "os.path.isfile", "prefetch_from_url.url_to_prefetch", "prefetch_from_file.file_to_prefetch" ]
[((678, 781), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '__doc__', 'formatter_class': 'argparse.RawDescriptionHelpFormatter'}), '(description=__doc__, formatter_class=argparse.\n RawDescriptionHelpFormatter)\n', (701, 781), False, 'import argparse\n'), ((1642, 1718), 'prefetch_from_f...
import numpy as np from src import const #TODO: should be imported from aguirregabiria_simple.py def period_profit(p: np.ndarray, lambdas: np.ndarray, betas_transition=const.betas_transition): """ Correct expected period return profit. See ReadMe for derivation """ constant_part = (p-const.c) * np.e *...
[ "numpy.log", "numpy.array", "numpy.allclose" ]
[((518, 538), 'numpy.array', 'np.array', (['[1.4, 1.2]'], {}), '([1.4, 1.2])\n', (526, 538), True, 'import numpy as np\n'), ((553, 578), 'numpy.array', 'np.array', (['[0.5, 0.4, 0.1]'], {}), '([0.5, 0.4, 0.1])\n', (561, 578), True, 'import numpy as np\n'), ((598, 791), 'numpy.array', 'np.array', (['[[np.e ** (-3.0 * 0....
import mock from pyramid.request import Request from .support import DummyRequest, unittest from kinto.core import authentication from kinto.core.authorization import RouteFactory, AuthorizationPolicy from kinto.core.storage import exceptions as storage_exceptions class RouteFactoryTest(unittest.TestCase): def...
[ "mock.patch", "mock.Mock", "pyramid.request.Request.blank", "kinto.core.authorization.RouteFactory", "kinto.core.authorization.AuthorizationPolicy", "mock.MagicMock" ]
[((3505, 3528), 'pyramid.request.Request.blank', 'Request.blank', ([], {'path': '"""/"""'}), "(path='/')\n", (3518, 3528), False, 'from pyramid.request import Request\n'), ((3556, 3578), 'mock.Mock', 'mock.Mock', ([], {'settings': '{}'}), '(settings={})\n', (3565, 3578), False, 'import mock\n'), ((3707, 3728), 'kinto.c...
import copy #from enum import IntFlag from time import sleep # I tried to use enum here, but I was having a problem with packages in the image, so I gave up as I just want to get it done class FieldValue: Empty = 0 Wall = 1 Player = 2 Box = 4 Goal = 8 class SenseHATColour: Red = (204, 4, 4) ...
[ "sense_hat.SenseHat", "time.sleep", "copy.deepcopy" ]
[((1862, 1872), 'sense_hat.SenseHat', 'SenseHat', ([], {}), '()\n', (1870, 1872), False, 'from sense_hat import SenseHat\n'), ((4169, 4189), 'copy.deepcopy', 'copy.deepcopy', (['level'], {}), '(level)\n', (4182, 4189), False, 'import copy\n'), ((6258, 6269), 'time.sleep', 'sleep', (['(0.25)'], {}), '(0.25)\n', (6263, 6...
# coding: utf-8 """ NiFi Rest Api The Rest Api provides programmatic access to command and control a NiFi instance in real time. Start and stop processors, monitor queues, query provenance data, and more. Each endpoint below includes a description, ...
[ "six.iteritems" ]
[((3606, 3633), 'six.iteritems', 'iteritems', (["params['kwargs']"], {}), "(params['kwargs'])\n", (3615, 3633), False, 'from six import iteritems\n'), ((8990, 9017), 'six.iteritems', 'iteritems', (["params['kwargs']"], {}), "(params['kwargs'])\n", (8999, 9017), False, 'from six import iteritems\n'), ((15208, 15235), 's...
#!/usr/bin/env python """ Usage: python vtk_reindent_code.py [--test] <file1> [<file2> ...] This script takes old-style "Whitesmiths" indented VTK source files as input, and re-indents the braces according to the new VTK style. Only the brace indentation is modified. If called with the --test option, then it ...
[ "re.compile", "os.path.split", "sys.stderr.write", "sys.exc_info", "sys.exit" ]
[((882, 902), 're.compile', 're.compile', (['"""[/"\']"""'], {}), '(\'[/"\\\']\')\n', (892, 902), False, 'import re\n'), ((958, 1002), 're.compile', 're.compile', (['"""\\\\/\\\\*(\\\\*(?!\\\\/)|[^*])*\\\\*\\\\/"""'], {}), "('\\\\/\\\\*(\\\\*(?!\\\\/)|[^*])*\\\\*\\\\/')\n", (968, 1002), False, 'import re\n'), ((1021, 1...
import os.path from scipy.optimize import fsolve import math import numpy as np from matplotlib import pyplot as plt import pandas as pd import utils_Florian as utils def equations(p, t_peak, t_half): x, y = p return (0.5 * (math.exp(-x * t_peak) - math.exp(-y * t_peak)) - (math.exp(-x * t_half) - math.exp...
[ "scipy.optimize.fsolve", "numpy.abs", "numpy.argmax", "numpy.exp", "numpy.linspace", "numpy.argmin", "pandas.DataFrame", "math.exp" ]
[((404, 418), 'pandas.DataFrame', 'pd.DataFrame', ([], {}), '()\n', (416, 418), True, 'import pandas as pd\n'), ((1088, 1155), 'scipy.optimize.fsolve', 'fsolve', (['equations', 'initial_conditions[alpha]'], {'args': '(t_peak, t_half)'}), '(equations, initial_conditions[alpha], args=(t_peak, t_half))\n', (1094, 1155), F...
import re import string import numpy as np from tqdm import tqdm from typing import List from docqa.triviaqa.read_data import TriviaQaQuestion from docqa.triviaqa.trivia_qa_eval import normalize_answer, f1_score from docqa.utils import flatten_iterable, split """ Tools for turning the aliases and answer strings from...
[ "docqa.triviaqa.trivia_qa_eval.normalize_answer", "docqa.utils.flatten_iterable", "re.compile", "tqdm.tqdm", "trivia_qa.build_span_corpus.TriviaQaWebDataset", "data_processing.text_utils.NltkAndPunctTokenizer", "numpy.array", "numpy.zeros", "multiprocessing.Pool", "docqa.triviaqa.trivia_qa_eval.f1...
[((10899, 10919), 'trivia_qa.build_span_corpus.TriviaQaWebDataset', 'TriviaQaWebDataset', ([], {}), '()\n', (10917, 10919), False, 'from trivia_qa.build_span_corpus import TriviaQaWebDataset\n'), ((5407, 5429), 'docqa.utils.flatten_iterable', 'flatten_iterable', (['para'], {}), '(para)\n', (5423, 5429), False, 'from do...
# Generated by Django 3.2.9 on 2021-11-13 14:43 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('base', '0005_authquery_password'), ] operations = [ migrations.AddField( model_name='profile', name='h...
[ "django.db.models.CharField" ]
[((348, 395), 'django.db.models.CharField', 'models.CharField', ([], {'default': '"""[]"""', 'max_length': '(1000)'}), "(default='[]', max_length=1000)\n", (364, 395), False, 'from django.db import migrations, models\n')]
import nifty.tools as nt import numpy as np import z5py from elf.label_multiset import deserialize_multiset from tqdm import trange def check_serialization(mset1, mset2): if len(mset1) != len(mset2): print("Serialization sizes disagree:", len(mset1), len(mset2)) return False if not np.array_...
[ "numpy.unique", "elf.label_multiset.deserialize_multiset", "nifty.tools.blocking", "z5py.File", "numpy.array_equal", "tqdm.trange" ]
[((1798, 1840), 'nifty.tools.blocking', 'nt.blocking', (['[0, 0, 0]', 'roi_end', '[1, 1, 1]'], {}), '([0, 0, 0], roi_end, [1, 1, 1])\n', (1809, 1840), True, 'import nifty.tools as nt\n'), ((1861, 1892), 'tqdm.trange', 'trange', (['blocking.numberOfBlocks'], {}), '(blocking.numberOfBlocks)\n', (1867, 1892), False, 'from...
import torch import torch.nn as nn import torch.optim as optim from torch.utils.data import DataLoader from torchvision.transforms import ToTensor from torchvision.datasets import MNIST from vis_util import visual_mnist ##### settings x_dim = 28 * 28 # size of mnist digit z_dim = 100 # random noise h_dim = 128 ...
[ "torch.nn.Sigmoid", "torch.nn.ReLU", "torch.randn", "torch.nn.BCELoss", "torch.nn.Linear", "torch.utils.data.DataLoader", "torchvision.transforms.ToTensor", "torch.zeros", "torch.ones" ]
[((548, 605), 'torch.utils.data.DataLoader', 'DataLoader', (['trainset'], {'batch_size': 'batch_size', 'shuffle': '(True)'}), '(trainset, batch_size=batch_size, shuffle=True)\n', (558, 605), False, 'from torch.utils.data import DataLoader\n'), ((686, 711), 'torch.ones', 'torch.ones', (['batch_size', '(1)'], {}), '(batc...
from django.shortcuts import render, redirect, reverse from django.core.paginator import Paginator, PageNotAnInteger, EmptyPage from django.db.models import Q from django.views.generic import View from .models import OrgInfo, CityInfo, TeacherInfo from operations.models import UserLove # Create your views here. clas...
[ "django.shortcuts.render", "django.db.models.Q", "django.core.paginator.Paginator" ]
[((1336, 1358), 'django.core.paginator.Paginator', 'Paginator', (['all_orgs', '(2)'], {}), '(all_orgs, 2)\n', (1345, 1358), False, 'from django.core.paginator import Paginator, PageNotAnInteger, EmptyPage\n'), ((1554, 1763), 'django.shortcuts.render', 'render', (['request', '"""orgs/org-list.html"""', "{'all_orgs': all...
from keras.callbacks import TensorBoard,EarlyStopping,TerminateOnNaN,ReduceLROnPlateau,ModelCheckpoint import os import sys import tensorflow as tf import keras.backend.tensorflow_backend as KTF file_abspath = os.path.abspath(sys.argv[0]) # exe所在文件地址 location = os.path.dirname(file_abspath) # exe所在文件夹目录地址 tbCallB...
[ "keras.callbacks.ModelCheckpoint", "keras.callbacks.ReduceLROnPlateau", "keras.callbacks.TerminateOnNaN", "keras.callbacks.TensorBoard", "os.path.dirname", "keras.callbacks.EarlyStopping", "os.path.abspath" ]
[((213, 241), 'os.path.abspath', 'os.path.abspath', (['sys.argv[0]'], {}), '(sys.argv[0])\n', (228, 241), False, 'import os\n'), ((266, 295), 'os.path.dirname', 'os.path.dirname', (['file_abspath'], {}), '(file_abspath)\n', (281, 295), False, 'import os\n'), ((327, 575), 'keras.callbacks.TensorBoard', 'TensorBoard', ([...
# Generated by Django 3.1.1 on 2020-12-26 20:50 from django.db import migrations class Migration(migrations.Migration): dependencies = [ ('dashboard', '0033_auto_20201226_2148'), ] operations = [ migrations.RenameField( model_name='sleepnight', old_name='data', ...
[ "django.db.migrations.RenameField" ]
[((228, 319), 'django.db.migrations.RenameField', 'migrations.RenameField', ([], {'model_name': '"""sleepnight"""', 'old_name': '"""data"""', 'new_name': '"""diary_day"""'}), "(model_name='sleepnight', old_name='data', new_name=\n 'diary_day')\n", (250, 319), False, 'from django.db import migrations\n')]
"""Fear and Greed Index Class""" __docformat__ = "numpy" from matplotlib import pyplot as plt from fear_greed_index import scrape_cnn from fear_greed_index.FearAndGreedIndicator import FearAndGreedIndicator class CNNFearAndGreedIndex: """CNN Fear and Greed Index Attributes ---------- junk_bond_deman...
[ "matplotlib.pyplot.imshow", "fear_greed_index.FearAndGreedIndicator.FearAndGreedIndicator", "fear_greed_index.scrape_cnn._get_chart", "matplotlib.pyplot.tight_layout", "fear_greed_index.scrape_cnn._get_fear_greed_index" ]
[((1459, 1500), 'fear_greed_index.FearAndGreedIndicator.FearAndGreedIndicator', 'FearAndGreedIndicator', (['"""Junk Bond Demand"""'], {}), "('Junk Bond Demand')\n", (1480, 1500), False, 'from fear_greed_index.FearAndGreedIndicator import FearAndGreedIndicator\n'), ((1534, 1576), 'fear_greed_index.FearAndGreedIndicator....
# -*- coding: utf-8 -*- """ idfy_rest_client.models.person_person_information This file was automatically generated for Idfy by APIMATIC v2.0 ( https://apimatic.io ) """ from idfy_rest_client.api_helper import APIHelper class PersonPersonInformation(object): """Implementation of the 'Person.P...
[ "idfy_rest_client.api_helper.APIHelper.RFC3339DateTime" ]
[((2794, 2825), 'idfy_rest_client.api_helper.APIHelper.RFC3339DateTime', 'APIHelper.RFC3339DateTime', (['dead'], {}), '(dead)\n', (2819, 2825), False, 'from idfy_rest_client.api_helper import APIHelper\n')]
import logging import subprocess # create logger logger = logging.getLogger('simple_example') logger.setLevel(logging.DEBUG) # create console handler and set level to INFO console_logger = logging.StreamHandler() console_logger.setLevel(logging.INFO) # create formatter formatter = logging.Formatter('%(asctime)s - %(...
[ "logging.getLogger", "logging.Formatter", "logging.StreamHandler", "subprocess.call" ]
[((59, 94), 'logging.getLogger', 'logging.getLogger', (['"""simple_example"""'], {}), "('simple_example')\n", (76, 94), False, 'import logging\n'), ((191, 214), 'logging.StreamHandler', 'logging.StreamHandler', ([], {}), '()\n', (212, 214), False, 'import logging\n'), ((285, 347), 'logging.Formatter', 'logging.Formatte...
import numpy as np from pandas import read_csv import pandas as pd from sklearn.model_selection import train_test_split import matplotlib.pyplot as plt from config import * from lib.preprocess.read_data import DataReader from lib.scaler.preprocessing_data.data_normalizer import DataNormalizer class DataPreprocessor:...
[ "lib.scaler.preprocessing_data.data_normalizer.DataNormalizer", "lib.preprocess.read_data.DataReader", "numpy.column_stack", "numpy.array", "numpy.concatenate" ]
[((717, 729), 'lib.preprocess.read_data.DataReader', 'DataReader', ([], {}), '()\n', (727, 729), False, 'from lib.preprocess.read_data import DataReader\n'), ((2432, 2461), 'lib.scaler.preprocessing_data.data_normalizer.DataNormalizer', 'DataNormalizer', (['scaler_method'], {}), '(scaler_method)\n', (2446, 2461), False...
# pylint: disable=arguments-differ,unused-argument,line-too-long """R2Plus1D, implemented in Gluon. https://arxiv.org/abs/1711.11248. Code partially borrowed from https://github.com/pytorch/vision/blob/master/torchvision/models/video/resnet.py.""" __all__ = ['R2Plus1D', 'r2plus1d_resnet18_kinetics400', 'r2...
[ "mxnet.gluon.nn.GlobalAvgPool3D", "mxnet.gluon.nn.HybridSequential", "mxnet.init.Normal", "mxnet.gluon.nn.Dropout", "mxnet.init.MSRAPrelu", "mxnet.context.cpu", "mxnet.gluon.nn.Conv3D", "mxnet.gluon.nn.Activation" ]
[((762, 967), 'mxnet.gluon.nn.Conv3D', 'nn.Conv3D', ([], {'in_channels': 'in_planes', 'channels': 'out_planes', 'kernel_size': '(3, 1, 1)', 'strides': '(temporal_stride, spatial_stride, spatial_stride)', 'padding': '(dilation, 0, 0)', 'dilation': 'dilation', 'use_bias': '(False)'}), '(in_channels=in_planes, channels=ou...
#!/usr/bin/env python3 import retrieve_author_ppn as autppn import retrieve_references as refs import zot_helpers as pyzot from itertools import islice researchers = autppn.constructOutput('test.csv') autppn.writeCsv('out.csv', researchers) for researcher in researchers: ppn = researcher['ppn'] creator_name...
[ "itertools.islice", "retrieve_author_ppn.constructOutput", "retrieve_references.getRefsByRole", "zot_helpers.create_collection", "retrieve_author_ppn.writeCsv", "retrieve_references.getReferences" ]
[((168, 202), 'retrieve_author_ppn.constructOutput', 'autppn.constructOutput', (['"""test.csv"""'], {}), "('test.csv')\n", (190, 202), True, 'import retrieve_author_ppn as autppn\n'), ((204, 243), 'retrieve_author_ppn.writeCsv', 'autppn.writeCsv', (['"""out.csv"""', 'researchers'], {}), "('out.csv', researchers)\n", (2...
#!/usr/bin/env python from pkg_resources import get_distribution from setuptools import setup, find_packages with open("README.md", "r") as f: long_description = f.read() version = get_distribution("autolabel").version setup( packages=find_packages(), install_requires=[ 'click', 'more-ite...
[ "setuptools.find_packages", "pkg_resources.get_distribution" ]
[((187, 216), 'pkg_resources.get_distribution', 'get_distribution', (['"""autolabel"""'], {}), "('autolabel')\n", (203, 216), False, 'from pkg_resources import get_distribution\n'), ((246, 261), 'setuptools.find_packages', 'find_packages', ([], {}), '()\n', (259, 261), False, 'from setuptools import setup, find_package...
import os from docker.models.containers import Container import docker import pytest from tox_docker.config import runas_name def find_container(instance_name: str) -> Container: # TODO: refactor this as a pytest fixture # this is running in a child-process of the tox instance which # spawned the conta...
[ "pytest.fail", "os.getppid", "docker.from_env" ]
[((472, 503), 'docker.from_env', 'docker.from_env', ([], {'version': '"""auto"""'}), "(version='auto')\n", (487, 503), False, 'import docker\n'), ((680, 752), 'pytest.fail', 'pytest.fail', (['f"""No running container with instance name {running_name!r}"""'], {}), "(f'No running container with instance name {running_nam...
#!usr/bin/env python3 # -*- coding: utf-8 -*- # # @author <NAME> # # get posts of a fb page, group or account # returns a json lines files with a line for each post # from argparse import ArgumentParser from datetime import date, datetime from facebook_scraper import get_posts from json import dumps, JSONEncoder clas...
[ "datetime.date.today", "json.dumps", "facebook_scraper.get_posts", "argparse.ArgumentParser" ]
[((1456, 1472), 'argparse.ArgumentParser', 'ArgumentParser', ([], {}), '()\n', (1470, 1472), False, 'from argparse import ArgumentParser\n'), ((1101, 1231), 'facebook_scraper.get_posts', 'get_posts', ([], {'group': 'account', 'cookies': 'cookies', 'pages': 'pages', 'extra_info': '(True)', 'options': "{'comments': comme...
import gi gi.require_version('Gtk', '3.0') from gi.repository import Gtk as gtk class Ked: def __init__(self, app_path, ufile=None): glade_layout = f'{app_path}/data/ked_layout.glade' self.builder = gtk.Builder() self.builder.add_from_file(glade_layout) win = self.builder.get_obje...
[ "gi.repository.Gtk.Builder", "gi.repository.Gtk.main", "gi.require_version" ]
[((11, 43), 'gi.require_version', 'gi.require_version', (['"""Gtk"""', '"""3.0"""'], {}), "('Gtk', '3.0')\n", (29, 43), False, 'import gi\n'), ((536, 546), 'gi.repository.Gtk.main', 'gtk.main', ([], {}), '()\n', (544, 546), True, 'from gi.repository import Gtk as gtk\n'), ((221, 234), 'gi.repository.Gtk.Builder', 'gtk....
#!/usr/bin/env python # vim: set fileencoding=utf-8 ts=4 sts=4 sw=4 et tw=80 : # # Compare an image file and its associated uncertainty image. # # <NAME> # Created: 2021-06-03 # Last modified: 2021-06-03 #-------------------------------------------------------------------------- #*********************************...
[ "logging.basicConfig", "logging.getLogger", "numpy.__version__.split", "numpy.abs", "os.getenv", "matplotlib.pyplot.gcf", "theil_sen.linefit", "imp.reload", "sys.stderr.write", "astropy.io.fits.getdata", "numpy.linspace", "matplotlib.pyplot.figure", "numpy.isnan", "sys.exit", "matplotlib...
[((514, 553), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.INFO'}), '(level=logging.INFO)\n', (533, 553), False, 'import logging\n'), ((563, 590), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (580, 590), False, 'import logging\n'), ((8752, 8782), 'astropy.io.fit...
import logging import os try: import simplejson as json except ImportError: import json from flask import Flask, request, make_response, Response from cStringIO import StringIO import zipfile def get_mocked_server(binary_directory): mocked_cb_server = Flask('cb') files = os.listdir(binary_directory...
[ "flask.request.args.get", "os.path.exists", "cStringIO.StringIO", "os.listdir", "json.loads", "zipfile.ZipFile", "flask.Flask", "json.dumps", "os.path.join", "flask.Response", "os.path.abspath", "flask.make_response" ]
[((268, 279), 'flask.Flask', 'Flask', (['"""cb"""'], {}), "('cb')\n", (273, 279), False, 'from flask import Flask, request, make_response, Response\n'), ((293, 321), 'os.listdir', 'os.listdir', (['binary_directory'], {}), '(binary_directory)\n', (303, 321), False, 'import os\n'), ((3167, 3215), 'os.path.join', 'os.path...
from common.parse_tex import ( BeginDocumentExtractor, BibitemExtractor, DocumentclassExtractor, EquationExtractor, MacroExtractor, PlaintextExtractor, ) from common.types import MacroDefinition from entities.sentences.extractor import SentenceExtractor def test_extract_plaintext_with_newlines...
[ "common.parse_tex.BeginDocumentExtractor", "common.parse_tex.DocumentclassExtractor", "entities.sentences.extractor.SentenceExtractor", "common.types.MacroDefinition", "common.parse_tex.EquationExtractor", "common.parse_tex.BibitemExtractor", "common.parse_tex.PlaintextExtractor", "common.parse_tex.Ma...
[((340, 360), 'common.parse_tex.PlaintextExtractor', 'PlaintextExtractor', ([], {}), '()\n', (358, 360), False, 'from common.parse_tex import BeginDocumentExtractor, BibitemExtractor, DocumentclassExtractor, EquationExtractor, MacroExtractor, PlaintextExtractor\n'), ((1023, 1042), 'entities.sentences.extractor.Sentence...
from webdriver_manager.driver import EdgeDriver, IEDriver from webdriver_manager.manager import DriverManager from webdriver_manager import utils class EdgeDriverManager(DriverManager): def __init__(self, version=None, os_type=utils.os_name()): super(EdgeDriverManager, self).__init__() ...
[ "webdriver_manager.utils.os_type", "webdriver_manager.utils.os_name", "webdriver_manager.driver.IEDriver", "webdriver_manager.driver.EdgeDriver" ]
[((250, 265), 'webdriver_manager.utils.os_name', 'utils.os_name', ([], {}), '()\n', (263, 265), False, 'from webdriver_manager import utils\n'), ((340, 384), 'webdriver_manager.driver.EdgeDriver', 'EdgeDriver', ([], {'version': 'version', 'os_type': 'os_type'}), '(version=version, os_type=os_type)\n', (350, 384), False...
"""Parent class DataN.""" import os import os.path from warnings import warn from typing import Union, NoReturn from pycifstar import Data, to_data from cryspy.A_functions_base.function_1_markdown import md_to_html from cryspy.A_functions_base.function_1_objects import \ get_functions_of_objet, get_table_html_for_...
[ "pycifstar.to_data", "os.path.isfile", "warnings.warn", "cryspy.A_functions_base.function_1_objects.get_table_html_for_variables", "cryspy.A_functions_base.function_1_objects.get_functions_of_objet", "pycifstar.Data" ]
[((13869, 13875), 'pycifstar.Data', 'Data', ([], {}), '()\n', (13873, 13875), False, 'from pycifstar import Data, to_data\n'), ((1332, 1366), 'cryspy.A_functions_base.function_1_objects.get_table_html_for_variables', 'get_table_html_for_variables', (['self'], {}), '(self)\n', (1360, 1366), False, 'from cryspy.A_functio...
import os import unittest def require(f): def skipit(*args, **kwargs): raise unittest.SkipTest('VIVBINS env var...') if os.getenv('VIVBINS') == None: return skipit return f
[ "unittest.SkipTest", "os.getenv" ]
[((90, 129), 'unittest.SkipTest', 'unittest.SkipTest', (['"""VIVBINS env var..."""'], {}), "('VIVBINS env var...')\n", (107, 129), False, 'import unittest\n'), ((138, 158), 'os.getenv', 'os.getenv', (['"""VIVBINS"""'], {}), "('VIVBINS')\n", (147, 158), False, 'import os\n')]
from setuptools import setup from setuptools.command.install import install as _install class Install(_install): def run(self): _install.do_egg_install(self) import nltk nltk.download("popular") setup( cmdclass={'install': Install}, install_requires=['nltk'], setup_requires=[...
[ "setuptools.setup", "nltk.download", "setuptools.command.install.install.do_egg_install" ]
[((227, 319), 'setuptools.setup', 'setup', ([], {'cmdclass': "{'install': Install}", 'install_requires': "['nltk']", 'setup_requires': "['nltk']"}), "(cmdclass={'install': Install}, install_requires=['nltk'],\n setup_requires=['nltk'])\n", (232, 319), False, 'from setuptools import setup\n'), ((142, 171), 'setuptool...