body stringlengths 26 98.2k | body_hash int64 -9,222,864,604,528,158,000 9,221,803,474B | docstring stringlengths 1 16.8k | path stringlengths 5 230 | name stringlengths 1 96 | repository_name stringlengths 7 89 | lang stringclasses 1
value | body_without_docstring stringlengths 20 98.2k |
|---|---|---|---|---|---|---|---|
def assign_only_product_as_production(db):
"Assign only product as reference product.\n\n Skips datasets that already have a reference product or no production exchanges. Production exchanges must have a ``name`` and an amount.\n\n Will replace the following activity fields, if not already specified:\n\n *... | -1,332,683,319,952,914,000 | Assign only product as reference product.
Skips datasets that already have a reference product or no production exchanges. Production exchanges must have a ``name`` and an amount.
Will replace the following activity fields, if not already specified:
* 'name' - name of reference product
* 'unit' - unit of reference p... | bw2io/strategies/generic.py | assign_only_product_as_production | pjamesjoyce/brightway2-io | python | def assign_only_product_as_production(db):
"Assign only product as reference product.\n\n Skips datasets that already have a reference product or no production exchanges. Production exchanges must have a ``name`` and an amount.\n\n Will replace the following activity fields, if not already specified:\n\n *... |
def link_technosphere_by_activity_hash(db, external_db_name=None, fields=None):
'Link technosphere exchanges using ``activity_hash`` function.\n\n If ``external_db_name``, link against a different database; otherwise link internally.\n\n If ``fields``, link using only certain fields.'
TECHNOSPHERE_TYPES =... | 4,406,278,707,064,293,400 | Link technosphere exchanges using ``activity_hash`` function.
If ``external_db_name``, link against a different database; otherwise link internally.
If ``fields``, link using only certain fields. | bw2io/strategies/generic.py | link_technosphere_by_activity_hash | pjamesjoyce/brightway2-io | python | def link_technosphere_by_activity_hash(db, external_db_name=None, fields=None):
'Link technosphere exchanges using ``activity_hash`` function.\n\n If ``external_db_name``, link against a different database; otherwise link internally.\n\n If ``fields``, link using only certain fields.'
TECHNOSPHERE_TYPES =... |
def set_code_by_activity_hash(db, overwrite=False):
"Use ``activity_hash`` to set dataset code.\n\n By default, won't overwrite existing codes, but will if ``overwrite`` is ``True``."
for ds in db:
if (('code' not in ds) or overwrite):
ds['code'] = activity_hash(ds)
return db | 8,632,796,420,951,601,000 | Use ``activity_hash`` to set dataset code.
By default, won't overwrite existing codes, but will if ``overwrite`` is ``True``. | bw2io/strategies/generic.py | set_code_by_activity_hash | pjamesjoyce/brightway2-io | python | def set_code_by_activity_hash(db, overwrite=False):
"Use ``activity_hash`` to set dataset code.\n\n By default, won't overwrite existing codes, but will if ``overwrite`` is ``True``."
for ds in db:
if (('code' not in ds) or overwrite):
ds['code'] = activity_hash(ds)
return db |
def drop_unlinked(db):
'This is the nuclear option - use at your own risk!'
for ds in db:
ds['exchanges'] = [obj for obj in ds['exchanges'] if obj.get('input')]
return db | 127,509,395,245,177,260 | This is the nuclear option - use at your own risk! | bw2io/strategies/generic.py | drop_unlinked | pjamesjoyce/brightway2-io | python | def drop_unlinked(db):
for ds in db:
ds['exchanges'] = [obj for obj in ds['exchanges'] if obj.get('input')]
return db |
def normalize_units(db):
'Normalize units in datasets and their exchanges'
for ds in db:
if ('unit' in ds):
ds['unit'] = normalize_units_function(ds['unit'])
for exc in ds.get('exchanges', []):
if ('unit' in exc):
exc['unit'] = normalize_units_function(exc... | 8,029,459,837,467,838 | Normalize units in datasets and their exchanges | bw2io/strategies/generic.py | normalize_units | pjamesjoyce/brightway2-io | python | def normalize_units(db):
for ds in db:
if ('unit' in ds):
ds['unit'] = normalize_units_function(ds['unit'])
for exc in ds.get('exchanges', []):
if ('unit' in exc):
exc['unit'] = normalize_units_function(exc['unit'])
for param in ds.get('parameters... |
def add_database_name(db, name):
'Add database name to datasets'
for ds in db:
ds['database'] = name
return db | 3,011,635,110,219,210,000 | Add database name to datasets | bw2io/strategies/generic.py | add_database_name | pjamesjoyce/brightway2-io | python | def add_database_name(db, name):
for ds in db:
ds['database'] = name
return db |
def convert_uncertainty_types_to_integers(db):
'Generic number conversion function convert to floats. Return to integers.'
for ds in db:
for exc in ds['exchanges']:
try:
exc['uncertainty type'] = int(exc['uncertainty type'])
except:
pass
return... | 7,904,900,524,443,326,000 | Generic number conversion function convert to floats. Return to integers. | bw2io/strategies/generic.py | convert_uncertainty_types_to_integers | pjamesjoyce/brightway2-io | python | def convert_uncertainty_types_to_integers(db):
for ds in db:
for exc in ds['exchanges']:
try:
exc['uncertainty type'] = int(exc['uncertainty type'])
except:
pass
return db |
def drop_falsey_uncertainty_fields_but_keep_zeros(db):
"Drop fields like '' but keep zero and NaN.\n\n Note that this doesn't strip `False`, which behaves *exactly* like 0.\n\n "
uncertainty_fields = ['minimum', 'maximum', 'scale', 'shape', 'loc']
def drop_if_appropriate(exc):
for field in un... | -8,108,741,207,312,171,000 | Drop fields like '' but keep zero and NaN.
Note that this doesn't strip `False`, which behaves *exactly* like 0. | bw2io/strategies/generic.py | drop_falsey_uncertainty_fields_but_keep_zeros | pjamesjoyce/brightway2-io | python | def drop_falsey_uncertainty_fields_but_keep_zeros(db):
"Drop fields like but keep zero and NaN.\n\n Note that this doesn't strip `False`, which behaves *exactly* like 0.\n\n "
uncertainty_fields = ['minimum', 'maximum', 'scale', 'shape', 'loc']
def drop_if_appropriate(exc):
for field in unce... |
def convert_activity_parameters_to_list(data):
'Convert activity parameters from dictionary to list of dictionaries'
def _(key, value):
dct = deepcopy(value)
dct['name'] = key
return dct
for ds in data:
if ('parameters' in ds):
ds['parameters'] = [_(x, y) for (x,... | -8,496,173,251,939,245,000 | Convert activity parameters from dictionary to list of dictionaries | bw2io/strategies/generic.py | convert_activity_parameters_to_list | pjamesjoyce/brightway2-io | python | def convert_activity_parameters_to_list(data):
def _(key, value):
dct = deepcopy(value)
dct['name'] = key
return dct
for ds in data:
if ('parameters' in ds):
ds['parameters'] = [_(x, y) for (x, y) in ds['parameters'].items()]
return data |
@property
def Ospfv3PseudoInterface(self):
'\n Returns\n -------\n - obj(uhd_restpy.testplatform.sessions.ixnetwork.topology.ospfv3pseudointerface_bbc932877888c8c8400661ec299754a8.Ospfv3PseudoInterface): An instance of the Ospfv3PseudoInterface class\n\n Raises\n ------\n -... | -7,806,945,575,050,585,000 | Returns
-------
- obj(uhd_restpy.testplatform.sessions.ixnetwork.topology.ospfv3pseudointerface_bbc932877888c8c8400661ec299754a8.Ospfv3PseudoInterface): An instance of the Ospfv3PseudoInterface class
Raises
------
- ServerError: The server has encountered an uncategorized error condition | uhd_restpy/testplatform/sessions/ixnetwork/topology/siminterfaceipv6config_189f3bfbc365f2b105e35cd8b9d542d6.py | Ospfv3PseudoInterface | rfrye-github/ixnetwork_restpy | python | @property
def Ospfv3PseudoInterface(self):
'\n Returns\n -------\n - obj(uhd_restpy.testplatform.sessions.ixnetwork.topology.ospfv3pseudointerface_bbc932877888c8c8400661ec299754a8.Ospfv3PseudoInterface): An instance of the Ospfv3PseudoInterface class\n\n Raises\n ------\n -... |
@property
def Count(self):
'\n Returns\n -------\n - number: Number of elements inside associated multiplier-scaled container object, e.g. number of devices inside a Device Group.\n '
return self._get_attribute(self._SDM_ATT_MAP['Count']) | 9,202,294,428,103,448,000 | Returns
-------
- number: Number of elements inside associated multiplier-scaled container object, e.g. number of devices inside a Device Group. | uhd_restpy/testplatform/sessions/ixnetwork/topology/siminterfaceipv6config_189f3bfbc365f2b105e35cd8b9d542d6.py | Count | rfrye-github/ixnetwork_restpy | python | @property
def Count(self):
'\n Returns\n -------\n - number: Number of elements inside associated multiplier-scaled container object, e.g. number of devices inside a Device Group.\n '
return self._get_attribute(self._SDM_ATT_MAP['Count']) |
@property
def DescriptiveName(self):
"\n Returns\n -------\n - str: Longer, more descriptive name for element. It's not guaranteed to be unique like -name-, but may offer more context.\n "
return self._get_attribute(self._SDM_ATT_MAP['DescriptiveName']) | 6,335,322,004,352,822,000 | Returns
-------
- str: Longer, more descriptive name for element. It's not guaranteed to be unique like -name-, but may offer more context. | uhd_restpy/testplatform/sessions/ixnetwork/topology/siminterfaceipv6config_189f3bfbc365f2b105e35cd8b9d542d6.py | DescriptiveName | rfrye-github/ixnetwork_restpy | python | @property
def DescriptiveName(self):
"\n Returns\n -------\n - str: Longer, more descriptive name for element. It's not guaranteed to be unique like -name-, but may offer more context.\n "
return self._get_attribute(self._SDM_ATT_MAP['DescriptiveName']) |
@property
def EnableIp(self):
'\n Returns\n -------\n - obj(uhd_restpy.multivalue.Multivalue): Enable IPv6\n '
from uhd_restpy.multivalue import Multivalue
return Multivalue(self, self._get_attribute(self._SDM_ATT_MAP['EnableIp'])) | 1,591,239,963,546,353,000 | Returns
-------
- obj(uhd_restpy.multivalue.Multivalue): Enable IPv6 | uhd_restpy/testplatform/sessions/ixnetwork/topology/siminterfaceipv6config_189f3bfbc365f2b105e35cd8b9d542d6.py | EnableIp | rfrye-github/ixnetwork_restpy | python | @property
def EnableIp(self):
'\n Returns\n -------\n - obj(uhd_restpy.multivalue.Multivalue): Enable IPv6\n '
from uhd_restpy.multivalue import Multivalue
return Multivalue(self, self._get_attribute(self._SDM_ATT_MAP['EnableIp'])) |
@property
def FromIP(self):
'\n Returns\n -------\n - obj(uhd_restpy.multivalue.Multivalue): 128 Bits IPv6 address.\n '
from uhd_restpy.multivalue import Multivalue
return Multivalue(self, self._get_attribute(self._SDM_ATT_MAP['FromIP'])) | 619,688,722,874,091,400 | Returns
-------
- obj(uhd_restpy.multivalue.Multivalue): 128 Bits IPv6 address. | uhd_restpy/testplatform/sessions/ixnetwork/topology/siminterfaceipv6config_189f3bfbc365f2b105e35cd8b9d542d6.py | FromIP | rfrye-github/ixnetwork_restpy | python | @property
def FromIP(self):
'\n Returns\n -------\n - obj(uhd_restpy.multivalue.Multivalue): 128 Bits IPv6 address.\n '
from uhd_restpy.multivalue import Multivalue
return Multivalue(self, self._get_attribute(self._SDM_ATT_MAP['FromIP'])) |
@property
def Name(self):
'\n Returns\n -------\n - str: Name of NGPF element, guaranteed to be unique in Scenario\n '
return self._get_attribute(self._SDM_ATT_MAP['Name']) | -1,824,082,867,023,513,900 | Returns
-------
- str: Name of NGPF element, guaranteed to be unique in Scenario | uhd_restpy/testplatform/sessions/ixnetwork/topology/siminterfaceipv6config_189f3bfbc365f2b105e35cd8b9d542d6.py | Name | rfrye-github/ixnetwork_restpy | python | @property
def Name(self):
'\n Returns\n -------\n - str: Name of NGPF element, guaranteed to be unique in Scenario\n '
return self._get_attribute(self._SDM_ATT_MAP['Name']) |
@property
def SubnetPrefixLength(self):
'\n Returns\n -------\n - obj(uhd_restpy.multivalue.Multivalue): Subnet Prefix Length\n '
from uhd_restpy.multivalue import Multivalue
return Multivalue(self, self._get_attribute(self._SDM_ATT_MAP['SubnetPrefixLength'])) | -8,855,745,252,700,621,000 | Returns
-------
- obj(uhd_restpy.multivalue.Multivalue): Subnet Prefix Length | uhd_restpy/testplatform/sessions/ixnetwork/topology/siminterfaceipv6config_189f3bfbc365f2b105e35cd8b9d542d6.py | SubnetPrefixLength | rfrye-github/ixnetwork_restpy | python | @property
def SubnetPrefixLength(self):
'\n Returns\n -------\n - obj(uhd_restpy.multivalue.Multivalue): Subnet Prefix Length\n '
from uhd_restpy.multivalue import Multivalue
return Multivalue(self, self._get_attribute(self._SDM_ATT_MAP['SubnetPrefixLength'])) |
@property
def ToIP(self):
'\n Returns\n -------\n - obj(uhd_restpy.multivalue.Multivalue): 128 Bits IPv6 address.\n '
from uhd_restpy.multivalue import Multivalue
return Multivalue(self, self._get_attribute(self._SDM_ATT_MAP['ToIP'])) | -1,024,568,113,818,389,900 | Returns
-------
- obj(uhd_restpy.multivalue.Multivalue): 128 Bits IPv6 address. | uhd_restpy/testplatform/sessions/ixnetwork/topology/siminterfaceipv6config_189f3bfbc365f2b105e35cd8b9d542d6.py | ToIP | rfrye-github/ixnetwork_restpy | python | @property
def ToIP(self):
'\n Returns\n -------\n - obj(uhd_restpy.multivalue.Multivalue): 128 Bits IPv6 address.\n '
from uhd_restpy.multivalue import Multivalue
return Multivalue(self, self._get_attribute(self._SDM_ATT_MAP['ToIP'])) |
def update(self, Name=None):
'Updates simInterfaceIPv6Config resource on the server.\n\n This method has some named parameters with a type: obj (Multivalue).\n The Multivalue class has documentation that details the possible values for those named parameters.\n\n Args\n ----\n - N... | 8,011,544,818,928,487,000 | Updates simInterfaceIPv6Config resource on the server.
This method has some named parameters with a type: obj (Multivalue).
The Multivalue class has documentation that details the possible values for those named parameters.
Args
----
- Name (str): Name of NGPF element, guaranteed to be unique in Scenario
Raises
----... | uhd_restpy/testplatform/sessions/ixnetwork/topology/siminterfaceipv6config_189f3bfbc365f2b105e35cd8b9d542d6.py | update | rfrye-github/ixnetwork_restpy | python | def update(self, Name=None):
'Updates simInterfaceIPv6Config resource on the server.\n\n This method has some named parameters with a type: obj (Multivalue).\n The Multivalue class has documentation that details the possible values for those named parameters.\n\n Args\n ----\n - N... |
def find(self, Count=None, DescriptiveName=None, Name=None):
"Finds and retrieves simInterfaceIPv6Config resources from the server.\n\n All named parameters are evaluated on the server using regex. The named parameters can be used to selectively retrieve simInterfaceIPv6Config resources from the server.\n ... | -1,710,958,394,887,526,400 | Finds and retrieves simInterfaceIPv6Config resources from the server.
All named parameters are evaluated on the server using regex. The named parameters can be used to selectively retrieve simInterfaceIPv6Config resources from the server.
To retrieve an exact match ensure the parameter value starts with ^ and ends wit... | uhd_restpy/testplatform/sessions/ixnetwork/topology/siminterfaceipv6config_189f3bfbc365f2b105e35cd8b9d542d6.py | find | rfrye-github/ixnetwork_restpy | python | def find(self, Count=None, DescriptiveName=None, Name=None):
"Finds and retrieves simInterfaceIPv6Config resources from the server.\n\n All named parameters are evaluated on the server using regex. The named parameters can be used to selectively retrieve simInterfaceIPv6Config resources from the server.\n ... |
def read(self, href):
'Retrieves a single instance of simInterfaceIPv6Config data from the server.\n\n Args\n ----\n - href (str): An href to the instance to be retrieved\n\n Returns\n -------\n - self: This instance with the simInterfaceIPv6Config resources from the server... | -3,705,495,493,069,741,000 | Retrieves a single instance of simInterfaceIPv6Config data from the server.
Args
----
- href (str): An href to the instance to be retrieved
Returns
-------
- self: This instance with the simInterfaceIPv6Config resources from the server available through an iterator or index
Raises
------
- NotFoundError: The request... | uhd_restpy/testplatform/sessions/ixnetwork/topology/siminterfaceipv6config_189f3bfbc365f2b105e35cd8b9d542d6.py | read | rfrye-github/ixnetwork_restpy | python | def read(self, href):
'Retrieves a single instance of simInterfaceIPv6Config data from the server.\n\n Args\n ----\n - href (str): An href to the instance to be retrieved\n\n Returns\n -------\n - self: This instance with the simInterfaceIPv6Config resources from the server... |
def get_device_ids(self, PortNames=None, EnableIp=None, FromIP=None, SubnetPrefixLength=None, ToIP=None):
'Base class infrastructure that gets a list of simInterfaceIPv6Config device ids encapsulated by this object.\n\n Use the optional regex parameters in the method to refine the list of device ids encapsul... | -2,168,971,069,518,069,800 | Base class infrastructure that gets a list of simInterfaceIPv6Config device ids encapsulated by this object.
Use the optional regex parameters in the method to refine the list of device ids encapsulated by this object.
Args
----
- PortNames (str): optional regex of port names
- EnableIp (str): optional regex of enabl... | uhd_restpy/testplatform/sessions/ixnetwork/topology/siminterfaceipv6config_189f3bfbc365f2b105e35cd8b9d542d6.py | get_device_ids | rfrye-github/ixnetwork_restpy | python | def get_device_ids(self, PortNames=None, EnableIp=None, FromIP=None, SubnetPrefixLength=None, ToIP=None):
'Base class infrastructure that gets a list of simInterfaceIPv6Config device ids encapsulated by this object.\n\n Use the optional regex parameters in the method to refine the list of device ids encapsul... |
def Abort(self):
'Executes the abort operation on the server.\n\n Abort CPF control plane (equals to demote to kUnconfigured state).\n\n Raises\n ------\n - NotFoundError: The requested resource does not exist on the server\n - ServerError: The server has encountered an uncategori... | 2,809,284,878,408,633,000 | Executes the abort operation on the server.
Abort CPF control plane (equals to demote to kUnconfigured state).
Raises
------
- NotFoundError: The requested resource does not exist on the server
- ServerError: The server has encountered an uncategorized error condition | uhd_restpy/testplatform/sessions/ixnetwork/topology/siminterfaceipv6config_189f3bfbc365f2b105e35cd8b9d542d6.py | Abort | rfrye-github/ixnetwork_restpy | python | def Abort(self):
'Executes the abort operation on the server.\n\n Abort CPF control plane (equals to demote to kUnconfigured state).\n\n Raises\n ------\n - NotFoundError: The requested resource does not exist on the server\n - ServerError: The server has encountered an uncategori... |
def Start(self):
'Executes the start operation on the server.\n\n Start CPF control plane (equals to promote to negotiated state).\n\n Raises\n ------\n - NotFoundError: The requested resource does not exist on the server\n - ServerError: The server has encountered an uncategorize... | 6,849,918,185,653,348,000 | Executes the start operation on the server.
Start CPF control plane (equals to promote to negotiated state).
Raises
------
- NotFoundError: The requested resource does not exist on the server
- ServerError: The server has encountered an uncategorized error condition | uhd_restpy/testplatform/sessions/ixnetwork/topology/siminterfaceipv6config_189f3bfbc365f2b105e35cd8b9d542d6.py | Start | rfrye-github/ixnetwork_restpy | python | def Start(self):
'Executes the start operation on the server.\n\n Start CPF control plane (equals to promote to negotiated state).\n\n Raises\n ------\n - NotFoundError: The requested resource does not exist on the server\n - ServerError: The server has encountered an uncategorize... |
def Stop(self):
'Executes the stop operation on the server.\n\n Stop CPF control plane (equals to demote to PreValidated-DoDDone state).\n\n Raises\n ------\n - NotFoundError: The requested resource does not exist on the server\n - ServerError: The server has encountered an uncate... | 328,731,283,898,110,100 | Executes the stop operation on the server.
Stop CPF control plane (equals to demote to PreValidated-DoDDone state).
Raises
------
- NotFoundError: The requested resource does not exist on the server
- ServerError: The server has encountered an uncategorized error condition | uhd_restpy/testplatform/sessions/ixnetwork/topology/siminterfaceipv6config_189f3bfbc365f2b105e35cd8b9d542d6.py | Stop | rfrye-github/ixnetwork_restpy | python | def Stop(self):
'Executes the stop operation on the server.\n\n Stop CPF control plane (equals to demote to PreValidated-DoDDone state).\n\n Raises\n ------\n - NotFoundError: The requested resource does not exist on the server\n - ServerError: The server has encountered an uncate... |
@logargs
@ValidateAndDefault({'account_name': (OptionalValueType(StrType), None), 'account_id': (OptionalValueType(SolidFireIDType), None), 'by_id': (BoolType, False), 'mvip': (IPv4AddressType, sfdefaults.mvip), 'username': (StrType, sfdefaults.username), 'password': (StrType, sfdefaults.password), 'output_format': (Op... | -4,768,445,726,859,416,000 | Show the list of volumes for an account
Args:
account_name: the name of the account
account_id: the ID of the account
by_id: show volume IDs instead of names
mvip: the management IP of the cluster
username: the admin user of the cluster
pa... | account_list_volumes.py | AccountListVolumes | cseelye/sfauto | python | @logargs
@ValidateAndDefault({'account_name': (OptionalValueType(StrType), None), 'account_id': (OptionalValueType(SolidFireIDType), None), 'by_id': (BoolType, False), 'mvip': (IPv4AddressType, sfdefaults.mvip), 'username': (StrType, sfdefaults.username), 'password': (StrType, sfdefaults.password), 'output_format': (Op... |
@cached_property
def openapi_types():
'\n This must be a method because a model may have properties that are\n of type self, this must run after the class is loaded\n\n Returns\n openapi_types (dict): The key is attribute name\n and the value is attribute type.\n ... | 4,470,210,284,486,104,600 | This must be a method because a model may have properties that are
of type self, this must run after the class is loaded
Returns
openapi_types (dict): The key is attribute name
and the value is attribute type. | intersight/model/storage_net_app_sensor_all_of.py | openapi_types | CiscoDevNet/intersight-python | python | @cached_property
def openapi_types():
'\n This must be a method because a model may have properties that are\n of type self, this must run after the class is loaded\n\n Returns\n openapi_types (dict): The key is attribute name\n and the value is attribute type.\n ... |
@convert_js_args_to_python_args
def __init__(self, *args, **kwargs):
'StorageNetAppSensorAllOf - a model defined in OpenAPI\n\n Args:\n\n Keyword Args:\n class_id (str): The fully-qualified name of the instantiated, concrete type. This property is used as a discriminator to identify the typ... | -5,255,135,685,495,371,000 | StorageNetAppSensorAllOf - a model defined in OpenAPI
Args:
Keyword Args:
class_id (str): The fully-qualified name of the instantiated, concrete type. This property is used as a discriminator to identify the type of the payload when marshaling and unmarshaling data.. defaults to "storage.NetAppSensor", must be on... | intersight/model/storage_net_app_sensor_all_of.py | __init__ | CiscoDevNet/intersight-python | python | @convert_js_args_to_python_args
def __init__(self, *args, **kwargs):
'StorageNetAppSensorAllOf - a model defined in OpenAPI\n\n Args:\n\n Keyword Args:\n class_id (str): The fully-qualified name of the instantiated, concrete type. This property is used as a discriminator to identify the typ... |
@export
def dbcluster_events(data: Any, cluster_columns: List[Any]=None, verbose: bool=False, normalize: bool=True, time_column: str='TimeCreatedUtc', max_cluster_distance: float=0.01, min_cluster_samples: int=2, **kwargs) -> Tuple[(pd.DataFrame, DBSCAN, np.ndarray)]:
"\n Cluster data set according to cluster_co... | -3,809,816,684,162,627,600 | Cluster data set according to cluster_columns features.
Parameters
----------
data : Any
Input data as a pandas DataFrame or numpy array
cluster_columns : List[Any], optional
List of columns to use for features
- for DataFrame this is a list of column names
- for numpy array this is a list of column in... | msticpy/analysis/eventcluster.py | dbcluster_events | 2xyo/msticpy | python | @export
def dbcluster_events(data: Any, cluster_columns: List[Any]=None, verbose: bool=False, normalize: bool=True, time_column: str='TimeCreatedUtc', max_cluster_distance: float=0.01, min_cluster_samples: int=2, **kwargs) -> Tuple[(pd.DataFrame, DBSCAN, np.ndarray)]:
"\n Cluster data set according to cluster_co... |
def _merge_clustered_items(cluster_set: np.array, labels: np.array, data: Union[(pd.DataFrame, np.array)], time_column: str, counts: np.array) -> pd.DataFrame:
'\n Merge outliers and core clusters into single DataFrame.\n\n Parameters\n ----------\n cluster_set : np.array\n The set of clusters\n ... | 413,818,286,560,348,900 | Merge outliers and core clusters into single DataFrame.
Parameters
----------
cluster_set : np.array
The set of clusters
labels : np.array
The cluster labels
data : Union[pd.DataFrame, np.array]
The source data
time_column : str
Name of the Time column
counts : np.array
The counts of members in eac... | msticpy/analysis/eventcluster.py | _merge_clustered_items | 2xyo/msticpy | python | def _merge_clustered_items(cluster_set: np.array, labels: np.array, data: Union[(pd.DataFrame, np.array)], time_column: str, counts: np.array) -> pd.DataFrame:
'\n Merge outliers and core clusters into single DataFrame.\n\n Parameters\n ----------\n cluster_set : np.array\n The set of clusters\n ... |
@export
def add_process_features(input_frame: pd.DataFrame, path_separator: str=None, force: bool=False) -> pd.DataFrame:
'\n Add numerical features based on patterns of command line and process name.\n\n Parameters\n ----------\n input_frame : pd.DataFrame\n The input dataframe\n path_separat... | -5,266,310,471,347,284,000 | Add numerical features based on patterns of command line and process name.
Parameters
----------
input_frame : pd.DataFrame
The input dataframe
path_separator : str, optional
Path separator. If not supplied, try to determine
from 'NewProcessName' column of first 10 rows
(the default is None)
force : bo... | msticpy/analysis/eventcluster.py | add_process_features | 2xyo/msticpy | python | @export
def add_process_features(input_frame: pd.DataFrame, path_separator: str=None, force: bool=False) -> pd.DataFrame:
'\n Add numerical features based on patterns of command line and process name.\n\n Parameters\n ----------\n input_frame : pd.DataFrame\n The input dataframe\n path_separat... |
def _add_processname_features(output_df: pd.DataFrame, force: bool, path_separator: str):
'\n Add process name default features.\n\n Parameters\n ----------\n output_df : pd.DataFrame\n The dataframe to add features to\n force : bool\n If True overwrite existing feature columns\n pat... | -2,476,954,572,613,136,000 | Add process name default features.
Parameters
----------
output_df : pd.DataFrame
The dataframe to add features to
force : bool
If True overwrite existing feature columns
path_separator : str
Path separator for OS | msticpy/analysis/eventcluster.py | _add_processname_features | 2xyo/msticpy | python | def _add_processname_features(output_df: pd.DataFrame, force: bool, path_separator: str):
'\n Add process name default features.\n\n Parameters\n ----------\n output_df : pd.DataFrame\n The dataframe to add features to\n force : bool\n If True overwrite existing feature columns\n pat... |
def _add_commandline_features(output_df: pd.DataFrame, force: bool):
'\n Add commandline default features.\n\n Parameters\n ----------\n output_df : pd.DataFrame\n The dataframe to add features to\n force : bool\n If True overwrite existing feature columns\n\n '
if (('commandline... | 9,055,919,210,541,722,000 | Add commandline default features.
Parameters
----------
output_df : pd.DataFrame
The dataframe to add features to
force : bool
If True overwrite existing feature columns | msticpy/analysis/eventcluster.py | _add_commandline_features | 2xyo/msticpy | python | def _add_commandline_features(output_df: pd.DataFrame, force: bool):
'\n Add commandline default features.\n\n Parameters\n ----------\n output_df : pd.DataFrame\n The dataframe to add features to\n force : bool\n If True overwrite existing feature columns\n\n '
if (('commandline... |
@export
@lru_cache(maxsize=1024)
def delim_count(value: str, delim_list: str='[\\s\\-\\\\/\\.,"\\\'|&:;%$()]') -> int:
'\n Count the delimiters in input column.\n\n Parameters\n ----------\n value : str\n Data to process\n delim_list : str, optional\n delimiters to use. (the default is ... | -5,264,490,467,521,285,000 | Count the delimiters in input column.
Parameters
----------
value : str
Data to process
delim_list : str, optional
delimiters to use. (the default is r'[\\s\\\\-\\\\\\\\/\.,"\\\\'|&:;%$()]')
Returns
-------
int
Count of delimiters in the string. | msticpy/analysis/eventcluster.py | delim_count | 2xyo/msticpy | python | @export
@lru_cache(maxsize=1024)
def delim_count(value: str, delim_list: str='[\\s\\-\\\\/\\.,"\\\'|&:;%$()]') -> int:
'\n Count the delimiters in input column.\n\n Parameters\n ----------\n value : str\n Data to process\n delim_list : str, optional\n delimiters to use. (the default is ... |
@export
@lru_cache(maxsize=1024)
def delim_hash(value: str, delim_list: str='[\\s\\-\\\\/\\.,"\\\'|&:;%$()]') -> int:
'\n Return a hash (CRC32) of the delimiters from input column.\n\n Parameters\n ----------\n value : str\n Data to process\n delim_list : str, optional\n delimiters to u... | -8,977,949,746,261,767,000 | Return a hash (CRC32) of the delimiters from input column.
Parameters
----------
value : str
Data to process
delim_list : str, optional
delimiters to use. (the default is r'[\\s\\\\-\\\\\\\\/\.,"\\\\'|&:;%$()]')
Returns
-------
int
Hash of delimiter set in the string. | msticpy/analysis/eventcluster.py | delim_hash | 2xyo/msticpy | python | @export
@lru_cache(maxsize=1024)
def delim_hash(value: str, delim_list: str='[\\s\\-\\\\/\\.,"\\\'|&:;%$()]') -> int:
'\n Return a hash (CRC32) of the delimiters from input column.\n\n Parameters\n ----------\n value : str\n Data to process\n delim_list : str, optional\n delimiters to u... |
@export
@lru_cache(maxsize=1024)
def char_ord_score(value: str, scale: int=1) -> int:
'\n Return sum of ord values of characters in string.\n\n Parameters\n ----------\n value : str\n Data to process\n scale : int, optional\n reduce the scale of the feature (reducing the\n influe... | -2,230,151,292,892,541,200 | Return sum of ord values of characters in string.
Parameters
----------
value : str
Data to process
scale : int, optional
reduce the scale of the feature (reducing the
influence of variations this feature on the clustering
algorithm (the default is 1)
Returns
-------
int
[description]
Notes
-----... | msticpy/analysis/eventcluster.py | char_ord_score | 2xyo/msticpy | python | @export
@lru_cache(maxsize=1024)
def char_ord_score(value: str, scale: int=1) -> int:
'\n Return sum of ord values of characters in string.\n\n Parameters\n ----------\n value : str\n Data to process\n scale : int, optional\n reduce the scale of the feature (reducing the\n influe... |
@export
@lru_cache(maxsize=1024)
def token_count(value: str, delimiter: str=' ') -> int:
"\n Return count of delimiter-separated tokens pd.Series column.\n\n Parameters\n ----------\n value : str\n Data to process\n delimiter : str, optional\n Delimiter used to split the column string.\... | -6,788,009,994,740,111,000 | Return count of delimiter-separated tokens pd.Series column.
Parameters
----------
value : str
Data to process
delimiter : str, optional
Delimiter used to split the column string.
(the default is ' ')
Returns
-------
int
count of tokens | msticpy/analysis/eventcluster.py | token_count | 2xyo/msticpy | python | @export
@lru_cache(maxsize=1024)
def token_count(value: str, delimiter: str=' ') -> int:
"\n Return count of delimiter-separated tokens pd.Series column.\n\n Parameters\n ----------\n value : str\n Data to process\n delimiter : str, optional\n Delimiter used to split the column string.\... |
def _string_score(input_str):
'Sum the ord(c) for characters in a string.'
return sum((ord(x) for x in input_str)) | 2,873,044,902,948,387,000 | Sum the ord(c) for characters in a string. | msticpy/analysis/eventcluster.py | _string_score | 2xyo/msticpy | python | def _string_score(input_str):
return sum((ord(x) for x in input_str)) |
@export
@lru_cache(maxsize=1024)
def crc32_hash(value: str) -> int:
'\n Return the CRC32 hash of the input column.\n\n Parameters\n ----------\n value : str\n Data to process\n\n Returns\n -------\n int\n CRC32 hash\n\n '
return crc32(bytes(value.encode('utf-8'))) | 4,065,869,139,596,183,000 | Return the CRC32 hash of the input column.
Parameters
----------
value : str
Data to process
Returns
-------
int
CRC32 hash | msticpy/analysis/eventcluster.py | crc32_hash | 2xyo/msticpy | python | @export
@lru_cache(maxsize=1024)
def crc32_hash(value: str) -> int:
'\n Return the CRC32 hash of the input column.\n\n Parameters\n ----------\n value : str\n Data to process\n\n Returns\n -------\n int\n CRC32 hash\n\n '
return crc32(bytes(value.encode('utf-8'))) |
@export
def delim_count_df(data: pd.DataFrame, column: str, delim_list: str='[\\s\\-\\\\/\\.,"\\\'|&:;%$()]') -> pd.Series:
'\n Count the delimiters in input column.\n\n Parameters\n ----------\n data : pd.DataFrame\n The DataFrame to process\n column : str\n The name of the column to p... | 1,464,786,969,080,396,500 | Count the delimiters in input column.
Parameters
----------
data : pd.DataFrame
The DataFrame to process
column : str
The name of the column to process
delim_list : str, optional
delimiters to use. (the default is r\'[\\s\\\\-\\\\\\\\/\.,"\\\\'|&:;%$()]\')
Returns
-------
pd.Series
Count of delimiters... | msticpy/analysis/eventcluster.py | delim_count_df | 2xyo/msticpy | python | @export
def delim_count_df(data: pd.DataFrame, column: str, delim_list: str='[\\s\\-\\\\/\\.,"\\\'|&:;%$()]') -> pd.Series:
'\n Count the delimiters in input column.\n\n Parameters\n ----------\n data : pd.DataFrame\n The DataFrame to process\n column : str\n The name of the column to p... |
@export
def char_ord_score_df(data: pd.DataFrame, column: str, scale: int=1) -> pd.Series:
'\n Return sum of ord values of characters in string.\n\n Parameters\n ----------\n data : pd.DataFrame\n The DataFrame to process\n column : str\n Column name to process\n scale : int, optiona... | 8,791,648,072,038,837,000 | Return sum of ord values of characters in string.
Parameters
----------
data : pd.DataFrame
The DataFrame to process
column : str
Column name to process
scale : int, optional
reduce the scale of the feature (reducing the
influence of variations this feature on the clustering
algorithm (the default ... | msticpy/analysis/eventcluster.py | char_ord_score_df | 2xyo/msticpy | python | @export
def char_ord_score_df(data: pd.DataFrame, column: str, scale: int=1) -> pd.Series:
'\n Return sum of ord values of characters in string.\n\n Parameters\n ----------\n data : pd.DataFrame\n The DataFrame to process\n column : str\n Column name to process\n scale : int, optiona... |
@export
def token_count_df(data: pd.DataFrame, column: str, delimiter: str=' ') -> pd.Series:
"\n Return count of delimiter-separated tokens pd.Series column.\n\n Parameters\n ----------\n data : pd.DataFrame\n The DataFrame to process\n column : str\n Column name to process\n delimi... | 6,190,990,618,236,187,000 | Return count of delimiter-separated tokens pd.Series column.
Parameters
----------
data : pd.DataFrame
The DataFrame to process
column : str
Column name to process
delimiter : str, optional
Delimiter used to split the column string.
(the default is ' ')
Returns
-------
pd.Series
count of tokens in... | msticpy/analysis/eventcluster.py | token_count_df | 2xyo/msticpy | python | @export
def token_count_df(data: pd.DataFrame, column: str, delimiter: str=' ') -> pd.Series:
"\n Return count of delimiter-separated tokens pd.Series column.\n\n Parameters\n ----------\n data : pd.DataFrame\n The DataFrame to process\n column : str\n Column name to process\n delimi... |
@export
def crc32_hash_df(data: pd.DataFrame, column: str) -> pd.Series:
'\n Return the CRC32 hash of the input column.\n\n Parameters\n ----------\n data : pd.DataFrame\n The DataFrame to process\n column : str\n Column name to process\n\n Returns\n -------\n pd.Series\n ... | 3,132,224,339,412,920,300 | Return the CRC32 hash of the input column.
Parameters
----------
data : pd.DataFrame
The DataFrame to process
column : str
Column name to process
Returns
-------
pd.Series
CRC32 hash of input column | msticpy/analysis/eventcluster.py | crc32_hash_df | 2xyo/msticpy | python | @export
def crc32_hash_df(data: pd.DataFrame, column: str) -> pd.Series:
'\n Return the CRC32 hash of the input column.\n\n Parameters\n ----------\n data : pd.DataFrame\n The DataFrame to process\n column : str\n Column name to process\n\n Returns\n -------\n pd.Series\n ... |
@export
def plot_cluster(db_cluster: DBSCAN, data: pd.DataFrame, x_predict: np.ndarray, plot_label: str=None, plot_features: Tuple[(int, int)]=(0, 1), verbose: bool=False, cut_off: int=3, xlabel: str=None, ylabel: str=None):
'\n Plot clustered data as scatter chart.\n\n Parameters\n ----------\n db_clus... | 463,275,847,204,298,500 | Plot clustered data as scatter chart.
Parameters
----------
db_cluster : DBSCAN
DBScan Cluster (from SkLearn DBSCAN).
data : pd.DataFrame
Dataframe containing original data.
x_predict : np.ndarray
The DBSCAN predict numpy array
plot_label : str, optional
If set the column to use to label data points
... | msticpy/analysis/eventcluster.py | plot_cluster | 2xyo/msticpy | python | @export
def plot_cluster(db_cluster: DBSCAN, data: pd.DataFrame, x_predict: np.ndarray, plot_label: str=None, plot_features: Tuple[(int, int)]=(0, 1), verbose: bool=False, cut_off: int=3, xlabel: str=None, ylabel: str=None):
'\n Plot clustered data as scatter chart.\n\n Parameters\n ----------\n db_clus... |
def get_store(factory=AuthStore, key=_store_registry_key, app=None):
"Returns an instance of :class:`AuthStore` from the app registry.\n\n It'll try to get it from the current app registry, and if it is not\n registered it'll be instantiated and registered. A second call to this\n function will return the ... | 6,624,406,828,909,451,000 | Returns an instance of :class:`AuthStore` from the app registry.
It'll try to get it from the current app registry, and if it is not
registered it'll be instantiated and registered. A second call to this
function will return the same instance.
:param factory:
The callable used to build and register the instance i... | Webapp2_samplesite/webapp2_extras/auth.py | get_store | Rockfish/PythonCourse | python | def get_store(factory=AuthStore, key=_store_registry_key, app=None):
"Returns an instance of :class:`AuthStore` from the app registry.\n\n It'll try to get it from the current app registry, and if it is not\n registered it'll be instantiated and registered. A second call to this\n function will return the ... |
def set_store(store, key=_store_registry_key, app=None):
'Sets an instance of :class:`AuthStore` in the app registry.\n\n :param store:\n An instance of :class:`AuthStore`.\n :param key:\n The key used to retrieve the instance from the registry. A default\n is used if it is not set.\n ... | 2,798,940,574,467,385,300 | Sets an instance of :class:`AuthStore` in the app registry.
:param store:
An instance of :class:`AuthStore`.
:param key:
The key used to retrieve the instance from the registry. A default
is used if it is not set.
:param request:
A :class:`webapp2.WSGIApplication` instance used to retrieve the
inst... | Webapp2_samplesite/webapp2_extras/auth.py | set_store | Rockfish/PythonCourse | python | def set_store(store, key=_store_registry_key, app=None):
'Sets an instance of :class:`AuthStore` in the app registry.\n\n :param store:\n An instance of :class:`AuthStore`.\n :param key:\n The key used to retrieve the instance from the registry. A default\n is used if it is not set.\n ... |
def get_auth(factory=Auth, key=_auth_registry_key, request=None):
"Returns an instance of :class:`Auth` from the request registry.\n\n It'll try to get it from the current request registry, and if it is not\n registered it'll be instantiated and registered. A second call to this\n function will return the ... | 2,269,344,620,877,514,200 | Returns an instance of :class:`Auth` from the request registry.
It'll try to get it from the current request registry, and if it is not
registered it'll be instantiated and registered. A second call to this
function will return the same instance.
:param factory:
The callable used to build and register the instanc... | Webapp2_samplesite/webapp2_extras/auth.py | get_auth | Rockfish/PythonCourse | python | def get_auth(factory=Auth, key=_auth_registry_key, request=None):
"Returns an instance of :class:`Auth` from the request registry.\n\n It'll try to get it from the current request registry, and if it is not\n registered it'll be instantiated and registered. A second call to this\n function will return the ... |
def set_auth(auth, key=_auth_registry_key, request=None):
'Sets an instance of :class:`Auth` in the request registry.\n\n :param auth:\n An instance of :class:`Auth`.\n :param key:\n The key used to retrieve the instance from the registry. A default\n is used if it is not set.\n :param... | -940,054,351,290,193,800 | Sets an instance of :class:`Auth` in the request registry.
:param auth:
An instance of :class:`Auth`.
:param key:
The key used to retrieve the instance from the registry. A default
is used if it is not set.
:param request:
A :class:`webapp2.Request` instance used to retrieve the instance. The
activ... | Webapp2_samplesite/webapp2_extras/auth.py | set_auth | Rockfish/PythonCourse | python | def set_auth(auth, key=_auth_registry_key, request=None):
'Sets an instance of :class:`Auth` in the request registry.\n\n :param auth:\n An instance of :class:`Auth`.\n :param key:\n The key used to retrieve the instance from the registry. A default\n is used if it is not set.\n :param... |
def __init__(self, app, config=None):
'Initializes the session store.\n\n :param app:\n A :class:`webapp2.WSGIApplication` instance.\n :param config:\n A dictionary of configuration values to be overridden. See\n the available keys in :data:`default_config`.\n '... | -9,101,138,287,151,809,000 | Initializes the session store.
:param app:
A :class:`webapp2.WSGIApplication` instance.
:param config:
A dictionary of configuration values to be overridden. See
the available keys in :data:`default_config`. | Webapp2_samplesite/webapp2_extras/auth.py | __init__ | Rockfish/PythonCourse | python | def __init__(self, app, config=None):
'Initializes the session store.\n\n :param app:\n A :class:`webapp2.WSGIApplication` instance.\n :param config:\n A dictionary of configuration values to be overridden. See\n the available keys in :data:`default_config`.\n '... |
@webapp2.cached_property
def session_attributes(self):
'The list of attributes stored in a session.\n\n This must be an ordered list of unique elements.\n '
seen = set()
attrs = (self._session_attributes + self.user_attributes)
return [a for a in attrs if ((a not in seen) and (not seen.add... | -3,796,023,426,019,284,000 | The list of attributes stored in a session.
This must be an ordered list of unique elements. | Webapp2_samplesite/webapp2_extras/auth.py | session_attributes | Rockfish/PythonCourse | python | @webapp2.cached_property
def session_attributes(self):
'The list of attributes stored in a session.\n\n This must be an ordered list of unique elements.\n '
seen = set()
attrs = (self._session_attributes + self.user_attributes)
return [a for a in attrs if ((a not in seen) and (not seen.add... |
@webapp2.cached_property
def user_attributes(self):
'The list of attributes retrieved from the user model.\n\n This must be an ordered list of unique elements.\n '
seen = set()
attrs = self.config['user_attributes']
return [a for a in attrs if ((a not in seen) and (not seen.add(a)))] | -1,450,585,386,066,368,500 | The list of attributes retrieved from the user model.
This must be an ordered list of unique elements. | Webapp2_samplesite/webapp2_extras/auth.py | user_attributes | Rockfish/PythonCourse | python | @webapp2.cached_property
def user_attributes(self):
'The list of attributes retrieved from the user model.\n\n This must be an ordered list of unique elements.\n '
seen = set()
attrs = self.config['user_attributes']
return [a for a in attrs if ((a not in seen) and (not seen.add(a)))] |
@webapp2.cached_property
def user_model(self):
'Configured user model.'
cls = self.config['user_model']
if isinstance(cls, str):
cls = self.config['user_model'] = webapp2.import_string(cls)
return cls | -9,142,937,729,856,513,000 | Configured user model. | Webapp2_samplesite/webapp2_extras/auth.py | user_model | Rockfish/PythonCourse | python | @webapp2.cached_property
def user_model(self):
cls = self.config['user_model']
if isinstance(cls, str):
cls = self.config['user_model'] = webapp2.import_string(cls)
return cls |
def user_to_dict(self, user):
"Returns a dictionary based on a user object.\n\n Extra attributes to be retrieved must be set in this module's\n configuration.\n\n :param user:\n User object: an instance the custom user model.\n :returns:\n A dictionary with user dat... | 395,166,271,577,502,900 | Returns a dictionary based on a user object.
Extra attributes to be retrieved must be set in this module's
configuration.
:param user:
User object: an instance the custom user model.
:returns:
A dictionary with user data. | Webapp2_samplesite/webapp2_extras/auth.py | user_to_dict | Rockfish/PythonCourse | python | def user_to_dict(self, user):
"Returns a dictionary based on a user object.\n\n Extra attributes to be retrieved must be set in this module's\n configuration.\n\n :param user:\n User object: an instance the custom user model.\n :returns:\n A dictionary with user dat... |
def get_user_by_auth_password(self, auth_id, password, silent=False):
'Returns a user dict based on auth_id and password.\n\n :param auth_id:\n Authentication id.\n :param password:\n User password.\n :param silent:\n If True, raises an exception if auth_id or p... | 3,215,861,248,836,423,000 | Returns a user dict based on auth_id and password.
:param auth_id:
Authentication id.
:param password:
User password.
:param silent:
If True, raises an exception if auth_id or password are invalid.
:returns:
A dictionary with user data.
:raises:
``InvalidAuthIdError`` or ``InvalidPasswordError``. | Webapp2_samplesite/webapp2_extras/auth.py | get_user_by_auth_password | Rockfish/PythonCourse | python | def get_user_by_auth_password(self, auth_id, password, silent=False):
'Returns a user dict based on auth_id and password.\n\n :param auth_id:\n Authentication id.\n :param password:\n User password.\n :param silent:\n If True, raises an exception if auth_id or p... |
def get_user_by_auth_token(self, user_id, token):
'Returns a user dict based on user_id and auth token.\n\n :param user_id:\n User id.\n :param token:\n Authentication token.\n :returns:\n A tuple ``(user_dict, token_timestamp)``. Both values can be None.\n ... | -490,687,262,392,920,260 | Returns a user dict based on user_id and auth token.
:param user_id:
User id.
:param token:
Authentication token.
:returns:
A tuple ``(user_dict, token_timestamp)``. Both values can be None.
The token timestamp will be None if the user is invalid or it
is valid but the token requires renewal. | Webapp2_samplesite/webapp2_extras/auth.py | get_user_by_auth_token | Rockfish/PythonCourse | python | def get_user_by_auth_token(self, user_id, token):
'Returns a user dict based on user_id and auth token.\n\n :param user_id:\n User id.\n :param token:\n Authentication token.\n :returns:\n A tuple ``(user_dict, token_timestamp)``. Both values can be None.\n ... |
def create_auth_token(self, user_id):
'Creates a new authentication token.\n\n :param user_id:\n Authentication id.\n :returns:\n A new authentication token.\n '
return self.user_model.create_auth_token(user_id) | 8,363,434,732,694,904,000 | Creates a new authentication token.
:param user_id:
Authentication id.
:returns:
A new authentication token. | Webapp2_samplesite/webapp2_extras/auth.py | create_auth_token | Rockfish/PythonCourse | python | def create_auth_token(self, user_id):
'Creates a new authentication token.\n\n :param user_id:\n Authentication id.\n :returns:\n A new authentication token.\n '
return self.user_model.create_auth_token(user_id) |
def delete_auth_token(self, user_id, token):
'Deletes an authentication token.\n\n :param user_id:\n User id.\n :param token:\n Authentication token.\n '
return self.user_model.delete_auth_token(user_id, token) | 7,380,778,098,418,366,000 | Deletes an authentication token.
:param user_id:
User id.
:param token:
Authentication token. | Webapp2_samplesite/webapp2_extras/auth.py | delete_auth_token | Rockfish/PythonCourse | python | def delete_auth_token(self, user_id, token):
'Deletes an authentication token.\n\n :param user_id:\n User id.\n :param token:\n Authentication token.\n '
return self.user_model.delete_auth_token(user_id, token) |
def get_session(self, request):
'Returns an auth session.\n\n :param request:\n A :class:`webapp2.Request` instance.\n :returns:\n A session dict.\n '
store = sessions.get_store(request=request)
return store.get_session(self.config['cookie_name'], backend=self.conf... | -5,181,140,631,074,181,000 | Returns an auth session.
:param request:
A :class:`webapp2.Request` instance.
:returns:
A session dict. | Webapp2_samplesite/webapp2_extras/auth.py | get_session | Rockfish/PythonCourse | python | def get_session(self, request):
'Returns an auth session.\n\n :param request:\n A :class:`webapp2.Request` instance.\n :returns:\n A session dict.\n '
store = sessions.get_store(request=request)
return store.get_session(self.config['cookie_name'], backend=self.conf... |
def serialize_session(self, data):
'Serializes values for a session.\n\n :param data:\n A dict with session data.\n :returns:\n A list with session data.\n '
assert (len(data) == len(self.session_attributes))
return [data.get(k) for k in self.session_attributes] | 4,561,535,077,663,560,000 | Serializes values for a session.
:param data:
A dict with session data.
:returns:
A list with session data. | Webapp2_samplesite/webapp2_extras/auth.py | serialize_session | Rockfish/PythonCourse | python | def serialize_session(self, data):
'Serializes values for a session.\n\n :param data:\n A dict with session data.\n :returns:\n A list with session data.\n '
assert (len(data) == len(self.session_attributes))
return [data.get(k) for k in self.session_attributes] |
def deserialize_session(self, data):
'Deserializes values for a session.\n\n :param data:\n A list with session data.\n :returns:\n A dict with session data.\n '
assert (len(data) == len(self.session_attributes))
return dict(list(zip(self.session_attributes, data))... | 6,139,114,933,462,788,000 | Deserializes values for a session.
:param data:
A list with session data.
:returns:
A dict with session data. | Webapp2_samplesite/webapp2_extras/auth.py | deserialize_session | Rockfish/PythonCourse | python | def deserialize_session(self, data):
'Deserializes values for a session.\n\n :param data:\n A list with session data.\n :returns:\n A dict with session data.\n '
assert (len(data) == len(self.session_attributes))
return dict(list(zip(self.session_attributes, data))... |
def set_password_validator(self, func):
'Sets the function used to perform password validation.\n\n :param func:\n A function that receives ``(store, auth_id, password)``\n and returns a user dict or None.\n '
self.validate_password = func.__get__(self, self.__class__) | -5,540,569,157,499,997,000 | Sets the function used to perform password validation.
:param func:
A function that receives ``(store, auth_id, password)``
and returns a user dict or None. | Webapp2_samplesite/webapp2_extras/auth.py | set_password_validator | Rockfish/PythonCourse | python | def set_password_validator(self, func):
'Sets the function used to perform password validation.\n\n :param func:\n A function that receives ``(store, auth_id, password)``\n and returns a user dict or None.\n '
self.validate_password = func.__get__(self, self.__class__) |
def set_token_validator(self, func):
'Sets the function used to perform token validation.\n\n :param func:\n A function that receives ``(store, user_id, token, token_ts)``\n and returns a tuple ``(user_dict, token)``.\n '
self.validate_token = func.__get__(self, self.__class_... | -8,153,688,014,519,468,000 | Sets the function used to perform token validation.
:param func:
A function that receives ``(store, user_id, token, token_ts)``
and returns a tuple ``(user_dict, token)``. | Webapp2_samplesite/webapp2_extras/auth.py | set_token_validator | Rockfish/PythonCourse | python | def set_token_validator(self, func):
'Sets the function used to perform token validation.\n\n :param func:\n A function that receives ``(store, user_id, token, token_ts)``\n and returns a tuple ``(user_dict, token)``.\n '
self.validate_token = func.__get__(self, self.__class_... |
def default_password_validator(self, auth_id, password, silent=False):
'Validates a password.\n\n Passwords are used to log-in using forms or to request auth tokens\n from services.\n\n :param auth_id:\n Authentication id.\n :param password:\n Password to be checked... | -5,114,930,920,428,402,000 | Validates a password.
Passwords are used to log-in using forms or to request auth tokens
from services.
:param auth_id:
Authentication id.
:param password:
Password to be checked.
:param silent:
If True, raises an exception if auth_id or password are invalid.
:returns:
user or None
:raises:
``Inva... | Webapp2_samplesite/webapp2_extras/auth.py | default_password_validator | Rockfish/PythonCourse | python | def default_password_validator(self, auth_id, password, silent=False):
'Validates a password.\n\n Passwords are used to log-in using forms or to request auth tokens\n from services.\n\n :param auth_id:\n Authentication id.\n :param password:\n Password to be checked... |
def default_token_validator(self, user_id, token, token_ts=None):
'Validates a token.\n\n Tokens are random strings used to authenticate temporarily. They are\n used to validate sessions or service requests.\n\n :param user_id:\n User id.\n :param token:\n Token to ... | -311,591,043,904,925,200 | Validates a token.
Tokens are random strings used to authenticate temporarily. They are
used to validate sessions or service requests.
:param user_id:
User id.
:param token:
Token to be checked.
:param token_ts:
Optional token timestamp used to pre-validate the token age.
:returns:
A tuple ``(user_dic... | Webapp2_samplesite/webapp2_extras/auth.py | default_token_validator | Rockfish/PythonCourse | python | def default_token_validator(self, user_id, token, token_ts=None):
'Validates a token.\n\n Tokens are random strings used to authenticate temporarily. They are\n used to validate sessions or service requests.\n\n :param user_id:\n User id.\n :param token:\n Token to ... |
def __init__(self, request):
'Initializes the auth provider for a request.\n\n :param request:\n A :class:`webapp2.Request` instance.\n '
self.request = request
self.store = get_store(app=request.app) | 5,832,942,312,021,142,000 | Initializes the auth provider for a request.
:param request:
A :class:`webapp2.Request` instance. | Webapp2_samplesite/webapp2_extras/auth.py | __init__ | Rockfish/PythonCourse | python | def __init__(self, request):
'Initializes the auth provider for a request.\n\n :param request:\n A :class:`webapp2.Request` instance.\n '
self.request = request
self.store = get_store(app=request.app) |
def get_user_by_session(self, save_session=True):
'Returns a user based on the current session.\n\n :param save_session:\n If True, saves the user in the session if authentication succeeds.\n :returns:\n A user dict or None.\n '
if (self._user is None):
data = ... | 3,667,674,903,363,847,000 | Returns a user based on the current session.
:param save_session:
If True, saves the user in the session if authentication succeeds.
:returns:
A user dict or None. | Webapp2_samplesite/webapp2_extras/auth.py | get_user_by_session | Rockfish/PythonCourse | python | def get_user_by_session(self, save_session=True):
'Returns a user based on the current session.\n\n :param save_session:\n If True, saves the user in the session if authentication succeeds.\n :returns:\n A user dict or None.\n '
if (self._user is None):
data = ... |
def get_user_by_token(self, user_id, token, token_ts=None, cache=None, cache_ts=None, remember=False, save_session=True):
'Returns a user based on an authentication token.\n\n :param user_id:\n User id.\n :param token:\n Authentication token.\n :param token_ts:\n ... | -5,872,988,609,794,356,000 | Returns a user based on an authentication token.
:param user_id:
User id.
:param token:
Authentication token.
:param token_ts:
Token timestamp, used to perform pre-validation.
:param cache:
Cached user data (from the session).
:param cache_ts:
Cache timestamp.
:param remember:
If True, saves pe... | Webapp2_samplesite/webapp2_extras/auth.py | get_user_by_token | Rockfish/PythonCourse | python | def get_user_by_token(self, user_id, token, token_ts=None, cache=None, cache_ts=None, remember=False, save_session=True):
'Returns a user based on an authentication token.\n\n :param user_id:\n User id.\n :param token:\n Authentication token.\n :param token_ts:\n ... |
def get_user_by_password(self, auth_id, password, remember=False, save_session=True, silent=False):
'Returns a user based on password credentials.\n\n :param auth_id:\n Authentication id.\n :param password:\n User password.\n :param remember:\n If True, saves pe... | 9,061,094,405,721,935,000 | Returns a user based on password credentials.
:param auth_id:
Authentication id.
:param password:
User password.
:param remember:
If True, saves permanent sessions.
:param save_session:
If True, saves the user in the session if authentication succeeds.
:param silent:
If True, raises an exception if... | Webapp2_samplesite/webapp2_extras/auth.py | get_user_by_password | Rockfish/PythonCourse | python | def get_user_by_password(self, auth_id, password, remember=False, save_session=True, silent=False):
'Returns a user based on password credentials.\n\n :param auth_id:\n Authentication id.\n :param password:\n User password.\n :param remember:\n If True, saves pe... |
@webapp2.cached_property
def session(self):
'Auth session.'
return self.store.get_session(self.request) | 5,353,006,842,283,151,000 | Auth session. | Webapp2_samplesite/webapp2_extras/auth.py | session | Rockfish/PythonCourse | python | @webapp2.cached_property
def session(self):
return self.store.get_session(self.request) |
def set_session(self, user, token=None, token_ts=None, cache_ts=None, remember=False, **session_args):
'Saves a user in the session.\n\n :param user:\n A dictionary with user data.\n :param token:\n A unique token to be persisted. If None, a new one is created.\n :param to... | -2,310,134,152,950,392,300 | Saves a user in the session.
:param user:
A dictionary with user data.
:param token:
A unique token to be persisted. If None, a new one is created.
:param token_ts:
Token timestamp. If None, a new one is created.
:param cache_ts:
Token cache timestamp. If None, a new one is created.
:remember:
If T... | Webapp2_samplesite/webapp2_extras/auth.py | set_session | Rockfish/PythonCourse | python | def set_session(self, user, token=None, token_ts=None, cache_ts=None, remember=False, **session_args):
'Saves a user in the session.\n\n :param user:\n A dictionary with user data.\n :param token:\n A unique token to be persisted. If None, a new one is created.\n :param to... |
def unset_session(self):
'Removes a user from the session and invalidates the auth token.'
self._user = None
data = self.get_session_data(pop=True)
if data:
self.store.delete_auth_token(data['user_id'], data['token']) | 4,615,528,405,671,208,000 | Removes a user from the session and invalidates the auth token. | Webapp2_samplesite/webapp2_extras/auth.py | unset_session | Rockfish/PythonCourse | python | def unset_session(self):
self._user = None
data = self.get_session_data(pop=True)
if data:
self.store.delete_auth_token(data['user_id'], data['token']) |
def get_session_data(self, pop=False):
'Returns the session data as a dictionary.\n\n :param pop:\n If True, removes the session.\n :returns:\n A deserialized session, or None.\n '
func = (self.session.pop if pop else self.session.get)
rv = func('_user', None)
... | 3,557,656,058,955,620,000 | Returns the session data as a dictionary.
:param pop:
If True, removes the session.
:returns:
A deserialized session, or None. | Webapp2_samplesite/webapp2_extras/auth.py | get_session_data | Rockfish/PythonCourse | python | def get_session_data(self, pop=False):
'Returns the session data as a dictionary.\n\n :param pop:\n If True, removes the session.\n :returns:\n A deserialized session, or None.\n '
func = (self.session.pop if pop else self.session.get)
rv = func('_user', None)
... |
def set_session_data(self, data, **session_args):
'Sets the session data as a list.\n\n :param data:\n Deserialized session data.\n :param session_args:\n Extra arguments for the session.\n '
self.session['_user'] = self.store.serialize_session(data)
self.session.c... | 7,625,603,847,048,672,000 | Sets the session data as a list.
:param data:
Deserialized session data.
:param session_args:
Extra arguments for the session. | Webapp2_samplesite/webapp2_extras/auth.py | set_session_data | Rockfish/PythonCourse | python | def set_session_data(self, data, **session_args):
'Sets the session data as a list.\n\n :param data:\n Deserialized session data.\n :param session_args:\n Extra arguments for the session.\n '
self.session['_user'] = self.store.serialize_session(data)
self.session.c... |
@pytest.fixture(scope='module')
def setup(duthosts, ptfhost, rand_selected_dut, rand_unselected_dut, tbinfo, ptfadapter):
'Gather all required test information from DUT and tbinfo.\n\n Args:\n duthosts: All DUTs belong to the testbed.\n rand_one_dut_hostname: hostname of a random chosen dut to run ... | 2,646,736,912,230,296,000 | Gather all required test information from DUT and tbinfo.
Args:
duthosts: All DUTs belong to the testbed.
rand_one_dut_hostname: hostname of a random chosen dut to run test.
tbinfo: A fixture to gather information about the testbed.
Yields:
A Dictionary with required test information. | tests/acl/test_acl.py | setup | KostiantynYarovyiBf/sonic-mgmt | python | @pytest.fixture(scope='module')
def setup(duthosts, ptfhost, rand_selected_dut, rand_unselected_dut, tbinfo, ptfadapter):
'Gather all required test information from DUT and tbinfo.\n\n Args:\n duthosts: All DUTs belong to the testbed.\n rand_one_dut_hostname: hostname of a random chosen dut to run ... |
@pytest.fixture(scope='module')
def populate_vlan_arp_entries(setup, ptfhost, duthosts, rand_one_dut_hostname, ip_version):
'Set up the ARP responder utility in the PTF container.'
duthost = duthosts[rand_one_dut_hostname]
if (setup['topo'] != 't0'):
def noop():
pass
(yield noop... | -6,976,798,631,610,011,000 | Set up the ARP responder utility in the PTF container. | tests/acl/test_acl.py | populate_vlan_arp_entries | KostiantynYarovyiBf/sonic-mgmt | python | @pytest.fixture(scope='module')
def populate_vlan_arp_entries(setup, ptfhost, duthosts, rand_one_dut_hostname, ip_version):
duthost = duthosts[rand_one_dut_hostname]
if (setup['topo'] != 't0'):
def noop():
pass
(yield noop)
return
addr_list = [DOWNSTREAM_DST_IP[ip_v... |
@pytest.fixture(scope='module', params=['ingress', 'egress'])
def stage(request, duthosts, rand_one_dut_hostname):
'Parametrize tests for Ingress/Egress stage testing.\n\n Args:\n request: A fixture to interact with Pytest data.\n duthosts: All DUTs belong to the testbed.\n rand_one_dut_host... | -784,252,154,036,044,000 | Parametrize tests for Ingress/Egress stage testing.
Args:
request: A fixture to interact with Pytest data.
duthosts: All DUTs belong to the testbed.
rand_one_dut_hostname: hostname of a random chosen dut to run test.
Returns:
str: The ACL stage to be tested. | tests/acl/test_acl.py | stage | KostiantynYarovyiBf/sonic-mgmt | python | @pytest.fixture(scope='module', params=['ingress', 'egress'])
def stage(request, duthosts, rand_one_dut_hostname):
'Parametrize tests for Ingress/Egress stage testing.\n\n Args:\n request: A fixture to interact with Pytest data.\n duthosts: All DUTs belong to the testbed.\n rand_one_dut_host... |
@pytest.fixture(scope='module')
def acl_table(duthosts, rand_one_dut_hostname, setup, stage, ip_version):
'Apply ACL table configuration and remove after tests.\n\n Args:\n duthosts: All DUTs belong to the testbed.\n rand_one_dut_hostname: hostname of a random chosen dut to run test.\n setup... | -3,063,090,470,390,482,000 | Apply ACL table configuration and remove after tests.
Args:
duthosts: All DUTs belong to the testbed.
rand_one_dut_hostname: hostname of a random chosen dut to run test.
setup: Parameters for the ACL tests.
stage: The ACL stage under test.
ip_version: The IP version under test
Yields:
The ACL ... | tests/acl/test_acl.py | acl_table | KostiantynYarovyiBf/sonic-mgmt | python | @pytest.fixture(scope='module')
def acl_table(duthosts, rand_one_dut_hostname, setup, stage, ip_version):
'Apply ACL table configuration and remove after tests.\n\n Args:\n duthosts: All DUTs belong to the testbed.\n rand_one_dut_hostname: hostname of a random chosen dut to run test.\n setup... |
@abstractmethod
def setup_rules(self, dut, acl_table, ip_version):
'Setup ACL rules for testing.\n\n Args:\n dut: The DUT having ACLs applied.\n acl_table: Configuration info for the ACL table.\n\n '
pass | -7,690,963,436,585,185,000 | Setup ACL rules for testing.
Args:
dut: The DUT having ACLs applied.
acl_table: Configuration info for the ACL table. | tests/acl/test_acl.py | setup_rules | KostiantynYarovyiBf/sonic-mgmt | python | @abstractmethod
def setup_rules(self, dut, acl_table, ip_version):
'Setup ACL rules for testing.\n\n Args:\n dut: The DUT having ACLs applied.\n acl_table: Configuration info for the ACL table.\n\n '
pass |
def post_setup_hook(self, dut, localhost, populate_vlan_arp_entries, tbinfo):
'Perform actions after rules have been applied.\n\n Args:\n dut: The DUT having ACLs applied.\n localhost: The host from which tests are run.\n populate_vlan_arp_entries: A function to populate ARP/... | 5,227,020,109,855,199,000 | Perform actions after rules have been applied.
Args:
dut: The DUT having ACLs applied.
localhost: The host from which tests are run.
populate_vlan_arp_entries: A function to populate ARP/FDB tables for VLAN interfaces. | tests/acl/test_acl.py | post_setup_hook | KostiantynYarovyiBf/sonic-mgmt | python | def post_setup_hook(self, dut, localhost, populate_vlan_arp_entries, tbinfo):
'Perform actions after rules have been applied.\n\n Args:\n dut: The DUT having ACLs applied.\n localhost: The host from which tests are run.\n populate_vlan_arp_entries: A function to populate ARP/... |
def teardown_rules(self, dut):
'Tear down ACL rules once the tests have completed.\n\n Args:\n dut: The DUT having ACLs applied.\n\n '
logger.info('Finished with tests, removing all ACL rules...')
dut.copy(src=os.path.join(FILES_DIR, ACL_REMOVE_RULES_FILE), dest=DUT_TMP_DIR)
rem... | -7,491,526,164,275,526,000 | Tear down ACL rules once the tests have completed.
Args:
dut: The DUT having ACLs applied. | tests/acl/test_acl.py | teardown_rules | KostiantynYarovyiBf/sonic-mgmt | python | def teardown_rules(self, dut):
'Tear down ACL rules once the tests have completed.\n\n Args:\n dut: The DUT having ACLs applied.\n\n '
logger.info('Finished with tests, removing all ACL rules...')
dut.copy(src=os.path.join(FILES_DIR, ACL_REMOVE_RULES_FILE), dest=DUT_TMP_DIR)
rem... |
@pytest.fixture(scope='class', autouse=True)
def acl_rules(self, duthosts, localhost, setup, acl_table, populate_vlan_arp_entries, tbinfo, ip_version):
'Setup/teardown ACL rules for the current set of tests.\n\n Args:\n duthosts: All DUTs belong to the testbed.\n rand_one_dut_hostname: ... | -2,370,236,786,902,345,000 | Setup/teardown ACL rules for the current set of tests.
Args:
duthosts: All DUTs belong to the testbed.
rand_one_dut_hostname: hostname of a random chosen dut to run test.
localhost: The host from which tests are run.
setup: Parameters for the ACL tests.
acl_table: Configuration info for the ACL tab... | tests/acl/test_acl.py | acl_rules | KostiantynYarovyiBf/sonic-mgmt | python | @pytest.fixture(scope='class', autouse=True)
def acl_rules(self, duthosts, localhost, setup, acl_table, populate_vlan_arp_entries, tbinfo, ip_version):
'Setup/teardown ACL rules for the current set of tests.\n\n Args:\n duthosts: All DUTs belong to the testbed.\n rand_one_dut_hostname: ... |
@pytest.yield_fixture(scope='class', autouse=True)
def counters_sanity_check(self, duthosts, acl_rules, acl_table):
'Validate that the counters for each rule in the rules list increased as expected.\n\n This fixture yields a list of rule IDs. The test case should add on to this list if\n it is require... | 3,925,598,669,342,153,700 | Validate that the counters for each rule in the rules list increased as expected.
This fixture yields a list of rule IDs. The test case should add on to this list if
it is required to check the rule for increased counters.
After the test cases pass, the fixture will wait for the ACL counters to update and then
check ... | tests/acl/test_acl.py | counters_sanity_check | KostiantynYarovyiBf/sonic-mgmt | python | @pytest.yield_fixture(scope='class', autouse=True)
def counters_sanity_check(self, duthosts, acl_rules, acl_table):
'Validate that the counters for each rule in the rules list increased as expected.\n\n This fixture yields a list of rule IDs. The test case should add on to this list if\n it is require... |
@pytest.fixture(params=['downlink->uplink', 'uplink->downlink'])
def direction(self, request):
'Parametrize test based on direction of traffic.'
return request.param | 3,840,651,992,981,872,000 | Parametrize test based on direction of traffic. | tests/acl/test_acl.py | direction | KostiantynYarovyiBf/sonic-mgmt | python | @pytest.fixture(params=['downlink->uplink', 'uplink->downlink'])
def direction(self, request):
return request.param |
@pytest.fixture(autouse=True)
def get_src_port(self, setup, direction):
'Get a source port for the current test.'
src_ports = (setup['downstream_port_ids'] if (direction == 'downlink->uplink') else setup['upstream_port_ids'])
src_port = random.choice(src_ports)
logger.info('Selected source port {}'.form... | -910,017,347,846,818,700 | Get a source port for the current test. | tests/acl/test_acl.py | get_src_port | KostiantynYarovyiBf/sonic-mgmt | python | @pytest.fixture(autouse=True)
def get_src_port(self, setup, direction):
src_ports = (setup['downstream_port_ids'] if (direction == 'downlink->uplink') else setup['upstream_port_ids'])
src_port = random.choice(src_ports)
logger.info('Selected source port {}'.format(src_port))
self.src_port = src_por... |
def get_dst_ports(self, setup, direction):
'Get the set of possible destination ports for the current test.'
return (setup['upstream_port_ids'] if (direction == 'downlink->uplink') else setup['downstream_port_ids']) | -7,723,670,512,636,729,000 | Get the set of possible destination ports for the current test. | tests/acl/test_acl.py | get_dst_ports | KostiantynYarovyiBf/sonic-mgmt | python | def get_dst_ports(self, setup, direction):
return (setup['upstream_port_ids'] if (direction == 'downlink->uplink') else setup['downstream_port_ids']) |
def get_dst_ip(self, direction, ip_version):
'Get the default destination IP for the current test.'
return (UPSTREAM_DST_IP[ip_version] if (direction == 'downlink->uplink') else DOWNSTREAM_DST_IP[ip_version]) | 7,964,747,422,421,615,000 | Get the default destination IP for the current test. | tests/acl/test_acl.py | get_dst_ip | KostiantynYarovyiBf/sonic-mgmt | python | def get_dst_ip(self, direction, ip_version):
return (UPSTREAM_DST_IP[ip_version] if (direction == 'downlink->uplink') else DOWNSTREAM_DST_IP[ip_version]) |
def tcp_packet(self, setup, direction, ptfadapter, ip_version, src_ip=None, dst_ip=None, proto=None, sport=17185, dport=81, flags=None):
'Generate a TCP packet for testing.'
src_ip = (src_ip or DEFAULT_SRC_IP[ip_version])
dst_ip = (dst_ip or self.get_dst_ip(direction, ip_version))
if (ip_version == 'ipv... | 7,023,518,864,380,091,000 | Generate a TCP packet for testing. | tests/acl/test_acl.py | tcp_packet | KostiantynYarovyiBf/sonic-mgmt | python | def tcp_packet(self, setup, direction, ptfadapter, ip_version, src_ip=None, dst_ip=None, proto=None, sport=17185, dport=81, flags=None):
src_ip = (src_ip or DEFAULT_SRC_IP[ip_version])
dst_ip = (dst_ip or self.get_dst_ip(direction, ip_version))
if (ip_version == 'ipv4'):
pkt = testutils.simple_... |
def udp_packet(self, setup, direction, ptfadapter, ip_version, src_ip=None, dst_ip=None, sport=1234, dport=80):
'Generate a UDP packet for testing.'
src_ip = (src_ip or DEFAULT_SRC_IP[ip_version])
dst_ip = (dst_ip or self.get_dst_ip(direction, ip_version))
if (ip_version == 'ipv4'):
return testu... | 7,555,449,266,937,661,000 | Generate a UDP packet for testing. | tests/acl/test_acl.py | udp_packet | KostiantynYarovyiBf/sonic-mgmt | python | def udp_packet(self, setup, direction, ptfadapter, ip_version, src_ip=None, dst_ip=None, sport=1234, dport=80):
src_ip = (src_ip or DEFAULT_SRC_IP[ip_version])
dst_ip = (dst_ip or self.get_dst_ip(direction, ip_version))
if (ip_version == 'ipv4'):
return testutils.simple_udp_packet(eth_dst=setup... |
def icmp_packet(self, setup, direction, ptfadapter, ip_version, src_ip=None, dst_ip=None, icmp_type=8, icmp_code=0):
'Generate an ICMP packet for testing.'
src_ip = (src_ip or DEFAULT_SRC_IP[ip_version])
dst_ip = (dst_ip or self.get_dst_ip(direction, ip_version))
if (ip_version == 'ipv4'):
retur... | -4,280,413,460,718,121,000 | Generate an ICMP packet for testing. | tests/acl/test_acl.py | icmp_packet | KostiantynYarovyiBf/sonic-mgmt | python | def icmp_packet(self, setup, direction, ptfadapter, ip_version, src_ip=None, dst_ip=None, icmp_type=8, icmp_code=0):
src_ip = (src_ip or DEFAULT_SRC_IP[ip_version])
dst_ip = (dst_ip or self.get_dst_ip(direction, ip_version))
if (ip_version == 'ipv4'):
return testutils.simple_icmp_packet(eth_dst... |
def expected_mask_routed_packet(self, pkt, ip_version):
'Generate the expected mask for a routed packet.'
exp_pkt = pkt.copy()
exp_pkt = mask.Mask(exp_pkt)
exp_pkt.set_do_not_care_scapy(packet.Ether, 'dst')
exp_pkt.set_do_not_care_scapy(packet.Ether, 'src')
if (ip_version == 'ipv4'):
exp... | -3,256,258,984,992,967,000 | Generate the expected mask for a routed packet. | tests/acl/test_acl.py | expected_mask_routed_packet | KostiantynYarovyiBf/sonic-mgmt | python | def expected_mask_routed_packet(self, pkt, ip_version):
exp_pkt = pkt.copy()
exp_pkt = mask.Mask(exp_pkt)
exp_pkt.set_do_not_care_scapy(packet.Ether, 'dst')
exp_pkt.set_do_not_care_scapy(packet.Ether, 'src')
if (ip_version == 'ipv4'):
exp_pkt.set_do_not_care_scapy(packet.IP, 'chksum')
... |
def test_ingress_unmatched_blocked(self, setup, direction, ptfadapter, ip_version, stage):
'Verify that unmatched packets are dropped for ingress.'
if (stage == 'egress'):
pytest.skip('Only run for ingress')
pkt = self.tcp_packet(setup, direction, ptfadapter, ip_version)
self._verify_acl_traffic... | 2,178,950,070,650,520,800 | Verify that unmatched packets are dropped for ingress. | tests/acl/test_acl.py | test_ingress_unmatched_blocked | KostiantynYarovyiBf/sonic-mgmt | python | def test_ingress_unmatched_blocked(self, setup, direction, ptfadapter, ip_version, stage):
if (stage == 'egress'):
pytest.skip('Only run for ingress')
pkt = self.tcp_packet(setup, direction, ptfadapter, ip_version)
self._verify_acl_traffic(setup, direction, ptfadapter, pkt, True, ip_version) |
def test_egress_unmatched_forwarded(self, setup, direction, ptfadapter, ip_version, stage):
'Verify that default egress rule allow all traffics'
if (stage == 'ingress'):
pytest.skip('Only run for egress')
pkt = self.tcp_packet(setup, direction, ptfadapter, ip_version)
self._verify_acl_traffic(se... | 5,269,488,178,567,495,000 | Verify that default egress rule allow all traffics | tests/acl/test_acl.py | test_egress_unmatched_forwarded | KostiantynYarovyiBf/sonic-mgmt | python | def test_egress_unmatched_forwarded(self, setup, direction, ptfadapter, ip_version, stage):
if (stage == 'ingress'):
pytest.skip('Only run for egress')
pkt = self.tcp_packet(setup, direction, ptfadapter, ip_version)
self._verify_acl_traffic(setup, direction, ptfadapter, pkt, False, ip_version) |
def test_source_ip_match_forwarded(self, setup, direction, ptfadapter, counters_sanity_check, ip_version):
'Verify that we can match and forward a packet on source IP.'
src_ip = ('0.0.0.0' if (ip_version == 'ipv4') else '0000:0000:0000:0000:0000:0000:0000:0000')
pkt = self.tcp_packet(setup, direction, ptfad... | 5,713,231,115,088,048,000 | Verify that we can match and forward a packet on source IP. | tests/acl/test_acl.py | test_source_ip_match_forwarded | KostiantynYarovyiBf/sonic-mgmt | python | def test_source_ip_match_forwarded(self, setup, direction, ptfadapter, counters_sanity_check, ip_version):
src_ip = ('0.0.0.0' if (ip_version == 'ipv4') else '0000:0000:0000:0000:0000:0000:0000:0000')
pkt = self.tcp_packet(setup, direction, ptfadapter, ip_version, src_ip=src_ip)
self._verify_acl_traffi... |
def test_rules_priority_forwarded(self, setup, direction, ptfadapter, counters_sanity_check, ip_version):
'Verify that we respect rule priorites in the forwarding case.'
src_ip = ('0.0.0.0' if (ip_version == 'ipv4') else '0000:0000:0000:0000:0000:0000:0000:0000')
pkt = self.tcp_packet(setup, direction, ptfa... | -8,105,410,999,960,000,000 | Verify that we respect rule priorites in the forwarding case. | tests/acl/test_acl.py | test_rules_priority_forwarded | KostiantynYarovyiBf/sonic-mgmt | python | def test_rules_priority_forwarded(self, setup, direction, ptfadapter, counters_sanity_check, ip_version):
src_ip = ('0.0.0.0' if (ip_version == 'ipv4') else '0000:0000:0000:0000:0000:0000:0000:0000')
pkt = self.tcp_packet(setup, direction, ptfadapter, ip_version, src_ip=src_ip)
self._verify_acl_traffic... |
def test_rules_priority_dropped(self, setup, direction, ptfadapter, counters_sanity_check, ip_version):
'Verify that we respect rule priorites in the drop case.'
src_ip = ('0.0.0.0' if (ip_version == 'ipv4') else '0000:0000:0000:0000:0000:0000:0000:0000')
pkt = self.tcp_packet(setup, direction, ptfadapter, ... | 959,052,328,267,444,600 | Verify that we respect rule priorites in the drop case. | tests/acl/test_acl.py | test_rules_priority_dropped | KostiantynYarovyiBf/sonic-mgmt | python | def test_rules_priority_dropped(self, setup, direction, ptfadapter, counters_sanity_check, ip_version):
src_ip = ('0.0.0.0' if (ip_version == 'ipv4') else '0000:0000:0000:0000:0000:0000:0000:0000')
pkt = self.tcp_packet(setup, direction, ptfadapter, ip_version, src_ip=src_ip)
self._verify_acl_traffic(s... |
def test_dest_ip_match_forwarded(self, setup, direction, ptfadapter, counters_sanity_check, ip_version):
'Verify that we can match and forward a packet on destination IP.'
dst_ip = (DOWNSTREAM_IP_TO_ALLOW[ip_version] if (direction == 'uplink->downlink') else UPSTREAM_IP_TO_ALLOW[ip_version])
pkt = self.tcp_... | 7,955,340,746,633,166,000 | Verify that we can match and forward a packet on destination IP. | tests/acl/test_acl.py | test_dest_ip_match_forwarded | KostiantynYarovyiBf/sonic-mgmt | python | def test_dest_ip_match_forwarded(self, setup, direction, ptfadapter, counters_sanity_check, ip_version):
dst_ip = (DOWNSTREAM_IP_TO_ALLOW[ip_version] if (direction == 'uplink->downlink') else UPSTREAM_IP_TO_ALLOW[ip_version])
pkt = self.tcp_packet(setup, direction, ptfadapter, ip_version, dst_ip=dst_ip)
... |
def test_dest_ip_match_dropped(self, setup, direction, ptfadapter, counters_sanity_check, ip_version):
'Verify that we can match and drop a packet on destination IP.'
dst_ip = (DOWNSTREAM_IP_TO_BLOCK[ip_version] if (direction == 'uplink->downlink') else UPSTREAM_IP_TO_BLOCK[ip_version])
pkt = self.tcp_packe... | -6,354,044,041,571,471,000 | Verify that we can match and drop a packet on destination IP. | tests/acl/test_acl.py | test_dest_ip_match_dropped | KostiantynYarovyiBf/sonic-mgmt | python | def test_dest_ip_match_dropped(self, setup, direction, ptfadapter, counters_sanity_check, ip_version):
dst_ip = (DOWNSTREAM_IP_TO_BLOCK[ip_version] if (direction == 'uplink->downlink') else UPSTREAM_IP_TO_BLOCK[ip_version])
pkt = self.tcp_packet(setup, direction, ptfadapter, ip_version, dst_ip=dst_ip)
... |
def test_source_ip_match_dropped(self, setup, direction, ptfadapter, counters_sanity_check, ip_version):
'Verify that we can match and drop a packet on source IP.'
src_ip = ('0.0.0.0' if (ip_version == 'ipv4') else '0000:0000:0000:0000:0000:0000:0000:0000')
pkt = self.tcp_packet(setup, direction, ptfadapter... | 5,289,088,510,342,193,000 | Verify that we can match and drop a packet on source IP. | tests/acl/test_acl.py | test_source_ip_match_dropped | KostiantynYarovyiBf/sonic-mgmt | python | def test_source_ip_match_dropped(self, setup, direction, ptfadapter, counters_sanity_check, ip_version):
src_ip = ('0.0.0.0' if (ip_version == 'ipv4') else '0000:0000:0000:0000:0000:0000:0000:0000')
pkt = self.tcp_packet(setup, direction, ptfadapter, ip_version, src_ip=src_ip)
self._verify_acl_traffic(... |
def test_udp_source_ip_match_forwarded(self, setup, direction, ptfadapter, counters_sanity_check, ip_version):
'Verify that we can match and forward a UDP packet on source IP.'
src_ip = ('0.0.0.0' if (ip_version == 'ipv4') else '0000:0000:0000:0000:0000:0000:0000:0000')
pkt = self.udp_packet(setup, directio... | -7,716,854,753,495,035,000 | Verify that we can match and forward a UDP packet on source IP. | tests/acl/test_acl.py | test_udp_source_ip_match_forwarded | KostiantynYarovyiBf/sonic-mgmt | python | def test_udp_source_ip_match_forwarded(self, setup, direction, ptfadapter, counters_sanity_check, ip_version):
src_ip = ('0.0.0.0' if (ip_version == 'ipv4') else '0000:0000:0000:0000:0000:0000:0000:0000')
pkt = self.udp_packet(setup, direction, ptfadapter, ip_version, src_ip=src_ip)
self._verify_acl_tr... |
def test_udp_source_ip_match_dropped(self, setup, direction, ptfadapter, counters_sanity_check, ip_version):
'Verify that we can match and drop a UDP packet on source IP.'
src_ip = ('0.0.0.0' if (ip_version == 'ipv4') else '0000:0000:0000:0000:0000:0000:0000:0000')
pkt = self.udp_packet(setup, direction, pt... | 3,425,280,445,662,462,500 | Verify that we can match and drop a UDP packet on source IP. | tests/acl/test_acl.py | test_udp_source_ip_match_dropped | KostiantynYarovyiBf/sonic-mgmt | python | def test_udp_source_ip_match_dropped(self, setup, direction, ptfadapter, counters_sanity_check, ip_version):
src_ip = ('0.0.0.0' if (ip_version == 'ipv4') else '0000:0000:0000:0000:0000:0000:0000:0000')
pkt = self.udp_packet(setup, direction, ptfadapter, ip_version, src_ip=src_ip)
self._verify_acl_traf... |
def test_icmp_source_ip_match_dropped(self, setup, direction, ptfadapter, counters_sanity_check, ip_version):
'Verify that we can match and drop an ICMP packet on source IP.'
src_ip = ('0.0.0.0' if (ip_version == 'ipv4') else '0000:0000:0000:0000:0000:0000:0000:0000')
pkt = self.icmp_packet(setup, direction... | -7,805,632,420,765,763,000 | Verify that we can match and drop an ICMP packet on source IP. | tests/acl/test_acl.py | test_icmp_source_ip_match_dropped | KostiantynYarovyiBf/sonic-mgmt | python | def test_icmp_source_ip_match_dropped(self, setup, direction, ptfadapter, counters_sanity_check, ip_version):
src_ip = ('0.0.0.0' if (ip_version == 'ipv4') else '0000:0000:0000:0000:0000:0000:0000:0000')
pkt = self.icmp_packet(setup, direction, ptfadapter, ip_version, src_ip=src_ip)
self._verify_acl_tr... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.