partition
stringclasses
3 values
func_name
stringlengths
1
134
docstring
stringlengths
1
46.9k
path
stringlengths
4
223
original_string
stringlengths
75
104k
code
stringlengths
75
104k
docstring_tokens
listlengths
1
1.97k
repo
stringlengths
7
55
language
stringclasses
1 value
url
stringlengths
87
315
code_tokens
listlengths
19
28.4k
sha
stringlengths
40
40
test
CarddavObject._get_webpages
:rtype: list(list(str))
khard/carddav_object.py
def _get_webpages(self): """ :rtype: list(list(str)) """ urls = [] for child in self.vcard.getChildren(): if child.name == "URL": urls.append(child.value) return sorted(urls)
def _get_webpages(self): """ :rtype: list(list(str)) """ urls = [] for child in self.vcard.getChildren(): if child.name == "URL": urls.append(child.value) return sorted(urls)
[ ":", "rtype", ":", "list", "(", "list", "(", "str", "))" ]
scheibler/khard
python
https://github.com/scheibler/khard/blob/0f69430c2680f1ff5f073a977a3c5b753b96cc17/khard/carddav_object.py#L670-L678
[ "def", "_get_webpages", "(", "self", ")", ":", "urls", "=", "[", "]", "for", "child", "in", "self", ".", "vcard", ".", "getChildren", "(", ")", ":", "if", "child", ".", "name", "==", "\"URL\"", ":", "urls", ".", "append", "(", "child", ".", "value", ")", "return", "sorted", "(", "urls", ")" ]
0f69430c2680f1ff5f073a977a3c5b753b96cc17
test
CarddavObject.get_anniversary
:returns: contacts anniversary or None if not available :rtype: datetime.datetime or str
khard/carddav_object.py
def get_anniversary(self): """:returns: contacts anniversary or None if not available :rtype: datetime.datetime or str """ # vcard 4.0 could contain a single text value try: if self.vcard.anniversary.params.get("VALUE")[0] == "text": return self.vcard.anniversary.value except (AttributeError, IndexError, TypeError): pass # else try to convert to a datetime object try: return helpers.string_to_date(self.vcard.anniversary.value) except (AttributeError, ValueError): # vcard 3.0: x-anniversary (private object) try: return helpers.string_to_date(self.vcard.x_anniversary.value) except (AttributeError, ValueError): pass return None
def get_anniversary(self): """:returns: contacts anniversary or None if not available :rtype: datetime.datetime or str """ # vcard 4.0 could contain a single text value try: if self.vcard.anniversary.params.get("VALUE")[0] == "text": return self.vcard.anniversary.value except (AttributeError, IndexError, TypeError): pass # else try to convert to a datetime object try: return helpers.string_to_date(self.vcard.anniversary.value) except (AttributeError, ValueError): # vcard 3.0: x-anniversary (private object) try: return helpers.string_to_date(self.vcard.x_anniversary.value) except (AttributeError, ValueError): pass return None
[ ":", "returns", ":", "contacts", "anniversary", "or", "None", "if", "not", "available", ":", "rtype", ":", "datetime", ".", "datetime", "or", "str" ]
scheibler/khard
python
https://github.com/scheibler/khard/blob/0f69430c2680f1ff5f073a977a3c5b753b96cc17/khard/carddav_object.py#L685-L704
[ "def", "get_anniversary", "(", "self", ")", ":", "# vcard 4.0 could contain a single text value", "try", ":", "if", "self", ".", "vcard", ".", "anniversary", ".", "params", ".", "get", "(", "\"VALUE\"", ")", "[", "0", "]", "==", "\"text\"", ":", "return", "self", ".", "vcard", ".", "anniversary", ".", "value", "except", "(", "AttributeError", ",", "IndexError", ",", "TypeError", ")", ":", "pass", "# else try to convert to a datetime object", "try", ":", "return", "helpers", ".", "string_to_date", "(", "self", ".", "vcard", ".", "anniversary", ".", "value", ")", "except", "(", "AttributeError", ",", "ValueError", ")", ":", "# vcard 3.0: x-anniversary (private object)", "try", ":", "return", "helpers", ".", "string_to_date", "(", "self", ".", "vcard", ".", "x_anniversary", ".", "value", ")", "except", "(", "AttributeError", ",", "ValueError", ")", ":", "pass", "return", "None" ]
0f69430c2680f1ff5f073a977a3c5b753b96cc17
test
CarddavObject.get_birthday
:returns: contacts birthday or None if not available :rtype: datetime.datetime or str
khard/carddav_object.py
def get_birthday(self): """:returns: contacts birthday or None if not available :rtype: datetime.datetime or str """ # vcard 4.0 could contain a single text value try: if self.vcard.bday.params.get("VALUE")[0] == "text": return self.vcard.bday.value except (AttributeError, IndexError, TypeError): pass # else try to convert to a datetime object try: return helpers.string_to_date(self.vcard.bday.value) except (AttributeError, ValueError): pass return None
def get_birthday(self): """:returns: contacts birthday or None if not available :rtype: datetime.datetime or str """ # vcard 4.0 could contain a single text value try: if self.vcard.bday.params.get("VALUE")[0] == "text": return self.vcard.bday.value except (AttributeError, IndexError, TypeError): pass # else try to convert to a datetime object try: return helpers.string_to_date(self.vcard.bday.value) except (AttributeError, ValueError): pass return None
[ ":", "returns", ":", "contacts", "birthday", "or", "None", "if", "not", "available", ":", "rtype", ":", "datetime", ".", "datetime", "or", "str" ]
scheibler/khard
python
https://github.com/scheibler/khard/blob/0f69430c2680f1ff5f073a977a3c5b753b96cc17/khard/carddav_object.py#L753-L768
[ "def", "get_birthday", "(", "self", ")", ":", "# vcard 4.0 could contain a single text value", "try", ":", "if", "self", ".", "vcard", ".", "bday", ".", "params", ".", "get", "(", "\"VALUE\"", ")", "[", "0", "]", "==", "\"text\"", ":", "return", "self", ".", "vcard", ".", "bday", ".", "value", "except", "(", "AttributeError", ",", "IndexError", ",", "TypeError", ")", ":", "pass", "# else try to convert to a datetime object", "try", ":", "return", "helpers", ".", "string_to_date", "(", "self", ".", "vcard", ".", "bday", ".", "value", ")", "except", "(", "AttributeError", ",", "ValueError", ")", ":", "pass", "return", "None" ]
0f69430c2680f1ff5f073a977a3c5b753b96cc17
test
CarddavObject._get_types_for_vcard_object
get list of types for phone number, email or post address :param object: vcard class object :type object: vobject.vCard :param default_type: use if the object contains no type :type default_type: str :returns: list of type labels :rtype: list(str)
khard/carddav_object.py
def _get_types_for_vcard_object(self, object, default_type): """ get list of types for phone number, email or post address :param object: vcard class object :type object: vobject.vCard :param default_type: use if the object contains no type :type default_type: str :returns: list of type labels :rtype: list(str) """ type_list = [] # try to find label group for custom value type if object.group: for label in self.vcard.getChildren(): if label.name == "X-ABLABEL" and label.group == object.group: custom_type = label.value.strip() if custom_type: type_list.append(custom_type) # then load type from params dict standard_types = object.params.get("TYPE") if standard_types is not None: if not isinstance(standard_types, list): standard_types = [standard_types] for type in standard_types: type = type.strip() if type and type.lower() != "pref": if not type.lower().startswith("x-"): type_list.append(type) elif type[2:].lower() not in [x.lower() for x in type_list]: # add x-custom type in case it's not already added by # custom label for loop above but strip x- before type_list.append(type[2:]) # try to get pref parameter from vcard version 4.0 try: type_list.append("pref=%d" % int(object.params.get("PREF")[0])) except (IndexError, TypeError, ValueError): # else try to determine, if type params contain pref attribute try: for x in object.params.get("TYPE"): if x.lower() == "pref" and "pref" not in type_list: type_list.append("pref") except TypeError: pass # return type_list or default type if type_list: return type_list return [default_type]
def _get_types_for_vcard_object(self, object, default_type): """ get list of types for phone number, email or post address :param object: vcard class object :type object: vobject.vCard :param default_type: use if the object contains no type :type default_type: str :returns: list of type labels :rtype: list(str) """ type_list = [] # try to find label group for custom value type if object.group: for label in self.vcard.getChildren(): if label.name == "X-ABLABEL" and label.group == object.group: custom_type = label.value.strip() if custom_type: type_list.append(custom_type) # then load type from params dict standard_types = object.params.get("TYPE") if standard_types is not None: if not isinstance(standard_types, list): standard_types = [standard_types] for type in standard_types: type = type.strip() if type and type.lower() != "pref": if not type.lower().startswith("x-"): type_list.append(type) elif type[2:].lower() not in [x.lower() for x in type_list]: # add x-custom type in case it's not already added by # custom label for loop above but strip x- before type_list.append(type[2:]) # try to get pref parameter from vcard version 4.0 try: type_list.append("pref=%d" % int(object.params.get("PREF")[0])) except (IndexError, TypeError, ValueError): # else try to determine, if type params contain pref attribute try: for x in object.params.get("TYPE"): if x.lower() == "pref" and "pref" not in type_list: type_list.append("pref") except TypeError: pass # return type_list or default type if type_list: return type_list return [default_type]
[ "get", "list", "of", "types", "for", "phone", "number", "email", "or", "post", "address", ":", "param", "object", ":", "vcard", "class", "object", ":", "type", "object", ":", "vobject", ".", "vCard", ":", "param", "default_type", ":", "use", "if", "the", "object", "contains", "no", "type", ":", "type", "default_type", ":", "str", ":", "returns", ":", "list", "of", "type", "labels", ":", "rtype", ":", "list", "(", "str", ")" ]
scheibler/khard
python
https://github.com/scheibler/khard/blob/0f69430c2680f1ff5f073a977a3c5b753b96cc17/khard/carddav_object.py#L1507-L1553
[ "def", "_get_types_for_vcard_object", "(", "self", ",", "object", ",", "default_type", ")", ":", "type_list", "=", "[", "]", "# try to find label group for custom value type", "if", "object", ".", "group", ":", "for", "label", "in", "self", ".", "vcard", ".", "getChildren", "(", ")", ":", "if", "label", ".", "name", "==", "\"X-ABLABEL\"", "and", "label", ".", "group", "==", "object", ".", "group", ":", "custom_type", "=", "label", ".", "value", ".", "strip", "(", ")", "if", "custom_type", ":", "type_list", ".", "append", "(", "custom_type", ")", "# then load type from params dict", "standard_types", "=", "object", ".", "params", ".", "get", "(", "\"TYPE\"", ")", "if", "standard_types", "is", "not", "None", ":", "if", "not", "isinstance", "(", "standard_types", ",", "list", ")", ":", "standard_types", "=", "[", "standard_types", "]", "for", "type", "in", "standard_types", ":", "type", "=", "type", ".", "strip", "(", ")", "if", "type", "and", "type", ".", "lower", "(", ")", "!=", "\"pref\"", ":", "if", "not", "type", ".", "lower", "(", ")", ".", "startswith", "(", "\"x-\"", ")", ":", "type_list", ".", "append", "(", "type", ")", "elif", "type", "[", "2", ":", "]", ".", "lower", "(", ")", "not", "in", "[", "x", ".", "lower", "(", ")", "for", "x", "in", "type_list", "]", ":", "# add x-custom type in case it's not already added by", "# custom label for loop above but strip x- before", "type_list", ".", "append", "(", "type", "[", "2", ":", "]", ")", "# try to get pref parameter from vcard version 4.0", "try", ":", "type_list", ".", "append", "(", "\"pref=%d\"", "%", "int", "(", "object", ".", "params", ".", "get", "(", "\"PREF\"", ")", "[", "0", "]", ")", ")", "except", "(", "IndexError", ",", "TypeError", ",", "ValueError", ")", ":", "# else try to determine, if type params contain pref attribute", "try", ":", "for", "x", "in", "object", ".", "params", ".", "get", "(", "\"TYPE\"", ")", ":", "if", "x", ".", "lower", "(", ")", "==", "\"pref\"", "and", "\"pref\"", "not", "in", "type_list", ":", "type_list", ".", "append", "(", "\"pref\"", ")", "except", "TypeError", ":", "pass", "# return type_list or default type", "if", "type_list", ":", "return", "type_list", "return", "[", "default_type", "]" ]
0f69430c2680f1ff5f073a977a3c5b753b96cc17
test
CarddavObject._parse_type_value
Parse type value of phone numbers, email and post addresses. :param types: list of type values :type types: list(str) :param value: the corresponding label, required for more verbose exceptions :type value: str :param supported_types: all allowed standard types :type supported_types: list(str) :returns: tuple of standard and custom types and pref integer :rtype: tuple(list(str), list(str), int)
khard/carddav_object.py
def _parse_type_value(types, value, supported_types): """Parse type value of phone numbers, email and post addresses. :param types: list of type values :type types: list(str) :param value: the corresponding label, required for more verbose exceptions :type value: str :param supported_types: all allowed standard types :type supported_types: list(str) :returns: tuple of standard and custom types and pref integer :rtype: tuple(list(str), list(str), int) """ custom_types = [] standard_types = [] pref = 0 for type in types: type = type.strip() if type: if type.lower() in supported_types: standard_types.append(type) elif type.lower() == "pref": pref += 1 elif re.match(r"^pref=\d{1,2}$", type.lower()): pref += int(type.split("=")[1]) else: if type.lower().startswith("x-"): custom_types.append(type[2:]) standard_types.append(type) else: custom_types.append(type) standard_types.append("X-{}".format(type)) return (standard_types, custom_types, pref)
def _parse_type_value(types, value, supported_types): """Parse type value of phone numbers, email and post addresses. :param types: list of type values :type types: list(str) :param value: the corresponding label, required for more verbose exceptions :type value: str :param supported_types: all allowed standard types :type supported_types: list(str) :returns: tuple of standard and custom types and pref integer :rtype: tuple(list(str), list(str), int) """ custom_types = [] standard_types = [] pref = 0 for type in types: type = type.strip() if type: if type.lower() in supported_types: standard_types.append(type) elif type.lower() == "pref": pref += 1 elif re.match(r"^pref=\d{1,2}$", type.lower()): pref += int(type.split("=")[1]) else: if type.lower().startswith("x-"): custom_types.append(type[2:]) standard_types.append(type) else: custom_types.append(type) standard_types.append("X-{}".format(type)) return (standard_types, custom_types, pref)
[ "Parse", "type", "value", "of", "phone", "numbers", "email", "and", "post", "addresses", "." ]
scheibler/khard
python
https://github.com/scheibler/khard/blob/0f69430c2680f1ff5f073a977a3c5b753b96cc17/khard/carddav_object.py#L1556-L1589
[ "def", "_parse_type_value", "(", "types", ",", "value", ",", "supported_types", ")", ":", "custom_types", "=", "[", "]", "standard_types", "=", "[", "]", "pref", "=", "0", "for", "type", "in", "types", ":", "type", "=", "type", ".", "strip", "(", ")", "if", "type", ":", "if", "type", ".", "lower", "(", ")", "in", "supported_types", ":", "standard_types", ".", "append", "(", "type", ")", "elif", "type", ".", "lower", "(", ")", "==", "\"pref\"", ":", "pref", "+=", "1", "elif", "re", ".", "match", "(", "r\"^pref=\\d{1,2}$\"", ",", "type", ".", "lower", "(", ")", ")", ":", "pref", "+=", "int", "(", "type", ".", "split", "(", "\"=\"", ")", "[", "1", "]", ")", "else", ":", "if", "type", ".", "lower", "(", ")", ".", "startswith", "(", "\"x-\"", ")", ":", "custom_types", ".", "append", "(", "type", "[", "2", ":", "]", ")", "standard_types", ".", "append", "(", "type", ")", "else", ":", "custom_types", ".", "append", "(", "type", ")", "standard_types", ".", "append", "(", "\"X-{}\"", ".", "format", "(", "type", ")", ")", "return", "(", "standard_types", ",", "custom_types", ",", "pref", ")" ]
0f69430c2680f1ff5f073a977a3c5b753b96cc17
test
list_to_string
converts list to string recursively so that nested lists are supported :param input: a list of strings and lists of strings (and so on recursive) :type input: list :param delimiter: the deimiter to use when joining the items :type delimiter: str :returns: the recursively joined list :rtype: str
khard/helpers.py
def list_to_string(input, delimiter): """converts list to string recursively so that nested lists are supported :param input: a list of strings and lists of strings (and so on recursive) :type input: list :param delimiter: the deimiter to use when joining the items :type delimiter: str :returns: the recursively joined list :rtype: str """ if isinstance(input, list): return delimiter.join( list_to_string(item, delimiter) for item in input) return input
def list_to_string(input, delimiter): """converts list to string recursively so that nested lists are supported :param input: a list of strings and lists of strings (and so on recursive) :type input: list :param delimiter: the deimiter to use when joining the items :type delimiter: str :returns: the recursively joined list :rtype: str """ if isinstance(input, list): return delimiter.join( list_to_string(item, delimiter) for item in input) return input
[ "converts", "list", "to", "string", "recursively", "so", "that", "nested", "lists", "are", "supported" ]
scheibler/khard
python
https://github.com/scheibler/khard/blob/0f69430c2680f1ff5f073a977a3c5b753b96cc17/khard/helpers.py#L59-L72
[ "def", "list_to_string", "(", "input", ",", "delimiter", ")", ":", "if", "isinstance", "(", "input", ",", "list", ")", ":", "return", "delimiter", ".", "join", "(", "list_to_string", "(", "item", ",", "delimiter", ")", "for", "item", "in", "input", ")", "return", "input" ]
0f69430c2680f1ff5f073a977a3c5b753b96cc17
test
string_to_date
Convert string to date object. :param input: the date string to parse :type input: str :returns: the parsed datetime object :rtype: datetime.datetime
khard/helpers.py
def string_to_date(input): """Convert string to date object. :param input: the date string to parse :type input: str :returns: the parsed datetime object :rtype: datetime.datetime """ # try date formats --mmdd, --mm-dd, yyyymmdd, yyyy-mm-dd and datetime # formats yyyymmddThhmmss, yyyy-mm-ddThh:mm:ss, yyyymmddThhmmssZ, # yyyy-mm-ddThh:mm:ssZ. for format_string in ("--%m%d", "--%m-%d", "%Y%m%d", "%Y-%m-%d", "%Y%m%dT%H%M%S", "%Y-%m-%dT%H:%M:%S", "%Y%m%dT%H%M%SZ", "%Y-%m-%dT%H:%M:%SZ"): try: return datetime.strptime(input, format_string) except ValueError: pass # try datetime formats yyyymmddThhmmsstz and yyyy-mm-ddThh:mm:sstz where tz # may look like -06:00. for format_string in ("%Y%m%dT%H%M%S%z", "%Y-%m-%dT%H:%M:%S%z"): try: return datetime.strptime(''.join(input.rsplit(":", 1)), format_string) except ValueError: pass raise ValueError
def string_to_date(input): """Convert string to date object. :param input: the date string to parse :type input: str :returns: the parsed datetime object :rtype: datetime.datetime """ # try date formats --mmdd, --mm-dd, yyyymmdd, yyyy-mm-dd and datetime # formats yyyymmddThhmmss, yyyy-mm-ddThh:mm:ss, yyyymmddThhmmssZ, # yyyy-mm-ddThh:mm:ssZ. for format_string in ("--%m%d", "--%m-%d", "%Y%m%d", "%Y-%m-%d", "%Y%m%dT%H%M%S", "%Y-%m-%dT%H:%M:%S", "%Y%m%dT%H%M%SZ", "%Y-%m-%dT%H:%M:%SZ"): try: return datetime.strptime(input, format_string) except ValueError: pass # try datetime formats yyyymmddThhmmsstz and yyyy-mm-ddThh:mm:sstz where tz # may look like -06:00. for format_string in ("%Y%m%dT%H%M%S%z", "%Y-%m-%dT%H:%M:%S%z"): try: return datetime.strptime(''.join(input.rsplit(":", 1)), format_string) except ValueError: pass raise ValueError
[ "Convert", "string", "to", "date", "object", "." ]
scheibler/khard
python
https://github.com/scheibler/khard/blob/0f69430c2680f1ff5f073a977a3c5b753b96cc17/khard/helpers.py#L81-L107
[ "def", "string_to_date", "(", "input", ")", ":", "# try date formats --mmdd, --mm-dd, yyyymmdd, yyyy-mm-dd and datetime", "# formats yyyymmddThhmmss, yyyy-mm-ddThh:mm:ss, yyyymmddThhmmssZ,", "# yyyy-mm-ddThh:mm:ssZ.", "for", "format_string", "in", "(", "\"--%m%d\"", ",", "\"--%m-%d\"", ",", "\"%Y%m%d\"", ",", "\"%Y-%m-%d\"", ",", "\"%Y%m%dT%H%M%S\"", ",", "\"%Y-%m-%dT%H:%M:%S\"", ",", "\"%Y%m%dT%H%M%SZ\"", ",", "\"%Y-%m-%dT%H:%M:%SZ\"", ")", ":", "try", ":", "return", "datetime", ".", "strptime", "(", "input", ",", "format_string", ")", "except", "ValueError", ":", "pass", "# try datetime formats yyyymmddThhmmsstz and yyyy-mm-ddThh:mm:sstz where tz", "# may look like -06:00.", "for", "format_string", "in", "(", "\"%Y%m%dT%H%M%S%z\"", ",", "\"%Y-%m-%dT%H:%M:%S%z\"", ")", ":", "try", ":", "return", "datetime", ".", "strptime", "(", "''", ".", "join", "(", "input", ".", "rsplit", "(", "\":\"", ",", "1", ")", ")", ",", "format_string", ")", "except", "ValueError", ":", "pass", "raise", "ValueError" ]
0f69430c2680f1ff5f073a977a3c5b753b96cc17
test
convert_to_yaml
converts a value list into yaml syntax :param name: name of object (example: phone) :type name: str :param value: object contents :type value: str, list(str), list(list(str)) :param indentation: indent all by number of spaces :type indentation: int :param indexOfColon: use to position : at the name string (-1 for no space) :type indexOfColon: int :param show_multi_line_character: option to hide "|" :type show_multi_line_character: boolean :returns: yaml formatted string array of name, value pair :rtype: list(str)
khard/helpers.py
def convert_to_yaml( name, value, indentation, indexOfColon, show_multi_line_character): """converts a value list into yaml syntax :param name: name of object (example: phone) :type name: str :param value: object contents :type value: str, list(str), list(list(str)) :param indentation: indent all by number of spaces :type indentation: int :param indexOfColon: use to position : at the name string (-1 for no space) :type indexOfColon: int :param show_multi_line_character: option to hide "|" :type show_multi_line_character: boolean :returns: yaml formatted string array of name, value pair :rtype: list(str) """ strings = [] if isinstance(value, list): # special case for single item lists: if len(value) == 1 \ and isinstance(value[0], str): # value = ["string"] should not be converted to # name: # - string # but to "name: string" instead value = value[0] elif len(value) == 1 \ and isinstance(value[0], list) \ and len(value[0]) == 1 \ and isinstance(value[0][0], str): # same applies to value = [["string"]] value = value[0][0] if isinstance(value, str): strings.append("%s%s%s: %s" % ( ' ' * indentation, name, ' ' * (indexOfColon-len(name)), indent_multiline_string(value, indentation+4, show_multi_line_character))) elif isinstance(value, list): strings.append("%s%s%s: " % ( ' ' * indentation, name, ' ' * (indexOfColon-len(name)))) for outer in value: # special case for single item sublists if isinstance(outer, list) \ and len(outer) == 1 \ and isinstance(outer[0], str): # outer = ["string"] should not be converted to # - # - string # but to "- string" instead outer = outer[0] if isinstance(outer, str): strings.append("%s- %s" % ( ' ' * (indentation+4), indent_multiline_string( outer, indentation+8, show_multi_line_character))) elif isinstance(outer, list): strings.append("%s- " % (' ' * (indentation+4))) for inner in outer: if isinstance(inner, str): strings.append("%s- %s" % ( ' ' * (indentation+8), indent_multiline_string( inner, indentation+12, show_multi_line_character))) return strings
def convert_to_yaml( name, value, indentation, indexOfColon, show_multi_line_character): """converts a value list into yaml syntax :param name: name of object (example: phone) :type name: str :param value: object contents :type value: str, list(str), list(list(str)) :param indentation: indent all by number of spaces :type indentation: int :param indexOfColon: use to position : at the name string (-1 for no space) :type indexOfColon: int :param show_multi_line_character: option to hide "|" :type show_multi_line_character: boolean :returns: yaml formatted string array of name, value pair :rtype: list(str) """ strings = [] if isinstance(value, list): # special case for single item lists: if len(value) == 1 \ and isinstance(value[0], str): # value = ["string"] should not be converted to # name: # - string # but to "name: string" instead value = value[0] elif len(value) == 1 \ and isinstance(value[0], list) \ and len(value[0]) == 1 \ and isinstance(value[0][0], str): # same applies to value = [["string"]] value = value[0][0] if isinstance(value, str): strings.append("%s%s%s: %s" % ( ' ' * indentation, name, ' ' * (indexOfColon-len(name)), indent_multiline_string(value, indentation+4, show_multi_line_character))) elif isinstance(value, list): strings.append("%s%s%s: " % ( ' ' * indentation, name, ' ' * (indexOfColon-len(name)))) for outer in value: # special case for single item sublists if isinstance(outer, list) \ and len(outer) == 1 \ and isinstance(outer[0], str): # outer = ["string"] should not be converted to # - # - string # but to "- string" instead outer = outer[0] if isinstance(outer, str): strings.append("%s- %s" % ( ' ' * (indentation+4), indent_multiline_string( outer, indentation+8, show_multi_line_character))) elif isinstance(outer, list): strings.append("%s- " % (' ' * (indentation+4))) for inner in outer: if isinstance(inner, str): strings.append("%s- %s" % ( ' ' * (indentation+8), indent_multiline_string( inner, indentation+12, show_multi_line_character))) return strings
[ "converts", "a", "value", "list", "into", "yaml", "syntax", ":", "param", "name", ":", "name", "of", "object", "(", "example", ":", "phone", ")", ":", "type", "name", ":", "str", ":", "param", "value", ":", "object", "contents", ":", "type", "value", ":", "str", "list", "(", "str", ")", "list", "(", "list", "(", "str", "))", ":", "param", "indentation", ":", "indent", "all", "by", "number", "of", "spaces", ":", "type", "indentation", ":", "int", ":", "param", "indexOfColon", ":", "use", "to", "position", ":", "at", "the", "name", "string", "(", "-", "1", "for", "no", "space", ")", ":", "type", "indexOfColon", ":", "int", ":", "param", "show_multi_line_character", ":", "option", "to", "hide", "|", ":", "type", "show_multi_line_character", ":", "boolean", ":", "returns", ":", "yaml", "formatted", "string", "array", "of", "name", "value", "pair", ":", "rtype", ":", "list", "(", "str", ")" ]
scheibler/khard
python
https://github.com/scheibler/khard/blob/0f69430c2680f1ff5f073a977a3c5b753b96cc17/khard/helpers.py#L120-L182
[ "def", "convert_to_yaml", "(", "name", ",", "value", ",", "indentation", ",", "indexOfColon", ",", "show_multi_line_character", ")", ":", "strings", "=", "[", "]", "if", "isinstance", "(", "value", ",", "list", ")", ":", "# special case for single item lists:", "if", "len", "(", "value", ")", "==", "1", "and", "isinstance", "(", "value", "[", "0", "]", ",", "str", ")", ":", "# value = [\"string\"] should not be converted to", "# name:", "# - string", "# but to \"name: string\" instead", "value", "=", "value", "[", "0", "]", "elif", "len", "(", "value", ")", "==", "1", "and", "isinstance", "(", "value", "[", "0", "]", ",", "list", ")", "and", "len", "(", "value", "[", "0", "]", ")", "==", "1", "and", "isinstance", "(", "value", "[", "0", "]", "[", "0", "]", ",", "str", ")", ":", "# same applies to value = [[\"string\"]]", "value", "=", "value", "[", "0", "]", "[", "0", "]", "if", "isinstance", "(", "value", ",", "str", ")", ":", "strings", ".", "append", "(", "\"%s%s%s: %s\"", "%", "(", "' '", "*", "indentation", ",", "name", ",", "' '", "*", "(", "indexOfColon", "-", "len", "(", "name", ")", ")", ",", "indent_multiline_string", "(", "value", ",", "indentation", "+", "4", ",", "show_multi_line_character", ")", ")", ")", "elif", "isinstance", "(", "value", ",", "list", ")", ":", "strings", ".", "append", "(", "\"%s%s%s: \"", "%", "(", "' '", "*", "indentation", ",", "name", ",", "' '", "*", "(", "indexOfColon", "-", "len", "(", "name", ")", ")", ")", ")", "for", "outer", "in", "value", ":", "# special case for single item sublists", "if", "isinstance", "(", "outer", ",", "list", ")", "and", "len", "(", "outer", ")", "==", "1", "and", "isinstance", "(", "outer", "[", "0", "]", ",", "str", ")", ":", "# outer = [\"string\"] should not be converted to", "# -", "# - string", "# but to \"- string\" instead", "outer", "=", "outer", "[", "0", "]", "if", "isinstance", "(", "outer", ",", "str", ")", ":", "strings", ".", "append", "(", "\"%s- %s\"", "%", "(", "' '", "*", "(", "indentation", "+", "4", ")", ",", "indent_multiline_string", "(", "outer", ",", "indentation", "+", "8", ",", "show_multi_line_character", ")", ")", ")", "elif", "isinstance", "(", "outer", ",", "list", ")", ":", "strings", ".", "append", "(", "\"%s- \"", "%", "(", "' '", "*", "(", "indentation", "+", "4", ")", ")", ")", "for", "inner", "in", "outer", ":", "if", "isinstance", "(", "inner", ",", "str", ")", ":", "strings", ".", "append", "(", "\"%s- %s\"", "%", "(", "' '", "*", "(", "indentation", "+", "8", ")", ",", "indent_multiline_string", "(", "inner", ",", "indentation", "+", "12", ",", "show_multi_line_character", ")", ")", ")", "return", "strings" ]
0f69430c2680f1ff5f073a977a3c5b753b96cc17
test
convert_to_vcard
converts user input into vcard compatible data structures :param name: object name, only required for error messages :type name: str :param value: user input :type value: str or list(str) :param allowed_object_type: set the accepted return type for vcard attribute :type allowed_object_type: enum of type ObjectType :returns: cleaned user input, ready for vcard or a ValueError :rtype: str or list(str)
khard/helpers.py
def convert_to_vcard(name, value, allowed_object_type): """converts user input into vcard compatible data structures :param name: object name, only required for error messages :type name: str :param value: user input :type value: str or list(str) :param allowed_object_type: set the accepted return type for vcard attribute :type allowed_object_type: enum of type ObjectType :returns: cleaned user input, ready for vcard or a ValueError :rtype: str or list(str) """ if isinstance(value, str): if allowed_object_type == ObjectType.list_with_strings: raise ValueError( "Error: " + name + " must not contain a single string.") else: return value.strip() elif isinstance(value, list): if allowed_object_type == ObjectType.string: raise ValueError( "Error: " + name + " must not contain a list.") else: for entry in value: if not isinstance(entry, str): raise ValueError( "Error: " + name + " must not contain a nested list") # filter out empty list items and strip leading and trailing space return [x.strip() for x in value if x] else: if allowed_object_type == ObjectType.string: raise ValueError( "Error: " + name + " must be a string.") elif allowed_object_type == ObjectType.list_with_strings: raise ValueError( "Error: " + name + " must be a list with strings.") else: raise ValueError( "Error: " + name + " must be a string or a list with strings.")
def convert_to_vcard(name, value, allowed_object_type): """converts user input into vcard compatible data structures :param name: object name, only required for error messages :type name: str :param value: user input :type value: str or list(str) :param allowed_object_type: set the accepted return type for vcard attribute :type allowed_object_type: enum of type ObjectType :returns: cleaned user input, ready for vcard or a ValueError :rtype: str or list(str) """ if isinstance(value, str): if allowed_object_type == ObjectType.list_with_strings: raise ValueError( "Error: " + name + " must not contain a single string.") else: return value.strip() elif isinstance(value, list): if allowed_object_type == ObjectType.string: raise ValueError( "Error: " + name + " must not contain a list.") else: for entry in value: if not isinstance(entry, str): raise ValueError( "Error: " + name + " must not contain a nested list") # filter out empty list items and strip leading and trailing space return [x.strip() for x in value if x] else: if allowed_object_type == ObjectType.string: raise ValueError( "Error: " + name + " must be a string.") elif allowed_object_type == ObjectType.list_with_strings: raise ValueError( "Error: " + name + " must be a list with strings.") else: raise ValueError( "Error: " + name + " must be a string or a list with strings.")
[ "converts", "user", "input", "into", "vcard", "compatible", "data", "structures", ":", "param", "name", ":", "object", "name", "only", "required", "for", "error", "messages", ":", "type", "name", ":", "str", ":", "param", "value", ":", "user", "input", ":", "type", "value", ":", "str", "or", "list", "(", "str", ")", ":", "param", "allowed_object_type", ":", "set", "the", "accepted", "return", "type", "for", "vcard", "attribute", ":", "type", "allowed_object_type", ":", "enum", "of", "type", "ObjectType", ":", "returns", ":", "cleaned", "user", "input", "ready", "for", "vcard", "or", "a", "ValueError", ":", "rtype", ":", "str", "or", "list", "(", "str", ")" ]
scheibler/khard
python
https://github.com/scheibler/khard/blob/0f69430c2680f1ff5f073a977a3c5b753b96cc17/khard/helpers.py#L185-L223
[ "def", "convert_to_vcard", "(", "name", ",", "value", ",", "allowed_object_type", ")", ":", "if", "isinstance", "(", "value", ",", "str", ")", ":", "if", "allowed_object_type", "==", "ObjectType", ".", "list_with_strings", ":", "raise", "ValueError", "(", "\"Error: \"", "+", "name", "+", "\" must not contain a single string.\"", ")", "else", ":", "return", "value", ".", "strip", "(", ")", "elif", "isinstance", "(", "value", ",", "list", ")", ":", "if", "allowed_object_type", "==", "ObjectType", ".", "string", ":", "raise", "ValueError", "(", "\"Error: \"", "+", "name", "+", "\" must not contain a list.\"", ")", "else", ":", "for", "entry", "in", "value", ":", "if", "not", "isinstance", "(", "entry", ",", "str", ")", ":", "raise", "ValueError", "(", "\"Error: \"", "+", "name", "+", "\" must not contain a nested list\"", ")", "# filter out empty list items and strip leading and trailing space", "return", "[", "x", ".", "strip", "(", ")", "for", "x", "in", "value", "if", "x", "]", "else", ":", "if", "allowed_object_type", "==", "ObjectType", ".", "string", ":", "raise", "ValueError", "(", "\"Error: \"", "+", "name", "+", "\" must be a string.\"", ")", "elif", "allowed_object_type", "==", "ObjectType", ".", "list_with_strings", ":", "raise", "ValueError", "(", "\"Error: \"", "+", "name", "+", "\" must be a list with strings.\"", ")", "else", ":", "raise", "ValueError", "(", "\"Error: \"", "+", "name", "+", "\" must be a string or a list with strings.\"", ")" ]
0f69430c2680f1ff5f073a977a3c5b753b96cc17
test
AddressBook._compare_uids
Calculate the minimum length of initial substrings of uid1 and uid2 for them to be different. :param uid1: first uid to compare :type uid1: str :param uid2: second uid to compare :type uid2: str :returns: the length of the shortes unequal initial substrings :rtype: int
khard/address_book.py
def _compare_uids(uid1, uid2): """Calculate the minimum length of initial substrings of uid1 and uid2 for them to be different. :param uid1: first uid to compare :type uid1: str :param uid2: second uid to compare :type uid2: str :returns: the length of the shortes unequal initial substrings :rtype: int """ sum = 0 for char1, char2 in zip(uid1, uid2): if char1 == char2: sum += 1 else: break return sum
def _compare_uids(uid1, uid2): """Calculate the minimum length of initial substrings of uid1 and uid2 for them to be different. :param uid1: first uid to compare :type uid1: str :param uid2: second uid to compare :type uid2: str :returns: the length of the shortes unequal initial substrings :rtype: int """ sum = 0 for char1, char2 in zip(uid1, uid2): if char1 == char2: sum += 1 else: break return sum
[ "Calculate", "the", "minimum", "length", "of", "initial", "substrings", "of", "uid1", "and", "uid2", "for", "them", "to", "be", "different", "." ]
scheibler/khard
python
https://github.com/scheibler/khard/blob/0f69430c2680f1ff5f073a977a3c5b753b96cc17/khard/address_book.py#L59-L76
[ "def", "_compare_uids", "(", "uid1", ",", "uid2", ")", ":", "sum", "=", "0", "for", "char1", ",", "char2", "in", "zip", "(", "uid1", ",", "uid2", ")", ":", "if", "char1", "==", "char2", ":", "sum", "+=", "1", "else", ":", "break", "return", "sum" ]
0f69430c2680f1ff5f073a977a3c5b753b96cc17
test
AddressBook._search_all
Search in all fields for contacts matching query. :param query: the query to search for :type query: str :yields: all found contacts :rtype: generator(carddav_object.CarddavObject)
khard/address_book.py
def _search_all(self, query): """Search in all fields for contacts matching query. :param query: the query to search for :type query: str :yields: all found contacts :rtype: generator(carddav_object.CarddavObject) """ regexp = re.compile(query, re.IGNORECASE | re.DOTALL) for contact in self.contacts.values(): # search in all contact fields contact_details = contact.print_vcard() if regexp.search(contact_details) is not None: yield contact else: # find phone numbers with special chars like / clean_contact_details = re.sub("[^a-zA-Z0-9\n]", "", contact_details) if regexp.search(clean_contact_details) is not None \ and len(re.sub("\D", "", query)) >= 3: yield contact
def _search_all(self, query): """Search in all fields for contacts matching query. :param query: the query to search for :type query: str :yields: all found contacts :rtype: generator(carddav_object.CarddavObject) """ regexp = re.compile(query, re.IGNORECASE | re.DOTALL) for contact in self.contacts.values(): # search in all contact fields contact_details = contact.print_vcard() if regexp.search(contact_details) is not None: yield contact else: # find phone numbers with special chars like / clean_contact_details = re.sub("[^a-zA-Z0-9\n]", "", contact_details) if regexp.search(clean_contact_details) is not None \ and len(re.sub("\D", "", query)) >= 3: yield contact
[ "Search", "in", "all", "fields", "for", "contacts", "matching", "query", "." ]
scheibler/khard
python
https://github.com/scheibler/khard/blob/0f69430c2680f1ff5f073a977a3c5b753b96cc17/khard/address_book.py#L78-L99
[ "def", "_search_all", "(", "self", ",", "query", ")", ":", "regexp", "=", "re", ".", "compile", "(", "query", ",", "re", ".", "IGNORECASE", "|", "re", ".", "DOTALL", ")", "for", "contact", "in", "self", ".", "contacts", ".", "values", "(", ")", ":", "# search in all contact fields", "contact_details", "=", "contact", ".", "print_vcard", "(", ")", "if", "regexp", ".", "search", "(", "contact_details", ")", "is", "not", "None", ":", "yield", "contact", "else", ":", "# find phone numbers with special chars like /", "clean_contact_details", "=", "re", ".", "sub", "(", "\"[^a-zA-Z0-9\\n]\"", ",", "\"\"", ",", "contact_details", ")", "if", "regexp", ".", "search", "(", "clean_contact_details", ")", "is", "not", "None", "and", "len", "(", "re", ".", "sub", "(", "\"\\D\"", ",", "\"\"", ",", "query", ")", ")", ">=", "3", ":", "yield", "contact" ]
0f69430c2680f1ff5f073a977a3c5b753b96cc17
test
AddressBook._search_names
Search in the name filed for contacts matching query. :param query: the query to search for :type query: str :yields: all found contacts :rtype: generator(carddav_object.CarddavObject)
khard/address_book.py
def _search_names(self, query): """Search in the name filed for contacts matching query. :param query: the query to search for :type query: str :yields: all found contacts :rtype: generator(carddav_object.CarddavObject) """ regexp = re.compile(query, re.IGNORECASE | re.DOTALL) for contact in self.contacts.values(): # only search in contact name if regexp.search(contact.get_full_name()) is not None: yield contact
def _search_names(self, query): """Search in the name filed for contacts matching query. :param query: the query to search for :type query: str :yields: all found contacts :rtype: generator(carddav_object.CarddavObject) """ regexp = re.compile(query, re.IGNORECASE | re.DOTALL) for contact in self.contacts.values(): # only search in contact name if regexp.search(contact.get_full_name()) is not None: yield contact
[ "Search", "in", "the", "name", "filed", "for", "contacts", "matching", "query", "." ]
scheibler/khard
python
https://github.com/scheibler/khard/blob/0f69430c2680f1ff5f073a977a3c5b753b96cc17/khard/address_book.py#L101-L114
[ "def", "_search_names", "(", "self", ",", "query", ")", ":", "regexp", "=", "re", ".", "compile", "(", "query", ",", "re", ".", "IGNORECASE", "|", "re", ".", "DOTALL", ")", "for", "contact", "in", "self", ".", "contacts", ".", "values", "(", ")", ":", "# only search in contact name", "if", "regexp", ".", "search", "(", "contact", ".", "get_full_name", "(", ")", ")", "is", "not", "None", ":", "yield", "contact" ]
0f69430c2680f1ff5f073a977a3c5b753b96cc17
test
AddressBook._search_uid
Search for contacts with a matching uid. :param query: the query to search for :type query: str :yields: all found contacts :rtype: generator(carddav_object.CarddavObject)
khard/address_book.py
def _search_uid(self, query): """Search for contacts with a matching uid. :param query: the query to search for :type query: str :yields: all found contacts :rtype: generator(carddav_object.CarddavObject) """ try: # First we treat the argument as a full UID and try to match it # exactly. yield self.contacts[query] except KeyError: # If that failed we look for all contacts whos UID start with the # given query. for uid in self.contacts: if uid.startswith(query): yield self.contacts[uid]
def _search_uid(self, query): """Search for contacts with a matching uid. :param query: the query to search for :type query: str :yields: all found contacts :rtype: generator(carddav_object.CarddavObject) """ try: # First we treat the argument as a full UID and try to match it # exactly. yield self.contacts[query] except KeyError: # If that failed we look for all contacts whos UID start with the # given query. for uid in self.contacts: if uid.startswith(query): yield self.contacts[uid]
[ "Search", "for", "contacts", "with", "a", "matching", "uid", "." ]
scheibler/khard
python
https://github.com/scheibler/khard/blob/0f69430c2680f1ff5f073a977a3c5b753b96cc17/khard/address_book.py#L116-L134
[ "def", "_search_uid", "(", "self", ",", "query", ")", ":", "try", ":", "# First we treat the argument as a full UID and try to match it", "# exactly.", "yield", "self", ".", "contacts", "[", "query", "]", "except", "KeyError", ":", "# If that failed we look for all contacts whos UID start with the", "# given query.", "for", "uid", "in", "self", ".", "contacts", ":", "if", "uid", ".", "startswith", "(", "query", ")", ":", "yield", "self", ".", "contacts", "[", "uid", "]" ]
0f69430c2680f1ff5f073a977a3c5b753b96cc17
test
AddressBook.search
Search this address book for contacts matching the query. The method can be one of "all", "name" and "uid". The backend for this address book migth be load()ed if needed. :param query: the query to search for :type query: str :param method: the type of fileds to use when seaching :type method: str :returns: all found contacts :rtype: list(carddav_object.CarddavObject)
khard/address_book.py
def search(self, query, method="all"): """Search this address book for contacts matching the query. The method can be one of "all", "name" and "uid". The backend for this address book migth be load()ed if needed. :param query: the query to search for :type query: str :param method: the type of fileds to use when seaching :type method: str :returns: all found contacts :rtype: list(carddav_object.CarddavObject) """ logging.debug('address book %s, searching with %s', self.name, query) if not self._loaded: self.load(query) if method == "all": search_function = self._search_all elif method == "name": search_function = self._search_names elif method == "uid": search_function = self._search_uid else: raise ValueError('Only the search methods "all", "name" and "uid" ' 'are supported.') return list(search_function(query))
def search(self, query, method="all"): """Search this address book for contacts matching the query. The method can be one of "all", "name" and "uid". The backend for this address book migth be load()ed if needed. :param query: the query to search for :type query: str :param method: the type of fileds to use when seaching :type method: str :returns: all found contacts :rtype: list(carddav_object.CarddavObject) """ logging.debug('address book %s, searching with %s', self.name, query) if not self._loaded: self.load(query) if method == "all": search_function = self._search_all elif method == "name": search_function = self._search_names elif method == "uid": search_function = self._search_uid else: raise ValueError('Only the search methods "all", "name" and "uid" ' 'are supported.') return list(search_function(query))
[ "Search", "this", "address", "book", "for", "contacts", "matching", "the", "query", "." ]
scheibler/khard
python
https://github.com/scheibler/khard/blob/0f69430c2680f1ff5f073a977a3c5b753b96cc17/khard/address_book.py#L136-L162
[ "def", "search", "(", "self", ",", "query", ",", "method", "=", "\"all\"", ")", ":", "logging", ".", "debug", "(", "'address book %s, searching with %s'", ",", "self", ".", "name", ",", "query", ")", "if", "not", "self", ".", "_loaded", ":", "self", ".", "load", "(", "query", ")", "if", "method", "==", "\"all\"", ":", "search_function", "=", "self", ".", "_search_all", "elif", "method", "==", "\"name\"", ":", "search_function", "=", "self", ".", "_search_names", "elif", "method", "==", "\"uid\"", ":", "search_function", "=", "self", ".", "_search_uid", "else", ":", "raise", "ValueError", "(", "'Only the search methods \"all\", \"name\" and \"uid\" '", "'are supported.'", ")", "return", "list", "(", "search_function", "(", "query", ")", ")" ]
0f69430c2680f1ff5f073a977a3c5b753b96cc17
test
AddressBook.get_short_uid_dict
Create a dictionary of shortend UIDs for all contacts. All arguments are only used if the address book is not yet initialized and will just be handed to self.load(). :param query: see self.load() :type query: str :returns: the contacts mapped by the shortes unique prefix of their UID :rtype: dict(str: CarddavObject)
khard/address_book.py
def get_short_uid_dict(self, query=None): """Create a dictionary of shortend UIDs for all contacts. All arguments are only used if the address book is not yet initialized and will just be handed to self.load(). :param query: see self.load() :type query: str :returns: the contacts mapped by the shortes unique prefix of their UID :rtype: dict(str: CarddavObject) """ if self._short_uids is None: if not self._loaded: self.load(query) if not self.contacts: self._short_uids = {} elif len(self.contacts) == 1: self._short_uids = {uid[0:1]: contact for uid, contact in self.contacts.items()} else: self._short_uids = {} sorted_uids = sorted(self.contacts) # Prepare for the loop; the first and last items are handled # seperatly. item0, item1 = sorted_uids[:2] same1 = self._compare_uids(item0, item1) self._short_uids[item0[:same1 + 1]] = self.contacts[item0] for item_new in sorted_uids[2:]: # shift the items and the common prefix lenght one further item0, item1 = item1, item_new same0, same1 = same1, self._compare_uids(item0, item1) # compute the final prefix length for item1 same = max(same0, same1) self._short_uids[item0[:same + 1]] = self.contacts[item0] # Save the last item. self._short_uids[item1[:same1 + 1]] = self.contacts[item1] return self._short_uids
def get_short_uid_dict(self, query=None): """Create a dictionary of shortend UIDs for all contacts. All arguments are only used if the address book is not yet initialized and will just be handed to self.load(). :param query: see self.load() :type query: str :returns: the contacts mapped by the shortes unique prefix of their UID :rtype: dict(str: CarddavObject) """ if self._short_uids is None: if not self._loaded: self.load(query) if not self.contacts: self._short_uids = {} elif len(self.contacts) == 1: self._short_uids = {uid[0:1]: contact for uid, contact in self.contacts.items()} else: self._short_uids = {} sorted_uids = sorted(self.contacts) # Prepare for the loop; the first and last items are handled # seperatly. item0, item1 = sorted_uids[:2] same1 = self._compare_uids(item0, item1) self._short_uids[item0[:same1 + 1]] = self.contacts[item0] for item_new in sorted_uids[2:]: # shift the items and the common prefix lenght one further item0, item1 = item1, item_new same0, same1 = same1, self._compare_uids(item0, item1) # compute the final prefix length for item1 same = max(same0, same1) self._short_uids[item0[:same + 1]] = self.contacts[item0] # Save the last item. self._short_uids[item1[:same1 + 1]] = self.contacts[item1] return self._short_uids
[ "Create", "a", "dictionary", "of", "shortend", "UIDs", "for", "all", "contacts", "." ]
scheibler/khard
python
https://github.com/scheibler/khard/blob/0f69430c2680f1ff5f073a977a3c5b753b96cc17/khard/address_book.py#L164-L200
[ "def", "get_short_uid_dict", "(", "self", ",", "query", "=", "None", ")", ":", "if", "self", ".", "_short_uids", "is", "None", ":", "if", "not", "self", ".", "_loaded", ":", "self", ".", "load", "(", "query", ")", "if", "not", "self", ".", "contacts", ":", "self", ".", "_short_uids", "=", "{", "}", "elif", "len", "(", "self", ".", "contacts", ")", "==", "1", ":", "self", ".", "_short_uids", "=", "{", "uid", "[", "0", ":", "1", "]", ":", "contact", "for", "uid", ",", "contact", "in", "self", ".", "contacts", ".", "items", "(", ")", "}", "else", ":", "self", ".", "_short_uids", "=", "{", "}", "sorted_uids", "=", "sorted", "(", "self", ".", "contacts", ")", "# Prepare for the loop; the first and last items are handled", "# seperatly.", "item0", ",", "item1", "=", "sorted_uids", "[", ":", "2", "]", "same1", "=", "self", ".", "_compare_uids", "(", "item0", ",", "item1", ")", "self", ".", "_short_uids", "[", "item0", "[", ":", "same1", "+", "1", "]", "]", "=", "self", ".", "contacts", "[", "item0", "]", "for", "item_new", "in", "sorted_uids", "[", "2", ":", "]", ":", "# shift the items and the common prefix lenght one further", "item0", ",", "item1", "=", "item1", ",", "item_new", "same0", ",", "same1", "=", "same1", ",", "self", ".", "_compare_uids", "(", "item0", ",", "item1", ")", "# compute the final prefix length for item1", "same", "=", "max", "(", "same0", ",", "same1", ")", "self", ".", "_short_uids", "[", "item0", "[", ":", "same", "+", "1", "]", "]", "=", "self", ".", "contacts", "[", "item0", "]", "# Save the last item.", "self", ".", "_short_uids", "[", "item1", "[", ":", "same1", "+", "1", "]", "]", "=", "self", ".", "contacts", "[", "item1", "]", "return", "self", ".", "_short_uids" ]
0f69430c2680f1ff5f073a977a3c5b753b96cc17
test
AddressBook.get_short_uid
Get the shortend UID for the given UID. :param uid: the full UID to shorten :type uid: str :returns: the shortend uid or the empty string :rtype: str
khard/address_book.py
def get_short_uid(self, uid): """Get the shortend UID for the given UID. :param uid: the full UID to shorten :type uid: str :returns: the shortend uid or the empty string :rtype: str """ if uid: short_uids = self.get_short_uid_dict() for length_of_uid in range(len(uid), 0, -1): if short_uids.get(uid[:length_of_uid]) is not None: return uid[:length_of_uid] return ""
def get_short_uid(self, uid): """Get the shortend UID for the given UID. :param uid: the full UID to shorten :type uid: str :returns: the shortend uid or the empty string :rtype: str """ if uid: short_uids = self.get_short_uid_dict() for length_of_uid in range(len(uid), 0, -1): if short_uids.get(uid[:length_of_uid]) is not None: return uid[:length_of_uid] return ""
[ "Get", "the", "shortend", "UID", "for", "the", "given", "UID", "." ]
scheibler/khard
python
https://github.com/scheibler/khard/blob/0f69430c2680f1ff5f073a977a3c5b753b96cc17/khard/address_book.py#L202-L215
[ "def", "get_short_uid", "(", "self", ",", "uid", ")", ":", "if", "uid", ":", "short_uids", "=", "self", ".", "get_short_uid_dict", "(", ")", "for", "length_of_uid", "in", "range", "(", "len", "(", "uid", ")", ",", "0", ",", "-", "1", ")", ":", "if", "short_uids", ".", "get", "(", "uid", "[", ":", "length_of_uid", "]", ")", "is", "not", "None", ":", "return", "uid", "[", ":", "length_of_uid", "]", "return", "\"\"" ]
0f69430c2680f1ff5f073a977a3c5b753b96cc17
test
VdirAddressBook._find_vcard_files
Find all vcard files inside this address book. If a search string is given only files which contents match that will be returned. :param search: a regular expression to limit the results :type search: str :param search_in_source_files: apply search regexp directly on the .vcf files to speed up parsing (less accurate) :type search_in_source_files: bool :returns: the paths of the vcard files :rtype: generator
khard/address_book.py
def _find_vcard_files(self, search=None, search_in_source_files=False): """Find all vcard files inside this address book. If a search string is given only files which contents match that will be returned. :param search: a regular expression to limit the results :type search: str :param search_in_source_files: apply search regexp directly on the .vcf files to speed up parsing (less accurate) :type search_in_source_files: bool :returns: the paths of the vcard files :rtype: generator """ files = glob.glob(os.path.join(self.path, "*.vcf")) if search and search_in_source_files: for filename in files: with open(filename, "r") as filehandle: if re.search(search, filehandle.read(), re.IGNORECASE | re.DOTALL): yield filename else: yield from files
def _find_vcard_files(self, search=None, search_in_source_files=False): """Find all vcard files inside this address book. If a search string is given only files which contents match that will be returned. :param search: a regular expression to limit the results :type search: str :param search_in_source_files: apply search regexp directly on the .vcf files to speed up parsing (less accurate) :type search_in_source_files: bool :returns: the paths of the vcard files :rtype: generator """ files = glob.glob(os.path.join(self.path, "*.vcf")) if search and search_in_source_files: for filename in files: with open(filename, "r") as filehandle: if re.search(search, filehandle.read(), re.IGNORECASE | re.DOTALL): yield filename else: yield from files
[ "Find", "all", "vcard", "files", "inside", "this", "address", "book", "." ]
scheibler/khard
python
https://github.com/scheibler/khard/blob/0f69430c2680f1ff5f073a977a3c5b753b96cc17/khard/address_book.py#L254-L276
[ "def", "_find_vcard_files", "(", "self", ",", "search", "=", "None", ",", "search_in_source_files", "=", "False", ")", ":", "files", "=", "glob", ".", "glob", "(", "os", ".", "path", ".", "join", "(", "self", ".", "path", ",", "\"*.vcf\"", ")", ")", "if", "search", "and", "search_in_source_files", ":", "for", "filename", "in", "files", ":", "with", "open", "(", "filename", ",", "\"r\"", ")", "as", "filehandle", ":", "if", "re", ".", "search", "(", "search", ",", "filehandle", ".", "read", "(", ")", ",", "re", ".", "IGNORECASE", "|", "re", ".", "DOTALL", ")", ":", "yield", "filename", "else", ":", "yield", "from", "files" ]
0f69430c2680f1ff5f073a977a3c5b753b96cc17
test
VdirAddressBook.load
Load all vcard files in this address book from disk. If a search string is given only files which contents match that will be loaded. :param query: a regular expression to limit the results :type query: str :param search_in_source_files: apply search regexp directly on the .vcf files to speed up parsing (less accurate) :type search_in_source_files: bool :returns: the number of successfully loaded cards and the number of errors :rtype: int, int :throws: AddressBookParseError
khard/address_book.py
def load(self, query=None, search_in_source_files=False): """Load all vcard files in this address book from disk. If a search string is given only files which contents match that will be loaded. :param query: a regular expression to limit the results :type query: str :param search_in_source_files: apply search regexp directly on the .vcf files to speed up parsing (less accurate) :type search_in_source_files: bool :returns: the number of successfully loaded cards and the number of errors :rtype: int, int :throws: AddressBookParseError """ if self._loaded: return logging.debug('Loading Vdir %s with query %s', self.name, query) errors = 0 for filename in self._find_vcard_files( search=query, search_in_source_files=search_in_source_files): try: card = CarddavObject.from_file(self, filename, self._private_objects, self._localize_dates) except (IOError, vobject.base.ParseError) as err: verb = "open" if isinstance(err, IOError) else "parse" logging.debug("Error: Could not %s file %s\n%s", verb, filename, err) if self._skip: errors += 1 else: # FIXME: This should throw an apropriate exception and the # sys.exit should be called somewhere closer to the command # line parsing. logging.error( "The vcard file %s of address book %s could not be " "parsed\nUse --debug for more information or " "--skip-unparsable to proceed", filename, self.name) sys.exit(2) else: uid = card.get_uid() if not uid: logging.warning("Card %s from address book %s has no UID " "and will not be availbale.", card, self.name) elif uid in self.contacts: logging.warning( "Card %s and %s from address book %s have the same " "UID. The former will not be availbale.", card, self.contacts[uid], self.name) else: self.contacts[uid] = card self._loaded = True if errors: logging.warning( "%d of %d vCard files of address book %s could not be parsed.", errors, len(self.contacts) + errors, self) logging.debug('Loded %s contacts from address book %s.', len(self.contacts), self.name)
def load(self, query=None, search_in_source_files=False): """Load all vcard files in this address book from disk. If a search string is given only files which contents match that will be loaded. :param query: a regular expression to limit the results :type query: str :param search_in_source_files: apply search regexp directly on the .vcf files to speed up parsing (less accurate) :type search_in_source_files: bool :returns: the number of successfully loaded cards and the number of errors :rtype: int, int :throws: AddressBookParseError """ if self._loaded: return logging.debug('Loading Vdir %s with query %s', self.name, query) errors = 0 for filename in self._find_vcard_files( search=query, search_in_source_files=search_in_source_files): try: card = CarddavObject.from_file(self, filename, self._private_objects, self._localize_dates) except (IOError, vobject.base.ParseError) as err: verb = "open" if isinstance(err, IOError) else "parse" logging.debug("Error: Could not %s file %s\n%s", verb, filename, err) if self._skip: errors += 1 else: # FIXME: This should throw an apropriate exception and the # sys.exit should be called somewhere closer to the command # line parsing. logging.error( "The vcard file %s of address book %s could not be " "parsed\nUse --debug for more information or " "--skip-unparsable to proceed", filename, self.name) sys.exit(2) else: uid = card.get_uid() if not uid: logging.warning("Card %s from address book %s has no UID " "and will not be availbale.", card, self.name) elif uid in self.contacts: logging.warning( "Card %s and %s from address book %s have the same " "UID. The former will not be availbale.", card, self.contacts[uid], self.name) else: self.contacts[uid] = card self._loaded = True if errors: logging.warning( "%d of %d vCard files of address book %s could not be parsed.", errors, len(self.contacts) + errors, self) logging.debug('Loded %s contacts from address book %s.', len(self.contacts), self.name)
[ "Load", "all", "vcard", "files", "in", "this", "address", "book", "from", "disk", "." ]
scheibler/khard
python
https://github.com/scheibler/khard/blob/0f69430c2680f1ff5f073a977a3c5b753b96cc17/khard/address_book.py#L278-L337
[ "def", "load", "(", "self", ",", "query", "=", "None", ",", "search_in_source_files", "=", "False", ")", ":", "if", "self", ".", "_loaded", ":", "return", "logging", ".", "debug", "(", "'Loading Vdir %s with query %s'", ",", "self", ".", "name", ",", "query", ")", "errors", "=", "0", "for", "filename", "in", "self", ".", "_find_vcard_files", "(", "search", "=", "query", ",", "search_in_source_files", "=", "search_in_source_files", ")", ":", "try", ":", "card", "=", "CarddavObject", ".", "from_file", "(", "self", ",", "filename", ",", "self", ".", "_private_objects", ",", "self", ".", "_localize_dates", ")", "except", "(", "IOError", ",", "vobject", ".", "base", ".", "ParseError", ")", "as", "err", ":", "verb", "=", "\"open\"", "if", "isinstance", "(", "err", ",", "IOError", ")", "else", "\"parse\"", "logging", ".", "debug", "(", "\"Error: Could not %s file %s\\n%s\"", ",", "verb", ",", "filename", ",", "err", ")", "if", "self", ".", "_skip", ":", "errors", "+=", "1", "else", ":", "# FIXME: This should throw an apropriate exception and the", "# sys.exit should be called somewhere closer to the command", "# line parsing.", "logging", ".", "error", "(", "\"The vcard file %s of address book %s could not be \"", "\"parsed\\nUse --debug for more information or \"", "\"--skip-unparsable to proceed\"", ",", "filename", ",", "self", ".", "name", ")", "sys", ".", "exit", "(", "2", ")", "else", ":", "uid", "=", "card", ".", "get_uid", "(", ")", "if", "not", "uid", ":", "logging", ".", "warning", "(", "\"Card %s from address book %s has no UID \"", "\"and will not be availbale.\"", ",", "card", ",", "self", ".", "name", ")", "elif", "uid", "in", "self", ".", "contacts", ":", "logging", ".", "warning", "(", "\"Card %s and %s from address book %s have the same \"", "\"UID. The former will not be availbale.\"", ",", "card", ",", "self", ".", "contacts", "[", "uid", "]", ",", "self", ".", "name", ")", "else", ":", "self", ".", "contacts", "[", "uid", "]", "=", "card", "self", ".", "_loaded", "=", "True", "if", "errors", ":", "logging", ".", "warning", "(", "\"%d of %d vCard files of address book %s could not be parsed.\"", ",", "errors", ",", "len", "(", "self", ".", "contacts", ")", "+", "errors", ",", "self", ")", "logging", ".", "debug", "(", "'Loded %s contacts from address book %s.'", ",", "len", "(", "self", ".", "contacts", ")", ",", "self", ".", "name", ")" ]
0f69430c2680f1ff5f073a977a3c5b753b96cc17
test
AddressBookCollection.get_abook
Get one of the backing abdress books by its name, :param name: the name of the address book to get :type name: str :returns: the matching address book or None :rtype: AddressBook or NoneType
khard/address_book.py
def get_abook(self, name): """Get one of the backing abdress books by its name, :param name: the name of the address book to get :type name: str :returns: the matching address book or None :rtype: AddressBook or NoneType """ for abook in self._abooks: if abook.name == name: return abook
def get_abook(self, name): """Get one of the backing abdress books by its name, :param name: the name of the address book to get :type name: str :returns: the matching address book or None :rtype: AddressBook or NoneType """ for abook in self._abooks: if abook.name == name: return abook
[ "Get", "one", "of", "the", "backing", "abdress", "books", "by", "its", "name" ]
scheibler/khard
python
https://github.com/scheibler/khard/blob/0f69430c2680f1ff5f073a977a3c5b753b96cc17/khard/address_book.py#L378-L389
[ "def", "get_abook", "(", "self", ",", "name", ")", ":", "for", "abook", "in", "self", ".", "_abooks", ":", "if", "abook", ".", "name", "==", "name", ":", "return", "abook" ]
0f69430c2680f1ff5f073a977a3c5b753b96cc17
test
SysHandler.get_table
This function is used in sys command (when user want to find a specific syscall) :param Architecture for syscall table; :param Searching pattern; :param Flag for verbose output :return Return a printable table of matched syscalls
shellen/syscalls/base_handler.py
def get_table(self, arch, pattern, colored=False, verbose=False): ''' This function is used in sys command (when user want to find a specific syscall) :param Architecture for syscall table; :param Searching pattern; :param Flag for verbose output :return Return a printable table of matched syscalls ''' rawtable = self.search(arch, pattern) if len(rawtable) == 0: return None used_hd = self.__fetch_used_headers(rawtable, verbose) table = [self.__make_colored_row(used_hd, 'yellow,bold', upper=True) if colored else used_hd] for command in rawtable: cur_tb_field = [] for hd in used_hd: value = command[hd] cur_tb_field.append(self.__make_colored_field(value, hd, verbose=verbose)) table.append(cur_tb_field) return DoubleTable(table)
def get_table(self, arch, pattern, colored=False, verbose=False): ''' This function is used in sys command (when user want to find a specific syscall) :param Architecture for syscall table; :param Searching pattern; :param Flag for verbose output :return Return a printable table of matched syscalls ''' rawtable = self.search(arch, pattern) if len(rawtable) == 0: return None used_hd = self.__fetch_used_headers(rawtable, verbose) table = [self.__make_colored_row(used_hd, 'yellow,bold', upper=True) if colored else used_hd] for command in rawtable: cur_tb_field = [] for hd in used_hd: value = command[hd] cur_tb_field.append(self.__make_colored_field(value, hd, verbose=verbose)) table.append(cur_tb_field) return DoubleTable(table)
[ "This", "function", "is", "used", "in", "sys", "command", "(", "when", "user", "want", "to", "find", "a", "specific", "syscall", ")" ]
merrychap/shellen
python
https://github.com/merrychap/shellen/blob/3514b7ed3a1b7b1660c3f846a52f58ef02f0954c/shellen/syscalls/base_handler.py#L33-L56
[ "def", "get_table", "(", "self", ",", "arch", ",", "pattern", ",", "colored", "=", "False", ",", "verbose", "=", "False", ")", ":", "rawtable", "=", "self", ".", "search", "(", "arch", ",", "pattern", ")", "if", "len", "(", "rawtable", ")", "==", "0", ":", "return", "None", "used_hd", "=", "self", ".", "__fetch_used_headers", "(", "rawtable", ",", "verbose", ")", "table", "=", "[", "self", ".", "__make_colored_row", "(", "used_hd", ",", "'yellow,bold'", ",", "upper", "=", "True", ")", "if", "colored", "else", "used_hd", "]", "for", "command", "in", "rawtable", ":", "cur_tb_field", "=", "[", "]", "for", "hd", "in", "used_hd", ":", "value", "=", "command", "[", "hd", "]", "cur_tb_field", ".", "append", "(", "self", ".", "__make_colored_field", "(", "value", ",", "hd", ",", "verbose", "=", "verbose", ")", ")", "table", ".", "append", "(", "cur_tb_field", ")", "return", "DoubleTable", "(", "table", ")" ]
3514b7ed3a1b7b1660c3f846a52f58ef02f0954c
test
Assembler.avail_archs
Initialize the dictionary of architectures for assembling via keystone
shellen/asms/asm.py
def avail_archs(self): ''' Initialize the dictionary of architectures for assembling via keystone''' return { ARM32: (KS_ARCH_ARM, KS_MODE_ARM), ARM64: (KS_ARCH_ARM64, KS_MODE_LITTLE_ENDIAN), ARM_TB: (KS_ARCH_ARM, KS_MODE_THUMB), HEXAGON: (KS_ARCH_HEXAGON, KS_MODE_BIG_ENDIAN), MIPS32: (KS_ARCH_MIPS, KS_MODE_MIPS32), MIPS64: (KS_ARCH_MIPS, KS_MODE_MIPS64), PPC32: (KS_ARCH_PPC, KS_MODE_PPC32), PPC64: (KS_ARCH_PPC, KS_MODE_PPC64), SPARC32: (KS_ARCH_SPARC, KS_MODE_SPARC32), SPARC64: (KS_ARCH_SPARC, KS_MODE_SPARC64), SYSTEMZ: (KS_ARCH_SYSTEMZ, KS_MODE_BIG_ENDIAN), X86_16: (KS_ARCH_X86, KS_MODE_16), X86_32: (KS_ARCH_X86, KS_MODE_32), X86_64: (KS_ARCH_X86, KS_MODE_64), }
def avail_archs(self): ''' Initialize the dictionary of architectures for assembling via keystone''' return { ARM32: (KS_ARCH_ARM, KS_MODE_ARM), ARM64: (KS_ARCH_ARM64, KS_MODE_LITTLE_ENDIAN), ARM_TB: (KS_ARCH_ARM, KS_MODE_THUMB), HEXAGON: (KS_ARCH_HEXAGON, KS_MODE_BIG_ENDIAN), MIPS32: (KS_ARCH_MIPS, KS_MODE_MIPS32), MIPS64: (KS_ARCH_MIPS, KS_MODE_MIPS64), PPC32: (KS_ARCH_PPC, KS_MODE_PPC32), PPC64: (KS_ARCH_PPC, KS_MODE_PPC64), SPARC32: (KS_ARCH_SPARC, KS_MODE_SPARC32), SPARC64: (KS_ARCH_SPARC, KS_MODE_SPARC64), SYSTEMZ: (KS_ARCH_SYSTEMZ, KS_MODE_BIG_ENDIAN), X86_16: (KS_ARCH_X86, KS_MODE_16), X86_32: (KS_ARCH_X86, KS_MODE_32), X86_64: (KS_ARCH_X86, KS_MODE_64), }
[ "Initialize", "the", "dictionary", "of", "architectures", "for", "assembling", "via", "keystone" ]
merrychap/shellen
python
https://github.com/merrychap/shellen/blob/3514b7ed3a1b7b1660c3f846a52f58ef02f0954c/shellen/asms/asm.py#L24-L42
[ "def", "avail_archs", "(", "self", ")", ":", "return", "{", "ARM32", ":", "(", "KS_ARCH_ARM", ",", "KS_MODE_ARM", ")", ",", "ARM64", ":", "(", "KS_ARCH_ARM64", ",", "KS_MODE_LITTLE_ENDIAN", ")", ",", "ARM_TB", ":", "(", "KS_ARCH_ARM", ",", "KS_MODE_THUMB", ")", ",", "HEXAGON", ":", "(", "KS_ARCH_HEXAGON", ",", "KS_MODE_BIG_ENDIAN", ")", ",", "MIPS32", ":", "(", "KS_ARCH_MIPS", ",", "KS_MODE_MIPS32", ")", ",", "MIPS64", ":", "(", "KS_ARCH_MIPS", ",", "KS_MODE_MIPS64", ")", ",", "PPC32", ":", "(", "KS_ARCH_PPC", ",", "KS_MODE_PPC32", ")", ",", "PPC64", ":", "(", "KS_ARCH_PPC", ",", "KS_MODE_PPC64", ")", ",", "SPARC32", ":", "(", "KS_ARCH_SPARC", ",", "KS_MODE_SPARC32", ")", ",", "SPARC64", ":", "(", "KS_ARCH_SPARC", ",", "KS_MODE_SPARC64", ")", ",", "SYSTEMZ", ":", "(", "KS_ARCH_SYSTEMZ", ",", "KS_MODE_BIG_ENDIAN", ")", ",", "X86_16", ":", "(", "KS_ARCH_X86", ",", "KS_MODE_16", ")", ",", "X86_32", ":", "(", "KS_ARCH_X86", ",", "KS_MODE_32", ")", ",", "X86_64", ":", "(", "KS_ARCH_X86", ",", "KS_MODE_64", ")", ",", "}" ]
3514b7ed3a1b7b1660c3f846a52f58ef02f0954c
test
Disassembler.avail_archs
Initialize the dictionary of architectures for disassembling via capstone
shellen/asms/disasm.py
def avail_archs(self): ''' Initialize the dictionary of architectures for disassembling via capstone''' return { ARM32: (CS_ARCH_ARM, CS_MODE_ARM), ARM64: (CS_ARCH_ARM64, CS_MODE_LITTLE_ENDIAN), ARM_TB: (CS_ARCH_ARM, CS_MODE_THUMB), MIPS32: (CS_ARCH_MIPS, CS_MODE_MIPS32), MIPS64: (CS_ARCH_MIPS, CS_MODE_MIPS64), SPARC32: (CS_ARCH_SPARC, CS_MODE_BIG_ENDIAN), SPARC64: (CS_ARCH_SPARC, CS_MODE_V9), SYSTEMZ: (CS_ARCH_SYSZ, CS_MODE_BIG_ENDIAN), X86_16: (CS_ARCH_X86, CS_MODE_16), X86_32: (CS_ARCH_X86, CS_MODE_32), X86_64: (CS_ARCH_X86, CS_MODE_64), }
def avail_archs(self): ''' Initialize the dictionary of architectures for disassembling via capstone''' return { ARM32: (CS_ARCH_ARM, CS_MODE_ARM), ARM64: (CS_ARCH_ARM64, CS_MODE_LITTLE_ENDIAN), ARM_TB: (CS_ARCH_ARM, CS_MODE_THUMB), MIPS32: (CS_ARCH_MIPS, CS_MODE_MIPS32), MIPS64: (CS_ARCH_MIPS, CS_MODE_MIPS64), SPARC32: (CS_ARCH_SPARC, CS_MODE_BIG_ENDIAN), SPARC64: (CS_ARCH_SPARC, CS_MODE_V9), SYSTEMZ: (CS_ARCH_SYSZ, CS_MODE_BIG_ENDIAN), X86_16: (CS_ARCH_X86, CS_MODE_16), X86_32: (CS_ARCH_X86, CS_MODE_32), X86_64: (CS_ARCH_X86, CS_MODE_64), }
[ "Initialize", "the", "dictionary", "of", "architectures", "for", "disassembling", "via", "capstone" ]
merrychap/shellen
python
https://github.com/merrychap/shellen/blob/3514b7ed3a1b7b1660c3f846a52f58ef02f0954c/shellen/asms/disasm.py#L25-L40
[ "def", "avail_archs", "(", "self", ")", ":", "return", "{", "ARM32", ":", "(", "CS_ARCH_ARM", ",", "CS_MODE_ARM", ")", ",", "ARM64", ":", "(", "CS_ARCH_ARM64", ",", "CS_MODE_LITTLE_ENDIAN", ")", ",", "ARM_TB", ":", "(", "CS_ARCH_ARM", ",", "CS_MODE_THUMB", ")", ",", "MIPS32", ":", "(", "CS_ARCH_MIPS", ",", "CS_MODE_MIPS32", ")", ",", "MIPS64", ":", "(", "CS_ARCH_MIPS", ",", "CS_MODE_MIPS64", ")", ",", "SPARC32", ":", "(", "CS_ARCH_SPARC", ",", "CS_MODE_BIG_ENDIAN", ")", ",", "SPARC64", ":", "(", "CS_ARCH_SPARC", ",", "CS_MODE_V9", ")", ",", "SYSTEMZ", ":", "(", "CS_ARCH_SYSZ", ",", "CS_MODE_BIG_ENDIAN", ")", ",", "X86_16", ":", "(", "CS_ARCH_X86", ",", "CS_MODE_16", ")", ",", "X86_32", ":", "(", "CS_ARCH_X86", ",", "CS_MODE_32", ")", ",", "X86_64", ":", "(", "CS_ARCH_X86", ",", "CS_MODE_64", ")", ",", "}" ]
3514b7ed3a1b7b1660c3f846a52f58ef02f0954c
test
getargspec_permissive
An `inspect.getargspec` with a relaxed sanity check to support Cython. Motivation: A Cython-compiled function is *not* an instance of Python's types.FunctionType. That is the sanity check the standard Py2 library uses in `inspect.getargspec()`. So, an exception is raised when calling `argh.dispatch_command(cythonCompiledFunc)`. However, the CyFunctions do have perfectly usable `.func_code` and `.func_defaults` which is all `inspect.getargspec` needs. This function just copies `inspect.getargspec()` from the standard library but relaxes the test to a more duck-typing one of having both `.func_code` and `.func_defaults` attributes.
argh/compat.py
def getargspec_permissive(func): """ An `inspect.getargspec` with a relaxed sanity check to support Cython. Motivation: A Cython-compiled function is *not* an instance of Python's types.FunctionType. That is the sanity check the standard Py2 library uses in `inspect.getargspec()`. So, an exception is raised when calling `argh.dispatch_command(cythonCompiledFunc)`. However, the CyFunctions do have perfectly usable `.func_code` and `.func_defaults` which is all `inspect.getargspec` needs. This function just copies `inspect.getargspec()` from the standard library but relaxes the test to a more duck-typing one of having both `.func_code` and `.func_defaults` attributes. """ if inspect.ismethod(func): func = func.im_func # Py2 Stdlib uses isfunction(func) which is too strict for Cython-compiled # functions though such have perfectly usable func_code, func_defaults. if not (hasattr(func, "func_code") and hasattr(func, "func_defaults")): raise TypeError('{!r} missing func_code or func_defaults'.format(func)) args, varargs, varkw = inspect.getargs(func.func_code) return inspect.ArgSpec(args, varargs, varkw, func.func_defaults)
def getargspec_permissive(func): """ An `inspect.getargspec` with a relaxed sanity check to support Cython. Motivation: A Cython-compiled function is *not* an instance of Python's types.FunctionType. That is the sanity check the standard Py2 library uses in `inspect.getargspec()`. So, an exception is raised when calling `argh.dispatch_command(cythonCompiledFunc)`. However, the CyFunctions do have perfectly usable `.func_code` and `.func_defaults` which is all `inspect.getargspec` needs. This function just copies `inspect.getargspec()` from the standard library but relaxes the test to a more duck-typing one of having both `.func_code` and `.func_defaults` attributes. """ if inspect.ismethod(func): func = func.im_func # Py2 Stdlib uses isfunction(func) which is too strict for Cython-compiled # functions though such have perfectly usable func_code, func_defaults. if not (hasattr(func, "func_code") and hasattr(func, "func_defaults")): raise TypeError('{!r} missing func_code or func_defaults'.format(func)) args, varargs, varkw = inspect.getargs(func.func_code) return inspect.ArgSpec(args, varargs, varkw, func.func_defaults)
[ "An", "inspect", ".", "getargspec", "with", "a", "relaxed", "sanity", "check", "to", "support", "Cython", "." ]
neithere/argh
python
https://github.com/neithere/argh/blob/dcd3253f2994400a6a58a700c118c53765bc50a4/argh/compat.py#L22-L48
[ "def", "getargspec_permissive", "(", "func", ")", ":", "if", "inspect", ".", "ismethod", "(", "func", ")", ":", "func", "=", "func", ".", "im_func", "# Py2 Stdlib uses isfunction(func) which is too strict for Cython-compiled", "# functions though such have perfectly usable func_code, func_defaults.", "if", "not", "(", "hasattr", "(", "func", ",", "\"func_code\"", ")", "and", "hasattr", "(", "func", ",", "\"func_defaults\"", ")", ")", ":", "raise", "TypeError", "(", "'{!r} missing func_code or func_defaults'", ".", "format", "(", "func", ")", ")", "args", ",", "varargs", ",", "varkw", "=", "inspect", ".", "getargs", "(", "func", ".", "func_code", ")", "return", "inspect", ".", "ArgSpec", "(", "args", ",", "varargs", ",", "varkw", ",", "func", ".", "func_defaults", ")" ]
dcd3253f2994400a6a58a700c118c53765bc50a4
test
dispatch
Parses given list of arguments using given parser, calls the relevant function and prints the result. The target function should expect one positional argument: the :class:`argparse.Namespace` object. However, if the function is decorated with :func:`~argh.decorators.plain_signature`, the positional and named arguments from the namespace object are passed to the function instead of the object itself. :param parser: the ArgumentParser instance. :param argv: a list of strings representing the arguments. If `None`, ``sys.argv`` is used instead. Default is `None`. :param add_help_command: if `True`, converts first positional argument "help" to a keyword argument so that ``help foo`` becomes ``foo --help`` and displays usage information for "foo". Default is `True`. :param output_file: A file-like object for output. If `None`, the resulting lines are collected and returned as a string. Default is ``sys.stdout``. :param errors_file: Same as `output_file` but for ``sys.stderr``. :param raw_output: If `True`, results are written to the output file raw, without adding whitespaces or newlines between yielded strings. Default is `False`. :param completion: If `True`, shell tab completion is enabled. Default is `True`. (You will also need to install it.) See :mod:`argh.completion`. :param skip_unknown_args: If `True`, unknown arguments do not cause an error (`ArgumentParser.parse_known_args` is used). :param namespace: An `argparse.Namespace`-like object. By default an :class:`ArghNamespace` object is used. Please note that support for combined default and nested functions may be broken if a different type of object is forced. By default the exceptions are not wrapped and will propagate. The only exception that is always wrapped is :class:`~argh.exceptions.CommandError` which is interpreted as an expected event so the traceback is hidden. You can also mark arbitrary exceptions as "wrappable" by using the :func:`~argh.decorators.wrap_errors` decorator.
argh/dispatching.py
def dispatch(parser, argv=None, add_help_command=True, completion=True, pre_call=None, output_file=sys.stdout, errors_file=sys.stderr, raw_output=False, namespace=None, skip_unknown_args=False): """ Parses given list of arguments using given parser, calls the relevant function and prints the result. The target function should expect one positional argument: the :class:`argparse.Namespace` object. However, if the function is decorated with :func:`~argh.decorators.plain_signature`, the positional and named arguments from the namespace object are passed to the function instead of the object itself. :param parser: the ArgumentParser instance. :param argv: a list of strings representing the arguments. If `None`, ``sys.argv`` is used instead. Default is `None`. :param add_help_command: if `True`, converts first positional argument "help" to a keyword argument so that ``help foo`` becomes ``foo --help`` and displays usage information for "foo". Default is `True`. :param output_file: A file-like object for output. If `None`, the resulting lines are collected and returned as a string. Default is ``sys.stdout``. :param errors_file: Same as `output_file` but for ``sys.stderr``. :param raw_output: If `True`, results are written to the output file raw, without adding whitespaces or newlines between yielded strings. Default is `False`. :param completion: If `True`, shell tab completion is enabled. Default is `True`. (You will also need to install it.) See :mod:`argh.completion`. :param skip_unknown_args: If `True`, unknown arguments do not cause an error (`ArgumentParser.parse_known_args` is used). :param namespace: An `argparse.Namespace`-like object. By default an :class:`ArghNamespace` object is used. Please note that support for combined default and nested functions may be broken if a different type of object is forced. By default the exceptions are not wrapped and will propagate. The only exception that is always wrapped is :class:`~argh.exceptions.CommandError` which is interpreted as an expected event so the traceback is hidden. You can also mark arbitrary exceptions as "wrappable" by using the :func:`~argh.decorators.wrap_errors` decorator. """ if completion: autocomplete(parser) if argv is None: argv = sys.argv[1:] if add_help_command: if argv and argv[0] == 'help': argv.pop(0) argv.append('--help') if skip_unknown_args: parse_args = parser.parse_known_args else: parse_args = parser.parse_args if not namespace: namespace = ArghNamespace() # this will raise SystemExit if parsing fails namespace_obj = parse_args(argv, namespace=namespace) function = _get_function_from_namespace_obj(namespace_obj) if function: lines = _execute_command(function, namespace_obj, errors_file, pre_call=pre_call) else: # no commands declared, can't dispatch; display help message lines = [parser.format_usage()] if output_file is None: # user wants a string; we create an internal temporary file-like object # and will return its contents as a string if sys.version_info < (3,0): f = compat.BytesIO() else: f = compat.StringIO() else: # normally this is stdout; can be any file f = output_file for line in lines: # print the line as soon as it is generated to ensure that it is # displayed to the user before anything else happens, e.g. # raw_input() is called io.dump(line, f) if not raw_output: # in most cases user wants one message per line io.dump('\n', f) if output_file is None: # user wanted a string; return contents of our temporary file-like obj f.seek(0) return f.read()
def dispatch(parser, argv=None, add_help_command=True, completion=True, pre_call=None, output_file=sys.stdout, errors_file=sys.stderr, raw_output=False, namespace=None, skip_unknown_args=False): """ Parses given list of arguments using given parser, calls the relevant function and prints the result. The target function should expect one positional argument: the :class:`argparse.Namespace` object. However, if the function is decorated with :func:`~argh.decorators.plain_signature`, the positional and named arguments from the namespace object are passed to the function instead of the object itself. :param parser: the ArgumentParser instance. :param argv: a list of strings representing the arguments. If `None`, ``sys.argv`` is used instead. Default is `None`. :param add_help_command: if `True`, converts first positional argument "help" to a keyword argument so that ``help foo`` becomes ``foo --help`` and displays usage information for "foo". Default is `True`. :param output_file: A file-like object for output. If `None`, the resulting lines are collected and returned as a string. Default is ``sys.stdout``. :param errors_file: Same as `output_file` but for ``sys.stderr``. :param raw_output: If `True`, results are written to the output file raw, without adding whitespaces or newlines between yielded strings. Default is `False`. :param completion: If `True`, shell tab completion is enabled. Default is `True`. (You will also need to install it.) See :mod:`argh.completion`. :param skip_unknown_args: If `True`, unknown arguments do not cause an error (`ArgumentParser.parse_known_args` is used). :param namespace: An `argparse.Namespace`-like object. By default an :class:`ArghNamespace` object is used. Please note that support for combined default and nested functions may be broken if a different type of object is forced. By default the exceptions are not wrapped and will propagate. The only exception that is always wrapped is :class:`~argh.exceptions.CommandError` which is interpreted as an expected event so the traceback is hidden. You can also mark arbitrary exceptions as "wrappable" by using the :func:`~argh.decorators.wrap_errors` decorator. """ if completion: autocomplete(parser) if argv is None: argv = sys.argv[1:] if add_help_command: if argv and argv[0] == 'help': argv.pop(0) argv.append('--help') if skip_unknown_args: parse_args = parser.parse_known_args else: parse_args = parser.parse_args if not namespace: namespace = ArghNamespace() # this will raise SystemExit if parsing fails namespace_obj = parse_args(argv, namespace=namespace) function = _get_function_from_namespace_obj(namespace_obj) if function: lines = _execute_command(function, namespace_obj, errors_file, pre_call=pre_call) else: # no commands declared, can't dispatch; display help message lines = [parser.format_usage()] if output_file is None: # user wants a string; we create an internal temporary file-like object # and will return its contents as a string if sys.version_info < (3,0): f = compat.BytesIO() else: f = compat.StringIO() else: # normally this is stdout; can be any file f = output_file for line in lines: # print the line as soon as it is generated to ensure that it is # displayed to the user before anything else happens, e.g. # raw_input() is called io.dump(line, f) if not raw_output: # in most cases user wants one message per line io.dump('\n', f) if output_file is None: # user wanted a string; return contents of our temporary file-like obj f.seek(0) return f.read()
[ "Parses", "given", "list", "of", "arguments", "using", "given", "parser", "calls", "the", "relevant", "function", "and", "prints", "the", "result", "." ]
neithere/argh
python
https://github.com/neithere/argh/blob/dcd3253f2994400a6a58a700c118c53765bc50a4/argh/dispatching.py#L65-L187
[ "def", "dispatch", "(", "parser", ",", "argv", "=", "None", ",", "add_help_command", "=", "True", ",", "completion", "=", "True", ",", "pre_call", "=", "None", ",", "output_file", "=", "sys", ".", "stdout", ",", "errors_file", "=", "sys", ".", "stderr", ",", "raw_output", "=", "False", ",", "namespace", "=", "None", ",", "skip_unknown_args", "=", "False", ")", ":", "if", "completion", ":", "autocomplete", "(", "parser", ")", "if", "argv", "is", "None", ":", "argv", "=", "sys", ".", "argv", "[", "1", ":", "]", "if", "add_help_command", ":", "if", "argv", "and", "argv", "[", "0", "]", "==", "'help'", ":", "argv", ".", "pop", "(", "0", ")", "argv", ".", "append", "(", "'--help'", ")", "if", "skip_unknown_args", ":", "parse_args", "=", "parser", ".", "parse_known_args", "else", ":", "parse_args", "=", "parser", ".", "parse_args", "if", "not", "namespace", ":", "namespace", "=", "ArghNamespace", "(", ")", "# this will raise SystemExit if parsing fails", "namespace_obj", "=", "parse_args", "(", "argv", ",", "namespace", "=", "namespace", ")", "function", "=", "_get_function_from_namespace_obj", "(", "namespace_obj", ")", "if", "function", ":", "lines", "=", "_execute_command", "(", "function", ",", "namespace_obj", ",", "errors_file", ",", "pre_call", "=", "pre_call", ")", "else", ":", "# no commands declared, can't dispatch; display help message", "lines", "=", "[", "parser", ".", "format_usage", "(", ")", "]", "if", "output_file", "is", "None", ":", "# user wants a string; we create an internal temporary file-like object", "# and will return its contents as a string", "if", "sys", ".", "version_info", "<", "(", "3", ",", "0", ")", ":", "f", "=", "compat", ".", "BytesIO", "(", ")", "else", ":", "f", "=", "compat", ".", "StringIO", "(", ")", "else", ":", "# normally this is stdout; can be any file", "f", "=", "output_file", "for", "line", "in", "lines", ":", "# print the line as soon as it is generated to ensure that it is", "# displayed to the user before anything else happens, e.g.", "# raw_input() is called", "io", ".", "dump", "(", "line", ",", "f", ")", "if", "not", "raw_output", ":", "# in most cases user wants one message per line", "io", ".", "dump", "(", "'\\n'", ",", "f", ")", "if", "output_file", "is", "None", ":", "# user wanted a string; return contents of our temporary file-like obj", "f", ".", "seek", "(", "0", ")", "return", "f", ".", "read", "(", ")" ]
dcd3253f2994400a6a58a700c118c53765bc50a4
test
_execute_command
Assumes that `function` is a callable. Tries different approaches to call it (with `namespace_obj` or with ordinary signature). Yields the results line by line. If :class:`~argh.exceptions.CommandError` is raised, its message is appended to the results (i.e. yielded by the generator as a string). All other exceptions propagate unless marked as wrappable by :func:`wrap_errors`.
argh/dispatching.py
def _execute_command(function, namespace_obj, errors_file, pre_call=None): """ Assumes that `function` is a callable. Tries different approaches to call it (with `namespace_obj` or with ordinary signature). Yields the results line by line. If :class:`~argh.exceptions.CommandError` is raised, its message is appended to the results (i.e. yielded by the generator as a string). All other exceptions propagate unless marked as wrappable by :func:`wrap_errors`. """ if pre_call: # XXX undocumented because I'm unsure if it's OK # Actually used in real projects: # * https://google.com/search?q=argh+dispatch+pre_call # * https://github.com/neithere/argh/issues/63 pre_call(namespace_obj) # the function is nested to catch certain exceptions (see below) def _call(): # Actually call the function if getattr(function, ATTR_EXPECTS_NAMESPACE_OBJECT, False): result = function(namespace_obj) else: # namespace -> dictionary _flat_key = lambda key: key.replace('-', '_') all_input = dict((_flat_key(k), v) for k,v in vars(namespace_obj).items()) # filter the namespace variables so that only those expected # by the actual function will pass spec = get_arg_spec(function) positional = [all_input[k] for k in spec.args] kwonly = getattr(spec, 'kwonlyargs', []) keywords = dict((k, all_input[k]) for k in kwonly) # *args if spec.varargs: positional += getattr(namespace_obj, spec.varargs) # **kwargs varkw = getattr(spec, 'varkw', getattr(spec, 'keywords', [])) if varkw: not_kwargs = [DEST_FUNCTION] + spec.args + [spec.varargs] + kwonly for k in vars(namespace_obj): if k.startswith('_') or k in not_kwargs: continue keywords[k] = getattr(namespace_obj, k) result = function(*positional, **keywords) # Yield the results if isinstance(result, (GeneratorType, list, tuple)): # yield each line ASAP, convert CommandError message to a line for line in result: yield line else: # yield non-empty non-iterable result as a single line if result is not None: yield result wrappable_exceptions = [CommandError] wrappable_exceptions += getattr(function, ATTR_WRAPPED_EXCEPTIONS, []) try: result = _call() for line in result: yield line except tuple(wrappable_exceptions) as e: processor = getattr(function, ATTR_WRAPPED_EXCEPTIONS_PROCESSOR, lambda e: '{0.__class__.__name__}: {0}'.format(e)) errors_file.write(compat.text_type(processor(e))) errors_file.write('\n')
def _execute_command(function, namespace_obj, errors_file, pre_call=None): """ Assumes that `function` is a callable. Tries different approaches to call it (with `namespace_obj` or with ordinary signature). Yields the results line by line. If :class:`~argh.exceptions.CommandError` is raised, its message is appended to the results (i.e. yielded by the generator as a string). All other exceptions propagate unless marked as wrappable by :func:`wrap_errors`. """ if pre_call: # XXX undocumented because I'm unsure if it's OK # Actually used in real projects: # * https://google.com/search?q=argh+dispatch+pre_call # * https://github.com/neithere/argh/issues/63 pre_call(namespace_obj) # the function is nested to catch certain exceptions (see below) def _call(): # Actually call the function if getattr(function, ATTR_EXPECTS_NAMESPACE_OBJECT, False): result = function(namespace_obj) else: # namespace -> dictionary _flat_key = lambda key: key.replace('-', '_') all_input = dict((_flat_key(k), v) for k,v in vars(namespace_obj).items()) # filter the namespace variables so that only those expected # by the actual function will pass spec = get_arg_spec(function) positional = [all_input[k] for k in spec.args] kwonly = getattr(spec, 'kwonlyargs', []) keywords = dict((k, all_input[k]) for k in kwonly) # *args if spec.varargs: positional += getattr(namespace_obj, spec.varargs) # **kwargs varkw = getattr(spec, 'varkw', getattr(spec, 'keywords', [])) if varkw: not_kwargs = [DEST_FUNCTION] + spec.args + [spec.varargs] + kwonly for k in vars(namespace_obj): if k.startswith('_') or k in not_kwargs: continue keywords[k] = getattr(namespace_obj, k) result = function(*positional, **keywords) # Yield the results if isinstance(result, (GeneratorType, list, tuple)): # yield each line ASAP, convert CommandError message to a line for line in result: yield line else: # yield non-empty non-iterable result as a single line if result is not None: yield result wrappable_exceptions = [CommandError] wrappable_exceptions += getattr(function, ATTR_WRAPPED_EXCEPTIONS, []) try: result = _call() for line in result: yield line except tuple(wrappable_exceptions) as e: processor = getattr(function, ATTR_WRAPPED_EXCEPTIONS_PROCESSOR, lambda e: '{0.__class__.__name__}: {0}'.format(e)) errors_file.write(compat.text_type(processor(e))) errors_file.write('\n')
[ "Assumes", "that", "function", "is", "a", "callable", ".", "Tries", "different", "approaches", "to", "call", "it", "(", "with", "namespace_obj", "or", "with", "ordinary", "signature", ")", ".", "Yields", "the", "results", "line", "by", "line", "." ]
neithere/argh
python
https://github.com/neithere/argh/blob/dcd3253f2994400a6a58a700c118c53765bc50a4/argh/dispatching.py#L210-L284
[ "def", "_execute_command", "(", "function", ",", "namespace_obj", ",", "errors_file", ",", "pre_call", "=", "None", ")", ":", "if", "pre_call", ":", "# XXX undocumented because I'm unsure if it's OK", "# Actually used in real projects:", "# * https://google.com/search?q=argh+dispatch+pre_call", "# * https://github.com/neithere/argh/issues/63", "pre_call", "(", "namespace_obj", ")", "# the function is nested to catch certain exceptions (see below)", "def", "_call", "(", ")", ":", "# Actually call the function", "if", "getattr", "(", "function", ",", "ATTR_EXPECTS_NAMESPACE_OBJECT", ",", "False", ")", ":", "result", "=", "function", "(", "namespace_obj", ")", "else", ":", "# namespace -> dictionary", "_flat_key", "=", "lambda", "key", ":", "key", ".", "replace", "(", "'-'", ",", "'_'", ")", "all_input", "=", "dict", "(", "(", "_flat_key", "(", "k", ")", ",", "v", ")", "for", "k", ",", "v", "in", "vars", "(", "namespace_obj", ")", ".", "items", "(", ")", ")", "# filter the namespace variables so that only those expected", "# by the actual function will pass", "spec", "=", "get_arg_spec", "(", "function", ")", "positional", "=", "[", "all_input", "[", "k", "]", "for", "k", "in", "spec", ".", "args", "]", "kwonly", "=", "getattr", "(", "spec", ",", "'kwonlyargs'", ",", "[", "]", ")", "keywords", "=", "dict", "(", "(", "k", ",", "all_input", "[", "k", "]", ")", "for", "k", "in", "kwonly", ")", "# *args", "if", "spec", ".", "varargs", ":", "positional", "+=", "getattr", "(", "namespace_obj", ",", "spec", ".", "varargs", ")", "# **kwargs", "varkw", "=", "getattr", "(", "spec", ",", "'varkw'", ",", "getattr", "(", "spec", ",", "'keywords'", ",", "[", "]", ")", ")", "if", "varkw", ":", "not_kwargs", "=", "[", "DEST_FUNCTION", "]", "+", "spec", ".", "args", "+", "[", "spec", ".", "varargs", "]", "+", "kwonly", "for", "k", "in", "vars", "(", "namespace_obj", ")", ":", "if", "k", ".", "startswith", "(", "'_'", ")", "or", "k", "in", "not_kwargs", ":", "continue", "keywords", "[", "k", "]", "=", "getattr", "(", "namespace_obj", ",", "k", ")", "result", "=", "function", "(", "*", "positional", ",", "*", "*", "keywords", ")", "# Yield the results", "if", "isinstance", "(", "result", ",", "(", "GeneratorType", ",", "list", ",", "tuple", ")", ")", ":", "# yield each line ASAP, convert CommandError message to a line", "for", "line", "in", "result", ":", "yield", "line", "else", ":", "# yield non-empty non-iterable result as a single line", "if", "result", "is", "not", "None", ":", "yield", "result", "wrappable_exceptions", "=", "[", "CommandError", "]", "wrappable_exceptions", "+=", "getattr", "(", "function", ",", "ATTR_WRAPPED_EXCEPTIONS", ",", "[", "]", ")", "try", ":", "result", "=", "_call", "(", ")", "for", "line", "in", "result", ":", "yield", "line", "except", "tuple", "(", "wrappable_exceptions", ")", "as", "e", ":", "processor", "=", "getattr", "(", "function", ",", "ATTR_WRAPPED_EXCEPTIONS_PROCESSOR", ",", "lambda", "e", ":", "'{0.__class__.__name__}: {0}'", ".", "format", "(", "e", ")", ")", "errors_file", ".", "write", "(", "compat", ".", "text_type", "(", "processor", "(", "e", ")", ")", ")", "errors_file", ".", "write", "(", "'\\n'", ")" ]
dcd3253f2994400a6a58a700c118c53765bc50a4
test
dispatch_command
A wrapper for :func:`dispatch` that creates a one-command parser. Uses :attr:`PARSER_FORMATTER`. This:: dispatch_command(foo) ...is a shortcut for:: parser = ArgumentParser() set_default_command(parser, foo) dispatch(parser) This function can be also used as a decorator.
argh/dispatching.py
def dispatch_command(function, *args, **kwargs): """ A wrapper for :func:`dispatch` that creates a one-command parser. Uses :attr:`PARSER_FORMATTER`. This:: dispatch_command(foo) ...is a shortcut for:: parser = ArgumentParser() set_default_command(parser, foo) dispatch(parser) This function can be also used as a decorator. """ parser = argparse.ArgumentParser(formatter_class=PARSER_FORMATTER) set_default_command(parser, function) dispatch(parser, *args, **kwargs)
def dispatch_command(function, *args, **kwargs): """ A wrapper for :func:`dispatch` that creates a one-command parser. Uses :attr:`PARSER_FORMATTER`. This:: dispatch_command(foo) ...is a shortcut for:: parser = ArgumentParser() set_default_command(parser, foo) dispatch(parser) This function can be also used as a decorator. """ parser = argparse.ArgumentParser(formatter_class=PARSER_FORMATTER) set_default_command(parser, function) dispatch(parser, *args, **kwargs)
[ "A", "wrapper", "for", ":", "func", ":", "dispatch", "that", "creates", "a", "one", "-", "command", "parser", ".", "Uses", ":", "attr", ":", "PARSER_FORMATTER", "." ]
neithere/argh
python
https://github.com/neithere/argh/blob/dcd3253f2994400a6a58a700c118c53765bc50a4/argh/dispatching.py#L287-L306
[ "def", "dispatch_command", "(", "function", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "parser", "=", "argparse", ".", "ArgumentParser", "(", "formatter_class", "=", "PARSER_FORMATTER", ")", "set_default_command", "(", "parser", ",", "function", ")", "dispatch", "(", "parser", ",", "*", "args", ",", "*", "*", "kwargs", ")" ]
dcd3253f2994400a6a58a700c118c53765bc50a4
test
dispatch_commands
A wrapper for :func:`dispatch` that creates a parser, adds commands to the parser and dispatches them. Uses :attr:`PARSER_FORMATTER`. This:: dispatch_commands([foo, bar]) ...is a shortcut for:: parser = ArgumentParser() add_commands(parser, [foo, bar]) dispatch(parser)
argh/dispatching.py
def dispatch_commands(functions, *args, **kwargs): """ A wrapper for :func:`dispatch` that creates a parser, adds commands to the parser and dispatches them. Uses :attr:`PARSER_FORMATTER`. This:: dispatch_commands([foo, bar]) ...is a shortcut for:: parser = ArgumentParser() add_commands(parser, [foo, bar]) dispatch(parser) """ parser = argparse.ArgumentParser(formatter_class=PARSER_FORMATTER) add_commands(parser, functions) dispatch(parser, *args, **kwargs)
def dispatch_commands(functions, *args, **kwargs): """ A wrapper for :func:`dispatch` that creates a parser, adds commands to the parser and dispatches them. Uses :attr:`PARSER_FORMATTER`. This:: dispatch_commands([foo, bar]) ...is a shortcut for:: parser = ArgumentParser() add_commands(parser, [foo, bar]) dispatch(parser) """ parser = argparse.ArgumentParser(formatter_class=PARSER_FORMATTER) add_commands(parser, functions) dispatch(parser, *args, **kwargs)
[ "A", "wrapper", "for", ":", "func", ":", "dispatch", "that", "creates", "a", "parser", "adds", "commands", "to", "the", "parser", "and", "dispatches", "them", ".", "Uses", ":", "attr", ":", "PARSER_FORMATTER", "." ]
neithere/argh
python
https://github.com/neithere/argh/blob/dcd3253f2994400a6a58a700c118c53765bc50a4/argh/dispatching.py#L309-L328
[ "def", "dispatch_commands", "(", "functions", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "parser", "=", "argparse", ".", "ArgumentParser", "(", "formatter_class", "=", "PARSER_FORMATTER", ")", "add_commands", "(", "parser", ",", "functions", ")", "dispatch", "(", "parser", ",", "*", "args", ",", "*", "*", "kwargs", ")" ]
dcd3253f2994400a6a58a700c118c53765bc50a4
test
safe_input
Prompts user for input. Correctly handles prompt message encoding.
argh/io.py
def safe_input(prompt): """ Prompts user for input. Correctly handles prompt message encoding. """ if sys.version_info < (3,0): if isinstance(prompt, compat.text_type): # Python 2.x: unicode → bytes encoding = locale.getpreferredencoding() or 'utf-8' prompt = prompt.encode(encoding) else: if not isinstance(prompt, compat.text_type): # Python 3.x: bytes → unicode prompt = prompt.decode() return _input(prompt)
def safe_input(prompt): """ Prompts user for input. Correctly handles prompt message encoding. """ if sys.version_info < (3,0): if isinstance(prompt, compat.text_type): # Python 2.x: unicode → bytes encoding = locale.getpreferredencoding() or 'utf-8' prompt = prompt.encode(encoding) else: if not isinstance(prompt, compat.text_type): # Python 3.x: bytes → unicode prompt = prompt.decode() return _input(prompt)
[ "Prompts", "user", "for", "input", ".", "Correctly", "handles", "prompt", "message", "encoding", "." ]
neithere/argh
python
https://github.com/neithere/argh/blob/dcd3253f2994400a6a58a700c118c53765bc50a4/argh/io.py#L32-L47
[ "def", "safe_input", "(", "prompt", ")", ":", "if", "sys", ".", "version_info", "<", "(", "3", ",", "0", ")", ":", "if", "isinstance", "(", "prompt", ",", "compat", ".", "text_type", ")", ":", "# Python 2.x: unicode → bytes", "encoding", "=", "locale", ".", "getpreferredencoding", "(", ")", "or", "'utf-8'", "prompt", "=", "prompt", ".", "encode", "(", "encoding", ")", "else", ":", "if", "not", "isinstance", "(", "prompt", ",", "compat", ".", "text_type", ")", ":", "# Python 3.x: bytes → unicode", "prompt", "=", "prompt", ".", "decode", "(", ")", "return", "_input", "(", "prompt", ")" ]
dcd3253f2994400a6a58a700c118c53765bc50a4
test
encode_output
Encodes given value so it can be written to given file object. Value may be Unicode, binary string or any other data type. The exact behaviour depends on the Python version: Python 3.x `sys.stdout` is a `_io.TextIOWrapper` instance that accepts `str` (unicode) and breaks on `bytes`. It is OK to simply assume that everything is Unicode unless special handling is introduced in the client code. Thus, no additional processing is performed. Python 2.x `sys.stdout` is a file-like object that accepts `str` (bytes) and breaks when `unicode` is passed to `sys.stdout.write()`. We can expect both Unicode and bytes. They need to be encoded so as to match the file object encoding. The output is binary if the object doesn't explicitly require Unicode.
argh/io.py
def encode_output(value, output_file): """ Encodes given value so it can be written to given file object. Value may be Unicode, binary string or any other data type. The exact behaviour depends on the Python version: Python 3.x `sys.stdout` is a `_io.TextIOWrapper` instance that accepts `str` (unicode) and breaks on `bytes`. It is OK to simply assume that everything is Unicode unless special handling is introduced in the client code. Thus, no additional processing is performed. Python 2.x `sys.stdout` is a file-like object that accepts `str` (bytes) and breaks when `unicode` is passed to `sys.stdout.write()`. We can expect both Unicode and bytes. They need to be encoded so as to match the file object encoding. The output is binary if the object doesn't explicitly require Unicode. """ if sys.version_info > (3,0): # Python 3: whatever → unicode return compat.text_type(value) else: # Python 2: handle special cases stream_encoding = getattr(output_file, 'encoding', None) if stream_encoding: if stream_encoding.upper() == 'UTF-8': return compat.text_type(value) else: return value.encode(stream_encoding, 'ignore') else: # no explicit encoding requirements; force binary if isinstance(value, compat.text_type): # unicode → binary return value.encode('utf-8') else: return str(value)
def encode_output(value, output_file): """ Encodes given value so it can be written to given file object. Value may be Unicode, binary string or any other data type. The exact behaviour depends on the Python version: Python 3.x `sys.stdout` is a `_io.TextIOWrapper` instance that accepts `str` (unicode) and breaks on `bytes`. It is OK to simply assume that everything is Unicode unless special handling is introduced in the client code. Thus, no additional processing is performed. Python 2.x `sys.stdout` is a file-like object that accepts `str` (bytes) and breaks when `unicode` is passed to `sys.stdout.write()`. We can expect both Unicode and bytes. They need to be encoded so as to match the file object encoding. The output is binary if the object doesn't explicitly require Unicode. """ if sys.version_info > (3,0): # Python 3: whatever → unicode return compat.text_type(value) else: # Python 2: handle special cases stream_encoding = getattr(output_file, 'encoding', None) if stream_encoding: if stream_encoding.upper() == 'UTF-8': return compat.text_type(value) else: return value.encode(stream_encoding, 'ignore') else: # no explicit encoding requirements; force binary if isinstance(value, compat.text_type): # unicode → binary return value.encode('utf-8') else: return str(value)
[ "Encodes", "given", "value", "so", "it", "can", "be", "written", "to", "given", "file", "object", "." ]
neithere/argh
python
https://github.com/neithere/argh/blob/dcd3253f2994400a6a58a700c118c53765bc50a4/argh/io.py#L50-L96
[ "def", "encode_output", "(", "value", ",", "output_file", ")", ":", "if", "sys", ".", "version_info", ">", "(", "3", ",", "0", ")", ":", "# Python 3: whatever → unicode", "return", "compat", ".", "text_type", "(", "value", ")", "else", ":", "# Python 2: handle special cases", "stream_encoding", "=", "getattr", "(", "output_file", ",", "'encoding'", ",", "None", ")", "if", "stream_encoding", ":", "if", "stream_encoding", ".", "upper", "(", ")", "==", "'UTF-8'", ":", "return", "compat", ".", "text_type", "(", "value", ")", "else", ":", "return", "value", ".", "encode", "(", "stream_encoding", ",", "'ignore'", ")", "else", ":", "# no explicit encoding requirements; force binary", "if", "isinstance", "(", "value", ",", "compat", ".", "text_type", ")", ":", "# unicode → binary", "return", "value", ".", "encode", "(", "'utf-8'", ")", "else", ":", "return", "str", "(", "value", ")" ]
dcd3253f2994400a6a58a700c118c53765bc50a4
test
dump
Writes given line to given output file. See :func:`encode_output` for details.
argh/io.py
def dump(raw_data, output_file): """ Writes given line to given output file. See :func:`encode_output` for details. """ data = encode_output(raw_data, output_file) output_file.write(data)
def dump(raw_data, output_file): """ Writes given line to given output file. See :func:`encode_output` for details. """ data = encode_output(raw_data, output_file) output_file.write(data)
[ "Writes", "given", "line", "to", "given", "output", "file", ".", "See", ":", "func", ":", "encode_output", "for", "details", "." ]
neithere/argh
python
https://github.com/neithere/argh/blob/dcd3253f2994400a6a58a700c118c53765bc50a4/argh/io.py#L99-L105
[ "def", "dump", "(", "raw_data", ",", "output_file", ")", ":", "data", "=", "encode_output", "(", "raw_data", ",", "output_file", ")", "output_file", ".", "write", "(", "data", ")" ]
dcd3253f2994400a6a58a700c118c53765bc50a4
test
autocomplete
Adds support for shell completion via argcomplete_ by patching given `argparse.ArgumentParser` (sub)class. If completion is not enabled, logs a debug-level message.
argh/completion.py
def autocomplete(parser): """ Adds support for shell completion via argcomplete_ by patching given `argparse.ArgumentParser` (sub)class. If completion is not enabled, logs a debug-level message. """ if COMPLETION_ENABLED: argcomplete.autocomplete(parser) elif 'bash' in os.getenv('SHELL', ''): logger.debug('Bash completion not available. Install argcomplete.')
def autocomplete(parser): """ Adds support for shell completion via argcomplete_ by patching given `argparse.ArgumentParser` (sub)class. If completion is not enabled, logs a debug-level message. """ if COMPLETION_ENABLED: argcomplete.autocomplete(parser) elif 'bash' in os.getenv('SHELL', ''): logger.debug('Bash completion not available. Install argcomplete.')
[ "Adds", "support", "for", "shell", "completion", "via", "argcomplete_", "by", "patching", "given", "argparse", ".", "ArgumentParser", "(", "sub", ")", "class", "." ]
neithere/argh
python
https://github.com/neithere/argh/blob/dcd3253f2994400a6a58a700c118c53765bc50a4/argh/completion.py#L84-L94
[ "def", "autocomplete", "(", "parser", ")", ":", "if", "COMPLETION_ENABLED", ":", "argcomplete", ".", "autocomplete", "(", "parser", ")", "elif", "'bash'", "in", "os", ".", "getenv", "(", "'SHELL'", ",", "''", ")", ":", "logger", ".", "debug", "(", "'Bash completion not available. Install argcomplete.'", ")" ]
dcd3253f2994400a6a58a700c118c53765bc50a4
test
ArghParser.parse_args
Wrapper for :meth:`argparse.ArgumentParser.parse_args`. If `namespace` is not defined, :class:`argh.dispatching.ArghNamespace` is used. This is required for functions to be properly used as commands.
argh/helpers.py
def parse_args(self, args=None, namespace=None): """ Wrapper for :meth:`argparse.ArgumentParser.parse_args`. If `namespace` is not defined, :class:`argh.dispatching.ArghNamespace` is used. This is required for functions to be properly used as commands. """ namespace = namespace or ArghNamespace() return super(ArghParser, self).parse_args(args, namespace)
def parse_args(self, args=None, namespace=None): """ Wrapper for :meth:`argparse.ArgumentParser.parse_args`. If `namespace` is not defined, :class:`argh.dispatching.ArghNamespace` is used. This is required for functions to be properly used as commands. """ namespace = namespace or ArghNamespace() return super(ArghParser, self).parse_args(args, namespace)
[ "Wrapper", "for", ":", "meth", ":", "argparse", ".", "ArgumentParser", ".", "parse_args", ".", "If", "namespace", "is", "not", "defined", ":", "class", ":", "argh", ".", "dispatching", ".", "ArghNamespace", "is", "used", ".", "This", "is", "required", "for", "functions", "to", "be", "properly", "used", "as", "commands", "." ]
neithere/argh
python
https://github.com/neithere/argh/blob/dcd3253f2994400a6a58a700c118c53765bc50a4/argh/helpers.py#L57-L64
[ "def", "parse_args", "(", "self", ",", "args", "=", "None", ",", "namespace", "=", "None", ")", ":", "namespace", "=", "namespace", "or", "ArghNamespace", "(", ")", "return", "super", "(", "ArghParser", ",", "self", ")", ".", "parse_args", "(", "args", ",", "namespace", ")" ]
dcd3253f2994400a6a58a700c118c53765bc50a4
test
CustomFormatter._expand_help
This method is copied verbatim from ArgumentDefaultsHelpFormatter with a couple of lines added just before the end. Reason: we need to `repr()` default values instead of simply inserting them as is. This helps notice, for example, an empty string as the default value; moreover, it prevents breaking argparse due to logical quirks inside of its formatters. Ideally this could be achieved by simply defining :attr:`DEFAULT_ARGUMENT_TEMPLATE` as ``{default!r}`` but unfortunately argparse only supports the old printf syntax.
argh/constants.py
def _expand_help(self, action): """ This method is copied verbatim from ArgumentDefaultsHelpFormatter with a couple of lines added just before the end. Reason: we need to `repr()` default values instead of simply inserting them as is. This helps notice, for example, an empty string as the default value; moreover, it prevents breaking argparse due to logical quirks inside of its formatters. Ideally this could be achieved by simply defining :attr:`DEFAULT_ARGUMENT_TEMPLATE` as ``{default!r}`` but unfortunately argparse only supports the old printf syntax. """ params = dict(vars(action), prog=self._prog) for name in list(params): if params[name] is argparse.SUPPRESS: del params[name] for name in list(params): if hasattr(params[name], '__name__'): params[name] = params[name].__name__ if params.get('choices') is not None: choices_str = ', '.join([str(c) for c in params['choices']]) params['choices'] = choices_str # XXX this is added in Argh vs. argparse.ArgumentDefaultsHelpFormatter # (avoiding empty strings, otherwise Argparse would die with # an IndexError in _format_action) # if 'default' in params: if params['default'] is None: params['default'] = '-' else: params['default'] = repr(params['default']) # # / return self._get_help_string(action) % params
def _expand_help(self, action): """ This method is copied verbatim from ArgumentDefaultsHelpFormatter with a couple of lines added just before the end. Reason: we need to `repr()` default values instead of simply inserting them as is. This helps notice, for example, an empty string as the default value; moreover, it prevents breaking argparse due to logical quirks inside of its formatters. Ideally this could be achieved by simply defining :attr:`DEFAULT_ARGUMENT_TEMPLATE` as ``{default!r}`` but unfortunately argparse only supports the old printf syntax. """ params = dict(vars(action), prog=self._prog) for name in list(params): if params[name] is argparse.SUPPRESS: del params[name] for name in list(params): if hasattr(params[name], '__name__'): params[name] = params[name].__name__ if params.get('choices') is not None: choices_str = ', '.join([str(c) for c in params['choices']]) params['choices'] = choices_str # XXX this is added in Argh vs. argparse.ArgumentDefaultsHelpFormatter # (avoiding empty strings, otherwise Argparse would die with # an IndexError in _format_action) # if 'default' in params: if params['default'] is None: params['default'] = '-' else: params['default'] = repr(params['default']) # # / return self._get_help_string(action) % params
[ "This", "method", "is", "copied", "verbatim", "from", "ArgumentDefaultsHelpFormatter", "with", "a", "couple", "of", "lines", "added", "just", "before", "the", "end", ".", "Reason", ":", "we", "need", "to", "repr", "()", "default", "values", "instead", "of", "simply", "inserting", "them", "as", "is", ".", "This", "helps", "notice", "for", "example", "an", "empty", "string", "as", "the", "default", "value", ";", "moreover", "it", "prevents", "breaking", "argparse", "due", "to", "logical", "quirks", "inside", "of", "its", "formatters", "." ]
neithere/argh
python
https://github.com/neithere/argh/blob/dcd3253f2994400a6a58a700c118c53765bc50a4/argh/constants.py#L55-L91
[ "def", "_expand_help", "(", "self", ",", "action", ")", ":", "params", "=", "dict", "(", "vars", "(", "action", ")", ",", "prog", "=", "self", ".", "_prog", ")", "for", "name", "in", "list", "(", "params", ")", ":", "if", "params", "[", "name", "]", "is", "argparse", ".", "SUPPRESS", ":", "del", "params", "[", "name", "]", "for", "name", "in", "list", "(", "params", ")", ":", "if", "hasattr", "(", "params", "[", "name", "]", ",", "'__name__'", ")", ":", "params", "[", "name", "]", "=", "params", "[", "name", "]", ".", "__name__", "if", "params", ".", "get", "(", "'choices'", ")", "is", "not", "None", ":", "choices_str", "=", "', '", ".", "join", "(", "[", "str", "(", "c", ")", "for", "c", "in", "params", "[", "'choices'", "]", "]", ")", "params", "[", "'choices'", "]", "=", "choices_str", "# XXX this is added in Argh vs. argparse.ArgumentDefaultsHelpFormatter", "# (avoiding empty strings, otherwise Argparse would die with", "# an IndexError in _format_action)", "#", "if", "'default'", "in", "params", ":", "if", "params", "[", "'default'", "]", "is", "None", ":", "params", "[", "'default'", "]", "=", "'-'", "else", ":", "params", "[", "'default'", "]", "=", "repr", "(", "params", "[", "'default'", "]", ")", "#", "# /", "return", "self", ".", "_get_help_string", "(", "action", ")", "%", "params" ]
dcd3253f2994400a6a58a700c118c53765bc50a4
test
_guess
Adds types, actions, etc. to given argument specification. For example, ``default=3`` implies ``type=int``. :param arg: a :class:`argh.utils.Arg` instance
argh/assembling.py
def _guess(kwargs): """ Adds types, actions, etc. to given argument specification. For example, ``default=3`` implies ``type=int``. :param arg: a :class:`argh.utils.Arg` instance """ guessed = {} # Parser actions that accept argument 'type' TYPE_AWARE_ACTIONS = 'store', 'append' # guess type/action from default value value = kwargs.get('default') if value is not None: if isinstance(value, bool): if kwargs.get('action') is None: # infer action from default value guessed['action'] = 'store_false' if value else 'store_true' elif kwargs.get('type') is None: # infer type from default value # (make sure that action handler supports this keyword) if kwargs.get('action', 'store') in TYPE_AWARE_ACTIONS: guessed['type'] = type(value) # guess type from choices (first item) if kwargs.get('choices') and 'type' not in list(guessed) + list(kwargs): guessed['type'] = type(kwargs['choices'][0]) return dict(kwargs, **guessed)
def _guess(kwargs): """ Adds types, actions, etc. to given argument specification. For example, ``default=3`` implies ``type=int``. :param arg: a :class:`argh.utils.Arg` instance """ guessed = {} # Parser actions that accept argument 'type' TYPE_AWARE_ACTIONS = 'store', 'append' # guess type/action from default value value = kwargs.get('default') if value is not None: if isinstance(value, bool): if kwargs.get('action') is None: # infer action from default value guessed['action'] = 'store_false' if value else 'store_true' elif kwargs.get('type') is None: # infer type from default value # (make sure that action handler supports this keyword) if kwargs.get('action', 'store') in TYPE_AWARE_ACTIONS: guessed['type'] = type(value) # guess type from choices (first item) if kwargs.get('choices') and 'type' not in list(guessed) + list(kwargs): guessed['type'] = type(kwargs['choices'][0]) return dict(kwargs, **guessed)
[ "Adds", "types", "actions", "etc", ".", "to", "given", "argument", "specification", ".", "For", "example", "default", "=", "3", "implies", "type", "=", "int", "." ]
neithere/argh
python
https://github.com/neithere/argh/blob/dcd3253f2994400a6a58a700c118c53765bc50a4/argh/assembling.py#L120-L149
[ "def", "_guess", "(", "kwargs", ")", ":", "guessed", "=", "{", "}", "# Parser actions that accept argument 'type'", "TYPE_AWARE_ACTIONS", "=", "'store'", ",", "'append'", "# guess type/action from default value", "value", "=", "kwargs", ".", "get", "(", "'default'", ")", "if", "value", "is", "not", "None", ":", "if", "isinstance", "(", "value", ",", "bool", ")", ":", "if", "kwargs", ".", "get", "(", "'action'", ")", "is", "None", ":", "# infer action from default value", "guessed", "[", "'action'", "]", "=", "'store_false'", "if", "value", "else", "'store_true'", "elif", "kwargs", ".", "get", "(", "'type'", ")", "is", "None", ":", "# infer type from default value", "# (make sure that action handler supports this keyword)", "if", "kwargs", ".", "get", "(", "'action'", ",", "'store'", ")", "in", "TYPE_AWARE_ACTIONS", ":", "guessed", "[", "'type'", "]", "=", "type", "(", "value", ")", "# guess type from choices (first item)", "if", "kwargs", ".", "get", "(", "'choices'", ")", "and", "'type'", "not", "in", "list", "(", "guessed", ")", "+", "list", "(", "kwargs", ")", ":", "guessed", "[", "'type'", "]", "=", "type", "(", "kwargs", "[", "'choices'", "]", "[", "0", "]", ")", "return", "dict", "(", "kwargs", ",", "*", "*", "guessed", ")" ]
dcd3253f2994400a6a58a700c118c53765bc50a4
test
set_default_command
Sets default command (i.e. a function) for given parser. If `parser.description` is empty and the function has a docstring, it is used as the description. .. note:: An attempt to set default command to a parser which already has subparsers (e.g. added with :func:`~argh.assembling.add_commands`) results in a `AssemblingError`. .. note:: If there are both explicitly declared arguments (e.g. via :func:`~argh.decorators.arg`) and ones inferred from the function signature (e.g. via :func:`~argh.decorators.command`), declared ones will be merged into inferred ones. If an argument does not conform function signature, `AssemblingError` is raised. .. note:: If the parser was created with ``add_help=True`` (which is by default), option name ``-h`` is silently removed from any argument.
argh/assembling.py
def set_default_command(parser, function): """ Sets default command (i.e. a function) for given parser. If `parser.description` is empty and the function has a docstring, it is used as the description. .. note:: An attempt to set default command to a parser which already has subparsers (e.g. added with :func:`~argh.assembling.add_commands`) results in a `AssemblingError`. .. note:: If there are both explicitly declared arguments (e.g. via :func:`~argh.decorators.arg`) and ones inferred from the function signature (e.g. via :func:`~argh.decorators.command`), declared ones will be merged into inferred ones. If an argument does not conform function signature, `AssemblingError` is raised. .. note:: If the parser was created with ``add_help=True`` (which is by default), option name ``-h`` is silently removed from any argument. """ if parser._subparsers: _require_support_for_default_command_with_subparsers() spec = get_arg_spec(function) declared_args = getattr(function, ATTR_ARGS, []) inferred_args = list(_get_args_from_signature(function)) if inferred_args and declared_args: # We've got a mixture of declared and inferred arguments # a mapping of "dest" strings to argument declarations. # # * a "dest" string is a normalized form of argument name, i.e.: # # '-f', '--foo' → 'foo' # 'foo-bar' → 'foo_bar' # # * argument declaration is a dictionary representing an argument; # it is obtained either from _get_args_from_signature() or from # an @arg decorator (as is). # dests = OrderedDict() for argspec in inferred_args: dest = _get_parser_param_kwargs(parser, argspec)['dest'] dests[dest] = argspec for declared_kw in declared_args: # an argument is declared via decorator dest = _get_dest(parser, declared_kw) if dest in dests: # the argument is already known from function signature # # now make sure that this declared arg conforms to the function # signature and therefore only refines an inferred arg: # # @arg('my-foo') maps to func(my_foo) # @arg('--my-bar') maps to func(my_bar=...) # either both arguments are positional or both are optional decl_positional = _is_positional(declared_kw['option_strings']) infr_positional = _is_positional(dests[dest]['option_strings']) if decl_positional != infr_positional: kinds = {True: 'positional', False: 'optional'} raise AssemblingError( '{func}: argument "{dest}" declared as {kind_i} ' '(in function signature) and {kind_d} (via decorator)' .format( func=function.__name__, dest=dest, kind_i=kinds[infr_positional], kind_d=kinds[decl_positional], )) # merge explicit argument declaration into the inferred one # (e.g. `help=...`) dests[dest].update(**declared_kw) else: # the argument is not in function signature varkw = getattr(spec, 'varkw', getattr(spec, 'keywords', [])) if varkw: # function accepts **kwargs; the argument goes into it dests[dest] = declared_kw else: # there's no way we can map the argument declaration # to function signature xs = (dests[x]['option_strings'] for x in dests) raise AssemblingError( '{func}: argument {flags} does not fit ' 'function signature: {sig}'.format( flags=', '.join(declared_kw['option_strings']), func=function.__name__, sig=', '.join('/'.join(x) for x in xs))) # pack the modified data back into a list inferred_args = dests.values() command_args = inferred_args or declared_args # add types, actions, etc. (e.g. default=3 implies type=int) command_args = [_guess(x) for x in command_args] for draft in command_args: draft = draft.copy() if 'help' not in draft: draft.update(help=DEFAULT_ARGUMENT_TEMPLATE) dest_or_opt_strings = draft.pop('option_strings') if parser.add_help and '-h' in dest_or_opt_strings: dest_or_opt_strings = [x for x in dest_or_opt_strings if x != '-h'] completer = draft.pop('completer', None) try: action = parser.add_argument(*dest_or_opt_strings, **draft) if COMPLETION_ENABLED and completer: action.completer = completer except Exception as e: raise type(e)('{func}: cannot add arg {args}: {msg}'.format( args='/'.join(dest_or_opt_strings), func=function.__name__, msg=e)) if function.__doc__ and not parser.description: parser.description = function.__doc__ parser.set_defaults(**{ DEST_FUNCTION: function, })
def set_default_command(parser, function): """ Sets default command (i.e. a function) for given parser. If `parser.description` is empty and the function has a docstring, it is used as the description. .. note:: An attempt to set default command to a parser which already has subparsers (e.g. added with :func:`~argh.assembling.add_commands`) results in a `AssemblingError`. .. note:: If there are both explicitly declared arguments (e.g. via :func:`~argh.decorators.arg`) and ones inferred from the function signature (e.g. via :func:`~argh.decorators.command`), declared ones will be merged into inferred ones. If an argument does not conform function signature, `AssemblingError` is raised. .. note:: If the parser was created with ``add_help=True`` (which is by default), option name ``-h`` is silently removed from any argument. """ if parser._subparsers: _require_support_for_default_command_with_subparsers() spec = get_arg_spec(function) declared_args = getattr(function, ATTR_ARGS, []) inferred_args = list(_get_args_from_signature(function)) if inferred_args and declared_args: # We've got a mixture of declared and inferred arguments # a mapping of "dest" strings to argument declarations. # # * a "dest" string is a normalized form of argument name, i.e.: # # '-f', '--foo' → 'foo' # 'foo-bar' → 'foo_bar' # # * argument declaration is a dictionary representing an argument; # it is obtained either from _get_args_from_signature() or from # an @arg decorator (as is). # dests = OrderedDict() for argspec in inferred_args: dest = _get_parser_param_kwargs(parser, argspec)['dest'] dests[dest] = argspec for declared_kw in declared_args: # an argument is declared via decorator dest = _get_dest(parser, declared_kw) if dest in dests: # the argument is already known from function signature # # now make sure that this declared arg conforms to the function # signature and therefore only refines an inferred arg: # # @arg('my-foo') maps to func(my_foo) # @arg('--my-bar') maps to func(my_bar=...) # either both arguments are positional or both are optional decl_positional = _is_positional(declared_kw['option_strings']) infr_positional = _is_positional(dests[dest]['option_strings']) if decl_positional != infr_positional: kinds = {True: 'positional', False: 'optional'} raise AssemblingError( '{func}: argument "{dest}" declared as {kind_i} ' '(in function signature) and {kind_d} (via decorator)' .format( func=function.__name__, dest=dest, kind_i=kinds[infr_positional], kind_d=kinds[decl_positional], )) # merge explicit argument declaration into the inferred one # (e.g. `help=...`) dests[dest].update(**declared_kw) else: # the argument is not in function signature varkw = getattr(spec, 'varkw', getattr(spec, 'keywords', [])) if varkw: # function accepts **kwargs; the argument goes into it dests[dest] = declared_kw else: # there's no way we can map the argument declaration # to function signature xs = (dests[x]['option_strings'] for x in dests) raise AssemblingError( '{func}: argument {flags} does not fit ' 'function signature: {sig}'.format( flags=', '.join(declared_kw['option_strings']), func=function.__name__, sig=', '.join('/'.join(x) for x in xs))) # pack the modified data back into a list inferred_args = dests.values() command_args = inferred_args or declared_args # add types, actions, etc. (e.g. default=3 implies type=int) command_args = [_guess(x) for x in command_args] for draft in command_args: draft = draft.copy() if 'help' not in draft: draft.update(help=DEFAULT_ARGUMENT_TEMPLATE) dest_or_opt_strings = draft.pop('option_strings') if parser.add_help and '-h' in dest_or_opt_strings: dest_or_opt_strings = [x for x in dest_or_opt_strings if x != '-h'] completer = draft.pop('completer', None) try: action = parser.add_argument(*dest_or_opt_strings, **draft) if COMPLETION_ENABLED and completer: action.completer = completer except Exception as e: raise type(e)('{func}: cannot add arg {args}: {msg}'.format( args='/'.join(dest_or_opt_strings), func=function.__name__, msg=e)) if function.__doc__ and not parser.description: parser.description = function.__doc__ parser.set_defaults(**{ DEST_FUNCTION: function, })
[ "Sets", "default", "command", "(", "i", ".", "e", ".", "a", "function", ")", "for", "given", "parser", "." ]
neithere/argh
python
https://github.com/neithere/argh/blob/dcd3253f2994400a6a58a700c118c53765bc50a4/argh/assembling.py#L188-L318
[ "def", "set_default_command", "(", "parser", ",", "function", ")", ":", "if", "parser", ".", "_subparsers", ":", "_require_support_for_default_command_with_subparsers", "(", ")", "spec", "=", "get_arg_spec", "(", "function", ")", "declared_args", "=", "getattr", "(", "function", ",", "ATTR_ARGS", ",", "[", "]", ")", "inferred_args", "=", "list", "(", "_get_args_from_signature", "(", "function", ")", ")", "if", "inferred_args", "and", "declared_args", ":", "# We've got a mixture of declared and inferred arguments", "# a mapping of \"dest\" strings to argument declarations.", "#", "# * a \"dest\" string is a normalized form of argument name, i.e.:", "#", "# '-f', '--foo' → 'foo'", "# 'foo-bar' → 'foo_bar'", "#", "# * argument declaration is a dictionary representing an argument;", "# it is obtained either from _get_args_from_signature() or from", "# an @arg decorator (as is).", "#", "dests", "=", "OrderedDict", "(", ")", "for", "argspec", "in", "inferred_args", ":", "dest", "=", "_get_parser_param_kwargs", "(", "parser", ",", "argspec", ")", "[", "'dest'", "]", "dests", "[", "dest", "]", "=", "argspec", "for", "declared_kw", "in", "declared_args", ":", "# an argument is declared via decorator", "dest", "=", "_get_dest", "(", "parser", ",", "declared_kw", ")", "if", "dest", "in", "dests", ":", "# the argument is already known from function signature", "#", "# now make sure that this declared arg conforms to the function", "# signature and therefore only refines an inferred arg:", "#", "# @arg('my-foo') maps to func(my_foo)", "# @arg('--my-bar') maps to func(my_bar=...)", "# either both arguments are positional or both are optional", "decl_positional", "=", "_is_positional", "(", "declared_kw", "[", "'option_strings'", "]", ")", "infr_positional", "=", "_is_positional", "(", "dests", "[", "dest", "]", "[", "'option_strings'", "]", ")", "if", "decl_positional", "!=", "infr_positional", ":", "kinds", "=", "{", "True", ":", "'positional'", ",", "False", ":", "'optional'", "}", "raise", "AssemblingError", "(", "'{func}: argument \"{dest}\" declared as {kind_i} '", "'(in function signature) and {kind_d} (via decorator)'", ".", "format", "(", "func", "=", "function", ".", "__name__", ",", "dest", "=", "dest", ",", "kind_i", "=", "kinds", "[", "infr_positional", "]", ",", "kind_d", "=", "kinds", "[", "decl_positional", "]", ",", ")", ")", "# merge explicit argument declaration into the inferred one", "# (e.g. `help=...`)", "dests", "[", "dest", "]", ".", "update", "(", "*", "*", "declared_kw", ")", "else", ":", "# the argument is not in function signature", "varkw", "=", "getattr", "(", "spec", ",", "'varkw'", ",", "getattr", "(", "spec", ",", "'keywords'", ",", "[", "]", ")", ")", "if", "varkw", ":", "# function accepts **kwargs; the argument goes into it", "dests", "[", "dest", "]", "=", "declared_kw", "else", ":", "# there's no way we can map the argument declaration", "# to function signature", "xs", "=", "(", "dests", "[", "x", "]", "[", "'option_strings'", "]", "for", "x", "in", "dests", ")", "raise", "AssemblingError", "(", "'{func}: argument {flags} does not fit '", "'function signature: {sig}'", ".", "format", "(", "flags", "=", "', '", ".", "join", "(", "declared_kw", "[", "'option_strings'", "]", ")", ",", "func", "=", "function", ".", "__name__", ",", "sig", "=", "', '", ".", "join", "(", "'/'", ".", "join", "(", "x", ")", "for", "x", "in", "xs", ")", ")", ")", "# pack the modified data back into a list", "inferred_args", "=", "dests", ".", "values", "(", ")", "command_args", "=", "inferred_args", "or", "declared_args", "# add types, actions, etc. (e.g. default=3 implies type=int)", "command_args", "=", "[", "_guess", "(", "x", ")", "for", "x", "in", "command_args", "]", "for", "draft", "in", "command_args", ":", "draft", "=", "draft", ".", "copy", "(", ")", "if", "'help'", "not", "in", "draft", ":", "draft", ".", "update", "(", "help", "=", "DEFAULT_ARGUMENT_TEMPLATE", ")", "dest_or_opt_strings", "=", "draft", ".", "pop", "(", "'option_strings'", ")", "if", "parser", ".", "add_help", "and", "'-h'", "in", "dest_or_opt_strings", ":", "dest_or_opt_strings", "=", "[", "x", "for", "x", "in", "dest_or_opt_strings", "if", "x", "!=", "'-h'", "]", "completer", "=", "draft", ".", "pop", "(", "'completer'", ",", "None", ")", "try", ":", "action", "=", "parser", ".", "add_argument", "(", "*", "dest_or_opt_strings", ",", "*", "*", "draft", ")", "if", "COMPLETION_ENABLED", "and", "completer", ":", "action", ".", "completer", "=", "completer", "except", "Exception", "as", "e", ":", "raise", "type", "(", "e", ")", "(", "'{func}: cannot add arg {args}: {msg}'", ".", "format", "(", "args", "=", "'/'", ".", "join", "(", "dest_or_opt_strings", ")", ",", "func", "=", "function", ".", "__name__", ",", "msg", "=", "e", ")", ")", "if", "function", ".", "__doc__", "and", "not", "parser", ".", "description", ":", "parser", ".", "description", "=", "function", ".", "__doc__", "parser", ".", "set_defaults", "(", "*", "*", "{", "DEST_FUNCTION", ":", "function", ",", "}", ")" ]
dcd3253f2994400a6a58a700c118c53765bc50a4
test
add_commands
Adds given functions as commands to given parser. :param parser: an :class:`argparse.ArgumentParser` instance. :param functions: a list of functions. A subparser is created for each of them. If the function is decorated with :func:`~argh.decorators.arg`, the arguments are passed to :class:`argparse.ArgumentParser.add_argument`. See also :func:`~argh.dispatching.dispatch` for requirements concerning function signatures. The command name is inferred from the function name. Note that the underscores in the name are replaced with hyphens, i.e. function name "foo_bar" becomes command name "foo-bar". :param namespace: an optional string representing the group of commands. For example, if a command named "hello" is added without the namespace, it will be available as "prog.py hello"; if the namespace if specified as "greet", then the command will be accessible as "prog.py greet hello". The namespace itself is not callable, so "prog.py greet" will fail and only display a help message. :param func_kwargs: a `dict` of keyword arguments to be passed to each nested ArgumentParser instance created per command (i.e. per function). Members of this dictionary have the highest priority, so a function's docstring is overridden by a `help` in `func_kwargs` (if present). :param namespace_kwargs: a `dict` of keyword arguments to be passed to the nested ArgumentParser instance under given `namespace`. Deprecated params that should be moved into `namespace_kwargs`: :param title: passed to :meth:`argparse.ArgumentParser.add_subparsers` as `title`. .. deprecated:: 0.26.0 Please use `namespace_kwargs` instead. :param description: passed to :meth:`argparse.ArgumentParser.add_subparsers` as `description`. .. deprecated:: 0.26.0 Please use `namespace_kwargs` instead. :param help: passed to :meth:`argparse.ArgumentParser.add_subparsers` as `help`. .. deprecated:: 0.26.0 Please use `namespace_kwargs` instead. .. note:: This function modifies the parser object. Generally side effects are bad practice but we don't seem to have any choice as ArgumentParser is pretty opaque. You may prefer :class:`~argh.helpers.ArghParser.add_commands` for a bit more predictable API. .. note:: An attempt to add commands to a parser which already has a default function (e.g. added with :func:`~argh.assembling.set_default_command`) results in `AssemblingError`.
argh/assembling.py
def add_commands(parser, functions, namespace=None, namespace_kwargs=None, func_kwargs=None, # deprecated args: title=None, description=None, help=None): """ Adds given functions as commands to given parser. :param parser: an :class:`argparse.ArgumentParser` instance. :param functions: a list of functions. A subparser is created for each of them. If the function is decorated with :func:`~argh.decorators.arg`, the arguments are passed to :class:`argparse.ArgumentParser.add_argument`. See also :func:`~argh.dispatching.dispatch` for requirements concerning function signatures. The command name is inferred from the function name. Note that the underscores in the name are replaced with hyphens, i.e. function name "foo_bar" becomes command name "foo-bar". :param namespace: an optional string representing the group of commands. For example, if a command named "hello" is added without the namespace, it will be available as "prog.py hello"; if the namespace if specified as "greet", then the command will be accessible as "prog.py greet hello". The namespace itself is not callable, so "prog.py greet" will fail and only display a help message. :param func_kwargs: a `dict` of keyword arguments to be passed to each nested ArgumentParser instance created per command (i.e. per function). Members of this dictionary have the highest priority, so a function's docstring is overridden by a `help` in `func_kwargs` (if present). :param namespace_kwargs: a `dict` of keyword arguments to be passed to the nested ArgumentParser instance under given `namespace`. Deprecated params that should be moved into `namespace_kwargs`: :param title: passed to :meth:`argparse.ArgumentParser.add_subparsers` as `title`. .. deprecated:: 0.26.0 Please use `namespace_kwargs` instead. :param description: passed to :meth:`argparse.ArgumentParser.add_subparsers` as `description`. .. deprecated:: 0.26.0 Please use `namespace_kwargs` instead. :param help: passed to :meth:`argparse.ArgumentParser.add_subparsers` as `help`. .. deprecated:: 0.26.0 Please use `namespace_kwargs` instead. .. note:: This function modifies the parser object. Generally side effects are bad practice but we don't seem to have any choice as ArgumentParser is pretty opaque. You may prefer :class:`~argh.helpers.ArghParser.add_commands` for a bit more predictable API. .. note:: An attempt to add commands to a parser which already has a default function (e.g. added with :func:`~argh.assembling.set_default_command`) results in `AssemblingError`. """ # FIXME "namespace" is a correct name but it clashes with the "namespace" # that represents arguments (argparse.Namespace and our ArghNamespace). # We should rename the argument here. if DEST_FUNCTION in parser._defaults: _require_support_for_default_command_with_subparsers() namespace_kwargs = namespace_kwargs or {} # FIXME remove this by 1.0 # if title: warnings.warn('argument `title` is deprecated in add_commands(),' ' use `parser_kwargs` instead', DeprecationWarning) namespace_kwargs['description'] = title if help: warnings.warn('argument `help` is deprecated in add_commands(),' ' use `parser_kwargs` instead', DeprecationWarning) namespace_kwargs['help'] = help if description: warnings.warn('argument `description` is deprecated in add_commands(),' ' use `parser_kwargs` instead', DeprecationWarning) namespace_kwargs['description'] = description # # / subparsers_action = get_subparsers(parser, create=True) if namespace: # Make a nested parser and init a deeper _SubParsersAction under it. # Create a named group of commands. It will be listed along with # root-level commands in ``app.py --help``; in that context its `title` # can be used as a short description on the right side of its name. # Normally `title` is shown above the list of commands # in ``app.py my-namespace --help``. subsubparser_kw = { 'help': namespace_kwargs.get('title'), } subsubparser = subparsers_action.add_parser(namespace, **subsubparser_kw) subparsers_action = subsubparser.add_subparsers(**namespace_kwargs) else: assert not namespace_kwargs, ('`parser_kwargs` only makes sense ' 'with `namespace`.') for func in functions: cmd_name, func_parser_kwargs = _extract_command_meta_from_func(func) # override any computed kwargs by manually supplied ones if func_kwargs: func_parser_kwargs.update(func_kwargs) # create and set up the parser for this command command_parser = subparsers_action.add_parser(cmd_name, **func_parser_kwargs) set_default_command(command_parser, func)
def add_commands(parser, functions, namespace=None, namespace_kwargs=None, func_kwargs=None, # deprecated args: title=None, description=None, help=None): """ Adds given functions as commands to given parser. :param parser: an :class:`argparse.ArgumentParser` instance. :param functions: a list of functions. A subparser is created for each of them. If the function is decorated with :func:`~argh.decorators.arg`, the arguments are passed to :class:`argparse.ArgumentParser.add_argument`. See also :func:`~argh.dispatching.dispatch` for requirements concerning function signatures. The command name is inferred from the function name. Note that the underscores in the name are replaced with hyphens, i.e. function name "foo_bar" becomes command name "foo-bar". :param namespace: an optional string representing the group of commands. For example, if a command named "hello" is added without the namespace, it will be available as "prog.py hello"; if the namespace if specified as "greet", then the command will be accessible as "prog.py greet hello". The namespace itself is not callable, so "prog.py greet" will fail and only display a help message. :param func_kwargs: a `dict` of keyword arguments to be passed to each nested ArgumentParser instance created per command (i.e. per function). Members of this dictionary have the highest priority, so a function's docstring is overridden by a `help` in `func_kwargs` (if present). :param namespace_kwargs: a `dict` of keyword arguments to be passed to the nested ArgumentParser instance under given `namespace`. Deprecated params that should be moved into `namespace_kwargs`: :param title: passed to :meth:`argparse.ArgumentParser.add_subparsers` as `title`. .. deprecated:: 0.26.0 Please use `namespace_kwargs` instead. :param description: passed to :meth:`argparse.ArgumentParser.add_subparsers` as `description`. .. deprecated:: 0.26.0 Please use `namespace_kwargs` instead. :param help: passed to :meth:`argparse.ArgumentParser.add_subparsers` as `help`. .. deprecated:: 0.26.0 Please use `namespace_kwargs` instead. .. note:: This function modifies the parser object. Generally side effects are bad practice but we don't seem to have any choice as ArgumentParser is pretty opaque. You may prefer :class:`~argh.helpers.ArghParser.add_commands` for a bit more predictable API. .. note:: An attempt to add commands to a parser which already has a default function (e.g. added with :func:`~argh.assembling.set_default_command`) results in `AssemblingError`. """ # FIXME "namespace" is a correct name but it clashes with the "namespace" # that represents arguments (argparse.Namespace and our ArghNamespace). # We should rename the argument here. if DEST_FUNCTION in parser._defaults: _require_support_for_default_command_with_subparsers() namespace_kwargs = namespace_kwargs or {} # FIXME remove this by 1.0 # if title: warnings.warn('argument `title` is deprecated in add_commands(),' ' use `parser_kwargs` instead', DeprecationWarning) namespace_kwargs['description'] = title if help: warnings.warn('argument `help` is deprecated in add_commands(),' ' use `parser_kwargs` instead', DeprecationWarning) namespace_kwargs['help'] = help if description: warnings.warn('argument `description` is deprecated in add_commands(),' ' use `parser_kwargs` instead', DeprecationWarning) namespace_kwargs['description'] = description # # / subparsers_action = get_subparsers(parser, create=True) if namespace: # Make a nested parser and init a deeper _SubParsersAction under it. # Create a named group of commands. It will be listed along with # root-level commands in ``app.py --help``; in that context its `title` # can be used as a short description on the right side of its name. # Normally `title` is shown above the list of commands # in ``app.py my-namespace --help``. subsubparser_kw = { 'help': namespace_kwargs.get('title'), } subsubparser = subparsers_action.add_parser(namespace, **subsubparser_kw) subparsers_action = subsubparser.add_subparsers(**namespace_kwargs) else: assert not namespace_kwargs, ('`parser_kwargs` only makes sense ' 'with `namespace`.') for func in functions: cmd_name, func_parser_kwargs = _extract_command_meta_from_func(func) # override any computed kwargs by manually supplied ones if func_kwargs: func_parser_kwargs.update(func_kwargs) # create and set up the parser for this command command_parser = subparsers_action.add_parser(cmd_name, **func_parser_kwargs) set_default_command(command_parser, func)
[ "Adds", "given", "functions", "as", "commands", "to", "given", "parser", "." ]
neithere/argh
python
https://github.com/neithere/argh/blob/dcd3253f2994400a6a58a700c118c53765bc50a4/argh/assembling.py#L321-L459
[ "def", "add_commands", "(", "parser", ",", "functions", ",", "namespace", "=", "None", ",", "namespace_kwargs", "=", "None", ",", "func_kwargs", "=", "None", ",", "# deprecated args:", "title", "=", "None", ",", "description", "=", "None", ",", "help", "=", "None", ")", ":", "# FIXME \"namespace\" is a correct name but it clashes with the \"namespace\"", "# that represents arguments (argparse.Namespace and our ArghNamespace).", "# We should rename the argument here.", "if", "DEST_FUNCTION", "in", "parser", ".", "_defaults", ":", "_require_support_for_default_command_with_subparsers", "(", ")", "namespace_kwargs", "=", "namespace_kwargs", "or", "{", "}", "# FIXME remove this by 1.0", "#", "if", "title", ":", "warnings", ".", "warn", "(", "'argument `title` is deprecated in add_commands(),'", "' use `parser_kwargs` instead'", ",", "DeprecationWarning", ")", "namespace_kwargs", "[", "'description'", "]", "=", "title", "if", "help", ":", "warnings", ".", "warn", "(", "'argument `help` is deprecated in add_commands(),'", "' use `parser_kwargs` instead'", ",", "DeprecationWarning", ")", "namespace_kwargs", "[", "'help'", "]", "=", "help", "if", "description", ":", "warnings", ".", "warn", "(", "'argument `description` is deprecated in add_commands(),'", "' use `parser_kwargs` instead'", ",", "DeprecationWarning", ")", "namespace_kwargs", "[", "'description'", "]", "=", "description", "#", "# /", "subparsers_action", "=", "get_subparsers", "(", "parser", ",", "create", "=", "True", ")", "if", "namespace", ":", "# Make a nested parser and init a deeper _SubParsersAction under it.", "# Create a named group of commands. It will be listed along with", "# root-level commands in ``app.py --help``; in that context its `title`", "# can be used as a short description on the right side of its name.", "# Normally `title` is shown above the list of commands", "# in ``app.py my-namespace --help``.", "subsubparser_kw", "=", "{", "'help'", ":", "namespace_kwargs", ".", "get", "(", "'title'", ")", ",", "}", "subsubparser", "=", "subparsers_action", ".", "add_parser", "(", "namespace", ",", "*", "*", "subsubparser_kw", ")", "subparsers_action", "=", "subsubparser", ".", "add_subparsers", "(", "*", "*", "namespace_kwargs", ")", "else", ":", "assert", "not", "namespace_kwargs", ",", "(", "'`parser_kwargs` only makes sense '", "'with `namespace`.'", ")", "for", "func", "in", "functions", ":", "cmd_name", ",", "func_parser_kwargs", "=", "_extract_command_meta_from_func", "(", "func", ")", "# override any computed kwargs by manually supplied ones", "if", "func_kwargs", ":", "func_parser_kwargs", ".", "update", "(", "func_kwargs", ")", "# create and set up the parser for this command", "command_parser", "=", "subparsers_action", ".", "add_parser", "(", "cmd_name", ",", "*", "*", "func_parser_kwargs", ")", "set_default_command", "(", "command_parser", ",", "func", ")" ]
dcd3253f2994400a6a58a700c118c53765bc50a4
test
add_subcommands
A wrapper for :func:`add_commands`. These examples are equivalent:: add_commands(parser, [get, put], namespace='db', namespace_kwargs={ 'title': 'database commands', 'help': 'CRUD for our silly database' }) add_subcommands(parser, 'db', [get, put], title='database commands', help='CRUD for our silly database')
argh/assembling.py
def add_subcommands(parser, namespace, functions, **namespace_kwargs): """ A wrapper for :func:`add_commands`. These examples are equivalent:: add_commands(parser, [get, put], namespace='db', namespace_kwargs={ 'title': 'database commands', 'help': 'CRUD for our silly database' }) add_subcommands(parser, 'db', [get, put], title='database commands', help='CRUD for our silly database') """ add_commands(parser, functions, namespace=namespace, namespace_kwargs=namespace_kwargs)
def add_subcommands(parser, namespace, functions, **namespace_kwargs): """ A wrapper for :func:`add_commands`. These examples are equivalent:: add_commands(parser, [get, put], namespace='db', namespace_kwargs={ 'title': 'database commands', 'help': 'CRUD for our silly database' }) add_subcommands(parser, 'db', [get, put], title='database commands', help='CRUD for our silly database') """ add_commands(parser, functions, namespace=namespace, namespace_kwargs=namespace_kwargs)
[ "A", "wrapper", "for", ":", "func", ":", "add_commands", "." ]
neithere/argh
python
https://github.com/neithere/argh/blob/dcd3253f2994400a6a58a700c118c53765bc50a4/argh/assembling.py#L484-L502
[ "def", "add_subcommands", "(", "parser", ",", "namespace", ",", "functions", ",", "*", "*", "namespace_kwargs", ")", ":", "add_commands", "(", "parser", ",", "functions", ",", "namespace", "=", "namespace", ",", "namespace_kwargs", "=", "namespace_kwargs", ")" ]
dcd3253f2994400a6a58a700c118c53765bc50a4
test
get_subparsers
Returns the :class:`argparse._SubParsersAction` instance for given :class:`ArgumentParser` instance as would have been returned by :meth:`ArgumentParser.add_subparsers`. The problem with the latter is that it only works once and raises an exception on the second attempt, and the public API seems to lack a method to get *existing* subparsers. :param create: If `True`, creates the subparser if it does not exist. Default if `False`.
argh/utils.py
def get_subparsers(parser, create=False): """ Returns the :class:`argparse._SubParsersAction` instance for given :class:`ArgumentParser` instance as would have been returned by :meth:`ArgumentParser.add_subparsers`. The problem with the latter is that it only works once and raises an exception on the second attempt, and the public API seems to lack a method to get *existing* subparsers. :param create: If `True`, creates the subparser if it does not exist. Default if `False`. """ # note that ArgumentParser._subparsers is *not* what is returned by # ArgumentParser.add_subparsers(). if parser._subparsers: actions = [a for a in parser._actions if isinstance(a, argparse._SubParsersAction)] assert len(actions) == 1 return actions[0] else: if create: return parser.add_subparsers()
def get_subparsers(parser, create=False): """ Returns the :class:`argparse._SubParsersAction` instance for given :class:`ArgumentParser` instance as would have been returned by :meth:`ArgumentParser.add_subparsers`. The problem with the latter is that it only works once and raises an exception on the second attempt, and the public API seems to lack a method to get *existing* subparsers. :param create: If `True`, creates the subparser if it does not exist. Default if `False`. """ # note that ArgumentParser._subparsers is *not* what is returned by # ArgumentParser.add_subparsers(). if parser._subparsers: actions = [a for a in parser._actions if isinstance(a, argparse._SubParsersAction)] assert len(actions) == 1 return actions[0] else: if create: return parser.add_subparsers()
[ "Returns", "the", ":", "class", ":", "argparse", ".", "_SubParsersAction", "instance", "for", "given", ":", "class", ":", "ArgumentParser", "instance", "as", "would", "have", "been", "returned", "by", ":", "meth", ":", "ArgumentParser", ".", "add_subparsers", ".", "The", "problem", "with", "the", "latter", "is", "that", "it", "only", "works", "once", "and", "raises", "an", "exception", "on", "the", "second", "attempt", "and", "the", "public", "API", "seems", "to", "lack", "a", "method", "to", "get", "*", "existing", "*", "subparsers", "." ]
neithere/argh
python
https://github.com/neithere/argh/blob/dcd3253f2994400a6a58a700c118c53765bc50a4/argh/utils.py#L21-L43
[ "def", "get_subparsers", "(", "parser", ",", "create", "=", "False", ")", ":", "# note that ArgumentParser._subparsers is *not* what is returned by", "# ArgumentParser.add_subparsers().", "if", "parser", ".", "_subparsers", ":", "actions", "=", "[", "a", "for", "a", "in", "parser", ".", "_actions", "if", "isinstance", "(", "a", ",", "argparse", ".", "_SubParsersAction", ")", "]", "assert", "len", "(", "actions", ")", "==", "1", "return", "actions", "[", "0", "]", "else", ":", "if", "create", ":", "return", "parser", ".", "add_subparsers", "(", ")" ]
dcd3253f2994400a6a58a700c118c53765bc50a4
test
get_arg_spec
Returns argument specification for given function. Omits special arguments of instance methods (`self`) and static methods (usually `cls` or something like this).
argh/utils.py
def get_arg_spec(function): """ Returns argument specification for given function. Omits special arguments of instance methods (`self`) and static methods (usually `cls` or something like this). """ while hasattr(function, '__wrapped__'): function = function.__wrapped__ spec = compat.getargspec(function) if inspect.ismethod(function): spec = spec._replace(args=spec.args[1:]) return spec
def get_arg_spec(function): """ Returns argument specification for given function. Omits special arguments of instance methods (`self`) and static methods (usually `cls` or something like this). """ while hasattr(function, '__wrapped__'): function = function.__wrapped__ spec = compat.getargspec(function) if inspect.ismethod(function): spec = spec._replace(args=spec.args[1:]) return spec
[ "Returns", "argument", "specification", "for", "given", "function", ".", "Omits", "special", "arguments", "of", "instance", "methods", "(", "self", ")", "and", "static", "methods", "(", "usually", "cls", "or", "something", "like", "this", ")", "." ]
neithere/argh
python
https://github.com/neithere/argh/blob/dcd3253f2994400a6a58a700c118c53765bc50a4/argh/utils.py#L46-L57
[ "def", "get_arg_spec", "(", "function", ")", ":", "while", "hasattr", "(", "function", ",", "'__wrapped__'", ")", ":", "function", "=", "function", ".", "__wrapped__", "spec", "=", "compat", ".", "getargspec", "(", "function", ")", "if", "inspect", ".", "ismethod", "(", "function", ")", ":", "spec", "=", "spec", ".", "_replace", "(", "args", "=", "spec", ".", "args", "[", "1", ":", "]", ")", "return", "spec" ]
dcd3253f2994400a6a58a700c118c53765bc50a4
test
named
Sets given string as command name instead of the function name. The string is used verbatim without further processing. Usage:: @named('load') def do_load_some_stuff_and_keep_the_original_function_name(args): ... The resulting command will be available only as ``load``. To add aliases without renaming the command, check :func:`aliases`. .. versionadded:: 0.19
argh/decorators.py
def named(new_name): """ Sets given string as command name instead of the function name. The string is used verbatim without further processing. Usage:: @named('load') def do_load_some_stuff_and_keep_the_original_function_name(args): ... The resulting command will be available only as ``load``. To add aliases without renaming the command, check :func:`aliases`. .. versionadded:: 0.19 """ def wrapper(func): setattr(func, ATTR_NAME, new_name) return func return wrapper
def named(new_name): """ Sets given string as command name instead of the function name. The string is used verbatim without further processing. Usage:: @named('load') def do_load_some_stuff_and_keep_the_original_function_name(args): ... The resulting command will be available only as ``load``. To add aliases without renaming the command, check :func:`aliases`. .. versionadded:: 0.19 """ def wrapper(func): setattr(func, ATTR_NAME, new_name) return func return wrapper
[ "Sets", "given", "string", "as", "command", "name", "instead", "of", "the", "function", "name", ".", "The", "string", "is", "used", "verbatim", "without", "further", "processing", "." ]
neithere/argh
python
https://github.com/neithere/argh/blob/dcd3253f2994400a6a58a700c118c53765bc50a4/argh/decorators.py#L24-L43
[ "def", "named", "(", "new_name", ")", ":", "def", "wrapper", "(", "func", ")", ":", "setattr", "(", "func", ",", "ATTR_NAME", ",", "new_name", ")", "return", "func", "return", "wrapper" ]
dcd3253f2994400a6a58a700c118c53765bc50a4
test
aliases
Defines alternative command name(s) for given function (along with its original name). Usage:: @aliases('co', 'check') def checkout(args): ... The resulting command will be available as ``checkout``, ``check`` and ``co``. .. note:: This decorator only works with a recent version of argparse (see `Python issue 9324`_ and `Python rev 4c0426`_). Such version ships with **Python 3.2+** and may be available in other environments as a separate package. Argh does not issue warnings and simply ignores aliases if they are not supported. See :attr:`~argh.assembling.SUPPORTS_ALIASES`. .. _Python issue 9324: http://bugs.python.org/issue9324 .. _Python rev 4c0426: http://hg.python.org/cpython/rev/4c0426261148/ .. versionadded:: 0.19
argh/decorators.py
def aliases(*names): """ Defines alternative command name(s) for given function (along with its original name). Usage:: @aliases('co', 'check') def checkout(args): ... The resulting command will be available as ``checkout``, ``check`` and ``co``. .. note:: This decorator only works with a recent version of argparse (see `Python issue 9324`_ and `Python rev 4c0426`_). Such version ships with **Python 3.2+** and may be available in other environments as a separate package. Argh does not issue warnings and simply ignores aliases if they are not supported. See :attr:`~argh.assembling.SUPPORTS_ALIASES`. .. _Python issue 9324: http://bugs.python.org/issue9324 .. _Python rev 4c0426: http://hg.python.org/cpython/rev/4c0426261148/ .. versionadded:: 0.19 """ def wrapper(func): setattr(func, ATTR_ALIASES, names) return func return wrapper
def aliases(*names): """ Defines alternative command name(s) for given function (along with its original name). Usage:: @aliases('co', 'check') def checkout(args): ... The resulting command will be available as ``checkout``, ``check`` and ``co``. .. note:: This decorator only works with a recent version of argparse (see `Python issue 9324`_ and `Python rev 4c0426`_). Such version ships with **Python 3.2+** and may be available in other environments as a separate package. Argh does not issue warnings and simply ignores aliases if they are not supported. See :attr:`~argh.assembling.SUPPORTS_ALIASES`. .. _Python issue 9324: http://bugs.python.org/issue9324 .. _Python rev 4c0426: http://hg.python.org/cpython/rev/4c0426261148/ .. versionadded:: 0.19 """ def wrapper(func): setattr(func, ATTR_ALIASES, names) return func return wrapper
[ "Defines", "alternative", "command", "name", "(", "s", ")", "for", "given", "function", "(", "along", "with", "its", "original", "name", ")", ".", "Usage", "::" ]
neithere/argh
python
https://github.com/neithere/argh/blob/dcd3253f2994400a6a58a700c118c53765bc50a4/argh/decorators.py#L46-L73
[ "def", "aliases", "(", "*", "names", ")", ":", "def", "wrapper", "(", "func", ")", ":", "setattr", "(", "func", ",", "ATTR_ALIASES", ",", "names", ")", "return", "func", "return", "wrapper" ]
dcd3253f2994400a6a58a700c118c53765bc50a4
test
arg
Declares an argument for given function. Does not register the function anywhere, nor does it modify the function in any way. The signature of the decorator matches that of :meth:`argparse.ArgumentParser.add_argument`, only some keywords are not required if they can be easily guessed (e.g. you don't have to specify type or action when an `int` or `bool` default value is supplied). Typical use cases: - In combination with :func:`expects_obj` (which is not recommended); - in combination with ordinary function signatures to add details that cannot be expressed with that syntax (e.g. help message). Usage:: from argh import arg @arg('path', help='path to the file to load') @arg('--format', choices=['yaml','json']) @arg('-v', '--verbosity', choices=range(0,3), default=2) def load(path, something=None, format='json', dry_run=False, verbosity=1): loaders = {'json': json.load, 'yaml': yaml.load} loader = loaders[args.format] data = loader(args.path) if not args.dry_run: if verbosity < 1: print('saving to the database') put_to_database(data) In this example: - `path` declaration is extended with `help`; - `format` declaration is extended with `choices`; - `dry_run` declaration is not duplicated; - `verbosity` is extended with `choices` and the default value is overridden. (If both function signature and `@arg` define a default value for an argument, `@arg` wins.) .. note:: It is recommended to avoid using this decorator unless there's no way to tune the argument's behaviour or presentation using ordinary function signatures. Readability counts, don't repeat yourself.
argh/decorators.py
def arg(*args, **kwargs): """ Declares an argument for given function. Does not register the function anywhere, nor does it modify the function in any way. The signature of the decorator matches that of :meth:`argparse.ArgumentParser.add_argument`, only some keywords are not required if they can be easily guessed (e.g. you don't have to specify type or action when an `int` or `bool` default value is supplied). Typical use cases: - In combination with :func:`expects_obj` (which is not recommended); - in combination with ordinary function signatures to add details that cannot be expressed with that syntax (e.g. help message). Usage:: from argh import arg @arg('path', help='path to the file to load') @arg('--format', choices=['yaml','json']) @arg('-v', '--verbosity', choices=range(0,3), default=2) def load(path, something=None, format='json', dry_run=False, verbosity=1): loaders = {'json': json.load, 'yaml': yaml.load} loader = loaders[args.format] data = loader(args.path) if not args.dry_run: if verbosity < 1: print('saving to the database') put_to_database(data) In this example: - `path` declaration is extended with `help`; - `format` declaration is extended with `choices`; - `dry_run` declaration is not duplicated; - `verbosity` is extended with `choices` and the default value is overridden. (If both function signature and `@arg` define a default value for an argument, `@arg` wins.) .. note:: It is recommended to avoid using this decorator unless there's no way to tune the argument's behaviour or presentation using ordinary function signatures. Readability counts, don't repeat yourself. """ def wrapper(func): declared_args = getattr(func, ATTR_ARGS, []) # The innermost decorator is called first but appears last in the code. # We need to preserve the expected order of positional arguments, so # the outermost decorator inserts its value before the innermost's: declared_args.insert(0, dict(option_strings=args, **kwargs)) setattr(func, ATTR_ARGS, declared_args) return func return wrapper
def arg(*args, **kwargs): """ Declares an argument for given function. Does not register the function anywhere, nor does it modify the function in any way. The signature of the decorator matches that of :meth:`argparse.ArgumentParser.add_argument`, only some keywords are not required if they can be easily guessed (e.g. you don't have to specify type or action when an `int` or `bool` default value is supplied). Typical use cases: - In combination with :func:`expects_obj` (which is not recommended); - in combination with ordinary function signatures to add details that cannot be expressed with that syntax (e.g. help message). Usage:: from argh import arg @arg('path', help='path to the file to load') @arg('--format', choices=['yaml','json']) @arg('-v', '--verbosity', choices=range(0,3), default=2) def load(path, something=None, format='json', dry_run=False, verbosity=1): loaders = {'json': json.load, 'yaml': yaml.load} loader = loaders[args.format] data = loader(args.path) if not args.dry_run: if verbosity < 1: print('saving to the database') put_to_database(data) In this example: - `path` declaration is extended with `help`; - `format` declaration is extended with `choices`; - `dry_run` declaration is not duplicated; - `verbosity` is extended with `choices` and the default value is overridden. (If both function signature and `@arg` define a default value for an argument, `@arg` wins.) .. note:: It is recommended to avoid using this decorator unless there's no way to tune the argument's behaviour or presentation using ordinary function signatures. Readability counts, don't repeat yourself. """ def wrapper(func): declared_args = getattr(func, ATTR_ARGS, []) # The innermost decorator is called first but appears last in the code. # We need to preserve the expected order of positional arguments, so # the outermost decorator inserts its value before the innermost's: declared_args.insert(0, dict(option_strings=args, **kwargs)) setattr(func, ATTR_ARGS, declared_args) return func return wrapper
[ "Declares", "an", "argument", "for", "given", "function", ".", "Does", "not", "register", "the", "function", "anywhere", "nor", "does", "it", "modify", "the", "function", "in", "any", "way", "." ]
neithere/argh
python
https://github.com/neithere/argh/blob/dcd3253f2994400a6a58a700c118c53765bc50a4/argh/decorators.py#L76-L132
[ "def", "arg", "(", "*", "args", ",", "*", "*", "kwargs", ")", ":", "def", "wrapper", "(", "func", ")", ":", "declared_args", "=", "getattr", "(", "func", ",", "ATTR_ARGS", ",", "[", "]", ")", "# The innermost decorator is called first but appears last in the code.", "# We need to preserve the expected order of positional arguments, so", "# the outermost decorator inserts its value before the innermost's:", "declared_args", ".", "insert", "(", "0", ",", "dict", "(", "option_strings", "=", "args", ",", "*", "*", "kwargs", ")", ")", "setattr", "(", "func", ",", "ATTR_ARGS", ",", "declared_args", ")", "return", "func", "return", "wrapper" ]
dcd3253f2994400a6a58a700c118c53765bc50a4
test
wrap_errors
Decorator. Wraps given exceptions into :class:`~argh.exceptions.CommandError`. Usage:: @wrap_errors([AssertionError]) def foo(x=None, y=None): assert x or y, 'x or y must be specified' If the assertion fails, its message will be correctly printed and the stack hidden. This helps to avoid boilerplate code. :param errors: A list of exception classes to catch. :param processor: A callable that expects the exception object and returns a string. For example, this renders all wrapped errors in red colour:: from termcolor import colored def failure(err): return colored(str(err), 'red') @wrap_errors(processor=failure) def my_command(...): ...
argh/decorators.py
def wrap_errors(errors=None, processor=None, *args): """ Decorator. Wraps given exceptions into :class:`~argh.exceptions.CommandError`. Usage:: @wrap_errors([AssertionError]) def foo(x=None, y=None): assert x or y, 'x or y must be specified' If the assertion fails, its message will be correctly printed and the stack hidden. This helps to avoid boilerplate code. :param errors: A list of exception classes to catch. :param processor: A callable that expects the exception object and returns a string. For example, this renders all wrapped errors in red colour:: from termcolor import colored def failure(err): return colored(str(err), 'red') @wrap_errors(processor=failure) def my_command(...): ... """ def wrapper(func): if errors: setattr(func, ATTR_WRAPPED_EXCEPTIONS, errors) if processor: setattr(func, ATTR_WRAPPED_EXCEPTIONS_PROCESSOR, processor) return func return wrapper
def wrap_errors(errors=None, processor=None, *args): """ Decorator. Wraps given exceptions into :class:`~argh.exceptions.CommandError`. Usage:: @wrap_errors([AssertionError]) def foo(x=None, y=None): assert x or y, 'x or y must be specified' If the assertion fails, its message will be correctly printed and the stack hidden. This helps to avoid boilerplate code. :param errors: A list of exception classes to catch. :param processor: A callable that expects the exception object and returns a string. For example, this renders all wrapped errors in red colour:: from termcolor import colored def failure(err): return colored(str(err), 'red') @wrap_errors(processor=failure) def my_command(...): ... """ def wrapper(func): if errors: setattr(func, ATTR_WRAPPED_EXCEPTIONS, errors) if processor: setattr(func, ATTR_WRAPPED_EXCEPTIONS_PROCESSOR, processor) return func return wrapper
[ "Decorator", ".", "Wraps", "given", "exceptions", "into", ":", "class", ":", "~argh", ".", "exceptions", ".", "CommandError", ".", "Usage", "::" ]
neithere/argh
python
https://github.com/neithere/argh/blob/dcd3253f2994400a6a58a700c118c53765bc50a4/argh/decorators.py#L135-L172
[ "def", "wrap_errors", "(", "errors", "=", "None", ",", "processor", "=", "None", ",", "*", "args", ")", ":", "def", "wrapper", "(", "func", ")", ":", "if", "errors", ":", "setattr", "(", "func", ",", "ATTR_WRAPPED_EXCEPTIONS", ",", "errors", ")", "if", "processor", ":", "setattr", "(", "func", ",", "ATTR_WRAPPED_EXCEPTIONS_PROCESSOR", ",", "processor", ")", "return", "func", "return", "wrapper" ]
dcd3253f2994400a6a58a700c118c53765bc50a4
test
confirm
A shortcut for typical confirmation prompt. :param action: a string describing the action, e.g. "Apply changes". A question mark will be appended. :param default: `bool` or `None`. Determines what happens when user hits :kbd:`Enter` without typing in a choice. If `True`, default choice is "yes". If `False`, it is "no". If `None` the prompt keeps reappearing until user types in a choice (not necessarily acceptable) or until the number of iteration reaches the limit. Default is `None`. :param skip: `bool`; if `True`, no interactive prompt is used and default choice is returned (useful for batch mode). Default is `False`. Usage:: def delete(key, silent=False): item = db.get(Item, args.key) if confirm('Delete '+item.title, default=True, skip=silent): item.delete() print('Item deleted.') else: print('Operation cancelled.') Returns `None` on `KeyboardInterrupt` event.
argh/interaction.py
def confirm(action, default=None, skip=False): """ A shortcut for typical confirmation prompt. :param action: a string describing the action, e.g. "Apply changes". A question mark will be appended. :param default: `bool` or `None`. Determines what happens when user hits :kbd:`Enter` without typing in a choice. If `True`, default choice is "yes". If `False`, it is "no". If `None` the prompt keeps reappearing until user types in a choice (not necessarily acceptable) or until the number of iteration reaches the limit. Default is `None`. :param skip: `bool`; if `True`, no interactive prompt is used and default choice is returned (useful for batch mode). Default is `False`. Usage:: def delete(key, silent=False): item = db.get(Item, args.key) if confirm('Delete '+item.title, default=True, skip=silent): item.delete() print('Item deleted.') else: print('Operation cancelled.') Returns `None` on `KeyboardInterrupt` event. """ MAX_ITERATIONS = 3 if skip: return default else: defaults = { None: ('y','n'), True: ('Y','n'), False: ('y','N'), } y, n = defaults[default] prompt = text_type('{action}? ({y}/{n})').format(**locals()) choice = None try: if default is None: cnt = 1 while not choice and cnt < MAX_ITERATIONS: choice = safe_input(prompt) cnt += 1 else: choice = safe_input(prompt) except KeyboardInterrupt: return None if choice in ('yes', 'y', 'Y'): return True if choice in ('no', 'n', 'N'): return False if default is not None: return default return None
def confirm(action, default=None, skip=False): """ A shortcut for typical confirmation prompt. :param action: a string describing the action, e.g. "Apply changes". A question mark will be appended. :param default: `bool` or `None`. Determines what happens when user hits :kbd:`Enter` without typing in a choice. If `True`, default choice is "yes". If `False`, it is "no". If `None` the prompt keeps reappearing until user types in a choice (not necessarily acceptable) or until the number of iteration reaches the limit. Default is `None`. :param skip: `bool`; if `True`, no interactive prompt is used and default choice is returned (useful for batch mode). Default is `False`. Usage:: def delete(key, silent=False): item = db.get(Item, args.key) if confirm('Delete '+item.title, default=True, skip=silent): item.delete() print('Item deleted.') else: print('Operation cancelled.') Returns `None` on `KeyboardInterrupt` event. """ MAX_ITERATIONS = 3 if skip: return default else: defaults = { None: ('y','n'), True: ('Y','n'), False: ('y','N'), } y, n = defaults[default] prompt = text_type('{action}? ({y}/{n})').format(**locals()) choice = None try: if default is None: cnt = 1 while not choice and cnt < MAX_ITERATIONS: choice = safe_input(prompt) cnt += 1 else: choice = safe_input(prompt) except KeyboardInterrupt: return None if choice in ('yes', 'y', 'Y'): return True if choice in ('no', 'n', 'N'): return False if default is not None: return default return None
[ "A", "shortcut", "for", "typical", "confirmation", "prompt", "." ]
neithere/argh
python
https://github.com/neithere/argh/blob/dcd3253f2994400a6a58a700c118c53765bc50a4/argh/interaction.py#L22-L84
[ "def", "confirm", "(", "action", ",", "default", "=", "None", ",", "skip", "=", "False", ")", ":", "MAX_ITERATIONS", "=", "3", "if", "skip", ":", "return", "default", "else", ":", "defaults", "=", "{", "None", ":", "(", "'y'", ",", "'n'", ")", ",", "True", ":", "(", "'Y'", ",", "'n'", ")", ",", "False", ":", "(", "'y'", ",", "'N'", ")", ",", "}", "y", ",", "n", "=", "defaults", "[", "default", "]", "prompt", "=", "text_type", "(", "'{action}? ({y}/{n})'", ")", ".", "format", "(", "*", "*", "locals", "(", ")", ")", "choice", "=", "None", "try", ":", "if", "default", "is", "None", ":", "cnt", "=", "1", "while", "not", "choice", "and", "cnt", "<", "MAX_ITERATIONS", ":", "choice", "=", "safe_input", "(", "prompt", ")", "cnt", "+=", "1", "else", ":", "choice", "=", "safe_input", "(", "prompt", ")", "except", "KeyboardInterrupt", ":", "return", "None", "if", "choice", "in", "(", "'yes'", ",", "'y'", ",", "'Y'", ")", ":", "return", "True", "if", "choice", "in", "(", "'no'", ",", "'n'", ",", "'N'", ")", ":", "return", "False", "if", "default", "is", "not", "None", ":", "return", "default", "return", "None" ]
dcd3253f2994400a6a58a700c118c53765bc50a4
test
Query.select
Select the provided column names from the model, do not return an entity, do not involve the rom session, just get the raw and/or processed column data from Redis. Keyword-only arguments: * *include_pk=False* - whether to include the primary key in the returned data (we need to get this in some cases, so we fetch it anyway; if you want it, we can return it to you - just be careful with the namedtuple option - see the warning below) * *decode=True* - whether to take a pass through normal data decoding in the model (will not return an entity/model) * *ff=_dict_data_factory* - the type of data to return from the select after all filters/limits/order_by are applied .. warning:: If ``include_pk = True`` and if you don't provide the primary key column, it will be appended to your list of columns. .. note:: if you want to provide a new factory function for the returned data, it must be of the form (below is the actual dict factory function) :: def _dict_data_factory(columns): _dict = dict _zip = zip def make(data): # do whatever you need to turn your tuple of columns plus # your list of data into whatever you want: return _dict(_zip(columns, data)) return make Available factory functions: * *``rom.query._dict_data_factory``* - default * *``rom.query._list_data_factory``* - lowest overhead, as the ``data`` passed in above is a list that you can do anything to * *``rom.query._tuple_data_factory``* - when you want tuples instead * *``rom.query._namedtuple_data_factory``* - get namedtuples, see see warning below .. warning:: If you use the ``_namedtuple_data_factory``, and your columns include underscore prefixes, they will be stripped. If this results in a name collision, you *will* get an exception. If you want differerent behavior, write your own 20 line factory function that does exactly what you want, and pass it; they are really easy!
rom/query.py
def select(self, *column_names, **kwargs): ''' Select the provided column names from the model, do not return an entity, do not involve the rom session, just get the raw and/or processed column data from Redis. Keyword-only arguments: * *include_pk=False* - whether to include the primary key in the returned data (we need to get this in some cases, so we fetch it anyway; if you want it, we can return it to you - just be careful with the namedtuple option - see the warning below) * *decode=True* - whether to take a pass through normal data decoding in the model (will not return an entity/model) * *ff=_dict_data_factory* - the type of data to return from the select after all filters/limits/order_by are applied .. warning:: If ``include_pk = True`` and if you don't provide the primary key column, it will be appended to your list of columns. .. note:: if you want to provide a new factory function for the returned data, it must be of the form (below is the actual dict factory function) :: def _dict_data_factory(columns): _dict = dict _zip = zip def make(data): # do whatever you need to turn your tuple of columns plus # your list of data into whatever you want: return _dict(_zip(columns, data)) return make Available factory functions: * *``rom.query._dict_data_factory``* - default * *``rom.query._list_data_factory``* - lowest overhead, as the ``data`` passed in above is a list that you can do anything to * *``rom.query._tuple_data_factory``* - when you want tuples instead * *``rom.query._namedtuple_data_factory``* - get namedtuples, see see warning below .. warning:: If you use the ``_namedtuple_data_factory``, and your columns include underscore prefixes, they will be stripped. If this results in a name collision, you *will* get an exception. If you want differerent behavior, write your own 20 line factory function that does exactly what you want, and pass it; they are really easy! ''' include_pk = kwargs.pop('include_pk', False) decode = kwargs.pop('decode', True) ff = kwargs.pop('ff', _dict_data_factory) if isinstance(column_names[0], (list, tuple)): column_names = column_names[0] if not column_names: raise QueryError("Must provide at least one column to query for raw data") if len(set(column_names)) != len(column_names): raise QueryError("Column names must be unique") missing = [c for c in column_names if c not in self._model._columns] if missing: raise QueryError("No such columns known: %r"%(missing,)) remove_last = False if self._model._pkey not in column_names: column_names += (self._model._pkey,) remove_last = not include_pk return self.replace(select=(column_names, decode, remove_last, ff))
def select(self, *column_names, **kwargs): ''' Select the provided column names from the model, do not return an entity, do not involve the rom session, just get the raw and/or processed column data from Redis. Keyword-only arguments: * *include_pk=False* - whether to include the primary key in the returned data (we need to get this in some cases, so we fetch it anyway; if you want it, we can return it to you - just be careful with the namedtuple option - see the warning below) * *decode=True* - whether to take a pass through normal data decoding in the model (will not return an entity/model) * *ff=_dict_data_factory* - the type of data to return from the select after all filters/limits/order_by are applied .. warning:: If ``include_pk = True`` and if you don't provide the primary key column, it will be appended to your list of columns. .. note:: if you want to provide a new factory function for the returned data, it must be of the form (below is the actual dict factory function) :: def _dict_data_factory(columns): _dict = dict _zip = zip def make(data): # do whatever you need to turn your tuple of columns plus # your list of data into whatever you want: return _dict(_zip(columns, data)) return make Available factory functions: * *``rom.query._dict_data_factory``* - default * *``rom.query._list_data_factory``* - lowest overhead, as the ``data`` passed in above is a list that you can do anything to * *``rom.query._tuple_data_factory``* - when you want tuples instead * *``rom.query._namedtuple_data_factory``* - get namedtuples, see see warning below .. warning:: If you use the ``_namedtuple_data_factory``, and your columns include underscore prefixes, they will be stripped. If this results in a name collision, you *will* get an exception. If you want differerent behavior, write your own 20 line factory function that does exactly what you want, and pass it; they are really easy! ''' include_pk = kwargs.pop('include_pk', False) decode = kwargs.pop('decode', True) ff = kwargs.pop('ff', _dict_data_factory) if isinstance(column_names[0], (list, tuple)): column_names = column_names[0] if not column_names: raise QueryError("Must provide at least one column to query for raw data") if len(set(column_names)) != len(column_names): raise QueryError("Column names must be unique") missing = [c for c in column_names if c not in self._model._columns] if missing: raise QueryError("No such columns known: %r"%(missing,)) remove_last = False if self._model._pkey not in column_names: column_names += (self._model._pkey,) remove_last = not include_pk return self.replace(select=(column_names, decode, remove_last, ff))
[ "Select", "the", "provided", "column", "names", "from", "the", "model", "do", "not", "return", "an", "entity", "do", "not", "involve", "the", "rom", "session", "just", "get", "the", "raw", "and", "/", "or", "processed", "column", "data", "from", "Redis", "." ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/query.py#L107-L180
[ "def", "select", "(", "self", ",", "*", "column_names", ",", "*", "*", "kwargs", ")", ":", "include_pk", "=", "kwargs", ".", "pop", "(", "'include_pk'", ",", "False", ")", "decode", "=", "kwargs", ".", "pop", "(", "'decode'", ",", "True", ")", "ff", "=", "kwargs", ".", "pop", "(", "'ff'", ",", "_dict_data_factory", ")", "if", "isinstance", "(", "column_names", "[", "0", "]", ",", "(", "list", ",", "tuple", ")", ")", ":", "column_names", "=", "column_names", "[", "0", "]", "if", "not", "column_names", ":", "raise", "QueryError", "(", "\"Must provide at least one column to query for raw data\"", ")", "if", "len", "(", "set", "(", "column_names", ")", ")", "!=", "len", "(", "column_names", ")", ":", "raise", "QueryError", "(", "\"Column names must be unique\"", ")", "missing", "=", "[", "c", "for", "c", "in", "column_names", "if", "c", "not", "in", "self", ".", "_model", ".", "_columns", "]", "if", "missing", ":", "raise", "QueryError", "(", "\"No such columns known: %r\"", "%", "(", "missing", ",", ")", ")", "remove_last", "=", "False", "if", "self", ".", "_model", ".", "_pkey", "not", "in", "column_names", ":", "column_names", "+=", "(", "self", ".", "_model", ".", "_pkey", ",", ")", "remove_last", "=", "not", "include_pk", "return", "self", ".", "replace", "(", "select", "=", "(", "column_names", ",", "decode", ",", "remove_last", ",", "ff", ")", ")" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
Query.replace
Copy the Query object, optionally replacing the filters, order_by, or limit information on the copy. This is mostly an internal detail that you can ignore.
rom/query.py
def replace(self, **kwargs): ''' Copy the Query object, optionally replacing the filters, order_by, or limit information on the copy. This is mostly an internal detail that you can ignore. ''' data = { 'model': self._model, 'filters': self._filters, 'order_by': self._order_by, 'limit': self._limit, 'select': self._select, } data.update(**kwargs) return Query(**data)
def replace(self, **kwargs): ''' Copy the Query object, optionally replacing the filters, order_by, or limit information on the copy. This is mostly an internal detail that you can ignore. ''' data = { 'model': self._model, 'filters': self._filters, 'order_by': self._order_by, 'limit': self._limit, 'select': self._select, } data.update(**kwargs) return Query(**data)
[ "Copy", "the", "Query", "object", "optionally", "replacing", "the", "filters", "order_by", "or", "limit", "information", "on", "the", "copy", ".", "This", "is", "mostly", "an", "internal", "detail", "that", "you", "can", "ignore", "." ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/query.py#L182-L196
[ "def", "replace", "(", "self", ",", "*", "*", "kwargs", ")", ":", "data", "=", "{", "'model'", ":", "self", ".", "_model", ",", "'filters'", ":", "self", ".", "_filters", ",", "'order_by'", ":", "self", ".", "_order_by", ",", "'limit'", ":", "self", ".", "_limit", ",", "'select'", ":", "self", ".", "_select", ",", "}", "data", ".", "update", "(", "*", "*", "kwargs", ")", "return", "Query", "(", "*", "*", "data", ")" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
Query.filter
Only columns/attributes that have been specified as having an index with the ``index=True`` option on the column definition can be filtered with this method. Prefix, suffix, and pattern match filters must be provided using the ``.startswith()``, ``.endswith()``, and the ``.like()`` methods on the query object, respectively. Geo location queries should be performed using the ``.near()`` method. Filters should be of the form:: # for numeric ranges, use None for open-ended ranges attribute=(min, max) # you can also query for equality by passing a single number attribute=value # for string searches, passing a plain string will require that # string to be in the index as a literal attribute=string # to perform an 'or' query on strings, you can pass a list of # strings attribute=[string1, string2] As an example, the following will return entities that have both ``hello`` and ``world`` in the ``String`` column ``scol`` and has a ``Numeric`` column ``ncol`` with value between 2 and 10 (including the endpoints):: results = MyModel.query \\ .filter(scol='hello') \\ .filter(scol='world') \\ .filter(ncol=(2, 10)) \\ .all() If you only want to match a single value as part of your range query, you can pass an integer, float, or Decimal object by itself, similar to the ``Model.get_by()`` method:: results = MyModel.query \\ .filter(ncol=5) \\ .execute() .. note:: Trying to use a range query `attribute=(min, max)` on indexed string columns won't return any results. .. note:: This method only filters columns that have been defined with ``index=True``.
rom/query.py
def filter(self, **kwargs): ''' Only columns/attributes that have been specified as having an index with the ``index=True`` option on the column definition can be filtered with this method. Prefix, suffix, and pattern match filters must be provided using the ``.startswith()``, ``.endswith()``, and the ``.like()`` methods on the query object, respectively. Geo location queries should be performed using the ``.near()`` method. Filters should be of the form:: # for numeric ranges, use None for open-ended ranges attribute=(min, max) # you can also query for equality by passing a single number attribute=value # for string searches, passing a plain string will require that # string to be in the index as a literal attribute=string # to perform an 'or' query on strings, you can pass a list of # strings attribute=[string1, string2] As an example, the following will return entities that have both ``hello`` and ``world`` in the ``String`` column ``scol`` and has a ``Numeric`` column ``ncol`` with value between 2 and 10 (including the endpoints):: results = MyModel.query \\ .filter(scol='hello') \\ .filter(scol='world') \\ .filter(ncol=(2, 10)) \\ .all() If you only want to match a single value as part of your range query, you can pass an integer, float, or Decimal object by itself, similar to the ``Model.get_by()`` method:: results = MyModel.query \\ .filter(ncol=5) \\ .execute() .. note:: Trying to use a range query `attribute=(min, max)` on indexed string columns won't return any results. .. note:: This method only filters columns that have been defined with ``index=True``. ''' cur_filters = list(self._filters) for attr, value in kwargs.items(): value = self._check(attr, value, which='filter') if isinstance(value, NUMERIC_TYPES): # for simple numeric equiality filters value = (value, value) if isinstance(value, six.string_types): cur_filters.append('%s:%s'%(attr, value)) elif six.PY3 and isinstance(value, bytes): cur_filters.append('%s:%s'%(attr, value.decode('latin-1'))) elif isinstance(value, tuple): if value is NOT_NULL: from .columns import OneToOne, ManyToOne ctype = type(self._model._columns[attr]) if not issubclass(ctype, (OneToOne, ManyToOne)): raise QueryError("Can only query for non-null column values " \ "on OneToOne or ManyToOne columns, %r is of type %r"%(attr, ctype)) if len(value) != 2: raise QueryError("Numeric ranges require 2 endpoints, you provided %s with %r"%(len(value), value)) tt = [] for v in value: if isinstance(v, date): v = dt2ts(v) if isinstance(v, dtime): v = t2ts(v) tt.append(v) value = tt cur_filters.append((attr, value[0], value[1])) elif isinstance(value, list) and value: cur_filters.append(['%s:%s'%(attr, _ts(v)) for v in value]) else: raise QueryError("Sorry, we don't know how to filter %r by %r"%(attr, value)) return self.replace(filters=tuple(cur_filters))
def filter(self, **kwargs): ''' Only columns/attributes that have been specified as having an index with the ``index=True`` option on the column definition can be filtered with this method. Prefix, suffix, and pattern match filters must be provided using the ``.startswith()``, ``.endswith()``, and the ``.like()`` methods on the query object, respectively. Geo location queries should be performed using the ``.near()`` method. Filters should be of the form:: # for numeric ranges, use None for open-ended ranges attribute=(min, max) # you can also query for equality by passing a single number attribute=value # for string searches, passing a plain string will require that # string to be in the index as a literal attribute=string # to perform an 'or' query on strings, you can pass a list of # strings attribute=[string1, string2] As an example, the following will return entities that have both ``hello`` and ``world`` in the ``String`` column ``scol`` and has a ``Numeric`` column ``ncol`` with value between 2 and 10 (including the endpoints):: results = MyModel.query \\ .filter(scol='hello') \\ .filter(scol='world') \\ .filter(ncol=(2, 10)) \\ .all() If you only want to match a single value as part of your range query, you can pass an integer, float, or Decimal object by itself, similar to the ``Model.get_by()`` method:: results = MyModel.query \\ .filter(ncol=5) \\ .execute() .. note:: Trying to use a range query `attribute=(min, max)` on indexed string columns won't return any results. .. note:: This method only filters columns that have been defined with ``index=True``. ''' cur_filters = list(self._filters) for attr, value in kwargs.items(): value = self._check(attr, value, which='filter') if isinstance(value, NUMERIC_TYPES): # for simple numeric equiality filters value = (value, value) if isinstance(value, six.string_types): cur_filters.append('%s:%s'%(attr, value)) elif six.PY3 and isinstance(value, bytes): cur_filters.append('%s:%s'%(attr, value.decode('latin-1'))) elif isinstance(value, tuple): if value is NOT_NULL: from .columns import OneToOne, ManyToOne ctype = type(self._model._columns[attr]) if not issubclass(ctype, (OneToOne, ManyToOne)): raise QueryError("Can only query for non-null column values " \ "on OneToOne or ManyToOne columns, %r is of type %r"%(attr, ctype)) if len(value) != 2: raise QueryError("Numeric ranges require 2 endpoints, you provided %s with %r"%(len(value), value)) tt = [] for v in value: if isinstance(v, date): v = dt2ts(v) if isinstance(v, dtime): v = t2ts(v) tt.append(v) value = tt cur_filters.append((attr, value[0], value[1])) elif isinstance(value, list) and value: cur_filters.append(['%s:%s'%(attr, _ts(v)) for v in value]) else: raise QueryError("Sorry, we don't know how to filter %r by %r"%(attr, value)) return self.replace(filters=tuple(cur_filters))
[ "Only", "columns", "/", "attributes", "that", "have", "been", "specified", "as", "having", "an", "index", "with", "the", "index", "=", "True", "option", "on", "the", "column", "definition", "can", "be", "filtered", "with", "this", "method", ".", "Prefix", "suffix", "and", "pattern", "match", "filters", "must", "be", "provided", "using", "the", ".", "startswith", "()", ".", "endswith", "()", "and", "the", ".", "like", "()", "methods", "on", "the", "query", "object", "respectively", ".", "Geo", "location", "queries", "should", "be", "performed", "using", "the", ".", "near", "()", "method", "." ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/query.py#L198-L291
[ "def", "filter", "(", "self", ",", "*", "*", "kwargs", ")", ":", "cur_filters", "=", "list", "(", "self", ".", "_filters", ")", "for", "attr", ",", "value", "in", "kwargs", ".", "items", "(", ")", ":", "value", "=", "self", ".", "_check", "(", "attr", ",", "value", ",", "which", "=", "'filter'", ")", "if", "isinstance", "(", "value", ",", "NUMERIC_TYPES", ")", ":", "# for simple numeric equiality filters", "value", "=", "(", "value", ",", "value", ")", "if", "isinstance", "(", "value", ",", "six", ".", "string_types", ")", ":", "cur_filters", ".", "append", "(", "'%s:%s'", "%", "(", "attr", ",", "value", ")", ")", "elif", "six", ".", "PY3", "and", "isinstance", "(", "value", ",", "bytes", ")", ":", "cur_filters", ".", "append", "(", "'%s:%s'", "%", "(", "attr", ",", "value", ".", "decode", "(", "'latin-1'", ")", ")", ")", "elif", "isinstance", "(", "value", ",", "tuple", ")", ":", "if", "value", "is", "NOT_NULL", ":", "from", ".", "columns", "import", "OneToOne", ",", "ManyToOne", "ctype", "=", "type", "(", "self", ".", "_model", ".", "_columns", "[", "attr", "]", ")", "if", "not", "issubclass", "(", "ctype", ",", "(", "OneToOne", ",", "ManyToOne", ")", ")", ":", "raise", "QueryError", "(", "\"Can only query for non-null column values \"", "\"on OneToOne or ManyToOne columns, %r is of type %r\"", "%", "(", "attr", ",", "ctype", ")", ")", "if", "len", "(", "value", ")", "!=", "2", ":", "raise", "QueryError", "(", "\"Numeric ranges require 2 endpoints, you provided %s with %r\"", "%", "(", "len", "(", "value", ")", ",", "value", ")", ")", "tt", "=", "[", "]", "for", "v", "in", "value", ":", "if", "isinstance", "(", "v", ",", "date", ")", ":", "v", "=", "dt2ts", "(", "v", ")", "if", "isinstance", "(", "v", ",", "dtime", ")", ":", "v", "=", "t2ts", "(", "v", ")", "tt", ".", "append", "(", "v", ")", "value", "=", "tt", "cur_filters", ".", "append", "(", "(", "attr", ",", "value", "[", "0", "]", ",", "value", "[", "1", "]", ")", ")", "elif", "isinstance", "(", "value", ",", "list", ")", "and", "value", ":", "cur_filters", ".", "append", "(", "[", "'%s:%s'", "%", "(", "attr", ",", "_ts", "(", "v", ")", ")", "for", "v", "in", "value", "]", ")", "else", ":", "raise", "QueryError", "(", "\"Sorry, we don't know how to filter %r by %r\"", "%", "(", "attr", ",", "value", ")", ")", "return", "self", ".", "replace", "(", "filters", "=", "tuple", "(", "cur_filters", ")", ")" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
Query.startswith
When provided with keyword arguments of the form ``col=prefix``, this will limit the entities returned to those that have a word with the provided prefix in the specified column(s). This requires that the ``prefix=True`` option was provided during column definition. Usage:: User.query.startswith(email='user@').execute()
rom/query.py
def startswith(self, **kwargs): ''' When provided with keyword arguments of the form ``col=prefix``, this will limit the entities returned to those that have a word with the provided prefix in the specified column(s). This requires that the ``prefix=True`` option was provided during column definition. Usage:: User.query.startswith(email='user@').execute() ''' new = [] for k, v in kwargs.items(): v = self._check(k, v, 'startswith') new.append(Prefix(k, v)) return self.replace(filters=self._filters+tuple(new))
def startswith(self, **kwargs): ''' When provided with keyword arguments of the form ``col=prefix``, this will limit the entities returned to those that have a word with the provided prefix in the specified column(s). This requires that the ``prefix=True`` option was provided during column definition. Usage:: User.query.startswith(email='user@').execute() ''' new = [] for k, v in kwargs.items(): v = self._check(k, v, 'startswith') new.append(Prefix(k, v)) return self.replace(filters=self._filters+tuple(new))
[ "When", "provided", "with", "keyword", "arguments", "of", "the", "form", "col", "=", "prefix", "this", "will", "limit", "the", "entities", "returned", "to", "those", "that", "have", "a", "word", "with", "the", "provided", "prefix", "in", "the", "specified", "column", "(", "s", ")", ".", "This", "requires", "that", "the", "prefix", "=", "True", "option", "was", "provided", "during", "column", "definition", "." ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/query.py#L293-L309
[ "def", "startswith", "(", "self", ",", "*", "*", "kwargs", ")", ":", "new", "=", "[", "]", "for", "k", ",", "v", "in", "kwargs", ".", "items", "(", ")", ":", "v", "=", "self", ".", "_check", "(", "k", ",", "v", ",", "'startswith'", ")", "new", ".", "append", "(", "Prefix", "(", "k", ",", "v", ")", ")", "return", "self", ".", "replace", "(", "filters", "=", "self", ".", "_filters", "+", "tuple", "(", "new", ")", ")" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
Query.endswith
When provided with keyword arguments of the form ``col=suffix``, this will limit the entities returned to those that have a word with the provided suffix in the specified column(s). This requires that the ``suffix=True`` option was provided during column definition. Usage:: User.query.endswith(email='@gmail.com').execute()
rom/query.py
def endswith(self, **kwargs): ''' When provided with keyword arguments of the form ``col=suffix``, this will limit the entities returned to those that have a word with the provided suffix in the specified column(s). This requires that the ``suffix=True`` option was provided during column definition. Usage:: User.query.endswith(email='@gmail.com').execute() ''' new = [] for k, v in kwargs.items(): v = self._check(k, v, 'endswith') new.append(Suffix(k, v[::-1])) return self.replace(filters=self._filters+tuple(new))
def endswith(self, **kwargs): ''' When provided with keyword arguments of the form ``col=suffix``, this will limit the entities returned to those that have a word with the provided suffix in the specified column(s). This requires that the ``suffix=True`` option was provided during column definition. Usage:: User.query.endswith(email='@gmail.com').execute() ''' new = [] for k, v in kwargs.items(): v = self._check(k, v, 'endswith') new.append(Suffix(k, v[::-1])) return self.replace(filters=self._filters+tuple(new))
[ "When", "provided", "with", "keyword", "arguments", "of", "the", "form", "col", "=", "suffix", "this", "will", "limit", "the", "entities", "returned", "to", "those", "that", "have", "a", "word", "with", "the", "provided", "suffix", "in", "the", "specified", "column", "(", "s", ")", ".", "This", "requires", "that", "the", "suffix", "=", "True", "option", "was", "provided", "during", "column", "definition", "." ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/query.py#L311-L327
[ "def", "endswith", "(", "self", ",", "*", "*", "kwargs", ")", ":", "new", "=", "[", "]", "for", "k", ",", "v", "in", "kwargs", ".", "items", "(", ")", ":", "v", "=", "self", ".", "_check", "(", "k", ",", "v", ",", "'endswith'", ")", "new", ".", "append", "(", "Suffix", "(", "k", ",", "v", "[", ":", ":", "-", "1", "]", ")", ")", "return", "self", ".", "replace", "(", "filters", "=", "self", ".", "_filters", "+", "tuple", "(", "new", ")", ")" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
Query.like
When provided with keyword arguments of the form ``col=pattern``, this will limit the entities returned to those that include the provided pattern. Note that 'like' queries require that the ``prefix=True`` option must have been provided as part of the column definition. Patterns allow for 4 wildcard characters, whose semantics are as follows: * *?* - will match 0 or 1 of any character * *\** - will match 0 or more of any character * *+* - will match 1 or more of any character * *!* - will match exactly 1 of any character As an example, imagine that you have enabled the required prefix matching on your ``User.email`` column. And lets say that you want to find everyone with an email address that contains the name 'frank' before the ``@`` sign. You can use either of the following patterns to discover those users. * *\*frank\*@* * *\*frank\*@* .. note:: Like queries implicitly start at the beginning of strings checked, so if you want to match a pattern that doesn't start at the beginning of a string, you should prefix it with one of the wildcard characters (like ``*`` as we did with the 'frank' pattern).
rom/query.py
def like(self, **kwargs): ''' When provided with keyword arguments of the form ``col=pattern``, this will limit the entities returned to those that include the provided pattern. Note that 'like' queries require that the ``prefix=True`` option must have been provided as part of the column definition. Patterns allow for 4 wildcard characters, whose semantics are as follows: * *?* - will match 0 or 1 of any character * *\** - will match 0 or more of any character * *+* - will match 1 or more of any character * *!* - will match exactly 1 of any character As an example, imagine that you have enabled the required prefix matching on your ``User.email`` column. And lets say that you want to find everyone with an email address that contains the name 'frank' before the ``@`` sign. You can use either of the following patterns to discover those users. * *\*frank\*@* * *\*frank\*@* .. note:: Like queries implicitly start at the beginning of strings checked, so if you want to match a pattern that doesn't start at the beginning of a string, you should prefix it with one of the wildcard characters (like ``*`` as we did with the 'frank' pattern). ''' new = [] for k, v in kwargs.items(): v = self._check(k, v, 'like') new.append(Pattern(k, v)) return self.replace(filters=self._filters+tuple(new))
def like(self, **kwargs): ''' When provided with keyword arguments of the form ``col=pattern``, this will limit the entities returned to those that include the provided pattern. Note that 'like' queries require that the ``prefix=True`` option must have been provided as part of the column definition. Patterns allow for 4 wildcard characters, whose semantics are as follows: * *?* - will match 0 or 1 of any character * *\** - will match 0 or more of any character * *+* - will match 1 or more of any character * *!* - will match exactly 1 of any character As an example, imagine that you have enabled the required prefix matching on your ``User.email`` column. And lets say that you want to find everyone with an email address that contains the name 'frank' before the ``@`` sign. You can use either of the following patterns to discover those users. * *\*frank\*@* * *\*frank\*@* .. note:: Like queries implicitly start at the beginning of strings checked, so if you want to match a pattern that doesn't start at the beginning of a string, you should prefix it with one of the wildcard characters (like ``*`` as we did with the 'frank' pattern). ''' new = [] for k, v in kwargs.items(): v = self._check(k, v, 'like') new.append(Pattern(k, v)) return self.replace(filters=self._filters+tuple(new))
[ "When", "provided", "with", "keyword", "arguments", "of", "the", "form", "col", "=", "pattern", "this", "will", "limit", "the", "entities", "returned", "to", "those", "that", "include", "the", "provided", "pattern", ".", "Note", "that", "like", "queries", "require", "that", "the", "prefix", "=", "True", "option", "must", "have", "been", "provided", "as", "part", "of", "the", "column", "definition", "." ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/query.py#L329-L362
[ "def", "like", "(", "self", ",", "*", "*", "kwargs", ")", ":", "new", "=", "[", "]", "for", "k", ",", "v", "in", "kwargs", ".", "items", "(", ")", ":", "v", "=", "self", ".", "_check", "(", "k", ",", "v", ",", "'like'", ")", "new", ".", "append", "(", "Pattern", "(", "k", ",", "v", ")", ")", "return", "self", ".", "replace", "(", "filters", "=", "self", ".", "_filters", "+", "tuple", "(", "new", ")", ")" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
Query.order_by
When provided with a column name, will sort the results of your query:: # returns all users, ordered by the created_at column in # descending order User.query.order_by('-created_at').execute()
rom/query.py
def order_by(self, column): ''' When provided with a column name, will sort the results of your query:: # returns all users, ordered by the created_at column in # descending order User.query.order_by('-created_at').execute() ''' cname = column.lstrip('-') col = self._check(cname) if type(col).__name__ in ('String', 'Text', 'Json') and col._keygen.__name__ not in _STRING_SORT_KEYGENS: warnings.warn("You are trying to order by a non-numeric column %r. " "Unless you have provided your own keygen or are using " "one of the sortable keygens: (%s), this probably won't " "work the way you expect it to."%(cname, STRING_SORT_KEYGENS_STR), stacklevel=2) return self.replace(order_by=column)
def order_by(self, column): ''' When provided with a column name, will sort the results of your query:: # returns all users, ordered by the created_at column in # descending order User.query.order_by('-created_at').execute() ''' cname = column.lstrip('-') col = self._check(cname) if type(col).__name__ in ('String', 'Text', 'Json') and col._keygen.__name__ not in _STRING_SORT_KEYGENS: warnings.warn("You are trying to order by a non-numeric column %r. " "Unless you have provided your own keygen or are using " "one of the sortable keygens: (%s), this probably won't " "work the way you expect it to."%(cname, STRING_SORT_KEYGENS_STR), stacklevel=2) return self.replace(order_by=column)
[ "When", "provided", "with", "a", "column", "name", "will", "sort", "the", "results", "of", "your", "query", "::" ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/query.py#L374-L391
[ "def", "order_by", "(", "self", ",", "column", ")", ":", "cname", "=", "column", ".", "lstrip", "(", "'-'", ")", "col", "=", "self", ".", "_check", "(", "cname", ")", "if", "type", "(", "col", ")", ".", "__name__", "in", "(", "'String'", ",", "'Text'", ",", "'Json'", ")", "and", "col", ".", "_keygen", ".", "__name__", "not", "in", "_STRING_SORT_KEYGENS", ":", "warnings", ".", "warn", "(", "\"You are trying to order by a non-numeric column %r. \"", "\"Unless you have provided your own keygen or are using \"", "\"one of the sortable keygens: (%s), this probably won't \"", "\"work the way you expect it to.\"", "%", "(", "cname", ",", "STRING_SORT_KEYGENS_STR", ")", ",", "stacklevel", "=", "2", ")", "return", "self", ".", "replace", "(", "order_by", "=", "column", ")" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
Query.count
Will return the total count of the objects that match the specified filters.:: # counts the number of users created in the last 24 hours User.query.filter(created_at=(time.time()-86400, time.time())).count()
rom/query.py
def count(self): ''' Will return the total count of the objects that match the specified filters.:: # counts the number of users created in the last 24 hours User.query.filter(created_at=(time.time()-86400, time.time())).count() ''' filters = self._filters if self._order_by: filters += (self._order_by.lstrip('-'),) if not filters: # We can actually count entities here... size = _connect(self._model).hlen(self._model._namespace + '::') limit = self._limit or (0, 2**64) size = max(size - max(limit[0], 0), 0) return min(size, limit[1]) return self._model._gindex.count(_connect(self._model), filters)
def count(self): ''' Will return the total count of the objects that match the specified filters.:: # counts the number of users created in the last 24 hours User.query.filter(created_at=(time.time()-86400, time.time())).count() ''' filters = self._filters if self._order_by: filters += (self._order_by.lstrip('-'),) if not filters: # We can actually count entities here... size = _connect(self._model).hlen(self._model._namespace + '::') limit = self._limit or (0, 2**64) size = max(size - max(limit[0], 0), 0) return min(size, limit[1]) return self._model._gindex.count(_connect(self._model), filters)
[ "Will", "return", "the", "total", "count", "of", "the", "objects", "that", "match", "the", "specified", "filters", ".", "::" ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/query.py#L402-L420
[ "def", "count", "(", "self", ")", ":", "filters", "=", "self", ".", "_filters", "if", "self", ".", "_order_by", ":", "filters", "+=", "(", "self", ".", "_order_by", ".", "lstrip", "(", "'-'", ")", ",", ")", "if", "not", "filters", ":", "# We can actually count entities here...", "size", "=", "_connect", "(", "self", ".", "_model", ")", ".", "hlen", "(", "self", ".", "_model", ".", "_namespace", "+", "'::'", ")", "limit", "=", "self", ".", "_limit", "or", "(", "0", ",", "2", "**", "64", ")", "size", "=", "max", "(", "size", "-", "max", "(", "limit", "[", "0", "]", ",", "0", ")", ",", "0", ")", "return", "min", "(", "size", ",", "limit", "[", "1", "]", ")", "return", "self", ".", "_model", ".", "_gindex", ".", "count", "(", "_connect", "(", "self", ".", "_model", ")", ",", "filters", ")" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
Query.iter_result
Iterate over the results of your query instead of getting them all with `.all()`. Will only perform a single query. If you expect that your processing will take more than 30 seconds to process 100 items, you should pass `timeout` and `pagesize` to reflect an appropriate timeout and page size to fetch at once. Usage:: for user in User.query.endswith(email='@gmail.com').iter_result(): # do something with user ...
rom/query.py
def iter_result(self, timeout=30, pagesize=100, no_hscan=False): ''' Iterate over the results of your query instead of getting them all with `.all()`. Will only perform a single query. If you expect that your processing will take more than 30 seconds to process 100 items, you should pass `timeout` and `pagesize` to reflect an appropriate timeout and page size to fetch at once. Usage:: for user in User.query.endswith(email='@gmail.com').iter_result(): # do something with user ... ''' if not self._filters and not self._order_by: if self._model._columns[self._model._pkey]._index: return self._iter_all_pkey() conn = _connect(self._model) version = list(map(int, conn.info()['redis_version'].split('.')[:2])) if version >= [2,8] and not no_hscan: return self._iter_all_hscan() return self._iter_all() return self._iter_results(timeout, pagesize)
def iter_result(self, timeout=30, pagesize=100, no_hscan=False): ''' Iterate over the results of your query instead of getting them all with `.all()`. Will only perform a single query. If you expect that your processing will take more than 30 seconds to process 100 items, you should pass `timeout` and `pagesize` to reflect an appropriate timeout and page size to fetch at once. Usage:: for user in User.query.endswith(email='@gmail.com').iter_result(): # do something with user ... ''' if not self._filters and not self._order_by: if self._model._columns[self._model._pkey]._index: return self._iter_all_pkey() conn = _connect(self._model) version = list(map(int, conn.info()['redis_version'].split('.')[:2])) if version >= [2,8] and not no_hscan: return self._iter_all_hscan() return self._iter_all() return self._iter_results(timeout, pagesize)
[ "Iterate", "over", "the", "results", "of", "your", "query", "instead", "of", "getting", "them", "all", "with", ".", "all", "()", ".", "Will", "only", "perform", "a", "single", "query", ".", "If", "you", "expect", "that", "your", "processing", "will", "take", "more", "than", "30", "seconds", "to", "process", "100", "items", "you", "should", "pass", "timeout", "and", "pagesize", "to", "reflect", "an", "appropriate", "timeout", "and", "page", "size", "to", "fetch", "at", "once", "." ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/query.py#L429-L452
[ "def", "iter_result", "(", "self", ",", "timeout", "=", "30", ",", "pagesize", "=", "100", ",", "no_hscan", "=", "False", ")", ":", "if", "not", "self", ".", "_filters", "and", "not", "self", ".", "_order_by", ":", "if", "self", ".", "_model", ".", "_columns", "[", "self", ".", "_model", ".", "_pkey", "]", ".", "_index", ":", "return", "self", ".", "_iter_all_pkey", "(", ")", "conn", "=", "_connect", "(", "self", ".", "_model", ")", "version", "=", "list", "(", "map", "(", "int", ",", "conn", ".", "info", "(", ")", "[", "'redis_version'", "]", ".", "split", "(", "'.'", ")", "[", ":", "2", "]", ")", ")", "if", "version", ">=", "[", "2", ",", "8", "]", "and", "not", "no_hscan", ":", "return", "self", ".", "_iter_all_hscan", "(", ")", "return", "self", ".", "_iter_all", "(", ")", "return", "self", ".", "_iter_results", "(", "timeout", ",", "pagesize", ")" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
Query.cached_result
This will execute the query, returning the key where a ZSET of your results will be stored for pagination, further operations, etc. The timeout must be a positive integer number of seconds for which to set the expiration time on the key (this is to ensure that any cached query results are eventually deleted, unless you make the explicit step to use the PERSIST command). .. note:: Limit clauses are ignored and not passed. Usage:: ukey = User.query.endswith(email='@gmail.com').cached_result(30) for i in xrange(0, conn.zcard(ukey), 100): # refresh the expiration conn.expire(ukey, 30) users = User.get(conn.zrange(ukey, i, i+99)) ...
rom/query.py
def cached_result(self, timeout): ''' This will execute the query, returning the key where a ZSET of your results will be stored for pagination, further operations, etc. The timeout must be a positive integer number of seconds for which to set the expiration time on the key (this is to ensure that any cached query results are eventually deleted, unless you make the explicit step to use the PERSIST command). .. note:: Limit clauses are ignored and not passed. Usage:: ukey = User.query.endswith(email='@gmail.com').cached_result(30) for i in xrange(0, conn.zcard(ukey), 100): # refresh the expiration conn.expire(ukey, 30) users = User.get(conn.zrange(ukey, i, i+99)) ... ''' if not (self._filters or self._order_by): raise QueryError("You are missing filter or order criteria") timeout = int(timeout) if timeout < 1: raise QueryError("You must specify a timeout >= 1, you gave %r"%timeout) return self._model._gindex.search( _connect(self._model), self._filters, self._order_by, timeout=timeout)
def cached_result(self, timeout): ''' This will execute the query, returning the key where a ZSET of your results will be stored for pagination, further operations, etc. The timeout must be a positive integer number of seconds for which to set the expiration time on the key (this is to ensure that any cached query results are eventually deleted, unless you make the explicit step to use the PERSIST command). .. note:: Limit clauses are ignored and not passed. Usage:: ukey = User.query.endswith(email='@gmail.com').cached_result(30) for i in xrange(0, conn.zcard(ukey), 100): # refresh the expiration conn.expire(ukey, 30) users = User.get(conn.zrange(ukey, i, i+99)) ... ''' if not (self._filters or self._order_by): raise QueryError("You are missing filter or order criteria") timeout = int(timeout) if timeout < 1: raise QueryError("You must specify a timeout >= 1, you gave %r"%timeout) return self._model._gindex.search( _connect(self._model), self._filters, self._order_by, timeout=timeout)
[ "This", "will", "execute", "the", "query", "returning", "the", "key", "where", "a", "ZSET", "of", "your", "results", "will", "be", "stored", "for", "pagination", "further", "operations", "etc", "." ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/query.py#L648-L675
[ "def", "cached_result", "(", "self", ",", "timeout", ")", ":", "if", "not", "(", "self", ".", "_filters", "or", "self", ".", "_order_by", ")", ":", "raise", "QueryError", "(", "\"You are missing filter or order criteria\"", ")", "timeout", "=", "int", "(", "timeout", ")", "if", "timeout", "<", "1", ":", "raise", "QueryError", "(", "\"You must specify a timeout >= 1, you gave %r\"", "%", "timeout", ")", "return", "self", ".", "_model", ".", "_gindex", ".", "search", "(", "_connect", "(", "self", ".", "_model", ")", ",", "self", ".", "_filters", ",", "self", ".", "_order_by", ",", "timeout", "=", "timeout", ")" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
Query.first
Returns only the first result from the query, if any.
rom/query.py
def first(self): ''' Returns only the first result from the query, if any. ''' lim = [0, 1] if self._limit: lim[0] = self._limit[0] if not self._filters and not self._order_by: for ent in self: return ent return None ids = self.limit(*lim)._search() if ids: return self._model.get(ids[0]) return None
def first(self): ''' Returns only the first result from the query, if any. ''' lim = [0, 1] if self._limit: lim[0] = self._limit[0] if not self._filters and not self._order_by: for ent in self: return ent return None ids = self.limit(*lim)._search() if ids: return self._model.get(ids[0]) return None
[ "Returns", "only", "the", "first", "result", "from", "the", "query", "if", "any", "." ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/query.py#L691-L705
[ "def", "first", "(", "self", ")", ":", "lim", "=", "[", "0", ",", "1", "]", "if", "self", ".", "_limit", ":", "lim", "[", "0", "]", "=", "self", ".", "_limit", "[", "0", "]", "if", "not", "self", ".", "_filters", "and", "not", "self", ".", "_order_by", ":", "for", "ent", "in", "self", ":", "return", "ent", "return", "None", "ids", "=", "self", ".", "limit", "(", "*", "lim", ")", ".", "_search", "(", ")", "if", "ids", ":", "return", "self", ".", "_model", ".", "get", "(", "ids", "[", "0", "]", ")", "return", "None" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
Query.delete
Will delete the entities that match at the time the query is executed. Used like:: MyModel.query.filter(email=...).delete() MyModel.query.endswith(email='@host.com').delete() .. warning:: can't be used on models on either side of a ``OneToMany``, ``ManyToOne``, or ``OneToOne`` relationship.
rom/query.py
def delete(self, blocksize=100): ''' Will delete the entities that match at the time the query is executed. Used like:: MyModel.query.filter(email=...).delete() MyModel.query.endswith(email='@host.com').delete() .. warning:: can't be used on models on either side of a ``OneToMany``, ``ManyToOne``, or ``OneToOne`` relationship. ''' from .columns import MODELS_REFERENCED if not self._model._no_fk or self._model._namespace in MODELS_REFERENCED: raise QueryError("Can't delete entities of models with foreign key relationships") de = [] i = 0 for result in self.iter_result(pagesize=blocksize): de.append(result) i += 1 if i >= blocksize: session.delete(de) # one round-trip to delete "chunk" items del de[:] i = 0 if de: session.delete(de)
def delete(self, blocksize=100): ''' Will delete the entities that match at the time the query is executed. Used like:: MyModel.query.filter(email=...).delete() MyModel.query.endswith(email='@host.com').delete() .. warning:: can't be used on models on either side of a ``OneToMany``, ``ManyToOne``, or ``OneToOne`` relationship. ''' from .columns import MODELS_REFERENCED if not self._model._no_fk or self._model._namespace in MODELS_REFERENCED: raise QueryError("Can't delete entities of models with foreign key relationships") de = [] i = 0 for result in self.iter_result(pagesize=blocksize): de.append(result) i += 1 if i >= blocksize: session.delete(de) # one round-trip to delete "chunk" items del de[:] i = 0 if de: session.delete(de)
[ "Will", "delete", "the", "entities", "that", "match", "at", "the", "time", "the", "query", "is", "executed", "." ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/query.py#L707-L735
[ "def", "delete", "(", "self", ",", "blocksize", "=", "100", ")", ":", "from", ".", "columns", "import", "MODELS_REFERENCED", "if", "not", "self", ".", "_model", ".", "_no_fk", "or", "self", ".", "_model", ".", "_namespace", "in", "MODELS_REFERENCED", ":", "raise", "QueryError", "(", "\"Can't delete entities of models with foreign key relationships\"", ")", "de", "=", "[", "]", "i", "=", "0", "for", "result", "in", "self", ".", "iter_result", "(", "pagesize", "=", "blocksize", ")", ":", "de", ".", "append", "(", "result", ")", "i", "+=", "1", "if", "i", ">=", "blocksize", ":", "session", ".", "delete", "(", "de", ")", "# one round-trip to delete \"chunk\" items", "del", "de", "[", ":", "]", "i", "=", "0", "if", "de", ":", "session", ".", "delete", "(", "de", ")" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
_on_delete
This function handles all on_delete semantics defined on OneToMany columns. This function only exists because 'cascade' is *very* hard to get right.
rom/columns.py
def _on_delete(ent): ''' This function handles all on_delete semantics defined on OneToMany columns. This function only exists because 'cascade' is *very* hard to get right. ''' seen_d = set([ent._pk]) to_delete = [ent] seen_s = set() to_save = [] def _set_default(ent, attr, de=NULL): pk = ent._pk if pk in seen_d: # going to be deleted, don't need to modify return col = ent.__class__._columns[attr] de = de if de is not NULL else col._default if de in (None, NULL): setattr(ent, attr, None) elif callable(col._default): setattr(ent, attr, col._default()) else: setattr(ent, attr, col._default) if pk not in seen_s: seen_s.add(pk) to_save.append(ent) for self in to_delete: for tbl, attr, action in MODELS_REFERENCED.get(self._namespace, ()): if action == 'no action': continue refs = MODELS[tbl].get_by(**{attr: self.id}) if not refs: continue if action == 'restrict': # raise the exception here for a better traceback raise _restrict(self, attr, refs) elif action == 'set null': for ref in refs: _set_default(ref, attr, None) continue elif action == 'set default': for ref in refs: _set_default(ref, attr) continue # otherwise col._on_delete == 'cascade' for ent in (refs if isinstance(refs, list) else [refs]): if ent._pk not in seen_d: seen_d.add(ent._pk) to_delete.append(ent) # If we got here, then to_delete includes all items to delete. Let's delete # them! for self in to_delete: self.delete(skip_on_delete_i_really_mean_it=SKIP_ON_DELETE) for self in to_save: # Careful not to resurrect deleted entities if self._pk not in seen_d: self.save()
def _on_delete(ent): ''' This function handles all on_delete semantics defined on OneToMany columns. This function only exists because 'cascade' is *very* hard to get right. ''' seen_d = set([ent._pk]) to_delete = [ent] seen_s = set() to_save = [] def _set_default(ent, attr, de=NULL): pk = ent._pk if pk in seen_d: # going to be deleted, don't need to modify return col = ent.__class__._columns[attr] de = de if de is not NULL else col._default if de in (None, NULL): setattr(ent, attr, None) elif callable(col._default): setattr(ent, attr, col._default()) else: setattr(ent, attr, col._default) if pk not in seen_s: seen_s.add(pk) to_save.append(ent) for self in to_delete: for tbl, attr, action in MODELS_REFERENCED.get(self._namespace, ()): if action == 'no action': continue refs = MODELS[tbl].get_by(**{attr: self.id}) if not refs: continue if action == 'restrict': # raise the exception here for a better traceback raise _restrict(self, attr, refs) elif action == 'set null': for ref in refs: _set_default(ref, attr, None) continue elif action == 'set default': for ref in refs: _set_default(ref, attr) continue # otherwise col._on_delete == 'cascade' for ent in (refs if isinstance(refs, list) else [refs]): if ent._pk not in seen_d: seen_d.add(ent._pk) to_delete.append(ent) # If we got here, then to_delete includes all items to delete. Let's delete # them! for self in to_delete: self.delete(skip_on_delete_i_really_mean_it=SKIP_ON_DELETE) for self in to_save: # Careful not to resurrect deleted entities if self._pk not in seen_d: self.save()
[ "This", "function", "handles", "all", "on_delete", "semantics", "defined", "on", "OneToMany", "columns", "." ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/columns.py#L51-L115
[ "def", "_on_delete", "(", "ent", ")", ":", "seen_d", "=", "set", "(", "[", "ent", ".", "_pk", "]", ")", "to_delete", "=", "[", "ent", "]", "seen_s", "=", "set", "(", ")", "to_save", "=", "[", "]", "def", "_set_default", "(", "ent", ",", "attr", ",", "de", "=", "NULL", ")", ":", "pk", "=", "ent", ".", "_pk", "if", "pk", "in", "seen_d", ":", "# going to be deleted, don't need to modify", "return", "col", "=", "ent", ".", "__class__", ".", "_columns", "[", "attr", "]", "de", "=", "de", "if", "de", "is", "not", "NULL", "else", "col", ".", "_default", "if", "de", "in", "(", "None", ",", "NULL", ")", ":", "setattr", "(", "ent", ",", "attr", ",", "None", ")", "elif", "callable", "(", "col", ".", "_default", ")", ":", "setattr", "(", "ent", ",", "attr", ",", "col", ".", "_default", "(", ")", ")", "else", ":", "setattr", "(", "ent", ",", "attr", ",", "col", ".", "_default", ")", "if", "pk", "not", "in", "seen_s", ":", "seen_s", ".", "add", "(", "pk", ")", "to_save", ".", "append", "(", "ent", ")", "for", "self", "in", "to_delete", ":", "for", "tbl", ",", "attr", ",", "action", "in", "MODELS_REFERENCED", ".", "get", "(", "self", ".", "_namespace", ",", "(", ")", ")", ":", "if", "action", "==", "'no action'", ":", "continue", "refs", "=", "MODELS", "[", "tbl", "]", ".", "get_by", "(", "*", "*", "{", "attr", ":", "self", ".", "id", "}", ")", "if", "not", "refs", ":", "continue", "if", "action", "==", "'restrict'", ":", "# raise the exception here for a better traceback", "raise", "_restrict", "(", "self", ",", "attr", ",", "refs", ")", "elif", "action", "==", "'set null'", ":", "for", "ref", "in", "refs", ":", "_set_default", "(", "ref", ",", "attr", ",", "None", ")", "continue", "elif", "action", "==", "'set default'", ":", "for", "ref", "in", "refs", ":", "_set_default", "(", "ref", ",", "attr", ")", "continue", "# otherwise col._on_delete == 'cascade'", "for", "ent", "in", "(", "refs", "if", "isinstance", "(", "refs", ",", "list", ")", "else", "[", "refs", "]", ")", ":", "if", "ent", ".", "_pk", "not", "in", "seen_d", ":", "seen_d", ".", "add", "(", "ent", ".", "_pk", ")", "to_delete", ".", "append", "(", "ent", ")", "# If we got here, then to_delete includes all items to delete. Let's delete", "# them!", "for", "self", "in", "to_delete", ":", "self", ".", "delete", "(", "skip_on_delete_i_really_mean_it", "=", "SKIP_ON_DELETE", ")", "for", "self", "in", "to_save", ":", "# Careful not to resurrect deleted entities", "if", "self", ".", "_pk", "not", "in", "seen_d", ":", "self", ".", "save", "(", ")" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
redis_prefix_lua
Performs the actual prefix, suffix, and pattern match operations.
rom/index.py
def redis_prefix_lua(conn, dest, index, prefix, is_first, pattern=None): ''' Performs the actual prefix, suffix, and pattern match operations. ''' tkey = '%s:%s'%(index.partition(':')[0], uuid.uuid4()) start, end = _start_end(prefix) return _redis_prefix_lua(conn, [dest, tkey, index], [start, end, pattern or prefix, int(pattern is not None), int(bool(is_first))] )
def redis_prefix_lua(conn, dest, index, prefix, is_first, pattern=None): ''' Performs the actual prefix, suffix, and pattern match operations. ''' tkey = '%s:%s'%(index.partition(':')[0], uuid.uuid4()) start, end = _start_end(prefix) return _redis_prefix_lua(conn, [dest, tkey, index], [start, end, pattern or prefix, int(pattern is not None), int(bool(is_first))] )
[ "Performs", "the", "actual", "prefix", "suffix", "and", "pattern", "match", "operations", "." ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/index.py#L386-L395
[ "def", "redis_prefix_lua", "(", "conn", ",", "dest", ",", "index", ",", "prefix", ",", "is_first", ",", "pattern", "=", "None", ")", ":", "tkey", "=", "'%s:%s'", "%", "(", "index", ".", "partition", "(", "':'", ")", "[", "0", "]", ",", "uuid", ".", "uuid4", "(", ")", ")", "start", ",", "end", "=", "_start_end", "(", "prefix", ")", "return", "_redis_prefix_lua", "(", "conn", ",", "[", "dest", ",", "tkey", ",", "index", "]", ",", "[", "start", ",", "end", ",", "pattern", "or", "prefix", ",", "int", "(", "pattern", "is", "not", "None", ")", ",", "int", "(", "bool", "(", "is_first", ")", ")", "]", ")" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
estimate_work_lua
Estimates the total work necessary to calculate the prefix match over the given index with the provided prefix.
rom/index.py
def estimate_work_lua(conn, index, prefix): ''' Estimates the total work necessary to calculate the prefix match over the given index with the provided prefix. ''' if index.endswith(':idx'): args = [] if not prefix else list(prefix) if args: args[0] = '-inf' if args[0] is None else repr(float(args[0])) args[1] = 'inf' if args[1] is None else repr(float(args[1])) return _estimate_work_lua(conn, [index], args, force_eval=True) elif index.endswith(':geo'): return _estimate_work_lua(conn, [index], filter(None, [prefix]), force_eval=True) start, end = _start_end(prefix) return _estimate_work_lua(conn, [index], [start, '(' + end], force_eval=True)
def estimate_work_lua(conn, index, prefix): ''' Estimates the total work necessary to calculate the prefix match over the given index with the provided prefix. ''' if index.endswith(':idx'): args = [] if not prefix else list(prefix) if args: args[0] = '-inf' if args[0] is None else repr(float(args[0])) args[1] = 'inf' if args[1] is None else repr(float(args[1])) return _estimate_work_lua(conn, [index], args, force_eval=True) elif index.endswith(':geo'): return _estimate_work_lua(conn, [index], filter(None, [prefix]), force_eval=True) start, end = _start_end(prefix) return _estimate_work_lua(conn, [index], [start, '(' + end], force_eval=True)
[ "Estimates", "the", "total", "work", "necessary", "to", "calculate", "the", "prefix", "match", "over", "the", "given", "index", "with", "the", "provided", "prefix", "." ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/index.py#L482-L497
[ "def", "estimate_work_lua", "(", "conn", ",", "index", ",", "prefix", ")", ":", "if", "index", ".", "endswith", "(", "':idx'", ")", ":", "args", "=", "[", "]", "if", "not", "prefix", "else", "list", "(", "prefix", ")", "if", "args", ":", "args", "[", "0", "]", "=", "'-inf'", "if", "args", "[", "0", "]", "is", "None", "else", "repr", "(", "float", "(", "args", "[", "0", "]", ")", ")", "args", "[", "1", "]", "=", "'inf'", "if", "args", "[", "1", "]", "is", "None", "else", "repr", "(", "float", "(", "args", "[", "1", "]", ")", ")", "return", "_estimate_work_lua", "(", "conn", ",", "[", "index", "]", ",", "args", ",", "force_eval", "=", "True", ")", "elif", "index", ".", "endswith", "(", "':geo'", ")", ":", "return", "_estimate_work_lua", "(", "conn", ",", "[", "index", "]", ",", "filter", "(", "None", ",", "[", "prefix", "]", ")", ",", "force_eval", "=", "True", ")", "start", ",", "end", "=", "_start_end", "(", "prefix", ")", "return", "_estimate_work_lua", "(", "conn", ",", "[", "index", "]", ",", "[", "start", ",", "'('", "+", "end", "]", ",", "force_eval", "=", "True", ")" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
GeneralIndex.search
Search for model ids that match the provided filters. Arguments: * *filters* - A list of filters that apply to the search of one of the following two forms: 1. ``'column:string'`` - a plain string will match a word in a text search on the column .. note:: Read the documentation about the ``Query`` object for what is actually passed during text search 2. ``('column', min, max)`` - a numeric column range search, between min and max (inclusive by default) .. note:: Read the documentation about the ``Query`` object for information about open-ended ranges 3. ``['column:string1', 'column:string2']`` - will match any of the provided words in a text search on the column 4. ``Prefix('column', 'prefix')`` - will match prefixes of words in a text search on the column 5. ``Suffix('column', 'suffix')`` - will match suffixes of words in a text search on the column 6. ``Pattern('column', 'pattern')`` - will match patterns over words in a text search on the column * *order_by* - A string that names the numeric column by which to sort the results by. Prefixing with '-' will return results in descending order .. note:: While you can technically pass a non-numeric index as an *order_by* clause, the results will basically be to order the results by string comparison of the ids (10 will come before 2). .. note:: If you omit the ``order_by`` argument, results will be ordered by the last filter. If the last filter was a text filter, see the previous note. If the last filter was numeric, then results will be ordered by that result. * *offset* - A numeric starting offset for results * *count* - The maximum number of results to return from the query
rom/index.py
def search(self, conn, filters, order_by, offset=None, count=None, timeout=None): ''' Search for model ids that match the provided filters. Arguments: * *filters* - A list of filters that apply to the search of one of the following two forms: 1. ``'column:string'`` - a plain string will match a word in a text search on the column .. note:: Read the documentation about the ``Query`` object for what is actually passed during text search 2. ``('column', min, max)`` - a numeric column range search, between min and max (inclusive by default) .. note:: Read the documentation about the ``Query`` object for information about open-ended ranges 3. ``['column:string1', 'column:string2']`` - will match any of the provided words in a text search on the column 4. ``Prefix('column', 'prefix')`` - will match prefixes of words in a text search on the column 5. ``Suffix('column', 'suffix')`` - will match suffixes of words in a text search on the column 6. ``Pattern('column', 'pattern')`` - will match patterns over words in a text search on the column * *order_by* - A string that names the numeric column by which to sort the results by. Prefixing with '-' will return results in descending order .. note:: While you can technically pass a non-numeric index as an *order_by* clause, the results will basically be to order the results by string comparison of the ids (10 will come before 2). .. note:: If you omit the ``order_by`` argument, results will be ordered by the last filter. If the last filter was a text filter, see the previous note. If the last filter was numeric, then results will be ordered by that result. * *offset* - A numeric starting offset for results * *count* - The maximum number of results to return from the query ''' # prepare the filters pipe, intersect, temp_id = self._prepare(conn, filters) # handle ordering if order_by: reverse = order_by and order_by.startswith('-') order_clause = '%s:%s:idx'%(self.namespace, order_by.lstrip('-')) intersect(temp_id, {temp_id:0, order_clause: -1 if reverse else 1}) # handle returning the temporary result key if timeout is not None: pipe.expire(temp_id, timeout) pipe.execute() return temp_id offset = offset if offset is not None else 0 end = (offset + count - 1) if count and count > 0 else -1 pipe.zrange(temp_id, offset, end) pipe.delete(temp_id) return pipe.execute()[-2]
def search(self, conn, filters, order_by, offset=None, count=None, timeout=None): ''' Search for model ids that match the provided filters. Arguments: * *filters* - A list of filters that apply to the search of one of the following two forms: 1. ``'column:string'`` - a plain string will match a word in a text search on the column .. note:: Read the documentation about the ``Query`` object for what is actually passed during text search 2. ``('column', min, max)`` - a numeric column range search, between min and max (inclusive by default) .. note:: Read the documentation about the ``Query`` object for information about open-ended ranges 3. ``['column:string1', 'column:string2']`` - will match any of the provided words in a text search on the column 4. ``Prefix('column', 'prefix')`` - will match prefixes of words in a text search on the column 5. ``Suffix('column', 'suffix')`` - will match suffixes of words in a text search on the column 6. ``Pattern('column', 'pattern')`` - will match patterns over words in a text search on the column * *order_by* - A string that names the numeric column by which to sort the results by. Prefixing with '-' will return results in descending order .. note:: While you can technically pass a non-numeric index as an *order_by* clause, the results will basically be to order the results by string comparison of the ids (10 will come before 2). .. note:: If you omit the ``order_by`` argument, results will be ordered by the last filter. If the last filter was a text filter, see the previous note. If the last filter was numeric, then results will be ordered by that result. * *offset* - A numeric starting offset for results * *count* - The maximum number of results to return from the query ''' # prepare the filters pipe, intersect, temp_id = self._prepare(conn, filters) # handle ordering if order_by: reverse = order_by and order_by.startswith('-') order_clause = '%s:%s:idx'%(self.namespace, order_by.lstrip('-')) intersect(temp_id, {temp_id:0, order_clause: -1 if reverse else 1}) # handle returning the temporary result key if timeout is not None: pipe.expire(temp_id, timeout) pipe.execute() return temp_id offset = offset if offset is not None else 0 end = (offset + count - 1) if count and count > 0 else -1 pipe.zrange(temp_id, offset, end) pipe.delete(temp_id) return pipe.execute()[-2]
[ "Search", "for", "model", "ids", "that", "match", "the", "provided", "filters", "." ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/index.py#L220-L288
[ "def", "search", "(", "self", ",", "conn", ",", "filters", ",", "order_by", ",", "offset", "=", "None", ",", "count", "=", "None", ",", "timeout", "=", "None", ")", ":", "# prepare the filters", "pipe", ",", "intersect", ",", "temp_id", "=", "self", ".", "_prepare", "(", "conn", ",", "filters", ")", "# handle ordering", "if", "order_by", ":", "reverse", "=", "order_by", "and", "order_by", ".", "startswith", "(", "'-'", ")", "order_clause", "=", "'%s:%s:idx'", "%", "(", "self", ".", "namespace", ",", "order_by", ".", "lstrip", "(", "'-'", ")", ")", "intersect", "(", "temp_id", ",", "{", "temp_id", ":", "0", ",", "order_clause", ":", "-", "1", "if", "reverse", "else", "1", "}", ")", "# handle returning the temporary result key", "if", "timeout", "is", "not", "None", ":", "pipe", ".", "expire", "(", "temp_id", ",", "timeout", ")", "pipe", ".", "execute", "(", ")", "return", "temp_id", "offset", "=", "offset", "if", "offset", "is", "not", "None", "else", "0", "end", "=", "(", "offset", "+", "count", "-", "1", ")", "if", "count", "and", "count", ">", "0", "else", "-", "1", "pipe", ".", "zrange", "(", "temp_id", ",", "offset", ",", "end", ")", "pipe", ".", "delete", "(", "temp_id", ")", "return", "pipe", ".", "execute", "(", ")", "[", "-", "2", "]" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
GeneralIndex.count
Returns the count of the items that match the provided filters. For the meaning of what the ``filters`` argument means, see the ``.search()`` method docs.
rom/index.py
def count(self, conn, filters): ''' Returns the count of the items that match the provided filters. For the meaning of what the ``filters`` argument means, see the ``.search()`` method docs. ''' pipe, intersect, temp_id = self._prepare(conn, filters) pipe.zcard(temp_id) pipe.delete(temp_id) return pipe.execute()[-2]
def count(self, conn, filters): ''' Returns the count of the items that match the provided filters. For the meaning of what the ``filters`` argument means, see the ``.search()`` method docs. ''' pipe, intersect, temp_id = self._prepare(conn, filters) pipe.zcard(temp_id) pipe.delete(temp_id) return pipe.execute()[-2]
[ "Returns", "the", "count", "of", "the", "items", "that", "match", "the", "provided", "filters", "." ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/index.py#L290-L300
[ "def", "count", "(", "self", ",", "conn", ",", "filters", ")", ":", "pipe", ",", "intersect", ",", "temp_id", "=", "self", ".", "_prepare", "(", "conn", ",", "filters", ")", "pipe", ".", "zcard", "(", "temp_id", ")", "pipe", ".", "delete", "(", "temp_id", ")", "return", "pipe", ".", "execute", "(", ")", "[", "-", "2", "]" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
_connect
Tries to get the _conn attribute from a model. Barring that, gets the global default connection using other methods.
rom/util.py
def _connect(obj): ''' Tries to get the _conn attribute from a model. Barring that, gets the global default connection using other methods. ''' from .columns import MODELS if isinstance(obj, MODELS['Model']): obj = obj.__class__ if hasattr(obj, '_conn'): return obj._conn if hasattr(obj, 'CONN'): return obj.CONN return get_connection()
def _connect(obj): ''' Tries to get the _conn attribute from a model. Barring that, gets the global default connection using other methods. ''' from .columns import MODELS if isinstance(obj, MODELS['Model']): obj = obj.__class__ if hasattr(obj, '_conn'): return obj._conn if hasattr(obj, 'CONN'): return obj.CONN return get_connection()
[ "Tries", "to", "get", "the", "_conn", "attribute", "from", "a", "model", ".", "Barring", "that", "gets", "the", "global", "default", "connection", "using", "other", "methods", "." ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/util.py#L181-L193
[ "def", "_connect", "(", "obj", ")", ":", "from", ".", "columns", "import", "MODELS", "if", "isinstance", "(", "obj", ",", "MODELS", "[", "'Model'", "]", ")", ":", "obj", "=", "obj", ".", "__class__", "if", "hasattr", "(", "obj", ",", "'_conn'", ")", ":", "return", "obj", ".", "_conn", "if", "hasattr", "(", "obj", ",", "'CONN'", ")", ":", "return", "obj", ".", "CONN", "return", "get_connection", "(", ")" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
FULL_TEXT
This is a basic full-text index keygen function. Words are lowercased, split by whitespace, and stripped of punctuation from both ends before an inverted index is created for term searching.
rom/util.py
def FULL_TEXT(val): ''' This is a basic full-text index keygen function. Words are lowercased, split by whitespace, and stripped of punctuation from both ends before an inverted index is created for term searching. ''' if isinstance(val, float): val = repr(val) elif val in (None, ''): return None elif not isinstance(val, six.string_types): if six.PY3 and isinstance(val, bytes): val = val.decode('latin-1') else: val = str(val) r = sorted(set([x for x in [s.lower().strip(string.punctuation) for s in val.split()] if x])) if not isinstance(val, str): # unicode on py2k return [s.encode('utf-8') for s in r] return r
def FULL_TEXT(val): ''' This is a basic full-text index keygen function. Words are lowercased, split by whitespace, and stripped of punctuation from both ends before an inverted index is created for term searching. ''' if isinstance(val, float): val = repr(val) elif val in (None, ''): return None elif not isinstance(val, six.string_types): if six.PY3 and isinstance(val, bytes): val = val.decode('latin-1') else: val = str(val) r = sorted(set([x for x in [s.lower().strip(string.punctuation) for s in val.split()] if x])) if not isinstance(val, str): # unicode on py2k return [s.encode('utf-8') for s in r] return r
[ "This", "is", "a", "basic", "full", "-", "text", "index", "keygen", "function", ".", "Words", "are", "lowercased", "split", "by", "whitespace", "and", "stripped", "of", "punctuation", "from", "both", "ends", "before", "an", "inverted", "index", "is", "created", "for", "term", "searching", "." ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/util.py#L254-L272
[ "def", "FULL_TEXT", "(", "val", ")", ":", "if", "isinstance", "(", "val", ",", "float", ")", ":", "val", "=", "repr", "(", "val", ")", "elif", "val", "in", "(", "None", ",", "''", ")", ":", "return", "None", "elif", "not", "isinstance", "(", "val", ",", "six", ".", "string_types", ")", ":", "if", "six", ".", "PY3", "and", "isinstance", "(", "val", ",", "bytes", ")", ":", "val", "=", "val", ".", "decode", "(", "'latin-1'", ")", "else", ":", "val", "=", "str", "(", "val", ")", "r", "=", "sorted", "(", "set", "(", "[", "x", "for", "x", "in", "[", "s", ".", "lower", "(", ")", ".", "strip", "(", "string", ".", "punctuation", ")", "for", "s", "in", "val", ".", "split", "(", ")", "]", "if", "x", "]", ")", ")", "if", "not", "isinstance", "(", "val", ",", "str", ")", ":", "# unicode on py2k", "return", "[", "s", ".", "encode", "(", "'utf-8'", ")", "for", "s", "in", "r", "]", "return", "r" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
SIMPLE
This is a basic case-sensitive "sorted order" index keygen function for strings. This will return a value that is suitable to be used for ordering by a 7-byte prefix of a string (that is 7 characters from a byte-string, and 1.75-7 characters from a unicode string, depending on character -> encoding length). .. warning:: Case sensitivity is based on the (encoded) byte prefixes of the strings/text being indexed, so ordering *may be different* than a native comparison ordering (especially if an order is different based on characters past the 7th encoded byte).
rom/util.py
def SIMPLE(val): ''' This is a basic case-sensitive "sorted order" index keygen function for strings. This will return a value that is suitable to be used for ordering by a 7-byte prefix of a string (that is 7 characters from a byte-string, and 1.75-7 characters from a unicode string, depending on character -> encoding length). .. warning:: Case sensitivity is based on the (encoded) byte prefixes of the strings/text being indexed, so ordering *may be different* than a native comparison ordering (especially if an order is different based on characters past the 7th encoded byte). ''' if not val: return None if not isinstance(val, six.string_types): if six.PY3 and isinstance(val, bytes): val = val.decode('latin-1') else: val = str(val) return {'': _prefix_score(val)}
def SIMPLE(val): ''' This is a basic case-sensitive "sorted order" index keygen function for strings. This will return a value that is suitable to be used for ordering by a 7-byte prefix of a string (that is 7 characters from a byte-string, and 1.75-7 characters from a unicode string, depending on character -> encoding length). .. warning:: Case sensitivity is based on the (encoded) byte prefixes of the strings/text being indexed, so ordering *may be different* than a native comparison ordering (especially if an order is different based on characters past the 7th encoded byte). ''' if not val: return None if not isinstance(val, six.string_types): if six.PY3 and isinstance(val, bytes): val = val.decode('latin-1') else: val = str(val) return {'': _prefix_score(val)}
[ "This", "is", "a", "basic", "case", "-", "sensitive", "sorted", "order", "index", "keygen", "function", "for", "strings", ".", "This", "will", "return", "a", "value", "that", "is", "suitable", "to", "be", "used", "for", "ordering", "by", "a", "7", "-", "byte", "prefix", "of", "a", "string", "(", "that", "is", "7", "characters", "from", "a", "byte", "-", "string", "and", "1", ".", "75", "-", "7", "characters", "from", "a", "unicode", "string", "depending", "on", "character", "-", ">", "encoding", "length", ")", "." ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/util.py#L278-L298
[ "def", "SIMPLE", "(", "val", ")", ":", "if", "not", "val", ":", "return", "None", "if", "not", "isinstance", "(", "val", ",", "six", ".", "string_types", ")", ":", "if", "six", ".", "PY3", "and", "isinstance", "(", "val", ",", "bytes", ")", ":", "val", "=", "val", ".", "decode", "(", "'latin-1'", ")", "else", ":", "val", "=", "str", "(", "val", ")", "return", "{", "''", ":", "_prefix_score", "(", "val", ")", "}" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
IDENTITY
This is a basic "equality" index keygen, primarily meant to be used for things like:: Model.query.filter(col='value') Where ``FULL_TEXT`` would transform a sentence like "A Simple Sentence" into an inverted index searchable by the words "a", "simple", and/or "sentence", ``IDENTITY`` will only be searchable by the orginal full sentence with the same capitalization - "A Simple Sentence". See ``IDENTITY_CI`` for the same function, only case-insensitive.
rom/util.py
def IDENTITY(val): ''' This is a basic "equality" index keygen, primarily meant to be used for things like:: Model.query.filter(col='value') Where ``FULL_TEXT`` would transform a sentence like "A Simple Sentence" into an inverted index searchable by the words "a", "simple", and/or "sentence", ``IDENTITY`` will only be searchable by the orginal full sentence with the same capitalization - "A Simple Sentence". See ``IDENTITY_CI`` for the same function, only case-insensitive. ''' if not val: return None if not isinstance(val, six.string_types_ex): val = str(val) return [val]
def IDENTITY(val): ''' This is a basic "equality" index keygen, primarily meant to be used for things like:: Model.query.filter(col='value') Where ``FULL_TEXT`` would transform a sentence like "A Simple Sentence" into an inverted index searchable by the words "a", "simple", and/or "sentence", ``IDENTITY`` will only be searchable by the orginal full sentence with the same capitalization - "A Simple Sentence". See ``IDENTITY_CI`` for the same function, only case-insensitive. ''' if not val: return None if not isinstance(val, six.string_types_ex): val = str(val) return [val]
[ "This", "is", "a", "basic", "equality", "index", "keygen", "primarily", "meant", "to", "be", "used", "for", "things", "like", "::" ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/util.py#L312-L329
[ "def", "IDENTITY", "(", "val", ")", ":", "if", "not", "val", ":", "return", "None", "if", "not", "isinstance", "(", "val", ",", "six", ".", "string_types_ex", ")", ":", "val", "=", "str", "(", "val", ")", "return", "[", "val", "]" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
refresh_indices
This utility function will iterate over all entities of a provided model, refreshing their indices. This is primarily useful after adding an index on a column. Arguments: * *model* - the model whose entities you want to reindex * *block_size* - the maximum number of entities you want to fetch from Redis at a time, defaulting to 100 This function will yield its progression through re-indexing all of your entities. Example use:: for progress, total in refresh_indices(MyModel, block_size=200): print "%s of %s"%(progress, total) .. note:: This uses the session object to handle index refresh via calls to ``.commit()``. If you have any outstanding entities known in the session, they will be committed.
rom/util.py
def refresh_indices(model, block_size=100): ''' This utility function will iterate over all entities of a provided model, refreshing their indices. This is primarily useful after adding an index on a column. Arguments: * *model* - the model whose entities you want to reindex * *block_size* - the maximum number of entities you want to fetch from Redis at a time, defaulting to 100 This function will yield its progression through re-indexing all of your entities. Example use:: for progress, total in refresh_indices(MyModel, block_size=200): print "%s of %s"%(progress, total) .. note:: This uses the session object to handle index refresh via calls to ``.commit()``. If you have any outstanding entities known in the session, they will be committed. ''' conn = _connect(model) max_id = int(conn.get('%s:%s:'%(model._namespace, model._pkey)) or '0') block_size = max(block_size, 10) for i in range(1, max_id+1, block_size): # fetches entities, keeping a record in the session models = model.get(list(range(i, i+block_size))) models # for pyflakes # re-save un-modified data, resulting in index-only updates session.commit(all=True) yield min(i+block_size, max_id), max_id
def refresh_indices(model, block_size=100): ''' This utility function will iterate over all entities of a provided model, refreshing their indices. This is primarily useful after adding an index on a column. Arguments: * *model* - the model whose entities you want to reindex * *block_size* - the maximum number of entities you want to fetch from Redis at a time, defaulting to 100 This function will yield its progression through re-indexing all of your entities. Example use:: for progress, total in refresh_indices(MyModel, block_size=200): print "%s of %s"%(progress, total) .. note:: This uses the session object to handle index refresh via calls to ``.commit()``. If you have any outstanding entities known in the session, they will be committed. ''' conn = _connect(model) max_id = int(conn.get('%s:%s:'%(model._namespace, model._pkey)) or '0') block_size = max(block_size, 10) for i in range(1, max_id+1, block_size): # fetches entities, keeping a record in the session models = model.get(list(range(i, i+block_size))) models # for pyflakes # re-save un-modified data, resulting in index-only updates session.commit(all=True) yield min(i+block_size, max_id), max_id
[ "This", "utility", "function", "will", "iterate", "over", "all", "entities", "of", "a", "provided", "model", "refreshing", "their", "indices", ".", "This", "is", "primarily", "useful", "after", "adding", "an", "index", "on", "a", "column", "." ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/util.py#L696-L729
[ "def", "refresh_indices", "(", "model", ",", "block_size", "=", "100", ")", ":", "conn", "=", "_connect", "(", "model", ")", "max_id", "=", "int", "(", "conn", ".", "get", "(", "'%s:%s:'", "%", "(", "model", ".", "_namespace", ",", "model", ".", "_pkey", ")", ")", "or", "'0'", ")", "block_size", "=", "max", "(", "block_size", ",", "10", ")", "for", "i", "in", "range", "(", "1", ",", "max_id", "+", "1", ",", "block_size", ")", ":", "# fetches entities, keeping a record in the session", "models", "=", "model", ".", "get", "(", "list", "(", "range", "(", "i", ",", "i", "+", "block_size", ")", ")", ")", "models", "# for pyflakes", "# re-save un-modified data, resulting in index-only updates", "session", ".", "commit", "(", "all", "=", "True", ")", "yield", "min", "(", "i", "+", "block_size", ",", "max_id", ")", ",", "max_id" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
clean_old_index
This utility function will clean out old index data that was accidentally left during item deletion in rom versions <= 0.27.0 . You should run this after you have upgraded all of your clients to version 0.28.0 or later. Arguments: * *model* - the model whose entities you want to reindex * *block_size* - the maximum number of items to check at a time defaulting to 100 This function will yield its progression through re-checking all of the data that could be left over. Example use:: for progress, total in clean_old_index(MyModel, block_size=200): print "%s of %s"%(progress, total)
rom/util.py
def clean_old_index(model, block_size=100, **kwargs): ''' This utility function will clean out old index data that was accidentally left during item deletion in rom versions <= 0.27.0 . You should run this after you have upgraded all of your clients to version 0.28.0 or later. Arguments: * *model* - the model whose entities you want to reindex * *block_size* - the maximum number of items to check at a time defaulting to 100 This function will yield its progression through re-checking all of the data that could be left over. Example use:: for progress, total in clean_old_index(MyModel, block_size=200): print "%s of %s"%(progress, total) ''' conn = _connect(model) version = list(map(int, conn.info()['redis_version'].split('.')[:2])) has_hscan = version >= [2, 8] pipe = conn.pipeline(True) prefix = '%s:'%model._namespace index = prefix + ':' block_size = max(block_size, 10) force_hscan = kwargs.get('force_hscan', False) if (has_hscan or force_hscan) and force_hscan is not None: max_id = conn.hlen(index) cursor = None scanned = 0 while cursor != b'0': cursor, remove = _scan_index_lua(conn, [index, prefix], [cursor or '0', block_size, 0, 0]) if remove: _clean_index_lua(conn, [model._namespace], remove) scanned += block_size if scanned > max_id: max_id = scanned + 1 yield scanned, max_id # need to scan over unique indexes :/ for uniq in chain(model._unique, model._cunique): name = uniq if isinstance(uniq, six.string_types) else ':'.join(uniq) idx = prefix + name + ':uidx' cursor = None while cursor != b'0': cursor, remove = _scan_index_lua(conn, [idx, prefix], [cursor or '0', block_size, 1, 0]) if remove: conn.hdel(idx, *remove) scanned += block_size if scanned > max_id: max_id = scanned + 1 yield scanned, max_id else: if model._unique or model._cunique: if has_hscan: warnings.warn("You have disabled the use of HSCAN to clean up indexes, this will prevent unique index cleanup", stacklevel=2) else: warnings.warn("Unique indexes cannot be cleaned up in Redis versions prior to 2.8", stacklevel=2) max_id = int(conn.get('%s%s:'%(prefix, model._pkey)) or '0') for i in range(1, max_id+1, block_size): ids = list(range(i, min(i+block_size, max_id+1))) for id in ids: pipe.exists(prefix + str(id)) pipe.hexists(index, id) result = iter(pipe.execute()) remove = [id for id, ent, ind in zip(ids, result, result) if ind and not ent] if remove: _clean_index_lua(conn, [model._namespace], remove) yield min(i+block_size, max_id-1), max_id yield max_id, max_id
def clean_old_index(model, block_size=100, **kwargs): ''' This utility function will clean out old index data that was accidentally left during item deletion in rom versions <= 0.27.0 . You should run this after you have upgraded all of your clients to version 0.28.0 or later. Arguments: * *model* - the model whose entities you want to reindex * *block_size* - the maximum number of items to check at a time defaulting to 100 This function will yield its progression through re-checking all of the data that could be left over. Example use:: for progress, total in clean_old_index(MyModel, block_size=200): print "%s of %s"%(progress, total) ''' conn = _connect(model) version = list(map(int, conn.info()['redis_version'].split('.')[:2])) has_hscan = version >= [2, 8] pipe = conn.pipeline(True) prefix = '%s:'%model._namespace index = prefix + ':' block_size = max(block_size, 10) force_hscan = kwargs.get('force_hscan', False) if (has_hscan or force_hscan) and force_hscan is not None: max_id = conn.hlen(index) cursor = None scanned = 0 while cursor != b'0': cursor, remove = _scan_index_lua(conn, [index, prefix], [cursor or '0', block_size, 0, 0]) if remove: _clean_index_lua(conn, [model._namespace], remove) scanned += block_size if scanned > max_id: max_id = scanned + 1 yield scanned, max_id # need to scan over unique indexes :/ for uniq in chain(model._unique, model._cunique): name = uniq if isinstance(uniq, six.string_types) else ':'.join(uniq) idx = prefix + name + ':uidx' cursor = None while cursor != b'0': cursor, remove = _scan_index_lua(conn, [idx, prefix], [cursor or '0', block_size, 1, 0]) if remove: conn.hdel(idx, *remove) scanned += block_size if scanned > max_id: max_id = scanned + 1 yield scanned, max_id else: if model._unique or model._cunique: if has_hscan: warnings.warn("You have disabled the use of HSCAN to clean up indexes, this will prevent unique index cleanup", stacklevel=2) else: warnings.warn("Unique indexes cannot be cleaned up in Redis versions prior to 2.8", stacklevel=2) max_id = int(conn.get('%s%s:'%(prefix, model._pkey)) or '0') for i in range(1, max_id+1, block_size): ids = list(range(i, min(i+block_size, max_id+1))) for id in ids: pipe.exists(prefix + str(id)) pipe.hexists(index, id) result = iter(pipe.execute()) remove = [id for id, ent, ind in zip(ids, result, result) if ind and not ent] if remove: _clean_index_lua(conn, [model._namespace], remove) yield min(i+block_size, max_id-1), max_id yield max_id, max_id
[ "This", "utility", "function", "will", "clean", "out", "old", "index", "data", "that", "was", "accidentally", "left", "during", "item", "deletion", "in", "rom", "versions", "<", "=", "0", ".", "27", ".", "0", ".", "You", "should", "run", "this", "after", "you", "have", "upgraded", "all", "of", "your", "clients", "to", "version", "0", ".", "28", ".", "0", "or", "later", "." ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/util.py#L731-L811
[ "def", "clean_old_index", "(", "model", ",", "block_size", "=", "100", ",", "*", "*", "kwargs", ")", ":", "conn", "=", "_connect", "(", "model", ")", "version", "=", "list", "(", "map", "(", "int", ",", "conn", ".", "info", "(", ")", "[", "'redis_version'", "]", ".", "split", "(", "'.'", ")", "[", ":", "2", "]", ")", ")", "has_hscan", "=", "version", ">=", "[", "2", ",", "8", "]", "pipe", "=", "conn", ".", "pipeline", "(", "True", ")", "prefix", "=", "'%s:'", "%", "model", ".", "_namespace", "index", "=", "prefix", "+", "':'", "block_size", "=", "max", "(", "block_size", ",", "10", ")", "force_hscan", "=", "kwargs", ".", "get", "(", "'force_hscan'", ",", "False", ")", "if", "(", "has_hscan", "or", "force_hscan", ")", "and", "force_hscan", "is", "not", "None", ":", "max_id", "=", "conn", ".", "hlen", "(", "index", ")", "cursor", "=", "None", "scanned", "=", "0", "while", "cursor", "!=", "b'0'", ":", "cursor", ",", "remove", "=", "_scan_index_lua", "(", "conn", ",", "[", "index", ",", "prefix", "]", ",", "[", "cursor", "or", "'0'", ",", "block_size", ",", "0", ",", "0", "]", ")", "if", "remove", ":", "_clean_index_lua", "(", "conn", ",", "[", "model", ".", "_namespace", "]", ",", "remove", ")", "scanned", "+=", "block_size", "if", "scanned", ">", "max_id", ":", "max_id", "=", "scanned", "+", "1", "yield", "scanned", ",", "max_id", "# need to scan over unique indexes :/", "for", "uniq", "in", "chain", "(", "model", ".", "_unique", ",", "model", ".", "_cunique", ")", ":", "name", "=", "uniq", "if", "isinstance", "(", "uniq", ",", "six", ".", "string_types", ")", "else", "':'", ".", "join", "(", "uniq", ")", "idx", "=", "prefix", "+", "name", "+", "':uidx'", "cursor", "=", "None", "while", "cursor", "!=", "b'0'", ":", "cursor", ",", "remove", "=", "_scan_index_lua", "(", "conn", ",", "[", "idx", ",", "prefix", "]", ",", "[", "cursor", "or", "'0'", ",", "block_size", ",", "1", ",", "0", "]", ")", "if", "remove", ":", "conn", ".", "hdel", "(", "idx", ",", "*", "remove", ")", "scanned", "+=", "block_size", "if", "scanned", ">", "max_id", ":", "max_id", "=", "scanned", "+", "1", "yield", "scanned", ",", "max_id", "else", ":", "if", "model", ".", "_unique", "or", "model", ".", "_cunique", ":", "if", "has_hscan", ":", "warnings", ".", "warn", "(", "\"You have disabled the use of HSCAN to clean up indexes, this will prevent unique index cleanup\"", ",", "stacklevel", "=", "2", ")", "else", ":", "warnings", ".", "warn", "(", "\"Unique indexes cannot be cleaned up in Redis versions prior to 2.8\"", ",", "stacklevel", "=", "2", ")", "max_id", "=", "int", "(", "conn", ".", "get", "(", "'%s%s:'", "%", "(", "prefix", ",", "model", ".", "_pkey", ")", ")", "or", "'0'", ")", "for", "i", "in", "range", "(", "1", ",", "max_id", "+", "1", ",", "block_size", ")", ":", "ids", "=", "list", "(", "range", "(", "i", ",", "min", "(", "i", "+", "block_size", ",", "max_id", "+", "1", ")", ")", ")", "for", "id", "in", "ids", ":", "pipe", ".", "exists", "(", "prefix", "+", "str", "(", "id", ")", ")", "pipe", ".", "hexists", "(", "index", ",", "id", ")", "result", "=", "iter", "(", "pipe", ".", "execute", "(", ")", ")", "remove", "=", "[", "id", "for", "id", ",", "ent", ",", "ind", "in", "zip", "(", "ids", ",", "result", ",", "result", ")", "if", "ind", "and", "not", "ent", "]", "if", "remove", ":", "_clean_index_lua", "(", "conn", ",", "[", "model", ".", "_namespace", "]", ",", "remove", ")", "yield", "min", "(", "i", "+", "block_size", ",", "max_id", "-", "1", ")", ",", "max_id", "yield", "max_id", ",", "max_id" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
show_progress
This utility function will print the progress of a passed iterator job as started by ``refresh_indices()`` and ``clean_old_index()``. Usage example:: class RomTest(Model): pass for i in xrange(1000): RomTest().save() util.show_progress(util.clean_old_index(RomTest))
rom/util.py
def show_progress(job): ''' This utility function will print the progress of a passed iterator job as started by ``refresh_indices()`` and ``clean_old_index()``. Usage example:: class RomTest(Model): pass for i in xrange(1000): RomTest().save() util.show_progress(util.clean_old_index(RomTest)) ''' start = time.time() last_print = 0 last_line = 0 for prog, total in chain(job, [(1, 1)]): # Only print a line when we start, finish, or every .1 seconds if (time.time() - last_print) > .1 or prog >= total: delta = (time.time() - start) or .0001 line = "%.1f%% complete, %.1f seconds elapsed, %.1f seconds remaining"%( 100. * prog / (total or 1), delta, total * delta / (prog or 1) - delta) length = len(line) # pad the line out with spaces just in case our line got shorter line += max(last_line - length, 0) * ' ' print(line, end="\r") last_line = length last_print = time.time() print()
def show_progress(job): ''' This utility function will print the progress of a passed iterator job as started by ``refresh_indices()`` and ``clean_old_index()``. Usage example:: class RomTest(Model): pass for i in xrange(1000): RomTest().save() util.show_progress(util.clean_old_index(RomTest)) ''' start = time.time() last_print = 0 last_line = 0 for prog, total in chain(job, [(1, 1)]): # Only print a line when we start, finish, or every .1 seconds if (time.time() - last_print) > .1 or prog >= total: delta = (time.time() - start) or .0001 line = "%.1f%% complete, %.1f seconds elapsed, %.1f seconds remaining"%( 100. * prog / (total or 1), delta, total * delta / (prog or 1) - delta) length = len(line) # pad the line out with spaces just in case our line got shorter line += max(last_line - length, 0) * ' ' print(line, end="\r") last_line = length last_print = time.time() print()
[ "This", "utility", "function", "will", "print", "the", "progress", "of", "a", "passed", "iterator", "job", "as", "started", "by", "refresh_indices", "()", "and", "clean_old_index", "()", "." ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/util.py#L814-L844
[ "def", "show_progress", "(", "job", ")", ":", "start", "=", "time", ".", "time", "(", ")", "last_print", "=", "0", "last_line", "=", "0", "for", "prog", ",", "total", "in", "chain", "(", "job", ",", "[", "(", "1", ",", "1", ")", "]", ")", ":", "# Only print a line when we start, finish, or every .1 seconds", "if", "(", "time", ".", "time", "(", ")", "-", "last_print", ")", ">", ".1", "or", "prog", ">=", "total", ":", "delta", "=", "(", "time", ".", "time", "(", ")", "-", "start", ")", "or", ".0001", "line", "=", "\"%.1f%% complete, %.1f seconds elapsed, %.1f seconds remaining\"", "%", "(", "100.", "*", "prog", "/", "(", "total", "or", "1", ")", ",", "delta", ",", "total", "*", "delta", "/", "(", "prog", "or", "1", ")", "-", "delta", ")", "length", "=", "len", "(", "line", ")", "# pad the line out with spaces just in case our line got shorter", "line", "+=", "max", "(", "last_line", "-", "length", ",", "0", ")", "*", "' '", "print", "(", "line", ",", "end", "=", "\"\\r\"", ")", "last_line", "=", "length", "last_print", "=", "time", ".", "time", "(", ")", "print", "(", ")" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
_script_load
Borrowed/modified from my book, Redis in Action: https://github.com/josiahcarlson/redis-in-action/blob/master/python/ch11_listing_source.py Used for Lua scripting support when writing against Redis 2.6+ to allow for multiple unique columns per model.
rom/util.py
def _script_load(script): ''' Borrowed/modified from my book, Redis in Action: https://github.com/josiahcarlson/redis-in-action/blob/master/python/ch11_listing_source.py Used for Lua scripting support when writing against Redis 2.6+ to allow for multiple unique columns per model. ''' script = script.encode('utf-8') if isinstance(script, six.text_type) else script sha = [None, sha1(script).hexdigest()] def call(conn, keys=[], args=[], force_eval=False): keys = tuple(keys) args = tuple(args) if not force_eval: if not sha[0]: try: # executing the script implicitly loads it return conn.execute_command( 'EVAL', script, len(keys), *(keys + args)) finally: # thread safe by re-using the GIL ;) del sha[:-1] try: return conn.execute_command( "EVALSHA", sha[0], len(keys), *(keys+args)) except redis.exceptions.ResponseError as msg: if not any(msg.args[0].startswith(nsm) for nsm in NO_SCRIPT_MESSAGES): raise return conn.execute_command( "EVAL", script, len(keys), *(keys+args)) return call
def _script_load(script): ''' Borrowed/modified from my book, Redis in Action: https://github.com/josiahcarlson/redis-in-action/blob/master/python/ch11_listing_source.py Used for Lua scripting support when writing against Redis 2.6+ to allow for multiple unique columns per model. ''' script = script.encode('utf-8') if isinstance(script, six.text_type) else script sha = [None, sha1(script).hexdigest()] def call(conn, keys=[], args=[], force_eval=False): keys = tuple(keys) args = tuple(args) if not force_eval: if not sha[0]: try: # executing the script implicitly loads it return conn.execute_command( 'EVAL', script, len(keys), *(keys + args)) finally: # thread safe by re-using the GIL ;) del sha[:-1] try: return conn.execute_command( "EVALSHA", sha[0], len(keys), *(keys+args)) except redis.exceptions.ResponseError as msg: if not any(msg.args[0].startswith(nsm) for nsm in NO_SCRIPT_MESSAGES): raise return conn.execute_command( "EVAL", script, len(keys), *(keys+args)) return call
[ "Borrowed", "/", "modified", "from", "my", "book", "Redis", "in", "Action", ":", "https", ":", "//", "github", ".", "com", "/", "josiahcarlson", "/", "redis", "-", "in", "-", "action", "/", "blob", "/", "master", "/", "python", "/", "ch11_listing_source", ".", "py" ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/util.py#L847-L881
[ "def", "_script_load", "(", "script", ")", ":", "script", "=", "script", ".", "encode", "(", "'utf-8'", ")", "if", "isinstance", "(", "script", ",", "six", ".", "text_type", ")", "else", "script", "sha", "=", "[", "None", ",", "sha1", "(", "script", ")", ".", "hexdigest", "(", ")", "]", "def", "call", "(", "conn", ",", "keys", "=", "[", "]", ",", "args", "=", "[", "]", ",", "force_eval", "=", "False", ")", ":", "keys", "=", "tuple", "(", "keys", ")", "args", "=", "tuple", "(", "args", ")", "if", "not", "force_eval", ":", "if", "not", "sha", "[", "0", "]", ":", "try", ":", "# executing the script implicitly loads it", "return", "conn", ".", "execute_command", "(", "'EVAL'", ",", "script", ",", "len", "(", "keys", ")", ",", "*", "(", "keys", "+", "args", ")", ")", "finally", ":", "# thread safe by re-using the GIL ;)", "del", "sha", "[", ":", "-", "1", "]", "try", ":", "return", "conn", ".", "execute_command", "(", "\"EVALSHA\"", ",", "sha", "[", "0", "]", ",", "len", "(", "keys", ")", ",", "*", "(", "keys", "+", "args", ")", ")", "except", "redis", ".", "exceptions", ".", "ResponseError", "as", "msg", ":", "if", "not", "any", "(", "msg", ".", "args", "[", "0", "]", ".", "startswith", "(", "nsm", ")", "for", "nsm", "in", "NO_SCRIPT_MESSAGES", ")", ":", "raise", "return", "conn", ".", "execute_command", "(", "\"EVAL\"", ",", "script", ",", "len", "(", "keys", ")", ",", "*", "(", "keys", "+", "args", ")", ")", "return", "call" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
EntityLock
Useful when you want exclusive access to an entity across all writers.:: # example import rom class Document(rom.Model): owner = rom.ManyToOne('User', on_delete='restrict') ... def change_owner(document, new_owner): with rom.util.EntityLock(document, 5, 90): document.owner = new_owner document.save()
rom/util.py
def EntityLock(entity, acquire_timeout, lock_timeout): ''' Useful when you want exclusive access to an entity across all writers.:: # example import rom class Document(rom.Model): owner = rom.ManyToOne('User', on_delete='restrict') ... def change_owner(document, new_owner): with rom.util.EntityLock(document, 5, 90): document.owner = new_owner document.save() ''' return Lock(entity._connection, entity._pk, acquire_timeout, lock_timeout)
def EntityLock(entity, acquire_timeout, lock_timeout): ''' Useful when you want exclusive access to an entity across all writers.:: # example import rom class Document(rom.Model): owner = rom.ManyToOne('User', on_delete='restrict') ... def change_owner(document, new_owner): with rom.util.EntityLock(document, 5, 90): document.owner = new_owner document.save() ''' return Lock(entity._connection, entity._pk, acquire_timeout, lock_timeout)
[ "Useful", "when", "you", "want", "exclusive", "access", "to", "an", "entity", "across", "all", "writers", ".", "::" ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/util.py#L1000-L1017
[ "def", "EntityLock", "(", "entity", ",", "acquire_timeout", ",", "lock_timeout", ")", ":", "return", "Lock", "(", "entity", ".", "_connection", ",", "entity", ".", "_pk", ",", "acquire_timeout", ",", "lock_timeout", ")" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
Session.add
Adds an entity to the session.
rom/util.py
def add(self, obj): ''' Adds an entity to the session. ''' if self.null_session: return self._init() pk = obj._pk if not pk.endswith(':None'): self.known[pk] = obj self.wknown[pk] = obj
def add(self, obj): ''' Adds an entity to the session. ''' if self.null_session: return self._init() pk = obj._pk if not pk.endswith(':None'): self.known[pk] = obj self.wknown[pk] = obj
[ "Adds", "an", "entity", "to", "the", "session", "." ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/util.py#L452-L462
[ "def", "add", "(", "self", ",", "obj", ")", ":", "if", "self", ".", "null_session", ":", "return", "self", ".", "_init", "(", ")", "pk", "=", "obj", ".", "_pk", "if", "not", "pk", ".", "endswith", "(", "':None'", ")", ":", "self", ".", "known", "[", "pk", "]", "=", "obj", "self", ".", "wknown", "[", "pk", "]", "=", "obj" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
Session.forget
Forgets about an entity (automatically called when an entity is deleted). Call this to ensure that an entity that you've modified is not automatically saved on ``session.commit()`` .
rom/util.py
def forget(self, obj): ''' Forgets about an entity (automatically called when an entity is deleted). Call this to ensure that an entity that you've modified is not automatically saved on ``session.commit()`` . ''' self._init() self.known.pop(obj._pk, None) self.wknown.pop(obj._pk, None)
def forget(self, obj): ''' Forgets about an entity (automatically called when an entity is deleted). Call this to ensure that an entity that you've modified is not automatically saved on ``session.commit()`` . ''' self._init() self.known.pop(obj._pk, None) self.wknown.pop(obj._pk, None)
[ "Forgets", "about", "an", "entity", "(", "automatically", "called", "when", "an", "entity", "is", "deleted", ")", ".", "Call", "this", "to", "ensure", "that", "an", "entity", "that", "you", "ve", "modified", "is", "not", "automatically", "saved", "on", "session", ".", "commit", "()", "." ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/util.py#L464-L472
[ "def", "forget", "(", "self", ",", "obj", ")", ":", "self", ".", "_init", "(", ")", "self", ".", "known", ".", "pop", "(", "obj", ".", "_pk", ",", "None", ")", "self", ".", "wknown", ".", "pop", "(", "obj", ".", "_pk", ",", "None", ")" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
Session.get
Fetches an entity from the session based on primary key.
rom/util.py
def get(self, pk): ''' Fetches an entity from the session based on primary key. ''' self._init() return self.known.get(pk) or self.wknown.get(pk)
def get(self, pk): ''' Fetches an entity from the session based on primary key. ''' self._init() return self.known.get(pk) or self.wknown.get(pk)
[ "Fetches", "an", "entity", "from", "the", "session", "based", "on", "primary", "key", "." ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/util.py#L474-L479
[ "def", "get", "(", "self", ",", "pk", ")", ":", "self", ".", "_init", "(", ")", "return", "self", ".", "known", ".", "get", "(", "pk", ")", "or", "self", ".", "wknown", ".", "get", "(", "pk", ")" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
Session.flush
Call ``.save()`` on all modified entities in the session. Use when you want to flush changes to Redis, but don't want to lose your local session cache. See the ``.commit()`` method for arguments and their meanings.
rom/util.py
def flush(self, full=False, all=False, force=False): ''' Call ``.save()`` on all modified entities in the session. Use when you want to flush changes to Redis, but don't want to lose your local session cache. See the ``.commit()`` method for arguments and their meanings. ''' self._init() return self.save(*self.known.values(), full=full, all=all, force=force)
def flush(self, full=False, all=False, force=False): ''' Call ``.save()`` on all modified entities in the session. Use when you want to flush changes to Redis, but don't want to lose your local session cache. See the ``.commit()`` method for arguments and their meanings. ''' self._init() return self.save(*self.known.values(), full=full, all=all, force=force)
[ "Call", ".", "save", "()", "on", "all", "modified", "entities", "in", "the", "session", ".", "Use", "when", "you", "want", "to", "flush", "changes", "to", "Redis", "but", "don", "t", "want", "to", "lose", "your", "local", "session", "cache", "." ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/util.py#L489-L499
[ "def", "flush", "(", "self", ",", "full", "=", "False", ",", "all", "=", "False", ",", "force", "=", "False", ")", ":", "self", ".", "_init", "(", ")", "return", "self", ".", "save", "(", "*", "self", ".", "known", ".", "values", "(", ")", ",", "full", "=", "full", ",", "all", "=", "all", ",", "force", "=", "force", ")" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
Session.commit
Call ``.save()`` on all modified entities in the session. Also forgets all known entities in the session, so this should only be called at the end of a request. Arguments: * *full* - pass ``True`` to force save full entities, not only changes * *all* - pass ``True`` to save all entities known, not only those entities that have been modified. * *full* - pass ``True`` to force-save all entities known, ignoring DataRaceError and EntityDeletedError exceptions
rom/util.py
def commit(self, full=False, all=False, force=False): ''' Call ``.save()`` on all modified entities in the session. Also forgets all known entities in the session, so this should only be called at the end of a request. Arguments: * *full* - pass ``True`` to force save full entities, not only changes * *all* - pass ``True`` to save all entities known, not only those entities that have been modified. * *full* - pass ``True`` to force-save all entities known, ignoring DataRaceError and EntityDeletedError exceptions ''' changes = self.flush(full, all, force) self.known = {} return changes
def commit(self, full=False, all=False, force=False): ''' Call ``.save()`` on all modified entities in the session. Also forgets all known entities in the session, so this should only be called at the end of a request. Arguments: * *full* - pass ``True`` to force save full entities, not only changes * *all* - pass ``True`` to save all entities known, not only those entities that have been modified. * *full* - pass ``True`` to force-save all entities known, ignoring DataRaceError and EntityDeletedError exceptions ''' changes = self.flush(full, all, force) self.known = {} return changes
[ "Call", ".", "save", "()", "on", "all", "modified", "entities", "in", "the", "session", ".", "Also", "forgets", "all", "known", "entities", "in", "the", "session", "so", "this", "should", "only", "be", "called", "at", "the", "end", "of", "a", "request", "." ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/util.py#L501-L518
[ "def", "commit", "(", "self", ",", "full", "=", "False", ",", "all", "=", "False", ",", "force", "=", "False", ")", ":", "changes", "=", "self", ".", "flush", "(", "full", ",", "all", ",", "force", ")", "self", ".", "known", "=", "{", "}", "return", "changes" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
Session.save
This method is an alternate API for saving many entities (possibly not tracked by the session). You can call:: session.save(obj) session.save(obj1, obj2, ...) session.save([obj1, obj2, ...]) And the entities will be flushed to Redis. You can pass the keyword arguments ``full``, ``all``, and ``force`` with the same meaning and semantics as the ``.commit()`` method.
rom/util.py
def save(self, *objects, **kwargs): ''' This method is an alternate API for saving many entities (possibly not tracked by the session). You can call:: session.save(obj) session.save(obj1, obj2, ...) session.save([obj1, obj2, ...]) And the entities will be flushed to Redis. You can pass the keyword arguments ``full``, ``all``, and ``force`` with the same meaning and semantics as the ``.commit()`` method. ''' from rom import Model full = kwargs.get('full') all = kwargs.get('all') force = kwargs.get('force') changes = 0 items = deque() items.extend(objects) while items: o = items.popleft() if isinstance(o, (list, tuple)): items.extendleft(reversed(o)) elif isinstance(o, Model): if not o._deleted and (all or o._modified): changes += o.save(full, force) else: raise ORMError( "Cannot save an object that is not an instance of a Model (you provided %r)"%( o,)) return changes
def save(self, *objects, **kwargs): ''' This method is an alternate API for saving many entities (possibly not tracked by the session). You can call:: session.save(obj) session.save(obj1, obj2, ...) session.save([obj1, obj2, ...]) And the entities will be flushed to Redis. You can pass the keyword arguments ``full``, ``all``, and ``force`` with the same meaning and semantics as the ``.commit()`` method. ''' from rom import Model full = kwargs.get('full') all = kwargs.get('all') force = kwargs.get('force') changes = 0 items = deque() items.extend(objects) while items: o = items.popleft() if isinstance(o, (list, tuple)): items.extendleft(reversed(o)) elif isinstance(o, Model): if not o._deleted and (all or o._modified): changes += o.save(full, force) else: raise ORMError( "Cannot save an object that is not an instance of a Model (you provided %r)"%( o,)) return changes
[ "This", "method", "is", "an", "alternate", "API", "for", "saving", "many", "entities", "(", "possibly", "not", "tracked", "by", "the", "session", ")", ".", "You", "can", "call", "::" ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/util.py#L520-L554
[ "def", "save", "(", "self", ",", "*", "objects", ",", "*", "*", "kwargs", ")", ":", "from", "rom", "import", "Model", "full", "=", "kwargs", ".", "get", "(", "'full'", ")", "all", "=", "kwargs", ".", "get", "(", "'all'", ")", "force", "=", "kwargs", ".", "get", "(", "'force'", ")", "changes", "=", "0", "items", "=", "deque", "(", ")", "items", ".", "extend", "(", "objects", ")", "while", "items", ":", "o", "=", "items", ".", "popleft", "(", ")", "if", "isinstance", "(", "o", ",", "(", "list", ",", "tuple", ")", ")", ":", "items", ".", "extendleft", "(", "reversed", "(", "o", ")", ")", "elif", "isinstance", "(", "o", ",", "Model", ")", ":", "if", "not", "o", ".", "_deleted", "and", "(", "all", "or", "o", ".", "_modified", ")", ":", "changes", "+=", "o", ".", "save", "(", "full", ",", "force", ")", "else", ":", "raise", "ORMError", "(", "\"Cannot save an object that is not an instance of a Model (you provided %r)\"", "%", "(", "o", ",", ")", ")", "return", "changes" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
Session.delete
This method offers the ability to delete multiple entities in a single round trip to Redis (assuming your models are all stored on the same server). You can call:: session.delete(obj) session.delete(obj1, obj2, ...) session.delete([obj1, obj2, ...]) The keyword argument ``force=True`` can be provided, which can force the deletion of an entitiy again, even if we believe it to already be deleted. If ``force=True``, we won't re-call the object's ``_before_delete()`` method, but we will re-call ``_after_delete()``. .. note:: Objects are automatically dropped from the session after delete for the sake of cache coherency.
rom/util.py
def delete(self, *objects, **kwargs): ''' This method offers the ability to delete multiple entities in a single round trip to Redis (assuming your models are all stored on the same server). You can call:: session.delete(obj) session.delete(obj1, obj2, ...) session.delete([obj1, obj2, ...]) The keyword argument ``force=True`` can be provided, which can force the deletion of an entitiy again, even if we believe it to already be deleted. If ``force=True``, we won't re-call the object's ``_before_delete()`` method, but we will re-call ``_after_delete()``. .. note:: Objects are automatically dropped from the session after delete for the sake of cache coherency. ''' force = kwargs.get('force') from .model import Model, SKIP_ON_DELETE flat = [] items = deque() items.extend(objects) types = set() # flatten what was passed in, more or less arbitrarily deep while items: o = items.popleft() if isinstance(o, (list, tuple)): items.extendleft(reversed(o)) elif isinstance(o, Model): if force or not o._deleted: flat.append(o) types.add(type(o)) # make sure we can bulk delete everything we've been requested to from .columns import MODELS_REFERENCED for t in types: if not t._no_fk or t._namespace in MODELS_REFERENCED: raise ORMError("Can't bulk delete entities of models with foreign key relationships") c2p = {} for o in flat: # prepare delete if not o._deleted: o._before_delete() # make sure we've got connections c = o._connection if c not in c2p: c2p[c] = c.pipeline() # use our existing delete, and pass through a pipeline :P o.delete(_conn=c2p[c], skip_on_delete_i_really_mean_it=SKIP_ON_DELETE) # actually delete the data in Redis for p in c2p.values(): p.execute() # remove the objects from the session forget = self.forget for o in flat: if o._deleted == 1: o._after_delete() o._deleted = 2 forget(o)
def delete(self, *objects, **kwargs): ''' This method offers the ability to delete multiple entities in a single round trip to Redis (assuming your models are all stored on the same server). You can call:: session.delete(obj) session.delete(obj1, obj2, ...) session.delete([obj1, obj2, ...]) The keyword argument ``force=True`` can be provided, which can force the deletion of an entitiy again, even if we believe it to already be deleted. If ``force=True``, we won't re-call the object's ``_before_delete()`` method, but we will re-call ``_after_delete()``. .. note:: Objects are automatically dropped from the session after delete for the sake of cache coherency. ''' force = kwargs.get('force') from .model import Model, SKIP_ON_DELETE flat = [] items = deque() items.extend(objects) types = set() # flatten what was passed in, more or less arbitrarily deep while items: o = items.popleft() if isinstance(o, (list, tuple)): items.extendleft(reversed(o)) elif isinstance(o, Model): if force or not o._deleted: flat.append(o) types.add(type(o)) # make sure we can bulk delete everything we've been requested to from .columns import MODELS_REFERENCED for t in types: if not t._no_fk or t._namespace in MODELS_REFERENCED: raise ORMError("Can't bulk delete entities of models with foreign key relationships") c2p = {} for o in flat: # prepare delete if not o._deleted: o._before_delete() # make sure we've got connections c = o._connection if c not in c2p: c2p[c] = c.pipeline() # use our existing delete, and pass through a pipeline :P o.delete(_conn=c2p[c], skip_on_delete_i_really_mean_it=SKIP_ON_DELETE) # actually delete the data in Redis for p in c2p.values(): p.execute() # remove the objects from the session forget = self.forget for o in flat: if o._deleted == 1: o._after_delete() o._deleted = 2 forget(o)
[ "This", "method", "offers", "the", "ability", "to", "delete", "multiple", "entities", "in", "a", "single", "round", "trip", "to", "Redis", "(", "assuming", "your", "models", "are", "all", "stored", "on", "the", "same", "server", ")", ".", "You", "can", "call", "::" ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/util.py#L556-L625
[ "def", "delete", "(", "self", ",", "*", "objects", ",", "*", "*", "kwargs", ")", ":", "force", "=", "kwargs", ".", "get", "(", "'force'", ")", "from", ".", "model", "import", "Model", ",", "SKIP_ON_DELETE", "flat", "=", "[", "]", "items", "=", "deque", "(", ")", "items", ".", "extend", "(", "objects", ")", "types", "=", "set", "(", ")", "# flatten what was passed in, more or less arbitrarily deep", "while", "items", ":", "o", "=", "items", ".", "popleft", "(", ")", "if", "isinstance", "(", "o", ",", "(", "list", ",", "tuple", ")", ")", ":", "items", ".", "extendleft", "(", "reversed", "(", "o", ")", ")", "elif", "isinstance", "(", "o", ",", "Model", ")", ":", "if", "force", "or", "not", "o", ".", "_deleted", ":", "flat", ".", "append", "(", "o", ")", "types", ".", "add", "(", "type", "(", "o", ")", ")", "# make sure we can bulk delete everything we've been requested to", "from", ".", "columns", "import", "MODELS_REFERENCED", "for", "t", "in", "types", ":", "if", "not", "t", ".", "_no_fk", "or", "t", ".", "_namespace", "in", "MODELS_REFERENCED", ":", "raise", "ORMError", "(", "\"Can't bulk delete entities of models with foreign key relationships\"", ")", "c2p", "=", "{", "}", "for", "o", "in", "flat", ":", "# prepare delete", "if", "not", "o", ".", "_deleted", ":", "o", ".", "_before_delete", "(", ")", "# make sure we've got connections", "c", "=", "o", ".", "_connection", "if", "c", "not", "in", "c2p", ":", "c2p", "[", "c", "]", "=", "c", ".", "pipeline", "(", ")", "# use our existing delete, and pass through a pipeline :P", "o", ".", "delete", "(", "_conn", "=", "c2p", "[", "c", "]", ",", "skip_on_delete_i_really_mean_it", "=", "SKIP_ON_DELETE", ")", "# actually delete the data in Redis", "for", "p", "in", "c2p", ".", "values", "(", ")", ":", "p", ".", "execute", "(", ")", "# remove the objects from the session", "forget", "=", "self", ".", "forget", "for", "o", "in", "flat", ":", "if", "o", ".", "_deleted", "==", "1", ":", "o", ".", "_after_delete", "(", ")", "o", ".", "_deleted", "=", "2", "forget", "(", "o", ")" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
Session.refresh
This method is an alternate API for refreshing many entities (possibly not tracked by the session). You can call:: session.refresh(obj) session.refresh(obj1, obj2, ...) session.refresh([obj1, obj2, ...]) And all provided entities will be reloaded from Redis. To force reloading for modified entities, you can pass ``force=True``.
rom/util.py
def refresh(self, *objects, **kwargs): ''' This method is an alternate API for refreshing many entities (possibly not tracked by the session). You can call:: session.refresh(obj) session.refresh(obj1, obj2, ...) session.refresh([obj1, obj2, ...]) And all provided entities will be reloaded from Redis. To force reloading for modified entities, you can pass ``force=True``. ''' self._init() from rom import Model force = kwargs.get('force') for o in objects: if isinstance(o, (list, tuple)): self.refresh(*o, force=force) elif isinstance(o, Model): if not o._new: o.refresh(force=force) else: # all objects are re-added to the session after refresh, # except for deleted entities... self.add(o) else: raise ORMError( "Cannot refresh an object that is not an instance of a Model (you provided %r)"%( o,))
def refresh(self, *objects, **kwargs): ''' This method is an alternate API for refreshing many entities (possibly not tracked by the session). You can call:: session.refresh(obj) session.refresh(obj1, obj2, ...) session.refresh([obj1, obj2, ...]) And all provided entities will be reloaded from Redis. To force reloading for modified entities, you can pass ``force=True``. ''' self._init() from rom import Model force = kwargs.get('force') for o in objects: if isinstance(o, (list, tuple)): self.refresh(*o, force=force) elif isinstance(o, Model): if not o._new: o.refresh(force=force) else: # all objects are re-added to the session after refresh, # except for deleted entities... self.add(o) else: raise ORMError( "Cannot refresh an object that is not an instance of a Model (you provided %r)"%( o,))
[ "This", "method", "is", "an", "alternate", "API", "for", "refreshing", "many", "entities", "(", "possibly", "not", "tracked", "by", "the", "session", ")", ".", "You", "can", "call", "::" ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/util.py#L627-L656
[ "def", "refresh", "(", "self", ",", "*", "objects", ",", "*", "*", "kwargs", ")", ":", "self", ".", "_init", "(", ")", "from", "rom", "import", "Model", "force", "=", "kwargs", ".", "get", "(", "'force'", ")", "for", "o", "in", "objects", ":", "if", "isinstance", "(", "o", ",", "(", "list", ",", "tuple", ")", ")", ":", "self", ".", "refresh", "(", "*", "o", ",", "force", "=", "force", ")", "elif", "isinstance", "(", "o", ",", "Model", ")", ":", "if", "not", "o", ".", "_new", ":", "o", ".", "refresh", "(", "force", "=", "force", ")", "else", ":", "# all objects are re-added to the session after refresh,", "# except for deleted entities...", "self", ".", "add", "(", "o", ")", "else", ":", "raise", "ORMError", "(", "\"Cannot refresh an object that is not an instance of a Model (you provided %r)\"", "%", "(", "o", ",", ")", ")" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
Session.refresh_all
This method is an alternate API for refreshing all entities tracked by the session. You can call:: session.refresh_all() session.refresh_all(force=True) And all entities known by the session will be reloaded from Redis. To force reloading for modified entities, you can pass ``force=True``.
rom/util.py
def refresh_all(self, *objects, **kwargs): ''' This method is an alternate API for refreshing all entities tracked by the session. You can call:: session.refresh_all() session.refresh_all(force=True) And all entities known by the session will be reloaded from Redis. To force reloading for modified entities, you can pass ``force=True``. ''' self.refresh(*self.known.values(), force=kwargs.get('force'))
def refresh_all(self, *objects, **kwargs): ''' This method is an alternate API for refreshing all entities tracked by the session. You can call:: session.refresh_all() session.refresh_all(force=True) And all entities known by the session will be reloaded from Redis. To force reloading for modified entities, you can pass ``force=True``. ''' self.refresh(*self.known.values(), force=kwargs.get('force'))
[ "This", "method", "is", "an", "alternate", "API", "for", "refreshing", "all", "entities", "tracked", "by", "the", "session", ".", "You", "can", "call", "::" ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/util.py#L658-L670
[ "def", "refresh_all", "(", "self", ",", "*", "objects", ",", "*", "*", "kwargs", ")", ":", "self", ".", "refresh", "(", "*", "self", ".", "known", ".", "values", "(", ")", ",", "force", "=", "kwargs", ".", "get", "(", "'force'", ")", ")" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
redis_writer_lua
... Actually write data to Redis. This is an internal detail. Please don't call me directly.
rom/model.py
def redis_writer_lua(conn, pkey, namespace, id, unique, udelete, delete, data, keys, scored, prefix, suffix, geo, old_data, is_delete): ''' ... Actually write data to Redis. This is an internal detail. Please don't call me directly. ''' ldata = [] for pair in data.items(): ldata.extend(pair) for item in prefix: item.append(_prefix_score(item[-1])) for item in suffix: item.append(_prefix_score(item[-1])) data = [json.dumps(x, default=_fix_bytes) for x in (unique, udelete, delete, ldata, keys, scored, prefix, suffix, geo, is_delete, old_data)] result = _redis_writer_lua(conn, [], [namespace, id] + data) if isinstance(conn, _Pipeline): # we're in a pipelined write situation, don't parse the pipeline :P return if six.PY3: result = result.decode() result = json.loads(result) if 'unique' in result: result = result['unique'] raise UniqueKeyViolation( "Value %r for %s:%s:uidx not distinct (failed for pk=%s)"%( unique[result], namespace, result, id), namespace, id) if 'race' in result: result = result['race'] if pkey in result: raise EntityDeletedError( "Entity %s:%s deleted by another writer; use .save(force=True) to re-save"%( namespace, id), namespace, id) raise DataRaceError( "%s:%s Column(s) %r updated by another writer, write aborted!"%( namespace, id, result), namespace, id)
def redis_writer_lua(conn, pkey, namespace, id, unique, udelete, delete, data, keys, scored, prefix, suffix, geo, old_data, is_delete): ''' ... Actually write data to Redis. This is an internal detail. Please don't call me directly. ''' ldata = [] for pair in data.items(): ldata.extend(pair) for item in prefix: item.append(_prefix_score(item[-1])) for item in suffix: item.append(_prefix_score(item[-1])) data = [json.dumps(x, default=_fix_bytes) for x in (unique, udelete, delete, ldata, keys, scored, prefix, suffix, geo, is_delete, old_data)] result = _redis_writer_lua(conn, [], [namespace, id] + data) if isinstance(conn, _Pipeline): # we're in a pipelined write situation, don't parse the pipeline :P return if six.PY3: result = result.decode() result = json.loads(result) if 'unique' in result: result = result['unique'] raise UniqueKeyViolation( "Value %r for %s:%s:uidx not distinct (failed for pk=%s)"%( unique[result], namespace, result, id), namespace, id) if 'race' in result: result = result['race'] if pkey in result: raise EntityDeletedError( "Entity %s:%s deleted by another writer; use .save(force=True) to re-save"%( namespace, id), namespace, id) raise DataRaceError( "%s:%s Column(s) %r updated by another writer, write aborted!"%( namespace, id, result), namespace, id)
[ "...", "Actually", "write", "data", "to", "Redis", ".", "This", "is", "an", "internal", "detail", ".", "Please", "don", "t", "call", "me", "directly", "." ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/model.py#L859-L904
[ "def", "redis_writer_lua", "(", "conn", ",", "pkey", ",", "namespace", ",", "id", ",", "unique", ",", "udelete", ",", "delete", ",", "data", ",", "keys", ",", "scored", ",", "prefix", ",", "suffix", ",", "geo", ",", "old_data", ",", "is_delete", ")", ":", "ldata", "=", "[", "]", "for", "pair", "in", "data", ".", "items", "(", ")", ":", "ldata", ".", "extend", "(", "pair", ")", "for", "item", "in", "prefix", ":", "item", ".", "append", "(", "_prefix_score", "(", "item", "[", "-", "1", "]", ")", ")", "for", "item", "in", "suffix", ":", "item", ".", "append", "(", "_prefix_score", "(", "item", "[", "-", "1", "]", ")", ")", "data", "=", "[", "json", ".", "dumps", "(", "x", ",", "default", "=", "_fix_bytes", ")", "for", "x", "in", "(", "unique", ",", "udelete", ",", "delete", ",", "ldata", ",", "keys", ",", "scored", ",", "prefix", ",", "suffix", ",", "geo", ",", "is_delete", ",", "old_data", ")", "]", "result", "=", "_redis_writer_lua", "(", "conn", ",", "[", "]", ",", "[", "namespace", ",", "id", "]", "+", "data", ")", "if", "isinstance", "(", "conn", ",", "_Pipeline", ")", ":", "# we're in a pipelined write situation, don't parse the pipeline :P", "return", "if", "six", ".", "PY3", ":", "result", "=", "result", ".", "decode", "(", ")", "result", "=", "json", ".", "loads", "(", "result", ")", "if", "'unique'", "in", "result", ":", "result", "=", "result", "[", "'unique'", "]", "raise", "UniqueKeyViolation", "(", "\"Value %r for %s:%s:uidx not distinct (failed for pk=%s)\"", "%", "(", "unique", "[", "result", "]", ",", "namespace", ",", "result", ",", "id", ")", ",", "namespace", ",", "id", ")", "if", "'race'", "in", "result", ":", "result", "=", "result", "[", "'race'", "]", "if", "pkey", "in", "result", ":", "raise", "EntityDeletedError", "(", "\"Entity %s:%s deleted by another writer; use .save(force=True) to re-save\"", "%", "(", "namespace", ",", "id", ")", ",", "namespace", ",", "id", ")", "raise", "DataRaceError", "(", "\"%s:%s Column(s) %r updated by another writer, write aborted!\"", "%", "(", "namespace", ",", "id", ",", "result", ")", ",", "namespace", ",", "id", ")" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
Model.save
Saves the current entity to Redis. Will only save changed data by default, but you can force a full save by passing ``full=True``. If the underlying entity was deleted and you want to re-save the entity, you can pass ``force=True`` to force a full re-save of the entity.
rom/model.py
def save(self, full=False, force=False): ''' Saves the current entity to Redis. Will only save changed data by default, but you can force a full save by passing ``full=True``. If the underlying entity was deleted and you want to re-save the entity, you can pass ``force=True`` to force a full re-save of the entity. ''' # handle the pre-commit hooks was_new = self._new if was_new: self._before_insert() else: self._before_update() new = self.to_dict() ret, data = self._apply_changes( self._last, new, full or self._new or force, is_new=self._new or force) self._last = data self._new = False self._modified = False self._deleted = False # handle the post-commit hooks if was_new: self._after_insert() else: self._after_update() return ret
def save(self, full=False, force=False): ''' Saves the current entity to Redis. Will only save changed data by default, but you can force a full save by passing ``full=True``. If the underlying entity was deleted and you want to re-save the entity, you can pass ``force=True`` to force a full re-save of the entity. ''' # handle the pre-commit hooks was_new = self._new if was_new: self._before_insert() else: self._before_update() new = self.to_dict() ret, data = self._apply_changes( self._last, new, full or self._new or force, is_new=self._new or force) self._last = data self._new = False self._modified = False self._deleted = False # handle the post-commit hooks if was_new: self._after_insert() else: self._after_update() return ret
[ "Saves", "the", "current", "entity", "to", "Redis", ".", "Will", "only", "save", "changed", "data", "by", "default", "but", "you", "can", "force", "a", "full", "save", "by", "passing", "full", "=", "True", "." ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/model.py#L475-L502
[ "def", "save", "(", "self", ",", "full", "=", "False", ",", "force", "=", "False", ")", ":", "# handle the pre-commit hooks", "was_new", "=", "self", ".", "_new", "if", "was_new", ":", "self", ".", "_before_insert", "(", ")", "else", ":", "self", ".", "_before_update", "(", ")", "new", "=", "self", ".", "to_dict", "(", ")", "ret", ",", "data", "=", "self", ".", "_apply_changes", "(", "self", ".", "_last", ",", "new", ",", "full", "or", "self", ".", "_new", "or", "force", ",", "is_new", "=", "self", ".", "_new", "or", "force", ")", "self", ".", "_last", "=", "data", "self", ".", "_new", "=", "False", "self", ".", "_modified", "=", "False", "self", ".", "_deleted", "=", "False", "# handle the post-commit hooks", "if", "was_new", ":", "self", ".", "_after_insert", "(", ")", "else", ":", "self", ".", "_after_update", "(", ")", "return", "ret" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
Model.delete
Deletes the entity immediately. Also performs any on_delete operations specified as part of column definitions.
rom/model.py
def delete(self, **kwargs): ''' Deletes the entity immediately. Also performs any on_delete operations specified as part of column definitions. ''' if kwargs.get('skip_on_delete_i_really_mean_it') is not SKIP_ON_DELETE: # handle the pre-commit hook self._before_delete() # handle any foreign key references + cascade options _on_delete(self) session.forget(self) self._apply_changes(self._last, {}, delete=True, _conn=kwargs.get('_conn')) self._modified = True self._deleted = True # handle the post-commit hooks if kwargs.get('skip_on_delete_i_really_mean_it') is not SKIP_ON_DELETE: self._after_delete()
def delete(self, **kwargs): ''' Deletes the entity immediately. Also performs any on_delete operations specified as part of column definitions. ''' if kwargs.get('skip_on_delete_i_really_mean_it') is not SKIP_ON_DELETE: # handle the pre-commit hook self._before_delete() # handle any foreign key references + cascade options _on_delete(self) session.forget(self) self._apply_changes(self._last, {}, delete=True, _conn=kwargs.get('_conn')) self._modified = True self._deleted = True # handle the post-commit hooks if kwargs.get('skip_on_delete_i_really_mean_it') is not SKIP_ON_DELETE: self._after_delete()
[ "Deletes", "the", "entity", "immediately", ".", "Also", "performs", "any", "on_delete", "operations", "specified", "as", "part", "of", "column", "definitions", "." ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/model.py#L504-L521
[ "def", "delete", "(", "self", ",", "*", "*", "kwargs", ")", ":", "if", "kwargs", ".", "get", "(", "'skip_on_delete_i_really_mean_it'", ")", "is", "not", "SKIP_ON_DELETE", ":", "# handle the pre-commit hook", "self", ".", "_before_delete", "(", ")", "# handle any foreign key references + cascade options", "_on_delete", "(", "self", ")", "session", ".", "forget", "(", "self", ")", "self", ".", "_apply_changes", "(", "self", ".", "_last", ",", "{", "}", ",", "delete", "=", "True", ",", "_conn", "=", "kwargs", ".", "get", "(", "'_conn'", ")", ")", "self", ".", "_modified", "=", "True", "self", ".", "_deleted", "=", "True", "# handle the post-commit hooks", "if", "kwargs", ".", "get", "(", "'skip_on_delete_i_really_mean_it'", ")", "is", "not", "SKIP_ON_DELETE", ":", "self", ".", "_after_delete", "(", ")" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
Model.copy
Creates a shallow copy of the given entity (any entities that can be retrieved from a OneToMany relationship will not be copied).
rom/model.py
def copy(self): ''' Creates a shallow copy of the given entity (any entities that can be retrieved from a OneToMany relationship will not be copied). ''' x = self.to_dict() x.pop(self._pkey) return self.__class__(**x)
def copy(self): ''' Creates a shallow copy of the given entity (any entities that can be retrieved from a OneToMany relationship will not be copied). ''' x = self.to_dict() x.pop(self._pkey) return self.__class__(**x)
[ "Creates", "a", "shallow", "copy", "of", "the", "given", "entity", "(", "any", "entities", "that", "can", "be", "retrieved", "from", "a", "OneToMany", "relationship", "will", "not", "be", "copied", ")", "." ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/model.py#L523-L530
[ "def", "copy", "(", "self", ")", ":", "x", "=", "self", ".", "to_dict", "(", ")", "x", ".", "pop", "(", "self", ".", "_pkey", ")", "return", "self", ".", "__class__", "(", "*", "*", "x", ")" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
Model.get
Will fetch one or more entities of this type from the session or Redis. Used like:: MyModel.get(5) MyModel.get([1, 6, 2, 4]) Passing a list or a tuple will return multiple entities, in the same order that the ids were passed.
rom/model.py
def get(cls, ids): ''' Will fetch one or more entities of this type from the session or Redis. Used like:: MyModel.get(5) MyModel.get([1, 6, 2, 4]) Passing a list or a tuple will return multiple entities, in the same order that the ids were passed. ''' conn = _connect(cls) # prepare the ids single = not isinstance(ids, (list, tuple, set, frozenset)) if single: ids = [ids] pks = ['%s:%s'%(cls._namespace, id) for id in map(int, ids)] # get from the session, if possible out = list(map(session.get, pks)) # if we couldn't get an instance from the session, load from Redis if None in out: pipe = conn.pipeline(True) idxs = [] # Fetch missing data for i, data in enumerate(out): if data is None: idxs.append(i) pipe.hgetall(pks[i]) # Update output list for i, data in zip(idxs, pipe.execute()): if data: if six.PY3: data = dict((k.decode(), v.decode()) for k, v in data.items()) out[i] = cls(_loading=True, **data) # Get rid of missing models out = [x for x in out if x] if single: return out[0] if out else None return out
def get(cls, ids): ''' Will fetch one or more entities of this type from the session or Redis. Used like:: MyModel.get(5) MyModel.get([1, 6, 2, 4]) Passing a list or a tuple will return multiple entities, in the same order that the ids were passed. ''' conn = _connect(cls) # prepare the ids single = not isinstance(ids, (list, tuple, set, frozenset)) if single: ids = [ids] pks = ['%s:%s'%(cls._namespace, id) for id in map(int, ids)] # get from the session, if possible out = list(map(session.get, pks)) # if we couldn't get an instance from the session, load from Redis if None in out: pipe = conn.pipeline(True) idxs = [] # Fetch missing data for i, data in enumerate(out): if data is None: idxs.append(i) pipe.hgetall(pks[i]) # Update output list for i, data in zip(idxs, pipe.execute()): if data: if six.PY3: data = dict((k.decode(), v.decode()) for k, v in data.items()) out[i] = cls(_loading=True, **data) # Get rid of missing models out = [x for x in out if x] if single: return out[0] if out else None return out
[ "Will", "fetch", "one", "or", "more", "entities", "of", "this", "type", "from", "the", "session", "or", "Redis", "." ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/model.py#L533-L573
[ "def", "get", "(", "cls", ",", "ids", ")", ":", "conn", "=", "_connect", "(", "cls", ")", "# prepare the ids", "single", "=", "not", "isinstance", "(", "ids", ",", "(", "list", ",", "tuple", ",", "set", ",", "frozenset", ")", ")", "if", "single", ":", "ids", "=", "[", "ids", "]", "pks", "=", "[", "'%s:%s'", "%", "(", "cls", ".", "_namespace", ",", "id", ")", "for", "id", "in", "map", "(", "int", ",", "ids", ")", "]", "# get from the session, if possible", "out", "=", "list", "(", "map", "(", "session", ".", "get", ",", "pks", ")", ")", "# if we couldn't get an instance from the session, load from Redis", "if", "None", "in", "out", ":", "pipe", "=", "conn", ".", "pipeline", "(", "True", ")", "idxs", "=", "[", "]", "# Fetch missing data", "for", "i", ",", "data", "in", "enumerate", "(", "out", ")", ":", "if", "data", "is", "None", ":", "idxs", ".", "append", "(", "i", ")", "pipe", ".", "hgetall", "(", "pks", "[", "i", "]", ")", "# Update output list", "for", "i", ",", "data", "in", "zip", "(", "idxs", ",", "pipe", ".", "execute", "(", ")", ")", ":", "if", "data", ":", "if", "six", ".", "PY3", ":", "data", "=", "dict", "(", "(", "k", ".", "decode", "(", ")", ",", "v", ".", "decode", "(", ")", ")", "for", "k", ",", "v", "in", "data", ".", "items", "(", ")", ")", "out", "[", "i", "]", "=", "cls", "(", "_loading", "=", "True", ",", "*", "*", "data", ")", "# Get rid of missing models", "out", "=", "[", "x", "for", "x", "in", "out", "if", "x", "]", "if", "single", ":", "return", "out", "[", "0", "]", "if", "out", "else", "None", "return", "out" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
Model.get_by
This method offers a simple query method for fetching entities of this type via attribute numeric ranges (such columns must be ``indexed``), or via ``unique`` columns. Some examples:: user = User.get_by(email_address='user@domain.com') # gets up to 25 users created in the last 24 hours users = User.get_by( created_at=(time.time()-86400, time.time()), _limit=(0, 25)) Optional keyword-only arguments: * *_limit* - A 2-tuple of (offset, count) that can be used to paginate or otherwise limit results returned by a numeric range query * *_numeric* - An optional boolean defaulting to False that forces the use of a numeric index for ``.get_by(col=val)`` queries even when ``col`` has an existing unique index If you would like to make queries against multiple columns or with multiple criteria, look into the Model.query class property. .. note:: rom will attempt to use a unique index first, then a numeric index if there was no unique index. You can explicitly tell rom to only use the numeric index by using ``.get_by(..., _numeric=True)``. .. note:: Ranged queries with `get_by(col=(start, end))` will only work with columns that use a numeric index.
rom/model.py
def get_by(cls, **kwargs): ''' This method offers a simple query method for fetching entities of this type via attribute numeric ranges (such columns must be ``indexed``), or via ``unique`` columns. Some examples:: user = User.get_by(email_address='user@domain.com') # gets up to 25 users created in the last 24 hours users = User.get_by( created_at=(time.time()-86400, time.time()), _limit=(0, 25)) Optional keyword-only arguments: * *_limit* - A 2-tuple of (offset, count) that can be used to paginate or otherwise limit results returned by a numeric range query * *_numeric* - An optional boolean defaulting to False that forces the use of a numeric index for ``.get_by(col=val)`` queries even when ``col`` has an existing unique index If you would like to make queries against multiple columns or with multiple criteria, look into the Model.query class property. .. note:: rom will attempt to use a unique index first, then a numeric index if there was no unique index. You can explicitly tell rom to only use the numeric index by using ``.get_by(..., _numeric=True)``. .. note:: Ranged queries with `get_by(col=(start, end))` will only work with columns that use a numeric index. ''' conn = _connect(cls) model = cls._namespace # handle limits and query requirements _limit = kwargs.pop('_limit', ()) if _limit and len(_limit) != 2: raise QueryError("Limit must include both 'offset' and 'count' parameters") elif _limit and not all(isinstance(x, six.integer_types) for x in _limit): raise QueryError("Limit arguments must both be integers") if len(kwargs) != 1: raise QueryError("We can only fetch object(s) by exactly one attribute, you provided %s"%(len(kwargs),)) _numeric = bool(kwargs.pop('_numeric', None)) for attr, value in kwargs.items(): plain_attr = attr.partition(':')[0] if isinstance(value, tuple) and len(value) != 2: raise QueryError("Range queries must include exactly two endpoints") # handle unique index lookups if attr in cls._unique and (plain_attr not in cls._index or not _numeric): if isinstance(value, tuple): raise QueryError("Cannot query a unique index with a range of values") single = not isinstance(value, list) if single: value = [value] qvalues = list(map(cls._columns[attr]._to_redis, value)) ids = [x for x in conn.hmget('%s:%s:uidx'%(model, attr), qvalues) if x] if not ids: return None if single else [] return cls.get(ids[0] if single else ids) if plain_attr not in cls._index: raise QueryError("Cannot query on a column without an index") if isinstance(value, NUMERIC_TYPES) and not isinstance(value, bool): value = (value, value) if isinstance(value, tuple): # this is a numeric range query, we'll just pull it directly args = list(value) for i, a in enumerate(args): # Handle the ranges where None is -inf on the left and inf # on the right when used in the context of a range tuple. args[i] = ('-inf', 'inf')[i] if a is None else cls._columns[attr]._to_redis(a) if _limit: args.extend(_limit) ids = conn.zrangebyscore('%s:%s:idx'%(model, attr), *args) if not ids: return [] return cls.get(ids) # defer other index lookups to the query object query = cls.query.filter(**{attr: value}) if _limit: query = query.limit(*_limit) return query.all()
def get_by(cls, **kwargs): ''' This method offers a simple query method for fetching entities of this type via attribute numeric ranges (such columns must be ``indexed``), or via ``unique`` columns. Some examples:: user = User.get_by(email_address='user@domain.com') # gets up to 25 users created in the last 24 hours users = User.get_by( created_at=(time.time()-86400, time.time()), _limit=(0, 25)) Optional keyword-only arguments: * *_limit* - A 2-tuple of (offset, count) that can be used to paginate or otherwise limit results returned by a numeric range query * *_numeric* - An optional boolean defaulting to False that forces the use of a numeric index for ``.get_by(col=val)`` queries even when ``col`` has an existing unique index If you would like to make queries against multiple columns or with multiple criteria, look into the Model.query class property. .. note:: rom will attempt to use a unique index first, then a numeric index if there was no unique index. You can explicitly tell rom to only use the numeric index by using ``.get_by(..., _numeric=True)``. .. note:: Ranged queries with `get_by(col=(start, end))` will only work with columns that use a numeric index. ''' conn = _connect(cls) model = cls._namespace # handle limits and query requirements _limit = kwargs.pop('_limit', ()) if _limit and len(_limit) != 2: raise QueryError("Limit must include both 'offset' and 'count' parameters") elif _limit and not all(isinstance(x, six.integer_types) for x in _limit): raise QueryError("Limit arguments must both be integers") if len(kwargs) != 1: raise QueryError("We can only fetch object(s) by exactly one attribute, you provided %s"%(len(kwargs),)) _numeric = bool(kwargs.pop('_numeric', None)) for attr, value in kwargs.items(): plain_attr = attr.partition(':')[0] if isinstance(value, tuple) and len(value) != 2: raise QueryError("Range queries must include exactly two endpoints") # handle unique index lookups if attr in cls._unique and (plain_attr not in cls._index or not _numeric): if isinstance(value, tuple): raise QueryError("Cannot query a unique index with a range of values") single = not isinstance(value, list) if single: value = [value] qvalues = list(map(cls._columns[attr]._to_redis, value)) ids = [x for x in conn.hmget('%s:%s:uidx'%(model, attr), qvalues) if x] if not ids: return None if single else [] return cls.get(ids[0] if single else ids) if plain_attr not in cls._index: raise QueryError("Cannot query on a column without an index") if isinstance(value, NUMERIC_TYPES) and not isinstance(value, bool): value = (value, value) if isinstance(value, tuple): # this is a numeric range query, we'll just pull it directly args = list(value) for i, a in enumerate(args): # Handle the ranges where None is -inf on the left and inf # on the right when used in the context of a range tuple. args[i] = ('-inf', 'inf')[i] if a is None else cls._columns[attr]._to_redis(a) if _limit: args.extend(_limit) ids = conn.zrangebyscore('%s:%s:idx'%(model, attr), *args) if not ids: return [] return cls.get(ids) # defer other index lookups to the query object query = cls.query.filter(**{attr: value}) if _limit: query = query.limit(*_limit) return query.all()
[ "This", "method", "offers", "a", "simple", "query", "method", "for", "fetching", "entities", "of", "this", "type", "via", "attribute", "numeric", "ranges", "(", "such", "columns", "must", "be", "indexed", ")", "or", "via", "unique", "columns", "." ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/model.py#L576-L663
[ "def", "get_by", "(", "cls", ",", "*", "*", "kwargs", ")", ":", "conn", "=", "_connect", "(", "cls", ")", "model", "=", "cls", ".", "_namespace", "# handle limits and query requirements", "_limit", "=", "kwargs", ".", "pop", "(", "'_limit'", ",", "(", ")", ")", "if", "_limit", "and", "len", "(", "_limit", ")", "!=", "2", ":", "raise", "QueryError", "(", "\"Limit must include both 'offset' and 'count' parameters\"", ")", "elif", "_limit", "and", "not", "all", "(", "isinstance", "(", "x", ",", "six", ".", "integer_types", ")", "for", "x", "in", "_limit", ")", ":", "raise", "QueryError", "(", "\"Limit arguments must both be integers\"", ")", "if", "len", "(", "kwargs", ")", "!=", "1", ":", "raise", "QueryError", "(", "\"We can only fetch object(s) by exactly one attribute, you provided %s\"", "%", "(", "len", "(", "kwargs", ")", ",", ")", ")", "_numeric", "=", "bool", "(", "kwargs", ".", "pop", "(", "'_numeric'", ",", "None", ")", ")", "for", "attr", ",", "value", "in", "kwargs", ".", "items", "(", ")", ":", "plain_attr", "=", "attr", ".", "partition", "(", "':'", ")", "[", "0", "]", "if", "isinstance", "(", "value", ",", "tuple", ")", "and", "len", "(", "value", ")", "!=", "2", ":", "raise", "QueryError", "(", "\"Range queries must include exactly two endpoints\"", ")", "# handle unique index lookups", "if", "attr", "in", "cls", ".", "_unique", "and", "(", "plain_attr", "not", "in", "cls", ".", "_index", "or", "not", "_numeric", ")", ":", "if", "isinstance", "(", "value", ",", "tuple", ")", ":", "raise", "QueryError", "(", "\"Cannot query a unique index with a range of values\"", ")", "single", "=", "not", "isinstance", "(", "value", ",", "list", ")", "if", "single", ":", "value", "=", "[", "value", "]", "qvalues", "=", "list", "(", "map", "(", "cls", ".", "_columns", "[", "attr", "]", ".", "_to_redis", ",", "value", ")", ")", "ids", "=", "[", "x", "for", "x", "in", "conn", ".", "hmget", "(", "'%s:%s:uidx'", "%", "(", "model", ",", "attr", ")", ",", "qvalues", ")", "if", "x", "]", "if", "not", "ids", ":", "return", "None", "if", "single", "else", "[", "]", "return", "cls", ".", "get", "(", "ids", "[", "0", "]", "if", "single", "else", "ids", ")", "if", "plain_attr", "not", "in", "cls", ".", "_index", ":", "raise", "QueryError", "(", "\"Cannot query on a column without an index\"", ")", "if", "isinstance", "(", "value", ",", "NUMERIC_TYPES", ")", "and", "not", "isinstance", "(", "value", ",", "bool", ")", ":", "value", "=", "(", "value", ",", "value", ")", "if", "isinstance", "(", "value", ",", "tuple", ")", ":", "# this is a numeric range query, we'll just pull it directly", "args", "=", "list", "(", "value", ")", "for", "i", ",", "a", "in", "enumerate", "(", "args", ")", ":", "# Handle the ranges where None is -inf on the left and inf", "# on the right when used in the context of a range tuple.", "args", "[", "i", "]", "=", "(", "'-inf'", ",", "'inf'", ")", "[", "i", "]", "if", "a", "is", "None", "else", "cls", ".", "_columns", "[", "attr", "]", ".", "_to_redis", "(", "a", ")", "if", "_limit", ":", "args", ".", "extend", "(", "_limit", ")", "ids", "=", "conn", ".", "zrangebyscore", "(", "'%s:%s:idx'", "%", "(", "model", ",", "attr", ")", ",", "*", "args", ")", "if", "not", "ids", ":", "return", "[", "]", "return", "cls", ".", "get", "(", "ids", ")", "# defer other index lookups to the query object", "query", "=", "cls", ".", "query", ".", "filter", "(", "*", "*", "{", "attr", ":", "value", "}", ")", "if", "_limit", ":", "query", "=", "query", ".", "limit", "(", "*", "_limit", ")", "return", "query", ".", "all", "(", ")" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
Model.update
Updates multiple attributes in a model. If ``args`` are provided, this method will assign attributes in the order returned by ``list(self._columns)`` until one or both are exhausted. If ``kwargs`` are provided, this method will assign attributes to the names provided, after ``args`` have been processed.
rom/model.py
def update(self, *args, **kwargs): ''' Updates multiple attributes in a model. If ``args`` are provided, this method will assign attributes in the order returned by ``list(self._columns)`` until one or both are exhausted. If ``kwargs`` are provided, this method will assign attributes to the names provided, after ``args`` have been processed. ''' sa = setattr for a, v in zip(self._columns, args): sa(self, a, v) for a, v in kwargs.items(): sa(self, a, v) return self
def update(self, *args, **kwargs): ''' Updates multiple attributes in a model. If ``args`` are provided, this method will assign attributes in the order returned by ``list(self._columns)`` until one or both are exhausted. If ``kwargs`` are provided, this method will assign attributes to the names provided, after ``args`` have been processed. ''' sa = setattr for a, v in zip(self._columns, args): sa(self, a, v) for a, v in kwargs.items(): sa(self, a, v) return self
[ "Updates", "multiple", "attributes", "in", "a", "model", ".", "If", "args", "are", "provided", "this", "method", "will", "assign", "attributes", "in", "the", "order", "returned", "by", "list", "(", "self", ".", "_columns", ")", "until", "one", "or", "both", "are", "exhausted", "." ]
josiahcarlson/rom
python
https://github.com/josiahcarlson/rom/blob/8b5607a856341df85df33422accc30ba9294dbdb/rom/model.py#L673-L687
[ "def", "update", "(", "self", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "sa", "=", "setattr", "for", "a", ",", "v", "in", "zip", "(", "self", ".", "_columns", ",", "args", ")", ":", "sa", "(", "self", ",", "a", ",", "v", ")", "for", "a", ",", "v", "in", "kwargs", ".", "items", "(", ")", ":", "sa", "(", "self", ",", "a", ",", "v", ")", "return", "self" ]
8b5607a856341df85df33422accc30ba9294dbdb
test
dump
Replacement for pickle.dump() using _LokyPickler.
loky/backend/reduction.py
def dump(obj, file, reducers=None, protocol=None): '''Replacement for pickle.dump() using _LokyPickler.''' global _LokyPickler _LokyPickler(file, reducers=reducers, protocol=protocol).dump(obj)
def dump(obj, file, reducers=None, protocol=None): '''Replacement for pickle.dump() using _LokyPickler.''' global _LokyPickler _LokyPickler(file, reducers=reducers, protocol=protocol).dump(obj)
[ "Replacement", "for", "pickle", ".", "dump", "()", "using", "_LokyPickler", "." ]
tomMoral/loky
python
https://github.com/tomMoral/loky/blob/dc2d941d8285a96f3a5b666a4bd04875b0b25984/loky/backend/reduction.py#L237-L240
[ "def", "dump", "(", "obj", ",", "file", ",", "reducers", "=", "None", ",", "protocol", "=", "None", ")", ":", "global", "_LokyPickler", "_LokyPickler", "(", "file", ",", "reducers", "=", "reducers", ",", "protocol", "=", "protocol", ")", ".", "dump", "(", "obj", ")" ]
dc2d941d8285a96f3a5b666a4bd04875b0b25984
test
_ReducerRegistry.register
Attach a reducer function to a given type in the dispatch table.
loky/backend/reduction.py
def register(cls, type, reduce_func): """Attach a reducer function to a given type in the dispatch table.""" if sys.version_info < (3,): # Python 2 pickler dispatching is not explicitly customizable. # Let us use a closure to workaround this limitation. def dispatcher(cls, obj): reduced = reduce_func(obj) cls.save_reduce(obj=obj, *reduced) cls.dispatch_table[type] = dispatcher else: cls.dispatch_table[type] = reduce_func
def register(cls, type, reduce_func): """Attach a reducer function to a given type in the dispatch table.""" if sys.version_info < (3,): # Python 2 pickler dispatching is not explicitly customizable. # Let us use a closure to workaround this limitation. def dispatcher(cls, obj): reduced = reduce_func(obj) cls.save_reduce(obj=obj, *reduced) cls.dispatch_table[type] = dispatcher else: cls.dispatch_table[type] = reduce_func
[ "Attach", "a", "reducer", "function", "to", "a", "given", "type", "in", "the", "dispatch", "table", "." ]
tomMoral/loky
python
https://github.com/tomMoral/loky/blob/dc2d941d8285a96f3a5b666a4bd04875b0b25984/loky/backend/reduction.py#L58-L68
[ "def", "register", "(", "cls", ",", "type", ",", "reduce_func", ")", ":", "if", "sys", ".", "version_info", "<", "(", "3", ",", ")", ":", "# Python 2 pickler dispatching is not explicitly customizable.", "# Let us use a closure to workaround this limitation.", "def", "dispatcher", "(", "cls", ",", "obj", ")", ":", "reduced", "=", "reduce_func", "(", "obj", ")", "cls", ".", "save_reduce", "(", "obj", "=", "obj", ",", "*", "reduced", ")", "cls", ".", "dispatch_table", "[", "type", "]", "=", "dispatcher", "else", ":", "cls", ".", "dispatch_table", "[", "type", "]", "=", "reduce_func" ]
dc2d941d8285a96f3a5b666a4bd04875b0b25984
test
_sem_open
Construct or retrieve a semaphore with the given name If value is None, try to retrieve an existing named semaphore. Else create a new semaphore with the given value
loky/backend/semlock.py
def _sem_open(name, value=None): """ Construct or retrieve a semaphore with the given name If value is None, try to retrieve an existing named semaphore. Else create a new semaphore with the given value """ if value is None: handle = pthread.sem_open(ctypes.c_char_p(name), 0) else: handle = pthread.sem_open(ctypes.c_char_p(name), SEM_OFLAG, SEM_PERM, ctypes.c_int(value)) if handle == SEM_FAILURE: e = ctypes.get_errno() if e == errno.EEXIST: raise FileExistsError("a semaphore named %s already exists" % name) elif e == errno.ENOENT: raise FileNotFoundError('cannot find semaphore named %s' % name) elif e == errno.ENOSYS: raise NotImplementedError('No semaphore implementation on this ' 'system') else: raiseFromErrno() return handle
def _sem_open(name, value=None): """ Construct or retrieve a semaphore with the given name If value is None, try to retrieve an existing named semaphore. Else create a new semaphore with the given value """ if value is None: handle = pthread.sem_open(ctypes.c_char_p(name), 0) else: handle = pthread.sem_open(ctypes.c_char_p(name), SEM_OFLAG, SEM_PERM, ctypes.c_int(value)) if handle == SEM_FAILURE: e = ctypes.get_errno() if e == errno.EEXIST: raise FileExistsError("a semaphore named %s already exists" % name) elif e == errno.ENOENT: raise FileNotFoundError('cannot find semaphore named %s' % name) elif e == errno.ENOSYS: raise NotImplementedError('No semaphore implementation on this ' 'system') else: raiseFromErrno() return handle
[ "Construct", "or", "retrieve", "a", "semaphore", "with", "the", "given", "name" ]
tomMoral/loky
python
https://github.com/tomMoral/loky/blob/dc2d941d8285a96f3a5b666a4bd04875b0b25984/loky/backend/semlock.py#L75-L99
[ "def", "_sem_open", "(", "name", ",", "value", "=", "None", ")", ":", "if", "value", "is", "None", ":", "handle", "=", "pthread", ".", "sem_open", "(", "ctypes", ".", "c_char_p", "(", "name", ")", ",", "0", ")", "else", ":", "handle", "=", "pthread", ".", "sem_open", "(", "ctypes", ".", "c_char_p", "(", "name", ")", ",", "SEM_OFLAG", ",", "SEM_PERM", ",", "ctypes", ".", "c_int", "(", "value", ")", ")", "if", "handle", "==", "SEM_FAILURE", ":", "e", "=", "ctypes", ".", "get_errno", "(", ")", "if", "e", "==", "errno", ".", "EEXIST", ":", "raise", "FileExistsError", "(", "\"a semaphore named %s already exists\"", "%", "name", ")", "elif", "e", "==", "errno", ".", "ENOENT", ":", "raise", "FileNotFoundError", "(", "'cannot find semaphore named %s'", "%", "name", ")", "elif", "e", "==", "errno", ".", "ENOSYS", ":", "raise", "NotImplementedError", "(", "'No semaphore implementation on this '", "'system'", ")", "else", ":", "raiseFromErrno", "(", ")", "return", "handle" ]
dc2d941d8285a96f3a5b666a4bd04875b0b25984
test
cpu_count
Return the number of CPUs the current process can use. The returned number of CPUs accounts for: * the number of CPUs in the system, as given by ``multiprocessing.cpu_count``; * the CPU affinity settings of the current process (available with Python 3.4+ on some Unix systems); * CFS scheduler CPU bandwidth limit (available on Linux only, typically set by docker and similar container orchestration systems); * the value of the LOKY_MAX_CPU_COUNT environment variable if defined. and is given as the minimum of these constraints. It is also always larger or equal to 1.
loky/backend/context.py
def cpu_count(): """Return the number of CPUs the current process can use. The returned number of CPUs accounts for: * the number of CPUs in the system, as given by ``multiprocessing.cpu_count``; * the CPU affinity settings of the current process (available with Python 3.4+ on some Unix systems); * CFS scheduler CPU bandwidth limit (available on Linux only, typically set by docker and similar container orchestration systems); * the value of the LOKY_MAX_CPU_COUNT environment variable if defined. and is given as the minimum of these constraints. It is also always larger or equal to 1. """ import math try: cpu_count_mp = mp.cpu_count() except NotImplementedError: cpu_count_mp = 1 # Number of available CPUs given affinity settings cpu_count_affinity = cpu_count_mp if hasattr(os, 'sched_getaffinity'): try: cpu_count_affinity = len(os.sched_getaffinity(0)) except NotImplementedError: pass # CFS scheduler CPU bandwidth limit # available in Linux since 2.6 kernel cpu_count_cfs = cpu_count_mp cfs_quota_fname = "/sys/fs/cgroup/cpu/cpu.cfs_quota_us" cfs_period_fname = "/sys/fs/cgroup/cpu/cpu.cfs_period_us" if os.path.exists(cfs_quota_fname) and os.path.exists(cfs_period_fname): with open(cfs_quota_fname, 'r') as fh: cfs_quota_us = int(fh.read()) with open(cfs_period_fname, 'r') as fh: cfs_period_us = int(fh.read()) if cfs_quota_us > 0 and cfs_period_us > 0: # Make sure this quantity is an int as math.ceil returns a # float in python2.7. (See issue #165) cpu_count_cfs = int(math.ceil(cfs_quota_us / cfs_period_us)) # User defined soft-limit passed as an loky specific environment variable. cpu_count_loky = int(os.environ.get('LOKY_MAX_CPU_COUNT', cpu_count_mp)) aggregate_cpu_count = min(cpu_count_mp, cpu_count_affinity, cpu_count_cfs, cpu_count_loky) return max(aggregate_cpu_count, 1)
def cpu_count(): """Return the number of CPUs the current process can use. The returned number of CPUs accounts for: * the number of CPUs in the system, as given by ``multiprocessing.cpu_count``; * the CPU affinity settings of the current process (available with Python 3.4+ on some Unix systems); * CFS scheduler CPU bandwidth limit (available on Linux only, typically set by docker and similar container orchestration systems); * the value of the LOKY_MAX_CPU_COUNT environment variable if defined. and is given as the minimum of these constraints. It is also always larger or equal to 1. """ import math try: cpu_count_mp = mp.cpu_count() except NotImplementedError: cpu_count_mp = 1 # Number of available CPUs given affinity settings cpu_count_affinity = cpu_count_mp if hasattr(os, 'sched_getaffinity'): try: cpu_count_affinity = len(os.sched_getaffinity(0)) except NotImplementedError: pass # CFS scheduler CPU bandwidth limit # available in Linux since 2.6 kernel cpu_count_cfs = cpu_count_mp cfs_quota_fname = "/sys/fs/cgroup/cpu/cpu.cfs_quota_us" cfs_period_fname = "/sys/fs/cgroup/cpu/cpu.cfs_period_us" if os.path.exists(cfs_quota_fname) and os.path.exists(cfs_period_fname): with open(cfs_quota_fname, 'r') as fh: cfs_quota_us = int(fh.read()) with open(cfs_period_fname, 'r') as fh: cfs_period_us = int(fh.read()) if cfs_quota_us > 0 and cfs_period_us > 0: # Make sure this quantity is an int as math.ceil returns a # float in python2.7. (See issue #165) cpu_count_cfs = int(math.ceil(cfs_quota_us / cfs_period_us)) # User defined soft-limit passed as an loky specific environment variable. cpu_count_loky = int(os.environ.get('LOKY_MAX_CPU_COUNT', cpu_count_mp)) aggregate_cpu_count = min(cpu_count_mp, cpu_count_affinity, cpu_count_cfs, cpu_count_loky) return max(aggregate_cpu_count, 1)
[ "Return", "the", "number", "of", "CPUs", "the", "current", "process", "can", "use", "." ]
tomMoral/loky
python
https://github.com/tomMoral/loky/blob/dc2d941d8285a96f3a5b666a4bd04875b0b25984/loky/backend/context.py#L104-L153
[ "def", "cpu_count", "(", ")", ":", "import", "math", "try", ":", "cpu_count_mp", "=", "mp", ".", "cpu_count", "(", ")", "except", "NotImplementedError", ":", "cpu_count_mp", "=", "1", "# Number of available CPUs given affinity settings", "cpu_count_affinity", "=", "cpu_count_mp", "if", "hasattr", "(", "os", ",", "'sched_getaffinity'", ")", ":", "try", ":", "cpu_count_affinity", "=", "len", "(", "os", ".", "sched_getaffinity", "(", "0", ")", ")", "except", "NotImplementedError", ":", "pass", "# CFS scheduler CPU bandwidth limit", "# available in Linux since 2.6 kernel", "cpu_count_cfs", "=", "cpu_count_mp", "cfs_quota_fname", "=", "\"/sys/fs/cgroup/cpu/cpu.cfs_quota_us\"", "cfs_period_fname", "=", "\"/sys/fs/cgroup/cpu/cpu.cfs_period_us\"", "if", "os", ".", "path", ".", "exists", "(", "cfs_quota_fname", ")", "and", "os", ".", "path", ".", "exists", "(", "cfs_period_fname", ")", ":", "with", "open", "(", "cfs_quota_fname", ",", "'r'", ")", "as", "fh", ":", "cfs_quota_us", "=", "int", "(", "fh", ".", "read", "(", ")", ")", "with", "open", "(", "cfs_period_fname", ",", "'r'", ")", "as", "fh", ":", "cfs_period_us", "=", "int", "(", "fh", ".", "read", "(", ")", ")", "if", "cfs_quota_us", ">", "0", "and", "cfs_period_us", ">", "0", ":", "# Make sure this quantity is an int as math.ceil returns a", "# float in python2.7. (See issue #165)", "cpu_count_cfs", "=", "int", "(", "math", ".", "ceil", "(", "cfs_quota_us", "/", "cfs_period_us", ")", ")", "# User defined soft-limit passed as an loky specific environment variable.", "cpu_count_loky", "=", "int", "(", "os", ".", "environ", ".", "get", "(", "'LOKY_MAX_CPU_COUNT'", ",", "cpu_count_mp", ")", ")", "aggregate_cpu_count", "=", "min", "(", "cpu_count_mp", ",", "cpu_count_affinity", ",", "cpu_count_cfs", ",", "cpu_count_loky", ")", "return", "max", "(", "aggregate_cpu_count", ",", "1", ")" ]
dc2d941d8285a96f3a5b666a4bd04875b0b25984
test
LokyContext.Queue
Returns a queue object
loky/backend/context.py
def Queue(self, maxsize=0, reducers=None): '''Returns a queue object''' from .queues import Queue return Queue(maxsize, reducers=reducers, ctx=self.get_context())
def Queue(self, maxsize=0, reducers=None): '''Returns a queue object''' from .queues import Queue return Queue(maxsize, reducers=reducers, ctx=self.get_context())
[ "Returns", "a", "queue", "object" ]
tomMoral/loky
python
https://github.com/tomMoral/loky/blob/dc2d941d8285a96f3a5b666a4bd04875b0b25984/loky/backend/context.py#L162-L166
[ "def", "Queue", "(", "self", ",", "maxsize", "=", "0", ",", "reducers", "=", "None", ")", ":", "from", ".", "queues", "import", "Queue", "return", "Queue", "(", "maxsize", ",", "reducers", "=", "reducers", ",", "ctx", "=", "self", ".", "get_context", "(", ")", ")" ]
dc2d941d8285a96f3a5b666a4bd04875b0b25984
test
LokyContext.SimpleQueue
Returns a queue object
loky/backend/context.py
def SimpleQueue(self, reducers=None): '''Returns a queue object''' from .queues import SimpleQueue return SimpleQueue(reducers=reducers, ctx=self.get_context())
def SimpleQueue(self, reducers=None): '''Returns a queue object''' from .queues import SimpleQueue return SimpleQueue(reducers=reducers, ctx=self.get_context())
[ "Returns", "a", "queue", "object" ]
tomMoral/loky
python
https://github.com/tomMoral/loky/blob/dc2d941d8285a96f3a5b666a4bd04875b0b25984/loky/backend/context.py#L168-L171
[ "def", "SimpleQueue", "(", "self", ",", "reducers", "=", "None", ")", ":", "from", ".", "queues", "import", "SimpleQueue", "return", "SimpleQueue", "(", "reducers", "=", "reducers", ",", "ctx", "=", "self", ".", "get_context", "(", ")", ")" ]
dc2d941d8285a96f3a5b666a4bd04875b0b25984
test
_get_chunks
Iterates over zip()ed iterables in chunks.
loky/process_executor.py
def _get_chunks(chunksize, *iterables): """Iterates over zip()ed iterables in chunks. """ if sys.version_info < (3, 3): it = itertools.izip(*iterables) else: it = zip(*iterables) while True: chunk = tuple(itertools.islice(it, chunksize)) if not chunk: return yield chunk
def _get_chunks(chunksize, *iterables): """Iterates over zip()ed iterables in chunks. """ if sys.version_info < (3, 3): it = itertools.izip(*iterables) else: it = zip(*iterables) while True: chunk = tuple(itertools.islice(it, chunksize)) if not chunk: return yield chunk
[ "Iterates", "over", "zip", "()", "ed", "iterables", "in", "chunks", "." ]
tomMoral/loky
python
https://github.com/tomMoral/loky/blob/dc2d941d8285a96f3a5b666a4bd04875b0b25984/loky/process_executor.py#L315-L325
[ "def", "_get_chunks", "(", "chunksize", ",", "*", "iterables", ")", ":", "if", "sys", ".", "version_info", "<", "(", "3", ",", "3", ")", ":", "it", "=", "itertools", ".", "izip", "(", "*", "iterables", ")", "else", ":", "it", "=", "zip", "(", "*", "iterables", ")", "while", "True", ":", "chunk", "=", "tuple", "(", "itertools", ".", "islice", "(", "it", ",", "chunksize", ")", ")", "if", "not", "chunk", ":", "return", "yield", "chunk" ]
dc2d941d8285a96f3a5b666a4bd04875b0b25984
test
_sendback_result
Safely send back the given result or exception
loky/process_executor.py
def _sendback_result(result_queue, work_id, result=None, exception=None): """Safely send back the given result or exception""" try: result_queue.put(_ResultItem(work_id, result=result, exception=exception)) except BaseException as e: exc = _ExceptionWithTraceback(e) result_queue.put(_ResultItem(work_id, exception=exc))
def _sendback_result(result_queue, work_id, result=None, exception=None): """Safely send back the given result or exception""" try: result_queue.put(_ResultItem(work_id, result=result, exception=exception)) except BaseException as e: exc = _ExceptionWithTraceback(e) result_queue.put(_ResultItem(work_id, exception=exc))
[ "Safely", "send", "back", "the", "given", "result", "or", "exception" ]
tomMoral/loky
python
https://github.com/tomMoral/loky/blob/dc2d941d8285a96f3a5b666a4bd04875b0b25984/loky/process_executor.py#L340-L347
[ "def", "_sendback_result", "(", "result_queue", ",", "work_id", ",", "result", "=", "None", ",", "exception", "=", "None", ")", ":", "try", ":", "result_queue", ".", "put", "(", "_ResultItem", "(", "work_id", ",", "result", "=", "result", ",", "exception", "=", "exception", ")", ")", "except", "BaseException", "as", "e", ":", "exc", "=", "_ExceptionWithTraceback", "(", "e", ")", "result_queue", ".", "put", "(", "_ResultItem", "(", "work_id", ",", "exception", "=", "exc", ")", ")" ]
dc2d941d8285a96f3a5b666a4bd04875b0b25984
test
_process_worker
Evaluates calls from call_queue and places the results in result_queue. This worker is run in a separate process. Args: call_queue: A ctx.Queue of _CallItems that will be read and evaluated by the worker. result_queue: A ctx.Queue of _ResultItems that will written to by the worker. initializer: A callable initializer, or None initargs: A tuple of args for the initializer process_management_lock: A ctx.Lock avoiding worker timeout while some workers are being spawned. timeout: maximum time to wait for a new item in the call_queue. If that time is expired, the worker will shutdown. worker_exit_lock: Lock to avoid flagging the executor as broken on workers timeout. current_depth: Nested parallelism level, to avoid infinite spawning.
loky/process_executor.py
def _process_worker(call_queue, result_queue, initializer, initargs, processes_management_lock, timeout, worker_exit_lock, current_depth): """Evaluates calls from call_queue and places the results in result_queue. This worker is run in a separate process. Args: call_queue: A ctx.Queue of _CallItems that will be read and evaluated by the worker. result_queue: A ctx.Queue of _ResultItems that will written to by the worker. initializer: A callable initializer, or None initargs: A tuple of args for the initializer process_management_lock: A ctx.Lock avoiding worker timeout while some workers are being spawned. timeout: maximum time to wait for a new item in the call_queue. If that time is expired, the worker will shutdown. worker_exit_lock: Lock to avoid flagging the executor as broken on workers timeout. current_depth: Nested parallelism level, to avoid infinite spawning. """ if initializer is not None: try: initializer(*initargs) except BaseException: _base.LOGGER.critical('Exception in initializer:', exc_info=True) # The parent will notice that the process stopped and # mark the pool broken return # set the global _CURRENT_DEPTH mechanism to limit recursive call global _CURRENT_DEPTH _CURRENT_DEPTH = current_depth _process_reference_size = None _last_memory_leak_check = None pid = os.getpid() mp.util.debug('Worker started with timeout=%s' % timeout) while True: try: call_item = call_queue.get(block=True, timeout=timeout) if call_item is None: mp.util.info("Shutting down worker on sentinel") except queue.Empty: mp.util.info("Shutting down worker after timeout %0.3fs" % timeout) if processes_management_lock.acquire(block=False): processes_management_lock.release() call_item = None else: mp.util.info("Could not acquire processes_management_lock") continue except BaseException as e: previous_tb = traceback.format_exc() try: result_queue.put(_RemoteTraceback(previous_tb)) except BaseException: # If we cannot format correctly the exception, at least print # the traceback. print(previous_tb) sys.exit(1) if call_item is None: # Notify queue management thread about clean worker shutdown result_queue.put(pid) with worker_exit_lock: return try: r = call_item() except BaseException as e: exc = _ExceptionWithTraceback(e) result_queue.put(_ResultItem(call_item.work_id, exception=exc)) else: _sendback_result(result_queue, call_item.work_id, result=r) del r # Free the resource as soon as possible, to avoid holding onto # open files or shared memory that is not needed anymore del call_item if _USE_PSUTIL: if _process_reference_size is None: # Make reference measurement after the first call _process_reference_size = _get_memory_usage(pid, force_gc=True) _last_memory_leak_check = time() continue if time() - _last_memory_leak_check > _MEMORY_LEAK_CHECK_DELAY: mem_usage = _get_memory_usage(pid) _last_memory_leak_check = time() if mem_usage - _process_reference_size < _MAX_MEMORY_LEAK_SIZE: # Memory usage stays within bounds: everything is fine. continue # Check again memory usage; this time take the measurement # after a forced garbage collection to break any reference # cycles. mem_usage = _get_memory_usage(pid, force_gc=True) _last_memory_leak_check = time() if mem_usage - _process_reference_size < _MAX_MEMORY_LEAK_SIZE: # The GC managed to free the memory: everything is fine. continue # The process is leaking memory: let the master process # know that we need to start a new worker. mp.util.info("Memory leak detected: shutting down worker") result_queue.put(pid) with worker_exit_lock: return else: # if psutil is not installed, trigger gc.collect events # regularly to limit potential memory leaks due to reference cycles if ((_last_memory_leak_check is None) or (time() - _last_memory_leak_check > _MEMORY_LEAK_CHECK_DELAY)): gc.collect() _last_memory_leak_check = time()
def _process_worker(call_queue, result_queue, initializer, initargs, processes_management_lock, timeout, worker_exit_lock, current_depth): """Evaluates calls from call_queue and places the results in result_queue. This worker is run in a separate process. Args: call_queue: A ctx.Queue of _CallItems that will be read and evaluated by the worker. result_queue: A ctx.Queue of _ResultItems that will written to by the worker. initializer: A callable initializer, or None initargs: A tuple of args for the initializer process_management_lock: A ctx.Lock avoiding worker timeout while some workers are being spawned. timeout: maximum time to wait for a new item in the call_queue. If that time is expired, the worker will shutdown. worker_exit_lock: Lock to avoid flagging the executor as broken on workers timeout. current_depth: Nested parallelism level, to avoid infinite spawning. """ if initializer is not None: try: initializer(*initargs) except BaseException: _base.LOGGER.critical('Exception in initializer:', exc_info=True) # The parent will notice that the process stopped and # mark the pool broken return # set the global _CURRENT_DEPTH mechanism to limit recursive call global _CURRENT_DEPTH _CURRENT_DEPTH = current_depth _process_reference_size = None _last_memory_leak_check = None pid = os.getpid() mp.util.debug('Worker started with timeout=%s' % timeout) while True: try: call_item = call_queue.get(block=True, timeout=timeout) if call_item is None: mp.util.info("Shutting down worker on sentinel") except queue.Empty: mp.util.info("Shutting down worker after timeout %0.3fs" % timeout) if processes_management_lock.acquire(block=False): processes_management_lock.release() call_item = None else: mp.util.info("Could not acquire processes_management_lock") continue except BaseException as e: previous_tb = traceback.format_exc() try: result_queue.put(_RemoteTraceback(previous_tb)) except BaseException: # If we cannot format correctly the exception, at least print # the traceback. print(previous_tb) sys.exit(1) if call_item is None: # Notify queue management thread about clean worker shutdown result_queue.put(pid) with worker_exit_lock: return try: r = call_item() except BaseException as e: exc = _ExceptionWithTraceback(e) result_queue.put(_ResultItem(call_item.work_id, exception=exc)) else: _sendback_result(result_queue, call_item.work_id, result=r) del r # Free the resource as soon as possible, to avoid holding onto # open files or shared memory that is not needed anymore del call_item if _USE_PSUTIL: if _process_reference_size is None: # Make reference measurement after the first call _process_reference_size = _get_memory_usage(pid, force_gc=True) _last_memory_leak_check = time() continue if time() - _last_memory_leak_check > _MEMORY_LEAK_CHECK_DELAY: mem_usage = _get_memory_usage(pid) _last_memory_leak_check = time() if mem_usage - _process_reference_size < _MAX_MEMORY_LEAK_SIZE: # Memory usage stays within bounds: everything is fine. continue # Check again memory usage; this time take the measurement # after a forced garbage collection to break any reference # cycles. mem_usage = _get_memory_usage(pid, force_gc=True) _last_memory_leak_check = time() if mem_usage - _process_reference_size < _MAX_MEMORY_LEAK_SIZE: # The GC managed to free the memory: everything is fine. continue # The process is leaking memory: let the master process # know that we need to start a new worker. mp.util.info("Memory leak detected: shutting down worker") result_queue.put(pid) with worker_exit_lock: return else: # if psutil is not installed, trigger gc.collect events # regularly to limit potential memory leaks due to reference cycles if ((_last_memory_leak_check is None) or (time() - _last_memory_leak_check > _MEMORY_LEAK_CHECK_DELAY)): gc.collect() _last_memory_leak_check = time()
[ "Evaluates", "calls", "from", "call_queue", "and", "places", "the", "results", "in", "result_queue", "." ]
tomMoral/loky
python
https://github.com/tomMoral/loky/blob/dc2d941d8285a96f3a5b666a4bd04875b0b25984/loky/process_executor.py#L350-L465
[ "def", "_process_worker", "(", "call_queue", ",", "result_queue", ",", "initializer", ",", "initargs", ",", "processes_management_lock", ",", "timeout", ",", "worker_exit_lock", ",", "current_depth", ")", ":", "if", "initializer", "is", "not", "None", ":", "try", ":", "initializer", "(", "*", "initargs", ")", "except", "BaseException", ":", "_base", ".", "LOGGER", ".", "critical", "(", "'Exception in initializer:'", ",", "exc_info", "=", "True", ")", "# The parent will notice that the process stopped and", "# mark the pool broken", "return", "# set the global _CURRENT_DEPTH mechanism to limit recursive call", "global", "_CURRENT_DEPTH", "_CURRENT_DEPTH", "=", "current_depth", "_process_reference_size", "=", "None", "_last_memory_leak_check", "=", "None", "pid", "=", "os", ".", "getpid", "(", ")", "mp", ".", "util", ".", "debug", "(", "'Worker started with timeout=%s'", "%", "timeout", ")", "while", "True", ":", "try", ":", "call_item", "=", "call_queue", ".", "get", "(", "block", "=", "True", ",", "timeout", "=", "timeout", ")", "if", "call_item", "is", "None", ":", "mp", ".", "util", ".", "info", "(", "\"Shutting down worker on sentinel\"", ")", "except", "queue", ".", "Empty", ":", "mp", ".", "util", ".", "info", "(", "\"Shutting down worker after timeout %0.3fs\"", "%", "timeout", ")", "if", "processes_management_lock", ".", "acquire", "(", "block", "=", "False", ")", ":", "processes_management_lock", ".", "release", "(", ")", "call_item", "=", "None", "else", ":", "mp", ".", "util", ".", "info", "(", "\"Could not acquire processes_management_lock\"", ")", "continue", "except", "BaseException", "as", "e", ":", "previous_tb", "=", "traceback", ".", "format_exc", "(", ")", "try", ":", "result_queue", ".", "put", "(", "_RemoteTraceback", "(", "previous_tb", ")", ")", "except", "BaseException", ":", "# If we cannot format correctly the exception, at least print", "# the traceback.", "print", "(", "previous_tb", ")", "sys", ".", "exit", "(", "1", ")", "if", "call_item", "is", "None", ":", "# Notify queue management thread about clean worker shutdown", "result_queue", ".", "put", "(", "pid", ")", "with", "worker_exit_lock", ":", "return", "try", ":", "r", "=", "call_item", "(", ")", "except", "BaseException", "as", "e", ":", "exc", "=", "_ExceptionWithTraceback", "(", "e", ")", "result_queue", ".", "put", "(", "_ResultItem", "(", "call_item", ".", "work_id", ",", "exception", "=", "exc", ")", ")", "else", ":", "_sendback_result", "(", "result_queue", ",", "call_item", ".", "work_id", ",", "result", "=", "r", ")", "del", "r", "# Free the resource as soon as possible, to avoid holding onto", "# open files or shared memory that is not needed anymore", "del", "call_item", "if", "_USE_PSUTIL", ":", "if", "_process_reference_size", "is", "None", ":", "# Make reference measurement after the first call", "_process_reference_size", "=", "_get_memory_usage", "(", "pid", ",", "force_gc", "=", "True", ")", "_last_memory_leak_check", "=", "time", "(", ")", "continue", "if", "time", "(", ")", "-", "_last_memory_leak_check", ">", "_MEMORY_LEAK_CHECK_DELAY", ":", "mem_usage", "=", "_get_memory_usage", "(", "pid", ")", "_last_memory_leak_check", "=", "time", "(", ")", "if", "mem_usage", "-", "_process_reference_size", "<", "_MAX_MEMORY_LEAK_SIZE", ":", "# Memory usage stays within bounds: everything is fine.", "continue", "# Check again memory usage; this time take the measurement", "# after a forced garbage collection to break any reference", "# cycles.", "mem_usage", "=", "_get_memory_usage", "(", "pid", ",", "force_gc", "=", "True", ")", "_last_memory_leak_check", "=", "time", "(", ")", "if", "mem_usage", "-", "_process_reference_size", "<", "_MAX_MEMORY_LEAK_SIZE", ":", "# The GC managed to free the memory: everything is fine.", "continue", "# The process is leaking memory: let the master process", "# know that we need to start a new worker.", "mp", ".", "util", ".", "info", "(", "\"Memory leak detected: shutting down worker\"", ")", "result_queue", ".", "put", "(", "pid", ")", "with", "worker_exit_lock", ":", "return", "else", ":", "# if psutil is not installed, trigger gc.collect events", "# regularly to limit potential memory leaks due to reference cycles", "if", "(", "(", "_last_memory_leak_check", "is", "None", ")", "or", "(", "time", "(", ")", "-", "_last_memory_leak_check", ">", "_MEMORY_LEAK_CHECK_DELAY", ")", ")", ":", "gc", ".", "collect", "(", ")", "_last_memory_leak_check", "=", "time", "(", ")" ]
dc2d941d8285a96f3a5b666a4bd04875b0b25984
test
_add_call_item_to_queue
Fills call_queue with _WorkItems from pending_work_items. This function never blocks. Args: pending_work_items: A dict mapping work ids to _WorkItems e.g. {5: <_WorkItem...>, 6: <_WorkItem...>, ...} work_ids: A queue.Queue of work ids e.g. Queue([5, 6, ...]). Work ids are consumed and the corresponding _WorkItems from pending_work_items are transformed into _CallItems and put in call_queue. call_queue: A ctx.Queue that will be filled with _CallItems derived from _WorkItems.
loky/process_executor.py
def _add_call_item_to_queue(pending_work_items, running_work_items, work_ids, call_queue): """Fills call_queue with _WorkItems from pending_work_items. This function never blocks. Args: pending_work_items: A dict mapping work ids to _WorkItems e.g. {5: <_WorkItem...>, 6: <_WorkItem...>, ...} work_ids: A queue.Queue of work ids e.g. Queue([5, 6, ...]). Work ids are consumed and the corresponding _WorkItems from pending_work_items are transformed into _CallItems and put in call_queue. call_queue: A ctx.Queue that will be filled with _CallItems derived from _WorkItems. """ while True: if call_queue.full(): return try: work_id = work_ids.get(block=False) except queue.Empty: return else: work_item = pending_work_items[work_id] if work_item.future.set_running_or_notify_cancel(): running_work_items += [work_id] call_queue.put(_CallItem(work_id, work_item.fn, work_item.args, work_item.kwargs), block=True) else: del pending_work_items[work_id] continue
def _add_call_item_to_queue(pending_work_items, running_work_items, work_ids, call_queue): """Fills call_queue with _WorkItems from pending_work_items. This function never blocks. Args: pending_work_items: A dict mapping work ids to _WorkItems e.g. {5: <_WorkItem...>, 6: <_WorkItem...>, ...} work_ids: A queue.Queue of work ids e.g. Queue([5, 6, ...]). Work ids are consumed and the corresponding _WorkItems from pending_work_items are transformed into _CallItems and put in call_queue. call_queue: A ctx.Queue that will be filled with _CallItems derived from _WorkItems. """ while True: if call_queue.full(): return try: work_id = work_ids.get(block=False) except queue.Empty: return else: work_item = pending_work_items[work_id] if work_item.future.set_running_or_notify_cancel(): running_work_items += [work_id] call_queue.put(_CallItem(work_id, work_item.fn, work_item.args, work_item.kwargs), block=True) else: del pending_work_items[work_id] continue
[ "Fills", "call_queue", "with", "_WorkItems", "from", "pending_work_items", "." ]
tomMoral/loky
python
https://github.com/tomMoral/loky/blob/dc2d941d8285a96f3a5b666a4bd04875b0b25984/loky/process_executor.py#L468-L505
[ "def", "_add_call_item_to_queue", "(", "pending_work_items", ",", "running_work_items", ",", "work_ids", ",", "call_queue", ")", ":", "while", "True", ":", "if", "call_queue", ".", "full", "(", ")", ":", "return", "try", ":", "work_id", "=", "work_ids", ".", "get", "(", "block", "=", "False", ")", "except", "queue", ".", "Empty", ":", "return", "else", ":", "work_item", "=", "pending_work_items", "[", "work_id", "]", "if", "work_item", ".", "future", ".", "set_running_or_notify_cancel", "(", ")", ":", "running_work_items", "+=", "[", "work_id", "]", "call_queue", ".", "put", "(", "_CallItem", "(", "work_id", ",", "work_item", ".", "fn", ",", "work_item", ".", "args", ",", "work_item", ".", "kwargs", ")", ",", "block", "=", "True", ")", "else", ":", "del", "pending_work_items", "[", "work_id", "]", "continue" ]
dc2d941d8285a96f3a5b666a4bd04875b0b25984
test
_queue_management_worker
Manages the communication between this process and the worker processes. This function is run in a local thread. Args: executor_reference: A weakref.ref to the ProcessPoolExecutor that owns this thread. Used to determine if the ProcessPoolExecutor has been garbage collected and that this function can exit. executor_flags: A ExecutorFlags holding internal states of the ProcessPoolExecutor. It permits to know if the executor is broken even the object has been gc. process: A list of the ctx.Process instances used as workers. pending_work_items: A dict mapping work ids to _WorkItems e.g. {5: <_WorkItem...>, 6: <_WorkItem...>, ...} work_ids_queue: A queue.Queue of work ids e.g. Queue([5, 6, ...]). call_queue: A ctx.Queue that will be filled with _CallItems derived from _WorkItems for processing by the process workers. result_queue: A ctx.SimpleQueue of _ResultItems generated by the process workers. thread_wakeup: A _ThreadWakeup to allow waking up the queue_manager_thread from the main Thread and avoid deadlocks caused by permanently locked queues.
loky/process_executor.py
def _queue_management_worker(executor_reference, executor_flags, processes, pending_work_items, running_work_items, work_ids_queue, call_queue, result_queue, thread_wakeup, processes_management_lock): """Manages the communication between this process and the worker processes. This function is run in a local thread. Args: executor_reference: A weakref.ref to the ProcessPoolExecutor that owns this thread. Used to determine if the ProcessPoolExecutor has been garbage collected and that this function can exit. executor_flags: A ExecutorFlags holding internal states of the ProcessPoolExecutor. It permits to know if the executor is broken even the object has been gc. process: A list of the ctx.Process instances used as workers. pending_work_items: A dict mapping work ids to _WorkItems e.g. {5: <_WorkItem...>, 6: <_WorkItem...>, ...} work_ids_queue: A queue.Queue of work ids e.g. Queue([5, 6, ...]). call_queue: A ctx.Queue that will be filled with _CallItems derived from _WorkItems for processing by the process workers. result_queue: A ctx.SimpleQueue of _ResultItems generated by the process workers. thread_wakeup: A _ThreadWakeup to allow waking up the queue_manager_thread from the main Thread and avoid deadlocks caused by permanently locked queues. """ executor = None def is_shutting_down(): # No more work items can be added if: # - The interpreter is shutting down OR # - The executor that own this worker is not broken AND # * The executor that owns this worker has been collected OR # * The executor that owns this worker has been shutdown. # If the executor is broken, it should be detected in the next loop. return (_global_shutdown or ((executor is None or executor_flags.shutdown) and not executor_flags.broken)) def shutdown_all_workers(): mp.util.debug("queue management thread shutting down") executor_flags.flag_as_shutting_down() # Create a list to avoid RuntimeError due to concurrent modification of # processes. nb_children_alive is thus an upper bound. Also release the # processes' _worker_exit_lock to accelerate the shutdown procedure, as # there is no need for hand-shake here. with processes_management_lock: n_children_alive = 0 for p in list(processes.values()): p._worker_exit_lock.release() n_children_alive += 1 n_children_to_stop = n_children_alive n_sentinels_sent = 0 # Send the right number of sentinels, to make sure all children are # properly terminated. while n_sentinels_sent < n_children_to_stop and n_children_alive > 0: for i in range(n_children_to_stop - n_sentinels_sent): try: call_queue.put_nowait(None) n_sentinels_sent += 1 except Full: break with processes_management_lock: n_children_alive = sum( p.is_alive() for p in list(processes.values()) ) # Release the queue's resources as soon as possible. Flag the feeder # thread for clean exit to avoid having the crash detection thread flag # the Executor as broken during the shutdown. This is safe as either: # * We don't need to communicate with the workers anymore # * There is nothing left in the Queue buffer except None sentinels mp.util.debug("closing call_queue") call_queue.close() mp.util.debug("joining processes") # If .join() is not called on the created processes then # some ctx.Queue methods may deadlock on Mac OS X. while processes: _, p = processes.popitem() p.join() mp.util.debug("queue management thread clean shutdown of worker " "processes: {}".format(list(processes))) result_reader = result_queue._reader wakeup_reader = thread_wakeup._reader readers = [result_reader, wakeup_reader] while True: _add_call_item_to_queue(pending_work_items, running_work_items, work_ids_queue, call_queue) # Wait for a result to be ready in the result_queue while checking # that all worker processes are still running, or for a wake up # signal send. The wake up signals come either from new tasks being # submitted, from the executor being shutdown/gc-ed, or from the # shutdown of the python interpreter. worker_sentinels = [p.sentinel for p in list(processes.values())] ready = wait(readers + worker_sentinels) broken = ("A worker process managed by the executor was unexpectedly " "terminated. This could be caused by a segmentation fault " "while calling the function or by an excessive memory usage " "causing the Operating System to kill the worker.", None, TerminatedWorkerError) if result_reader in ready: try: result_item = result_reader.recv() broken = None if isinstance(result_item, _RemoteTraceback): broken = ("A task has failed to un-serialize. Please " "ensure that the arguments of the function are " "all picklable.", result_item.tb, BrokenProcessPool) except BaseException as e: tb = getattr(e, "__traceback__", None) if tb is None: _, _, tb = sys.exc_info() broken = ("A result has failed to un-serialize. Please " "ensure that the objects returned by the function " "are always picklable.", traceback.format_exception(type(e), e, tb), BrokenProcessPool) elif wakeup_reader in ready: broken = None result_item = None thread_wakeup.clear() if broken is not None: msg, cause_tb, exc_type = broken if (issubclass(exc_type, TerminatedWorkerError) and (sys.platform != "win32")): # In Windows, introspecting terminated workers exitcodes seems # unstable, therefore they are not appended in the exception # message. msg += " The exit codes of the workers are {}".format( get_exitcodes_terminated_worker(processes)) bpe = exc_type(msg) if cause_tb is not None: bpe = set_cause(bpe, _RemoteTraceback( "\n'''\n{}'''".format(''.join(cause_tb)))) # Mark the process pool broken so that submits fail right now. executor_flags.flag_as_broken(bpe) # All futures in flight must be marked failed for work_id, work_item in pending_work_items.items(): work_item.future.set_exception(bpe) # Delete references to object. See issue16284 del work_item pending_work_items.clear() # Terminate remaining workers forcibly: the queues or their # locks may be in a dirty state and block forever. while processes: _, p = processes.popitem() mp.util.debug('terminate process {}'.format(p.name)) try: recursive_terminate(p) except ProcessLookupError: # pragma: no cover pass shutdown_all_workers() return if isinstance(result_item, int): # Clean shutdown of a worker using its PID, either on request # by the executor.shutdown method or by the timeout of the worker # itself: we should not mark the executor as broken. with processes_management_lock: p = processes.pop(result_item, None) # p can be None is the executor is concurrently shutting down. if p is not None: p._worker_exit_lock.release() p.join() del p # Make sure the executor have the right number of worker, even if a # worker timeout while some jobs were submitted. If some work is # pending or there is less processes than running items, we need to # start a new Process and raise a warning. n_pending = len(pending_work_items) n_running = len(running_work_items) if (n_pending - n_running > 0 or n_running > len(processes)): executor = executor_reference() if (executor is not None and len(processes) < executor._max_workers): warnings.warn( "A worker stopped while some jobs were given to the " "executor. This can be caused by a too short worker " "timeout or by a memory leak.", UserWarning ) executor._adjust_process_count() executor = None elif result_item is not None: work_item = pending_work_items.pop(result_item.work_id, None) # work_item can be None if another process terminated if work_item is not None: if result_item.exception: work_item.future.set_exception(result_item.exception) else: work_item.future.set_result(result_item.result) # Delete references to object. See issue16284 del work_item running_work_items.remove(result_item.work_id) # Delete reference to result_item del result_item # Check whether we should start shutting down. executor = executor_reference() # No more work items can be added if: # - The interpreter is shutting down OR # - The executor that owns this worker has been collected OR # - The executor that owns this worker has been shutdown. if is_shutting_down(): # bpo-33097: Make sure that the executor is flagged as shutting # down even if it is shutdown by the interpreter exiting. with executor_flags.shutdown_lock: executor_flags.shutdown = True if executor_flags.kill_workers: while pending_work_items: _, work_item = pending_work_items.popitem() work_item.future.set_exception(ShutdownExecutorError( "The Executor was shutdown before this job could " "complete.")) del work_item # Terminate remaining workers forcibly: the queues or their # locks may be in a dirty state and block forever. while processes: _, p = processes.popitem() recursive_terminate(p) shutdown_all_workers() return # Since no new work items can be added, it is safe to shutdown # this thread if there are no pending work items. if not pending_work_items: shutdown_all_workers() return elif executor_flags.broken: return executor = None
def _queue_management_worker(executor_reference, executor_flags, processes, pending_work_items, running_work_items, work_ids_queue, call_queue, result_queue, thread_wakeup, processes_management_lock): """Manages the communication between this process and the worker processes. This function is run in a local thread. Args: executor_reference: A weakref.ref to the ProcessPoolExecutor that owns this thread. Used to determine if the ProcessPoolExecutor has been garbage collected and that this function can exit. executor_flags: A ExecutorFlags holding internal states of the ProcessPoolExecutor. It permits to know if the executor is broken even the object has been gc. process: A list of the ctx.Process instances used as workers. pending_work_items: A dict mapping work ids to _WorkItems e.g. {5: <_WorkItem...>, 6: <_WorkItem...>, ...} work_ids_queue: A queue.Queue of work ids e.g. Queue([5, 6, ...]). call_queue: A ctx.Queue that will be filled with _CallItems derived from _WorkItems for processing by the process workers. result_queue: A ctx.SimpleQueue of _ResultItems generated by the process workers. thread_wakeup: A _ThreadWakeup to allow waking up the queue_manager_thread from the main Thread and avoid deadlocks caused by permanently locked queues. """ executor = None def is_shutting_down(): # No more work items can be added if: # - The interpreter is shutting down OR # - The executor that own this worker is not broken AND # * The executor that owns this worker has been collected OR # * The executor that owns this worker has been shutdown. # If the executor is broken, it should be detected in the next loop. return (_global_shutdown or ((executor is None or executor_flags.shutdown) and not executor_flags.broken)) def shutdown_all_workers(): mp.util.debug("queue management thread shutting down") executor_flags.flag_as_shutting_down() # Create a list to avoid RuntimeError due to concurrent modification of # processes. nb_children_alive is thus an upper bound. Also release the # processes' _worker_exit_lock to accelerate the shutdown procedure, as # there is no need for hand-shake here. with processes_management_lock: n_children_alive = 0 for p in list(processes.values()): p._worker_exit_lock.release() n_children_alive += 1 n_children_to_stop = n_children_alive n_sentinels_sent = 0 # Send the right number of sentinels, to make sure all children are # properly terminated. while n_sentinels_sent < n_children_to_stop and n_children_alive > 0: for i in range(n_children_to_stop - n_sentinels_sent): try: call_queue.put_nowait(None) n_sentinels_sent += 1 except Full: break with processes_management_lock: n_children_alive = sum( p.is_alive() for p in list(processes.values()) ) # Release the queue's resources as soon as possible. Flag the feeder # thread for clean exit to avoid having the crash detection thread flag # the Executor as broken during the shutdown. This is safe as either: # * We don't need to communicate with the workers anymore # * There is nothing left in the Queue buffer except None sentinels mp.util.debug("closing call_queue") call_queue.close() mp.util.debug("joining processes") # If .join() is not called on the created processes then # some ctx.Queue methods may deadlock on Mac OS X. while processes: _, p = processes.popitem() p.join() mp.util.debug("queue management thread clean shutdown of worker " "processes: {}".format(list(processes))) result_reader = result_queue._reader wakeup_reader = thread_wakeup._reader readers = [result_reader, wakeup_reader] while True: _add_call_item_to_queue(pending_work_items, running_work_items, work_ids_queue, call_queue) # Wait for a result to be ready in the result_queue while checking # that all worker processes are still running, or for a wake up # signal send. The wake up signals come either from new tasks being # submitted, from the executor being shutdown/gc-ed, or from the # shutdown of the python interpreter. worker_sentinels = [p.sentinel for p in list(processes.values())] ready = wait(readers + worker_sentinels) broken = ("A worker process managed by the executor was unexpectedly " "terminated. This could be caused by a segmentation fault " "while calling the function or by an excessive memory usage " "causing the Operating System to kill the worker.", None, TerminatedWorkerError) if result_reader in ready: try: result_item = result_reader.recv() broken = None if isinstance(result_item, _RemoteTraceback): broken = ("A task has failed to un-serialize. Please " "ensure that the arguments of the function are " "all picklable.", result_item.tb, BrokenProcessPool) except BaseException as e: tb = getattr(e, "__traceback__", None) if tb is None: _, _, tb = sys.exc_info() broken = ("A result has failed to un-serialize. Please " "ensure that the objects returned by the function " "are always picklable.", traceback.format_exception(type(e), e, tb), BrokenProcessPool) elif wakeup_reader in ready: broken = None result_item = None thread_wakeup.clear() if broken is not None: msg, cause_tb, exc_type = broken if (issubclass(exc_type, TerminatedWorkerError) and (sys.platform != "win32")): # In Windows, introspecting terminated workers exitcodes seems # unstable, therefore they are not appended in the exception # message. msg += " The exit codes of the workers are {}".format( get_exitcodes_terminated_worker(processes)) bpe = exc_type(msg) if cause_tb is not None: bpe = set_cause(bpe, _RemoteTraceback( "\n'''\n{}'''".format(''.join(cause_tb)))) # Mark the process pool broken so that submits fail right now. executor_flags.flag_as_broken(bpe) # All futures in flight must be marked failed for work_id, work_item in pending_work_items.items(): work_item.future.set_exception(bpe) # Delete references to object. See issue16284 del work_item pending_work_items.clear() # Terminate remaining workers forcibly: the queues or their # locks may be in a dirty state and block forever. while processes: _, p = processes.popitem() mp.util.debug('terminate process {}'.format(p.name)) try: recursive_terminate(p) except ProcessLookupError: # pragma: no cover pass shutdown_all_workers() return if isinstance(result_item, int): # Clean shutdown of a worker using its PID, either on request # by the executor.shutdown method or by the timeout of the worker # itself: we should not mark the executor as broken. with processes_management_lock: p = processes.pop(result_item, None) # p can be None is the executor is concurrently shutting down. if p is not None: p._worker_exit_lock.release() p.join() del p # Make sure the executor have the right number of worker, even if a # worker timeout while some jobs were submitted. If some work is # pending or there is less processes than running items, we need to # start a new Process and raise a warning. n_pending = len(pending_work_items) n_running = len(running_work_items) if (n_pending - n_running > 0 or n_running > len(processes)): executor = executor_reference() if (executor is not None and len(processes) < executor._max_workers): warnings.warn( "A worker stopped while some jobs were given to the " "executor. This can be caused by a too short worker " "timeout or by a memory leak.", UserWarning ) executor._adjust_process_count() executor = None elif result_item is not None: work_item = pending_work_items.pop(result_item.work_id, None) # work_item can be None if another process terminated if work_item is not None: if result_item.exception: work_item.future.set_exception(result_item.exception) else: work_item.future.set_result(result_item.result) # Delete references to object. See issue16284 del work_item running_work_items.remove(result_item.work_id) # Delete reference to result_item del result_item # Check whether we should start shutting down. executor = executor_reference() # No more work items can be added if: # - The interpreter is shutting down OR # - The executor that owns this worker has been collected OR # - The executor that owns this worker has been shutdown. if is_shutting_down(): # bpo-33097: Make sure that the executor is flagged as shutting # down even if it is shutdown by the interpreter exiting. with executor_flags.shutdown_lock: executor_flags.shutdown = True if executor_flags.kill_workers: while pending_work_items: _, work_item = pending_work_items.popitem() work_item.future.set_exception(ShutdownExecutorError( "The Executor was shutdown before this job could " "complete.")) del work_item # Terminate remaining workers forcibly: the queues or their # locks may be in a dirty state and block forever. while processes: _, p = processes.popitem() recursive_terminate(p) shutdown_all_workers() return # Since no new work items can be added, it is safe to shutdown # this thread if there are no pending work items. if not pending_work_items: shutdown_all_workers() return elif executor_flags.broken: return executor = None
[ "Manages", "the", "communication", "between", "this", "process", "and", "the", "worker", "processes", "." ]
tomMoral/loky
python
https://github.com/tomMoral/loky/blob/dc2d941d8285a96f3a5b666a4bd04875b0b25984/loky/process_executor.py#L508-L757
[ "def", "_queue_management_worker", "(", "executor_reference", ",", "executor_flags", ",", "processes", ",", "pending_work_items", ",", "running_work_items", ",", "work_ids_queue", ",", "call_queue", ",", "result_queue", ",", "thread_wakeup", ",", "processes_management_lock", ")", ":", "executor", "=", "None", "def", "is_shutting_down", "(", ")", ":", "# No more work items can be added if:", "# - The interpreter is shutting down OR", "# - The executor that own this worker is not broken AND", "# * The executor that owns this worker has been collected OR", "# * The executor that owns this worker has been shutdown.", "# If the executor is broken, it should be detected in the next loop.", "return", "(", "_global_shutdown", "or", "(", "(", "executor", "is", "None", "or", "executor_flags", ".", "shutdown", ")", "and", "not", "executor_flags", ".", "broken", ")", ")", "def", "shutdown_all_workers", "(", ")", ":", "mp", ".", "util", ".", "debug", "(", "\"queue management thread shutting down\"", ")", "executor_flags", ".", "flag_as_shutting_down", "(", ")", "# Create a list to avoid RuntimeError due to concurrent modification of", "# processes. nb_children_alive is thus an upper bound. Also release the", "# processes' _worker_exit_lock to accelerate the shutdown procedure, as", "# there is no need for hand-shake here.", "with", "processes_management_lock", ":", "n_children_alive", "=", "0", "for", "p", "in", "list", "(", "processes", ".", "values", "(", ")", ")", ":", "p", ".", "_worker_exit_lock", ".", "release", "(", ")", "n_children_alive", "+=", "1", "n_children_to_stop", "=", "n_children_alive", "n_sentinels_sent", "=", "0", "# Send the right number of sentinels, to make sure all children are", "# properly terminated.", "while", "n_sentinels_sent", "<", "n_children_to_stop", "and", "n_children_alive", ">", "0", ":", "for", "i", "in", "range", "(", "n_children_to_stop", "-", "n_sentinels_sent", ")", ":", "try", ":", "call_queue", ".", "put_nowait", "(", "None", ")", "n_sentinels_sent", "+=", "1", "except", "Full", ":", "break", "with", "processes_management_lock", ":", "n_children_alive", "=", "sum", "(", "p", ".", "is_alive", "(", ")", "for", "p", "in", "list", "(", "processes", ".", "values", "(", ")", ")", ")", "# Release the queue's resources as soon as possible. Flag the feeder", "# thread for clean exit to avoid having the crash detection thread flag", "# the Executor as broken during the shutdown. This is safe as either:", "# * We don't need to communicate with the workers anymore", "# * There is nothing left in the Queue buffer except None sentinels", "mp", ".", "util", ".", "debug", "(", "\"closing call_queue\"", ")", "call_queue", ".", "close", "(", ")", "mp", ".", "util", ".", "debug", "(", "\"joining processes\"", ")", "# If .join() is not called on the created processes then", "# some ctx.Queue methods may deadlock on Mac OS X.", "while", "processes", ":", "_", ",", "p", "=", "processes", ".", "popitem", "(", ")", "p", ".", "join", "(", ")", "mp", ".", "util", ".", "debug", "(", "\"queue management thread clean shutdown of worker \"", "\"processes: {}\"", ".", "format", "(", "list", "(", "processes", ")", ")", ")", "result_reader", "=", "result_queue", ".", "_reader", "wakeup_reader", "=", "thread_wakeup", ".", "_reader", "readers", "=", "[", "result_reader", ",", "wakeup_reader", "]", "while", "True", ":", "_add_call_item_to_queue", "(", "pending_work_items", ",", "running_work_items", ",", "work_ids_queue", ",", "call_queue", ")", "# Wait for a result to be ready in the result_queue while checking", "# that all worker processes are still running, or for a wake up", "# signal send. The wake up signals come either from new tasks being", "# submitted, from the executor being shutdown/gc-ed, or from the", "# shutdown of the python interpreter.", "worker_sentinels", "=", "[", "p", ".", "sentinel", "for", "p", "in", "list", "(", "processes", ".", "values", "(", ")", ")", "]", "ready", "=", "wait", "(", "readers", "+", "worker_sentinels", ")", "broken", "=", "(", "\"A worker process managed by the executor was unexpectedly \"", "\"terminated. This could be caused by a segmentation fault \"", "\"while calling the function or by an excessive memory usage \"", "\"causing the Operating System to kill the worker.\"", ",", "None", ",", "TerminatedWorkerError", ")", "if", "result_reader", "in", "ready", ":", "try", ":", "result_item", "=", "result_reader", ".", "recv", "(", ")", "broken", "=", "None", "if", "isinstance", "(", "result_item", ",", "_RemoteTraceback", ")", ":", "broken", "=", "(", "\"A task has failed to un-serialize. Please \"", "\"ensure that the arguments of the function are \"", "\"all picklable.\"", ",", "result_item", ".", "tb", ",", "BrokenProcessPool", ")", "except", "BaseException", "as", "e", ":", "tb", "=", "getattr", "(", "e", ",", "\"__traceback__\"", ",", "None", ")", "if", "tb", "is", "None", ":", "_", ",", "_", ",", "tb", "=", "sys", ".", "exc_info", "(", ")", "broken", "=", "(", "\"A result has failed to un-serialize. Please \"", "\"ensure that the objects returned by the function \"", "\"are always picklable.\"", ",", "traceback", ".", "format_exception", "(", "type", "(", "e", ")", ",", "e", ",", "tb", ")", ",", "BrokenProcessPool", ")", "elif", "wakeup_reader", "in", "ready", ":", "broken", "=", "None", "result_item", "=", "None", "thread_wakeup", ".", "clear", "(", ")", "if", "broken", "is", "not", "None", ":", "msg", ",", "cause_tb", ",", "exc_type", "=", "broken", "if", "(", "issubclass", "(", "exc_type", ",", "TerminatedWorkerError", ")", "and", "(", "sys", ".", "platform", "!=", "\"win32\"", ")", ")", ":", "# In Windows, introspecting terminated workers exitcodes seems", "# unstable, therefore they are not appended in the exception", "# message.", "msg", "+=", "\" The exit codes of the workers are {}\"", ".", "format", "(", "get_exitcodes_terminated_worker", "(", "processes", ")", ")", "bpe", "=", "exc_type", "(", "msg", ")", "if", "cause_tb", "is", "not", "None", ":", "bpe", "=", "set_cause", "(", "bpe", ",", "_RemoteTraceback", "(", "\"\\n'''\\n{}'''\"", ".", "format", "(", "''", ".", "join", "(", "cause_tb", ")", ")", ")", ")", "# Mark the process pool broken so that submits fail right now.", "executor_flags", ".", "flag_as_broken", "(", "bpe", ")", "# All futures in flight must be marked failed", "for", "work_id", ",", "work_item", "in", "pending_work_items", ".", "items", "(", ")", ":", "work_item", ".", "future", ".", "set_exception", "(", "bpe", ")", "# Delete references to object. See issue16284", "del", "work_item", "pending_work_items", ".", "clear", "(", ")", "# Terminate remaining workers forcibly: the queues or their", "# locks may be in a dirty state and block forever.", "while", "processes", ":", "_", ",", "p", "=", "processes", ".", "popitem", "(", ")", "mp", ".", "util", ".", "debug", "(", "'terminate process {}'", ".", "format", "(", "p", ".", "name", ")", ")", "try", ":", "recursive_terminate", "(", "p", ")", "except", "ProcessLookupError", ":", "# pragma: no cover", "pass", "shutdown_all_workers", "(", ")", "return", "if", "isinstance", "(", "result_item", ",", "int", ")", ":", "# Clean shutdown of a worker using its PID, either on request", "# by the executor.shutdown method or by the timeout of the worker", "# itself: we should not mark the executor as broken.", "with", "processes_management_lock", ":", "p", "=", "processes", ".", "pop", "(", "result_item", ",", "None", ")", "# p can be None is the executor is concurrently shutting down.", "if", "p", "is", "not", "None", ":", "p", ".", "_worker_exit_lock", ".", "release", "(", ")", "p", ".", "join", "(", ")", "del", "p", "# Make sure the executor have the right number of worker, even if a", "# worker timeout while some jobs were submitted. If some work is", "# pending or there is less processes than running items, we need to", "# start a new Process and raise a warning.", "n_pending", "=", "len", "(", "pending_work_items", ")", "n_running", "=", "len", "(", "running_work_items", ")", "if", "(", "n_pending", "-", "n_running", ">", "0", "or", "n_running", ">", "len", "(", "processes", ")", ")", ":", "executor", "=", "executor_reference", "(", ")", "if", "(", "executor", "is", "not", "None", "and", "len", "(", "processes", ")", "<", "executor", ".", "_max_workers", ")", ":", "warnings", ".", "warn", "(", "\"A worker stopped while some jobs were given to the \"", "\"executor. This can be caused by a too short worker \"", "\"timeout or by a memory leak.\"", ",", "UserWarning", ")", "executor", ".", "_adjust_process_count", "(", ")", "executor", "=", "None", "elif", "result_item", "is", "not", "None", ":", "work_item", "=", "pending_work_items", ".", "pop", "(", "result_item", ".", "work_id", ",", "None", ")", "# work_item can be None if another process terminated", "if", "work_item", "is", "not", "None", ":", "if", "result_item", ".", "exception", ":", "work_item", ".", "future", ".", "set_exception", "(", "result_item", ".", "exception", ")", "else", ":", "work_item", ".", "future", ".", "set_result", "(", "result_item", ".", "result", ")", "# Delete references to object. See issue16284", "del", "work_item", "running_work_items", ".", "remove", "(", "result_item", ".", "work_id", ")", "# Delete reference to result_item", "del", "result_item", "# Check whether we should start shutting down.", "executor", "=", "executor_reference", "(", ")", "# No more work items can be added if:", "# - The interpreter is shutting down OR", "# - The executor that owns this worker has been collected OR", "# - The executor that owns this worker has been shutdown.", "if", "is_shutting_down", "(", ")", ":", "# bpo-33097: Make sure that the executor is flagged as shutting", "# down even if it is shutdown by the interpreter exiting.", "with", "executor_flags", ".", "shutdown_lock", ":", "executor_flags", ".", "shutdown", "=", "True", "if", "executor_flags", ".", "kill_workers", ":", "while", "pending_work_items", ":", "_", ",", "work_item", "=", "pending_work_items", ".", "popitem", "(", ")", "work_item", ".", "future", ".", "set_exception", "(", "ShutdownExecutorError", "(", "\"The Executor was shutdown before this job could \"", "\"complete.\"", ")", ")", "del", "work_item", "# Terminate remaining workers forcibly: the queues or their", "# locks may be in a dirty state and block forever.", "while", "processes", ":", "_", ",", "p", "=", "processes", ".", "popitem", "(", ")", "recursive_terminate", "(", "p", ")", "shutdown_all_workers", "(", ")", "return", "# Since no new work items can be added, it is safe to shutdown", "# this thread if there are no pending work items.", "if", "not", "pending_work_items", ":", "shutdown_all_workers", "(", ")", "return", "elif", "executor_flags", ".", "broken", ":", "return", "executor", "=", "None" ]
dc2d941d8285a96f3a5b666a4bd04875b0b25984
test
ProcessPoolExecutor._ensure_executor_running
ensures all workers and management thread are running
loky/process_executor.py
def _ensure_executor_running(self): """ensures all workers and management thread are running """ with self._processes_management_lock: if len(self._processes) != self._max_workers: self._adjust_process_count() self._start_queue_management_thread()
def _ensure_executor_running(self): """ensures all workers and management thread are running """ with self._processes_management_lock: if len(self._processes) != self._max_workers: self._adjust_process_count() self._start_queue_management_thread()
[ "ensures", "all", "workers", "and", "management", "thread", "are", "running" ]
tomMoral/loky
python
https://github.com/tomMoral/loky/blob/dc2d941d8285a96f3a5b666a4bd04875b0b25984/loky/process_executor.py#L1011-L1017
[ "def", "_ensure_executor_running", "(", "self", ")", ":", "with", "self", ".", "_processes_management_lock", ":", "if", "len", "(", "self", ".", "_processes", ")", "!=", "self", ".", "_max_workers", ":", "self", ".", "_adjust_process_count", "(", ")", "self", ".", "_start_queue_management_thread", "(", ")" ]
dc2d941d8285a96f3a5b666a4bd04875b0b25984
test
ProcessPoolExecutor.map
Returns an iterator equivalent to map(fn, iter). Args: fn: A callable that will take as many arguments as there are passed iterables. timeout: The maximum number of seconds to wait. If None, then there is no limit on the wait time. chunksize: If greater than one, the iterables will be chopped into chunks of size chunksize and submitted to the process pool. If set to one, the items in the list will be sent one at a time. Returns: An iterator equivalent to: map(func, *iterables) but the calls may be evaluated out-of-order. Raises: TimeoutError: If the entire result iterator could not be generated before the given timeout. Exception: If fn(*args) raises for any values.
loky/process_executor.py
def map(self, fn, *iterables, **kwargs): """Returns an iterator equivalent to map(fn, iter). Args: fn: A callable that will take as many arguments as there are passed iterables. timeout: The maximum number of seconds to wait. If None, then there is no limit on the wait time. chunksize: If greater than one, the iterables will be chopped into chunks of size chunksize and submitted to the process pool. If set to one, the items in the list will be sent one at a time. Returns: An iterator equivalent to: map(func, *iterables) but the calls may be evaluated out-of-order. Raises: TimeoutError: If the entire result iterator could not be generated before the given timeout. Exception: If fn(*args) raises for any values. """ timeout = kwargs.get('timeout', None) chunksize = kwargs.get('chunksize', 1) if chunksize < 1: raise ValueError("chunksize must be >= 1.") results = super(ProcessPoolExecutor, self).map( partial(_process_chunk, fn), _get_chunks(chunksize, *iterables), timeout=timeout) return _chain_from_iterable_of_lists(results)
def map(self, fn, *iterables, **kwargs): """Returns an iterator equivalent to map(fn, iter). Args: fn: A callable that will take as many arguments as there are passed iterables. timeout: The maximum number of seconds to wait. If None, then there is no limit on the wait time. chunksize: If greater than one, the iterables will be chopped into chunks of size chunksize and submitted to the process pool. If set to one, the items in the list will be sent one at a time. Returns: An iterator equivalent to: map(func, *iterables) but the calls may be evaluated out-of-order. Raises: TimeoutError: If the entire result iterator could not be generated before the given timeout. Exception: If fn(*args) raises for any values. """ timeout = kwargs.get('timeout', None) chunksize = kwargs.get('chunksize', 1) if chunksize < 1: raise ValueError("chunksize must be >= 1.") results = super(ProcessPoolExecutor, self).map( partial(_process_chunk, fn), _get_chunks(chunksize, *iterables), timeout=timeout) return _chain_from_iterable_of_lists(results)
[ "Returns", "an", "iterator", "equivalent", "to", "map", "(", "fn", "iter", ")", "." ]
tomMoral/loky
python
https://github.com/tomMoral/loky/blob/dc2d941d8285a96f3a5b666a4bd04875b0b25984/loky/process_executor.py#L1046-L1076
[ "def", "map", "(", "self", ",", "fn", ",", "*", "iterables", ",", "*", "*", "kwargs", ")", ":", "timeout", "=", "kwargs", ".", "get", "(", "'timeout'", ",", "None", ")", "chunksize", "=", "kwargs", ".", "get", "(", "'chunksize'", ",", "1", ")", "if", "chunksize", "<", "1", ":", "raise", "ValueError", "(", "\"chunksize must be >= 1.\"", ")", "results", "=", "super", "(", "ProcessPoolExecutor", ",", "self", ")", ".", "map", "(", "partial", "(", "_process_chunk", ",", "fn", ")", ",", "_get_chunks", "(", "chunksize", ",", "*", "iterables", ")", ",", "timeout", "=", "timeout", ")", "return", "_chain_from_iterable_of_lists", "(", "results", ")" ]
dc2d941d8285a96f3a5b666a4bd04875b0b25984
test
wrap_non_picklable_objects
Wrapper for non-picklable object to use cloudpickle to serialize them. Note that this wrapper tends to slow down the serialization process as it is done with cloudpickle which is typically slower compared to pickle. The proper way to solve serialization issues is to avoid defining functions and objects in the main scripts and to implement __reduce__ functions for complex classes.
loky/cloudpickle_wrapper.py
def wrap_non_picklable_objects(obj, keep_wrapper=True): """Wrapper for non-picklable object to use cloudpickle to serialize them. Note that this wrapper tends to slow down the serialization process as it is done with cloudpickle which is typically slower compared to pickle. The proper way to solve serialization issues is to avoid defining functions and objects in the main scripts and to implement __reduce__ functions for complex classes. """ if not cloudpickle: raise ImportError("could not import cloudpickle. Please install " "cloudpickle to allow extended serialization. " "(`pip install cloudpickle`).") # If obj is a class, create a CloudpickledClassWrapper which instantiates # the object internally and wrap it directly in a CloudpickledObjectWrapper if inspect.isclass(obj): class CloudpickledClassWrapper(CloudpickledObjectWrapper): def __init__(self, *args, **kwargs): self._obj = obj(*args, **kwargs) self._keep_wrapper = keep_wrapper CloudpickledClassWrapper.__name__ = obj.__name__ return CloudpickledClassWrapper # If obj is an instance of a class, just wrap it in a regular # CloudpickledObjectWrapper return _wrap_non_picklable_objects(obj, keep_wrapper=keep_wrapper)
def wrap_non_picklable_objects(obj, keep_wrapper=True): """Wrapper for non-picklable object to use cloudpickle to serialize them. Note that this wrapper tends to slow down the serialization process as it is done with cloudpickle which is typically slower compared to pickle. The proper way to solve serialization issues is to avoid defining functions and objects in the main scripts and to implement __reduce__ functions for complex classes. """ if not cloudpickle: raise ImportError("could not import cloudpickle. Please install " "cloudpickle to allow extended serialization. " "(`pip install cloudpickle`).") # If obj is a class, create a CloudpickledClassWrapper which instantiates # the object internally and wrap it directly in a CloudpickledObjectWrapper if inspect.isclass(obj): class CloudpickledClassWrapper(CloudpickledObjectWrapper): def __init__(self, *args, **kwargs): self._obj = obj(*args, **kwargs) self._keep_wrapper = keep_wrapper CloudpickledClassWrapper.__name__ = obj.__name__ return CloudpickledClassWrapper # If obj is an instance of a class, just wrap it in a regular # CloudpickledObjectWrapper return _wrap_non_picklable_objects(obj, keep_wrapper=keep_wrapper)
[ "Wrapper", "for", "non", "-", "picklable", "object", "to", "use", "cloudpickle", "to", "serialize", "them", "." ]
tomMoral/loky
python
https://github.com/tomMoral/loky/blob/dc2d941d8285a96f3a5b666a4bd04875b0b25984/loky/cloudpickle_wrapper.py#L86-L113
[ "def", "wrap_non_picklable_objects", "(", "obj", ",", "keep_wrapper", "=", "True", ")", ":", "if", "not", "cloudpickle", ":", "raise", "ImportError", "(", "\"could not import cloudpickle. Please install \"", "\"cloudpickle to allow extended serialization. \"", "\"(`pip install cloudpickle`).\"", ")", "# If obj is a class, create a CloudpickledClassWrapper which instantiates", "# the object internally and wrap it directly in a CloudpickledObjectWrapper", "if", "inspect", ".", "isclass", "(", "obj", ")", ":", "class", "CloudpickledClassWrapper", "(", "CloudpickledObjectWrapper", ")", ":", "def", "__init__", "(", "self", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "self", ".", "_obj", "=", "obj", "(", "*", "args", ",", "*", "*", "kwargs", ")", "self", ".", "_keep_wrapper", "=", "keep_wrapper", "CloudpickledClassWrapper", ".", "__name__", "=", "obj", ".", "__name__", "return", "CloudpickledClassWrapper", "# If obj is an instance of a class, just wrap it in a regular", "# CloudpickledObjectWrapper", "return", "_wrap_non_picklable_objects", "(", "obj", ",", "keep_wrapper", "=", "keep_wrapper", ")" ]
dc2d941d8285a96f3a5b666a4bd04875b0b25984