INSTRUCTION
stringlengths
1
8.43k
RESPONSE
stringlengths
75
104k
Parse persons from given datafield.
def _parse_persons(self, datafield, subfield, roles=["aut"]): """ Parse persons from given datafield. Args: datafield (str): code of datafield ("010", "730", etc..) subfield (char): code of subfield ("a", "z", "4", etc..) role (list of str): set to ["any"] for any role, ["aut"] for authors, etc.. For details see http://www.loc.gov/marc/relators/relaterm.html Main records for persons are: "100", "600" and "700", subrecords "c". Returns: list: Person objects. """ # parse authors parsed_persons = [] raw_persons = self.get_subfields(datafield, subfield) for person in raw_persons: # check if person have at least one of the roles specified in # 'roles' parameter of function other_subfields = person.other_subfields if "4" in other_subfields and roles != ["any"]: person_roles = other_subfields["4"] # list of role parameters relevant = any(map(lambda role: role in roles, person_roles)) # skip non-relevant persons if not relevant: continue # result of .strip() is string, so ind1/2 in MARCSubrecord are lost ind1 = person.i1 ind2 = person.i2 person = person.strip() name = "" second_name = "" surname = "" title = "" # here it gets nasty - there is lot of options in ind1/ind2 # parameters if ind1 == "1" and ind2 == " ": if "," in person: surname, name = person.split(",", 1) elif " " in person: surname, name = person.split(" ", 1) else: surname = person if "c" in other_subfields: title = ",".join(other_subfields["c"]) elif ind1 == "0" and ind2 == " ": name = person.strip() if "b" in other_subfields: second_name = ",".join(other_subfields["b"]) if "c" in other_subfields: surname = ",".join(other_subfields["c"]) elif ind1 == "1" and ind2 == "0" or ind1 == "0" and ind2 == "0": name = person.strip() if "c" in other_subfields: title = ",".join(other_subfields["c"]) parsed_persons.append( Person( name.strip(), second_name.strip(), surname.strip(), title.strip() ) ) return parsed_persons
Args: undefined ( optional ): Argument which will be returned if the subname record is not found.
def get_subname(self, undefined=""): """ Args: undefined (optional): Argument, which will be returned if the `subname` record is not found. Returns: str: Subname of the book or `undefined` if `subname` is not \ found. """ return _undefined_pattern( "".join(self.get_subfields("245", "b")), lambda x: x.strip() == "", undefined )
Args: undefined ( optional ): Argument which will be returned if the price record is not found.
def get_price(self, undefined=""): """ Args: undefined (optional): Argument, which will be returned if the `price` record is not found. Returns: str: Price of the book (with currency) or `undefined` if `price` \ is not found. """ return _undefined_pattern( "".join(self.get_subfields("020", "c")), lambda x: x.strip() == "", undefined )
Args: undefined ( optional ): Argument which will be returned if the part record is not found.
def get_part(self, undefined=""): """ Args: undefined (optional): Argument, which will be returned if the `part` record is not found. Returns: str: Which part of the book series is this record or `undefined` \ if `part` is not found. """ return _undefined_pattern( "".join(self.get_subfields("245", "p")), lambda x: x.strip() == "", undefined )
Args: undefined ( optional ): Argument which will be returned if the part_name record is not found.
def get_part_name(self, undefined=""): """ Args: undefined (optional): Argument, which will be returned if the `part_name` record is not found. Returns: str: Name of the part of the series. or `undefined` if `part_name`\ is not found. """ return _undefined_pattern( "".join(self.get_subfields("245", "n")), lambda x: x.strip() == "", undefined )
Args: undefined ( optional ): Argument which will be returned if the publisher record is not found.
def get_publisher(self, undefined=""): """ Args: undefined (optional): Argument, which will be returned if the `publisher` record is not found. Returns: str: Name of the publisher ("``Grada``" for example) or \ `undefined` if `publisher` is not found. """ publishers = set([ remove_hairs_fn(publisher) for publisher in self["260b "] + self["264b"] ]) return _undefined_pattern( ", ".join(publishers), lambda x: x.strip() == "", undefined )
Args: undefined ( optional ): Argument which will be returned if the pub_date record is not found.
def get_pub_date(self, undefined=""): """ Args: undefined (optional): Argument, which will be returned if the `pub_date` record is not found. Returns: str: Date of publication (month and year usually) or `undefined` \ if `pub_date` is not found. """ dates = self["260c "] + self["264c"] def clean_date(date): """ Clean the `date` strings from special characters, but leave sequences of numbers followed by -. So: [2015]- -> 2015 2015- -> 2015- """ out = "" was_digit = False for c in date: if c.isdigit() or (c == "-" and was_digit) or c == " ": out += c was_digit = c.isdigit() return out # clean all the date strings dates = set([ clean_date(date) for date in self["260c "] + self["264c"] ]) return _undefined_pattern( ", ".join(dates), lambda x: x.strip() == "", undefined )
Args: undefined ( optional ): Argument which will be returned if the pub_order record is not found.
def get_pub_order(self, undefined=""): """ Args: undefined (optional): Argument, which will be returned if the `pub_order` record is not found. Returns: str: Information about order in which was the book published or \ `undefined` if `pub_order` is not found. """ return _undefined_pattern( "".join(self.get_subfields("901", "f")), lambda x: x.strip() == "", undefined )
Args: undefined ( optional ): Argument which will be returned if the pub_place record is not found.
def get_pub_place(self, undefined=""): """ Args: undefined (optional): Argument, which will be returned if the `pub_place` record is not found. Returns: str: Name of city/country where the book was published or \ `undefined` if `pub_place` is not found. """ places = set([ remove_hairs_fn(place) for place in self["260a "] + self["264a"] ]) return _undefined_pattern( ", ".join(places), lambda x: x.strip() == "", undefined )
Args: undefined ( optional ): Argument which will be returned if the format record is not found.
def get_format(self, undefined=""): """ Args: undefined (optional): Argument, which will be returned if the `format` record is not found. Returns: str: Dimensions of the book ('``23 cm``' for example) or `undefined` if `format` is not found. """ return _undefined_pattern( "".join(self.get_subfields("300", "c")), lambda x: x.strip() == "", undefined )
Returns: list: Authors represented as: class:. Person objects.
def get_authors(self): """ Returns: list: Authors represented as :class:`.Person` objects. """ authors = self._parse_persons("100", "a") authors += self._parse_persons("600", "a") authors += self._parse_persons("700", "a") authors += self._parse_persons("800", "a") return authors
Args: roles ( list optional ): Specify which types of corporations you need. Set to [ any ] for any role [ dst ] for distributors etc..
def get_corporations(self, roles=["dst"]): """ Args: roles (list, optional): Specify which types of corporations you need. Set to ``["any"]`` for any role, ``["dst"]`` for distributors, etc.. Note: See http://www.loc.gov/marc/relators/relaterm.html for details. Returns: list: :class:`.Corporation` objects specified by roles parameter. """ corporations = self._parse_corporations("110", "a", roles) corporations += self._parse_corporations("610", "a", roles) corporations += self._parse_corporations("710", "a", roles) corporations += self._parse_corporations("810", "a", roles) return corporations
Get list of VALID ISBN.
def get_ISBNs(self): """ Get list of VALID ISBN. Returns: list: List with *valid* ISBN strings. """ invalid_isbns = set(self.get_invalid_ISBNs()) valid_isbns = [ self._clean_isbn(isbn) for isbn in self["020a"] if self._clean_isbn(isbn) not in invalid_isbns ] if valid_isbns: return valid_isbns # this is used sometimes in czech national library return [ self._clean_isbn(isbn) for isbn in self["901i"] ]
Get list of VALID ISSNs ( 022a ).
def get_ISSNs(self): """ Get list of VALID ISSNs (``022a``). Returns: list: List with *valid* ISSN strings. """ invalid_issns = set(self.get_invalid_ISSNs()) return [ self._clean_isbn(issn) for issn in self["022a"] if self._clean_isbn(issn) not in invalid_issns ]
Filter binding from ISBN record. In MARC XML/ OAI the binding information is stored in same subrecord as ISBN.
def _filter_binding(self, binding): """ Filter binding from ISBN record. In MARC XML / OAI, the binding information is stored in same subrecord as ISBN. Example: ``<subfield code="a">80-251-0225-4 (brož.) :</subfield>`` -> ``brož.``. """ binding = binding.strip().split(" ", 1)[-1] # isolate bind. from ISBN binding = remove_hairs_fn(binding) # remove special chars from binding return binding.split(":")[-1].strip()
Content of field 856u42. Typically URL pointing to producers homepage.
def get_urls(self): """ Content of field ``856u42``. Typically URL pointing to producers homepage. Returns: list: List of URLs defined by producer. """ urls = self.get_subfields("856", "u", i1="4", i2="2") return map(lambda x: x.replace("&amp;", "&"), urls)
URL s which may point to edeposit aleph kramerius and so on.
def get_internal_urls(self): """ URL's, which may point to edeposit, aleph, kramerius and so on. Fields ``856u40``, ``998a`` and ``URLu``. Returns: list: List of internal URLs. """ internal_urls = self.get_subfields("856", "u", i1="4", i2="0") internal_urls.extend(self.get_subfields("998", "a")) internal_urls.extend(self.get_subfields("URL", "u")) return map(lambda x: x.replace("&amp;", "&"), internal_urls)
Returns: PublicationType:: class:. PublicationType enum ** value **.
def get_pub_type(self): """ Returns: PublicationType: :class:`.PublicationType` enum **value**. """ INFO_CHAR_INDEX = 6 SECOND_INFO_CHAR_I = 18 if not len(self.leader) >= INFO_CHAR_INDEX + 1: return PublicationType.monographic if self.controlfields.get("FMT") == "SE": return PublicationType.continuing info_char = self.leader[INFO_CHAR_INDEX] multipart_n = self.get_subfields("245", "n", exception=False) multipart_p = self.get_subfields("245", "p", exception=False) if info_char in "acd": return PublicationType.monographic elif info_char in "bis": return PublicationType.continuing elif info_char == "m" and (multipart_n or multipart_p): return PublicationType.multipart_monograph elif info_char == "m" and len(self.leader) >= SECOND_INFO_CHAR_I + 1: if self.leader[SECOND_INFO_CHAR_I] == "a": return PublicationType.multipart_monograph elif self.leader[SECOND_INFO_CHAR_I] == " ": return PublicationType.single_unit return PublicationType.monographic
Standard dict - like. get () method.
def get(self, item, alt=None): """ Standard dict-like .get() method. Args: item (str): See :meth:`.__getitem__` for details. alt (default None): Alternative value, if item is not found. Returns: obj: `item` or `alt`, if item is not found. """ try: val = self[item] except ValueError: return alt return val if val is not None else alt
r Create a callable that implements a PID controller.
def pid(kp=0., ki=0., kd=0., smooth=0.1): r'''Create a callable that implements a PID controller. A PID controller returns a control signal :math:`u(t)` given a history of error measurements :math:`e(0) \dots e(t)`, using proportional (P), integral (I), and derivative (D) terms, according to: .. math:: u(t) = kp * e(t) + ki * \int_{s=0}^t e(s) ds + kd * \frac{de(s)}{ds}(t) The proportional term is just the current error, the integral term is the sum of all error measurements, and the derivative term is the instantaneous derivative of the error measurement. Parameters ---------- kp : float The weight associated with the proportional term of the PID controller. ki : float The weight associated with the integral term of the PID controller. kd : float The weight associated with the derivative term of the PID controller. smooth : float in [0, 1] Derivative values will be smoothed with this exponential average. A value of 1 never incorporates new derivative information, a value of 0.5 uses the mean of the historic and new information, and a value of 0 discards historic information (i.e., the derivative in this case will be unsmoothed). The default is 0.1. Returns ------- controller : callable (float, float) -> float Returns a function that accepts an error measurement and a delta-time value since the previous measurement, and returns a control signal. ''' state = dict(p=0, i=0, d=0) def control(error, dt=1): state['d'] = smooth * state['d'] + (1 - smooth) * (error - state['p']) / dt state['i'] += error * dt state['p'] = error return kp * state['p'] + ki * state['i'] + kd * state['d'] return control
Given a sequence of sequences return a flat numpy array.
def as_flat_array(iterables): '''Given a sequence of sequences, return a flat numpy array. Parameters ---------- iterables : sequence of sequence of number A sequence of tuples or lists containing numbers. Typically these come from something that represents each joint in a skeleton, like angle. Returns ------- ndarray : An array of flattened data from each of the source iterables. ''' arr = [] for x in iterables: arr.extend(x) return np.array(arr)
Load a skeleton definition from a file.
def load(self, source, **kwargs): '''Load a skeleton definition from a file. Parameters ---------- source : str or file A filename or file-like object that contains text information describing a skeleton. See :class:`pagoda.parser.Parser` for more information about the format of the text file. ''' if hasattr(source, 'endswith') and source.lower().endswith('.asf'): self.load_asf(source, **kwargs) else: self.load_skel(source, **kwargs)
Load a skeleton definition from a text file.
def load_skel(self, source, **kwargs): '''Load a skeleton definition from a text file. Parameters ---------- source : str or file A filename or file-like object that contains text information describing a skeleton. See :class:`pagoda.parser.BodyParser` for more information about the format of the text file. ''' logging.info('%s: parsing skeleton configuration', source) if hasattr(source, 'read'): p = parser.parse(source, self.world, self.jointgroup, **kwargs) else: with open(source) as handle: p = parser.parse(handle, self.world, self.jointgroup, **kwargs) self.bodies = p.bodies self.joints = p.joints self.set_pid_params(kp=0.999 / self.world.dt)
Load a skeleton definition from an ASF text file.
def load_asf(self, source, **kwargs): '''Load a skeleton definition from an ASF text file. Parameters ---------- source : str or file A filename or file-like object that contains text information describing a skeleton, in ASF format. ''' if hasattr(source, 'read'): p = parser.parse_asf(source, self.world, self.jointgroup, **kwargs) else: with open(source) as handle: p = parser.parse_asf(handle, self.world, self.jointgroup, **kwargs) self.bodies = p.bodies self.joints = p.joints self.set_pid_params(kp=0.999 / self.world.dt)
Set PID parameters for all joints in the skeleton.
def set_pid_params(self, *args, **kwargs): '''Set PID parameters for all joints in the skeleton. Parameters for this method are passed directly to the `pid` constructor. ''' for joint in self.joints: joint.target_angles = [None] * joint.ADOF joint.controllers = [pid(*args, **kwargs) for i in range(joint.ADOF)]
Get a list of all current joint torques in the skeleton.
def joint_torques(self): '''Get a list of all current joint torques in the skeleton.''' return as_flat_array(getattr(j, 'amotor', j).feedback[-1][:j.ADOF] for j in self.joints)
Get a list of the indices for a specific joint.
def indices_for_joint(self, name): '''Get a list of the indices for a specific joint. Parameters ---------- name : str The name of the joint to look up. Returns ------- list of int : A list of the index values for quantities related to the named joint. Often useful for getting, say, the angles for a specific joint in the skeleton. ''' j = 0 for joint in self.joints: if joint.name == name: return list(range(j, j + joint.ADOF)) j += joint.ADOF return []
Get a list of the indices for a specific body.
def indices_for_body(self, name, step=3): '''Get a list of the indices for a specific body. Parameters ---------- name : str The name of the body to look up. step : int, optional The number of numbers for each body. Defaults to 3, should be set to 4 for body rotation (since quaternions have 4 values). Returns ------- list of int : A list of the index values for quantities related to the named body. ''' for j, body in enumerate(self.bodies): if body.name == name: return list(range(j * step, (j + 1) * step)) return []
Get the current joint separations for the skeleton.
def joint_distances(self): '''Get the current joint separations for the skeleton. Returns ------- distances : list of float A list expressing the distance between the two joint anchor points, for each joint in the skeleton. These quantities describe how "exploded" the bodies in the skeleton are; a value of 0 indicates that the constraints are perfectly satisfied for that joint. ''' return [((np.array(j.anchor) - j.anchor2) ** 2).sum() for j in self.joints]
Enable the joint motors in this skeleton.
def enable_motors(self, max_force): '''Enable the joint motors in this skeleton. This method sets the maximum force that can be applied by each joint to attain the desired target velocities. It also enables torque feedback for all joint motors. Parameters ---------- max_force : float The maximum force that each joint is allowed to apply to attain its target velocity. ''' for joint in self.joints: amotor = getattr(joint, 'amotor', joint) amotor.max_forces = max_force if max_force > 0: amotor.enable_feedback() else: amotor.disable_feedback()
Move each joint toward a target angle.
def set_target_angles(self, angles): '''Move each joint toward a target angle. This method uses a PID controller to set a target angular velocity for each degree of freedom in the skeleton, based on the difference between the current and the target angle for the respective DOF. PID parameters are by default set to achieve a tiny bit less than complete convergence in one time step, using only the P term (i.e., the P coefficient is set to 1 - \delta, while I and D coefficients are set to 0). PID parameters can be updated by calling the `set_pid_params` method. Parameters ---------- angles : list of float A list of the target angles for every joint in the skeleton. ''' j = 0 for joint in self.joints: velocities = [ ctrl(tgt - cur, self.world.dt) for cur, tgt, ctrl in zip(joint.angles, angles[j:j+joint.ADOF], joint.controllers)] joint.velocities = velocities j += joint.ADOF
Add torques for each degree of freedom in the skeleton.
def add_torques(self, torques): '''Add torques for each degree of freedom in the skeleton. Parameters ---------- torques : list of float A list of the torques to add to each degree of freedom in the skeleton. ''' j = 0 for joint in self.joints: joint.add_torques( list(torques[j:j+joint.ADOF]) + [0] * (3 - joint.ADOF)) j += joint.ADOF
Return the names of our marker labels in canonical order.
def labels(self): '''Return the names of our marker labels in canonical order.''' return sorted(self.channels, key=lambda c: self.channels[c])
Load marker data from a CSV file.
def load_csv(self, filename, start_frame=10, max_frames=int(1e300)): '''Load marker data from a CSV file. The file will be imported using Pandas, which must be installed to use this method. (``pip install pandas``) The first line of the CSV file will be used for header information. The "time" column will be used as the index for the data frame. There must be columns named 'markerAB-foo-x','markerAB-foo-y','markerAB-foo-z', and 'markerAB-foo-c' for marker 'foo' to be included in the model. Parameters ---------- filename : str Name of the CSV file to load. ''' import pandas as pd compression = None if filename.endswith('.gz'): compression = 'gzip' df = pd.read_csv(filename, compression=compression).set_index('time').fillna(-1) # make sure the data frame's time index matches our world. assert self.world.dt == pd.Series(df.index).diff().mean() markers = [] for c in df.columns: m = re.match(r'^marker\d\d-(.*)-c$', c) if m: markers.append(m.group(1)) self.channels = self._map_labels_to_channels(markers) cols = [c for c in df.columns if re.match(r'^marker\d\d-.*-[xyzc]$', c)] self.data = df[cols].values.reshape((len(df), len(markers), 4))[start_frame:] self.data[:, :, [1, 2]] = self.data[:, :, [2, 1]] logging.info('%s: loaded marker data %s', filename, self.data.shape) self.process_data() self.create_bodies()
Load marker data from a C3D file.
def load_c3d(self, filename, start_frame=0, max_frames=int(1e300)): '''Load marker data from a C3D file. The file will be imported using the c3d module, which must be installed to use this method. (``pip install c3d``) Parameters ---------- filename : str Name of the C3D file to load. start_frame : int, optional Discard the first N frames. Defaults to 0. max_frames : int, optional Maximum number of frames to load. Defaults to loading all frames. ''' import c3d with open(filename, 'rb') as handle: reader = c3d.Reader(handle) logging.info('world frame rate %s, marker frame rate %s', 1 / self.world.dt, reader.point_rate) # set up a map from marker label to index in the data stream. self.channels = self._map_labels_to_channels([ s.strip() for s in reader.point_labels]) # read the actual c3d data into a numpy array. data = [] for i, (_, frame, _) in enumerate(reader.read_frames()): if i >= start_frame: data.append(frame[:, [0, 1, 2, 4]]) if len(data) > max_frames: break self.data = np.array(data) # scale the data to meters -- mm is a very common C3D unit. if reader.get('POINT:UNITS').string_value.strip().lower() == 'mm': logging.info('scaling point data from mm to m') self.data[:, :, :3] /= 1000. logging.info('%s: loaded marker data %s', filename, self.data.shape) self.process_data() self.create_bodies()
Process data to produce velocity and dropout information.
def process_data(self): '''Process data to produce velocity and dropout information.''' self.visibility = self.data[:, :, 3] self.positions = self.data[:, :, :3] self.velocities = np.zeros_like(self.positions) + 1000 for frame_no in range(1, len(self.data) - 1): prev = self.data[frame_no - 1] next = self.data[frame_no + 1] for c in range(self.num_markers): if -1 < prev[c, 3] < 100 and -1 < next[c, 3] < 100: self.velocities[frame_no, c] = ( next[c, :3] - prev[c, :3]) / (2 * self.world.dt) self.cfms = np.zeros_like(self.visibility) + self.DEFAULT_CFM
Create physics bodies corresponding to each marker in our data.
def create_bodies(self): '''Create physics bodies corresponding to each marker in our data.''' self.bodies = {} for label in self.channels: body = self.world.create_body( 'sphere', name='marker:{}'.format(label), radius=0.02) body.is_kinematic = True body.color = 0.9, 0.1, 0.1, 0.5 self.bodies[label] = body
Load attachment configuration from the given text source.
def load_attachments(self, source, skeleton): '''Load attachment configuration from the given text source. The attachment configuration file has a simple format. After discarding Unix-style comments (any part of a line that starts with the pound (#) character), each line in the file is then expected to have the following format:: marker-name body-name X Y Z The marker name must correspond to an existing "channel" in our marker data. The body name must correspond to a rigid body in the skeleton. The X, Y, and Z coordinates specify the body-relative offsets where the marker should be attached: 0 corresponds to the center of the body along the given axis, while -1 and 1 correspond to the minimal (maximal, respectively) extent of the body's bounding box along the corresponding dimension. Parameters ---------- source : str or file-like A filename or file-like object that we can use to obtain text configuration that describes how markers are attached to skeleton bodies. skeleton : :class:`pagoda.skeleton.Skeleton` The skeleton to attach our marker data to. ''' self.targets = {} self.offsets = {} filename = source if isinstance(source, str): source = open(source) else: filename = '(file-{})'.format(id(source)) for i, line in enumerate(source): tokens = line.split('#')[0].strip().split() if not tokens: continue label = tokens.pop(0) if label not in self.channels: logging.info('%s:%d: unknown marker %s', filename, i, label) continue if not tokens: continue name = tokens.pop(0) bodies = [b for b in skeleton.bodies if b.name == name] if len(bodies) != 1: logging.info('%s:%d: %d skeleton bodies match %s', filename, i, len(bodies), name) continue b = self.targets[label] = bodies[0] o = self.offsets[label] = \ np.array(list(map(float, tokens))) * b.dimensions / 2 logging.info('%s <--> %s, offset %s', label, b.name, o)
Attach marker bodies to the corresponding skeleton bodies.
def attach(self, frame_no): '''Attach marker bodies to the corresponding skeleton bodies. Attachments are only made for markers that are not in a dropout state in the given frame. Parameters ---------- frame_no : int The frame of data we will use for attaching marker bodies. ''' assert not self.joints for label, j in self.channels.items(): target = self.targets.get(label) if target is None: continue if self.visibility[frame_no, j] < 0: continue if np.linalg.norm(self.velocities[frame_no, j]) > 10: continue joint = ode.BallJoint(self.world.ode_world, self.jointgroup) joint.attach(self.bodies[label].ode_body, target.ode_body) joint.setAnchor1Rel([0, 0, 0]) joint.setAnchor2Rel(self.offsets[label]) joint.setParam(ode.ParamCFM, self.cfms[frame_no, j]) joint.setParam(ode.ParamERP, self.erp) joint.name = label self.joints[label] = joint self._frame_no = frame_no
Reposition markers to a specific frame of data.
def reposition(self, frame_no): '''Reposition markers to a specific frame of data. Parameters ---------- frame_no : int The frame of data where we should reposition marker bodies. Markers will be positioned in the appropriate places in world coordinates. In addition, linear velocities of the markers will be set according to the data as long as there are no dropouts in neighboring frames. ''' for label, j in self.channels.items(): body = self.bodies[label] body.position = self.positions[frame_no, j] body.linear_velocity = self.velocities[frame_no, j]
Get a list of the distances between markers and their attachments.
def distances(self): '''Get a list of the distances between markers and their attachments. Returns ------- distances : ndarray of shape (num-markers, 3) Array of distances for each marker joint in our attachment setup. If a marker does not currently have an associated joint (e.g. because it is not currently visible) this will contain NaN for that row. ''' distances = [] for label in self.labels: joint = self.joints.get(label) distances.append([np.nan, np.nan, np.nan] if joint is None else np.array(joint.getAnchor()) - joint.getAnchor2()) return np.array(distances)
Return an array of the forces exerted by marker springs.
def forces(self, dx_tm1=None): '''Return an array of the forces exerted by marker springs. Notes ----- The forces exerted by the marker springs can be approximated by:: F = kp * dx where ``dx`` is the current array of marker distances. An even more accurate value is computed by approximating the velocity of the spring displacement:: F = kp * dx + kd * (dx - dx_tm1) / dt where ``dx_tm1`` is an array of distances from the previous time step. Parameters ---------- dx_tm1 : ndarray An array of distances from markers to their attachment targets, measured at the previous time step. Returns ------- F : ndarray An array of forces that the markers are exerting on the skeleton. ''' cfm = self.cfms[self._frame_no][:, None] kp = self.erp / (cfm * self.world.dt) kd = (1 - self.erp) / cfm dx = self.distances() F = kp * dx if dx_tm1 is not None: bad = np.isnan(dx) | np.isnan(dx_tm1) F[~bad] += (kd * (dx - dx_tm1) / self.world.dt)[~bad] return F
Create and configure a skeleton in our model.
def load_skeleton(self, filename, pid_params=None): '''Create and configure a skeleton in our model. Parameters ---------- filename : str The name of a file containing skeleton configuration data. pid_params : dict, optional If given, use this dictionary to set the PID controller parameters on each joint in the skeleton. See :func:`pagoda.skeleton.pid` for more information. ''' self.skeleton = skeleton.Skeleton(self) self.skeleton.load(filename, color=(0.3, 0.5, 0.9, 0.8)) if pid_params: self.skeleton.set_pid_params(**pid_params) self.skeleton.erp = 0.1 self.skeleton.cfm = 0
Load marker data and attachment preferences into the model.
def load_markers(self, filename, attachments, max_frames=1e100): '''Load marker data and attachment preferences into the model. Parameters ---------- filename : str The name of a file containing marker data. This currently needs to be either a .C3D or a .CSV file. CSV files must adhere to a fairly strict column naming convention; see :func:`Markers.load_csv` for more information. attachments : str The name of a text file specifying how markers are attached to skeleton bodies. max_frames : number, optional Only read in this many frames of marker data. By default, the entire data file is read into memory. Returns ------- markers : :class:`Markers` Returns a markers object containing loaded marker data as well as skeleton attachment configuration. ''' self.markers = Markers(self) fn = filename.lower() if fn.endswith('.c3d'): self.markers.load_c3d(filename, max_frames=max_frames) elif fn.endswith('.csv') or fn.endswith('.csv.gz'): self.markers.load_csv(filename, max_frames=max_frames) else: logging.fatal('%s: not sure how to load markers!', filename) self.markers.load_attachments(attachments, self.skeleton)
Advance the physics world by one step.
def step(self, substeps=2): '''Advance the physics world by one step. Typically this is called as part of a :class:`pagoda.viewer.Viewer`, but it can also be called manually (or some other stepping mechanism entirely can be used). ''' # by default we step by following our loaded marker data. self.frame_no += 1 try: next(self.follower) except (AttributeError, StopIteration) as err: self.reset()
Settle the skeleton to our marker data at a specific frame.
def settle_to_markers(self, frame_no=0, max_distance=0.05, max_iters=300, states=None): '''Settle the skeleton to our marker data at a specific frame. Parameters ---------- frame_no : int, optional Settle the skeleton to marker data at this frame. Defaults to 0. max_distance : float, optional The settling process will stop when the mean marker distance falls below this threshold. Defaults to 0.1m (10cm). Setting this too small prevents the settling process from finishing (it will loop indefinitely), and setting it too large prevents the skeleton from settling to a stable state near the markers. max_iters : int, optional Attempt to settle markers for at most this many iterations. Defaults to 1000. states : list of body states, optional If given, set the bodies in our skeleton to these kinematic states before starting the settling process. ''' if states is not None: self.skeleton.set_body_states(states) dist = None for _ in range(max_iters): for _ in self._step_to_marker_frame(frame_no): pass dist = np.nanmean(abs(self.markers.distances())) logging.info('settling to frame %d: marker distance %.3f', frame_no, dist) if dist < max_distance: return self.skeleton.get_body_states() for b in self.skeleton.bodies: b.linear_velocity = 0, 0, 0 b.angular_velocity = 0, 0, 0 return states
Iterate over a set of marker data dragging its skeleton along.
def follow_markers(self, start=0, end=1e100, states=None): '''Iterate over a set of marker data, dragging its skeleton along. Parameters ---------- start : int, optional Start following marker data after this frame. Defaults to 0. end : int, optional Stop following marker data after this frame. Defaults to the end of the marker data. states : list of body states, optional If given, set the states of the skeleton bodies to these values before starting to follow the marker data. ''' if states is not None: self.skeleton.set_body_states(states) for frame_no, frame in enumerate(self.markers): if frame_no < start: continue if frame_no >= end: break for states in self._step_to_marker_frame(frame_no): yield states
Update the simulator to a specific frame of marker data.
def _step_to_marker_frame(self, frame_no, dt=None): '''Update the simulator to a specific frame of marker data. This method returns a generator of body states for the skeleton! This generator must be exhausted (e.g., by consuming this call in a for loop) for the simulator to work properly. This process involves the following steps: - Move the markers to their new location: - Detach from the skeleton - Update marker locations - Reattach to the skeleton - Detect ODE collisions - Yield the states of the bodies in the skeleton - Advance the ODE world one step Parameters ---------- frame_no : int Step to this frame of marker data. dt : float, optional Step with this time duration. Defaults to ``self.dt``. Returns ------- states : sequence of state tuples A generator of a sequence of one body state for the skeleton. This generator must be exhausted for the simulation to work properly. ''' # update the positions and velocities of the markers. self.markers.detach() self.markers.reposition(frame_no) self.markers.attach(frame_no) # detect collisions. self.ode_space.collide(None, self.on_collision) # record the state of each skeleton body. states = self.skeleton.get_body_states() self.skeleton.set_body_states(states) # yield the current simulation state to our caller. yield states # update the ode world. self.ode_world.step(dt or self.dt) # clear out contact joints to prepare for the next frame. self.ode_contactgroup.empty()
Follow a set of marker data yielding kinematic joint angles.
def inverse_kinematics(self, start=0, end=1e100, states=None, max_force=20): '''Follow a set of marker data, yielding kinematic joint angles. Parameters ---------- start : int, optional Start following marker data after this frame. Defaults to 0. end : int, optional Stop following marker data after this frame. Defaults to the end of the marker data. states : list of body states, optional If given, set the states of the skeleton bodies to these values before starting to follow the marker data. max_force : float, optional Allow each degree of freedom in the skeleton to exert at most this force when attempting to maintain its equilibrium position. This defaults to 20N. Set this value higher to simulate a stiff skeleton while following marker data. Returns ------- angles : sequence of angle frames Returns a generator of joint angle data for the skeleton. One set of joint angles will be generated for each frame of marker data between `start` and `end`. ''' zeros = None if max_force > 0: self.skeleton.enable_motors(max_force) zeros = np.zeros(self.skeleton.num_dofs) for _ in self.follow_markers(start, end, states): if zeros is not None: self.skeleton.set_target_angles(zeros) yield self.skeleton.joint_angles
Follow a set of angle data yielding dynamic joint torques.
def inverse_dynamics(self, angles, start=0, end=1e100, states=None, max_force=100): '''Follow a set of angle data, yielding dynamic joint torques. Parameters ---------- angles : ndarray (num-frames x num-dofs) Follow angle data provided by this array of angle values. start : int, optional Start following angle data after this frame. Defaults to the start of the angle data. end : int, optional Stop following angle data after this frame. Defaults to the end of the angle data. states : list of body states, optional If given, set the states of the skeleton bodies to these values before starting to follow the marker data. max_force : float, optional Allow each degree of freedom in the skeleton to exert at most this force when attempting to follow the given joint angles. Defaults to 100N. Setting this value to be large results in more accurate following but can cause oscillations in the PID controllers, resulting in noisy torques. Returns ------- torques : sequence of torque frames Returns a generator of joint torque data for the skeleton. One set of joint torques will be generated for each frame of angle data between `start` and `end`. ''' if states is not None: self.skeleton.set_body_states(states) for frame_no, frame in enumerate(angles): if frame_no < start: continue if frame_no >= end: break self.ode_space.collide(None, self.on_collision) states = self.skeleton.get_body_states() self.skeleton.set_body_states(states) # joseph's stability fix: step to compute torques, then reset the # skeleton to the start of the step, and then step using computed # torques. thus any numerical errors between the body states after # stepping using angle constraints will be removed, because we # will be stepping the model using the computed torques. self.skeleton.enable_motors(max_force) self.skeleton.set_target_angles(angles[frame_no]) self.ode_world.step(self.dt) torques = self.skeleton.joint_torques self.skeleton.disable_motors() self.skeleton.set_body_states(states) self.skeleton.add_torques(torques) yield torques self.ode_world.step(self.dt) self.ode_contactgroup.empty()
Move the body according to a set of torque data.
def forward_dynamics(self, torques, start=0, states=None): '''Move the body according to a set of torque data.''' if states is not None: self.skeleton.set_body_states(states) for frame_no, torque in enumerate(torques): if frame_no < start: continue if frame_no >= end: break self.ode_space.collide(None, self.on_collision) self.skeleton.add_torques(torque) self.ode_world.step(self.dt) yield self.ode_contactgroup.empty()
Sort values but put numbers after alphabetically sorted words.
def resorted(values): """ Sort values, but put numbers after alphabetically sorted words. This function is here to make outputs diff-compatible with Aleph. Example:: >>> sorted(["b", "1", "a"]) ['1', 'a', 'b'] >>> resorted(["b", "1", "a"]) ['a', 'b', '1'] Args: values (iterable): any iterable object/list/tuple/whatever. Returns: list of sorted values, but with numbers after words """ if not values: return values values = sorted(values) # look for first word first_word = next( (cnt for cnt, val in enumerate(values) if val and not val[0].isdigit()), None ) # if not found, just return the values if first_word is None: return values words = values[first_word:] numbers = values[:first_word] return words + numbers
Draw all bodies in the world.
def render(self, dt): '''Draw all bodies in the world.''' for frame in self._frozen: for body in frame: self.draw_body(body) for body in self.world.bodies: self.draw_body(body) if hasattr(self.world, 'markers'): # draw line between anchor1 and anchor2 for marker joints. window.glColor4f(0.9, 0.1, 0.1, 0.9) window.glLineWidth(3) for j in self.world.markers.joints.values(): window.glBegin(window.GL_LINES) window.glVertex3f(*j.getAnchor()) window.glVertex3f(*j.getAnchor2()) window.glEnd()
Get room stream to listen for messages.
def get_stream(self, error_callback=None, live=True): """ Get room stream to listen for messages. Kwargs: error_callback (func): Callback to call when an error occurred (parameters: exception) live (bool): If True, issue a live stream, otherwise an offline stream Returns: :class:`Stream`. Stream """ self.join() return Stream(self, error_callback=error_callback, live=live)
Get list of users in the room.
def get_users(self, sort=True): """ Get list of users in the room. Kwargs: sort (bool): If True, sort rooms by name Returns: array. List of users """ self._load() if sort: self.users.sort(key=operator.itemgetter("name")) return self.users
Recent messages.
def recent(self, message_id=None, limit=None): """ Recent messages. Kwargs: message_id (int): If specified, return messages since the specified message ID limit (int): If specified, limit the number of messages Returns: array. Messages """ parameters = {} if message_id: parameters["since_message_id"] = message_id if limit: parameters["limit"] = limit messages = self._connection.get("room/%s/recent" % self.id, key="messages", parameters=parameters) if messages: messages = [Message(self._campfire, message) for message in messages] return messages
Set the room name.
def set_name(self, name): """ Set the room name. Args: name (str): Name Returns: bool. Success """ if not self._campfire.get_user().admin: return False result = self._connection.put("room/%s" % self.id, {"room": {"name": name}}) if result["success"]: self._load() return result["success"]
Set the room topic.
def set_topic(self, topic): """ Set the room topic. Args: topic (str): Topic Returns: bool. Success """ if not topic: topic = '' result = self._connection.put("room/%s" % self.id, {"room": {"topic": topic}}) if result["success"]: self._load() return result["success"]
Post a message.
def speak(self, message): """ Post a message. Args: message (:class:`Message` or string): Message Returns: bool. Success """ campfire = self.get_campfire() if not isinstance(message, Message): message = Message(campfire, message) result = self._connection.post( "room/%s/speak" % self.id, {"message": message.get_data()}, parse_data=True, key="message" ) if result["success"]: return Message(campfire, result["data"]) return result["success"]
Recent messages.
def transcript(self, for_date=None): """ Recent messages. Kwargs: for_date (date): If specified, get the transcript for this specific date Returns: array. Messages """ url = "room/%s/transcript" % self.id if for_date: url = "%s/%d/%d/%d" % (url, for_date.year, for_date.month, for_date.day) messages = self._connection.get(url, key="messages") if messages: messages = [Message(self._campfire, message) for message in messages] return messages
Create a new thread to upload a file ( thread should be then started with start () to perform upload. )
def upload(self, path, progress_callback=None, finished_callback=None, error_callback=None): """ Create a new thread to upload a file (thread should be then started with start() to perform upload.) Args: path (str): Path to file Kwargs: progress_callback (func): Callback to call as file is uploaded (parameters: current, total) finished_callback (func): Callback to call when upload is finished error_callback (func): Callback to call when an error occurred (parameters: exception) Returns: :class:`Upload`. Upload thread """ return Upload( self, {"upload": path}, progress_callback = progress_callback, finished_callback = finished_callback, error_callback = error_callback )
Build a call to use the new get_config function from args passed to Config. __init__.
def get_new_call(group_name, app_name, search_path, filename, require_load, version, secure): # type: (str, str, Optional[str], str, bool, Optional[str], bool) -> str ''' Build a call to use the new ``get_config`` function from args passed to ``Config.__init__``. ''' new_call_kwargs = { 'group_name': group_name, 'filename': filename } # type: Dict[str, Any] new_call_lookup_options = {} # type: Dict[str, Any] new_call_lookup_options['secure'] = secure if search_path: new_call_lookup_options['search_path'] = search_path if require_load: new_call_lookup_options['require_load'] = require_load if version: new_call_lookup_options['version'] = version if new_call_lookup_options: new_call_kwargs['lookup_options'] = new_call_lookup_options output = build_call_str('get_config', (app_name,), new_call_kwargs) return output
Build a callable Python string for a function call. The output will be combined similar to this template::
def build_call_str(prefix, args, kwargs): # type: (str, Any, Any) -> str ''' Build a callable Python string for a function call. The output will be combined similar to this template:: <prefix>(<args>, <kwargs>) Example:: >>> build_call_str('foo', (1, 2), {'a': '10'}) "foo(1, 2, a='10')" ''' kwargs_str = ', '.join(['%s=%r' % (key, value) for key, value in kwargs.items()]) args_str = ', '.join([repr(arg) for arg in args]) output = [prefix, '('] if args: output.append(args_str) if args and kwargs: output.append(', ') if kwargs: output.append(kwargs_str) output.append(')') return ''.join(output)
Returns a list of paths specified by the XDG_CONFIG_DIRS environment variable or the appropriate default.
def get_xdg_dirs(self): # type: () -> List[str] """ Returns a list of paths specified by the XDG_CONFIG_DIRS environment variable or the appropriate default. The list is sorted by precedence, with the most important item coming *last* (required by the existing config_resolver logic). """ config_dirs = getenv('XDG_CONFIG_DIRS', '') if config_dirs: self._log.debug('XDG_CONFIG_DIRS is set to %r', config_dirs) output = [] for path in reversed(config_dirs.split(':')): output.append(join(path, self.group_name, self.app_name)) return output return ['/etc/xdg/%s/%s' % (self.group_name, self.app_name)]
Returns the value specified in the XDG_CONFIG_HOME environment variable or the appropriate default.
def get_xdg_home(self): # type: () -> str """ Returns the value specified in the XDG_CONFIG_HOME environment variable or the appropriate default. """ config_home = getenv('XDG_CONFIG_HOME', '') if config_home: self._log.debug('XDG_CONFIG_HOME is set to %r', config_home) return expanduser(join(config_home, self.group_name, self.app_name)) return expanduser('~/.config/%s/%s' % (self.group_name, self.app_name))
Returns the filename which is effectively used by the application. If overridden by an environment variable it will return that filename.
def _effective_filename(self): # type: () -> str """ Returns the filename which is effectively used by the application. If overridden by an environment variable, it will return that filename. """ # same logic for the configuration filename. First, check if we were # initialized with a filename... config_filename = '' if self.filename: config_filename = self.filename # ... next, take the value from the environment env_filename = getenv(self.env_filename_name) if env_filename: self._log.info('Configuration filename was overridden with %r ' 'by the environment variable %s.', env_filename, self.env_filename_name) config_filename = env_filename return config_filename
Returns a list of paths to search for config files in reverse order of precedence. In other words: the last path element will override the settings from the first one.
def _effective_path(self): # type: () -> List[str] """ Returns a list of paths to search for config files in reverse order of precedence. In other words: the last path element will override the settings from the first one. """ # default search path path = (['/etc/%s/%s' % (self.group_name, self.app_name)] + self.get_xdg_dirs() + [expanduser('~/.%s/%s' % (self.group_name, self.app_name)), self.get_xdg_home(), join(getcwd(), '.{}'.format(self.group_name), self.app_name)]) # If a path was passed directly to this instance, override the path. if self.search_path: path = self.search_path.split(pathsep) # Next, consider the environment variables... env_path = getenv(self.env_path_name) if env_path and env_path.startswith('+'): # If prefixed with a '+', append the path elements additional_paths = env_path[1:].split(pathsep) self._log.info('Search path extended with %r by the environment ' 'variable %s.', additional_paths, self.env_path_name) path.extend(additional_paths) elif env_path: # Otherwise, override again. This takes absolute precedence. self._log.info("Configuration search path was overridden with " "%r by the environment variable %r.", env_path, self.env_path_name) path = env_path.split(pathsep) return path
Check if filename can be read. Will return boolean which is True if the file can be read False otherwise.
def check_file(self, filename): # type: (str) -> bool """ Check if ``filename`` can be read. Will return boolean which is True if the file can be read, False otherwise. """ if not exists(filename): return False # Check if the file is version-compatible with this instance. new_config = ConfigResolverBase() new_config.read(filename) if self.version and not new_config.has_option('meta', 'version'): # self.version is set, so we MUST have a version in the file! raise NoVersionError( "The config option 'meta.version' is missing in {}. The " "application expects version {}!".format(filename, self.version)) elif not self.version and new_config.has_option('meta', 'version'): # Automatically "lock-in" a version number if one is found. # This prevents loading a chain of config files with incompatible # version numbers! self.version = StrictVersion(new_config.get('meta', 'version')) self._log.info('%r contains a version number, but the config ' 'instance was not created with a version ' 'restriction. Will set version number to "%s" to ' 'prevent accidents!', filename, self.version) elif self.version: # This instance expected a certain version. We need to check the # version in the file and compare. file_version = new_config.get('meta', 'version') major, minor, _ = StrictVersion(file_version).version expected_major, expected_minor, _ = self.version.version if expected_major != major: self._log.error( 'Invalid major version number in %r. Expected %r, got %r!', abspath(filename), str(self.version), file_version) return False if expected_minor != minor: self._log.warning( 'Mismatching minor version number in %r. ' 'Expected %r, got %r!', abspath(filename), str(self.version), file_version) return True return True
Overrides: py: meth: configparser. ConfigParser. get.
def get(self, section, option, **kwargs): # type: ignore # type: (str, str, Any) -> Any """ Overrides :py:meth:`configparser.ConfigParser.get`. In addition to ``section`` and ``option``, this call takes an optional ``default`` value. This behaviour works in *addition* to the :py:class:`configparser.ConfigParser` default mechanism. Note that a default value from ``ConfigParser`` takes precedence. The reason this additional functionality is added, is because the defaults of :py:class:`configparser.ConfigParser` are not dependent on sections. If you specify a default for the option ``test``, then this value will be returned for both ``section1.test`` and for ``section2.test``. Using the default on the ``get`` call gives you more fine-grained control over this. Also note, that if a default value was used, it will be logged with level ``logging.DEBUG``. :param section: The config file section. :param option: The option name. :param kwargs: These keyword args are passed through to :py:meth:`configparser.ConfigParser.get`. """ if "default" in kwargs: default = kwargs.pop("default") new_kwargs = {'fallback': default} new_kwargs.update(kwargs) new_call = build_call_str('.get', (section, option), new_kwargs) warn('Using the "default" argument to Config.get() will no ' 'longer work in config_resolver 5.0! Version 5 will return ' 'standard Python ConfigParser instances which use "fallback" ' 'instead of "default". Replace your code with "%s"' % new_call, DeprecationWarning, stacklevel=2) have_default = True else: have_default = False try: value = super(Config, self).get(section, option, **kwargs) return value except (NoSectionError, NoOptionError) as exc: if have_default: self._log.debug("%s: Returning default value %r", exc, default) return default else: raise
Searches for an appropriate config file. If found loads the file into the current instance. This method can also be used to reload a configuration. Note that you may want to set reload to True to clear the configuration before loading in that case. Without doing that values will remain available even if they have been removed from the config files.
def load(self, reload=False, require_load=False): # type: (bool, bool) -> None """ Searches for an appropriate config file. If found, loads the file into the current instance. This method can also be used to reload a configuration. Note that you may want to set ``reload`` to ``True`` to clear the configuration before loading in that case. Without doing that, values will remain available even if they have been removed from the config files. :param reload: if set to ``True``, the existing values are cleared before reloading. :param require_load: If set to ``True`` this will raise a :py:exc:`IOError` if no config file has been found to load. """ if reload: # pragma: no cover self.config = None # only load the config if necessary (or explicitly requested) if self.config: # pragma: no cover self._log.debug('Returning cached config instance. Use ' '``reload=True`` to avoid caching!') return path = self._effective_path() config_filename = self._effective_filename() # Next, use the resolved path to find the filenames. Keep track of # which files we loaded in order to inform the user. self._active_path = [join(_, config_filename) for _ in path] for dirname in path: conf_name = join(dirname, config_filename) readable = self.check_file(conf_name) if readable: action = 'Updating' if self._loaded_files else 'Loading initial' self._log.info('%s config from %s', action, conf_name) self.read(conf_name) if conf_name == expanduser("~/.%s/%s/%s" % ( self.group_name, self.app_name, self.filename)): self._log.warning( "DEPRECATION WARNING: The file " "'%s/.%s/%s/app.ini' was loaded. The XDG " "Basedir standard requires this file to be in " "'%s/.config/%s/%s/app.ini'! This location " "will no longer be parsed in a future version of " "config_resolver! You can already (and should) move " "the file!", expanduser("~"), self.group_name, self.app_name, expanduser("~"), self.group_name, self.app_name) self._loaded_files.append(conf_name) if not self._loaded_files and not require_load: self._log.warning( "No config file named %s found! Search path was %r", config_filename, path) elif not self._loaded_files and require_load: raise IOError("No config file named %s found! Search path " "was %r" % (config_filename, path))
Overrides: py: meth:. Config. check_file
def check_file(self, filename): # type: (str) -> bool """ Overrides :py:meth:`.Config.check_file` """ can_read = super(SecuredConfig, self).check_file(filename) if not can_read: return False mode = get_stat(filename).st_mode if (mode & stat.S_IRGRP) or (mode & stat.S_IROTH): msg = "File %r is not secure enough. Change it's mode to 600" self._log.warning(msg, filename) return False return True
https:// www. python. org/ dev/ peps/ pep - 0333/ #environ - variables
def setup_environ(self): """https://www.python.org/dev/peps/pep-0333/#environ-variables""" # Set up base environment env = self.base_environ = {} env['SERVER_NAME'] = self.server_name env['GATEWAY_INTERFACE'] = 'CGI/1.1' env['SERVER_PORT'] = str(self.server_port) env['REMOTE_HOST'] = '' env['CONTENT_LENGTH'] = '' env['SCRIPT_NAME'] = ''
https:// www. python. org/ dev/ peps/ pep - 0333/ #environ - variables
def get_environ(self): """https://www.python.org/dev/peps/pep-0333/#environ-variables""" env = self.base_environ.copy() env['REQUEST_METHOD'] = self.request_method if '?' in self.path: path, query = self.path.split('?', 1) else: path, query = self.path, '' env['PATH_INFO'] = urllib.parse.unquote(path) env['QUERY_STRING'] = query env['CONTENT_TYPE'] = self.headers.get('Content-Type', '') env['CONTENT_LENGTH'] = self.headers.get('Content-Length', '0') env['SERVER_PROTOCOL'] = self.request_version env['REMOTE_ADDR'] = self.client_address[0] env['REMOTE_PORT'] = self.client_address[1] env['wsgi.version'] = (1, 0) env['wsgi.url_scheme'] = 'http' env['wsgi.input'] = io.BytesIO(self.raw_request) env['wsgi.errors'] = sys.stderr env['wsgi.multithread'] = False env['wsgi.multiprocess'] = True env['wsgi.run_once'] = False for k, v in self.headers.items(): k = k.replace('-', '_').upper() if k in env: continue env['HTTP_' + k] = v return env
Get styles.
def get(self, q=None, page=None): """Get styles.""" # Check cache to exit early if needed etag = generate_etag(current_ext.content_version.encode('utf8')) self.check_etag(etag, weak=True) # Build response res = jsonify(current_ext.styles) res.set_etag(etag) return res
Create a connection with given settings.
def create_from_settings(settings): """ Create a connection with given settings. Args: settings (dict): A dictionary of settings Returns: :class:`Connection`. The connection """ return Connection( settings["url"], settings["base_url"], settings["user"], settings["password"], authorizations = settings["authorizations"], debug = settings["debug"] )
Issue a PUT request.
def delete(self, url=None, post_data={}, parse_data=False, key=None, parameters=None): """ Issue a PUT request. Kwargs: url (str): Destination URL post_data (dict): Dictionary of parameter and values parse_data (bool): If true, parse response data key (string): If parse_data==True, look for this key when parsing data parameters (dict): Additional GET parameters to append to the URL Returns: dict. Response (a dict with keys: success, data, info, body) Raises: AuthenticationError, ConnectionError, urllib2.HTTPError, ValueError, Exception """ return self._fetch("DELETE", url, post_data=post_data, parse_data=parse_data, key=key, parameters=parameters, full_return=True)
Issue a POST request.
def post(self, url=None, post_data={}, parse_data=False, key=None, parameters=None, listener=None): """ Issue a POST request. Kwargs: url (str): Destination URL post_data (dict): Dictionary of parameter and values parse_data (bool): If true, parse response data key (string): If parse_data==True, look for this key when parsing data parameters (dict): Additional GET parameters to append to the URL listener (func): callback called when uploading a file Returns: dict. Response (a dict with keys: success, data, info, body) Raises: AuthenticationError, ConnectionError, urllib2.HTTPError, ValueError, Exception """ return self._fetch("POST", url, post_data=post_data, parse_data=parse_data, key=key, parameters=parameters, listener=listener, full_return=True)
Issue a GET request.
def get(self, url=None, parse_data=True, key=None, parameters=None): """ Issue a GET request. Kwargs: url (str): Destination URL parse_data (bool): If true, parse response data key (string): If parse_data==True, look for this key when parsing data parameters (dict): Additional GET parameters to append to the URL Returns: dict. Response (a dict with keys: success, data, info, body) Raises: AuthenticationError, ConnectionError, urllib2.HTTPError, ValueError, Exception """ return self._fetch("GET", url, post_data=None, parse_data=parse_data, key=key, parameters=parameters)
Get headers.
def get_headers(self): """ Get headers. Returns: tuple: Headers """ headers = { "User-Agent": "kFlame 1.0" } password_url = self._get_password_url() if password_url and password_url in self._settings["authorizations"]: headers["Authorization"] = self._settings["authorizations"][password_url] return headers
Get URL used for authentication
def _get_password_url(self): """ Get URL used for authentication Returns: string: URL """ password_url = None if self._settings["user"] or self._settings["authorization"]: if self._settings["url"]: password_url = self._settings["url"] elif self._settings["base_url"]: password_url = self._settings["base_url"] return password_url
Parses a response.
def parse(self, text, key=None): """ Parses a response. Args: text (str): Text to parse Kwargs: key (str): Key to look for, if any Returns: Parsed value Raises: ValueError """ try: data = json.loads(text) except ValueError as e: raise ValueError("%s: Value: [%s]" % (e, text)) if data and key: if key not in data: raise ValueError("Invalid response (key %s not found): %s" % (key, data)) data = data[key] return data
Build a request for twisted
def build_twisted_request(self, method, url, extra_headers={}, body_producer=None, full_url=False): """ Build a request for twisted Args: method (str): Request method (GET/POST/PUT/DELETE/etc.) If not specified, it will be POST if post_data is not None url (str): Destination URL (full, or relative) Kwargs: extra_headers (dict): Headers (override default connection headers, if any) body_producer (:class:`twisted.web.iweb.IBodyProducer`): Object producing request body full_url (bool): If False, URL is relative Returns: tuple. Tuple with two elements: reactor, and request """ uri = url if full_url else self._url(url) raw_headers = self.get_headers() if extra_headers: raw_headers.update(extra_headers) headers = http_headers.Headers() for header in raw_headers: headers.addRawHeader(header, raw_headers[header]) agent = client.Agent(reactor) request = agent.request(method, uri, headers, body_producer) return (reactor, request)
Issue a request.
def _fetch(self, method, url=None, post_data=None, parse_data=True, key=None, parameters=None, listener=None, full_return=False): """ Issue a request. Args: method (str): Request method (GET/POST/PUT/DELETE/etc.) If not specified, it will be POST if post_data is not None Kwargs: url (str): Destination URL post_data (str): A string of what to POST parse_data (bool): If true, parse response data key (string): If parse_data==True, look for this key when parsing data parameters (dict): Additional GET parameters to append to the URL listener (func): callback called when uploading a file full_return (bool): If set to True, get a full response (with success, data, info, body) Returns: dict. Response. If full_return==True, a dict with keys: success, data, info, body, otherwise the parsed data Raises: AuthenticationError, ConnectionError, urllib2.HTTPError, ValueError """ headers = self.get_headers() headers["Content-Type"] = "application/json" handlers = [] debuglevel = int(self._settings["debug"]) handlers.append(urllib2.HTTPHandler(debuglevel=debuglevel)) if hasattr(httplib, "HTTPS"): handlers.append(urllib2.HTTPSHandler(debuglevel=debuglevel)) handlers.append(urllib2.HTTPCookieProcessor(cookielib.CookieJar())) password_url = self._get_password_url() if password_url and "Authorization" not in headers: pwd_manager = urllib2.HTTPPasswordMgrWithDefaultRealm() pwd_manager.add_password(None, password_url, self._settings["user"], self._settings["password"]) handlers.append(HTTPBasicAuthHandler(pwd_manager)) opener = urllib2.build_opener(*handlers) if post_data is not None: post_data = json.dumps(post_data) uri = self._url(url, parameters) request = RESTRequest(uri, method=method, headers=headers) if post_data is not None: request.add_data(post_data) response = None try: response = opener.open(request) body = response.read() if password_url and password_url not in self._settings["authorizations"] and request.has_header("Authorization"): self._settings["authorizations"][password_url] = request.get_header("Authorization") except urllib2.HTTPError as e: if e.code == 401: raise AuthenticationError("Access denied while trying to access %s" % uri) elif e.code == 404: raise ConnectionError("URL not found: %s" % uri) else: raise except urllib2.URLError as e: raise ConnectionError("Error while fetching from %s: %s" % (uri, e)) finally: if response: response.close() opener.close() data = None if parse_data: if not key: key = string.split(url, "/")[0] data = self.parse(body, key) if full_return: info = response.info() if response else None status = int(string.split(info["status"])[0]) if (info and "status" in info) else None return { "success": (status >= 200 and status < 300), "data": data, "info": info, "body": body } return data
Build destination URL.
def _url(self, url=None, parameters=None): """ Build destination URL. Kwargs: url (str): Destination URL parameters (dict): Additional GET parameters to append to the URL Returns: str. URL """ uri = url or self._settings["url"] if url and self._settings["base_url"]: uri = "%s/%s" % (self._settings["base_url"], url) uri += ".json" if parameters: uri += "?%s" % urllib.urlencode(parameters) return uri
Tells if this message is a text message.
def is_text(self): """ Tells if this message is a text message. Returns: bool. Success """ return self.type in [ self._TYPE_PASTE, self._TYPE_TEXT, self._TYPE_TWEET ]
Get rooms list.
def get_rooms(self, sort=True): """ Get rooms list. Kwargs: sort (bool): If True, sort rooms by name Returns: array. List of rooms (each room is a dict) """ rooms = self._connection.get("rooms") if sort: rooms.sort(key=operator.itemgetter("name")) return rooms
Get a room by name.
def get_room_by_name(self, name): """ Get a room by name. Returns: :class:`Room`. Room Raises: RoomNotFoundException """ rooms = self.get_rooms() for room in rooms or []: if room["name"] == name: return self.get_room(room["id"]) raise RoomNotFoundException("Room %s not found" % name)
Get room.
def get_room(self, id): """ Get room. Returns: :class:`Room`. Room """ if id not in self._rooms: self._rooms[id] = Room(self, id) return self._rooms[id]
Get user.
def get_user(self, id = None): """ Get user. Returns: :class:`User`. User """ if not id: id = self._user.id if id not in self._users: self._users[id] = self._user if id == self._user.id else User(self, id) return self._users[id]
Search transcripts.
def search(self, terms): """ Search transcripts. Args: terms (str): Terms for search Returns: array. Messages """ messages = self._connection.get("search/%s" % urllib.quote_plus(terms), key="messages") if messages: messages = [Message(self, message) for message in messages] return messages
: rtype: Cookie. SimpleCookie
def cookie_dump(key, value='', max_age=None, expires=None, path='/', domain=None, secure=False, httponly=False): """ :rtype: ``Cookie.SimpleCookie`` """ cookie = SimpleCookie() cookie[key] = value for attr in ('max_age', 'expires', 'path', 'domain', 'secure', 'httponly'): attr_key = attr.replace('_', '-') attr_value = locals()[attr] if attr_value: cookie[key][attr_key] = attr_value return cookie
e. g. 200 OK
def response_status_string(code): """e.g. ``200 OK`` """ mean = HTTP_STATUS_CODES.get(code, 'unknown').upper() return '{code} {mean}'.format(code=code, mean=mean)
Request cookies
def cookies(self): """Request cookies :rtype: dict """ http_cookie = self.environ.get('HTTP_COOKIE', '') _cookies = { k: v.value for (k, v) in SimpleCookie(http_cookie).items() } return _cookies
Attach an observer.
def attach(self, observer): """ Attach an observer. Args: observer (func): A function to be called when new messages arrive Returns: :class:`Stream`. Current instance to allow chaining """ if not observer in self._observers: self._observers.append(observer) return self
Called when incoming messages arrive.
def incoming(self, messages): """ Called when incoming messages arrive. Args: messages (tuple): Messages (each message is a dict) """ if self._observers: campfire = self._room.get_campfire() for message in messages: for observer in self._observers: observer(Message(campfire, message))
Called by the thread it runs the process.
def run(self): """ Called by the thread, it runs the process. NEVER call this method directly. Instead call start() to start the thread. To stop, call stop(), and then join() """ if self._live: self._use_process = True self._abort = False campfire = self._room.get_campfire() if self._live: process = LiveStreamProcess(campfire.get_connection().get_settings(), self._room.id) else: process = StreamProcess(campfire.get_connection().get_settings(), self._room.id, pause=self._pause) if not self._use_process: process.set_callback(self.incoming) if self._use_process: queue = Queue() process.set_queue(queue) process.start() if not process.is_alive(): return self._streaming = True while not self._abort: if self._use_process: if not process.is_alive(): self._abort = True break try: incoming = queue.get_nowait() if isinstance(incoming, list): self.incoming(incoming) elif isinstance(incoming, Exception): self._abort = True if self._error_callback: self._error_callback(incoming, self._room) except Empty: time.sleep(self._pause) pass else: process.fetch() time.sleep(self._pause) self._streaming = False if self._use_process and self._abort and not process.is_alive() and self._error_callback: self._error_callback(Exception("Streaming process was killed"), self._room) if self._use_process: queue.close() if process.is_alive(): process.stop() process.terminate() process.join()
Called by the process it runs it.
def run(self): """ Called by the process, it runs it. NEVER call this method directly. Instead call start() to start the separate process. If you don't want to use a second process, then call fetch() directly on this istance. To stop, call terminate() """ if not self._queue: raise Exception("No queue available to send messages") while True: self.fetch() time.sleep(self._pause)
Fetch new messages.
def fetch(self): """ Fetch new messages. """ try: if not self._last_message_id: messages = self._connection.get("room/%s/recent" % self._room_id, key="messages", parameters={ "limit": 1 }) self._last_message_id = messages[-1]["id"] messages = self._connection.get("room/%s/recent" % self._room_id, key="messages", parameters={ "since_message_id": self._last_message_id }) except: messages = [] if messages: self._last_message_id = messages[-1]["id"] self.received(messages)
Called when new messages arrive.
def received(self, messages): """ Called when new messages arrive. Args: messages (tuple): Messages """ if messages: if self._queue: self._queue.put_nowait(messages) if self._callback: self._callback(messages)
Called by the process it runs it.
def run(self): """ Called by the process, it runs it. NEVER call this method directly. Instead call start() to start the separate process. If you don't want to use a second process, then call fetch() directly on this istance. To stop, call terminate() """ if not self._queue: raise Exception("No queue available to send messages") factory = LiveStreamFactory(self) self._reactor.connectSSL("streaming.campfirenow.com", 443, factory, ssl.ClientContextFactory()) self._reactor.run()