file_name
large_stringlengths
4
140
prefix
large_stringlengths
0
12.1k
suffix
large_stringlengths
0
12k
middle
large_stringlengths
0
7.51k
fim_type
large_stringclasses
4 values
utils.py
if(isinstance(objs, dict)): stringToReturn += "\n" for key in objs: stringToReturn += currentDepth*" "+key+": "+get_result_as_tree(objs[key], depth, currentDepth+1, key) return stringToReturn if (isinstance(objs, tuple)): return str(objs[0]) + "\n" if(objs is None): return stringToReturn mydict = objs.__dict__ stringToReturn += "\n" for key in mydict: stringToReturn += currentDepth*" " stringToReturn += key+": "+get_result_as_tree(mydict[key], depth, currentDepth+1, key) return stringToReturn """ def filter_objects_from_simple_keypaths(objs, simpleKeyPaths): # First, we assemble the key paths. # They start out like this: # [accouts.username, accounts.initiator_secret.secret, accounts.status] # and become like this: # {"accounts":{"username":True, "initiator_secret":{"secret":True}, "status":True} keyPaths = dict() for simpleKeyPath in simpleKeyPaths: currentLevel = keyPaths keyPathArray = simpleKeyPath.split('.') for i in range(len(keyPathArray)): if(i<(len(keyPathArray) - 1)): if currentLevel.get(keyPathArray[i]) is None: currentLevel[keyPathArray[i]] = dict() else: currentLevel[keyPathArray[i]] = True currentLevel = currentLevel[keyPathArray[i]] # Then we pass it in to filter objects. return filter_objects(objs, keyPaths) # Keypaths is arranged as follows: # it is a nested dict with the order of the keys. def filter_objects(objs, keyPaths): # Otherwise, we keep recursing deeper. # Because there are deeper keys, we know that we can go deeper. # This means we are dealing with either an array or a dict. # If keyPaths looks like this: # {"username": True, "volumes": {"Id": True}} # The keys in this sequence will be username and volumes. # When we recurse into volumes, the keys will be Id. finalFilteredObjects = dict() if keyPaths == True and type(objs) is not list: return objs # If we've found a list, we recurse deeper to pull out the objs. # We do not advance our keyPath recursion because this is just a list. if type(objs) is list: # If we have a list of objects, we will need to assemble and return a list of stuff. filteredObjsDict = [None]*len(objs) for i in range(len(objs)): # Each element could be a string, dict, or list. filteredObjsDict[i] = filter_objects(objs[i], keyPaths) return filteredObjsDict dictionaryOfInterest = None if type(objs) is dict: dictionaryOfInterest = objs else: dictionaryOfInterest = objs.__dict__ for key in keyPaths: # If we've found a dict, we recurse deeper to pull out the objs. # Because this is a dict, we must advance our keyPaths recursion. # Consider the following example: if key not in dictionaryOfInterest: raise ValueError("'"+key+"' is not a valid key for this level. Valid keys are: "+','.join(dictionaryOfInterest.keys())) finalFilteredObjects[key] = filter_objects(dictionaryOfInterest[key], keyPaths[key]) return finalFilteredObjects def print_result_as_table(objs, keyPaths): filteredDictionary = filter_objects(objs, keyPaths) def print_result_as_tree(objs, depth=1): print(get_result_as_tree(objs, depth)) def establish_connection(ctx): # Verify that the mvip does not contain the port number: if ctx.mvip and ":" in ctx.mvip: ctx.logger.error('Please provide the port using the port parameter.') exit(1) cfg = None # Arguments take precedence regardless of env settings if ctx.mvip: if ctx.username is None: ctx.username = getpass.getpass("Username:") if ctx.password is None: ctx.password = getpass.getpass("Password:") cfg = {'mvip': ctx.mvip, 'username': "b'"+encrypt(ctx.username).decode('utf-8')+"'", 'password': "b'"+encrypt(ctx.password).decode('utf-8')+"'", 'port': ctx.port, 'url': 'https://%s:%s' % (ctx.mvip, ctx.port), 'version': ctx.version, 'verifyssl': ctx.verifyssl, 'timeout': ctx.timeout} try: ctx.element = ElementFactory.create(cfg["mvip"],decrypt(cfg["username"]),decrypt(cfg["password"]),port=cfg["port"],version=cfg["version"],verify_ssl=cfg["verifyssl"],timeout=cfg["timeout"]) ctx.version = ctx.element._api_version cfg["version"] = ctx.element._api_version except Exception as e: ctx.logger.error(e.__str__()) exit(1) # If someone accidentally passed in an argument, but didn't specify everything, throw an error. elif ctx.username or ctx.password: ctx.logger.error("In order to manually connect, please provide an mvip, a username, AND a password") # If someone asked for a given connection or we need to default to using the connection at index 0 if it exists: else: if ctx.connectionindex is None and ctx.name is None: cfg = get_default_connection(ctx) elif ctx.connectionindex is not None: connections = get_connections(ctx) if int(ctx.connectionindex) > (len(connections)-1) or int(ctx.connectionindex) < (-len(connections)): ctx.logger.error("Connection "+str(ctx.connectionindex)+" Please provide an index between "+str(-len(connections))+" and "+str(len(connections)-1)) exit(1) cfg = connections[ctx.connectionindex] elif ctx.name is not None: connections = get_connections(ctx) filteredCfg = [connection for connection in connections if connection["name"] == ctx.name] if(len(filteredCfg) > 1): ctx.logger.error("Your connections.csv file has become corrupted. There are two connections of the same name.") exit() if(len(filteredCfg) < 1): ctx.logger.error("Could not find a connection named "+ctx.name) exit() cfg = filteredCfg[0] # If we managed to find the connection we were looking for, we must try to establish the connection. if cfg is not None: # Finally, we need to establish our connection via elementfactory: try: if int(cfg["port"]) != 443: address = cfg["mvip"] + ":" + cfg["port"] else: address = cfg["mvip"] ctx.element = ElementFactory.create(address, decrypt(cfg["username"]), decrypt(cfg["password"]), cfg["version"], verify_ssl=cfg["verifyssl"]) if int(cfg["timeout"]) != 30: ctx.element.timeout(cfg["timeout"]) except Exception as e: ctx.logger.error(e.__str__()) ctx.logger.error("The connection is corrupt. Run 'sfcli connection prune' to try and remove all broken connections or use 'sfcli connection remove -n name'") ctx.logger.error(cfg) exit(1) # If we want the json output directly from the source, we'll have to override the send request method in the sdk: # This is so that we can circumvent the python objects and get exactly what the json-rpc returns. if ctx.json and ctx.element: def new_send_request(*args, **kwargs): return ctx.element.__class__.__bases__[0].send_request(ctx.element, return_response_raw=True, *args, **kwargs) ctx.element.send_request = new_send_request # The only time it is none is when we're asking for help or we're trying to store a connection. # If that's not what we're doing, we catch it later. if cfg is not None: cfg["port"] = int(cfg["port"]) ctx.cfg = cfg cfg["name"] = cfg.get("name", "default") if not ctx.nocache: write_default_connection(ctx, cfg) if ctx.element is None: ctx.logger.error("You must establish at least one connection and specify which you intend to use.") exit() # this needs to be atomic. def get_connections(ctx): connectionsCsvLocation = resource_filename(Requirement.parse("solidfire-cli"), "connections.csv") connectionsLock = resource_filename(Requirement.parse("solidfire-cli"), "connectionsLock") if os.path.exists(connectionsCsvLocation): try: with FileLock(connectionsLock): with open(connectionsCsvLocation, 'r') as connectionFile: connections = list(csv.DictReader(connectionFile, delimiter=',')) except Exception as e: ctx.logger.error("Problem reading "+connectionsCsvLocation+" because: "+str(e.args)+" Try changing the permissions of that file.") exit(1) else: connections = [] for connection in connections: connection["version"] = float(connection["version"]) if connection.get("verifyssl") == "True": connection["verifyssl"] = True else:
connection["verifyssl"] = False
conditional_block
utils.py
=False): #print(jsonpickle.encode(objs)) nestedDict = serializer.loads(jsonpickle.encode(objs)) filteredDict = type(nestedDict)() if(pickle==False): remove_pickling(nestedDict, filteredDict) else: filteredDict = nestedDict print(serializer.dumps(filteredDict,indent=4)) def remove_pickling(nestedDict, filteredDict): if type(nestedDict) is dict: #foreach key, if list, recurse, if dict, recurse, if string recurse unless py/obj is key. for key in nestedDict: if key == "py/object": continue else: filteredDict[key] = type(nestedDict[key])() filteredDict[key] = remove_pickling(nestedDict[key], filteredDict[key]) return filteredDict if type(nestedDict) is list: # foreach item for i in range(len(nestedDict)): filteredDict.append(type(nestedDict[i])()) filteredDict[i] = remove_pickling(nestedDict[i], filteredDict[i]) return filteredDict return nestedDict """ SDK1.6 Note: Commenting this as print_tree is not supported in SDK 1.6. """ def get_result_as_tree(objs, depth=1, currentDepth=0, lastKey = ""): print("print_tree is not supported in SDK1.6") """stringToReturn = "" if(currentDepth > depth): return "<to see more details, increase depth>\n" if(type(objs) is str or type(objs) is bool or type(objs) is int or type(objs) is type(u'') or objs is None or type(objs) is float):# or (sys.version_info[0]<3 and type(objs) is long)): return str(objs) + "\n" if(type(objs) is list): stringToReturn += "\n" for i in range(len(objs)): obj = objs[i] stringToReturn += currentDepth*" "+get_result_as_tree(obj, depth, currentDepth+1, lastKey) return stringToReturn if(isinstance(objs, dict)): stringToReturn += "\n" for key in objs: stringToReturn += currentDepth*" "+key+": "+get_result_as_tree(objs[key], depth, currentDepth+1, key) return stringToReturn if (isinstance(objs, tuple)): return str(objs[0]) + "\n" if(objs is None): return stringToReturn mydict = objs.__dict__ stringToReturn += "\n" for key in mydict: stringToReturn += currentDepth*" " stringToReturn += key+": "+get_result_as_tree(mydict[key], depth, currentDepth+1, key) return stringToReturn
def filter_objects_from_simple_keypaths(objs, simpleKeyPaths): # First, we assemble the key paths. # They start out like this: # [accouts.username, accounts.initiator_secret.secret, accounts.status] # and become like this: # {"accounts":{"username":True, "initiator_secret":{"secret":True}, "status":True} keyPaths = dict() for simpleKeyPath in simpleKeyPaths: currentLevel = keyPaths keyPathArray = simpleKeyPath.split('.') for i in range(len(keyPathArray)): if(i<(len(keyPathArray) - 1)): if currentLevel.get(keyPathArray[i]) is None: currentLevel[keyPathArray[i]] = dict() else: currentLevel[keyPathArray[i]] = True currentLevel = currentLevel[keyPathArray[i]] # Then we pass it in to filter objects. return filter_objects(objs, keyPaths) # Keypaths is arranged as follows: # it is a nested dict with the order of the keys. def filter_objects(objs, keyPaths): # Otherwise, we keep recursing deeper. # Because there are deeper keys, we know that we can go deeper. # This means we are dealing with either an array or a dict. # If keyPaths looks like this: # {"username": True, "volumes": {"Id": True}} # The keys in this sequence will be username and volumes. # When we recurse into volumes, the keys will be Id. finalFilteredObjects = dict() if keyPaths == True and type(objs) is not list: return objs # If we've found a list, we recurse deeper to pull out the objs. # We do not advance our keyPath recursion because this is just a list. if type(objs) is list: # If we have a list of objects, we will need to assemble and return a list of stuff. filteredObjsDict = [None]*len(objs) for i in range(len(objs)): # Each element could be a string, dict, or list. filteredObjsDict[i] = filter_objects(objs[i], keyPaths) return filteredObjsDict dictionaryOfInterest = None if type(objs) is dict: dictionaryOfInterest = objs else: dictionaryOfInterest = objs.__dict__ for key in keyPaths: # If we've found a dict, we recurse deeper to pull out the objs. # Because this is a dict, we must advance our keyPaths recursion. # Consider the following example: if key not in dictionaryOfInterest: raise ValueError("'"+key+"' is not a valid key for this level. Valid keys are: "+','.join(dictionaryOfInterest.keys())) finalFilteredObjects[key] = filter_objects(dictionaryOfInterest[key], keyPaths[key]) return finalFilteredObjects def print_result_as_table(objs, keyPaths): filteredDictionary = filter_objects(objs, keyPaths) def print_result_as_tree(objs, depth=1): print(get_result_as_tree(objs, depth)) def establish_connection(ctx): # Verify that the mvip does not contain the port number: if ctx.mvip and ":" in ctx.mvip: ctx.logger.error('Please provide the port using the port parameter.') exit(1) cfg = None # Arguments take precedence regardless of env settings if ctx.mvip: if ctx.username is None: ctx.username = getpass.getpass("Username:") if ctx.password is None: ctx.password = getpass.getpass("Password:") cfg = {'mvip': ctx.mvip, 'username': "b'"+encrypt(ctx.username).decode('utf-8')+"'", 'password': "b'"+encrypt(ctx.password).decode('utf-8')+"'", 'port': ctx.port, 'url': 'https://%s:%s' % (ctx.mvip, ctx.port), 'version': ctx.version, 'verifyssl': ctx.verifyssl, 'timeout': ctx.timeout} try: ctx.element = ElementFactory.create(cfg["mvip"],decrypt(cfg["username"]),decrypt(cfg["password"]),port=cfg["port"],version=cfg["version"],verify_ssl=cfg["verifyssl"],timeout=cfg["timeout"]) ctx.version = ctx.element._api_version cfg["version"] = ctx.element._api_version except Exception as e: ctx.logger.error(e.__str__()) exit(1) # If someone accidentally passed in an argument, but didn't specify everything, throw an error. elif ctx.username or ctx.password: ctx.logger.error("In order to manually connect, please provide an mvip, a username, AND a password") # If someone asked for a given connection or we need to default to using the connection at index 0 if it exists: else: if ctx.connectionindex is None and ctx.name is None: cfg = get_default_connection(ctx) elif ctx.connectionindex is not None: connections = get_connections(ctx) if int(ctx.connectionindex) > (len(connections)-1) or int(ctx.connectionindex) < (-len(connections)): ctx.logger.error("Connection "+str(ctx.connectionindex)+" Please provide an index between "+str(-len(connections))+" and "+str(len(connections)-1)) exit(1) cfg = connections[ctx.connectionindex] elif ctx.name is not None: connections = get_connections(ctx) filteredCfg = [connection for connection in connections if connection["name"] == ctx.name] if(len(filteredCfg) > 1): ctx.logger.error("Your connections.csv file has become corrupted. There are two connections of the same name.") exit() if(len(filteredCfg) < 1): ctx.logger.error("Could not find a connection named "+ctx.name) exit() cfg = filteredCfg[0] # If we managed to find the connection we were looking for, we must try to establish the connection. if cfg is not None: # Finally, we need to establish our connection via elementfactory: try: if int(cfg["port"]) != 443: address = cfg["mvip"] + ":" + cfg["port"] else: address = cfg["mvip"] ctx.element = ElementFactory.create(address, decrypt(cfg["username"]), decrypt(cfg["password"]), cfg["version"], verify_ssl=cfg["verifyssl"]) if int(cfg["timeout"]) != 30: ctx.element.timeout(cfg["timeout"]) except Exception as e: ctx.logger.error(e.__str__()) ctx.logger
"""
random_line_split
assets.py
(self, sid_group): """Retrieve the most recent symbol for a set of sids. Parameters ---------- sid_group : iterable[int] The sids to lookup. The length of this sequence must be less than or equal to SQLITE_MAX_VARIABLE_NUMBER because the sids will be passed in as sql bind params. Returns ------- sel : Selectable The sqlalchemy selectable that will query for the most recent symbol for each sid. Notes ----- This is implemented as an inner select of the columns of interest ordered by the end date of the (sid, symbol) mapping. We then group that inner select on the sid with no aggregations to select the last row per group which gives us the most recently active symbol for all of the sids. """ cols = self.equity_symbol_mappings.c # These are the columns we actually want. data_cols = (cols.sid,) + tuple(cols[name] for name in symbol_columns) # Also select the max of end_date so that all non-grouped fields take # on the value associated with the max end_date. The SQLite docs say # this: # # When the min() or max() aggregate functions are used in an aggregate # query, all bare columns in the result set take values from the input # row which also contains the minimum or maximum. Only the built-in # min() and max() functions work this way. # # See https://www.sqlite.org/lang_select.html#resultset, for more info. to_select = data_cols + (sa.func.max(cols.end_date),) return sa.select( to_select, ).where( cols.sid.in_(map(int, sid_group)) ).group_by( cols.sid, ) def _lookup_most_recent_symbols(self, sids): return { row.sid: {c: row[c] for c in symbol_columns} for row in concat( self.engine.execute( self._select_most_recent_symbols_chunk(sid_group), ).fetchall() for sid_group in partition_all( SQLITE_MAX_VARIABLE_NUMBER, sids ) ) } def _retrieve_asset_dicts(self, sids, asset_tbl, querying_equities): if not sids: return if querying_equities: def mkdict(row, exchanges=self.exchange_info, symbols=self._lookup_most_recent_symbols(sids)): d = dict(row) d['exchange_info'] = exchanges[d.pop('exchange')] # we are not required to have a symbol for every asset, if # we don't have any symbols we will just use the empty string return merge(d, symbols.get(row['sid'], {})) else: def mkdict(row, exchanges=self.exchange_info): d = dict(row) d['exchange_info'] = exchanges[d.pop('exchange')] return d for assets in group_into_chunks(sids): # Load misses from the db. query = self._select_assets_by_sid(asset_tbl, assets) for row in query.execute().fetchall(): yield _convert_asset_timestamp_fields(mkdict(row)) def _retrieve_assets(self, sids, asset_tbl, asset_type): """ Internal function for loading assets from a table. This should be the only method of `AssetFinder` that writes Assets into self._asset_cache. Parameters --------- sids : iterable of int Asset ids to look up. asset_tbl : sqlalchemy.Table Table from which to query assets. asset_type : type Type of asset to be constructed. Returns ------- assets : dict[int -> Asset] Dict mapping requested sids to the retrieved assets. """ # Fastpath for empty request. if not sids: return {} cache = self._asset_cache hits = {} querying_equities = issubclass(asset_type, Equity) filter_kwargs = ( _filter_equity_kwargs if querying_equities else _filter_future_kwargs ) rows = self._retrieve_asset_dicts(sids, asset_tbl, querying_equities) for row in rows: sid = row['sid'] asset = asset_type(**filter_kwargs(row)) hits[sid] = cache[sid] = asset # If we get here, it means something in our code thought that a # particular sid was an equity/future and called this function with a # concrete type, but we couldn't actually resolve the asset. This is # an error in our code, not a user-input error. misses = tuple(set(sids) - viewkeys(hits)) if misses: if querying_equities: raise EquitiesNotFound(sids=misses) else: raise FutureContractsNotFound(sids=misses) return hits def lookup_symbol(self, symbol): """Lookup an equity by symbol. This method can only resolve the equity if exactly one equity has ever owned the ticker. Parameters ---------- symbol : str The ticker symbol to resolve. Returns ------- equity : Equity The equity identified by ``symbol``. Raises ------ SymbolNotFound Raised when no equity has ever held the given symbol. MultipleSymbolsFound Raised when more than one equity has held ``symbol`` or when the symbol is ambiguous across multiple countries. Notes ----- The resolution algorithm is as follows: - Split the symbol into the company and share class component. - Do a dictionary lookup of the ``(company_symbol, share_class_symbol)`` in the provided ownership map. - If there is no entry in the dictionary, we don't know about this symbol so raise a ``SymbolNotFound`` error. - If more there is more than one owner, raise ``MultipleSymbolsFound`` - Otherwise, because the list mapped to a symbol cannot be empty, return the single asset. """ if symbol is None: raise TypeError("Cannot lookup asset for symbol of None") mapping = self.symbol_ownership_map # split the symbol into the components, if there are no # company/share class parts then share_class_symbol will be empty company_symbol, share_class_symbol = split_delimited_symbol(symbol) try: owners = mapping[company_symbol, share_class_symbol] assert owners, 'empty owners list for %r' % symbol except KeyError: # no equity has ever held this symbol raise SymbolNotFound(symbol=symbol) # exactly one equity has ever held this symbol if len(owners) == 1: return self.retrieve_asset(owners[0].sid) options = {self.retrieve_asset(owner.sid) for owner in owners} # more than one equity has held this ticker, this # is ambiguous raise MultipleSymbolsFound(symbol=symbol, options=options) def lookup_future_symbol(self, symbol): """Lookup a future contract by symbol. Parameters ---------- symbol : str The symbol of the desired contract. Returns ------- future : Future The future contract referenced by ``symbol``. Raises ------ SymbolNotFound Raised when no contract named 'symbol' is found. """ data = self._select_asset_by_symbol(self.futures_contracts, symbol)\ .execute().fetchone() # If no data found, raise an exception if not data: raise SymbolNotFound(symbol=symbol) return self.retrieve_asset(data['sid']) def _get_contract_sids(self, root_symbol): fc_cols = self.futures_contracts.c return [r.sid for r in list(sa.select((fc_cols.sid,)).where( (fc_cols.root_symbol == root_symbol) & (pd.notnull(fc_cols.start_date))).order_by( fc_cols.auto_close_date).execute().fetchall())] def _get_root_symbol_exchange(self, root_symbol): fc_cols = self.futures_root_symbols.c fields = (fc_cols.exchange,) exchange = sa.select(fields).where( fc_cols.root_symbol == root_symbol).execute().scalar() if exchange is not None: return exchange else: raise SymbolNotFound(symbol=root_symbol) def get_ordered_contracts(self, root_symbol): try: return self._ordered_contracts[root_symbol] except KeyError: contract_sids = self._get_contract_sids(root_symbol) contracts = deque(self.retrieve_all(contract_sids)) chain_predicate = self._future_chain_predicates.get(root_symbol, None) oc = OrderedContracts(root_symbol, contracts, chain_predicate) self._ordered_contracts[root_symbol] = oc return oc def create_continuous_future(self, root_symbol, offset, roll_style, adjustment):
if adjustment not in ADJUSTMENT_STYLES: raise ValueError( 'Invalid adjustment style {!r}. Allowed adjustment styles are ' '{}.'.format(adjustment, list(ADJUSTMENT_STYLES)) ) oc = self.get_ordered_contracts(root_symbol) exchange = self._get_root_symbol_exchange(root_symbol) sid = _encode_continuous_future_sid(root_symbol, offset, roll_style, None) mul_sid = _encode_continuous_future_sid(root_symbol, offset, roll_style, 'div') add_sid = _encode_continuous_future_sid(root_symbol, offset, roll_style, 'add') cf_template = partial(
identifier_body
assets.py
', 'sid start end') class AssetFinder(object): """ An AssetFinder is an interface to a database of Asset metadata written by an ``AssetDBWriter``. This class provides methods for looking up assets by unique integer id or by symbol. For historical reasons, we refer to these unique ids as 'sids'. Parameters ---------- engine : str or SQLAlchemy.engine An engine with a connection to the asset database to use, or a string that can be parsed by SQLAlchemy as a URI. future_chain_predicates : dict A dict mapping future root symbol to a predicate function which accepts a contract as a parameter and returns whether or not the contract should be included in the chain. See Also -------- :class:`zipline.assets.AssetDBWriter` """ @preprocess(engine=coerce_string_to_eng(require_exists=True)) def __init__(self, engine, future_chain_predicates=None): self.engine = engine metadata = sa.MetaData(bind=engine) metadata.reflect(only=asset_db_table_names) for table_name in asset_db_table_names: setattr(self, table_name, metadata.tables[table_name]) # Check the version info of the db for compatibility check_version_info(engine, self.version_info, ASSET_DB_VERSION) # Cache for lookup of assets by sid, the objects in the asset lookup # may be shared with the results from equity and future lookup caches. # # The top level cache exists to minimize lookups on the asset type # routing. # # The caches are read through, i.e. accessing an asset through # retrieve_asset will populate the cache on first retrieval. self._asset_cache = {} self._asset_type_cache = {} self._caches = (self._asset_cache, self._asset_type_cache) self._future_chain_predicates = future_chain_predicates \ if future_chain_predicates is not None else {} self._ordered_contracts = {} # Populated on first call to `lifetimes`. self._asset_lifetimes = {} # Stores the max end_date of the bundle self._bundle_end_date = None self._sids_to_real_sids = {} self._real_sids_to_sids = {} @lazyval def
(self): es = sa.select(self.exchanges.c).execute().fetchall() return { name: ExchangeInfo(name, canonical_name, country_code) for name, canonical_name, country_code in es } @lazyval def symbol_ownership_map(self): return build_ownership_map( table=self.equity_symbol_mappings, key_from_row=( lambda row: (row.company_symbol, row.share_class_symbol) ), value_from_row=lambda row: row.symbol, ) @lazyval def country_codes(self): return tuple([c for (c,) in sa.select( sa.distinct(self.exchanges.c.country_code, )).execute().fetchall()]) def lookup_asset_types(self, sids): """ Retrieve asset types for a list of sids. Parameters ---------- sids : list[int] Returns ------- types : dict[sid -> str or None] Asset types for the provided sids. """ found = {} missing = set() for sid in sids: try: found[sid] = self._asset_type_cache[sid] except KeyError: missing.add(sid) if not missing: return found router_cols = self.asset_router.c for assets in group_into_chunks(missing): query = sa.select((router_cols.sid, router_cols.asset_type)).where( self.asset_router.c.sid.in_(map(int, assets)) ) for sid, type_ in query.execute().fetchall(): missing.remove(sid) found[sid] = self._asset_type_cache[sid] = type_ for sid in missing: found[sid] = self._asset_type_cache[sid] = None return found def group_by_type(self, sids): """ Group a list of sids by asset type. Parameters ---------- sids : list[int] Returns ------- types : dict[str or None -> list[int]] A dict mapping unique asset types to lists of sids drawn from sids. If we fail to look up an asset, we assign it a key of None. """ return invert(self.lookup_asset_types(sids)) def retrieve_asset(self, sid, default_none=False): """ Retrieve the Asset for a given sid. """ try: asset = self._asset_cache[sid] if asset is None and not default_none: raise SidsNotFound(sids=[sid]) return asset except KeyError: return self.retrieve_all((sid,), default_none=default_none)[0] def retrieve_all(self, sids, default_none=False): """ Retrieve all assets in `sids`. Parameters ---------- sids : iterable of int Assets to retrieve. default_none : bool If True, return None for failed lookups. If False, raise `SidsNotFound`. Returns ------- assets : list[Asset or None] A list of the same length as `sids` containing Assets (or Nones) corresponding to the requested sids. Raises ------ SidsNotFound When a requested sid is not found and default_none=False. """ sids = list(sids) hits, missing, failures = {}, set(), [] for sid in sids: try: asset = self._asset_cache[sid] if not default_none and asset is None: # Bail early if we've already cached that we don't know # about an asset. raise SidsNotFound(sids=[sid]) hits[sid] = asset except KeyError: missing.add(sid) # All requests were cache hits. Return requested sids in order. if not missing: return [hits[sid] for sid in sids] update_hits = hits.update # Look up cache misses by type. type_to_assets = self.group_by_type(missing) # Handle failures failures = {failure: None for failure in type_to_assets.pop(None, ())} update_hits(failures) self._asset_cache.update(failures) if failures and not default_none: raise SidsNotFound(sids=list(failures)) # We don't update the asset cache here because it should already be # updated by `self.retrieve_equities`. update_hits(self.retrieve_equities(type_to_assets.pop('equity', ()))) update_hits( self.retrieve_futures_contracts(type_to_assets.pop('future', ())) ) # We shouldn't know about any other asset types. if type_to_assets: raise AssertionError( "Found asset types: %s" % list(type_to_assets.keys()) ) return [hits[sid] for sid in sids] def retrieve_equities(self, sids): """ Retrieve Equity objects for a list of sids. Users generally shouldn't need to this method (instead, they should prefer the more general/friendly `retrieve_assets`), but it has a documented interface and tests because it's used upstream. Parameters ---------- sids : iterable[int] Returns ------- equities : dict[int -> Equity] Raises ------ EquitiesNotFound When any requested asset isn't found. """ return self._retrieve_assets(sids, self.equities, Equity) def _retrieve_equity(self, sid): return self.retrieve_equities((sid,))[sid] def retrieve_futures_contracts(self, sids): """ Retrieve Future objects for an iterable of sids. Users generally shouldn't need to this method (instead, they should prefer the more general/friendly `retrieve_assets`), but it has a documented interface and tests because it's used upstream. Parameters ---------- sids : iterable[int] Returns ------- equities : dict[int -> Equity] Raises ------ EquitiesNotFound When any requested asset isn't found. """ return self._retrieve_assets(sids, self.futures_contracts, Future) @staticmethod def _select_assets_by_sid(asset_tbl, sids): return sa.select([asset_tbl]).where( asset_tbl.c.sid.in_(map(int, sids)) ) @staticmethod def _select_asset_by_symbol(asset_tbl, symbol): return sa.select([asset_tbl]).where(asset_tbl.c.symbol == symbol) def _select_most_recent_symbols_chunk(self, sid_group): """Retrieve the most recent symbol for a set of sids. Parameters ---------- sid_group : iterable[int] The sids to lookup. The length of this sequence must be less than or equal to SQLITE_MAX_VARIABLE_NUMBER because the sids will be passed in as sql bind params. Returns ------- sel : Selectable The sqlalchemy selectable that will query for the most recent symbol for each sid. Notes ----- This is implemented as an inner select of the columns of interest ordered by the end date of the (sid, symbol) mapping. We then group that inner select on the sid with no
exchange_info
identifier_name
assets.py
_by_type(self, sids): """ Group a list of sids by asset type. Parameters ---------- sids : list[int] Returns ------- types : dict[str or None -> list[int]] A dict mapping unique asset types to lists of sids drawn from sids. If we fail to look up an asset, we assign it a key of None. """ return invert(self.lookup_asset_types(sids)) def retrieve_asset(self, sid, default_none=False): """ Retrieve the Asset for a given sid. """ try: asset = self._asset_cache[sid] if asset is None and not default_none: raise SidsNotFound(sids=[sid]) return asset except KeyError: return self.retrieve_all((sid,), default_none=default_none)[0] def retrieve_all(self, sids, default_none=False): """ Retrieve all assets in `sids`. Parameters ---------- sids : iterable of int Assets to retrieve. default_none : bool If True, return None for failed lookups. If False, raise `SidsNotFound`. Returns ------- assets : list[Asset or None] A list of the same length as `sids` containing Assets (or Nones) corresponding to the requested sids. Raises ------ SidsNotFound When a requested sid is not found and default_none=False. """ sids = list(sids) hits, missing, failures = {}, set(), [] for sid in sids: try: asset = self._asset_cache[sid] if not default_none and asset is None: # Bail early if we've already cached that we don't know # about an asset. raise SidsNotFound(sids=[sid]) hits[sid] = asset except KeyError: missing.add(sid) # All requests were cache hits. Return requested sids in order. if not missing: return [hits[sid] for sid in sids] update_hits = hits.update # Look up cache misses by type. type_to_assets = self.group_by_type(missing) # Handle failures failures = {failure: None for failure in type_to_assets.pop(None, ())} update_hits(failures) self._asset_cache.update(failures) if failures and not default_none: raise SidsNotFound(sids=list(failures)) # We don't update the asset cache here because it should already be # updated by `self.retrieve_equities`. update_hits(self.retrieve_equities(type_to_assets.pop('equity', ()))) update_hits( self.retrieve_futures_contracts(type_to_assets.pop('future', ())) ) # We shouldn't know about any other asset types. if type_to_assets: raise AssertionError( "Found asset types: %s" % list(type_to_assets.keys()) ) return [hits[sid] for sid in sids] def retrieve_equities(self, sids): """ Retrieve Equity objects for a list of sids. Users generally shouldn't need to this method (instead, they should prefer the more general/friendly `retrieve_assets`), but it has a documented interface and tests because it's used upstream. Parameters ---------- sids : iterable[int] Returns ------- equities : dict[int -> Equity] Raises ------ EquitiesNotFound When any requested asset isn't found. """ return self._retrieve_assets(sids, self.equities, Equity) def _retrieve_equity(self, sid): return self.retrieve_equities((sid,))[sid] def retrieve_futures_contracts(self, sids): """ Retrieve Future objects for an iterable of sids. Users generally shouldn't need to this method (instead, they should prefer the more general/friendly `retrieve_assets`), but it has a documented interface and tests because it's used upstream. Parameters ---------- sids : iterable[int] Returns ------- equities : dict[int -> Equity] Raises ------ EquitiesNotFound When any requested asset isn't found. """ return self._retrieve_assets(sids, self.futures_contracts, Future) @staticmethod def _select_assets_by_sid(asset_tbl, sids): return sa.select([asset_tbl]).where( asset_tbl.c.sid.in_(map(int, sids)) ) @staticmethod def _select_asset_by_symbol(asset_tbl, symbol): return sa.select([asset_tbl]).where(asset_tbl.c.symbol == symbol) def _select_most_recent_symbols_chunk(self, sid_group): """Retrieve the most recent symbol for a set of sids. Parameters ---------- sid_group : iterable[int] The sids to lookup. The length of this sequence must be less than or equal to SQLITE_MAX_VARIABLE_NUMBER because the sids will be passed in as sql bind params. Returns ------- sel : Selectable The sqlalchemy selectable that will query for the most recent symbol for each sid. Notes ----- This is implemented as an inner select of the columns of interest ordered by the end date of the (sid, symbol) mapping. We then group that inner select on the sid with no aggregations to select the last row per group which gives us the most recently active symbol for all of the sids. """ cols = self.equity_symbol_mappings.c # These are the columns we actually want. data_cols = (cols.sid,) + tuple(cols[name] for name in symbol_columns) # Also select the max of end_date so that all non-grouped fields take # on the value associated with the max end_date. The SQLite docs say # this: # # When the min() or max() aggregate functions are used in an aggregate # query, all bare columns in the result set take values from the input # row which also contains the minimum or maximum. Only the built-in # min() and max() functions work this way. # # See https://www.sqlite.org/lang_select.html#resultset, for more info. to_select = data_cols + (sa.func.max(cols.end_date),) return sa.select( to_select, ).where( cols.sid.in_(map(int, sid_group)) ).group_by( cols.sid, ) def _lookup_most_recent_symbols(self, sids): return { row.sid: {c: row[c] for c in symbol_columns} for row in concat( self.engine.execute( self._select_most_recent_symbols_chunk(sid_group), ).fetchall() for sid_group in partition_all( SQLITE_MAX_VARIABLE_NUMBER, sids ) ) } def _retrieve_asset_dicts(self, sids, asset_tbl, querying_equities): if not sids: return if querying_equities: def mkdict(row, exchanges=self.exchange_info, symbols=self._lookup_most_recent_symbols(sids)): d = dict(row) d['exchange_info'] = exchanges[d.pop('exchange')] # we are not required to have a symbol for every asset, if # we don't have any symbols we will just use the empty string return merge(d, symbols.get(row['sid'], {})) else: def mkdict(row, exchanges=self.exchange_info): d = dict(row) d['exchange_info'] = exchanges[d.pop('exchange')] return d for assets in group_into_chunks(sids): # Load misses from the db. query = self._select_assets_by_sid(asset_tbl, assets) for row in query.execute().fetchall(): yield _convert_asset_timestamp_fields(mkdict(row)) def _retrieve_assets(self, sids, asset_tbl, asset_type): """ Internal function for loading assets from a table. This should be the only method of `AssetFinder` that writes Assets into self._asset_cache. Parameters --------- sids : iterable of int Asset ids to look up. asset_tbl : sqlalchemy.Table Table from which to query assets. asset_type : type Type of asset to be constructed. Returns ------- assets : dict[int -> Asset] Dict mapping requested sids to the retrieved assets. """ # Fastpath for empty request. if not sids: return {} cache = self._asset_cache hits = {} querying_equities = issubclass(asset_type, Equity) filter_kwargs = ( _filter_equity_kwargs if querying_equities else _filter_future_kwargs ) rows = self._retrieve_asset_dicts(sids, asset_tbl, querying_equities) for row in rows: sid = row['sid'] asset = asset_type(**filter_kwargs(row)) hits[sid] = cache[sid] = asset # If we get here, it means something in our code thought that a # particular sid was an equity/future and called this function with a # concrete type, but we couldn't actually resolve the asset. This is # an error in our code, not a user-input error. misses = tuple(set(sids) - viewkeys(hits)) if misses: if querying_equities:
raise EquitiesNotFound(sids=misses)
conditional_block
assets.py
etimes', 'sid start end') class AssetFinder(object): """ An AssetFinder is an interface to a database of Asset metadata written by an ``AssetDBWriter``. This class provides methods for looking up assets by unique integer id or by symbol. For historical reasons, we refer to these unique ids as 'sids'. Parameters ---------- engine : str or SQLAlchemy.engine An engine with a connection to the asset database to use, or a string that can be parsed by SQLAlchemy as a URI. future_chain_predicates : dict A dict mapping future root symbol to a predicate function which accepts a contract as a parameter and returns whether or not the contract should be included in the chain. See Also -------- :class:`zipline.assets.AssetDBWriter` """ @preprocess(engine=coerce_string_to_eng(require_exists=True)) def __init__(self, engine, future_chain_predicates=None): self.engine = engine metadata = sa.MetaData(bind=engine) metadata.reflect(only=asset_db_table_names) for table_name in asset_db_table_names: setattr(self, table_name, metadata.tables[table_name]) # Check the version info of the db for compatibility check_version_info(engine, self.version_info, ASSET_DB_VERSION) # Cache for lookup of assets by sid, the objects in the asset lookup # may be shared with the results from equity and future lookup caches. # # The top level cache exists to minimize lookups on the asset type # routing. # # The caches are read through, i.e. accessing an asset through # retrieve_asset will populate the cache on first retrieval. self._asset_cache = {} self._asset_type_cache = {} self._caches = (self._asset_cache, self._asset_type_cache) self._future_chain_predicates = future_chain_predicates \ if future_chain_predicates is not None else {} self._ordered_contracts = {} # Populated on first call to `lifetimes`. self._asset_lifetimes = {} # Stores the max end_date of the bundle self._bundle_end_date = None self._sids_to_real_sids = {} self._real_sids_to_sids = {} @lazyval def exchange_info(self): es = sa.select(self.exchanges.c).execute().fetchall() return { name: ExchangeInfo(name, canonical_name, country_code) for name, canonical_name, country_code in es } @lazyval def symbol_ownership_map(self): return build_ownership_map( table=self.equity_symbol_mappings, key_from_row=( lambda row: (row.company_symbol, row.share_class_symbol) ), value_from_row=lambda row: row.symbol, ) @lazyval def country_codes(self): return tuple([c for (c,) in sa.select( sa.distinct(self.exchanges.c.country_code, )).execute().fetchall()]) def lookup_asset_types(self, sids): """ Retrieve asset types for a list of sids. Parameters ---------- sids : list[int] Returns ------- types : dict[sid -> str or None] Asset types for the provided sids. """ found = {} missing = set() for sid in sids: try: found[sid] = self._asset_type_cache[sid] except KeyError: missing.add(sid) if not missing: return found router_cols = self.asset_router.c for assets in group_into_chunks(missing): query = sa.select((router_cols.sid, router_cols.asset_type)).where( self.asset_router.c.sid.in_(map(int, assets)) ) for sid, type_ in query.execute().fetchall(): missing.remove(sid) found[sid] = self._asset_type_cache[sid] = type_ for sid in missing: found[sid] = self._asset_type_cache[sid] = None return found def group_by_type(self, sids): """ Group a list of sids by asset type. Parameters ---------- sids : list[int] Returns ------- types : dict[str or None -> list[int]] A dict mapping unique asset types to lists of sids drawn from sids. If we fail to look up an asset, we assign it a key of None. """ return invert(self.lookup_asset_types(sids)) def retrieve_asset(self, sid, default_none=False): """ Retrieve the Asset for a given sid. """ try: asset = self._asset_cache[sid] if asset is None and not default_none: raise SidsNotFound(sids=[sid]) return asset except KeyError: return self.retrieve_all((sid,), default_none=default_none)[0] def retrieve_all(self, sids, default_none=False): """ Retrieve all assets in `sids`. Parameters ---------- sids : iterable of int Assets to retrieve. default_none : bool If True, return None for failed lookups. If False, raise `SidsNotFound`. Returns ------- assets : list[Asset or None] A list of the same length as `sids` containing Assets (or Nones) corresponding to the requested sids. Raises ------ SidsNotFound When a requested sid is not found and default_none=False. """ sids = list(sids) hits, missing, failures = {}, set(), [] for sid in sids: try: asset = self._asset_cache[sid] if not default_none and asset is None: # Bail early if we've already cached that we don't know
missing.add(sid) # All requests were cache hits. Return requested sids in order. if not missing: return [hits[sid] for sid in sids] update_hits = hits.update # Look up cache misses by type. type_to_assets = self.group_by_type(missing) # Handle failures failures = {failure: None for failure in type_to_assets.pop(None, ())} update_hits(failures) self._asset_cache.update(failures) if failures and not default_none: raise SidsNotFound(sids=list(failures)) # We don't update the asset cache here because it should already be # updated by `self.retrieve_equities`. update_hits(self.retrieve_equities(type_to_assets.pop('equity', ()))) update_hits( self.retrieve_futures_contracts(type_to_assets.pop('future', ())) ) # We shouldn't know about any other asset types. if type_to_assets: raise AssertionError( "Found asset types: %s" % list(type_to_assets.keys()) ) return [hits[sid] for sid in sids] def retrieve_equities(self, sids): """ Retrieve Equity objects for a list of sids. Users generally shouldn't need to this method (instead, they should prefer the more general/friendly `retrieve_assets`), but it has a documented interface and tests because it's used upstream. Parameters ---------- sids : iterable[int] Returns ------- equities : dict[int -> Equity] Raises ------ EquitiesNotFound When any requested asset isn't found. """ return self._retrieve_assets(sids, self.equities, Equity) def _retrieve_equity(self, sid): return self.retrieve_equities((sid,))[sid] def retrieve_futures_contracts(self, sids): """ Retrieve Future objects for an iterable of sids. Users generally shouldn't need to this method (instead, they should prefer the more general/friendly `retrieve_assets`), but it has a documented interface and tests because it's used upstream. Parameters ---------- sids : iterable[int] Returns ------- equities : dict[int -> Equity] Raises ------ EquitiesNotFound When any requested asset isn't found. """ return self._retrieve_assets(sids, self.futures_contracts, Future) @staticmethod def _select_assets_by_sid(asset_tbl, sids): return sa.select([asset_tbl]).where( asset_tbl.c.sid.in_(map(int, sids)) ) @staticmethod def _select_asset_by_symbol(asset_tbl, symbol): return sa.select([asset_tbl]).where(asset_tbl.c.symbol == symbol) def _select_most_recent_symbols_chunk(self, sid_group): """Retrieve the most recent symbol for a set of sids. Parameters ---------- sid_group : iterable[int] The sids to lookup. The length of this sequence must be less than or equal to SQLITE_MAX_VARIABLE_NUMBER because the sids will be passed in as sql bind params. Returns ------- sel : Selectable The sqlalchemy selectable that will query for the most recent symbol for each sid. Notes ----- This is implemented as an inner select of the columns of interest ordered by the end date of the (sid, symbol) mapping. We then group that inner select on the sid with no aggreg
# about an asset. raise SidsNotFound(sids=[sid]) hits[sid] = asset except KeyError:
random_line_split
game.rs
f, "Callback<{}>", std::any::type_name::<T>()) } } impl Game { /// Create a new game alongside a handle to thet game. pub fn new() -> (Game, GameHandle) { let (sender, receiver) = mpsc::channel(1024); let world = logic::create_world(logic::WorldKind::WithObjects); let schedule = logic::add_systems(Default::default(), logic::SystemSet::Everything); let executor = logic::Executor::new(schedule); let game = Game { players: BTreeMap::new(), receiver, world, executor, snapshots: SnapshotEncoder::new(), time: 0, }; let handle = GameHandle { sender }; (game, handle) } /// Run the game to completion (either the handle is dropped or a fatal error occurs). pub async fn run(&mut self) { let mut timer = time::interval(time::Duration::from_secs(1) / TICK_RATE); loop { tokio::select! { _ = timer.tick() => { self.tick(); } command = self.receiver.recv() => match command { None => { log::info!("game handle dropped"); break; }, Some(command) => { log::debug!("got command: {:?}", command); self.execute_command(command); } } }; } } fn tick(&mut self) { self.executor.tick(&mut self.world); self.snapshots.update_mapping(&self.world); self.check_win_condition(); let mut events = Vec::<EventKind>::new(); let snapshot = Arc::new(self.snapshot()); events.push(snapshot.into()); for event in events { self.broadcast(event); } self.time = self.time.wrapping_add(1); } fn broadcast<T>(&mut self, kind: T) where T: Into<EventKind>, { let event = Event { time: self.time, kind: kind.into(), }; let mut dead = Vec::new(); for (&id, player) in &mut self.players { match player.events.try_send(event.clone()) { Ok(()) => {} Err(TrySendError::Full(_)) => { log::warn!("player {}'s event buffer is full", id); dead.push(id); // TODO: request full client resync } Err(TrySendError::Closed(_)) => { log::info!("player {} stopped listening for events", id); dead.push(id); // TODO: stop attempting to send events to this player, and potentially // disconnect them. } } } for player in dead { self.remove_player(player); } } fn remove_player(&mut self, player: PlayerId) -> Option<PlayerData> { let data = self.players.remove(&player)?; self.world.delete(data.entity); self.world .resources .get_mut::<DeadEntities>() .unwrap() .entities .push(data.network_id); Some(data) } /// Check if any player has won or lost. fn check_win_condition(&mut self) { let dead = self.world.resources.get::<DeadEntities>().unwrap(); let mut losers = Vec::new(); for (&player, data) in &self.players { if dead.entities.contains(&data.network_id) { losers.push(player); } } drop(dead); for loser in losers { let mut player = self.players.remove(&loser).unwrap(); let event = Event { time: self.time, kind: EventKind::GameOver(GameOver::Loser), }; tokio::spawn(async move { player.events.send(event).await }); if self.players.len() == 1 { let winner = *self.players.keys().next().unwrap(); let mut player = self.remove_player(winner).unwrap(); let event = Event { time: self.time, kind: EventKind::GameOver(GameOver::Winner), }; tokio::spawn(async move { player.events.send(event).await }); } } } /// Execute a command. fn execute_command(&mut self, command: Command) { match command { Command::RegisterPlayer { callback } => { callback.send(self.register_player()); } Command::DisconnectPlayer(player) => { self.remove_player(player); } Command::Request { callback, request } => { let message = self.handle_request(request); callback.send(message); } Command::Snapshot { callback } => { let snapshot = self.snapshot(); callback.send(snapshot); } Command::PerformAction { action, player } => self.perform_action(action, player), } } /// Create and register a new player fn register_player(&mut self) -> PlayerHandle { let player = self.next_player_id(); let entity = logic::add_player(&mut self.world, player); let (sender, receiver) = mpsc::channel(EVENT_BUFFER_SIZE); let network_id = *self.world.get_component::<EntityId>(entity).unwrap(); let data = PlayerData { network_id, entity, events: sender, }; self.players.insert(player, data); PlayerHandle { player, events: receiver, } } /// Find the next available player id fn next_player_id(&self) -> PlayerId { let mut id = 1; for player in self.players.keys() { if player.0 == id { id += 1; } else { break; } } PlayerId(id) } /// Perform the request and return the result in a message fn handle_request(&mut self, request: Request) -> Response { let kind = match request.kind { RequestKind::Ping => protocol::Pong.into(), RequestKind::Init => { let error = "Requested 'Init' on already initialized player"; ResponseKind::Error(error.into()) } }; Response { channel: request.channel, kind, } } /// Get a snapshot of the current game state. fn snapshot(&self) -> Snapshot { self.snapshots.make_snapshot(&self.world) } /// Perform an action for a player. fn perform_action(&mut self, action: Action, player: PlayerId) { match action.kind { ActionKind::Move(new) => { || -> Option<()> { let data = self.players.get(&player)?; let mut movement = self.world.get_component_mut::<Movement>(data.entity)?; movement.direction = new.direction; Some(()) }(); } ActionKind::Break(breaking) => { || -> Option<()> { let data = self.players.get(&player)?; let breaking = breaking .entity .and_then(|breaking| self.snapshots.lookup(breaking)); self.world .get_component_mut::<WorldInteraction>(data.entity)? .breaking = breaking; Some(()) }(); } ActionKind::Throw(throwing) => { if let Some(data) = self.players.get(&player) { logic::events::throw(&mut self.world, data.entity, throwing.target); } } } } } impl GameHandle { /// Register a new client and return it's id. pub async fn register_player(&mut self) -> crate::Result<PlayerHandle> { self.send_with(|callback| Command::RegisterPlayer { callback }) .await } /// Remove a player from the game. pub async fn disconnect_player(&mut self, player: PlayerId) -> crate::Result<()> { self.sender.send(Command::DisconnectPlayer(player)).await?; Ok(()) } /// Handle a request made by a player. pub async fn handle_request(&mut self, request: Request) -> crate::Result<Response> { self.send_with(move |callback| Command::Request { request, callback }) .await } /// Get a snapshot of the current game state. pub async fn snapshot(&mut self) -> crate::Result<Snapshot> { self.send_with(|callback| Command::Snapshot { callback }) .await } /// Handle an action performed by a player pub async fn handle_action(&mut self, action: Action, player: PlayerId) -> crate::Result<()> { self.sender .send(Command::PerformAction { action, player }) .await?; Ok(()) } /// Send a command to the game with the specified callback and then return the value passed into /// the callback. async fn send_with<F, O>(&mut self, to_command: F) -> crate::Result<O> where F: FnOnce(Callback<O>) -> Command, { let (callback, value) = Callback::new(); let command = to_command(callback); self.sender.send(command).await?; value.await.map_err(Into::into) } } impl PlayerHandle { /// Get the id of this player pub fn id(&self) -> PlayerId { self.player } pub async fn poll_event(&mut self) -> Option<Event> { self.events.recv().await } } impl<T> Callback<T> { /// Create a new callback pub fn new() -> (Callback<T>, oneshot::Receiver<T>)
{ let (sender, receiver) = oneshot::channel(); (Callback { sender }, receiver) }
identifier_body
game.rs
32 = 60; /// The maximum number of events to buffer per player. const EVENT_BUFFER_SIZE: usize = 1024; pub struct Game { players: BTreeMap<PlayerId, PlayerData>, receiver: mpsc::Receiver<Command>, world: World, executor: logic::Executor, snapshots: SnapshotEncoder, time: u32, } #[derive(Debug, Clone)] struct PlayerData { entity: Entity, network_id: EntityId, events: mpsc::Sender<Event>, } #[derive(Debug)] pub struct PlayerHandle { player: PlayerId, events: mpsc::Receiver<Event>, } #[derive(Debug, Clone)] pub struct GameHandle { sender: mpsc::Sender<Command>, } #[derive(Debug)] enum Command { Request { request: Request, callback: Callback<Response>, }, RegisterPlayer { callback: Callback<PlayerHandle>, }, DisconnectPlayer(PlayerId), Snapshot { callback: Callback<Snapshot>, }, PerformAction { action: Action, player: PlayerId, }, } struct Callback<T> { sender: oneshot::Sender<T>, } // We don't care what the callback contains, simply print the expected return type. impl<T> Debug for Callback<T> { fn fmt(&self, f: &mut Formatter) -> fmt::Result { write!(f, "Callback<{}>", std::any::type_name::<T>()) } } impl Game { /// Create a new game alongside a handle to thet game. pub fn new() -> (Game, GameHandle) { let (sender, receiver) = mpsc::channel(1024); let world = logic::create_world(logic::WorldKind::WithObjects); let schedule = logic::add_systems(Default::default(), logic::SystemSet::Everything); let executor = logic::Executor::new(schedule); let game = Game { players: BTreeMap::new(), receiver, world, executor, snapshots: SnapshotEncoder::new(), time: 0, }; let handle = GameHandle { sender }; (game, handle) } /// Run the game to completion (either the handle is dropped or a fatal error occurs). pub async fn run(&mut self) { let mut timer = time::interval(time::Duration::from_secs(1) / TICK_RATE); loop { tokio::select! { _ = timer.tick() => { self.tick(); } command = self.receiver.recv() => match command { None => { log::info!("game handle dropped"); break; }, Some(command) => { log::debug!("got command: {:?}", command); self.execute_command(command); } } }; } } fn tick(&mut self) { self.executor.tick(&mut self.world); self.snapshots.update_mapping(&self.world); self.check_win_condition(); let mut events = Vec::<EventKind>::new(); let snapshot = Arc::new(self.snapshot()); events.push(snapshot.into()); for event in events { self.broadcast(event); } self.time = self.time.wrapping_add(1); } fn broadcast<T>(&mut self, kind: T) where T: Into<EventKind>, { let event = Event { time: self.time, kind: kind.into(), }; let mut dead = Vec::new(); for (&id, player) in &mut self.players { match player.events.try_send(event.clone()) { Ok(()) => {} Err(TrySendError::Full(_)) => { log::warn!("player {}'s event buffer is full", id); dead.push(id); // TODO: request full client resync } Err(TrySendError::Closed(_)) => { log::info!("player {} stopped listening for events", id); dead.push(id); // TODO: stop attempting to send events to this player, and potentially // disconnect them. } } } for player in dead { self.remove_player(player); } } fn remove_player(&mut self, player: PlayerId) -> Option<PlayerData> { let data = self.players.remove(&player)?; self.world.delete(data.entity); self.world .resources .get_mut::<DeadEntities>() .unwrap() .entities .push(data.network_id); Some(data) } /// Check if any player has won or lost. fn check_win_condition(&mut self) { let dead = self.world.resources.get::<DeadEntities>().unwrap(); let mut losers = Vec::new(); for (&player, data) in &self.players { if dead.entities.contains(&data.network_id) { losers.push(player); } } drop(dead); for loser in losers { let mut player = self.players.remove(&loser).unwrap(); let event = Event { time: self.time, kind: EventKind::GameOver(GameOver::Loser), }; tokio::spawn(async move { player.events.send(event).await }); if self.players.len() == 1 { let winner = *self.players.keys().next().unwrap(); let mut player = self.remove_player(winner).unwrap(); let event = Event { time: self.time, kind: EventKind::GameOver(GameOver::Winner), }; tokio::spawn(async move { player.events.send(event).await }); } } } /// Execute a command. fn execute_command(&mut self, command: Command) { match command { Command::RegisterPlayer { callback } => { callback.send(self.register_player()); } Command::DisconnectPlayer(player) => { self.remove_player(player); } Command::Request { callback, request } => { let message = self.handle_request(request); callback.send(message); } Command::Snapshot { callback } => { let snapshot = self.snapshot(); callback.send(snapshot); } Command::PerformAction { action, player } => self.perform_action(action, player), } } /// Create and register a new player fn register_player(&mut self) -> PlayerHandle { let player = self.next_player_id(); let entity = logic::add_player(&mut self.world, player); let (sender, receiver) = mpsc::channel(EVENT_BUFFER_SIZE); let network_id = *self.world.get_component::<EntityId>(entity).unwrap(); let data = PlayerData { network_id, entity, events: sender, }; self.players.insert(player, data); PlayerHandle { player, events: receiver, } } /// Find the next available player id fn next_player_id(&self) -> PlayerId { let mut id = 1; for player in self.players.keys() { if player.0 == id { id += 1; } else { break; } } PlayerId(id) } /// Perform the request and return the result in a message fn handle_request(&mut self, request: Request) -> Response { let kind = match request.kind { RequestKind::Ping => protocol::Pong.into(), RequestKind::Init => { let error = "Requested 'Init' on already initialized player"; ResponseKind::Error(error.into()) } }; Response { channel: request.channel, kind, } } /// Get a snapshot of the current game state. fn snapshot(&self) -> Snapshot { self.snapshots.make_snapshot(&self.world) } /// Perform an action for a player. fn perform_action(&mut self, action: Action, player: PlayerId) { match action.kind { ActionKind::Move(new) => { || -> Option<()> { let data = self.players.get(&player)?; let mut movement = self.world.get_component_mut::<Movement>(data.entity)?; movement.direction = new.direction; Some(()) }(); } ActionKind::Break(breaking) => { || -> Option<()> { let data = self.players.get(&player)?; let breaking = breaking .entity .and_then(|breaking| self.snapshots.lookup(breaking)); self.world .get_component_mut::<WorldInteraction>(data.entity)? .breaking = breaking; Some(()) }(); } ActionKind::Throw(throwing) => { if let Some(data) = self.players.get(&player) { logic::events::throw(&mut self.world, data.entity, throwing.target); } } } } } impl GameHandle { /// Register a new client and return it's id. pub async fn register_player(&mut self) -> crate::Result<PlayerHandle> { self.send_with(|callback| Command::RegisterPlayer { callback }) .await } /// Remove a player from the game. pub async fn disconnect_player(&mut self, player: PlayerId) -> crate::Result<()> { self.sender.send(Command::DisconnectPlayer(player)).await?; Ok(()) } /// Handle a request made by a player. pub async fn handle_request(&mut self, request: Request) -> crate::Result<Response> { self.send_with(move |callback| Command::Request { request, callback }) .await } /// Get a snapshot of the current game state. pub async fn
(&mut self) -> crate::Result<Snapshot>
snapshot
identifier_name
game.rs
32 = 60; /// The maximum number of events to buffer per player. const EVENT_BUFFER_SIZE: usize = 1024; pub struct Game { players: BTreeMap<PlayerId, PlayerData>, receiver: mpsc::Receiver<Command>, world: World, executor: logic::Executor, snapshots: SnapshotEncoder, time: u32, } #[derive(Debug, Clone)] struct PlayerData { entity: Entity, network_id: EntityId, events: mpsc::Sender<Event>, } #[derive(Debug)] pub struct PlayerHandle { player: PlayerId, events: mpsc::Receiver<Event>, } #[derive(Debug, Clone)]
#[derive(Debug)] enum Command { Request { request: Request, callback: Callback<Response>, }, RegisterPlayer { callback: Callback<PlayerHandle>, }, DisconnectPlayer(PlayerId), Snapshot { callback: Callback<Snapshot>, }, PerformAction { action: Action, player: PlayerId, }, } struct Callback<T> { sender: oneshot::Sender<T>, } // We don't care what the callback contains, simply print the expected return type. impl<T> Debug for Callback<T> { fn fmt(&self, f: &mut Formatter) -> fmt::Result { write!(f, "Callback<{}>", std::any::type_name::<T>()) } } impl Game { /// Create a new game alongside a handle to thet game. pub fn new() -> (Game, GameHandle) { let (sender, receiver) = mpsc::channel(1024); let world = logic::create_world(logic::WorldKind::WithObjects); let schedule = logic::add_systems(Default::default(), logic::SystemSet::Everything); let executor = logic::Executor::new(schedule); let game = Game { players: BTreeMap::new(), receiver, world, executor, snapshots: SnapshotEncoder::new(), time: 0, }; let handle = GameHandle { sender }; (game, handle) } /// Run the game to completion (either the handle is dropped or a fatal error occurs). pub async fn run(&mut self) { let mut timer = time::interval(time::Duration::from_secs(1) / TICK_RATE); loop { tokio::select! { _ = timer.tick() => { self.tick(); } command = self.receiver.recv() => match command { None => { log::info!("game handle dropped"); break; }, Some(command) => { log::debug!("got command: {:?}", command); self.execute_command(command); } } }; } } fn tick(&mut self) { self.executor.tick(&mut self.world); self.snapshots.update_mapping(&self.world); self.check_win_condition(); let mut events = Vec::<EventKind>::new(); let snapshot = Arc::new(self.snapshot()); events.push(snapshot.into()); for event in events { self.broadcast(event); } self.time = self.time.wrapping_add(1); } fn broadcast<T>(&mut self, kind: T) where T: Into<EventKind>, { let event = Event { time: self.time, kind: kind.into(), }; let mut dead = Vec::new(); for (&id, player) in &mut self.players { match player.events.try_send(event.clone()) { Ok(()) => {} Err(TrySendError::Full(_)) => { log::warn!("player {}'s event buffer is full", id); dead.push(id); // TODO: request full client resync } Err(TrySendError::Closed(_)) => { log::info!("player {} stopped listening for events", id); dead.push(id); // TODO: stop attempting to send events to this player, and potentially // disconnect them. } } } for player in dead { self.remove_player(player); } } fn remove_player(&mut self, player: PlayerId) -> Option<PlayerData> { let data = self.players.remove(&player)?; self.world.delete(data.entity); self.world .resources .get_mut::<DeadEntities>() .unwrap() .entities .push(data.network_id); Some(data) } /// Check if any player has won or lost. fn check_win_condition(&mut self) { let dead = self.world.resources.get::<DeadEntities>().unwrap(); let mut losers = Vec::new(); for (&player, data) in &self.players { if dead.entities.contains(&data.network_id) { losers.push(player); } } drop(dead); for loser in losers { let mut player = self.players.remove(&loser).unwrap(); let event = Event { time: self.time, kind: EventKind::GameOver(GameOver::Loser), }; tokio::spawn(async move { player.events.send(event).await }); if self.players.len() == 1 { let winner = *self.players.keys().next().unwrap(); let mut player = self.remove_player(winner).unwrap(); let event = Event { time: self.time, kind: EventKind::GameOver(GameOver::Winner), }; tokio::spawn(async move { player.events.send(event).await }); } } } /// Execute a command. fn execute_command(&mut self, command: Command) { match command { Command::RegisterPlayer { callback } => { callback.send(self.register_player()); } Command::DisconnectPlayer(player) => { self.remove_player(player); } Command::Request { callback, request } => { let message = self.handle_request(request); callback.send(message); } Command::Snapshot { callback } => { let snapshot = self.snapshot(); callback.send(snapshot); } Command::PerformAction { action, player } => self.perform_action(action, player), } } /// Create and register a new player fn register_player(&mut self) -> PlayerHandle { let player = self.next_player_id(); let entity = logic::add_player(&mut self.world, player); let (sender, receiver) = mpsc::channel(EVENT_BUFFER_SIZE); let network_id = *self.world.get_component::<EntityId>(entity).unwrap(); let data = PlayerData { network_id, entity, events: sender, }; self.players.insert(player, data); PlayerHandle { player, events: receiver, } } /// Find the next available player id fn next_player_id(&self) -> PlayerId { let mut id = 1; for player in self.players.keys() { if player.0 == id { id += 1; } else { break; } } PlayerId(id) } /// Perform the request and return the result in a message fn handle_request(&mut self, request: Request) -> Response { let kind = match request.kind { RequestKind::Ping => protocol::Pong.into(), RequestKind::Init => { let error = "Requested 'Init' on already initialized player"; ResponseKind::Error(error.into()) } }; Response { channel: request.channel, kind, } } /// Get a snapshot of the current game state. fn snapshot(&self) -> Snapshot { self.snapshots.make_snapshot(&self.world) } /// Perform an action for a player. fn perform_action(&mut self, action: Action, player: PlayerId) { match action.kind { ActionKind::Move(new) => { || -> Option<()> { let data = self.players.get(&player)?; let mut movement = self.world.get_component_mut::<Movement>(data.entity)?; movement.direction = new.direction; Some(()) }(); } ActionKind::Break(breaking) => { || -> Option<()> { let data = self.players.get(&player)?; let breaking = breaking .entity .and_then(|breaking| self.snapshots.lookup(breaking)); self.world .get_component_mut::<WorldInteraction>(data.entity)? .breaking = breaking; Some(()) }(); } ActionKind::Throw(throwing) => { if let Some(data) = self.players.get(&player) { logic::events::throw(&mut self.world, data.entity, throwing.target); } } } } } impl GameHandle { /// Register a new client and return it's id. pub async fn register_player(&mut self) -> crate::Result<PlayerHandle> { self.send_with(|callback| Command::RegisterPlayer { callback }) .await } /// Remove a player from the game. pub async fn disconnect_player(&mut self, player: PlayerId) -> crate::Result<()> { self.sender.send(Command::DisconnectPlayer(player)).await?; Ok(()) } /// Handle a request made by a player. pub async fn handle_request(&mut self, request: Request) -> crate::Result<Response> { self.send_with(move |callback| Command::Request { request, callback }) .await } /// Get a snapshot of the current game state. pub async fn snapshot(&mut self) -> crate::Result<Snapshot> {
pub struct GameHandle { sender: mpsc::Sender<Command>, }
random_line_split
tmsExport_generated.go
), metric.value) if err != nil { log.Entry().WithError(err).Error("Error persisting influx environment.") errCount++ } } if errCount > 0 { log.Entry().Error("failed to persist Influx environment") } } // TmsExportCommand This step allows you to export an MTA file (multi-target application archive) and multiple MTA extension descriptors into a TMS (SAP Cloud Transport Management service) landscape for further TMS-controlled distribution through a TMS-configured landscape. func TmsExportCommand() *cobra.Command { const STEP_NAME = "tmsExport" metadata := tmsExportMetadata() var stepConfig tmsExportOptions var startTime time.Time var influx tmsExportInflux var logCollector *log.CollectorHook var splunkClient *splunk.Splunk telemetryClient := &telemetry.Telemetry{} var createTmsExportCmd = &cobra.Command{ Use: STEP_NAME, Short: "This step allows you to export an MTA file (multi-target application archive) and multiple MTA extension descriptors into a TMS (SAP Cloud Transport Management service) landscape for further TMS-controlled distribution through a TMS-configured landscape.", Long: `This step allows you to export an MTA file (multi-target application archive) and multiple MTA extension descriptors into a TMS (SAP Cloud Transport Management service) landscape for further TMS-controlled distribution through a TMS-configured landscape. The MTA file is attached to a new transport request which is added to the import queues of the follow-on transport nodes of the specified export node. TMS lets you manage transports between SAP Business Technology Platform accounts in Neo and Cloud Foundry, such as from DEV to TEST and PROD accounts. For more information, see [official documentation of SAP Cloud Transport Management service](https://help.sap.com/viewer/p/TRANSPORT_MANAGEMENT_SERVICE) !!! note "Prerequisites" * You have subscribed to and set up TMS, as described in [Initial Setup](https://help.sap.com/viewer/7f7160ec0d8546c6b3eab72fb5ad6fd8/Cloud/en-US/66fd7283c62f48adb23c56fb48c84a60.html), which includes the configuration of your transport landscape. * A corresponding service key has been created, as described in [Set Up the Environment to Transport Content Archives directly in an Application](https://help.sap.com/viewer/7f7160ec0d8546c6b3eab72fb5ad6fd8/Cloud/en-US/8d9490792ed14f1bbf8a6ac08a6bca64.html). This service key (JSON) must be stored as a secret text within the Jenkins secure store or provided as value of tmsServiceKey parameter.`, PreRunE: func(cmd *cobra.Command, _ []string) error { startTime = time.Now() log.SetStepName(STEP_NAME) log.SetVerbose(GeneralConfig.Verbose) GeneralConfig.GitHubAccessTokens = ResolveAccessTokens(GeneralConfig.GitHubTokens) path, _ := os.Getwd() fatalHook := &log.FatalHook{CorrelationID: GeneralConfig.CorrelationID, Path: path} log.RegisterHook(fatalHook) err := PrepareConfig(cmd, &metadata, STEP_NAME, &stepConfig, config.OpenPiperFile) if err != nil { log.SetErrorCategory(log.ErrorConfiguration) return err } log.RegisterSecret(stepConfig.TmsServiceKey) if len(GeneralConfig.HookConfig.SentryConfig.Dsn) > 0 { sentryHook := log.NewSentryHook(GeneralConfig.HookConfig.SentryConfig.Dsn, GeneralConfig.CorrelationID) log.RegisterHook(&sentryHook) } if len(GeneralConfig.HookConfig.SplunkConfig.Dsn) > 0 { splunkClient = &splunk.Splunk{} logCollector = &log.CollectorHook{CorrelationID: GeneralConfig.CorrelationID} log.RegisterHook(logCollector) } if err = log.RegisterANSHookIfConfigured(GeneralConfig.CorrelationID); err != nil { log.Entry().WithError(err).Warn("failed to set up SAP Alert Notification Service log hook") } validation, err := validation.New(validation.WithJSONNamesForStructFields(), validation.WithPredefinedErrorMessages()) if err != nil { return err } if err = validation.ValidateStruct(stepConfig); err != nil { log.SetErrorCategory(log.ErrorConfiguration) return err } return nil }, Run: func(_ *cobra.Command, _ []string) { stepTelemetryData := telemetry.CustomData{} stepTelemetryData.ErrorCode = "1" handler := func() { influx.persist(GeneralConfig.EnvRootPath, "influx") config.RemoveVaultSecretFiles() stepTelemetryData.Duration = fmt.Sprintf("%v", time.Since(startTime).Milliseconds()) stepTelemetryData.ErrorCategory = log.GetErrorCategory().String() stepTelemetryData.PiperCommitHash = GitCommit telemetryClient.SetData(&stepTelemetryData) telemetryClient.Send() if len(GeneralConfig.HookConfig.SplunkConfig.Dsn) > 0 { splunkClient.Initialize(GeneralConfig.CorrelationID, GeneralConfig.HookConfig.SplunkConfig.Dsn, GeneralConfig.HookConfig.SplunkConfig.Token, GeneralConfig.HookConfig.SplunkConfig.Index, GeneralConfig.HookConfig.SplunkConfig.SendLogs) splunkClient.Send(telemetryClient.GetData(), logCollector) } if len(GeneralConfig.HookConfig.SplunkConfig.ProdCriblEndpoint) > 0
} log.DeferExitHandler(handler) defer handler() telemetryClient.Initialize(GeneralConfig.NoTelemetry, STEP_NAME) tmsExport(stepConfig, &stepTelemetryData, &influx) stepTelemetryData.ErrorCode = "0" log.Entry().Info("SUCCESS") }, } addTmsExportFlags(createTmsExportCmd, &stepConfig) return createTmsExportCmd } func addTmsExportFlags(cmd *cobra.Command, stepConfig *tmsExportOptions) { cmd.Flags().StringVar(&stepConfig.TmsServiceKey, "tmsServiceKey", os.Getenv("PIPER_tmsServiceKey"), "Service key JSON string to access the SAP Cloud Transport Management service instance APIs. If not specified and if pipeline is running on Jenkins, service key, stored under ID provided with credentialsId parameter, is used.") cmd.Flags().StringVar(&stepConfig.CustomDescription, "customDescription", os.Getenv("PIPER_customDescription"), "Can be used as the description of a transport request. Will overwrite the default, which is corresponding Git commit ID.") cmd.Flags().StringVar(&stepConfig.NamedUser, "namedUser", `Piper-Pipeline`, "Defines the named user to execute transport request with. The default value is 'Piper-Pipeline'. If pipeline is running on Jenkins, the name of the user, who started the job, is tried to be used at first.") cmd.Flags().StringVar(&stepConfig.NodeName, "nodeName", os.Getenv("PIPER_nodeName"), "Defines the name of the export node - starting node in TMS landscape. The transport request is added to the queues of the follow-on nodes of export node.") cmd.Flags().StringVar(&stepConfig.MtaPath, "mtaPath", os.Getenv("PIPER_mtaPath"), "Defines the relative path to *.mtar file for the export to the SAP Cloud Transport Management service. If not specified, it will use the *.mtar file created in mtaBuild.") cmd.Flags().StringVar(&stepConfig.MtaVersion, "mtaVersion", `*`, "Defines the version of the MTA for which the MTA extension descriptor will be used. You can use an asterisk (*) to accept any MTA version, or use a specific version compliant with SemVer 2.0, e.g. 1.0.0 (see semver.org). If the parameter is not configured, an asterisk is used.") cmd.Flags().StringVar(&stepConfig.Proxy, "proxy", os.Getenv("PIPER_proxy"), "Proxy URL which should be used for communication with the SAP Cloud Transport Management service backend.") cmd.MarkFlagRequired("tmsServiceKey") cmd.MarkFlagRequired("nodeName") } // retrieve step metadata func tmsExportMetadata() config.StepData { var theMetaData = config.StepData{ Metadata: config.StepMetadata{ Name: "tmsExport", Aliases: []config.Alias{}, Description: "This step allows you to export an MTA file (multi-target application archive) and multiple MTA extension descriptors into a TMS (S
{ splunkClient.Initialize(GeneralConfig.CorrelationID, GeneralConfig.HookConfig.SplunkConfig.ProdCriblEndpoint, GeneralConfig.HookConfig.SplunkConfig.ProdCriblToken, GeneralConfig.HookConfig.SplunkConfig.ProdCriblIndex, GeneralConfig.HookConfig.SplunkConfig.SendLogs) splunkClient.Send(telemetryClient.GetData(), logCollector) }
conditional_block
tmsExport_generated.go
(path, resourceName string) { measurementContent := []struct { measurement string valType string name string value interface{} }{ {valType: config.InfluxField, measurement: "step_data", name: "tms", value: i.step_data.fields.tms}, } errCount := 0 for _, metric := range measurementContent { err := piperenv.SetResourceParameter(path, resourceName, filepath.Join(metric.measurement, fmt.Sprintf("%vs", metric.valType), metric.name), metric.value) if err != nil { log.Entry().WithError(err).Error("Error persisting influx environment.") errCount++ } } if errCount > 0 { log.Entry().Error("failed to persist Influx environment") } } // TmsExportCommand This step allows you to export an MTA file (multi-target application archive) and multiple MTA extension descriptors into a TMS (SAP Cloud Transport Management service) landscape for further TMS-controlled distribution through a TMS-configured landscape. func TmsExportCommand() *cobra.Command { const STEP_NAME = "tmsExport" metadata := tmsExportMetadata() var stepConfig tmsExportOptions var startTime time.Time var influx tmsExportInflux var logCollector *log.CollectorHook var splunkClient *splunk.Splunk telemetryClient := &telemetry.Telemetry{} var createTmsExportCmd = &cobra.Command{ Use: STEP_NAME, Short: "This step allows you to export an MTA file (multi-target application archive) and multiple MTA extension descriptors into a TMS (SAP Cloud Transport Management service) landscape for further TMS-controlled distribution through a TMS-configured landscape.", Long: `This step allows you to export an MTA file (multi-target application archive) and multiple MTA extension descriptors into a TMS (SAP Cloud Transport Management service) landscape for further TMS-controlled distribution through a TMS-configured landscape. The MTA file is attached to a new transport request which is added to the import queues of the follow-on transport nodes of the specified export node. TMS lets you manage transports between SAP Business Technology Platform accounts in Neo and Cloud Foundry, such as from DEV to TEST and PROD accounts. For more information, see [official documentation of SAP Cloud Transport Management service](https://help.sap.com/viewer/p/TRANSPORT_MANAGEMENT_SERVICE) !!! note "Prerequisites" * You have subscribed to and set up TMS, as described in [Initial Setup](https://help.sap.com/viewer/7f7160ec0d8546c6b3eab72fb5ad6fd8/Cloud/en-US/66fd7283c62f48adb23c56fb48c84a60.html), which includes the configuration of your transport landscape. * A corresponding service key has been created, as described in [Set Up the Environment to Transport Content Archives directly in an Application](https://help.sap.com/viewer/7f7160ec0d8546c6b3eab72fb5ad6fd8/Cloud/en-US/8d9490792ed14f1bbf8a6ac08a6bca64.html). This service key (JSON) must be stored as a secret text within the Jenkins secure store or provided as value of tmsServiceKey parameter.`, PreRunE: func(cmd *cobra.Command, _ []string) error { startTime = time.Now() log.SetStepName(STEP_NAME) log.SetVerbose(GeneralConfig.Verbose) GeneralConfig.GitHubAccessTokens = ResolveAccessTokens(GeneralConfig.GitHubTokens) path, _ := os.Getwd() fatalHook := &log.FatalHook{CorrelationID: GeneralConfig.CorrelationID, Path: path} log.RegisterHook(fatalHook) err := PrepareConfig(cmd, &metadata, STEP_NAME, &stepConfig, config.OpenPiperFile) if err != nil { log.SetErrorCategory(log.ErrorConfiguration) return err } log.RegisterSecret(stepConfig.TmsServiceKey) if len(GeneralConfig.HookConfig.SentryConfig.Dsn) > 0 { sentryHook := log.NewSentryHook(GeneralConfig.HookConfig.SentryConfig.Dsn, GeneralConfig.CorrelationID) log.RegisterHook(&sentryHook) } if len(GeneralConfig.HookConfig.SplunkConfig.Dsn) > 0 { splunkClient = &splunk.Splunk{} logCollector = &log.CollectorHook{CorrelationID: GeneralConfig.CorrelationID} log.RegisterHook(logCollector) } if err = log.RegisterANSHookIfConfigured(GeneralConfig.CorrelationID); err != nil { log.Entry().WithError(err).Warn("failed to set up SAP Alert Notification Service log hook") } validation, err := validation.New(validation.WithJSONNamesForStructFields(), validation.WithPredefinedErrorMessages()) if err != nil { return err } if err = validation.ValidateStruct(stepConfig); err != nil { log.SetErrorCategory(log.ErrorConfiguration) return err } return nil }, Run: func(_ *cobra.Command, _ []string) { stepTelemetryData := telemetry.CustomData{} stepTelemetryData.ErrorCode = "1" handler := func() { influx.persist(GeneralConfig.EnvRootPath, "influx") config.RemoveVaultSecretFiles() stepTelemetryData.Duration = fmt.Sprintf("%v", time.Since(startTime).Milliseconds()) stepTelemetryData.ErrorCategory = log.GetErrorCategory().String() stepTelemetryData.PiperCommitHash = GitCommit telemetryClient.SetData(&stepTelemetryData) telemetryClient.Send() if len(GeneralConfig.HookConfig.SplunkConfig.Dsn) > 0 { splunkClient.Initialize(GeneralConfig.CorrelationID, GeneralConfig.HookConfig.SplunkConfig.Dsn, GeneralConfig.HookConfig.SplunkConfig.Token, GeneralConfig.HookConfig.SplunkConfig.Index, GeneralConfig.HookConfig.SplunkConfig.SendLogs) splunkClient.Send(telemetryClient.GetData(), logCollector) } if len(GeneralConfig.HookConfig.SplunkConfig.ProdCriblEndpoint) > 0 { splunkClient.Initialize(GeneralConfig.CorrelationID, GeneralConfig.HookConfig.SplunkConfig.ProdCriblEndpoint, GeneralConfig.HookConfig.SplunkConfig.ProdCriblToken, GeneralConfig.HookConfig.SplunkConfig.ProdCriblIndex, GeneralConfig.HookConfig.SplunkConfig.SendLogs) splunkClient.Send(telemetryClient.GetData(), logCollector) } } log.DeferExitHandler(handler) defer handler() telemetryClient.Initialize(GeneralConfig.NoTelemetry, STEP_NAME) tmsExport(stepConfig, &stepTelemetryData, &influx) stepTelemetryData.ErrorCode = "0" log.Entry().Info("SUCCESS") }, } addTmsExportFlags(createTmsExportCmd, &stepConfig) return createTmsExportCmd } func addTmsExportFlags(cmd *cobra.Command, stepConfig *tmsExportOptions) { cmd.Flags().StringVar(&stepConfig.TmsServiceKey, "tmsServiceKey", os.Getenv("PIPER_tmsServiceKey"), "Service key JSON string to access the SAP Cloud Transport Management service instance APIs. If not specified and if pipeline is running on Jenkins, service key, stored under ID provided with credentialsId parameter, is used.") cmd.Flags().StringVar(&stepConfig.CustomDescription, "customDescription", os.Getenv("PIPER_customDescription"), "Can be used as the description of a transport request. Will overwrite the default, which is corresponding Git commit ID.") cmd.Flags().StringVar(&stepConfig.NamedUser, "namedUser", `Piper-Pipeline`, "Defines the named user to execute transport request with. The default value is 'Piper-Pipeline'. If pipeline is running on Jenkins, the name of the user, who started the job, is tried to be used at first.") cmd.Flags().StringVar(&stepConfig.NodeName, "nodeName", os.Getenv("PIPER_nodeName"), "Defines the name of the export node - starting node in TMS landscape. The transport request is added to the queues of the follow-on nodes of export node.") cmd.Flags().StringVar(&stepConfig.MtaPath, "mtaPath", os.Getenv("PIPER_mtaPath"), "Defines the relative path to *.mtar file for the export to the SAP Cloud Transport Management service. If not specified, it will use the *.mtar file created in mtaBuild.") cmd.Flags().StringVar(&stepConfig.MtaVersion, "mtaVersion", `*`, "Defines the version of the MTA for which the MTA extension descriptor will be used. You can use an asterisk (*) to accept any MTA version, or use a specific version compliant with SemVer 2.0, e.g. 1.0.0 (see semver.org). If the parameter is not configured, an asterisk is used.") cmd.Flags().StringVar(&stepConfig.Proxy, "proxy", os.Getenv("PIPER_proxy"), "Proxy URL which should
persist
identifier_name
tmsExport_generated.go
var startTime time.Time var influx tmsExportInflux var logCollector *log.CollectorHook var splunkClient *splunk.Splunk telemetryClient := &telemetry.Telemetry{} var createTmsExportCmd = &cobra.Command{ Use: STEP_NAME, Short: "This step allows you to export an MTA file (multi-target application archive) and multiple MTA extension descriptors into a TMS (SAP Cloud Transport Management service) landscape for further TMS-controlled distribution through a TMS-configured landscape.", Long: `This step allows you to export an MTA file (multi-target application archive) and multiple MTA extension descriptors into a TMS (SAP Cloud Transport Management service) landscape for further TMS-controlled distribution through a TMS-configured landscape. The MTA file is attached to a new transport request which is added to the import queues of the follow-on transport nodes of the specified export node. TMS lets you manage transports between SAP Business Technology Platform accounts in Neo and Cloud Foundry, such as from DEV to TEST and PROD accounts. For more information, see [official documentation of SAP Cloud Transport Management service](https://help.sap.com/viewer/p/TRANSPORT_MANAGEMENT_SERVICE) !!! note "Prerequisites" * You have subscribed to and set up TMS, as described in [Initial Setup](https://help.sap.com/viewer/7f7160ec0d8546c6b3eab72fb5ad6fd8/Cloud/en-US/66fd7283c62f48adb23c56fb48c84a60.html), which includes the configuration of your transport landscape. * A corresponding service key has been created, as described in [Set Up the Environment to Transport Content Archives directly in an Application](https://help.sap.com/viewer/7f7160ec0d8546c6b3eab72fb5ad6fd8/Cloud/en-US/8d9490792ed14f1bbf8a6ac08a6bca64.html). This service key (JSON) must be stored as a secret text within the Jenkins secure store or provided as value of tmsServiceKey parameter.`, PreRunE: func(cmd *cobra.Command, _ []string) error { startTime = time.Now() log.SetStepName(STEP_NAME) log.SetVerbose(GeneralConfig.Verbose) GeneralConfig.GitHubAccessTokens = ResolveAccessTokens(GeneralConfig.GitHubTokens) path, _ := os.Getwd() fatalHook := &log.FatalHook{CorrelationID: GeneralConfig.CorrelationID, Path: path} log.RegisterHook(fatalHook) err := PrepareConfig(cmd, &metadata, STEP_NAME, &stepConfig, config.OpenPiperFile) if err != nil { log.SetErrorCategory(log.ErrorConfiguration) return err } log.RegisterSecret(stepConfig.TmsServiceKey) if len(GeneralConfig.HookConfig.SentryConfig.Dsn) > 0 { sentryHook := log.NewSentryHook(GeneralConfig.HookConfig.SentryConfig.Dsn, GeneralConfig.CorrelationID) log.RegisterHook(&sentryHook) } if len(GeneralConfig.HookConfig.SplunkConfig.Dsn) > 0 { splunkClient = &splunk.Splunk{} logCollector = &log.CollectorHook{CorrelationID: GeneralConfig.CorrelationID} log.RegisterHook(logCollector) } if err = log.RegisterANSHookIfConfigured(GeneralConfig.CorrelationID); err != nil { log.Entry().WithError(err).Warn("failed to set up SAP Alert Notification Service log hook") } validation, err := validation.New(validation.WithJSONNamesForStructFields(), validation.WithPredefinedErrorMessages()) if err != nil { return err } if err = validation.ValidateStruct(stepConfig); err != nil { log.SetErrorCategory(log.ErrorConfiguration) return err } return nil }, Run: func(_ *cobra.Command, _ []string) { stepTelemetryData := telemetry.CustomData{} stepTelemetryData.ErrorCode = "1" handler := func() { influx.persist(GeneralConfig.EnvRootPath, "influx") config.RemoveVaultSecretFiles() stepTelemetryData.Duration = fmt.Sprintf("%v", time.Since(startTime).Milliseconds()) stepTelemetryData.ErrorCategory = log.GetErrorCategory().String() stepTelemetryData.PiperCommitHash = GitCommit telemetryClient.SetData(&stepTelemetryData) telemetryClient.Send() if len(GeneralConfig.HookConfig.SplunkConfig.Dsn) > 0 { splunkClient.Initialize(GeneralConfig.CorrelationID, GeneralConfig.HookConfig.SplunkConfig.Dsn, GeneralConfig.HookConfig.SplunkConfig.Token, GeneralConfig.HookConfig.SplunkConfig.Index, GeneralConfig.HookConfig.SplunkConfig.SendLogs) splunkClient.Send(telemetryClient.GetData(), logCollector) } if len(GeneralConfig.HookConfig.SplunkConfig.ProdCriblEndpoint) > 0 { splunkClient.Initialize(GeneralConfig.CorrelationID, GeneralConfig.HookConfig.SplunkConfig.ProdCriblEndpoint, GeneralConfig.HookConfig.SplunkConfig.ProdCriblToken, GeneralConfig.HookConfig.SplunkConfig.ProdCriblIndex, GeneralConfig.HookConfig.SplunkConfig.SendLogs) splunkClient.Send(telemetryClient.GetData(), logCollector) } } log.DeferExitHandler(handler) defer handler() telemetryClient.Initialize(GeneralConfig.NoTelemetry, STEP_NAME) tmsExport(stepConfig, &stepTelemetryData, &influx) stepTelemetryData.ErrorCode = "0" log.Entry().Info("SUCCESS") }, } addTmsExportFlags(createTmsExportCmd, &stepConfig) return createTmsExportCmd } func addTmsExportFlags(cmd *cobra.Command, stepConfig *tmsExportOptions) { cmd.Flags().StringVar(&stepConfig.TmsServiceKey, "tmsServiceKey", os.Getenv("PIPER_tmsServiceKey"), "Service key JSON string to access the SAP Cloud Transport Management service instance APIs. If not specified and if pipeline is running on Jenkins, service key, stored under ID provided with credentialsId parameter, is used.") cmd.Flags().StringVar(&stepConfig.CustomDescription, "customDescription", os.Getenv("PIPER_customDescription"), "Can be used as the description of a transport request. Will overwrite the default, which is corresponding Git commit ID.") cmd.Flags().StringVar(&stepConfig.NamedUser, "namedUser", `Piper-Pipeline`, "Defines the named user to execute transport request with. The default value is 'Piper-Pipeline'. If pipeline is running on Jenkins, the name of the user, who started the job, is tried to be used at first.") cmd.Flags().StringVar(&stepConfig.NodeName, "nodeName", os.Getenv("PIPER_nodeName"), "Defines the name of the export node - starting node in TMS landscape. The transport request is added to the queues of the follow-on nodes of export node.") cmd.Flags().StringVar(&stepConfig.MtaPath, "mtaPath", os.Getenv("PIPER_mtaPath"), "Defines the relative path to *.mtar file for the export to the SAP Cloud Transport Management service. If not specified, it will use the *.mtar file created in mtaBuild.") cmd.Flags().StringVar(&stepConfig.MtaVersion, "mtaVersion", `*`, "Defines the version of the MTA for which the MTA extension descriptor will be used. You can use an asterisk (*) to accept any MTA version, or use a specific version compliant with SemVer 2.0, e.g. 1.0.0 (see semver.org). If the parameter is not configured, an asterisk is used.") cmd.Flags().StringVar(&stepConfig.Proxy, "proxy", os.Getenv("PIPER_proxy"), "Proxy URL which should be used for communication with the SAP Cloud Transport Management service backend.") cmd.MarkFlagRequired("tmsServiceKey") cmd.MarkFlagRequired("nodeName") } // retrieve step metadata func tmsExportMetadata() config.StepData
{ var theMetaData = config.StepData{ Metadata: config.StepMetadata{ Name: "tmsExport", Aliases: []config.Alias{}, Description: "This step allows you to export an MTA file (multi-target application archive) and multiple MTA extension descriptors into a TMS (SAP Cloud Transport Management service) landscape for further TMS-controlled distribution through a TMS-configured landscape.", }, Spec: config.StepSpec{ Inputs: config.StepInputs{ Secrets: []config.StepSecrets{ {Name: "credentialsId", Description: "Jenkins 'Secret text' credentials ID containing service key for SAP Cloud Transport Management service.", Type: "jenkins"}, }, Resources: []config.StepResources{ {Name: "buildResult", Type: "stash"}, }, Parameters: []config.StepParameters{ { Name: "tmsServiceKey", ResourceRef: []config.ResourceReference{ {
identifier_body
tmsExport_generated.go
.name), metric.value) if err != nil { log.Entry().WithError(err).Error("Error persisting influx environment.") errCount++ } } if errCount > 0 { log.Entry().Error("failed to persist Influx environment") } } // TmsExportCommand This step allows you to export an MTA file (multi-target application archive) and multiple MTA extension descriptors into a TMS (SAP Cloud Transport Management service) landscape for further TMS-controlled distribution through a TMS-configured landscape. func TmsExportCommand() *cobra.Command { const STEP_NAME = "tmsExport" metadata := tmsExportMetadata() var stepConfig tmsExportOptions var startTime time.Time var influx tmsExportInflux var logCollector *log.CollectorHook var splunkClient *splunk.Splunk telemetryClient := &telemetry.Telemetry{} var createTmsExportCmd = &cobra.Command{ Use: STEP_NAME, Short: "This step allows you to export an MTA file (multi-target application archive) and multiple MTA extension descriptors into a TMS (SAP Cloud Transport Management service) landscape for further TMS-controlled distribution through a TMS-configured landscape.", Long: `This step allows you to export an MTA file (multi-target application archive) and multiple MTA extension descriptors into a TMS (SAP Cloud Transport Management service) landscape for further TMS-controlled distribution through a TMS-configured landscape. The MTA file is attached to a new transport request which is added to the import queues of the follow-on transport nodes of the specified export node. TMS lets you manage transports between SAP Business Technology Platform accounts in Neo and Cloud Foundry, such as from DEV to TEST and PROD accounts. For more information, see [official documentation of SAP Cloud Transport Management service](https://help.sap.com/viewer/p/TRANSPORT_MANAGEMENT_SERVICE) !!! note "Prerequisites" * You have subscribed to and set up TMS, as described in [Initial Setup](https://help.sap.com/viewer/7f7160ec0d8546c6b3eab72fb5ad6fd8/Cloud/en-US/66fd7283c62f48adb23c56fb48c84a60.html), which includes the configuration of your transport landscape. * A corresponding service key has been created, as described in [Set Up the Environment to Transport Content Archives directly in an Application](https://help.sap.com/viewer/7f7160ec0d8546c6b3eab72fb5ad6fd8/Cloud/en-US/8d9490792ed14f1bbf8a6ac08a6bca64.html). This service key (JSON) must be stored as a secret text within the Jenkins secure store or provided as value of tmsServiceKey parameter.`, PreRunE: func(cmd *cobra.Command, _ []string) error { startTime = time.Now() log.SetStepName(STEP_NAME) log.SetVerbose(GeneralConfig.Verbose) GeneralConfig.GitHubAccessTokens = ResolveAccessTokens(GeneralConfig.GitHubTokens) path, _ := os.Getwd() fatalHook := &log.FatalHook{CorrelationID: GeneralConfig.CorrelationID, Path: path} log.RegisterHook(fatalHook) err := PrepareConfig(cmd, &metadata, STEP_NAME, &stepConfig, config.OpenPiperFile) if err != nil { log.SetErrorCategory(log.ErrorConfiguration) return err
log.RegisterSecret(stepConfig.TmsServiceKey) if len(GeneralConfig.HookConfig.SentryConfig.Dsn) > 0 { sentryHook := log.NewSentryHook(GeneralConfig.HookConfig.SentryConfig.Dsn, GeneralConfig.CorrelationID) log.RegisterHook(&sentryHook) } if len(GeneralConfig.HookConfig.SplunkConfig.Dsn) > 0 { splunkClient = &splunk.Splunk{} logCollector = &log.CollectorHook{CorrelationID: GeneralConfig.CorrelationID} log.RegisterHook(logCollector) } if err = log.RegisterANSHookIfConfigured(GeneralConfig.CorrelationID); err != nil { log.Entry().WithError(err).Warn("failed to set up SAP Alert Notification Service log hook") } validation, err := validation.New(validation.WithJSONNamesForStructFields(), validation.WithPredefinedErrorMessages()) if err != nil { return err } if err = validation.ValidateStruct(stepConfig); err != nil { log.SetErrorCategory(log.ErrorConfiguration) return err } return nil }, Run: func(_ *cobra.Command, _ []string) { stepTelemetryData := telemetry.CustomData{} stepTelemetryData.ErrorCode = "1" handler := func() { influx.persist(GeneralConfig.EnvRootPath, "influx") config.RemoveVaultSecretFiles() stepTelemetryData.Duration = fmt.Sprintf("%v", time.Since(startTime).Milliseconds()) stepTelemetryData.ErrorCategory = log.GetErrorCategory().String() stepTelemetryData.PiperCommitHash = GitCommit telemetryClient.SetData(&stepTelemetryData) telemetryClient.Send() if len(GeneralConfig.HookConfig.SplunkConfig.Dsn) > 0 { splunkClient.Initialize(GeneralConfig.CorrelationID, GeneralConfig.HookConfig.SplunkConfig.Dsn, GeneralConfig.HookConfig.SplunkConfig.Token, GeneralConfig.HookConfig.SplunkConfig.Index, GeneralConfig.HookConfig.SplunkConfig.SendLogs) splunkClient.Send(telemetryClient.GetData(), logCollector) } if len(GeneralConfig.HookConfig.SplunkConfig.ProdCriblEndpoint) > 0 { splunkClient.Initialize(GeneralConfig.CorrelationID, GeneralConfig.HookConfig.SplunkConfig.ProdCriblEndpoint, GeneralConfig.HookConfig.SplunkConfig.ProdCriblToken, GeneralConfig.HookConfig.SplunkConfig.ProdCriblIndex, GeneralConfig.HookConfig.SplunkConfig.SendLogs) splunkClient.Send(telemetryClient.GetData(), logCollector) } } log.DeferExitHandler(handler) defer handler() telemetryClient.Initialize(GeneralConfig.NoTelemetry, STEP_NAME) tmsExport(stepConfig, &stepTelemetryData, &influx) stepTelemetryData.ErrorCode = "0" log.Entry().Info("SUCCESS") }, } addTmsExportFlags(createTmsExportCmd, &stepConfig) return createTmsExportCmd } func addTmsExportFlags(cmd *cobra.Command, stepConfig *tmsExportOptions) { cmd.Flags().StringVar(&stepConfig.TmsServiceKey, "tmsServiceKey", os.Getenv("PIPER_tmsServiceKey"), "Service key JSON string to access the SAP Cloud Transport Management service instance APIs. If not specified and if pipeline is running on Jenkins, service key, stored under ID provided with credentialsId parameter, is used.") cmd.Flags().StringVar(&stepConfig.CustomDescription, "customDescription", os.Getenv("PIPER_customDescription"), "Can be used as the description of a transport request. Will overwrite the default, which is corresponding Git commit ID.") cmd.Flags().StringVar(&stepConfig.NamedUser, "namedUser", `Piper-Pipeline`, "Defines the named user to execute transport request with. The default value is 'Piper-Pipeline'. If pipeline is running on Jenkins, the name of the user, who started the job, is tried to be used at first.") cmd.Flags().StringVar(&stepConfig.NodeName, "nodeName", os.Getenv("PIPER_nodeName"), "Defines the name of the export node - starting node in TMS landscape. The transport request is added to the queues of the follow-on nodes of export node.") cmd.Flags().StringVar(&stepConfig.MtaPath, "mtaPath", os.Getenv("PIPER_mtaPath"), "Defines the relative path to *.mtar file for the export to the SAP Cloud Transport Management service. If not specified, it will use the *.mtar file created in mtaBuild.") cmd.Flags().StringVar(&stepConfig.MtaVersion, "mtaVersion", `*`, "Defines the version of the MTA for which the MTA extension descriptor will be used. You can use an asterisk (*) to accept any MTA version, or use a specific version compliant with SemVer 2.0, e.g. 1.0.0 (see semver.org). If the parameter is not configured, an asterisk is used.") cmd.Flags().StringVar(&stepConfig.Proxy, "proxy", os.Getenv("PIPER_proxy"), "Proxy URL which should be used for communication with the SAP Cloud Transport Management service backend.") cmd.MarkFlagRequired("tmsServiceKey") cmd.MarkFlagRequired("nodeName") } // retrieve step metadata func tmsExportMetadata() config.StepData { var theMetaData = config.StepData{ Metadata: config.StepMetadata{ Name: "tmsExport", Aliases: []config.Alias{}, Description: "This step allows you to export an MTA file (multi-target application archive) and multiple MTA extension descriptors into a TMS (SAP Cloud
}
random_line_split
par_granges.rs
- Optional argument to modify the default size ration of the channel that `R::P` is sent on. /// formula is: ((BYTES_INA_GIGABYTE * channel_size_modifier) * threads) / size_of(R::P) /// * `processor`- Something that implements [`RegionProcessor`](RegionProcessor) pub fn new( reads: PathBuf, ref_fasta: Option<PathBuf>, regions_bed: Option<PathBuf>, regions_bcf: Option<PathBuf>, threads: Option<usize>, chunksize: Option<u32>, channel_size_modifier: Option<f64>, processor: R, ) -> Self { let threads = if let Some(threads) = threads { threads } else { num_cpus::get() }; // Keep two around for main thread and thread running the pool let threads = std::cmp::max(threads.checked_sub(2).unwrap_or(0), 1); let pool = rayon::ThreadPoolBuilder::new() .num_threads(threads) .build() .unwrap(); info!("Using {} worker threads.", threads); Self { reads, ref_fasta, regions_bed, regions_bcf, threads, chunksize: chunksize.unwrap_or(CHUNKSIZE), channel_size_modifier: channel_size_modifier.unwrap_or(CHANNEL_SIZE_MODIFIER), pool, processor, } } /// Process each region. /// /// This method splits the sequences in the BAM/CRAM header into `chunksize` * `self.threads` regions (aka 'super chunks'). /// It then queries that 'super chunk' against the intervals (either the BED file, or the whole genome broken up into `chunksize` /// regions). The results of that query are then processed by a pool of workers that apply `process_region` to reach interval to /// do perbase analysis on. The collected result for each region is then sent back over the returned `Receiver<R::P>` channel /// for the caller to use. The results will be returned in order according to the order of the intervals used to drive this method. /// /// While one 'super chunk' is being worked on by all workers, the last 'super chunks' results are being printed to either to /// a file or to STDOUT, in order. /// /// Note, a common use case of this will be to fetch a region and do a pileup. The bounds of bases being looked at should still be /// checked since a fetch will pull all reads that overlap the region in question. pub fn process(self) -> Result<Receiver<R::P>> { let channel_size: usize = ((BYTES_INA_GIGABYTE as f64 * self.channel_size_modifier).floor() as usize / std::mem::size_of::<R::P>()) * self.threads; info!( "Creating channel of length {:?} (* 120 bytes to get mem)", channel_size ); let (snd, rxv) = bounded(channel_size); thread::spawn(move || { self.pool.install(|| { info!("Reading from {:?}", self.reads); let mut reader = IndexedReader::from_path(&self.reads).expect("Indexed BAM/CRAM"); // If passed add ref_fasta if let Some(ref_fasta) = &self.ref_fasta { reader.set_reference(ref_fasta).expect("Set ref"); } // Get a copy of the header let header = reader.header().to_owned(); // Work out if we are restricted to a subset of sites let bed_intervals = if let Some(regions_bed) = &self.regions_bed { Some( Self::bed_to_intervals(&header, regions_bed) .expect("Parsed BED to intervals"), ) } else { None }; let bcf_intervals = if let Some(regions_bcf) = &self.regions_bcf
else { None }; let restricted_ivs = match (bed_intervals, bcf_intervals) { (Some(bed_ivs), Some(bcf_ivs)) => Some(Self::merge_intervals(bed_ivs, bcf_ivs)), (Some(bed_ivs), None) => Some(bed_ivs), (None, Some(bcf_ivs)) => Some(bcf_ivs), (None, None) => None, }; let intervals = if let Some(restricted) = restricted_ivs { restricted } else { Self::header_to_intervals(&header, self.chunksize) .expect("Parsed BAM/CRAM header to intervals") }; // The number positions to try to process in one batch let serial_step_size = self .chunksize .checked_mul(self.threads as u32) .unwrap_or(u32::MAX); // aka superchunk for (tid, intervals) in intervals.into_iter().enumerate() { let tid: u32 = tid as u32; let tid_end: u32 = header.target_len(tid).unwrap().try_into().unwrap(); // Result holds the processed positions to be sent to writer let mut result = vec![]; for chunk_start in (0..tid_end).step_by(serial_step_size as usize) { let tid_name = std::str::from_utf8(header.tid2name(tid)).unwrap(); let chunk_end = std::cmp::min(chunk_start as u32 + serial_step_size, tid_end); info!( "Batch Processing {}:{}-{}", tid_name, chunk_start, chunk_end ); let (r, _) = rayon::join( || { // Must be a vec so that par_iter works and results stay in order let ivs: Vec<Interval<u32, ()>> = Lapper::<u32, ()>::find(&intervals, chunk_start, chunk_end) // Truncate intervals that extend forward or backward of chunk in question .map(|iv| Interval { start: std::cmp::max(iv.start, chunk_start), stop: std::cmp::min(iv.stop, chunk_end), val: (), }) .collect(); ivs.into_par_iter() .flat_map(|iv| { info!("Processing {}:{}-{}", tid_name, iv.start, iv.stop); self.processor.process_region(tid, iv.start, iv.stop) }) .collect() }, || { result.into_iter().for_each(|p| { snd.send(p).expect("Sent a serializable to writer") }) }, ); result = r; } // Send final set of results result .into_iter() .for_each(|p| snd.send(p).expect("Sent a serializable to writer")); } }); }); Ok(rxv) } // Convert the header into intervals of equally sized chunks. The last interval may be short. fn header_to_intervals(header: &HeaderView, chunksize: u32) -> Result<Vec<Lapper<u32, ()>>> { let mut intervals = vec![vec![]; header.target_count() as usize]; for tid in 0..(header.target_count()) { let tid_len: u32 = header.target_len(tid).unwrap().try_into().unwrap(); for start in (0..tid_len).step_by(chunksize as usize) { let stop = std::cmp::min(start as u32 + chunksize, tid_len); intervals[tid as usize].push(Interval { start: start as u32, stop: stop, val: (), }); } } Ok(intervals.into_iter().map(|ivs| Lapper::new(ivs)).collect()) } /// Read a bed file into a vector of lappers with the index representing the TID // TODO add a proper error message fn bed_to_intervals(header: &HeaderView, bed_file: &PathBuf) -> Result<Vec<Lapper<u32, ()>>> { let mut bed_reader = bed::Reader::from_file(bed_file)?; let mut intervals = vec![vec![]; header.target_count() as usize]; for record in bed_reader.records() { let record = record?; let tid = header .tid(record.chrom().as_bytes()) .expect("Chromosome not found in BAM/CRAM header"); intervals[tid as usize].push(Interval { start: record.start().try_into().unwrap(), stop: record.end().try_into().unwrap(), val: (), }); } Ok(intervals .into_iter() .map(|ivs| { let mut lapper = Lapper::new(ivs); lapper.merge_overlaps(); lapper }) .collect()) } /// Read a BCF/VCF file into a vector of lappers with index representing the TID fn bcf_to_intervals(header: &HeaderView, bcf_file: &PathBuf) -> Result<Vec<Lapper<u32, ()>>> { let mut bcf_reader = Reader::from_path(bcf_file).expect("Error opening BCF/VCF file."); let bcf_header_reader = Reader::from_path(bcf_file).expect("Error opening BCF/VCF file
{ Some( Self::bcf_to_intervals(&header, regions_bcf) .expect("Parsed BCF/VCF to intervals"), ) }
conditional_block
par_granges.rs
=> Some(Self::merge_intervals(bed_ivs, bcf_ivs)), (Some(bed_ivs), None) => Some(bed_ivs), (None, Some(bcf_ivs)) => Some(bcf_ivs), (None, None) => None, }; let intervals = if let Some(restricted) = restricted_ivs { restricted } else { Self::header_to_intervals(&header, self.chunksize) .expect("Parsed BAM/CRAM header to intervals") }; // The number positions to try to process in one batch let serial_step_size = self .chunksize .checked_mul(self.threads as u32) .unwrap_or(u32::MAX); // aka superchunk for (tid, intervals) in intervals.into_iter().enumerate() { let tid: u32 = tid as u32; let tid_end: u32 = header.target_len(tid).unwrap().try_into().unwrap(); // Result holds the processed positions to be sent to writer let mut result = vec![]; for chunk_start in (0..tid_end).step_by(serial_step_size as usize) { let tid_name = std::str::from_utf8(header.tid2name(tid)).unwrap(); let chunk_end = std::cmp::min(chunk_start as u32 + serial_step_size, tid_end); info!( "Batch Processing {}:{}-{}", tid_name, chunk_start, chunk_end ); let (r, _) = rayon::join( || { // Must be a vec so that par_iter works and results stay in order let ivs: Vec<Interval<u32, ()>> = Lapper::<u32, ()>::find(&intervals, chunk_start, chunk_end) // Truncate intervals that extend forward or backward of chunk in question .map(|iv| Interval { start: std::cmp::max(iv.start, chunk_start), stop: std::cmp::min(iv.stop, chunk_end), val: (), }) .collect(); ivs.into_par_iter() .flat_map(|iv| { info!("Processing {}:{}-{}", tid_name, iv.start, iv.stop); self.processor.process_region(tid, iv.start, iv.stop) }) .collect() }, || { result.into_iter().for_each(|p| { snd.send(p).expect("Sent a serializable to writer") }) }, ); result = r; } // Send final set of results result .into_iter() .for_each(|p| snd.send(p).expect("Sent a serializable to writer")); } }); }); Ok(rxv) } // Convert the header into intervals of equally sized chunks. The last interval may be short. fn header_to_intervals(header: &HeaderView, chunksize: u32) -> Result<Vec<Lapper<u32, ()>>> { let mut intervals = vec![vec![]; header.target_count() as usize]; for tid in 0..(header.target_count()) { let tid_len: u32 = header.target_len(tid).unwrap().try_into().unwrap(); for start in (0..tid_len).step_by(chunksize as usize) { let stop = std::cmp::min(start as u32 + chunksize, tid_len); intervals[tid as usize].push(Interval { start: start as u32, stop: stop, val: (), }); } } Ok(intervals.into_iter().map(|ivs| Lapper::new(ivs)).collect()) } /// Read a bed file into a vector of lappers with the index representing the TID // TODO add a proper error message fn bed_to_intervals(header: &HeaderView, bed_file: &PathBuf) -> Result<Vec<Lapper<u32, ()>>> { let mut bed_reader = bed::Reader::from_file(bed_file)?; let mut intervals = vec![vec![]; header.target_count() as usize]; for record in bed_reader.records() { let record = record?; let tid = header .tid(record.chrom().as_bytes()) .expect("Chromosome not found in BAM/CRAM header"); intervals[tid as usize].push(Interval { start: record.start().try_into().unwrap(), stop: record.end().try_into().unwrap(), val: (), }); } Ok(intervals .into_iter() .map(|ivs| { let mut lapper = Lapper::new(ivs); lapper.merge_overlaps(); lapper }) .collect()) } /// Read a BCF/VCF file into a vector of lappers with index representing the TID fn bcf_to_intervals(header: &HeaderView, bcf_file: &PathBuf) -> Result<Vec<Lapper<u32, ()>>> { let mut bcf_reader = Reader::from_path(bcf_file).expect("Error opening BCF/VCF file."); let bcf_header_reader = Reader::from_path(bcf_file).expect("Error opening BCF/VCF file."); let bcf_header = bcf_header_reader.header(); let mut intervals = vec![vec![]; header.target_count() as usize]; // TODO: validate the headers against eachother for record in bcf_reader.records() { let record = record?; let record_rid = bcf_header.rid2name(record.rid().unwrap()).unwrap(); let tid = header .tid(record_rid) .expect("Chromosome not found in BAM/CRAM header"); let pos: u32 = record .pos() .try_into() .expect("Got a negative value for pos"); intervals[tid as usize].push(Interval { start: pos, stop: pos + 1, val: (), }); } Ok(intervals .into_iter() .map(|ivs| { let mut lapper = Lapper::new(ivs); lapper.merge_overlaps(); lapper }) .collect()) } /// Merge two sets of restriction intervals together fn merge_intervals( a_ivs: Vec<Lapper<u32, ()>>, b_ivs: Vec<Lapper<u32, ()>>, ) -> Vec<Lapper<u32, ()>> { let mut intervals = vec![vec![]; a_ivs.len()]; for (i, (a_lapper, b_lapper)) in a_ivs.into_iter().zip(b_ivs.into_iter()).enumerate() { intervals[i] = a_lapper.into_iter().chain(b_lapper.into_iter()).collect(); } intervals .into_iter() .map(|ivs| { let mut lapper = Lapper::new(ivs); lapper.merge_overlaps(); lapper }) .collect() } } #[cfg(test)] mod test { use super::*; use bio::io::bed; use num_cpus; use proptest::prelude::*; use rust_htslib::{bam, bcf}; use rust_lapper::{Interval, Lapper}; use std::collections::{HashMap, HashSet}; use tempfile::tempdir; // The purpose of these tests is to demonstrate that positions are covered once under a variety of circumstances prop_compose! { fn arb_iv_start(max_iv: u64)(start in 0..max_iv/2) -> u64 { start } } prop_compose! { fn arb_iv_size(max_iv: u64)(size in 1..max_iv/2) -> u64 { size } } prop_compose! { // Create an arbitrary interval where the min size == max_iv / 2 fn arb_iv(max_iv: u64)(start in arb_iv_start(max_iv), size in arb_iv_size(max_iv)) -> Interval<u64, ()> { Interval {start, stop: start + size, val: ()} } } // Create an arbitrary number of intervals along with the expected number of positions they cover fn arb_ivs( max_iv: u64, // max iv size max_ivs: usize, // max number of intervals ) -> impl Strategy<Value = (Vec<Interval<u64, ()>>, u64, u64)> { prop::collection::vec(arb_iv(max_iv), 0..max_ivs).prop_map(|vec| { let mut furthest_right = 0; let lapper = Lapper::new(vec.clone()); let expected = lapper.cov(); for iv in vec.iter() { if iv.stop > furthest_right { furthest_right = iv.stop; } } (vec, expected, furthest_right) }) } // Create arbitrary number of contigs with arbitrary intervals each fn arb_chrs( max_chr: usize, // number of chromosomes to use max_iv: u64, // max interval size max_ivs: usize, // max number of intervals ) -> impl Strategy<Value = Vec<(Vec<Interval<u64, ()>>, u64, u64)>>
{ prop::collection::vec(arb_ivs(max_iv, max_ivs), 0..max_chr) }
identifier_body
par_granges.rs
on::ThreadPoolBuilder::new() .num_threads(threads) .build() .unwrap(); info!("Using {} worker threads.", threads); Self { reads, ref_fasta, regions_bed, regions_bcf, threads, chunksize: chunksize.unwrap_or(CHUNKSIZE), channel_size_modifier: channel_size_modifier.unwrap_or(CHANNEL_SIZE_MODIFIER), pool, processor, } } /// Process each region. /// /// This method splits the sequences in the BAM/CRAM header into `chunksize` * `self.threads` regions (aka 'super chunks'). /// It then queries that 'super chunk' against the intervals (either the BED file, or the whole genome broken up into `chunksize` /// regions). The results of that query are then processed by a pool of workers that apply `process_region` to reach interval to /// do perbase analysis on. The collected result for each region is then sent back over the returned `Receiver<R::P>` channel /// for the caller to use. The results will be returned in order according to the order of the intervals used to drive this method. /// /// While one 'super chunk' is being worked on by all workers, the last 'super chunks' results are being printed to either to /// a file or to STDOUT, in order. /// /// Note, a common use case of this will be to fetch a region and do a pileup. The bounds of bases being looked at should still be /// checked since a fetch will pull all reads that overlap the region in question. pub fn process(self) -> Result<Receiver<R::P>> { let channel_size: usize = ((BYTES_INA_GIGABYTE as f64 * self.channel_size_modifier).floor() as usize / std::mem::size_of::<R::P>()) * self.threads; info!( "Creating channel of length {:?} (* 120 bytes to get mem)", channel_size ); let (snd, rxv) = bounded(channel_size); thread::spawn(move || { self.pool.install(|| { info!("Reading from {:?}", self.reads); let mut reader = IndexedReader::from_path(&self.reads).expect("Indexed BAM/CRAM"); // If passed add ref_fasta if let Some(ref_fasta) = &self.ref_fasta { reader.set_reference(ref_fasta).expect("Set ref"); } // Get a copy of the header let header = reader.header().to_owned(); // Work out if we are restricted to a subset of sites let bed_intervals = if let Some(regions_bed) = &self.regions_bed { Some( Self::bed_to_intervals(&header, regions_bed) .expect("Parsed BED to intervals"), ) } else { None }; let bcf_intervals = if let Some(regions_bcf) = &self.regions_bcf { Some( Self::bcf_to_intervals(&header, regions_bcf) .expect("Parsed BCF/VCF to intervals"), ) } else { None }; let restricted_ivs = match (bed_intervals, bcf_intervals) { (Some(bed_ivs), Some(bcf_ivs)) => Some(Self::merge_intervals(bed_ivs, bcf_ivs)), (Some(bed_ivs), None) => Some(bed_ivs), (None, Some(bcf_ivs)) => Some(bcf_ivs), (None, None) => None, }; let intervals = if let Some(restricted) = restricted_ivs { restricted } else { Self::header_to_intervals(&header, self.chunksize) .expect("Parsed BAM/CRAM header to intervals") }; // The number positions to try to process in one batch let serial_step_size = self .chunksize .checked_mul(self.threads as u32) .unwrap_or(u32::MAX); // aka superchunk for (tid, intervals) in intervals.into_iter().enumerate() { let tid: u32 = tid as u32; let tid_end: u32 = header.target_len(tid).unwrap().try_into().unwrap(); // Result holds the processed positions to be sent to writer let mut result = vec![]; for chunk_start in (0..tid_end).step_by(serial_step_size as usize) { let tid_name = std::str::from_utf8(header.tid2name(tid)).unwrap(); let chunk_end = std::cmp::min(chunk_start as u32 + serial_step_size, tid_end); info!( "Batch Processing {}:{}-{}", tid_name, chunk_start, chunk_end ); let (r, _) = rayon::join( || { // Must be a vec so that par_iter works and results stay in order let ivs: Vec<Interval<u32, ()>> = Lapper::<u32, ()>::find(&intervals, chunk_start, chunk_end) // Truncate intervals that extend forward or backward of chunk in question .map(|iv| Interval { start: std::cmp::max(iv.start, chunk_start), stop: std::cmp::min(iv.stop, chunk_end), val: (), }) .collect(); ivs.into_par_iter() .flat_map(|iv| { info!("Processing {}:{}-{}", tid_name, iv.start, iv.stop); self.processor.process_region(tid, iv.start, iv.stop) }) .collect() }, || { result.into_iter().for_each(|p| { snd.send(p).expect("Sent a serializable to writer") }) }, ); result = r; } // Send final set of results result .into_iter() .for_each(|p| snd.send(p).expect("Sent a serializable to writer")); } }); }); Ok(rxv) } // Convert the header into intervals of equally sized chunks. The last interval may be short. fn header_to_intervals(header: &HeaderView, chunksize: u32) -> Result<Vec<Lapper<u32, ()>>> { let mut intervals = vec![vec![]; header.target_count() as usize]; for tid in 0..(header.target_count()) { let tid_len: u32 = header.target_len(tid).unwrap().try_into().unwrap(); for start in (0..tid_len).step_by(chunksize as usize) { let stop = std::cmp::min(start as u32 + chunksize, tid_len); intervals[tid as usize].push(Interval { start: start as u32, stop: stop, val: (), }); } } Ok(intervals.into_iter().map(|ivs| Lapper::new(ivs)).collect()) } /// Read a bed file into a vector of lappers with the index representing the TID // TODO add a proper error message fn bed_to_intervals(header: &HeaderView, bed_file: &PathBuf) -> Result<Vec<Lapper<u32, ()>>> { let mut bed_reader = bed::Reader::from_file(bed_file)?; let mut intervals = vec![vec![]; header.target_count() as usize]; for record in bed_reader.records() { let record = record?; let tid = header .tid(record.chrom().as_bytes()) .expect("Chromosome not found in BAM/CRAM header"); intervals[tid as usize].push(Interval { start: record.start().try_into().unwrap(), stop: record.end().try_into().unwrap(), val: (), }); } Ok(intervals .into_iter() .map(|ivs| { let mut lapper = Lapper::new(ivs); lapper.merge_overlaps(); lapper }) .collect()) } /// Read a BCF/VCF file into a vector of lappers with index representing the TID fn bcf_to_intervals(header: &HeaderView, bcf_file: &PathBuf) -> Result<Vec<Lapper<u32, ()>>> { let mut bcf_reader = Reader::from_path(bcf_file).expect("Error opening BCF/VCF file."); let bcf_header_reader = Reader::from_path(bcf_file).expect("Error opening BCF/VCF file."); let bcf_header = bcf_header_reader.header(); let mut intervals = vec![vec![]; header.target_count() as usize]; // TODO: validate the headers against eachother for record in bcf_reader.records() { let record = record?; let record_rid = bcf_header.rid2name(record.rid().unwrap()).unwrap(); let tid = header .tid(record_rid) .expect("Chromosome not found in BAM/CRAM header"); let pos: u32 = record .pos() .try_into() .expect("Got a negative value for pos"); intervals[tid as usize].push(Interval { start: pos, stop: pos + 1, val: (), }); } Ok(intervals .into_iter() .map(|ivs| { let mut lapper = Lapper::new(ivs);
lapper.merge_overlaps(); lapper })
random_line_split
par_granges.rs
val: (), }); } Ok(intervals .into_iter() .map(|ivs| { let mut lapper = Lapper::new(ivs); lapper.merge_overlaps(); lapper }) .collect()) } /// Merge two sets of restriction intervals together fn merge_intervals( a_ivs: Vec<Lapper<u32, ()>>, b_ivs: Vec<Lapper<u32, ()>>, ) -> Vec<Lapper<u32, ()>> { let mut intervals = vec![vec![]; a_ivs.len()]; for (i, (a_lapper, b_lapper)) in a_ivs.into_iter().zip(b_ivs.into_iter()).enumerate() { intervals[i] = a_lapper.into_iter().chain(b_lapper.into_iter()).collect(); } intervals .into_iter() .map(|ivs| { let mut lapper = Lapper::new(ivs); lapper.merge_overlaps(); lapper }) .collect() } } #[cfg(test)] mod test { use super::*; use bio::io::bed; use num_cpus; use proptest::prelude::*; use rust_htslib::{bam, bcf}; use rust_lapper::{Interval, Lapper}; use std::collections::{HashMap, HashSet}; use tempfile::tempdir; // The purpose of these tests is to demonstrate that positions are covered once under a variety of circumstances prop_compose! { fn arb_iv_start(max_iv: u64)(start in 0..max_iv/2) -> u64 { start } } prop_compose! { fn arb_iv_size(max_iv: u64)(size in 1..max_iv/2) -> u64 { size } } prop_compose! { // Create an arbitrary interval where the min size == max_iv / 2 fn arb_iv(max_iv: u64)(start in arb_iv_start(max_iv), size in arb_iv_size(max_iv)) -> Interval<u64, ()> { Interval {start, stop: start + size, val: ()} } } // Create an arbitrary number of intervals along with the expected number of positions they cover fn arb_ivs( max_iv: u64, // max iv size max_ivs: usize, // max number of intervals ) -> impl Strategy<Value = (Vec<Interval<u64, ()>>, u64, u64)> { prop::collection::vec(arb_iv(max_iv), 0..max_ivs).prop_map(|vec| { let mut furthest_right = 0; let lapper = Lapper::new(vec.clone()); let expected = lapper.cov(); for iv in vec.iter() { if iv.stop > furthest_right { furthest_right = iv.stop; } } (vec, expected, furthest_right) }) } // Create arbitrary number of contigs with arbitrary intervals each fn arb_chrs( max_chr: usize, // number of chromosomes to use max_iv: u64, // max interval size max_ivs: usize, // max number of intervals ) -> impl Strategy<Value = Vec<(Vec<Interval<u64, ()>>, u64, u64)>> { prop::collection::vec(arb_ivs(max_iv, max_ivs), 0..max_chr) } // An empty BAM with correct header // A BED file with the randomly generated intervals (with expected number of positions) // proptest generate random chunksize, cpus proptest! { #[test] // add random chunksize and random cpus // NB: using any larger numbers for this tends to blow up the test runtime fn interval_set(chromosomes in arb_chrs(4, 10_000, 1_000), chunksize in any::<u32>(), cpus in 0..num_cpus::get(), use_bed in any::<bool>(), use_vcf in any::<bool>()) { let tempdir = tempdir().unwrap(); let bam_path = tempdir.path().join("test.bam"); let bed_path = tempdir.path().join("test.bed"); let vcf_path = tempdir.path().join("test.vcf"); // Build a BAM let mut header = bam::header::Header::new(); for (i,chr) in chromosomes.iter().enumerate() { let mut chr_rec = bam::header::HeaderRecord::new(b"SQ"); chr_rec.push_tag(b"SN", &i.to_string()); chr_rec.push_tag(b"LN", &chr.2.to_string()); // set len as max observed header.push_record(&chr_rec); } let writer = bam::Writer::from_path(&bam_path, &header, bam::Format::BAM).expect("Opened test.bam for writing"); drop(writer); // force flush the writer so the header info is written bam::index::build(&bam_path, None, bam::index::Type::BAI, 1).unwrap(); // Build a bed let mut writer = bed::Writer::to_file(&bed_path).expect("Opened test.bed for writing"); for (i, chr) in chromosomes.iter().enumerate() { for iv in chr.0.iter() { let mut record = bed::Record::new(); record.set_start(iv.start); record.set_end(iv.stop); record.set_chrom(&i.to_string()); record.set_score(&0.to_string()); writer.write(&record).expect("Wrote to test.bed"); } } drop(writer); // force flush // Build a VCF file let mut vcf_truth = HashMap::new(); let mut header = bcf::header::Header::new(); for (i,chr) in chromosomes.iter().enumerate() { header.push_record(format!("##contig=<ID={},length={}>", &i.to_string(), &chr.2.to_string()).as_bytes()); } let mut writer = bcf::Writer::from_path(&vcf_path, &header, true, bcf::Format::VCF).expect("Failed to open test.vcf for writing"); let mut record = writer.empty_record(); for (i, chr) in chromosomes.iter().enumerate() { record.set_rid(Some(i as u32)); let counter = vcf_truth.entry(i).or_insert(0); let mut seen = HashSet::new(); for iv in chr.0.iter() { if !seen.contains(&iv.start) { *counter += 1; seen.insert(iv.start); } record.set_pos(iv.start as i64); writer.write(&record).expect("Failed to write to test.vcf") } } drop(writer); // force flush // Create the processor with a dumb impl of processing that just returns positions with no counting let test_processor = TestProcessor {}; let par_granges_runner = ParGranges::new( bam_path, None, if use_bed { Some(bed_path) } else { None }, // do one with regions if use_vcf { Some(vcf_path) } else { None }, // do one with vcf regions Some(cpus), Some(chunksize), Some(0.002), test_processor ); let receiver = par_granges_runner.process().expect("Launch ParGranges Process"); let mut chrom_counts = HashMap::new(); receiver.into_iter().for_each(|p: PileupPosition| { let positions = chrom_counts.entry(p.ref_seq.parse::<usize>().expect("parsed chr")).or_insert(0u64); *positions += 1 }); // Validate that for each chr we get the expected number of bases for (chrom, positions) in chrom_counts.iter() { if use_bed && !use_vcf { // if this was with bed, should be equal to .1 prop_assert_eq!(chromosomes[*chrom].1, *positions, "chr: {}, expected: {}, found: {}", chrom, chromosomes[*chrom].1, positions); } else if use_bed && use_vcf { // if this was with bed, should be equal to .1, bed restrictions and vcf restrctions should overlap prop_assert_eq!(chromosomes[*chrom].1, *positions, "chr: {}, expected: {}, found: {}", chrom, chromosomes[*chrom].1, positions); } else if use_vcf && !use_bed { // total positions should be equal to the number of records for that chr in the vcf prop_assert_eq!(vcf_truth.get(chrom).unwrap(), positions, "chr: {}, expected: {}, found: {}", chrom, chromosomes[*chrom].1, positions); } else { // if this was bam only, should be equal to rightmost postion prop_assert_eq!(chromosomes[*chrom].2, *positions, "chr: {}, expected: {}, found: {}", chrom, chromosomes[*chrom].2, positions); } } } } use crate::position::{pileup_position::PileupPosition, Position}; use smartstring::SmartString; struct TestProcessor {} impl RegionProcessor for TestProcessor { type P = PileupPosition; fn
process_region
identifier_name
medrs.rs
= File::open(filename).expect(format!("no such file: {}", filename).as_str()); let buf = BufReader::new(file); buf.lines() .map(|l| l.expect("Could not parse line")) .collect() } fn read_file_to_string(filename: &str) -> String { let mut file = match File::open(filename) { Ok(file) => file, Err(_) => panic!("no such file"), }; let mut file_contents = String::new(); file.read_to_string(&mut file_contents) .ok() .expect("failed to read!"); file_contents } fn replace_sparql_placeolder(pattern: &str, sparql: &String, lines: &Vec<String>) -> String { let rep: String = if lines.is_empty() { "".to_string() } else { "wd:".to_string() + &lines.join(" wd:") }; sparql.replace(pattern, &rep) } fn output_sparql_result_items(sparql: &String) { let api = Api::new("https://www.wikidata.org/w/api.php").expect("Can't connect to Wikidata"); let result = api.sparql_query(&sparql).expect("SPARQL query failed"); let varname = result["head"]["vars"][0] .as_str() .expect("Can't find first variable name in SPARQL result"); let entities = api.entities_from_sparql_result(&result, &varname); println!("{}", entities.join("\n")); } /* fn get_all_from_stdin() -> String { let mut payload = Vec::new(); io::stdin().read_to_end(&mut payload).unwrap(); let s = match str::from_utf8(&payload) { Ok(v) => v, Err(e) => panic!("Invalid UTF-8 sequence: {}", e), }; s.to_string() } */ fn command_query(args: &Args) { if args.arg_query.is_empty() { println!("Requires SPARQL query"); } let sparql = &args.arg_query; output_sparql_result_items(&sparql); } fn command_run(args: &Args) { let articles = lines_from_file(&args.flag_articles); let reviews = lines_from_file(&args.flag_reviews); let topics = lines_from_file(&args.flag_topics); let journals = lines_from_file(&args.flag_journals); let publishers = lines_from_file(&args.flag_publishers); let mut sparql = read_file_to_string(&args.flag_sparql); sparql = replace_sparql_placeolder("%%ARTICLES%%", &sparql, &articles); sparql = replace_sparql_placeolder("%%REVIEWS%%", &sparql, &reviews); sparql = replace_sparql_placeolder("%%TOPICS%%", &sparql, &topics); sparql = replace_sparql_placeolder("%%JOURNALS%%", &sparql, &journals); sparql = replace_sparql_placeolder("%%PUBLISHERS%%", &sparql, &publishers); output_sparql_result_items(&sparql); } fn get_api_url_for_wiki(wiki: &String) -> Option<String> { // Get site matrix from wikidata let api = Api::new("https://www.wikidata.org/w/api.php").expect("Can't connect to Wikidata"); let params = api.params_into(&vec![("action", "sitematrix")]); let site_matrix = api .get_query_api_json(&params) .expect("Can't load sitematrix from wikidata API"); //println!("{:#?}", &site_matrix); // Go through the "normal" objects let mut ret: Option<String> = None; site_matrix["sitematrix"] .as_object() .expect("sitematrix is not an object") .iter() .for_each(|(_, data)| { match data["site"] .as_array() .unwrap_or(&vec![]) .iter() .filter_map(|x| { if x["dbname"].as_str().unwrap_or("") == wiki { x["url"].as_str() } else { None } }) .next() { Some(url) => { ret = Some(url.to_string() + "/w/api.php"); } None => {} } }); // Try the "specials" site_matrix["sitematrix"]["specials"] .as_array() .unwrap_or(&vec![]) .iter() .for_each(|x| { if x["dbname"].as_str().unwrap_or("") == wiki { ret = Some(x["url"].as_str().unwrap_or("").to_string() + "/w/api.php"); } }); ret } fn get_external_urls(api_url: &String, title: &String) -> Vec<String> { let api = Api::new(&api_url).expect(&format!("Can't connect to {}", &api_url)); let params = api.params_into(&vec![
("action", "query"), ("prop", "extlinks"), ("ellimit", "500"), ("titles", title.as_str()), ]); let result = api .get_query_api_json_all(&params) .expect("query.extlinks failed"); let mut urls: Vec<String> = vec![]; result["query"]["pages"] .as_object() .expect("query.pages in result not an object") .iter() .for_each(|(_page_id, data)| { data["extlinks"] .as_array() .expect("extlinks not an array") .iter() .for_each(|x| urls.push(x["*"].as_str().expect("* not a string").to_string())); }); urls } fn get_paper_q(api: &Api, id: &GenericWorkIdentifier) -> Option<String> { let wdp = WikidataPapers::new(); match &id.work_type { GenericWorkType::Property(prop) => { let result = wdp.search_external_id(&prop, &id.id, api); result.get(0).map(|s| s.to_owned()) // First one will do } _ => None, } /* wdp.add_adapter(Box::new(Pubmed2Wikidata::new())); wdp.add_adapter(Box::new(Crossref2Wikidata::new())); wdp.add_adapter(Box::new(Semanticscholar2Wikidata::new())); wdp.add_adapter(Box::new(Orcid2Wikidata::new())); let ids = vec![id.to_owned()]; let ids = wdp.update_from_paper_ids(&ids); let q = ids .iter() .filter_map(|x| match x.work_type { GenericWorkType::Item => Some(x.id.to_owned()), _ => None, }) .next(); q*/ } fn command_refs(args: &Args) { if args.arg_wiki.is_empty() { panic!("wiki code (e.g. 'enwiki') is required"); } if args.arg_title.is_empty() { panic!("article title is required"); } let wiki = &args.arg_wiki; let title = &args.arg_title; // Get the API URL for the wiki let api_url = match get_api_url_for_wiki(&wiki) { Some(url) => url, None => panic!("Can't find API URL for {}", &wiki), }; // Get all external URLs from that page, on that wiki let urls = get_external_urls(&api_url, &title); //println!("{:#?}", &urls); lazy_static! { static ref RE_DOI: Regex = Regex::new(r#"^*.?//doi.org/(.+)$"#).unwrap(); static ref RE_PMID: Regex = Regex::new(r#"^*.?//www.ncbi.nlm.nih.gov/pubmed/(\d+)$"#).unwrap(); static ref RE_PMCID: Regex = Regex::new(r#"^*.?//www.ncbi.nlm.nih.gov/pmc/articles/PMC(\d+)$"#).unwrap(); } let mut ids: Vec<GenericWorkIdentifier> = vec![]; for url in urls { match RE_DOI.captures(&url) { Some(caps) => { let id = caps.get(1).unwrap().as_str(); match urlencoding::decode(&id) { Ok(id) => { ids.push(GenericWorkIdentifier::new_prop(PROP_DOI, &id)); } _ => {} } } None => {} } match RE_PMID.captures(&url) { Some(caps) => { let id = caps.get(1).unwrap().as_str(); ids.push(GenericWorkIdentifier::new_prop(PROP_PMID, id)); } None => {} } match RE_PMCID.captures(&url) { Some(caps) => { let id = caps.get(1).unwrap().as_str(); ids.push(GenericWorkIdentifier::new_prop(PROP_PMCID, id)); } None => {} } } let api = Api::new("https://www.wikidata.org/w/api.php").expect("Can't connect to Wikidata"); for id in ids { match get_paper_q(&api, &id
random_line_split
medrs.rs
(filename: &str) -> Vec<String> { if filename.is_empty() { return vec![]; } let file = File::open(filename).expect(format!("no such file: {}", filename).as_str()); let buf = BufReader::new(file); buf.lines() .map(|l| l.expect("Could not parse line")) .collect() } fn read_file_to_string(filename: &str) -> String { let mut file = match File::open(filename) { Ok(file) => file, Err(_) => panic!("no such file"), }; let mut file_contents = String::new(); file.read_to_string(&mut file_contents) .ok() .expect("failed to read!"); file_contents } fn replace_sparql_placeolder(pattern: &str, sparql: &String, lines: &Vec<String>) -> String { let rep: String = if lines.is_empty() { "".to_string() } else { "wd:".to_string() + &lines.join(" wd:") }; sparql.replace(pattern, &rep) } fn output_sparql_result_items(sparql: &String) { let api = Api::new("https://www.wikidata.org/w/api.php").expect("Can't connect to Wikidata"); let result = api.sparql_query(&sparql).expect("SPARQL query failed"); let varname = result["head"]["vars"][0] .as_str() .expect("Can't find first variable name in SPARQL result"); let entities = api.entities_from_sparql_result(&result, &varname); println!("{}", entities.join("\n")); } /* fn get_all_from_stdin() -> String { let mut payload = Vec::new(); io::stdin().read_to_end(&mut payload).unwrap(); let s = match str::from_utf8(&payload) { Ok(v) => v, Err(e) => panic!("Invalid UTF-8 sequence: {}", e), }; s.to_string() } */ fn command_query(args: &Args) { if args.arg_query.is_empty() { println!("Requires SPARQL query"); } let sparql = &args.arg_query; output_sparql_result_items(&sparql); } fn command_run(args: &Args) { let articles = lines_from_file(&args.flag_articles); let reviews = lines_from_file(&args.flag_reviews); let topics = lines_from_file(&args.flag_topics); let journals = lines_from_file(&args.flag_journals); let publishers = lines_from_file(&args.flag_publishers); let mut sparql = read_file_to_string(&args.flag_sparql); sparql = replace_sparql_placeolder("%%ARTICLES%%", &sparql, &articles); sparql = replace_sparql_placeolder("%%REVIEWS%%", &sparql, &reviews); sparql = replace_sparql_placeolder("%%TOPICS%%", &sparql, &topics); sparql = replace_sparql_placeolder("%%JOURNALS%%", &sparql, &journals); sparql = replace_sparql_placeolder("%%PUBLISHERS%%", &sparql, &publishers); output_sparql_result_items(&sparql); } fn get_api_url_for_wiki(wiki: &String) -> Option<String> { // Get site matrix from wikidata let api = Api::new("https://www.wikidata.org/w/api.php").expect("Can't connect to Wikidata"); let params = api.params_into(&vec![("action", "sitematrix")]); let site_matrix = api .get_query_api_json(&params) .expect("Can't load sitematrix from wikidata API"); //println!("{:#?}", &site_matrix); // Go through the "normal" objects let mut ret: Option<String> = None; site_matrix["sitematrix"] .as_object() .expect("sitematrix is not an object") .iter() .for_each(|(_, data)| { match data["site"] .as_array() .unwrap_or(&vec![]) .iter() .filter_map(|x| { if x["dbname"].as_str().unwrap_or("") == wiki { x["url"].as_str() } else { None } }) .next() { Some(url) => { ret = Some(url.to_string() + "/w/api.php"); } None => {} } }); // Try the "specials" site_matrix["sitematrix"]["specials"] .as_array() .unwrap_or(&vec![]) .iter() .for_each(|x| { if x["dbname"].as_str().unwrap_or("") == wiki { ret = Some(x["url"].as_str().unwrap_or("").to_string() + "/w/api.php"); } }); ret } fn get_external_urls(api_url: &String, title: &String) -> Vec<String> { let api = Api::new(&api_url).expect(&format!("Can't connect to {}", &api_url)); let params = api.params_into(&vec![ ("action", "query"), ("prop", "extlinks"), ("ellimit", "500"), ("titles", title.as_str()), ]); let result = api .get_query_api_json_all(&params) .expect("query.extlinks failed"); let mut urls: Vec<String> = vec![]; result["query"]["pages"] .as_object() .expect("query.pages in result not an object") .iter() .for_each(|(_page_id, data)| { data["extlinks"] .as_array() .expect("extlinks not an array") .iter() .for_each(|x| urls.push(x["*"].as_str().expect("* not a string").to_string())); }); urls } fn get_paper_q(api: &Api, id: &GenericWorkIdentifier) -> Option<String> { let wdp = WikidataPapers::new(); match &id.work_type { GenericWorkType::Property(prop) => { let result = wdp.search_external_id(&prop, &id.id, api); result.get(0).map(|s| s.to_owned()) // First one will do } _ => None, } /* wdp.add_adapter(Box::new(Pubmed2Wikidata::new())); wdp.add_adapter(Box::new(Crossref2Wikidata::new())); wdp.add_adapter(Box::new(Semanticscholar2Wikidata::new())); wdp.add_adapter(Box::new(Orcid2Wikidata::new())); let ids = vec![id.to_owned()]; let ids = wdp.update_from_paper_ids(&ids); let q = ids .iter() .filter_map(|x| match x.work_type { GenericWorkType::Item => Some(x.id.to_owned()), _ => None, }) .next(); q*/ } fn command_refs(args: &Args) { if args.arg_wiki.is_empty() { panic!("wiki code (e.g. 'enwiki') is required"); } if args.arg_title.is_empty() { panic!("article title is required"); } let wiki = &args.arg_wiki; let title = &args.arg_title; // Get the API URL for the wiki let api_url = match get_api_url_for_wiki(&wiki) { Some(url) => url, None => panic!("Can't find API URL for {}", &wiki), }; // Get all external URLs from that page, on that wiki let urls = get_external_urls(&api_url, &title); //println!("{:#?}", &urls); lazy_static! { static ref RE_DOI: Regex = Regex::new(r#"^*.?//doi.org/(.+)$"#).unwrap(); static ref RE_PMID: Regex = Regex::new(r#"^*.?//www.ncbi.nlm.nih.gov/pubmed/(\d+)$"#).unwrap(); static ref RE_PMCID: Regex = Regex::new(r#"^*.?//www.ncbi.nlm.nih.gov/pmc/articles/PMC(\d+)$"#).unwrap(); } let mut ids: Vec<GenericWorkIdentifier> = vec![]; for url in urls { match RE_DOI.captures(&url) { Some(caps) => { let id = caps.get(1).unwrap().as_str(); match urlencoding::decode(&id) { Ok(id) => { ids.push(GenericWorkIdentifier::new_prop(PROP_DOI, &id)); } _ => {} } } None => {} } match RE_PMID.captures(&url) { Some(caps) => { let id = caps.get(1).unwrap().as_str(); ids.push(GenericWorkIdentifier::new_prop(PROP_PMID, id)); } None => {} } match RE_PMCID.captures(&url) { Some(caps) => { let id = caps.get(1).unwrap().as_str(); ids.push(GenericWorkIdentifier::new_prop(PROP_PMCID, id)); } None => {} } } let api = Api::new("https://www.wikidata
lines_from_file
identifier_name
medrs.rs
File::open(filename).expect(format!("no such file: {}", filename).as_str()); let buf = BufReader::new(file); buf.lines() .map(|l| l.expect("Could not parse line")) .collect() } fn read_file_to_string(filename: &str) -> String { let mut file = match File::open(filename) { Ok(file) => file, Err(_) => panic!("no such file"), }; let mut file_contents = String::new(); file.read_to_string(&mut file_contents) .ok() .expect("failed to read!"); file_contents } fn replace_sparql_placeolder(pattern: &str, sparql: &String, lines: &Vec<String>) -> String
fn output_sparql_result_items(sparql: &String) { let api = Api::new("https://www.wikidata.org/w/api.php").expect("Can't connect to Wikidata"); let result = api.sparql_query(&sparql).expect("SPARQL query failed"); let varname = result["head"]["vars"][0] .as_str() .expect("Can't find first variable name in SPARQL result"); let entities = api.entities_from_sparql_result(&result, &varname); println!("{}", entities.join("\n")); } /* fn get_all_from_stdin() -> String { let mut payload = Vec::new(); io::stdin().read_to_end(&mut payload).unwrap(); let s = match str::from_utf8(&payload) { Ok(v) => v, Err(e) => panic!("Invalid UTF-8 sequence: {}", e), }; s.to_string() } */ fn command_query(args: &Args) { if args.arg_query.is_empty() { println!("Requires SPARQL query"); } let sparql = &args.arg_query; output_sparql_result_items(&sparql); } fn command_run(args: &Args) { let articles = lines_from_file(&args.flag_articles); let reviews = lines_from_file(&args.flag_reviews); let topics = lines_from_file(&args.flag_topics); let journals = lines_from_file(&args.flag_journals); let publishers = lines_from_file(&args.flag_publishers); let mut sparql = read_file_to_string(&args.flag_sparql); sparql = replace_sparql_placeolder("%%ARTICLES%%", &sparql, &articles); sparql = replace_sparql_placeolder("%%REVIEWS%%", &sparql, &reviews); sparql = replace_sparql_placeolder("%%TOPICS%%", &sparql, &topics); sparql = replace_sparql_placeolder("%%JOURNALS%%", &sparql, &journals); sparql = replace_sparql_placeolder("%%PUBLISHERS%%", &sparql, &publishers); output_sparql_result_items(&sparql); } fn get_api_url_for_wiki(wiki: &String) -> Option<String> { // Get site matrix from wikidata let api = Api::new("https://www.wikidata.org/w/api.php").expect("Can't connect to Wikidata"); let params = api.params_into(&vec![("action", "sitematrix")]); let site_matrix = api .get_query_api_json(&params) .expect("Can't load sitematrix from wikidata API"); //println!("{:#?}", &site_matrix); // Go through the "normal" objects let mut ret: Option<String> = None; site_matrix["sitematrix"] .as_object() .expect("sitematrix is not an object") .iter() .for_each(|(_, data)| { match data["site"] .as_array() .unwrap_or(&vec![]) .iter() .filter_map(|x| { if x["dbname"].as_str().unwrap_or("") == wiki { x["url"].as_str() } else { None } }) .next() { Some(url) => { ret = Some(url.to_string() + "/w/api.php"); } None => {} } }); // Try the "specials" site_matrix["sitematrix"]["specials"] .as_array() .unwrap_or(&vec![]) .iter() .for_each(|x| { if x["dbname"].as_str().unwrap_or("") == wiki { ret = Some(x["url"].as_str().unwrap_or("").to_string() + "/w/api.php"); } }); ret } fn get_external_urls(api_url: &String, title: &String) -> Vec<String> { let api = Api::new(&api_url).expect(&format!("Can't connect to {}", &api_url)); let params = api.params_into(&vec![ ("action", "query"), ("prop", "extlinks"), ("ellimit", "500"), ("titles", title.as_str()), ]); let result = api .get_query_api_json_all(&params) .expect("query.extlinks failed"); let mut urls: Vec<String> = vec![]; result["query"]["pages"] .as_object() .expect("query.pages in result not an object") .iter() .for_each(|(_page_id, data)| { data["extlinks"] .as_array() .expect("extlinks not an array") .iter() .for_each(|x| urls.push(x["*"].as_str().expect("* not a string").to_string())); }); urls } fn get_paper_q(api: &Api, id: &GenericWorkIdentifier) -> Option<String> { let wdp = WikidataPapers::new(); match &id.work_type { GenericWorkType::Property(prop) => { let result = wdp.search_external_id(&prop, &id.id, api); result.get(0).map(|s| s.to_owned()) // First one will do } _ => None, } /* wdp.add_adapter(Box::new(Pubmed2Wikidata::new())); wdp.add_adapter(Box::new(Crossref2Wikidata::new())); wdp.add_adapter(Box::new(Semanticscholar2Wikidata::new())); wdp.add_adapter(Box::new(Orcid2Wikidata::new())); let ids = vec![id.to_owned()]; let ids = wdp.update_from_paper_ids(&ids); let q = ids .iter() .filter_map(|x| match x.work_type { GenericWorkType::Item => Some(x.id.to_owned()), _ => None, }) .next(); q*/ } fn command_refs(args: &Args) { if args.arg_wiki.is_empty() { panic!("wiki code (e.g. 'enwiki') is required"); } if args.arg_title.is_empty() { panic!("article title is required"); } let wiki = &args.arg_wiki; let title = &args.arg_title; // Get the API URL for the wiki let api_url = match get_api_url_for_wiki(&wiki) { Some(url) => url, None => panic!("Can't find API URL for {}", &wiki), }; // Get all external URLs from that page, on that wiki let urls = get_external_urls(&api_url, &title); //println!("{:#?}", &urls); lazy_static! { static ref RE_DOI: Regex = Regex::new(r#"^*.?//doi.org/(.+)$"#).unwrap(); static ref RE_PMID: Regex = Regex::new(r#"^*.?//www.ncbi.nlm.nih.gov/pubmed/(\d+)$"#).unwrap(); static ref RE_PMCID: Regex = Regex::new(r#"^*.?//www.ncbi.nlm.nih.gov/pmc/articles/PMC(\d+)$"#).unwrap(); } let mut ids: Vec<GenericWorkIdentifier> = vec![]; for url in urls { match RE_DOI.captures(&url) { Some(caps) => { let id = caps.get(1).unwrap().as_str(); match urlencoding::decode(&id) { Ok(id) => { ids.push(GenericWorkIdentifier::new_prop(PROP_DOI, &id)); } _ => {} } } None => {} } match RE_PMID.captures(&url) { Some(caps) => { let id = caps.get(1).unwrap().as_str(); ids.push(GenericWorkIdentifier::new_prop(PROP_PMID, id)); } None => {} } match RE_PMCID.captures(&url) { Some(caps) => { let id = caps.get(1).unwrap().as_str(); ids.push(GenericWorkIdentifier::new_prop(PROP_PMCID, id)); } None => {} } } let api = Api::new("https://www.wikidata.org/w/api.php").expect("Can't connect to Wikidata"); for id in ids { match get_paper_q(&api, &
{ let rep: String = if lines.is_empty() { "".to_string() } else { "wd:".to_string() + &lines.join(" wd:") }; sparql.replace(pattern, &rep) }
identifier_body
views.py
< time.time(): url = "https://api.weixin.qq.com/cgi-bin/ticket/getticket?type=jsapi&access_token=%s" % (self.getAccessToken()) response = requests.get(url) jsapi_ticket = json.loads(response.text)['ticket'] data['jsapi_ticket'] = jsapi_ticket data['expire_time'] = int(time.time()) + 7000 fopen = open('jsapi_ticket.json', 'w') fopen.write(json.dumps(data)) fopen.close() return jsapi_ticket def getAccessToken(self): data = json.loads(open(os.getcwd()+'/access_token.json').read()) access_token = data['access_token'] if data['expire_time'] < time.time(): url = "https://api.weixin.qq.com/cgi-bin/token?grant_type=client_credential&appid=%s&secret=%s" % (self.appId, self.appSecret) response = requests.get(url) access_token = json.loads(response.text)['access_token'] data['access_token'] = access_token data['expire_time'] = int(time.time()) + 7000 fopen = open('access_token.json', 'w') fopen.write(json.dumps(data)) fopen.close() return access_token def get_client_ip(request): x_forwarded_for = request.META.get('HTTP_X_FORWARDED_FOR') if x_forwarded_for: ip = x_forwarded_for.split(',')[0] else: ip = request.META.get('REMOTE_ADDR') return ip def add_corners(im, rad): circle = Image.new('L', (rad * 2, rad * 2), 0) draw = ImageDraw.Draw(circle) draw.ellipse((0, 0, rad * 2, rad * 2), fill=255) alpha = Image.new('L', im.size, 100) w, h = im.size alpha.paste(circle.crop((0, 0, rad, rad)), (0, 0)) alpha.paste(circle.crop((0, rad, rad, rad * 2)), (0, h - rad)) alpha.paste(circle.crop((rad, 0, rad * 2, rad)), (w - rad, 0)) alpha.paste(circle.crop((rad, rad, rad * 2, rad * 2)), (w - rad, h - rad)) im.putalpha(alpha) return im def guozuIndex(request): return render_to_response('guozu/index.html',{'random': random.randrange(0, 10)}) def guozuSaveImage(request): # if request.is_ajax(): # img = request.POST.get("img") # openid = request.POST.get("openid") # file = cStringIO.StringIO(urllib.urlopen(img).read()) # im = Image.open(file) # im.thumbnail((100, 100), Image.ANTIALIAS) # im.save("/home/project/project/template/guozu/upload/wx_"+openid+".png") return HttpResponse('1') def guozuImage(request): ba = Image.open("/Users/dongli/Documents/Outsourcing/project/project/template/guozu/images/logo.png") im = Image.open('/Users/dongli/Documents/Outsourcing/project/project/template/guozu/images/test.png') bg = Image.open('/Users/dongli/Documents/Outsourcing/project/project/template/guozu/images/img.png') ba.paste(im, (199,85)) ba.paste(bg, (199,85), mask=bg) draw = ImageDraw.Draw(ba) font = ImageFont.truetype("/Users/dongli/Documents/Outsourcing/project/project/template/guozu/font.ttf", 20) font1 = ImageFont.truetype("/Users/dongli/Documents/Outsourcing/project/project/template/guozu/font.ttf", 34) font2 = ImageFont.truetype("/Users/dongli/Documents/Outsourcing/project/project/template/guozu/font.ttf", 30) draw.text((194, 401),"123213",(255,240,8),font=font) draw.text((75, 230),unicode('只要血液一天是红色的,','utf-8'),(255,255,255),font=font1)
draw.text((110, 230),unicode('昆明,你能为我们','utf-8'),(255,255,255),font=font1) draw.text((90, 290),unicode('带来第一场胜利吗?!','utf-8'),(255,255,255),font=font1) draw.text((130, 230),unicode('莫愁长征无知己,','utf-8'),(255,255,255),font=font1) draw.text((60, 290),unicode('天涯海角永相随!国足必胜!','utf-8'),(255,255,255),font=font2) draw.text((55, 230),unicode('国足,我们真的不想再失望!','utf-8'),(255,255,255),font=font2) draw.text((90, 290),unicode('拿出斗志,拿下一场!','utf-8'),(255,255,255),font=font1) ba.save("/Users/dongli/Documents/Outsourcing/project/project/template/guozu/images/new.png") return HttpResponse(1) def guozuGetUserInfo(request): if request.is_ajax(): code = request.POST.get("code") r = requests.get('https://api.weixin.qq.com/sns/oauth2/access_token?appid=wx7dfb9e650ce540b1&secret=538b5277fe1d37660ed4b218f7057c86&code='+code+'&grant_type=authorization_code') access_token = json.loads(r.text)['access_token'] openid = json.loads(r.text)['openid'] q = requests.get('https://api.weixin.qq.com/sns/userinfo?access_token='+access_token+'&openid='+openid+'&lang=zh_CN ') q.encoding = 'utf-8' return HttpResponse(json.dumps({"nickname": json.loads(q.text)['nickname'], "openid": json.loads(q.text)['openid']})) def guozuSubmit(request): if request.is_ajax(): ip = get_client_ip(request) try: reader = geoip2.database.Reader('/Users/dongli/Documents/Outsourcing/project/project/template/guozu/GeoLite2-City.mmdb') r = reader.city(ip) city = r.city.names.get('zh-CN', '') except: city = u"北京" return HttpResponse(city) # img = request.POST.get('img') # word = request.POST.get('word') # name = request.POST.get('name') # openid = request.POST.get('openid') # p = GuozuRecord(openid=openid, name=name, city=city, image=img, word=word, date=datetime.datetime.now()) # p.save() # if GuozuCity.objects.filter(name=city).count(): # count = GuozuCity.objects.get(name=city).count # GuozuCity.objects.filter(name=city).update(count=count+1) # else: # p = GuozuCity(name=city, count=1) # p.save() # return HttpResponse(json.dumps({"city": city, "record": serializers.serialize('json', GuozuRecord.objects.filter(openid=openid).order_by('-id'))})) def guozuGetCode(request): if request.is_ajax(): phone = request.POST.get('phone') AppKey = '1a9c8ee74f8ba02ddfe5c934d83a4cb5'; AppSecret = '0f1e0d91b368'; p = ServerAPI(AppKey,AppSecret, UseSSLLib = True); a = p.sendSmsCode(phone,'') return HttpResponse(a['obj']); def guozuGetMsg(request): if request.is_ajax(): phone = request.POST.get('phone') count = GuozuCount.objects.get(id=1).count count = count + 1
draw.text((110, 290),unicode('心就一天是国足的!','utf-8'),(255,255,255),font=font1) draw.text((50, 230),unicode('赢了一起狂,输了一起扛!','utf-8'),(255,255,255),font=font1) draw.text((180, 290),unicode('国足雄起!','utf-8'),(255,255,255),font=font1)
random_line_split
views.py
.randrange(0, 10)}) def guozuSaveImage(request): # if request.is_ajax(): # img = request.POST.get("img") # openid = request.POST.get("openid") # file = cStringIO.StringIO(urllib.urlopen(img).read()) # im = Image.open(file) # im.thumbnail((100, 100), Image.ANTIALIAS) # im.save("/home/project/project/template/guozu/upload/wx_"+openid+".png") return HttpResponse('1') def guozuImage(request): ba = Image.open("/Users/dongli/Documents/Outsourcing/project/project/template/guozu/images/logo.png") im = Image.open('/Users/dongli/Documents/Outsourcing/project/project/template/guozu/images/test.png') bg = Image.open('/Users/dongli/Documents/Outsourcing/project/project/template/guozu/images/img.png') ba.paste(im, (199,85)) ba.paste(bg, (199,85), mask=bg) draw = ImageDraw.Draw(ba) font = ImageFont.truetype("/Users/dongli/Documents/Outsourcing/project/project/template/guozu/font.ttf", 20) font1 = ImageFont.truetype("/Users/dongli/Documents/Outsourcing/project/project/template/guozu/font.ttf", 34) font2 = ImageFont.truetype("/Users/dongli/Documents/Outsourcing/project/project/template/guozu/font.ttf", 30) draw.text((194, 401),"123213",(255,240,8),font=font) draw.text((75, 230),unicode('只要血液一天是红色的,','utf-8'),(255,255,255),font=font1) draw.text((110, 290),unicode('心就一天是国足的!','utf-8'),(255,255,255),font=font1) draw.text((50, 230),unicode('赢了一起狂,输了一起扛!','utf-8'),(255,255,255),font=font1) draw.text((180, 290),unicode('国足雄起!','utf-8'),(255,255,255),font=font1) draw.text((110, 230),unicode('昆明,你能为我们','utf-8'),(255,255,255),font=font1) draw.text((90, 290),unicode('带来第一场胜利吗?!','utf-8'),(255,255,255),font=font1) draw.text((130, 230),unicode('莫愁长征无知己,','utf-8'),(255,255,255),font=font1) draw.text((60, 290),unicode('天涯海角永相随!国足必胜!','utf-8'),(255,255,255),font=font2) draw.text((55, 230),unicode('国足,我们真的不想再失望!','utf-8'),(255,255,255),font=font2) draw.text((90, 290),unicode('拿出斗志,拿下一场!','utf-8'),(255,255,255),font=font1) ba.save("/Users/dongli/Documents/Outsourcing/project/project/template/guozu/images/new.png") return HttpResponse(1) def guozuGetUserInfo(request): if request.is_ajax(): code = request.POST.get("code") r = requests.get('https://api.weixin.qq.com/sns/oauth2/access_token?appid=wx7dfb9e650ce540b1&secret=538b5277fe1d37660ed4b218f7057c86&code='+code+'&grant_type=authorization_code') access_token = json.loads(r.text)['access_token'] openid = json.loads(r.text)['openid'] q = requests.get('https://api.weixin.qq.com/sns/userinfo?access_token='+access_token+'&openid='+openid+'&lang=zh_CN ') q.encoding = 'utf-8' return HttpResponse(json.dumps({"nickname": json.loads(q.text)['nickname'], "openid": json.loads(q.text)['openid']})) def guozuSubmit(request): if request.is_ajax(): ip = get_client_ip(request) try: reader = geoip2.database.Reader('/Users/dongli/Documents/Outsourcing/project/project/template/guozu/GeoLite2-City.mmdb') r = reader.city(ip) city = r.city.names.get('zh-CN', '') except: city = u"北京" return HttpResponse(city) # img = request.POST.get('img') # word = request.POST.get('word') # name = request.POST.get('name') # openid = request.POST.get('openid') # p = GuozuRecord(openid=openid, name=name, city=city, image=img, word=word, date=datetime.datetime.now()) # p.save() # if GuozuCity.objects.filter(name=city).count(): # count = GuozuCity.objects.get(name=city).count # GuozuCity.objects.filter(name=city).update(count=count+1) # else: # p = GuozuCity(name=city, count=1) # p.save() # return HttpResponse(json.dumps({"city": city, "record": serializers.serialize('json', GuozuRecord.objects.filter(openid=openid).order_by('-id'))})) def guozuGetCode(request): if request.is_ajax(): phone = request.POST.get('phone') AppKey = '1a9c8ee74f8ba02ddfe5c934d83a4cb5'; AppSecret = '0f1e0d91b368'; p = ServerAPI(AppKey,AppSecret, UseSSLLib = True); a = p.sendSmsCode(phone,'') return HttpResponse(a['obj']); def guozuGetMsg(request): if request.is_ajax(): phone = request.POST.get('phone') count = GuozuCount.objects.get(id=1).count count = count + 1 GuozuCount.objects.filter(id=1).update(count=count) number = GuozuList.objects.get(id=count).number AppKey = '1a9c8ee74f8ba02ddfe5c934d83a4cb5'; AppSecret = '0f1e0d91b368'; p = ServerAPI(AppKey,AppSecret, UseSSLLib = True); return HttpResponse(number); def guozuGetData(request): if request.is_ajax(): openid = request.POST.get('openid') serialized_obj1 = serializers.serialize('json', GuozuCity.objects.all().order_by('-count')) if GuozuRecord.objects.filter(openid=openid).count(): record = serializers.serialize('json', GuozuRecord.objects.filter(openid=openid).order_by('-id')) rtype = 1 else: record = serializers.serialize('json', GuozuRecord.objects.all().order_by('-id')[0:10]) rtype = 2 data = serializers.serialize('json', GuozuRecord.objects.all().order_by('-id')[0:10]) return HttpResponse(json.dumps({"city": serialized_obj1, "record": record, "data": data, "total": GuozuRecord.objects.all().count(), "type": rtype})) def guozuUpload(request): if request.method == 'POST': imageid = request.POST.get('openid') file_ext = request.FILES['file'].name.split('.')[1] user_upload_folder = '/Users/dongli/Documents/Outsourcing/project/project/template/guozu/upload/' file_upload = open( os.path.join(user_upload_folder, imageid+'.'+file_ext), 'w') file_upload.write(request.FILES['file'].read()) file_upload.close() im = Image.open(os.path.join(user_upload_folder, imageid+'.'+file_ext)) orientation = 274 try: exif = im._getexif() if exif: exif = dict(exif.items()) if exif[orientation] == 3: im = im.rotate(180, expand=True) elif exif[orientation] == 6: im = im.rotate(270, expand=True) elif exif[orientation] == 8: im = im.rotate(90, expand=True) except: pass im.thumb
nail((100, 100), Image.ANTIALIAS
conditional_block
views.py
time.time(): url = "https://api.weixin.qq.com/cgi-bin/ticket/getticket?type=jsapi&access_token=%s" % (self.getAccessToken()) response = requests.get(url) jsapi_ticket = json.loads(response.text)['ticket'] data['jsapi_ticket'] = jsapi_ticket data['expire_time'] = int(time.time()) + 7000 fopen = open('jsapi_ticket.json', 'w') fopen.write(json.dumps(data)) fopen.close() return jsapi_ticket def getAccessToken(self): data = json.loads(open(os.getcwd()+'/access_token.json').read()) access_token = data['access_token'] if data['expire_time'] < time.time(): url = "https://api.weixin.qq.com/cgi-bin/token?grant_type=client_credential&appid=%s&secret=%s" % (self.appId, self.appSecret) response = requests.get(url) access_token = json.loads(response.text)['access_token'] data['access_token'] = access_token data['expire_time'] = int(time.time()) + 7000 fopen = open('access_token.json', 'w') fopen.write(json.dumps(data)) fopen.close() return access_token def get_client_ip(request): x_forwarded_for = request.META.get('HTTP_X_FORWARDED_FOR') if x_forwarded_for: ip = x_forwarded_for.split(',')[0] else: ip = request.META.get('REMOTE_ADDR') return ip def add_corners(im, rad): circle = Image.new('L', (rad * 2, rad * 2), 0) draw = ImageDraw.Draw(circle) draw.ellipse((0, 0, rad * 2, rad * 2), fill=255) alpha = Image.new('L', im.size, 100) w, h = im.size alpha.paste(circle.crop((0, 0, rad, rad)), (0, 0)) alpha.paste(circle.crop((0, rad, rad, rad * 2)), (0, h - rad)) alpha.paste(circle.crop((rad, 0, rad * 2, rad)), (w - rad, 0)) alpha.paste(circle.crop((rad, rad, rad * 2, rad * 2)), (w - rad, h - rad)) im.putalpha(alpha) return im def guozuIndex(request): return render_to_response('guozu/index.html',{'random': random.randrange(0, 10)}) def guozuSaveImage(request): # if request.is_ajax(): # img = request.POST.get("img") # openid = request.POST.get("openid") # file = cStringIO.StringIO(urllib.urlopen(img).read()) # im = Image.open(file) # im.thumbnail((100, 100), Image.ANTIALIAS) # im.save("/home/project/project/template/guozu/upload/wx_"+openid+".png")
def guozuImage(request): ba = Image.open("/Users/dongli/Documents/Outsourcing/project/project/template/guozu/images/logo.png") im = Image.open('/Users/dongli/Documents/Outsourcing/project/project/template/guozu/images/test.png') bg = Image.open('/Users/dongli/Documents/Outsourcing/project/project/template/guozu/images/img.png') ba.paste(im, (199,85)) ba.paste(bg, (199,85), mask=bg) draw = ImageDraw.Draw(ba) font = ImageFont.truetype("/Users/dongli/Documents/Outsourcing/project/project/template/guozu/font.ttf", 20) font1 = ImageFont.truetype("/Users/dongli/Documents/Outsourcing/project/project/template/guozu/font.ttf", 34) font2 = ImageFont.truetype("/Users/dongli/Documents/Outsourcing/project/project/template/guozu/font.ttf", 30) draw.text((194, 401),"123213",(255,240,8),font=font) draw.text((75, 230),unicode('只要血液一天是红色的,','utf-8'),(255,255,255),font=font1) draw.text((110, 290),unicode('心就一天是国足的!','utf-8'),(255,255,255),font=font1) draw.text((50, 230),unicode('赢了一起狂,输了一起扛!','utf-8'),(255,255,255),font=font1) draw.text((180, 290),unicode('国足雄起!','utf-8'),(255,255,255),font=font1) draw.text((110, 230),unicode('昆明,你能为我们','utf-8'),(255,255,255),font=font1) draw.text((90, 290),unicode('带来第一场胜利吗?!','utf-8'),(255,255,255),font=font1) draw.text((130, 230),unicode('莫愁长征无知己,','utf-8'),(255,255,255),font=font1) draw.text((60, 290),unicode('天涯海角永相随!国足必胜!','utf-8'),(255,255,255),font=font2) draw.text((55, 230),unicode('国足,我们真的不想再失望!','utf-8'),(255,255,255),font=font2) draw.text((90, 290),unicode('拿出斗志,拿下一场!','utf-8'),(255,255,255),font=font1) ba.save("/Users/dongli/Documents/Outsourcing/project/project/template/guozu/images/new.png") return HttpResponse(1) def guozuGetUserInfo(request): if request.is_ajax(): code = request.POST.get("code") r = requests.get('https://api.weixin.qq.com/sns/oauth2/access_token?appid=wx7dfb9e650ce540b1&secret=538b5277fe1d37660ed4b218f7057c86&code='+code+'&grant_type=authorization_code') access_token = json.loads(r.text)['access_token'] openid = json.loads(r.text)['openid'] q = requests.get('https://api.weixin.qq.com/sns/userinfo?access_token='+access_token+'&openid='+openid+'&lang=zh_CN ') q.encoding = 'utf-8' return HttpResponse(json.dumps({"nickname": json.loads(q.text)['nickname'], "openid": json.loads(q.text)['openid']})) def guozuSubmit(request): if request.is_ajax(): ip = get_client_ip(request) try: reader = geoip2.database.Reader('/Users/dongli/Documents/Outsourcing/project/project/template/guozu/GeoLite2-City.mmdb') r = reader.city(ip) city = r.city.names.get('zh-CN', '') except: city = u"北京" return HttpResponse(city) # img = request.POST.get('img') # word = request.POST.get('word') # name = request.POST.get('name') # openid = request.POST.get('openid') # p = GuozuRecord(openid=openid, name=name, city=city, image=img, word=word, date=datetime.datetime.now()) # p.save() # if GuozuCity.objects.filter(name=city).count(): # count = GuozuCity.objects.get(name=city).count # GuozuCity.objects.filter(name=city).update(count=count+1) # else: # p = GuozuCity(name=city, count=1) # p.save() # return HttpResponse(json.dumps({"city": city, "record": serializers.serialize('json', GuozuRecord.objects.filter(openid=openid).order_by('-id'))})) def guozuGetCode(request): if request.is_ajax(): phone = request.POST.get('phone') AppKey = '1a9c8ee74f8ba02ddfe5c934d83a4cb5'; AppSecret = '0f1e0d91b368'; p = ServerAPI(AppKey,AppSecret, UseSSLLib = True); a = p.sendSmsCode(phone,'') return HttpResponse(a['obj']); def guozuGetMsg(request): if request.is_ajax(): phone = request.POST.get('phone') count = GuozuCount.objects.get(id=1).count count = count + 1
return HttpResponse('1')
identifier_body
views.py
= "https://api.weixin.qq.com/cgi-bin/token?grant_type=client_credential&appid=%s&secret=%s" % (self.appId, self.appSecret) response = requests.get(url) access_token = json.loads(response.text)['access_token'] data['access_token'] = access_token data['expire_time'] = int(time.time()) + 7000 fopen = open('access_token.json', 'w') fopen.write(json.dumps(data)) fopen.close() return access_token def get_client_ip(request): x_forwarded_for = request.META.get('HTTP_X_FORWARDED_FOR') if x_forwarded_for: ip = x_forwarded_for.split(',')[0] else: ip = request.META.get('REMOTE_ADDR') return ip def add_corners(im, rad): circle = Image.new('L', (rad * 2, rad * 2), 0) draw = ImageDraw.Draw(circle) draw.ellipse((0, 0, rad * 2, rad * 2), fill=255) alpha = Image.new('L', im.size, 100) w, h = im.size alpha.paste(circle.crop((0, 0, rad, rad)), (0, 0)) alpha.paste(circle.crop((0, rad, rad, rad * 2)), (0, h - rad)) alpha.paste(circle.crop((rad, 0, rad * 2, rad)), (w - rad, 0)) alpha.paste(circle.crop((rad, rad, rad * 2, rad * 2)), (w - rad, h - rad)) im.putalpha(alpha) return im def guozuIndex(request): return render_to_response('guozu/index.html',{'random': random.randrange(0, 10)}) def guozuSaveImage(request): # if request.is_ajax(): # img = request.POST.get("img") # openid = request.POST.get("openid") # file = cStringIO.StringIO(urllib.urlopen(img).read()) # im = Image.open(file) # im.thumbnail((100, 100), Image.ANTIALIAS) # im.save("/home/project/project/template/guozu/upload/wx_"+openid+".png") return HttpResponse('1') def guozuImage(request): ba = Image.open("/Users/dongli/Documents/Outsourcing/project/project/template/guozu/images/logo.png") im = Image.open('/Users/dongli/Documents/Outsourcing/project/project/template/guozu/images/test.png') bg = Image.open('/Users/dongli/Documents/Outsourcing/project/project/template/guozu/images/img.png') ba.paste(im, (199,85)) ba.paste(bg, (199,85), mask=bg) draw = ImageDraw.Draw(ba) font = ImageFont.truetype("/Users/dongli/Documents/Outsourcing/project/project/template/guozu/font.ttf", 20) font1 = ImageFont.truetype("/Users/dongli/Documents/Outsourcing/project/project/template/guozu/font.ttf", 34) font2 = ImageFont.truetype("/Users/dongli/Documents/Outsourcing/project/project/template/guozu/font.ttf", 30) draw.text((194, 401),"123213",(255,240,8),font=font) draw.text((75, 230),unicode('只要血液一天是红色的,','utf-8'),(255,255,255),font=font1) draw.text((110, 290),unicode('心就一天是国足的!','utf-8'),(255,255,255),font=font1) draw.text((50, 230),unicode('赢了一起狂,输了一起扛!','utf-8'),(255,255,255),font=font1) draw.text((180, 290),unicode('国足雄起!','utf-8'),(255,255,255),font=font1) draw.text((110, 230),unicode('昆明,你能为我们','utf-8'),(255,255,255),font=font1) draw.text((90, 290),unicode('带来第一场胜利吗?!','utf-8'),(255,255,255),font=font1) draw.text((130, 230),unicode('莫愁长征无知己,','utf-8'),(255,255,255),font=font1) draw.text((60, 290),unicode('天涯海角永相随!国足必胜!','utf-8'),(255,255,255),font=font2) draw.text((55, 230),unicode('国足,我们真的不想再失望!','utf-8'),(255,255,255),font=font2) draw.text((90, 290),unicode('拿出斗志,拿下一场!','utf-8'),(255,255,255),font=font1) ba.save("/Users/dongli/Documents/Outsourcing/project/project/template/guozu/images/new.png") return HttpResponse(1) def guozuGetUserInfo(request): if request.is_ajax(): code = request.POST.get("code") r = requests.get('https://api.weixin.qq.com/sns/oauth2/access_token?appid=wx7dfb9e650ce540b1&secret=538b5277fe1d37660ed4b218f7057c86&code='+code+'&grant_type=authorization_code') access_token = json.loads(r.text)['access_token'] openid = json.loads(r.text)['openid'] q = requests.get('https://api.weixin.qq.com/sns/userinfo?access_token='+access_token+'&openid='+openid+'&lang=zh_CN ') q.encoding = 'utf-8' return HttpResponse(json.dumps({"nickname": json.loads(q.text)['nickname'], "openid": json.loads(q.text)['openid']})) def guozuSubmit(request): if request.is_ajax(): ip = get_client_ip(request) try: reader = geoip2.database.Reader('/Users/dongli/Documents/Outsourcing/project/project/template/guozu/GeoLite2-City.mmdb') r = reader.city(ip) city = r.city.names.get('zh-CN', '') except: city = u"北京" return HttpResponse(city) # img = request.POST.get('img') # word = request.POST.get('word') # name = request.POST.get('name') # openid = request.POST.get('openid') # p = GuozuRecord(openid=openid, name=name, city=city, image=img, word=word, date=datetime.datetime.now()) # p.save() # if GuozuCity.objects.filter(name=city).count(): # count = GuozuCity.objects.get(name=city).count # GuozuCity.objects.filter(name=city).update(count=count+1) # else: # p = GuozuCity(name=city, count=1) # p.save() # return HttpResponse(json.dumps({"city": city, "record": serializers.serialize('json', GuozuRecord.objects.filter(openid=openid).order_by('-id'))})) def guozuGetCode(request): if request.is_ajax(): phone = request.POST.get('phone') AppKey = '1a9c8ee74f8ba02ddfe5c934d83a4cb5'; AppSecret = '0f1e0d91b368'; p = ServerAPI(AppKey,AppSecret, UseSSLLib = True); a = p.sendSmsCode(phone,'') return HttpResponse(a['obj']); def guozuGetMsg(request): if request.is_ajax(): phone = request.POST.get('phone') count = GuozuCount.objects.get(id=1).count count = count + 1 GuozuCount.objects.filter(id=1).update(count=count) number = GuozuList.objects.get(id=count).number AppKey = '1a9c8ee74f8ba02ddfe5c934d83a4cb5'; AppSecret = '0f1e0d91b368'; p = ServerAPI(AppKey,AppSecret, UseSSLLib = True); return HttpResponse(number); def guozuGetData(request): if request.is_ajax(): openid = request.POST.get('openid') serialized_obj1 = serializers.serialize('json', GuozuCity.objects.all().order_by('-count')) if GuozuRecord.objects
.filter(open
identifier_name
host_window.py
NavigationToolbar import pylab def _configure_device(): """ Configure and get the USB device running. Returns device class if success and None if failed""" vendor_id = 0x04D8 # These ids are microchip's libusb based device product_id = 0x0204 # ids dev = usb.core.find(idVendor=vendor_id, idProduct = product_id) try: dev.set_configuration() return dev except: return None class VSDataAquisition(object): """ A place holder for collecting data from the ADC of the device. This class will also control sample/hold bit of the device.""" def __init__(self): """ Configure the device and set class properties""" self.data0 = [] # This will hold data from ADC0 self.data1 = [] # This will hold data from ADC1 self.dev = _configure_device() def get_data(self): """ Get the next data from ADC0. For ADC1, use get_dc_offset()""" self.dev.write(1, 'A0') digit1, digit2 = self.dev.read(0x81, 64)[:2] # Save the data as voltage between 0.0 and 5.0 self.data0.append((digit1 + 256*digit2)*5.0/1024) def get_dc_offset(self): """ Get the initial DC offset of the analog output""" self.dev.write(1, 'A1') digit1, digit2 = self.dev.read(0x81, 64)[:2] # Save the data as voltage between 0.0 and 5.0 self.data1.append((digit1 + 256*digit2)*5.0/1024) def sample(self): """ Set the sample bit for getting intial DC offset""" self.dev.write(1, 'S') def hold(self): """ Clear the sample bit for consecutive data aquisition""" self.dev.write(1, 'H') class VSControlBox(wx.Panel): """ A static box for controlling the start and stop of the device and displaying the final result of the venous flow measurement.""" def __init__(self, parent, ID, label): wx.Panel.__init__(self, parent, ID) # Create two box sizers, one for the button and one for the status # message and final output reading. button_box = wx.StaticBox(self, -1, label = 'Device Control') info_box = wx.StaticBox(self, -1, label = 'Information') box = wx.StaticBox(self, -1, label) main_sizer = wx.StaticBoxSizer(box, wx.HORIZONTAL) button_sizer = wx.StaticBoxSizer(button_box, orient=wx.HORIZONTAL) info_sizer = wx.StaticBoxSizer(info_box, orient=wx.HORIZONTAL) # Create a start/stop measurement button. self.start_button = wx.Button(self, label = 'Start measurement',) # Create a result and information text box. self.result_box = wx.StaticText(self) self.txt_info_box = wx.StaticText(self, size=(200, -1)) self.result_box.SetLabel("0.00")
info_sizer.Add(self.txt_info_box, 0, wx.ALL, 10) # Add the sizers to main sizer main_sizer.Add(button_sizer, flag=wx.ALIGN_CENTER_VERTICAL) main_sizer.AddSpacer(20) main_sizer.Add(info_sizer, flag=wx.ALIGN_CENTER_VERTICAL|wx.EXPAND) # Bind events to the button self.start_button.Bind(wx.EVT_BUTTON, parent.Parent.start_stop) # Finally, make a fit self.SetSizer(main_sizer) main_sizer.Fit(self) def start_stop(self, event): """ Bind a rudimentary event now. Will update it later.""" self.start_button.SetLabel('Measuring') self.start_button.Enable = False # Do nothing as of now. Will call measuring functions later. self.txt_info_box.SetLabel('Starting measurement.') time.sleep(2) self.start_button.SetLabel('Start measurement') self.start_button.Enable = True self.txt_info_box.SetLabel('Completed measurement.') self.result_box.SetLabel("100.00") class VSGraphFrame(wx.Frame): """ Main frame for the measurement application""" title = "Venous flow calculation" def __init__(self): wx.Frame.__init__(self, None, -1, self.title) self.SAMPLING_TIME = 10000.0 # Set the sampling time here. self.daq = VSDataAquisition() if self.daq == None: help_string = ''' Device is not connected. Please connect the device and restart the software''' wx.MessageBox(help_string, 'Device not found', wx.OK | wx.ICON_INFORMATION) self.Destroy() return self.create_menu() # Create the menu self.create_status_bar () # Add a status bar. Could use for debugging self.create_main_panel() # The main panel # We will use a timer for getting our samples. self.redraw_timer = wx.Timer(self, wx.ID_ANY) # Sampling duration itself is another timer. This timer is a # oneshot timer and runs for SAMPLING_TIME time. self.sampling_timer = wx.Timer(self, wx.ID_ANY) self.Bind(wx.EVT_TIMER, self.on_redraw_timer, self.redraw_timer) self.Bind(wx.EVT_TIMER, self.on_sampling_timer, self.sampling_timer) self.redraw_timer.Start(100) def create_menu(self): """ Add menu bar items. One File and one About""" self.menubar = wx.MenuBar() menu_file = wx.Menu() menu_help = wx.Menu() # Add save and exit to File menu menu_save = menu_file.Append(-1, '&Save plot\tCtrl-S', 'Save plot to a file') menu_file.AppendSeparator() menu_exit = menu_file.Append(-1, '&Exit\tCtrl-X', 'Exit the program') self.Bind(wx.EVT_MENU, self.on_save, menu_save) self.Bind(wx.EVT_MENU, self.on_exit, menu_exit) # Add an about in the Help menu. Will update later. help_about = menu_help.Append(-1, '&About', 'About the program') menu_help.AppendSeparator() self.Bind(wx.EVT_MENU, self.on_about, help_about) # Add them both to menubar self.menubar.Append(menu_file, '&File') self.menubar.Append(menu_help, '&Help') self.SetMenuBar(self.menubar) def create_main_panel(self): """ Create the main panel which will show the dynamic plot.""" self.panel = wx.Panel(self) self.init_plot() self.canvas = FigCanvas(self.panel, -1, self.fig) self.control_box = VSControlBox(self.panel, -1, 'Information board') self.vbox = wx.BoxSizer(wx.VERTICAL) self.vbox.Add(self.canvas, 1, wx.LEFT | wx.TOP | wx.GROW) self.vbox.Add(self.control_box, 0, wx.ALIGN_LEFT | wx.TOP | wx.EXPAND) self.panel.SetSizer(self.vbox) self.vbox.Fit(self) def create_status_bar(self): """ Create a status bar. Will use it in future for debugging""" self.statusbar = self.CreateStatusBar() def init_plot(self): """ Initialize the plot canvas""" self.dpi = 100 self.fig = Figure((5.0, 5.0), dpi = self.dpi) self.main_plot = self.fig.add_subplot(111) self.main_plot.set_axis_bgcolor('black') self.main_plot.set_title('Dynamic venous flow view', size = 12) pylab.setp(self.main_plot.get_xticklabels(), fontsize = 8) pylab.setp(self.main_plot.get_yticklabels(), fontsize = 8) # Plot the data as a green line self.plot_data = self.main_plot.plot( self.daq.data0, linewidth = 1, color = (0, 1, 0), )[0] self.main_plot.grid(True, color='gray') def draw_plot(self): """ Redraw the plot after every data aquisition.""" # X axis is auto follow. XLEN = 100 xmax = max(len(self.daq.data0), XLEN) xmin = xmax - XLEN # The Y value will lie between 0.0 and 5.0 volts ymax = 5.0 ymin = 0.0 self.main_plot.set_xbound(lower=xmin, upper=xmax) self.main_plot.set_ybound(lower=ymin, upper=ymax) # Add the grid. Grid looks cool and is actually very helpful. self.main_plot.grid(True, color='gray') pylab.setp(self.main_plot.get_xticklabels(), visible=True) self.plot_data.set
# Add the items to sizers button_sizer.Add(self.start_button, 0, wx.ALL, 10) info_sizer.Add(self.result_box, 0, wx.ALL, 10)
random_line_split
host_window.py
digit1, digit2 = self.dev.read(0x81, 64)[:2] # Save the data as voltage between 0.0 and 5.0 self.data1.append((digit1 + 256*digit2)*5.0/1024) def sample(self): """ Set the sample bit for getting intial DC offset""" self.dev.write(1, 'S') def hold(self): """ Clear the sample bit for consecutive data aquisition""" self.dev.write(1, 'H') class VSControlBox(wx.Panel): """ A static box for controlling the start and stop of the device and displaying the final result of the venous flow measurement.""" def __init__(self, parent, ID, label): wx.Panel.__init__(self, parent, ID) # Create two box sizers, one for the button and one for the status # message and final output reading. button_box = wx.StaticBox(self, -1, label = 'Device Control') info_box = wx.StaticBox(self, -1, label = 'Information') box = wx.StaticBox(self, -1, label) main_sizer = wx.StaticBoxSizer(box, wx.HORIZONTAL) button_sizer = wx.StaticBoxSizer(button_box, orient=wx.HORIZONTAL) info_sizer = wx.StaticBoxSizer(info_box, orient=wx.HORIZONTAL) # Create a start/stop measurement button. self.start_button = wx.Button(self, label = 'Start measurement',) # Create a result and information text box. self.result_box = wx.StaticText(self) self.txt_info_box = wx.StaticText(self, size=(200, -1)) self.result_box.SetLabel("0.00") # Add the items to sizers button_sizer.Add(self.start_button, 0, wx.ALL, 10) info_sizer.Add(self.result_box, 0, wx.ALL, 10) info_sizer.Add(self.txt_info_box, 0, wx.ALL, 10) # Add the sizers to main sizer main_sizer.Add(button_sizer, flag=wx.ALIGN_CENTER_VERTICAL) main_sizer.AddSpacer(20) main_sizer.Add(info_sizer, flag=wx.ALIGN_CENTER_VERTICAL|wx.EXPAND) # Bind events to the button self.start_button.Bind(wx.EVT_BUTTON, parent.Parent.start_stop) # Finally, make a fit self.SetSizer(main_sizer) main_sizer.Fit(self) def start_stop(self, event): """ Bind a rudimentary event now. Will update it later.""" self.start_button.SetLabel('Measuring') self.start_button.Enable = False # Do nothing as of now. Will call measuring functions later. self.txt_info_box.SetLabel('Starting measurement.') time.sleep(2) self.start_button.SetLabel('Start measurement') self.start_button.Enable = True self.txt_info_box.SetLabel('Completed measurement.') self.result_box.SetLabel("100.00") class VSGraphFrame(wx.Frame): """ Main frame for the measurement application""" title = "Venous flow calculation" def __init__(self): wx.Frame.__init__(self, None, -1, self.title) self.SAMPLING_TIME = 10000.0 # Set the sampling time here. self.daq = VSDataAquisition() if self.daq == None: help_string = ''' Device is not connected. Please connect the device and restart the software''' wx.MessageBox(help_string, 'Device not found', wx.OK | wx.ICON_INFORMATION) self.Destroy() return self.create_menu() # Create the menu self.create_status_bar () # Add a status bar. Could use for debugging self.create_main_panel() # The main panel # We will use a timer for getting our samples. self.redraw_timer = wx.Timer(self, wx.ID_ANY) # Sampling duration itself is another timer. This timer is a # oneshot timer and runs for SAMPLING_TIME time. self.sampling_timer = wx.Timer(self, wx.ID_ANY) self.Bind(wx.EVT_TIMER, self.on_redraw_timer, self.redraw_timer) self.Bind(wx.EVT_TIMER, self.on_sampling_timer, self.sampling_timer) self.redraw_timer.Start(100) def create_menu(self): """ Add menu bar items. One File and one About""" self.menubar = wx.MenuBar() menu_file = wx.Menu() menu_help = wx.Menu() # Add save and exit to File menu menu_save = menu_file.Append(-1, '&Save plot\tCtrl-S', 'Save plot to a file') menu_file.AppendSeparator() menu_exit = menu_file.Append(-1, '&Exit\tCtrl-X', 'Exit the program') self.Bind(wx.EVT_MENU, self.on_save, menu_save) self.Bind(wx.EVT_MENU, self.on_exit, menu_exit) # Add an about in the Help menu. Will update later. help_about = menu_help.Append(-1, '&About', 'About the program') menu_help.AppendSeparator() self.Bind(wx.EVT_MENU, self.on_about, help_about) # Add them both to menubar self.menubar.Append(menu_file, '&File') self.menubar.Append(menu_help, '&Help') self.SetMenuBar(self.menubar) def create_main_panel(self): """ Create the main panel which will show the dynamic plot.""" self.panel = wx.Panel(self) self.init_plot() self.canvas = FigCanvas(self.panel, -1, self.fig) self.control_box = VSControlBox(self.panel, -1, 'Information board') self.vbox = wx.BoxSizer(wx.VERTICAL) self.vbox.Add(self.canvas, 1, wx.LEFT | wx.TOP | wx.GROW) self.vbox.Add(self.control_box, 0, wx.ALIGN_LEFT | wx.TOP | wx.EXPAND) self.panel.SetSizer(self.vbox) self.vbox.Fit(self) def create_status_bar(self): """ Create a status bar. Will use it in future for debugging""" self.statusbar = self.CreateStatusBar() def init_plot(self): """ Initialize the plot canvas""" self.dpi = 100 self.fig = Figure((5.0, 5.0), dpi = self.dpi) self.main_plot = self.fig.add_subplot(111) self.main_plot.set_axis_bgcolor('black') self.main_plot.set_title('Dynamic venous flow view', size = 12) pylab.setp(self.main_plot.get_xticklabels(), fontsize = 8) pylab.setp(self.main_plot.get_yticklabels(), fontsize = 8) # Plot the data as a green line self.plot_data = self.main_plot.plot( self.daq.data0, linewidth = 1, color = (0, 1, 0), )[0] self.main_plot.grid(True, color='gray') def draw_plot(self): """ Redraw the plot after every data aquisition.""" # X axis is auto follow. XLEN = 100 xmax = max(len(self.daq.data0), XLEN) xmin = xmax - XLEN # The Y value will lie between 0.0 and 5.0 volts ymax = 5.0 ymin = 0.0 self.main_plot.set_xbound(lower=xmin, upper=xmax) self.main_plot.set_ybound(lower=ymin, upper=ymax) # Add the grid. Grid looks cool and is actually very helpful. self.main_plot.grid(True, color='gray') pylab.setp(self.main_plot.get_xticklabels(), visible=True) self.plot_data.set_xdata(arange(len(self.daq.data0))) self.plot_data.set_ydata(array(self.daq.data0)) self.canvas.draw() def on_save(self, event): """ Method for saving a plot """ file_choices = "PNG (*.png)|*.png" dlg = wx.FileDialog( self, message="Save plot as...", defaultDir=os.getcwd(), defaultFile="plot.png", wildcard=file_choices, style=wx.SAVE) if dlg.ShowModal() == wx.ID_OK: path = dlg.GetPath() self.canvas.print_figure(path, dpi=self.dpi) self.flash_status_message("Saved to %s" % path) def start_stop(self, event): """ Restart measurements and complete calculations""" self.daq.data0= [] self.control_box.txt_info_box.SetLabel('Starting measurement') self.sampling_timer.Start(self.SAMPLING_TIME, oneShot=True) def on_redraw_timer(self, event): """ Update the plot whenever data is obtained """ if self.sampling_timer.IsRunning(): self.daq.get_data() self.draw_plot() else: self.control_box.txt_info_box.SetLabel('Measurement complete') self.calculate() return def on_sampling_timer(self, event): """ Stop the timer when sampling is complete.""" self.sampling_timer.Stop() def calculate(self): """Calculate the venous flow rate. Dummy now.""" if self.sampling_timer.IsRunning(): return if self.daq.data0 == []:
average = 0.0
conditional_block
host_window.py
NavigationToolbar import pylab def _configure_device(): """ Configure and get the USB device running. Returns device class if success and None if failed""" vendor_id = 0x04D8 # These ids are microchip's libusb based device product_id = 0x0204 # ids dev = usb.core.find(idVendor=vendor_id, idProduct = product_id) try: dev.set_configuration() return dev except: return None class VSDataAquisition(object): """ A place holder for collecting data from the ADC of the device. This class will also control sample/hold bit of the device.""" def __init__(self): """ Configure the device and set class properties""" self.data0 = [] # This will hold data from ADC0 self.data1 = [] # This will hold data from ADC1 self.dev = _configure_device() def get_data(self): """ Get the next data from ADC0. For ADC1, use get_dc_offset()""" self.dev.write(1, 'A0') digit1, digit2 = self.dev.read(0x81, 64)[:2] # Save the data as voltage between 0.0 and 5.0 self.data0.append((digit1 + 256*digit2)*5.0/1024) def get_dc_offset(self): """ Get the initial DC offset of the analog output""" self.dev.write(1, 'A1') digit1, digit2 = self.dev.read(0x81, 64)[:2] # Save the data as voltage between 0.0 and 5.0 self.data1.append((digit1 + 256*digit2)*5.0/1024) def sample(self): """ Set the sample bit for getting intial DC offset""" self.dev.write(1, 'S') def hold(self): """ Clear the sample bit for consecutive data aquisition""" self.dev.write(1, 'H') class VSControlBox(wx.Panel):
# Add the items to sizers button_sizer.Add(self.start_button, 0, wx.ALL, 10) info_sizer.Add(self.result_box, 0, wx.ALL, 10) info_sizer.Add(self.txt_info_box, 0, wx.ALL, 10) # Add the sizers to main sizer main_sizer.Add(button_sizer, flag=wx.ALIGN_CENTER_VERTICAL) main_sizer.AddSpacer(20) main_sizer.Add(info_sizer, flag=wx.ALIGN_CENTER_VERTICAL|wx.EXPAND) # Bind events to the button self.start_button.Bind(wx.EVT_BUTTON, parent.Parent.start_stop) # Finally, make a fit self.SetSizer(main_sizer) main_sizer.Fit(self) def start_stop(self, event): """ Bind a rudimentary event now. Will update it later.""" self.start_button.SetLabel('Measuring') self.start_button.Enable = False # Do nothing as of now. Will call measuring functions later. self.txt_info_box.SetLabel('Starting measurement.') time.sleep(2) self.start_button.SetLabel('Start measurement') self.start_button.Enable = True self.txt_info_box.SetLabel('Completed measurement.') self.result_box.SetLabel("100.00") class VSGraphFrame(wx.Frame): """ Main frame for the measurement application""" title = "Venous flow calculation" def __init__(self): wx.Frame.__init__(self, None, -1, self.title) self.SAMPLING_TIME = 10000.0 # Set the sampling time here. self.daq = VSDataAquisition() if self.daq == None: help_string = ''' Device is not connected. Please connect the device and restart the software''' wx.MessageBox(help_string, 'Device not found', wx.OK | wx.ICON_INFORMATION) self.Destroy() return self.create_menu() # Create the menu self.create_status_bar () # Add a status bar. Could use for debugging self.create_main_panel() # The main panel # We will use a timer for getting our samples. self.redraw_timer = wx.Timer(self, wx.ID_ANY) # Sampling duration itself is another timer. This timer is a # oneshot timer and runs for SAMPLING_TIME time. self.sampling_timer = wx.Timer(self, wx.ID_ANY) self.Bind(wx.EVT_TIMER, self.on_redraw_timer, self.redraw_timer) self.Bind(wx.EVT_TIMER, self.on_sampling_timer, self.sampling_timer) self.redraw_timer.Start(100) def create_menu(self): """ Add menu bar items. One File and one About""" self.menubar = wx.MenuBar() menu_file = wx.Menu() menu_help = wx.Menu() # Add save and exit to File menu menu_save = menu_file.Append(-1, '&Save plot\tCtrl-S', 'Save plot to a file') menu_file.AppendSeparator() menu_exit = menu_file.Append(-1, '&Exit\tCtrl-X', 'Exit the program') self.Bind(wx.EVT_MENU, self.on_save, menu_save) self.Bind(wx.EVT_MENU, self.on_exit, menu_exit) # Add an about in the Help menu. Will update later. help_about = menu_help.Append(-1, '&About', 'About the program') menu_help.AppendSeparator() self.Bind(wx.EVT_MENU, self.on_about, help_about) # Add them both to menubar self.menubar.Append(menu_file, '&File') self.menubar.Append(menu_help, '&Help') self.SetMenuBar(self.menubar) def create_main_panel(self): """ Create the main panel which will show the dynamic plot.""" self.panel = wx.Panel(self) self.init_plot() self.canvas = FigCanvas(self.panel, -1, self.fig) self.control_box = VSControlBox(self.panel, -1, 'Information board') self.vbox = wx.BoxSizer(wx.VERTICAL) self.vbox.Add(self.canvas, 1, wx.LEFT | wx.TOP | wx.GROW) self.vbox.Add(self.control_box, 0, wx.ALIGN_LEFT | wx.TOP | wx.EXPAND) self.panel.SetSizer(self.vbox) self.vbox.Fit(self) def create_status_bar(self): """ Create a status bar. Will use it in future for debugging""" self.statusbar = self.CreateStatusBar() def init_plot(self): """ Initialize the plot canvas""" self.dpi = 100 self.fig = Figure((5.0, 5.0), dpi = self.dpi) self.main_plot = self.fig.add_subplot(111) self.main_plot.set_axis_bgcolor('black') self.main_plot.set_title('Dynamic venous flow view', size = 12) pylab.setp(self.main_plot.get_xticklabels(), fontsize = 8) pylab.setp(self.main_plot.get_yticklabels(), fontsize = 8) # Plot the data as a green line self.plot_data = self.main_plot.plot( self.daq.data0, linewidth = 1, color = (0, 1, 0), )[0] self.main_plot.grid(True, color='gray') def draw_plot(self): """ Redraw the plot after every data aquisition.""" # X axis is auto follow. XLEN = 100 xmax = max(len(self.daq.data0), XLEN) xmin = xmax - XLEN # The Y value will lie between 0.0 and 5.0 volts ymax = 5.0 ymin = 0.0 self.main_plot.set_xbound(lower=xmin, upper=xmax) self.main_plot.set_ybound(lower=ymin, upper=ymax) # Add the grid. Grid looks cool and is actually very helpful. self.main_plot.grid(True, color='gray') pylab.setp(self.main_plot.get_xticklabels(), visible=True) self.plot_data.set
""" A static box for controlling the start and stop of the device and displaying the final result of the venous flow measurement.""" def __init__(self, parent, ID, label): wx.Panel.__init__(self, parent, ID) # Create two box sizers, one for the button and one for the status # message and final output reading. button_box = wx.StaticBox(self, -1, label = 'Device Control') info_box = wx.StaticBox(self, -1, label = 'Information') box = wx.StaticBox(self, -1, label) main_sizer = wx.StaticBoxSizer(box, wx.HORIZONTAL) button_sizer = wx.StaticBoxSizer(button_box, orient=wx.HORIZONTAL) info_sizer = wx.StaticBoxSizer(info_box, orient=wx.HORIZONTAL) # Create a start/stop measurement button. self.start_button = wx.Button(self, label = 'Start measurement',) # Create a result and information text box. self.result_box = wx.StaticText(self) self.txt_info_box = wx.StaticText(self, size=(200, -1)) self.result_box.SetLabel("0.00")
identifier_body
host_window.py
): """ Get the next data from ADC0. For ADC1, use get_dc_offset()""" self.dev.write(1, 'A0') digit1, digit2 = self.dev.read(0x81, 64)[:2] # Save the data as voltage between 0.0 and 5.0 self.data0.append((digit1 + 256*digit2)*5.0/1024) def get_dc_offset(self): """ Get the initial DC offset of the analog output""" self.dev.write(1, 'A1') digit1, digit2 = self.dev.read(0x81, 64)[:2] # Save the data as voltage between 0.0 and 5.0 self.data1.append((digit1 + 256*digit2)*5.0/1024) def sample(self): """ Set the sample bit for getting intial DC offset""" self.dev.write(1, 'S') def hold(self): """ Clear the sample bit for consecutive data aquisition""" self.dev.write(1, 'H') class VSControlBox(wx.Panel): """ A static box for controlling the start and stop of the device and displaying the final result of the venous flow measurement.""" def __init__(self, parent, ID, label): wx.Panel.__init__(self, parent, ID) # Create two box sizers, one for the button and one for the status # message and final output reading. button_box = wx.StaticBox(self, -1, label = 'Device Control') info_box = wx.StaticBox(self, -1, label = 'Information') box = wx.StaticBox(self, -1, label) main_sizer = wx.StaticBoxSizer(box, wx.HORIZONTAL) button_sizer = wx.StaticBoxSizer(button_box, orient=wx.HORIZONTAL) info_sizer = wx.StaticBoxSizer(info_box, orient=wx.HORIZONTAL) # Create a start/stop measurement button. self.start_button = wx.Button(self, label = 'Start measurement',) # Create a result and information text box. self.result_box = wx.StaticText(self) self.txt_info_box = wx.StaticText(self, size=(200, -1)) self.result_box.SetLabel("0.00") # Add the items to sizers button_sizer.Add(self.start_button, 0, wx.ALL, 10) info_sizer.Add(self.result_box, 0, wx.ALL, 10) info_sizer.Add(self.txt_info_box, 0, wx.ALL, 10) # Add the sizers to main sizer main_sizer.Add(button_sizer, flag=wx.ALIGN_CENTER_VERTICAL) main_sizer.AddSpacer(20) main_sizer.Add(info_sizer, flag=wx.ALIGN_CENTER_VERTICAL|wx.EXPAND) # Bind events to the button self.start_button.Bind(wx.EVT_BUTTON, parent.Parent.start_stop) # Finally, make a fit self.SetSizer(main_sizer) main_sizer.Fit(self) def start_stop(self, event): """ Bind a rudimentary event now. Will update it later.""" self.start_button.SetLabel('Measuring') self.start_button.Enable = False # Do nothing as of now. Will call measuring functions later. self.txt_info_box.SetLabel('Starting measurement.') time.sleep(2) self.start_button.SetLabel('Start measurement') self.start_button.Enable = True self.txt_info_box.SetLabel('Completed measurement.') self.result_box.SetLabel("100.00") class VSGraphFrame(wx.Frame): """ Main frame for the measurement application""" title = "Venous flow calculation" def __init__(self): wx.Frame.__init__(self, None, -1, self.title) self.SAMPLING_TIME = 10000.0 # Set the sampling time here. self.daq = VSDataAquisition() if self.daq == None: help_string = ''' Device is not connected. Please connect the device and restart the software''' wx.MessageBox(help_string, 'Device not found', wx.OK | wx.ICON_INFORMATION) self.Destroy() return self.create_menu() # Create the menu self.create_status_bar () # Add a status bar. Could use for debugging self.create_main_panel() # The main panel # We will use a timer for getting our samples. self.redraw_timer = wx.Timer(self, wx.ID_ANY) # Sampling duration itself is another timer. This timer is a # oneshot timer and runs for SAMPLING_TIME time. self.sampling_timer = wx.Timer(self, wx.ID_ANY) self.Bind(wx.EVT_TIMER, self.on_redraw_timer, self.redraw_timer) self.Bind(wx.EVT_TIMER, self.on_sampling_timer, self.sampling_timer) self.redraw_timer.Start(100) def create_menu(self): """ Add menu bar items. One File and one About""" self.menubar = wx.MenuBar() menu_file = wx.Menu() menu_help = wx.Menu() # Add save and exit to File menu menu_save = menu_file.Append(-1, '&Save plot\tCtrl-S', 'Save plot to a file') menu_file.AppendSeparator() menu_exit = menu_file.Append(-1, '&Exit\tCtrl-X', 'Exit the program') self.Bind(wx.EVT_MENU, self.on_save, menu_save) self.Bind(wx.EVT_MENU, self.on_exit, menu_exit) # Add an about in the Help menu. Will update later. help_about = menu_help.Append(-1, '&About', 'About the program') menu_help.AppendSeparator() self.Bind(wx.EVT_MENU, self.on_about, help_about) # Add them both to menubar self.menubar.Append(menu_file, '&File') self.menubar.Append(menu_help, '&Help') self.SetMenuBar(self.menubar) def create_main_panel(self): """ Create the main panel which will show the dynamic plot.""" self.panel = wx.Panel(self) self.init_plot() self.canvas = FigCanvas(self.panel, -1, self.fig) self.control_box = VSControlBox(self.panel, -1, 'Information board') self.vbox = wx.BoxSizer(wx.VERTICAL) self.vbox.Add(self.canvas, 1, wx.LEFT | wx.TOP | wx.GROW) self.vbox.Add(self.control_box, 0, wx.ALIGN_LEFT | wx.TOP | wx.EXPAND) self.panel.SetSizer(self.vbox) self.vbox.Fit(self) def create_status_bar(self): """ Create a status bar. Will use it in future for debugging""" self.statusbar = self.CreateStatusBar() def init_plot(self): """ Initialize the plot canvas""" self.dpi = 100 self.fig = Figure((5.0, 5.0), dpi = self.dpi) self.main_plot = self.fig.add_subplot(111) self.main_plot.set_axis_bgcolor('black') self.main_plot.set_title('Dynamic venous flow view', size = 12) pylab.setp(self.main_plot.get_xticklabels(), fontsize = 8) pylab.setp(self.main_plot.get_yticklabels(), fontsize = 8) # Plot the data as a green line self.plot_data = self.main_plot.plot( self.daq.data0, linewidth = 1, color = (0, 1, 0), )[0] self.main_plot.grid(True, color='gray') def draw_plot(self): """ Redraw the plot after every data aquisition.""" # X axis is auto follow. XLEN = 100 xmax = max(len(self.daq.data0), XLEN) xmin = xmax - XLEN # The Y value will lie between 0.0 and 5.0 volts ymax = 5.0 ymin = 0.0 self.main_plot.set_xbound(lower=xmin, upper=xmax) self.main_plot.set_ybound(lower=ymin, upper=ymax) # Add the grid. Grid looks cool and is actually very helpful. self.main_plot.grid(True, color='gray') pylab.setp(self.main_plot.get_xticklabels(), visible=True) self.plot_data.set_xdata(arange(len(self.daq.data0))) self.plot_data.set_ydata(array(self.daq.data0)) self.canvas.draw() def on_save(self, event): """ Method for saving a plot """ file_choices = "PNG (*.png)|*.png" dlg = wx.FileDialog( self, message="Save plot as...", defaultDir=os.getcwd(), defaultFile="plot.png", wildcard=file_choices, style=wx.SAVE) if dlg.ShowModal() == wx.ID_OK: path = dlg.GetPath() self.canvas.print_figure(path, dpi=self.dpi) self.flash_status_message("Saved to %s" % path) def start_stop(self, event): """ Restart measurements and complete calculations""" self.daq.data0= [] self.control_box.txt_info_box.SetLabel('Starting measurement') self.sampling_timer.Start(self.SAMPLING_TIME, oneShot=True) def
on_redraw_timer
identifier_name
plot_materials.py
f.seek(160); x_axis_type = np.fromfile(f, dtype=np.uint8, count=1)[0] print(x_axis_type) f.seek(166); x_start, x_end = np.fromfile(f, dtype=np.float32, count=2) print(x_start) print(x_end) ## Load the n, k data f.seek(174); raw_eps = np.fromfile(f, dtype=np.float32, count=datalength*2) f.close eps = raw_eps[::2] + 1j*raw_eps[1::2] from scipy.constants import h, c, eV if x_axis_type == 2: # 'eV' freq = np.linspace(x_start*eV/h, x_end*eV/h, datalength) elif x_axis_type == 3: # 'um' wavelength = np.linspace(x_start*1e-6, x_end*1e-6, datalength) freq = c/wavelength elif x_axis_type == 0: # 'cm-1' freq = np.linspace(x_start*100*c, x_end*100*c, datalength) return freq, eps #}}} def load_SOPRA_permittivity(filename):#{{{ data = [] with open(filename) as f: for line in f.readlines(): if line[:5] == 'DATA1': data.append(map(lambda x: float(x), line.split('*')[2:5])) wl, n, k = np.array(data).T eps = ((n+1j*k)**2)[::-1] freq = (2.998e8 / (wl*1e-9))[::-1] return freq, eps #}}} def load_n_k_eps(filename):#{{{ lambda_angstr, n, k = np.loadtxt(data_source, usecols=[0,1,2], unpack=True, comments=';') eps = (n+1j*k)**2 freq = 2.997e8 / (lambda_angstr*1e-10) return freq, n, k, eps #}}} def load_eps(filename):#{{{ return freq, eps #}}} ## List of sources for given materials ## == TiO2 ==#{{{ TiO2_files = [ meep_materials.material_TiO2(where = None), meep_materials.material_TiO2(where = None, extraordinary=1.), ## (rutile is anisotropic) meep_materials.material_TiO2(where = None, extraordinary=0.), "sopra/TIO2.MAT", "sopra/TIO2B.MAT", "luxpop/TiO2.nk", "luxpop/TiO2-e.nk", "luxpop/TiO2_llnl_cxro.nk", "luxpop/TiO2-e_palik.nk", "scout/TiO2 (amorph).b", "scout/TiO2 I (Jellison).b", "scout/TiO2 II (Jellison).b", "scout/TiO2 II (Jellison).b", "other/TiO2-Mounaix_polycryst.nk", ] #}}} ## == SiC == #{{{ SiC_files = [ meep_materials.material_SiC(where = None), "sopra/SIC.MAT", "luxpop/SiC_llnl_cxro.nk", "luxpop/SiC_osantowski.nk", "luxpop/SiC_palik.nk", "luxpop/SiC_windt.nk", "luxpop/SiC_yanagihara.nk", 'scout/SiC (MIR).b', 'scout/SiC (NIR-UV).b', #'scout_reverse_engineering/scout/SiC (simple infrared model).b' ]#}}} ## == SiO2 == #{{{ SiO2_files = [ 'scout/SiO2 (MIR-VUV).b', 'scout/SiO2 [micron].b', 'scout/SiO2 (fused).b', 'sopra/SIO2.MAT', #'kitamura/SiO2/Koike1989.n', meep_materials.material_SiO2(where = None) ] ##SiO2_files += ['kitamura/SiO2/'+f for f in os.listdir('kitamura/SiO2/') if (f[-2:]=='.n' or f[-2:]=='.k')] #}}} ## == Au == #{{{ Au_files = [ meep_materials.material_Au(), 'sopra/AU.MAT', 'scout/Au (J, & C,, L, & H,).b', 'scout/Au (JC).b', 'scout/Au (MQ).b', 'scout/Au [micron].b', 'scout/Au.b', 'scout/Au model.b', ] #}}} ## == Si == #{{{ Si_files = [ meep_materials.material_Si_NIR(), meep_materials.material_Si_MIR(), 'other/Si_Dai2003.k', 'other/Si_Dai2003.n', 'sopra/SI100_2.MAT', 'scout/Si (100).b', 'scout/Si (Aspnes).b', 'scout/Si (Vis-UV, Brendel model).b', 'scout/Si (cryst,).b', 'scout/Si (infrared).b', 'scout/Si (crystalline, MIR-VUV).b', ] #}}} ## == InP == #{{{ InP_files = [ meep_materials.material_InP(), 'sopra/INP.MAT', 'scout/InP.b', 'scout/InP (IR model).b', 'scout/InP (Jellison).b', ] #}}} ## == GaAs == #{{{ GaAs_files = [ meep_materials.material_GaAs(), 'sopra/GAAS.MAT', 'sopra/GAAS031T.MAT', 'scout/GaAs (100).b', 'scout/GaAs.b', 'scout/GaAs (31 deg C).b', ] #}}} ## == InSb == #{{{ InSb_files = [ 'sopra/INSB.MAT', 'scout/InSb (Jellison).b', 'scout/InSb.b', ] #}}} ## == STO == #{{{ STO_files = [ 'other/N_model_STO_300K.nk', 'other/N_STO_300K.nk', 'other/STO_Neville1972_300K.epsilon', #'other/STO_Neville1972_090K.epsilon', #'other/STO_Neville1972_004K.epsilon', #meep_materials.material_STO_THz(), meep_materials.material_STO(), ] #}}} ## == Al2O3 == #{{{ Al2O3_files = [ meep_materials.material_Sapphire(), meep_materials.material_Sapphire(ordinary=.66), meep_materials.material_Sapphire(ordinary=0), 'other/N_Al2O3_c-cut.nk', 'other/N_model_Al2O3_c-cut.nk', 'scout/Al2O3.b', 'scout/Al2O3 (Palik).b', 'sopra/AL2O3.MAT', 'sopra/AL2O3P.MAT', #'other/', ] #}}} ## == Fe2O3 == #{{{ Fe2O3_files = [ 'scout/Fe2O3.b', ] #}}} ## == Fe2O3 == #{{{ Fe3O4_files = [ 'scout/Fe3O4.b', ] #}}} ## == Fe2O3 == #{{{ Cu_files = [ 'scout/Cu.b', ] #}}} ## == Fe2O3 == #{{{ Ni_files = [ 'scout/Ni.b', ] #}}} ## == Fe2O3 == #{{{ ZnO_files = [ 'scout/ZnO.b', ] #}}} ## Dictionary of materials materials = { 'TiO2': TiO2_files, 'SiC': SiC_files
f.seek(151); datalength = np.fromfile(f, dtype=np.uint16, count=1)[0] print(datalength)
random_line_split
plot_materials.py
scout/Au (J, & C,, L, & H,).b', 'scout/Au (JC).b', 'scout/Au (MQ).b', 'scout/Au [micron].b', 'scout/Au.b', 'scout/Au model.b', ] #}}} ## == Si == #{{{ Si_files = [ meep_materials.material_Si_NIR(), meep_materials.material_Si_MIR(), 'other/Si_Dai2003.k', 'other/Si_Dai2003.n', 'sopra/SI100_2.MAT', 'scout/Si (100).b', 'scout/Si (Aspnes).b', 'scout/Si (Vis-UV, Brendel model).b', 'scout/Si (cryst,).b', 'scout/Si (infrared).b', 'scout/Si (crystalline, MIR-VUV).b', ] #}}} ## == InP == #{{{ InP_files = [ meep_materials.material_InP(), 'sopra/INP.MAT', 'scout/InP.b', 'scout/InP (IR model).b', 'scout/InP (Jellison).b', ] #}}} ## == GaAs == #{{{ GaAs_files = [ meep_materials.material_GaAs(), 'sopra/GAAS.MAT', 'sopra/GAAS031T.MAT', 'scout/GaAs (100).b', 'scout/GaAs.b', 'scout/GaAs (31 deg C).b', ] #}}} ## == InSb == #{{{ InSb_files = [ 'sopra/INSB.MAT', 'scout/InSb (Jellison).b', 'scout/InSb.b', ] #}}} ## == STO == #{{{ STO_files = [ 'other/N_model_STO_300K.nk', 'other/N_STO_300K.nk', 'other/STO_Neville1972_300K.epsilon', #'other/STO_Neville1972_090K.epsilon', #'other/STO_Neville1972_004K.epsilon', #meep_materials.material_STO_THz(), meep_materials.material_STO(), ] #}}} ## == Al2O3 == #{{{ Al2O3_files = [ meep_materials.material_Sapphire(), meep_materials.material_Sapphire(ordinary=.66), meep_materials.material_Sapphire(ordinary=0), 'other/N_Al2O3_c-cut.nk', 'other/N_model_Al2O3_c-cut.nk', 'scout/Al2O3.b', 'scout/Al2O3 (Palik).b', 'sopra/AL2O3.MAT', 'sopra/AL2O3P.MAT', #'other/', ] #}}} ## == Fe2O3 == #{{{ Fe2O3_files = [ 'scout/Fe2O3.b', ] #}}} ## == Fe2O3 == #{{{ Fe3O4_files = [ 'scout/Fe3O4.b', ] #}}} ## == Fe2O3 == #{{{ Cu_files = [ 'scout/Cu.b', ] #}}} ## == Fe2O3 == #{{{ Ni_files = [ 'scout/Ni.b', ] #}}} ## == Fe2O3 == #{{{ ZnO_files = [ 'scout/ZnO.b', ] #}}} ## Dictionary of materials materials = { 'TiO2': TiO2_files, 'SiC': SiC_files, 'SiO2': SiO2_files, 'Au': Au_files, 'Si': Si_files, 'InP': InP_files, 'GaAs': GaAs_files, 'InSb': InSb_files, 'STO': STO_files, 'Al2O3': Al2O3_files, 'Fe2O3': Fe2O3_files, 'Fe3O4': Fe3O4_files, 'Ni': Ni_files, 'ZnO': ZnO_files, 'Cu': Cu_files, } ##and not (len(sys.argv)>2 and (sys.argv[2] == "--big")) if len(sys.argv)>1 and (sys.argv[1] in materials.keys()): data_sources = materials[sys.argv[1]] output_file = sys.argv[1] else: print("Error: the first parameter may be one of the following materials:", materials.keys()) exit(1) if len(sys.argv)>2 and (sys.argv[2] == "--big"): output_file+="_big"; plt.figure(figsize=(16,16)) else: plt.figure(figsize=(8,8)) freq_range=(1e11, 1e16) for number, data_source in enumerate(data_sources): try: if type(data_source)==str: plotlabel = data_source ## Select the file type to be loaded from n, eps, k = None, None, None print("Loading ", data_source) if data_source.endswith('.b'): freq, eps = load_SCOUT_permittivity(data_source) n, k = (eps**.5).real, (eps**.5).imag print(freq) print(eps) print(n) print(k) elif data_source.endswith('.MAT'): freq, eps = load_SOPRA_permittivity(data_source) n, k = (eps**.5).real, (eps**.5).imag elif data_source.endswith('.epsilon'): freq, eps = np.loadtxt(data_source, usecols=[0,1], unpack=True, comments='#') n, k = (eps**.5).real, (eps**.5).imag print(freq, n) elif data_source.endswith('.nk'): freq, n, k, eps = load_n_k_eps(data_source) elif data_source.endswith('.n'): lambda_micron, n = np.loadtxt(data_source, usecols=[0,1], unpack=True, comments='#') freq = 2.997e8 / (lambda_micron*1e-6) k, eps = None, None elif data_source.endswith('.k'): lambda_micron, k = np.loadtxt(data_source, usecols=[0,1], unpack=True, comments='#') freq = 2.997e8 / (lambda_micron*1e-6) print(freq) n, eps = None, None ## Read the MEEP's material model if hasattr(data_source, 'pol'): freq = 10**np.arange(np.log10(freq_range[0]), np.log10(freq_range[1]), .003) eps = analytic_eps(data_source, freq) n = (analytic_eps(data_source, freq)**.5).real k = (analytic_eps(data_source, freq)**.5).imag plotlabel = "model " + getattr(data_source, 'shortname', data_source.name) ## Plot the data if hasattr(data_source, 'pol'): color = 'black' else: color = matplotlib.cm.hsv(float(number)/len(data_sources)) if eps.all()!=None or (n.all()!=None and k.all()!=None): if eps.all()==None: eps=(n+1j*k)**.5 plt.subplot(3,1,1) print(" Plotting epsilon for '%s' with %d data points" % (plotlabel, len(eps))) plt.plot(freq, eps+float(number)/100, color=color, marker='o', markersize=0, label=plotlabel) plt.plot(freq, eps.imag+float(number)/100, color=color, marker='s', markersize=0, ls='--') plt.xlim(freq_range); plt.grid(True) if n.all()!=None: plt.subplot(3,1,2) print(" Plotting N for '%s' with %d data points" % (plotlabel, len(n))) plt.plot(freq, n, color=color, marker='o', markersize=0, label=plotlabel) plt.xlim(freq_range); plt.grid(True) if k.all()!=None:
plt.subplot(3,1,3) print(" Plotting k for '%s' with %d data points" % (plotlabel, len(k))) plt.plot(freq, k, color=color, marker='o', markersize=0, label=plotlabel) plt.xlim(freq_range); plt.grid(True)
conditional_block
plot_materials.py
(filename):#{{{ """ Reads the permittivity function from a given file with SCOUT binary format The SCOUT database of materials is supplied with the SCOUT program and may be freely downloaded from http://www.mtheiss.com/download/scout.zip Different files use different units for the x-axis (179 files with "eV", 118 "micron", 49 "1/cm"), so we have to build the frequency axis accordingly. The data were randomly verified against the SOPRA and luxpop data. """ ## Open the file for binary access f = open(filename, "rb") ## Load the number of data points, type of x axis and its boundaries f.seek(151); datalength = np.fromfile(f, dtype=np.uint16, count=1)[0] print(datalength) f.seek(160); x_axis_type = np.fromfile(f, dtype=np.uint8, count=1)[0] print(x_axis_type) f.seek(166); x_start, x_end = np.fromfile(f, dtype=np.float32, count=2) print(x_start) print(x_end) ## Load the n, k data f.seek(174); raw_eps = np.fromfile(f, dtype=np.float32, count=datalength*2) f.close eps = raw_eps[::2] + 1j*raw_eps[1::2] from scipy.constants import h, c, eV if x_axis_type == 2: # 'eV' freq = np.linspace(x_start*eV/h, x_end*eV/h, datalength) elif x_axis_type == 3: # 'um' wavelength = np.linspace(x_start*1e-6, x_end*1e-6, datalength) freq = c/wavelength elif x_axis_type == 0: # 'cm-1' freq = np.linspace(x_start*100*c, x_end*100*c, datalength) return freq, eps #}}} def load_SOPRA_permittivity(filename):#{{{ data = [] with open(filename) as f: for line in f.readlines(): if line[:5] == 'DATA1': data.append(map(lambda x: float(x), line.split('*')[2:5])) wl, n, k = np.array(data).T eps = ((n+1j*k)**2)[::-1] freq = (2.998e8 / (wl*1e-9))[::-1] return freq, eps #}}} def load_n_k_eps(filename):#{{{ lambda_angstr, n, k = np.loadtxt(data_source, usecols=[0,1,2], unpack=True, comments=';') eps = (n+1j*k)**2 freq = 2.997e8 / (lambda_angstr*1e-10) return freq, n, k, eps #}}} def load_eps(filename):#{{{ return freq, eps #}}} ## List of sources for given materials ## == TiO2 ==#{{{ TiO2_files = [ meep_materials.material_TiO2(where = None), meep_materials.material_TiO2(where = None, extraordinary=1.), ## (rutile is anisotropic) meep_materials.material_TiO2(where = None, extraordinary=0.), "sopra/TIO2.MAT", "sopra/TIO2B.MAT", "luxpop/TiO2.nk", "luxpop/TiO2-e.nk", "luxpop/TiO2_llnl_cxro.nk", "luxpop/TiO2-e_palik.nk", "scout/TiO2 (amorph).b", "scout/TiO2 I (Jellison).b", "scout/TiO2 II (Jellison).b", "scout/TiO2 II (Jellison).b", "other/TiO2-Mounaix_polycryst.nk", ] #}}} ## == SiC == #{{{ SiC_files = [ meep_materials.material_SiC(where = None), "sopra/SIC.MAT", "luxpop/SiC_llnl_cxro.nk", "luxpop/SiC_osantowski.nk", "luxpop/SiC_palik.nk", "luxpop/SiC_windt.nk", "luxpop/SiC_yanagihara.nk", 'scout/SiC (MIR).b', 'scout/SiC (NIR-UV).b', #'scout_reverse_engineering/scout/SiC (simple infrared model).b' ]#}}} ## == SiO2 == #{{{ SiO2_files = [ 'scout/SiO2 (MIR-VUV).b', 'scout/SiO2 [micron].b', 'scout/SiO2 (fused).b', 'sopra/SIO2.MAT', #'kitamura/SiO2/Koike1989.n', meep_materials.material_SiO2(where = None) ] ##SiO2_files += ['kitamura/SiO2/'+f for f in os.listdir('kitamura/SiO2/') if (f[-2:]=='.n' or f[-2:]=='.k')] #}}} ## == Au == #{{{ Au_files = [ meep_materials.material_Au(), 'sopra/AU.MAT', 'scout/Au (J, & C,, L, & H,).b', 'scout/Au (JC).b', 'scout/Au (MQ).b', 'scout/Au [micron].b', 'scout/Au.b', 'scout/Au model.b', ] #}}} ## == Si == #{{{ Si_files = [ meep_materials.material_Si_NIR(), meep_materials.material_Si_MIR(), 'other/Si_Dai2003.k', 'other/Si_Dai2003.n', 'sopra/SI100_2.MAT', 'scout/Si (100).b', 'scout/Si (Aspnes).b', 'scout/Si (Vis-UV, Brendel model).b', 'scout/Si (cryst,).b', 'scout/Si (infrared).b', 'scout/Si (crystalline, MIR-VUV).b', ] #}}} ## == InP == #{{{ InP_files = [ meep_materials.material_InP(), 'sopra/INP.MAT', 'scout/InP.b', 'scout/InP (IR model).b', 'scout/InP (Jellison).b', ] #}}} ## == GaAs == #{{{ GaAs_files = [ meep_materials.material_GaAs(), 'sopra/GAAS.MAT', 'sopra/GAAS031T.MAT', 'scout/GaAs (100).b', 'scout/GaAs.b', 'scout/GaAs (31 deg C).b', ] #}}} ## == InSb == #{{{ InSb_files = [ 'sopra/INSB.MAT', 'scout/InSb (Jellison).b', 'scout/InSb.b', ] #}}} ## == STO == #{{{ STO_files = [ 'other/N_model_STO_300K.nk', 'other/N_STO_300K.nk', 'other/STO_Neville1972_300K.epsilon', #'other/STO_Neville1972_090K.epsilon', #'other/STO_Neville1972_004K.epsilon', #meep_materials.material_STO_THz(), meep_materials.material_STO(), ] #}}} ## == Al2O3 == #{{{ Al2O3_files = [ meep_materials.material_Sapphire(), meep_materials.material_Sapphire(ordinary=.66), meep_materials.material_Sapphire(ordinary=0), 'other/N_Al2O3_c-cut.nk', 'other/N_model_Al2O3_c-cut.nk', 'scout/Al2O3.b', 'scout/Al2O3 (Palik).b', 'sopra/AL2O3.MAT', 'sopra/AL2O3P.MAT', #'other/', ] #}}} ## == Fe2O3 == #{{{ Fe2O3_files = [
load_SCOUT_permittivity
identifier_name
plot_materials.py
#}}} ## Functions loading data from different format files ## Note: you must obtain the SOPRA and SCOUT databases from the web, if you want to use them def load_SCOUT_permittivity(filename):#{{{ """ Reads the permittivity function from a given file with SCOUT binary format The SCOUT database of materials is supplied with the SCOUT program and may be freely downloaded from http://www.mtheiss.com/download/scout.zip Different files use different units for the x-axis (179 files with "eV", 118 "micron", 49 "1/cm"), so we have to build the frequency axis accordingly. The data were randomly verified against the SOPRA and luxpop data. """ ## Open the file for binary access f = open(filename, "rb") ## Load the number of data points, type of x axis and its boundaries f.seek(151); datalength = np.fromfile(f, dtype=np.uint16, count=1)[0] print(datalength) f.seek(160); x_axis_type = np.fromfile(f, dtype=np.uint8, count=1)[0] print(x_axis_type) f.seek(166); x_start, x_end = np.fromfile(f, dtype=np.float32, count=2) print(x_start) print(x_end) ## Load the n, k data f.seek(174); raw_eps = np.fromfile(f, dtype=np.float32, count=datalength*2) f.close eps = raw_eps[::2] + 1j*raw_eps[1::2] from scipy.constants import h, c, eV if x_axis_type == 2: # 'eV' freq = np.linspace(x_start*eV/h, x_end*eV/h, datalength) elif x_axis_type == 3: # 'um' wavelength = np.linspace(x_start*1e-6, x_end*1e-6, datalength) freq = c/wavelength elif x_axis_type == 0: # 'cm-1' freq = np.linspace(x_start*100*c, x_end*100*c, datalength) return freq, eps #}}} def load_SOPRA_permittivity(filename):#{{{ data = [] with open(filename) as f: for line in f.readlines(): if line[:5] == 'DATA1': data.append(map(lambda x: float(x), line.split('*')[2:5])) wl, n, k = np.array(data).T eps = ((n+1j*k)**2)[::-1] freq = (2.998e8 / (wl*1e-9))[::-1] return freq, eps #}}} def load_n_k_eps(filename):#{{{ lambda_angstr, n, k = np.loadtxt(data_source, usecols=[0,1,2], unpack=True, comments=';') eps = (n+1j*k)**2 freq = 2.997e8 / (lambda_angstr*1e-10) return freq, n, k, eps #}}} def load_eps(filename):#{{{ return freq, eps #}}} ## List of sources for given materials ## == TiO2 ==#{{{ TiO2_files = [ meep_materials.material_TiO2(where = None), meep_materials.material_TiO2(where = None, extraordinary=1.), ## (rutile is anisotropic) meep_materials.material_TiO2(where = None, extraordinary=0.), "sopra/TIO2.MAT", "sopra/TIO2B.MAT", "luxpop/TiO2.nk", "luxpop/TiO2-e.nk", "luxpop/TiO2_llnl_cxro.nk", "luxpop/TiO2-e_palik.nk", "scout/TiO2 (amorph).b", "scout/TiO2 I (Jellison).b", "scout/TiO2 II (Jellison).b", "scout/TiO2 II (Jellison).b", "other/TiO2-Mounaix_polycryst.nk", ] #}}} ## == SiC == #{{{ SiC_files = [ meep_materials.material_SiC(where = None), "sopra/SIC.MAT", "luxpop/SiC_llnl_cxro.nk", "luxpop/SiC_osantowski.nk", "luxpop/SiC_palik.nk", "luxpop/SiC_windt.nk", "luxpop/SiC_yanagihara.nk", 'scout/SiC (MIR).b', 'scout/SiC (NIR-UV).b', #'scout_reverse_engineering/scout/SiC (simple infrared model).b' ]#}}} ## == SiO2 == #{{{ SiO2_files = [ 'scout/SiO2 (MIR-VUV).b', 'scout/SiO2 [micron].b', 'scout/SiO2 (fused).b', 'sopra/SIO2.MAT', #'kitamura/SiO2/Koike1989.n', meep_materials.material_SiO2(where = None) ] ##SiO2_files += ['kitamura/SiO2/'+f for f in os.listdir('kitamura/SiO2/') if (f[-2:]=='.n' or f[-2:]=='.k')] #}}} ## == Au == #{{{ Au_files = [ meep_materials.material_Au(), 'sopra/AU.MAT', 'scout/Au (J, & C,, L, & H,).b', 'scout/Au (JC).b', 'scout/Au (MQ).b', 'scout/Au [micron].b', 'scout/Au.b', 'scout/Au model.b', ] #}}} ## == Si == #{{{ Si_files = [ meep_materials.material_Si_NIR(), meep_materials.material_Si_MIR(), 'other/Si_Dai2003.k', 'other/Si_Dai2003.n', 'sopra/SI100_2.MAT', 'scout/Si (100).b', 'scout/Si (Aspnes).b', 'scout/Si (Vis-UV, Brendel model).b', 'scout/Si (cryst,).b', 'scout/Si (infrared).b', 'scout/Si (crystalline, MIR-VUV).b', ] #}}} ## == InP == #{{{ InP_files = [ meep_materials.material_InP(), 'sopra/INP.MAT', 'scout/InP.b', 'scout/InP (IR model).b', 'scout/InP (Jellison).b', ] #}}} ## == GaAs == #{{{ GaAs_files = [ meep_materials.material_GaAs(), 'sopra/GAAS.MAT', 'sopra/GAAS031T.MAT', 'scout/GaAs (100).b', 'scout/GaAs.b', 'scout/GaAs (31 deg C).b', ] #}}} ## == InSb == #{{{ InSb_files = [ 'sopra/INSB.MAT', 'scout/InSb (Jellison).b', 'scout/InSb.b', ] #}}} ## == STO == #{{{ STO_files = [ 'other/N_model_STO_300K.nk', 'other/N_STO_300K.nk', 'other/STO_Neville1972_300K.epsilon', #'other/STO_Neville1972_090K.epsilon', #'other/STO_Neville1972_004K.epsilon', #meep_materials.material_STO_THz(), meep_materials.material_STO(), ] #}}} ## == Al2O3 == #{{{ Al2O3_files = [ meep_materials.material_Sapphire(), meep_materials.material_Sapphire(ordinary=.66), meep_materials.material_Sapphire(ordinary=0),
complex_eps = mat.eps for polariz in mat.pol: complex_eps += polariz['sigma'] * polariz['omega']**2 / (polariz['omega']**2 - freq**2 - 1j*freq*polariz['gamma']) return complex_eps # + sum(0)
identifier_body
generate_concept_dicts.py
", "http://hl7.org/fhir/sid/ndc": "NDC", "urn:iso:std:iso:11073:10101": "MDC", "doi:10.1016/S0735-1097(99)00126-6": "BARI", "http://www.nlm.nih.gov/research/umls": "UMLS", "http://pubchem.ncbi.nlm.nih.gov": "PUBCHEM_CID", "http://braininfo.rprc.washington.edu/aboutBrainInfo.aspx#NeuroNames": "NEU", "http://www.itis.gov": "ITIS_TSN", "http://arxiv.org/abs/1612.07003": "IBSI", "http://www.nlm.nih.gov/research/umls/rxnorm": "RXNORM", } DOC_LINES = [ f"# Auto-generated by {os.path.basename(__file__)}.\n", "# -*- coding: utf-8 -*-\n", "\n", ] def camel_case(s): leave_alone = ( "mm", "cm", "km", "um", "ms", # 'us'?-doesn't seem to be there "ml", "mg", "kg", ) # ... probably need others return "".join( word.capitalize() if word != word.upper() and word not in leave_alone else word for word in re.split(r"\W", s, flags=re.UNICODE) if word.isalnum() ) def keyword_from_meaning(name): """Return a camel case valid python identifier""" # Try to adhere to keyword scheme in DICOM (CP850) # singular/plural alternative forms are made plural # e.g., “Physician(s) of Record” becomes “PhysiciansOfRecord” name = name.replace("(s)", "s") # “Patient’s Name” -> “PatientName” # “Operators’ Name” -> “OperatorsName” name = name.replace("’s ", " ") name = name.replace("'s ", " ") name = name.replace("s’ ", "s ") name = name.replace("s' ", "s ") # Mathematical symbols name = name.replace("%", " Percent ") name = name.replace(">", " Greater Than ") name = name.replace("=", " Equals ") name = name.replace("<", " Lesser Than ") name = re.sub(r"([0-9]+)\.([0-9]+)", "\\1 Point \\2", name) name = re.sub(r"\s([0-9.]+)-([0-9.]+)\s", " \\1 To \\2 ", name) name = re.sub(r"([0-9]+)day", "\\1 Day", name) name = re.sub(r"([0-9]+)y", "\\1 Years", name) # Remove category modifiers, such as "(specimen)", "(procedure)", # "(body structure)", etc. name = re.sub(r"^(.+) \([a-z ]+\)$", "\\1", name) name = camel_case(name.strip()) # Python variables must not begin with a number. if re.match(r"[0-9]", name): name = "_" + name return name def download_fhir_value_sets(local_dir): ftp_host = "medical.nema.org" if not os.path.exists(local_dir): os.makedirs(local_dir) logger.info("storing files in " + local_dir) logger.info(f'log into FTP server "{ftp_host}"') ftp = ftplib.FTP(ftp_host, timeout=60) ftp.login("anonymous") ftp_path = "medical/dicom/resources/valuesets/fhir/json" logger.info(f'list files in directory "{ftp_path}"') fhir_value_set_files = ftp.nlst(ftp_path) try: for ftp_filepath in fhir_value_set_files: ftp_filename = os.path.basename(ftp_filepath) logger.info(f'retrieve value set file "{ftp_filename}"') with BytesIO() as fp: ftp.retrbinary(f"RETR {ftp_filepath}", fp.write) content = fp.getvalue() local_filename = os.path.join(local_dir, ftp_filename) with open(local_filename, "wb") as f_local: f_local.write(content) finally: ftp.quit() def _parse_html(content): # from lxml import html # doc = html.document_fromstring(content) return ET.fromstring(content, pa
response = urllib_request.urlopen(url) return response.read() def _get_text(element): text = "".join(element.itertext()) return text.strip() def get_table_o1(): logger.info("process Table O1") url = "http://dicom.nema.org/medical/dicom/current/output/chtml/part16/chapter_O.html#table_O-1" root = _parse_html(_download_html(url)) namespaces = {"w3": root.tag.split("}")[0].strip("{")} body = root.find("w3:body", namespaces=namespaces) table = body.findall(".//w3:tbody", namespaces=namespaces)[0] rows = table.findall("./w3:tr", namespaces=namespaces) data = [] for row in rows: data.append( ( _get_text(row[0].findall(".//w3:p", namespaces=namespaces)[-1]), _get_text(row[1].findall(".//w3:p", namespaces=namespaces)[0]), _get_text(row[2].findall(".//w3:p", namespaces=namespaces)[0]), ) ) return data def get_table_d1(): logger.info("process Table D1") url = "http://dicom.nema.org/medical/dicom/current/output/chtml/part16/chapter_D.html#table_D-1" root = _parse_html(_download_html(url)) namespaces = {"w3": root.tag.split("}")[0].strip("{")} body = root.find("w3:body", namespaces=namespaces) table = body.findall(".//w3:tbody", namespaces=namespaces)[0] rows = table.findall("./w3:tr", namespaces=namespaces) return [ ( _get_text(row[0].findall(".//w3:p", namespaces=namespaces)[0]), _get_text(row[1].findall(".//w3:p", namespaces=namespaces)[0]), ) for row in rows ] def write_concepts(concepts, cid_concepts, cid_lists, name_for_cid): lines = DOC_LINES + [ "# Dict with scheme designator keys; value format is:\n", "# {keyword: {code1: (meaning, cid_list), code2: ...}\n", "#\n", "# Most keyword identifiers map to a single code, but not all\n", "\n", ] with open("_concepts_dict.py", "w", encoding="UTF8") as f_concepts: f_concepts.writelines(lines) f_concepts.write("concepts = {}\n") # start with empty dict for scheme, value in concepts.items(): f_concepts.write(f"\nconcepts['{scheme}'] = \\\n") pprint(value, f_concepts) lines = DOC_LINES + [ "# Dict with cid number as keys; value format is:\n", "# {scheme designator: <list of keywords for current cid>\n", "# scheme_designator: ...}\n", "\n", ] with open("_cid_dict.py", "w", encoding="UTF8") as f_cid: f_cid.writelines(lines) f_cid.write("name_for_cid = {}\n") f_cid.write("cid_concepts = {}\n") for cid, value in cid_lists.items(): f_cid.write(f"\nname_for_cid[{cid}] = '{name_for_cid[cid]}'\n") f_cid.write(f"cid_concepts[{cid}] = \\\n") pprint(value, f_cid) def write_snomed_mapping(snomed_codes): with open("_snomed_dict.py", "w", encoding="UTF8") as f_concepts: lines = DOC_LINES + [ "# Dict with scheme designator keys; value format is:\n", "# {concept_id1: snomed_id1, concept_id2: ...}\n", "# or\n", "# {snomed_id1: concept_id1, snomed_id2: ...}\n", "\n", ] f_concepts.writelines(lines) f_concepts.write("mapping = {}\n") # start with empty dict f_concepts.write("\nmapping['{}'] = {{\n".format("SCT")) for sct, srt, meaning in snomed_codes: f_concepts.write(f" '{sct}': '{srt}',\n") f_concepts.write("}\n") f_concepts.write("\nmapping['{}'] = {{\n".format("SRT")) for sct, srt, meaning in snomed_codes:
rser=ET.XMLParser(encoding="utf-8")) def _download_html(url):
identifier_body
generate_concept_dicts.py
SI", "http://www.nlm.nih.gov/research/umls/rxnorm": "RXNORM", } DOC_LINES = [ f"# Auto-generated by {os.path.basename(__file__)}.\n", "# -*- coding: utf-8 -*-\n", "\n", ] def camel_case(s): leave_alone = ( "mm", "cm", "km", "um", "ms", # 'us'?-doesn't seem to be there "ml", "mg", "kg", ) # ... probably need others return "".join( word.capitalize() if word != word.upper() and word not in leave_alone else word for word in re.split(r"\W", s, flags=re.UNICODE) if word.isalnum() ) def keyword_from_meaning(name): """Return a camel case valid python identifier""" # Try to adhere to keyword scheme in DICOM (CP850) # singular/plural alternative forms are made plural # e.g., “Physician(s) of Record” becomes “PhysiciansOfRecord” name = name.replace("(s)", "s") # “Patient’s Name” -> “PatientName” # “Operators’ Name” -> “OperatorsName” name = name.replace("’s ", " ") name = name.replace("'s ", " ") name = name.replace("s’ ", "s ") name = name.replace("s' ", "s ") # Mathematical symbols name = name.replace("%", " Percent ") name = name.replace(">", " Greater Than ") name = name.replace("=", " Equals ") name = name.replace("<", " Lesser Than ") name = re.sub(r"([0-9]+)\.([0-9]+)", "\\1 Point \\2", name) name = re.sub(r"\s([0-9.]+)-([0-9.]+)\s", " \\1 To \\2 ", name) name = re.sub(r"([0-9]+)day", "\\1 Day", name) name = re.sub(r"([0-9]+)y", "\\1 Years", name) # Remove category modifiers, such as "(specimen)", "(procedure)", # "(body structure)", etc. name = re.sub(r"^(.+) \([a-z ]+\)$", "\\1", name) name = camel_case(name.strip()) # Python variables must not begin with a number. if re.match(r"[0-9]", name): name = "_" + name return name def download_fhir_value_sets(local_dir): ftp_host = "medical.nema.org" if not os.path.exists(local_dir): os.makedirs(local_dir) logger.info("storing files in " + local_dir) logger.info(f'log into FTP server "{ftp_host}"') ftp = ftplib.FTP(ftp_host, timeout=60) ftp.login("anonymous") ftp_path = "medical/dicom/resources/valuesets/fhir/json" logger.info(f'list files in directory "{ftp_path}"') fhir_value_set_files = ftp.nlst(ftp_path) try: for ftp_filepath in fhir_value_set_files: ftp_filename = os.path.basename(ftp_filepath) logger.info(f'retrieve value set file "{ftp_filename}"') with BytesIO() as fp: ftp.retrbinary(f"RETR {ftp_filepath}", fp.write) content = fp.getvalue() local_filename = os.path.join(local_dir, ftp_filename) with open(local_filename, "wb") as f_local: f_local.write(content) finally: ftp.quit() def _parse_html(content): # from lxml import html # doc = html.document_fromstring(content) return ET.fromstring(content, parser=ET.XMLParser(encoding="utf-8")) def _download_html(url): response = urllib_request.urlopen(url) return response.read() def _get_text(element): text = "".join(element.itertext()) return text.strip() def get_table_o1(): logger.info("process Table O1") url = "http://dicom.nema.org/medical/dicom/current/output/chtml/part16/chapter_O.html#table_O-1" root = _parse_html(_download_html(url)) namespaces = {"w3": root.tag.split("}")[0].strip("{")} body = root.find("w3:body", namespaces=namespaces) table = body.findall(".//w3:tbody", namespaces=namespaces)[0] rows = table.findall("./w3:tr", namespaces=namespaces) data = [] for row in rows: data.append( ( _get_text(row[0].findall(".//w3:p", namespaces=namespaces)[-1]), _get_text(row[1].findall(".//w3:p", namespaces=namespaces)[0]), _get_text(row[2].findall(".//w3:p", namespaces=namespaces)[0]), ) ) return data def get_table_d1(): logger.info("process Table D1") url = "http://dicom.nema.org/medical/dicom/current/output/chtml/part16/chapter_D.html#table_D-1" root = _parse_html(_download_html(url)) namespaces = {"w3": root.tag.split("}")[0].strip("{")} body = root.find("w3:body", namespaces=namespaces) table = body.findall(".//w3:tbody", namespaces=namespaces)[0] rows = table.findall("./w3:tr", namespaces=namespaces) return [ ( _get_text(row[0].findall(".//w3:p", namespaces=namespaces)[0]), _get_text(row[1].findall(".//w3:p", namespaces=namespaces)[0]), ) for row in rows ] def write_concepts(concepts, cid_concepts, cid_lists, name_for_cid): lines = DOC_LINES + [ "# Dict with scheme designator keys; value format is:\n", "# {keyword: {code1: (meaning, cid_list), code2: ...}\n", "#\n", "# Most keyword identifiers map to a single code, but not all\n", "\n", ] with open("_concepts_dict.py", "w", encoding="UTF8") as f_concepts: f_concepts.writelines(lines) f_concepts.write("concepts = {}\n") # start with empty dict for scheme, value in concepts.items(): f_concepts.write(f"\nconcepts['{scheme}'] = \\\n") pprint(value, f_concepts) lines = DOC_LINES + [ "# Dict with cid number as keys; value format is:\n", "# {scheme designator: <list of keywords for current cid>\n", "# scheme_designator: ...}\n", "\n", ] with open("_cid_dict.py", "w", encoding="UTF8") as f_cid: f_cid.writelines(lines) f_cid.write("name_for_cid = {}\n") f_cid.write("cid_concepts = {}\n") for cid, value in cid_lists.items(): f_cid.write(f"\nname_for_cid[{cid}] = '{name_for_cid[cid]}'\n") f_cid.write(f"cid_concepts[{cid}] = \\\n") pprint(value, f_cid) def write_snomed_mapping(snomed_codes): with open("_snomed_dict.py", "w", encoding="UTF8") as f_concepts: lines = DOC_LINES + [ "# Dict with scheme designator keys; value format is:\n", "# {concept_id1: snomed_id1, concept_id2: ...}\n", "# or\n", "# {snomed_id1: concept_id1, snomed_id2: ...}\n", "\n", ] f_concepts.writelines(lines) f_concepts.write("mapping = {}\n") # start with empty dict f_concepts.write("\nmapping['{}'] = {{\n".format("SCT")) for sct, srt, meaning in snomed_codes: f_concepts.write(f" '{sct}': '{srt}',\n") f_concepts.write("}\n") f_concepts.write("\nmapping['{}'] = {{\n".format("SRT")) for sct, srt, meaning in snomed_codes: f_concepts.write(f" '{srt}': '{sct}',\n") f_concepts.write("}") if __name__ == "__main__": logging.basicConfig(level=logging.INFO) local_dir = tempfile.gettempdir() fhir_dir = os.path.join(local_dir, "fhir") if not os.path.exists(fhir_dir) or not os.listdir(fhir_dir): download_fhir_value_sets(fhir_dir) else: msg = "Using locally downloaded files\n" msg += "from directory " + fhir_dir logging.info(msg) fhir_value_set_files = glob.glob(os.path.join(fhir_dir, "*")) cid_pattern = re.compile("^dicom-cid-([0-9]+)-[a-zA-Z]+") concepts = dict() cid_lists = dict() name_for_cid = dict()
random_line_split
generate_concept_dicts.py
UM", "http://hl7.org/fhir/sid/ndc": "NDC", "urn:iso:std:iso:11073:10101": "MDC", "doi:10.1016/S0735-1097(99)00126-6": "BARI", "http://www.nlm.nih.gov/research/umls": "UMLS", "http://pubchem.ncbi.nlm.nih.gov": "PUBCHEM_CID", "http://braininfo.rprc.washington.edu/aboutBrainInfo.aspx#NeuroNames": "NEU", "http://www.itis.gov": "ITIS_TSN", "http://arxiv.org/abs/1612.07003": "IBSI", "http://www.nlm.nih.gov/research/umls/rxnorm": "RXNORM", } DOC_LINES = [ f"# Auto-generated by {os.path.basename(__file__)}.\n", "# -*- coding: utf-8 -*-\n", "\n", ] def camel_case(s): leave_alone = ( "mm", "cm", "km", "um", "ms", # 'us'?-doesn't seem to be there "ml", "mg", "kg", ) # ... probably need others return "".join( word.capitalize() if word != word.upper() and word not in leave_alone else word for word in re.split(r"\W", s, flags=re.UNICODE) if word.isalnum() ) def keyword_from_meaning(name): """Return a camel case valid python identifier""" # Try to adhere to keyword scheme in DICOM (CP850) # singular/plural alternative forms are made plural # e.g., “Physician(s) of Record” becomes “PhysiciansOfRecord” name = name.replace("(s)", "s") # “Patient’s Name” -> “PatientName” # “Operators’ Name” -> “OperatorsName” name = name.replace("’s ", " ") name = name.replace("'s ", " ") name = name.replace("s’ ", "s ") name = name.replace("s' ", "s ") # Mathematical symbols name = name.replace("%", " Percent ") name = name.replace(">", " Greater Than ") name = name.replace("=", " Equals ") name = name.replace("<", " Lesser Than ") name = re.sub(r"([0-9]+)\.([0-9]+)", "\\1 Point \\2", name) name = re.sub(r"\s([0-9.]+)-([0-9.]+)\s", " \\1 To \\2 ", name) name = re.sub(r"([0-9]+)day", "\\1 Day", name) name = re.sub(r"([0-9]+)y", "\\1 Years", name) # Remove category modifiers, such as "(specimen)", "(procedure)", # "(body structure)", etc. name = re.sub(r"^(.+) \([a-z ]+\)$", "\\1", name) name = camel_case(name.strip()) # Python variables must not begin with a number. if re.match(r"[0-9]", name): name = "_" + name return na
_fhir_value_sets(local_dir): ftp_host = "medical.nema.org" if not os.path.exists(local_dir): os.makedirs(local_dir) logger.info("storing files in " + local_dir) logger.info(f'log into FTP server "{ftp_host}"') ftp = ftplib.FTP(ftp_host, timeout=60) ftp.login("anonymous") ftp_path = "medical/dicom/resources/valuesets/fhir/json" logger.info(f'list files in directory "{ftp_path}"') fhir_value_set_files = ftp.nlst(ftp_path) try: for ftp_filepath in fhir_value_set_files: ftp_filename = os.path.basename(ftp_filepath) logger.info(f'retrieve value set file "{ftp_filename}"') with BytesIO() as fp: ftp.retrbinary(f"RETR {ftp_filepath}", fp.write) content = fp.getvalue() local_filename = os.path.join(local_dir, ftp_filename) with open(local_filename, "wb") as f_local: f_local.write(content) finally: ftp.quit() def _parse_html(content): # from lxml import html # doc = html.document_fromstring(content) return ET.fromstring(content, parser=ET.XMLParser(encoding="utf-8")) def _download_html(url): response = urllib_request.urlopen(url) return response.read() def _get_text(element): text = "".join(element.itertext()) return text.strip() def get_table_o1(): logger.info("process Table O1") url = "http://dicom.nema.org/medical/dicom/current/output/chtml/part16/chapter_O.html#table_O-1" root = _parse_html(_download_html(url)) namespaces = {"w3": root.tag.split("}")[0].strip("{")} body = root.find("w3:body", namespaces=namespaces) table = body.findall(".//w3:tbody", namespaces=namespaces)[0] rows = table.findall("./w3:tr", namespaces=namespaces) data = [] for row in rows: data.append( ( _get_text(row[0].findall(".//w3:p", namespaces=namespaces)[-1]), _get_text(row[1].findall(".//w3:p", namespaces=namespaces)[0]), _get_text(row[2].findall(".//w3:p", namespaces=namespaces)[0]), ) ) return data def get_table_d1(): logger.info("process Table D1") url = "http://dicom.nema.org/medical/dicom/current/output/chtml/part16/chapter_D.html#table_D-1" root = _parse_html(_download_html(url)) namespaces = {"w3": root.tag.split("}")[0].strip("{")} body = root.find("w3:body", namespaces=namespaces) table = body.findall(".//w3:tbody", namespaces=namespaces)[0] rows = table.findall("./w3:tr", namespaces=namespaces) return [ ( _get_text(row[0].findall(".//w3:p", namespaces=namespaces)[0]), _get_text(row[1].findall(".//w3:p", namespaces=namespaces)[0]), ) for row in rows ] def write_concepts(concepts, cid_concepts, cid_lists, name_for_cid): lines = DOC_LINES + [ "# Dict with scheme designator keys; value format is:\n", "# {keyword: {code1: (meaning, cid_list), code2: ...}\n", "#\n", "# Most keyword identifiers map to a single code, but not all\n", "\n", ] with open("_concepts_dict.py", "w", encoding="UTF8") as f_concepts: f_concepts.writelines(lines) f_concepts.write("concepts = {}\n") # start with empty dict for scheme, value in concepts.items(): f_concepts.write(f"\nconcepts['{scheme}'] = \\\n") pprint(value, f_concepts) lines = DOC_LINES + [ "# Dict with cid number as keys; value format is:\n", "# {scheme designator: <list of keywords for current cid>\n", "# scheme_designator: ...}\n", "\n", ] with open("_cid_dict.py", "w", encoding="UTF8") as f_cid: f_cid.writelines(lines) f_cid.write("name_for_cid = {}\n") f_cid.write("cid_concepts = {}\n") for cid, value in cid_lists.items(): f_cid.write(f"\nname_for_cid[{cid}] = '{name_for_cid[cid]}'\n") f_cid.write(f"cid_concepts[{cid}] = \\\n") pprint(value, f_cid) def write_snomed_mapping(snomed_codes): with open("_snomed_dict.py", "w", encoding="UTF8") as f_concepts: lines = DOC_LINES + [ "# Dict with scheme designator keys; value format is:\n", "# {concept_id1: snomed_id1, concept_id2: ...}\n", "# or\n", "# {snomed_id1: concept_id1, snomed_id2: ...}\n", "\n", ] f_concepts.writelines(lines) f_concepts.write("mapping = {}\n") # start with empty dict f_concepts.write("\nmapping['{}'] = {{\n".format("SCT")) for sct, srt, meaning in snomed_codes: f_concepts.write(f" '{sct}': '{srt}',\n") f_concepts.write("}\n") f_concepts.write("\nmapping['{}'] = {{\n".format("SRT")) for sct, srt, meaning in snomed_codes:
me def download
conditional_block
generate_concept_dicts.py
", "http://hl7.org/fhir/sid/ndc": "NDC", "urn:iso:std:iso:11073:10101": "MDC", "doi:10.1016/S0735-1097(99)00126-6": "BARI", "http://www.nlm.nih.gov/research/umls": "UMLS", "http://pubchem.ncbi.nlm.nih.gov": "PUBCHEM_CID", "http://braininfo.rprc.washington.edu/aboutBrainInfo.aspx#NeuroNames": "NEU", "http://www.itis.gov": "ITIS_TSN", "http://arxiv.org/abs/1612.07003": "IBSI", "http://www.nlm.nih.gov/research/umls/rxnorm": "RXNORM", } DOC_LINES = [ f"# Auto-generated by {os.path.basename(__file__)}.\n", "# -*- coding: utf-8 -*-\n", "\n", ] def camel_case(s): leave_alone = ( "mm", "cm", "km", "um", "ms", # 'us'?-doesn't seem to be there "ml", "mg", "kg", ) # ... probably need others return "".join( word.capitalize() if word != word.upper() and word not in leave_alone else word for word in re.split(r"\W", s, flags=re.UNICODE) if word.isalnum() ) def keyword_from_meaning(name): """Return a camel case valid python identifier""" # Try to adhere to keyword scheme in DICOM (CP850) # singular/plural alternative forms are made plural # e.g., “Physician(s) of Record” becomes “PhysiciansOfRecord” name = name.replace("(s)", "s") # “Patient’s Name” -> “PatientName” # “Operators’ Name” -> “OperatorsName” name = name.replace("’s ", " ") name = name.replace("'s ", " ") name = name.replace("s’ ", "s ") name = name.replace("s' ", "s ") # Mathematical symbols name = name.replace("%", " Percent ") name = name.replace(">", " Greater Than ") name = name.replace("=", " Equals ") name = name.replace("<", " Lesser Than ") name = re.sub(r"([0-9]+)\.([0-9]+)", "\\1 Point \\2", name) name = re.sub(r"\s([0-9.]+)-([0-9.]+)\s", " \\1 To \\2 ", name) name = re.sub(r"([0-9]+)day", "\\1 Day", name) name = re.sub(r"([0-9]+)y", "\\1 Years", name) # Remove category modifiers, such as "(specimen)", "(procedure)", # "(body structure)", etc. name = re.sub(r"^(.+) \([a-z ]+\)$", "\\1", name) name = camel_case(name.strip()) # Python variables must not begin with a number. if re.match(r"[0-9]", name): name = "_" + name return name def download_fhir_value_sets(local_dir): ftp_host = "medical.nema.org" if not os.path.exists(local_dir): os.makedirs(local_dir) logger.info("storing files in " + local_dir) logger.info(f'log into FTP server "{ftp_host}"') ftp = ftplib.FTP(ftp_host, timeout=60) ftp.login("anonymous") ftp_path = "medical/dicom/resources/valuesets/fhir/json" logger.info(f'list files in directory "{ftp_path}"') fhir_value_set_files = ftp.nlst(ftp_path) try: for ftp_filepath in fhir_value_set_files: ftp_filename = os.path.basename(ftp_filepath) logger.info(f'retrieve value set file "{ftp_filename}"') with BytesIO() as fp: ftp.retrbinary(f"RETR {ftp_filepath}", fp.write) content = fp.getvalue() local_filename = os.path.join(local_dir, ftp_filename) with open(local_filename, "wb") as f_local: f_local.write(content) finally: ftp.quit() def _parse_html(content): # from lxml import html # doc = html.document_fromstring(content) return ET.fromstring(content, parser=ET.XMLParser(encoding="utf-8")) def _download_html(url): response = urllib_request.urlopen(url) return response.read() def _get_text(element): text = "".join(element.itertext()) return text.strip() def get_table_o1(): logger.info("process Table O1") url = "http://dicom.nema.org/medical/dicom/current/output/chtml/part16/chapter_O.html#table_O-1" root = _parse_html(_download_html(url)) namespaces = {"w3": root.tag.split("}")[0].strip("{")} body = root.find("w3:body", namespaces=namespaces) table = body.findall(".//w3:tbody", namespaces=namespaces)[0] rows = table.findall("./w3:tr", namespaces=namespaces) data = [] for row in rows: data.append( ( _get_text(row[0].findall(".//w3:p", namespaces=namespaces)[-1]), _get_text(row[1].findall(".//w3:p", namespaces=namespaces)[0]), _get_text(row[2].findall(".//w3:p", namespaces=namespaces)[0]), ) ) return data def get_table_d1(): logger.info(
le D1") url = "http://dicom.nema.org/medical/dicom/current/output/chtml/part16/chapter_D.html#table_D-1" root = _parse_html(_download_html(url)) namespaces = {"w3": root.tag.split("}")[0].strip("{")} body = root.find("w3:body", namespaces=namespaces) table = body.findall(".//w3:tbody", namespaces=namespaces)[0] rows = table.findall("./w3:tr", namespaces=namespaces) return [ ( _get_text(row[0].findall(".//w3:p", namespaces=namespaces)[0]), _get_text(row[1].findall(".//w3:p", namespaces=namespaces)[0]), ) for row in rows ] def write_concepts(concepts, cid_concepts, cid_lists, name_for_cid): lines = DOC_LINES + [ "# Dict with scheme designator keys; value format is:\n", "# {keyword: {code1: (meaning, cid_list), code2: ...}\n", "#\n", "# Most keyword identifiers map to a single code, but not all\n", "\n", ] with open("_concepts_dict.py", "w", encoding="UTF8") as f_concepts: f_concepts.writelines(lines) f_concepts.write("concepts = {}\n") # start with empty dict for scheme, value in concepts.items(): f_concepts.write(f"\nconcepts['{scheme}'] = \\\n") pprint(value, f_concepts) lines = DOC_LINES + [ "# Dict with cid number as keys; value format is:\n", "# {scheme designator: <list of keywords for current cid>\n", "# scheme_designator: ...}\n", "\n", ] with open("_cid_dict.py", "w", encoding="UTF8") as f_cid: f_cid.writelines(lines) f_cid.write("name_for_cid = {}\n") f_cid.write("cid_concepts = {}\n") for cid, value in cid_lists.items(): f_cid.write(f"\nname_for_cid[{cid}] = '{name_for_cid[cid]}'\n") f_cid.write(f"cid_concepts[{cid}] = \\\n") pprint(value, f_cid) def write_snomed_mapping(snomed_codes): with open("_snomed_dict.py", "w", encoding="UTF8") as f_concepts: lines = DOC_LINES + [ "# Dict with scheme designator keys; value format is:\n", "# {concept_id1: snomed_id1, concept_id2: ...}\n", "# or\n", "# {snomed_id1: concept_id1, snomed_id2: ...}\n", "\n", ] f_concepts.writelines(lines) f_concepts.write("mapping = {}\n") # start with empty dict f_concepts.write("\nmapping['{}'] = {{\n".format("SCT")) for sct, srt, meaning in snomed_codes: f_concepts.write(f" '{sct}': '{srt}',\n") f_concepts.write("}\n") f_concepts.write("\nmapping['{}'] = {{\n".format("SRT")) for sct, srt, meaning in snomed_codes:
"process Tab
identifier_name
lobby.go
{ log.Panic(err) } } /* Game table in DB will need Host and Port stored */ func newSession(g Game, sessId int, host *HostUser) (*Session, error) { session := Session{ sessionId: sessId, game: g, LobbyHost: host, userMap: make(map[string]*AnonUser), PlayerMap: make([]User, 0, g.MaxUsers() + 1), //PlayerMap: make(map[int]*User, g.MaxUsers() + 1), Send: make(chan []byte), Receive: make(chan interface{}), Data: make(chan interface{}), Exit: make(chan bool, 1), } host.SetSession(&session) host.SetPlayer(0) session.PlayerMap = append(session.PlayerMap, session.LobbyHost) //set index 0 as host //session.PlayerMap[0] = session.LobbyHost go session.dataHandler() return &session, nil } func
(uId int, user string) (*HostUser, error) { hostUser := HostUser{ userId: uId, username: user, Send: make(chan interface{}), Receive: make(chan interface{}), } return &hostUser, nil } func newAnonUser(nick string) *AnonUser { anon := AnonUser{ Nickname: nick, Ready: false, Send: make(chan interface{}), Receive: make(chan interface{}), } return &anon } func (s Session) SessionId() int { return s.sessionId } /* Adds an anon user to the session, and sets user's pointer to the session */ func (s Session) addAnonUser(id string, anon *AnonUser) error { s.Lock() defer s.Unlock() _, ok := s.userMap[id] if !ok { return errors.New("lobby: user already exists.") } if len(s.PlayerMap) >= cap(s.PlayerMap) { return errors.New("lobby: lobby is full.") } s.userMap[id] = anon s.PlayerMap = append(s.PlayerMap, anon) anon.SetPlayer(len(s.PlayerMap) - 1) anon.SetSession(&s) s.emitAnonUserData() return nil } func (s Session) removeAnonUser(id, reason string) error { s.Lock() defer s.Unlock() u, ok := s.userMap[id] if !ok { return errors.New("lobby: user does not exist.") } index := u.Player() //user, ok := s.PlayerMap[index] //if !ok { //return errors.New("removePlayer: no player found at index") //} //if u != user { //return errors.New("removePlayer: player does not match retrieved") //} //delete(s.PlayerMap, index) s.removePlayer(index) delete(s.userMap, id) var msg Command if reason == "" { msg.Cmd = C_Leave } else { msg.Cmd = C_Kick msg.Data = reason } u.Send <- msg s.emitAnonUserData() return nil } /* Removes a player from the PlayerMap slice - No Memory Leak: Example: Removing index 2 from: [0][1][2][3][4] Append: [0][1] [3][4] Final memory block still has redundant data: [0][1][3][4][4] Overwrite with nil: [0][1][3][4][nil] */ func (s Session) removePlayer(index int) { s.PlayerMap, s.PlayerMap[len(s.PlayerMap)-1] = append(s.PlayerMap[:index], s.PlayerMap[index+1:]...), nil //Update all player numbers greater than deleted index for i := index; i < len(s.PlayerMap); i++ { s.PlayerMap[i].SetPlayer(i) } } func (s Session) connectSession(addr string) error { tcpAddr, err := net.ResolveTCPAddr("tcp", addr) if err != nil { return err } conn, err := net.DialTCP("tcp", nil, tcpAddr) if err != nil { return err } s.gameTCPConn = conn go s.receiveHandler() go s.sendHandler() return nil } func (s Session) requestSession() { addr := s.game.host + ":" + s.game.port err := s.connectSession(addr) if err != nil { //return nil, err log.Print(err) } request := make(map[string]interface{}) request["event"] = "new" request["players"] = s.CurrentUserCount() request["maxplayers"] = s.game.MaxUsers() jsonMsg, err := json.Marshal(request) if err != nil { log.Print(err) } s.Send <- jsonMsg //MAKE A TIMEOUT go s.requestTimeout() go s.sendTimeout() } func (s Session) sendTimeout() { time.Sleep(10 * time.Second) s.timeout <- true } func (s Session) requestTimeout() { s.timeout = make(chan bool, 1) var gs GameStart select { case t := <- s.timeout: if t { //server timed out gs = GameStart{ Response: false, Feedback: "Server was unable to host.", } } else { //game created gs = GameStart{ Response: true, Feedback: "Application started succesfully.", } } } s.LobbyHost.Send <- gs } func (s Session) AnonUserById(id string) (*AnonUser, bool) { a, ok := s.userMap[id] return a, ok } func (s Session) CurrentUserCount() int { return len(s.userMap) } /* Flip ready bool */ func (s Session) setAnonUserReady(n string, r bool) { s.Lock() defer s.Unlock() s.userMap[n].Ready = r s.emitAnonUserData() } func (s Session) emitAnonUserData() { var list []LobbyUser players := s.PlayerMap[1:] //slice out host index /* for i := 1; i < len(s.PlayerMap); i++ { p := s.PlayerMap[i] user := LobbyUser{ Player: i, Nickname: p.Nickname Ready: p.Ready } list = append(list, user) } */ for i, p := range players { p := p.(AnonUser) user := LobbyUser{ Player: i, Nickname: p.Nickname, Ready: p.Ready, } list = append(list, user) } s.LobbyHost.Send <- list } /* Main goroutine for handling messages to the game server */ func (s Session) sendHandler() { for { select { case data := <-s.Send: s.gameTCPConn.Write(data) case <- s.Exit: return } } } /* Main goroutine for handling messages from the game server */ func (s Session) receiveHandler() { decoder := json.NewDecoder(s.gameTCPConn) for { var gMsg GameMessage err := decoder.Decode(&gMsg) if err != nil { log.Print(err) } //check event switch gMsg.Event { case "created": s.Data <- gMsg case "gamestart": s.LobbyHost.Send <- gMsg case "gameterminate": s.Data <- gMsg case "msgplayer": if gMsg.Player == 0 { user := s.PlayerMap[gMsg.Player].(HostUser) user.Send <- gMsg } else { user := s.PlayerMap[gMsg.Player].(AnonUser) user.Send <- gMsg } case "msgall": s.LobbyHost.Send <- gMsg } } } /* Main goroutine for processing lobby commands */ func (s Session) dataHandler() { for { select { case data := <-s.Data: switch jsonType := data.(type) { case SetReady: s.setAnonUserReady(jsonType.Nickname, jsonType.Ready) case RemovedUser: err := s.removeAnonUser(jsonType.Nickname, jsonType.Reason) if err != nil { log.Panic(err) } case GameMessage: switch jsonType.Event { case "created": s.timeout <- false log.Printf("Game successfully created: %s", jsonType.Msg) } default: log.Print("Session dataHandler: unknown type received") } case <- s.Exit: return } } } func (u HostUser) UserId() int { return u.userId } func (u HostUser) SetSession(s *Session) { u.Sess = s } func (u HostUser) SetSocket(sock socketio.Socket) { u.socket = sock } func (u HostUser) Player() int { return u.player } /* Only to be called while Session is locked */ func (u HostUser) SetPlayer(number int) { u.player = number } /* Joins the user's socket namespace, and the session namespace */ func (u HostUser) Setup() { //u
newHostUser
identifier_name
lobby.go
{ log.Panic(err) } } /* Game table in DB will need Host and Port stored */ func newSession(g Game, sessId int, host *HostUser) (*Session, error) { session := Session{ sessionId: sessId, game: g, LobbyHost: host, userMap: make(map[string]*AnonUser), PlayerMap: make([]User, 0, g.MaxUsers() + 1), //PlayerMap: make(map[int]*User, g.MaxUsers() + 1), Send: make(chan []byte), Receive: make(chan interface{}), Data: make(chan interface{}), Exit: make(chan bool, 1), } host.SetSession(&session) host.SetPlayer(0) session.PlayerMap = append(session.PlayerMap, session.LobbyHost) //set index 0 as host //session.PlayerMap[0] = session.LobbyHost go session.dataHandler() return &session, nil } func newHostUser(uId int, user string) (*HostUser, error) { hostUser := HostUser{ userId: uId, username: user, Send: make(chan interface{}), Receive: make(chan interface{}), } return &hostUser, nil } func newAnonUser(nick string) *AnonUser { anon := AnonUser{ Nickname: nick, Ready: false, Send: make(chan interface{}), Receive: make(chan interface{}), } return &anon } func (s Session) SessionId() int { return s.sessionId } /* Adds an anon user to the session, and sets user's pointer to the session */ func (s Session) addAnonUser(id string, anon *AnonUser) error { s.Lock() defer s.Unlock() _, ok := s.userMap[id] if !ok { return errors.New("lobby: user already exists.") } if len(s.PlayerMap) >= cap(s.PlayerMap) { return errors.New("lobby: lobby is full.") } s.userMap[id] = anon s.PlayerMap = append(s.PlayerMap, anon) anon.SetPlayer(len(s.PlayerMap) - 1) anon.SetSession(&s) s.emitAnonUserData() return nil } func (s Session) removeAnonUser(id, reason string) error { s.Lock() defer s.Unlock() u, ok := s.userMap[id] if !ok { return errors.New("lobby: user does not exist.") } index := u.Player() //user, ok := s.PlayerMap[index] //if !ok { //return errors.New("removePlayer: no player found at index") //} //if u != user { //return errors.New("removePlayer: player does not match retrieved") //} //delete(s.PlayerMap, index) s.removePlayer(index) delete(s.userMap, id) var msg Command if reason == "" { msg.Cmd = C_Leave } else { msg.Cmd = C_Kick msg.Data = reason } u.Send <- msg s.emitAnonUserData() return nil } /* Removes a player from the PlayerMap slice - No Memory Leak: Example: Removing index 2 from: [0][1][2][3][4] Append: [0][1] [3][4] Final memory block still has redundant data: [0][1][3][4][4] Overwrite with nil: [0][1][3][4][nil] */ func (s Session) removePlayer(index int) { s.PlayerMap, s.PlayerMap[len(s.PlayerMap)-1] = append(s.PlayerMap[:index], s.PlayerMap[index+1:]...), nil //Update all player numbers greater than deleted index for i := index; i < len(s.PlayerMap); i++ { s.PlayerMap[i].SetPlayer(i) } } func (s Session) connectSession(addr string) error { tcpAddr, err := net.ResolveTCPAddr("tcp", addr) if err != nil { return err } conn, err := net.DialTCP("tcp", nil, tcpAddr) if err != nil { return err } s.gameTCPConn = conn go s.receiveHandler() go s.sendHandler() return nil } func (s Session) requestSession() { addr := s.game.host + ":" + s.game.port err := s.connectSession(addr) if err != nil { //return nil, err log.Print(err) } request := make(map[string]interface{}) request["event"] = "new" request["players"] = s.CurrentUserCount() request["maxplayers"] = s.game.MaxUsers() jsonMsg, err := json.Marshal(request) if err != nil { log.Print(err) } s.Send <- jsonMsg //MAKE A TIMEOUT go s.requestTimeout() go s.sendTimeout() } func (s Session) sendTimeout() { time.Sleep(10 * time.Second) s.timeout <- true } func (s Session) requestTimeout() { s.timeout = make(chan bool, 1) var gs GameStart select { case t := <- s.timeout: if t { //server timed out gs = GameStart{ Response: false, Feedback: "Server was unable to host.", } } else { //game created gs = GameStart{ Response: true, Feedback: "Application started succesfully.", } } } s.LobbyHost.Send <- gs } func (s Session) AnonUserById(id string) (*AnonUser, bool) { a, ok := s.userMap[id] return a, ok } func (s Session) CurrentUserCount() int { return len(s.userMap) } /* Flip ready bool */ func (s Session) setAnonUserReady(n string, r bool) { s.Lock() defer s.Unlock() s.userMap[n].Ready = r s.emitAnonUserData() } func (s Session) emitAnonUserData() { var list []LobbyUser players := s.PlayerMap[1:] //slice out host index /* for i := 1; i < len(s.PlayerMap); i++ { p := s.PlayerMap[i] user := LobbyUser{ Player: i, Nickname: p.Nickname Ready: p.Ready } list = append(list, user) } */ for i, p := range players { p := p.(AnonUser) user := LobbyUser{ Player: i, Nickname: p.Nickname, Ready: p.Ready, } list = append(list, user) } s.LobbyHost.Send <- list } /* Main goroutine for handling messages to the game server */ func (s Session) sendHandler() { for { select { case data := <-s.Send: s.gameTCPConn.Write(data) case <- s.Exit: return } } } /* Main goroutine for handling messages from the game server */ func (s Session) receiveHandler() { decoder := json.NewDecoder(s.gameTCPConn) for { var gMsg GameMessage err := decoder.Decode(&gMsg) if err != nil { log.Print(err) } //check event switch gMsg.Event { case "created": s.Data <- gMsg case "gamestart": s.LobbyHost.Send <- gMsg case "gameterminate": s.Data <- gMsg case "msgplayer": if gMsg.Player == 0
else { user := s.PlayerMap[gMsg.Player].(AnonUser) user.Send <- gMsg } case "msgall": s.LobbyHost.Send <- gMsg } } } /* Main goroutine for processing lobby commands */ func (s Session) dataHandler() { for { select { case data := <-s.Data: switch jsonType := data.(type) { case SetReady: s.setAnonUserReady(jsonType.Nickname, jsonType.Ready) case RemovedUser: err := s.removeAnonUser(jsonType.Nickname, jsonType.Reason) if err != nil { log.Panic(err) } case GameMessage: switch jsonType.Event { case "created": s.timeout <- false log.Printf("Game successfully created: %s", jsonType.Msg) } default: log.Print("Session dataHandler: unknown type received") } case <- s.Exit: return } } } func (u HostUser) UserId() int { return u.userId } func (u HostUser) SetSession(s *Session) { u.Sess = s } func (u HostUser) SetSocket(sock socketio.Socket) { u.socket = sock } func (u HostUser) Player() int { return u.player } /* Only to be called while Session is locked */ func (u HostUser) SetPlayer(number int) { u.player = number } /* Joins the user's socket namespace, and the session namespace */ func (u HostUser) Setup() { //u
{ user := s.PlayerMap[gMsg.Player].(HostUser) user.Send <- gMsg }
conditional_block
lobby.go
("removePlayer: player does not match retrieved") //} //delete(s.PlayerMap, index) s.removePlayer(index) delete(s.userMap, id) var msg Command if reason == "" { msg.Cmd = C_Leave } else { msg.Cmd = C_Kick msg.Data = reason } u.Send <- msg s.emitAnonUserData() return nil } /* Removes a player from the PlayerMap slice - No Memory Leak: Example: Removing index 2 from: [0][1][2][3][4] Append: [0][1] [3][4] Final memory block still has redundant data: [0][1][3][4][4] Overwrite with nil: [0][1][3][4][nil] */ func (s Session) removePlayer(index int) { s.PlayerMap, s.PlayerMap[len(s.PlayerMap)-1] = append(s.PlayerMap[:index], s.PlayerMap[index+1:]...), nil //Update all player numbers greater than deleted index for i := index; i < len(s.PlayerMap); i++ { s.PlayerMap[i].SetPlayer(i) } } func (s Session) connectSession(addr string) error { tcpAddr, err := net.ResolveTCPAddr("tcp", addr) if err != nil { return err } conn, err := net.DialTCP("tcp", nil, tcpAddr) if err != nil { return err } s.gameTCPConn = conn go s.receiveHandler() go s.sendHandler() return nil } func (s Session) requestSession() { addr := s.game.host + ":" + s.game.port err := s.connectSession(addr) if err != nil { //return nil, err log.Print(err) } request := make(map[string]interface{}) request["event"] = "new" request["players"] = s.CurrentUserCount() request["maxplayers"] = s.game.MaxUsers() jsonMsg, err := json.Marshal(request) if err != nil { log.Print(err) } s.Send <- jsonMsg //MAKE A TIMEOUT go s.requestTimeout() go s.sendTimeout() } func (s Session) sendTimeout() { time.Sleep(10 * time.Second) s.timeout <- true } func (s Session) requestTimeout() { s.timeout = make(chan bool, 1) var gs GameStart select { case t := <- s.timeout: if t { //server timed out gs = GameStart{ Response: false, Feedback: "Server was unable to host.", } } else { //game created gs = GameStart{ Response: true, Feedback: "Application started succesfully.", } } } s.LobbyHost.Send <- gs } func (s Session) AnonUserById(id string) (*AnonUser, bool) { a, ok := s.userMap[id] return a, ok } func (s Session) CurrentUserCount() int { return len(s.userMap) } /* Flip ready bool */ func (s Session) setAnonUserReady(n string, r bool) { s.Lock() defer s.Unlock() s.userMap[n].Ready = r s.emitAnonUserData() } func (s Session) emitAnonUserData() { var list []LobbyUser players := s.PlayerMap[1:] //slice out host index /* for i := 1; i < len(s.PlayerMap); i++ { p := s.PlayerMap[i] user := LobbyUser{ Player: i, Nickname: p.Nickname Ready: p.Ready } list = append(list, user) } */ for i, p := range players { p := p.(AnonUser) user := LobbyUser{ Player: i, Nickname: p.Nickname, Ready: p.Ready, } list = append(list, user) } s.LobbyHost.Send <- list } /* Main goroutine for handling messages to the game server */ func (s Session) sendHandler() { for { select { case data := <-s.Send: s.gameTCPConn.Write(data) case <- s.Exit: return } } } /* Main goroutine for handling messages from the game server */ func (s Session) receiveHandler() { decoder := json.NewDecoder(s.gameTCPConn) for { var gMsg GameMessage err := decoder.Decode(&gMsg) if err != nil { log.Print(err) } //check event switch gMsg.Event { case "created": s.Data <- gMsg case "gamestart": s.LobbyHost.Send <- gMsg case "gameterminate": s.Data <- gMsg case "msgplayer": if gMsg.Player == 0 { user := s.PlayerMap[gMsg.Player].(HostUser) user.Send <- gMsg } else { user := s.PlayerMap[gMsg.Player].(AnonUser) user.Send <- gMsg } case "msgall": s.LobbyHost.Send <- gMsg } } } /* Main goroutine for processing lobby commands */ func (s Session) dataHandler() { for { select { case data := <-s.Data: switch jsonType := data.(type) { case SetReady: s.setAnonUserReady(jsonType.Nickname, jsonType.Ready) case RemovedUser: err := s.removeAnonUser(jsonType.Nickname, jsonType.Reason) if err != nil { log.Panic(err) } case GameMessage: switch jsonType.Event { case "created": s.timeout <- false log.Printf("Game successfully created: %s", jsonType.Msg) } default: log.Print("Session dataHandler: unknown type received") } case <- s.Exit: return } } } func (u HostUser) UserId() int { return u.userId } func (u HostUser) SetSession(s *Session) { u.Sess = s } func (u HostUser) SetSocket(sock socketio.Socket) { u.socket = sock } func (u HostUser) Player() int { return u.player } /* Only to be called while Session is locked */ func (u HostUser) SetPlayer(number int) { u.player = number } /* Joins the user's socket namespace, and the session namespace */ func (u HostUser) Setup() { //u.socket.Join(u.username) //not necessary socket ID namespace u.socket.Join(fmt.Sprintf("%d", u.Sess.SessionId())) go u.sendHandler() u.receiveHandler() } /* Emits socket.io messages to the namespace */ func (u HostUser) sendHandler() { sessionNamespace := fmt.Sprintf("%d", u.Sess.SessionId()) for { select { case data := <-u.Send: switch dataType := data.(type) { case []LobbyUser: /*msg, err := json.Marshal(dataType) if err != nil { log.Panic("send lobby user list: error") } */ u.socket.BroadcastTo(sessionNamespace, "updatelobby", dataType) u.socket.Emit("updatelobby", dataType) case GameStart: /*msg, err := json.Marshal(dataType) if err != nil { log.Panic("unable to marshal message") } */ u.socket.BroadcastTo(sessionNamespace, "gamestart", dataType) u.socket.Emit("gamestart", dataType) case GameMessage: /*msg, err := json.Marshal(dataType.Msg) if err != nil { log.Panic("unable to marshal message") } */ switch dataType.Event{ case "msgplayer": u.socket.Emit("msgplayer", dataType.Msg) case "msgall": u.socket.BroadcastTo(sessionNamespace, "msgall", dataType.Msg) u.socket.Emit("msgall", dataType.Msg) } default: log.Print("HostUser sendHandler: unknown type received") } } } } /* Main goroutine for handling messages for host user */ func (u HostUser) receiveHandler() { //Tell server the applet has loaded and ready to communicate //Used to initially ping server and pass any preliminary host information //u.socket.Of(fmt.Sprintf("/%s", u.socket.Id())).On("kick", func(msg []byte) { u.socket.On("kick", func(msg map[string]interface{}) { var data RemovedUser /*err := json.Unmarshal(msg, &data) if err != nil { log.Panic(err) } */ data = RemovedUser{ Nickname: msg["nickname"].(string), Reason: msg["reason"].(string), } u.Sess.Data <- data }) //launch the game with the current users in lobby, server should respond if successful u.socket.On("start", func() { start := Command{ Cmd: C_Start, } jsonMsg, err := json.Marshal(start) if err != nil { log.Print(err) } u.Sess.Send <- jsonMsg }) //send to game server, server should give response to be emitted
random_line_split
lobby.go
{ log.Panic(err) } } /* Game table in DB will need Host and Port stored */ func newSession(g Game, sessId int, host *HostUser) (*Session, error) { session := Session{ sessionId: sessId, game: g, LobbyHost: host, userMap: make(map[string]*AnonUser), PlayerMap: make([]User, 0, g.MaxUsers() + 1), //PlayerMap: make(map[int]*User, g.MaxUsers() + 1), Send: make(chan []byte), Receive: make(chan interface{}), Data: make(chan interface{}), Exit: make(chan bool, 1), } host.SetSession(&session) host.SetPlayer(0) session.PlayerMap = append(session.PlayerMap, session.LobbyHost) //set index 0 as host //session.PlayerMap[0] = session.LobbyHost go session.dataHandler() return &session, nil } func newHostUser(uId int, user string) (*HostUser, error) { hostUser := HostUser{ userId: uId, username: user, Send: make(chan interface{}), Receive: make(chan interface{}), } return &hostUser, nil } func newAnonUser(nick string) *AnonUser { anon := AnonUser{ Nickname: nick, Ready: false, Send: make(chan interface{}), Receive: make(chan interface{}), } return &anon } func (s Session) SessionId() int { return s.sessionId } /* Adds an anon user to the session, and sets user's pointer to the session */ func (s Session) addAnonUser(id string, anon *AnonUser) error { s.Lock() defer s.Unlock() _, ok := s.userMap[id] if !ok { return errors.New("lobby: user already exists.") } if len(s.PlayerMap) >= cap(s.PlayerMap) { return errors.New("lobby: lobby is full.") } s.userMap[id] = anon s.PlayerMap = append(s.PlayerMap, anon) anon.SetPlayer(len(s.PlayerMap) - 1) anon.SetSession(&s) s.emitAnonUserData() return nil } func (s Session) removeAnonUser(id, reason string) error { s.Lock() defer s.Unlock() u, ok := s.userMap[id] if !ok { return errors.New("lobby: user does not exist.") } index := u.Player() //user, ok := s.PlayerMap[index] //if !ok { //return errors.New("removePlayer: no player found at index") //} //if u != user { //return errors.New("removePlayer: player does not match retrieved") //} //delete(s.PlayerMap, index) s.removePlayer(index) delete(s.userMap, id) var msg Command if reason == "" { msg.Cmd = C_Leave } else { msg.Cmd = C_Kick msg.Data = reason } u.Send <- msg s.emitAnonUserData() return nil } /* Removes a player from the PlayerMap slice - No Memory Leak: Example: Removing index 2 from: [0][1][2][3][4] Append: [0][1] [3][4] Final memory block still has redundant data: [0][1][3][4][4] Overwrite with nil: [0][1][3][4][nil] */ func (s Session) removePlayer(index int) { s.PlayerMap, s.PlayerMap[len(s.PlayerMap)-1] = append(s.PlayerMap[:index], s.PlayerMap[index+1:]...), nil //Update all player numbers greater than deleted index for i := index; i < len(s.PlayerMap); i++ { s.PlayerMap[i].SetPlayer(i) } } func (s Session) connectSession(addr string) error { tcpAddr, err := net.ResolveTCPAddr("tcp", addr) if err != nil { return err } conn, err := net.DialTCP("tcp", nil, tcpAddr) if err != nil { return err } s.gameTCPConn = conn go s.receiveHandler() go s.sendHandler() return nil } func (s Session) requestSession()
func (s Session) sendTimeout() { time.Sleep(10 * time.Second) s.timeout <- true } func (s Session) requestTimeout() { s.timeout = make(chan bool, 1) var gs GameStart select { case t := <- s.timeout: if t { //server timed out gs = GameStart{ Response: false, Feedback: "Server was unable to host.", } } else { //game created gs = GameStart{ Response: true, Feedback: "Application started succesfully.", } } } s.LobbyHost.Send <- gs } func (s Session) AnonUserById(id string) (*AnonUser, bool) { a, ok := s.userMap[id] return a, ok } func (s Session) CurrentUserCount() int { return len(s.userMap) } /* Flip ready bool */ func (s Session) setAnonUserReady(n string, r bool) { s.Lock() defer s.Unlock() s.userMap[n].Ready = r s.emitAnonUserData() } func (s Session) emitAnonUserData() { var list []LobbyUser players := s.PlayerMap[1:] //slice out host index /* for i := 1; i < len(s.PlayerMap); i++ { p := s.PlayerMap[i] user := LobbyUser{ Player: i, Nickname: p.Nickname Ready: p.Ready } list = append(list, user) } */ for i, p := range players { p := p.(AnonUser) user := LobbyUser{ Player: i, Nickname: p.Nickname, Ready: p.Ready, } list = append(list, user) } s.LobbyHost.Send <- list } /* Main goroutine for handling messages to the game server */ func (s Session) sendHandler() { for { select { case data := <-s.Send: s.gameTCPConn.Write(data) case <- s.Exit: return } } } /* Main goroutine for handling messages from the game server */ func (s Session) receiveHandler() { decoder := json.NewDecoder(s.gameTCPConn) for { var gMsg GameMessage err := decoder.Decode(&gMsg) if err != nil { log.Print(err) } //check event switch gMsg.Event { case "created": s.Data <- gMsg case "gamestart": s.LobbyHost.Send <- gMsg case "gameterminate": s.Data <- gMsg case "msgplayer": if gMsg.Player == 0 { user := s.PlayerMap[gMsg.Player].(HostUser) user.Send <- gMsg } else { user := s.PlayerMap[gMsg.Player].(AnonUser) user.Send <- gMsg } case "msgall": s.LobbyHost.Send <- gMsg } } } /* Main goroutine for processing lobby commands */ func (s Session) dataHandler() { for { select { case data := <-s.Data: switch jsonType := data.(type) { case SetReady: s.setAnonUserReady(jsonType.Nickname, jsonType.Ready) case RemovedUser: err := s.removeAnonUser(jsonType.Nickname, jsonType.Reason) if err != nil { log.Panic(err) } case GameMessage: switch jsonType.Event { case "created": s.timeout <- false log.Printf("Game successfully created: %s", jsonType.Msg) } default: log.Print("Session dataHandler: unknown type received") } case <- s.Exit: return } } } func (u HostUser) UserId() int { return u.userId } func (u HostUser) SetSession(s *Session) { u.Sess = s } func (u HostUser) SetSocket(sock socketio.Socket) { u.socket = sock } func (u HostUser) Player() int { return u.player } /* Only to be called while Session is locked */ func (u HostUser) SetPlayer(number int) { u.player = number } /* Joins the user's socket namespace, and the session namespace */ func (u HostUser) Setup() { //
{ addr := s.game.host + ":" + s.game.port err := s.connectSession(addr) if err != nil { //return nil, err log.Print(err) } request := make(map[string]interface{}) request["event"] = "new" request["players"] = s.CurrentUserCount() request["maxplayers"] = s.game.MaxUsers() jsonMsg, err := json.Marshal(request) if err != nil { log.Print(err) } s.Send <- jsonMsg //MAKE A TIMEOUT go s.requestTimeout() go s.sendTimeout() }
identifier_body
session.go
nil { s.handler.HandleSessionError(err) return // go home handler, you're drunk! } } if err := s.respondOK(strings.Join(helloParts, " ")); err != nil { s.handler.HandleSessionError(err) return // communication problem, most likely? } for { if keepGoing := s.serveOne(); !keepGoing { return } } } // serveOne handles each command-response interaction with the client. The // boolean return value indicates whether the communication with the client // should continue or not. func (s *session) serveOne() bool { if s.state == stateTerminateConnection { return false } readBy := time.Now().Add(s.server.Timeout) if err := s.conn.SetReadDeadline(readBy); err != nil { return s.handleError(err, false) } line, err := s.reader.ReadLine() if err != nil { return s.handleError(err, false) // communication problem, most likely? } args := strings.Split(line, " ") command := strings.ToUpper(args[0]) cmdValidator, exists := validators[command] if !exists { return s.handleError(errInvalidSyntax, true) // unknown command } if err := cmdValidator.validate(s, args[1:]); err != nil { return s.handleError(err, true) } return s.handleError(operationHandlers[command](s, args[1:]), true) } // handleCAPA is a callback for capability listing. // RFC 2449, page 2. func (s *session) handleCAPA(args []string) error { if err := s.respondOK("Capability list follows"); err != nil { return err } dotWriter := s.writer.DotWriter() defer s.closeOrReport(dotWriter) for _, capability := range s.server.capabilities { if _, err := fmt.Fprintln(dotWriter, capability); err != nil { return err } } return nil } // handleAPOP is a callback for an APOP authentication mechanism. // RFC 1939, page 15. func (s *session) handleAPOP(args []string) error { if !s.server.APOP { return NewReportableError("server does not support APOP") } if err := s.handler.AuthenticateAPOP(args[0], args[1]); err != nil { return err } return s.signIn() } // handleDELE is a callback for a single message deletion. // RFC 1939, page 8. func (s *session) handleDELE(args []string) error { return s.withMessageDo(args[0], func(msgId uint64) error { s.markedDeleted[msgId] = struct{}{} return s.respondOK("message %d deleted", msgId) }) } // handleAPOP is a callback for listing one or more messages. // RFC 1939, page 6. func (s *session) handleLIST(args []string) error { if len(args) == 1 { return s.withMessageDo(args[0], func(msgId uint64) error { return s.respondOK("%d %d", msgId, s.msgSizes[msgId]) }) } return s.forEachMessage(func(msgId uint64) (string, error) { return fmt.Sprintf("%d %d", msgId, s.msgSizes[msgId]), nil }) } // handleNOOP is a callback for a no-op (timeout reset) command. // RFC 1939, page 9. func (s *session) handleNOOP(args []string) error { return s.respondOK("doing nothing") } // handlePASS is a callback for the client providing password ("PASS" command). // This must have been preceded by a "USER" command where the client provides // its username. // RFC 1939, page 14. func (s *session) handlePASS(args []string) error { if s.username == ""
if err := s.handler.AuthenticatePASS(s.username, args[0]); err != nil { return err } return s.signIn() } // handleQUIT is a callback for the client terminating the session. It will do // slightly different things depending on the current state of the transaction. // RFC 1939, pages 5 (in authorization state) and 10 (in transaction state). func (s *session) handleQUIT(args []string) error { bye := func() error { s.state = stateTerminateConnection return s.respondOK("dewey POP3 server signing off") } if s.state == stateAuthorization { return bye() } s.state = stateUpdate // so that no future calls will succeed var delMsg []uint64 for key := range s.markedDeleted { delMsg = append(delMsg, key) } if err := s.handler.DeleteMessages(delMsg); err != nil { return err } return bye() } // handleRETR is a callback for the client requesting full content of a a single // message. // RFC 1939, page 8. func (s *session) handleRETR(args []string) (err error) { return s.withMessageDo(args[0], func(msgId uint64) error { if err := s.respondOK("%d octets", s.msgSizes[msgId]); err != nil { return err } readCloser, err := s.handler.GetMessageReader(msgId) if err != nil { return err } defer s.closeOrReport(readCloser) dotWriter := s.writer.DotWriter() defer s.closeOrReport(dotWriter) _, err = io.Copy(dotWriter, readCloser) return err }) } // handleRSET is a callback for the client requesting the session to be reset. // This essentially means undeleting all messages previously marked for // deletion. // RFC 1939, page 9. func (s *session) handleRSET(args []string) error { s.markedDeleted = make(map[uint64]struct{}) return s.respondOK( "maildrop has %d messages (%d octets)", s.getMessageCount(), s.getMaildropSize(), ) } // handleRETR is a callback for the client requesting full content of a a single // message. // RFC 1939, page 8. func (s *session) handleSTAT(args []string) error { return s.respondOK("%d %d", s.getMessageCount(), s.getMaildropSize()) } // handleTOP is a callback for the client requesting a number of lines from the // top of a single message. // RFC 1939, page 11. func (s *session) handleTOP(args []string) error { return s.withMessageDo(args[0], func(msgId uint64) error { noLines, err := strconv.ParseUint(args[1], 10, 64) if err != nil { return errInvalidSyntax } if err := s.writer.PrintfLine("+OK"); err != nil { return err } readCloser, err := s.handler.GetMessageReader(msgId) if err != nil { return err } defer s.closeOrReport(readCloser) dotWriter := s.writer.DotWriter() defer s.closeOrReport(dotWriter) protoReader := textproto.NewReader(bufio.NewReader(readCloser)) for i := uint64(0); i < noLines; i++ { line, readErr := protoReader.ReadLineBytes() if err := printTopLine(line, readErr, dotWriter); err != nil { return err } } return nil }) } func printTopLine(line []byte, readErr error, writer io.Writer) error { if readErr == io.EOF || readErr == nil { if err := writeWithError(writer, line); err != nil { return err } } if readErr != nil { return readErr } return writeWithError(writer, []byte{'\n'}) } func writeWithError(w io.Writer, content []byte) error { _, err := w.Write(content) return err } // handleUIDL is a callback for the client unique message identifiers for // either one or all messages. // RFC 1939, page 12. func (s *session) handleUIDL(args []string) (err error) { if len(args) == 1 { return s.withMessageDo(args[0], func(msgId uint64) error { uidl, err := s.handler.GetMessageID(msgId) if err != nil { return err } return s.respondOK("%d %s", msgId, uidl) }) } return s.forEachMessage(func(msgId uint64) (string, error) { uidl, err := s.handler.GetMessageID(msgId) if err != nil { return "", err } return fmt.Sprintf("%d %s", msgId, uidl), nil }) } // handleUSER is a callback for the client providing it's username. This must be // followed by a "PASS" command with a corresponding password. // RFC 1939, page 13. func (s *session) handleUSER(args
{ return NewReportableError("please provide username first") }
conditional_block
session.go
const ( stateAuthorization = iota stateTransaction stateUpdate stateTerminateConnection ) type operationHandler func(s *session, args []string) error var ( operationHandlers = map[string]operationHandler{ "APOP": (*session).handleAPOP, "CAPA": (*session).handleCAPA, "DELE": (*session).handleDELE, "LIST": (*session).handleLIST, "NOOP": (*session).handleNOOP, "PASS": (*session).handlePASS, "QUIT": (*session).handleQUIT, "RETR": (*session).handleRETR, "RSET": (*session).handleRSET, "STAT": (*session).handleSTAT, "TOP": (*session).handleTOP, "UIDL": (*session).handleUIDL, "USER": (*session).handleUSER, } ) type session struct { server *Server handler Handler conn net.Conn state int username string markedDeleted map[uint64]struct{} msgSizes map[uint64]uint64 reader *textproto.Reader writer *textproto.Writer } func newSession(server *Server, handler Handler, conn net.Conn) *session { return &session{ server: server, handler: handler, conn: conn, markedDeleted: make(map[uint64]struct{}), msgSizes: make(map[uint64]uint64), reader: textproto.NewReader(bufio.NewReader(conn)), writer: textproto.NewWriter(bufio.NewWriter(conn)), } } // serve method handles the entire session which after the first message from // the server is a series of command-response interactions. func (s *session) serve() { defer s.conn.Close() defer s.unlock() // unlock maildrop if locked no matter what helloParts := []string{"POP3 server ready"} if s.server.APOP { banner := s.server.getBanner() helloParts = append(helloParts, banner) if err := s.handler.SetBanner(banner); err != nil { s.handler.HandleSessionError(err) return // go home handler, you're drunk! } } if err := s.respondOK(strings.Join(helloParts, " ")); err != nil { s.handler.HandleSessionError(err) return // communication problem, most likely? } for { if keepGoing := s.serveOne(); !keepGoing { return } } } // serveOne handles each command-response interaction with the client. The // boolean return value indicates whether the communication with the client // should continue or not. func (s *session) serveOne() bool { if s.state == stateTerminateConnection { return false } readBy := time.Now().Add(s.server.Timeout) if err := s.conn.SetReadDeadline(readBy); err != nil { return s.handleError(err, false) } line, err := s.reader.ReadLine() if err != nil { return s.handleError(err, false) // communication problem, most likely? } args := strings.Split(line, " ") command := strings.ToUpper(args[0]) cmdValidator, exists := validators[command] if !exists { return s.handleError(errInvalidSyntax, true) // unknown command } if err := cmdValidator.validate(s, args[1:]); err != nil { return s.handleError(err, true) } return s.handleError(operationHandlers[command](s, args[1:]), true) } // handleCAPA is a callback for capability listing. // RFC 2449, page 2. func (s *session) handleCAPA(args []string) error { if err := s.respondOK("Capability list follows"); err != nil { return err } dotWriter := s.writer.DotWriter() defer s.closeOrReport(dotWriter) for _, capability := range s.server.capabilities { if _, err := fmt.Fprintln(dotWriter, capability); err != nil { return err } } return nil } // handleAPOP is a callback for an APOP authentication mechanism. // RFC 1939, page 15. func (s *session) handleAPOP(args []string) error { if !s.server.APOP { return NewReportableError("server does not support APOP") } if err := s.handler.AuthenticateAPOP(args[0], args[1]); err != nil { return err } return s.signIn() } // handleDELE is a callback for a single message deletion. // RFC 1939, page 8. func (s *session) handleDELE(args []string) error { return s.withMessageDo(args[0], func(msgId uint64) error { s.markedDeleted[msgId] = struct{}{} return s.respondOK("message %d deleted", msgId) }) } // handleAPOP is a callback for listing one or more messages. // RFC 1939, page 6. func (s *session) handleLIST(args []string) error { if len(args) == 1 { return s.withMessageDo(args[0], func(msgId uint64) error { return s.respondOK("%d %d", msgId, s.msgSizes[msgId]) }) } return s.forEachMessage(func(msgId uint64) (string, error) { return fmt.Sprintf("%d %d", msgId, s.msgSizes[msgId]), nil }) } // handleNOOP is a callback for a no-op (timeout reset) command. // RFC 1939, page 9. func (s *session) handleNOOP(args []string) error { return s.respondOK("doing nothing") } // handlePASS is a callback for the client providing password ("PASS" command). // This must have been preceded by a "USER" command where the client provides // its username. // RFC 1939, page 14. func (s *session) handlePASS(args []string) error { if s.username == "" { return NewReportableError("please provide username first") } if err := s.handler.AuthenticatePASS(s.username, args[0]); err != nil { return err } return s.signIn() } // handleQUIT is a callback for the client terminating the session. It will do // slightly different things depending on the current state of the transaction. // RFC 1939, pages 5 (in authorization state) and 10 (in transaction state). func (s *session) handleQUIT(args []string) error { bye := func() error { s.state = stateTerminateConnection return s.respondOK("dewey POP3 server signing off") } if s.state == stateAuthorization { return bye() } s.state = stateUpdate // so that no future calls will succeed var delMsg []uint64 for key := range s.markedDeleted { delMsg = append(delMsg, key) } if err := s.handler.DeleteMessages(delMsg); err != nil { return err } return bye() } // handleRETR is a callback for the client requesting full content of a a single // message. // RFC 1939, page 8. func (s *session) handleRETR(args []string) (err error) { return s.withMessageDo(args[0], func(msgId uint64) error { if err := s.respondOK("%d octets", s.msgSizes[msgId]); err != nil { return err } readCloser, err := s.handler.GetMessageReader(msgId) if err != nil { return err } defer s.closeOrReport(readCloser) dotWriter := s.writer.DotWriter() defer s.closeOrReport(dotWriter) _, err = io.Copy(dotWriter, readCloser) return err }) } // handleRSET is a callback for the client requesting the session to be reset. // This essentially means undeleting all messages previously marked for // deletion. // RFC 1939, page 9. func (s *session) handleRSET(args []string) error { s.markedDeleted = make(map[uint64]struct{}) return s.respondOK( "maildrop has %d messages (%d octets)", s.getMessageCount(), s.getMaildropSize(), ) } // handleRETR is a callback for the client requesting full content of a a single // message. // RFC 1939, page 8. func (s *session) handleSTAT(args []string) error { return s.respondOK("%d %d", s.getMessageCount(), s.getMaildropSize()) } // handleTOP is a callback for the client requesting a number of lines from the // top of a single message. // RFC 1939, page 11. func (s *session) handleTOP(args []string) error { return s.withMessageDo(args[0], func(msgId uint64) error { noLines, err := strconv.ParseUint(args[1], 10, 64) if err != nil { return errInvalidSyntax } if err := s.writer.PrintfLine("+OK"); err != nil { return err } readCloser, err := s.handler.GetMessageReader(msgId)
"time" )
random_line_split
session.go
nil { s.handler.HandleSessionError(err) return // go home handler, you're drunk! } } if err := s.respondOK(strings.Join(helloParts, " ")); err != nil { s.handler.HandleSessionError(err) return // communication problem, most likely? } for { if keepGoing := s.serveOne(); !keepGoing { return } } } // serveOne handles each command-response interaction with the client. The // boolean return value indicates whether the communication with the client // should continue or not. func (s *session) serveOne() bool { if s.state == stateTerminateConnection { return false } readBy := time.Now().Add(s.server.Timeout) if err := s.conn.SetReadDeadline(readBy); err != nil { return s.handleError(err, false) } line, err := s.reader.ReadLine() if err != nil { return s.handleError(err, false) // communication problem, most likely? } args := strings.Split(line, " ") command := strings.ToUpper(args[0]) cmdValidator, exists := validators[command] if !exists { return s.handleError(errInvalidSyntax, true) // unknown command } if err := cmdValidator.validate(s, args[1:]); err != nil { return s.handleError(err, true) } return s.handleError(operationHandlers[command](s, args[1:]), true) } // handleCAPA is a callback for capability listing. // RFC 2449, page 2. func (s *session) handleCAPA(args []string) error { if err := s.respondOK("Capability list follows"); err != nil { return err } dotWriter := s.writer.DotWriter() defer s.closeOrReport(dotWriter) for _, capability := range s.server.capabilities { if _, err := fmt.Fprintln(dotWriter, capability); err != nil { return err } } return nil } // handleAPOP is a callback for an APOP authentication mechanism. // RFC 1939, page 15. func (s *session) handleAPOP(args []string) error { if !s.server.APOP { return NewReportableError("server does not support APOP") } if err := s.handler.AuthenticateAPOP(args[0], args[1]); err != nil { return err } return s.signIn() } // handleDELE is a callback for a single message deletion. // RFC 1939, page 8. func (s *session) handleDELE(args []string) error { return s.withMessageDo(args[0], func(msgId uint64) error { s.markedDeleted[msgId] = struct{}{} return s.respondOK("message %d deleted", msgId) }) } // handleAPOP is a callback for listing one or more messages. // RFC 1939, page 6. func (s *session) handleLIST(args []string) error { if len(args) == 1 { return s.withMessageDo(args[0], func(msgId uint64) error { return s.respondOK("%d %d", msgId, s.msgSizes[msgId]) }) } return s.forEachMessage(func(msgId uint64) (string, error) { return fmt.Sprintf("%d %d", msgId, s.msgSizes[msgId]), nil }) } // handleNOOP is a callback for a no-op (timeout reset) command. // RFC 1939, page 9. func (s *session) handleNOOP(args []string) error { return s.respondOK("doing nothing") } // handlePASS is a callback for the client providing password ("PASS" command). // This must have been preceded by a "USER" command where the client provides // its username. // RFC 1939, page 14. func (s *session) handlePASS(args []string) error { if s.username == "" { return NewReportableError("please provide username first") } if err := s.handler.AuthenticatePASS(s.username, args[0]); err != nil { return err } return s.signIn() } // handleQUIT is a callback for the client terminating the session. It will do // slightly different things depending on the current state of the transaction. // RFC 1939, pages 5 (in authorization state) and 10 (in transaction state). func (s *session) handleQUIT(args []string) error { bye := func() error { s.state = stateTerminateConnection return s.respondOK("dewey POP3 server signing off") } if s.state == stateAuthorization { return bye() } s.state = stateUpdate // so that no future calls will succeed var delMsg []uint64 for key := range s.markedDeleted { delMsg = append(delMsg, key) } if err := s.handler.DeleteMessages(delMsg); err != nil { return err } return bye() } // handleRETR is a callback for the client requesting full content of a a single // message. // RFC 1939, page 8. func (s *session)
(args []string) (err error) { return s.withMessageDo(args[0], func(msgId uint64) error { if err := s.respondOK("%d octets", s.msgSizes[msgId]); err != nil { return err } readCloser, err := s.handler.GetMessageReader(msgId) if err != nil { return err } defer s.closeOrReport(readCloser) dotWriter := s.writer.DotWriter() defer s.closeOrReport(dotWriter) _, err = io.Copy(dotWriter, readCloser) return err }) } // handleRSET is a callback for the client requesting the session to be reset. // This essentially means undeleting all messages previously marked for // deletion. // RFC 1939, page 9. func (s *session) handleRSET(args []string) error { s.markedDeleted = make(map[uint64]struct{}) return s.respondOK( "maildrop has %d messages (%d octets)", s.getMessageCount(), s.getMaildropSize(), ) } // handleRETR is a callback for the client requesting full content of a a single // message. // RFC 1939, page 8. func (s *session) handleSTAT(args []string) error { return s.respondOK("%d %d", s.getMessageCount(), s.getMaildropSize()) } // handleTOP is a callback for the client requesting a number of lines from the // top of a single message. // RFC 1939, page 11. func (s *session) handleTOP(args []string) error { return s.withMessageDo(args[0], func(msgId uint64) error { noLines, err := strconv.ParseUint(args[1], 10, 64) if err != nil { return errInvalidSyntax } if err := s.writer.PrintfLine("+OK"); err != nil { return err } readCloser, err := s.handler.GetMessageReader(msgId) if err != nil { return err } defer s.closeOrReport(readCloser) dotWriter := s.writer.DotWriter() defer s.closeOrReport(dotWriter) protoReader := textproto.NewReader(bufio.NewReader(readCloser)) for i := uint64(0); i < noLines; i++ { line, readErr := protoReader.ReadLineBytes() if err := printTopLine(line, readErr, dotWriter); err != nil { return err } } return nil }) } func printTopLine(line []byte, readErr error, writer io.Writer) error { if readErr == io.EOF || readErr == nil { if err := writeWithError(writer, line); err != nil { return err } } if readErr != nil { return readErr } return writeWithError(writer, []byte{'\n'}) } func writeWithError(w io.Writer, content []byte) error { _, err := w.Write(content) return err } // handleUIDL is a callback for the client unique message identifiers for // either one or all messages. // RFC 1939, page 12. func (s *session) handleUIDL(args []string) (err error) { if len(args) == 1 { return s.withMessageDo(args[0], func(msgId uint64) error { uidl, err := s.handler.GetMessageID(msgId) if err != nil { return err } return s.respondOK("%d %s", msgId, uidl) }) } return s.forEachMessage(func(msgId uint64) (string, error) { uidl, err := s.handler.GetMessageID(msgId) if err != nil { return "", err } return fmt.Sprintf("%d %s", msgId, uidl), nil }) } // handleUSER is a callback for the client providing it's username. This must be // followed by a "PASS" command with a corresponding password. // RFC 1939, page 13. func (s *session) handleUSER(args []
handleRETR
identifier_name
session.go
nil { s.handler.HandleSessionError(err) return // go home handler, you're drunk! } } if err := s.respondOK(strings.Join(helloParts, " ")); err != nil { s.handler.HandleSessionError(err) return // communication problem, most likely? } for { if keepGoing := s.serveOne(); !keepGoing { return } } } // serveOne handles each command-response interaction with the client. The // boolean return value indicates whether the communication with the client // should continue or not. func (s *session) serveOne() bool { if s.state == stateTerminateConnection { return false } readBy := time.Now().Add(s.server.Timeout) if err := s.conn.SetReadDeadline(readBy); err != nil { return s.handleError(err, false) } line, err := s.reader.ReadLine() if err != nil { return s.handleError(err, false) // communication problem, most likely? } args := strings.Split(line, " ") command := strings.ToUpper(args[0]) cmdValidator, exists := validators[command] if !exists { return s.handleError(errInvalidSyntax, true) // unknown command } if err := cmdValidator.validate(s, args[1:]); err != nil { return s.handleError(err, true) } return s.handleError(operationHandlers[command](s, args[1:]), true) } // handleCAPA is a callback for capability listing. // RFC 2449, page 2. func (s *session) handleCAPA(args []string) error { if err := s.respondOK("Capability list follows"); err != nil { return err } dotWriter := s.writer.DotWriter() defer s.closeOrReport(dotWriter) for _, capability := range s.server.capabilities { if _, err := fmt.Fprintln(dotWriter, capability); err != nil { return err } } return nil } // handleAPOP is a callback for an APOP authentication mechanism. // RFC 1939, page 15. func (s *session) handleAPOP(args []string) error { if !s.server.APOP { return NewReportableError("server does not support APOP") } if err := s.handler.AuthenticateAPOP(args[0], args[1]); err != nil { return err } return s.signIn() } // handleDELE is a callback for a single message deletion. // RFC 1939, page 8. func (s *session) handleDELE(args []string) error { return s.withMessageDo(args[0], func(msgId uint64) error { s.markedDeleted[msgId] = struct{}{} return s.respondOK("message %d deleted", msgId) }) } // handleAPOP is a callback for listing one or more messages. // RFC 1939, page 6. func (s *session) handleLIST(args []string) error { if len(args) == 1 { return s.withMessageDo(args[0], func(msgId uint64) error { return s.respondOK("%d %d", msgId, s.msgSizes[msgId]) }) } return s.forEachMessage(func(msgId uint64) (string, error) { return fmt.Sprintf("%d %d", msgId, s.msgSizes[msgId]), nil }) } // handleNOOP is a callback for a no-op (timeout reset) command. // RFC 1939, page 9. func (s *session) handleNOOP(args []string) error { return s.respondOK("doing nothing") } // handlePASS is a callback for the client providing password ("PASS" command). // This must have been preceded by a "USER" command where the client provides // its username. // RFC 1939, page 14. func (s *session) handlePASS(args []string) error { if s.username == "" { return NewReportableError("please provide username first") } if err := s.handler.AuthenticatePASS(s.username, args[0]); err != nil { return err } return s.signIn() } // handleQUIT is a callback for the client terminating the session. It will do // slightly different things depending on the current state of the transaction. // RFC 1939, pages 5 (in authorization state) and 10 (in transaction state). func (s *session) handleQUIT(args []string) error { bye := func() error { s.state = stateTerminateConnection return s.respondOK("dewey POP3 server signing off") } if s.state == stateAuthorization { return bye() } s.state = stateUpdate // so that no future calls will succeed var delMsg []uint64 for key := range s.markedDeleted { delMsg = append(delMsg, key) } if err := s.handler.DeleteMessages(delMsg); err != nil { return err } return bye() } // handleRETR is a callback for the client requesting full content of a a single // message. // RFC 1939, page 8. func (s *session) handleRETR(args []string) (err error) { return s.withMessageDo(args[0], func(msgId uint64) error { if err := s.respondOK("%d octets", s.msgSizes[msgId]); err != nil { return err } readCloser, err := s.handler.GetMessageReader(msgId) if err != nil { return err } defer s.closeOrReport(readCloser) dotWriter := s.writer.DotWriter() defer s.closeOrReport(dotWriter) _, err = io.Copy(dotWriter, readCloser) return err }) } // handleRSET is a callback for the client requesting the session to be reset. // This essentially means undeleting all messages previously marked for // deletion. // RFC 1939, page 9. func (s *session) handleRSET(args []string) error { s.markedDeleted = make(map[uint64]struct{}) return s.respondOK( "maildrop has %d messages (%d octets)", s.getMessageCount(), s.getMaildropSize(), ) } // handleRETR is a callback for the client requesting full content of a a single // message. // RFC 1939, page 8. func (s *session) handleSTAT(args []string) error { return s.respondOK("%d %d", s.getMessageCount(), s.getMaildropSize()) } // handleTOP is a callback for the client requesting a number of lines from the // top of a single message. // RFC 1939, page 11. func (s *session) handleTOP(args []string) error
return err } } return nil }) } func printTopLine(line []byte, readErr error, writer io.Writer) error { if readErr == io.EOF || readErr == nil { if err := writeWithError(writer, line); err != nil { return err } } if readErr != nil { return readErr } return writeWithError(writer, []byte{'\n'}) } func writeWithError(w io.Writer, content []byte) error { _, err := w.Write(content) return err } // handleUIDL is a callback for the client unique message identifiers for // either one or all messages. // RFC 1939, page 12. func (s *session) handleUIDL(args []string) (err error) { if len(args) == 1 { return s.withMessageDo(args[0], func(msgId uint64) error { uidl, err := s.handler.GetMessageID(msgId) if err != nil { return err } return s.respondOK("%d %s", msgId, uidl) }) } return s.forEachMessage(func(msgId uint64) (string, error) { uidl, err := s.handler.GetMessageID(msgId) if err != nil { return "", err } return fmt.Sprintf("%d %s", msgId, uidl), nil }) } // handleUSER is a callback for the client providing it's username. This must be // followed by a "PASS" command with a corresponding password. // RFC 1939, page 13. func (s *session) handleUSER(args []
{ return s.withMessageDo(args[0], func(msgId uint64) error { noLines, err := strconv.ParseUint(args[1], 10, 64) if err != nil { return errInvalidSyntax } if err := s.writer.PrintfLine("+OK"); err != nil { return err } readCloser, err := s.handler.GetMessageReader(msgId) if err != nil { return err } defer s.closeOrReport(readCloser) dotWriter := s.writer.DotWriter() defer s.closeOrReport(dotWriter) protoReader := textproto.NewReader(bufio.NewReader(readCloser)) for i := uint64(0); i < noLines; i++ { line, readErr := protoReader.ReadLineBytes() if err := printTopLine(line, readErr, dotWriter); err != nil {
identifier_body
game.py
(BaseModel): """ Class representing a Game. A game only contains basic information about the game and the scores. """ id = PrimaryKeyField() game_acbid = IntegerField(unique=True, index=True) team_home_id = ForeignKeyField(Team, related_name='games_home', index=True, null=True) team_away_id = ForeignKeyField(Team, related_name='games_away', index=True, null=True) season = IntegerField(null=False) competition_phase = CharField(max_length=255, null=True) round_phase = CharField(max_length=255, null=True) journey = IntegerField(null=False) score_home = IntegerField(null=True) score_away = IntegerField(null=True) score_home_first = IntegerField(null=True) score_away_first = IntegerField(null=True) score_home_second = IntegerField(null=True) score_away_second = IntegerField(null=True) score_home_third = IntegerField(null=True) score_away_third = IntegerField(null=True) score_home_fourth = IntegerField(null=True) score_away_fourth = IntegerField(null=True) score_home_extra = IntegerField(null=True) score_away_extra = IntegerField(null=True) venue = CharField(max_length=255, null=True) attendance = IntegerField(null=True) kickoff_time = DateTimeField(index=True) referee_1 = CharField(max_length=255, null=True) referee_2 = CharField(max_length=255, null=True) referee_3 = CharField(max_length=255, null=True) db_flag = BooleanField(null=True) @staticmethod def save_games(season, logging_level=logging.INFO): """ Method for saving locally the games of a season. :param season: int :param logging_level: logging object :return: """ logger.info('Starting the download of games...') if season.season == get_current_season(): current_game_events_ids = season.get_current_game_events_ids() game_ids_list = list(current_game_events_ids.values()) else: game_ids_list = season.get_game_ids() n_checkpoints = 4 checkpoints = [round(i * float(len(game_ids_list)) / n_checkpoints) for i in range(n_checkpoints + 1)] for i in range(len(game_ids_list)): game_id = int(game_ids_list[i]) % 1000 url2 = BASE_URL + "/fichas/LACB{}.php".format(game_ids_list[i]) filename = os.path.join(season.GAMES_PATH, str(game_id)+"-" +str(game_ids_list[i]) + '.html') open_or_download(file_path=filename, url=url2) if i in checkpoints: logger.info('{}% already downloaded'.format(round(float(i * 100) / len(game_ids_list)))) logger.info('Download finished! (new {} games in {})\n'.format(len(game_ids_list), season.GAMES_PATH)) @staticmethod def save_games_copa(season, logging_level=logging.INFO): """ Method for saving locally the games of a season. :param season: int :param logging_level: logging object :return: """ logging.basicConfig(level=logging_level) logger = logging.getLogger(__name__) logger.info('Starting the download of games...') if season.season == get_current_season(): current_game_events_ids = season.get_current_game_events_ids_copa() game_ids_list = list(current_game_events_ids.values()) else: game_ids_list=season.get_game_ids_copa() n_checkpoints = 4 checkpoints = [round(i * float(len(game_ids_list)) / n_checkpoints) for i in range(n_checkpoints + 1)] for i in range(len(game_ids_list)): game_id=int(game_ids_list[i]) % 1000 url2 = BASE_URL + "/fichas/CREY{}.php".format(game_ids_list[i]) filename = os.path.join(season.GAMES_COPA_PATH, str(game_id)+"-" +str(game_ids_list[i]) + '.html') open_or_download(file_path=filename, url=url2) if i in checkpoints: logger.info('{}% already downloaded'.format(round(float(i * 100) / len(game_ids_list)))) logger.info('Download finished! (new {} games in {})\n'.format(len(game_ids_list), season.GAMES_COPA_PATH)) @staticmethod def sanity_check(season, logging_level=logging.INFO): sanity_check_game(season.GAMES_PATH, logging_level) @staticmethod def sanity_check_copa(season, logging_level=logging.INFO): sanity_check_game_copa(season.GAMES_COPA_PATH, logging_level) @staticmethod def create_instance(raw_game, game_acbid, season, competition_phase,round_phase=None): """ Extract all the information regarding the game such as the date, attendance, venue, score per quarter or teams. Therefore, we need first to extract and insert the teams in the database in order to get the references to the db. :param raw_game: String :param game_acbid: int :param season: Season :param competition_phase: String :param round_phase: String :return: Game object """ # There are two different statistics table in acb.com. I assume they created the new one to introduce the +/- stat. estadisticas_tag = '.estadisticasnew' if re.search(r'<table class="estadisticasnew"', raw_game) else '.estadisticas' doc = pq(raw_game) game_dict = dict() """ Each game has an unique id in acb.com. The id has 5 digits, where the first two digits are the season code (the oldest season in 1956 has code 1) and the three last are the number of the game (a simple counter since the beginning of the season). This id can be used to access the concrete game within the link 'http://www.acb.com/fichas/LACBXXYYY.php' """ game_dict['game_acbid'] = game_acbid game_dict['season'] = season.season game_dict['competition_phase'] = competition_phase game_dict['round_phase'] = round_phase # Information about the teams. info_teams_data = doc(estadisticas_tag).eq(1) home_team_name = None away_team_name = None """ We only have the names of the teams (text) within the doc. We will look for its associated id by looking in our teamname table, where we have all the historical official names for each team and season. However the ACB sometimes doesn't agree in the names and writes them in different ways depending on the game (sometimes taking older names or making small changes). For instance VALENCIA BASKET CLUB instead of VALENCIA BASKET. So if there is not such a direct correspondance we will take the closest match. """ for i in [0, 2]: team_data = info_teams_data('.estverde').eq(i)('td').eq(0).text() team_name = re.search("(.*) [0-9]", team_data).groups()[0] try: ## In case the name of the team is exactly the same as one stated in our database for a season team_acbid = TeamName.get(TeamName.name == team_name).team_id.team_acbid team = Team.get(Team.team_acbid == team_acbid) except TeamName.DoesNotExist: ## In case there is not an exact correspondance within our database, let's find the closest match. query = TeamName.select(TeamName.team_id, TeamName.name) teams_names_ids = dict() for q in query: teams_names_ids[q.name] = q.team_id.id most_likely_team = difflib.get_close_matches(team_name, teams_names_ids.keys(), 1, 0.4)[0] team = Team.get(Team.id == teams_names_ids[most_likely_team]) if most_likely_team not in season.mismatched_teams: # debug info to check the correctness. season.mismatched_teams.append(most_likely_team) logger.info('Season {} -> {} has been matched to: {}'.format(season.season, team_name, most_likely_team)) # TeamName.get_or_create(**{'team': team, 'name': team_name, 'season': season.season}) game_dict['team_home_id' if i == 0 else 'team_away_id'] = team home_team_name = team_name if i == 0 else home_team_name away_team_name = team_name if i != 0 else away_team_name # Information about the game. info_game_data = doc(estadisticas_tag).eq(0) scheduling_data = info_game_data('.estnegro')('td').eq(0).text() scheduling_data = scheduling_data.split("|") journey, date, time, venue, attendance = list(map(lambda x: x.strip(), scheduling_data)) # Remove extra spaces. if date and time: day, month, year = list(map(int, date.split("/"))) hour, minute = list(map(int, time.split(":")))
Game
identifier_name
game.py
IntegerField(null=True) score_home_fourth = IntegerField(null=True) score_away_fourth = IntegerField(null=True) score_home_extra = IntegerField(null=True) score_away_extra = IntegerField(null=True) venue = CharField(max_length=255, null=True) attendance = IntegerField(null=True) kickoff_time = DateTimeField(index=True) referee_1 = CharField(max_length=255, null=True) referee_2 = CharField(max_length=255, null=True) referee_3 = CharField(max_length=255, null=True) db_flag = BooleanField(null=True) @staticmethod def save_games(season, logging_level=logging.INFO): """ Method for saving locally the games of a season. :param season: int :param logging_level: logging object :return: """ logger.info('Starting the download of games...') if season.season == get_current_season(): current_game_events_ids = season.get_current_game_events_ids() game_ids_list = list(current_game_events_ids.values()) else: game_ids_list = season.get_game_ids() n_checkpoints = 4 checkpoints = [round(i * float(len(game_ids_list)) / n_checkpoints) for i in range(n_checkpoints + 1)] for i in range(len(game_ids_list)): game_id = int(game_ids_list[i]) % 1000 url2 = BASE_URL + "/fichas/LACB{}.php".format(game_ids_list[i]) filename = os.path.join(season.GAMES_PATH, str(game_id)+"-" +str(game_ids_list[i]) + '.html') open_or_download(file_path=filename, url=url2) if i in checkpoints: logger.info('{}% already downloaded'.format(round(float(i * 100) / len(game_ids_list)))) logger.info('Download finished! (new {} games in {})\n'.format(len(game_ids_list), season.GAMES_PATH)) @staticmethod def save_games_copa(season, logging_level=logging.INFO): """ Method for saving locally the games of a season. :param season: int :param logging_level: logging object :return: """ logging.basicConfig(level=logging_level) logger = logging.getLogger(__name__) logger.info('Starting the download of games...') if season.season == get_current_season(): current_game_events_ids = season.get_current_game_events_ids_copa() game_ids_list = list(current_game_events_ids.values()) else: game_ids_list=season.get_game_ids_copa() n_checkpoints = 4 checkpoints = [round(i * float(len(game_ids_list)) / n_checkpoints) for i in range(n_checkpoints + 1)] for i in range(len(game_ids_list)): game_id=int(game_ids_list[i]) % 1000 url2 = BASE_URL + "/fichas/CREY{}.php".format(game_ids_list[i]) filename = os.path.join(season.GAMES_COPA_PATH, str(game_id)+"-" +str(game_ids_list[i]) + '.html') open_or_download(file_path=filename, url=url2) if i in checkpoints: logger.info('{}% already downloaded'.format(round(float(i * 100) / len(game_ids_list)))) logger.info('Download finished! (new {} games in {})\n'.format(len(game_ids_list), season.GAMES_COPA_PATH)) @staticmethod def sanity_check(season, logging_level=logging.INFO): sanity_check_game(season.GAMES_PATH, logging_level) @staticmethod def sanity_check_copa(season, logging_level=logging.INFO): sanity_check_game_copa(season.GAMES_COPA_PATH, logging_level) @staticmethod def create_instance(raw_game, game_acbid, season, competition_phase,round_phase=None): """ Extract all the information regarding the game such as the date, attendance, venue, score per quarter or teams. Therefore, we need first to extract and insert the teams in the database in order to get the references to the db. :param raw_game: String :param game_acbid: int :param season: Season :param competition_phase: String :param round_phase: String :return: Game object """ # There are two different statistics table in acb.com. I assume they created the new one to introduce the +/- stat. estadisticas_tag = '.estadisticasnew' if re.search(r'<table class="estadisticasnew"', raw_game) else '.estadisticas' doc = pq(raw_game) game_dict = dict() """ Each game has an unique id in acb.com. The id has 5 digits, where the first two digits are the season code (the oldest season in 1956 has code 1) and the three last are the number of the game (a simple counter since the beginning of the season). This id can be used to access the concrete game within the link 'http://www.acb.com/fichas/LACBXXYYY.php' """ game_dict['game_acbid'] = game_acbid game_dict['season'] = season.season game_dict['competition_phase'] = competition_phase game_dict['round_phase'] = round_phase # Information about the teams. info_teams_data = doc(estadisticas_tag).eq(1) home_team_name = None away_team_name = None """ We only have the names of the teams (text) within the doc. We will look for its associated id by looking in our teamname table, where we have all the historical official names for each team and season. However the ACB sometimes doesn't agree in the names and writes them in different ways depending on the game (sometimes taking older names or making small changes). For instance VALENCIA BASKET CLUB instead of VALENCIA BASKET. So if there is not such a direct correspondance we will take the closest match. """ for i in [0, 2]: team_data = info_teams_data('.estverde').eq(i)('td').eq(0).text() team_name = re.search("(.*) [0-9]", team_data).groups()[0] try: ## In case the name of the team is exactly the same as one stated in our database for a season team_acbid = TeamName.get(TeamName.name == team_name).team_id.team_acbid team = Team.get(Team.team_acbid == team_acbid) except TeamName.DoesNotExist: ## In case there is not an exact correspondance within our database, let's find the closest match. query = TeamName.select(TeamName.team_id, TeamName.name) teams_names_ids = dict() for q in query: teams_names_ids[q.name] = q.team_id.id most_likely_team = difflib.get_close_matches(team_name, teams_names_ids.keys(), 1, 0.4)[0] team = Team.get(Team.id == teams_names_ids[most_likely_team]) if most_likely_team not in season.mismatched_teams: # debug info to check the correctness. season.mismatched_teams.append(most_likely_team) logger.info('Season {} -> {} has been matched to: {}'.format(season.season, team_name, most_likely_team)) # TeamName.get_or_create(**{'team': team, 'name': team_name, 'season': season.season}) game_dict['team_home_id' if i == 0 else 'team_away_id'] = team home_team_name = team_name if i == 0 else home_team_name away_team_name = team_name if i != 0 else away_team_name # Information about the game. info_game_data = doc(estadisticas_tag).eq(0) scheduling_data = info_game_data('.estnegro')('td').eq(0).text() scheduling_data = scheduling_data.split("|") journey, date, time, venue, attendance = list(map(lambda x: x.strip(), scheduling_data)) # Remove extra spaces. if date and time: day, month, year = list(map(int, date.split("/"))) hour, minute = list(map(int, time.split(":"))) game_dict['kickoff_time'] = datetime.datetime(year=year, month=month, day=day, hour=hour, minute=minute) if attendance: try: game_dict['attendance'] = int(attendance.split(":")[1]) except ValueError: pass if venue: game_dict['venue'] = venue if journey: game_dict['journey'] = journey.split(" ")[1] if competition_phase=='cup': if int(journey.split(" ")[1])==1: game_dict['round_phase'] ="quarter_final" elif int(journey.split(" ")[1])==2: game_dict['round_phase'] ="semi_final" elif int(journey.split(" ")[1])==3: game_dict['round_phase'] ="final" for i in range(2, 7): score_home_attribute = '' score_away_attribute = '' if i == 2:
score_home_attribute = 'score_home_first' score_away_attribute = 'score_away_first'
conditional_block
game.py
games_away', index=True, null=True) season = IntegerField(null=False) competition_phase = CharField(max_length=255, null=True) round_phase = CharField(max_length=255, null=True) journey = IntegerField(null=False) score_home = IntegerField(null=True) score_away = IntegerField(null=True) score_home_first = IntegerField(null=True) score_away_first = IntegerField(null=True) score_home_second = IntegerField(null=True) score_away_second = IntegerField(null=True) score_home_third = IntegerField(null=True) score_away_third = IntegerField(null=True) score_home_fourth = IntegerField(null=True) score_away_fourth = IntegerField(null=True) score_home_extra = IntegerField(null=True) score_away_extra = IntegerField(null=True) venue = CharField(max_length=255, null=True) attendance = IntegerField(null=True) kickoff_time = DateTimeField(index=True) referee_1 = CharField(max_length=255, null=True) referee_2 = CharField(max_length=255, null=True) referee_3 = CharField(max_length=255, null=True) db_flag = BooleanField(null=True) @staticmethod def save_games(season, logging_level=logging.INFO): """ Method for saving locally the games of a season. :param season: int :param logging_level: logging object :return: """ logger.info('Starting the download of games...') if season.season == get_current_season(): current_game_events_ids = season.get_current_game_events_ids() game_ids_list = list(current_game_events_ids.values()) else: game_ids_list = season.get_game_ids() n_checkpoints = 4 checkpoints = [round(i * float(len(game_ids_list)) / n_checkpoints) for i in range(n_checkpoints + 1)] for i in range(len(game_ids_list)): game_id = int(game_ids_list[i]) % 1000 url2 = BASE_URL + "/fichas/LACB{}.php".format(game_ids_list[i]) filename = os.path.join(season.GAMES_PATH, str(game_id)+"-" +str(game_ids_list[i]) + '.html') open_or_download(file_path=filename, url=url2) if i in checkpoints: logger.info('{}% already downloaded'.format(round(float(i * 100) / len(game_ids_list)))) logger.info('Download finished! (new {} games in {})\n'.format(len(game_ids_list), season.GAMES_PATH)) @staticmethod def save_games_copa(season, logging_level=logging.INFO): """ Method for saving locally the games of a season. :param season: int :param logging_level: logging object :return: """ logging.basicConfig(level=logging_level) logger = logging.getLogger(__name__) logger.info('Starting the download of games...') if season.season == get_current_season(): current_game_events_ids = season.get_current_game_events_ids_copa() game_ids_list = list(current_game_events_ids.values()) else: game_ids_list=season.get_game_ids_copa() n_checkpoints = 4 checkpoints = [round(i * float(len(game_ids_list)) / n_checkpoints) for i in range(n_checkpoints + 1)] for i in range(len(game_ids_list)): game_id=int(game_ids_list[i]) % 1000 url2 = BASE_URL + "/fichas/CREY{}.php".format(game_ids_list[i]) filename = os.path.join(season.GAMES_COPA_PATH, str(game_id)+"-" +str(game_ids_list[i]) + '.html') open_or_download(file_path=filename, url=url2) if i in checkpoints: logger.info('{}% already downloaded'.format(round(float(i * 100) / len(game_ids_list)))) logger.info('Download finished! (new {} games in {})\n'.format(len(game_ids_list), season.GAMES_COPA_PATH)) @staticmethod def sanity_check(season, logging_level=logging.INFO): sanity_check_game(season.GAMES_PATH, logging_level) @staticmethod def sanity_check_copa(season, logging_level=logging.INFO):
@staticmethod def create_instance(raw_game, game_acbid, season, competition_phase,round_phase=None): """ Extract all the information regarding the game such as the date, attendance, venue, score per quarter or teams. Therefore, we need first to extract and insert the teams in the database in order to get the references to the db. :param raw_game: String :param game_acbid: int :param season: Season :param competition_phase: String :param round_phase: String :return: Game object """ # There are two different statistics table in acb.com. I assume they created the new one to introduce the +/- stat. estadisticas_tag = '.estadisticasnew' if re.search(r'<table class="estadisticasnew"', raw_game) else '.estadisticas' doc = pq(raw_game) game_dict = dict() """ Each game has an unique id in acb.com. The id has 5 digits, where the first two digits are the season code (the oldest season in 1956 has code 1) and the three last are the number of the game (a simple counter since the beginning of the season). This id can be used to access the concrete game within the link 'http://www.acb.com/fichas/LACBXXYYY.php' """ game_dict['game_acbid'] = game_acbid game_dict['season'] = season.season game_dict['competition_phase'] = competition_phase game_dict['round_phase'] = round_phase # Information about the teams. info_teams_data = doc(estadisticas_tag).eq(1) home_team_name = None away_team_name = None """ We only have the names of the teams (text) within the doc. We will look for its associated id by looking in our teamname table, where we have all the historical official names for each team and season. However the ACB sometimes doesn't agree in the names and writes them in different ways depending on the game (sometimes taking older names or making small changes). For instance VALENCIA BASKET CLUB instead of VALENCIA BASKET. So if there is not such a direct correspondance we will take the closest match. """ for i in [0, 2]: team_data = info_teams_data('.estverde').eq(i)('td').eq(0).text() team_name = re.search("(.*) [0-9]", team_data).groups()[0] try: ## In case the name of the team is exactly the same as one stated in our database for a season team_acbid = TeamName.get(TeamName.name == team_name).team_id.team_acbid team = Team.get(Team.team_acbid == team_acbid) except TeamName.DoesNotExist: ## In case there is not an exact correspondance within our database, let's find the closest match. query = TeamName.select(TeamName.team_id, TeamName.name) teams_names_ids = dict() for q in query: teams_names_ids[q.name] = q.team_id.id most_likely_team = difflib.get_close_matches(team_name, teams_names_ids.keys(), 1, 0.4)[0] team = Team.get(Team.id == teams_names_ids[most_likely_team]) if most_likely_team not in season.mismatched_teams: # debug info to check the correctness. season.mismatched_teams.append(most_likely_team) logger.info('Season {} -> {} has been matched to: {}'.format(season.season, team_name, most_likely_team)) # TeamName.get_or_create(**{'team': team, 'name': team_name, 'season': season.season}) game_dict['team_home_id' if i == 0 else 'team_away_id'] = team home_team_name = team_name if i == 0 else home_team_name away_team_name = team_name if i != 0 else away_team_name # Information about the game. info_game_data = doc(estadisticas_tag).eq(0) scheduling_data = info_game_data('.estnegro')('td').eq(0).text() scheduling_data = scheduling_data.split("|") journey, date, time, venue, attendance = list(map(lambda x: x.strip(), scheduling_data)) # Remove extra spaces. if date and time: day, month, year = list(map(int, date.split("/"))) hour, minute = list(map(int, time.split(":"))) game_dict['kickoff_time'] = datetime.datetime(year=year, month=month, day=day, hour=hour, minute=minute) if attendance: try: game_dict['attendance'] = int(attendance.split(":")[1]) except ValueError: pass if venue: game_dict['venue'] = venue if journey: game_dict['jour
sanity_check_game_copa(season.GAMES_COPA_PATH, logging_level)
identifier_body
game.py
=True) @staticmethod def save_games(season, logging_level=logging.INFO): """ Method for saving locally the games of a season. :param season: int :param logging_level: logging object :return: """ logger.info('Starting the download of games...') if season.season == get_current_season(): current_game_events_ids = season.get_current_game_events_ids() game_ids_list = list(current_game_events_ids.values()) else: game_ids_list = season.get_game_ids() n_checkpoints = 4 checkpoints = [round(i * float(len(game_ids_list)) / n_checkpoints) for i in range(n_checkpoints + 1)] for i in range(len(game_ids_list)): game_id = int(game_ids_list[i]) % 1000 url2 = BASE_URL + "/fichas/LACB{}.php".format(game_ids_list[i]) filename = os.path.join(season.GAMES_PATH, str(game_id)+"-" +str(game_ids_list[i]) + '.html') open_or_download(file_path=filename, url=url2) if i in checkpoints: logger.info('{}% already downloaded'.format(round(float(i * 100) / len(game_ids_list)))) logger.info('Download finished! (new {} games in {})\n'.format(len(game_ids_list), season.GAMES_PATH)) @staticmethod def save_games_copa(season, logging_level=logging.INFO): """ Method for saving locally the games of a season. :param season: int :param logging_level: logging object :return: """ logging.basicConfig(level=logging_level) logger = logging.getLogger(__name__) logger.info('Starting the download of games...') if season.season == get_current_season(): current_game_events_ids = season.get_current_game_events_ids_copa() game_ids_list = list(current_game_events_ids.values()) else: game_ids_list=season.get_game_ids_copa() n_checkpoints = 4 checkpoints = [round(i * float(len(game_ids_list)) / n_checkpoints) for i in range(n_checkpoints + 1)] for i in range(len(game_ids_list)): game_id=int(game_ids_list[i]) % 1000 url2 = BASE_URL + "/fichas/CREY{}.php".format(game_ids_list[i]) filename = os.path.join(season.GAMES_COPA_PATH, str(game_id)+"-" +str(game_ids_list[i]) + '.html') open_or_download(file_path=filename, url=url2) if i in checkpoints: logger.info('{}% already downloaded'.format(round(float(i * 100) / len(game_ids_list)))) logger.info('Download finished! (new {} games in {})\n'.format(len(game_ids_list), season.GAMES_COPA_PATH)) @staticmethod def sanity_check(season, logging_level=logging.INFO): sanity_check_game(season.GAMES_PATH, logging_level) @staticmethod def sanity_check_copa(season, logging_level=logging.INFO): sanity_check_game_copa(season.GAMES_COPA_PATH, logging_level) @staticmethod def create_instance(raw_game, game_acbid, season, competition_phase,round_phase=None): """ Extract all the information regarding the game such as the date, attendance, venue, score per quarter or teams. Therefore, we need first to extract and insert the teams in the database in order to get the references to the db. :param raw_game: String :param game_acbid: int :param season: Season :param competition_phase: String :param round_phase: String :return: Game object """ # There are two different statistics table in acb.com. I assume they created the new one to introduce the +/- stat. estadisticas_tag = '.estadisticasnew' if re.search(r'<table class="estadisticasnew"', raw_game) else '.estadisticas' doc = pq(raw_game) game_dict = dict() """ Each game has an unique id in acb.com. The id has 5 digits, where the first two digits are the season code (the oldest season in 1956 has code 1) and the three last are the number of the game (a simple counter since the beginning of the season). This id can be used to access the concrete game within the link 'http://www.acb.com/fichas/LACBXXYYY.php' """ game_dict['game_acbid'] = game_acbid game_dict['season'] = season.season game_dict['competition_phase'] = competition_phase game_dict['round_phase'] = round_phase # Information about the teams. info_teams_data = doc(estadisticas_tag).eq(1) home_team_name = None away_team_name = None """ We only have the names of the teams (text) within the doc. We will look for its associated id by looking in our teamname table, where we have all the historical official names for each team and season. However the ACB sometimes doesn't agree in the names and writes them in different ways depending on the game (sometimes taking older names or making small changes). For instance VALENCIA BASKET CLUB instead of VALENCIA BASKET. So if there is not such a direct correspondance we will take the closest match. """ for i in [0, 2]: team_data = info_teams_data('.estverde').eq(i)('td').eq(0).text() team_name = re.search("(.*) [0-9]", team_data).groups()[0] try: ## In case the name of the team is exactly the same as one stated in our database for a season team_acbid = TeamName.get(TeamName.name == team_name).team_id.team_acbid team = Team.get(Team.team_acbid == team_acbid) except TeamName.DoesNotExist: ## In case there is not an exact correspondance within our database, let's find the closest match. query = TeamName.select(TeamName.team_id, TeamName.name) teams_names_ids = dict() for q in query: teams_names_ids[q.name] = q.team_id.id most_likely_team = difflib.get_close_matches(team_name, teams_names_ids.keys(), 1, 0.4)[0] team = Team.get(Team.id == teams_names_ids[most_likely_team]) if most_likely_team not in season.mismatched_teams: # debug info to check the correctness. season.mismatched_teams.append(most_likely_team) logger.info('Season {} -> {} has been matched to: {}'.format(season.season, team_name, most_likely_team)) # TeamName.get_or_create(**{'team': team, 'name': team_name, 'season': season.season}) game_dict['team_home_id' if i == 0 else 'team_away_id'] = team home_team_name = team_name if i == 0 else home_team_name away_team_name = team_name if i != 0 else away_team_name # Information about the game. info_game_data = doc(estadisticas_tag).eq(0) scheduling_data = info_game_data('.estnegro')('td').eq(0).text() scheduling_data = scheduling_data.split("|") journey, date, time, venue, attendance = list(map(lambda x: x.strip(), scheduling_data)) # Remove extra spaces. if date and time: day, month, year = list(map(int, date.split("/"))) hour, minute = list(map(int, time.split(":"))) game_dict['kickoff_time'] = datetime.datetime(year=year, month=month, day=day, hour=hour, minute=minute) if attendance: try: game_dict['attendance'] = int(attendance.split(":")[1]) except ValueError: pass if venue: game_dict['venue'] = venue if journey: game_dict['journey'] = journey.split(" ")[1] if competition_phase=='cup': if int(journey.split(" ")[1])==1: game_dict['round_phase'] ="quarter_final" elif int(journey.split(" ")[1])==2: game_dict['round_phase'] ="semi_final" elif int(journey.split(" ")[1])==3: game_dict['round_phase'] ="final" for i in range(2, 7): score_home_attribute = '' score_away_attribute = '' if i == 2: score_home_attribute = 'score_home_first' score_away_attribute = 'score_away_first' elif i == 3: score_home_attribute = 'score_home_second' score_away_attribute = 'score_away_second' elif i == 4: score_home_attribute = 'score_home_third' score_away_attribute = 'score_away_third' elif i == 5: score_home_attribute = 'score_home_fourth' score_away_attribute = 'score_away_fourth'
elif i == 6: score_home_attribute = 'score_home_extra' score_away_attribute = 'score_away_extra' quarter_data = info_game_data('.estnaranja')('td').eq(i).text()
random_line_split
utils.py
if val: if attr.endswith('date'): display_value = repr(val.value) else: display_value = str(val.value) attr_confidence = val.confidence else: display_value = '' attr_confidence = '1' obj_data[attr] = display_value obj_data[attr + '_confidence'] = attr_confidence # Differentiate id/text for complex attributes for attr in self.complex_attrs: try: val = getattr(obj, attr).get_value() except AttributeError: val = None if val: val_id = val.id val_text = str(val.value) attr_confidence = val.confidence else: val_id, val_text = '', '' attr_confidence = '1' obj_data[attr] = {} obj_data[attr]['id'] = val_id obj_data[attr]['text'] = val_text obj_data[attr + '_confidence'] = attr_confidence # Add optional attributes that are lists for attr in self.list_attrs: try: lst = getattr(obj, attr).get_list() except AttributeError: lst = [] lst_no_nulls = [inst.get_value() for inst in lst if inst.get_value()] if any(lst_no_nulls): lst_confidence = lst_no_nulls[0].confidence else: lst_confidence = '1' cleaned_lst = [] for inst in lst_no_nulls: if attr != 'classification': cleaned_lst.append({ 'id': inst.id, 'text': str(inst.value) }) else: # For classificaitons, we want to get the Classification # model, not the OrganizationClassification model cleaned_lst.append({ 'id': inst.value.id, 'text': str(inst.value) }) obj_data[attr] = cleaned_lst obj_data[attr + '_confidence'] = lst_confidence # Add objects corresponding to foreign keys for attr, fkey in self.set_attrs.items(): try: lst = getattr(obj, attr).all() except AttributeError: lst = [] lst_refs = [getattr(inst.object_ref, fkey) for inst in lst if getattr(inst.object_ref, fkey, None)] lst_values = [inst.get_value().value for inst in lst_refs if inst.get_value()] # We need to traverse the relationship again due to the particular # membership relationships on complex fields lst_values = [inst.get_value() for inst in lst_values if inst.get_value()] if any(lst_values): lst_confidence = lst_values[0].confidence else: lst_confidence = '1' cleaned_lst = [] for inst in lst_values: cleaned_lst.append({ 'id': inst.id, 'text': str(inst.value) }) obj_data[attr] = cleaned_lst obj_data[attr + '_confidence'] = lst_confidence collected_attrs.append(obj_data) return collected_attrs class DictDiffer(object): """ Calculate the difference between two dictionaries as: (1) items added (2) items removed (3) keys same in both but changed values (4) keys same in both and unchanged values """ def __init__(self, current_dict, past_dict): self.current_dict, self.past_dict = current_dict, past_dict self.set_current, self.set_past = set(current_dict.keys()), set(past_dict.keys()) self.intersect = self.set_current.intersection(self.set_past) def added(self): return self.set_current - self.intersect def removed(self): return self.set_past - self.intersect def changed(self): return set(o for o in self.intersect if self.past_dict[o] != self.current_dict[o]) def unchanged(self): return set(o for o in self.intersect if self.past_dict[o] == self.current_dict[o]) class VersionsMixin(object): ''' Model mixin to get version diff for a given model instance ''' def _getDiff(self, differ): skip_fields = ['date_updated', 'id'] def makeIt(change_type): for field in getattr(differ, change_type)(): if field not in skip_fields: if change_type == 'changed': yield { 'field_name': field, 'to': differ.current_dict[field], 'from': differ.past_dict[field], } elif change_type == 'added': yield { 'field_name': field, 'to': differ.current_dict[field], 'from': None } elif change_type == 'removed': yield { 'field_name': field, 'to': None, 'from': differ.past_dict[field] } additions = [a for a in makeIt('added')] changes = [c for c in makeIt('changed')] removals = [r for r in makeIt('removed')] return additions, changes, removals def getRevisions(self, versions): from source.models import Source, AccessPoint revisions = [] for version in versions: complete_revision = { 'id': version.revision.id } revision_meta = { 'modification_date': version.revision.date_created, 'comment': version.revision.comment, 'user': version.revision.user, } complex_list_models = [c.field_model._meta.model_name for c in getattr(self, 'complex_lists', [])] for object_property in version.revision.version_set.all(): if object_property.object != self or isinstance(self, Source): serialized_data = json.loads(object_property.serialized_data)[0] # a bit of a hack in order to get sources and access points # to work field_names = [] if 'value' in serialized_data['fields']: field_names.append((serialized_data['fields']['value'], serialized_data['model'].split('.')[1])) else: for field in serialized_data['fields']: field_names.append((serialized_data['fields'][field], field)) for value, field_name in field_names: if field_name in complex_list_models: try: complete_revision[field_name].add(value) except KeyError: complete_revision[field_name] = {value} else: complete_revision[field_name] = value revisions.append((complete_revision, version.revision)) return revisions def getDifferences(self, revisions): differences = [] for index, (version, revision) in enumerate(revisions): if (index - 1) > 0: try: previous, previous_revision = revisions[index - 1] except (IndexError, AssertionError): continue else: continue differ = DictDiffer(previous, version) fields_added, fields_changed, fields_removed = self._getDiff(differ) diff = { 'modification_date': previous_revision.date_created, 'comment': previous_revision.comment, 'user': previous_revision.user, 'from_id': version['id'], 'to_id': previous_revision.id, 'fields_added': fields_added, 'fields_changed': fields_changed, 'fields_removed': fields_removed, 'model': self._meta.object_name, } differences.append(diff) return differences def getVersions(self, versions=None): if not versions: versions = Version.objects.get_for_object(self) revisions = self.getRevisions(versions) return self.getDifferences(revisions) def execute_sql(file_path): ''' Execute arbitrary SQL code from a file location. ''' with open(file_path) as f: statements = f.read().split(';') with connection.cursor() as c: for statement in statements: if statement.strip(): c.execute(statement.strip()) def class_for_name(class_name, module_name="person.models"): # Check for irregular class names (names where we cannot infer the class # name by capitalizing the first letter of class_name) irregular_names = ( ('Membershipperson', 'MembershipPerson'), ('Membershiporganization', 'MembershipOrganization'), ('Personextra', 'PersonExtra'), ('Personbiography', 'PersonBiography') ) for name, formatted_name in irregular_names: if class_name == name: class_name = formatted_name break if class_name not in settings.ALLOWED_CLASS_FOR_NAME: raise Exception("Unallowed class for name") module = importlib.import_module(module_name) class_ = getattr(module, class_name) return class_ def get_osm_by_id(osm_id): osm_feature = None cursor = connection.cursor() query = ''' SELECT ST_X(ST_Centroid(geometry)), ST_Y(ST_Centroid(geometry)), * FROM osm_data WHERE id = {osm_id} '''.format(osm_id=osm_id) cursor.execute(query) columns = [c[0] for c in cursor.description] results_tuple = namedtuple('OSMFeature', columns) row = cursor.fetchone() if row: osm_feature = results_tuple(*row) return osm_feature def get_hierarchy_by_id(osm_id): hierarchy = ''' SELECT parents.* FROM osm_data AS parents JOIN ( SELECT UNNEST(hierarchy) AS h_id, localname, tags, admin_level,
val = getattr(obj, attr).get_value() except AttributeError: val = None
random_line_split
utils.py
_name) return class_ def get_osm_by_id(osm_id): osm_feature = None cursor = connection.cursor() query = ''' SELECT ST_X(ST_Centroid(geometry)), ST_Y(ST_Centroid(geometry)), * FROM osm_data WHERE id = {osm_id} '''.format(osm_id=osm_id) cursor.execute(query) columns = [c[0] for c in cursor.description] results_tuple = namedtuple('OSMFeature', columns) row = cursor.fetchone() if row: osm_feature = results_tuple(*row) return osm_feature def get_hierarchy_by_id(osm_id): hierarchy = ''' SELECT parents.* FROM osm_data AS parents JOIN ( SELECT UNNEST(hierarchy) AS h_id, localname, tags, admin_level, name, geometry FROM osm_data WHERE id = %s ) AS child ON parents.id = child.h_id::integer ''' cursor = connection.cursor() cursor.execute(hierarchy, [osm_id]) columns = [c[0] for c in cursor.description] results_tuple = namedtuple('OSMFeature', columns) hierarchy = [results_tuple(*r) for r in cursor] return hierarchy def generate_hierarchy(query, q_args, rel_field, sources=False): cursor = connection.cursor() cursor.execute(query, q_args) columns = [c[0] for c in cursor.description] results_tuple = namedtuple('Organization', columns) hierarchy = [(idx, results_tuple(*r)) for idx, r in enumerate(cursor)] trimmed_hierarchy = [] for org_id, orgs in itertools.groupby(hierarchy, key=lambda x: x[1].id): group = list(orgs) lowest_index = min(g[0] for g in group) orgs = [o[1] for o in group] start_date = None if orgs[0].start_date: start_date = orgs[0].start_date end_date = None if orgs[0].end_date: end_date = orgs[0].end_date # Create a label, which we display on the charts for person and unit "parents." label = '<b>' + orgs[0].name + '</b>' + '\n\n' + _('Unknown commander') if orgs[0].commander: label = '<b>' + orgs[0].name + '</b>' + '\n\n' + orgs[0].commander trimmed = { 'id': org_id, 'label': str(label), 'detail_id': str(orgs[0].org_org_id), 'name': orgs[0].name, 'other_names': list({o.alias.strip() for o in orgs if o.alias}), 'classifications': list({o.classification.strip() for o in orgs if o.classification}), 'division_id': orgs[0].division_id, 'date_first_cited': start_date, 'date_last_cited': end_date, 'commander': orgs[0].commander, } trimmed[rel_field] = getattr(orgs[0], rel_field) if sources: trimmed['sources'] = [] source_ids = [] for o in orgs: org_source = json.loads(o.source) if org_source['id'] not in source_ids: trimmed['sources'].append(org_source) source_ids.append(org_source['id']) trimmed['confidence'] = REVERSE_CONFIDENCE[int(orgs[0].confidence)].title() trimmed_hierarchy.append((lowest_index, trimmed)) hierarchy = [i[1] for i in sorted(trimmed_hierarchy, key=lambda x: x[0])] return hierarchy # this makes an edge list that shows the parent relationships (see child_id) solr = pysolr.Solr(settings.SOLR_URL) def get_org_hierarchy_by_id(org_id, when=None, sources=False, direction='up', authenticated=False): ''' org_id: uuid for the organization when: date for limiting the search ''' base_url = settings.SOLR_URL if direction == 'up': from_ = 'child' to = 'parent' elif direction == 'down': from_ = 'parent' to = 'child' filter_query = '{!graph from=composition_%s_id_s_fct to=composition_%s_id_s_fct returnRoot=true}composition_%s_id_s_fct:%s' % (from_, to, from_, org_id) if when: filter_query += ' AND {!field f=composition_daterange_dr op=contains}%s' % when if not authenticated: filter_query += ' AND published_b:T' results = solr.search('*:*', fq=filter_query) if when: for result in results: for key in [from_, to]: org = result['composition_{}_id_s_fct'.format(key)] args = (org, when) query = 'commander_org_id_s_fct:%s AND {!field f=commander_assignment_range_dr op=contains}%s' % args if not authenticated: query += ' AND published_b:T' commanders = solr.search(query) # We need to deduplicate commanders and then throw out the open ended date ranges. result['commanders-{}'.format(key)] = [] for commander in commanders: label_fmt = '{name} ({start} - {end})' assignment_range = commander['commander_assignment_range_dr'] start, end = assignment_range.replace('[', '').replace(']', '').split(' TO ') if start == '*': start = '?' if end == '*': end = '?' label = label_fmt.format(name=commander['commander_person_name_s'], start=start, end=end) commander['label'] = label result['commanders-{}'.format(key)].append(commander) return results def get_child_orgs_by_id(org_id, when=None, sources=False): hierarchy = ''' WITH RECURSIVE parents AS ( SELECT o.*, NULL::VARCHAR AS parent_id, NULL::VARCHAR AS parent_name, NULL::DATE AS start_date, NULL::DATE AS end_date, NULL::VARCHAR AS comp_open_ended, NULL::VARCHAR AS source, NULL::VARCHAR AS confidence FROM organization As o WHERE id = %s UNION SELECT o.*, h.parent_id::VARCHAR AS parent_id, parents.name AS parent_name, h.start_date::date, h.end_date::date, h.open_ended AS comp_open_ended, row_to_json(ss.*)::VARCHAR AS source, ccc.confidence FROM organization AS o JOIN composition AS h ON o.id = h.child_id JOIN composition_compositionchild AS ccc ON h.id = ccc.object_ref_id LEFT JOIN composition_compositionchild_sources AS cccs ON ccc.id = cccs.compositionchild_id LEFT JOIN source_source AS ss ON cccs.source_id = ss.uuid JOIN parents ON parents.id = h.parent_id ) SELECT * FROM parents WHERE id != %s ''' q_args = [org_id, org_id] if when: hierarchy = ''' {} AND CASE WHEN (start_date IS NOT NULL AND end_date IS NOT NULL AND comp_open_ended IN ('N', 'E')) THEN (%s::date BETWEEN start_date::date AND end_date::date) WHEN (start_date IS NOT NULL AND end_date IS NOT NULL AND comp_open_ended = 'Y') THEN (%s::date BETWEEN start_date::date AND NOW()::date) WHEN (start_date IS NOT NULL AND end_date IS NULL AND comp_open_ended IN ('N', 'E')) THEN (start_date::date = %s::date) WHEN (start_date IS NOT NULL AND end_date IS NULL AND comp_open_ended = 'Y') THEN (%s::date BETWEEN start_date::date AND NOW()::date) WHEN (start_date IS NULL AND end_date IS NOT NULL AND comp_open_ended IN ('N', 'E')) THEN (end_date::date = %s) WHEN (start_date IS NULL AND end_date IS NOT NULL AND comp_open_ended = 'Y') THEN TRUE END '''.format(hierarchy) q_args.extend([when] * 5) hierarchy = '{} ORDER BY id'.format(hierarchy) hierarchy = generate_hierarchy(hierarchy, q_args, 'parent_id', sources=sources) return hierarchy def deleted_in_str(objects): index = 0 for obj in objects: if isinstance(obj, list): objects[index] = deleted_in_str(obj) else: if hasattr(obj, 'field_name'): name = obj.field_name + ": " + str(obj) else: name = type(obj).__name__ + ": " + str(obj) if '_sources' in name: objects[index] = "Object sources" else:
objects[index] = name
conditional_block
utils.py
()] if any(lst_values): lst_confidence = lst_values[0].confidence else: lst_confidence = '1' cleaned_lst = [] for inst in lst_values: cleaned_lst.append({ 'id': inst.id, 'text': str(inst.value) }) obj_data[attr] = cleaned_lst obj_data[attr + '_confidence'] = lst_confidence collected_attrs.append(obj_data) return collected_attrs class DictDiffer(object): """ Calculate the difference between two dictionaries as: (1) items added (2) items removed (3) keys same in both but changed values (4) keys same in both and unchanged values """ def __init__(self, current_dict, past_dict): self.current_dict, self.past_dict = current_dict, past_dict self.set_current, self.set_past = set(current_dict.keys()), set(past_dict.keys()) self.intersect = self.set_current.intersection(self.set_past) def added(self): return self.set_current - self.intersect def removed(self): return self.set_past - self.intersect def changed(self): return set(o for o in self.intersect if self.past_dict[o] != self.current_dict[o]) def unchanged(self): return set(o for o in self.intersect if self.past_dict[o] == self.current_dict[o]) class VersionsMixin(object): ''' Model mixin to get version diff for a given model instance ''' def _getDiff(self, differ): skip_fields = ['date_updated', 'id'] def
(change_type): for field in getattr(differ, change_type)(): if field not in skip_fields: if change_type == 'changed': yield { 'field_name': field, 'to': differ.current_dict[field], 'from': differ.past_dict[field], } elif change_type == 'added': yield { 'field_name': field, 'to': differ.current_dict[field], 'from': None } elif change_type == 'removed': yield { 'field_name': field, 'to': None, 'from': differ.past_dict[field] } additions = [a for a in makeIt('added')] changes = [c for c in makeIt('changed')] removals = [r for r in makeIt('removed')] return additions, changes, removals def getRevisions(self, versions): from source.models import Source, AccessPoint revisions = [] for version in versions: complete_revision = { 'id': version.revision.id } revision_meta = { 'modification_date': version.revision.date_created, 'comment': version.revision.comment, 'user': version.revision.user, } complex_list_models = [c.field_model._meta.model_name for c in getattr(self, 'complex_lists', [])] for object_property in version.revision.version_set.all(): if object_property.object != self or isinstance(self, Source): serialized_data = json.loads(object_property.serialized_data)[0] # a bit of a hack in order to get sources and access points # to work field_names = [] if 'value' in serialized_data['fields']: field_names.append((serialized_data['fields']['value'], serialized_data['model'].split('.')[1])) else: for field in serialized_data['fields']: field_names.append((serialized_data['fields'][field], field)) for value, field_name in field_names: if field_name in complex_list_models: try: complete_revision[field_name].add(value) except KeyError: complete_revision[field_name] = {value} else: complete_revision[field_name] = value revisions.append((complete_revision, version.revision)) return revisions def getDifferences(self, revisions): differences = [] for index, (version, revision) in enumerate(revisions): if (index - 1) > 0: try: previous, previous_revision = revisions[index - 1] except (IndexError, AssertionError): continue else: continue differ = DictDiffer(previous, version) fields_added, fields_changed, fields_removed = self._getDiff(differ) diff = { 'modification_date': previous_revision.date_created, 'comment': previous_revision.comment, 'user': previous_revision.user, 'from_id': version['id'], 'to_id': previous_revision.id, 'fields_added': fields_added, 'fields_changed': fields_changed, 'fields_removed': fields_removed, 'model': self._meta.object_name, } differences.append(diff) return differences def getVersions(self, versions=None): if not versions: versions = Version.objects.get_for_object(self) revisions = self.getRevisions(versions) return self.getDifferences(revisions) def execute_sql(file_path): ''' Execute arbitrary SQL code from a file location. ''' with open(file_path) as f: statements = f.read().split(';') with connection.cursor() as c: for statement in statements: if statement.strip(): c.execute(statement.strip()) def class_for_name(class_name, module_name="person.models"): # Check for irregular class names (names where we cannot infer the class # name by capitalizing the first letter of class_name) irregular_names = ( ('Membershipperson', 'MembershipPerson'), ('Membershiporganization', 'MembershipOrganization'), ('Personextra', 'PersonExtra'), ('Personbiography', 'PersonBiography') ) for name, formatted_name in irregular_names: if class_name == name: class_name = formatted_name break if class_name not in settings.ALLOWED_CLASS_FOR_NAME: raise Exception("Unallowed class for name") module = importlib.import_module(module_name) class_ = getattr(module, class_name) return class_ def get_osm_by_id(osm_id): osm_feature = None cursor = connection.cursor() query = ''' SELECT ST_X(ST_Centroid(geometry)), ST_Y(ST_Centroid(geometry)), * FROM osm_data WHERE id = {osm_id} '''.format(osm_id=osm_id) cursor.execute(query) columns = [c[0] for c in cursor.description] results_tuple = namedtuple('OSMFeature', columns) row = cursor.fetchone() if row: osm_feature = results_tuple(*row) return osm_feature def get_hierarchy_by_id(osm_id): hierarchy = ''' SELECT parents.* FROM osm_data AS parents JOIN ( SELECT UNNEST(hierarchy) AS h_id, localname, tags, admin_level, name, geometry FROM osm_data WHERE id = %s ) AS child ON parents.id = child.h_id::integer ''' cursor = connection.cursor() cursor.execute(hierarchy, [osm_id]) columns = [c[0] for c in cursor.description] results_tuple = namedtuple('OSMFeature', columns) hierarchy = [results_tuple(*r) for r in cursor] return hierarchy def generate_hierarchy(query, q_args, rel_field, sources=False): cursor = connection.cursor() cursor.execute(query, q_args) columns = [c[0] for c in cursor.description] results_tuple = namedtuple('Organization', columns) hierarchy = [(idx, results_tuple(*r)) for idx, r in enumerate(cursor)] trimmed_hierarchy = [] for org_id, orgs in itertools.groupby(hierarchy, key=lambda x: x[1].id): group = list(orgs) lowest_index = min(g[0] for g in group) orgs = [o[1] for o in group] start_date = None if orgs[0].start_date: start_date = orgs[0].start_date end_date = None if orgs[0].end_date: end_date = orgs[0].end_date # Create a label, which we display on the charts for person and unit "parents." label = '<b>' + orgs[0].name + '</b>' + '\n\n' + _('Unknown commander') if orgs[0].commander: label = '<b>' + orgs[0].name + '</b>' + '\n\n' + orgs[0].commander trimmed = { 'id': org_id, 'label': str(label), 'detail_id': str(orgs[0].org_org_id), 'name': orgs[0].name, 'other_names': list({o.alias.strip() for o in orgs if o.alias}), 'classifications': list({o.classification.strip() for o in orgs if o.classification}), 'division_id': orgs[0].division_id, 'date_first_cited': start_date, 'date_last_cited': end_date, 'commander': orgs[0].commander, } trimmed[rel_field] = getattr(orgs[0], rel_field) if sources: trimmed['sources'] = [] source_ids = [] for o in orgs: org_source = json.loads(o.source) if org_source['id'] not in source_ids:
makeIt
identifier_name
utils.py
osm_data WHERE id = {osm_id} '''.format(osm_id=osm_id) cursor.execute(query) columns = [c[0] for c in cursor.description] results_tuple = namedtuple('OSMFeature', columns) row = cursor.fetchone() if row: osm_feature = results_tuple(*row) return osm_feature def get_hierarchy_by_id(osm_id): hierarchy = ''' SELECT parents.* FROM osm_data AS parents JOIN ( SELECT UNNEST(hierarchy) AS h_id, localname, tags, admin_level, name, geometry FROM osm_data WHERE id = %s ) AS child ON parents.id = child.h_id::integer ''' cursor = connection.cursor() cursor.execute(hierarchy, [osm_id]) columns = [c[0] for c in cursor.description] results_tuple = namedtuple('OSMFeature', columns) hierarchy = [results_tuple(*r) for r in cursor] return hierarchy def generate_hierarchy(query, q_args, rel_field, sources=False): cursor = connection.cursor() cursor.execute(query, q_args) columns = [c[0] for c in cursor.description] results_tuple = namedtuple('Organization', columns) hierarchy = [(idx, results_tuple(*r)) for idx, r in enumerate(cursor)] trimmed_hierarchy = [] for org_id, orgs in itertools.groupby(hierarchy, key=lambda x: x[1].id): group = list(orgs) lowest_index = min(g[0] for g in group) orgs = [o[1] for o in group] start_date = None if orgs[0].start_date: start_date = orgs[0].start_date end_date = None if orgs[0].end_date: end_date = orgs[0].end_date # Create a label, which we display on the charts for person and unit "parents." label = '<b>' + orgs[0].name + '</b>' + '\n\n' + _('Unknown commander') if orgs[0].commander: label = '<b>' + orgs[0].name + '</b>' + '\n\n' + orgs[0].commander trimmed = { 'id': org_id, 'label': str(label), 'detail_id': str(orgs[0].org_org_id), 'name': orgs[0].name, 'other_names': list({o.alias.strip() for o in orgs if o.alias}), 'classifications': list({o.classification.strip() for o in orgs if o.classification}), 'division_id': orgs[0].division_id, 'date_first_cited': start_date, 'date_last_cited': end_date, 'commander': orgs[0].commander, } trimmed[rel_field] = getattr(orgs[0], rel_field) if sources: trimmed['sources'] = [] source_ids = [] for o in orgs: org_source = json.loads(o.source) if org_source['id'] not in source_ids: trimmed['sources'].append(org_source) source_ids.append(org_source['id']) trimmed['confidence'] = REVERSE_CONFIDENCE[int(orgs[0].confidence)].title() trimmed_hierarchy.append((lowest_index, trimmed)) hierarchy = [i[1] for i in sorted(trimmed_hierarchy, key=lambda x: x[0])] return hierarchy # this makes an edge list that shows the parent relationships (see child_id) solr = pysolr.Solr(settings.SOLR_URL) def get_org_hierarchy_by_id(org_id, when=None, sources=False, direction='up', authenticated=False): ''' org_id: uuid for the organization when: date for limiting the search ''' base_url = settings.SOLR_URL if direction == 'up': from_ = 'child' to = 'parent' elif direction == 'down': from_ = 'parent' to = 'child' filter_query = '{!graph from=composition_%s_id_s_fct to=composition_%s_id_s_fct returnRoot=true}composition_%s_id_s_fct:%s' % (from_, to, from_, org_id) if when: filter_query += ' AND {!field f=composition_daterange_dr op=contains}%s' % when if not authenticated: filter_query += ' AND published_b:T' results = solr.search('*:*', fq=filter_query) if when: for result in results: for key in [from_, to]: org = result['composition_{}_id_s_fct'.format(key)] args = (org, when) query = 'commander_org_id_s_fct:%s AND {!field f=commander_assignment_range_dr op=contains}%s' % args if not authenticated: query += ' AND published_b:T' commanders = solr.search(query) # We need to deduplicate commanders and then throw out the open ended date ranges. result['commanders-{}'.format(key)] = [] for commander in commanders: label_fmt = '{name} ({start} - {end})' assignment_range = commander['commander_assignment_range_dr'] start, end = assignment_range.replace('[', '').replace(']', '').split(' TO ') if start == '*': start = '?' if end == '*': end = '?' label = label_fmt.format(name=commander['commander_person_name_s'], start=start, end=end) commander['label'] = label result['commanders-{}'.format(key)].append(commander) return results def get_child_orgs_by_id(org_id, when=None, sources=False): hierarchy = ''' WITH RECURSIVE parents AS ( SELECT o.*, NULL::VARCHAR AS parent_id, NULL::VARCHAR AS parent_name, NULL::DATE AS start_date, NULL::DATE AS end_date, NULL::VARCHAR AS comp_open_ended, NULL::VARCHAR AS source, NULL::VARCHAR AS confidence FROM organization As o WHERE id = %s UNION SELECT o.*, h.parent_id::VARCHAR AS parent_id, parents.name AS parent_name, h.start_date::date, h.end_date::date, h.open_ended AS comp_open_ended, row_to_json(ss.*)::VARCHAR AS source, ccc.confidence FROM organization AS o JOIN composition AS h ON o.id = h.child_id JOIN composition_compositionchild AS ccc ON h.id = ccc.object_ref_id LEFT JOIN composition_compositionchild_sources AS cccs ON ccc.id = cccs.compositionchild_id LEFT JOIN source_source AS ss ON cccs.source_id = ss.uuid JOIN parents ON parents.id = h.parent_id ) SELECT * FROM parents WHERE id != %s ''' q_args = [org_id, org_id] if when: hierarchy = ''' {} AND CASE WHEN (start_date IS NOT NULL AND end_date IS NOT NULL AND comp_open_ended IN ('N', 'E')) THEN (%s::date BETWEEN start_date::date AND end_date::date) WHEN (start_date IS NOT NULL AND end_date IS NOT NULL AND comp_open_ended = 'Y') THEN (%s::date BETWEEN start_date::date AND NOW()::date) WHEN (start_date IS NOT NULL AND end_date IS NULL AND comp_open_ended IN ('N', 'E')) THEN (start_date::date = %s::date) WHEN (start_date IS NOT NULL AND end_date IS NULL AND comp_open_ended = 'Y') THEN (%s::date BETWEEN start_date::date AND NOW()::date) WHEN (start_date IS NULL AND end_date IS NOT NULL AND comp_open_ended IN ('N', 'E')) THEN (end_date::date = %s) WHEN (start_date IS NULL AND end_date IS NOT NULL AND comp_open_ended = 'Y') THEN TRUE END '''.format(hierarchy) q_args.extend([when] * 5) hierarchy = '{} ORDER BY id'.format(hierarchy) hierarchy = generate_hierarchy(hierarchy, q_args, 'parent_id', sources=sources) return hierarchy def deleted_in_str(objects): index = 0 for obj in objects: if isinstance(obj, list): objects[index] = deleted_in_str(obj) else: if hasattr(obj, 'field_name'): name = obj.field_name + ": " + str(obj) else: name = type(obj).__name__ + ": " + str(obj) if '_sources' in name: objects[index] = "Object sources" else: objects[index] = name index += 1 return objects def import_class(cl):
d = cl.rfind('.') classname = cl[d+1:len(cl)] m = __import__(cl[0:d], globals(), locals(), [classname]) return getattr(m, classname)
identifier_body
listing-ctrl.js
.Object.extend("listing"); var userquery = new Parse.Query(Parse.User); var query = new Parse.Query(Listing); var listing = $stateParams.listingId; console.log(listing); query.get(listing, { success: function(listing) { console.log(listing); $scope.listing = listing; $scope.listing.listingid = listing.id; $scope.listing.ownerId = listing.get('ownerId'); $scope.listing.apps = listing.get('applications'); $scope.listing.appids = _.pluck($scope.listing.apps, 'id'); $scope.listing.address = listing.get("address"); $scope.pageTitle = $scope.listing.address; $scope.place = { types: ['street_address'] }; $scope.listing.beds = listing.get("beds"); $scope.listing.baths = listing.get("baths"); $scope.listing.rent = listing.get("rent"); $scope.listing.aptnum = listing.get("aptnum") $scope.listing.tempdt = listing.get("availableDate"); $scope.listing.dt = $scope.listing.tempdt.toLocaleDateString(); $scope.listing.tempenddt = listing.get("enddate"); $scope.listing.enddt = $scope.listing.tempenddt.toLocaleDateString(); $scope.listing.gender = listing.get("gender"); $scope.listing.desc = listing.get("description"); $scope.listing.templaundry = listing.get("laundry"); $scope.listing.tempair = listing.get("air"); $scope.listing.appcount = listing.get('applications').length; $scope.pictureFiles[0] = $scope.listing.get("photo0"); $scope.pictureFiles[1] = $scope.listing.get("photo1"); $scope.pictureFiles[2] = $scope.listing.get("photo2"); $scope.pictureFiles[3] = $scope.listing.get("photo3"); $scope.pictureFiles[4] = $scope.listing.get("photo4"); $scope.picUrls[0] = $scope.pictureFiles[0].url(); $scope.picUrls[1] = $scope.pictureFiles[1].url(); $scope.picUrls[2] = $scope.pictureFiles[2].url(); $scope.picUrls[3] = $scope.pictureFiles[3].url(); $scope.picUrls[4] = $scope.pictureFiles[4].url(); $scope.share = {}; var desclen = $scope.listing.desc.length; var desccut = ""; if (desclen < 141) { desccut = $scope.listing.desc; } else { desccut = $scope.listing.desc.substr(0, 140) + "..."; }; $scope.share.title = $scope.listing.beds + " bed, " + $scope.listing.baths + "bath for $" + $scope.listing.rent + "/month"; $scope.share.desc = "Located at " + $scope.listing.address + " and available starting " + $scope.listing.dt + ". " + desccut; $scope.share.url = "http://piip.parseapp.com/#!/listing/" + $scope.listing.listingid; $scope.share.geopoint = $scope.listing.get('geopoint'); //$scope.share.url = "http://piip.parseapp.com/?_escaped_fragment_=%2Flisting%2F" + $scope.listing.listingid; $scope.share.image = $scope.picUrls[0]; $scope.$emit('metaUpdate', $scope.share); $("<img/>").attr("src", $scope.picUrls[0]).load(function(){ s = {w:this.width, h:this.height}; $scope.share.imgw = s.w $scope.share.imgh = s.h; $scope.$emit('metaUpdate', $scope.share); console.log('$emit triggered'); }); if ($scope.listing.gender == 'both') { $scope.listing.gender = 'Male and Female'; } if ($scope.listing.templaundry == true) { $scope.listing.laundry = 'In Building'; } if ($scope.listing.tempair == true) { $scope.listing.air = 'A/C'; } $scope.map = { center: { latitude: $scope.listing.attributes.geopoint._latitude, longitude: $scope.listing.attributes.geopoint._longitude }, zoom: 13 // options: { // types: "locality, neighborhood, postal_town, colloquial_area, political, postal_code" // } }; $scope.marker = { id: listing.id, coords: { latitude: $scope.listing.attributes.geopoint._latitude, longitude: $scope.listing.attributes.geopoint._longitude }, }; uiGmapGoogleMapApi.then(function(maps) { }); userquery.get($scope.listing.ownerId, { success: function(profile) { $scope.profile = profile; $scope.profile.name = profile.get('name'); $scope.profile.email = profile.get('email'); $scope.profile.gender = profile.get('gender'); $scope.profile.age = profile.get('age'); $scope.profile.education = profile.get('education').reverse(); $scope.profile.work = profile.get('work'); $scope.profile.propics = profile.get('imgUser'); }, error: function(error) { console.log(error); } }) }, error: function(error) { console.log(error); } }); $scope.listflag = function () { $scope.listing.increment('flag'); $scope.listing.save(); $scope.flagged = true; } $scope.apply = function () { if ($rootScope.currentUser) { var tempindex = _.indexOf($scope.listing.appids, $rootScope.currentUser.id); if (tempindex == -1) { $scope.listing.addUnique('applications', { id: $rootScope.currentUser.id, appstatus: 'new', chatstatus: false, modifieddt: new Date() }); $rootScope.currentUser.addUnique('applications', $scope.listing.listingid); $scope.listing.save(null, { success: function(result) { // Notification Email Parse.Cloud.run('emailNewApp', { recname: $scope.profile.name, recemail: $scope.profile.email, sender: $rootScope.currentUser.get('name') }, { success: function(message) { }, error: function(result, error) { console.log(error) } }); }, error: function(result, error) { console.log(error); } }); $rootScope.currentUser.save(null, { success: function(result) { $scope.applyalert = "Applied!"; }, error: function(error) { console.log(error); } }); } else { $scope.applyalert = "Applied!"; }; }; } $scope.fbConnect = function () { // NB: this is a contrived example for demo purposes, you would never write the following code in a real app // normally you would define a User.js data module for all your user objects and the method below would be on the user, e.g. $rootScope.currentUser.fbConnect() Parse.FacebookUtils.logIn("public_profile,email,user_birthday,user_education_history,user_photos,user_work_history", {}).then(function(user) { console.log('facebook connected!'); $rootScope.currentUser = Parse.User.current(); updateProf(); }, function(error) { console.log('something went wrong, try again'); }); } $scope.fbShare = function() { FB.ui({ method: 'share', url: 'http://piip.parseapp.com/#!/listing/' + $scope.listing.listingid, title: $scope.share.title, description: $scope.share.desc, image: $scope.share.image, href: 'http://piip.parseapp.com/#!/listing/' + $scope.listing.listingid }, function(response){}); } var user = {}; function
(dateString) { var today = new Date(); var birthDate = new Date(dateString); var age = today.getFullYear() - birthDate.getFullYear(); var m = today.getMonth() - birthDate.getMonth(); if (m < 0 || (m === 0 && today.getDate() < birthDate.getDate())) { age--; } return age; } // FACEBOOK PROFILE UPDATE updateProf = function () { FB.apiAngular('me?fields=id,name,first_name,last_name,email,birthday,education,gender,work,albums') .then(function (profileData) { user.profileData = profileData; user.age = calculateAge(profileData.birthday); }) .then(function (resList) { user.profilePhotos = []; for (var i = user.profileData.albums.data.length - 1; i >= 0; i--) { if (user.profileData.albums.data[i].name == "Profile Pictures") { user.albumid = user.profileData.albums.data[i].
calculateAge
identifier_name
listing-ctrl.js
"); var userquery = new Parse.Query(Parse.User); var query = new Parse.Query(Listing); var listing = $stateParams.listingId; console.log(listing); query.get(listing, { success: function(listing) { console.log(listing); $scope.listing = listing; $scope.listing.listingid = listing.id; $scope.listing.ownerId = listing.get('ownerId'); $scope.listing.apps = listing.get('applications'); $scope.listing.appids = _.pluck($scope.listing.apps, 'id'); $scope.listing.address = listing.get("address"); $scope.pageTitle = $scope.listing.address; $scope.place = { types: ['street_address'] }; $scope.listing.beds = listing.get("beds"); $scope.listing.baths = listing.get("baths"); $scope.listing.rent = listing.get("rent"); $scope.listing.aptnum = listing.get("aptnum") $scope.listing.tempdt = listing.get("availableDate"); $scope.listing.dt = $scope.listing.tempdt.toLocaleDateString(); $scope.listing.tempenddt = listing.get("enddate"); $scope.listing.enddt = $scope.listing.tempenddt.toLocaleDateString(); $scope.listing.gender = listing.get("gender"); $scope.listing.desc = listing.get("description"); $scope.listing.templaundry = listing.get("laundry"); $scope.listing.tempair = listing.get("air"); $scope.listing.appcount = listing.get('applications').length; $scope.pictureFiles[0] = $scope.listing.get("photo0"); $scope.pictureFiles[1] = $scope.listing.get("photo1"); $scope.pictureFiles[2] = $scope.listing.get("photo2"); $scope.pictureFiles[3] = $scope.listing.get("photo3"); $scope.pictureFiles[4] = $scope.listing.get("photo4"); $scope.picUrls[0] = $scope.pictureFiles[0].url(); $scope.picUrls[1] = $scope.pictureFiles[1].url(); $scope.picUrls[2] = $scope.pictureFiles[2].url(); $scope.picUrls[3] = $scope.pictureFiles[3].url(); $scope.picUrls[4] = $scope.pictureFiles[4].url(); $scope.share = {}; var desclen = $scope.listing.desc.length; var desccut = ""; if (desclen < 141) { desccut = $scope.listing.desc; } else { desccut = $scope.listing.desc.substr(0, 140) + "..."; }; $scope.share.title = $scope.listing.beds + " bed, " + $scope.listing.baths + "bath for $" + $scope.listing.rent + "/month"; $scope.share.desc = "Located at " + $scope.listing.address + " and available starting " + $scope.listing.dt + ". " + desccut; $scope.share.url = "http://piip.parseapp.com/#!/listing/" + $scope.listing.listingid; $scope.share.geopoint = $scope.listing.get('geopoint'); //$scope.share.url = "http://piip.parseapp.com/?_escaped_fragment_=%2Flisting%2F" + $scope.listing.listingid; $scope.share.image = $scope.picUrls[0]; $scope.$emit('metaUpdate', $scope.share); $("<img/>").attr("src", $scope.picUrls[0]).load(function(){ s = {w:this.width, h:this.height}; $scope.share.imgw = s.w $scope.share.imgh = s.h; $scope.$emit('metaUpdate', $scope.share); console.log('$emit triggered'); }); if ($scope.listing.gender == 'both') { $scope.listing.gender = 'Male and Female'; } if ($scope.listing.templaundry == true) { $scope.listing.laundry = 'In Building'; } if ($scope.listing.tempair == true) { $scope.listing.air = 'A/C'; } $scope.map = { center: { latitude: $scope.listing.attributes.geopoint._latitude, longitude: $scope.listing.attributes.geopoint._longitude }, zoom: 13 // options: { // types: "locality, neighborhood, postal_town, colloquial_area, political, postal_code" // } }; $scope.marker = { id: listing.id, coords: { latitude: $scope.listing.attributes.geopoint._latitude, longitude: $scope.listing.attributes.geopoint._longitude }, }; uiGmapGoogleMapApi.then(function(maps) { }); userquery.get($scope.listing.ownerId, { success: function(profile) { $scope.profile = profile; $scope.profile.name = profile.get('name'); $scope.profile.email = profile.get('email'); $scope.profile.gender = profile.get('gender'); $scope.profile.age = profile.get('age'); $scope.profile.education = profile.get('education').reverse(); $scope.profile.work = profile.get('work'); $scope.profile.propics = profile.get('imgUser'); }, error: function(error) { console.log(error); } }) }, error: function(error) { console.log(error); } }); $scope.listflag = function () { $scope.listing.increment('flag'); $scope.listing.save(); $scope.flagged = true; } $scope.apply = function () { if ($rootScope.currentUser) { var tempindex = _.indexOf($scope.listing.appids, $rootScope.currentUser.id); if (tempindex == -1) { $scope.listing.addUnique('applications', { id: $rootScope.currentUser.id, appstatus: 'new', chatstatus: false, modifieddt: new Date() }); $rootScope.currentUser.addUnique('applications', $scope.listing.listingid); $scope.listing.save(null, { success: function(result) { // Notification Email Parse.Cloud.run('emailNewApp', { recname: $scope.profile.name, recemail: $scope.profile.email, sender: $rootScope.currentUser.get('name') }, { success: function(message) { }, error: function(result, error) { console.log(error) } }); }, error: function(result, error) { console.log(error); } }); $rootScope.currentUser.save(null, { success: function(result) { $scope.applyalert = "Applied!"; }, error: function(error) { console.log(error); } }); } else { $scope.applyalert = "Applied!"; }; }; } $scope.fbConnect = function () { // NB: this is a contrived example for demo purposes, you would never write the following code in a real app // normally you would define a User.js data module for all your user objects and the method below would be on the user, e.g. $rootScope.currentUser.fbConnect() Parse.FacebookUtils.logIn("public_profile,email,user_birthday,user_education_history,user_photos,user_work_history", {}).then(function(user) { console.log('facebook connected!'); $rootScope.currentUser = Parse.User.current(); updateProf(); }, function(error) { console.log('something went wrong, try again'); }); } $scope.fbShare = function() { FB.ui({ method: 'share', url: 'http://piip.parseapp.com/#!/listing/' + $scope.listing.listingid, title: $scope.share.title, description: $scope.share.desc, image: $scope.share.image, href: 'http://piip.parseapp.com/#!/listing/' + $scope.listing.listingid }, function(response){}); } var user = {}; function calculateAge(dateString) { var today = new Date(); var birthDate = new Date(dateString); var age = today.getFullYear() - birthDate.getFullYear(); var m = today.getMonth() - birthDate.getMonth(); if (m < 0 || (m === 0 && today.getDate() < birthDate.getDate())) { age--; } return age; } // FACEBOOK PROFILE UPDATE updateProf = function () { FB.apiAngular('me?fields=id,name,first_name,last_name,email,birthday,education,gender,work,albums') .then(function (profileData) { user.profileData = profileData; user.age = calculateAge(profileData.birthday); }) .then(function (resList) { user.profilePhotos = []; for (var i = user.profileData.albums.data.length - 1; i >= 0; i--) { if (user.profileData.albums.data[i].name == "Profile Pictures")
{ user.albumid = user.profileData.albums.data[i].id; }
conditional_block
listing-ctrl.js
v: '3.17', libraries: 'places' }); }) .controller("ListingCtrl", function($scope, $rootScope, $state, $stateParams, $modal, fbookFactory, FacebookAngularPatch, uiGmapGoogleMapApi, markersFactory) { $scope.pictureFiles = []; $scope.picUrls = []; var Listing = Parse.Object.extend("listing"); var userquery = new Parse.Query(Parse.User); var query = new Parse.Query(Listing); var listing = $stateParams.listingId; console.log(listing); query.get(listing, { success: function(listing) { console.log(listing); $scope.listing = listing; $scope.listing.listingid = listing.id; $scope.listing.ownerId = listing.get('ownerId'); $scope.listing.apps = listing.get('applications'); $scope.listing.appids = _.pluck($scope.listing.apps, 'id'); $scope.listing.address = listing.get("address"); $scope.pageTitle = $scope.listing.address; $scope.place = { types: ['street_address'] }; $scope.listing.beds = listing.get("beds"); $scope.listing.baths = listing.get("baths"); $scope.listing.rent = listing.get("rent"); $scope.listing.aptnum = listing.get("aptnum") $scope.listing.tempdt = listing.get("availableDate"); $scope.listing.dt = $scope.listing.tempdt.toLocaleDateString(); $scope.listing.tempenddt = listing.get("enddate"); $scope.listing.enddt = $scope.listing.tempenddt.toLocaleDateString(); $scope.listing.gender = listing.get("gender"); $scope.listing.desc = listing.get("description"); $scope.listing.templaundry = listing.get("laundry"); $scope.listing.tempair = listing.get("air"); $scope.listing.appcount = listing.get('applications').length; $scope.pictureFiles[0] = $scope.listing.get("photo0"); $scope.pictureFiles[1] = $scope.listing.get("photo1"); $scope.pictureFiles[2] = $scope.listing.get("photo2"); $scope.pictureFiles[3] = $scope.listing.get("photo3"); $scope.pictureFiles[4] = $scope.listing.get("photo4"); $scope.picUrls[0] = $scope.pictureFiles[0].url(); $scope.picUrls[1] = $scope.pictureFiles[1].url(); $scope.picUrls[2] = $scope.pictureFiles[2].url(); $scope.picUrls[3] = $scope.pictureFiles[3].url(); $scope.picUrls[4] = $scope.pictureFiles[4].url(); $scope.share = {}; var desclen = $scope.listing.desc.length; var desccut = ""; if (desclen < 141) { desccut = $scope.listing.desc; } else { desccut = $scope.listing.desc.substr(0, 140) + "..."; }; $scope.share.title = $scope.listing.beds + " bed, " + $scope.listing.baths + "bath for $" + $scope.listing.rent + "/month"; $scope.share.desc = "Located at " + $scope.listing.address + " and available starting " + $scope.listing.dt + ". " + desccut; $scope.share.url = "http://piip.parseapp.com/#!/listing/" + $scope.listing.listingid; $scope.share.geopoint = $scope.listing.get('geopoint'); //$scope.share.url = "http://piip.parseapp.com/?_escaped_fragment_=%2Flisting%2F" + $scope.listing.listingid; $scope.share.image = $scope.picUrls[0]; $scope.$emit('metaUpdate', $scope.share); $("<img/>").attr("src", $scope.picUrls[0]).load(function(){ s = {w:this.width, h:this.height}; $scope.share.imgw = s.w $scope.share.imgh = s.h; $scope.$emit('metaUpdate', $scope.share); console.log('$emit triggered'); }); if ($scope.listing.gender == 'both') { $scope.listing.gender = 'Male and Female'; } if ($scope.listing.templaundry == true) { $scope.listing.laundry = 'In Building'; } if ($scope.listing.tempair == true) { $scope.listing.air = 'A/C'; } $scope.map = { center: { latitude: $scope.listing.attributes.geopoint._latitude, longitude: $scope.listing.attributes.geopoint._longitude }, zoom: 13 // options: { // types: "locality, neighborhood, postal_town, colloquial_area, political, postal_code" // } }; $scope.marker = { id: listing.id, coords: { latitude: $scope.listing.attributes.geopoint._latitude, longitude: $scope.listing.attributes.geopoint._longitude }, }; uiGmapGoogleMapApi.then(function(maps) { }); userquery.get($scope.listing.ownerId, { success: function(profile) { $scope.profile = profile; $scope.profile.name = profile.get('name'); $scope.profile.email = profile.get('email'); $scope.profile.gender = profile.get('gender'); $scope.profile.age = profile.get('age'); $scope.profile.education = profile.get('education').reverse(); $scope.profile.work = profile.get('work'); $scope.profile.propics = profile.get('imgUser'); }, error: function(error) { console.log(error); } }) }, error: function(error) { console.log(error); } }); $scope.listflag = function () { $scope.listing.increment('flag'); $scope.listing.save(); $scope.flagged = true; } $scope.apply = function () { if ($rootScope.currentUser) { var tempindex = _.indexOf($scope.listing.appids, $rootScope.currentUser.id); if (tempindex == -1) { $scope.listing.addUnique('applications', { id: $rootScope.currentUser.id, appstatus: 'new', chatstatus: false, modifieddt: new Date() }); $rootScope.currentUser.addUnique('applications', $scope.listing.listingid); $scope.listing.save(null, { success: function(result) { // Notification Email Parse.Cloud.run('emailNewApp', { recname: $scope.profile.name, recemail: $scope.profile.email, sender: $rootScope.currentUser.get('name') }, { success: function(message) { }, error: function(result, error) { console.log(error) } }); }, error: function(result, error) { console.log(error); } }); $rootScope.currentUser.save(null, { success: function(result) { $scope.applyalert = "Applied!"; }, error: function(error) { console.log(error); } }); } else { $scope.applyalert = "Applied!"; }; }; } $scope.fbConnect = function () { // NB: this is a contrived example for demo purposes, you would never write the following code in a real app // normally you would define a User.js data module for all your user objects and the method below would be on the user, e.g. $rootScope.currentUser.fbConnect() Parse.FacebookUtils.logIn("public_profile,email,user_birthday,user_education_history,user_photos,user_work_history", {}).then(function(user) { console.log('facebook connected!'); $rootScope.currentUser = Parse.User.current(); updateProf(); }, function(error) { console.log('something went wrong, try again'); }); } $scope.fbShare = function() { FB.ui({ method: 'share', url: 'http://piip.parseapp.com/#!/listing/' + $scope.listing.listingid, title: $scope.share.title, description: $scope.share.desc, image: $scope.share.image, href: 'http://piip.parseapp.com/#!/listing/' + $scope.listing.listingid }, function(response){}); } var user = {}; function calculateAge(dateString) { var today = new Date(); var birthDate = new Date(dateString); var age = today.getFullYear() - birthDate.getFullYear(); var m = today.getMonth() - birthDate.getMonth(); if (m < 0 || (m === 0 && today.getDate() < birthDate.getDate())) { age--; } return age; } // FACEBOOK PROFILE UPDATE updateProf =
.config(function(uiGmapGoogleMapApiProvider) { console.log('config maps'); uiGmapGoogleMapApiProvider.configure({ key: 'AIzaSyCCMEJsPzyGW-oLOShTOJw_Pe9Qv2YMAZo',
random_line_split
listing-ctrl.js
.Object.extend("listing"); var userquery = new Parse.Query(Parse.User); var query = new Parse.Query(Listing); var listing = $stateParams.listingId; console.log(listing); query.get(listing, { success: function(listing) { console.log(listing); $scope.listing = listing; $scope.listing.listingid = listing.id; $scope.listing.ownerId = listing.get('ownerId'); $scope.listing.apps = listing.get('applications'); $scope.listing.appids = _.pluck($scope.listing.apps, 'id'); $scope.listing.address = listing.get("address"); $scope.pageTitle = $scope.listing.address; $scope.place = { types: ['street_address'] }; $scope.listing.beds = listing.get("beds"); $scope.listing.baths = listing.get("baths"); $scope.listing.rent = listing.get("rent"); $scope.listing.aptnum = listing.get("aptnum") $scope.listing.tempdt = listing.get("availableDate"); $scope.listing.dt = $scope.listing.tempdt.toLocaleDateString(); $scope.listing.tempenddt = listing.get("enddate"); $scope.listing.enddt = $scope.listing.tempenddt.toLocaleDateString(); $scope.listing.gender = listing.get("gender"); $scope.listing.desc = listing.get("description"); $scope.listing.templaundry = listing.get("laundry"); $scope.listing.tempair = listing.get("air"); $scope.listing.appcount = listing.get('applications').length; $scope.pictureFiles[0] = $scope.listing.get("photo0"); $scope.pictureFiles[1] = $scope.listing.get("photo1"); $scope.pictureFiles[2] = $scope.listing.get("photo2"); $scope.pictureFiles[3] = $scope.listing.get("photo3"); $scope.pictureFiles[4] = $scope.listing.get("photo4"); $scope.picUrls[0] = $scope.pictureFiles[0].url(); $scope.picUrls[1] = $scope.pictureFiles[1].url(); $scope.picUrls[2] = $scope.pictureFiles[2].url(); $scope.picUrls[3] = $scope.pictureFiles[3].url(); $scope.picUrls[4] = $scope.pictureFiles[4].url(); $scope.share = {}; var desclen = $scope.listing.desc.length; var desccut = ""; if (desclen < 141) { desccut = $scope.listing.desc; } else { desccut = $scope.listing.desc.substr(0, 140) + "..."; }; $scope.share.title = $scope.listing.beds + " bed, " + $scope.listing.baths + "bath for $" + $scope.listing.rent + "/month"; $scope.share.desc = "Located at " + $scope.listing.address + " and available starting " + $scope.listing.dt + ". " + desccut; $scope.share.url = "http://piip.parseapp.com/#!/listing/" + $scope.listing.listingid; $scope.share.geopoint = $scope.listing.get('geopoint'); //$scope.share.url = "http://piip.parseapp.com/?_escaped_fragment_=%2Flisting%2F" + $scope.listing.listingid; $scope.share.image = $scope.picUrls[0]; $scope.$emit('metaUpdate', $scope.share); $("<img/>").attr("src", $scope.picUrls[0]).load(function(){ s = {w:this.width, h:this.height}; $scope.share.imgw = s.w $scope.share.imgh = s.h; $scope.$emit('metaUpdate', $scope.share); console.log('$emit triggered'); }); if ($scope.listing.gender == 'both') { $scope.listing.gender = 'Male and Female'; } if ($scope.listing.templaundry == true) { $scope.listing.laundry = 'In Building'; } if ($scope.listing.tempair == true) { $scope.listing.air = 'A/C'; } $scope.map = { center: { latitude: $scope.listing.attributes.geopoint._latitude, longitude: $scope.listing.attributes.geopoint._longitude }, zoom: 13 // options: { // types: "locality, neighborhood, postal_town, colloquial_area, political, postal_code" // } }; $scope.marker = { id: listing.id, coords: { latitude: $scope.listing.attributes.geopoint._latitude, longitude: $scope.listing.attributes.geopoint._longitude }, }; uiGmapGoogleMapApi.then(function(maps) { }); userquery.get($scope.listing.ownerId, { success: function(profile) { $scope.profile = profile; $scope.profile.name = profile.get('name'); $scope.profile.email = profile.get('email'); $scope.profile.gender = profile.get('gender'); $scope.profile.age = profile.get('age'); $scope.profile.education = profile.get('education').reverse(); $scope.profile.work = profile.get('work'); $scope.profile.propics = profile.get('imgUser'); }, error: function(error) { console.log(error); } }) }, error: function(error) { console.log(error); } }); $scope.listflag = function () { $scope.listing.increment('flag'); $scope.listing.save(); $scope.flagged = true; } $scope.apply = function () { if ($rootScope.currentUser) { var tempindex = _.indexOf($scope.listing.appids, $rootScope.currentUser.id); if (tempindex == -1) { $scope.listing.addUnique('applications', { id: $rootScope.currentUser.id, appstatus: 'new', chatstatus: false, modifieddt: new Date() }); $rootScope.currentUser.addUnique('applications', $scope.listing.listingid); $scope.listing.save(null, { success: function(result) { // Notification Email Parse.Cloud.run('emailNewApp', { recname: $scope.profile.name, recemail: $scope.profile.email, sender: $rootScope.currentUser.get('name') }, { success: function(message) { }, error: function(result, error) { console.log(error) } }); }, error: function(result, error) { console.log(error); } }); $rootScope.currentUser.save(null, { success: function(result) { $scope.applyalert = "Applied!"; }, error: function(error) { console.log(error); } }); } else { $scope.applyalert = "Applied!"; }; }; } $scope.fbConnect = function () { // NB: this is a contrived example for demo purposes, you would never write the following code in a real app // normally you would define a User.js data module for all your user objects and the method below would be on the user, e.g. $rootScope.currentUser.fbConnect() Parse.FacebookUtils.logIn("public_profile,email,user_birthday,user_education_history,user_photos,user_work_history", {}).then(function(user) { console.log('facebook connected!'); $rootScope.currentUser = Parse.User.current(); updateProf(); }, function(error) { console.log('something went wrong, try again'); }); } $scope.fbShare = function() { FB.ui({ method: 'share', url: 'http://piip.parseapp.com/#!/listing/' + $scope.listing.listingid, title: $scope.share.title, description: $scope.share.desc, image: $scope.share.image, href: 'http://piip.parseapp.com/#!/listing/' + $scope.listing.listingid }, function(response){}); } var user = {}; function calculateAge(dateString)
// FACEBOOK PROFILE UPDATE updateProf = function () { FB.apiAngular('me?fields=id,name,first_name,last_name,email,birthday,education,gender,work,albums') .then(function (profileData) { user.profileData = profileData; user.age = calculateAge(profileData.birthday); }) .then(function (resList) { user.profilePhotos = []; for (var i = user.profileData.albums.data.length - 1; i >= 0; i--) { if (user.profileData.albums.data[i].name == "Profile Pictures") { user.albumid = user.profileData.albums.data[i
{ var today = new Date(); var birthDate = new Date(dateString); var age = today.getFullYear() - birthDate.getFullYear(); var m = today.getMonth() - birthDate.getMonth(); if (m < 0 || (m === 0 && today.getDate() < birthDate.getDate())) { age--; } return age; }
identifier_body
main.rs
will not be /// carried over to the output file because there is not a one-for-one feature correspondence between the /// two files due to the joins and splits of stream segments. Instead the output attribute table will /// only contain a feature ID (FID) entry. /// /// > Note: this tool should be used to pre-process vector streams that are input to the /// > `VectorStreamNetworkAnalysis` tool. /// /// # See Also /// `VectorStreamNetworkAnalysis`, `FixDanglingArcs` fn main() { let args: Vec<String> = env::args().collect(); if args[1].trim() == "run" { match run(&args) { Ok(_) => {} Err(e) => panic!("{:?}", e), } } if args.len() <= 1 || args[1].trim() == "help" { // print help help(); } if args[1].trim() == "version" { // print version information version(); } } fn help() { let mut ext = ""; if cfg!(target_os = "windows") { ext = ".exe"; } let exe_name = &format!("correct_stream_vector_direction{}", ext); let sep: String = path::MAIN_SEPARATOR.to_string(); let s = r#" This tool resolves topological errors and inconsistencies associated with digitized vector streams. The following commands are recognized: help Prints help information. run Runs the tool. version Prints the tool version information. The following flags can be used with the 'run' command: --routes Name of the input routes vector file. -o, --output Name of the output HTML file. --length Maximum segment length (m). --dist Search distance, in grid cells, used in visibility analysis.
working directory contained in the WhiteboxTools settings.json file. Example Usage: >> .*EXE_NAME run --routes=footpath.shp --dem=DEM.tif -o=assessedRoutes.shp --length=50.0 --dist=200 Note: Use of this tool requires a valid license. To obtain a license, contact Whitebox Geospatial Inc. ([email protected]). "# .replace("*", &sep) .replace("EXE_NAME", exe_name); println!("{}", s); } fn version() { const VERSION: Option<&'static str> = option_env!("CARGO_PKG_VERSION"); println!( "correct_stream_vector_direction v{} by Dr. John B. Lindsay (c) 2023.", VERSION.unwrap_or("Unknown version") ); } fn get_tool_name() -> String { String::from("CorrectStreamVectorDirection") // This should be camel case and is a reference to the tool name. } fn run(args: &Vec<String>) -> Result<(), std::io::Error> { let tool_name = get_tool_name(); let sep: String = path::MAIN_SEPARATOR.to_string(); // Read in the environment variables and get the necessary values let configurations = whitebox_common::configs::get_configs()?; let mut working_directory = configurations.working_directory.clone(); if !working_directory.is_empty() && !working_directory.ends_with(&sep) { working_directory += &sep; } // read the arguments let mut input_file = String::new(); let mut outlet_file = String::new(); let mut output_file: String = String::new(); let mut snap_dist = 1.0; if args.len() <= 1 { return Err(Error::new( ErrorKind::InvalidInput, "Tool run with too few parameters.", )); } for i in 0..args.len() { let mut arg = args[i].replace("\"", ""); arg = arg.replace("\'", ""); let cmd = arg.split("="); // in case an equals sign was used let vec = cmd.collect::<Vec<&str>>(); let mut keyval = false; if vec.len() > 1 { keyval = true; } let flag_val = vec[0].to_lowercase().replace("--", "-"); if flag_val == "-i" || flag_val == "-input" { input_file = if keyval { vec[1].to_string() } else { args[i + 1].to_string() }; } else if flag_val == "-outlet" { outlet_file = if keyval { vec[1].to_string() } else { args[i + 1].to_string() }; } else if flag_val == "-o" || flag_val == "-output" { output_file = if keyval { vec[1].to_string() } else { args[i + 1].to_string() }; } else if flag_val == "-snap" || flag_val == "-dist" { snap_dist = if keyval { vec[1] .to_string() .parse::<f64>() .expect(&format!("Error parsing {}", flag_val)) } else { args[i + 1] .to_string() .parse::<f64>() .expect(&format!("Error parsing {}", flag_val)) }; } } if configurations.verbose_mode { let welcome_len = format!("* Welcome to {} *", tool_name).len().max(28); // 28 = length of the 'Powered by' by statement. println!("{}", "*".repeat(welcome_len)); println!("* Welcome to {} {}*", tool_name, " ".repeat(welcome_len - 15 - tool_name.len())); println!("* Powered by WhiteboxTools {}*", " ".repeat(welcome_len - 28)); println!("* www.whiteboxgeo.com {}*", " ".repeat(welcome_len - 23)); println!("{}", "*".repeat(welcome_len)); } // let mut progress: usize; let mut old_progress: usize = 1; let start = Instant::now(); if !input_file.contains(path::MAIN_SEPARATOR) && !input_file.contains("/") { input_file = format!("{}{}", working_directory, input_file); } if !outlet_file.contains(path::MAIN_SEPARATOR) && !outlet_file.contains("/") { outlet_file = format!("{}{}", working_directory, outlet_file); } if output_file.is_empty() { output_file = input_file .clone() .replace(".shp", "_corrected.shp") .replace(".SHP", "_corrected.shp"); } if !output_file.contains(path::MAIN_SEPARATOR) && !output_file.contains("/") { output_file = format!("{}{}", working_directory, output_file); } if snap_dist <= 0f64 { if configurations.verbose_mode { wrapped_print("Error: The snap distance must be greater than 0.0.", 50); } } let input = Shapefile::read(&input_file).expect("Error reading file"); //?; // Make sure the input vector file is of polyline type if input.header.shape_type.base_shape_type() != ShapeType::PolyLine { // return Err(Error::new( // ErrorKind::InvalidInput, // "The vector data must be of PolyLine base shape type.", // )); panic!("The vector stream data must be of PolyLine base shape type."); } let outlets = Shapefile::read(&outlet_file).expect("Error reading file"); //?; // Make sure the input vector file is of polyline type if outlets.header.shape_type.base_shape_type() != ShapeType::Point { panic!("The vector outlets data must be of POINT base shape type."); } let mut progress: usize; // Read each line segment into an rtree. type Location = GeomWithData<[f64; 2], (usize, bool)>; let mut end_nodes = vec![]; let (mut part_start, mut part_end): (usize, usize); let mut fid = 0usize; // fid is unique to each part in the vector let mut polylines = vec![]; for record_num in 0..input.num_records { let record = input.get_record(record_num); for part in 0..record.num_parts as usize { part_start = record.parts[part] as usize; part_end = if part < record.num_parts as usize - 1 { record.parts[part + 1] as usize - 1 } else { record.num_points as usize - 1 }; polylines.push( Polyline::new( &record.points[part_start..=part_end], record_num ) ); end_nodes.push(Location::new( [record.points[part_start].x, record.points[part_start].y], (fid, true) )); end_nodes.push(Location::new( [record.points[part_end].x, record.points[part_end].y], (fid, false) )); fid += 1; } if configurations.verbose_mode { progress = (100.0_f64 * (record_num + 1) as f64 / input.num_records as f64) as usize; if progress != old_progress { println!("Reading vector: {}
Input/output file names can be fully qualified, or can rely on the
random_line_split
main.rs
not be /// carried over to the output file because there is not a one-for-one feature correspondence between the /// two files due to the joins and splits of stream segments. Instead the output attribute table will /// only contain a feature ID (FID) entry. /// /// > Note: this tool should be used to pre-process vector streams that are input to the /// > `VectorStreamNetworkAnalysis` tool. /// /// # See Also /// `VectorStreamNetworkAnalysis`, `FixDanglingArcs` fn main() { let args: Vec<String> = env::args().collect(); if args[1].trim() == "run" { match run(&args) { Ok(_) => {} Err(e) => panic!("{:?}", e), } } if args.len() <= 1 || args[1].trim() == "help" { // print help help(); } if args[1].trim() == "version" { // print version information version(); } } fn help() { let mut ext = ""; if cfg!(target_os = "windows") { ext = ".exe"; } let exe_name = &format!("correct_stream_vector_direction{}", ext); let sep: String = path::MAIN_SEPARATOR.to_string(); let s = r#" This tool resolves topological errors and inconsistencies associated with digitized vector streams. The following commands are recognized: help Prints help information. run Runs the tool. version Prints the tool version information. The following flags can be used with the 'run' command: --routes Name of the input routes vector file. -o, --output Name of the output HTML file. --length Maximum segment length (m). --dist Search distance, in grid cells, used in visibility analysis. Input/output file names can be fully qualified, or can rely on the working directory contained in the WhiteboxTools settings.json file. Example Usage: >> .*EXE_NAME run --routes=footpath.shp --dem=DEM.tif -o=assessedRoutes.shp --length=50.0 --dist=200 Note: Use of this tool requires a valid license. To obtain a license, contact Whitebox Geospatial Inc. ([email protected]). "# .replace("*", &sep) .replace("EXE_NAME", exe_name); println!("{}", s); } fn version() { const VERSION: Option<&'static str> = option_env!("CARGO_PKG_VERSION"); println!( "correct_stream_vector_direction v{} by Dr. John B. Lindsay (c) 2023.", VERSION.unwrap_or("Unknown version") ); } fn get_tool_name() -> String { String::from("CorrectStreamVectorDirection") // This should be camel case and is a reference to the tool name. } fn run(args: &Vec<String>) -> Result<(), std::io::Error> { let tool_name = get_tool_name(); let sep: String = path::MAIN_SEPARATOR.to_string(); // Read in the environment variables and get the necessary values let configurations = whitebox_common::configs::get_configs()?; let mut working_directory = configurations.working_directory.clone(); if !working_directory.is_empty() && !working_directory.ends_with(&sep) { working_directory += &sep; } // read the arguments let mut input_file = String::new(); let mut outlet_file = String::new(); let mut output_file: String = String::new(); let mut snap_dist = 1.0; if args.len() <= 1 { return Err(Error::new( ErrorKind::InvalidInput, "Tool run with too few parameters.", )); } for i in 0..args.len() { let mut arg = args[i].replace("\"", ""); arg = arg.replace("\'", ""); let cmd = arg.split("="); // in case an equals sign was used let vec = cmd.collect::<Vec<&str>>(); let mut keyval = false; if vec.len() > 1 { keyval = true; } let flag_val = vec[0].to_lowercase().replace("--", "-"); if flag_val == "-i" || flag_val == "-input"
else if flag_val == "-outlet" { outlet_file = if keyval { vec[1].to_string() } else { args[i + 1].to_string() }; } else if flag_val == "-o" || flag_val == "-output" { output_file = if keyval { vec[1].to_string() } else { args[i + 1].to_string() }; } else if flag_val == "-snap" || flag_val == "-dist" { snap_dist = if keyval { vec[1] .to_string() .parse::<f64>() .expect(&format!("Error parsing {}", flag_val)) } else { args[i + 1] .to_string() .parse::<f64>() .expect(&format!("Error parsing {}", flag_val)) }; } } if configurations.verbose_mode { let welcome_len = format!("* Welcome to {} *", tool_name).len().max(28); // 28 = length of the 'Powered by' by statement. println!("{}", "*".repeat(welcome_len)); println!("* Welcome to {} {}*", tool_name, " ".repeat(welcome_len - 15 - tool_name.len())); println!("* Powered by WhiteboxTools {}*", " ".repeat(welcome_len - 28)); println!("* www.whiteboxgeo.com {}*", " ".repeat(welcome_len - 23)); println!("{}", "*".repeat(welcome_len)); } // let mut progress: usize; let mut old_progress: usize = 1; let start = Instant::now(); if !input_file.contains(path::MAIN_SEPARATOR) && !input_file.contains("/") { input_file = format!("{}{}", working_directory, input_file); } if !outlet_file.contains(path::MAIN_SEPARATOR) && !outlet_file.contains("/") { outlet_file = format!("{}{}", working_directory, outlet_file); } if output_file.is_empty() { output_file = input_file .clone() .replace(".shp", "_corrected.shp") .replace(".SHP", "_corrected.shp"); } if !output_file.contains(path::MAIN_SEPARATOR) && !output_file.contains("/") { output_file = format!("{}{}", working_directory, output_file); } if snap_dist <= 0f64 { if configurations.verbose_mode { wrapped_print("Error: The snap distance must be greater than 0.0.", 50); } } let input = Shapefile::read(&input_file).expect("Error reading file"); //?; // Make sure the input vector file is of polyline type if input.header.shape_type.base_shape_type() != ShapeType::PolyLine { // return Err(Error::new( // ErrorKind::InvalidInput, // "The vector data must be of PolyLine base shape type.", // )); panic!("The vector stream data must be of PolyLine base shape type."); } let outlets = Shapefile::read(&outlet_file).expect("Error reading file"); //?; // Make sure the input vector file is of polyline type if outlets.header.shape_type.base_shape_type() != ShapeType::Point { panic!("The vector outlets data must be of POINT base shape type."); } let mut progress: usize; // Read each line segment into an rtree. type Location = GeomWithData<[f64; 2], (usize, bool)>; let mut end_nodes = vec![]; let (mut part_start, mut part_end): (usize, usize); let mut fid = 0usize; // fid is unique to each part in the vector let mut polylines = vec![]; for record_num in 0..input.num_records { let record = input.get_record(record_num); for part in 0..record.num_parts as usize { part_start = record.parts[part] as usize; part_end = if part < record.num_parts as usize - 1 { record.parts[part + 1] as usize - 1 } else { record.num_points as usize - 1 }; polylines.push( Polyline::new( &record.points[part_start..=part_end], record_num ) ); end_nodes.push(Location::new( [record.points[part_start].x, record.points[part_start].y], (fid, true) )); end_nodes.push(Location::new( [record.points[part_end].x, record.points[part_end].y], (fid, false) )); fid += 1; } if configurations.verbose_mode { progress = (100.0_f64 * (record_num + 1) as f64 / input.num_records as f64) as usize; if progress != old_progress { println!("Reading vector:
{ input_file = if keyval { vec[1].to_string() } else { args[i + 1].to_string() }; }
conditional_block
main.rs
not be /// carried over to the output file because there is not a one-for-one feature correspondence between the /// two files due to the joins and splits of stream segments. Instead the output attribute table will /// only contain a feature ID (FID) entry. /// /// > Note: this tool should be used to pre-process vector streams that are input to the /// > `VectorStreamNetworkAnalysis` tool. /// /// # See Also /// `VectorStreamNetworkAnalysis`, `FixDanglingArcs` fn main() { let args: Vec<String> = env::args().collect(); if args[1].trim() == "run" { match run(&args) { Ok(_) => {} Err(e) => panic!("{:?}", e), } } if args.len() <= 1 || args[1].trim() == "help" { // print help help(); } if args[1].trim() == "version" { // print version information version(); } } fn help() { let mut ext = ""; if cfg!(target_os = "windows") { ext = ".exe"; } let exe_name = &format!("correct_stream_vector_direction{}", ext); let sep: String = path::MAIN_SEPARATOR.to_string(); let s = r#" This tool resolves topological errors and inconsistencies associated with digitized vector streams. The following commands are recognized: help Prints help information. run Runs the tool. version Prints the tool version information. The following flags can be used with the 'run' command: --routes Name of the input routes vector file. -o, --output Name of the output HTML file. --length Maximum segment length (m). --dist Search distance, in grid cells, used in visibility analysis. Input/output file names can be fully qualified, or can rely on the working directory contained in the WhiteboxTools settings.json file. Example Usage: >> .*EXE_NAME run --routes=footpath.shp --dem=DEM.tif -o=assessedRoutes.shp --length=50.0 --dist=200 Note: Use of this tool requires a valid license. To obtain a license, contact Whitebox Geospatial Inc. ([email protected]). "# .replace("*", &sep) .replace("EXE_NAME", exe_name); println!("{}", s); } fn version()
fn get_tool_name() -> String { String::from("CorrectStreamVectorDirection") // This should be camel case and is a reference to the tool name. } fn run(args: &Vec<String>) -> Result<(), std::io::Error> { let tool_name = get_tool_name(); let sep: String = path::MAIN_SEPARATOR.to_string(); // Read in the environment variables and get the necessary values let configurations = whitebox_common::configs::get_configs()?; let mut working_directory = configurations.working_directory.clone(); if !working_directory.is_empty() && !working_directory.ends_with(&sep) { working_directory += &sep; } // read the arguments let mut input_file = String::new(); let mut outlet_file = String::new(); let mut output_file: String = String::new(); let mut snap_dist = 1.0; if args.len() <= 1 { return Err(Error::new( ErrorKind::InvalidInput, "Tool run with too few parameters.", )); } for i in 0..args.len() { let mut arg = args[i].replace("\"", ""); arg = arg.replace("\'", ""); let cmd = arg.split("="); // in case an equals sign was used let vec = cmd.collect::<Vec<&str>>(); let mut keyval = false; if vec.len() > 1 { keyval = true; } let flag_val = vec[0].to_lowercase().replace("--", "-"); if flag_val == "-i" || flag_val == "-input" { input_file = if keyval { vec[1].to_string() } else { args[i + 1].to_string() }; } else if flag_val == "-outlet" { outlet_file = if keyval { vec[1].to_string() } else { args[i + 1].to_string() }; } else if flag_val == "-o" || flag_val == "-output" { output_file = if keyval { vec[1].to_string() } else { args[i + 1].to_string() }; } else if flag_val == "-snap" || flag_val == "-dist" { snap_dist = if keyval { vec[1] .to_string() .parse::<f64>() .expect(&format!("Error parsing {}", flag_val)) } else { args[i + 1] .to_string() .parse::<f64>() .expect(&format!("Error parsing {}", flag_val)) }; } } if configurations.verbose_mode { let welcome_len = format!("* Welcome to {} *", tool_name).len().max(28); // 28 = length of the 'Powered by' by statement. println!("{}", "*".repeat(welcome_len)); println!("* Welcome to {} {}*", tool_name, " ".repeat(welcome_len - 15 - tool_name.len())); println!("* Powered by WhiteboxTools {}*", " ".repeat(welcome_len - 28)); println!("* www.whiteboxgeo.com {}*", " ".repeat(welcome_len - 23)); println!("{}", "*".repeat(welcome_len)); } // let mut progress: usize; let mut old_progress: usize = 1; let start = Instant::now(); if !input_file.contains(path::MAIN_SEPARATOR) && !input_file.contains("/") { input_file = format!("{}{}", working_directory, input_file); } if !outlet_file.contains(path::MAIN_SEPARATOR) && !outlet_file.contains("/") { outlet_file = format!("{}{}", working_directory, outlet_file); } if output_file.is_empty() { output_file = input_file .clone() .replace(".shp", "_corrected.shp") .replace(".SHP", "_corrected.shp"); } if !output_file.contains(path::MAIN_SEPARATOR) && !output_file.contains("/") { output_file = format!("{}{}", working_directory, output_file); } if snap_dist <= 0f64 { if configurations.verbose_mode { wrapped_print("Error: The snap distance must be greater than 0.0.", 50); } } let input = Shapefile::read(&input_file).expect("Error reading file"); //?; // Make sure the input vector file is of polyline type if input.header.shape_type.base_shape_type() != ShapeType::PolyLine { // return Err(Error::new( // ErrorKind::InvalidInput, // "The vector data must be of PolyLine base shape type.", // )); panic!("The vector stream data must be of PolyLine base shape type."); } let outlets = Shapefile::read(&outlet_file).expect("Error reading file"); //?; // Make sure the input vector file is of polyline type if outlets.header.shape_type.base_shape_type() != ShapeType::Point { panic!("The vector outlets data must be of POINT base shape type."); } let mut progress: usize; // Read each line segment into an rtree. type Location = GeomWithData<[f64; 2], (usize, bool)>; let mut end_nodes = vec![]; let (mut part_start, mut part_end): (usize, usize); let mut fid = 0usize; // fid is unique to each part in the vector let mut polylines = vec![]; for record_num in 0..input.num_records { let record = input.get_record(record_num); for part in 0..record.num_parts as usize { part_start = record.parts[part] as usize; part_end = if part < record.num_parts as usize - 1 { record.parts[part + 1] as usize - 1 } else { record.num_points as usize - 1 }; polylines.push( Polyline::new( &record.points[part_start..=part_end], record_num ) ); end_nodes.push(Location::new( [record.points[part_start].x, record.points[part_start].y], (fid, true) )); end_nodes.push(Location::new( [record.points[part_end].x, record.points[part_end].y], (fid, false) )); fid += 1; } if configurations.verbose_mode { progress = (100.0_f64 * (record_num + 1) as f64 / input.num_records as f64) as usize; if progress != old_progress { println!("Reading vector
{ const VERSION: Option<&'static str> = option_env!("CARGO_PKG_VERSION"); println!( "correct_stream_vector_direction v{} by Dr. John B. Lindsay (c) 2023.", VERSION.unwrap_or("Unknown version") ); }
identifier_body
main.rs
mut old_progress: usize = 1; let start = Instant::now(); if !input_file.contains(path::MAIN_SEPARATOR) && !input_file.contains("/") { input_file = format!("{}{}", working_directory, input_file); } if !outlet_file.contains(path::MAIN_SEPARATOR) && !outlet_file.contains("/") { outlet_file = format!("{}{}", working_directory, outlet_file); } if output_file.is_empty() { output_file = input_file .clone() .replace(".shp", "_corrected.shp") .replace(".SHP", "_corrected.shp"); } if !output_file.contains(path::MAIN_SEPARATOR) && !output_file.contains("/") { output_file = format!("{}{}", working_directory, output_file); } if snap_dist <= 0f64 { if configurations.verbose_mode { wrapped_print("Error: The snap distance must be greater than 0.0.", 50); } } let input = Shapefile::read(&input_file).expect("Error reading file"); //?; // Make sure the input vector file is of polyline type if input.header.shape_type.base_shape_type() != ShapeType::PolyLine { // return Err(Error::new( // ErrorKind::InvalidInput, // "The vector data must be of PolyLine base shape type.", // )); panic!("The vector stream data must be of PolyLine base shape type."); } let outlets = Shapefile::read(&outlet_file).expect("Error reading file"); //?; // Make sure the input vector file is of polyline type if outlets.header.shape_type.base_shape_type() != ShapeType::Point { panic!("The vector outlets data must be of POINT base shape type."); } let mut progress: usize; // Read each line segment into an rtree. type Location = GeomWithData<[f64; 2], (usize, bool)>; let mut end_nodes = vec![]; let (mut part_start, mut part_end): (usize, usize); let mut fid = 0usize; // fid is unique to each part in the vector let mut polylines = vec![]; for record_num in 0..input.num_records { let record = input.get_record(record_num); for part in 0..record.num_parts as usize { part_start = record.parts[part] as usize; part_end = if part < record.num_parts as usize - 1 { record.parts[part + 1] as usize - 1 } else { record.num_points as usize - 1 }; polylines.push( Polyline::new( &record.points[part_start..=part_end], record_num ) ); end_nodes.push(Location::new( [record.points[part_start].x, record.points[part_start].y], (fid, true) )); end_nodes.push(Location::new( [record.points[part_end].x, record.points[part_end].y], (fid, false) )); fid += 1; } if configurations.verbose_mode { progress = (100.0_f64 * (record_num + 1) as f64 / input.num_records as f64) as usize; if progress != old_progress { println!("Reading vector: {}%", progress); old_progress = progress; } } } let num_polylines = polylines.len(); // will be updated after the joins. let snap_dist_sq = snap_dist * snap_dist; let endnode_tree = RTree::bulk_load(end_nodes); let precision = EPSILON * 10f64; let mut outlet_pt: Point2D; let mut p1: Point2D; let mut visited = vec![false; num_polylines]; let mut reverse = vec![false; num_polylines]; for record_num in 0..outlets.num_records { let record = outlets.get_record(record_num); if record.shape_type != ShapeType::Null { for p in 0..record.points.len() { outlet_pt = record.points[p]; let ret = endnode_tree.locate_within_distance([outlet_pt.x, outlet_pt.y], snap_dist_sq); for pt in ret { let (fid, is_start) = pt.data; if !visited[fid] { visited[fid] = true; // now find all connected line segments let mut queue: VecDeque<(usize, bool)> = VecDeque::with_capacity(num_polylines); queue.push_back((fid, is_start)); while !queue.is_empty() { let (fid2, is_start2) = queue.pop_front().unwrap(); // Find the point associated with the other end of this polyline p1 = if !is_start2 { polylines[fid2].get_first_node() } else { // To get here means that you first encountered the beginning of the polyline, which // shouldn't happen if it is correctly directed, since we are doing a bottom-up // scan of the network. Therefore, reverse the line in the output. reverse[fid2] = true; polylines[fid2].get_last_node() }; // Find the neighbouring endnodes of p1 let ret2 = endnode_tree.locate_within_distance([p1.x, p1.y], precision); for pt2 in ret2 { let (fid_n, is_start_n) = pt2.data; if fid_n != fid2 && !visited[fid_n] { // Add this newly encountered polyline to the queue queue.push_back((fid_n, is_start_n)); visited[fid_n] = true; } } } } } } } if configurations.verbose_mode { progress = (100.0_f64 * (fid + 1) as f64 / num_polylines as f64) as usize; if progress != old_progress { println!("Looking for reverse-oriented arcs: {}%", progress); old_progress = progress; } } } let mut num_reversed = 0; for fid in 0..polylines.len() { if reverse[fid] { let mut line = polylines[fid].vertices.clone(); line.reverse(); polylines[fid].vertices = line; num_reversed += 1; } } println!("num. reversed arcs: {num_reversed}"); // create output file let mut output = Shapefile::initialize_using_file(&output_file.replace(".shp", "_reversed_arcs.shp"), &input, ShapeType::PolyLine, true).expect("Error creating output file"); let mut sfg: ShapefileGeometry; for fid in 0..polylines.len() { if reverse[fid] { sfg = ShapefileGeometry::new(ShapeType::PolyLine); sfg.add_part(&polylines[fid].vertices); output.add_record(sfg); let record_num = polylines[fid].id; let att_data = input.attributes.get_record(record_num); output.attributes.add_record(att_data.clone(), false); } } output.write().expect("Error writing file."); // create output file let mut output = Shapefile::initialize_using_file(&output_file, &input, ShapeType::PolyLine, true).expect("Error creating output file"); //?; // add the attributes // let in_atts = input.attributes.get_fields(); let mut sfg: ShapefileGeometry; for poly_id in 0..polylines.len() { sfg = ShapefileGeometry::new(ShapeType::PolyLine); sfg.add_part(&polylines[poly_id].vertices); output.add_record(sfg); let record_num = polylines[poly_id].id; let att_data = input.attributes.get_record(record_num); output.attributes.add_record(att_data.clone(), false); if configurations.verbose_mode { progress = (100.0_f64 * (poly_id + 1) as f64 / polylines.len() as f64) as usize; if progress != old_progress { println!("Looking for dangling arcs: {}%", progress); old_progress = progress; } } } if configurations.verbose_mode { println!("Saving data...") }; output.write().expect("Error writing file."); let elapsed_time = get_formatted_elapsed_time(start); if configurations.verbose_mode { println!( "\n{}", &format!("Elapsed Time (Including I/O): {}", elapsed_time) ); } Ok(()) } #[derive(Default, Clone, Debug)] struct Polyline { vertices: Vec<Point2D>, id: usize, } impl Index<usize> for Polyline { type Output = Point2D; fn index<'a>(&'a self, index: usize) -> &'a Point2D { &self.vertices[index] } } impl Polyline { // Creates a new Polyline from vertices fn new(vertices: &[Point2D], id: usize) -> Self { Polyline { vertices: vertices.clone().to_vec(), id, } } fn
get_first_node
identifier_name
spmc.rs
Ptr::new(single_node.as_ptr()), incin, }), }; (sender, receiver) } /// The [`Sender`] handle of a SPMC channel. Created by [`create`] or /// [`with_incin`] function. pub struct
<T> { back: NonNull<Node<T>>, } impl<T> Sender<T> { /// Sends a message and if the receiver disconnected, an error is returned. pub fn send(&mut self, message: T) -> Result<(), NoRecv<T>> { // First we allocate the node for our message. let alloc = OwnedAlloc::new(Node { message: Removable::new(message), next: AtomicPtr::new(null_mut()), }); let nnptr = alloc.into_raw(); // This dereferral is safe because the queue has at least one node. We // possess a single node in the back, and if the queue has just one // node, it is stored in the back (and in the front). Also, we are the // only ones with access to the back. let res = unsafe { // We try to update the back's next pointer. We want to catch any // bit marking here. A marked lower bit means the receiver // disconnected. self.back.as_ref().next.compare_exchange( null_mut(), nnptr.as_ptr(), Release, Relaxed, ) }; if res.is_ok() { // If we succeeded, let's update the back so we keep the invariant // "the back has a single node". self.back = nnptr; Ok(()) } else { // If we failed, receiver disconnected. It is safe to dealloc // because this is the node we just allocated, and we did not share // it with anyone (cas failed). let mut alloc = unsafe { OwnedAlloc::from_raw(nnptr) }; let message = alloc.message.replace(None).unwrap(); Err(NoRecv { message }) } } /// Tests if there are any [`Receiver`]s still connected. There are no /// guarantees that [`send`](Sender::send) will succeed if this method /// returns `true` because the [`Receiver`] may disconnect meanwhile. pub fn is_connected(&self) -> bool { // Safe because we always have at least one node, which is only dropped // in the last side to disconnect's drop. let back = unsafe { self.back.as_ref() }; back.next.load(Relaxed).is_null() } } impl<T> Drop for Sender<T> { fn drop(&mut self) { // This dereferral is safe because the queue always have at least one // node. This single node is only dropped when the last side to // disconnect drops. let res = unsafe { // Let's try to mark next's bit so that receiver will see we // disconnected, if it hasn't disconnected by itself. It is ok to // just swap, since we have only two possible values (null and // null | 1) and we everyone will be setting to the same value // (null | 1). self.back .as_ref() .next .swap((null_mut::<Node<T>>() as usize | 1) as *mut _, Relaxed) }; // If the previously stored value was not null, receiver has already // disconnected. It is safe to drop because we are the only ones that // have a pointer to the node. if !res.is_null() { unsafe { OwnedAlloc::from_raw(self.back) }; } } } impl<T> fmt::Debug for Sender<T> { fn fmt(&self, fmtr: &mut fmt::Formatter) -> fmt::Result { fmtr.write_str("spmc::Sender") } } unsafe impl<T> Send for Sender<T> where T: Send {} unsafe impl<T> Sync for Sender<T> where T: Send {} /// The [`Receiver`] handle of a SPMC channel. Created by [`create`] or /// [`with_incin`] function. It is clonable and does not require mutability. pub struct Receiver<T> { inner: Arc<ReceiverInner<T>>, } impl<T> Receiver<T> { /// Tries to receive a message. If no message is available, /// [`Err`]`(`[`RecvErr::NoMessage`]`)` is returned. If the sender /// disconnected, [`Err`]`(`[`RecvErr::NoSender`]`)` is returned. #[allow(unused_must_use)] pub fn recv(&self) -> Result<T, RecvErr> { // We have to pause the incinerator due to ABA problem. This channel // suffers from it, yeah. let pause = self.inner.incin.inner.pause(); // Bypassing null check is safe because we never store null in // the front. let mut front_nnptr = unsafe { // First we load pointer stored in the front. bypass_null(self.inner.front.load(Relaxed)) }; loop { // Let's remove the node logically first. Safe to derefer this // pointer because we paused the incinerator and we only // delete nodes via incinerator. match unsafe { front_nnptr.as_ref().message.take(AcqRel) } { Some(val) => { // Safe to call because we passed a pointer from the front // which was loaded during the very same pause we are // passing. unsafe { self.try_clear_first(front_nnptr, &pause) }; break Ok(val); }, // Safe to call because we passed a pointer from the front // which was loaded during the very same pause we are passing. None => unsafe { front_nnptr = self.try_clear_first(front_nnptr, &pause)?; }, } } } /// Tests if there are any [`Sender`]s still connected. There are no /// guarantees that [`recv`](Receiver::recv) will succeed if this method /// returns `true` because the [`Receiver`] may disconnect meanwhile. /// This method may also return `true` if the [`Sender`] disconnected /// but there are messages pending in the buffer. Note that another /// [`Receiver`] may pop out the pending messages after this method was /// called. pub fn is_connected(&self) -> bool { // We need this pause because of use-after-free. let _pause = self.inner.incin.inner.pause(); // Safe to derefer this pointer because we paused the incinerator and we // only delete nodes via incinerator. let front = unsafe { &*self.inner.front.load(Relaxed) }; front.message.is_present(Relaxed) || front.next.load(Relaxed) as usize & 1 == 0 } /// The shared incinerator used by this [`Receiver`]. pub fn incin(&self) -> SharedIncin<T> { self.inner.incin.clone() } // This function is unsafe because passing the wrong pointer will lead to // undefined behavior. The pointer must have been loaded from the front // during the passed pause. unsafe fn try_clear_first( &self, expected: NonNull<Node<T>>, pause: &Pause<OwnedAlloc<Node<T>>>, ) -> Result<NonNull<Node<T>>, RecvErr> { let next = expected.as_ref().next.load(Acquire); if next as usize & 1 == 1 { // If the next is bit flagged, sender disconnected, no more messages // ever. Err(RecvErr::NoSender) } else if next.is_null() { // No bit flag means sender is still there but we have no message. Err(RecvErr::NoMessage) } else { let ptr = expected.as_ptr(); // We are not oblied to succeed. This is just cleanup and some other // thread might do it. let next = match self .inner .front .compare_exchange(ptr, next, Relaxed, Relaxed) { Ok(_) => { // Only deleting nodes via incinerator due to ABA // problem and use-after-frees. pause.add_to_incin(OwnedAlloc::from_raw(expected)); next }, Err(found) => found, }; // Safe to by-pass the check since we only store non-null // pointers on the front. Ok(bypass_null(next)) } } } impl<T> Clone for Receiver<T> { fn clone(&self) -> Self { Self { inner: self.inner.clone() } } } impl<T> fmt::Debug for Receiver<T> { fn fmt(&self, fmtr: &mut fmt::Formatter) -> fmt::Result { write!(fmtr, "spmc::Receiver {} ptr: {:p} {}", '{', self.inner, '}') } } unsafe impl<T> Send for Receiver<T> where T: Send {} unsafe impl<T> Sync for Receiver<T> where T: Send {} struct ReceiverInner<T> { // never null front: AtomicPtr<Node<T>>, incin: SharedIncin<T>, } impl<T> Drop for ReceiverInner<T> { fn drop(&mut self) { let front = self.front.get_mut(); loop { // This null-check-by-pass is safe because we never store null in // the
Sender
identifier_name
spmc.rs
Ptr::new(single_node.as_ptr()), incin, }), }; (sender, receiver) } /// The [`Sender`] handle of a SPMC channel. Created by [`create`] or /// [`with_incin`] function. pub struct Sender<T> { back: NonNull<Node<T>>, } impl<T> Sender<T> { /// Sends a message and if the receiver disconnected, an error is returned. pub fn send(&mut self, message: T) -> Result<(), NoRecv<T>> { // First we allocate the node for our message. let alloc = OwnedAlloc::new(Node { message: Removable::new(message), next: AtomicPtr::new(null_mut()), }); let nnptr = alloc.into_raw(); // This dereferral is safe because the queue has at least one node. We // possess a single node in the back, and if the queue has just one // node, it is stored in the back (and in the front). Also, we are the // only ones with access to the back. let res = unsafe { // We try to update the back's next pointer. We want to catch any // bit marking here. A marked lower bit means the receiver // disconnected. self.back.as_ref().next.compare_exchange( null_mut(), nnptr.as_ptr(), Release, Relaxed, ) }; if res.is_ok() { // If we succeeded, let's update the back so we keep the invariant // "the back has a single node". self.back = nnptr; Ok(()) } else { // If we failed, receiver disconnected. It is safe to dealloc // because this is the node we just allocated, and we did not share // it with anyone (cas failed). let mut alloc = unsafe { OwnedAlloc::from_raw(nnptr) }; let message = alloc.message.replace(None).unwrap(); Err(NoRecv { message }) } } /// Tests if there are any [`Receiver`]s still connected. There are no /// guarantees that [`send`](Sender::send) will succeed if this method /// returns `true` because the [`Receiver`] may disconnect meanwhile. pub fn is_connected(&self) -> bool
} impl<T> Drop for Sender<T> { fn drop(&mut self) { // This dereferral is safe because the queue always have at least one // node. This single node is only dropped when the last side to // disconnect drops. let res = unsafe { // Let's try to mark next's bit so that receiver will see we // disconnected, if it hasn't disconnected by itself. It is ok to // just swap, since we have only two possible values (null and // null | 1) and we everyone will be setting to the same value // (null | 1). self.back .as_ref() .next .swap((null_mut::<Node<T>>() as usize | 1) as *mut _, Relaxed) }; // If the previously stored value was not null, receiver has already // disconnected. It is safe to drop because we are the only ones that // have a pointer to the node. if !res.is_null() { unsafe { OwnedAlloc::from_raw(self.back) }; } } } impl<T> fmt::Debug for Sender<T> { fn fmt(&self, fmtr: &mut fmt::Formatter) -> fmt::Result { fmtr.write_str("spmc::Sender") } } unsafe impl<T> Send for Sender<T> where T: Send {} unsafe impl<T> Sync for Sender<T> where T: Send {} /// The [`Receiver`] handle of a SPMC channel. Created by [`create`] or /// [`with_incin`] function. It is clonable and does not require mutability. pub struct Receiver<T> { inner: Arc<ReceiverInner<T>>, } impl<T> Receiver<T> { /// Tries to receive a message. If no message is available, /// [`Err`]`(`[`RecvErr::NoMessage`]`)` is returned. If the sender /// disconnected, [`Err`]`(`[`RecvErr::NoSender`]`)` is returned. #[allow(unused_must_use)] pub fn recv(&self) -> Result<T, RecvErr> { // We have to pause the incinerator due to ABA problem. This channel // suffers from it, yeah. let pause = self.inner.incin.inner.pause(); // Bypassing null check is safe because we never store null in // the front. let mut front_nnptr = unsafe { // First we load pointer stored in the front. bypass_null(self.inner.front.load(Relaxed)) }; loop { // Let's remove the node logically first. Safe to derefer this // pointer because we paused the incinerator and we only // delete nodes via incinerator. match unsafe { front_nnptr.as_ref().message.take(AcqRel) } { Some(val) => { // Safe to call because we passed a pointer from the front // which was loaded during the very same pause we are // passing. unsafe { self.try_clear_first(front_nnptr, &pause) }; break Ok(val); }, // Safe to call because we passed a pointer from the front // which was loaded during the very same pause we are passing. None => unsafe { front_nnptr = self.try_clear_first(front_nnptr, &pause)?; }, } } } /// Tests if there are any [`Sender`]s still connected. There are no /// guarantees that [`recv`](Receiver::recv) will succeed if this method /// returns `true` because the [`Receiver`] may disconnect meanwhile. /// This method may also return `true` if the [`Sender`] disconnected /// but there are messages pending in the buffer. Note that another /// [`Receiver`] may pop out the pending messages after this method was /// called. pub fn is_connected(&self) -> bool { // We need this pause because of use-after-free. let _pause = self.inner.incin.inner.pause(); // Safe to derefer this pointer because we paused the incinerator and we // only delete nodes via incinerator. let front = unsafe { &*self.inner.front.load(Relaxed) }; front.message.is_present(Relaxed) || front.next.load(Relaxed) as usize & 1 == 0 } /// The shared incinerator used by this [`Receiver`]. pub fn incin(&self) -> SharedIncin<T> { self.inner.incin.clone() } // This function is unsafe because passing the wrong pointer will lead to // undefined behavior. The pointer must have been loaded from the front // during the passed pause. unsafe fn try_clear_first( &self, expected: NonNull<Node<T>>, pause: &Pause<OwnedAlloc<Node<T>>>, ) -> Result<NonNull<Node<T>>, RecvErr> { let next = expected.as_ref().next.load(Acquire); if next as usize & 1 == 1 { // If the next is bit flagged, sender disconnected, no more messages // ever. Err(RecvErr::NoSender) } else if next.is_null() { // No bit flag means sender is still there but we have no message. Err(RecvErr::NoMessage) } else { let ptr = expected.as_ptr(); // We are not oblied to succeed. This is just cleanup and some other // thread might do it. let next = match self .inner .front .compare_exchange(ptr, next, Relaxed, Relaxed) { Ok(_) => { // Only deleting nodes via incinerator due to ABA // problem and use-after-frees. pause.add_to_incin(OwnedAlloc::from_raw(expected)); next }, Err(found) => found, }; // Safe to by-pass the check since we only store non-null // pointers on the front. Ok(bypass_null(next)) } } } impl<T> Clone for Receiver<T> { fn clone(&self) -> Self { Self { inner: self.inner.clone() } } } impl<T> fmt::Debug for Receiver<T> { fn fmt(&self, fmtr: &mut fmt::Formatter) -> fmt::Result { write!(fmtr, "spmc::Receiver {} ptr: {:p} {}", '{', self.inner, '}') } } unsafe impl<T> Send for Receiver<T> where T: Send {} unsafe impl<T> Sync for Receiver<T> where T: Send {} struct ReceiverInner<T> { // never null front: AtomicPtr<Node<T>>, incin: SharedIncin<T>, } impl<T> Drop for ReceiverInner<T> { fn drop(&mut self) { let front = self.front.get_mut(); loop { // This null-check-by-pass is safe because we never store null in //
{ // Safe because we always have at least one node, which is only dropped // in the last side to disconnect's drop. let back = unsafe { self.back.as_ref() }; back.next.load(Relaxed).is_null() }
identifier_body
spmc.rs
Ptr::new(single_node.as_ptr()), incin, }), }; (sender, receiver) } /// The [`Sender`] handle of a SPMC channel. Created by [`create`] or /// [`with_incin`] function. pub struct Sender<T> { back: NonNull<Node<T>>, } impl<T> Sender<T> { /// Sends a message and if the receiver disconnected, an error is returned. pub fn send(&mut self, message: T) -> Result<(), NoRecv<T>> { // First we allocate the node for our message. let alloc = OwnedAlloc::new(Node { message: Removable::new(message), next: AtomicPtr::new(null_mut()), }); let nnptr = alloc.into_raw(); // This dereferral is safe because the queue has at least one node. We // possess a single node in the back, and if the queue has just one // node, it is stored in the back (and in the front). Also, we are the // only ones with access to the back. let res = unsafe { // We try to update the back's next pointer. We want to catch any // bit marking here. A marked lower bit means the receiver // disconnected. self.back.as_ref().next.compare_exchange( null_mut(), nnptr.as_ptr(), Release, Relaxed, ) }; if res.is_ok() { // If we succeeded, let's update the back so we keep the invariant // "the back has a single node". self.back = nnptr; Ok(()) } else { // If we failed, receiver disconnected. It is safe to dealloc // because this is the node we just allocated, and we did not share // it with anyone (cas failed). let mut alloc = unsafe { OwnedAlloc::from_raw(nnptr) }; let message = alloc.message.replace(None).unwrap(); Err(NoRecv { message }) } } /// Tests if there are any [`Receiver`]s still connected. There are no /// guarantees that [`send`](Sender::send) will succeed if this method /// returns `true` because the [`Receiver`] may disconnect meanwhile. pub fn is_connected(&self) -> bool { // Safe because we always have at least one node, which is only dropped // in the last side to disconnect's drop. let back = unsafe { self.back.as_ref() }; back.next.load(Relaxed).is_null() } } impl<T> Drop for Sender<T> { fn drop(&mut self) { // This dereferral is safe because the queue always have at least one // node. This single node is only dropped when the last side to // disconnect drops. let res = unsafe { // Let's try to mark next's bit so that receiver will see we // disconnected, if it hasn't disconnected by itself. It is ok to // just swap, since we have only two possible values (null and // null | 1) and we everyone will be setting to the same value // (null | 1). self.back .as_ref() .next .swap((null_mut::<Node<T>>() as usize | 1) as *mut _, Relaxed) }; // If the previously stored value was not null, receiver has already // disconnected. It is safe to drop because we are the only ones that // have a pointer to the node. if !res.is_null() { unsafe { OwnedAlloc::from_raw(self.back) }; } } } impl<T> fmt::Debug for Sender<T> { fn fmt(&self, fmtr: &mut fmt::Formatter) -> fmt::Result { fmtr.write_str("spmc::Sender") } } unsafe impl<T> Send for Sender<T> where T: Send {} unsafe impl<T> Sync for Sender<T> where T: Send {} /// The [`Receiver`] handle of a SPMC channel. Created by [`create`] or /// [`with_incin`] function. It is clonable and does not require mutability. pub struct Receiver<T> { inner: Arc<ReceiverInner<T>>, } impl<T> Receiver<T> { /// Tries to receive a message. If no message is available, /// [`Err`]`(`[`RecvErr::NoMessage`]`)` is returned. If the sender /// disconnected, [`Err`]`(`[`RecvErr::NoSender`]`)` is returned. #[allow(unused_must_use)] pub fn recv(&self) -> Result<T, RecvErr> { // We have to pause the incinerator due to ABA problem. This channel // suffers from it, yeah. let pause = self.inner.incin.inner.pause(); // Bypassing null check is safe because we never store null in // the front. let mut front_nnptr = unsafe { // First we load pointer stored in the front. bypass_null(self.inner.front.load(Relaxed)) }; loop { // Let's remove the node logically first. Safe to derefer this // pointer because we paused the incinerator and we only // delete nodes via incinerator. match unsafe { front_nnptr.as_ref().message.take(AcqRel) } { Some(val) => { // Safe to call because we passed a pointer from the front // which was loaded during the very same pause we are // passing. unsafe { self.try_clear_first(front_nnptr, &pause) }; break Ok(val); }, // Safe to call because we passed a pointer from the front // which was loaded during the very same pause we are passing. None => unsafe { front_nnptr = self.try_clear_first(front_nnptr, &pause)?; }, } } } /// Tests if there are any [`Sender`]s still connected. There are no /// guarantees that [`recv`](Receiver::recv) will succeed if this method /// returns `true` because the [`Receiver`] may disconnect meanwhile. /// This method may also return `true` if the [`Sender`] disconnected /// but there are messages pending in the buffer. Note that another /// [`Receiver`] may pop out the pending messages after this method was /// called. pub fn is_connected(&self) -> bool { // We need this pause because of use-after-free. let _pause = self.inner.incin.inner.pause(); // Safe to derefer this pointer because we paused the incinerator and we // only delete nodes via incinerator. let front = unsafe { &*self.inner.front.load(Relaxed) }; front.message.is_present(Relaxed) || front.next.load(Relaxed) as usize & 1 == 0 } /// The shared incinerator used by this [`Receiver`]. pub fn incin(&self) -> SharedIncin<T> { self.inner.incin.clone() } // This function is unsafe because passing the wrong pointer will lead to // undefined behavior. The pointer must have been loaded from the front // during the passed pause. unsafe fn try_clear_first( &self, expected: NonNull<Node<T>>, pause: &Pause<OwnedAlloc<Node<T>>>, ) -> Result<NonNull<Node<T>>, RecvErr> { let next = expected.as_ref().next.load(Acquire); if next as usize & 1 == 1 { // If the next is bit flagged, sender disconnected, no more messages // ever. Err(RecvErr::NoSender) } else if next.is_null() { // No bit flag means sender is still there but we have no message. Err(RecvErr::NoMessage) } else
// Safe to by-pass the check since we only store non-null // pointers on the front. Ok(bypass_null(next)) } } } impl<T> Clone for Receiver<T> { fn clone(&self) -> Self { Self { inner: self.inner.clone() } } } impl<T> fmt::Debug for Receiver<T> { fn fmt(&self, fmtr: &mut fmt::Formatter) -> fmt::Result { write!(fmtr, "spmc::Receiver {} ptr: {:p} {}", '{', self.inner, '}') } } unsafe impl<T> Send for Receiver<T> where T: Send {} unsafe impl<T> Sync for Receiver<T> where T: Send {} struct ReceiverInner<T> { // never null front: AtomicPtr<Node<T>>, incin: SharedIncin<T>, } impl<T> Drop for ReceiverInner<T> { fn drop(&mut self) { let front = self.front.get_mut(); loop { // This null-check-by-pass is safe because we never store null in // the
{ let ptr = expected.as_ptr(); // We are not oblied to succeed. This is just cleanup and some other // thread might do it. let next = match self .inner .front .compare_exchange(ptr, next, Relaxed, Relaxed) { Ok(_) => { // Only deleting nodes via incinerator due to ABA // problem and use-after-frees. pause.add_to_incin(OwnedAlloc::from_raw(expected)); next }, Err(found) => found, };
conditional_block
spmc.rs
AtomicPtr::new(single_node.as_ptr()), incin, }), }; (sender, receiver) } /// The [`Sender`] handle of a SPMC channel. Created by [`create`] or /// [`with_incin`] function. pub struct Sender<T> { back: NonNull<Node<T>>, } impl<T> Sender<T> { /// Sends a message and if the receiver disconnected, an error is returned. pub fn send(&mut self, message: T) -> Result<(), NoRecv<T>> { // First we allocate the node for our message. let alloc = OwnedAlloc::new(Node { message: Removable::new(message), next: AtomicPtr::new(null_mut()), }); let nnptr = alloc.into_raw(); // This dereferral is safe because the queue has at least one node. We // possess a single node in the back, and if the queue has just one // node, it is stored in the back (and in the front). Also, we are the // only ones with access to the back. let res = unsafe { // We try to update the back's next pointer. We want to catch any // bit marking here. A marked lower bit means the receiver // disconnected. self.back.as_ref().next.compare_exchange( null_mut(), nnptr.as_ptr(), Release, Relaxed, ) }; if res.is_ok() { // If we succeeded, let's update the back so we keep the invariant // "the back has a single node". self.back = nnptr; Ok(()) } else { // If we failed, receiver disconnected. It is safe to dealloc // because this is the node we just allocated, and we did not share // it with anyone (cas failed). let mut alloc = unsafe { OwnedAlloc::from_raw(nnptr) }; let message = alloc.message.replace(None).unwrap(); Err(NoRecv { message }) } } /// Tests if there are any [`Receiver`]s still connected. There are no /// guarantees that [`send`](Sender::send) will succeed if this method /// returns `true` because the [`Receiver`] may disconnect meanwhile. pub fn is_connected(&self) -> bool { // Safe because we always have at least one node, which is only dropped // in the last side to disconnect's drop. let back = unsafe { self.back.as_ref() }; back.next.load(Relaxed).is_null() } } impl<T> Drop for Sender<T> { fn drop(&mut self) { // This dereferral is safe because the queue always have at least one // node. This single node is only dropped when the last side to // disconnect drops. let res = unsafe { // Let's try to mark next's bit so that receiver will see we // disconnected, if it hasn't disconnected by itself. It is ok to // just swap, since we have only two possible values (null and // null | 1) and we everyone will be setting to the same value // (null | 1). self.back .as_ref() .next .swap((null_mut::<Node<T>>() as usize | 1) as *mut _, Relaxed) }; // If the previously stored value was not null, receiver has already // disconnected. It is safe to drop because we are the only ones that // have a pointer to the node. if !res.is_null() { unsafe { OwnedAlloc::from_raw(self.back) }; } } } impl<T> fmt::Debug for Sender<T> { fn fmt(&self, fmtr: &mut fmt::Formatter) -> fmt::Result { fmtr.write_str("spmc::Sender") } } unsafe impl<T> Send for Sender<T> where T: Send {} unsafe impl<T> Sync for Sender<T> where T: Send {} /// The [`Receiver`] handle of a SPMC channel. Created by [`create`] or /// [`with_incin`] function. It is clonable and does not require mutability. pub struct Receiver<T> { inner: Arc<ReceiverInner<T>>, } impl<T> Receiver<T> { /// Tries to receive a message. If no message is available, /// [`Err`]`(`[`RecvErr::NoMessage`]`)` is returned. If the sender /// disconnected, [`Err`]`(`[`RecvErr::NoSender`]`)` is returned. #[allow(unused_must_use)] pub fn recv(&self) -> Result<T, RecvErr> { // We have to pause the incinerator due to ABA problem. This channel
// the front. let mut front_nnptr = unsafe { // First we load pointer stored in the front. bypass_null(self.inner.front.load(Relaxed)) }; loop { // Let's remove the node logically first. Safe to derefer this // pointer because we paused the incinerator and we only // delete nodes via incinerator. match unsafe { front_nnptr.as_ref().message.take(AcqRel) } { Some(val) => { // Safe to call because we passed a pointer from the front // which was loaded during the very same pause we are // passing. unsafe { self.try_clear_first(front_nnptr, &pause) }; break Ok(val); }, // Safe to call because we passed a pointer from the front // which was loaded during the very same pause we are passing. None => unsafe { front_nnptr = self.try_clear_first(front_nnptr, &pause)?; }, } } } /// Tests if there are any [`Sender`]s still connected. There are no /// guarantees that [`recv`](Receiver::recv) will succeed if this method /// returns `true` because the [`Receiver`] may disconnect meanwhile. /// This method may also return `true` if the [`Sender`] disconnected /// but there are messages pending in the buffer. Note that another /// [`Receiver`] may pop out the pending messages after this method was /// called. pub fn is_connected(&self) -> bool { // We need this pause because of use-after-free. let _pause = self.inner.incin.inner.pause(); // Safe to derefer this pointer because we paused the incinerator and we // only delete nodes via incinerator. let front = unsafe { &*self.inner.front.load(Relaxed) }; front.message.is_present(Relaxed) || front.next.load(Relaxed) as usize & 1 == 0 } /// The shared incinerator used by this [`Receiver`]. pub fn incin(&self) -> SharedIncin<T> { self.inner.incin.clone() } // This function is unsafe because passing the wrong pointer will lead to // undefined behavior. The pointer must have been loaded from the front // during the passed pause. unsafe fn try_clear_first( &self, expected: NonNull<Node<T>>, pause: &Pause<OwnedAlloc<Node<T>>>, ) -> Result<NonNull<Node<T>>, RecvErr> { let next = expected.as_ref().next.load(Acquire); if next as usize & 1 == 1 { // If the next is bit flagged, sender disconnected, no more messages // ever. Err(RecvErr::NoSender) } else if next.is_null() { // No bit flag means sender is still there but we have no message. Err(RecvErr::NoMessage) } else { let ptr = expected.as_ptr(); // We are not oblied to succeed. This is just cleanup and some other // thread might do it. let next = match self .inner .front .compare_exchange(ptr, next, Relaxed, Relaxed) { Ok(_) => { // Only deleting nodes via incinerator due to ABA // problem and use-after-frees. pause.add_to_incin(OwnedAlloc::from_raw(expected)); next }, Err(found) => found, }; // Safe to by-pass the check since we only store non-null // pointers on the front. Ok(bypass_null(next)) } } } impl<T> Clone for Receiver<T> { fn clone(&self) -> Self { Self { inner: self.inner.clone() } } } impl<T> fmt::Debug for Receiver<T> { fn fmt(&self, fmtr: &mut fmt::Formatter) -> fmt::Result { write!(fmtr, "spmc::Receiver {} ptr: {:p} {}", '{', self.inner, '}') } } unsafe impl<T> Send for Receiver<T> where T: Send {} unsafe impl<T> Sync for Receiver<T> where T: Send {} struct ReceiverInner<T> { // never null front: AtomicPtr<Node<T>>, incin: SharedIncin<T>, } impl<T> Drop for ReceiverInner<T> { fn drop(&mut self) { let front = self.front.get_mut(); loop { // This null-check-by-pass is safe because we never store null in // the front.
// suffers from it, yeah. let pause = self.inner.incin.inner.pause(); // Bypassing null check is safe because we never store null in
random_line_split
threeHandle.js
this.width = 500 this.height = 500 this.scene = null this.light = null this.camera = null this.controls = null this.renderer = null this.fov = 60 this.mixer = null this.Stats = null this.manager = null this.crossOrigin = 'anonymous' this.requestHeader = {} this.near = 1 this.far = 2000 } async init( { container, width, height } ) { this.scene = await this.setScene() this.light = await this.setLgiht() await this.scene.add( this.light ) this.width = width || 500 this.height = height || 500 this.camera = await this.setCamera() this.renderer = await this.setRenderer() this.renderer.setClearColor( 'rgb(92,92,92)', 1.0 ) this.controls = await this.setControls() container.appendChild( this.renderer.domElement ) await this.setStats( container ) await this.setClock() window.addEventListener( 'resize', this.onWindowResize, false ) return { scene : this.scene, light : this.light, camera : this.camera, renderer : this.renderer, controls : this.controls, stats : this.stats, clock : this.clock } } setScene() { const scene = new THREE.Scene() scene.add( new THREE.AmbientLight( 0x999999 ) ) return scene } setLgiht() { const light = new THREE.DirectionalLight( 0xdfebff, 0.45 ) light.position.set( 50, 200, 100 ) light.position.multiplyScalar( 0.3 ) return light } setCamera() { const camera = new THREE.PerspectiveCamera( this.fov, this.width / this.height, this.near, this.far ) camera.position.set( 10, 90, 65 ) camera.up.set( 0, 1, 0 ) camera.lookAt( this.scene.position ) return camera } setRenderer() { const renderer = new THREE.WebGLRenderer( { alpha : true } ) // 为了兼容高清屏幕 renderer.setPixelRatio( window.devicePixelRatio ) renderer.setSize( this.width, this.height ) renderer.shadowMap.enabled = true return renderer } setControls() { const controls = new OrbitControls( this.camera, this.renderer.domElement ) controls.target.set( 0, 0, 0 ) controls.minDistance = 20 controls.maxDistance = 100000 controls.maxPolarAngle = Math.PI / 3 controls.update() controls.enableDamping = true controls.dampingFactor = 0.25 controls.enableZoom = true controls.autoRotate = false controls.enablePan = true return controls } setStats( container ) { this.stats = new Stats() this.stats.domElement.style.position = 'absolute' this.stats.domElement.style.left = '5px' this.stats.domElement.style.top = '5px' this.stats.domElement.style.display = 'none' container.appendChild( this.stats.dom ) } setClock() { this.clock = new THREE.Clock() } onWindowResize() { this.camera.aspect = this.width / this.height this.camera.updateProjectionMatrix() this.renderer.setSize( this.width, this.height ) } render() { this.renderer.render( this.scene, this.camera ) } fbxAnimate() { const delta = this.clock.getDelta() if ( this.mixer ) this.mixer.update( delta ) this.stats.update() } // 线条绘制立方体 // v6----- v5 // /| /| // v1------v0| // | | | | // | |v7---| |v4 // | / | / // v2------v3 // v0 - v1 - v2 - v3 - v0 - v5 - v4 - v4 - v0 - v5 - v6 - v1 - v2 - v7 - v4 - v5 - v6 - v7 async drawCubeByLines( { width, height, depth } ) { const objects = [] const geometryBox = await this.box( width, height, depth ) const lineSegments = await new THREE.LineSegments( geometryBox, await new THREE.LineDashedMaterial( { color : 0xffaa00, dashSize : 3, gapSize : 1 } ) ) await lineSegments.computeLineDistances() await objects.push( lineSegments ) await this.scene.add( lineSegments ) return lineSegments } async box( width, height, depth ) { const geometry = await new THREE.BufferGeometry() const position = [ -width, -height, -depth, -width, height, -depth, -width, height, -depth, width, height, -depth, width, height, -depth, width, -height, -depth, width, -height, -depth, -width, -height, -depth, -width, -height, depth, -width, height, depth, -width, height, depth, width, height, depth, width, height, depth, width, -height, depth, width, -height, depth, -width, -height, depth, -width, -height, -depth, -width, -height, depth, -width, height, -depth, -width, height, depth, width, height, -depth, width, height, depth, width, -height, -depth, width, -height, depth ] geometry.setAttribute( 'position', await new THREE.Float32BufferAttribute( position, 3 ) ) return geometry } updateCamera( { x, y, z } ) { this.camera.position.set( x, y, z ) this.camera.lookAt( this.scene.position ) } setGridHelper( size = 224, divisions = 50, color1 = 0x303030, color2 = 0x303030 ) { const gridHelper = new THREE.GridHelper( size, divisions, color1, color2 ) return gridHelper } addAnimateFbx( object ) { this.mixer = new THREE.AnimationMixer( object ) const len = object.animations.length // 动画的数量 if ( len > 0 ) { const action = this.mixer.clipAction( object.animations[0] ) action.play() } return this.mixer } loadFbx( baseUrl, fn ) { const loader = new FBXLoader( this.manager ) loader.setCrossOrigin( this.crossOrigin ) loader.setRequestHeader( this.requestHeader ) const that = this loader.load( baseUrl, ( object ) => { fn && fn( object ) }, that.onProgress, that.onError ) } // fbx模型加载贴图 loadImage( url ) { const loader = new THREE.TextureLoader() const texturePlante = loader.load( url ) const material = new THREE.MeshPhongMaterial( { map : texturePlante } ) return material } loadObj( baseUrl, objUrl, materials, fn ) { const loader = new OBJLoader( this.manager ) loader.setRequestHeader( this.requestHeader ) const that = this if ( materials ) { loader .setMaterials( materials ) .setPath( baseUrl ) .load( objUrl, ( object ) => { fn && fn( object ) }, that.onProgress, that.onError ) } else { loader.setPath( baseUrl ).load( objUrl, ( object ) => { fn && fn( object ) }, that.onProgress, that.onError ) } } loadMtl( baseUrl, mtlUrl, fn ) { const loader = new MTLLoader( this.manager ) loader.setCrossOrigin( this.crossOrigin ) loader.setRequestHeader( this.requestHeader ) const that = this loader.setPath( baseUrl ).load( mtlUrl, async( materials ) => { materials.preload() fn && fn( materials ) }, that.onProgress, that.onError ) } loadJSON( baseUrl, fn ) { const that = this const loader = new ObjectLoader() loader
import { OrbitControls } from 'three/examples/jsm/controls/OrbitControls' class ThreeHandle { constructor() {
random_line_split
threeHandle.js
立方体 // v6----- v5 // /| /| // v1------v0| // | | | | // | |v7---| |v4 // | / | / // v2------v3 // v0 - v1 - v2 - v3 - v0 - v5 - v4 - v4 - v0 - v5 - v6 - v1 - v2 - v7 - v4 - v5 - v6 - v7 async drawCubeByLines( { width, height, depth } ) { const objects = [] const geometryBox = await this.box( width, height, depth ) const lineSegments = await new THREE.LineSegments( geometryBox, await new THREE.LineDashedMaterial( { color : 0xffaa00, dashSize : 3, gapSize : 1 } ) ) await lineSegments.computeLineDistances() await objects.push( lineSegments ) await this.scene.add( lineSegments ) return lineSegments } async box( width, height, depth ) { const geometry = await new THREE.BufferGeometry() const position = [ -width, -height, -depth, -width, height, -depth, -width, height, -depth, width, height, -depth, width, height, -depth, width, -height, -depth, width, -height, -depth, -width, -height, -depth, -width, -height, depth, -width, height, depth, -width, height, depth, width, height, depth, width, height, depth, width, -height, depth, width, -height, depth, -width, -height, depth, -width, -height, -depth, -width, -height, depth, -width, height, -depth, -width, height, depth, width, height, -depth, width, height, depth, width, -height, -depth, width, -height, depth ] geometry.setAttribute( 'position', await new THREE.Float32BufferAttribute( position, 3 ) ) return geometry } updateCamera( { x, y, z } ) { this.camera.position.set( x, y, z ) this.camera.lookAt( this.scene.position ) } setGridHelper( size = 224, divisions = 50, color1 = 0x303030, color2 = 0x303030 ) { const gridHelper = new THREE.GridHelper( size, divisions, color1, color2 ) return gridHelper } addAnimateFbx( object ) { this.mixer = new THREE.AnimationMixer( object ) const len = object.animations.length // 动画的数量 if ( len > 0 ) { const action = this.mixer.clipAction( object.animations[0] ) action.play() } return this.mixer } loadFbx( baseUrl, fn ) { const loader = new FBXLoader( this.manager ) loader.setCrossOrigin( this.crossOrigin ) loader.setRequestHeader( this.requestHeader ) const that = this loader.load( baseUrl, ( object ) => { fn && fn( object ) }, that.onProgress, that.onError ) } // fbx模型加载贴图 loadImage( url ) { const loader = new THREE.TextureLoader() const texturePlante = loader.load( url ) const material = new THREE.MeshPhongMaterial( { map : texturePlante } ) return material } loadObj( baseUrl, objUrl, materials, fn ) { const loader = new OBJLoader( this.manager ) loader.setRequestHeader( this.requestHeader ) const that = this if ( materials ) { loader .setMaterials( materials ) .setPath( baseUrl ) .load( objUrl, ( object ) => { fn && fn( object ) }, that.onProgress, that.onError ) } else { loader.setPath( baseUrl ).load( objUrl, ( object ) => { fn && fn( object ) }, that.onProgress, that.onError ) } } loadMtl( baseUrl, mtlUrl, fn ) { const loader = new MTLLoader( this.manager ) loader.setCrossOrigin( this.crossOrigin ) loader.setRequestHeader( this.requestHeader ) const that = this loader.setPath( baseUrl ).load( mtlUrl, async( materials ) => { materials.preload() fn && fn( materials ) }, that.onProgress, that.onError ) } loadJSON( baseUrl, fn ) { const that = this const loader = new ObjectLoader() loader.setCrossOrigin( that.crossOrigin ) loader.setRequestHeader( that.requestHeader ) loader.load( baseUrl, ( object ) => { fn && fn( object ) }, that.onProgress, that.onError ) } loadStl( baseUrl, fn ) { const that = this const loader = new STLLoader() loader.load( baseUrl, ( geometry ) => { const material = new THREE.MeshPhongMaterial( { color : 0xff5533, specular : 0x111111, shininess : 200 } ) const object = new THREE.Mesh( geometry, material ) fn && fn( object ) }, that.onProgress, that.onError ) } loadPly( baseUrl, fn ) { const that = this const loader = new PLYLoader() loader.load( baseUrl, ( geometry ) => { // console.log( 'loadPly', geometry ) geometry.computeVertexNormals() const loadGeometry = geometry.clone() var material = new THREE.MeshStandardMaterial( { color : 0x7fff00, flatShading : true } ) var mesh = new THREE.Mesh( loadGeometry, material ) mesh.translateZ( 5 ) mesh.translateY( 2 ) // scene.add(mesh); // eslint-disable-next-line no-unused-vars const mateial = new THREE.PointsMaterial( { color : 0xffffff, size : 0.05, opacity : 0.5, transparent : true, blending : THREE.AdditiveBlending // map: generateSprite() } ) fn && fn( mesh ) }, that.onProgress, that.onError ) } // https://threejs.org/docs/index.html?q=LoadingManager#api/en/loaders/managers/LoadingManager loadManager( type = 'dds' ) { let reg let loader // eslint-disable-next-line default-case switch ( type ) { case 'dds': reg = /\.dds$/i loader = new DDSLoader() break case 'tga': reg = /\.tga$/i loader = new TGALoader() break } this.manager = new LoadingManager() this.manager.addHandler( reg, loader ) } async distoryEvent() { window.removeEventListener( 'resize', this.onWindowResize, false ) } async clearScene() { const that = this const groups = this.scene.children.filter( ( item ) => item.type == 'Group' ) const LineSegments = this.scene.children.filter( ( item ) => item.type == 'LineSegments' ) that.getGroup( groups ) that.getGroup( LineSegments ) } getGroup( groups ) { const that = this if ( groups.length > 0 ) { for ( let i = 0; i < groups.length; i++ ) { const currObj = groups[i] if ( currObj instanceof THREE.Scene ) { const children = currObj.children for ( let j = 0; j < children.length; j++ ) { that.deleteGroup( children[j] ) } } else { that.deleteGroup( currObj ) } that.scene.remove( currObj ) } } } deleteGroup( group ) { if ( !group ) return group.traverse( function( item ) { if ( item instanceof THREE.Mesh ) { item.geometry.dispose() item.material.dispose() } } ) } onProgress( xhr ) { console.log( ( xhr.loaded / xhr.total ) * 100 + '% loaded' ) if ( xhr.lengthComputable ) { // const percentComplete = xhr.loaded / xhr.total * 100 // console.log( Math.round( percentComplete, 2 ) + '% downloaded' ) } } onError() { console.error( '加载模型出错,请检查你的模型文件是否合法
' ) }
identifier_name
threeHandle.js
.camera = await this.setCamera() this.renderer = await this.setRenderer() this.renderer.setClearColor( 'rgb(92,92,92)', 1.0 ) this.controls = await this.setControls() container.appendChild( this.renderer.domElement ) await this.setStats( container ) await this.setClock() window.addEventListener( 'resize', this.onWindowResize, false ) return { scene : this.scene, light : this.light, camera : this.camera, renderer : this.renderer, controls : this.controls, stats : this.stats, clock : this.clock } } setScene() { const scene = new THREE.Scene() scene.add( new THREE.AmbientLight( 0x999999 ) ) return scene } setLgiht() { const light = new THREE.DirectionalLight( 0xdfebff, 0.45 ) light.position.set( 50, 200, 100 ) light.position.multiplyScalar( 0.3 ) return light } setCamera() { const camera = new THREE.PerspectiveCamera( this.fov, this.width / this.height, this.near, this.far ) camera.position.set( 10, 90, 65 ) camera.up.set( 0, 1, 0 ) camera.lookAt( this.scene.position ) return camera } setRenderer() { const renderer = new THREE.WebGLRenderer( { alpha : true } ) // 为了兼容高清屏幕 renderer.setPixelRatio( window.devicePixelRatio ) renderer.setSize( this.width, this.height ) renderer.shadowMap.enabled = true return renderer } setControls() { const controls = new OrbitControls( this.camera, this.renderer.domElement ) controls.target.set( 0, 0, 0 ) controls.minDistance = 20 controls.maxDistance = 100000 controls.maxPolarAngle = Math.PI / 3 controls.update() controls.enableDamping = true controls.dampingFactor = 0.25 controls.enableZoom = true controls.autoRotate = false controls.enablePan = true return controls } setStats( container ) { this.stats = new Stats() this.stats.domElement.style.position = 'absolute' this.stats.domElement.style.left = '5px' this.stats.domElement.style.top = '5px' this.stats.domElement.style.display = 'none' container.appendChild( this.stats.dom ) } setClock() { this.clock = new THREE.Clock() } onWindowResize() { this.camera.aspect = this.width / this.height this.camera.updateProjectionMatrix() this.renderer.setSize( this.width, this.height ) } render() { this.renderer.render( this.scene, this.camera ) } fbxAnimate() { const delta = this.clock.getDelta() if ( this.mixer ) this.mixer.update( delta ) this.stats.update() } // 线条绘制立方体 // v6----- v5 // /| /| // v1------v0| // | | | | // | |v7---| |v4 // | / | / // v2------v3 // v0 - v1 - v2 - v3 - v0 - v5 - v4 - v4 - v0 - v5 - v6 - v1 - v2 - v7 - v4 - v5 - v6 - v7 async drawCubeByLines( { width, height, depth } ) { const objects = [] const geometryBox = await this.box( width, height, depth ) const lineSegments = await new THREE.LineSegments( geometryBox, await new THREE.LineDashedMaterial( { color : 0xffaa00, dashSize : 3, gapSize : 1 } ) ) await lineSegments.computeLineDistances() await objects.push( lineSegments ) await this.scene.add( lineSegments ) return lineSegments } async box( width, height, depth ) { const geometry = await new THREE.BufferGeometry() const position = [ -width, -height, -depth, -width, height, -depth, -width, height, -depth, width, height, -depth, width, height, -depth, width, -height, -depth, width, -height, -depth, -width, -height, -depth, -width, -height, depth, -width, height, depth, -width, height, depth, width, height, depth, width, height, depth, width, -height, depth, width, -height, depth, -width, -height, depth, -width, -height, -depth, -width, -height, depth, -width, height, -depth, -width, height, depth, width, height, -depth, width, height, depth, width, -height, -depth, width, -height, depth ] geometry.setAttribute( 'position', await new THREE.Float32BufferAttribute( position, 3 ) ) return geometry } updateCamera( { x, y, z } ) { this.camera.position.set( x, y, z ) this.camera.lookAt( this.scene.position ) } setGridHelper( size = 224, divisions = 50, color1 = 0x303030, color2 = 0x303030 ) { const gridHelper = new THREE.GridHelper( size, divisions, color1, color2 ) return gridHelper } addAnimateFbx( object ) { this.mixer = new THREE.AnimationMixer( object ) const len = object.animations.length // 动画的数量 if ( len > 0 ) { const action = this.mixer.clipAction( object.animations[0] ) action.play() } return this.mixer } loadFbx( baseUrl, fn ) { const loader = new FBXLoader( this
const loader = new THREE.TextureLoader() const texturePlante = loader.load( url ) const material = new THREE.MeshPhongMaterial( { map : texturePlante } ) return material } loadObj( baseUrl, objUrl, materials, fn ) { const loader = new OBJLoader( this.manager ) loader.setRequestHeader( this.requestHeader ) const that = this if ( materials ) { loader .setMaterials( materials ) .setPath( baseUrl ) .load( objUrl, ( object ) => { fn && fn( object ) }, that.onProgress, that.onError ) } else { loader.setPath( baseUrl ).load( objUrl, ( object ) => { fn && fn( object ) }, that.onProgress, that.onError ) } } loadMtl( baseUrl, mtlUrl, fn ) { const loader = new MTLLoader( this.manager ) loader.setCrossOrigin( this.crossOrigin ) loader.setRequestHeader( this.requestHeader ) const that = this loader.setPath( baseUrl ).load( mtlUrl, async( materials ) => { materials.preload() fn && fn( materials ) }, that.onProgress, that.onError ) } loadJSON( baseUrl, fn ) { const that = this const loader = new ObjectLoader() loader.setCrossOrigin( that.crossOrigin ) loader.setRequestHeader( that.requestHeader ) loader.load( baseUrl, ( object ) => { fn && fn( object ) }, that.onProgress, that.onError ) } loadStl( baseUrl, fn ) { const that = this const loader = new STLLoader() loader.load( baseUrl, ( geometry ) => { const material = new THREE.MeshPhongMaterial( { color : 0xff5533, specular : 0x111111, shininess : 200 } ) const object = new THREE.Mesh( geometry, material ) fn && fn( object ) }, that.onProgress, that.onError ) } loadPly( baseUrl, fn ) { const that = this const loader = new PLYLoader() loader.load(
.manager ) loader.setCrossOrigin( this.crossOrigin ) loader.setRequestHeader( this.requestHeader ) const that = this loader.load( baseUrl, ( object ) => { fn && fn( object ) }, that.onProgress, that.onError ) } // fbx模型加载贴图 loadImage( url ) {
identifier_body
demo.py
translate is effectively canceled out, # leaving a rotation and then a translation. # Translate inverse(Translate) Rotate Translate # # Translate inverse(Translate) = Identity. i.e. 5 * 1/5 = 1, # so we really just need to do a rotation first, and then a translation, # but this can be counterintuitive at first because we like to think # in relative terms. # To understand why the code in this demo works, you can think # about it in one of two ways. Either there is a sequence # of function calls, all of which happen relative to the global # origin; or, you can read the transformations backwards, # where instead of doing operations on points, the operations # all modify the current axis to a new relative axis, # and all subsequent functions move those relative axises to # new relative axises. # Strong suggestion for computer graphics, especially from # modelspace to global space: # Read the transformations in the latter. # See the transformations below, and the associated animated gifs. import sys import os import numpy as np import math from OpenGL.GL import * import glfw if not glfw.init(): sys.exit() glfw.window_hint(glfw.CONTEXT_VERSION_MAJOR,1) glfw.window_hint(glfw.CONTEXT_VERSION_MINOR,4) window = glfw.create_window(500, 500, "ModelViewProjection Demo 9", None, None) if not window: glfw.terminate() sys.exit() # Make the window's context current glfw.make_context_current(window) # Install a key handler def on_key(window, key, scancode, action, mods): if key == glfw.KEY_ESCAPE and action == glfw.PRESS: glfw.set_window_should_close(window,1) glfw.set_key_callback(window, on_key) glClearColor(0.0, 0.0, 0.0, 1.0) glMatrixMode(GL_PROJECTION); glLoadIdentity(); glMatrixMode(GL_MODELVIEW); glLoadIdentity(); def draw_in_square_viewport(): # clear to gray. glClearColor(0.2, #r 0.2, #g 0.2, #b 1.0) #a glClear(GL_COLOR_BUFFER_BIT) width, height = glfw.get_framebuffer_size(window) # figure out the minimum dimension of the window min = width if width < height else height # per pixel than just it's current color. glEnable(GL_SCISSOR_TEST) glScissor(int((width - min)/2.0), #min x int((height - min)/2.0), #min y min, #width x min) #width y glClearColor(0.0, #r 0.0, #g 0.0, #b 1.0) #a # gl clear will only update the square to black values. glClear(GL_COLOR_BUFFER_BIT) # disable the scissor test, so now any opengl calls will # happen as usual. glDisable(GL_SCISSOR_TEST) # But, we only want to draw within the black square. # We set the viewport, so that the NDC coordinates # will be mapped the the region of screen coordinates # that we care about, which is the black square. glViewport(int(0.0 + (width - min)/2.0), #min x int(0.0 + (height - min)/2.0), #min y min, #width x min) #width y class Vertex: def __init__(self,x,y): self.x = x self.y = y def __repr__(self): return f"Vertex(x={repr(self.x)},y={repr(self.y)})" def translate(self, tx, ty): return Vertex(x=self.x + tx, y=self.y + ty) def scale(self, scale_x, scale_y): return Vertex(x=self.x * scale_x, y=self.y * scale_y) def rotate(self,angle_in_radians): return Vertex(x= self.x * math.cos(angle_in_radians) - self.y * math.sin(angle_in_radians), y= self.x * math.sin(angle_in_radians) + self.y * math.cos(angle_in_radians)) # NEW # removed rotate_around, as it was useless for our purpose class Paddle: def __init__(self,vertices, r, g, b, initial_position, rotation=0.0, input_offset_x=0.0, input_offset_y=0.0): self.vertices = vertices self.r = r self.g = g self.b = b self.rotation = rotation self.input_offset_x = input_offset_x self.input_offset_y = input_offset_y self.initial_position = initial_position def __repr__(self): return f"Paddle(vertices={repr(self.vertices)},r={repr(self.r)},g={repr(self.g)},b={repr(self.b)},initial_position={repr(self.initial_position)},rotation={repr(self.rotation)},input_offset_x={repr(self.input_offset_x)},input_offset_y={repr({self.input_offset_y})})" paddle1 = Paddle(vertices=[Vertex(x=-10.0, y=-30.0), Vertex(x= 10.0, y=-30.0), Vertex(x= 10.0, y=30.0), Vertex(x=-10.0, y=30.0)], r=0.578123, g=0.0, b=1.0, initial_position=Vertex(-90.0,0.0)) paddle2 = Paddle(vertices=[Vertex(x=-10.0, y=-30.0), Vertex(x= 10.0, y=-30.0), Vertex(x= 10.0, y=30.0), Vertex(x=-10.0, y=30.0)], r=1.0, g=0.0, b=0.0, initial_position=Vertex(90.0,0.0)) def handle_movement_of_paddles():
paddle2.rotation -= 0.1 TARGET_FRAMERATE = 60 # fps # to try to standardize on 60 fps, compare times between frames time_at_beginning_of_previous_frame = glfw.get_time() # Loop until the user closes the window while not glfw.window_should_close(window): # poll the time to try to get a constant framerate while glfw.get_time() < time_at_beginning_of_previous_frame + 1.0/TARGET_FRAMERATE: pass # set for comparison on the next frame time_at_beginning_of_previous_frame = glfw.get_time() # Poll for and process events glfw.poll_events() width, height = glfw.get_framebuffer_size(window) glViewport(0, 0, width, height) glClear(GL_COLOR_BUFFER_BIT | GL_DEPTH_BUFFER_BIT) # render scene draw_in_square_viewport() handle_movement_of_paddles() # draw paddle1 glColor3f(paddle1.r, paddle1.g, paddle1.b) # if you read the operations below as rotate, translate1, translate2, # you should imagine it as follows # eog ../images/rotation1F.gif # if instead you read them backwards, imagine the transformations # as follows # eog ../images/rotation1B.gif # side note. Typically I use a debugger as an interactive evaluator, # in order to understand how code which I do not understand works. # In computer graphics, the debugger is of limited help because # the transformations on the individual points is not worth # thinking about, and therefore the intermediat results # are worthless for reasoning. # # In order to be successful, I highly recommend reading the transformations # backwards, with a moving/rotating/scaled axises. # # (This advise will be modified when I introduce transformation stacks, # but the same principle will apply. Also, on
global paddle1, paddle2 if glfw.get_key(window, glfw.KEY_S) == glfw.PRESS: paddle1.input_offset_y -= 10.0 if glfw.get_key(window, glfw.KEY_W) == glfw.PRESS: paddle1.input_offset_y += 10.0 if glfw.get_key(window, glfw.KEY_K) == glfw.PRESS: paddle2.input_offset_y -= 10.0 if glfw.get_key(window, glfw.KEY_I) == glfw.PRESS: paddle2.input_offset_y += 10.0 global paddle_1_rotation, paddle_2_rotation if glfw.get_key(window, glfw.KEY_A) == glfw.PRESS: paddle1.rotation += 0.1 if glfw.get_key(window, glfw.KEY_D) == glfw.PRESS: paddle1.rotation -= 0.1 if glfw.get_key(window, glfw.KEY_J) == glfw.PRESS: paddle2.rotation += 0.1 if glfw.get_key(window, glfw.KEY_L) == glfw.PRESS:
identifier_body
demo.py
translate is effectively canceled out, # leaving a rotation and then a translation. # Translate inverse(Translate) Rotate Translate # # Translate inverse(Translate) = Identity. i.e. 5 * 1/5 = 1, # so we really just need to do a rotation first, and then a translation, # but this can be counterintuitive at first because we like to think # in relative terms. # To understand why the code in this demo works, you can think # about it in one of two ways. Either there is a sequence # of function calls, all of which happen relative to the global # origin; or, you can read the transformations backwards, # where instead of doing operations on points, the operations # all modify the current axis to a new relative axis, # and all subsequent functions move those relative axises to # new relative axises. # Strong suggestion for computer graphics, especially from # modelspace to global space: # Read the transformations in the latter. # See the transformations below, and the associated animated gifs. import sys import os import numpy as np import math from OpenGL.GL import * import glfw if not glfw.init(): sys.exit() glfw.window_hint(glfw.CONTEXT_VERSION_MAJOR,1) glfw.window_hint(glfw.CONTEXT_VERSION_MINOR,4) window = glfw.create_window(500, 500, "ModelViewProjection Demo 9", None, None) if not window: glfw.terminate() sys.exit() # Make the window's context current glfw.make_context_current(window) # Install a key handler def on_key(window, key, scancode, action, mods): if key == glfw.KEY_ESCAPE and action == glfw.PRESS: glfw.set_window_should_close(window,1) glfw.set_key_callback(window, on_key) glClearColor(0.0, 0.0, 0.0, 1.0) glMatrixMode(GL_PROJECTION); glLoadIdentity(); glMatrixMode(GL_MODELVIEW); glLoadIdentity(); def draw_in_square_viewport(): # clear to gray. glClearColor(0.2, #r 0.2, #g 0.2, #b 1.0) #a glClear(GL_COLOR_BUFFER_BIT) width, height = glfw.get_framebuffer_size(window) # figure out the minimum dimension of the window min = width if width < height else height # per pixel than just it's current color. glEnable(GL_SCISSOR_TEST) glScissor(int((width - min)/2.0), #min x int((height - min)/2.0), #min y min, #width x min) #width y glClearColor(0.0, #r 0.0, #g 0.0, #b 1.0) #a # gl clear will only update the square to black values. glClear(GL_COLOR_BUFFER_BIT) # disable the scissor test, so now any opengl calls will # happen as usual. glDisable(GL_SCISSOR_TEST) # But, we only want to draw within the black square. # We set the viewport, so that the NDC coordinates # will be mapped the the region of screen coordinates # that we care about, which is the black square. glViewport(int(0.0 + (width - min)/2.0), #min x int(0.0 + (height - min)/2.0), #min y min, #width x min) #width y class Vertex: def __init__(self,x,y): self.x = x self.y = y def __repr__(self): return f"Vertex(x={repr(self.x)},y={repr(self.y)})" def translate(self, tx, ty): return Vertex(x=self.x + tx, y=self.y + ty) def scale(self, scale_x, scale_y): return Vertex(x=self.x * scale_x, y=self.y * scale_y) def rotate(self,angle_in_radians): return Vertex(x= self.x * math.cos(angle_in_radians) - self.y * math.sin(angle_in_radians), y= self.x * math.sin(angle_in_radians) + self.y * math.cos(angle_in_radians)) # NEW # removed rotate_around, as it was useless for our purpose class Paddle: def __init__(self,vertices, r, g, b, initial_position, rotation=0.0, input_offset_x=0.0, input_offset_y=0.0): self.vertices = vertices self.r = r self.g = g self.b = b self.rotation = rotation self.input_offset_x = input_offset_x self.input_offset_y = input_offset_y self.initial_position = initial_position def __repr__(self): return f"Paddle(vertices={repr(self.vertices)},r={repr(self.r)},g={repr(self.g)},b={repr(self.b)},initial_position={repr(self.initial_position)},rotation={repr(self.rotation)},input_offset_x={repr(self.input_offset_x)},input_offset_y={repr({self.input_offset_y})})" paddle1 = Paddle(vertices=[Vertex(x=-10.0, y=-30.0), Vertex(x= 10.0, y=-30.0),
Vertex(x= 10.0, y=30.0), Vertex(x=-10.0, y=30.0)], r=0.578123, g=0.0, b=1.0, initial_position=Vertex(-90.0,0.0)) paddle2 = Paddle(vertices=[Vertex(x=-10.0, y=-30.0), Vertex(x= 10.0, y=-30.0), Vertex(x= 10.0, y=30.0), Vertex(x=-10.0, y=30.0)], r=1.0, g=0.0, b=0.0, initial_position=Vertex(90.0,0.0)) def handle_movement_of_paddles(): global paddle1, paddle2 if glfw.get_key(window, glfw.KEY_S) == glfw.PRESS: paddle1.input_offset_y -= 10.0 if glfw.get_key(window, glfw.KEY_W) == glfw.PRESS: paddle1.input_offset_y += 10.0 if glfw.get_key(window, glfw.KEY_K) == glfw.PRESS: paddle2.input_offset_y -= 10.0 if glfw.get_key(window, glfw.KEY_I) == glfw.PRESS: paddle2.input_offset_y += 10.0 global paddle_1_rotation, paddle_2_rotation if glfw.get_key(window, glfw.KEY_A) == glfw.PRESS: paddle1.rotation += 0.1 if glfw.get_key(window, glfw.KEY_D) == glfw.PRESS: paddle1.rotation -= 0.1 if glfw.get_key(window, glfw.KEY_J) == glfw.PRESS: paddle2.rotation += 0.1 if glfw.get_key(window, glfw.KEY_L) == glfw.PRESS: paddle2.rotation -= 0.1 TARGET_FRAMERATE = 60 # fps # to try to standardize on 60 fps, compare times between frames time_at_beginning_of_previous_frame = glfw.get_time() # Loop until the user closes the window while not glfw.window_should_close(window): # poll the time to try to get a constant framerate while glfw.get_time() < time_at_beginning_of_previous_frame + 1.0/TARGET_FRAMERATE: pass # set for comparison on the next frame time_at_beginning_of_previous_frame = glfw.get_time() # Poll for and process events glfw.poll_events() width, height = glfw.get_framebuffer_size(window) glViewport(0, 0, width, height) glClear(GL_COLOR_BUFFER_BIT | GL_DEPTH_BUFFER_BIT) # render scene draw_in_square_viewport() handle_movement_of_paddles() # draw paddle1 glColor3f(paddle1.r, paddle1.g, paddle1.b) # if you read the operations below as rotate, translate1, translate2, # you should imagine it as follows # eog ../images/rotation1F.gif # if instead you read them backwards, imagine the transformations # as follows # eog ../images/rotation1B.gif # side note. Typically I use a debugger as an interactive evaluator, # in order to understand how code which I do not understand works. # In computer graphics, the debugger is of limited help because # the transformations on the individual points is not worth # thinking about, and therefore the intermediat results # are worthless for reasoning. # # In order to be successful, I highly recommend reading the transformations # backwards, with a moving/rotating/scaled axises. # # (This advise will be modified when I introduce transformation stacks, # but the same principle will apply. Also, on
random_line_split
demo.py
translate is effectively canceled out, # leaving a rotation and then a translation. # Translate inverse(Translate) Rotate Translate # # Translate inverse(Translate) = Identity. i.e. 5 * 1/5 = 1, # so we really just need to do a rotation first, and then a translation, # but this can be counterintuitive at first because we like to think # in relative terms. # To understand why the code in this demo works, you can think # about it in one of two ways. Either there is a sequence # of function calls, all of which happen relative to the global # origin; or, you can read the transformations backwards, # where instead of doing operations on points, the operations # all modify the current axis to a new relative axis, # and all subsequent functions move those relative axises to # new relative axises. # Strong suggestion for computer graphics, especially from # modelspace to global space: # Read the transformations in the latter. # See the transformations below, and the associated animated gifs. import sys import os import numpy as np import math from OpenGL.GL import * import glfw if not glfw.init(): sys.exit() glfw.window_hint(glfw.CONTEXT_VERSION_MAJOR,1) glfw.window_hint(glfw.CONTEXT_VERSION_MINOR,4) window = glfw.create_window(500, 500, "ModelViewProjection Demo 9", None, None) if not window: glfw.terminate() sys.exit() # Make the window's context current glfw.make_context_current(window) # Install a key handler def on_key(window, key, scancode, action, mods): if key == glfw.KEY_ESCAPE and action == glfw.PRESS: glfw.set_window_should_close(window,1) glfw.set_key_callback(window, on_key) glClearColor(0.0, 0.0, 0.0, 1.0) glMatrixMode(GL_PROJECTION); glLoadIdentity(); glMatrixMode(GL_MODELVIEW); glLoadIdentity(); def draw_in_square_viewport(): # clear to gray. glClearColor(0.2, #r 0.2, #g 0.2, #b 1.0) #a glClear(GL_COLOR_BUFFER_BIT) width, height = glfw.get_framebuffer_size(window) # figure out the minimum dimension of the window min = width if width < height else height # per pixel than just it's current color. glEnable(GL_SCISSOR_TEST) glScissor(int((width - min)/2.0), #min x int((height - min)/2.0), #min y min, #width x min) #width y glClearColor(0.0, #r 0.0, #g 0.0, #b 1.0) #a # gl clear will only update the square to black values. glClear(GL_COLOR_BUFFER_BIT) # disable the scissor test, so now any opengl calls will # happen as usual. glDisable(GL_SCISSOR_TEST) # But, we only want to draw within the black square. # We set the viewport, so that the NDC coordinates # will be mapped the the region of screen coordinates # that we care about, which is the black square. glViewport(int(0.0 + (width - min)/2.0), #min x int(0.0 + (height - min)/2.0), #min y min, #width x min) #width y class Vertex: def __init__(self,x,y): self.x = x self.y = y def
(self): return f"Vertex(x={repr(self.x)},y={repr(self.y)})" def translate(self, tx, ty): return Vertex(x=self.x + tx, y=self.y + ty) def scale(self, scale_x, scale_y): return Vertex(x=self.x * scale_x, y=self.y * scale_y) def rotate(self,angle_in_radians): return Vertex(x= self.x * math.cos(angle_in_radians) - self.y * math.sin(angle_in_radians), y= self.x * math.sin(angle_in_radians) + self.y * math.cos(angle_in_radians)) # NEW # removed rotate_around, as it was useless for our purpose class Paddle: def __init__(self,vertices, r, g, b, initial_position, rotation=0.0, input_offset_x=0.0, input_offset_y=0.0): self.vertices = vertices self.r = r self.g = g self.b = b self.rotation = rotation self.input_offset_x = input_offset_x self.input_offset_y = input_offset_y self.initial_position = initial_position def __repr__(self): return f"Paddle(vertices={repr(self.vertices)},r={repr(self.r)},g={repr(self.g)},b={repr(self.b)},initial_position={repr(self.initial_position)},rotation={repr(self.rotation)},input_offset_x={repr(self.input_offset_x)},input_offset_y={repr({self.input_offset_y})})" paddle1 = Paddle(vertices=[Vertex(x=-10.0, y=-30.0), Vertex(x= 10.0, y=-30.0), Vertex(x= 10.0, y=30.0), Vertex(x=-10.0, y=30.0)], r=0.578123, g=0.0, b=1.0, initial_position=Vertex(-90.0,0.0)) paddle2 = Paddle(vertices=[Vertex(x=-10.0, y=-30.0), Vertex(x= 10.0, y=-30.0), Vertex(x= 10.0, y=30.0), Vertex(x=-10.0, y=30.0)], r=1.0, g=0.0, b=0.0, initial_position=Vertex(90.0,0.0)) def handle_movement_of_paddles(): global paddle1, paddle2 if glfw.get_key(window, glfw.KEY_S) == glfw.PRESS: paddle1.input_offset_y -= 10.0 if glfw.get_key(window, glfw.KEY_W) == glfw.PRESS: paddle1.input_offset_y += 10.0 if glfw.get_key(window, glfw.KEY_K) == glfw.PRESS: paddle2.input_offset_y -= 10.0 if glfw.get_key(window, glfw.KEY_I) == glfw.PRESS: paddle2.input_offset_y += 10.0 global paddle_1_rotation, paddle_2_rotation if glfw.get_key(window, glfw.KEY_A) == glfw.PRESS: paddle1.rotation += 0.1 if glfw.get_key(window, glfw.KEY_D) == glfw.PRESS: paddle1.rotation -= 0.1 if glfw.get_key(window, glfw.KEY_J) == glfw.PRESS: paddle2.rotation += 0.1 if glfw.get_key(window, glfw.KEY_L) == glfw.PRESS: paddle2.rotation -= 0.1 TARGET_FRAMERATE = 60 # fps # to try to standardize on 60 fps, compare times between frames time_at_beginning_of_previous_frame = glfw.get_time() # Loop until the user closes the window while not glfw.window_should_close(window): # poll the time to try to get a constant framerate while glfw.get_time() < time_at_beginning_of_previous_frame + 1.0/TARGET_FRAMERATE: pass # set for comparison on the next frame time_at_beginning_of_previous_frame = glfw.get_time() # Poll for and process events glfw.poll_events() width, height = glfw.get_framebuffer_size(window) glViewport(0, 0, width, height) glClear(GL_COLOR_BUFFER_BIT | GL_DEPTH_BUFFER_BIT) # render scene draw_in_square_viewport() handle_movement_of_paddles() # draw paddle1 glColor3f(paddle1.r, paddle1.g, paddle1.b) # if you read the operations below as rotate, translate1, translate2, # you should imagine it as follows # eog ../images/rotation1F.gif # if instead you read them backwards, imagine the transformations # as follows # eog ../images/rotation1B.gif # side note. Typically I use a debugger as an interactive evaluator, # in order to understand how code which I do not understand works. # In computer graphics, the debugger is of limited help because # the transformations on the individual points is not worth # thinking about, and therefore the intermediat results # are worthless for reasoning. # # In order to be successful, I highly recommend reading the transformations # backwards, with a moving/rotating/scaled axises. # # (This advise will be modified when I introduce transformation stacks, # but the same principle will apply. Also,
__repr__
identifier_name
demo.py
translate is effectively canceled out, # leaving a rotation and then a translation. # Translate inverse(Translate) Rotate Translate # # Translate inverse(Translate) = Identity. i.e. 5 * 1/5 = 1, # so we really just need to do a rotation first, and then a translation, # but this can be counterintuitive at first because we like to think # in relative terms. # To understand why the code in this demo works, you can think # about it in one of two ways. Either there is a sequence # of function calls, all of which happen relative to the global # origin; or, you can read the transformations backwards, # where instead of doing operations on points, the operations # all modify the current axis to a new relative axis, # and all subsequent functions move those relative axises to # new relative axises. # Strong suggestion for computer graphics, especially from # modelspace to global space: # Read the transformations in the latter. # See the transformations below, and the associated animated gifs. import sys import os import numpy as np import math from OpenGL.GL import * import glfw if not glfw.init(): sys.exit() glfw.window_hint(glfw.CONTEXT_VERSION_MAJOR,1) glfw.window_hint(glfw.CONTEXT_VERSION_MINOR,4) window = glfw.create_window(500, 500, "ModelViewProjection Demo 9", None, None) if not window: glfw.terminate() sys.exit() # Make the window's context current glfw.make_context_current(window) # Install a key handler def on_key(window, key, scancode, action, mods): if key == glfw.KEY_ESCAPE and action == glfw.PRESS: glfw.set_window_should_close(window,1) glfw.set_key_callback(window, on_key) glClearColor(0.0, 0.0, 0.0, 1.0) glMatrixMode(GL_PROJECTION); glLoadIdentity(); glMatrixMode(GL_MODELVIEW); glLoadIdentity(); def draw_in_square_viewport(): # clear to gray. glClearColor(0.2, #r 0.2, #g 0.2, #b 1.0) #a glClear(GL_COLOR_BUFFER_BIT) width, height = glfw.get_framebuffer_size(window) # figure out the minimum dimension of the window min = width if width < height else height # per pixel than just it's current color. glEnable(GL_SCISSOR_TEST) glScissor(int((width - min)/2.0), #min x int((height - min)/2.0), #min y min, #width x min) #width y glClearColor(0.0, #r 0.0, #g 0.0, #b 1.0) #a # gl clear will only update the square to black values. glClear(GL_COLOR_BUFFER_BIT) # disable the scissor test, so now any opengl calls will # happen as usual. glDisable(GL_SCISSOR_TEST) # But, we only want to draw within the black square. # We set the viewport, so that the NDC coordinates # will be mapped the the region of screen coordinates # that we care about, which is the black square. glViewport(int(0.0 + (width - min)/2.0), #min x int(0.0 + (height - min)/2.0), #min y min, #width x min) #width y class Vertex: def __init__(self,x,y): self.x = x self.y = y def __repr__(self): return f"Vertex(x={repr(self.x)},y={repr(self.y)})" def translate(self, tx, ty): return Vertex(x=self.x + tx, y=self.y + ty) def scale(self, scale_x, scale_y): return Vertex(x=self.x * scale_x, y=self.y * scale_y) def rotate(self,angle_in_radians): return Vertex(x= self.x * math.cos(angle_in_radians) - self.y * math.sin(angle_in_radians), y= self.x * math.sin(angle_in_radians) + self.y * math.cos(angle_in_radians)) # NEW # removed rotate_around, as it was useless for our purpose class Paddle: def __init__(self,vertices, r, g, b, initial_position, rotation=0.0, input_offset_x=0.0, input_offset_y=0.0): self.vertices = vertices self.r = r self.g = g self.b = b self.rotation = rotation self.input_offset_x = input_offset_x self.input_offset_y = input_offset_y self.initial_position = initial_position def __repr__(self): return f"Paddle(vertices={repr(self.vertices)},r={repr(self.r)},g={repr(self.g)},b={repr(self.b)},initial_position={repr(self.initial_position)},rotation={repr(self.rotation)},input_offset_x={repr(self.input_offset_x)},input_offset_y={repr({self.input_offset_y})})" paddle1 = Paddle(vertices=[Vertex(x=-10.0, y=-30.0), Vertex(x= 10.0, y=-30.0), Vertex(x= 10.0, y=30.0), Vertex(x=-10.0, y=30.0)], r=0.578123, g=0.0, b=1.0, initial_position=Vertex(-90.0,0.0)) paddle2 = Paddle(vertices=[Vertex(x=-10.0, y=-30.0), Vertex(x= 10.0, y=-30.0), Vertex(x= 10.0, y=30.0), Vertex(x=-10.0, y=30.0)], r=1.0, g=0.0, b=0.0, initial_position=Vertex(90.0,0.0)) def handle_movement_of_paddles(): global paddle1, paddle2 if glfw.get_key(window, glfw.KEY_S) == glfw.PRESS: paddle1.input_offset_y -= 10.0 if glfw.get_key(window, glfw.KEY_W) == glfw.PRESS: paddle1.input_offset_y += 10.0 if glfw.get_key(window, glfw.KEY_K) == glfw.PRESS: paddle2.input_offset_y -= 10.0 if glfw.get_key(window, glfw.KEY_I) == glfw.PRESS:
global paddle_1_rotation, paddle_2_rotation if glfw.get_key(window, glfw.KEY_A) == glfw.PRESS: paddle1.rotation += 0.1 if glfw.get_key(window, glfw.KEY_D) == glfw.PRESS: paddle1.rotation -= 0.1 if glfw.get_key(window, glfw.KEY_J) == glfw.PRESS: paddle2.rotation += 0.1 if glfw.get_key(window, glfw.KEY_L) == glfw.PRESS: paddle2.rotation -= 0.1 TARGET_FRAMERATE = 60 # fps # to try to standardize on 60 fps, compare times between frames time_at_beginning_of_previous_frame = glfw.get_time() # Loop until the user closes the window while not glfw.window_should_close(window): # poll the time to try to get a constant framerate while glfw.get_time() < time_at_beginning_of_previous_frame + 1.0/TARGET_FRAMERATE: pass # set for comparison on the next frame time_at_beginning_of_previous_frame = glfw.get_time() # Poll for and process events glfw.poll_events() width, height = glfw.get_framebuffer_size(window) glViewport(0, 0, width, height) glClear(GL_COLOR_BUFFER_BIT | GL_DEPTH_BUFFER_BIT) # render scene draw_in_square_viewport() handle_movement_of_paddles() # draw paddle1 glColor3f(paddle1.r, paddle1.g, paddle1.b) # if you read the operations below as rotate, translate1, translate2, # you should imagine it as follows # eog ../images/rotation1F.gif # if instead you read them backwards, imagine the transformations # as follows # eog ../images/rotation1B.gif # side note. Typically I use a debugger as an interactive evaluator, # in order to understand how code which I do not understand works. # In computer graphics, the debugger is of limited help because # the transformations on the individual points is not worth # thinking about, and therefore the intermediat results # are worthless for reasoning. # # In order to be successful, I highly recommend reading the transformations # backwards, with a moving/rotating/scaled axises. # # (This advise will be modified when I introduce transformation stacks, # but the same principle will apply. Also,
paddle2.input_offset_y += 10.0
conditional_block
workplace_preparation.py
images') if path.exists(path_to_images) and path.isdir(path_to_images): shutil.rmtree(path_to_images) makedirs(path_to_images) # split the given video into images subprocess.run(['ffmpeg', '-i', temp_video, '-r', str(number_of_images / video_duration), '-f', 'image2', path.join(path_to_images, 'image%d.jpg')]) # remove extra files remove_extra_images(path_to_images, number_of_images) remove(temp_video) def create_temp_model(temp_dir_path: str) -> str: """ The function prepares the images for our model based on a given video :param temp_dir_path: video in h264 format :return number_of_images: path to temporary model folder """ # create temp images folder path_to_temp_model = path.join(temp_dir_path, 'temp_model') path_to_temp_images = path.join(path_to_temp_model, 'temp_images') # remove old temporary folder if exists if path.exists(path_to_temp_model) and path.isdir(path_to_temp_model): shutil.rmtree(path_to_temp_model) number_of_temp_images = 0 path_to_images = path.join(temp_dir_path, 'images') # take only part of the images for the temp model while number_of_temp_images < number_of_images_in_temp_model: try: number_of_temp_images = len([name for name in listdir(path_to_images) if name.endswith('.jpg')]) except FileNotFoundError: number_of_temp_images = 0 # copy subdirectory example shutil.copytree(path_to_images, path_to_temp_images) # run colmap to create model for the first 10 images in video subprocess.run(['colmap', 'automatic_reconstructor', '--workspace_path', path_to_temp_model, '--image_path', path_to_temp_images, '--data_type=video', '--quality=extreme']) return path_to_temp_model def quaternion_to_rotation_matrix(q0, q1, q2, q3) -> np: """ The function converts the quaternion vector to a rotation matrix https://automaticaddison.com/how-to-convert-a-quaternion-to-a-rotation-matrix/ :param q0: the value of qw :param q1: the value of qx :param q2: the value of qy :param q3: the value of qz :return rot_matrix: rotation matrix 3x3 as NumPy array """ # First row of the rotation matrix r00 = 2 * (q0 * q0 + q1 * q1) - 1 r01 = 2 * (q1 * q2 - q0 * q3) r02 = 2 * (q1 * q3 + q0 * q2) # Second row of the rotation matrix r10 = 2 * (q1 * q2 + q0 * q3) r11 = 2 * (q0 * q0 + q2 * q2) - 1 r12 = 2 * (q2 * q3 - q0 * q1) # Third row of the rotation matrix r20 = 2 * (q1 * q3 - q0 * q2) r21 = 2 * (q2 * q3 + q0 * q1) r22 = 2 * (q0 * q0 + q3 * q3) - 1 # 3x3 rotation matrix rot_matrix = np.array([[r00, r01, r02], [r10, r11, r12], [r20, r21, r22]]) return rot_matrix def rotation_matrix_to_quaternion(rotation_matrix: np) -> object: """ The function converts rotation matrix to quaternion vector https://learnopencv.com/rotation-matrix-to-euler-angles/ :param rotation_matrix: rotation matrix 3x3 represented by NumPy array :return quaternion vector: defined by (qx, qy, qz, qw) """ cosine_for_pitch = math.sqrt(rotation_matrix[0][0] ** 2 + rotation_matrix[1][0] ** 2) is_singular = cosine_for_pitch < 10 ** -6 if not is_singular: yaw = math.atan2(rotation_matrix[1][0], rotation_matrix[0][0]) pitch = math.atan2(-rotation_matrix[2][0], cosine_for_pitch) roll = math.atan2(rotation_matrix[2][1], rotation_matrix[2][2]) else: yaw = math.atan2(-rotation_matrix[1][2], rotation_matrix[1][1]) pitch = math.atan2(-rotation_matrix[2][0], cosine_for_pitch) roll = 0 e = (yaw, pitch, roll) return euler_to_quaternion(e) def euler_to_quaternion(euler: tuple) -> object: """ The function convert Euler angle to quaternion object :param Euler: angle represented by yaw, pitch, roll :return quaternion vector: defined by (qx, qy, qz, qw) """ (yaw, pitch, roll) = (euler[0], euler[1], euler[2]) qy = np.sin(roll / 2) * np.cos(pitch / 2) * np.cos(yaw / 2) - np.cos(roll / 2) * np.sin(pitch / 2) * np.sin(yaw / 2) qx = np.cos(roll / 2) * np.sin(pitch / 2) * np.cos(yaw / 2) + np.sin(roll / 2) * np.cos(pitch / 2) * np.sin(yaw / 2) qw = np.cos(roll / 2) * np.cos(pitch / 2) * np.sin(yaw / 2) - np.sin(roll / 2) * np.sin(pitch / 2) * np.cos(yaw / 2) qz = np.cos(roll / 2) * np.cos(pitch / 2) * np.cos(yaw / 2) + np.sin(roll / 2) * np.sin(pitch / 2) * np.sin(yaw / 2) return qx, qy, qz, qw def get_first_image_pose(image_src: str) -> list:
qz = float(columns[3]) rotation_matrix = quaternion_to_rotation_matrix(qw, qx, qy, qz) tx = float(columns[5]) ty = float(columns[7]) tz = float(columns[6]) translation_vector = np.array([tx, ty, tz]) return [rotation_matrix, translation_vector] return [] def draw_rel_camera_pose(image: int, origin: list, camera_pose: list, plot_dir_path: str) -> None: """ Debug function for plotting the relative camera poses :param image: number of current image :param origin: list of [x,y,z] of the origin :param camera_pose: list of [x1,y1,z1][x2,y2,z2] of the camera pose (three 2d vectors) :param plot_dir_path: path to plot directory """ fig = plt.figure() ax = fig.add_subplot(111, projection='3d') ax.view_init(elev=10.) ax.set_title('camera pose image: %d' % image) scale = 7 ax.set_xlim3d(-scale, scale) ax.set_ylim3d(-scale, scale) ax.set_zlim3d(-scale, scale) # replace the Y-Axis with Z-Axis ax.scatter(origin[0], origin[2], origin[1], c='black') for i in range(3): ax.plot([origin[0], camera_pose[i][0]], [origin[2], camera_pose[i][2]], [origin[1], camera_pose[i][1]]) i += 1 fig.savefig(f'{plot_dir_path}/%d.png' % image) plt.close(fig) plt.clf() def compute_absolut_camera_pose(camera_pose_rel_dict: dict, first_image_pose: list, workspace_path: str, do_plot=False) -> dict: """ The function return a dictionary with recovered R&T for each image :param camera_pose_rel_dict: dictionary of relative camera poses for each image
""" The function return the absolut R & T for the first image in temp model :param image_src: path to image file (colmap output) :return R&T: R = list[0], T list[1] or None if image1 not exists """ # read images file with open(image_src, 'r') as file: lines = file.readlines()[4::2] # create absolut camera pose dictionary for line in lines: columns = line.split() image_name = columns[9].split('.')[0] image_id = int(image_name.split('e')[1]) # convert and return the camera pose for the first image in model if image_id == 1: qw = float(columns[1]) qx = float(columns[2]) qy = float(columns[4])
identifier_body
workplace_preparation.py
.rmtree(path_to_temp_model) number_of_temp_images = 0 path_to_images = path.join(temp_dir_path, 'images') # take only part of the images for the temp model while number_of_temp_images < number_of_images_in_temp_model: try: number_of_temp_images = len([name for name in listdir(path_to_images) if name.endswith('.jpg')]) except FileNotFoundError: number_of_temp_images = 0 # copy subdirectory example shutil.copytree(path_to_images, path_to_temp_images) # run colmap to create model for the first 10 images in video subprocess.run(['colmap', 'automatic_reconstructor', '--workspace_path', path_to_temp_model, '--image_path', path_to_temp_images, '--data_type=video', '--quality=extreme']) return path_to_temp_model def quaternion_to_rotation_matrix(q0, q1, q2, q3) -> np: """ The function converts the quaternion vector to a rotation matrix https://automaticaddison.com/how-to-convert-a-quaternion-to-a-rotation-matrix/ :param q0: the value of qw :param q1: the value of qx :param q2: the value of qy :param q3: the value of qz :return rot_matrix: rotation matrix 3x3 as NumPy array """ # First row of the rotation matrix r00 = 2 * (q0 * q0 + q1 * q1) - 1 r01 = 2 * (q1 * q2 - q0 * q3) r02 = 2 * (q1 * q3 + q0 * q2) # Second row of the rotation matrix r10 = 2 * (q1 * q2 + q0 * q3) r11 = 2 * (q0 * q0 + q2 * q2) - 1 r12 = 2 * (q2 * q3 - q0 * q1) # Third row of the rotation matrix r20 = 2 * (q1 * q3 - q0 * q2) r21 = 2 * (q2 * q3 + q0 * q1) r22 = 2 * (q0 * q0 + q3 * q3) - 1 # 3x3 rotation matrix rot_matrix = np.array([[r00, r01, r02], [r10, r11, r12], [r20, r21, r22]]) return rot_matrix def rotation_matrix_to_quaternion(rotation_matrix: np) -> object: """ The function converts rotation matrix to quaternion vector https://learnopencv.com/rotation-matrix-to-euler-angles/ :param rotation_matrix: rotation matrix 3x3 represented by NumPy array :return quaternion vector: defined by (qx, qy, qz, qw) """ cosine_for_pitch = math.sqrt(rotation_matrix[0][0] ** 2 + rotation_matrix[1][0] ** 2) is_singular = cosine_for_pitch < 10 ** -6 if not is_singular: yaw = math.atan2(rotation_matrix[1][0], rotation_matrix[0][0]) pitch = math.atan2(-rotation_matrix[2][0], cosine_for_pitch) roll = math.atan2(rotation_matrix[2][1], rotation_matrix[2][2]) else: yaw = math.atan2(-rotation_matrix[1][2], rotation_matrix[1][1]) pitch = math.atan2(-rotation_matrix[2][0], cosine_for_pitch) roll = 0 e = (yaw, pitch, roll) return euler_to_quaternion(e) def euler_to_quaternion(euler: tuple) -> object: """ The function convert Euler angle to quaternion object :param Euler: angle represented by yaw, pitch, roll :return quaternion vector: defined by (qx, qy, qz, qw) """ (yaw, pitch, roll) = (euler[0], euler[1], euler[2]) qy = np.sin(roll / 2) * np.cos(pitch / 2) * np.cos(yaw / 2) - np.cos(roll / 2) * np.sin(pitch / 2) * np.sin(yaw / 2) qx = np.cos(roll / 2) * np.sin(pitch / 2) * np.cos(yaw / 2) + np.sin(roll / 2) * np.cos(pitch / 2) * np.sin(yaw / 2) qw = np.cos(roll / 2) * np.cos(pitch / 2) * np.sin(yaw / 2) - np.sin(roll / 2) * np.sin(pitch / 2) * np.cos(yaw / 2) qz = np.cos(roll / 2) * np.cos(pitch / 2) * np.cos(yaw / 2) + np.sin(roll / 2) * np.sin(pitch / 2) * np.sin(yaw / 2) return qx, qy, qz, qw def get_first_image_pose(image_src: str) -> list: """ The function return the absolut R & T for the first image in temp model :param image_src: path to image file (colmap output) :return R&T: R = list[0], T list[1] or None if image1 not exists """ # read images file with open(image_src, 'r') as file: lines = file.readlines()[4::2] # create absolut camera pose dictionary for line in lines: columns = line.split() image_name = columns[9].split('.')[0] image_id = int(image_name.split('e')[1]) # convert and return the camera pose for the first image in model if image_id == 1: qw = float(columns[1]) qx = float(columns[2]) qy = float(columns[4]) qz = float(columns[3]) rotation_matrix = quaternion_to_rotation_matrix(qw, qx, qy, qz) tx = float(columns[5]) ty = float(columns[7]) tz = float(columns[6]) translation_vector = np.array([tx, ty, tz]) return [rotation_matrix, translation_vector] return [] def draw_rel_camera_pose(image: int, origin: list, camera_pose: list, plot_dir_path: str) -> None: """ Debug function for plotting the relative camera poses :param image: number of current image :param origin: list of [x,y,z] of the origin :param camera_pose: list of [x1,y1,z1][x2,y2,z2] of the camera pose (three 2d vectors) :param plot_dir_path: path to plot directory """ fig = plt.figure() ax = fig.add_subplot(111, projection='3d') ax.view_init(elev=10.) ax.set_title('camera pose image: %d' % image) scale = 7 ax.set_xlim3d(-scale, scale) ax.set_ylim3d(-scale, scale) ax.set_zlim3d(-scale, scale) # replace the Y-Axis with Z-Axis ax.scatter(origin[0], origin[2], origin[1], c='black') for i in range(3): ax.plot([origin[0], camera_pose[i][0]], [origin[2], camera_pose[i][2]], [origin[1], camera_pose[i][1]]) i += 1 fig.savefig(f'{plot_dir_path}/%d.png' % image) plt.close(fig) plt.clf() def compute_absolut_camera_pose(camera_pose_rel_dict: dict, first_image_pose: list, workspace_path: str, do_plot=False) -> dict: """ The function return a dictionary with recovered R&T for each image :param camera_pose_rel_dict: dictionary of relative camera poses for each image :param first_image_pose: absolute R&T of the first image :param workspace_path: path to workspace_path :param do_plot: boolean flag for debug purpose :return: camera_pose_recover dictionary """ # create directory for reference plots ref_pose_images_path = path.join(workspace_path, 'ref_images') if do_plot: makedirs(ref_pose_images_path) # initialize parameters for computing absolut camera poses camera_pose_recover = {} rotation = first_image_pose[0] translation = first_image_pose[1] is_first = True prev_rotation = np.identity(3) prev_translation = np.zeros(3) # foreach image compute the absolut pose out of the reference pose for image in camera_pose_rel_dict.keys(): rel_rotation = camera_pose_rel_dict[image][0] rel_translation = camera_pose_rel_dict[image][1] # for the first image, take the values from the temporary model if not is_first:
rotation = rel_rotation @ np.linalg.inv(prev_rotation.T) translation = rel_translation + prev_translation
conditional_block
workplace_preparation.py
2 * (q2 * q3 - q0 * q1) # Third row of the rotation matrix r20 = 2 * (q1 * q3 - q0 * q2) r21 = 2 * (q2 * q3 + q0 * q1) r22 = 2 * (q0 * q0 + q3 * q3) - 1 # 3x3 rotation matrix rot_matrix = np.array([[r00, r01, r02], [r10, r11, r12], [r20, r21, r22]]) return rot_matrix def rotation_matrix_to_quaternion(rotation_matrix: np) -> object: """ The function converts rotation matrix to quaternion vector https://learnopencv.com/rotation-matrix-to-euler-angles/ :param rotation_matrix: rotation matrix 3x3 represented by NumPy array :return quaternion vector: defined by (qx, qy, qz, qw) """ cosine_for_pitch = math.sqrt(rotation_matrix[0][0] ** 2 + rotation_matrix[1][0] ** 2) is_singular = cosine_for_pitch < 10 ** -6 if not is_singular: yaw = math.atan2(rotation_matrix[1][0], rotation_matrix[0][0]) pitch = math.atan2(-rotation_matrix[2][0], cosine_for_pitch) roll = math.atan2(rotation_matrix[2][1], rotation_matrix[2][2]) else: yaw = math.atan2(-rotation_matrix[1][2], rotation_matrix[1][1]) pitch = math.atan2(-rotation_matrix[2][0], cosine_for_pitch) roll = 0 e = (yaw, pitch, roll) return euler_to_quaternion(e) def euler_to_quaternion(euler: tuple) -> object: """ The function convert Euler angle to quaternion object :param Euler: angle represented by yaw, pitch, roll :return quaternion vector: defined by (qx, qy, qz, qw) """ (yaw, pitch, roll) = (euler[0], euler[1], euler[2]) qy = np.sin(roll / 2) * np.cos(pitch / 2) * np.cos(yaw / 2) - np.cos(roll / 2) * np.sin(pitch / 2) * np.sin(yaw / 2) qx = np.cos(roll / 2) * np.sin(pitch / 2) * np.cos(yaw / 2) + np.sin(roll / 2) * np.cos(pitch / 2) * np.sin(yaw / 2) qw = np.cos(roll / 2) * np.cos(pitch / 2) * np.sin(yaw / 2) - np.sin(roll / 2) * np.sin(pitch / 2) * np.cos(yaw / 2) qz = np.cos(roll / 2) * np.cos(pitch / 2) * np.cos(yaw / 2) + np.sin(roll / 2) * np.sin(pitch / 2) * np.sin(yaw / 2) return qx, qy, qz, qw def get_first_image_pose(image_src: str) -> list: """ The function return the absolut R & T for the first image in temp model :param image_src: path to image file (colmap output) :return R&T: R = list[0], T list[1] or None if image1 not exists """ # read images file with open(image_src, 'r') as file: lines = file.readlines()[4::2] # create absolut camera pose dictionary for line in lines: columns = line.split() image_name = columns[9].split('.')[0] image_id = int(image_name.split('e')[1]) # convert and return the camera pose for the first image in model if image_id == 1: qw = float(columns[1]) qx = float(columns[2]) qy = float(columns[4]) qz = float(columns[3]) rotation_matrix = quaternion_to_rotation_matrix(qw, qx, qy, qz) tx = float(columns[5]) ty = float(columns[7]) tz = float(columns[6]) translation_vector = np.array([tx, ty, tz]) return [rotation_matrix, translation_vector] return [] def draw_rel_camera_pose(image: int, origin: list, camera_pose: list, plot_dir_path: str) -> None: """ Debug function for plotting the relative camera poses :param image: number of current image :param origin: list of [x,y,z] of the origin :param camera_pose: list of [x1,y1,z1][x2,y2,z2] of the camera pose (three 2d vectors) :param plot_dir_path: path to plot directory """ fig = plt.figure() ax = fig.add_subplot(111, projection='3d') ax.view_init(elev=10.) ax.set_title('camera pose image: %d' % image) scale = 7 ax.set_xlim3d(-scale, scale) ax.set_ylim3d(-scale, scale) ax.set_zlim3d(-scale, scale) # replace the Y-Axis with Z-Axis ax.scatter(origin[0], origin[2], origin[1], c='black') for i in range(3): ax.plot([origin[0], camera_pose[i][0]], [origin[2], camera_pose[i][2]], [origin[1], camera_pose[i][1]]) i += 1 fig.savefig(f'{plot_dir_path}/%d.png' % image) plt.close(fig) plt.clf() def compute_absolut_camera_pose(camera_pose_rel_dict: dict, first_image_pose: list, workspace_path: str, do_plot=False) -> dict: """ The function return a dictionary with recovered R&T for each image :param camera_pose_rel_dict: dictionary of relative camera poses for each image :param first_image_pose: absolute R&T of the first image :param workspace_path: path to workspace_path :param do_plot: boolean flag for debug purpose :return: camera_pose_recover dictionary """ # create directory for reference plots ref_pose_images_path = path.join(workspace_path, 'ref_images') if do_plot: makedirs(ref_pose_images_path) # initialize parameters for computing absolut camera poses camera_pose_recover = {} rotation = first_image_pose[0] translation = first_image_pose[1] is_first = True prev_rotation = np.identity(3) prev_translation = np.zeros(3) # foreach image compute the absolut pose out of the reference pose for image in camera_pose_rel_dict.keys(): rel_rotation = camera_pose_rel_dict[image][0] rel_translation = camera_pose_rel_dict[image][1] # for the first image, take the values from the temporary model if not is_first: rotation = rel_rotation @ np.linalg.inv(prev_rotation.T) translation = rel_translation + prev_translation # compute the absolut camera pose camera_pose = rotation + translation if do_plot: draw_rel_camera_pose(image, translation, camera_pose, ref_pose_images_path) # save the values foreach image (in R & T format) camera_pose_recover[image] = [rotation, translation] prev_rotation = rotation prev_translation = translation is_first = False return camera_pose_recover def write_camera_pose_to_file(camera_pose_abs_dict: dict, pose_dir_path: str) -> None: """ The function write the recovered camera poses according to COLMAP documentation :param camera_pose_abs_dict: A dictionary of recovered camera poses for each image :param pose_dir_path: path to image file """ image_dst = path.join(pose_dir_path, 'images.txt') with open(image_dst, 'w+') as file: file.write('# Image list with two lines of data per image:\n') file.write('# IMAGE_ID, QW, QX, QY, QZ, TX, TY, TZ, CAMERA_ID, NAME\n') file.write('# POINTS2D[] as (X, Y, POINT3D_ID)\n') file.write(f'# Number of images: {len(camera_pose_abs_dict.keys())}\n') # write each camera pose to file for image in camera_pose_abs_dict.keys(): image_pose_data = [] t_vector = camera_pose_abs_dict[image][1] qx, qy, qz, qw = rotation_matrix_to_quaternion(camera_pose_abs_dict[image][0]) image_pose_data.append(str(image)) # image_pose_data.append(f'{qw} {qx} {qy} {qz}') image_pose_data.append(f'{qz} {qy} {qx} {qw}') image_pose_data.append(' '.join(map(str, t_vector))) image_pose_data.append('1') image_pose_data.append(f'image{image}.jpg') file.write(' '.join(image_pose_data) + '\n\n') def
clear_workspace
identifier_name
workplace_preparation.py
'images') if path.exists(path_to_images) and path.isdir(path_to_images): shutil.rmtree(path_to_images) makedirs(path_to_images) # split the given video into images subprocess.run(['ffmpeg', '-i', temp_video, '-r', str(number_of_images / video_duration), '-f', 'image2', path.join(path_to_images, 'image%d.jpg')]) # remove extra files remove_extra_images(path_to_images, number_of_images) remove(temp_video) def create_temp_model(temp_dir_path: str) -> str:
""" # create temp images folder path_to_temp_model = path.join(temp_dir_path, 'temp_model') path_to_temp_images = path.join(path_to_temp_model, 'temp_images') # remove old temporary folder if exists if path.exists(path_to_temp_model) and path.isdir(path_to_temp_model): shutil.rmtree(path_to_temp_model) number_of_temp_images = 0 path_to_images = path.join(temp_dir_path, 'images') # take only part of the images for the temp model while number_of_temp_images < number_of_images_in_temp_model: try: number_of_temp_images = len([name for name in listdir(path_to_images) if name.endswith('.jpg')]) except FileNotFoundError: number_of_temp_images = 0 # copy subdirectory example shutil.copytree(path_to_images, path_to_temp_images) # run colmap to create model for the first 10 images in video subprocess.run(['colmap', 'automatic_reconstructor', '--workspace_path', path_to_temp_model, '--image_path', path_to_temp_images, '--data_type=video', '--quality=extreme']) return path_to_temp_model def quaternion_to_rotation_matrix(q0, q1, q2, q3) -> np: """ The function converts the quaternion vector to a rotation matrix https://automaticaddison.com/how-to-convert-a-quaternion-to-a-rotation-matrix/ :param q0: the value of qw :param q1: the value of qx :param q2: the value of qy :param q3: the value of qz :return rot_matrix: rotation matrix 3x3 as NumPy array """ # First row of the rotation matrix r00 = 2 * (q0 * q0 + q1 * q1) - 1 r01 = 2 * (q1 * q2 - q0 * q3) r02 = 2 * (q1 * q3 + q0 * q2) # Second row of the rotation matrix r10 = 2 * (q1 * q2 + q0 * q3) r11 = 2 * (q0 * q0 + q2 * q2) - 1 r12 = 2 * (q2 * q3 - q0 * q1) # Third row of the rotation matrix r20 = 2 * (q1 * q3 - q0 * q2) r21 = 2 * (q2 * q3 + q0 * q1) r22 = 2 * (q0 * q0 + q3 * q3) - 1 # 3x3 rotation matrix rot_matrix = np.array([[r00, r01, r02], [r10, r11, r12], [r20, r21, r22]]) return rot_matrix def rotation_matrix_to_quaternion(rotation_matrix: np) -> object: """ The function converts rotation matrix to quaternion vector https://learnopencv.com/rotation-matrix-to-euler-angles/ :param rotation_matrix: rotation matrix 3x3 represented by NumPy array :return quaternion vector: defined by (qx, qy, qz, qw) """ cosine_for_pitch = math.sqrt(rotation_matrix[0][0] ** 2 + rotation_matrix[1][0] ** 2) is_singular = cosine_for_pitch < 10 ** -6 if not is_singular: yaw = math.atan2(rotation_matrix[1][0], rotation_matrix[0][0]) pitch = math.atan2(-rotation_matrix[2][0], cosine_for_pitch) roll = math.atan2(rotation_matrix[2][1], rotation_matrix[2][2]) else: yaw = math.atan2(-rotation_matrix[1][2], rotation_matrix[1][1]) pitch = math.atan2(-rotation_matrix[2][0], cosine_for_pitch) roll = 0 e = (yaw, pitch, roll) return euler_to_quaternion(e) def euler_to_quaternion(euler: tuple) -> object: """ The function convert Euler angle to quaternion object :param Euler: angle represented by yaw, pitch, roll :return quaternion vector: defined by (qx, qy, qz, qw) """ (yaw, pitch, roll) = (euler[0], euler[1], euler[2]) qy = np.sin(roll / 2) * np.cos(pitch / 2) * np.cos(yaw / 2) - np.cos(roll / 2) * np.sin(pitch / 2) * np.sin(yaw / 2) qx = np.cos(roll / 2) * np.sin(pitch / 2) * np.cos(yaw / 2) + np.sin(roll / 2) * np.cos(pitch / 2) * np.sin(yaw / 2) qw = np.cos(roll / 2) * np.cos(pitch / 2) * np.sin(yaw / 2) - np.sin(roll / 2) * np.sin(pitch / 2) * np.cos(yaw / 2) qz = np.cos(roll / 2) * np.cos(pitch / 2) * np.cos(yaw / 2) + np.sin(roll / 2) * np.sin(pitch / 2) * np.sin(yaw / 2) return qx, qy, qz, qw def get_first_image_pose(image_src: str) -> list: """ The function return the absolut R & T for the first image in temp model :param image_src: path to image file (colmap output) :return R&T: R = list[0], T list[1] or None if image1 not exists """ # read images file with open(image_src, 'r') as file: lines = file.readlines()[4::2] # create absolut camera pose dictionary for line in lines: columns = line.split() image_name = columns[9].split('.')[0] image_id = int(image_name.split('e')[1]) # convert and return the camera pose for the first image in model if image_id == 1: qw = float(columns[1]) qx = float(columns[2]) qy = float(columns[4]) qz = float(columns[3]) rotation_matrix = quaternion_to_rotation_matrix(qw, qx, qy, qz) tx = float(columns[5]) ty = float(columns[7]) tz = float(columns[6]) translation_vector = np.array([tx, ty, tz]) return [rotation_matrix, translation_vector] return [] def draw_rel_camera_pose(image: int, origin: list, camera_pose: list, plot_dir_path: str) -> None: """ Debug function for plotting the relative camera poses :param image: number of current image :param origin: list of [x,y,z] of the origin :param camera_pose: list of [x1,y1,z1][x2,y2,z2] of the camera pose (three 2d vectors) :param plot_dir_path: path to plot directory """ fig = plt.figure() ax = fig.add_subplot(111, projection='3d') ax.view_init(elev=10.) ax.set_title('camera pose image: %d' % image) scale = 7 ax.set_xlim3d(-scale, scale) ax.set_ylim3d(-scale, scale) ax.set_zlim3d(-scale, scale) # replace the Y-Axis with Z-Axis ax.scatter(origin[0], origin[2], origin[1], c='black') for i in range(3): ax.plot([origin[0], camera_pose[i][0]], [origin[2], camera_pose[i][2]], [origin[1], camera_pose[i][1]]) i += 1 fig.savefig(f'{plot_dir_path}/%d.png' % image) plt.close(fig) plt.clf() def compute_absolut_camera_pose(camera_pose_rel_dict: dict, first_image_pose: list, workspace_path: str, do_plot=False) -> dict: """ The function return a dictionary with recovered R&T for each image :param camera_pose_rel_dict: dictionary of relative camera poses for each image
""" The function prepares the images for our model based on a given video :param temp_dir_path: video in h264 format :return number_of_images: path to temporary model folder
random_line_split
common.go
errors.New("cannot set invalid user agent") errHTTPClientInvalid = errors.New("custom http client cannot be nil") zeroValueUnix = time.Unix(0, 0) // ErrTypeAssertFailure defines an error when type assertion fails ErrTypeAssertFailure = errors.New("type assert failure") ) // MatchesEmailPattern ensures that the string is an email address by regexp check func MatchesEmailPattern(value string) bool { if len(value) < 3 || len(value) > 254 { return false } return emailRX.MatchString(value) } // SetHTTPClientWithTimeout sets a new *http.Client with different timeout // settings func SetHTTPClientWithTimeout(t time.Duration) error { if t <= 0 { return errCannotSetInvalidTimeout } m.Lock() _HTTPClient = NewHTTPClientWithTimeout(t) m.Unlock() return nil } // SetHTTPUserAgent sets the user agent which will be used for all common HTTP // requests. func SetHTTPUserAgent(agent string) error { if agent == "" { return errUserAgentInvalid } m.Lock() _HTTPUserAgent = agent m.Unlock() return nil } // SetHTTPClient sets a custom HTTP client. func SetHTTPClient(client *http.Client) error { if client == nil { return errHTTPClientInvalid } m.Lock() _HTTPClient = client m.Unlock() return nil } // NewHTTPClientWithTimeout initialises a new HTTP client and its underlying // transport IdleConnTimeout with the specified timeout duration func NewHTTPClientWithTimeout(t time.Duration) *http.Client { tr := &http.Transport{ // Added IdleConnTimeout to reduce the time of idle connections which // could potentially slow macOS reconnection when there is a sudden // network disconnection/issue IdleConnTimeout: t, Proxy: http.ProxyFromEnvironment, } h := &http.Client{ Transport: tr, Timeout: t} return h } // StringSliceDifference concatenates slices together based on its index and // returns an individual string array func StringSliceDifference(slice1, slice2 []string) []string { var diff []string for i := 0; i < 2; i++ { for _, s1 := range slice1 { found := false for _, s2 := range slice2 { if s1 == s2 { found = true break } } if !found { diff = append(diff, s1) } } if i == 0 { slice1, slice2 = slice2, slice1 } } return diff } // StringDataContains checks the substring array with an input and returns a bool func StringDataContains(haystack []string, needle string) bool { data := strings.Join(haystack, ",") return strings.Contains(data, needle) } // StringDataCompare data checks the substring array with an input and returns a bool func StringDataCompare(haystack []string, needle string) bool { for x := range haystack { if haystack[x] == needle { return true } } return false } // StringDataCompareInsensitive data checks the substring array with an input and returns // a bool irrespective of lower or upper case strings func StringDataCompareInsensitive(haystack []string, needle string) bool { for x := range haystack { if strings.EqualFold(haystack[x], needle) { return true } } return false } // StringDataContainsInsensitive checks the substring array with an input and returns // a bool irrespective of lower or upper case strings func StringDataContainsInsensitive(haystack []string, needle string) bool { for _, data := range haystack { if strings.Contains(strings.ToUpper(data), strings.ToUpper(needle)) { return true } } return false } // IsEnabled takes in a boolean param and returns a string if it is enabled // or disabled func IsEnabled(isEnabled bool) string { if isEnabled { return "Enabled" } return "Disabled" } // IsValidCryptoAddress validates your cryptocurrency address string using the // regexp package // Validation issues occurring because "3" is contained in // litecoin and Bitcoin addresses - non-fatal func IsValidCryptoAddress(address, crypto string) (bool, error) {
case "eth": return regexp.MatchString("^0x[a-km-z0-9]{40}$", address) default: return false, fmt.Errorf("%w %s", errInvalidCryptoCurrency, crypto) } } // YesOrNo returns a boolean variable to check if input is "y" or "yes" func YesOrNo(input string) bool { if strings.EqualFold(input, "y") || strings.EqualFold(input, "yes") { return true } return false } // SendHTTPRequest sends a request using the http package and returns the body // contents func SendHTTPRequest(ctx context.Context, method, urlPath string, headers map[string]string, body io.Reader, verbose bool) ([]byte, error) { method = strings.ToUpper(method) if method != http.MethodOptions && method != http.MethodGet && method != http.MethodHead && method != http.MethodPost && method != http.MethodPut && method != http.MethodDelete && method != http.MethodTrace && method != http.MethodConnect { return nil, errors.New("invalid HTTP method specified") } req, err := http.NewRequestWithContext(ctx, method, urlPath, body) if err != nil { return nil, err } for k, v := range headers { req.Header.Add(k, v) } if verbose { log.Debugf(log.Global, "Request path: %s", urlPath) for k, d := range req.Header { log.Debugf(log.Global, "Request header [%s]: %s", k, d) } log.Debugf(log.Global, "Request type: %s", method) if body != nil { log.Debugf(log.Global, "Request body: %v", body) } } m.RLock() if _HTTPUserAgent != "" && req.Header.Get("User-Agent") == "" { req.Header.Add("User-Agent", _HTTPUserAgent) } if _HTTPClient == nil { m.RUnlock() m.Lock() // Set *http.Client with default timeout if not populated. _HTTPClient = NewHTTPClientWithTimeout(defaultTimeout) m.Unlock() m.RLock() } resp, err := _HTTPClient.Do(req) m.RUnlock() if err != nil { return nil, err } defer resp.Body.Close() contents, err := io.ReadAll(resp.Body) if verbose { log.Debugf(log.Global, "HTTP status: %s, Code: %v", resp.Status, resp.StatusCode) log.Debugf(log.Global, "Raw response: %s", string(contents)) } return contents, err } // EncodeURLValues concatenates url values onto a url string and returns a // string func EncodeURLValues(urlPath string, values url.Values) string { u := urlPath if len(values) > 0 { u += "?" + values.Encode() } return u } // ExtractHost returns the hostname out of a string func ExtractHost(address string) string { host, _, _ := net.SplitHostPort(address) if host == "" { return "localhost" } return host } // ExtractPort returns the port name out of a string func ExtractPort(host string) int { _, port, _ := net.SplitHostPort(host) if port == "" { return 80 } portInt, _ := strconv.Atoi(port) return portInt } // GetURIPath returns the path of a URL given a URI func GetURIPath(uri string) string { urip, err := url.Parse(uri) if err != nil { return "" } if urip.RawQuery != "" { return urip.Path + "?" + urip.RawQuery } return urip.Path } // GetExecutablePath returns the executables launch path func GetExecutablePath() (string, error) { ex, err := os.Executable() if err != nil { return "", err } return filepath.Dir(ex), nil } // GetDefaultDataDir returns the default data directory // Windows - C:\Users\%USER%\AppData\Roaming\GoCryptoTrader // Linux/Unix or OSX - $HOME/.gocryptotrader func GetDefaultDataDir(env string) string { if env == "windows" { return filepath.Join(os.Getenv("APPDATA"), "GoCryptoTrader") } usr, err := user.Current() if err == nil { return filepath.Join(usr.HomeDir, ".gocryptotrader") } dir, err := os.UserHomeDir() if err != nil { log.Warnln(log.Global, "Environment variable unset, defaulting to current directory") dir = "." } return filepath.Join(dir, ".gocryptotrader") } // CreateDir creates a directory based on the supplied parameter func CreateDir(dir string)
switch strings.ToLower(crypto) { case "btc": return regexp.MatchString("^(bc1|[13])[a-zA-HJ-NP-Z0-9]{25,90}$", address) case "ltc": return regexp.MatchString("^[L3M][a-km-zA-HJ-NP-Z1-9]{25,34}$", address)
random_line_split
common.go
.New("cannot set invalid user agent") errHTTPClientInvalid = errors.New("custom http client cannot be nil") zeroValueUnix = time.Unix(0, 0) // ErrTypeAssertFailure defines an error when type assertion fails ErrTypeAssertFailure = errors.New("type assert failure") ) // MatchesEmailPattern ensures that the string is an email address by regexp check func MatchesEmailPattern(value string) bool { if len(value) < 3 || len(value) > 254 { return false } return emailRX.MatchString(value) } // SetHTTPClientWithTimeout sets a new *http.Client with different timeout // settings func SetHTTPClientWithTimeout(t time.Duration) error { if t <= 0 { return errCannotSetInvalidTimeout } m.Lock() _HTTPClient = NewHTTPClientWithTimeout(t) m.Unlock() return nil } // SetHTTPUserAgent sets the user agent which will be used for all common HTTP // requests. func SetHTTPUserAgent(agent string) error
// SetHTTPClient sets a custom HTTP client. func SetHTTPClient(client *http.Client) error { if client == nil { return errHTTPClientInvalid } m.Lock() _HTTPClient = client m.Unlock() return nil } // NewHTTPClientWithTimeout initialises a new HTTP client and its underlying // transport IdleConnTimeout with the specified timeout duration func NewHTTPClientWithTimeout(t time.Duration) *http.Client { tr := &http.Transport{ // Added IdleConnTimeout to reduce the time of idle connections which // could potentially slow macOS reconnection when there is a sudden // network disconnection/issue IdleConnTimeout: t, Proxy: http.ProxyFromEnvironment, } h := &http.Client{ Transport: tr, Timeout: t} return h } // StringSliceDifference concatenates slices together based on its index and // returns an individual string array func StringSliceDifference(slice1, slice2 []string) []string { var diff []string for i := 0; i < 2; i++ { for _, s1 := range slice1 { found := false for _, s2 := range slice2 { if s1 == s2 { found = true break } } if !found { diff = append(diff, s1) } } if i == 0 { slice1, slice2 = slice2, slice1 } } return diff } // StringDataContains checks the substring array with an input and returns a bool func StringDataContains(haystack []string, needle string) bool { data := strings.Join(haystack, ",") return strings.Contains(data, needle) } // StringDataCompare data checks the substring array with an input and returns a bool func StringDataCompare(haystack []string, needle string) bool { for x := range haystack { if haystack[x] == needle { return true } } return false } // StringDataCompareInsensitive data checks the substring array with an input and returns // a bool irrespective of lower or upper case strings func StringDataCompareInsensitive(haystack []string, needle string) bool { for x := range haystack { if strings.EqualFold(haystack[x], needle) { return true } } return false } // StringDataContainsInsensitive checks the substring array with an input and returns // a bool irrespective of lower or upper case strings func StringDataContainsInsensitive(haystack []string, needle string) bool { for _, data := range haystack { if strings.Contains(strings.ToUpper(data), strings.ToUpper(needle)) { return true } } return false } // IsEnabled takes in a boolean param and returns a string if it is enabled // or disabled func IsEnabled(isEnabled bool) string { if isEnabled { return "Enabled" } return "Disabled" } // IsValidCryptoAddress validates your cryptocurrency address string using the // regexp package // Validation issues occurring because "3" is contained in // litecoin and Bitcoin addresses - non-fatal func IsValidCryptoAddress(address, crypto string) (bool, error) { switch strings.ToLower(crypto) { case "btc": return regexp.MatchString("^(bc1|[13])[a-zA-HJ-NP-Z0-9]{25,90}$", address) case "ltc": return regexp.MatchString("^[L3M][a-km-zA-HJ-NP-Z1-9]{25,34}$", address) case "eth": return regexp.MatchString("^0x[a-km-z0-9]{40}$", address) default: return false, fmt.Errorf("%w %s", errInvalidCryptoCurrency, crypto) } } // YesOrNo returns a boolean variable to check if input is "y" or "yes" func YesOrNo(input string) bool { if strings.EqualFold(input, "y") || strings.EqualFold(input, "yes") { return true } return false } // SendHTTPRequest sends a request using the http package and returns the body // contents func SendHTTPRequest(ctx context.Context, method, urlPath string, headers map[string]string, body io.Reader, verbose bool) ([]byte, error) { method = strings.ToUpper(method) if method != http.MethodOptions && method != http.MethodGet && method != http.MethodHead && method != http.MethodPost && method != http.MethodPut && method != http.MethodDelete && method != http.MethodTrace && method != http.MethodConnect { return nil, errors.New("invalid HTTP method specified") } req, err := http.NewRequestWithContext(ctx, method, urlPath, body) if err != nil { return nil, err } for k, v := range headers { req.Header.Add(k, v) } if verbose { log.Debugf(log.Global, "Request path: %s", urlPath) for k, d := range req.Header { log.Debugf(log.Global, "Request header [%s]: %s", k, d) } log.Debugf(log.Global, "Request type: %s", method) if body != nil { log.Debugf(log.Global, "Request body: %v", body) } } m.RLock() if _HTTPUserAgent != "" && req.Header.Get("User-Agent") == "" { req.Header.Add("User-Agent", _HTTPUserAgent) } if _HTTPClient == nil { m.RUnlock() m.Lock() // Set *http.Client with default timeout if not populated. _HTTPClient = NewHTTPClientWithTimeout(defaultTimeout) m.Unlock() m.RLock() } resp, err := _HTTPClient.Do(req) m.RUnlock() if err != nil { return nil, err } defer resp.Body.Close() contents, err := io.ReadAll(resp.Body) if verbose { log.Debugf(log.Global, "HTTP status: %s, Code: %v", resp.Status, resp.StatusCode) log.Debugf(log.Global, "Raw response: %s", string(contents)) } return contents, err } // EncodeURLValues concatenates url values onto a url string and returns a // string func EncodeURLValues(urlPath string, values url.Values) string { u := urlPath if len(values) > 0 { u += "?" + values.Encode() } return u } // ExtractHost returns the hostname out of a string func ExtractHost(address string) string { host, _, _ := net.SplitHostPort(address) if host == "" { return "localhost" } return host } // ExtractPort returns the port name out of a string func ExtractPort(host string) int { _, port, _ := net.SplitHostPort(host) if port == "" { return 80 } portInt, _ := strconv.Atoi(port) return portInt } // GetURIPath returns the path of a URL given a URI func GetURIPath(uri string) string { urip, err := url.Parse(uri) if err != nil { return "" } if urip.RawQuery != "" { return urip.Path + "?" + urip.RawQuery } return urip.Path } // GetExecutablePath returns the executables launch path func GetExecutablePath() (string, error) { ex, err := os.Executable() if err != nil { return "", err } return filepath.Dir(ex), nil } // GetDefaultDataDir returns the default data directory // Windows - C:\Users\%USER%\AppData\Roaming\GoCryptoTrader // Linux/Unix or OSX - $HOME/.gocryptotrader func GetDefaultDataDir(env string) string { if env == "windows" { return filepath.Join(os.Getenv("APPDATA"), "GoCryptoTrader") } usr, err := user.Current() if err == nil { return filepath.Join(usr.HomeDir, ".gocryptotrader") } dir, err := os.UserHomeDir() if err != nil { log.Warnln(log.Global, "Environment variable unset, defaulting to current directory") dir = "." } return filepath.Join(dir, ".gocryptotrader") } // CreateDir creates a directory based on the supplied parameter func CreateDir(dir string
{ if agent == "" { return errUserAgentInvalid } m.Lock() _HTTPUserAgent = agent m.Unlock() return nil }
identifier_body
common.go
aystack []string, needle string) bool { for x := range haystack { if strings.EqualFold(haystack[x], needle) { return true } } return false } // StringDataContainsInsensitive checks the substring array with an input and returns // a bool irrespective of lower or upper case strings func StringDataContainsInsensitive(haystack []string, needle string) bool { for _, data := range haystack { if strings.Contains(strings.ToUpper(data), strings.ToUpper(needle)) { return true } } return false } // IsEnabled takes in a boolean param and returns a string if it is enabled // or disabled func IsEnabled(isEnabled bool) string { if isEnabled { return "Enabled" } return "Disabled" } // IsValidCryptoAddress validates your cryptocurrency address string using the // regexp package // Validation issues occurring because "3" is contained in // litecoin and Bitcoin addresses - non-fatal func IsValidCryptoAddress(address, crypto string) (bool, error) { switch strings.ToLower(crypto) { case "btc": return regexp.MatchString("^(bc1|[13])[a-zA-HJ-NP-Z0-9]{25,90}$", address) case "ltc": return regexp.MatchString("^[L3M][a-km-zA-HJ-NP-Z1-9]{25,34}$", address) case "eth": return regexp.MatchString("^0x[a-km-z0-9]{40}$", address) default: return false, fmt.Errorf("%w %s", errInvalidCryptoCurrency, crypto) } } // YesOrNo returns a boolean variable to check if input is "y" or "yes" func YesOrNo(input string) bool { if strings.EqualFold(input, "y") || strings.EqualFold(input, "yes") { return true } return false } // SendHTTPRequest sends a request using the http package and returns the body // contents func SendHTTPRequest(ctx context.Context, method, urlPath string, headers map[string]string, body io.Reader, verbose bool) ([]byte, error) { method = strings.ToUpper(method) if method != http.MethodOptions && method != http.MethodGet && method != http.MethodHead && method != http.MethodPost && method != http.MethodPut && method != http.MethodDelete && method != http.MethodTrace && method != http.MethodConnect { return nil, errors.New("invalid HTTP method specified") } req, err := http.NewRequestWithContext(ctx, method, urlPath, body) if err != nil { return nil, err } for k, v := range headers { req.Header.Add(k, v) } if verbose { log.Debugf(log.Global, "Request path: %s", urlPath) for k, d := range req.Header { log.Debugf(log.Global, "Request header [%s]: %s", k, d) } log.Debugf(log.Global, "Request type: %s", method) if body != nil { log.Debugf(log.Global, "Request body: %v", body) } } m.RLock() if _HTTPUserAgent != "" && req.Header.Get("User-Agent") == "" { req.Header.Add("User-Agent", _HTTPUserAgent) } if _HTTPClient == nil { m.RUnlock() m.Lock() // Set *http.Client with default timeout if not populated. _HTTPClient = NewHTTPClientWithTimeout(defaultTimeout) m.Unlock() m.RLock() } resp, err := _HTTPClient.Do(req) m.RUnlock() if err != nil { return nil, err } defer resp.Body.Close() contents, err := io.ReadAll(resp.Body) if verbose { log.Debugf(log.Global, "HTTP status: %s, Code: %v", resp.Status, resp.StatusCode) log.Debugf(log.Global, "Raw response: %s", string(contents)) } return contents, err } // EncodeURLValues concatenates url values onto a url string and returns a // string func EncodeURLValues(urlPath string, values url.Values) string { u := urlPath if len(values) > 0 { u += "?" + values.Encode() } return u } // ExtractHost returns the hostname out of a string func ExtractHost(address string) string { host, _, _ := net.SplitHostPort(address) if host == "" { return "localhost" } return host } // ExtractPort returns the port name out of a string func ExtractPort(host string) int { _, port, _ := net.SplitHostPort(host) if port == "" { return 80 } portInt, _ := strconv.Atoi(port) return portInt } // GetURIPath returns the path of a URL given a URI func GetURIPath(uri string) string { urip, err := url.Parse(uri) if err != nil { return "" } if urip.RawQuery != "" { return urip.Path + "?" + urip.RawQuery } return urip.Path } // GetExecutablePath returns the executables launch path func GetExecutablePath() (string, error) { ex, err := os.Executable() if err != nil { return "", err } return filepath.Dir(ex), nil } // GetDefaultDataDir returns the default data directory // Windows - C:\Users\%USER%\AppData\Roaming\GoCryptoTrader // Linux/Unix or OSX - $HOME/.gocryptotrader func GetDefaultDataDir(env string) string { if env == "windows" { return filepath.Join(os.Getenv("APPDATA"), "GoCryptoTrader") } usr, err := user.Current() if err == nil { return filepath.Join(usr.HomeDir, ".gocryptotrader") } dir, err := os.UserHomeDir() if err != nil { log.Warnln(log.Global, "Environment variable unset, defaulting to current directory") dir = "." } return filepath.Join(dir, ".gocryptotrader") } // CreateDir creates a directory based on the supplied parameter func CreateDir(dir string) error { _, err := os.Stat(dir) if !os.IsNotExist(err) { return nil } log.Warnf(log.Global, "Directory %s does not exist.. creating.\n", dir) return os.MkdirAll(dir, file.DefaultPermissionOctal) } // ChangePermission lists all the directories and files in an array func ChangePermission(directory string) error { return filepath.Walk(directory, func(path string, info os.FileInfo, err error) error { if err != nil { return err } if info.Mode().Perm() != file.DefaultPermissionOctal { return os.Chmod(path, file.DefaultPermissionOctal) } return nil }) } // SplitStringSliceByLimit splits a slice of strings into slices by input limit and returns a slice of slice of strings func SplitStringSliceByLimit(in []string, limit uint) [][]string { var stringSlice []string sliceSlice := make([][]string, 0, len(in)/int(limit)+1) for len(in) >= int(limit) { stringSlice, in = in[:limit], in[limit:] sliceSlice = append(sliceSlice, stringSlice) } if len(in) > 0 { sliceSlice = append(sliceSlice, in) } return sliceSlice } // AddPaddingOnUpperCase adds padding to a string when detecting an upper case letter. If // there are multiple upper case items like `ThisIsHTTPExample`, it will only // pad between like this `This Is HTTP Example`. func AddPaddingOnUpperCase(s string) string { if s == "" { return "" } var result []string left := 0 for x := 0; x < len(s); x++ { if x == 0 { continue } if unicode.IsUpper(rune(s[x])) { if !unicode.IsUpper(rune(s[x-1])) { result = append(result, s[left:x]) left = x } } else if x > 1 && unicode.IsUpper(rune(s[x-1])) { if s[left:x-1] == "" { continue } result = append(result, s[left:x-1]) left = x - 1 } } result = append(result, s[left:]) return strings.Join(result, " ") } // InArray checks if _val_ belongs to _array_ func InArray(val, array interface{}) (exists bool, index int) { exists = false index = -1 if array == nil { return } switch reflect.TypeOf(array).Kind() { case reflect.Array, reflect.Slice: s := reflect.ValueOf(array) for i := 0; i < s.Len(); i++ { if reflect.DeepEqual(val, s.Index(i).Interface()) { index = i exists = true return } } } return } // multiError holds all the errors as a slice, this is unexported, so it forces // inbuilt error handling. type multiError struct { loadedErrors []error offset *int } // AppendError appends error in a more idiomatic way. This can start out as a // standard error e.g. err := errors.New("random error") // err = AppendError(err, errors.New("another random error")) func
AppendError
identifier_name
common.go
t} return h } // StringSliceDifference concatenates slices together based on its index and // returns an individual string array func StringSliceDifference(slice1, slice2 []string) []string { var diff []string for i := 0; i < 2; i++ { for _, s1 := range slice1 { found := false for _, s2 := range slice2 { if s1 == s2 { found = true break } } if !found { diff = append(diff, s1) } } if i == 0 { slice1, slice2 = slice2, slice1 } } return diff } // StringDataContains checks the substring array with an input and returns a bool func StringDataContains(haystack []string, needle string) bool { data := strings.Join(haystack, ",") return strings.Contains(data, needle) } // StringDataCompare data checks the substring array with an input and returns a bool func StringDataCompare(haystack []string, needle string) bool { for x := range haystack { if haystack[x] == needle { return true } } return false } // StringDataCompareInsensitive data checks the substring array with an input and returns // a bool irrespective of lower or upper case strings func StringDataCompareInsensitive(haystack []string, needle string) bool { for x := range haystack { if strings.EqualFold(haystack[x], needle) { return true } } return false } // StringDataContainsInsensitive checks the substring array with an input and returns // a bool irrespective of lower or upper case strings func StringDataContainsInsensitive(haystack []string, needle string) bool { for _, data := range haystack { if strings.Contains(strings.ToUpper(data), strings.ToUpper(needle)) { return true } } return false } // IsEnabled takes in a boolean param and returns a string if it is enabled // or disabled func IsEnabled(isEnabled bool) string { if isEnabled { return "Enabled" } return "Disabled" } // IsValidCryptoAddress validates your cryptocurrency address string using the // regexp package // Validation issues occurring because "3" is contained in // litecoin and Bitcoin addresses - non-fatal func IsValidCryptoAddress(address, crypto string) (bool, error) { switch strings.ToLower(crypto) { case "btc": return regexp.MatchString("^(bc1|[13])[a-zA-HJ-NP-Z0-9]{25,90}$", address) case "ltc": return regexp.MatchString("^[L3M][a-km-zA-HJ-NP-Z1-9]{25,34}$", address) case "eth": return regexp.MatchString("^0x[a-km-z0-9]{40}$", address) default: return false, fmt.Errorf("%w %s", errInvalidCryptoCurrency, crypto) } } // YesOrNo returns a boolean variable to check if input is "y" or "yes" func YesOrNo(input string) bool { if strings.EqualFold(input, "y") || strings.EqualFold(input, "yes") { return true } return false } // SendHTTPRequest sends a request using the http package and returns the body // contents func SendHTTPRequest(ctx context.Context, method, urlPath string, headers map[string]string, body io.Reader, verbose bool) ([]byte, error) { method = strings.ToUpper(method) if method != http.MethodOptions && method != http.MethodGet && method != http.MethodHead && method != http.MethodPost && method != http.MethodPut && method != http.MethodDelete && method != http.MethodTrace && method != http.MethodConnect { return nil, errors.New("invalid HTTP method specified") } req, err := http.NewRequestWithContext(ctx, method, urlPath, body) if err != nil { return nil, err } for k, v := range headers { req.Header.Add(k, v) } if verbose { log.Debugf(log.Global, "Request path: %s", urlPath) for k, d := range req.Header { log.Debugf(log.Global, "Request header [%s]: %s", k, d) } log.Debugf(log.Global, "Request type: %s", method) if body != nil { log.Debugf(log.Global, "Request body: %v", body) } } m.RLock() if _HTTPUserAgent != "" && req.Header.Get("User-Agent") == "" { req.Header.Add("User-Agent", _HTTPUserAgent) } if _HTTPClient == nil { m.RUnlock() m.Lock() // Set *http.Client with default timeout if not populated. _HTTPClient = NewHTTPClientWithTimeout(defaultTimeout) m.Unlock() m.RLock() } resp, err := _HTTPClient.Do(req) m.RUnlock() if err != nil { return nil, err } defer resp.Body.Close() contents, err := io.ReadAll(resp.Body) if verbose { log.Debugf(log.Global, "HTTP status: %s, Code: %v", resp.Status, resp.StatusCode) log.Debugf(log.Global, "Raw response: %s", string(contents)) } return contents, err } // EncodeURLValues concatenates url values onto a url string and returns a // string func EncodeURLValues(urlPath string, values url.Values) string { u := urlPath if len(values) > 0 { u += "?" + values.Encode() } return u } // ExtractHost returns the hostname out of a string func ExtractHost(address string) string { host, _, _ := net.SplitHostPort(address) if host == "" { return "localhost" } return host } // ExtractPort returns the port name out of a string func ExtractPort(host string) int { _, port, _ := net.SplitHostPort(host) if port == "" { return 80 } portInt, _ := strconv.Atoi(port) return portInt } // GetURIPath returns the path of a URL given a URI func GetURIPath(uri string) string { urip, err := url.Parse(uri) if err != nil { return "" } if urip.RawQuery != "" { return urip.Path + "?" + urip.RawQuery } return urip.Path } // GetExecutablePath returns the executables launch path func GetExecutablePath() (string, error) { ex, err := os.Executable() if err != nil { return "", err } return filepath.Dir(ex), nil } // GetDefaultDataDir returns the default data directory // Windows - C:\Users\%USER%\AppData\Roaming\GoCryptoTrader // Linux/Unix or OSX - $HOME/.gocryptotrader func GetDefaultDataDir(env string) string { if env == "windows" { return filepath.Join(os.Getenv("APPDATA"), "GoCryptoTrader") } usr, err := user.Current() if err == nil { return filepath.Join(usr.HomeDir, ".gocryptotrader") } dir, err := os.UserHomeDir() if err != nil { log.Warnln(log.Global, "Environment variable unset, defaulting to current directory") dir = "." } return filepath.Join(dir, ".gocryptotrader") } // CreateDir creates a directory based on the supplied parameter func CreateDir(dir string) error { _, err := os.Stat(dir) if !os.IsNotExist(err) { return nil } log.Warnf(log.Global, "Directory %s does not exist.. creating.\n", dir) return os.MkdirAll(dir, file.DefaultPermissionOctal) } // ChangePermission lists all the directories and files in an array func ChangePermission(directory string) error { return filepath.Walk(directory, func(path string, info os.FileInfo, err error) error { if err != nil { return err } if info.Mode().Perm() != file.DefaultPermissionOctal { return os.Chmod(path, file.DefaultPermissionOctal) } return nil }) } // SplitStringSliceByLimit splits a slice of strings into slices by input limit and returns a slice of slice of strings func SplitStringSliceByLimit(in []string, limit uint) [][]string { var stringSlice []string sliceSlice := make([][]string, 0, len(in)/int(limit)+1) for len(in) >= int(limit) { stringSlice, in = in[:limit], in[limit:] sliceSlice = append(sliceSlice, stringSlice) } if len(in) > 0 { sliceSlice = append(sliceSlice, in) } return sliceSlice } // AddPaddingOnUpperCase adds padding to a string when detecting an upper case letter. If // there are multiple upper case items like `ThisIsHTTPExample`, it will only // pad between like this `This Is HTTP Example`. func AddPaddingOnUpperCase(s string) string { if s == "" { return "" } var result []string left := 0 for x := 0; x < len(s); x++ { if x == 0 { continue } if unicode.IsUpper(rune(s[x])) { if !unicode.IsUpper(rune(s[x-1]))
{ result = append(result, s[left:x]) left = x }
conditional_block
grafananet.go
GrafanaNetConfig: could not read schemasFile %q: %s", schemasFile, err.Error()) } if aggregationFile != "" { _, err = conf.ReadAggregations(aggregationFile) if err != nil { return GrafanaNetConfig{}, fmt.Errorf("NewGrafanaNetConfig: could not read aggregationFile %q: %s", aggregationFile, err.Error()) } } return GrafanaNetConfig{ Addr: addr, ApiKey: apiKey, SchemasFile: schemasFile, AggregationFile: aggregationFile, BufSize: 1e7, // since a message is typically around 100B this is 1GB FlushMaxNum: 5000, FlushMaxWait: time.Second / 2, Timeout: 10 * time.Second, Concurrency: 100, OrgID: 1, SSLVerify: true, Blocking: false, Spool: false, ErrBackoffMin: 100 * time.Millisecond, ErrBackoffFactor: 1.5, }, nil } type GrafanaNet struct { baseRoute Cfg GrafanaNetConfig schemas persister.WhisperSchemas aggregation conf.Aggregations schemasStr string aggregationStr string addrMetrics string addrSchemas string addrAggregation string dispatch func(chan []byte, []byte, metrics.Gauge, metrics.Counter) in []chan []byte shutdown chan struct{} wg *sync.WaitGroup client *http.Client numErrFlush metrics.Counter numOut metrics.Counter // metrics successfully written to our buffered conn (no flushing yet) numDropBuffFull metrics.Counter // metric drops due to queue full durationTickFlush metrics.Timer // only updated after successful flush durationManuFlush metrics.Timer // only updated after successful flush. not implemented yet tickFlushSize metrics.Histogram // only updated after successful flush manuFlushSize metrics.Histogram // only updated after successful flush. not implemented yet numBuffered metrics.Gauge bufferSize metrics.Gauge } // getGrafanaNetAddr returns the metrics, schemas and aggregation address (URL) for a given config URL // The URL we instruct customers to use is the url to post metrics to, so that one is obvious // but we support posting both to both /graphite/metrics and /metrics , whereas the schemas and // aggregation URL should always get the /graphite prefix. func getGrafanaNetAddr(addr string) (string, string, string) { if strings.HasSuffix(addr, "/") { addr = addr[:len(addr)-1] } if !strings.HasSuffix(addr, "/metrics") { panic("getAddr called on an addr that does not end on /metrics or /metrics/ - this is not supported. Normally NewGrafanaNetConfig would already have validated this") } addrMetrics := addr baseAddr := strings.TrimSuffix(addrMetrics, "/metrics") if strings.HasSuffix(baseAddr, "/graphite") { baseAddr = strings.TrimSuffix(baseAddr, "/graphite") } addrSchemas := baseAddr + "/graphite/config/storageSchema" addrAggregation := baseAddr + "/graphite/config/storageAggregation" return addrMetrics, addrSchemas, addrAggregation } // NewGrafanaNet creates a special route that writes to a grafana.net datastore // We will automatically run the route and the destination func NewGrafanaNet(key string, matcher matcher.Matcher, cfg GrafanaNetConfig) (Route, error) { schemas, err := getSchemas(cfg.SchemasFile) if err != nil { return nil, err } schemasStr := schemas.String() var aggregation conf.Aggregations var aggregationStr string if cfg.AggregationFile != "" { aggregation, err = conf.ReadAggregations(cfg.AggregationFile) if err != nil { return nil, err } aggregationStr = aggregation.String() } cleanAddr := util.AddrToPath(cfg.Addr) r := &GrafanaNet{ baseRoute: baseRoute{"GrafanaNet", sync.Mutex{}, atomic.Value{}, key}, Cfg: cfg, schemas: schemas, schemasStr: schemasStr, aggregation: aggregation, aggregationStr: aggregationStr, in: make([]chan []byte, cfg.Concurrency), shutdown: make(chan struct{}), wg: new(sync.WaitGroup), numErrFlush: stats.Counter("dest=" + cleanAddr + ".unit=Err.type=flush"), numOut: stats.Counter("dest=" + cleanAddr + ".unit=Metric.direction=out"), durationTickFlush: stats.Timer("dest=" + cleanAddr + ".what=durationFlush.type=ticker"), durationManuFlush: stats.Timer("dest=" + cleanAddr + ".what=durationFlush.type=manual"), tickFlushSize: stats.Histogram("dest=" + cleanAddr + ".unit=B.what=FlushSize.type=ticker"), manuFlushSize: stats.Histogram("dest=" + cleanAddr + ".unit=B.what=FlushSize.type=manual"), numBuffered: stats.Gauge("dest=" + cleanAddr + ".unit=Metric.what=numBuffered"), bufferSize: stats.Gauge("dest=" + cleanAddr + ".unit=Metric.what=bufferSize"), numDropBuffFull: stats.Counter("dest=" + cleanAddr + ".unit=Metric.action=drop.reason=queue_full"), } r.addrMetrics, r.addrSchemas, r.addrAggregation = getGrafanaNetAddr(cfg.Addr) r.bufferSize.Update(int64(cfg.BufSize)) if cfg.Blocking
else { r.dispatch = dispatchNonBlocking } r.wg.Add(cfg.Concurrency) for i := 0; i < cfg.Concurrency; i++ { r.in[i] = make(chan []byte, cfg.BufSize/cfg.Concurrency) go r.run(r.in[i]) } r.config.Store(baseConfig{matcher, make([]*dest.Destination, 0)}) // start off with a transport the same as Go's DefaultTransport transport := &http.Transport{ Proxy: http.ProxyFromEnvironment, DialContext: (&net.Dialer{ Timeout: 30 * time.Second, KeepAlive: 30 * time.Second, DualStack: true, }).DialContext, MaxIdleConns: cfg.Concurrency, MaxIdleConnsPerHost: cfg.Concurrency, IdleConnTimeout: 90 * time.Second, TLSHandshakeTimeout: 10 * time.Second, ExpectContinueTimeout: 1 * time.Second, } // disable http 2.0 because there seems to be a compatibility problem between nginx hosts and the golang http2 implementation // which would occasionally result in bogus `400 Bad Request` errors. transport.TLSNextProto = make(map[string]func(authority string, c *tls.Conn) http.RoundTripper) if !cfg.SSLVerify { transport.TLSClientConfig = &tls.Config{ InsecureSkipVerify: true, } } r.client = &http.Client{ Timeout: cfg.Timeout, Transport: transport, } go r.updateSchemas() if cfg.AggregationFile != "" { go r.updateAggregation() } return r, nil } // run manages incoming and outgoing data for a shard func (route *GrafanaNet) run(in chan []byte) { var metrics []*schema.MetricData buffer := new(bytes.Buffer) timer := time.NewTimer(route.Cfg.FlushMaxWait) for { select { case buf := <-in: route.numBuffered.Dec(1) md, err := parseMetric(buf, route.schemas, route.Cfg.OrgID) if err != nil { log.Errorf("RouteGrafanaNet: parseMetric failed: %s. skipping metric", err) continue } md.SetId() metrics = append(metrics, md) if len(metrics) == route.Cfg.FlushMaxNum { metrics = route.retryFlush(metrics, buffer) // reset our timer if !timer.Stop() { <-timer.C } timer.Reset(route.Cfg.FlushMaxWait) } case <-timer.C: timer.Reset(route.Cfg.FlushMaxWait) metrics = route.retryFlush(metrics, buffer) case <-route.shutdown: metrics = route.retryFlush(metrics, buffer) return } } route.wg.Done() } func (route *GrafanaNet) retryFlush(metrics []*schema.MetricData, buffer *bytes.Buffer) []*schema.MetricData { if len(metrics) == 0 { return metrics } mda := schema.MetricDataArray(metrics) data, err := msg.CreateMsg(mda, 0, msg.FormatMetricDataArrayMsgp) if err != nil { panic(err) } route.numOut.Inc(int64(len(metrics))) buffer.Reset() snappyBody := snappy
{ r.dispatch = dispatchBlocking }
conditional_block
grafananet.go
er{ Timeout: 30 * time.Second, KeepAlive: 30 * time.Second, DualStack: true, }).DialContext, MaxIdleConns: cfg.Concurrency, MaxIdleConnsPerHost: cfg.Concurrency, IdleConnTimeout: 90 * time.Second, TLSHandshakeTimeout: 10 * time.Second, ExpectContinueTimeout: 1 * time.Second, } // disable http 2.0 because there seems to be a compatibility problem between nginx hosts and the golang http2 implementation // which would occasionally result in bogus `400 Bad Request` errors. transport.TLSNextProto = make(map[string]func(authority string, c *tls.Conn) http.RoundTripper) if !cfg.SSLVerify { transport.TLSClientConfig = &tls.Config{ InsecureSkipVerify: true, } } r.client = &http.Client{ Timeout: cfg.Timeout, Transport: transport, } go r.updateSchemas() if cfg.AggregationFile != "" { go r.updateAggregation() } return r, nil } // run manages incoming and outgoing data for a shard func (route *GrafanaNet) run(in chan []byte) { var metrics []*schema.MetricData buffer := new(bytes.Buffer) timer := time.NewTimer(route.Cfg.FlushMaxWait) for { select { case buf := <-in: route.numBuffered.Dec(1) md, err := parseMetric(buf, route.schemas, route.Cfg.OrgID) if err != nil { log.Errorf("RouteGrafanaNet: parseMetric failed: %s. skipping metric", err) continue } md.SetId() metrics = append(metrics, md) if len(metrics) == route.Cfg.FlushMaxNum { metrics = route.retryFlush(metrics, buffer) // reset our timer if !timer.Stop() { <-timer.C } timer.Reset(route.Cfg.FlushMaxWait) } case <-timer.C: timer.Reset(route.Cfg.FlushMaxWait) metrics = route.retryFlush(metrics, buffer) case <-route.shutdown: metrics = route.retryFlush(metrics, buffer) return } } route.wg.Done() } func (route *GrafanaNet) retryFlush(metrics []*schema.MetricData, buffer *bytes.Buffer) []*schema.MetricData { if len(metrics) == 0 { return metrics } mda := schema.MetricDataArray(metrics) data, err := msg.CreateMsg(mda, 0, msg.FormatMetricDataArrayMsgp) if err != nil { panic(err) } route.numOut.Inc(int64(len(metrics))) buffer.Reset() snappyBody := snappy.NewWriter(buffer) snappyBody.Write(data) snappyBody.Close() body := buffer.Bytes() req, err := http.NewRequest("POST", route.addrMetrics, bytes.NewReader(body)) if err != nil { panic(err) } req.Header.Add("Authorization", "Bearer "+route.Cfg.ApiKey) req.Header.Add("Content-Type", "rt-metric-binary-snappy") req.Header.Add("User-Agent", UserAgent) req.Header.Add("Carbon-Relay-NG-Instance", Instance) boff := &backoff.Backoff{ Min: route.Cfg.ErrBackoffMin, Max: 30 * time.Second, Factor: route.Cfg.ErrBackoffFactor, Jitter: true, } var dur time.Duration for { dur, err = route.flush(mda, req) if err == nil { break } route.numErrFlush.Inc(1) b := boff.Duration() log.Warnf("GrafanaNet failed to submit data to %s: %s - will try again in %s (this attempt took %s)", route.addrMetrics, err.Error(), b, dur) time.Sleep(b) // re-instantiate body, since the previous .Do() attempt would have Read it all the way req.Body = ioutil.NopCloser(bytes.NewReader(body)) } log.Debugf("GrafanaNet sent metrics in %s -msg size %d", dur, len(metrics)) route.durationTickFlush.Update(dur) route.tickFlushSize.Update(int64(len(metrics))) return metrics[:0] } func (route *GrafanaNet) flush(mda schema.MetricDataArray, req *http.Request) (time.Duration, error) { pre := time.Now() resp, err := route.client.Do(req) dur := time.Since(pre) if err != nil { return dur, err } if resp.StatusCode >= 200 && resp.StatusCode < 300 { bod, err := ioutil.ReadAll(resp.Body) resp.Body.Close() if err != nil { log.Warnf("GrafanaNet remote said %q, but could not read its response: %s", resp.Status, err.Error()) return dur, nil } var mResp MetricsResponse err = json.Unmarshal(bod, &mResp) if err != nil { log.Warnf("GrafanaNet remote returned %q, but could not parse its response: %s", resp.Status, err.Error()) return dur, nil } if mResp.Invalid != 0 { var b strings.Builder fmt.Fprintf(&b, "request contained %d invalid metrics that were dropped (%d valid metrics were published in this request)\n", mResp.Invalid, mResp.Published) for key, vErr := range mResp.ValidationErrors { fmt.Fprintf(&b, " %q : %d metrics. Examples:\n", key, vErr.Count) for _, idx := range vErr.ExampleIds { fmt.Fprintf(&b, " - %#v\n", mda[idx]) } } log.Warn(b.String()) } return dur, nil } buf := make([]byte, 300) n, _ := resp.Body.Read(buf) ioutil.ReadAll(resp.Body) resp.Body.Close() return dur, fmt.Errorf("http %d - %s", resp.StatusCode, buf[:n]) } // Dispatch takes in the requested buf or drops it if blocking mode and queue of the shard is full func (route *GrafanaNet) Dispatch(buf []byte) { // should return as quickly as possible log.Tracef("route %s sending to dest %s: %s", route.key, route.addrMetrics, buf) buf = bytes.TrimSpace(buf) index := bytes.Index(buf, []byte(" ")) if index == -1 { log.Error("RouteGrafanaNet: invalid message") return } key := buf[:index] hasher := fnv.New32a() hasher.Write(key) shard := int(hasher.Sum32() % uint32(route.Cfg.Concurrency)) route.dispatch(route.in[shard], buf, route.numBuffered, route.numDropBuffFull) } func (route *GrafanaNet) Flush() error { //conf := route.config.Load().(Config) // no-op. Flush() is currently not called by anything. return nil } func (route *GrafanaNet) updateSchemas() { route.postConfig(route.addrSchemas, route.schemasStr) for range time.Tick(6 * time.Hour) { route.postConfig(route.addrSchemas, route.schemasStr) } } func (route *GrafanaNet) updateAggregation() { route.postConfig(route.addrAggregation, route.aggregationStr) for range time.Tick(6 * time.Hour) { route.postConfig(route.addrAggregation, route.aggregationStr) } } func (route *GrafanaNet) postConfig(path, cfg string) { boff := &backoff.Backoff{ Min: route.Cfg.ErrBackoffMin, Max: 30 * time.Minute, Factor: route.Cfg.ErrBackoffFactor, Jitter: true, } for { req, err := http.NewRequest("POST", path, strings.NewReader(cfg)) if err != nil { panic(err) } req.Header.Add("Authorization", "Bearer "+route.Cfg.ApiKey) req.Header.Add("User-Agent", UserAgent) req.Header.Add("Carbon-Relay-NG-Instance", Instance) resp, err := route.client.Do(req) if err != nil { log.Warnf("got error for %s: %s", path, err.Error()) time.Sleep(boff.Duration()) continue } boff.Reset() if resp.StatusCode == http.StatusNotFound { // if grafana cloud is not updated yet for this new feature. // we are still done with our work. no need to log anything } else if resp.StatusCode >= 200 && resp.StatusCode < 300 { // it got accepted, we're done. log.Infof("GrafanaNet %s submitted", path)
} else { // if it's neither of the above, let's log it, but make it look not too scary log.Infof("GrafanaNet %s resulted in code %s (should be harmless)", path, resp.Status)
random_line_split
grafananet.go
if err != nil { return GrafanaNetConfig{}, fmt.Errorf("NewGrafanaNetConfig: could not read aggregationFile %q: %s", aggregationFile, err.Error()) } } return GrafanaNetConfig{ Addr: addr, ApiKey: apiKey, SchemasFile: schemasFile, AggregationFile: aggregationFile, BufSize: 1e7, // since a message is typically around 100B this is 1GB FlushMaxNum: 5000, FlushMaxWait: time.Second / 2, Timeout: 10 * time.Second, Concurrency: 100, OrgID: 1, SSLVerify: true, Blocking: false, Spool: false, ErrBackoffMin: 100 * time.Millisecond, ErrBackoffFactor: 1.5, }, nil } type GrafanaNet struct { baseRoute Cfg GrafanaNetConfig schemas persister.WhisperSchemas aggregation conf.Aggregations schemasStr string aggregationStr string addrMetrics string addrSchemas string addrAggregation string dispatch func(chan []byte, []byte, metrics.Gauge, metrics.Counter) in []chan []byte shutdown chan struct{} wg *sync.WaitGroup client *http.Client numErrFlush metrics.Counter numOut metrics.Counter // metrics successfully written to our buffered conn (no flushing yet) numDropBuffFull metrics.Counter // metric drops due to queue full durationTickFlush metrics.Timer // only updated after successful flush durationManuFlush metrics.Timer // only updated after successful flush. not implemented yet tickFlushSize metrics.Histogram // only updated after successful flush manuFlushSize metrics.Histogram // only updated after successful flush. not implemented yet numBuffered metrics.Gauge bufferSize metrics.Gauge } // getGrafanaNetAddr returns the metrics, schemas and aggregation address (URL) for a given config URL // The URL we instruct customers to use is the url to post metrics to, so that one is obvious // but we support posting both to both /graphite/metrics and /metrics , whereas the schemas and // aggregation URL should always get the /graphite prefix. func getGrafanaNetAddr(addr string) (string, string, string) { if strings.HasSuffix(addr, "/") { addr = addr[:len(addr)-1] } if !strings.HasSuffix(addr, "/metrics") { panic("getAddr called on an addr that does not end on /metrics or /metrics/ - this is not supported. Normally NewGrafanaNetConfig would already have validated this") } addrMetrics := addr baseAddr := strings.TrimSuffix(addrMetrics, "/metrics") if strings.HasSuffix(baseAddr, "/graphite") { baseAddr = strings.TrimSuffix(baseAddr, "/graphite") } addrSchemas := baseAddr + "/graphite/config/storageSchema" addrAggregation := baseAddr + "/graphite/config/storageAggregation" return addrMetrics, addrSchemas, addrAggregation } // NewGrafanaNet creates a special route that writes to a grafana.net datastore // We will automatically run the route and the destination func NewGrafanaNet(key string, matcher matcher.Matcher, cfg GrafanaNetConfig) (Route, error) { schemas, err := getSchemas(cfg.SchemasFile) if err != nil { return nil, err } schemasStr := schemas.String() var aggregation conf.Aggregations var aggregationStr string if cfg.AggregationFile != "" { aggregation, err = conf.ReadAggregations(cfg.AggregationFile) if err != nil { return nil, err } aggregationStr = aggregation.String() } cleanAddr := util.AddrToPath(cfg.Addr) r := &GrafanaNet{ baseRoute: baseRoute{"GrafanaNet", sync.Mutex{}, atomic.Value{}, key}, Cfg: cfg, schemas: schemas, schemasStr: schemasStr, aggregation: aggregation, aggregationStr: aggregationStr, in: make([]chan []byte, cfg.Concurrency), shutdown: make(chan struct{}), wg: new(sync.WaitGroup), numErrFlush: stats.Counter("dest=" + cleanAddr + ".unit=Err.type=flush"), numOut: stats.Counter("dest=" + cleanAddr + ".unit=Metric.direction=out"), durationTickFlush: stats.Timer("dest=" + cleanAddr + ".what=durationFlush.type=ticker"), durationManuFlush: stats.Timer("dest=" + cleanAddr + ".what=durationFlush.type=manual"), tickFlushSize: stats.Histogram("dest=" + cleanAddr + ".unit=B.what=FlushSize.type=ticker"), manuFlushSize: stats.Histogram("dest=" + cleanAddr + ".unit=B.what=FlushSize.type=manual"), numBuffered: stats.Gauge("dest=" + cleanAddr + ".unit=Metric.what=numBuffered"), bufferSize: stats.Gauge("dest=" + cleanAddr + ".unit=Metric.what=bufferSize"), numDropBuffFull: stats.Counter("dest=" + cleanAddr + ".unit=Metric.action=drop.reason=queue_full"), } r.addrMetrics, r.addrSchemas, r.addrAggregation = getGrafanaNetAddr(cfg.Addr) r.bufferSize.Update(int64(cfg.BufSize)) if cfg.Blocking { r.dispatch = dispatchBlocking } else { r.dispatch = dispatchNonBlocking } r.wg.Add(cfg.Concurrency) for i := 0; i < cfg.Concurrency; i++ { r.in[i] = make(chan []byte, cfg.BufSize/cfg.Concurrency) go r.run(r.in[i]) } r.config.Store(baseConfig{matcher, make([]*dest.Destination, 0)}) // start off with a transport the same as Go's DefaultTransport transport := &http.Transport{ Proxy: http.ProxyFromEnvironment, DialContext: (&net.Dialer{ Timeout: 30 * time.Second, KeepAlive: 30 * time.Second, DualStack: true, }).DialContext, MaxIdleConns: cfg.Concurrency, MaxIdleConnsPerHost: cfg.Concurrency, IdleConnTimeout: 90 * time.Second, TLSHandshakeTimeout: 10 * time.Second, ExpectContinueTimeout: 1 * time.Second, } // disable http 2.0 because there seems to be a compatibility problem between nginx hosts and the golang http2 implementation // which would occasionally result in bogus `400 Bad Request` errors. transport.TLSNextProto = make(map[string]func(authority string, c *tls.Conn) http.RoundTripper) if !cfg.SSLVerify { transport.TLSClientConfig = &tls.Config{ InsecureSkipVerify: true, } } r.client = &http.Client{ Timeout: cfg.Timeout, Transport: transport, } go r.updateSchemas() if cfg.AggregationFile != "" { go r.updateAggregation() } return r, nil } // run manages incoming and outgoing data for a shard func (route *GrafanaNet) run(in chan []byte) { var metrics []*schema.MetricData buffer := new(bytes.Buffer) timer := time.NewTimer(route.Cfg.FlushMaxWait) for { select { case buf := <-in: route.numBuffered.Dec(1) md, err := parseMetric(buf, route.schemas, route.Cfg.OrgID) if err != nil { log.Errorf("RouteGrafanaNet: parseMetric failed: %s. skipping metric", err) continue } md.SetId() metrics = append(metrics, md) if len(metrics) == route.Cfg.FlushMaxNum { metrics = route.retryFlush(metrics, buffer) // reset our timer if !timer.Stop() { <-timer.C } timer.Reset(route.Cfg.FlushMaxWait) } case <-timer.C: timer.Reset(route.Cfg.FlushMaxWait) metrics = route.retryFlush(metrics, buffer) case <-route.shutdown: metrics = route.retryFlush(metrics, buffer) return } } route.wg.Done() } func (route *GrafanaNet) retryFlush(metrics []*schema.MetricData, buffer *bytes.Buffer) []*schema.MetricData
{ if len(metrics) == 0 { return metrics } mda := schema.MetricDataArray(metrics) data, err := msg.CreateMsg(mda, 0, msg.FormatMetricDataArrayMsgp) if err != nil { panic(err) } route.numOut.Inc(int64(len(metrics))) buffer.Reset() snappyBody := snappy.NewWriter(buffer) snappyBody.Write(data) snappyBody.Close() body := buffer.Bytes() req, err := http.NewRequest("POST", route.addrMetrics, bytes.NewReader(body)) if err != nil { panic(err)
identifier_body
grafananet.go
GrafanaNetConfig: could not read schemasFile %q: %s", schemasFile, err.Error()) } if aggregationFile != "" { _, err = conf.ReadAggregations(aggregationFile) if err != nil { return GrafanaNetConfig{}, fmt.Errorf("NewGrafanaNetConfig: could not read aggregationFile %q: %s", aggregationFile, err.Error()) } } return GrafanaNetConfig{ Addr: addr, ApiKey: apiKey, SchemasFile: schemasFile, AggregationFile: aggregationFile, BufSize: 1e7, // since a message is typically around 100B this is 1GB FlushMaxNum: 5000, FlushMaxWait: time.Second / 2, Timeout: 10 * time.Second, Concurrency: 100, OrgID: 1, SSLVerify: true, Blocking: false, Spool: false, ErrBackoffMin: 100 * time.Millisecond, ErrBackoffFactor: 1.5, }, nil } type GrafanaNet struct { baseRoute Cfg GrafanaNetConfig schemas persister.WhisperSchemas aggregation conf.Aggregations schemasStr string aggregationStr string addrMetrics string addrSchemas string addrAggregation string dispatch func(chan []byte, []byte, metrics.Gauge, metrics.Counter) in []chan []byte shutdown chan struct{} wg *sync.WaitGroup client *http.Client numErrFlush metrics.Counter numOut metrics.Counter // metrics successfully written to our buffered conn (no flushing yet) numDropBuffFull metrics.Counter // metric drops due to queue full durationTickFlush metrics.Timer // only updated after successful flush durationManuFlush metrics.Timer // only updated after successful flush. not implemented yet tickFlushSize metrics.Histogram // only updated after successful flush manuFlushSize metrics.Histogram // only updated after successful flush. not implemented yet numBuffered metrics.Gauge bufferSize metrics.Gauge } // getGrafanaNetAddr returns the metrics, schemas and aggregation address (URL) for a given config URL // The URL we instruct customers to use is the url to post metrics to, so that one is obvious // but we support posting both to both /graphite/metrics and /metrics , whereas the schemas and // aggregation URL should always get the /graphite prefix. func getGrafanaNetAddr(addr string) (string, string, string) { if strings.HasSuffix(addr, "/") { addr = addr[:len(addr)-1] } if !strings.HasSuffix(addr, "/metrics") { panic("getAddr called on an addr that does not end on /metrics or /metrics/ - this is not supported. Normally NewGrafanaNetConfig would already have validated this") } addrMetrics := addr baseAddr := strings.TrimSuffix(addrMetrics, "/metrics") if strings.HasSuffix(baseAddr, "/graphite") { baseAddr = strings.TrimSuffix(baseAddr, "/graphite") } addrSchemas := baseAddr + "/graphite/config/storageSchema" addrAggregation := baseAddr + "/graphite/config/storageAggregation" return addrMetrics, addrSchemas, addrAggregation } // NewGrafanaNet creates a special route that writes to a grafana.net datastore // We will automatically run the route and the destination func NewGrafanaNet(key string, matcher matcher.Matcher, cfg GrafanaNetConfig) (Route, error) { schemas, err := getSchemas(cfg.SchemasFile) if err != nil { return nil, err } schemasStr := schemas.String() var aggregation conf.Aggregations var aggregationStr string if cfg.AggregationFile != "" { aggregation, err = conf.ReadAggregations(cfg.AggregationFile) if err != nil { return nil, err } aggregationStr = aggregation.String() } cleanAddr := util.AddrToPath(cfg.Addr) r := &GrafanaNet{ baseRoute: baseRoute{"GrafanaNet", sync.Mutex{}, atomic.Value{}, key}, Cfg: cfg, schemas: schemas, schemasStr: schemasStr, aggregation: aggregation, aggregationStr: aggregationStr, in: make([]chan []byte, cfg.Concurrency), shutdown: make(chan struct{}), wg: new(sync.WaitGroup), numErrFlush: stats.Counter("dest=" + cleanAddr + ".unit=Err.type=flush"), numOut: stats.Counter("dest=" + cleanAddr + ".unit=Metric.direction=out"), durationTickFlush: stats.Timer("dest=" + cleanAddr + ".what=durationFlush.type=ticker"), durationManuFlush: stats.Timer("dest=" + cleanAddr + ".what=durationFlush.type=manual"), tickFlushSize: stats.Histogram("dest=" + cleanAddr + ".unit=B.what=FlushSize.type=ticker"), manuFlushSize: stats.Histogram("dest=" + cleanAddr + ".unit=B.what=FlushSize.type=manual"), numBuffered: stats.Gauge("dest=" + cleanAddr + ".unit=Metric.what=numBuffered"), bufferSize: stats.Gauge("dest=" + cleanAddr + ".unit=Metric.what=bufferSize"), numDropBuffFull: stats.Counter("dest=" + cleanAddr + ".unit=Metric.action=drop.reason=queue_full"), } r.addrMetrics, r.addrSchemas, r.addrAggregation = getGrafanaNetAddr(cfg.Addr) r.bufferSize.Update(int64(cfg.BufSize)) if cfg.Blocking { r.dispatch = dispatchBlocking } else { r.dispatch = dispatchNonBlocking } r.wg.Add(cfg.Concurrency) for i := 0; i < cfg.Concurrency; i++ { r.in[i] = make(chan []byte, cfg.BufSize/cfg.Concurrency) go r.run(r.in[i]) } r.config.Store(baseConfig{matcher, make([]*dest.Destination, 0)}) // start off with a transport the same as Go's DefaultTransport transport := &http.Transport{ Proxy: http.ProxyFromEnvironment, DialContext: (&net.Dialer{ Timeout: 30 * time.Second, KeepAlive: 30 * time.Second, DualStack: true, }).DialContext, MaxIdleConns: cfg.Concurrency, MaxIdleConnsPerHost: cfg.Concurrency, IdleConnTimeout: 90 * time.Second, TLSHandshakeTimeout: 10 * time.Second, ExpectContinueTimeout: 1 * time.Second, } // disable http 2.0 because there seems to be a compatibility problem between nginx hosts and the golang http2 implementation // which would occasionally result in bogus `400 Bad Request` errors. transport.TLSNextProto = make(map[string]func(authority string, c *tls.Conn) http.RoundTripper) if !cfg.SSLVerify { transport.TLSClientConfig = &tls.Config{ InsecureSkipVerify: true, } } r.client = &http.Client{ Timeout: cfg.Timeout, Transport: transport, } go r.updateSchemas() if cfg.AggregationFile != "" { go r.updateAggregation() } return r, nil } // run manages incoming and outgoing data for a shard func (route *GrafanaNet) run(in chan []byte) { var metrics []*schema.MetricData buffer := new(bytes.Buffer) timer := time.NewTimer(route.Cfg.FlushMaxWait) for { select { case buf := <-in: route.numBuffered.Dec(1) md, err := parseMetric(buf, route.schemas, route.Cfg.OrgID) if err != nil { log.Errorf("RouteGrafanaNet: parseMetric failed: %s. skipping metric", err) continue } md.SetId() metrics = append(metrics, md) if len(metrics) == route.Cfg.FlushMaxNum { metrics = route.retryFlush(metrics, buffer) // reset our timer if !timer.Stop() { <-timer.C } timer.Reset(route.Cfg.FlushMaxWait) } case <-timer.C: timer.Reset(route.Cfg.FlushMaxWait) metrics = route.retryFlush(metrics, buffer) case <-route.shutdown: metrics = route.retryFlush(metrics, buffer) return } } route.wg.Done() } func (route *GrafanaNet)
(metrics []*schema.MetricData, buffer *bytes.Buffer) []*schema.MetricData { if len(metrics) == 0 { return metrics } mda := schema.MetricDataArray(metrics) data, err := msg.CreateMsg(mda, 0, msg.FormatMetricDataArrayMsgp) if err != nil { panic(err) } route.numOut.Inc(int64(len(metrics))) buffer.Reset() snappyBody := snappy
retryFlush
identifier_name
process.go
return allResults[0], nil } // GetProcesses lists all processes given some filter criteria. func (context Context) GetProcesses( filters F, limit int, cursor string, sortBy string, order string, ) (result []types.Process, cm types.CollectionMetadata, err error) { c := context.Session.DB(context.DBName).C("processes") // Validate the parameters. if sortBy != "" && sortBy != "id" && sortBy != "process-id" && sortBy != "host-id" && sortBy != "host-address" && sortBy != "start-time" && sortBy != "type" && sortBy != "resource" && sortBy != "status" { err = errors.Wrapf(ErrBadInput, "cannot sort by \"%s\"", sortBy) return } if order != "" && order != "asc" && order != "desc" { err = errors.Wrapf(ErrBadInput, "order can be either \"asc\" or \"desc\", not \"%s\"", order) return } if order == "" { order = "asc" } // We currently don't limit access to this collection. Everyone can see it. // Build a query given the parameters. query := bson.M{} for k, v := range filters { switch k { case "id": setDefault(&query, "_id", bson.M{}) query["_id"].(bson.M)["$in"] = v.([]bson.ObjectId) case "process-id": setDefault(&query, k, bson.M{}) query[k].(bson.M)["$eq"] = v.(uint64) case "host-id", "host-address", "type", "resource", "status": setDefault(&query, k, bson.M{}) query[k].(bson.M)["$eq"] = v.(string) default: err = errors.Wrap(ErrBadInput, "invalid value of argument filters") return } } // We count the result size given the filters. This is before pagination. var resultSize int resultSize, err = c.Find(query).Count() if err != nil { err = errors.Wrap(err, "mongo find failed") return } // If a cursor was specified then we have to do a range query. if cursor != "" { comparer := "$gt" if order == "desc" { comparer = "$lt" } // If there is no sorting then the cursor only points to the _id field. if sortBy != "" && sortBy != "id" { splits := strings.Split(cursor, "-") cursor = splits[1] var decoded []byte decoded, err = hex.DecodeString(splits[0]) if err != nil { err = errors.Wrap(err, "hex decode string failed") return } var otherCursor interface{} switch sortBy { case "host-id", "host-address", "type", "resource", "status": otherCursor = string(decoded) case "process-id": otherCursor = binary.BigEndian.Uint64(decoded) case "start-time": var t time.Time t.GobDecode(decoded) otherCursor = t } setDefault(&query, "$or", bson.M{}) query["$or"] = []bson.M{ bson.M{sortBy: bson.M{comparer: otherCursor}}, bson.M{sortBy: bson.M{"$eq": otherCursor}, "_id": bson.M{comparer: bson.ObjectIdHex(cursor)}}, } } else { if bson.IsObjectIdHex(cursor) == false { err = errors.Wrap(ErrBadInput, "invalid cursor") return } setDefault(&query, "_id", bson.M{}) query["_id"].(bson.M)[comparer] = bson.ObjectIdHex(cursor) } } // Execute the query. q := c.Find(query) // We always sort by _id, but we may also sort by a specific field. if sortBy == "" || sortBy == "id" { if order == "asc" { q = q.Sort("_id") } else { q = q.Sort("-_id") } } else { if order == "asc" { q = q.Sort(sortBy, "_id") } else { q = q.Sort("-"+sortBy, "-_id") } } if limit > 0 { q = q.Limit(limit) } // Collect the results. var allResults []types.Process err = q.All(&allResults) if err != nil { err = errors.Wrap(err, "mongo find failed") return } // Compute the next cursor. nextCursor := "" if limit > 0 && len(allResults) == limit { lastResult := allResults[len(allResults)-1] nextCursor = lastResult.ID.Hex() if sortBy != "" { var encoded string var b []byte switch sortBy { case "id": b = []byte(lastResult.ID) case "process-id": b = make([]byte, 4) binary.BigEndian.PutUint64(b, lastResult.ProcessID) case "host-id": b = []byte(lastResult.HostID) case "host-address": b = []byte(lastResult.HostAddress) case "start-time": b, err = lastResult.StartTime.GobEncode() case "type": b = []byte(lastResult.Type) case "resource": b = []byte(lastResult.Resource) case "status": b = []byte(lastResult.Status) } encoded = hex.EncodeToString(b) nextCursor = encoded + "-" + nextCursor } } // Assemble the results. result = allResults cm = types.CollectionMetadata{ TotalResultSize: resultSize, ReturnedResultSize: len(result), NextPageCursor: nextCursor, } return } // CountProcesses is the same as GetProcesses but returns only the count, not the actual processes. func (context Context) CountProcesses(filters F) (count int, err error) { c := context.Session.DB(context.DBName).C("processes") // Build a query given the parameters. query := bson.M{} for k, v := range filters { switch k { case "id": setDefault(&query, "_id", bson.M{}) query["_id"].(bson.M)["$in"] = v.([]bson.ObjectId) case "process-id": setDefault(&query, k, bson.M{}) query[k].(bson.M)["$eq"] = v.(uint64) case "host-id", "host-address", "type", "resource", "status": setDefault(&query, k, bson.M{}) query[k].(bson.M)["$eq"] = v.(string) default: err = errors.Wrap(ErrBadInput, "invalid value of argument filters") return } } // We count the number of tasks that satisfy the filter criteria. count, err = c.Find(query).Count() if err != nil { err = errors.Wrap(err, "mongo find failed") } return } // CreateProcess adds a given process to the database. func (context Context) CreateProcess(proc types.Process) (result types.Process, err error) { // This action is only permitted for the root user. if context.User.IsRoot() == false { err = types.ErrUnauthorized return } if proc.Type != types.ProcController && proc.Type != types.ProcWorker && proc.Type != types.ProcScheduler { err = errors.Wrapf(ErrBadInput, "value of type can be \"%s\", \"%s\" or \"%s\", but found \"%s\"", types.ProcController, types.ProcWorker, types.ProcScheduler, proc.Type) return } // Find the first candidate ordinal. We do this optimistically assuming there is no race conditions. proc.RunningOrinal, err = context.findCandidateOrdinal(proc.Type) if err != nil { err = errors.Wrap(err, "find candidate ordinal failed") return } // Give default values to some fields. proc.ID = bson.NewObjectId() proc.Status = types.ProcIdle proc.StartTime = time.Now() c := context.Session.DB(context.DBName).C("processes") err = c.Insert(proc) if err != nil { lastError := err.(*mgo.LastError) if lastError.Code == 11000 { err = types.ErrIdentifierTaken return } err = errors.Wrap(err, "mongo insert failed") return } // Check if the ordinal was accepted. If not, then there was a race condition so we need to handle it. var accepted bool accepted, err = context.isOrdinalAccepted(proc.Type, proc.ID, proc.RunningOrinal) if err != nil { err = errors.Wrap(err, "is ordinal accepted failed") return } for accepted == false { // As long as the ordinal isn't accepted we need to try again. proc.RunningOrinal, err = context.findCandidateOrdinal(proc.Type) if err != nil { err = errors.Wrap(err, "find candidate ordinal failed")
{ err = ErrNotFound return }
conditional_block
process.go
nextCursor := "" if limit > 0 && len(allResults) == limit { lastResult := allResults[len(allResults)-1] nextCursor = lastResult.ID.Hex() if sortBy != "" { var encoded string var b []byte switch sortBy { case "id": b = []byte(lastResult.ID) case "process-id": b = make([]byte, 4) binary.BigEndian.PutUint64(b, lastResult.ProcessID) case "host-id": b = []byte(lastResult.HostID) case "host-address": b = []byte(lastResult.HostAddress) case "start-time": b, err = lastResult.StartTime.GobEncode() case "type": b = []byte(lastResult.Type) case "resource": b = []byte(lastResult.Resource) case "status": b = []byte(lastResult.Status) } encoded = hex.EncodeToString(b) nextCursor = encoded + "-" + nextCursor } } // Assemble the results. result = allResults cm = types.CollectionMetadata{ TotalResultSize: resultSize, ReturnedResultSize: len(result), NextPageCursor: nextCursor, } return } // CountProcesses is the same as GetProcesses but returns only the count, not the actual processes. func (context Context) CountProcesses(filters F) (count int, err error) { c := context.Session.DB(context.DBName).C("processes") // Build a query given the parameters. query := bson.M{} for k, v := range filters { switch k { case "id": setDefault(&query, "_id", bson.M{}) query["_id"].(bson.M)["$in"] = v.([]bson.ObjectId) case "process-id": setDefault(&query, k, bson.M{}) query[k].(bson.M)["$eq"] = v.(uint64) case "host-id", "host-address", "type", "resource", "status": setDefault(&query, k, bson.M{}) query[k].(bson.M)["$eq"] = v.(string) default: err = errors.Wrap(ErrBadInput, "invalid value of argument filters") return } } // We count the number of tasks that satisfy the filter criteria. count, err = c.Find(query).Count() if err != nil { err = errors.Wrap(err, "mongo find failed") } return } // CreateProcess adds a given process to the database. func (context Context) CreateProcess(proc types.Process) (result types.Process, err error) { // This action is only permitted for the root user. if context.User.IsRoot() == false { err = types.ErrUnauthorized return } if proc.Type != types.ProcController && proc.Type != types.ProcWorker && proc.Type != types.ProcScheduler { err = errors.Wrapf(ErrBadInput, "value of type can be \"%s\", \"%s\" or \"%s\", but found \"%s\"", types.ProcController, types.ProcWorker, types.ProcScheduler, proc.Type) return } // Find the first candidate ordinal. We do this optimistically assuming there is no race conditions. proc.RunningOrinal, err = context.findCandidateOrdinal(proc.Type) if err != nil { err = errors.Wrap(err, "find candidate ordinal failed") return } // Give default values to some fields. proc.ID = bson.NewObjectId() proc.Status = types.ProcIdle proc.StartTime = time.Now() c := context.Session.DB(context.DBName).C("processes") err = c.Insert(proc) if err != nil { lastError := err.(*mgo.LastError) if lastError.Code == 11000 { err = types.ErrIdentifierTaken return } err = errors.Wrap(err, "mongo insert failed") return } // Check if the ordinal was accepted. If not, then there was a race condition so we need to handle it. var accepted bool accepted, err = context.isOrdinalAccepted(proc.Type, proc.ID, proc.RunningOrinal) if err != nil { err = errors.Wrap(err, "is ordinal accepted failed") return } for accepted == false { // As long as the ordinal isn't accepted we need to try again. proc.RunningOrinal, err = context.findCandidateOrdinal(proc.Type) if err != nil { err = errors.Wrap(err, "find candidate ordinal failed") return } // Update the process with the new ordinal. c := context.Session.DB(context.DBName).C("processes") err = c.Update(bson.M{"_id": proc.ID}, bson.M{"$set": bson.M{"running-ordinal": proc.RunningOrinal}}) if err != nil { err = errors.Wrap(err, "mongo update failed") return } // Check if the ordinal was accepted. accepted, err = context.isOrdinalAccepted(proc.Type, proc.ID, proc.RunningOrinal) if err != nil { err = errors.Wrap(err, "is ordinal accepted failed") return } } return proc, nil } func (context Context) findCandidateOrdinal(processType string) (ordinal int, err error) { c := context.Session.DB(context.DBName).C("processes") query := bson.M{"type": processType, "status": bson.M{"$ne": types.ProcTerminated}} var ordinals []struct { Ordinal int `bson:"running-ordinal"` } q := c.Find(query).Select(bson.M{"running-ordinal": 1}).Sort("running-ordinal") err = q.All(&ordinals) if err != nil { return -1, err } ordinal = 1 for i := range ordinals { if ordinal == ordinals[i].Ordinal { ordinal++ } } return } func (context Context) isOrdinalAccepted(processType string, processID bson.ObjectId, ordinal int) (result bool, err error) { c := context.Session.DB(context.DBName).C("processes") query := bson.M{"type": processType, "status": bson.M{"$ne": types.ProcTerminated}, "running-ordinal": ordinal} var ordinals []struct { ObjectID bson.ObjectId `bson:"_id"` Ordinal int `bson:"running-ordinal"` } q := c.Find(query).Select(bson.M{"_id": 1, "running-ordinal": 1}) err = q.All(&ordinals) if err != nil { return false, err } result = true for i := range ordinals { if processID > ordinals[i].ObjectID { result = false break } } return } // UpdateProcess updates the information about a given process. func (context Context) UpdateProcess(id bson.ObjectId, updates map[string]interface{}) (result types.Process, err error) { // This action is only permitted for the root user. if context.User.IsRoot() == false { err = types.ErrUnauthorized return } // Build the update document. Validate values. valueUpdates := bson.M{} for k, v := range updates { switch k { case "status": status := v.(string) if status != types.ProcIdle && status != types.ProcWorking && status != types.ProcTerminated { err = errors.Wrapf(ErrBadInput, "value of status can be \"%s\", \"%s\" or \"%s\", but found \"%s\"", types.ProcIdle, types.ProcWorking, types.ProcTerminated, status) return } valueUpdates["status"] = status case "last-keepalive": valueUpdates["last-keepalive"] = v.(time.Time) default: err = errors.Wrap(ErrBadInput, "invalid value of parameter updates") return } } // If there were no updates, then we can skip this step. if len(valueUpdates) > 0 { c := context.Session.DB(context.DBName).C("processes") err = c.Update(bson.M{"_id": id}, bson.M{"$set": valueUpdates}) if err != nil { err = errors.Wrap(err, "mongo update failed") return } } // Get the updated process and update cache if needed. result, err = context.GetProcessByID(id) if err != nil { err = errors.Wrap(err, "process get by ID failed") return } return } // StartProcess starts a process of a given type and initializes all other fields automatically. func (context Context) StartProcess(processType string) (result types.Process, err error)
{ if processType != types.ProcController && processType != types.ProcScheduler && processType != types.ProcWorker { panic("invalid processType") } var hostID string hostID, err = os.Hostname() if err != nil { err = errors.Wrap(err, "get hostname from os failed") return } var hostAddress string hostAddress, err = getOutboundIP() if err != nil { //err = errors.Wrap(err, "get outbound ip failed") //return hostAddress = "localhost" }
identifier_body
process.go
", "type", "resource", "status": setDefault(&query, k, bson.M{}) query[k].(bson.M)["$eq"] = v.(string) default: err = errors.Wrap(ErrBadInput, "invalid value of argument filters") return } } // We count the result size given the filters. This is before pagination. var resultSize int resultSize, err = c.Find(query).Count() if err != nil { err = errors.Wrap(err, "mongo find failed") return }
// If a cursor was specified then we have to do a range query. if cursor != "" { comparer := "$gt" if order == "desc" { comparer = "$lt" } // If there is no sorting then the cursor only points to the _id field. if sortBy != "" && sortBy != "id" { splits := strings.Split(cursor, "-") cursor = splits[1] var decoded []byte decoded, err = hex.DecodeString(splits[0]) if err != nil { err = errors.Wrap(err, "hex decode string failed") return } var otherCursor interface{} switch sortBy { case "host-id", "host-address", "type", "resource", "status": otherCursor = string(decoded) case "process-id": otherCursor = binary.BigEndian.Uint64(decoded) case "start-time": var t time.Time t.GobDecode(decoded) otherCursor = t } setDefault(&query, "$or", bson.M{}) query["$or"] = []bson.M{ bson.M{sortBy: bson.M{comparer: otherCursor}}, bson.M{sortBy: bson.M{"$eq": otherCursor}, "_id": bson.M{comparer: bson.ObjectIdHex(cursor)}}, } } else { if bson.IsObjectIdHex(cursor) == false { err = errors.Wrap(ErrBadInput, "invalid cursor") return } setDefault(&query, "_id", bson.M{}) query["_id"].(bson.M)[comparer] = bson.ObjectIdHex(cursor) } } // Execute the query. q := c.Find(query) // We always sort by _id, but we may also sort by a specific field. if sortBy == "" || sortBy == "id" { if order == "asc" { q = q.Sort("_id") } else { q = q.Sort("-_id") } } else { if order == "asc" { q = q.Sort(sortBy, "_id") } else { q = q.Sort("-"+sortBy, "-_id") } } if limit > 0 { q = q.Limit(limit) } // Collect the results. var allResults []types.Process err = q.All(&allResults) if err != nil { err = errors.Wrap(err, "mongo find failed") return } // Compute the next cursor. nextCursor := "" if limit > 0 && len(allResults) == limit { lastResult := allResults[len(allResults)-1] nextCursor = lastResult.ID.Hex() if sortBy != "" { var encoded string var b []byte switch sortBy { case "id": b = []byte(lastResult.ID) case "process-id": b = make([]byte, 4) binary.BigEndian.PutUint64(b, lastResult.ProcessID) case "host-id": b = []byte(lastResult.HostID) case "host-address": b = []byte(lastResult.HostAddress) case "start-time": b, err = lastResult.StartTime.GobEncode() case "type": b = []byte(lastResult.Type) case "resource": b = []byte(lastResult.Resource) case "status": b = []byte(lastResult.Status) } encoded = hex.EncodeToString(b) nextCursor = encoded + "-" + nextCursor } } // Assemble the results. result = allResults cm = types.CollectionMetadata{ TotalResultSize: resultSize, ReturnedResultSize: len(result), NextPageCursor: nextCursor, } return } // CountProcesses is the same as GetProcesses but returns only the count, not the actual processes. func (context Context) CountProcesses(filters F) (count int, err error) { c := context.Session.DB(context.DBName).C("processes") // Build a query given the parameters. query := bson.M{} for k, v := range filters { switch k { case "id": setDefault(&query, "_id", bson.M{}) query["_id"].(bson.M)["$in"] = v.([]bson.ObjectId) case "process-id": setDefault(&query, k, bson.M{}) query[k].(bson.M)["$eq"] = v.(uint64) case "host-id", "host-address", "type", "resource", "status": setDefault(&query, k, bson.M{}) query[k].(bson.M)["$eq"] = v.(string) default: err = errors.Wrap(ErrBadInput, "invalid value of argument filters") return } } // We count the number of tasks that satisfy the filter criteria. count, err = c.Find(query).Count() if err != nil { err = errors.Wrap(err, "mongo find failed") } return } // CreateProcess adds a given process to the database. func (context Context) CreateProcess(proc types.Process) (result types.Process, err error) { // This action is only permitted for the root user. if context.User.IsRoot() == false { err = types.ErrUnauthorized return } if proc.Type != types.ProcController && proc.Type != types.ProcWorker && proc.Type != types.ProcScheduler { err = errors.Wrapf(ErrBadInput, "value of type can be \"%s\", \"%s\" or \"%s\", but found \"%s\"", types.ProcController, types.ProcWorker, types.ProcScheduler, proc.Type) return } // Find the first candidate ordinal. We do this optimistically assuming there is no race conditions. proc.RunningOrinal, err = context.findCandidateOrdinal(proc.Type) if err != nil { err = errors.Wrap(err, "find candidate ordinal failed") return } // Give default values to some fields. proc.ID = bson.NewObjectId() proc.Status = types.ProcIdle proc.StartTime = time.Now() c := context.Session.DB(context.DBName).C("processes") err = c.Insert(proc) if err != nil { lastError := err.(*mgo.LastError) if lastError.Code == 11000 { err = types.ErrIdentifierTaken return } err = errors.Wrap(err, "mongo insert failed") return } // Check if the ordinal was accepted. If not, then there was a race condition so we need to handle it. var accepted bool accepted, err = context.isOrdinalAccepted(proc.Type, proc.ID, proc.RunningOrinal) if err != nil { err = errors.Wrap(err, "is ordinal accepted failed") return } for accepted == false { // As long as the ordinal isn't accepted we need to try again. proc.RunningOrinal, err = context.findCandidateOrdinal(proc.Type) if err != nil { err = errors.Wrap(err, "find candidate ordinal failed") return } // Update the process with the new ordinal. c := context.Session.DB(context.DBName).C("processes") err = c.Update(bson.M{"_id": proc.ID}, bson.M{"$set": bson.M{"running-ordinal": proc.RunningOrinal}}) if err != nil { err = errors.Wrap(err, "mongo update failed") return } // Check if the ordinal was accepted. accepted, err = context.isOrdinalAccepted(proc.Type, proc.ID, proc.RunningOrinal) if err != nil { err = errors.Wrap(err, "is ordinal accepted failed") return } } return proc, nil } func (context Context) findCandidateOrdinal(processType string) (ordinal int, err error) { c := context.Session.DB(context.DBName).C("processes") query := bson.M{"type": processType, "status": bson.M{"$ne": types.ProcTerminated}} var ordinals []struct { Ordinal int `bson:"running-ordinal"` } q := c.Find(query).Select(bson.M{"running-ordinal": 1}).Sort("running-ordinal") err = q.All(&ordinals) if err != nil { return -1, err } ordinal = 1 for i := range ordinals { if ordinal == ordinals[i].Ordinal { ordinal++ } } return } func (context Context) isOrdinalAccepted(processType string, processID bson.ObjectId, ordinal int) (result bool, err error) { c := context.Session.DB(context.DBName).C("processes") query := bson.M{"type": processType, "status": bson.M{"$ne": types.ProcTerminated
random_line_split
process.go
( filters F, limit int, cursor string, sortBy string, order string, ) (result []types.Process, cm types.CollectionMetadata, err error) { c := context.Session.DB(context.DBName).C("processes") // Validate the parameters. if sortBy != "" && sortBy != "id" && sortBy != "process-id" && sortBy != "host-id" && sortBy != "host-address" && sortBy != "start-time" && sortBy != "type" && sortBy != "resource" && sortBy != "status" { err = errors.Wrapf(ErrBadInput, "cannot sort by \"%s\"", sortBy) return } if order != "" && order != "asc" && order != "desc" { err = errors.Wrapf(ErrBadInput, "order can be either \"asc\" or \"desc\", not \"%s\"", order) return } if order == "" { order = "asc" } // We currently don't limit access to this collection. Everyone can see it. // Build a query given the parameters. query := bson.M{} for k, v := range filters { switch k { case "id": setDefault(&query, "_id", bson.M{}) query["_id"].(bson.M)["$in"] = v.([]bson.ObjectId) case "process-id": setDefault(&query, k, bson.M{}) query[k].(bson.M)["$eq"] = v.(uint64) case "host-id", "host-address", "type", "resource", "status": setDefault(&query, k, bson.M{}) query[k].(bson.M)["$eq"] = v.(string) default: err = errors.Wrap(ErrBadInput, "invalid value of argument filters") return } } // We count the result size given the filters. This is before pagination. var resultSize int resultSize, err = c.Find(query).Count() if err != nil { err = errors.Wrap(err, "mongo find failed") return } // If a cursor was specified then we have to do a range query. if cursor != "" { comparer := "$gt" if order == "desc" { comparer = "$lt" } // If there is no sorting then the cursor only points to the _id field. if sortBy != "" && sortBy != "id" { splits := strings.Split(cursor, "-") cursor = splits[1] var decoded []byte decoded, err = hex.DecodeString(splits[0]) if err != nil { err = errors.Wrap(err, "hex decode string failed") return } var otherCursor interface{} switch sortBy { case "host-id", "host-address", "type", "resource", "status": otherCursor = string(decoded) case "process-id": otherCursor = binary.BigEndian.Uint64(decoded) case "start-time": var t time.Time t.GobDecode(decoded) otherCursor = t } setDefault(&query, "$or", bson.M{}) query["$or"] = []bson.M{ bson.M{sortBy: bson.M{comparer: otherCursor}}, bson.M{sortBy: bson.M{"$eq": otherCursor}, "_id": bson.M{comparer: bson.ObjectIdHex(cursor)}}, } } else { if bson.IsObjectIdHex(cursor) == false { err = errors.Wrap(ErrBadInput, "invalid cursor") return } setDefault(&query, "_id", bson.M{}) query["_id"].(bson.M)[comparer] = bson.ObjectIdHex(cursor) } } // Execute the query. q := c.Find(query) // We always sort by _id, but we may also sort by a specific field. if sortBy == "" || sortBy == "id" { if order == "asc" { q = q.Sort("_id") } else { q = q.Sort("-_id") } } else { if order == "asc" { q = q.Sort(sortBy, "_id") } else { q = q.Sort("-"+sortBy, "-_id") } } if limit > 0 { q = q.Limit(limit) } // Collect the results. var allResults []types.Process err = q.All(&allResults) if err != nil { err = errors.Wrap(err, "mongo find failed") return } // Compute the next cursor. nextCursor := "" if limit > 0 && len(allResults) == limit { lastResult := allResults[len(allResults)-1] nextCursor = lastResult.ID.Hex() if sortBy != "" { var encoded string var b []byte switch sortBy { case "id": b = []byte(lastResult.ID) case "process-id": b = make([]byte, 4) binary.BigEndian.PutUint64(b, lastResult.ProcessID) case "host-id": b = []byte(lastResult.HostID) case "host-address": b = []byte(lastResult.HostAddress) case "start-time": b, err = lastResult.StartTime.GobEncode() case "type": b = []byte(lastResult.Type) case "resource": b = []byte(lastResult.Resource) case "status": b = []byte(lastResult.Status) } encoded = hex.EncodeToString(b) nextCursor = encoded + "-" + nextCursor } } // Assemble the results. result = allResults cm = types.CollectionMetadata{ TotalResultSize: resultSize, ReturnedResultSize: len(result), NextPageCursor: nextCursor, } return } // CountProcesses is the same as GetProcesses but returns only the count, not the actual processes. func (context Context) CountProcesses(filters F) (count int, err error) { c := context.Session.DB(context.DBName).C("processes") // Build a query given the parameters. query := bson.M{} for k, v := range filters { switch k { case "id": setDefault(&query, "_id", bson.M{}) query["_id"].(bson.M)["$in"] = v.([]bson.ObjectId) case "process-id": setDefault(&query, k, bson.M{}) query[k].(bson.M)["$eq"] = v.(uint64) case "host-id", "host-address", "type", "resource", "status": setDefault(&query, k, bson.M{}) query[k].(bson.M)["$eq"] = v.(string) default: err = errors.Wrap(ErrBadInput, "invalid value of argument filters") return } } // We count the number of tasks that satisfy the filter criteria. count, err = c.Find(query).Count() if err != nil { err = errors.Wrap(err, "mongo find failed") } return } // CreateProcess adds a given process to the database. func (context Context) CreateProcess(proc types.Process) (result types.Process, err error) { // This action is only permitted for the root user. if context.User.IsRoot() == false { err = types.ErrUnauthorized return } if proc.Type != types.ProcController && proc.Type != types.ProcWorker && proc.Type != types.ProcScheduler { err = errors.Wrapf(ErrBadInput, "value of type can be \"%s\", \"%s\" or \"%s\", but found \"%s\"", types.ProcController, types.ProcWorker, types.ProcScheduler, proc.Type) return } // Find the first candidate ordinal. We do this optimistically assuming there is no race conditions. proc.RunningOrinal, err = context.findCandidateOrdinal(proc.Type) if err != nil { err = errors.Wrap(err, "find candidate ordinal failed") return } // Give default values to some fields. proc.ID = bson.NewObjectId() proc.Status = types.ProcIdle proc.StartTime = time.Now() c := context.Session.DB(context.DBName).C("processes") err = c.Insert(proc) if err != nil { lastError := err.(*mgo.LastError) if lastError.Code == 11000 { err = types.ErrIdentifierTaken return } err = errors.Wrap(err, "mongo insert failed") return } // Check if the ordinal was accepted. If not, then there was a race condition so we need to handle it. var accepted bool accepted, err = context.isOrdinalAccepted(proc.Type, proc.ID, proc.RunningOrinal) if err != nil { err = errors.Wrap(err, "is ordinal accepted failed") return } for accepted == false { // As long as the ordinal isn't accepted we need to try again. proc.RunningOrinal, err = context.findCandidateOrdinal(proc.Type) if err != nil { err = errors.Wrap(err, "find candidate ordinal failed") return } // Update the process with the new ordinal. c := context.Session.DB(context.DBName).C("processes") err = c.Update(bson
GetProcesses
identifier_name
intcode.rs
Pos = 0, Imm = 1, } impl TryFrom<isize> for AddrMode { type Error = &'static str; fn try_from(num: isize) -> Result<Self, Self::Error> { match num { 0 => Ok(Self::Pos), 1 => Ok(Self::Imm), _ => Err("invalid address mode value"), } } } #[derive(Debug)] enum IPChange { Delta(isize), New(usize), Halt, } /// Parse instruction will take a full instruction, and split it into the original instruction /// along with addressing modes for each argument. fn parse_instruction(word: isize) -> Result<(OpCode, AddrMode, AddrMode, AddrMode), &'static str> { if word <= 0 { return Err("instruction word must be greater than zero"); } Ok(( OpCode::try_from(word % 100)?, // first two digits are op AddrMode::try_from(word / 100 % 10)?, // 100s place AddrMode::try_from(word / 1000 % 10)?, // 1000s place AddrMode::try_from(word / 10000 % 10)?, // 10000s place )) } /// Trait is used by interpret for reading information interactively pub trait Input { fn get_isize(&mut self) -> isize; } /// Trait is used by `interpret` for writing information interactively pub trait Output { fn write_isize(&mut self, val: isize) -> (); } // Implementations for Input trait impl Input for () { fn get_isize(&mut self) -> isize { panic!("Program requested input, but input source was ()"); } } impl Input for isize { fn get_isize(&mut self) -> isize { *self } } // Implementations for Output trait impl Output for () { fn write_isize(&mut self, _val: isize) -> () { panic!("Program attempted to write value, but out was ()"); } } impl Output for &mut Vec<isize> { fn write_isize(&mut self, val: isize) -> () { self.push(val) } } /// Interpret array as an IntCode program. ///
/// `mem` is the initial machine memory state, it is modified during the run /// /// Will panic if it encounters an unknown opcode pub fn interpret(mut mem: &mut [isize], mut input: impl Input, mut output: impl Output) -> isize { let mut ip: usize = 0; loop { match step(&mut mem, ip, &mut input, &mut output) { IPChange::Delta(delta) => ip = (ip as isize + delta) as usize, IPChange::New(new) => ip = new, IPChange::Halt => break, } } mem[0] } fn step( mem: &mut [isize], ip: usize, input: &mut impl Input, output: &mut impl Output, ) -> IPChange { use AddrMode::*; use OpCode::*; let (op, addr1, addr2, addr3) = match parse_instruction(mem[ip]) { Ok(val) => val, Err(err) => { println!( "State:\n\tIP: {}\n\tVals: {:?}, {:?}, {:?}, {:?}", ip, mem.get(ip), mem.get(ip + 1), mem.get(ip + 2), mem.get(ip + 3) ); panic!(format!("Encountered unrecoverable error: {}", err)); } }; // placing Halt check here so that args can be extracted without duplicating their code all // over the place if op == Halt { return IPChange::Halt; } // HACK: this whole block is a hack, need to wrap memory up in a new type and provide accessors // that understand addressing modes let arg1 = match addr1 { Imm => mem.get(ip + 1), Pos => mem.get(*mem.get(ip + 1).unwrap_or(&0) as usize), } .unwrap_or(&-1337); let arg2 = match addr2 { Imm => mem.get(ip + 2), Pos => mem.get(*mem.get(ip + 2).unwrap_or(&0) as usize), } .unwrap_or(&-1337); match op { Add => { mem[mem[ip + 3] as usize] = arg1 + arg2; IPChange::Delta(4) } Multiply => { mem[mem[ip + 3] as usize] = arg1 * arg2; IPChange::Delta(4) } ReadIn => { mem[mem[ip + 1] as usize] = input.get_isize(); IPChange::Delta(2) } WriteOut => { output.write_isize(mem[mem[ip + 1] as usize]); IPChange::Delta(2) } JmpIfTrue => { if *arg1 != 0 { IPChange::New(usize::try_from(*arg2).unwrap()) } else { IPChange::Delta(3) } } JmpIfFalse => { if *arg1 == 0 { IPChange::New(usize::try_from(*arg2).unwrap()) } else { IPChange::Delta(3) } } LessThan => { mem[mem[ip + 3] as usize] = if arg1 < arg2 { 1 } else { 0 }; IPChange::Delta(4) } Equals => { mem[mem[ip + 3] as usize] = if arg1 == arg2 { 1 } else { 0 }; IPChange::Delta(4) } Halt => unreachable!(), } } #[cfg(test)] mod test { use super::*; #[test] fn interpret_day2_examples() { let mut programs = vec![ vec![1, 9, 10, 3, 2, 3, 11, 0, 99, 30, 40, 50], vec![1, 0, 0, 0, 99], vec![2, 3, 0, 3, 99], vec![2, 4, 4, 5, 99, 0], vec![1, 1, 1, 4, 99, 5, 6, 0, 99], ]; let outputs = vec![ vec![3500, 9, 10, 70, 2, 3, 11, 0, 99, 30, 40, 50], vec![2, 0, 0, 0, 99], vec![2, 3, 0, 6, 99], vec![2, 4, 4, 5, 99, 9801], vec![30, 1, 1, 4, 2, 5, 6, 0, 99], ]; for i in 0..programs.len() { assert_eq!(interpret(&mut programs[i], (), ()), outputs[i][0]); assert_eq!(programs[i], outputs[i]); } } #[test] fn test_parse_instruction() { use AddrMode::*; use OpCode::*; type Output = (OpCode, AddrMode, AddrMode, AddrMode); fn eq(left: Output, right: Output) -> bool { left.0 == right.0 && left.1 == right.1 && left.2 == right.2 && left.3 == right.3 } // from day 5 examples assert!(eq( parse_instruction(1002).unwrap(), (Multiply, Pos, Imm, Pos) )); // synthetic assert!(eq(parse_instruction(2).unwrap(), (Multiply, Pos, Pos, Pos))); assert!(eq(parse_instruction(11101).unwrap(), (Add, Imm, Imm, Imm))); assert!(eq(parse_instruction(10101).unwrap(), (Add, Imm, Pos, Imm))); assert!(eq( parse_instruction(104).unwrap(), (WriteOut, Imm, Pos, Pos) )); assert!(eq( parse_instruction(10003).unwrap(), (ReadIn, Pos, Pos, Imm) )); } #[test] fn day5_snippets() { // This tests immediate and positional addressing and negative immediate support // Should: find (100 + -1), store result @4 let mut simple_prog = vec![1101, 100, -1, 4, 0]; interpret(&mut simple_prog, (), ()); assert_eq!(simple_prog[4], 99); // This should save whatever it gets from input to @0, then print it back out let arb_input = 1
random_line_split
intcode.rs
} #[derive(Debug, PartialEq)] enum AddrMode { Pos = 0, Imm = 1, } impl TryFrom<isize> for AddrMode { type Error = &'static str; fn try_from(num: isize) -> Result<Self, Self::Error> { match num { 0 => Ok(Self::Pos), 1 => Ok(Self::Imm), _ => Err("invalid address mode value"), } } } #[derive(Debug)] enum IPChange { Delta(isize), New(usize), Halt, } /// Parse instruction will take a full instruction, and split it into the original instruction /// along with addressing modes for each argument. fn parse_instruction(word: isize) -> Result<(OpCode, AddrMode, AddrMode, AddrMode), &'static str> { if word <= 0 { return Err("instruction word must be greater than zero"); } Ok(( OpCode::try_from(word % 100)?, // first two digits are op AddrMode::try_from(word / 100 % 10)?, // 100s place AddrMode::try_from(word / 1000 % 10)?, // 1000s place AddrMode::try_from(word / 10000 % 10)?, // 10000s place )) } /// Trait is used by interpret for reading information interactively pub trait Input { fn get_isize(&mut self) -> isize; } /// Trait is used by `interpret` for writing information interactively pub trait Output { fn write_isize(&mut self, val: isize) -> (); } // Implementations for Input trait impl Input for () { fn get_isize(&mut self) -> isize { panic!("Program requested input, but input source was ()"); } } impl Input for isize { fn get_isize(&mut self) -> isize { *self } } // Implementations for Output trait impl Output for () { fn write_isize(&mut self, _val: isize) -> () { panic!("Program attempted to write value, but out was ()"); } } impl Output for &mut Vec<isize> { fn write_isize(&mut self, val: isize) -> () { self.push(val) } } /// Interpret array as an IntCode program. /// /// `mem` is the initial machine memory state, it is modified during the run /// /// Will panic if it encounters an unknown opcode pub fn interpret(mut mem: &mut [isize], mut input: impl Input, mut output: impl Output) -> isize { let mut ip: usize = 0; loop { match step(&mut mem, ip, &mut input, &mut output) { IPChange::Delta(delta) => ip = (ip as isize + delta) as usize, IPChange::New(new) => ip = new, IPChange::Halt => break, } } mem[0] } fn step( mem: &mut [isize], ip: usize, input: &mut impl Input, output: &mut impl Output, ) -> IPChange { use AddrMode::*; use OpCode::*; let (op, addr1, addr2, addr3) = match parse_instruction(mem[ip]) { Ok(val) => val, Err(err) => { println!( "State:\n\tIP: {}\n\tVals: {:?}, {:?}, {:?}, {:?}", ip, mem.get(ip), mem.get(ip + 1), mem.get(ip + 2), mem.get(ip + 3) ); panic!(format!("Encountered unrecoverable error: {}", err)); } }; // placing Halt check here so that args can be extracted without duplicating their code all // over the place if op == Halt { return IPChange::Halt; } // HACK: this whole block is a hack, need to wrap memory up in a new type and provide accessors // that understand addressing modes let arg1 = match addr1 { Imm => mem.get(ip + 1), Pos => mem.get(*mem.get(ip + 1).unwrap_or(&0) as usize), } .unwrap_or(&-1337); let arg2 = match addr2 { Imm => mem.get(ip + 2), Pos => mem.get(*mem.get(ip + 2).unwrap_or(&0) as usize), } .unwrap_or(&-1337); match op { Add => { mem[mem[ip + 3] as usize] = arg1 + arg2; IPChange::Delta(4) } Multiply => { mem[mem[ip + 3] as usize] = arg1 * arg2; IPChange::Delta(4) } ReadIn => { mem[mem[ip + 1] as usize] = input.get_isize(); IPChange::Delta(2) } WriteOut => { output.write_isize(mem[mem[ip + 1] as usize]); IPChange::Delta(2) } JmpIfTrue => { if *arg1 != 0 { IPChange::New(usize::try_from(*arg2).unwrap()) } else { IPChange::Delta(3) } } JmpIfFalse => { if *arg1 == 0 { IPChange::New(usize::try_from(*arg2).unwrap()) } else { IPChange::Delta(3) } } LessThan => { mem[mem[ip + 3] as usize] = if arg1 < arg2 { 1 } else { 0 }; IPChange::Delta(4) } Equals => { mem[mem[ip + 3] as usize] = if arg1 == arg2 { 1 } else { 0 }; IPChange::Delta(4) } Halt => unreachable!(), } } #[cfg(test)] mod test { use super::*; #[test] fn interpret_day2_examples() { let mut programs = vec![ vec![1, 9, 10, 3, 2, 3, 11, 0, 99, 30, 40, 50], vec![1, 0, 0, 0, 99], vec![2, 3, 0, 3, 99], vec![2, 4, 4, 5, 99, 0], vec![1, 1, 1, 4, 99, 5, 6, 0, 99], ]; let outputs = vec![ vec![3500, 9, 10, 70, 2, 3, 11, 0, 99, 30, 40, 50], vec![2, 0, 0, 0, 99], vec![2, 3, 0, 6, 99], vec![2, 4, 4, 5, 99, 9801], vec![30, 1, 1, 4, 2, 5, 6, 0, 99], ]; for i in 0..programs.len() { assert_eq!(interpret(&mut programs[i], (), ()), outputs[i][0]); assert_eq!(programs[i], outputs[i]); } } #[test] fn test_parse_instruction() { use AddrMode::*; use OpCode::*; type Output = (OpCode, AddrMode, AddrMode, AddrMode); fn eq(left: Output, right: Output) -> bool { left.0 == right.0 && left.1 == right.1 && left.2 == right.2 && left.3 == right.3 } // from day 5 examples assert!(eq( parse_instruction(1002).unwrap(), (Multiply, Pos, Imm, Pos) )); // synthetic assert!(eq(parse_instruction(2).unwrap(), (Multiply, Pos, Pos, Pos))); assert!(eq(parse_instruction(11101).unwrap(), (Add, Imm, Imm, Imm))); assert!(eq(parse_instruction(10101).unwrap(), (Add, Imm, Pos, Imm))); assert!(eq( parse_instruction(104).unwrap(), (WriteOut, Imm, Pos, Pos) )); assert!(eq( parse_instruction(10003).unwrap(), (ReadIn, Pos, Pos, Imm)
{ match num { 1 => Ok(Self::Add), 2 => Ok(Self::Multiply), 3 => Ok(Self::ReadIn), 4 => Ok(Self::WriteOut), 5 => Ok(Self::JmpIfTrue), 6 => Ok(Self::JmpIfFalse), 7 => Ok(Self::LessThan), 8 => Ok(Self::Equals), 99 => Ok(Self::Halt), _ => Err("invalid opcode value"), } }
identifier_body
intcode.rs
(num: isize) -> Result<Self, Self::Error> { match num { 1 => Ok(Self::Add), 2 => Ok(Self::Multiply), 3 => Ok(Self::ReadIn), 4 => Ok(Self::WriteOut), 5 => Ok(Self::JmpIfTrue), 6 => Ok(Self::JmpIfFalse), 7 => Ok(Self::LessThan), 8 => Ok(Self::Equals), 99 => Ok(Self::Halt), _ => Err("invalid opcode value"), } } } #[derive(Debug, PartialEq)] enum AddrMode { Pos = 0, Imm = 1, } impl TryFrom<isize> for AddrMode { type Error = &'static str; fn try_from(num: isize) -> Result<Self, Self::Error> { match num { 0 => Ok(Self::Pos), 1 => Ok(Self::Imm), _ => Err("invalid address mode value"), } } } #[derive(Debug)] enum IPChange { Delta(isize), New(usize), Halt, } /// Parse instruction will take a full instruction, and split it into the original instruction /// along with addressing modes for each argument. fn parse_instruction(word: isize) -> Result<(OpCode, AddrMode, AddrMode, AddrMode), &'static str> { if word <= 0 { return Err("instruction word must be greater than zero"); } Ok(( OpCode::try_from(word % 100)?, // first two digits are op AddrMode::try_from(word / 100 % 10)?, // 100s place AddrMode::try_from(word / 1000 % 10)?, // 1000s place AddrMode::try_from(word / 10000 % 10)?, // 10000s place )) } /// Trait is used by interpret for reading information interactively pub trait Input { fn get_isize(&mut self) -> isize; } /// Trait is used by `interpret` for writing information interactively pub trait Output { fn write_isize(&mut self, val: isize) -> (); } // Implementations for Input trait impl Input for () { fn get_isize(&mut self) -> isize { panic!("Program requested input, but input source was ()"); } } impl Input for isize { fn get_isize(&mut self) -> isize { *self } } // Implementations for Output trait impl Output for () { fn write_isize(&mut self, _val: isize) -> () { panic!("Program attempted to write value, but out was ()"); } } impl Output for &mut Vec<isize> { fn write_isize(&mut self, val: isize) -> () { self.push(val) } } /// Interpret array as an IntCode program. /// /// `mem` is the initial machine memory state, it is modified during the run /// /// Will panic if it encounters an unknown opcode pub fn interpret(mut mem: &mut [isize], mut input: impl Input, mut output: impl Output) -> isize { let mut ip: usize = 0; loop { match step(&mut mem, ip, &mut input, &mut output) { IPChange::Delta(delta) => ip = (ip as isize + delta) as usize, IPChange::New(new) => ip = new, IPChange::Halt => break, } } mem[0] } fn step( mem: &mut [isize], ip: usize, input: &mut impl Input, output: &mut impl Output, ) -> IPChange { use AddrMode::*; use OpCode::*; let (op, addr1, addr2, addr3) = match parse_instruction(mem[ip]) { Ok(val) => val, Err(err) => { println!( "State:\n\tIP: {}\n\tVals: {:?}, {:?}, {:?}, {:?}", ip, mem.get(ip), mem.get(ip + 1), mem.get(ip + 2), mem.get(ip + 3) ); panic!(format!("Encountered unrecoverable error: {}", err)); } }; // placing Halt check here so that args can be extracted without duplicating their code all // over the place if op == Halt { return IPChange::Halt; } // HACK: this whole block is a hack, need to wrap memory up in a new type and provide accessors // that understand addressing modes let arg1 = match addr1 { Imm => mem.get(ip + 1), Pos => mem.get(*mem.get(ip + 1).unwrap_or(&0) as usize), } .unwrap_or(&-1337); let arg2 = match addr2 { Imm => mem.get(ip + 2), Pos => mem.get(*mem.get(ip + 2).unwrap_or(&0) as usize), } .unwrap_or(&-1337); match op { Add => { mem[mem[ip + 3] as usize] = arg1 + arg2; IPChange::Delta(4) } Multiply => { mem[mem[ip + 3] as usize] = arg1 * arg2; IPChange::Delta(4) } ReadIn => { mem[mem[ip + 1] as usize] = input.get_isize(); IPChange::Delta(2) } WriteOut => { output.write_isize(mem[mem[ip + 1] as usize]); IPChange::Delta(2) } JmpIfTrue => { if *arg1 != 0 { IPChange::New(usize::try_from(*arg2).unwrap()) } else { IPChange::Delta(3) } } JmpIfFalse => { if *arg1 == 0 { IPChange::New(usize::try_from(*arg2).unwrap()) } else { IPChange::Delta(3) } } LessThan => { mem[mem[ip + 3] as usize] = if arg1 < arg2 { 1 } else { 0 }; IPChange::Delta(4) } Equals => { mem[mem[ip + 3] as usize] = if arg1 == arg2 { 1 } else { 0 }; IPChange::Delta(4) } Halt => unreachable!(), } } #[cfg(test)] mod test { use super::*; #[test] fn interpret_day2_examples() { let mut programs = vec![ vec![1, 9, 10, 3, 2, 3, 11, 0, 99, 30, 40, 50], vec![1, 0, 0, 0, 99], vec![2, 3, 0, 3, 99], vec![2, 4, 4, 5, 99, 0], vec![1, 1, 1, 4, 99, 5, 6, 0, 99], ]; let outputs = vec![ vec![3500, 9, 10, 70, 2, 3, 11, 0, 99, 30, 40, 50], vec![2, 0, 0, 0, 99], vec![2, 3, 0, 6, 99], vec![2, 4, 4, 5, 99, 9801], vec![30, 1, 1, 4, 2, 5, 6, 0, 99], ]; for i in 0..programs.len() { assert_eq!(interpret(&mut programs[i], (), ()), outputs[i][0]); assert_eq!(programs[i], outputs[i]); } } #[test] fn test_parse_instruction() { use AddrMode::*; use OpCode::*; type Output = (OpCode, AddrMode, AddrMode, AddrMode); fn eq(left: Output, right: Output) -> bool { left.0 == right.0 && left.1 == right.1 && left.2 == right.2 && left.3 == right.3 } // from day 5 examples assert!(eq( parse_instruction(1002).unwrap(), (Multiply, Pos, Imm, Pos) )); // synthetic assert!(eq(parse_instruction(2).unwrap(), (Multiply, Pos, Pos, Pos))); assert!(eq(parse_instruction(11101).unwrap(), (Add, Imm, Imm, Imm))); assert!(eq(parse_instruction(10101).unwrap(), (Add, Imm, Pos, Imm))); assert!(eq( parse_instruction(104).unwrap(), (WriteOut, Imm, Pos, Pos) )); assert!(eq( parse_instruction(10003).
try_from
identifier_name
main.go
[:0] } continue } requestChan <- js tmpbyte = tmpbyte[:0] } } } func serverStart() { eventLoop() } func eventLoop() { for { js := <-requestChan if js.Method == "LoginQq" { LoginQQ, _ = js.Params.getInt64("loginqq") logger.Printf(">>> %d\n", LoginQQ) continue } subtype, _ := js.Params.getInt64("subtype") switch js.Method { case "GroupMessage": go groupMessageHandle(js.Params) case "DiscussMessage": go discussMessageHandle(js.Params) case "PrivateMessage": go privateMessageHandle(js.Params) case "GroupMemberJoin": qqNum, _ := js.Params.getInt64("qqnum") groupNum, _ := js.Params.getInt64("groupnum") beingOperateQQ, _ := js.Params.getInt64("opqqnum") message := welcomeNewMember(subtype, groupNum, qqNum, beingOperateQQ) sendGroupMessage(groupNum, message) getGroupMemberList(groupNum) case "GroupMemberLeave": qqNum, _ := js.Params.getInt64("qqnum") groupNum, _ := js.Params.getInt64("groupnum") if groupNum == 196656732 { continue } //beingOperateQQ := js.Params.GetInt64("opqqnum") if v, ok := groups.Load(groupNum); ok { group := v.(Group) members := group.Members if members == nil { continue } if v, ok := members.Load(qqNum); ok { member := v.(Member) if subtype == 1 { message := fmt.Sprintf("群员:[%s] 退群了!!!", member.Nickname) sendGroupMessage(groupNum, message) } else if subtype == 2 { message := fmt.Sprintf("群员:[%s] 被 某个管理员 踢了!!!", member.Nickname) sendGroupMessage(groupNum, message) } } } case "RequestAddFriend": responseFlag, _ := js.Params.getString("response_flag") addFriend(responseFlag, 1, "") case "RequestAddGroup": responseFlag, _ := js.Params.getString("response_flag") if subtype == 2 { addGroup(responseFlag, 2, 1, "") } getGroupList() case "GroupMemberList": var groupMemberList []Member if err := js.Params.UnmarshalGroupMemberList(&groupMemberList); err != nil { logger.Printf(">>> get group member list faild: %v", err) } logger.Printf(">>> %v\n", groupMemberList) go updateGroupMember(groupMemberList) case "GroupList": var grouplist []Group if err := js.Params.UnmarshalGroupList(&grouplist); err != nil { logger.Printf(">>> get group list faild: %v", err) } //logger.Printf(">>> %v\n", grouplist) go updateGroupList(grouplist) case "GroupMemberInfo": var memberInfo Member if err := js.Params.UnmarshalGroupMemberInfo(&memberInfo); err != nil { logger.Printf(">>> get member info faild: %v", err) } logger.Printf(">>> %v\n", memberInfo) go updateMemberInfo(memberInfo) default: logger.Printf("未处理:%s\n", js) } } } func GetGroupListFromDB() { rows, err := db.Query("select id, group_number, name from groups") if err != nil { panic(err) } defer rows.Close() for rows.Next() { var groupName string var groupID int64 var groupNum int64 rows.Scan(&groupID, &groupNum, &groupName) if _, ok := groups.Load(groupNum); !ok { group := Group{} group.ID = groupID group.GroupNum = groupNum group.GroupName = groupName group.Members = GetGroupMembersFromDB(group.ID, group.GroupNum) groups.Store(groupNum, group) } } } func GetGroupMembersFromDB(groupId int64, groupNumbber int64) *sync.Map { memberList := new(sync.Map) rows, err := db.Query("select m.id, m.user_id, u.qq_number, m.Nickname, m.Rights from group_members m join users u on m.user_id = u.id where m.group_id = $1", groupId) if err != nil { panic(err) } defer rows.Close() for rows.Next() { var id int64 var userId int64 var qq_number int64 var nickname string var rights int32 rows.Scan(&id, &userId, &qq_number, &nickname, &rights) m := Member{} m.ID = id m.UserID = userId m.GroupID = groupId m.GroupNum = groupNumbber m.QQNum = qq_number m.Nickname = nickname m.Permission = rights memberList.Store(qq_number, m) } return memberList } func updateGroupList(groupList []Group) { if len(groupList) == 0 { return } GetGroupListFromDB() var groupNums []int64 = []int64{} for _, ng := range groupList { groupNums = append(groupNums, ng.GroupNum) if v, ok := groups.Load(ng.GroupNum); ok { og := v.(Group) if og.GroupName != ng.GroupName { og.GroupName = ng.GroupName groups.Store(ng.GroupNum, og) trans, err := db.Begin() if err != nil { //reportError(err) continue } _, err = trans.Exec("update groups set name = $1 where Id = $2", ng.GroupName, og.ID) if err != nil { //reportError(err) trans.Rollback() continue } else { trans.Commit() } } } else { var groupID int64 trans, err := db.Begin() if err != nil { //reportError(err) continue } err = trans.QueryRow("insert into groups(group_number, name) values($1, $2) returning id", ng.GroupNum, ng.GroupName).Scan(&groupID) if err != nil { //reportError(err) trans.Rollback() continue } else { trans.Commit() } groups.Store(ng.GroupNum, Group{ID: groupID, GroupNum: ng.GroupNum, GroupName: ng.GroupName, Members: nil}) //getGroupMemberList(ng.GroupNum) } } var waitForDeleteGroupNums []int64 = []int64{} groups.Range(func(key, value interface{}) bool { found := false for _, num := range groupNums { if num == key { found = true break } } if !found { g := value.(Group) waitForDeleteGroupNums = append(waitForDeleteGroupNums, g.GroupNum) trans, err := db.Begin() if err != nil { return true } _, err = trans.Exec("delete from replies where group_id = $1", g.ID) if err != nil { trans.Rollback() return true } _, err = trans.Exec("delete from group_members where group_id = $1", g.ID) if err != nil { trans.Rollback() return true } _, err = trans.Exec("delete from groups where id = $1", g.ID) if err != nil { trans.Rollback() } else { trans.Commit() } } return true }) for _, num := range waitForDeleteGroupNums { groups.Delete(num) } groups.Range(func(key, value interface{}) bool { log.Printf(">>> new groups: %v\n", value.(Group)) return true }) } func updateGroupMember(groupMemberList []Member) { flag := true for _, nm := range gr
oupMemberList { if v, ok := groups.Load(nm.GroupNum); ok { g := v.(Group) if flag { g.Members = GetGroupMembersFromDB(g.ID, g.GroupNum) flag = false } if g.Members == nil { g.Members = new(sync.Map) } if _, ok := g.Members.Load(nm.QQNum); !ok { g.Members.Store(nm.QQNum, nm) trans, err := db.Begin() if err != nil { //reportError(err) continue } var userID int64 var count int64 err = trans.QueryRow("select count(1) from users where qq_numer = $1", nm.QQNum).Scan(&count)
identifier_body
main.go
return } s := strings.SplitN(strings.TrimSpace(cmd), ":", 2) if len(s) == 2 { groupNum, err := strconv.ParseInt(strings.TrimSpace(s[0]), 10, 64) if err != nil { fmt.Println(err) return } if g, ok := groups.Load(groupNum); ok { sendGroupMessage(g.(Group).GroupNum, s[1]) } else { fmt.Println("group not found!") } } } func handleRequest(conn net.Conn) { defer conn.Close() tmpbyte := make([]byte, 2<<22) tmpbyte = tmpbyte[:0] buf := make([]byte, 2<<20) for { // Read the incoming connection into the buffer. reqLen, err := conn.Read(buf) if err != nil { logger.Println("Error reading:", err.Error()) return } if reqLen == 0 { continue } scanner := bufio.NewScanner(bytes.NewReader(buf[:reqLen])) for scanner.Scan() { b := bytes.TrimSpace(scanner.Bytes()) if len(b) == 0 { logger.Println("len(b)==0", string(b)) continue } else if len(tmpbyte) > 0 && len(tmpbyte)+len(b) < 4096 { b = append(tmpbyte, b...) logger.Printf("retry b := %s\n", string(b)) } else { tmpbyte = tmpbyte[:0] } var js Notification err = json.Unmarshal(b, &js) if err != nil { if serr, ok := err.(*json.SyntaxError); ok { logger.Printf("%s, %s\n", serr.Error(), string(b)) if errStr := serr.Error(); strings.HasPrefix(errStr, "invalid character") && strings.HasSuffix(errStr, "in string literal") { tmpbyte = tmpbyte[:0] continue } tmpbyte = append(tmpbyte, b...) } else { logger.Printf("%s, %s\n", err.Error(), string(b)) tmpbyte = tmpbyte[:0] } continue } requestChan <- js tmpbyte = tmpbyte[:0] } } } func serverStart() { eventLoop() } func eventLoop() { for { js := <-requestChan if js.Method == "LoginQq" { LoginQQ, _ = js.Params.getInt64("loginqq") logger.Printf(">>> %d\n", LoginQQ) continue } subtype, _ := js.Params.getInt64("subtype") switch js.Method { case "GroupMessage": go groupMessageHandle(js.Params) case "DiscussMessage": go discussMessageHandle(js.Params) case "PrivateMessage": go privateMessageHandle(js.Params) case "GroupMemberJoin": qqNum, _ := js.Params.getInt64("qqnum") groupNum, _ := js.Params.getInt64("groupnum") beingOperateQQ, _ := js.Params.getInt64("opqqnum") message := welcomeNewMember(subtype, groupNum, qqNum, beingOperateQQ) sendGroupMessage(groupNum, message) getGroupMemberList(groupNum) case "GroupMemberLeave": qqNum, _ := js.Params.getInt64("qqnum") groupNum, _ := js.Params.getInt64("groupnum") if groupNum == 196656732 { continue } //beingOperateQQ := js.Params.GetInt64("opqqnum") if v, ok := groups.Load(groupNum); ok { group := v.(Group) members := group.Members if members == nil { continue } if v, ok := members.Load(qqNum); ok { member := v.(Member) if subtype == 1 { message := fmt.Sprintf("群员:[%s] 退群了!!!", member.Nickname) sendGroupMessage(groupNum, message) } else if subtype == 2 { message := fmt.Sprintf("群员:[%s] 被 某个管理员 踢了!!!", member.Nickname) sendGroupMessage(groupNum, message) } } } case "RequestAddFriend": responseFlag, _ := js.Params.getString("response_flag") addFriend(responseFlag, 1, "") case "RequestAddGroup": responseFlag, _ := js.Params.getString("response_flag") if subtype == 2 { addGroup(responseFlag, 2, 1, "") } getGroupList() case "GroupMemberList": var groupMemberList []Member if err := js.Params.UnmarshalGroupMemberList(&groupMemberList); err != nil { logger.Printf(">>> get group member list faild: %v", err) } logger.Printf(">>> %v\n", groupMemberList) go updateGroupMember(groupMemberList) case "GroupList": var grouplist []Group if err := js.Params.UnmarshalGroupList(&grouplist); err != nil { logger.Printf(">>> get group list faild: %v", err) } //logger.Printf(">>> %v\n", grouplist) go updateGroupList(grouplist) case "GroupMemberInfo": var memberInfo Member if err := js.Params.UnmarshalGroupMemberInfo(&memberInfo); err != nil { logger.Printf(">>> get member info faild: %v", err) } logger.Printf(">>> %v\n", memberInfo) go updateMemberInfo(memberInfo) default: logger.Printf("未处理:%s\n", js) } } } func GetGroupListFromDB() { rows, err := db.Query("select id, group_number, name from groups") if err != nil { panic(err) } defer rows.Close() for rows.Next() { var groupName string var groupID int64 var groupNum int64 rows.Scan(&groupID, &groupNum, &groupName) if _, ok := groups.Load(groupNum); !ok { group := Group{} group.ID = groupID group.GroupNum = groupNum group.GroupName = groupName group.Members = GetGroupMembersFromDB(group.ID, group.GroupNum) groups.Store(groupNum, group) } } } func GetGroupMembersFromDB(groupId int64, groupNumbber int64) *sync.Map { memberList := new(sync.Map) rows, err := db.Query("select m.id, m.user_id, u.qq_number, m.Nickname, m.Rights from group_members m join users u on m.user_id = u.id where m.group_id = $1", groupId) if err != nil { panic(err) } defer rows.Close() for rows.Next() { var id int64 var userId int64 var qq_number int64 var nickname string var rights int32 rows.Scan(&id, &userId, &qq_number, &nickname, &rights) m := Member{} m.ID = id m.UserID = userId m.GroupID = groupId m.GroupNum = groupNumbber m.QQNum = qq_number m.Nickname = nickname m.Permission = rights memberList.Store(qq_number, m) } return memberList } func updateGroupList(groupList []Group) { if len(groupList) == 0 { return } GetGroupListFromDB() var groupNums []int64 = []int64{} for _, ng := range groupList { groupNums = append(groupNums, ng.GroupNum) if v, ok := groups.Load(ng.GroupNum); ok { og := v.(Group) if og.GroupName != ng.GroupName { og.GroupName = ng.GroupName groups.Store(ng.GroupNum, og) trans, err := db.Begin() if err != nil { //reportError(err) continue } _, err = trans.Exec("update groups set name = $1 where Id = $2", ng.GroupName, og.ID) if err != nil { //reportError(err) trans.Rollback() continue } else { trans.Commit() } } } else { var groupID int64 trans, err := db.Begin() if err != nil { //reportError(err) continue } err = trans.QueryRow("insert into groups(group_number, name) values($1, $2) returning id", ng.GroupNum, ng.GroupName).Scan(&groupID) if err != nil { //reportError(err) trans.Rollback() continue } else { trans.Commit() } groups.Store(ng.GroupNum, Group{ID: groupID, GroupNum: ng.GroupNum, GroupName: ng.GroupName, Members: nil}) //getGroupMemberList(ng.GroupNum) } } var waitFor
{ leaveGroup(groupNum) getGroupList() }
conditional_block
main.go
0] continue } tmpbyte = append(tmpbyte, b...) } else { logger.Printf("%s, %s\n", err.Error(), string(b)) tmpbyte = tmpbyte[:0] } continue } requestChan <- js tmpbyte = tmpbyte[:0] } } } func serverStart() { eventLoop() } func eventLoop() { for { js := <-requestChan if js.Method == "LoginQq" { LoginQQ, _ = js.Params.getInt64("loginqq") logger.Printf(">>> %d\n", LoginQQ) continue } subtype, _ := js.Params.getInt64("subtype") switch js.Method { case "GroupMessage": go groupMessageHandle(js.Params) case "DiscussMessage": go discussMessageHandle(js.Params) case "PrivateMessage": go privateMessageHandle(js.Params) case "GroupMemberJoin": qqNum, _ := js.Params.getInt64("qqnum") groupNum, _ := js.Params.getInt64("groupnum") beingOperateQQ, _ := js.Params.getInt64("opqqnum") message := welcomeNewMember(subtype, groupNum, qqNum, beingOperateQQ) sendGroupMessage(groupNum, message) getGroupMemberList(groupNum) case "GroupMemberLeave": qqNum, _ := js.Params.getInt64("qqnum") groupNum, _ := js.Params.getInt64("groupnum") if groupNum == 196656732 { continue } //beingOperateQQ := js.Params.GetInt64("opqqnum") if v, ok := groups.Load(groupNum); ok { group := v.(Group) members := group.Members if members == nil { continue } if v, ok := members.Load(qqNum); ok { member := v.(Member) if subtype == 1 { message := fmt.Sprintf("群员:[%s] 退群了!!!", member.Nickname) sendGroupMessage(groupNum, message) } else if subtype == 2 { message := fmt.Sprintf("群员:[%s] 被 某个管理员 踢了!!!", member.Nickname) sendGroupMessage(groupNum, message) } } } case "RequestAddFriend": responseFlag, _ := js.Params.getString("response_flag") addFriend(responseFlag, 1, "") case "RequestAddGroup": responseFlag, _ := js.Params.getString("response_flag") if subtype == 2 { addGroup(responseFlag, 2, 1, "") } getGroupList() case "GroupMemberList": var groupMemberList []Member if err := js.Params.UnmarshalGroupMemberList(&groupMemberList); err != nil { logger.Printf(">>> get group member list faild: %v", err) } logger.Printf(">>> %v\n", groupMemberList) go updateGroupMember(groupMemberList) case "GroupList": var grouplist []Group if err := js.Params.UnmarshalGroupList(&grouplist); err != nil { logger.Printf(">>> get group list faild: %v", err) } //logger.Printf(">>> %v\n", grouplist) go updateGroupList(grouplist) case "GroupMemberInfo": var memberInfo Member if err := js.Params.UnmarshalGroupMemberInfo(&memberInfo); err != nil { logger.Printf(">>> get member info faild: %v", err) } logger.Printf(">>> %v\n", memberInfo) go updateMemberInfo(memberInfo) default: logger.Printf("未处理:%s\n", js) } } } func GetGroupListFromDB() { rows, err := db.Query("select id, group_number, name from groups") if err != nil { panic(err) } defer rows.Close() for rows.Next() { var groupName string var groupID int64 var groupNum int64 rows.Scan(&groupID, &groupNum, &groupName) if _, ok := groups.Load(groupNum); !ok { group := Group{} group.ID = groupID group.GroupNum = groupNum group.GroupName = groupName group.Members = GetGroupMembersFromDB(group.ID, group.GroupNum) groups.Store(groupNum, group) } } } func GetGroupMembersFromDB(groupId int64, groupNumbber int64) *sync.Map { memberList := new(sync.Map) rows, err := db.Query("select m.id, m.user_id, u.qq_number, m.Nickname, m.Rights from group_members m join users u on m.user_id = u.id where m.group_id = $1", groupId) if err != nil { panic(err) } defer rows.Close() for rows.Next() { var id int64 var userId int64 var qq_number int64 var nickname string var rights int32 rows.Scan(&id, &userId, &qq_number, &nickname, &rights) m := Member{} m.ID = id m.UserID = userId m.GroupID = groupId m.GroupNum = groupNumbber m.QQNum = qq_number m.Nickname = nickname m.Permission = rights memberList.Store(qq_number, m) } return memberList } func updateGroupList(groupList []Group) { if len(groupList) == 0 { return } GetGroupListFromDB() var groupNums []int64 = []int64{} for _, ng := range groupList { groupNums = append(groupNums, ng.GroupNum) if v, ok := groups.Load(ng.GroupNum); ok { og := v.(Group) if og.GroupName != ng.GroupName { og.GroupName = ng.GroupName groups.Store(ng.GroupNum, og) trans, err := db.Begin() if err != nil { //reportError(err) continue } _, err = trans.Exec("update groups set name = $1 where Id = $2", ng.GroupName, og.ID) if err != nil { //reportError(err) trans.Rollback() continue } else { trans.Commit() } } } else { var groupID int64 trans, err := db.Begin() if err != nil { //reportError(err) continue } err = trans.QueryRow("insert into groups(group_number, name) values($1, $2) returning id", ng.GroupNum, ng.GroupName).Scan(&groupID) if err != nil { //reportError(err) trans.Rollback() continue } else { trans.Commit() } groups.Store(ng.GroupNum, Group{ID: groupID, GroupNum: ng.GroupNum, GroupName: ng.GroupName, Members: nil}) //getGroupMemberList(ng.GroupNum) } } var waitForDeleteGroupNums []int64 = []int64{} groups.Range(func(key, value interface{}) bool { found := false for _, num := range groupNums { if num == key { found = true break } } if !found { g := value.(Group) waitForDeleteGroupNums = append(waitForDeleteGroupNums, g.GroupNum) trans, err := db.Begin() if err != nil { return true } _, err = trans.Exec("delete from replies where group_id = $1", g.ID) if err != nil { trans.Rollback() return true } _, err = trans.Exec("delete from group_members where group_id = $1", g.ID) if err != nil { trans.Rollback() return true } _, err = trans.Exec("delete from groups where id = $1", g.ID) if err != nil { trans.Rollback() } else { trans.Commit() } } return true }) for _, num := range waitForDeleteGroupNums { groups.Delete(num) } groups.Range(func(key, value interface{}) bool { log.Printf(">>> new groups: %v\n", value.(Group)) return true }) } func updateGroupMember(groupMemberList []Member) { flag := true for _, nm := range groupMemberList { if v, ok := groups.Load(nm.GroupNum); ok { g := v.(Group) if flag { g.Members = GetGroupMembersFromDB(g.ID, g.GroupNum) flag = false } if g.Members == nil { g.Members = new(sync.Map) } if _, ok := g.Members.Load(nm.QQNum); !ok { g.Members.Store(nm.QQNum, nm) trans, err := db.Begin() if err != nil { //reportError(err)
continue }
random_line_split
main.go
(cmd string) { if cmd == "exit" { close(closeSignChan) return } else if cmd == "init" { getLoginQQ() getGroupList() return } else if strings.HasPrefix(cmd, "random:") { var i int32 = 0 var target = rand.Int31n(300) groups.Range(func(key, value interface{}) bool { if i == target { sendGroupMessage(value.(Group).GroupNum, cmd[7:]) return false } i++ return true }) return } else if strings.HasPrefix(cmd, "level:") { groupNum, err := strconv.ParseInt(strings.TrimSpace(cmd[6:]), 10, 64) if err != nil { fmt.Println(err) return } if _, ok := groups.Load(groupNum); ok { leaveGroup(groupNum) getGroupList() } return } s := strings.SplitN(strings.TrimSpace(cmd), ":", 2) if len(s) == 2 { groupNum, err := strconv.ParseInt(strings.TrimSpace(s[0]), 10, 64) if err != nil { fmt.Println(err) return } if g, ok := groups.Load(groupNum); ok { sendGroupMessage(g.(Group).GroupNum, s[1]) } else { fmt.Println("group not found!") } } } func handleRequest(conn net.Conn) { defer conn.Close() tmpbyte := make([]byte, 2<<22) tmpbyte = tmpbyte[:0] buf := make([]byte, 2<<20) for { // Read the incoming connection into the buffer. reqLen, err := conn.Read(buf) if err != nil { logger.Println("Error reading:", err.Error()) return } if reqLen == 0 { continue } scanner := bufio.NewScanner(bytes.NewReader(buf[:reqLen])) for scanner.Scan() { b := bytes.TrimSpace(scanner.Bytes()) if len(b) == 0 { logger.Println("len(b)==0", string(b)) continue } else if len(tmpbyte) > 0 && len(tmpbyte)+len(b) < 4096 { b = append(tmpbyte, b...) logger.Printf("retry b := %s\n", string(b)) } else { tmpbyte = tmpbyte[:0] } var js Notification err = json.Unmarshal(b, &js) if err != nil { if serr, ok := err.(*json.SyntaxError); ok { logger.Printf("%s, %s\n", serr.Error(), string(b)) if errStr := serr.Error(); strings.HasPrefix(errStr, "invalid character") && strings.HasSuffix(errStr, "in string literal") { tmpbyte = tmpbyte[:0] continue } tmpbyte = append(tmpbyte, b...) } else { logger.Printf("%s, %s\n", err.Error(), string(b)) tmpbyte = tmpbyte[:0] } continue } requestChan <- js tmpbyte = tmpbyte[:0] } } } func serverStart() { eventLoop() } func eventLoop() { for { js := <-requestChan if js.Method == "LoginQq" { LoginQQ, _ = js.Params.getInt64("loginqq") logger.Printf(">>> %d\n", LoginQQ) continue } subtype, _ := js.Params.getInt64("subtype") switch js.Method { case "GroupMessage": go groupMessageHandle(js.Params) case "DiscussMessage": go discussMessageHandle(js.Params) case "PrivateMessage": go privateMessageHandle(js.Params) case "GroupMemberJoin": qqNum, _ := js.Params.getInt64("qqnum") groupNum, _ := js.Params.getInt64("groupnum") beingOperateQQ, _ := js.Params.getInt64("opqqnum") message := welcomeNewMember(subtype, groupNum, qqNum, beingOperateQQ) sendGroupMessage(groupNum, message) getGroupMemberList(groupNum) case "GroupMemberLeave": qqNum, _ := js.Params.getInt64("qqnum") groupNum, _ := js.Params.getInt64("groupnum") if groupNum == 196656732 { continue } //beingOperateQQ := js.Params.GetInt64("opqqnum") if v, ok := groups.Load(groupNum); ok { group := v.(Group) members := group.Members if members == nil { continue } if v, ok := members.Load(qqNum); ok { member := v.(Member) if subtype == 1 { message := fmt.Sprintf("群员:[%s] 退群了!!!", member.Nickname) sendGroupMessage(groupNum, message) } else if subtype == 2 { message := fmt.Sprintf("群员:[%s] 被 某个管理员 踢了!!!", member.Nickname) sendGroupMessage(groupNum, message) } } } case "RequestAddFriend": responseFlag, _ := js.Params.getString("response_flag") addFriend(responseFlag, 1, "") case "RequestAddGroup": responseFlag, _ := js.Params.getString("response_flag") if subtype == 2 { addGroup(responseFlag, 2, 1, "") } getGroupList() case "GroupMemberList": var groupMemberList []Member if err := js.Params.UnmarshalGroupMemberList(&groupMemberList); err != nil { logger.Printf(">>> get group member list faild: %v", err) } logger.Printf(">>> %v\n", groupMemberList) go updateGroupMember(groupMemberList) case "GroupList": var grouplist []Group if err := js.Params.UnmarshalGroupList(&grouplist); err != nil { logger.Printf(">>> get group list faild: %v", err) } //logger.Printf(">>> %v\n", grouplist) go updateGroupList(grouplist) case "GroupMemberInfo": var memberInfo Member if err := js.Params.UnmarshalGroupMemberInfo(&memberInfo); err != nil { logger.Printf(">>> get member info faild: %v", err) } logger.Printf(">>> %v\n", memberInfo) go updateMemberInfo(memberInfo) default: logger.Printf("未处理:%s\n", js) } } } func GetGroupListFromDB() { rows, err := db.Query("select id, group_number, name from groups") if err != nil { panic(err) } defer rows.Close() for rows.Next() { var groupName string var groupID int64 var groupNum int64 rows.Scan(&groupID, &groupNum, &groupName) if _, ok := groups.Load(groupNum); !ok { group := Group{} group.ID = groupID group.GroupNum = groupNum group.GroupName = groupName group.Members = GetGroupMembersFromDB(group.ID, group.GroupNum) groups.Store(groupNum, group) } } } func GetGroupMembersFromDB(groupId int64, groupNumbber int64) *sync.Map { memberList := new(sync.Map) rows, err := db.Query("select m.id, m.user_id, u.qq_number, m.Nickname, m.Rights from group_members m join users u on m.user_id = u.id where m.group_id = $1", groupId) if err != nil { panic(err) } defer rows.Close() for rows.Next() { var id int64 var userId int64 var qq_number int64 var nickname string var rights int32 rows.Scan(&id, &userId, &qq_number, &nickname, &rights) m := Member{} m.ID = id m.UserID = userId m.GroupID = groupId m.GroupNum = groupNumbber m.QQNum = qq_number m.Nickname = nickname m.Permission = rights memberList.Store(qq_number, m) } return memberList } func updateGroupList(groupList []Group) { if len(groupList) == 0 { return } GetGroupListFromDB() var groupNums []int64 = []int64{} for _, ng := range groupList { groupNums = append(groupNums, ng.GroupNum) if v, ok := groups.Load(ng.GroupNum); ok { og := v.(Group) if og.GroupName != ng.GroupName { og.GroupName = ng.GroupName groups.Store(ng.GroupNum, og) trans, err := db.Begin() if err != nil { //reportError(err) continue } _, err = trans.Exec("update groups set name = $1 where Id = $2", ng.GroupName, og.ID) if
cmd
identifier_name
main.rs
pub mod shader; pub mod text; pub mod texture; // needed for font atlas but needed things can be ported out pub mod timer; pub mod util; pub use display::Display; pub use input::Handler; pub use loader::Loader; pub use render::{RenderMgr, }; pub use shader::Shader; pub use timer::Timer; fn main() { // Test code for parsing fnt files // use text::metafile::test_noms; // test_noms(); // Specify OpenGL version let gl_request = glutin::GlRequest::Specific(glutin::Api::OpenGl, (4, 3)); let gl_profile = glutin::GlProfile::Core; // Create a window let el = EventLoop::new(); let wb = glutin::window::WindowBuilder::new() .with_title("RaumEn SysInfo") .with_inner_size(glutin::dpi::LogicalSize::new(1024.0, 768.0)) .with_maximized(false); let windowed_context = glutin::ContextBuilder::new() .with_gl(gl_request) .with_gl_profile(gl_profile) .build_windowed(wb, &el) .unwrap(); let windowed_context = unsafe { windowed_context.make_current().unwrap() }; // Set up OpenGL unsafe { load_with(|symbol| windowed_context.context().get_proc_address(symbol) as *const _); ClearColor(0.0, 1.0, 0.0, 1.0); } let mut render_mgr = RenderMgr::new(); let mut mgr = render_mgr.mgr.clone(); let mut system = sysinfo::System::new(); let cpu = cpu_name(); let ram = get_ram_total(&mut system); let cpu_ram = mk_cpu_ram_str(&cpu, &ram, &mut system); let hdd = get_hdd(); let mut fps: f32 = 30.0; let mut once_per_sec = false; let mut clean_up_time = false; { // Here, we're getting the size of the window in pixels // and passing it to the update_size() method. It in turn // updates the Projection Matrix and passes that to // ALL THE SHADERS, so if you add a SHADER, you need // to REMEMBER to add that shader to the update_size() // method near the bottom of this file. // let dpi = windowed_context.window().get_hidpi_factor(); let size: glutin::dpi::PhysicalSize<u32> = windowed_context.window().inner_size(); mgr.update_size(size.into()); } { let _textmgr = mgr.clone().textmgr.take().unwrap(); let mut textmgr = _textmgr.lock().unwrap(); textmgr.add_font(mgr.clone(), "pirate"); textmgr.add_font(mgr.clone(), "sans"); textmgr.new_text(mgr.clone(), "Title", "SysInfo", "pirate", 4.0, 0.0, 0.0, 1.0, true, true); textmgr.new_text(mgr.clone(), "CPU RAM HDD", &[cpu_ram, hdd.clone()].join(""), "sans", 2.0, 0.0, 0.4, 1.0, true, true); textmgr.new_text(mgr.clone(), "FPS", "FPS: 0.0", "sans", 1.5, 0.0, 0.0, 0.3, false, true); } // Game loop! println!("Starting main loop."); el.run(move |event, _, control_flow| { // *control_flow = ControlFlow::Wait; match event { Event::LoopDestroyed => { println!("Cleaning Up..."); // Clean up render_mgr.clean_up(); clean_up_time = true; return } Event::WindowEvent { event, .. } => match event { WindowEvent::CloseRequested => { *control_flow = ControlFlow::Exit; }, WindowEvent::Resized(size) => { windowed_context.resize(size); mgr.update_size(size.into()); }, _ => { mgr.handler_do(|handler| { handler.window_event(&event); }); } }, DeviceEvent{ event, ..} => { mgr.handler_do(|handler| { handler.device_event(&event); }); } Event::NewEvents( _time ) => { // Emitted when new events arrive from the OS to be processed. // // This event type is useful as a place to put code that should be done before you start processing events, such as // updating frame timing information for benchmarking or checking the StartCause][crate::event::StartCause] to see // if a timer set by [ControlFlow::WaitUntil has elapsed. } Event::MainEventsCleared => { // Emitted when all of the event loop's input events have been processed and redraw processing is about to begin. // This event is useful as a place to put your code that should be run after all state-changing events have been // handled and you want to do stuff (updating state, performing calculations, etc) that happens as the "main body" // of your event loop. // If your program draws graphics, it's usually better to do it in response to Event::RedrawRequested, which gets // emitted immediately after this event. { let mut handler = mgr.handler.lock().unwrap(); handler.timer.tick(); handler.reset_delta(); if handler.timer.once_per_sec() { fps = handler.timer.fps; once_per_sec = true; } } if once_per_sec { once_per_sec = false; println!("Once per second FPS: {}", &format!("FPS: {:.3}", (fps * 1000.0).round() / 1000.0 ) ); let cpu_ram = mk_cpu_ram_str(&cpu, &ram, &mut system); let _textmgr = mgr.clone().textmgr.take().unwrap(); let mut textmgr = _textmgr.lock().unwrap(); textmgr.update_text(mgr.clone(), "CPU RAM HDD", &[cpu_ram, hdd.clone()].join("")); textmgr.update_text(mgr.clone(), "FPS", &format!("FPS: {:.3}", (fps * 1000.0).round() / 1000.0 ) ); } windowed_context.window().request_redraw(); } Event::RedrawRequested(_) => { // Emitted after MainEventsCleared when a window should be redrawn. // This gets triggered in two scenarios: // - The OS has performed an operation that's invalidated the window's contents (such as resizing the window). // - The application has explicitly requested a redraw via Window::request_redraw. // During each iteration of the event loop, Winit will aggregate duplicate redraw requests into a single event, // to help avoid duplicating rendering work. if clean_up_time { return; } // *** Drawing phase render_mgr.render(); // _fbo_final.blit_to_screen(&world); // Write the new frame to the screen! windowed_context.swap_buffers().unwrap(); } Event::RedrawEventsCleared => { // Emitted after all RedrawRequested events have been processed and control flow is about to be taken away from // the program. If there are no RedrawRequested events, it is emitted immediately after MainEventsCleared. // This event is useful for doing any cleanup or bookkeeping work after all the rendering tasks have been completed. } e => println!("Other Event:\n{:?}", e) } }); } pub const EOF: &str = "\04"; pub fn eof(string: &str) -> String { [string, EOF].join("") } // pub fn call_cmd(cmd: &str) -> Result<String, String> { // use subprocess::{Exec,Redirection}; // let out = Exec::shell(cmd) // .stdout(Redirection::Pipe) // .capture().map_err(|e|e.to_string())? // .stdout_str(); // return Ok(out.trim().to_owned()); // } // use nom::{multispace, rest_s}; // named!(_cpu_name<&str, String>, // do_parse!( // tag!("Name") >> multispace >> out: rest_s >> // ( out.to_owned() ) // ) // ); fn mk_cpu_ram_str(cpu: &str, ram: &str, system: &mut sysinfo::System) -> String
fn cpu_name() -> String { // use cupid; let info = cupid::master(); match info { Some(x) => { match x.brand_string() { Some(x) => { ["CPU: ".to_owned(), x.to_owned()].join("") } _ => { "Could not get CPU Name".to_owned() } } } _ => { "Could not get CPU Name".to_owned() } } } fn get_ram_total(system: &mut sysinfo::System) -> String { system.refresh_all(); let
{ let ram_used = get_ram_used(system); [cpu.to_owned(), ram.to_owned(), ram_used].join("\n") }
identifier_body
main.rs
) .unwrap(); let windowed_context = unsafe { windowed_context.make_current().unwrap() }; // Set up OpenGL unsafe { load_with(|symbol| windowed_context.context().get_proc_address(symbol) as *const _); ClearColor(0.0, 1.0, 0.0, 1.0); } let mut render_mgr = RenderMgr::new(); let mut mgr = render_mgr.mgr.clone(); let mut system = sysinfo::System::new(); let cpu = cpu_name(); let ram = get_ram_total(&mut system); let cpu_ram = mk_cpu_ram_str(&cpu, &ram, &mut system); let hdd = get_hdd(); let mut fps: f32 = 30.0; let mut once_per_sec = false; let mut clean_up_time = false; { // Here, we're getting the size of the window in pixels // and passing it to the update_size() method. It in turn // updates the Projection Matrix and passes that to // ALL THE SHADERS, so if you add a SHADER, you need // to REMEMBER to add that shader to the update_size() // method near the bottom of this file. // let dpi = windowed_context.window().get_hidpi_factor(); let size: glutin::dpi::PhysicalSize<u32> = windowed_context.window().inner_size(); mgr.update_size(size.into()); } { let _textmgr = mgr.clone().textmgr.take().unwrap(); let mut textmgr = _textmgr.lock().unwrap(); textmgr.add_font(mgr.clone(), "pirate"); textmgr.add_font(mgr.clone(), "sans"); textmgr.new_text(mgr.clone(), "Title", "SysInfo", "pirate", 4.0, 0.0, 0.0, 1.0, true, true); textmgr.new_text(mgr.clone(), "CPU RAM HDD", &[cpu_ram, hdd.clone()].join(""), "sans", 2.0, 0.0, 0.4, 1.0, true, true); textmgr.new_text(mgr.clone(), "FPS", "FPS: 0.0", "sans", 1.5, 0.0, 0.0, 0.3, false, true); } // Game loop! println!("Starting main loop."); el.run(move |event, _, control_flow| { // *control_flow = ControlFlow::Wait; match event { Event::LoopDestroyed => { println!("Cleaning Up..."); // Clean up render_mgr.clean_up(); clean_up_time = true; return } Event::WindowEvent { event, .. } => match event { WindowEvent::CloseRequested => { *control_flow = ControlFlow::Exit; }, WindowEvent::Resized(size) => { windowed_context.resize(size); mgr.update_size(size.into()); }, _ => { mgr.handler_do(|handler| { handler.window_event(&event); }); } }, DeviceEvent{ event, ..} => { mgr.handler_do(|handler| { handler.device_event(&event); }); } Event::NewEvents( _time ) => { // Emitted when new events arrive from the OS to be processed. // // This event type is useful as a place to put code that should be done before you start processing events, such as // updating frame timing information for benchmarking or checking the StartCause][crate::event::StartCause] to see // if a timer set by [ControlFlow::WaitUntil has elapsed. } Event::MainEventsCleared => { // Emitted when all of the event loop's input events have been processed and redraw processing is about to begin. // This event is useful as a place to put your code that should be run after all state-changing events have been // handled and you want to do stuff (updating state, performing calculations, etc) that happens as the "main body" // of your event loop. // If your program draws graphics, it's usually better to do it in response to Event::RedrawRequested, which gets // emitted immediately after this event. { let mut handler = mgr.handler.lock().unwrap(); handler.timer.tick(); handler.reset_delta(); if handler.timer.once_per_sec() { fps = handler.timer.fps; once_per_sec = true; } } if once_per_sec { once_per_sec = false; println!("Once per second FPS: {}", &format!("FPS: {:.3}", (fps * 1000.0).round() / 1000.0 ) ); let cpu_ram = mk_cpu_ram_str(&cpu, &ram, &mut system); let _textmgr = mgr.clone().textmgr.take().unwrap(); let mut textmgr = _textmgr.lock().unwrap(); textmgr.update_text(mgr.clone(), "CPU RAM HDD", &[cpu_ram, hdd.clone()].join("")); textmgr.update_text(mgr.clone(), "FPS", &format!("FPS: {:.3}", (fps * 1000.0).round() / 1000.0 ) ); } windowed_context.window().request_redraw(); } Event::RedrawRequested(_) => { // Emitted after MainEventsCleared when a window should be redrawn. // This gets triggered in two scenarios: // - The OS has performed an operation that's invalidated the window's contents (such as resizing the window). // - The application has explicitly requested a redraw via Window::request_redraw. // During each iteration of the event loop, Winit will aggregate duplicate redraw requests into a single event, // to help avoid duplicating rendering work. if clean_up_time { return; } // *** Drawing phase render_mgr.render(); // _fbo_final.blit_to_screen(&world); // Write the new frame to the screen! windowed_context.swap_buffers().unwrap(); } Event::RedrawEventsCleared => { // Emitted after all RedrawRequested events have been processed and control flow is about to be taken away from // the program. If there are no RedrawRequested events, it is emitted immediately after MainEventsCleared. // This event is useful for doing any cleanup or bookkeeping work after all the rendering tasks have been completed. } e => println!("Other Event:\n{:?}", e) } }); } pub const EOF: &str = "\04"; pub fn eof(string: &str) -> String { [string, EOF].join("") } // pub fn call_cmd(cmd: &str) -> Result<String, String> { // use subprocess::{Exec,Redirection}; // let out = Exec::shell(cmd) // .stdout(Redirection::Pipe) // .capture().map_err(|e|e.to_string())? // .stdout_str(); // return Ok(out.trim().to_owned()); // } // use nom::{multispace, rest_s}; // named!(_cpu_name<&str, String>, // do_parse!( // tag!("Name") >> multispace >> out: rest_s >> // ( out.to_owned() ) // ) // ); fn mk_cpu_ram_str(cpu: &str, ram: &str, system: &mut sysinfo::System) -> String { let ram_used = get_ram_used(system); [cpu.to_owned(), ram.to_owned(), ram_used].join("\n") } fn cpu_name() -> String { // use cupid; let info = cupid::master(); match info { Some(x) => { match x.brand_string() { Some(x) => { ["CPU: ".to_owned(), x.to_owned()].join("") } _ => { "Could not get CPU Name".to_owned() } } } _ => { "Could not get CPU Name".to_owned() } } } fn get_ram_total(system: &mut sysinfo::System) -> String { system.refresh_all(); let ram_total = ((system.get_total_memory() as f32 / 1024.0) / 1024.0).round(); format!("Total Memory: {} GB", ram_total ) } fn get_ram_used(system: &mut sysinfo::System) -> String { system.refresh_all(); let ram_used = (((system.get_used_memory() as f32 / 1024.0) / 1024.0) * 1000.0).round() / 1000.0; format!("Used Memory : {:.3} GB", ram_used ) } fn get_hdd() -> String { let sys = System::new(); let mut out = String::new(); match sys.mounts() { Ok(mounts) => { for mount in mounts.iter() { if mount.total == bytesize::ByteSize::b(0) { continue; } let mnt = mount.fs_mounted_on.clone(); if is_dir_or_subdir_linux(&mnt, "/boot") { continue; } if is_dir_or_subdir_linux(&mnt, "/dev") { continue; } if is_dir_or_subdir_linux(&mnt, "/run")
{ continue; }
conditional_block
main.rs
let gl_profile = glutin::GlProfile::Core; // Create a window let el = EventLoop::new(); let wb = glutin::window::WindowBuilder::new() .with_title("RaumEn SysInfo") .with_inner_size(glutin::dpi::LogicalSize::new(1024.0, 768.0)) .with_maximized(false); let windowed_context = glutin::ContextBuilder::new() .with_gl(gl_request) .with_gl_profile(gl_profile) .build_windowed(wb, &el) .unwrap(); let windowed_context = unsafe { windowed_context.make_current().unwrap() }; // Set up OpenGL unsafe { load_with(|symbol| windowed_context.context().get_proc_address(symbol) as *const _); ClearColor(0.0, 1.0, 0.0, 1.0); } let mut render_mgr = RenderMgr::new(); let mut mgr = render_mgr.mgr.clone(); let mut system = sysinfo::System::new(); let cpu = cpu_name(); let ram = get_ram_total(&mut system); let cpu_ram = mk_cpu_ram_str(&cpu, &ram, &mut system); let hdd = get_hdd(); let mut fps: f32 = 30.0; let mut once_per_sec = false; let mut clean_up_time = false; { // Here, we're getting the size of the window in pixels // and passing it to the update_size() method. It in turn // updates the Projection Matrix and passes that to // ALL THE SHADERS, so if you add a SHADER, you need // to REMEMBER to add that shader to the update_size() // method near the bottom of this file. // let dpi = windowed_context.window().get_hidpi_factor(); let size: glutin::dpi::PhysicalSize<u32> = windowed_context.window().inner_size(); mgr.update_size(size.into()); } { let _textmgr = mgr.clone().textmgr.take().unwrap(); let mut textmgr = _textmgr.lock().unwrap(); textmgr.add_font(mgr.clone(), "pirate"); textmgr.add_font(mgr.clone(), "sans"); textmgr.new_text(mgr.clone(), "Title", "SysInfo", "pirate", 4.0, 0.0, 0.0, 1.0, true, true); textmgr.new_text(mgr.clone(), "CPU RAM HDD", &[cpu_ram, hdd.clone()].join(""), "sans", 2.0, 0.0, 0.4, 1.0, true, true); textmgr.new_text(mgr.clone(), "FPS", "FPS: 0.0", "sans", 1.5, 0.0, 0.0, 0.3, false, true); } // Game loop! println!("Starting main loop."); el.run(move |event, _, control_flow| { // *control_flow = ControlFlow::Wait; match event { Event::LoopDestroyed => { println!("Cleaning Up..."); // Clean up render_mgr.clean_up(); clean_up_time = true; return } Event::WindowEvent { event, .. } => match event { WindowEvent::CloseRequested => { *control_flow = ControlFlow::Exit; }, WindowEvent::Resized(size) => { windowed_context.resize(size); mgr.update_size(size.into()); }, _ => { mgr.handler_do(|handler| { handler.window_event(&event); }); } }, DeviceEvent{ event, ..} => { mgr.handler_do(|handler| { handler.device_event(&event); }); } Event::NewEvents( _time ) => { // Emitted when new events arrive from the OS to be processed. // // This event type is useful as a place to put code that should be done before you start processing events, such as // updating frame timing information for benchmarking or checking the StartCause][crate::event::StartCause] to see // if a timer set by [ControlFlow::WaitUntil has elapsed. } Event::MainEventsCleared => { // Emitted when all of the event loop's input events have been processed and redraw processing is about to begin. // This event is useful as a place to put your code that should be run after all state-changing events have been // handled and you want to do stuff (updating state, performing calculations, etc) that happens as the "main body" // of your event loop. // If your program draws graphics, it's usually better to do it in response to Event::RedrawRequested, which gets // emitted immediately after this event. { let mut handler = mgr.handler.lock().unwrap(); handler.timer.tick(); handler.reset_delta(); if handler.timer.once_per_sec() { fps = handler.timer.fps; once_per_sec = true; } } if once_per_sec { once_per_sec = false; println!("Once per second FPS: {}", &format!("FPS: {:.3}", (fps * 1000.0).round() / 1000.0 ) ); let cpu_ram = mk_cpu_ram_str(&cpu, &ram, &mut system); let _textmgr = mgr.clone().textmgr.take().unwrap(); let mut textmgr = _textmgr.lock().unwrap(); textmgr.update_text(mgr.clone(), "CPU RAM HDD", &[cpu_ram, hdd.clone()].join("")); textmgr.update_text(mgr.clone(), "FPS", &format!("FPS: {:.3}", (fps * 1000.0).round() / 1000.0 ) ); } windowed_context.window().request_redraw(); } Event::RedrawRequested(_) => { // Emitted after MainEventsCleared when a window should be redrawn. // This gets triggered in two scenarios: // - The OS has performed an operation that's invalidated the window's contents (such as resizing the window). // - The application has explicitly requested a redraw via Window::request_redraw. // During each iteration of the event loop, Winit will aggregate duplicate redraw requests into a single event, // to help avoid duplicating rendering work. if clean_up_time { return; } // *** Drawing phase render_mgr.render(); // _fbo_final.blit_to_screen(&world); // Write the new frame to the screen! windowed_context.swap_buffers().unwrap(); } Event::RedrawEventsCleared => { // Emitted after all RedrawRequested events have been processed and control flow is about to be taken away from // the program. If there are no RedrawRequested events, it is emitted immediately after MainEventsCleared. // This event is useful for doing any cleanup or bookkeeping work after all the rendering tasks have been completed. } e => println!("Other Event:\n{:?}", e) } }); } pub const EOF: &str = "\04"; pub fn eof(string: &str) -> String { [string, EOF].join("") } // pub fn call_cmd(cmd: &str) -> Result<String, String> { // use subprocess::{Exec,Redirection}; // let out = Exec::shell(cmd) // .stdout(Redirection::Pipe) // .capture().map_err(|e|e.to_string())? // .stdout_str(); // return Ok(out.trim().to_owned()); // } // use nom::{multispace, rest_s}; // named!(_cpu_name<&str, String>, // do_parse!( // tag!("Name") >> multispace >> out: rest_s >> // ( out.to_owned() ) // ) // ); fn mk_cpu_ram_str(cpu: &str, ram: &str, system: &mut sysinfo::System) -> String { let ram_used = get_ram_used(system); [cpu.to_owned(), ram.to_owned(), ram_used].join("\n") } fn cpu_name() -> String { // use cupid; let info = cupid::master(); match info { Some(x) => { match x.brand_string() { Some(x) => { ["CPU: ".to_owned(), x.to_owned()].join("") } _ => { "Could not get CPU Name".to_owned() } } } _ => { "Could not get CPU Name".to_owned() } } } fn get_ram_total(system: &mut sysinfo::System) -> String { system.refresh_all(); let ram_total = ((system.get_total_memory() as f32 / 1024.0) / 1024.0).round(); format!("Total Memory: {} GB", ram_total ) } fn get_ram_used(system: &mut sysinfo::System) -> String { system.refresh_all(); let ram_used = (((system.get_used_memory() as f32 / 1024.0) / 1024.0) * 1000.0).round() / 1000.0; format!("Used Memory : {:.3} GB", ram_used ) } fn
get_hdd
identifier_name
main.rs
(|symbol| windowed_context.context().get_proc_address(symbol) as *const _); ClearColor(0.0, 1.0, 0.0, 1.0); } let mut render_mgr = RenderMgr::new(); let mut mgr = render_mgr.mgr.clone(); let mut system = sysinfo::System::new(); let cpu = cpu_name(); let ram = get_ram_total(&mut system); let cpu_ram = mk_cpu_ram_str(&cpu, &ram, &mut system); let hdd = get_hdd(); let mut fps: f32 = 30.0; let mut once_per_sec = false; let mut clean_up_time = false; { // Here, we're getting the size of the window in pixels // and passing it to the update_size() method. It in turn // updates the Projection Matrix and passes that to // ALL THE SHADERS, so if you add a SHADER, you need // to REMEMBER to add that shader to the update_size() // method near the bottom of this file. // let dpi = windowed_context.window().get_hidpi_factor(); let size: glutin::dpi::PhysicalSize<u32> = windowed_context.window().inner_size(); mgr.update_size(size.into()); } { let _textmgr = mgr.clone().textmgr.take().unwrap(); let mut textmgr = _textmgr.lock().unwrap(); textmgr.add_font(mgr.clone(), "pirate"); textmgr.add_font(mgr.clone(), "sans"); textmgr.new_text(mgr.clone(), "Title", "SysInfo", "pirate", 4.0, 0.0, 0.0, 1.0, true, true); textmgr.new_text(mgr.clone(), "CPU RAM HDD", &[cpu_ram, hdd.clone()].join(""), "sans", 2.0, 0.0, 0.4, 1.0, true, true); textmgr.new_text(mgr.clone(), "FPS", "FPS: 0.0", "sans", 1.5, 0.0, 0.0, 0.3, false, true); } // Game loop! println!("Starting main loop."); el.run(move |event, _, control_flow| { // *control_flow = ControlFlow::Wait; match event { Event::LoopDestroyed => { println!("Cleaning Up..."); // Clean up render_mgr.clean_up(); clean_up_time = true; return } Event::WindowEvent { event, .. } => match event { WindowEvent::CloseRequested => { *control_flow = ControlFlow::Exit; }, WindowEvent::Resized(size) => { windowed_context.resize(size); mgr.update_size(size.into()); }, _ => { mgr.handler_do(|handler| { handler.window_event(&event); }); } }, DeviceEvent{ event, ..} => { mgr.handler_do(|handler| { handler.device_event(&event); }); } Event::NewEvents( _time ) => { // Emitted when new events arrive from the OS to be processed. // // This event type is useful as a place to put code that should be done before you start processing events, such as // updating frame timing information for benchmarking or checking the StartCause][crate::event::StartCause] to see // if a timer set by [ControlFlow::WaitUntil has elapsed. } Event::MainEventsCleared => { // Emitted when all of the event loop's input events have been processed and redraw processing is about to begin. // This event is useful as a place to put your code that should be run after all state-changing events have been // handled and you want to do stuff (updating state, performing calculations, etc) that happens as the "main body" // of your event loop. // If your program draws graphics, it's usually better to do it in response to Event::RedrawRequested, which gets // emitted immediately after this event. { let mut handler = mgr.handler.lock().unwrap(); handler.timer.tick(); handler.reset_delta(); if handler.timer.once_per_sec() { fps = handler.timer.fps; once_per_sec = true; } } if once_per_sec { once_per_sec = false; println!("Once per second FPS: {}", &format!("FPS: {:.3}", (fps * 1000.0).round() / 1000.0 ) ); let cpu_ram = mk_cpu_ram_str(&cpu, &ram, &mut system); let _textmgr = mgr.clone().textmgr.take().unwrap(); let mut textmgr = _textmgr.lock().unwrap(); textmgr.update_text(mgr.clone(), "CPU RAM HDD", &[cpu_ram, hdd.clone()].join("")); textmgr.update_text(mgr.clone(), "FPS", &format!("FPS: {:.3}", (fps * 1000.0).round() / 1000.0 ) ); } windowed_context.window().request_redraw(); } Event::RedrawRequested(_) => { // Emitted after MainEventsCleared when a window should be redrawn. // This gets triggered in two scenarios: // - The OS has performed an operation that's invalidated the window's contents (such as resizing the window). // - The application has explicitly requested a redraw via Window::request_redraw. // During each iteration of the event loop, Winit will aggregate duplicate redraw requests into a single event, // to help avoid duplicating rendering work. if clean_up_time { return; } // *** Drawing phase render_mgr.render(); // _fbo_final.blit_to_screen(&world); // Write the new frame to the screen! windowed_context.swap_buffers().unwrap(); } Event::RedrawEventsCleared => { // Emitted after all RedrawRequested events have been processed and control flow is about to be taken away from // the program. If there are no RedrawRequested events, it is emitted immediately after MainEventsCleared. // This event is useful for doing any cleanup or bookkeeping work after all the rendering tasks have been completed. } e => println!("Other Event:\n{:?}", e) } }); } pub const EOF: &str = "\04"; pub fn eof(string: &str) -> String { [string, EOF].join("") } // pub fn call_cmd(cmd: &str) -> Result<String, String> { // use subprocess::{Exec,Redirection}; // let out = Exec::shell(cmd) // .stdout(Redirection::Pipe) // .capture().map_err(|e|e.to_string())? // .stdout_str(); // return Ok(out.trim().to_owned()); // } // use nom::{multispace, rest_s}; // named!(_cpu_name<&str, String>, // do_parse!( // tag!("Name") >> multispace >> out: rest_s >> // ( out.to_owned() ) // ) // ); fn mk_cpu_ram_str(cpu: &str, ram: &str, system: &mut sysinfo::System) -> String { let ram_used = get_ram_used(system); [cpu.to_owned(), ram.to_owned(), ram_used].join("\n") } fn cpu_name() -> String { // use cupid; let info = cupid::master(); match info { Some(x) => { match x.brand_string() { Some(x) => { ["CPU: ".to_owned(), x.to_owned()].join("") } _ => { "Could not get CPU Name".to_owned() } } } _ => { "Could not get CPU Name".to_owned() } } } fn get_ram_total(system: &mut sysinfo::System) -> String { system.refresh_all(); let ram_total = ((system.get_total_memory() as f32 / 1024.0) / 1024.0).round(); format!("Total Memory: {} GB", ram_total ) } fn get_ram_used(system: &mut sysinfo::System) -> String { system.refresh_all(); let ram_used = (((system.get_used_memory() as f32 / 1024.0) / 1024.0) * 1000.0).round() / 1000.0; format!("Used Memory : {:.3} GB", ram_used ) } fn get_hdd() -> String { let sys = System::new(); let mut out = String::new(); match sys.mounts() { Ok(mounts) => { for mount in mounts.iter() { if mount.total == bytesize::ByteSize::b(0) { continue; } let mnt = mount.fs_mounted_on.clone(); if is_dir_or_subdir_linux(&mnt, "/boot") { continue; } if is_dir_or_subdir_linux(&mnt, "/dev") { continue; } if is_dir_or_subdir_linux(&mnt, "/run") { continue; } if is_dir_or_subdir_linux(&mnt, "/snap") { continue; } if is_dir_or_subdir_linux(&mnt, "/sys") { continue; }
random_line_split
lib.rs
c >= 'a' && c <= 'f') || (c >= 'A' && c <= 'F') ).map(|c| if c >= '0' && c <= '9' { c as u64 - '0' as u64 } else if c >= 'a' && c <= 'f' { 10 + c as u64 - 'a' as u64 } else { 10 + c as u64 - 'A' as u64 } as u32 ) } fn unicode_char<'a>() -> impl Parser<&'a str, Output = Option<char>> { c_hx_do!{ __ <- string(r#"\u"#), d3 <- parse_hex(), d2 <- parse_hex(), d1 <- parse_hex(), d0 <- parse_hex(); { let unicode = d0 + 0x10 * d1 + 0x100 * d2 + 0x1000 * d3; char::try_from(unicode).ok() } } } #[derive(PartialEq)] enum StringPiece<'a > { Ref(&'a str), Char(Option<char>) } fn braced_parser<'a, PBL, P, PBR, O>(pbl: PBL, p: P, pbr: PBR) -> impl Parser<&'a str, Output = O> where PBL: Parser<&'a str>, PBR: Parser<&'a str>, P: Parser<&'a str, Output = O> { between( c_compre![c; c <- pbl, __ <- skip_many(space())], c_compre![c; __ <- skip_many(space()), c <- pbr], p ) } fn string_part<'a>() -> impl Parser<&'a str, Output = Vec<StringPiece<'a >>> { many( choice( ( attempt(take_while1(|c: char| c != '\\' && c != '"' && c != '\n' && c != '\r' && c != '\t') .map(|chars: &str| StringPiece::Ref(chars))), attempt(string("\\\"").map(|_|StringPiece::Ref("\""))), attempt(string("\\\\").map(|_|StringPiece::Ref("\\"))), attempt(string("\\n").map(|_|StringPiece::Ref("\n"))), attempt(string("\\t").map(|_|StringPiece::Ref("\t"))), attempt(string("\\/").map(|_|StringPiece::Ref("/"))), attempt(string("\\r").map(|_|StringPiece::Ref("\r"))), attempt(string("\\f").map(|_|StringPiece::Ref("\u{000c}"))), attempt(string("\\b").map(|_|StringPiece::Ref("\u{0008}"))), attempt(unicode_char().map(|s|StringPiece::Char(s))), ) ) ) } fn string_parser_inner<'a>() -> impl Parser<&'a str, Output = SmolStr> { c_hx_do! { x <- between(char('"'), char('"'), string_part()); { let cap = x.iter().fold(0, |acc, s| acc + match s { StringPiece::Ref(strref) => strref.len(), StringPiece::Char(c) => c.map(|c_inner| c_inner.len_utf8()).unwrap_or(0) } ); if cap <= 22 { let mut buf: [u8; 22] = [0; 22]; let mut offset = 0; for s in x.iter() { match s { StringPiece::Ref(strref) => { for &b in strref.as_bytes() { buf[offset] = b; offset += 1; } }, StringPiece::Char(c) => { if let Some(chr) = c { chr.encode_utf8(&mut buf[offset..]); offset += chr.len_utf8(); } } } } return unsafe { SmolStr::new(str::from_utf8_unchecked(&buf[0..cap])) }; } let mut str = String::with_capacity(cap); for s in x.iter() { match s { StringPiece::Ref(strref) => str.push_str(strref), StringPiece::Char(c) => if let Some(chr) = c { str.push(*chr); } } } SmolStr::new(str) } } } fn string_parser<'a>() -> impl Parser<&'a str, Output = Node> { string_parser_inner().map(|x| Node::String(x)) } fn digit_sequence<'a>() -> impl Parser<&'a str, Output = &'a str> { take_while1(|c: char| c >= '0' && c <= '9') } #[inline(always)] fn power(lhs: f64, rhs: f64) -> f64 { lhs.powf(rhs) } fn trailing_digit_sequence<'a>() -> impl Parser<&'a str, Output = &'a str> { c_hx_do! { __ <- char('.'), rest <- digit_sequence(); rest } }
__ <- satisfy(|c: char| c == 'e' || c == 'E'), sign_char <- optional(satisfy(|c: char| c == '+' || c == '-')), digits <- digit_sequence(); { let sign = match sign_char { Some('-') => -1.0, _ => 1.0 }; let mut acc = 0; for c in digits.as_bytes() { acc = acc * 10 + (c - b'0') as u64; } power(10.0, sign * acc as f64) } } } #[derive(PartialEq, Copy, Clone)] enum NumberPrefix<'a > { LeadingZero, Digits(char, &'a str) } fn leading_zero_parser <'a>() -> impl Parser<&'a str, Output = NumberPrefix<'a >> { char('0').map(|_| NumberPrefix::LeadingZero) } fn leading_digits_parser <'a>() -> impl Parser<&'a str, Output = NumberPrefix<'a >> { c_hx_do! { leading_digit <- satisfy(|c: char| c >= '1' && c <= '9'), digs <- optional(digit_sequence()); NumberPrefix::Digits(leading_digit, digs.unwrap_or("")) } } fn leading_parser <'a>() -> impl Parser<&'a str, Output = NumberPrefix<'a >> { choice(( attempt(leading_digits_parser()), attempt(leading_zero_parser()), )) } fn number_parser<'a>() -> impl Parser<&'a str, Output = Node> { c_hx_do! { minus_sign <- optional(char('-')), leading <- leading_parser(), trail <- optional(trailing_digit_sequence()), exp <- optional(exponent_parser()); { Node::Number({ let mut acc = match leading { NumberPrefix::LeadingZero => 0.0, NumberPrefix::Digits(leading_digit, l_digs) => { let mut l = (leading_digit as u8 - b'0') as u64; for c in l_digs.as_bytes() { l = l * 10 + (c - b'0') as u64; } l as f64 } }; if let Some(t_digs) = trail { let mut divider = 1.0; for c in t_digs.as_bytes() { divider /= 10.0; acc += (c - b'0') as f64 * divider; } } if let Some(exponent) = exp { acc *= exponent; } if let Some(_) = minus_sign { -acc } else { acc } }) } } } fn bool_parser<'a>() -> impl Parser<&'a str, Output = Node> { c_hx_do!{ word <- string("true").or(string("false")); match word { "true" => Node::Boolean(true), _ => Node::Boolean(false) } } } fn null_parser<'a>() -> impl Parser<&'a str, Output = Node> { c_hx_do!{ _word <- string("null"); Node::Null } } macro_rules! ref_parser { ($parser_fn:ident) => { parser(|input| { let _: &mut &str = input; $parser_fn().parse_stream(input).into_result() }) } } fn primitive_parser<'a>() -> impl Parser<&'a str, Output = Node> { let possible_parser = bool_parser() .or(number_parser()) .or(string_parser()) .or(null_parser()) .or(ref_parser!(array_parser)) .or(ref_parser!(dictionary_parser)); c_hx_do! { __ <- skip_many(space()), pars <- possible_parser, ___ <- skip_many(space()); pars } } fn array_parser<'a>() -> impl Parser<&'a str
fn exponent_parser<'a>() -> impl Parser<&'a str, Output = f64> { c_hx_do!{
random_line_split
lib.rs
c >= 'a' && c <= 'f') || (c >= 'A' && c <= 'F') ).map(|c| if c >= '0' && c <= '9' { c as u64 - '0' as u64 } else if c >= 'a' && c <= 'f' { 10 + c as u64 - 'a' as u64 } else { 10 + c as u64 - 'A' as u64 } as u32 ) } fn unicode_char<'a>() -> impl Parser<&'a str, Output = Option<char>> { c_hx_do!{ __ <- string(r#"\u"#), d3 <- parse_hex(), d2 <- parse_hex(), d1 <- parse_hex(), d0 <- parse_hex(); { let unicode = d0 + 0x10 * d1 + 0x100 * d2 + 0x1000 * d3; char::try_from(unicode).ok() } } } #[derive(PartialEq)] enum StringPiece<'a > { Ref(&'a str), Char(Option<char>) } fn braced_parser<'a, PBL, P, PBR, O>(pbl: PBL, p: P, pbr: PBR) -> impl Parser<&'a str, Output = O> where PBL: Parser<&'a str>, PBR: Parser<&'a str>, P: Parser<&'a str, Output = O> { between( c_compre![c; c <- pbl, __ <- skip_many(space())], c_compre![c; __ <- skip_many(space()), c <- pbr], p ) } fn string_part<'a>() -> impl Parser<&'a str, Output = Vec<StringPiece<'a >>> { many( choice( ( attempt(take_while1(|c: char| c != '\\' && c != '"' && c != '\n' && c != '\r' && c != '\t') .map(|chars: &str| StringPiece::Ref(chars))), attempt(string("\\\"").map(|_|StringPiece::Ref("\""))), attempt(string("\\\\").map(|_|StringPiece::Ref("\\"))), attempt(string("\\n").map(|_|StringPiece::Ref("\n"))), attempt(string("\\t").map(|_|StringPiece::Ref("\t"))), attempt(string("\\/").map(|_|StringPiece::Ref("/"))), attempt(string("\\r").map(|_|StringPiece::Ref("\r"))), attempt(string("\\f").map(|_|StringPiece::Ref("\u{000c}"))), attempt(string("\\b").map(|_|StringPiece::Ref("\u{0008}"))), attempt(unicode_char().map(|s|StringPiece::Char(s))), ) ) ) } fn string_parser_inner<'a>() -> impl Parser<&'a str, Output = SmolStr> { c_hx_do! { x <- between(char('"'), char('"'), string_part()); { let cap = x.iter().fold(0, |acc, s| acc + match s { StringPiece::Ref(strref) => strref.len(), StringPiece::Char(c) => c.map(|c_inner| c_inner.len_utf8()).unwrap_or(0) } ); if cap <= 22 { let mut buf: [u8; 22] = [0; 22]; let mut offset = 0; for s in x.iter() { match s { StringPiece::Ref(strref) => { for &b in strref.as_bytes() { buf[offset] = b; offset += 1; } }, StringPiece::Char(c) => { if let Some(chr) = c { chr.encode_utf8(&mut buf[offset..]); offset += chr.len_utf8(); } } } } return unsafe { SmolStr::new(str::from_utf8_unchecked(&buf[0..cap])) }; } let mut str = String::with_capacity(cap); for s in x.iter() { match s { StringPiece::Ref(strref) => str.push_str(strref), StringPiece::Char(c) => if let Some(chr) = c { str.push(*chr); } } } SmolStr::new(str) } } } fn string_parser<'a>() -> impl Parser<&'a str, Output = Node> { string_parser_inner().map(|x| Node::String(x)) } fn digit_sequence<'a>() -> impl Parser<&'a str, Output = &'a str> { take_while1(|c: char| c >= '0' && c <= '9') } #[inline(always)] fn power(lhs: f64, rhs: f64) -> f64 { lhs.powf(rhs) } fn trailing_digit_sequence<'a>() -> impl Parser<&'a str, Output = &'a str> { c_hx_do! { __ <- char('.'), rest <- digit_sequence(); rest } } fn exponent_parser<'a>() -> impl Parser<&'a str, Output = f64> { c_hx_do!{ __ <- satisfy(|c: char| c == 'e' || c == 'E'), sign_char <- optional(satisfy(|c: char| c == '+' || c == '-')), digits <- digit_sequence(); { let sign = match sign_char { Some('-') => -1.0, _ => 1.0 }; let mut acc = 0; for c in digits.as_bytes() { acc = acc * 10 + (c - b'0') as u64; } power(10.0, sign * acc as f64) } } } #[derive(PartialEq, Copy, Clone)] enum NumberPrefix<'a > { LeadingZero, Digits(char, &'a str) } fn leading_zero_parser <'a>() -> impl Parser<&'a str, Output = NumberPrefix<'a >> { char('0').map(|_| NumberPrefix::LeadingZero) } fn leading_digits_parser <'a>() -> impl Parser<&'a str, Output = NumberPrefix<'a >> { c_hx_do! { leading_digit <- satisfy(|c: char| c >= '1' && c <= '9'), digs <- optional(digit_sequence()); NumberPrefix::Digits(leading_digit, digs.unwrap_or("")) } } fn leading_parser <'a>() -> impl Parser<&'a str, Output = NumberPrefix<'a >> { choice(( attempt(leading_digits_parser()), attempt(leading_zero_parser()), )) } fn number_parser<'a>() -> impl Parser<&'a str, Output = Node> { c_hx_do! { minus_sign <- optional(char('-')), leading <- leading_parser(), trail <- optional(trailing_digit_sequence()), exp <- optional(exponent_parser()); { Node::Number({ let mut acc = match leading { NumberPrefix::LeadingZero => 0.0, NumberPrefix::Digits(leading_digit, l_digs) => { let mut l = (leading_digit as u8 - b'0') as u64; for c in l_digs.as_bytes() { l = l * 10 + (c - b'0') as u64; } l as f64 } }; if let Some(t_digs) = trail { let mut divider = 1.0; for c in t_digs.as_bytes() { divider /= 10.0; acc += (c - b'0') as f64 * divider; } } if let Some(exponent) = exp { acc *= exponent; } if let Some(_) = minus_sign { -acc } else { acc } }) } } } fn bool_parser<'a>() -> impl Parser<&'a str, Output = Node>
fn null_parser<'a>() -> impl Parser<&'a str, Output = Node> { c_hx_do!{ _word <- string("null"); Node::Null } } macro_rules! ref_parser { ($parser_fn:ident) => { parser(|input| { let _: &mut &str = input; $parser_fn().parse_stream(input).into_result() }) } } fn primitive_parser<'a>() -> impl Parser<&'a str, Output = Node> { let possible_parser = bool_parser() .or(number_parser()) .or(string_parser()) .or(null_parser()) .or(ref_parser!(array_parser)) .or(ref_parser!(dictionary_parser)); c_hx_do! { __ <- skip_many(space()), pars <- possible_parser, ___ <- skip_many(space()); pars } } fn array_parser<'a>() -> impl Parser<&'
{ c_hx_do!{ word <- string("true").or(string("false")); match word { "true" => Node::Boolean(true), _ => Node::Boolean(false) } } }
identifier_body