file_name
large_stringlengths 4
140
| prefix
large_stringlengths 0
12.1k
| suffix
large_stringlengths 0
12k
| middle
large_stringlengths 0
7.51k
| fim_type
large_stringclasses 4
values |
---|---|---|---|---|
index.ts | (v: T) => TResult | Promise<TResult>,
onRejected?: (reason: any) => Promise<never>,
): Promise<TResult> {
return this._promise.then<TResult>(onFulfilled, onRejected);
}
public catch<TResult>(onRejected?: (reason: any) => Promise<TResult>): Promise<TResult | T> {
return this._promise.catch(onRejected);
}
public finally(onFinally?: (() => void) | null): Promise<T> {
return this._promise.finally(onFinally);
}
// tslint:enable:promise-function-async
// tslint:enable:async-suffix
get [Symbol.toStringTag](): 'Promise' {
return this._promise[Symbol.toStringTag];
}
}
export class | {
protected _abiEncoderByFunctionSignature: AbiEncoderByFunctionSignature;
protected _web3Wrapper: Web3Wrapper;
public abi: ContractAbi;
public address: string;
public contractName: string;
public constructorArgs: any[] = [];
public _deployedBytecodeIfExists?: Buffer;
private _evmIfExists?: VM;
private _evmAccountIfExists?: Buffer;
protected static _formatABIDataItemList(
abis: DataItem[],
values: any[],
formatter: (type: string, value: any) => any,
): any {
return values.map((value: any, i: number) => formatABIDataItem(abis[i], value, formatter));
}
protected static _lowercaseAddress(type: string, value: string): string {
return type === 'address' ? value.toLowerCase() : value;
}
protected static _bigNumberToString(_type: string, value: any): any {
return BigNumber.isBigNumber(value) ? value.toString() : value;
}
protected static _lookupConstructorAbi(abi: ContractAbi): ConstructorAbi {
const constructorAbiIfExists = abi.find(
(abiDefinition: AbiDefinition) => abiDefinition.type === AbiType.Constructor,
// tslint:disable-next-line:no-unnecessary-type-assertion
) as ConstructorAbi | undefined;
if (constructorAbiIfExists !== undefined) {
return constructorAbiIfExists;
} else {
// If the constructor is not explicitly defined, it won't be included in the ABI. It is
// still callable however, so we construct what the ABI would look like were it to exist.
const defaultConstructorAbi: ConstructorAbi = {
type: AbiType.Constructor,
stateMutability: 'nonpayable',
payable: false,
inputs: [],
};
return defaultConstructorAbi;
}
}
protected static _throwIfCallResultIsRevertError(rawCallResult: string): void {
// Try to decode the call result as a revert error.
let revert: RevertError;
try {
revert = decodeBytesAsRevertError(rawCallResult);
} catch (err) {
// Can't decode it as a revert error, so assume it didn't revert.
return;
}
throw revert;
}
protected static _throwIfThrownErrorIsRevertError(error: Error): void {
// Try to decode a thrown error.
let revertError: RevertError;
try {
revertError = decodeThrownErrorAsRevertError(error);
} catch (err) {
// Can't decode it.
return;
}
// Re-cast StringRevertErrors as plain Errors for backwards-compatibility.
if (revertError instanceof StringRevertError) {
throw new Error(revertError.values.message as string);
}
throw revertError;
}
protected static _throwIfUnexpectedEmptyCallResult(rawCallResult: string, methodAbi: AbiEncoder.Method): void {
// With live nodes, we will receive an empty call result if:
// 1. The function has no return value.
// 2. The contract reverts without data.
// 3. The contract reverts with an invalid opcode (`assert(false)` or `invalid()`).
if (!rawCallResult || rawCallResult === '0x') {
const returnValueDataItem = methodAbi.getReturnValueDataItem();
if (returnValueDataItem.components === undefined || returnValueDataItem.components.length === 0) {
// Expected no result (which makes it hard to tell if the call reverted).
return;
}
throw new Error(`Function "${methodAbi.getSignature()}" reverted with no data`);
}
}
// Throws if the given arguments cannot be safely/correctly encoded based on
// the given inputAbi. An argument may not be considered safely encodeable
// if it overflows the corresponding Solidity type, there is a bug in the
// encoder, or the encoder performs unsafe type coercion.
public static strictArgumentEncodingCheck(inputAbi: DataItem[], args: any[]): string {
const abiEncoder = AbiEncoder.create(inputAbi);
const params = abiUtils.parseEthersParams(inputAbi);
const rawEncoded = abiEncoder.encode(args);
const rawDecoded = abiEncoder.decodeAsArray(rawEncoded);
for (let i = 0; i < rawDecoded.length; i++) {
const original = args[i];
const decoded = rawDecoded[i];
if (!abiUtils.isAbiDataEqual(params.names[i], params.types[i], original, decoded)) {
throw new Error(
`Cannot safely encode argument: ${params.names[i]} (${original}) of type ${
params.types[i]
}. (Possible type overflow or other encoding error)`,
);
}
}
return rawEncoded;
}
protected static async _applyDefaultsToContractTxDataAsync<T extends Partial<TxData | TxDataPayable>>(
txData: T,
estimateGasAsync?: (txData: T) => Promise<number>,
): Promise<TxData> {
const txDataWithDefaults = BaseContract._removeUndefinedProperties<T>(txData);
if (txDataWithDefaults.gas === undefined && estimateGasAsync !== undefined) {
txDataWithDefaults.gas = await estimateGasAsync(txDataWithDefaults);
}
if (txDataWithDefaults.from !== undefined) {
txDataWithDefaults.from = txDataWithDefaults.from.toLowerCase();
}
return txDataWithDefaults as TxData;
}
protected static _assertCallParams(callData: Partial<CallData>, defaultBlock?: BlockParam): void {
assert.doesConformToSchema('callData', callData, schemas.callDataSchema, [
schemas.addressSchema,
schemas.numberSchema,
schemas.jsNumber,
]);
if (defaultBlock !== undefined) {
assert.isBlockParam('defaultBlock', defaultBlock);
}
}
private static _removeUndefinedProperties<T>(props: any): T {
const clonedProps = { ...props };
Object.keys(clonedProps).forEach(key => clonedProps[key] === undefined && delete clonedProps[key]);
return clonedProps;
}
protected _promiseWithTransactionHash(
txHashPromise: Promise<string>,
opts: AwaitTransactionSuccessOpts,
): PromiseWithTransactionHash<TransactionReceiptWithDecodedLogs> {
return new PromiseWithTransactionHash<TransactionReceiptWithDecodedLogs>(
txHashPromise,
(async (): Promise<TransactionReceiptWithDecodedLogs> => {
// When the transaction hash resolves, wait for it to be mined.
return this._web3Wrapper.awaitTransactionSuccessAsync(
await txHashPromise,
opts.pollingIntervalMs,
opts.timeoutMs,
);
})(),
);
}
protected async _applyDefaultsToTxDataAsync<T extends Partial<TxData | TxDataPayable>>(
txData: T,
estimateGasAsync?: (txData: T) => Promise<number>,
): Promise<TxData> {
// Gas amount sourced with the following priorities:
// 1. Optional param passed in to public method call
// 2. Global config passed in at library instantiation
// 3. Gas estimate calculation + safety margin
// tslint:disable-next-line:no-object-literal-type-assertion
const txDataWithDefaults = {
to: this.address,
...this._web3Wrapper.getContractDefaults(),
...BaseContract._removeUndefinedProperties(txData),
} as T;
if (txDataWithDefaults.gas === undefined && estimateGasAsync !== undefined) {
txDataWithDefaults.gas = await estimateGasAsync(txDataWithDefaults);
}
if (txDataWithDefaults.from !== undefined) {
txDataWithDefaults.from = txDataWithDefaults.from.toLowerCase();
}
return txDataWithDefaults as TxData;
}
protected async _evmExecAsync(encodedData: string): Promise<string> {
const encodedDataBytes = Buffer.from(encodedData.substr(2), 'hex');
const addressBuf = Buffer.from(this.address.substr(2), 'hex');
// should only run once, the first time it is called
if (this._evmIfExists === undefined) {
const vm = new VM({});
const psm = new PStateManager(vm.stateManager);
// create an account with 1 ETH
const accountPk = Buffer.from(ARBITRARY_PRIVATE_KEY, 'hex');
const accountAddress = util.privateToAddress(accountPk);
const account = new Account({ balance: 1e18 });
await psm.putAccount(accountAddress, account);
// 'deploy' the contract
if (this._deployedBytecodeIfExists === undefined | BaseContract | identifier_name |
index.ts | ContractEvent,
SendTransactionOpts,
AwaitTransactionSuccessOpts,
ContractFunctionObj,
ContractTxFunctionObj,
SubscriptionErrors,
} from './types';
export interface AbiEncoderByFunctionSignature {
[key: string]: AbiEncoder.Method;
}
const ARBITRARY_PRIVATE_KEY = 'e331b6d69882b4cb4ea581d88e0b604039a3de5967688d3dcffdd2270c0fd109';
// tslint:disable: max-classes-per-file
/**
* @dev A promise-compatible type that exposes a `txHash` field.
* Not used by BaseContract, but generated contracts will return it in
* `awaitTransactionSuccessAsync()`.
* Maybe there's a better place for this.
*/
export class PromiseWithTransactionHash<T> implements Promise<T> {
public readonly txHashPromise: Promise<string>;
private readonly _promise: Promise<T>;
constructor(txHashPromise: Promise<string>, promise: Promise<T>) {
this.txHashPromise = txHashPromise;
this._promise = promise;
}
// tslint:disable:promise-function-async
// tslint:disable:async-suffix
public then<TResult>(
onFulfilled?: (v: T) => TResult | Promise<TResult>,
onRejected?: (reason: any) => Promise<never>,
): Promise<TResult> {
return this._promise.then<TResult>(onFulfilled, onRejected);
}
public catch<TResult>(onRejected?: (reason: any) => Promise<TResult>): Promise<TResult | T> {
return this._promise.catch(onRejected);
}
public finally(onFinally?: (() => void) | null): Promise<T> {
return this._promise.finally(onFinally);
}
// tslint:enable:promise-function-async
// tslint:enable:async-suffix
get [Symbol.toStringTag](): 'Promise' {
return this._promise[Symbol.toStringTag];
}
}
export class BaseContract {
protected _abiEncoderByFunctionSignature: AbiEncoderByFunctionSignature;
protected _web3Wrapper: Web3Wrapper;
public abi: ContractAbi;
public address: string;
public contractName: string;
public constructorArgs: any[] = [];
public _deployedBytecodeIfExists?: Buffer;
private _evmIfExists?: VM;
private _evmAccountIfExists?: Buffer;
protected static _formatABIDataItemList(
abis: DataItem[],
values: any[],
formatter: (type: string, value: any) => any,
): any {
return values.map((value: any, i: number) => formatABIDataItem(abis[i], value, formatter));
}
protected static _lowercaseAddress(type: string, value: string): string {
return type === 'address' ? value.toLowerCase() : value;
}
protected static _bigNumberToString(_type: string, value: any): any {
return BigNumber.isBigNumber(value) ? value.toString() : value;
}
protected static _lookupConstructorAbi(abi: ContractAbi): ConstructorAbi {
const constructorAbiIfExists = abi.find(
(abiDefinition: AbiDefinition) => abiDefinition.type === AbiType.Constructor,
// tslint:disable-next-line:no-unnecessary-type-assertion
) as ConstructorAbi | undefined;
if (constructorAbiIfExists !== undefined) {
return constructorAbiIfExists;
} else {
// If the constructor is not explicitly defined, it won't be included in the ABI. It is
// still callable however, so we construct what the ABI would look like were it to exist.
const defaultConstructorAbi: ConstructorAbi = {
type: AbiType.Constructor,
stateMutability: 'nonpayable',
payable: false,
inputs: [],
};
return defaultConstructorAbi;
}
}
protected static _throwIfCallResultIsRevertError(rawCallResult: string): void {
// Try to decode the call result as a revert error.
let revert: RevertError;
try {
revert = decodeBytesAsRevertError(rawCallResult);
} catch (err) {
// Can't decode it as a revert error, so assume it didn't revert.
return;
}
throw revert;
}
protected static _throwIfThrownErrorIsRevertError(error: Error): void {
// Try to decode a thrown error.
let revertError: RevertError;
try {
revertError = decodeThrownErrorAsRevertError(error);
} catch (err) {
// Can't decode it.
return;
}
// Re-cast StringRevertErrors as plain Errors for backwards-compatibility.
if (revertError instanceof StringRevertError) {
throw new Error(revertError.values.message as string);
}
throw revertError;
}
protected static _throwIfUnexpectedEmptyCallResult(rawCallResult: string, methodAbi: AbiEncoder.Method): void {
// With live nodes, we will receive an empty call result if:
// 1. The function has no return value.
// 2. The contract reverts without data.
// 3. The contract reverts with an invalid opcode (`assert(false)` or `invalid()`).
if (!rawCallResult || rawCallResult === '0x') {
const returnValueDataItem = methodAbi.getReturnValueDataItem();
if (returnValueDataItem.components === undefined || returnValueDataItem.components.length === 0) {
// Expected no result (which makes it hard to tell if the call reverted).
return;
}
throw new Error(`Function "${methodAbi.getSignature()}" reverted with no data`);
}
}
// Throws if the given arguments cannot be safely/correctly encoded based on
// the given inputAbi. An argument may not be considered safely encodeable
// if it overflows the corresponding Solidity type, there is a bug in the
// encoder, or the encoder performs unsafe type coercion.
public static strictArgumentEncodingCheck(inputAbi: DataItem[], args: any[]): string {
const abiEncoder = AbiEncoder.create(inputAbi);
const params = abiUtils.parseEthersParams(inputAbi);
const rawEncoded = abiEncoder.encode(args);
const rawDecoded = abiEncoder.decodeAsArray(rawEncoded);
for (let i = 0; i < rawDecoded.length; i++) {
const original = args[i];
const decoded = rawDecoded[i];
if (!abiUtils.isAbiDataEqual(params.names[i], params.types[i], original, decoded)) {
throw new Error(
`Cannot safely encode argument: ${params.names[i]} (${original}) of type ${
params.types[i]
}. (Possible type overflow or other encoding error)`,
);
}
}
return rawEncoded;
}
protected static async _applyDefaultsToContractTxDataAsync<T extends Partial<TxData | TxDataPayable>>(
txData: T,
estimateGasAsync?: (txData: T) => Promise<number>,
): Promise<TxData> {
const txDataWithDefaults = BaseContract._removeUndefinedProperties<T>(txData);
if (txDataWithDefaults.gas === undefined && estimateGasAsync !== undefined) {
txDataWithDefaults.gas = await estimateGasAsync(txDataWithDefaults);
}
if (txDataWithDefaults.from !== undefined) {
txDataWithDefaults.from = txDataWithDefaults.from.toLowerCase();
}
return txDataWithDefaults as TxData;
}
protected static _assertCallParams(callData: Partial<CallData>, defaultBlock?: BlockParam): void {
assert.doesConformToSchema('callData', callData, schemas.callDataSchema, [
schemas.addressSchema,
schemas.numberSchema,
schemas.jsNumber,
]);
if (defaultBlock !== undefined) {
assert.isBlockParam('defaultBlock', defaultBlock);
}
}
private static _removeUndefinedProperties<T>(props: any): T {
const clonedProps = { ...props };
Object.keys(clonedProps).forEach(key => clonedProps[key] === undefined && delete clonedProps[key]);
return clonedProps;
}
protected _promiseWithTransactionHash(
txHashPromise: Promise<string>,
opts: AwaitTransactionSuccessOpts,
): PromiseWithTransactionHash<TransactionReceiptWithDecodedLogs> {
return new PromiseWithTransactionHash<TransactionReceiptWithDecodedLogs>(
txHashPromise,
(async (): Promise<TransactionReceiptWithDecodedLogs> => {
// When the transaction hash resolves, wait for it to be mined.
return this._web3Wrapper.awaitTransactionSuccessAsync(
await txHashPromise,
opts.pollingIntervalMs,
opts.timeoutMs,
);
})(),
);
}
protected async _applyDefaultsToTxDataAsync<T extends Partial<TxData | TxDataPayable>>(
txData: T,
estimateGasAsync?: (txData: T) => Promise<number>,
): Promise<TxData> {
// Gas amount sourced with the following priorities:
// 1. Optional param passed in to public method call
// 2. Global config passed in at library instantiation
// 3. Gas estimate calculation + safety margin
// tslint:disable-next-line:no-object-literal-type-assertion
const txDataWithDefaults = {
to: this.address,
...this._ | export { | random_line_split |
|
cluster.go | PgStorageSpec `json:replicastorage`
BackrestStorage PgStorageSpec `json:backreststorage`
// Resources behaves just like the "Requests" section of a Kubernetes
// container definition. You can set individual items such as "cpu" and
// "memory", e.g. "{ cpu: "0.5", memory: "2Gi" }"
Resources v1.ResourceList `json:"resources"`
// Limits stores the CPU/memory limits to use with PostgreSQL instances
//
// A long note on memory limits.
//
// We want to avoid the OOM killer coming for the PostgreSQL process or any
// of their backends per lots of guidance from the PostgreSQL documentation.
// Based on Kubernetes' behavior with limits, the best thing is to not set
// them. However, if they ever do set, we suggest that you have
// Request == Limit to get the Guaranteed QoS
//
// Guaranteed QoS prevents a backend from being first in line to be killed if
// the *Node* has memory pressure, but if there is, say
// a runaway client backend that causes the *Pod* to exceed its memory
// limit, a backend can still be killed by the OOM killer, which is not
// great.
//
// As such, given the choice, the preference is for the Pod to be evicted
// and have a failover event, vs. having an individual client backend killed
// and causing potential "bad things."
//
// For more info on PostgreSQL and Kubernetes memory management, see:
//
// https://www.postgresql.org/docs/current/kernel-resources.html#LINUX-MEMORY-OVERCOMMIT
// https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/#how-pods-with-resource-limits-are-run
Limits v1.ResourceList `json:"limits"`
// BackrestResources, if specified, contains the container request resources
// for the pgBackRest Deployment for this PostgreSQL cluster
BackrestResources v1.ResourceList `json:"backrestResources"`
// BackrestLimits, if specified, contains the container resource limits
// for the pgBackRest Deployment for this PostgreSQL cluster
BackrestLimits v1.ResourceList `json:"backrestLimits"`
// PgBouncer contains all of the settings to properly maintain a pgBouncer
// implementation
PgBouncer PgBouncerSpec `json:"pgBouncer"`
User string `json:"user"`
Database string `json:"database"`
Replicas string `json:"replicas"`
UserSecretName string `json:"usersecretname"`
RootSecretName string `json:"rootsecretname"`
PrimarySecretName string `json:"primarysecretname"`
CollectSecretName string `json:"collectSecretName"`
Status string `json:"status"`
CustomConfig string `json:"customconfig"`
UserLabels map[string]string `json:"userlabels"`
PodAntiAffinity PodAntiAffinitySpec `json:"podAntiAffinity"`
SyncReplication *bool `json:"syncReplication"`
BackrestS3Bucket string `json:"backrestS3Bucket"`
BackrestS3Region string `json:"backrestS3Region"`
BackrestS3Endpoint string `json:"backrestS3Endpoint"`
BackrestS3URIStyle string `json:"backrestS3URIStyle"`
BackrestS3VerifyTLS string `json:"backrestS3VerifyTLS"`
BackrestRepoPath string `json:"backrestRepoPath"`
TablespaceMounts map[string]PgStorageSpec `json:"tablespaceMounts"`
TLS TLSSpec `json:"tls"`
TLSOnly bool `json:"tlsOnly"`
Standby bool `json:"standby"`
Shutdown bool `json:"shutdown"`
PGDataSource PGDataSourceSpec `json:"pgDataSource"`
}
// PGDataSourceSpec defines the data source that should be used to populate the initial PGDATA
// directory when bootstrapping a new PostgreSQL cluster
// swagger:ignore
type PGDataSourceSpec struct {
RestoreFrom string `json:"restoreFrom"`
RestoreOpts string `json:"restoreOpts"`
}
// PgclusterList is the CRD that defines a Crunchy PG Cluster List
// swagger:ignore
// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
type PgclusterList struct {
metav1.TypeMeta `json:",inline"`
metav1.ListMeta `json:"metadata"`
Items []Pgcluster `json:"items"`
}
// PgclusterStatus is the CRD that defines PG Cluster Status
// swagger:ignore
type PgclusterStatus struct {
State PgclusterState `json:"state,omitempty"`
Message string `json:"message,omitempty"`
}
// PgclusterState is the crd that defines PG Cluster Stage
// swagger:ignore
type PgclusterState string
// PodAntiAffinityDeployment distinguishes between the different types of
// Deployments that can leverage PodAntiAffinity
type PodAntiAffinityDeployment int
// PodAntiAffinityType defines the different types of type of anti-affinity rules applied to pg
// clusters when utilizing the default pod anti-affinity rules provided by the PostgreSQL Operator,
// which are enabled for a new pg cluster by default. Valid Values include "required" for
// requiredDuringSchedulingIgnoredDuringExecution anti-affinity, "preferred" for
// preferredDuringSchedulingIgnoredDuringExecution anti-affinity, and "disabled" to disable the
// default pod anti-affinity rules for the pg cluster all together.
type PodAntiAffinityType string
// PodAntiAffinitySpec provides multiple configurations for how pod
// anti-affinity can be set.
// - "Default" is the default rule that applies to all Pods that are a part of
// the PostgreSQL cluster
// - "PgBackrest" applies just to the pgBackRest repository Pods in said
// Deployment
// - "PgBouncer" applies to just pgBouncer Pods in said Deployment
// swaggier:ignore
type PodAntiAffinitySpec struct {
Default PodAntiAffinityType `json:"default"`
PgBackRest PodAntiAffinityType `json:"pgBackRest"`
PgBouncer PodAntiAffinityType `json:"pgBouncer"`
}
// PgBouncerSpec is a struct that is used within the Cluster specification that
// provides the attributes for managing a PgBouncer implementation, including:
// - is it enabled?
// - what resources it should consume
// - the total number of replicas
type PgBouncerSpec struct {
// Replicas represents the total number of Pods to deploy with pgBouncer,
// which effectively enables/disables the pgBouncer.
//
// if it is set to 0 or less, it is disabled.
//
// if it is set to 1 or more, it is enabled
Replicas int32 `json:"replicas"`
// Resources, if specified, contains the container request resources
// for any pgBouncer Deployments that are part of a PostgreSQL cluster
Resources v1.ResourceList `json:"resources"`
// Limits, if specified, contains the container resource limits
// for any pgBouncer Deployments that are part of a PostgreSQL cluster
Limits v1.ResourceList `json:"limits"`
}
// Enabled returns true if the pgBouncer is enabled for the cluster, i.e. there
// is at least one replica set
func (s *PgBouncerSpec) Enabled() bool {
return s.Replicas > 0
}
// TLSSpec contains the information to set up a TLS-enabled PostgreSQL cluster
type TLSSpec struct {
// CASecret contains the name of the secret to use as the trusted CA for the
// TLSSecret
// This is our own format and should contain at least one key: "ca.crt"
// It can also contain a key "ca.crl" which is the certificate revocation list
CASecret string `json:"caSecret"`
// ReplicationTLSSecret contains the name of the secret that specifies a TLS
// keypair that can be used by the replication user (e.g. "primaryuser") to
// perform certificate based authentication between replicas.
// The keypair must be considered valid by the CA specified in the CASecret
ReplicationTLSSecret string `json:"replicationTLSSecret"`
// TLSSecret contains the name of the secret to use that contains the TLS
// keypair for the PostgreSQL server
// This follows the Kubernetes secret format ("kubernetes.io/tls") which has
// two keys: tls.crt and tls.key
TLSSecret string `json:"tlsSecret"`
}
// IsTLSEnabled returns true if the cluster is TLS enabled, i.e. both the TLS
// secret name and the CA secret name are available
func (t TLSSpec) IsTLSEnabled() bool {
return (t.TLSSecret != "" && t.CASecret != "")
}
const (
// PgclusterStateCreated ...
PgclusterStateCreated PgclusterState = "pgcluster Created"
// PgclusterStateProcessed ...
PgclusterStateProcessed PgclusterState = "pgcluster Processed"
// PgclusterStateInitialized ...
PgclusterStateInitialized PgclusterState = "pgcluster Initialized" | random_line_split |
||
cluster.go | badgerport"`
ExporterPort string `json:"exporterport"`
PrimaryStorage PgStorageSpec `json:primarystorage`
WALStorage PgStorageSpec `json:walstorage`
ArchiveStorage PgStorageSpec `json:archivestorage`
ReplicaStorage PgStorageSpec `json:replicastorage`
BackrestStorage PgStorageSpec `json:backreststorage`
// Resources behaves just like the "Requests" section of a Kubernetes
// container definition. You can set individual items such as "cpu" and
// "memory", e.g. "{ cpu: "0.5", memory: "2Gi" }"
Resources v1.ResourceList `json:"resources"`
// Limits stores the CPU/memory limits to use with PostgreSQL instances
//
// A long note on memory limits.
//
// We want to avoid the OOM killer coming for the PostgreSQL process or any
// of their backends per lots of guidance from the PostgreSQL documentation.
// Based on Kubernetes' behavior with limits, the best thing is to not set
// them. However, if they ever do set, we suggest that you have
// Request == Limit to get the Guaranteed QoS
//
// Guaranteed QoS prevents a backend from being first in line to be killed if
// the *Node* has memory pressure, but if there is, say
// a runaway client backend that causes the *Pod* to exceed its memory
// limit, a backend can still be killed by the OOM killer, which is not
// great.
//
// As such, given the choice, the preference is for the Pod to be evicted
// and have a failover event, vs. having an individual client backend killed
// and causing potential "bad things."
//
// For more info on PostgreSQL and Kubernetes memory management, see:
//
// https://www.postgresql.org/docs/current/kernel-resources.html#LINUX-MEMORY-OVERCOMMIT
// https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/#how-pods-with-resource-limits-are-run
Limits v1.ResourceList `json:"limits"`
// BackrestResources, if specified, contains the container request resources
// for the pgBackRest Deployment for this PostgreSQL cluster
BackrestResources v1.ResourceList `json:"backrestResources"`
// BackrestLimits, if specified, contains the container resource limits
// for the pgBackRest Deployment for this PostgreSQL cluster
BackrestLimits v1.ResourceList `json:"backrestLimits"`
// PgBouncer contains all of the settings to properly maintain a pgBouncer
// implementation
PgBouncer PgBouncerSpec `json:"pgBouncer"`
User string `json:"user"`
Database string `json:"database"`
Replicas string `json:"replicas"`
UserSecretName string `json:"usersecretname"`
RootSecretName string `json:"rootsecretname"`
PrimarySecretName string `json:"primarysecretname"`
CollectSecretName string `json:"collectSecretName"`
Status string `json:"status"`
CustomConfig string `json:"customconfig"`
UserLabels map[string]string `json:"userlabels"`
PodAntiAffinity PodAntiAffinitySpec `json:"podAntiAffinity"`
SyncReplication *bool `json:"syncReplication"`
BackrestS3Bucket string `json:"backrestS3Bucket"`
BackrestS3Region string `json:"backrestS3Region"`
BackrestS3Endpoint string `json:"backrestS3Endpoint"`
BackrestS3URIStyle string `json:"backrestS3URIStyle"`
BackrestS3VerifyTLS string `json:"backrestS3VerifyTLS"`
BackrestRepoPath string `json:"backrestRepoPath"`
TablespaceMounts map[string]PgStorageSpec `json:"tablespaceMounts"`
TLS TLSSpec `json:"tls"`
TLSOnly bool `json:"tlsOnly"`
Standby bool `json:"standby"`
Shutdown bool `json:"shutdown"`
PGDataSource PGDataSourceSpec `json:"pgDataSource"`
}
// PGDataSourceSpec defines the data source that should be used to populate the initial PGDATA
// directory when bootstrapping a new PostgreSQL cluster
// swagger:ignore
type PGDataSourceSpec struct {
RestoreFrom string `json:"restoreFrom"`
RestoreOpts string `json:"restoreOpts"`
}
// PgclusterList is the CRD that defines a Crunchy PG Cluster List
// swagger:ignore
// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
type PgclusterList struct {
metav1.TypeMeta `json:",inline"`
metav1.ListMeta `json:"metadata"`
Items []Pgcluster `json:"items"`
}
// PgclusterStatus is the CRD that defines PG Cluster Status
// swagger:ignore
type PgclusterStatus struct {
State PgclusterState `json:"state,omitempty"`
Message string `json:"message,omitempty"`
}
// PgclusterState is the crd that defines PG Cluster Stage
// swagger:ignore
type PgclusterState string
// PodAntiAffinityDeployment distinguishes between the different types of
// Deployments that can leverage PodAntiAffinity
type PodAntiAffinityDeployment int
// PodAntiAffinityType defines the different types of type of anti-affinity rules applied to pg
// clusters when utilizing the default pod anti-affinity rules provided by the PostgreSQL Operator,
// which are enabled for a new pg cluster by default. Valid Values include "required" for
// requiredDuringSchedulingIgnoredDuringExecution anti-affinity, "preferred" for
// preferredDuringSchedulingIgnoredDuringExecution anti-affinity, and "disabled" to disable the
// default pod anti-affinity rules for the pg cluster all together.
type PodAntiAffinityType string
// PodAntiAffinitySpec provides multiple configurations for how pod
// anti-affinity can be set.
// - "Default" is the default rule that applies to all Pods that are a part of
// the PostgreSQL cluster
// - "PgBackrest" applies just to the pgBackRest repository Pods in said
// Deployment
// - "PgBouncer" applies to just pgBouncer Pods in said Deployment
// swaggier:ignore
type PodAntiAffinitySpec struct {
Default PodAntiAffinityType `json:"default"`
PgBackRest PodAntiAffinityType `json:"pgBackRest"`
PgBouncer PodAntiAffinityType `json:"pgBouncer"`
}
// PgBouncerSpec is a struct that is used within the Cluster specification that
// provides the attributes for managing a PgBouncer implementation, including:
// - is it enabled?
// - what resources it should consume
// - the total number of replicas
type PgBouncerSpec struct {
// Replicas represents the total number of Pods to deploy with pgBouncer,
// which effectively enables/disables the pgBouncer.
//
// if it is set to 0 or less, it is disabled.
//
// if it is set to 1 or more, it is enabled
Replicas int32 `json:"replicas"`
// Resources, if specified, contains the container request resources
// for any pgBouncer Deployments that are part of a PostgreSQL cluster
Resources v1.ResourceList `json:"resources"`
// Limits, if specified, contains the container resource limits
// for any pgBouncer Deployments that are part of a PostgreSQL cluster
Limits v1.ResourceList `json:"limits"`
}
// Enabled returns true if the pgBouncer is enabled for the cluster, i.e. there
// is at least one replica set
func (s *PgBouncerSpec) Enabled() bool {
return s.Replicas > 0
}
// TLSSpec contains the information to set up a TLS-enabled PostgreSQL cluster
type TLSSpec struct {
// CASecret contains the name of the secret to use as the trusted CA for the
// TLSSecret
// This is our own format and should contain at least one key: "ca.crt"
// It can also contain a key "ca.crl" which is the certificate revocation list
CASecret string `json:"caSecret"`
// ReplicationTLSSecret contains the name of the secret that specifies a TLS
// keypair that can be used by the replication user (e.g. "primaryuser") to
// perform certificate based authentication between replicas.
// The keypair must be considered valid by the CA specified in the CASecret
ReplicationTLSSecret string `json:"replicationTLSSecret"`
// TLSSecret contains the name of the secret to use that contains the TLS
// keypair for the PostgreSQL server
// This follows the Kubernetes secret format ("kubernetes.io/tls") which has
// two keys: tls.crt and tls.key
TLSSecret string `json:"tlsSecret"`
}
// IsTLSEnabled returns true if the cluster is TLS enabled, i.e. both the TLS
// secret name and the CA secret name are available
func (t TLSSpec) IsTLSEnabled() bool | {
return (t.TLSSecret != "" && t.CASecret != "")
} | identifier_body |
|
cluster.go | Crunchy PG Cluster Spec
// swagger:ignore
type PgclusterSpec struct {
Namespace string `json:"namespace"`
Name string `json:"name"`
ClusterName string `json:"clustername"`
Policies string `json:"policies"`
CCPImage string `json:"ccpimage"`
CCPImageTag string `json:"ccpimagetag"`
CCPImagePrefix string `json:"ccpimageprefix"`
PGOImagePrefix string `json:"pgoimageprefix"`
Port string `json:"port"`
PGBadgerPort string `json:"pgbadgerport"`
ExporterPort string `json:"exporterport"`
PrimaryStorage PgStorageSpec `json:primarystorage`
WALStorage PgStorageSpec `json:walstorage`
ArchiveStorage PgStorageSpec `json:archivestorage`
ReplicaStorage PgStorageSpec `json:replicastorage`
BackrestStorage PgStorageSpec `json:backreststorage`
// Resources behaves just like the "Requests" section of a Kubernetes
// container definition. You can set individual items such as "cpu" and
// "memory", e.g. "{ cpu: "0.5", memory: "2Gi" }"
Resources v1.ResourceList `json:"resources"`
// Limits stores the CPU/memory limits to use with PostgreSQL instances
//
// A long note on memory limits.
//
// We want to avoid the OOM killer coming for the PostgreSQL process or any
// of their backends per lots of guidance from the PostgreSQL documentation.
// Based on Kubernetes' behavior with limits, the best thing is to not set
// them. However, if they ever do set, we suggest that you have
// Request == Limit to get the Guaranteed QoS
//
// Guaranteed QoS prevents a backend from being first in line to be killed if
// the *Node* has memory pressure, but if there is, say
// a runaway client backend that causes the *Pod* to exceed its memory
// limit, a backend can still be killed by the OOM killer, which is not
// great.
//
// As such, given the choice, the preference is for the Pod to be evicted
// and have a failover event, vs. having an individual client backend killed
// and causing potential "bad things."
//
// For more info on PostgreSQL and Kubernetes memory management, see:
//
// https://www.postgresql.org/docs/current/kernel-resources.html#LINUX-MEMORY-OVERCOMMIT
// https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/#how-pods-with-resource-limits-are-run
Limits v1.ResourceList `json:"limits"`
// BackrestResources, if specified, contains the container request resources
// for the pgBackRest Deployment for this PostgreSQL cluster
BackrestResources v1.ResourceList `json:"backrestResources"`
// BackrestLimits, if specified, contains the container resource limits
// for the pgBackRest Deployment for this PostgreSQL cluster
BackrestLimits v1.ResourceList `json:"backrestLimits"`
// PgBouncer contains all of the settings to properly maintain a pgBouncer
// implementation
PgBouncer PgBouncerSpec `json:"pgBouncer"`
User string `json:"user"`
Database string `json:"database"`
Replicas string `json:"replicas"`
UserSecretName string `json:"usersecretname"`
RootSecretName string `json:"rootsecretname"`
PrimarySecretName string `json:"primarysecretname"`
CollectSecretName string `json:"collectSecretName"`
Status string `json:"status"`
CustomConfig string `json:"customconfig"`
UserLabels map[string]string `json:"userlabels"`
PodAntiAffinity PodAntiAffinitySpec `json:"podAntiAffinity"`
SyncReplication *bool `json:"syncReplication"`
BackrestS3Bucket string `json:"backrestS3Bucket"`
BackrestS3Region string `json:"backrestS3Region"`
BackrestS3Endpoint string `json:"backrestS3Endpoint"`
BackrestS3URIStyle string `json:"backrestS3URIStyle"`
BackrestS3VerifyTLS string `json:"backrestS3VerifyTLS"`
BackrestRepoPath string `json:"backrestRepoPath"`
TablespaceMounts map[string]PgStorageSpec `json:"tablespaceMounts"`
TLS TLSSpec `json:"tls"`
TLSOnly bool `json:"tlsOnly"`
Standby bool `json:"standby"`
Shutdown bool `json:"shutdown"`
PGDataSource PGDataSourceSpec `json:"pgDataSource"`
}
// PGDataSourceSpec defines the data source that should be used to populate the initial PGDATA
// directory when bootstrapping a new PostgreSQL cluster
// swagger:ignore
type PGDataSourceSpec struct {
RestoreFrom string `json:"restoreFrom"`
RestoreOpts string `json:"restoreOpts"`
}
// PgclusterList is the CRD that defines a Crunchy PG Cluster List
// swagger:ignore
// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
type PgclusterList struct {
metav1.TypeMeta `json:",inline"`
metav1.ListMeta `json:"metadata"`
Items []Pgcluster `json:"items"`
}
// PgclusterStatus is the CRD that defines PG Cluster Status
// swagger:ignore
type PgclusterStatus struct {
State PgclusterState `json:"state,omitempty"`
Message string `json:"message,omitempty"`
}
// PgclusterState is the crd that defines PG Cluster Stage
// swagger:ignore
type PgclusterState string
// PodAntiAffinityDeployment distinguishes between the different types of
// Deployments that can leverage PodAntiAffinity
type PodAntiAffinityDeployment int
// PodAntiAffinityType defines the different types of type of anti-affinity rules applied to pg
// clusters when utilizing the default pod anti-affinity rules provided by the PostgreSQL Operator,
// which are enabled for a new pg cluster by default. Valid Values include "required" for
// requiredDuringSchedulingIgnoredDuringExecution anti-affinity, "preferred" for
// preferredDuringSchedulingIgnoredDuringExecution anti-affinity, and "disabled" to disable the
// default pod anti-affinity rules for the pg cluster all together.
type PodAntiAffinityType string
// PodAntiAffinitySpec provides multiple configurations for how pod
// anti-affinity can be set.
// - "Default" is the default rule that applies to all Pods that are a part of
// the PostgreSQL cluster
// - "PgBackrest" applies just to the pgBackRest repository Pods in said
// Deployment
// - "PgBouncer" applies to just pgBouncer Pods in said Deployment
// swaggier:ignore
type PodAntiAffinitySpec struct {
Default PodAntiAffinityType `json:"default"`
PgBackRest PodAntiAffinityType `json:"pgBackRest"`
PgBouncer PodAntiAffinityType `json:"pgBouncer"`
}
// PgBouncerSpec is a struct that is used within the Cluster specification that
// provides the attributes for managing a PgBouncer implementation, including:
// - is it enabled?
// - what resources it should consume
// - the total number of replicas
type PgBouncerSpec struct {
// Replicas represents the total number of Pods to deploy with pgBouncer,
// which effectively enables/disables the pgBouncer.
//
// if it is set to 0 or less, it is disabled.
//
// if it is set to 1 or more, it is enabled
Replicas int32 `json:"replicas"`
// Resources, if specified, contains the container request resources
// for any pgBouncer Deployments that are part of a PostgreSQL cluster
Resources v1.ResourceList `json:"resources"`
// Limits, if specified, contains the container resource limits
// for any pgBouncer Deployments that are part of a PostgreSQL cluster
Limits v1.ResourceList `json:"limits"`
}
// Enabled returns true if the pgBouncer is enabled for the cluster, i.e. there
// is at least one replica set
func (s *PgBouncerSpec) | () bool {
return s.Replicas > 0
}
// TLSSpec contains the information to set up a TLS-enabled PostgreSQL cluster
type TLSSpec struct {
// CASecret contains the name of the secret to use as the trusted CA for the
// TLSSecret
// This is our own format and should contain at least one key: "ca.crt"
// It can also contain a key "ca.crl" which is the certificate revocation list
CASecret string `json:"caSecret"`
// ReplicationTLSSecret contains the name of the secret that specifies a TLS
// keypair that can be used by the replication user (e.g. "primaryuser") to
// perform certificate based authentication between replicas.
// The keypair must be considered valid by the CA specified in the CASecret
ReplicationTLSSecret string `json:"replication | Enabled | identifier_name |
service.go | err != nil {
return errors.Wrap(err, "failed to process channels")
}
}
}
}
// RSSFeed generates RSS feed for given channel
func (s *Service) RSSFeed(fi FeedInfo) (string, error) | }
duration := ""
if entry.Duration > 0 {
duration = fmt.Sprintf("%d", entry.Duration)
}
items = append(items, rssfeed.Item{
Title: entry.Title,
Description: entry.Media.Description,
Link: entry.Link.Href,
PubDate: entry.Published.In(time.UTC).Format(time.RFC1123Z),
GUID: entry.ChannelID + "::" + entry.VideoID,
Author: entry.Author.Name,
Enclosure: rssfeed.Enclosure{
URL: fileURL,
Type: "audio/mpeg",
Length: fileSize,
},
Duration: duration,
DT: time.Now(),
})
}
rss := rssfeed.Rss2{
Version: "2.0",
NsItunes: "http://www.itunes.com/dtds/podcast-1.0.dtd",
NsMedia: "http://search.yahoo.com/mrss/",
ItemList: items,
Title: fi.Name,
Description: "generated by feed-master",
Link: entries[0].Author.URI,
PubDate: items[0].PubDate,
LastBuildDate: time.Now().Format(time.RFC1123Z),
Language: fi.Language,
ItunesAuthor: entries[0].Author.Name,
ItunesExplicit: "no",
}
// set image from channel as rss thumbnail
// TODO: we may want to load it locally in case if youtube doesn't like such remote usage of images
if image := entries[0].Media.Thumbnail.URL; image != "" {
rss.ItunesImage = &rssfeed.ItunesImg{URL: image}
rss.MediaThumbnail = &rssfeed.MediaThumbnail{URL: image}
}
if fi.Type == ytfeed.FTPlaylist {
rss.Link = "https://www.youtube.com/playlist?list=" + fi.ID
}
b, err := xml.MarshalIndent(&rss, "", " ")
if err != nil {
return "", errors.Wrap(err, "failed to marshal rss")
}
res := string(b)
// this hack to avoid having different items for marshal and unmarshal due to "itunes" namespace
res = strings.Replace(res, "<duration>", "<itunes:duration>", -1)
res = strings.Replace(res, "</duration>", "</itunes:duration>", -1)
return res, nil
}
// procChannels processes all channels, downloads audio, updates metadata and stores RSS
func (s *Service) procChannels(ctx context.Context) error {
var allStats stats
for _, feedInfo := range s.Feeds {
entries, err := s.ChannelService.Get(ctx, feedInfo.ID, feedInfo.Type)
if err != nil {
log.Printf("[WARN] failed to get channel entries for %s: %s", feedInfo.ID, err)
continue
}
log.Printf("[INFO] got %d entries for %s, limit to %d", len(entries), feedInfo.Name, s.keep(feedInfo))
changed, processed := false, 0
for i, entry := range entries {
// exit right away if context is done
select {
case <-ctx.Done():
return ctx.Err()
default:
}
allStats.entries++
if processed >= s.keep(feedInfo) {
break
}
isAllowed, err := s.isAllowed(entry, feedInfo)
if err != nil {
return errors.Wrapf(err, "failed to check if entry %s is relevant", entry.VideoID)
}
if !isAllowed {
log.Printf("[DEBUG] skipping filtered %s", entry.String())
allStats.ignored++
continue
}
ok, err := s.isNew(entry, feedInfo)
if err != nil {
return errors.Wrapf(err, "failed to check if entry %s exists", entry.VideoID)
}
if !ok {
allStats.skipped++
processed++
continue
}
// got new entry, but with very old timestamp. skip it if we have already reached max capacity
// (this is to eliminate the initial load) and this entry is older than the oldest one we have.
// Also marks it as processed as we don't want to process it again
oldestEntry := s.oldestEntry()
if entry.Published.Before(oldestEntry.Published) && s.countAllEntries() >= s.totalEntriesToKeep() {
allStats.ignored++
log.Printf("[INFO] skipping entry %s as it is older than the oldest one we have %s",
entry.String(), oldestEntry.String())
if procErr := s.Store.SetProcessed(entry); procErr != nil {
log.Printf("[WARN] failed to set processed status for %s: %v", entry.VideoID, procErr)
}
continue
}
log.Printf("[INFO] new entry [%d] %s, %s, %s, %s", i+1, entry.VideoID, entry.Title, feedInfo.Name, entry.String())
file, downErr := s.Downloader.Get(ctx, entry.VideoID, s.makeFileName(entry))
if downErr != nil {
allStats.ignored++
if downErr == ytfeed.ErrSkip { // downloader decided to skip this entry
log.Printf("[INFO] skipping %s", entry.String())
continue
}
log.Printf("[WARN] failed to download %s: %s", entry.VideoID, downErr)
continue
}
if short, duration := s.isShort(file); short {
allStats.ignored++
log.Printf("[INFO] skip short file %s (%v): %s, %s", file, duration, entry.VideoID, entry.String())
if procErr := s.Store.SetProcessed(entry); procErr != nil {
log.Printf("[WARN] failed to set processed status for %s: %v", entry.VideoID, procErr)
}
continue
}
// update metadata
if tagsErr := s.updateMp3Tags(file, entry, feedInfo); tagsErr != nil {
log.Printf("[WARN] failed to update metadata for %s: %s", entry.VideoID, tagsErr)
}
processed++
fsize := 0
if fi, err := os.Stat(file); err == nil {
fsize = int(fi.Size())
} else {
log.Printf("[WARN] failed to get file size for %s: %v", file, err)
}
log.Printf("[INFO] downloaded %s (%s) to %s, size: %d, channel: %+v", entry.VideoID, entry.Title, file, fsize, feedInfo)
entry = s.update(entry, file, feedInfo)
ok, saveErr := s.Store.Save(entry)
if saveErr != nil {
return errors.Wrapf(saveErr, "failed to save entry %+v", entry)
}
if !ok {
log.Printf("[WARN] attempt to save dup entry %+v", entry)
}
changed = true
if procErr := s.Store.SetProcessed(entry); procErr != nil {
log.Printf("[WARN] failed to set processed status for %s: %v", entry.VideoID, procErr)
}
allStats.added++
log.Printf("[INFO] saved %s (%s) to %s, channel: %+v", entry.VideoID, entry.Title, file, feedInfo)
}
allStats.processed += processed
if changed {
removed := s.removeOld(feedInfo)
allStats.removed += removed
// save rss feed to fs if there are new entries
rss, rssErr := s.RSSFeed(feedInfo)
if rssErr != nil {
log.Printf("[WARN] failed to generate rss for %s: %s", feedInfo.Name, rssErr)
} else {
if err := s.RSSFileStore.Save(feedInfo.ID, rss); err != nil {
log.Printf("[WARN] failed to save rss for %s: %s", feedInfo.Name, err)
}
}
}
}
log.Printf("[INFO] all channels processed - channels: %d, %s, lifetime: %d, feed size: %d",
len(s.Feeds), allStats.String(), s.Store.CountProcessed(), s.countAllEntries())
newestEntry | {
entries, err := s.Store.Load(fi.ID, s.keep(fi))
if err != nil {
return "", errors.Wrap(err, "failed to get channel entries")
}
if len(entries) == 0 {
return "", nil
}
items := []rssfeed.Item{}
for _, entry := range entries {
fileURL := s.RootURL + "/" + path.Base(entry.File)
var fileSize int
if fileInfo, fiErr := os.Stat(entry.File); fiErr != nil {
log.Printf("[WARN] failed to get file size for %s (%s %s): %v", entry.File, entry.VideoID, entry.Title, fiErr)
} else {
fileSize = int(fileInfo.Size()) | identifier_body |
service.go |
for _, f := range s.Feeds {
log.Printf("[INFO] youtube feed %+v", f)
}
tick := time.NewTicker(s.CheckDuration)
defer tick.Stop()
if err := s.procChannels(ctx); err != nil {
return errors.Wrap(err, "failed to process channels")
}
for {
select {
case <-ctx.Done():
return ctx.Err()
case <-tick.C:
if err := s.procChannels(ctx); err != nil {
return errors.Wrap(err, "failed to process channels")
}
}
}
}
// RSSFeed generates RSS feed for given channel
func (s *Service) RSSFeed(fi FeedInfo) (string, error) {
entries, err := s.Store.Load(fi.ID, s.keep(fi))
if err != nil {
return "", errors.Wrap(err, "failed to get channel entries")
}
if len(entries) == 0 {
return "", nil
}
items := []rssfeed.Item{}
for _, entry := range entries {
fileURL := s.RootURL + "/" + path.Base(entry.File)
var fileSize int
if fileInfo, fiErr := os.Stat(entry.File); fiErr != nil {
log.Printf("[WARN] failed to get file size for %s (%s %s): %v", entry.File, entry.VideoID, entry.Title, fiErr)
} else {
fileSize = int(fileInfo.Size())
}
duration := ""
if entry.Duration > 0 {
duration = fmt.Sprintf("%d", entry.Duration)
}
items = append(items, rssfeed.Item{
Title: entry.Title,
Description: entry.Media.Description,
Link: entry.Link.Href,
PubDate: entry.Published.In(time.UTC).Format(time.RFC1123Z),
GUID: entry.ChannelID + "::" + entry.VideoID,
Author: entry.Author.Name,
Enclosure: rssfeed.Enclosure{
URL: fileURL,
Type: "audio/mpeg",
Length: fileSize,
},
Duration: duration,
DT: time.Now(),
})
}
rss := rssfeed.Rss2{
Version: "2.0",
NsItunes: "http://www.itunes.com/dtds/podcast-1.0.dtd",
NsMedia: "http://search.yahoo.com/mrss/",
ItemList: items,
Title: fi.Name,
Description: "generated by feed-master",
Link: entries[0].Author.URI,
PubDate: items[0].PubDate,
LastBuildDate: time.Now().Format(time.RFC1123Z),
Language: fi.Language,
ItunesAuthor: entries[0].Author.Name,
ItunesExplicit: "no",
}
// set image from channel as rss thumbnail
// TODO: we may want to load it locally in case if youtube doesn't like such remote usage of images
if image := entries[0].Media.Thumbnail.URL; image != "" {
rss.ItunesImage = &rssfeed.ItunesImg{URL: image}
rss.MediaThumbnail = &rssfeed.MediaThumbnail{URL: image}
}
if fi.Type == ytfeed.FTPlaylist {
rss.Link = "https://www.youtube.com/playlist?list=" + fi.ID
}
b, err := xml.MarshalIndent(&rss, "", " ")
if err != nil {
return "", errors.Wrap(err, "failed to marshal rss")
}
res := string(b)
// this hack to avoid having different items for marshal and unmarshal due to "itunes" namespace
res = strings.Replace(res, "<duration>", "<itunes:duration>", -1)
res = strings.Replace(res, "</duration>", "</itunes:duration>", -1)
return res, nil
}
// procChannels processes all channels, downloads audio, updates metadata and stores RSS
func (s *Service) procChannels(ctx context.Context) error {
var allStats stats
for _, feedInfo := range s.Feeds {
entries, err := s.ChannelService.Get(ctx, feedInfo.ID, feedInfo.Type)
if err != nil {
log.Printf("[WARN] failed to get channel entries for %s: %s", feedInfo.ID, err)
continue
}
log.Printf("[INFO] got %d entries for %s, limit to %d", len(entries), feedInfo.Name, s.keep(feedInfo))
changed, processed := false, 0
for i, entry := range entries {
// exit right away if context is done
select {
case <-ctx.Done():
return ctx.Err()
default:
}
allStats.entries++
if processed >= s.keep(feedInfo) {
break
}
isAllowed, err := s.isAllowed(entry, feedInfo)
if err != nil {
return errors.Wrapf(err, "failed to check if entry %s is relevant", entry.VideoID)
}
if !isAllowed {
log.Printf("[DEBUG] skipping filtered %s", entry.String())
allStats.ignored++
continue
}
ok, err := s.isNew(entry, feedInfo)
if err != nil {
return errors.Wrapf(err, "failed to check if entry %s exists", entry.VideoID)
}
if !ok {
allStats.skipped++
processed++
continue
}
// got new entry, but with very old timestamp. skip it if we have already reached max capacity
// (this is to eliminate the initial load) and this entry is older than the oldest one we have.
// Also marks it as processed as we don't want to process it again
oldestEntry := s.oldestEntry()
if entry.Published.Before(oldestEntry.Published) && s.countAllEntries() >= s.totalEntriesToKeep() {
allStats.ignored++
log.Printf("[INFO] skipping entry %s as it is older than the oldest one we have %s",
entry.String(), oldestEntry.String())
if procErr := s.Store.SetProcessed(entry); procErr != nil {
log.Printf("[WARN] failed to set processed status for %s: %v", entry.VideoID, procErr)
}
continue
}
log.Printf("[INFO] new entry [%d] %s, %s, %s, %s", i+1, entry.VideoID, entry.Title, feedInfo.Name, entry.String())
file, downErr := s.Downloader.Get(ctx, entry.VideoID, s.makeFileName(entry))
if downErr != nil {
allStats.ignored++
if downErr == ytfeed.ErrSkip { // downloader decided to skip this entry
log.Printf("[INFO] skipping %s", entry.String())
continue
}
log.Printf("[WARN] failed to download %s: %s", entry.VideoID, downErr)
continue
}
if short, duration := s.isShort(file); short {
allStats.ignored++
log.Printf("[INFO] skip short file %s (%v): %s, %s", file, duration, entry.VideoID, entry.String())
if procErr := s.Store.SetProcessed(entry); procErr != nil {
log.Printf("[WARN] failed to set processed status for %s: %v", entry.VideoID, procErr)
}
continue
}
// update metadata
if tagsErr := s.updateMp3Tags(file, entry, feedInfo); tagsErr != nil {
log.Printf("[WARN] failed to update metadata for %s: %s", entry.VideoID, tagsErr)
}
processed++
fsize := 0
if fi, err := os.Stat(file); err == nil {
fsize = int(fi.Size())
} else {
log.Printf("[WARN] failed to get file size for %s: %v", file, err)
}
log.Printf("[INFO] downloaded %s (%s) to %s, size: %d, channel: %+v", entry.VideoID, entry.Title, file, fsize, feedInfo)
entry = s.update(entry, file, feedInfo)
ok, saveErr := s.Store.Save(entry)
if saveErr != nil {
return errors.Wrapf(saveErr, "failed to save entry %+v", entry)
}
if !ok {
log.Printf("[WARN] attempt to save dup entry %+v", entry)
}
changed = true
if procErr := s.Store.SetProcessed(entry); procErr != nil {
log.Printf("[WARN] failed to set processed status for %s: %v", entry.VideoID, procErr)
}
allStats.added++
log.Printf("[INFO] saved %s (%s) to %s, channel: %+v", entry.VideoID, entry.Title, file, feedInfo)
}
allStats.processed += processed
if changed {
removed := s.removeOld(feedInfo)
allStats.removed += removed
// save rss feed to fs if there are new entries
rss, rssErr := s.RSSFeed(feedInfo)
if rssErr != nil {
log.Printf("[WARN] failed to generate rss for %s: %s", | {
log.Printf("[DEBUG] skip youtube episodes shorter than %v", s.SkipShorts)
} | conditional_block |
|
service.go | INFO] new entry [%d] %s, %s, %s, %s", i+1, entry.VideoID, entry.Title, feedInfo.Name, entry.String())
file, downErr := s.Downloader.Get(ctx, entry.VideoID, s.makeFileName(entry))
if downErr != nil {
allStats.ignored++
if downErr == ytfeed.ErrSkip { // downloader decided to skip this entry
log.Printf("[INFO] skipping %s", entry.String())
continue
}
log.Printf("[WARN] failed to download %s: %s", entry.VideoID, downErr)
continue
}
if short, duration := s.isShort(file); short {
allStats.ignored++
log.Printf("[INFO] skip short file %s (%v): %s, %s", file, duration, entry.VideoID, entry.String())
if procErr := s.Store.SetProcessed(entry); procErr != nil {
log.Printf("[WARN] failed to set processed status for %s: %v", entry.VideoID, procErr)
}
continue
}
// update metadata
if tagsErr := s.updateMp3Tags(file, entry, feedInfo); tagsErr != nil {
log.Printf("[WARN] failed to update metadata for %s: %s", entry.VideoID, tagsErr)
}
processed++
fsize := 0
if fi, err := os.Stat(file); err == nil {
fsize = int(fi.Size())
} else {
log.Printf("[WARN] failed to get file size for %s: %v", file, err)
}
log.Printf("[INFO] downloaded %s (%s) to %s, size: %d, channel: %+v", entry.VideoID, entry.Title, file, fsize, feedInfo)
entry = s.update(entry, file, feedInfo)
ok, saveErr := s.Store.Save(entry)
if saveErr != nil {
return errors.Wrapf(saveErr, "failed to save entry %+v", entry)
}
if !ok {
log.Printf("[WARN] attempt to save dup entry %+v", entry)
}
changed = true
if procErr := s.Store.SetProcessed(entry); procErr != nil {
log.Printf("[WARN] failed to set processed status for %s: %v", entry.VideoID, procErr)
}
allStats.added++
log.Printf("[INFO] saved %s (%s) to %s, channel: %+v", entry.VideoID, entry.Title, file, feedInfo)
}
allStats.processed += processed
if changed {
removed := s.removeOld(feedInfo)
allStats.removed += removed
// save rss feed to fs if there are new entries
rss, rssErr := s.RSSFeed(feedInfo)
if rssErr != nil {
log.Printf("[WARN] failed to generate rss for %s: %s", feedInfo.Name, rssErr)
} else {
if err := s.RSSFileStore.Save(feedInfo.ID, rss); err != nil {
log.Printf("[WARN] failed to save rss for %s: %s", feedInfo.Name, err)
}
}
}
}
log.Printf("[INFO] all channels processed - channels: %d, %s, lifetime: %d, feed size: %d",
len(s.Feeds), allStats.String(), s.Store.CountProcessed(), s.countAllEntries())
newestEntry := s.newestEntry()
log.Printf("[INFO] last entry: %s", newestEntry.String())
return nil
}
// StoreRSS saves RSS feed to file
func (s *Service) StoreRSS(chanID, rss string) error {
return s.RSSFileStore.Save(chanID, rss)
}
// RemoveEntry deleted entry from store. Doesn't removes file
func (s *Service) RemoveEntry(entry ytfeed.Entry) error {
if err := s.Store.ResetProcessed(entry); err != nil {
return errors.Wrapf(err, "failed to reset processed entry %s", entry.VideoID)
}
if err := s.Store.Remove(entry); err != nil {
return errors.Wrapf(err, "failed to remove entry %s", entry.VideoID)
}
return nil
}
// isNew checks if entry already processed
func (s *Service) isNew(entry ytfeed.Entry, fi FeedInfo) (ok bool, err error) {
// check if entry already exists in store
// this method won't work after migration to locally altered published ts but have to stay for now
// to avoid false-positives on old entries what never got set with SetProcessed
exists, exErr := s.Store.Exist(entry)
if err != nil {
return false, errors.Wrapf(exErr, "failed to check if entry %s exists", entry.VideoID)
}
if exists {
return false, nil
}
// check if we already processed this entry.
// this is needed to avoid infinite get/remove loop when the original feed is updated in place.
// after migration to locally altered published ts, it is also the primary way to detect already processed entries
found, _, procErr := s.Store.CheckProcessed(entry)
if procErr != nil {
log.Printf("[WARN] can't get processed status for %s, %+v", entry.VideoID, fi)
}
if procErr == nil && found {
return false, nil
}
return true, nil
}
// isAllowed checks if entry matches all filters for the channel feed
func (s *Service) isAllowed(entry ytfeed.Entry, fi FeedInfo) (ok bool, err error) {
matchedIncludeFilter := true
if fi.Filter.Include != "" {
matchedIncludeFilter, err = regexp.MatchString(fi.Filter.Include, entry.Title)
if err != nil {
return false, errors.Wrapf(err, "failed to check if entry %s matches include filter", entry.VideoID)
}
}
matchedExcludeFilter := false
if fi.Filter.Exclude != "" {
matchedExcludeFilter, err = regexp.MatchString(fi.Filter.Exclude, entry.Title)
if err != nil {
return false, errors.Wrapf(err, "failed to check if entry %s matches exclude filter", entry.VideoID)
}
}
return matchedIncludeFilter && !matchedExcludeFilter, nil
}
func (s *Service) isShort(file string) (bool, time.Duration) {
if s.SkipShorts.Seconds() > 0 {
// skip shorts if duration is less than SkipShorts
duration := s.DurationService.File(file)
if duration > 0 && duration < int(s.SkipShorts.Seconds()) {
return true, time.Duration(duration) * time.Second
}
}
return false, 0
}
// update sets entry file name and reset published ts
func (s *Service) update(entry ytfeed.Entry, file string, fi FeedInfo) ytfeed.Entry {
entry.File = file
// only reset time if published not too long ago
// this is done to avoid initial set of entries added with a new channel to the top of the feed
if time.Since(entry.Published) < time.Hour*24 {
log.Printf("[DEBUG] reset published time for %s, from %s to %s (%v), %s",
entry.VideoID, entry.Published.Format(time.RFC3339), time.Now().Format(time.RFC3339),
time.Since(entry.Published), entry.String())
entry.Published = time.Now() // reset published ts to prevent possible out-of-order entries
} else {
log.Printf("[DEBUG] keep published time for %s, %s", entry.VideoID, entry.Published.Format(time.RFC3339))
}
if !strings.Contains(entry.Title, fi.Name) { // if title doesn't contains channel name add it
entry.Title = fi.Name + ": " + entry.Title
}
entry.Duration = s.DurationService.File(file)
log.Printf("[DEBUG] updated entry: %s", entry.String())
return entry
}
// removeOld deletes old entries from store and corresponding files
func (s *Service) removeOld(fi FeedInfo) int {
removed := 0
keep := s.keep(fi)
files, err := s.Store.RemoveOld(fi.ID, keep+1)
if err != nil { // even with error we get a list of files to remove
log.Printf("[WARN] failed to remove some old meta data for %s, %v", fi.ID, err)
}
for _, f := range files {
if e := os.Remove(f); e != nil {
log.Printf("[WARN] failed to remove file %s: %v", f, e)
continue
}
removed++
log.Printf("[INFO] removed %s for %s (%s)", f, fi.ID, fi.Name)
}
return removed
}
func (s *Service) keep(fi FeedInfo) int {
keep := s.KeepPerChannel
if fi.Keep > 0 {
keep = fi.Keep
}
return keep
}
func (s *Service) makeFileName(entry ytfeed.Entry) string {
h := sha1.New()
if _, err := h.Write([]byte(entry.UID())); err != nil {
return uuid.New().String()
}
return fmt.Sprintf("%x", h.Sum(nil))
}
// totalEntriesToKeep returns total number of entries to keep, summing all channels' keep values
func (s *Service) | totalEntriesToKeep | identifier_name |
|
service.go | return errors.Wrapf(err, "failed to check if entry %s is relevant", entry.VideoID)
}
if !isAllowed {
log.Printf("[DEBUG] skipping filtered %s", entry.String())
allStats.ignored++
continue
}
ok, err := s.isNew(entry, feedInfo)
if err != nil {
return errors.Wrapf(err, "failed to check if entry %s exists", entry.VideoID)
}
if !ok {
allStats.skipped++
processed++
continue
}
// got new entry, but with very old timestamp. skip it if we have already reached max capacity
// (this is to eliminate the initial load) and this entry is older than the oldest one we have.
// Also marks it as processed as we don't want to process it again
oldestEntry := s.oldestEntry()
if entry.Published.Before(oldestEntry.Published) && s.countAllEntries() >= s.totalEntriesToKeep() {
allStats.ignored++
log.Printf("[INFO] skipping entry %s as it is older than the oldest one we have %s",
entry.String(), oldestEntry.String())
if procErr := s.Store.SetProcessed(entry); procErr != nil {
log.Printf("[WARN] failed to set processed status for %s: %v", entry.VideoID, procErr)
}
continue
}
log.Printf("[INFO] new entry [%d] %s, %s, %s, %s", i+1, entry.VideoID, entry.Title, feedInfo.Name, entry.String())
file, downErr := s.Downloader.Get(ctx, entry.VideoID, s.makeFileName(entry))
if downErr != nil {
allStats.ignored++
if downErr == ytfeed.ErrSkip { // downloader decided to skip this entry
log.Printf("[INFO] skipping %s", entry.String())
continue
}
log.Printf("[WARN] failed to download %s: %s", entry.VideoID, downErr)
continue
}
if short, duration := s.isShort(file); short {
allStats.ignored++
log.Printf("[INFO] skip short file %s (%v): %s, %s", file, duration, entry.VideoID, entry.String())
if procErr := s.Store.SetProcessed(entry); procErr != nil {
log.Printf("[WARN] failed to set processed status for %s: %v", entry.VideoID, procErr)
}
continue
}
// update metadata
if tagsErr := s.updateMp3Tags(file, entry, feedInfo); tagsErr != nil {
log.Printf("[WARN] failed to update metadata for %s: %s", entry.VideoID, tagsErr)
}
processed++
fsize := 0
if fi, err := os.Stat(file); err == nil {
fsize = int(fi.Size())
} else {
log.Printf("[WARN] failed to get file size for %s: %v", file, err)
}
log.Printf("[INFO] downloaded %s (%s) to %s, size: %d, channel: %+v", entry.VideoID, entry.Title, file, fsize, feedInfo)
entry = s.update(entry, file, feedInfo)
ok, saveErr := s.Store.Save(entry)
if saveErr != nil {
return errors.Wrapf(saveErr, "failed to save entry %+v", entry)
}
if !ok {
log.Printf("[WARN] attempt to save dup entry %+v", entry)
}
changed = true
if procErr := s.Store.SetProcessed(entry); procErr != nil {
log.Printf("[WARN] failed to set processed status for %s: %v", entry.VideoID, procErr)
}
allStats.added++
log.Printf("[INFO] saved %s (%s) to %s, channel: %+v", entry.VideoID, entry.Title, file, feedInfo)
}
allStats.processed += processed
if changed {
removed := s.removeOld(feedInfo)
allStats.removed += removed
// save rss feed to fs if there are new entries
rss, rssErr := s.RSSFeed(feedInfo)
if rssErr != nil {
log.Printf("[WARN] failed to generate rss for %s: %s", feedInfo.Name, rssErr)
} else {
if err := s.RSSFileStore.Save(feedInfo.ID, rss); err != nil {
log.Printf("[WARN] failed to save rss for %s: %s", feedInfo.Name, err)
}
}
}
}
log.Printf("[INFO] all channels processed - channels: %d, %s, lifetime: %d, feed size: %d",
len(s.Feeds), allStats.String(), s.Store.CountProcessed(), s.countAllEntries())
newestEntry := s.newestEntry()
log.Printf("[INFO] last entry: %s", newestEntry.String())
return nil
}
// StoreRSS saves RSS feed to file
func (s *Service) StoreRSS(chanID, rss string) error {
return s.RSSFileStore.Save(chanID, rss)
}
// RemoveEntry deleted entry from store. Doesn't removes file
func (s *Service) RemoveEntry(entry ytfeed.Entry) error {
if err := s.Store.ResetProcessed(entry); err != nil {
return errors.Wrapf(err, "failed to reset processed entry %s", entry.VideoID)
}
if err := s.Store.Remove(entry); err != nil {
return errors.Wrapf(err, "failed to remove entry %s", entry.VideoID)
}
return nil
}
// isNew checks if entry already processed
func (s *Service) isNew(entry ytfeed.Entry, fi FeedInfo) (ok bool, err error) {
// check if entry already exists in store
// this method won't work after migration to locally altered published ts but have to stay for now
// to avoid false-positives on old entries what never got set with SetProcessed
exists, exErr := s.Store.Exist(entry)
if err != nil {
return false, errors.Wrapf(exErr, "failed to check if entry %s exists", entry.VideoID)
}
if exists {
return false, nil
}
// check if we already processed this entry.
// this is needed to avoid infinite get/remove loop when the original feed is updated in place.
// after migration to locally altered published ts, it is also the primary way to detect already processed entries
found, _, procErr := s.Store.CheckProcessed(entry)
if procErr != nil {
log.Printf("[WARN] can't get processed status for %s, %+v", entry.VideoID, fi)
}
if procErr == nil && found {
return false, nil
}
return true, nil
}
// isAllowed checks if entry matches all filters for the channel feed
func (s *Service) isAllowed(entry ytfeed.Entry, fi FeedInfo) (ok bool, err error) {
matchedIncludeFilter := true
if fi.Filter.Include != "" {
matchedIncludeFilter, err = regexp.MatchString(fi.Filter.Include, entry.Title)
if err != nil {
return false, errors.Wrapf(err, "failed to check if entry %s matches include filter", entry.VideoID)
}
}
matchedExcludeFilter := false
if fi.Filter.Exclude != "" {
matchedExcludeFilter, err = regexp.MatchString(fi.Filter.Exclude, entry.Title)
if err != nil {
return false, errors.Wrapf(err, "failed to check if entry %s matches exclude filter", entry.VideoID)
}
}
return matchedIncludeFilter && !matchedExcludeFilter, nil
}
func (s *Service) isShort(file string) (bool, time.Duration) {
if s.SkipShorts.Seconds() > 0 {
// skip shorts if duration is less than SkipShorts
duration := s.DurationService.File(file)
if duration > 0 && duration < int(s.SkipShorts.Seconds()) {
return true, time.Duration(duration) * time.Second
}
}
return false, 0
}
// update sets entry file name and reset published ts
func (s *Service) update(entry ytfeed.Entry, file string, fi FeedInfo) ytfeed.Entry {
entry.File = file
// only reset time if published not too long ago
// this is done to avoid initial set of entries added with a new channel to the top of the feed
if time.Since(entry.Published) < time.Hour*24 {
log.Printf("[DEBUG] reset published time for %s, from %s to %s (%v), %s",
entry.VideoID, entry.Published.Format(time.RFC3339), time.Now().Format(time.RFC3339),
time.Since(entry.Published), entry.String())
entry.Published = time.Now() // reset published ts to prevent possible out-of-order entries
} else {
log.Printf("[DEBUG] keep published time for %s, %s", entry.VideoID, entry.Published.Format(time.RFC3339))
}
if !strings.Contains(entry.Title, fi.Name) { // if title doesn't contains channel name add it
entry.Title = fi.Name + ": " + entry.Title
}
entry.Duration = s.DurationService.File(file) | log.Printf("[DEBUG] updated entry: %s", entry.String()) | random_line_split |
|
mod.rs | _mem: usize,
show_gc: bool,
show_allocates: bool,
show_heap_map: bool,
show_free_list: bool,
}
impl<'a> IntoIterator for &'a Memory {
type Item = usize;
type IntoIter = MemoryIntoIterator<'a>;
fn into_iter(self) -> Self::IntoIter {
MemoryIntoIterator {
mem: self,
scan: 0,
free: 0,
}
}
}
pub struct MemoryIntoIterator<'a> {
mem: &'a Memory,
scan: usize,
free: usize,
}
impl<'a> Iterator for MemoryIntoIterator<'a> {
type Item = usize;
fn next(&mut self) -> Option<Self::Item> {
if self.scan == 0 {
self.scan = 1;
self.free = self.mem.head;
} else {
self.scan = self.mem.next_object_in_heap(self.scan);
}
while self.scan == self.free {
self.scan = self.mem.next_object_in_heap(self.free);
self.free = self.mem.get_fl_next(self.free);
}
if self.scan >= MAX_MEMORY_SLOTS - 1 {
return None;
} else {
return Some(self.scan);
}
}
}
#[cfg(feature = "dynamic_mem")]
fn im() -> Mem {
return vec![0; MAX_MEMORY_SLOTS];
}
#[cfg(not(feature = "dynamic_mem"))]
fn im() -> Mem {
return [0; MAX_MEMORY_SLOTS];
}
impl Memory {
pub fn initialze_memory() -> Memory {
let mut mem = Memory {
head: 1,
mem: im(),
mark_bits: [0; MARK_BITS],
roots: Vec::new(),
gc_count: 0,
allocates: 0,
lastgc_live_mem: 0,
lastgc_free_mem: 0,
last_gc_ms: 0,
total_gc_ms: 0,
show_gc: false,
show_allocates: false,
show_heap_map: false,
show_free_list: false,
};
mem.set_size(0, MAX_MEMORY_SLOTS); // magic memory at zero is heap_size
mem.set_size(mem.head, MAX_MEMORY_SLOTS - 2); // set initial object size as all heap
mem.set_fl_next(mem.head, 0);
mem
}
// objects API
// allocate_object (size) --- size is number of indexable slots
// add/remote_root () --- add to or remove from gc root set.
// element_size() - number of indexable slots - get_size() - OBJECT_HEADER_SLOTS
// at_put - store into object slot at index
// at -- fetch object slot at index
pub fn allocate_object(&mut self, unrounded_size: usize) -> usize {
self.allocates += 1;
let mut result = self.allocate_object_nocompress(unrounded_size);
if result == 0 {
self.gc();
result = self.allocate_object_nocompress(unrounded_size);
if result == 0 {
self.print_freelist();
self.print_heap();
panic!("out of memory");
}
}
result
}
pub fn live_objects(&self) -> MemoryIntoIterator {
return self.into_iter();
}
pub fn add_root(&mut self, obj: usize) {
self.roots.push(obj);
}
pub fn remove_root(&mut self, obj: usize) {
for i in 0..self.roots.len() {
if obj == self.roots[i] {
self.roots.remove(i);
return;
}
}
}
pub fn at_put(&mut self, obj: usize, index: usize, value: usize) {
let slots = self.mem[obj];
let base = obj+OBJECT_HEADER_SLOTS;
let object =&mut self.mem[ base.. base + slots ];
object[index] = value;
}
pub fn at(&self, obj: usize, index: usize) -> usize {
let slots = self.mem[obj];
let base = obj+OBJECT_HEADER_SLOTS;
let object =&self.mem[ base.. base + slots ];
return object[index];
}
pub fn element_size(&self, obj: usize) -> usize {
return self.mem[obj] - OBJECT_HEADER_SLOTS;
}
pub fn enable_show_heap_map(&mut self, enabled: bool) {
self.show_heap_map = enabled;
}
pub fn enable_show_freelist(&mut self, enabled: bool) {
self.show_free_list = enabled;
}
pub fn enable_show_gc(&mut self, enabled: bool) {
self.show_gc = enabled;
}
pub fn enable_show_allocates(&mut self, enabled: bool) {
self.show_allocates = enabled;
}
fn rounded_size(unrounded_size: usize) -> usize {
(unrounded_size + 1) & !(1) // rounded to 2
}
fn get_size(&self, obj: usize) -> usize {
return self.mem[obj];
}
fn set_size(&mut self, obj: usize, size: usize) {
self.mem[obj] = size;
}
fn next_object_in_heap(&self, obj: usize) -> usize {
return obj + self.get_size(obj);
}
//free list is linked off the first slot
fn get_fl_next(&self, obj: usize) -> usize {
return self.mem[obj + 1];
}
fn | (&mut self, obj: usize, next: usize) {
self.mem[obj + 1] = next;
}
fn mark_object(&mut self, obj: usize) {
self.mark_bits[obj / MARK_BITS_PER_SLOT] |= 1 << (obj % MARK_BITS_PER_SLOT);
}
fn unmark_object(&mut self, obj: usize) {
self.mark_bits[obj / MARK_BITS_PER_SLOT] &= !(1 << (obj % MARK_BITS_PER_SLOT));
}
fn is_marked(&self, obj: usize) -> bool {
(self.mark_bits[obj / MARK_BITS_PER_SLOT] & (1 << (obj % MARK_BITS_PER_SLOT))) != 0
}
fn allocate_object_nocompress(&mut self, unrounded_size: usize) -> usize {
let size = Memory::rounded_size(unrounded_size + OBJECT_HEADER_SLOTS);
let mut free = self.head;
while free != 0 {
let avail = self.get_size(free);
if avail > size {
let newsize = avail - size;
if newsize < 2 {
panic!("remaining size is less than 2");
}
// shrink current free to smaller size
self.set_size(free, newsize);
// new object is on the end of current free object
let new_object = free + newsize;
self.set_size(new_object, size);
for index in 0..self.element_size(new_object) {
self.at_put(new_object, index, 0);
}
if self.show_allocates {
println!(
"Success: allocate_object returning -> {} size {}",
new_object, size
);
}
if self.head != free {
if self.show_allocates {
println!("Reset head past intermediate free blocks \n");
let mut show = self.head;
while show != free {
println!("Abandon {} size {}\n", show, self.get_size(show));
show = self.get_fl_next(show);
}
}
self.head = free;
}
return new_object;
}
free = self.get_fl_next(free);
}
0
}
pub fn gc(&mut self) {
let start = SystemTime::now();
for i in 0..self.roots.len() {
self.mark_and_scan(self.roots[i]);
}
self.sweep();
self.gc_count += 1;
if self.show_gc {
self.print_gc_stats();
}
match start.elapsed() {
Ok(elapsed) => {
self.last_gc_ms = elapsed.as_millis();
self.total_gc_ms += self.last_gc_ms;
}
Err(e) => {
println!("Error: {:?}", e);
}
}
}
fn sweep(&mut self) {
let mut scan = 1;
self.head = 0;
let mut tail = self.head;
self.lastgc_free_mem = 0;
self.lastgc_live_mem = 0;
while scan < MAX_MEMORY_SLOTS - 1 {
if self.is_marked(scan) {
self.unmark_object(scan);
self.lastgc_live_mem += self.get_size(scan);
} else {
self.lastgc_free_mem += self.get_size(scan);
if tail == 0 {
self.head = scan;
self.set_fl_next(scan, 0);
tail = scan;
} else {
if self.next_object_in_heap(tail) == scan {
self.set_size(tail, self.get_size(tail) + self.get_size(scan));
} else {
self.set_fl_next(tail, scan);
self.set_fl_next(scan, 0);
tail = scan;
}
}
}
scan = self.next_object_in_heap(scan);
}
if self.show_free_list {
self.print_freelist();
}
if self.show_heap_map {
self.print_heap();
}
}
fn mark_and_scan(&mut | set_fl_next | identifier_name |
mod.rs | #[cfg(not(feature="dynamic_mem"))]
const MAX_MEMORY_SLOTS: usize = 1024 * 128;
type Bits = u128;
const MARK_BITS_PER_SLOT: usize = mem::size_of::<Bits>();
const MARK_BITS: usize = MAX_MEMORY_SLOTS / MARK_BITS_PER_SLOT;
#[cfg(feature="dynamic_mem")]
type Mem = Vec<usize>;
#[cfg(not(feature="dynamic_mem"))]
type Mem = [usize; MAX_MEMORY_SLOTS] ;
pub const OBJECT_HEADER_SLOTS: usize = 1;
pub struct Memory {
head: usize,
mem: Mem,
mark_bits: [u128; MARK_BITS],
roots: Vec<usize>,
gc_count: usize,
allocates: usize,
last_gc_ms: u128,
total_gc_ms: u128,
lastgc_live_mem: usize,
lastgc_free_mem: usize,
show_gc: bool,
show_allocates: bool,
show_heap_map: bool,
show_free_list: bool,
}
impl<'a> IntoIterator for &'a Memory {
type Item = usize;
type IntoIter = MemoryIntoIterator<'a>;
fn into_iter(self) -> Self::IntoIter {
MemoryIntoIterator {
mem: self,
scan: 0,
free: 0,
}
}
}
pub struct MemoryIntoIterator<'a> {
mem: &'a Memory,
scan: usize,
free: usize,
}
impl<'a> Iterator for MemoryIntoIterator<'a> {
type Item = usize;
fn next(&mut self) -> Option<Self::Item> {
if self.scan == 0 {
self.scan = 1;
self.free = self.mem.head;
} else {
self.scan = self.mem.next_object_in_heap(self.scan);
}
while self.scan == self.free {
self.scan = self.mem.next_object_in_heap(self.free);
self.free = self.mem.get_fl_next(self.free);
}
if self.scan >= MAX_MEMORY_SLOTS - 1 {
return None;
} else {
return Some(self.scan);
}
}
}
#[cfg(feature = "dynamic_mem")]
fn im() -> Mem {
return vec![0; MAX_MEMORY_SLOTS];
}
#[cfg(not(feature = "dynamic_mem"))]
fn im() -> Mem {
return [0; MAX_MEMORY_SLOTS];
}
impl Memory {
pub fn initialze_memory() -> Memory {
let mut mem = Memory {
head: 1,
mem: im(),
mark_bits: [0; MARK_BITS],
roots: Vec::new(),
gc_count: 0,
allocates: 0,
lastgc_live_mem: 0,
lastgc_free_mem: 0,
last_gc_ms: 0,
total_gc_ms: 0,
show_gc: false,
show_allocates: false,
show_heap_map: false,
show_free_list: false,
};
mem.set_size(0, MAX_MEMORY_SLOTS); // magic memory at zero is heap_size
mem.set_size(mem.head, MAX_MEMORY_SLOTS - 2); // set initial object size as all heap
mem.set_fl_next(mem.head, 0);
mem
}
// objects API
// allocate_object (size) --- size is number of indexable slots
// add/remote_root () --- add to or remove from gc root set.
// element_size() - number of indexable slots - get_size() - OBJECT_HEADER_SLOTS
// at_put - store into object slot at index
// at -- fetch object slot at index
pub fn allocate_object(&mut self, unrounded_size: usize) -> usize {
self.allocates += 1;
let mut result = self.allocate_object_nocompress(unrounded_size);
if result == 0 {
self.gc();
result = self.allocate_object_nocompress(unrounded_size);
if result == 0 {
self.print_freelist();
self.print_heap();
panic!("out of memory");
}
}
result
}
pub fn live_objects(&self) -> MemoryIntoIterator {
return self.into_iter();
}
pub fn add_root(&mut self, obj: usize) {
self.roots.push(obj);
}
pub fn remove_root(&mut self, obj: usize) {
for i in 0..self.roots.len() {
if obj == self.roots[i] {
self.roots.remove(i);
return;
}
}
}
pub fn at_put(&mut self, obj: usize, index: usize, value: usize) {
let slots = self.mem[obj];
let base = obj+OBJECT_HEADER_SLOTS;
let object =&mut self.mem[ base.. base + slots ];
object[index] = value;
}
pub fn at(&self, obj: usize, index: usize) -> usize {
let slots = self.mem[obj];
let base = obj+OBJECT_HEADER_SLOTS;
let object =&self.mem[ base.. base + slots ];
return object[index];
}
pub fn element_size(&self, obj: usize) -> usize {
return self.mem[obj] - OBJECT_HEADER_SLOTS;
}
pub fn enable_show_heap_map(&mut self, enabled: bool) {
self.show_heap_map = enabled;
}
pub fn enable_show_freelist(&mut self, enabled: bool) {
self.show_free_list = enabled;
}
pub fn enable_show_gc(&mut self, enabled: bool) {
self.show_gc = enabled;
}
pub fn enable_show_allocates(&mut self, enabled: bool) {
self.show_allocates = enabled;
}
fn rounded_size(unrounded_size: usize) -> usize {
(unrounded_size + 1) & !(1) // rounded to 2
}
fn get_size(&self, obj: usize) -> usize {
return self.mem[obj];
}
fn set_size(&mut self, obj: usize, size: usize) {
self.mem[obj] = size;
}
fn next_object_in_heap(&self, obj: usize) -> usize {
return obj + self.get_size(obj);
}
//free list is linked off the first slot
fn get_fl_next(&self, obj: usize) -> usize {
return self.mem[obj + 1];
}
fn set_fl_next(&mut self, obj: usize, next: usize) {
self.mem[obj + 1] = next;
}
fn mark_object(&mut self, obj: usize) {
self.mark_bits[obj / MARK_BITS_PER_SLOT] |= 1 << (obj % MARK_BITS_PER_SLOT);
}
fn unmark_object(&mut self, obj: usize) {
self.mark_bits[obj / MARK_BITS_PER_SLOT] &= !(1 << (obj % MARK_BITS_PER_SLOT));
}
fn is_marked(&self, obj: usize) -> bool {
(self.mark_bits[obj / MARK_BITS_PER_SLOT] & (1 << (obj % MARK_BITS_PER_SLOT))) != 0
}
fn allocate_object_nocompress(&mut self, unrounded_size: usize) -> usize {
let size = Memory::rounded_size(unrounded_size + OBJECT_HEADER_SLOTS);
let mut free = self.head;
while free != 0 {
let avail = self.get_size(free);
if avail > size {
let newsize = avail - size;
if newsize < 2 {
panic!("remaining size is less than 2");
}
// shrink current free to smaller size
self.set_size(free, newsize);
// new object is on the end of current free object
let new_object = free + newsize;
self.set_size(new_object, size);
for index in 0..self.element_size(new_object) {
self.at_put(new_object, index, 0);
}
if self.show_allocates {
println!(
"Success: allocate_object returning -> {} size {}",
new_object, size
);
}
if self.head != free {
if self.show_allocates {
println!("Reset head past intermediate free blocks \n");
let mut show = self.head;
while show != free {
println!("Abandon {} size {}\n", show, self.get_size(show));
show = self.get_fl_next(show);
}
}
self.head = free;
}
return new_object;
}
free = self.get_fl_next(free);
}
0
}
pub fn gc(&mut self) {
let start = SystemTime::now();
for i in 0..self.roots.len() {
self.mark_and_scan(self.roots[i]);
}
self.sweep();
self.gc_count += 1;
if self.show_gc {
self.print_gc_stats();
}
match start.elapsed() {
Ok(elapsed) => {
self.last_gc_ms = elapsed.as_millis();
self.total_gc_ms += self.last_gc_ms;
}
Err(e) => {
println!("Error: {:?}", e);
}
}
}
fn sweep(&mut self) {
let mut scan = 1;
self.head = 0;
let mut tail = self.head;
|
#[cfg(feature="dynamic_mem")]
const MAX_MEMORY_SLOTS: usize = 1024 * 1024 * 2; | random_line_split |
|
mod.rs | _free_list: bool,
}
impl<'a> IntoIterator for &'a Memory {
type Item = usize;
type IntoIter = MemoryIntoIterator<'a>;
fn into_iter(self) -> Self::IntoIter {
MemoryIntoIterator {
mem: self,
scan: 0,
free: 0,
}
}
}
pub struct MemoryIntoIterator<'a> {
mem: &'a Memory,
scan: usize,
free: usize,
}
impl<'a> Iterator for MemoryIntoIterator<'a> {
type Item = usize;
fn next(&mut self) -> Option<Self::Item> {
if self.scan == 0 {
self.scan = 1;
self.free = self.mem.head;
} else {
self.scan = self.mem.next_object_in_heap(self.scan);
}
while self.scan == self.free {
self.scan = self.mem.next_object_in_heap(self.free);
self.free = self.mem.get_fl_next(self.free);
}
if self.scan >= MAX_MEMORY_SLOTS - 1 {
return None;
} else {
return Some(self.scan);
}
}
}
#[cfg(feature = "dynamic_mem")]
fn im() -> Mem {
return vec![0; MAX_MEMORY_SLOTS];
}
#[cfg(not(feature = "dynamic_mem"))]
fn im() -> Mem {
return [0; MAX_MEMORY_SLOTS];
}
impl Memory {
pub fn initialze_memory() -> Memory {
let mut mem = Memory {
head: 1,
mem: im(),
mark_bits: [0; MARK_BITS],
roots: Vec::new(),
gc_count: 0,
allocates: 0,
lastgc_live_mem: 0,
lastgc_free_mem: 0,
last_gc_ms: 0,
total_gc_ms: 0,
show_gc: false,
show_allocates: false,
show_heap_map: false,
show_free_list: false,
};
mem.set_size(0, MAX_MEMORY_SLOTS); // magic memory at zero is heap_size
mem.set_size(mem.head, MAX_MEMORY_SLOTS - 2); // set initial object size as all heap
mem.set_fl_next(mem.head, 0);
mem
}
// objects API
// allocate_object (size) --- size is number of indexable slots
// add/remote_root () --- add to or remove from gc root set.
// element_size() - number of indexable slots - get_size() - OBJECT_HEADER_SLOTS
// at_put - store into object slot at index
// at -- fetch object slot at index
pub fn allocate_object(&mut self, unrounded_size: usize) -> usize {
self.allocates += 1;
let mut result = self.allocate_object_nocompress(unrounded_size);
if result == 0 {
self.gc();
result = self.allocate_object_nocompress(unrounded_size);
if result == 0 {
self.print_freelist();
self.print_heap();
panic!("out of memory");
}
}
result
}
pub fn live_objects(&self) -> MemoryIntoIterator {
return self.into_iter();
}
pub fn add_root(&mut self, obj: usize) {
self.roots.push(obj);
}
pub fn remove_root(&mut self, obj: usize) {
for i in 0..self.roots.len() {
if obj == self.roots[i] {
self.roots.remove(i);
return;
}
}
}
pub fn at_put(&mut self, obj: usize, index: usize, value: usize) {
let slots = self.mem[obj];
let base = obj+OBJECT_HEADER_SLOTS;
let object =&mut self.mem[ base.. base + slots ];
object[index] = value;
}
pub fn at(&self, obj: usize, index: usize) -> usize {
let slots = self.mem[obj];
let base = obj+OBJECT_HEADER_SLOTS;
let object =&self.mem[ base.. base + slots ];
return object[index];
}
pub fn element_size(&self, obj: usize) -> usize {
return self.mem[obj] - OBJECT_HEADER_SLOTS;
}
pub fn enable_show_heap_map(&mut self, enabled: bool) {
self.show_heap_map = enabled;
}
pub fn enable_show_freelist(&mut self, enabled: bool) {
self.show_free_list = enabled;
}
pub fn enable_show_gc(&mut self, enabled: bool) {
self.show_gc = enabled;
}
pub fn enable_show_allocates(&mut self, enabled: bool) {
self.show_allocates = enabled;
}
fn rounded_size(unrounded_size: usize) -> usize {
(unrounded_size + 1) & !(1) // rounded to 2
}
fn get_size(&self, obj: usize) -> usize {
return self.mem[obj];
}
fn set_size(&mut self, obj: usize, size: usize) {
self.mem[obj] = size;
}
fn next_object_in_heap(&self, obj: usize) -> usize {
return obj + self.get_size(obj);
}
//free list is linked off the first slot
fn get_fl_next(&self, obj: usize) -> usize {
return self.mem[obj + 1];
}
fn set_fl_next(&mut self, obj: usize, next: usize) {
self.mem[obj + 1] = next;
}
fn mark_object(&mut self, obj: usize) {
self.mark_bits[obj / MARK_BITS_PER_SLOT] |= 1 << (obj % MARK_BITS_PER_SLOT);
}
fn unmark_object(&mut self, obj: usize) {
self.mark_bits[obj / MARK_BITS_PER_SLOT] &= !(1 << (obj % MARK_BITS_PER_SLOT));
}
fn is_marked(&self, obj: usize) -> bool {
(self.mark_bits[obj / MARK_BITS_PER_SLOT] & (1 << (obj % MARK_BITS_PER_SLOT))) != 0
}
fn allocate_object_nocompress(&mut self, unrounded_size: usize) -> usize {
let size = Memory::rounded_size(unrounded_size + OBJECT_HEADER_SLOTS);
let mut free = self.head;
while free != 0 {
let avail = self.get_size(free);
if avail > size {
let newsize = avail - size;
if newsize < 2 {
panic!("remaining size is less than 2");
}
// shrink current free to smaller size
self.set_size(free, newsize);
// new object is on the end of current free object
let new_object = free + newsize;
self.set_size(new_object, size);
for index in 0..self.element_size(new_object) {
self.at_put(new_object, index, 0);
}
if self.show_allocates {
println!(
"Success: allocate_object returning -> {} size {}",
new_object, size
);
}
if self.head != free {
if self.show_allocates {
println!("Reset head past intermediate free blocks \n");
let mut show = self.head;
while show != free {
println!("Abandon {} size {}\n", show, self.get_size(show));
show = self.get_fl_next(show);
}
}
self.head = free;
}
return new_object;
}
free = self.get_fl_next(free);
}
0
}
pub fn gc(&mut self) {
let start = SystemTime::now();
for i in 0..self.roots.len() {
self.mark_and_scan(self.roots[i]);
}
self.sweep();
self.gc_count += 1;
if self.show_gc {
self.print_gc_stats();
}
match start.elapsed() {
Ok(elapsed) => {
self.last_gc_ms = elapsed.as_millis();
self.total_gc_ms += self.last_gc_ms;
}
Err(e) => {
println!("Error: {:?}", e);
}
}
}
fn sweep(&mut self) {
let mut scan = 1;
self.head = 0;
let mut tail = self.head;
self.lastgc_free_mem = 0;
self.lastgc_live_mem = 0;
while scan < MAX_MEMORY_SLOTS - 1 {
if self.is_marked(scan) {
self.unmark_object(scan);
self.lastgc_live_mem += self.get_size(scan);
} else {
self.lastgc_free_mem += self.get_size(scan);
if tail == 0 {
self.head = scan;
self.set_fl_next(scan, 0);
tail = scan;
} else {
if self.next_object_in_heap(tail) == scan {
self.set_size(tail, self.get_size(tail) + self.get_size(scan));
} else {
self.set_fl_next(tail, scan);
self.set_fl_next(scan, 0);
tail = scan;
}
}
}
scan = self.next_object_in_heap(scan);
}
if self.show_free_list {
self.print_freelist();
}
if self.show_heap_map {
self.print_heap();
}
}
fn mark_and_scan(&mut self, object: usize) {
if object == 0 || self.is_marked(object) | {
return;
} | conditional_block |
|
utils.py | ignore_columns)
columns = list(set(df.columns) - set(ignore_columns))
## define possible strategies
if how == "binarizer":
strategy = lambda x: pd.get_dummies(x)
elif how == "ordinal":
enc = OrdinalEncoder()
strategy = lambda x: pd.DataFrame(enc.fit_transform(x), columns = x.columns)
else:
print("strategy not implemented (yet!). Using pandas.get_dummies() instead!")
strategy = lambda x: pd.get_dummies(x)
cols = []
## go through all remaining columns, check if 'object' features exist
for column in columns:
if pd.api.types.is_string_dtype(df[column]):
cols.append(column)
## get all other columns from data
other_columns = list(set(df.columns) - set(cols))
## get both subdatasets - the encoded one and the remaining original one
encoded_data_raw = strategy(df[cols])
data_raw = df[other_columns]
## merge both subdatasets
encoded_data = pd.concat([encoded_data_raw, data_raw], axis = 1)
return encoded_data
def check_nan(data:pd.Series) -> bool:
'''
checks whether given data contains NaN's
Parameters:
- data: data to check [pandas.Series], can also be pandas.DataFrame
Returns:
- nan's: True, if data contains NaN's, otherwise False [Boolean]
'''
## make sure not to overwrite given data
df = data.copy()
if (not type(df) == pd.DataFrame) and (not type(df) == pd.Series):
print("data is no pandas.DataFrame, no check for NaN's done")
return False
if type(df) == pd.DataFrame:
return data.isna().sum().sum().astype(bool)
return data.isna().sum().astype(bool)
def add_nan(data:pd.DataFrame, amount:float = 0.05) -> pd.DataFrame:
'''
taking the given DataFrame and randomly adds the given amount of NaN's into it
Parameters:
- data: given data to add NaN's to [pandas.DataFrame]
- amount: desired amount of NaN's [Float, default = 0.05]
Returns:
- nan_data: data containing desired amount of NaN's [pandas.DataFrame]
'''
## set a numpy array with <amount> number of `True`s in the shape of data
nan_array = np.random.random(data.shape) < amount
## mask every element in 'data' with an NaN, when that element in 'nan_array' is set to True
nan_data = data.mask(nan_array)
return nan_data
def check_numeric(data:pd.DataFrame, ignore_columns:list = []) -> pd.DataFrame:
'''
function that converts all columns in DataFrame into numeric ones. Deletes all columns where `pandas.to_numeric()` fails (as they seem to be Strings)
Parameters:
- data: DataFrame with all different kinds of column types [pandas.DataFrame]
- ignore_columns: List of columns that shall be ignored [List, default = []]
Returns:
- df: DataFrame with converted columns and without deleted columns [pandas.DataFrame]
'''
## make sure not to overwrite given data
df = data.copy()
## check if data is pandas.DataFrame
if not type(df) == pd.DataFrame:
if type(df) == pd.Series:
num_df = pd.to_numeric(df, errors = "coerce")
if num_df.isna().sum() > 0:
print("data cannot be converted to numerical data, you have to encode it")
return df
return num_df
print("data is no pandas.DataFrame, cannot be further processed")
return df
## remaining columns that shall be checked (all - ignore_columns)
columns = list(set(df.columns) - set(ignore_columns))
## iterate over all columns, convert them to numeric ones (float or int)
for col in tqdm(columns, desc="make all columns numerical"):
## if error, then fill with NaN
df[col] = pd.to_numeric(df[col], errors="coerce")
## drop all columns that contain NaN's
df = df.dropna(axis=1)
return df
def iter_columns(data:pd.DataFrame, columns:list, trait:str) -> str:
'''
iterator, going over all columns in DataFrame, checking their content for desired trait
Parameters:
- data: DataFrame to iterate over [pandas.DataFrame]
- columns: columns to check [List]
- trait: what shall the column be checked for. Possible values [String]
- Unique: check for unique values per column, returns columns consisting of the same value over all samples --> 'unique'
Returns:
- col: column that contains only one different value [String]
'''
## iterate over all given columns
for col in tqdm(columns, desc=f"handle {trait}'s'"):
## check for Unique's
if trait == "unique":
## check if column contains more than one different value
if data[col].unique().__len__() == 1:
## if yes, return that column
yield col
def handle_nans(data:pd.DataFrame, strategy:str = "null", ignore_columns:list = []) -> pd.DataFrame:
'''
function that drops all columns (=features) that only contain one different value
Parameters:
- data: DataFrame [pandas.DataFrame]
- strategy: strategy to fill in the dataset. Possible values are [String]
- 0: fill all with Zero --> 'null' = default
- Mean: fill all with mean of respective feature --> 'mean'
- Median: fill all with median of respective feature --> 'median'
- Max: fill all with max of respective feature --> 'max'
- Min: fill all with min of respective feature --> 'min'
- ignore_columns: List of columns that shall be ignored [List, default = []]
Returns:
- df: DataFrame without deleted columns [pandas.DataFrame]
'''
## make sure not to overwrite given data
df = data.copy()
## check if data is data contains NaN's
if not check_nan(df):
print("no NaN's inside data")
return df
## remaining columns that shall be checked (all - ignore_columns)
columns = list(set(df.columns) - set(ignore_columns))
## init columns to drop
cols = []
## check strategy, calculate filling value(s)
if strategy == "null":
value = [0 for _ in range(columns.__len__())]
elif strategy == "mean":
value = df[columns].mean()
elif strategy == "median":
value = df[columns].median()
elif strategy == "min":
value = df[columns].min()
elif strategy == "max":
value = df[columns].max()
else:
print("strategy not implemented (yet). Filling with 0")
value = [0 for _ in range(columns.__len__())]
df = df.fillna(dict(zip(columns, value)))
## drop columns that ONLY contain NaN's, no matter what 'ignore_columns' says
df = df.dropna(how = "all", axis = 1)
return df
def handle_uniques(data:pd.DataFrame, ignore_columns:list = []) -> pd.DataFrame:
'''
function that handles all columns (=features) that only contain one different value by dropping them --> they do not contain helpful (any) information
Parameters:
- data: DataFrame [pandas.DataFrame]
- ignore_columns: List of columns that shall be ignored [List, default = []]
Returns:
- df: DataFrame without deleted columns [pandas.DataFrame]
'''
## make sure not to overwrite given data
df = data.copy()
## check if data is pandas.DataFrame
if not type(df) == pd.DataFrame:
print("data is no pandas.DataFrame, cannot be further processed")
return df
## remaining columns that shall be checked (all - ignore_columns)
columns = list(set(df.columns) - set(ignore_columns))
## init columns to drop
cols = []
## make sure not to overwrite given data
df = data.copy()
for col in iter_columns(df, columns, "unique"):
cols.append(col)
df = df.drop(cols, axis=1)
return df
def drop_features(data:pd.DataFrame, columns:list = []) -> pd.DataFrame:
'''
function that drops all columns are given by `columns`
Parameters:
- data: DataFrame with time columns [pandas.DataFrame]
- columns: List of columns that shall be deleted [List, default = []]
Returns:
- df: DataFrame without deleted columns [pandas.DataFrame]
'''
## make sure not to overwrite given data or the given columns
cols = columns.copy()
df = data.copy()
## check if data is pandas.DataFrame
if not type(df) == pd.DataFrame:
print("data is no pandas.DataFrame, cannot be further processed")
return df
df = df.drop(cols, axis=1)
return df
def flatten(X:np.array) -> np.array:
| '''
flattens a 3D array into 2D array
Parameters:
- X: a 3D array with shape samples x width x height [numpy.array]
Returns:
- flattened_X: 2D array with shape sample x width*height [numpy.array]
'''
flattened_X = X.reshape(X.shape[0], -1)
return flattened_X | identifier_body |
|
utils.py | def check_nan(data:pd.Series) -> bool:
'''
checks whether given data contains NaN's
Parameters:
- data: data to check [pandas.Series], can also be pandas.DataFrame
Returns:
- nan's: True, if data contains NaN's, otherwise False [Boolean]
'''
## make sure not to overwrite given data
df = data.copy()
if (not type(df) == pd.DataFrame) and (not type(df) == pd.Series):
print("data is no pandas.DataFrame, no check for NaN's done")
return False
if type(df) == pd.DataFrame:
return data.isna().sum().sum().astype(bool)
return data.isna().sum().astype(bool)
def add_nan(data:pd.DataFrame, amount:float = 0.05) -> pd.DataFrame:
'''
taking the given DataFrame and randomly adds the given amount of NaN's into it
Parameters:
- data: given data to add NaN's to [pandas.DataFrame]
- amount: desired amount of NaN's [Float, default = 0.05]
Returns:
- nan_data: data containing desired amount of NaN's [pandas.DataFrame]
'''
## set a numpy array with <amount> number of `True`s in the shape of data
nan_array = np.random.random(data.shape) < amount
## mask every element in 'data' with an NaN, when that element in 'nan_array' is set to True
nan_data = data.mask(nan_array)
return nan_data
def check_numeric(data:pd.DataFrame, ignore_columns:list = []) -> pd.DataFrame:
'''
function that converts all columns in DataFrame into numeric ones. Deletes all columns where `pandas.to_numeric()` fails (as they seem to be Strings)
Parameters:
- data: DataFrame with all different kinds of column types [pandas.DataFrame]
- ignore_columns: List of columns that shall be ignored [List, default = []]
Returns:
- df: DataFrame with converted columns and without deleted columns [pandas.DataFrame]
'''
## make sure not to overwrite given data
df = data.copy()
## check if data is pandas.DataFrame
if not type(df) == pd.DataFrame:
if type(df) == pd.Series:
num_df = pd.to_numeric(df, errors = "coerce")
if num_df.isna().sum() > 0:
print("data cannot be converted to numerical data, you have to encode it")
return df
return num_df
print("data is no pandas.DataFrame, cannot be further processed")
return df
## remaining columns that shall be checked (all - ignore_columns)
columns = list(set(df.columns) - set(ignore_columns))
## iterate over all columns, convert them to numeric ones (float or int)
for col in tqdm(columns, desc="make all columns numerical"):
## if error, then fill with NaN
df[col] = pd.to_numeric(df[col], errors="coerce")
## drop all columns that contain NaN's
df = df.dropna(axis=1)
return df
def iter_columns(data:pd.DataFrame, columns:list, trait:str) -> str:
'''
iterator, going over all columns in DataFrame, checking their content for desired trait
Parameters:
- data: DataFrame to iterate over [pandas.DataFrame]
- columns: columns to check [List]
- trait: what shall the column be checked for. Possible values [String]
- Unique: check for unique values per column, returns columns consisting of the same value over all samples --> 'unique'
Returns:
- col: column that contains only one different value [String]
'''
## iterate over all given columns
for col in tqdm(columns, desc=f"handle {trait}'s'"):
## check for Unique's
if trait == "unique":
## check if column contains more than one different value
if data[col].unique().__len__() == 1:
## if yes, return that column
yield col
def handle_nans(data:pd.DataFrame, strategy:str = "null", ignore_columns:list = []) -> pd.DataFrame:
'''
function that drops all columns (=features) that only contain one different value
Parameters:
- data: DataFrame [pandas.DataFrame]
- strategy: strategy to fill in the dataset. Possible values are [String]
- 0: fill all with Zero --> 'null' = default
- Mean: fill all with mean of respective feature --> 'mean'
- Median: fill all with median of respective feature --> 'median'
- Max: fill all with max of respective feature --> 'max'
- Min: fill all with min of respective feature --> 'min'
- ignore_columns: List of columns that shall be ignored [List, default = []]
Returns:
- df: DataFrame without deleted columns [pandas.DataFrame]
'''
## make sure not to overwrite given data
df = data.copy()
## check if data is data contains NaN's
if not check_nan(df):
print("no NaN's inside data")
return df
## remaining columns that shall be checked (all - ignore_columns)
columns = list(set(df.columns) - set(ignore_columns))
## init columns to drop
cols = []
## check strategy, calculate filling value(s)
if strategy == "null":
value = [0 for _ in range(columns.__len__())]
elif strategy == "mean":
value = df[columns].mean()
elif strategy == "median":
value = df[columns].median()
elif strategy == "min":
value = df[columns].min()
elif strategy == "max":
value = df[columns].max()
else:
print("strategy not implemented (yet). Filling with 0")
value = [0 for _ in range(columns.__len__())]
df = df.fillna(dict(zip(columns, value)))
## drop columns that ONLY contain NaN's, no matter what 'ignore_columns' says
df = df.dropna(how = "all", axis = 1)
return df
def handle_uniques(data:pd.DataFrame, ignore_columns:list = []) -> pd.DataFrame:
'''
function that handles all columns (=features) that only contain one different value by dropping them --> they do not contain helpful (any) information
Parameters:
- data: DataFrame [pandas.DataFrame]
- ignore_columns: List of columns that shall be ignored [List, default = []]
Returns:
- df: DataFrame without deleted columns [pandas.DataFrame]
'''
## make sure not to overwrite given data
df = data.copy()
## check if data is pandas.DataFrame
if not type(df) == pd.DataFrame:
print("data is no pandas.DataFrame, cannot be further processed")
return df
## remaining columns that shall be checked (all - ignore_columns)
columns = list(set(df.columns) - set(ignore_columns))
## init columns to drop
cols = []
## make sure not to overwrite given data
df = data.copy()
for col in iter_columns(df, columns, "unique"):
cols.append(col)
df = df.drop(cols, axis=1)
return df
def drop_features(data:pd.DataFrame, columns:list = []) -> pd.DataFrame:
'''
function that drops all columns are given by `columns`
Parameters:
- data: DataFrame with time columns [pandas.DataFrame]
- columns: List of columns that shall be deleted [List, default = []]
Returns:
- df: DataFrame without deleted columns [pandas.DataFrame]
'''
## make sure not to overwrite given data or the given columns
cols = columns.copy()
df = data.copy()
## check if data is pandas.DataFrame
if not type(df) == pd.DataFrame:
print("data is no pandas.DataFrame, cannot be further processed")
return df
df = df.drop(cols, axis=1)
return df
def flatten(X:np.array) -> np.array:
'''
flattens a 3D array into 2D array
Parameters:
- X: a 3D array with shape samples x width x height [numpy.array]
Returns:
- flattened_X: 2D array with shape sample x width*height [numpy.array]
'''
flattened_X = X.reshape(X.shape[0], -1)
return flattened_X
def iter_scale(X:np.array, scaler:object) -> np.array:
'''
iterates over fiven X, scales given 3D array using given (trained) scaler
Parameters:
- X: 3D array with shape samples x length x features [numpy.array]
- scaler: scaler object, e.g., sklearn.preprocessing.StandardScaler, sklearn.preprocessing.normalize [object]
Returns:
- scaled_X: scaled 3D array of same shape [numpy.array]
'''
## copy X to make sure not to overwrite the original data
scaled_X = X.copy()
for X_to_scale in tqdm(scaled_X):
yield scaler.transform(X_to_scale.reshape(1, -1)).reshape(X_to_scale.shape)
def pca(X:np.array, y:np.array) -> (np.array, np.array):
''' | plots a scatter showing the transformed dataset (if it is 2D) with different coloring for the different classes
Parameters:
- X: Array containing the original x values [numpy.array]
- y: Array containing the labels [numpy.array]
Returns: | random_line_split |
|
utils.py | List of columns that shall be ignored [List, default = []]
- how: strategy to encode. The following are possible [String]
- Binarize: every unique value gets own column, filled with 0's and 1's, using pandas.get_dummies() --> 'binarizer' = Default
- OrdinalEncoder: unique values get replaced by increasing number (same amount of features) using sklearn's OrdinalEncoder --> 'ordinal'
Returns:
- encoded_data: encoded DataFrame [pandas.DataFrame]
'''
## make sure not to overwrite given data
df = data.copy()
## check if data is pandas.DataFrame
if not type(df) == pd.DataFrame:
if type(df) == pd.Series:
df = pd.DataFrame(data)
df.columns = ["Series_Data"]
else:
print("data is no pandas.DataFrame, cannot be further processed")
return df
## remaining columns that shall be checked (all - ignore_columns)
columns = list(set(df.columns) - set(ignore_columns))
## define possible strategies
if how == "binarizer":
strategy = lambda x: pd.get_dummies(x)
elif how == "ordinal":
enc = OrdinalEncoder()
strategy = lambda x: pd.DataFrame(enc.fit_transform(x), columns = x.columns)
else:
print("strategy not implemented (yet!). Using pandas.get_dummies() instead!")
strategy = lambda x: pd.get_dummies(x)
cols = []
## go through all remaining columns, check if 'object' features exist
for column in columns:
if pd.api.types.is_string_dtype(df[column]):
cols.append(column)
## get all other columns from data
other_columns = list(set(df.columns) - set(cols))
## get both subdatasets - the encoded one and the remaining original one
encoded_data_raw = strategy(df[cols])
data_raw = df[other_columns]
## merge both subdatasets
encoded_data = pd.concat([encoded_data_raw, data_raw], axis = 1)
return encoded_data
def check_nan(data:pd.Series) -> bool:
'''
checks whether given data contains NaN's
Parameters:
- data: data to check [pandas.Series], can also be pandas.DataFrame
Returns:
- nan's: True, if data contains NaN's, otherwise False [Boolean]
'''
## make sure not to overwrite given data
df = data.copy()
if (not type(df) == pd.DataFrame) and (not type(df) == pd.Series):
print("data is no pandas.DataFrame, no check for NaN's done")
return False
if type(df) == pd.DataFrame:
return data.isna().sum().sum().astype(bool)
return data.isna().sum().astype(bool)
def add_nan(data:pd.DataFrame, amount:float = 0.05) -> pd.DataFrame:
'''
taking the given DataFrame and randomly adds the given amount of NaN's into it
Parameters:
- data: given data to add NaN's to [pandas.DataFrame]
- amount: desired amount of NaN's [Float, default = 0.05]
Returns:
- nan_data: data containing desired amount of NaN's [pandas.DataFrame]
'''
## set a numpy array with <amount> number of `True`s in the shape of data
nan_array = np.random.random(data.shape) < amount
## mask every element in 'data' with an NaN, when that element in 'nan_array' is set to True
nan_data = data.mask(nan_array)
return nan_data
def check_numeric(data:pd.DataFrame, ignore_columns:list = []) -> pd.DataFrame:
'''
function that converts all columns in DataFrame into numeric ones. Deletes all columns where `pandas.to_numeric()` fails (as they seem to be Strings)
Parameters:
- data: DataFrame with all different kinds of column types [pandas.DataFrame]
- ignore_columns: List of columns that shall be ignored [List, default = []]
Returns:
- df: DataFrame with converted columns and without deleted columns [pandas.DataFrame]
'''
## make sure not to overwrite given data
df = data.copy()
## check if data is pandas.DataFrame
if not type(df) == pd.DataFrame:
if type(df) == pd.Series:
num_df = pd.to_numeric(df, errors = "coerce")
if num_df.isna().sum() > 0:
print("data cannot be converted to numerical data, you have to encode it")
return df
return num_df
print("data is no pandas.DataFrame, cannot be further processed")
return df
## remaining columns that shall be checked (all - ignore_columns)
columns = list(set(df.columns) - set(ignore_columns))
## iterate over all columns, convert them to numeric ones (float or int)
for col in tqdm(columns, desc="make all columns numerical"):
## if error, then fill with NaN
df[col] = pd.to_numeric(df[col], errors="coerce")
## drop all columns that contain NaN's
df = df.dropna(axis=1)
return df
def iter_columns(data:pd.DataFrame, columns:list, trait:str) -> str:
'''
iterator, going over all columns in DataFrame, checking their content for desired trait
Parameters:
- data: DataFrame to iterate over [pandas.DataFrame]
- columns: columns to check [List]
- trait: what shall the column be checked for. Possible values [String]
- Unique: check for unique values per column, returns columns consisting of the same value over all samples --> 'unique'
Returns:
- col: column that contains only one different value [String]
'''
## iterate over all given columns
for col in tqdm(columns, desc=f"handle {trait}'s'"):
## check for Unique's
if trait == "unique":
## check if column contains more than one different value
if data[col].unique().__len__() == 1:
## if yes, return that column
yield col
def handle_nans(data:pd.DataFrame, strategy:str = "null", ignore_columns:list = []) -> pd.DataFrame:
'''
function that drops all columns (=features) that only contain one different value
Parameters:
- data: DataFrame [pandas.DataFrame]
- strategy: strategy to fill in the dataset. Possible values are [String]
- 0: fill all with Zero --> 'null' = default
- Mean: fill all with mean of respective feature --> 'mean'
- Median: fill all with median of respective feature --> 'median'
- Max: fill all with max of respective feature --> 'max'
- Min: fill all with min of respective feature --> 'min'
- ignore_columns: List of columns that shall be ignored [List, default = []]
Returns:
- df: DataFrame without deleted columns [pandas.DataFrame]
'''
## make sure not to overwrite given data
df = data.copy()
## check if data is data contains NaN's
if not check_nan(df):
print("no NaN's inside data")
return df
## remaining columns that shall be checked (all - ignore_columns)
columns = list(set(df.columns) - set(ignore_columns))
## init columns to drop
cols = []
## check strategy, calculate filling value(s)
if strategy == "null":
|
elif strategy == "mean":
value = df[columns].mean()
elif strategy == "median":
value = df[columns].median()
elif strategy == "min":
value = df[columns].min()
elif strategy == "max":
value = df[columns].max()
else:
print("strategy not implemented (yet). Filling with 0")
value = [0 for _ in range(columns.__len__())]
df = df.fillna(dict(zip(columns, value)))
## drop columns that ONLY contain NaN's, no matter what 'ignore_columns' says
df = df.dropna(how = "all", axis = 1)
return df
def handle_uniques(data:pd.DataFrame, ignore_columns:list = []) -> pd.DataFrame:
'''
function that handles all columns (=features) that only contain one different value by dropping them --> they do not contain helpful (any) information
Parameters:
- data: DataFrame [pandas.DataFrame]
- ignore_columns: List of columns that shall be ignored [List, default = []]
Returns:
- df: DataFrame without deleted columns [pandas.DataFrame]
'''
## make sure not to overwrite given data
df = data.copy()
## check if data is pandas.DataFrame
if not type(df) == pd.DataFrame:
print("data is no pandas.DataFrame, cannot be further processed")
return df
## remaining columns that shall be checked (all - ignore_columns)
columns = list(set(df.columns) - set(ignore_columns))
## init columns to drop
cols = []
## make sure not to overwrite given data
df = data.copy()
for col in iter_columns(df, columns, "unique"):
cols.append(col)
df = df.drop(cols, axis=1)
return df
def drop_features(data:pd.DataFrame, columns:list = []) -> pd.DataFrame:
'''
function that drops all columns are given by `columns`
Parameters:
- data: DataFrame with time columns [pandas.DataFrame]
- | value = [0 for _ in range(columns.__len__())] | conditional_block |
utils.py | (data:pd.DataFrame, approach:str = "reuse") -> pd.DataFrame:
'''
function to balance the appearance of samples from different classes -> tackle distribution shift
Parameters:
- data: data with distribution shift [pandas.DataFrame]
- approach: strategy to tackle the distribution shift. Possible values are [String]
- reusing minor class samples --> 'reuse' = default
- mixing both approaches by using the size of the median common class --> "mix"
- constraining the number of major class samples --> 'constrain'
Returns:
- df: data without distribution shift [pandas.DataFrame]
'''
## make sure not to overwrite given data
df = data.copy()
## get all labels that exist
labels = data.loc[:, "Label"]
## get appearances of each label
counted_labels = Counter(labels).most_common()
## get max num of samples (valid for all classes)
if approach == "reuse":
## take appearance value of most common label
sample_size = counted_labels[0][1]
elif approach == "mix":
sample_size = counted_labels[int(counted_labels.__len__()*0.5)][1]
elif approach == "constrain":
## take appearance value of least common label
sample_size = counted_labels[-1][1]
else:
print("approach not implemented (yet)! Using 'resue' instead!")
## take appearance value of most common label
sample_size = counted_labels[0][1]
## take a 'subset' or 'superset' of every class
sampled_data = [df[df.Label == label].sample(n = sample_size, replace = True) for label in np.unique(labels)]
## return merged data
return pd.concat(sampled_data).reset_index(drop = True)
def encode_objects(data:pd.DataFrame, ignore_columns:list = [], how:str = "binarizer") -> pd.DataFrame:
'''
goes through given dataset, encodes all object columns into numerical data
Parameters:
- data: DataFrame to anaylse [pandas.DataFrame]
- ignore_columns: List of columns that shall be ignored [List, default = []]
- how: strategy to encode. The following are possible [String]
- Binarize: every unique value gets own column, filled with 0's and 1's, using pandas.get_dummies() --> 'binarizer' = Default
- OrdinalEncoder: unique values get replaced by increasing number (same amount of features) using sklearn's OrdinalEncoder --> 'ordinal'
Returns:
- encoded_data: encoded DataFrame [pandas.DataFrame]
'''
## make sure not to overwrite given data
df = data.copy()
## check if data is pandas.DataFrame
if not type(df) == pd.DataFrame:
if type(df) == pd.Series:
df = pd.DataFrame(data)
df.columns = ["Series_Data"]
else:
print("data is no pandas.DataFrame, cannot be further processed")
return df
## remaining columns that shall be checked (all - ignore_columns)
columns = list(set(df.columns) - set(ignore_columns))
## define possible strategies
if how == "binarizer":
strategy = lambda x: pd.get_dummies(x)
elif how == "ordinal":
enc = OrdinalEncoder()
strategy = lambda x: pd.DataFrame(enc.fit_transform(x), columns = x.columns)
else:
print("strategy not implemented (yet!). Using pandas.get_dummies() instead!")
strategy = lambda x: pd.get_dummies(x)
cols = []
## go through all remaining columns, check if 'object' features exist
for column in columns:
if pd.api.types.is_string_dtype(df[column]):
cols.append(column)
## get all other columns from data
other_columns = list(set(df.columns) - set(cols))
## get both subdatasets - the encoded one and the remaining original one
encoded_data_raw = strategy(df[cols])
data_raw = df[other_columns]
## merge both subdatasets
encoded_data = pd.concat([encoded_data_raw, data_raw], axis = 1)
return encoded_data
def check_nan(data:pd.Series) -> bool:
'''
checks whether given data contains NaN's
Parameters:
- data: data to check [pandas.Series], can also be pandas.DataFrame
Returns:
- nan's: True, if data contains NaN's, otherwise False [Boolean]
'''
## make sure not to overwrite given data
df = data.copy()
if (not type(df) == pd.DataFrame) and (not type(df) == pd.Series):
print("data is no pandas.DataFrame, no check for NaN's done")
return False
if type(df) == pd.DataFrame:
return data.isna().sum().sum().astype(bool)
return data.isna().sum().astype(bool)
def add_nan(data:pd.DataFrame, amount:float = 0.05) -> pd.DataFrame:
'''
taking the given DataFrame and randomly adds the given amount of NaN's into it
Parameters:
- data: given data to add NaN's to [pandas.DataFrame]
- amount: desired amount of NaN's [Float, default = 0.05]
Returns:
- nan_data: data containing desired amount of NaN's [pandas.DataFrame]
'''
## set a numpy array with <amount> number of `True`s in the shape of data
nan_array = np.random.random(data.shape) < amount
## mask every element in 'data' with an NaN, when that element in 'nan_array' is set to True
nan_data = data.mask(nan_array)
return nan_data
def check_numeric(data:pd.DataFrame, ignore_columns:list = []) -> pd.DataFrame:
'''
function that converts all columns in DataFrame into numeric ones. Deletes all columns where `pandas.to_numeric()` fails (as they seem to be Strings)
Parameters:
- data: DataFrame with all different kinds of column types [pandas.DataFrame]
- ignore_columns: List of columns that shall be ignored [List, default = []]
Returns:
- df: DataFrame with converted columns and without deleted columns [pandas.DataFrame]
'''
## make sure not to overwrite given data
df = data.copy()
## check if data is pandas.DataFrame
if not type(df) == pd.DataFrame:
if type(df) == pd.Series:
num_df = pd.to_numeric(df, errors = "coerce")
if num_df.isna().sum() > 0:
print("data cannot be converted to numerical data, you have to encode it")
return df
return num_df
print("data is no pandas.DataFrame, cannot be further processed")
return df
## remaining columns that shall be checked (all - ignore_columns)
columns = list(set(df.columns) - set(ignore_columns))
## iterate over all columns, convert them to numeric ones (float or int)
for col in tqdm(columns, desc="make all columns numerical"):
## if error, then fill with NaN
df[col] = pd.to_numeric(df[col], errors="coerce")
## drop all columns that contain NaN's
df = df.dropna(axis=1)
return df
def iter_columns(data:pd.DataFrame, columns:list, trait:str) -> str:
'''
iterator, going over all columns in DataFrame, checking their content for desired trait
Parameters:
- data: DataFrame to iterate over [pandas.DataFrame]
- columns: columns to check [List]
- trait: what shall the column be checked for. Possible values [String]
- Unique: check for unique values per column, returns columns consisting of the same value over all samples --> 'unique'
Returns:
- col: column that contains only one different value [String]
'''
## iterate over all given columns
for col in tqdm(columns, desc=f"handle {trait}'s'"):
## check for Unique's
if trait == "unique":
## check if column contains more than one different value
if data[col].unique().__len__() == 1:
## if yes, return that column
yield col
def handle_nans(data:pd.DataFrame, strategy:str = "null", ignore_columns:list = []) -> pd.DataFrame:
'''
function that drops all columns (=features) that only contain one different value
Parameters:
- data: DataFrame [pandas.DataFrame]
- strategy: strategy to fill in the dataset. Possible values are [String]
- 0: fill all with Zero --> 'null' = default
- Mean: fill all with mean of respective feature --> 'mean'
- Median: fill all with median of respective feature --> 'median'
- Max: fill all with max of respective feature --> 'max'
- Min: fill all with min of respective feature --> 'min'
- ignore_columns: List of columns that shall be ignored [List, default = []]
Returns:
- df: DataFrame without deleted columns [pandas.DataFrame]
'''
## make sure not to overwrite given data
df = data.copy()
## check if data is data contains NaN's
if not check_nan(df):
print("no NaN's inside data")
return df
## remaining columns that shall be checked (all - ignore_columns)
columns = list(set(df.columns) - set(ignore_columns))
## init columns to drop
cols = []
## check strategy, calculate filling value(s)
if strategy == "null | tackle_distribution_shift | identifier_name |
|
models.go | exported.
ReleasePolicy *KeyReleasePolicy
// Application specific metadata in the form of key-value pairs.
Tags map[string]*string
}
// DeletedKey - A DeletedKey consisting of a WebKey plus its Attributes and deletion info
type DeletedKey struct {
// The key management attributes.
Attributes *KeyAttributes
// The Json web key.
Key *JSONWebKey
// The url of the recovery object, used to identify and recover the deleted key.
RecoveryID *string
// The policy rules under which the key can be exported.
ReleasePolicy *KeyReleasePolicy
// Application specific metadata in the form of key-value pairs.
Tags map[string]*string
// READ-ONLY; The time when the key was deleted, in UTC
DeletedDate *time.Time
// READ-ONLY; True if the key's lifetime is managed by key vault. If this is a key backing a certificate, then managed will
// be true.
Managed *bool
// READ-ONLY; The time when the key is scheduled to be purged, in UTC
ScheduledPurgeDate *time.Time
}
// DeletedKeyProperties - The deleted key item containing the deleted key metadata and information about deletion.
type DeletedKeyProperties struct {
// The key management attributes.
Attributes *KeyAttributes
// Key identifier.
KID *ID
// The url of the recovery object, used to identify and recover the deleted key.
RecoveryID *string
// Application specific metadata in the form of key-value pairs.
Tags map[string]*string
// READ-ONLY; The time when the key was deleted, in UTC
DeletedDate *time.Time
// READ-ONLY; True if the key's lifetime is managed by key vault. If this is a key backing a certificate, then managed will
// be true.
Managed *bool
// READ-ONLY; The time when the key is scheduled to be purged, in UTC
ScheduledPurgeDate *time.Time
}
// DeletedKeyPropertiesListResult - A list of keys that have been deleted in this vault.
type DeletedKeyPropertiesListResult struct {
// READ-ONLY; The URL to get the next set of deleted keys.
NextLink *string
// READ-ONLY; A response message containing a list of deleted keys in the vault along with a link to the next page of deleted
// keys
Value []*DeletedKeyProperties
}
// GetRandomBytesParameters - The get random bytes request object.
type GetRandomBytesParameters struct {
// REQUIRED; The requested number of random bytes.
Count *int32
}
// ImportKeyParameters - The key import parameters.
type ImportKeyParameters struct {
// REQUIRED; The Json web key
Key *JSONWebKey
// Whether to import as a hardware key (HSM) or software key.
HSM *bool
// The key management attributes.
KeyAttributes *KeyAttributes
// The policy rules under which the key can be exported.
ReleasePolicy *KeyReleasePolicy
// Application specific metadata in the form of key-value pairs.
Tags map[string]*string
}
// JSONWebKey - As of http://tools.ietf.org/html/draft-ietf-jose-json-web-key-18
type JSONWebKey struct {
// Elliptic curve name.
Crv *CurveName
// RSA private exponent, or the D component of an EC private key.
D []byte
// RSA private key parameter.
DP []byte
// RSA private key parameter.
DQ []byte
// RSA public exponent.
E []byte
// Symmetric key.
K []byte
// Key identifier.
KID *ID
KeyOps []*KeyOperation
// JsonWebKey Key Type (kty), as defined in https://tools.ietf.org/html/draft-ietf-jose-json-web-algorithms-40.
Kty *KeyType
// RSA modulus.
N []byte
// RSA secret prime.
P []byte
// RSA secret prime, with p < q.
Q []byte
// RSA private key parameter.
QI []byte
// Protected Key, used with 'Bring Your Own Key'.
T []byte
// X component of an EC public key.
X []byte
// Y component of an EC public key.
Y []byte
}
// KeyAttributes - The attributes of a key managed by the key vault service.
type KeyAttributes struct {
// Determines whether the object is enabled.
Enabled *bool
// Expiry date in UTC.
Expires *time.Time
// Indicates if the private key can be exported. Release policy must be provided when creating the first version of an exportable
// key.
Exportable *bool
// Not before date in UTC.
NotBefore *time.Time
// READ-ONLY; Creation time in UTC.
Created *time.Time
// READ-ONLY; softDelete data retention days. Value should be >=7 and <=90 when softDelete enabled, otherwise 0.
RecoverableDays *int32
// READ-ONLY; Reflects the deletion recovery level currently in effect for keys in the current vault. If it contains 'Purgeable'
// the key can be permanently deleted by a privileged user; otherwise, only the system
// can purge the key, at the end of the retention interval.
RecoveryLevel *string
// READ-ONLY; Last updated time in UTC.
Updated *time.Time
}
// KeyBundle - A KeyBundle consisting of a WebKey plus its attributes.
type KeyBundle struct {
// The key management attributes.
Attributes *KeyAttributes
// The Json web key.
Key *JSONWebKey
// The policy rules under which the key can be exported.
ReleasePolicy *KeyReleasePolicy
// Application specific metadata in the form of key-value pairs.
Tags map[string]*string
// READ-ONLY; True if the key's lifetime is managed by key vault. If this is a key backing a certificate, then managed will
// be true.
Managed *bool
}
// KeyOperationParameters - The key operations parameters.
type KeyOperationParameters struct {
// REQUIRED; algorithm identifier
Algorithm *EncryptionAlgorithm
// REQUIRED
Value []byte
// Additional data to authenticate but not encrypt/decrypt when using authenticated crypto algorithms.
AdditionalAuthenticatedData []byte
// The tag to authenticate when performing decryption with an authenticated algorithm.
AuthenticationTag []byte
// Cryptographically random, non-repeating initialization vector for symmetric algorithms.
IV []byte
}
// KeyOperationResult - The key operation result.
type KeyOperationResult struct {
// READ-ONLY
AdditionalAuthenticatedData []byte
// READ-ONLY
AuthenticationTag []byte
// READ-ONLY
IV []byte
// READ-ONLY; Key identifier
KID *ID
// READ-ONLY
Result []byte
}
// KeyProperties - The key item containing key metadata.
type KeyProperties struct {
// The key management attributes.
Attributes *KeyAttributes
// Key identifier.
KID *ID
// Application specific metadata in the form of key-value pairs.
Tags map[string]*string
// READ-ONLY; True if the key's lifetime is managed by key vault. If this is a key backing a certificate, then managed will
// be true.
Managed *bool
}
// KeyPropertiesListResult - The key list result.
type KeyPropertiesListResult struct {
// READ-ONLY; The URL to get the next set of keys.
NextLink *string
// READ-ONLY; A response message containing a list of keys in the key vault along with a link to the next page of keys.
Value []*KeyProperties
}
// KeyReleasePolicy - The policy rules under which the key can be exported.
type KeyReleasePolicy struct {
// Content type and version of key release policy
ContentType *string
// Blob encoding the policy rules under which the key can be released. Blob must be base64 URL encoded.
EncodedPolicy []byte
// Defines the mutability state of the policy. Once marked immutable, this flag cannot be reset and the policy cannot be changed
// under any circumstances.
Immutable *bool
}
// KeyReleaseResult - The release result, containing the released key.
type KeyReleaseResult struct {
// READ-ONLY; A signed object containing the released key.
Value *string
}
// KeyRotationPolicy - Management policy for a key.
type KeyRotationPolicy struct {
// The key rotation policy attributes.
Attributes *KeyRotationPolicyAttributes
// Actions that will be performed by Key Vault over the lifetime of a key. For preview, lifetimeActions can only have two
// items at maximum: one for rotate, one for notify. Notification time would be
// default to 30 days before expiry and it is not configurable.
LifetimeActions []*LifetimeAction
// READ-ONLY; The key policy id.
ID *string
}
// KeyRotationPolicyAttributes - The key rotation policy attributes.
type KeyRotationPolicyAttributes struct {
// The expiryTime will be applied on the new key version. It should be at least 28 days. It will be in ISO 8601 Format. Examples:
// 90 days: P90D, 3 months: P3M, 48 hours: PT48H, 1 year and 10 days: P1Y10D | ExpiryTime *string
// READ-ONLY; The key rotation policy created time in UTC. | random_line_split |
|
stateful.go | state will then be the set of rules used for matching
// until another Push or Pop is encountered.
func Push(state string) Action {
return ActionPush{state}
}
type include struct{ state string }
func (i include) applyAction(lexer *Lexer, groups []string) error { panic("should not be called") }
func (i include) applyRules(state string, rule int, rules compiledRules) error {
includedRules, ok := rules[i.state]
if !ok {
return fmt.Errorf("invalid include state %q", i.state)
}
clone := make([]compiledRule, len(includedRules))
copy(clone, includedRules)
rules[state] = append(rules[state][:rule], append(clone, rules[state][rule+1:]...)...) // nolint: makezero
return nil
}
// Include rules from another state in this one.
func Include(state string) Rule {
return Rule{Action: include{state}}
}
// Definition is the lexer.Definition.
type Definition struct {
rules compiledRules
symbols map[string]rune
// Map of key->*regexp.Regexp
backrefCache sync.Map
initialState string
}
// MustSimple creates a new lexer definition based on a single state described by `rules`.
// panics if the rules trigger an error
func MustSimple(rules []Rule, options ...Option) *Definition {
def, err := NewSimple(rules, options...)
if err != nil {
panic(err)
}
return def
}
// Must creates a new stateful lexer and panics if it is incorrect.
func Must(rules Rules, options ...Option) *Definition {
def, err := New(rules, options...)
if err != nil {
panic(err)
}
return def
}
// NewSimple creates a new stateful lexer with a single "Root" state.
func NewSimple(rules []Rule, options ...Option) (*Definition, error) {
return New(Rules{"Root": rules}, options...)
}
// New constructs a new stateful lexer from rules.
func New(rules Rules, options ...Option) (*Definition, error) {
compiled := compiledRules{}
for key, set := range rules {
for i, rule := range set {
pattern := "^(?:" + rule.Pattern + ")"
var (
re *regexp.Regexp
err error
)
var match = backrefReplace.FindStringSubmatch(rule.Pattern)
if match == nil || len(match[1])%2 == 0 {
re, err = regexp.Compile(pattern)
if err != nil {
return nil, fmt.Errorf("%s.%d: %s", key, i, err)
}
}
compiled[key] = append(compiled[key], compiledRule{
Rule: rule,
ignore: len(rule.Name) > 0 && unicode.IsLower(rune(rule.Name[0])),
RE: re,
})
}
}
restart:
for state, rules := range compiled {
for i, rule := range rules {
if action, ok := rule.Action.(RulesAction); ok {
if err := action.applyRules(state, i, compiled); err != nil {
return nil, fmt.Errorf("%s.%d: %s", state, i, err)
}
goto restart
}
}
}
keys := make([]string, 0, len(compiled))
for key := range compiled {
keys = append(keys, key)
}
symbols := map[string]rune{
"EOF": lexer.EOF,
}
sort.Strings(keys)
duplicates := map[string]compiledRule{}
rn := lexer.EOF - 1
for _, key := range keys {
for i, rule := range compiled[key] {
if dup, ok := duplicates[rule.Name]; ok && rule.Pattern != dup.Pattern {
panic(fmt.Sprintf("duplicate key %q with different patterns %q != %q", rule.Name, rule.Pattern, dup.Pattern))
}
duplicates[rule.Name] = rule
compiled[key][i] = rule
symbols[rule.Name] = rn
rn--
}
}
d := &Definition{
initialState: "Root",
rules: compiled,
symbols: symbols,
}
for _, option := range options {
option(d)
}
return d, nil
}
// Rules returns the user-provided Rules used to construct the lexer.
func (d *Definition) Rules() Rules {
out := Rules{}
for state, rules := range d.rules {
for _, rule := range rules {
out[state] = append(out[state], rule.Rule)
}
}
return out
}
func (d *Definition) LexString(filename string, s string) (lexer.Lexer, error) { // nolint: golint
return &Lexer{
def: d,
data: s,
stack: []lexerState{{name: d.initialState}},
pos: lexer.Position{
Filename: filename,
Line: 1,
Column: 1,
},
}, nil
}
func (d *Definition) Lex(filename string, r io.Reader) (lexer.Lexer, error) { // nolint: golint
w := &strings.Builder{}
_, err := io.Copy(w, r)
if err != nil {
return nil, err
}
return d.LexString(filename, w.String())
}
func (d *Definition) Symbols() map[string]rune { // nolint: golint
return d.symbols
}
type lexerState struct {
name string
groups []string
}
// Lexer implementation.
type Lexer struct {
stack []lexerState
def *Definition
data string
pos lexer.Position
}
func (l *Lexer) Next() (lexer.Token, error) { // nolint: golint
parent := l.stack[len(l.stack)-1]
rules := l.def.rules[parent.name]
next:
for len(l.data) > 0 {
var (
rule *compiledRule
match []int
)
for _, candidate := range rules {
// Special case "Return()".
if candidate.Rule == ReturnRule {
l.stack = l.stack[:len(l.stack)-1]
parent = l.stack[len(l.stack)-1]
rules = l.def.rules[parent.name]
continue next
}
re, err := l.getPattern(candidate)
if err != nil {
return lexer.Token{}, participle.Wrapf(l.pos, err, "rule %q", candidate.Name)
}
match = re.FindStringSubmatchIndex(l.data)
if match != nil {
rule = &candidate // nolint
break
}
}
if match == nil || rule == nil {
sample := []rune(l.data)
if len(sample) > 16 {
sample = append(sample[:16], []rune("...")...)
}
return lexer.Token{}, participle.Errorf(l.pos, "invalid input text %q", string(sample))
}
if rule.Action != nil {
groups := make([]string, 0, len(match)/2)
for i := 0; i < len(match); i += 2 {
groups = append(groups, l.data[match[i]:match[i+1]])
}
if err := rule.Action.applyAction(l, groups); err != nil {
return lexer.Token{}, participle.Errorf(l.pos, "rule %q: %s", rule.Name, err)
}
} else if match[0] == match[1] {
return lexer.Token{}, participle.Errorf(l.pos, "rule %q did not match any input", rule.Name)
}
span := l.data[match[0]:match[1]]
l.data = l.data[match[1]:]
// l.groups = groups
// Update position.
pos := l.pos
l.pos.Offset += match[1]
lines := strings.Count(span, "\n")
l.pos.Line += lines
// Update column.
if lines == 0 {
l.pos.Column += utf8.RuneCountInString(span)
} else {
l.pos.Column = utf8.RuneCountInString(span[strings.LastIndex(span, "\n"):])
}
if rule.ignore {
parent = l.stack[len(l.stack)-1]
rules = l.def.rules[parent.name]
continue
}
return lexer.Token{
Type: l.def.symbols[rule.Name],
Value: span,
Pos: pos,
}, nil
}
return lexer.EOFToken(l.pos), nil
}
func (l *Lexer) getPattern(candidate compiledRule) (*regexp.Regexp, error) | {
if candidate.RE != nil {
return candidate.RE, nil
}
// We don't have a compiled RE. This means there are back-references
// that need to be substituted first.
parent := l.stack[len(l.stack)-1]
key := candidate.Pattern + "\000" + strings.Join(parent.groups, "\000")
cached, ok := l.def.backrefCache.Load(key)
if ok {
return cached.(*regexp.Regexp), nil
}
var (
re *regexp.Regexp
err error
)
pattern := backrefReplace.ReplaceAllStringFunc(candidate.Pattern, func(s string) string {
var rematch = backrefReplace.FindStringSubmatch(s) | identifier_body |
|
stateful.go | "
"github.com/alecthomas/participle/v2"
"github.com/alecthomas/participle/v2/lexer"
)
var (
backrefReplace = regexp.MustCompile(`(\\+)(\d)`)
)
// Option for modifying how the Lexer works.
type Option func(d *Definition)
// InitialState overrides the default initial state of "Root".
func InitialState(state string) Option {
return func(d *Definition) {
d.initialState = state
}
}
// A Rule matching input and possibly changing state.
type Rule struct {
Name string
Pattern string
Action Action
}
// Rules grouped by name.
type Rules map[string][]Rule
// compiledRule is a Rule with its pattern compiled.
type compiledRule struct {
Rule
ignore bool
RE *regexp.Regexp
}
// compiledRules grouped by name.
type compiledRules map[string][]compiledRule
// A Action is applied when a rule matches.
type Action interface {
// Actions are responsible for validating the match. ie. if they consumed any input.
applyAction(lexer *Lexer, groups []string) error
}
// RulesAction is an optional interface that Actions can implement.
//
// It is applied during rule construction to mutate the rule map.
type RulesAction interface {
applyRules(state string, rule int, rules compiledRules) error
}
// ActionPop pops to the previous state when the Rule matches. |
func (p ActionPop) applyAction(lexer *Lexer, groups []string) error {
if groups[0] == "" {
return errors.New("did not consume any input")
}
lexer.stack = lexer.stack[:len(lexer.stack)-1]
return nil
}
// Pop to the previous state.
func Pop() Action {
return ActionPop{}
}
// ReturnRule signals the lexer to return immediately.
var ReturnRule = Rule{"returnToParent", "", nil}
// Return to the parent state.
//
// Useful as the last rule in a sub-state.
func Return() Rule { return ReturnRule }
// ActionPush pushes the current state and switches to "State" when the Rule matches.
type ActionPush struct{ State string }
func (p ActionPush) applyAction(lexer *Lexer, groups []string) error {
if groups[0] == "" {
return errors.New("did not consume any input")
}
lexer.stack = append(lexer.stack, lexerState{name: p.State, groups: groups})
return nil
}
// Push to the given state.
//
// The target state will then be the set of rules used for matching
// until another Push or Pop is encountered.
func Push(state string) Action {
return ActionPush{state}
}
type include struct{ state string }
func (i include) applyAction(lexer *Lexer, groups []string) error { panic("should not be called") }
func (i include) applyRules(state string, rule int, rules compiledRules) error {
includedRules, ok := rules[i.state]
if !ok {
return fmt.Errorf("invalid include state %q", i.state)
}
clone := make([]compiledRule, len(includedRules))
copy(clone, includedRules)
rules[state] = append(rules[state][:rule], append(clone, rules[state][rule+1:]...)...) // nolint: makezero
return nil
}
// Include rules from another state in this one.
func Include(state string) Rule {
return Rule{Action: include{state}}
}
// Definition is the lexer.Definition.
type Definition struct {
rules compiledRules
symbols map[string]rune
// Map of key->*regexp.Regexp
backrefCache sync.Map
initialState string
}
// MustSimple creates a new lexer definition based on a single state described by `rules`.
// panics if the rules trigger an error
func MustSimple(rules []Rule, options ...Option) *Definition {
def, err := NewSimple(rules, options...)
if err != nil {
panic(err)
}
return def
}
// Must creates a new stateful lexer and panics if it is incorrect.
func Must(rules Rules, options ...Option) *Definition {
def, err := New(rules, options...)
if err != nil {
panic(err)
}
return def
}
// NewSimple creates a new stateful lexer with a single "Root" state.
func NewSimple(rules []Rule, options ...Option) (*Definition, error) {
return New(Rules{"Root": rules}, options...)
}
// New constructs a new stateful lexer from rules.
func New(rules Rules, options ...Option) (*Definition, error) {
compiled := compiledRules{}
for key, set := range rules {
for i, rule := range set {
pattern := "^(?:" + rule.Pattern + ")"
var (
re *regexp.Regexp
err error
)
var match = backrefReplace.FindStringSubmatch(rule.Pattern)
if match == nil || len(match[1])%2 == 0 {
re, err = regexp.Compile(pattern)
if err != nil {
return nil, fmt.Errorf("%s.%d: %s", key, i, err)
}
}
compiled[key] = append(compiled[key], compiledRule{
Rule: rule,
ignore: len(rule.Name) > 0 && unicode.IsLower(rune(rule.Name[0])),
RE: re,
})
}
}
restart:
for state, rules := range compiled {
for i, rule := range rules {
if action, ok := rule.Action.(RulesAction); ok {
if err := action.applyRules(state, i, compiled); err != nil {
return nil, fmt.Errorf("%s.%d: %s", state, i, err)
}
goto restart
}
}
}
keys := make([]string, 0, len(compiled))
for key := range compiled {
keys = append(keys, key)
}
symbols := map[string]rune{
"EOF": lexer.EOF,
}
sort.Strings(keys)
duplicates := map[string]compiledRule{}
rn := lexer.EOF - 1
for _, key := range keys {
for i, rule := range compiled[key] {
if dup, ok := duplicates[rule.Name]; ok && rule.Pattern != dup.Pattern {
panic(fmt.Sprintf("duplicate key %q with different patterns %q != %q", rule.Name, rule.Pattern, dup.Pattern))
}
duplicates[rule.Name] = rule
compiled[key][i] = rule
symbols[rule.Name] = rn
rn--
}
}
d := &Definition{
initialState: "Root",
rules: compiled,
symbols: symbols,
}
for _, option := range options {
option(d)
}
return d, nil
}
// Rules returns the user-provided Rules used to construct the lexer.
func (d *Definition) Rules() Rules {
out := Rules{}
for state, rules := range d.rules {
for _, rule := range rules {
out[state] = append(out[state], rule.Rule)
}
}
return out
}
func (d *Definition) LexString(filename string, s string) (lexer.Lexer, error) { // nolint: golint
return &Lexer{
def: d,
data: s,
stack: []lexerState{{name: d.initialState}},
pos: lexer.Position{
Filename: filename,
Line: 1,
Column: 1,
},
}, nil
}
func (d *Definition) Lex(filename string, r io.Reader) (lexer.Lexer, error) { // nolint: golint
w := &strings.Builder{}
_, err := io.Copy(w, r)
if err != nil {
return nil, err
}
return d.LexString(filename, w.String())
}
func (d *Definition) Symbols() map[string]rune { // nolint: golint
return d.symbols
}
type lexerState struct {
name string
groups []string
}
// Lexer implementation.
type Lexer struct {
stack []lexerState
def *Definition
data string
pos lexer.Position
}
func (l *Lexer) Next() (lexer.Token, error) { // nolint: golint
parent := l.stack[len(l.stack)-1]
rules := l.def.rules[parent.name]
next:
for len(l.data) > 0 {
var (
rule *compiledRule
match []int
)
for _, candidate := range rules {
// Special case "Return()".
if candidate.Rule == ReturnRule {
l.stack = l.stack[:len(l.stack)-1]
parent = l.stack[len(l.stack)-1]
rules = l.def.rules[parent.name]
continue next
}
re, err := l.getPattern(candidate)
if err != nil {
return lexer.Token{}, participle.Wrapf(l.pos, err, "rule %q", candidate.Name)
}
match = re.FindStringSubmatchIndex(l.data)
if match != nil {
rule = &candidate // nolint
break
}
}
if match == nil || rule == nil {
sample := []rune(l.data)
if len(sample) > 16 {
sample = append(sample[:16], []rune("...")...)
}
return lexer.Token{}, participle.Errorf(l.pos, "invalid input text %q", string(sample))
| type ActionPop struct{} | random_line_split |
stateful.go | "
"github.com/alecthomas/participle/v2"
"github.com/alecthomas/participle/v2/lexer"
)
var (
backrefReplace = regexp.MustCompile(`(\\+)(\d)`)
)
// Option for modifying how the Lexer works.
type Option func(d *Definition)
// InitialState overrides the default initial state of "Root".
func InitialState(state string) Option {
return func(d *Definition) {
d.initialState = state
}
}
// A Rule matching input and possibly changing state.
type Rule struct {
Name string
Pattern string
Action Action
}
// Rules grouped by name.
type Rules map[string][]Rule
// compiledRule is a Rule with its pattern compiled.
type compiledRule struct {
Rule
ignore bool
RE *regexp.Regexp
}
// compiledRules grouped by name.
type compiledRules map[string][]compiledRule
// A Action is applied when a rule matches.
type Action interface {
// Actions are responsible for validating the match. ie. if they consumed any input.
applyAction(lexer *Lexer, groups []string) error
}
// RulesAction is an optional interface that Actions can implement.
//
// It is applied during rule construction to mutate the rule map.
type RulesAction interface {
applyRules(state string, rule int, rules compiledRules) error
}
// ActionPop pops to the previous state when the Rule matches.
type ActionPop struct{}
func (p ActionPop) applyAction(lexer *Lexer, groups []string) error {
if groups[0] == "" {
return errors.New("did not consume any input")
}
lexer.stack = lexer.stack[:len(lexer.stack)-1]
return nil
}
// Pop to the previous state.
func Pop() Action {
return ActionPop{}
}
// ReturnRule signals the lexer to return immediately.
var ReturnRule = Rule{"returnToParent", "", nil}
// Return to the parent state.
//
// Useful as the last rule in a sub-state.
func Return() Rule { return ReturnRule }
// ActionPush pushes the current state and switches to "State" when the Rule matches.
type ActionPush struct{ State string }
func (p ActionPush) applyAction(lexer *Lexer, groups []string) error {
if groups[0] == "" {
return errors.New("did not consume any input")
}
lexer.stack = append(lexer.stack, lexerState{name: p.State, groups: groups})
return nil
}
// Push to the given state.
//
// The target state will then be the set of rules used for matching
// until another Push or Pop is encountered.
func Push(state string) Action {
return ActionPush{state}
}
type include struct{ state string }
func (i include) applyAction(lexer *Lexer, groups []string) error { panic("should not be called") }
func (i include) applyRules(state string, rule int, rules compiledRules) error {
includedRules, ok := rules[i.state]
if !ok {
return fmt.Errorf("invalid include state %q", i.state)
}
clone := make([]compiledRule, len(includedRules))
copy(clone, includedRules)
rules[state] = append(rules[state][:rule], append(clone, rules[state][rule+1:]...)...) // nolint: makezero
return nil
}
// Include rules from another state in this one.
func Include(state string) Rule {
return Rule{Action: include{state}}
}
// Definition is the lexer.Definition.
type Definition struct {
rules compiledRules
symbols map[string]rune
// Map of key->*regexp.Regexp
backrefCache sync.Map
initialState string
}
// MustSimple creates a new lexer definition based on a single state described by `rules`.
// panics if the rules trigger an error
func MustSimple(rules []Rule, options ...Option) *Definition {
def, err := NewSimple(rules, options...)
if err != nil {
panic(err)
}
return def
}
// Must creates a new stateful lexer and panics if it is incorrect.
func Must(rules Rules, options ...Option) *Definition {
def, err := New(rules, options...)
if err != nil {
panic(err)
}
return def
}
// NewSimple creates a new stateful lexer with a single "Root" state.
func NewSimple(rules []Rule, options ...Option) (*Definition, error) {
return New(Rules{"Root": rules}, options...)
}
// New constructs a new stateful lexer from rules.
func New(rules Rules, options ...Option) (*Definition, error) {
compiled := compiledRules{}
for key, set := range rules {
for i, rule := range set {
pattern := "^(?:" + rule.Pattern + ")"
var (
re *regexp.Regexp
err error
)
var match = backrefReplace.FindStringSubmatch(rule.Pattern)
if match == nil || len(match[1])%2 == 0 {
re, err = regexp.Compile(pattern)
if err != nil {
return nil, fmt.Errorf("%s.%d: %s", key, i, err)
}
}
compiled[key] = append(compiled[key], compiledRule{
Rule: rule,
ignore: len(rule.Name) > 0 && unicode.IsLower(rune(rule.Name[0])),
RE: re,
})
}
}
restart:
for state, rules := range compiled {
for i, rule := range rules {
if action, ok := rule.Action.(RulesAction); ok {
if err := action.applyRules(state, i, compiled); err != nil {
return nil, fmt.Errorf("%s.%d: %s", state, i, err)
}
goto restart
}
}
}
keys := make([]string, 0, len(compiled))
for key := range compiled {
keys = append(keys, key)
}
symbols := map[string]rune{
"EOF": lexer.EOF,
}
sort.Strings(keys)
duplicates := map[string]compiledRule{}
rn := lexer.EOF - 1
for _, key := range keys {
for i, rule := range compiled[key] {
if dup, ok := duplicates[rule.Name]; ok && rule.Pattern != dup.Pattern {
panic(fmt.Sprintf("duplicate key %q with different patterns %q != %q", rule.Name, rule.Pattern, dup.Pattern))
}
duplicates[rule.Name] = rule
compiled[key][i] = rule
symbols[rule.Name] = rn
rn--
}
}
d := &Definition{
initialState: "Root",
rules: compiled,
symbols: symbols,
}
for _, option := range options {
option(d)
}
return d, nil
}
// Rules returns the user-provided Rules used to construct the lexer.
func (d *Definition) Rules() Rules {
out := Rules{}
for state, rules := range d.rules {
for _, rule := range rules {
out[state] = append(out[state], rule.Rule)
}
}
return out
}
func (d *Definition) LexString(filename string, s string) (lexer.Lexer, error) { // nolint: golint
return &Lexer{
def: d,
data: s,
stack: []lexerState{{name: d.initialState}},
pos: lexer.Position{
Filename: filename,
Line: 1,
Column: 1,
},
}, nil
}
func (d *Definition) Lex(filename string, r io.Reader) (lexer.Lexer, error) { // nolint: golint
w := &strings.Builder{}
_, err := io.Copy(w, r)
if err != nil {
return nil, err
}
return d.LexString(filename, w.String())
}
func (d *Definition) | () map[string]rune { // nolint: golint
return d.symbols
}
type lexerState struct {
name string
groups []string
}
// Lexer implementation.
type Lexer struct {
stack []lexerState
def *Definition
data string
pos lexer.Position
}
func (l *Lexer) Next() (lexer.Token, error) { // nolint: golint
parent := l.stack[len(l.stack)-1]
rules := l.def.rules[parent.name]
next:
for len(l.data) > 0 {
var (
rule *compiledRule
match []int
)
for _, candidate := range rules {
// Special case "Return()".
if candidate.Rule == ReturnRule {
l.stack = l.stack[:len(l.stack)-1]
parent = l.stack[len(l.stack)-1]
rules = l.def.rules[parent.name]
continue next
}
re, err := l.getPattern(candidate)
if err != nil {
return lexer.Token{}, participle.Wrapf(l.pos, err, "rule %q", candidate.Name)
}
match = re.FindStringSubmatchIndex(l.data)
if match != nil {
rule = &candidate // nolint
break
}
}
if match == nil || rule == nil {
sample := []rune(l.data)
if len(sample) > 16 {
sample = append(sample[:16], []rune("...")...)
}
return lexer.Token{}, participle.Errorf(l.pos, "invalid input text %q", string(sample))
| Symbols | identifier_name |
stateful.go | "
"github.com/alecthomas/participle/v2"
"github.com/alecthomas/participle/v2/lexer"
)
var (
backrefReplace = regexp.MustCompile(`(\\+)(\d)`)
)
// Option for modifying how the Lexer works.
type Option func(d *Definition)
// InitialState overrides the default initial state of "Root".
func InitialState(state string) Option {
return func(d *Definition) {
d.initialState = state
}
}
// A Rule matching input and possibly changing state.
type Rule struct {
Name string
Pattern string
Action Action
}
// Rules grouped by name.
type Rules map[string][]Rule
// compiledRule is a Rule with its pattern compiled.
type compiledRule struct {
Rule
ignore bool
RE *regexp.Regexp
}
// compiledRules grouped by name.
type compiledRules map[string][]compiledRule
// A Action is applied when a rule matches.
type Action interface {
// Actions are responsible for validating the match. ie. if they consumed any input.
applyAction(lexer *Lexer, groups []string) error
}
// RulesAction is an optional interface that Actions can implement.
//
// It is applied during rule construction to mutate the rule map.
type RulesAction interface {
applyRules(state string, rule int, rules compiledRules) error
}
// ActionPop pops to the previous state when the Rule matches.
type ActionPop struct{}
func (p ActionPop) applyAction(lexer *Lexer, groups []string) error {
if groups[0] == "" {
return errors.New("did not consume any input")
}
lexer.stack = lexer.stack[:len(lexer.stack)-1]
return nil
}
// Pop to the previous state.
func Pop() Action {
return ActionPop{}
}
// ReturnRule signals the lexer to return immediately.
var ReturnRule = Rule{"returnToParent", "", nil}
// Return to the parent state.
//
// Useful as the last rule in a sub-state.
func Return() Rule { return ReturnRule }
// ActionPush pushes the current state and switches to "State" when the Rule matches.
type ActionPush struct{ State string }
func (p ActionPush) applyAction(lexer *Lexer, groups []string) error {
if groups[0] == "" |
lexer.stack = append(lexer.stack, lexerState{name: p.State, groups: groups})
return nil
}
// Push to the given state.
//
// The target state will then be the set of rules used for matching
// until another Push or Pop is encountered.
func Push(state string) Action {
return ActionPush{state}
}
type include struct{ state string }
func (i include) applyAction(lexer *Lexer, groups []string) error { panic("should not be called") }
func (i include) applyRules(state string, rule int, rules compiledRules) error {
includedRules, ok := rules[i.state]
if !ok {
return fmt.Errorf("invalid include state %q", i.state)
}
clone := make([]compiledRule, len(includedRules))
copy(clone, includedRules)
rules[state] = append(rules[state][:rule], append(clone, rules[state][rule+1:]...)...) // nolint: makezero
return nil
}
// Include rules from another state in this one.
func Include(state string) Rule {
return Rule{Action: include{state}}
}
// Definition is the lexer.Definition.
type Definition struct {
rules compiledRules
symbols map[string]rune
// Map of key->*regexp.Regexp
backrefCache sync.Map
initialState string
}
// MustSimple creates a new lexer definition based on a single state described by `rules`.
// panics if the rules trigger an error
func MustSimple(rules []Rule, options ...Option) *Definition {
def, err := NewSimple(rules, options...)
if err != nil {
panic(err)
}
return def
}
// Must creates a new stateful lexer and panics if it is incorrect.
func Must(rules Rules, options ...Option) *Definition {
def, err := New(rules, options...)
if err != nil {
panic(err)
}
return def
}
// NewSimple creates a new stateful lexer with a single "Root" state.
func NewSimple(rules []Rule, options ...Option) (*Definition, error) {
return New(Rules{"Root": rules}, options...)
}
// New constructs a new stateful lexer from rules.
func New(rules Rules, options ...Option) (*Definition, error) {
compiled := compiledRules{}
for key, set := range rules {
for i, rule := range set {
pattern := "^(?:" + rule.Pattern + ")"
var (
re *regexp.Regexp
err error
)
var match = backrefReplace.FindStringSubmatch(rule.Pattern)
if match == nil || len(match[1])%2 == 0 {
re, err = regexp.Compile(pattern)
if err != nil {
return nil, fmt.Errorf("%s.%d: %s", key, i, err)
}
}
compiled[key] = append(compiled[key], compiledRule{
Rule: rule,
ignore: len(rule.Name) > 0 && unicode.IsLower(rune(rule.Name[0])),
RE: re,
})
}
}
restart:
for state, rules := range compiled {
for i, rule := range rules {
if action, ok := rule.Action.(RulesAction); ok {
if err := action.applyRules(state, i, compiled); err != nil {
return nil, fmt.Errorf("%s.%d: %s", state, i, err)
}
goto restart
}
}
}
keys := make([]string, 0, len(compiled))
for key := range compiled {
keys = append(keys, key)
}
symbols := map[string]rune{
"EOF": lexer.EOF,
}
sort.Strings(keys)
duplicates := map[string]compiledRule{}
rn := lexer.EOF - 1
for _, key := range keys {
for i, rule := range compiled[key] {
if dup, ok := duplicates[rule.Name]; ok && rule.Pattern != dup.Pattern {
panic(fmt.Sprintf("duplicate key %q with different patterns %q != %q", rule.Name, rule.Pattern, dup.Pattern))
}
duplicates[rule.Name] = rule
compiled[key][i] = rule
symbols[rule.Name] = rn
rn--
}
}
d := &Definition{
initialState: "Root",
rules: compiled,
symbols: symbols,
}
for _, option := range options {
option(d)
}
return d, nil
}
// Rules returns the user-provided Rules used to construct the lexer.
func (d *Definition) Rules() Rules {
out := Rules{}
for state, rules := range d.rules {
for _, rule := range rules {
out[state] = append(out[state], rule.Rule)
}
}
return out
}
func (d *Definition) LexString(filename string, s string) (lexer.Lexer, error) { // nolint: golint
return &Lexer{
def: d,
data: s,
stack: []lexerState{{name: d.initialState}},
pos: lexer.Position{
Filename: filename,
Line: 1,
Column: 1,
},
}, nil
}
func (d *Definition) Lex(filename string, r io.Reader) (lexer.Lexer, error) { // nolint: golint
w := &strings.Builder{}
_, err := io.Copy(w, r)
if err != nil {
return nil, err
}
return d.LexString(filename, w.String())
}
func (d *Definition) Symbols() map[string]rune { // nolint: golint
return d.symbols
}
type lexerState struct {
name string
groups []string
}
// Lexer implementation.
type Lexer struct {
stack []lexerState
def *Definition
data string
pos lexer.Position
}
func (l *Lexer) Next() (lexer.Token, error) { // nolint: golint
parent := l.stack[len(l.stack)-1]
rules := l.def.rules[parent.name]
next:
for len(l.data) > 0 {
var (
rule *compiledRule
match []int
)
for _, candidate := range rules {
// Special case "Return()".
if candidate.Rule == ReturnRule {
l.stack = l.stack[:len(l.stack)-1]
parent = l.stack[len(l.stack)-1]
rules = l.def.rules[parent.name]
continue next
}
re, err := l.getPattern(candidate)
if err != nil {
return lexer.Token{}, participle.Wrapf(l.pos, err, "rule %q", candidate.Name)
}
match = re.FindStringSubmatchIndex(l.data)
if match != nil {
rule = &candidate // nolint
break
}
}
if match == nil || rule == nil {
sample := []rune(l.data)
if len(sample) > 16 {
sample = append(sample[:16], []rune("...")...)
}
return lexer.Token{}, participle.Errorf(l.pos, "invalid input text %q", string(sample | {
return errors.New("did not consume any input")
} | conditional_block |
gaiatools.py | rometric_n_good_obs_al,astrometric_n_bad_obs_al,"+\
"astrometric_gof_al,astrometric_chi2_al,astrometric_excess_noise,astrometric_excess_noise_sig,"+\
"astrometric_params_solved,astrometric_primary_flag,astrometric_weight_al,"+\
"astrometric_pseudo_colour,astrometric_pseudo_colour_error,"+\
"mean_varpi_factor_al,astrometric_matched_observations,visibility_periods_used,"+\
"astrometric_sigma5d_max,frame_rotator_object_type,matched_observations"
cols_phot = "phot_g_mean_mag,phot_bp_mean_mag,phot_rp_mean_mag,phot_variable_flag"
ecol_phot = "phot_g_mean_flux,phot_bp_mean_flux,phot_rp_mean_flux,"+\
"phot_g_mean_flux_error,phot_g_mean_flux_over_error,"+\
"phot_bp_mean_flux_error,phot_bp_mean_flux_over_error,"+\
"phot_rp_mean_flux_error,phot_rp_mean_flux_over_error"
qual_phot = "phot_g_n_obs,phot_bp_n_obs,phot_rp_n_obs,phot_bp_rp_excess_factor,phot_proc_mode"
cols_redd = "bp_rp,bp_g,g_rp,a_g_val,e_bp_min_rp_val,"+\
"a_g_percentile_lower,a_g_percentile_upper,"+\
"e_bp_min_rp_percentile_lower,e_bp_min_rp_percentile_upper"
cols_spec = "radial_velocity,radial_velocity_error"
qual_spec = "rv_template_teff,rv_template_logg,rv_template_fe_h,rv_nb_transits"
cols_star = "teff_val,radius_val,lum_val"
ecol_star = "teff_percentile_lower,teff_percentile_upper,"+\
"radius_percentile_lower,radius_percentile_upper,"+\
"lum_percentile_lower,lum_percentile_upper"
cols_rave = ""
ecol_rave = ""
all_columns = ",".join(["source_id", cols_astrometry, ecol_astrometry, qual_astrometry,
cols_phot, ecol_phot, qual_phot, cols_redd, cols_spec, qual_spec, cols_star, ecol_star])
## This is a full set of things that I think will be useful
full_columns = ",".join(["source_id", cols_astrometry, ecol_astrometry,
cols_phot, ecol_phot, cols_redd, cols_spec, qual_spec, cols_star, ecol_star])
## This is a minimal set of things that I think will be useful
default_columns = ",".join(["source_id",cols_astrometry,ecol_astrometry,
cols_phot, cols_spec, cols_star])
def create_source_query_from_ids(ids, columns=default_columns,
source="gaiaedr3.gaia_source"):
out = "SELECT {} FROM {} WHERE ".format(
columns, source)
idstrs = " or ".join(["source_id = {}".format(x) for x in ids])
out += idstrs
return out
def create_source_query_from(coords, radius=1*u.arcsec,
columns=default_columns,
source="gaiaedr3.gaia_source",
Nmax=None):
"""
Generate a string selecting specific list of coordinates.
Built from https://gist.github.com/mfouesneau/b6b25ed645eab9da4710153fcf9a4cb8
"""
N = len(coords)
if Nmax is None: Nmax = 2*N
out = "SELECT TOP {} {} FROM {} WHERE ".format(
Nmax, columns, source)
def _make_contains_str(c):
cstr = "CONTAINS(POINT('ICRS',{0:}.ra,{0:}.dec),CIRCLE('ICRS',{1:},{2:},{3:}))=1".format(
source, c.ra.deg, c.dec.deg, radius.to("deg").value)
return cstr
cstrs = map(_make_contains_str, coords)
out += " or ".join(cstrs)
return out
def create_samples(Nsamp,mu,cov):
Nstars,Nparams = mu.shape
assert Nstars == len(cov)
assert Nparams == cov.shape[1]
output = np.zeros((Nsamp*Nstars, Nparams))
for i in range(Nstars):
i1 = Nsamp*i
i2 = Nsamp*(i+1)
output[i1:i2,:] = np.random.multivariate_normal(mu[i,:],cov[i,:,:],Nsamp)
output = output.reshape(Nstars, Nsamp, Nparams)
return output
def get_gc_frame():
v_sun = coord.CartesianDifferential([11.1, 250, 7.25]*u.km/u.s)
#gc_frame = coord.Galactocentric(galcen_distance=8.3*u.kpc,
# z_sun=0*u.pc,
# galcen_v_sun=v_sun)
gc_frame = coord.Galactocentric()
return gc_frame
def get_gccoo_w0(coo):
gc_frame = get_gc_frame()
gccoo = coo.transform_to(gc_frame)
w0 = gd.PhaseSpacePosition(gccoo.data)
return gccoo, w0
def get_orbit_params(orbits):
N = orbits.shape[1]
pers = []
apos = []
eccs = []
for i in range(N):
orbit = orbits[:,i]
rp, ra = orbit.pericenter(), orbit.apocenter()
pers.append(rp)
apos.append(ra)
eccs.append((ra - rp) / (ra + rp))
return u.Quantity(pers), u.Quantity(apos), u.Quantity(eccs)
def get_orbit_params_fast(orbits):
try:
N = orbits.shape[1]
except IndexError:
orbit = orbits
r = np.sqrt(np.sum(orbits.xyz**2,axis=0))
rp, ra = np.min(r), np.max(r)
return u.Quantity(rp), u.Quantity(ra), u.Quantity((ra-rp)/(ra+rp))
pers = []
apos = []
eccs = []
for i in range(N):
orbit = orbits[:,i]
r = np.sqrt(np.sum(orbit.xyz**2,axis=0))
rp, ra = np.min(r), np.max(r)
pers.append(rp)
apos.append(ra)
eccs.append((ra - rp) / (ra + rp))
return u.Quantity(pers), u.Quantity(apos), u.Quantity(eccs)
def calc_vtan_error(pmra, pmdec, parallax):
d = u.kpc / parallax.value
pmra = pmra.to(u.rad/u.yr, u.dimensionless_angles())
pmdec= pmdec.to(u.rad/u.yr, u.dimensionless_angles())
vtan = d * np.sqrt(pmra**2 + pmdec**2)
vtan = vtan.to(u.km/u.s, u.dimensionless_angles())
return vtan
def avgstd(x,ignore_nan=False, axis=None):
mean = np.nanmean if ignore_nan else np.mean
stdev = np.nanstd if ignore_nan else np.std
kws = {}
if axis is not None: kws['axis'] = axis
mu = mean(x,**kws)
sig = stdev(x,**kws)
return np.vstack([mu,sig]).T
def medscat(x,sigma=2,ignore_nan=False, axis=None, for_errorbar_plot=False):
percentile = np.nanpercentile if ignore_nan else np.percentile
pdict = {1:[16,50,84],2:[5,50,95],3:[.1,50,99.9]}
assert sigma in pdict
kws = {}
if axis is not None: kws['axis'] = axis
p1,p2,p3 = percentile(x, pdict[sigma], **kws)
e1 = p1-p2
e2 = p3-p2
if for_errorbar_plot:
e1 = -e1
return p2, np.stack([e1,e2])
return np.stack([e1,p2,e2])
def modefinder(x, bins="auto", dropna=True):
"""
Estimates the mode of a sample of points.
Assumes a unimodal system.
Take a histogram of the data and return the bin with the largest value.
TODO If an initial value is specified, find the local maximum closest to that value.
"""
if dropna: |
h,x = np.histogram(x, bins=bins)
xm = (x[1:]+x[:-1])/2.
ix = np.argmax(h)
return xm[ix]
def get_finite(x,y):
""" Get x and y that are both finite """
finite = np.logical_and(np.isfinite(x), np.isfinite(y))
xf = x[finite]; yf = y[finite]
return xf, yf
def fit_spline(x, y, **kwargs):
""" A simple wrapper to scipy.interpolate.UnivariateSpline (remove nan, sort x) """
xf, yf = get_finite(x,y)
| x = x[np.isfinite(x)] | conditional_block |
gaiatools.py | rometric_n_good_obs_al,astrometric_n_bad_obs_al,"+\
"astrometric_gof_al,astrometric_chi2_al,astrometric_excess_noise,astrometric_excess_noise_sig,"+\
"astrometric_params_solved,astrometric_primary_flag,astrometric_weight_al,"+\
"astrometric_pseudo_colour,astrometric_pseudo_colour_error,"+\
"mean_varpi_factor_al,astrometric_matched_observations,visibility_periods_used,"+\
"astrometric_sigma5d_max,frame_rotator_object_type,matched_observations"
cols_phot = "phot_g_mean_mag,phot_bp_mean_mag,phot_rp_mean_mag,phot_variable_flag"
ecol_phot = "phot_g_mean_flux,phot_bp_mean_flux,phot_rp_mean_flux,"+\
"phot_g_mean_flux_error,phot_g_mean_flux_over_error,"+\
"phot_bp_mean_flux_error,phot_bp_mean_flux_over_error,"+\
"phot_rp_mean_flux_error,phot_rp_mean_flux_over_error"
qual_phot = "phot_g_n_obs,phot_bp_n_obs,phot_rp_n_obs,phot_bp_rp_excess_factor,phot_proc_mode"
cols_redd = "bp_rp,bp_g,g_rp,a_g_val,e_bp_min_rp_val,"+\
"a_g_percentile_lower,a_g_percentile_upper,"+\
"e_bp_min_rp_percentile_lower,e_bp_min_rp_percentile_upper"
cols_spec = "radial_velocity,radial_velocity_error"
qual_spec = "rv_template_teff,rv_template_logg,rv_template_fe_h,rv_nb_transits"
cols_star = "teff_val,radius_val,lum_val"
ecol_star = "teff_percentile_lower,teff_percentile_upper,"+\
"radius_percentile_lower,radius_percentile_upper,"+\
"lum_percentile_lower,lum_percentile_upper"
cols_rave = ""
ecol_rave = ""
all_columns = ",".join(["source_id", cols_astrometry, ecol_astrometry, qual_astrometry,
cols_phot, ecol_phot, qual_phot, cols_redd, cols_spec, qual_spec, cols_star, ecol_star])
## This is a full set of things that I think will be useful
full_columns = ",".join(["source_id", cols_astrometry, ecol_astrometry,
cols_phot, ecol_phot, cols_redd, cols_spec, qual_spec, cols_star, ecol_star])
## This is a minimal set of things that I think will be useful
default_columns = ",".join(["source_id",cols_astrometry,ecol_astrometry,
cols_phot, cols_spec, cols_star])
def create_source_query_from_ids(ids, columns=default_columns,
source="gaiaedr3.gaia_source"):
out = "SELECT {} FROM {} WHERE ".format(
columns, source)
idstrs = " or ".join(["source_id = {}".format(x) for x in ids])
out += idstrs
return out
def create_source_query_from(coords, radius=1*u.arcsec,
columns=default_columns,
source="gaiaedr3.gaia_source",
Nmax=None):
"""
Generate a string selecting specific list of coordinates.
Built from https://gist.github.com/mfouesneau/b6b25ed645eab9da4710153fcf9a4cb8
"""
N = len(coords)
if Nmax is None: Nmax = 2*N
out = "SELECT TOP {} {} FROM {} WHERE ".format(
Nmax, columns, source)
def | (c):
cstr = "CONTAINS(POINT('ICRS',{0:}.ra,{0:}.dec),CIRCLE('ICRS',{1:},{2:},{3:}))=1".format(
source, c.ra.deg, c.dec.deg, radius.to("deg").value)
return cstr
cstrs = map(_make_contains_str, coords)
out += " or ".join(cstrs)
return out
def create_samples(Nsamp,mu,cov):
Nstars,Nparams = mu.shape
assert Nstars == len(cov)
assert Nparams == cov.shape[1]
output = np.zeros((Nsamp*Nstars, Nparams))
for i in range(Nstars):
i1 = Nsamp*i
i2 = Nsamp*(i+1)
output[i1:i2,:] = np.random.multivariate_normal(mu[i,:],cov[i,:,:],Nsamp)
output = output.reshape(Nstars, Nsamp, Nparams)
return output
def get_gc_frame():
v_sun = coord.CartesianDifferential([11.1, 250, 7.25]*u.km/u.s)
#gc_frame = coord.Galactocentric(galcen_distance=8.3*u.kpc,
# z_sun=0*u.pc,
# galcen_v_sun=v_sun)
gc_frame = coord.Galactocentric()
return gc_frame
def get_gccoo_w0(coo):
gc_frame = get_gc_frame()
gccoo = coo.transform_to(gc_frame)
w0 = gd.PhaseSpacePosition(gccoo.data)
return gccoo, w0
def get_orbit_params(orbits):
N = orbits.shape[1]
pers = []
apos = []
eccs = []
for i in range(N):
orbit = orbits[:,i]
rp, ra = orbit.pericenter(), orbit.apocenter()
pers.append(rp)
apos.append(ra)
eccs.append((ra - rp) / (ra + rp))
return u.Quantity(pers), u.Quantity(apos), u.Quantity(eccs)
def get_orbit_params_fast(orbits):
try:
N = orbits.shape[1]
except IndexError:
orbit = orbits
r = np.sqrt(np.sum(orbits.xyz**2,axis=0))
rp, ra = np.min(r), np.max(r)
return u.Quantity(rp), u.Quantity(ra), u.Quantity((ra-rp)/(ra+rp))
pers = []
apos = []
eccs = []
for i in range(N):
orbit = orbits[:,i]
r = np.sqrt(np.sum(orbit.xyz**2,axis=0))
rp, ra = np.min(r), np.max(r)
pers.append(rp)
apos.append(ra)
eccs.append((ra - rp) / (ra + rp))
return u.Quantity(pers), u.Quantity(apos), u.Quantity(eccs)
def calc_vtan_error(pmra, pmdec, parallax):
d = u.kpc / parallax.value
pmra = pmra.to(u.rad/u.yr, u.dimensionless_angles())
pmdec= pmdec.to(u.rad/u.yr, u.dimensionless_angles())
vtan = d * np.sqrt(pmra**2 + pmdec**2)
vtan = vtan.to(u.km/u.s, u.dimensionless_angles())
return vtan
def avgstd(x,ignore_nan=False, axis=None):
mean = np.nanmean if ignore_nan else np.mean
stdev = np.nanstd if ignore_nan else np.std
kws = {}
if axis is not None: kws['axis'] = axis
mu = mean(x,**kws)
sig = stdev(x,**kws)
return np.vstack([mu,sig]).T
def medscat(x,sigma=2,ignore_nan=False, axis=None, for_errorbar_plot=False):
percentile = np.nanpercentile if ignore_nan else np.percentile
pdict = {1:[16,50,84],2:[5,50,95],3:[.1,50,99.9]}
assert sigma in pdict
kws = {}
if axis is not None: kws['axis'] = axis
p1,p2,p3 = percentile(x, pdict[sigma], **kws)
e1 = p1-p2
e2 = p3-p2
if for_errorbar_plot:
e1 = -e1
return p2, np.stack([e1,e2])
return np.stack([e1,p2,e2])
def modefinder(x, bins="auto", dropna=True):
"""
Estimates the mode of a sample of points.
Assumes a unimodal system.
Take a histogram of the data and return the bin with the largest value.
TODO If an initial value is specified, find the local maximum closest to that value.
"""
if dropna: x = x[np.isfinite(x)]
h,x = np.histogram(x, bins=bins)
xm = (x[1:]+x[:-1])/2.
ix = np.argmax(h)
return xm[ix]
def get_finite(x,y):
""" Get x and y that are both finite """
finite = np.logical_and(np.isfinite(x), np.isfinite(y))
xf = x[finite]; yf = y[finite]
return xf, yf
def fit_spline(x, y, **kwargs):
""" A simple wrapper to scipy.interpolate.UnivariateSpline (remove nan, sort x) """
xf, yf = get_finite(x,y)
| _make_contains_str | identifier_name |
gaiatools.py | rometric_n_good_obs_al,astrometric_n_bad_obs_al,"+\
"astrometric_gof_al,astrometric_chi2_al,astrometric_excess_noise,astrometric_excess_noise_sig,"+\
"astrometric_params_solved,astrometric_primary_flag,astrometric_weight_al,"+\
"astrometric_pseudo_colour,astrometric_pseudo_colour_error,"+\
"mean_varpi_factor_al,astrometric_matched_observations,visibility_periods_used,"+\
"astrometric_sigma5d_max,frame_rotator_object_type,matched_observations"
cols_phot = "phot_g_mean_mag,phot_bp_mean_mag,phot_rp_mean_mag,phot_variable_flag"
ecol_phot = "phot_g_mean_flux,phot_bp_mean_flux,phot_rp_mean_flux,"+\
"phot_g_mean_flux_error,phot_g_mean_flux_over_error,"+\
"phot_bp_mean_flux_error,phot_bp_mean_flux_over_error,"+\
"phot_rp_mean_flux_error,phot_rp_mean_flux_over_error"
qual_phot = "phot_g_n_obs,phot_bp_n_obs,phot_rp_n_obs,phot_bp_rp_excess_factor,phot_proc_mode"
cols_redd = "bp_rp,bp_g,g_rp,a_g_val,e_bp_min_rp_val,"+\
"a_g_percentile_lower,a_g_percentile_upper,"+\
"e_bp_min_rp_percentile_lower,e_bp_min_rp_percentile_upper"
cols_spec = "radial_velocity,radial_velocity_error"
qual_spec = "rv_template_teff,rv_template_logg,rv_template_fe_h,rv_nb_transits"
cols_star = "teff_val,radius_val,lum_val"
ecol_star = "teff_percentile_lower,teff_percentile_upper,"+\
"radius_percentile_lower,radius_percentile_upper,"+\
"lum_percentile_lower,lum_percentile_upper"
cols_rave = ""
ecol_rave = ""
all_columns = ",".join(["source_id", cols_astrometry, ecol_astrometry, qual_astrometry,
cols_phot, ecol_phot, qual_phot, cols_redd, cols_spec, qual_spec, cols_star, ecol_star])
## This is a full set of things that I think will be useful
full_columns = ",".join(["source_id", cols_astrometry, ecol_astrometry,
cols_phot, ecol_phot, cols_redd, cols_spec, qual_spec, cols_star, ecol_star])
## This is a minimal set of things that I think will be useful
default_columns = ",".join(["source_id",cols_astrometry,ecol_astrometry,
cols_phot, cols_spec, cols_star])
def create_source_query_from_ids(ids, columns=default_columns,
source="gaiaedr3.gaia_source"):
out = "SELECT {} FROM {} WHERE ".format(
columns, source)
idstrs = " or ".join(["source_id = {}".format(x) for x in ids])
out += idstrs
return out
def create_source_query_from(coords, radius=1*u.arcsec,
columns=default_columns,
source="gaiaedr3.gaia_source",
Nmax=None):
"""
Generate a string selecting specific list of coordinates.
Built from https://gist.github.com/mfouesneau/b6b25ed645eab9da4710153fcf9a4cb8
"""
N = len(coords)
if Nmax is None: Nmax = 2*N
out = "SELECT TOP {} {} FROM {} WHERE ".format(
Nmax, columns, source)
def _make_contains_str(c):
cstr = "CONTAINS(POINT('ICRS',{0:}.ra,{0:}.dec),CIRCLE('ICRS',{1:},{2:},{3:}))=1".format(
source, c.ra.deg, c.dec.deg, radius.to("deg").value)
return cstr
cstrs = map(_make_contains_str, coords)
out += " or ".join(cstrs)
return out
def create_samples(Nsamp,mu,cov):
Nstars,Nparams = mu.shape
assert Nstars == len(cov)
assert Nparams == cov.shape[1]
output = np.zeros((Nsamp*Nstars, Nparams))
for i in range(Nstars):
i1 = Nsamp*i
i2 = Nsamp*(i+1)
output[i1:i2,:] = np.random.multivariate_normal(mu[i,:],cov[i,:,:],Nsamp)
output = output.reshape(Nstars, Nsamp, Nparams)
return output
def get_gc_frame():
v_sun = coord.CartesianDifferential([11.1, 250, 7.25]*u.km/u.s)
#gc_frame = coord.Galactocentric(galcen_distance=8.3*u.kpc,
# z_sun=0*u.pc,
# galcen_v_sun=v_sun)
gc_frame = coord.Galactocentric()
return gc_frame
def get_gccoo_w0(coo):
gc_frame = get_gc_frame()
gccoo = coo.transform_to(gc_frame)
w0 = gd.PhaseSpacePosition(gccoo.data)
return gccoo, w0
def get_orbit_params(orbits):
N = orbits.shape[1]
pers = []
apos = []
eccs = []
for i in range(N):
orbit = orbits[:,i]
rp, ra = orbit.pericenter(), orbit.apocenter()
pers.append(rp)
apos.append(ra)
eccs.append((ra - rp) / (ra + rp))
return u.Quantity(pers), u.Quantity(apos), u.Quantity(eccs)
def get_orbit_params_fast(orbits):
try:
N = orbits.shape[1]
except IndexError:
orbit = orbits
r = np.sqrt(np.sum(orbits.xyz**2,axis=0))
rp, ra = np.min(r), np.max(r)
return u.Quantity(rp), u.Quantity(ra), u.Quantity((ra-rp)/(ra+rp))
pers = []
apos = []
eccs = []
for i in range(N):
orbit = orbits[:,i]
r = np.sqrt(np.sum(orbit.xyz**2,axis=0))
rp, ra = np.min(r), np.max(r)
pers.append(rp)
apos.append(ra)
eccs.append((ra - rp) / (ra + rp))
return u.Quantity(pers), u.Quantity(apos), u.Quantity(eccs)
def calc_vtan_error(pmra, pmdec, parallax):
d = u.kpc / parallax.value
pmra = pmra.to(u.rad/u.yr, u.dimensionless_angles())
pmdec= pmdec.to(u.rad/u.yr, u.dimensionless_angles())
vtan = d * np.sqrt(pmra**2 + pmdec**2)
vtan = vtan.to(u.km/u.s, u.dimensionless_angles())
return vtan
| stdev = np.nanstd if ignore_nan else np.std
kws = {}
if axis is not None: kws['axis'] = axis
mu = mean(x,**kws)
sig = stdev(x,**kws)
return np.vstack([mu,sig]).T
def medscat(x,sigma=2,ignore_nan=False, axis=None, for_errorbar_plot=False):
percentile = np.nanpercentile if ignore_nan else np.percentile
pdict = {1:[16,50,84],2:[5,50,95],3:[.1,50,99.9]}
assert sigma in pdict
kws = {}
if axis is not None: kws['axis'] = axis
p1,p2,p3 = percentile(x, pdict[sigma], **kws)
e1 = p1-p2
e2 = p3-p2
if for_errorbar_plot:
e1 = -e1
return p2, np.stack([e1,e2])
return np.stack([e1,p2,e2])
def modefinder(x, bins="auto", dropna=True):
"""
Estimates the mode of a sample of points.
Assumes a unimodal system.
Take a histogram of the data and return the bin with the largest value.
TODO If an initial value is specified, find the local maximum closest to that value.
"""
if dropna: x = x[np.isfinite(x)]
h,x = np.histogram(x, bins=bins)
xm = (x[1:]+x[:-1])/2.
ix = np.argmax(h)
return xm[ix]
def get_finite(x,y):
""" Get x and y that are both finite """
finite = np.logical_and(np.isfinite(x), np.isfinite(y))
xf = x[finite]; yf = y[finite]
return xf, yf
def fit_spline(x, y, **kwargs):
""" A simple wrapper to scipy.interpolate.UnivariateSpline (remove nan, sort x) """
xf, yf = get_finite(x,y)
| def avgstd(x,ignore_nan=False, axis=None):
mean = np.nanmean if ignore_nan else np.mean | random_line_split |
gaiatools.py | ICRS',{1:},{2:},{3:}))=1".format(
source, c.ra.deg, c.dec.deg, radius.to("deg").value)
return cstr
cstrs = map(_make_contains_str, coords)
out += " or ".join(cstrs)
return out
def create_samples(Nsamp,mu,cov):
Nstars,Nparams = mu.shape
assert Nstars == len(cov)
assert Nparams == cov.shape[1]
output = np.zeros((Nsamp*Nstars, Nparams))
for i in range(Nstars):
i1 = Nsamp*i
i2 = Nsamp*(i+1)
output[i1:i2,:] = np.random.multivariate_normal(mu[i,:],cov[i,:,:],Nsamp)
output = output.reshape(Nstars, Nsamp, Nparams)
return output
def get_gc_frame():
v_sun = coord.CartesianDifferential([11.1, 250, 7.25]*u.km/u.s)
#gc_frame = coord.Galactocentric(galcen_distance=8.3*u.kpc,
# z_sun=0*u.pc,
# galcen_v_sun=v_sun)
gc_frame = coord.Galactocentric()
return gc_frame
def get_gccoo_w0(coo):
gc_frame = get_gc_frame()
gccoo = coo.transform_to(gc_frame)
w0 = gd.PhaseSpacePosition(gccoo.data)
return gccoo, w0
def get_orbit_params(orbits):
N = orbits.shape[1]
pers = []
apos = []
eccs = []
for i in range(N):
orbit = orbits[:,i]
rp, ra = orbit.pericenter(), orbit.apocenter()
pers.append(rp)
apos.append(ra)
eccs.append((ra - rp) / (ra + rp))
return u.Quantity(pers), u.Quantity(apos), u.Quantity(eccs)
def get_orbit_params_fast(orbits):
try:
N = orbits.shape[1]
except IndexError:
orbit = orbits
r = np.sqrt(np.sum(orbits.xyz**2,axis=0))
rp, ra = np.min(r), np.max(r)
return u.Quantity(rp), u.Quantity(ra), u.Quantity((ra-rp)/(ra+rp))
pers = []
apos = []
eccs = []
for i in range(N):
orbit = orbits[:,i]
r = np.sqrt(np.sum(orbit.xyz**2,axis=0))
rp, ra = np.min(r), np.max(r)
pers.append(rp)
apos.append(ra)
eccs.append((ra - rp) / (ra + rp))
return u.Quantity(pers), u.Quantity(apos), u.Quantity(eccs)
def calc_vtan_error(pmra, pmdec, parallax):
d = u.kpc / parallax.value
pmra = pmra.to(u.rad/u.yr, u.dimensionless_angles())
pmdec= pmdec.to(u.rad/u.yr, u.dimensionless_angles())
vtan = d * np.sqrt(pmra**2 + pmdec**2)
vtan = vtan.to(u.km/u.s, u.dimensionless_angles())
return vtan
def avgstd(x,ignore_nan=False, axis=None):
mean = np.nanmean if ignore_nan else np.mean
stdev = np.nanstd if ignore_nan else np.std
kws = {}
if axis is not None: kws['axis'] = axis
mu = mean(x,**kws)
sig = stdev(x,**kws)
return np.vstack([mu,sig]).T
def medscat(x,sigma=2,ignore_nan=False, axis=None, for_errorbar_plot=False):
percentile = np.nanpercentile if ignore_nan else np.percentile
pdict = {1:[16,50,84],2:[5,50,95],3:[.1,50,99.9]}
assert sigma in pdict
kws = {}
if axis is not None: kws['axis'] = axis
p1,p2,p3 = percentile(x, pdict[sigma], **kws)
e1 = p1-p2
e2 = p3-p2
if for_errorbar_plot:
e1 = -e1
return p2, np.stack([e1,e2])
return np.stack([e1,p2,e2])
def modefinder(x, bins="auto", dropna=True):
"""
Estimates the mode of a sample of points.
Assumes a unimodal system.
Take a histogram of the data and return the bin with the largest value.
TODO If an initial value is specified, find the local maximum closest to that value.
"""
if dropna: x = x[np.isfinite(x)]
h,x = np.histogram(x, bins=bins)
xm = (x[1:]+x[:-1])/2.
ix = np.argmax(h)
return xm[ix]
def get_finite(x,y):
""" Get x and y that are both finite """
finite = np.logical_and(np.isfinite(x), np.isfinite(y))
xf = x[finite]; yf = y[finite]
return xf, yf
def fit_spline(x, y, **kwargs):
""" A simple wrapper to scipy.interpolate.UnivariateSpline (remove nan, sort x) """
xf, yf = get_finite(x,y)
iisort = np.argsort(xf)
return interpolate.UnivariateSpline(xf[iisort],yf[iisort], **kwargs)
def bin_medscat(x, y, percentiles=[5,50,95], for_errorbar_plot=False, dropna=True, bins="auto", **kwargs):
"""
Histogram x into bins.
Then in those bins, take percentiles of y.
"""
if dropna: x, y = get_finite(x, y)
h, xe = np.histogram(x, bins=bins, **kwargs)
xout = (xe[1:]+xe[:-1])/2.
indices = np.digitize(x, xe)
yout = np.zeros((len(xe)-1,len(percentiles)))+np.nan
for ix in np.unique(indices):
# Skip things outside the bin range
if ix >= len(yout): continue
# Percentile in this bin
ii = ix==indices
yout[ix,:] = np.percentile(y[ii], percentiles)
if for_errorbar_plot:
e1 = yout[:,1] - yout[:,0]
e2 = yout[:,2] - yout[:,1]
return xout, yout[:,1], [e1,e2]
return xout, yout
def calculate_actions(w0,pot=gp.MilkyWayPotential(), dt=0.5, n_steps=10000, full_output=False):
""" Approximate actions following https://github.com/adrn/gala/blob/master/docs/dynamics/actionangle.rst """
assert len(w0.shape)==0
w = gp.Hamiltonian(pot).integrate_orbit(w0, dt=dt, n_steps=n_steps)
toy_potential = gd.fit_isochrone(w)
toy_actions, toy_angles, toy_freqs = toy_potential.action_angle(w)
with warnings.catch_warnings(record=True):
warnings.simplefilter("ignore")
result = gd.find_actions(w, N_max=8, toy_potential=toy_potential)
if full_output: return result, w
return result["actions"]
def query_and_match(coo, match_radius=1, columns=full_columns):
"""
Query gaia given coordinates
Return a table that is sorted, and an array saying which rows actually matched an object in gaia
"""
from pyia import GaiaDataNew
query = create_source_query_from(coo, columns=columns)
gaia = GaiaDataNew.from_query(query)
gcoo = SkyCoord(gaia.ra, gaia.dec)
idx, d2d, _ = coo.match_to_catalog_sky(gcoo)
iimatch = d2d.arcsec < match_radius
gtab = gaia.data[idx]
if iimatch.sum() != len(gtab):
print("Warning: only matched {}/{} stars".format(iimatch.sum(),len(gtab)))
return gtab, iimatch
def query_and_match_sourceid(source_ids, match_radius=1, columns=full_columns):
| """
Query gaia given source_ids
Return a table in the order of the source_ids
"""
from pyia import GaiaDataNew
unique_arr, indexes = np.unique(source_ids, return_inverse=True)
assert len(unique_arr) == len(source_ids), "Not all IDs are unique"
query = create_source_query_from_ids(source_ids, columns=columns)
gaia = GaiaDataNew.from_query(query)
# Sort by source id, find indices, then resort
gdat = gaia.data
gdat.sort("source_id")
assert np.all(unique_arr == gdat["source_id"])
gdat = gdat[indexes]
assert np.all(gdat["source_id"]==source_ids)
return gdat | identifier_body |
|
langviews.js | = $(this.config.platformSelector).val();
if (endDate.diff(startDate, 'days') === 0) {
startDate.subtract(3, 'days');
endDate.add(3, 'days');
}
return `/pageviews#start=${startDate.format('YYYY-MM-DD')}` +
`&end=${endDate.format('YYYY-MM-DD')}&project=${lang}.${project}.org&platform=${platform}&pages=${article}`;
}
/**
* Loop through given interwiki data and query the pageviews API for each
* Also updates this.langData with result
* @param {Object} interWikiData - as given by the getInterwikiData promise
* @return {Deferred} - Promise resolving with data ready to be rendered to view
*/
getPageViewsData(interWikiData) {
const startDate = this.daterangepicker.startDate.startOf('day'),
endDate = this.daterangepicker.endDate.startOf('day'),
interWikiKeys = Object.keys(interWikiData);
// XXX: throttling
let dfd = $.Deferred(), promises = [], count = 0, hadFailure, failureRetries = {},
totalRequestCount = interWikiKeys.length, failedPages = [];
/** clear out existing data */
this.langData = [];
const makeRequest = dbName => {
const data = interWikiData[dbName],
uriEncodedPageName = encodeURIComponent(data.title);
const url = (
`https://wikimedia.org/api/rest_v1/metrics/pageviews/per-article/${data.lang}.${this.baseProject}` +
`/${$(this.config.platformSelector).val()}/${$(this.config.agentSelector).val()}/${uriEncodedPageName}/daily` +
`/${startDate.format(this.config.timestampFormat)}/${endDate.format(this.config.timestampFormat)}`
);
const promise = $.ajax({ url, dataType: 'json' });
promises.push(promise);
promise.done(pvData => {
const viewCounts = pvData.items.map(el => el.views),
views = viewCounts.reduce((a, b) => a + b);
this.langData.push({
badges: data.badges,
dbName,
lang: data.lang,
pageName: data.title,
views,
url: data.url,
average: Math.round(views / this.numDaysInRange())
});
/** keep track of unique badges/titles and total pageviews */
this.totals.views += views;
if (!this.totals.titles.includes(data.title)) {
this.totals.titles.push(data.title);
}
data.badges.forEach(badge => {
if (this.totals.badges[badge] === undefined) {
this.totals.badges[badge] = 1;
} else {
this.totals.badges[badge] += 1;
}
});
}).fail(errorData => {
// XXX: throttling
/** first detect if this was a Cassandra backend error, and if so, schedule a re-try */
const cassandraError = errorData.responseJSON.title === 'Error in Cassandra table storage backend',
failedPageLink = this.getPageLink(data.title, `${data.lang}.${this.baseProject}.org`);
if (cassandraError) {
if (failureRetries[dbName]) {
failureRetries[dbName]++;
} else {
failureRetries[dbName] = 1;
}
/** maximum of 3 retries */
if (failureRetries[dbName] < 3) {
totalRequestCount++;
return this.rateLimit(makeRequest, 100, this)(dbName);
}
/** retries exceeded */
failedPages.push(failedPageLink);
} else {
this.writeMessage(
`${failedPageLink}: ${$.i18n('api-error', 'Pageviews API')} - ${errorData.responseJSON.title}`
);
}
hadFailure = true; // don't treat this series of requests as being cached by server
}).always(() => {
this.updateProgressBar((++count / totalRequestCount) * 100);
// XXX: throttling, totalRequestCount can just be interWikiKeys.length
if (count === totalRequestCount) {
dfd.resolve(this.langData);
if (failedPages.length) {
this.writeMessage($.i18n(
'api-error-timeout',
'<ul>' +
failedPages.map(failedPage => `<li>${failedPage}</li>`).join('') +
'</ul>'
));
}
/**
* if there were no failures, assume the resource is now cached by the server
* and save this assumption to our own cache so we don't throttle the same requests
*/
// XXX: throttling
if (!hadFailure) {
simpleStorage.set(this.getCacheKey(), true, {TTL: 600000});
}
}
});
};
/**
* We don't want to throttle requests for cached resources. However in our case,
* we're unable to check response headers to see if the resource was cached,
* so we use simpleStorage to keep track of what the user has recently queried.
*/
// XXX: throttling
const requestFn = this.isRequestCached() ? makeRequest : this.rateLimit(makeRequest, 100, this);
interWikiKeys.forEach((dbName, index) => {
requestFn(dbName);
});
return dfd;
}
/**
* Return cache key for current params
* @return {String} key
*/
getCacheKey() {
return `lv-cache-${this.hashCode(
JSON.stringify(this.getParams(true))
)}`;
}
/**
* Check simple storage to see if a request with the current params would be cached
* @return {Boolean} cached or not
*/
isRequestCached() {
return simpleStorage.hasKey(this.getCacheKey());
}
/**
* Query Wikidata to find data about a given page across all sister projects
* @param {String} dbName - database name of source project
* @param {String} pageName - name of page we want to get data about
* @return {Deferred} - Promise resolving with interwiki data
*/
getInterwikiData(dbName, pageName) {
const dfd = $.Deferred();
const url = `https://www.wikidata.org/w/api.php?action=wbgetentities&sites=${dbName}` +
`&titles=${encodeURIComponent(pageName)}&props=sitelinks/urls|datatype&format=json&callback=?`;
$.getJSON(url).done(data => {
if (data.error) {
return dfd.reject(`${$.i18n('api-error', 'Wikidata')}: ${data.error.info}`);
} else if (data.entities['-1']) {
return dfd.reject(
`<a href='${this.getPageURL(pageName)}'>${pageName.descore()}</a> - ${$.i18n('api-error-no-data')}`
);
}
const key = Object.keys(data.entities)[0],
sitelinks = data.entities[key].sitelinks,
filteredLinks = {},
matchRegex = new RegExp(`^https://[\\w-]+\\.${this.baseProject}\\.org`);
/** restrict to selected base project (e.g. wikipedias, not wikipedias and wikivoyages) */
Object.keys(sitelinks).forEach(key => {
const siteMapKey = sitelinks[key].site.replace(/-/g, '_');
if (matchRegex.test(sitelinks[key].url) && siteMap[siteMapKey]) {
sitelinks[key].lang = siteMap[siteMapKey].replace(/\.wiki.*$/, '');
filteredLinks[key] = sitelinks[key];
}
});
return dfd.resolve(filteredLinks);
});
return dfd;
}
/**
* Parse wiki URL for the page name
* @param {String} url - full URL to a wiki page
* @return {String|null} page name
*/
getPageNameFromURL(url) {
if (url.includes('?')) {
return url.match(/\?(?:.*\b)?title=(.*?)(?:&|$)/)[1];
} else {
return url.match(/\/wiki\/(.*?)(?:\?|$)/)[1];
}
}
/**
* Parses the URL query string and sets all the inputs accordingly
* Should only be called on initial page load, until we decide to support pop states (probably never)
* @returns {null} nothing
*/
popParams() | {
let startDate, endDate, params = this.parseQueryString('pages');
$(this.config.projectInput).val(params.project || this.config.defaults.project);
if (this.validateProject()) return;
// FIXME: only run this when they actually submit
this.patchUsage('lv');
/**
* Check if we're using a valid range, and if so ignore any start/end dates.
* If an invalid range, throw and error and use default dates.
*/
if (params.range) {
if (!this.setSpecialRange(params.range)) {
this.addSiteNotice('danger', $.i18n('param-error-3'), $.i18n('invalid-params'), true);
this.setSpecialRange(this.config.defaults.dateRange);
}
} else if (params.start) {
startDate = moment(params.start || moment().subtract(this.config.defaults.daysAgo, 'days')); | identifier_body |
|
langviews.js | `);
if (cassandraError) {
if (failureRetries[dbName]) {
failureRetries[dbName]++;
} else {
failureRetries[dbName] = 1;
}
/** maximum of 3 retries */
if (failureRetries[dbName] < 3) {
totalRequestCount++;
return this.rateLimit(makeRequest, 100, this)(dbName);
}
/** retries exceeded */
failedPages.push(failedPageLink);
} else {
this.writeMessage(
`${failedPageLink}: ${$.i18n('api-error', 'Pageviews API')} - ${errorData.responseJSON.title}`
);
}
hadFailure = true; // don't treat this series of requests as being cached by server
}).always(() => {
this.updateProgressBar((++count / totalRequestCount) * 100);
// XXX: throttling, totalRequestCount can just be interWikiKeys.length
if (count === totalRequestCount) {
dfd.resolve(this.langData);
if (failedPages.length) {
this.writeMessage($.i18n(
'api-error-timeout',
'<ul>' +
failedPages.map(failedPage => `<li>${failedPage}</li>`).join('') +
'</ul>'
));
}
/**
* if there were no failures, assume the resource is now cached by the server
* and save this assumption to our own cache so we don't throttle the same requests
*/
// XXX: throttling
if (!hadFailure) {
simpleStorage.set(this.getCacheKey(), true, {TTL: 600000});
}
}
});
};
/**
* We don't want to throttle requests for cached resources. However in our case,
* we're unable to check response headers to see if the resource was cached,
* so we use simpleStorage to keep track of what the user has recently queried.
*/
// XXX: throttling
const requestFn = this.isRequestCached() ? makeRequest : this.rateLimit(makeRequest, 100, this);
interWikiKeys.forEach((dbName, index) => {
requestFn(dbName);
});
return dfd;
}
/**
* Return cache key for current params
* @return {String} key
*/
getCacheKey() {
return `lv-cache-${this.hashCode(
JSON.stringify(this.getParams(true))
)}`;
}
/**
* Check simple storage to see if a request with the current params would be cached
* @return {Boolean} cached or not
*/
isRequestCached() {
return simpleStorage.hasKey(this.getCacheKey());
}
/**
* Query Wikidata to find data about a given page across all sister projects
* @param {String} dbName - database name of source project
* @param {String} pageName - name of page we want to get data about
* @return {Deferred} - Promise resolving with interwiki data
*/
getInterwikiData(dbName, pageName) {
const dfd = $.Deferred();
const url = `https://www.wikidata.org/w/api.php?action=wbgetentities&sites=${dbName}` +
`&titles=${encodeURIComponent(pageName)}&props=sitelinks/urls|datatype&format=json&callback=?`;
$.getJSON(url).done(data => {
if (data.error) {
return dfd.reject(`${$.i18n('api-error', 'Wikidata')}: ${data.error.info}`);
} else if (data.entities['-1']) {
return dfd.reject(
`<a href='${this.getPageURL(pageName)}'>${pageName.descore()}</a> - ${$.i18n('api-error-no-data')}`
);
}
const key = Object.keys(data.entities)[0],
sitelinks = data.entities[key].sitelinks,
filteredLinks = {},
matchRegex = new RegExp(`^https://[\\w-]+\\.${this.baseProject}\\.org`);
/** restrict to selected base project (e.g. wikipedias, not wikipedias and wikivoyages) */
Object.keys(sitelinks).forEach(key => {
const siteMapKey = sitelinks[key].site.replace(/-/g, '_');
if (matchRegex.test(sitelinks[key].url) && siteMap[siteMapKey]) {
sitelinks[key].lang = siteMap[siteMapKey].replace(/\.wiki.*$/, '');
filteredLinks[key] = sitelinks[key];
}
});
return dfd.resolve(filteredLinks);
});
return dfd;
}
/**
* Parse wiki URL for the page name
* @param {String} url - full URL to a wiki page
* @return {String|null} page name
*/
getPageNameFromURL(url) {
if (url.includes('?')) {
return url.match(/\?(?:.*\b)?title=(.*?)(?:&|$)/)[1];
} else {
return url.match(/\/wiki\/(.*?)(?:\?|$)/)[1];
}
}
/**
* Parses the URL query string and sets all the inputs accordingly
* Should only be called on initial page load, until we decide to support pop states (probably never)
* @returns {null} nothing
*/
popParams() {
let startDate, endDate, params = this.parseQueryString('pages');
$(this.config.projectInput).val(params.project || this.config.defaults.project);
if (this.validateProject()) return;
// FIXME: only run this when they actually submit
this.patchUsage('lv');
/**
* Check if we're using a valid range, and if so ignore any start/end dates.
* If an invalid range, throw and error and use default dates.
*/
if (params.range) {
if (!this.setSpecialRange(params.range)) {
this.addSiteNotice('danger', $.i18n('param-error-3'), $.i18n('invalid-params'), true);
this.setSpecialRange(this.config.defaults.dateRange);
}
} else if (params.start) {
startDate = moment(params.start || moment().subtract(this.config.defaults.daysAgo, 'days'));
endDate = moment(params.end || Date.now());
if (startDate < this.config.minDate || endDate < this.config.minDate) {
this.addSiteNotice('danger', $.i18n('param-error-1', `${$.i18n('july')} 2015`), $.i18n('invalid-params'), true);
return;
} else if (startDate > endDate) {
this.addSiteNotice('warning', $.i18n('param-error-2'), $.i18n('invalid-params'), true);
return;
}
this.daterangepicker.setStartDate(startDate);
this.daterangepicker.setEndDate(endDate);
} else {
this.setSpecialRange(this.config.defaults.dateRange);
}
$(this.config.platformSelector).val(params.platform || 'all-access');
$(this.config.agentSelector).val(params.agent || 'user');
this.sort = params.sort || this.config.defaults.params.sort;
this.direction = params.direction || this.config.defaults.params.direction;
/** start up processing if page name is present */
if (params.page) {
$(this.config.articleInput).val(decodeURIComponent(params.page).descore());
this.processArticle();
}
}
getState() {
const classList = $('main')[0].classList;
return this.config.formStates.filter(stateName => {
return classList.contains(stateName);
})[0];
}
/**
* Helper to set a CSS class on the `main` node,
* styling the document based on a 'state'
* @param {String} state - class to be added;
* should be one of ['initial', 'processing', 'complete']
* @returns {null} nothing
*/
setState(state) {
$('main').removeClass(this.config.formStates.join(' ')).addClass(state);
switch (state) {
case 'initial':
this.clearMessages();
this.assignDefaults();
if (this.typeahead) this.typeahead.hide();
$(this.config.articleInput).val('').focus();
break;
case 'processing':
this.processStarted();
this.clearMessages();
document.activeElement.blur();
$('.progress-bar').addClass('active');
break;
case 'complete':
this.processEnded();
/** stop hidden animation for slight performance improvement */
this.updateProgressBar(0);
$('.progress-bar').removeClass('active');
break;
case 'invalid':
break;
}
}
/**
* sets up the daterange selector and adds listeners
* @returns {null} - nothing
*/
setupDateRangeSelector() {
super.setupDateRangeSelector();
$(this.config.dateRangeSelector).on('apply.daterangepicker', (e, action) => {
if (action.chosenLabel === $.i18n('custom-range')) {
this.specialRange = null;
/** force events to re-fire since apply.daterangepicker occurs before 'change' event */
this.daterangepicker.updateElement();
}
});
}
/**
* Process the langviews for the article and options entered
* Called when submitting the form
* @return {null} nothing
*/
| processArticle | identifier_name |
|
langviews.js | -badge' src='${badgeImage}' alt='${badgeName}' title='${badgeName}' />`;
}
/**
* Render list of langviews into view
* @returns {null} nothing
*/
renderData() {
/** sort ascending by current sort setting */
const sortedLangViews = this.langData.sort((a, b) => {
const before = this.getSortProperty(a, this.sort),
after = this.getSortProperty(b, this.sort);
if (before < after) {
return this.direction;
} else if (before > after) {
return -this.direction;
} else {
return 0;
}
});
$('.sort-link span').removeClass('glyphicon-sort-by-alphabet-alt glyphicon-sort-by-alphabet').addClass('glyphicon-sort');
const newSortClassName = parseInt(this.direction) === 1 ? 'glyphicon-sort-by-alphabet-alt' : 'glyphicon-sort-by-alphabet';
$(`.sort-link--${this.sort} span`).addClass(newSortClassName).removeClass('glyphicon-sort');
const totalBadgesMarkup = Object.keys(this.totals.badges).map(badge => {
return `<span class='nowrap'>${this.getBadgeMarkup(badge)} × ${this.totals.badges[badge]}</span>`;
}).join(', ');
$('.output-totals').html(
`<th scope='row'>${$.i18n('totals')}</th>
<th>${$.i18n('num-languages', this.langData.length)}</th>
<th>${$.i18n('unique-titles', this.totals.titles.length)}</th>
<th>${totalBadgesMarkup}</th>
<th>${this.n(this.totals.views)}</th>
<th>${this.n(Math.round(this.totals.views / this.numDaysInRange()))} / ${$.i18n('day')}</th>`
);
$('#lang_list').html('');
sortedLangViews.forEach((item, index) => {
let badgeMarkup = '';
if (item.badges) {
badgeMarkup = item.badges.map(this.getBadgeMarkup.bind(this)).join();
}
$('#lang_list').append(
`<tr>
<th scope='row'>${index + 1}</th>
<td>${item.lang}</td>
<td><a href="${item.url}" target="_blank">${item.pageName}</a></td>
<td>${badgeMarkup}</td>
<td><a href='${this.getPageviewsURL(item.lang, this.baseProject, item.pageName)}'>${this.n(item.views)}</a></td>
<td>${this.n(item.average)} / ${$.i18n('day')}</td>
</tr>`
);
});
this.pushParams();
this.setState('complete');
}
/**
* Get value of given langview entry for the purposes of column sorting
* @param {object} item - langview entry within this.langData
* @param {String} type - type of property to get
* @return {String|Number} - value
*/
getSortProperty(item, type) {
switch (type) {
case 'lang':
return item.lang;
case 'title':
return item.pageName;
case 'badges':
return item.badges.sort().join('');
case 'views':
return Number(item.views);
}
}
/**
* Link to /pageviews for given article and chosen daterange
* @param {String} lang - two character language code
* @param {String} project - base project without lang, e.g. wikipedia.org
* @param {String} article - page name
* @returns {String} URL
*/
// FIXME: should include agent and platform, and use special ranges as currently specified
getPageviewsURL(lang, project, article) {
let startDate = moment(this.daterangepicker.startDate),
endDate = moment(this.daterangepicker.endDate);
const platform = $(this.config.platformSelector).val();
if (endDate.diff(startDate, 'days') === 0) {
startDate.subtract(3, 'days');
endDate.add(3, 'days');
}
return `/pageviews#start=${startDate.format('YYYY-MM-DD')}` +
`&end=${endDate.format('YYYY-MM-DD')}&project=${lang}.${project}.org&platform=${platform}&pages=${article}`;
}
/**
* Loop through given interwiki data and query the pageviews API for each
* Also updates this.langData with result
* @param {Object} interWikiData - as given by the getInterwikiData promise
* @return {Deferred} - Promise resolving with data ready to be rendered to view
*/
getPageViewsData(interWikiData) {
const startDate = this.daterangepicker.startDate.startOf('day'),
endDate = this.daterangepicker.endDate.startOf('day'),
interWikiKeys = Object.keys(interWikiData);
// XXX: throttling
let dfd = $.Deferred(), promises = [], count = 0, hadFailure, failureRetries = {},
totalRequestCount = interWikiKeys.length, failedPages = [];
/** clear out existing data */
this.langData = [];
const makeRequest = dbName => {
const data = interWikiData[dbName],
uriEncodedPageName = encodeURIComponent(data.title);
const url = (
`https://wikimedia.org/api/rest_v1/metrics/pageviews/per-article/${data.lang}.${this.baseProject}` +
`/${$(this.config.platformSelector).val()}/${$(this.config.agentSelector).val()}/${uriEncodedPageName}/daily` +
`/${startDate.format(this.config.timestampFormat)}/${endDate.format(this.config.timestampFormat)}`
);
const promise = $.ajax({ url, dataType: 'json' });
promises.push(promise);
promise.done(pvData => {
const viewCounts = pvData.items.map(el => el.views),
views = viewCounts.reduce((a, b) => a + b);
this.langData.push({
badges: data.badges,
dbName,
lang: data.lang,
pageName: data.title,
views,
url: data.url,
average: Math.round(views / this.numDaysInRange())
});
/** keep track of unique badges/titles and total pageviews */
this.totals.views += views;
if (!this.totals.titles.includes(data.title)) {
this.totals.titles.push(data.title);
}
data.badges.forEach(badge => {
if (this.totals.badges[badge] === undefined) {
this.totals.badges[badge] = 1;
} else {
this.totals.badges[badge] += 1;
}
});
}).fail(errorData => {
// XXX: throttling
/** first detect if this was a Cassandra backend error, and if so, schedule a re-try */
const cassandraError = errorData.responseJSON.title === 'Error in Cassandra table storage backend',
failedPageLink = this.getPageLink(data.title, `${data.lang}.${this.baseProject}.org`);
if (cassandraError) {
if (failureRetries[dbName]) {
failureRetries[dbName]++;
} else { | totalRequestCount++;
return this.rateLimit(makeRequest, 100, this)(dbName);
}
/** retries exceeded */
failedPages.push(failedPageLink);
} else {
this.writeMessage(
`${failedPageLink}: ${$.i18n('api-error', 'Pageviews API')} - ${errorData.responseJSON.title}`
);
}
hadFailure = true; // don't treat this series of requests as being cached by server
}).always(() => {
this.updateProgressBar((++count / totalRequestCount) * 100);
// XXX: throttling, totalRequestCount can just be interWikiKeys.length
if (count === totalRequestCount) {
dfd.resolve(this.langData);
if (failedPages.length) {
this.writeMessage($.i18n(
'api-error-timeout',
'<ul>' +
failedPages.map(failedPage => `<li>${failedPage}</li>`).join('') +
'</ul>'
));
}
/**
* if there were no failures, assume the resource is now cached by the server
* and save this assumption to our own cache so we don't throttle the same requests
*/
// XXX: throttling
if (!hadFailure) {
simpleStorage.set(this.getCacheKey(), true, {TTL: 600000});
}
}
});
};
/**
* We don't want to throttle requests for cached resources. However in our case,
* we're unable to check response headers to see if the resource was cached,
* so we use simpleStorage to keep track of what the user has recently queried.
*/
// XXX: throttling
const requestFn = this.isRequestCached() ? makeRequest : this.rateLimit(makeRequest, 100, this);
interWikiKeys.forEach((dbName, index) => {
requestFn(dbName);
});
return dfd;
}
/**
* Return cache key for current params
* @return {String} key
*/
| failureRetries[dbName] = 1;
}
/** maximum of 3 retries */
if (failureRetries[dbName] < 3) { | random_line_split |
langviews.js | of column sorting
* @param {object} item - langview entry within this.langData
* @param {String} type - type of property to get
* @return {String|Number} - value
*/
getSortProperty(item, type) {
switch (type) {
case 'lang':
return item.lang;
case 'title':
return item.pageName;
case 'badges':
return item.badges.sort().join('');
case 'views':
return Number(item.views);
}
}
/**
* Link to /pageviews for given article and chosen daterange
* @param {String} lang - two character language code
* @param {String} project - base project without lang, e.g. wikipedia.org
* @param {String} article - page name
* @returns {String} URL
*/
// FIXME: should include agent and platform, and use special ranges as currently specified
getPageviewsURL(lang, project, article) {
let startDate = moment(this.daterangepicker.startDate),
endDate = moment(this.daterangepicker.endDate);
const platform = $(this.config.platformSelector).val();
if (endDate.diff(startDate, 'days') === 0) {
startDate.subtract(3, 'days');
endDate.add(3, 'days');
}
return `/pageviews#start=${startDate.format('YYYY-MM-DD')}` +
`&end=${endDate.format('YYYY-MM-DD')}&project=${lang}.${project}.org&platform=${platform}&pages=${article}`;
}
/**
* Loop through given interwiki data and query the pageviews API for each
* Also updates this.langData with result
* @param {Object} interWikiData - as given by the getInterwikiData promise
* @return {Deferred} - Promise resolving with data ready to be rendered to view
*/
getPageViewsData(interWikiData) {
const startDate = this.daterangepicker.startDate.startOf('day'),
endDate = this.daterangepicker.endDate.startOf('day'),
interWikiKeys = Object.keys(interWikiData);
// XXX: throttling
let dfd = $.Deferred(), promises = [], count = 0, hadFailure, failureRetries = {},
totalRequestCount = interWikiKeys.length, failedPages = [];
/** clear out existing data */
this.langData = [];
const makeRequest = dbName => {
const data = interWikiData[dbName],
uriEncodedPageName = encodeURIComponent(data.title);
const url = (
`https://wikimedia.org/api/rest_v1/metrics/pageviews/per-article/${data.lang}.${this.baseProject}` +
`/${$(this.config.platformSelector).val()}/${$(this.config.agentSelector).val()}/${uriEncodedPageName}/daily` +
`/${startDate.format(this.config.timestampFormat)}/${endDate.format(this.config.timestampFormat)}`
);
const promise = $.ajax({ url, dataType: 'json' });
promises.push(promise);
promise.done(pvData => {
const viewCounts = pvData.items.map(el => el.views),
views = viewCounts.reduce((a, b) => a + b);
this.langData.push({
badges: data.badges,
dbName,
lang: data.lang,
pageName: data.title,
views,
url: data.url,
average: Math.round(views / this.numDaysInRange())
});
/** keep track of unique badges/titles and total pageviews */
this.totals.views += views;
if (!this.totals.titles.includes(data.title)) {
this.totals.titles.push(data.title);
}
data.badges.forEach(badge => {
if (this.totals.badges[badge] === undefined) {
this.totals.badges[badge] = 1;
} else {
this.totals.badges[badge] += 1;
}
});
}).fail(errorData => {
// XXX: throttling
/** first detect if this was a Cassandra backend error, and if so, schedule a re-try */
const cassandraError = errorData.responseJSON.title === 'Error in Cassandra table storage backend',
failedPageLink = this.getPageLink(data.title, `${data.lang}.${this.baseProject}.org`);
if (cassandraError) {
if (failureRetries[dbName]) {
failureRetries[dbName]++;
} else {
failureRetries[dbName] = 1;
}
/** maximum of 3 retries */
if (failureRetries[dbName] < 3) {
totalRequestCount++;
return this.rateLimit(makeRequest, 100, this)(dbName);
}
/** retries exceeded */
failedPages.push(failedPageLink);
} else {
this.writeMessage(
`${failedPageLink}: ${$.i18n('api-error', 'Pageviews API')} - ${errorData.responseJSON.title}`
);
}
hadFailure = true; // don't treat this series of requests as being cached by server
}).always(() => {
this.updateProgressBar((++count / totalRequestCount) * 100);
// XXX: throttling, totalRequestCount can just be interWikiKeys.length
if (count === totalRequestCount) {
dfd.resolve(this.langData);
if (failedPages.length) {
this.writeMessage($.i18n(
'api-error-timeout',
'<ul>' +
failedPages.map(failedPage => `<li>${failedPage}</li>`).join('') +
'</ul>'
));
}
/**
* if there were no failures, assume the resource is now cached by the server
* and save this assumption to our own cache so we don't throttle the same requests
*/
// XXX: throttling
if (!hadFailure) {
simpleStorage.set(this.getCacheKey(), true, {TTL: 600000});
}
}
});
};
/**
* We don't want to throttle requests for cached resources. However in our case,
* we're unable to check response headers to see if the resource was cached,
* so we use simpleStorage to keep track of what the user has recently queried.
*/
// XXX: throttling
const requestFn = this.isRequestCached() ? makeRequest : this.rateLimit(makeRequest, 100, this);
interWikiKeys.forEach((dbName, index) => {
requestFn(dbName);
});
return dfd;
}
/**
* Return cache key for current params
* @return {String} key
*/
getCacheKey() {
return `lv-cache-${this.hashCode(
JSON.stringify(this.getParams(true))
)}`;
}
/**
* Check simple storage to see if a request with the current params would be cached
* @return {Boolean} cached or not
*/
isRequestCached() {
return simpleStorage.hasKey(this.getCacheKey());
}
/**
* Query Wikidata to find data about a given page across all sister projects
* @param {String} dbName - database name of source project
* @param {String} pageName - name of page we want to get data about
* @return {Deferred} - Promise resolving with interwiki data
*/
getInterwikiData(dbName, pageName) {
const dfd = $.Deferred();
const url = `https://www.wikidata.org/w/api.php?action=wbgetentities&sites=${dbName}` +
`&titles=${encodeURIComponent(pageName)}&props=sitelinks/urls|datatype&format=json&callback=?`;
$.getJSON(url).done(data => {
if (data.error) {
return dfd.reject(`${$.i18n('api-error', 'Wikidata')}: ${data.error.info}`);
} else if (data.entities['-1']) {
return dfd.reject(
`<a href='${this.getPageURL(pageName)}'>${pageName.descore()}</a> - ${$.i18n('api-error-no-data')}`
);
}
const key = Object.keys(data.entities)[0],
sitelinks = data.entities[key].sitelinks,
filteredLinks = {},
matchRegex = new RegExp(`^https://[\\w-]+\\.${this.baseProject}\\.org`);
/** restrict to selected base project (e.g. wikipedias, not wikipedias and wikivoyages) */
Object.keys(sitelinks).forEach(key => {
const siteMapKey = sitelinks[key].site.replace(/-/g, '_');
if (matchRegex.test(sitelinks[key].url) && siteMap[siteMapKey]) {
sitelinks[key].lang = siteMap[siteMapKey].replace(/\.wiki.*$/, '');
filteredLinks[key] = sitelinks[key];
}
});
return dfd.resolve(filteredLinks);
});
return dfd;
}
/**
* Parse wiki URL for the page name
* @param {String} url - full URL to a wiki page
* @return {String|null} page name
*/
getPageNameFromURL(url) {
if (url.includes('?')) {
return url.match(/\?(?:.*\b)?title=(.*?)(?:&|$)/)[1];
} else | {
return url.match(/\/wiki\/(.*?)(?:\?|$)/)[1];
} | conditional_block |
|
gaia_executor.rs | store_config
} else {
StoreConfig::init()
}
};
let (alive_id, partitions) = get_init_info(&store_config);
info!("alive_id: {:?}, partitions: {:?}", alive_id, partitions);
store_config.update_alive_id(alive_id);
info!("{:?}", store_config);
let worker_num = store_config.timely_worker_per_process;
let store_config = Arc::new(store_config);
if store_config.graph_type.to_lowercase().eq(VINEYARD_GRAPH) {
if cfg!(target_os = "linux") {
info!(
"Start executor with vineyard graph object id {:?}",
store_config.vineyard_graph_id
);
use maxgraph_runtime::store::ffi::FFIGraphStore;
let ffi_store = FFIGraphStore::new(store_config.vineyard_graph_id, worker_num as i32);
let partition_manager = ffi_store.get_partition_manager();
run_main(
store_config,
Arc::new(ffi_store),
Arc::new(partition_manager),
);
} else {
unimplemented!("Mac not support vineyard graph")
}
} else {
unimplemented!("only start vineyard graph from executor")
}
}
fn run_main<V, VI, E, EI>(
store_config: Arc<StoreConfig>,
graph: Arc<GlobalGraphQuery<V = V, E = E, VI = VI, EI = EI>>,
partition_manager: Arc<GraphPartitionManager>,
) where
V: Vertex + 'static,
VI: Iterator<Item = V> + Send + 'static,
E: Edge + 'static,
EI: Iterator<Item = E> + Send + 'static,
{
let process_partition_list = partition_manager.get_process_partition_list();
info!("process_partition_list: {:?}", process_partition_list);
let runtime_info = Arc::new(Mutex::new(RuntimeInfo::new(
store_config.timely_worker_per_process,
process_partition_list,
)));
let runtime_info_clone = runtime_info.clone();
let (hb_resp_sender, hb_resp_receiver) = channel();
let signal = Arc::new(AtomicBool::new(false));
let gaia_server_manager =
GaiaServerManager::new(hb_resp_receiver, runtime_info, signal.clone());
let partition_worker_mapping = gaia_server_manager.get_partition_worker_mapping();
let worker_partition_list_mapping = gaia_server_manager.get_worker_partition_list_mapping();
let server_manager = Box::new(gaia_server_manager);
let _manager_guards = ServerManager::start_server(
server_manager,
store_config.clone(),
Box::new(recover_prepare),
)
.unwrap();
let gaia_service = GaiaService::new(
store_config.clone(),
graph.clone(),
partition_manager.clone(),
partition_worker_mapping,
worker_partition_list_mapping,
);
let (_, gaia_rpc_service_port) = gaia_service.start_rpc_service();
let store_context = StoreContext::new(graph, partition_manager);
start_hb_rpc_service(
runtime_info_clone,
store_config,
gaia_rpc_service_port,
hb_resp_sender,
store_context,
);
thread::sleep(Duration::from_secs(u64::max_value()));
}
fn recover_prepare(prepared: &[u8]) -> Result<Vec<u8>, String> {
::protobuf::parse_from_bytes::<QueryFlow>(prepared)
.map_err(|err| err.to_string())
.and_then(move |desc| {
info!("parse {} bytes to {:?} ", prepared.len(), desc);
Ok(desc.write_to_bytes().expect("query flow to bytes"))
})
}
fn start_hb_rpc_service<VV, VVI, EE, EEI>(
runtime_info: Arc<Mutex<RuntimeInfo>>,
store_config: Arc<StoreConfig>,
gaia_service_port: u16,
hb_resp_sender: Sender<Arc<ServerHBResp>>,
store_context: StoreContext<VV, VVI, EE, EEI>,
) where
VV: 'static + Vertex,
VVI: 'static + Iterator<Item = VV> + Send,
EE: 'static + Edge,
EEI: 'static + Iterator<Item = EE> + Send,
{
// build hb information
let mut hb_providers = Vec::new();
let mut hb_resp_senders = Vec::new();
let hb_provider = move |ref mut server_hb_req: &mut ServerHBReq| {
server_hb_req.set_runtimeReq(build_runtime_req(runtime_info.clone()));
server_hb_req
.mut_endpoint()
.set_runtimCtrlAndAsyncPort(gaia_service_port as i32);
};
hb_providers.push(Box::new(hb_provider));
hb_resp_senders.push(hb_resp_sender);
let store_config_clone = store_config.clone();
init_with_rpc_service(
store_config_clone,
hb_providers,
hb_resp_senders,
store_context,
);
}
fn build_runtime_req(runtime_info: Arc<Mutex<RuntimeInfo>>) -> RuntimeHBReq {
let hb_req = runtime_info.lock().expect("Lock runtime hb req failed");
let mut runtime_req = RuntimeHBReq::new();
runtime_req.set_serverStatus(hb_req.get_server_status());
runtime_req.set_runtimePort(hb_req.get_server_port() as i32);
runtime_req.set_worker_num_per_process(hb_req.get_worker_num_per_process());
runtime_req.set_process_partition_list(hb_req.get_process_partition_list().to_vec());
debug!("Build runtime request {:?} in heartbeat", &runtime_req);
runtime_req
}
/// return: (aliveId, partiiton assignments)
fn get_init_info(config: &StoreConfig) -> (u64, Vec<PartitionId>) {
use maxgraph_common::proto::data_grpc::*;
use maxgraph_common::util::ip;
use maxgraph_server::client::ZKClient;
let zk_url = format!("{}/{}", config.zk_url, config.graph_name);
let zk = ZKClient::new(&zk_url, config.zk_timeout_ms, config.get_zk_auth());
let addr = zk.get_coordinator_addr();
let channel =
ChannelBuilder::new(Arc::new(EnvBuilder::new().build())).connect(addr.to_string().as_str());
let client = ServerDataApiClient::new(channel);
let mut request = GetExecutorAliveIdRequest::new();
request.set_serverId(config.worker_id);
request.set_ip(ip::get_local_ip());
let response = client.get_executor_alive_id(&request).unwrap();
let alive_id = response.get_aliveId();
let mut request = GetPartitionAssignmentRequest::new();
request.set_serverId(config.worker_id);
let response = client.get_partition_assignment(&request).unwrap();
let partitions = response.get_partitionId().to_vec();
(alive_id, partitions)
}
pub struct GaiaService<V, VI, E, EI>
where
V: Vertex + 'static,
VI: Iterator<Item = V> + Send + 'static,
E: Edge + 'static,
EI: Iterator<Item = E> + Send + 'static,
{
store_config: Arc<StoreConfig>,
graph: Arc<dyn GlobalGraphQuery<V = V, E = E, VI = VI, EI = EI>>,
partition_manager: Arc<dyn GraphPartitionManager>,
// mapping of partition id -> worker id
partition_worker_mapping: Arc<RwLock<Option<HashMap<u32, u32>>>>,
// mapping of worker id -> partition list
worker_partition_list_mapping: Arc<RwLock<Option<HashMap<u32, Vec<u32>>>>>,
rpc_runtime: Runtime,
}
impl<V, VI, E, EI> GaiaService<V, VI, E, EI>
where
V: Vertex + 'static,
VI: Iterator<Item = V> + Send + 'static,
E: Edge + 'static,
EI: Iterator<Item = E> + Send + 'static,
{
pub fn new(
store_config: Arc<StoreConfig>,
graph: Arc<dyn GlobalGraphQuery<V = V, E = E, VI = VI, EI = EI>>,
partition_manager: Arc<dyn GraphPartitionManager>,
partition_worker_mapping: Arc<RwLock<Option<HashMap<u32, u32>>>>,
worker_partition_list_mapping: Arc<RwLock<Option<HashMap<u32, Vec<u32>>>>>,
) -> GaiaService<V, VI, E, EI> {
GaiaService {
store_config,
graph,
partition_manager,
partition_worker_mapping,
worker_partition_list_mapping,
rpc_runtime: Runtime::new().unwrap(),
}
}
pub fn start_rpc_service(&self) -> (String, u16) | {
let rpc_port = self.rpc_runtime.block_on(async {
let query_vineyard = QueryVineyard::new(
self.graph.clone(),
self.partition_manager.clone(),
self.partition_worker_mapping.clone(),
self.worker_partition_list_mapping.clone(),
self.store_config.worker_num as usize,
self.store_config.worker_id as u64,
);
let job_compiler = query_vineyard.initialize_job_compiler();
let service = Service::new(job_compiler);
let addr = format!("{}:{}", "0.0.0.0", self.store_config.rpc_port);
let local_addr = start_rpc_server(addr.parse().unwrap(), service, true, false)
.await
.unwrap();
local_addr.port()
});
let ip = get_local_ip();
info!("start rpc server on {} {}", ip, rpc_port); | identifier_body |
|
gaia_executor.rs | crate maxgraph_server;
extern crate maxgraph_store;
extern crate pegasus;
extern crate pegasus_server;
extern crate protobuf;
extern crate structopt;
use gaia_runtime::server::init_with_rpc_service;
use gaia_runtime::server::manager::GaiaServerManager;
use grpcio::ChannelBuilder;
use grpcio::EnvBuilder;
use gs_gremlin::{InitializeJobCompiler, QueryVineyard};
use maxgraph_common::proto::data::*;
use maxgraph_common::proto::hb::*;
use maxgraph_common::proto::query_flow::*;
use maxgraph_common::util;
use maxgraph_common::util::get_local_ip;
use maxgraph_common::util::log4rs::init_log4rs;
use maxgraph_runtime::server::manager::*;
use maxgraph_runtime::server::RuntimeInfo;
use maxgraph_server::StoreContext;
use maxgraph_store::api::graph_partition::GraphPartitionManager;
use maxgraph_store::api::prelude::*;
use maxgraph_store::config::{StoreConfig, VINEYARD_GRAPH};
use pegasus_server::rpc::start_rpc_server;
use pegasus_server::service::Service;
use protobuf::Message;
use std::collections::HashMap;
use std::env;
use std::sync::atomic::AtomicBool;
use std::sync::mpsc::{channel, Sender};
use std::sync::{Arc, Mutex, RwLock};
use std::thread;
use std::time::Duration;
use tokio::runtime::Runtime;
fn main() {
if let Some(_) = env::args().find(|arg| arg == "--show-build-info") |
init_log4rs();
let mut store_config = {
let args: Vec<String> = std::env::args().collect();
if args.len() <= 6 && args[1] == "--config" {
let mut store_config = StoreConfig::init_from_file(&args[2], &args[4]);
if args.len() == 6 {
store_config.graph_name = (&args[5]).to_string();
}
store_config
} else {
StoreConfig::init()
}
};
let (alive_id, partitions) = get_init_info(&store_config);
info!("alive_id: {:?}, partitions: {:?}", alive_id, partitions);
store_config.update_alive_id(alive_id);
info!("{:?}", store_config);
let worker_num = store_config.timely_worker_per_process;
let store_config = Arc::new(store_config);
if store_config.graph_type.to_lowercase().eq(VINEYARD_GRAPH) {
if cfg!(target_os = "linux") {
info!(
"Start executor with vineyard graph object id {:?}",
store_config.vineyard_graph_id
);
use maxgraph_runtime::store::ffi::FFIGraphStore;
let ffi_store = FFIGraphStore::new(store_config.vineyard_graph_id, worker_num as i32);
let partition_manager = ffi_store.get_partition_manager();
run_main(
store_config,
Arc::new(ffi_store),
Arc::new(partition_manager),
);
} else {
unimplemented!("Mac not support vineyard graph")
}
} else {
unimplemented!("only start vineyard graph from executor")
}
}
fn run_main<V, VI, E, EI>(
store_config: Arc<StoreConfig>,
graph: Arc<GlobalGraphQuery<V = V, E = E, VI = VI, EI = EI>>,
partition_manager: Arc<GraphPartitionManager>,
) where
V: Vertex + 'static,
VI: Iterator<Item = V> + Send + 'static,
E: Edge + 'static,
EI: Iterator<Item = E> + Send + 'static,
{
let process_partition_list = partition_manager.get_process_partition_list();
info!("process_partition_list: {:?}", process_partition_list);
let runtime_info = Arc::new(Mutex::new(RuntimeInfo::new(
store_config.timely_worker_per_process,
process_partition_list,
)));
let runtime_info_clone = runtime_info.clone();
let (hb_resp_sender, hb_resp_receiver) = channel();
let signal = Arc::new(AtomicBool::new(false));
let gaia_server_manager =
GaiaServerManager::new(hb_resp_receiver, runtime_info, signal.clone());
let partition_worker_mapping = gaia_server_manager.get_partition_worker_mapping();
let worker_partition_list_mapping = gaia_server_manager.get_worker_partition_list_mapping();
let server_manager = Box::new(gaia_server_manager);
let _manager_guards = ServerManager::start_server(
server_manager,
store_config.clone(),
Box::new(recover_prepare),
)
.unwrap();
let gaia_service = GaiaService::new(
store_config.clone(),
graph.clone(),
partition_manager.clone(),
partition_worker_mapping,
worker_partition_list_mapping,
);
let (_, gaia_rpc_service_port) = gaia_service.start_rpc_service();
let store_context = StoreContext::new(graph, partition_manager);
start_hb_rpc_service(
runtime_info_clone,
store_config,
gaia_rpc_service_port,
hb_resp_sender,
store_context,
);
thread::sleep(Duration::from_secs(u64::max_value()));
}
fn recover_prepare(prepared: &[u8]) -> Result<Vec<u8>, String> {
::protobuf::parse_from_bytes::<QueryFlow>(prepared)
.map_err(|err| err.to_string())
.and_then(move |desc| {
info!("parse {} bytes to {:?} ", prepared.len(), desc);
Ok(desc.write_to_bytes().expect("query flow to bytes"))
})
}
fn start_hb_rpc_service<VV, VVI, EE, EEI>(
runtime_info: Arc<Mutex<RuntimeInfo>>,
store_config: Arc<StoreConfig>,
gaia_service_port: u16,
hb_resp_sender: Sender<Arc<ServerHBResp>>,
store_context: StoreContext<VV, VVI, EE, EEI>,
) where
VV: 'static + Vertex,
VVI: 'static + Iterator<Item = VV> + Send,
EE: 'static + Edge,
EEI: 'static + Iterator<Item = EE> + Send,
{
// build hb information
let mut hb_providers = Vec::new();
let mut hb_resp_senders = Vec::new();
let hb_provider = move |ref mut server_hb_req: &mut ServerHBReq| {
server_hb_req.set_runtimeReq(build_runtime_req(runtime_info.clone()));
server_hb_req
.mut_endpoint()
.set_runtimCtrlAndAsyncPort(gaia_service_port as i32);
};
hb_providers.push(Box::new(hb_provider));
hb_resp_senders.push(hb_resp_sender);
let store_config_clone = store_config.clone();
init_with_rpc_service(
store_config_clone,
hb_providers,
hb_resp_senders,
store_context,
);
}
fn build_runtime_req(runtime_info: Arc<Mutex<RuntimeInfo>>) -> RuntimeHBReq {
let hb_req = runtime_info.lock().expect("Lock runtime hb req failed");
let mut runtime_req = RuntimeHBReq::new();
runtime_req.set_serverStatus(hb_req.get_server_status());
runtime_req.set_runtimePort(hb_req.get_server_port() as i32);
runtime_req.set_worker_num_per_process(hb_req.get_worker_num_per_process());
runtime_req.set_process_partition_list(hb_req.get_process_partition_list().to_vec());
debug!("Build runtime request {:?} in heartbeat", &runtime_req);
runtime_req
}
/// return: (aliveId, partiiton assignments)
fn get_init_info(config: &StoreConfig) -> (u64, Vec<PartitionId>) {
use maxgraph_common::proto::data_grpc::*;
use maxgraph_common::util::ip;
use maxgraph_server::client::ZKClient;
let zk_url = format!("{}/{}", config.zk_url, config.graph_name);
let zk = ZKClient::new(&zk_url, config.zk_timeout_ms, config.get_zk_auth());
let addr = zk.get_coordinator_addr();
let channel =
ChannelBuilder::new(Arc::new(EnvBuilder::new().build())).connect(addr.to_string().as_str());
let client = ServerDataApiClient::new(channel);
let mut request = GetExecutorAliveIdRequest::new();
request.set_serverId(config.worker_id);
request.set_ip(ip::get_local_ip());
let response = client.get_executor_alive_id(&request).unwrap();
let alive_id = response.get_aliveId();
let mut request = GetPartitionAssignmentRequest::new();
request.set_serverId(config.worker_id);
let response = client.get_partition_assignment(&request).unwrap();
let partitions = response.get_partitionId().to_vec();
(alive_id, partitions)
}
pub struct GaiaService<V, VI, E, EI>
where
V: Vertex + 'static,
VI: Iterator<Item = V> + Send + 'static,
E: Edge + 'static,
EI: Iterator<Item = E> + Send + 'static,
{
store_config: Arc<StoreConfig>,
graph: Arc<dyn GlobalGraphQuery<V = V, E = E, VI = VI, EI = EI>>,
partition_manager: Arc<dyn GraphPartitionManager>,
// mapping of partition id -> worker id
partition_worker_mapping: Arc<RwLock<Option<HashMap<u32, u32>>>>,
// mapping of worker id -> partition | {
util::get_build_info();
return;
} | conditional_block |
gaia_executor.rs | extern crate maxgraph_server;
extern crate maxgraph_store;
extern crate pegasus;
extern crate pegasus_server;
extern crate protobuf;
extern crate structopt;
use gaia_runtime::server::init_with_rpc_service;
use gaia_runtime::server::manager::GaiaServerManager;
use grpcio::ChannelBuilder;
use grpcio::EnvBuilder;
use gs_gremlin::{InitializeJobCompiler, QueryVineyard};
use maxgraph_common::proto::data::*;
use maxgraph_common::proto::hb::*;
use maxgraph_common::proto::query_flow::*;
use maxgraph_common::util;
use maxgraph_common::util::get_local_ip;
use maxgraph_common::util::log4rs::init_log4rs;
use maxgraph_runtime::server::manager::*;
use maxgraph_runtime::server::RuntimeInfo;
use maxgraph_server::StoreContext;
use maxgraph_store::api::graph_partition::GraphPartitionManager;
use maxgraph_store::api::prelude::*;
use maxgraph_store::config::{StoreConfig, VINEYARD_GRAPH};
use pegasus_server::rpc::start_rpc_server;
use pegasus_server::service::Service;
use protobuf::Message;
use std::collections::HashMap;
use std::env;
use std::sync::atomic::AtomicBool;
use std::sync::mpsc::{channel, Sender};
use std::sync::{Arc, Mutex, RwLock};
use std::thread;
use std::time::Duration;
use tokio::runtime::Runtime;
fn main() {
if let Some(_) = env::args().find(|arg| arg == "--show-build-info") {
util::get_build_info();
return;
}
init_log4rs();
let mut store_config = {
let args: Vec<String> = std::env::args().collect();
if args.len() <= 6 && args[1] == "--config" {
let mut store_config = StoreConfig::init_from_file(&args[2], &args[4]);
if args.len() == 6 {
store_config.graph_name = (&args[5]).to_string();
}
store_config
} else {
StoreConfig::init()
}
};
let (alive_id, partitions) = get_init_info(&store_config);
info!("alive_id: {:?}, partitions: {:?}", alive_id, partitions);
store_config.update_alive_id(alive_id);
info!("{:?}", store_config);
let worker_num = store_config.timely_worker_per_process;
let store_config = Arc::new(store_config);
if store_config.graph_type.to_lowercase().eq(VINEYARD_GRAPH) {
if cfg!(target_os = "linux") {
info!(
"Start executor with vineyard graph object id {:?}",
store_config.vineyard_graph_id
);
use maxgraph_runtime::store::ffi::FFIGraphStore;
let ffi_store = FFIGraphStore::new(store_config.vineyard_graph_id, worker_num as i32);
let partition_manager = ffi_store.get_partition_manager();
run_main(
store_config,
Arc::new(ffi_store),
Arc::new(partition_manager),
);
} else {
unimplemented!("Mac not support vineyard graph")
}
} else {
unimplemented!("only start vineyard graph from executor")
}
}
fn run_main<V, VI, E, EI>(
store_config: Arc<StoreConfig>,
graph: Arc<GlobalGraphQuery<V = V, E = E, VI = VI, EI = EI>>,
partition_manager: Arc<GraphPartitionManager>,
) where
V: Vertex + 'static,
VI: Iterator<Item = V> + Send + 'static,
E: Edge + 'static,
EI: Iterator<Item = E> + Send + 'static,
{
let process_partition_list = partition_manager.get_process_partition_list();
info!("process_partition_list: {:?}", process_partition_list);
let runtime_info = Arc::new(Mutex::new(RuntimeInfo::new(
store_config.timely_worker_per_process,
process_partition_list,
)));
let runtime_info_clone = runtime_info.clone();
let (hb_resp_sender, hb_resp_receiver) = channel();
let signal = Arc::new(AtomicBool::new(false));
let gaia_server_manager =
GaiaServerManager::new(hb_resp_receiver, runtime_info, signal.clone());
let partition_worker_mapping = gaia_server_manager.get_partition_worker_mapping();
let worker_partition_list_mapping = gaia_server_manager.get_worker_partition_list_mapping();
let server_manager = Box::new(gaia_server_manager);
let _manager_guards = ServerManager::start_server( | server_manager,
store_config.clone(),
Box::new(recover_prepare),
)
.unwrap();
let gaia_service = GaiaService::new(
store_config.clone(),
graph.clone(),
partition_manager.clone(),
partition_worker_mapping,
worker_partition_list_mapping,
);
let (_, gaia_rpc_service_port) = gaia_service.start_rpc_service();
let store_context = StoreContext::new(graph, partition_manager);
start_hb_rpc_service(
runtime_info_clone,
store_config,
gaia_rpc_service_port,
hb_resp_sender,
store_context,
);
thread::sleep(Duration::from_secs(u64::max_value()));
}
fn recover_prepare(prepared: &[u8]) -> Result<Vec<u8>, String> {
::protobuf::parse_from_bytes::<QueryFlow>(prepared)
.map_err(|err| err.to_string())
.and_then(move |desc| {
info!("parse {} bytes to {:?} ", prepared.len(), desc);
Ok(desc.write_to_bytes().expect("query flow to bytes"))
})
}
fn start_hb_rpc_service<VV, VVI, EE, EEI>(
runtime_info: Arc<Mutex<RuntimeInfo>>,
store_config: Arc<StoreConfig>,
gaia_service_port: u16,
hb_resp_sender: Sender<Arc<ServerHBResp>>,
store_context: StoreContext<VV, VVI, EE, EEI>,
) where
VV: 'static + Vertex,
VVI: 'static + Iterator<Item = VV> + Send,
EE: 'static + Edge,
EEI: 'static + Iterator<Item = EE> + Send,
{
// build hb information
let mut hb_providers = Vec::new();
let mut hb_resp_senders = Vec::new();
let hb_provider = move |ref mut server_hb_req: &mut ServerHBReq| {
server_hb_req.set_runtimeReq(build_runtime_req(runtime_info.clone()));
server_hb_req
.mut_endpoint()
.set_runtimCtrlAndAsyncPort(gaia_service_port as i32);
};
hb_providers.push(Box::new(hb_provider));
hb_resp_senders.push(hb_resp_sender);
let store_config_clone = store_config.clone();
init_with_rpc_service(
store_config_clone,
hb_providers,
hb_resp_senders,
store_context,
);
}
fn build_runtime_req(runtime_info: Arc<Mutex<RuntimeInfo>>) -> RuntimeHBReq {
let hb_req = runtime_info.lock().expect("Lock runtime hb req failed");
let mut runtime_req = RuntimeHBReq::new();
runtime_req.set_serverStatus(hb_req.get_server_status());
runtime_req.set_runtimePort(hb_req.get_server_port() as i32);
runtime_req.set_worker_num_per_process(hb_req.get_worker_num_per_process());
runtime_req.set_process_partition_list(hb_req.get_process_partition_list().to_vec());
debug!("Build runtime request {:?} in heartbeat", &runtime_req);
runtime_req
}
/// return: (aliveId, partiiton assignments)
fn get_init_info(config: &StoreConfig) -> (u64, Vec<PartitionId>) {
use maxgraph_common::proto::data_grpc::*;
use maxgraph_common::util::ip;
use maxgraph_server::client::ZKClient;
let zk_url = format!("{}/{}", config.zk_url, config.graph_name);
let zk = ZKClient::new(&zk_url, config.zk_timeout_ms, config.get_zk_auth());
let addr = zk.get_coordinator_addr();
let channel =
ChannelBuilder::new(Arc::new(EnvBuilder::new().build())).connect(addr.to_string().as_str());
let client = ServerDataApiClient::new(channel);
let mut request = GetExecutorAliveIdRequest::new();
request.set_serverId(config.worker_id);
request.set_ip(ip::get_local_ip());
let response = client.get_executor_alive_id(&request).unwrap();
let alive_id = response.get_aliveId();
let mut request = GetPartitionAssignmentRequest::new();
request.set_serverId(config.worker_id);
let response = client.get_partition_assignment(&request).unwrap();
let partitions = response.get_partitionId().to_vec();
(alive_id, partitions)
}
pub struct GaiaService<V, VI, E, EI>
where
V: Vertex + 'static,
VI: Iterator<Item = V> + Send + 'static,
E: Edge + 'static,
EI: Iterator<Item = E> + Send + 'static,
{
store_config: Arc<StoreConfig>,
graph: Arc<dyn GlobalGraphQuery<V = V, E = E, VI = VI, EI = EI>>,
partition_manager: Arc<dyn GraphPartitionManager>,
// mapping of partition id -> worker id
partition_worker_mapping: Arc<RwLock<Option<HashMap<u32, u32>>>>,
// mapping of worker id -> partition list | random_line_split |
|
gaia_executor.rs | crate maxgraph_server;
extern crate maxgraph_store;
extern crate pegasus;
extern crate pegasus_server;
extern crate protobuf;
extern crate structopt;
use gaia_runtime::server::init_with_rpc_service;
use gaia_runtime::server::manager::GaiaServerManager;
use grpcio::ChannelBuilder;
use grpcio::EnvBuilder;
use gs_gremlin::{InitializeJobCompiler, QueryVineyard};
use maxgraph_common::proto::data::*;
use maxgraph_common::proto::hb::*;
use maxgraph_common::proto::query_flow::*;
use maxgraph_common::util;
use maxgraph_common::util::get_local_ip;
use maxgraph_common::util::log4rs::init_log4rs;
use maxgraph_runtime::server::manager::*;
use maxgraph_runtime::server::RuntimeInfo;
use maxgraph_server::StoreContext;
use maxgraph_store::api::graph_partition::GraphPartitionManager;
use maxgraph_store::api::prelude::*;
use maxgraph_store::config::{StoreConfig, VINEYARD_GRAPH};
use pegasus_server::rpc::start_rpc_server;
use pegasus_server::service::Service;
use protobuf::Message;
use std::collections::HashMap;
use std::env;
use std::sync::atomic::AtomicBool;
use std::sync::mpsc::{channel, Sender};
use std::sync::{Arc, Mutex, RwLock};
use std::thread;
use std::time::Duration;
use tokio::runtime::Runtime;
fn main() {
if let Some(_) = env::args().find(|arg| arg == "--show-build-info") {
util::get_build_info();
return;
}
init_log4rs();
let mut store_config = {
let args: Vec<String> = std::env::args().collect();
if args.len() <= 6 && args[1] == "--config" {
let mut store_config = StoreConfig::init_from_file(&args[2], &args[4]);
if args.len() == 6 {
store_config.graph_name = (&args[5]).to_string();
}
store_config
} else {
StoreConfig::init()
}
};
let (alive_id, partitions) = get_init_info(&store_config);
info!("alive_id: {:?}, partitions: {:?}", alive_id, partitions);
store_config.update_alive_id(alive_id);
info!("{:?}", store_config);
let worker_num = store_config.timely_worker_per_process;
let store_config = Arc::new(store_config);
if store_config.graph_type.to_lowercase().eq(VINEYARD_GRAPH) {
if cfg!(target_os = "linux") {
info!(
"Start executor with vineyard graph object id {:?}",
store_config.vineyard_graph_id
);
use maxgraph_runtime::store::ffi::FFIGraphStore;
let ffi_store = FFIGraphStore::new(store_config.vineyard_graph_id, worker_num as i32);
let partition_manager = ffi_store.get_partition_manager();
run_main(
store_config,
Arc::new(ffi_store),
Arc::new(partition_manager),
);
} else {
unimplemented!("Mac not support vineyard graph")
}
} else {
unimplemented!("only start vineyard graph from executor")
}
}
fn | <V, VI, E, EI>(
store_config: Arc<StoreConfig>,
graph: Arc<GlobalGraphQuery<V = V, E = E, VI = VI, EI = EI>>,
partition_manager: Arc<GraphPartitionManager>,
) where
V: Vertex + 'static,
VI: Iterator<Item = V> + Send + 'static,
E: Edge + 'static,
EI: Iterator<Item = E> + Send + 'static,
{
let process_partition_list = partition_manager.get_process_partition_list();
info!("process_partition_list: {:?}", process_partition_list);
let runtime_info = Arc::new(Mutex::new(RuntimeInfo::new(
store_config.timely_worker_per_process,
process_partition_list,
)));
let runtime_info_clone = runtime_info.clone();
let (hb_resp_sender, hb_resp_receiver) = channel();
let signal = Arc::new(AtomicBool::new(false));
let gaia_server_manager =
GaiaServerManager::new(hb_resp_receiver, runtime_info, signal.clone());
let partition_worker_mapping = gaia_server_manager.get_partition_worker_mapping();
let worker_partition_list_mapping = gaia_server_manager.get_worker_partition_list_mapping();
let server_manager = Box::new(gaia_server_manager);
let _manager_guards = ServerManager::start_server(
server_manager,
store_config.clone(),
Box::new(recover_prepare),
)
.unwrap();
let gaia_service = GaiaService::new(
store_config.clone(),
graph.clone(),
partition_manager.clone(),
partition_worker_mapping,
worker_partition_list_mapping,
);
let (_, gaia_rpc_service_port) = gaia_service.start_rpc_service();
let store_context = StoreContext::new(graph, partition_manager);
start_hb_rpc_service(
runtime_info_clone,
store_config,
gaia_rpc_service_port,
hb_resp_sender,
store_context,
);
thread::sleep(Duration::from_secs(u64::max_value()));
}
fn recover_prepare(prepared: &[u8]) -> Result<Vec<u8>, String> {
::protobuf::parse_from_bytes::<QueryFlow>(prepared)
.map_err(|err| err.to_string())
.and_then(move |desc| {
info!("parse {} bytes to {:?} ", prepared.len(), desc);
Ok(desc.write_to_bytes().expect("query flow to bytes"))
})
}
fn start_hb_rpc_service<VV, VVI, EE, EEI>(
runtime_info: Arc<Mutex<RuntimeInfo>>,
store_config: Arc<StoreConfig>,
gaia_service_port: u16,
hb_resp_sender: Sender<Arc<ServerHBResp>>,
store_context: StoreContext<VV, VVI, EE, EEI>,
) where
VV: 'static + Vertex,
VVI: 'static + Iterator<Item = VV> + Send,
EE: 'static + Edge,
EEI: 'static + Iterator<Item = EE> + Send,
{
// build hb information
let mut hb_providers = Vec::new();
let mut hb_resp_senders = Vec::new();
let hb_provider = move |ref mut server_hb_req: &mut ServerHBReq| {
server_hb_req.set_runtimeReq(build_runtime_req(runtime_info.clone()));
server_hb_req
.mut_endpoint()
.set_runtimCtrlAndAsyncPort(gaia_service_port as i32);
};
hb_providers.push(Box::new(hb_provider));
hb_resp_senders.push(hb_resp_sender);
let store_config_clone = store_config.clone();
init_with_rpc_service(
store_config_clone,
hb_providers,
hb_resp_senders,
store_context,
);
}
fn build_runtime_req(runtime_info: Arc<Mutex<RuntimeInfo>>) -> RuntimeHBReq {
let hb_req = runtime_info.lock().expect("Lock runtime hb req failed");
let mut runtime_req = RuntimeHBReq::new();
runtime_req.set_serverStatus(hb_req.get_server_status());
runtime_req.set_runtimePort(hb_req.get_server_port() as i32);
runtime_req.set_worker_num_per_process(hb_req.get_worker_num_per_process());
runtime_req.set_process_partition_list(hb_req.get_process_partition_list().to_vec());
debug!("Build runtime request {:?} in heartbeat", &runtime_req);
runtime_req
}
/// return: (aliveId, partiiton assignments)
fn get_init_info(config: &StoreConfig) -> (u64, Vec<PartitionId>) {
use maxgraph_common::proto::data_grpc::*;
use maxgraph_common::util::ip;
use maxgraph_server::client::ZKClient;
let zk_url = format!("{}/{}", config.zk_url, config.graph_name);
let zk = ZKClient::new(&zk_url, config.zk_timeout_ms, config.get_zk_auth());
let addr = zk.get_coordinator_addr();
let channel =
ChannelBuilder::new(Arc::new(EnvBuilder::new().build())).connect(addr.to_string().as_str());
let client = ServerDataApiClient::new(channel);
let mut request = GetExecutorAliveIdRequest::new();
request.set_serverId(config.worker_id);
request.set_ip(ip::get_local_ip());
let response = client.get_executor_alive_id(&request).unwrap();
let alive_id = response.get_aliveId();
let mut request = GetPartitionAssignmentRequest::new();
request.set_serverId(config.worker_id);
let response = client.get_partition_assignment(&request).unwrap();
let partitions = response.get_partitionId().to_vec();
(alive_id, partitions)
}
pub struct GaiaService<V, VI, E, EI>
where
V: Vertex + 'static,
VI: Iterator<Item = V> + Send + 'static,
E: Edge + 'static,
EI: Iterator<Item = E> + Send + 'static,
{
store_config: Arc<StoreConfig>,
graph: Arc<dyn GlobalGraphQuery<V = V, E = E, VI = VI, EI = EI>>,
partition_manager: Arc<dyn GraphPartitionManager>,
// mapping of partition id -> worker id
partition_worker_mapping: Arc<RwLock<Option<HashMap<u32, u32>>>>,
// mapping of worker id -> partition list | run_main | identifier_name |
conn.go | out.
func (c *Conn) SetReadDeadline(t time.Time) error {
return c.conn.SetReadDeadline(t)
}
// SetWriteDeadline sets the write deadline on the underlying connection.
// A zero value for t means Write will not time out.
// After a Write has timed out, the Noise state is corrupt and all future writes will return the same error.
func (c *Conn) SetWriteDeadline(t time.Time) error {
return c.conn.SetWriteDeadline(t)
}
// Write writes data to the connection.
func (c *Conn) Write(b []byte) (int, error) {
//
if hp := c.config.Pattern; !c.isClient && len(hp.Messages) < 2 {
return 0, errors.New("A server should not write on one-way patterns")
}
// Make sure to go through the handshake first
if err := c.Handshake(); err != nil {
return 0, err
}
// Lock the write socket
c.outLock.Lock()
defer c.outLock.Unlock()
// process the data in a loop
var n int
data := b
for len(data) > 0 {
// fragment the data
m := len(data)
if m > MaxMsgLen {
m = MaxMsgLen
}
// Encrypt
ciphertext := c.out.Encrypt(nil, nil, data[:m])
// header (length)
length := []byte{byte(len(ciphertext) >> 8), byte(len(ciphertext) % 256)}
// Send data
_, err := c.conn.Write(append(length, ciphertext...))
if err != nil {
return n, err
}
n += m
data = data[m:]
}
return n, nil
}
// Read can be made to time out and return a net.Error with Timeout() == true
// after a fixed time limit; see SetDeadline and SetReadDeadline.
func (c *Conn) Read(b []byte) (int, error) {
var err error
// Make sure to go through the handshake first
if err = c.Handshake(); err != nil {
return 0, err
}
// Put this after Handshake, in case people were calling
// Read(nil) for the side effect of the Handshake.
if len(b) == 0 {
return 0, err
}
// If this is a one-way pattern, do some checks
if hp := c.config.Pattern; !c.isClient && len(hp.Messages) < 2 {
return 0, errors.New("A client should not read on one-way patterns")
}
// Lock the read socket
c.inLock.Lock()
defer c.inLock.Unlock()
// read whatever there is to read in the buffer
readSoFar := 0
if len(c.inputBuffer) > 0 {
copy(b, c.inputBuffer)
if len(c.inputBuffer) >= len(b) {
c.inputBuffer = c.inputBuffer[len(b):]
return len(b), nil
}
readSoFar += len(c.inputBuffer)
c.inputBuffer = c.inputBuffer[:0]
}
// read header from socket
bufHeader, err := readBytes(c.conn, 2)
if err != nil {
return 0, err
}
length := (int(bufHeader[0]) << 8) | int(bufHeader[1])
if length > MaxMsgLen {
return 2, errors.New("Noise: Noise message received exceeds NoiseMessageLength")
}
// read noise message from socket
noiseMessage, err := readBytes(c.conn, length)
if err != nil {
return 0, err
}
// decrypt
plaintext, err := c.in.Decrypt(nil, nil, noiseMessage)
if err != nil {
return 0, err
}
// append to the input buffer
c.inputBuffer = append(c.inputBuffer, plaintext...)
// read whatever we can read
rest := len(b) - readSoFar
copy(b[readSoFar:], c.inputBuffer)
if len(c.inputBuffer) >= rest {
c.inputBuffer = c.inputBuffer[rest:]
return len(b), nil
}
// we haven't filled the buffer
readSoFar += len(c.inputBuffer)
c.inputBuffer = c.inputBuffer[:0]
return readSoFar, nil
}
// Close closes the connection.
func (c *Conn) Close() error {
return c.conn.Close()
}
// Noise-related functions
// Handshake runs the client or server handshake protocol if
// it has not yet been run.
// Most uses of this package need not call Handshake explicitly: | // the first Read or Write will call it automatically.
func (c *Conn) Handshake() (err error) {
c.handshakeMutex.Lock()
defer c.handshakeMutex.Unlock()
if c.handshakeComplete {
return nil
}
var remoteKeyPair *DHKey
if c.config.PeerStatic != nil {
if len(c.config.PeerStatic) != 32 {
return errors.New("noise: the provided remote key is not 32-byte")
}
remoteKeyPair = &DHKey{}
copy(remoteKeyPair.Public[:], c.config.PeerStatic)
}
c.hs, err = NewHandshakeState(*c.config)
if err != nil {
return err
}
// start handshake
var c1, c2 *CipherState
var state bool
var msg []byte
state = c.isClient
for _ = range c.config.Pattern.Messages {
if state {
msg, c1, c2, err = c.hs.WriteMessage(nil, nil)
if err != nil {
return err
}
// header (length)
length := []byte{byte(len(msg) >> 8), byte(len(msg) % 256)}
// write
_, err = c.conn.Write(append(length, msg...))
if err != nil {
return err
}
} else {
bufHeader, err := readBytes(c.conn, 2)
if err != nil {
return err
}
length := (int(bufHeader[0]) << 8) | int(bufHeader[1])
if length > MaxMsgLen {
return errors.New("Noise: Noise message received exceeds NoiseMessageLength")
}
msg, err = readBytes(c.conn, length)
if err != nil {
return err
}
_, c1, c2, err = c.hs.ReadMessage(nil, msg)
if err != nil {
return err
}
}
state = !state
}
if c.isClient {
c.out, c.in = c1, c2
} else {
c.out, c.in = c2, c1
}
c.handshakeComplete = true
return nil
}
// IsRemoteAuthenticated can be used to check if the remote peer has been
// properly authenticated. It serves no real purpose for the moment as the
// handshake will not go through if a peer is not properly authenticated in
// patterns where the peer needs to be authenticated.
func (c *Conn) IsRemoteAuthenticated() bool {
return c.isRemoteAuthenticated
}
// RemoteKey returns the static key of the remote peer.
// It is useful in case the static key is only transmitted during the handshake.
func (c *Conn) RemoteKey() ([]byte, error) {
if !c.handshakeComplete {
return nil, errors.New("handshake not completed")
}
return c.hs.rs, nil
}
// Server returns a new Noise server side connection
// using net.Conn as the underlying transport.
// The configuration config must be non-nil and must include
// at least one certificate or else set GetCertificate.
func Server(conn net.Conn, config *Config) *Conn {
return &Conn{conn: conn, config: config, isClient: false}
}
// Client returns a new Noise client side connection
// using conn as the underlying transport.
// The config cannot be nil: users must set either ServerName or
// InsecureSkipVerify in the config.
func Client(conn net.Conn, config *Config) *Conn {
return &Conn{conn: conn, config: config, isClient: true}
}
// A Listener implements a network Listener (net.Listener) for Noise connections.
type Listener struct {
net.Listener
config *Config
}
// Accept waits for and returns the next incoming Noise connection.
// The returned connection is of type *Conn.
func (l *Listener) Accept() (net.Conn, error) {
c, err := l.Listener.Accept()
if err != nil {
return &Conn{}, err
}
return Server(c, l.config), nil
}
// Close closes the listener.
// Any blocked Accept operations will be unblocked and return errors.
func (l *Listener) Close() error {
return l.Listener.Close()
}
// Addr returns the listener's network address.
func (l *Listener) Addr() net.Addr {
return l.Listener.Addr()
}
// Listen creates a Noise Listener accepting connections on the
// given network address using net.Listen.
// The configuration config must be non-nil.
func Listen(network, laddr string, config *Config) (net.Listener, error) {
if config == nil {
return &Listener{}, errors.New("Noise: no Config set")
}
l, err := net.Listen(network, laddr)
if err != nil {
return &Listener{}, err
}
noiseListener := &Listener{}
noiseListener.Listener = l
noise | random_line_split |
|
conn.go | .
func (c *Conn) SetReadDeadline(t time.Time) error {
return c.conn.SetReadDeadline(t)
}
// SetWriteDeadline sets the write deadline on the underlying connection.
// A zero value for t means Write will not time out.
// After a Write has timed out, the Noise state is corrupt and all future writes will return the same error.
func (c *Conn) SetWriteDeadline(t time.Time) error {
return c.conn.SetWriteDeadline(t)
}
// Write writes data to the connection.
func (c *Conn) Write(b []byte) (int, error) {
//
if hp := c.config.Pattern; !c.isClient && len(hp.Messages) < 2 {
return 0, errors.New("A server should not write on one-way patterns")
}
// Make sure to go through the handshake first
if err := c.Handshake(); err != nil {
return 0, err
}
// Lock the write socket
c.outLock.Lock()
defer c.outLock.Unlock()
// process the data in a loop
var n int
data := b
for len(data) > 0 {
// fragment the data
m := len(data)
if m > MaxMsgLen {
m = MaxMsgLen
}
// Encrypt
ciphertext := c.out.Encrypt(nil, nil, data[:m])
// header (length)
length := []byte{byte(len(ciphertext) >> 8), byte(len(ciphertext) % 256)}
// Send data
_, err := c.conn.Write(append(length, ciphertext...))
if err != nil {
return n, err
}
n += m
data = data[m:]
}
return n, nil
}
// Read can be made to time out and return a net.Error with Timeout() == true
// after a fixed time limit; see SetDeadline and SetReadDeadline.
func (c *Conn) Read(b []byte) (int, error) {
var err error
// Make sure to go through the handshake first
if err = c.Handshake(); err != nil {
return 0, err
}
// Put this after Handshake, in case people were calling
// Read(nil) for the side effect of the Handshake.
if len(b) == 0 {
return 0, err
}
// If this is a one-way pattern, do some checks
if hp := c.config.Pattern; !c.isClient && len(hp.Messages) < 2 {
return 0, errors.New("A client should not read on one-way patterns")
}
// Lock the read socket
c.inLock.Lock()
defer c.inLock.Unlock()
// read whatever there is to read in the buffer
readSoFar := 0
if len(c.inputBuffer) > 0 {
copy(b, c.inputBuffer)
if len(c.inputBuffer) >= len(b) {
c.inputBuffer = c.inputBuffer[len(b):]
return len(b), nil
}
readSoFar += len(c.inputBuffer)
c.inputBuffer = c.inputBuffer[:0]
}
// read header from socket
bufHeader, err := readBytes(c.conn, 2)
if err != nil {
return 0, err
}
length := (int(bufHeader[0]) << 8) | int(bufHeader[1])
if length > MaxMsgLen {
return 2, errors.New("Noise: Noise message received exceeds NoiseMessageLength")
}
// read noise message from socket
noiseMessage, err := readBytes(c.conn, length)
if err != nil {
return 0, err
}
// decrypt
plaintext, err := c.in.Decrypt(nil, nil, noiseMessage)
if err != nil {
return 0, err
}
// append to the input buffer
c.inputBuffer = append(c.inputBuffer, plaintext...)
// read whatever we can read
rest := len(b) - readSoFar
copy(b[readSoFar:], c.inputBuffer)
if len(c.inputBuffer) >= rest {
c.inputBuffer = c.inputBuffer[rest:]
return len(b), nil
}
// we haven't filled the buffer
readSoFar += len(c.inputBuffer)
c.inputBuffer = c.inputBuffer[:0]
return readSoFar, nil
}
// Close closes the connection.
func (c *Conn) Close() error {
return c.conn.Close()
}
// Noise-related functions
// Handshake runs the client or server handshake protocol if
// it has not yet been run.
// Most uses of this package need not call Handshake explicitly:
// the first Read or Write will call it automatically.
func (c *Conn) Handshake() (err error) {
c.handshakeMutex.Lock()
defer c.handshakeMutex.Unlock()
if c.handshakeComplete {
return nil
}
var remoteKeyPair *DHKey
if c.config.PeerStatic != nil {
if len(c.config.PeerStatic) != 32 {
return errors.New("noise: the provided remote key is not 32-byte")
}
remoteKeyPair = &DHKey{}
copy(remoteKeyPair.Public[:], c.config.PeerStatic)
}
c.hs, err = NewHandshakeState(*c.config)
if err != nil {
return err
}
// start handshake
var c1, c2 *CipherState
var state bool
var msg []byte
state = c.isClient
for _ = range c.config.Pattern.Messages {
if state {
msg, c1, c2, err = c.hs.WriteMessage(nil, nil)
if err != nil {
return err
}
// header (length)
length := []byte{byte(len(msg) >> 8), byte(len(msg) % 256)}
// write
_, err = c.conn.Write(append(length, msg...))
if err != nil {
return err
}
} else {
bufHeader, err := readBytes(c.conn, 2)
if err != nil {
return err
}
length := (int(bufHeader[0]) << 8) | int(bufHeader[1])
if length > MaxMsgLen {
return errors.New("Noise: Noise message received exceeds NoiseMessageLength")
}
msg, err = readBytes(c.conn, length)
if err != nil {
return err
}
_, c1, c2, err = c.hs.ReadMessage(nil, msg)
if err != nil {
return err
}
}
state = !state
}
if c.isClient {
c.out, c.in = c1, c2
} else {
c.out, c.in = c2, c1
}
c.handshakeComplete = true
return nil
}
// IsRemoteAuthenticated can be used to check if the remote peer has been
// properly authenticated. It serves no real purpose for the moment as the
// handshake will not go through if a peer is not properly authenticated in
// patterns where the peer needs to be authenticated.
func (c *Conn) | () bool {
return c.isRemoteAuthenticated
}
// RemoteKey returns the static key of the remote peer.
// It is useful in case the static key is only transmitted during the handshake.
func (c *Conn) RemoteKey() ([]byte, error) {
if !c.handshakeComplete {
return nil, errors.New("handshake not completed")
}
return c.hs.rs, nil
}
// Server returns a new Noise server side connection
// using net.Conn as the underlying transport.
// The configuration config must be non-nil and must include
// at least one certificate or else set GetCertificate.
func Server(conn net.Conn, config *Config) *Conn {
return &Conn{conn: conn, config: config, isClient: false}
}
// Client returns a new Noise client side connection
// using conn as the underlying transport.
// The config cannot be nil: users must set either ServerName or
// InsecureSkipVerify in the config.
func Client(conn net.Conn, config *Config) *Conn {
return &Conn{conn: conn, config: config, isClient: true}
}
// A Listener implements a network Listener (net.Listener) for Noise connections.
type Listener struct {
net.Listener
config *Config
}
// Accept waits for and returns the next incoming Noise connection.
// The returned connection is of type *Conn.
func (l *Listener) Accept() (net.Conn, error) {
c, err := l.Listener.Accept()
if err != nil {
return &Conn{}, err
}
return Server(c, l.config), nil
}
// Close closes the listener.
// Any blocked Accept operations will be unblocked and return errors.
func (l *Listener) Close() error {
return l.Listener.Close()
}
// Addr returns the listener's network address.
func (l *Listener) Addr() net.Addr {
return l.Listener.Addr()
}
// Listen creates a Noise Listener accepting connections on the
// given network address using net.Listen.
// The configuration config must be non-nil.
func Listen(network, laddr string, config *Config) (net.Listener, error) {
if config == nil {
return &Listener{}, errors.New("Noise: no Config set")
}
l, err := net.Listen(network, laddr)
if err != nil {
return &Listener{}, err
}
noiseListener := &Listener{}
noiseListener.Listener = l
noise | IsRemoteAuthenticated | identifier_name |
conn.go | .
func (c *Conn) SetReadDeadline(t time.Time) error {
return c.conn.SetReadDeadline(t)
}
// SetWriteDeadline sets the write deadline on the underlying connection.
// A zero value for t means Write will not time out.
// After a Write has timed out, the Noise state is corrupt and all future writes will return the same error.
func (c *Conn) SetWriteDeadline(t time.Time) error {
return c.conn.SetWriteDeadline(t)
}
// Write writes data to the connection.
func (c *Conn) Write(b []byte) (int, error) {
//
if hp := c.config.Pattern; !c.isClient && len(hp.Messages) < 2 {
return 0, errors.New("A server should not write on one-way patterns")
}
// Make sure to go through the handshake first
if err := c.Handshake(); err != nil {
return 0, err
}
// Lock the write socket
c.outLock.Lock()
defer c.outLock.Unlock()
// process the data in a loop
var n int
data := b
for len(data) > 0 {
// fragment the data
m := len(data)
if m > MaxMsgLen {
m = MaxMsgLen
}
// Encrypt
ciphertext := c.out.Encrypt(nil, nil, data[:m])
// header (length)
length := []byte{byte(len(ciphertext) >> 8), byte(len(ciphertext) % 256)}
// Send data
_, err := c.conn.Write(append(length, ciphertext...))
if err != nil {
return n, err
}
n += m
data = data[m:]
}
return n, nil
}
// Read can be made to time out and return a net.Error with Timeout() == true
// after a fixed time limit; see SetDeadline and SetReadDeadline.
func (c *Conn) Read(b []byte) (int, error) {
var err error
// Make sure to go through the handshake first
if err = c.Handshake(); err != nil {
return 0, err
}
// Put this after Handshake, in case people were calling
// Read(nil) for the side effect of the Handshake.
if len(b) == 0 {
return 0, err
}
// If this is a one-way pattern, do some checks
if hp := c.config.Pattern; !c.isClient && len(hp.Messages) < 2 {
return 0, errors.New("A client should not read on one-way patterns")
}
// Lock the read socket
c.inLock.Lock()
defer c.inLock.Unlock()
// read whatever there is to read in the buffer
readSoFar := 0
if len(c.inputBuffer) > 0 {
copy(b, c.inputBuffer)
if len(c.inputBuffer) >= len(b) {
c.inputBuffer = c.inputBuffer[len(b):]
return len(b), nil
}
readSoFar += len(c.inputBuffer)
c.inputBuffer = c.inputBuffer[:0]
}
// read header from socket
bufHeader, err := readBytes(c.conn, 2)
if err != nil {
return 0, err
}
length := (int(bufHeader[0]) << 8) | int(bufHeader[1])
if length > MaxMsgLen {
return 2, errors.New("Noise: Noise message received exceeds NoiseMessageLength")
}
// read noise message from socket
noiseMessage, err := readBytes(c.conn, length)
if err != nil {
return 0, err
}
// decrypt
plaintext, err := c.in.Decrypt(nil, nil, noiseMessage)
if err != nil {
return 0, err
}
// append to the input buffer
c.inputBuffer = append(c.inputBuffer, plaintext...)
// read whatever we can read
rest := len(b) - readSoFar
copy(b[readSoFar:], c.inputBuffer)
if len(c.inputBuffer) >= rest {
c.inputBuffer = c.inputBuffer[rest:]
return len(b), nil
}
// we haven't filled the buffer
readSoFar += len(c.inputBuffer)
c.inputBuffer = c.inputBuffer[:0]
return readSoFar, nil
}
// Close closes the connection.
func (c *Conn) Close() error |
// Noise-related functions
// Handshake runs the client or server handshake protocol if
// it has not yet been run.
// Most uses of this package need not call Handshake explicitly:
// the first Read or Write will call it automatically.
func (c *Conn) Handshake() (err error) {
c.handshakeMutex.Lock()
defer c.handshakeMutex.Unlock()
if c.handshakeComplete {
return nil
}
var remoteKeyPair *DHKey
if c.config.PeerStatic != nil {
if len(c.config.PeerStatic) != 32 {
return errors.New("noise: the provided remote key is not 32-byte")
}
remoteKeyPair = &DHKey{}
copy(remoteKeyPair.Public[:], c.config.PeerStatic)
}
c.hs, err = NewHandshakeState(*c.config)
if err != nil {
return err
}
// start handshake
var c1, c2 *CipherState
var state bool
var msg []byte
state = c.isClient
for _ = range c.config.Pattern.Messages {
if state {
msg, c1, c2, err = c.hs.WriteMessage(nil, nil)
if err != nil {
return err
}
// header (length)
length := []byte{byte(len(msg) >> 8), byte(len(msg) % 256)}
// write
_, err = c.conn.Write(append(length, msg...))
if err != nil {
return err
}
} else {
bufHeader, err := readBytes(c.conn, 2)
if err != nil {
return err
}
length := (int(bufHeader[0]) << 8) | int(bufHeader[1])
if length > MaxMsgLen {
return errors.New("Noise: Noise message received exceeds NoiseMessageLength")
}
msg, err = readBytes(c.conn, length)
if err != nil {
return err
}
_, c1, c2, err = c.hs.ReadMessage(nil, msg)
if err != nil {
return err
}
}
state = !state
}
if c.isClient {
c.out, c.in = c1, c2
} else {
c.out, c.in = c2, c1
}
c.handshakeComplete = true
return nil
}
// IsRemoteAuthenticated can be used to check if the remote peer has been
// properly authenticated. It serves no real purpose for the moment as the
// handshake will not go through if a peer is not properly authenticated in
// patterns where the peer needs to be authenticated.
func (c *Conn) IsRemoteAuthenticated() bool {
return c.isRemoteAuthenticated
}
// RemoteKey returns the static key of the remote peer.
// It is useful in case the static key is only transmitted during the handshake.
func (c *Conn) RemoteKey() ([]byte, error) {
if !c.handshakeComplete {
return nil, errors.New("handshake not completed")
}
return c.hs.rs, nil
}
// Server returns a new Noise server side connection
// using net.Conn as the underlying transport.
// The configuration config must be non-nil and must include
// at least one certificate or else set GetCertificate.
func Server(conn net.Conn, config *Config) *Conn {
return &Conn{conn: conn, config: config, isClient: false}
}
// Client returns a new Noise client side connection
// using conn as the underlying transport.
// The config cannot be nil: users must set either ServerName or
// InsecureSkipVerify in the config.
func Client(conn net.Conn, config *Config) *Conn {
return &Conn{conn: conn, config: config, isClient: true}
}
// A Listener implements a network Listener (net.Listener) for Noise connections.
type Listener struct {
net.Listener
config *Config
}
// Accept waits for and returns the next incoming Noise connection.
// The returned connection is of type *Conn.
func (l *Listener) Accept() (net.Conn, error) {
c, err := l.Listener.Accept()
if err != nil {
return &Conn{}, err
}
return Server(c, l.config), nil
}
// Close closes the listener.
// Any blocked Accept operations will be unblocked and return errors.
func (l *Listener) Close() error {
return l.Listener.Close()
}
// Addr returns the listener's network address.
func (l *Listener) Addr() net.Addr {
return l.Listener.Addr()
}
// Listen creates a Noise Listener accepting connections on the
// given network address using net.Listen.
// The configuration config must be non-nil.
func Listen(network, laddr string, config *Config) (net.Listener, error) {
if config == nil {
return &Listener{}, errors.New("Noise: no Config set")
}
l, err := net.Listen(network, laddr)
if err != nil {
return &Listener{}, err
}
noiseListener := &Listener{}
noiseListener.Listener = l
| {
return c.conn.Close()
} | identifier_body |
conn.go | .
func (c *Conn) SetReadDeadline(t time.Time) error {
return c.conn.SetReadDeadline(t)
}
// SetWriteDeadline sets the write deadline on the underlying connection.
// A zero value for t means Write will not time out.
// After a Write has timed out, the Noise state is corrupt and all future writes will return the same error.
func (c *Conn) SetWriteDeadline(t time.Time) error {
return c.conn.SetWriteDeadline(t)
}
// Write writes data to the connection.
func (c *Conn) Write(b []byte) (int, error) {
//
if hp := c.config.Pattern; !c.isClient && len(hp.Messages) < 2 {
return 0, errors.New("A server should not write on one-way patterns")
}
// Make sure to go through the handshake first
if err := c.Handshake(); err != nil {
return 0, err
}
// Lock the write socket
c.outLock.Lock()
defer c.outLock.Unlock()
// process the data in a loop
var n int
data := b
for len(data) > 0 {
// fragment the data
m := len(data)
if m > MaxMsgLen {
m = MaxMsgLen
}
// Encrypt
ciphertext := c.out.Encrypt(nil, nil, data[:m])
// header (length)
length := []byte{byte(len(ciphertext) >> 8), byte(len(ciphertext) % 256)}
// Send data
_, err := c.conn.Write(append(length, ciphertext...))
if err != nil {
return n, err
}
n += m
data = data[m:]
}
return n, nil
}
// Read can be made to time out and return a net.Error with Timeout() == true
// after a fixed time limit; see SetDeadline and SetReadDeadline.
func (c *Conn) Read(b []byte) (int, error) {
var err error
// Make sure to go through the handshake first
if err = c.Handshake(); err != nil {
return 0, err
}
// Put this after Handshake, in case people were calling
// Read(nil) for the side effect of the Handshake.
if len(b) == 0 {
return 0, err
}
// If this is a one-way pattern, do some checks
if hp := c.config.Pattern; !c.isClient && len(hp.Messages) < 2 {
return 0, errors.New("A client should not read on one-way patterns")
}
// Lock the read socket
c.inLock.Lock()
defer c.inLock.Unlock()
// read whatever there is to read in the buffer
readSoFar := 0
if len(c.inputBuffer) > 0 {
copy(b, c.inputBuffer)
if len(c.inputBuffer) >= len(b) {
c.inputBuffer = c.inputBuffer[len(b):]
return len(b), nil
}
readSoFar += len(c.inputBuffer)
c.inputBuffer = c.inputBuffer[:0]
}
// read header from socket
bufHeader, err := readBytes(c.conn, 2)
if err != nil {
return 0, err
}
length := (int(bufHeader[0]) << 8) | int(bufHeader[1])
if length > MaxMsgLen {
return 2, errors.New("Noise: Noise message received exceeds NoiseMessageLength")
}
// read noise message from socket
noiseMessage, err := readBytes(c.conn, length)
if err != nil {
return 0, err
}
// decrypt
plaintext, err := c.in.Decrypt(nil, nil, noiseMessage)
if err != nil {
return 0, err
}
// append to the input buffer
c.inputBuffer = append(c.inputBuffer, plaintext...)
// read whatever we can read
rest := len(b) - readSoFar
copy(b[readSoFar:], c.inputBuffer)
if len(c.inputBuffer) >= rest {
c.inputBuffer = c.inputBuffer[rest:]
return len(b), nil
}
// we haven't filled the buffer
readSoFar += len(c.inputBuffer)
c.inputBuffer = c.inputBuffer[:0]
return readSoFar, nil
}
// Close closes the connection.
func (c *Conn) Close() error {
return c.conn.Close()
}
// Noise-related functions
// Handshake runs the client or server handshake protocol if
// it has not yet been run.
// Most uses of this package need not call Handshake explicitly:
// the first Read or Write will call it automatically.
func (c *Conn) Handshake() (err error) {
c.handshakeMutex.Lock()
defer c.handshakeMutex.Unlock()
if c.handshakeComplete {
return nil
}
var remoteKeyPair *DHKey
if c.config.PeerStatic != nil {
if len(c.config.PeerStatic) != 32 {
return errors.New("noise: the provided remote key is not 32-byte")
}
remoteKeyPair = &DHKey{}
copy(remoteKeyPair.Public[:], c.config.PeerStatic)
}
c.hs, err = NewHandshakeState(*c.config)
if err != nil {
return err
}
// start handshake
var c1, c2 *CipherState
var state bool
var msg []byte
state = c.isClient
for _ = range c.config.Pattern.Messages {
if state {
msg, c1, c2, err = c.hs.WriteMessage(nil, nil)
if err != nil {
return err
}
// header (length)
length := []byte{byte(len(msg) >> 8), byte(len(msg) % 256)}
// write
_, err = c.conn.Write(append(length, msg...))
if err != nil {
return err
}
} else {
bufHeader, err := readBytes(c.conn, 2)
if err != nil {
return err
}
length := (int(bufHeader[0]) << 8) | int(bufHeader[1])
if length > MaxMsgLen {
return errors.New("Noise: Noise message received exceeds NoiseMessageLength")
}
msg, err = readBytes(c.conn, length)
if err != nil |
_, c1, c2, err = c.hs.ReadMessage(nil, msg)
if err != nil {
return err
}
}
state = !state
}
if c.isClient {
c.out, c.in = c1, c2
} else {
c.out, c.in = c2, c1
}
c.handshakeComplete = true
return nil
}
// IsRemoteAuthenticated can be used to check if the remote peer has been
// properly authenticated. It serves no real purpose for the moment as the
// handshake will not go through if a peer is not properly authenticated in
// patterns where the peer needs to be authenticated.
func (c *Conn) IsRemoteAuthenticated() bool {
return c.isRemoteAuthenticated
}
// RemoteKey returns the static key of the remote peer.
// It is useful in case the static key is only transmitted during the handshake.
func (c *Conn) RemoteKey() ([]byte, error) {
if !c.handshakeComplete {
return nil, errors.New("handshake not completed")
}
return c.hs.rs, nil
}
// Server returns a new Noise server side connection
// using net.Conn as the underlying transport.
// The configuration config must be non-nil and must include
// at least one certificate or else set GetCertificate.
func Server(conn net.Conn, config *Config) *Conn {
return &Conn{conn: conn, config: config, isClient: false}
}
// Client returns a new Noise client side connection
// using conn as the underlying transport.
// The config cannot be nil: users must set either ServerName or
// InsecureSkipVerify in the config.
func Client(conn net.Conn, config *Config) *Conn {
return &Conn{conn: conn, config: config, isClient: true}
}
// A Listener implements a network Listener (net.Listener) for Noise connections.
type Listener struct {
net.Listener
config *Config
}
// Accept waits for and returns the next incoming Noise connection.
// The returned connection is of type *Conn.
func (l *Listener) Accept() (net.Conn, error) {
c, err := l.Listener.Accept()
if err != nil {
return &Conn{}, err
}
return Server(c, l.config), nil
}
// Close closes the listener.
// Any blocked Accept operations will be unblocked and return errors.
func (l *Listener) Close() error {
return l.Listener.Close()
}
// Addr returns the listener's network address.
func (l *Listener) Addr() net.Addr {
return l.Listener.Addr()
}
// Listen creates a Noise Listener accepting connections on the
// given network address using net.Listen.
// The configuration config must be non-nil.
func Listen(network, laddr string, config *Config) (net.Listener, error) {
if config == nil {
return &Listener{}, errors.New("Noise: no Config set")
}
l, err := net.Listen(network, laddr)
if err != nil {
return &Listener{}, err
}
noiseListener := &Listener{}
noiseListener.Listener = l
| {
return err
} | conditional_block |
barnacle_vcf.py | +base2
return alt1, alt2
elif breakpoint1.endswith('(down)') and breakpoint2.endswith('(down)'):
alt1 = base1+']'+chrom2+':'+str(pos2)+']'
alt2 = base2+']'+chrom1+':'+str(pos1)+']'
return alt1, alt2
#one breakpoint is at the start of the contig region and other breakpoint is at the end
if breakpoint1.endswith('(up)'):
alt1 = ']'+chrom2+':'+pos2+']'+base1
alt2 = base2+'['+chrom1+':'+pos1+'['
else:
alt1 = base1+'['+chrom2+':'+pos2+'['
alt2 = ']'+chrom1+':'+pos1+']'+base2
return alt1, alt2
def get_QUAL():
return '.'
def get_FILT():
return 'PASS'
def get_INFO(member, id1, id2):
overlap = str(member.meta_fields['ctg_overlap'])
svtype = 'FND'
dp = int(member.avg_read_to_ctg_unique)
if int(overlap) > 0:
return 'SVTYPE='+svtype+';MATEID='+str(id2)+'b;CIPOS=0,'+overlap+';SR='+str(dp)+';CTG=', svtype+';MATEID='+str(id1)+'a;CIPOS=0,'+overlap+';SR='+str(dp)+';CTG='
else:
return 'SVTYPE='+svtype+';MATEID='+str(id2)+'b;SR='+str(dp)+';CTG=', svtype+';MATEID='+str(id1)+'a;SR='+str(dp)+';CTG='
#header output method
def write_header(GIN_user, GIN_pass, LIMS_user, LIMS_pass, refseq_flag, library, filetype_flag, out_file, contig=None):
#file format
out_file.write('##fileformat=VCFv4.1\n')
#file date
out_file.write('##filedate='+time.strftime("%Y%m%d")+'\n')
#tcga version
out_file.write('##tcgaversion=1.0\n')
#genome reference; need to use URL
if refseq_flag == 'hg19':
out_file.write('##reference=<ID=hg19,Source=http://www.bcgsc.ca/downloads/genomes/Homo_sapiens/hg19/1000genomes/bwa_ind/genome/\n')
elif refseq_flag == 'hg18':
out_file.write('##reference=<ID=hg18,Source=http://www.bcgsc.ca/downloads/genomes/Homo_sapiens/hg18/bwa_ind/genome/tcga_ref/>\n')
elif refseq_flag == 'mm9':
out_file.write('##reference=<ID=mm9,Source=/projects/transabyss/trans-ABySS/annotations/mm9/201107/genome.fa>\n')
#contig assembly tags, need to use URL
out_file.write('##assembly='+contig+'\n')
#center
out_file.write('##center="BCGSC"\n')
#phasing
out_file.write('##phasing=none\n')
info_format = {
'svtype':'##INFO=<ID=SVTYPE,Number=1,Type=String,Description="Type of structural variant">\n',
'mateid':'##INFO=<ID=MATEID,Number=1,Type=String,Description="ID of mate breakends">\n',
'event':'##INFO=<ID=EVENT,Number=1,Type=String,Description="ID of breakend event">\n',
'cipos':'##INFO=<ID=CIPOS,Number=2,Type=Integer,Description="Confidence interval around POS for imprecise variants">\n',
'svlen':'##INFO=<ID=SVLEN,Number=1,Type=Integer,Description="Difference in length between REF and ALT alleles">\n',
'end':'##INFO=<ID=END,Number=1,Type=Integer,Description="End position of the variant described in this record">\n',
'inv':'##ALT=<ID=INV,Description="Inversion">\n',
'del':'##ALT=<ID=DEL,Description="Deletion">\n',
'duptan':'##ALT=<ID=DUP:TANDEM,Description="Tandem Duplication">\n',
'sr':'##INFO=<ID=SR,Number=1,Type=Integer,Description="Spanning reads">\n',
'dp':'##INFO=<ID=DP,Number=1,Type=Integer,Description="Read depth">\n',
'CTG':'##INFO=<ID=CTG,Number=.,Type=String,Description="Contig ID">\n'
}
fusion = ['svtype', 'mateid', 'cipos', 'CTG', 'sr']
duplication = ['svtype', 'duptan', 'CTG', 'sr']
if filetype_flag == 'fusion':
for item in fusion:
if item in info_format:
out_file.write(info_format[item])
elif filetype_flag == 'itd' or filetype_flag == 'ptd':
for item in duplication:
if item in info_format:
out_file.write(info_format[item])
sample_info = commands.getoutput("python get_tcga_sample_info.py --username "+GIN_user+" --password "+GIN_pass+" --LIMS_user "+LIMS_user+" --LIMS_pass "+LIMS_pass+" --library "+library)
#sample info
sample_info = sample_info.split(',')
patient = sample_info[0]
sample_id = sample_info[1]
sample_desc = sample_info[2]
platform = sample_info[3]
accession = sample_info[4]
out_file.write('##SAMPLE=<ID='+sample_id+',Individual='+patient+',Description="'+sample_desc+'",Platform='+platform+',Accession='+accession+'>\n')
#pedigree
out_file.write('##PEDIGREE=<Name_0='+sample_id+'>\n')
fields_line = '#CHROM\t'+'POS\t'+'ID\t'+'REF\t'+'ALT\t'+'QUAL\t'+'FILTER\t'+'INFO\n'
out_file.write(fields_line)
def create_fusion_dict(output_dir, in_file, refseq, sequence_dict):
fusion_dict = {}
groupParser = CandidateGroupParserCls(in_file)
id1, id2 = 1, 1
key1, key2 = 1, 2
ctg_dir = output_dir.strip('vcf')
contigfa = open(ctg_dir+'fa', "w")
for group in groupParser:
member = group.members[0]
chrom1, chrom2 = get_CHROM(member)
pos1, pos2 = get_POS(member)
ref1, ref2 = get_REF(member, refseq)
alt1, alt2 = get_ALT(member, refseq)
qual = get_QUAL()
filt = get_FILT()
info1, info2 = get_INFO(member, id1, id2)
fusion1 = chrom1+'\t'+pos1+'\t'+str(id1)+'a'+'\t'+ref1+'\t'+alt1+'\t'+qual+'\t'+filt+'\t'+info1
fusion2 = chrom2+'\t'+pos2+'\t'+str(id2)+'b'+'\t'+ref2+'\t'+alt2+'\t'+qual+'\t'+filt+'\t'+info2
counter = 0
for m in group.members:
contig = m.contig_info.ToString()
contig = contig.replace(':', '_')
contig = contig.partition('(')[0]
if counter > 0:
fusion1 += ','+contig
fusion2 += ','+contig
else:
fusion1 += contig
fusion2 += contig
sequence = sequence_dict['>'+contig]
contigfa.write('>'+contig+'\n'+sequence+'\n')
counter += 1
fusion_dict[key1] = fusion1+'\n'
fusion_dict[key2] = fusion2+'\n'
id1 += 1
id2 += 1
key1 += 2
key2 += 2
return fusion_dict
def dup_CHROM(member):
breakpoint = member.breakpointA.ToString()
breakpoint = breakpoint.partition(':')
chrom = breakpoint[0]
chrom = chrom[3:]
return chrom
def dup_POS(member):
breakpoint = member.breakpointA.ToString()
breakpoint = breakpoint.partition(':')
position = breakpoint[2]
position = position.partition('(')[0]
return position
def | (member, refseq):
pos = dup_POS(member)
chrom = dup_CHROM(member)
seq = refseq.GetSequence(chrom, int(pos), int(pos))
return seq
def dup_ALT():
alt = '<DUP:TANDEM>'
return alt
def dup_INFO(member):
pos = int(dup_POS(member))
svtype = 'FND'
length | dup_REF | identifier_name |
barnacle_vcf.py | str(dp)+';CTG='
#header output method
def write_header(GIN_user, GIN_pass, LIMS_user, LIMS_pass, refseq_flag, library, filetype_flag, out_file, contig=None):
#file format
out_file.write('##fileformat=VCFv4.1\n')
#file date
out_file.write('##filedate='+time.strftime("%Y%m%d")+'\n')
#tcga version
out_file.write('##tcgaversion=1.0\n')
#genome reference; need to use URL
if refseq_flag == 'hg19':
out_file.write('##reference=<ID=hg19,Source=http://www.bcgsc.ca/downloads/genomes/Homo_sapiens/hg19/1000genomes/bwa_ind/genome/\n')
elif refseq_flag == 'hg18':
out_file.write('##reference=<ID=hg18,Source=http://www.bcgsc.ca/downloads/genomes/Homo_sapiens/hg18/bwa_ind/genome/tcga_ref/>\n')
elif refseq_flag == 'mm9':
out_file.write('##reference=<ID=mm9,Source=/projects/transabyss/trans-ABySS/annotations/mm9/201107/genome.fa>\n')
#contig assembly tags, need to use URL
out_file.write('##assembly='+contig+'\n')
#center
out_file.write('##center="BCGSC"\n')
#phasing
out_file.write('##phasing=none\n')
info_format = {
'svtype':'##INFO=<ID=SVTYPE,Number=1,Type=String,Description="Type of structural variant">\n',
'mateid':'##INFO=<ID=MATEID,Number=1,Type=String,Description="ID of mate breakends">\n',
'event':'##INFO=<ID=EVENT,Number=1,Type=String,Description="ID of breakend event">\n',
'cipos':'##INFO=<ID=CIPOS,Number=2,Type=Integer,Description="Confidence interval around POS for imprecise variants">\n',
'svlen':'##INFO=<ID=SVLEN,Number=1,Type=Integer,Description="Difference in length between REF and ALT alleles">\n',
'end':'##INFO=<ID=END,Number=1,Type=Integer,Description="End position of the variant described in this record">\n',
'inv':'##ALT=<ID=INV,Description="Inversion">\n',
'del':'##ALT=<ID=DEL,Description="Deletion">\n',
'duptan':'##ALT=<ID=DUP:TANDEM,Description="Tandem Duplication">\n',
'sr':'##INFO=<ID=SR,Number=1,Type=Integer,Description="Spanning reads">\n',
'dp':'##INFO=<ID=DP,Number=1,Type=Integer,Description="Read depth">\n',
'CTG':'##INFO=<ID=CTG,Number=.,Type=String,Description="Contig ID">\n'
}
fusion = ['svtype', 'mateid', 'cipos', 'CTG', 'sr']
duplication = ['svtype', 'duptan', 'CTG', 'sr']
if filetype_flag == 'fusion':
for item in fusion:
if item in info_format:
out_file.write(info_format[item])
elif filetype_flag == 'itd' or filetype_flag == 'ptd':
for item in duplication:
if item in info_format:
out_file.write(info_format[item])
sample_info = commands.getoutput("python get_tcga_sample_info.py --username "+GIN_user+" --password "+GIN_pass+" --LIMS_user "+LIMS_user+" --LIMS_pass "+LIMS_pass+" --library "+library)
#sample info
sample_info = sample_info.split(',')
patient = sample_info[0]
sample_id = sample_info[1]
sample_desc = sample_info[2]
platform = sample_info[3]
accession = sample_info[4]
out_file.write('##SAMPLE=<ID='+sample_id+',Individual='+patient+',Description="'+sample_desc+'",Platform='+platform+',Accession='+accession+'>\n')
#pedigree
out_file.write('##PEDIGREE=<Name_0='+sample_id+'>\n')
fields_line = '#CHROM\t'+'POS\t'+'ID\t'+'REF\t'+'ALT\t'+'QUAL\t'+'FILTER\t'+'INFO\n'
out_file.write(fields_line)
def create_fusion_dict(output_dir, in_file, refseq, sequence_dict):
fusion_dict = {}
groupParser = CandidateGroupParserCls(in_file)
id1, id2 = 1, 1
key1, key2 = 1, 2
ctg_dir = output_dir.strip('vcf')
contigfa = open(ctg_dir+'fa', "w")
for group in groupParser:
member = group.members[0]
chrom1, chrom2 = get_CHROM(member)
pos1, pos2 = get_POS(member)
ref1, ref2 = get_REF(member, refseq)
alt1, alt2 = get_ALT(member, refseq)
qual = get_QUAL()
filt = get_FILT()
info1, info2 = get_INFO(member, id1, id2)
fusion1 = chrom1+'\t'+pos1+'\t'+str(id1)+'a'+'\t'+ref1+'\t'+alt1+'\t'+qual+'\t'+filt+'\t'+info1
fusion2 = chrom2+'\t'+pos2+'\t'+str(id2)+'b'+'\t'+ref2+'\t'+alt2+'\t'+qual+'\t'+filt+'\t'+info2
counter = 0
for m in group.members:
contig = m.contig_info.ToString()
contig = contig.replace(':', '_')
contig = contig.partition('(')[0]
if counter > 0:
fusion1 += ','+contig
fusion2 += ','+contig
else:
fusion1 += contig
fusion2 += contig
sequence = sequence_dict['>'+contig]
contigfa.write('>'+contig+'\n'+sequence+'\n')
counter += 1
fusion_dict[key1] = fusion1+'\n'
fusion_dict[key2] = fusion2+'\n'
id1 += 1
id2 += 1
key1 += 2
key2 += 2
return fusion_dict
def dup_CHROM(member):
breakpoint = member.breakpointA.ToString()
breakpoint = breakpoint.partition(':')
chrom = breakpoint[0]
chrom = chrom[3:]
return chrom
def dup_POS(member):
breakpoint = member.breakpointA.ToString()
breakpoint = breakpoint.partition(':')
position = breakpoint[2]
position = position.partition('(')[0]
return position
def dup_REF(member, refseq):
pos = dup_POS(member)
chrom = dup_CHROM(member)
seq = refseq.GetSequence(chrom, int(pos), int(pos))
return seq
def dup_ALT():
alt = '<DUP:TANDEM>'
return alt
def dup_INFO(member):
pos = int(dup_POS(member))
svtype = 'FND'
length = len(member.event_seq)
end = pos + length - 1
info = 'SVTYPE='+svtype+';END='+str(end)
return info
def ptd_POS(member):
breakpoint1 = member.breakpointA.ToString()
breakpoint2 = member.breakpointB.ToString()
if breakpoint1.endswith('(down)'):
end = breakpoint2
start = breakpoint1.partition(':')[2]
else:
end = breakpoint1
start = breakpoint2.partition(':')[2]
start = start.partition('(')[0]
return start
def ptd_INFO(member):
breakpoint1 = member.breakpointA.ToString()
breakpoint2 = member.breakpointB.ToString()
if breakpoint1.endswith('(down)'):
end = breakpoint2
start = breakpoint1.partition(':')[2]
else:
end = breakpoint1
start = breakpoint2.partition(':')[2]
start = start.partition('(')[0]
endpos = end.partition(':')[2]
endpos = endpos.partition('(')[0]
svtype = 'FND'
info = 'SVTYPE='+svtype+';END='+endpos
return info
def create_dup_dict(in_file, refseq, filetype_flag):
dup_dict = {}
groupParser = CandidateGroupParserCls(in_file)
id1 = 1
for group in groupParser:
member = group.members[0]
chrom = dup_CHROM(member)
qual = get_QUAL()
filt = get_FILT()
if filetype_flag == 'itd':
pos = dup_POS(member)
else:
pos = ptd_POS(member)
ref = dup_REF(member, refseq)
alt = dup_ALT()
if filetype_flag == 'itd':
info = dup_INFO(member)
else:
| info = ptd_INFO(member) | conditional_block |
|
barnacle_vcf.py | ['+base2
return alt1, alt2
elif breakpoint1.endswith('(down)') and breakpoint2.endswith('(down)'):
alt1 = base1+']'+chrom2+':'+str(pos2)+']'
alt2 = base2+']'+chrom1+':'+str(pos1)+']'
return alt1, alt2
#one breakpoint is at the start of the contig region and other breakpoint is at the end
if breakpoint1.endswith('(up)'):
alt1 = ']'+chrom2+':'+pos2+']'+base1
alt2 = base2+'['+chrom1+':'+pos1+'['
else:
alt1 = base1+'['+chrom2+':'+pos2+'['
alt2 = ']'+chrom1+':'+pos1+']'+base2
return alt1, alt2
def get_QUAL():
return '.'
def get_FILT():
return 'PASS'
def get_INFO(member, id1, id2):
overlap = str(member.meta_fields['ctg_overlap'])
svtype = 'FND'
dp = int(member.avg_read_to_ctg_unique)
if int(overlap) > 0:
return 'SVTYPE='+svtype+';MATEID='+str(id2)+'b;CIPOS=0,'+overlap+';SR='+str(dp)+';CTG=', svtype+';MATEID='+str(id1)+'a;CIPOS=0,'+overlap+';SR='+str(dp)+';CTG='
else:
return 'SVTYPE='+svtype+';MATEID='+str(id2)+'b;SR='+str(dp)+';CTG=', svtype+';MATEID='+str(id1)+'a;SR='+str(dp)+';CTG='
#header output method
def write_header(GIN_user, GIN_pass, LIMS_user, LIMS_pass, refseq_flag, library, filetype_flag, out_file, contig=None):
#file format
out_file.write('##fileformat=VCFv4.1\n')
#file date
out_file.write('##filedate='+time.strftime("%Y%m%d")+'\n')
#tcga version
out_file.write('##tcgaversion=1.0\n')
#genome reference; need to use URL
if refseq_flag == 'hg19':
out_file.write('##reference=<ID=hg19,Source=http://www.bcgsc.ca/downloads/genomes/Homo_sapiens/hg19/1000genomes/bwa_ind/genome/\n')
elif refseq_flag == 'hg18':
out_file.write('##reference=<ID=hg18,Source=http://www.bcgsc.ca/downloads/genomes/Homo_sapiens/hg18/bwa_ind/genome/tcga_ref/>\n')
elif refseq_flag == 'mm9':
out_file.write('##reference=<ID=mm9,Source=/projects/transabyss/trans-ABySS/annotations/mm9/201107/genome.fa>\n')
#contig assembly tags, need to use URL
out_file.write('##assembly='+contig+'\n')
#center
out_file.write('##center="BCGSC"\n')
#phasing
out_file.write('##phasing=none\n')
info_format = {
'svtype':'##INFO=<ID=SVTYPE,Number=1,Type=String,Description="Type of structural variant">\n',
'mateid':'##INFO=<ID=MATEID,Number=1,Type=String,Description="ID of mate breakends">\n',
'event':'##INFO=<ID=EVENT,Number=1,Type=String,Description="ID of breakend event">\n',
'cipos':'##INFO=<ID=CIPOS,Number=2,Type=Integer,Description="Confidence interval around POS for imprecise variants">\n',
'svlen':'##INFO=<ID=SVLEN,Number=1,Type=Integer,Description="Difference in length between REF and ALT alleles">\n',
'end':'##INFO=<ID=END,Number=1,Type=Integer,Description="End position of the variant described in this record">\n',
'inv':'##ALT=<ID=INV,Description="Inversion">\n',
'del':'##ALT=<ID=DEL,Description="Deletion">\n',
'duptan':'##ALT=<ID=DUP:TANDEM,Description="Tandem Duplication">\n',
'sr':'##INFO=<ID=SR,Number=1,Type=Integer,Description="Spanning reads">\n',
'dp':'##INFO=<ID=DP,Number=1,Type=Integer,Description="Read depth">\n',
'CTG':'##INFO=<ID=CTG,Number=.,Type=String,Description="Contig ID">\n'
}
fusion = ['svtype', 'mateid', 'cipos', 'CTG', 'sr']
duplication = ['svtype', 'duptan', 'CTG', 'sr']
if filetype_flag == 'fusion':
for item in fusion:
if item in info_format:
out_file.write(info_format[item])
elif filetype_flag == 'itd' or filetype_flag == 'ptd':
for item in duplication:
if item in info_format:
out_file.write(info_format[item])
sample_info = commands.getoutput("python get_tcga_sample_info.py --username "+GIN_user+" --password "+GIN_pass+" --LIMS_user "+LIMS_user+" --LIMS_pass "+LIMS_pass+" --library "+library)
#sample info
sample_info = sample_info.split(',')
patient = sample_info[0]
sample_id = sample_info[1]
sample_desc = sample_info[2]
platform = sample_info[3]
accession = sample_info[4]
out_file.write('##SAMPLE=<ID='+sample_id+',Individual='+patient+',Description="'+sample_desc+'",Platform='+platform+',Accession='+accession+'>\n')
#pedigree
out_file.write('##PEDIGREE=<Name_0='+sample_id+'>\n')
fields_line = '#CHROM\t'+'POS\t'+'ID\t'+'REF\t'+'ALT\t'+'QUAL\t'+'FILTER\t'+'INFO\n'
out_file.write(fields_line)
def create_fusion_dict(output_dir, in_file, refseq, sequence_dict):
fusion_dict = {}
groupParser = CandidateGroupParserCls(in_file)
id1, id2 = 1, 1
key1, key2 = 1, 2
ctg_dir = output_dir.strip('vcf')
contigfa = open(ctg_dir+'fa', "w")
for group in groupParser:
member = group.members[0]
chrom1, chrom2 = get_CHROM(member)
pos1, pos2 = get_POS(member)
ref1, ref2 = get_REF(member, refseq)
alt1, alt2 = get_ALT(member, refseq)
qual = get_QUAL()
filt = get_FILT()
info1, info2 = get_INFO(member, id1, id2)
fusion1 = chrom1+'\t'+pos1+'\t'+str(id1)+'a'+'\t'+ref1+'\t'+alt1+'\t'+qual+'\t'+filt+'\t'+info1
fusion2 = chrom2+'\t'+pos2+'\t'+str(id2)+'b'+'\t'+ref2+'\t'+alt2+'\t'+qual+'\t'+filt+'\t'+info2
counter = 0
for m in group.members:
contig = m.contig_info.ToString()
contig = contig.replace(':', '_')
contig = contig.partition('(')[0]
if counter > 0:
fusion1 += ','+contig
fusion2 += ','+contig
else:
fusion1 += contig
fusion2 += contig
sequence = sequence_dict['>'+contig]
contigfa.write('>'+contig+'\n'+sequence+'\n')
counter += 1
fusion_dict[key1] = fusion1+'\n'
fusion_dict[key2] = fusion2+'\n'
id1 += 1
id2 += 1
key1 += 2
key2 += 2
return fusion_dict
def dup_CHROM(member):
breakpoint = member.breakpointA.ToString()
breakpoint = breakpoint.partition(':')
chrom = breakpoint[0]
chrom = chrom[3:]
return chrom
def dup_POS(member):
breakpoint = member.breakpointA.ToString()
breakpoint = breakpoint.partition(':')
position = breakpoint[2]
position = position.partition('(')[0]
return position | seq = refseq.GetSequence(chrom, int(pos), int(pos))
return seq
def dup_ALT():
alt = '<DUP:TANDEM>'
return alt
def dup_INFO(member):
pos = int(dup_POS(member))
svtype = 'FND'
length |
def dup_REF(member, refseq):
pos = dup_POS(member)
chrom = dup_CHROM(member) | random_line_split |
barnacle_vcf.py | ['+base2
return alt1, alt2
elif breakpoint1.endswith('(down)') and breakpoint2.endswith('(down)'):
alt1 = base1+']'+chrom2+':'+str(pos2)+']'
alt2 = base2+']'+chrom1+':'+str(pos1)+']'
return alt1, alt2
#one breakpoint is at the start of the contig region and other breakpoint is at the end
if breakpoint1.endswith('(up)'):
alt1 = ']'+chrom2+':'+pos2+']'+base1
alt2 = base2+'['+chrom1+':'+pos1+'['
else:
alt1 = base1+'['+chrom2+':'+pos2+'['
alt2 = ']'+chrom1+':'+pos1+']'+base2
return alt1, alt2
def get_QUAL():
return '.'
def get_FILT():
return 'PASS'
def get_INFO(member, id1, id2):
|
#header output method
def write_header(GIN_user, GIN_pass, LIMS_user, LIMS_pass, refseq_flag, library, filetype_flag, out_file, contig=None):
#file format
out_file.write('##fileformat=VCFv4.1\n')
#file date
out_file.write('##filedate='+time.strftime("%Y%m%d")+'\n')
#tcga version
out_file.write('##tcgaversion=1.0\n')
#genome reference; need to use URL
if refseq_flag == 'hg19':
out_file.write('##reference=<ID=hg19,Source=http://www.bcgsc.ca/downloads/genomes/Homo_sapiens/hg19/1000genomes/bwa_ind/genome/\n')
elif refseq_flag == 'hg18':
out_file.write('##reference=<ID=hg18,Source=http://www.bcgsc.ca/downloads/genomes/Homo_sapiens/hg18/bwa_ind/genome/tcga_ref/>\n')
elif refseq_flag == 'mm9':
out_file.write('##reference=<ID=mm9,Source=/projects/transabyss/trans-ABySS/annotations/mm9/201107/genome.fa>\n')
#contig assembly tags, need to use URL
out_file.write('##assembly='+contig+'\n')
#center
out_file.write('##center="BCGSC"\n')
#phasing
out_file.write('##phasing=none\n')
info_format = {
'svtype':'##INFO=<ID=SVTYPE,Number=1,Type=String,Description="Type of structural variant">\n',
'mateid':'##INFO=<ID=MATEID,Number=1,Type=String,Description="ID of mate breakends">\n',
'event':'##INFO=<ID=EVENT,Number=1,Type=String,Description="ID of breakend event">\n',
'cipos':'##INFO=<ID=CIPOS,Number=2,Type=Integer,Description="Confidence interval around POS for imprecise variants">\n',
'svlen':'##INFO=<ID=SVLEN,Number=1,Type=Integer,Description="Difference in length between REF and ALT alleles">\n',
'end':'##INFO=<ID=END,Number=1,Type=Integer,Description="End position of the variant described in this record">\n',
'inv':'##ALT=<ID=INV,Description="Inversion">\n',
'del':'##ALT=<ID=DEL,Description="Deletion">\n',
'duptan':'##ALT=<ID=DUP:TANDEM,Description="Tandem Duplication">\n',
'sr':'##INFO=<ID=SR,Number=1,Type=Integer,Description="Spanning reads">\n',
'dp':'##INFO=<ID=DP,Number=1,Type=Integer,Description="Read depth">\n',
'CTG':'##INFO=<ID=CTG,Number=.,Type=String,Description="Contig ID">\n'
}
fusion = ['svtype', 'mateid', 'cipos', 'CTG', 'sr']
duplication = ['svtype', 'duptan', 'CTG', 'sr']
if filetype_flag == 'fusion':
for item in fusion:
if item in info_format:
out_file.write(info_format[item])
elif filetype_flag == 'itd' or filetype_flag == 'ptd':
for item in duplication:
if item in info_format:
out_file.write(info_format[item])
sample_info = commands.getoutput("python get_tcga_sample_info.py --username "+GIN_user+" --password "+GIN_pass+" --LIMS_user "+LIMS_user+" --LIMS_pass "+LIMS_pass+" --library "+library)
#sample info
sample_info = sample_info.split(',')
patient = sample_info[0]
sample_id = sample_info[1]
sample_desc = sample_info[2]
platform = sample_info[3]
accession = sample_info[4]
out_file.write('##SAMPLE=<ID='+sample_id+',Individual='+patient+',Description="'+sample_desc+'",Platform='+platform+',Accession='+accession+'>\n')
#pedigree
out_file.write('##PEDIGREE=<Name_0='+sample_id+'>\n')
fields_line = '#CHROM\t'+'POS\t'+'ID\t'+'REF\t'+'ALT\t'+'QUAL\t'+'FILTER\t'+'INFO\n'
out_file.write(fields_line)
def create_fusion_dict(output_dir, in_file, refseq, sequence_dict):
fusion_dict = {}
groupParser = CandidateGroupParserCls(in_file)
id1, id2 = 1, 1
key1, key2 = 1, 2
ctg_dir = output_dir.strip('vcf')
contigfa = open(ctg_dir+'fa', "w")
for group in groupParser:
member = group.members[0]
chrom1, chrom2 = get_CHROM(member)
pos1, pos2 = get_POS(member)
ref1, ref2 = get_REF(member, refseq)
alt1, alt2 = get_ALT(member, refseq)
qual = get_QUAL()
filt = get_FILT()
info1, info2 = get_INFO(member, id1, id2)
fusion1 = chrom1+'\t'+pos1+'\t'+str(id1)+'a'+'\t'+ref1+'\t'+alt1+'\t'+qual+'\t'+filt+'\t'+info1
fusion2 = chrom2+'\t'+pos2+'\t'+str(id2)+'b'+'\t'+ref2+'\t'+alt2+'\t'+qual+'\t'+filt+'\t'+info2
counter = 0
for m in group.members:
contig = m.contig_info.ToString()
contig = contig.replace(':', '_')
contig = contig.partition('(')[0]
if counter > 0:
fusion1 += ','+contig
fusion2 += ','+contig
else:
fusion1 += contig
fusion2 += contig
sequence = sequence_dict['>'+contig]
contigfa.write('>'+contig+'\n'+sequence+'\n')
counter += 1
fusion_dict[key1] = fusion1+'\n'
fusion_dict[key2] = fusion2+'\n'
id1 += 1
id2 += 1
key1 += 2
key2 += 2
return fusion_dict
def dup_CHROM(member):
breakpoint = member.breakpointA.ToString()
breakpoint = breakpoint.partition(':')
chrom = breakpoint[0]
chrom = chrom[3:]
return chrom
def dup_POS(member):
breakpoint = member.breakpointA.ToString()
breakpoint = breakpoint.partition(':')
position = breakpoint[2]
position = position.partition('(')[0]
return position
def dup_REF(member, refseq):
pos = dup_POS(member)
chrom = dup_CHROM(member)
seq = refseq.GetSequence(chrom, int(pos), int(pos))
return seq
def dup_ALT():
alt = '<DUP:TANDEM>'
return alt
def dup_INFO(member):
pos = int(dup_POS(member))
svtype = 'FND'
length | overlap = str(member.meta_fields['ctg_overlap'])
svtype = 'FND'
dp = int(member.avg_read_to_ctg_unique)
if int(overlap) > 0:
return 'SVTYPE='+svtype+';MATEID='+str(id2)+'b;CIPOS=0,'+overlap+';SR='+str(dp)+';CTG=', svtype+';MATEID='+str(id1)+'a;CIPOS=0,'+overlap+';SR='+str(dp)+';CTG='
else:
return 'SVTYPE='+svtype+';MATEID='+str(id2)+'b;SR='+str(dp)+';CTG=', svtype+';MATEID='+str(id1)+'a;SR='+str(dp)+';CTG=' | identifier_body |
MakeGoodGamesBot.py | tit_for_tat(),
# 2) RT followers,
# 3) RT highest ratio of number_of_RT/number_of_followers of previous day Statuses.
def run_schedule(dt=get_date(),ky='#indiedev',mx=150,clean=False,folow=False):
if clean: tit_for_tat()
if folow: RT_followers(key_=ky,max_=mx)
RT_last_day(dt,key_=ky)
def twice():
run_schedule(folow=True)
run_schedule(ky='#indiegame',clean=True,folow=True)
def loop_schedule(date):
while True:
for ky in ['#indiedev','#indiegame']:
print 'Day '+str(date)+' and keyword '+str(ky)
run_schedule(dt=date,ky=ky)
d = get_date()
if date != d:
date = d
break
#Main Functions for 'run_schedule()'
#Keeps track of who doesn't follow back. If an ID appears twice in this category then it
#unfollows. Bot follows ALL followers. Filter for undesirable accounts will be implemented soon.
def | ():
print 'Tit for Tat!'
follow_me = twitter_client.followers_ids() #who follows me
follow_you = twitter_client.friends_ids() #who do I follow
erros_ids = []
fol_len = len([1 for id_ in follow_me if id_ not in follow_you])
print 'Following '+str(fol_len)+' new users.'
for id_ in follow_me:
if id_ not in follow_you:
try:
twitter_client.create_friendship(id_)
time.sleep(5)
except:
erros_ids.append(id_)
unfollow,rem_len = remember_follow()
print 'Unfollowing '+str(len(unfollow))+'. Remembering '+str(rem_len)+'.'
for id_ in follow_you:
if id_ in unfollow:
try:
twitter_client.destroy_friendship(id_)
time.sleep(5)
except:
erros_ids.append(id_)
#Take previous day tweets. Rank by higher RT of smaller accounts. Try to RT the underdog!
def RT_last_day(date,key_='#indiedev'):
print 'RT '+str(key_)+' most relevant tweets from yesterday!'
d = latest_tweets(date=date,key_=key_)
d = rank_sort(d)
plot_distro(d[:5000])
a = RT_this(d[:10000],sleep_t=180)
return a
#Take timelines from followers and looks for keyword (default: #indiedev. RTs top tweets (default=2).
def RT_followers(key_='#indiedev',max_=150,rts_=2): #900/1500 Rate limit
print 'RT '+str(key_)+' within followers!'
clct,twtn = [],0
friends = twitter_client.followers_ids()
#Get collection of tweets of followers.
for f in friends:
c=[]
try:
c = twitter_client.user_timeline(f,count=100)
except:
#print 'Cant retweet follower.'
pass
tcl = [ci for ci in c if '#indiedev' in ci.text and ci.in_reply_to_status_id == None]
ttcl = []
for t in tcl:
keep = True
for word in banwords:
if word in t.entities['hashtags']:
keep = False
if keep == True:
ttcl.append(t)
tcl = ttcl
if len(tcl) > 0:
dc = {str(i.id):int(i.retweet_count) for i in tcl}
dfc = pd.DataFrame.from_dict(dc,orient='index')
dfc = dfc.sort(0,ascending=False)
#Final collection of RTs considers only top statuses
clct = clct + list(dfc.index[:rts_])
#After selection of most desirable RTs, we randomly RT them.
rd.shuffle(clct)
print 'Going for '+str(len(clct[:max_]))+' tweets.'
for id_ in clct:
if twtn >= max_: break
try:
twtn+=1
twitter_client.retweet(id_)
print 'Tweeted '+str(twtn)
time.sleep(120)
twitter_client.create_favorite(id_)
except:
pass
#RT statuses from a given DataFrame d.
def RT_this(d,sleep_t=60,stop_at=500,allow_like=False,checkmedia=False):
err,twts,iters=0,0,0
for tweet in d.values:
if twts == stop_at:
print 'Got '+str(stop_at)+' tweets. Stopping.'
break
try:
like,publish,i = True,True,0
if type(tweet[11]) != None:
try:
if tweet[11] in banRT: like,publish = False,False
except:
pass
i=1
if len(tweet[27]) > 14:
if tweet[27][:15] in banTXT: like,publish = False,False
i=2
like = False
if like and rd.random() < 0.05:
twitter_client.create_favorite(tweet[8])
print 'Liked '+tweet[27]
if publish == False:
time.sleep(60)
i=3
if checkmedia and publish:
publish = filter_gif(tweet)
if publish:
try:
twitter_client.retweet(tweet[8])
i,twts=4,twts+1
print 'RTed : '+str(twts)+' at '+str(time.ctime())
time.sleep(sleep_t)
i=5
if type(tweet[11]) != None:
banRT.append(tweet[11])
if len(tweet[27]) > 14:
banTXT.append(tweet[27][:15])
except:
pass
try:
u_fol,u_fri=tweet[29]['followers_count'],tweet[29]['friends_count']
if (u_fol > 500 and u_fol < 10000) or (u_fol > 1.5*u_fri):
if i==4:
time.sleep(sleep_t)
i=6
twitter_client.create_friendship(tweet[29]['id'])
except:
pass
if like and allow_like:
try:
twitter_client.create_favorite(tweet[8])
print 'Liked '+tweet[27]
time.sleep(sleep_t/3)
except:
print 'Couldnt like'
save_banDF()
iters+=1
except:
err+=1
#StreamListener: DoggoListener is not currently on use as RTs of previous day and followers
#use most of the bot's daytime.
#List of words to avoid and possible tweet lexic
banword = ["porn","pron","p0rn","pr0n"]
doggolex = ['*doggehyped*']
class DoggoListener(StreamListener):
def on_data(self, data):
tweet = json.loads(data)
i,like,publish = 0,True,True
try:
for word in banword:
if word in tweet['text'].lower(): like,publish = False,False
i=1
if tweet.get('lang') and tweet.get('lang') != 'en': like,publish = False,False
i=2
try:
if type(tweet['user']['description']) != None:
if 'indie' not in tweet['user']['description'] or 'dev' not in tweet['user']['description'] or 'developer' not in tweet['user']['description']:
like = False
if tweet['user']['followers_count'] < 1000: publish = False
else:
like,publish = False,False
except:
like,publish = False,False
i=3
if type(tweet['in_reply_to_status_id']) != None:
if tweet['in_reply_to_status_id'] in banRT:
like,publish = False,False
i=4
if len(tweet['text']) > 14:
if tweet['text'][:15] in banTXT:
like,publish = False,False
i=5
if like:
twitter_client.create_favorite(tweet['id'])
print 'Liked '+tweet['text']
if publish == False:
time.sleep(10)
i=6
if publish:
twitter_client.retweet(tweet['id'])
#Some console output to check if stream is RTweeting.
try:
print 'RTd: '+str(tweet['text'])
except:
print '*Woof*'
i='t'
if type(tweet['in_reply_to_status_id']) != None:
i=7
banRT.append(tweet['in_reply_to_status_id'])
if len(tweet['text']) > 14:
i=8
banTXT.append(tweet['text'][:15])
save_banDF()
time.sleep(60)
except:
print i #For debugging purposes
return True
def run_doggo_run(): #Streams the doggolistener()
if | tit_for_tat | identifier_name |
MakeGoodGamesBot.py | tit_for_tat(),
# 2) RT followers,
# 3) RT highest ratio of number_of_RT/number_of_followers of previous day Statuses.
def run_schedule(dt=get_date(),ky='#indiedev',mx=150,clean=False,folow=False):
if clean: tit_for_tat()
if folow: RT_followers(key_=ky,max_=mx)
RT_last_day(dt,key_=ky)
def twice():
|
def loop_schedule(date):
while True:
for ky in ['#indiedev','#indiegame']:
print 'Day '+str(date)+' and keyword '+str(ky)
run_schedule(dt=date,ky=ky)
d = get_date()
if date != d:
date = d
break
#Main Functions for 'run_schedule()'
#Keeps track of who doesn't follow back. If an ID appears twice in this category then it
#unfollows. Bot follows ALL followers. Filter for undesirable accounts will be implemented soon.
def tit_for_tat():
print 'Tit for Tat!'
follow_me = twitter_client.followers_ids() #who follows me
follow_you = twitter_client.friends_ids() #who do I follow
erros_ids = []
fol_len = len([1 for id_ in follow_me if id_ not in follow_you])
print 'Following '+str(fol_len)+' new users.'
for id_ in follow_me:
if id_ not in follow_you:
try:
twitter_client.create_friendship(id_)
time.sleep(5)
except:
erros_ids.append(id_)
unfollow,rem_len = remember_follow()
print 'Unfollowing '+str(len(unfollow))+'. Remembering '+str(rem_len)+'.'
for id_ in follow_you:
if id_ in unfollow:
try:
twitter_client.destroy_friendship(id_)
time.sleep(5)
except:
erros_ids.append(id_)
#Take previous day tweets. Rank by higher RT of smaller accounts. Try to RT the underdog!
def RT_last_day(date,key_='#indiedev'):
print 'RT '+str(key_)+' most relevant tweets from yesterday!'
d = latest_tweets(date=date,key_=key_)
d = rank_sort(d)
plot_distro(d[:5000])
a = RT_this(d[:10000],sleep_t=180)
return a
#Take timelines from followers and looks for keyword (default: #indiedev. RTs top tweets (default=2).
def RT_followers(key_='#indiedev',max_=150,rts_=2): #900/1500 Rate limit
print 'RT '+str(key_)+' within followers!'
clct,twtn = [],0
friends = twitter_client.followers_ids()
#Get collection of tweets of followers.
for f in friends:
c=[]
try:
c = twitter_client.user_timeline(f,count=100)
except:
#print 'Cant retweet follower.'
pass
tcl = [ci for ci in c if '#indiedev' in ci.text and ci.in_reply_to_status_id == None]
ttcl = []
for t in tcl:
keep = True
for word in banwords:
if word in t.entities['hashtags']:
keep = False
if keep == True:
ttcl.append(t)
tcl = ttcl
if len(tcl) > 0:
dc = {str(i.id):int(i.retweet_count) for i in tcl}
dfc = pd.DataFrame.from_dict(dc,orient='index')
dfc = dfc.sort(0,ascending=False)
#Final collection of RTs considers only top statuses
clct = clct + list(dfc.index[:rts_])
#After selection of most desirable RTs, we randomly RT them.
rd.shuffle(clct)
print 'Going for '+str(len(clct[:max_]))+' tweets.'
for id_ in clct:
if twtn >= max_: break
try:
twtn+=1
twitter_client.retweet(id_)
print 'Tweeted '+str(twtn)
time.sleep(120)
twitter_client.create_favorite(id_)
except:
pass
#RT statuses from a given DataFrame d.
def RT_this(d,sleep_t=60,stop_at=500,allow_like=False,checkmedia=False):
err,twts,iters=0,0,0
for tweet in d.values:
if twts == stop_at:
print 'Got '+str(stop_at)+' tweets. Stopping.'
break
try:
like,publish,i = True,True,0
if type(tweet[11]) != None:
try:
if tweet[11] in banRT: like,publish = False,False
except:
pass
i=1
if len(tweet[27]) > 14:
if tweet[27][:15] in banTXT: like,publish = False,False
i=2
like = False
if like and rd.random() < 0.05:
twitter_client.create_favorite(tweet[8])
print 'Liked '+tweet[27]
if publish == False:
time.sleep(60)
i=3
if checkmedia and publish:
publish = filter_gif(tweet)
if publish:
try:
twitter_client.retweet(tweet[8])
i,twts=4,twts+1
print 'RTed : '+str(twts)+' at '+str(time.ctime())
time.sleep(sleep_t)
i=5
if type(tweet[11]) != None:
banRT.append(tweet[11])
if len(tweet[27]) > 14:
banTXT.append(tweet[27][:15])
except:
pass
try:
u_fol,u_fri=tweet[29]['followers_count'],tweet[29]['friends_count']
if (u_fol > 500 and u_fol < 10000) or (u_fol > 1.5*u_fri):
if i==4:
time.sleep(sleep_t)
i=6
twitter_client.create_friendship(tweet[29]['id'])
except:
pass
if like and allow_like:
try:
twitter_client.create_favorite(tweet[8])
print 'Liked '+tweet[27]
time.sleep(sleep_t/3)
except:
print 'Couldnt like'
save_banDF()
iters+=1
except:
err+=1
#StreamListener: DoggoListener is not currently on use as RTs of previous day and followers
#use most of the bot's daytime.
#List of words to avoid and possible tweet lexic
banword = ["porn","pron","p0rn","pr0n"]
doggolex = ['*doggehyped*']
class DoggoListener(StreamListener):
def on_data(self, data):
tweet = json.loads(data)
i,like,publish = 0,True,True
try:
for word in banword:
if word in tweet['text'].lower(): like,publish = False,False
i=1
if tweet.get('lang') and tweet.get('lang') != 'en': like,publish = False,False
i=2
try:
if type(tweet['user']['description']) != None:
if 'indie' not in tweet['user']['description'] or 'dev' not in tweet['user']['description'] or 'developer' not in tweet['user']['description']:
like = False
if tweet['user']['followers_count'] < 1000: publish = False
else:
like,publish = False,False
except:
like,publish = False,False
i=3
if type(tweet['in_reply_to_status_id']) != None:
if tweet['in_reply_to_status_id'] in banRT:
like,publish = False,False
i=4
if len(tweet['text']) > 14:
if tweet['text'][:15] in banTXT:
like,publish = False,False
i=5
if like:
twitter_client.create_favorite(tweet['id'])
print 'Liked '+tweet['text']
if publish == False:
time.sleep(10)
i=6
if publish:
twitter_client.retweet(tweet['id'])
#Some console output to check if stream is RTweeting.
try:
print 'RTd: '+str(tweet['text'])
except:
print '*Woof*'
i='t'
if type(tweet['in_reply_to_status_id']) != None:
i=7
banRT.append(tweet['in_reply_to_status_id'])
if len(tweet['text']) > 14:
i=8
banTXT.append(tweet['text'][:15])
save_banDF()
time.sleep(60)
except:
print i #For debugging purposes
return True
def run_doggo_run(): #Streams the doggolistener()
if __ | run_schedule(folow=True)
run_schedule(ky='#indiegame',clean=True,folow=True) | identifier_body |
MakeGoodGamesBot.py | tit_for_tat(),
# 2) RT followers,
# 3) RT highest ratio of number_of_RT/number_of_followers of previous day Statuses.
def run_schedule(dt=get_date(),ky='#indiedev',mx=150,clean=False,folow=False):
if clean: |
if folow: RT_followers(key_=ky,max_=mx)
RT_last_day(dt,key_=ky)
def twice():
run_schedule(folow=True)
run_schedule(ky='#indiegame',clean=True,folow=True)
def loop_schedule(date):
while True:
for ky in ['#indiedev','#indiegame']:
print 'Day '+str(date)+' and keyword '+str(ky)
run_schedule(dt=date,ky=ky)
d = get_date()
if date != d:
date = d
break
#Main Functions for 'run_schedule()'
#Keeps track of who doesn't follow back. If an ID appears twice in this category then it
#unfollows. Bot follows ALL followers. Filter for undesirable accounts will be implemented soon.
def tit_for_tat():
print 'Tit for Tat!'
follow_me = twitter_client.followers_ids() #who follows me
follow_you = twitter_client.friends_ids() #who do I follow
erros_ids = []
fol_len = len([1 for id_ in follow_me if id_ not in follow_you])
print 'Following '+str(fol_len)+' new users.'
for id_ in follow_me:
if id_ not in follow_you:
try:
twitter_client.create_friendship(id_)
time.sleep(5)
except:
erros_ids.append(id_)
unfollow,rem_len = remember_follow()
print 'Unfollowing '+str(len(unfollow))+'. Remembering '+str(rem_len)+'.'
for id_ in follow_you:
if id_ in unfollow:
try:
twitter_client.destroy_friendship(id_)
time.sleep(5)
except:
erros_ids.append(id_)
#Take previous day tweets. Rank by higher RT of smaller accounts. Try to RT the underdog!
def RT_last_day(date,key_='#indiedev'):
print 'RT '+str(key_)+' most relevant tweets from yesterday!'
d = latest_tweets(date=date,key_=key_)
d = rank_sort(d)
plot_distro(d[:5000])
a = RT_this(d[:10000],sleep_t=180)
return a
#Take timelines from followers and looks for keyword (default: #indiedev. RTs top tweets (default=2).
def RT_followers(key_='#indiedev',max_=150,rts_=2): #900/1500 Rate limit
print 'RT '+str(key_)+' within followers!'
clct,twtn = [],0
friends = twitter_client.followers_ids()
#Get collection of tweets of followers.
for f in friends:
c=[]
try:
c = twitter_client.user_timeline(f,count=100)
except:
#print 'Cant retweet follower.'
pass
tcl = [ci for ci in c if '#indiedev' in ci.text and ci.in_reply_to_status_id == None]
ttcl = []
for t in tcl:
keep = True
for word in banwords:
if word in t.entities['hashtags']:
keep = False
if keep == True:
ttcl.append(t)
tcl = ttcl
if len(tcl) > 0:
dc = {str(i.id):int(i.retweet_count) for i in tcl}
dfc = pd.DataFrame.from_dict(dc,orient='index')
dfc = dfc.sort(0,ascending=False)
#Final collection of RTs considers only top statuses
clct = clct + list(dfc.index[:rts_])
#After selection of most desirable RTs, we randomly RT them.
rd.shuffle(clct)
print 'Going for '+str(len(clct[:max_]))+' tweets.'
for id_ in clct:
if twtn >= max_: break
try:
twtn+=1
twitter_client.retweet(id_)
print 'Tweeted '+str(twtn)
time.sleep(120)
twitter_client.create_favorite(id_)
except:
pass
#RT statuses from a given DataFrame d.
def RT_this(d,sleep_t=60,stop_at=500,allow_like=False,checkmedia=False):
err,twts,iters=0,0,0
for tweet in d.values:
if twts == stop_at:
print 'Got '+str(stop_at)+' tweets. Stopping.'
break
try:
like,publish,i = True,True,0
if type(tweet[11]) != None:
try:
if tweet[11] in banRT: like,publish = False,False
except:
pass
i=1
if len(tweet[27]) > 14:
if tweet[27][:15] in banTXT: like,publish = False,False
i=2
like = False
if like and rd.random() < 0.05:
twitter_client.create_favorite(tweet[8])
print 'Liked '+tweet[27]
if publish == False:
time.sleep(60)
i=3
if checkmedia and publish:
publish = filter_gif(tweet)
if publish:
try:
twitter_client.retweet(tweet[8])
i,twts=4,twts+1
print 'RTed : '+str(twts)+' at '+str(time.ctime())
time.sleep(sleep_t)
i=5
if type(tweet[11]) != None:
banRT.append(tweet[11])
if len(tweet[27]) > 14:
banTXT.append(tweet[27][:15])
except:
pass
try:
u_fol,u_fri=tweet[29]['followers_count'],tweet[29]['friends_count']
if (u_fol > 500 and u_fol < 10000) or (u_fol > 1.5*u_fri):
if i==4:
time.sleep(sleep_t)
i=6
twitter_client.create_friendship(tweet[29]['id'])
except:
pass
if like and allow_like:
try:
twitter_client.create_favorite(tweet[8])
print 'Liked '+tweet[27]
time.sleep(sleep_t/3)
except:
print 'Couldnt like'
save_banDF()
iters+=1
except:
err+=1
#StreamListener: DoggoListener is not currently on use as RTs of previous day and followers
#use most of the bot's daytime.
#List of words to avoid and possible tweet lexic
banword = ["porn","pron","p0rn","pr0n"]
doggolex = ['*doggehyped*']
class DoggoListener(StreamListener):
def on_data(self, data):
tweet = json.loads(data)
i,like,publish = 0,True,True
try:
for word in banword:
if word in tweet['text'].lower(): like,publish = False,False
i=1
if tweet.get('lang') and tweet.get('lang') != 'en': like,publish = False,False
i=2
try:
if type(tweet['user']['description']) != None:
if 'indie' not in tweet['user']['description'] or 'dev' not in tweet['user']['description'] or 'developer' not in tweet['user']['description']:
like = False
if tweet['user']['followers_count'] < 1000: publish = False
else:
like,publish = False,False
except:
like,publish = False,False
i=3
if type(tweet['in_reply_to_status_id']) != None:
if tweet['in_reply_to_status_id'] in banRT:
like,publish = False,False
i=4
if len(tweet['text']) > 14:
if tweet['text'][:15] in banTXT:
like,publish = False,False
i=5
if like:
twitter_client.create_favorite(tweet['id'])
print 'Liked '+tweet['text']
if publish == False:
time.sleep(10)
i=6
if publish:
twitter_client.retweet(tweet['id'])
#Some console output to check if stream is RTweeting.
try:
print 'RTd: '+str(tweet['text'])
except:
print '*Woof*'
i='t'
if type(tweet['in_reply_to_status_id']) != None:
i=7
banRT.append(tweet['in_reply_to_status_id'])
if len(tweet['text']) > 14:
i=8
banTXT.append(tweet['text'][:15])
save_banDF()
time.sleep(60)
except:
print i #For debugging purposes
return True
def run_doggo_run(): #Streams the doggolistener()
| tit_for_tat() | conditional_block |
MakeGoodGamesBot.py | tit_for_tat(),
# 2) RT followers,
# 3) RT highest ratio of number_of_RT/number_of_followers of previous day Statuses.
def run_schedule(dt=get_date(),ky='#indiedev',mx=150,clean=False,folow=False):
if clean: tit_for_tat()
if folow: RT_followers(key_=ky,max_=mx)
RT_last_day(dt,key_=ky)
def twice():
run_schedule(folow=True)
run_schedule(ky='#indiegame',clean=True,folow=True)
def loop_schedule(date):
while True:
for ky in ['#indiedev','#indiegame']:
print 'Day '+str(date)+' and keyword '+str(ky)
run_schedule(dt=date,ky=ky)
d = get_date()
if date != d:
date = d
break
#Main Functions for 'run_schedule()'
#Keeps track of who doesn't follow back. If an ID appears twice in this category then it
#unfollows. Bot follows ALL followers. Filter for undesirable accounts will be implemented soon.
def tit_for_tat():
print 'Tit for Tat!'
follow_me = twitter_client.followers_ids() #who follows me
follow_you = twitter_client.friends_ids() #who do I follow
erros_ids = []
fol_len = len([1 for id_ in follow_me if id_ not in follow_you])
print 'Following '+str(fol_len)+' new users.'
for id_ in follow_me:
if id_ not in follow_you:
try:
twitter_client.create_friendship(id_)
time.sleep(5)
except:
erros_ids.append(id_)
unfollow,rem_len = remember_follow()
print 'Unfollowing '+str(len(unfollow))+'. Remembering '+str(rem_len)+'.'
for id_ in follow_you:
if id_ in unfollow:
try:
twitter_client.destroy_friendship(id_)
time.sleep(5)
except:
erros_ids.append(id_)
#Take previous day tweets. Rank by higher RT of smaller accounts. Try to RT the underdog!
def RT_last_day(date,key_='#indiedev'):
print 'RT '+str(key_)+' most relevant tweets from yesterday!'
d = latest_tweets(date=date,key_=key_)
d = rank_sort(d)
plot_distro(d[:5000])
a = RT_this(d[:10000],sleep_t=180)
return a
#Take timelines from followers and looks for keyword (default: #indiedev. RTs top tweets (default=2).
def RT_followers(key_='#indiedev',max_=150,rts_=2): #900/1500 Rate limit
print 'RT '+str(key_)+' within followers!'
clct,twtn = [],0
friends = twitter_client.followers_ids()
#Get collection of tweets of followers.
for f in friends:
c=[]
try:
c = twitter_client.user_timeline(f,count=100)
except:
#print 'Cant retweet follower.'
pass
tcl = [ci for ci in c if '#indiedev' in ci.text and ci.in_reply_to_status_id == None]
ttcl = []
for t in tcl:
keep = True
for word in banwords:
if word in t.entities['hashtags']:
keep = False
if keep == True:
ttcl.append(t)
tcl = ttcl
if len(tcl) > 0:
dc = {str(i.id):int(i.retweet_count) for i in tcl}
dfc = pd.DataFrame.from_dict(dc,orient='index')
dfc = dfc.sort(0,ascending=False)
#Final collection of RTs considers only top statuses
clct = clct + list(dfc.index[:rts_])
#After selection of most desirable RTs, we randomly RT them.
rd.shuffle(clct)
print 'Going for '+str(len(clct[:max_]))+' tweets.'
for id_ in clct:
if twtn >= max_: break
try:
twtn+=1
twitter_client.retweet(id_)
print 'Tweeted '+str(twtn)
time.sleep(120)
twitter_client.create_favorite(id_)
except:
pass
#RT statuses from a given DataFrame d.
def RT_this(d,sleep_t=60,stop_at=500,allow_like=False,checkmedia=False):
err,twts,iters=0,0,0
for tweet in d.values:
if twts == stop_at:
print 'Got '+str(stop_at)+' tweets. Stopping.'
break
try:
like,publish,i = True,True,0
if type(tweet[11]) != None:
try:
if tweet[11] in banRT: like,publish = False,False
except:
pass
i=1
if len(tweet[27]) > 14:
if tweet[27][:15] in banTXT: like,publish = False,False
i=2
like = False
if like and rd.random() < 0.05:
twitter_client.create_favorite(tweet[8])
print 'Liked '+tweet[27]
if publish == False:
time.sleep(60)
i=3
if checkmedia and publish:
publish = filter_gif(tweet) | i,twts=4,twts+1
print 'RTed : '+str(twts)+' at '+str(time.ctime())
time.sleep(sleep_t)
i=5
if type(tweet[11]) != None:
banRT.append(tweet[11])
if len(tweet[27]) > 14:
banTXT.append(tweet[27][:15])
except:
pass
try:
u_fol,u_fri=tweet[29]['followers_count'],tweet[29]['friends_count']
if (u_fol > 500 and u_fol < 10000) or (u_fol > 1.5*u_fri):
if i==4:
time.sleep(sleep_t)
i=6
twitter_client.create_friendship(tweet[29]['id'])
except:
pass
if like and allow_like:
try:
twitter_client.create_favorite(tweet[8])
print 'Liked '+tweet[27]
time.sleep(sleep_t/3)
except:
print 'Couldnt like'
save_banDF()
iters+=1
except:
err+=1
#StreamListener: DoggoListener is not currently on use as RTs of previous day and followers
#use most of the bot's daytime.
#List of words to avoid and possible tweet lexic
banword = ["porn","pron","p0rn","pr0n"]
doggolex = ['*doggehyped*']
class DoggoListener(StreamListener):
def on_data(self, data):
tweet = json.loads(data)
i,like,publish = 0,True,True
try:
for word in banword:
if word in tweet['text'].lower(): like,publish = False,False
i=1
if tweet.get('lang') and tweet.get('lang') != 'en': like,publish = False,False
i=2
try:
if type(tweet['user']['description']) != None:
if 'indie' not in tweet['user']['description'] or 'dev' not in tweet['user']['description'] or 'developer' not in tweet['user']['description']:
like = False
if tweet['user']['followers_count'] < 1000: publish = False
else:
like,publish = False,False
except:
like,publish = False,False
i=3
if type(tweet['in_reply_to_status_id']) != None:
if tweet['in_reply_to_status_id'] in banRT:
like,publish = False,False
i=4
if len(tweet['text']) > 14:
if tweet['text'][:15] in banTXT:
like,publish = False,False
i=5
if like:
twitter_client.create_favorite(tweet['id'])
print 'Liked '+tweet['text']
if publish == False:
time.sleep(10)
i=6
if publish:
twitter_client.retweet(tweet['id'])
#Some console output to check if stream is RTweeting.
try:
print 'RTd: '+str(tweet['text'])
except:
print '*Woof*'
i='t'
if type(tweet['in_reply_to_status_id']) != None:
i=7
banRT.append(tweet['in_reply_to_status_id'])
if len(tweet['text']) > 14:
i=8
banTXT.append(tweet['text'][:15])
save_banDF()
time.sleep(60)
except:
print i #For debugging purposes
return True
def run_doggo_run(): #Streams the doggolistener()
if __ | if publish:
try:
twitter_client.retweet(tweet[8]) | random_line_split |
calibrate_gripper_g1r1.py | amp_temp':{
'type': 'adc_linear_3V3', #3V3 supply, no divider
'name': 'Microchip TC1047',
'cb_mV_at_25C': 750.0,
'cb_mV_per_C': 10.0,
'cb_scale': 1.0,
'cb_bias': 0.0,
},
'motor_temp':{
'type': 'adc_linear_3V3', #5V supply, no divider
'name': 'Analog TMP36',
'cb_mV_at_25C': 750.0,
'cb_mV_per_C': 10.0,
'cb_scale': 1.0,
'cb_bias': 0.0},
'torque':{
'type': 'adc_poly',
'name': 'Allegro A1321',
'cb_inv_torque': [1,0],
'cb_torque': [1,0],
'cb_scale': 1.0,
'cb_bias': 0.0},
'current':{
'type': 'none',
'cb_scale': 0.0,
'cb_bias': 0.0},
},
'param':{
'max_amp_temp': 100.0,
'max_current': 800,
'max_motor_temp': 75.0,
'max_tq': 150.0,
'min_tq': -30.0,
'thetadot_deadband': 1.0
},
'param_internal':
{
'calib_tq_degree':1,
'pwm_theta':[-800,800],
'pwm_torque':[-1000,-1000],
'joint_limits':[0,315.0]
}
}
config_default_g1_j1={
'calib':{
'motor':{
'name': 'Maxon RE13 2.5W 24V',
'winding_resistance': 53.2,#Ohm
'winding_inductance':1.79,#mH
'torque_constant':19.7, #mNm/A
'thermal_resistance_housing_ambient': 33.0,#K/W
'thermal_resistance_rotor_housing': 7.0,#K/W
'max_winding_temp': 85, #C
'gear_ratio': 275.0,
'thermal_time_constant_winding': 4.85, #S
'thermal_time_constant_motor':346, #S
'temp_sensor_type':'housing'
},
'theta':{
'type': 'ma3_12bit',
'name': 'US Digital MA3',
'cb_scale': 1.0,
'cb_bias': 0.0},
'amp_temp':{
'type': 'adc_linear_3V3', #3V3 supply, no divider
'name': 'Microchip TC1047',
'cb_mV_at_25C': 750.0,
'cb_mV_per_C': 10.0,
'cb_scale': 1.0,
'cb_bias': 0.0,
},
'motor_temp':{
'type': 'adc_linear_3V3', #5V supply, no divider=3V3
'name': 'Analog TMP36',
'cb_mV_at_25C': 750.0,
'cb_mV_per_C': 10.0,
'cb_scale': 1.0,
'cb_bias': 0.0},
'torque':{
'type': 'adc_poly',
'name': 'Allegro A1321',
'cb_inv_torque': [1,0],
'cb_torque': [1,0],
'cb_scale': 1.0,
'cb_bias': 0.0},
'current':{
'type': 'none',
'cb_scale': 0.0,
'cb_bias': 0.0},
},
'param':{
'max_amp_temp': 100.0,
'max_current': 800,
'max_motor_temp': 75.0,
'max_tq': 150.0,
'min_tq': -30.0,
'thetadot_deadband': 1.0
},
'param_internal':
{
'calib_tq_degree':1,
'pwm_theta':[-800,800],
'pwm_torque':[-1000,-1000],
'joint_limits':[0,315.0]
}
}
class M3Calibrate_Gripper_G1R1(M3CalibrateActuatorEcR1):
def __init__(self):
M3CalibrateActuatorEcR1.__init__(self)
self.joint_names=['Left Digit J0',
'Right Digit J1']
self.config_default=[
config_default_g1_j0,
config_default_g1_j1]
def start(self,ctype):
if not M3CalibrateActuatorEcR1.start(self,ctype):
return False
self.jid=int(self.comp_ec.name[self.comp_ec.name.find('_j')+2:])
self.calib_default=self.config_default[self.jid]['calib']
self.param_default=self.config_default[self.jid]['param']
self.param_internal=self.config_default[self.jid]['param_internal']
print 'Calibrating joint',self.joint_names[self.jid]
return True
def do_task(self,ct):
if ct=='ch':
self.reset_sensor('torque')
self.calibrate_torque()
self.write_config()
return True
if ct=='tt':
self.reset_sensor('theta')
self.calibrate_theta()
self.write_config()
return True
if M3CalibrateActuatorEc.do_task(self,ct):
return True
return False
def print_tasks(self):
M3CalibrateActuatorEcR1.print_tasks(self)
print 'ch: calibrate torque'
print 'tt: calibrate theta'
def display_sensors(self):
M3CalibrateActuatorEcR1.display_sensors(self) | print 'Pos: (mm) : '+'%3.3f'%pos+' Qei On '+'%d'%q_on+' Qei Period '+'%d'%q_p+' Qei Rollover '+'%d'%q_r
raw=self.comp_ec.status.adc_torque
c=self.torque.raw_2_mNm(self.comp_rt.config['calib']['torque'],raw)
mN=c/self.comp_j.config['calib']['cb_drive_radius_m']
print 'Force: (g) : '+'%3.2f'%m3u.mN2g(mN)+' (mN): '+'%3.2f'%mN+' (ADC) '+'%d'%raw
def calibrate_torque(self):
self.proxy.publish_command(self.comp_rt)
self.proxy.publish_param(self.comp_rt)
self.proxy.make_operational(self.name_rt)
self.step()
print 'Make sure other digit is all the way open'
print 'Place digit in zero load condition'
print 'Hit enter when ready'
raw_input()
self.step()
raw_a=int(self.get_sensor_list_avg(['adc_torque'],1.0)['adc_torque'])
load_a=0
print 'Hang 1Kg weight from gripper near slider'
print 'Hit enter to move joint in first direction.'
raw_input()
self.comp_rt.set_mode_pwm()
print 'Desired pwm? [',self.param_internal['pwm_torque'][0],']?'
p=int(m3t.get_float(self.param_internal['pwm_theta'][0]))
self.comp_rt.set_pwm(p)
self.step()
print 'Hit any key when ready to sample'
raw_input()
raw_b=int(self.get_sensor_list_avg(['adc_torque'],1.0)['adc_torque'])
print 'Was load in the opening direction [y]?'
if m3t.get_yes_no('y'):
load_b=m3u.g2mN(1000.0)*self.comp_j.config['calib']['cb_drive_radius_m']
else:
load_b=m3u.g2mN(-1000.0)*self.comp_j.config['calib']['cb_drive | q_on=self.comp_ec.status.qei_on
q_p=self.comp_ec.status.qei_period
q_r=self.comp_ec.status.qei_rollover
c=self.theta.raw_2_deg(self.comp_rt.config['calib']['theta'],q_on,q_p,q_r)
pos=1000.0*math.pi*2*self.comp_j.config['calib']['cb_drive_radius_m']*c/360.0 | random_line_split |
calibrate_gripper_g1r1.py | _temp':{
'type': 'adc_linear_3V3', #3V3 supply, no divider
'name': 'Microchip TC1047',
'cb_mV_at_25C': 750.0,
'cb_mV_per_C': 10.0,
'cb_scale': 1.0,
'cb_bias': 0.0,
},
'motor_temp':{
'type': 'adc_linear_3V3', #5V supply, no divider
'name': 'Analog TMP36',
'cb_mV_at_25C': 750.0,
'cb_mV_per_C': 10.0,
'cb_scale': 1.0,
'cb_bias': 0.0},
'torque':{
'type': 'adc_poly',
'name': 'Allegro A1321',
'cb_inv_torque': [1,0],
'cb_torque': [1,0],
'cb_scale': 1.0,
'cb_bias': 0.0},
'current':{
'type': 'none',
'cb_scale': 0.0,
'cb_bias': 0.0},
},
'param':{
'max_amp_temp': 100.0,
'max_current': 800,
'max_motor_temp': 75.0,
'max_tq': 150.0,
'min_tq': -30.0,
'thetadot_deadband': 1.0
},
'param_internal':
{
'calib_tq_degree':1,
'pwm_theta':[-800,800],
'pwm_torque':[-1000,-1000],
'joint_limits':[0,315.0]
}
}
config_default_g1_j1={
'calib':{
'motor':{
'name': 'Maxon RE13 2.5W 24V',
'winding_resistance': 53.2,#Ohm
'winding_inductance':1.79,#mH
'torque_constant':19.7, #mNm/A
'thermal_resistance_housing_ambient': 33.0,#K/W
'thermal_resistance_rotor_housing': 7.0,#K/W
'max_winding_temp': 85, #C
'gear_ratio': 275.0,
'thermal_time_constant_winding': 4.85, #S
'thermal_time_constant_motor':346, #S
'temp_sensor_type':'housing'
},
'theta':{
'type': 'ma3_12bit',
'name': 'US Digital MA3',
'cb_scale': 1.0,
'cb_bias': 0.0},
'amp_temp':{
'type': 'adc_linear_3V3', #3V3 supply, no divider
'name': 'Microchip TC1047',
'cb_mV_at_25C': 750.0,
'cb_mV_per_C': 10.0,
'cb_scale': 1.0,
'cb_bias': 0.0,
},
'motor_temp':{
'type': 'adc_linear_3V3', #5V supply, no divider=3V3
'name': 'Analog TMP36',
'cb_mV_at_25C': 750.0,
'cb_mV_per_C': 10.0,
'cb_scale': 1.0,
'cb_bias': 0.0},
'torque':{
'type': 'adc_poly',
'name': 'Allegro A1321',
'cb_inv_torque': [1,0],
'cb_torque': [1,0],
'cb_scale': 1.0,
'cb_bias': 0.0},
'current':{
'type': 'none',
'cb_scale': 0.0,
'cb_bias': 0.0},
},
'param':{
'max_amp_temp': 100.0,
'max_current': 800,
'max_motor_temp': 75.0,
'max_tq': 150.0,
'min_tq': -30.0,
'thetadot_deadband': 1.0
},
'param_internal':
{
'calib_tq_degree':1,
'pwm_theta':[-800,800],
'pwm_torque':[-1000,-1000],
'joint_limits':[0,315.0]
}
}
class M3Calibrate_Gripper_G1R1(M3CalibrateActuatorEcR1):
def __init__(self):
M3CalibrateActuatorEcR1.__init__(self)
self.joint_names=['Left Digit J0',
'Right Digit J1']
self.config_default=[
config_default_g1_j0,
config_default_g1_j1]
def start(self,ctype):
if not M3CalibrateActuatorEcR1.start(self,ctype):
return False
self.jid=int(self.comp_ec.name[self.comp_ec.name.find('_j')+2:])
self.calib_default=self.config_default[self.jid]['calib']
self.param_default=self.config_default[self.jid]['param']
self.param_internal=self.config_default[self.jid]['param_internal']
print 'Calibrating joint',self.joint_names[self.jid]
return True
def do_task(self,ct):
if ct=='ch':
|
if ct=='tt':
self.reset_sensor('theta')
self.calibrate_theta()
self.write_config()
return True
if M3CalibrateActuatorEc.do_task(self,ct):
return True
return False
def print_tasks(self):
M3CalibrateActuatorEcR1.print_tasks(self)
print 'ch: calibrate torque'
print 'tt: calibrate theta'
def display_sensors(self):
M3CalibrateActuatorEcR1.display_sensors(self)
q_on=self.comp_ec.status.qei_on
q_p=self.comp_ec.status.qei_period
q_r=self.comp_ec.status.qei_rollover
c=self.theta.raw_2_deg(self.comp_rt.config['calib']['theta'],q_on,q_p,q_r)
pos=1000.0*math.pi*2*self.comp_j.config['calib']['cb_drive_radius_m']*c/360.0
print 'Pos: (mm) : '+'%3.3f'%pos+' Qei On '+'%d'%q_on+' Qei Period '+'%d'%q_p+' Qei Rollover '+'%d'%q_r
raw=self.comp_ec.status.adc_torque
c=self.torque.raw_2_mNm(self.comp_rt.config['calib']['torque'],raw)
mN=c/self.comp_j.config['calib']['cb_drive_radius_m']
print 'Force: (g) : '+'%3.2f'%m3u.mN2g(mN)+' (mN): '+'%3.2f'%mN+' (ADC) '+'%d'%raw
def calibrate_torque(self):
self.proxy.publish_command(self.comp_rt)
self.proxy.publish_param(self.comp_rt)
self.proxy.make_operational(self.name_rt)
self.step()
print 'Make sure other digit is all the way open'
print 'Place digit in zero load condition'
print 'Hit enter when ready'
raw_input()
self.step()
raw_a=int(self.get_sensor_list_avg(['adc_torque'],1.0)['adc_torque'])
load_a=0
print 'Hang 1Kg weight from gripper near slider'
print 'Hit enter to move joint in first direction.'
raw_input()
self.comp_rt.set_mode_pwm()
print 'Desired pwm? [',self.param_internal['pwm_torque'][0],']?'
p=int(m3t.get_float(self.param_internal['pwm_theta'][0]))
self.comp_rt.set_pwm(p)
self.step()
print 'Hit any key when ready to sample'
raw_input()
raw_b=int(self.get_sensor_list_avg(['adc_torque'],1.0)['adc_torque'])
print 'Was load in the opening direction [y]?'
if m3t.get_yes_no('y'):
load_b=m3u.g2mN(1000.0)*self.comp_j.config['calib']['cb_drive_radius_m']
else:
load_b=m3u.g2mN(-1000.0)*self.comp_j.config['calib']['cb_drive | self.reset_sensor('torque')
self.calibrate_torque()
self.write_config()
return True | conditional_block |
calibrate_gripper_g1r1.py | _temp':{
'type': 'adc_linear_3V3', #3V3 supply, no divider
'name': 'Microchip TC1047',
'cb_mV_at_25C': 750.0,
'cb_mV_per_C': 10.0,
'cb_scale': 1.0,
'cb_bias': 0.0,
},
'motor_temp':{
'type': 'adc_linear_3V3', #5V supply, no divider
'name': 'Analog TMP36',
'cb_mV_at_25C': 750.0,
'cb_mV_per_C': 10.0,
'cb_scale': 1.0,
'cb_bias': 0.0},
'torque':{
'type': 'adc_poly',
'name': 'Allegro A1321',
'cb_inv_torque': [1,0],
'cb_torque': [1,0],
'cb_scale': 1.0,
'cb_bias': 0.0},
'current':{
'type': 'none',
'cb_scale': 0.0,
'cb_bias': 0.0},
},
'param':{
'max_amp_temp': 100.0,
'max_current': 800,
'max_motor_temp': 75.0,
'max_tq': 150.0,
'min_tq': -30.0,
'thetadot_deadband': 1.0
},
'param_internal':
{
'calib_tq_degree':1,
'pwm_theta':[-800,800],
'pwm_torque':[-1000,-1000],
'joint_limits':[0,315.0]
}
}
config_default_g1_j1={
'calib':{
'motor':{
'name': 'Maxon RE13 2.5W 24V',
'winding_resistance': 53.2,#Ohm
'winding_inductance':1.79,#mH
'torque_constant':19.7, #mNm/A
'thermal_resistance_housing_ambient': 33.0,#K/W
'thermal_resistance_rotor_housing': 7.0,#K/W
'max_winding_temp': 85, #C
'gear_ratio': 275.0,
'thermal_time_constant_winding': 4.85, #S
'thermal_time_constant_motor':346, #S
'temp_sensor_type':'housing'
},
'theta':{
'type': 'ma3_12bit',
'name': 'US Digital MA3',
'cb_scale': 1.0,
'cb_bias': 0.0},
'amp_temp':{
'type': 'adc_linear_3V3', #3V3 supply, no divider
'name': 'Microchip TC1047',
'cb_mV_at_25C': 750.0,
'cb_mV_per_C': 10.0,
'cb_scale': 1.0,
'cb_bias': 0.0,
},
'motor_temp':{
'type': 'adc_linear_3V3', #5V supply, no divider=3V3
'name': 'Analog TMP36',
'cb_mV_at_25C': 750.0,
'cb_mV_per_C': 10.0,
'cb_scale': 1.0,
'cb_bias': 0.0},
'torque':{
'type': 'adc_poly',
'name': 'Allegro A1321',
'cb_inv_torque': [1,0],
'cb_torque': [1,0],
'cb_scale': 1.0,
'cb_bias': 0.0},
'current':{
'type': 'none',
'cb_scale': 0.0,
'cb_bias': 0.0},
},
'param':{
'max_amp_temp': 100.0,
'max_current': 800,
'max_motor_temp': 75.0,
'max_tq': 150.0,
'min_tq': -30.0,
'thetadot_deadband': 1.0
},
'param_internal':
{
'calib_tq_degree':1,
'pwm_theta':[-800,800],
'pwm_torque':[-1000,-1000],
'joint_limits':[0,315.0]
}
}
class M3Calibrate_Gripper_G1R1(M3CalibrateActuatorEcR1):
def __init__(self):
M3CalibrateActuatorEcR1.__init__(self)
self.joint_names=['Left Digit J0',
'Right Digit J1']
self.config_default=[
config_default_g1_j0,
config_default_g1_j1]
def start(self,ctype):
if not M3CalibrateActuatorEcR1.start(self,ctype):
return False
self.jid=int(self.comp_ec.name[self.comp_ec.name.find('_j')+2:])
self.calib_default=self.config_default[self.jid]['calib']
self.param_default=self.config_default[self.jid]['param']
self.param_internal=self.config_default[self.jid]['param_internal']
print 'Calibrating joint',self.joint_names[self.jid]
return True
def do_task(self,ct):
|
def print_tasks(self):
M3CalibrateActuatorEcR1.print_tasks(self)
print 'ch: calibrate torque'
print 'tt: calibrate theta'
def display_sensors(self):
M3CalibrateActuatorEcR1.display_sensors(self)
q_on=self.comp_ec.status.qei_on
q_p=self.comp_ec.status.qei_period
q_r=self.comp_ec.status.qei_rollover
c=self.theta.raw_2_deg(self.comp_rt.config['calib']['theta'],q_on,q_p,q_r)
pos=1000.0*math.pi*2*self.comp_j.config['calib']['cb_drive_radius_m']*c/360.0
print 'Pos: (mm) : '+'%3.3f'%pos+' Qei On '+'%d'%q_on+' Qei Period '+'%d'%q_p+' Qei Rollover '+'%d'%q_r
raw=self.comp_ec.status.adc_torque
c=self.torque.raw_2_mNm(self.comp_rt.config['calib']['torque'],raw)
mN=c/self.comp_j.config['calib']['cb_drive_radius_m']
print 'Force: (g) : '+'%3.2f'%m3u.mN2g(mN)+' (mN): '+'%3.2f'%mN+' (ADC) '+'%d'%raw
def calibrate_torque(self):
self.proxy.publish_command(self.comp_rt)
self.proxy.publish_param(self.comp_rt)
self.proxy.make_operational(self.name_rt)
self.step()
print 'Make sure other digit is all the way open'
print 'Place digit in zero load condition'
print 'Hit enter when ready'
raw_input()
self.step()
raw_a=int(self.get_sensor_list_avg(['adc_torque'],1.0)['adc_torque'])
load_a=0
print 'Hang 1Kg weight from gripper near slider'
print 'Hit enter to move joint in first direction.'
raw_input()
self.comp_rt.set_mode_pwm()
print 'Desired pwm? [',self.param_internal['pwm_torque'][0],']?'
p=int(m3t.get_float(self.param_internal['pwm_theta'][0]))
self.comp_rt.set_pwm(p)
self.step()
print 'Hit any key when ready to sample'
raw_input()
raw_b=int(self.get_sensor_list_avg(['adc_torque'],1.0)['adc_torque'])
print 'Was load in the opening direction [y]?'
if m3t.get_yes_no('y'):
load_b=m3u.g2mN(1000.0)*self.comp_j.config['calib']['cb_drive_radius_m']
else:
load_b=m3u.g2mN(-1000.0)*self.comp_j.config['calib']['cb_drive | if ct=='ch':
self.reset_sensor('torque')
self.calibrate_torque()
self.write_config()
return True
if ct=='tt':
self.reset_sensor('theta')
self.calibrate_theta()
self.write_config()
return True
if M3CalibrateActuatorEc.do_task(self,ct):
return True
return False | identifier_body |
calibrate_gripper_g1r1.py | _temp':{
'type': 'adc_linear_3V3', #3V3 supply, no divider
'name': 'Microchip TC1047',
'cb_mV_at_25C': 750.0,
'cb_mV_per_C': 10.0,
'cb_scale': 1.0,
'cb_bias': 0.0,
},
'motor_temp':{
'type': 'adc_linear_3V3', #5V supply, no divider
'name': 'Analog TMP36',
'cb_mV_at_25C': 750.0,
'cb_mV_per_C': 10.0,
'cb_scale': 1.0,
'cb_bias': 0.0},
'torque':{
'type': 'adc_poly',
'name': 'Allegro A1321',
'cb_inv_torque': [1,0],
'cb_torque': [1,0],
'cb_scale': 1.0,
'cb_bias': 0.0},
'current':{
'type': 'none',
'cb_scale': 0.0,
'cb_bias': 0.0},
},
'param':{
'max_amp_temp': 100.0,
'max_current': 800,
'max_motor_temp': 75.0,
'max_tq': 150.0,
'min_tq': -30.0,
'thetadot_deadband': 1.0
},
'param_internal':
{
'calib_tq_degree':1,
'pwm_theta':[-800,800],
'pwm_torque':[-1000,-1000],
'joint_limits':[0,315.0]
}
}
config_default_g1_j1={
'calib':{
'motor':{
'name': 'Maxon RE13 2.5W 24V',
'winding_resistance': 53.2,#Ohm
'winding_inductance':1.79,#mH
'torque_constant':19.7, #mNm/A
'thermal_resistance_housing_ambient': 33.0,#K/W
'thermal_resistance_rotor_housing': 7.0,#K/W
'max_winding_temp': 85, #C
'gear_ratio': 275.0,
'thermal_time_constant_winding': 4.85, #S
'thermal_time_constant_motor':346, #S
'temp_sensor_type':'housing'
},
'theta':{
'type': 'ma3_12bit',
'name': 'US Digital MA3',
'cb_scale': 1.0,
'cb_bias': 0.0},
'amp_temp':{
'type': 'adc_linear_3V3', #3V3 supply, no divider
'name': 'Microchip TC1047',
'cb_mV_at_25C': 750.0,
'cb_mV_per_C': 10.0,
'cb_scale': 1.0,
'cb_bias': 0.0,
},
'motor_temp':{
'type': 'adc_linear_3V3', #5V supply, no divider=3V3
'name': 'Analog TMP36',
'cb_mV_at_25C': 750.0,
'cb_mV_per_C': 10.0,
'cb_scale': 1.0,
'cb_bias': 0.0},
'torque':{
'type': 'adc_poly',
'name': 'Allegro A1321',
'cb_inv_torque': [1,0],
'cb_torque': [1,0],
'cb_scale': 1.0,
'cb_bias': 0.0},
'current':{
'type': 'none',
'cb_scale': 0.0,
'cb_bias': 0.0},
},
'param':{
'max_amp_temp': 100.0,
'max_current': 800,
'max_motor_temp': 75.0,
'max_tq': 150.0,
'min_tq': -30.0,
'thetadot_deadband': 1.0
},
'param_internal':
{
'calib_tq_degree':1,
'pwm_theta':[-800,800],
'pwm_torque':[-1000,-1000],
'joint_limits':[0,315.0]
}
}
class M3Calibrate_Gripper_G1R1(M3CalibrateActuatorEcR1):
def __init__(self):
M3CalibrateActuatorEcR1.__init__(self)
self.joint_names=['Left Digit J0',
'Right Digit J1']
self.config_default=[
config_default_g1_j0,
config_default_g1_j1]
def start(self,ctype):
if not M3CalibrateActuatorEcR1.start(self,ctype):
return False
self.jid=int(self.comp_ec.name[self.comp_ec.name.find('_j')+2:])
self.calib_default=self.config_default[self.jid]['calib']
self.param_default=self.config_default[self.jid]['param']
self.param_internal=self.config_default[self.jid]['param_internal']
print 'Calibrating joint',self.joint_names[self.jid]
return True
def | (self,ct):
if ct=='ch':
self.reset_sensor('torque')
self.calibrate_torque()
self.write_config()
return True
if ct=='tt':
self.reset_sensor('theta')
self.calibrate_theta()
self.write_config()
return True
if M3CalibrateActuatorEc.do_task(self,ct):
return True
return False
def print_tasks(self):
M3CalibrateActuatorEcR1.print_tasks(self)
print 'ch: calibrate torque'
print 'tt: calibrate theta'
def display_sensors(self):
M3CalibrateActuatorEcR1.display_sensors(self)
q_on=self.comp_ec.status.qei_on
q_p=self.comp_ec.status.qei_period
q_r=self.comp_ec.status.qei_rollover
c=self.theta.raw_2_deg(self.comp_rt.config['calib']['theta'],q_on,q_p,q_r)
pos=1000.0*math.pi*2*self.comp_j.config['calib']['cb_drive_radius_m']*c/360.0
print 'Pos: (mm) : '+'%3.3f'%pos+' Qei On '+'%d'%q_on+' Qei Period '+'%d'%q_p+' Qei Rollover '+'%d'%q_r
raw=self.comp_ec.status.adc_torque
c=self.torque.raw_2_mNm(self.comp_rt.config['calib']['torque'],raw)
mN=c/self.comp_j.config['calib']['cb_drive_radius_m']
print 'Force: (g) : '+'%3.2f'%m3u.mN2g(mN)+' (mN): '+'%3.2f'%mN+' (ADC) '+'%d'%raw
def calibrate_torque(self):
self.proxy.publish_command(self.comp_rt)
self.proxy.publish_param(self.comp_rt)
self.proxy.make_operational(self.name_rt)
self.step()
print 'Make sure other digit is all the way open'
print 'Place digit in zero load condition'
print 'Hit enter when ready'
raw_input()
self.step()
raw_a=int(self.get_sensor_list_avg(['adc_torque'],1.0)['adc_torque'])
load_a=0
print 'Hang 1Kg weight from gripper near slider'
print 'Hit enter to move joint in first direction.'
raw_input()
self.comp_rt.set_mode_pwm()
print 'Desired pwm? [',self.param_internal['pwm_torque'][0],']?'
p=int(m3t.get_float(self.param_internal['pwm_theta'][0]))
self.comp_rt.set_pwm(p)
self.step()
print 'Hit any key when ready to sample'
raw_input()
raw_b=int(self.get_sensor_list_avg(['adc_torque'],1.0)['adc_torque'])
print 'Was load in the opening direction [y]?'
if m3t.get_yes_no('y'):
load_b=m3u.g2mN(1000.0)*self.comp_j.config['calib']['cb_drive_radius_m']
else:
load_b=m3u.g2mN(-1000.0)*self.comp_j.config['calib']['cb | do_task | identifier_name |
getfood_base.py | mine = 4
def __init__(self,
agent_start_pos=(1, 1),
health_cap=50,
food_rate=4,
grid_size=8,
obs_vision=False,
reward_type='delta',
fully_observed=False,
only_partial_obs=False,
can_die=True,
one_hot_obs=True,
mixing_time_periods=[],
mixing_time_period_length=120,
**kwargs
):
self.agent_start_pos = agent_start_pos
# self.agent_start_dir = agent_start_dir
self.food_rate = food_rate
self.health_cap = health_cap
self.health = health_cap
self.last_health = self.health
self.obs_vision = obs_vision
self.reward_type = reward_type
self.fully_observed = fully_observed
self.only_partial_obs = only_partial_obs
self.can_die = can_die
self.one_hot_obs = one_hot_obs
# for conditional entropy of s' | s
self.transition_count = {}
self.prev_obs_string = ''
# for mixing time
self.mixing_time_periods = mixing_time_periods
self.max_mixing_time_period = max(mixing_time_periods) if mixing_time_periods else 0
self.mixing_time_period_length = mixing_time_period_length
self.obs_counts = []
if not hasattr(self, 'actions'):
self.actions = FoodEnvBase.Actions
super().__init__(
# Set this to True for maximum speed
see_through_walls=True,
grid_size=grid_size,
**kwargs
)
def _reward(self):
if self.reward_type == 'survival':
rwd = 1
elif self.reward_type == 'delta':
rwd = self.health - self.last_health
elif self.reward_type == 'health':
rwd = self.health
else:
assert False, "Reward type not matched"
self.last_health = self.health
return rwd
def _gen_grid(self, width, height):
# Create an empty grid
self.grid = GridAbsolute(width, height)
# Generate the surrounding walls
self.grid.wall_rect(0, 0, width, height)
# Place a goal square in the bottom-right corner
# self.grid.set(width - 2, height - 2, Goal())
self.extra_gen_grid()
# Place the agent
if self.agent_start_pos is not None:
self.start_pos = self.agent_start_pos
# self.start_dir = self.agent_start_dir
else:
self.place_agent()
self.mission = None
def step(self, action, incl_health=True):
done = False
matched = super().step(action, override=True)
# subclass-defined extra actions. if not caught by that, then unknown action
if not self.extra_step(action, matched):
assert False, "unknown action %d" % action
# decrease health bar
self.decay_health()
# generate new food
self.place_items()
# generate obs after action is caught and food is placed. generate reward before death check
img = self.get_img(onehot=self.one_hot_obs)
full_img = self.get_full_img(scale=1 if self.fully_observed else 1 / 8, onehot=self.one_hot_obs)
# NOTE: below not nec due to onehot being passed into two func calls above. but leaving here for now in case.
# if self.one_hot_obs:
# # ignore second channel since redundant (due to one-to-one mapping btwn color and obj type for now)
# img = np.concatenate([np.eye(len(self.object_to_idx))[ch].transpose(2, 0, 1) for ch in img[:1]])
# full_img = np.concatenate([np.eye(len(self.object_to_idx))[ch].transpose(2, 0, 1) for ch in full_img[:1]])
rwd = self._reward()
# tick on each grid item
to_remove = []
for j in range(0, self.grid.height):
for i in range(0, self.grid.width):
cell = self.grid.get(i, j)
if cell is not None:
if not cell.step():
self.dead_obj(i, j, cell)
to_remove.append((i, j))
for idxs in to_remove:
self.grid.set(*idxs, None)
# dead.
if self.dead():
done = True
if self.fully_observed:
if incl_health:
obs = np.concatenate((full_img.flatten(), np.array([self.health]), np.array(self.agent_pos)))
else:
obs = np.concatenate((full_img.flatten(), np.array(self.agent_pos)))
elif self.only_partial_obs:
if incl_health:
|
else:
obs = img.flatten()
else:
if incl_health:
obs = np.concatenate((img.flatten(), full_img.flatten(), np.array([self.health])))
else:
obs = np.concatenate((img.flatten(), full_img.flatten()))
obs_string = obs.tostring()
# transition count stuff
self.transition_count.setdefault(hash(self.prev_obs_string), {})
self.transition_count[hash(self.prev_obs_string)][hash(obs_string)] = 1 + self.transition_count[hash(self.prev_obs_string)].get(hash(obs_string), 0)
# mixing time stuff
if self.step_count % self.mixing_time_period_length == 0:
self.obs_counts.append(self.obs_count.copy())
if hasattr(self, 'obs_count') and self.mixing_time_periods and len(self.obs_counts) > self.max_mixing_time_period:
self.obs_counts = self.obs_counts[-(self.max_mixing_time_period+1):]
self.prev_obs_string = obs_string
return obs, rwd, done, {}
def reset(self, incl_health=True):
super().reset()
self.health = self.health_cap
self.extra_reset()
img = self.get_img(onehot=self.one_hot_obs)
full_img = self.get_full_img(onehot=self.one_hot_obs)
self.transition_count = {}
# if self.one_hot_obs:
# # ignore second channel since redundant (due to one-to-one mapping btwn color and obj type for now)
# img = np.concatenate([np.eye(len(self.object_to_idx))[ch].transpose(2, 0, 1) for ch in img[:1]])
# full_img = np.concatenate([np.eye(len(self.object_to_idx))[ch].transpose(2, 0, 1) for ch in full_img[:1]])
if self.fully_observed:
if incl_health:
obs = np.concatenate((full_img.flatten(), np.array([self.health]), np.array(self.agent_pos)))
else:
obs = np.concatenate((full_img.flatten(), np.array(self.agent_pos)))
elif self.only_partial_obs:
if incl_health:
obs = np.concatenate((img.flatten(), np.array([self.health])))
else:
obs = img.flatten()
else:
if incl_health:
obs = np.concatenate((img.flatten(), full_img.flatten(), np.array([self.health])))
else:
obs = np.concatenate((img.flatten(), full_img.flatten()))
self.prev_obs_string = obs.tostring()
return obs
def get_full_img(self, scale=1 / 8, onehot=False):
""" Return the whole grid view """
if self.obs_vision:
full_img = self.get_full_obs_render(scale=scale)
else:
full_img = self.grid.encode(self, onehot=onehot)
# NOTE: in case need to scale here instead of in above func call: return cv2.resize(full_img, (0, 0), fx=0.125, fy=0.125, interpolation=cv2.INTER_AREA)
return full_img
def get_img(self, onehot=False):
""" Return the agent view """
if self.obs_vision:
img = self.gen_obs(onehot=False)
img = self.get_obs_render(img, CELL_PIXELS // 4)
else:
img = self.gen_obs(onehot=onehot)
return img
def extra_step(self, action, matched):
return matched
def extra_reset(self):
pass
def place_items(self):
pass
def extra_gen_grid(self):
pass
def place_prob(self, obj, prob, top=None, size=None):
if np.random.binomial(1, prob):
pos = self.place_obj(obj, top, size)
obj.cur_pos = pos
return True
return False
def decay_health(self):
self.add_health(-1)
def add_health(self, num):
# clip health between 0 and cap after adjustment
self.health = max(0, min(self.health_cap, self.health + num))
def count_type(self, type):
count = 0
for i in range(self.grid_size):
for j in range(self.grid_size):
cell = self.grid.get(i, j)
if type is None and cell is None or cell is not None and cell.type == type:
count += 1
return count
def count_all_types(self):
counts = {}
for i in range(self.grid_size):
for j in range(self.grid_size):
cell = self.grid.get(i, j)
type = cell.type if cell is not None else ''
counts[type] = counts.get | obs = np.concatenate((img.flatten(), np.array([self.health]))) | conditional_block |
getfood_base.py | for conditional entropy of s' | s
self.transition_count = {}
self.prev_obs_string = ''
# for mixing time
self.mixing_time_periods = mixing_time_periods
self.max_mixing_time_period = max(mixing_time_periods) if mixing_time_periods else 0
self.mixing_time_period_length = mixing_time_period_length
self.obs_counts = []
if not hasattr(self, 'actions'):
self.actions = FoodEnvBase.Actions
super().__init__(
# Set this to True for maximum speed
see_through_walls=True,
grid_size=grid_size,
**kwargs
)
def _reward(self):
if self.reward_type == 'survival':
rwd = 1
elif self.reward_type == 'delta':
rwd = self.health - self.last_health
elif self.reward_type == 'health':
rwd = self.health
else:
assert False, "Reward type not matched"
self.last_health = self.health
return rwd
def _gen_grid(self, width, height):
# Create an empty grid
self.grid = GridAbsolute(width, height)
# Generate the surrounding walls
self.grid.wall_rect(0, 0, width, height)
# Place a goal square in the bottom-right corner
# self.grid.set(width - 2, height - 2, Goal())
self.extra_gen_grid()
# Place the agent
if self.agent_start_pos is not None:
self.start_pos = self.agent_start_pos
# self.start_dir = self.agent_start_dir
else:
self.place_agent()
self.mission = None
def step(self, action, incl_health=True):
done = False
matched = super().step(action, override=True)
# subclass-defined extra actions. if not caught by that, then unknown action
if not self.extra_step(action, matched):
assert False, "unknown action %d" % action
# decrease health bar
self.decay_health()
# generate new food
self.place_items()
# generate obs after action is caught and food is placed. generate reward before death check
img = self.get_img(onehot=self.one_hot_obs)
full_img = self.get_full_img(scale=1 if self.fully_observed else 1 / 8, onehot=self.one_hot_obs)
# NOTE: below not nec due to onehot being passed into two func calls above. but leaving here for now in case.
# if self.one_hot_obs:
# # ignore second channel since redundant (due to one-to-one mapping btwn color and obj type for now)
# img = np.concatenate([np.eye(len(self.object_to_idx))[ch].transpose(2, 0, 1) for ch in img[:1]])
# full_img = np.concatenate([np.eye(len(self.object_to_idx))[ch].transpose(2, 0, 1) for ch in full_img[:1]])
rwd = self._reward()
# tick on each grid item
to_remove = []
for j in range(0, self.grid.height):
for i in range(0, self.grid.width):
cell = self.grid.get(i, j)
if cell is not None:
if not cell.step():
self.dead_obj(i, j, cell)
to_remove.append((i, j))
for idxs in to_remove:
self.grid.set(*idxs, None)
# dead.
if self.dead():
done = True
if self.fully_observed:
if incl_health:
obs = np.concatenate((full_img.flatten(), np.array([self.health]), np.array(self.agent_pos)))
else:
obs = np.concatenate((full_img.flatten(), np.array(self.agent_pos)))
elif self.only_partial_obs:
if incl_health:
obs = np.concatenate((img.flatten(), np.array([self.health])))
else:
obs = img.flatten()
else:
if incl_health:
obs = np.concatenate((img.flatten(), full_img.flatten(), np.array([self.health])))
else:
obs = np.concatenate((img.flatten(), full_img.flatten()))
obs_string = obs.tostring()
# transition count stuff
self.transition_count.setdefault(hash(self.prev_obs_string), {})
self.transition_count[hash(self.prev_obs_string)][hash(obs_string)] = 1 + self.transition_count[hash(self.prev_obs_string)].get(hash(obs_string), 0)
# mixing time stuff
if self.step_count % self.mixing_time_period_length == 0:
self.obs_counts.append(self.obs_count.copy())
if hasattr(self, 'obs_count') and self.mixing_time_periods and len(self.obs_counts) > self.max_mixing_time_period:
self.obs_counts = self.obs_counts[-(self.max_mixing_time_period+1):]
self.prev_obs_string = obs_string
return obs, rwd, done, {}
def reset(self, incl_health=True):
super().reset()
self.health = self.health_cap
self.extra_reset()
img = self.get_img(onehot=self.one_hot_obs)
full_img = self.get_full_img(onehot=self.one_hot_obs)
self.transition_count = {}
# if self.one_hot_obs:
# # ignore second channel since redundant (due to one-to-one mapping btwn color and obj type for now)
# img = np.concatenate([np.eye(len(self.object_to_idx))[ch].transpose(2, 0, 1) for ch in img[:1]])
# full_img = np.concatenate([np.eye(len(self.object_to_idx))[ch].transpose(2, 0, 1) for ch in full_img[:1]])
if self.fully_observed:
if incl_health:
obs = np.concatenate((full_img.flatten(), np.array([self.health]), np.array(self.agent_pos)))
else:
obs = np.concatenate((full_img.flatten(), np.array(self.agent_pos)))
elif self.only_partial_obs:
if incl_health:
obs = np.concatenate((img.flatten(), np.array([self.health])))
else:
obs = img.flatten()
else:
if incl_health:
obs = np.concatenate((img.flatten(), full_img.flatten(), np.array([self.health])))
else:
obs = np.concatenate((img.flatten(), full_img.flatten()))
self.prev_obs_string = obs.tostring()
return obs
def get_full_img(self, scale=1 / 8, onehot=False):
""" Return the whole grid view """
if self.obs_vision:
full_img = self.get_full_obs_render(scale=scale)
else:
full_img = self.grid.encode(self, onehot=onehot)
# NOTE: in case need to scale here instead of in above func call: return cv2.resize(full_img, (0, 0), fx=0.125, fy=0.125, interpolation=cv2.INTER_AREA)
return full_img
def get_img(self, onehot=False):
""" Return the agent view """
if self.obs_vision:
img = self.gen_obs(onehot=False)
img = self.get_obs_render(img, CELL_PIXELS // 4)
else:
img = self.gen_obs(onehot=onehot)
return img
def extra_step(self, action, matched):
return matched
def extra_reset(self):
pass
def place_items(self):
pass
def extra_gen_grid(self):
pass
def place_prob(self, obj, prob, top=None, size=None):
if np.random.binomial(1, prob):
pos = self.place_obj(obj, top, size)
obj.cur_pos = pos
return True
return False
def decay_health(self):
self.add_health(-1)
def add_health(self, num):
# clip health between 0 and cap after adjustment
self.health = max(0, min(self.health_cap, self.health + num))
def count_type(self, type):
count = 0
for i in range(self.grid_size):
for j in range(self.grid_size):
cell = self.grid.get(i, j)
if type is None and cell is None or cell is not None and cell.type == type:
count += 1
return count
def count_all_types(self):
counts = {}
for i in range(self.grid_size):
for j in range(self.grid_size):
cell = self.grid.get(i, j)
type = cell.type if cell is not None else ''
counts[type] = counts.get(type, 0) + 1
if hasattr(self, 'monsters'):
counts['monster'] = len(self.monsters)
return counts
def exists_type(self, type):
""" Check if object of type TYPE exists in current grid. """
for i in range(1, self.grid_size - 1):
for j in range(1, self.grid_size - 1):
obj = self.grid.get(i, j)
if obj and obj.type == type:
return True
return False
def dead(self):
return self.can_die and self.health <= 0
def dead_obj(self, i, j, obj):
""" Called when OBJ dies at position (i, j). """
pass
def __getstate__(self):
d = self.__dict__.copy()
del d['grid_render']
return d
def __setstate__(self, d):
self.__dict__.update(d)
class | FoodEnvEmptyFullObs | identifier_name |
|
getfood_base.py | mine = 4
def __init__(self,
agent_start_pos=(1, 1),
health_cap=50,
food_rate=4,
grid_size=8,
obs_vision=False,
reward_type='delta',
fully_observed=False,
only_partial_obs=False,
can_die=True,
one_hot_obs=True,
mixing_time_periods=[],
mixing_time_period_length=120,
**kwargs
):
self.agent_start_pos = agent_start_pos
# self.agent_start_dir = agent_start_dir
self.food_rate = food_rate
self.health_cap = health_cap
self.health = health_cap
self.last_health = self.health
self.obs_vision = obs_vision
self.reward_type = reward_type
self.fully_observed = fully_observed
self.only_partial_obs = only_partial_obs
self.can_die = can_die
self.one_hot_obs = one_hot_obs
# for conditional entropy of s' | s
self.transition_count = {}
self.prev_obs_string = ''
# for mixing time
self.mixing_time_periods = mixing_time_periods
self.max_mixing_time_period = max(mixing_time_periods) if mixing_time_periods else 0
self.mixing_time_period_length = mixing_time_period_length
self.obs_counts = []
if not hasattr(self, 'actions'):
self.actions = FoodEnvBase.Actions
super().__init__(
# Set this to True for maximum speed
see_through_walls=True,
grid_size=grid_size,
**kwargs
)
def _reward(self):
if self.reward_type == 'survival':
rwd = 1
elif self.reward_type == 'delta':
rwd = self.health - self.last_health
elif self.reward_type == 'health':
rwd = self.health
else:
assert False, "Reward type not matched"
self.last_health = self.health
return rwd
def _gen_grid(self, width, height):
# Create an empty grid
self.grid = GridAbsolute(width, height)
# Generate the surrounding walls
self.grid.wall_rect(0, 0, width, height)
# Place a goal square in the bottom-right corner
# self.grid.set(width - 2, height - 2, Goal())
self.extra_gen_grid()
# Place the agent
if self.agent_start_pos is not None:
self.start_pos = self.agent_start_pos
# self.start_dir = self.agent_start_dir
else:
self.place_agent()
self.mission = None
def step(self, action, incl_health=True):
done = False
matched = super().step(action, override=True)
# subclass-defined extra actions. if not caught by that, then unknown action
if not self.extra_step(action, matched):
assert False, "unknown action %d" % action
# decrease health bar
self.decay_health()
# generate new food
self.place_items()
# generate obs after action is caught and food is placed. generate reward before death check
img = self.get_img(onehot=self.one_hot_obs)
full_img = self.get_full_img(scale=1 if self.fully_observed else 1 / 8, onehot=self.one_hot_obs)
# NOTE: below not nec due to onehot being passed into two func calls above. but leaving here for now in case.
# if self.one_hot_obs:
# # ignore second channel since redundant (due to one-to-one mapping btwn color and obj type for now)
# img = np.concatenate([np.eye(len(self.object_to_idx))[ch].transpose(2, 0, 1) for ch in img[:1]])
# full_img = np.concatenate([np.eye(len(self.object_to_idx))[ch].transpose(2, 0, 1) for ch in full_img[:1]])
rwd = self._reward()
# tick on each grid item
to_remove = []
for j in range(0, self.grid.height):
for i in range(0, self.grid.width):
cell = self.grid.get(i, j)
if cell is not None:
if not cell.step():
self.dead_obj(i, j, cell)
to_remove.append((i, j))
for idxs in to_remove:
self.grid.set(*idxs, None)
# dead.
if self.dead():
done = True
if self.fully_observed:
if incl_health:
obs = np.concatenate((full_img.flatten(), np.array([self.health]), np.array(self.agent_pos)))
else:
obs = np.concatenate((full_img.flatten(), np.array(self.agent_pos)))
elif self.only_partial_obs:
if incl_health:
obs = np.concatenate((img.flatten(), np.array([self.health])))
else:
obs = img.flatten()
else:
if incl_health:
obs = np.concatenate((img.flatten(), full_img.flatten(), np.array([self.health])))
else:
obs = np.concatenate((img.flatten(), full_img.flatten()))
obs_string = obs.tostring()
# transition count stuff
self.transition_count.setdefault(hash(self.prev_obs_string), {})
self.transition_count[hash(self.prev_obs_string)][hash(obs_string)] = 1 + self.transition_count[hash(self.prev_obs_string)].get(hash(obs_string), 0)
# mixing time stuff
if self.step_count % self.mixing_time_period_length == 0:
self.obs_counts.append(self.obs_count.copy())
if hasattr(self, 'obs_count') and self.mixing_time_periods and len(self.obs_counts) > self.max_mixing_time_period:
self.obs_counts = self.obs_counts[-(self.max_mixing_time_period+1):]
self.prev_obs_string = obs_string
return obs, rwd, done, {}
def reset(self, incl_health=True):
super().reset()
self.health = self.health_cap
self.extra_reset()
img = self.get_img(onehot=self.one_hot_obs)
full_img = self.get_full_img(onehot=self.one_hot_obs)
self.transition_count = {}
# if self.one_hot_obs: | if self.fully_observed:
if incl_health:
obs = np.concatenate((full_img.flatten(), np.array([self.health]), np.array(self.agent_pos)))
else:
obs = np.concatenate((full_img.flatten(), np.array(self.agent_pos)))
elif self.only_partial_obs:
if incl_health:
obs = np.concatenate((img.flatten(), np.array([self.health])))
else:
obs = img.flatten()
else:
if incl_health:
obs = np.concatenate((img.flatten(), full_img.flatten(), np.array([self.health])))
else:
obs = np.concatenate((img.flatten(), full_img.flatten()))
self.prev_obs_string = obs.tostring()
return obs
def get_full_img(self, scale=1 / 8, onehot=False):
""" Return the whole grid view """
if self.obs_vision:
full_img = self.get_full_obs_render(scale=scale)
else:
full_img = self.grid.encode(self, onehot=onehot)
# NOTE: in case need to scale here instead of in above func call: return cv2.resize(full_img, (0, 0), fx=0.125, fy=0.125, interpolation=cv2.INTER_AREA)
return full_img
def get_img(self, onehot=False):
""" Return the agent view """
if self.obs_vision:
img = self.gen_obs(onehot=False)
img = self.get_obs_render(img, CELL_PIXELS // 4)
else:
img = self.gen_obs(onehot=onehot)
return img
def extra_step(self, action, matched):
return matched
def extra_reset(self):
pass
def place_items(self):
pass
def extra_gen_grid(self):
pass
def place_prob(self, obj, prob, top=None, size=None):
if np.random.binomial(1, prob):
pos = self.place_obj(obj, top, size)
obj.cur_pos = pos
return True
return False
def decay_health(self):
self.add_health(-1)
def add_health(self, num):
# clip health between 0 and cap after adjustment
self.health = max(0, min(self.health_cap, self.health + num))
def count_type(self, type):
count = 0
for i in range(self.grid_size):
for j in range(self.grid_size):
cell = self.grid.get(i, j)
if type is None and cell is None or cell is not None and cell.type == type:
count += 1
return count
def count_all_types(self):
counts = {}
for i in range(self.grid_size):
for j in range(self.grid_size):
cell = self.grid.get(i, j)
type = cell.type if cell is not None else ''
counts[type] = counts.get(type | # # ignore second channel since redundant (due to one-to-one mapping btwn color and obj type for now)
# img = np.concatenate([np.eye(len(self.object_to_idx))[ch].transpose(2, 0, 1) for ch in img[:1]])
# full_img = np.concatenate([np.eye(len(self.object_to_idx))[ch].transpose(2, 0, 1) for ch in full_img[:1]]) | random_line_split |
getfood_base.py | mine = 4
def __init__(self,
agent_start_pos=(1, 1),
health_cap=50,
food_rate=4,
grid_size=8,
obs_vision=False,
reward_type='delta',
fully_observed=False,
only_partial_obs=False,
can_die=True,
one_hot_obs=True,
mixing_time_periods=[],
mixing_time_period_length=120,
**kwargs
):
self.agent_start_pos = agent_start_pos
# self.agent_start_dir = agent_start_dir
self.food_rate = food_rate
self.health_cap = health_cap
self.health = health_cap
self.last_health = self.health
self.obs_vision = obs_vision
self.reward_type = reward_type
self.fully_observed = fully_observed
self.only_partial_obs = only_partial_obs
self.can_die = can_die
self.one_hot_obs = one_hot_obs
# for conditional entropy of s' | s
self.transition_count = {}
self.prev_obs_string = ''
# for mixing time
self.mixing_time_periods = mixing_time_periods
self.max_mixing_time_period = max(mixing_time_periods) if mixing_time_periods else 0
self.mixing_time_period_length = mixing_time_period_length
self.obs_counts = []
if not hasattr(self, 'actions'):
self.actions = FoodEnvBase.Actions
super().__init__(
# Set this to True for maximum speed
see_through_walls=True,
grid_size=grid_size,
**kwargs
)
def _reward(self):
if self.reward_type == 'survival':
rwd = 1
elif self.reward_type == 'delta':
rwd = self.health - self.last_health
elif self.reward_type == 'health':
rwd = self.health
else:
assert False, "Reward type not matched"
self.last_health = self.health
return rwd
def _gen_grid(self, width, height):
# Create an empty grid
self.grid = GridAbsolute(width, height)
# Generate the surrounding walls
self.grid.wall_rect(0, 0, width, height)
# Place a goal square in the bottom-right corner
# self.grid.set(width - 2, height - 2, Goal())
self.extra_gen_grid()
# Place the agent
if self.agent_start_pos is not None:
self.start_pos = self.agent_start_pos
# self.start_dir = self.agent_start_dir
else:
self.place_agent()
self.mission = None
def step(self, action, incl_health=True):
done = False
matched = super().step(action, override=True)
# subclass-defined extra actions. if not caught by that, then unknown action
if not self.extra_step(action, matched):
assert False, "unknown action %d" % action
# decrease health bar
self.decay_health()
# generate new food
self.place_items()
# generate obs after action is caught and food is placed. generate reward before death check
img = self.get_img(onehot=self.one_hot_obs)
full_img = self.get_full_img(scale=1 if self.fully_observed else 1 / 8, onehot=self.one_hot_obs)
# NOTE: below not nec due to onehot being passed into two func calls above. but leaving here for now in case.
# if self.one_hot_obs:
# # ignore second channel since redundant (due to one-to-one mapping btwn color and obj type for now)
# img = np.concatenate([np.eye(len(self.object_to_idx))[ch].transpose(2, 0, 1) for ch in img[:1]])
# full_img = np.concatenate([np.eye(len(self.object_to_idx))[ch].transpose(2, 0, 1) for ch in full_img[:1]])
rwd = self._reward()
# tick on each grid item
to_remove = []
for j in range(0, self.grid.height):
for i in range(0, self.grid.width):
cell = self.grid.get(i, j)
if cell is not None:
if not cell.step():
self.dead_obj(i, j, cell)
to_remove.append((i, j))
for idxs in to_remove:
self.grid.set(*idxs, None)
# dead.
if self.dead():
done = True
if self.fully_observed:
if incl_health:
obs = np.concatenate((full_img.flatten(), np.array([self.health]), np.array(self.agent_pos)))
else:
obs = np.concatenate((full_img.flatten(), np.array(self.agent_pos)))
elif self.only_partial_obs:
if incl_health:
obs = np.concatenate((img.flatten(), np.array([self.health])))
else:
obs = img.flatten()
else:
if incl_health:
obs = np.concatenate((img.flatten(), full_img.flatten(), np.array([self.health])))
else:
obs = np.concatenate((img.flatten(), full_img.flatten()))
obs_string = obs.tostring()
# transition count stuff
self.transition_count.setdefault(hash(self.prev_obs_string), {})
self.transition_count[hash(self.prev_obs_string)][hash(obs_string)] = 1 + self.transition_count[hash(self.prev_obs_string)].get(hash(obs_string), 0)
# mixing time stuff
if self.step_count % self.mixing_time_period_length == 0:
self.obs_counts.append(self.obs_count.copy())
if hasattr(self, 'obs_count') and self.mixing_time_periods and len(self.obs_counts) > self.max_mixing_time_period:
self.obs_counts = self.obs_counts[-(self.max_mixing_time_period+1):]
self.prev_obs_string = obs_string
return obs, rwd, done, {}
def reset(self, incl_health=True):
super().reset()
self.health = self.health_cap
self.extra_reset()
img = self.get_img(onehot=self.one_hot_obs)
full_img = self.get_full_img(onehot=self.one_hot_obs)
self.transition_count = {}
# if self.one_hot_obs:
# # ignore second channel since redundant (due to one-to-one mapping btwn color and obj type for now)
# img = np.concatenate([np.eye(len(self.object_to_idx))[ch].transpose(2, 0, 1) for ch in img[:1]])
# full_img = np.concatenate([np.eye(len(self.object_to_idx))[ch].transpose(2, 0, 1) for ch in full_img[:1]])
if self.fully_observed:
if incl_health:
obs = np.concatenate((full_img.flatten(), np.array([self.health]), np.array(self.agent_pos)))
else:
obs = np.concatenate((full_img.flatten(), np.array(self.agent_pos)))
elif self.only_partial_obs:
if incl_health:
obs = np.concatenate((img.flatten(), np.array([self.health])))
else:
obs = img.flatten()
else:
if incl_health:
obs = np.concatenate((img.flatten(), full_img.flatten(), np.array([self.health])))
else:
obs = np.concatenate((img.flatten(), full_img.flatten()))
self.prev_obs_string = obs.tostring()
return obs
def get_full_img(self, scale=1 / 8, onehot=False):
""" Return the whole grid view """
if self.obs_vision:
full_img = self.get_full_obs_render(scale=scale)
else:
full_img = self.grid.encode(self, onehot=onehot)
# NOTE: in case need to scale here instead of in above func call: return cv2.resize(full_img, (0, 0), fx=0.125, fy=0.125, interpolation=cv2.INTER_AREA)
return full_img
def get_img(self, onehot=False):
""" Return the agent view """
if self.obs_vision:
img = self.gen_obs(onehot=False)
img = self.get_obs_render(img, CELL_PIXELS // 4)
else:
img = self.gen_obs(onehot=onehot)
return img
def extra_step(self, action, matched):
return matched
def extra_reset(self):
|
def place_items(self):
pass
def extra_gen_grid(self):
pass
def place_prob(self, obj, prob, top=None, size=None):
if np.random.binomial(1, prob):
pos = self.place_obj(obj, top, size)
obj.cur_pos = pos
return True
return False
def decay_health(self):
self.add_health(-1)
def add_health(self, num):
# clip health between 0 and cap after adjustment
self.health = max(0, min(self.health_cap, self.health + num))
def count_type(self, type):
count = 0
for i in range(self.grid_size):
for j in range(self.grid_size):
cell = self.grid.get(i, j)
if type is None and cell is None or cell is not None and cell.type == type:
count += 1
return count
def count_all_types(self):
counts = {}
for i in range(self.grid_size):
for j in range(self.grid_size):
cell = self.grid.get(i, j)
type = cell.type if cell is not None else ''
counts[type] = counts.get(type | pass | identifier_body |
codeHost.ts | without bleeding CSS to other code hosts (GitLab also uses .file-actions elements).
fileActions.classList.add('sg-github-file-actions')
// Old GitHub Enterprise PR views have a "☑ show comments" text that we want to insert *after*
const showCommentsElement = codeView.querySelector('.show-file-notes')
if (showCommentsElement) {
showCommentsElement.after(mountElement)
} else {
fileActions.prepend(mountElement)
}
return mountElement
}
const toolbarButtonProps = {
className: 'btn btn-sm tooltipped tooltipped-s',
}
const diffCodeView: Omit<CodeView, 'element'> = {
dom: diffDomFunctions,
getToolbarMount: createFileActionsToolbarMount,
resolveFileInfo: resolveDiffFileInfo,
toolbarButtonProps,
getScrollBoundaries: codeView => {
const fileHeader = codeView.querySelector<HTMLElement>('.file-header')
if (!fileHeader) {
throw new Error('Could not find .file-header element in GitHub PR code view')
}
return [fileHeader]
},
}
const diffConversationCodeView: Omit<CodeView, 'element'> = {
...diffCodeView,
getToolbarMount: undefined,
}
const singleFileCodeView: Omit<CodeView, 'element'> = {
dom: singleFileDOMFunctions,
getToolbarMount: createFileActionsToolbarMount,
resolveFileInfo,
toolbarButtonProps,
getSelections: getSelectionsFromHash,
observeSelections: observeSelectionsFromHash,
}
/**
* Some code snippets get leading white space trimmed. This adjusts based on
* this. See an example here https://github.com/sourcegraph/browser-extensions/issues/188.
*/
const getSnippetPositionAdjuster = (
requestGraphQL: PlatformContext['requestGraphQL']
): PositionAdjuster<RepoSpec & RevisionSpec & FileSpec & ResolvedRevisionSpec> => ({ direction, codeView, position }) =>
fetchBlobContentLines({ ...position, requestGraphQL }).pipe(
map(lines => {
const codeElement = singleFileDOMFunctions.getCodeElementFromLineNumber(
codeView,
position.line,
position.part
)
if (!codeElement) {
| const actualLine = lines[position.line - 1]
const documentLine = codeElement.textContent || ''
const actualLeadingWhiteSpace = actualLine.length - trimStart(actualLine).length
const documentLeadingWhiteSpace = documentLine.length - trimStart(documentLine).length
const modifier = direction === AdjustmentDirection.ActualToCodeView ? -1 : 1
const delta = Math.abs(actualLeadingWhiteSpace - documentLeadingWhiteSpace) * modifier
return {
line: position.line,
character: position.character + delta,
}
})
)
const searchResultCodeViewResolver = toCodeViewResolver('.code-list-item', {
dom: searchCodeSnippetDOMFunctions,
getPositionAdjuster: getSnippetPositionAdjuster,
resolveFileInfo: resolveSnippetFileInfo,
toolbarButtonProps,
})
const snippetCodeView: Omit<CodeView, 'element'> = {
dom: singleFileDOMFunctions,
resolveFileInfo: resolveSnippetFileInfo,
getPositionAdjuster: getSnippetPositionAdjuster,
}
export const createFileLineContainerToolbarMount: NonNullable<CodeView['getToolbarMount']> = (
codeViewElement: HTMLElement
): HTMLElement => {
const className = 'sourcegraph-github-file-code-view-toolbar-mount'
const existingMount = codeViewElement.querySelector(`.${className}`) as HTMLElement
if (existingMount) {
return existingMount
}
const mountElement = document.createElement('div')
mountElement.style.display = 'inline-flex'
mountElement.style.verticalAlign = 'middle'
mountElement.style.alignItems = 'center'
mountElement.className = className
const rawURLLink = codeViewElement.querySelector('#raw-url')
const buttonGroup = rawURLLink?.closest('.BtnGroup')
if (!buttonGroup?.parentNode) {
throw new Error('File actions not found')
}
buttonGroup.parentNode.insertBefore(mountElement, buttonGroup)
return mountElement
}
/**
* Matches the modern single-file code view, or snippets embedded in comments.
*
*/
export const fileLineContainerResolver: ViewResolver<CodeView> = {
selector: '.js-file-line-container',
resolveView: (fileLineContainer: HTMLElement): CodeView | null => {
const embeddedBlobWrapper = fileLineContainer.closest('.blob-wrapper-embedded')
if (embeddedBlobWrapper) {
// This is a snippet embedded in a comment.
// Resolve to `.blob-wrapper-embedded`'s parent element,
// the smallest element that contains both the code and
// the HTML anchor allowing to resolve the file info.
const element = embeddedBlobWrapper.parentElement!
return {
element,
...snippetCodeView,
}
}
const { pageType } = parseURL()
if (pageType !== 'blob') {
// this is not a single-file code view
return null
}
const repositoryContent = fileLineContainer.closest('.repository-content')
if (!repositoryContent) {
throw new Error('Could not find repository content element')
}
return {
element: repositoryContent as HTMLElement,
...singleFileCodeView,
getToolbarMount: createFileLineContainerToolbarMount,
}
},
}
const genericCodeViewResolver: ViewResolver<CodeView> = {
selector: target => {
const codeViews = new Set<HTMLElement>()
// Logic to support large diffs that are loaded asynchronously:
// https://github.com/sourcegraph/sourcegraph/issues/18337
// - Don't return `.file` elements that have yet to be loaded (loading is triggered by user)
// - When the user triggers diff loading, the mutation observer will tell us about
// .js-blob-wrapper, since the actual '.file' has been in the DOM the whole time. Return
// the closest ancestor '.file'
for (const file of querySelectorAllOrSelf<HTMLElement>(target, '.file')) {
if (file.querySelectorAll('.js-diff-load-container').length === 0) {
codeViews.add(file)
}
}
for (const blobWrapper of querySelectorAllOrSelf(target, '.js-blob-wrapper')) {
const file = blobWrapper.closest('.file')
if (file instanceof HTMLElement) {
codeViews.add(file)
}
}
return [...codeViews]
},
resolveView: (element: HTMLElement): CodeView | null => {
if (element.querySelector('article.markdown-body')) {
// This code view is rendered markdown, we shouldn't add code intelligence
return null
}
// This is a suggested change on a GitHub PR
if (element.closest('.js-suggested-changes-blob')) {
return null
}
const { pageType } = parseURL()
const isSingleCodeFile =
pageType === 'blob' &&
document.querySelectorAll('.file').length === 1 &&
document.querySelectorAll('.diff-view').length === 0
if (isSingleCodeFile) {
return { element, ...singleFileCodeView }
}
if (element.closest('.discussion-item-body') || element.classList.contains('js-comment-container')) {
// This code view is embedded on a PR conversation page.
return { element, ...diffConversationCodeView }
}
return { element, ...diffCodeView }
},
}
/**
* Returns true if the current page is GitHub Enterprise.
*/
export function checkIsGitHubEnterprise(): boolean {
const ogSiteName = document.head.querySelector<HTMLMetaElement>('meta[property="og:site_name"]')
return (
!!ogSiteName &&
// GitHub Enterprise v2.14.11 has "GitHub" as og:site_name
(ogSiteName.content === 'GitHub Enterprise' || ogSiteName.content === 'GitHub') &&
document.body.classList.contains('enterprise')
)
}
/**
* Returns true if the current page is github.com.
*/
export const checkIsGitHubDotCom = (url = window.location.href): boolean => /^https?:\/\/(www\.)?github\.com/.test(url)
/**
* Returns true if the current page is either github.com or GitHub Enterprise.
*/
export const checkIsGitHub = (): boolean => checkIsGitHubDotCom() || checkIsGitHubEnterprise()
const OPEN_ON_SOURCEGRAPH_ID = 'open-on-sourcegraph'
export const createOpenOnSourcegraphIfNotExists: MountGetter = (container: HTMLElement): HTMLElement | null => {
const pageheadActions = querySelectorOrSelf(container, '.pagehead-actions')
// If ran on page that isn't under a repository namespace.
if (!pageheadActions || pageheadActions.children.length === 0) {
return null
}
// Check for existing
let mount = pageheadActions.querySelector<HTMLElement>('#' + OPEN_ON_SOURCEGRAPH_ID)
if (mount) {
return mount
}
// Create new
mount = document.createElement('li')
mount.id = OPEN_ON_SOURCEGRAPH_ID
pageheadActions.prepend(mount)
return mount
}
const nativeTooltipResolver: ViewResolver<NativeTooltip> = {
selector: '.js-tagsearch-popover',
resolveView: element => ({ element }),
}
const iconClassName = 'icon--github v-align-text-bottom'
const notificationClassNames = {
[NotificationType.Log]: 'flash',
[NotificationType.Success]: 'flash flash-success',
[NotificationType.Info]: 'flash',
[Notification | throw new Error('(adjustPosition) could not find code element for line provided')
}
| conditional_block |
codeHost.ts | without bleeding CSS to other code hosts (GitLab also uses .file-actions elements).
fileActions.classList.add('sg-github-file-actions')
// Old GitHub Enterprise PR views have a "☑ show comments" text that we want to insert *after*
const showCommentsElement = codeView.querySelector('.show-file-notes')
if (showCommentsElement) {
showCommentsElement.after(mountElement)
} else {
fileActions.prepend(mountElement)
}
return mountElement
}
const toolbarButtonProps = {
className: 'btn btn-sm tooltipped tooltipped-s',
}
const diffCodeView: Omit<CodeView, 'element'> = {
dom: diffDomFunctions,
getToolbarMount: createFileActionsToolbarMount,
resolveFileInfo: resolveDiffFileInfo,
toolbarButtonProps,
getScrollBoundaries: codeView => {
const fileHeader = codeView.querySelector<HTMLElement>('.file-header')
if (!fileHeader) {
throw new Error('Could not find .file-header element in GitHub PR code view')
}
return [fileHeader]
},
}
const diffConversationCodeView: Omit<CodeView, 'element'> = {
...diffCodeView,
getToolbarMount: undefined,
}
const singleFileCodeView: Omit<CodeView, 'element'> = {
dom: singleFileDOMFunctions,
getToolbarMount: createFileActionsToolbarMount,
resolveFileInfo,
toolbarButtonProps,
getSelections: getSelectionsFromHash,
observeSelections: observeSelectionsFromHash,
}
/**
* Some code snippets get leading white space trimmed. This adjusts based on
* this. See an example here https://github.com/sourcegraph/browser-extensions/issues/188.
*/
const getSnippetPositionAdjuster = (
requestGraphQL: PlatformContext['requestGraphQL']
): PositionAdjuster<RepoSpec & RevisionSpec & FileSpec & ResolvedRevisionSpec> => ({ direction, codeView, position }) =>
fetchBlobContentLines({ ...position, requestGraphQL }).pipe(
map(lines => {
const codeElement = singleFileDOMFunctions.getCodeElementFromLineNumber(
codeView,
position.line,
position.part
)
if (!codeElement) {
throw new Error('(adjustPosition) could not find code element for line provided')
}
const actualLine = lines[position.line - 1]
const documentLine = codeElement.textContent || ''
const actualLeadingWhiteSpace = actualLine.length - trimStart(actualLine).length
const documentLeadingWhiteSpace = documentLine.length - trimStart(documentLine).length
const modifier = direction === AdjustmentDirection.ActualToCodeView ? -1 : 1
const delta = Math.abs(actualLeadingWhiteSpace - documentLeadingWhiteSpace) * modifier
return {
line: position.line,
character: position.character + delta,
}
})
)
const searchResultCodeViewResolver = toCodeViewResolver('.code-list-item', {
dom: searchCodeSnippetDOMFunctions,
getPositionAdjuster: getSnippetPositionAdjuster,
resolveFileInfo: resolveSnippetFileInfo,
toolbarButtonProps,
})
const snippetCodeView: Omit<CodeView, 'element'> = {
dom: singleFileDOMFunctions,
resolveFileInfo: resolveSnippetFileInfo,
getPositionAdjuster: getSnippetPositionAdjuster,
}
export const createFileLineContainerToolbarMount: NonNullable<CodeView['getToolbarMount']> = (
codeViewElement: HTMLElement
): HTMLElement => {
const className = 'sourcegraph-github-file-code-view-toolbar-mount'
const existingMount = codeViewElement.querySelector(`.${className}`) as HTMLElement
if (existingMount) {
return existingMount
}
const mountElement = document.createElement('div')
mountElement.style.display = 'inline-flex'
mountElement.style.verticalAlign = 'middle'
mountElement.style.alignItems = 'center'
mountElement.className = className
const rawURLLink = codeViewElement.querySelector('#raw-url')
const buttonGroup = rawURLLink?.closest('.BtnGroup')
if (!buttonGroup?.parentNode) {
throw new Error('File actions not found')
}
buttonGroup.parentNode.insertBefore(mountElement, buttonGroup)
return mountElement
}
/**
* Matches the modern single-file code view, or snippets embedded in comments.
*
*/
export const fileLineContainerResolver: ViewResolver<CodeView> = {
selector: '.js-file-line-container',
resolveView: (fileLineContainer: HTMLElement): CodeView | null => {
const embeddedBlobWrapper = fileLineContainer.closest('.blob-wrapper-embedded')
if (embeddedBlobWrapper) {
// This is a snippet embedded in a comment.
// Resolve to `.blob-wrapper-embedded`'s parent element,
// the smallest element that contains both the code and
// the HTML anchor allowing to resolve the file info.
const element = embeddedBlobWrapper.parentElement!
return {
element,
...snippetCodeView,
}
}
const { pageType } = parseURL()
if (pageType !== 'blob') {
// this is not a single-file code view
return null
}
const repositoryContent = fileLineContainer.closest('.repository-content')
if (!repositoryContent) {
throw new Error('Could not find repository content element')
}
return {
element: repositoryContent as HTMLElement,
...singleFileCodeView,
getToolbarMount: createFileLineContainerToolbarMount,
}
},
}
const genericCodeViewResolver: ViewResolver<CodeView> = {
selector: target => {
const codeViews = new Set<HTMLElement>()
// Logic to support large diffs that are loaded asynchronously:
// https://github.com/sourcegraph/sourcegraph/issues/18337
// - Don't return `.file` elements that have yet to be loaded (loading is triggered by user)
// - When the user triggers diff loading, the mutation observer will tell us about
// .js-blob-wrapper, since the actual '.file' has been in the DOM the whole time. Return
// the closest ancestor '.file'
for (const file of querySelectorAllOrSelf<HTMLElement>(target, '.file')) {
if (file.querySelectorAll('.js-diff-load-container').length === 0) {
codeViews.add(file)
}
}
for (const blobWrapper of querySelectorAllOrSelf(target, '.js-blob-wrapper')) {
const file = blobWrapper.closest('.file')
if (file instanceof HTMLElement) {
codeViews.add(file)
}
}
return [...codeViews]
},
resolveView: (element: HTMLElement): CodeView | null => {
if (element.querySelector('article.markdown-body')) {
// This code view is rendered markdown, we shouldn't add code intelligence
return null
}
// This is a suggested change on a GitHub PR
if (element.closest('.js-suggested-changes-blob')) {
return null
}
const { pageType } = parseURL()
const isSingleCodeFile =
pageType === 'blob' &&
document.querySelectorAll('.file').length === 1 &&
document.querySelectorAll('.diff-view').length === 0
if (isSingleCodeFile) {
return { element, ...singleFileCodeView }
}
if (element.closest('.discussion-item-body') || element.classList.contains('js-comment-container')) {
// This code view is embedded on a PR conversation page.
return { element, ...diffConversationCodeView }
}
return { element, ...diffCodeView }
},
}
/**
* Returns true if the current page is GitHub Enterprise.
*/
export function checkIsGitHubEnterprise(): boolean {
| /**
* Returns true if the current page is github.com.
*/
export const checkIsGitHubDotCom = (url = window.location.href): boolean => /^https?:\/\/(www\.)?github\.com/.test(url)
/**
* Returns true if the current page is either github.com or GitHub Enterprise.
*/
export const checkIsGitHub = (): boolean => checkIsGitHubDotCom() || checkIsGitHubEnterprise()
const OPEN_ON_SOURCEGRAPH_ID = 'open-on-sourcegraph'
export const createOpenOnSourcegraphIfNotExists: MountGetter = (container: HTMLElement): HTMLElement | null => {
const pageheadActions = querySelectorOrSelf(container, '.pagehead-actions')
// If ran on page that isn't under a repository namespace.
if (!pageheadActions || pageheadActions.children.length === 0) {
return null
}
// Check for existing
let mount = pageheadActions.querySelector<HTMLElement>('#' + OPEN_ON_SOURCEGRAPH_ID)
if (mount) {
return mount
}
// Create new
mount = document.createElement('li')
mount.id = OPEN_ON_SOURCEGRAPH_ID
pageheadActions.prepend(mount)
return mount
}
const nativeTooltipResolver: ViewResolver<NativeTooltip> = {
selector: '.js-tagsearch-popover',
resolveView: element => ({ element }),
}
const iconClassName = 'icon--github v-align-text-bottom'
const notificationClassNames = {
[NotificationType.Log]: 'flash',
[NotificationType.Success]: 'flash flash-success',
[NotificationType.Info]: 'flash',
[Notification | const ogSiteName = document.head.querySelector<HTMLMetaElement>('meta[property="og:site_name"]')
return (
!!ogSiteName &&
// GitHub Enterprise v2.14.11 has "GitHub" as og:site_name
(ogSiteName.content === 'GitHub Enterprise' || ogSiteName.content === 'GitHub') &&
document.body.classList.contains('enterprise')
)
}
| identifier_body |
codeHost.ts | (codeView: HTMLElement): HTMLElement {
const className = 'github-file-actions-toolbar-mount'
const existingMount = codeView.querySelector('.' + className) as HTMLElement
if (existingMount) {
return existingMount
}
const mountElement = document.createElement('div')
mountElement.className = className
const fileActions = codeView.querySelector('.file-actions')
if (!fileActions) {
throw new Error('Could not find GitHub file actions with selector .file-actions')
}
// Add a class to the .file-actions element, so that we can reliably match it in
// stylesheets without bleeding CSS to other code hosts (GitLab also uses .file-actions elements).
fileActions.classList.add('sg-github-file-actions')
// Old GitHub Enterprise PR views have a "☑ show comments" text that we want to insert *after*
const showCommentsElement = codeView.querySelector('.show-file-notes')
if (showCommentsElement) {
showCommentsElement.after(mountElement)
} else {
fileActions.prepend(mountElement)
}
return mountElement
}
const toolbarButtonProps = {
className: 'btn btn-sm tooltipped tooltipped-s',
}
const diffCodeView: Omit<CodeView, 'element'> = {
dom: diffDomFunctions,
getToolbarMount: createFileActionsToolbarMount,
resolveFileInfo: resolveDiffFileInfo,
toolbarButtonProps,
getScrollBoundaries: codeView => {
const fileHeader = codeView.querySelector<HTMLElement>('.file-header')
if (!fileHeader) {
throw new Error('Could not find .file-header element in GitHub PR code view')
}
return [fileHeader]
},
}
const diffConversationCodeView: Omit<CodeView, 'element'> = {
...diffCodeView,
getToolbarMount: undefined,
}
const singleFileCodeView: Omit<CodeView, 'element'> = {
dom: singleFileDOMFunctions,
getToolbarMount: createFileActionsToolbarMount,
resolveFileInfo,
toolbarButtonProps,
getSelections: getSelectionsFromHash,
observeSelections: observeSelectionsFromHash,
}
/**
* Some code snippets get leading white space trimmed. This adjusts based on
* this. See an example here https://github.com/sourcegraph/browser-extensions/issues/188.
*/
const getSnippetPositionAdjuster = (
requestGraphQL: PlatformContext['requestGraphQL']
): PositionAdjuster<RepoSpec & RevisionSpec & FileSpec & ResolvedRevisionSpec> => ({ direction, codeView, position }) =>
fetchBlobContentLines({ ...position, requestGraphQL }).pipe(
map(lines => {
const codeElement = singleFileDOMFunctions.getCodeElementFromLineNumber(
codeView,
position.line,
position.part
)
if (!codeElement) {
throw new Error('(adjustPosition) could not find code element for line provided')
}
const actualLine = lines[position.line - 1]
const documentLine = codeElement.textContent || ''
const actualLeadingWhiteSpace = actualLine.length - trimStart(actualLine).length
const documentLeadingWhiteSpace = documentLine.length - trimStart(documentLine).length
const modifier = direction === AdjustmentDirection.ActualToCodeView ? -1 : 1
const delta = Math.abs(actualLeadingWhiteSpace - documentLeadingWhiteSpace) * modifier
return {
line: position.line,
character: position.character + delta,
}
})
)
const searchResultCodeViewResolver = toCodeViewResolver('.code-list-item', {
dom: searchCodeSnippetDOMFunctions,
getPositionAdjuster: getSnippetPositionAdjuster,
resolveFileInfo: resolveSnippetFileInfo,
toolbarButtonProps,
})
const snippetCodeView: Omit<CodeView, 'element'> = {
dom: singleFileDOMFunctions,
resolveFileInfo: resolveSnippetFileInfo,
getPositionAdjuster: getSnippetPositionAdjuster,
}
export const createFileLineContainerToolbarMount: NonNullable<CodeView['getToolbarMount']> = (
codeViewElement: HTMLElement
): HTMLElement => {
const className = 'sourcegraph-github-file-code-view-toolbar-mount'
const existingMount = codeViewElement.querySelector(`.${className}`) as HTMLElement
if (existingMount) {
return existingMount
}
const mountElement = document.createElement('div')
mountElement.style.display = 'inline-flex'
mountElement.style.verticalAlign = 'middle'
mountElement.style.alignItems = 'center'
mountElement.className = className
const rawURLLink = codeViewElement.querySelector('#raw-url')
const buttonGroup = rawURLLink?.closest('.BtnGroup')
if (!buttonGroup?.parentNode) {
throw new Error('File actions not found')
}
buttonGroup.parentNode.insertBefore(mountElement, buttonGroup)
return mountElement
}
/**
* Matches the modern single-file code view, or snippets embedded in comments.
*
*/
export const fileLineContainerResolver: ViewResolver<CodeView> = {
selector: '.js-file-line-container',
resolveView: (fileLineContainer: HTMLElement): CodeView | null => {
const embeddedBlobWrapper = fileLineContainer.closest('.blob-wrapper-embedded')
if (embeddedBlobWrapper) {
// This is a snippet embedded in a comment.
// Resolve to `.blob-wrapper-embedded`'s parent element,
// the smallest element that contains both the code and
// the HTML anchor allowing to resolve the file info.
const element = embeddedBlobWrapper.parentElement!
return {
element,
...snippetCodeView,
}
}
const { pageType } = parseURL()
if (pageType !== 'blob') {
// this is not a single-file code view
return null
}
const repositoryContent = fileLineContainer.closest('.repository-content')
if (!repositoryContent) {
throw new Error('Could not find repository content element')
}
return {
element: repositoryContent as HTMLElement,
...singleFileCodeView,
getToolbarMount: createFileLineContainerToolbarMount,
}
},
}
const genericCodeViewResolver: ViewResolver<CodeView> = {
selector: target => {
const codeViews = new Set<HTMLElement>()
// Logic to support large diffs that are loaded asynchronously:
// https://github.com/sourcegraph/sourcegraph/issues/18337
// - Don't return `.file` elements that have yet to be loaded (loading is triggered by user)
// - When the user triggers diff loading, the mutation observer will tell us about
// .js-blob-wrapper, since the actual '.file' has been in the DOM the whole time. Return
// the closest ancestor '.file'
for (const file of querySelectorAllOrSelf<HTMLElement>(target, '.file')) {
if (file.querySelectorAll('.js-diff-load-container').length === 0) {
codeViews.add(file)
}
}
for (const blobWrapper of querySelectorAllOrSelf(target, '.js-blob-wrapper')) {
const file = blobWrapper.closest('.file')
if (file instanceof HTMLElement) {
codeViews.add(file)
}
}
return [...codeViews]
},
resolveView: (element: HTMLElement): CodeView | null => {
if (element.querySelector('article.markdown-body')) {
// This code view is rendered markdown, we shouldn't add code intelligence
return null
}
// This is a suggested change on a GitHub PR
if (element.closest('.js-suggested-changes-blob')) {
return null
}
const { pageType } = parseURL()
const isSingleCodeFile =
pageType === 'blob' &&
document.querySelectorAll('.file').length === 1 &&
document.querySelectorAll('.diff-view').length === 0
if (isSingleCodeFile) {
return { element, ...singleFileCodeView }
}
if (element.closest('.discussion-item-body') || element.classList.contains('js-comment-container')) {
// This code view is embedded on a PR conversation page.
return { element, ...diffConversationCodeView }
}
return { element, ...diffCodeView }
},
}
/**
* Returns true if the current page is GitHub Enterprise.
*/
export function checkIsGitHubEnterprise(): boolean {
const ogSiteName = document.head.querySelector<HTMLMetaElement>('meta[property="og:site_name"]')
return (
!!ogSiteName &&
// GitHub Enterprise v2.14.11 has "GitHub" as og:site_name
(ogSiteName.content === 'GitHub Enterprise' || ogSiteName.content === 'GitHub') &&
document.body.classList.contains('enterprise')
)
}
/**
* Returns true if the current page is github.com.
*/
export const checkIsGitHubDotCom = (url = window.location.href): boolean => /^https?:\/\/(www\.)?github\.com/.test(url)
/**
* Returns true if the current page is either github.com or GitHub Enterprise.
*/
export const checkIsGitHub = (): boolean => checkIsGitHubDotCom() || checkIsGitHubEnterprise()
const OPEN_ON_SOURCEGRAPH_ID = 'open-on-sourcegraph'
export const createOpenOnSourcegraphIfNotExists: MountGetter = (container: HTMLElement): HTMLElement | null => {
const pageheadActions = querySelectorOrSelf(container, '.pagehead-actions')
// If ran on page that isn't under a repository namespace.
if (!pageheadActions || pageheadActions.children.length === 0) {
return null
}
// Check for existing
let mount = pageheadActions.querySelector<HTMLElement>('#' + OPEN_ON_SOURCEGRAPH_ID)
| createFileActionsToolbarMount | identifier_name |
|
codeHost.ts | ',
resolveView: (fileLineContainer: HTMLElement): CodeView | null => {
const embeddedBlobWrapper = fileLineContainer.closest('.blob-wrapper-embedded')
if (embeddedBlobWrapper) {
// This is a snippet embedded in a comment.
// Resolve to `.blob-wrapper-embedded`'s parent element,
// the smallest element that contains both the code and
// the HTML anchor allowing to resolve the file info.
const element = embeddedBlobWrapper.parentElement!
return {
element,
...snippetCodeView,
}
}
const { pageType } = parseURL()
if (pageType !== 'blob') {
// this is not a single-file code view
return null
}
const repositoryContent = fileLineContainer.closest('.repository-content')
if (!repositoryContent) {
throw new Error('Could not find repository content element')
}
return {
element: repositoryContent as HTMLElement,
...singleFileCodeView,
getToolbarMount: createFileLineContainerToolbarMount,
}
},
}
const genericCodeViewResolver: ViewResolver<CodeView> = {
selector: target => {
const codeViews = new Set<HTMLElement>()
// Logic to support large diffs that are loaded asynchronously:
// https://github.com/sourcegraph/sourcegraph/issues/18337
// - Don't return `.file` elements that have yet to be loaded (loading is triggered by user)
// - When the user triggers diff loading, the mutation observer will tell us about
// .js-blob-wrapper, since the actual '.file' has been in the DOM the whole time. Return
// the closest ancestor '.file'
for (const file of querySelectorAllOrSelf<HTMLElement>(target, '.file')) {
if (file.querySelectorAll('.js-diff-load-container').length === 0) {
codeViews.add(file)
}
}
for (const blobWrapper of querySelectorAllOrSelf(target, '.js-blob-wrapper')) {
const file = blobWrapper.closest('.file')
if (file instanceof HTMLElement) {
codeViews.add(file)
}
}
return [...codeViews]
},
resolveView: (element: HTMLElement): CodeView | null => {
if (element.querySelector('article.markdown-body')) {
// This code view is rendered markdown, we shouldn't add code intelligence
return null
}
// This is a suggested change on a GitHub PR
if (element.closest('.js-suggested-changes-blob')) {
return null
}
const { pageType } = parseURL()
const isSingleCodeFile =
pageType === 'blob' &&
document.querySelectorAll('.file').length === 1 &&
document.querySelectorAll('.diff-view').length === 0
if (isSingleCodeFile) {
return { element, ...singleFileCodeView }
}
if (element.closest('.discussion-item-body') || element.classList.contains('js-comment-container')) {
// This code view is embedded on a PR conversation page.
return { element, ...diffConversationCodeView }
}
return { element, ...diffCodeView }
},
}
/**
* Returns true if the current page is GitHub Enterprise.
*/
export function checkIsGitHubEnterprise(): boolean {
const ogSiteName = document.head.querySelector<HTMLMetaElement>('meta[property="og:site_name"]')
return (
!!ogSiteName &&
// GitHub Enterprise v2.14.11 has "GitHub" as og:site_name
(ogSiteName.content === 'GitHub Enterprise' || ogSiteName.content === 'GitHub') &&
document.body.classList.contains('enterprise')
)
}
/**
* Returns true if the current page is github.com.
*/
export const checkIsGitHubDotCom = (url = window.location.href): boolean => /^https?:\/\/(www\.)?github\.com/.test(url)
/**
* Returns true if the current page is either github.com or GitHub Enterprise.
*/
export const checkIsGitHub = (): boolean => checkIsGitHubDotCom() || checkIsGitHubEnterprise()
const OPEN_ON_SOURCEGRAPH_ID = 'open-on-sourcegraph'
export const createOpenOnSourcegraphIfNotExists: MountGetter = (container: HTMLElement): HTMLElement | null => {
const pageheadActions = querySelectorOrSelf(container, '.pagehead-actions')
// If ran on page that isn't under a repository namespace.
if (!pageheadActions || pageheadActions.children.length === 0) {
return null
}
// Check for existing
let mount = pageheadActions.querySelector<HTMLElement>('#' + OPEN_ON_SOURCEGRAPH_ID)
if (mount) {
return mount
}
// Create new
mount = document.createElement('li')
mount.id = OPEN_ON_SOURCEGRAPH_ID
pageheadActions.prepend(mount)
return mount
}
const nativeTooltipResolver: ViewResolver<NativeTooltip> = {
selector: '.js-tagsearch-popover',
resolveView: element => ({ element }),
}
const iconClassName = 'icon--github v-align-text-bottom'
const notificationClassNames = {
[NotificationType.Log]: 'flash',
[NotificationType.Success]: 'flash flash-success',
[NotificationType.Info]: 'flash',
[NotificationType.Warning]: 'flash flash-warn',
[NotificationType.Error]: 'flash flash-error',
}
export const githubCodeHost: CodeHost = {
type: 'github',
name: checkIsGitHubEnterprise() ? 'GitHub Enterprise' : 'GitHub',
codeViewResolvers: [genericCodeViewResolver, fileLineContainerResolver, searchResultCodeViewResolver],
contentViewResolvers: [markdownBodyViewResolver],
nativeTooltipResolvers: [nativeTooltipResolver],
getContext: () => {
const repoHeaderHasPrivateMarker =
!!document.querySelector('.repohead .private') ||
!!document.querySelector('h1 .octicon-lock ~ [itemprop="author"] ~ [itemprop="name"]') ||
!!(
document
.querySelector('h1 [itemprop="author"] ~ [itemprop="name"] ~ .Label')
?.textContent?.trim()
.toLowerCase() === 'private'
)
const parsedURL = parseURL()
return {
...parsedURL,
revision:
parsedURL.pageType === 'blob' || parsedURL.pageType === 'tree'
? resolveFileInfo().blob.revision
: undefined,
privateRepository: window.location.hostname !== 'github.com' || repoHeaderHasPrivateMarker,
}
},
isLightTheme: defer(() => {
const mode = document.documentElement.dataset.colorMode as 'auto' | 'light' | 'dark' | undefined
if (mode === 'auto') {
return observeSystemIsLightTheme()
}
return of(mode !== 'dark')
}),
getViewContextOnSourcegraphMount: createOpenOnSourcegraphIfNotExists,
viewOnSourcegraphButtonClassProps: {
className: 'btn btn-sm tooltipped tooltipped-s',
iconClassName,
},
check: checkIsGitHub,
getCommandPaletteMount,
notificationClassNames,
commandPaletteClassProps: {
buttonClassName: 'Header-link',
popoverClassName: 'Box',
formClassName: 'p-1',
inputClassName: 'form-control input-sm header-search-input jump-to-field',
listClassName: 'p-0 m-0 js-navigation-container jump-to-suggestions-results-container',
selectedListItemClassName: 'navigation-focus',
listItemClassName:
'd-flex flex-justify-start flex-items-center p-0 f5 navigation-item js-navigation-item js-jump-to-scoped-search',
actionItemClassName:
'command-palette-action-item--github no-underline d-flex flex-auto flex-items-center jump-to-suggestions-path p-2',
noResultsClassName: 'd-flex flex-auto flex-items-center jump-to-suggestions-path p-2',
iconClassName,
},
codeViewToolbarClassProps: {
className: 'code-view-toolbar--github',
listItemClass: 'code-view-toolbar__item--github BtnGroup',
actionItemClass: 'btn btn-sm tooltipped tooltipped-s BtnGroup-item action-item--github',
actionItemPressedClass: 'selected',
actionItemIconClass: 'icon--github v-align-text-bottom',
},
hoverOverlayClassProps: {
className: 'Box',
actionItemClassName: 'btn btn-secondary',
actionItemPressedClassName: 'active',
closeButtonClassName: 'btn-octicon p-0 hover-overlay__close-button--github',
badgeClassName: 'label hover-overlay__badge--github',
getAlertClassName: createNotificationClassNameGetter(notificationClassNames, 'flash-full'),
iconClassName,
},
setElementTooltip,
linkPreviewContentClass: 'text-small text-gray p-1 mx-1 border rounded-1 bg-gray text-gray-dark',
urlToFile: (sourcegraphURL, target, context) => {
if (target.viewState) {
// A view state means that a panel must be shown, and panels are currently only supported on
// Sourcegraph (not code hosts).
return toAbsoluteBlobURL(sourcegraphURL, target)
}
// Make sure the location is also on this github instance, return an absolute URL otherwise.
const sameCodeHost = target.rawRepoName.startsWith(window.location.hostname)
if (!sameCodeHost) {
return toAbsoluteBlobURL(sourcegraphURL, target)
}
const revision = target.revision || 'HEAD'
// If we're provided options, we can make the j2d URL more specific.
const { rawRepoName } = parseURL() | random_line_split |
||
edgehub.rs | , BrokerBuilder, BrokerHandle, BrokerReady, BrokerSnapshot, FilePersistor,
MakeMqttPacketProcessor, Message, Persist, Server, ServerCertificate, SystemEvent,
VersionedFileFormat,
};
use mqtt_edgehub::{
auth::{
EdgeHubAuthenticator, EdgeHubAuthorizer, LocalAuthenticator, LocalAuthorizer,
PolicyAuthorizer,
},
command::{
AuthorizedIdentitiesCommand, BridgeUpdateCommand, CommandHandler, DisconnectCommand,
PolicyUpdateCommand,
},
connection::MakeEdgeHubPacketProcessor,
settings::Settings,
};
use super::{shutdown, Bootstrap};
const DEVICE_ID_ENV: &str = "IOTEDGE_DEVICEID";
const IOTHUB_HOSTNAME_ENV: &str = "IOTEDGE_IOTHUBHOSTNAME";
#[derive(Default)]
pub struct EdgeHubBootstrap {
broker_ready: BrokerReady,
}
#[async_trait]
impl Bootstrap for EdgeHubBootstrap {
type Settings = Settings;
fn load_config<P: AsRef<Path>>(&self, path: P) -> Result<Self::Settings> {
info!("loading settings from a file {}", path.as_ref().display());
Ok(Self::Settings::from_file(path)?)
}
type Authorizer = LocalAuthorizer<EdgeHubAuthorizer<PolicyAuthorizer>>;
async fn make_broker(
&self,
settings: &Self::Settings,
) -> Result<(Broker<Self::Authorizer>, FilePersistor<VersionedFileFormat>)> {
info!("loading state...");
let persistence_config = settings.broker().persistence();
let state_dir = persistence_config.file_path();
fs::create_dir_all(state_dir.clone())?;
let mut persistor = FilePersistor::new(state_dir, VersionedFileFormat::default());
let state = persistor.load().await?;
info!("state loaded.");
let device_id = env::var(DEVICE_ID_ENV).context(DEVICE_ID_ENV)?;
let iothub_id = env::var(IOTHUB_HOSTNAME_ENV).context(IOTHUB_HOSTNAME_ENV)?;
let authorizer = LocalAuthorizer::new(EdgeHubAuthorizer::new(
PolicyAuthorizer::new(device_id.clone(), self.broker_ready.handle()),
device_id,
iothub_id,
self.broker_ready.handle(),
));
let broker = BrokerBuilder::default()
.with_authorizer(authorizer)
.with_state(state.unwrap_or_default())
.with_config(settings.broker().clone())
.build();
Ok((broker, persistor))
}
fn snapshot_interval(&self, settings: &Self::Settings) -> StdDuration {
settings.broker().persistence().time_interval()
}
fn session_expiration(&self, settings: &Self::Settings) -> StdDuration {
settings.broker().session().expiration()
}
fn session_cleanup_interval(&self, settings: &Self::Settings) -> StdDuration {
settings.broker().session().cleanup_interval()
}
async fn run(
self,
config: Self::Settings,
broker: Broker<Self::Authorizer>,
) -> Result<BrokerSnapshot> {
let broker_handle = broker.handle();
let sidecars = make_sidecars(&broker_handle, &config)?;
info!("starting server...");
let server = make_server(config, broker, self.broker_ready).await?;
let shutdown_signal = shutdown_signal(&server);
let server = tokio::spawn(server.serve(shutdown_signal));
info!("starting sidecars...");
let mut shutdowns = Vec::new();
let mut sidecar_joins = Vec::new();
for sidecar in sidecars {
shutdowns.push(sidecar.shutdown_handle()?);
sidecar_joins.push(tokio::spawn(sidecar.run()));
}
let state = match future::select(server, future::select_all(sidecar_joins)).await {
// server exited first
Either::Left((snapshot, sidecars)) => {
// send shutdown event to each sidecar
let shutdowns = shutdowns.into_iter().map(SidecarShutdownHandle::shutdown);
future::join_all(shutdowns).await;
// awaits for at least one to finish
let (_res, _stopped, sidecars) = sidecars.await;
|
snapshot??
}
// one of sidecars exited first
Either::Right(((res, stopped, sidecars), server)) => {
debug!("a sidecar has stopped. shutting down all sidecars...");
if let Err(e) = res {
error!(message = "failed waiting for sidecar shutdown", error = %e);
}
// send shutdown event to each of the rest sidecars
shutdowns.remove(stopped);
let shutdowns = shutdowns.into_iter().map(SidecarShutdownHandle::shutdown);
future::join_all(shutdowns).await;
// wait for the rest to exit
future::join_all(sidecars).await;
// signal server
broker_handle.send(Message::System(SystemEvent::Shutdown))?;
server.await??
}
};
Ok(state)
}
}
async fn make_server<Z>(
config: Settings,
broker: Broker<Z>,
broker_ready: BrokerReady,
) -> Result<Server<Z, MakeEdgeHubPacketProcessor<MakeMqttPacketProcessor>>>
where
Z: Authorizer + Send + 'static,
{
let broker_handle = broker.handle();
let make_processor = MakeEdgeHubPacketProcessor::new_default(broker_handle.clone());
let mut server = Server::from_broker(broker).with_packet_processor(make_processor);
// Add system transport to allow communication between edgehub components
let authenticator = LocalAuthenticator::new();
server.with_tcp(config.listener().system().addr(), authenticator, None)?;
// Add regular MQTT over TCP transport
let authenticator = EdgeHubAuthenticator::new(config.auth().url());
if let Some(tcp) = config.listener().tcp() {
let broker_ready = Some(broker_ready.signal());
server.with_tcp(tcp.addr(), authenticator.clone(), broker_ready)?;
}
// Add regular MQTT over TLS transport
if let Some(tls) = config.listener().tls() {
let identity = if let Some(config) = tls.certificate() {
info!("loading identity from {}", config.cert_path().display());
ServerCertificate::from_pem(config.cert_path(), config.private_key_path())
.with_context(|| {
ServerCertificateLoadError::File(
config.cert_path().to_path_buf(),
config.private_key_path().to_path_buf(),
)
})?
} else {
info!("downloading identity from edgelet");
download_server_certificate()
.await
.with_context(|| ServerCertificateLoadError::Edgelet)?
};
let broker_ready = Some(broker_ready.signal());
server.with_tls(tls.addr(), identity, authenticator.clone(), broker_ready)?;
};
Ok(server)
}
fn make_sidecars(
broker_handle: &BrokerHandle,
config: &Settings,
) -> Result<Vec<Box<dyn Sidecar + Send>>> {
let mut sidecars: Vec<Box<dyn Sidecar + Send>> = Vec::new();
let system_address = config.listener().system().addr().to_string();
let device_id = env::var(DEVICE_ID_ENV).context(DEVICE_ID_ENV)?;
let settings = BridgeSettings::new()?;
let bridge_controller =
BridgeController::new(system_address.clone(), device_id.to_owned(), settings);
let bridge_controller_handle = bridge_controller.handle();
sidecars.push(Box::new(bridge_controller));
let mut command_handler = CommandHandler::new(system_address, &device_id);
command_handler.add_command(DisconnectCommand::new(&broker_handle));
command_handler.add_command(AuthorizedIdentitiesCommand::new(&broker_handle));
command_handler.add_command(PolicyUpdateCommand::new(broker_handle));
command_handler.add_command(BridgeUpdateCommand::new(bridge_controller_handle));
sidecars.push(Box::new(command_handler));
Ok(sidecars)
}
pub const WORKLOAD_URI: &str = "IOTEDGE_WORKLOADURI";
pub const EDGE_DEVICE_HOST_NAME: &str = "EdgeDeviceHostName";
pub const MODULE_ID: &str = "IOTEDGE_MODULEID";
pub const MODULE_GENERATION_ID: &str = "IOTEDGE_MODULEGENERATIONID";
pub const CERTIFICATE_VALIDITY_DAYS: i64 = 90;
async fn download_server_certificate() -> Result<ServerCertificate> {
let uri = env::var(WORKLOAD_URI).context(WORKLOAD_URI)?;
let hostname = env::var(EDGE_DEVICE_HOST_NAME).context(EDGE_DEVICE_HOST_NAME)?;
let module_id = env::var(MODULE_ID).context(MODULE_ID)?;
let generation_id = env::var(MODULE_GENERATION_ID).context(MODULE_GENERATION_ID)?;
let expiration = Utc::now() + Duration::days(CERTIFICATE_VALIDITY_DAYS);
let client = edgelet_client::workload(&uri)?;
let cert = client
.create_server_cert(&module_id, &generation_id, &hostname, expiration)
.await?;
if cert.private_key().type_() != "key" {
bail!(
"unknown type of private key: {}",
cert.private_key().type_()
);
}
if let Some(private_key) = cert.private_key().bytes() {
let identity = ServerCertificate::from_pem_pair(cert.certificate(), private_key)?;
Ok(identity)
} else {
bail!("missing private key");
}
}
fn shutdown_signal<Z, P>(server: &Server<Z, P | // wait for the rest to exit
future::join_all(sidecars).await; | random_line_split |
edgehub.rs | , BrokerBuilder, BrokerHandle, BrokerReady, BrokerSnapshot, FilePersistor,
MakeMqttPacketProcessor, Message, Persist, Server, ServerCertificate, SystemEvent,
VersionedFileFormat,
};
use mqtt_edgehub::{
auth::{
EdgeHubAuthenticator, EdgeHubAuthorizer, LocalAuthenticator, LocalAuthorizer,
PolicyAuthorizer,
},
command::{
AuthorizedIdentitiesCommand, BridgeUpdateCommand, CommandHandler, DisconnectCommand,
PolicyUpdateCommand,
},
connection::MakeEdgeHubPacketProcessor,
settings::Settings,
};
use super::{shutdown, Bootstrap};
const DEVICE_ID_ENV: &str = "IOTEDGE_DEVICEID";
const IOTHUB_HOSTNAME_ENV: &str = "IOTEDGE_IOTHUBHOSTNAME";
#[derive(Default)]
pub struct EdgeHubBootstrap {
broker_ready: BrokerReady,
}
#[async_trait]
impl Bootstrap for EdgeHubBootstrap {
type Settings = Settings;
fn load_config<P: AsRef<Path>>(&self, path: P) -> Result<Self::Settings> {
info!("loading settings from a file {}", path.as_ref().display());
Ok(Self::Settings::from_file(path)?)
}
type Authorizer = LocalAuthorizer<EdgeHubAuthorizer<PolicyAuthorizer>>;
async fn make_broker(
&self,
settings: &Self::Settings,
) -> Result<(Broker<Self::Authorizer>, FilePersistor<VersionedFileFormat>)> {
info!("loading state...");
let persistence_config = settings.broker().persistence();
let state_dir = persistence_config.file_path();
fs::create_dir_all(state_dir.clone())?;
let mut persistor = FilePersistor::new(state_dir, VersionedFileFormat::default());
let state = persistor.load().await?;
info!("state loaded.");
let device_id = env::var(DEVICE_ID_ENV).context(DEVICE_ID_ENV)?;
let iothub_id = env::var(IOTHUB_HOSTNAME_ENV).context(IOTHUB_HOSTNAME_ENV)?;
let authorizer = LocalAuthorizer::new(EdgeHubAuthorizer::new(
PolicyAuthorizer::new(device_id.clone(), self.broker_ready.handle()),
device_id,
iothub_id,
self.broker_ready.handle(),
));
let broker = BrokerBuilder::default()
.with_authorizer(authorizer)
.with_state(state.unwrap_or_default())
.with_config(settings.broker().clone())
.build();
Ok((broker, persistor))
}
fn snapshot_interval(&self, settings: &Self::Settings) -> StdDuration |
fn session_expiration(&self, settings: &Self::Settings) -> StdDuration {
settings.broker().session().expiration()
}
fn session_cleanup_interval(&self, settings: &Self::Settings) -> StdDuration {
settings.broker().session().cleanup_interval()
}
async fn run(
self,
config: Self::Settings,
broker: Broker<Self::Authorizer>,
) -> Result<BrokerSnapshot> {
let broker_handle = broker.handle();
let sidecars = make_sidecars(&broker_handle, &config)?;
info!("starting server...");
let server = make_server(config, broker, self.broker_ready).await?;
let shutdown_signal = shutdown_signal(&server);
let server = tokio::spawn(server.serve(shutdown_signal));
info!("starting sidecars...");
let mut shutdowns = Vec::new();
let mut sidecar_joins = Vec::new();
for sidecar in sidecars {
shutdowns.push(sidecar.shutdown_handle()?);
sidecar_joins.push(tokio::spawn(sidecar.run()));
}
let state = match future::select(server, future::select_all(sidecar_joins)).await {
// server exited first
Either::Left((snapshot, sidecars)) => {
// send shutdown event to each sidecar
let shutdowns = shutdowns.into_iter().map(SidecarShutdownHandle::shutdown);
future::join_all(shutdowns).await;
// awaits for at least one to finish
let (_res, _stopped, sidecars) = sidecars.await;
// wait for the rest to exit
future::join_all(sidecars).await;
snapshot??
}
// one of sidecars exited first
Either::Right(((res, stopped, sidecars), server)) => {
debug!("a sidecar has stopped. shutting down all sidecars...");
if let Err(e) = res {
error!(message = "failed waiting for sidecar shutdown", error = %e);
}
// send shutdown event to each of the rest sidecars
shutdowns.remove(stopped);
let shutdowns = shutdowns.into_iter().map(SidecarShutdownHandle::shutdown);
future::join_all(shutdowns).await;
// wait for the rest to exit
future::join_all(sidecars).await;
// signal server
broker_handle.send(Message::System(SystemEvent::Shutdown))?;
server.await??
}
};
Ok(state)
}
}
async fn make_server<Z>(
config: Settings,
broker: Broker<Z>,
broker_ready: BrokerReady,
) -> Result<Server<Z, MakeEdgeHubPacketProcessor<MakeMqttPacketProcessor>>>
where
Z: Authorizer + Send + 'static,
{
let broker_handle = broker.handle();
let make_processor = MakeEdgeHubPacketProcessor::new_default(broker_handle.clone());
let mut server = Server::from_broker(broker).with_packet_processor(make_processor);
// Add system transport to allow communication between edgehub components
let authenticator = LocalAuthenticator::new();
server.with_tcp(config.listener().system().addr(), authenticator, None)?;
// Add regular MQTT over TCP transport
let authenticator = EdgeHubAuthenticator::new(config.auth().url());
if let Some(tcp) = config.listener().tcp() {
let broker_ready = Some(broker_ready.signal());
server.with_tcp(tcp.addr(), authenticator.clone(), broker_ready)?;
}
// Add regular MQTT over TLS transport
if let Some(tls) = config.listener().tls() {
let identity = if let Some(config) = tls.certificate() {
info!("loading identity from {}", config.cert_path().display());
ServerCertificate::from_pem(config.cert_path(), config.private_key_path())
.with_context(|| {
ServerCertificateLoadError::File(
config.cert_path().to_path_buf(),
config.private_key_path().to_path_buf(),
)
})?
} else {
info!("downloading identity from edgelet");
download_server_certificate()
.await
.with_context(|| ServerCertificateLoadError::Edgelet)?
};
let broker_ready = Some(broker_ready.signal());
server.with_tls(tls.addr(), identity, authenticator.clone(), broker_ready)?;
};
Ok(server)
}
fn make_sidecars(
broker_handle: &BrokerHandle,
config: &Settings,
) -> Result<Vec<Box<dyn Sidecar + Send>>> {
let mut sidecars: Vec<Box<dyn Sidecar + Send>> = Vec::new();
let system_address = config.listener().system().addr().to_string();
let device_id = env::var(DEVICE_ID_ENV).context(DEVICE_ID_ENV)?;
let settings = BridgeSettings::new()?;
let bridge_controller =
BridgeController::new(system_address.clone(), device_id.to_owned(), settings);
let bridge_controller_handle = bridge_controller.handle();
sidecars.push(Box::new(bridge_controller));
let mut command_handler = CommandHandler::new(system_address, &device_id);
command_handler.add_command(DisconnectCommand::new(&broker_handle));
command_handler.add_command(AuthorizedIdentitiesCommand::new(&broker_handle));
command_handler.add_command(PolicyUpdateCommand::new(broker_handle));
command_handler.add_command(BridgeUpdateCommand::new(bridge_controller_handle));
sidecars.push(Box::new(command_handler));
Ok(sidecars)
}
pub const WORKLOAD_URI: &str = "IOTEDGE_WORKLOADURI";
pub const EDGE_DEVICE_HOST_NAME: &str = "EdgeDeviceHostName";
pub const MODULE_ID: &str = "IOTEDGE_MODULEID";
pub const MODULE_GENERATION_ID: &str = "IOTEDGE_MODULEGENERATIONID";
pub const CERTIFICATE_VALIDITY_DAYS: i64 = 90;
async fn download_server_certificate() -> Result<ServerCertificate> {
let uri = env::var(WORKLOAD_URI).context(WORKLOAD_URI)?;
let hostname = env::var(EDGE_DEVICE_HOST_NAME).context(EDGE_DEVICE_HOST_NAME)?;
let module_id = env::var(MODULE_ID).context(MODULE_ID)?;
let generation_id = env::var(MODULE_GENERATION_ID).context(MODULE_GENERATION_ID)?;
let expiration = Utc::now() + Duration::days(CERTIFICATE_VALIDITY_DAYS);
let client = edgelet_client::workload(&uri)?;
let cert = client
.create_server_cert(&module_id, &generation_id, &hostname, expiration)
.await?;
if cert.private_key().type_() != "key" {
bail!(
"unknown type of private key: {}",
cert.private_key().type_()
);
}
if let Some(private_key) = cert.private_key().bytes() {
let identity = ServerCertificate::from_pem_pair(cert.certificate(), private_key)?;
Ok(identity)
} else {
bail!("missing private key");
}
}
fn shutdown_signal<Z, P>(server: &Server<Z, | {
settings.broker().persistence().time_interval()
} | identifier_body |
edgehub.rs | , BrokerBuilder, BrokerHandle, BrokerReady, BrokerSnapshot, FilePersistor,
MakeMqttPacketProcessor, Message, Persist, Server, ServerCertificate, SystemEvent,
VersionedFileFormat,
};
use mqtt_edgehub::{
auth::{
EdgeHubAuthenticator, EdgeHubAuthorizer, LocalAuthenticator, LocalAuthorizer,
PolicyAuthorizer,
},
command::{
AuthorizedIdentitiesCommand, BridgeUpdateCommand, CommandHandler, DisconnectCommand,
PolicyUpdateCommand,
},
connection::MakeEdgeHubPacketProcessor,
settings::Settings,
};
use super::{shutdown, Bootstrap};
const DEVICE_ID_ENV: &str = "IOTEDGE_DEVICEID";
const IOTHUB_HOSTNAME_ENV: &str = "IOTEDGE_IOTHUBHOSTNAME";
#[derive(Default)]
pub struct EdgeHubBootstrap {
broker_ready: BrokerReady,
}
#[async_trait]
impl Bootstrap for EdgeHubBootstrap {
type Settings = Settings;
fn load_config<P: AsRef<Path>>(&self, path: P) -> Result<Self::Settings> {
info!("loading settings from a file {}", path.as_ref().display());
Ok(Self::Settings::from_file(path)?)
}
type Authorizer = LocalAuthorizer<EdgeHubAuthorizer<PolicyAuthorizer>>;
async fn make_broker(
&self,
settings: &Self::Settings,
) -> Result<(Broker<Self::Authorizer>, FilePersistor<VersionedFileFormat>)> {
info!("loading state...");
let persistence_config = settings.broker().persistence();
let state_dir = persistence_config.file_path();
fs::create_dir_all(state_dir.clone())?;
let mut persistor = FilePersistor::new(state_dir, VersionedFileFormat::default());
let state = persistor.load().await?;
info!("state loaded.");
let device_id = env::var(DEVICE_ID_ENV).context(DEVICE_ID_ENV)?;
let iothub_id = env::var(IOTHUB_HOSTNAME_ENV).context(IOTHUB_HOSTNAME_ENV)?;
let authorizer = LocalAuthorizer::new(EdgeHubAuthorizer::new(
PolicyAuthorizer::new(device_id.clone(), self.broker_ready.handle()),
device_id,
iothub_id,
self.broker_ready.handle(),
));
let broker = BrokerBuilder::default()
.with_authorizer(authorizer)
.with_state(state.unwrap_or_default())
.with_config(settings.broker().clone())
.build();
Ok((broker, persistor))
}
fn snapshot_interval(&self, settings: &Self::Settings) -> StdDuration {
settings.broker().persistence().time_interval()
}
fn session_expiration(&self, settings: &Self::Settings) -> StdDuration {
settings.broker().session().expiration()
}
fn | (&self, settings: &Self::Settings) -> StdDuration {
settings.broker().session().cleanup_interval()
}
async fn run(
self,
config: Self::Settings,
broker: Broker<Self::Authorizer>,
) -> Result<BrokerSnapshot> {
let broker_handle = broker.handle();
let sidecars = make_sidecars(&broker_handle, &config)?;
info!("starting server...");
let server = make_server(config, broker, self.broker_ready).await?;
let shutdown_signal = shutdown_signal(&server);
let server = tokio::spawn(server.serve(shutdown_signal));
info!("starting sidecars...");
let mut shutdowns = Vec::new();
let mut sidecar_joins = Vec::new();
for sidecar in sidecars {
shutdowns.push(sidecar.shutdown_handle()?);
sidecar_joins.push(tokio::spawn(sidecar.run()));
}
let state = match future::select(server, future::select_all(sidecar_joins)).await {
// server exited first
Either::Left((snapshot, sidecars)) => {
// send shutdown event to each sidecar
let shutdowns = shutdowns.into_iter().map(SidecarShutdownHandle::shutdown);
future::join_all(shutdowns).await;
// awaits for at least one to finish
let (_res, _stopped, sidecars) = sidecars.await;
// wait for the rest to exit
future::join_all(sidecars).await;
snapshot??
}
// one of sidecars exited first
Either::Right(((res, stopped, sidecars), server)) => {
debug!("a sidecar has stopped. shutting down all sidecars...");
if let Err(e) = res {
error!(message = "failed waiting for sidecar shutdown", error = %e);
}
// send shutdown event to each of the rest sidecars
shutdowns.remove(stopped);
let shutdowns = shutdowns.into_iter().map(SidecarShutdownHandle::shutdown);
future::join_all(shutdowns).await;
// wait for the rest to exit
future::join_all(sidecars).await;
// signal server
broker_handle.send(Message::System(SystemEvent::Shutdown))?;
server.await??
}
};
Ok(state)
}
}
async fn make_server<Z>(
config: Settings,
broker: Broker<Z>,
broker_ready: BrokerReady,
) -> Result<Server<Z, MakeEdgeHubPacketProcessor<MakeMqttPacketProcessor>>>
where
Z: Authorizer + Send + 'static,
{
let broker_handle = broker.handle();
let make_processor = MakeEdgeHubPacketProcessor::new_default(broker_handle.clone());
let mut server = Server::from_broker(broker).with_packet_processor(make_processor);
// Add system transport to allow communication between edgehub components
let authenticator = LocalAuthenticator::new();
server.with_tcp(config.listener().system().addr(), authenticator, None)?;
// Add regular MQTT over TCP transport
let authenticator = EdgeHubAuthenticator::new(config.auth().url());
if let Some(tcp) = config.listener().tcp() {
let broker_ready = Some(broker_ready.signal());
server.with_tcp(tcp.addr(), authenticator.clone(), broker_ready)?;
}
// Add regular MQTT over TLS transport
if let Some(tls) = config.listener().tls() {
let identity = if let Some(config) = tls.certificate() {
info!("loading identity from {}", config.cert_path().display());
ServerCertificate::from_pem(config.cert_path(), config.private_key_path())
.with_context(|| {
ServerCertificateLoadError::File(
config.cert_path().to_path_buf(),
config.private_key_path().to_path_buf(),
)
})?
} else {
info!("downloading identity from edgelet");
download_server_certificate()
.await
.with_context(|| ServerCertificateLoadError::Edgelet)?
};
let broker_ready = Some(broker_ready.signal());
server.with_tls(tls.addr(), identity, authenticator.clone(), broker_ready)?;
};
Ok(server)
}
fn make_sidecars(
broker_handle: &BrokerHandle,
config: &Settings,
) -> Result<Vec<Box<dyn Sidecar + Send>>> {
let mut sidecars: Vec<Box<dyn Sidecar + Send>> = Vec::new();
let system_address = config.listener().system().addr().to_string();
let device_id = env::var(DEVICE_ID_ENV).context(DEVICE_ID_ENV)?;
let settings = BridgeSettings::new()?;
let bridge_controller =
BridgeController::new(system_address.clone(), device_id.to_owned(), settings);
let bridge_controller_handle = bridge_controller.handle();
sidecars.push(Box::new(bridge_controller));
let mut command_handler = CommandHandler::new(system_address, &device_id);
command_handler.add_command(DisconnectCommand::new(&broker_handle));
command_handler.add_command(AuthorizedIdentitiesCommand::new(&broker_handle));
command_handler.add_command(PolicyUpdateCommand::new(broker_handle));
command_handler.add_command(BridgeUpdateCommand::new(bridge_controller_handle));
sidecars.push(Box::new(command_handler));
Ok(sidecars)
}
pub const WORKLOAD_URI: &str = "IOTEDGE_WORKLOADURI";
pub const EDGE_DEVICE_HOST_NAME: &str = "EdgeDeviceHostName";
pub const MODULE_ID: &str = "IOTEDGE_MODULEID";
pub const MODULE_GENERATION_ID: &str = "IOTEDGE_MODULEGENERATIONID";
pub const CERTIFICATE_VALIDITY_DAYS: i64 = 90;
async fn download_server_certificate() -> Result<ServerCertificate> {
let uri = env::var(WORKLOAD_URI).context(WORKLOAD_URI)?;
let hostname = env::var(EDGE_DEVICE_HOST_NAME).context(EDGE_DEVICE_HOST_NAME)?;
let module_id = env::var(MODULE_ID).context(MODULE_ID)?;
let generation_id = env::var(MODULE_GENERATION_ID).context(MODULE_GENERATION_ID)?;
let expiration = Utc::now() + Duration::days(CERTIFICATE_VALIDITY_DAYS);
let client = edgelet_client::workload(&uri)?;
let cert = client
.create_server_cert(&module_id, &generation_id, &hostname, expiration)
.await?;
if cert.private_key().type_() != "key" {
bail!(
"unknown type of private key: {}",
cert.private_key().type_()
);
}
if let Some(private_key) = cert.private_key().bytes() {
let identity = ServerCertificate::from_pem_pair(cert.certificate(), private_key)?;
Ok(identity)
} else {
bail!("missing private key");
}
}
fn shutdown_signal<Z, P>(server: &Server<Z, P | session_cleanup_interval | identifier_name |
edgehub.rs | , BrokerBuilder, BrokerHandle, BrokerReady, BrokerSnapshot, FilePersistor,
MakeMqttPacketProcessor, Message, Persist, Server, ServerCertificate, SystemEvent,
VersionedFileFormat,
};
use mqtt_edgehub::{
auth::{
EdgeHubAuthenticator, EdgeHubAuthorizer, LocalAuthenticator, LocalAuthorizer,
PolicyAuthorizer,
},
command::{
AuthorizedIdentitiesCommand, BridgeUpdateCommand, CommandHandler, DisconnectCommand,
PolicyUpdateCommand,
},
connection::MakeEdgeHubPacketProcessor,
settings::Settings,
};
use super::{shutdown, Bootstrap};
const DEVICE_ID_ENV: &str = "IOTEDGE_DEVICEID";
const IOTHUB_HOSTNAME_ENV: &str = "IOTEDGE_IOTHUBHOSTNAME";
#[derive(Default)]
pub struct EdgeHubBootstrap {
broker_ready: BrokerReady,
}
#[async_trait]
impl Bootstrap for EdgeHubBootstrap {
type Settings = Settings;
fn load_config<P: AsRef<Path>>(&self, path: P) -> Result<Self::Settings> {
info!("loading settings from a file {}", path.as_ref().display());
Ok(Self::Settings::from_file(path)?)
}
type Authorizer = LocalAuthorizer<EdgeHubAuthorizer<PolicyAuthorizer>>;
async fn make_broker(
&self,
settings: &Self::Settings,
) -> Result<(Broker<Self::Authorizer>, FilePersistor<VersionedFileFormat>)> {
info!("loading state...");
let persistence_config = settings.broker().persistence();
let state_dir = persistence_config.file_path();
fs::create_dir_all(state_dir.clone())?;
let mut persistor = FilePersistor::new(state_dir, VersionedFileFormat::default());
let state = persistor.load().await?;
info!("state loaded.");
let device_id = env::var(DEVICE_ID_ENV).context(DEVICE_ID_ENV)?;
let iothub_id = env::var(IOTHUB_HOSTNAME_ENV).context(IOTHUB_HOSTNAME_ENV)?;
let authorizer = LocalAuthorizer::new(EdgeHubAuthorizer::new(
PolicyAuthorizer::new(device_id.clone(), self.broker_ready.handle()),
device_id,
iothub_id,
self.broker_ready.handle(),
));
let broker = BrokerBuilder::default()
.with_authorizer(authorizer)
.with_state(state.unwrap_or_default())
.with_config(settings.broker().clone())
.build();
Ok((broker, persistor))
}
fn snapshot_interval(&self, settings: &Self::Settings) -> StdDuration {
settings.broker().persistence().time_interval()
}
fn session_expiration(&self, settings: &Self::Settings) -> StdDuration {
settings.broker().session().expiration()
}
fn session_cleanup_interval(&self, settings: &Self::Settings) -> StdDuration {
settings.broker().session().cleanup_interval()
}
async fn run(
self,
config: Self::Settings,
broker: Broker<Self::Authorizer>,
) -> Result<BrokerSnapshot> {
let broker_handle = broker.handle();
let sidecars = make_sidecars(&broker_handle, &config)?;
info!("starting server...");
let server = make_server(config, broker, self.broker_ready).await?;
let shutdown_signal = shutdown_signal(&server);
let server = tokio::spawn(server.serve(shutdown_signal));
info!("starting sidecars...");
let mut shutdowns = Vec::new();
let mut sidecar_joins = Vec::new();
for sidecar in sidecars {
shutdowns.push(sidecar.shutdown_handle()?);
sidecar_joins.push(tokio::spawn(sidecar.run()));
}
let state = match future::select(server, future::select_all(sidecar_joins)).await {
// server exited first
Either::Left((snapshot, sidecars)) => {
// send shutdown event to each sidecar
let shutdowns = shutdowns.into_iter().map(SidecarShutdownHandle::shutdown);
future::join_all(shutdowns).await;
// awaits for at least one to finish
let (_res, _stopped, sidecars) = sidecars.await;
// wait for the rest to exit
future::join_all(sidecars).await;
snapshot??
}
// one of sidecars exited first
Either::Right(((res, stopped, sidecars), server)) => |
};
Ok(state)
}
}
async fn make_server<Z>(
config: Settings,
broker: Broker<Z>,
broker_ready: BrokerReady,
) -> Result<Server<Z, MakeEdgeHubPacketProcessor<MakeMqttPacketProcessor>>>
where
Z: Authorizer + Send + 'static,
{
let broker_handle = broker.handle();
let make_processor = MakeEdgeHubPacketProcessor::new_default(broker_handle.clone());
let mut server = Server::from_broker(broker).with_packet_processor(make_processor);
// Add system transport to allow communication between edgehub components
let authenticator = LocalAuthenticator::new();
server.with_tcp(config.listener().system().addr(), authenticator, None)?;
// Add regular MQTT over TCP transport
let authenticator = EdgeHubAuthenticator::new(config.auth().url());
if let Some(tcp) = config.listener().tcp() {
let broker_ready = Some(broker_ready.signal());
server.with_tcp(tcp.addr(), authenticator.clone(), broker_ready)?;
}
// Add regular MQTT over TLS transport
if let Some(tls) = config.listener().tls() {
let identity = if let Some(config) = tls.certificate() {
info!("loading identity from {}", config.cert_path().display());
ServerCertificate::from_pem(config.cert_path(), config.private_key_path())
.with_context(|| {
ServerCertificateLoadError::File(
config.cert_path().to_path_buf(),
config.private_key_path().to_path_buf(),
)
})?
} else {
info!("downloading identity from edgelet");
download_server_certificate()
.await
.with_context(|| ServerCertificateLoadError::Edgelet)?
};
let broker_ready = Some(broker_ready.signal());
server.with_tls(tls.addr(), identity, authenticator.clone(), broker_ready)?;
};
Ok(server)
}
fn make_sidecars(
broker_handle: &BrokerHandle,
config: &Settings,
) -> Result<Vec<Box<dyn Sidecar + Send>>> {
let mut sidecars: Vec<Box<dyn Sidecar + Send>> = Vec::new();
let system_address = config.listener().system().addr().to_string();
let device_id = env::var(DEVICE_ID_ENV).context(DEVICE_ID_ENV)?;
let settings = BridgeSettings::new()?;
let bridge_controller =
BridgeController::new(system_address.clone(), device_id.to_owned(), settings);
let bridge_controller_handle = bridge_controller.handle();
sidecars.push(Box::new(bridge_controller));
let mut command_handler = CommandHandler::new(system_address, &device_id);
command_handler.add_command(DisconnectCommand::new(&broker_handle));
command_handler.add_command(AuthorizedIdentitiesCommand::new(&broker_handle));
command_handler.add_command(PolicyUpdateCommand::new(broker_handle));
command_handler.add_command(BridgeUpdateCommand::new(bridge_controller_handle));
sidecars.push(Box::new(command_handler));
Ok(sidecars)
}
pub const WORKLOAD_URI: &str = "IOTEDGE_WORKLOADURI";
pub const EDGE_DEVICE_HOST_NAME: &str = "EdgeDeviceHostName";
pub const MODULE_ID: &str = "IOTEDGE_MODULEID";
pub const MODULE_GENERATION_ID: &str = "IOTEDGE_MODULEGENERATIONID";
pub const CERTIFICATE_VALIDITY_DAYS: i64 = 90;
async fn download_server_certificate() -> Result<ServerCertificate> {
let uri = env::var(WORKLOAD_URI).context(WORKLOAD_URI)?;
let hostname = env::var(EDGE_DEVICE_HOST_NAME).context(EDGE_DEVICE_HOST_NAME)?;
let module_id = env::var(MODULE_ID).context(MODULE_ID)?;
let generation_id = env::var(MODULE_GENERATION_ID).context(MODULE_GENERATION_ID)?;
let expiration = Utc::now() + Duration::days(CERTIFICATE_VALIDITY_DAYS);
let client = edgelet_client::workload(&uri)?;
let cert = client
.create_server_cert(&module_id, &generation_id, &hostname, expiration)
.await?;
if cert.private_key().type_() != "key" {
bail!(
"unknown type of private key: {}",
cert.private_key().type_()
);
}
if let Some(private_key) = cert.private_key().bytes() {
let identity = ServerCertificate::from_pem_pair(cert.certificate(), private_key)?;
Ok(identity)
} else {
bail!("missing private key");
}
}
fn shutdown_signal<Z, P>(server: &Server<Z, | {
debug!("a sidecar has stopped. shutting down all sidecars...");
if let Err(e) = res {
error!(message = "failed waiting for sidecar shutdown", error = %e);
}
// send shutdown event to each of the rest sidecars
shutdowns.remove(stopped);
let shutdowns = shutdowns.into_iter().map(SidecarShutdownHandle::shutdown);
future::join_all(shutdowns).await;
// wait for the rest to exit
future::join_all(sidecars).await;
// signal server
broker_handle.send(Message::System(SystemEvent::Shutdown))?;
server.await??
} | conditional_block |
utils.ts | city) {
const hex = hexColor.replace('#', '');
const r = parseInt(hex.substring(0, 2), 16);
const g = parseInt(hex.substring(2, 4), 16);
const b = parseInt(hex.substring(4, 6), 16);
const result = `rgba(${r}, ${g}, ${b}, ${opacity})`;
return result;
}
/**
*
* @param target {object}
*/
export function isObject(target) {
return (
Object.prototype.toString.call(target).toLowerCase() === '[object object]'
);
}
/**
*
* @param target {string | any}
*/
export function isString(target) {
return (
Object.prototype.toString.call(target).toLowerCase() === '[object string]'
);
}
/**
*
* @param target {object}
* @param options {object}
*/
export function mergeOptions(target, options) {
if (!isObject(target) || !isObject(options)) {
throw new Error('params must be object');
}
Object.keys(options).forEach(key => {
if (options[key] && !target[key]) {
target[key] = options[key];
return;
}
if (isObject(target[key]) && isObject(options[key])) {
mergeOptions(target[key], options[key]);
} else {
const isType =
Object.prototype.toString.call(options[key]).toLowerCase() ===
Object.prototype.toString.call(target[key]).toLowerCase();
if (!isType && target[key] != undefined) {
throw new Error(`params ${key} must be ${typeof target[key]}`);
} else {
target[key] = options[key];
}
}
});
}
/**
*
* @param ctx
* @param vertex
* @param radius
* @param options
*/
export const drawCircle = (
ctx: any,
vertex: IVertex,
radius,
options: DrawCircleOption
) => {
if (!options.hasOwnProperty('dashed') || options.dashed === false) {
ctx.setLineDash([]);
} else {
const _dashedConfig =
options.dashedConfig && options.dashedConfig.length
? options.dashedConfig
: [5, 5, 5];
ctx.setLineDash(_dashedConfig);
}
const { lineColor, weight, opacity, fillColor } = options;
ctx.beginPath();
ctx.arc(vertex.x, vertex.y, radius, 0, 2 * Math.PI, false);
ctx.fillStyle = hex2rgba(fillColor, opacity);
ctx.fill();
ctx.lineWidth = weight;
ctx.strokeStyle = lineColor;
ctx.stroke();
};
/**
*
* @param ctx
* @param vertexes
* @param drawPolyOption
*/
export const drawPolygon = (
ctx: any,
vertexes: IVertex[],
drawPolyOption: DrawPolygonOption
) => {
const { lineColor, fillColor, weight, opacity } = drawPolyOption;
ctx.fillStyle = hex2rgba(fillColor, opacity);
ctx.lineWidth = weight;
ctx.strokeStyle = hex2rgba(lineColor, opacity);
ctx.beginPath();
ctx.moveTo(vertexes[0].x, vertexes[0].y);
for (let i = 1; i < vertexes.length; i++) {
const { x, y } = vertexes[i];
ctx.lineTo(x, y);
}
ctx.lineTo(vertexes[0].x, vertexes[0].y);
ctx.closePath();
ctx.fill();
ctx.stroke();
};
/**
*
* @param ctx
* @param start
* @param end
* @param drawLineOption
*/
export const drawLine = (
ctx: any,
start: IVertex,
end: IVertex,
drawLineOption: DrawLineOption
): any => {
const { color, weight, opacity, strokeStyle, noStrokeStyle } = drawLineOption;
ctx.beginPath();
ctx.moveTo(start.x, start.y);
ctx.lineTo(end.x, end.y);
if (!noStrokeStyle) {
const style = hex2rgba(color, opacity);
ctx.strokeStyle = strokeStyle ? strokeStyle : style;
}
ctx.lineWidth = weight;
ctx.stroke();
return {
start,
end
};
};
export function drawRect(
ctx: any,
rect: [IVertex, IVertex],
drawRectOption: DrawRectOption
) {
const {
lineColor,
fillColor,
weight,
opacity,
dashed,
dashedConfig
} = drawRectOption;
ctx.beginPath();
// 虚线设置
if (dashed) {
const _dashedConfig =
dashedConfig && dashedConfig.length ? dashedConfig : [5, 5, 5];
ctx.setLineDash(_dashedConfig);
} else {
ctx.setLineDash([]);
}
ctx.lineWidth = weight;
ctx.strokeStyle = hex2rgba(lineColor, opacity);
if (fillColor) { ctx.fillStyle = hex2rgba(fillColor, opacity) };
const [{ x: startX, y: startY }, { x: endX, y: endY }] = rect;
const width = endX - startX;
const height = endY - startY;
ctx.rect(startX, startY, width, height);
fillColor && ctx.fill()
ctx.stroke();
return rect;
}
export function drawEllipse(
ctx: any,
ellipse: Ellipse,
drawEllipseOption: DrawEllipseOption
) {
const {
vertex: { x, y },
radius: { minorAxis, macroAxis }
} = ellipse;
const {
fillColor,
opacity,
linearFlag,
linearStartPoint,
linearEndPoint,
startRadius,
endRadius,
colorItems
} = drawEllipseOption;
ctx.save();
let grd = null;
if (linearFlag) {
grd = ctx.createLinearGradient(
linearStartPoint.x,
linearStartPoint.y,
linearEndPoint.x,
linearEndPoint.y
);
colorItems.forEach(item => {
grd.addColorStop(item.position, item.color);
});
} else {
grd = ctx.createRadialGradient(x, y, startRadius, x, y, endRadius);
colorItems.forEach(item => {
grd.addColorStop(item.position, item.color);
});
}
ctx.fillStyle = grd;
var step = minorAxis > macroAxis ? 1 / minorAxis : 1 / macroAxis;
ctx.beginPath();
ctx.moveTo(x + minorAxis, y);
for (let i = 0; i < 2 * Math.PI; i += step) {
ctx.lineTo(x + minorAxis * Math.cos(i), y + macroAxis * Math.sin(i));
}
ctx.closePath();
ctx.fill();
}
/**
*
* @param ctx
* @param rect
*/
export const clearCanvas = (
ctx: any,
rect: { width: number; height: number }
) => {
const { width, height } = rect;
ctx.clearRect(0, 0, width, height);
};
/**
*
* @param event {mouseEvent}
*/
export function transformVertex(event, zoom = 1): IVertex {
const { offsetX: x, offsetY: y } = event;
return { x: x / zoom, y: y / zoom };
}
export function isInRect(vertex: IVertex, rect: [IVertex, IVertex]): boolean {
if (
vertex.x > rect[0].x &&
vertex.x < rect[1].x &&
vertex.y > rect[0].y &&
vertex.y < rect[1].y
) {
return true;
}
return false;
}
/**
* 在 canvas 上下文新建一个渐变区域
* @param ctx
* @param option
*/
export const createLinearGradient = (
ctx: CanvasRenderingContext2D,
option: LinearGradientOption
) => {
const { scope, colorSteps } = option;
const [start, end] = scope;
const gradient = ctx.createLinearGradient(start.x, start.y, end.x, end.y);
colorSteps.forEach((step, index) => {
gradient.addColorStop(step.distance, step.color);
});
return gradient;
};
/**
* 获取两个点连线的方向
* @param start IVetex
* @param end IVetex
*/
export function getDirection(start: IVertex, end: IVertex) {
let left: boolean = false;
let top: boolean = false;
let right: boolean = false;
let bottom: boolean = false;
if (start.x <= end.x) {
right = true;
} else {
left = true;
}
if (start.y <= end.y) {
bottom = true;
} else {
top = true;
}
return {
left,
top,
right,
bottom
};
}
/**
* @param start IVertex
* @param end IVertex
* 知道两点, 获取两点连线 斜率 和 纵轴交点
*/
export function getSlopeAndB(
start: IVertex,
end: IVertex
): { slope: number; b: number } {
const { y: y1, x: x1 } = start;
const { y: y2, x: x2 } = end;
const xDistance = x1 - x2;
const yDistance = y1 - y2;
const b = y1 - x1 * (yDistance / xDistance);
return {
| lor, opa | identifier_name |
|
utils.ts | .dashedConfig && options.dashedConfig.length
? options.dashedConfig
: [5, 5, 5];
ctx.setLineDash(_dashedConfig);
}
const { lineColor, weight, opacity, fillColor } = options;
ctx.beginPath();
ctx.arc(vertex.x, vertex.y, radius, 0, 2 * Math.PI, false);
ctx.fillStyle = hex2rgba(fillColor, opacity);
ctx.fill();
ctx.lineWidth = weight;
ctx.strokeStyle = lineColor;
ctx.stroke();
};
/**
*
* @param ctx
* @param vertexes
* @param drawPolyOption
*/
export const drawPolygon = (
ctx: any,
vertexes: IVertex[],
drawPolyOption: DrawPolygonOption
) => {
const { lineColor, fillColor, weight, opacity } = drawPolyOption;
ctx.fillStyle = hex2rgba(fillColor, opacity);
ctx.lineWidth = weight;
ctx.strokeStyle = hex2rgba(lineColor, opacity);
ctx.beginPath();
ctx.moveTo(vertexes[0].x, vertexes[0].y);
for (let i = 1; i < vertexes.length; i++) {
const { x, y } = vertexes[i];
ctx.lineTo(x, y);
}
ctx.lineTo(vertexes[0].x, vertexes[0].y);
ctx.closePath();
ctx.fill();
ctx.stroke();
};
/**
*
* @param ctx
* @param start
* @param end
* @param drawLineOption
*/
export const drawLine = (
ctx: any,
start: IVertex,
end: IVertex,
drawLineOption: DrawLineOption
): any => {
const { color, weight, opacity, strokeStyle, noStrokeStyle } = drawLineOption;
ctx.beginPath();
ctx.moveTo(start.x, start.y);
ctx.lineTo(end.x, end.y);
if (!noStrokeStyle) {
const style = hex2rgba(color, opacity);
ctx.strokeStyle = strokeStyle ? strokeStyle : style;
}
ctx.lineWidth = weight;
ctx.stroke();
return {
start,
end
};
};
export function drawRect(
ctx: any,
rect: [IVertex, IVertex],
drawRectOption: DrawRectOption
) {
const {
lineColor,
fillColor,
weight,
opacity,
dashed,
dashedConfig
} = drawRectOption;
ctx.beginPath();
// 虚线设置
if (dashed) {
const _dashedConfig =
dashedConfig && dashedConfig.length ? dashedConfig : [5, 5, 5];
ctx.setLineDash(_dashedConfig);
} else {
ctx.setLineDash([]);
}
ctx.lineWidth = weight;
ctx.strokeStyle = hex2rgba(lineColor, opacity);
if (fillColor) { ctx.fillStyle = hex2rgba(fillColor, opacity) };
const [{ x: startX, y: startY }, { x: endX, y: endY }] = rect;
const width = endX - startX;
const height = endY - startY;
ctx.rect(startX, startY, width, height);
fillColor && ctx.fill()
ctx.stroke();
return rect;
}
export function drawEllipse(
ctx: any,
ellipse: Ellipse,
drawEllipseOption: DrawEllipseOption
) {
const {
vertex: { x, y },
radius: { minorAxis, macroAxis }
} = ellipse;
const {
fillColor,
opacity,
linearFlag,
linearStartPoint,
linearEndPoint,
startRadius,
endRadius,
colorItems
} = drawEllipseOption;
ctx.save();
let grd = null;
if (linearFlag) {
grd = ctx.createLinearGradient(
linearStartPoint.x,
linearStartPoint.y,
linearEndPoint.x,
linearEndPoint.y
);
colorItems.forEach(item => {
grd.addColorStop(item.position, item.color);
});
} else {
grd = ctx.createRadialGradient(x, y, startRadius, x, y, endRadius);
colorItems.forEach(item => {
grd.addColorStop(item.position, item.color);
});
}
ctx.fillStyle = grd;
var step = minorAxis > macroAxis ? 1 / minorAxis : 1 / macroAxis;
ctx.beginPath();
ctx.moveTo(x + minorAxis, y);
for (let i = 0; i < 2 * Math.PI; i += step) {
ctx.lineTo(x + minorAxis * Math.cos(i), y + macroAxis * Math.sin(i));
}
ctx.closePath();
ctx.fill();
}
/**
*
* @param ctx
* @param rect
*/
export const clearCanvas = (
ctx: any,
rect: { width: number; height: number }
) => {
const { width, height } = rect;
ctx.clearRect(0, 0, width, height);
};
/**
*
* @param event {mouseEvent}
*/
export function transformVertex(event, zoom = 1): IVertex {
const { offsetX: x, offsetY: y } = event;
return { x: x / zoom, y: y / zoom };
}
export function isInRect(vertex: IVertex, rect: [IVertex, IVertex]): boolean {
if (
vertex.x > rect[0].x &&
vertex.x < rect[1].x &&
vertex.y > rect[0].y &&
vertex.y < rect[1].y
) {
return true;
}
return false;
}
/**
* 在 canvas 上下文新建一个渐变区域
* @param ctx
* @param option
*/
export const createLinearGradient = (
ctx: CanvasRenderingContext2D,
option: LinearGradientOption
) => {
const { scope, colorSteps } = option;
const [start, end] = scope;
const gradient = ctx.createLinearGradient(start.x, start.y, end.x, end.y);
colorSteps.forEach((step, index) => {
gradient.addColorStop(step.distance, step.color);
});
return gradient;
};
/**
* 获取两个点连线的方向
* @param start IVetex
* @param end IVetex | export function getDirection(start: IVertex, end: IVertex) {
let left: boolean = false;
let top: boolean = false;
let right: boolean = false;
let bottom: boolean = false;
if (start.x <= end.x) {
right = true;
} else {
left = true;
}
if (start.y <= end.y) {
bottom = true;
} else {
top = true;
}
return {
left,
top,
right,
bottom
};
}
/**
* @param start IVertex
* @param end IVertex
* 知道两点, 获取两点连线 斜率 和 纵轴交点
*/
export function getSlopeAndB(
start: IVertex,
end: IVertex
): { slope: number; b: number } {
const { y: y1, x: x1 } = start;
const { y: y2, x: x2 } = end;
const xDistance = x1 - x2;
const yDistance = y1 - y2;
const b = y1 - x1 * (yDistance / xDistance);
return {
b,
slope: yDistance / xDistance
};
}
/**
* 按照斜率方程(斜率/纵轴交点) 和 步数以及总步数来 算出下一个点的坐标
* @param start: IVertex,
* @param end: IVertex,
* @param slope: number,
* @param stepIndex: number,
* @param totalStep: number,
* @param b: number
*/
export function getNextVertex(
start: IVertex,
end: IVertex,
slope: number,
stepIndex: number,
totalStep: number,
b: number
): IVertex {
const { y: y1, x: x1 } = start;
const { y: y2, x: x2 } = end;
const xDistance = Math.abs(x1 - x2) * (stepIndex / totalStep);
const direction = getDirection(start, end);
let x = x1 + xDistance;
if (direction.left) {
x = x1 - xDistance;
}
const y = slope * x + b;
return { x, y };
}
export interface ArrowOptions {
strokeStyle: string;
colorFill: [string, string];
}
const defaultArrowOpitons: ArrowOptions = {
strokeStyle: '#68cdfa',
colorFill: ['#68cdfa', '#68cdfa']
};
/* 获取画箭头两端的点 */
export function getArrowPoint(
pixelStart: IVertex,
pixelEnd: IVertex,
length: number = 15
) {
// 绘制箭头的函数
// const length = 12;
const angleValue = Math.PI / 7;
const angle = angleValue; // 箭头和主线的夹角
const r = length; // r/Math.sin(angle)代表箭头长度
let delta = 0; // 主线斜率,垂直时无斜率
let param = 0; // 代码简洁考虑 | */ | random_line_split |
utils.ts |
/**
*
* @param ctx
* @param vertex
* @param radius
* @param options
*/
export const drawCircle = (
ctx: any,
vertex: IVertex,
radius,
options: DrawCircleOption
) => {
if (!options.hasOwnProperty('dashed') || options.dashed === false) {
ctx.setLineDash([]);
} else {
const _dashedConfig =
options.dashedConfig && options.dashedConfig.length
? options.dashedConfig
: [5, 5, 5];
ctx.setLineDash(_dashedConfig);
}
const { lineColor, weight, opacity, fillColor } = options;
ctx.beginPath();
ctx.arc(vertex.x, vertex.y, radius, 0, 2 * Math.PI, false);
ctx.fillStyle = hex2rgba(fillColor, opacity);
ctx.fill();
ctx.lineWidth = weight;
ctx.strokeStyle = lineColor;
ctx.stroke();
};
/**
*
* @param ctx
* @param vertexes
* @param drawPolyOption
*/
export const drawPolygon = (
ctx: any,
vertexes: IVertex[],
drawPolyOption: DrawPolygonOption
) => {
const { lineColor, fillColor, weight, opacity } = drawPolyOption;
ctx.fillStyle = hex2rgba(fillColor, opacity);
ctx.lineWidth = weight;
ctx.strokeStyle = hex2rgba(lineColor, opacity);
ctx.beginPath();
ctx.moveTo(vertexes[0].x, vertexes[0].y);
for (let i = 1; i < vertexes.length; i++) {
const { x, y } = vertexes[i];
ctx.lineTo(x, y);
}
ctx.lineTo(vertexes[0].x, vertexes[0].y);
ctx.closePath();
ctx.fill();
ctx.stroke();
};
/**
*
* @param ctx
* @param start
* @param end
* @param drawLineOption
*/
export const drawLine = (
ctx: any,
start: IVertex,
end: IVertex,
drawLineOption: DrawLineOption
): any => {
const { color, weight, opacity, strokeStyle, noStrokeStyle } = drawLineOption;
ctx.beginPath();
ctx.moveTo(start.x, start.y);
ctx.lineTo(end.x, end.y);
if (!noStrokeStyle) {
const style = hex2rgba(color, opacity);
ctx.strokeStyle = strokeStyle ? strokeStyle : style;
}
ctx.lineWidth = weight;
ctx.stroke();
return {
start,
end
};
};
export function drawRect(
ctx: any,
rect: [IVertex, IVertex],
drawRectOption: DrawRectOption
) {
const {
lineColor,
fillColor,
weight,
opacity,
dashed,
dashedConfig
} = drawRectOption;
ctx.beginPath();
// 虚线设置
if (dashed) {
const _dashedConfig =
dashedConfig && dashedConfig.length ? dashedConfig : [5, 5, 5];
ctx.setLineDash(_dashedConfig);
} else {
ctx.setLineDash([]);
}
ctx.lineWidth = weight;
ctx.strokeStyle = hex2rgba(lineColor, opacity);
if (fillColor) { ctx.fillStyle = hex2rgba(fillColor, opacity) };
const [{ x: startX, y: startY }, { x: endX, y: endY }] = rect;
const width = endX - startX;
const height = endY - startY;
ctx.rect(startX, startY, width, height);
fillColor && ctx.fill()
ctx.stroke();
return rect;
}
export function drawEllipse(
ctx: any,
ellipse: Ellipse,
drawEllipseOption: DrawEllipseOption
) {
const {
vertex: { x, y },
radius: { minorAxis, macroAxis }
} = ellipse;
const {
fillColor,
opacity,
linearFlag,
linearStartPoint,
linearEndPoint,
startRadius,
endRadius,
colorItems
} = drawEllipseOption;
ctx.save();
let grd = null;
if (linearFlag) {
grd = ctx.createLinearGradient(
linearStartPoint.x,
linearStartPoint.y,
linearEndPoint.x,
linearEndPoint.y
);
colorItems.forEach(item => {
grd.addColorStop(item.position, item.color);
});
} else {
grd = ctx.createRadialGradient(x, y, startRadius, x, y, endRadius);
colorItems.forEach(item => {
grd.addColorStop(item.position, item.color);
});
}
ctx.fillStyle = grd;
var step = minorAxis > macroAxis ? 1 / minorAxis : 1 / macroAxis;
ctx.beginPath();
ctx.moveTo(x + minorAxis, y);
for (let i = 0; i < 2 * Math.PI; i += step) {
ctx.lineTo(x + minorAxis * Math.cos(i), y + macroAxis * Math.sin(i));
}
ctx.closePath();
ctx.fill();
}
/**
*
* @param ctx
* @param rect
*/
export const clearCanvas = (
ctx: any,
rect: { width: number; height: number }
) => {
const { width, height } = rect;
ctx.clearRect(0, 0, width, height);
};
/**
*
* @param event {mouseEvent}
*/
export function transformVertex(event, zoom = 1): IVertex {
const { offsetX: x, offsetY: y } = event;
return { x: x / zoom, y: y / zoom };
}
export function isInRect(vertex: IVertex, rect: [IVertex, IVertex]): boolean {
if (
vertex.x > rect[0].x &&
vertex.x < rect[1].x &&
vertex.y > rect[0].y &&
vertex.y < rect[1].y
) {
return true;
}
return false;
}
/**
* 在 canvas 上下文新建一个渐变区域
* @param ctx
* @param option
*/
export const createLinearGradient = (
ctx: CanvasRenderingContext2D,
option: LinearGradientOption
) => {
const { scope, colorSteps } = option;
const [start, end] = scope;
const gradient = ctx.createLinearGradient(start.x, start.y, end.x, end.y);
colorSteps.forEach((step, index) => {
gradient.addColorStop(step.distance, step.color);
});
return gradient;
};
/**
* 获取两个点连线的方向
* @param start IVetex
* @param end IVetex
*/
export function getDirection(start: IVertex, end: IVertex) {
let left: boolean = false;
let top: boolean = false;
let right: boolean = false;
let bottom: boolean = false;
if (start.x <= end.x) {
right = true;
} else {
left = true;
}
if (start.y <= end.y) {
bottom = true;
} else {
top = true;
}
return {
left,
top,
right,
bottom
};
}
/**
* @param start IVertex
* @param end IVertex
* 知道两点, 获取两点连线 斜率 和 纵轴交点
*/
export function getSlopeAndB(
start: IVertex,
end: IVertex
): { slope: number; b: number } {
const { y: y1, x: x1 } = start;
const { y: y2, x: x2 } = end;
const xDistance = x1 - x2;
const yDistance = y1 - y2;
const b = y1 - x1 * (yDistance / xDistance);
return {
b,
slope: yDistance / xDistance
};
}
/**
* 按照斜率方程(斜率/纵轴交点) 和 步数以及总步数来 算出下一个点的坐标
* @param start: IVertex,
* @param end: IVertex,
* @param slope: number,
* @param stepIndex: number,
* @param totalStep: number,
* @param b: number
*/
export function getNextVertex(
start: IVertex,
end: IVertex,
slope: number,
stepIndex: number,
totalStep: number,
b: number
): IVertex {
const { y: y1, x: x1 } = start;
const { y: y2, x: x2 } = end;
const xDistance = Math.abs(x1 - x2) * (stepIndex / totalStep);
const direction = getDirection(start, end);
let x = x1 + xDistance;
if (direction.left) {
x = x1 - xDistance;
}
const y = slope * x + b;
return { x, y };
}
export interface ArrowOptions {
strokeStyle: string;
colorFill: [string, string];
}
const defaultArrowOpitons: ArrowOptions = {
strokeStyle: '#68cdfa',
colorFill: ['#68cdfa', '#68cdfa']
};
/* 获取画箭头两端的点 */
export function getArrowPoint(
pixelStart: IVertex,
pixelEnd: IVertex,
length | et[key] = options[key];
}
}
});
} | conditional_block |
|
utils.ts | ashedConfig && options.dashedConfig.length
? options.dashedConfig
: [5, 5, 5];
ctx.setLineDash(_dashedConfig);
}
const { lineColor, weight, opacity, fillColor } = options;
ctx.beginPath();
ctx.arc(vertex.x, vertex.y, radius, 0, 2 * Math.PI, false);
ctx.fillStyle = hex2rgba(fillColor, opacity);
ctx.fill();
ctx.lineWidth = weight;
ctx.strokeStyle = lineColor;
ctx.stroke();
};
/**
*
* @param ctx
* @param vertexes
* @param drawPolyOption
*/
export const drawPolygon = (
ctx: any,
vertexes: IVertex[],
drawPolyOption: DrawPolygonOption
) => {
const { lineColor, fillColor, weight, opacity } = drawPolyOption;
ctx.fillStyle = hex2rgba(fillColor, opacity);
ctx.lineWidth = weight;
ctx.strokeStyle = hex2rgba(lineColor, opacity);
ctx.beginPath();
ctx.moveTo(vertexes[0].x, vertexes[0].y);
for (let i = 1; i < vertexes.length; i++) {
const { x, y } = vertexes[i];
ctx.lineTo(x, y);
}
ctx.lineTo(vertexes[0].x, vertexes[0].y);
ctx.closePath();
ctx.fill();
ctx.stroke();
};
/**
*
* @param ctx
* @param start
* @param end
* @param drawLineOption
*/
export const drawLine = (
ctx: any,
start: IVertex,
end: IVertex,
drawLineOption: DrawLineOption
): any => {
const { color, weight, opacity, strokeStyle, noStrokeStyle } = drawLineOption;
ctx.beginPath();
ctx.moveTo(start.x, start.y);
ctx.lineTo(end.x, end.y);
if (!noStrokeStyle) {
const style = hex2rgba(color, opacity);
ctx.strokeStyle = strokeStyle ? strokeStyle : style;
}
ctx.lineWidth = weight;
ctx.stroke();
return {
start,
end
};
};
export function drawRect(
ctx: any,
rect: [IVertex, IVertex],
drawRectOption: DrawRectOption
) {
const {
lineColor,
fillColor,
weight,
opacity,
dashed,
dashedConfig
} = drawRectOption;
ctx.beginPath();
// 虚线设置
if (dashed) {
const _dashedConfig =
dashedConfig && dashedConfig.length ? dashedConfig : [5, 5, 5];
ctx.setLineDash(_dashedConfig);
} else {
ctx.setLineDash([]);
}
ctx.lineWidth = weight;
ctx.strokeStyle = hex2rgba(lineColor, opacity);
if (fillColor) { ctx.fillStyle = hex2rgba(fillColor, opacity) };
const [{ x: startX, y: startY }, { x: endX, y: endY }] = rect;
const width = endX - startX;
const height = endY - startY;
ctx.rect(startX, startY, width, height);
fillColor && ctx.fill()
ctx.stroke();
return rect;
}
export function drawEllipse(
ctx: any,
ellipse: Ellipse,
drawEllipseOption: DrawEllipseOption
) {
const {
vertex: { x, y },
radius: { minorAxis, macroAxis }
} = ellipse;
const {
fillColor,
opacity,
linearFlag,
linearStartPoint,
linearEndPoint,
startRadius,
endRadius,
colorItems
} = drawEllipseOption;
ctx.save();
let grd = null;
if (linearFlag) {
grd = ctx.createLinearGradient(
linearStartPoint.x,
linearStartPoint.y,
linearEndPoint.x,
linearEndPoint.y
);
colorItems.forEach(item => {
grd.addColorStop(item.position, item.color);
});
} else {
grd = ctx.createRadialGradient(x, y, startRadius, x, y, endRadius);
colorItems.forEach(item => {
grd.addColorStop(item.position, item.color);
});
}
ctx.fillStyle = grd;
var step = minorAxis > macroAxis ? 1 / minorAxis : 1 / macroAxis;
ctx.beginPath();
ctx.moveTo(x + minorAxis, y);
for (let i = 0; i < 2 * Math.PI; i += step) {
ctx.lineTo(x + minorAxis * Math.cos(i), y + macroAxis * Math.sin(i));
}
ctx.closePath();
ctx.fill();
}
/**
*
* @param ctx
* @param rect
*/
export const clearCanvas = (
ctx: any,
rect: { width: number; height: number }
) => {
const { width, height } = rect;
ctx.clearRect(0, 0, width, height);
};
/**
*
* @param event {mouseEvent}
*/
export function transformVertex(event, zoom = 1): IVertex {
const { offsetX: x | Rect(vertex: IVertex, rect: [IVertex, IVertex]): boolean {
if (
vertex.x > rect[0].x &&
vertex.x < rect[1].x &&
vertex.y > rect[0].y &&
vertex.y < rect[1].y
) {
return true;
}
return false;
}
/**
* 在 canvas 上下文新建一个渐变区域
* @param ctx
* @param option
*/
export const createLinearGradient = (
ctx: CanvasRenderingContext2D,
option: LinearGradientOption
) => {
const { scope, colorSteps } = option;
const [start, end] = scope;
const gradient = ctx.createLinearGradient(start.x, start.y, end.x, end.y);
colorSteps.forEach((step, index) => {
gradient.addColorStop(step.distance, step.color);
});
return gradient;
};
/**
* 获取两个点连线的方向
* @param start IVetex
* @param end IVetex
*/
export function getDirection(start: IVertex, end: IVertex) {
let left: boolean = false;
let top: boolean = false;
let right: boolean = false;
let bottom: boolean = false;
if (start.x <= end.x) {
right = true;
} else {
left = true;
}
if (start.y <= end.y) {
bottom = true;
} else {
top = true;
}
return {
left,
top,
right,
bottom
};
}
/**
* @param start IVertex
* @param end IVertex
* 知道两点, 获取两点连线 斜率 和 纵轴交点
*/
export function getSlopeAndB(
start: IVertex,
end: IVertex
): { slope: number; b: number } {
const { y: y1, x: x1 } = start;
const { y: y2, x: x2 } = end;
const xDistance = x1 - x2;
const yDistance = y1 - y2;
const b = y1 - x1 * (yDistance / xDistance);
return {
b,
slope: yDistance / xDistance
};
}
/**
* 按照斜率方程(斜率/纵轴交点) 和 步数以及总步数来 算出下一个点的坐标
* @param start: IVertex,
* @param end: IVertex,
* @param slope: number,
* @param stepIndex: number,
* @param totalStep: number,
* @param b: number
*/
export function getNextVertex(
start: IVertex,
end: IVertex,
slope: number,
stepIndex: number,
totalStep: number,
b: number
): IVertex {
const { y: y1, x: x1 } = start;
const { y: y2, x: x2 } = end;
const xDistance = Math.abs(x1 - x2) * (stepIndex / totalStep);
const direction = getDirection(start, end);
let x = x1 + xDistance;
if (direction.left) {
x = x1 - xDistance;
}
const y = slope * x + b;
return { x, y };
}
export interface ArrowOptions {
strokeStyle: string;
colorFill: [string, string];
}
const defaultArrowOpitons: ArrowOptions = {
strokeStyle: '#68cdfa',
colorFill: ['#68cdfa', '#68cdfa']
};
/* 获取画箭头两端的点 */
export function getArrowPoint(
pixelStart: IVertex,
pixelEnd: IVertex,
length: number = 15
) {
// 绘制箭头的函数
// const length = 12;
const angleValue = Math.PI / 7;
const angle = angleValue; // 箭头和主线的夹角
const r = length; // r/Math.sin(angle)代表箭头长度
let delta = 0; // 主线斜率,垂直时无斜率
let param = 0; // 代码简洁考虑 | , offsetY: y } = event;
return { x: x / zoom, y: y / zoom };
}
export function isIn | identifier_body |
v4-signature.go |
// Sign - local variables
type Sign struct {
accessKeyID string
secretAccessKey string
region string
httpRequest *http.Request
extractedSignedHeaders http.Header
}
// AWS Signature Version '4' constants.
const (
signV4Algorithm = "AWS4-HMAC-SHA256"
iso8601Format = "20060102T150405Z"
yyyymmdd = "20060102"
)
// New - initialize a new authorization checkes.
func New(accessKeyID, secretAccessKey, region string) (*Sign, *probe.Error) {
if !isValidAccessKey.MatchString(accessKeyID) {
return nil, ErrInvalidAccessKeyID("Invalid access key id.", accessKeyID).Trace(accessKeyID)
}
if !isValidSecretKey.MatchString(secretAccessKey) {
return nil, ErrInvalidAccessKeyID("Invalid secret key.", secretAccessKey).Trace(secretAccessKey)
}
if region == "" {
return nil, ErrRegionISEmpty("Region is empty.").Trace()
}
signature := &Sign{
accessKeyID: accessKeyID,
secretAccessKey: secretAccessKey,
region: region,
}
return signature, nil
}
// SetHTTPRequestToVerify - sets the http request which needs to be verified.
func (s *Sign) SetHTTPRequestToVerify(r *http.Request) *Sign {
// Do not set http request if its 'nil'.
if r == nil {
return s
}
s.httpRequest = r
return s
}
// getCanonicalHeaders generate a list of request headers with their values
func (s Sign) getCanonicalHeaders(signedHeaders http.Header) string {
var headers []string
vals := make(http.Header)
for k, vv := range signedHeaders {
headers = append(headers, strings.ToLower(k))
vals[strings.ToLower(k)] = vv
}
headers = append(headers, "host")
sort.Strings(headers)
var buf bytes.Buffer
for _, k := range headers {
buf.WriteString(k)
buf.WriteByte(':')
switch {
case k == "host":
buf.WriteString(s.httpRequest.Host)
fallthrough
default:
for idx, v := range vals[k] {
if idx > 0 {
buf.WriteByte(',')
}
buf.WriteString(v)
}
buf.WriteByte('\n')
}
}
return buf.String()
}
// getSignedHeaders generate a string i.e alphabetically sorted, semicolon-separated list of lowercase request header names
func (s Sign) getSignedHeaders(signedHeaders http.Header) string {
var headers []string
for k := range signedHeaders {
headers = append(headers, strings.ToLower(k))
}
headers = append(headers, "host")
sort.Strings(headers)
return strings.Join(headers, ";")
}
// getCanonicalRequest generate a canonical request of style
//
// canonicalRequest =
// <HTTPMethod>\n
// <CanonicalURI>\n
// <CanonicalQueryString>\n
// <CanonicalHeaders>\n
// <SignedHeaders>\n
// <HashedPayload>
//
func (s *Sign) getCanonicalRequest() string {
payload := s.httpRequest.Header.Get(http.CanonicalHeaderKey("x-amz-content-sha256"))
s.httpRequest.URL.RawQuery = strings.Replace(s.httpRequest.URL.Query().Encode(), "+", "%20", -1)
encodedPath := getURLEncodedName(s.httpRequest.URL.Path)
// Convert any space strings back to "+".
encodedPath = strings.Replace(encodedPath, "+", "%20", -1)
canonicalRequest := strings.Join([]string{
s.httpRequest.Method,
encodedPath,
s.httpRequest.URL.RawQuery,
s.getCanonicalHeaders(s.extractedSignedHeaders),
s.getSignedHeaders(s.extractedSignedHeaders),
payload,
}, "\n")
return canonicalRequest
}
// getCanonicalRequest generate a canonical request of style
//
// canonicalRequest =
// <HTTPMethod>\n
// <CanonicalURI>\n
// <CanonicalQueryString>\n
// <CanonicalHeaders>\n
// <SignedHeaders>\n
// <HashedPayload>
//
func (s Sign) getPresignedCanonicalRequest(presignedQuery string) string {
rawQuery := strings.Replace(presignedQuery, "+", "%20", -1)
encodedPath := getURLEncodedName(s.httpRequest.URL.Path)
// Convert any space strings back to "+".
encodedPath = strings.Replace(encodedPath, "+", "%20", -1)
canonicalRequest := strings.Join([]string{
s.httpRequest.Method,
encodedPath,
rawQuery,
s.getCanonicalHeaders(s.extractedSignedHeaders),
s.getSignedHeaders(s.extractedSignedHeaders),
"UNSIGNED-PAYLOAD",
}, "\n")
return canonicalRequest
}
// getScope generate a string of a specific date, an AWS region, and a service.
func (s Sign) getScope(t time.Time) string {
scope := strings.Join([]string{
t.Format(yyyymmdd),
s.region,
"s3",
"aws4_request",
}, "/")
return scope
}
// getStringToSign a string based on selected query values.
func (s Sign) getStringToSign(canonicalRequest string, t time.Time) string {
stringToSign := signV4Algorithm + "\n" + t.Format(iso8601Format) + "\n"
stringToSign = stringToSign + s.getScope(t) + "\n"
canonicalRequestBytes := sha256.Sum256([]byte(canonicalRequest))
stringToSign = stringToSign + hex.EncodeToString(canonicalRequestBytes[:])
return stringToSign
}
// getSigningKey hmac seed to calculate final signature.
func (s Sign) getSigningKey(t time.Time) []byte {
secret := s.secretAccessKey
date := sumHMAC([]byte("AWS4"+secret), []byte(t.Format(yyyymmdd)))
region := sumHMAC(date, []byte(s.region))
service := sumHMAC(region, []byte("s3"))
signingKey := sumHMAC(service, []byte("aws4_request"))
return signingKey
}
// getSignature final signature in hexadecimal form.
func (s Sign) getSignature(signingKey []byte, stringToSign string) string {
return hex.EncodeToString(sumHMAC(signingKey, []byte(stringToSign)))
}
// DoesPolicySignatureMatch - Verify query headers with post policy
// - http://docs.aws.amazon.com/AmazonS3/latest/API/sigv4-HTTPPOSTConstructPolicy.html
// returns true if matches, false otherwise. if error is not nil then it is always false
func (s *Sign) DoesPolicySignatureMatch(formValues map[string]string) (bool, *probe.Error) {
// Parse credential tag.
credential, err := parseCredential("Credential=" + formValues["X-Amz-Credential"])
if err != nil {
return false, err.Trace(formValues["X-Amz-Credential"])
}
// Verify if the access key id matches.
if credential.accessKeyID != s.accessKeyID {
return false, ErrInvalidAccessKeyID("Access key id does not match with our records.", credential.accessKeyID).Trace(credential.accessKeyID)
}
// Verify if the region is valid.
reqRegion := credential.scope.region
if !isValidRegion(reqRegion, s.region) {
return false, ErrInvalidRegion("Requested region is not recognized.", reqRegion).Trace(reqRegion)
}
// Save region.
s.region = reqRegion
// Parse date string.
t, e := time.Parse(iso8601Format, formValues["X-Amz-Date"])
if e != nil {
return false, probe.NewError(e)
}
signingKey := s.getSigningKey(t)
newSignature := s.getSignature(signingKey, formValues["Policy"])
if newSignature != formValues["X-Amz-Signature"] {
return false, nil
}
return true, nil
}
// DoesPresignedSignatureMatch - Verify query headers with presigned signature
// - http://docs.aws.amazon.com/AmazonS3/latest/API/sigv4-query-string-auth.html
// returns true if matches, false otherwise. if error is not nil then it is always false
func (s *Sign) DoesPresignedSignatureMatch() (bool, *probe.Error) {
// Parse request query string.
preSignValues, err := parsePreSignV4(s.httpRequest.URL.Query())
if err != nil {
return false, err.Trace(s.httpRequest.URL.String())
}
// Verify if the access key id matches.
if preSignValues.Credential.accessKeyID != s.accessKeyID {
return false, ErrInvalidAccessKeyID("Access key id does not match with our records.", preSignValues.Credential.accessKeyID).Trace(preSignValues.Credential.accessKeyID)
}
// Verify if region is valid.
reqRegion := preSignValues.Credential.scope.region
if !isValidRegion(reqRegion, s.region) {
return false, ErrInvalidRegion("Requested region is not recognized.", reqRegion).Trace(reqRegion)
}
// Save region.
s.region = reqRegion
// Extract all the signed headers along with its values.
s.extractedSignedHeaders = extractSignedHeaders(preSignValues.SignedHeaders, s.httpRequest.Header)
// Construct new query | "github.com/minio/minio/pkg/probe"
) | random_line_split |
|
v4-signature.go | es.
func New(accessKeyID, secretAccessKey, region string) (*Sign, *probe.Error) {
if !isValidAccessKey.MatchString(accessKeyID) {
return nil, ErrInvalidAccessKeyID("Invalid access key id.", accessKeyID).Trace(accessKeyID)
}
if !isValidSecretKey.MatchString(secretAccessKey) {
return nil, ErrInvalidAccessKeyID("Invalid secret key.", secretAccessKey).Trace(secretAccessKey)
}
if region == "" {
return nil, ErrRegionISEmpty("Region is empty.").Trace()
}
signature := &Sign{
accessKeyID: accessKeyID,
secretAccessKey: secretAccessKey,
region: region,
}
return signature, nil
}
// SetHTTPRequestToVerify - sets the http request which needs to be verified.
func (s *Sign) SetHTTPRequestToVerify(r *http.Request) *Sign {
// Do not set http request if its 'nil'.
if r == nil {
return s
}
s.httpRequest = r
return s
}
// getCanonicalHeaders generate a list of request headers with their values
func (s Sign) getCanonicalHeaders(signedHeaders http.Header) string {
var headers []string
vals := make(http.Header)
for k, vv := range signedHeaders {
headers = append(headers, strings.ToLower(k))
vals[strings.ToLower(k)] = vv
}
headers = append(headers, "host")
sort.Strings(headers)
var buf bytes.Buffer
for _, k := range headers {
buf.WriteString(k)
buf.WriteByte(':')
switch {
case k == "host":
buf.WriteString(s.httpRequest.Host)
fallthrough
default:
for idx, v := range vals[k] {
if idx > 0 {
buf.WriteByte(',')
}
buf.WriteString(v)
}
buf.WriteByte('\n')
}
}
return buf.String()
}
// getSignedHeaders generate a string i.e alphabetically sorted, semicolon-separated list of lowercase request header names
func (s Sign) getSignedHeaders(signedHeaders http.Header) string {
var headers []string
for k := range signedHeaders {
headers = append(headers, strings.ToLower(k))
}
headers = append(headers, "host")
sort.Strings(headers)
return strings.Join(headers, ";")
}
// getCanonicalRequest generate a canonical request of style
//
// canonicalRequest =
// <HTTPMethod>\n
// <CanonicalURI>\n
// <CanonicalQueryString>\n
// <CanonicalHeaders>\n
// <SignedHeaders>\n
// <HashedPayload>
//
func (s *Sign) getCanonicalRequest() string {
payload := s.httpRequest.Header.Get(http.CanonicalHeaderKey("x-amz-content-sha256"))
s.httpRequest.URL.RawQuery = strings.Replace(s.httpRequest.URL.Query().Encode(), "+", "%20", -1)
encodedPath := getURLEncodedName(s.httpRequest.URL.Path)
// Convert any space strings back to "+".
encodedPath = strings.Replace(encodedPath, "+", "%20", -1)
canonicalRequest := strings.Join([]string{
s.httpRequest.Method,
encodedPath,
s.httpRequest.URL.RawQuery,
s.getCanonicalHeaders(s.extractedSignedHeaders),
s.getSignedHeaders(s.extractedSignedHeaders),
payload,
}, "\n")
return canonicalRequest
}
// getCanonicalRequest generate a canonical request of style
//
// canonicalRequest =
// <HTTPMethod>\n
// <CanonicalURI>\n
// <CanonicalQueryString>\n
// <CanonicalHeaders>\n
// <SignedHeaders>\n
// <HashedPayload>
//
func (s Sign) | (presignedQuery string) string {
rawQuery := strings.Replace(presignedQuery, "+", "%20", -1)
encodedPath := getURLEncodedName(s.httpRequest.URL.Path)
// Convert any space strings back to "+".
encodedPath = strings.Replace(encodedPath, "+", "%20", -1)
canonicalRequest := strings.Join([]string{
s.httpRequest.Method,
encodedPath,
rawQuery,
s.getCanonicalHeaders(s.extractedSignedHeaders),
s.getSignedHeaders(s.extractedSignedHeaders),
"UNSIGNED-PAYLOAD",
}, "\n")
return canonicalRequest
}
// getScope generate a string of a specific date, an AWS region, and a service.
func (s Sign) getScope(t time.Time) string {
scope := strings.Join([]string{
t.Format(yyyymmdd),
s.region,
"s3",
"aws4_request",
}, "/")
return scope
}
// getStringToSign a string based on selected query values.
func (s Sign) getStringToSign(canonicalRequest string, t time.Time) string {
stringToSign := signV4Algorithm + "\n" + t.Format(iso8601Format) + "\n"
stringToSign = stringToSign + s.getScope(t) + "\n"
canonicalRequestBytes := sha256.Sum256([]byte(canonicalRequest))
stringToSign = stringToSign + hex.EncodeToString(canonicalRequestBytes[:])
return stringToSign
}
// getSigningKey hmac seed to calculate final signature.
func (s Sign) getSigningKey(t time.Time) []byte {
secret := s.secretAccessKey
date := sumHMAC([]byte("AWS4"+secret), []byte(t.Format(yyyymmdd)))
region := sumHMAC(date, []byte(s.region))
service := sumHMAC(region, []byte("s3"))
signingKey := sumHMAC(service, []byte("aws4_request"))
return signingKey
}
// getSignature final signature in hexadecimal form.
func (s Sign) getSignature(signingKey []byte, stringToSign string) string {
return hex.EncodeToString(sumHMAC(signingKey, []byte(stringToSign)))
}
// DoesPolicySignatureMatch - Verify query headers with post policy
// - http://docs.aws.amazon.com/AmazonS3/latest/API/sigv4-HTTPPOSTConstructPolicy.html
// returns true if matches, false otherwise. if error is not nil then it is always false
func (s *Sign) DoesPolicySignatureMatch(formValues map[string]string) (bool, *probe.Error) {
// Parse credential tag.
credential, err := parseCredential("Credential=" + formValues["X-Amz-Credential"])
if err != nil {
return false, err.Trace(formValues["X-Amz-Credential"])
}
// Verify if the access key id matches.
if credential.accessKeyID != s.accessKeyID {
return false, ErrInvalidAccessKeyID("Access key id does not match with our records.", credential.accessKeyID).Trace(credential.accessKeyID)
}
// Verify if the region is valid.
reqRegion := credential.scope.region
if !isValidRegion(reqRegion, s.region) {
return false, ErrInvalidRegion("Requested region is not recognized.", reqRegion).Trace(reqRegion)
}
// Save region.
s.region = reqRegion
// Parse date string.
t, e := time.Parse(iso8601Format, formValues["X-Amz-Date"])
if e != nil {
return false, probe.NewError(e)
}
signingKey := s.getSigningKey(t)
newSignature := s.getSignature(signingKey, formValues["Policy"])
if newSignature != formValues["X-Amz-Signature"] {
return false, nil
}
return true, nil
}
// DoesPresignedSignatureMatch - Verify query headers with presigned signature
// - http://docs.aws.amazon.com/AmazonS3/latest/API/sigv4-query-string-auth.html
// returns true if matches, false otherwise. if error is not nil then it is always false
func (s *Sign) DoesPresignedSignatureMatch() (bool, *probe.Error) {
// Parse request query string.
preSignValues, err := parsePreSignV4(s.httpRequest.URL.Query())
if err != nil {
return false, err.Trace(s.httpRequest.URL.String())
}
// Verify if the access key id matches.
if preSignValues.Credential.accessKeyID != s.accessKeyID {
return false, ErrInvalidAccessKeyID("Access key id does not match with our records.", preSignValues.Credential.accessKeyID).Trace(preSignValues.Credential.accessKeyID)
}
// Verify if region is valid.
reqRegion := preSignValues.Credential.scope.region
if !isValidRegion(reqRegion, s.region) {
return false, ErrInvalidRegion("Requested region is not recognized.", reqRegion).Trace(reqRegion)
}
// Save region.
s.region = reqRegion
// Extract all the signed headers along with its values.
s.extractedSignedHeaders = extractSignedHeaders(preSignValues.SignedHeaders, s.httpRequest.Header)
// Construct new query.
query := make(url.Values)
query.Set("X-Amz-Algorithm", signV4Algorithm)
if time.Now().UTC().Sub(preSignValues.Date) > time.Duration(preSignValues.Expires) {
return false, ErrExpiredPresignRequest("Presigned request already expired, please initiate a new request.")
}
// Save the date and expires.
t := preSignValues.Date
expireSeconds := int(time.Duration(preSignValues.Expires) / time.Second)
// Construct the query.
query.Set("X-Amz-Date", t.Format(iso8601Format))
query.Set("X-Amz-Expires", strconv.Itoa | getPresignedCanonicalRequest | identifier_name |
v4-signature.go | secretAccessKey).Trace(secretAccessKey)
}
if region == "" {
return nil, ErrRegionISEmpty("Region is empty.").Trace()
}
signature := &Sign{
accessKeyID: accessKeyID,
secretAccessKey: secretAccessKey,
region: region,
}
return signature, nil
}
// SetHTTPRequestToVerify - sets the http request which needs to be verified.
func (s *Sign) SetHTTPRequestToVerify(r *http.Request) *Sign {
// Do not set http request if its 'nil'.
if r == nil {
return s
}
s.httpRequest = r
return s
}
// getCanonicalHeaders generate a list of request headers with their values
func (s Sign) getCanonicalHeaders(signedHeaders http.Header) string {
var headers []string
vals := make(http.Header)
for k, vv := range signedHeaders {
headers = append(headers, strings.ToLower(k))
vals[strings.ToLower(k)] = vv
}
headers = append(headers, "host")
sort.Strings(headers)
var buf bytes.Buffer
for _, k := range headers {
buf.WriteString(k)
buf.WriteByte(':')
switch {
case k == "host":
buf.WriteString(s.httpRequest.Host)
fallthrough
default:
for idx, v := range vals[k] {
if idx > 0 {
buf.WriteByte(',')
}
buf.WriteString(v)
}
buf.WriteByte('\n')
}
}
return buf.String()
}
// getSignedHeaders generate a string i.e alphabetically sorted, semicolon-separated list of lowercase request header names
func (s Sign) getSignedHeaders(signedHeaders http.Header) string {
var headers []string
for k := range signedHeaders {
headers = append(headers, strings.ToLower(k))
}
headers = append(headers, "host")
sort.Strings(headers)
return strings.Join(headers, ";")
}
// getCanonicalRequest generate a canonical request of style
//
// canonicalRequest =
// <HTTPMethod>\n
// <CanonicalURI>\n
// <CanonicalQueryString>\n
// <CanonicalHeaders>\n
// <SignedHeaders>\n
// <HashedPayload>
//
func (s *Sign) getCanonicalRequest() string {
payload := s.httpRequest.Header.Get(http.CanonicalHeaderKey("x-amz-content-sha256"))
s.httpRequest.URL.RawQuery = strings.Replace(s.httpRequest.URL.Query().Encode(), "+", "%20", -1)
encodedPath := getURLEncodedName(s.httpRequest.URL.Path)
// Convert any space strings back to "+".
encodedPath = strings.Replace(encodedPath, "+", "%20", -1)
canonicalRequest := strings.Join([]string{
s.httpRequest.Method,
encodedPath,
s.httpRequest.URL.RawQuery,
s.getCanonicalHeaders(s.extractedSignedHeaders),
s.getSignedHeaders(s.extractedSignedHeaders),
payload,
}, "\n")
return canonicalRequest
}
// getCanonicalRequest generate a canonical request of style
//
// canonicalRequest =
// <HTTPMethod>\n
// <CanonicalURI>\n
// <CanonicalQueryString>\n
// <CanonicalHeaders>\n
// <SignedHeaders>\n
// <HashedPayload>
//
func (s Sign) getPresignedCanonicalRequest(presignedQuery string) string {
rawQuery := strings.Replace(presignedQuery, "+", "%20", -1)
encodedPath := getURLEncodedName(s.httpRequest.URL.Path)
// Convert any space strings back to "+".
encodedPath = strings.Replace(encodedPath, "+", "%20", -1)
canonicalRequest := strings.Join([]string{
s.httpRequest.Method,
encodedPath,
rawQuery,
s.getCanonicalHeaders(s.extractedSignedHeaders),
s.getSignedHeaders(s.extractedSignedHeaders),
"UNSIGNED-PAYLOAD",
}, "\n")
return canonicalRequest
}
// getScope generate a string of a specific date, an AWS region, and a service.
func (s Sign) getScope(t time.Time) string {
scope := strings.Join([]string{
t.Format(yyyymmdd),
s.region,
"s3",
"aws4_request",
}, "/")
return scope
}
// getStringToSign a string based on selected query values.
func (s Sign) getStringToSign(canonicalRequest string, t time.Time) string {
stringToSign := signV4Algorithm + "\n" + t.Format(iso8601Format) + "\n"
stringToSign = stringToSign + s.getScope(t) + "\n"
canonicalRequestBytes := sha256.Sum256([]byte(canonicalRequest))
stringToSign = stringToSign + hex.EncodeToString(canonicalRequestBytes[:])
return stringToSign
}
// getSigningKey hmac seed to calculate final signature.
func (s Sign) getSigningKey(t time.Time) []byte {
secret := s.secretAccessKey
date := sumHMAC([]byte("AWS4"+secret), []byte(t.Format(yyyymmdd)))
region := sumHMAC(date, []byte(s.region))
service := sumHMAC(region, []byte("s3"))
signingKey := sumHMAC(service, []byte("aws4_request"))
return signingKey
}
// getSignature final signature in hexadecimal form.
func (s Sign) getSignature(signingKey []byte, stringToSign string) string {
return hex.EncodeToString(sumHMAC(signingKey, []byte(stringToSign)))
}
// DoesPolicySignatureMatch - Verify query headers with post policy
// - http://docs.aws.amazon.com/AmazonS3/latest/API/sigv4-HTTPPOSTConstructPolicy.html
// returns true if matches, false otherwise. if error is not nil then it is always false
func (s *Sign) DoesPolicySignatureMatch(formValues map[string]string) (bool, *probe.Error) {
// Parse credential tag.
credential, err := parseCredential("Credential=" + formValues["X-Amz-Credential"])
if err != nil {
return false, err.Trace(formValues["X-Amz-Credential"])
}
// Verify if the access key id matches.
if credential.accessKeyID != s.accessKeyID {
return false, ErrInvalidAccessKeyID("Access key id does not match with our records.", credential.accessKeyID).Trace(credential.accessKeyID)
}
// Verify if the region is valid.
reqRegion := credential.scope.region
if !isValidRegion(reqRegion, s.region) {
return false, ErrInvalidRegion("Requested region is not recognized.", reqRegion).Trace(reqRegion)
}
// Save region.
s.region = reqRegion
// Parse date string.
t, e := time.Parse(iso8601Format, formValues["X-Amz-Date"])
if e != nil {
return false, probe.NewError(e)
}
signingKey := s.getSigningKey(t)
newSignature := s.getSignature(signingKey, formValues["Policy"])
if newSignature != formValues["X-Amz-Signature"] {
return false, nil
}
return true, nil
}
// DoesPresignedSignatureMatch - Verify query headers with presigned signature
// - http://docs.aws.amazon.com/AmazonS3/latest/API/sigv4-query-string-auth.html
// returns true if matches, false otherwise. if error is not nil then it is always false
func (s *Sign) DoesPresignedSignatureMatch() (bool, *probe.Error) {
// Parse request query string.
preSignValues, err := parsePreSignV4(s.httpRequest.URL.Query())
if err != nil {
return false, err.Trace(s.httpRequest.URL.String())
}
// Verify if the access key id matches.
if preSignValues.Credential.accessKeyID != s.accessKeyID {
return false, ErrInvalidAccessKeyID("Access key id does not match with our records.", preSignValues.Credential.accessKeyID).Trace(preSignValues.Credential.accessKeyID)
}
// Verify if region is valid.
reqRegion := preSignValues.Credential.scope.region
if !isValidRegion(reqRegion, s.region) {
return false, ErrInvalidRegion("Requested region is not recognized.", reqRegion).Trace(reqRegion)
}
// Save region.
s.region = reqRegion
// Extract all the signed headers along with its values.
s.extractedSignedHeaders = extractSignedHeaders(preSignValues.SignedHeaders, s.httpRequest.Header)
// Construct new query.
query := make(url.Values)
query.Set("X-Amz-Algorithm", signV4Algorithm)
if time.Now().UTC().Sub(preSignValues.Date) > time.Duration(preSignValues.Expires) {
return false, ErrExpiredPresignRequest("Presigned request already expired, please initiate a new request.")
}
// Save the date and expires.
t := preSignValues.Date
expireSeconds := int(time.Duration(preSignValues.Expires) / time.Second)
// Construct the query.
query.Set("X-Amz-Date", t.Format(iso8601Format))
query.Set("X-Amz-Expires", strconv.Itoa(expireSeconds))
query.Set("X-Amz-SignedHeaders", s.getSignedHeaders(s.extractedSignedHeaders))
query.Set("X-Amz-Credential", s.accessKeyID+"/"+s.getScope(t))
// Save other headers available in the request parameters.
for k, v := range s.httpRequest.URL.Query() {
if strings.HasPrefix(strings.ToLower(k), "x-amz") | {
continue
} | conditional_block |
|
v4-signature.go | .
func New(accessKeyID, secretAccessKey, region string) (*Sign, *probe.Error) |
// SetHTTPRequestToVerify - sets the http request which needs to be verified.
func (s *Sign) SetHTTPRequestToVerify(r *http.Request) *Sign {
// Do not set http request if its 'nil'.
if r == nil {
return s
}
s.httpRequest = r
return s
}
// getCanonicalHeaders generate a list of request headers with their values
func (s Sign) getCanonicalHeaders(signedHeaders http.Header) string {
var headers []string
vals := make(http.Header)
for k, vv := range signedHeaders {
headers = append(headers, strings.ToLower(k))
vals[strings.ToLower(k)] = vv
}
headers = append(headers, "host")
sort.Strings(headers)
var buf bytes.Buffer
for _, k := range headers {
buf.WriteString(k)
buf.WriteByte(':')
switch {
case k == "host":
buf.WriteString(s.httpRequest.Host)
fallthrough
default:
for idx, v := range vals[k] {
if idx > 0 {
buf.WriteByte(',')
}
buf.WriteString(v)
}
buf.WriteByte('\n')
}
}
return buf.String()
}
// getSignedHeaders generate a string i.e alphabetically sorted, semicolon-separated list of lowercase request header names
func (s Sign) getSignedHeaders(signedHeaders http.Header) string {
var headers []string
for k := range signedHeaders {
headers = append(headers, strings.ToLower(k))
}
headers = append(headers, "host")
sort.Strings(headers)
return strings.Join(headers, ";")
}
// getCanonicalRequest generate a canonical request of style
//
// canonicalRequest =
// <HTTPMethod>\n
// <CanonicalURI>\n
// <CanonicalQueryString>\n
// <CanonicalHeaders>\n
// <SignedHeaders>\n
// <HashedPayload>
//
func (s *Sign) getCanonicalRequest() string {
payload := s.httpRequest.Header.Get(http.CanonicalHeaderKey("x-amz-content-sha256"))
s.httpRequest.URL.RawQuery = strings.Replace(s.httpRequest.URL.Query().Encode(), "+", "%20", -1)
encodedPath := getURLEncodedName(s.httpRequest.URL.Path)
// Convert any space strings back to "+".
encodedPath = strings.Replace(encodedPath, "+", "%20", -1)
canonicalRequest := strings.Join([]string{
s.httpRequest.Method,
encodedPath,
s.httpRequest.URL.RawQuery,
s.getCanonicalHeaders(s.extractedSignedHeaders),
s.getSignedHeaders(s.extractedSignedHeaders),
payload,
}, "\n")
return canonicalRequest
}
// getCanonicalRequest generate a canonical request of style
//
// canonicalRequest =
// <HTTPMethod>\n
// <CanonicalURI>\n
// <CanonicalQueryString>\n
// <CanonicalHeaders>\n
// <SignedHeaders>\n
// <HashedPayload>
//
func (s Sign) getPresignedCanonicalRequest(presignedQuery string) string {
rawQuery := strings.Replace(presignedQuery, "+", "%20", -1)
encodedPath := getURLEncodedName(s.httpRequest.URL.Path)
// Convert any space strings back to "+".
encodedPath = strings.Replace(encodedPath, "+", "%20", -1)
canonicalRequest := strings.Join([]string{
s.httpRequest.Method,
encodedPath,
rawQuery,
s.getCanonicalHeaders(s.extractedSignedHeaders),
s.getSignedHeaders(s.extractedSignedHeaders),
"UNSIGNED-PAYLOAD",
}, "\n")
return canonicalRequest
}
// getScope generate a string of a specific date, an AWS region, and a service.
func (s Sign) getScope(t time.Time) string {
scope := strings.Join([]string{
t.Format(yyyymmdd),
s.region,
"s3",
"aws4_request",
}, "/")
return scope
}
// getStringToSign a string based on selected query values.
func (s Sign) getStringToSign(canonicalRequest string, t time.Time) string {
stringToSign := signV4Algorithm + "\n" + t.Format(iso8601Format) + "\n"
stringToSign = stringToSign + s.getScope(t) + "\n"
canonicalRequestBytes := sha256.Sum256([]byte(canonicalRequest))
stringToSign = stringToSign + hex.EncodeToString(canonicalRequestBytes[:])
return stringToSign
}
// getSigningKey hmac seed to calculate final signature.
func (s Sign) getSigningKey(t time.Time) []byte {
secret := s.secretAccessKey
date := sumHMAC([]byte("AWS4"+secret), []byte(t.Format(yyyymmdd)))
region := sumHMAC(date, []byte(s.region))
service := sumHMAC(region, []byte("s3"))
signingKey := sumHMAC(service, []byte("aws4_request"))
return signingKey
}
// getSignature final signature in hexadecimal form.
func (s Sign) getSignature(signingKey []byte, stringToSign string) string {
return hex.EncodeToString(sumHMAC(signingKey, []byte(stringToSign)))
}
// DoesPolicySignatureMatch - Verify query headers with post policy
// - http://docs.aws.amazon.com/AmazonS3/latest/API/sigv4-HTTPPOSTConstructPolicy.html
// returns true if matches, false otherwise. if error is not nil then it is always false
func (s *Sign) DoesPolicySignatureMatch(formValues map[string]string) (bool, *probe.Error) {
// Parse credential tag.
credential, err := parseCredential("Credential=" + formValues["X-Amz-Credential"])
if err != nil {
return false, err.Trace(formValues["X-Amz-Credential"])
}
// Verify if the access key id matches.
if credential.accessKeyID != s.accessKeyID {
return false, ErrInvalidAccessKeyID("Access key id does not match with our records.", credential.accessKeyID).Trace(credential.accessKeyID)
}
// Verify if the region is valid.
reqRegion := credential.scope.region
if !isValidRegion(reqRegion, s.region) {
return false, ErrInvalidRegion("Requested region is not recognized.", reqRegion).Trace(reqRegion)
}
// Save region.
s.region = reqRegion
// Parse date string.
t, e := time.Parse(iso8601Format, formValues["X-Amz-Date"])
if e != nil {
return false, probe.NewError(e)
}
signingKey := s.getSigningKey(t)
newSignature := s.getSignature(signingKey, formValues["Policy"])
if newSignature != formValues["X-Amz-Signature"] {
return false, nil
}
return true, nil
}
// DoesPresignedSignatureMatch - Verify query headers with presigned signature
// - http://docs.aws.amazon.com/AmazonS3/latest/API/sigv4-query-string-auth.html
// returns true if matches, false otherwise. if error is not nil then it is always false
func (s *Sign) DoesPresignedSignatureMatch() (bool, *probe.Error) {
// Parse request query string.
preSignValues, err := parsePreSignV4(s.httpRequest.URL.Query())
if err != nil {
return false, err.Trace(s.httpRequest.URL.String())
}
// Verify if the access key id matches.
if preSignValues.Credential.accessKeyID != s.accessKeyID {
return false, ErrInvalidAccessKeyID("Access key id does not match with our records.", preSignValues.Credential.accessKeyID).Trace(preSignValues.Credential.accessKeyID)
}
// Verify if region is valid.
reqRegion := preSignValues.Credential.scope.region
if !isValidRegion(reqRegion, s.region) {
return false, ErrInvalidRegion("Requested region is not recognized.", reqRegion).Trace(reqRegion)
}
// Save region.
s.region = reqRegion
// Extract all the signed headers along with its values.
s.extractedSignedHeaders = extractSignedHeaders(preSignValues.SignedHeaders, s.httpRequest.Header)
// Construct new query.
query := make(url.Values)
query.Set("X-Amz-Algorithm", signV4Algorithm)
if time.Now().UTC().Sub(preSignValues.Date) > time.Duration(preSignValues.Expires) {
return false, ErrExpiredPresignRequest("Presigned request already expired, please initiate a new request.")
}
// Save the date and expires.
t := preSignValues.Date
expireSeconds := int(time.Duration(preSignValues.Expires) / time.Second)
// Construct the query.
query.Set("X-Amz-Date", t.Format(iso8601Format))
query.Set("X-Amz-Expires", strconv.Itoa | {
if !isValidAccessKey.MatchString(accessKeyID) {
return nil, ErrInvalidAccessKeyID("Invalid access key id.", accessKeyID).Trace(accessKeyID)
}
if !isValidSecretKey.MatchString(secretAccessKey) {
return nil, ErrInvalidAccessKeyID("Invalid secret key.", secretAccessKey).Trace(secretAccessKey)
}
if region == "" {
return nil, ErrRegionISEmpty("Region is empty.").Trace()
}
signature := &Sign{
accessKeyID: accessKeyID,
secretAccessKey: secretAccessKey,
region: region,
}
return signature, nil
} | identifier_body |
policy_row.ts | _element.js';
import {loadTimeData} from 'chrome://resources/js/load_time_data.js';
import {Conflict, PolicyConflictElement} from './policy_conflict.js';
import {getTemplate} from './policy_row.html.js';
export interface Policy {
ignored?: boolean;
name: string;
level: string;
link?: string;
scope: string;
source: string;
error: string;
warning: string;
info: string;
value: any;
deprecated?: boolean;
future?: boolean;
allSourcesMerged?: boolean;
conflicts?: Conflict[];
superseded?: Conflict[];
forSigninScreen: boolean;
isExtension: boolean;
}
export class | extends CustomElement {
static override get template() {
return getTemplate();
}
policy: Policy;
private unset_: boolean;
private hasErrors_: boolean;
private hasWarnings_: boolean;
private hasInfos_: boolean;
private hasConflicts_: boolean;
private hasSuperseded_: boolean;
private isMergedValue_: boolean;
private deprecated_: boolean;
private future_: boolean;
connectedCallback() {
const toggle = this.shadowRoot!.querySelector('.policy.row .toggle');
toggle!.addEventListener('click', () => this.toggleExpanded());
const copy = this.shadowRoot!.querySelector('.copy-value');
copy!.addEventListener('click', () => this.copyValue_());
this.setAttribute('role', 'rowgroup');
this.classList.add('policy-data');
}
initialize(policy: Policy) {
this.policy = policy;
this.unset_ = policy.value === undefined;
this.hasErrors_ = !!policy.error;
this.hasWarnings_ = !!policy.warning;
this.hasInfos_ = !!policy.info;
this.hasConflicts_ = !!policy.conflicts;
this.hasSuperseded_ = !!policy.superseded;
this.isMergedValue_ = !!policy.allSourcesMerged;
this.deprecated_ = !!policy.deprecated;
this.future_ = !!policy.future;
// Populate the name column.
const nameDisplay = this.shadowRoot!.querySelector('.name .link span');
nameDisplay!.textContent = policy.name;
if (policy.link) {
const link =
this.shadowRoot!.querySelector('.name .link') as HTMLAnchorElement;
link.href = policy.link;
link.title = loadTimeData.getStringF('policyLearnMore', policy.name);
this.toggleAttribute('no-help-link', false);
} else {
this.toggleAttribute('no-help-link', true);
}
// Populate the remaining columns with policy scope, level and value if a
// value has been set. Otherwise, leave them blank.
if (!this.unset_) {
const scopeDisplay = this.shadowRoot!.querySelector('.scope');
let scope = 'scopeDevice';
if (policy.scope === 'user') {
scope = 'scopeUser';
} else if (policy.scope === 'allUsers') {
scope = 'scopeAllUsers';
}
scopeDisplay!.textContent = loadTimeData.getString(scope);
// Display scope and level as rows instead of columns on space constraint
// devices.
// <if expr="is_android or is_ios">
const scopeRowContentDisplay =
this.shadowRoot!.querySelector('.scope.row .value');
scopeRowContentDisplay!.textContent = loadTimeData.getString(scope);
const levelRowContentDisplay =
this.shadowRoot!.querySelector('.level.row .value');
levelRowContentDisplay!.textContent = loadTimeData.getString(
policy.level === 'recommended' ? 'levelRecommended' :
'levelMandatory');
// </if>
const levelDisplay = this.shadowRoot!.querySelector('.level');
levelDisplay!.textContent = loadTimeData.getString(
policy.level === 'recommended' ? 'levelRecommended' :
'levelMandatory');
const sourceDisplay = this.shadowRoot!.querySelector('.source');
sourceDisplay!.textContent = loadTimeData.getString(policy.source);
// Reduces load on the DOM for long values;
const convertValue = (value: string, format?: boolean) => {
// Skip 'string' policy to avoid unnecessary conversions.
if (typeof value == 'string') {
return value;
}
if (format) {
return JSON.stringify(value, null, 2);
} else {
return JSON.stringify(value, null);
}
};
// If value is longer than 256 characters, truncate and add ellipsis.
const policyValueStr = convertValue(policy.value);
const truncatedValue = policyValueStr.length > 256 ?
`${policyValueStr.substring(0, 256)}\u2026` :
policyValueStr;
const valueDisplay = this.shadowRoot!.querySelector('.value');
valueDisplay!.textContent = truncatedValue;
const copyLink =
this.shadowRoot!.querySelector('.copy .link') as HTMLElement;
copyLink!.title = loadTimeData.getStringF('policyCopyValue', policy.name);
const valueRowContentDisplay =
this.shadowRoot!.querySelector('.value.row .value');
// Expanded policy value is formatted.
valueRowContentDisplay!.textContent =
convertValue(policy.value, /*format=*/ true);
const errorRowContentDisplay =
this.shadowRoot!.querySelector('.errors.row .value');
errorRowContentDisplay!.textContent = policy.error;
const warningRowContentDisplay =
this.shadowRoot!.querySelector('.warnings.row .value');
warningRowContentDisplay!.textContent = policy.warning;
const infoRowContentDisplay =
this.shadowRoot!.querySelector('.infos.row .value');
infoRowContentDisplay!.textContent = policy.info;
const messagesDisplay = this.shadowRoot!.querySelector('.messages');
const errorsNotice =
this.hasErrors_ ? loadTimeData.getString('error') : '';
const deprecationNotice =
this.deprecated_ ? loadTimeData.getString('deprecated') : '';
const futureNotice = this.future_ ? loadTimeData.getString('future') : '';
const warningsNotice =
this.hasWarnings_ ? loadTimeData.getString('warning') : '';
const conflictsNotice = this.hasConflicts_ && !this.isMergedValue_ ?
loadTimeData.getString('conflict') :
'';
const ignoredNotice =
this.policy.ignored ? loadTimeData.getString('ignored') : '';
let notice =
[
errorsNotice,
deprecationNotice,
futureNotice,
warningsNotice,
ignoredNotice,
conflictsNotice,
].filter(x => !!x)
.join(', ') ||
loadTimeData.getString('ok');
const supersededNotice = this.hasSuperseded_ && !this.isMergedValue_ ?
loadTimeData.getString('superseding') :
'';
if (supersededNotice) {
// Include superseded notice regardless of other notices
notice += `, ${supersededNotice}`;
}
messagesDisplay!.textContent = notice;
if (policy.conflicts) {
policy.conflicts.forEach(conflict => {
const row = document.createElement('policy-conflict') as
PolicyConflictElement;
row.initialize(conflict, 'conflictValue');
row.classList!.add('policy-conflict-data');
this.shadowRoot!.appendChild(row);
});
}
if (policy.superseded) {
policy.superseded.forEach(superseded => {
const row = document.createElement('policy-conflict') as
PolicyConflictElement;
row.initialize(superseded, 'supersededValue');
row.classList!.add('policy-superseded-data');
this.shadowRoot!.appendChild(row);
});
}
} else {
const messagesDisplay = this.shadowRoot!.querySelector('.messages');
messagesDisplay!.textContent = loadTimeData.getString('unset');
}
}
// Copies the policy's value to the clipboard.
private copyValue_() {
const policyValueDisplay =
this.shadowRoot!.querySelector('.value.row .value');
// Select the text that will be copied.
const selection = window.getSelection();
const range = window.document.createRange();
range.selectNodeContents(policyValueDisplay as Node);
selection!.removeAllRanges();
selection!.addRange(range);
// Copy the policy value to the clipboard.
navigator.clipboard
.writeText((policyValueDisplay as CustomElement)!.innerText)
.catch(error => {
console.error('Unable to copy policy value to clipboard:', error);
});
}
// Toggle the visibility of an additional row containing the complete text.
private toggleExpanded() {
const warningRowDisplay =
this.shadowRoot!.querySelector('.warnings.row') as CustomElement;
const errorRowDisplay =
this.shadowRoot!.querySelector('.errors.row') as CustomElement;
const infoRowDisplay =
this.shadowRoot!.querySelector('.infos.row') as CustomElement;
const valueRowDisplay =
this.shadowRoot!.querySelector('.value.row') as CustomElement;
// <if expr="is_android or is_ios">
const scopeRowDisplay =
this.shadowRoot!.querySelector('.scope.row') as CustomElement;
scopeRowDisplay.hidden = !scopeRowDisplay.hidden;
const levelRowDisplay =
this.shadowRoot!.querySelector('.level.row') as CustomElement;
levelRowDisplay.hidden = !levelRowDisplay.hidden;
// </if>
valueRowDisplay.hidden = !valueRowDisplay.hidden;
this.classList!.toggle('expanded', !valueRowDisplay.hidden);
this.shadowRoot!.querySelector<CustomElement>('.show-more')!.hidden =
!valueRowDisplay.hidden;
this.shadowRoot!.querySelector<CustomElement>('.show-less')!.hidden =
valueRowDisplay!.hidden;
| PolicyRowElement | identifier_name |
policy_row.ts | _element.js';
import {loadTimeData} from 'chrome://resources/js/load_time_data.js';
import {Conflict, PolicyConflictElement} from './policy_conflict.js';
import {getTemplate} from './policy_row.html.js';
export interface Policy {
ignored?: boolean;
name: string;
level: string;
link?: string;
scope: string;
source: string;
error: string;
warning: string;
info: string;
value: any;
deprecated?: boolean;
future?: boolean;
allSourcesMerged?: boolean;
conflicts?: Conflict[];
superseded?: Conflict[];
forSigninScreen: boolean;
isExtension: boolean;
}
export class PolicyRowElement extends CustomElement {
static override get template() {
return getTemplate();
}
policy: Policy;
private unset_: boolean;
private hasErrors_: boolean;
private hasWarnings_: boolean;
private hasInfos_: boolean;
private hasConflicts_: boolean;
private hasSuperseded_: boolean;
private isMergedValue_: boolean;
private deprecated_: boolean;
private future_: boolean;
connectedCallback() {
const toggle = this.shadowRoot!.querySelector('.policy.row .toggle');
toggle!.addEventListener('click', () => this.toggleExpanded());
const copy = this.shadowRoot!.querySelector('.copy-value');
copy!.addEventListener('click', () => this.copyValue_());
this.setAttribute('role', 'rowgroup');
this.classList.add('policy-data');
}
initialize(policy: Policy) {
this.policy = policy;
this.unset_ = policy.value === undefined;
this.hasErrors_ = !!policy.error;
this.hasWarnings_ = !!policy.warning;
this.hasInfos_ = !!policy.info;
this.hasConflicts_ = !!policy.conflicts;
this.hasSuperseded_ = !!policy.superseded;
this.isMergedValue_ = !!policy.allSourcesMerged;
this.deprecated_ = !!policy.deprecated;
this.future_ = !!policy.future;
// Populate the name column.
const nameDisplay = this.shadowRoot!.querySelector('.name .link span');
nameDisplay!.textContent = policy.name;
if (policy.link) {
const link =
this.shadowRoot!.querySelector('.name .link') as HTMLAnchorElement;
link.href = policy.link;
link.title = loadTimeData.getStringF('policyLearnMore', policy.name);
this.toggleAttribute('no-help-link', false);
} else {
this.toggleAttribute('no-help-link', true);
}
// Populate the remaining columns with policy scope, level and value if a
// value has been set. Otherwise, leave them blank.
if (!this.unset_) {
const scopeDisplay = this.shadowRoot!.querySelector('.scope');
let scope = 'scopeDevice';
if (policy.scope === 'user') {
scope = 'scopeUser';
} else if (policy.scope === 'allUsers') {
scope = 'scopeAllUsers';
}
scopeDisplay!.textContent = loadTimeData.getString(scope);
// Display scope and level as rows instead of columns on space constraint
// devices.
// <if expr="is_android or is_ios">
const scopeRowContentDisplay =
this.shadowRoot!.querySelector('.scope.row .value');
scopeRowContentDisplay!.textContent = loadTimeData.getString(scope);
const levelRowContentDisplay =
this.shadowRoot!.querySelector('.level.row .value');
levelRowContentDisplay!.textContent = loadTimeData.getString(
policy.level === 'recommended' ? 'levelRecommended' :
'levelMandatory');
// </if>
const levelDisplay = this.shadowRoot!.querySelector('.level');
levelDisplay!.textContent = loadTimeData.getString(
policy.level === 'recommended' ? 'levelRecommended' :
'levelMandatory');
const sourceDisplay = this.shadowRoot!.querySelector('.source');
sourceDisplay!.textContent = loadTimeData.getString(policy.source);
// Reduces load on the DOM for long values;
const convertValue = (value: string, format?: boolean) => {
// Skip 'string' policy to avoid unnecessary conversions.
if (typeof value == 'string') |
if (format) {
return JSON.stringify(value, null, 2);
} else {
return JSON.stringify(value, null);
}
};
// If value is longer than 256 characters, truncate and add ellipsis.
const policyValueStr = convertValue(policy.value);
const truncatedValue = policyValueStr.length > 256 ?
`${policyValueStr.substring(0, 256)}\u2026` :
policyValueStr;
const valueDisplay = this.shadowRoot!.querySelector('.value');
valueDisplay!.textContent = truncatedValue;
const copyLink =
this.shadowRoot!.querySelector('.copy .link') as HTMLElement;
copyLink!.title = loadTimeData.getStringF('policyCopyValue', policy.name);
const valueRowContentDisplay =
this.shadowRoot!.querySelector('.value.row .value');
// Expanded policy value is formatted.
valueRowContentDisplay!.textContent =
convertValue(policy.value, /*format=*/ true);
const errorRowContentDisplay =
this.shadowRoot!.querySelector('.errors.row .value');
errorRowContentDisplay!.textContent = policy.error;
const warningRowContentDisplay =
this.shadowRoot!.querySelector('.warnings.row .value');
warningRowContentDisplay!.textContent = policy.warning;
const infoRowContentDisplay =
this.shadowRoot!.querySelector('.infos.row .value');
infoRowContentDisplay!.textContent = policy.info;
const messagesDisplay = this.shadowRoot!.querySelector('.messages');
const errorsNotice =
this.hasErrors_ ? loadTimeData.getString('error') : '';
const deprecationNotice =
this.deprecated_ ? loadTimeData.getString('deprecated') : '';
const futureNotice = this.future_ ? loadTimeData.getString('future') : '';
const warningsNotice =
this.hasWarnings_ ? loadTimeData.getString('warning') : '';
const conflictsNotice = this.hasConflicts_ && !this.isMergedValue_ ?
loadTimeData.getString('conflict') :
'';
const ignoredNotice =
this.policy.ignored ? loadTimeData.getString('ignored') : '';
let notice =
[
errorsNotice,
deprecationNotice,
futureNotice,
warningsNotice,
ignoredNotice,
conflictsNotice,
].filter(x => !!x)
.join(', ') ||
loadTimeData.getString('ok');
const supersededNotice = this.hasSuperseded_ && !this.isMergedValue_ ?
loadTimeData.getString('superseding') :
'';
if (supersededNotice) {
// Include superseded notice regardless of other notices
notice += `, ${supersededNotice}`;
}
messagesDisplay!.textContent = notice;
if (policy.conflicts) {
policy.conflicts.forEach(conflict => {
const row = document.createElement('policy-conflict') as
PolicyConflictElement;
row.initialize(conflict, 'conflictValue');
row.classList!.add('policy-conflict-data');
this.shadowRoot!.appendChild(row);
});
}
if (policy.superseded) {
policy.superseded.forEach(superseded => {
const row = document.createElement('policy-conflict') as
PolicyConflictElement;
row.initialize(superseded, 'supersededValue');
row.classList!.add('policy-superseded-data');
this.shadowRoot!.appendChild(row);
});
}
} else {
const messagesDisplay = this.shadowRoot!.querySelector('.messages');
messagesDisplay!.textContent = loadTimeData.getString('unset');
}
}
// Copies the policy's value to the clipboard.
private copyValue_() {
const policyValueDisplay =
this.shadowRoot!.querySelector('.value.row .value');
// Select the text that will be copied.
const selection = window.getSelection();
const range = window.document.createRange();
range.selectNodeContents(policyValueDisplay as Node);
selection!.removeAllRanges();
selection!.addRange(range);
// Copy the policy value to the clipboard.
navigator.clipboard
.writeText((policyValueDisplay as CustomElement)!.innerText)
.catch(error => {
console.error('Unable to copy policy value to clipboard:', error);
});
}
// Toggle the visibility of an additional row containing the complete text.
private toggleExpanded() {
const warningRowDisplay =
this.shadowRoot!.querySelector('.warnings.row') as CustomElement;
const errorRowDisplay =
this.shadowRoot!.querySelector('.errors.row') as CustomElement;
const infoRowDisplay =
this.shadowRoot!.querySelector('.infos.row') as CustomElement;
const valueRowDisplay =
this.shadowRoot!.querySelector('.value.row') as CustomElement;
// <if expr="is_android or is_ios">
const scopeRowDisplay =
this.shadowRoot!.querySelector('.scope.row') as CustomElement;
scopeRowDisplay.hidden = !scopeRowDisplay.hidden;
const levelRowDisplay =
this.shadowRoot!.querySelector('.level.row') as CustomElement;
levelRowDisplay.hidden = !levelRowDisplay.hidden;
// </if>
valueRowDisplay.hidden = !valueRowDisplay.hidden;
this.classList!.toggle('expanded', !valueRowDisplay.hidden);
this.shadowRoot!.querySelector<CustomElement>('.show-more')!.hidden =
!valueRowDisplay.hidden;
this.shadowRoot!.querySelector<CustomElement>('.show-less')!.hidden =
valueRowDisplay!.hidden;
| {
return value;
} | conditional_block |
policy_row.ts | _element.js';
import {loadTimeData} from 'chrome://resources/js/load_time_data.js';
import {Conflict, PolicyConflictElement} from './policy_conflict.js';
import {getTemplate} from './policy_row.html.js';
export interface Policy {
ignored?: boolean;
name: string;
level: string;
link?: string;
scope: string;
source: string;
error: string;
warning: string;
info: string;
value: any;
deprecated?: boolean;
future?: boolean;
allSourcesMerged?: boolean;
conflicts?: Conflict[];
superseded?: Conflict[];
forSigninScreen: boolean;
isExtension: boolean;
}
export class PolicyRowElement extends CustomElement {
static override get template() |
policy: Policy;
private unset_: boolean;
private hasErrors_: boolean;
private hasWarnings_: boolean;
private hasInfos_: boolean;
private hasConflicts_: boolean;
private hasSuperseded_: boolean;
private isMergedValue_: boolean;
private deprecated_: boolean;
private future_: boolean;
connectedCallback() {
const toggle = this.shadowRoot!.querySelector('.policy.row .toggle');
toggle!.addEventListener('click', () => this.toggleExpanded());
const copy = this.shadowRoot!.querySelector('.copy-value');
copy!.addEventListener('click', () => this.copyValue_());
this.setAttribute('role', 'rowgroup');
this.classList.add('policy-data');
}
initialize(policy: Policy) {
this.policy = policy;
this.unset_ = policy.value === undefined;
this.hasErrors_ = !!policy.error;
this.hasWarnings_ = !!policy.warning;
this.hasInfos_ = !!policy.info;
this.hasConflicts_ = !!policy.conflicts;
this.hasSuperseded_ = !!policy.superseded;
this.isMergedValue_ = !!policy.allSourcesMerged;
this.deprecated_ = !!policy.deprecated;
this.future_ = !!policy.future;
// Populate the name column.
const nameDisplay = this.shadowRoot!.querySelector('.name .link span');
nameDisplay!.textContent = policy.name;
if (policy.link) {
const link =
this.shadowRoot!.querySelector('.name .link') as HTMLAnchorElement;
link.href = policy.link;
link.title = loadTimeData.getStringF('policyLearnMore', policy.name);
this.toggleAttribute('no-help-link', false);
} else {
this.toggleAttribute('no-help-link', true);
}
// Populate the remaining columns with policy scope, level and value if a
// value has been set. Otherwise, leave them blank.
if (!this.unset_) {
const scopeDisplay = this.shadowRoot!.querySelector('.scope');
let scope = 'scopeDevice';
if (policy.scope === 'user') {
scope = 'scopeUser';
} else if (policy.scope === 'allUsers') {
scope = 'scopeAllUsers';
}
scopeDisplay!.textContent = loadTimeData.getString(scope);
// Display scope and level as rows instead of columns on space constraint
// devices.
// <if expr="is_android or is_ios">
const scopeRowContentDisplay =
this.shadowRoot!.querySelector('.scope.row .value');
scopeRowContentDisplay!.textContent = loadTimeData.getString(scope);
const levelRowContentDisplay =
this.shadowRoot!.querySelector('.level.row .value');
levelRowContentDisplay!.textContent = loadTimeData.getString(
policy.level === 'recommended' ? 'levelRecommended' :
'levelMandatory');
// </if>
const levelDisplay = this.shadowRoot!.querySelector('.level');
levelDisplay!.textContent = loadTimeData.getString(
policy.level === 'recommended' ? 'levelRecommended' :
'levelMandatory');
const sourceDisplay = this.shadowRoot!.querySelector('.source');
sourceDisplay!.textContent = loadTimeData.getString(policy.source);
// Reduces load on the DOM for long values;
const convertValue = (value: string, format?: boolean) => {
// Skip 'string' policy to avoid unnecessary conversions.
if (typeof value == 'string') {
return value;
}
if (format) {
return JSON.stringify(value, null, 2);
} else {
return JSON.stringify(value, null);
}
};
// If value is longer than 256 characters, truncate and add ellipsis.
const policyValueStr = convertValue(policy.value);
const truncatedValue = policyValueStr.length > 256 ?
`${policyValueStr.substring(0, 256)}\u2026` :
policyValueStr;
const valueDisplay = this.shadowRoot!.querySelector('.value');
valueDisplay!.textContent = truncatedValue;
const copyLink =
this.shadowRoot!.querySelector('.copy .link') as HTMLElement;
copyLink!.title = loadTimeData.getStringF('policyCopyValue', policy.name);
const valueRowContentDisplay =
this.shadowRoot!.querySelector('.value.row .value');
// Expanded policy value is formatted.
valueRowContentDisplay!.textContent =
convertValue(policy.value, /*format=*/ true);
const errorRowContentDisplay =
this.shadowRoot!.querySelector('.errors.row .value');
errorRowContentDisplay!.textContent = policy.error;
const warningRowContentDisplay =
this.shadowRoot!.querySelector('.warnings.row .value');
warningRowContentDisplay!.textContent = policy.warning;
const infoRowContentDisplay =
this.shadowRoot!.querySelector('.infos.row .value');
infoRowContentDisplay!.textContent = policy.info;
const messagesDisplay = this.shadowRoot!.querySelector('.messages');
const errorsNotice =
this.hasErrors_ ? loadTimeData.getString('error') : '';
const deprecationNotice =
this.deprecated_ ? loadTimeData.getString('deprecated') : '';
const futureNotice = this.future_ ? loadTimeData.getString('future') : '';
const warningsNotice =
this.hasWarnings_ ? loadTimeData.getString('warning') : '';
const conflictsNotice = this.hasConflicts_ && !this.isMergedValue_ ?
loadTimeData.getString('conflict') :
'';
const ignoredNotice =
this.policy.ignored ? loadTimeData.getString('ignored') : '';
let notice =
[
errorsNotice,
deprecationNotice,
futureNotice,
warningsNotice,
ignoredNotice,
conflictsNotice,
].filter(x => !!x)
.join(', ') ||
loadTimeData.getString('ok');
const supersededNotice = this.hasSuperseded_ && !this.isMergedValue_ ?
loadTimeData.getString('superseding') :
'';
if (supersededNotice) {
// Include superseded notice regardless of other notices
notice += `, ${supersededNotice}`;
}
messagesDisplay!.textContent = notice;
if (policy.conflicts) {
policy.conflicts.forEach(conflict => {
const row = document.createElement('policy-conflict') as
PolicyConflictElement;
row.initialize(conflict, 'conflictValue');
row.classList!.add('policy-conflict-data');
this.shadowRoot!.appendChild(row);
});
}
if (policy.superseded) {
policy.superseded.forEach(superseded => {
const row = document.createElement('policy-conflict') as
PolicyConflictElement;
row.initialize(superseded, 'supersededValue');
row.classList!.add('policy-superseded-data');
this.shadowRoot!.appendChild(row);
});
}
} else {
const messagesDisplay = this.shadowRoot!.querySelector('.messages');
messagesDisplay!.textContent = loadTimeData.getString('unset');
}
}
// Copies the policy's value to the clipboard.
private copyValue_() {
const policyValueDisplay =
this.shadowRoot!.querySelector('.value.row .value');
// Select the text that will be copied.
const selection = window.getSelection();
const range = window.document.createRange();
range.selectNodeContents(policyValueDisplay as Node);
selection!.removeAllRanges();
selection!.addRange(range);
// Copy the policy value to the clipboard.
navigator.clipboard
.writeText((policyValueDisplay as CustomElement)!.innerText)
.catch(error => {
console.error('Unable to copy policy value to clipboard:', error);
});
}
// Toggle the visibility of an additional row containing the complete text.
private toggleExpanded() {
const warningRowDisplay =
this.shadowRoot!.querySelector('.warnings.row') as CustomElement;
const errorRowDisplay =
this.shadowRoot!.querySelector('.errors.row') as CustomElement;
const infoRowDisplay =
this.shadowRoot!.querySelector('.infos.row') as CustomElement;
const valueRowDisplay =
this.shadowRoot!.querySelector('.value.row') as CustomElement;
// <if expr="is_android or is_ios">
const scopeRowDisplay =
this.shadowRoot!.querySelector('.scope.row') as CustomElement;
scopeRowDisplay.hidden = !scopeRowDisplay.hidden;
const levelRowDisplay =
this.shadowRoot!.querySelector('.level.row') as CustomElement;
levelRowDisplay.hidden = !levelRowDisplay.hidden;
// </if>
valueRowDisplay.hidden = !valueRowDisplay.hidden;
this.classList!.toggle('expanded', !valueRowDisplay.hidden);
this.shadowRoot!.querySelector<CustomElement>('.show-more')!.hidden =
!valueRowDisplay.hidden;
this.shadowRoot!.querySelector<CustomElement>('.show-less')!.hidden =
valueRowDisplay!.hidden;
| {
return getTemplate();
} | identifier_body |
policy_row.ts | /custom_element.js';
import {loadTimeData} from 'chrome://resources/js/load_time_data.js';
import {Conflict, PolicyConflictElement} from './policy_conflict.js';
import {getTemplate} from './policy_row.html.js';
export interface Policy {
ignored?: boolean;
name: string;
level: string;
link?: string;
scope: string;
source: string;
error: string;
warning: string;
info: string;
value: any;
deprecated?: boolean;
future?: boolean;
allSourcesMerged?: boolean;
conflicts?: Conflict[];
superseded?: Conflict[];
forSigninScreen: boolean;
isExtension: boolean;
}
export class PolicyRowElement extends CustomElement {
static override get template() {
return getTemplate();
}
policy: Policy;
private unset_: boolean;
private hasErrors_: boolean;
private hasWarnings_: boolean;
private hasInfos_: boolean;
private hasConflicts_: boolean;
private hasSuperseded_: boolean;
private isMergedValue_: boolean;
private deprecated_: boolean;
private future_: boolean;
connectedCallback() {
const toggle = this.shadowRoot!.querySelector('.policy.row .toggle');
toggle!.addEventListener('click', () => this.toggleExpanded());
const copy = this.shadowRoot!.querySelector('.copy-value');
copy!.addEventListener('click', () => this.copyValue_());
this.setAttribute('role', 'rowgroup');
this.classList.add('policy-data');
}
initialize(policy: Policy) {
this.policy = policy;
this.unset_ = policy.value === undefined;
this.hasErrors_ = !!policy.error;
this.hasWarnings_ = !!policy.warning;
this.hasInfos_ = !!policy.info;
this.hasConflicts_ = !!policy.conflicts;
this.hasSuperseded_ = !!policy.superseded;
this.isMergedValue_ = !!policy.allSourcesMerged;
this.deprecated_ = !!policy.deprecated;
this.future_ = !!policy.future;
// Populate the name column.
const nameDisplay = this.shadowRoot!.querySelector('.name .link span');
nameDisplay!.textContent = policy.name;
if (policy.link) {
const link =
this.shadowRoot!.querySelector('.name .link') as HTMLAnchorElement;
link.href = policy.link;
link.title = loadTimeData.getStringF('policyLearnMore', policy.name);
this.toggleAttribute('no-help-link', false);
} else {
this.toggleAttribute('no-help-link', true);
}
// Populate the remaining columns with policy scope, level and value if a
// value has been set. Otherwise, leave them blank.
if (!this.unset_) {
const scopeDisplay = this.shadowRoot!.querySelector('.scope');
let scope = 'scopeDevice';
if (policy.scope === 'user') {
scope = 'scopeUser';
} else if (policy.scope === 'allUsers') {
scope = 'scopeAllUsers';
}
scopeDisplay!.textContent = loadTimeData.getString(scope);
// Display scope and level as rows instead of columns on space constraint
// devices.
// <if expr="is_android or is_ios">
const scopeRowContentDisplay =
this.shadowRoot!.querySelector('.scope.row .value');
scopeRowContentDisplay!.textContent = loadTimeData.getString(scope);
const levelRowContentDisplay =
this.shadowRoot!.querySelector('.level.row .value');
levelRowContentDisplay!.textContent = loadTimeData.getString(
policy.level === 'recommended' ? 'levelRecommended' :
'levelMandatory');
// </if>
const levelDisplay = this.shadowRoot!.querySelector('.level');
levelDisplay!.textContent = loadTimeData.getString(
policy.level === 'recommended' ? 'levelRecommended' :
'levelMandatory');
const sourceDisplay = this.shadowRoot!.querySelector('.source');
sourceDisplay!.textContent = loadTimeData.getString(policy.source);
// Reduces load on the DOM for long values;
const convertValue = (value: string, format?: boolean) => { | }
if (format) {
return JSON.stringify(value, null, 2);
} else {
return JSON.stringify(value, null);
}
};
// If value is longer than 256 characters, truncate and add ellipsis.
const policyValueStr = convertValue(policy.value);
const truncatedValue = policyValueStr.length > 256 ?
`${policyValueStr.substring(0, 256)}\u2026` :
policyValueStr;
const valueDisplay = this.shadowRoot!.querySelector('.value');
valueDisplay!.textContent = truncatedValue;
const copyLink =
this.shadowRoot!.querySelector('.copy .link') as HTMLElement;
copyLink!.title = loadTimeData.getStringF('policyCopyValue', policy.name);
const valueRowContentDisplay =
this.shadowRoot!.querySelector('.value.row .value');
// Expanded policy value is formatted.
valueRowContentDisplay!.textContent =
convertValue(policy.value, /*format=*/ true);
const errorRowContentDisplay =
this.shadowRoot!.querySelector('.errors.row .value');
errorRowContentDisplay!.textContent = policy.error;
const warningRowContentDisplay =
this.shadowRoot!.querySelector('.warnings.row .value');
warningRowContentDisplay!.textContent = policy.warning;
const infoRowContentDisplay =
this.shadowRoot!.querySelector('.infos.row .value');
infoRowContentDisplay!.textContent = policy.info;
const messagesDisplay = this.shadowRoot!.querySelector('.messages');
const errorsNotice =
this.hasErrors_ ? loadTimeData.getString('error') : '';
const deprecationNotice =
this.deprecated_ ? loadTimeData.getString('deprecated') : '';
const futureNotice = this.future_ ? loadTimeData.getString('future') : '';
const warningsNotice =
this.hasWarnings_ ? loadTimeData.getString('warning') : '';
const conflictsNotice = this.hasConflicts_ && !this.isMergedValue_ ?
loadTimeData.getString('conflict') :
'';
const ignoredNotice =
this.policy.ignored ? loadTimeData.getString('ignored') : '';
let notice =
[
errorsNotice,
deprecationNotice,
futureNotice,
warningsNotice,
ignoredNotice,
conflictsNotice,
].filter(x => !!x)
.join(', ') ||
loadTimeData.getString('ok');
const supersededNotice = this.hasSuperseded_ && !this.isMergedValue_ ?
loadTimeData.getString('superseding') :
'';
if (supersededNotice) {
// Include superseded notice regardless of other notices
notice += `, ${supersededNotice}`;
}
messagesDisplay!.textContent = notice;
if (policy.conflicts) {
policy.conflicts.forEach(conflict => {
const row = document.createElement('policy-conflict') as
PolicyConflictElement;
row.initialize(conflict, 'conflictValue');
row.classList!.add('policy-conflict-data');
this.shadowRoot!.appendChild(row);
});
}
if (policy.superseded) {
policy.superseded.forEach(superseded => {
const row = document.createElement('policy-conflict') as
PolicyConflictElement;
row.initialize(superseded, 'supersededValue');
row.classList!.add('policy-superseded-data');
this.shadowRoot!.appendChild(row);
});
}
} else {
const messagesDisplay = this.shadowRoot!.querySelector('.messages');
messagesDisplay!.textContent = loadTimeData.getString('unset');
}
}
// Copies the policy's value to the clipboard.
private copyValue_() {
const policyValueDisplay =
this.shadowRoot!.querySelector('.value.row .value');
// Select the text that will be copied.
const selection = window.getSelection();
const range = window.document.createRange();
range.selectNodeContents(policyValueDisplay as Node);
selection!.removeAllRanges();
selection!.addRange(range);
// Copy the policy value to the clipboard.
navigator.clipboard
.writeText((policyValueDisplay as CustomElement)!.innerText)
.catch(error => {
console.error('Unable to copy policy value to clipboard:', error);
});
}
// Toggle the visibility of an additional row containing the complete text.
private toggleExpanded() {
const warningRowDisplay =
this.shadowRoot!.querySelector('.warnings.row') as CustomElement;
const errorRowDisplay =
this.shadowRoot!.querySelector('.errors.row') as CustomElement;
const infoRowDisplay =
this.shadowRoot!.querySelector('.infos.row') as CustomElement;
const valueRowDisplay =
this.shadowRoot!.querySelector('.value.row') as CustomElement;
// <if expr="is_android or is_ios">
const scopeRowDisplay =
this.shadowRoot!.querySelector('.scope.row') as CustomElement;
scopeRowDisplay.hidden = !scopeRowDisplay.hidden;
const levelRowDisplay =
this.shadowRoot!.querySelector('.level.row') as CustomElement;
levelRowDisplay.hidden = !levelRowDisplay.hidden;
// </if>
valueRowDisplay.hidden = !valueRowDisplay.hidden;
this.classList!.toggle('expanded', !valueRowDisplay.hidden);
this.shadowRoot!.querySelector<CustomElement>('.show-more')!.hidden =
!valueRowDisplay.hidden;
this.shadowRoot!.querySelector<CustomElement>('.show-less')!.hidden =
valueRowDisplay!.hidden;
| // Skip 'string' policy to avoid unnecessary conversions.
if (typeof value == 'string') {
return value; | random_line_split |
server.go | ok := privateKey.(*ecdsa.PrivateKey); ok {
keyType = cert.ECPrivateKeyBlockType
bytes, err = x509.MarshalECPrivateKey(key)
} else if key, ok := privateKey.(*rsa.PrivateKey); ok {
keyType = cert.RSAPrivateKeyBlockType
bytes = x509.MarshalPKCS1PrivateKey(key)
} else {
keyType = cert.PrivateKeyBlockType
bytes, err = x509.MarshalPKCS8PrivateKey(privateKey)
}
if err != nil {
logrus.Errorf("Unable to marshal private key: %v", err)
}
return keyType, bytes, err
}
func newPrivateKey() (crypto.Signer, error) {
caKeyBytes, err := cert.MakeEllipticPrivateKeyPEM()
if err != nil {
return nil, err
}
caKeyIFace, err := cert.ParsePrivateKeyPEM(caKeyBytes)
if err != nil {
return nil, err
}
return caKeyIFace.(crypto.Signer), nil
}
func (s *server) save() (_err error) {
defer func() {
if _err != nil {
logrus.Errorf("Saving cert error: %s", _err)
}
}()
certStr, err := certToString(s.tlsCert)
if err != nil {
return err
}
cfg, err := s.listenConfigStorage.Get()
if err != nil {
return err
}
cfg.GeneratedCerts = map[string]string{s.cn: certStr}
_, err = s.listenConfigStorage.Set(cfg)
return err
}
func (s *server) userConfigure() error {
if s.userConfig.HTTPSPort == 0 {
s.userConfig.HTTPSPort = 8443
}
for _, d := range s.userConfig.Domains {
s.domains[d] = true
}
for _, ip := range s.userConfig.KnownIPs {
if netIP := net.ParseIP(ip); netIP != nil {
s.ips[ip] = true
}
}
if bindAddress := net.ParseIP(s.userConfig.BindAddress); bindAddress != nil {
s.ips[s.userConfig.BindAddress] = true
}
if s.activeCA == nil && s.activeCAKey == nil {
if s.userConfig.CACerts != "" && s.userConfig.CAKey != "" {
ca, err := cert.ParseCertsPEM([]byte(s.userConfig.CACerts))
if err != nil {
return err
}
key, err := cert.ParsePrivateKeyPEM([]byte(s.userConfig.CAKey))
if err != nil {
return err
}
s.activeCA = ca[0]
s.activeCAKey = key.(crypto.Signer)
} else {
ca, key, err := genCA()
if err != nil {
return err
}
s.activeCA = ca
s.activeCAKey = key
}
}
return nil
}
func genCA() (*x509.Certificate, crypto.Signer, error) {
caKey, err := newPrivateKey()
if err != nil {
return nil, nil, err
}
caCert, err := cert.NewSelfSignedCACert(cert.Config{
CommonName: "k3s-ca",
Organization: []string{"k3s-org"},
}, caKey)
if err != nil {
return nil, nil, err
}
return caCert, caKey, nil
}
func (s *server) Update(status *ListenerStatus) (_err error) {
s.Lock()
defer func() {
s.Unlock()
if _err != nil {
logrus.Errorf("Update cert error: %s", _err)
}
if s.tlsCert == nil {
s.getCertificate(&tls.ClientHelloInfo{ServerName: "localhost"})
}
}()
certString := status.GeneratedCerts[s.cn]
tlsCert, err := stringToCert(certString)
if err != nil {
logrus.Errorf("Update cert unable to convert string to cert: %s", err)
s.tlsCert = nil
}
if tlsCert != nil {
s.tlsCert = tlsCert
for i, certBytes := range tlsCert.Certificate {
cert, err := x509.ParseCertificate(certBytes)
if err != nil {
logrus.Errorf("Update cert %d parse error: %s", i, err)
s.tlsCert = nil
break
}
ips := map[string]bool{}
for _, ip := range cert.IPAddresses {
ips[ip.String()] = true
}
domains := map[string]bool{}
for _, domain := range cert.DNSNames {
domains[domain] = true
}
if !(reflect.DeepEqual(ips, s.ips) && reflect.DeepEqual(domains, s.domains)) {
subset := true
for ip := range s.ips {
if !ips[ip] {
subset = false
break
}
}
if subset {
for domain := range s.domains {
if !domains[domain] {
subset = false
break
}
}
}
if !subset {
s.tlsCert = nil
}
for ip := range ips {
s.ips[ip] = true
}
for domain := range domains {
s.domains[domain] = true
}
}
}
}
return s.reload()
}
func (s *server) shutdown() error {
for _, listener := range s.listeners {
if err := listener.Close(); err != nil {
return err
}
}
s.listeners = nil
for _, server := range s.servers {
go server.Shutdown(context.Background())
}
s.servers = nil
return nil
}
func (s *server) reload() error {
if len(s.listeners) > 0 {
return nil
}
if err := s.shutdown(); err != nil |
if err := s.serveHTTPS(); err != nil {
return err
}
return nil
}
func (s *server) getCertificate(hello *tls.ClientHelloInfo) (_servingCert *tls.Certificate, _err error) {
s.Lock()
changed := false
defer func() {
defer s.Unlock()
if _err != nil {
logrus.Errorf("Get certificate error: %s", _err)
return
}
if changed {
s.save()
}
}()
if hello.ServerName != "" && !s.domains[hello.ServerName] {
s.tlsCert = nil
s.domains[hello.ServerName] = true
}
if s.tlsCert != nil {
return s.tlsCert, nil
}
ips := []net.IP{}
for ipStr := range s.ips {
if ip := net.ParseIP(ipStr); ip != nil {
ips = append(ips, ip)
}
}
dnsNames := []string{}
for domain := range s.domains {
dnsNames = append(dnsNames, domain)
}
cfg := cert.Config{
CommonName: s.cn,
Organization: s.activeCA.Subject.Organization,
Usages: []x509.ExtKeyUsage{x509.ExtKeyUsageServerAuth},
AltNames: cert.AltNames{
DNSNames: dnsNames,
IPs: ips,
},
}
key, err := newPrivateKey()
if err != nil {
return nil, err
}
cert, err := cert.NewSignedCert(cfg, key, s.activeCA, s.activeCAKey)
if err != nil {
return nil, err
}
tlsCert := &tls.Certificate{
Certificate: [][]byte{
cert.Raw,
},
PrivateKey: key,
}
changed = true
s.tlsCert = tlsCert
return tlsCert, nil
}
func (s *server) cacheHandler(handler http.Handler) http.Handler {
return http.HandlerFunc(func(resp http.ResponseWriter, req *http.Request) {
h, _, err := net.SplitHostPort(req.Host)
if err != nil {
h = req.Host
}
s.Lock()
if ip := net.ParseIP(h); ip != nil {
if !s.ips[h] {
s.ips[h] = true
s.tlsCert = nil
}
} else {
if !s.domains[h] {
s.domains[h] = true
s.tlsCert = nil
}
}
s.Unlock()
handler.ServeHTTP(resp, req)
})
}
func (s *server) serveHTTPS() error {
conf := &tls.Config{
ClientAuth: tls.RequestClientCert,
GetCertificate: s.getCertificate,
PreferServerCipherSuites: true,
}
listener, err := s.newListener(s.userConfig.BindAddress, s.userConfig.HTTPSPort, conf)
if err != nil {
return err
}
logger := logrus.StandardLogger()
server := &http.Server{
Handler: s.cacheHandler(s.Handler()),
ErrorLog: log.New(logger.WriterLevel(logrus.DebugLevel), "", log.LstdFlags),
}
s.servers = append(s.servers, server)
s.start | {
return err
} | conditional_block |
server.go | if err := s.userConfigure(); err != nil {
return nil, err
}
lc, err := listenConfigStorage.Get()
if err != nil {
return nil, err
}
return s, s.Update(lc)
}
func (s *server) CACert() (string, error) {
if s.userConfig.NoCACerts {
return "", nil
}
if s.userConfig.CACerts != "" {
return s.userConfig.CACerts, nil
}
return "", fmt.Errorf("ca cert not found")
}
func marshalPrivateKey(privateKey crypto.Signer) (string, []byte, error) {
var (
keyType string
bytes []byte
err error
)
if key, ok := privateKey.(*ecdsa.PrivateKey); ok {
keyType = cert.ECPrivateKeyBlockType
bytes, err = x509.MarshalECPrivateKey(key)
} else if key, ok := privateKey.(*rsa.PrivateKey); ok {
keyType = cert.RSAPrivateKeyBlockType
bytes = x509.MarshalPKCS1PrivateKey(key)
} else {
keyType = cert.PrivateKeyBlockType
bytes, err = x509.MarshalPKCS8PrivateKey(privateKey)
}
if err != nil {
logrus.Errorf("Unable to marshal private key: %v", err)
}
return keyType, bytes, err
}
func newPrivateKey() (crypto.Signer, error) {
caKeyBytes, err := cert.MakeEllipticPrivateKeyPEM()
if err != nil {
return nil, err
}
caKeyIFace, err := cert.ParsePrivateKeyPEM(caKeyBytes)
if err != nil {
return nil, err
}
return caKeyIFace.(crypto.Signer), nil
}
func (s *server) save() (_err error) {
defer func() {
if _err != nil {
logrus.Errorf("Saving cert error: %s", _err)
}
}()
certStr, err := certToString(s.tlsCert)
if err != nil {
return err
}
cfg, err := s.listenConfigStorage.Get()
if err != nil {
return err
}
cfg.GeneratedCerts = map[string]string{s.cn: certStr}
_, err = s.listenConfigStorage.Set(cfg)
return err
}
func (s *server) userConfigure() error {
if s.userConfig.HTTPSPort == 0 {
s.userConfig.HTTPSPort = 8443
}
for _, d := range s.userConfig.Domains {
s.domains[d] = true
}
for _, ip := range s.userConfig.KnownIPs {
if netIP := net.ParseIP(ip); netIP != nil {
s.ips[ip] = true
}
}
if bindAddress := net.ParseIP(s.userConfig.BindAddress); bindAddress != nil {
s.ips[s.userConfig.BindAddress] = true
}
if s.activeCA == nil && s.activeCAKey == nil {
if s.userConfig.CACerts != "" && s.userConfig.CAKey != "" {
ca, err := cert.ParseCertsPEM([]byte(s.userConfig.CACerts))
if err != nil {
return err
}
key, err := cert.ParsePrivateKeyPEM([]byte(s.userConfig.CAKey))
if err != nil {
return err
}
s.activeCA = ca[0]
s.activeCAKey = key.(crypto.Signer)
} else {
ca, key, err := genCA()
if err != nil {
return err
}
s.activeCA = ca
s.activeCAKey = key
}
}
return nil
}
func genCA() (*x509.Certificate, crypto.Signer, error) {
caKey, err := newPrivateKey()
if err != nil {
return nil, nil, err
}
caCert, err := cert.NewSelfSignedCACert(cert.Config{
CommonName: "k3s-ca",
Organization: []string{"k3s-org"},
}, caKey)
if err != nil {
return nil, nil, err
}
return caCert, caKey, nil
}
func (s *server) Update(status *ListenerStatus) (_err error) {
s.Lock()
defer func() {
s.Unlock()
if _err != nil {
logrus.Errorf("Update cert error: %s", _err)
}
if s.tlsCert == nil {
s.getCertificate(&tls.ClientHelloInfo{ServerName: "localhost"})
}
}()
certString := status.GeneratedCerts[s.cn]
tlsCert, err := stringToCert(certString)
if err != nil {
logrus.Errorf("Update cert unable to convert string to cert: %s", err)
s.tlsCert = nil
}
if tlsCert != nil {
s.tlsCert = tlsCert
for i, certBytes := range tlsCert.Certificate {
cert, err := x509.ParseCertificate(certBytes)
if err != nil {
logrus.Errorf("Update cert %d parse error: %s", i, err)
s.tlsCert = nil
break
}
ips := map[string]bool{}
for _, ip := range cert.IPAddresses {
ips[ip.String()] = true
}
domains := map[string]bool{}
for _, domain := range cert.DNSNames {
domains[domain] = true
}
if !(reflect.DeepEqual(ips, s.ips) && reflect.DeepEqual(domains, s.domains)) {
subset := true
for ip := range s.ips {
if !ips[ip] {
subset = false
break
}
}
if subset {
for domain := range s.domains {
if !domains[domain] {
subset = false
break
}
}
}
if !subset {
s.tlsCert = nil
}
for ip := range ips {
s.ips[ip] = true
}
for domain := range domains {
s.domains[domain] = true
}
}
}
}
return s.reload()
}
func (s *server) shutdown() error {
for _, listener := range s.listeners {
if err := listener.Close(); err != nil {
return err
}
}
s.listeners = nil
for _, server := range s.servers {
go server.Shutdown(context.Background())
}
s.servers = nil
return nil
}
func (s *server) reload() error {
if len(s.listeners) > 0 {
return nil
}
if err := s.shutdown(); err != nil {
return err
}
if err := s.serveHTTPS(); err != nil {
return err
}
return nil
}
func (s *server) getCertificate(hello *tls.ClientHelloInfo) (_servingCert *tls.Certificate, _err error) {
s.Lock()
changed := false
defer func() {
defer s.Unlock()
if _err != nil {
logrus.Errorf("Get certificate error: %s", _err)
return
}
if changed {
s.save()
}
}()
if hello.ServerName != "" && !s.domains[hello.ServerName] {
s.tlsCert = nil
s.domains[hello.ServerName] = true
}
if s.tlsCert != nil {
return s.tlsCert, nil
}
ips := []net.IP{}
for ipStr := range s.ips {
if ip := net.ParseIP(ipStr); ip != nil {
ips = append(ips, ip)
}
}
dnsNames := []string{}
for domain := range s.domains {
dnsNames = append(dnsNames, domain)
}
cfg := cert.Config{
CommonName: s.cn,
Organization: s.activeCA.Subject.Organization,
Usages: []x509.ExtKeyUsage{x509.ExtKeyUsageServerAuth},
AltNames: cert.AltNames{
DNSNames: dnsNames,
IPs: ips,
},
}
key, err := newPrivateKey()
if err != nil {
return nil, err
}
cert, err := cert.NewSignedCert(cfg, key, s.activeCA, s.activeCAKey)
if err != nil {
return nil, err
}
tlsCert := &tls.Certificate{
Certificate: [][]byte{
cert.Raw,
},
PrivateKey: key,
}
changed = true
s.tlsCert = tlsCert
return tlsCert, nil
}
func (s *server) cacheHandler(handler http.Handler) http.Handler {
return http.HandlerFunc(func(resp http.ResponseWriter, req *http.Request) {
h, _, err := net.SplitHostPort(req.Host)
if err != nil {
h = req.Host
}
s.Lock()
if ip := net.ParseIP(h); ip != nil {
if !s.ips[h] {
s.ips[h] = true
s.tlsCert = nil
}
} else {
if !s.domains[h] {
s.domains[h] = | s.ips = map[string]bool{}
s.domains = map[string]bool{}
| random_line_split |
|
server.go | , ok := privateKey.(*ecdsa.PrivateKey); ok {
keyType = cert.ECPrivateKeyBlockType
bytes, err = x509.MarshalECPrivateKey(key)
} else if key, ok := privateKey.(*rsa.PrivateKey); ok {
keyType = cert.RSAPrivateKeyBlockType
bytes = x509.MarshalPKCS1PrivateKey(key)
} else {
keyType = cert.PrivateKeyBlockType
bytes, err = x509.MarshalPKCS8PrivateKey(privateKey)
}
if err != nil {
logrus.Errorf("Unable to marshal private key: %v", err)
}
return keyType, bytes, err
}
func newPrivateKey() (crypto.Signer, error) {
caKeyBytes, err := cert.MakeEllipticPrivateKeyPEM()
if err != nil {
return nil, err
}
caKeyIFace, err := cert.ParsePrivateKeyPEM(caKeyBytes)
if err != nil {
return nil, err
}
return caKeyIFace.(crypto.Signer), nil
}
func (s *server) save() (_err error) {
defer func() {
if _err != nil {
logrus.Errorf("Saving cert error: %s", _err)
}
}()
certStr, err := certToString(s.tlsCert)
if err != nil {
return err
}
cfg, err := s.listenConfigStorage.Get()
if err != nil {
return err
}
cfg.GeneratedCerts = map[string]string{s.cn: certStr}
_, err = s.listenConfigStorage.Set(cfg)
return err
}
func (s *server) | () error {
if s.userConfig.HTTPSPort == 0 {
s.userConfig.HTTPSPort = 8443
}
for _, d := range s.userConfig.Domains {
s.domains[d] = true
}
for _, ip := range s.userConfig.KnownIPs {
if netIP := net.ParseIP(ip); netIP != nil {
s.ips[ip] = true
}
}
if bindAddress := net.ParseIP(s.userConfig.BindAddress); bindAddress != nil {
s.ips[s.userConfig.BindAddress] = true
}
if s.activeCA == nil && s.activeCAKey == nil {
if s.userConfig.CACerts != "" && s.userConfig.CAKey != "" {
ca, err := cert.ParseCertsPEM([]byte(s.userConfig.CACerts))
if err != nil {
return err
}
key, err := cert.ParsePrivateKeyPEM([]byte(s.userConfig.CAKey))
if err != nil {
return err
}
s.activeCA = ca[0]
s.activeCAKey = key.(crypto.Signer)
} else {
ca, key, err := genCA()
if err != nil {
return err
}
s.activeCA = ca
s.activeCAKey = key
}
}
return nil
}
func genCA() (*x509.Certificate, crypto.Signer, error) {
caKey, err := newPrivateKey()
if err != nil {
return nil, nil, err
}
caCert, err := cert.NewSelfSignedCACert(cert.Config{
CommonName: "k3s-ca",
Organization: []string{"k3s-org"},
}, caKey)
if err != nil {
return nil, nil, err
}
return caCert, caKey, nil
}
func (s *server) Update(status *ListenerStatus) (_err error) {
s.Lock()
defer func() {
s.Unlock()
if _err != nil {
logrus.Errorf("Update cert error: %s", _err)
}
if s.tlsCert == nil {
s.getCertificate(&tls.ClientHelloInfo{ServerName: "localhost"})
}
}()
certString := status.GeneratedCerts[s.cn]
tlsCert, err := stringToCert(certString)
if err != nil {
logrus.Errorf("Update cert unable to convert string to cert: %s", err)
s.tlsCert = nil
}
if tlsCert != nil {
s.tlsCert = tlsCert
for i, certBytes := range tlsCert.Certificate {
cert, err := x509.ParseCertificate(certBytes)
if err != nil {
logrus.Errorf("Update cert %d parse error: %s", i, err)
s.tlsCert = nil
break
}
ips := map[string]bool{}
for _, ip := range cert.IPAddresses {
ips[ip.String()] = true
}
domains := map[string]bool{}
for _, domain := range cert.DNSNames {
domains[domain] = true
}
if !(reflect.DeepEqual(ips, s.ips) && reflect.DeepEqual(domains, s.domains)) {
subset := true
for ip := range s.ips {
if !ips[ip] {
subset = false
break
}
}
if subset {
for domain := range s.domains {
if !domains[domain] {
subset = false
break
}
}
}
if !subset {
s.tlsCert = nil
}
for ip := range ips {
s.ips[ip] = true
}
for domain := range domains {
s.domains[domain] = true
}
}
}
}
return s.reload()
}
func (s *server) shutdown() error {
for _, listener := range s.listeners {
if err := listener.Close(); err != nil {
return err
}
}
s.listeners = nil
for _, server := range s.servers {
go server.Shutdown(context.Background())
}
s.servers = nil
return nil
}
func (s *server) reload() error {
if len(s.listeners) > 0 {
return nil
}
if err := s.shutdown(); err != nil {
return err
}
if err := s.serveHTTPS(); err != nil {
return err
}
return nil
}
func (s *server) getCertificate(hello *tls.ClientHelloInfo) (_servingCert *tls.Certificate, _err error) {
s.Lock()
changed := false
defer func() {
defer s.Unlock()
if _err != nil {
logrus.Errorf("Get certificate error: %s", _err)
return
}
if changed {
s.save()
}
}()
if hello.ServerName != "" && !s.domains[hello.ServerName] {
s.tlsCert = nil
s.domains[hello.ServerName] = true
}
if s.tlsCert != nil {
return s.tlsCert, nil
}
ips := []net.IP{}
for ipStr := range s.ips {
if ip := net.ParseIP(ipStr); ip != nil {
ips = append(ips, ip)
}
}
dnsNames := []string{}
for domain := range s.domains {
dnsNames = append(dnsNames, domain)
}
cfg := cert.Config{
CommonName: s.cn,
Organization: s.activeCA.Subject.Organization,
Usages: []x509.ExtKeyUsage{x509.ExtKeyUsageServerAuth},
AltNames: cert.AltNames{
DNSNames: dnsNames,
IPs: ips,
},
}
key, err := newPrivateKey()
if err != nil {
return nil, err
}
cert, err := cert.NewSignedCert(cfg, key, s.activeCA, s.activeCAKey)
if err != nil {
return nil, err
}
tlsCert := &tls.Certificate{
Certificate: [][]byte{
cert.Raw,
},
PrivateKey: key,
}
changed = true
s.tlsCert = tlsCert
return tlsCert, nil
}
func (s *server) cacheHandler(handler http.Handler) http.Handler {
return http.HandlerFunc(func(resp http.ResponseWriter, req *http.Request) {
h, _, err := net.SplitHostPort(req.Host)
if err != nil {
h = req.Host
}
s.Lock()
if ip := net.ParseIP(h); ip != nil {
if !s.ips[h] {
s.ips[h] = true
s.tlsCert = nil
}
} else {
if !s.domains[h] {
s.domains[h] = true
s.tlsCert = nil
}
}
s.Unlock()
handler.ServeHTTP(resp, req)
})
}
func (s *server) serveHTTPS() error {
conf := &tls.Config{
ClientAuth: tls.RequestClientCert,
GetCertificate: s.getCertificate,
PreferServerCipherSuites: true,
}
listener, err := s.newListener(s.userConfig.BindAddress, s.userConfig.HTTPSPort, conf)
if err != nil {
return err
}
logger := logrus.StandardLogger()
server := &http.Server{
Handler: s.cacheHandler(s.Handler()),
ErrorLog: log.New(logger.WriterLevel(logrus.DebugLevel), "", log.LstdFlags),
}
s.servers = append(s.servers, server)
s.start | userConfigure | identifier_name |
server.go | ok := privateKey.(*ecdsa.PrivateKey); ok {
keyType = cert.ECPrivateKeyBlockType
bytes, err = x509.MarshalECPrivateKey(key)
} else if key, ok := privateKey.(*rsa.PrivateKey); ok {
keyType = cert.RSAPrivateKeyBlockType
bytes = x509.MarshalPKCS1PrivateKey(key)
} else {
keyType = cert.PrivateKeyBlockType
bytes, err = x509.MarshalPKCS8PrivateKey(privateKey)
}
if err != nil {
logrus.Errorf("Unable to marshal private key: %v", err)
}
return keyType, bytes, err
}
func newPrivateKey() (crypto.Signer, error) {
caKeyBytes, err := cert.MakeEllipticPrivateKeyPEM()
if err != nil {
return nil, err
}
caKeyIFace, err := cert.ParsePrivateKeyPEM(caKeyBytes)
if err != nil {
return nil, err
}
return caKeyIFace.(crypto.Signer), nil
}
func (s *server) save() (_err error) {
defer func() {
if _err != nil {
logrus.Errorf("Saving cert error: %s", _err)
}
}()
certStr, err := certToString(s.tlsCert)
if err != nil {
return err
}
cfg, err := s.listenConfigStorage.Get()
if err != nil {
return err
}
cfg.GeneratedCerts = map[string]string{s.cn: certStr}
_, err = s.listenConfigStorage.Set(cfg)
return err
}
func (s *server) userConfigure() error {
if s.userConfig.HTTPSPort == 0 {
s.userConfig.HTTPSPort = 8443
}
for _, d := range s.userConfig.Domains {
s.domains[d] = true
}
for _, ip := range s.userConfig.KnownIPs {
if netIP := net.ParseIP(ip); netIP != nil {
s.ips[ip] = true
}
}
if bindAddress := net.ParseIP(s.userConfig.BindAddress); bindAddress != nil {
s.ips[s.userConfig.BindAddress] = true
}
if s.activeCA == nil && s.activeCAKey == nil {
if s.userConfig.CACerts != "" && s.userConfig.CAKey != "" {
ca, err := cert.ParseCertsPEM([]byte(s.userConfig.CACerts))
if err != nil {
return err
}
key, err := cert.ParsePrivateKeyPEM([]byte(s.userConfig.CAKey))
if err != nil {
return err
}
s.activeCA = ca[0]
s.activeCAKey = key.(crypto.Signer)
} else {
ca, key, err := genCA()
if err != nil {
return err
}
s.activeCA = ca
s.activeCAKey = key
}
}
return nil
}
func genCA() (*x509.Certificate, crypto.Signer, error) {
caKey, err := newPrivateKey()
if err != nil {
return nil, nil, err
}
caCert, err := cert.NewSelfSignedCACert(cert.Config{
CommonName: "k3s-ca",
Organization: []string{"k3s-org"},
}, caKey)
if err != nil {
return nil, nil, err
}
return caCert, caKey, nil
}
func (s *server) Update(status *ListenerStatus) (_err error) {
s.Lock()
defer func() {
s.Unlock()
if _err != nil {
logrus.Errorf("Update cert error: %s", _err)
}
if s.tlsCert == nil {
s.getCertificate(&tls.ClientHelloInfo{ServerName: "localhost"})
}
}()
certString := status.GeneratedCerts[s.cn]
tlsCert, err := stringToCert(certString)
if err != nil {
logrus.Errorf("Update cert unable to convert string to cert: %s", err)
s.tlsCert = nil
}
if tlsCert != nil {
s.tlsCert = tlsCert
for i, certBytes := range tlsCert.Certificate {
cert, err := x509.ParseCertificate(certBytes)
if err != nil {
logrus.Errorf("Update cert %d parse error: %s", i, err)
s.tlsCert = nil
break
}
ips := map[string]bool{}
for _, ip := range cert.IPAddresses {
ips[ip.String()] = true
}
domains := map[string]bool{}
for _, domain := range cert.DNSNames {
domains[domain] = true
}
if !(reflect.DeepEqual(ips, s.ips) && reflect.DeepEqual(domains, s.domains)) {
subset := true
for ip := range s.ips {
if !ips[ip] {
subset = false
break
}
}
if subset {
for domain := range s.domains {
if !domains[domain] {
subset = false
break
}
}
}
if !subset {
s.tlsCert = nil
}
for ip := range ips {
s.ips[ip] = true
}
for domain := range domains {
s.domains[domain] = true
}
}
}
}
return s.reload()
}
func (s *server) shutdown() error {
for _, listener := range s.listeners {
if err := listener.Close(); err != nil {
return err
}
}
s.listeners = nil
for _, server := range s.servers {
go server.Shutdown(context.Background())
}
s.servers = nil
return nil
}
func (s *server) reload() error |
func (s *server) getCertificate(hello *tls.ClientHelloInfo) (_servingCert *tls.Certificate, _err error) {
s.Lock()
changed := false
defer func() {
defer s.Unlock()
if _err != nil {
logrus.Errorf("Get certificate error: %s", _err)
return
}
if changed {
s.save()
}
}()
if hello.ServerName != "" && !s.domains[hello.ServerName] {
s.tlsCert = nil
s.domains[hello.ServerName] = true
}
if s.tlsCert != nil {
return s.tlsCert, nil
}
ips := []net.IP{}
for ipStr := range s.ips {
if ip := net.ParseIP(ipStr); ip != nil {
ips = append(ips, ip)
}
}
dnsNames := []string{}
for domain := range s.domains {
dnsNames = append(dnsNames, domain)
}
cfg := cert.Config{
CommonName: s.cn,
Organization: s.activeCA.Subject.Organization,
Usages: []x509.ExtKeyUsage{x509.ExtKeyUsageServerAuth},
AltNames: cert.AltNames{
DNSNames: dnsNames,
IPs: ips,
},
}
key, err := newPrivateKey()
if err != nil {
return nil, err
}
cert, err := cert.NewSignedCert(cfg, key, s.activeCA, s.activeCAKey)
if err != nil {
return nil, err
}
tlsCert := &tls.Certificate{
Certificate: [][]byte{
cert.Raw,
},
PrivateKey: key,
}
changed = true
s.tlsCert = tlsCert
return tlsCert, nil
}
func (s *server) cacheHandler(handler http.Handler) http.Handler {
return http.HandlerFunc(func(resp http.ResponseWriter, req *http.Request) {
h, _, err := net.SplitHostPort(req.Host)
if err != nil {
h = req.Host
}
s.Lock()
if ip := net.ParseIP(h); ip != nil {
if !s.ips[h] {
s.ips[h] = true
s.tlsCert = nil
}
} else {
if !s.domains[h] {
s.domains[h] = true
s.tlsCert = nil
}
}
s.Unlock()
handler.ServeHTTP(resp, req)
})
}
func (s *server) serveHTTPS() error {
conf := &tls.Config{
ClientAuth: tls.RequestClientCert,
GetCertificate: s.getCertificate,
PreferServerCipherSuites: true,
}
listener, err := s.newListener(s.userConfig.BindAddress, s.userConfig.HTTPSPort, conf)
if err != nil {
return err
}
logger := logrus.StandardLogger()
server := &http.Server{
Handler: s.cacheHandler(s.Handler()),
ErrorLog: log.New(logger.WriterLevel(logrus.DebugLevel), "", log.LstdFlags),
}
s.servers = append(s.servers, server)
s.start | {
if len(s.listeners) > 0 {
return nil
}
if err := s.shutdown(); err != nil {
return err
}
if err := s.serveHTTPS(); err != nil {
return err
}
return nil
} | identifier_body |
cli.rs | mut remaining_v: libc::tcflag_t = v;
let mut s = String::new();
for &(fname, fvalue) in fs {
if (v & fvalue) != 0 {
let was_empty = s.is_empty();
write!(
&mut s,
"{}{}",
match was_empty {
true => "",
false => " | ",
},
fname
)?;
remaining_v &= !v;
}
}
if remaining_v != 0 {
let was_empty = s.is_empty();
write!(
&mut s,
"{}(extra: {:x})",
match was_empty {
true => "",
false => " ",
},
remaining_v
)?;
}
Ok(s)
}
fn debug_format_c_cc_field(c_cc: &[libc::cc_t; 32]) -> std::result::Result<String, fmt::Error> {
use fmt::Write;
const INDICES: &'static [(&'static str, usize)] = &[
("VDISCARD", libc::VDISCARD),
("VEOF", libc::VEOF),
("VEOL", libc::VEOL),
("VEOL2", libc::VEOL2),
("VERASE", libc::VERASE),
("VINTR", libc::VINTR),
("VKILL", libc::VKILL),
("VLNEXT", libc::VLNEXT),
("VMIN", libc::VMIN),
("VQUIT", libc::VQUIT),
("VREPRINT", libc::VREPRINT),
("VSTART", libc::VSTART),
("VSTOP", libc::VSTOP),
("VSUSP", libc::VSUSP),
("VSWTC", libc::VSWTC),
("VTIME", libc::VTIME),
("VWERASE", libc::VWERASE),
];
let mut s = String::new();
for &(name, idx) in INDICES {
let was_empty = s.is_empty();
write!(
&mut s,
"{}{}:{}",
match was_empty {
true => "",
false => ", ",
},
name,
c_cc[idx]
)?;
}
Ok(s)
}
impl fmt::Debug for TerminalAttributes {
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
f.debug_struct("TerminalAttributes")
.field(
"c_iflag",
&debug_format_flag_field(
self.inner.c_iflag,
&[
("IGNBRK", libc::IGNBRK),
("BRKINT", libc::BRKINT),
("IGNPAR", libc::IGNPAR),
("PARMRK", libc::PARMRK),
("INPCK", libc::INPCK),
("ISTRIP", libc::ISTRIP),
("INLCR", libc::INLCR),
("IGNCR", libc::IGNCR),
("ICRNL", libc::ICRNL),
("IXON", libc::IXON),
("IXANY", libc::IXANY),
("IXOFF", libc::IXOFF),
("IMAXBEL", libc::IMAXBEL),
("IUTF8", libc::IUTF8),
],
)?,
)
.field(
"c_oflag",
&debug_format_flag_field(
self.inner.c_oflag,
&[
("OPOST", libc::OPOST),
("OLCUC", libc::OLCUC),
("ONLCR", libc::ONLCR),
("ONOCR", libc::ONOCR),
("ONLRET", libc::ONLRET),
("OFILL", libc::OFILL),
("OFDEL", libc::OFDEL),
("NLDLY", libc::NLDLY),
("CRDLY", libc::CRDLY),
("TABDLY", libc::TABDLY),
("BSDLY", libc::BSDLY),
("VTDLY", libc::VTDLY),
("FFDLY", libc::FFDLY),
],
)?,
)
.field(
"c_cflag",
&debug_format_flag_field(
self.inner.c_cflag,
&[
("CBAUD", libc::CBAUD),
("CBAUDEX", libc::CBAUDEX),
("CSIZE", libc::CSIZE),
("CSTOPB", libc::CSTOPB),
("CREAD", libc::CREAD),
("PARENB", libc::PARENB),
("PARODD", libc::PARODD),
("HUPCL", libc::HUPCL),
("CLOCAL", libc::CLOCAL),
("CIBAUD", libc::CIBAUD),
("CMSPAR", libc::CMSPAR),
("CRTSCTS", libc::CRTSCTS),
],
)?,
)
.field(
"c_lflag",
&debug_format_flag_field(
self.inner.c_lflag,
&[
("ISIG", libc::ISIG),
("ICANON", libc::ICANON),
("ECHO", libc::ECHO),
("ECHOE", libc::ECHOE),
("ECHOK", libc::ECHOK),
("ECHONL", libc::ECHONL),
("ECHOCTL", libc::ECHOCTL),
("ECHOPRT", libc::ECHOPRT),
("ECHOKE", libc::ECHOKE),
("FLUSHO", libc::FLUSHO),
("NOFLSH", libc::NOFLSH),
("TOSTOP", libc::TOSTOP),
("PENDIN", libc::PENDIN),
("IEXTEN", libc::IEXTEN),
],
)?,
)
.field("c_cc", &debug_format_c_cc_field(&self.inner.c_cc)?)
.field("c_ispeed", unsafe { &libc::cfgetispeed(&self.inner) })
.field("c_ospeed", unsafe { &libc::cfgetospeed(&self.inner) })
.finish()
}
}
impl AbstractTerminalAttributes for TerminalAttributes {
fn enable(&mut self, flag: TerminalFlag) {
self.inner.c_lflag |= flag.to_value();
}
fn disable(&mut self, flag: TerminalFlag) {
self.inner.c_lflag &= !flag.to_value();
}
}
/// This trait describes an abstract input or output stream.
///
/// This trait primarily exists for testing purposes. In almost all cases, users
/// will instead just use the concrete type `Stream` defined below.
pub trait AbstractStream {
/// A type which describes the attributes of this stream / terminal.
type Attributes: AbstractTerminalAttributes + fmt::Debug;
/// Returns whether or not this stream refers to an interactive terminal (a
/// TTY), as opposed to, for example, a pipe.
fn isatty(&self) -> bool;
/// Retrieve the current attributes of this stream / terminal.
fn get_attributes(&self) -> IoResult<Self::Attributes>;
/// Modify this stream's / terminal's attributes to match the given state.
fn set_attributes(&mut self, attributes: &Self::Attributes) -> IoResult<()>;
/// Return a `Read` for this stream, if reading is supported.
fn as_reader(&self) -> Option<Box<dyn Read>>;
/// Return a `Write` for this stream, if writing is supported.
fn as_writer(&self) -> Option<Box<dyn Write>>;
}
/// Standard input / output streams.
#[derive(Debug)]
pub enum Stream {
/// Standard output.
Stdout,
/// Standard error.
Stderr,
/// Standard input.
Stdin,
}
impl Stream {
fn to_fd(&self) -> c_int |
}
impl AbstractStream for Stream {
type Attributes = TerminalAttributes;
fn isatty(&self) -> bool {
let ret = unsafe { libc::isatty(self.to_fd()) };
let error: i32 = errno::errno().into();
match ret {
1 => true,
0 => match error {
libc::EBADF => false,
libc::ENOTTY => false,
_ => {
debug!(
"Unrecognized isatty errno: {}; assuming {:?} is not a TTY",
error, *self
);
false
}
},
_ => {
debug!(
"Unrecognized isatty return code: {}; assuming {:?} is not a TTY",
ret, *self
);
false
}
}
}
fn get_attributes(&self) -> IoResult<Self::Attributes> {
TerminalAttributes::new(self.to_fd())
}
fn set_attributes(&mut self, attributes: &Self::Attributes) -> IoResult<()> {
let ret = attributes.apply(self.to_fd());
debug_assert!(ret.is_err() || *attributes == Self::Attributes::new(self.to_fd()).unwrap());
ret
}
fn as_reader(&self) -> Option<Box<dyn Read>> {
match *self {
Stream::Stdin => Some(Box::new(io::stdin())),
_ => None,
}
}
fn as_writer(&self) -> Option<Box<dyn Write>> {
match *self {
Stream::Std | {
match *self {
Stream::Stdout => libc::STDOUT_FILENO,
Stream::Stderr => libc::STDERR_FILENO,
Stream::Stdin => libc::STDIN_FILENO,
}
} | identifier_body |
cli.rs | SPAR", libc::CMSPAR),
("CRTSCTS", libc::CRTSCTS),
],
)?,
)
.field(
"c_lflag",
&debug_format_flag_field(
self.inner.c_lflag,
&[
("ISIG", libc::ISIG),
("ICANON", libc::ICANON),
("ECHO", libc::ECHO),
("ECHOE", libc::ECHOE),
("ECHOK", libc::ECHOK),
("ECHONL", libc::ECHONL),
("ECHOCTL", libc::ECHOCTL),
("ECHOPRT", libc::ECHOPRT),
("ECHOKE", libc::ECHOKE),
("FLUSHO", libc::FLUSHO),
("NOFLSH", libc::NOFLSH),
("TOSTOP", libc::TOSTOP),
("PENDIN", libc::PENDIN),
("IEXTEN", libc::IEXTEN),
],
)?,
)
.field("c_cc", &debug_format_c_cc_field(&self.inner.c_cc)?)
.field("c_ispeed", unsafe { &libc::cfgetispeed(&self.inner) })
.field("c_ospeed", unsafe { &libc::cfgetospeed(&self.inner) })
.finish()
}
}
impl AbstractTerminalAttributes for TerminalAttributes {
fn enable(&mut self, flag: TerminalFlag) {
self.inner.c_lflag |= flag.to_value();
}
fn disable(&mut self, flag: TerminalFlag) {
self.inner.c_lflag &= !flag.to_value();
}
}
/// This trait describes an abstract input or output stream.
///
/// This trait primarily exists for testing purposes. In almost all cases, users
/// will instead just use the concrete type `Stream` defined below.
pub trait AbstractStream {
/// A type which describes the attributes of this stream / terminal.
type Attributes: AbstractTerminalAttributes + fmt::Debug;
/// Returns whether or not this stream refers to an interactive terminal (a
/// TTY), as opposed to, for example, a pipe.
fn isatty(&self) -> bool;
/// Retrieve the current attributes of this stream / terminal.
fn get_attributes(&self) -> IoResult<Self::Attributes>;
/// Modify this stream's / terminal's attributes to match the given state.
fn set_attributes(&mut self, attributes: &Self::Attributes) -> IoResult<()>;
/// Return a `Read` for this stream, if reading is supported.
fn as_reader(&self) -> Option<Box<dyn Read>>;
/// Return a `Write` for this stream, if writing is supported.
fn as_writer(&self) -> Option<Box<dyn Write>>;
}
/// Standard input / output streams.
#[derive(Debug)]
pub enum Stream {
/// Standard output.
Stdout,
/// Standard error.
Stderr,
/// Standard input.
Stdin,
}
impl Stream {
fn to_fd(&self) -> c_int {
match *self {
Stream::Stdout => libc::STDOUT_FILENO,
Stream::Stderr => libc::STDERR_FILENO,
Stream::Stdin => libc::STDIN_FILENO,
}
}
}
impl AbstractStream for Stream {
type Attributes = TerminalAttributes;
fn isatty(&self) -> bool {
let ret = unsafe { libc::isatty(self.to_fd()) };
let error: i32 = errno::errno().into();
match ret {
1 => true,
0 => match error {
libc::EBADF => false,
libc::ENOTTY => false,
_ => {
debug!(
"Unrecognized isatty errno: {}; assuming {:?} is not a TTY",
error, *self
);
false
}
},
_ => {
debug!(
"Unrecognized isatty return code: {}; assuming {:?} is not a TTY",
ret, *self
);
false
}
}
}
fn get_attributes(&self) -> IoResult<Self::Attributes> {
TerminalAttributes::new(self.to_fd())
}
fn set_attributes(&mut self, attributes: &Self::Attributes) -> IoResult<()> {
let ret = attributes.apply(self.to_fd());
debug_assert!(ret.is_err() || *attributes == Self::Attributes::new(self.to_fd()).unwrap());
ret
}
fn as_reader(&self) -> Option<Box<dyn Read>> {
match *self {
Stream::Stdin => Some(Box::new(io::stdin())),
_ => None,
}
}
fn as_writer(&self) -> Option<Box<dyn Write>> {
match *self {
Stream::Stdout => Some(Box::new(io::stdout())),
Stream::Stderr => Some(Box::new(io::stderr())),
_ => None,
}
}
}
/// This structure handles a) disabling the echoing of characters typed to
/// `Stdin`, and b) remembering to reset the terminal attributes afterwards
/// (via `Drop`).
struct DisableEcho<'s, S: AbstractStream> {
stream: &'s mut S,
initial_attributes: S::Attributes,
}
impl<'s, S: AbstractStream> DisableEcho<'s, S> {
fn new(stream: &'s mut S) -> Result<Self> {
let initial_attributes = stream.get_attributes()?;
debug!("Initial stream attributes: {:#?}", initial_attributes);
let mut attributes = stream.get_attributes()?;
// Don't echo characters typed to stdin.
attributes.disable(TerminalFlag::Echo);
// But, *do* echo the newline when the user hits ENTER.
attributes.enable(TerminalFlag::EchoNewlines);
debug!("Setting attributes to: {:#?}", attributes);
stream.set_attributes(&attributes)?;
Ok(DisableEcho {
stream: stream,
initial_attributes: initial_attributes,
})
}
}
impl<'s, S: AbstractStream> Drop for DisableEcho<'s, S> {
fn drop(&mut self) {
self.stream
.set_attributes(&self.initial_attributes)
.unwrap();
}
}
fn require_isatty<S: AbstractStream>(s: &mut S) -> Result<()> {
if !s.isatty() {
Err(Error::Precondition(format!(
"cannot prompt interactively when the I/O streams are not TTYs"
)))
} else {
Ok(())
}
}
fn build_input_reader<IS: AbstractStream>(
input_stream: &mut IS,
) -> Result<io::BufReader<Box<dyn Read>>> {
require_isatty(input_stream)?;
Ok(io::BufReader::new(match input_stream.as_reader() {
None => {
return Err(Error::Precondition(format!(
"the given input stream must support `Read`"
)))
}
Some(r) => r,
}))
}
fn remove_newline(mut s: String) -> Result<String> {
// Remove the trailing newline (if any - not finding one is an error).
if !s.ends_with('\n') {
return Err(io::Error::new(io::ErrorKind::UnexpectedEof, "unexpected end of input").into());
}
s.pop();
// If this is windows and so there's also a \r, remove that too.
if s.ends_with('\r') {
s.pop();
}
Ok(s)
}
fn prompt_for_string_impl<IS: AbstractStream, OS: AbstractStream>(
input_stream: &mut IS,
// We have to take the reader as a parameter, since it must be "global",
// even if this function is e.g. called in a loop. Otherwise, because it's
// buffered, we might buffer some input and then discard it.
input_reader: &mut io::BufReader<Box<dyn Read>>,
output_stream: &mut OS,
prompt: &str,
is_sensitive: bool,
) -> Result<String> {
use io::BufRead;
require_isatty(output_stream)?;
// It's fine to construct a separate writer, potentially on each loop
// iteration or whatever, because we flush immediately, and don't do any
// buffering.
let mut writer = match output_stream.as_writer() {
None => {
return Err(Error::Precondition(format!(
"the given output stream must support `Write`"
)))
}
Some(w) => w,
};
write!(writer, "{}", prompt)?;
// We have to flush so the user sees the prompt immediately.
writer.flush()?;
Ok({
let _disable_echo = match is_sensitive {
false => None,
true => Some(DisableEcho::new(input_stream)?),
};
let mut ret = String::new();
input_reader.read_line(&mut ret)?;
remove_newline(ret)?
})
}
/// Prompt the user for a string (read from the given input stream) using the
/// given output stream (typically standard output or standard error) to display
/// the given prompt message.
///
/// If `is_sensitive` is true, then the users characters will not be echoed back
/// (e.g. this will behave like a password prompt).
///
/// Note that there are various requirements for the given streams, and this
/// function will return an error if any of them are not met:
///
/// - Both `input_stream` and `output_stream` must be TTYs.
/// - `input_stream` must return a valid `Read` instance.
/// - `output_stream` must return a valid `Write` instance.
pub fn | prompt_for_string | identifier_name |
|
cli.rs | && self.inner.c_oflag == other.inner.c_oflag
&& self.inner.c_cflag == other.inner.c_cflag
&& self.inner.c_lflag == other.inner.c_lflag
&& self.inner.c_line == other.inner.c_line
&& self.inner.c_cc == other.inner.c_cc
&& self.inner.c_ispeed == other.inner.c_ispeed
&& self.inner.c_ospeed == other.inner.c_ospeed
}
}
impl Eq for TerminalAttributes {}
fn debug_format_flag_field(
v: libc::tcflag_t,
fs: &'static [(&'static str, libc::tcflag_t)],
) -> std::result::Result<String, fmt::Error> {
use fmt::Write;
let mut remaining_v: libc::tcflag_t = v;
let mut s = String::new();
for &(fname, fvalue) in fs {
if (v & fvalue) != 0 {
let was_empty = s.is_empty();
write!(
&mut s,
"{}{}",
match was_empty {
true => "",
false => " | ",
},
fname
)?;
remaining_v &= !v;
}
}
if remaining_v != 0 {
let was_empty = s.is_empty();
write!(
&mut s,
"{}(extra: {:x})",
match was_empty {
true => "",
false => " ",
},
remaining_v
)?;
}
Ok(s)
}
fn debug_format_c_cc_field(c_cc: &[libc::cc_t; 32]) -> std::result::Result<String, fmt::Error> {
use fmt::Write;
const INDICES: &'static [(&'static str, usize)] = &[
("VDISCARD", libc::VDISCARD),
("VEOF", libc::VEOF),
("VEOL", libc::VEOL),
("VEOL2", libc::VEOL2),
("VERASE", libc::VERASE),
("VINTR", libc::VINTR),
("VKILL", libc::VKILL),
("VLNEXT", libc::VLNEXT),
("VMIN", libc::VMIN),
("VQUIT", libc::VQUIT),
("VREPRINT", libc::VREPRINT),
("VSTART", libc::VSTART),
("VSTOP", libc::VSTOP),
("VSUSP", libc::VSUSP),
("VSWTC", libc::VSWTC),
("VTIME", libc::VTIME),
("VWERASE", libc::VWERASE),
];
let mut s = String::new();
for &(name, idx) in INDICES {
let was_empty = s.is_empty();
write!(
&mut s,
"{}{}:{}",
match was_empty {
true => "",
false => ", ",
},
name,
c_cc[idx]
)?;
}
Ok(s)
}
impl fmt::Debug for TerminalAttributes {
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
f.debug_struct("TerminalAttributes")
.field(
"c_iflag",
&debug_format_flag_field(
self.inner.c_iflag,
&[
("IGNBRK", libc::IGNBRK),
("BRKINT", libc::BRKINT),
("IGNPAR", libc::IGNPAR),
("PARMRK", libc::PARMRK),
("INPCK", libc::INPCK),
("ISTRIP", libc::ISTRIP),
("INLCR", libc::INLCR),
("IGNCR", libc::IGNCR),
("ICRNL", libc::ICRNL),
("IXON", libc::IXON),
("IXANY", libc::IXANY),
("IXOFF", libc::IXOFF),
("IMAXBEL", libc::IMAXBEL),
("IUTF8", libc::IUTF8),
],
)?,
)
.field(
"c_oflag",
&debug_format_flag_field(
self.inner.c_oflag,
&[
("OPOST", libc::OPOST),
("OLCUC", libc::OLCUC),
("ONLCR", libc::ONLCR),
("ONOCR", libc::ONOCR),
("ONLRET", libc::ONLRET),
("OFILL", libc::OFILL),
("OFDEL", libc::OFDEL),
("NLDLY", libc::NLDLY),
("CRDLY", libc::CRDLY),
("TABDLY", libc::TABDLY),
("BSDLY", libc::BSDLY),
("VTDLY", libc::VTDLY),
("FFDLY", libc::FFDLY),
],
)?,
)
.field(
"c_cflag",
&debug_format_flag_field(
self.inner.c_cflag,
&[
("CBAUD", libc::CBAUD),
("CBAUDEX", libc::CBAUDEX),
("CSIZE", libc::CSIZE),
("CSTOPB", libc::CSTOPB),
("CREAD", libc::CREAD),
("PARENB", libc::PARENB),
("PARODD", libc::PARODD),
("HUPCL", libc::HUPCL),
("CLOCAL", libc::CLOCAL),
("CIBAUD", libc::CIBAUD),
("CMSPAR", libc::CMSPAR),
("CRTSCTS", libc::CRTSCTS),
],
)?,
)
.field(
"c_lflag",
&debug_format_flag_field(
self.inner.c_lflag,
&[
("ISIG", libc::ISIG),
("ICANON", libc::ICANON),
("ECHO", libc::ECHO),
("ECHOE", libc::ECHOE),
("ECHOK", libc::ECHOK),
("ECHONL", libc::ECHONL),
("ECHOCTL", libc::ECHOCTL),
("ECHOPRT", libc::ECHOPRT),
("ECHOKE", libc::ECHOKE),
("FLUSHO", libc::FLUSHO),
("NOFLSH", libc::NOFLSH),
("TOSTOP", libc::TOSTOP),
("PENDIN", libc::PENDIN),
("IEXTEN", libc::IEXTEN),
],
)?,
)
.field("c_cc", &debug_format_c_cc_field(&self.inner.c_cc)?)
.field("c_ispeed", unsafe { &libc::cfgetispeed(&self.inner) })
.field("c_ospeed", unsafe { &libc::cfgetospeed(&self.inner) })
.finish()
}
}
impl AbstractTerminalAttributes for TerminalAttributes {
fn enable(&mut self, flag: TerminalFlag) {
self.inner.c_lflag |= flag.to_value();
}
fn disable(&mut self, flag: TerminalFlag) {
self.inner.c_lflag &= !flag.to_value();
}
}
/// This trait describes an abstract input or output stream.
///
/// This trait primarily exists for testing purposes. In almost all cases, users
/// will instead just use the concrete type `Stream` defined below.
pub trait AbstractStream {
/// A type which describes the attributes of this stream / terminal.
type Attributes: AbstractTerminalAttributes + fmt::Debug;
/// Returns whether or not this stream refers to an interactive terminal (a
/// TTY), as opposed to, for example, a pipe.
fn isatty(&self) -> bool;
/// Retrieve the current attributes of this stream / terminal.
fn get_attributes(&self) -> IoResult<Self::Attributes>;
/// Modify this stream's / terminal's attributes to match the given state.
fn set_attributes(&mut self, attributes: &Self::Attributes) -> IoResult<()>;
/// Return a `Read` for this stream, if reading is supported.
fn as_reader(&self) -> Option<Box<dyn Read>>;
/// Return a `Write` for this stream, if writing is supported.
fn as_writer(&self) -> Option<Box<dyn Write>>;
}
/// Standard input / output streams.
#[derive(Debug)]
pub enum Stream {
/// Standard output.
Stdout,
/// Standard error.
Stderr,
/// Standard input.
Stdin,
}
impl Stream {
fn to_fd(&self) -> c_int {
match *self {
Stream::Stdout => libc::STDOUT_FILENO,
Stream::Stderr => libc::STDERR_FILENO,
Stream::Stdin => libc::STDIN_FILENO,
}
}
}
impl AbstractStream for Stream {
type Attributes = TerminalAttributes;
fn isatty(&self) -> bool {
let ret = unsafe { libc::isatty(self.to_fd()) };
let error: i32 = errno::errno().into();
match ret {
1 => true,
0 => match error {
libc::EBADF => false,
libc::ENOTTY => false,
_ => {
debug!(
"Unrecognized isatty errno: {}; assuming {:?} is not a TTY",
error, *self
);
false
}
},
_ => {
debug!(
"Unrecognized isatty return code: {}; assuming {:?} is not a TTY",
ret, *self
);
false
| random_line_split |
||
helmrepoutils.go | ("package", "utils", "method", "GetHelmRepoClient")
httpClient := http.DefaultClient
transport := &http.Transport{
Proxy: http.ProxyFromEnvironment,
DialContext: (&net.Dialer{
Timeout: 30 * time.Second,
KeepAlive: 30 * time.Second,
DualStack: true,
}).DialContext,
MaxIdleConns: 100,
IdleConnTimeout: 90 * time.Second,
TLSHandshakeTimeout: 10 * time.Second,
ExpectContinueTimeout: 1 * time.Second,
TLSClientConfig: &tls.Config{
InsecureSkipVerify: false,
},
}
if configMap != nil {
configData := configMap.Data
srLogger.Info("ConfigRef retrieved", "configMap.Data", configData)
if configData["insecureSkipVerify"] != "" {
b, err := strconv.ParseBool(configData["insecureSkipVerify"])
if err != nil {
if errors.IsNotFound(err) {
return nil, nil
}
srLogger.Error(err, "Unable to parse", "insecureSkipVerify", configData["insecureSkipVerify"])
return nil, err
}
srLogger.Info("Set InsecureSkipVerify", "insecureSkipVerify", b)
transport.TLSClientConfig.InsecureSkipVerify = b
} else {
srLogger.Info("insecureSkipVerify is not specified")
}
} else {
srLogger.Info("configMap is nil")
}
httpClient.Transport = transport
srLogger.Info("InsecureSkipVerify equal", "InsecureSkipVerify", transport.TLSClientConfig.InsecureSkipVerify)
return httpClient, nil
}
//GetConfigMap search the config map containing the helm repo client configuration.
func GetConfigMap(client client.Client, parentNamespace string, configMapRef *corev1.ObjectReference) (configMap *corev1.ConfigMap, err error) {
srLogger := log.WithValues("package", "utils", "method", "getConfigMap")
if configMapRef != nil {
srLogger.Info("Retrieve configMap ", "parentNamespace", parentNamespace, "configMapRef.Name", configMapRef.Name)
ns := configMapRef.Namespace
if ns == "" {
ns = parentNamespace
}
configMap = &corev1.ConfigMap{}
err = client.Get(context.TODO(), types.NamespacedName{Namespace: ns, Name: configMapRef.Name}, configMap)
if err != nil {
if errors.IsNotFound(err) {
srLogger.Error(err, "ConfigMap not found ", "Name:", configMapRef.Name, " on namespace: ", ns)
return nil, nil
}
srLogger.Error(err, "Failed to get configMap ", "Name:", configMapRef.Name, " on namespace: ", ns)
return nil, err
}
srLogger.Info("ConfigMap found ", "Name:", configMapRef.Name, " on namespace: ", ns)
} else {
srLogger.Info("no configMapRef defined ", "parentNamespace", parentNamespace)
}
return configMap, err
}
//GetSecret returns the secret to access the helm-repo
func | (client client.Client, parentNamespace string, secretRef *corev1.ObjectReference) (secret *corev1.Secret, err error) {
srLogger := log.WithValues("package", "utils", "method", "getSecret")
if secretRef != nil {
srLogger.Info("Retreive secret", "parentNamespace", parentNamespace, "secretRef", secretRef)
ns := secretRef.Namespace
if ns == "" {
ns = parentNamespace
}
secret = &corev1.Secret{}
err = client.Get(context.TODO(), types.NamespacedName{Namespace: ns, Name: secretRef.Name}, secret)
if err != nil {
srLogger.Error(err, "Failed to get secret ", "Name:", secretRef.Name, " on namespace: ", secretRef.Namespace)
return nil, err
}
srLogger.Info("Secret found ", "Name:", secretRef.Name, " on namespace: ", secretRef.Namespace)
} else {
srLogger.Info("No secret defined", "parentNamespace", parentNamespace)
}
return secret, err
}
func DownloadChart(configMap *corev1.ConfigMap, secret *corev1.Secret, chartsDir string, s *appv1alpha1.HelmRelease) (chartDir string, err error) {
switch strings.ToLower(string(s.Spec.Source.SourceType)) {
case string(appv1alpha1.HelmRepoSourceType):
return DownloadChartFromHelmRepo(configMap, secret, chartsDir, s)
case string(appv1alpha1.GitHubSourceType):
return DownloadChartFromGitHub(configMap, secret, chartsDir, s)
default:
return "", fmt.Errorf("SourceType '%s' unsupported", s.Spec.Source.SourceType)
}
}
//DownloadChartFromGitHub downloads a chart into the charsDir
func DownloadChartFromGitHub(configMap *corev1.ConfigMap, secret *corev1.Secret, chartsDir string, s *appv1alpha1.HelmRelease) (chartDir string, err error) {
srLogger := log.WithValues("HelmRelease.Namespace", s.Namespace, "SubscrptionRelease.Name", s.Name)
if s.Spec.Source.GitHub == nil {
err := fmt.Errorf("GitHub type but Spec.GitHub is not defined")
return "", err
}
if _, err := os.Stat(chartsDir); os.IsNotExist(err) {
err := os.MkdirAll(chartsDir, 0755)
if err != nil {
srLogger.Error(err, "Unable to create chartDir: ", "chartsDir", chartsDir)
return "", err
}
}
destRepo := filepath.Join(chartsDir, s.Spec.ReleaseName, s.Namespace, s.Spec.ChartName)
for _, url := range s.Spec.Source.GitHub.Urls {
options := &git.CloneOptions{
URL: url,
Depth: 1,
SingleBranch: true,
RecurseSubmodules: git.DefaultSubmoduleRecursionDepth,
}
if secret != nil && secret.Data != nil {
srLogger.Info("Add credentials")
options.Auth = &githttp.BasicAuth{
Username: string(secret.Data["user"]),
Password: string(secret.Data["password"]),
}
}
if s.Spec.Source.GitHub.Branch == "" {
options.ReferenceName = plumbing.Master
} else {
options.ReferenceName = plumbing.ReferenceName(s.Spec.Source.GitHub.Branch)
}
os.RemoveAll(chartDir)
_, err = git.PlainClone(destRepo, false, options)
if err != nil {
os.RemoveAll(destRepo)
srLogger.Error(err, "Clone failed", "url", url)
continue
}
}
if err != nil {
srLogger.Error(err, "All urls failed")
}
chartDir = filepath.Join(destRepo, s.Spec.Source.GitHub.ChartPath)
return chartDir, err
}
//DownloadChartFromHelmRepo downloads a chart into the charsDir
func DownloadChartFromHelmRepo(configMap *corev1.ConfigMap, secret *corev1.Secret, chartsDir string, s *appv1alpha1.HelmRelease) (chartDir string, err error) {
srLogger := log.WithValues("HelmRelease.Namespace", s.Namespace, "SubscrptionRelease.Name", s.Name)
if s.Spec.Source.HelmRepo == nil {
err := fmt.Errorf("HelmRepo type but Spec.HelmRepo is not defined")
return "", err
}
if _, err := os.Stat(chartsDir); os.IsNotExist(err) {
err := os.MkdirAll(chartsDir, 0755)
if err != nil {
srLogger.Error(err, "Unable to create chartDir: ", "chartsDir", chartsDir)
return "", err
}
}
httpClient, err := GetHelmRepoClient(s.Namespace, configMap)
if err != nil {
srLogger.Error(err, "Failed to create httpClient ", "sr.Spec.SecretRef.Name", s.Spec.SecretRef.Name)
return "", err
}
var downloadErr error
for _, urlelem := range s.Spec.Source.HelmRepo.Urls {
var URLP *url.URL
URLP, downloadErr = url.Parse(urlelem)
if err != nil {
srLogger.Error(downloadErr, "url", urlelem)
continue
}
fileName := filepath.Base(URLP.Path)
// Create the file
chartZip := filepath.Join(chartsDir, fileName)
if _, err := os.Stat(chartZip); os.IsNotExist(err) {
var req *http.Request
req, downloadErr = http.NewRequest(http.MethodGet, urlelem, nil)
if downloadErr != nil {
srLogger.Error(downloadErr, "Can not build request: ", "urlelem", urlelem)
continue
}
if secret != nil && secret.Data != nil {
req.SetBasicAuth(string(secret.Data["user"]), string(secret.Data["password"]))
}
var resp *http.Response
resp, downloadErr = httpClient.Do(req)
if downloadErr != nil {
srLogger.Error(downloadErr, | GetSecret | identifier_name |
helmrepoutils.go | Values("package", "utils", "method", "GetHelmRepoClient")
httpClient := http.DefaultClient
transport := &http.Transport{
Proxy: http.ProxyFromEnvironment,
DialContext: (&net.Dialer{
Timeout: 30 * time.Second,
KeepAlive: 30 * time.Second,
DualStack: true,
}).DialContext,
MaxIdleConns: 100,
IdleConnTimeout: 90 * time.Second,
TLSHandshakeTimeout: 10 * time.Second,
ExpectContinueTimeout: 1 * time.Second,
TLSClientConfig: &tls.Config{
InsecureSkipVerify: false,
},
}
if configMap != nil {
configData := configMap.Data
srLogger.Info("ConfigRef retrieved", "configMap.Data", configData)
if configData["insecureSkipVerify"] != "" {
b, err := strconv.ParseBool(configData["insecureSkipVerify"])
if err != nil {
if errors.IsNotFound(err) {
return nil, nil
}
srLogger.Error(err, "Unable to parse", "insecureSkipVerify", configData["insecureSkipVerify"])
return nil, err
}
srLogger.Info("Set InsecureSkipVerify", "insecureSkipVerify", b)
transport.TLSClientConfig.InsecureSkipVerify = b
} else {
srLogger.Info("insecureSkipVerify is not specified")
}
} else {
srLogger.Info("configMap is nil")
}
httpClient.Transport = transport
srLogger.Info("InsecureSkipVerify equal", "InsecureSkipVerify", transport.TLSClientConfig.InsecureSkipVerify)
return httpClient, nil
}
//GetConfigMap search the config map containing the helm repo client configuration.
func GetConfigMap(client client.Client, parentNamespace string, configMapRef *corev1.ObjectReference) (configMap *corev1.ConfigMap, err error) {
srLogger := log.WithValues("package", "utils", "method", "getConfigMap")
if configMapRef != nil {
srLogger.Info("Retrieve configMap ", "parentNamespace", parentNamespace, "configMapRef.Name", configMapRef.Name)
ns := configMapRef.Namespace
if ns == "" {
ns = parentNamespace
}
configMap = &corev1.ConfigMap{}
err = client.Get(context.TODO(), types.NamespacedName{Namespace: ns, Name: configMapRef.Name}, configMap)
if err != nil {
if errors.IsNotFound(err) {
srLogger.Error(err, "ConfigMap not found ", "Name:", configMapRef.Name, " on namespace: ", ns)
return nil, nil
}
srLogger.Error(err, "Failed to get configMap ", "Name:", configMapRef.Name, " on namespace: ", ns)
return nil, err
}
srLogger.Info("ConfigMap found ", "Name:", configMapRef.Name, " on namespace: ", ns)
} else {
srLogger.Info("no configMapRef defined ", "parentNamespace", parentNamespace)
}
return configMap, err
}
//GetSecret returns the secret to access the helm-repo
func GetSecret(client client.Client, parentNamespace string, secretRef *corev1.ObjectReference) (secret *corev1.Secret, err error) {
srLogger := log.WithValues("package", "utils", "method", "getSecret")
if secretRef != nil {
srLogger.Info("Retreive secret", "parentNamespace", parentNamespace, "secretRef", secretRef)
ns := secretRef.Namespace
if ns == "" {
ns = parentNamespace
}
secret = &corev1.Secret{}
err = client.Get(context.TODO(), types.NamespacedName{Namespace: ns, Name: secretRef.Name}, secret)
if err != nil {
srLogger.Error(err, "Failed to get secret ", "Name:", secretRef.Name, " on namespace: ", secretRef.Namespace)
return nil, err
}
srLogger.Info("Secret found ", "Name:", secretRef.Name, " on namespace: ", secretRef.Namespace)
} else {
srLogger.Info("No secret defined", "parentNamespace", parentNamespace)
}
return secret, err
}
func DownloadChart(configMap *corev1.ConfigMap, secret *corev1.Secret, chartsDir string, s *appv1alpha1.HelmRelease) (chartDir string, err error) {
switch strings.ToLower(string(s.Spec.Source.SourceType)) {
case string(appv1alpha1.HelmRepoSourceType):
return DownloadChartFromHelmRepo(configMap, secret, chartsDir, s)
case string(appv1alpha1.GitHubSourceType):
return DownloadChartFromGitHub(configMap, secret, chartsDir, s)
default:
return "", fmt.Errorf("SourceType '%s' unsupported", s.Spec.Source.SourceType)
}
}
//DownloadChartFromGitHub downloads a chart into the charsDir
func DownloadChartFromGitHub(configMap *corev1.ConfigMap, secret *corev1.Secret, chartsDir string, s *appv1alpha1.HelmRelease) (chartDir string, err error) {
srLogger := log.WithValues("HelmRelease.Namespace", s.Namespace, "SubscrptionRelease.Name", s.Name)
if s.Spec.Source.GitHub == nil {
err := fmt.Errorf("GitHub type but Spec.GitHub is not defined")
return "", err
}
if _, err := os.Stat(chartsDir); os.IsNotExist(err) {
err := os.MkdirAll(chartsDir, 0755)
if err != nil {
srLogger.Error(err, "Unable to create chartDir: ", "chartsDir", chartsDir)
return "", err
}
}
destRepo := filepath.Join(chartsDir, s.Spec.ReleaseName, s.Namespace, s.Spec.ChartName)
for _, url := range s.Spec.Source.GitHub.Urls {
options := &git.CloneOptions{
URL: url,
Depth: 1,
SingleBranch: true,
RecurseSubmodules: git.DefaultSubmoduleRecursionDepth,
}
if secret != nil && secret.Data != nil {
srLogger.Info("Add credentials")
options.Auth = &githttp.BasicAuth{
Username: string(secret.Data["user"]),
Password: string(secret.Data["password"]),
}
}
if s.Spec.Source.GitHub.Branch == "" {
options.ReferenceName = plumbing.Master
} else {
options.ReferenceName = plumbing.ReferenceName(s.Spec.Source.GitHub.Branch)
}
os.RemoveAll(chartDir)
_, err = git.PlainClone(destRepo, false, options)
if err != nil {
os.RemoveAll(destRepo)
srLogger.Error(err, "Clone failed", "url", url)
continue
}
}
if err != nil {
srLogger.Error(err, "All urls failed")
}
chartDir = filepath.Join(destRepo, s.Spec.Source.GitHub.ChartPath)
return chartDir, err
}
//DownloadChartFromHelmRepo downloads a chart into the charsDir | srLogger := log.WithValues("HelmRelease.Namespace", s.Namespace, "SubscrptionRelease.Name", s.Name)
if s.Spec.Source.HelmRepo == nil {
err := fmt.Errorf("HelmRepo type but Spec.HelmRepo is not defined")
return "", err
}
if _, err := os.Stat(chartsDir); os.IsNotExist(err) {
err := os.MkdirAll(chartsDir, 0755)
if err != nil {
srLogger.Error(err, "Unable to create chartDir: ", "chartsDir", chartsDir)
return "", err
}
}
httpClient, err := GetHelmRepoClient(s.Namespace, configMap)
if err != nil {
srLogger.Error(err, "Failed to create httpClient ", "sr.Spec.SecretRef.Name", s.Spec.SecretRef.Name)
return "", err
}
var downloadErr error
for _, urlelem := range s.Spec.Source.HelmRepo.Urls {
var URLP *url.URL
URLP, downloadErr = url.Parse(urlelem)
if err != nil {
srLogger.Error(downloadErr, "url", urlelem)
continue
}
fileName := filepath.Base(URLP.Path)
// Create the file
chartZip := filepath.Join(chartsDir, fileName)
if _, err := os.Stat(chartZip); os.IsNotExist(err) {
var req *http.Request
req, downloadErr = http.NewRequest(http.MethodGet, urlelem, nil)
if downloadErr != nil {
srLogger.Error(downloadErr, "Can not build request: ", "urlelem", urlelem)
continue
}
if secret != nil && secret.Data != nil {
req.SetBasicAuth(string(secret.Data["user"]), string(secret.Data["password"]))
}
var resp *http.Response
resp, downloadErr = httpClient.Do(req)
if downloadErr != nil {
srLogger.Error(downloadErr, " | func DownloadChartFromHelmRepo(configMap *corev1.ConfigMap, secret *corev1.Secret, chartsDir string, s *appv1alpha1.HelmRelease) (chartDir string, err error) { | random_line_split |
helmrepoutils.go | namespace: ", ns)
} else {
srLogger.Info("no configMapRef defined ", "parentNamespace", parentNamespace)
}
return configMap, err
}
//GetSecret returns the secret to access the helm-repo
func GetSecret(client client.Client, parentNamespace string, secretRef *corev1.ObjectReference) (secret *corev1.Secret, err error) {
srLogger := log.WithValues("package", "utils", "method", "getSecret")
if secretRef != nil {
srLogger.Info("Retreive secret", "parentNamespace", parentNamespace, "secretRef", secretRef)
ns := secretRef.Namespace
if ns == "" {
ns = parentNamespace
}
secret = &corev1.Secret{}
err = client.Get(context.TODO(), types.NamespacedName{Namespace: ns, Name: secretRef.Name}, secret)
if err != nil {
srLogger.Error(err, "Failed to get secret ", "Name:", secretRef.Name, " on namespace: ", secretRef.Namespace)
return nil, err
}
srLogger.Info("Secret found ", "Name:", secretRef.Name, " on namespace: ", secretRef.Namespace)
} else {
srLogger.Info("No secret defined", "parentNamespace", parentNamespace)
}
return secret, err
}
func DownloadChart(configMap *corev1.ConfigMap, secret *corev1.Secret, chartsDir string, s *appv1alpha1.HelmRelease) (chartDir string, err error) {
switch strings.ToLower(string(s.Spec.Source.SourceType)) {
case string(appv1alpha1.HelmRepoSourceType):
return DownloadChartFromHelmRepo(configMap, secret, chartsDir, s)
case string(appv1alpha1.GitHubSourceType):
return DownloadChartFromGitHub(configMap, secret, chartsDir, s)
default:
return "", fmt.Errorf("SourceType '%s' unsupported", s.Spec.Source.SourceType)
}
}
//DownloadChartFromGitHub downloads a chart into the charsDir
func DownloadChartFromGitHub(configMap *corev1.ConfigMap, secret *corev1.Secret, chartsDir string, s *appv1alpha1.HelmRelease) (chartDir string, err error) {
srLogger := log.WithValues("HelmRelease.Namespace", s.Namespace, "SubscrptionRelease.Name", s.Name)
if s.Spec.Source.GitHub == nil {
err := fmt.Errorf("GitHub type but Spec.GitHub is not defined")
return "", err
}
if _, err := os.Stat(chartsDir); os.IsNotExist(err) {
err := os.MkdirAll(chartsDir, 0755)
if err != nil {
srLogger.Error(err, "Unable to create chartDir: ", "chartsDir", chartsDir)
return "", err
}
}
destRepo := filepath.Join(chartsDir, s.Spec.ReleaseName, s.Namespace, s.Spec.ChartName)
for _, url := range s.Spec.Source.GitHub.Urls {
options := &git.CloneOptions{
URL: url,
Depth: 1,
SingleBranch: true,
RecurseSubmodules: git.DefaultSubmoduleRecursionDepth,
}
if secret != nil && secret.Data != nil {
srLogger.Info("Add credentials")
options.Auth = &githttp.BasicAuth{
Username: string(secret.Data["user"]),
Password: string(secret.Data["password"]),
}
}
if s.Spec.Source.GitHub.Branch == "" {
options.ReferenceName = plumbing.Master
} else {
options.ReferenceName = plumbing.ReferenceName(s.Spec.Source.GitHub.Branch)
}
os.RemoveAll(chartDir)
_, err = git.PlainClone(destRepo, false, options)
if err != nil {
os.RemoveAll(destRepo)
srLogger.Error(err, "Clone failed", "url", url)
continue
}
}
if err != nil {
srLogger.Error(err, "All urls failed")
}
chartDir = filepath.Join(destRepo, s.Spec.Source.GitHub.ChartPath)
return chartDir, err
}
//DownloadChartFromHelmRepo downloads a chart into the charsDir
func DownloadChartFromHelmRepo(configMap *corev1.ConfigMap, secret *corev1.Secret, chartsDir string, s *appv1alpha1.HelmRelease) (chartDir string, err error) {
srLogger := log.WithValues("HelmRelease.Namespace", s.Namespace, "SubscrptionRelease.Name", s.Name)
if s.Spec.Source.HelmRepo == nil {
err := fmt.Errorf("HelmRepo type but Spec.HelmRepo is not defined")
return "", err
}
if _, err := os.Stat(chartsDir); os.IsNotExist(err) {
err := os.MkdirAll(chartsDir, 0755)
if err != nil {
srLogger.Error(err, "Unable to create chartDir: ", "chartsDir", chartsDir)
return "", err
}
}
httpClient, err := GetHelmRepoClient(s.Namespace, configMap)
if err != nil {
srLogger.Error(err, "Failed to create httpClient ", "sr.Spec.SecretRef.Name", s.Spec.SecretRef.Name)
return "", err
}
var downloadErr error
for _, urlelem := range s.Spec.Source.HelmRepo.Urls {
var URLP *url.URL
URLP, downloadErr = url.Parse(urlelem)
if err != nil {
srLogger.Error(downloadErr, "url", urlelem)
continue
}
fileName := filepath.Base(URLP.Path)
// Create the file
chartZip := filepath.Join(chartsDir, fileName)
if _, err := os.Stat(chartZip); os.IsNotExist(err) {
var req *http.Request
req, downloadErr = http.NewRequest(http.MethodGet, urlelem, nil)
if downloadErr != nil {
srLogger.Error(downloadErr, "Can not build request: ", "urlelem", urlelem)
continue
}
if secret != nil && secret.Data != nil {
req.SetBasicAuth(string(secret.Data["user"]), string(secret.Data["password"]))
}
var resp *http.Response
resp, downloadErr = httpClient.Do(req)
if downloadErr != nil {
srLogger.Error(downloadErr, "Http request failed: ", "urlelem", urlelem)
continue
}
srLogger.Info("Get suceeded: ", "urlelem", urlelem)
defer resp.Body.Close()
var out *os.File
out, downloadErr = os.Create(chartZip)
if downloadErr != nil {
srLogger.Error(downloadErr, "Failed to create: ", "chartZip", chartZip)
continue
}
defer out.Close()
// Write the body to file
_, downloadErr = io.Copy(out, resp.Body)
if downloadErr != nil {
srLogger.Error(downloadErr, "Failed to copy body: ", "chartZip", chartZip)
continue
}
}
var r *os.File
r, downloadErr = os.Open(chartZip)
if downloadErr != nil {
srLogger.Error(downloadErr, "Failed to open: ", "chartZip", chartZip)
continue
}
chartDirUnzip := filepath.Join(chartsDir, s.Spec.ReleaseName, s.Namespace)
chartDir = filepath.Join(chartDirUnzip, s.Spec.ChartName)
//Clean before untar
os.RemoveAll(chartDirUnzip)
downloadErr = Untar(chartDirUnzip, r)
if downloadErr != nil {
//Remove zip because failed to untar and so probably corrupted
os.RemoveAll(chartZip)
srLogger.Error(downloadErr, "Failed to unzip: ", "chartZip", chartZip)
continue
}
}
return chartDir, downloadErr
}
//Untar untars the reader into the dst directory
func Untar(dst string, r io.Reader) error {
srLogger := log.WithValues("destination", dst)
gzr, err := gzip.NewReader(r)
if err != nil {
srLogger.Error(err, "")
return err
}
defer gzr.Close()
tr := tar.NewReader(gzr)
for {
header, err := tr.Next()
switch {
// if no more files are found return
case err == io.EOF:
return nil
// return any other error
case err != nil:
srLogger.Error(err, "")
return err
// if the header is nil, just skip it (not sure how this happens)
case header == nil:
continue
}
// the target location where the dir/file should be created
target := filepath.Join(dst, header.Name)
// the following switch could also be done using fi.Mode(), not sure if there
// a benefit of using one vs. the other.
// fi := header.FileInfo()
// check the file type
switch header.Typeflag {
// if its a dir and it doesn't exist create it
case tar.TypeDir:
if _, err := os.Stat(target); err != nil {
if err := os.MkdirAll(target, 0755); err != nil | {
srLogger.Error(err, "")
return err
} | conditional_block |
|
helmrepoutils.go | ("package", "utils", "method", "GetHelmRepoClient")
httpClient := http.DefaultClient
transport := &http.Transport{
Proxy: http.ProxyFromEnvironment,
DialContext: (&net.Dialer{
Timeout: 30 * time.Second,
KeepAlive: 30 * time.Second,
DualStack: true,
}).DialContext,
MaxIdleConns: 100,
IdleConnTimeout: 90 * time.Second,
TLSHandshakeTimeout: 10 * time.Second,
ExpectContinueTimeout: 1 * time.Second,
TLSClientConfig: &tls.Config{
InsecureSkipVerify: false,
},
}
if configMap != nil {
configData := configMap.Data
srLogger.Info("ConfigRef retrieved", "configMap.Data", configData)
if configData["insecureSkipVerify"] != "" {
b, err := strconv.ParseBool(configData["insecureSkipVerify"])
if err != nil {
if errors.IsNotFound(err) {
return nil, nil
}
srLogger.Error(err, "Unable to parse", "insecureSkipVerify", configData["insecureSkipVerify"])
return nil, err
}
srLogger.Info("Set InsecureSkipVerify", "insecureSkipVerify", b)
transport.TLSClientConfig.InsecureSkipVerify = b
} else {
srLogger.Info("insecureSkipVerify is not specified")
}
} else {
srLogger.Info("configMap is nil")
}
httpClient.Transport = transport
srLogger.Info("InsecureSkipVerify equal", "InsecureSkipVerify", transport.TLSClientConfig.InsecureSkipVerify)
return httpClient, nil
}
//GetConfigMap search the config map containing the helm repo client configuration.
func GetConfigMap(client client.Client, parentNamespace string, configMapRef *corev1.ObjectReference) (configMap *corev1.ConfigMap, err error) {
srLogger := log.WithValues("package", "utils", "method", "getConfigMap")
if configMapRef != nil {
srLogger.Info("Retrieve configMap ", "parentNamespace", parentNamespace, "configMapRef.Name", configMapRef.Name)
ns := configMapRef.Namespace
if ns == "" {
ns = parentNamespace
}
configMap = &corev1.ConfigMap{}
err = client.Get(context.TODO(), types.NamespacedName{Namespace: ns, Name: configMapRef.Name}, configMap)
if err != nil {
if errors.IsNotFound(err) {
srLogger.Error(err, "ConfigMap not found ", "Name:", configMapRef.Name, " on namespace: ", ns)
return nil, nil
}
srLogger.Error(err, "Failed to get configMap ", "Name:", configMapRef.Name, " on namespace: ", ns)
return nil, err
}
srLogger.Info("ConfigMap found ", "Name:", configMapRef.Name, " on namespace: ", ns)
} else {
srLogger.Info("no configMapRef defined ", "parentNamespace", parentNamespace)
}
return configMap, err
}
//GetSecret returns the secret to access the helm-repo
func GetSecret(client client.Client, parentNamespace string, secretRef *corev1.ObjectReference) (secret *corev1.Secret, err error) {
srLogger := log.WithValues("package", "utils", "method", "getSecret")
if secretRef != nil {
srLogger.Info("Retreive secret", "parentNamespace", parentNamespace, "secretRef", secretRef)
ns := secretRef.Namespace
if ns == "" {
ns = parentNamespace
}
secret = &corev1.Secret{}
err = client.Get(context.TODO(), types.NamespacedName{Namespace: ns, Name: secretRef.Name}, secret)
if err != nil {
srLogger.Error(err, "Failed to get secret ", "Name:", secretRef.Name, " on namespace: ", secretRef.Namespace)
return nil, err
}
srLogger.Info("Secret found ", "Name:", secretRef.Name, " on namespace: ", secretRef.Namespace)
} else {
srLogger.Info("No secret defined", "parentNamespace", parentNamespace)
}
return secret, err
}
func DownloadChart(configMap *corev1.ConfigMap, secret *corev1.Secret, chartsDir string, s *appv1alpha1.HelmRelease) (chartDir string, err error) {
switch strings.ToLower(string(s.Spec.Source.SourceType)) {
case string(appv1alpha1.HelmRepoSourceType):
return DownloadChartFromHelmRepo(configMap, secret, chartsDir, s)
case string(appv1alpha1.GitHubSourceType):
return DownloadChartFromGitHub(configMap, secret, chartsDir, s)
default:
return "", fmt.Errorf("SourceType '%s' unsupported", s.Spec.Source.SourceType)
}
}
//DownloadChartFromGitHub downloads a chart into the charsDir
func DownloadChartFromGitHub(configMap *corev1.ConfigMap, secret *corev1.Secret, chartsDir string, s *appv1alpha1.HelmRelease) (chartDir string, err error) | }
if secret != nil && secret.Data != nil {
srLogger.Info("Add credentials")
options.Auth = &githttp.BasicAuth{
Username: string(secret.Data["user"]),
Password: string(secret.Data["password"]),
}
}
if s.Spec.Source.GitHub.Branch == "" {
options.ReferenceName = plumbing.Master
} else {
options.ReferenceName = plumbing.ReferenceName(s.Spec.Source.GitHub.Branch)
}
os.RemoveAll(chartDir)
_, err = git.PlainClone(destRepo, false, options)
if err != nil {
os.RemoveAll(destRepo)
srLogger.Error(err, "Clone failed", "url", url)
continue
}
}
if err != nil {
srLogger.Error(err, "All urls failed")
}
chartDir = filepath.Join(destRepo, s.Spec.Source.GitHub.ChartPath)
return chartDir, err
}
//DownloadChartFromHelmRepo downloads a chart into the charsDir
func DownloadChartFromHelmRepo(configMap *corev1.ConfigMap, secret *corev1.Secret, chartsDir string, s *appv1alpha1.HelmRelease) (chartDir string, err error) {
srLogger := log.WithValues("HelmRelease.Namespace", s.Namespace, "SubscrptionRelease.Name", s.Name)
if s.Spec.Source.HelmRepo == nil {
err := fmt.Errorf("HelmRepo type but Spec.HelmRepo is not defined")
return "", err
}
if _, err := os.Stat(chartsDir); os.IsNotExist(err) {
err := os.MkdirAll(chartsDir, 0755)
if err != nil {
srLogger.Error(err, "Unable to create chartDir: ", "chartsDir", chartsDir)
return "", err
}
}
httpClient, err := GetHelmRepoClient(s.Namespace, configMap)
if err != nil {
srLogger.Error(err, "Failed to create httpClient ", "sr.Spec.SecretRef.Name", s.Spec.SecretRef.Name)
return "", err
}
var downloadErr error
for _, urlelem := range s.Spec.Source.HelmRepo.Urls {
var URLP *url.URL
URLP, downloadErr = url.Parse(urlelem)
if err != nil {
srLogger.Error(downloadErr, "url", urlelem)
continue
}
fileName := filepath.Base(URLP.Path)
// Create the file
chartZip := filepath.Join(chartsDir, fileName)
if _, err := os.Stat(chartZip); os.IsNotExist(err) {
var req *http.Request
req, downloadErr = http.NewRequest(http.MethodGet, urlelem, nil)
if downloadErr != nil {
srLogger.Error(downloadErr, "Can not build request: ", "urlelem", urlelem)
continue
}
if secret != nil && secret.Data != nil {
req.SetBasicAuth(string(secret.Data["user"]), string(secret.Data["password"]))
}
var resp *http.Response
resp, downloadErr = httpClient.Do(req)
if downloadErr != nil {
srLogger.Error(downloadErr, | {
srLogger := log.WithValues("HelmRelease.Namespace", s.Namespace, "SubscrptionRelease.Name", s.Name)
if s.Spec.Source.GitHub == nil {
err := fmt.Errorf("GitHub type but Spec.GitHub is not defined")
return "", err
}
if _, err := os.Stat(chartsDir); os.IsNotExist(err) {
err := os.MkdirAll(chartsDir, 0755)
if err != nil {
srLogger.Error(err, "Unable to create chartDir: ", "chartsDir", chartsDir)
return "", err
}
}
destRepo := filepath.Join(chartsDir, s.Spec.ReleaseName, s.Namespace, s.Spec.ChartName)
for _, url := range s.Spec.Source.GitHub.Urls {
options := &git.CloneOptions{
URL: url,
Depth: 1,
SingleBranch: true,
RecurseSubmodules: git.DefaultSubmoduleRecursionDepth, | identifier_body |
ResNet50_augmentation.py |
def get_dataset(ds_path):
ds_path = pathlib.Path(ds_path)
images = list(ds_path.glob('*/*.jpg'))
images = [str(path) for path in images]
total_images = len(images)
labels = sorted(item.name for item in ds_path.glob('*/') if item.is_dir())
classes = labels
labels = dict((name, index) for index, name in enumerate(labels))
labels = [labels[pathlib.Path(path).parent.name] for path in images]
labels = tf.keras.utils.to_categorical(labels, num_classes=len(classes), dtype='float32')
return images, labels, classes
def aug_fn(image):
data = {"image":image}
aug_data = transforms(**data)
aug_img = aug_data["image"]
aug_img = tf.cast(aug_img, tf.float32) / 255.0
aug_img = tf.image.per_image_standardization(aug_img)
# aug_img = tf.keras.applications.resnet.preprocess_input(aug_img)
return aug_img
def process_data(image, label):
aug_img = tf.numpy_function(func=aug_fn, inp=[image], Tout=tf.float32)
return aug_img, label
def make_tf_data(images, labels, augmentation):
images = tf.data.Dataset.from_tensor_slices(images)
images = images.map(preprocess_image, num_parallel_calls=AUTOTUNE)
labels = tf.data.Dataset.from_tensor_slices(labels)
dataset = tf.data.Dataset.zip((images, labels))
dataset = dataset.repeat()
if augmentation: |
dataset = dataset.batch(BATCH_SIZE)
dataset = dataset.prefetch(AUTOTUNE)
return dataset
def residual_block(x, filters, kernel_size=3, stride=1, conv_shortcut=True, name=None):
if conv_shortcut:
shortcut = tf.keras.layers.Conv2D(4 * filters, 1, strides=stride, name=name+'_0_conv', kernel_initializer='he_uniform', bias_initializer='he_uniform', kernel_regularizer=tf.keras.regularizers.l2(0.0001))(x)
shortcut = tf.keras.layers.BatchNormalization(axis=3, name=name+'_0_bn')(shortcut)
else:
shortcut = x
x = tf.keras.layers.Conv2D(filters, 1, strides=stride, name=name + '_1_conv', kernel_initializer='he_uniform', bias_initializer='he_uniform', kernel_regularizer=tf.keras.regularizers.l2(0.0001))(x)
x = tf.keras.layers.BatchNormalization(axis=3, name=name + '_1_bn')(x)
x = tf.keras.layers.Activation('relu', name=name + '_1_relu')(x)
x = tf.keras.layers.Conv2D(filters, kernel_size, padding='SAME', name=name + '_2_conv', kernel_initializer='he_uniform', bias_initializer='he_uniform', kernel_regularizer=tf.keras.regularizers.l2(0.0001))(x)
x = tf.keras.layers.BatchNormalization(axis=3, name=name + '_2_bn')(x)
x = tf.keras.layers.Activation('relu', name=name + '_2_relu')(x)
x = tf.keras.layers.Conv2D(4 * filters, 1, name=name + '_3_conv', kernel_initializer='he_uniform', bias_initializer='he_uniform', kernel_regularizer=tf.keras.regularizers.l2(0.0001))(x)
x = tf.keras.layers.BatchNormalization(axis=3, name=name + '_3_bn')(x)
x = tf.keras.layers.Add(name=name + '_add')([shortcut, x])
x = tf.keras.layers.Activation('relu', name=name + '_out')(x)
return x
def residual_stack(x, filters, blocks, stride1=2, name=None):
x = residual_block(x, filters, stride=stride1, name=name + '_block1')
for i in range(2, blocks + 1):
x = residual_block(x, filters, conv_shortcut=False, name=name + '_block' + str(i))
return x
def ResNet50():
inputs = tf.keras.layers.Input(shape=INPUT_SHAPE)
x = tf.keras.layers.ZeroPadding2D(padding=((3, 3), (3, 3)), name='conv1_pad')(inputs)
x = tf.keras.layers.Conv2D(64, 7, strides=2, use_bias=True, name='conv1_conv', kernel_initializer='he_uniform', bias_initializer='he_uniform', kernel_regularizer=tf.keras.regularizers.l2(0.0001))(x)
x = tf.keras.layers.BatchNormalization(axis=3, name='conv1_bn')(x)
x = tf.keras.layers.Activation('relu', name='conv1_relu')(x)
x = tf.keras.layers.ZeroPadding2D(padding=((1, 1), (1, 1)), name='pool1_pad')(x)
x = tf.keras.layers.MaxPooling2D(3, strides=2, name='pool1_pool')(x)
x = residual_stack(x, 64, 3, stride1=1, name='conv2')
x = residual_stack(x, 128, 4, name='conv3')
x = residual_stack(x, 256, 6, name='conv4')
x = residual_stack(x, 512, 3, name='conv5')
x = tf.keras.layers.GlobalAveragePooling2D(name='avg_pool')(x)
outputs = tf.keras.layers.Dense(n_classes, activation='softmax', name='predictions')(x)
model = tf.keras.Model(inputs=inputs, outputs=outputs)
return model
@tf.function
def train(model, images, labels):
with tf.GradientTape() as tape:
y_pred = model(images, training=True)
loss = tf.reduce_mean(cost_fn(labels, y_pred))
grads = tape.gradient(loss, model.trainable_variables)
optimizer.apply_gradients(grads_and_vars=zip(grads, model.trainable_variables))
train_acc.update_state(labels, y_pred)
train_loss.update_state(labels, y_pred)
@tf.function
def validation(model, images, labels):
y_pred = model(images, training=False)
loss = tf.reduce_mean(cost_fn(labels, y_pred))
val_acc.update_state(labels, y_pred)
val_loss.update_state(labels, y_pred)
def lrfn():
if epoch < LR_RAMPUP_EPOCHS:
lr = (LR_MAX - LR_START) / LR_RAMPUP_EPOCHS * epoch + LR_START
elif epoch < LR_RAMPUP_EPOCHS + LR_SUSTAIN_EPOCHS:
lr = LR_MAX
else:
lr = (LR_MAX - LR_MIN) * LR_EXP_DECAY**(epoch - LR_RAMPUP_EPOCHS - LR_SUSTAIN_EPOCHS) + LR_MIN
return lr
def tf_data_visualize(augmentation_element, name):
row, col, idx = 5, 4, 0
row = min(row, BATCH_SIZE // col)
for (image, label) in augmentation_element:
print(image.shape, label.shape)
image = image / 255.0
plt.figure(figsize=(15, int(15 * row / col)))
for j in range(row * col):
plt.subplot(row, col, j + 1)
plt.axis('off')
plt.imshow(image[j, ])
# plt.savefig(f'{SAVED_PATH}/{LOG_TIME}/{name}_{idx}.jpg')
plt.show()
idx += 1
if idx == 3:
break
if __name__ == "__main__":
# hyper parameters
AUTOTUNE = tf.data.experimental.AUTOTUNE
IMG_SIZE = 224
INPUT_SHAPE = (IMG_SIZE, IMG_SIZE, 3)
BATCH_SIZE = 32
EPOCHS = 1000
# learning rate scheduler
LR_START = 0.001
LR_MAX = 0.005 * strategy.num_replicas_in_sync
LR_MIN = 0.001
LR_RAMPUP_EPOCHS = 5
LR_SUSTAIN_EPOCHS = 0
LR_EXP_DECAY = .8
# early stopping
PATIENCE = 3
EARLY_STOPPING = True
minimum_loss = float(2147000000)
total_images, total_labels, CLASSES = get_dataset('/home/v100/tf_workspace/datasets/natural_images/natural_images')
n_classes = len(CLASSES)
train_images, valid_images, train_labels, valid_labels = train_test_split(total_images, total_labels, test_size=.3, shuffle=True, random_state=777)
TRAIN_STEPS_PER_EPOCH = int(tf.math.ceil(len(train_images) / BATCH_SIZE).numpy())
VALID_STEP_PER_EPOCH = int(tf.math.ceil(len(valid_images) / BATCH_SIZE).numpy())
cost_fn = tf.keras.losses.CategoricalCrossentropy()
# optimizer = tf.keras.optimizers.Adam(learning_rate=lrfn)
optimizer = tf.keras.optimizers.Adam(learning_rate=0.00001)
inputs = tf.keras.Input(shape=(INPUT_SHAPE))
model = ResNet50()
model(inputs=inputs)
model.summary()
# tf.keras.utils.plot_model(model, show_shapes=True)
train_acc = | dataset = dataset.map(partial(process_data), num_parallel_calls=AUTOTUNE) | random_line_split |
ResNet50_augmentation.py |
def get_dataset(ds_path):
ds_path = pathlib.Path(ds_path)
images = list(ds_path.glob('*/*.jpg'))
images = [str(path) for path in images]
total_images = len(images)
labels = sorted(item.name for item in ds_path.glob('*/') if item.is_dir())
classes = labels
labels = dict((name, index) for index, name in enumerate(labels))
labels = [labels[pathlib.Path(path).parent.name] for path in images]
labels = tf.keras.utils.to_categorical(labels, num_classes=len(classes), dtype='float32')
return images, labels, classes
def aug_fn(image):
data = {"image":image}
aug_data = transforms(**data)
aug_img = aug_data["image"]
aug_img = tf.cast(aug_img, tf.float32) / 255.0
aug_img = tf.image.per_image_standardization(aug_img)
# aug_img = tf.keras.applications.resnet.preprocess_input(aug_img)
return aug_img
def process_data(image, label):
|
def make_tf_data(images, labels, augmentation):
images = tf.data.Dataset.from_tensor_slices(images)
images = images.map(preprocess_image, num_parallel_calls=AUTOTUNE)
labels = tf.data.Dataset.from_tensor_slices(labels)
dataset = tf.data.Dataset.zip((images, labels))
dataset = dataset.repeat()
if augmentation:
dataset = dataset.map(partial(process_data), num_parallel_calls=AUTOTUNE)
dataset = dataset.batch(BATCH_SIZE)
dataset = dataset.prefetch(AUTOTUNE)
return dataset
def residual_block(x, filters, kernel_size=3, stride=1, conv_shortcut=True, name=None):
if conv_shortcut:
shortcut = tf.keras.layers.Conv2D(4 * filters, 1, strides=stride, name=name+'_0_conv', kernel_initializer='he_uniform', bias_initializer='he_uniform', kernel_regularizer=tf.keras.regularizers.l2(0.0001))(x)
shortcut = tf.keras.layers.BatchNormalization(axis=3, name=name+'_0_bn')(shortcut)
else:
shortcut = x
x = tf.keras.layers.Conv2D(filters, 1, strides=stride, name=name + '_1_conv', kernel_initializer='he_uniform', bias_initializer='he_uniform', kernel_regularizer=tf.keras.regularizers.l2(0.0001))(x)
x = tf.keras.layers.BatchNormalization(axis=3, name=name + '_1_bn')(x)
x = tf.keras.layers.Activation('relu', name=name + '_1_relu')(x)
x = tf.keras.layers.Conv2D(filters, kernel_size, padding='SAME', name=name + '_2_conv', kernel_initializer='he_uniform', bias_initializer='he_uniform', kernel_regularizer=tf.keras.regularizers.l2(0.0001))(x)
x = tf.keras.layers.BatchNormalization(axis=3, name=name + '_2_bn')(x)
x = tf.keras.layers.Activation('relu', name=name + '_2_relu')(x)
x = tf.keras.layers.Conv2D(4 * filters, 1, name=name + '_3_conv', kernel_initializer='he_uniform', bias_initializer='he_uniform', kernel_regularizer=tf.keras.regularizers.l2(0.0001))(x)
x = tf.keras.layers.BatchNormalization(axis=3, name=name + '_3_bn')(x)
x = tf.keras.layers.Add(name=name + '_add')([shortcut, x])
x = tf.keras.layers.Activation('relu', name=name + '_out')(x)
return x
def residual_stack(x, filters, blocks, stride1=2, name=None):
x = residual_block(x, filters, stride=stride1, name=name + '_block1')
for i in range(2, blocks + 1):
x = residual_block(x, filters, conv_shortcut=False, name=name + '_block' + str(i))
return x
def ResNet50():
inputs = tf.keras.layers.Input(shape=INPUT_SHAPE)
x = tf.keras.layers.ZeroPadding2D(padding=((3, 3), (3, 3)), name='conv1_pad')(inputs)
x = tf.keras.layers.Conv2D(64, 7, strides=2, use_bias=True, name='conv1_conv', kernel_initializer='he_uniform', bias_initializer='he_uniform', kernel_regularizer=tf.keras.regularizers.l2(0.0001))(x)
x = tf.keras.layers.BatchNormalization(axis=3, name='conv1_bn')(x)
x = tf.keras.layers.Activation('relu', name='conv1_relu')(x)
x = tf.keras.layers.ZeroPadding2D(padding=((1, 1), (1, 1)), name='pool1_pad')(x)
x = tf.keras.layers.MaxPooling2D(3, strides=2, name='pool1_pool')(x)
x = residual_stack(x, 64, 3, stride1=1, name='conv2')
x = residual_stack(x, 128, 4, name='conv3')
x = residual_stack(x, 256, 6, name='conv4')
x = residual_stack(x, 512, 3, name='conv5')
x = tf.keras.layers.GlobalAveragePooling2D(name='avg_pool')(x)
outputs = tf.keras.layers.Dense(n_classes, activation='softmax', name='predictions')(x)
model = tf.keras.Model(inputs=inputs, outputs=outputs)
return model
@tf.function
def train(model, images, labels):
with tf.GradientTape() as tape:
y_pred = model(images, training=True)
loss = tf.reduce_mean(cost_fn(labels, y_pred))
grads = tape.gradient(loss, model.trainable_variables)
optimizer.apply_gradients(grads_and_vars=zip(grads, model.trainable_variables))
train_acc.update_state(labels, y_pred)
train_loss.update_state(labels, y_pred)
@tf.function
def validation(model, images, labels):
y_pred = model(images, training=False)
loss = tf.reduce_mean(cost_fn(labels, y_pred))
val_acc.update_state(labels, y_pred)
val_loss.update_state(labels, y_pred)
def lrfn():
if epoch < LR_RAMPUP_EPOCHS:
lr = (LR_MAX - LR_START) / LR_RAMPUP_EPOCHS * epoch + LR_START
elif epoch < LR_RAMPUP_EPOCHS + LR_SUSTAIN_EPOCHS:
lr = LR_MAX
else:
lr = (LR_MAX - LR_MIN) * LR_EXP_DECAY**(epoch - LR_RAMPUP_EPOCHS - LR_SUSTAIN_EPOCHS) + LR_MIN
return lr
def tf_data_visualize(augmentation_element, name):
row, col, idx = 5, 4, 0
row = min(row, BATCH_SIZE // col)
for (image, label) in augmentation_element:
print(image.shape, label.shape)
image = image / 255.0
plt.figure(figsize=(15, int(15 * row / col)))
for j in range(row * col):
plt.subplot(row, col, j + 1)
plt.axis('off')
plt.imshow(image[j, ])
# plt.savefig(f'{SAVED_PATH}/{LOG_TIME}/{name}_{idx}.jpg')
plt.show()
idx += 1
if idx == 3:
break
if __name__ == "__main__":
# hyper parameters
AUTOTUNE = tf.data.experimental.AUTOTUNE
IMG_SIZE = 224
INPUT_SHAPE = (IMG_SIZE, IMG_SIZE, 3)
BATCH_SIZE = 32
EPOCHS = 1000
# learning rate scheduler
LR_START = 0.001
LR_MAX = 0.005 * strategy.num_replicas_in_sync
LR_MIN = 0.001
LR_RAMPUP_EPOCHS = 5
LR_SUSTAIN_EPOCHS = 0
LR_EXP_DECAY = .8
# early stopping
PATIENCE = 3
EARLY_STOPPING = True
minimum_loss = float(2147000000)
total_images, total_labels, CLASSES = get_dataset('/home/v100/tf_workspace/datasets/natural_images/natural_images')
n_classes = len(CLASSES)
train_images, valid_images, train_labels, valid_labels = train_test_split(total_images, total_labels, test_size=.3, shuffle=True, random_state=777)
TRAIN_STEPS_PER_EPOCH = int(tf.math.ceil(len(train_images) / BATCH_SIZE).numpy())
VALID_STEP_PER_EPOCH = int(tf.math.ceil(len(valid_images) / BATCH_SIZE).numpy())
cost_fn = tf.keras.losses.CategoricalCrossentropy()
# optimizer = tf.keras.optimizers.Adam(learning_rate=lrfn)
optimizer = tf.keras.optimizers.Adam(learning_rate=0.00001)
inputs = tf.keras.Input(shape=(INPUT_SHAPE))
model = ResNet50()
model(inputs=inputs)
model.summary()
# tf.keras.utils.plot_model(model, show_shapes=True)
train_acc = tf | aug_img = tf.numpy_function(func=aug_fn, inp=[image], Tout=tf.float32)
return aug_img, label | identifier_body |
ResNet50_augmentation.py |
def get_dataset(ds_path):
ds_path = pathlib.Path(ds_path)
images = list(ds_path.glob('*/*.jpg'))
images = [str(path) for path in images]
total_images = len(images)
labels = sorted(item.name for item in ds_path.glob('*/') if item.is_dir())
classes = labels
labels = dict((name, index) for index, name in enumerate(labels))
labels = [labels[pathlib.Path(path).parent.name] for path in images]
labels = tf.keras.utils.to_categorical(labels, num_classes=len(classes), dtype='float32')
return images, labels, classes
def aug_fn(image):
data = {"image":image}
aug_data = transforms(**data)
aug_img = aug_data["image"]
aug_img = tf.cast(aug_img, tf.float32) / 255.0
aug_img = tf.image.per_image_standardization(aug_img)
# aug_img = tf.keras.applications.resnet.preprocess_input(aug_img)
return aug_img
def process_data(image, label):
aug_img = tf.numpy_function(func=aug_fn, inp=[image], Tout=tf.float32)
return aug_img, label
def make_tf_data(images, labels, augmentation):
images = tf.data.Dataset.from_tensor_slices(images)
images = images.map(preprocess_image, num_parallel_calls=AUTOTUNE)
labels = tf.data.Dataset.from_tensor_slices(labels)
dataset = tf.data.Dataset.zip((images, labels))
dataset = dataset.repeat()
if augmentation:
dataset = dataset.map(partial(process_data), num_parallel_calls=AUTOTUNE)
dataset = dataset.batch(BATCH_SIZE)
dataset = dataset.prefetch(AUTOTUNE)
return dataset
def residual_block(x, filters, kernel_size=3, stride=1, conv_shortcut=True, name=None):
if conv_shortcut:
shortcut = tf.keras.layers.Conv2D(4 * filters, 1, strides=stride, name=name+'_0_conv', kernel_initializer='he_uniform', bias_initializer='he_uniform', kernel_regularizer=tf.keras.regularizers.l2(0.0001))(x)
shortcut = tf.keras.layers.BatchNormalization(axis=3, name=name+'_0_bn')(shortcut)
else:
shortcut = x
x = tf.keras.layers.Conv2D(filters, 1, strides=stride, name=name + '_1_conv', kernel_initializer='he_uniform', bias_initializer='he_uniform', kernel_regularizer=tf.keras.regularizers.l2(0.0001))(x)
x = tf.keras.layers.BatchNormalization(axis=3, name=name + '_1_bn')(x)
x = tf.keras.layers.Activation('relu', name=name + '_1_relu')(x)
x = tf.keras.layers.Conv2D(filters, kernel_size, padding='SAME', name=name + '_2_conv', kernel_initializer='he_uniform', bias_initializer='he_uniform', kernel_regularizer=tf.keras.regularizers.l2(0.0001))(x)
x = tf.keras.layers.BatchNormalization(axis=3, name=name + '_2_bn')(x)
x = tf.keras.layers.Activation('relu', name=name + '_2_relu')(x)
x = tf.keras.layers.Conv2D(4 * filters, 1, name=name + '_3_conv', kernel_initializer='he_uniform', bias_initializer='he_uniform', kernel_regularizer=tf.keras.regularizers.l2(0.0001))(x)
x = tf.keras.layers.BatchNormalization(axis=3, name=name + '_3_bn')(x)
x = tf.keras.layers.Add(name=name + '_add')([shortcut, x])
x = tf.keras.layers.Activation('relu', name=name + '_out')(x)
return x
def residual_stack(x, filters, blocks, stride1=2, name=None):
x = residual_block(x, filters, stride=stride1, name=name + '_block1')
for i in range(2, blocks + 1):
x = residual_block(x, filters, conv_shortcut=False, name=name + '_block' + str(i))
return x
def ResNet50():
inputs = tf.keras.layers.Input(shape=INPUT_SHAPE)
x = tf.keras.layers.ZeroPadding2D(padding=((3, 3), (3, 3)), name='conv1_pad')(inputs)
x = tf.keras.layers.Conv2D(64, 7, strides=2, use_bias=True, name='conv1_conv', kernel_initializer='he_uniform', bias_initializer='he_uniform', kernel_regularizer=tf.keras.regularizers.l2(0.0001))(x)
x = tf.keras.layers.BatchNormalization(axis=3, name='conv1_bn')(x)
x = tf.keras.layers.Activation('relu', name='conv1_relu')(x)
x = tf.keras.layers.ZeroPadding2D(padding=((1, 1), (1, 1)), name='pool1_pad')(x)
x = tf.keras.layers.MaxPooling2D(3, strides=2, name='pool1_pool')(x)
x = residual_stack(x, 64, 3, stride1=1, name='conv2')
x = residual_stack(x, 128, 4, name='conv3')
x = residual_stack(x, 256, 6, name='conv4')
x = residual_stack(x, 512, 3, name='conv5')
x = tf.keras.layers.GlobalAveragePooling2D(name='avg_pool')(x)
outputs = tf.keras.layers.Dense(n_classes, activation='softmax', name='predictions')(x)
model = tf.keras.Model(inputs=inputs, outputs=outputs)
return model
@tf.function
def train(model, images, labels):
with tf.GradientTape() as tape:
y_pred = model(images, training=True)
loss = tf.reduce_mean(cost_fn(labels, y_pred))
grads = tape.gradient(loss, model.trainable_variables)
optimizer.apply_gradients(grads_and_vars=zip(grads, model.trainable_variables))
train_acc.update_state(labels, y_pred)
train_loss.update_state(labels, y_pred)
@tf.function
def validation(model, images, labels):
y_pred = model(images, training=False)
loss = tf.reduce_mean(cost_fn(labels, y_pred))
val_acc.update_state(labels, y_pred)
val_loss.update_state(labels, y_pred)
def | ():
if epoch < LR_RAMPUP_EPOCHS:
lr = (LR_MAX - LR_START) / LR_RAMPUP_EPOCHS * epoch + LR_START
elif epoch < LR_RAMPUP_EPOCHS + LR_SUSTAIN_EPOCHS:
lr = LR_MAX
else:
lr = (LR_MAX - LR_MIN) * LR_EXP_DECAY**(epoch - LR_RAMPUP_EPOCHS - LR_SUSTAIN_EPOCHS) + LR_MIN
return lr
def tf_data_visualize(augmentation_element, name):
row, col, idx = 5, 4, 0
row = min(row, BATCH_SIZE // col)
for (image, label) in augmentation_element:
print(image.shape, label.shape)
image = image / 255.0
plt.figure(figsize=(15, int(15 * row / col)))
for j in range(row * col):
plt.subplot(row, col, j + 1)
plt.axis('off')
plt.imshow(image[j, ])
# plt.savefig(f'{SAVED_PATH}/{LOG_TIME}/{name}_{idx}.jpg')
plt.show()
idx += 1
if idx == 3:
break
if __name__ == "__main__":
# hyper parameters
AUTOTUNE = tf.data.experimental.AUTOTUNE
IMG_SIZE = 224
INPUT_SHAPE = (IMG_SIZE, IMG_SIZE, 3)
BATCH_SIZE = 32
EPOCHS = 1000
# learning rate scheduler
LR_START = 0.001
LR_MAX = 0.005 * strategy.num_replicas_in_sync
LR_MIN = 0.001
LR_RAMPUP_EPOCHS = 5
LR_SUSTAIN_EPOCHS = 0
LR_EXP_DECAY = .8
# early stopping
PATIENCE = 3
EARLY_STOPPING = True
minimum_loss = float(2147000000)
total_images, total_labels, CLASSES = get_dataset('/home/v100/tf_workspace/datasets/natural_images/natural_images')
n_classes = len(CLASSES)
train_images, valid_images, train_labels, valid_labels = train_test_split(total_images, total_labels, test_size=.3, shuffle=True, random_state=777)
TRAIN_STEPS_PER_EPOCH = int(tf.math.ceil(len(train_images) / BATCH_SIZE).numpy())
VALID_STEP_PER_EPOCH = int(tf.math.ceil(len(valid_images) / BATCH_SIZE).numpy())
cost_fn = tf.keras.losses.CategoricalCrossentropy()
# optimizer = tf.keras.optimizers.Adam(learning_rate=lrfn)
optimizer = tf.keras.optimizers.Adam(learning_rate=0.00001)
inputs = tf.keras.Input(shape=(INPUT_SHAPE))
model = ResNet50()
model(inputs=inputs)
model.summary()
# tf.keras.utils.plot_model(model, show_shapes=True)
train_acc = | lrfn | identifier_name |
ResNet50_augmentation.py | =3, name=name + '_2_bn')(x)
x = tf.keras.layers.Activation('relu', name=name + '_2_relu')(x)
x = tf.keras.layers.Conv2D(4 * filters, 1, name=name + '_3_conv', kernel_initializer='he_uniform', bias_initializer='he_uniform', kernel_regularizer=tf.keras.regularizers.l2(0.0001))(x)
x = tf.keras.layers.BatchNormalization(axis=3, name=name + '_3_bn')(x)
x = tf.keras.layers.Add(name=name + '_add')([shortcut, x])
x = tf.keras.layers.Activation('relu', name=name + '_out')(x)
return x
def residual_stack(x, filters, blocks, stride1=2, name=None):
x = residual_block(x, filters, stride=stride1, name=name + '_block1')
for i in range(2, blocks + 1):
x = residual_block(x, filters, conv_shortcut=False, name=name + '_block' + str(i))
return x
def ResNet50():
inputs = tf.keras.layers.Input(shape=INPUT_SHAPE)
x = tf.keras.layers.ZeroPadding2D(padding=((3, 3), (3, 3)), name='conv1_pad')(inputs)
x = tf.keras.layers.Conv2D(64, 7, strides=2, use_bias=True, name='conv1_conv', kernel_initializer='he_uniform', bias_initializer='he_uniform', kernel_regularizer=tf.keras.regularizers.l2(0.0001))(x)
x = tf.keras.layers.BatchNormalization(axis=3, name='conv1_bn')(x)
x = tf.keras.layers.Activation('relu', name='conv1_relu')(x)
x = tf.keras.layers.ZeroPadding2D(padding=((1, 1), (1, 1)), name='pool1_pad')(x)
x = tf.keras.layers.MaxPooling2D(3, strides=2, name='pool1_pool')(x)
x = residual_stack(x, 64, 3, stride1=1, name='conv2')
x = residual_stack(x, 128, 4, name='conv3')
x = residual_stack(x, 256, 6, name='conv4')
x = residual_stack(x, 512, 3, name='conv5')
x = tf.keras.layers.GlobalAveragePooling2D(name='avg_pool')(x)
outputs = tf.keras.layers.Dense(n_classes, activation='softmax', name='predictions')(x)
model = tf.keras.Model(inputs=inputs, outputs=outputs)
return model
@tf.function
def train(model, images, labels):
with tf.GradientTape() as tape:
y_pred = model(images, training=True)
loss = tf.reduce_mean(cost_fn(labels, y_pred))
grads = tape.gradient(loss, model.trainable_variables)
optimizer.apply_gradients(grads_and_vars=zip(grads, model.trainable_variables))
train_acc.update_state(labels, y_pred)
train_loss.update_state(labels, y_pred)
@tf.function
def validation(model, images, labels):
y_pred = model(images, training=False)
loss = tf.reduce_mean(cost_fn(labels, y_pred))
val_acc.update_state(labels, y_pred)
val_loss.update_state(labels, y_pred)
def lrfn():
if epoch < LR_RAMPUP_EPOCHS:
lr = (LR_MAX - LR_START) / LR_RAMPUP_EPOCHS * epoch + LR_START
elif epoch < LR_RAMPUP_EPOCHS + LR_SUSTAIN_EPOCHS:
lr = LR_MAX
else:
lr = (LR_MAX - LR_MIN) * LR_EXP_DECAY**(epoch - LR_RAMPUP_EPOCHS - LR_SUSTAIN_EPOCHS) + LR_MIN
return lr
def tf_data_visualize(augmentation_element, name):
row, col, idx = 5, 4, 0
row = min(row, BATCH_SIZE // col)
for (image, label) in augmentation_element:
print(image.shape, label.shape)
image = image / 255.0
plt.figure(figsize=(15, int(15 * row / col)))
for j in range(row * col):
plt.subplot(row, col, j + 1)
plt.axis('off')
plt.imshow(image[j, ])
# plt.savefig(f'{SAVED_PATH}/{LOG_TIME}/{name}_{idx}.jpg')
plt.show()
idx += 1
if idx == 3:
break
if __name__ == "__main__":
# hyper parameters
AUTOTUNE = tf.data.experimental.AUTOTUNE
IMG_SIZE = 224
INPUT_SHAPE = (IMG_SIZE, IMG_SIZE, 3)
BATCH_SIZE = 32
EPOCHS = 1000
# learning rate scheduler
LR_START = 0.001
LR_MAX = 0.005 * strategy.num_replicas_in_sync
LR_MIN = 0.001
LR_RAMPUP_EPOCHS = 5
LR_SUSTAIN_EPOCHS = 0
LR_EXP_DECAY = .8
# early stopping
PATIENCE = 3
EARLY_STOPPING = True
minimum_loss = float(2147000000)
total_images, total_labels, CLASSES = get_dataset('/home/v100/tf_workspace/datasets/natural_images/natural_images')
n_classes = len(CLASSES)
train_images, valid_images, train_labels, valid_labels = train_test_split(total_images, total_labels, test_size=.3, shuffle=True, random_state=777)
TRAIN_STEPS_PER_EPOCH = int(tf.math.ceil(len(train_images) / BATCH_SIZE).numpy())
VALID_STEP_PER_EPOCH = int(tf.math.ceil(len(valid_images) / BATCH_SIZE).numpy())
cost_fn = tf.keras.losses.CategoricalCrossentropy()
# optimizer = tf.keras.optimizers.Adam(learning_rate=lrfn)
optimizer = tf.keras.optimizers.Adam(learning_rate=0.00001)
inputs = tf.keras.Input(shape=(INPUT_SHAPE))
model = ResNet50()
model(inputs=inputs)
model.summary()
# tf.keras.utils.plot_model(model, show_shapes=True)
train_acc = tf.metrics.CategoricalAccuracy()
train_loss = tf.metrics.CategoricalCrossentropy()
val_acc = tf.metrics.CategoricalAccuracy()
val_loss = tf.metrics.CategoricalCrossentropy()
transforms = A.Compose([
# A.Resize(IMG_SIZE, IMG_SIZE, 3, p=1),
A.OneOf([
A.HorizontalFlip(p=0.6),
A.VerticalFlip(p=0.6),
], p=0.7),
# A.Cutout(num_holes=15, max_h_size=4, max_w_size=4, fill_value=[0, 0, 0], p=0.7),
A.OneOf([
A.RandomRotate90(p=0.6),
A.ShiftScaleRotate(p=0.6, border_mode=1)
], p=0.7),
# A.RandomBrightness(limit=0.1, p=0.5),
# A.HueSaturationValue(hue_shift_limit=20, sat_shift_limit=30, val_shift_limit=20, p=0.5),
# A.RandomContrast(limit=0.2, p=0.5),
])
# tf_data_visualize(make_tf_data(train_images, train_labels, True), 'train')
stateful_matrices = ['train_acc', 'train_loss', 'valid_acc', 'valid_loss']
print()
print('Learning started. It takes sometime.')
for epoch in range(EPOCHS):
print("Current Learning Rate : ", optimizer._decayed_lr('float32').numpy())
tf.print("Epoch {}/{}".format(epoch + 1, EPOCHS))
prog_bar = tf.keras.utils.Progbar(target=TRAIN_STEPS_PER_EPOCH, stateful_metrics=stateful_matrices)
train_acc.reset_states()
train_loss.reset_states()
val_acc.reset_states()
val_loss.reset_states()
for idx, (images, labels) in enumerate(make_tf_data(train_images, train_labels, True)):
train(model, images, labels)
values=[('train_loss', train_loss.result().numpy()), ('train_acc', train_acc.result().numpy())]
prog_bar.update(idx, values=values)
if idx+1 >= TRAIN_STEPS_PER_EPOCH:
break
for idx, (images, labels) in enumerate(make_tf_data(valid_images, valid_labels, True)):
validation(model, images, labels)
if idx+1 >= VALID_STEP_PER_EPOCH:
break
values = [('train_loss', train_loss.result().numpy()), ('train_acc', train_acc.result().numpy()), ('valid_loss', val_loss.result().numpy()), ('valid_acc', val_acc.result().numpy())]
prog_bar.update(TRAIN_STEPS_PER_EPOCH, values=values, finalize=True)
if EARLY_STOPPING:
tmp_loss = (val_loss.result().numpy())
if tmp_loss < minimum_loss:
minimum_loss = tmp_loss
PATIENCE = 3
else:
| PATIENCE -= 1
if PATIENCE == 0:
break | conditional_block |
|
sriov.go | {
return nil, "", fmt.Errorf("failed to load netconf: %v", err)
}
if n.VlanId < 0 || n.VlanId > 4094 {
return nil, "", fmt.Errorf(`invalid VLAN ID %d (must be between 0 and 4095 inclusive)`, n.VlanId)
}
return n, n.CNIVersion, nil
}
func | (conf *NetConf, ifName string, netns ns.NetNS) (*current.Interface, error) {
vf := ¤t.Interface{}
// 申请一个可用的Virtual Function
m, vfIdx, vfDevName, err := allocFreeVF()
if err != nil {
return nil, err
}
fmt.Fprintf(os.Stderr, "***********CNI SR-IOV 成功申请%v网卡的第%v个VF, 名称为: %v\n", m.Attrs().Name, vfIdx, vfDevName)
vfDev, err := netlink.LinkByName(vfDevName)
if err != nil {
return nil, fmt.Errorf("failed to lookup vf device %q: %v", vfDevName, err)
}
if conf.MTU <= 0 {
conf.MTU = m.Attrs().MTU
}
if err = netlink.LinkSetVfHardwareAddr(m, vfIdx, vfDev.Attrs().HardwareAddr); err != nil {
return nil, fmt.Errorf("failed to set vf %d macaddress: %v", vfIdx, err)
}
if err = netlink.LinkSetVfVlan(m, vfIdx, conf.VlanId); err != nil {
return nil, fmt.Errorf("failed to set vf %d vlan: %v", vfIdx, err)
}
if err = netlink.LinkSetUp(vfDev); err != nil {
return nil, fmt.Errorf("failed to setup vf %d device: %v", vfIdx, err)
}
// move VF device to ns
if err = netlink.LinkSetNsFd(vfDev, int(netns.Fd())); err != nil {
return nil, fmt.Errorf("failed to move vf %d to netns: %v", vfIdx, err)
}
err = netns.Do(func(_ ns.NetNS) error {
err := ip.RenameLink(vfDevName, ifName)
if err != nil {
return fmt.Errorf("failed to rename vlan to %q: %v", ifName, err)
}
vf.Name = ifName
// Re-fetch interface to get all properties/attributes
contVF, err := netlink.LinkByName(vf.Name)
if err != nil {
return fmt.Errorf("failed to refetch vlan %q: %v", vf.Name, err)
}
vf.Mac = contVF.Attrs().HardwareAddr.String()
vf.Sandbox = netns.Path()
return nil
})
if err != nil {
return nil, err
}
return vf, nil
}
func allocFreeVF() (netlink.Link, int, string, error) {
vfIdx := -1
devName := ""
// 获取机器可用物理网卡(PF)列表
links, err := netlink.LinkList()
if err != nil {
return nil, -1, "", fmt.Errorf("获取可用物理网卡失败: %v", err)
}
for _, link := range links {
if link.Type() == "device" && link.Attrs().OperState == netlink.OperUp {
master := link.Attrs().Name
sriovFile := fmt.Sprintf("/sys/class/net/%s/device/sriov_numvfs", master)
if _, err := os.Lstat(sriovFile); err != nil {
return nil, -1, "", fmt.Errorf("failed to open the sriov_numfs of device %q: %v", master, err)
}
data, err := ioutil.ReadFile(sriovFile)
if err != nil {
return nil, -1, "", fmt.Errorf("failed to read the sriov_numfs of device %q: %v", master, err)
}
if len(data) == 0 {
return nil, -1, "", fmt.Errorf("no data in the file %q", sriovFile)
}
sriovNumfs := strings.TrimSpace(string(data))
vfTotal, err := strconv.Atoi(sriovNumfs)
if err != nil {
return nil, -1, "", fmt.Errorf("failed to convert sriov_numfs(byte value) to int of device %q: %v", master, err)
}
if vfTotal <= 0 {
return nil, -1, "", fmt.Errorf("no virtual function in the device %q: %v", master)
}
for vf := 0; vf < vfTotal; vf++ {
devName, err = getVFDeviceName(master, vf)
// got a free vf
if err == nil {
vfIdx = vf
break
}
}
if vfIdx == -1 {
return nil, -1, "", fmt.Errorf("can not get a free virtual function in directory %s", master)
}
return link, vfIdx, devName, nil
}
}
return nil, vfIdx, devName, fmt.Errorf("该主机无可用物理网卡")
}
func getVFDeviceName(master string, vf int) (string, error) {
vfDir := fmt.Sprintf("/sys/class/net/%s/device/virtfn%d/net", master, vf)
if _, err := os.Lstat(vfDir); err != nil {
return "", fmt.Errorf("failed to open the virtfn%d dir of the device %q: %v", vf, master, err)
}
infos, err := ioutil.ReadDir(vfDir)
if err != nil {
return "", fmt.Errorf("failed to read the virtfn%d dir of the device %q: %v", vf, master, err)
}
if len(infos) != 1 {
return "", fmt.Errorf("no network device in directory %s", vfDir)
}
return infos[0].Name(), nil
}
func releaseVF(conf *NetConf, ifName string, netns ns.NetNS) error {
initns, err := ns.GetCurrentNS()
fmt.Fprintf(os.Stderr, "***********CNI SR-IOV releaseVF initns = %v\n", initns)
if err != nil {
return fmt.Errorf("failed to get init netns: %v", err)
}
// for IPAM in cmdDel
return netns.Do(func(_ ns.NetNS) error {
// get VF device
vfDev, err := netlink.LinkByName(ifName)
if err != nil {
return fmt.Errorf("failed to lookup device %s: %v", ifName, err)
}
// device name in init netns
index := vfDev.Attrs().Index
devName := fmt.Sprintf("dev%d", index)
fmt.Fprintf(os.Stderr, "***********CNI SR-IOV releaseVF index = %v devName = %v vfDev = %v\n", index, devName, vfDev)
// shutdown VF device
if err = netlink.LinkSetDown(vfDev); err != nil {
return fmt.Errorf("failed to down device: %v", err)
}
// rename VF device
err = ip.RenameLink(ifName, devName)
if err != nil {
return fmt.Errorf("failed to rename device %s to %s: %v", ifName, devName, err)
}
// move VF device to init netns
if err = netlink.LinkSetNsFd(vfDev, int(initns.Fd())); err != nil {
return fmt.Errorf("failed to move device %s to init netns: %v", ifName, err)
}
return nil
})
}
func cmdAdd(args *skel.CmdArgs) error {
fmt.Fprintf(os.Stderr, "***********CNI SR-IOV cmdAdd args.ContainerID = %v\n", args.ContainerID)
fmt.Fprintf(os.Stderr, "***********CNI SR-IOV cmdAdd args.Netns = %v\n", args.Netns)
fmt.Fprintf(os.Stderr, "***********CNI SR-IOV cmdAdd args.IfName = %v\n", args.IfName)
fmt.Fprintf(os.Stderr, "***********CNI SR-IOV cmdAdd args.Args = %v\n", args.Args)
fmt.Fprintf(os.Stderr, "***********CNI SR-IOV cmdAdd args.Path = %v\n", args.Path)
fmt.Fprintf(os.Stderr, "***********CNI SR-IOV cmdAdd args.StdinData = %v\n", string(args.StdinData))
n, cniVersion, err := loadConf(args.StdinData)
if err != nil {
return err
}
netns, err := ns.GetNS(args.Netns)
if err != nil {
return fmt.Errorf("failed to open netns %q: %v", args.Netns, err)
}
defer netns.Close()
vfInterface, err := setupVF(n, args.IfName, netns)
if err != nil {
return err
}
// run the IPAM plugin and get back the config to apply
r, err := ipam.ExecAdd(n.IPAM.Type, args.Stdin | setupVF | identifier_name |
sriov.go | ().Name, vfIdx, vfDevName)
vfDev, err := netlink.LinkByName(vfDevName)
if err != nil {
return nil, fmt.Errorf("failed to lookup vf device %q: %v", vfDevName, err)
}
if conf.MTU <= 0 {
conf.MTU = m.Attrs().MTU
}
if err = netlink.LinkSetVfHardwareAddr(m, vfIdx, vfDev.Attrs().HardwareAddr); err != nil {
return nil, fmt.Errorf("failed to set vf %d macaddress: %v", vfIdx, err)
}
if err = netlink.LinkSetVfVlan(m, vfIdx, conf.VlanId); err != nil {
return nil, fmt.Errorf("failed to set vf %d vlan: %v", vfIdx, err)
}
if err = netlink.LinkSetUp(vfDev); err != nil {
return nil, fmt.Errorf("failed to setup vf %d device: %v", vfIdx, err)
}
// move VF device to ns
if err = netlink.LinkSetNsFd(vfDev, int(netns.Fd())); err != nil {
return nil, fmt.Errorf("failed to move vf %d to netns: %v", vfIdx, err)
}
err = netns.Do(func(_ ns.NetNS) error {
err := ip.RenameLink(vfDevName, ifName)
if err != nil {
return fmt.Errorf("failed to rename vlan to %q: %v", ifName, err)
}
vf.Name = ifName
// Re-fetch interface to get all properties/attributes
contVF, err := netlink.LinkByName(vf.Name)
if err != nil {
return fmt.Errorf("failed to refetch vlan %q: %v", vf.Name, err)
}
vf.Mac = contVF.Attrs().HardwareAddr.String()
vf.Sandbox = netns.Path()
return nil
})
if err != nil {
return nil, err
}
return vf, nil
}
func allocFreeVF() (netlink.Link, int, string, error) {
vfIdx := -1
devName := ""
// 获取机器可用物理网卡(PF)列表
links, err := netlink.LinkList()
if err != nil {
return nil, -1, "", fmt.Errorf("获取可用物理网卡失败: %v", err)
}
for _, link := range links {
if link.Type() == "device" && link.Attrs().OperState == netlink.OperUp {
master := link.Attrs().Name
sriovFile := fmt.Sprintf("/sys/class/net/%s/device/sriov_numvfs", master)
if _, err := os.Lstat(sriovFile); err != nil {
return nil, -1, "", fmt.Errorf("failed to open the sriov_numfs of device %q: %v", master, err)
}
data, err := ioutil.ReadFile(sriovFile)
if err != nil {
return nil, -1, "", fmt.Errorf("failed to read the sriov_numfs of device %q: %v", master, err)
}
if len(data) == 0 {
return nil, -1, "", fmt.Errorf("no data in the file %q", sriovFile)
}
sriovNumfs := strings.TrimSpace(string(data))
vfTotal, err := strconv.Atoi(sriovNumfs)
if err != nil {
return nil, -1, "", fmt.Errorf("failed to convert sriov_numfs(byte value) to int of device %q: %v", master, err)
}
if vfTotal <= 0 {
return nil, -1, "", fmt.Errorf("no virtual function in the device %q: %v", master)
}
for vf := 0; vf < vfTotal; vf++ {
devName, err = getVFDeviceName(master, vf)
// got a free vf
if err == nil {
vfIdx = vf
break
}
}
if vfIdx == -1 {
return nil, -1, "", fmt.Errorf("can not get a free virtual function in directory %s", master)
}
return link, vfIdx, devName, nil
}
}
return nil, vfIdx, devName, fmt.Errorf("该主机无可用物理网卡")
}
func getVFDeviceName(master string, vf int) (string, error) {
vfDir := fmt.Sprintf("/sys/class/net/%s/device/virtfn%d/net", master, vf)
if _, err := os.Lstat(vfDir); err != nil {
return "", fmt.Errorf("failed to open the virtfn%d dir of the device %q: %v", vf, master, err)
}
infos, err := ioutil.ReadDir(vfDir)
if err != nil {
return "", fmt.Errorf("failed to read the virtfn%d dir of the device %q: %v", vf, master, err)
}
if len(infos) != 1 {
return "", fmt.Errorf("no network device in directory %s", vfDir)
}
return infos[0].Name(), nil
}
func releaseVF(conf *NetConf, ifName string, netns ns.NetNS) error {
initns, err := ns.GetCurrentNS()
fmt.Fprintf(os.Stderr, "***********CNI SR-IOV releaseVF initns = %v\n", initns)
if err != nil {
return fmt.Errorf("failed to get init netns: %v", err)
}
// for IPAM in cmdDel
return netns.Do(func(_ ns.NetNS) error {
// get VF device
vfDev, err := netlink.LinkByName(ifName)
if err != nil {
return fmt.Errorf("failed to lookup device %s: %v", ifName, err)
}
// device name in init netns
index := vfDev.Attrs().Index
devName := fmt.Sprintf("dev%d", index)
fmt.Fprintf(os.Stderr, "***********CNI SR-IOV releaseVF index = %v devName = %v vfDev = %v\n", index, devName, vfDev)
// shutdown VF device
if err = netlink.LinkSetDown(vfDev); err != nil {
return fmt.Errorf("failed to down device: %v", err)
}
// rename VF device
err = ip.RenameLink(ifName, devName)
if err != nil {
return fmt.Errorf("failed to rename device %s to %s: %v", ifName, devName, err)
}
// move VF device to init netns
if err = netlink.LinkSetNsFd(vfDev, int(initns.Fd())); err != nil {
return fmt.Errorf("failed to move device %s to init netns: %v", ifName, err)
}
return nil
})
}
func cmdAdd(args *skel.CmdArgs) error {
fmt.Fprintf(os.Stderr, "***********CNI SR-IOV cmdAdd args.ContainerID = %v\n", args.ContainerID)
fmt.Fprintf(os.Stderr, "***********CNI SR-IOV cmdAdd args.Netns = %v\n", args.Netns)
fmt.Fprintf(os.Stderr, "***********CNI SR-IOV cmdAdd args.IfName = %v\n", args.IfName)
fmt.Fprintf(os.Stderr, "***********CNI SR-IOV cmdAdd args.Args = %v\n", args.Args)
fmt.Fprintf(os.Stderr, "***********CNI SR-IOV cmdAdd args.Path = %v\n", args.Path)
fmt.Fprintf(os.Stderr, "***********CNI SR-IOV cmdAdd args.StdinData = %v\n", string(args.StdinData))
n, cniVersion, err := loadConf(args.StdinData)
if err != nil {
return err
}
netns, err := ns.GetNS(args.Netns)
if err != nil {
return fmt.Errorf("failed to open netns %q: %v", args.Netns, err)
}
defer netns.Close()
vfInterface, err := setupVF(n, args.IfName, netns)
if err != nil {
return err
}
// run the IPAM plugin and get back the config to apply
r, err := ipam.ExecAdd(n.IPAM.Type, args.StdinData)
if err != nil {
return err
}
// Convert whatever the IPAM result was into the current Result type
result, err := current.NewResultFromResult(r)
if err != nil {
return err
}
if len(result.IPs) == 0 {
return errors.New("IPAM plugin returned missing IP config")
}
for _, ipc := range result.IPs {
// All addresses belong to the vlan interface
ipc.Interface = current.Int(0)
}
result.Interfaces = []*current.Interface{vfInterface}
err = netns.Do(func(_ ns.NetNS) error {
return ipam.ConfigureIface(args.IfName, result)
})
if err != nil {
return err
}
| result.DNS = n.DNS
fmt.Fprintf(os.Stderr, "***********CNI SR-IOV cmdAdd result = %v\n", result)
return types.PrintResult(result, cniVersion) | random_line_split |
|
sriov.go | return nil, "", fmt.Errorf("failed to load netconf: %v", err)
}
if n.VlanId < 0 || n.VlanId > 4094 {
return nil, "", fmt.Errorf(`invalid VLAN ID %d (must be between 0 and 4095 inclusive)`, n.VlanId)
}
return n, n.CNIVersion, nil
}
func setupVF(conf *NetConf, ifName string, netns ns.NetNS) (*current.Interface, error) {
vf := ¤t.Interface{}
// 申请一个可用的Virtual Function
m, vfIdx, vfDevName, err := allocFreeVF()
if err != nil {
return nil, err
}
fmt.Fprintf(os.Stderr, "***********CNI SR-IOV 成功申请%v网卡的第%v个VF, 名称为: %v\n", m.Attrs().Name, vfIdx, vfDevName)
vfDev, err := netlink.LinkByName(vfDevName)
if err != nil {
return nil, fmt.Errorf("failed to lookup vf device %q: %v", vfDevName, err)
}
if conf.MTU <= 0 {
conf.MTU = m.Attrs().MTU
}
if err = netlink.LinkSetVfHardwareAddr(m, vfIdx, vfDev.Attrs().HardwareAddr); err != nil {
return nil, fmt.Errorf("failed to set vf %d macaddress: %v", vfIdx, err)
}
if err = netlink.LinkSetVfVlan(m, vfIdx, conf.VlanId); err != nil {
return nil, fmt.Errorf("failed to set vf %d vlan: %v", vfIdx, err)
}
if err = netlink.LinkSetUp(vfDev); err != nil {
return nil, fmt.Errorf("failed to setup vf %d device: %v", vfIdx, err)
}
// move VF device to ns
if err = netlink.LinkSetNsFd(vfDev, int(netns.Fd())); err != nil {
return nil, fmt.Errorf("failed to move vf %d to netns: %v", vfIdx, err)
}
err = netns.Do(func(_ ns.NetNS) error {
err := ip.RenameLink(vfDevName, ifName)
if err != nil {
return fmt.Errorf("failed to rename vlan to %q: %v", ifName, err)
}
vf.Name = ifName
// Re-fetch interface to get all properties/attributes
contVF, err := netlink.LinkByName(vf.Name)
if err != nil {
return fmt.Errorf("failed to refetch vlan %q: %v", vf.Name, err)
}
vf.Mac = contVF.Attrs().HardwareAddr.String()
vf.Sandbox = netns.Path()
return nil
})
if err != nil {
return nil, err
}
return vf, nil
}
func allocFreeVF() (netlink.Link, int, string, error) {
vfIdx := -1
devName := ""
// 获取机器可用物理网卡(PF)列表
links, err := netlink.LinkList()
if err != nil {
return nil, -1, "", fmt.Errorf("获取可用物理网卡失败: %v", err)
}
for _, link := range links {
if link.Type() == "device" && link.Attrs().OperState == netlink.OperUp {
master := link.Attrs().Name
sriovFile := fmt.Sprintf("/sys/class/net/%s/device/sriov_numvfs", master)
if _, err := os.Lstat(sriovFile); err != nil {
return nil, -1, "", fmt.Errorf("failed to open the sriov_numfs of device %q: %v", master, err)
}
data, err := ioutil.ReadFile(sriovFile)
if err != nil {
return nil, -1, "", fmt.Errorf("failed to read the sriov_numfs of device %q: %v", master, err)
}
if len(data) == 0 {
return nil, -1, "", fmt.Errorf("no data in the file %q", sriovFile)
}
sriovNumfs := strings.TrimSpace(string(data))
vfTotal, err := strconv.Atoi(sriovNumfs)
if err != nil {
return nil, -1, "", fmt.Errorf("failed to convert sriov_numfs(byte value) to int of device %q: %v", master, err)
}
if vfTotal <= 0 {
return nil, -1, "", fmt.Errorf("no virtual function in the device %q: %v", master)
}
for vf := 0; vf < vfTotal; vf++ {
devName, err = getVFDeviceName(master, vf)
// got a free vf
if err == nil {
vfIdx = vf
break
}
}
if vfIdx == -1 {
return nil, -1, "", fmt.Errorf("can not get a free virtual function in directory %s", master)
}
return link, vfIdx, devName, nil
}
}
return nil, vfIdx, devName, fmt.Errorf("该主机无可用物理网卡")
}
func getVFDeviceName(master string, vf int) (string, error) {
vfDir := fmt.Sprintf("/sys/class/net/%s/device/virtfn%d/net", master, vf)
if _, err := os.Lstat(vfDir); err != nil {
return "", fmt.Errorf("failed to open the virtfn%d dir of the device %q: %v", vf, master, err)
}
infos, err := ioutil.ReadDir(vfDir)
if err != nil {
return "", fmt.Errorf("failed to read the virtfn%d dir of the device %q: %v", vf, master, err)
}
if len(infos) != 1 {
return "", fmt.Errorf("no network device in directory %s", vfDir)
}
return infos[0].Name(), nil
}
func releaseVF(conf *NetConf, ifName string, netns ns.NetNS) error {
initns, err := ns.GetCurrentNS()
fmt.Fprintf(os.Stderr, "***********CNI SR-IOV releaseVF initns = %v\n", initns)
if err != nil {
return fmt.Errorf("failed to get init netns: %v", err)
}
// for IPAM in cmdDel
return netns.Do(func(_ ns.NetNS) error {
// get VF device
vfDev, err := netlink.LinkByName(ifName)
if err != nil {
return fmt.Errorf("failed to lookup device %s: %v", ifName, err)
}
// device name in init netns
index := vfDev.Attrs().Index
devName := fmt.Sprintf("dev%d", index)
fmt.Fprintf(os.Stderr, "***********CNI SR-IOV releaseVF index = %v devName = %v vfDev = %v\n", index, devName, vfDev)
// shutdown VF device
if err = netlink.LinkSetDown(vfDev); err != nil {
return fmt.Errorf("failed to down device: %v", err)
}
// rename VF device
err = ip.RenameLink(ifName, devName)
if err != nil {
return fmt.Errorf("failed to rename device %s to %s: %v", ifName, devName, err)
}
// move VF device to init netns
if err = netlink.LinkSetNsFd(vfDev, int(initns.Fd())); err != nil {
return fmt.Errorf("failed to move device %s to init netns: %v", ifName, err)
}
return nil
})
}
func cmdAdd(args *skel.CmdArgs) error {
fmt.Fprintf(os.Stderr, "***********CNI SR-IOV cmdAdd args.ContainerID = %v\n", args.ContainerID)
fmt.Fprintf(os.Stderr, "***********CNI SR-IOV cmdAdd args.Netns = %v\n", args.Netns)
fmt.Fprintf(os.Stderr, "***********CNI SR-IOV cmdAdd args.IfName = %v\n", args.IfName)
fmt.Fprintf(os.Stderr, "***********CNI SR-IOV cmdAdd args.Args = %v\n", args.Args)
fmt.Fprintf(os.Stderr, "***********CNI SR-IOV cmdAdd args.Path = %v\n", args.Path)
fmt.Fprintf(os.Stderr, "***********CNI SR-IOV cmdAdd args.StdinData = %v\n", string(args.StdinData))
n, cniVersion, err := loadConf(args.StdinData)
if err != nil {
return err
}
netns, err := ns.GetNS(args.Netns)
if err != nil {
return fmt.Errorf("failed to open netns %q: %v", args.Netns, err)
}
defer netns.Close()
vfInterface, err := setupVF(n, args.IfName, netns)
if err != nil {
return err
}
// run the IPAM plugin and get back the config to apply
r, err := ipam.ExecAdd(n. | tdin | IPAM.Type, args.S | conditional_block |
sriov.go | {
return nil, "", fmt.Errorf("failed to load netconf: %v", err)
}
if n.VlanId < 0 || n.VlanId > 4094 {
return nil, "", fmt.Errorf(`invalid VLAN ID %d (must be between 0 and 4095 inclusive)`, n.VlanId)
}
return n, n.CNIVersion, nil
}
func setupVF(conf *NetConf, ifName string, netns ns.NetNS) (*current.Interface, error) {
vf := ¤t.Interface{}
// 申请一个可用的Virtual Function
m, vfIdx, vfDevName, err := allocFreeVF()
if err != nil {
return nil, err
}
fmt.Fprintf(os.Stderr, "***********CNI SR-IOV 成功申请%v网卡的第%v个VF, 名称为: %v\n", m.Attrs().Name, vfIdx, vfDevName)
vfDev, err := netlink.LinkByName(vfDevName)
if err != nil {
return nil, fmt.Errorf("failed to lookup vf device %q: %v", vfDevName, err)
}
if conf.MTU <= 0 {
conf.MTU = m.Attrs().MTU
}
if err = netlink.LinkSetVfHardwareAddr(m, vfIdx, vfDev.Attrs().HardwareAddr); err != nil {
return nil, fmt.Errorf("failed to set vf %d macaddress: %v", vfIdx, err)
}
if err = netlink.LinkSetVfVlan(m, vfIdx, conf.VlanId); err != nil {
return nil, fmt.Errorf("failed to set vf %d vlan: %v", vfIdx, err)
}
if err = netlink.LinkSetUp(vfDev); err != nil {
return nil, fmt.Errorf("failed to setup vf %d device: %v", vfIdx, err)
}
// move VF device to ns
if err = netlink.LinkSetNsFd(vfDev, int(netns.Fd())); err != nil {
return nil, fmt.Errorf("failed to move vf %d to netns: %v", vfIdx, err)
}
err = netns.Do(func(_ ns.NetNS) error {
err := ip.RenameLink(vfDevName, ifName)
if err != nil {
return fmt.Errorf("failed to rename vlan to %q: %v", ifName, err)
}
vf.Name = ifName
// Re-fetch interface to get all properties/attributes
contVF, err := netlink.LinkByName(vf.Name)
if err != nil {
return fmt.Errorf("failed to refetch vlan %q: %v", vf.Name, err)
}
vf.Mac = contVF.Attrs().HardwareAddr.String()
vf.Sandbox = netns.Path()
return nil
})
if err != nil {
return nil, err
}
return vf, nil
}
func allocFreeVF() (netlink.Link, int, string, error) {
vfIdx := -1
devName := ""
// 获取机器可用物理网卡(PF)列表
links, err := netlink.LinkList()
if err != nil {
return nil, -1, "", fmt.Errorf("获取可用物理网卡失败: %v", err)
}
for _, link := range links {
if link.Type() == "device" && link.Attrs().OperState == netlink.OperUp {
master := link.Attrs().Name
sriovFile := fmt.Sprintf("/sys/class/net/%s/device/sriov_numvfs", master)
if _, err := os.Lstat(sriovFile); err != nil {
return nil, -1, "", fmt.Errorf("failed to open the sriov_numfs of device %q: %v", master, err)
}
data, err := ioutil.ReadFile(sriovFile)
if err != nil {
return nil, -1, "", fmt.Errorf("failed to read the sriov_numfs of device %q: %v", master, err)
}
if len(data) == 0 {
return nil, -1, "", fmt.Errorf("no data in the file %q", sriovFile)
}
sriovNumfs := strings.TrimSpace(string(data))
vfTotal, err := strconv.Atoi(sriovNumfs)
if err != nil {
return nil, -1, "", fmt.Errorf("failed to convert sriov_numfs(byte value) to int of device %q: %v", master, err)
}
if vfTotal <= 0 {
return nil, -1, "", fmt.Errorf("no virtual function in the device %q: %v", master)
}
for vf := 0; vf < vfTotal; vf++ {
devName, err = getVFDeviceName(master, vf)
// got a free vf
if err == nil {
vfIdx = vf
break
}
}
if vfIdx == -1 {
return nil, -1, "", fmt.Errorf("can not get a free virtual function in directory %s", master)
}
return link, vfIdx, devName, nil
}
}
return nil, vfIdx, devName, fmt.Errorf("该主机无可用物理网卡")
}
func getVFDeviceName(master string, vf int) (string, error) {
vfDir := fmt.Sprintf("/sys/class/net/%s/device/virtfn%d/net", master, vf)
if _, err := os.Lstat(vf | ()
fmt.Fprintf(os.Stderr, "***********CNI SR-IOV releaseVF initns = %v\n", initns)
if err != nil {
return fmt.Errorf("failed to get init netns: %v", err)
}
// for IPAM in cmdDel
return netns.Do(func(_ ns.NetNS) error {
// get VF device
vfDev, err := netlink.LinkByName(ifName)
if err != nil {
return fmt.Errorf("failed to lookup device %s: %v", ifName, err)
}
// device name in init netns
index := vfDev.Attrs().Index
devName := fmt.Sprintf("dev%d", index)
fmt.Fprintf(os.Stderr, "***********CNI SR-IOV releaseVF index = %v devName = %v vfDev = %v\n", index, devName, vfDev)
// shutdown VF device
if err = netlink.LinkSetDown(vfDev); err != nil {
return fmt.Errorf("failed to down device: %v", err)
}
// rename VF device
err = ip.RenameLink(ifName, devName)
if err != nil {
return fmt.Errorf("failed to rename device %s to %s: %v", ifName, devName, err)
}
// move VF device to init netns
if err = netlink.LinkSetNsFd(vfDev, int(initns.Fd())); err != nil {
return fmt.Errorf("failed to move device %s to init netns: %v", ifName, err)
}
return nil
})
}
func cmdAdd(args *skel.CmdArgs) error {
fmt.Fprintf(os.Stderr, "***********CNI SR-IOV cmdAdd args.ContainerID = %v\n", args.ContainerID)
fmt.Fprintf(os.Stderr, "***********CNI SR-IOV cmdAdd args.Netns = %v\n", args.Netns)
fmt.Fprintf(os.Stderr, "***********CNI SR-IOV cmdAdd args.IfName = %v\n", args.IfName)
fmt.Fprintf(os.Stderr, "***********CNI SR-IOV cmdAdd args.Args = %v\n", args.Args)
fmt.Fprintf(os.Stderr, "***********CNI SR-IOV cmdAdd args.Path = %v\n", args.Path)
fmt.Fprintf(os.Stderr, "***********CNI SR-IOV cmdAdd args.StdinData = %v\n", string(args.StdinData))
n, cniVersion, err := loadConf(args.StdinData)
if err != nil {
return err
}
netns, err := ns.GetNS(args.Netns)
if err != nil {
return fmt.Errorf("failed to open netns %q: %v", args.Netns, err)
}
defer netns.Close()
vfInterface, err := setupVF(n, args.IfName, netns)
if err != nil {
return err
}
// run the IPAM plugin and get back the config to apply
r, err := ipam.ExecAdd(n.IPAM.Type, args.StdinData | Dir); err != nil {
return "", fmt.Errorf("failed to open the virtfn%d dir of the device %q: %v", vf, master, err)
}
infos, err := ioutil.ReadDir(vfDir)
if err != nil {
return "", fmt.Errorf("failed to read the virtfn%d dir of the device %q: %v", vf, master, err)
}
if len(infos) != 1 {
return "", fmt.Errorf("no network device in directory %s", vfDir)
}
return infos[0].Name(), nil
}
func releaseVF(conf *NetConf, ifName string, netns ns.NetNS) error {
initns, err := ns.GetCurrentNS | identifier_body |
ngx-grid-table.component.ts | End();
this.api.showLoadingOverlay();
this.setDataLoading(true);
this.dataSource(rowQuery)
.pipe(
takeUntil(this.destroy$),
catchError((err) => {
Console.collapse('grid-table.component RefreshRowsData', 'redBg', 'ERROR', 'redOutline');
console.log(err);
console.groupEnd();
return of({
total: 0,
records: [],
size: rowQuery.pageSize,
current: rowQuery.pageNum,
statistics: [],
} as IPage<any>);
}),
)
.subscribe((resultPage) => {
this.cur_page = resultPage;
this.api.setRowData(resultPage.records);
this.statistics = resultPage.statistics || [];
if (!resultPage.records.length) {
this.api.showNoRowsOverlay();
} else {
this.api.hideOverlay();
}
this.setDataLoading(false);
});
}
onPageSizeChange(size: number): void {
this.setPageSize(size, true);
this.query(1, size);
}
onPageIndexChange(idx: number): void {
this.setPageIndex(idx, true);
this.query(idx, this.pageSize);
}
/**
*
* @param maxRequestCount 最大并发查询数
* @param pageSize 每次查询条数,默认采用当前分页大小
*/
exportAllPageData(maxRequestCount: number = 3, pageSize?: number): void {
const initPageNum = 1;
pageSize = pageSize || this.pageSize;
if (this.dataLoadModel !== 'pageable') {
// 暂时只支持客户端模式导出全部数据,因为服务端模式下由于缓存大小问题setRowData方法无效
console.warn('pageable 模式才能前端导出!');
// 导出当前缓存的数据
this.api.exportDataAsExcel({ customFooter: this.exportStatisticsFooter() });
return;
}
const rowQuery: IRowQuery = clientSideAsRowQuery(this.api, this.columnApi, initPageNum, pageSize, this.filters());
const params = { percent: 0, status: 'active' };
const confirm = this.modal.confirm({
nzContent: this.progressTmpl,
nzTitle: this.translateService.instant('grid.export.confirm'),
nzComponentParams: params,
nzOkText: this.translateService.instant('grid.export.start'),
nzOnOk: () => {
params.percent = 0.01;
params.status = 'active';
confirm.updateConfig({
nzTitle: undefined,
nzContent: this.progressTmpl,
nzComponentParams: params,
nzOkText: null,
nzClosable: false,
nzCloseIcon: undefined,
nzIconType: undefined,
nzMaskClosable: false,
nzCancelText: null,
});
let statisticsFooter: Array<GridStatistics> = [];
return this.dataSource(rowQuery)
.pipe(
mergeMap((page: IPage<any>) => {
const { total, size, current, records, statistics } = page;
statisticsFooter = statistics || [];
const totalPage = Math.ceil(total / size);
if (totalPage > current) {
const step = parseFloat((100 / totalPage).toFixed(2));
params.percent = step;
return range(current + 1, totalPage).pipe(
mergeMap((index: number) => {
return this.dataSource(Object.assign({}, rowQuery, { pageNum: index }));
}, maxRequestCount),
pluck('records'),
tap((next) => {
params.percent = parseFloat((params.percent + step).toFixed(2));
}),
reduce((acc, val) => acc.concat(val || []), records),
);
} else {
return of([]);
}
}),
)
.toPromise()
.then((next) => {
params.status = '';
params.percent = 100;
this.api.setRowData(next);
this.api.exportDataAsExcel({ customFooter: this.exportStatisticsFooter(statisticsFooter) });
this.refresh();
return true;
})
.catch((err) => {
Console.collapse('grid-table.component ExportAllPageData', 'redBg', 'ERROR', 'redOutline');
console.error(err);
console.groupEnd();
params.status = 'exception';
confirm.updateConfig({
nzCancelText: undefined,
nzOkText: this.translateService.instant('grid.export.retry'),
});
return false;
});
},
});
}
private exportStatisticsFooter(statistics?: Array<GridStatistics>): ExcelCell[][] {
const footers: Array<Array<ExcelCell>> = [[]];
statistics = statistics || this.statistics;
if (this.showStatistics && statistics && statistics.length) {
const footer = statistics
.filter((items) => items.skipExport !== true)
.map((items) => {
const rows: Array<ExcelCell> = [];
rows.push({
styleId: 'bigHeader',
data: {
type: 'String',
value: items.label || '',
},
});
items.fields.forEach((item) => {
rows.push({
styleId: 'bigHeader',
data: {
type: 'String',
value: `${item.label}:${item.value}`,
},
} as ExcelCell);
});
return rows;
});
footers.push(...footer);
}
return footers;
}
ngOnDestroy(): void {
this.destroy$.next();
this.destroy$.complete();
}
public setData(data: IPage<any>): void {
if (this.dataLoadModel === 'pageable') {
this.cur_page = data;
this.setPageIndex(this.cur_page.current, false);
this.setPageSize(this.cur_page.size, false);
this.api.setRowData(this.cur_page.records);
} else {
console.warn('只有 tableModel === ‘pageable’ 才允许直接存值');
}
}
// 添加菜单
public addMenu(item: MenuItem): void {
this.additionMenu.push(item);
}
/**
* 刷新表格
*/
public refresh(): Observable<void> {
if (this.dataLoadModel === 'pageable') {
this.query(this.pageIndex, this.pageSize);
} else {
this.api.purgeServerSideCache();
}
// 当loading状态变更为false的时候,可以视为本次数据加载已经完成,返回这个Observable,供其他业务订阅
return this.dataLoadingChange.asObservable().pipe(
filter((status) => !status),
take(1),
map(() => {}),
);
}
/**
* 获取所有范围选中(range)的行数据
*/
public getSelectionData<U>(): U[] {
return this.getSelection((s) => s.data);
}
/**
* 获取所有范围选中(range)的行,并对其进行转换
*/
public getSelection<U>(hand: (value: RowNode) => U): U[] {
const range: CellRange = this.api.getCellRanges()[0];
const nodes: RowNode[] = this.api.getRenderedNodes();
if (range && range.startRow && range.endRow) {
const r = [];
for (let i = range.startRow.rowIndex; i <= range.endRow.rowIndex; i++) {
r.push(hand(nodes[i]));
}
return r;
} else {
return [];
}
}
/**
* 获取所有checkbox选择的行数据
*/
public getCheckedData<U>(): U[] {
return this.getChecked((s) => s.data);
}
/**
* 获取所有checkbox选择的行,并对其进行转换
*/
public getChecked<U>(hand: (value: RowNode) => U): U[] {
const nodes: RowNode[] = this.api.getSelectedNodes();
return nodes.map(hand);
}
/**
* 重置表单
*/
resetForm(): void {
this.form.reset();
}
get pageViewTmpl(): TemplateRef<any> {
return this.customPageView ? this.customPageView : this.defaultPageTmpl;
}
// ================================== 数据绑定 ====================================
toggleFullscreen(): void {
this.setFullscreen(!this.fullscreen);
}
setFullscreen(fullscreen: boolean): void {
this.fullscreen = fullscreen;
this.fullscreenChange.emit(this.fullscreen);
}
setPageIndex(pageIndex: number, emit: boolean): void {
this.pageIndex = pageIndex;
if (emit) {
this.pageIndexChange.emit(pageIndex);
}
}
setPageSize(pageSize: number, emit: boolean): void {
this.pageSize = pageSize;
if (emit) {
this.pageSizeChange.emit(pageSize);
}
}
private setDataLoading(loading: boolean): void {
this.dataLoading = loading;
this.dataLoadingChange.emit(this.dataLoading);
}
toggleDataModel(): void {
if ('pageable' === this.dataLoadModel) {
this.setDataMode('infinite');
} else {
this.setDataMode('pageable');
}
}
setDataMode(model: 'pageable' | 'infinite'): void {
this.dataLoadModel = model;
repairRowModeType(this.gridOptions, this.dataLoadModel);
// TODO 刷新表格
this.repaint();
this.dataLoadModelChange.emit(this.dataLoadModel);
}
repaint(): void {
this.__show__ = false;
setTimeout(() => (this.__show__ = true), 200);
}
}
| conditional_block |
||
ngx-grid-table.component.ts | (this.showPagination !== false) {
this.showPagination = {
...NgxGridTableConstants.DEFAULT_PAGINATION,
...this.showPagination,
};
}
this.gridOptions = initGridOptions(this.gridOptions, this.rowSelection!, (event) => this.onGridReady(event));
}
private onGridReady(event: GridReadyEvent): void {
this.api = event.api;
this.columnApi = event.columnApi;
if (this.dataLoadModel === 'infinite') {
this.api.setServerSideDatasource(this.infiniteDataSource());
}
if (this.haveInit) {
this.gridReLoadReady.emit({ event, gridTable: this });
} else {
this.gridReady.emit({ event, gridTable: this });
this.haveInit = true;
}
// 当网格数据就绪时
// this.api.addEventListener('firstDataRendered', () => {
// this.firstDataRenderedTime = new Date().getTime();
// if (cellActionColumnDefs.length) {
// this.columnApi.autoSizeColumn('_action');
// }
// });
if (this.initLoadData) {
this.refresh();
}
}
private filters(): IFilter[] {
let filters = [] as IFilter[];
if (this.form) {
filters = [...this.form.filter];
}
if (this.filterHand) {
filters = [...this.filterHand(filters, this.form)];
}
return filters;
}
private infiniteDataSource(): IServerSideDatasource {
const getRows = (params: IServerSideGetRowsParams) => {
const rowQuery = serverSideAsRowQuery(params, this.treeData, this.filters());
Console.collapse('grid-table.component getData', 'indigoBg', 'queryParams', 'indigoOutline');
console.log(rowQuery);
console.groupEnd();
this.setDataLoading(true);
this.api.showLoadingOverlay();
this.dataSource(rowQuery)
.pipe(
takeUntil(this.destroy$),
catchError((err) => {
Console.collapse('grid-table.component RefreshRowsData', 'redBg', 'ERROR', 'redOutline');
console.error(err);
console.groupEnd();
return of({} as IPage<any>);
}),
)
.subscribe((resultPage) => {
if (resultPage.records) {
params.successCallback(resultPage.records, resultPage.total);
this.statistics = resultPage.statistics || [];
this.api.hideOverlay();
} else {
this.api.showNoRowsOverlay();
params.failCallback();
}
this.setDataLoading(false);
});
};
return { getRows };
}
// private doDelete(): void {
// const data: any[] = this.getCheckedData();
// if (!data.length) {
// return;
// }
// let mapper: (v: any) => string;
// if (typeof this.gridOptions.getRowNodeId === 'undefined') {
// if (typeof data[0].id !== 'undefined') {
// mapper = (v) => v.id;
// } else {
// console.warn(
// '删除操作无法获取键,默认情况下将获取id作为键,如果没有id字段或不希望使用id作为删除键,请配置 gridOptions.getRowNodeId',
// );
// return;
// }
// } else {
// mapper = this.gridOptions.getRowNodeId;
// }
//
// const ids: string[] = data.map(mapper);
//
// this.http.delete(`/api/${this.currentUrl}/deleteByKeys`, { keys: ids }).subscribe((value) => {
// this.deleted.emit(value);
// });
// }
/**
* 查询
*/
query(pageNum: number, pageSize: number): void {
this.api.clearRangeSelection();
this.api.deselectAll();
if (this.dataLoadModel !== 'pageable') {
console.warn('pageable 模式才能进行分页查询');
return;
}
const rowQuery: IRowQuery = clientSideAsRowQuery(this.api, this.columnApi, pageNum, pageSize, this.filters());
Console.collapse('grid-table.component getData', 'indigoBg', 'queryParams', 'indigoOutline');
console.log(rowQuery);
console.groupEnd();
this.api.showLoadingOverlay();
this.setDataLoading(true);
this.dataSource(rowQuery)
.pipe(
takeUntil(this.destroy$),
catchError((err) => {
Console.collapse('grid-table.component RefreshRowsData', 'redBg', 'ERROR', 'redOutline');
console.log(err);
console.groupEnd();
return of({
total: 0,
records: [],
size: rowQuery.pageSize,
current: rowQuery.pageNum,
statistics: [],
} as IPage<any>);
}),
)
.subscribe((resultPage) => {
this.cur_page = resultPage;
this.api.setRowData(resultPage.records);
this.statistics = resultPage.statistics || [];
if (!resultPage.records.length) {
this.api.showNoRowsOverlay();
} else {
this.api.hideOverlay();
}
this.setDataLoading(false);
});
}
onPageSizeChange(size: number): void {
this.setPageSize(size, true);
this.query(1, size);
}
onPageIndexChange(idx: number): void {
this.setPageIndex(idx, true);
this.query(idx, this.pageSize);
}
/**
*
* @param maxRequestCount 最大并发查询数
* @param pageSize 每次查询条数,默认采用当前分页大小
*/
exportAllPageData(maxRequestCount: number = 3, pageSize?: number): void {
const initPageNum = 1;
pageSize = pageSize || this.pageSize;
if (this.dataLoadModel !== 'pageable') {
// 暂时只支持客户端模式导出全部数据,因为服务端模式下由于缓存大小问题setRowData方法无效
console.warn('pageable 模式才能前端导出!');
// 导出当前缓存的数据
this.api.exportDataAsExcel({ customFooter: this.exportStatisticsFooter() });
return;
}
const rowQuery: IRowQuery = clientSideAsRowQuery(this.api, this.columnApi, initPageNum, pageSize, this.filters());
const params = { percent: 0, status: 'active' };
const confirm = this.modal.confirm({
nzContent: this.progressTmpl,
nzTitle: this.translateService.instant('grid.export.confirm'),
nzComponentParams: params,
nzOkText: this.translateService.instant('grid.export.start'),
nzOnOk: () => {
params.percent = 0.01;
params.status = 'active';
confirm.updateConfig({
nzTitle: undefined,
nzContent: this.progressTmpl,
nzComponentParams: params,
nzOkText: null,
nzClosable: false,
nzCloseIcon: undefined,
nzIconType: undefined,
nzMaskClosable: false,
nzCancelText: null,
});
let statisticsFooter: Array<GridStatistics> = [];
return this.dataSource(rowQuery)
.pipe(
mergeMap((page: IPage<any>) => {
const { total, size, current, records, statistics } = page;
statisticsFooter = statistics || [];
const totalPage = Math.ceil(total / size);
if (totalPage > current) {
const step = parseFloat((100 / totalPage).toFixed(2));
params.percent = step;
return range(current + 1, totalPage).pipe(
mergeMap((index: number) => {
return this.dataSource(Object.assign({}, rowQuery, { pageNum: index }));
}, maxRequestCount),
pluck('records'),
tap((next) => {
params.percent = parseFloat((params.percent + step).toFixed(2));
}),
reduce((acc, val) => acc.concat(val || []), records),
);
} else {
return of([]);
}
}),
)
.toPromise()
.then((next) => {
params.status = '';
params.percent = 100;
this.api.setRowData(next);
this.api.exportDataAsExcel({ customFooter: this.exportStatisticsFooter(statisticsFooter) });
this.refresh();
return true;
})
.catch((err) => {
Console.collapse('grid-table.component ExportAllPageData', 'redBg', 'ERROR', 'redOutline');
console.error(err);
console.groupEnd();
params.status = 'exception';
confirm.updateConfig({
nzCancelText: undefined,
nzOkText: this.translateService.instant('grid.export.retry'),
});
return false;
});
},
});
}
private exportStatisticsFooter(statistics?: Array<GridStatistics>): ExcelCell[][] {
const footers: Array<Array<ExcelCell>> = [[]];
statistics = statistics || this.statistics;
if (this.showStatistics && statistics && statistics.length) {
const footer = statistics
.filter((items) => items.skipExport !== true)
.map((items) => {
const rows: Array<ExcelCell> = [];
rows.push({
styleId: 'bigHeader',
data: {
type: 'String',
value: items.label || '',
},
});
items.fields.forEach((item) => {
rows.push({
styleId: 'bigHeader', | data: { | random_line_split |
|
ngx-grid-table.component.ts | `, { keys: ids }).subscribe((value) => {
// this.deleted.emit(value);
// });
// }
/**
* 查询
*/
query(pageNum: number, pageSize: number): void {
this.api.clearRangeSelection();
this.api.deselectAll();
if (this.dataLoadModel !== 'pageable') {
console.warn('pageable 模式才能进行分页查询');
return;
}
const rowQuery: IRowQuery = clientSideAsRowQuery(this.api, this.columnApi, pageNum, pageSize, this.filters());
Console.collapse('grid-table.component getData', 'indigoBg', 'queryParams', 'indigoOutline');
console.log(rowQuery);
console.groupEnd();
this.api.showLoadingOverlay();
this.setDataLoading(true);
this.dataSource(rowQuery)
.pipe(
takeUntil(this.destroy$),
catchError((err) => {
Console.collapse('grid-table.component RefreshRowsData', 'redBg', 'ERROR', 'redOutline');
console.log(err);
console.groupEnd();
return of({
total: 0,
records: [],
size: rowQuery.pageSize,
current: rowQuery.pageNum,
statistics: [],
} as IPage<any>);
}),
)
.subscribe((resultPage) => {
this.cur_page = resultPage;
this.api.setRowData(resultPage.records);
this.statistics = resultPage.statistics || [];
if (!resultPage.records.length) {
this.api.showNoRowsOverlay();
} else {
this.api.hideOverlay();
}
this.setDataLoading(false);
});
}
onPageSizeChange(size: number): void {
this.setPageSize(size, true);
this.query(1, size);
}
onPageIndexChange(idx: number): void {
this.setPageIndex(idx, true);
this.query(idx, this.pageSize);
}
/**
*
* @param maxRequestCount 最大并发查询数
* @param pageSize 每次查询条数,默认采用当前分页大小
*/
exportAllPageData(maxRequestCount: number = 3, pageSize?: number): void {
const initPageNum = 1;
pageSize = pageSize || this.pageSize;
if (this.dataLoadModel !== 'pageable') {
// 暂时只支持客户端模式导出全部数据,因为服务端模式下由于缓存大小问题setRowData方法无效
console.warn('pageable 模式才能前端导出!');
// 导出当前缓存的数据
this.api.exportDataAsExcel({ customFooter: this.exportStatisticsFooter() });
return;
}
const rowQuery: IRowQuery = clientSideAsRowQuery(this.api, this.columnApi, initPageNum, pageSize, this.filters());
const params = { percent: 0, status: 'active' };
const confirm = this.modal.confirm({
nzContent: this.progressTmpl,
nzTitle: this.translateService.instant('grid.export.confirm'),
nzComponentParams: params,
nzOkText: this.translateService.instant('grid.export.start'),
nzOnOk: () => {
params.percent = 0.01;
params.status = 'active';
confirm.updateConfig({
nzTitle: undefined,
nzContent: this.progressTmpl,
nzComponentParams: params,
nzOkText: null,
nzClosable: false,
nzCloseIcon: undefined,
nzIconType: undefined,
nzMaskClosable: false,
nzCancelText: null,
});
let statisticsFooter: Array<GridStatistics> = [];
return this.dataSource(rowQuery)
.pipe(
mergeMap((page: IPage<any>) => {
const { total, size, current, records, statistics } = page;
statisticsFooter = statistics || [];
const totalPage = Math.ceil(total / size);
if (totalPage > current) {
const step = parseFloat((100 / totalPage).toFixed(2));
params.percent = step;
return range(current + 1, totalPage).pipe(
mergeMap((index: number) => {
return this.dataSource(Object.assign({}, rowQuery, { pageNum: index }));
}, maxRequestCount),
pluck('records'),
tap((next) => {
params.percent = parseFloat((params.percent + step).toFixed(2));
}),
reduce((acc, val) => acc.concat(val || []), records),
);
} else {
return of([]);
}
}),
)
.toPromise()
.then((next) => {
params.status = '';
params.percent = 100;
this.api.setRowData(next);
this.api.exportDataAsExcel({ customFooter: this.exportStatisticsFooter(statisticsFooter) });
this.refresh();
return true;
})
.catch((err) => {
Console.collapse('grid-table.component ExportAllPageData', 'redBg', 'ERROR', 'redOutline');
console.error(err);
console.groupEnd();
params.status = 'exception';
confirm.updateConfig({
nzCancelText: undefined,
nzOkText: this.translateService.instant('grid.export.retry'),
});
return false;
});
},
});
}
private exportStatisticsFooter(statistics?: Array<GridStatistics>): ExcelCell[][] {
const footers: Array<Array<ExcelCell>> = [[]];
statistics = statistics || this.statistics;
if (this.showStatistics && statistics && statistics.length) {
const footer = statistics
.filter((items) => items.skipExport !== true)
.map((items) => {
const rows: Array<ExcelCell> = [];
rows.push({
styleId: 'bigHeader',
data: {
type: 'String',
value: items.label || '',
},
});
items.fields.forEach((item) => {
rows.push({
styleId: 'bigHeader',
data: {
type: 'String',
value: `${item.label}:${item.value}`,
},
} as ExcelCell);
});
return rows;
});
footers.push(...footer);
}
return footers;
}
ngOnDestroy(): void {
this.destroy$.next();
this.destroy$.complete();
}
public setData(data: IPage<any>): void {
if (this.dataLoadModel === 'pageable') {
this.cur_page = data;
this.setPageIndex(this.cur_page.current, false);
this.setPageSize(this.cur_page.size, false);
this.api.setRowData(this.cur_page.records);
} else {
console.warn('只有 tableModel === ‘pageable’ 才允许直接存值');
}
}
// 添加菜单
public addMenu(item: MenuItem): void {
this.additionMenu.push(item);
}
/**
* 刷新表格
*/
public refresh(): Observable<void> {
if (this.dataLoadModel === 'pageable') {
this.query(this.pageIndex, this.pageSize);
} else {
this.api.purgeServerSideCache();
}
// 当loading状态变更为false的时候,可以视为本次数据加载已经完成,返回这个Observable,供其他业务订阅
return this.dataLoadingChange.asObservable().pipe(
filter((status) => !status),
take(1),
map(() => {}),
);
}
/**
* 获取所有范围选中(range)的行数据
*/
public getSelectionData<U>(): U[] {
return this.getSelection((s) => s.data);
}
/**
* 获取所有范围选中(range)的行,并对其进行转换
*/
public getSelection<U>(hand: (value: RowNode) => U): U[] {
const range: CellRange = this.api.getCellRanges()[0];
const nodes: RowNode[] = this.api.getRenderedNodes();
if (range && range.startRow && range.endRow) {
const r = [];
for (let i = range.startRow.rowIndex; i <= range.endRow.rowIndex; i++) {
r.push(hand(nodes[i]));
}
return r;
} else {
return [];
}
}
/**
* 获取所有checkbox选择的行数据
*/
public getCheckedData<U>(): U[] {
return this.getChecked((s) => s.data);
}
/**
* 获取所有checkbox选择的行,并对其进行转换
*/
public getChecked<U>(hand: (value: RowNode) => U): U[] {
const nodes: RowNode[] = this.api.getSelectedNodes();
return nodes.map(hand);
}
/**
* 重置表单
*/
resetForm(): void {
this.form.reset();
}
get pageViewTmpl(): TemplateRef<any> {
return this.customPageView ? this.customPageView : this.defaultPageTmpl;
}
// ================================== 数据绑定 ====================================
toggleFullscreen(): void {
this.setFullscreen(!this.fullscreen);
}
setFullscreen(fullscreen: boolean): void {
this.fullscreen = fullscreen;
this.fullscreenChange.emit(this.fullscreen);
}
setPageIndex(pageIndex: number, emit: boolean): void {
this.pageIndex = pageIndex;
if (emit) {
this.pageIndexChange.emit(pageIndex);
}
}
setPageSize(pageSize: number, emit: boolean): void {
this.pageSize = pageSize;
if (emit) {
this.pageSizeChange.emit(pageSize);
}
}
private setDataLoading(loading: boolean): void {
| this.dataL | identifier_name |
|
ngx-grid-table.component.ts | = 1;
/** 页大小 */
@Input() pageSize = 20;
/** 表格基础配置 */
@Input() gridOptions!: GridOptions;
/** 表格基础配置 */
@Input() colACLTmpl!: string;
/** 表格主题 */
@Input() gridTheme = 'ag-theme-balham';
/** 表格CSS */
@Input() gridTableClass = [];
/** 数据表格样式 */
@Input() gridTableStyle: { [key: string]: any } = { width: '100%', height: '70%' };
/** 是否在表格初始化后立即执行一次查询 */
@Input() initLoadData = true;
/** 分页还是无限 */
@Input() dataLoadModel: 'pageable' | 'infinite' = 'pageable';
/** 是否显示分页控件 */
@Input() showPagination: false | PaginationCfg = {};
/** 是否显示默认状态栏, 展示用户选中项状态数据 */
@Input() defaultStatusBar = false;
/** 单行还是多行 */
@Input() rowSelection: undefined | 'single' | 'multiple' = undefined;
/** 是否显示统计 */
@Input() showStatistics = false;
// /** 是否展示删除菜单 */
// @Input() deleteMenu = false;
/** 默认是否可以改变列宽 */
@Input() resizable = false;
/** 分组时checkbox是否影响自己的子项 */
@Input() groupSelectsChildren = true;
/** 操作列模板 */
@Input() optionCell!: TemplateRef<any>;
/** 数据源 */
@Input() dataSource!: IGridDataSource<any>;
/** 表单schema */
@Input() searchSchema!: SFSchema;
/** 初始表单数据 */
@Input() initFormData!: any;
@Input() customPageView!: TemplateRef<any>;
@Input() filterHand!: (filters: IFilter[], form: SfQueryFormComponent) => IFilter[];
@Input() topToolPanel!: TemplateRef<any>;
@Input() bottomToolPanel!: TemplateRef<any>;
// ============================== 事件 ============================
@Output() fullscreenChange = new EventEmitter<boolean>();
@Output() pageIndexChange = new EventEmitter<number>();
@Output() pageSizeChange = new EventEmitter<number>();
/** 表格就绪事件 */
@Output() gridReady = new EventEmitter<{ event: GridReadyEvent; gridTable: NgxGridTableComponent }>();
@Output() gridReLoadReady = new EventEmitter<{ event: GridReadyEvent; gridTable: NgxGridTableComponent }>();
/** 删除事件 */
@Output() deleted = new EventEmitter<any>();
@Output() dataLoadingChange = new EventEmitter<boolean>();
@Output() dataLoadModelChange = new EventEmitter<'pageable' | 'infinite'>();
// ============================= 组件 =======================
@ViewChild(SfQueryFormComponent) form!: SfQueryFormComponent;
@ViewChild('defaultPageTmpl') defaultPageTmpl!: TemplateRef<any>;
@ViewChild('progressTmpl') progressTmpl!: TemplateRef<any>;
constructor(
private translateService: TranslateService,
private msg: NzMessageService,
private modal: NzModalService,
private reuseTabService: ReuseTabService,
private activatedRoute: ActivatedRoute,
private router: Router,
private http: _HttpClient,
private aclService: ACLService,
) {
this.currentUrl = this.router.url;
}
ngOnInit(): void {
if (!this.haveInit) {
this.initGridOptions();
}
}
private initGridOptions() {
// api 获取方法,用于给functions传递api对象
const apiGetter: ApiGetter = { get: () => ({ api: this.api, columnApi: this.columnApi }) };
buildFrameworkComponents(this.gridOptions);
// 构建菜单
buildMenus(
this.gridOptions,
this.translateService,
this.aclService,
this.additionMenu,
// this.deleteMenu,
this.destroy$,
apiGetter,
() => this.getSelectionData(),
// () => this.doDelete(),
);
buildTreeDataCfg(this.gridOptions, this.treeData);
buildStatusBar(this.gridOptions, this.defaultStatusBar);
buildSideBar(this.gridOptions) |
if (this.dataLoadModel === 'infinite') {
this.api.setServerSideDatasource(this.infiniteDataSource());
}
if (this.haveInit) {
this.gridReLoadReady.emit({ event, gridTable: this });
} else {
this.gridReady.emit({ event, gridTable: this });
this.haveInit = true;
}
// 当网格数据就绪时
// this.api.addEventListener('firstDataRendered', () => {
// this.firstDataRenderedTime = new Date().getTime();
// if (cellActionColumnDefs.length) {
// this.columnAp
i.autoSizeColumn('_action');
// }
// });
if (this.initLoadData) {
this.refresh();
}
}
private filters(): IFilter[] {
let filters = [] as IFilter[];
if (this.form) {
filters = [...this.form.filter];
}
if (this.filterHand) {
filters = [...this.filterHand(filters, this.form)];
}
return filters;
}
private infiniteDataSource(): IServerSideDatasource {
const getRows = (params: IServerSideGetRowsParams) => {
const rowQuery = serverSideAsRowQuery(params, this.treeData, this.filters());
Console.collapse('grid-table.component getData', 'indigoBg', 'queryParams', 'indigoOutline');
console.log(rowQuery);
console.groupEnd();
this.setDataLoading(true);
this.api.showLoadingOverlay();
this.dataSource(rowQuery)
.pipe(
takeUntil(this.destroy$),
catchError((err) => {
Console.collapse('grid-table.component RefreshRowsData', 'redBg', 'ERROR', 'redOutline');
console.error(err);
console.groupEnd();
return of({} as IPage<any>);
}),
)
.subscribe((resultPage) => {
if (resultPage.records) {
params.successCallback(resultPage.records, resultPage.total);
this.statistics = resultPage.statistics || [];
this.api.hideOverlay();
} else {
this.api.showNoRowsOverlay();
params.failCallback();
}
this.setDataLoading(false);
});
};
return { getRows };
}
// private doDelete(): void {
// const data: any[] = this.getCheckedData();
// if (!data.length) {
// return;
// }
// let mapper: (v: any) => string;
// if (typeof this.gridOptions.getRowNodeId === 'undefined') {
// if (typeof data[0].id !== 'undefined') {
// mapper = (v) => v.id;
// } else {
// console.warn(
// '删除操作无法获取键,默认情况下将获取id作为键,如果没有id字段或不希望使用id作为删除键,请配置 gridOptions.getRowNodeId',
// );
// return;
// }
// } else {
// mapper = this.gridOptions.getRowNodeId;
// }
//
// const ids: string[] = data.map(mapper);
//
// this.http.delete(`/api/${this.currentUrl}/deleteByKeys`, { keys: ids }).subscribe((value) => {
// this.deleted.emit(value);
// });
// }
/**
* 查询
*/
query(pageNum: number, pageSize: number): void {
this.api.clearRangeSelection();
this.api.deselectAll();
if (this.dataLoadModel !== 'pageable') {
console.warn('pageable 模式才能进行分页查询');
return;
}
const rowQuery: IRowQuery = clientSideAsRowQuery(this.api, this.columnApi, pageNum, pageSize, this.filters());
Console.collapse('grid-table.component getData', 'indigoBg', 'queryParams', 'indigoOutline');
console.log(rowQuery);
console.groupEnd();
this.api.showLoadingOverlay();
this.setDataLoading(true);
this.dataSource(rowQuery)
.pipe(
takeUntil(this.destroy$),
catchError((err) => {
Console.collapse('grid-table.component RefreshRowsData', 'redBg', 'ERROR', 'redOutline');
console.log(err);
console.groupEnd();
return of({
total: 0,
records: [],
size: rowQuery.pageSize,
current: rowQuery.pageNum,
statistics: [],
} as I | ;
buildResizable(this.gridOptions, this.resizable);
buildOptionField(this.gridOptions, this.optionCell);
reuseTabFix(this.router, this.currentUrl, this.destroy$, apiGetter);
repairRowModeType(this.gridOptions, this.dataLoadModel);
buildColACL(this.gridOptions, this.aclService, this.colACLTmpl);
if (this.showPagination !== false) {
this.showPagination = {
...NgxGridTableConstants.DEFAULT_PAGINATION,
...this.showPagination,
};
}
this.gridOptions = initGridOptions(this.gridOptions, this.rowSelection!, (event) => this.onGridReady(event));
}
private onGridReady(event: GridReadyEvent): void {
this.api = event.api;
this.columnApi = event.columnApi; | identifier_body |
normalizations.py | """Normalization layers."""
from typing import List, Optional, Tuple
import jax
from jax import numpy as jnp
from lingvo.jax import base_layer
from lingvo.jax import py_utils
from lingvo.jax import pytypes
NestedMap = py_utils.NestedMap
WeightInit = py_utils.WeightInit
weight_params = py_utils.weight_params
InstantiableParams = py_utils.InstantiableParams
JTensor = pytypes.JTensor
def compute_moments(
inputs: JTensor,
padding: JTensor,
reduce_over_dims: List[int],
enable_cross_replica_sum_on_tpu: bool = False,
keepdims: bool = False,
) -> Tuple[JTensor, JTensor]:
"""Computes mean and variance over the valid data points in inputs."""
assert inputs.ndim == padding.ndim
rank = inputs.ndim
assert all([0 <= dim < rank for dim in reduce_over_dims])
mask = 1.0 - padding
sum_v = jnp.sum(inputs * mask, axis=reduce_over_dims, keepdims=keepdims)
count_v = jnp.sum(
jnp.ones_like(inputs) * mask, axis=reduce_over_dims, keepdims=keepdims)
if enable_cross_replica_sum_on_tpu:
# TODO(shafey, yonghui): Fetch axis_name from globals.
sum_v = jax.lax.psum(sum_v, axis_name='batch')
count_v = jax.lax.psum(count_v, axis_name='batch')
count_v = jnp.maximum(count_v, 1.0)
mean = sum_v / count_v
sum_vv = jnp.sum(
(inputs - mean) * (inputs - mean) * mask,
axis=reduce_over_dims,
keepdims=keepdims)
if enable_cross_replica_sum_on_tpu:
# TODO(shafey, yonghui): Fetch axis_name from globals.
sum_vv = jax.lax.psum(sum_vv, axis_name='batch')
variance = sum_vv / count_v
return mean, variance
class BatchNorm(base_layer.BaseLayer):
"""Batch normalization layer."""
@classmethod
def Params(cls) -> InstantiableParams:
p = super().Params()
p.Define('dim', 0, 'Depth of the input/output.')
p.Define(
'decay', 0.999,
'Decay in updating the mean and variance moving average used in'
' batch normalization.')
p.Define(
'enable_cross_replica_sum_on_tpu', False,
'If true, computes global mean and variance across all replicas.'
'Only effective for tpu.')
p.Define(
'use_moving_avg_in_training', False,
'If True, use global moving avg (mean, variance) during training'
' to avoid mismatch between train and eval, which then'
' essentially acts as an adaptive normalization step. When this is'
' set to True, it also disables the use of beta and gamma variables.')
p.Define('set_padded_output_to_zero', True,
'If True, sets the padded outputs to zero.')
return p
def __init__(self, params: InstantiableParams) -> None:
super().__init__(params)
p = self.params
self._epsilon = 0.001
self._decay = p.decay
def _get_weight_shape(self) -> JTensor:
return [self.params.dim]
def create_layer_variables(self) -> None:
p = self.params
beta_pc = weight_params(
shape=self._get_weight_shape(),
init=WeightInit.Constant(0.0),
dtype=p.dtype)
self.create_variable('beta', beta_pc)
# gamma = theta.gamma + 1.0
gamma_pc = weight_params(
shape=self._get_weight_shape(),
init=WeightInit.Constant(0.0),
dtype=p.dtype)
self.create_variable('gamma', gamma_pc)
mva = weight_params(
shape=[p.dim],
init=WeightInit.Constant(0.0),
dtype=p.dtype,
collections=[base_layer.REQUIRES_MEAN_SYNC])
self.create_variable('moving_mean', mva, trainable=False)
mvv = weight_params(
shape=[p.dim],
init=WeightInit.Constant(1.0),
dtype=p.dtype,
collections=[base_layer.REQUIRES_MEAN_SYNC])
self.create_variable('moving_variance', mvv, trainable=False)
def _get_default_paddings(self, inputs: JTensor) -> JTensor:
"""Gets the default paddings for an input."""
in_shape = list(inputs.shape)
assert len(in_shape) > 1
in_shape[-1] = 1
return jnp.zeros(in_shape, dtype=inputs.dtype)
def _get_beta_gamma(self, theta: NestedMap) -> Tuple[JTensor, JTensor]:
p = self.params
if p.use_moving_avg_in_training:
beta = 0.0
gamma = 1.0
else:
beta = theta.beta
gamma = theta.gamma + 1.0
return beta, gamma
def compute_and_update_moments(
self, theta: NestedMap, inputs: JTensor,
paddings: JTensor) -> Tuple[JTensor, JTensor, JTensor, JTensor]:
"""Computes moments and updates state.
Args:
theta: A `.NestedMap` object containing weights' values of this layer and
its children layers.
inputs: The inputs JTensor. Shaped [..., dim].
paddings: The paddings JTensor. Shaped [..., 1], with the same rank as
the input JTensor.
Returns:
Tuple of (mean, variance, beta, gamma).
"""
p = self.params
if self.do_eval:
# The mean and variance used for normalization.
norm_mean, norm_variance = theta.moving_mean, theta.moving_variance
base_layer.add_summary('moving_mean', theta.moving_mean)
base_layer.add_summary('moving_variance', theta.moving_variance)
else:
rank = inputs.ndim
reduce_over_dims = list(range(0, rank - 1))
mean, variance = compute_moments(
inputs,
paddings,
reduce_over_dims,
enable_cross_replica_sum_on_tpu=p.enable_cross_replica_sum_on_tpu,
keepdims=True)
new_moving_mean = theta.moving_mean * p.decay + mean * (1.0 - p.decay)
self.forward_update_var('moving_mean', new_moving_mean)
new_moving_variance = (
theta.moving_variance * p.decay + variance * (1.0 - p.decay))
self.forward_update_var('moving_variance', new_moving_variance)
# Add some summaries for visualization.
base_layer.add_summary('mean', mean)
base_layer.add_summary('variance', variance)
base_layer.add_summary('moving_mean', theta.moving_mean)
base_layer.add_summary('moving_variance', theta.moving_variance)
if p.use_moving_avg_in_training:
# Use the global statistics for normalization.
|
else:
# Use the batch statistics for normalization.
norm_mean = mean
norm_variance = variance
beta, gamma = self._get_beta_gamma(theta)
return norm_mean, norm_variance, beta, gamma
def fprop(self,
theta: NestedMap,
inputs: JTensor,
paddings: Optional[JTensor] = None) -> JTensor:
"""Apply batch normalization.
Args:
theta: A `.NestedMap` object containing weights' values of this layer and
its children layers.
inputs: The inputs JTensor. Shaped [..., dim].
paddings: The paddings JTensor. Shaped [..., 1].
Returns:
Output after applying batch normalization, with the same shape as
'inputs'.
"""
p = self.params
inputs, paddings = self._cast_to_fprop_dtype((inputs, paddings))
if paddings is None:
paddings = self._get_default_paddings(inputs)
assert inputs.ndim == paddings.ndim
assert paddings.shape[-1] == 1
norm_mean, norm_variance, beta, gamma = self.compute_and_update_moments(
theta, inputs, paddings)
inv = gamma / jnp.sqrt(norm_variance + self._epsilon)
bn_output = (inputs - norm_mean) * inv + beta
if p.set_padded_output_to_zero:
bn_output *= 1.0 - paddings
return bn_output
class LayerNorm(base_layer.BaseLayer):
"""Layer normalization."""
@classmethod
def Params(cls) -> InstantiableParams:
p = super().Params()
p.Define('input_dims', 0, 'Depth of the input to the network.')
p.Define('epsilon', 1e-6, 'Tiny value to guard rsqrt.')
p.Define('scale', True, 'Whether to use a learned scaling.')
p.Define('bias', True, 'Whether to use bias.')
return p
def create_layer_variables(self) -> None:
super().create_layer_variables()
p = self.params
wp = p.weight_split_dims_mapping
wp_scale = wp.wt
| norm_mean = theta.moving_mean
norm_variance = theta.moving_variance | conditional_block |
normalizations.py | """Normalization layers."""
from typing import List, Optional, Tuple
import jax
from jax import numpy as jnp
from lingvo.jax import base_layer
from lingvo.jax import py_utils
from lingvo.jax import pytypes
NestedMap = py_utils.NestedMap
WeightInit = py_utils.WeightInit
weight_params = py_utils.weight_params
InstantiableParams = py_utils.InstantiableParams
JTensor = pytypes.JTensor
def compute_moments(
inputs: JTensor,
padding: JTensor,
reduce_over_dims: List[int],
enable_cross_replica_sum_on_tpu: bool = False,
keepdims: bool = False,
) -> Tuple[JTensor, JTensor]:
|
if enable_cross_replica_sum_on_tpu:
# TODO(shafey, yonghui): Fetch axis_name from globals.
sum_vv = jax.lax.psum(sum_vv, axis_name='batch')
variance = sum_vv / count_v
return mean, variance
class BatchNorm(base_layer.BaseLayer):
"""Batch normalization layer."""
@classmethod
def Params(cls) -> InstantiableParams:
p = super().Params()
p.Define('dim', 0, 'Depth of the input/output.')
p.Define(
'decay', 0.999,
'Decay in updating the mean and variance moving average used in'
' batch normalization.')
p.Define(
'enable_cross_replica_sum_on_tpu', False,
'If true, computes global mean and variance across all replicas.'
'Only effective for tpu.')
p.Define(
'use_moving_avg_in_training', False,
'If True, use global moving avg (mean, variance) during training'
' to avoid mismatch between train and eval, which then'
' essentially acts as an adaptive normalization step. When this is'
' set to True, it also disables the use of beta and gamma variables.')
p.Define('set_padded_output_to_zero', True,
'If True, sets the padded outputs to zero.')
return p
def __init__(self, params: InstantiableParams) -> None:
super().__init__(params)
p = self.params
self._epsilon = 0.001
self._decay = p.decay
def _get_weight_shape(self) -> JTensor:
return [self.params.dim]
def create_layer_variables(self) -> None:
p = self.params
beta_pc = weight_params(
shape=self._get_weight_shape(),
init=WeightInit.Constant(0.0),
dtype=p.dtype)
self.create_variable('beta', beta_pc)
# gamma = theta.gamma + 1.0
gamma_pc = weight_params(
shape=self._get_weight_shape(),
init=WeightInit.Constant(0.0),
dtype=p.dtype)
self.create_variable('gamma', gamma_pc)
mva = weight_params(
shape=[p.dim],
init=WeightInit.Constant(0.0),
dtype=p.dtype,
collections=[base_layer.REQUIRES_MEAN_SYNC])
self.create_variable('moving_mean', mva, trainable=False)
mvv = weight_params(
shape=[p.dim],
init=WeightInit.Constant(1.0),
dtype=p.dtype,
collections=[base_layer.REQUIRES_MEAN_SYNC])
self.create_variable('moving_variance', mvv, trainable=False)
def _get_default_paddings(self, inputs: JTensor) -> JTensor:
"""Gets the default paddings for an input."""
in_shape = list(inputs.shape)
assert len(in_shape) > 1
in_shape[-1] = 1
return jnp.zeros(in_shape, dtype=inputs.dtype)
def _get_beta_gamma(self, theta: NestedMap) -> Tuple[JTensor, JTensor]:
p = self.params
if p.use_moving_avg_in_training:
beta = 0.0
gamma = 1.0
else:
beta = theta.beta
gamma = theta.gamma + 1.0
return beta, gamma
def compute_and_update_moments(
self, theta: NestedMap, inputs: JTensor,
paddings: JTensor) -> Tuple[JTensor, JTensor, JTensor, JTensor]:
"""Computes moments and updates state.
Args:
theta: A `.NestedMap` object containing weights' values of this layer and
its children layers.
inputs: The inputs JTensor. Shaped [..., dim].
paddings: The paddings JTensor. Shaped [..., 1], with the same rank as
the input JTensor.
Returns:
Tuple of (mean, variance, beta, gamma).
"""
p = self.params
if self.do_eval:
# The mean and variance used for normalization.
norm_mean, norm_variance = theta.moving_mean, theta.moving_variance
base_layer.add_summary('moving_mean', theta.moving_mean)
base_layer.add_summary('moving_variance', theta.moving_variance)
else:
rank = inputs.ndim
reduce_over_dims = list(range(0, rank - 1))
mean, variance = compute_moments(
inputs,
paddings,
reduce_over_dims,
enable_cross_replica_sum_on_tpu=p.enable_cross_replica_sum_on_tpu,
keepdims=True)
new_moving_mean = theta.moving_mean * p.decay + mean * (1.0 - p.decay)
self.forward_update_var('moving_mean', new_moving_mean)
new_moving_variance = (
theta.moving_variance * p.decay + variance * (1.0 - p.decay))
self.forward_update_var('moving_variance', new_moving_variance)
# Add some summaries for visualization.
base_layer.add_summary('mean', mean)
base_layer.add_summary('variance', variance)
base_layer.add_summary('moving_mean', theta.moving_mean)
base_layer.add_summary('moving_variance', theta.moving_variance)
if p.use_moving_avg_in_training:
# Use the global statistics for normalization.
norm_mean = theta.moving_mean
norm_variance = theta.moving_variance
else:
# Use the batch statistics for normalization.
norm_mean = mean
norm_variance = variance
beta, gamma = self._get_beta_gamma(theta)
return norm_mean, norm_variance, beta, gamma
def fprop(self,
theta: NestedMap,
inputs: JTensor,
paddings: Optional[JTensor] = None) -> JTensor:
"""Apply batch normalization.
Args:
theta: A `.NestedMap` object containing weights' values of this layer and
its children layers.
inputs: The inputs JTensor. Shaped [..., dim].
paddings: The paddings JTensor. Shaped [..., 1].
Returns:
Output after applying batch normalization, with the same shape as
'inputs'.
"""
p = self.params
inputs, paddings = self._cast_to_fprop_dtype((inputs, paddings))
if paddings is None:
paddings = self._get_default_paddings(inputs)
assert inputs.ndim == paddings.ndim
assert paddings.shape[-1] == 1
norm_mean, norm_variance, beta, gamma = self.compute_and_update_moments(
theta, inputs, paddings)
inv = gamma / jnp.sqrt(norm_variance + self._epsilon)
bn_output = (inputs - norm_mean) * inv + beta
if p.set_padded_output_to_zero:
bn_output *= 1.0 - paddings
return bn_output
class LayerNorm(base_layer.BaseLayer):
"""Layer normalization."""
@classmethod
def Params(cls) -> InstantiableParams:
p = super().Params()
p.Define('input_dims', 0, 'Depth of the input to the network.')
p.Define('epsilon', 1e-6, 'Tiny value to guard rsqrt.')
p.Define('scale', True, 'Whether to use a learned scaling.')
p.Define('bias', True, 'Whether to use bias.')
return p
def create_layer_variables(self) -> None:
super().create_layer_variables()
p = self.params
wp = p.weight_split_dims_mapping
wp_scale = wp.wt | """Computes mean and variance over the valid data points in inputs."""
assert inputs.ndim == padding.ndim
rank = inputs.ndim
assert all([0 <= dim < rank for dim in reduce_over_dims])
mask = 1.0 - padding
sum_v = jnp.sum(inputs * mask, axis=reduce_over_dims, keepdims=keepdims)
count_v = jnp.sum(
jnp.ones_like(inputs) * mask, axis=reduce_over_dims, keepdims=keepdims)
if enable_cross_replica_sum_on_tpu:
# TODO(shafey, yonghui): Fetch axis_name from globals.
sum_v = jax.lax.psum(sum_v, axis_name='batch')
count_v = jax.lax.psum(count_v, axis_name='batch')
count_v = jnp.maximum(count_v, 1.0)
mean = sum_v / count_v
sum_vv = jnp.sum(
(inputs - mean) * (inputs - mean) * mask,
axis=reduce_over_dims,
keepdims=keepdims) | identifier_body |
normalizations.py | """Normalization layers."""
from typing import List, Optional, Tuple
import jax
from jax import numpy as jnp
from lingvo.jax import base_layer
from lingvo.jax import py_utils
from lingvo.jax import pytypes
NestedMap = py_utils.NestedMap
WeightInit = py_utils.WeightInit
weight_params = py_utils.weight_params
InstantiableParams = py_utils.InstantiableParams
JTensor = pytypes.JTensor
def compute_moments(
inputs: JTensor,
padding: JTensor,
reduce_over_dims: List[int],
enable_cross_replica_sum_on_tpu: bool = False,
keepdims: bool = False,
) -> Tuple[JTensor, JTensor]:
"""Computes mean and variance over the valid data points in inputs."""
assert inputs.ndim == padding.ndim
rank = inputs.ndim
assert all([0 <= dim < rank for dim in reduce_over_dims])
mask = 1.0 - padding
sum_v = jnp.sum(inputs * mask, axis=reduce_over_dims, keepdims=keepdims)
count_v = jnp.sum(
jnp.ones_like(inputs) * mask, axis=reduce_over_dims, keepdims=keepdims)
if enable_cross_replica_sum_on_tpu:
# TODO(shafey, yonghui): Fetch axis_name from globals.
sum_v = jax.lax.psum(sum_v, axis_name='batch')
count_v = jax.lax.psum(count_v, axis_name='batch')
count_v = jnp.maximum(count_v, 1.0)
mean = sum_v / count_v
sum_vv = jnp.sum(
(inputs - mean) * (inputs - mean) * mask,
axis=reduce_over_dims,
keepdims=keepdims)
if enable_cross_replica_sum_on_tpu:
# TODO(shafey, yonghui): Fetch axis_name from globals.
sum_vv = jax.lax.psum(sum_vv, axis_name='batch')
variance = sum_vv / count_v
return mean, variance
class BatchNorm(base_layer.BaseLayer):
"""Batch normalization layer."""
@classmethod
def Params(cls) -> InstantiableParams:
p = super().Params()
p.Define('dim', 0, 'Depth of the input/output.')
p.Define(
'decay', 0.999,
'Decay in updating the mean and variance moving average used in'
' batch normalization.')
p.Define(
'enable_cross_replica_sum_on_tpu', False,
'If true, computes global mean and variance across all replicas.'
'Only effective for tpu.')
p.Define(
'use_moving_avg_in_training', False,
'If True, use global moving avg (mean, variance) during training'
' to avoid mismatch between train and eval, which then'
' essentially acts as an adaptive normalization step. When this is'
' set to True, it also disables the use of beta and gamma variables.')
p.Define('set_padded_output_to_zero', True,
'If True, sets the padded outputs to zero.')
return p
def __init__(self, params: InstantiableParams) -> None:
super().__init__(params)
p = self.params
self._epsilon = 0.001
self._decay = p.decay
def _get_weight_shape(self) -> JTensor:
return [self.params.dim]
def create_layer_variables(self) -> None:
p = self.params
beta_pc = weight_params(
shape=self._get_weight_shape(),
init=WeightInit.Constant(0.0),
dtype=p.dtype)
self.create_variable('beta', beta_pc)
# gamma = theta.gamma + 1.0
gamma_pc = weight_params(
shape=self._get_weight_shape(),
init=WeightInit.Constant(0.0),
dtype=p.dtype)
self.create_variable('gamma', gamma_pc)
mva = weight_params(
shape=[p.dim],
init=WeightInit.Constant(0.0),
dtype=p.dtype,
collections=[base_layer.REQUIRES_MEAN_SYNC])
self.create_variable('moving_mean', mva, trainable=False)
mvv = weight_params(
shape=[p.dim],
init=WeightInit.Constant(1.0),
dtype=p.dtype,
collections=[base_layer.REQUIRES_MEAN_SYNC])
self.create_variable('moving_variance', mvv, trainable=False)
def _get_default_paddings(self, inputs: JTensor) -> JTensor:
"""Gets the default paddings for an input."""
in_shape = list(inputs.shape)
assert len(in_shape) > 1
in_shape[-1] = 1
return jnp.zeros(in_shape, dtype=inputs.dtype)
def _get_beta_gamma(self, theta: NestedMap) -> Tuple[JTensor, JTensor]:
p = self.params
if p.use_moving_avg_in_training:
beta = 0.0
gamma = 1.0
else:
beta = theta.beta
gamma = theta.gamma + 1.0
return beta, gamma
def compute_and_update_moments(
self, theta: NestedMap, inputs: JTensor,
paddings: JTensor) -> Tuple[JTensor, JTensor, JTensor, JTensor]:
"""Computes moments and updates state.
Args:
theta: A `.NestedMap` object containing weights' values of this layer and
its children layers.
inputs: The inputs JTensor. Shaped [..., dim].
paddings: The paddings JTensor. Shaped [..., 1], with the same rank as
the input JTensor.
Returns:
Tuple of (mean, variance, beta, gamma).
"""
p = self.params
if self.do_eval:
# The mean and variance used for normalization.
norm_mean, norm_variance = theta.moving_mean, theta.moving_variance
base_layer.add_summary('moving_mean', theta.moving_mean)
base_layer.add_summary('moving_variance', theta.moving_variance)
else:
rank = inputs.ndim
reduce_over_dims = list(range(0, rank - 1))
mean, variance = compute_moments(
inputs,
paddings,
reduce_over_dims,
enable_cross_replica_sum_on_tpu=p.enable_cross_replica_sum_on_tpu,
keepdims=True)
new_moving_mean = theta.moving_mean * p.decay + mean * (1.0 - p.decay)
self.forward_update_var('moving_mean', new_moving_mean)
new_moving_variance = (
theta.moving_variance * p.decay + variance * (1.0 - p.decay))
self.forward_update_var('moving_variance', new_moving_variance)
# Add some summaries for visualization.
base_layer.add_summary('mean', mean)
base_layer.add_summary('variance', variance)
base_layer.add_summary('moving_mean', theta.moving_mean)
base_layer.add_summary('moving_variance', theta.moving_variance)
if p.use_moving_avg_in_training:
# Use the global statistics for normalization.
norm_mean = theta.moving_mean
norm_variance = theta.moving_variance
else:
# Use the batch statistics for normalization.
norm_mean = mean
norm_variance = variance
beta, gamma = self._get_beta_gamma(theta)
return norm_mean, norm_variance, beta, gamma
def fprop(self,
theta: NestedMap,
inputs: JTensor,
paddings: Optional[JTensor] = None) -> JTensor:
"""Apply batch normalization.
Args:
theta: A `.NestedMap` object containing weights' values of this layer and
its children layers.
inputs: The inputs JTensor. Shaped [..., dim]. | """
p = self.params
inputs, paddings = self._cast_to_fprop_dtype((inputs, paddings))
if paddings is None:
paddings = self._get_default_paddings(inputs)
assert inputs.ndim == paddings.ndim
assert paddings.shape[-1] == 1
norm_mean, norm_variance, beta, gamma = self.compute_and_update_moments(
theta, inputs, paddings)
inv = gamma / jnp.sqrt(norm_variance + self._epsilon)
bn_output = (inputs - norm_mean) * inv + beta
if p.set_padded_output_to_zero:
bn_output *= 1.0 - paddings
return bn_output
class LayerNorm(base_layer.BaseLayer):
"""Layer normalization."""
@classmethod
def Params(cls) -> InstantiableParams:
p = super().Params()
p.Define('input_dims', 0, 'Depth of the input to the network.')
p.Define('epsilon', 1e-6, 'Tiny value to guard rsqrt.')
p.Define('scale', True, 'Whether to use a learned scaling.')
p.Define('bias', True, 'Whether to use bias.')
return p
def create_layer_variables(self) -> None:
super().create_layer_variables()
p = self.params
wp = p.weight_split_dims_mapping
wp_scale = wp.wt
| paddings: The paddings JTensor. Shaped [..., 1].
Returns:
Output after applying batch normalization, with the same shape as
'inputs'. | random_line_split |
normalizations.py | """Normalization layers."""
from typing import List, Optional, Tuple
import jax
from jax import numpy as jnp
from lingvo.jax import base_layer
from lingvo.jax import py_utils
from lingvo.jax import pytypes
NestedMap = py_utils.NestedMap
WeightInit = py_utils.WeightInit
weight_params = py_utils.weight_params
InstantiableParams = py_utils.InstantiableParams
JTensor = pytypes.JTensor
def compute_moments(
inputs: JTensor,
padding: JTensor,
reduce_over_dims: List[int],
enable_cross_replica_sum_on_tpu: bool = False,
keepdims: bool = False,
) -> Tuple[JTensor, JTensor]:
"""Computes mean and variance over the valid data points in inputs."""
assert inputs.ndim == padding.ndim
rank = inputs.ndim
assert all([0 <= dim < rank for dim in reduce_over_dims])
mask = 1.0 - padding
sum_v = jnp.sum(inputs * mask, axis=reduce_over_dims, keepdims=keepdims)
count_v = jnp.sum(
jnp.ones_like(inputs) * mask, axis=reduce_over_dims, keepdims=keepdims)
if enable_cross_replica_sum_on_tpu:
# TODO(shafey, yonghui): Fetch axis_name from globals.
sum_v = jax.lax.psum(sum_v, axis_name='batch')
count_v = jax.lax.psum(count_v, axis_name='batch')
count_v = jnp.maximum(count_v, 1.0)
mean = sum_v / count_v
sum_vv = jnp.sum(
(inputs - mean) * (inputs - mean) * mask,
axis=reduce_over_dims,
keepdims=keepdims)
if enable_cross_replica_sum_on_tpu:
# TODO(shafey, yonghui): Fetch axis_name from globals.
sum_vv = jax.lax.psum(sum_vv, axis_name='batch')
variance = sum_vv / count_v
return mean, variance
class BatchNorm(base_layer.BaseLayer):
"""Batch normalization layer."""
@classmethod
def Params(cls) -> InstantiableParams:
p = super().Params()
p.Define('dim', 0, 'Depth of the input/output.')
p.Define(
'decay', 0.999,
'Decay in updating the mean and variance moving average used in'
' batch normalization.')
p.Define(
'enable_cross_replica_sum_on_tpu', False,
'If true, computes global mean and variance across all replicas.'
'Only effective for tpu.')
p.Define(
'use_moving_avg_in_training', False,
'If True, use global moving avg (mean, variance) during training'
' to avoid mismatch between train and eval, which then'
' essentially acts as an adaptive normalization step. When this is'
' set to True, it also disables the use of beta and gamma variables.')
p.Define('set_padded_output_to_zero', True,
'If True, sets the padded outputs to zero.')
return p
def __init__(self, params: InstantiableParams) -> None:
super().__init__(params)
p = self.params
self._epsilon = 0.001
self._decay = p.decay
def _get_weight_shape(self) -> JTensor:
return [self.params.dim]
def create_layer_variables(self) -> None:
p = self.params
beta_pc = weight_params(
shape=self._get_weight_shape(),
init=WeightInit.Constant(0.0),
dtype=p.dtype)
self.create_variable('beta', beta_pc)
# gamma = theta.gamma + 1.0
gamma_pc = weight_params(
shape=self._get_weight_shape(),
init=WeightInit.Constant(0.0),
dtype=p.dtype)
self.create_variable('gamma', gamma_pc)
mva = weight_params(
shape=[p.dim],
init=WeightInit.Constant(0.0),
dtype=p.dtype,
collections=[base_layer.REQUIRES_MEAN_SYNC])
self.create_variable('moving_mean', mva, trainable=False)
mvv = weight_params(
shape=[p.dim],
init=WeightInit.Constant(1.0),
dtype=p.dtype,
collections=[base_layer.REQUIRES_MEAN_SYNC])
self.create_variable('moving_variance', mvv, trainable=False)
def _get_default_paddings(self, inputs: JTensor) -> JTensor:
"""Gets the default paddings for an input."""
in_shape = list(inputs.shape)
assert len(in_shape) > 1
in_shape[-1] = 1
return jnp.zeros(in_shape, dtype=inputs.dtype)
def | (self, theta: NestedMap) -> Tuple[JTensor, JTensor]:
p = self.params
if p.use_moving_avg_in_training:
beta = 0.0
gamma = 1.0
else:
beta = theta.beta
gamma = theta.gamma + 1.0
return beta, gamma
def compute_and_update_moments(
self, theta: NestedMap, inputs: JTensor,
paddings: JTensor) -> Tuple[JTensor, JTensor, JTensor, JTensor]:
"""Computes moments and updates state.
Args:
theta: A `.NestedMap` object containing weights' values of this layer and
its children layers.
inputs: The inputs JTensor. Shaped [..., dim].
paddings: The paddings JTensor. Shaped [..., 1], with the same rank as
the input JTensor.
Returns:
Tuple of (mean, variance, beta, gamma).
"""
p = self.params
if self.do_eval:
# The mean and variance used for normalization.
norm_mean, norm_variance = theta.moving_mean, theta.moving_variance
base_layer.add_summary('moving_mean', theta.moving_mean)
base_layer.add_summary('moving_variance', theta.moving_variance)
else:
rank = inputs.ndim
reduce_over_dims = list(range(0, rank - 1))
mean, variance = compute_moments(
inputs,
paddings,
reduce_over_dims,
enable_cross_replica_sum_on_tpu=p.enable_cross_replica_sum_on_tpu,
keepdims=True)
new_moving_mean = theta.moving_mean * p.decay + mean * (1.0 - p.decay)
self.forward_update_var('moving_mean', new_moving_mean)
new_moving_variance = (
theta.moving_variance * p.decay + variance * (1.0 - p.decay))
self.forward_update_var('moving_variance', new_moving_variance)
# Add some summaries for visualization.
base_layer.add_summary('mean', mean)
base_layer.add_summary('variance', variance)
base_layer.add_summary('moving_mean', theta.moving_mean)
base_layer.add_summary('moving_variance', theta.moving_variance)
if p.use_moving_avg_in_training:
# Use the global statistics for normalization.
norm_mean = theta.moving_mean
norm_variance = theta.moving_variance
else:
# Use the batch statistics for normalization.
norm_mean = mean
norm_variance = variance
beta, gamma = self._get_beta_gamma(theta)
return norm_mean, norm_variance, beta, gamma
def fprop(self,
theta: NestedMap,
inputs: JTensor,
paddings: Optional[JTensor] = None) -> JTensor:
"""Apply batch normalization.
Args:
theta: A `.NestedMap` object containing weights' values of this layer and
its children layers.
inputs: The inputs JTensor. Shaped [..., dim].
paddings: The paddings JTensor. Shaped [..., 1].
Returns:
Output after applying batch normalization, with the same shape as
'inputs'.
"""
p = self.params
inputs, paddings = self._cast_to_fprop_dtype((inputs, paddings))
if paddings is None:
paddings = self._get_default_paddings(inputs)
assert inputs.ndim == paddings.ndim
assert paddings.shape[-1] == 1
norm_mean, norm_variance, beta, gamma = self.compute_and_update_moments(
theta, inputs, paddings)
inv = gamma / jnp.sqrt(norm_variance + self._epsilon)
bn_output = (inputs - norm_mean) * inv + beta
if p.set_padded_output_to_zero:
bn_output *= 1.0 - paddings
return bn_output
class LayerNorm(base_layer.BaseLayer):
"""Layer normalization."""
@classmethod
def Params(cls) -> InstantiableParams:
p = super().Params()
p.Define('input_dims', 0, 'Depth of the input to the network.')
p.Define('epsilon', 1e-6, 'Tiny value to guard rsqrt.')
p.Define('scale', True, 'Whether to use a learned scaling.')
p.Define('bias', True, 'Whether to use bias.')
return p
def create_layer_variables(self) -> None:
super().create_layer_variables()
p = self.params
wp = p.weight_split_dims_mapping
wp_scale = wp.wt
| _get_beta_gamma | identifier_name |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.