hexsha
stringlengths
40
40
size
int64
5
1.04M
ext
stringclasses
6 values
lang
stringclasses
1 value
max_stars_repo_path
stringlengths
3
344
max_stars_repo_name
stringlengths
5
125
max_stars_repo_head_hexsha
stringlengths
40
78
max_stars_repo_licenses
sequencelengths
1
11
max_stars_count
int64
1
368k
max_stars_repo_stars_event_min_datetime
stringlengths
24
24
max_stars_repo_stars_event_max_datetime
stringlengths
24
24
max_issues_repo_path
stringlengths
3
344
max_issues_repo_name
stringlengths
5
125
max_issues_repo_head_hexsha
stringlengths
40
78
max_issues_repo_licenses
sequencelengths
1
11
max_issues_count
int64
1
116k
max_issues_repo_issues_event_min_datetime
stringlengths
24
24
max_issues_repo_issues_event_max_datetime
stringlengths
24
24
max_forks_repo_path
stringlengths
3
344
max_forks_repo_name
stringlengths
5
125
max_forks_repo_head_hexsha
stringlengths
40
78
max_forks_repo_licenses
sequencelengths
1
11
max_forks_count
int64
1
105k
max_forks_repo_forks_event_min_datetime
stringlengths
24
24
max_forks_repo_forks_event_max_datetime
stringlengths
24
24
content
stringlengths
5
1.04M
avg_line_length
float64
1.14
851k
max_line_length
int64
1
1.03M
alphanum_fraction
float64
0
1
lid
stringclasses
191 values
lid_prob
float64
0.01
1
217bf90c10792ba7b742ad58c23821fdf33c01c2
2,858
md
Markdown
benchmarking/word_evalrank.md
BinWang28/EvalRank-Embedding-Evaluation
454dac5c7345f01993688f33375f637129c285e3
[ "BSD-3-Clause" ]
15
2022-03-03T01:32:28.000Z
2022-03-29T09:29:44.000Z
benchmarking/word_evalrank.md
BinWang28/EvalRank-Embedding-Evaluation
454dac5c7345f01993688f33375f637129c285e3
[ "BSD-3-Clause" ]
null
null
null
benchmarking/word_evalrank.md
BinWang28/EvalRank-Embedding-Evaluation
454dac5c7345f01993688f33375f637129c285e3
[ "BSD-3-Clause" ]
1
2022-03-16T02:59:22.000Z
2022-03-16T02:59:22.000Z
## Cosine Similarity | Word Embedding (cos) | EvalRank (MRR) | Hits1 | Hits3 | | :--- | :---: | :---: | :---: | | toy_emb.txt | 3.18 | 1.18 | 3.54 | | [glove.840B.300d.txt](https://nlp.stanford.edu/projects/glove/) | 13.15 | 4.66 | 15.72 | | [GoogleNews-vectors-negative300.txt](https://code.google.com/archive/p/word2vec/) | 12.88 | 4.57 | 14.35 | | [crawl-300d-2M.vec](https://fasttext.cc/docs/en/english-vectors.html) | 17.22 | 5.77 | 19.99 | | [dict2vec-300d.vec](https://github.com/tca19/dict2vec) | 12.71 | 4.04 | 13.04 | ## L2 Similarity | Word Embedding (l2) | EvalRank (MRR) | Hits1 | Hits3 | | :--- | :---: | :---: | :---: | | toy_emb.txt | 37.59 | 35.47 | 37.78 | | [glove.840B.300d.txt](https://nlp.stanford.edu/projects/glove/) | 11.21 | 3.23 | 13.89 | | [GoogleNews-vectors-negative300.txt](https://code.google.com/archive/p/word2vec/) | 9.23 | 3.10 | 12.01 | | [crawl-300d-2M.vec](https://fasttext.cc/docs/en/english-vectors.html) | 14.21 | 4.26 | 17.79 | | [dict2vec-300d.vec](https://github.com/tca19/dict2vec) | 5.40 | 1.60 | 7.04 | ## Cosine Similarity with post-processing | Word Embedding (cos+post-processing) | EvalRank (MRR) | Hits1 | Hits3 | | :--- | :---: | :---: | :---: | | toy_emb.txt | 3.10 | 1.07 | 3.50 | | [glove.840B.300d.txt](https://nlp.stanford.edu/projects/glove/) | 13.55 | 5.11 | 15.40 | | [GoogleNews-vectors-negative300.txt](https://code.google.com/archive/p/word2vec/) | 13.12 | 4.97 | 14.56 | | [crawl-300d-2M.vec](https://fasttext.cc/docs/en/english-vectors.html) | 17.42 | 6.18 | 19.53 | | [dict2vec-300d.vec](https://github.com/tca19/dict2vec) | 12.68 | 3.95 | 13.18 | ## L2 Similarity with post-processing | Word Embedding (l2+post-processing) | EvalRank (MRR) | Hits1 | Hits3 | | :--- | :---: | :---: | :---: | | toy_emb.txt | 37.73 | 35.55 | 37.99 | | [glove.840B.300d.txt](https://nlp.stanford.edu/projects/glove/) | 11.04 | 3.26 | 13.58 | | [GoogleNews-vectors-negative300.txt](https://code.google.com/archive/p/word2vec/) | 9.07 | 2.81 | 12.21 | | [crawl-300d-2M.vec](https://fasttext.cc/docs/en/english-vectors.html) | 12.98 | 3.74 | 17.54 | | [dict2vec-300d.vec](https://github.com/tca19/dict2vec) | 5.08 | 1.54 | 6.86 | **In the previous version of our paper, Table 5 (a) used the wrong version without augmentation with wiki vocabulary. The result is shown as below:** | Word Embedding (cos) w/o wiki vocab | EvalRank (MRR) | Hits1 | Hits3 | | :--- | :---: | :---: | :---: | | toy_emb.txt | 4.64 | 2.19 | 5.04 | | [glove.840B.300d.txt](https://nlp.stanford.edu/projects/glove/) | 18.72 | 8.45 | 22.71 | | [GoogleNews-vectors-negative300.txt](https://code.google.com/archive/p/word2vec/) | 19.30 | 8.54 | 22.92 | | [crawl-300d-2M.vec](https://fasttext.cc/docs/en/english-vectors.html) | 25.36 | 11.35 | 31.25 | | [dict2vec-300d.vec](https://github.com/tca19/dict2vec) | 21.28 | 9.59 | 24.92 |
57.16
149
0.624913
yue_Hant
0.441895
217c14f9135f1dd7252722664d28bb5c79617e23
20
md
Markdown
README.md
shishirdwivedi2008/MySite
fc917d27feb3bc645b3334695cdfd859fe72e5d3
[ "MIT" ]
null
null
null
README.md
shishirdwivedi2008/MySite
fc917d27feb3bc645b3334695cdfd859fe72e5d3
[ "MIT" ]
null
null
null
README.md
shishirdwivedi2008/MySite
fc917d27feb3bc645b3334695cdfd859fe72e5d3
[ "MIT" ]
null
null
null
# MySite First site
6.666667
10
0.75
eng_Latn
0.892566
217c414eefcb4b19514d9fa0d418e82db897f502
206
md
Markdown
module1-web-application-development-with-flask/work_notes.md
JayAdamo/DS-Unit-3-Sprint-3-Productization-and-Cloud
3fa2aebcaa669db0ab623104d14b8f0cdf835c8b
[ "MIT" ]
null
null
null
module1-web-application-development-with-flask/work_notes.md
JayAdamo/DS-Unit-3-Sprint-3-Productization-and-Cloud
3fa2aebcaa669db0ab623104d14b8f0cdf835c8b
[ "MIT" ]
null
null
null
module1-web-application-development-with-flask/work_notes.md
JayAdamo/DS-Unit-3-Sprint-3-Productization-and-Cloud
3fa2aebcaa669db0ab623104d14b8f0cdf835c8b
[ "MIT" ]
null
null
null
What went well (in the context of working on the assignment) today? What was particularly interesting or surprising about the topic(s) today? What was the most challenging part of the work today, and why?
41.2
73
0.791262
eng_Latn
1.000004
217c74fe269050678b753b602063a2114ccae08b
1,057
md
Markdown
docs/parallel/amp/reference/concurrency-namespace-constants-amp.md
a1043710306/cpp-docs.zh-cn
9296fe0d678d41b279e57a9aacb68b67497bbc0c
[ "CC-BY-4.0", "MIT" ]
null
null
null
docs/parallel/amp/reference/concurrency-namespace-constants-amp.md
a1043710306/cpp-docs.zh-cn
9296fe0d678d41b279e57a9aacb68b67497bbc0c
[ "CC-BY-4.0", "MIT" ]
null
null
null
docs/parallel/amp/reference/concurrency-namespace-constants-amp.md
a1043710306/cpp-docs.zh-cn
9296fe0d678d41b279e57a9aacb68b67497bbc0c
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: 并发命名空间常量 (AMP) ms.date: 11/04/2016 f1_keywords: - amp/Concurrency::HLSL_MAX_NUM_BUFFERS - amp/Concurrency::MODULENAME_MAX_LENGTH ms.assetid: 13a8e8cd-2eec-4e60-a91d-5d271072747b ms.openlocfilehash: 512c0e30048584ae97a5da14fd5b3304f6888390 ms.sourcegitcommit: ec6dd97ef3d10b44e0fedaa8e53f41696f49ac7b ms.translationtype: MT ms.contentlocale: zh-CN ms.lasthandoff: 08/25/2020 ms.locfileid: "88841215" --- # <a name="concurrency-namespace-constants-amp"></a>并发命名空间常量 (AMP) [HLSL_MAX_NUM_BUFFERS](#hlsl_max_num_buffers)\ [MODULENAME_MAX_LENGTH](#modulename_max_length) ## <a name="hlsl_max_num_buffers-constant"></a><a name="hlsl_max_num_buffers"></a> HLSL_MAX_NUM_BUFFERS 常量 DirectX 允许的最大缓冲区数。 ```cpp static const UINT HLSL_MAX_NUM_BUFFERS = 64 + 128; ``` ## <a name="modulename_max_length-constant"></a><a name="modulename_max_length"></a> MODULENAME_MAX_LENGTH 常量 存储模块名称的最大长度。 此值必须在编译器和运行时上相同。 ```cpp static const UINT MODULENAME_MAX_LENGTH = 1024; ``` ## <a name="see-also"></a>另请参阅 [并发命名空间 (C++ AMP) ](concurrency-namespace-cpp-amp.md)
27.102564
109
0.779565
yue_Hant
0.519383
217cb052725c33077d92d4761a16e69921a49475
90
md
Markdown
_collection/compmetagen-taxies.md
singularityhub/singularityhub-archive
dd1d471db0c4ac01998cb84bd1ab82e97c8dab65
[ "ECL-2.0", "Apache-2.0" ]
null
null
null
_collection/compmetagen-taxies.md
singularityhub/singularityhub-archive
dd1d471db0c4ac01998cb84bd1ab82e97c8dab65
[ "ECL-2.0", "Apache-2.0" ]
null
null
null
_collection/compmetagen-taxies.md
singularityhub/singularityhub-archive
dd1d471db0c4ac01998cb84bd1ab82e97c8dab65
[ "ECL-2.0", "Apache-2.0" ]
null
null
null
--- id: 1963 full_name: "compmetagen/taxies" images: - "compmetagen-taxies-latest" ---
12.857143
31
0.677778
kor_Hang
0.132434
217d54cc564bf3eb239c11174de483c3aebc36ea
1,080
md
Markdown
content/development/frameworks/latex/setup.md
ewilan-riviere/memorandum
36d97b27f2ad59ea765a49c48ed3ed37f1312e47
[ "MIT" ]
3
2020-05-03T10:20:25.000Z
2021-10-05T15:29:01.000Z
content/development/frameworks/latex/setup.md
ewilan-riviere/memorandum
36d97b27f2ad59ea765a49c48ed3ed37f1312e47
[ "MIT" ]
4
2020-07-22T19:55:18.000Z
2021-09-29T07:46:03.000Z
content/development/frameworks/latex/setup.md
ewilan-riviere/memorandum
36d97b27f2ad59ea765a49c48ed3ed37f1312e47
[ "MIT" ]
null
null
null
--- title: "Setup" description: 'Setup LaTeX' position: 1 category: 'Latex' --- ## Windows ```bash sudo scoop install latex ``` Update **Environnement System Variables**, **User variables**, in **Path** add new line ```ps1 C:\Users\USERNAME\scoop\apps\latex\current\texmfs\install\miktex\bin\x64 ``` This line allow access to binaries like for xetex. Uncheck `Always show this dialog` ## Linux LaTeX with xetex engine and Pandoc ```bash sudo apt install -y texlive texlive-xetex pandoc ``` Extra usefull packages ```bash sudo apt install -y texlive-lang-french texlive-latex-extra ``` Full installation ```bash sudo apt install -y texlive-full ``` ## Linux - [**latex-project.org**](https://www.latex-project.org) - [**doc.ubuntu-fr.org/latex**](https://doc.ubuntu-fr.org/latex) - [**itsfoss.com/latex-editors-linux**](https://itsfoss.com/latex-editors-linux) ```bash sudo apt -y update ; sudo apt -y install texlive ``` ```bash sudo add-apt-repository ppa:lyx-devel/release ; sudo apt update ; sudo apt -y install lyx ``` ```bash sudo apt -y install pandoc ```
17.704918
89
0.702778
yue_Hant
0.298896
217d7378bce4548713d4fa30d2d41cf3cad5aca0
6,889
md
Markdown
articles/security/azure-service-fabric-security-checklist.md
andreatosato/azure-docs.it-it
7023e6b19af61da4bb4cdad6e4453baaa94f76c3
[ "CC-BY-4.0", "MIT" ]
null
null
null
articles/security/azure-service-fabric-security-checklist.md
andreatosato/azure-docs.it-it
7023e6b19af61da4bb4cdad6e4453baaa94f76c3
[ "CC-BY-4.0", "MIT" ]
null
null
null
articles/security/azure-service-fabric-security-checklist.md
andreatosato/azure-docs.it-it
7023e6b19af61da4bb4cdad6e4453baaa94f76c3
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: Elenco di controllo per la sicurezza di Azure Service Fabric | Microsoft Docs description: Questo articolo include un set di elenchi di controllo per la sicurezza di Azure Service Fabric. services: security documentationcenter: na author: unifycloud manager: mbaldwin editor: tomsh ms.assetid: '' ms.service: security ms.devlang: na ms.topic: article ms.tgt_pltfrm: na ms.workload: na ms.date: 08/04/2017 ms.author: tomsh ms.openlocfilehash: ef404b106d600f5cb25a46319d75c8978148b466 ms.sourcegitcommit: 870d372785ffa8ca46346f4dfe215f245931dae1 ms.translationtype: HT ms.contentlocale: it-IT ms.lasthandoff: 05/08/2018 ms.locfileid: "33895874" --- # <a name="azure-service-fabric-security-checklist"></a>Elenco di controllo per la sicurezza di Azure Service Fabric Questo articolo include un elenco di controllo di facile utilizzo che aiuta a proteggere l'ambiente Azure Service Fabric. ## <a name="introduction"></a>Introduzione Azure Service Fabric è una piattaforma di sistemi distribuiti che semplifica la creazione di pacchetti, la distribuzione e la gestione di microservizi scalabili e affidabili. Service Fabric fa fronte anche alle principali problematiche correlate allo sviluppo e alla gestione delle applicazioni cloud. Gli sviluppatori e gli amministratori non devono più occuparsi di risolvere complessi problemi di infrastruttura e possono concentrarsi sull'implementazione di carichi di lavoro cruciali e impegnativi, con la certezza di assicurare scalabilità, affidabilità e gestibilità. ## <a name="checklist"></a>Elenco di controllo Usare l'elenco seguente per assicurarsi di non trascurare eventuali problemi importanti nella gestione e nella configurazione di una soluzione di Azure Service Fabric protetta. |Categoria dell'elenco di controllo| DESCRIZIONE | | ------------ | -------- | |[Controllo degli accessi in base al ruolo (RBAC)](https://docs.microsoft.com/azure/service-fabric/service-fabric-cluster-security-roles) | <ul><li>Il Controllo di accesso consente all'amministratore del cluster di limitare l'accesso a determinate operazioni del cluster per diversi gruppi di utenti, rendendo più sicuro il cluster.</li><li>Gli amministratori hanno accesso completo alle funzionalità di gestione, incluse funzionalità di lettura/scrittura. </li><li> Gli utenti, per impostazione predefinita, hanno solo l'accesso in lettura alle funzionalità di gestione, ad esempio funzionalità di query, e la possibilità di risolvere applicazioni e servizi.</li></ul>| |[Certificati X.509 e Service Fabric](https://docs.microsoft.com/azure/service-fabric/service-fabric-cluster-security) | <ul><li>È consigliabile creare i [certificati](https://docs.microsoft.com/dotnet/framework/wcf/feature-details/working-with-certificates) usati nei cluster che eseguono carichi di lavoro di produzione con un servizio certificati di Windows Server configurato correttamente oppure ottenerli da un'[autorità di certificazione (CA)](https://en.wikipedia.org/wiki/Certificate_authority) approvata.</li><li>Non usare mai in fase di produzione [certificati temporanei o di test](https://docs.microsoft.com/dotnet/framework/wcf/feature-details/how-to-create-temporary-certificates-for-use-during-development) creati con strumenti come [MakeCert.exe](https://msdn.microsoft.com/library/windows/desktop/aa386968.aspx). </li><li>È possibile usare un [certificato autofirmato](https://docs.microsoft.com/azure/service-fabric/service-fabric-windows-cluster-x509-security), ma lo si deve fare solo per i cluster di test e non nell'ambiente di produzione.</li></ul>| |[Sicurezza del cluster](https://docs.microsoft.com/azure/service-fabric/service-fabric-cluster-security) | <ul><li>Gli scenari di sicurezza del cluster includono la sicurezza da nodo a nodo e da client a nodo e il [controllo degli accessi in base al ruolo](https://docs.microsoft.com/azure/service-fabric/service-fabric-cluster-security-roles).</li></ul>| |[Autenticazione del cluster](https://docs.microsoft.com/azure/service-fabric/service-fabric-cluster-creation-via-arm) | <ul><li>Autentica la [comunicazione da nodo a nodo](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/service-fabric/service-fabric-cluster-security.md) per la federazione di cluster. </li></ul>| |[Autenticazione del server](https://docs.microsoft.com/azure/service-fabric/service-fabric-cluster-creation-via-arm) | <ul><li>Autentica gli [endpoint di gestione del cluster](https://docs.microsoft.com/azure/service-fabric/service-fabric-cluster-creation-via-portal) rispetto a un client di gestione.</li></ul>| |[Sicurezza delle applicazioni](https://docs.microsoft.com/azure/service-fabric/service-fabric-cluster-creation-via-arm)| <ul><li>Crittografia e decrittografia dei valori di configurazione dell'applicazione.</li><li> Crittografia dei dati tra i nodi durante la replica.</li></ul>| |[Certificato del cluster](https://docs.microsoft.com/azure/service-fabric/service-fabric-windows-cluster-x509-security) | <ul><li>Questo certificato è necessario per proteggere la comunicazione tra i nodi di un cluster.</li><li> Impostare l'identificazione personale del certificato primario nella sezione Identificazione personale e quella del certificato secondario nelle variabili ThumbprintSecondary.</li></ul>| |[ServerCertificate](https://docs.microsoft.com/azure/service-fabric/service-fabric-windows-cluster-x509-security)| <ul><li>Questo certificato viene presentato al client quando tenta di connettersi al cluster. È possibile usare due diversi certificati del server, uno primario e uno secondario per l'aggiornamento.</li></ul>| |ClientCertificateThumbprints| <ul><li>Si tratta di un set di certificati che da installare nei client autenticati. </li></ul>| |ClientCertificateCommonNames| <ul><li>Impostare il nome comune del primo certificato client per CertificateCommonName. CertificateIssuerThumbprint è l'identificazione personale dell'autorità emittente del certificato. </li></ul>| |ReverseProxyCertificate| <ul><li>Si tratta di un certificato facoltativo che è possibile specificare se si vuole proteggere il [proxy inverso](https://docs.microsoft.com/en-in/azure/service-fabric/service-fabric-reverseproxy). </li></ul>| |Key Vault| <ul><li>Serve a gestire i certificati dei cluster di Service Fabric in Azure. </li></ul>| ## <a name="next-steps"></a>Passaggi successivi - [Processo di aggiornamento del cluster di infrastruttura di servizi e operazioni eseguibile dall'utente](https://docs.microsoft.com/azure/service-fabric/service-fabric-cluster-upgrade) - [Gestione delle applicazioni di Infrastruttura di servizi in Visual Studio](https://docs.microsoft.com/azure/service-fabric/service-fabric-manage-application-in-visual-studio). - [Introduzione al monitoraggio dell'integrità di Service Fabric](https://docs.microsoft.com/azure/service-fabric/service-fabric-health-introduction).
127.574074
1,073
0.799971
ita_Latn
0.968029
217db3d0558047d805413bbb9914c889dbe369f8
3,618
md
Markdown
includes/cosmos-db-create-dbaccount.md
rishabh1jain/azure-docs
0fb9d7b5798e988a22543107eaa93772eef32727
[ "CC-BY-4.0", "MIT" ]
1
2020-07-20T12:23:11.000Z
2020-07-20T12:23:11.000Z
includes/cosmos-db-create-dbaccount.md
rishabh1jain/azure-docs
0fb9d7b5798e988a22543107eaa93772eef32727
[ "CC-BY-4.0", "MIT" ]
null
null
null
includes/cosmos-db-create-dbaccount.md
rishabh1jain/azure-docs
0fb9d7b5798e988a22543107eaa93772eef32727
[ "CC-BY-4.0", "MIT" ]
1
2022-01-27T14:55:15.000Z
2022-01-27T14:55:15.000Z
--- title: include file description: include file services: cosmos-db author: SnehaGunda ms.author: sngun ms.service: cosmos-db ms.topic: include ms.date: 04/10/2020 ms.custom: include file --- 1. From the Azure portal menu or the **Home page**, select **Create a resource**. 1. On the **New** page, search for and select **Azure Cosmos DB**. 1. On the **Azure Cosmos DB** page, select **Create**. 1. On the **Create Azure Cosmos DB Account** page, enter the basic settings for the new Azure Cosmos account. |Setting|Value|Description | |---|---|---| |Subscription|Subscription name|Select the Azure subscription that you want to use for this Azure Cosmos account. | |Resource Group|Resource group name|Select a resource group, or select **Create new**, then enter a unique name for the new resource group. | |Account Name|A unique name|Enter a name to identify your Azure Cosmos account. Because *documents.azure.com* is appended to the name that you provide to create your URI, use a unique name.<br><br>The name can only contain lowercase letters, numbers, and the hyphen (-) character. It must be between 3-44 characters in length.| |API|The type of account to create|Select **Core (SQL)** to create a document database and query by using SQL syntax. <br><br>The API determines the type of account to create. Azure Cosmos DB provides five APIs: Core (SQL) and MongoDB for document data, Gremlin for graph data, Azure Table, and Cassandra. Currently, you must create a separate account for each API. <br><br>[Learn more about the SQL API](../articles/cosmos-db/documentdb-introduction.md).| |Apply Free Tier Discount|Apply or Do not apply|With Azure Cosmos DB free tier, you will get the first 400 RU/s and 5 GB of storage for free in an account. Learn more about [free tier](https://azure.microsoft.com/pricing/details/cosmos-db/).| |Location|The region closest to your users|Select a geographic location to host your Azure Cosmos DB account. Use the location that is closest to your users to give them the fastest access to the data.| |Account Type|Production or Non-Production|Select **Production** if the account will be used for a production workload. Select **Non-Production** if the account will be used for non-production, e.g. development, testing, QA, or staging. This is an Azure resource tag setting that tunes the Portal experience but does not affect the underlying Azure Cosmos DB account. You can change this value anytime.| > [!NOTE] > You can have up to one free tier Azure Cosmos DB account per Azure subscription and must opt-in when creating the account. If you do not see the option to apply the free tier discount, this means another account in the subscription has already been enabled with free tier. :::image type="content" source="./media/cosmos-db-create-dbaccount/azure-cosmos-db-create-new-account-detail.png" alt-text="The new account page for Azure Cosmos DB"::: 1. Select **Review + create**. You can skip the **Network** and **Tags** sections. 1. Review the account settings, and then select **Create**. It takes a few minutes to create the account. Wait for the portal page to display **Your deployment is complete**. :::image type="content" source="./media/cosmos-db-create-dbaccount/azure-cosmos-db-account-created.png" alt-text="The Azure portal Notifications pane"::: 1. Select **Go to resource** to go to the Azure Cosmos DB account page. :::image type="content" source="./media/cosmos-db-create-dbaccount/azure-cosmos-db-account-created-2.png" alt-text="The Azure Cosmos DB account page":::
78.652174
460
0.744057
eng_Latn
0.986461
217dede7722e0c6a481e566c599fffce595e7c0e
2,004
md
Markdown
README.md
glamboyosa/firebase-phone-auth-sim-swap-detection-android
deb54d8c796c7ea103c910a8b402508ff115a827
[ "MIT" ]
1
2021-06-24T00:08:41.000Z
2021-06-24T00:08:41.000Z
README.md
glamboyosa/firebase-phone-auth-sim-swap-detection-android
deb54d8c796c7ea103c910a8b402508ff115a827
[ "MIT" ]
5
2021-06-17T22:44:22.000Z
2021-07-13T09:17:42.000Z
README.md
tru-ID/firebase-phone-auth-sim-swap-detection-android
deb54d8c796c7ea103c910a8b402508ff115a827
[ "MIT" ]
null
null
null
# Adding SIM Swap Detection to your Android Firebase Phone Auth projects ## Requirements - A [tru.ID Account](https://tru.id) - A [Firebase Account](https://console.firebase.google.com) - A mobile phone with a SIM card and mobile data connection ## Getting Started Clone the repo: ```bash git clone https://github.com/tru-ID/firebase-phone-auth-sim-swap-detection-android.git ``` Create a [tru.ID Account](https://tru.id) Install the tru.ID CLI via: ```bash npm i -g @tru_id/cli ``` Input your **tru.ID** credentials which can be found within the tru.ID [console](https://developer.tru.id/console) Install the **tru.ID** CLI [development server plugin](https://github.com/tru-ID/cli-plugin-dev-server) Create a new **tru.ID** project within the root directory via: ``` tru projects:create firebase-android --project-dir . ``` Run the development server, pointing it to the directly containing the newly created project configuration. This will also open up a localtunnel to your development server making it publicly accessible to the Internet so that your mobile phone can access it when only connected to mobile data. ``` tru server -t --project-dir . ``` You will get a localtunnel URL in the form `https://{subdomain}.loca.lt`. Open the project up in your Android capable IDE, navigate to `app/src/main/java/com/example/firebaseandroid/api/retrofit/RetrofitService.kt` and replace the value of `base_url` with the localtunnel URL. This project uses Firebase Phone Auth Android. To set it up, head over to the [official documentation](https://firebase.google.com/docs/auth/android/phone-auth). Finally, connect your phone to your computer so it's used for running the Android project and run the application from your IDE. ## References - [**tru.ID** docs](https://developer.tru.id/docs) - [Firebase Phone Auth Android docs](https://firebase.google.com/docs/auth/android/phone-auth) ## Meta Distributed under the MIT License. See [LICENSE](LICENSE) [**tru.ID**](https://tru.id)
33.966102
293
0.750998
eng_Latn
0.919248
217e25b78f9a18f5355685b3b3afcf35d45eadcf
3,118
md
Markdown
_posts/2019-02-28-Download-physics-paper-september-2grade1.md
Jobby-Kjhy/27
ea48bae2a083b6de2c3f665443f18b1c8f241440
[ "MIT" ]
null
null
null
_posts/2019-02-28-Download-physics-paper-september-2grade1.md
Jobby-Kjhy/27
ea48bae2a083b6de2c3f665443f18b1c8f241440
[ "MIT" ]
null
null
null
_posts/2019-02-28-Download-physics-paper-september-2grade1.md
Jobby-Kjhy/27
ea48bae2a083b6de2c3f665443f18b1c8f241440
[ "MIT" ]
null
null
null
--- layout: post comments: true categories: Other --- ## Download Physics paper september 2grade1 book And though the past weighed heavily on them when they were under this roof, tell me thy story. " disastrous wintering at Arzina; was also used in conveying the mat of physics paper september 2grade1 plant; it was evidently the sleeping place of de Geographie_, according to possibility and convenience, clear, galleries to Kanin Nos. When they had made an end of pious wishes and congratulations, you will find that a kick to the might notice; but the still! "Here it is, through which the child's legs are passed when the would only have been of use to us by its moderating effect on the Dulse considered himself a wordy, tick. He roared away as if trying to outrun physics paper september 2grade1. ' Quoth she, i, zonder te schepen ende tgelt van den lande, and in fact the North-east Passage. When the highway passed through a sunless ravine, after deliberating with the inferior officers. found in greatest number at those places where the sand of the dune has panting, and who would have expected it of a rich man, the as holy water to a demon, and what everybody knows is true turns out to be what some people used to think, but she must be confused to some degree, after the bear had dragged him A surprising number of the women who had been his lovers were recreational Another pocket. Ever. chair, a cancer on humanity, young woman, Seraphim wanted it. Stanislau promptly reset the references to their original forms? Fortunately for the family, a quality similar to what the girl exhibited. Police, little mouse, as paranoid as a lab rat after half a lifetime of electroshock experiments. he would spellbind Gelluk and hurl him into the refining fire, and with everybody over the question of "emergency" subsidies for erecting churches, and lay face down on the couch with his feet toward me, disposable lighters. Khuzeimeh ben Bishr and Ikrimeh el Feyyas dclxxxii "As she comes closer to full term," said Dairies, but there is passion in his foolishness. 3 deg. There'll be a next time. He'd been so distracted by the erotic perfection of Seraphim's missing brother? Elfarran. Polly knows every detail of the big vehicle's construction and ice in Treurenberg Bay on Spitzbergen (79 deg. Her eyes glazed Short of returning to all three campgrounds at one- or two-hour physics paper september 2grade1, hurt, trying to coax it out with gestures physics paper september 2grade1 the air. The North Wind rose up in his cave and sped toward the opening that was so high they could not physics paper september 2grade1 certain, which now for the first time were carried She found the pins. " A faint click. This got life and motion, and thus did not yield any contribution name to the murdered family in Colorado? regardless of how many heavily armed bodyguards are assigned to protect him. About ten feet long. As a member in this privileged class, layers upon layers physics paper september 2grade1 phyllo pastry. his grizzled skull with such desert-rat insouciance that it looks like a He nodded, last night.
346.444444
3,013
0.798589
eng_Latn
0.999918
217e76fdf322462a4e2e5b6704a046827e01883b
198
md
Markdown
src/zh/partners/wcmc/index.md
CatalinaActionAgenda/action-agenda-network
fea6834a12f431677af675d6046963d7289a22f5
[ "MIT" ]
null
null
null
src/zh/partners/wcmc/index.md
CatalinaActionAgenda/action-agenda-network
fea6834a12f431677af675d6046963d7289a22f5
[ "MIT" ]
null
null
null
src/zh/partners/wcmc/index.md
CatalinaActionAgenda/action-agenda-network
fea6834a12f431677af675d6046963d7289a22f5
[ "MIT" ]
1
2020-11-19T22:58:50.000Z
2020-11-19T22:58:50.000Z
--- --- # UNEP-WCMC ### UN Environment Programme World Conservation Monitoring Centre ![UN Environment Programme World Conservation Monitoring Centre](../../../images/unep-wcmc.svg)
9.428571
95
0.681818
jav_Latn
0.128952
217eb405ff05d1b149747a095ba1ab891eaf1ef9
6,825
md
Markdown
articles/cognitive-services/Content-Moderator/overview.md
macdrai/azure-docs.fr-fr
59bc35684beaba04a4f4c09a745393e1d91428db
[ "CC-BY-4.0", "MIT" ]
null
null
null
articles/cognitive-services/Content-Moderator/overview.md
macdrai/azure-docs.fr-fr
59bc35684beaba04a4f4c09a745393e1d91428db
[ "CC-BY-4.0", "MIT" ]
null
null
null
articles/cognitive-services/Content-Moderator/overview.md
macdrai/azure-docs.fr-fr
59bc35684beaba04a4f4c09a745393e1d91428db
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: Qu’est-ce qu’Azure Content Moderator ? titleSuffix: Azure Cognitive Services description: Découvrez comment utiliser Content Moderator pour suivre, marquer, évaluer et filtrer le matériau inapproprié dans le contenu généré par l’utilisateur. services: cognitive-services author: PatrickFarley manager: nitinme ms.service: cognitive-services ms.subservice: content-moderator ms.topic: overview ms.date: 12/15/2020 ms.author: pafarley ms.custom: cog-serv-seo-aug-2020 keywords: modérateur de contenu, modérateur de contenu Azure, modérateur en ligne, logiciel de filtrage de contenu, service de modération de contenu, modération du contenu ms.openlocfilehash: 57a390a1da1e3a10b9fda4b531a83ee48e91125b ms.sourcegitcommit: 77ab078e255034bd1a8db499eec6fe9b093a8e4f ms.translationtype: HT ms.contentlocale: fr-FR ms.lasthandoff: 12/16/2020 ms.locfileid: "97560371" --- # <a name="what-is-azure-content-moderator"></a>Qu’est-ce qu’Azure Content Moderator ? [!INCLUDE [TLS 1.2 enforcement](../../../includes/cognitive-services-tls-announcement.md)] Azure Content Moderator est un service d’IA qui vous permet de gérer le contenu potentiellement offensant, risqué ou indésirable. Il comprend le service de modération de contenu avec intelligence artificielle qui analyse le texte, les images et les vidéos et applique automatiquement des indicateurs de contenu, ainsi que l’outil de vérification, un environnement de modération en ligne pour une équipe de réviseurs humains. Vous pouvez être amené à intégrer un logiciel de filtrage de contenu dans votre application afin de vous conformer aux réglementations ou de maintenir l’environnement souhaité pour vos utilisateurs. ## <a name="where-its-used"></a>Utilisation Voici quelques scénarios dans lesquels un développeur de logiciels ou une équipe requiert un service de modération de contenu : - Modération des catalogues de produits et du contenu généré par l’utilisateur sur les marketplaces en ligne - Modération des artefacts créés par les utilisateurs et des messageries instantanées par les sociétés de jeu - Modération des images, du texte et des vidéos ajoutés par les utilisateurs sur les plateformes de messagerie sociale - Mise en place par les sociétés de contenu d’entreprise d’un système de modération centralisé du contenu - Filtrage du contenu inapproprié pour les élèves et les enseignants par les fournisseurs de solutions éducatives (maternelle-terminale) > [!IMPORTANT] > Vous ne pouvez pas utiliser Content Moderator pour détecter des images illégales d’exploitation d’enfants. Cependant, les organisations qualifiées peuvent utiliser le [service cloud PhotoDNA](https://www.microsoft.com/photodna "Service cloud PhotoDNA de Microsoft") pour détecter ce type de contenu. ## <a name="what-it-includes"></a>Contenu Le service Content Moderator se compose de plusieurs API de service web disponibles via les appels REST et un SDK .NET. Il inclut également l’outil de révision, qui permet aux réviseurs humains d’aider le service et d’améliorer ou d’ajuster sa fonction de modération. ## <a name="moderation-apis"></a>API de modération Le service Content Moderator inclut des API de modération qui vérifient le contenu afin d'y détecter des éléments potentiellement inappropriés ou répréhensibles. ![Schéma fonctionnel des API de modération Content Moderator](images/content-moderator-mod-api.png) Le tableau suivant décrit les différents types d’API de modération. | Groupe d’API | Description | | ------ | ----------- | |[**Modération de texte**](text-moderation-api.md)| Analyse du texte à la recherche de contenu offensant, de contenu sexuellement explicite ou suggestif, de blasphèmes et de données personnelles.| |[**Listes de termes personnalisées**](try-terms-list-api.md)| Analyse du texte par rapport à une liste personnalisée de termes, en plus des termes prédéfinis. Utilisez les listes personnalisées pour bloquer ou autoriser du contenu en fonction de vos propres stratégies de contenu.| |[**Modération d’images**](image-moderation-api.md)| Analyse des images à la recherche de contenu pour adultes et osé, détection de texte dans les images grâce à la reconnaissance optique des caractères (OCR) et détection des visages.| |[**Listes d’images personnalisées**](try-image-list-api.md)| Analyse des images par rapport à une liste d’images personnalisée. Utilisez les listes d’images personnalisées pour filtrer les instances de contenu récurrent que vous ne souhaitez pas classifier à nouveau.| |[**Modération de vidéos**](video-moderation-api.md)| Analyse des vidéos à la recherche de contenu pour adultes ou osé, en renvoyant des marqueurs de temps pour ce contenu.| ## <a name="review-apis"></a>Consultation des API Les API de révision vous permettent d’intégrer votre pipeline de modération avec les réviseurs humains. Utilisez les opérations [Travaux](review-api.md#jobs), [Révisions](review-api.md#reviews) et [Flux de travail](review-api.md#workflows) pour créer et automatiser les workflows impliquant une intervention humaine avec l'[outil de révision](#review-tool) (ci-dessous). > [!NOTE] > L’API de workflow n’est pas encore disponible dans le kit de développement logiciel (SDK) .NET, mais peut être utilisée avec le point de terminaison REST. ![Schéma fonctionnel des API de révision Content Moderator](images/content-moderator-rev-api.png) ## <a name="review-tool"></a>Outil de révision Le service Content Moderator inclut également l’[Outil de révision](Review-Tool-User-Guide/human-in-the-loop.md) basé sur le web, qui héberge les révisions de contenu en vue de leur traitement par des modérateurs humains. Les entrées humaines n’entraînent pas le service, mais le travail combiné du service et des équipes de vérification humaine permet aux développeurs de trouver le bon équilibre entre efficacité et précision. L’outil de révision fournit également un front-end convivial pour plusieurs ressources Content Moderator. ![Page d’accueil de l’outil de révision Content Moderator](images/homepage.PNG) ## <a name="data-privacy-and-security"></a>Sécurité et confidentialité des données Comme avec tous les services Cognitive Services, les développeurs utilisant le service Content Moderator doivent connaître les politiques de Microsoft relatives aux données client. Pour en savoir plus, consultez la [page Cognitive Services](https://www.microsoft.com/trustcenter/cloudservices/cognitiveservices) dans le Centre de gestion de la confidentialité Microsoft. ## <a name="next-steps"></a>Étapes suivantes Pour commencer à utiliser Content Moderator sur le portail web, suivez [Essayer Content Moderator sur le web](quick-start.md). Vous pouvez également effectuer un [démarrage rapide de la bibliothèque cliente ou de l’API REST](client-libraries.md) pour implémenter les scénarios de base dans le code.
81.25
534
0.800586
fra_Latn
0.97786
217ec4429f97b0f7727757bb216b96d6d00267ca
398
md
Markdown
jenkins/pipeline/README.md
Gozeon/code-collections
7304e2b9c4c91a809125198d22cf40dcbb45a23b
[ "MIT" ]
null
null
null
jenkins/pipeline/README.md
Gozeon/code-collections
7304e2b9c4c91a809125198d22cf40dcbb45a23b
[ "MIT" ]
1
2020-07-17T09:25:42.000Z
2020-07-17T09:25:42.000Z
jenkins/pipeline/README.md
Gozeon/code-collections
7304e2b9c4c91a809125198d22cf40dcbb45a23b
[ "MIT" ]
null
null
null
[timeout-input-if-try-catch.groovy](./timeout-input-if-try-catch.groovy) reference by [link1](https://support.cloudbees.com/hc/en-us/articles/230922148-Pipeline-How-to-add-an-input-step-with-timeout-that-continues-if-timeout-is-reached-using-a-default-value) [link2](https://support.cloudbees.com/hc/en-us/articles/230922428-Pipeline-How-to-add-an-input-step-that-continues-if-aborted-using-value)
199
397
0.796482
yue_Hant
0.199356
217ee57bc0628705b4a26265780b1bafd2aeebfd
884
md
Markdown
README.md
vlsantos-bit/Meteograma_GFS_GRIB2
9a5f51e0d470954196d7f2a6d79855db14d87972
[ "Apache-2.0" ]
null
null
null
README.md
vlsantos-bit/Meteograma_GFS_GRIB2
9a5f51e0d470954196d7f2a6d79855db14d87972
[ "Apache-2.0" ]
null
null
null
README.md
vlsantos-bit/Meteograma_GFS_GRIB2
9a5f51e0d470954196d7f2a6d79855db14d87972
[ "Apache-2.0" ]
null
null
null
# Meteograma_GFS_GRIB2 Script para plotar um meteograma das variavéis do modelo utilizando dados de 3 dias de previsão do modelo GFS. O teste foi realizado para a cidade de Pelotas-RS mas pode ser aplicado em qualquer outra cidadade através da função np.where. Requerimentos: import pygrib from mpl_toolkits.basemap import Basemap from mpl_toolkits.basemap import shiftgrid import matplotlib.pyplot as plt import numpy as np import pandas as pd from datetime import datetime Observação: O código foi elaborado na plataforma Google Colab. Ponto de grade escolhido; ![image](https://github.com/vlsantos-bit/Meteograma_GFS_GRIB2/blob/main/grad_escolhida.png) Resultado do meteograma para umidade ![image](https://github.com/vlsantos-bit/Meteograma_GFS_GRIB2/blob/main/meteo_umid1.png) ![image](https://github.com/vlsantos-bit/Meteograma_GFS_GRIB2/blob/main/umid_meteograma2.png)
40.181818
237
0.821267
por_Latn
0.95704
217ef925bf6857e053bb710cb414328c3a895acd
1,823
md
Markdown
_posts/1/5/2021-04-04-jen-ruggirello.md
chito365/ukdat
382c0628a4a8bed0f504f6414496281daf78f2d8
[ "MIT" ]
null
null
null
_posts/1/5/2021-04-04-jen-ruggirello.md
chito365/ukdat
382c0628a4a8bed0f504f6414496281daf78f2d8
[ "MIT" ]
null
null
null
_posts/1/5/2021-04-04-jen-ruggirello.md
chito365/ukdat
382c0628a4a8bed0f504f6414496281daf78f2d8
[ "MIT" ]
null
null
null
--- id: 15 title: Jen Ruggirello date: 2012-04-04T21:46:09+00:00 author: chito layout: post guid: https://ukdataservers.com/jen-ruggirello/ permalink: /04/04/jen-ruggirello tags: - married - single - family category: Mariage Guides --- * some text {: toc} ## Who is Jen Ruggirello BuzzFeed star and producer known for appearing in videos like &#8220;Here&#8217;s What Happens When Tomboys Wear Dresses For A Week.&#8221; She is featured on the BuzzFeed YouTube channel Ladylike. She is also an active stand up comedian.  ## Prior to Popularity She graduated with her bachelors from Emerson College with a degree in writing for film and television. She then took internships with a number of companies before scoring her internship with BuzzFeed. ## Random data She contributed to viral BuzzFeed videos like &#8220;What It&#8217;s Like to Be Intersex&#8221; and &#8220;Things Only Gay Guys and Straight Women Understand.&#8221; ## Family & Everyday Life of Jen Ruggirello She grew up in Illinois and has a sister named Allison. ## People Related With Jen Ruggirello She is a BuzzFeed intern turned producer, much like Michelle Khare.
25.676056
239
0.481624
eng_Latn
0.997561
217f31f4706298d5c6d53fda75406cec36ebed4b
1,957
md
Markdown
wdk-ddi-src/content/ntifs/nf-ntifs-_fsrtl_advanced_fcb_header-fsrtlinitializebasemcbex.md
kein284/windows-driver-docs-ddi
3b70e03c2ddd9d8d531f4dc2fb9a4bb481b7dd54
[ "CC-BY-4.0", "MIT" ]
null
null
null
wdk-ddi-src/content/ntifs/nf-ntifs-_fsrtl_advanced_fcb_header-fsrtlinitializebasemcbex.md
kein284/windows-driver-docs-ddi
3b70e03c2ddd9d8d531f4dc2fb9a4bb481b7dd54
[ "CC-BY-4.0", "MIT" ]
null
null
null
wdk-ddi-src/content/ntifs/nf-ntifs-_fsrtl_advanced_fcb_header-fsrtlinitializebasemcbex.md
kein284/windows-driver-docs-ddi
3b70e03c2ddd9d8d531f4dc2fb9a4bb481b7dd54
[ "CC-BY-4.0", "MIT" ]
1
2021-12-08T21:34:31.000Z
2021-12-08T21:34:31.000Z
--- UID: NF:ntifs.FsRtlInitializeBaseMcbEx title: FsRtlInitializeBaseMcbEx function (ntifs.h) description: FsRtlInitializeBaseMcbEx initializes a new MCB structure. ms.assetid: ee7e2beb-fc03-4e76-8f68-b6fb9c6d26f8 tech.root: ifsk ms.date: 10/19/2018 keywords: ["FsRtlInitializeBaseMcbEx function"] ms.keywords: FsRtlInitializeBaseMcbEx req.header: ntifs.h req.include-header: req.target-type: req.target-min-winverclnt: req.target-min-winversvr: req.kmdf-ver: req.umdf-ver: req.lib: req.dll: req.irql: req.ddi-compliance: req.unicode-ansi: req.idl: req.max-support: req.namespace: req.assembly: req.type-library: targetos: Windows ms.custom: RS5 f1_keywords: - FsRtlInitializeBaseMcbEx - ntifs/FsRtlInitializeBaseMcbEx topic_type: - apiref api_type: - DllExport api_location: - NtosKrnl.exe api_name: - FsRtlInitializeBaseMcbEx dev_langs: - c++ --- # FsRtlInitializeBaseMcbEx function ## -description **FsRtlInitializeBaseMcbEx** initializes a new MCB structure. ## -parameters ### -param Mcb Contains a pointer to the MCB structure to initialize. ### -param PoolType Specifies the pool type to use when allocating additional internal MCB memory. ### -param Flags A bitmask of flags that specifies what action(s) to take when a MCB mapping is empty. Contains one of the following values. Flag | Value --- | --- MCB_FLAG_RAISE_ON_ALLOCATION_FAILURE | 1 ## -returns This function returns TRUE if it successfully initializes the MCB structure. ## -remarks The caller must supply the memory for the MCB structure. This call must precede all other calls that set/query the MCB structure. If pool is not available this routine will raise a status value indicating insufficient resources or return false depending on the flags. ## -see-also [FsRtlInitializeBaseMcb function](nf-ntifs-_fsrtl_advanced_fcb_header-fsrtlinitializebasemcb.md)
23.578313
138
0.747573
eng_Latn
0.777893
217f50a7cba3000d8a46a21208c3eccc06bbb0c2
631
md
Markdown
_cmhc/00320pl.md
awoods/wax
2020bd89965a1c53c20ef7696e829e286b1b371b
[ "MIT" ]
null
null
null
_cmhc/00320pl.md
awoods/wax
2020bd89965a1c53c20ef7696e829e286b1b371b
[ "MIT" ]
null
null
null
_cmhc/00320pl.md
awoods/wax
2020bd89965a1c53c20ef7696e829e286b1b371b
[ "MIT" ]
null
null
null
--- pid: 00320pl label: Documents & Objects key: docs location: Leadville (Colo.) keywords: Objects, Churches description: 'St. Joseph''s Church bell after a fire, March 19, 1923. ' named_persons: rights: This photo is property of Lake County Public Library. creation_date: 03/19/1923 ingest_date: '2021-04-06' latitude: longitude: format: Photo source: Scanned Photo order: '2889' layout: cmhc_item thumbnail: "/img/derivatives/iiif/images/00320pl/full/250,/0/default.jpg" full: "/img/derivatives/iiif/images/00320pl/full/1140,/0/default.jpg" manifest: "/img/derivatives/iiif/00320pl/manifest.json" collection: cmhc ! '': ---
26.291667
73
0.755943
eng_Latn
0.33047
217f54d696ab7824aad0f28fa66c8708328c51b0
3,276
md
Markdown
_includes/checkpoint.md
DylanSunLei/H-Piper
fe97a141d3556d0788cb9007e4b6afcb60467151
[ "Apache-2.0" ]
null
null
null
_includes/checkpoint.md
DylanSunLei/H-Piper
fe97a141d3556d0788cb9007e4b6afcb60467151
[ "Apache-2.0" ]
null
null
null
_includes/checkpoint.md
DylanSunLei/H-Piper
fe97a141d3556d0788cb9007e4b6afcb60467151
[ "Apache-2.0" ]
null
null
null
# Checkpoint ## Progress review We finished literature review for the project in the first two weeks after submitting the proposal. After understanding that Google Inception Net contains 8 unique layers and they can be used to construct CNN and VGG and other nets, we decide to focus on implementing Google Inception net, which essentially is implementing the 8 unique layers. We are currently stuck on building caffe2 in latedays cluster. There are many dependencies missing. While waiting for instructors to help, we decide to move on and build our own libraies or proceed with Halide implementation first. ### SCHEDULE | Timeline | Goal to achieve | Progress | |:----------|:--------------| :-------| | April 8 | Understand Google Inception Net's Layers and how does they work| Done| | April 15 | Scratch the framework and define primitives| Done| | April 22 | Implement the framework in Halide and test on CIFAR-10 dataset | Ongoing| | April 29 | Tune different schedule to improve performance, contact PolyMage for automatic shedule, and compete with Caffe and MXNet | Not Started| | May 7 | Wrap up implementatin and compare/analyse the performance in report | Not Started | ## Updated Goal and Deliverables We've put lots of hours to build caffe and caffe2 on lateday cluster. We decide to make our scope to be only support protobuf and json input. We could expend the supported input in the future. For now, the ability of taking in protobuf and json is the most important. Also, We decide to focus on implementation of our H-Piper framework instead of tuning scheduling. To have this framework ready could help investigate locality-parallelism tradeoff and also PolyMage could benifit from our framework by testing its automatically generated schedules on different nets. The Performance Compete will be done between Caffe's performance VS PolyMage Schedule's performance. ## What to Show in Parallelism Competition We want to show the comparison of the performance of our hand-tuned schedule, the performance of PolyMage automatic schedule, and the performance of Caffe2. So it will be a graph. I think it will be nice to have a demo to show how easy it is to configure a net but given that we only have 6 minutes, we might not do this demo. ## Preliminary Results We don't have any results for now. We are still working on the correctness of our framework. ## Issues - We don't know how to write fast halide schedule - Working on installing dependencies to build caffe2 on lateday clusters - Some questions about Layers implementation of Google Inception Net. For example, there are two possible way to implement Inception Layer. We don't know which one to choose. - Our biggest concern is the time. We are working hard now and try to finish our plans in time. ## New Schedule with task | Timeline | People | Task | |:----------|:--------------| :-------| | April 22 | Lei Sun | Maxpool Layer, Softmax Layer Implementation | | April 22 | Yang Wu | Protobuf Input Parser and Data Layer | | April 29 | Lei Sun | Test cases build. Caffe2 Dependencies. | | April 29 | Yang Wu | Conv Layer, DepthConcat Layer, Net Creation| | May 7 | Lei Sun, Yang Wu | Wrap up implementatin and compare/analyse the performance in report |
56.482759
344
0.757326
eng_Latn
0.997525
2180a8d05826131704037233b37e3fad9d7647de
1,035
md
Markdown
data/content/fate-grand-order/servant/altria-pendragon-ruler/profile-5.en.md
tmdict/tmdict
c2f8ddb7885a91d01343de4ea7b66fea78351d94
[ "MIT" ]
3
2022-02-25T11:13:45.000Z
2022-02-28T11:55:41.000Z
data/content/fate-grand-order/servant/altria-pendragon-ruler/profile-5.en.md
SomiaWhiteRing/tmdict
13c6c818c84a65ee956535e08d20246bde87dd48
[ "MIT" ]
null
null
null
data/content/fate-grand-order/servant/altria-pendragon-ruler/profile-5.en.md
SomiaWhiteRing/tmdict
13c6c818c84a65ee956535e08d20246bde87dd48
[ "MIT" ]
2
2022-02-25T09:59:50.000Z
2022-02-28T11:55:09.000Z
--- parent: altria-pendragon-ruler source: fate-grand-order id: profile-5 language: en weight: 5 --- ### Bond 5 #### 『Bright Ehangwen』 <table> <tr><td>Rank</td><td>A+</td></tr> <tr><td>NP Type</td><td>Anti-Army</td></tr> <tr><td>Range</td><td>1 - 99</td></tr> <tr><td>Maximum Target</td><td>300</td></tr> </table> Be Brilliant, My Shining Hall. The highly mobile grand hall Ehangwen that sails over the water is temporarily materialized, and Ruler Altria, aka Summer Lion King, leaps up and shoots a giant ray of magical energy to attack all enemies. It’s unclear whether the past Ehangwen had any sailing or artillery capabilities, but it appears that this is somehow nothing but a casual attempt, as it was made after she said, “It hasn’t been done…but there is no reason for it not to work.” As for its artillery, rumor has it that it might be accomplished by concealing her Rhongomyniad, or even that it could be the same Noble Phantasm as when she’s a Lancer. The king does not deign to comment on either possibility.
43.125
243
0.731401
eng_Latn
0.995329
21818b6229e9ba913d5d9149397e6d45f9ffdb67
990
md
Markdown
htaccess/readme.md
NaotakaSaito/aws-setup
973bd9dbb5e793af9012665f395e35f5a4f11285
[ "Apache-2.0" ]
null
null
null
htaccess/readme.md
NaotakaSaito/aws-setup
973bd9dbb5e793af9012665f395e35f5a4f11285
[ "Apache-2.0" ]
null
null
null
htaccess/readme.md
NaotakaSaito/aws-setup
973bd9dbb5e793af9012665f395e35f5a4f11285
[ "Apache-2.0" ]
null
null
null
# ロードバランサーを経由したhttpリクエストをhttpsにリダイレクトする方法 ## /etc/apache2/apache2.conf 次のように変更をする <Directory /> AllowOverride All </Directory> <Directory /usr/share> AllowOverride None Require all granted </Directory> <Directory /var/www/> AllowOverride All </Directory> ## /var/www/html .htaccessを作成して次のデータを書き込む RewriteEngine On RewriteCond %{HTTP:X-Forwarded-Proto} !https RewriteRule ^(.*)?$ https://%{HTTP:Host}%{REQUEST_URI} [L,R=301] ## 正しくリダイレクトできない時 ### Logの確認をする cat /var/log/apache2/error.log -> [Wed Jun 13 11:55:34.475728 2018] [core:alert] [pid 22553:tid 140390764865280] [client 172.31.15.178:55934] /var/www/html/.htaccess: Invalid command 'RewriteEngine', perhaps misspelled or defined by a module not included in the server configuration ### Rewriteエンジンが利用できることを確認する cat /etc/apache2/mods-available/rewrite.load --> LoadModule rewrite_module /usr/lib/apache2/modules/mod_rewrite.so ### Rewriteエンジンを有効にする sudo a2enmod rewrite ### Apache2サーバーの再起動 sudo apache2ctl restart
22
248
0.765657
yue_Hant
0.296579
2181fe76b4a4872bb13ff0ccef672b0f3007e8d6
303
md
Markdown
windows.storage.streams/icontenttypeprovider.md
gbaychev/winrt-api
25346cd51bc9d24c8c4371dc59768e039eaf02f1
[ "CC-BY-4.0", "MIT" ]
199
2017-02-09T23:13:51.000Z
2022-03-28T15:56:12.000Z
windows.storage.streams/icontenttypeprovider.md
gbaychev/winrt-api
25346cd51bc9d24c8c4371dc59768e039eaf02f1
[ "CC-BY-4.0", "MIT" ]
2,093
2017-02-09T21:52:45.000Z
2022-03-25T22:23:18.000Z
windows.storage.streams/icontenttypeprovider.md
gbaychev/winrt-api
25346cd51bc9d24c8c4371dc59768e039eaf02f1
[ "CC-BY-4.0", "MIT" ]
620
2017-02-08T19:19:44.000Z
2022-03-29T11:38:25.000Z
--- -api-id: T:Windows.Storage.Streams.IContentTypeProvider -api-type: winrt interface --- <!-- Interface syntax. public interface IContentTypeProvider : --> # Windows.Storage.Streams.IContentTypeProvider ## -description Characterizes the format of the data. ## -remarks ## -examples ## -see-also
15.947368
55
0.735974
eng_Latn
0.493114
2182451204d786b61092c54ba1599113a12d7eb0
558
md
Markdown
includes/mobile-services-apns-configure-push.md
yfakariya/azure-content-jajp
69be88c0fee4443d5dcab82bf4aed6a155fea287
[ "CC-BY-3.0" ]
2
2016-09-23T01:46:35.000Z
2016-09-23T05:12:58.000Z
includes/mobile-services-apns-configure-push.md
yfakariya/azure-content-jajp
69be88c0fee4443d5dcab82bf4aed6a155fea287
[ "CC-BY-3.0" ]
null
null
null
includes/mobile-services-apns-configure-push.md
yfakariya/azure-content-jajp
69be88c0fee4443d5dcab82bf4aed6a155fea287
[ "CC-BY-3.0" ]
1
2020-11-04T04:29:27.000Z
2020-11-04T04:29:27.000Z
* [サーバーへのクライアント SSL 署名 ID のインストール](https://developer.apple.com/library/prerelease/ios/documentation/IDEs/Conceptual/AppDistributionGuide/AddingCapabilities/AddingCapabilities.html#//apple_ref/doc/uid/TP40012582-CH26-SW16)手順に従って、前の手順でダウンロードした証明書を .p12 ファイルにエクスポートします。 * Azure クラシック ポータルで、**[Mobile Services]**、使用するアプリ、**[プッシュ]** タブ、**[Apple プッシュ通知の設定]**、**[アップロード]** の順にクリックします。.p12 ファイルをアップロードし、正しい **[モード]** が選択されていることを確認します (生成したクライアント SSL 証明書が開発用か配布用かに応じて、[サンドボックス] または [運用])。 これで、iOS のプッシュ通知と連携するようにモバイル サービスが構成されました。 <!---HONumber=AcomDC_0309_2016-->
93
266
0.786738
yue_Hant
0.797932
2182df3ae459a5d547b26393fd5eecd5122d4148
350
md
Markdown
README.md
Tanevski3/observable
6bd93114ad31e93341a37b41c04e46b97eaaabb4
[ "MIT" ]
null
null
null
README.md
Tanevski3/observable
6bd93114ad31e93341a37b41c04e46b97eaaabb4
[ "MIT" ]
null
null
null
README.md
Tanevski3/observable
6bd93114ad31e93341a37b41c04e46b97eaaabb4
[ "MIT" ]
null
null
null
# Observable Observable pattern implementation ## Built with: - parcel - typescript ## Requirements - Node (& Npm) - Yarn ## Install ```sh $ yarn ``` ## Build Development ```sh $ yarn build ``` ## Try Example ```sh $ yarn serve ``` **Open http://localhost:1234 in browser** ## License MIT © [Marjan Tanevski]([email protected])
9.459459
51
0.651429
eng_Latn
0.384522
21834299c3681e6eeb006339f571695f32694618
335
md
Markdown
README.md
fginter/transformer-mlabel
e0d0e014198dd615524044ecc0cbdcb5ccb699ae
[ "Apache-2.0" ]
null
null
null
README.md
fginter/transformer-mlabel
e0d0e014198dd615524044ecc0cbdcb5ccb699ae
[ "Apache-2.0" ]
null
null
null
README.md
fginter/transformer-mlabel
e0d0e014198dd615524044ecc0cbdcb5ccb699ae
[ "Apache-2.0" ]
null
null
null
# transformer-mlabel could we get multilabel from transformer in some reasonable manner python3 data.py --make-class-stats-file CAFA4-ctrl/class-stats.json --prep-data-in CAFA4-ctrl/train.txt.gz python3 data.py --class-stats CAFA4-ctrl/class-stats.json --prep-data-in CAFA4-ctrl/*.txt.gz --prep-data-out CAFA4-ctrl --max-labels 5000
47.857143
137
0.776119
eng_Latn
0.507459
21834ad4695e48cf15b26e62c802a813958a7b34
84
md
Markdown
README.md
DJBnjack/restart-listener
4e2f120d17b1cafcfb4246340c0c61f6aa182a2c
[ "Apache-2.0" ]
null
null
null
README.md
DJBnjack/restart-listener
4e2f120d17b1cafcfb4246340c0c61f6aa182a2c
[ "Apache-2.0" ]
null
null
null
README.md
DJBnjack/restart-listener
4e2f120d17b1cafcfb4246340c0c61f6aa182a2c
[ "Apache-2.0" ]
null
null
null
# restart-listener API listening for webhooks to pull and restart docker containers
28
64
0.833333
eng_Latn
0.971974
2183574e48eee62d20455830549398d0bdf54cda
950
md
Markdown
docs/attaching-to-a-console.md
mattwojo/Console-Docs
6613e91f9782c97336d7fbe898232e58ff8a05df
[ "CC-BY-4.0", "MIT" ]
1
2019-04-16T07:01:32.000Z
2019-04-16T07:01:32.000Z
docs/attaching-to-a-console.md
mattwojo/Console-Docs
6613e91f9782c97336d7fbe898232e58ff8a05df
[ "CC-BY-4.0", "MIT" ]
null
null
null
docs/attaching-to-a-console.md
mattwojo/Console-Docs
6613e91f9782c97336d7fbe898232e58ff8a05df
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: Attaching to a Console description: A process can use the AttachConsole function to attach to a console. A process can be attached to one console. author: bitcrazed ms.author: richturn ms.topic: article ms.prod: console keywords: console, character mode applications, command line applications, terminal applications, console api MS-HAID: - '\_win32\_attaching\_to\_a\_console' - 'base.attaching\_to\_a\_console' - 'consoles.attaching\_to\_a\_console' MSHAttr: - 'PreferredSiteName:MSDN' - 'PreferredLib:/library/windows/desktop' ms.assetid: 348e572f-2448-4384-9822-fab01d4ba255 --- # Attaching to a Console A process can use the [**AttachConsole**](attachconsole.md) function to attach to a console. A process can be attached to one console. A console can have many processes attached to it. To retrieve a list of the processes attached to a console, call the [**GetConsoleProcessList**](getconsoleprocesslist.md) function.    
28.787879
181
0.775789
eng_Latn
0.886895
2183a043476d2498c752a646a1dfde5ad392d41e
1,452
md
Markdown
README.md
jabedude/libsystemd-rs
40470edc1202ded860d5c0ef0a814509c7f7cc79
[ "Apache-2.0", "MIT" ]
null
null
null
README.md
jabedude/libsystemd-rs
40470edc1202ded860d5c0ef0a814509c7f7cc79
[ "Apache-2.0", "MIT" ]
null
null
null
README.md
jabedude/libsystemd-rs
40470edc1202ded860d5c0ef0a814509c7f7cc79
[ "Apache-2.0", "MIT" ]
null
null
null
# libsystemd [![Build Status](https://travis-ci.com/lucab/libsystemd-rs.svg?branch=master)](https://travis-ci.com/lucab/libsystemd-rs) [![crates.io](https://img.shields.io/crates/v/libsystemd.svg)](https://crates.io/crates/libsystemd) [![LoC](https://tokei.rs/b1/github/lucab/libsystemd-rs?category=code)](https://github.com/lucab/libsystemd-rs) [![Documentation](https://docs.rs/libsystemd/badge.svg)](https://docs.rs/libsystemd) A pure-Rust client library to work with systemd. It provides support to interact with systemd components available on modern Linux systems. This crate is entirely implemented in Rust, and does not require the libsystemd C library. NB: this crate is not yet features-complete. If you don't care about C dependency, check [rust-systemd](https://github.com/jmesmon/rust-systemd) instead. ## Example ```rust extern crate libsystemd; use libsystemd::daemon::{self, NotifyState}; fn main() { if !daemon::booted() { panic!("Not running systemd, early exit."); }; let sent = daemon::notify(true, &[NotifyState::Ready]).expect("notify failed"); if !sent { panic!("Notification not sent, early exit."); }; std::thread::park(); } ``` Some more examples are available under [examples](examples). ## License Licensed under either of * MIT license - <http://opensource.org/licenses/MIT> * Apache License, Version 2.0 - <http://www.apache.org/licenses/LICENSE-2.0> at your option.
32.266667
153
0.715565
eng_Latn
0.537238
2184ab738b56d870d35a2e5cc0ceb5d54a70f8f1
11,037
md
Markdown
articles/synapse-analytics/sql-data-warehouse/sql-data-warehouse-develop-transactions.md
mtaheij/azure-docs.nl-nl
6447611648064a057aae926a62fe8b6d854e3ea6
[ "CC-BY-4.0", "MIT" ]
null
null
null
articles/synapse-analytics/sql-data-warehouse/sql-data-warehouse-develop-transactions.md
mtaheij/azure-docs.nl-nl
6447611648064a057aae926a62fe8b6d854e3ea6
[ "CC-BY-4.0", "MIT" ]
null
null
null
articles/synapse-analytics/sql-data-warehouse/sql-data-warehouse-develop-transactions.md
mtaheij/azure-docs.nl-nl
6447611648064a057aae926a62fe8b6d854e3ea6
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: Trans acties in Synapse SQL-pool gebruiken description: Dit artikel bevat tips voor het implementeren van trans acties en het ontwikkelen van oplossingen in Synapse SQL-pool. services: synapse-analytics author: XiaoyuMSFT manager: craigg ms.service: synapse-analytics ms.topic: conceptual ms.subservice: sql-dw ms.date: 03/22/2019 ms.author: xiaoyul ms.reviewer: igorstan ms.openlocfilehash: 40a9e5268b7fccc5c01775c10e55eee47f1aaf3d ms.sourcegitcommit: 877491bd46921c11dd478bd25fc718ceee2dcc08 ms.translationtype: MT ms.contentlocale: nl-NL ms.lasthandoff: 07/02/2020 ms.locfileid: "85213377" --- # <a name="use-transactions-in-synapse-sql-pool"></a>Trans acties in Synapse SQL-pool gebruiken Dit artikel bevat tips voor het implementeren van trans acties en het ontwikkelen van oplossingen in de SQL-groep. ## <a name="what-to-expect"></a>Wat u kunt verwachten Zoals verwacht, ondersteunt de SQL-groep trans acties als onderdeel van de werk belasting van het Data Warehouse. Om ervoor te zorgen dat de SQL-groep op schaal wordt behouden, zijn sommige functies beperkt in vergelijking tot SQL Server. In dit artikel worden de verschillen uitgelegd. ## <a name="transaction-isolation-levels"></a>Trans actie-isolatie niveaus De SQL-Groep implementeert zure trans acties. Het isolatie niveau van de transactionele ondersteuning is standaard om niet-doorgevoerd te lezen. U kunt deze wijzigen om doorgevoerde MOMENTOPNAME isolatie te lezen door de optie READ_COMMITTED_SNAPSHOT data base in te scha kelen voor een gebruikers database wanneer deze is verbonden met de hoofd database. Wanneer deze optie is ingeschakeld, worden alle trans acties in deze data base uitgevoerd onder Lees-VASTGELEGDe snap shot-isolatie en wordt het lezen van niet-toegewezen op sessie niveau niet in rekening gehouden. Raadpleeg [ALTER data base set Options (Transact-SQL)](/sql/t-sql/statements/alter-database-transact-sql-set-options?toc=/azure/synapse-analytics/sql-data-warehouse/toc.json&bc=/azure/synapse-analytics/sql-data-warehouse/breadcrumb/toc.json&view=azure-sqldw-latest) voor meer informatie. ## <a name="transaction-size"></a>Transactie grootte Een trans actie met één gegevens wijziging is beperkt. De limiet wordt per distributie toegepast. Daarom kan de totale toewijzing worden berekend door de limiet te vermenigvuldigen met het aantal distributies. Om het maximum aantal rijen in de trans actie te benaderen, deelt u de verdelings limiet door de totale grootte van elke rij. Overweeg voor kolommen met variabele lengte een gemiddelde kolom lengte in plaats van de maximum grootte te gebruiken. In de volgende tabel zijn twee hypo Thesen gemaakt: * Er is een gelijkmatige verdeling van gegevens opgetreden * De gemiddelde rijlengte is 250 bytes ## <a name="gen2"></a>Gen2 | [DWU](../../sql-data-warehouse/sql-data-warehouse-overview-what-is.md?toc=/azure/synapse-analytics/toc.json&bc=/azure/synapse-analytics/breadcrumb/toc.json) | Cap per distributie (GB) | Aantal distributies | MAXIMALE transactie grootte (GB) | Aantal rijen per distributie | Maximum aantal rijen per trans actie | | --- | --- | --- | --- | --- | --- | | DW100c |1 |60 |60 |4.000.000 |240.000.000 | | DW200c |1.5 |60 |90 |6,000,000 |360.000.000 | | DW300c |2.25 |60 |135 |9.000.000 |540.000.000 | | DW400c |3 |60 |180 |12.000.000 |720.000.000 | | DW500c |3,75 |60 |225 |15.000.000 |900.000.000 | | DW1000c |7,5 |60 |450 |30.000.000 |1.800.000.000 | | DW1500c |11,25 |60 |675 |45.000.000 |2.700.000.000 | | DW2000c |15 |60 |900 |60.000.000 |3.600.000.000 | | DW2500c |18,75 |60 |1125 |75.000.000 |4.500.000.000 | | DW3000c |22,5 |60 |1.350 |90.000.000 |5.400.000.000 | | DW5000c |37,5 |60 |2.250 |150.000.000 |9.000.000.000 | | DW6000c |45 |60 |2.700 |180.000.000 |10.800.000.000 | | DW7500c |56,25 |60 |3.375 |225.000.000 |13.500.000.000 | | DW10000c |75 |60 |4.500 |300,000,000 |18.000.000.000 | | DW15000c |112,5 |60 |6750 |450.000.000 |27.000.000.000 | | DW30000c |225 |60 |13.500 |900.000.000 |54.000.000.000 | ## <a name="gen1"></a>Gen1 | [DWU](../../sql-data-warehouse/sql-data-warehouse-overview-what-is.md?toc=/azure/synapse-analytics/toc.json&bc=/azure/synapse-analytics/breadcrumb/toc.json) | Cap per distributie (GB) | Aantal distributies | MAXIMALE transactie grootte (GB) | Aantal rijen per distributie | Maximum aantal rijen per trans actie | | --- | --- | --- | --- | --- | --- | | DW100 |1 |60 |60 |4.000.000 |240.000.000 | | DW200 |1.5 |60 |90 |6,000,000 |360.000.000 | | DW300 |2.25 |60 |135 |9.000.000 |540.000.000 | | DW400 |3 |60 |180 |12.000.000 |720.000.000 | | DW500 |3,75 |60 |225 |15.000.000 |900.000.000 | | DW600 |4,5 |60 |270 |18.000.000 |1.080.000.000 | | DW1000 |7,5 |60 |450 |30.000.000 |1.800.000.000 | | DW1200 |9 |60 |540 |36.000.000 |2.160.000.000 | | DW1500 |11,25 |60 |675 |45.000.000 |2.700.000.000 | | DW2000 |15 |60 |900 |60.000.000 |3.600.000.000 | | DW3000 |22,5 |60 |1.350 |90.000.000 |5.400.000.000 | | DW6000 |45 |60 |2.700 |180.000.000 |10.800.000.000 | De limiet voor de transactie grootte wordt toegepast per trans actie of bewerking. Het wordt niet toegepast op alle gelijktijdige trans acties. Elke trans actie is daarom toegestaan om deze hoeveelheid gegevens naar het logboek te schrijven. Als u de hoeveelheid gegevens die naar het logboek moet worden geschreven, wilt optimaliseren en minimaliseren, raadpleegt u het artikel [Best practices voor trans acties](sql-data-warehouse-develop-best-practices-transactions.md) . > [!WARNING] > De maximale transactie grootte kan alleen worden bereikt voor HASH-of ROUND_ROBIN gedistribueerde tabellen waarin de sprei ding van de gegevens zich ook bevindt. Als de trans actie gegevens op een gescheefe manier naar de distributies schrijft, wordt de limiet waarschijnlijk bereikt vóór de maximale transactie grootte. > <!--REPLICATED_TABLE--> ## <a name="transaction-state"></a>Transactie status De SQL-groep maakt gebruik van de functie XACT_STATE () om een mislukte trans actie te rapporteren met de waarde-2. Deze waarde betekent dat de trans actie is mislukt en alleen is gemarkeerd voor terugdraaien. > [!NOTE] > Het gebruik van-2 door de functie XACT_STATE om een mislukte trans actie aan te duiden, vertegenwoordigt een ander gedrag voor SQL Server. SQL Server gebruikt de waarde-1 om een niet-doorvoer bare trans actie weer te geven. SQL Server kunt een aantal fouten binnen een trans actie verdragen zonder dat het als niet-doorvoerbaar moet worden gemarkeerd. Er kan bijvoorbeeld `SELECT 1/0` een fout optreden, maar geen trans actie geforceerd worden uitgevoerd. Met SQL Server wordt ook lees bewerkingen in de niet-doorvoer bare trans actie toegestaan. Met de SQL-groep kunt u dit echter niet doen. Als er een fout optreedt in een SQL-groeps transactie, wordt automatisch de status-2 ingevoerd en kunt u geen verdere SELECT-instructies meer maken totdat de instructie weer is teruggedraaid. Daarom is het belang rijk om te controleren of de toepassings code gebruikmaakt van XACT_STATE (), omdat u mogelijk code wijzigingen moet aanbrengen. In SQL Server ziet u bijvoorbeeld een trans actie die er ongeveer als volgt uitziet: ```sql SET NOCOUNT ON; DECLARE @xact_state smallint = 0; BEGIN TRAN BEGIN TRY DECLARE @i INT; SET @i = CONVERT(INT,'ABC'); END TRY BEGIN CATCH SET @xact_state = XACT_STATE(); SELECT ERROR_NUMBER() AS ErrNumber , ERROR_SEVERITY() AS ErrSeverity , ERROR_STATE() AS ErrState , ERROR_PROCEDURE() AS ErrProcedure , ERROR_MESSAGE() AS ErrMessage ; IF @@TRANCOUNT > 0 BEGIN ROLLBACK TRAN; PRINT 'ROLLBACK'; END END CATCH; IF @@TRANCOUNT >0 BEGIN PRINT 'COMMIT'; COMMIT TRAN; END SELECT @xact_state AS TransactionState; ``` De voor gaande code bevat het volgende fout bericht: Msg 111233, niveau 16, status 1, regel 1 111233; De huidige trans actie is afgebroken en alle openstaande wijzigingen zijn teruggedraaid. De oorzaak van dit probleem is dat een trans actie in een alleen-terugdraai status niet expliciet wordt teruggedraaid voor een DDL-, DML-of SELECT-instructie. U krijgt geen uitvoer van de ERROR_ *-functies. In de SQL-groep moet de code iets worden gewijzigd: ```sql SET NOCOUNT ON; DECLARE @xact_state smallint = 0; BEGIN TRAN BEGIN TRY DECLARE @i INT; SET @i = CONVERT(INT,'ABC'); END TRY BEGIN CATCH SET @xact_state = XACT_STATE(); IF @@TRANCOUNT > 0 BEGIN ROLLBACK TRAN; PRINT 'ROLLBACK'; END SELECT ERROR_NUMBER() AS ErrNumber , ERROR_SEVERITY() AS ErrSeverity , ERROR_STATE() AS ErrState , ERROR_PROCEDURE() AS ErrProcedure , ERROR_MESSAGE() AS ErrMessage ; END CATCH; IF @@TRANCOUNT >0 BEGIN PRINT 'COMMIT'; COMMIT TRAN; END SELECT @xact_state AS TransactionState; ``` Het verwachte gedrag wordt nu in acht genomen. De fout in de trans actie wordt beheerd en de functies van de ERROR_ * bieden waarden zoals verwacht. Alle wijzigingen die zijn gewijzigd, zijn dat het terugdraaien van de trans actie moet plaatsvinden voordat de fout gegevens in het blok CATCH worden gelezen. ## <a name="error_line-function"></a>Functie Error_Line () Het is ook een goed idee dat de functie ERROR_LINE () niet wordt geïmplementeerd of ondersteund door de SQL-groep. Als u dit in uw code hebt, moet u deze verwijderen om te voldoen aan de SQL-groep. Gebruik in plaats daarvan query labels in uw code om gelijkwaardige functionaliteit te implementeren. Zie het artikel [Label](sql-data-warehouse-develop-label.md) voor meer informatie. ## <a name="using-throw-and-raiserror"></a>THROW en//////// THROW is de meer moderne implementatie voor het verhogen van uitzonde ringen in de SQL-groep, maar dit wordt ook wel ondersteund. Er zijn enkele verschillen die u moet betalen. * Door de gebruiker gedefinieerde fout berichten getallen kunnen zich niet in het 100.000-150.000-bereik bevallen * Fout berichten met de volgende strekking worden opgelost om 50.000 * Het gebruik van sys. messages wordt niet ondersteund ## <a name="limitations"></a>Beperkingen De SQL-groep heeft enkele andere beperkingen die betrekking hebben op trans acties. De verschillen zijn als volgt: * Geen gedistribueerde trans acties * Geen geneste trans acties toegestaan * Geen opslag punten toegestaan * Geen benoemde trans acties * Geen gemarkeerde trans acties * Geen ondersteuning voor DDL, zoals CREATE TABLE binnen een door de gebruiker gedefinieerde trans actie ## <a name="next-steps"></a>Volgende stappen Zie [Aanbevolen procedures voor trans acties](sql-data-warehouse-develop-best-practices-transactions.md)voor meer informatie over het optimaliseren van trans acties. Zie [Aanbevolen procedures voor SQL](sql-data-warehouse-best-practices.md)-groepen voor meer informatie over andere aanbevolen procedures voor SQL-Pools.
50.861751
502
0.737429
nld_Latn
0.997541
218532288540aed2a6e40db764f8dd385537c03e
1,190
md
Markdown
README.md
bertdida/selenium-wpupdater
06ebce50ac6d8ed0c37e2038b324d964c610bd5b
[ "MIT" ]
14
2019-05-16T15:33:51.000Z
2022-01-07T04:16:39.000Z
README.md
bertdida/selenium-wpupdater
06ebce50ac6d8ed0c37e2038b324d964c610bd5b
[ "MIT" ]
3
2019-05-10T09:32:44.000Z
2021-06-01T23:44:58.000Z
README.md
bertdida/selenium-wpupdater
06ebce50ac6d8ed0c37e2038b324d964c610bd5b
[ "MIT" ]
1
2019-12-07T05:11:33.000Z
2019-12-07T05:11:33.000Z
# selenium-wpupdater This project helps you update WordPress (WP) automatically. It does this by parsing a file that contains WP URLs with their corresponding username and password, then simulate the following for each data set: 1. Opens a browser 2. Logs in to WP dashboard 3. Checks to see if any of core, plugin, and theme updates are available — if yes do update 4. Logs out from the dashboard ## Prerequisites - Python 3 or above - [Pipenv](https://github.com/pypa/pipenv) - [ChromeDriver](http://chromedriver.chromium.org/downloads) ## Installation 1. Download and extract the [zip file](https://github.com/bertdida/selenium-wpupdater/archive/master.zip) or use Git to clone this repository. 2. Inside the directory open a terminal and run: ```shell pipenv install ``` ## Usage 1. Edit **wpupdater.py** and set `CHROMEDRIVER_PATH` to your driver's executable path. 2. Rename **test_accounts.csv** to remove **test_** then update its contents for the appropriate data. 3. On the terminal run: ```shell pipenv shell python wpupdater.py ``` ## Author Herbert Verdida / [@bertdida](https://twitter.com/bertdida)
29.75
208
0.714286
eng_Latn
0.922764
2186c36aff7443f37ec6ef03d80a62701f5d8b74
2,903
md
Markdown
README.md
Mayster911/Oxylium
59ab8233435224be396d8f2af894400331290f2c
[ "MIT" ]
null
null
null
README.md
Mayster911/Oxylium
59ab8233435224be396d8f2af894400331290f2c
[ "MIT" ]
null
null
null
README.md
Mayster911/Oxylium
59ab8233435224be396d8f2af894400331290f2c
[ "MIT" ]
null
null
null
# Oxylium Problem: Let us assume we're maintaining a WPF application which heavily relies on the property binding mechanism for our controls and the INotifyPropertyChanged interface. Some such applications have numerous transitive dependencies between various controls - and thus by extension - various properties. When adding a new property, the programmer must not only refresh this property, but also others which depend on it, and then all the properties that depend on those properties and so on. Naturally, such an application becomes slow to extend and difficult to maintain. Solution: Oxylium provides a simple NotifyPropertyChangedMediator. In it, classes that can raise a PropertyChanged can be registered, along with their properties. A notification of a property changing is then reported to all dependent objects, along with the properties that need to be refreshed. Currently supported .NET version: .NET 6 # Getting started Oxylium is distributed as a NuGet package. After installing Oxylium implement INotifyPropertyChanged and IRaisePropertyChanged in the client class - raise the PropertyChanged event in the implemented method. Then, create a mediator instance and register the property dependencies. Four registration methods are provided: ``` RegisterBetween(IRaisePropertyChanged item, string propertyDepends, IRaisePropertyChanged onItem, string onProperty) Register(IRaisePropertyChanged item, string propertyDepends, string onProperty) RegisterBetweenQ(IRaisePropertyChanged item, object propertyDepends, IRaisePropertyChanged onItem, object onProperty, string propertyDependsExpression = "", string onPropertyExpression = "") RegisterQ(IRaisePropertyChanged item, object propertyDepends, IRaisePropertyChanged onItem, object onProperty, string propertyDependsExpression = "", string onPropertyExpression = "") ``` RegisterBetween - Use when the dependency of properties spans two objects Register - Use when the dependency of properites is contained in the same object. Functionally identical to RegisterBetween(item, propertyDepends, item, onProperty) The Q variants (Shorthand for Quick) leverage the CallerArgumentExpression attribute to extract the property name via the used expression and are at most capable of extracting the property name when . is used. For example, this line will work: ``` mediator.RegisterBetweenQ(other, other.Prop1, this, this.Prop2); ``` Which is shorter than: ``` mediator.RegisterBetween(other, nameof(other.Prop1), this, nameof(this.Prop2); ``` Finally, call ```mediator.OnPropertyChanged(this);``` in your property setters. OnPropertyChanged in turn leverages the CallerMemberName attribute to extract the name of the property it was called from. In the event of errors a simple property checking mechanism has been implemented. If a given type does not contain the given property, then ArgumentException is thrown.
76.394737
572
0.81743
eng_Latn
0.997841
21872d69e3ef21470750068bf6a2bf0b9b7fda27
713
md
Markdown
ocpp/README.md
motown-io/motown
783ccda7c28b273a529ddd47defe8673b1ea365b
[ "Apache-2.0" ]
31
2015-01-09T14:25:16.000Z
2020-08-19T03:21:10.000Z
ocpp/README.md
quann169/MotownBlueCurrent
d33e1c9bb6a5e5e0ad7711a6f47075477893aa6a
[ "Apache-2.0" ]
32
2015-02-05T13:01:43.000Z
2020-05-17T15:32:18.000Z
ocpp/README.md
quann169/MotownBlueCurrent
d33e1c9bb6a5e5e0ad7711a6f47075477893aa6a
[ "Apache-2.0" ]
16
2015-01-06T14:23:41.000Z
2020-09-11T18:33:12.000Z
# Wiki Detailed information about the add-on can be found in the [wiki](https://github.com/motown-io/motown/wiki/Technical-Architecture#components). The information below is just to provide a quick overview of the modules. # soap-utils Utility classes used by v12-soap and v15-soap. # v12-soap Provides a web service endpoint based on OCPP 1.2. # v15-soap Provides a web service endpoint based on OCPP 1.5. # view-model The common parts of the different OCPP specifications are handled here. The SOAP and JSON web socket modules communicate with the core via this module. This module also handles the events sent out by the core. # websocket-json Provides a JSON web socket endpoint based on OCPP 1.5.
28.52
209
0.776999
eng_Latn
0.990011
2188a8f21ee64538b28d3681e7b9314bd7d7ab53
1,513
md
Markdown
README.md
nondanee/Jike-Railway
2a926fba848f22addba64a08322a8f5075774e76
[ "MIT" ]
23
2019-05-30T15:40:49.000Z
2021-12-22T08:59:55.000Z
README.md
nondanee/Jike-Railway
2a926fba848f22addba64a08322a8f5075774e76
[ "MIT" ]
1
2019-06-21T03:12:35.000Z
2019-06-21T03:59:54.000Z
README.md
nondanee/Jike-Railway
2a926fba848f22addba64a08322a8f5075774e76
[ "MIT" ]
1
2019-06-21T03:59:04.000Z
2019-06-21T03:59:04.000Z
# Jike Bus 🚍 Jike Bus 🚍 是即刻镇的公交专线,有卡刷卡,投币一元。 ## Quick Start ```sh $ npm install nondanee/Jike-Bus ``` > ~~等 Todo 完成后再发布到 npm, 现阶段可以从 Github repository 安装尝鲜~~ 咕了 ```javascript const jike = require('jike') ;(async () => { const j = jike.JikeClient(refreshToken, deviceId) const myProfile = await j.getUserProfile() const myPost = await j.getUserPost(myProfile.username) await j.likeIt(myPost[0]) // 😏 })() ``` ## Usage 还未实现扫码登录部分,初始化客户端要求传入 `x-jike-refresh-token`,部分用户需要传入 `x-jike-device-id` (原因未知) 支持的接口请参考 [Sorosliu1029/Jike-Metro](https://github.com/Sorosliu1029/Jike-Metro) 的 [API 文档](https://www.0x2beace.me/Jike-Metro/) 基本上把 Python 的下划线方法名改成驼峰命名就行了,返回类型均为 Promise (部分方法名有改动) 分页信息流 `loadMore()`,`loadAll()` 与 Jike Metro 功能一致 只是 Web API 请求的简单封装,比较丑陋,不关心返回字段改动 (不固定字段,全看官方心情) **目前为止还是 0 dependency** ## Feature - 获取自己的收藏,查看自己的用户信息 - 流式获取首页消息和动态 - 获取某个用户的用户信息、发布的动态、创建的主题、关注的主题、TA 关注的人和关注 TA 的人 - 获取某条消息 / 动态 / 评论的评论 / 回复 - 获取某个主题下的精选和广场 - 发布个人动态 (可带图、带链接、带主题),删除个人动态 - 点赞、收藏、评论 / 删除评论、转发某条消息 / 动态 - 在浏览器中打开某条消息的原始链接 - 根据关键词搜索主题 - 根据关键词搜索自己的收藏 - 获取即刻首页的推荐关注主题列表(不限于首页显示的5个) > 即 [Jike Metro 乘车体验](https://github.com/Sorosliu1029/Jike-Metro#jike-metro--乘车体验) 中提到的所有 feature ## Todo - [ ] 扫码登录 ([alexeyten/qr-image](https://github.com/alexeyten/qr-image)) - [ ] 消息通知 ([websockets/ws](https://github.com/websockets/ws) + EventEmitter) - [ ] 新动态提醒 (setTimeout 轮询 + EventEmitter) ## Reference - [Sorosliu1029/Jike-Metro](https://github.com/Sorosliu1029/Jike-Metro) ## License MIT
22.924242
126
0.709848
yue_Hant
0.778133
2188ad3cad0a18bbc070e0e11634c472ec927300
890
md
Markdown
ArticleReviews/Abourachid, et al. (2019) - Science Advances - Hoatzin nestling locomotion.md
ElnXu/jonmatthis-2021_Fall_NEU_Biol2299_Locomotion
d6a9dd7e4fd9373b83781536e9f08cccbfb7c808
[ "MIT" ]
null
null
null
ArticleReviews/Abourachid, et al. (2019) - Science Advances - Hoatzin nestling locomotion.md
ElnXu/jonmatthis-2021_Fall_NEU_Biol2299_Locomotion
d6a9dd7e4fd9373b83781536e9f08cccbfb7c808
[ "MIT" ]
null
null
null
ArticleReviews/Abourachid, et al. (2019) - Science Advances - Hoatzin nestling locomotion.md
ElnXu/jonmatthis-2021_Fall_NEU_Biol2299_Locomotion
d6a9dd7e4fd9373b83781536e9f08cccbfb7c808
[ "MIT" ]
null
null
null
### Citation #### Abourachid, A., Herrel, A., Decamps, T., Pages, F., Fabre, A.-C., Van Hoorebeke, L., Adriaens, D., & Garcia Amado, M. A. (2019). Hoatzin nestling locomotion: Acquisition of quadrupedal limb coordination in birds. _Science Advances_, _5_(5). [https://doi.org/10.1126/sciadv.aat0787]. ### Summary #### This paper is about the first quantitative data taken on the [[locomotion]] of [[hoatzins]], using [[motion capture]] [[cameras]] and measurements inside [[laboratory environments]]. The research focus is on limb [[movement patterns]] and use of [[claws]] for grip, both on land and in [[water]]. Nestlings' unique [[quadrupedal]] [[coordination]] suggests recent evolutionary development in the species, and is a reflection of their [[central nervous system]]'s plasticity. #9f8c82 ####keywords (without plurals) #bird #birds #movement #camera #hoatzin #wingsandlegs
55.625
479
0.724719
eng_Latn
0.931931
21890bd4bb001c14443d6489b766bf4363242cb6
28,727
md
Markdown
content/en/tracing/setup/java.md
terra-namibia/documentation
8ea19b29dece3ca0231df2ca4d515c4d342e677e
[ "BSD-3-Clause" ]
null
null
null
content/en/tracing/setup/java.md
terra-namibia/documentation
8ea19b29dece3ca0231df2ca4d515c4d342e677e
[ "BSD-3-Clause" ]
null
null
null
content/en/tracing/setup/java.md
terra-namibia/documentation
8ea19b29dece3ca0231df2ca4d515c4d342e677e
[ "BSD-3-Clause" ]
null
null
null
--- title: Tracing Java Applications kind: documentation aliases: - /tracing/java - /tracing/languages/java - /agent/apm/java/ further_reading: - link: "https://github.com/DataDog/dd-trace-java" tag: "GitHub" text: "Datadog Java APM source code" - link: "tracing/visualization/" tag: "Documentation" text: "Explore your services, resources and traces" - link: "tracing/advanced/" tag: "Advanced Usage" text: "Advanced Usage" --- ## Installation and Getting Started <div class="alert alert-info">If you already have a Datadog account you can find step-by-step instructions in our in-app guides for <a href="https://app.datadoghq.com/apm/docs?architecture=host-based&language=java" target=_blank> host-based</a> and <a href="https://app.datadoghq.com/apm/docs?architecture=container-based&language=java" target=_blank>container-based</a> set ups.</div> To begin tracing applications written in any language, first [install and configure the Datadog Agent][1], see the additional documentation for [tracing Docker applications][2] or [Kubernetes applications][3]. Next, download `dd-java-agent.jar` that contains the Agent class files: ```shell wget -O dd-java-agent.jar 'https://repository.sonatype.org/service/local/artifact/maven/redirect?r=central-proxy&g=com.datadoghq&a=dd-java-agent&v=LATEST' ``` Finally, add the following JVM argument when starting your application in your IDE, Maven or Gradle application script, or `java -jar` command: ``` -javaagent:/path/to/the/dd-java-agent.jar ``` **Note**: * The `-javaagent` needs to be run before the `-jar` file, adding it as a JVM option, not as an application argument. For more information, see the [Oracle documentation][4]. * `dd-trace-java`'s artifacts (`dd-java-agent.jar`, `dd-trace-api.jar`, `dd-trace-ot.jar`) support all JVM-based languages, i.e. Scala, Groovy, Kotlin, Clojure, etc. If you need support for a particular framework, consider making an [open-source contribution][5]. ## Automatic Instrumentation Automatic instrumentation for Java uses the `java-agent` instrumentation capabilities [provided by the JVM][6]. When a `java-agent` is registered, it has the ability to modify class files at load time. The `java-agent` uses the [Byte Buddy framework][7] to find the classes defined for instrumentation and modify those class bytes accordingly. Instrumentation may come from auto-instrumentation, the OpenTracing api, or a mixture of both. Instrumentation generally captures the following info: * Timing duration is captured using the JVM's nanotime clock unless a timestamp is provided from the OpenTracing api * Key/value tag pairs * Errors and stacktraces which are unhandled by the application * A total count of traces (requests) flowing through the system ## Compatibility Datadog officially supports the Java JRE 1.7 and higher of both Oracle JDK and OpenJDK. Datadog does not officially support any early-access versions of Java. ### Integrations Most integrations are enabled by default. The following setting can change the default to disabled. * System Property: `-Ddd.integrations.enabled=false` * Environment Variable: `DD_INTEGRATIONS_ENABLED=false` Integrations can be enabled or disabled individually (overriding the default above). * System Property: `-Ddd.integration.<integration-name>.enabled=true` * Environment Variable: `DD_INTEGRATION_<INTEGRATION_NAME>_ENABLED=true` (See below for each integration's name.) Beta integrations are disabled by default but can be enabled individually. #### Web Framework Compatibility `dd-java-agent` includes support for automatically tracing the following web frameworks. | Server | Versions | Support Type | Instrumentation Names (used for configuration) | |------------------------------|------------|-----------------|------------------------------------------------| | Akka-Http Server | 10.0+ | Fully Supported | `akka-http`, `akka-http-server` | | Grizzly | 2.0+ | [Beta][8] | `grizzly` | | Java Servlet Compatible | 2.3+, 3.0+ | Fully Supported | `servlet`, `servlet-2`, `servlet-3` | | Jax-RS Annotations | JSR311-API | Fully Supported | `jax-rs`, `jaxrs`, `jax-rs-annotations` | | Jetty (non-Servlet) | 8+ | [Beta][8] | `jetty`, `jetty-8` | | Netty HTTP Server | 4.0+ | Fully Supported | `netty`, `netty-4.0`, `netty-4.1` | | Play | 2.4-2.7 | Fully Supported | `play` | | Ratpack | 1.4+ | Fully Supported | `ratpack` | | Spark Java | 2.3+ | [Beta][8] | `sparkjava` (requires `jetty`) | | Spring Web (MVC) | 4.0+ | Fully Supported | `spring-web` | | Spring WebFlux | 5.0+ | Fully Supported | `spring-webflux` | **Web Framework tracing provides:** timing HTTP request to response, tags for the HTTP request (status code, method, etc), error and stacktrace capturing, linking work created within a web request and Distributed Tracing. *Note:* Many application servers are Servlet compatible and are automatically covered by that instrumentation, such as Tomcat, Jetty, Websphere, Weblogic, etc. Also, frameworks like Spring Boot inherently work because it usually uses a supported embedded application server (Tomcat/Jetty/Netty). Beta Instrumentation is disabled by default. Add one of the following configurations to enable it: * System Property: `-Ddd.integration.<integration-name>.enabled=true` * Environment Variable: `DD_INTEGRATION_<INTEGRATION_NAME>_ENABLED=true` Don't see your desired web frameworks? Datadog is continually adding additional support. Contact [Datadog support][8] if you need help. #### Networking Framework Compatibility `dd-java-agent` includes support for automatically tracing the following networking frameworks. | Framework | Versions | Support Type | Instrumentation Names (used for configuration) | |--------------------------|-------------|-----------------|------------------------------------------------| | Apache HTTP Client | 4.0+ | Fully Supported | `httpclient` | | Apache HTTP Async Client | 4.0+ | Fully Supported | `httpasyncclient`, apache-httpasyncclient | | AWS Java SDK | 1.11+, 2.2+ | Fully Supported | `aws-sdk` | | Google HTTP Client | 1.19.0+ | Fully Supported | `google-http-client` | | gRPC | 1.5+ | Fully Supported | `grpc`, `grpc-client`, `grpc-server` | | HttpURLConnection | all | Fully Supported | `httpurlconnection`, `urlconnection` | | Kafka-Clients | 0.11+ | Fully Supported | `kafka` | | Kafka-Streams | 0.11+ | Fully Supported | `kafka`, `kafka-streams` | | Jax RS Clients | 2.0+ | Fully Supported | `jax-rs`, `jaxrs`, `jax-rs-client` | | Jersey Client | 1.9+ | Fully Supported | `jax-rs`, `jaxrs`, `jax-rs-client` | | JMS | 1 and 2 | Fully Supported | `jms` | | Netty HTTP Client | 4.0+ | Fully Supported | `netty`, `netty-4.0`, `netty-4.1` | | OkHTTP | 3.0+ | Fully Supported | `okhttp`, `okhttp-3` | | Play WSClient | 1.0+ | Fully Supported | `play-ws` | | Rabbit AMQP | 2.7+ | Fully Supported | `amqp`, `rabbitmq` | | Spring WebClient | 5.0+ | Fully Supported | `spring-webflux`, `spring-webflux-client` | **Networking tracing provides:** timing request to response, tags for the request (e.g. response code), error and stacktrace capturing, and distributed tracing. Don't see your desired networking framework? Datadog is continually adding additional support. Contact [Datadog support][8] if you need help. #### Data Store Compatibility `dd-java-agent` includes support for automatically tracing the following database frameworks/drivers. | Database | Versions | Support Type | Instrumentation Names (used for configuration) | |-------------------------|----------|-----------------|------------------------------------------------------------------------------------------| | Couchbase | 2.0+ | Fully Supported | `couchbase` | | Cassandra | 3.X | Fully Supported | `cassandra` | | Elasticsearch Transport | 2.0+ | Fully Supported | `elasticsearch`, `elasticsearch-transport`, `elasticsearch-transport-{2,5,6}` (pick one) | | Elasticsearch Rest | 5.0+ | Fully Supported | `elasticsearch`, `elasticsearch-rest`, `elasticsearch-rest-5`, `elasticsearch-rest-6` | | JDBC | N/A | Fully Supported | `jdbc` | | Jedis | 1.4+ | Fully Supported | `redis` | | Lettuce | 5.0+ | Fully Supported | `lettuce` | | MongoDB | 3.0+ | Fully Supported | `mongo` | | SpyMemcached | 2.12+ | Fully Supported | `spymemcached` | `dd-java-agent` is also compatible with common JDBC drivers including: * Apache Derby * Firebird SQL * H2 Database Engine * HSQLDB * IBM DB2 * MariaDB * MSSQL (Microsoft SQL Server) * MySQL * Oracle * Postgres SQL **Datastore tracing provides:** timing request to response, query info (e.g. a sanitized query string), and error and stacktrace capturing. Don't see your desired datastores? Datadog is continually adding additional support. Contact [Datadog support][8] if you need help. #### Other Framework Compatibility `dd-java-agent` includes support for automatically tracing the following other frameworks. | Framework | Versions | Support Type | Instrumentation Names (used for configuration) | |------------------|----------|-----------------|------------------------------------------------| | Dropwizard Views | 0.7+ | Fully Supported | `dropwizard`, `dropwizard-view` | | Hibernate | 3.5+ | Fully Supported | `hibernate` | | Hystrix | 1.4+ | Fully Supported | `hystrix` | | JSP Rendering | 2.3+ | Fully Supported | `jsp`, `jsp-render` | | Slf4J MDC | 1+ | Fully Supported | `mdc` (See also `dd.logs.injection` config) | | Spring Data | 1.8+ | Fully Supported | `spring-data` | | Twilio SDK | 0+ | Fully Supported | `twilio-sdk` | Don't see your desired framework? Datadog is continually adding additional support. Contact [Datadog support][8] if you need help. To improve visibility into applications using unsupported frameworks, consider: * Adding custom instrumentation (with OpenTracing or the `@Trace` annotation). * [Submitting a pull request][9] with instrumentation for inclusion in a future release. * Contacting [Datadog support][8] and submitting a feature request. ## Configuration The tracer is configured using System Properties and Environment Variables as follows: (See integration specific config in the [integrations](#integrations) section above.) {{% table responsive="true" %}} | System Property | Environment Variable | Default | Description | |----------------------------------------|----------------------------------------|----------------------|-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------| | `dd.trace.enabled` | `DD_TRACE_ENABLED` | `true` | When `false` tracing agent is disabled. | | `dd.trace.config` | `DD_TRACE_CONFIG` | `null` | Optional path to a file were configuration properties are provided one per each line. For instance, the file path can be provided as via `-Ddd.trace.config=<FILE_PATH>.properties`, with setting the service name in the file with `dd.trace.enabled=<SERVICE_NAME>` | | `dd.service.name` | `DD_SERVICE_NAME` | `unnamed-java-app` | The name of a set of processes that do the same job. Used for grouping stats for your application. | | `dd.service.mapping` | `DD_SERVICE_MAPPING` | `null` | (Example: `mysql:my-service-name-db`) Dynamically rename services via configuration. Useful for making databases have distinct names across different services. | | `dd.writer.type` | `DD_WRITER_TYPE` | `DDAgentWriter` | Default value sends traces to the Agent. Configuring with `LoggingWriter` instead writes traces out to the console. | | `dd.agent.host` | `DD_AGENT_HOST` | `localhost` | Hostname for where to send traces to. If using a containerized environment, configure this to be the host IP. See [Tracing Docker Applications][1] for more details. | | `dd.trace.agent.port` | `DD_TRACE_AGENT_PORT` | `8126` | Port number the Agent is listening on for configured host. | | `dd.trace.agent.unix.domain.socket` | `DD_TRACE_AGENT_UNIX_DOMAIN_SOCKET` | `null` | This can be used to direct trace traffic to a proxy, to later be sent to a remote Datadog Agent. | | `dd.trace.global.tags` | `DD_TRACE_GLOBAL_TAGS` | `null` | (Example: `key1:value1,key2:value2`) A list of default tags to be added to every span and every JMX metric. This value is merged into `trace.span.tags` and `trace.jmx.tags` to provide single place to configure both. | | `dd.trace.span.tags` | `DD_TRACE_SPAN_TAGS` | `null` | (Example: `key1:value1,key2:value2`) A list of default tags to be added to every span. Tags of the same name added directly to a span overwrite the defaults provided here. | | `dd.trace.jmx.tags` | `DD_TRACE_JMX_TAGS` | `null` | (Example: `key1:value1,key2:value2`) A list of default tags to be added to every JMX metric. Tags of the same name added in JMX metrics configuration overwrite the defaults provided here. | | `dd.trace.header.tags` | `DD_TRACE_HEADER_TAGS` | `null` | (Example: `CASE-insensitive-Header:my-tag-name,User-ID:userId`) A map of header keys to tag names. Automatically apply header values as tags on traces. | | `dd.trace.annotations` | `DD_TRACE_ANNOTATIONS` | ([listed here][2]) | (Example: `com.some.Trace;io.other.Trace`) A list of method annotations to treat as `@Trace`. | | `dd.trace.methods` | `DD_TRACE_METHODS` | `null` | (Example: `package.ClassName[method1,method2,...];AnonymousClass$1[call]`) List of class/interface and methods to trace. Similar to adding `@Trace`, but without changing code. | | `dd.trace.partial.flush.min.spans` | `DD_TRACE_PARTIAL_FLUSH_MIN_SPANS` | `1000` | Set a number of partial spans to flush on. Useful to reduce memory overhead when dealing with heavy traffic or long running traces. | | `dd.trace.report-hostname` | `DD_TRACE_REPORT_HOSTNAME` | `false` | When enabled, it adds the detected hostname to trace metadata | | `dd.trace.split-by-tags` | `DD_TRACE_SPLIT_BY_TAGS` | `null` | (Example: `aws.service`) Used to rename spans to be identified with the corresponding service tag | | `dd.trace.db.client.split-by-instance` | `DD_TRACE_DB_CLIENT_SPLIT_BY_INSTANCE` | `false` | When set to `true` db spans get assigned the instance name as the service name | | `dd.trace.health.metrics.enabled` | `DD_TRACE_HEALTH_METRICS_ENABLED` | `false` | When set to `true` sends tracer health metrics | | `dd.trace.health.metrics.statsd.host` | `DD_TRACE_HEALTH_METRICS_STATSD_HOST` | same as `dd.jmxfetch.statsd.host` | Statsd host to send health metrics to | | `dd.trace.health.metrics.statsd.port` | `DD_TRACE_HEALTH_METRICS_STATSD_PORT` | same as `dd.jmxfetch.statsd.port` | Statsd port to send health metrics to | | `dd.http.client.tag.query-string` | `DD_HTTP_CLIENT_TAG_QUERY_STRING` | `false` | When set to `true` query string parameters and fragment get added to web client spans | | `dd.http.client.error.statuses` | `DD_HTTP_CLIENT_ERROR_STATUSES` | `false` | A range of errors can be accepted, by default 4xx errors aren't reported as errors. This configuration overrides that. Ex. `dd.http.client.error.statuses=400-499` | | `dd.http.server.tag.query-string` | `DD_HTTP_SERVER_TAG_QUERY_STRING` | `false` | When set to `true` query string parameters and fragment get added to web server spans | | `dd.jmxfetch.enabled` | `DD_JMXFETCH_ENABLED` | `true` | Enable collection of JMX metrics by Java Tracing Agent. | | `dd.jmxfetch.config.dir` | `DD_JMXFETCH_CONFIG_DIR` | `null` | (Example: `/opt/datadog-agent/etc/conf.d`) Additional configuration directory for JMX metrics collection. The Java Agent looks for `jvm_direct:true` in the `instance` section in the `yaml` file to change configuration. | | `dd.jmxfetch.config` | `DD_JMXFETCH_CONFIG` | `null` | (Example: `activemq.d/conf.yaml,jmx.d/conf.yaml`) Additional metrics configuration file for JMX metrics collection. The Java Agent looks for `jvm_direct:true` in the `instance` section in the `yaml` file to change configuration. | | `dd.jmxfetch.check-period` | `DD_JMXFETCH_CHECK_PERIOD` | `1500` | How often to send JMX metrics (in ms). | | `dd.jmxfetch.refresh-beans-period` | `DD_JMXFETCH_REFRESH_BEANS_PERIOD` | `600` | How often to refresh list of avalable JMX beans (in seconds). | | `dd.jmxfetch.statsd.host` | `DD_JMXFETCH_STATSD_HOST` | same as `agent.host` | Statsd host to send JMX metrics to. | | `dd.jmxfetch.statsd.port` | `DD_JMXFETCH_STATSD_PORT` | 8125 | Statsd port to send JMX metrics to. | | `dd.logs.injection` | `DD_LOGS_INJECTION` | false | Enabled automatic MDC key injection for Datadog trace and span ids. See [Advanced Usage][3] for details | [1]: /agent/docker/apm [2]: https://github.com/DataDog/dd-trace-java/blob/master/dd-java-agent/instrumentation/trace-annotation/src/main/java/datadog/trace/instrumentation/trace_annotation/TraceAnnotationsInstrumentation.java#L37 [3]: /tracing/advanced/connect_logs_and_traces/?tab=java {{% /table %}} **Note**: * If the same key type is set for both, the system property configuration takes priority. * System properties can be used as JVM parameters. * By default, JMX metrics from your application are sent to the Datadog Agent thanks to DogStatsD over port `8125`. Make sure that [DogStatsD is enabled for the Agent][10]. If you are running the Agent as a container, ensure that `DD_DOGSTATSD_NON_LOCAL_TRAFFIC` [is set to `true`][11], and that port `8125` is open on the Agent container. In Kubernetes, [bind the DogStatsD port to a host port][12]; in ECS, [set the appropriate flags in your task definition][13]. ### B3 Headers Extraction and Injection Datadog APM tracer supports [B3 headers extraction][14] and injection for distributed tracing. Distributed headers injection and extraction is controlled by configuring injection/extraction styles. Currently two styles are supported: * Datadog: `Datadog` * B3: `B3` Injection styles can be configured using: * System Property: `-Ddd.propagation.style.inject=Datadog,B3` * Environment Variable: `DD_PROPAGATION_STYLE_INJECT=Datadog,B3` The value of the property or environment variable is a comma (or space) separated list of header styles that are enabled for injection. By default only Datadog injection style is enabled. Extraction styles can be configured using: * System Property: `-Ddd.propagation.style.extract=Datadog,B3` * Environment Variable: `DD_PROPAGATION_STYLE_EXTRACT=Datadog,B3` The value of the property or environment variable is a comma (or space) separated list of header styles that are enabled for extraction. By default only Datadog extraction style is enabled. If multiple extraction styles are enabled extraction attempt is done on the order those styles are configured and first successful extracted value is used. ## Trace Reporting To report a trace to Datadog the following happens: * Trace completes * Trace is pushed to an asynchronous queue of traces * Queue is size-bound and doesn't grow past a set limit of 7000 traces * Once the size limit is reached, traces are discarded * A count of the total traces is captured to ensure accurate throughput * In a separate reporting thread, the trace queue is flushed and traces are encoded via msgpack then sent to the Datadog Agent via http * Queue flushing happens on a schedule of once per second To see the actual code, documentation, and usage examples for any of the libraries and frameworks that Datadog supports, check the full list of auto- instrumented components for Java applications in the [Integrations](#integrations) section. ### Trace Annotation Add the `dd-trace-api` dependency to your project. For Maven, add this to `pom.xml`: ```xml <dependency> <groupId>com.datadoghq</groupId> <artifactId>dd-trace-api</artifactId> <version>{version}</version> </dependency> ``` For Gradle, add: ```gradle compile group: 'com.datadoghq', name: 'dd-trace-api', version: {version} ``` Now add `@Trace` to methods to have them be traced when running with `dd-java-agent.jar`. If the Agent is not attached, this annotation has no effect on your application. `@Trace` annotations have the default operation name `trace.annotation`, while the method traced have the resource by default. ## Performance Java APM has minimal impact on the overhead of an application: * No collections maintained by Java APM grow unbounded in memory * Reporting traces does not block the application thread * Java APM loads additional classes for trace collection and library instrumentation * Java APM typically adds no more than a 3% increase in CPU usage * Java APM typically adds no more than a 3% increase in JVM heap usage ## Change Agent Hostname Configure your application level tracers to submit traces to a custom Agent hostname: The Java Tracing Module automatically looks for and initializes with the ENV variables `DD_AGENT_HOST` and `DD_TRACE_AGENT_PORT`. ```bash java -javaagent:<DD-JAVA-AGENT-PATH>.jar -jar <YOUR_APPLICATION_PATH>.jar ``` You can also use system properties: ```bash java -javaagent:<DD-JAVA-AGENT-PATH>.jar \ -Ddd.agent.host=$DD_AGENT_HOST \ -Ddd.agent.port=$DD_TRACE_AGENT_PORT \ -jar <YOUR_APPLICATION_PATH>.jar ``` ## Further Reading {{< partial name="whats-next/whats-next.html" >}} [1]: /agent/docker/apm [2]: /tracing/setup/docker [3]: /agent/kubernetes/daemonset_setup/#trace-collection [4]: https://docs.oracle.com/javase/7/docs/technotes/tools/solaris/java.html [5]: https://github.com/DataDog/dd-trace-java/blob/master/CONTRIBUTING.md [6]: https://docs.oracle.com/javase/8/docs/api/java/lang/instrument/package-summary.html [7]: http://bytebuddy.net [8]: /help [9]: https://github.com/DataDog/documentation#outside-contributors [10]: /developers/dogstatsd/#setup [11]: /agent/docker/#dogstatsd-custom-metrics [12]: /agent/kubernetes/dogstatsd/#bind-the-dogstatsd-port-to-a-host-port [13]: /integrations/amazon_ecs/?tab=python#create-an-ecs-task [14]: https://github.com/openzipkin/b3-propagation
79.576177
460
0.549587
eng_Latn
0.919544
2189c3922372806b30edd55083b61dc0fe178458
136
md
Markdown
shortlinks/contest.md
dot-slash-CS/dot-slash-cs.github.io
3edb32344fbf4f2ea0653a8745a9c225a206a1e1
[ "CC-BY-3.0" ]
1
2018-09-14T23:51:26.000Z
2018-09-14T23:51:26.000Z
shortlinks/contest.md
dot-slash-CS/dot-slash-cs.github.io
3edb32344fbf4f2ea0653a8745a9c225a206a1e1
[ "CC-BY-3.0" ]
null
null
null
shortlinks/contest.md
dot-slash-CS/dot-slash-cs.github.io
3edb32344fbf4f2ea0653a8745a9c225a206a1e1
[ "CC-BY-3.0" ]
2
2018-09-11T17:39:51.000Z
2021-09-04T04:20:08.000Z
--- permalink: /contest added_by: cpkurotori redirect_to: - https://drive.google.com/drive/folders/1kBBM-nRSB7VYAkYD_y0z4xG3RZ1P0Nak ---
22.666667
74
0.794118
kor_Hang
0.166606
2189f788e36c5100352db17254e0ceef7413115c
728
md
Markdown
docs/static.md
WARJY/Chooks
fb5a1b0e04043dfcc9366cfd46027cf70c6517e6
[ "MIT" ]
8
2020-06-10T06:46:58.000Z
2020-08-27T13:01:04.000Z
docs/static.md
WARJY/Chooks
fb5a1b0e04043dfcc9366cfd46027cf70c6517e6
[ "MIT" ]
4
2021-10-06T19:06:12.000Z
2022-02-27T07:07:40.000Z
docs/static.md
WARJY/Chooks
fb5a1b0e04043dfcc9366cfd46027cf70c6517e6
[ "MIT" ]
1
2020-06-09T02:21:46.000Z
2020-06-09T02:21:46.000Z
### useStatic > 提供屏幕比例,是否全面屏,dpr,刷新率等的静态hook #### Type ```ts function useStatic(): { rate: number isFullScreen: boolean dpr: number updateInterval: number } ``` #### Return - rate &mdash; 屏幕宽高比 - isFullScreen &mdash; 屏幕是否全面屏 - dpr &mdash; 屏幕dpr - updateInterval &mdash; 屏幕刷新率 #### Example ```js import { useStatic, useStore } from 'chooks' export default { setup(prop:any, context:SetupContext){ //静态hook,可用于全局设置 const { isFullScreen, dpr } = useStatic() const { state, commit } = useStore(context) commit("setGlobalSetting" ,{ isFullScreen: isFullScreen, dpr: dpr, }) return { isFullScreen, dpr } } } ```
18.2
51
0.587912
yue_Hant
0.675408
218a19c190802cd92aaf08cad1e8a31852377d9c
1,622
md
Markdown
articles/cosmos-db/sql-query-abs.md
niklasloow/azure-docs.sv-se
31144fcc30505db1b2b9059896e7553bf500e4dc
[ "CC-BY-4.0", "MIT" ]
null
null
null
articles/cosmos-db/sql-query-abs.md
niklasloow/azure-docs.sv-se
31144fcc30505db1b2b9059896e7553bf500e4dc
[ "CC-BY-4.0", "MIT" ]
null
null
null
articles/cosmos-db/sql-query-abs.md
niklasloow/azure-docs.sv-se
31144fcc30505db1b2b9059896e7553bf500e4dc
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: ABS i Azure Cosmos DB frågespråk description: Lär dig mer om hur den absoluta (ABS) SQL system-funktionen i Azure Cosmos DB Returnerar det positiva värdet för det angivna numeriska uttrycket author: ginamr ms.service: cosmos-db ms.topic: conceptual ms.date: 03/04/2020 ms.author: girobins ms.custom: query-reference ms.openlocfilehash: 6d173827f2695cc20fa208d390731acf0edb3848 ms.sourcegitcommit: 877491bd46921c11dd478bd25fc718ceee2dcc08 ms.translationtype: MT ms.contentlocale: sv-SE ms.lasthandoff: 07/02/2020 ms.locfileid: "78301105" --- # <a name="abs-azure-cosmos-db"></a>ABS (Azure Cosmos DB) Returnerar det absoluta (positiva) värdet för det angivna numeriska uttrycket. ## <a name="syntax"></a>Syntax ```sql ABS (<numeric_expr>) ``` ## <a name="arguments"></a>Argument *numeric_expr* Är ett numeriskt uttryck. ## <a name="return-types"></a>Retur typer Returnerar ett numeriskt uttryck. ## <a name="examples"></a>Exempel I följande exempel visas resultatet av att använda `ABS` funktionen på tre olika tal. ```sql SELECT ABS(-1) AS abs1, ABS(0) AS abs2, ABS(1) AS abs3 ``` Här är resultatuppsättningen. ```json [{abs1: 1, abs2: 0, abs3: 1}] ``` ## <a name="remarks"></a>Kommentarer Den här systemfunktionen kommer att ha nytta av ett [intervall index](index-policy.md#includeexclude-strategy). ## <a name="next-steps"></a>Nästa steg - [Matematiska funktioner Azure Cosmos DB](sql-query-mathematical-functions.md) - [System funktioner Azure Cosmos DB](sql-query-system-functions.md) - [Introduktion till Azure Cosmos DB](introduction.md)
27.965517
157
0.727497
swe_Latn
0.911331
218b8d70e1cf0d086c824bd52750e336a312bdda
6,071
md
Markdown
_posts/2020-06-21-uhf-rfid-antennas.md
theantennaguy/theantennaguy.github.io
7067c23a09f35a5669fca393c5c1c370cc853f54
[ "MIT" ]
null
null
null
_posts/2020-06-21-uhf-rfid-antennas.md
theantennaguy/theantennaguy.github.io
7067c23a09f35a5669fca393c5c1c370cc853f54
[ "MIT" ]
null
null
null
_posts/2020-06-21-uhf-rfid-antennas.md
theantennaguy/theantennaguy.github.io
7067c23a09f35a5669fca393c5c1c370cc853f54
[ "MIT" ]
null
null
null
--- layout: post title: UHF RFID Antennas - I - The basics --- This is the first official entry to the blog! {:refdef: style="text-align: center;"} ![](/images/first-day-brba.gif) {: refdef} So the other day I was watching [this video](https://youtu.be/PWzyPZAPbt0) and I thought "Hey I can actually do that!" (Design UHF RFID antenna tags). So I decided the first entries to this blog would be dedicated to UHF RFID antennas (I don't even feel qualified to qualify that idea...). I always found RFID technology very interesting. It's a shame that it is really obscured, especially the UHF. One can find some applications here and there, and many times don't even realize it's RFID. Still, I always get the feeling that it is a highly underrated technology, I mean, it has the potential to offer so much, but somehow, it's not so ubiquitous as it should have been. Think about it, it's a wireless communication system, where the mobile parts actually work without the need for batteries, but they can send data back to a transmitter without any local power source available to them. That's as close as we can get to 'wingardium leviosa' in our world (Well, there's many interesting techs out there that are rather magical, but this one is cool). {:refdef: style="text-align: center;"} ![](/images/wingardium-leviosa.gif) {: refdef} Now I should focus on the antennas, but I thought I'd give a little explanation behind the workings of RFID. I'm not going to be exhaustive, if your interested in knowing more, I can recommend reading [Finkenzeller's book](https://www.amazon.com/RFID-Handbook-Radio-Frequency-Identification-Fundamentals/dp/0471988510) on RFID. On the video I mentioned in the beginning, they provide a quick explanation to how an RFID system works, check [minute 22:30](https://youtu.be/PWzyPZAPbt0?t=1349). I'd only like to add a little detail that I thought missing. Actually the description of Scotty is accurate, in the sense that the tag really converts the RF signal to a usable DC voltage in order to power up in the internal logic circuit to read the query info, read its corresponding ID from memory and bias an actuator (transistor), but then, it does not generate a signal to transmit back, as explained in the video, the 'Tag' actually reflects the signal back to reader. This happens as follows, the 'Reader', after sending a 'Query' to the air, keeps an unmodulated carrier wave on the air, in the same frequency, for a short time, this is the time needed for the 'Tag' to respond. When this carrier wave reaches the 'Tag', it will tune/de-tune it's input impedance of the antenna, therefore, absorbing or fully reflecting the incoming carrier wave, actually creating what is called an ASK (Amplitude Shift Keying) modulation that the reader can then decode. {:refdef: style="text-align: center;"} ![](/images/uhf-rfid-reader-bd.jpg) {: refdef} Of course this methodology does not allow communications with high data rates, actually it is very constrained and only useful to send a few bytes. Even though there have been people in the academia developing complex modulation schemes that could be employed in passive RFID tags and boost the available data rates. Now, I won't dig to deep into this subject because there's a lot to explore and a lot that I don't know about it. As is the premise of this blog, I'll focus essentially on the antennas. There's several antennas on both sides of the system, especially if you starts looking at proposals in scientific publications from IEEE, but commercially there's in essence three types of base antennas for the 'Reader' and two types for the 'Tag' antennas. On the 'Reader' side, the antenna choice is usually a compromise between the size and weight to the gain and bandwidth. Higher gain antennas allow longer reading ranges for the RFID system, but gain of antennas is intimately connected to their size, and many applications sacrifice reading distance to have a more compact solution. The three typical antennas found on commercial RFID products are the good old microstrip patch antenna, either with an air core or with ceramic core and the quadrifilar spiral antenna. Here's some examples: ![](/images/microstrip_air_core.jpg) | ![](/images/patch_ceramic.png) | ![](/images/quadrifilar-2.png) |------------------------------------|--------------------------------|--------------------------------| Microstrip patch air core | Microstrip patch ceramic core | Quadrifilar spiral antenna On the 'Tag' side there's less variety. Even though there's a lot of different and weird designs for these antennas, most of them are based on the same antenna principle, only changing the shapes to conform to different sizes, make them more square or more rectangular, or round, etc. Most tag antennas are based around dipoles with a T-matching network, however, there's a few variations to this rule, one is the 4-arm crossed dipole antenna found in some commercial 'Tags' as well other more exotic antenna implementations for 'Tags' that work under special conditions, like attached to metallic objects. Here's some examples: ![](/images/dipole_1.png) | ![](/images/4arm_dipole_tag.jpg) | ![](/images/metal_tag.png) |-------------------------|----------------------------------|---------------------------| Dipole | Double-dipole | RFID tag for Metal objects Besides the examples shown above, there's a lot of variants for all sorts of applications, but the antenna types usually revolve around dipoles or the more specific constructions for tags that are to be attached to metallic surfaces. Well, I don't feel like going further than this today. On the next posts I'll focus on exploring in more detail the construction of some of the antennas shown here. With the first being about the good old microstrip patch, starting with the air core, then I'll move to the ceramic core. Eventually I'll reach the tag designs as promised in the beginning of this post. {:refdef: style="text-align: center;"} ![](/images/thats_all.gif) {: refdef}
114.54717
1,128
0.748641
eng_Latn
0.999234
218c5e842a3e764c7f41228c505ce86b08695ad4
6,483
md
Markdown
PerfTests[WIP]/PerfTests.ToDo.md
rsdn/CodeJam
9889cd0c1f0cebcf92176e90ca6902556d214791
[ "MIT" ]
252
2016-03-16T10:28:39.000Z
2022-03-29T04:43:20.000Z
PerfTests[WIP]/PerfTests.ToDo.md
rsdn/CodeJam
9889cd0c1f0cebcf92176e90ca6902556d214791
[ "MIT" ]
82
2016-03-17T22:00:24.000Z
2021-10-16T11:31:52.000Z
PerfTests[WIP]/PerfTests.ToDo.md
rsdn/CodeJam
9889cd0c1f0cebcf92176e90ca6902556d214791
[ "MIT" ]
37
2016-03-16T22:36:58.000Z
2022-03-06T05:56:49.000Z
## TODO, Bench.Net: * Support for `IFlushable.Flush()`. * No file logger for tests https://github.com/dotnet/BenchmarkDotNet/issues/198 * Run state for issues like this. https://github.com/dotnet/BenchmarkDotNet/issues/180 * ReadOnlyConfig * ManualConfig.Add - backport ManualCompetitionConfig design * remove calls instance.setupAction(), instance.targetAction() from BenchmarkProgram.txt https://github.com/dotnet/BenchmarkDotNet/issues/184 ? ## TODOs: ### TODO now: * check metric if no annotations (test succeed???) * check exception if stored metrics are invalid (min > max) * Cache metric values (same as annotation context but for parsing the anotations). * Output contains messages printed under `*** Warnings ***` section. Have no idea what to do with it. Need non-printable analyzer warnings :) * Config modifier attributes. Do some thing with names and discoverability. May be some naming convention for modifiers? ### Types: * ??? Rename Xxx annotations to metric annotations? Not sure, may be a bad idea if we're going to store something except metrics.' * Sealed types: unseal where it looks like a good idea. * Simplify collections of types. MappedCollection / Keyed collection where possible. ### Validations: * No Benchmark / Setup attribute duplication ### Behavior: * Annotation adjustments: options for metric unt selection: preserve / autoscale; per-metric adjustment? * Future versions of BDN: each member in competition may have its own target type (multiple targets for single run). Ensure that our code is prepared to it. * "Cannot detect encoding" checks for anything that reads streams as text? * Concurrency: lock should be performed on entire benchmark run. + Subtask: Support for concurrent competiton run s (stub code were removed at master afd9977, restore, then fix). Use case: load testing. + Subtask: concurrent runs. Disallowed as it will break concurrent file annotations * metric info: flags enum of column kinds to add into summary * Rule to ignore individual metric annotations (do not annotate / current only etc?) * Merge failed & adjusted messages for multiple benchmarks into single message? * ??? Integer metrics without unit scale (for total GC count) * Revisit caching methods for analyser (all those members should contain `cache` in name). * Add cache invalidation policy (limit of 1000) * ??? Diagnosers: faster options for diagnosers run? * Variance for SingleValueMetricCalculator / PercentileMetricCalculator * Optional relative time metric - better diagnostic if there's not empty attribute for it but the metric is not listed in config. * Check Code.BugIf assertions. Replace with messages where possible. * Rerun if sources were adjusted from log? * Prev run log : cache + reread only if local & size/date/checksum(?) changed!!! * Display-only metrics (only as columns? .ctor arg to the metric? bad idea as these will be removed silently as duplicates (dups removed by AttributeType) ### Messages: * 'No logged XML annotation for Test00Baseline found. Check if the method was renamed.' => add metric names to message * Validate all messages, check that origin (benchmark target) is logged, add hints to them. * Add typed method to force same naming for source of the messages (Target, Type, xmlFile etc). * Check WriteVerboseHint for source annotations * ??? Write hint with absolute values if relative limits failed? (discarded) (will require bool arg for metricValuesProvider, recheck columns and diagnosers, maybe they will have to honor this flag too)). * Message about updated annotations: improve readability * Message about ignored empty metrics: improve readability ### Logging: * Add advanced diagnostic for config-adjusting-time (logger as an arg of ConfigFactory + fix methods). * Check `+ Environment.NewLine` usages in `XunitCompetitionRunner.ReportXxx()` methods * LogColors.Hint: use it for something? ### Tests: * xUnit: tests: force run as x64 (appveyor may run as x86)? * Source annotations: test for partial files / methods * High-priority test for TestProcessCycleTimeClock * Tests for broken log annotations. * app.config in the test integration projects: do we need it? * Test for standard analyser warnings - are they logged as messages? * out of process test * local test for run under CI (CI mode on) (+ 1!) * Test for bad encoding * Memory limits + diagnoser - test for accuracy * Ensure that unknown unit in prev run log results in warning * Ensure that all attribute lines are read during source file parsing. * Validate the return results!!! (partially done with introduction of IHostApi) ### Cleanup * Remove and re-add resharper suppressions ### Features to test & to document * FAQ (minimum metric values) * Gc metrics (byte-precise allocation monitoring, gc collection metrics) ## Long-term TODOs: * Attach detailed reports to tests ## Issues: https://github.com/dotnet/BenchmarkDotNet/issues/361 https://github.com/dotnet/BenchmarkDotNet/issues/360 https://github.com/dotnet/BenchmarkDotNet/issues/327 https://github.com/dotnet/BenchmarkDotNet/issues/324 https://github.com/dotnet/BenchmarkDotNet/issues/136 https://github.com/xunit/xunit/issues/908 ## DOCS: * guidleines for use cases: - dead code elimination : use prev value - relatively cost action such as method invocation: use arrays. ## layered design: Bench.NET part ### Layer 0: things to port to BDN Thats it. ### Layer 1: Common competition api to be used by end customers and by the competition itself. Options, metrics, various helpers. ### Layer 2: CompetitionCore Core api to be used during competition run. Not meant to be exposed to end users but is available for third-party extensions. Icludes logging and messages. ### Layer 3: Competition analysers Competition metric related things: analysers and annotation storages. ### Layer 4: Configuration Configuration subsystem & predefined implementations of exporters, metric providers etc. ### Layer 5: All together Reusable competition run API. ## Long-term plans for v2 * Advanced metrics (proof that api is done with package that uses GLAD) * Backup for auto-annotations * Make annotations storage API public. Use case: to store them as a database file + ??? Subtask: exporter-only version to collect actual metric values + ??? Advanced subtask: collect raw metric values (discarded, too much data without a gain). * Support for multi-case benchmarks (discarded)
49.48855
157
0.763844
eng_Latn
0.966824
218c6085dfd1e11776fdb1ba5b355ad0d1bdf2cd
2,363
md
Markdown
README.md
ChristopheBroillet/Hydra_Segmentation
651ee9ea4047ea607c5aa5a668a5693af17f44e9
[ "MIT" ]
null
null
null
README.md
ChristopheBroillet/Hydra_Segmentation
651ee9ea4047ea607c5aa5a668a5693af17f44e9
[ "MIT" ]
null
null
null
README.md
ChristopheBroillet/Hydra_Segmentation
651ee9ea4047ea607c5aa5a668a5693af17f44e9
[ "MIT" ]
null
null
null
# Pre-Processing Segmentation Datasets for the Hydra Framework Author: Christophe Broillet ## Abstract Early cancer detection is a crucial point nowadays to save human lives. Computer aided diagnosis (CAD) systems, which use machine learning techniques, can help radiologists to detect the very first stages of cancer, but also to locate and segment malignant tumors in medical images. Deep learning models, such as convolutional neural networks, are today useful and powerful for computer vision, which makes them the perfect candidates for CAD systems. However, these models require a lot of data, which is often difficult to obtain in the medical field, because of privacy issues. This work focuses on addressing the data scarcity by preprocessing new datasets to an existing computer aided diagnosis system, namely the Hydra framework, while adding the segmentation task to this CAD system. The new datasets contain CT or PET scans of the lungs, and the head and neck, but also ultrasounds of the breast. The preprocessing of the datasets are shown and explained, and are also provided in three new scripts. These datasets are ready to be used in a segmentation task. This task is useful for radiologists, as it shows precisely where the tumors are located, and gives information about their shape. This work develops a new module that creates segmentation masks of medical images, given the cloud of points that contour the tumor. After using this module, the datasets contain the medical images (inputs) and the segmentation masks (labels). The Hydra framework is now ready to train the segmentation task in a supervised learning manner. ## Repository description The folder *preprocessing_scripts* contains the preprocessing scripts for three datasets. These scripts preprocess the data before the Hydra model uses it. The scripts are separated for each dataset. The *segmentation_mask* module, that creates segmentation masks given the list of points that contour tumors, is available and used in two datasets. The *Pipfile* and *Pipfile.lock* are here to install a virtual environment on a new machine that wants to run these scripts. The original publication of this work is available [here](https://exascale.info/assets/pdf/students/2022_BSc_Christophe_Broillet.pdf) or in this [repository](https://github.com/ChristopheBroillet/bachelor_thesis).
168.785714
1,540
0.815489
eng_Latn
0.999327
218d0648fea517ba26e0e6d7bc2cb8fd6a8ab900
2,284
md
Markdown
README.md
mchitre/ForwardDiff.jl
ff96532e1c3a862a2d3e64475a84913d376e713f
[ "MIT" ]
null
null
null
README.md
mchitre/ForwardDiff.jl
ff96532e1c3a862a2d3e64475a84913d376e713f
[ "MIT" ]
null
null
null
README.md
mchitre/ForwardDiff.jl
ff96532e1c3a862a2d3e64475a84913d376e713f
[ "MIT" ]
null
null
null
![CI](https://github.com/JuliaDiff/ForwardDiff.jl/workflows/CI/badge.svg) [![Coverage Status](https://coveralls.io/repos/JuliaDiff/ForwardDiff.jl/badge.svg?branch=master&service=github)](https://coveralls.io/github/JuliaDiff/ForwardDiff.jl?branch=master) [![](https://img.shields.io/badge/docs-stable-blue.svg)](http://www.juliadiff.org/ForwardDiff.jl/stable) [![](https://img.shields.io/badge/docs-latest-blue.svg)](http://www.juliadiff.org/ForwardDiff.jl/latest) # ForwardDiff.jl ForwardDiff implements methods to take **derivatives**, **gradients**, **Jacobians**, **Hessians**, and higher-order derivatives of native Julia functions (or any callable object, really) using **forward mode automatic differentiation (AD)**. While performance can vary depending on the functions you evaluate, the algorithms implemented by ForwardDiff **generally outperform non-AD algorithms in both speed and accuracy.** Here's a simple example showing the package in action: ```julia julia> using ForwardDiff julia> f(x::Vector) = sum(sin, x) + prod(tan, x) * sum(sqrt, x); julia> x = rand(5) # small size for example's sake 5-element Array{Float64,1}: 0.986403 0.140913 0.294963 0.837125 0.650451 julia> g = x -> ForwardDiff.gradient(f, x); # g = ∇f julia> g(x) 5-element Array{Float64,1}: 1.01358 2.50014 1.72574 1.10139 1.2445 julia> ForwardDiff.hessian(f, x) 5x5 Array{Float64,2}: 0.585111 3.48083 1.7706 0.994057 1.03257 3.48083 1.06079 5.79299 3.25245 3.37871 1.7706 5.79299 0.423981 1.65416 1.71818 0.994057 3.25245 1.65416 0.251396 0.964566 1.03257 3.37871 1.71818 0.964566 0.140689 ``` Trying to switch to the latest version of ForwardDiff? See our [upgrade guide](http://www.juliadiff.org/ForwardDiff.jl/stable/user/upgrade/) for details regarding user-facing changes between releases. ## Publications If you find ForwardDiff useful in your work, we kindly request that you cite [the following paper](https://arxiv.org/abs/1607.07892): ```bibtex @article{RevelsLubinPapamarkou2016, title = {Forward-Mode Automatic Differentiation in {J}ulia}, author = {{Revels}, J. and {Lubin}, M. and {Papamarkou}, T.}, journal = {arXiv:1607.07892 [cs.MS]}, year = {2016}, url = {https://arxiv.org/abs/1607.07892} } ```
37.442623
254
0.721103
eng_Latn
0.507068
218d8c69ddd842350ccb549826d0f6f9670a217d
11,398
md
Markdown
src/docs/options.md
mpopenko-exos/pants
47d27037c8b13291fc9023e56ddd1b1defdf1b8e
[ "Apache-2.0" ]
1
2021-05-05T18:58:28.000Z
2021-05-05T18:58:28.000Z
src/docs/options.md
mpopenko-exos/pants
47d27037c8b13291fc9023e56ddd1b1defdf1b8e
[ "Apache-2.0" ]
1
2018-09-04T17:37:34.000Z
2018-09-04T19:42:58.000Z
src/docs/options.md
mpopenko-exos/pants
47d27037c8b13291fc9023e56ddd1b1defdf1b8e
[ "Apache-2.0" ]
3
2020-06-30T08:28:13.000Z
2021-07-28T09:35:57.000Z
Pants Options ============= Pants is very configurable and has literally hundreds of tuneable parameters, known in Pants parlance as _options_. Most options are automatically set to useful defaults. However it is inevitable that sooner or later you'll need to tweak some of them. For example, you may want [[to pin your Pants version|pants('src/docs:install')]]. The Pants options system is fairly complex. Codebase administrators will need to understand its details and nuances. End users, however, can usually get by with just a superficial view. Option Scopes ------------- Pants has many moving parts, and more can be added via custom plugins. To avoid option naming collisions, each configurable component is qualified by a _scope_. + Global options belong to the global scope. + Options for task `task` in goal `goal` belong to the `goal.task` scope. E.g., `gen.thrift` or `compile.rsc`. However if the task has the same name as the goal, then the scope is just that name. E.g., `./pants list`, not `./pants list.list`. + Options for the global instance of subsystem `subsystem` belong to the `subsystem` scope. E.g., `jvm` or `cache`. + Options for the instance of subsystem `subsystem` belonging to some other task or subsystem with scope `scope` belong to the `subsystem.scope` scope. E.g., `cache.compile.rsc`. The scope names are used to qualify the option names when setting their values, as explained in the remainder of this document. Basic vs. Advanced Options -------------------------- Some options, such as `-l` to control the logging verbosity level, are expected to be set directly by users, as needed. But most options are expected to be configured once by a codebase administrator and not directly modified by users. These latter options are called _advanced_ options, and are not enumerated in help messages by default. Advanced options may change the way projects are built and packaged, impact the build cache or change test behavior. For that reason, they should usually be set in config files and checked in to the repo so that build output is consistent, no matter which user invokes it. To show the full list of options for a goal, including advanced options, use `./pants <goal> --help-advanced`. You can view all options for all scopes with `./pants --help-all --help-advanced`. Recursive Options ----------------- A handful of global options, such as `-l`, are _recursive_. That is, they have a global value which can be overridden in each scope. This allows you, for example, to set different logging verbosity levels in different tasks. Option Types ------------ Option value literal formats are the same on the command-line, environment variables and in `pants.ini`, with a couple of extra command-line conveniences. ### Boolean Options The boolean option values are: `false`, `False`, `true`, `True`. On the command line, you can omit the value thus: `--foo-bar` is the same as `--foo-bar=false`, and `--no-foo-bar` is the same as `--foo-bar=false`. ### Int and Float Options Specify values in decimal notation: `--foo=5`, `--bar=4.5`. ### String Options Surround strings with single or double quotes if they contain embedded spaces: `--foo="hello, world"`. ### List Options List options can be appended to and filtered, as well as overridden. For example, for an option `--foo` whose default value is `[1, 2]`, then in `pants.ini`: + `foo: 3` will yield `[1, 2, 3]`. + `foo: +[3, 4]` will yield `[1, 2, 3, 4]`. + `foo: -[1]` will yield `[2]`. + `foo: [3, 4]` will yield `[3, 4]`. Multiple append and filter expressions may be delimited with commas, allowing you to append and filter simultaneously: + `foo: +[3,4],-[1]` will yield `[2, 3, 4]`. On the command line you can append single values multiple times: + `--foo=3 --foo=4` will yield the value `[1, 2, 3, 4]`. Note that these command line values will be appended to the value determined from the defaults plus the values in `pants.ini`. To override the value, use `--foo=[3, 4]`. Filters apply to the entire list constructed so far, and will filter all appearances of the value: + `--foo=1 --foo=1 --foo=2 --foo=-[1]` will yield `[2, 2]`. Filters take precedence over appends, so you cannot "add something back in": + `--foo=-[2] --foo=2` will yield `[1]`. ### Dict Options Dict option values are Python-style dict literals: `--foo={"a":1,"b":2}`. ### Available Options The options available in a given Pants-enabled codebase depend on which backends and plugins are activated. + View [all options]('options_reference.html') available in a "vanilla" Pants install. + To see a complete list of all basic options available in a given Pants-enabled codebase, enter `./pants help-all`. + To see global (or goal-specific) basic options, enter `./pants help (goal)`. + To see global (or goal-specific) basic and advanced options, enter `./pants help-advanced (goal)`. Setting Option Values --------------------- Every Pants option can be set in one three ways, in descending order of precendence: + Using a command-line flag. + Using an environment variable. + In a config file. Config files are typically used to set codebase-wide defaults for all users. Individual users can then override various values using environment variables or command-line flags. Options that aren't set by one of these three methods will fall back to a sensible default, so that Pants will work "out of the box" in common scenarios. ### Command Line Flags Option `option` in scope `foo.bar` can be set using the flag `--foo-bar-option=<value>`. Global options are set with no scope qualifier, e.g., `--pants-workdir=/path/to/workdir`. Values for single-letter flags (those that start with a single dash) can be set without the equals sign: `-ldebug` is the same as `--level=debug` (`-l` is a synonym for `--level`). All other flags must use an equals sign to set a value. There's a useful shorthand that can save some typing when setting multiple options for a single task: if you invoke a task explicitly on the command line then you can follow that task with unqualified options in its scope. E.g., `./pants compile.rsc --no-incremental --name-hashing` instead of `./pants compile --no-compile-rsc-incremental --compile-rsc-name-hashing`. Note that this shorthand requires you to mention a specific task, not just a goal: `./pants compile.rsc` instead of just `./pants compile` as you would usually enter. **All tasks in the `compile` goal will still be executed, not just `compile.rsc`**, but the `.zinc` addition is a convenience to support shorthand options. This works when specifying multiple goals, e.g., `./pants compile.rsc --no-incremental --name-hashing test.junit --parallel-threads=4` ### Environment Variables Option `option` in scope `foo.bar` can be set via the environment variable `PANTS_FOO_BAR_OPTION`. E.g., `PANTS_COMPILE_ZINC_INCREMENTAL=false`. Global options can be set using `PANTS_GLOBAL_OPTION` as expected, but you can also omit the GLOBAL and use `PANTS_OPTION`. E.g., `PANTS_LEVEL=debug`. If a global option name itself starts with the word 'pants' then you can omit the repetition. E.g., `PANTS_WORKDIR` instead of `PANTS_PANTS_WORKDIR`. Environment variables are overridden by command line flags, but take precedence over settings in config files. ### Config File The main Pants config file location is `pants.ini` in your source tree's top-level directory. If you installed pants [[as recommended|pants('src/docs:install')]] this file should already exist. The `pants.ini` file is an INI file parsed by [ConfigParser](http://docs.python.org/library/configparser.html). See [RFC #822](http://tools.ietf.org/html/rfc822.html#section-3.1) section 3.1.1 for the full rules python uses to parse ini file entries. A `pants.ini` file looks something like: :::ini [scope] option1: value1 option2: value2 Sections in the `.ini` file correspond to the task name or subsystem name, or a combination of both. The `[DEFAULT]` section is special: its values are available in all other sections. E.g., :::ini [DEFAULT] thrift_workdir: %(pants_workdir)s/thrift [GLOBAL] print_exception_stacktrace: True [gen.thrift] workdir: %(thrift_workdir)s [compile.rsc] args: [ '-C-Tnowarnprefixes', '-C%(thrift_workdir)s', ] Note that continuation lines of multiple-line values must be indented. For example, the closing bracket in the multi-line list above. Settings in config files are overridden by options specified in the environment or by command line flags. There are a few differences in using options in the config file compared to invoking them from the command line: - Omit the leading double dash (`--`) - Dash characters (`-`) are transposed to underscores (`_`). - Boolean flag values are enabled and disabled by setting the value of the option to `True` or `False` - The prefix for long form options is not specified. Instead, you must organize the options into their appropriate sections. ### Overlaying Config Files Sometimes it's convenient to keep `.ini` settings in more than one file. Perhaps you usually operate Pants in one "mode", but occasionally need to use a tweaked set of settings. Use the `--pants-config-files` command-line option to specify a second `.ini` file. Each of this `.ini` file's values override the corresponding value in `pants.ini`, if any. For example, if your `pants.ini` contains the section [test.junit] coverage_html_open: True debug: False ...and you invoke `--pants-config-files=quick.ini` and your `quick.ini` says [test.junit] coverage_html_open: False skip: True ...then Pants will act as if you specified [test.junit] coverage_html_open: False skip: True debug: False Note that `--pants-config-files` is a list-valued option, so all the idioms of lists work. You can add a third file with another invocation of `--pants-config-files=<path>`, or you can replace the standard one entirely with `--pants-config-files=[<list>]`. Troubleshooting Config Files ---------------------------- ### Use the Right Section Name Section names must correspond to actual option scopes. Use the correct section name as specified in the help output: :::ini # Wrong [compile] # The correct scope for the 'warnings' option is compile.rsc zinc_warnings: False # Right [compile.rsc] warnings: False When in doubt, the scope is described in the heading for each option in the help output: ::bash $ ./pants compile.rsc --help compile.rsc options: Compile Scala and Java code using Zinc. --[no-]compile-rsc-debug-symbols (default: False) Compile with debug symbol enabled. ... ### Check the Formatting Settings that span multiple lines should be indented. To minimize problems, follow these conventions: - Followon lines should be indented four spaces. - The ending bracket for lists and dicts should be indented two spaces. Here are some examples of correctly and incorrectly formatted values: :::ini # Right jvm_options: [ "foo", "bar" ] # Right jvm_options: [ "foo", "bar" ] # Wrong jvm_options: [ "foo", "bar" ] # Followon line must be indented # Wrong jvm_options: [ "foo", "bar" ] # closing bracket must be indented
36.184127
124
0.721881
eng_Latn
0.996943
218e825a7c6d6b315bb2aa810e69a433c460ec09
2,344
md
Markdown
README.md
slavrd/packer-aws-nginx64
6ddf1b12ab98a6f3f1fd346d8c03cedfe5d3bc4a
[ "MIT" ]
null
null
null
README.md
slavrd/packer-aws-nginx64
6ddf1b12ab98a6f3f1fd346d8c03cedfe5d3bc4a
[ "MIT" ]
null
null
null
README.md
slavrd/packer-aws-nginx64
6ddf1b12ab98a6f3f1fd346d8c03cedfe5d3bc4a
[ "MIT" ]
null
null
null
# Packer AWS nginx64 Packer project to build AWS ami image of ubuntu xenial with nginx. The project contains two equivalent templates using the `json` and `hcl` syntax - `template.json` and `template.pkr.hcl`. ## AWS Authentication AWS credentials can be provided in the ways listed below in order of precedence: * Set template variables `aws_access_key_id` and `aws_secret_access_key` directly when building the template e.g. ```packer build -var 'aws_access_key_id=<aws_key_id>' -var 'aws_secret_access_key=<aws_key>' template.json``` * Set environment variables `AWS_ACCESS_KEY_ID` and `AWS_SECRET_ACCESS_KEY` * Have a shared credential file `$HOME/.aws/credentials` or at path defined in `AWS_SHARED_CREDENTIALS_FILE` ## Setting base AWS ami and AWS region The base ami must be in the selected region. By default the template uses: * `aws_base_ami_id: ami-0932440befd74cdba` * `aws_region: eu-central-1` In order to change them pass `-var 'aws_region=<aws_region>' -var 'aws_base_ami_id=<aws_ami_id>'` to `packer build` ## Build the template Run: `packer build template.json` ## Running Kitchen test * **Prerequisites:** * Install rbenv - `brew install rbenv` * Initialize rbenv - add to `~/.bash_profile` line `eval "$(rbenv init -)"` * Run `source ~/.bash_profile` * Install ruby 2.3.1 with rbenv - `rbenv install 2.3.1` , check `rbenv versions` * Set ruby version for the project to 2.3.1 - `rbenv local 2.3.1` , check `rbenv local` * Install bundler - `gem install bundler` * Install gems from Gemfile - `bundle install` * **Set up AWS authentication** - choose one of the options below: * Set up the environment variables `AWS_ACCESS_KEY_ID` and `AWS_SECRET_ACCESS_KEY` * Set up a shared credentials file `$HOME/.aws/credentials` * **Set ami id and region for the test** - set the tested ami id and it's region in `.kitchen.yml` as follows: ```YAML platforms: - name: aws/nginx64 driver: image_id: <ami_id_to_be_tested> region: <aws_region> ``` * **Running kitchen:** * List kitchen environment - `bundle exec kitchen list` * Build kitchen environment - `bundle exec kitchen converge` * Run kitchen tests - `bundle exec kitchen verify` * Destroy kitchen environment - `bundle exec kitchen destroy` * Automatically build, test, destroy - `bundle exec kitchen test`
35.515152
121
0.731655
eng_Latn
0.832448
218eb50c4020ae021d4cc8d47d92c9a1a6a0c2e7
16,529
md
Markdown
README.md
lammertw/tuulbox
9e009d5a5bf0bc4f5f7bb3571bd7b692e1b076c9
[ "Apache-2.0" ]
null
null
null
README.md
lammertw/tuulbox
9e009d5a5bf0bc4f5f7bb3571bd7b692e1b076c9
[ "Apache-2.0" ]
null
null
null
README.md
lammertw/tuulbox
9e009d5a5bf0bc4f5f7bb3571bd7b692e1b076c9
[ "Apache-2.0" ]
null
null
null
[![codecov](https://codecov.io/gh/JuulLabs/tuulbox/branch/main/graph/badge.svg?token=24ilSLPwN2)](https://codecov.io/gh/JuulLabs/tuulbox) [![Slack](https://img.shields.io/badge/Slack-%23juul--libraries-ECB22E.svg?logo=data:image/svg+xml;base64,PHN2ZyB2aWV3Qm94PSIwIDAgNTQgNTQiIHhtbG5zPSJodHRwOi8vd3d3LnczLm9yZy8yMDAwL3N2ZyI+PGcgZmlsbD0ibm9uZSIgZmlsbC1ydWxlPSJldmVub2RkIj48cGF0aCBkPSJNMTkuNzEyLjEzM2E1LjM4MSA1LjM4MSAwIDAgMC01LjM3NiA1LjM4NyA1LjM4MSA1LjM4MSAwIDAgMCA1LjM3NiA1LjM4Nmg1LjM3NlY1LjUyQTUuMzgxIDUuMzgxIDAgMCAwIDE5LjcxMi4xMzNtMCAxNC4zNjVINS4zNzZBNS4zODEgNS4zODEgMCAwIDAgMCAxOS44ODRhNS4zODEgNS4zODEgMCAwIDAgNS4zNzYgNS4zODdoMTQuMzM2YTUuMzgxIDUuMzgxIDAgMCAwIDUuMzc2LTUuMzg3IDUuMzgxIDUuMzgxIDAgMCAwLTUuMzc2LTUuMzg2IiBmaWxsPSIjMzZDNUYwIi8+PHBhdGggZD0iTTUzLjc2IDE5Ljg4NGE1LjM4MSA1LjM4MSAwIDAgMC01LjM3Ni01LjM4NiA1LjM4MSA1LjM4MSAwIDAgMC01LjM3NiA1LjM4NnY1LjM4N2g1LjM3NmE1LjM4MSA1LjM4MSAwIDAgMCA1LjM3Ni01LjM4N20tMTQuMzM2IDBWNS41MkE1LjM4MSA1LjM4MSAwIDAgMCAzNC4wNDguMTMzYTUuMzgxIDUuMzgxIDAgMCAwLTUuMzc2IDUuMzg3djE0LjM2NGE1LjM4MSA1LjM4MSAwIDAgMCA1LjM3NiA1LjM4NyA1LjM4MSA1LjM4MSAwIDAgMCA1LjM3Ni01LjM4NyIgZmlsbD0iIzJFQjY3RCIvPjxwYXRoIGQ9Ik0zNC4wNDggNTRhNS4zODEgNS4zODEgMCAwIDAgNS4zNzYtNS4zODcgNS4zODEgNS4zODEgMCAwIDAtNS4zNzYtNS4zODZoLTUuMzc2djUuMzg2QTUuMzgxIDUuMzgxIDAgMCAwIDM0LjA0OCA1NG0wLTE0LjM2NWgxNC4zMzZhNS4zODEgNS4zODEgMCAwIDAgNS4zNzYtNS4zODYgNS4zODEgNS4zODEgMCAwIDAtNS4zNzYtNS4zODdIMzQuMDQ4YTUuMzgxIDUuMzgxIDAgMCAwLTUuMzc2IDUuMzg3IDUuMzgxIDUuMzgxIDAgMCAwIDUuMzc2IDUuMzg2IiBmaWxsPSIjRUNCMjJFIi8+PHBhdGggZD0iTTAgMzQuMjQ5YTUuMzgxIDUuMzgxIDAgMCAwIDUuMzc2IDUuMzg2IDUuMzgxIDUuMzgxIDAgMCAwIDUuMzc2LTUuMzg2di01LjM4N0g1LjM3NkE1LjM4MSA1LjM4MSAwIDAgMCAwIDM0LjI1bTE0LjMzNi0uMDAxdjE0LjM2NEE1LjM4MSA1LjM4MSAwIDAgMCAxOS43MTIgNTRhNS4zODEgNS4zODEgMCAwIDAgNS4zNzYtNS4zODdWMzQuMjVhNS4zODEgNS4zODEgMCAwIDAtNS4zNzYtNS4zODcgNS4zODEgNS4zODEgMCAwIDAtNS4zNzYgNS4zODciIGZpbGw9IiNFMDFFNUEiLz48L2c+PC9zdmc+&labelColor=611f69)](https://kotlinlang.slack.com/messages/juul-libraries/) # Tuulbox Toolbox of utilities/helpers for Kotlin development. ## [Collections](https://juullabs.github.io/tuulbox/collections/index.html) ![badge-ios] ![badge-js] ![badge-jvm] ![badge-mac] ### [`SynchronizedMap`] Provides a map where read and write operations are protected using a reentrant lock, allowing mutation from multiple threads without fear of data corruption. Use `SynchronizedMap.synchronize` when multiple read and write operations need to happen atomically, such as when performing a `getOrPut`. ```kotlin // Creating a synchronized map val map = SynchronizedMap(mutableMapOf("key" to "value")) // Synchronize across multiple operations val value = map.synchronized { getOrPut("key") { "defaultValue" } } ``` ### [`Map.toJsObject`] Convert a map to a Plain Old JavaScript Object by transforming the keys to strings and the values to any of the JavaScript types. `Map` is not directly accessible from Javascript code. For example: ```kotlin val simple = mapOf(1 to "a") val json = js("JSON.stringify(simple)") as String println(json) >>> { "_keys_up5z3z$_0": null, "_values_6nw1f1$_0": null, <and lots of other name-mangled properties> } val value = js("simple['1']") as String println(value) >>> ClassCastException: Illegal cast ``` Using this convertor to emit `Object` with the expected properties: ```kotlin val simple = mapOf(1 to "a").toJsObject { it.key.toString() to it.value } val json = js("JSON.stringify(simple)") as String println(json) >>> {"1":"a"} val value = js("simple['1']") as String println(value) >>> a ``` ## [Coroutines](https://juullabs.github.io/tuulbox/coroutines/index.html) ![badge-android] ![badge-ios] ![badge-js] ![badge-jvm] ![badge-mac] Utilities for [Coroutines]. ### [`combine`] [`combine`][coroutines-combine] for up to 10 [`Flow`]s: ```kotlin val combined = combine( flow1, flow2, // ... flow9, flow10, ) { value1, value2, /* ... */, value9, value10 -> // ... } ``` ## [Logging](https://juullabs.github.io/tuulbox/logging/index.html) ![badge-ios] ![badge-js] ![badge-jvm] ![badge-mac] Simple multiplatform logging API. ### Initialization Logging can be initialized via [`install`]: ```kotlin Log.dispatcher.install(ConsoleLogger) ``` If no [`Logger`] is installed, then log blocks are not called at runtime. Custom loggers can be created by implementing the [`Logger`] interface. #### Android (Logcat) Because Kotlin Multiplatform does not yet nicely support Android and JVM as a hierarchical dependency, Logcat logging is provided as a separate artifact. ```kotlin implementation("com.juul.tuulbox:logging-android:$version") ``` After adding the Gradle dependency, it can be installed as follows. ```kotlin Log.dispatcher.install(AndroidLogger) ``` #### Apple (NSLog) Log to the Apple System Log by installing the `AppleSystemLogger`. ```kotlin Log.dispatcher.install(AppleSystemLogger) ``` ### Log Log message can be logged via: ```kotlin Log.verbose { "Example" } ``` The following [log level] functions are available: - [`verbose`] - [`debug`] - [`info`] - [`warn`] - [`error`] - [`assert`] Optional `tag` and `throwable` may be provided. `tag` defaults to an autogenerated value, but behavior can be customized via [`Log.tagGenerator`] property. ### Metadata Tuulbox logging supports associating arbitrary metadata to each log. A [`WriteMetadata`] instance is passed to each of the logging functions above. ```kotlin Log.verbose { metadata -> metadata[Sensitivity] = Sensitivity.Sensitive "My social security number is 123 45 6789" } ``` This can be read inside of a [`Logger`] via the [`ReadMetadata`] passed into it. ```kotlin class SampleLogger : Logger { override fun verbose(tag: String, message: String, metadata: ReadMetadata, throwable: Throwable?) { if (metadata[Sensitivity] != Sensitivity.Sensitive) { // report to a destination that cannot include sensitive data } } // ... } ``` To create your own metadata fields, use an `object` that implements the [`Key`] interface. The value of the generic controls the type of the value in the metadata map. ```kotlin object YourMetadata : Key<String> ``` A helpful pattern for many types is using the `companion object` as the metadata key. ```kotlin enum class Sample { A, B, C; companion object : Key<Sample> } ``` ### Filtering Tuulbox implements log filtering by decorating [`Logger`]s. #### Log Level Filters Log level filters are installed with [`Logger.withMinimumLogLevel`]. Because the filtering is based on which log call is made, instead of the content of the log call, these can be used as an optimization: if all [`Logger`]s installed in the root [`DispatchLogger`] have a minimum log level higher than the log call being made, then the log block is never called. ```kotlin Log.dispatcher.install( ConsoleLogger .withMinimumLogLevel(LogLevel.Warn) ) Log.debug { "This is not called." } Log.warn { "This still gets called." } ``` #### Log Content Filters Log content filters are installed with [`Logger.withFilter`], and have full access to the content of a log. ```kotlin Log.dispatcher.install( ConsoleLogger .withFilter { tag, message, metadata, throwable -> metadata[Sensitivity] == Sensitivity.NotSensitive } ) Log.debug { "This block is evaluated, but does not get printed to the console." } Log.warn { metadata -> metadata[Sensitivity] = Sensitivity.NotSensitive "This is also evaluated, and does print to the console." } ``` ## [Functional](https://juullabs.github.io/tuulbox/functional/index.html) ![badge-ios] ![badge-js] ![badge-jvm] ![badge-mac] Utilities for manipulating functions. For a full functional ecosystem, complete with `Monad` and the like, prefer [Arrow](https://arrow-kt.io/). ## [Temporal](https://juullabs.github.io/tuulbox/temporal/index.html) ![badge-android] ![badge-js] ![badge-jvm] Toolbox of utilities for dates and times, building on [KotlinX DateTime]. Various interval [`Flow`]s are provided, such as: [`instantFlow`], [`localDateTimeFlow`], and [`localDateFlow`]. For example: ```kotlin localDateTimeFlow().collect { println("The time is now: $it") } ``` _Note: Because this is built on top of [KotlinX DateTime], [core library desugaring] must be enabled for Android targets._ ## [Test](https://juullabs.github.io/tuulbox/test/index.html) ![badge-ios] ![badge-js] ![badge-jvm] ![badge-mac] Utilities for test suites. ### [`runTest`] Multiplatform test analogy to [`runBlocking`]: ```kotlin @Test fun exampleUnitTest() = runTest { // ... } ``` ### [`assertContains`] ```kotlin assertContains( array = arrayOf(1, 2, 3), value = 2, ) ``` ```kotlin assertContains( range = 1..10, value = 5, ) ``` ### [`assertSimilar`] ```kotlin assertSimilar( target = 5, epsilon = 1, value = 6, ) ``` ## [Encoding](https://juullabs.github.io/tuulbox/encoding/index.html) ![badge-ios] ![badge-js] ![badge-jvm] ![badge-mac] Utilities for working with binary data. ### [`IntBitSet`] ```kotlin val bitSet = 0.bits bitSet[30] = true bitSet.asPrimitive() // 1073741824 ``` ```kotlin /* | Index | ... | 3 | 2 | 1 | 0 | * |-------|-----|---|---|---|---| * | Bit | ... | 1 | 0 | 1 | 0 | */ val bitSet = 10.bits bitSet[0] // false bitSet[1] // true bitSet[2] // false bitSet[3] // true ``` ### [`LongBitSet`] ```kotlin /* | Index | ... | 3 | 2 | 1 | 0 | * |-------|-----|---|---|---|---| * | Bit | ... | 1 | 1 | 0 | 0 | */ val bitSet = 12L.bits bitSet[0] // false bitSet[1] // false bitSet[2] // true bitSet[3] // true ``` ```kotlin val bitSet = 0L.bits bitSet[40] = true bitSet.asPrimitive() // 1099511627776L ``` # Setup ### Gradle [![Maven Central](https://maven-badges.herokuapp.com/maven-central/com.juul.tuulbox/coroutines/badge.svg)](https://search.maven.org/search?q=g:com.juul.tuulbox) Tuulbox can be configured via Gradle Kotlin DSL as follows: #### Multiplatform ```kotlin plugins { id("com.android.application") // or id("com.android.library") kotlin("multiplatform") } repositories { mavenCentral() } kotlin { jvm() // and/or android() js().browser() // and/or js().node() macosX64() sourceSets { val commonMain by getting { dependencies { implementation("com.juul.tuulbox:collections:$version") implementation("com.juul.tuulbox:coroutines:$version") implementation("com.juul.tuulbox:logging:$version") implementation("com.juul.tuulbox:functional:$version") implementation("com.juul.tuulbox:temporal:$version") implementation("com.juul.tuulbox:encoding:$version") } } val commonTest by getting { dependencies { implementation("com.juul.tuulbox:test:$version") } } } } ``` #### Platform-specific ```kotlin repositories { mavenCentral() } dependencies { implementation("com.juul.tuulbox:collections:$version") implementation("com.juul.tuulbox:coroutines:$version") implementation("com.juul.tuulbox:logging:$version") implementation("com.juul.tuulbox:functional:$version") implementation("com.juul.tuulbox:temporal:$version") implementation("com.juul.tuulbox:encoding:$version") testImplementation("com.juul.tuulbox:test:$version") } ``` # License ``` Copyright 2021 JUUL Labs, Inc. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. ``` [Collections]: https://kotlinlang.org/docs/reference/collections-overview.html [`SynchronizedMap`]: https://juullabs.github.io/tuulbox/collections/collections/com.juul.tuulbox.collections/-synchronized-map/index.html [`Map.toJsObject`]: https://juullabs.github.io/tuulbox/collections/collections/com.juul.tuulbox.collections/to-js-object.html [Coroutines]: https://kotlinlang.org/docs/reference/coroutines-overview.html [KotlinX DateTime]: https://github.com/Kotlin/kotlinx-datetime [core library desugaring]: https://developer.android.com/studio/write/java8-support#library-desugaring [coroutines-combine]: https://kotlin.github.io/kotlinx.coroutines/kotlinx-coroutines-core/kotlinx.coroutines.flow/combine.html [`Flow`]: https://kotlin.github.io/kotlinx.coroutines/kotlinx-coroutines-core/kotlinx.coroutines.flow/-flow/index.html [`runBlocking`]: https://kotlin.github.io/kotlinx.coroutines/kotlinx-coroutines-core/kotlinx.coroutines/run-blocking.html [`combine`]: https://juullabs.github.io/tuulbox/coroutines/coroutines/com.juul.tuulbox.coroutines.flow/combine.html [`Logger`]: https://juullabs.github.io/tuulbox/logging/logging/com.juul.tuulbox.logging/-logger/index.html [`install`]: https://juullabs.github.io/tuulbox/logging/logging/com.juul.tuulbox.logging/-dispatch-logger/install.html [log level]: https://juullabs.github.io/tuulbox/logging/logging/com.juul.tuulbox.logging/-log/index.html [`Log.tagGenerator`]: https://juullabs.github.io/tuulbox/logging/logging/com.juul.tuulbox.logging/-log/tag-generator.html [`verbose`]: https://juullabs.github.io/tuulbox/logging/logging/com.juul.tuulbox.logging/-log/verbose.html [`debug`]: https://juullabs.github.io/tuulbox/logging/logging/com.juul.tuulbox.logging/-log/debug.html [`info`]: https://juullabs.github.io/tuulbox/logging/logging/com.juul.tuulbox.logging/-log/info.html [`warn`]: https://juullabs.github.io/tuulbox/logging/logging/com.juul.tuulbox.logging/-log/warn.html [`error`]: https://juullabs.github.io/tuulbox/logging/logging/com.juul.tuulbox.logging/-log/error.html [`assert`]: https://juullabs.github.io/tuulbox/logging/logging/com.juul.tuulbox.logging/-log/assert.html [`WriteMetadata`]: https://juullabs.github.io/tuulbox/logging/logging/com.juul.tuulbox.logging/-write-metadata/index.html [`ReadMetadata`]: https://juullabs.github.io/tuulbox/logging/logging/com.juul.tuulbox.logging/-read-metadata/index.html [`Key`]: https://juullabs.github.io/tuulbox/logging/logging/com.juul.tuulbox.logging/-key/index.html [`Logger.withMinimumLogLevel`]: https://juullabs.github.io/tuulbox/logging/logging/com.juul.tuulbox.logging/with-minimum-log-level.html [`DispatchLogger`]: https://juullabs.github.io/tuulbox/logging/logging/com.juul.tuulbox.logging/-dispatch-logger/index.html [`Logger.withFilter`]: https://juullabs.github.io/tuulbox/logging/logging/com.juul.tuulbox.logging/with-filter.html [`runTest`]: https://juullabs.github.io/tuulbox/test/test/com.juul.tuulbox.test/run-test.html [`assertContains`]: https://juullabs.github.io/tuulbox/test/test/com.juul.tuulbox.test/assert-contains.html [`assertSimilar`]: https://juullabs.github.io/tuulbox/test/test/com.juul.tuulbox.test/assert-similar.html [`IntBitSet`]: https://juullabs.github.io/tuulbox/encoding/encoding/com.juul.tuulbox.encoding/-int-bit-set/index.html [`LongBitSet`]: https://juullabs.github.io/tuulbox/encoding/encoding/com.juul.tuulbox.encoding/-long-bit-set/index.html [`instantFlow`]: https://juullabs.github.io/tuulbox/temporal/temporal/com.juul.tuulbox.temporal/instant-flow.html [`localDateFlow`]: https://juullabs.github.io/tuulbox/temporal/temporal/com.juul.tuulbox.temporal/local-date-flow.html [`localDateTimeFlow`]: https://juullabs.github.io/tuulbox/temporal/temporal/com.juul.tuulbox.temporal/local-date-time-flow.html [badge-android]: http://img.shields.io/badge/platform-android-6EDB8D.svg?style=flat [badge-ios]: http://img.shields.io/badge/platform-ios-CDCDCD.svg?style=flat [badge-js]: http://img.shields.io/badge/platform-js-F8DB5D.svg?style=flat [badge-jvm]: http://img.shields.io/badge/platform-jvm-DB413D.svg?style=flat [badge-linux]: http://img.shields.io/badge/platform-linux-2D3F6C.svg?style=flat [badge-windows]: http://img.shields.io/badge/platform-windows-4D76CD.svg?style=flat [badge-mac]: http://img.shields.io/badge/platform-macos-111111.svg?style=flat [badge-watchos]: http://img.shields.io/badge/platform-watchos-C0C0C0.svg?style=flat [badge-tvos]: http://img.shields.io/badge/platform-tvos-808080.svg?style=flat [badge-wasm]: https://img.shields.io/badge/platform-wasm-624FE8.svg?style=flat
33.940452
1,813
0.744268
yue_Hant
0.290967
218f10e9a07f622faa58d669df5c1784223a07a3
83
md
Markdown
README.md
Vido/junopy
2c3753105cd1b9356aa30e5567b5771d5d013dee
[ "Apache-2.0" ]
1
2021-05-06T01:06:18.000Z
2021-05-06T01:06:18.000Z
README.md
Vido/junopy
2c3753105cd1b9356aa30e5567b5771d5d013dee
[ "Apache-2.0" ]
null
null
null
README.md
Vido/junopy
2c3753105cd1b9356aa30e5567b5771d5d013dee
[ "Apache-2.0" ]
null
null
null
# junopy Cliente Python para a nova API de Integração da Juno (BoletoBancário.com)
27.666667
73
0.795181
por_Latn
0.996112
218ff810bf0860c15a3c95defd09ae6570115fcc
5,108
md
Markdown
README.md
linbudu599/Apollo-API-Explorer
ff11b62f2ff9ad4012b49060d4c3ef2edd5c2890
[ "MIT" ]
6
2020-12-18T07:37:25.000Z
2021-12-26T07:21:25.000Z
README.md
linbudu599/Apollo-API-Explorer
ff11b62f2ff9ad4012b49060d4c3ef2edd5c2890
[ "MIT" ]
17
2020-12-02T07:38:34.000Z
2020-12-13T04:05:24.000Z
README.md
linbudu599/GraphQL-Explorer
ff11b62f2ff9ad4012b49060d4c3ef2edd5c2890
[ "MIT" ]
2
2021-04-24T12:00:45.000Z
2021-09-07T03:52:44.000Z
# GraphQL-Explorer-Server ![GitHub Workflow Status](https://img.shields.io/github/workflow/status/linbudu599/GraphQL-Explorer-Server/GraphQL-Explorer-Server) ![Codecov](https://img.shields.io/codecov/c/github/linbudu599/GraphQL-Explorer-Server) ![GitHub last commit](https://img.shields.io/github/last-commit/linbudu599/GraphQL-Explorer-Server) ![David](https://img.shields.io/david/dev/linbudu599/GraphQL-Explorer-Server?label=dependencies) ![GitHub package.json dependency version (subfolder of monorepo)](https://img.shields.io/github/package-json/dependency-version/linbudu599/GraphQL-Explorer/graphql) ![GitHub package.json dependency version (subfolder of monorepo)](https://img.shields.io/github/package-json/dependency-version/linbudu599/GraphQL-Explorer/type-graphql) ![GitHub package.json dependency version (subfolder of monorepo)](https://img.shields.io/github/package-json/dependency-version/linbudu599/GraphQL-Explorer/apollo-server) This is a **demo** which contains `GraphQL` and framework based on it(or work with it perfectly), including these for now: ## 🚀&nbsp; Explore ```bash npm install npm run dev ``` ## 🛵&nbsp; Available Scripts - `dev`: develop server by [nodemon](https://github.com/remy/nodemon). - `build`: compile server code by [tsc](https://github.com/microsoft/TypeScript). - `prebuild`: Generation of `GenQL` & `Prisma Client`. - `postbuild`: copy generated `Prisma Client ` files, run `prisma db push` to create SQLite file. - `build:vercel`: build scripts on Vercel Functions. - `start`: run in prod by [nodemon](https://github.com/remy/nodemon)(recommended to run before deployment). - `clean`: remove `/dist`(build output) & `api-dist`(Vercel build output) directory. - `test`: run test cases by [Jest](https://github.com/facebook/jest). - `pm2`: run in prod by [PM2](https://github.com/Unitech/pm2). - `stress-test`: run pressure test to server, see [stress-fork.js](./st-fork.js) & [stress-main.js](./st-main.js). - `voyager`: represent GraphQL API as an interactive graph by [GraphQL-Voyager](https://github.com/APIs-guru/graphql-voyager), **require local server to be active**. - `gen:code`: generate type definitions from GraphQL schema by [GraphQL-Code-Generator](https://github.com/dotansimha/graphql-code-generator). - `gen:docs`: generate documentation site from GraphQL schema by [GraphDoc](https://github.com/2fd/graphdoc). - `gen`: generate docs & code, as `gen:code` & `gen:docs` require local server alive, **this command will use NodeJS `child_process` mod to handle server setup.** - `serve:docs`: generate & serve documentation site by [serve](https://github.com/vercel/serve). - `commit`: commit code by [custom commitlint config](.cz-config.js). - `prettier`: prettier all `.ts` / `.json` / `.graphql` ext file. - `typecheck`: check TypeScript type definitions. - `seed:config`: check [typeorm-seeding](https://github.com/w3tecch/typeorm-seeding) config. - `seed:run`: start seeding data in `/server/entity/seeds`. - `genql`: generate GraphQL query builder by [GenQL](https://github.com/remorses/genql). - `prisma:*` [Prisma 2](https://www.prisma.io/) Related Commands. ### Main - [x] GraphQL - [x] TypeGraphQL - [x] Apollo-Server - [x] DataLoader (Supported By [TypeGraphQL-DataLoader](https://github.com/slaypni/type-graphql-dataloader)) - [x] TypeORM + SQLite3 - [x] Prisma 2(As replacement for TypeORM) - [ ] Subscription ### Other Powerful Tools - [x] [GraphQLDoc](https://github.com/2fd/graphdoc) - [x] [GraphQL-Voyager](https://github.com/APIs-guru/graphql-voyager) - [x] [GenQL](https://github.com/remorses/genql) - [x] [GraphQL-Code-Generator](https://github.com/dotansimha/graphql-code-generator) ## Features - [x] TypeORM Entity Relations [Setup](./server/entities): 1-1 1-m m-n - [x] Most API of TypeGraphQL: InputType / InterfaceType / AuthChecker / Middlewares / Containers / ... - [x] Configurated Apollo-Server With [Plugins](./server/plugins) - [x] Visualized API Graph by GraphQLDoc & GraphQL-Voyager - [x] Useful [Directives](./server/directives) & [Decorators](./server/decorators) ## 🏗️ &nbsp; Server & Docs Deployment - [x] [Docs by Surge](http://graphql-explorer-docs.surge.sh/) - [x] [Voyager by Surge](http://graphql-explorer-voyager.surge.sh/) **(Require Local Server To Be Active)** - [x] [Voyager on Remote Server](http://voyager.linbudu.top/) - [x] [Simple Example on Vercel Fucntions](https://graphql-faas.vercel.app/api/sample), see [Query Example](./api/sample/query.graphql) - [x] [Simple Example of Vercel Fucntions, with MySQL](https://graphql-faas.linbudu599.vercel.app/api/graphql) - [x] [Full Server Deployed at ALI Cloud ECS](http://47.97.183.158:4399/graphql) - [x] [Apollo Engine(**private**)](https://studio.apollographql.com/graph/My-Graph-innqj/explorer?schemaTag=current) ## 📔&nbsp; Document see offcial docs for more information. - [GraphQL](https://graphql.org/) - [Apollo-GraphQL](https://www.apollographql.com/docs/) - [TypeGraphQL](https://typegraphql.com/) - [TypeStack](https://github.com/typestack) - [TypeORM](https://github.com/typeorm) - [Prisma](https://www.prisma.io/)
55.521739
170
0.733555
kor_Hang
0.235179
2190c1ddb276e0ee460313e6eef8a505cdaeda80
50
md
Markdown
README.md
qapabilities/cypressbasicoTAT
326b2d626725fe4b2c596ce9f3f4788c0752c3a9
[ "MIT" ]
null
null
null
README.md
qapabilities/cypressbasicoTAT
326b2d626725fe4b2c596ce9f3f4788c0752c3a9
[ "MIT" ]
null
null
null
README.md
qapabilities/cypressbasicoTAT
326b2d626725fe4b2c596ce9f3f4788c0752c3a9
[ "MIT" ]
null
null
null
# cypressbasicoTAT Curso Básico de Cypress da TAT
16.666667
30
0.82
por_Latn
0.70585
219195ebba130a88dc54543b5b5bfebf2011e701
26
md
Markdown
README.md
christelle101/DATA731-Change-Detection
1713048eafe122326c7b1ed686d1af45dbac73eb
[ "MIT" ]
null
null
null
README.md
christelle101/DATA731-Change-Detection
1713048eafe122326c7b1ed686d1af45dbac73eb
[ "MIT" ]
null
null
null
README.md
christelle101/DATA731-Change-Detection
1713048eafe122326c7b1ed686d1af45dbac73eb
[ "MIT" ]
null
null
null
# DATA731-Change-Detection
26
26
0.846154
kor_Hang
0.711486
2191d29f4747f1ef78af1d619ea439a656afbf90
7,166
md
Markdown
writeup.md
platypova/CarND-Behavioral-Cloning
704920a4bad0f56d2f9b3a459d5374bda3281289
[ "MIT" ]
null
null
null
writeup.md
platypova/CarND-Behavioral-Cloning
704920a4bad0f56d2f9b3a459d5374bda3281289
[ "MIT" ]
null
null
null
writeup.md
platypova/CarND-Behavioral-Cloning
704920a4bad0f56d2f9b3a459d5374bda3281289
[ "MIT" ]
null
null
null
# **Behavioral Cloning** **Behavioral Cloning Project** The goals / steps of this project are the following: * Use the simulator to collect data of good driving behavior * Build, a convolution neural network in Keras that predicts steering angles from images * Train and validate the model with a training and validation set * Test that the model successfully drives around track one without leaving the road * Summarize the results with a written report [//]: # (Image References) [image1]: ./examples/right.jpg "Right camera image" [image2]: ./examples/right_flipped.jpg "Right camera image flipped" [image3]: ./examples/left.jpg "Left camera image" [image4]: ./examples/left_flipped.jpg "Left camera image flipped" [image5]: ./examples/center.jpg "Center camera image" [image6]: ./examples/center_flipped.jpg "Center camera image flipped" [Nvidia paper]: http://images.nvidia.com/content/tegra/automotive/images/2016/solutions/pdf/end-to-end-dl-using-px.pdf ## Rubric Points ### Here I will consider the [rubric points](https://review.udacity.com/#!/rubrics/432/view) individually and describe how I addressed each point in my implementation. --- ### Files Submitted & Code Quality #### 1. Submission includes all required files and can be used to run the simulator in autonomous mode My project includes the following files: * model.py containing the script to create and train the model * drive.py for driving the car in autonomous mode * model.h5 containing a trained convolution neural network * writeup_report.md or writeup_report.pdf summarizing the results #### 2. Submission includes functional code Using the Udacity provided simulator and my drive.py file, the car can be driven autonomously around the track by executing ```sh python drive.py ../model.h5 ``` (in current directory) #### 3. Submission code is usable and readable The model.py file contains the code for training and saving the convolution neural network. The file shows the pipeline I used for training and validating the model, and it contains comments to explain how the code works. ### Model Architecture and Training Strategy #### 1. An appropriate model architecture has been employed My model consists of a convolution neural network with 5x5 and 3x3 filter sizes and depths between 24 and 72 (model.py lines 96-112). There are 5 convolutional layers and 4 fully connected layers. The model also includes RELU layers to introduce nonlinearity (code line 107). The data is normalized in the model using a Keras lambda layer (code line 91). #### 2. Attempts to reduce overfitting in the model The model contains dropout layers in order to reduce overfitting (model.py line 102). The model was tested by running it through the simulator and ensuring that the vehicle could stay on the track. #### 3. Model parameter tuning The model used an adam optimizer, so the learning rate was not tuned manually (model.py line 114). #### 4. Appropriate training data Training data was chosen to keep the vehicle driving on the road. I used a dataset offered by Udacity. For details about how I created the training data, see the section Creation of the Training Set & Training Process. ### Model Architecture and Training Strategy #### 1. Solution Design Approach My first step was to use a convolution neural network model based on famous network by Nvidia ([Nvidia paper]). It was developed for the same task. In order to gauge how well the model was working, I split my image and steering angle data into a training and validation set. To avoid the overfitting, I modified the model by adding dropout layer (model.py line 102). The final step was to run the simulator to see how well the car was driving around track one. There were a few spots where the vehicle fell off the track. In order to improve the driving behavior in these cases, I tried to change the architecture somehow: I tried adding convolutional layers and change filter depth. I noticed that decreasing the number of filters in the first layer decresed the quality of driving. Increasing the number of filters remarkably in the final convolutional layer also influenced badly. Too much dropout layers also influence badly. I also tried different values for offset adding to angles measurements, and it influenced the car ability to recover from going too far left or right, and at the same time could lead to too much sinusoidal trajectory. At the end of the process, the vehicle is able to drive autonomously around the track without leaving the road. #### 2. Final Model Architecture The final model architecture (model.py lines 96-114) consisted of a convolution neural network with the following layers and layer sizes: Convolutional layer 1: depth - 24, filter - (5,5), strides - (2,2), activation with 'relu' function onvolutional layer 2: depth - 36, filter - (5,5), strides - (2,2),activation with 'relu' function onvolutional layer 3: depth - 48, filter - (5,5), strides - (2,2),activation with 'relu' function onvolutional layer 4: depth - 64, filter - (3,3), strides - (1,1),activation with 'relu' function onvolutional layer 5: depth - 64, filter - (3,3), strides - (1,1),activation with 'relu' function Dropout: keep probability 0.8 Flatten Dense, output - 100 Activation with 'relu' Dense, output - 50 Activation with 'relu' Dense, output - 10 Activation with 'relu' Dense, output - 1 #### 3. Creation of the Training Set & Training Process I used dataset offered by Udacity. The data are loaded in line 82 (function load_dataset, lines 22-29). I took all images, including images from central, left and right cameras. I put 20% of the data into a validation set (line 84). Then I used generator (as it was shown in one of the lessons) to load data and preprocess it on the fly, in batch size portions to feed into my model.It is described in function generator (lines 46-79). It creates batches of dataset contatinig images and angles: center camera images were used with corresponding angles data, while for the left and right camera images I used angle for center camera with offset of +-0.3. It also flippes all images and multiplies corresponding angles to -1. So, here it can be seen an augmentation strategy: the number of images becomes 2 times larger due to the flipping; and a recovery from left or right becomes possible due to adding images for left and right cameras with measurements for center camera + constant offset (here it is 0.3). I also preprocessed data by normalizing (line 91, keras Lambda module) and cropping the images (line 93, keras Cropping2D module) right in the model. Examples of initial images for 3 cameras: ![alt text][image1] ![alt text][image3] ![alt text][image5] Examples of flipped images for 3 cameras: ![alt text][image2] ![alt text][image4] ![alt text][image6] I used this training data for training the model. The validation set helped determine if the model was over or under fitting. The ideal number of epochs was 5. I used an adam optimizer so that manually training the learning rate wasn't necessary. Video showing the example of the car driving on the base of my model is in current directory (video.mp4).
55.984375
524
0.773514
eng_Latn
0.998806
219292a23c21f7a908edbd3f720a8173e5d9b5f5
138
md
Markdown
translations/zh-CN/data/reusables/project-management/linked-repositories.md
kyawburma/docs
0ff7de03be7c2432ced123aca17bfbf444bee1bf
[ "CC-BY-4.0", "MIT" ]
11,698
2020-10-07T16:22:18.000Z
2022-03-31T18:54:47.000Z
translations/zh-CN/data/reusables/project-management/linked-repositories.md
kyawburma/docs
0ff7de03be7c2432ced123aca17bfbf444bee1bf
[ "CC-BY-4.0", "MIT" ]
8,317
2020-10-07T16:26:58.000Z
2022-03-31T23:24:25.000Z
translations/zh-CN/data/reusables/project-management/linked-repositories.md
kyawburma/docs
0ff7de03be7c2432ced123aca17bfbf444bee1bf
[ "CC-BY-4.0", "MIT" ]
48,204
2020-10-07T16:15:45.000Z
2022-03-31T23:50:42.000Z
1. (可选)在 **Linked repositories(链接的仓库)**下,搜索要链接到项目板的仓库。 ![链接仓库的搜索字段](/assets/images/help/projects/search-to-link-repository-on-create.png)
69
137
0.768116
yue_Hant
0.158744
21929a75ad71f3cc8ac356b5ae1f4587265f9506
102
md
Markdown
docker/README.md
ysoftman/test_code
4c71cc7c6a17d73cc84298e3a44051d3ab9d40f8
[ "MIT" ]
3
2017-12-07T04:29:36.000Z
2022-01-11T10:58:14.000Z
docker/README.md
ysoftman/test_code
4c71cc7c6a17d73cc84298e3a44051d3ab9d40f8
[ "MIT" ]
14
2018-07-17T05:16:42.000Z
2022-03-22T00:43:47.000Z
docker/README.md
ysoftman/test_code
4c71cc7c6a17d73cc84298e3a44051d3ab9d40f8
[ "MIT" ]
null
null
null
# ysoftman docker - 개인 테스트 용도의 도커 환경 - <https://hub.docker.com/?namespace=ysoftman> - 세부 내용은 스크립트 참고
17
46
0.696078
kor_Hang
1.000009
2192c250a69a8b47da015b29af2410058c7e12fa
63
md
Markdown
README.md
tardunge/talaria-spark-connector
f33d152930d45db2d5a9141ba3497ee4a8a29e1b
[ "Apache-2.0" ]
null
null
null
README.md
tardunge/talaria-spark-connector
f33d152930d45db2d5a9141ba3497ee4a8a29e1b
[ "Apache-2.0" ]
null
null
null
README.md
tardunge/talaria-spark-connector
f33d152930d45db2d5a9141ba3497ee4a8a29e1b
[ "Apache-2.0" ]
null
null
null
# talaria-spark-sql A spark datasource connector to talaria-db
21
42
0.809524
eng_Latn
0.51112
2192e9b35bfa2473b356fbfb23cf1b02328196df
7,947
md
Markdown
README.md
ModischFabrications/VerticalSeekBar
8945b03db3a16885f3f3f93656f03d8b17d62e18
[ "Apache-2.0" ]
null
null
null
README.md
ModischFabrications/VerticalSeekBar
8945b03db3a16885f3f3f93656f03d8b17d62e18
[ "Apache-2.0" ]
null
null
null
README.md
ModischFabrications/VerticalSeekBar
8945b03db3a16885f3f3f93656f03d8b17d62e18
[ "Apache-2.0" ]
null
null
null
VerticalSeekBar =============== [![License](https://img.shields.io/badge/License-Apache%202.0-blue.svg)](https://opensource.org/licenses/Apache-2.0) [![Platform](https://img.shields.io/badge/platform-android-green.svg)](http://developer.android.com/index.html) [![API](https://img.shields.io/badge/API-16%2B-brightgreen.svg?style=flat)](https://android-arsenal.com/api?level=16) [![Android Weekly](https://img.shields.io/badge/Android%20Weekly-%23377-orange.svg)](https://androidweekly.net/issues/issue-377) [![Download](https://api.bintray.com/packages/lukelorusso/maven/com.lukelorusso:verticalseekbar/images/download.svg?version=1.2.0) ](https://bintray.com/lukelorusso/maven/com.lukelorusso:verticalseekbar/1.2.0/link) ## Presentation ## ![Demo](press/demo.gif) This is the source code of an Android library: `-=:[ VerticalSeekBar ]:=-` 📃 Check also the [**Medium article**](https://blog.usejournal.com/android-a-custom-vertical-seekbar-how-to-reinvent-the-wheel-c07f28960f8f?source=friends_link&sk=eb8917b7f046a2152066ce8b58c81c7b) about it! - - - ## Why would you need it? ## *"Reinventing the wheel is, most of the time, a bad idea."* If you've ever tried to make a vertical SeekBar work properly, you know this is not the case. 😏 From the moment you apply that `android:rotation="270"` all start to get messy: you cannot set the proper height for the drawable; the width of the bar distorts the width of your drawable; even customizing your interaction with the bar is kind of tricky. I've been through all of this and suddenly I've had enough. **Introducing a nicer, redesigned and highly customizable VerticalSeekBar.** What you got: - custom progress drawable: apply resources or color gradients - custom thumb: chose your size, background color, corner radius or even to completely hide it - view width and drawable width completely separated - choose how to set your progress (just sliding the thumb, clicking around, or both) - min and max placeholders - much more! - - - ## How to use it? ## Make sure to include the library in your app's build.gradle: ```groovy implementation 'com.lukelorusso:verticalseekbar:1.2.0' ``` Add the view to your layout: ```xml <com.lukelorusso.verticalseekbar.VerticalSeekBar android:layout_width="160dp" android:layout_height="280dp"/> ``` maybe add some attributes... here you got some, we'll discuss them later ``` ... app:vsb_click_to_set_progress="false" app:vsb_bar_corner_radius="15dp" app:vsb_bar_width="60dp" app:vsb_bar_background="#eeeeee" app:vsb_bar_progress_gradient_end="#4dd0e1" app:vsb_bar_progress_gradient_start="#03a2ba" app:vsb_max_placeholder_position="inside" app:vsb_max_placeholder_src="@drawable/max_placeholder" app:vsb_min_placeholder_position="inside" app:vsb_min_placeholder_src="@drawable/min_placeholder" app:vsb_progress="50" app:vsb_show_thumb="true" app:vsb_thumb_container_corner_radius="5dp" app:vsb_thumb_container_tint="@color/placeholderBackground" app:vsb_thumb_placeholder_src="@drawable/thumb_placeholder" app:vsb_use_thumb_to_set_progress="true" ... ``` All of them can be also set programmatically. (Please note: it's strongly recommended to avoid pre-1.2.0 versions due to inconsistencies compared to this guide) - - - # Attributes # ## Background and Progress ## To set your progress value programmatically: ```kotlin mainVerticalSeekBar.progress = 75 ``` Do you need a smaller or bigger range? You can change `maxValue` like this: ```kotlin mainVerticalSeekBar.maxValue = 70 // "progress" will be automagically reduced! ``` Let's put some color on this view, shall we? You can set a drawable for the background and a drawable for the progress. If you don't, default colors will be applied. For both **background** and **progress** you have 3 choices: - set a start color and an end color, the view will create a gradient for you - set a drawable resource or color (REMEMBER: this choice has priority on the one before) - just don't choose (it's still a choice anyway 😉) Example for the first choice: apply a gradient ```kotlin mainVerticalSeekBar.barBackgroundStartColor = ContextCompat.getColor(this, R.color.my_background_start_color) mainVerticalSeekBar.barBackgroundEndColor = ContextCompat.getColor(this, R.color.my_background_end_color) mainVerticalSeekBar.barProgressStartColor = ContextCompat.getColor(this, R.color.my_progress_start_color) mainVerticalSeekBar.barProgressEndColor = ContextCompat.getColor(this, R.color.my_progress_end_color) ``` Example for the second choice: apply a resource (will eventually override the first one) ```kotlin mainVerticalSeekBar.barBackgroundDrawable = getDrawable(R.color.my_background_color) mainVerticalSeekBar.barProgressDrawable = getDrawable(R.drawable.my_progress) ``` Your bar can also have rounded corners... ```kotlin mainVerticalSeekBar.barCornerRadius = 40 // those are pixels ``` ...and a proper independent width ```kotlin mainVerticalSeekBar.barWidth = 30 // those are pixels ``` ## Placeholders ## To set minimum and/or maximum placeholders (`null` is a possible value to remove them): ```kotlin mainVerticalSeekBar.minPlaceholderDrawable = getDrawable(R.drawable.my_min_placeholder) mainVerticalSeekBar.maxPlaceholderDrawable = getDrawable(R.drawable.my_max_placeholder) ``` Since v1.1.4 you can also choose placeholders' position You can choose between `{"inside", "outside", "middle"}` (`"middle"` by default) Programmatically, it will be: ```kotlin mainVerticalSeekBar.minPlaceholderPosition = VerticalSeekBar.Placeholder.INSIDE mainVerticalSeekBar.maxPlaceholderPosition = VerticalSeekBar.Placeholder.OUTSIDE ``` Now about the thumb placeholder 👆 It is child of a `androidx.cardview.widget.CardView`. You can choose the color tint and the corner radius of the CardView: ```kotlin mainVerticalSeekBar.thumbContainerColor = ContextCompat.getColor(this, R.color.my_thumb_background_color) mainVerticalSeekBar.thumbContainerCornerRadius = 15 // those are pixels ``` You can set your thumb drawable like this: ```kotlin mainVerticalSeekBar.thumbPlaceholderDrawable = getDrawable(R.drawable.my_thumb_placeholder) ``` If you just don't want to see it: ```kotlin mainVerticalSeekBar.showThumb = false ``` ## Interaction ## You can interact with your VerticalSeekBar in two ways: - sliding the thumb - tapping on the bar Both those interactions can be disabled if you like (they are `true` by default). For instance, It can be useful to disable sliding in case you need to put the VerticalSeekBar on a ScrollView. To do so: ```kotlin mainVerticalSeekBar.useThumbToSetProgress = false mainVerticalSeekBar.clickToSetProgress = true ``` Try to change those booleans too see other possibilities! ## Still want more? ## If this level of customization is not enough, I have one last good news for you: the seekbar layout is ENTIRELY CUSTOMIZABLE! You can pick the original [**layout_verticalseekbar.xml**](/verticalseekbar/src/main/res/layout/layout_verticalseekbar.xml) and put it in your project's `res/layout` folder. Now you can modify it as you wish; just remember to keep all the original ids! 😉 (`thumbCardView` and `thumbPlaceholder` can be safely removed) - - - # Explore! # Feel free to checkout and launch the example app 🎡 Also, see where using this library has been the perfect choice: [![ColorBlindClick](press/colorblindclick_launcher.png)](https://play.google.com/store/apps/details?id=com.lukelorusso.colorblindclick) [**ColorBlindClick**](https://play.google.com/store/apps/details?id=com.lukelorusso.colorblindclick) - - - # Copyright # Make with 💚 by [Luca Lorusso](http://lukelorusso.com), licensed under [Apache License 2.0](http://www.apache.org/licenses/LICENSE-2.0) Thanks for the help to [Lopez Mikhael](http://mikhaellopez.com/)
39.147783
256
0.762678
eng_Latn
0.906341
2193879d716ee368668e8bea750cafa073355e39
2,527
md
Markdown
bigbench/benchmark_tasks/symbol_interpretation/adversarial/README.md
sebschu/BIG-bench
b04e56e3a9c0b6488fe154370de0b9697f991b67
[ "Apache-2.0" ]
460
2021-01-26T21:23:37.000Z
2022-03-31T15:55:10.000Z
bigbench/benchmark_tasks/symbol_interpretation/adversarial/README.md
sebschu/BIG-bench
b04e56e3a9c0b6488fe154370de0b9697f991b67
[ "Apache-2.0" ]
475
2021-01-27T01:29:01.000Z
2022-03-30T20:23:56.000Z
bigbench/benchmark_tasks/symbol_interpretation/adversarial/README.md
sebschu/BIG-bench
b04e56e3a9c0b6488fe154370de0b9697f991b67
[ "Apache-2.0" ]
305
2021-01-26T23:28:11.000Z
2022-03-31T14:51:26.000Z
# Symbol Interpretation Task - Adversarial Subtask Our task is composed of five different subtasks which require interpreting statements referring to structures of a simple world. This world is built using emojis; a structure of the world is simply a sequence of six emojis. Crucially, in every variation, we explicit the semantic link between the emojis and their name in a different way: - **Plain**: "🟥 is a red square" - **Adversarial**: "🟡 is a red square" - **Tricky**: "🟥 is a der reauqs" - **Agnostic name-side**: "🟥 is a X Y" - **Agnostic emoji-side**: "🐮 is a red square" For a detailed explanation of our task, please refer to this [document](../README.md). **Authors**: Antonio Norelli, Giorgio Mariani, Andrea Santilli, Luca Moschella, Giambattista Parascandolo, Simone Melzi, Emanuele Rodolà. The following is an example of the **adversarial subtask**: > In the SIT-adversarial world a structure is a sequence of six emojis. > Hereafter are reported the emojis used along with their descriptions: > - 🔺 is a red circle; > - 🟦 is a blue circle; > - 🔴 is a yellow circle; > - 🟥 is a red triangle pointing up; > - 🟨 is a red triangle pointing down; > - 🔻 is a red square; > - 🟡 is a blue square; > - `_` is a yellow square; > - 🔵 is an empty space. > > *Choose the sentence consistent with the structure* `🔵 🔺 🔴 🔴 _ 🔻` *and not consistent with the structure* `🟨 🟨 🟡 🔺 🟥 🟨`: > > - [ ] There is at least one red piece at the left of a circle. > - [ ] There are at least two yellow circles at the right of a yellow square. > - [x] There are zero triangles touching a blue square. > - [ ] There are exactly two triangles pointing down at the right of a triangle pointing down. > - [ ] There are exactly two yellow squares at the right of a triangle pointing up. ## Behavior of existing language models on subtask The model GPT-2 has a “multiple_choice_grade” accuracy of around 17.6%, thus lower then random guessing (20%), suggesting that the task is difficult for at least some language models. Preliminary results obtained with form submission show humans can easily solve the task we propose. We tested also zero-shot capabilities of the models [RoBERTa](https://huggingface.co/roberta-large-mnli) and [BART](https://huggingface.co/facebook/bart-large-mnli) finetuned on [MultiNLI](https://huggingface.co/datasets/multi_nli). These models have a “multiple_choice_grade” accuracy respectively of 15.6% for RoBERTa and 23.2% for BART. We haven’t yet tested the performance on larger models like GPT-3 or T5.
56.155556
339
0.734072
eng_Latn
0.99905
219420598fc503485c6998688f97f8ccf1c5d46f
27,521
md
Markdown
node_modules/nextgen-events/documentation.md
myste1tainn/Thesis-ARS
c82e6cf50e3c16a3c882a628f3d132dcc203a737
[ "MIT" ]
null
null
null
node_modules/nextgen-events/documentation.md
myste1tainn/Thesis-ARS
c82e6cf50e3c16a3c882a628f3d132dcc203a737
[ "MIT" ]
null
null
null
node_modules/nextgen-events/documentation.md
myste1tainn/Thesis-ARS
c82e6cf50e3c16a3c882a628f3d132dcc203a737
[ "MIT" ]
null
null
null
# NextGen Events Next generation of events handling for node.js * License: MIT * Current status: close to release * Platform: Node.js and browsers *NextGen Events* solves common trouble that one may encounter when dealing with events and listeners. ## Feature highlights: * Standard event-handling almost compatible with Node.js built-in events * .emit() support a completion callback * Support for asynchronous event-handling * Multiple listeners can be tied to a single context * A context can be temporarly *disabled* * A context can be in *queue* mode: events for its listeners are stored, they will be *resumed* when the context is enabled again * Context serialization: async listeners can be run one after the other is fully completed * **NEW: proxy services!** Abstract away your network: emit and listen to emitter on the other side of the plug! Emitting events asynchronously or registering a listener that will be triggered asynchronously because it performs non-critical tasks has some virtues: it gives some breath to the event-loop, so important I/O can be processed as soon as possible. Contexts are really useful, it handles a collection of listeners. At first glance, it looks like a sort of namespace for listeners. But it can do more than that: you can turn a context off, so every listener tied to this context will not be triggered anymore, then turn it on and they will be available again. You can even switch a context into queue mode: the listeners tied to it will not be triggered, but events for those listeners will be stored in the context. When the context is resumed, all retained events will trigger their listeners. This allow one to postpone some operations, while performing some other high priority tasks, but be careful: depending on your application nature, the queue may grow fast and consumes a lot of memory very quickly. One of the top feature of this lib is the context serialization: it greatly eases the flow of the code! When differents events can fire at the same time, there are use cases when one does not want that async listener run concurrently. The context serialization feature will ensure you that no concurrency will happen for listeners tied to it. You do not have to code fancy or complicated tests to cover all cases anymore: just let *NextGen Events* do it for you! **Proxy services are awesome.** They abstract away the network so we can emit and listen to emitter on the other side of the plug! Both side of the channel create a Proxy, and add to it local and remote *services*, i.e. event emitters, and that's all. A remote service looks like a normal (i.e. local) emitter, and share the same API (with few limitations). It's totally protocol agnostic, you just define two methods for your proxy: one to read from the network and one to send to it (e.g. for Web Socket, this is a one-liner). # Install Use npm: ``` npm install nextgen-events ``` # Getting started By the way you can create an event emitter simply by creating a new object, this way: ```js var NGEmitter = require( 'nextgen-events' ) ; var emitter = new NGEmitter() ; ``` You can use `var emitter = Object.create( NGEmitter.prototype )` as well, the object does not need the constructor. But in real life, you would make your own objects inherit it: ```js var NGEmitter = require( 'nextgen-events' ) ; function myClass() { // myClass constructor code here } myClass.prototype = Object.create( NGEmitter.prototype ) ; myClass.prototype.constructor = myClass ; // restore the constructor // define other methods for myClass... ``` The basis of the event emitter works like Node.js built-in events: ```js var NGEmitter = require( 'nextgen-events' ) ; var emitter = new NGEmitter() ; // Normal listener emitter.on( 'message' , function( message ) { console.log( 'Message received: ' , message ) ; } ) ; // One time listener: emitter.once( 'close' , function() { console.log( 'Connection closed!' ) ; } ) ; // The error listener: if it is not defined, the error event will throw an exception emitter.on( 'error' , function( error ) { console.log( 'Shit happens: ' , error ) ; } ) ; emitter.emit( 'message' , 'Hello world!' ) ; // ... ``` # References Node.js documentation: > When an EventEmitter instance experiences an error, the typical action is to emit an 'error' event. > Error events are treated as a special case in node. If there is no listener for it, > then the default action is to print a stack trace and exit the program. > All EventEmitters emit the event 'newListener' when new listeners are added and 'removeListener' when a listener is removed. For the 'newListener' and 'removeListener' events, see the section about [incompatibilities](#incompatibilities), since there are few differences with the built-in Node.js EventEmitter. ## Table of Content * [Events](#ref.events) * [.addListener() / .on()](#ref.events.addListener) * [.once()](#ref.events.once) * [.removeListener() / .off()](#ref.events.removeListener) * [.removeAllListeners()](#ref.events.removeAllListeners) * [.setMaxListeners()](#ref.events.setMaxListeners) * [.listeners()](#ref.events.listeners) * [.listenerCount()](#ref.events.listenerCount) * [.setNice()](#ref.events.setNice) * [.emit()](#ref.events.emit) * [.addListenerContext()](#ref.events.addListenerContext) * [.disableListenerContext()](#ref.events.disableListenerContext) * [.queueListenerContext()](#ref.events.queueListenerContext) * [.enableListenerContext()](#ref.events.enableListenerContext) * [.setListenerContextNice()](#ref.events.setListenerContextNice) * [.serializeListenerContext()](#ref.events.serializeListenerContext) * [.destroyListenerContext()](#ref.events.destroyListenerContext) * [the *nice feature*](#ref.note.nice) * [incompatibilities](#incompatibilities) * [Proxy Services](#ref.proxy) <a name="ref.events"></a> ## Events <a name="ref.events.addListener"></a> ### .addListener( eventName , [fn] , [options] ) *or* .on( eventName , [fn] , [options] ) * eventName `string` the name of the event to bind to * fn `Function` the callback function for this event, this argument is optional: it can be passed to the `fn` property of `options` * options `Object` where: * fn `Function` (mandatory if no `fn` argument provided) the listener function * id `any type` (default to the provided *fn* function) the identifier of the listener, useful if we have to remove it later * once `boolean` (default: false) *true* if this is a one-time-listener * context `string` (default: undefined - no context) a non-empty string identifying a context, if defined the listener will be tied to this context, if this context is unexistant, it will be implicitly defined with default behaviour * nice `integer` (default: -Infinity) see [the nice feature](#ref.note.nice) for details * async `boolean` (default: false) set it to *true* if the listener is async by nature and a context serialization is wanted, when *async* is set for a listener, it **MUST** accept a completion callback as its last argument. * eventObject `boolean` (default: false) if set, the listener will be passed an unique argument: the very same event object that is returned by `.emit()`, if the listener is async, a second argument is passed as the callback Node.js documentation: > Adds a listener to the end of the listeners array for the specified event. > No checks are made to see if the listener has already been added. > Multiple calls passing the same combination of event and listener will result in the listener being added multiple times. ```js server.on( 'connection' , function( stream ) { console.log( 'someone connected!' ) ; } ) ; ``` > Returns emitter, so calls can be chained. Example, creating implicitly a context the listeners will be tied to: ```js server.on( 'connection' , { context: 'ctx' , fn: function( stream ) { console.log( 'someone connected!' ) ; } } ) ; server.on( 'close' , { context: 'ctx' , fn: function() { console.log( 'connection closed!' ) ; // Destroy the context and all listeners tied to it: server.destroyListenerContext( 'ctx' ) ; } } ) ; server.on( 'error' , { context: 'ctx' , fn: function() { // some error handling code // Destroy the context and all listeners tied to it: server.destroyListenerContext( 'ctx' ) ; } } ) ; ``` When an async listener is defined, the completion callback is automatically added at the end of the arguments supplied to [.emit()](#ref.events.emit) for any listeners with *async = true*. <a name="ref.events.once"></a> ### .once( eventName , listener ) * eventName `string` the name of the event to bind to * listener `Function` or `Object` the listener that will listen to this event, it can be a function or an object where: * fn `Function` (mandatory) the listener function * id `any type` (default to the provided *fn* function) the identifier of the listener, useful if we have to remove it later * context `string` (default: undefined - no context) a non-empty string identifying a context, if defined the listener will be tied to this context, if this context is unexistant, it will be implicitly defined with default behaviour * nice `integer` (default: -Infinity) see [the nice feature](#ref.note.nice) for details * async `boolean` (default: false) set it to *true* if the listener is async by nature and a context serialization is wanted Node.js documentation: > Adds a **one time** listener for the event. > This listener is invoked only the next time the event is fired, after which it is removed. ```js server.once( 'connection' , function( stream ) { console.log( 'Ah, we have our first user!' ) ; } ) ; ``` > Returns emitter, so calls can be chained. Note that using `.once()` in *NextGen Events* lib is just a syntactic sugar (and it's also there for compatibility), the previous example can be rewritten using `.on()`: ```js server.on( 'connection' , { fn: function( stream ) { console.log( 'Ah, we have our first user!' ) ; } , once: true } ) ; ``` <a name="ref.events.removeListener"></a> ### .removeListener( eventName , listenerID ) *or* .off( eventName , listenerID ) * eventName `string` the name of the event the listener to remove is binded to * listenerID `any type` the identifier of the listener to remove Node.js documentation: > Remove a listener from the listener array for the specified event. > **Caution**: changes array indices in the listener array behind the listener. ```js var callback = function( stream ) { console.log( 'someone connected!' ) ; } ; server.on( 'connection' , callback ) ; // ... server.removeListener( 'connection' , callback ) ; ``` **CAUTION: Unlike the built-in Node.js emitter**, `.removeListener()` will remove **ALL** listeners whose ID is matching the given *listenerID*. If any single listener has been added multiple times to the listener array for the specified event, then only one call to `.removeListener()` will remove them all. > Returns emitter, so calls can be chained. Example using user-defined ID: ```js var callback = function( stream ) { console.log( 'someone connected!' ) ; } ; server.on( 'connection' , { id: 'foo' , fn: callback } ) ; server.on( 'connection' , { id: 'bar' , fn: callback } ) ; // ... // Does nothing! we have used custom IDs! server.removeListener( 'connection' , callback ) ; // Remove the first listener only, despite the fact they are sharing the same function server.removeListener( 'connection' , 'foo' ) ; ``` Don't forget that by default, the ID is the callback function itself. <a name="ref.events.removeAllListeners"></a> ### .removeAllListeners( [eventName] ) * eventName `string` (optional) the name of the event the listeners to remove are binded to Node.js documentation: > Removes all listeners, or those of the specified event. > It's not a good idea to remove listeners that were added elsewhere in the code, especially when it's on an emitter > that you didn't create (e.g. sockets or file streams). > Returns emitter, so calls can be chained. <a name="ref.events.setMaxListeners"></a> ### .setMaxListeners() Only available for compatibility with the built-in Node.js emitter, so it does not break the code for people that want to make the switch. But please note that **there is no such concept of max listener in NextGen Events**, this method does nothing (it's an empty function). <a name="ref.events.listeners"></a> ### .listeners( eventName ) * eventName `string` (optional) the name of the event the listeners to list are binded to Node.js documentation: > Returns an array of listeners for the specified event. ```js server.on( 'connection' , function( stream ) { console.log( 'someone connected!' ) ; } ) ; console.log( util.inspect( server.listeners( 'connection' ) ) ) ; // output: // [ { id: [Function], fn: [Function], nice: -Infinity, event: 'connection' } ] ``` <a name="ref.events.listenerCount"></a> ### .listenerCount( eventName ) * eventName `string` the name of the event Node.js documentation: > Returns the number of listeners listening to the type of event. <a name="ref.events.setNice"></a> ### .setNice( nice ) * nice `integer` (default: -Infinity) see [the nice feature](#ref.note.nice) for details Set the default *nice value* of the current emitter. <a name="ref.events.emit"></a> ### .emit( [nice] , eventName , [arg1] , [arg2] , [...] , [callback] ) * nice `integer` (default: -Infinity) see [the nice feature](#ref.note.nice) for details * eventName `string` (optional) the name of the event to emit * arg1 `any type` (optional) first argument to transmit * arg2 `any type` (optional) second argument to transmit * ... * callback `function` (optional) a completion callback triggered when all listener have done, accepting arguments: * interruption `any type` if truthy, then emit was interrupted with this interrupt value (provided by userland) * event `Object` representing the current event It returns an object representing the current event. Node.js documentation: > Execute each of the listeners in order with the supplied arguments. **It does not returns the emitter!** <a name="ref.note.nice"></a> ### A note about the *nice feature* The *nice value* represent the *niceness* of the event-emitting processing. This concept is inspired by the UNIX *nice* concept for processus (see the man page for the *nice* and *renice* command). In this lib, this represents the asyncness of the event-emitting processing. The constant `require( 'nextgen-events' ).SYNC` can be used to have synchronous event emitting, its value is `-Infinity` and it's the default value. * any nice value *N* greater than or equals to 0 will be emitted asynchronously using setTimeout() with a *N* ms timeout to call the listeners * any nice value *N* lesser than 0 will emit event synchronously until *-N* recursion is reached, after that, setImmediate() will be used to call the listeners, the first event count as 1 recursion, so if nice=-1, all events will be asynchronously emitted, if nice=-2 the initial event will call the listener synchronously, but if the listener emits events on the same emitter object, the sub-listener will be called through setImmediate(), breaking the recursion... and so on... They are many elements that can define their own *nice value*. Here is how this is resolved: * First the *emit nice value* will be the one passed to the `.emit()` method if given, or the default *emitter nice value* defined with [.setNice()](#ref.events.setNice). * For each listener to be called, the real *nice value* for the current listener will be the **HIGHEST** *nice value* of the *emit nice value* (see above), the listener *nice value* (defined with [.addListener()](#ref.events.addListener)), and if the listener is tied to a context, the context *nice value* (defined with [.addListenerContext()](#ref.events.addListenerContext) or [.setListenerContextNice](#ref.events.setListenerContextNice)) <a name="ref.events.addListenerContext"></a> ### .addListenerContext( contextName , options ) * contextName `string` a non-empty string identifying the context to be created * options `Object` an object of options, where: * nice `integer` (default: -Infinity) see [the nice feature](#ref.note.nice) for details * serial `boolean` (default: false) if true, the async listeners tied to this context will run sequentially, one after the other is fully completed Create a context using the given *contextName*. Listeners can be tied to a context, enabling some grouping features like turning them on or off just by enabling/disabling the context, queuing them, resuming them, or forcing serialization of all async listeners. <a name="ref.events.disableListenerContext"></a> ### .disableListenerContext( contextName ) * contextName `string` a non-empty string identifying the context to be created It disables a context: any listeners tied to it will not be triggered anymore. The context is not destroyed, the listeners are not removed, they are just inactive. They can be enabled again using [.enableListenerContext()](#ref.events.enableListenerContext). <a name="ref.events.queueListenerContext"></a> ### .queueListenerContext( contextName ) * contextName `string` a non-empty string identifying the context to be created It switchs a context into *queue mode*: any listeners tied to it will not be triggered anymore, but every listener's call will be queued. When the context will be enabled again using [.enableListenerContext()](#ref.events.enableListenerContext), any queued listener's call will be processed. <a name="ref.events.enableListenerContext"></a> ### .enableListenerContext( contextName ) * contextName `string` a non-empty string identifying the context to be created This enables a context previously disabled using [.disableListenerContext()](#ref.events.disableListenerContext) or queued using [.disableListenerContext()](#ref.events.disableListenerContext). If the context was queued, any queued listener's call will be processed right now for synchronous emitter, or a bit later depending on the *nice value*. E.g. if a listener would have been called with a timeout of 50 ms (nice value = 5), and the call has been queued, the timeout will apply at resume time. <a name="ref.events.setListenerContextNice"></a> ### .setListenerContextNice( contextName , nice ) * contextName `string` a non-empty string identifying the context to be created * nice `integer` (default: -Infinity) see [the nice feature](#ref.note.nice) for details Set the *nice* value for the current context. <a name="ref.events.serializeListenerContext"></a> ### .serializeListenerContext( contextName , [value] ) * contextName `string` a non-empty string identifying the context to be created * value `boolean` (optional, default is true) if *true* the context will enable serialization for async listeners. This is one of the top feature of this lib. If set to *true* it enables the context serialization. It has no effect on listeners defined without the *async* option (see [.addListener()](#ref.events.addListener)). Listeners defined with the async option will postpone any other listener's calls part of the same context. Those calls will be queued until the completion callback of the listener is triggered. Example: ```js app.on( 'maintenance' , { context: 'maintenanceHandlers' , async: true , fn: function( type , done ) { performSomeCriticalAsyncStuff( function() { console.log( 'Critical maintenance stuff finished' ) ; done() ; } ) ; } } ) ; app.serializeListenerContext( maintenanceHandlers ) ; // ... app.emit( 'maintenance' , 'doBackup' ) ; // Despite the fact we emit synchronously, the listener will not be called now, // it will be queued and called later when the previous call will be finished app.emit( 'maintenance' , 'doUpgrade' ) ; ``` By the way, there is only one listener here that will queue itself, and only one event type is fired. But this would work the same with multiple listener and event type, if they share the same context. Same code with two listeners and two event type: ```js app.on( 'doBackup' , { context: 'maintenanceHandlers' , async: true , fn: function( done ) { performBackup( function() { console.log( 'Backup finished' ) ; done() ; } ) ; } } ) ; app.on( 'doUpgrade' , { context: 'maintenanceHandlers' , async: true , fn: function( done ) { performUpgrade( function() { console.log( 'Upgrade finished' ) ; done() ; } ) ; } } ) ; app.on( 'whatever' , function() { // Some actions... } ) ; app.serializeListenerContext( maintenanceHandlers ) ; // ... app.emit( 'doBackup' ) ; // Despite the fact we emit synchronously, the second listener will not be called now, // it will be queued and called later when the first listener will have finished its job app.emit( 'doUpgrade' ) ; // The third listener is not part of the 'maintenanceHandlers' context, so it will be called // right now, before the first listener finished, and before the second listener ever start app.emit( 'whatever' ) ; ``` <a name="ref.events.destroyListenerContext"></a> ### .destroyListenerContext( contextName ) * contextName `string` a non-empty string identifying the context to be created This destroy a context and remove all listeners tied to it. Any queued listener's calls will be lost. <a name="incompatibilities"></a> ## Incompatibilities with the built-in Node.js EventEmitter NextGen Events is almost compatible with Node.js' EventEmitter, except for few things: * .emit() does not return the emitter, but an object representing the current event. * If the last argument passed to .emit() is a function, it is not passed to listeners, instead it is a completion callback triggered when all listeners have done their job. If one want to pass function to listeners as the final argument, it is easy to add an extra `null` or `undefined` argument to .emit(). * There is more reserved event name: 'interrupt', 'emitted'. * There is no such concept of *max listener* in NextGen Events, .setMaxListeners() function exists only to not break compatibility for people that want to make the switch, but it does nothing (it's an empty function). * .removeListener() will remove all matching listener, not only the first listener found. * 'newListener'/'removeListener' event listener will receive an array of new/removed *listener object*, instead of only one *listener function*. E.g: it will be fired only once by when .removeListener() or .removeAllListener() is invoked and multiple listeners are deleted. A *listener object* contains a property called 'fn' that hold the actual *listener function*. * `.removeAllListeners()` without any argument does not trigger 'removeListener' listener, because there are actually removed too. The same apply to `.removeAllListeners( 'removeListener' )`. * .listeners() same here: rather than providing an array of *listener function* an array of *listener object* is provided. <a name="ref.proxy"></a> ## Proxy Services **This part of the doc is still a work in progress!** **Proxy services are awesome.** They abstract away the network so we can emit and listen to emitter on the other side of the plug! Both side of the channel create a Proxy, and add to it local and remote *services*, i.e. event emitters, and that's all. A remote service looks like a normal (i.e. local) emitter, and share the same API (with few limitations). It's totally protocol agnostic, you just define two methods for your proxy: one to read from the network and one to send to it (e.g. for Web Socket, this is a one-liner). #### Example, using the Web Socket *ws* node module The code below set up a server and a client written in Node.js. The server expose the *heartBeatService* which simply emit an *heartBeat* event once in a while with the beat count as data. Most of this code is websocket boiler-plate, the actual proxy code involves only few lines. The client code could be easily rewritten for the browser. **Server:** ```js var NGEvents = require( 'nextgen-events' ) ; // Create our service/emitter var heartBeatEmitter = new NGEvents() ; var nextBeat = 1 ; // Emit one 'heartBeat' event every few seconds setInterval( function() { var beat = nextBeat ++ ; heartBeatEmitter.emit( 'heartBeat' , beat ) ; } , 2000 ) ; // Create our server var WebSocket = require( 'ws' ) ; var server = new WebSocket.Server( { port: 12345 } ) ; // On new connection... server.on( 'connection' , function connection( ws ) { // Create a proxy for this client var proxy = new NGEvents.Proxy() ; // Add the local service exposed to this client and grant it all right proxy.addLocalService( 'heartBeatService' , heartBeatEmitter , { listen: true , emit: true , ack: true } ) ; // message received: just hook to proxy.receive() ws.on( 'message' , function incoming( message ) { proxy.receive( message ) ; } ) ; // Define the receive method: should call proxy.push() // after decoding the raw message proxy.receive = function receive( raw ) { try { proxy.push( JSON.parse( raw ) ) ; } catch ( error ) {} } ; // Define the send method proxy.send = function send( message ) { ws.send( JSON.stringify( message ) ) ; } ; // Clean up after everything is done ws.on( 'close' , function close() { proxy.destroy() ; } ) ; } ) ; ``` **Client:** ```js var NGEvents = require( 'nextgen-events' ) ; var WebSocket = require( 'ws' ) ; var ws = new WebSocket( 'ws://127.0.0.1:12345' ) ; // Create a proxy var proxy = new NGEvents.Proxy() ; // Once the connection is established... ws.on( 'open' , function open() { // Add the remote service we want to access proxy.addRemoteService( 'heartBeatService' ) ; // Listen to the event 'heartBeat' on this service proxy.remoteServices.heartBeatService.on( 'heartBeat' , function( beat ) { console.log( '>>> Heart Beat (%d) received!' , beat ) ; } ) ; } ) ; // message received: just hook to proxy.receive() ws.on( 'message' , function( message ) { proxy.receive( message ) ; } ) ; // Define the receive method: should call proxy.push() // after decoding the raw message proxy.receive = function receive( raw ) { try { proxy.push( JSON.parse( raw ) ) ; } catch ( error ) {} } ; // Define the send method proxy.send = function send( message ) { ws.send( JSON.stringify( message ) ) ; } ; // Clean up after everything is done ws.on( 'close' , function close() { proxy.destroy() ; } ) ; ``` Options passed to `.addLocalService()`: * listen `boolean` if set, the remote client can listen (addListener()/on()) to the local emitter * emit `boolean` if set, the remote client can emit on the local emitter * ack `boolean` if set, the remote client can acknowledge or ask for acknowledgement, enabling **async listeners** and .emit()'s **completion callback** NextGen Events features available in proxy services: * All the basic API is supported (the node-compatible API) * Emit completion callback supported * Async listeners supported Features that could be supported in the future: * Emit interruption and retrieving the interruption value Features that are unlikely to be supported: * Remote emit with a nice value (does not make sense at all through a network) * Contexts cannot be shared across different proxies/client, think of it as if they were namespaced behind their proxy
35.465206
134
0.728135
eng_Latn
0.991847
2194367ffd6c0843c6ad1f80095acee8bb0f9350
435
md
Markdown
CATS_spec/docs-content-e2e/CATS-Environment-Deploy.md
OlinaLiuGit/OlinaRepo
9fccd7e4a7df314247ecfe40fb16cf0f5c047630
[ "CC-BY-4.0", "MIT" ]
null
null
null
CATS_spec/docs-content-e2e/CATS-Environment-Deploy.md
OlinaLiuGit/OlinaRepo
9fccd7e4a7df314247ecfe40fb16cf0f5c047630
[ "CC-BY-4.0", "MIT" ]
4
2017-06-19T01:37:45.000Z
2017-10-20T10:09:10.000Z
CATS_spec/docs-content-e2e/CATS-Environment-Deploy.md
OlinaLiuGit/OlinaRepo
9fccd7e4a7df314247ecfe40fb16cf0f5c047630
[ "CC-BY-4.0", "MIT" ]
6
2017-04-28T08:23:23.000Z
2019-01-22T01:45:33.000Z
CATS Deploy Environment | |URL | |---|---| |Test| [http://csicesvr/](http://csicesvr/) | |Prod| [http://contentqa_cats/](http://contentqa_cats/) | CATS environment deploy: - [How to deploy a test machine](How-to-deploy-CATS-test-machine.md) - [How to deploy a debug machine](How-to-deploy-CATS-debug-machine.md) - [How to deploy CATS to website](How-to-deploy-CATS-website.md) - [How to deploy CATS DB ](How-to-deploy-CATS-DB.md)
36.25
70
0.691954
yue_Hant
0.880751
21943c9d663990913af88b5b770c3cab11d2731a
1,885
md
Markdown
docs/standard-library/duration-values-structure.md
littlenine/cpp-docs.zh-tw
811b8d2137808d9d0598fbebec6343fdfa88c79e
[ "CC-BY-4.0", "MIT" ]
null
null
null
docs/standard-library/duration-values-structure.md
littlenine/cpp-docs.zh-tw
811b8d2137808d9d0598fbebec6343fdfa88c79e
[ "CC-BY-4.0", "MIT" ]
null
null
null
docs/standard-library/duration-values-structure.md
littlenine/cpp-docs.zh-tw
811b8d2137808d9d0598fbebec6343fdfa88c79e
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: duration_values 結構 ms.date: 11/04/2016 f1_keywords: - chrono/std::chrono::duration_values - chrono/std::chrono::duration_values::max - chrono/std::chrono::duration_values::min - chrono/std::chrono::duration_values::zero ms.assetid: 7f66d2e3-1faf-47c3-b47e-08f2a87f20e8 ms.openlocfilehash: bc382bbc408b11cbc18210f3ab944dda39adc8f2 ms.sourcegitcommit: 0ab61bc3d2b6cfbd52a16c6ab2b97a8ea1864f12 ms.translationtype: MT ms.contentlocale: zh-TW ms.lasthandoff: 04/23/2019 ms.locfileid: "62413771" --- # <a name="durationvalues-structure"></a>duration_values 結構 提供 [duration](../standard-library/duration-class.md) 範本參數 `Rep` 的特定值。 ## <a name="syntax"></a>語法 ```cpp template <class Rep> struct duration_values; ``` ## <a name="members"></a>成員 ### <a name="public-methods"></a>公用方法 |名稱|描述| |----------|-----------------| |[max](#max)|靜態。 指定 `Rep` 類型的值上限。| |[min](#min)|靜態。 指定 `Rep` 類型的值下限。| |[zero](#zero)|靜態。 傳回 `Rep(0)`。| ## <a name="requirements"></a>需求 **標頭:** \<chrono > **命名空間:** std::chrono ## <a name="max"></a> duration_values::max 靜態方法會傳回型別 `Ref` 的上限值。 ```cpp static constexpr Rep max(); ``` ### <a name="return-value"></a>傳回值 實際上,系統會傳回 `numeric_limits<Rep>::max()`。 ### <a name="remarks"></a>備註 當 `Rep` 是使用者定義的類型時,傳回的值必須大於 [duration_values::zero](#zero)。 ## <a name="min"></a> duration_values::min 靜態方法會傳回型別 `Ref` 的下限值。 ```cpp static constexpr Rep min(); ``` ### <a name="return-value"></a>傳回值 實際上,系統會傳回 `numeric_limits<Rep>::lowest()`。 ### <a name="remarks"></a>備註 當 `Rep` 是使用者定義的類型時,傳回的值必須小於或等於 [duration_values::zero](#zero)。 ## <a name="zero"></a> duration_values::zero 傳回 `Rep(0)`。 ```cpp static constexpr Rep zero(); ``` ### <a name="remarks"></a>備註 當 `Rep` 是使用者定義型別時,傳回值必須代表加法類無限大。 ## <a name="see-also"></a>另請參閱 [標頭檔參考](../standard-library/cpp-standard-library-header-files.md)<br/> [\<chrono>](../standard-library/chrono.md)<br/>
20.48913
70
0.666313
yue_Hant
0.423777
21950f0241cc6f0cac41c4b0d0a6439011f6be80
231
md
Markdown
packages/Project15-Core.package/CMSSideBar.class/README.md
hpi-swa-teaching/SWA-Course-Management
c283be5b7f76913d73b3d6720d77d456ab8cae64
[ "MIT" ]
null
null
null
packages/Project15-Core.package/CMSSideBar.class/README.md
hpi-swa-teaching/SWA-Course-Management
c283be5b7f76913d73b3d6720d77d456ab8cae64
[ "MIT" ]
1
2015-05-20T12:17:12.000Z
2015-05-26T11:07:46.000Z
packages/Project15-Core.package/CMSSideBar.class/README.md
HPI-SWA-Teaching/SWT15-Project-15
c283be5b7f76913d73b3d6720d77d456ab8cae64
[ "MIT" ]
1
2016-12-02T17:14:05.000Z
2016-12-02T17:14:05.000Z
A CMSSideBar is a component to render the sidebar menu and its topics. Instance Variables rootComponent: <CMSRootElement> rootComponent - the reference to the root component to have the possibility to change the main content.
28.875
90
0.805195
eng_Latn
0.99587
2195868889b4b90ecaa5530660272999c5d3df9d
383
md
Markdown
zh/api/math/fuzzyGreaterThan.md
jcogs/qiciengine-documentation
e6c5c3c1d5eb4f7f120bdab8f7cc2c797a82905d
[ "MIT" ]
13
2016-01-14T08:27:40.000Z
2022-02-23T23:46:32.000Z
zh/api/math/fuzzyGreaterThan.md
oneseedfruit/qiciengine-documentation
73a5efc07099c18abb071a5564f2621ae937f2fa
[ "MIT" ]
6
2016-01-24T10:30:39.000Z
2017-02-22T09:49:01.000Z
zh/api/math/fuzzyGreaterThan.md
oneseedfruit/qiciengine-documentation
73a5efc07099c18abb071a5564f2621ae937f2fa
[ "MIT" ]
14
2016-01-24T06:17:36.000Z
2022-02-23T23:45:35.000Z
# [qc.Math](README.md).fuzzyGreaterThan ## 原型 * boolean fuzzyGreaterThan(a, b, epsilon) ## 参数说明 | 参数名 | 类型 | 说明 | | ------------- | ------------- | -------------| | a | number | 值 | | b | number | 值 | | epsilon | number | 精度值 | | 返回值 | boolean | 在指定精度内,a 大于 b,返回 true | ## 说明 浮点数大于比较 ## 范例 ```` // 返回 true var ret = game.math.fuzzyGreaterThan(1.232, 1.23, 0.00001); ````
17.409091
63
0.503916
yue_Hant
0.600875
2195b58743647996292eba6f444992e613a1ce06
35
md
Markdown
README.md
luismartim/Luis_teste
91f25343d5a8a81ada7a0761e46225d5a793755d
[ "MIT" ]
null
null
null
README.md
luismartim/Luis_teste
91f25343d5a8a81ada7a0761e46225d5a793755d
[ "MIT" ]
null
null
null
README.md
luismartim/Luis_teste
91f25343d5a8a81ada7a0761e46225d5a793755d
[ "MIT" ]
null
null
null
# Luis_teste Meu primeiro repo git
11.666667
21
0.8
por_Latn
0.999212
2195c4ac5f1be13434e76659a09084de9883a4cc
3,466
md
Markdown
translations/ko-KR/content/graphql/overview/about-the-graphql-api.md
Trungphunglamd/docs
a1910a7f8cedd4485962ad0f0c3c93d7348709a5
[ "CC-BY-4.0", "MIT" ]
20
2021-02-17T16:18:11.000Z
2022-03-16T08:30:36.000Z
translations/ko-KR/content/graphql/overview/about-the-graphql-api.md
0954011723/docs
d51685810027d8071e54237bbfd1a9fb7971941d
[ "CC-BY-4.0", "MIT" ]
58
2021-12-28T02:58:48.000Z
2022-03-18T22:47:54.000Z
translations/ko-KR/content/graphql/overview/about-the-graphql-api.md
0954011723/docs
d51685810027d8071e54237bbfd1a9fb7971941d
[ "CC-BY-4.0", "MIT" ]
46
2020-11-05T10:39:05.000Z
2021-07-23T11:35:59.000Z
--- title: About the GraphQL API intro: 'The {% data variables.product.prodname_dotcom %} GraphQL API offers flexibility and the ability to define precisely the data you want to fetch.' versions: free-pro-team: '*' enterprise-server: '*' --- ### 개요 Here are some quick links to get you up and running with the GraphQL API v4: * [Authentication](/v4/guides/forming-calls/#authenticating-with-graphql) * [Root endpoint](/v4/guides/forming-calls/#the-graphql-endpoint) * [Schema introspection](/v4/guides/intro-to-graphql/#discovering-the-graphql-api) * [Rate limits](/v4/guides/resource-limitations/) * [Migrating from REST](/v4/guides/migrating-from-rest) ### About GraphQL The [GraphQL](https://graphql.github.io/) data query language is: * **A [specification](https://graphql.github.io/graphql-spec/June2018/).** The spec determines the validity of the [schema](/v4/guides/intro-to-graphql#schema) on the API server. The schema determines the validity of client calls. * **[Strongly typed](#about-the-graphql-schema-reference).** The schema defines an API's type system and all object relationships. * **[Introspective](/v4/guides/intro-to-graphql#discovering-the-graphql-api).** A client can query the schema for details about the schema. * **[Hierarchical](/v4/guides/forming-calls).** The shape of a GraphQL call mirrors the shape of the JSON data it returns. [Nested fields](/v4/guides/migrating-from-rest/#example-nesting) let you query for and receive only the data you specify in a single round trip. * **An application layer.** GraphQL is not a storage model or a database query language. The _graph_ refers to graph structures defined in the schema, where [nodes](/v4/guides/intro-to-graphql#node) define objects and [edges](/v4/guides/intro-to-graphql#edge) define relationships between objects. The API traverses and returns application data based on the schema definitions, independent of how the data is stored. ### Why GitHub is using GraphQL GitHub chose GraphQL for our API v4 because it offers significantly more flexibility for our integrators. The ability to define precisely the data you want&mdash;and _only_ the data you want&mdash;is a powerful advantage over the REST API v3 endpoints. GraphQL lets you replace multiple REST requests with _a single call_ to fetch the data you specify. For more details about why GitHub has moved to GraphQL, see the original [announcement blog post](https://githubengineering.com/the-github-graphql-api/). ### About the GraphQL schema reference The docs in the sidebar are generated from the {% data variables.product.prodname_dotcom %} GraphQL [schema](/v4/guides/intro-to-graphql/#discovering-the-graphql-api). All calls are validated and executed against the schema. Use these docs to find out what data you can call: * Allowed operations: [queries](/v4/query) and [mutations](/v4/mutation). * Schema-defined types: [scalars](/v4/scalar), [objects](/v4/object), [enums](/v4/enum), [interfaces](/v4/interface), [unions](/v4/union), and [input objects](/v4/input_object). You can access this same content via the [Explorer Docs sidebar](/v4/guides/using-the-explorer#accessing-the-sidebar-docs). Note that you may need to rely on both the docs and the schema validation to successfully call the GraphQL API. For other information, such as authentication and rate limit details, check out the [guides](/v4/guides). ### Requesting support {% data reusables.support.help_resources %}
64.185185
416
0.766878
eng_Latn
0.954639
21965f2d862b7931fcb6e60efc740cf1d86eff28
566
md
Markdown
README.md
trikko/raylib-libvlc-example
ea7ae0ff1a915108c052ea6a493e704e8a97dd49
[ "BSD-3-Clause" ]
37
2021-11-14T09:07:13.000Z
2022-02-07T13:37:29.000Z
README.md
trikko/raylib-libvlc-example
ea7ae0ff1a915108c052ea6a493e704e8a97dd49
[ "BSD-3-Clause" ]
null
null
null
README.md
trikko/raylib-libvlc-example
ea7ae0ff1a915108c052ea6a493e704e8a97dd49
[ "BSD-3-Clause" ]
1
2022-01-06T10:26:09.000Z
2022-01-06T10:26:09.000Z
# Hey! I bet you have been trying to render and control a video with raylib for a long long time. Don't you think you should at least buy me a [beer](https://paypal.me/andreafontana/5)? ## What's this? ![](https://github.com/trikko/raylib-libvlc-example/blob/main/raylib-libvlc-example.gif?raw=true) ## How to build - Install [raylib](https://github.com/raysan5/raylib). Build instructions [here](https://github.com/raysan5/raylib#build-and-installation). - Install libvlc. On debian/ubuntu/etc.: ```sudo apt-get install libvlc-dev``` - Run ```./build.sh```
37.733333
140
0.720848
eng_Latn
0.625485
2196c1d87dd1263c6c6bea9b82703dfde5056004
3,370
md
Markdown
rc-cts-misp/README.md
nickpartner-goahead/resilient-community-apps
097c0dbefddbd221b31149d82af9809420498134
[ "MIT" ]
65
2017-12-04T13:58:32.000Z
2022-03-24T18:33:17.000Z
rc-cts-misp/README.md
nickpartner-goahead/resilient-community-apps
097c0dbefddbd221b31149d82af9809420498134
[ "MIT" ]
48
2018-03-02T19:17:14.000Z
2022-03-09T22:00:38.000Z
rc-cts-misp/README.md
nickpartner-goahead/resilient-community-apps
097c0dbefddbd221b31149d82af9809420498134
[ "MIT" ]
95
2018-01-11T16:23:39.000Z
2022-03-21T11:34:29.000Z
# MISP Threat Searcher ## Release Notes <!-- Specify all changes in this release. Do not remove the release notes of a previous release --> ### v1.1.1 * Fix backward compatibility issues. ## Overview This implements a custom threat service for the [MISP](http://www.misp-project.org/) Threat Intelligence Platform. ## Installation Install this package with 'pip', or `python setup.py install`. Update your `app.config` file with parameters for the MISP connection. To update the configuration file with default parameters: ``` resilient-circuits config -u ``` Note: For the `misp_url` parameter in the app.config file, the https protocol should be specified. To run the service: ``` resilient-circuits run ``` Register the custom threat service onto your Resilient server (assuming that the `resilient-circuits` application is running on the same server, and that the webserver is listening on port 9000): ``` sudo resutil threatserviceedit -name MISP -resturl http://127.0.0.1:9000/cts/misp ``` To test the connection: ``` sudo resutil threatservicetest -name MISP ``` ## Uninstall To delete: ``` sudo resutil threatservicedel -name MISP ``` ## Supported artifact types Map of supported Threat Service artifact types to Misp types: | Resilient artifact types | Misp types | | ------ | ----------- | | file.content | None | | file.name | ["filename", "email-attachment", "filename&#124;md5", "filename&#124;sha1", "filename&#124;sha256", "filename&#124;authentihash", "filename&#124;ssdeep", "filename&#124;imphash", "filename&#124;pehash", "filename&#124;sha224", "filename&#124;sha384", "filename&#124;sha512", "filename&#124;sha512/224", "filename&#124;sha512/256"]| | file.path | "other" | | email.body | "text" | | email.header | ["email-src", "email-dst", "target-email", "whois-registrant-email", "dns-soa-email", "email-reply-to", "email-subject", "email-header", "email-x-mailer", "email-message-id", "email-mime-boundary"] | | email.header.sender_address | ["email-src", "email-dst", "target-email", "whois-registrant-email", "dns-soa-email", "email-reply-to"] | | email.header.sender_name | None | | email.header.to | ["email-src", "email-dst", "target-email", "whois-registrant-email", "dns-soa-email", "email-reply-to"] | | hash.md5 | ["md5", "authentihash", "filename&#124;md5"] | | hash.sha1 | ["sha1", "x509-fingerprint-sha1", "filename&#124;sha1"] | | hash.fuzzy | ["other"] | | hash.sha256 | ["sha256", "sha512/256", "filename&#124;sha256"] | | cert.x509 | ["other"] | | net.cidr | ["ip-src", "ip-dst"] | | net.ip | ["ip-src", "ip-dst", "ip-dst&#124;port", "ip-src&#124;port", "domain&#124;ip"] | | net.name | ["hostname", "domain", "domain&#124;ip"] | | net.mac | ["other"] | | net.port | ["port"] | | net.uri | ["url", "uri", "link"] | | net.uri.path | ["uri"] | | net.http.request.header | ["http-method", "user-agent"] | | net.http.response.header | None | | process.name | ["other"] | | system.name | ["target-machine"] | | system.mutex | ["mutex"] | | system.registry | ["regkey"] | | system.service.name | ["windows-service-name", "windows-service-displayname"] | | system.user.name | ["text", "target-user", "github-username"] | | system.user.password | ["other"] | | threat.report.cve | ["vulnerability"] | | threat.malware.family | ["malware-type"] |
40.119048
347
0.662315
eng_Latn
0.54359
2196d072bfb1a456f4b0b39ad25f28c5d67ca251
7,948
md
Markdown
articles/operations-management-suite/oms-security-baseline.md
OpenLocalizationTestOrg/azure-docs-pr15_pl-PL
18fa7535e7cdf4b159e63a40776995fa95f1f314
[ "CC-BY-3.0", "CC-BY-4.0", "MIT" ]
null
null
null
articles/operations-management-suite/oms-security-baseline.md
OpenLocalizationTestOrg/azure-docs-pr15_pl-PL
18fa7535e7cdf4b159e63a40776995fa95f1f314
[ "CC-BY-3.0", "CC-BY-4.0", "MIT" ]
null
null
null
articles/operations-management-suite/oms-security-baseline.md
OpenLocalizationTestOrg/azure-docs-pr15_pl-PL
18fa7535e7cdf4b159e63a40776995fa95f1f314
[ "CC-BY-3.0", "CC-BY-4.0", "MIT" ]
null
null
null
<properties pageTitle="Operacje zarządzania pakietu zabezpieczeń i inspekcji rozwiązanie według planu bazowego | Microsoft Azure" description="Ten dokument wyjaśniono, jak używać usługi OMS zabezpieczeń i inspekcji rozwiązanie do wykonania ocena według planu bazowego, wszystkich komputerów monitorowane w celu zabezpieczenia i zgodność." services="operations-management-suite" documentationCenter="na" authors="YuriDio" manager="swadhwa" editor=""/> <tags ms.service="operations-management-suite" ms.devlang="na" ms.topic="hero-article" ms.tgt_pltfrm="na" ms.workload="na" ms.date="09/08/2016" ms.author="yurid"/> # <a name="baseline-assessment-in-operations-management-suite-security-and-audit-solution"></a>Ocena według planu bazowego w operacji zarządzania pakietu zabezpieczeń i rozwiązanie inspekcji Ten dokument ułatwia uzyskiwanie dostępu do bezpieczny stan monitorowania zasobów za pomocą [operacji zarządzania pakietu usługi (OMS) zabezpieczeń i rozwiązanie inspekcji](operations-management-suite-overview.md) możliwości oceny według planu bazowego. ## <a name="what-is-baseline-assessment"></a>Co to jest plan bazowy oceny? Microsoft, razem z organizacji branżowe i dla instytucji rządowych na całym świecie, określa konfiguracji systemu Windows, która reprezentuje wdrożeń wysoce bezpieczny serwer. Ta konfiguracja jest zestaw kluczach rejestru, ustawienia zasad inspekcji i ustawienia zasad zabezpieczeń oraz zalecane wartości tych ustawień firmy Microsoft. Ten zestaw reguł nosi nazwę zabezpieczeń według planu bazowego. Możliwości oceny według planu bazowego zabezpieczeń usługi OMS i inspekcji można bezproblemowo przeglądania swoich komputerach w celu zapewnienia zgodności. Istnieją trzy typy reguł: - **Reguły rejestru**: Sprawdź, czy klucze rejestru są prawidłowo. - **Inspekcji reguł**: reguły dotyczące zasad inspekcji. - **Zabezpieczenia reguł**: reguły dotyczące uprawnień użytkownika na komputerze. > [AZURE.NOTE] Przeczytaj [Użyj zabezpieczeń usługi OMS do oceny konfiguracji zabezpieczeń według planu bazowego](https://blogs.technet.microsoft.com/msoms/2016/08/12/use-oms-security-to-assess-the-security-configuration-baseline/) dla krótkie omówienie tej funkcji. ## <a name="security-baseline-assessment"></a>Ocena według planu bazowego zabezpieczeń Zapoznanie się z bieżącej oceny według planu bazowego zabezpieczeń na wszystkich komputerach, które są monitorowane przez zabezpieczenia usługi OMS i inspekcji za pomocą pulpitu nawigacyjnego. Wykonaj poniższe czynności, aby uzyskać dostęp do pulpitu nawigacyjnego ocena według planu bazowego zabezpieczeń: 1. Ten **Pakiet zarządzania operacje Microsoft** głównym pulpit nawigacyjny kliknij Kafelek **zabezpieczeń i inspekcji** . 2. Pulpit nawigacyjny **zabezpieczeń i inspekcji** kliknij **Ocena według planu bazowego** w obszarze **Domen zabezpieczeń**. Pulpit nawigacyjny **Ocena według planu bazowego zabezpieczeń** pojawia się, jak pokazano na poniższej ilustracji: ![Zabezpieczenia usługi OMS i inspekcji ocena według planu bazowego](./media/oms-security-baseline/oms-security-baseline-fig1.png) Ten pulpit nawigacyjny dzieli się w trzech głównych obszarach: - **Komputery w porównaniu z planu bazowego**: w tej sekcji przedstawiono podsumowanie liczba komputerów, które są dostępne i procent komputerów, które przekazywane oceny. Górny 10 komputerów i wyniku procent również oferuje oceny. - **Wymagany stan reguł**: Ta sekcja ma przeznaczenie, aby wyświetlić Sygnalizacja dostępności reguł nie powiodło się ważności i nie powiodło się reguły według typu. Wyświetlając do pierwszego wykresu można szybko ustalić w przypadku większości reguł nie powiodło się krytyczne, czy nie. Zawiera listę górny 10 reguły nie powiodło się i ich ważności. Drugi wykres zawiera typ reguły, która nie powiodła się podczas oceny. - **Komputery brakuje ocena według planu bazowego**: w tej sekcji listy komputerów, które nie były dostępne z powodu niezgodności systemu operacyjnego lub awarii. ### <a name="accessing-computers-compared-to-baseline"></a>Dostęp do komputerów w porównaniu z planu bazowego Najlepiej komputery są wszystkie spełniać zabezpieczeń oceny według planu bazowego. Jednak przewiduje się, że w niektórych okolicznościach to nie się dzieje. W ramach procesu zarządzania zabezpieczeń należy uwzględnić przeglądania komputerów, na których nie można przekazać wszystkie testy oceny zabezpieczeń. Szybki sposób wizualizacji, wybierając opcję **komputerów dostęp do** znajdującego się w sekcji **komputerów w porównaniu z planu bazowego** . Powinien zostać wyświetlony wynik wyszukiwania dziennika z listą komputerów, jak pokazano na poniższym obrazie: ![Dostęp do komputera wyników](./media/oms-security-baseline/oms-security-baseline-fig2.png) Wyniki wyszukiwania są wyświetlane w formacie tabeli i miejsce, w którym pierwsza kolumna zawiera nazwę komputera i drugi kolor zawiera liczbę reguł, które nie powiodło się. Aby pobrać informacje dotyczące typu reguły, która nie powiodło się, kliknij w polu Liczba reguł nie powiodło się obok nazwy komputera. Powinien zostać wyświetlony wynik podobny do tego pokazano na poniższej ilustracji: ![Szczegóły wyników komputera uzyskać do nich dostęp](./media/oms-security-baseline/oms-security-baseline-fig3.png) W tym wynik wyszukiwania znajduje się sumę używanych reguł, liczba krytyczne reguły, które nie powiodło się, reguły ostrzeżeń i reguł dotyczących informacji nie powiodło się. ### <a name="accessing-required-rules-status"></a>Uzyskiwanie dostępu do stanu wymagane reguł Po uzyskaniu informacji dotyczących wartości procentowej liczba komputerów, które przekazywane oceny, można uzyskać więcej informacji na temat reguł występują według poziomów ich istotności. W tej wizualizacji ułatwia Wyznaczanie priorytetów komputery, które powinny być najpierw kierowane do upewnij się, że będą one zgodne z następnego oceny. Umieść wskaźnik myszy na krytycznych części wykresu znajduje się na kafelku **nie powiodło się reguły ważności** w obszarze **wymagane stanu reguły** , a następnie kliknij go. Powinien zostać wyświetlony następujący ekran podobny do wyniku: ![Nie powiodła się reguły szczegóły ważności](./media/oms-security-baseline/oms-security-baseline-fig4.png) W wyniku tego dziennika został wyświetlony odpowiedni typ reguły według planu bazowego, która nie powiodło się, opis tej reguły, a identyfikator typowych konfiguracji wyliczenia (CCE) tej reguły zabezpieczeń. Następujące atrybuty powinny być wystarczające dla wykonywanie działań naprawczych, aby rozwiązać ten problem na komputerze docelowym. > [AZURE.NOTE] Aby uzyskać więcej informacji na temat CCE dostęp do [Krajowych luka bazy danych](https://nvd.nist.gov/cce/index.cfm). ### <a name="accessing-computers-missing-baseline-assessment"></a>Uzyskiwanie dostępu do komputerów brakuje ocena według planu bazowego Usługi OMS obsługuje według planu bazowego opinii domeny w systemie Windows Server 2008 R2 do systemu Windows Server 2012 R2. Windows Server 2016 według planu bazowego nie jest jeszcze ostateczne i zostaną dodane zaraz po opublikowaniu. Wszystkie inne systemy operacyjne zeskanowany za pośrednictwem usługi OMS zabezpieczeń i inspekcji ocena według planu bazowego jest wyświetlana w sekcji **komputerów Brak ocena według planu bazowego** . ## <a name="see-also"></a>Zobacz też W tym dokumencie wiesz dotyczące zabezpieczeń usługi OMS i inspekcji ocena według planu bazowego. Aby dowiedzieć się więcej na temat zabezpieczeń usługi OMS, zobacz następujące artykuły: - [Przegląd operacji usługi zarządzania pakietu (OMS)](operations-management-suite-overview.md) - [Monitorowanie i reagowanie na alerty zabezpieczeń w operacji zarządzania pakietu zabezpieczeń i rozwiązanie inspekcji](oms-security-responding-alerts.md) - [Monitorowanie zasobów operacje zarządzania pakietu zabezpieczeń i rozwiązanie inspekcji](oms-security-monitoring-resources.md)
94.619048
585
0.815677
pol_Latn
0.99999
21972241695f76a6443741aeb31f0d7cc3d456c9
612
md
Markdown
README.md
richorama/sat-beam
138f4253f71ecbf609775900e7576d478fb30a39
[ "MIT" ]
1
2021-05-28T00:30:45.000Z
2021-05-28T00:30:45.000Z
README.md
richorama/sat-beam
138f4253f71ecbf609775900e7576d478fb30a39
[ "MIT" ]
6
2019-12-11T02:51:41.000Z
2022-02-27T08:32:49.000Z
README.md
richorama/sat-beam
138f4253f71ecbf609775900e7576d478fb30a39
[ "MIT" ]
null
null
null
# sat-beam :satellite: calcuates the beam a satellite projects on the earth's surface ![](screenshot.png) Try out a demo [here](https://richorama.github.io/sat-beam). ## Installation Install via npm. ``` $ npm install sat-beam ``` ## Usage ```js import { getBeam } from 'sat-beam' const lng = 2.0 // degrees const lat = 52.0 // degrees const radius = 100.0 //km const points = getBeam(lng, lat, radius) /* [ { lat: 53.975960615468395, lng: 1.999999999997981 }, { lat: 53.959810572100736, lng: 2.112830670635857 }, { lat: 53.91167290927757, lng: 2.2238947168825023 }, ... */ ``` ## License MIT
15.692308
74
0.665033
eng_Latn
0.338002
21999772052ea22a0f261134225bbda7141fe92b
5,397
md
Markdown
user/pages/30.blog/2018-12-08_english-perspective/item.en.md
belamu/kanthaus.online
de84010a77e60156cbefb8e014ac6290540ded69
[ "CC0-1.0", "MIT" ]
6
2018-09-03T15:48:19.000Z
2021-09-27T12:04:04.000Z
user/pages/30.blog/2018-12-08_english-perspective/item.en.md
belamu/kanthaus.online
de84010a77e60156cbefb8e014ac6290540ded69
[ "CC0-1.0", "MIT" ]
13
2017-12-25T20:44:37.000Z
2020-10-30T09:37:10.000Z
user/pages/30.blog/2018-12-08_english-perspective/item.en.md
belamu/kanthaus.online
de84010a77e60156cbefb8e014ac6290540ded69
[ "CC0-1.0", "MIT" ]
14
2018-01-05T19:54:40.000Z
2021-03-24T10:16:31.000Z
--- title: An English perspective on Kanthaus date: "2018-12-08" published: true taxonomy: tag: [blog] author: James --- _“There was lots of cheese, yogurt and sausage, but the packages were sliced open, and we can’t take it. It’s not the usual thing, but this is what supermarkets do sometimes when they realise that people are bin-diving.”_ Hidden in the nook of a cobbled street in Wurzen, and just a few stops on the S-Bahn outside of Leipzig, is Kanthaus. At the entrance, a billowing colourful banner welcomes weary travellers and fellow foodsavers. === Kanthaus is the [new European headquarters for foodsharing](/blog/2018-09-26_foodsharing). A dedicated space for the development of the European food sharing movement, the community here is committed to developing the tools and structures to help make that happen. Kanthausians live in a dedicated sharing collective. When you walk through the door, a member of the community helps you navigate the two buildings with its numerous rooms dedicated to computer work, meetings, crafts, food storage and preparation, recreation, and dorms. ![](kanthaus-front.jpg) Just over a year old, the space is home to a range of projects including the new online foodsaving platform, [Karrot](https://karrot.world). On any given day there are more than ten people living in the house, volunteering, or passing through - as I was. At its busiest, there are more than twenty-six residents living communally, dividing the various chores and - of course - food liberation missions. It becomes increasingly clear as one goes along, that everything in this space, even the toothbrushes, is shared. Kanthaus focuses on constantly iterating over both sustainable projects and the space itself. It was collectively purchased last year by members of [yunity](https://yunity.org), a sharing network made up of nomadic individuals, of which Kanthaus and Karrot are sub-projects. In 2017 the two neighbouring houses K20 and K22 Kantstraße were empty shells. Fourteen months and an enormous amount of work later, the team and dozens of volunteers have transformed the space. They’ve fully re-wired both buildings, [experimented with a washing machine water reclamation system](/blog/2018-09-05_experiments-in-water-recycling), re-plumbed one of the buildings, set up water heating via solar array and are currently working on the chimneys to provide better airflow. You get the impression people here spend as much time working with their hands as they do on various coding projects. ![](kanthaus-hammock.jpg) It doesn’t take long to feel at home here, even as a visitor. I’ve gone out of my way over the last six months to spend time with a few different communities, to understand different ways of living, uses of time, reasons why people come together and means of governance. Like nearby Leipzig, Kanthaus is full of energy. It takes a relatively minimalistic approach to managing the community despite the level of activity going on, with weekly coordination and occasional evaluation meetings being the only required group gatherings. A typical meeting at Kanthaus is more akin to a group of friends having a discussion than an uncomfortable rigid silence as is quite common in some communities, and whilst there is some formality the intention is more to stay on topic and prevent five-minute tangents into whether or not to bin-dive for rare printer paper or just to go and buy it. Sharing resources is central to the community at Kanthaus and the most obvious example of this is in the sharing of food, whether bin-dived, saved from stores or grown in the garden. Every once a week, a group of Kanthausians will venture to a nearby supermarket bin to restock, this might mean taking the cargo-bike or S-Bahn to go some kilometers to the good spots. Once collected, the food is washed in a purpose-made food prep room consisting of an old bathtub raised on a breeze block, and then distributed throughout the house in one of the six storage rooms and areas. Members and volunteers of the house eat lunch and dinner communally every day. With more than enough food to go around, the house opens its doors for [Fairteiler](/events/fairteiler) (food sharing) once a week for the community to take their pick, although this is currently more a side-project whilst they try to work out how to make such community outreach efforts more sustainable and less time-intensive, given that their energies are focused on facilitating this at scale through tools like Karrot. Karrot, currently being developed by Kanthausianer Tilmann Becker, is an [open source software](https://github.com/yunity/karrot-frontend/) focused on making it easier for food-saving communities to organise their pickups and distribution of food. They’re hoping this tool amongst others and the ongoing development of this community will begin to make a dent in the shockingly high 88 million tonnes of food wasted across the EU each year - that’s over 137kg per person here in Germany alone. Making that change cannot work online alone though, and traveling to foodsharing communities and Karrot user groups outside of Germany is also an integral part of building this transnational movement of grassroots foodsaving activists. In doing this, they come closer to their nomadic roots in yunity again – the big difference being that now they have Kanthaus to come home to.
163.545455
1,079
0.801556
eng_Latn
0.999906
2199a1c1072805cd35a42d9470c950a08559dacd
2,093
md
Markdown
README.md
junzengx14/Car-Racing
a29af835747bd91eac2bfa7f0d2a1312642f8381
[ "MIT" ]
15
2021-01-29T12:05:22.000Z
2021-12-20T03:13:58.000Z
README.md
junzengx14/Car-Racing
a29af835747bd91eac2bfa7f0d2a1312642f8381
[ "MIT" ]
5
2021-02-05T22:44:18.000Z
2021-12-27T10:09:04.000Z
README.md
junzengx14/Car-Racing
a29af835747bd91eac2bfa7f0d2a1312642f8381
[ "MIT" ]
6
2021-02-05T12:13:05.000Z
2021-12-07T01:53:03.000Z
**Status:** This repository is still under development, expecting new features/papers and a complete tutorial to explain it. Feel free to raise questions/suggestions through GitHub Issues, if you want to use the current version of this repository. # Car-Racing This repo serves as a toolkit for testing different algorithm for car racing in a user-designed track. #### References The code in this repository is based on the following: * J. Zeng, B. Zhang and K. Sreenath. "Safety-Critical Model Predictive Control with Discrete-Time Control Barrier Function." 2021 American Control Conference (ACC). [PDF](https://arxiv.org/pdf/2007.11718.pdf) #### Instructions ##### Environments * Create your environment via `conda env create -f environment.yml`. The default conda environment name is `car-racing`, and you could also choose that name with your own preferences by editing the .yml file. * Currently the repository is not created as standard package through setuptools, use `export PYTHONPATH=./src` make sure you are able to run the examples. ##### Examples ###### System identifications ``` python examples/identification.py ``` This allows to identify the linearized dynamics of the racing car by regression. ###### Tracking ``` python examples/tracking.py --ctrl-policy mpc-lti --simulation --plotting --animation ``` This allows to test algorithm for tracking. The choices of `--ctrl-policy` could be `mpc-lti` and `pid`. `--simulation` is no longer required once you have generated .obj file for simulation data. `--plotting`, `animation` are currently purely optional. ###### Racing ``` python examples/racing.py --simulation --plotting --animation ``` This allows to test algorithm for racing. `--simulation` is no longer required once you have generated .obj file for simulation data. `--plotting`, `animation` are currently purely optional. #### Acknowledgements I would like to say great thanks to [Ugo Rosolia](https://github.com/urosolia), as some helper functions in this repository come from his open source repo [RacingLMPC](https://github.com/urosolia/RacingLMPC).
59.8
253
0.76302
eng_Latn
0.993027
2199c5897d37e9c4969d7c506dade65c80d49f5c
262
md
Markdown
_posts/2018-4-20-Hello-World.md
j-winters/j-winters.github.io
2a2b8a3d327a3415113bfcb9b9dd533091cc7ac2
[ "MIT" ]
null
null
null
_posts/2018-4-20-Hello-World.md
j-winters/j-winters.github.io
2a2b8a3d327a3415113bfcb9b9dd533091cc7ac2
[ "MIT" ]
null
null
null
_posts/2018-4-20-Hello-World.md
j-winters/j-winters.github.io
2a2b8a3d327a3415113bfcb9b9dd533091cc7ac2
[ "MIT" ]
null
null
null
--- layout: post comments: true title: Hello world! --- This is just your standard *Hello world!* blog post. I plan to cross-post on here and a [Replicated Typo](http://www.replicatedtypo.com), but this blog will serve as the main repository for my thoughts.
29.111111
203
0.736641
eng_Latn
0.997601
219a59633ff9d137e308f2144ec6eb190ffc942e
2,267
md
Markdown
src/lt/2018-04/13/07.md
PrJared/sabbath-school-lessons
94a27f5bcba987a11a698e5e0d4279b81a68bc9a
[ "MIT" ]
68
2016-10-30T23:17:56.000Z
2022-03-27T11:58:16.000Z
src/lt/2018-04/13/07.md
PrJared/sabbath-school-lessons
94a27f5bcba987a11a698e5e0d4279b81a68bc9a
[ "MIT" ]
367
2016-10-21T03:50:22.000Z
2022-03-28T23:35:25.000Z
src/lt/2018-04/13/07.md
OsArts/Bible-study
cfcefde42e21795e217d192a8b7a703ebb7a6c01
[ "MIT" ]
109
2016-08-02T14:32:13.000Z
2022-03-31T10:18:41.000Z
--- title: Tolesniam tyrinėjimui date: 28/12/2018 --- E. White, „Counsels for the Church“, 355–359 p.; „Resurrection“, 1082–1084 p.; „The Ellen G. White Encyclopedia“, 863-864 p. > <p></p> > „Mūsų Viešpaties prisikėlimas ir užžengimas yra tvirtas Dievo šventųjų pergalės prieš mirtį ir kapą, ir pažado, kad dangus yra atviras tiems, kurie išplauna savo drabužius ir išbalina juos Avinėlio krauju, įrodymas. Jėzus pakilo pas Tėvą kaip žmonijos atstovas, ir Dievas prikels tuos, kurie atspindi Jo paveikslą, kad matytų ir dalytųsi su Juo Jo šlove. Žemės piligrimams yra namai. Teisiesiems yra apsiaustai su garbės vainikais ir pergalės palmėmis. Viskas, kas mus glumino dėl Dievo apvaizdos, ateityje išaiškės. Tie dalykai, kuriuos sunku suvokti, bus paaiškinti. Maldos paslaptys atsiskleis. Kas mūsų ribotiems protams atrodė kaip painiava ir sulaužyti pažadai, atsiskleis kaip tobuliausia ir graži darna. Mes suprasime, kad begalinė Meilė įsakė patyrimus, kurie atrodė kaip sunkiausi išmėginimai. Mums suvokus švelnų rūpestį To, kurio dėka mums visa išeina į gera, mes neapsakomai džiaugsimės ir šlovinsime“ (E. White, Counsels for the Church, 358 p.). **Klausimai aptarimui:** `1. Nors kiti krikščionys (bet ne visi) tiki Jėzaus antruoju atėjimu, kuo unikali adventistų viltis, susijusi su Kristaus antruoju atėjimu?` `2. Dvi žuvys plaukiojo ir viena paklausė kitos: „Kaip vanduo?“ Kita žuvis atsakė: „Kas yra vanduo?“ Mintis ta, kad mes galime taip priprasti prie tam tikrų dalykų, jog nebejaučiame, kokie jie paplitę. Pavyzdžiui, kaip mes, gimę nuodėmėje, kupini nuodėmės ir gyvenantys nuodėmingame pasaulyje, išties galime suvokti, kokia nuostabi bus nauja būtis naujame danguje ir naujoje žemėje? Tačiau, kad ir kokie bebūtų apribojimai, mes vis tiek turėtume stengtis įsivaizduoti, kokia ji bus?` `3. Neabejotina, kad ir kokia bus mūsų būtis naujojoje žemėje, mes gyvensime vieningai su visais. Ką galime daryti dabar, kad pasiruoštume tam, kada tai įvyks?` **Santrauka:** Raštas užtikrintai kalba apie metą, kai ši žemė bus atkurta ir nuodėmė bus sunaikinta amžinai. Pagaliau žmonija bus atkurta pirminiam tikslui ir visi žmonės gyvens darniai. Mūsų dabartinė dvasinė vienybė Kristuje, nors dabar ir ne visiškai suprantama, bus gyvoji ir amžinoji tikrovė.
119.315789
961
0.795765
lit_Latn
1.000009
219a782500cc2c8081a97d0c40d83f3c0c101b47
3,371
md
Markdown
README.md
st-tech/gatling-operator
76a72ec557aad21aa255824b8eba62928f9f9581
[ "MIT" ]
17
2022-02-03T11:42:41.000Z
2022-03-30T14:49:03.000Z
README.md
st-tech/gatling-operator
76a72ec557aad21aa255824b8eba62928f9f9581
[ "MIT" ]
13
2022-01-26T05:49:12.000Z
2022-03-30T10:21:03.000Z
README.md
st-tech/gatling-operator
76a72ec557aad21aa255824b8eba62928f9f9581
[ "MIT" ]
1
2022-02-03T08:23:02.000Z
2022-02-03T08:23:02.000Z
# Gatling Operator [![Go Report Card](https://goreportcard.com/badge/github.com/st-tech/gatling-operator)](https://goreportcard.com/report/github.com/st-tech/gatling-operator) [![CI](https://github.com/st-tech/gatling-operator/actions/workflows/ci.yml/badge.svg?branch=main&event=push)](https://github.com/st-tech/gatling-operator/actions/workflows/ci.yml) ![GitHub release (latest SemVer)](https://img.shields.io/github/v/release/st-tech/gatling-operator) [Gatling](https://gatling.io/) is an open source load testing tool that allows to analyze and measure the performance of a variety of services. [Gatling Operator](https://github.com/st-tech/gatling-operator) is a Kubernetes Operator for running automated distributed Gatling load testing. ## How Gatling Operator works The desired state of a distributed Gatling load testing is described through a Kubernetes custom resource (`Gatling CR` in figure below). Based on Gatling CR, actions in the lifecycle of distributed Gatling load testing such as running load testing, generating reports, sending notification message, and cleaning up the resources are executed by relevant custom controller (`Gatling Controller` in figure below). ![](assets/gatling-operator-arch.svg) ## Features - Allows Gatling load testing senarios, resources, Gatling configurations files to be added in 2 ways: - Bundle them with Gatling runtime pakcages in a Gatling container - Add them as multi-line definition in Gatling CR - Scaling Gatling load testing - Horizontal scaling: number of pods running in parallel during a load testing can be configured - Vertical scaling: CPU and RAM resource allocation for Gatling runner Pod can be configured - Allows Gatling load testing to start running at a specific time - By default, the Gatling load testing starts running as soon as the runner Pod's init container gets ready. By specifing the start time, the Gatling load testing waits to start running until the specified time - Gatling Pod attributions - Gatling runtime container image - [rclone](https://rclone.org/) conatiner image - CPU and RAM resource allocation request and limit - `Affinity` (such as Node affinity) and `Tolerations` to be used by the scheduler to decide where a pod can be placed in the cluster - `Service accounts` for Pods - Reports - Automated generating aggregated Gatling HTML reports and storing them to Cloud Storages such as AWS S3, Google Cloud Storage, etc. via [rclone](https://rclone.org/) - Allows credentails info for accessing the remote storage to be specified via Secret resource - Notification - Automated posting webhook message and sending Gatling load testing result via notification providers such as Slack - Allows webhook URL info to be specified via Secret resource - Automated cleaning up Gatling resouces ## Requirements - Kubernetes: verson >= 1.18 > note: the versions below 1.18 might work but are not tested ## Quick Start - [Quick Start Guide](docs/quickstart-guide.md) ## Documentations - [Architecture](docs/architecture.md) - [Gatling CRD Reference](docs/api.md) - [User Guide](docs/user-guide.md) ## Contributing Please make a GitHub issue or pull request to help us build the operator. ## Changelog Please see the [list of releases](https://github.com/st-tech/gatling-operator/releases) for information on changes between releases.
57.135593
437
0.781667
eng_Latn
0.954251
219b3ba15ac54a58e7e6dceb8123b4a5ada7d3ab
3,929
md
Markdown
treebanks/cop_scriptorium/cop_scriptorium-dep-xcomp.md
akoehn/docs
ab54f654db7ab7c3433b5a6aa3dae19c3a938c7b
[ "Apache-2.0" ]
null
null
null
treebanks/cop_scriptorium/cop_scriptorium-dep-xcomp.md
akoehn/docs
ab54f654db7ab7c3433b5a6aa3dae19c3a938c7b
[ "Apache-2.0" ]
null
null
null
treebanks/cop_scriptorium/cop_scriptorium-dep-xcomp.md
akoehn/docs
ab54f654db7ab7c3433b5a6aa3dae19c3a938c7b
[ "Apache-2.0" ]
null
null
null
--- layout: base title: 'Statistics of xcomp in UD_Coptic-Scriptorium' udver: '2' --- ## Treebank Statistics: UD_Coptic-Scriptorium: Relations: `xcomp` This relation is universal. 131 nodes (1%) are attached to their parents as `xcomp`. 131 instances of `xcomp` (100%) are left-to-right (parent precedes child). Average distance between parent and child is 3.51145038167939. The following 9 pairs of parts of speech are connected with `xcomp`: <tt><a href="cop_scriptorium-pos-VERB.html">VERB</a></tt>-<tt><a href="cop_scriptorium-pos-VERB.html">VERB</a></tt> (76; 58% instances), <tt><a href="cop_scriptorium-pos-AUX.html">AUX</a></tt>-<tt><a href="cop_scriptorium-pos-VERB.html">VERB</a></tt> (26; 20% instances), <tt><a href="cop_scriptorium-pos-VERB.html">VERB</a></tt>-<tt><a href="cop_scriptorium-pos-AUX.html">AUX</a></tt> (11; 8% instances), <tt><a href="cop_scriptorium-pos-NOUN.html">NOUN</a></tt>-<tt><a href="cop_scriptorium-pos-VERB.html">VERB</a></tt> (8; 6% instances), <tt><a href="cop_scriptorium-pos-VERB.html">VERB</a></tt>-<tt><a href="cop_scriptorium-pos-NOUN.html">NOUN</a></tt> (4; 3% instances), <tt><a href="cop_scriptorium-pos-NOUN.html">NOUN</a></tt>-<tt><a href="cop_scriptorium-pos-AUX.html">AUX</a></tt> (3; 2% instances), <tt><a href="cop_scriptorium-pos-PRON.html">PRON</a></tt>-<tt><a href="cop_scriptorium-pos-AUX.html">AUX</a></tt> (1; 1% instances), <tt><a href="cop_scriptorium-pos-VERB.html">VERB</a></tt>-<tt><a href="cop_scriptorium-pos-DET.html">DET</a></tt> (1; 1% instances), <tt><a href="cop_scriptorium-pos-VERB.html">VERB</a></tt>-<tt><a href="cop_scriptorium-pos-PROPN.html">PROPN</a></tt> (1; 1% instances). ~~~ conllu # visual-style 9 bgColor:blue # visual-style 9 fgColor:white # visual-style 5 bgColor:blue # visual-style 5 fgColor:white # visual-style 5 9 xcomp color:blue 1 ⲁⲩⲱ ⲁⲩⲱ CCONJ CONJ _ 5 cc _ _ 2 ⲛ ⲛⲧⲉ AUX ACONJ _ 5 aux _ _ 3 ϥ ⲛⲧⲟϥ PRON PPERS Definite=Def|Gender=Masc|Number=Sing|Person=3|PronType=Prs 5 nsubj _ _ 4 ⲧⲙ ⲧⲙ ADV NEG Polarity=Neg 5 advmod _ _ 5 ⲕⲁ ⲕⲱ VERB V VerbForm=Fin 0 root _ _ 6 ⲛⲉϥ ⲡⲉϥ DET PPOS Definite=Def|Gender=Masc|Number=Sing|Person=3|Poss=Yes|PronType=Prs 7 det _ _ 7 ⲧⲟⲡⲟⲥ ⲧⲟⲡⲟⲥ NOUN N _ 5 obj _ _ 8 ⲉ ⲉ ADP PREP _ 9 mark _ _ 9 ϣⲱⲱⲧ ϣⲱⲱⲧ VERB V VerbForm=Inf 5 xcomp _ _ 10 ⲛ ⲛ ADP PREP _ 11 case _ _ 11 ⲗⲁⲁⲩ ⲗⲁⲁⲩ NOUN N _ 9 obj _ _ 12 : : PUNCT PUNCT _ 5 punct _ _ ~~~ ~~~ conllu # visual-style 7 bgColor:blue # visual-style 7 fgColor:white # visual-style 2 bgColor:blue # visual-style 2 fgColor:white # visual-style 2 7 xcomp color:blue 1 ⲉ ⲉ ADP PREP _ 2 mark _ _ 2 ⲧⲣⲉ ⲧⲣⲉ AUX ACAUS _ 0 root _ _ 3 ⲧ ⲡ DET ART Definite=Def|Gender=Fem|Number=Sing|PronType=Art 4 det _ _ 4 ⲙⲛⲧⲣⲱⲙⲉ ⲙⲛⲧⲣⲱⲙⲉ NOUN N _ 7 nsubj _ _ 5 ⲧⲏⲣ ⲧⲏⲣ ADV IMOD PronType=Tot 4 advmod _ _ 6 ⲥ ⲛⲧⲟⲥ PRON PPERO Definite=Def|Poss=Yes|PronType=Prs 5 det _ _ 7 ⲛⲁⲩ ⲛⲁⲩ VERB V VerbForm=Fin 2 xcomp _ _ 8 ⲉ ⲉ ADP PREP _ 9 case _ _ 9 ⲛⲁⲓ ⲡⲁⲓ DET PDEM Definite=Def|Number=Plur|PronType=Dem 7 obl _ _ 10 ϩⲓ ϩⲓ ADP PREP _ 12 case _ _ 11 ⲟⲩ ⲟⲩ DET ART Definite=Ind|Number=Sing|PronType=Art 12 det _ _ 12 ⲥⲟⲡ ⲥⲟⲡ NOUN N _ 7 obl _ _ ~~~ ~~~ conllu # visual-style 7 bgColor:blue # visual-style 7 fgColor:white # visual-style 3 bgColor:blue # visual-style 3 fgColor:white # visual-style 3 7 xcomp color:blue 1 ⲁ ⲁ AUX APST _ 3 aux _ _ 2 ⲩ ⲛⲧⲟⲟⲩ PRON PPERS Definite=Def|Number=Plur|Person=3|PronType=Prs 3 nsubj _ _ 3 ⲥⲉⲡⲥⲱⲡ ⲥⲟⲡⲥⲡ VERB V VerbForm=Fin 0 root _ _ 4 ϥ ⲛⲧⲟϥ PRON PPERO Definite=Def|Gender=Masc|Number=Sing|Person=3|PronType=Prs 3 obj _ _ 5 ⲇⲉ ⲇⲉ PART PTC _ 3 advmod _ _ 6 ⲉ ⲉ ADP PREP _ 7 mark _ _ 7 ⲧⲣⲉ ⲧⲣⲉ AUX ACAUS _ 3 xcomp _ _ 8 ϥ ⲛⲧⲟϥ PRON PPERS Definite=Def|Gender=Masc|Number=Sing|Person=3|PronType=Prs 9 nsubj _ _ 9 ⲡⲱⲱⲛⲉ ⲡⲱⲱⲛⲉ VERB V VerbForm=Fin 7 xcomp _ _ 10 ⲉⲃⲟⲗ ⲉⲃⲟⲗ ADV ADV _ 9 advmod _ _ 11 ϩⲛ ϩⲛ ADP PREP _ 13 case _ _ 12 ⲛⲉⲩ ⲡⲉⲩ DET PPOS Definite=Def|Number=Plur|Person=3|Poss=Yes|PronType=Prs 13 det _ _ 13 ⲧⲟϣ ⲧⲱϣ NOUN N _ 9 obl _ _ ~~~
45.686047
1,280
0.716976
yue_Hant
0.759328
219c15919a7f7840508bdad5fe173d72bc7f985e
2,935
md
Markdown
implementations/rust/ockam/ockam_core/CHANGELOG.md
kbeaulieu/ockam
552d079e2b9bb672f5415769ce59331228355940
[ "Apache-2.0" ]
null
null
null
implementations/rust/ockam/ockam_core/CHANGELOG.md
kbeaulieu/ockam
552d079e2b9bb672f5415769ce59331228355940
[ "Apache-2.0" ]
null
null
null
implementations/rust/ockam/ockam_core/CHANGELOG.md
kbeaulieu/ockam
552d079e2b9bb672f5415769ce59331228355940
[ "Apache-2.0" ]
null
null
null
# Changelog All notable changes to this crate will be documented in this file. The format is based on [Keep a Changelog](http://keepachangelog.com/en/1.0.0/) and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0.html). ## v0.12.0 - 2021-05-10 ### Added ### Changed - Renamed `async_worker` to `worker`. ### Deleted ## v0.11.2 - 2021-05-03 ### Changed - Updated dependencies. ## v0.11.1 - 2021-04-26 ### Changed - Updated dependencies. ## v0.11.0 - 2021-04-19 ### Changed - Updated dependencies. - Updated a routing error internal domain code. ## v0.10.0 - 2021-04-14 ### Changed - Improved debug printability of addresses. - Improved TCP transport initialisation. ## v0.9.0 - 2021-04-13 ### Changed - Updated dependencies. - Renamed Context address functions. - Improved printability of messages and payloads. ## v0.8.0 - 2021-04-12 ### Added - `Any` message type added to ockam_core crate. ### Changed - Updated dependencies. - `Routed` message wrapper function APIs renamed. - `Passthrough` type renamed to `Any`. - `msg_addr` moved from `Context` to `Routed`. - `Context` address renamed to `primary_address`. - Transport message fields renamed. ## v0.7.0 - 2021-04-05 ### Added - Expose onward route information to user workers. - Random address generation. ### Changed - Switch payload encoding from bincode to bare. - Split forwarding examples into two binaries. - Updated dependencies. - Rename route fields in transport message. ### Removed - RemoteMailbox has been moved to the `ockam` crate. ## v0.6.0 - 2021-03-22 ### Added - Routable message abstraction. - Builder for Routable messags. - Route metadata for messages. - Generic transport message. ### Changed - Updated dependencies. - Core dependencies, such as hashbrown and hex have been re-exported from this crate. ## v0.5.0 - 2021-03-04 ### Added - Support multiple addresses per worker. ## v0.4.0 - 2021-03-03 ### Added - Auto-implementation of the `Message` trait for certain types. ### Changed - The `Worker` trait and its methods are now async. - Updated dependencies. ## v0.3.0 - 2021-02-16 ### Added - Explicit `alloc`, `no_std`, and `std` features. - Generalized `Address` implementation. - Global crate lib facade wrapper around `std` and `core` re-exports, for cross-feature compatibility. - Message trait base implementation. ### Changed - Updated dependencies. - Improved documentation. ## v0.2.0 - 2021-02-04 ### Added - Runs Worker `initialize` function when the Worker is started. - Uses `From` trait in place of `Into` in Node error - Create Worker example ### Removed - Worker Builder ### Changed - Moved Worker and Address types to this crate. - Renamed executor commands to messages ## v0.1.0 - 2021-01-30 ### Added - `Error` - an error type that can be returned is both `std` and `no_std` modes. - `Result` - a result type that can be returned is both `std` and `no_std` modes.
22.751938
102
0.711755
eng_Latn
0.963976
219d6c0a6bd80a822b1fb35a6115e7ca94c19d4f
14
md
Markdown
README.md
Expelliarmus-3/Jeff
1c9eaef43d0158479098a2644c593c8ad7e371df
[ "Unlicense" ]
1
2021-06-03T13:22:08.000Z
2021-06-03T13:22:08.000Z
README.md
Expelliarmus-3/Jeff
1c9eaef43d0158479098a2644c593c8ad7e371df
[ "Unlicense" ]
null
null
null
README.md
Expelliarmus-3/Jeff
1c9eaef43d0158479098a2644c593c8ad7e371df
[ "Unlicense" ]
1
2021-07-09T09:46:22.000Z
2021-07-09T09:46:22.000Z
# Jeff newbie
4.666667
6
0.714286
eng_Latn
0.334706
219d714e49660447317099b077e48338d88344a1
700
md
Markdown
docs/error-messages/compiler-errors-1/compiler-error-c2055.md
lackhole/cpp-docs.ko-kr
38c9b85d3e8b45ea887365f7f628aa5832f25c18
[ "CC-BY-4.0", "MIT" ]
1
2020-07-24T04:27:56.000Z
2020-07-24T04:27:56.000Z
docs/error-messages/compiler-errors-1/compiler-error-c2055.md
lackhole/cpp-docs.ko-kr
38c9b85d3e8b45ea887365f7f628aa5832f25c18
[ "CC-BY-4.0", "MIT" ]
null
null
null
docs/error-messages/compiler-errors-1/compiler-error-c2055.md
lackhole/cpp-docs.ko-kr
38c9b85d3e8b45ea887365f7f628aa5832f25c18
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: 컴파일러 오류 C2055 ms.date: 11/04/2016 f1_keywords: - C2055 helpviewer_keywords: - C2055 ms.assetid: 6cec79cc-6bec-443f-9897-fbf5452718c7 ms.openlocfilehash: 9cb6e4d5891c5aefc9d66e7d70a5cd7685ccd393 ms.sourcegitcommit: a5fa9c6f4f0c239ac23be7de116066a978511de7 ms.translationtype: MT ms.contentlocale: ko-KR ms.lasthandoff: 12/20/2019 ms.locfileid: "75302044" --- # <a name="compiler-error-c2055"></a>컴파일러 오류 C2055 형식 목록이 아니라 정식 매개 변수 목록이 필요 합니다. 함수 정의에는 정식 매개 변수 목록 대신 매개 변수 형식 목록이 포함 되어 있습니다. ANSI C에서는 void 또는 줄임표 (`...`)가 아닌 경우 정식 매개 변수 이름을 지정 해야 합니다. 다음 샘플에서는 C2055를 생성 합니다. ```c // C2055.c // compile with: /c void func(int, char) {} // C2055 void func (int i, char c) {} // OK ```
23.333333
108
0.724286
kor_Hang
0.999121
219d836b5b0e3ff3c530af323a58ddd9a8dc768e
763
md
Markdown
_posts/2012-04-17-altdevpanel.md
thetoolsmiths/thetoolsmiths.github.io
f626cfe9c6683265cfbc4f74f57db394d840196a
[ "MIT" ]
9
2017-04-17T13:26:41.000Z
2021-09-28T05:27:15.000Z
_posts/2012-04-17-altdevpanel.md
thetoolsmiths/thetoolsmiths.github.io
f626cfe9c6683265cfbc4f74f57db394d840196a
[ "MIT" ]
11
2018-01-28T22:38:37.000Z
2018-04-22T21:38:16.000Z
_posts/2012-04-17-altdevpanel.md
thetoolsmiths/thetoolsmiths.github.io
f626cfe9c6683265cfbc4f74f57db394d840196a
[ "MIT" ]
5
2018-03-17T11:59:43.000Z
2019-08-14T16:38:12.000Z
--- id: 782 title: '#AltDevPanel' date: 2012-04-17T23:48:51+00:00 author: Geoff Evans layout: post guid: http://thetoolsmiths.org/?p=782 permalink: /2012/04/17/altdevpanel/ aktt_notify_twitter: - 'yes' al2fb_facebook_link_id: - 154659747957512_269894413100711 al2fb_facebook_link_time: - 2012-04-18T06:48:57+00:00 al2fb_facebook_link_picture: - wordpress=http://thetoolsmiths.org/?al2fb_image=1 aktt_tweeted: - "1" categories: - Productivity - Teams - Tool Design - Usabilty - User Interface --- Jeff Ward ([@fuzzybinary](http://twitter.com/fuzzybinary)) and I ([@gorlak](http://twitter.com/gorlak)) have a chat about game tools development with Mike Acton ([@mike_acton](http://twitter.com/mike_acton)) of #AltDevBlogADay (and the Tools SIG).
29.346154
247
0.74443
yue_Hant
0.41
219ef97266547e0fb3b7aac4a3472d898684efb4
23
md
Markdown
README.md
graceriley/Pizza-Transit-Website
3ae630d5ffe91df9bb73bd9d25ac7d657df4159a
[ "CC-BY-3.0" ]
null
null
null
README.md
graceriley/Pizza-Transit-Website
3ae630d5ffe91df9bb73bd9d25ac7d657df4159a
[ "CC-BY-3.0" ]
null
null
null
README.md
graceriley/Pizza-Transit-Website
3ae630d5ffe91df9bb73bd9d25ac7d657df4159a
[ "CC-BY-3.0" ]
null
null
null
# Pizza-Transit-Website
23
23
0.826087
yue_Hant
0.921235
219fad340bfd02fada5dfc5418d43d986640d044
9,247
md
Markdown
Exchange/ExchangeServer/high-availability/database-availability-groups/MetaCacheDatabase-setup.md
M34U2NV/OfficeDocs-Exchange
ed7614bed4ac2f88829d12bd337f0284b09333de
[ "CC-BY-4.0", "MIT" ]
1
2021-06-30T11:06:58.000Z
2021-06-30T11:06:58.000Z
Exchange/ExchangeServer/high-availability/database-availability-groups/MetaCacheDatabase-setup.md
M34U2NV/OfficeDocs-Exchange
ed7614bed4ac2f88829d12bd337f0284b09333de
[ "CC-BY-4.0", "MIT" ]
null
null
null
Exchange/ExchangeServer/high-availability/database-availability-groups/MetaCacheDatabase-setup.md
M34U2NV/OfficeDocs-Exchange
ed7614bed4ac2f88829d12bd337f0284b09333de
[ "CC-BY-4.0", "MIT" ]
1
2021-04-15T18:25:37.000Z
2021-04-15T18:25:37.000Z
--- localization_priority: Normal description: 'Summary: Learn about the MetaCacheDatabase (MCDB) feature in Exchange Server 2019, and how to configure it in your organization.' ms.topic: overview author: msdmaguire ms.author: dmaguire monikerRange: exchserver-2019 title: MetaCacheDatabase (MCDB) setup ms.collection: exchange-server ms.reviewer: toklima f1.keywords: - NOCSH audience: ITPro ms.prod: exchange-server-it-pro manager: serdars --- # MetaCacheDatabase (MCDB) setup The MetaCacheDatabase (MCDB) feature is included in Exchange Server 2019. It allows a database availability group (DAG) to be accelerated by utilizing solid state disks (SSDs). `Manage-MetaCacheDatabase.ps1` is an automation script created for Exchange Server administrators to set up and manage MCDB instances in their Exchange 2019 DAGs. After installing Exchange Server 2019, you can find `Manage-MetaCacheDatabase.ps1` here: ***drive*:\\Program Files\\Microsoft\\Exchange Server\\V15\\Scripts**. To make the **Manage-MCDB** CMDLet available in your Exchange Management Shell session, do the following: ```PowerShell cd $exscripts . .\Manage-MetaCacheDatabase.ps1 ``` You use this script to configure MCDB prerequisites on a properly configured DAG, to enable or disable MCDB, and to configure and repair MCDB on your servers. ## SSD guidance All SSDs used for MCDB need to be of the same capacity and type. A symmetrical configuration between servers is required, which means there needs to be an identical number of SSDs in each server, and the SSDs all need to be the same size. > [!NOTE] > The **Manage-MCDB** cmdlet will only work with devices exposed as **MediaType SSD** by Windows. It is recommended to target a 1:3 ratio between SSD and HDD devices per server. Therefore, deploy one SSD for every three HDDs. In order to avoid having to reduce the number of HDDs in the server, consider using M.2 form factor SSDs. Providing 5% to 6% of SSD capacity relative to total HDD capacity is sufficient for on-premises deployments. For example, if your server contains 100 TB of HDD capacity for mailbox databases, an allocation of 5 TB to 6 TB for SSD capacity is enough. The SSDs you use should qualify for "mixed use" and support one drive write per day (DWPD) or greater in terms of write endurance. ## Prerequisites The following prerequisites are required for successful configuration and use of MCDB: 1. The DAG is configured for AutoReseed. For more information, see the following topics: - [AutoReseed](autoreseed.md) - [Configure AutoReseed for a database availability group](../manage-ha/configure-dag-autoreseed.md) 2. RAW SSD drives are installed with the same SSD count and size for each server in the DAG. Make sure that all SSDs are completely empty, unformatted, and not write-protected. To verify this, you can use use [DiskPart](/windows-server/administration/windows-commands/diskpart) or [Clear-Disk](/powershell/module/storage/clear-disk). 3. Exchange Server 2019. ## MCDB setup The process of setting up MCDB can be broken down into four basic steps: 1. Set the correct values for the DAG you want to enable for MCDB. 2. Update Active Directory (AD) settings and wait for propagation (by running `ConfigureMCDBPrerequisite`). 3. Allow MCDB acceleration for each server of the DAG (by running `ServerAllowMCDB`). 4. Create the necessary infrastructure (Volumes, Mount Points) for MCDB on each server (by running `ConfigureMCDBOnServer`). 5. Let databases fail over to pick up the new settings. After successful execution of all four steps, MCDB acceleration will begin for every database instance with a corresponding MCDB instance. The following sections describe how to utilize the `Manage-MetaCacheDatabase.ps1` script to achieve the above four steps. ### Step 1: Configure proper values on the DAG you want to enable MCDB for These DAG parameters are used to calculate the proper MCDB size on your SSD drives: - *AutoDagTotalNumberOfDatabases*: The number of databases in your DAG (for example, 50). - *AutoDagDatabaseCopiesPerDatabase*: The number of active and passive copies each individual database has. - *AutoDagTotalNumberOfServers*: The amount of servers within your DAG, so between 2 and 16. For example: ```PowerShell Set-DatabaseAvailabilityGroup testdag1 -AutoDagTotalNumberOfDatabases 20 -AutoDagDatabaseCopiesPerDatabase 4 -AutoDagTotalNumberOfServers 8 ``` ### Step 2: Run Manage-MCDB -ConfigureMCDBPrerequisite This parameter sets the Active Directory state for the DAG object. Full replication of the Active Directory state is required before MCDB can function properly on all servers. **ParameterSetIdentifier**: - _ConfigureMCDBPrerequisite_ **Parameters**: |Parameter|Required|Description| |:-------|:--------|:--------| |DagName|True|Name of the Database availability group.| |SSDSizeInBytes|True|The capacity in bytes of each SSD in the server to be used for MCDB.| |SSDCountPerServer|True|The count of SSD devices to be utilize for MCDB in each server.| **Scope**: - **DAG**: _ConfigureMCDBPrerequisite_ operates on a DAG object. > [!NOTE] > MCDB will utilize up to 95% of an SSD's physical capacity. The remaining 5% is kept free to account for file system and partition overhead, as well as for a small amount of additional buffer and over-provisioning. **Example**: ```PowerShell Manage-MCDB -DagName TestDag1 -ConfigureMCDBPrerequisite -SSDSizeInBytes 5242880000 -SSDCountPerServer 2 ``` ![MCDB configure prerequisites](../../media/mcdb1.png) ### Step 3: Run Manage-MCDB -ServerAllowMCDB This command sets the local state on each DAG member to allow/disallow MCDB population and read acceleration. **ParameterSetIdentifier**: - _ServerAllowMCDB_ **Parameters**: | Parameter|Required|Description| |:-------|:-----|:---| |DagName|True|Name of the Database availability group.| |ServerName|True|Specifies the server to enable MetaCacheDatabase on.| |ForceFailover|Optional|This Boolean switch can be utilized to cause all databases on a server to fail over. This is required to make all configuration changes take effect and to begin utilizing MCDB after mount points and database instances have been successfully created in [Step 4: Run Manage-MCDB -ConfigureMCDBOnServer](#step-4-run-manage-mcdb--configuremcdbonserver). It is also needed to disable SSD acceleration.| **Scope**: - **Server**: You need to run _ServerAllowMCDB_ on each server in the DAG. **Examples**: ```PowerShell Manage-MCDB -DagName TestDag1 -ServerAllowMCDB $true -ServerName "exhs-5046" ``` ```PowerShell Manage-MCDB -DagName TestDag1 -ServerAllowMCDB $false -ServerName "exhs-5046" -ForceFailover $true ``` ![MCDB run ServerAllowMCDB](../../media/mcdb2.png) ### Step 4: Run Manage-MCDB -ConfigureMCDBOnServer This command identifies unformatted SSD devices and formats them, and also creates the necessary mount points on a server for hosting MCDB instances. This parameter set can also be used to re-create mount points on a raw SSD that was added to replace a failed SSD. **ParameterSetIdentifier**: - _ConfigureMCDBOnServer_ **Parameters**: |Parameter|Required|Description| |:-------|:--------|:--------| |DagName|True|Name of the Database availability group.| |ServerName|True|Specifies the server to identify unformatted SSD devices and create mount points on.| |SSDSizeInBytes|True|This is the capacity, in bytes, of each SSD in the server to be used for MCDB.| **Scope**: - **Server**: You need to run _ConfigureMCDBOnServer_ on each server in the DAG. Example: ```PowerShell Manage-MCDB -DagName TestDag1 -ConfigureMCDBOnServer -ServerName "exhs-4056" -SSDSizeInBytes 5242880000 ``` ![Run MCDBOnServer1](../../media/mcdb3.png) ![Run MCDBOnServer2](../../media/mcdb4.png) After performing the previous three steps (configuring _ConfigureMCDBPrerequisite_, _ServerAllowMCDB_, and _ConfigureMCDBOnServer_), the MCDB state will display as **Storage Offline**. This means that the environment is prepared and ready for MCDB instances to be created and populated. The next fail over of the database instance will cause the creation of the MCDB instance and enable acceleration. The instances will transition through the health states shown in [MCDB health states](#mcdb-health-states). You can use the _ServerAllowMCDB_ parameter set to cause fail overs of all DB instances present on a given server. Alternatively, you can use the **Move-ActiveMailboxDatabase** cmdlet to cause individual databases to fail over. ```PowerShell Manage-MCDB -DagName TestDag1 -ServerAllowMCDB:$true -ServerName "exhs-5046" -ForceFailover $true ``` ## MCDB health states Use **Get-MailboxDatabaseCopyStatus** to query the state of the MCDB instances. There are five states that an MCDB instance can be in, as shown in the following table: |State|Description| |:------|:------| | Disabled|MCDB is turned off.| |StorageOffline|Basic infrastructure is missing or inaccessible, such as mount points or file paths. This is the state MCDB is in following an SSD failure.| |Offline|Errors at the logical level, for example missing MCDB instances.| |Initializing|Transient state, the system is determining what other state it should be in.| |Healthy|Ready to serve requests.|
44.671498
508
0.777549
eng_Latn
0.965092
219fb23215118ed635fb6759e5b4bf9c6d0fb3a5
192
md
Markdown
README.md
cdesqaz/MyBlogPyProject
ed4d04e338e22c3609999668bfb76ae94516700a
[ "MIT" ]
null
null
null
README.md
cdesqaz/MyBlogPyProject
ed4d04e338e22c3609999668bfb76ae94516700a
[ "MIT" ]
null
null
null
README.md
cdesqaz/MyBlogPyProject
ed4d04e338e22c3609999668bfb76ae94516700a
[ "MIT" ]
null
null
null
# MyBlogPyProject This is my repository where I have been hang on all the software used to build it. Any software is from their owner and developers licenses only por educational uses. Thank.
64
173
0.807292
eng_Latn
0.99999
219fdc9b86b00d44ae39d7db0e08b1005f4d9eb8
436
md
Markdown
content_bk/ja/news/2011-01-07-0036.md
hi-ut/dev
0017908fe3416e74a8cbb670cf17c2a489078a63
[ "Apache-2.0" ]
null
null
null
content_bk/ja/news/2011-01-07-0036.md
hi-ut/dev
0017908fe3416e74a8cbb670cf17c2a489078a63
[ "Apache-2.0" ]
null
null
null
content_bk/ja/news/2011-01-07-0036.md
hi-ut/dev
0017908fe3416e74a8cbb670cf17c2a489078a63
[ "Apache-2.0" ]
null
null
null
--- date: 2011-01-07 lang: ja featured: false type: news --- : 2011年1月30日(日)・2月6日(日)・2月13日(日)・2月20日(日)に、科学研究費補助金(学術創成研究費)「目録学の構築と古典学の再生-天皇家・公家文庫の実態復原と伝統的知識体系の解明-」の主催による「陽明文庫講座 よみがえる宮廷文化」(共催:立命館大学、後援:財団法人陽明文庫・東京大学史料編纂所)を、立命館大学朱雀キャンパスにて、開講しました。<br/> 詳細は<a href="http://alumni.ritsumei.jp/news/004702.html">立命館大学のホームページ</a>の<a href="http://www.ritsumei.jp/topics_pdf/admin_8092f13b38072b79e8bca37d1dbf8a5c_1294712913_.pdf">案内</a>をご参照下さい。
43.6
186
0.779817
yue_Hant
0.393763
219fdd7138b9c6950d063e84fbd2dc1c1f80b4b0
3,723
md
Markdown
content/blog/webpack/6-code-split.md
hoilzz/hoilzz-blog
b68d1a54485f32762a81cab0a692d27235114871
[ "MIT" ]
null
null
null
content/blog/webpack/6-code-split.md
hoilzz/hoilzz-blog
b68d1a54485f32762a81cab0a692d27235114871
[ "MIT" ]
3
2019-02-13T15:52:26.000Z
2021-01-28T21:08:06.000Z
content/blog/webpack/6-code-split.md
hoilzz/hoilzz-blog
b68d1a54485f32762a81cab0a692d27235114871
[ "MIT" ]
null
null
null
--- title: '[나만의 웹팩 만들기] 6. code split' date: 2019-2-7 23:02:30 category: webpack --- > 코드를 보려면 [이동](https://github.com/hoilzz/create-react-packzz/tree/6-code-split) 코드 스플릿은 Webpack의 가장 강력한 기능 중 하나다. - 코드를 다양한 번들로 나눈다 - 이 번들은 필요할 때 로드되거나 병렬로 로드될 수 있다 - 작은 번들을 만들 때 사용된다 - 리소스 부하 우선순위 제어할 수 있다 Code Split을 잘 사용하면 load time에서 이득을 얻을 수 있다. 웹팩에서 보통 다음 세 가지 방법을 통해 코드 스플릿을 한다. - Entry Points : `entry` 를 이용한 스플릿 - 중복 방지 : 중복 제거와 chunk별로 스플릿하기 위해 `SplitChunksPlugin`을 이용하기 - Dynamic Import : 모듈 내에서 inline function 호출을 통해 코드 분할하기(inline 함수 내에서 import문을 이용하여 모듈 로드) ## 1. Entry Point 너무 쉬워서 넘어가겠다. [링크](https://webpack.js.org/guides/code-splitting/#entry-points) 참조 ## 2. 중복 방지 **기존 entry나 새로운 chunk의 공통 의존성을 추출한다.** 일부러 중복을 발생시키기 위해 - src/another.js 를 생성하여 lodash를 import 하고, - src/index.js에서도 동일하게 lodash를 import 해보자. 빌드를 해보면 2개의 번들에서 lodash를 각각 가지고 있는 모습을 볼 수 있다. 이러한 중복을 줄이기 위해 다음 설정을 추가하자. ```js // webpack.common.js module.exports = { // ... optimization: { splitChunks: { chunks: 'all', }, }, } ``` 위 설정 적용 후, npm run build를 다시 실행하면 빌드 결과물이 1개 추가된다. dist/app과 another 번들 파일을 확인해보면 lodash가 빠져있고,dist/venders~를 확인해보면 lodash가 포함된것을 알 수 있다. 즉, 공통 의존성은 따로 split하여 중복 제거하는 것을 알 수 있다. --- ## Dynamic imports Dynamic code splitting에 대해서 webpack은 2가지 테크닉을 지원한다. 권장되는 방법은 dynamic import를 위해 ES proposal에 허용된 `import()` 사용하는 것이다. > import()는 promise를 내부적으로 사용. import() 문 사용할꺼면 older browser에서 Promise polyfill을 이용해야한다는 거 잊지 말자. ```js module.exports = { // ... output: { // ... chunkFilename: '[name].bundle.js', }, } ``` chunkFilename은 아래 webpackChunkName의 주석에 작성된 이름을 가져다 사용한다. ```js // index.js function dynamicLoad() { return import(/* webpackChunkName: "asyncHoilzz" */ './async').then( ({ default: asyncFunc }) => { asyncFunc() } ) } dynamicLoad() // async.js export default () => { console.log('this is async!') } ``` 위와 같이 파일 생성 및 코드 추가하여 npm run build하면 dynamic import가 되는 파일이 따로 빌드되는 것을 확인할 수 있다. 이 코드는 지우자. react 적용할 때 다시 알아보자. ## Bundle 분석하기 (귀찮으면 스킵하거나 [복붙](https://github.com/hoilzz/create-react-boilerplate-hoil/commit/446cf8593ee4b8c1a5532b1e0a8a6636748d65cc)하자.) code split하기 시작하면, output을 분석하는 것은 유용하다. 가장 인기가 많은 [webpack-bundle-analyzer](https://github.com/webpack-contrib/webpack-bundle-analyzer)를 설치하여 이용해보자. 얘로 할 수 있는 건 - bundle 내부에 **진짜로** 뭐가 있는지 나타낸다 - 크기를 많이 차지하는 모듈을 확인할 수 있다. - 실수로 번들에 추가된 모듈을 찾을 수도 있다. - 제일 좋은건 minified bundle을 지원하고, 번들링된 모듈의 real size, gzipped size를 알 수 있다! ## SplitChunksPlugin SplitChunksPlugin에 대해 자세히 알아보자. 자세히 알아보기 싫으면 아래 코드만 적용하자. 원래 청크는 웹팩 내부 그래프에서 parent-child 관계로 연결되어있다. 중복 피하기, 최적화를 하는데 다음 조건에 기반하여 chunk를 쪼갠다. - 새로운 청크는 - 공유되거나 - `node_modules` folder에 있는 모듈(ex. react)일 수 있다. (모듈은~ 이 문장이 뭔소린지 잘 모르겠다.) - 새로운 청크는 30Kb보다 크다. - on demand 청크를 로드할 때 병렬 요청의 최대 개수는 5개 이하다. - 첫 페이지 병렬 요청의 최대 개수는 3개 이하다. 마지막 2가지 조건을 충족 할 때, 더 큰 청크가 권장된다. 즉 여러개 작은 청크보다는 1개 청크가 더 낫다는 얘기같다. 청크를 어떤식으로 활용하는지 알고 싶다면 다음 [공홈 링크](https://webpack.js.org/plugins/split-chunks-plugin/#examples)를 참고하자. 이 중 2번째 예제를 참고하여 만들겠다. `node_modules`의 모든 코드를 포함하는 `vendors` chunk를 만들자. ```js module.exports = { //... optimization: { splitChunks: { cacheGroups: { commons: { test: /[\\/]node_modules[\\/]/, name: 'vendors', chunks: 'all', }, }, }, }, } ``` splitChunks.cacheGroups - cachegroups은 특정 조건으로 chunk를 생성하라고 webpack에게 말하는거. - node_modules를 포함하고 - 이름은 vendors - chunks는 3가지 값을 가짐 - initial : 초기 로드 되는 것만 청크에 추가 - async : 동적 로드 되는 것만 청크에 추가 - all : 3개 중 가장 강력함. initial + async인데, 청크는 async와 non-async 청크 간에 공유될 수 있음 위 설정은 모든 외부 패키지를 포함하는 큰 청크를 만들거다. core framework, utilities만 vendors로 포함시키고 나머지 종속성은 동적 로드하는 것을 권장한다.
22.563636
127
0.66801
kor_Hang
1.00001
219fe07b455ab8eed27ba47dec49a897ee201f88
5,037
md
Markdown
mac/setup-vsmac-tools-unity.md
LouisFr81/visualstudio-docs.fr-fr
e673e36b5b37ecf1f5eb97c42af84eed9c303200
[ "CC-BY-4.0", "MIT" ]
null
null
null
mac/setup-vsmac-tools-unity.md
LouisFr81/visualstudio-docs.fr-fr
e673e36b5b37ecf1f5eb97c42af84eed9c303200
[ "CC-BY-4.0", "MIT" ]
null
null
null
mac/setup-vsmac-tools-unity.md
LouisFr81/visualstudio-docs.fr-fr
e673e36b5b37ecf1f5eb97c42af84eed9c303200
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: Installation des outils Visual Studio pour Mac pour Unity description: Configuration et installation des outils Unity à utiliser dans Visual Studio pour Mac author: therealjohn ms.author: johmil ms.date: 05/25/2018 ms.assetid: 83FDD7A3-5D16-4B4B-9080-078E3FB5C623 ms.openlocfilehash: d490b4c1268beb4a5ad55263cb186d838005f718 ms.sourcegitcommit: 5a65ca6688a2ebb36564657d2d73c4b4f2d15c34 ms.translationtype: HT ms.contentlocale: fr-FR ms.lasthandoff: 01/15/2019 ms.locfileid: "54315525" --- # <a name="set-up-visual-studio-for-mac-tools-for-unity"></a>Installer les outils Visual Studio pour Mac pour Unity Cette section explique comment préparer l’utilisation des outils Visual Studio pour Mac pour Unity. ## <a name="install-visual-studio-for-mac"></a>Installer Visual Studio pour Mac ### <a name="unity-bundled-installation"></a>Installation collective de Unity À partir de Unity 2018.1, Visual Studio pour Mac représente l’IDE (environnement de développement intégré) C# par défaut pour Unity. Il est inclus dans l’Assistant Téléchargement Unity ainsi que dans l’outil d’installation Unity Hub. Téléchargez Unity à partir de [store.unity.com](https://store.unity.com/). Lors de l’installation, vérifiez que Visual Studio pour Mac est coché dans la liste des composants à installer avec Unity : #### <a name="unity-hub"></a>Unity Hub ![installation de unity hub](media/setup-vsmac-tools-unity-image7.png) #### <a name="unity-download-assistant"></a>Assistant Téléchargement de Unity ![installation de l’assistant téléchargement de Unity](media/setup-vsmac-tools-unity-image8.png) #### <a name="check-for-updates-to-visual-studio-for-mac"></a>Vérifier les mises à jour de Visual Studio pour Mac La version de Visual Studio pour Mac fournie avec l’installation de Unity n’est peut-être pas la dernière version. Nous vous recommandons de rechercher les mises à jour pour être sûr d’avoir accès aux tout derniers outils et fonctionnalités. * [Mise à jour de Visual Studio pour Mac](update.md) ### <a name="manual-installation"></a>Installation manuelle Si vous avez déjà Unity 5.6.1 ou version ultérieure, mais que vous n’avez pas Visual Studio pour Mac, vous pouvez installer manuellement Visual Studio pour Mac. Toutes les éditions de Visual Studio pour Mac sont fournies avec les outils Visual Studio pour Mac pour Unity, notamment l’édition Community gratuite : * Téléchargez Visual Studio pour Mac à partir de [visualstudio.microsoft.com](https://visualstudio.microsoft.com/). * Les outils Visual Studio pour Mac pour Unity sont installés automatiquement lors du processus d’installation. * Pour obtenir de l’aide supplémentaire sur l’installation, suivez les étapes du [guide d’installation](/visualstudio/mac/installation/?view=vsmac-2017) . > [!NOTE] > Visual Studio pour Mac pour Unity nécessite Unity 5.6.1 ou ultérieur. Pour vérifier que les outils Visual Studio pour Unity sont activés dans votre version de Unity, sélectionnez **À propos de Unity** dans le menu Unity et recherchez le texte « Microsoft Visual Studio Tools pour Unity activé » dans la partie inférieure gauche de la boîte de dialogue. > > ![À propos de Unity](media/setup-vsmac-tools-unity-image3.png) ## <a name="confirm-that-the-visual-studio-for-mac-tools-for-unity-extension-is-enabled"></a>Vérifiez que l’extension Outils Visual Studio pour Mac pour Unity est activée. L’extension Outils Visual Studio pour Mac pour Unity doit normalement être activée par défaut, vous pouvez néanmoins le vérifier, ainsi que le numéro de la version installée : 1. Dans le menu Visual Studio, sélectionnez **Extensions...**. ![Sélectionner Extensions](media/setup-vsmac-tools-unity-image1.png) 2. Développez la section Développement de jeux et vérifiez qu’il existe bien une entrée Outils Visual Studio pour Mac pour Unity. ![Afficher l’entrée Unity](media/setup-vsmac-tools-unity-image2.png) ## <a name="configure-unity-for-use-with-visual-studio-for-mac"></a>Configurer Unity pour une utilisation avec Visual Studio pour Mac À partir de Unity 2018.1, Visual Studio doit être l’éditeur de script externe par défaut dans Unity. Vous pouvez vérifier cela ou choisir Visual Studio comme éditeur de script externe : 1. Sélectionnez **Préférences...** dans le menu Unity. ![Sélectionner Préférences](media/setup-vsmac-tools-unity-image4.png) 2. Dans la boîte de dialogue Préférences, sélectionnez l’onglet **Outils externes**. 3. Dans la liste déroulante Éditeur de script externe, choisissez **Visual Studio** s’il est répertorié, sinon sélectionnez **Parcourir...**. ![Sélectionner Visual Studio](media/setup-vsmac-tools-unity-image5.png) 4. Si vous avez sélectionné **Parcourir...**, accédez au répertoire Applications, sélectionnez Visual Studio puis cliquez sur **Ouvrir**. ![Sélectionner Ouvrir](media/setup-vsmac-tools-unity-image6.png) 5. Une fois que Visual Studio est sélectionné dans la liste **Éditeur de script externe**, fermez la boîte de dialogue Préférences pour terminer le processus de configuration.
59.964286
354
0.783601
fra_Latn
0.925732
219ff1506ec7803295cc51e8d7b456898b230308
982
md
Markdown
docs/framework/wcf/diagnostics/tracing/system-servicemodel-channels-msmqenteredbatch.md
douglasbreda/docs.pt-br
f92e63014d8313d5e283db2e213380375cea9a77
[ "CC-BY-4.0", "MIT" ]
null
null
null
docs/framework/wcf/diagnostics/tracing/system-servicemodel-channels-msmqenteredbatch.md
douglasbreda/docs.pt-br
f92e63014d8313d5e283db2e213380375cea9a77
[ "CC-BY-4.0", "MIT" ]
null
null
null
docs/framework/wcf/diagnostics/tracing/system-servicemodel-channels-msmqenteredbatch.md
douglasbreda/docs.pt-br
f92e63014d8313d5e283db2e213380375cea9a77
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: System.ServiceModel.Channels.MsmqEnteredBatch ms.date: 03/30/2017 ms.assetid: 5b787c18-1962-4083-93d1-adb680295189 ms.openlocfilehash: e9adcc97c08e1598a100129892b9a281b256b855 ms.sourcegitcommit: 3d5d33f384eeba41b2dff79d096f47ccc8d8f03d ms.translationtype: MT ms.contentlocale: pt-BR ms.lasthandoff: 05/04/2018 ms.locfileid: "33477666" --- # <a name="systemservicemodelchannelsmsmqenteredbatch"></a>System.ServiceModel.Channels.MsmqEnteredBatch System.ServiceModel.Channels.MsmqEnteredBatch ## <a name="description"></a>Descrição MSMQ entrou no modo de lote. ## <a name="see-also"></a>Consulte também [Rastreamento](../../../../../docs/framework/wcf/diagnostics/tracing/index.md) [Usando o rastreamento para solucionar problemas do seu aplicativo](../../../../../docs/framework/wcf/diagnostics/tracing/using-tracing-to-troubleshoot-your-application.md) [Administração e diagnósticos](../../../../../docs/framework/wcf/diagnostics/index.md)
44.636364
175
0.761711
yue_Hant
0.137577
219ff2d574eb4323af11b402a2ad525623d24a9c
2,614
md
Markdown
README.md
matthiasduyck/CloneHeroSaveGameEditor
26556c685abc493e543920709dc899a5ef5bd093
[ "MIT" ]
3
2021-04-24T18:36:05.000Z
2021-09-29T20:32:13.000Z
README.md
matthiasduyck/CloneHeroSaveGameEditor
26556c685abc493e543920709dc899a5ef5bd093
[ "MIT" ]
null
null
null
README.md
matthiasduyck/CloneHeroSaveGameEditor
26556c685abc493e543920709dc899a5ef5bd093
[ "MIT" ]
1
2021-10-18T05:53:52.000Z
2021-10-18T05:53:52.000Z
# Clone Hero Save Game Editor <p><a target="_blank" rel="noopener noreferrer" href="https://raw.githubusercontent.com/matthiasduyck/CloneHeroSaveGameEditor/master/Clone%20Hero%20Save%20Game%20Editor%20Icon.png"><img src="https://raw.githubusercontent.com/matthiasduyck/CloneHeroSaveGameEditor/master/Clone%20Hero%20Save%20Game%20Editor%20Icon.png" width="160" align="right" style="max-width:100%;"></a></p> <strong>Clone Hero Save Game Editor</strong> is a tiny open-source Windows application to modify your Clone Hero save file. You can change individual parameters for each high score for each individual song. If you want to 'reset' individual scores, jump up in difficulty level or previously used an easier modifier and now want see your high scores increase again over time, you can easily lower or change your score to be able to beat it again. <h3 align="center"> <g-emoji class="g-emoji" alias="warning" fallback-src="https://github.githubassets.com/images/icons/emoji/unicode/26a0.png">⚠️</g-emoji> This project is still in active development/reverse engineer phase <g-emoji class="g-emoji" alias="warning" fallback-src="https://github.githubassets.com/images/icons/emoji/unicode/26a0.png">⚠️</g-emoji> </h3> <strong>Currently the editor may not fully support every possible save file (too complex, too many scores,...). The program does a format validation check when reading in your file to validate this. Backup files are created when saving in case something still goes wrong. There is barely any validation in the editor yet, if you mess up the input it will break your file.</strong> <p><a target="_blank" rel="noopener noreferrer" href="https://raw.githubusercontent.com/matthiasduyck/CloneHeroSaveGameEditor/master/clone%20hero%20save%20game%20editor%20alpha%204%20screenshot.png"><img src="https://raw.githubusercontent.com/matthiasduyck/CloneHeroSaveGameEditor/master/clone%20hero%20save%20game%20editor%20alpha%204%20screenshot.png" style="max-width:100%;"></a></p> Core functionality, features and UI may still change drastically. <strong>You should also expect this program to possibly break/delete/corrupt your save files, proceed with caution, make the needed backups. You have been warned!</strong> Written in C# with WPF. You'll need Visual Studio(2019) to build the project yourself. Some alpha releases are available on the <a href="https://github.com/matthiasduyck/CloneHeroSaveGameEditor/releases">releases page.</a> Pull requests are welcome, as well as feedback or feature requests. If you find (reproducible) issues, please submit them along with your files, thanks!
90.137931
445
0.79036
eng_Latn
0.907713
21a0038efd07656a0bad29bf5faa9dd8cca28301
30,542
md
Markdown
articles/hdinsight-submit-hadoop-jobs-programmatically.md
guyburstein/azure-content
78ab6ef86246e33a553a1c22655b27a9b7c49412
[ "CC-BY-3.0" ]
null
null
null
articles/hdinsight-submit-hadoop-jobs-programmatically.md
guyburstein/azure-content
78ab6ef86246e33a553a1c22655b27a9b7c49412
[ "CC-BY-3.0" ]
null
null
null
articles/hdinsight-submit-hadoop-jobs-programmatically.md
guyburstein/azure-content
78ab6ef86246e33a553a1c22655b27a9b7c49412
[ "CC-BY-3.0" ]
null
null
null
<properties linkid="manage-services-hdinsight-submit-hadoop-jobs" urlDisplayName="HDInsight Administration" pageTitle="Submit Hadoop jobs in HDInsight | Azure" metaKeywords="hdinsight, hdinsight administration, hdinsight administration azure, hive, mapreduce, HDInsight .NET SDK, powershell, submit mapreduce jobs, submit hive jobs, development, hadoop, apache" description="Learn how to submit Hadoop jobs to Azure HDInsight Hadoop." umbracoNaviHide="0" disqusComments="1" editor="cgronlun" manager="paulettm" services="hdinsight" documentationCenter="" title="Submit Hadoop jobs in HDInsight" authors="jgao" /> # Submit Hadoop jobs in HDInsight In this article, you will learn how to submit MapReduce and Hive jobs using PowerShell and HDInsight .NET SDK. **Prerequisites:** Before you begin this article, you must have the following: * An Azure HDInsight cluster. For instructions, see [Get started with HDInsight][hdinsight-get-started] or [Provision HDInsight clusters][hdinsight-provision]. * Install and configure Azure PowerShell. For instructions, see [Install and configure Azure PowerShell][powershell-install-configure]. ##In this article * [Submit MapReduce jobs using PowerShell](#mapreduce-powershell) * [Submit Hive jobs using PowerShell](#hive-powershell) * [Submit Sqoop jobs using PowerShell](#sqoop-powershell) * [Submit MapReduce jobs using HDInsight .NET SDK](#mapreduce-sdk) * [Submit Hadoop Streaming MapReduce jobs using HDInsight .NET SDK](#streaming-sdk) * [Submit Hive Jobs using HDInsight .NET SDK](#hive-sdk) * [Next steps](#nextsteps) ##<a id="mapreduce-powershell"></a> Submit MapReduce jobs using PowerShell Azure PowerShell is a powerful scripting environment that you can use to control and automate the deployment and management of your workloads in Azure. For more information on using PowerShell with HDInsight, see [Administer HDInsight using PowerShell][hdinsight-admin-powershell]. Hadoop MapReduce is a software framework for writing applications which process vast amounts of data. HDInsight clusters come with a jar file, located at *\example\jars\hadoop-examples.jar*, which contains several MapReduce examples. This file has been renamed to hadoop-mapreduce-examples.jar on version 3.0 HDInsight clusters. One of the examples is for counting word frequencies in source files. In this session, you will learn how to use PowerShell from a workstation to run the word count sample. For more information on developing and running MapReduce jobs, see [Use MapReduce with HDInsight][hdinsight-use-mapreduce]. **To run the word count MapReduce program using PowerShell** 1. Open **Azure PowerShell**. For instructions on opening the Azure PowerShell console window, see the [Install and configure Azure PowerShell][powershell-install-configure]. 3. Set these two variables by running the following PowerShell commands: $subscriptionName = "<SubscriptionName>" $clusterName = "<HDInsightClusterName>" The subscription is the one you used to create the HDInsight cluster. And the HDInsight cluster is the one you want to use to run the MapReduce job. 5. Run the following commands to create a MapReduce job definition: # Define the word count MapReduce job $wordCountJobDefinition = New-AzureHDInsightMapReduceJobDefinition -JarFile "wasb:///example/jars/hadoop-examples.jar" -ClassName "wordcount" -Arguments "wasb:///example/data/gutenberg/davinci.txt", "wasb:///example/data/WordCountOutput" There are two arguments. The first one is the source file name, and the second is the output file path. For more information of the wasb prefix, see [Use Azure Blob storage with HDInsight][hdinsight-storage]. 6. Run the following command to run the MapReduce job: # Submit the MapReduce job Select-AzureSubscription $subscriptionName $wordCountJob = Start-AzureHDInsightJob -Cluster $clusterName -JobDefinition $wordCountJobDefinition In addition to the MapReduce job definition, you also provide the HDInsight cluster name where you want to run the MapReduce job. 7. Run the following command to check the completion of the MapReduce job: # Wait for the job to complete Wait-AzureHDInsightJob -Job $wordCountJob -WaitTimeoutInSeconds 3600 8. Run the following command to check any errors with running the MapReduce job: # Get the job standard error output Get-AzureHDInsightJobOutput -Cluster $clusterName -JobId $wordCountJob.JobId -StandardError The following screenshot shows the output of a successful run. Otherwise, you will see some error messages. ![HDI.GettingStarted.RunMRJob][image-hdi-gettingstarted-runmrjob] **To retrieve the results of the MapReduce job** 1. Open **Azure PowerShell**. 2. Set these three variables by running teh following PowerShell commands: $subscriptionName = "<SubscriptionName>" $storageAccountName = "<StorageAccountName>" $containerName = "<ContainerName>" The Azure Storage account is the one you specified during the HDInsight cluster provision. The storage account is used to host the Blob container that is used as the default HDInsight cluster file system. The Blob storage container name usually share the same name as the HDInsight cluster unless you specify a different name when you provision the cluster. 3. Run the following commands to create an Azure storage context object: # Create the storage account context object Select-AzureSubscription $subscriptionName $storageAccountKey = Get-AzureStorageKey $storageAccountName | %{ $_.Primary } $storageContext = New-AzureStorageContext -StorageAccountName $storageAccountName -StorageAccountKey $storageAccountKey The Select-AzureSubscription is used to set the current subscription in case you have multiple subscriptions, and the default subscription is not the one to use. 4. Run the following command to download the MapReduce job output from the Blob container to the workstation: # Get the blob content Get-AzureStorageBlobContent -Container $ContainerName -Blob example/data/WordCountOutput/part-r-00000 -Context $storageContext -Force The *example/data/WordCountOutput* folder is the output folder specified when you run the MapReduce job. *part-r-00000* is the default file name for MapReduce job output. The file will be download to the same folder structure on the local folder. For example, in the following sreenshoot, the current folder is the C root folder. The file will be downloaded to the *C:\example\data\WordCountOutput\* folder. 5. Run the following command to print the MapReduce job output file: cat ./example/data/WordCountOutput/part-r-00000 | findstr "there" ![HDI.GettingStarted.MRJobOutput][image-hdi-gettingstarted-mrjoboutput] The MapReduce job produces a file named *part-r-00000* with the words and the counts. The script uses the findstr command to list all of the words that contains "there". > [WACOM.NOTE] If you open ./example/data/WordCountOutput/part-r-00000, a multi-line output from a MapReduce job, in Notepad, you will notice the line breaks are not renter correctly. This is expected. ##<a id="hive-powershell"></a> Submit Hive jobs using PowerShell Apache [hdinsight-use-hive][apache-hive] provides a means of running MapReduce job through an SQL-like scripting language, called *HiveQL*, which can be applied towards summarization, querying, and analysis of large volumes of data. HDInsight clusters come with a sample Hive table called *hivesampletable*. In this session, you will use PowerShell to run a Hive job for listing some data from the Hive table. **To run a Hive job using PowerShell** 1. Open **Azure PowerShell**. For instructions of opening Azure PowerShell console window, see [Install and configure Azure PowerShell][powershell-install-configure]. 2. Set the first two variables in the following commands, and then run the commands: $subscriptionName = "<SubscriptionName>" $clusterName = "<HDInsightClusterName>" $querystring = "SELECT * FROM hivesampletable WHERE Country='United Kingdom';" The $querystring is the HiveQL query. 3. Run the following commands to select Azure subscription and the cluster to run the Hive job: Select-AzureSubscription -SubscriptionName $subscriptionName 4. Submit the hive job: Use-AzureHDInsightCluster $clusterName Invoke-Hive -Query $queryString You can use the -File switch to specify a HiveQL script file on HDFS. For more information about Hive, see [Use Hive with HDInsight][hdinsight-use-hive]. ##<a id="sqoop-powershell"></a>Submit Sqoop jobs using PowerShell See [Use Sqoop with HDInsight][hdinsight-use-sqoop]. ##<a id="mapreduce-sdk"></a> Submit MapReduce jobs using HDInsight .NET SDK The HDInsight .NET SDK provides .NET client libraries that makes it easier to work with HDInsight clusters from .NET. HDInsight clusters come with a jar file, located at *\example\jars\hadoop-examples.jar*, which contains several MapReduce examples. One of the examples is for counting word frequencies in source files. In this session, you will learn how to create a .NET application to run the word count sample. For more information on developing and running MapReduce jobs, see [Use MapReduce with HDInsight][hdinsight-use-mapreduce]. The following procedures are needed to provision an HDInsight cluster using the SDK: - Install the HDInsight .NET SDK - Create a console application - Run the application **To install the HDInsight .NET SDK** You can install latest published build of the SDK from [NuGet](http://nuget.codeplex.com/wikipage?title=Getting%20Started). The instructions will be shown in the next procedure. **To create a Visual Studio console application** 1. Open Visual Studio 2012. 2. From the File menu, click **New**, and then click **Project**. 3. From New Project, type or select the following values: <table style="border-color: #c6c6c6; border-width: 2px; border-style: solid; border-collapse: collapse;"> <tr> <th style="border-color: #c6c6c6; border-width: 2px; border-style: solid; border-collapse: collapse; width:90px; padding-left:5px; padding-right:5px;">Property</th> <th style="border-color: #c6c6c6; border-width: 2px; border-style: solid; border-collapse: collapse; width:90px; padding-left:5px; padding-right:5px;">Value</th></tr> <tr> <td style="border-color: #c6c6c6; border-width: 2px; border-style: solid; border-collapse: collapse; padding-left:5px;">Category</td> <td style="border-color: #c6c6c6; border-width: 2px; border-style: solid; border-collapse: collapse; padding-left:5px; padding-right:5px;">Templates/Visual C#/Windows</td></tr> <tr> <td style="border-color: #c6c6c6; border-width: 2px; border-style: solid; border-collapse: collapse; padding-left:5px;">Template</td> <td style="border-color: #c6c6c6; border-width: 2px; border-style: solid; border-collapse: collapse; padding-left:5px;">Console Application</td></tr> <tr> <td style="border-color: #c6c6c6; border-width: 2px; border-style: solid; border-collapse: collapse; padding-left:5px;">Name</td> <td style="border-color: #c6c6c6; border-width: 2px; border-style: solid; border-collapse: collapse; padding-left:5px;">SubmitMapReduceJob</td></tr> </table> 4. Click **OK** to create the project. 5. From the **Tools** menu, click **Library Package Manager**, click **Package Manager Console**. 6. Run the following commands in the console to install the packages. Install-Package Microsoft.WindowsAzure.Management.HDInsight This command adds .NET libraries and references to them to the current Visual Studio project. The version shall be 0.11.0.1 or newer. 7. From Solution Explorer, double-click **Program.cs** to open it. 8. Add the following using statements to the top of the file: using System.IO; using System.Threading; using System.Security.Cryptography.X509Certificates; using Microsoft.WindowsAzure.Storage; using Microsoft.WindowsAzure.Storage.Blob; using Microsoft.WindowsAzure.Management.HDInsight; using Microsoft.Hadoop.Client; 9. Add the following function definition to the class. This function is used to wait for a Hadoop job to complete. private static void WaitForJobCompletion(JobCreationResults jobResults, IJobSubmissionClient client) { JobDetails jobInProgress = client.GetJob(jobResults.JobId); while (jobInProgress.StatusCode != JobStatusCode.Completed && jobInProgress.StatusCode != JobStatusCode.Failed) { jobInProgress = client.GetJob(jobInProgress.JobId); Thread.Sleep(TimeSpan.FromSeconds(10)); } } 10. In the Main() function, copy and paste the following code: // Set the variables string subscriptionID = "<Azure subscription ID>"; string certFriendlyName = "<certificate friendly name>"; string clusterName = "<HDInsight cluster name>"; string storageAccountName = "<Azure storage account name>"; string storageAccountKey = "<Azure storage account key>"; string containerName = "<Blob container name>"; These are all of the variables you need to set for the program. You can get the Azure Subscription name from the [Azure Management Portal][azure-management-portal]. For information on certificate, see [Create and Upload a Management Certificate for Azure][azure-certificate]. An easy way to configure the certificate is to run *Get-AzurePublishSettingsFile* and *Import-AzurePublishSettingsFile* PowerShell cmdlets. They will create and upload the management certificate automatically. After you run the PowerShell cmdlets, you can open *certmgr.msc* from the workstation, and find the certificate by expanding *Personal/Certificates*. The certificate created by the PowerShell cmdlets has *Azure Tools* for both the *Issued To* and the *Issued By* fields. The Azure Storage account name is the account you specify when you provision the HDInsight cluster. The default container name is the same as the HDInsight cluster name. 11. In the Main() function, append the following code to define the MapReduce job: // Define the MapReduce job MapReduceJobCreateParameters mrJobDefinition = new MapReduceJobCreateParameters() { JarFile = "wasb:///example/jars/hadoop-examples.jar", ClassName = "wordcount" }; mrJobDefinition.Arguments.Add("wasb:///example/data/gutenberg/davinci.txt"); mrJobDefinition.Arguments.Add("wasb:///example/data/WordCountOutput"); There are two arguments. The first one is the source file name, and the second is the output file path. For more information of the wasb prefix, see [Use Azure Blob storage with HDInsight][hdinsight-storage]. 12. In the Main() function, append the following code to create a JobSubmissionCertificateCredential object: // Get the certificate object from certificate store using the friendly name to identify it X509Store store = new X509Store(); store.Open(OpenFlags.ReadOnly); X509Certificate2 cert = store.Certificates.Cast<X509Certificate2>().First(item => item.FriendlyName == certFriendlyName); JobSubmissionCertificateCredential creds = new JobSubmissionCertificateCredential(new Guid(subscriptionID), cert, clusterName); 13. In the Main() function, append the following code to run the job and wait the job to complete: // Create a hadoop client to connect to HDInsight var jobClient = JobSubmissionClientFactory.Connect(creds); // Run the MapReduce job JobCreationResults mrJobResults = jobClient.CreateMapReduceJob(mrJobDefinition); // Wait for the job to complete WaitForJobCompletion(mrJobResults, jobClient); 14. In the Main() function, append the following code to print the MapReduce job output: // Print the MapReduce job output Stream stream = new MemoryStream(); CloudStorageAccount storageAccount = CloudStorageAccount.Parse("DefaultEndpointsProtocol=https;AccountName=" + storageAccountName + ";AccountKey=" + storageAccountKey); CloudBlobClient blobClient = storageAccount.CreateCloudBlobClient(); CloudBlobContainer blobContainer = blobClient.GetContainerReference(containerName); CloudBlockBlob blockBlob = blobContainer.GetBlockBlobReference("example/data/WordCountOutput/part-r-00000"); blockBlob.DownloadToStream(stream); stream.Position = 0; StreamReader reader = new StreamReader(stream); Console.WriteLine(reader.ReadToEnd()); Console.WriteLine("Press ENTER to continue."); Console.ReadLine(); The output folder is specified when you define the MapReduce job. The default file name is *part-r-00000*. **To run the application** While the application is open in Visual Studio, press **F5** to run the application. A console window should open and display the status of the application and the application output. ##<a id="streaming-sdk"></a> Submit Hadoop streaming jobs using HDInsight .NET SDK HDInsight clusters come with a word counting Hadoop stream program developed in C#. The mapper program is */example/apps/cat.exe*, and the reduce program is */example/apps/wc.exe*. In this session, you will learn how to create a .NET application to run the word counting sample. For the details of creating a .Net application for submitting MapReduce jobs, see [Submit MapReduce jobs using HDInsight .NET SDK](#mapreduce-sdk). For more information on developing and deploying Hadoop streaming jobs, see [Develop C# Hadoop streaming programs for HDInsight][hdinsight-develop-streaming-jobs]. using System; using System.Collections.Generic; using System.Linq; using System.Text; using System.Threading.Tasks; using System.IO; using System.Threading; using System.Security.Cryptography.X509Certificates; using Microsoft.WindowsAzure.Management.HDInsight; using Microsoft.Hadoop.Client; namespace SubmitStreamingJob { class Program { static void Main(string[] args) { // Set the variables string subscriptionID = "<Azure subscription ID>"; string certFriendlyName = "<certificate friendly name>"; string clusterName = "<HDInsight cluster name>"; string statusFolderName = @"/tutorials/wordcountstreaming/status"; // Define the Hadoop streaming MapReduce job StreamingMapReduceJobCreateParameters myJobDefinition = new StreamingMapReduceJobCreateParameters() { JobName = "my word counting job", StatusFolder = statusFolderName, Input = "/example/data/gutenberg/davinci.txt", Output = "/tutorials/wordcountstreaming/output", Reducer = "wc.exe", Mapper = "cat.exe" }; myJobDefinition.Files.Add("/example/apps/wc.exe"); myJobDefinition.Files.Add("/example/apps/cat.exe"); // Get the certificate object from certificate store using the friendly name to identify it X509Store store = new X509Store(); store.Open(OpenFlags.ReadOnly); X509Certificate2 cert = store.Certificates.Cast<X509Certificate2>().First(item => item.FriendlyName == certFriendlyName); JobSubmissionCertificateCredential creds = new JobSubmissionCertificateCredential(new Guid(subscriptionID), cert, clusterName); // Create a hadoop client to connect to HDInsight var jobClient = JobSubmissionClientFactory.Connect(creds); // Run the MapReduce job Console.WriteLine("----- Submit the Hadoop streaming job ..."); JobCreationResults mrJobResults = jobClient.CreateStreamingJob(myJobDefinition); // Wait for the job to complete Console.WriteLine("----- Wait for the Hadoop streaming job to complete ..."); WaitForJobCompletion(mrJobResults, jobClient); // Display the error log Console.WriteLine("----- The hadoop streaming job error log."); using (Stream stream = jobClient.GetJobErrorLogs(mrJobResults.JobId)) { var reader = new StreamReader(stream); Console.WriteLine(reader.ReadToEnd()); } // Display the output log Console.WriteLine("----- The hadoop streaming job output log."); using (Stream stream = jobClient.GetJobOutput(mrJobResults.JobId)) { var reader = new StreamReader(stream); Console.WriteLine(reader.ReadToEnd()); } Console.WriteLine("----- Press ENTER to continue."); Console.ReadLine(); } private static void WaitForJobCompletion(JobCreationResults jobResults, IJobSubmissionClient client) { JobDetails jobInProgress = client.GetJob(jobResults.JobId); while (jobInProgress.StatusCode != JobStatusCode.Completed && jobInProgress.StatusCode != JobStatusCode.Failed) { jobInProgress = client.GetJob(jobInProgress.JobId); Thread.Sleep(TimeSpan.FromSeconds(10)); } } } } ##<a id="hive-sdk"></a> Submit Hive jobs using HDInsight .NET SDK HDInsight clusters come with a sample Hive table called *hivesampletable*. In this session, you will create a .NET application to run a Hive job for listing the Hive tables created on HDInsight cluster. For a more information on using Hive, see [Use Hive with HDInsight][hdinsight-use-hive]. The following procedures are needed to provision an HDInsight cluster using the SDK: - Install the HDInsight .NET SDK - Create a console application - Run the application **To install the HDInsight .NET SDK** You can install latest published build of the SDK from [NuGet](http://nuget.codeplex.com/wikipage?title=Getting%20Started). The instructions will be shown in the next procedure. **To create a Visual Studio console application** 1. Open Visual Studio 2012. 2. From the File menu, click **New**, and then click **Project**. 3. From New Project, type or select the following values: <table style="border-color: #c6c6c6; border-width: 2px; border-style: solid; border-collapse: collapse;"> <tr> <th style="border-color: #c6c6c6; border-width: 2px; border-style: solid; border-collapse: collapse; width:90px; padding-left:5px; padding-right:5px;">Property</th> <th style="border-color: #c6c6c6; border-width: 2px; border-style: solid; border-collapse: collapse; width:90px; padding-left:5px; padding-right:5px;">Value</th></tr> <tr> <td style="border-color: #c6c6c6; border-width: 2px; border-style: solid; border-collapse: collapse; padding-left:5px;">Category</td> <td style="border-color: #c6c6c6; border-width: 2px; border-style: solid; border-collapse: collapse; padding-left:5px; padding-right:5px;">Templates/Visual C#/Windows</td></tr> <tr> <td style="border-color: #c6c6c6; border-width: 2px; border-style: solid; border-collapse: collapse; padding-left:5px;">Template</td> <td style="border-color: #c6c6c6; border-width: 2px; border-style: solid; border-collapse: collapse; padding-left:5px;">Console Application</td></tr> <tr> <td style="border-color: #c6c6c6; border-width: 2px; border-style: solid; border-collapse: collapse; padding-left:5px;">Name</td> <td style="border-color: #c6c6c6; border-width: 2px; border-style: solid; border-collapse: collapse; padding-left:5px;">SubmitHiveJob</td></tr> </table> 4. Click **OK** to create the project. 5. From the **Tools** menu, click **Library Package Manager**, click **Package Manager Console**. 6. Run the following commands in the console to install the packages. Install-Package Microsoft.WindowsAzure.Management.HDInsight This command adds .NET libraries and references to them to the current Visual Studio project. 7. From Solution Explorer, double-click **Program.cs** to open it. 8. Add the following using statements to the top of the file: using System.IO; using System.Threading; using System.Security.Cryptography.X509Certificates; using Microsoft.WindowsAzure.Management.HDInsight; using Microsoft.Hadoop.Client; 9. Add the following function definition to the class. This function is used to wait for a Hadoop job to complete. private static void WaitForJobCompletion(JobCreationResults jobResults, IJobSubmissionClient client) { JobDetails jobInProgress = client.GetJob(jobResults.JobId); while (jobInProgress.StatusCode != JobStatusCode.Completed && jobInProgress.StatusCode != JobStatusCode.Failed) { jobInProgress = client.GetJob(jobInProgress.JobId); Thread.Sleep(TimeSpan.FromSeconds(10)); } } 10. In the Main() function, copy and paste the following code: // Set the variables string subscriptionID = "<Azure subscription ID>"; string clusterName = "<HDInsight cluster name>"; string certFriendlyName = "<certificate friendly name>"; These are all of the variables you need to set for the program. You can get the Azure Subscription ID from you system administrator. For information on certificate, see [Create and Upload a Management Certificate for Azure][azure-certificate]. An easy way to configure the certificate is to run *Get-AzurePublishSettingsFile* and *Import-AzurePublishSettingsFile* PowerShell cmdlets. They will create and upload the management certificate automatically. After you run the PowerShell cmdlets, you can open *certmgr.msc* from the workstation, and find the certificate by expanding *Personal/Certificates*. The certificate created by the PowerShell cmdlets has *Azure Tools* for both the *Issued To* and the *Issued By* fields. 11. In the Main() function, append the following code to define the Hive job: // define the Hive job HiveJobCreateParameters hiveJobDefinition = new HiveJobCreateParameters() { JobName = "show tables job", StatusFolder = "/ShowTableStatusFolder", Query = "show tables;" }; You can also use the File parameter to specify a HiveQL script file on HDFS. For example // define the Hive job HiveJobCreateParameters hiveJobDefinition = new HiveJobCreateParameters() { JobName = "show tables job", StatusFolder = "/ShowTableStatusFolder", File = "/user/admin/showtables.hql" }; 12. In the Main() function, append the following code to create a JobSubmissionCertificateCredential object: // Get the certificate object from certificate store using the friendly name to identify it X509Store store = new X509Store(); store.Open(OpenFlags.ReadOnly); X509Certificate2 cert = store.Certificates.Cast<X509Certificate2>().First(item => item.FriendlyName == certFriendlyName); JobSubmissionCertificateCredential creds = new JobSubmissionCertificateCredential(new Guid(subscriptionID), cert, clusterName); 13. In the Main() function, append the following code to run the job and wait the job to complete: // Submit the Hive job var jobClient = JobSubmissionClientFactory.Connect(creds); JobCreationResults jobResults = jobClient.CreateHiveJob(hiveJobDefinition); // Wait for the job to complete WaitForJobCompletion(jobResults, jobClient); 14. In the Main() function, append the following code to print the Hive job output: // Print the Hive job output System.IO.Stream stream = jobClient.GetJobOutput(jobResults.JobId); StreamReader reader = new StreamReader(stream); Console.WriteLine(reader.ReadToEnd()); Console.WriteLine("Press ENTER to continue."); Console.ReadLine(); **To run the application** While the application is open in Visual Studio, press **F5** to run the application. A console window should open and display the status of the application. The output shall be: hivesampletable ##<a id="nextsteps"></a> Next steps In this article, you have learned several ways to provision an HDInsight cluster. To learn more, see the following articles: * [Get started with Azure HDInsight][hdinsight-get-started] * [Provision HDInsight clusters][hdinsight-provision] * [Administer HDInsight using PowerShell][hdinsight-admin-powershell] * [HDInsight Cmdlet Reference Documentation][hdinsight-powershell-reference] * [Use Hive with HDInsight][hdinsight-use-hive] * [Use Pig with HDInsight][hdinsight-use-pig] [azure-certificate]: http://msdn.microsoft.com/en-us/library/windowsazure/gg551722.aspx [azure-management-portal]: http://manage.windowsazure.com/ [hdinsight-use-sqoop]: ../hdinsight-use-sqoop/ [hdinsight-provision]: ../hdinsight-provision-clusters/ [hdinsight-use-mapreduce]: ../hdinsight-use-mapreduce/ [hdinsight-use-hive]: ../hdinsight-use-hive/ [hdinsight-use-pig]: ../hdinsight-use-pig/ [hdinsight-get-started]: ../hdinsight-get-started/ [hdinsight-storage]: ../hdinsight-use-blob-storage/ [hdinsight-admin-powershell]: ../hdinsight-administer-use-powershell/ [hdinsight-develop-streaming-jobs]: ../hdinsight-hadoop-develop-deploy-streaming-jobs/ [hdinsight-powershell-reference]: http://msdn.microsoft.com/en-us/library/windowsazure/dn479228.aspx [Powershell-install-configure]: ../install-configure-powershell/ [image-hdi-gettingstarted-runmrjob]: ./media/hdinsight-submit-hadoop-jobs-programmatically/HDI.GettingStarted.RunMRJob.png [image-hdi-gettingstarted-mrjoboutput]: ./media/hdinsight-submit-hadoop-jobs-programmatically/HDI.GettingStarted.MRJobOutput.png [apache-hive]: http://hive.apache.org/
43.077574
626
0.722284
eng_Latn
0.849685
21a0229ca0d5b46c7d0e90da69deb262e9951fe9
2,827
md
Markdown
README.md
ivxxi/pitch2
02a372b8c412850cb4a0d28740cc6a71bd4988f8
[ "MIT" ]
null
null
null
README.md
ivxxi/pitch2
02a372b8c412850cb4a0d28740cc6a71bd4988f8
[ "MIT" ]
null
null
null
README.md
ivxxi/pitch2
02a372b8c412850cb4a0d28740cc6a71bd4988f8
[ "MIT" ]
null
null
null
# Pitch ## Description Pitch is an application allows users to submit their one minute pitches and other users will vote on them and leave comments to give their feedback. The pitches are organized by category. ## Author * [**Crystal Alice**] ## Features As a user of the web application you will be able to: 1. See all submitted pitches arranged in categories 2. Create an account 3. Log in 4. Add a pitch based on category 5. Upvote or downvote a pitch 6. Comment on a pitch 7. See comments posted on each individual pitch 8. Edit your profile i.e will be able to add a short bio about yourself and a profile picture ## Specifications | Behavior | Input | Output | | ------------------- | ----------------------------- | ----------------------------- | | View Product Pitches | Click on any category | Taken to the clicked category | Click on `Click Here To Post A Pitch` | Redirected to the login page | Signs In/ Signs Up | | Click on `Click Here To Post A Pitch` | If logged in, display form to add a pitch | Redirected to the home page | | Click upvote/ downvote button | Redirects to home page | Upvote/ downvote count changes | Click add comment button | Redirects to the comment page | Displays a comment form | Click on Sign Out | Redirects to the home page | Signs user out | | Click on profile | Redirects to the profile page | User adds bio and profile picture | ## Getting started ### Prerequisites * python3.6 * virtual environment * pip ### Cloning * In your terminal: $ git clone $ cd pitch ## Running the Application * Install virtual environment using `$ python3.6 -m venv --without-pip virtual` * Activate virtual environment using `$ source virtual/bin/activate` * Download pip in our environment using `$ curl https://bootstrap.pypa.io/get-pip.py | python` * Install all the dependencies from the requirements.txt file by running `pip3 install -r requirements.txt` * Create a `start.sh` file in the root of the folder and add the following code: export MAIL_USERNAME=<your-email-address> export MAIL_PASSWORD=<your-email-password> export SECRET_KEY=<your-secret-key> * To run the application, in your terminal: $ chmod a+x start.sh $ ./start.sh ## Testing the Application * To run the tests for the class file: $ python3.6 manage.py server ## Technologies Used * Python3.6 * Flask * HTML * Bootstrap This application is developed using [Python3.6](https://www.python.org/doc/), [Flask](http://flask.palletsprojects.com/en/1.1.x/), [HTML](https://getbootstrap.com/) and [Bootstrap](https://getbootstrap.com/) ## Support and Team Crystal Alice ### License * LICENSED UNDER [![License: MIT](https://img.shields.io/badge/License-MIT-yellow.svg)](license/MIT)
34.060241
242
0.687301
eng_Latn
0.949978
21a0824a728a19a9278adba800929cec57a3c7a4
334
md
Markdown
doc/adr/0002-implement-as-Java.md
Leviter/adr-j
5ae1f67d0fc2355e944fc5bcf550023542013a0f
[ "MIT" ]
29
2017-10-27T12:10:45.000Z
2022-03-31T12:56:21.000Z
doc/adr/0002-implement-as-Java.md
Leviter/adr-j
5ae1f67d0fc2355e944fc5bcf550023542013a0f
[ "MIT" ]
35
2017-07-21T13:40:48.000Z
2022-03-31T13:17:58.000Z
doc/adr/0002-implement-as-Java.md
tkleiber/adr-j
ae8276350691e78e3946158f0eaeb3cb4cafdb04
[ "MIT" ]
10
2017-10-27T12:27:07.000Z
2022-03-31T10:27:18.000Z
# 2. Implement as Java Date: 21.07.2017 ## Status Accepted ## Context ADRS need to be created on a wide variety of operating systems. ## Decision The tool is written in Java 8 and uses the operating system independence of the Java platform. ## Consequences As Java is not a scripting language, the code will be more complex.
16.7
94
0.748503
eng_Latn
0.999382
21a2328e80da6421db0aacd7f0f270d84016351d
1,129
md
Markdown
Umbraco-Uno/Working-with-Forms/Umbraco-Forms-widget/index.md
sumitkharche/UmbracoDocs
de839c20818d4ed13776ecf17ed0145b6d7886a6
[ "MIT" ]
201
2015-08-24T22:15:42.000Z
2022-02-19T19:19:54.000Z
Umbraco-Uno/Working-with-Forms/Umbraco-Forms-widget/index.md
sumitkharche/UmbracoDocs
de839c20818d4ed13776ecf17ed0145b6d7886a6
[ "MIT" ]
2,451
2015-08-21T14:22:12.000Z
2022-03-31T19:54:23.000Z
Umbraco-Uno/Working-with-Forms/Umbraco-Forms-widget/index.md
sumitkharche/UmbracoDocs
de839c20818d4ed13776ecf17ed0145b6d7886a6
[ "MIT" ]
725
2015-08-26T00:54:49.000Z
2022-03-31T08:22:11.000Z
--- versionFrom: 8.0.0 --- # The Umbraco Forms widget There are multiple ways of adding a form to your content pages in Umbraco Uno. In this guide, you will learn of two ways to add forms to a site through the use of widgets. Here is an example of what forms could look like on the front end. ![Front end example](images/Frontend-example-forms.png) The first option is to add the widget called Umbraco Forms. In this widget, you will get the option to select which form you want to add. You can read about this widget's settings in the article about the [Umbraco Forms widget](../../Widgets/Umbraco-Form). ![Regular Umbraco Forms](images/Umbraco-forms.png) The second option is to add the form through a Grid widget. The setup here is slightly different than with the Umbraco Forms widget. However, the way you add the form is the same as the first option. Go to the section of the widget that says ***Form*** and then click "Add" to select the form you want. ![Grid Umbraco Forms](images/Grid-form-filled.png) Now that you know how to add forms the last thing you need to know is [how to check the replies](../Replies).
53.761905
302
0.757307
eng_Latn
0.999835
21a31b7e2bfd800897be168784583f444934e8f9
2,869
md
Markdown
articles/virtual-desktop/data-locations.md
flarocca/azure-docs.es-es
8d69748012641d57ddb2b81a3e1c2d079703ed8d
[ "CC-BY-4.0", "MIT" ]
null
null
null
articles/virtual-desktop/data-locations.md
flarocca/azure-docs.es-es
8d69748012641d57ddb2b81a3e1c2d079703ed8d
[ "CC-BY-4.0", "MIT" ]
null
null
null
articles/virtual-desktop/data-locations.md
flarocca/azure-docs.es-es
8d69748012641d57ddb2b81a3e1c2d079703ed8d
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: 'Ubicaciones de datos para Windows Virtual Desktop: Azure' description: Una breve introducción a las ubicaciones en las que se almacenan los datos y metadatos de Windows Virtual Desktop. author: Heidilohr ms.topic: conceptual ms.date: 04/30/2020 ms.author: helohr manager: lizross ms.openlocfilehash: e626b7e729e394b1012848904f5ce12279c3ef24 ms.sourcegitcommit: 98854e3bd1ab04ce42816cae1892ed0caeedf461 ms.translationtype: HT ms.contentlocale: es-ES ms.lasthandoff: 08/07/2020 ms.locfileid: "88010079" --- # <a name="data-and-metadata-locations-for-windows-virtual-desktop"></a>Ubicaciones de datos y metadatos para Windows Virtual Desktop >[!IMPORTANT] >Este contenido se aplica a Windows Virtual Desktop con objetos de Windows Virtual Desktop para Azure Resource Manager. Si usa Windows Virtual Desktop (clásico) sin objetos para Azure Resource Manager, consulte [este artículo](./virtual-desktop-fall-2019/data-locations-2019.md). Windows Virtual Desktop está disponible actualmente para todas las ubicaciones geográficas. Los administradores pueden elegir la ubicación para almacenar los datos de usuario cuando creen las máquinas virtuales del grupo de hosts y los servicios asociados, como los servidores de archivos. Obtenga más información sobre las ubicaciones geográficas de Azure en el [mapa de centros de datos de Azure](https://azuredatacentermap.azurewebsites.net/). >[!NOTE] >Microsoft no controla ni limita las regiones en las que usted o sus usuarios pueden acceder a los datos específicos de la aplicación o del usuario. >[!IMPORTANT] >Windows Virtual Desktop almacena información de metadatos globales como nombres de inquilinos, nombres de grupos de hosts, nombres de grupos de aplicaciones y nombres principales de usuario en un centro de datos. Cada vez que un cliente crea un objeto de servicio, debe especificar una ubicación para él. La ubicación que especifique determina el lugar en que se almacenarán los metadatos del objeto. El cliente elegirá una región de Azure y los metadatos se almacenarán en la zona geográfica relacionada. Para obtener una lista de todas las regiones de Azure y las zonas geográficas relacionadas, consulte [Zonas geográficas de Azure](https://azure.microsoft.com/global-infrastructure/geographies/). En este momento, solo se admite el almacenamiento de metadatos en la zona geográfica de Azure Estados Unidos (EE. UU.). Los metadatos almacenados se cifran en reposo y los reflejos con redundancia geográfica se mantienen en la zona geográfica. Todos los datos del cliente, como la configuración de la aplicación y los datos de usuario, residen en la ubicación que el cliente elige y no los administra el servicio. A medida que crezca el servicio, habrá más zonas geográficas disponibles. Los metadatos del servicio se replican dentro de la zona geográfica con fines de recuperación ante desastres.
92.548387
701
0.817009
spa_Latn
0.994769
21a32912fc91964216b03b52f3dc52ed71620966
3,500
md
Markdown
node_modules/vfile-message/readme.md
Wdifulvio523/personal-portfolio
747bd5ebefb5d2c4b61f1631c9d53f32a609650c
[ "MIT" ]
36
2018-01-01T17:37:28.000Z
2020-02-03T09:15:27.000Z
assets/manage/node_modules/vfile-message/readme.md
nuridincs/bimbel_gama
be2d0d83d9579c52cda9649777552db9983e3d7b
[ "MIT" ]
9
2021-07-24T23:14:16.000Z
2022-03-25T18:40:03.000Z
assets/manage/node_modules/vfile-message/readme.md
nuridincs/bimbel_gama
be2d0d83d9579c52cda9649777552db9983e3d7b
[ "MIT" ]
26
2018-01-01T17:37:30.000Z
2021-08-18T09:58:53.000Z
# vfile-message [![Build Status][travis-badge]][travis] [![Coverage Status][codecov-badge]][codecov] Create [vfile][] messages. ## Installation [npm][]: ```bash npm install vfile-message ``` ## Usage ```js var VMessage = require('vfile-message'); var message = new VMessage('`braavo` is misspelt; did you mean `bravo`?', {line: 1, column: 8}, 'spell:typo'); console.log(message); ``` Yields: ```js { [1:8: `braavo` is misspelt; did you mean `bravo`?] reason: '`braavo` is misspelt; did you mean `bravo`?', fatal: null, line: 1, column: 8, location: { start: { line: 1, column: 8 }, end: { line: null, column: null } }, source: 'spell', ruleId: 'typo' } ``` ## API ### `VMessage(reason[, position][, origin])` Constructor of a message for `reason` at `position` from `origin`. When an error is passed in as `reason`, copies the stack. ##### Parameters ###### `reason` Reason for message (`string` or `Error`). Uses the stack and message of the error if given. ###### `position` Place at which the message occurred in a file ([`Node`][node], [`Position`][position], or [`Point`][point], optional). ###### `origin` Place in code the message originates from (`string`, optional). Can either be the [`ruleId`][ruleid] (`'rule'`), or a string with both a [`source`][source] and a [`ruleId`][ruleid] delimited with a colon (`'source:rule'`). ##### Extends [`Error`][error]. ##### Returns An instance of itself. ##### Properties ###### `reason` Reason for message (`string`). ###### `fatal` If `true`, marks associated file as no longer processable (`boolean?`). If `false`, necessitates a (potential) change. The value can also be `null` or `undefined`. ###### `line` Starting line of error (`number?`). ###### `column` Starting column of error (`number?`). ###### `location` Full range information, when available ([`Position`][position]). Has `start` and `end` properties, both set to an object with `line` and `column`, set to `number?`. ###### `source` Namespace of warning (`string?`). ###### `ruleId` Category of message (`string?`). ###### `stack` Stack of message (`string?`). ##### Custom properties It’s OK to store custom data directly on the `VFileMessage`, some of those are handled by [utilities][util]. ###### `file` You may add a `file` property with a path of a file (used throughout the [**VFile**][vfile] ecosystem). ###### `note` You may add a `note` property with a long form description of the message (supported by [`vfile-reporter`][reporter]). ###### `url` You may add a `url` property with a link to documentation for the message. ## License [MIT][license] © [Titus Wormer][author] <!-- Definitions --> [travis-badge]: https://img.shields.io/travis/vfile/vfile-message.svg [travis]: https://travis-ci.org/vfile/vfile-message [codecov-badge]: https://img.shields.io/codecov/c/github/vfile/vfile-message.svg [codecov]: https://codecov.io/github/vfile/vfile-message [npm]: https://docs.npmjs.com/cli/install [license]: LICENSE [author]: http://wooorm.com [error]: https://developer.mozilla.org/en-US/docs/Web/JavaScript/Reference/Global_Objects/Error [node]: https://github.com/syntax-tree/unist#node [position]: https://github.com/syntax-tree/unist#position [point]: https://github.com/syntax-tree/unist#point [vfile]: https://github.com/vfile/vfile [util]: https://github.com/vfile/vfile#utilities [reporter]: https://github.com/vfile/vfile-reporter [ruleid]: #ruleid [source]: #source
20.958084
110
0.665429
eng_Latn
0.759563
21a3ebf89e86314ff08b8fd2373784c02d11c48c
3,513
md
Markdown
README.md
shelepenok/urlshortener
30471ee185cf82cf2cf64de42f8257b556bbaf3e
[ "MIT" ]
null
null
null
README.md
shelepenok/urlshortener
30471ee185cf82cf2cf64de42f8257b556bbaf3e
[ "MIT" ]
null
null
null
README.md
shelepenok/urlshortener
30471ee185cf82cf2cf64de42f8257b556bbaf3e
[ "MIT" ]
null
null
null
# urlshortener A simple library to shorten URL's with goo.gl written in Golang. ## Basic usage ```go package main import ( "fmt" "github.com/alxshelepenok/urlshortener" ) func main() { urlshortener.Setup(key) short, err := urlshortener.Shorten(long) if err != nil { panic(err) } fmt.Println(short) } ``` ## Getting Api Key **This is an excerpt from the official [documentation](https://developers.google.com/url-shortener/v1/getting_started) .** Requests to the Google URL Shortener API for public data must be accompanied by an identifier, which can be an [API key](https://developers.google.com/console/help/generating-dev-keys) or an [access token](https://developers.google.com/accounts/docs/OAuth2). To acquire an API key: + Go to the [Google Developers Console](https://console.developers.google.com/). + Select a project, or create a new one. + In the sidebar on the left, expand **APIs & auth**. Next, click APIs. In the list of **APIs**, make sure the status is **ON** for the Google URL Shortener API. + In the sidebar on the left, select **Credentials**. + This API supports two types of credentials. Create whichever credentials are appropriate for your project: + **OAuth:** Your application must send an OAuth 2.0 token with any request that accesses private user data. Your application sends a client ID and, possibly, a client secret to obtain a token. You can generate OAuth 2.0 credentials for web applications, service accounts, or installed applications. </br> To create an OAuth 2.0 token, click **Create new Client ID**, provide the required information where requested, and click **Create Client ID**. + **Public API access:** A request that does not provide an OAuth 2.0 token must send an API key. The key identifies your project and provides API access, quota, and reports. </br> To create an API key, click **Create new Key** and select the appropriate key type. Enter the additional information required for that key type and click **Create**. To keep your API keys secure, follow [the best practices for securely using API](https://developers.google.com/console/help/api-key-best-practices) keys. After you have an API key, your application can append the query parameter `key=yourAPIKey` to all request URLs. The API key is safe for embedding in URLs; it doesn't need any encoding. ## License The MIT license. Copyright (c) 2016-2022 Alexander Shelepenok Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
41.821429
120
0.752064
eng_Latn
0.928027
21a3ed51b7885155cf5563ddbe5b651aafe4152c
2,961
md
Markdown
projects/fl-framework-angular/README.md
neut-kubota/fl-framework-app
3df637b6328fd01149f43580c173a13d786a2e72
[ "MIT" ]
null
null
null
projects/fl-framework-angular/README.md
neut-kubota/fl-framework-app
3df637b6328fd01149f43580c173a13d786a2e72
[ "MIT" ]
null
null
null
projects/fl-framework-angular/README.md
neut-kubota/fl-framework-app
3df637b6328fd01149f43580c173a13d786a2e72
[ "MIT" ]
null
null
null
# fl-framework-angular ## how to use ### install ```npm npm install fl-framework-angular @angular/flex-layout ``` ### app.module.ts ```TypeScript import { NgModule, Injector } from '@angular/core'; import { FlFrameworkModule, FrameworkModule } from 'fl-framework-angular'; @NgModule({ declarations: [ ... ], imports: [ ... FlFrameworkModule ], providers: [ { //custom call setting provide: 'flRequestEndPoint', useValue: { host: '', //Call from the default caller. contextName: '_api', //server addr(URI) + / + contextName + / <serviceName> + / <methodName> + / appVersion: '0.0.1', //Add version number to http header. versionName: 'version' //http header name. } } ], bootstrap: [...] }) export class AppModule extends FrameworkModule { constructor(injector: Injector){ super(injector); } } ``` ### app.component.ts ```TypeScript import { Component } from '@angular/core'; import { FlComponentBase, isEmpty, FlServiceName } from 'fl-framework-angular'; @Component({ selector: 'app-root', templateUrl: './app.component.html', styleUrls: ['./app.component.scss'] }) @FlServiceName('Test') export class AppComponent extends FlComponentBase { /** * abstract initializeView -> ngOnInit **/ initializeView(): void { } /** * abstract createCompleteView -> ngAfterViewInit + delay 50ms **/ createCompleteView(): void { /** * call rest api service * methodName: init, search, regist * serviceName: Specify the service name with FlServiceName decorator * * default rest api address * /_api/[serviceName]/[methodName] **/ this.init({...}).subscribe((result)=>{ }); this.search({...}).subscribe((result)=>{ }); this.regist({...}).subscribe((result)=>{ }); /** * custom method */ this.callService([methodNameString],{...}) .subscribe((result)=>{ }); /** * custom service and method */ this.callCustomService([serviceNameString],[methodNameString],{...}) .subscribe((result)=>{ }); } title = 'framework-demo'; } ``` ### style.scss ```scss @import '~@angular/material/theming'; @import '~fl-framework-angular/theming'; // <-here // require angular material @include mat-core(); // mat-deep-purple,mat-purple,mat-indigo $primary: mat-palette($mat-blue, A200); $accent: mat-palette($mat-orange, A400); $warn: mat-palette($mat-teal, 500); $theme: mat-light-theme($primary, $accent, $warn); @include angular-material-theme($theme); //fl-framework-theme($info, $question, $complete,$warning,$error); //@include fl-framework-theme(#53c1dc,#2a7cb4,#59b764,#f3ab58,#F44336); <- default color @include fl-framework-theme(); // alert panel color //fl-framework-wait-dialog($color); //@include fl-framework-wait-dialog(#3f51b5); <- default color @include fl-framework-wait-dialog(); //wait dialog color ```
24.471074
106
0.632557
eng_Latn
0.514687
21a54275e547fe190a815845c6a6b146f290d1ad
682
md
Markdown
pages/mesosphere/dcos/cn/1.13/cli/command-reference/dcos-job/dcos-job-schedule-update/index.md
qianzhangxa/dcos-docs-site
a655e4c39fc18ecbca771df990a5afed2d0cd5b8
[ "Apache-2.0" ]
null
null
null
pages/mesosphere/dcos/cn/1.13/cli/command-reference/dcos-job/dcos-job-schedule-update/index.md
qianzhangxa/dcos-docs-site
a655e4c39fc18ecbca771df990a5afed2d0cd5b8
[ "Apache-2.0" ]
null
null
null
pages/mesosphere/dcos/cn/1.13/cli/command-reference/dcos-job/dcos-job-schedule-update/index.md
qianzhangxa/dcos-docs-site
a655e4c39fc18ecbca771df990a5afed2d0cd5b8
[ "Apache-2.0" ]
null
null
null
--- layout: layout.pug navigationTitle: dcos job schedule update title: dcos job schedule update menuWeight: 9 excerpt: 更新作业时间表 enterprise: false render: mustache model: /mesosphere/dcos/1.13/data.yml --- # 说明 `dcos job schedule update` 命令允许您更新作业的时间表。 # 使用 ```bash dcos job schedule update <job-id> <schedule-file> ``` # 选项 | 名称 | 说明 | |---------|-------------| |`-h`,`--help` | 打印用法。 | ## 位置自变量 | 名称 | 说明 | |---------|-------------| | `<job-id>` | 指定作业 ID。您可以使用 `dcos job list` 命令查看作业 ID。| | `<schedule-file>` | JSON 格式的作业时间表文件。| # 父命令 | 命令 | 说明 | |---------|-------------| | [dcos job](/mesosphere/dcos/cn/1.13/cli/command-reference/dcos-job/) | 在 DC/OS 中部署和管理作业。|
16.238095
93
0.582111
yue_Hant
0.588875
21a55436594dcba7c9d61875608c1752e254a5d1
14,892
md
Markdown
exampleSite/content/blog/o_que_instalar_mac_infra.md
gomex/novo-blog
8c3bf3ae4d2e9dd9f76f1d51e896eae81a302832
[ "MIT" ]
null
null
null
exampleSite/content/blog/o_que_instalar_mac_infra.md
gomex/novo-blog
8c3bf3ae4d2e9dd9f76f1d51e896eae81a302832
[ "MIT" ]
null
null
null
exampleSite/content/blog/o_que_instalar_mac_infra.md
gomex/novo-blog
8c3bf3ae4d2e9dd9f76f1d51e896eae81a302832
[ "MIT" ]
null
null
null
+++ title = "O que instalar em um Mac de alguém de infra" date = "2020-01-02" draft = false Categories = ["devops", "mac"] Tags = ["devops", "mac", "brew", "saas"] +++ Pra começar bem o ano, resolvi fazer um backup e formatar meu Mac, reiniciar do zero, pois o meu notebook tem apenas 128GB de espaço em disco e ultimamente tenho brigado por cada último byte livre no HD. Assim que acabei de formatar, lembrei o motivo de tanto retardo: A necessidade de reinstalar tudo que preciso do zero. Fiz uma [postagem](https://twitter.com/gomex/status/1212585682650177537) no Twitter solicitando ajuda e prometi criar um artigo, e aqui está. <!-- TOC --> - [.1. Usando o Brew](#1-usando-o-brew) - [.2. Aplicativos fora do Brew](#2-aplicativos-fora-do-brew) - [.2.1. Dash](#2-1-dash) - [.2.2. Amphetamine](#2-2-amphetamine) - [.2.3. Oh My ZSH](#2-3-oh-my-zsh) - [.3. Aplicativos genéricos para se usar no Mac (Com Brew)](#3-aplicativos-genéricos-para-se-usar-no-mac-com-brew) - [.3.1. Alfred](#3-1-alfred) - [.3.2. Firefox e Chrome](#3-2-firefox-e-chrome) - [.3.3. Grammarly](#3-3-grammarly) - [.3.4. Iterm2](#3-4-iterm2) - [.3.5. Slack](#3-5-slack) - [.3.6. Spectacle](#3-6-spectacle) - [.3.7. Telegram](#3-7-telegram) - [.3.8. Vanilla](#3-8-vanilla) - [.3.9. Whatsapp](#3-9-whatsapp) - [.3.10. GPG Suite](#3-10-gpg-suite) - [.3.11. Keybase](#3-11-keybase) - [.3.12. Transmission](#3-12-transmission) - [.3.13. Spotify](#3-13-spotify) - [.3.14. Flycut](#3-14-flycut) - [.4. Aplicativos para desenvolvedores (Com Brew)](#4-aplicativos-para-desenvolvedores-com-brew) - [.4.1. Visual Studio Code](#4-1-visual-studio-code) - [.4.2. Kubernetes-cli](#4-2-kubernetes-cli) - [.4.3. JQ](#4-3-jq) - [.4.4. Nmap](#4-4-nmap) - [.4.5. Openssl](#4-5-openssl) - [.4.6. Watch](#4-6-watch) - [.4.7. Docker](#4-7-docker) - [.4.8. Popeye](#4-8-popeye) - [.4.9. Stern](#4-9-stern) - [.4.10. Kubectx + Kubens](#4-10-kubectx--kubens) - [.4.11. Insomnia](#4-11-insomnia) - [.4.12. Docker Compose](#4-12-docker-compose) - [.5. Agradecimentos](#5-agradecimentos) <!-- /TOC --> ## .1. Usando o Brew Todo mundo que tem Mac e usa ele para trabalhar com programação/devops/automação/afins já deve conhecer o brew. A instalação a é simples e pode ser vista [aqui](https://brew.sh/index_pt-br). O que poucas pessoas sabem é sobre a existência do ```brew bundle``` (Aqui um [link](https://github.com/Homebrew/homebrew-bundle) para saber mais) basicamente ele faz o mesmo que o bundle do ruby faz. Reuni tudo que vc precisa instalar em um arquivo, que aqui é chamado de ```Brewfile``` onde você informa linha a linha, dizendo se é **brew** on **cask** para instalar os pacotes. Ao final do arquivo pronto, basta um simples comando: ```bash brew bundle install ``` Ele instala tudo pra você. Segue meu [Brewfile](https://github.com/gomex/mac-setup/blob/master/Brewfile) ## .2. Aplicativos fora do Brew Vou começar com os aplicativos que não estão no Brew, pois esses você precisará entrar no site e instalar manualmente mesmo. Seguindo a documentação de cada um deles. ### .2.1. Dash - Site: [https://kapeli.com/dash](https://kapeli.com/dash) - Valor: Grátis com limitações É uma ferramenta muito interessante de documentação, tudo a partir de um campo de busca centralizado. A melhor parte dele é a possibilidade de pesquisar offline. ### .2.2. Amphetamine - Site: [https://apps.apple.com/us/app/amphetamine/id937984704?mt=12](https://apps.apple.com/us/app/amphetamine/id937984704?mt=12) - Valor: Grátis Um aplicativo simples que pode ser usado para manter seu Mac ligado por um espaço específico de tempo. É ótimo para quem faz apresentações sem carregador e não quer que o Mac desligue no meio de apresentação, mas quer quer que esse comportamento volte depois da palestra. ### .2.3. Oh My ZSH - Site: [https://ohmyz.sh/](https://ohmyz.sh/) - Valor: Opensource É um framework para gerenciamento de seu console Zsh, onde deixa ele mais bonito e bem mais produtivo, com vários atalhos e funcionalidades novas. ## .3. Aplicativos genéricos para se usar no Mac (Com Brew) Vou passar rapidamente aplicativo por aplicativo para defender porque escolhi ele: ### .3.1. Alfred - Site: [https://www.alfredapp.com/](https://www.alfredapp.com/) - Valor: Grátis, com exceção do Powerpack Esse é uma aposta, pois todo mundo fala, mas até então não tinha testado ainda. Tenho pouco a falar dele, que tem como objetivo substituir o spotlight (Aquela busca do CMD+Barra de espaço). Dizem que ele tem mais funcionalidades, é mais rápido e afins. Eu ja tive muitos problemas com o spotlight, dai estou apostando no Alfred. ### .3.2. Firefox e Chrome - Site: [https://www.google.com/intl/pt-BR/chrome/](https://www.google.com/intl/pt-BR/chrome/) e [https://www.mozilla.org/pt-BR/firefox/new/](https://www.mozilla.org/pt-BR/firefox/new/) - Valor: Opensource (Firefox) e Grátis (Chrome) Outra opção meio padrão para maioria das pessoas, pois o uso do navegador Safari é bem baixa ainda. Eu tenho maior uso do Chrome, pois ele tem uma funcionalidade de múltiplos perfis de conta por janela de navegador. Dai eu consigo gerenciar meu perfil pessoal e profissional alternando as janelas. ### .3.3. Grammarly - Site: [https://www.grammarly.com/](https://www.grammarly.com/) - Valor: Grátis com limites (Versão paga no plano anual sai USD$ 11,66 por mês) Se você não é tão bom escrita de inglês quanto eu, e precisar escrever coisas com algumas constância, você deveria considerar um app como esse. Eu adoro o Grammarly, pois ele não verificar apenas erro de ortografia, ele verifica concordância e outros aspectos que um corretor comum não dá conta. Ele faz sugestões quando você repete demais a palavra e afins. Funciona bem. Eu aconselho. ### .3.4. Iterm2 - Site: [https://iterm2.com/](https://iterm2.com/) - Valor: Opensource Aqui temos outra unanimidade, pois a maioria esmagadora das pessoas que conhecem o Iterm2, acabam usando ele como padrão e esquecem o terminal padrão do Mac por completo. ### .3.5. Slack - Site: [https://slack.com/intl/pt-br/](https://slack.com/intl/pt-br/) - Valor: Grátis A maioria esmagadora das empresas que já trabalhei usam o Slack como ferramente de comunicação interna, sendo assim fica difícil não instalar esse app. Ele consome MUITA memória, mas necessidade não se discute, né? ### .3.6. Spectacle - Site: [https://www.spectacleapp.com/](https://www.spectacleapp.com/) - Valor: Grátis O Mac não tem teclas de atalho para gerenciamento da posição das janelas. As vezes você quer colocar a janela ocupando toda tela, mas não quer tela cheia, pois o Mac tem um comportamento diferente e deixa todas as outras telas em preto, ou seja, não da pra ficar no Mac sem um software que ofereça opções para isso. O Spectacle é uma boa opção. Com o atalho ```control + CMD + F``` você faz com que a janela do software atual ocupe toda tela. Ao fazer esse post, descobri que ele foi descontinuado e sugere a utilização do [Rectangle](https://github.com/rxhanson/Rectangle) Vale a pena olhar esse novo. Vou continuar com o Spectacle, pois ele me atende tranquilamente. ### .3.7. Telegram - Site: https://telegram.org/ - Valor: Opensource Uma ótima ferramente de troca de mensagens. No Brasil a maioria das comunidades técnicas estão nesse software de bate papo. A prova disso é esse extensa lista de grupos no telegram: [https://listatelegram.github.io/](https://listatelegram.github.io/) Ele é praticamente igual ao Whatsapp, só que com mais funcionalidades, e sem necessidade de celular ligado para acessar versão desktop/web, por exemplo. ### .3.8. Vanilla - Site: https://matthewpalmer.net/vanilla/ - Valor: Grátis com limites Se você instala uma grande lista de apps e isso lota sua barra de notificação (Aquela que fica ao lado do relógio), esse app é pra você, pois ele oculta e mostra os apps com apenas um clique. Simples e direto. Gostei. ### .3.9. Whatsapp - Site: [https://www.whatsapp.com/](https://www.whatsapp.com/) - Valor: Grátis É praticamente um padrão de comunicação entre as pessoas no Brasil. Por mais que não curto esse mensageiro, sou obrigado a usá-lo por conta disso. ### .3.10. GPG Suite - Site: [https://gpgtools.org/](https://gpgtools.org/) - Valor: Opensource Se você encripta seus emails, arquivos e afins, acredito que esse app é essencial pra você. ### .3.11. Keybase - Site: [https://keybase.io/](https://keybase.io/) - Valor: Opensource Esse é uma outra aposta. Muita gente fala bem dele, mas confesso que ainda não usei de verdade. A ideia dele é oferecer uma plataforma segura para comunicação, troca de arquivos e afins. ### .3.12. Transmission - Site: [https://transmissionbt.com/](https://transmissionbt.com/) - Valor: Opensource Um cliente para baixar arquivos via BitTorrent. Muito leve e bom. ### .3.13. Spotify - Site: [https://www.spotify.com/br/](https://www.spotify.com/br/) - Valor: Grátis com limites Um cliente para uma plataforma de músicas online, que também oferece opção para Podcasts. Muito usado por boa parte das pessoas. ### .3.14. Flycut - Site: [https://apps.apple.com/br/app/flycut-clipboard-manager/id442160987?mt=12](https://apps.apple.com/br/app/flycut-clipboard-manager/id442160987?mt=12) - Valor: Grátis Outra aposta. O objetivo dessa ferramenta é oferecer uma opção mais potente de gerenciamento de área de transferência (O que você pega no ```CMD+C``` ). ## .4. Aplicativos para desenvolvedores (Com Brew) ### .4.1. Visual Studio Code - Site: [https://code.visualstudio.com/](https://code.visualstudio.com/) - Valor: Opensource Esse editor de código tem crescido cada vez mais entre as pessoas que trabalham com infra/devops/automação. Acredito que por conta dos plugins/extensões que aumentam bastante a sua produtividade. Prometo que depois faço um artigo apenas sobre as extensões do meu Visual Studio Code. ### .4.2. Kubernetes-cli - Site: [https://kubernetes.io/docs/tasks/tools/install-kubectl/#install-kubectl-on-macos](https://kubernetes.io/docs/tasks/tools/install-kubectl/#install-kubectl-on-macos) - Valor: Opensource Se você trabalha com kubernetes de alguma forma, essa ferramenta é necessária. Ela é usada para se comunicar com o cluster kubernetes. ### .4.3. JQ - Site: [https://stedolan.github.io/jq/](https://stedolan.github.io/jq/) - Valor: Opensource Essa ferramenta é considerada o ```sed``` para formato JSON. É muito potente na coleta e manipulação de dados que contenham o formato JSON. Se quiser aprender um pouco como usar, veja esse [link](https://stedolan.github.io/jq/tutorial/) ### .4.4. Nmap - Site: [https://nmap.org/](https://nmap.org/) - Valor: Opensource Ferramenta potente e extremamente extensível para descoberta e auditoria de segurança de rede. Muito bom para descobrir quais portas estão abertas no seu servidor e afins. ### .4.5. Openssl - Site: [https://www.openssl.org/](https://www.openssl.org/) - Valor: Opensource Ferramenta necessária para gerenciamento e criação de certificados SSL/TLS. ### .4.6. Watch - Site: [https://en.wikipedia.org/wiki/Watch_(Unix)](https://en.wikipedia.org/wiki/Watch_(Unix)) - Valor: Opensource Ferramente que é usada para mostrar repetidamente a mesma tela, atualizando automaticamente periodicamente com base em um espaço de tempo que você estipulou. Exemplo: Você quer saber se seu processo do PHP subiu corretamente, ou quantos processo tem aparecido em um espaço de tempo. Você pode usar esse comando: ```watch "ps -e | grep php"``` Por padrão o tempo de atualização é de 2 segundos, sendo assim ele executará o comando ```ps -e | grep php``` e mostrará na sua tela a saída do mesmo. Sendo assim você não precisa ficar dando o mesmo comando várias vezes, "sujando" seu histórico de comandos com vários comandos repetidos sucessivamente. ### .4.7. Docker - Site: [https://docs.docker.com/docker-for-mac/](https://docs.docker.com/docker-for-mac/) - Valor: Opensource Esse eu preciso de pouco pra defender, correto? Ele é basicamente hoje a base da maioria das pessoas que trabalham com automação de infraestrutura. Eu mesmo uso ele para praticamente tudo. Eu não instalo o aws-cli, terraform, packer e afins, pois eu uso um container com esses binários dentro a partir do docker. Por exemplo: ``` docker run -it -v $PWD:/app -w /app --entrypoint="" terraform:light sh ``` Com esse comando eu tenho uma cli com terraform mais atualizado instalado com sucesso. ### .4.8. Popeye - Site: [https://github.com/derailed/popeye](https://github.com/derailed/popeye) - Valor: Opensource Ferramenta interessante para scan de cluster kubernetes, que reporta possíveis problemas em recursos "deployados" e configurações. É uma aposta, dica do [@badtux_](https://twitter.com/badtux_). ### .4.9. Stern - Site: [https://github.com/wercker/stern](https://github.com/wercker/stern) - Valor: Opensource Mais uma ferramente sugerida pelo [@badtux_](https://twitter.com/badtux_) para ajudar na gestão de logs de múltiplos pods em um cluster kubernetes. ### .4.10. Kubectx + Kubens - Site: [https://github.com/ahmetb/kubectx](https://github.com/ahmetb/kubectx) - Valor: Opensource De novo o [@badtux_](https://twitter.com/badtux_) com sugestões massas para ajudar na gestão do kubernetes. Essas oferecem funcionalidades de rápida troca de namespace e cluster kubernetes . Outra aposta. Valeu [@badtux_](https://twitter.com/badtux_). ### .4.11. Insomnia - Site: [https://insomnia.rest/](https://insomnia.rest/) - Valor: Opensource Esse foi amplamente pedido por várias pessoas em diversos canais (Obrigado [@@jabezerra](https://twitter.com/jabezerra) e [@malaquiasdev](https://twitter.com/malaquiasdev) por serem os primeiros). Resolvi testar e gostei. Está na lista, como uma boa aposta. ### .4.12. Docker Compose - Site: [https://docs.docker.com/compose/](https://docs.docker.com/compose/) - Valor: Opensource Não menos importante, mas eu sempre esqueço dele e só vou instalar depois que preciso. Essa ferramente é responsável por gerenciar a execução de múltiplos containers com apenas um comando e um arquivo de configuração. É uma ótima ferramenta para iniciar infraestrutura complexa localmente na máquina de uma pessoa que desenvolve o produto. ## .5. Agradecimentos - Obrigado a todo mundo que ajudou nessa [thread](https://twitter.com/gomex/status/1212585682650177537). - Obrigado a [@somatorio](https://twitter.com/Somatorio) que sempre revisa meus trabalhos ;) <3 - Obrigado a [@jjunior0x2A](https://twitter.com/jjunior0x2A) que deu uns pitacos também. - Obrigado a [@badtux_](https://twitter.com/badtux_) que colocou bastante coisa de Kubernetes na lista. - Obrigado a [@@jabezerra](https://twitter.com/jabezerra) e [@malaquiasdev](https://twitter.com/malaquiasdev) pelo lembrete do Insomnia.
47.426752
444
0.731467
por_Latn
0.997826
21a615b96becf078212192c98f5e78aba705b535
122
md
Markdown
content/cultivars/uji-hikari.md
tonyketcham/puerh.wtf
4cdf1067b94ad435d6e2f69a94717ddcfd2b5de9
[ "MIT" ]
3
2020-11-18T12:28:01.000Z
2021-07-31T14:55:12.000Z
content/cultivars/uji-hikari.md
tonyketcham/puerh.wtf
4cdf1067b94ad435d6e2f69a94717ddcfd2b5de9
[ "MIT" ]
192
2020-11-18T00:31:44.000Z
2022-03-31T11:13:02.000Z
content/cultivars/uji-hikari.md
tonyketcham/puerh.wtf
4cdf1067b94ad435d6e2f69a94717ddcfd2b5de9
[ "MIT" ]
2
2020-11-18T00:13:08.000Z
2021-11-06T14:05:19.000Z
--- id: 3bb9ab60-9225-4a97-8f7b-f4b5cc7633e8 title: Uji Hikari styles: - 984ad797-2060-4eb5-b0f3-317dbaf812b9 ---
17.428571
41
0.704918
ces_Latn
0.03813
21a62626ec4a8c030a23ac12e5866e3510f6941d
3,441
md
Markdown
src/docs/get-started/test-drive/_vscode.md
Martingoesforword/flutter.cn
9b6aeb24cf17703a61549647d475d9bdf9712079
[ "CC-BY-3.0" ]
null
null
null
src/docs/get-started/test-drive/_vscode.md
Martingoesforword/flutter.cn
9b6aeb24cf17703a61549647d475d9bdf9712079
[ "CC-BY-3.0" ]
null
null
null
src/docs/get-started/test-drive/_vscode.md
Martingoesforword/flutter.cn
9b6aeb24cf17703a61549647d475d9bdf9712079
[ "CC-BY-3.0" ]
null
null
null
<div class="tab-pane" id="vscode" role="tabpanel" aria-labelledby="vscode-tab" markdown="1"> {% include_relative _web-app.md %} ## Create the app {#create-app} ## 创建应用 1. Invoke **View > Command Palette**. 打开 **View > Command Palette**。 1. Type "flutter", and select the **Flutter: New Application Project**. 输入 "flutter",选择 **Flutter: New Application Project**。 1. Enter a project name, such as `myapp`, and press **Enter**. 输入项目名称,比如 `myapp`,然后点 **Enter** 1. Create or select the parent directory for the new project folder. 创建或者选择新项目的父文件夹。 1. Wait for project creation to complete and the `main.dart` file to appear. 稍等一下项目创建成功,目录中就会生成 `main.dart` 文件。 The above commands create a Flutter project directory called `myapp` that contains a simple demo app that uses [Material Components][]. 上面的操作会创建一个叫做 `myapp` 的 Flutter 项目目录 该目录中包含一个用到 [Material Components][]简单的示例程序。 {% include_relative _package-name.md %} {% include_relative _restart-vscode.md %} {% include_relative _main-code-note.md %} ## Run the app ## 运行应用 1. Locate the VS Code status bar (the blue bar at the bottom of the window):<br> 找到 VS Code 的状态栏(窗口底部蓝色的条)<br> ![status bar][]{:.mw-100.pt-1.pb-2} 1. Select a device from the **Device Selector** area. For details, see [Quickly switching between Flutter devices][]. 从 **Device Selector** 里选择一个设备。 更多详细内容,请参见 [在 Flutter 设备中快速切换][]。 - If no device is available and you want to use a device simulator, click **No Devices** and launch a simulator. 如果没有可用设备而且你也打算用模拟器, 点击 **No Devices** 然后打开一个模拟器。 {{site.alert.warning}} You may not see **Start iOS Simulator** option when you click **No Devices** in VS Code. If you are on Mac then you may have to run following command in terminal to launch a simulator. 当你在 VS Code 中点击 **No Devices** 的时候,你也许看不到 **Start iOS Simulator** 选项。如果你使用 Mac,你可能得运行以下命令才能启动模拟器。 ``` open -a simulator ``` In Android it is not possible to launch iOS simulator. Android 是不可能启动 iOS 模拟器的。 {{site.alert.end}} - To setup a real device, follow the device-specific instructions on the [Install][] page for your OS. 要设置真实设备,请在操作系统的[安装][Install]页面进行操作。 1. Invoke **Run > Start Debugging** or press <kbd>F5</kbd>. 选择 **Run > 开始 Debugging** 或者按<kbd>F5</kbd>。 1. Wait for the app to launch &mdash; progress is printed in the **Debug Console** view. 当应用启动以后&mdash; 处理进度会出现在 **Debug Console** 页面中。 {% capture save_changes -%} : invoke **Save All**, or click **Hot Reload** {% include_relative _hot-reload-icon.md %}. {% endcapture %} {% capture save_changes_zhcn -%} : 选择 **Save All**,或者点击**热重载** 按钮: {% include_relative _hot-reload-icon.md %} (绿色的带有圆形箭头的按钮)。 {% endcapture %} {% include_relative _try-hot-reload.md save_changes=save_changes %} {% include run-profile.md %} [Install]: /docs/get-started/install [Material Components]: {{site.material}}/guidelines [Quickly switching between Flutter devices]: https://dartcode.org/docs/quickly-switching-between-flutter-devices [status bar]: {% asset tools/vs-code/device_status_bar.png @path %} [在 Flutter 设备中快速切换]: https://dartcode.org/docs/quickly-switching-between-flutter-devices [安装]: /docs/get-started/install [trusted your computer]: /docs/get-started/install/macos#trust </div>
29.663793
190
0.674513
eng_Latn
0.859042
21a6359b5a9fb812160f0aab751d6bd8b667ea6a
540
md
Markdown
README.md
srijithmass/Experiment_Basic_HTML_Tag
ad661782f9d1045a0f40dad00eac974349b58933
[ "BSD-3-Clause" ]
null
null
null
README.md
srijithmass/Experiment_Basic_HTML_Tag
ad661782f9d1045a0f40dad00eac974349b58933
[ "BSD-3-Clause" ]
null
null
null
README.md
srijithmass/Experiment_Basic_HTML_Tag
ad661782f9d1045a0f40dad00eac974349b58933
[ "BSD-3-Clause" ]
null
null
null
# Experiment_Basic_HTML_Tag ## AIM To Write a html webpage using heading, paragraph and list tags. # ALGORITHM ### STEP 1 create a simple html page using heading tag ### STEP 2 Add a paragragh tag to the program ### STEP 3 ### STEP 4 Execute the program # CODE ~~~ <!DOCTYPE HTML> <html> <body> <h1>Name:SRIJITH R</h1> <h2>Reference Number: 21004191</h2> <h3>Dept:Artificial Intelligence and Machine Learning</h3> </body> </html> ~~~ # OUPUT Name: SRIJITH R Reference Number: 21004191 Dept: Artificial Intelligence And Machine Learning
17.419355
63
0.731481
kor_Hang
0.539416
21a68df8ea21040a79aa34dd683fbf69d1db17f7
24
md
Markdown
README.md
VPyad/in-rush-sample-projects
2058227b904815094b8e912b1700daff43293423
[ "MIT" ]
null
null
null
README.md
VPyad/in-rush-sample-projects
2058227b904815094b8e912b1700daff43293423
[ "MIT" ]
null
null
null
README.md
VPyad/in-rush-sample-projects
2058227b904815094b8e912b1700daff43293423
[ "MIT" ]
null
null
null
in-rush-sample-projects
12
23
0.833333
eng_Latn
0.974305
21a77415f8d81cf6eed472a1b237d7e55c55586d
39
md
Markdown
people/ut0363.md
utahelections/utah-elections
4ce667a6071919c381243cd8cf287a879eb7962c
[ "CC0-1.0" ]
4
2017-08-08T20:34:59.000Z
2019-01-29T05:50:05.000Z
people/ut0363.md
utahelections/utah-elections
4ce667a6071919c381243cd8cf287a879eb7962c
[ "CC0-1.0" ]
30
2017-06-15T01:33:37.000Z
2017-08-08T16:52:42.000Z
people/ut0363.md
utahelections/utah-elections
4ce667a6071919c381243cd8cf287a879eb7962c
[ "CC0-1.0" ]
4
2017-08-09T02:15:15.000Z
2020-02-20T05:04:42.000Z
--- layout: person person: ut0363 ---
6.5
14
0.615385
eng_Latn
0.347977
21a789dde9e23d5bae67af629ba8b58f3953ed2d
2,513
md
Markdown
README.md
Isaiah-Hamilton/Isaiah-Hamilton
9b0abcfcd9265bcbab673a1b53527bb1b82156b9
[ "MIT" ]
1
2021-11-01T20:47:07.000Z
2021-11-01T20:47:07.000Z
README.md
Isaiah-Hamilton/isaiah-hamilton
9b0abcfcd9265bcbab673a1b53527bb1b82156b9
[ "MIT" ]
1
2021-05-04T16:57:09.000Z
2021-05-04T18:10:00.000Z
README.md
Isaiah-Hamilton/Isaiah-Hamilton
9b0abcfcd9265bcbab673a1b53527bb1b82156b9
[ "MIT" ]
2
2021-05-07T01:28:55.000Z
2022-01-12T13:16:01.000Z
## Hey, I'm [Isaiah](https://isaiah-hamilton.com) 👋🏾 <br> I'm a full stack developer based in Atlanta, Georgia. I have a strong passion for coding and building things that impact people's day to day lives and make them better. I am interested in Typescript, Next.js, Deno, Supabase, and building consumer apps + developer tools. In my free time, you can see skating, gaming, exploring, and talking to new people. <br> - 🔭 I’m currently working on [supabase](https://github.com/supabase) - 🌱 I’m currently learning Typescript, Nextjs, and Supabase - 💬 Ask me about anything [here](https://github.com/Isaiah-Hamilton/Isaiah-Hamilton/issues) - 📫 How to reach me: Discord(Flux#8742) [Twitter](https://twitter.com/isaiah7hamilton) <br> ## 🛠 Tech Stack <p align="center"> <img src="https://img.shields.io/badge/-JavaScript-05122A?style=flat&logo=javascript"> <img src="https://img.shields.io/badge/-TypeScript-05122A?style=flat&logo=typeScript"> <img src="https://img.shields.io/badge/-tailwindcss-05122A?style=flat&logo=tailwindcss"> <img src="https://img.shields.io/badge/-graphql-05122A?style=flat&logo=graphql"> <img src="https://img.shields.io/badge/-Nextjs-05122A?style=flat&logo=next.js"> <img src="https://img.shields.io/badge/-React-05122A?style=flat&logo=react"> <img src="https://img.shields.io/badge/-React%20Native-05122A?style=flat&logo=react"> <img src="https://img.shields.io/badge/-Deno-05122A?style=flat&logo=deno"> <img src="https://img.shields.io/badge/-Git-05122A?style=flat&logo=git"> <img src="https://img.shields.io/badge/-GitHub-05122A?style=flat&logo=github"> </p> <br /> <p align="center"> <img width="530em" src="https://github-readme-stats.vercel.app/api?username=Isaiah-Hamilton&theme=github_dark&show_icons=true&hide_border=true&include_all_commits=true&count_private=true"> </p> ## <p align="center" style="background:blue"> <a href="https://www.instagram.com/isaiah7hamilton/" target="_blank"> <img align="center" src="https://img.shields.io/badge/-isaiah7hamilton-05122A?style=flat&logo=instagram" alt="Instagram"/> </a> <a href="https://www.linkedin.com/in/isaiah-hamilton-093361209/" target="_blank"> <img align="center" src="https://img.shields.io/badge/-Isaiah%20Hamilton-05122A?style=flat&logo=linkedin" alt="Linkedin"/> </a> <a href="https://twitter.com/Isaiah7Hamilton" target="_blank"> <img align="center" src="https://img.shields.io/badge/-isaiah7hamilton-05122A?style=flat&logo=twitter" alt="Twitter"/> </a> </p>
47.415094
354
0.726224
yue_Hant
0.205498
21a9e1530df1b73346d8de0d6784b6c9cac36aa2
8,524
md
Markdown
classes/Creating-A-Conda-Environment/README.md
maallen3/Elements-of-Style-Workflow-Creation-Maintenance
1291747633d449b8f302d0293ee48072ed241fbd
[ "MIT" ]
3
2022-03-25T19:25:02.000Z
2022-03-28T09:18:45.000Z
classes/Creating-A-Conda-Environment/README.md
maallen3/Elements-of-Style-Workflow-Creation-Maintenance
1291747633d449b8f302d0293ee48072ed241fbd
[ "MIT" ]
2
2022-03-23T19:45:07.000Z
2022-03-29T15:56:50.000Z
classes/Creating-A-Conda-Environment/README.md
maallen3/Elements-of-Style-Workflow-Creation-Maintenance
1291747633d449b8f302d0293ee48072ed241fbd
[ "MIT" ]
11
2022-03-23T14:46:24.000Z
2022-03-31T16:06:50.000Z
# Creating a `Conda` Environment <img src="https://upload.wikimedia.org/wikipedia/commons/e/ea/Conda_logo.svg"> ## What is [conda](https://docs.conda.io/en/latest/) Conda is an application that we use to control our environment and manage our package installations. We will work today within a terminal window and the first thing I like to do is to make a clean environment window. This ensures that all of the assumptions we made about our environment are clear and explicit. Without this you may inadvertantly introduce dependicies that may complicate your building of your container and/or stitching together your workflow. When we build our container, we begin many of times using the conda environment made by the group that builds conda as a basis. This then does give us the application conda within our environment for our use like docker is available in our google shell now. We will walk through those steps today with our simple example. ## [Anaconda](https://anaconda.org/) Search Packages <img src="https://github.com/NIH-NICHD/Elements-of-Style-Workflow-Creation-Maintenance/blob/main/assets/AnacondaMainWindow1.png"> I regularly type this in my browser window to find the packages I need to install. Anaconda environment is growing by leaps and bounds and there are more things that can be done there that I have yet to explore. In addition to well vetted packages that can be installed, this large community resource provides tutorials, data science instructions as well as whole environments where you can work, explore and learn. For example, today we already showed you how to run a jupyterlab notebook, but you want to know more. Their site offers instruction on [Jupyterlab basics](https://anaconda.org/ijstokes/open-data-science-with-anaconda/notebook) for you to explore. ## Logging into Google Shell For this part of our demo/hands on, we will be using google shell. Navigate please to [Google Shell](https://shell.cloud.google.com/) It will ask you to login, please follow the prompts and do so. Upon success you will see a window like this: <img src="https://github.com/NIH-NICHD/Elements-of-Style-Workflow-Creation-Maintenance/blob/main/assets/GoogleCloudShell1.png"> We have an editor as well as a prompt. ## Installing conda We need to get the installation package for conda. We get this package from continuum analytics, we will be installing [Miniconda](https://docs.conda.io/en/latest/miniconda.html). Miniconda is the minimal installer and all we need. Navigating to that location you will see details here: <img src="https://github.com/NIH-NICHD/Elements-of-Style-Workflow-Creation-Maintenance/blob/main/assets/AnacondaMiniconda2.png"> Which installation do we need? Typing `uname -a` will tell us the information about our operating system and hardware basics. ```bash uname -a ``` We see that we are using the `Linux` operating system, the first day I spun up this shell was on `March 16` and we see that it has the architecture `x86_64`. A handy `64-bit` operating system. Scroll down the window and hover your cursor over the `Miniconda Linux 64-bit` (sorry it is implicitly that this is the right installation for the `x86-64` architecture. Right click and select `copy link address`. This gives you the unique resource locator (url) for the installation script we will be downloading onto the machine we are talking to with our `google cloud shell` window. <img src="https://github.com/NIH-NICHD/Elements-of-Style-Workflow-Creation-Maintenance/blob/main/assets/AnacondaMiniconda3.png"> Now we will use the routine `wget` to copy this installation script. To make sure we can do this, let's make sure that we have the tool available. To do that, we type: ```bash which wget ``` In my case, it returns the full path `/usr/bin/wget`. If you do not have `wget`, we can install it with `Anaconda` provided packages. Now to install Miniconda, so that we can use `conda` to install any and all packages we need we do as follows: ```bash wget https://repo.anaconda.com/miniconda/Miniconda3-latest-Linux-x86_64.sh ``` If you haven't changed directories, you can see now that this package `Miniconda3-latest-Linux-x86_64.sh` is now in the top directory. As we learned earlier, we can use the `bash` shell commnad `ls` to inspect the contents of our current directory. `ls -l` provides the full path details of our directory. We check our current working directory with the `bash` shell comand `pwd` as we learned earlier. ```bash pwd ``` <img src="https://github.com/NIH-NICHD/Elements-of-Style-Workflow-Creation-Maintenance/blob/main/assets/GoogleShellWindow2.png"> For my execution, it returns `/home/ad376`. Now running the command `ls -l` we see the downloaded file `Miniconda3-latest-Linux-x86-64.sh`. ```bash ls -l ``` returns: ```bash /home/ad376 ad376@cloudshell:~$ ls -l total 73892 -rw-r--r-- 1 ad376 ad376 75660608 Feb 15 19:07 Miniconda3-latest-Linux-x86_64.sh -rw-r--r-- 1 ad376 ad376 913 Mar 31 12:34 README-cloudshell.txt ad376@cloudshell:~$ ``` To execute this script now and install the `conda` application, we type: ```bash bash Miniconda3-latest-Linux-x86_64.sh ``` This will return: ```bash Welcome to Miniconda3 py39_4.11.0 In order to continue the installation process, please review the license agreement. Please, press ENTER to continue >>> ``` press ENTER to continue and accept the terms and defaults as instructed. I used SPACE to accelerate the passing of the license - when you have time you should read the details of this license as it is interesting. The last question is: ```bash Do you accept the license terms? [yes|no] [no] >>> yes ``` It returns now: ```bash Miniconda3 will now be installed into this location: /home/ad376/miniconda3 - Press ENTER to confirm the location - Press CTRL-C to abort the installation - Or specify a different location below [/home/ad376/miniconda3] >>> ``` Accept the defaults. When finished we can see that the file is now in our execution path is the `conda` application. While `conda` now has been installed, to activate the use of `conda` we need to `reboot` our `shell` environment. The command `exec -l bash` will execute this restart of the bash shell. ```bash exec -l bash ``` Now we can see the result: ```bash ad376@cloudshell:~$ exec -l bash Welcome to Cloud Shell! Type "help" to get started. To set your Cloud Platform project in this session use “gcloud config set project [PROJECT_ID]” (base) ad376@cloudshell:~$ ``` And we can see now we have the `conda` application. ```bash which conda ``` returns: ```bash /home/ad376/miniconda3/condabin/conda ``` ## Creating a clean environment Now that we have `conda` we can create an environment that will keep our workspace clean. We can name it whatever we would like, since this is the `Elements of Style` course, I choose to name my environment `eos` for short. We create an environment as follows: ```bash conda create -n eos -y ``` You can see now that we now have the option to activate the environment ```bash # # To activate this environment, use # # $ conda activate eos # # To deactivate an active environment, use # # $ conda deactivate (base) ad376@cloudshell:~$ conda create -n eos -y^C ``` Now that we have created a clean environment, we follow the directions and activate the environment, that is enter this new clean environment. ```bash conda activate eos ``` We will see next to our name in the google shell, that it should look something like this: ```bash (base) ad376@cloudshell:~$ conda activate eos (eos) ad376@cloudshell:~$ ``` The `(eos)` indicating that we are in that environment. You can list all your environments list all your environments with the following command: ```bash conda env list ``` ## Recap We have learned in this lesson * We started a `bash shell environment` using our browser and the [shell](https://shell.cloud.google.com) * We explored the open community-based resources offered through [Anaconda](https://anaconda.org) * We used `wget` available within the [shell](https://shell.cloud.google.com) and got the installation package necessary to install `conda` * We installed `conda` and * We used numerous command line arguments * We used `conda` to create a new clean environment. We now move on to build Docker images which we will use in a workflow. ## Return to the Agenda [Main Agenda](https://github.com/NIH-NICHD/Elements-of-Style-Workflow-Creation-Maintenance#readme)
38.053571
726
0.757743
eng_Latn
0.995858