repo_name
stringlengths 4
136
| issue_id
stringlengths 5
10
| text
stringlengths 37
4.84M
|
---|---|---|
geospace-code/h5fortran | 757772065 | Title: gfortran compile example in the install instructions does not seem to be enough
Question:
username_0: I tried "if not using CMake FetchContent as is suggested" under https://github.com/geospace-code/h5fortran/blob/master/Install.md#optional-install and this seems to be lacking libraries.
Later I found this: "On Ubuntu it looks like" https://github.com/geospace-code/h5fortran/tree/master/Examples#non-cmake-build which looks more complete. It seems that the two examples do not match.
(working on https://github.com/openjournals/joss-reviews/issues/2842)
Answers:
username_1: Yes thank you. I removed the inconsistencies and deduplicated text: https://github.com/geospace-code/h5fortran/blob/main/Install.md#use-h5fortran-from-your-project
The new h5fortran packaging method automatically seeks HDF5 library, the user doesn't need to manually specify that in their own project's CMakeLists.txt
username_0: Thank you for resolving this.
Status: Issue closed
|
saesrpg/saesrpg-gang | 624443793 | Title: Underground Empire Updates.
Question:
username_0: Gang safe: {2127.73, 1740.39, 20.39}
ATM: {2201.59,1597.9,68.4}
[Markers](https://pastebin.com/30V1F9U2).
[Base file](https://mega.nz/file/KBtDAZrS), encryption key will be sent in gm chat. I've just copy pasted the updated object table into the file Licano sent me to make it easier (hopefully).
Please update the regular spawn to the following:
```lua
<spawn id="Underground Empire" skin="240,124,186" x="2192.517" y="1600.65" z="68.74" rot="180" weapons="2,1;24,250;27,250;29,350;31,450;34,150;46,1;16,10;41,4000"
description="Underground Empire is an organized crime syndicate." r="80" g="0" b="0" restricted="gang,UndergroundEmpire" ganglevelrestriction="0" />```
We'd also like to have skin 124 replaced with 172
Answers:
username_1: :+1:
Status: Issue closed
|
vim-jp/issues | 198653652 | Title: コマンドラインの動作が勝手にスクロールする様になった
Question:
username_0: ## 質問・報告の内容
以前は発生していなかったが、おそらく job/channel/timer が入ったあたりで画面再描画が走る様になっている?

再現方法は、コマンドラインで複数行のマルチバイトを入力します。てか1行目から変ですが。以前(job/channel/timerが入る前の頃?)はちゃんと動いていました。
## Vimのバージョン
8.0.134
## OSの種類/ディストリ/バージョン
* Windows 7 64bit
## 使用している or 関係していそうなプラグイン
なし
Answers:
username_0: ちなみに gvim.exe でも発生します。
username_1: `cmdheight` の値はいくつになっていますか?
手元の `cmdheight=2` の 8.0.142 では再現しなかったので。
username_1: vim.exe では再現できました。なんか…MS IMEの描画が位置がアレなのかなぁ?
username_0: gvim の場合のキャプチャです。

username_1: ありがとうございます。
でも僕の手元では再現できていませんね。
…なにがちがうんだろう?
**Google IME でのキャプチャ**

**MS IME でのキャプチャ**

username_0: むむむーーーーーーーーーーー
username_0: bisect してます。。。
username_2: `call timer_start(1000, 'strlen', { 'repeat': -1 })` したらmacOS (Vim 8.0.142) でも再現しましたが、何も発動せずに起こる事象ならば別問題かもしれません。
username_0: そして HEAD まで戻したら再現しないというアレ
username_0: 再現しなくなってしまったので閉じます。
Status: Issue closed
|
aws/aws-sdk-js | 268767521 | Title: Error while tagging sqs resources
Question:
username_0: P.S. working with [API_TagResources](http://docs.aws.amazon.com/resourcegroupstagging/latest/APIReference/API_TagResources.html)
P.S.S. [email protected]
Answers:
username_1: @username_0
According to [this list](http://docs.aws.amazon.com/awsconsolehelpdocs/latest/gsg/supported-resources.html), SQS is not one of the resources that supports tagging.
Please create a feature request on the [SQS forums](https://forums.aws.amazon.com/forum.jspa?forumID=12) letting them know you'd like for them to support tagging.
Status: Issue closed
username_0: @username_1
According to this [changelog](https://github.com/aws/aws-sdk-js/blob/master/CHANGELOG.md#21370) QSQ **_is_** supported, and I can easily tag it via AWS CLI like this `aws sqs tag-queue --queue-url="https://sqs.[region].amazonaws.com/[account]/[queue]" --tags="Key=Value"`
Please reopen the issue.
username_1: P.S. working with [API_TagResources](http://docs.aws.amazon.com/resourcegroupstagging/latest/APIReference/API_TagResources.html)
P.S.S. [email protected]
username_0: @username_1
Apologies from my side too...
I am using ResourceGroupsTaggingAPI and I thought that it should support SQS too.
Will try to make a workaround with SQS.tagQueue and let you know.
username_0: As a summary:
1. AWS-CLI tagging (one queue) - supported
2. AWS-SDK tagging (one queue) - supported
3. GroupsTaggingAPI - is **NOT** supported **BUT** you can use a workaround which worked for me:
```
let promises = queueUrls.map(url => {
return this._sqs.tagQueue({
QueueUrl: [queue url],
Tags: [tags payload]
}).promise();
});
Promise.all(promises).then(res => {}).catch(err => {});
```
@username_1 Thank you for explanations.
Status: Issue closed
|
rahvaalgatus/kestame | 366719291 | Title: Lootuste ja hirmude visuaal
Question:
username_0: Uus Eakuse-laadselt, kus ringi keskel kolm kategooriat ning ümber probleemid (suurus määratud olulisusest) ja nende pakutud lahendused lisakliki all.
Answers:
username_1: siin on värvid jaotatud nõnda, et iga valdkond (kokku on kolm valdkonda - sündivus, tööhõive ja ränne) on oma värvi. sündivus: #8b7883, tööhõive: #65a580; ränne: #2b7aa1
@username_2 kas peaks täpsemalt soovitud tulemust kirjeldama või on sul pilt olemas peas?
username_2: @username_1, gaks juhuks kinnitaksin siiski üle, kas sain nägemusest õigesti aru. :)
- [Uus Eakuse](https://uuseakus.rahvaalgatus.ee/)-laadselt sõõrik kolme kategooriaga keskel.
- Kas sõõriku sees peaks kategooriat kirjeldav tekst asetsema?
- Sõõriku ümber väiksemad (ka värvitud, eks?) ringid, mis kujutavad erinevaid aruteludel tõstatatud probleeme.
- Ringide keskel pakutud lahenduste arv.
- Ringide (ehk probleemide) olulisus tuli pakutud lahenduste arvust või peaks kuidagi teie poolt käsitsi määratud olema?
- Ringidele vajutades või hiirega üle minnes kuvatakse aruteludel pakutud lahendusi. Seal edasi kuskile klikkida ei saa, eks?
Mida väiksemal ekraanil kuvada? Kas peidame üldse, kuvame nimekirja kujul või ootame, et külastaja zoomib sisse ja navigeerib visualisatsiooni ümber?
username_1: Vastan iga punkti kohta eraldi:
Uus Eakuse-laadselt sõõrik kolme kategooriaga keskel. - JAH
Kas sõõriku sees peaks kategooriat kirjeldav tekst asetsema? - täpselt ma ei mõista, kas see tähendab, et nt “sündimus” oleks ühe lausega lahti kirjutatud? Kas see visuaalselt mahub? Kui, siis igaks juhuks panen kohe kaasa ka tekstid, mida kasutame kolme fookusteema kirjeldamiseks.
Sõõriku ümber väiksemad (ka värvitud, eks?) ringid, mis kujutavad erinevaid aruteludel tõstatatud probleeme - jah
Ringide keskel pakutud lahenduste arv - jah
Ringide (ehk probleemide) olulisus tuli pakutud lahenduste arvust või peaks kuidagi teie poolt käsitsi määratud olema? - see tuleb ankeedist. Igalt arutelult saame tulemusena teada, millised on igal teemal kolm kõige olulisemaks hääletatud probleemi
Ringidele vajutades või hiirega üle minnes kuvatakse aruteludel pakutud lahendusi. Seal edasi kuskile klikkida ei saa, eks? - jah. Iseküsimus on muidugi selles, kuidas lahendada, et kõigile probleemidele ei olegi pakutud lahendusi. Kuigi need võivad olla väga oluliseks peetud?
*********
SÜNDIMUSE TÕSTMINE
Ideed, mis aitavad Eesti peredel jõuda kahe ja enama lapseni
TARK RÄNNE
Ideed, mis aitavad väljarännet vähendada, tagasirännet soodustada ning kaasata väljast tulnud inimeste potentsiaali
INIMVARA TÕHUSAM KAASAMINE
Ideed, kuidas tööturult eemal olevaid inimesi paremini kaasata, tööviljakust tõsta, töid automatiseerida ning ümber- ja täiendõpet soodustada
username_2: Nende mõne toimunud arutelu pealt saame natuke juba andmeid mängimiseks, eks? Tegin arutelude tulemuste tabelisse lehe "Probleemid" struktuuri ettepanekuga: Igal real üks probleem, kuid lahendused ühes tulbas reavahega eraldatud. Olulisus lihtsalt eraldi tulbas numbrina.
Kas see haakuks teie töövooga või on mõni teine struktuur käepärasem täitjale?
username_0: Oleme juba üsna lähedal:
Kiiruga vaatan - kui joonistasime kavandit, siis arutasime, et:
- ringi/kasti suurus = probleemi olulisus
- ringi/kasti pealkiri = probleemi nimi
- ringis/kastis pakutud lahenduste arv ja klikates nimekiri
username_2: Kuhu "ringi pealkiri" läheb?
Ringide asemel kastid tekstiga on kindlasti parem, aga 24+ probleemi ei mahu praeguste pikkade fraasidega täies ära. Siis tasub keskmisest sõõrikust ära öelda ja olemegi tagasi kolme tulba või kaardi-laadse kujunduse juures (Rahvaalgatuse esilehe laadselt). :)
username_0: Algne kavand ka siia

username_2: Oh, ikka võid torkida, sest sinu jaoks peab täitmine ju mugav olema. :) Kui aga pakutud lahendused on igaüks niikuinii teises tulbas eraldi ridadel, siis on nende arv sealt tuletatav ja eraldi tulpa arvuga ignoreeritakse.
username_3: Praegu ei ole lahendused igaüks eraldi real, vaid veerus E komadega eraldatud. Kas panen need eraldi ridadele?
username_2: Pakkusin https://github.com/rahvaalgatus/kestame/issues/13#issuecomment-432053924 -s esialgu reavahed välja, aga suurt vahet ei ole. Komadega on teil tõenäoliselt raske, sest need kipuvad eesti keele lausetes esinema. Võid ise valida, kas semikoolonid, reavahed või mõni muu märgis.
username_3: Kui jääb olemasolev lahendus, saan ettepanekud nummerdada, nt: 1. Inimesed olgu õnnelikud. 2. Elu saagu hea jne. Siis saab ehk mugavamalt üle tõsta?
username_2: Numbrid teeksid selle isegi pisut raskemaks, sest siis tuleb neid eristada teistest lauses esinevatest numbritest, eemaldada nad enne kuvamist, et HTML oma tööd teha saaks jms. :) Kõige lihtsamad ongi kas reavahed või harva esinev interpunktsioon. Kui soovid, võime alati igale probleemi lahendusele tabelis oma rea pühendada, aga siis peaksid lahenduse ridade ees ka probleemid kirjas olema ja see võib sul/teil tülikas hallata olla.
username_1: @username_2 ma võtan enda peale joonistamise, et sa saaksid natuke parema ettenägemuse, mida saame veel korra põrgutada ja siis juba töösse lasta. jäin haigeks, aga katsun täna täpsema versiooni valmis teha, sest tuli soov kõiki tulemusi eraldi lehele paigutada ja mitte kuvada neid avalehel. nii et ma katsun selle ideekorje mustandi teha.
username_1: @username_2 võtsime aega ja mõtlesime asjad läbi.
hetkel tundub mõistlik niimoodi teha: veits muudame Avalehe.
Avalehele lisandub nupp "tutvu tulemustega", mis viib lehele, mis hetkel kannab nime "Ideekorje". Värv: #ff4800
[Avaleht 30 10 2018.pdf](https://github.com/rahvaalgatus/kestame/files/2538659/Avaleht.30.10.2018.pdf)
nüüd kõik tulemused viskame hetkel "Ideekorje" lehele. Nimetame selle lehe "Tulemused" ümber. Sisu võiks olla nagu mock-up'il.
Koosneb neljast osast:
I osa jaoks teeb Maris teksti. Nupp tuleb laivi alles 23.11, aga võiksid selle planeerida (ta viitab google ankeedile, mida ei hakka embeddima).
II osa on meie juba tehtud sõnapilved.
III osa on graafikud, mida seni polnud. Hetkel on tehtud pildina, aga võiks selle automaatiseerida (andmed on meie samast tulemuste tabelist, teema-probleem-olulisus tulbad). Kas on võimalik seda teha visuaalselt nii nagu pildil - et füüsiliselt võtab graafik ainult niipalju ruumi, aga klikitavate rubriigide all on kaks ülejäänud teemat? Iga teema tulbad on oma värvi, samu värve kasutame ka IV. osas. Probleem, ehk graafiku selgitav tekst, võiks mahuda kas tervikuna, või olla nähtav kui hiirega selle peale liigud. Kõik suurused ja font on joonise peal illustratiivsed - ehk peaksid olema meie tavalised.
IV osa - kiiruse ja lihtsuse huvides otsustasime siiski tavaliste tulbade kasuks. Näidatud võiksid olla ainult need, millele on pakutud lahendused. Selle osa kujundusest on visuaalselt täpsem. Meie värvid: Sündivus: #ff4800; Ränne: #808284; Tööhõive: #65a580. Lahenduste osa on hetkel tehtud mul sama värviga, lihtsalt opacity vähendatud (pildil on 18%). Shadow on must (x:0/y:2/blur:4). Nooli shadow on ka must (x:0/y:0/blur:5). Mõlemaid aga võid oma nägemuse järgi sättida. Ristküliku nurgad on raadiusega 4, kui avaneb aken selgitustega, siis alumised muutuvad 0ks.
[tulemused copy.pdf](https://github.com/rahvaalgatus/kestame/files/2538717/tulemused.copy.pdf)
username_1: Uued tekstid "Ideekorje" lehe jaoks on .md failis juba olemas.
username_1: @username_2 Andri, on sul kommentaare või küsimuse? saame me sellega edasi minna?
username_2: Kas numbrid värvitud ribadel viitavad pakutud lahenduste arvule või olulisusele? Eeldan, et nii number kui ka riba pikkus viitab lahenduste arvule ning olulisus määrab vaid järjekorra.
Kuidas see moblas välja võiks näha? Portree-tahvlis vist mahub isegi ära.
Ideenupp tuleb paari nädala pärast, aga tulemuste visualiseerimise ise võiks ASAP valmis teha ja laivi saada, eks?
username_2: Mobiili küsimus võib kehtida ka lahenduste tulpade kohta, kuid eeldan, et ega seal me midagi enamat välja ei mõtle kui lihtsalt kolm tulpa üksteise järel. :)
username_2: Mobiilis on ~320px horisontaalset ruumi ehk u kolmandik PDFi joonisest. See mahutab horisontaalselt enam-vähem kuus-seitse sõna 16px/17px fondiga. Probleemid mahuvad kahele reale küll, aga riba neist paremale mitte. Värvilise riba saaks iga probleemi alla tuua. Teksti ei tahaks väiksemaks teha, sest kahtlustan, et kõik lugejaskond pole esimeses nooruses.
username_1: see tundub ka sobilik lahendus!
username_1: kas tahad, et me selle enne sinuga joonisel vaatame, või teeksid valmis ja siis vaatame? mulle tundub, et selline lahendus on OK, kui olen õigesti tekstist aru saanud :)
username_2: Praegu pole vaja. :) Vaatame, mis välja tuleb!
username_2: Panin esimese redaktsiooni probleemide graafikutest https://kestame.rahvaalgatus.ee/tulemused/ lehele üles. Kombineerin ideekorje lehega siis, kui oleme kõik rahul. Puudu on lahenduste kastidel kolmnurk vms avamise märgis. Lisan selle varsti.
Tegin hetkel nii probleemide graafiku kui lahenduste tulbad ilma pealkirja pikkuse piiramiseta. Lööb küll natuke vertikaalse rütmi sassi, kuid ma ei ole veendunud, et ellips hea lahendus on. Lõpuks on probleemid ju lugemiseks. Avatud soovitusteks. Probleemide graafikul jätsin ka joonestiku ja joonlaua alt välja. Tundub see jätkuvalt vajalik?
Kui tahame lahendused loetavamaks teha --- näiteks kuvada neid nummerdatult --- peame tabelis nad teineteisest eraldama reavahe või mõne muu harvaesineva interpunktsiooniga, nagu ennist (https://github.com/rahvaalgatus/kestame/issues/13#issuecomment-434514626) kirjutasin. Praegu on lihtsalt üks mass teksti, kui mõni probleem tulbas avada. :)
username_1: täname sind! korjasime tagasisidet ja siin ta on tervikuna:
1) tunnetatud probleemide olulisus:
- saab täitsa nii olla, aga prooviks äkki teksti veidi väiksema fondiga teha (nt 16px 18px asemel)
- pealkirja “Tunnetatud probleemide olulisus” - kirjutaks konkreetsemaks “Milliseid probleeme märgiti olulisemaks?"
2) lahendused:
a) iga pakutud lahendus võiks ikkagi olla nummerdatud, sest hetkel tõepoolest ei ole selge, kus üks algab ja teine lõpeb. Maris tegeleb faili korrigeerimisega.
b) ideedega kasti pikkus võiks olla piiratud ning kui teksti on rohkem kui see piir, siis pigem tekib scrolling bar. sest kui oletada, et ideede arv suureneb veelgi, siis kardan, et kast võib liiga pikaks saada.
c) hetkel on tulbad ja tekst paigast ära - tasub ühtlustada.
PS meil on tekstid ka olemas, Maris täiendab neid veel tänase jooksul (need on Ideekorje.md failis).
username_2: Mõtled probleemide kastide kõrgused ei ole võrdsed?
username_3: Tahaks semantilist loendit jah. Võtsin järjekorranumbrid ära ja lisasin reavahed (cmd-enter). Vaatan nüüd lubatud tekstid üle.
username_2: Jaa. Ülal lingitud `/tulemused` leht on vaid meile arutamiseks. Avalikkusele tuleb ta kenasti Ideekorjega kokku nagu soovitud.
username_1: @username_2 Andri, meil oleks vaja tulemustega lehte nädala lõppuni, muidu see kaotab mõtet. On tehtav, eks?
username_2: Tehtav! Teen ka täna ülalmainitud täiendused ära!
username_1: <img width="300" alt="screenshot 2018-11-13 16 31 39" src="https://user-images.githubusercontent.com/41726243/48419844-9e640880-e761-11e8-9fd8-c35dad142c35.png">
tekst ja esimene oranžikas tulb ei asu vertikaalselt ühel liinil
username_2: See peakski olema 16px ehk sama suur kui tekst muudel lehtede. Kui mõtlesid, et paar pikslit väiksem font vähendab probleemide ridade arvu, siis see kahjuks ei tundu see nii olevat. Kõrgus jääb suhteliselt muutumatuks.. Võid Developer Toolsiga Firefoxis või Chrome'is proovida `<div id="#problems">` elemendi `font-size`-i paari ühiku võrra vähendada (`1.6rem` tähendab zoomimata brauseris 16px) .
username_1: proovisin, et 13px peal on näha erinevust, aga siis on vist tõesti tekst liiga väike. Nii et las ta olla nagu on.
Status: Issue closed
username_2: Laivis: https://kestame.rahvaalgatus.ee/ideekorje/
Kuna praegu lahtisi küsimusi ei jäänud, sulgen _issue_. Kui midagi kerkib, võime uuesti avada või uue teha. :)
username_2: Monitoorimise skript jälgib nüüd ka probleemide tabelit. Nii arutelude kui probleemide muudatused tabelis kajastuvad hiljemalt 15 minutit peale muudatust lehel. |
nestjs/elasticsearch | 1111681311 | Title: Build fails on non UNIX systems
Question:
username_0: ### Did you read the migration guide?
- [X] I have read the whole migration guide
### Is there an existing issue that is already proposing this?
- [X] I have searched the existing issues
### Potential Commit/PR that introduced the regression
PR #727
### Versions
_No response_
### Describe the regression
Package fails to build on Windows.
### Minimum reproduction code
```ts
```
### Expected behavior
Package builds on Windows.
### Other
_No response_
Answers:
username_1: Let's track this here https://github.com/nestjs/elasticsearch/pull/727
Status: Issue closed
|
lark-parser/lark | 602798806 | Title: Help needed: How to deal with operator precedence (my first time using parsers)
Question:
username_0: **Problem statement**
Hello, I'm trying to parse a language similar to a tiny subset of Java.
I only need to parse expressions that can apply functions and use the logical and numerical comparison operators, and math operations.
I'm having trouble with precedence, read all the docs (even took the basic math part from there) but couldn't figure it out yet.
**What was tried and examples**
However, I already started using precedence with "value_comparison_op.1", "bool_comparison_op.2" to force priority, but to no avail. Worse, if I use identifiers that access members of the classes I get a different interpretation of the same basic form.
Below are two examples, but my objective first:
```
Input: a == u or b == v
Desired: (a == u) or (b == v)
```
However it is interpreting as `a == (u or (b==v))`
And if I add member access (nested identifier), it's even more broken. See the 2 examples below:

Here's the relevant part (I believe):
```
rule: expression
?expression: value
| sum
| funcall
| "(" expression ")"
| coallesce
| logical_expression
| comment
logical_expression: comparison
| negation
comparison: bool_comparison
?bool_comparison: value_comparison
| expression bool_comparison_op expression
?value_comparison: expression value_comparison_op expression
negation: "!" expression
value_comparison_op: EQ
| NEQ
| GE
| LE
| GT
| LT
bool_comparison_op: or | and
```
I think all the code is suffering from such problems, including `funcall`'s method name before the parenthesis, being identified as an identifier and not a function call and its arglist.
Answers:
username_1: The general idea is to clearly define the tree structure you want to have at the end. If you want this:
```
and_expression:
comparison:
sum:
number
number
...
```
you 'just' have to clearly tell the parser which rules are 'below' which other ones:
```
?and_expression: [and_expression "and"] comparison
?comparison: [sum value_comparison_op] sum
?sum: [sum "+"] number
```
I would suggest looking at the calculator example for more examples.
username_0: Thanks so much @username_1 !
I think it now makes more sense to me :)
Will try to apply that :)
Status: Issue closed
username_0: **Problem statement**
Hello, I'm trying to parse a language similar to a tiny subset of Java.
I only need to parse expressions that can apply functions and use the logical and numerical comparison operators, and math operations.
I'm having trouble with precedence, read all the docs (even took the basic math part from there) but couldn't figure it out yet.
**What was tried and examples**
However, I already started using precedence with "value_comparison_op.1", "bool_comparison_op.2" to force priority, but to no avail. Worse, if I use identifiers that access members of the classes I get a different interpretation of the same basic form.
Below are two examples, but my objective first:
```
Input: a == u or b == v
Desired: (a == u) or (b == v)
```
However it is interpreting as `a == (u or (b==v))`
And if I add member access (nested identifier), it's even more broken. See the 2 examples below:

Here's the relevant part (I believe):
```
rule: expression
?expression: value
| sum
| funcall
| "(" expression ")"
| coallesce
| logical_expression
| comment
logical_expression: comparison
| negation
comparison: bool_comparison
?bool_comparison: value_comparison
| expression bool_comparison_op expression
?value_comparison: expression value_comparison_op expression
negation: "!" expression
value_comparison_op: EQ
| NEQ
| GE
| LE
| GT
| LT
bool_comparison_op: or | and
```
I think all the code is suffering from such problems, including `funcall`'s method name before the parenthesis, being identified as an identifier and not a function call and its arglist.
Thanks!
username_2: Echoing the above comment, I found [this chapter](https://www.craftinginterpreters.com/parsing-expressions.html) of Crafting Interpreters to be very helpful.
Status: Issue closed
|
internetee/registry | 232852717 | Title: papertrail
Question:
username_0: #426
Answers:
username_0: #426
username_0: @username_1 Please provide requirement list.
username_0: https://docs.google.com/document/d/1GFs_kxVElYoCicZ69vAaUIuol4ri23_WbxJNir0RV6M/edit?ts=59364fb7#heading=h.57eceh79hphg
username_0: @username_1 Please provide requirements for 3rd list item in "Requirements" section
username_1: domain:
* date + period
* domain name
* contact
* contact name
* contact code (personal, business id, birthday)
* contact country
* contact type
* status
* host name
* host ip
* registrar
* epp requests (create, update, renew, transfer, delete)
contact:
* date + period
* contact name
* contact code (personal, business id, birthday)
* contact country
* contact type
* status
* registrar
* epp requests (create, update, renew, transfer, delete) + related epp requests (domain update, domain create)
* domain (domains that contact is associated to)
registrar:
* date + period
* registrar name
* business id
* domain
username_0: @ratM1n This needs your input regarding JSON support in Postgres
username_2: @username_0 Postgres does support indexing on JSON fields.
username_0: https://www.postgresql.org/docs/9.4/static/tutorial-inheritance.html
username_0: https://wiki.postgresql.org/wiki/Audit_trigger
username_0: http://propelorm.org/blog/2011/08/29/introducing-archivable-behavior-and-why-soft-delete-is-deprecated.html |
aleksanderwozniak/table_calendar | 490366656 | Title: RenderFlex overflow
Question:
username_0: When I add the calendar like shown in the example I get following error Message:
════════ Exception Caught By rendering library ═════════════════════════════════════════════════════
The following assertion was thrown during layout:
A RenderFlex overflowed by 62 pixels on the bottom.
The overflowing RenderFlex has an orientation of Axis.vertical.
The edge of the RenderFlex that is overflowing has been marked in the rendering with a yellow and black striped pattern. This is usually caused by the contents being too big for the RenderFlex.
Consider applying a flex factor (e.g. using an Expanded widget) to force the children of the RenderFlex to fit within the available space instead of being sized to their natural size.
This is considered an error condition because it indicates that there is content that cannot be seen. If the content is legitimately bigger than the available space, consider clipping it with a ClipRect widget before putting it in the flex, or using a scrollable container rather than a Flex, like a ListView.
The yellow and black bars show up on the screen for a split second but than disappear.
What am I doing wrong?
Answers:
username_1: I'm having the same problem, this happens for example on a iPad and you are on Portrait mode.
username_2: This could be related to #38 or #60. Try specifying `rowHeight`.
username_3: @username_2 Thanks for the awesome package, first off.
I have a question about this issue. So for me changing rowHeight doesn't seem appropriate. What I would like to do is apply a scale to the output Table. I entered a flutter question on this actually:
https://github.com/flutter/flutter/issues/56687
Do you know of any way to maybe do this?
username_4: I'm also struggling to make table calendar scale to fit when there isn't enough room. I have a screen where there isn't quite enough room when there are 5 rows in a month, and the only solution I can find is to wrap it in a `SingleChildScrollView`. I've tried wrapping in combinations of `FittedBox` and `AspectRatio`, which seems like it should work, but the calendar refuses to respond to these size constraints.
This is an awesome library, but seems like there are elements which use fixed sizes, which unfortunately means ignore parent box constraints.
username_5: The easiest solution I found is to add `FlexibleSpaceBar` widget to the `flexibleSpace` of the `AppBar`. Something like this:
```
AppBar(
backgroundColor: constants.RED_COLOR,
title: Text(
formattedDate
),
flexibleSpace: FlexibleSpaceBar(
title: MyCalendarWidget
)),
```
username_2: v3.0.0 introduces `shouldFillViewport` property, which I think might solve this issue without having to use `rowHeight`. There is a lot of changes in that release, but you can quickly test it by running [examples](https://github.com/username_2/table_calendar/tree/3.0.0-beta/example/lib) from `3.0.0-beta` branch locally, and modifying the UI to imitate your situation.
If you want to use `3.0.0-beta` branch in your project, add this to `pubspec.yaml`:
```yaml
table_calendar:
git:
url: git://github.com/username_2/table_calendar.git
ref: 3.0.0-beta
```
username_2: 3.0.0 prerelease with updated API has been uploaded to pub.dev: https://pub.dev/packages/table_calendar/versions/3.0.0-nullsafety.0
Status: Issue closed
|
Alfresco/alfresco-ng2-components | 202462320 | Title: upload dialog Wrong number on File Upload Count
Question:
username_0: Currently File Upload Count Value is Taking from Internal Queus,where Each file is in Queue ,When Ever I have Uploaded new File or Folder its always Display Count from Previous File
for Example: Lets say i have uploaded 1 file and Close File-Dialogue After that i have upload a Folder So. it should display Current Progress i.e 1 but Currenly it showing 2 one for Previous File and 1 For Folder
Answers:
username_1: @username_0 please next time keep the template :)
username_2: A few scenarios where the file upload count might want to be reset:
1. When a new folder is navigated into
2. If the upload dialog is closed
3. If user clicks "refresh count" button
Status: Issue closed
|
awslabs/amplify-video | 1023383071 | Title: Video player fails to play signed urls
Question:
username_0: **Describe the bug**
When signed url is passed to the suggested video player (video.js), it fails with:
No compatible source was found for this media.
```
<VideoPlayer
autoplay
controls
sources={[{
src: 'https://xxxxxxxx.cloudfront.net/{path}/{path.m3u8}{token}',
type: 'application/x-mpegURL'
}]}
```
Answers:
username_0: I resolved the issue using [React Player](https://www.npmjs.com/package/react-player). if anyone else is dealing with playing HLS with authorisation header. here is the way I did it:
```
<ReactPlayer
url={source} //ttps://xxxxxxxx.cloudfront.net/{path}/{path.m3u8}
controls={true}
width="100%"
height="100%"
config={{
file: {
hlsOptions: {
xhrSetup: function xhrSetup(xhr, url) {
xhr.setRequestHeader(
"Access-Control-Allow-Headers",
"Content-Type, Accept, X-Requested-With"
);
xhr.setRequestHeader(
"Access-Control-Allow-Origin",
"*"
);
xhr.setRequestHeader(
"Access-Control-Allow-Credentials",
"true"
);
xhr.open('GET', url + token); // this is your token: ?Policy=foo&Key-Pair-Id=bar&Signature=foobar
}
}
}
}}
/>
```
Status: Issue closed
username_0: @username_1 I am having problem with playing videos using both react player and video js on iOS (Safari). IS there a specific config that needs to be set? I have already tried this but didn't work
```
{
html5: {
hls: {
overrideNative: overrideNative
},
nativeVideoTracks: !overrideNative,
nativeAudioTracks: !overrideNative,
nativeTextTracks: !overrideNative
}
}
```
username_0: https://github.com/awslabs/amplify-video/issues/323 |
WIPACrepo/iceprod | 153873648 | Title: web-based pool management
Question:
username_0: (imported from trac: [#1496](http://code.icecube.wisc.edu/projects/icecube/ticket/1496))
Following on from [#1484](http://code.icecube.wisc.edu/projects/icecube/ticket/1484), make pool management web-based.
Master:
* place to add sites to the pool (returns the passkey for the new site)
Site:
* place to add the master url and the site passkey
Status: Issue closed
Answers:
username_0: not relevant due to 2.4 refactor |
att/rserve-js | 102662459 | Title: decoding barfs on NAs in string arrays
Question:
username_0: what's especially upsetting is that it reports: `Error: URI malformed`
huh? what URI? well, it's doing this:
```{js}
read_string_array: function(attributes, length) {
var a = this.read_stream(length).make(Uint8Array);
var result = [];
var current_str = "";
for (var i=0; i<a.length; ++i)
if (a[i] === 0) {
current_str = decodeURIComponent(escape(current_str));
result.push(current_str);
current_str = "";
} else {
current_str = current_str + String.fromCharCode(a[i]);
}
return [Rserve.Robj.string_array(result, attributes), length];
},
```
we get the sequence of bytes `255,0` for NA, `current_str = "ÿ"`, and it can't be decoded.
my question for @username_1 is: do we want URI decoding here? is Rserve URI encoding the strings?
then i suggest mapping this sequence to null.
[given that the array i'm trying decode has 4071492 bytes, this code is troublesome from a performance perspective, too, but i know how to deal with that: create arrays and join them rather than string `+=`]
Answers:
username_1: Yes, that is a special case for 'NA' - it is encoded as a string containing just 0xff - since it can never occur in UTF8 in is unambiguous. And, yes, IMHO it should map to null (and vice versa).
username_1: And, no Rserve doesn't URI-encode strings - it's not necessary since it's all UTF8.
username_0: Cool, should be an easy fix, and a small performance boost then.
username_0: I bet there is something that depends on the URI decoding, but I'm still inclined to remove it because it's wrong.
username_1: Hm.. you mean this: `decodeURIComponent(escape(...))` - it doesn't decode, I think it's a JS trick to handle something that I can't recall - it may be byte stream to wide char or something like that ...
username_0: Meh, okay I'll leave that and fix the NA thing for RCloud 1.5 then.
username_1: I don't think it can be fixed at application level since by then you cannot distinguish the special sequence from different yet valid \u00ff ...
username_0: Naw, I meant fixing it here in rserve.js where it's crashing. I guess it's a safe change but I'd deploy it to develop/1.5 RCloud.
username_0: It appears this code also barfs on Western Latin ISO-8859 characters:


username_1: Strings must be UTF-8 encoded, so using Latin-1 encoded payload in strings is illegal (i.e., that is not a valid string).
username_0: So this is a tl2010 bug?
username_1: No idea - you didn't say how you got that illegal payload - it has certainly nothing to do with rserve-js.
Perhaps rserve.conf in RCloud needs
encoding utf8
?
username_0: I meant, so is it a bug of an R library to produce such strings, or an RCloud bug? I'll try that, thanks.
username_1: R allows three encodings of strings: native (=same as the locale), UTF-8, Latin1 and bytes (see ?Encoding). We don't support the last one, because the semantics of that are "no idea what this is, but it's some unknown encoding" so there is no way to convert that to a valid string - it's not intended to store binary payload. Rserve is designed to make sure that the client gets only one encoding by re-coding what else comes its way. By default it uses native encoding on the locale unless told otherwise.
So please file a separate bug with more details, because there too many pieces involved here. It could be that DBF (which where tl2010 gets the strings from) flags strings incorrectly or Rserve somehow fails to re-code it properly. I don't even know where that string is passed / how it arrives on the JS side...
username_0: Okay, I should be able to create a simple notebook that reproduces it, and file an RCloud bug. I will try the rserve.conf flag first, though.
Status: Issue closed
|
matthewbdaly/grunt-blogbuilder | 183086285 | Title: Gulp plugin?
Question:
username_0: Thought about making this a Gulp plugin as well? I could see quite a bit of value for that as well, myself included
Answers:
username_1: I have thought about that and I'd like to do it, but it's a bit fiddlier because you have to pass through both the content and templates whereas with Grunt you can just specify them as configuration variables. Unfortunately I'm not likely to have time to do it anytime soon.
username_0: @username_1 Status?
Status: Issue closed
username_1: Not likely to happen any time soon, I'm afraid. If you want to build one yourself based on this I'm happy to link to it, but I won't have the chance to do this myself. |
oneleif/BFTJ-API | 621805744 | Title: Create basic models
Question:
username_0: ## Models
- Post ([Example](https://breadforthejourneyomaha.org/story/congratulations-to-oneleif/))
- Member ([Board](https://breadforthejourneyomaha.org/board/))
- Event ([Example](https://breadforthejourneyomaha.org/events/2020-bread-for-the-journey-fall-table/))<issue_closed>
Status: Issue closed |
ian-ross/canbando | 849943127 | Title: Fix persistence when editing and deleting label definitions
Question:
username_0: - [ ] When editing label definitions, the changes don't get persisted properly.
- [ ] When deleting label definitions, you can end up with cards persisted without an associated list (i.e. the IndexedDB entry has `list: undefined`).
Answers:
username_0: Second part was just a mistake in the test data setup.
Status: Issue closed
|
openstreetmap/iD | 540432465 | Title: standalone query builder
Question:
username_0: hi
I would like to create an generic UI overpass query builder that is user friendly.
how can I separately and independently use the part of iD that deals with building overpass queries? Is there a possibility that we can separate as a independ component?
I would like to create one using Bootstrap or other as a base
Answers:
username_1: Hi @username_0, iD doesn't actually build nor run overpass queries. Tag suggestion in iD's combo dropdowns are returned from the [TagInfo API](https://wiki.openstreetmap.org/wiki/Taginfo/API). See [this file](https://github.com/openstreetmap/iD/blob/2.x/modules/services/taginfo.js) for our implementation.
I'd try looking into the [Overpass Turbo project](https://github.com/tyrasd/overpass-turbo) which has a wizard for building queries. Hope that helps, good luck!
Status: Issue closed
username_0: @username_1 thank you so much for the tips!
I apologize for writing an issue here, but iD's tag suggestion is the closest thing to what I had in mind
username_1: @username_0 Happy to help! I know it can be hard at first to figure out how all the pieces fit together and what questions to ask. You can also check out our Slack group where some overpass experts hang out: https://slack.openstreetmap.us/ |
Matthew-Hsu/PiPass | 95762854 | Title: PiPass not passing
Question:
username_0: For the last week, I've seen no streetpasses on any of the 6 3DS systems that regularly use my relay. I've checked configuration, logs and settings on all devices. The system seems to have just "stopped".
I'm using a Pi rev B
Asustek RT2573 wifi card
Syslog on Pi shows when a device associates, no errors in the PiPass logs.
I can connect a 3DS to the accesspoint (Pi) but it will not access the internet unless I put google DNS directly in the 3DS config.
I can however connect my phone to the accesspoint (Pi) and browse the internet.
The Pi itself does resolve DNS.
I would say it looks like the Pi isn't passing DNS but if it is, it's only an issue passing DNS for 3DS consoles.
Since the latest image says it's only for the Pi2 I've reflashed my card with 1.52, which use to work fine but now does not. I've also updated from 1.52 to 1.6 and I've been unsuccessful in getting this to work. I tried switching to different spreadsheets today and was able to get it working for about 20 min and then nothing again.
I'm a little stumped here. I can't imagine an issue with the Pi itself causing this, and the wifi card has been working like a champ since I set this up.
Any insight would be appreciated.
Answers:
username_0: Ok. So..... This is a little embarrassing. Aparrently my Active Directory Server was on the blink. I noticed the internet seemed a little slow and I was having trouble with DNS resolution on my own network. Since my DC is the first DNS on my network I rebooted it and I could browse the internet on my test console again. Apparently this happened when my UPS dropped my servers last weekend. My DC server died and I resurrected it with a spare chassis, it was authenticating users on my network but apparently wasnt fully operational. I'll be quietly working on that now.
Sorry for wasting a post.
Status: Issue closed
username_1: Glad that you were able to figure out what was going on with your network! Sometimes this stuff is a bit tricky, so don't worry about "wasting a post".
I hope you enjoy PiPass! |
AlphaWallet/alpha-wallet-android | 599228856 | Title: Unified function name, Networks
Question:
username_0: 
Change 'Networks' to'Select Active Networks'
Answers:
username_1: I solved it on Android. Victor check if iOS needs the same to be done to it.
username_0: Yes, I have an issue for iOS as well.
Status: Issue closed
|
redhat-developer-demos/knative-tutorial | 462609424 | Title: Incorrect docker version when building
Question:
username_0: **Describe the bug**
A clear and concise description of what the bug is.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to '...'
2. Click on '....'
3. Scroll down to '....'
4. See error
**Expected behavior**
A clear and concise description of what you expected to happen.
**Screenshots**
If applicable, add screenshots to help explain your problem.
**Desktop (please complete the following information):**
- OS: [e.g. iOS]
- Browser [e.g. chrome, safari]
- Version [e.g. 22]
**Smartphone (please complete the following information):**
- Device: [e.g. iPhone6]
- OS: [e.g. iOS8.1]
- Browser [e.g. stock browser, safari]
- Version [e.g. 22]
**Additional context**
Add any other context about the problem here.
Answers:
username_1: minishift does not have Docker version that supports multistage build
Status: Issue closed
username_2: I encountered this same issue trying to setup the knative tutorial here on macos with minishift.
https://redhat-developer-demos.github.io/knative-tutorial/knative-tutorial-basics/0.5.0/01-setup.html
Is there a reasonable workaround for the multi-stage docker limitation with minishift? If not, then the tutorial should be updated to reflect that it isn't supported for minishift. There are currently an entire set of specific minishift directions at the above link, which appear to work fine until the multistage docker build step.
username_1: I should soon start to update of tutorial to v0.7 will have fresh and updated instructions for OpenShift. 🤞🏻
Get Outlook for iOS<https://aka.ms/o0ukef> |
bl-lia/kktAPK | 476079870 | Title: TimelinePresenter.kt line 126
Question:
username_0: #### in
* Number of crashes: 1
* Impacted devices: 1
There's a lot more information about this crash on crashlytics.com:
[https://fabric.io/bllias-projects/android/apps/com.bl_lia.kirakiratter/issues/c4a7d48bd80c83068f605d7141793766?utm_medium=service_hooks-github&utm_source=issue_impact](https://fabric.io/bllias-projects/android/apps/com.bl_lia.kirakiratter/issues/c4a7d48bd80c83068f605d7141793766?utm_medium=service_hooks-github&utm_source=issue_impact) |
mapgears/ol3-google-maps | 328058487 | Title: Help: Adding a Search Functionality in Google Maps
Question:
username_0: I wanted to add a search box functionality wherein a user could enter a location and as per the location entered the map should pan and zoom to the location.
Answers:
username_1: @username_0 Your question doesn't seem related to olgm. See: https://developers.google.com/maps/documentation/javascript/examples/places-searchbox
Please, ask your questions on Stack Overflow, using the "google-maps" tag.
Status: Issue closed
|
plotly/plotly.py | 349178234 | Title: Add type annotation stubs
Question:
username_0: Hi,
I was wondering if you would be interested in PRs that add type annotation stubs for the plotly.py API. That would enable static type checkers like mypy and Pyre to statically type check Python projects using plotly, and as well as enhancing the IDE experience for editors able to offer richer experiences from type annotations (just as VSCode and PyCharM).
Answers:
username_1: Hi @username_0, I would love to do this... but we still support Python 2.7.
That said, in all of our new generated code we use numpydoc docstrings with return types (see, for example, `plotly/graph_objs/_bar.py`). This is enough for PyCharm to figure things out (I haven't really tested other editors very much) and it does provide a nice experience when working with objects in the `graph_objs` hierarchy.
I would be interested in gaining a better understanding of the most broadly compatible way to specify type information for a mixed Python 2/3 project!
username_0: Great to hear that. The stubs can live in separate files that sit parallel
to the original Python source files, which preserves Python 2
compatibility.
I recently went through this with the Visdom project. See the
`__init__.pyi` file in
https://github.com/facebookresearch/visdom/tree/17f28cb258ad874d6fc5fcce5a4479abe4377738/py/visdom,
which lives alongside the `__init__.py` that it annotates.
This is the same system as Typescript's `.d.ts` files, if you're familiar.
username_1: That looks really interesting. I don't see any downside to adding these to at least the code generation output, because that would automatically stay up to date (and this covers the majority of the public API).
What has your experience been like keeping these up to date as a codebase evolves (for the human generated code 🙂)? What kind of tooling/testing would you recommend to that help flag inconsistencies?
username_0: Well, I won't exaggerate - I became excited about static type checking for Python and started using it for my own research code, but I don't have have direct experience applying it to large multi-developer codebases. I'd look forward to trying and learning, however.
In terms of help from tooling, the only thing I know of is https://github.com/Instagram/MonkeyType. It attempts to produce stubs automatically based on the types it observes variables take during runtime, but I haven't tried it.
The good news is the worse-case scenario for an incorrect stub is a user gets an incorrect warning (or lack of warning) from their type checker and reports it as an issue. Actual runtime behavior can never be affected by incorrect stubs.
username_1: If you're interested in digging in, I'd like to start by adding stub generation to the `graph_obj` code generation logic. The current logic is in `codegen/datatypes.py`. As I said, this is the majority of the public API, and we can keep it correct for free.
Once we have this, I'd like to do a bit of an editor survey to get a full list of the Python editors that support completion across the full nested `graph_objs` hierarchy because of the addition of the stubs (as I mentioned, PyCharm does fine with just the types in the numpydoc strings).
Then I think we'd have enough information to make an informed decision about how much benefit users would get from the slight increase in maintenance burden.
How does that sound?
username_0: Sounds great.
I should mention that there's a formalized generalization of PyCharm's docstring-aware approach whereby type annotations can be encoded into strings and put as comments inline with a source file, instead of as a separate stub (https://www.python.org/dev/peps/pep-0484/#suggested-syntax-for-python-2-7-and-straddling-code). I personally feel this is less elegant than separate stubs that get to use the nice new Python 3 syntax for type annotations and I'm probably less motivated personally to contribute annotations in that format, but it's a possibility.
username_0: I'm going to link to https://github.com/pytorch/pytorch/pull/12500 as an example of a big project that went through the process of adding stubs and is discussing testing strategies etc - might be a useful reference for me in the future or whoever else wants to take a stab at this.
username_2: Hey, I'm wondering if there is any news on this? With powerful static typecheckers such as Pylance (the new language server in VSCode) becoming widely available and adopted, it would be very nice to have decent typing information available for plotly objects.
username_3: There's no news at this time, but if someone wants to pick up the torch and work on a PR or proposal for how to get this done, we'd happily consider it and help you out :)
username_1: Also, depending on how large they are (we have a *lot* or graph object classes), we may need to ship them in a separate package.
username_4: Since python 2 has reached EOL now, I guess maintaining support for it should not be part of the consideration any more?
username_5: As pointed out by @technic in (this comment)[https://github.com/plotly/plotly.py/issues/1682#issuecomment-922361704] from #1682, it looks like a simple workaround is to make use to the [`TYPE_CHECKING`](https://docs.python.org/3/library/typing.html#typing.TYPE_CHECKING) which is specifically designed to handle lazy imports. |
FasterXML/jackson-module-kotlin | 660048588 | Title: JsonProperty is being ignored when property name in camelCase starts with a single letter
Question:
username_0: **Describe the bug**
JsonProperty is being ignored when property name in camelCase starts with a single letter.
**To Reproduce**
Here is an example with two properties of same type but different name. Kotlin compiler places @JsonProperty on same place: constructor param. However, only `originalName` property name is replaced but `aName` is not.
```kotlin
class JsonAnnotationTest {
private val mapper = ObjectMapper().registerModule(KotlinModule())
@Test
fun `JsonProperty annotation test`() {
assertThat(mapper.writeValueAsString(EntityWithJacksonAnnotation()))
.isEqualTo("""{"changedName1":"value","changedName2":"value"}""")
}
}
data class EntityWithJacksonAnnotation(
@JsonProperty("changedName1")
val originalName: String? = "value",
@JsonProperty("changedName2")
val aName: String? = "value"
)
```
**Expected behavior**
```
Expecting:
<"{"changedName1":"value","aname2":"value"}">
to be equal to:
<"{"changedName1":"value","changedName2":"value"}">
but was not.
Expected :{"changedName1":"value","changedName2":"value"}
Actual :{"changedName1":"value","aname2":"value"}
```
**Versions**
Kotlin: 1.3
Jackson-module-kotlin: 2.11.1
Jackson-databind: 2.11.1
Answers:
username_1: I have found the same issue too.
`MyClass(val eTag: String)` the eTag value is parsed as `etag` removing the capital letter T
This can be a big problem if a back-end sends me a json with capital letters in the second place of the properties, because Jackson can not find the property and fails. |
osmlab/name-suggestion-index | 370795906 | Title: Research needed for: amenity/pharmacy|The Generics Pharmacy
Question:
username_0: #### :thinking: We need help deciding what to do with "amenity/pharmacy|The Generics Pharmacy"
:point_right: Read the [CONTRIBUTING.md](https://github.com/osmlab/name-suggestion-index/blob/master/CONTRIBUTING.md) guide for more info, and feel free to ask questions on this issue!
### What's needed
First, do some research on `amenity/pharmacy|The Generics Pharmacy` by searching Google and Wikipedia.
#### If it is an actual brand:
* Edit `config/canonical.json`:
* Find the entry for `amenity/pharmacy|The Generics Pharmacy`
* Add tags for `brand:wikidata` and `brand:wikipedia`
* Optionally add `"countryCodes"` or other tags
* `npm run build`
* Commit your changes and submit a pull request
#### If it is not an actual brand:
* Edit `config/filters.json` to add an expression that discards it
* Edit `config/canonical.json` to remove the entry for `amenity/pharmacy|The Generics Pharmacy`
* `npm run build`
* Commit your changes and submit a pull request
#### If the situation is complicated:
* Leave a comment on this issue with what you find
:sparkles: Thanks!
_This is an autogenerated ticket_
Answers:
username_1: Closed in #2373
Status: Issue closed
|
vzaccaria/pandoc-mm | 185034427 | Title: The user has to be aware of placement distances to avoid text overlapping.
Question:
username_0: The user has to be aware of placement distances to avoid text overlapping.
Placement distance (or placement altogether) could probably be abstracted away from user responsibility. This is particularly annoying when adding new mind map entries - which causes all previous positioning to break.
Example:
Try adding the following line to to example given in the README to see text overlaps:
`* TEST`
Status: Issue closed
Answers:
username_0: Should be fixed now |
hookom/undead-darts | 295007807 | Title: Tooltips for stat descriptions
Question:
username_0: Already using React-Tooltip to display pointsOutOfFirst on playerName hover. Do the same with descriptions of the statistics on hover of the column headers. Save the descriptions as strings in one of the objects in helpers.js.
Answers:
username_0: Stat Value (which is already stored in an object in helpers.js) should also be displayed in the Tooltip
Status: Issue closed
|
developmentseed/osm-teams | 376538507 | Title: Implement authentication to API
Question:
username_0: We need to implement authentication to osm-teams, there are two main components:
- Authenticate requests to the "authorization server" by proving you are the resource owner, using your OSM id. This would be typically done using passport / openID in an express server. We need to figure out how to do this step using fastify. Once we have "logged in users" we can have them grant authorization to clients to sign requests on behalf of users. This is (B) in the diagram below
- Authenticate requests from client applications using access tokens. This means implementing an authorization server and the authorization code grant for down stream applications. It involves issuing tokens, and then protecting the API using those tokens.

## Possible implementation paths:
### Log in
- If we were to use passport, then we have an OAuth1.0 connect for OSM that we can re-use, we can just re-use the code in scoreboard
- If we don't use passport, we have to figure out how to adapt the OAuth1.0 connect code for the fastify server
### OAuth Server
- [OAuth2orize](https://github.com/jaredhanson/oauth2orize) is the passport way of implementing an OAuth2 server. It has a nice API (that I understand) and [nice examples](https://github.com/gerges-beshay/oauth2orize-examples). The problem is that it is a connect middleware and we chose fastify which is probably incompatible with it
- [https://github.com/ToonvanStrijp/fastify-oauth-server] is a fastify plugin for [oauth2-server](https://github.com/thomseddon/node-oauth2-server). I think this is going to be easier to integrate with fastify, but their API/docs are a bit obtuse and I have yet to see an example with login. I found an article [here](https://tech.zilverline.com/2017/03/17/nodejs-oauth2-provider) that could help but it's a bit short on details
## References:
- [RFC 6749 - Section 4.1: The Authorization Code Grant](https://tools.ietf.org/html/rfc6749.html#section-4.1)
- [RFC 6759 - Section 2: Authenticated Requests using Bearer tokens](https://tools.ietf.org/html/rfc6750.html#section-2)
- [OAuth2Server docs](https://oauth2-server.readthedocs.io/en/latest/api/oauth2-server.html)
- [OAuth2orize docs](https://github.com/jaredhanson/oauth2orize)
- [Building an OAuth2 server in node](https://tech.zilverline.com/2017/03/17/nodejs-oauth2-provider)
## Next steps
- [x] See if we can implement login using passport, if not then we need to figure out another way to login using osm ids #2
- [ ] See if we can adapt `oauth2orize`, if we can't then implement `oath2-server`<issue_closed>
Status: Issue closed |
getsentry/sentry-unity | 920470392 | Title: Special defaults for Desktop vs Mobile vs Xyz
Question:
username_0: We're adding defaults that make sense on Mobile. For example #230. This features does make sense on Desktop though so we could have different defaults for Mobile vs Desktop.
- [ ] On desktop environments default `IsEnvironmentUser` to true.
Answers:
username_0: We could just flip this flag in the Sentry native Windows support configure method.
So if a user opts-in to SendDefaultPii, we'll take the username of the user logged in to the machine. |
LeetCode-Feedback/LeetCode-Feedback | 733783335 | Title: 288. Unique Word Abbreviation - Explain the problem
Question:
username_0: #### Your LeetCode username
ajayvv
#### Category of the bug
- [x] Question
#### Description of the bug
The description is quite vague and doesn't mention what exactly is the expectation. Only after submitting and seeing various tests fail, we get to know what's expected. The following point is incomprehensible:
_"Else, for all words in dictionary such that their abbreviation is equal to the abbreviation of word those words are equal to word."_
Please make it clear, or better add at least one more meaningful example.
Answers:
username_1: Hi @username_0
Thank you for reaching out to us. I've relayed this issue to our team to investigate.
username_1: Hi @username_0
Thank you for your time. We've used your feedback to update the problem. Your LeetCode account has received 100 LeetCoins as the reward for this feedback. We appreciate your support!
Status: Issue closed
|
primefaces/primeng-quickstart-cli | 426493594 | Title: Can't resolve '@angular/cdk/scrolling'
Question:
username_0: ERROR in ./node_modules/primeng/components/dropdown/dropdown.js
Module not found: Error: Can't resolve '@angular/cdk/scrolling' in 'xx\primeng\node_modules\primeng\components\dropdown'
Answers:
username_1: run `npm install @angular/cdk --save`
Status: Issue closed
username_0: thanks, shouldn't be in the package json?
username_2: Please add this to readme
username_3: Just install a primeng version that is OK with your cdk, other way around. But changing your cdk version for primeng could mess with other dependencies and packages that you've already installed
username_4: thankyou..
username_5: Thanks!!!!!
username_6: Thanks
username_7: Thanks so much !
username_8: For me it didn't solve
Primeng "^9.0.0-rc.2"
username_9: ERROR in node_modules/@angular/cdk/coercion/array.d.ts(10,60): error TS1005: ',' expected.
node_modules/@angular/cdk/coercion/array.d.ts(10,61): error TS1005: ',' expected.
node_modules/@angular/cdk/coercion/array.d.ts(10,75): error TS1144: '{' or ';' expected.
node_modules/@angular/cdk/coercion/array.d.ts(10,77): error TS1011: An element access expression should take an argument.
username_10: how did you resolve this ? I am facing same issue
username_11: didn't solve For angular 9.0.6
I already installed angualr/cdk
still errors:
Failed to compile entry-point primeng/dropdown (es2015 as esm2015) due to compilation errors:
node_modules/@angular/cdk/scrolling/scrolling-module.d.ts:10:22 - error NG6002: Appears in the NgModule.imports of DropdownModule, but could not be resolved to an NgModule class.
This likely means that the library (@angular/cdk/scrolling) which declares ScrollingModule has not been processed correctly by ngcc, or is not compatible with Angular Ivy. Check if a newer version of the library is available, and update if so. Also consider checking with the library's authors to see if the library is expected to be compatible with Ivy.
10 export declare class ScrollingModule {
~~~~~~~~~~~~~~~
username_12: I would really like to use VirtualScrolling but I'm getting the same issue. I've tried installing/uninstalling angular cdk every possible way I can think of. Even deleting the whole node_modules folder and running `npm install` again doesn't fix it.
I gave up and tried to use the DataView component instead but it gives me similar errors. Since this issue is closed I think we need to open another one.
username_13: Failed to compiler angular project, how to resolver it??
Error: node_modules/primeng/dropdown/dropdown.d.ts:1:42 - error TS2307: Cannot find module '@angular/cdk/scrolling' or its corresponding type declarations.
1 import { CdkVirtualScrollViewport } from '@angular/cdk/scrolling'; |
rust-lang/rust-by-example | 261171821 | Title: Add implicit reference conversion to Strings section
Question:
username_0: Talk about functions that can accept
- `&str`
- `String`
- `T: Into<String>`
- `T: AsRef<str>`
- `T: Into<Cow<'a, str>>`
Based on
- http://hermanradtke.com/2015/05/06/creating-a-rust-function-that-accepts-string-or-str.html
- https://users.rust-lang.org/t/idiomatic-string-parmeter-types-str-vs-asref-str-vs-into-string/7934
Answers:
username_0: Not sure if this would be like #894 or just too much for this resource but I feel like this should be captured somewhere. |
dorseysen/One-Date-One-Question | 497531294 | Title: 2019-09-27:数组缺失成员 - 综合 (腾讯)。难度 ★★★☆
Question:
username_0: 数组缺失成员 - 综合 (腾讯)。
难度 ★★★☆
Answers:
username_0: ```js
// 2019-09-27:数组缺失成员 - 综合 (腾讯)
// 难度 ★★★☆
// 有一组数字,从1到n(假设n=10000),从中任意删除了3个数,顺序也被打乱,剩余数字放在一个n-3的数组里,请找出丢失的数字,要求算法比较快。
const handle = n => {
// 创建原数组
let arr = Array.from({length: n}, (item, index) => index + 1);
// 随机删除3个成员
let defectArr = arr.map(item => item).sort(() => Math.random() - 0.5).splice(0, n - 3);
// 找出丢失的数字
console.time('2019-09-27 算法题时间测试');
let map = {};
defectArr.forEach(item => map[item] = 1);
console.log(arr.filter(item => map[item] !== 1));
console.timeEnd('2019-09-27 算法题时间测试');
return '2019-09-27 算法题时间测试';
}
return handle(100000);
``` |
type-challenges/type-challenges | 1103767472 | Title: 11 - 元组转换为对象
Question:
username_0: <!--
小贴士:
🎉 恭喜你成功解决了挑战,很高兴看到你愿意分享你的答案!
由于用户数量的增加,Issue 池可能会很快被答案填满。为了保证 Issue 讨论的效率,在提交 Issue 前,请利用搜索查看是否有其他人分享过类似的档案。
你可以为其点赞,或者在 Issue 下追加你的想法和评论。如果您认为自己有不同的解法,欢迎新开 Issue 进行讨论并分享你的解题思路!
谢谢!
-->
```ts
// 你的答案
// 这里我看很多答案都是使用原来的 any[],但是这满足不了test-case那个应该报错的error case,这里应该约束元素里的类型仅能
// 为 string symbol number这三个可以为对象键的类型
type TupleToObject<T extends readonly (string | symbol | number)[]> = {
[P in T[number]]: P
}
```
Answers:
username_1: 但是keyof any出来的就是string | number | symbol啊,那为什么any不行(key of any)[]就行呢?
username_0: 确实,我也发现了这个有趣的写法,经过一些查阅,我觉得原理应该是这样的:keyof操作符是用于提取后面对象键的操作,那么keyof any,意思就是提取一个any类型的对象键,但是对象键只可能是 string number symbol,所以即使对于any,也就是任何类型进行keyof提取对象键,只可能出现这三种类型,大概是下面这个意思
```js
let a: any;
a['a'] //ok
a[0] // ok
a[Symbol()] //ok
a[{}] // error
```
所以在any上运行keyof,ts就是给出了全部可能的对象键类型 |
iptv-org/iptv | 936146681 | Title: Fix: Cheddar
Question:
username_0: <!-- Please fill out the information in this issue template so that we can
efficiently process your request -->
<!-- IMPORTANT: An issue may contain a request for only one channel, otherwise it will be closed -->
**_Channel Name:_** Cheddar
**_Broken Link (from playlist):_** https://dai2.xumo.com/amagi_hls_data_xumo1212A-redboxcheddar/CDN/playlist.m3u8?p=redbox&deviceid=&is_lat=&subp=RedboxdesktopWebWindows
**_Possible Replacement (optional):_**
**_Notes (optional):_** Will play in FFplay but not VLC
Status: Issue closed
Answers:
username_1: Closed in favor of #3457 |
googleads/googleads-perl-lib | 79160491 | Title: AdWords client is not compatible with perl v5.18
Question:
username_0: perl v5.18 changes the behavior of $class->isa. See search.cpan.org/dist/perl-5.18.0/pod/perldelta.pod, search for #47113 and http://www.nntp.perl.org/group/perl.perl5.porters/2012/06/msg188459.html. Google/Ads/SOAP/Typelib/ComplexType.pm, line 87, uses $type->isa('UNIVERSAL') to determine whether or not to load the $type module. Perl 5.18 has changed to return 1 whether or not the $type module is loaded.
This issue can be demonstrated in a couple ways:
1) Demonstrating how perl changed
```
jpickard ~/tmp/AdWords $ perl -v
This is perl 5, version 18, subversion 2 (v5.18.2) built for x86_64-linux
Copyright 1987-2013, <NAME>
Perl may be copied only under the terms of either the Artistic License or the
GNU General Public License, which may be found in the Perl 5 source kit.
Complete documentation for Perl, including FAQ lists, should be found on
this system using "man perl" or "perldoc perl". If you have access to the
Internet, point your browser at http://www.perl.org/, the Perl Home Page.
jpickard ~/tmp/AdWords $ perl -e'$type = "Google::Ads::AdWords::v201502::Budget::BudgetDeliveryMethod"; print $type->isa("UNIVERSAL") ? "loaded\n" : "NOT loaded\n"'
loaded
jpickard ~/tmp/AdWords $ export PATH=/usr/local/bin:/usr/bin:/bin:/opt/bin:/usr/x86_64-pc-linux-gnu/gcc-bin/4.6.3:/opt/hadoop/bin:/opt/node/bin
jpickard ~/tmp/AdWords $ perl -v
This is perl 5, version 16, subversion 3 (v5.16.3) built for x86_64-linux
(with 22 registered patches, see perl -V for more detail)
Copyright 1987-2012, <NAME>
Perl may be copied only under the terms of either the Artistic License or the
GNU General Public License, which may be found in the Perl 5 source kit.
Complete documentation for Perl, including FAQ lists, should be found on
this system using "man perl" or "perldoc perl". If you have access to the
Internet, point your browser at http://www.perl.org/, the Perl Home Page.
jpickard ~/tmp/AdWords $ perl -e'$type = "Google::Ads::AdWords::v201502::Budget::BudgetDeliveryMethod"; print $type->isa("UNIVERSAL") ? "loaded\n" : "NOT loaded\n"'
NOT loaded
jpickard ~/tmp/AdWords $
```
2) Demonstrating how AdWords code has this issue. v5.18 dies because a module was not loaded. v5.16 dies after that when making a call to Google:
```
jpickard ~/tmp/AdWords $ perl -v
This is perl 5, version 18, subversion 2 (v5.18.2) built for x86_64-linux
Copyright 1987-2013, <NAME>
Perl may be copied only under the terms of either the Artistic License or the
GNU General Public License, which may be found in the Perl 5 source kit.
Complete documentation for Perl, including FAQ lists, should be found on
this system using "man perl" or "perldoc perl". If you have access to the
Internet, point your browser at http://www.perl.org/, the Perl Home Page.
jpickard ~/tmp/AdWords $ perl googleads-perl-lib/examples/v201502/campaign_management/get_campaigns_by_label.pl
Use of uninitialized value in concatenation (.) or string at /opt/perl5/perls/perl-5.18.2/lib/site_perl/5.18.2/Google/Ads/Common/HTTPTransport.pm line 32.
Can't locate object method "new" via package "Google::Ads::AdWords::v201502::Predicate::Operator" (perhaps you forgot to load "Google::Ads::AdWords::v201502::Predicate::Operator"?) at /opt/perl5/perls/perl-5.18.2/lib/site_perl/5.18.2/Google/Ads/SOAP/Typelib/ComplexType.pm line 119.
jpickard ~/tmp/AdWords $ export PATH=/usr/local/bin:/usr/bin:/bin:/opt/bin:/usr/x86_64-pc-linux-gnu/gcc-bin/4.6.3:/opt/hadoop/bin:/opt/node/bin
jpickard ~/tmp/AdWords $ perl -v
This is perl 5, version 16, subversion 3 (v5.16.3) built for x86_64-linux
(with 22 registered patches, see perl -V for more detail)
Copyright 1987-2012, <NAME>
Perl may be copied only under the terms of either the Artistic License or the
GNU General Public License, which may be found in the Perl 5 source kit.
Complete documentation for Perl, including FAQ lists, should be found on
this system using "man perl" or "perldoc perl". If you have access to the
Internet, point your browser at http://www.perl.org/, the Perl Home Page.
jpickard ~/tmp/AdWords $ perl googleads-perl-lib/examples/v201502/campaign_management/get_campaigns_by_label.pl
Use of uninitialized value in concatenation (.) or string at /usr/lib64/perl5/vendor_perl/5.16.3/Google/Ads/Common/HTTPTransport.pm line 32.
[21 May 2015 16:38:07,353 - INFO ] Outgoing Request:
<SOAP-ENV:Envelope xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:SOAP-ENV="http://schemas.xmlsoap.org/soap/envelope/" ><SOAP-ENV:Header><RequestHeader xmlns="https://adwords.google.com/api/adwords/cm/v201502"><userAgent xmlns="https://adwords.google.com/api/adwords/cm/v201502">googleads-perl-lib/examples/v201502/campaign_management/get_campaigns_by_label.pl (AwApi-Perl/3.3.0, Common-Perl/3.1.1, SOAP-WSDL/2.00.10, libwww-perl/6.05, perl/5.016003)</userAgent><validateOnly xmlns="https://adwords.google.com/api/adwords/cm/v201502">false</validateOnly><partialFailure xmlns="https://adwords.google.com/api/adwords/cm/v201502">false</partialFailure></RequestHeader></SOAP-ENV:Header><SOAP-ENV:Body><get xmlns="https://adwords.google.com/api/adwords/cm/v201502"><serviceSelector><fields>Id</fields><fields>Name</fields><fields>Labels</fields><predicates><field>Labels</field><operator>CONTAINS_ANY</operator><values>INSERT_LABEL_ID_HERE</values></predicates><ordering><field>Name</field><sortOrder>ASCENDING</sortOrder></ordering><paging><startIndex>0</startIndex><numberResults>500</numberResults></paging></serviceSelector></get></SOAP-ENV:Body></SOAP-ENV:Envelope>
The library couldn't find any authorization mechanism set up to properly sign the requests against the API. Please read the following guide on how to setup OAuth2 https://github.com/googleads/googleads-perl-lib/wiki/Using-OAuth-2.0 at /usr/lib64/perl5/vendor_perl/5.16.3/Google/Ads/Common/HTTPTransport.pm line 47.
```
Answers:
username_1: We're looking into fixing this issue. Thank you for providing such a thorough and detailed explanation of the issue. It was extremely helpful.
Cheers,
<NAME>, AdWords API Team
username_1: This is now fixed with version 4.0.0.
Status: Issue closed
|
html-next/flexi | 266342964 | Title: Error: Cannot find module 'debug'
Question:
username_0: When I carried out “npm run server” in windows7,the page show the following error
Error: Cannot find module 'debug'
at Function.Module._resolveFilename (module.js:527:15)
at Function.Module._load (module.js:476:23)
at Module.require (module.js:568:17)
at require (internal/module.js:11:18)
at Object.<anonymous> (C:\Users\xxx\node_modules\[email protected]@s
elenium-standalone\bin\selenium-standalone:3:13)
at Module._compile (module.js:624:30)
at Object.Module._extensions..js (module.js:635:10)
at Module.load (module.js:545:32)
at tryModuleLoad (module.js:508:12)
at Function.Module._load (module.js:500:3)
npm ERR! code ELIFECYCLE
npm ERR! errno 1
npm ERR! [email protected] server: `selenium-standalone start`
npm ERR! Exit status 1
npm ERR!
npm ERR! Failed at the [email protected] server script.
npm ERR! This is probably not a problem with npm. There is likely additional log
ging output above.
npm ERR! A complete log of this run can be found in:
npm ERR! C:\Users\xxx\AppData\Roaming\npm-cache\_logs\2017-10-18T01_35_34_529
Z-debug.log
Answers:
username_1: @username_0 can you give more details please? An empty description does not help us reproduce.
Status: Issue closed
|
faisalman/ua-parser-js | 638907373 | Title: undefined
Question:
username_0: { vendor: undefined, model: undefined, type: undefined } Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/81.0.4044.138 Safari/537.36
why?
Status: Issue closed
Answers:
username_1: Answered here: https://github.com/username_1/ua-parser-js/issues/182 |
influxdata/helm-charts | 916785964 | Title: S3 optional endpoint-url is not optional
Question:
username_0: In Values.yaml
```
## Optional. Specify if you're using an alternate S3 endpoint.
```
But if you don't specify an `endpoint-url` this is what you get:
```
aws: error: argument --endpoint-url: expected one argument
```
If optional, [this line](https://github.com/influxdata/helm-charts/blob/0b7a0a1c8d590f7a27f88b855464945b8e40bb01/charts/influxdb/files/backup-retention-script.sh#L59) should be modified to prevent this error. Happy to submit a PR if you agree with that.
Answers:
username_1: Apologies for the delayed response. PR is always appreciated :) |
SerenityOS/serenity | 684244367 | Title: LibTLS: Record MAC check is completely broken
Question:
username_0: This has possibly been broke from the start, and just brought to light by 4acdb60.
repro: try to open any HTTPS connection. it _should_ fail with "integrity check failed".
The aforementioned equality check:
https://github.com/SerenityOS/serenity/blob/f0ef283f3c6c16e593a6773e9b9ea542507db43b/Libraries/LibTLS/Record.cpp#L242
It should be noted that while there _is_ a logical slice error, it does not seem to be the only issue.
(`decrypted_span` contains the MAC itself, it should not)
https://github.com/SerenityOS/serenity/blob/f0ef283f3c6c16e593a6773e9b9ea542507db43b/Libraries/LibTLS/Record.cpp#L240<issue_closed>
Status: Issue closed |
WikiWatershed/model-my-watershed | 939007671 | Title: Upgrade Flake8, Improve API Docs
Question:
username_0: As of [`d854ca9` (#3399)](https://github.com/WikiWatershed/model-my-watershed/pull/3399/commits/d854ca98e8b4c65d87da4ec5e77e491bfdcf174c), we have fixed the broken links in the technical documentation. Unfortunately, that requires very long lines, which trips flake8 warnings, which are currently ignored with `# NOQA`. Unfortunately, that ignore message shows up in the API docs:

The way to ignore this cleanly is to use `per-file-ignores` in the flake8 configuration, as described in the docs: https://flake8.pycqa.org/en/latest/user/options.html#cmdoption-flake8-per-file-ignores. This new option is only available in flake8 3.7+, and we are currently on 2.2.5, with the latest being 3.9.2.
- [ ] Upgrade flake8 to the latest possible (given the Python 2.7 requirement)
- [ ] Add a flake8 configuration file ignoring `E501` errors in `geoprocessing_api/views.py`
- [ ] Remove the `# NOQA` lines from the API docs |
ISWPOLI/Poli-antifraude | 211768643 | Title: Generar preguntas de seguridad
Question:
username_0: Rol: Como un administrativo de la universidad PG.
Caracteristica/Funcionalidad: Necesito que el sistema poli Antifraude me genere preguntas de seguridad personalizadas
Razon/Finalidad: Con la finalidad de poder hacer un reconocimiento de la persona mediante uso de su informacion personal.
Answers:
username_1: - [ ] Formulario con un máximo de 5 preguntas aleatorias de las preguntas generadas en la DB
Status: Issue closed
|
cgpu/ion-somatic-variant-calling | 480552177 | Title: Consider adjusting very strict discovery parameters: --max_alt_alleles_in_normal_count --max_alt_allele_in_normal_fraction
Question:
username_0: `--max_alt_alleles_in_normal_count 1000`
`--max_alt_allele_in_normal_fraction 0.10`
https://best-practices-for-processing-hts-data.readthedocs.io/en/latest/mutect2_pitfalls.html
 |
akshitadixit/Structurex | 801229433 | Title: Add Speed and Number of bars to the Sorting Visualizer
Question:
username_0: I would like to work on this issue. Please assign this to me
Answers:
username_0: I would like to work on this issue. Please assign this to me
username_1: Sure. But did you activate gh-pages on your repo? Please do that so it is easier for me to view changes.
username_0: Yes i had activated the gh-pages but that branch was deleted.
now i have changed the branch. Sorry for the inconvenience caused.
Status: Issue closed
|
graphql-python/graphene | 230016442 | Title: SQLAlchemy and Nested Schema
Question:
username_0: I have two models, User and Deal.
```python
class User(SQLAlchemyObjectType):
class Meta:
model = UserModel
class Deal(SQLAlchemyObjectType):
class Meta:
model = DealModel
```
Now I can query the DB using Graphql. One such query is:
```graphql
query {
trendingDeals {
title
user {
name
location
deals {
title
}
}
}
}
```
I can keep on nesting fields and SQLAlchemy-Graphene will keep on traversing the relationships associated with them. However, I do not want the **user** field of **Deal** model to contain the whole User schema. Instead, only some particular fields. But, if I use `exclude_fields` in the **User** schema, that field is becomes hidden. I want to traverse the said field when the root query is just `users`. Like this -
This should be valid.
```graphql
{
user {
name
deals {
title
}
}
}
```
This should not be valid.
```graphql
query {
Deals {
title
user {
name
location
deals { # user should not contain the deals field when nested in some other query
title
}
}
}
}
```
In essence, how can I give alternate representation of some fields in a model without remapping the whole model to a new schema.
Is there any way?
Thanks.
Answers:
username_1: @username_0 Same here, i want to restrict the fields in the nested query. But how to do? Any Sol?
username_2: I worked around this by setting up my own resolvers for those fields that used DataLoader to batch the queries together. Then nested, redundant fields would use the same data, avoiding the N+1 issue.
Status: Issue closed
username_3: This issue looks like it should be raised on [graphene-sqlalchemy](https://github.com/graphql-python/graphene-sqlalchemy) |
tensorflow/tensorflow | 597149574 | Title: Google Developers Certification
Question:
username_0: Google Developers Certification plugin is not available in PyCharm 2020.1 (it is needed to take the Tensorfow Certification exam):

Answers:
username_1: I faced the same issue. Downgraded PyCharm to pycharm-community-2019.3.4
username_2: Hello,
The Google Developers Certification plugin is now available in PyCharm 2020.
Status: Issue closed
|
dbeaver/dbeaver | 984825453 | Title: Hana: create procedure fails with syntax error at END if there is a loop inside.
Question:
username_0: #### System information:
- Win10
- DBeaver 21.2.0
#### Connection specification:
Sap Hana
#### Describe the problem you're observing:
Simple procedure cannot be created due to supposed syntax errors.
#### Steps to reproduce, if exist:
Run the SQL Script:
```
create or replace procedure abc
(in x SMALLINT,
in y SMALLINT)
as
begin
declare loop_x SMALLINT;
declare loop_y SMALLINT;
declare maxID BIGINT;
for loop_y in 1..:y do
for loop_x in 1..:x do
maxID := maxID + 2;
end for;
end for;
end;
```
It seems the parser sends only the string
```
create or replace procedure abc
(in x SMALLINT,
in y SMALLINT)
as
begin
declare loop_x SMALLINT;
declare loop_y SMALLINT;
declare maxID BIGINT;
for loop_y in 1..:y do
for loop_x in 1..:x do
maxID := maxID + 2;
end for
```
truncating the statement after the end-for's semicolon char.
Error message is
```
SQL Error [257] [HY000]: SAP DBTech JDBC: [257] (at 302): sql syntax error: line 14 col 13 (at pos 302)
```
and the create statement is obviously valid, it works in Hana Studio.
Answers:
username_1: Hello @username_0
Thanks for the bug report!
You can try to highlight your procedure statement in SQL Editor and use Ctrl+Enter to avoid block parsing problems.
Now we, unfortunately, cannot test this case, because the connection to the database you gave us is no longer working. If you have the opportunity to back our access to the database - it would be very cool!
username_0: I have shutdown the server but can start it at any time. It is just that I do no longer need it 24x7. Please send me an email when and for how long I should start the server and I will do so.
username_1: Thanks for your help. Your database was very useful for us.
Anyway. I see a blank line in your statement.
Can you please tell me - the "Blank line is statement delimiter" is enabled or disabled in your preferences?

username_0: Yes, it was checked. It tried with and without (opening a new editor after the change to be safe).
I also removed the blank lines from the procedure code.
Behavior is always the same, the code sent to the database ends at the line `end for;`
username_1: Fixed in DBeaver 21.2.2
Status: Issue closed
username_2: Veified |
open-gamma-ray-astro/gamma-astro-data-formats | 344735343 | Title: Optional column for a gammaness value
Question:
username_0: Most experiments assign each event a score that represents how likely an event is a gamma-ray. (which is not-necessearily or even more commonly not a probability)
I think there should be an optional column in the standard for this parameter and possibly also responses for different thresholds in this parameter.
Answers:
username_1: See #34 .
Can you please look at that, and if you think for v0.2 something should be added or changed, send a PR?
Otherwise I'd suggest you comment on #34 and we close this one as duplicate and leave #34 open as future work.
Status: Issue closed
username_1: Closing this issue now. Let's continue the discussion in #34 . |
AdoptOpenJDK/openjdk-docker | 640164341 | Title: Ubuntu 20.04
Question:
username_0: Now that Ubuntu 20.04 LTS is out, what's the roadmap towards updating the Ubuntu based images to use the current LTS?
Answers:
username_1: @dinogun Any way we can get some nightly images build using the Ubuntu 20.04 base?
username_2: This came up recently when a user of a downstream image was unable to use GitHub actions because the Git that comes from the 18.04 repo was too old.
https://github.com/username_2/docker-gradle/issues/151
Status: Issue closed
|
ElektraInitiative/libelektra | 527562407 | Title: Build: AWK Warnings
Question:
username_0: # Steps to Reproduce the Problem
1. Build the [Docker image for Debian sid](https://github.com/ElektraInitiative/libelektra/blob/master/scripts/docker/README.md)
2. Run the container
3. Build Elektra using the commands below:
```sh
mkdir build
cmake -G Ninja -B build
cmake --build build
```
# Expected Result
The last command does not print any warning or error messages.
# Actual Result
The command `cmake --build build` prints the following warning messages:
```
Generating ../../../../include/gen/templates.hpp
awk: cmd. line:12: warning: regexp escape sequence `\"' is not a known regexp operator
awk: cmd. line:12: warning: regexp escape sequence `\"' is not a known regexp operator
awk: cmd. line:12: warning: regexp escape sequence `\"' is not a known regexp operator
awk: cmd. line:12: warning: regexp escape sequence `\"' is not a known regexp operator
awk: cmd. line:12: warning: regexp escape sequence `\"' is not a known regexp operator
awk: cmd. line:12: warning: regexp escape sequence `\"' is not a known regexp operator
awk: cmd. line:12: warning: regexp escape sequence `\"' is not a known regexp operator
awk: cmd. line:12: warning: regexp escape sequence `\"' is not a known regexp operator
awk: cmd. line:12: warning: regexp escape sequence `\"' is not a known regexp operator
awk: cmd. line:12: warning: regexp escape sequence `\"' is not a known regexp operator
awk: cmd. line:12: warning: regexp escape sequence `\"' is not a known regexp operator
awk: cmd. line:12: warning: regexp escape sequence `\"' is not a known regexp operator
awk: cmd. line:12: warning: regexp escape sequence `\"' is not a known regexp operator
awk: cmd. line:12: warning: regexp escape sequence `\"' is not a known regexp operator
```
# System Information
- Elektra Version: `master`
- Docker `2.1.0.5`
# Further Information
As far as I can tell the problem is the line:
https://github.com/ElektraInitiative/libelektra/blob/ba8ffe00c0735610a4b6c5f4af57e6ca7c207dfd/src/tools/kdb/gen/templates/collect.sh#L18
. Looks like we should remove the `\` before the first `"`. I am not sure that is the right option though, since I do not understand why the character set (`[]`) contains multiple double and single quotes.
Answers:
username_1: Thank you for reporting the problem!
I do not know if this is a "good first issue" if you are not sure yourself what is meant?
username_0: I am pretty sure it would be a “good first issue“ for an `awk` programmer 😊. Anyway, I removed the label for now.
username_2: The line in question is part of a single quoted string in a shell script. Therefore the sequence `'"'"'` closes the single quoted string `'`, opens a double quoted one `"`, adds a single quote to the double quoted string `'`, closes the double quoted string `"` and opens a single quoted string `'`. Together this just adds the a single quote to the single quoted string.
The double quote before that is there to add a double quote to the character set.
It seems the warning wants us to replace `\"` with `"`. If we do that we need to make sure that it still works with both `gawk` and `mawk`. Sometimes they disagree. The POSIX standard says `\"` is correct. ([here](https://pubs.opengroup.org/onlinepubs/9699919799/utilities/awk.html) the table under Regular Expressions)
username_1: If the POSIX standard says contrary to the warning it might be better to ignore the warning.
username_0: While following POSIX sounds good, at least the `awk` implementations I tested (`nawk`, `gawk`, `bioawk`, and `mawk`) seem to have no problem with unescaped double quotes. At least the script
```sh
#!/usr/bin/env sh
# BWK awk/nawk/
# One true awk gawk
set -- /usr/bin/awk /usr/local/bin/awk bioawk mawk
for awk; do
printf '"bla\nblubb' | "$awk" '$0 ~ /["].*/ { print $0 }'
done
```
prints the same line:
```
"bla
```
four times on my machine. I already removed the backslash locally and will probably open a pull request that contains this minor “fix” and some other minor improvements in the next days.
username_1: Good! :sparkle:
Status: Issue closed
|
SahilJ97/Explainable-Stance-Detection | 1047593165 | Title: Installing requirements.txt fails
Question:
username_0: Installing the provided requirements.txt fails due to dependency problems.
I was able to install by changing the versions of following libraries:
```
tensorboard==2.3.0
numpy==1.17.0
```
Answers:
username_1: @username_0 Thanks for bringing that to my attention! I'll update requirements.txt accordingly.
Status: Issue closed
|
firemodels/fds | 455882866 | Title: mpiexec.c (1901): assert (exitcodes !=NULL) failed
Question:
username_0: Error encountered running sample case with 4 meshes, 3 mpi processes. Following is error:
type helpfds for help on running fds
C:\Users\firemodel>cd\fds6_projects
C:\FDS6_projects>cd mpi_test
C:\FDS6_projects\mpi_test>where mpiexec
C:\Program Files\firemodels\FDS6\bin\mpi\mpiexec.exe
C:\FDS6_projects\mpi_test>mpiexec -hosts 3 firemodel-7259 1 firemodel-7260 1 firemodel-7261 1 test_mpi
Hello world: rank 0 of 3 running on
FIREMODEL-7259
Hello world: rank 1 of 3 running on
FIREMODEL-7260
Hello world: rank 2 of 3 running on
FIREMODEL-7261
C:\FDS6_projects\mpi_test>mpiexec -n 3 firemodel-7259 2 firemodel-7260 1 firemodel-7261 1 -env OMP_NUM_THREADS 4 fds layer_4mesha.fds
[mpiexec@FIREMODEL-7259] mpiexec.c (1901): assert (exitcodes != NULL) failed
The error is consistent.
FDS file is modification of layer_4mesh.fds
&MESH ID='Mesh01', IJK=20,20,5, XB=0.0,2.0,0.0,2.0,0.0,0.5, MPI_PROCESS=0/
&MESH ID='Mesh02', IJK=20,20,5, XB=0.0,2.0,0.0,2.0,0.5,1.0, MPI_PROCESS=0/
&MESH ID='Mesh03', IJK=20,20,5, XB=0.0,2.0,0.0,2.0,1.0,1.5, MPI_PROCESS=1/
&MESH ID='Mesh04', IJK=20,20,5, XB=0.0,2.0,0.0,2.0,1.5,2.0, MPI_PROCESS=2/
System information as follows:
• OS is Windows 10
• FDS 6.7.1
• The working directory is shared with the ‘Everyone’ with full control;
• PyroSim hydra service is stopped for all;
Help appreciated.
Answers:
username_1: Can you successfully run this case using 4 MPI processes?
username_1: Why are you allocating 4 processes for this job?
```
C:\FDS6_projects\mpi_test>mpiexec -n 3 firemodel-7259 2 firemodel-7260 1 firemodel-7261 1 -env OMP_NUM_THREADS 4 fds layer_4mesha.fds
```
username_0: The case is a sample from FDS downloads.
The example has 4 meshes. It just happens that we have 4 identical computers in our cluster and I intend to setup and test the parallel processing with these machines.
When I test run the test_mpi as shown in the earlier attached screen shot, it was a smooth run with 3 machines. Once I called the 4th machine, the last process (not necessarily any particular machine) took much longer time to return and the system hangs without going back to the DOS prompt. Following is the screen shot.
*************************************
type helpfds for help on running fds
C:\Users\firemodel>mpiexec -hosts 4 firemodel-7259 1 firemodel-7260 1 firemodel-7261 1 firemodel-7262 1 test_mpi
Hello world: rank 0 of 4 running on
FIREMODEL-7259
Hello world: rank 1 of 4 running on
FIREMODEL-7260
Hello world: rank 2 of 4 running on
FIREMODEL-7261
Hello world: rank 3 of 4 running on
FIREMODEL-7262
**************************************
So I decided to use 3 machines to run the 4 processes, also intending to test out an unbalanced process assignment between machines.
I have just tested with 3 meshes assigned to 3 processes, yet I have the same error.
<NAME> M.Sc., P.Eng., CP
Senior Associate
4th Floor, 780 Beatty Street
Vancouver, BC V6B 2M1
[LMDG 30-year logo larger font 300dpi]
604-682-7146 ext 410
<EMAIL><mailto:<EMAIL>>
www.lmdg.com<http://www.lmdg.com/>
[96dpi OQM-certified-wordmark-FINAL]
username_1: Can you run the job with 4 MPI processes and 1 OpenMP thread per MPI process?
username_0: I have the same error message:
*********************************
type helpfds for help on running fds
C:\Users\firemodel>cd\fds6_projects\mpi_test
C:\FDS6_projects\mpi_test>mpiexec -n 4 firemodel-7259 2 firemodel-7260 1 firemodel-7261 1 -env OMP_NUM_THREADS 1 fds layer_4mesh.fds
[mpiexec@FIREMODEL-7259] mpiexec.c (1901): assert (exitcodes != NULL) failed
C:\FDS6_projects\mpi_test>mpiexec -n 4 firemodel-7259 1 firemodel-7260 1 firemodel-7261 1 firemodel-7262 1 -env OMP_NUM_THREADS 1 fds layer_4mesh.fds
[mpiexec@FIREMODEL-7259] mpiexec.c (1901): assert (exitcodes != NULL) failed
C:\FDS6_projects\mpi_test>
***********************************
Below is the input file:
**************************************
layer_4mesha.fds
Generated by PyroSim - Version 2019.1.0515
Jun 11, 2019 10:25:17 AM
&HEAD CHID='layer_4mesha', TITLE='layer_4mesh'/
&TIME T_END=60.0/
&DUMP RENDER_FILE='layer_4mesha.ge1', DT_DEVC=2.0/
&MESH ID='Mesh01', IJK=20,20,5, XB=0.0,2.0,0.0,2.0,0.0,0.5, MPI_PROCESS=0/
&MESH ID='Mesh02', IJK=20,20,5, XB=0.0,2.0,0.0,2.0,0.5,1.0, MPI_PROCESS=1/
&MESH ID='Mesh03', IJK=20,20,5, XB=0.0,2.0,0.0,2.0,1.0,1.5, MPI_PROCESS=2/
&MESH ID='Mesh04', IJK=20,20,5, XB=0.0,2.0,0.0,2.0,1.5,2.0, MPI_PROCESS=3/
&REAC ID='Reaction1',
FUEL='PROPANE',
SOOT_YIELD=0.01,
RADIATIVE_FRACTION=0.35/
&DEVC ID='z_int', QUANTITY='LAYER HEIGHT', XB=1.8,1.8,1.8,1.8,0.0,2.0/
&DEVC ID='T_up', QUANTITY='UPPER TEMPERATURE', XB=1.8,1.8,1.8,1.8,0.0,2.0/
&DEVC ID='T_low', QUANTITY='LOWER TEMPERATURE', XB=1.8,1.8,1.8,1.8,0.0,2.0/
&SURF ID='FIRE',
COLOR='RED',
HRRPUA=200.0/
&VENT ID='Vent #1', SURF_ID='OPEN', XB=0.0,0.0,0.4,1.6,0.8,1.2/
&VENT ID='Vent #2', SURF_ID='OPEN', XB=2.0,2.0,0.4,1.6,0.8,1.2/
&VENT ID='Vent #3', SURF_ID='OPEN', XB=0.4,1.6,0.0,0.0,0.8,1.2/
&VENT ID='Vent #4', SURF_ID='OPEN', XB=0.4,1.6,2.0,2.0,0.8,1.2/
&VENT ID='Vent #5', SURF_ID='FIRE', XB=0.8,1.2,0.8,1.2,0.0,0.0/
&SLCF QUANTITY='TEMPERATURE', VECTOR=.TRUE., PBY=1.0/
&TAIL /
****************************************************
I imported the example into PyroSim to view and adjust the mpi_process. Then I exported it back to FDS to run.
I have the same error with the original fds without going through PyroSim.
<NAME> M.Sc., P.Eng., CP
Senior Associate
4th Floor, 780 Beatty Street
Vancouver, BC V6B 2M1
[LMDG 30-year logo larger font 300dpi]
604-682-7146 ext 410
<EMAIL><mailto:<EMAIL>>
www.lmdg.com<http://www.lmdg.com/>
[96dpi OQM-certified-wordmark-FINAL]
username_1: You need the option -wdir to indicate a shared directory where the input file resides. The other computers do not know where to write output.
username_0: I am afraid I have the same error.
C:\FDS6_projects\mpi_test>mpiexec -n 4 firemodel-7259 2 firemodel-7260 1 firemodel-7261 1 -wdir \\firemodel-7259\mpi_test -env OMP_NUM_THREADS 1 fds layer_4meshd.fds
[mpiexec@FIREMODEL-7259] mpiexec.c (1901): assert (exitcodes != NULL) failed
C:\FDS6_projects\mpi_test>mpiexec -n 4 firemodel-7259 2 firemodel-7260 1 firemodel-7261 1 -wdir \\firemodel-7259\fds6_projects\mpi_test\ -env OMP_NUM_THREADS 1 fds layer_4meshd.fds
[mpiexec@FIREMODEL-7259] mpiexec.c (1901): assert (exitcodes != NULL) failed
C:\FDS6_projects\mpi_test>mpiexec -n 4 firemodel-7259 2 firemodel-7260 1 firemodel-7261 1 -wdir \\firemodel-7259\mpi_test\ -env OMP_NUM_THREADS 1 fds layer_4meshd.fds
[mpiexec@FIREMODEL-7259] mpiexec.c (1901): assert (exitcodes != NULL) failed
C:\FDS6_projects\mpi_test>mpiexec -n 4 firemodel-7259 2 firemodel-7260 1 firemodel-7261 1 -wdir \\10.2.1.98\mpi_test\ -env OMP_NUM_THREADS 1 fds layer_4meshd.fds
[mpiexec@FIREMODEL-7259] mpiexec.c (1901): assert (exitcodes != NULL) failed
C:\FDS6_projects\mpi_test>mpiexec -n 4 firemodel-7259 2 firemodel-7260 1 firemodel-7261 1 -wdir \\10.2.1.98\fds6_projects\mpi_test\ -env OMP_NUM_THREADS 1 fds layer_4meshd.fds
[mpiexec@FIREMODEL-7259] mpiexec.c (1901): assert (exitcodes != NULL) failed
C:\FDS6_projects\mpi_test>mpiexec -n 4 firemodel-7259 2 firemodel-7260 1 firemodel-7261 1 -wdir \\10.2.1.98\fds6_projects\mpi_test -env OMP_NUM_THREADS 1 fds layer_4meshd.fds
[mpiexec@FIREMODEL-7259] mpiexec.c (1901): assert (exitcodes != NULL) failed
C:\FDS6_projects\mpi_test>
<NAME> M.Sc., P.Eng., CP
Senior Associate
4th Floor, 780 Beatty Street
Vancouver, BC V6B 2M1
[LMDG 30-year logo larger font 300dpi]
604-682-7146 ext 410
<EMAIL><mailto:<EMAIL>>
www.lmdg.com<http://www.lmdg.com/>
[96dpi OQM-certified-wordmark-FINAL]
username_1: If you login to all computers, can they "see" your working directory, using the exact form of the address?
Can you run the job like this
```
mpiexec -localonly -n 4 fds jobname.fds
```
Do you have the same version of FDS installed on all computers?
You might want to open a Discussion thread. I am running out of ideas. Running FDS across multiple Windows computers is tricky. Most people use linux clusters to run FDS in parallel.
username_0: Hi Kevin
Thanks very much for your help. Just before I go, I tested the mpiexec localonly. It returns the same error message. Apparently, I am not able to run mpiexec other than the test_mpi that came with the installation.
I am able to run fds_local.
All the computers have the same version of FDS and they can see the working directory.
If you can throw me some more ideas before I turn to the discussion group.
Thans.
************************************
type helpfds for help on running fds
C:\Users\firemodel>cd\fds6_projects\mpi_test
C:\FDS6_projects\mpi_test>dir
Volume in drive C is Windows
Volume Serial Number is 5463-D411
Directory of C:\FDS6_projects\mpi_test
06/14/2019 10:29 AM <DIR> .
06/14/2019 10:29 AM <DIR> ..
06/11/2019 10:25 AM 1,313 layer_4mesha.fds
06/13/2019 02:38 PM 967 layer_4meshd.fds
2 File(s) 2,280 bytes
2 Dir(s) 407,688,237,056 bytes free
C:\FDS6_projects\mpi_test>mpiexec localonly -n 4 fds layer_4meshd.fds
[mpiexec@FIREMODEL-7260] mpiexec.c (1901): assert (exitcodes != NULL) failed
C:\FDS6_projects\mpi_test>
*********************************************
<NAME> M.Sc., P.Eng., CP
Senior Associate
4th Floor, 780 Beatty Street
Vancouver, BC V6B 2M1
[LMDG 30-year logo larger font 300dpi]
604-682-7146 ext 410
<EMAIL><mailto:<EMAIL>>
www.lmdg.com<http://www.lmdg.com/>
[96dpi OQM-certified-wordmark-FINAL]
username_1: What do you see when you type:
```
where mpiexec
```
The command
```
fds_local -p 4 job_name.fds
```
is equivalent to
```
mpiexec -n 4 -localonly fds job_name.fds
```
Both should work for you, or neither should. You need not provide login credentials to run these commands.
username_0: Response to your earlier email:
1. where mpiexec shows the correct location of the exe file
2. fds_local works fine
3. mpiexec localonly returns the error
****************************************
type helpfds for help on running fds
C:\Users\firemodel>cd\fds6_projects
C:\FDS6_projects>cd mpi_test
C:\FDS6_projects\mpi_test>where mpiexec
C:\Program Files\firemodels\FDS6\bin\mpi\mpiexec.exe
C:\FDS6_projects\mpi_test>mpiexec localonly -n 4 fds layer_4meshd.fds
[mpiexec@FIREMODEL-7259] mpiexec.c (1901): assert (exitcodes != NULL) failed
C:\FDS6_projects\mpi_test>fds_local layer_4meshd.fds
Reading input file ...
WARNING: MPI_PROCESS set for MESH 2 and only one MPI process exists
WARNING: MPI_PROCESS set for MESH 3 and only one MPI process exists
WARNING: MPI_PROCESS set for MESH 4 and only one MPI process exists
Fire Dynamics Simulator
Current Date : June 14, 2019 12:04:20
Revision : FDS6.7.1-0-g14cc738-HEAD
Revision Date : Mon Feb 4 12:26:25 2019 -0500
Compiler : Intel ifort 172.16.17.32
Compilation Date : Wed 02/13/2019 03:33 PM
MPI Enabled; Number of MPI Processes: 1
OpenMP Enabled; Number of OpenMP Threads: 4
MPI version: 3.1
MPI library version: Intel(R) MPI Library 2019 for Windows* OS
Job TITLE :
Job ID string : layer_4meshd
Time Step: 1, Simulation Time: 0.11 s
*******************************************
Following is the path from CMDfds prompt:
C:\FDS6_projects\mpi_test>path
PATH=C:\Program Files\firemodels\FDS6\bin\\mpi;C:\Program Files\firemodels\SMV6;C:\Program Files\firemodels\FDS6\bin; . ….
<NAME>., P.Eng., CP
Senior Associate
4th Floor, 780 Beatty Street
Vancouver, BC V6B 2M1
[LMDG 30-year logo larger font 300dpi]
604-682-7146 ext 410
<EMAIL><mailto:<EMAIL>>
www.lmdg.com<http://www.lmdg.com/>
[96dpi OQM-certified-wordmark-FINAL]
username_1: You did not type the commands like I asked you to do.
username_0: I was working from a remote station and somehow I lost your last email and couldn’t retrieve it. I tried to follow it by memory but obviously I missed out something. Would you be so kind to resend your last email to me.
Much appreciated.
<NAME>.Sc., P.Eng., CP
Senior Associate
4th Floor, 780 Beatty Street
Vancouver, BC V6B 2M1
[LMDG 30-year logo larger font 300dpi]
604-682-7146 ext 410
<EMAIL><mailto:<EMAIL>>
www.lmdg.com<http://www.lmdg.com/>
[96dpi OQM-certified-wordmark-FINAL]
username_2: just click on the link at the bottom of this (or your email) and you should
go to a page containing all emails in this thread.
On Mon, Jun 17, 2019 at 12:40 PM username_0 <<EMAIL>>
wrote:
> I was working from a remote station and somehow I lost your last email and
> couldn’t retrieve it. I tried to follow it by memory but obviously I missed
> out something. Would you be so kind to resend your last email to me.
>
> Much appreciated.
>
> <NAME> M.Sc., P.Eng., CP
> Senior Associate
> 4th Floor, 780 Beatty Street
> Vancouver, BC V6B 2M1
> [LMDG 30-year logo larger font 300dpi]
> 604-682-7146 ext 410
> <EMAIL><mailto:<EMAIL>>
> www.lmdg.com<http://www.lmdg.com/>
>
> [96dpi OQM-certified-wordmark-FINAL]
>
>
username_0: Hi Glenn
Thanks for your response.
In response to Kevin’s earlier email, I have the following info:
1. All the computers can see the shared working directory
2. ‘where mpiexec’ returns
C:\FDS6_projects\mpi_test>where mpiexec
C:\Program Files\firemodels\FDS6\bin\mpi\mpiexec.exe
3. Both mpiexec localonly and fds_local run the sample file
****************************
C:\mpi_projects>mpiexec -n 4 -localonly fds layer_4meshd.fds
Reading input file ...
Fire Dynamics Simulator
Current Date : June 17, 2019 10:50:34
Revision : FDS6.7.1-0-g14cc738-HEAD
Revision Date : Mon Feb 4 12:26:25 2019 -0500
Compiler : Intel ifort 172.16.17.32
Compilation Date : Wed 02/13/2019 03:33 PM
MPI Enabled; Number of MPI Processes: 4
OpenMP Enabled; Number of OpenMP Threads: 4
MPI version: 3.1
MPI library version: Intel(R) MPI Library 2019 for Windows* OS
Job TITLE :
Job ID string : layer_4meshd
************************************
*************************************
C:\mpi_projects>fds_local -p 4 layer_4meshd.fds
Reading input file ...
Fire Dynamics Simulator
Current Date : June 17, 2019 10:52:21
Revision : FDS6.7.1-0-g14cc738-HEAD
Revision Date : Mon Feb 4 12:26:25 2019 -0500
Compiler : Intel ifort 172.16.17.32
Compilation Date : Wed 02/13/2019 03:33 PM
MPI Enabled; Number of MPI Processes: 4
OpenMP Enabled; Number of OpenMP Threads: 1
MPI version: 3.1
MPI library version: Intel(R) MPI Library 2019 for Windows* OS
Job TITLE :
Job ID string : layer_4meshd
*******************************************
4. Running with 4 hosts returns the 1901 error
[Truncated]
wrote:
> I was working from a remote station and somehow I lost your last email and
> couldn’t retrieve it. I tried to follow it by memory but obviously I missed
> out something. Would you be so kind to resend your last email to me.
>
> Much appreciated.
>
> <NAME> M.Sc., P.Eng., CP
> Senior Associate
> 4th Floor, 780 Beatty Street
> Vancouver, BC V6B 2M1
> [LMDG 30-year logo larger font 300dpi]
> 604-682-7146 ext 410
> <EMAIL><mailto:<EMAIL><mailto:<EMAIL>%3cmailto:<EMAIL>>>
> www.lmdg.com<http://www.lmdg.com/<http://www.lmdg.com%3chttp:/www.lmdg.com/>>
>
> [96dpi OQM-certified-wordmark-FINAL]
>
>
username_1: This is all I know to do. It is possible that there are additional security features on your network that might cause things not to work.
username_0: Thank you.
<NAME> M.Sc., P.Eng., CP
Senior Associate
4th Floor, 780 Beatty Street
Vancouver, BC V6B 2M1
[LMDG 30-year logo larger font 300dpi]
604-682-7146 ext 410
<EMAIL><mailto:<EMAIL>>
www.lmdg.com<http://www.lmdg.com/>
[96dpi OQM-certified-wordmark-FINAL]
username_0: Finally, I solved the problem. Apparently, the -n argument does not work for Windows. The following command gets 4 machines each with 4 threads working. (I did specify an OMP_STACKSIZE in the environment for each machine though).
***********************************
C:\mpi_projects>mpiexec -hosts 4 firemodel-7259 1 firemodel-7260 1 firemodel-7261 1 firemodel-7262 1 -wdir \\firemodel-7259\mpi_projects -env OMP_NUM_THREADS 4 fds layer_4meshd.fds
Reading input file ...
Fire Dynamics Simulator
Current Date : June 18, 2019 14:01:28
Revision : FDS6.7.1-0-g14cc738-HEAD
Revision Date : Mon Feb 4 12:26:25 2019 -0500
Compiler : Intel ifort 172.16.17.32
Compilation Date : Wed 02/13/2019 03:33 PM
MPI Enabled; Number of MPI Processes: 4
OpenMP Enabled; Number of OpenMP Threads: 4
MPI version: 3.1
MPI library version: Intel(R) MPI Library 2019 for Windows* OS
*************************************
<NAME> M.Sc., P.Eng., CP
Senior Associate
4th Floor, 780 Beatty Street
Vancouver, BC V6B 2M1
[LMDG 30-year logo larger font 300dpi]
604-682-7146 ext 410
<EMAIL><mailto:<EMAIL>>
www.lmdg.com<http://www.lmdg.com/>
[96dpi OQM-certified-wordmark-FINAL]
Status: Issue closed
username_1: I think the -n does work, but not when you have explicitly listed the computers. There are so many options for mpiexec under the different OS that it is hard to know how to use them together. Thanks for pointing that out. |
NaturalNode/natural | 285370530 | Title: Help with Vietnamese
Question:
username_0: add vietnamese of tokenizers
`
var Tokenizer = require('./tokenizer'),
util = require('util');
var AggressiveTokenizer = function() {
Tokenizer.call(this);
};
util.inherits(AggressiveTokenizer, Tokenizer);
module.exports = AggressiveTokenizer;
AggressiveTokenizer.prototype.tokenize = function(text) {
// break a string up into an array of tokens by anything non-word
return this.trim(text.split(/[^a-z0-9àáảãạăắằẳẵặâấầẩẫậéèẻẽẹêếềểễệíìỉĩịóòỏõọôốồổỗộơớờởỡợúùủũụưứừửữựýỳỷỹỵđ]+/i));
};
`
Answers:
username_1: +1
username_2: Added your tokenizer to natural. See [#425](https://github.com/NaturalNode/natural/pull/425).
Thanks!
Status: Issue closed
|
olsh/dotnet-reqube | 832349048 | Title: 1.2.0+ complains about solution file not exists.
Question:
username_0: Here is the error message:
`
22:17:18 INF] Reading input file ReSharperReport.xml...
Unhandled exception. System.IO.FileNotFoundException: Solution file /var/jenkins/workspace/Efficiency-COOP_sonar-test_qa/cop\Cop.sln does not exist
`
Answers:
username_0: This error is caused by the path format of my ReSharperReport.xml file.
like this: \<Solution\>cop/Cop.sln\</Solution\>
username_1: Hi @username_0
Thank you for the feedback. I don't use R# CLT on *nix personally, so I'd be happy to accept a PR.
username_1: This way you'll get the correct directory separators in the R# report file.
Let's keep the issue open until the bug is resolved on the CLT side.
username_2: This solution leads to a error:
`Unhandled exception. System.ArgumentException: Absolute paths are not allowed with -output, when converting a sln.`
username_1: Hi @username_2
Can't reproduce the issue. I use the following command
```
./inspectcode.sh /project/path/Project.sln -o="/output/directory/result.xml" --no-build -a
```
The result.xml contains absolute paths.
JetBrains Inspect Code 2021.3.2 |
oblador/react-native-vector-icons | 268293176 | Title: How to fill with color transparent place in icon
Question:
username_0: For example, I have bookmark check icon, and need transparent check mark fill in some color
```
<MaterialIcons
name = 'bookmark-check'
size = {30}
color = {'blue'}
/>
```

Answers:
username_1: I have problems with this too! I'm using "logo-youtube" from Ionicons and I want the middle play triangle to be white while the surrounding part is red, however if I add a white backgroundColor there will be a white box around the icon, and changing the padding to 0 does nothing at all.
username_2: Also having this issue with circle icons such as `play-circle`. Tried to add border radius and it doesn't seem to work (on iOS, using props, style, Text or View wrapper, or Icon.Button). So yeah, there's box behind the icon that would be nice to figure out how to not show.
username_3: Don't think this is possible with the current implementation I'm afraid :-/
username_4: did anyone got the solution for the above issue..please let me know..i am facing the same
username_5: Also having this issue
username_6: @username_3 Is there any plans to change that? Or will it be "as is"? I think an ability to fill color is must have feature. Now it's not possible to do in any way (at least I couldn't find) neither with this repo, nor with React Native component (like imaginable overlay or whatever).
username_7: Well, it seems that the icon is already surrounded with a View on it's implementation, and the component got a style prop that is referred to that View. So you can style like you normally do, according to your need.
```
<TouchableOpacity>
<FontAwesome name="facebook-official" size = {48} color= "#05f" style={iconStyle}/>
</TouchableOpacity>
const styles = {
iconStyle: {
backgroundColor:"#fff",
height: 44,
width:43,
borderWidth: 4,
borderRadius: 12,
borderColor: '#212121',
borderBottomWidth:0,
borderBottomColor: 'transparent'
}
}
```
username_8: I also having this problem. Maybe adding a view behind the icon can fix the problem temporary.
username_9: Good idea. I will use the solution before author fix the bug, but maybe is a feature. XD
username_10: <View style={{ backgroundColor: '#fff', borderRadius: 25 }}>
<MaterialIcons
name = 'bookmark-check'
size = {30}
color = {'blue'}
/>
</View>
try this out ... give borderRadius: 25 or until hide it. Note: no need for width and height if u already write sizze={30}
Hope this helps!
username_11: use reverse attribute
username_12: This worked for me, based off of natejenkins answer
```
<View style={{ width: size, height: size, padding: 5 }}>
<View style={{
backgroundColor: 'white', height: '100%', width: '100%', borderRadius: 50,
}} />
</View>
<VectorIcon color={color} name={name} size={size} style={{ position: 'absolute', top: 0, left: 0 }} />
``` |
inveniosoftware/invenio-oauthclient | 975621826 | Title: ReadTheDocs version does not match repository
Question:
username_0: Package version (if known): current (
## Describe the bug
In the _contributing_ (rtfm) page, _pull request guidelines_ (https://invenio-oauthclient.readthedocs.io/en/latest/contributing.html#pull-request-guidelines), there is a link to Travis that is broken. Whereas in the contributing (github) document (https://github.com/inveniosoftware/invenio-oauthclient/blob/master/CONTRIBUTING.rst) the corresponding link points to GitHub CI workflows.
Also, I suppose the Python versions therein listed (`2.7, 3.3, 3.4 and 3.5`) is outdated.
## Steps to Reproduce
1. Go to https://invenio-oauthclient.readthedocs.io/en/latest/contributing.html#pull-request-guidelines
2. Click on the link (currently, https://travis-ci.com/inveniosoftware/invenio-oauthclient/pull_requests)
4. See error
## Expected behavior
To land on a working page after clicking the above mentioned link.
Regarding the Python versions `invenio-oauthclient` should support I would not expect to see `2.x` but `3.x` anymore (for new features, at least).
## Additional context
Checking the version of the `latest` version of the docs, at https://readthedocs.org/projects/invenio-oauthclient/, I noticed the latest version was built 2.5+ years ago. |
tmenier/Flurl | 163073176 | Title: 302 with set-cookie error
Question:
username_0: when a request is 302 with a set-cookie header, cookie is not set correctly
Answers:
username_1: Please provide a code example or unit test to demonstrate the issue. Keep in mind Flurl follows 302 redirects by default (it uses [HttpClientHandler](https://msdn.microsoft.com/en-us/library/system.net.http.httpclienthandler.aspx) under the hood, which has this behavior), so cookie support needs to be explicitly enabled before the HTTP call by calling `EnableCookies()` on the FlurlClient.
username_0: I get the cookie by do this trick
```cs
var content = new
{
j_password = <PASSWORD>,
j_username = username + ",undergraduate"
};
var ret = await ElearningUris.ValidateAccount
.WithClient(client)
.PostUrlEncodedAsync(content);
var response = JsonConvert.DeserializeObject<LoginReturn>(await ret.Content.ReadAsStringAsync());
var c1 = client.GetCookies();
var uri = ret.RequestMessage.RequestUri.AbsoluteUri;
var cookie = new Cookie("JSESSIONID", Regex.Match(uri, "jsessionid=(.*)").Groups[1].Value);
client.WithCookie(cookie);
var c2 = client.GetCookies();
```
```
HTTP/1.1 302 Found
Server: Apache-Coyote/1.1
Set-Cookie: JSESSIONID=<KEY>; Path=/choose_courses/; HttpOnly
Location: http://elearning.ustb.edu.cn/choose_courses/loginsucc.action;jsessionid=<KEY>FB18C37144D1
Content-Length: 0
Date: Sat, 02 Jul 2016 07:57:05 GMT
```
but you can see

I test it by flask
```python
from flask import Flask, session, redirect, url_for, escape, request, make_response
app = Flask(__name__)
app.debug = True
@app.route('/')
def index():
if 'user_cookie' in request.cookies:
return 'Logged in as %s' % escape(request.cookies.get('user_cookie'))
return 'You are not logged in'
@app.route("/login", methods=['GET', 'POST'])
def login():
if request.method == 'POST':
redirect_to_index = redirect(url_for('index'))
response = app.make_response(redirect_to_index)
response.set_cookie("user_cookie", request.form['username'])
return response
return '''
<form action="" method="post">
<p><input type=text name=username>
<p><input type=submit value=Login>
</form>
'''
[Truncated]
Location: http://127.0.0.1:5000/
Set-Cookie: user_cookie=sad; Path=/
Server: Werkzeug/0.10.4 Python/3.5.0
Date: Sat, 02 Jul 2016 08:24:17 GMT
<!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 3.2 Final//EN">
<title>Redirecting...</title>
<h1>Redirecting...</h1>
<p>You should be redirected automatically to target URL: <a href="/">/</a>. If not click the link.
```
```
HTTP/1.0 200 OK
Content-Type: text/html; charset=utf-8
Content-Length: 16
Server: Werkzeug/0.10.4 Python/3.5.0
Date: Sat, 02 Jul 2016 08:24:17 GMT
Logged in as sad
```
username_0: I'm not sure what the problem is
username_0: ```cs
private FlurlClient client = new FlurlClient(ElearningUris.Server)
.WithHeader("User-Agent", "Mozilla/5.0 (Windows NT 10.0; WOW64; rv:46.0) Gecko/20100101 Firefox/46.0")
.EnableCookies();
```
Miss some code
username_1: Much work was done with cookie support for Flurl.Http 1.0. `GetCookies()` has been removed, replaced by `FlurlClient.Cookies` (dictionary). Please try it and see if it fixes your issue:
https://www.nuget.org/packages/Flurl.Http/1.0.0-beta7
username_0: http://stackoverflow.com/questions/15103513/httpwebresponse-cookies-empty-despite-set-cookie-header-no-redirect
username_0: They look very similar.
username_1: [This test](https://github.com/username_1/Flurl/blob/dev/Test/Flurl.Test.Shared/Http/RealHttpTests.cs#L45-L50) demonstrates a cookie being set property with a 302. In a browser, the initial request to https://httpbin.org/cookies/set?z=999 responds with these headers:
```
HTTP/1.1 302 FOUND
Server: nginx
Date: Tue, 12 Jul 2016 20:28:38 GMT
Content-Type: text/html; charset=utf-8
Content-Length: 223
Connection: keep-alive
Location: /cookies
Set-Cookie: z=999; Path=/
Access-Control-Allow-Origin: *
Access-Control-Allow-Credentials: true
```
So I don't know what to do about this issue. If you can demonstrate it in a test with a **_real_** URL provided, I can look into it further. Otherwise I'll have to close this as not reproducible.
username_0: Ok, I success make it reproduce
1. First step is not neccessory.( only for fiddler to grap the package)
add
```
127.0.0.1 test
```
to host
2.
use python 3 and flask
```python
from flask import Flask, session, redirect, url_for, escape, request, make_response
app = Flask(__name__)
app.debug = True
@app.route('/test_path/test.action;jsessionid=<id>')
def index(id):
return 'jsessionid is %s' % id
@app.route("/login", methods=['GET', 'POST'])
def login():
if request.method == 'POST':
redirect_to_index = redirect("/test_path/test.action;jsessionid=A53FEA63CB70CCA5623EFB18C37144D1")
response = app.make_response(redirect_to_index)
response.set_cookie("JSESSIONID", "A53FEA63CB70CCA5623EFB18C37144D1",path='/test_path/')
return response
return '''
<form action="" method="post">
<p><input type=text name=username>
<p><input type=submit value=Login>
</form>
'''
@app.route('/logout')
def logout():
session.pop('username', None)
return redirect(url_for('index'))
if __name__ == '__main__':
app.run()
```
3. Add test
```csharp
[Test]
public async Task get_my_cookie()
{
const string url = "http://test:5000/login";
var data = new { username = "456789" };
var client = new FlurlClient(url).EnableCookies();
var re = await url.WithClient(client).AllowAnyHttpStatus().PostUrlEncodedAsync(data).ReceiveString();
var cookies = client.Cookies;
Assert.Equals(cookies.Count, 1);
}
```
[Truncated]
Content-Length: 339
Location: http://127.0.0.1:5000/test_path/test.action;jsessionid=<KEY>
Set-Cookie: JSESSIONID=<KEY>; Path=/test_path/
Server: Werkzeug/0.10.4 Python/3.5.0
Date: Wed, 13 Jul 2016 10:35:35 GMT
<!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 3.2 Final//EN">
<title>Redirecting...</title>
<h1>Redirecting...</h1>
<p>You should be redirected automatically to target URL: <a href="/test_path/test.action;jsessionid=<KEY>">/test_path/test.action;jsessionid=<KEY></a>. If not click the link.
```
In Chrome it's ok, and restsharp is also ok(not test, but orgin problem don't occur in restsharp).

about how to fix:
http://stackoverflow.com/questions/15103513/httpwebresponse-cookies-empty-despite-set-cookie-header-no-redirect
http://stackoverflow.com/questions/1055853/how-to-parse-httpwebresponse-headers-keys-for-a-set-cookie-session-id-returned/1055883#1055883
A lot of people meet this issue, it seems a cookiecontainer bug.
Status: Issue closed
|
F5Networks/f5-aws-cloudformation | 535766105 | Title: S3 permission issue with latest CFT
Question:
username_0: ## Do you already have an issue opened with F5 support?
no
## Description
S3 error: Access Denied For more information check http://docs.aws.amazon.com/AmazonS3/latest/API/ErrorResponses.html
## Template
https://github.com/F5Networks/f5-aws-cloudformation/tree/master/supported/autoscale/waf/via-lb/1nic/existing-stack/payg
and possibly other templates
## Severity Level
3
Answers:
username_0: Also suggest to add a test for post template release to validate S3 permissions across all templates. This will help avoid catching in the field.
username_1: Thanks Jeff. We have updated the readme's to use the correct s3 url. We have added the request to test on release into our internal tasks.
https://github.com/F5Networks/f5-aws-cloudformation/tree/master/supported/autoscale/waf/via-lb/1nic/existing-stack/payg#launching-the-template-using-the-aws-launch-stack-buttons
Status: Issue closed
username_0: This is happening again ever since latest templates were released today. AWS is throwing access denied S3 errors. |
ONSdigital/eq-survey-runner | 269633748 | Title: Generated requirements.txt contains test dependencies
Question:
username_0: ### Expected behaviour
The dev dependencies should not in included in the requirements.txt when the app is deployed
### Actual behaviour
The test dependencies are in the requirements.txt
### Steps to reproduce the behaviour
`piping lock -r > requirements.txt`
Answers:
username_1: This seems to be intended behaviour. See: https://github.com/kennethreitz/pipenv/issues/245.
We could use this script as a workaround (from the above): https://github.com/kennethreitz/pipenv/issues/245#issuecomment-283694215
Or something like: https://github.com/UnitedIncome/serverless-python-requirements/issues/87#issuecomment-335454523
username_1: Fixed in latest version of Pipenv: kennethreitz/pipenv#972
Upgrading Pipenv on Jenkins would solve this issue.
Status: Issue closed
|
vivet/GoogleApi | 394732760 | Title: FEATURE REQUEST: User settable BaseURL for API calls
Question:
username_0: I'd be interested if you could make the change to allow a user settable BaseURL for the different APIs. The reason I'm interested in this is I want to have requests sent to one of my web servers for caching to reduce duplicate API requests over time.
Thanks, really liking this project.
Answers:
username_1: Hi
I am not sure i completely understand what you are trying to accomplish, (and I would to before adding the feature).
So basically you have a "mirror" of the Google service, and you want to try that before actually requesting at Google? If that's the case then I am pretty sure you would violate the terms of service of Google. It depends a bit on what kind of data you are caching, as it's different for each api. Here are some info:
* https://cloud.google.com/maps-platform/terms/
* https://cloud.google.com/maps-platform/terms/maps-service-terms/
username_1: Hey I hope you figured out how to override the request class to accomplish your use case.
Let me know if you still need something from me.
Will close the issue.
Status: Issue closed
|
kubecost/cost-analyzer-helm-chart | 599557245 | Title: Run as nonroot user (PSP ON)
Question:
username_0: Hello,
I'm using the latest chart with PSP ON on my EKS cluster:
`eks.privileged false RunAsAny MustRunAsNonRoot MustRunAs MustRunAs false configMap,emptyDir,secret
`
**What problem are you trying to solve?**
Running this chart as nonroot user
**Describe the solution you'd like**
The image gcr.io/kubecost1/checks:prod-1.55.1 on kind: CronJob (cost-analyzer-checks) must run asnonroot user (todo: set empty securityContext)
Answers:
username_1: Thanks for reporting, Adrien! We'll investigate this in our upcoming sprint but let us know if you're interested in collaborating on this improvement.
Also, for additional context, this pod powers alerts & email updates. Here's an [architecture overview](https://github.com/kubecost/docs/blob/master/architecture.md).
username_2: I think this should be as simple as setting the SecurityContext on the checks pod-- no root capabilities are needed as far as I know.
username_1: @username_0, was ajay's comment helpful? Let us know if it would be useful for us to add to our helm chart still.
username_3: Hi @username_0 this PR https://github.com/kubecost/cost-analyzer-helm-chart/pull/756/files should fix the issue.
username_3: @username_2 @username_1 version 1.73.0 fix this issue. Very good! 👍
username_3: But I've still a problem with Network Policy for kubecostCheck. Ref: https://github.com/kubecost/cost-analyzer-helm-chart/issues/761
username_1: This core issue should be addressed. Closing while we look at #761.
Please let us know if there are questions.
Status: Issue closed
|
geduldig/TwitterAPI | 864114169 | Title: [Question] Handling of too many requests
Question:
username_0: I was wondering if this package handles errors arising from too many requests made to a Twitter API made in a given time slot.
Answers:
username_1: If you are using the `TwitterAPI.request` method, you should catch the appropriate error and throttle your requests. If your are using the `TwitterPager` helper class, it will automatically throttle for you. The class also takes an optional `wait` argument to set the interval between requests.
The docs are sorely out-of-date. I'll be working on this in the next couple of weeks. Exceptions to be caught are listed [here](https://twitterapi.readthedocs.io/en/latest/twittererror.html).
Status: Issue closed
|
raszi/node-tmp | 97660496 | Title: On Mac 10.9 with node 0.10.40, throws an error cannot find module rimraf
Question:
username_0: The rimraf js is present in the node_modules/rimraf/ , in the same folder as temp.js and at the global scale too. Still, the module loading is failing.
Doesn't seem to happen on ubuntu 12.04 with node 0.10.38
Status: Issue closed
Answers:
username_1: `tmp` does not depend on `rimraf`, the `temp` module depends on `rimraf` but it is a different project, please open the issue [there](https://github.com/bruce/node-temp/issues).
username_0: My Bad. Thanks for redirecting me to correct repo. |
TorXakis/TorXakis | 321143443 | Title: Shouldn't we remove all linux commands from semaphore CI configuration?
Question:
username_0: And instead put them in scripts. Otherwise:
- Things are defined at two places.
- Semaphore CI is not versioned
For instance the $PATH variable could be set in a script

Does it make sense to put the exports as part of `setup.sh`?
Answers:
username_1: The reason for writing `export`s as separate commands was that they don't work when they are in a script. The exported variables were only available within the script, not for following scripts.
If that behaviour is changed, then we should put them back into relevant scripts.
Status: Issue closed
username_0: Ok, clear. |
YangCatalog/search | 374961800 | Title: Impact graph "ERROR! Loop found {} <=> {}" errors
Question:
username_0: Following on from #13, I tried explicitly mentioning a submodule and get what you see below.



Answers:
username_1: fixed with #15
Status: Issue closed
|
getgrav/grav-plugin-custom-css | 221822823 | Title: Wrapper for highlighting css / less / sass / scss
Question:
username_0: Nice plugin, for development I really like this plugin.
To make it even better I'd like to use a "wrapper" so the inline css has highlighting.
When looking at the files me as rookie really don't know how to add it.
I have just started to work with Grav.
I think this is a good highlighter:
http://prismjs.com/index.html
It looks like all you need to do is adding a css and js file and the textarea needs a class.
```<pre><code class="language-css"><textarea...```
```
<link href="themes/prism.css" rel="stylesheet" />
<script src="prism.js"></script>
```
Example here: https://jsfiddle.net/gnev4yd1/
Answers:
username_0: Ok so it seems its already available. Now it only needs to work :)
When changing the blueprint.yaml nothing happens unfortunately.
```
css_inline:
type: editor
label: Inline CSS
autofocus: true
codemirror:
mode: 'yaml'
indentUnit: 4
autofocus: true
indentWithTabs: false
lineNumbers: true
styleActiveLine: true
gutters: ['CodeMirror-lint-markers']
lint: true
```
username_0: Somehow its not updating the label and not working.
It looks more like an update/cache issue, down see what else can be wrong.
Status: Issue closed
|
dropbox/dropbox-sdk-js | 263742030 | Title: Incomplete Documentation For filesGetThumbnail
Question:
username_0: The docs do not describe how the thumbnail data is returned. Via debugging I see a string property named fileBinary. The docs have no mention of this property anywhere.
Answers:
username_1: Thanks for pointing that out! Using `fileBinary` is the correct way of accessing the resulting data in node. (In the browser, it's `fileBlob`.) I'll ask the team to add this to the documentation.
username_2: This is a limitation / issue with the Stone API specification. The Dropbox JS SDK manually adds `fileBinary`/`fileBlob` to the object because that property isn't in the Stone API specification. As a corollary, the TypeScript definition files lack these properties, and [require hacky typecasting to 'any' to compile](https://github.com/dropbox/dropbox-sdk-js/blob/master/examples/typescript/download/download.ts#L10).
If you fix this in a generic manner (through e.g. [arguments to the Stone generators](https://github.com/dropbox/dropbox-sdk-js/blob/master/generator/generate_routes.py#L74) that pass in library-specific augmentations to the Stone specification), then you could kill two birds with one stone -- improve the TypeScript definition files *and* correct the documentation. |
PaddlePaddle/PaddleOCR | 1101313413 | Title: 训练train.py卡住不动,偶尔会到训练迭代1-2次在卡住不动
Question:
username_0: 请提供下述完整信息以便快速定位问题/Please provide the following information to quickly locate the problem
- 系统环境/System Environment:windows10
- 版本号/Version:Paddle:2.2
- 运行指令/Command Code:--use_gpu --cfg ./configs/deeplabv3p_xception65_optic.yaml
- 完整报错/Complete Error Message:

启动训练后在这里卡住不动,神奇的是偶尔会迭代两次然后再次卡住!显存没满,内存也没满,都还剩下很多。这是怎么回事
Answers:
username_1: 看配置文件,你跑的并不是OCR的模型吧,可以去PaddleSeg下提issue:https://github.com/PaddlePaddle/PaddleSeg/issues |
TheCoder4eu/BootsFaces-OSP | 53959264 | Title: Combine the Javascript and CSS files
Question:
username_0: We should give our users a hint in the documentation how to combine the Javascript and CSS file (by using http://showcase.omnifaces.org/resourcehandlers/CombinedResourceHandler). Maybe we should even implement such a resource handler ourselves.
Answers:
username_0: I modified several components of BootsFaces so that the CombinedResourceHandler detects their JS files and combines them into a single file.
username_0: Done. The ```CombinedResourceHandler``` is already active on the staging area, so I assume we can call the feature "tested".
The documentation will be available at http://www.bootsfaces.net/integration/OmniFaces.jsf as soon as the new version of the showcase has been released.
Status: Issue closed
|
hermanjustnu/scoop-emulators | 879355221 | Title: [email protected]: hash check failed
Question:
username_0: ```
Checking hash of RetroArch.7z ... ERROR Hash check failed!
App: emulators/retroarch
URL: https://buildbot.libretro.com/stable/1.9.2/windows/x86_64/RetroArch.7z
First bytes: 37 7A BC AF 27 1C 00 04
Expected: 43295d3c71580a4bac1370cd19e5927a63577850348dab29d258b0ca36cbbfd2
Actual: 86e8a4777cbe95f93d881568f5d13e9ff3bb3038a82cf7fc88937eebfabcb421
```<issue_closed>
Status: Issue closed |
RIP21/react-simplemde-editor | 545339574 | Title: How to set 'Tab' shortcut to next Editor?
Question:
username_0: Tab to next input field.
How can i do this?
Answers:
username_1: Probably you can override some event and on Tab click, you can change the focus on the next input by its ref or something. Never did that, so I can just guess. Recommend you to visit the easymde docs to see a bit more possibilities of fine-tuning and other features that this editor may have.
Status: Issue closed
|
trackreco/mkFit | 354989627 | Title: Migrate to multi-threaded ROOT6 for validation
Question:
username_0: Not so much an issue, more of a sticky for an update to the validation. At the moment, PlotValidation.cpp (the main code for reading the trees and making the physics performance plots) is all serial and takes quite a bit of time to run.
With the developments in ROOT6 for parallelism + multithreading, I want to at some point update the code to take advantage of this. The tutorials from ROOT provide some good examples: https://root.cern.ch/doc/v612/group__tutorial__multicore.html
Namely, parallelize the loop over entries when reading the tree by using a TTreeReaderValue, making a vector of TThreadedObjects, or something like this.
This is rather low priority for me. This would be a fun side project when I get time (don't know when...) to dive into this and multithread the validation code. Considering our project is all about MT, this would be nice exercise into what ROOT has under the hood for doing these sorts of tasks.<issue_closed>
Status: Issue closed |
rongcloud/rongcloud-im-flutter-sdk | 600691080 | Title: 请问有推送功能吗
Question:
username_0: 请问有推送功能吗 融云的远程推送
Answers:
username_1: IM Flutter Plugin 核心是 iOS/Android 的原生 SDK,所以远程推送与原生 SDK 的方式一致
[iOS 远程推送文档](https://docs.rongcloud.cn/v3/views/im/noui/guide/private/notify/push/ios.html)
[Android 远程推送文档](https://docs.rongcloud.cn/v3/views/im/noui/guide/private/notify/push/rong.html)
Status: Issue closed
|
WayofTime/BloodMagic | 106356228 | Title: Dupe bug with Vajra
Question:
username_0: Apparently mining a belljar with Vajra in silk touch mode (you know, the right click thing) gives you back two belljars with the exact same amount of reagent.
Status: Issue closed
Answers:
username_1: Seeing as it is item-specific (the Vajra), this would have to be a bug on their end. Every other mining method works as intended. |
jambolo/Chess | 394934229 | Title: Static evaluator could consider threats and protection
Question:
username_0: The computer doesn't develop its pieces. It justs lets them sit on the back row for most of the game. This is probably because it doesn't value threats and protection in the SEF.
Answers:
username_0: Done in 4930137eba7b9c6d24584a29c1049beb2ff50658.
Status: Issue closed
|
emmetio/livestyle-sublime | 18770879 | Title: Firefox support!
Question:
username_0: <bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/1267503-firefox-support?utm_campaign=plugin&utm_content=tracker%2F305388&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F305388&utm_medium=issues&utm_source=github).
</bountysource-plugin>
Answers:
username_1: +1
username_2: +1
username_3: +1
username_4: +100000000
username_5: There’s basic, one way (editor→browser) support via Remote View feature, new in LiveStyle 0.9. It can be actually used with any browser (IE, mobile, etc).
**Instructions:**
1. Download LiveStyle app from http://livestyle.io
2. Run LiveStyle app: it acts like plugin installer and Remote View client. Install both Chrome and Sublime Text plugins. Note that these are dev plugins from different repos, you *must* remove old LiveStyle plugins from Chrome and Sublime Text.
3. In Chrome, go to your web-site, then open LiveStyle popup (it’s a LS icon in toolbar) and enable Remote View (requires Google account).
4. Remote View will issue a publicly available domain name that connected to your local web-site and powered with LiveStyle updates. Use this public domain name in Firefox (and any other browser: IE, mobile, etc.) to get live updates from your editor.
Note that you still have to control LiveStyle mappings in Chrome and use DevTools only for browser→editor live editing (they are reflected in Remote View domains as well). Full Firefox support will be a part of crowdfunding campaign.
PS: LiveStyle 0.9 support LESS and SCSS, check it out: http://livestyle.emmet.io/alpha/
username_4: yeaaaah don't want to have to download chrome, plus I really want to use the browser -> editor feature.
FF support should definitely be a priority...
username_6: 1
username_7: +1
username_4: +1 (again)
username_4: Bump
username_8: +1 again
username_9: +1
Also because when coding for Firefox you don't have to be watching for all the bugs like in chrome, I'm surprised there is no support for it. At least say so in the site.
username_10: +10
username_11: It is a shame. Just don't like Chrome....
username_12: I have a solution haha i very simple actually,
Install Chrome Store Foxified
Then goto LiveStyle Alpha: [https://chrome.google.com/webstore/detail/livestyle-alpha/obipchajaiohjoohongibhgbfgchblei](url)
Install it by signing in the crx file and save it as xpi
then install it with firefox, thats it! hahahahahhahahah |
EdenServer/community | 573461083 | Title: Windurst 5-2 The Shadows Await
Question:
username_0: ### Checklist
<!--
Don't edit or delete this section, but tick the boxes after you have submitted your issue.
If there are unticked boxes a developer may not address the issue.
Make sure you comply with the checklist and then start writing in the details section below.
-->
- [/] I have searched for existing issues for issues like this one. The issue has not been posted. (Duplicate reports slow down development.)
- [/] I have provided reproducable steps. (No "as the title says" posts please. Provide reproducable steps even if it seems like a waste of time.)
- [/] I have provided my client version in the details. (type /ver into your game window)
### Details
Version 30200217_0
When receiving the cutscene for 5-2 at the door in the Throne Room it begins the cut scene, but then your character will just continue to run into the door. Issue happens on non-retail clients too (^: Using prt scrn does not fix the problem and you'll be stuck until you force close your client.
Status: Issue closed
Answers:
username_1: have not been able to replicate this unless on a different version, as the ids have shifted and the door stays closed as a result. using eden install does not seem to encounter this issue, so closing. |
boostorg/variant2 | 830248098 | Title: Inheriting constructors issue with older compilers
Question:
username_0: With clang < 4 and gcc < 7 and probably latest MSVC, the following code and [godbolt](https://godbolt.org/z/fs44cz) fail to compile:
```c++
#include <boost/variant2/variant.hpp>
template <typename ...T>
class any_type : boost::variant2::variant<T...>
{
using parent_t = boost::variant2::variant<T...>;
using parent_t::parent_t;
};
struct foo
{
foo() {}
foo(int i_arg) : i(i_arg) {}
foo(foo const& rhs) {}
template <typename T>
foo(T const& rhs) {
rhs.bar();
}
int i{0};
};
int main(int argc, char *argv[])
{
using any_t = any_type<int, double, foo>;
any_t a(1);
any_t b(a);
}
```
The issue comes from `foo`'s converting constructor.
Related to boostorg/gil#526
Answers:
username_1: Interesting, I hadn't noticed that the derivation was even private. Yes, this fails with MSVC (and it fails with their own `std::variant` too.) I'm not quite sure what makes the newer GCC and Clang compile it.
I can make this work, but I need to think a bit more about it, make sure there's nothing legitimate that the fix would break.
Status: Issue closed
username_1: Merged the fix to develop.
username_0: Awesome thanks! Since you mention MSVC's `std::variant`, I took the initiative to report the issue there as well. |
keboola/ex-email-attachments | 399697827 | Title: mysterious data1 table
Question:
username_0: vypadá to, že pokud je v inboxu víc zpráv, generuje to na výstupu víc souborů - a pak to chce importovat jako víc tabulek. v dokumentaci jsem nic nenašel, ve zdrojáku toto
https://github.com/keboola/ex-email-attachments/blob/dbf9188de202f002eae7f8a02cfbb32f55ae4195/src/Keboola/ExEmailAttachments/Action/RunAction.php#L182
je to feature nebo bug? například vypadl jeden pravidelnej job, v inboxu jsou dva maily a už to nemůžu nijak rozběhnout.
https://connection.keboola.com/admin/projects/232/jobs/477809311

```
{
"parameters": {
"email": "***<EMAIL>",
"delimiter": ";",
"enclosure": "\"",
"primaryKey": [],
"incremental": true
}
}
```
Answers:
username_0: několik jobů během poslední noci vyfailovalo kvůli docker runneru, těm se neuložil state a zítra budou importovat víc souborů - všem teda vznikne nějaká `data1` tabulka a v `data` budou nekompletní importy (bude chybět to, co je v `data1`).
username_1: Myslím že to bylo ohledně tohoto issue: https://github.com/keboola/ex-email-attachments/issues/18
username_1: https://github.com/keboola/ex-email-attachments/commit/cc1f6de6cb6d6cffaf2a357226456aee27623610
username_0: ref https://keboola.zendesk.com/agent/tickets/8772
ref https://keboola.zendesk.com/agent/tickets/8692
username_1: Doplněno do dokumentace: https://github.com/keboola/ex-email-attachments/pull/27
Status: Issue closed
|
denoland/deno | 577907911 | Title: Deno compile needs revision
Question:
username_0: **Deno version:** 0.35.0
**Trying to compile a `.js` file results in `Uncaught Error: Unexpected skip of the emit.`**
Attempting the following
`./app.js`
```javascript
console.log("Compile me");
```
`./compile_app.js`
```javascript
const [compile_diag, compile_res] = await Deno.compile(
"./app.jsx",
undefined,
{
allowJs: false,
checkJs: false,
lib: ["es2018"],
sourceMap: false,
target: "es2018",
}
);
compile_diag && console.log(Deno.formatDiagnostics(compile_diag));
const truncate_filename = filename => filename.replace('file:///', '');
for (const [filename, text] of Object.entries(compile_res)) {
console.log(`${truncate_filename(filename)}`);
console.log(text.split('\n').map(line => `\t${line}`).join('\n'));
}
```
Results in:

Quite probably because the resulting filename matches the original filename, however since `Deno,compile` does not write anything into the filesystem this should not be a problem.
Same thing happens when attempting to compile a JS and TS file named the same on the same route. Ex: `mod.js` & `mod.ts`,
Answers:
username_1: CC @username_2
username_2: We are ignoring an error in the TypeScript compiler which is protecting possibly overwriting the source, which we do all the time in the internal of Deno. I am not totally sure it is an error, though we need to surface up what is there, because even if `Deno.compile()` doesn't physically write out files, the logic I think is sound.
What happens when you set `out` to some other path?
username_0: `out` makes the program implode. Setting `outDir` however solves the problem at certaing extent.
`.js` files won't fail. But naming two files the same with JS and TS extension keeps being a problem. However I don't see a solution that doesn't involve changing `Deno.compile` behavior.
username_2: `something.js` and `something.ts` in the same path is always a bad idea. The Deno runtime sort of allows it, but the is no way of disambiguating the output of a compile.
So I don't think this is a bug in functionality, but needs better error messages instead of the obscure emit skipped.
username_2: Writing it down while it is on my mind. Currently we squash a whole set of errors from the TypeScript compiler that cover over things that aren't relevant in the Deno runtime, but `Deno.compile()` and `Deno.bundle()` is really targeted at generating code that would work in _other_ runtimes, so we shouldn't just hide those issues away. So we need two different sets of errors that we squash and allow the invoker to decide if those are useful or not to ignore.
username_0: Not sure about a different set of errors, but definitely more clearer ones.
It might be the moment to discuss something that's been bugging me for a while. Renaming `Deno.compile` to fit something more closer to it's real behavior.
username_2: Like what?
username_0: Deno.transpile
If compilation through Deno becomes a feature inside the program a rename would be necessary, to not mention the fact that this is not compiling at all
Also, `Deno.transpileOnly` seems to me like an over simplified version of `Deno.compile` which could be achieved through TS config. Am I wrong?
username_2: Transpile, to me, means to convert from one syntax to another. `Deno.compile()` and `Deno.bundle()` do type checking and resolution, before it does a transpile. So to me it is less representative to name it `.transpile()`.
`Deno.transpileOnly()` only erases the types and transform syntax, similar to how Babel handles TypeScript. You can't use compiler options to mimic it. In fact there is a request in TypeScript to add the feature to `tsc` because it can't be fully replicated via compiler options.
username_0: In the most traditional meaning of the word, a compile process realizes both transpiling and bundling to deliver one single target(binary or not)
Deno compile doesn't really works by itself currently since it can't erase references to other modules in the spirit of not breaking anything. And Deno bundle doesn't make any real sintax transpilation other than converting module references to embedded code.
Deno compile should do both. Since current behavior doesn't actually deliver a functional output into say ES3.
username_3: Is there a sensible way to merge them both into `Deno.transpile()` and opt in to type checking and resolving as options? (Or are they too fundamentally different?)
username_0: I see no use case on Bundling without specyfing a target runtime.
Thumbs up for unification.
username_2: `Deno.transpileOnly()` uses a totally different TypeScript compiler API than `Deno.compile()`/`Deno.bundle()` does. `Deno.bundle()` is not only a default configuration of `Deno.compile()` it injects code to make the output a fully stand alone JavaScript file. `Deno.transpile()` does not exist (because without the `Only` is was more confusing).
username_2: My biggest concern about combining `Deno.compile()` and `Deno.bundle()` would be that the return types are dramatically different. The output filename of `Deno.bundle()` makes 0 sense in that context, while as with `Deno.compile()` the output file name(s) are meaningful. Also, it is more consistent with `deno run` and `deno bundle`. Alignment to the concept in the command line is important.
username_0: Consistency is important, but oversimplify the operations behind this commands only diminishes the worth they might have.
Deno `bundle | compile` in my opinion should mimic the behavior of compilers such as parcel or webpack in Node Land, in fact that was my first impression when I first met this project.
Commands such as those should translate into one or more operations that resemble what we are trying to achieve manually here, and not expect the user to make manual scripts for building their projects unless they require a really specific setting. Without leading into magic territory, to resolve an app structure and produce and efficient build should be no manual task
username_0: You gotta forgive my ignorance in this matter. I don't know if TypeScript compiler is the one producing the final standalone build or which parts is Deno playing with to allow us to do so.
username_2: So a bit of clarity what happens... First some terminology:
- Deno compiler, a specific "private" worker which contains the TypeScript compiler, implements a "compiler host" which gives the TypeScript compiler access to the "outside world" as well as some other things that we do to make things work that aren't implemented in Rust.
- TypeScript compiler, in particular an implementation of the compiler services to create a TypeScript `ts.Program` and emitting that program. Commonly these are called the compiler services and are effectively what is implemented in `tsc`. This is different than the language services, which are what power things like VSCode and other editors.
- There is a whole bunch of Rust logic which supports the compiler, as well as handles the caching of modules and deals with figuring out what needs to be done to a module before it is injected in the runtime. These don't always have clean names.
Now the different ways these work together:
- `deno run` or `deno fetch`:
- If the main module is a TypeScript file (or checkJs is enabled) and the JavaScript emit for that file isn't in the cache, Rust spins up the snapshot of the Deno compiler.
- The main module is passed to the Deno compiler. The Deno compiler uses a specific API from TypeScript to "preprocess" the files before an AST transform is done to identify all the dependencies.
- The Deno compiler requests the resolution of the modules names for any dependencies of the main module, and then requests that module content from Rust. If Rust doesn't have the "source" for the module, it goes and fetches it before returning it to the Deno compiler. This then recursively happens for all the code for the program, populating the source content in memory awaiting the rest of the process.
- Once all dependencies have been loaded, the Deno compiler does a `ts.createProgram()` passing in the main module. This effectively does an AST parse of all of the code. There are several classes of TypeScript errors that can surface at this point, mostly having to do with the configuration of TypeScript. They aren't very common for us because of the opinionated way we do this. This resolves all the types in memory, but doesn't "enforce" the type checking.
- We attempt to get any diagnostics from the TypeScript compiler, and if there are any that we aren't ignoring, we send them to Rust and they are treated as terminal. Rust logs them out to stderr.
- We then do an emit of the program. We don't actually care about the emit result directly, because the TypeScript compiler attempt to "write out" each file in the emit, and we hook that in the Deno compiler and write that to the cache.
- `deno run` will continue on, loading the main module as an ESM into the isolate and filling any dependencies of that module from the cache.
- `deno bundle`:
- Very similar to `deno run` except that the Deno compiler is always spun up.
- We do care about the emit, as the emit is a single file, which we then add a loader to and analyse the exports of the main module, and re-export those out of the bundle module. The output of the emit is passed back to Rust to be either written to file or output to stdout.
- `Deno.compile()`:
- Works like `deno fetch` but has a slightly different default configuration of the compiler so it doesn't short circuit when some of the content is JavaScript, in that it always analyses JavaScript dependencies and allows JavaScript as part of the program. It also supports a situation where all the sources are provided, instead of being fetched externally to Deno. If the sources are provided, then it does not use the Deno caching mechanisms in Rust (or the Deno module resolution logic).
- When we introduced the API we discussed using `Deno.fetch()` but the problem is that `deno fetch` is actually really poorly named, and so doing that doesn't make sense, and `deno compile` isn't right either, because that would be as said be expected to generate a stand alone binary. `Deno.transpile()` is also imperfect as well, since it does more than just transpile. The justification for `Deno.compile()` though is that this is what we call it internally (it is a `RuntimeCompile` versus a `Compile`) and it aligns to generally what `tsc main.ts` does (though a lot more opinionated).
- `Deno.bundle()`:
- Works like `deno bundle` but again slightly different config and the support for internal sources. This is though properly aligned and effectively `deno bundle` and `Deno.bundle()` are pretty much the same.
- `Deno.transpileOnly()`:
- Not something we use elsewhere in Deno, but a common workflow that is likely needed or useful in user land.
- Uses the `ts.transpileModule()` API, which basically just does an AST parse and an emit transform without any syntax/type checking by the TypeScript compiler. Again, this is like the Babel support for TypeScript, it just strips the types and transforms the syntax. This is likely the first API we could move directly to SWC and not use the compiler at all. It would be super super fast if we did that. We can't move other bits to SWC any time soon, because of the overhead of double parsing, so I would be really reticent to merge the APIs at the moment as it really complicates it and likely there would be differences in the output between the two if we did that.
username_0: If merging the API's is such a hassle then probably we shouldn't do that, but rather provide a simplified version (maybe named differently and treat `compile`, `bundle` and `transpileOnly` as low level APIs) that uses them to accomplish this same goal.
username_0: In the meanwhile, some problems regarding the `compiler` should be addressed. Like not being able to use import-maps to make remote code work. That renders compilation only being able to reach the application and not the modules it depends on.
username_2: This was resolved in Deno 1.5.1.
Status: Issue closed
|
swaywm/swaylock | 481882941 | Title: swaylock shows cursor
Question:
username_0: When the screen has been locked with swaylock, the cusor still remains on the screen (but cant be moved).
From this release issue on sway https://github.com/swaywm/sway/issues/1735 I thought I could use the `hide_cursor` feature mentioned to work round this issue. But it seems like it might have been removed before sway 1.0 was released because I cant seem to find it in the man page or get it working with `swaymsg`
Might be related to #32 but the cursor does not dissappear once moved like mentioned there.
Answers:
username_1: I have a related issue where the cursor is shown, and can be moved; because the mouse is connected to the monitor which I shut off typically when I lock the screen, and when that monitor shuts off, so does the built-in USB hub. |
apollographql/apollo-server | 1096080910 | Title: It shows It appears that you might be offline. POST to this endpoint to query your graph when I access http://localhost:9852/graphql
Question:
username_0: After I upgrade apollo-server from
"apollo-server": "^2.25.0",
"apollo-server-express": "^2.25.0",
To
"apollo-server": "^3.6.1",
"apollo-server-core": "^3.6.1",
"apollo-server-express": "^3.6.1",
When I access http://localhost:9852/graphql to check queries. I got
-------------------------
Welcome to Apollo Server
It appears that you might be offline. POST to this endpoint to query your graph:
curl --request POST \
--header 'content-type: application/json' \
--url '' \
--data '{"query":"query { __typename }"}'
-------------------------
I don't know how to do next step, anyone help me to resolve the issues.
Best wishes
Answers:
username_0: I find the issue becuase helmet blocked some js come from CDN. close it
Status: Issue closed
|
oppia/oppia | 273164826 | Title: We should link directly from exploration gallery tiles to the editor
Question:
username_0: Currently if you want to edit an exploration in the gallery, you have to click on it to open the learner view, then click on the NavBar to get to the editor view (which will also open in a new tab).
It would be helpful to have a direct link to the editor, at least for explorations over which I have edit rights. (I have a feeling we used to have this at some point in the past.)
Status: Issue closed
Answers:
username_1: I don't think we should do this, actually. The creator dashboard is the place to go for creators to access their explorations, not the library, and if that page is not intuitive then we should make it more so rather than adding functionality to the library page. (Note, btw, that the "gallery" is now called the "library".)
So, closing for now, but please open a new issue if you have suggestions for improving the creator dashboard and making it more usable as a creator's "home page". |
GIS4WRF/gis4wrf | 418800190 | Title: Standard longitude bug
Question:
username_0: **Describe the bug**
I have lambert grid where the standard longitude is -10, whereas the center_point longitude = 2. In GIS4WRF, I only have the possibility to set the center_longitude. Inspection of namelist.wps learned me that the standard longitude is set to 2 as well. This is incorrect.
Changing the standard longitude in namelist.wps, and then "import from namelist" did not show the correct grid, aligned with the -10 longitude line.
**Expected behavior**
Extra box to provide the standard longitude. In the above example, the grid should be aligned (by rotation) with the longitude line -10, and the center_longitude of the grid should be equal to 2.
Answers:
username_1: You're absolutely right. Thanks for reporting this. I'll do a few more fixes and then attach a pre-release package here (a zip file that you can install in QGIS), so that you can test it before release.
username_1: @username_0 I fixed the issue and also added support for Mercator and Polar Stereographic projections. If you want to try it before the official release, please download the following zip file:
[gis4wrf-0.14.0-preview.zip](https://github.com/GIS4WRF/gis4wrf/files/2948574/gis4wrf-0.14.0-preview.zip). To install it, go to the plugin manager of QGIS and on the left side do "Install from ZIP".
Status: Issue closed
username_0: Thanks! It works like a charm :) |
xcmyz/FastSpeech | 520683018 | Title: minor issue that might prevent some trouble later on
Question:
username_0: Hello,
Thanks a lot for publishing this amzaing repo.
I just have a little concern that I would like to let you know.
I see that under the 'paper' directory you have uploaded some papers.
I don't think that is a safe thing to do regarding copyrights.
The copyrights for each papers might differ, but just to be safe, how about replacing the papers with a link to it or just name of the papers???
Answers:
username_1: Thank you very much
Status: Issue closed
|
geopandas/contextily | 996792046 | Title: urllib3 connection error
Question:
username_0: I am suddenly having issues with base maps timing out. Nothing has changed in my code but this is now happening from home and the office (so, not a location thing). Odd, any ideas? Have you seen this?
````
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/jamie/reportgeneration/venv/lib/python3.8/site-packages/requests/adapters.py", line 439, in send
resp = conn.urlopen(
File "/home/jamie/reportgeneration/venv/lib/python3.8/site-packages/urllib3/connectionpool.py", line 755, in urlopen
retries = retries.increment(
File "/home/jamie/reportgeneration/venv/lib/python3.8/site-packages/urllib3/util/retry.py", line 574, in increment
raise MaxRetryError(_pool, url, error or ResponseError(cause))
urllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='a.basemaps.cartocdn.com', port=443): Max retries exceeded with url: /light_all/14/8250/5410.png (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f7dd93c6850>: Failed to establish a new connection: [Errno 110] Connection timed out'))
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "code/app.py", line 38, in <module>
landcoverMap(report, reportData)
File "/home/jamie/reportgeneration/code/maps/landcover.py", line 8, in landcoverMap
map1.addBasemap(None)
File "/home/jamie/reportgeneration/code/map.py", line 64, in addBasemap
cx.add_basemap(ax, source=baseimage, crs="EPSG:27700", zorder=1, alpha=1, attribution_size=2)
File "/home/jamie/reportgeneration/venv/lib/python3.8/site-packages/contextily/plotting.py", line 143, in add_basemap
image, extent = bounds2img(
File "/home/jamie/reportgeneration/venv/lib/python3.8/site-packages/contextily/tile.py", line 248, in bounds2img
image = _fetch_tile(tile_url, wait, max_retries)
File "/home/jamie/reportgeneration/venv/lib/python3.8/site-packages/joblib/memory.py", line 591, in __call__
return self._cached_call(args, kwargs)[0]
File "/home/jamie/reportgeneration/venv/lib/python3.8/site-packages/joblib/memory.py", line 534, in _cached_call
out, metadata = self.call(*args, **kwargs)
File "/home/jamie/reportgeneration/venv/lib/python3.8/site-packages/joblib/memory.py", line 761, in call
output = self.func(*args, **kwargs)
File "/home/jamie/reportgeneration/venv/lib/python3.8/site-packages/contextily/tile.py", line 303, in _fetch_tile
request = _retryer(tile_url, wait, max_retries)
File "/home/jamie/reportgeneration/venv/lib/python3.8/site-packages/contextily/tile.py", line 446, in _retryer
request = requests.get(tile_url, headers={"user-agent": USER_AGENT})
File "/home/jamie/reportgeneration/venv/lib/python3.8/site-packages/requests/api.py", line 75, in get
return request('get', url, params=params, **kwargs)
File "/home/jamie/reportgeneration/venv/lib/python3.8/site-packages/requests/api.py", line 61, in request
return session.request(method=method, url=url, **kwargs)
File "/home/jamie/reportgeneration/venv/lib/python3.8/site-packages/requests/sessions.py", line 542, in request
resp = self.send(prep, **send_kwargs)
File "/home/jamie/reportgeneration/venv/lib/python3.8/site-packages/requests/sessions.py", line 655, in send
r = adapter.send(request, **kwargs)
File "/home/jamie/reportgeneration/venv/lib/python3.8/site-packages/requests/adapters.py", line 516, in send
raise ConnectionError(e, request=request)
requests.exceptions.ConnectionError: HTTPSConnectionPool(host='a.basemaps.cartocdn.com', port=443): Max retries exceeded with url: /light_all/14/8250/5410.png (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f7dd93c6850>: Failed to establish a new connection: [Errno 110] Connection timed out'))
````
Answers:
username_0: It looks like an issue with the fastly cdn, now and again it works for me but it is very rare.
username_1: Thanks for reporting @username_0. As I understand, that is related to the provider rather than `contextily`. It might be worth to track availability over at [`xyzservices`](https://github.com/geopandas/xyzservices)? I'm closing for now but feel free to open another issue over there.
Status: Issue closed
|
ErwinKomen/RU-passim | 780271205 | Title: Literature: Zotero syncing not working?
Status: Issue closed
Question:
username_0: This has been resolved, together with issue #117
Answers:
username_0: This has been resolved, together with issue #117
username_0: This is not resolved yet, unfortunately.
Problem:
1. Synchronisation goes well
2. Data ends up in `data` field of the record, but `full` and `abbr` are not filled (nor is `year` filled)
1. In fact, there are numerous entrances where short, abbr and full are empty - how can these be found?
2. I guess the calculation of these fields should take place right after performing a synchronisation or even as part of synchronisation
Status: Issue closed
|
davidgohel/ggiraph | 234062941 | Title: Possible bug with Shiny Flexdashboard - Chart does not display even though it displays in its ggplot format.
Question:
username_0: I'm trying to use `ggiraph` in a shiny flexdashboard, but I was running into an issue where the `ggiraph` chart does not appear even though the `ggplot` chart does.
Here's a reproducible example of the issue I'm running into:
---
title: "Untitled"
output:
flexdashboard::flex_dashboard:
orientation: columns
vertical_layout: fill
runtime: shiny
---
```{r setup, include=FALSE}
library(flexdashboard)
library(ggiraph)
library(magrittr)
dataset = iris
dataset$tooltip = dataset$Species
dataset$clickjs = paste0("alert(\"",dataset$Species, "\")" )
```
Inputs {.sidebar}
---
```{r}
selectInput("Species", "Species", choices = c("setosa", "versicolor", "virginica"))
filtered_df <- reactive({
dataset[dataset$Species == input$Species, ]
})
```
Column
-----------------------------------------------------------------------
### Chart
```{r}
plot <- reactive({
ggplot(filtered_df(), aes(x = Sepal.Length, y = Petal.Width,
color = Species, tooltip = tooltip, onclick = clickjs) ) +
geom_point_interactive() + theme_light()
})
renderggiraph({
ggiraph(code = {print(plot())}, flexdashboard = TRUE)
})
```
Please let me know if you need more details.
Answers:
username_1: Hello,
Thanks for reporting that. I don't know yet if this can be fixed, it works when not programming with Rmd but with R standard code.
David
username_1: Hello
Problem should be fixed now (you don't need `flexdashboard=TRUE` anymore)
---
title: "Untitled"
output:
flexdashboard::flex_dashboard:
orientation: columns
vertical_layout: fill
runtime: shiny
---
```{r setup, include=FALSE}
library(flexdashboard)
library(ggiraph)
library(magrittr)
dataset = iris
dataset$tooltip = dataset$Species
dataset$clickjs = paste0("alert(\"",dataset$Species, "\")" )
```
Inputs {.sidebar}
---
```{r}
selectInput("Species", "Species", choices = c("setosa", "versicolor", "virginica"))
filtered_df <- reactive({
dataset[dataset$Species == input$Species, ]
})
```
Column
-----------------------------------------------------------------------
### Chart
```{r}
plot <- reactive({
ggplot(filtered_df(), aes(x = Sepal.Length, y = Petal.Width,
color = Species, tooltip = tooltip, onclick = clickjs) ) +
geom_point_interactive()
})
renderggiraph({
ggiraph(code = {print(plot())})
})
```
Status: Issue closed
|
ClickHouse/ClickHouse | 946100143 | Title: remove experimental status of window functions
Question:
username_0: Blocking issues:
[ ] broken null handling, need to refactor to add deeper separation between window and aggregate functions https://github.com/ClickHouse/ClickHouse/issues/26115
[ ] expression analyzer woes, not all expression dependencies are handled properly https://github.com/ClickHouse/ClickHouse/issues/24892
[ ] a mystery ubsan failure that doesn't reproduce https://github.com/ClickHouse/ClickHouse/issues/24184
Answers:
username_1: After this, will window functions also distribute across nodes?
username_0: We don't have plans for this yet. I think one easy way to distribute would be to calculate window function on shards when the `PARTITION BY` of the window matches the sharding key.
Status: Issue closed
|
GuangchuangYu/scatterpie | 278597197 | Title: Plot order of pies
Question:
username_0: Is it possible to change the order that the pies are plotted?
I'd like to plot in the order of decreasing pie radius to ensure that the smallest pies are always printed on top of the larger pies. I tried to order the data frame according to decreasing radius but it appears as if the points are plotted in the order of the x values?
Can I change the plot order?
Thanks for a nice package!
Answers:
username_1: `sorted_by_radius` parameter added in v = 0.0.8.
```r
d = data.frame(x = c(2,2), y = c(2,2.5), A= c(1, 4), B=c(4, 1), r = c(1, 3), g = letters[1:2])
ggplot() + geom_scatterpie(aes(x, y, group=g, r=r), data=d, cols=c('A', 'B'))
ggplot() + geom_scatterpie(aes(x, y, group=g, r=r), data=d, cols=c('A', 'B'), sorted_by_radius=T)
```
username_0: Wonderful. Thanks!
Status: Issue closed
|
rancher/rke2 | 961829767 | Title: RKE2 not starting up services on RHEL8
Question:
username_0: **Environmental Info:**
RKE2 Version:
[myuser@vm1 ~]$ rke2 -v
rke2 version v1.21.3+rke2r1 (2ed0b0d1b6924af4414393cd1796c174a1ff5352)
go version go1.16.6b7
Node(s) CPU architecture, OS, and Version:
Linux vm1 4.18.0-147.51.2.el8_1.x86_64 #1 SMP Thu Jul 8 06:09:25 EDT 2021 x86_64 x86_64 x86_64 GNU/Linux
Cluster Configuration:
1 VM, Red Hat Enterprise Linux 8, latest patch level, 2
vCPU, 8GB RAM
/etc/rancher/rke2/config.yaml:
debug: true
selinux: true
**Describe the bug:**
rke2-server does not seem to be able to start its components.
**Steps To Reproduce:**
**Added some iptable entries**
sudo iptables -A INPUT -p tcp --dport 6443 -j ACCEPT
sudo iptables -A INPUT -p tcp --dport 2379 -j ACCEPT
sudo iptables -A INPUT -p tcp --dport 2380 -j ACCEPT
**Installed Docker**
sudo dnf config-manager --add-repo=https://download.docker.com/linux/centos/docker-ce.repo
sudo dnf install docker-ce --nobest -y
sudo systemctl start docker
sudo systemctl enable docker
sudo usermod -aG docker $USER
**Installed and started RKE2 server**
Run the installer: curl -sfL https://get.rke2.io | sh -
Enable the rke2-server service: systemctl enable rke2-server.service
Start the service: systemctl start rke2-server.service
**Expected behavior:**
RKE2 server to startup components like etcd, kube-apiserver
**Actual behavior:**
connection errors in log (see attached log file) for 2379 (etcd) and 6443 (kube-apiserver)
**Additional context / logs:**
[log-rke2 - Copy.txt](https://github.com/rancher/rke2/files/6938937/log-rke2.-.Copy.txt)
Answers:
username_1: Why are you installing and starting Docker before RKE2? RKE2 uses its own embedded containerd; there is no need to install docker beforehand and in fact you are better off not.
It also looks like you've not installed the required selinux packages; normally the installer does this for you so I'm confused how this could happen:
`Aug 05 13:15:38 vm1 rke2[15268]: time="2021-08-05T13:15:38Z" level=warning msg="SELinux is enabled for rke2 but process is not running in context 'container_runtime_t', rke2-selinux policy may need to be applied"`
If removing docker and installing the selinux packages does not resolve the error, see if there's anything interesting in the containerd log at `/var/lib/rancher/rke2/agent/containerd/containerd.log`.
username_0: @username_1 thats good to know. I was just expecting it, as it was a requirement for rke (binary) and rke in rancher, as far as I know.
I now completly started from scratch, these are the steps I did:
[rke2-server.log](https://github.com/rancher/rke2/files/6943593/rke2-server.log)
systemctl disable firewalld
vim /etc/NetworkManager/conf.d/rke2-canal.conf
systemctl reload NetworkManager
dnf upgrade -y
reboot
curl -sfL https://get.rke2.io | sh -
systemctl enable rke2-server.service
systemctl start rke2-server.service
I found another possible issue:
username_1: etcd still isn't starting, can you check the containerd log file as requested above?
username_0: something seems to be off for CNI.
[containerd.log](https://github.com/rancher/rke2/files/6953037/containerd.log)
@username_1 have you ever experienced something similar?
username_1: ```console
time="2021-08-06T06:36:38.554910732Z" level=info msg="CreateContainer within sandbox \"214255c37689a276a424b37dffd2b03b9f7c641b68045f2180b2413dd5275d51\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"84b2f6932660193fa18d0a555b43e1b4444592ff7b6898<KEY>cc05bd\""
time="2021-08-06T06:36:38.555421235Z" level=info msg="StartContainer for \"84b2f6932660193fa18d0a555b43e1b4444592ff7b6898aaf<KEY>cc05bd\""
time="2021-08-06T06:36:38.722580775Z" level=info msg="StartContainer for \"84b2f6932660193fa18d0a555b43e1b4444592ff7b6898aaf9e5c86f01cc05bd\" returns successfully"
time="2021-08-06T06:36:52.714149951Z" level=info msg="RunPodsandbox for &PodSandboxMetadata{Name:etcd-adbsg-fzag-k8s-vm1,Uid:985840a449fab27fbd1831f57843061a,Namespace:kube-system,Attempt:0,}"
time="2021-08-06T06:36:52.792435138Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/k3s/containerd/io.containerd.runtime.v2.task/k8s.io/e26f8170d80ada8ab2531a87962fec1edbfd7b708c416ff914d2c5e1d6cf662e pid=6507
time="2021-08-06T06:36:52.908778362Z" level=info msg="shim disconnected" id=e26f8170d80ada8ab2531a87962fec1edbfd7b708c416ff914d2c5e1d6cf662e
time="2021-08-06T06:36:52.908851662Z" level=error msg="copy shim log" error="read /proc/self/fd/28: file already closed"
time="2021-08-06T06:36:52.943644879Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:etcd-adbsg-fzag-k8s-vm1,Uid:985840a449fab27fbd1831f57843061a,Namespace:kube-system,Attempt:0,} failed, error" error="failed to create containerd task: OCI runtime create failed: container_linux.go:380: starting container process caused: process_linux.go:545: container init caused: failed to set /proc/self/attr/keycreate on procfs: write /proc/self/attr/keycreate: invalid argument: unknown"
```
I've only seen this one other time, in https://github.com/rancher/rke2/issues/851#issuecomment-814508133 - and this user was setting a custom-data dir and also using a standalone containerd, both of which complicate things on selinux-enabled hosts. Can you start over on a clean host that doesn't have system docker or containerd installed, and ensure that you're using the RPM install with all the required selinux packages?
username_0: I did start again from a clean RHEL 8.1 VM in Azure.
I used the script in the quick start guide to install it. That uses rpm as well, right?
username_2: Do we need to mention this in the support matrix or rke2 docs?
Looks like we need RHEL 8.5 to get SELinux working in this case.
username_3: Hi all,
i'm also facing same issue while deploying rke2 using ansible on RHEL 8.2,
does anyone get some solution for that?
username_4: I am having the same issue on RedHat 8.4 on AWS, it looks like the error is:
```bash
sudo systemctl status rke2-server
● rke2-server.service - Rancher Kubernetes Engine v2 (server)
Loaded: loaded (/usr/lib/systemd/system/rke2-server.service; enabled; vendor preset: disabled)
Active: activating (auto-restart) (Result: exit-code) since Thu 2022-03-03 15:36:37 UTC; 728ms ago
Docs: https://github.com/rancher/rke2#readme
Process: 17941 ExecStopPost=/bin/sh -c systemd-cgls /system.slice/rke2-server.service | grep -Eo '[0-9]+ (containerd|kubelet)' | awk '{print $1}' | xargs -r kill (code=exited, status=0/SUCCESS)
Process: 17938 ExecStartPre=/bin/sh -xc ! /usr/bin/systemctl is-enabled --quiet nm-cloud-setup.service (code=exited, status=1/FAILURE)
```
It looks like what is failing is the command before the execution of rke2:
```
[ec2-user@ip-172-31-44-245 ~]$ /bin/sh -xc ! /usr/bin/systemctl is-enabled --quiet nm-cloud-setup.service
[ec2-user@ip-172-31-44-245 ~]$ echo $?
1
```
causing the service's executable to not run ?
On the same node, just running `rke2 server` results in a successful bootstrapping.
## Steps to reproduce:
- AWS EC2 instance using the AMI `ami-06ec8443c2a35b0ba` on `eu-central-1` , this instance is based on RHEL 8.4 with kernel `4.18.0-305.el8.x86_64` .
- All settings default: selinux is enforcing, firewalld is not enabled out-of-the-box
- Then try to install RKE2:
```bash
[ec2-user@ip-172-31-44-245 ~]$ curl -sfL https://get.rke2.io | sudo sh -
[ec2-user@ip-172-31-44-245 ~]$ sudo systemctl enable rke2-server
[ec2-user@ip-172-31-44-245 ~]$ sudo systemctl start rke2-server
```
Using this user-data should result in same behavior:
```yaml
#cloud-config
runcmd:
- curl -sfL https://get.rke2.io | sudo sh -
- sudo systemctl enable rke2-server
- sudo systemctl start rke2-server
```
username_2: @username_4 Not the same issue we discussed in the original thread.
Looks like you need to make sure the known issues are addressed as per the rke2 doc.
https://docs.rke2.io/known_issues/#networkmanager
```
In some operating systems like RHEL 8.4, NetworkManager includes two extra services
called nm-cloud-setup.service and nm-cloud-setup.timer.
These services add a routing table that interfere with the CNI plugin's configuration.
Unfortunately, there is no config that can avoid that as explained in the [issue](https://github.com/rancher/rke2/issues/1053).
Therefore, if those services exist, they should be disabled and the node must be rebooted.
```
username_4: Thanks for the clarification! |
thingsboard/thingsboard | 937707638 | Title: How to check the type of an Entity and change dashboard state based on the Type
Question:
username_0: If I have an entity list which contains multiple entities, of different types, which all show a single measurement (so all can be shown no matter the type) is there a way to switch dashboard states based on the type of the clicked entity so that I can display the diagnostic data for each entity (which differ significantly) using a type based state for each differing type?
There's an answer to a similar question here which implies its possible, but unfortunately didn't elaborate on how to do it: https://github.com/thingsboard/thingsboard/issues/3758#issuecomment-731016259
An example:
EntityList
-----------
Thermometer 25C
Thermometer2 45C
Light Sensor 100lux
When I click any of the thermometers I want to navigate to a dashboard state (thermometer), which has specific data and graphs for a thermometer (and specifically the one selected).. but if I select a Light Sensor.. then I want to go to a (Light Sensor state), with different data and attributes.
Answers:
username_1: Hello, @username_0
To switch states depending on the device profile you need to create a custom action. The following code will allow you to get the type (profile) of a device and then open an appropriate dashboard state:
```
let $injector = widgetContext.$scope.$injector;
$injector.get(widgetContext.servicesMap.get('entityService')).getEntity(entityId.entityType,
entityId.id).subscribe(function(entity) {
let entitySubType = entity.type;
if (entitySubType === 'therm') {
openDashboardState('thermometer_state');
} else if (entitySubType === 'light') {
openDashboardState('light_sensor_state');
}
});
function openDashboardState(statedId) {
let params = {
entityId: entityId,
entityName: entityName,
entityLabel: entityLabel
};
widgetContext.stateController.openState(statedId,
params, false);
}
```
username_0: Hi, @username_1
That's a perfect answer, thank you very much for that, I had been struggling for a long time and had come up with a method that partially worked, using a complete reload and some JSON crafting to make the state parameters.. but your way is much much nicer and doesn't require the reload (or hardcoded dashboard addressing)
Thanks,
Sam
Status: Issue closed
|
sandstorm-io/vagrant-spk | 338412846 | Title: Error when using VirtualBox 5.2 and Vagrant 2.1.2
Question:
username_0: I installed VirtualBox 5.2 on a Windows machine and run `vagrant-spk`, then `Vagrant` says 5.2 is not supported. So I update `Vagrant` to 2.1.2, then `vagrant-spk vm up` throws error:
```
C:\Users\Romulus\WebstormProjects\sandmd>vagrant-spk vm up
Calling 'vagrant' 'up' in C:\Users\Romulus\WebstormProjects\sandmd\.sandstorm
There was an error loading a Vagrantfile. The file being loaded
and the error message are shown below. This is usually caused by
a syntax error.
Path: C:/Users/Romulus/WebstormProjects/sandmd/.sandstorm/Vagrantfile
Line number: 13
Message: RuntimeError: can't modify frozen String
[31mCommand failed with a non-zero exit status (1).[0m
```
That line seems to be a hack, so I remove it. Then the error becomes:
```
C:\Users\Romulus\WebstormProjects\sandmd>vagrant-spk vm up
Calling 'vagrant' 'up' in C:\Users\Romulus\WebstormProjects\sandmd\.sandstorm
Bringing machine 'default' up with 'virtualbox' provider...
The version of powershell currently installed on this host is less than
the required minimum version. Please upgrade the installed version of
powershell to the minimum required version and run the command again.
Installed version: 2
Minimum required version: 3
[31mCommand failed with a non-zero exit status (1).[0m
```
I can't find where to update Powershell, but it seems like `vagrant-spk` needs an update?
Answers:
username_1: @kentonv hasn't pushed a Windows release with the fix for the first issue. Removing it is an adequate fix, I did this: https://github.com/sandstorm-io/vagrant-spk/commit/e0c1d4c086ad207af53f7b7f49ec6f656aec4b8d
Your PowerShell issue seems to be an issue with the version of Windows you are on, and how Vagrant feels about it. You can either upgrade your PowerShell version or downgrade Vagrant to 1.9.6. Here's where you can find the WMF download for your version of Windows: https://docs.microsoft.com/en-us/powershell/scripting/setup/installing-windows-powershell?view=powershell-6
Status: Issue closed
username_0: @username_1 Thank you, I updated PowerShell to 5.0 and the error is fixed. |
zalando/postgres-operator | 975490631 | Title: K8s: Restart in the middle of reinit of replica db node -> manual fixing of db node needed
Question:
username_0: Such needs manual fixing.
Larger the database → Higher the probability for these kinds of problems.
In such situation, Patroni REST interface of the problematic DB pod / DB node does not seem to be accessible. ( To be verified again still. )
E.g. such access would be needed, if we would like to implement our own automatic reinitialization script for such case.
Q: What could be done for this problem?
Attached Logs, CRD and manifest files here for reference :
[reinit-error.txt](https://github.com/zalando/postgres-operator/files/7020839/reinit-error.txt)
[manifest.txt](https://github.com/zalando/postgres-operator/files/7020879/manifest.txt)
[crd.txt](https://github.com/zalando/postgres-operator/files/7020911/crd.txt) |
DataViva/dataviva-site | 54319652 | Title: Missing colour scale in tree maps
Question:
username_0: When you change the colour visualization of the tree map, the colour scale is missing. The user is not informed what the shades of red and green means.

Answers:
username_1: I just realized this as well, as I was incorporating the new d3plus.
We're going to be using the built-in color scale of the new d3plus for this. Here's a screenshot of my working environment to show what it will look like (I already got it hooked up!):

Status: Issue closed
|
ropensci/RNeXML | 771272427 | Title: Warning message: select_() is deprecated as of dplyr 0.7.0
Question:
username_0: R version 4.0.2 (2020-06-22)
Platform: x86_64-apple-darwin17.0 (64-bit)
Running under: macOS Catalina 10.15.7
Matrix products: default
BLAS: /System/Library/Frameworks/Accelerate.framework/Versions/A/Frameworks/vecLib.framework/Versions/A/libBLAS.dylib
LAPACK: /Library/Frameworks/R.framework/Versions/4.0/Resources/lib/libRlapack.dylib
locale:
[1] en_US.UTF-8/en_US.UTF-8/en_US.UTF-8/C/en_US.UTF-8/en_US.UTF-8
attached base packages:
[1] stats graphics grDevices utils datasets methods base
other attached packages:
[1] rphenoscape_0.2.15
loaded via a namespace (and not attached):
[1] Rcpp_1.0.4.6 xml2_1.3.2 magrittr_1.5 uuid_0.1-4 tidyselect_1.1.0 ape_5.3 lattice_0.20-41
[8] R6_2.4.1 rlang_0.4.8 stringr_1.4.0 httr_1.4.1 plyr_1.8.6 dplyr_1.0.2 tools_4.0.2
[15] parallel_4.0.2 RNeXML_2.4.3 grid_4.0.2 nlme_3.1-148 ellipsis_0.3.0 lazyeval_0.2.2 tibble_3.0.1
[22] lifecycle_0.2.0 crayon_1.3.4 tidyr_1.0.2 purrr_0.3.4 reshape2_1.4.4 vctrs_0.3.4 curl_4.3
[29] glue_1.4.0 stringi_1.4.6 compiler_4.0.2 pillar_1.4.4 generics_0.0.2 XML_3.99-0.3 jsonlite_1.6.1
[36] pkgconfig_2.0.3
``` |
GABowers/Capstone-Application | 708155544 | Title: Fix System.FormatException: Input string was not in a correct format. in ContainerController.ParseOperation (String input) (H:\Users\GAB\Documents\GitHub\Capstone-Application\Capstone Application\ContainerController.cs:68)
Question:
username_0: ### Version 1.0.0.0(1.0.0.0) ###
### Stacktrace ###
Capstone_Application.ContainerController.ParseOperation(String input) in H:\Users\GAB\Documents\GitHub\Capstone-Application\Capstone Application\ContainerController.cs:line 68;Capstone_Application
Capstone_Application.ContainerController..ctor(AgentContainerSetting input) in H:\Users\GAB\Documents\GitHub\Capstone-Application\Capstone Application\ContainerController.cs:line 22;Capstone_Application.ContainerController
Capstone_Application.AgentController..ctor(Int32 agentX, Int32 agentY, Int32 state, CA parent, BlankGrid cell) in H:\Users\GAB\Documents\GitHub\Capstone-Application\Capstone Application\AgentController.cs:line 45;Capstone_Application.AgentController
Capstone_Application.ControllerScript.EditGrid(Int32 xValue, Int32 yValue, PixelBox container, Int32 buttonPressed, Int32 state) in H:\Users\GAB\Documents\GitHub\Capstone-Application\Capstone Application\ControllerScript.cs:line 587;Capstone_Application
Capstone_Application.Form1.innerPictureBox_MouseUp(Object sender, MouseEventArgs e) in H:\Users\GAB\Documents\GitHub\Capstone-Application\Capstone Application\Form1.cs:line 282;Capstone_Application
Capstone_Application.Program.Main() in H:\Users\GAB\Documents\GitHub\Capstone-Application\Capstone Application\Program.cs:line 25;Capstone_Application
### Reason ###
System.FormatException: Input string was not in a correct format.
### Link to App Center ###
* [https://appcenter.ms/users/username_0/apps/Capstone/crashes/errors/4132482664u](https://appcenter.ms/users/username_0/apps/Capstone/crashes/errors/4132482664u)
Status: Issue closed
Answers:
username_0: 242b33f79e811c46ef8439ed1eb74bc06fbf7d6d |
stoicflame/enunciate | 129817559 | Title: [Regression] Only one ResourceMethodSignature query param appears in documentation.
Question:
username_0: When I use custom query parameters @RSParam + @ResourceMethodSignature only the first explitcit param appears in the documentation.
For example
```java
@Path("permissions")
public class PermissionDao {
/**
* Gets all {@link Permission} satisfying the given query restriction. If no
* query parameters are specified, all entities will be returned.<br>
*
* @generated
* @param queryParams the QueryParams map with the parameters to use in the query
* @param outMetadata the IOutMetadata with the response metadata
* @RSParam l (optional) Max amount of response elements (limit). Example <i>?l=10</i>
* @RSParam o (optional) Offset in the data storage of the first response element. Example <i>?o=10</i>
* @RSParam by (optional) Ordering criteria of response elements. Example ?by=aFieldName|desc</i>
* @RSParam filter (optional) Filter criteria of the response. Example <i>?aFieldName=1|2|3</i>
* @return a list of {@link Permission}
*/
@ResponseHeaders({
@ResponseHeader(name = "meta-totalCount", description = "Number of existing elements with the query restriction."),
@ResponseHeader(name = "meta-partialCount", description = "Number of elements in the response."),
@ResponseHeader(name = "meta-offset", description = "Offset in the data storage of the first element of the response."),
@ResponseHeader(name = "meta-next", description = "Offset of the next existing element in data storage not included in the response.")
})
@ResourceMethodSignature(queryParams = { @QueryParam("l"), @QueryParam("o"), @QueryParam("by"), @QueryParam("filter") },
output = WorkflowModelPackage.PermissionXMLContainer.class
)
@Override
@GET
@Produces({ MediaType.APPLICATION_JSON, MediaType.APPLICATION_XML })
@NonNull
public List<Permission> getAll(@Context @NonNull IQueryParams queryParams, @Context @Nullable IOutMetadata outMetadata) {
return super.getAll(queryParams, outMetadata);
}
}
```
This is a regression due #76 fix commit.
Annotation that comes ResourceMethodSignature (wrapped in a ExplicitResourceParameter) results always equals between them, so only one of them is is putted in the params TreeSet in `ResourceMethod.loadResourceParameters() `method. The equals/hashCode/compareTo methods of ResourceParameter does not consider extended class that override parameterName ed typeName properties.
Answers:
username_1: Thanks for the report. Assigned to 2.3.
username_0: I have a patch but now all Request Param has the same description the resource method. This is due to no reference to `ExplicitResourceParameter.getDovValue()` that has the correct description. It's taken parsing again the javadoc using JavaDoc structure in the `ParameterImpl`
Status: Issue closed
username_1: [Enunciate 2.3 has been released](https://github.com/username_1/enunciate/releases/tag/v2.3.0). |
cloudnativelabs/kube-router | 338382561 | Title: Documentation clarification regarding --allow-privileged=true ?
Question:
username_0: [1] https://kubernetes.io/docs/concepts/workloads/pods/pod/#privileged-mode-for-pod-containers
[2] https://github.com/cloudnativelabs/kube-router/blob/master/docs/generic.md#configuring-the-kubelet
Thanks,
Answers:
username_1: IMHO, instead of using privileged:true, only the required capabilities (https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.11/#capabilities-v1-core ) like NET_ADMIN and probably some few others should be used and documented.
username_2: Can anyone tell which capabilities would be needed?
username_3: Closing as stale.
Status: Issue closed
|
openwrt/luci | 839317920 | Title: Possible to display encryption methods in use for wireless stations?
Question:
username_0: It would be quite useful for troubleshooting problematic stations that do not behave correctly with WPA2/WPA3 transition mode, as not all of the devices would display the exact WPA* mode that they are using. I've tried to acquire this information on the AP with `hostapd_cli` but I only get an empty list. I know it can indeed talk to the hostapd running on the system because `hostapd_cli -iwlan1 status` reports the correct info, but `hostapd_cli -iwlan1 list_sta` returns blank without any errors. Any ideas? |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.