repo_name
stringlengths 4
136
| issue_id
stringlengths 5
10
| text
stringlengths 37
4.84M
|
---|---|---|
mwaskom/seaborn | 87131009 | Title: None
Question:
username_0: Can't you do this already using `FacetGrid.map`? Just pass a function that computes whatever statistic you want and then draw it with `plt.text`.
Answers:
username_0: Oh I see, the question of how to access both levels of the hue variable is orthogonal to whether you use them to draw a plot or compute a statistic.
I would define a little wrapper function that takes three positional inputs: x, y, and hue (similarly to e.g. boxplot) and then have that plot both traces and compute the stat.
username_0: Right, you'd pass that function to `FacetGrid.map`. The trick is just to delegate the "hue" faceting to the mapped function rather than to `FacetGrid` itself.
username_0: Don't use `hue` in the `FacetGrid` constructor.
username_0: But now you have the numeric Series and the categorical Series inside the mapped function, so you can use them to plot both CDFs separately and and compute the test between the two levels of the hue variable.
username_0: Did you get this figured out @olgabot?
Status: Issue closed
|
pjohnkeane/laurajanescott | 104423531 | Title: Line up top of slideshow with the first black line above 'basket'
Question:
username_0: Probably need to adjust margin/padding on page div.
Answers:
username_0: More complicated than that because of the variability of the sidebar height.
Status: Issue closed
username_0: Resolved under https://github.com/username_0/laurajanescott/commit/0732868bb7e3b4b98c3969f21e53e07748e896d1 |
sketchplugins/plugin-directory | 120693329 | Title: Automate-Sketch
Question:
username_0: https://github.com/username_0/Automate-Sketch/
- Add Space Between CJK and Western Character
- Change Typeface for the Latin character set
- Artboard Form Group
- Artboard Form Layers
- Clear Slices
- New Layer Base Slice
- Clean Layer Name
- Clear Guides
Answers:
username_1: This was added in ef33dc256172b4353349a426cfd840c1eebcf31f, a few days ago : )
Status: Issue closed
|
DevZest/WpfDocking | 1007254450 | Title: When holding down SHIFT and mouse left click dockItem header cause NullReferenceException.
Question:
username_0: I comment out `TestDrop` method in `IsShiftKeyDown` property in file `WpfDocking\Src\Windows\Docking\Primitives\DockManager.DragHandler.cs`, it seems ok but I don't know if it has other impacts.
``` c#
private bool IsShiftKeyDown
{
set
{
bool oldValue = GetIsShiftKeyDown(DockControl);
if (oldValue == value)
return;
SetIsShiftKeyDown(DockControl, value);
//TestDrop();
}
}
``` |
wenzhixin/bootstrap-table | 404183033 | Title: Error Show/Hide column with Cookie extension: Cannot set property 'visible' of undefined
Question:
username_0: I'm using the version 1.13.3
And I'm using the Cookie extension because in some table I need it, but in other no.
If I try to hide some column into the table, it return me this error:

You can see it here:

The link of the example is here:
https://jsfiddle.net/7ofnd1s2/9
Thank you for your help.
Answers:
username_1: Duplicate of #4141
Status: Issue closed
|
pulibrary/figgy | 282223253 | Title: Visibility and rights statement based on copyright year
Question:
username_0: For the Brittle Books project, they would like to be able to create a ScannedResource with a bib number, and have the visibility and rights statement set based on the date of the resource (1923 or ealier: open/public domain, 1924 or later: private/in copyright)
Answers:
username_1: Questions:
1. Do we have example records with both side of this date range?
2. Is this only for brittle books? How are they identified? A member of a collection? I assume it isn't everything because we probably don't want to auto-mark everything we have from post-1924 as private.
username_0: Sample records:
* pre 1924 (public domain): 4609321 and 5241365
* 1924- (in copyright): 1605804 and 1234567
This was requested by a single project, and we don't want this for everything (because there are other ways we can have rights to display things). @roelmoon and @joycebcat can correct me, but I think we would only want this for Brittle Books. We can create a Brittle Books collection and have that trigger this behavior. This would probably be a good way to trigger this: a form to create a member of a specific collection could be a good place to hang this logic.
username_1: Are they all going to be in the same collection?
If so, we can key this logic off whether or not an item is in that collection.
Status: Issue closed
|
microsoft/vscode-azurecache | 661986772 | Title: Redis connections not disposed after signing out of Azure account
Question:
username_0: When you sign out of your Azure account, existing Redis connections still continue to be alive, pinging away every 60 seconds.
Need a way to listen for sign out events to disconnect those connections. See https://github.com/microsoft/vscode-azuretools/pull/772<issue_closed>
Status: Issue closed |
tymondesigns/jwt-auth | 83528264 | Title: res is null. $scope.signup angular error
Question:
username_0: 
I have setup this project in local, but on submitting I am getting this error from the angularjs.
Please help me understand this.
Also, I have implemented roles and permission from Entrust package, How can I implement JWT to work with Entrust roles package?
Any help/guide will be highly appreciated.
Answers:
username_1: Looks like CORS is the cause of your issue.
This package is great for this https://github.com/barryvdh/laravel-cors
Status: Issue closed
|
Azure/azure-sdk-for-java | 564353180 | Title: Fit and finish
Question:
username_0: - IndexingParametersExtensions equivalent is missing
- Need to support @search.coverage in AutocompletePagedResponse (it's in the Swagger already)
- Add support for FieldBuilder
- Need to set Accept to application/json;odata.metadata=none in all Documents operations
Answers:
username_1: `@search.coverage` is being added to `AutocompletePagedResponse` in #8303
username_1: Setting `application/json;odata.metadata=none` was added in #8772.
username_1: Spun off `FieldBuilder` into issue #8903.
username_1: @username_2 would static helpers be sufficient to cover `IndexingParametersExtensions`?
username_2: @username_1 That's a question for @username_3
username_3: I'm sorry, I don't see anything named `IndexingParametersExtensions`. Where's that?
username_1: @username_3 It is in the Track 1 SDK for .NET, [IndexingParametersExtensions](https://github.com/Azure/azure-sdk-for-net/blob/master/sdk/search/Microsoft.Azure.Search.Service/src/Customizations/Indexers/Models/IndexingParametersExtensions.cs), looking at that class again I realize it is a static helper class and not a partial class as were a few other change requests.
I'm going to spin this off into a new issue so this one can be closed since half the requests are completed and the other two are disparate concepts that should be tracked separately. This is the new issue to add a helper class for `IndexingParameters`, https://github.com/Azure/azure-sdk-for-java/issues/8908.
Status: Issue closed
username_3: @username_1 thanks for the clarification, I was missing context on this one. Makes sense. |
convox/rack | 206194971 | Title: Lambda Custom Topic Off Node 0.10
Question:
username_0: data: {u'text': u'Another idea: Can we update lambda functions and runtimes in place outside of cloudformation?', u'list': {u'name': u'Big Projects', u'id': u'5874264096c4913948a0ac7d'}, u'board': {u'shortLink': u'XxvWeICx', u'id': u'587425638297353d93b2d4af', u'name': u'Roadmap'}, u'card': {u'idShort': 383, u'shortLink': u'KJsy1hqI', u'id': u'586fa7b88f453663d6688afa', u'name': u'Lambda Custom Topic Off Node 0.10'}}
date: 2017-01-16T23:32:58.898Z
id: 587d582a4fc05ad9fa547bc2
idMemberCreator: 4f26df78114e3e040d62d52d
memberCreator: {u'username': u'nzoschke', u'fullName': u'<NAME>', u'avatarHash': u'53c860bb51950517f3cf01e346cea68d', u'id': u'4f26df78114e3e040d62d52d', u'initials': u'NZ'}
type: commentCard
https://trello.com/c/KJsy1hqI/383-lambda-custom-topic-off-node-0-10<issue_closed>
Status: Issue closed |
muwuding/bill-pnt4 | 940440172 | Title: 测试发布
Question:
username_0:  <br/><b>Click the link below to watch the video: </b><br/>https://drive.google.com/file/d/1h2kq-r2eO4DBUaC1AKm5ISxEql2aSVsH/preview<br/>这就是我的测试发布 |
botpress/botpress | 909929869 | Title: [FEATURE] Display Bot ID in addition to Bot Name
Question:
username_0: **Is your feature request related to a problem? Please describe.**
Currently the BotID is displayed in the Bot URL and in the Config File. I am looking for a way to get the bot ID from the bot name, or from the UI. This is because the bot ID needs to be provided in the REST API request, which would be called in an action.
**Describe the solution you'd like**
We can have the BotID displayed in the Config page where the name, description, status and language are shown:

**Describe alternatives you've considered**
N/A
**Additional context**
N/A
Answers:
username_1: Having a read-only field with botID that is copyable would be perfect here.
Status: Issue closed
username_1: closed by https://github.com/botpress/studio/pull/64 |
homelleon/GameEngine | 247674766 | Title: Textures loaded as GUI don't keep it's natural scale
Question:
username_0: After rendering GUITexture it covers whole the screen if scale was set to (1.0f, 1.0f).
Need to define its scale due to the real size of texture automatically and use users parameters to change it proportionally.
Answers:
username_0: fixed by previous commit and last big GUI update.
https://github.com/username_0/GameEngine/commit/e18daf032881ce5ab46766b0bec790052cad8580
Status: Issue closed
|
Automattic/node-canvas | 46699126 | Title: Question: How expansive is it to create the canvas object?
Question:
username_0: I am currently using **node-canvas** in the context of an Express.js application, which basically generates an image every time a certain endpoint gets called.
After using it for a while and doing multiple load tests, I noticed that creating a canvas and drawing stuff on it, is very expansive and causes the CPU to easily jump to 100% consumption.
Do you have any tips how to us the library in a web server context?
Or avoid these CPU usage peaks?
One of my ideas was to have a canvas object pool or only instantiating only one canvas, not sure if this is doable or would help.
Any feedback is welcome
Thanks username_0
Answers:
username_1: I would recommend having a pool of a few canvas that you can reuse, should work pretty good. I believe the async library has some helper methods for this...
Status: Issue closed
|
dominiksalvet/asus-fan-control | 642336623 | Title: Add support for ASUS TUF FX505DY
Question:
username_0: addresses 1335
temperatures 55 60 62 65 68 72 76 80
tested false
**Additional context**
I have tried finding the right adresses with afc-scout but I haven't found anything, only zeros between adresses 0 and 5000.
Answers:
username_1: Hello @username_0 and welcome to the asus-fan-control community! :wave:
1. Once you set values with `set-temps`, does `get-temps` report them?
2. Have you tried different versions of acpi_call?
3. Do other fan control solutions work (at least partially)?
username_0: I have tried pretty much everything from https://wiki.archlinux.org/index.php/Fan_speed_control but none of it works. I have come to a conclusion that I don't have PWM. The only thing that can change my fan speeds is `echo X >> /sys/devices/platform/asus-nb-wmi/throttle_thermal_policy` where X={0,1,2}. Pressing Fn+f5 cycles through these values similar to Windows. 2 - quiet (but for some reason not as quiet as on W10), 1 - turbo (full speed), and 0 - balanced (about half speed). This only works since Kernel 5.6.
I originally thought that this should work on my laptop because FX505GM is supported but that laptop is Intel+Nvidia and mine is AMD+AMD meaning the important parts are most likely completely different.
This is output from `sensors`: https://pastebin.com/crFTM686
Fan speed is reported correctly and CPU temps are also reported.
username_1: Yeah, it seems so. Even ASUS distinguishes individual models when [downloading drivers](https://www.asus.com/Laptops/ASUS-TUF-Gaming-FX505/HelpDesk_Download/) for FX505 series - see *Please select your CPU or BIOS model name*.
What's more, your model belongs to another series. The mentioned series is only **FX505** series, you have **FX505DY** and those are different. It has its own [support page](https://www.asus.com/Laptops/ASUS-TUF-Gaming-FX505DY/HelpDesk_Download/). Interesting. :thinking:
username_1: Hello, BTW have you tried https://github.com/leonmaxx/asus_fanmode? :-)
username_0: Thanks for the suggestion, I will try it and post results.
username_0: So, the project you linked allows you to automatically select one of the existing profiles (quiet,balanced,turbo) based on the temperature. For me the quiet profile is already too loud and editing it to turn on the fans at a higher temperature is what I would like to do. (For example, now my temperature is ~40° and I can hear the fans from across the room.) Unfortunately setting manual fan speeds is only possible on more expensive models from the ROG line.
username_1: Well... And have you tried https://github.com/FlyGoat/RyzenAdj (there should be more options)? For more information, see #56.
username_0: I have tried it, but there are no options related to fan control. This might be a coincidence, but around the time I used the tool, my power button stopped working (while the laptop is on). I had to reset my BIOS to get it working again so the tool might be touching something it is not supposed to. It might be a good idea to report it there.
username_1: Okay, I will close this issue but feel free to continue posting relevant updates if you want.
Status: Issue closed
|
ianagbip1oti/greedorama | 647044798 | Title: IMPROVE: Adding Attribution
Question:
username_0: I can see that greedorama is using iexcloud and that would necessiate attribution as outlined [here](https://iexcloud.io/docs/api/#attribution). I've taken a look in the repository and have not seen anything regarding that, hence the ticket.<issue_closed>
Status: Issue closed |
projectdiscovery/dnsx | 962630928 | Title: [issue] Missing support for EDNS records
Question:
username_0: **Describe the bug**
`dnsx` has some issue or has no support for `EDNS` record. As you can see from dig command `; EDNS: version: 0, flags:; udp: 1232` which mean that EDNS record are being used and DNS message length is over `512` bytes.
**Dnsx version**
`v1.0.5`
**Screenshot of the error or bug**
- `dig` command for `poc.username_0.xyz`

- `dnsx` command for `poc.username_0.xyz`
<issue_closed>
Status: Issue closed |
ptillet/triton | 838594012 | Title: undefined symbol: compress2
Question:
username_0: I installed the triton by pip install triton==0.2.3; but when I run the code
it return Import Error :
triton/_C/libtriton.so: undefined symbol: compress2
I have libz.so.1 in /usr/lib64 and /usr/lib64 in my $LD_LIBRARY_PATH.
Answers:
username_1: Triton <1 is very old and not really supported anymore. I am working with deepspeed to get it to work with the incoming release.
Can you try pip install -U --pre triton to see if it works with the new tutorials?
username_0: After use `pip install -U --pre triton` it works. But not work with DeepSpeed.
username_1: Yes, I am in the process of getting DeepSpeed to work with Triton v1. There has been a few breaking API changes.
username_0: Thanks for reply. I install triton 0.2.3 by llvm10 successfully. Looking forward your new Deepspeed, thank!
Status: Issue closed
|
code4tottori/curry-naru-kosodate | 135177910 | Title: 保護者はシステムにユーザー登録できる
Question:
username_0: 集める情報はかなりデリケートな内容になるので注意がいるでしょう。特にマイナンバーはかなりまずそう。
それはともかく、このユーザーと、実際に預ける子どもさんがまだ整理されていないように感じました。
タイトル通りユーザーは(アカウント共有するかはともかく)保護者と考えていて、子どもを登録するならまた別と考えていましたけど、子どもをユーザーとして登録する方がいいでしょうか。
Answers:
username_1: 自分の子供を預ける時に預け先に知っていて欲しい情報ですかね
(病児保育は利用したことないので利用時に実際にどのような情報を求められるかはわからないですが)
* 子供の名前
* 子供の生年月日(年齢)
* アレルギーの有無
* その他注意して欲しいこと
* 緊急連絡先:名前 + 電話番号(職場の電話番号なら連絡先も)
* どれかに必ず繋がるように3つくらい?
* 親2人分と祖父、祖母あたりの電話番号
* 緊急でない連絡先:メールとか
username_2: @username_1 さんのコメントをみて、マイナンバーとかキーワードにあってもよさそうだなぁと頭をよぎりました
実際の利用するときの[申請書](http://www.ncn-t.net/morigen/koguma/sinseisho2013.pdf)と[問診票](http://www.ncn-t.net/morigen/koguma/kogumamonsin2013.pdf)
申請書は初回のみ、問診票は毎回利用するときに必要
これベースのほうが仕様がわかりやすいかもしれません
username_0: 集める情報はかなりデリケートな内容になるので注意がいるでしょう。特にマイナンバーはかなりまずそう。
それはともかく、このユーザーと、実際に預ける子どもさんがまだ整理されていないように感じました。
タイトル通りユーザーは(アカウント共有するかはともかく)保護者と考えていて、子どもを登録するならまた別と考えていましたけど、子どもをユーザーとして登録する方がいいでしょうか。
username_1: 保護者が保育利用の主体なので、ユーザは保護者で子供登録は別でいいかなあと思いました。
username_0: 実運用ではいろいろとややこしいことも起こりそうですし、両者は分けて考えた方がよさそうですね。
今回の実装に反映するかはともかくとして。
username_3: 今回は親がユーザーで何人の子を預けたいというところまででいいのではないでしょうか?
子供の登録もあると嬉しいかもしれませんが複雑化するかもしれないので |
Azure/azure-sdk-for-net | 1170756475 | Title: Calling Server Outbound Calls and Direct Routing
Question:
username_0: ### Library name and version
Azure.Communication.CallingServer 1.0.0-beta.3
### Query/Question
Hello.
I am working on a telephony bot using the **Bot Framework**. The bot runs exclusively on **SBC** connectivity through **Direct Routing** (no phone numbers in the Communication Service Resource).
This works fine as the bot is triggered through inbound calls. But in one particular case, I need to make an outbound call (not a call transfer).
So I came across the **Azure.Communication.CallingServer** library and the **CreateCallConnectionAsync** method. I was wondering if this library supports **Direct Routing**.
Here is what I did:
```
var connectionString = “xxxxx";
var callingServerClient = new CallingServerClient(connectionString);
var client = new CommunicationIdentityClient(connectionString);
var userResponse = await client.CreateUserAsync();
var createCallOption = new CreateCallOptions(
new Uri("https://xxx.ngrok.io/api/callingserver?secret=xxxx"),
new[] { MediaType.Audio },
new[]
{
EventSubscriptionType.ParticipantsUpdated,
EventSubscriptionType.DtmfReceived
}
)
//{
// AlternateCallerId = new PhoneNumberIdentifier("+33xxx") // <-- using AlternateCallerId raises error: 403 Forbidden. Not allowed to impersonate.
//}
;
var callConnection = await callingServerClient.CreateCallConnectionAsync(
source: new CommunicationUserIdentifier(userResponse.Value.Id),
targets: new List<CommunicationIdentifier>() { new PhoneNumberIdentifier("+33xxx") },
options: createCallOption
);
```
As a result, I see a **"callConnectionState":"connecting"** in the first callback request.
Immediately followed by second request with **"callConnectionState":"disconnected"**
Any assistance would be greatly appreciated
Cheers.
### Environment
- Azure.Communication.CallingServer 1.0.0-beta.3
- .NET Core 3.1
- Windows 10 Enterprise
Answers:
username_1: Label prediction was below confidence level `0.6` for Model:`ServiceLabels`: 'Azure.Core:0.1684931,Azure.Identity:0.11274838,Communication:0.106337726'
username_2: @username_0 We are looking into it and get back to you.
username_3: Hello @username_0 '
Have you referred to this documentation: [Session Border Controllers and voice routing](https://docs.microsoft.com/en-us/azure/communication-services/concepts/telephony/direct-routing-provisioning)
[List of Session Border Controllers certified for Azure Communication Services direct routing](https://docs.microsoft.com/en-us/azure/communication-services/concepts/telephony/certified-session-border-controllers)
username_4: Hello @username_3,
Yes I have. As mentionned, my project (which is a Bot) already works fine using **Direct Routing**. That is for inbound calls and call transfers. What I'm struggling with is outbound calls. From what I understand, **Azure.Communication.CallingServer** could be the way to go.
Thanks.
username_0: Sorry, I wasn't signed in with the correct user account. |
inclusive-design/wecount.inclusivedesign.ca | 734095314 | Title: netlify-plugin-checklinks errors
Question:
username_0: ## Describe the bug
```
9:04:13 PM: ✖ FAIL load https://5f9f4ccbc726ab107271d1ae--wecount.netlify.app/404.html
9:04:13 PM: | operator: load
9:04:13 PM: | expected: 200 https://5f9f4ccbc726ab107271d1ae--wecount.netlify.app/404.html
9:04:13 PM: | actual: HTTP 404 Not Found
9:04:13 PM: | at: dist/404.html:14:938 <meta property="og:url" content="https://5f9f4ccbc726ab107271d1ae--wecount.netlify.app/404.html">
9:04:13 PM: ✖ FAIL load dist/learn/index.html
9:04:13 PM: | operator: load
9:04:13 PM: | expected: 200 dist/learn/index.html
9:04:13 PM: | actual: Invalid URL: http://AI%20for%20Accessibility%20grants
9:04:13 PM: | at: dist/learn
9:04:13 PM: ✖ FAIL load https://5f9f4ccbc726ab107271d1ae--wecount.netlify.app/
9:04:13 PM: | operator: load
9:04:13 PM: | expected: 200 https://5f9f4ccbc726ab107271d1ae--wecount.netlify.app/
9:04:13 PM: | actual: HTTP 404 Not Found
9:04:13 PM: | at: dist/index.html:14:1104 <meta property="og:url" content="https://5f9f4ccbc726ab107271d1ae--wecount.netlify.app/">
9:04:13 PM: ✖ FAIL load dist/initiatives/recbrkgunidv1ymec
9:04:13 PM: | operator: load
9:04:13 PM: | expected: 200 dist/initiatives/recbrkgunidv1ymec
9:04:13 PM: | actual: ENOENT: no such file or directory, open '/opt/build/repo/dist/initiatives/recbrkgunidv1ymec/'
9:04:13 PM: | at: dist/initiatives/index.html:14:22670 <a href="https://wecount.inclusivedesign.ca/initiatives/recbrkgunidv1ymec/">...</a>
9:04:13 PM: ✖ FAIL load https://5f9f4ccbc726ab107271d1ae--wecount.netlify.app/about/
9:04:13 PM: | operator: load
9:04:13 PM: | expected: 200 https://5f9f4ccbc726ab107271d1ae--wecount.netlify.app/about/
9:04:13 PM: | actual: HTTP 404 Not Found
9:04:13 PM: | at: dist/about/index.html:14:1216 <meta property="og:url" content="https://5f9f4ccbc726ab107271d1ae--wecount.netlify.app/about/">
9:04:13 PM: ✖ FAIL load https://5f9f4ccbc726ab107271d1ae--wecount.netlify.app/news/
9:04:13 PM: | operator: load
9:04:13 PM: | expected: 200 https://5f9f4ccbc726ab107271d1ae--wecount.netlify.app/news/
9:04:13 PM: | actual: HTTP 404 Not Found
9:04:13 PM: | at: dist/news/index.html:14:930 <meta property="og:url" content="https://5f9f4ccbc726ab107271d1ae--wecount.netlify.app/news/">
9:04:13 PM: ✖ FAIL load https://5f9f4ccbc726ab107271d1ae--wecount.netlify.app/inclusion-challenges/
9:04:13 PM: | operator: load
9:04:13 PM: | expected: 200 https://5f9f4ccbc726ab107271d1ae--wecount.netlify.app/inclusion-challenges/
9:04:13 PM: | actual: HTTP 404 Not Found
9:04:13 PM: | at: dist/inclusion-challenges/index.html:14:1198 <meta property="og:url" content="https://5f9f4ccbc726ab107271d1ae--wecount.netlify.app/inclusion-challenges/">
9:04:13 PM: ✖ FAIL load https://5f9f4ccbc726ab107271d1ae--wecount.netlify.app/badges/
9:04:13 PM: | operator: load
9:04:13 PM: | expected: 200 https://5f9f4ccbc726ab107271d1ae--wecount.netlify.app/badges/
9:04:13 PM: | actual: HTTP 404 Not Found
9:04:13 PM: | at: dist/badges/index.html:14:1200 <meta property="og:url" content="https://5f9f4ccbc726ab107271d1ae--wecount.netlify.app/badges/">
9:04:13 PM: ✖ FAIL load https://5f9f4ccbc726ab107271d1ae--wecount.netlify.app/initiatives/
9:04:13 PM: | operator: load
9:04:13 PM: | expected: 200 https://5f9f4ccbc726ab107271d1ae--wecount.netlify.app/initiatives/
9:04:13 PM: | actual: HTTP 404 Not Found
9:04:13 PM: | at: dist/initiatives/index.html:14:1142 <meta property="og:url" content="https://5f9f4ccbc726ab107271d1ae--wecount.netlify.app/initiatives/">
9:04:13 PM: ✖ FAIL load https://5f9f4ccbc726ab107271d1ae--wecount.netlify.app/views/
9:04:13 PM: | operator: load
9:04:13 PM: | expected: 200 https://5f9f4ccbc726ab107271d1ae--wecount.netlify.app/views/
9:04:13 PM: | actual: HTTP 404 Not Found
9:04:13 PM: | at: dist/views/index.html:14:932 <meta property="og:url" content="https://5f9f4ccbc726ab107271d1ae--wecount.netlify.app/views/">
9:04:13 PM: ✖ FAIL load https://5f9f4ccbc726ab107271d1ae--wecount.netlify.app/initiatives/recBRKGUNiDV1ymEC/
9:04:13 PM: | operator: load
9:04:13 PM: | expected: 200 https://5f9f4ccbc726ab107271d1ae--wecount.netlify.app/initiatives/recBRKGUNiDV1ymEC/
9:04:13 PM: | actual: HTTP 404 Not Found
9:04:13 PM: | at: dist/initiatives/recBRKGUNiDV1ymEC/index.html:14:992 <meta property="og:url" content="https://5f9f4ccbc726ab107271d1ae--wecount.netlify.app/initiatives/recBRKGUNiDV1ymEC/">
9:04:13 PM: ✖ FAIL load https://5f9f4ccbc726ab107271d1ae--wecount.netlify.app/initiatives/rechCtfzoi4gYQKvU/
9:04:13 PM: | operator: load
[Truncated]
9:04:13 PM: | actual: HTTP 404 Not Found
9:04:13 PM: | at: dist/views/the-importance-of-peripheral-vision/index.html:14:1158 <meta property="og:url" content="https://5f9f4ccbc726ab107271d1ae--wecount.netlify.app/views/the-importance-of-peripheral-vision/">
9:04:13 PM: ✖ FAIL load https://5f9f4ccbc726ab107271d1ae--wecount.netlify.app/tags/inclusion/
9:04:13 PM: | operator: load
9:04:13 PM: | expected: 200 https://5f9f4ccbc726ab107271d1ae--wecount.netlify.app/tags/inclusion/
9:04:13 PM: | actual: HTTP 404 Not Found
9:04:13 PM: | at: dist/tags/inclusion/index.html:14:972 <meta property="og:url" content="https://5f9f4ccbc726ab107271d1ae--wecount.netlify.app/tags/inclusion/">
9:04:13 PM: ✖ FAIL load https://5f9f4ccbc726ab107271d1ae--wecount.netlify.app/tags/smart-cities/
9:04:13 PM: | operator: load
9:04:13 PM: | expected: 200 https://5f9f4ccbc726ab107271d1ae--wecount.netlify.app/tags/smart-cities/
9:04:13 PM: | actual: HTTP 404 Not Found
9:04:13 PM: | at: dist/tags/smart-cities/index.html:14:978 <meta property="og:url" content="https://5f9f4ccbc726ab107271d1ae--wecount.netlify.app/tags/smart-cities/">
9:04:13 PM: 140 tests
9:04:13 PM: 117 passed
9:04:13 PM: 23 failed
```
## Screenshots
https://app.netlify.com/sites/wecount/deploys/5f9f4ccbc726ab107271d1ae
Answers:
username_1: These errors are reported when the deploy is triggered in the context of the production branch and deploy previews, but are not reported in the context of branch deploys.
An issue is filed against https://github.com/Munter/netlify-plugin-checklinks repository: https://github.com/Munter/netlify-plugin-checklinks/issues/278
username_2: I think I've figured out why this happens. In the production branch, those URLs reference pages which exist. However, in the deploy preview context, the URLs reference the deploy preview. But the deploy preview isn't published when the build plugin runs, so those pages don't exist yet… and then the build fails.
I'm thinking about some ways to work around this as I think it's a problem more with our test expectations than with the build plugin (I ran into it on the IDRC site too). |
jkjung-avt/tensorrt_demos | 753917728 | Title: TensorRT includes and libs
Question:
username_0: I compile the plugin code for yolo support. I'm compiling on Xavier NX running jetpack 4.4. The TensorRT includes and libraries are not in /usr/local. Do you know where they are located in jetpack 4.4?
Answers:
username_1: Check if you could find these files:
* /usr/include/aarch64-linux-gnu/NvInfer.h
* /usr/lib/aarch64-linux-gpu/libnvinfer.so
Status: Issue closed
username_0: Thank you! |
milvus-io/milvus | 522652111 | Title: [BUG] Search does not use GPU when index type is FLAT
Question:
username_0: **Describe the bug**
Search does not use GPU when index type is FLAT
**Steps/Code to reproduce behavior**
1. set "search_resources" in server_config.yaml to contains only gpu
search_resources:
- gpu0
2. create index with IndexType.FLAT
3. search performance seems as same as CPU search
**Expected behavior**
FLAT GPU performance should be much faster then FLAT CPU
**Environment details**
N/A
**Screenshots**
N/A
**Additional context**
Status: Issue closed
Answers:
username_1: done |
razorpay/razorpay-magento | 834603394 | Title: Again No Order ID after Capture Payment ( Web Book already enabled )
Question:
username_0: Guys,
What you doing with Update... No testing ... Today i missed 3 Order
pay_Gnqe1ypJK44MCa order_GnqdnneVWpIpV7 ₹ 1,199.00 ₹ - Indian Rupee (INR)
pay_GnmLCXx0Pg17YO order_Gnla4dsenjbGTI ₹ 4,615.00 ₹ - Indian Rupee (INR)
pay_GnlXCf3ziXxrW5 order_GnlWfI1KTGBQ0W ₹ 2,399.00 ₹ - Indian Rupee (INR)
--------------------------------------------------------------------------
Latest Log
[2021-03-18 04:36:42] main.INFO: Razorpay Webhook: Quote order is inactive for quoteID: 85484 and Razorpay payment_id(:pay_Gnk2NVCUXXgGy8) with Maze OrderID (:100017238) [] []
[2021-03-18 04:36:49] main.INFO: Razorpay Webhook processing started. [] []
[2021-03-18 04:36:51] main.INFO: Razorpay Webhook processing started. [] []
[2021-03-18 04:36:51] main.INFO: Razorpay Webhook: Quote order is inactive for quoteID: 84665 and Razorpay payment_id(:pay_GnlNNT60Z9Ymvd) with Maze OrderID (:100017239) [] []
[2021-03-18 04:36:56] main.INFO: Razorpay Webhook processing started. [] []
[2021-03-18 04:36:56] main.INFO: Razorpay Webhook: Order processing is active for quoteID: 64435 and Razorpay payment_id(:pay_GnlXCf3ziXxrW5) and webhook attempt: 1 [] []
[2021-03-18 04:36:59] main.INFO: Razorpay Webhook processing started. [] []
[2021-03-18 04:37:09] main.INFO: Razorpay Webhook processing started. [] []
[2021-03-18 04:37:09] main.INFO: Razorpay Webhook processing started. [] []
[2021-03-18 04:37:09] main.INFO: Razorpay Webhook: Order processing is active for quoteID: 85490 and Razorpay payment_id(:pay_GnmLCXx0Pg17YO) and webhook attempt: 1 [] []
[2021-03-18 04:37:15] main.INFO: Razorpay Webhook processing started. [] []
[2021-03-18 04:37:15] main.INFO: Razorpay Webhook: Quote order is inactive for quoteID: 85508 and Razorpay payment_id(:pay_GnpEQglgpndlxf) with Maze OrderID (:100017240) [] []
[2021-03-18 04:37:16] main.INFO: Razorpay Webhook processing started. [] []
[2021-03-18 04:37:27] main.INFO: Razorpay Webhook processing started. [] []
[2021-03-18 04:37:27] main.INFO: Razorpay Webhook: Order processing is active for quoteID: 85512 and Razorpay payment_id(:pay_Gnqe1ypJK44MCa) and webhook attempt: 1 [] []
Answers:
username_0: No reply from Razorpay Dashboard also...
Ticket ID #5212733 | Category: Transaction related
username_1: @username_0 , can you please share the log file to email.
username_0: @username_1 Sent
username_0: Please recover all order soon as possible. Webhook not working properly.
Status: Issue closed
username_2: @username_0 how did u recovered missed order
username_1: @username_2 , Please use latest version of plugin and enable webhook. It should resolve the issue.
username_0: July Month Order Not recovered. it's around 40k Missing order . Payment is capture but order id not generated in Magento admin.
username_0: @username_1 & @username_2
username_0: July Month Order Not recovered. it's around 40k Missing order . Payment is capture but order id not generated in Magento admin.
@username_1 & @username_2
username_1: @username_0 , Please make sure you are in latest version of the razorpay module and webhook enabled. If you still facing issue please let me know.
That much missing order means some issue in your site or webhook disabled.
username_0: It was Old order When old Extension was installed.
username_0: Why not create Pending Order If payment failed then Can i have change to recover order. Right now i have no choice of recover order if webhook failed some condition.
username_0: There are no bugs, Some time website/Mobile Hang because you are using same JS popup in Magento platform.. When device have low memory or something else...
username_2: @username_0 and @username_1 i have updated module using zip mode downloaded from - https://github.com/razorpay/razorpay-magento
from where should i download latest Razorpay/Razorpay folder
username_1: @username_2, please find the latest release https://github.com/razorpay/razorpay-magento/releases/tag/3.6.3 |
jlippold/tweakCompatible | 348911773 | Title: `mySB` working on iOS 11.3.1
Question:
username_0: ```
{
"packageId": "com.junesiphone.mysb",
"action": "working",
"userInfo": {
"arch32": false,
"packageId": "com.junesiphone.mysb",
"deviceId": "iPhone10,6",
"url": "http://cydia.saurik.com/package/com.junesiphone.mysb/",
"iOSVersion": "11.3.1",
"packageVersionIndexed": true,
"packageName": "mySB",
"category": "Addons (FrontPage)",
"repository": "junesiphone.com",
"name": "mySB",
"installed": "0.6",
"packageIndexed": true,
"packageStatusExplaination": "This package version has been marked as Working based on feedback from users in the community. The current positive rating is 100% with 2 working reports.",
"id": "com.junesiphone.mysb",
"commercial": false,
"packageInstalled": true,
"tweakCompatVersion": "0.1.0",
"shortDescription": "mySB for FrontPage",
"latest": "0.6",
"author": "JunesiPhone",
"packageStatus": "Working"
},
"base64": "<KEY>",
"chosenStatus": "working",
"notes": ""
}
``` |
onfido/horza | 105353806 | Title: Consistent collection queries
Question:
username_0: When calling any? on an empty Horza::Entities::Collection, the result returns false. However when the collection has elements in it the following error is raised
LocalJumpError: no block given (yield)
When calling blank? on a collection that contains elements, the result is correctly false. However when calling blank? on an empty collection the result is false also. |
SHoltzen/dice | 549282510 | Title: Undocumented dependency: m4.
Question:
username_0: When trying to install your tool i found a small problem, namely
that m4 is needed to build. Including it in the dependencies removes the error:
```
[ERROR] The compilation of conf-m4 failed at ...
```
Status: Issue closed
Answers:
username_1: Added to the list of dependencies, thanks! |
InterImmCenter/feed | 128188554 | Title: Blizzard Bears Down on U.S. East Coast NASA-NOAA's Suomi NPP satellite snapped this image of the blizzard approaching the U.S. East coast around 2:35 a.m. EST on Jan. 22 2016 using the Visible Infrared Imaging Radiometer Suite (VIIRS) instrument's Day-Night band. | http://ift.tt/1NpA4CQ | January 22 2016 | http://ift.tt/1Jp9yyG
Question:
username_0: <img src="http://ift.tt/1Jp9yyG"><br><b>Blizzard Bears Down on U.S. East Coast</b><br>
NASA-NOAA's Suomi NPP satellite snapped this image of the blizzard approaching the U.S. East coast around 2:35 a.m. EST on Jan. 22, 2016 using the Visible Infrared Imaging Radiometer Suite (VIIRS) instrument's Day-Night band.<br><br>
January 22, 2016<br>
via NASA http://ift.tt/1NpA4CQ |
silnrsi/palaso-python | 463743306 | Title: cmptxtrender does not show names of SMP characters
Question:
username_0: The output of `cmptxtrender` shows the input text of the string being tested. It does this by showing the glyph names of the input characters (as opposed to the USV of the characters) in a column called _Original_ when called from `smith test` or `smith xtest`
If an input character is a non-BMP character, then `.notdef` is shown. The glyph name should be shown instead. The correct glyph names are shown in the rendered output.
You can see examples of this in the `xtest` output for [Narnoor](https://github.com/silnrsi/font-narnoor) at https://build.palaso.org/viewType.html?buildTypeId=Fonts_Narnoor Look for `.html` files in the folder `tests/xtest` in a built artefact.
I suspect the string `.notdef` is coming from https://github.com/silnrsi/palaso-python/blob/master/scripts/font/cmptxtrender#L46 or maybe the next line. Since BMP characters work correctly, I wonder if there is a bug in binary/text strings related to how SMP characters are stored. This issue happens on both Python 2 and 3.
Status: Issue closed
Answers:
username_1: Thanks! |
aws/sagemaker-tensorflow-extensions | 594802927 | Title: TF 2.0 support?
Question:
username_0: Hi, will this extensions work with TensorFlow 2?
Answers:
username_1: Thanks for reaching out!
Absolutely, take a look at the versions that start with 2 =)
https://pypi.org/project/sagemaker-tensorflow/#history
Example: https://pypi.org/project/sagemaker-tensorflow/2.1.0.1.0.0/
Hope that helps!
Status: Issue closed
username_2: Hey there!
Based on this and #56, there still seems to be confusion around whether TF2 is supported - mainly because the README on master still says "TensorFlow versions 1.7-1.15.2". This was initially why we've been holding back on a TF 2.0 migration for one of our models at work 😂..
I see there's an updated readme (corresponding to the latest PyPI release descriptions) in the tf-2 branch.
I know there must be some reason the branches aren't merged, but maybe there could be some simple explanation until that time, like "[...] TensorFlow versions 1.7-1.15.2. TF 2 support is also available in the tf-2 branch and on PyPI (all the versions that start with 2: https://pypi.org/project/sagemaker-tensorflow/#history)."
What do you think? If I hadn't gone into the closed issues and seen these two threads, I would've again concluded that we're still waiting on TF 2 here 🤷 |
rdaly525/MetaMapper | 473079839 | Title: Task List
Question:
username_0: - [ ] Figure out a plan to pass semantic information from Halide to CoreIR (What is Data, what is Valid, what is reset)
- [ ] There is an issue with getting into random states after configuration. Need to replace control registers to reset-able registers.
- [ ] Write a plan for how to specify pipelining/retiming |
18F/its70-fs-epermit-scale-up | 298991535 | Title: Forest Service GitHub Enterprise Repositories
Question:
username_0: # Question/Comment on the Forest Service RFP
## Name and affiliation
<NAME>, TechTrend
## Section of RFP documents
Draft RFP Response Questions
## Question/Comment
My company, TechTrend, develops many applications in support of the Forest Service. All of our work in support of the Forest Service is done via the Forest Service's GitHub Enterprise (FS GHE), which is not publicly accessible. Can we include links to repositories that are located within the Forest Service's GitHub Enterprise? Since this is a Forest Service solicitation, I assume the evaluators can be given access to the FS GHE if they don't already have accounts. Same question for links to Github Profile's, which are also located in the FS GHE. Assuming the final RFP has the same requirement for links to repos and user handles, the same question will apply to that as well.
Answers:
username_1: We prefer open repositories as they demonstrate a vendor’s ability to work in the open. The technical quote response form will, however, include an opportunity to provide repository credentials. Please take a look at [Section 3.i.f of the Questions.](https://github.com/18F/its70-fs-epermit-scale-up/blob/master/solicitation-documents/technical-quote-questions.md)
Status: Issue closed
username_0: Great, thanks for the information.
username_1: I'm gonna keep this open in case others have a similar question. Thanks again for your interest!
username_1: # Question/Comment on the Forest Service RFP
## Name and affiliation
<NAME>, TechTrend
## Section of RFP documents
Draft RFP Response Questions
## Question/Comment
My company, TechTrend, develops many applications in support of the Forest Service. All of our work in support of the Forest Service is done on the Forest Service's GitHub Enterprise (FS GHE), which is not publicly accessible. Can we include links to repositories that are located within the Forest Service's GitHub Enterprise? Since this is a Forest Service solicitation, I assume the evaluators can be given access to the FS GHE if they don't already have accounts. Same question for links to Github Profile's, which are also located in the FS GHE. Assuming the final RFP has the same requirement for links to repos and user handles, the same question will apply to that as well.
username_2: Our company, Biswas Information Technology Solutions, Inc. (BITS) also has developed open source applications (Clinical Trial Management Systems) but it is on NIH's github site. Our EPA web API source code is also in EPA's repository. This makes it very hard to demonstrate our qualifications as these are not public facing projects. This prohibits us from answering the RFI questions, but hopefully, we can provide information on the projects when responding to the actual RFP.
username_1: Thanks for the comment. The RFP will afford an opportunity to provide credentials for private repos, in case that’s a possibility. Additionally, we are also open to viewing repos from non-government open-source projects as long as they fit the other [requirements](https://github.com/18F/its70-fs-epermit-scale-up/blob/master/solicitation-documents/Draft-RFP.md#51-written-quote-submission) stated in the RFQ. |
KingdomOfSaudiArabia/royal-orders | 223595103 | Title: صدور عدد من الأوامر الملكية إضافة أربعون
Question:
username_0: الرقم : أ / 198
التاريخ : 25 / 7 / 1438هـ
بعون الله تعالى
نحن سلمان بن عبدالعزيز آل سعود
ملك المملكة العربية السعودية
بعد الاطلاع على النظام الأساسي للحكم الصادر بالأمر الملكي رقم ( أ / 90 ) بتاريخ 27 / 8 / 1412هـ.
وبعد الاطلاع على نظام الوزراء ونواب الوزراء وموظفي المرتبة الممتازة الصادر بالمرسوم الملكي رقم ( م / 10 ) بتاريخ 18 / 3 / 1391هـ .
وبعد الاطلاع على الأمر الملكي رقم ( أ / 14 ) بتاريخ 3 / 3 / 1414هـ .
أمرنا بما هو آت:
أولاً : يعين الدكتور خليل بن مصلح الثقفي رئيساً عاماً للهيئة العامة للأرصاد وحماية البيئة بالمرتبة الممتازة .
ثانياً : يبلغ أمرنا هذا للجهات المختصة لاعتماده وتنفيذه .
سلمان بن عبدالعزيز آل سعود
// يتبع //
00:30ت م
www.spa.gov.sa/1619263 |
Medium/medium-sdk-nodejs | 110400673 | Title: Readme.md provides wrong information
Question:
username_0: According to Readme file, we need to run `npm install medium-sdk-nodejs` and then `var medium = require('medium-sdk-nodejs')`.
Both of them don't work and correct usage is `npm install medium-sdk` and then `var medium = require('medium-sdk')`.
Not giving a pull request because it'll take time to register.
Answers:
username_1: Thanks a lot username_0 it fixed the registry issue.
Status: Issue closed
|
crystal-lang/shards | 974166078 | Title: No warning for circular dependencies
Question:
username_0: I came across an issue today where we had a circular dependency
https://github.com/PlaceOS/driver/issues/105
however an earlier version of this shard didn't have the circular dependency so was being installed without warning
I couldn't work out why shards wasn't installing the latest version, I bumped my shards yml to force its hand and then got the error about circular dependencies.
It would be nice if shards would hint as to why it didn't select the latest version.
Like `Installing habitat (0.4.7) - There is a circular dependency between placeos-driver and placeos-core with 0.5.0`
unsure how difficult this would be |
zzzprojects/EntityFramework-Extensions | 404576037 | Title: System.Exception A default DbContext context must exist, or a context factory must be provided (EntityFrameworkManager.ContextFactory). This setting is required for some additional features. Read more: http://entityframework-extensions.net/context-factory
Question:
username_0: ` <PackageReference Include="Z.EntityFramework.Extensions.EFCore" Version="2.1.47" />`
```
services.AddDbContext<ApplicationDbContext>(options =>
options.UseSqlServer(Configuration.GetConnectionString("DefaultConnection")));
```
```
var auxs = new List<ConciliacionBancariaAux>();
foreach (var aux in conciliacionViewModel.conciliacionBancariaAux)
{
var conciliado= !string.IsNullOrEmpty(aux.Uuid);
//if(aux.IsNew)
{
auxs.Add(new ConciliacionBancariaAux
{
ProcesoBancoId = procesoBancoId,
Fecha = aux.Fecha,
Referencia = aux.Referencia,
TipoMovimientoId = aux.TipoMovimientoId,
Debito = aux.Debito,
Credito = aux.Credito,
EstadoId = aux.EstadoId,
Uuid = aux.Uuid,
Conciliado = conciliado,
TableInfo = aux.TableInfo,
IdOrigen = aux.IdOrigen,
IdRef = aux.IdRef,
Arrastrado = !aux.IsNew,
IsNew = aux.IsNew,
ProcesoBancoIdOld = aux.ProcesoBancoIdOld
});
}
}
db.BulkInsert(auxs);
```
Status: Issue closed
Answers:
username_1: @username_2 Hi, I came to this page when trying to solve the same problem. I've looked at the pages you show username_0 but I still can't figure out the solution.
All the examples shown on the page and on .NET Fiddle uses a constructor without parameter but my constructor takes DbContextOption:
`public class AppDbContext : DbContext
{
public DbSet<Employee> Employees { get; set; }
public AppDbContext(DbContextOptions<AppDbContext> options) : base(options) {}
}`
So this does not work for me as no parameter is passed to the constructor.
`return new AppDbContextChild();`
Is there anyway to solve this?
Thank you!
username_2: Hello @username_1 ,
Here is an example:
```csharp
EntityFrameworkManager.ContextFactory = context =>
{
var optionsBuilder = new DbContextOptionsBuilder<EntityContext>();
optionsBuilder.UseSqlServer(My.ConnectionString);
return new EntityContext(optionsBuilder.Options);
};
```
You simply do what you always do when creating a context.
Let me know if that answers your question.
Best Regards,
Jon |
barentsen/did-tabbys-star-fade | 128308758 | Title: Define the correct form of the upper limits likelihood
Question:
username_0: We need to include upper limits, however because we are including this in addition to the standard data it needs to be properly normalized.
The regular log-likelihood is
```python
def lnlike(theta, x, y, yerr):
m, b, lnf = theta
model = m * x + b
npt_lc = np.share(y)[0]
err_jit2 = yerr**2 + (np.e**lnf)**2
loglc = (
- (npt_lc/2.)*np.log(2.*np.pi)
- 0.5 * np.sum(np.log(err_jit2))
- 0.5 * np.sum((model - y)**2 / err_jit2)
)
return loglc
```
Answers:
username_1: Oh. I need to think about the normalization. I've simply been doing this:
```Python
def lnlike_limit(theta, x_limit, y_limit):
"""Non-detections."""
m, b = theta
sigma = 0.1 # uncertainty on the limit
model = m * x_limit + b
return norm.logcdf(model - y_limit, scale=sigma)
```
and then
```Python
def lnprob(theta, x, y, yerr, x_limit, y_limit):
"""Returns the log posterior."""
lp = lnprior(theta)
if not np.isfinite(lp):
return -np.inf
return lp + lnlike(theta, x, y, yerr) + lnlike_limit(theta, x_limit, y_limit)
```
username_0: There is a paper here that may have the answer:
Some Aspects of Measurement Error in Linear Regression of Astronomical Data
<NAME>
http://arxiv.org/pdf/0705.2774v1.pdf
username_1: If we use `scipy.norm.logpdf` and `scipy.norm.logcdf` then normalisation will be ok. And we may or may not be faster.
Thanks for the link. Will deffo check tomo. Now time for :zzz:
username_0: I don't think that paper is what I was after.
username_0: I don't think `scipy.norm.logpdf' does the right thing. It doesn't seem to take into account the number of data points.
username_0: I think it's a simple as
```python
loglc = -npt_lc * np.log(2) + np.sum(
np.log(1 + scipy.special.erf((model - y) / (yerr * 2**0.5))))
```
username_1: What is `npt_lc`?
username_0: number of observations
username_1: Can you say a few more words about how the function is derived? I'm a bit lost why the number of observations is in it.
username_1: Ah, nevermind, you brought `np.log(2)` outside of the sum and hence needed to multiply with the length of the sum.
Good news, the result of your expression is identical to `np.sum(norm.logcdf(model - y, scale=yerr))`:
```Python
In [40]: y = np.array([1, 2, 3])
In [41]: yerr = np.array([0.1, 0.1, 0.1])
In [42]: model = np.array([1.1, 2.1, 3.1])
In [43]: -3 * np.log(2) + np.sum(
np.log(1 + scipy.special.erf((model - y) / (yerr * 2**0.5))))
Out[43]: -0.51826133707034927
In [44]: np.sum(norm.logcdf(model - y, scale=yerr))
Out[44]: -0.51826133707034894
``` |
moo-man/FVTT-DD-Import | 996171072 | Title: Import fails with Loading failed for worlds/undefined/
Question:
username_0: Failed to load resource: the server responded with a status of 404 ()
foundry.js:35541 Foundry VTT | Loading failed for worlds/undefined/xxx.webp (100%): undefined
_onError @ foundry.js:35541
Foundry 0.8.9
FVTT-DD-Import 2.2.0
Answers:
username_0: user issue...
In the import dialog, just edit correctly the "upload path" field to an existing folder
Leaving this issue in case it helps someone.
Maybe as a feature it could default to the current world?
Status: Issue closed
username_1: It used to, but 0.8 changed where the world name is stored and the importer needs to use the new location
username_0: But it's confusing as it was working in 0.7 without any custom path ...
username_1: Right, as I said 0.8 changed where your world name is stored, used to be `game.world.name` (which the importer uses) and now it's `game.world.data.name` hence the undefined |
98oktay/oktaybaskus | 637068370 | Title: Improvement changing votes
Question:
username_0: Hi, an improvement for the site would be to show when a member changes their vote to show that he changed and what it was and what it has become.
I think it is safe to assume when not everyone is changing there vote, that it is still the same voting round. you could even put a button on the site [New vote] or [Clear votes]
Thanks
richard |
kartik-v/yii2-editable | 577076812 | Title: Wrong Editable is selected when choosing an editable to edit
Question:
username_0: ## Prerequisites
- [X] I have searched for similar issues in both open and closed tickets and cannot find a duplicate.
- [X] The issue still exists against the latest `master` branch of yii2-editable.
- [X] This is not an usage question. I confirm having gone through and read the [documentation and demos](http://demos.krajee.com/editable).
- [X] This is not a general programming / coding question. (Those should be directed to the [webtips Q & A forum](http://webtips.krajee.com/questions)).
- [X] I have attempted to find the simplest possible steps to reproduce the issue.
- [X] I have included a failing test as a pull request (Optional).
## Steps to reproduce the issue
1. Create two (or more) editables, setting their model attribute to two different instances of the same model class.
2. Set the attributes to the same. This way we have multiple models with the same attributes being edited.
3. Click any editable other than the first one.
## Expected behavior and actual behavior
Expected: You begin editing the editable you clicked
When I follow those steps, I see...
The dialog appears on the FIRST Editable following the same configuration on the page, and any edits affect that one.
I was expecting...
To edit the one I clicked.
## Environment
#### Browsers
- [ X ] Google Chrome
- [ X ] Mozilla Firefox
- [ ] Internet Explorer
- [ ] Safari
#### Operating System
- [ ] Windows
- [ ] Mac OS X
- [ X ] Linux
- [ ] Mobile
#### Libraries
- jQuery version: 3.4.1
- yii2-editable version: @dev
## Isolating the problem
- [ ] This bug happens [on the demos page](https://demos.krajee.com/editable)
- [ X ] The bug happens consistently across all tested browsers
- [ X ] This bug happens when using yii2-editable without other plugins.
Status: Issue closed
Answers:
username_0: Using the ['options' => ['id' => 'UNIQUE_ID']] does fix this. I suppose this might not even be a bug. |
clearlinux/distribution | 425858476 | Title: Is Clear Linux a general purpose Operating System?
Question:
username_0: Is Clear Linux a general purpose Operating System?
If so, do I download the `installer.img.xz` file?
Answers:
username_1: You need to uncompress the img file, then something like Etcher can write the IMG to a USB. If you need an ISO, uncompress the iso.xz file. Read the instructions on clearlinux.org, or the Documentation. It covers how to approach a lot of use cases.
Based only on my experience, I'd answer "no" to the general purpose question. I believe Clear categorizes itself as an "infrastructure OS", but that doesn't have to constrain anyone's own usage, and a skilled Linux user can mold it to their needs. See clr-on-clr tools.
username_2: we certainly aim to be an instructure OS as well as a developer / desktop OS.
on the later we are making improvements but we know we're not ready for general use.
username_0: Thanks for your help.
username_3: Please open a new issue if you have additional questions! thanks.
Status: Issue closed
username_0: No worries. |
toanpv/vjreport | 374998579 | Title: Vở bài tập
Question:
username_0: Description:
---
Các môn
Device info:
---
<table>
<tr><td>App version</td><td>1.1.7</td></tr>
<tr><td>App version code</td><td>18092701</td></tr>
<tr><td>Android build version</td><td>1533411376</td></tr>
<tr><td>Android release version</td><td>6.0</td></tr>
<tr><td>Android SDK version</td><td>23</td></tr>
<tr><td>Android build ID</td><td>CPH1609EX_11_A.24_180805</td></tr>
<tr><td>Device brand</td><td>OPPO</td></tr>
<tr><td>Device manufacturer</td><td>OPPO</td></tr>
<tr><td>Device name</td><td>CPH1609</td></tr>
<tr><td>Device model</td><td>CPH1609</td></tr>
<tr><td>Device product name</td><td>CPH1609</td></tr>
<tr><td>Device hardware name</td><td>mt6755</td></tr>
<tr><td>ABIs</td><td>[arm64-v8a, armeabi-v7a, armeabi]</td></tr>
<tr><td>ABIs (32bit)</td><td>[armeabi-v7a, armeabi]</td></tr>
<tr><td>ABIs (64bit)</td><td>[arm64-v8a]</td></tr>
</table> |
netbox-community/netbox-docker | 667873195 | Title: netbox_1 do not connect error
Question:
username_0: Hi everybody,
I did the steps to put netbox on by docker-compose on CentOS 7 and I got this error on docker-compose logs:
netbox_1 | self.fetch_command(subcommand).run_from_argv(self.argv)
netbox_1 | File "/usr/local/lib/python3.7/site-packages/django/core/management/base.py", line 328, in run_from_argv
netbox_1 | self.execute(*args, **cmd_options)
netbox_1 | File "/usr/local/lib/python3.7/site-packages/django/core/management/base.py", line 369, in execute
netbox_1 | output = self.handle(*args, **options)
netbox_1 | File "/usr/local/lib/python3.7/site-packages/django/core/management/base.py", line 83, in wrapped
netbox_1 | res = handle_func(*args, **kwargs)
netbox_1 | File "/usr/local/lib/python3.7/site-packages/django/core/management/commands/migrate.py", line 86, in handle
netbox_1 | executor = MigrationExecutor(connection, self.migration_progress_callback)
netbox_1 | File "/usr/local/lib/python3.7/site-packages/django/db/migrations/executor.py", line 18, in __init__
netbox_1 | self.loader = MigrationLoader(self.connection)
netbox_1 | File "/usr/local/lib/python3.7/site-packages/django/db/migrations/loader.py", line 49, in __init__
netbox_1 | self.build_graph()
netbox_1 | File "/usr/local/lib/python3.7/site-packages/django/db/migrations/loader.py", line 212, in build_graph
netbox_1 | self.applied_migrations = recorder.applied_migrations()
netbox_1 | File "/usr/local/lib/python3.7/site-packages/django/db/migrations/recorder.py", line 76, in applied_migrations
netbox_1 | if self.has_table():
netbox_1 | File "/usr/local/lib/python3.7/site-packages/django/db/migrations/recorder.py", line 56, in has_table
netbox_1 | return self.Migration._meta.db_table in self.connection.introspection.table_names(self.connection.cursor())
netbox_1 | File "/usr/local/lib/python3.7/site-packages/django/utils/asyncio.py", line 26, in inner
netbox_1 | return func(*args, **kwargs)
netbox_1 | File "/usr/local/lib/python3.7/site-packages/django/db/backends/base/base.py", line 260, in cursor
netbox_1 | return self._cursor()
netbox_1 | File "/usr/local/lib/python3.7/site-packages/django/db/backends/base/base.py", line 236, in _cursor
netbox_1 | self.ensure_connection()
netbox_1 | File "/usr/local/lib/python3.7/site-packages/django/utils/asyncio.py", line 26, in inner
netbox_1 | return func(*args, **kwargs)
netbox_1 | File "/usr/local/lib/python3.7/site-packages/django/db/backends/base/base.py", line 220, in ensure_connection
netbox_1 | self.connect()
netbox_1 | File "/usr/local/lib/python3.7/site-packages/django/db/utils.py", line 90, in __exit__
netbox_1 | raise dj_exc_value.with_traceback(traceback) from exc_value
netbox_1 | File "/usr/local/lib/python3.7/site-packages/django/db/backends/base/base.py", line 220, in ensure_connection
netbox_1 | self.connect()
netbox_1 | File "/usr/local/lib/python3.7/site-packages/django/utils/asyncio.py", line 26, in inner
netbox_1 | return func(*args, **kwargs)
netbox_1 | File "/usr/local/lib/python3.7/site-packages/django/db/backends/base/base.py", line 197, in connect
netbox_1 | self.connection = self.get_new_connection(conn_params)
netbox_1 | File "/usr/local/lib/python3.7/site-packages/django/utils/asyncio.py", line 26, in inner
netbox_1 | return func(*args, **kwargs)
netbox_1 | File "/usr/local/lib/python3.7/site-packages/django/db/backends/postgresql/base.py", line 185, in get_new_connection
netbox_1 | connection = Database.connect(**conn_params)
netbox_1 | File "/usr/local/lib/python3.7/site-packages/psycopg2/__init__.py", line 127, in connect
netbox_1 | conn = _connect(dsn, connection_factory=connection_factory, **kwasync)
netbox_1 | django.db.utils.OperationalError: could not connect to server: Host is unreachable
netbox_1 | Is the server running on host "postgres" (172.23.0.3) and accepting
netbox_1 | TCP/IP connections on port 5432?
netbox_1 |
netbox_1 | ⏳ Waiting on DB... (3s / 30s)
Can someone help me ?
Status: Issue closed
Answers:
username_1: Please use our Slack channel to ask questions or fill in a proper bug report, i.e. use the provided template. |
JoiaApp/email | 258249795 | Title: Forget Password flow & email
Question:
username_0: @dansftown @joshbodily
### App Flow
Follow this prototype for app flow: https://marvelapp.com/6cfd43b/screen/32488632
1. User taps "Forgot Password"
2. Enters email address
3. User receives email with a unique code (not too long, maybe 5–6 random characters)
4. User returns to app, drops in temporary password and logs in
5. User is immediately prompted to set new password by entering it twice.
### Email
The email template has been added to Mandrill with a merge tag for the user's first name. The only other dynamic content is the random temporary password code.
 |
jsvine/pdfplumber | 1185053311 | Title: Add documentation re. common table-extraction challenges
Question:
username_0: Perhaps a new section in the README.md file, at least for now, identifying common issues that people run into, especially with table extraction. For example:
- Invisible lines causing problems
- #311 (and h/t to @jfuruness in that thread for the inspiration to create such documentation)
- Extracting tables with missing bottom borders
- https://github.com/username_0/pdfplumber/discussions/631
- https://github.com/username_0/pdfplumber/discussions/625
- Extracting tables with missing side borders
- https://github.com/username_0/pdfplumber/discussions/625
Other suggested topics? Other examples to add to the topics above? |
phetsims/rosetta | 136482545 | Title: remember the most recent locale for a translator and auto-select it
Question:
username_0: This was requested by a user. Seem like a good idea, but low priority for now. @oliver-phet will let the user know.
Answers:
username_0: This issue has been on my plate for over four years but as a low priority. Basically, low priority means I'll never get to it, because there is never a time where I don't have a lot of high priority things to address.
@muedli has been working on Rosetta recently, so I'm going to assign this to him and he and I will discuss an implementation plan and see if we can make it happen that way. |
jelovirt/org.lwdita | 285608944 | Title: DITA + Markdown publishing fail - "No renderer configured for com.vladsch.flexmark.ext.typographic.TypographicSmarts"
Question:
username_0: See issue here for more details and samples:
https://github.com/dita-ot/dita-ot/issues/2866
Answers:
username_1: Smart quote processing should be disabled
Status: Issue closed
username_1: See issue here for more details and samples:
https://github.com/dita-ot/dita-ot/issues/2866
Status: Issue closed
|
polonel/trudesk | 371914494 | Title: Duplicate key mongo error on installations step.
Question:
username_0: **Is this a BUG REPORT or FEATURE REQUEST?**:
<!-- PUT AN x IN THE BOX -->
- [x] BUG
- [ ] FEATURE
**What happened**:
{"success":false,"error":{"name":"MongoError","message":"E11000 duplicate key error collection: trudesk.counters index: _id_ dup key: { : \"tickets\" }","driver":true,"index":0,"code":11000,"errmsg":"E11000 duplicate key error collection: trudesk.counters index: _id_ dup key: { : \"tickets\" }"}}
**What did you expect to happen**:
It should redirect to home/dashboard page
**How to reproduce it (as minimally and precisely as possible)**:
On installation Step:
1. Click on `Get Start`
2. Give configurations of mongo db server
3. Fill form for `Admin User Create`
4. Error occured
**Anything else we need to know?**:
**Environment**:
- Trudesk Version:
- OS (e.g. from /etc/os-release):
- Node.JS Version:
- MongoDB Version:
- Is this hosted on cloud.trudesk.io:
Answers:
username_1: This occurs because the install was interrupted during the population of
the mongo database. The error states that it already has a duplicate entry
which means you had previously ran the install wizard and it did not
complete.
To resolve you’ll need to empty the mongo database and start fresh.
username_0: @username_1 Thanks for your reply.
It is resolved now.
Actually, the issue was due to I did not restart server after installation complete.
Status: Issue closed
|
luoto/chingu-frontend | 353908003 | Title: Quirky design-for-delight touch
Question:
username_0: This could be a fun way to pour a little of that gourmet chingu playfulness into the client user-journey and potentially add a little flavor to the video walkthrough. Let me know what you think ;p
**If teamHealth = GREEN, then** [https://giphy.com/gifs/1hMkbiQUNL3WSsgcSw](https://giphy.com/gifs/1hMkbiQUNL3WSsgcSw)
**If teamHealth = YELLOW, then** [http://g.recordit.co/oKyBqr2mXY.gif](http://g.recordit.co/oKyBqr2mXY.gif)
Answers:
username_0:  |
linked-art/linked.art | 234942247 | Title: Consider .1 properties for scope and model
Question:
username_0: Thanks to @username_1, in #51 ...
Property | Property subtype property
------------|----------
`P3 has note` | `P3.1 has type`
`P14 carried out by` | `P14.1 in the role of`
`P16 used specific object`|`P16.1 mode of use`
`P19 was intended use of`|`P19.1 mode of use`
`P62 depicts`|`P62.1 mode of depiction`
`P67 refers to`|`P67.1 has type`
`P69 has association with`|`P69.1 has type`
`P102 has title`|`P102.1 has type`
`P107 has current or former member`|`P107.1 kind of member`
`P130 shows features of`|`P130.1 kind of similarity`
`P136 was based on`|`P136.1 in the taxonomic role`
`P137 exemplifies`|`P137.1 in the taxonomic role`
`P138 represents`|`P138.1 mode of representation`
`P139 has alternative form`|`P139.1 has type`
`P144 joined with`|`P144.1 kind of member`
- [X] `P3.1` We don't use P3
- [X] `P14.1` Covered by #51
- [X] `P16.1` Should also be covered by the same technique as #51 -- just make a sub-activity with a P2 that is about the mode of use of the object
- [X] `P19.1` Ditto, but not sure that we have any use cases for P19 anyway
- [ ] `P62.1` No use cases?
- [ ] `P67.1` Probably a P2 on the referring text
- [X] `P69.1` E29 (the domain and range) isn't used
- [X] `P102.1` Title isn't used(!); just Appellation
- [ ] `P107.1` No use cases?
- [ ] `P130.1` No use cases?
- [X] `P136.1` E83, P136 not used
- [X] `P137.1` Ditto
- [X] `P138.1` See P62.1
- [ ] `P139.1` Use cases?
- [ ] `P144.1` Use cases?
Answers:
username_1: What's the story with `P3 has note`? In my experiment with the Collections API of Museum Victoria, I used it a couple of times. From that and other data sets I've seen it seemed to me that relatively informal "notes" would be a fairly common data type in museum CMS. Do such things end up elsewhere in the Linked.Art profile? Or just excluded?
username_0: The big issue is that it changes semantics depending on which class it's used with, which might be okay in the serialization-free CRM, but in RDF is a no-no. For example, it's a generic note on most classes, but a transcription for an `Inscription`. So if I want to have a note associated with an `Inscription`, I can't programmatically distinguish between transcription of the original and note about the resource.
So ... we use `rdf:value` for all "values" of things, and a `referred_to_by` `LinguisticObject` for notes / statements. That way 3.1 is the classified_as of the LinguisticObject, and the value is just another use of `rdf:value`.
username_0: So for your cases, I think we'd do a Linguistic object with an aat classification for each of object description versus physical description, per http://linked.art/model/base/#3-statements-about-a-resource
username_1: I hadn't known about the recommendation to use `P3 has note` for transcriptions. I agree that is awful. I think the CIDOC should purge the model of these "metaproperties" and align their metamodel with RDF. The extra complexity of the metamodel is not worth it, and pragmatically, if it's not expressible in RDF, it's a major barrier to adoption.
username_0: `P3_has_note` is a kludge of the worst sort 😿 How can it be both not expressed in terms of CRM and then used for specific, different cases under E33, E34 and E90 ... none of which are actually "notes" but instead the value of the resource. One despairs!
username_2: P144.1 has come up in #62, as well as P107.1. I lean toward thinking that this is outside of the scope of what we're doing here (we're also not modeling family relationships, or other intracacies of human relationships), but perhaps we should be explicit about that?
username_0: Yes, agreed. It's an art project (heh), not a person project. Put it in model/actor ?
username_2: I'd be fine with that. something like: "The nuances of person-to-person relationships are outside the scope of this project, beyond simple membership in groups. This includes the details of family relationships, group roles, and other similar reifications of the nature of relationships"
username_2: Probably worth checking out this (https://www.w3.org/TR/vocab-org/) when we decide to revisit.
username_3: Hello everyone,
As p67.1 is not yet marked as solved, I would have a use case. In our data E38 or an E22 use p67 to refer to an actor E39. In all cases I know in what kind of type/role they refer to this actor, eg. the actor could be the architect of the depicted/discussed work. Would you in this case use p67.1 or is there a better way to express this information?
And forgive me for asking this question here, I was searching your git for p67 and decided to post in this issue, instead of the slack channel.
Many thanks,
Leo
username_0: I think the intended use is to clarify the "about-ness" of the intellectual content to the resource that is being referenced by it. For example, a text that is a citation of another text is referring to the second text by citing it, a text might refer to a Dimension by providing evidence for it, or an image might refer to a literary concept by being an allegory for it. The content refers to the resource in some particular way. If you knew that all already, apologies for mansplaining / mecspliquer.
I would express 'architect of' by adding a P2 has type / `classified_as` property to a Production activity that produced the MMO. Which avoids the problem but doesn't solve it.
The referring E89 could be partitioned, with the part having a P2 that identifies the type of content and has the P67 relationship. For example:
```
{
"type": "LinguisticObject",
"includes": [
{
"type": "LinguisticObject",
"classified_as": [ "aat:citation" ],
"refers_to": ["some-target-uri-goes-here"]
}
]
}
```
Saying ... there is a part of this content that is a citation that refers to the target resource.
Thoughts?
username_3: Dear @username_0
This helps a lot. In putting the lingustic object in-between I get more flexibility. Is there an issue in using too many linguistic objects?
Many thanks and kind regards,
Leo
username_1: It turns out that the CRM working group has in fact produced a better RDF mapping for these `.1` properties to RDF than the one which they have been pushing; esssentially reifying the properties-which-have-properties as ordinary resources, which can therefore have their own properties and be associated with `E55 Types` and so on. They have called it the "Property Class" technique.
http://www.cidoc-crm.org/Issue/ID-266-reified-association-vs-sub-event is an archive of a fairly relevant discussion on the CRM mail-list fro 2014.
http://www.cidoc-crm.org/sites/default/files/Roles.pdf presents a few slides showing options for modelling properties of properties, including the "Property Class" approach.
The RDFS expression of "Property Classes" is here: http://www.cidoc-crm.org/sites/default/files/CRMpc_v1.1_0.rdfs
These couple of slides include a nice illustration of the "Property Class" approach defined in the RDFS: http://www.cidoc-crm.org/sites/default/files/20160802PropertiesOfProperties.pptx
username_0: P62.1 is okay for two reasons:
* we don't use P62 in favor of the long-cut
* the VI in the long-cut ( P65 / P138) can have a classification of the mode
And similarly P65.1, P138.1 ... these are types on the domain/range.
P67.1 is the root cause for issue #159. Given that this is the only time it has come up, and we have two possible solutions, it seems done?
P130.1 almost certainly lacks data that would be used to implement it. Should seek use cases in a time box and if none, then mark okay?
P139.1 - the type of alternate form is surely the type of the ... alternate form Appellation?
username_0: I propose to close this issue for two reasons:
1. We approach the problem from a use-case-first pattern, and then look for a solution, rather than attempting to align all of the properties in CRM. Thus if we need a .1 property, then we will figure out how to do it.
2. We discussed it ad nauseum in #107
3. There's a potential for a solution in the pattern described in #229
username_4: WG discussion 28-1-2020: we agree it is better to solve the individual situations on a case-by-case basis, when a proper use case come for them
Status: Issue closed
username_0: Closing, per WG discussion. |
liuyueyi/hexblog | 588880702 | Title: 200326-MongoDb系列教程六:文档 Document 插入姿势 - 一灰灰Blog
Question:
username_0: https://blog.hhui.top/hexblog/2020/03/26/200326-MongoDb%E7%B3%BB%E5%88%97%E6%95%99%E7%A8%8B%E5%85%AD-%E6%96%87%E6%A1%A3-Document-%E6%8F%92%E5%85%A5%E5%A7%BF%E5%8A%BF/
文档相当于关系数据库中数据行,也是我们最关心的数据本身;以BSON格式存储(和json区别不大) 我们通常所说业务开发者的CURD四大技能,在mongodb中,就是针对Document而言,接下来我们先看一下文档的新增使用姿势 |
csrgxtu/TemplateEngine | 253315586 | Title: url route
Question:
username_0: 
Answers:
username_0:  |
GC-spigot/Tokens | 776424748 | Title: [Issue] Receivers of /tokens add/pay get no messages
Question:
username_0: **Describe the bug**
_A clear and concise description of what the bug is._
When the console executes /tokens add or pay, the receiver does not get a message showing that they have gotten the tokens.
I've tried resetting the messages.yml file after using the latest tokens version on 1.16.4.
**To Reproduce**
A way to reproduce the video
N/A
**Screenshots / Video**
If applicable, add screenshots to help explain your problem.
N/A
**Server Information (please complete the following information):**
- Minecraft Version: Tuinity 1.16.4 (latest)
- Plugin Version: 2.9.2
Answers:
username_1: Try updating, doesn't happen in testing
Status: Issue closed
|
EDSM-NET/FrontEnd | 874847843 | Title: All types of globe molluscs (codex_ent_small_org_moll01_v5_*) seem missing in dump file codex.json.gz
Question:
username_0: For a statistical project, I make heavy use of your databases.
I seems that in EDSM's dump file "codex.json.gz" all types of globe molluscs are missing, even though some definately has been found.
The globe molluscs should be found under the "type" **codex_ent_small_org_moll01_v5_\*** or the "name" **\* Globe Mollusc**.
Grepping each of this strings on the codex.json leads to no results.
I wonder if there are no such data in your db or if there may be a bug in the dump routine?
Answers:
username_0: Additional comment:
I wondered if those globe molluscs were present in EDSM's Codex DB.
According to Canonn's database, globe molluscs are only found in the Region 9, the "Inner Scutum-Centaurus Arm".
There are about 270 entries, created between 2020-12-22 and 2021-03-17 in the Area "Eol Flyou".
On EDSM's Codex page [https://www.edsm.net/de/codex/discoveries/idR/9/region/Inner+Scutum-Centaurus+Arm/index/idC/20000/nameC/Biological+and+Geological/idS/22000/nameS/Organic+Structures](https://www.edsm.net/de/codex/discoveries/idR/9/region/Inner+Scutum-Centaurus+Arm/index/idC/20000/nameC/Biological+and+Geological/idS/22000/nameS/Organic+Structures) are no globe molluscs at all.
Maybe there are just no entries of globe molluscs in the EDSM database?
username_1: If they were send to the API, they'll be added soon.
While I'm processing the 'ew one from Odyssey, i will most likely integrate them.
Also transferring to the Alias repo as it where they will get added. |
haystack/eyebrowse-server | 177059670 | Title: status 500 twitter connection
Question:
username_0: I got a status 500 when trying to connect my twitter account
http://eyebrowse.csail.mit.edu/accounts/profile/sync_twitter?oauth_token=ep<KEY>&oauth_verifier=<KEY>
It's probably in the function sync_twitter in views.py
Answers:
username_1: traceback:
```
Traceback (most recent call last):
File "python2.7/site-packages/django/core/handlers/base.py", line 111, in get_response
response = callback(request, *callback_args, **callback_kwargs)
File "python2.7/site-packages/django/contrib/auth/decorators.py", line 20, in _wrapped_view
return view_func(request, *args, **kwargs)
File "python2.7/site-packages/annoying/decorators.py", line 74, in wrapper
output = function(request, *args, **kwargs)
File "/eyebrowse-server/accounts/views.py", line 293, in sync_twitter
access_token=token, access_token_secret=secret)
File "python2.7/site-packages/django/db/models/manager.py", line 137, in create
return self.get_query_set().create(**kwargs)
File "python2.7/site-packages/django/db/models/query.py", line 377, in create
obj.save(force_insert=True, using=self.db)
File "python2.7/site-packages/django/db/models/base.py", line 463, in save
self.save_base(using=using, force_insert=force_insert, force_update=force_update)
File "python2.7/site-packages/django/db/models/base.py", line 551, in save_base
result = manager._insert([self], fields=fields, return_id=update_pk, using=using, raw=raw)
File "python2.7/site-packages/django/db/models/manager.py", line 203, in _insert
return insert_query(self.model, objs, fields, **kwargs)
File "python2.7/site-packages/django/db/models/query.py", line 1593, in insert_query
return query.get_compiler(using=using).execute_sql(return_id)
File "python2.7/site-packages/django/db/models/sql/compiler.py", line 912, in execute_sql
cursor.execute(sql, params)
File "python2.7/site-packages/django/db/backends/postgresql_psycopg2/base.py", line 52, in execute
return self.cursor.execute(query, args)
DatabaseError: integer out of range
``` |
oppia/oppia | 536750921 | Title: RUN_E2E_TESTS_EXPLORATION_FEEDBACK_TAB flaky e2e test
Question:
username_0: There is a flaky e2e test:
`ExplorationFeedback should add feedback to an exploration`
Full log is [log.txt](https://github.com/oppia/oppia/files/3953817/log.txt).
Answers:
username_1: @username_0 does this still happen? If not, let's close this.
Status: Issue closed
username_0: Nope, don't see this any more. |
cypress-io/cypress | 326594252 | Title: Tests run in Chrome with baseUrl set navigate directly to baseUrl instead of Test Runner URL
Question:
username_0: ### Current behavior:
When I set a `baseUrl` and attempt to run the tests in Chrome, instead of the test runner it attempts to directly go to that URL.

Removing the `baseUrl` property will load Cypress as normal, as does having a value set but running the tests using Electron.
### Desired behavior:
When I run the tests with Chrome, it should navigate to Cypress running on localhost and show me the Cypress Test Runner. This is how it behaves when using Electron as the browser.
### Steps to reproduce:
1. Add `"baseUrl": "https://doesntmatteranyurl"` to `Cypress.json`
2. Open Cypress
3. Set browser to Chrome
4. Run any tests
### Versions
**Cypress:** 2.1.0 (happened on previous versions too)
**OS:** Windows 10 1803
Answers:
username_1: This is the intended behavior of `baseUrl` as outlined [here](https://docs.cypress.io/guides/references/best-practices.html#Setting-a-global-baseUrl)
How are you seeing the main url in the Electron browser? - the main url bar does not display for me in Electron.
Are you having an issue running the tests in Chrome versus Electron? Are the tests not running / displaying an error?
username_1: I'm unable to recreate this in my environment using Cypress 2.1.0 on Mac. Tried Chrome 66, Chromium 65, Canary 68 and Electron 59. The only other obvious difference I see is Windows, which I would have to test.
Do you have any other configuration set up? Anything in your `cypress.json` or env vars?
username_0: It could be, my colleages also had it but they were all running Windows 7 or 10 too.
Here's the full `cypress.json`, no environment variables set.
```json
{
"baseUrl": "https://e2e-webapp-dev-brighthr.azurewebsites.net",
"viewportWidth": 1440,
"viewportHeight": 900,
"env": {
"apiUrl": "https://e2e-brighthr-api-dev.azurewebsites.net",
"region": "uk"
},
"reporter": "mochawesome",
"reporterOptions": {
"inline": true,
"reportDir": "cypress/report",
"reportFilename": "e2e.html",
"configFile": "cypress.json"
},
"reporterEnabled": "mochawesome,@cypress/mocha-teamcity-reporter",
"mochawesomeReporterOptions": {
"inline": true,
"json": false,
"quiet": true,
"reportDir": "test-results/e2e",
"reportFilename": "index.html"
}
}
```
username_0: @username_1: Just asked a colleague on a Mac to run it and it's fine for them, so it appears to be a Windows issue. Here's what I see:

And here's what my colleague sees (note the Test Runner UI in the background):

username_0: Still an issue on 3.0.1.
username_2: I'm experiencing the same issue on **Fedora 28**:
With **Chrome 68**: the `baseUrl` is opened instead of the test runner and I get a 404 error.
When I use **Chromium 67** or **Electron 59** everything is working as expected.
username_2: Another hint: the error occurs in Cypress version `2.1.0` (where it already worked before) and in version `3.0.3`.
This seems to indicate that something changed in a newer Chrome version which causes the problem.
username_1: Oftentimes Windows has a different major Chrome release version than Linux/Mac machines, so while it is only happening on Windows - it may be a specific Chrome version issue - would need full versions of Chrome it is happening on.
I'm still unable to replicate on:
- Chrome Version 70.0.3518.0
- Chrome Version 68.0.3440.106
- Chrome Version 67.0.3396.0
These just happen to be the chrome versions on my machine at the moment though.
username_3: I have the same issue...
Chrome 68, Cypress 3.1.0, Mac OS.
username_0: Still an issue with Cypress 3.1.4 and Chrome 71.0.3578.98 (Official Build) (64-bit) on Windows 10 1803.
username_4: I have this issue as well on Windows 7, Cypress 3.1.4 and 71.0.3578.98 (Official Build) (32-bit).
username_1: Can any of you experiencing this issue, try to add the code below and respond whether it fixes the issue or not?
Add the code below to your `cypress/plugins/index.js` file:
```js
module.exports = (on) => {
on('before:browser:launch', (browser = {}, args) => {
if (browser.name === 'chrome') {
// ^ make sure this is your browser name, you may
// be using 'canary' or 'chromium' for example, so change it to match!
args.push('--proxy-bypass-list=<-loopback>')
return args
}
})
}
```
username_5: Hi @username_1, I confirm that adding these lines fixes the issue :D
username_0: @username_1 yes, that seems to fix it for me.
username_1: This PR will automatically pass these flags through to Chrome and should resolve this issue when released: https://github.com/cypress-io/cypress/pull/3049
Status: Issue closed
username_0: @username_1 interestingly, this doesn't work for every machine I have. It worked on my home machine yesterday, but when I try it on my work machine I'm still getting this. I've tried with and without your code snippet and tried 3.1.5 too, but no luck. I've also upgraded to Chrome 72.0.3626.81 (Official Build) (64-bit), reinstalled it from scratch and completely wiped Chrome's settings, but no change.
username_1: Hey @username_0, can you try setting `"modifyObstructiveCode": false` in your `cypress.json`? Does that fix the issue?
username_0: @username_1 just tried it, no difference.
username_7: This issue happens to me also. The only way to get around is restarting my machine. I am using a MacBook
username_8: **Config**
- My `baseUrl` is set to `http://localhost:3000`
- Tested on Cypress 3.1.5 and 3.2.0
- Ubuntu 18.04
**Preview**

username_8: For those who really need to move forward, here is my to cents hack to bypass this issue:
```es6
/**
* Open a link not covered by `baseUrl`.
* See https://github.com/cypress-io/cypress/issues/1777
*/
Cypress.Commands.add('openExternalLink', url => {
cy.visit('/').then(window => {
const linkIdentifier = `__TMP_CY_EXTERNAL_LINK__${btoa(url)}`;
const link = window.document.createElement('a');
link.innerHTML = linkIdentifier;
link.setAttribute('href', url);
link.setAttribute('id', linkIdentifier);
window.document.body.appendChild(link);
cy.get(`#${linkIdentifier}`).click();
});
});
```
username_1: I'm not able to replicate this issue by simply visiting 2 localhosts, with 1 of the localhosts being set as my `baseUrl`.
In Chrome Version 73.0.3683.86

**test code**
```js
it('visits 3030 port', () => {
cy.visit('http://localhost:3030')
})
it('visits 2222 port', () => {
cy.visit('http://localhost:2222')
})
```
**cypress.json**
```json
{
"baseUrl": "http://localhost:3030"
}
```
Can you verify that visiting the second localhost directly - with no baseUrl set - loads the page properly? I want to make sure there is not some misdirection - that the page loads usually within Cypress.
Anyways, we will need a reproducible example in order to reopen the issue unfortunately.
username_9: I have similar issue - I can visit my baseUrl, but not with a relative path
<img width="1152" alt="cypress" src="https://user-images.githubusercontent.com/595192/58246120-2862de80-7d4e-11e9-9d37-fe2f0afaa951.png">
```
cy.visit("http://localhost:4200/index.html"); //OK
cy.visit("index.html"); // fails
```
cypress.json:
```
{
"baseUrl": "http://localhost:4200"
}
```
username_9: so it appears I have missed a subtle difference between `cy.request` and `cy.visit`
beginner's error!
username_10: I'm having the issue where my baseUrl is `http://localhost:4200/myapp`
But cypress with chrome is opening `http://localhost:4200/` which nothing exists there.
 |
arjitjohar/Group10Project | 841274764 | Title: Bug: User Cannot Input Enter Title Of Choice
Question:
username_0: Error Name: User Cannot Enter Title Of Choice
Version: Master – last commit id 733a623
Platform: Windows 10, x64
Description: Once the user selects the input file, and if they desire to choose a different title for their compostion, the converter does not pick up the changes.
Steps to Reproduce:
1. Run The Application
2. Select the tablature folder "Testing Inputs"
3. Select any input file.
4. Type a new title in the title box.
5. Click convert.
Expected: The musicXML file should have the title requested inside the musicXML file.
Actual: The standard title named "Classical Guitar" shows.
Priority: 5
Severity: Low
Screenshot:
The window once the user has typed their personalized title:

Status: Issue closed
Answers:
username_1: Hey, username_0, thank you for the bug report. This issue has now been fixed. |
SAP/node-rfc | 1070689152 | Title: NodeJS Process Termination w/o exception when passing empty string key as rfmParams in client.call
Question:
username_0: **Describe the bug**
Nothing urgent, but when you pass an object with an empty string key in it for the parameter rfmParams to client.call (see screenshot) it will terminate the NodeJS Process without throwing an exception.
**To Reproduce**
Installed via npm locally. Test script:
``` js
const noderfc = require("node-rfc");
const pool = new noderfc.Pool({ connectionParameters: {
dest: "MME", // sap test system connection params
USER:"BPINST",
PASSWD:"<PASSWORD>",
ASHOST:"",
SYSNR:"00",
CLIENT:"100",
LANG:"EN",
} });
(async () => {
try {
// get a client connection instance
const client = await pool.acquire();
const result = await client.call("RFC_PING", {
"": ""
})
console.log(result);
} catch (err) {
// connection and invocation errors
console.error(err);
}
})();
```
**Screenshots**

**Environment**

**Additional context**
Add any other context about the problem here.
Answers:
username_1: Thank you for reporting this bug. It will be fixed in next release.
Status: Issue closed
|
Assistant/ModAssistant | 468433871 | Title: Mods stop working after booting Beat Saber a second time
Question:
username_0: Mods were working great for a little while after I first installed Modusername_1. I followed the instructions, launching Beat Saber without mods first, and then installing them. During this time, I opened and closed the game several times. After a couple of hours, I launched the game again to find there was no mod button in the menu anymore. I uninstalled the game and mods and repeated the whole installation process: launch without mods, close game, install mods. launch again, mods are there, close game, relaunch game, mods are no longer working. I've tried deleting game files, using other mod managers, uninstalling all mods except the required ones. But the mod button in the menu always disappears when I boot the game a second time after installing the mods. During this time, there were no updates to the game.
Status: Issue closed
Answers:
username_1: Mod username_1 simply installs the mod files, it's not responsible in any way for their operation. |
elastic/elasticsearch-dsl-py | 480395162 | Title: Update Document.save documentation (<= 6.4.0)
Question:
username_0: save method can also update a document passing in `id` as a keyword, just like this:
`my_document.save(id='an-id-here')`, this `id` will be passed in to `es.index` here https://github.com/elastic/elasticsearch-dsl-py/blob/6.4.0/elasticsearch_dsl/document.py#L441 and `elasticsearch-py` will handle it as a `POST` here https://github.com/elastic/elasticsearch-py/commit/8e92b857bcb91fcfcb2c37f8525c39d434406620#diff-2696ade2db2bc9506566dc21df9e00e8L358 |
bottlerocket-os/bottlerocket | 993677035 | Title: infrasys: add cross account support
Question:
username_0: <!--
Tips:
- Please search for similar requests, including closed issues.
- Please include details about the environment you're running in.
-->
**What I'd like:**
To resources in different accounts.
**Current Ideas:**
Everywhere we are changing the region, we can change the account.
We can use account switching code, we just need to figure out where we want to put account information in Infra.toml. |
libwfa/libwfa | 172641524 | Title: Inconsistency RASSI/RASSCF for doublets
Question:
username_0: &WFA
h5file $Project.rasscf.h5
&RASSI
NROFJOBIPHS
1 2
1 2
TRD1
&WFA
h5file $Project.rassi.h5
~~~~
Answers:
username_0: &RASSI
NROFJOBIPHS
2 2 2
1 2
1 2
TRD1
&WFA
H5file $Project.rassi.h5
refst 3
~~~~
username_0: This is correct. Spin-tracing will not yield the same result as considering alpha and beta individually
Status: Issue closed
|
LaurentMazare/tch-rs | 748293398 | Title: Setting quantization engine
Question:
username_0: Hello,
I am interested in using tch-rs to run inference on arm devices (such as a Raspberry Pi). After working out how to get libtorch to build for aarch64, things are working great! However, I am having a little bit of difficulty getting quantized models to run. When I try to run a quantized torchscript module I get an error saying:
```
RuntimeError: Didn't find engine for operation quantized::conv2d_prepack NoQEngine
```
I believe the solution in python would be to run the following before loading the model:
```python
torch.backends.quantized.engine = 'qnnpack'
```
And I think the equivalent in C++ (according to [this](https://github.com/pytorch/pytorch/issues/28945#issuecomment-549399690) issue) is something like:
```cpp
auto qengines = at::globalContext().supportedQEngines();
if (std::find(qengines.begin(), qengines.end(), at::QEngine::QNNPACK) != qengines.end()) {
at::globalContext().setQEngine(at::QEngine::QNNPACK);
}
```
Is there any sort of equivalent I can use in tch-rs? Sorry if its a silly question I'm very new to rust (and even newer to using C++ bindings in rust).
## Getting a quantized model (python)
```python
import torch
class Model(torch.nn.Module):
def __init__(self):
super().__init__()
self.quant = torch.quantization.QuantStub()
self.conv = torch.nn.Conv2d(3, 16, 3)
self.relu = torch.nn.ReLU()
self.dequant = torch.quantization.DeQuantStub()
def forward(self, x):
x = self.quant(x)
x = self.conv(x)
x = self.relu(x)
x = self.dequant(x)
return x
# Get fp32 model
model_fp32 = Model()
model_fp32.eval()
# Get int8 model
model_fp32.qconfig = torch.quantization.get_default_qconfig('qnnpack')
model_fp32_fused = torch.quantization.fuse_modules(model_fp32, [['conv', 'relu']])
model_fp32_prepared = torch.quantization.prepare(model_fp32_fused)
model_fp32_prepared(torch.randn(1, 3, 224, 224))
model_int8 = torch.quantization.convert(model_fp32_prepared)
# Get and save int8 torchscript model
model_int8_script = torch.jit.script(model_int8)
model_int8(torch.randn(1, 3, 224, 224))
torch.jit.save(model_int8_script, "model_int8.pth")
```
## Running inference (rust)
```rust
use std::time::{SystemTime};
extern crate tch;
use tch::CModule;
use tch::vision::{imagenet};
pub fn main() {
let module = CModule::load("model_int8.pth").unwrap();
let image = imagenet::load_image_and_resize224("cat.jpg").unwrap();
let now = SystemTime::now();
for _i in 1..10 {
let _out = module.forward_ts(&[image.unsqueeze(0)]).unwrap();
}
println!("{}", now.elapsed().unwrap().as_millis() / 10);
}
```
Answers:
username_1: Hello,
I just merged some code that should make it possible to select the quantization engine, e.g. by using the following:
```rust
tch::QEngine::QNNPACK.set()?;
```
It's very nice that you've been able to get all this to work on arm platforms, could be worth writing some documentation/blog post about it as it might be useful to others.
username_0: Wow amazing, thank you! Sure, once I've got things cleaned up a bit I'd be happy to write it up.
username_0: Raised a [small PR](https://github.com/username_1/tch-rs/pull/289) to add a quantized JIT model example.
username_1: Closing this now that the changes have been merged, feel free to re-open if needed.
Status: Issue closed
|
sony/nmos-cpp | 400850252 | Title: nmos-cpp-registry spends a lot of time in regex code
Question:
username_0: When running continuous registry updates from my node, I see the following in "perf top"
```
4.47% nmos-cpp-registry [.] _ZZNKSt8__detail15_BracketMatcherINSt7__cxx1112regex_traitsIcEELb0ELb0EE8_M_applyEcSt17integral_constantIbLb0EEENKUlvE_clEv
3.20% libc-2.27.so [.] cfree@GLIBC_2.2.5
3.14% libstdc++.so.6.0.25 [.] __dynamic_cast
2.62% libstdc++.so.6.0.25 [.] _ZNK10__cxxabiv121__vmi_class_type_info12__do_dyncastElNS_17__class_type_info10__sub_kindEPKS1_PKvS4_S6_RNS1_16__dyncast_resultE
1.84% libstdc++.so.6.0.25 [.] _ZNKSt7__cxx117collateIcE12do_transformEPKcS3_
1.84% libc-2.27.so [.] malloc
1.54% libc-2.27.so [.] __GI___strcmp_ssse3
1.36% nmos-cpp-registry [.] _ZNSt6vectorIcSaIcEE4dataEv
```
Repeatedly breaking on the BracketMatcher function show two principal contributors. The biggest one is the JSON validator, which appears to be recompiling its regexps at run time, possibly because the it's rebuilding the registration API frequently(?!):
[sorry about the long comments, but I haven't had any luck attaching files to github tickets]
```
#0 std::__detail::_BracketMatcher<std::__cxx11::regex_traits<char>, false, false>::_M_apply (this=0x7fffdd7ddd20, __ch=-125 '\203') at /usr/include/c++/7/bits/regex_compiler.tcc:612
#1 0x0000555555742164 in std::__detail::_BracketMatcher<std::__cxx11::regex_traits<char>, false, false>::_M_make_cache (this=0x7fffdd7ddd20) at /usr/include/c++/7/bits/regex_compiler.h:556
#2 0x000055555573dc61 in std::__detail::_BracketMatcher<std::__cxx11::regex_traits<char>, false, false>::_M_ready (this=0x7fffdd7ddd20) at /usr/include/c++/7/bits/regex_compiler.h:525
#3 0x000055555573e969 in std::__detail::_Compiler<std::__cxx11::regex_traits<char> >::_M_insert_bracket_matcher<false, false> (this=0x7fffdd7de190, __neg=true) at /usr/include/c++/7/bits/regex_compiler.tcc:444
#4 0x000055555573ad67 in std::__detail::_Compiler<std::__cxx11::regex_traits<char> >::_M_bracket_expression (this=0x7fffdd7de190) at /usr/include/c++/7/bits/regex_compiler.tcc:363
#5 0x0000555555735e79 in std::__detail::_Compiler<std::__cxx11::regex_traits<char> >::_M_atom (this=0x7fffdd7de190) at /usr/include/c++/7/bits/regex_compiler.tcc:349
#6 0x0000555555732999 in std::__detail::_Compiler<std::__cxx11::regex_traits<char> >::_M_term (this=0x7fffdd7de190) at /usr/include/c++/7/bits/regex_compiler.tcc:139
#7 0x000055555572fa66 in std::__detail::_Compiler<std::__cxx11::regex_traits<char> >::_M_alternative (this=0x7fffdd7de190) at /usr/include/c++/7/bits/regex_compiler.tcc:121
#8 0x000055555572fa89 in std::__detail::_Compiler<std::__cxx11::regex_traits<char> >::_M_alternative (this=0x7fffdd7de190) at /usr/include/c++/7/bits/regex_compiler.tcc:124
#9 0x000055555572d6d8 in std::__detail::_Compiler<std::__cxx11::regex_traits<char> >::_M_disjunction (this=0x7fffdd7de190) at /usr/include/c++/7/bits/regex_compiler.tcc:97
#10 0x000055555572bb77 in std::__detail::_Compiler<std::__cxx11::regex_traits<char> >::_Compiler (this=0x7fffdd7de190, __b=0x7fff7000c570 "^[^\\s\\/]+\\/[^\\s\\/]+$", __e=0x7fff7000c584 "", __loc=..., __flags=(unknown: 16))
at /usr/include/c++/7/bits/regex_compiler.tcc:82
#11 0x0000555555729d0c in std::__detail::__compile_nfa<char const*, std::__cxx11::regex_traits<char> > (__first=0x7fff7000c570 "^[^\\s\\/]+\\/[^\\s\\/]+$", __last=0x7fff7000c584 "", __loc=..., __flags=(unknown: 16)) at /usr/include/c++/7/bits/regex_compiler.h:203
#12 0x000055555572728a in std::__cxx11::basic_regex<char, std::__cxx11::regex_traits<char> >::basic_regex<char const*> (this=0x7fffdd7de4a0, __first=0x7fff7000c570 "^[^\\s\\/]+\\/[^\\s\\/]+$", __last=0x7fff7000c584 "", __loc=..., __f=(unknown: 16))
at /usr/include/c++/7/bits/regex.h:767
#13 0x00005555557255a7 in std::__cxx11::basic_regex<char, std::__cxx11::regex_traits<char> >::basic_regex<char const*> (this=0x7fffdd7de4a0, __first=0x7fff7000c570 "^[^\\s\\/]+\\/[^\\s\\/]+$", __last=0x7fff7000c584 "", __f=(unknown: 16))
at /usr/include/c++/7/bits/regex.h:512
#14 0x00005555557235f1 in std::__cxx11::basic_regex<char, std::__cxx11::regex_traits<char> >::basic_regex<std::char_traits<char>, std::allocator<char> > (this=0x7fffdd7de4a0, __s="^[^\\s\\/]+\\/[^\\s\\/]+$", __f=(unknown: 16)) at /usr/include/c++/7/bits/regex.h:493
#15 0x00005555559ab8a9 in nlohmann::json_schema_draft4::json_validator::validate_string (this=0x555555f5cbe8, instance=..., schema=..., name="root.data.caps.media_types[1]") at /home/sable/bt/git/nmos-cpp/Development/third_party/nlohmann/json-validator.cpp:766
#16 0x00005555559a9160 in nlohmann::json_schema_draft4::json_validator::validate (this=0x555555f5cbe8, instance=..., schema_=..., name="root.data.caps.media_types[1]") at /home/sable/bt/git/nmos-cpp/Development/third_party/nlohmann/json-validator.cpp:466
#17 0x00005555559a8e51 in nlohmann::json_schema_draft4::json_validator::validate (this=0x555555f5cbe8, instance=..., schema_=..., name="root.data.caps.media_types[1]") at /home/sable/bt/git/nmos-cpp/Development/third_party/nlohmann/json-validator.cpp:438
#18 0x00005555559a9e54 in nlohmann::json_schema_draft4::json_validator::validate_array (this=0x555555f5cbe8, instance=..., schema=..., name="root.data.caps.media_types") at /home/sable/bt/git/nmos-cpp/Development/third_party/nlohmann/json-validator.cpp:569
#19 0x00005555559a9137 in nlohmann::json_schema_draft4::json_validator::validate (this=0x555555f5cbe8, instance=..., schema_=..., name="root.data.caps.media_types") at /home/sable/bt/git/nmos-cpp/Development/third_party/nlohmann/json-validator.cpp:462
#20 0x00005555559aa699 in nlohmann::json_schema_draft4::json_validator::validate_object (this=0x555555f5cbe8, instance=..., schema=..., name="root.data.caps") at /home/sable/bt/git/nmos-cpp/Development/third_party/nlohmann/json-validator.cpp:649
#21 0x00005555559a910e in nlohmann::json_schema_draft4::json_validator::validate (this=0x555555f5cbe8, instance=..., schema_=..., name="root.data.caps") at /home/sable/bt/git/nmos-cpp/Development/third_party/nlohmann/json-validator.cpp:458
#22 0x00005555559aa699 in nlohmann::json_schema_draft4::json_validator::validate_object (this=0x555555f5cbe8, instance=..., schema=..., name="root.data") at /home/sable/bt/git/nmos-cpp/Development/third_party/nlohmann/json-validator.cpp:649
#23 0x00005555559a910e in nlohmann::json_schema_draft4::json_validator::validate (this=0x555555f5cbe8, instance=..., schema_=..., name="root.data") at /home/sable/bt/git/nmos-cpp/Development/third_party/nlohmann/json-validator.cpp:458
#24 0x00005555559a8e51 in nlohmann::json_schema_draft4::json_validator::validate (this=0x555555f5cbe8, instance=..., schema_=..., name="root.data") at /home/sable/bt/git/nmos-cpp/Development/third_party/nlohmann/json-validator.cpp:438
#25 0x00005555559a8e51 in nlohmann::json_schema_draft4::json_validator::validate (this=0x555555f5cbe8, instance=..., schema_=..., name="root.data") at /home/sable/bt/git/nmos-cpp/Development/third_party/nlohmann/json-validator.cpp:438
#26 0x00005555559aa699 in nlohmann::json_schema_draft4::json_validator::validate_object (this=0x555555f5cbe8, instance=..., schema=..., name="root") at /home/sable/bt/git/nmos-cpp/Development/third_party/nlohmann/json-validator.cpp:649
#27 0x00005555559a910e in nlohmann::json_schema_draft4::json_validator::validate (this=0x555555f5cbe8, instance=..., schema_=..., name="root") at /home/sable/bt/git/nmos-cpp/Development/third_party/nlohmann/json-validator.cpp:458
#28 0x00005555559a8e51 in nlohmann::json_schema_draft4::json_validator::validate (this=0x555555f5cbe8, instance=..., schema_=..., name="root") at /home/sable/bt/git/nmos-cpp/Development/third_party/nlohmann/json-validator.cpp:438
#29 0x00005555559a86c0 in nlohmann::json_schema_draft4::json_validator::validate (this=0x555555f5cbe8, instance=...) at /home/sable/bt/git/nmos-cpp/Development/third_party/nlohmann/json-validator.cpp:359
#30 0x0000555555946adc in web::json::experimental::details::json_validator_impl::validate (this=0x555555df8f10, value=..., id=...) at /home/sable/bt/git/nmos-cpp/Development/cpprest/json_validator_impl.cpp:115
#31 0x0000555555944c6a in web::json::experimental::json_validator::validate (this=0x555555f53918, value=..., id=...) at /home/sable/bt/git/nmos-cpp/Development/cpprest/json_validator_impl.cpp:137
#32 0x000055555590a1db in nmos::make_unmounted_registration_api(nmos::registry_model&, slog::base_gate&)::{lambda(web::http::http_request, web::http::http_response, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const&, std::unordered_map<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >, std::hash<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > >, std::equal_to<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > >, std::allocator<std::pair<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > > > > const&)#4}::operator()(web::http::http_request, web::http::http_response, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const&, std::unordered_map<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >, std::hash<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > >, std::equal_to<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > >, std::allocator<std::pair<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > > > > const&) const::{lambda(web::json::value)#1}::operator()(web::json)
(__closure=0x7fff5c003920, body=...) at /home/sable/bt/git/nmos-cpp/Development/nmos/registration_api.cpp:161
#33 0x00005555559139f7 in std::_Function_handler<bool (web::json::value), nmos::make_unmounted_registration_api(nmos::registry_model&, slog::base_gate&)::{lambda(web::http::http_request, web::http::http_response, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const&, std::unordered_map<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >, std::hash<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > >, std::equal_to<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > >, std::allocator<std::pair<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > > > > const&)#4}::operator()(web::http::http_request, web::http::http_response, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const&, std::unordered_map<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >, std::hash<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > >, std::equal_to<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > >, std::allocator<std::pair<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > > > > const&) const::{lambda(web::json::value)#1}>::_M_invoke(std::_Any_data const&, web::json::value&&) (__functor=..., __args#0=...) at /usr/include/c++/7/bits/std_function.h:302
#34 0x00005555558e0e85 in std::function<bool (web::json::value)>::operator()(web::json::value) const (this=0x7fffdd7e18c0, __args#0=...) at /usr/include/c++/7/bits/std_function.h:706
#35 0x00005555559138e3 in pplx::task<web::json::value>::_ContinuationTaskHandle<web::json::value, bool, nmos::make_unmounted_registration_api(nmos::registry_model&, slog::base_gate&)::{lambda(web::http::http_request, web::http::http_response, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const&, std::unordered_map<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >, std::hash<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > >, std::equal_to<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > >, std::allocator<std::pair<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > > > > const&)#4}::operator()(web::http::http_request, web::http::http_response, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const&, std::unordered_map<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >, std::hash<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > >, std::equal_to<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > >, std::allocator<std::pair<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > > > > const&) const::{lambda(web::json::value)#1}, std::integral_constant<bool, false>, pplx::details::_TypeSelectorNoAsync>::_LogWorkItemAndInvokeUserLambda<std::function<bool (web::json::value)>, web::json::value>(std::function<bool (web::json::value)>&&, bool (&&)(web::json::value)) const (this=0x7fff5c009710, _func=..., _value=...) at /usr/local/include/pplx/pplxtasks.h:3859
#36 0x0000555555913719 in pplx::task<web::json::value>::_ContinuationTaskHandle<web::json::value, bool, nmos::make_unmounted_registration_api(nmos::registry_model&, slog::base_gate&)::{lambda(web::http::http_request, web::http::http_response, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const&, std::unordered_map<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >, std::hash<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > >, std::equal_to<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > >, std::allocator<std::pair<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > > > > const&)#4}::operator()(web::http::http_request, web::http::http_response, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const&, std::unordered_map<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >, std::hash<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > >, std::equal_to<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > >, std::allocator<std::pair<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > > > > const&) const::{lambda(web::json::value)#1}, std::integral_constant<bool, false>, pplx::details::_TypeSelectorNoAsync>::_Continue(nmos::make_unmounted_registration_api(nmos::registry_model&, slog::base_gate&)::{lambda(web::http::http_request, web::http::http_response, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const&, std::unordered_map<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >, std::hash<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > >, std::equal_to<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > >, std::allocator<std::pair<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > > > > const&)#4}::operator()(web::http::http_request, web::http::http_response, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const&, std::unordered_map<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >, std::hash<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > >, std::equal_to<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > >, std::allocator<std::pair<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > > > > const&) const::{lambda(web::json::value)#1}, std::integral_constant<bool, false>) const (this=0x7fff5c009710) at /usr/local/include/pplx/pplxtasks.h:3891
#37 0x000055555591357d in pplx::task<web::json::value>::_ContinuationTaskHandle<web::json::value, bool, nmos::make_unmounted_registration_api(nmos::registry_model&, slog::base_gate&)::{lambda(web::http::http_request, web::http::http_response, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const&, std::unordered_map<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >, std::hash<std::__cxx11::basi---Type <return> to continue, or q <return> to quit---
[Truncated]
at /usr/include/c++/7/bits/std_function.h:316
#79 0x00007ffff78483cb in std::function<void (web::http::http_request)>::operator()(web::http::http_request) const () from /usr/local/lib/libcpprest.so.2.10
#80 0x00007ffff7846126 in web::http::experimental::listener::details::http_listener_impl::handle_request(web::http::http_request) () from /usr/local/lib/libcpprest.so.2.10
#81 0x00007ffff79f4f51 in (anonymous namespace)::asio_server_connection::dispatch_request_to_listener() () from /usr/local/lib/libcpprest.so.2.10
#82 0x00007ffff79f39a0 in (anonymous namespace)::asio_server_connection::handle_headers() () from /usr/local/lib/libcpprest.so.2.10
#83 0x00007ffff79f2f89 in (anonymous namespace)::asio_server_connection::handle_http_line(boost::system::error_code const&) () from /usr/local/lib/libcpprest.so.2.10
#84 0x00007ffff79f2243 in (anonymous namespace)::asio_server_connection::start_request_response()::{lambda(boost::system::error_code const&, unsigned long)#2}::operator()(boost::system::error_code const&, unsigned long) const () from /usr/local/lib/libcpprest.so.2.10
#85 0x00007ffff79fa9cb in boost::asio::detail::read_until_match_op<boost::asio::basic_stream_socket<boost::asio::ip::tcp, boost::asio::stream_socket_service<boost::asio::ip::tcp> >, std::allocator<char>, (anonymous namespace)::crlfcrlf_nonascii_searcher_t, (anonymous namespace)::asio_server_connection::start_request_response()::{lambda(boost::system::error_code const&, unsigned long)#2}>::operator()(boost::system::error_code const&, unsigned long, int) () from /usr/local/lib/libcpprest.so.2.10
#86 0x00007ffff7a14c32 in boost::asio::detail::binder2<boost::asio::detail::read_until_match_op<boost::asio::basic_stream_socket<boost::asio::ip::tcp, boost::asio::stream_socket_service<boost::asio::ip::tcp> >, std::allocator<char>, (anonymous namespace)::crlfcrlf_nonascii_searcher_t, (anonymous namespace)::asio_server_connection::start_request_response()::{lambda(boost::system::error_code const&, unsigned long)#2}>, boost::system::error_code, unsigned long>::operator()() () from /usr/local/lib/libcpprest.so.2.10
#87 0x00007ffff7a1239d in void boost::asio::asio_handler_invoke<boost::asio::detail::binder2<boost::asio::detail::read_until_match_op<boost::asio::basic_stream_socket<boost::asio::ip::tcp, boost::asio::stream_socket_service<boost::asio::ip::tcp> >, std::allocator<char>, (anonymous namespace)::crlfcrlf_nonascii_searcher_t, (anonymous namespace)::asio_server_connection::start_request_response()::{lambda(boost::system::error_code const&, unsigned long)#2}>, boost::system::error_code, unsigned long> >(boost::asio::detail::binder2<boost::asio::detail::read_until_match_op<boost::asio::basic_stream_socket<boost::asio::ip::tcp, boost::asio::stream_socket_service<boost::asio::ip::tcp> >, std::allocator<char>, (anonymous namespace)::crlfcrlf_nonascii_searcher_t, (anonymous namespace)::asio_server_connection::start_request_response()::{lambda(boost::system::error_code const&, unsigned long)#2}>, boost::system::error_code, unsigned long>&, ...) () from /usr/local/lib/libcpprest.so.2.10
#88 0x00007ffff7a109f4 in void boost_asio_handler_invoke_helpers::invoke<boost::asio::detail::binder2<boost::asio::detail::read_until_match_op<boost::asio::basic_stream_socket<boost::asio::ip::tcp, boost::asio::stream_socket_service<boost::asio::ip::tcp> >, std::allocator<char>, (anonymous namespace)::crlfcrlf_nonascii_searcher_t, (anonymous namespace)::asio_server_connection::start_request_response()::{lambda(boost::system::error_code const&, unsigned long)#2}>, boost::system::error_code, unsigned long>, {lambda(boost::system::error_code const&, unsigned long)#2}>(boost::asio::detail::binder2<boost::asio::detail::read_until_match_op<boost::asio::basic_stream_socket<boost::asio::ip::tcp, boost::asio::stream_socket_service<boost::asio::ip::tcp> >, std::allocator<char>, (anonymous namespace)::crlfcrlf_nonascii_searcher_t, (anonymous namespace)::asio_server_connection::start_request_response()::{lambda(boost::system::error_code const&, unsigned long)#2}>, boost::system::error_code, unsigned long>&, {lambda(boost::system::error_code const&, unsigned long)#2}&) ()
from /usr/local/lib/libcpprest.so.2.10
#89 0x00007ffff7a0e682 in void boost::asio::detail::asio_handler_invoke<boost::asio::detail::binder2<boost::asio::detail::read_until_match_op<boost::asio::basic_stream_socket<boost::asio::ip::tcp, boost::asio::stream_socket_service<boost::asio::ip::tcp> >, std::allocator<char>, (anonymous namespace)::crlfcrlf_nonascii_searcher_t, (anonymous namespace)::asio_server_connection::start_request_response()::{lambda(boost::system::error_code const&, unsigned long)#2}>, boost::system::error_code, unsigned long>, boost::asio::basic_stream_socket<boost::asio::ip::tcp, boost::asio::stream_socket_service<boost::asio::ip::tcp> >, std::allocator<char>, (anonymous namespace)::crlfcrlf_nonascii_searcher_t, {lambda(boost::system::error_code const&, unsigned long)#2}>(boost::asio::detail::binder2<boost::asio::detail::read_until_match_op<boost::asio::basic_stream_socket<boost::asio::ip::tcp, boost::asio::stream_socket_service<boost::asio::ip::tcp> >, std::allocator<char>, (anonymous namespace)::crlfcrlf_nonascii_searcher_t, (anonymous namespace)::asio_server_connection::start_request_response()::{lambda(boost::system::error_code const&, unsigned long)#2}>, boost::system::error_code, unsigned long>&, boost::asio::detail::read_until_match_op<boost::asio::basic_stream_socket<boost::asio::ip::tcp, boost::asio::stream_socket_service<boost::asio::ip::tcp> >, std::allocator<char>, (anonymous namespace)::crlfcrlf_nonascii_searcher_t, {lambda(boost::system::error_code const&, unsigned long)#2}>*) () from /usr/local/lib/libcpprest.so.2.10
#90 0x00007ffff7a0bfbe in void boost_asio_handler_invoke_helpers::invoke<boost::asio::detail::binder2<boost::asio::detail::read_until_match_op<boost::asio::basic_stream_socket<boost::asio::ip::tcp, boost::asio::stream_socket_service<boost::asio::ip::tcp> >, std::allocator<char>, (anonymous namespace)::crlfcrlf_nonascii_searcher_t, (anonymous namespace)::asio_server_connection::start_request_response()::{lambda(boost::system::error_code const&, unsigned long)#2}>, boost::system::error_code, unsigned long>, (anonymous namespace)::asio_server_connection::start_request_response()::{lambda(boost::system::error_code const&, unsigned long)#2}>(boost::asio::detail::binder2<boost::asio::detail::read_until_match_op<boost::asio::basic_stream_socket<boost::asio::ip::tcp, boost::asio::stream_socket_service<boost::asio::ip::tcp> >, std::allocator<char>, (anonymous namespace)::crlfcrlf_nonascii_searcher_t, (anonymous namespace)::asio_server_connection::start_request_response()::{lambda(boost::system::error_code const&, unsigned long)#2}>, boost::system::error_code, unsigned long>&, (anonymous namespace)::asio_server_connection::start_request_response()::{lambda(boost::system::error_code const&, unsigned long)#2}&) () from /usr/local/lib/libcpprest.so.2.10
#91 0x00007ffff7a09c9d in boost::asio::detail::reactive_socket_recv_op<boost::asio::mutable_buffers_1, boost::asio::detail::read_until_match_op<boost::asio::basic_stream_socket<boost::asio::ip::tcp, boost::asio::stream_socket_service<boost::asio::ip::tcp> >, std::allocator<char>, (anonymous namespace)::crlfcrlf_nonascii_searcher_t, (anonymous namespace)::asio_server_connection::start_request_response()::{lambda(boost::system::error_code const&, unsigned long)#2}> >::do_complete(boost::asio::detail::task_io_service*, boost::asio::detail::task_io_service_operation*, boost::system::error_code const&, unsigned long) () from /usr/local/lib/libcpprest.so.2.10
#92 0x0000555555759242 in boost::asio::detail::task_io_service_operation::complete (this=0x7fff7c017430, owner=..., ec=..., bytes_transferred=0) at /usr/include/boost/asio/detail/task_io_service_operation.hpp:38
#93 0x000055555575a20f in boost::asio::detail::task_io_service::do_run_one (this=0x555555f49cd0, lock=..., this_thread=..., ec=...) at /usr/include/boost/asio/detail/impl/task_io_service.ipp:372
#94 0x0000555555759ebf in boost::asio::detail::task_io_service::run (this=0x555555f49cd0, ec=...) at /usr/include/boost/asio/detail/impl/task_io_service.ipp:149
#95 0x000055555575a50c in boost::asio::io_service::run (this=0x7ffff7dd4868 <crossplat::threadpool::shared_instance()::s_shared+8>) at /usr/include/boost/asio/impl/io_service.ipp:59
```
Answers:
username_1: I've checked the ``api_router`` class I implemented for nmos-cpp and it already caches prebuilt regexes, phew, but as the first call-stack shows, [third_party/nlohmann/json-validator.cpp](https://github.com/sony/nmos-cpp/blob/master/Development/third_party/nlohmann/json-validator.cpp#L656) we adopted just doesn't. That's bad. (It isn't "rebuilding the Registration API"!)
On the other hand, the second call-stack which has a ``regex_search`` with a prebuilt regex (at ``#65``) is awfully deep within that, and I don't understand why it's hitting ``_BracketMatcher`` at the top either. Will see what I can repro (though it's going to depend on the regex implementation.
username_1: https://github.com/pboettch/json-schema-validator has had a "complete rewrite of the validator - aiming at a 2.0-release" and one of the things <NAME> has done is to cache the built regexes. I'll have a play with the latest version to see how compatible and how stable it is.
username_1: Hi @username_0. I've updated to the rewritten json-schema-validator on a branch - https://github.com/sony/nmos-cpp/tree/pboettch-json-schema-validator-update. I've tested it works. I still need to fix some compile warnings for it, but if you were to have a chance to compare performance in your environment, that would be wonderful. Understand if you've moved on to other things! Thanks.
username_0: That's a big improvement! CPU usage goes down from 90%+ to 20%+
Status: Issue closed
username_1: I've got impatient and committed patched versions of nlohmann/json version 3.5.0 and pboettch/json-schema-validator based on my PRs on those projects:
* https://github.com/nlohmann/json/pull/1469
* https://github.com/pboettch/json-schema-validator/pull/49
I will keep an eye out for official releases of these dependencies, but I'm going to close this issue based on our testing. Thanks, @username_0. |
Aviana/LunaUnitFrames | 912351030 | Title: Can't restore old fonts and bars
Question:
username_0: Hi, I've recently updated LUF to the latest version since it's on Curseforge too now.
The only issue I have though is that the old fonts and bars are gone, or at least the ones I used myself:
- DorisPP
- Healbot
I extracted them from the previous releases and put them into the corresponding folders but nothing, the game doesn't show them up.
Is there a way I can solve this?
Thanks, in advance!
Answers:
username_1: You need to install https://www.curseforge.com/wow/addons/sharedmedia as it is mentioned in the readme.
Status: Issue closed
|
krzyzanowskim/CryptoSwift | 445712080 | Title: Sounds not playing in decrypted video
Question:
username_0: Hello, i am using following code to decrypted .mp4 video
import Foundation
import CryptoSwift
extension Data {
func aesEncrypt(key: String, iv: String) throws -> Data{
let encypted = try AES(key: Array(key.utf8), blockMode: CTR(iv: Array(iv.utf8)), padding: .noPadding).encrypt(self.bytes)
return Data(bytes: encypted)
}
func aesDecrypt(key: String, iv: String) throws -> Data {
AES.Variant.init(rawValue: 128)
let decrypted = try AES(key: Array(key.utf8), blockMode: CTR(iv: Array(iv.utf8)), padding: .noPadding).decrypt(self.bytes)
return Data(bytes: decrypted)
}
}
But when i play video not getting soud of video
Answers:
username_1: so you think it's related to CryptoSwift? how?
username_0: Because its playing in normal video but when i decrypt it sound does not playing from video
username_2: @username_0 @username_1 I'm also facing this same issue. After decrypting, the audio is not playing in the video.
username_1: Guys. I can't talk on a video/audio level about cryptography ;) If you could make it down to bytes, that'd give us some insight. So, compare bytes before encryption and after decryption and see if there is a difference.
Status: Issue closed
|
Torniojaws/vortech-backend | 258247094 | Title: Release download counter endpoints and methods
Question:
username_0: The routes:
- [ ] Implement the endpoint ``/downloads/releases``
- [ ] Implement the endpoint ``/downloads/releases/:id``
The GET methods:
- [ ] ``GET /downloads/releases`` should return the download counts of all releases
- [ ] ``GET /downloads/releases/:id`` should return the download count of a specific release
The POST methods:
- [ ] ``POST /downloads/releases`` with a valid JSON should increment the count of a release specified in the JSON
Answers:
username_0: Something like:
```
DownloadsReleases(int `DownloadID`, int `ReleaseID`, datetime `DownloadDate`)
```
Where the amount of downloads is ``COUNT(*)`` of DownloadDates where ReleaseID = ?
For the current data, this would be roughly 23k downloads * 30 bytes in the row = 690 KB of data in the DB, so nothing major.
Status: Issue closed
username_0: Done. |
wix/react-native-navigation | 180619654 | Title: this.props.navigator.showLightBox not working
Question:
username_0: ### Issue Description
Hi,
I tried to use showLightBox but i got the following error
undefined is not a function (evaluating '_platformSpecificDeprecated2.default.showListBox(params)')
when i replaced it with showModel, it worked as expected
_onPressDescription(description){
this.props.navigator.showLightBox({
screen: 'example.DisplayDescription',
style: {
backgroundBlur: "dark", // 'dark' / 'light' / 'xlight' / 'none' - the type of blur on the background
backgroundColor: "#ff000080" // tint color for the background, you can specify alpha here (optional)
},
passProps: {
DisplayText: description
}
});
}
---
### Environment
* React Native Navigation version: 2
* React Native version: 0.25
* Platform(s) (iOS, Android, or both?): tested on Android only
* Device info (Simulator/Device? OS version? Debug/Release?): LG G3 Debug, OS version 5.0
Answers:
username_1: Hey @username_0, lightbox is not implemented on Android.
I'm not entirely sure what out plans are for lightbox on Android, @DanielZlotin can probably shed some light on this matter.
username_2: I actually did some research on this for Android. Android has a heads up notification which I personally dislike the user experience (for example, whatsapp). If you perform a google search for this, you will see all relevant results are about people asking how to disable it. I think it would be good to make iOS/Android consistent but the user experience of such should be noted. The main problem on Android is that the heads up notification could block the top of the app unless you do a swipe action and if you have a whatsapp group that is in conversation. It's almost impossible to get to the top of your app...
username_0: I'm sorry @username_2 but how that related to the light box?
username_2: @username_0 Ouch, sorry about that. We were implementing something on Android that mimic the heads up notification thinking about using lightbox. You could just ignore my previous comment.
username_3: I am using a modal with transparent overlay on Android. Only difference is it comes with its own nav bar. Would be nice to have a native lightbox on Android but its not a big deal for me.
username_0: @username_3 well you can hide the navigation bar from the navigation style, you just need to set
navigatorStyle: {
navBarHidden: true
}
username_4: @username_3 @username_0 Could you please share how you configured modal with transparent overlay. I am trying to show a modal with transparent/transluscent background but I am not able to get it to work. I tried setting the background color for the modal screen to transparent but that didn't work. Here is my modal screen component
```js
import React, {Component} from 'react';
import {
StyleSheet,
Text,
View,
TouchableOpacity,
Dimensions
} from 'react-native';
import navigatorStyles from '../styles/navigationStyles'
export default class MessageModalScreen extends Component {
static navigatorStyle = {
navBarHidden: true
}
constructor(props) {
super(props);
}
render() {
return (
<View style={styles.container}>
<Text style={styles.message}>
{this.props.message}
</Text>
<TouchableOpacity onPress={ this.onDismissPress.bind(this) }>
<Text style={styles.button}>Dismiss</Text>
</TouchableOpacity>
</View>
);
}
onDismissPress() {
this.props.navigator.dismissModal();
}
}
var styles = StyleSheet.create({
container: {
flex: 1,
width: Dimensions.get('window').width,
justifyContent: 'center',
alignItems: 'center',
backgroundColor: 'rgba(0, 0, 0, 0.5)' // Tried 'transparent' also
},
message: {
fontSize: 20,
textAlign: 'center',
margin: 10,
paddingTop: 20
},
button: {
textAlign: 'center',
fontSize: 18,
marginBottom: 10,
marginTop:10,
color: '#4692ad'
}
});
```
It would be of great help if you could let me know how can I make the background of the modals transparent such that the user can still the app in the background.
username_4: Okay, it looks like adding transparency to the background color works on Android but not on iOS. Any pointers on how to make the modals background transparent for iOS as well.
username_3: @username_4 I'm using a lightbox on iOS, which takes its own style parameters e.g:
```
this.props.navigator.showLightBox({
screen: "myApp.showModal",
title: "My Modal",
style: {
backgroundBlur: "none",
backgroundColor: "#888888DD",
},
navigatorStyle: { ... }
})
```
username_4: @username_3 Thanks for the response 👍 . I didn't realize that the solution was right in front of me. :)
username_1: Lightbox is available on Android now.
Status: Issue closed
username_5: @username_1 but this.props.navigator.dismissLightBox(); not work on Android. |
An-Sar/PrimalCore | 878182350 | Title: need help with crafttweaker cauldron
Question:
username_0: Hi, I'm rather new to crafttweaker and I'm trying to add a Primal Cauldron fluid recipe for Rustic's ale wort.
I'm working off the wiki examples of adding the recipe but I guess I have problems with syntax and the recipe doesn't get added.
I'm not sure if this is the place to ask this but I'd love to know what I'm doing wrong.
pastebin link https://pastebin.com/ALQ3YPam
(ignore much of the pastebin, the betterwithmods recipe works; I tried doing the primal one in sub-paragraphs like it shows on the wiki but I'm thinking it doesn't change much so it's all in one line, let me know if it does) #
Answers:
username_1: remove "String", "int", "ILiquidStack", "IIngredient" and "IItemStack" from the recipe
username_0: it works, thanks :)
Status: Issue closed
|
numpy/numpy | 52964587 | Title: np.equal not implemented for string arrays?
Question:
username_0: https://github.com/scikit-learn/scikit-learn/issues/4014
```
a = np.array(['eggs', 'spam', 'spam', 'eggs', 'spam', 'spam', 'spam', 'spam',
'spam', 'spam', 'spam', 'eggs', 'eggs', 'spam', 'eggs', 'eggs',
'eggs', 'eggs', 'eggs', 'spam'],
dtype='|S4')
b = np.array(['eggs', 'spam', 'spam', 'eggs', 'eggs', 'spam', 'spam', 'spam',
'spam', 'eggs', 'spam', 'eggs', 'spam', 'eggs', 'spam', 'spam',
'eggs', 'spam', 'spam', 'eggs'],
dtype='|S4')
print(a == b)
print(np.equal(a, b))
```
```
[ True True True True False True True True True False True True
False False False False True False False False]
NotImplemented
```
Answers:
username_1: Duplicate of #4470, please close.
Status: Issue closed
username_2: Thanks @username_1 . |
rust-lang/rust | 795031564 | Title: `cargo fmt` and other tools don't work for a linked toolchain.
Question:
username_0: This is derived from [this issue](https://github.com/rust-lang/rustup/issues/2548), in which we concluded that the bug should be filed against rustc, as it's an `x.py` problem.
In rustc:
```
$ ./x.py build --stage 1 src/tools/rustfmt/
...
$ find 'build' -name rustfmt
build/x86_64-unknown-linux-gnu/stage1-tools/x86_64-unknown-linux-gnu/release/rustfmt
build/x86_64-unknown-linux-gnu/stage1-tools-bin/rustfmt
build/x86_64-unknown-linux-gnu/stage0/share/doc/rustfmt
build/x86_64-unknown-linux-gnu/stage0/bin/rustfmt
build/tmp-rustbuild-tests/main/tmp/tarball/rustfmt
build/tmp-rustbuild-tests/main/tmp/tarball/rustfmt/B/image/share/doc/rustfmt
```
Then in a project using this toolchain via a rustup link:
```
$ cargo fmt
error: 'cargo-fmt' is not installed for the toolchain 'yk-stage1'
note: this is a custom toolchain, which cannot use `rustup component add`
help: if you built this toolchain from source, and used `rustup toolchain link`, then you may be able to build the component with `x.py`
```
But note that `cp build/x86_64-unknown-linux-gnu/stage1-tools-bin/rustfmt build/x86_64-unknown-linux-gnu/stage1/bin/rustfmt` does not make this work either, so there is probably more to it?
Thanks
Answers:
username_0: A load of comments from me. Hopefully all small.
In the interest of moving on, your next PR could add more tests?
username_1: (I think you posted that comment in the wrong place.)
username_0: Oops, yes.
username_0: Can anyone comment on what the correct fix for this might be?
I notice that if I copy `cargo-clippy` and `clippy-driver` from `stage1-tools-bin/` into `stage1/bin` then I'm able to run `cargo clippy` via the linked toolchain.
I initially considered changing rustup to add `stage1-tools-bin` to the PATH, but that seems incorrect as we would be adding something outside of the toolchain root to PATH.
So should x.py copy or move the binaries from `stage1-tools-bin` into `stage1/bin`?
I'm willing to (attempt to) do the work if we can agree on what the correct fix is.
username_1: I'm not really knowledgeable about `x.py`. You might want to ask in `#t-compiler` on Zulip. |
bosancz/bosan.cz | 435766977 | Title: V programu nejde upravit akce
Question:
username_0: Nejde přidat vedoucí akce - já jsem je přidala přes Správu obsahu, ale mělo by to jít i rovnou z Interní sekce, ne?
Answers:
username_1: Funguje v Přehledu (interni/moje/prehled) ale už ne při manuálním vyhledání sebe sama v sekci Akce (interni/obsah/akce) - testováno uživatelem test1.
username_2: Tohle by mělo být vyřešeno posledním updatem, můžete někdo ověřit? @username_0 , @username_1 , @username_4 ?
username_0: Pořad mi to nejde, resp. jde vybrat vedoucí, ale pak se tam nezobrazí a ani neuloží
username_3: Mobilní verze: Ve Správě obsahu v Akcích při výběru vedoucího nejde zavřít ten modal s výběrem vedoucích, ani tlačítko zpět nefunguje. Dal bych vedle tlačítka Vybrat ještě tlačítko zavřít nebo tak něco.. Zavřít to můžu jen refreshnutím stránky. Ale možná to je jen problém malého displaye, že nejde kouknout mimo ten modal.
username_2: @username_0 : počkala jsi na načtení nové verze? Tj. aktualizovat a počkat, než se zobrazí nahoře oranžový pruh a aktualizovat znovu? (chápu, je to debilní)
username_0: Nepočkala, funguje to 🙂
username_2: @username_3 : Správa obsahu není vyvíjena, opravuji jen ty, co jsou pod loginem, byla by to duplicita práce. Časem se tam přesune vše. Akce se konkrétně vyřeší tím, že bude stejné editační okno, jako je to v Moje akce a tohle se smaže.
Status: Issue closed
|
Mailu/Mailu | 876582291 | Title: smtp plain text authentication permitted
Question:
username_0: Thank you for opening an issue with Mailu. Please understand that issues are meant for bugs and enhancement-requests.
For **user-support questions**, reach out to us on [matrix](https://matrix.to/#/#mailu:tedomum.net).
To be able to help you best, we need some more information.
## Before you open your issue
- [X] Check if no issue or pull-request for this already exists.
- [X] Check [documentation](https://mailu.io/master/) and [FAQ](https://mailu.io/master/faq.html). (Tip, use the search function on the documentation page)
- [X] You understand `Mailu` is made by volunteers in their **free time** — be conscise, civil and accept that delays can occur.
- [X] The title of the issue should be short and simple. It should contain specific terms related to the actual issue. Be specific while writing the title.
## Environment & Versions
### Environment
- [X] docker-compose
- [ ] kubernetes
- [ ] docker swarm
### Versions
To find your version, get the image name of a mailu container and read the version from the tag (example for version 1.7).
```
$> docker ps -a | grep mailu
140b09d4b09c mailu/roundcube:1.7 "docker-php-entrypoi…" 2 weeks ago Up 2 days (healthy) 80/tcp
$> grep MAILU_VERSION docker-compose.yml mailu.env
d0d3826162f4 mailu/roundcube:master "docker-php-entrypoi…" 6 days ago Up 6 days (healthy) 80/tcp
docker-compose.yml: image: ${DOCKER_ORG:-mailu}/${DOCKER_PREFIX:-}nginx:${MAILU_VERSION:-master}
docker-compose.yml: image: ${DOCKER_ORG:-mailu}/${DOCKER_PREFIX:-}unbound:${MAILU_VERSION:-master}
docker-compose.yml: image: ${DOCKER_ORG:-mailu}/${DOCKER_PREFIX:-}admin:${MAILU_VERSION:-master}
docker-compose.yml: image: ${DOCKER_ORG:-mailu}/${DOCKER_PREFIX:-}dovecot:${MAILU_VERSION:-master}
docker-compose.yml: image: ${DOCKER_ORG:-mailu}/${DOCKER_PREFIX:-}postfix:${MAILU_VERSION:-master}
docker-compose.yml: image: ${DOCKER_ORG:-mailu}/${DOCKER_PREFIX:-}rspamd:${MAILU_VERSION:-master}
docker-compose.yml: image: ${DOCKER_ORG:-mailu}/${DOCKER_PREFIX:-}clamav:${MAILU_VERSION:-master}
docker-compose.yml: image: ${DOCKER_ORG:-mailu}/${DOCKER_PREFIX:-}radicale:${MAILU_VERSION:-master}
docker-compose.yml: image: ${DOCKER_ORG:-mailu}/${DOCKER_PREFIX:-}fetchmail:${MAILU_VERSION:-master}
docker-compose.yml: image: ${DOCKER_ORG:-mailu}/${DOCKER_PREFIX:-}roundcube:${MAILU_VERSION:-master}
```
## Description
Further explain the bug in a few words. It should be clear what the unexpected behaviour is. Share it in an easy-to-understand language.
I've been working through the results of our first penetration test and have one item I need some assistance on.
I have a low priority result "SMTP Service Cleartext Login Permitted" received. The IP address this is tagged to is our email's public IP address. Unfortunately, the report does not specify which port this is being tagged on.
Description: The remote host is running an SMTP server that advertises that it allows cleartext logins over unencrypted connections. An attacker may be able to uncover user names and passwords by sniffing traffic to the server if a less secure authentication mechanism (i.e. LOGIN or PLAIN) is used.
Solution: Configure the service to support less secure authentication mechanisms only over an encrypted channel.
I'm not really sure what I need to do. I have an SSL on the mail server and the password stuff should all be encrypted. What might I be missing? Any guidance would be appreciated.
## Replication Steps
Steps for replicating your issue
I sought help
at https://matrix.to/#/#mailu:tedomum.net and I was instructed to view the dovecot documentation (https://doc.dovecot.org/configuration_manual/authentication/).
It is already with this configuration made in the environment but it still presents this flaw in my email environment (mailu).
[Truncated]
admin_1_c1b8d3c6db94 | File "/usr/lib/python3.8/site-packages/flask/app.py", line 2292, in wsgi_app
admin_1_c1b8d3c6db94 | response = self.full_dispatch_request()
admin_1_c1b8d3c6db94 | File "/usr/lib/python3.8/site-packages/flask/app.py", line 1815, in full_dispatch_request
admin_1_c1b8d3c6db94 | rv = self.handle_user_exception(e)
admin_1_c1b8d3c6db94 | File "/usr/lib/python3.8/site-packages/flask/app.py", line 1718, in handle_user_exception
admin_1_c1b8d3c6db94 | reraise(exc_type, exc_value, tb)
admin_1_c1b8d3c6db94 | File "/usr/lib/python3.8/site-packages/flask/_compat.py", line 35, in reraise
admin_1_c1b8d3c6db94 | raise value
admin_1_c1b8d3c6db94 | File "/usr/lib/python3.8/site-packages/flask/app.py", line 1813, in full_dispatch_request
admin_1_c1b8d3c6db94 | rv = self.dispatch_request()
admin_1_c1b8d3c6db94 | File "/usr/lib/python3.8/site-packages/flask/app.py", line 1799, in dispatch_request
admin_1_c1b8d3c6db94 | return self.view_functions[rule.endpoint](**req.view_args)
admin_1_c1b8d3c6db94 | File "/app/mailu/internal/views/auth.py", line 24, in nginx_authentication
admin_1_c1b8d3c6db94 | headers = nginx.handle_authentication(flask.request.headers)
admin_1_c1b8d3c6db94 | File "/app/mailu/internal/nginx.py", line 45, in handle_authentication
admin_1_c1b8d3c6db94 | user_email = raw_user_email.encode("iso8859-1").decode("utf8")
admin_1_c1b8d3c6db94 | UnicodeDecodeError: 'utf-8' codec can't decode byte 0x99 in position 0: invalid start byte
```
```` |
jaegertracing/helm-charts | 571489647 | Title: [Jaeger] Add a batch job to clean up ElasticSearch indices
Question:
username_0: I was wondering if there has been any discussion regarding adding a K8s job to the Helm chart for Jeager to clean up elasticsearch indices (for example: delete indices that are n days old, since each ES index is timestamped by day).
The following image is already published.
https://hub.docker.com/r/jaegertracing/jaeger-es-index-cleaner
Answers:
username_0: @username_1 Thoughts ?
username_1: That's a great point, we could allow for the creation of https://github.com/helm/charts/tree/master/stable/elasticsearch-curator as a subchart or use the recommended practice of ILM which would be configured in whatever manner the elastic chart recommend. But IMHO there should be no additional template to this jaeger chart.
username_1: On second thought, if jaeger are maintaining an image specifically for this, perhaps we could indeed add it as a cronjob. 🤔
@jpkrohling @yurishkuro Is there a reason operator does not recommend using elastic curator or built int ILM of elasticsearch for index deletion, but resort to a custom tool?
username_0: Also, if we end up deciding on a CronJob, i already have one for my purposes that i am using and can make a PR with that. 👍
username_1: Sounds good @username_0 , let's use the cronjob if that's what the operator does. It seems to run es-curator in any case.
username_0: This PR addresses this issue: https://github.com/jaegertracing/helm-charts/pull/60
username_0: @username_1 Should we just close this issue ?
Status: Issue closed
username_1: Yep! Solved via #60
username_2: It would be very helpful if we can add support for passing ES_USERNAME & ES_PASSWORD similar to collector deploy and spark cronjob?
username_1: 💯 @username_2 Would you be willing to make a PR to add this?
username_2: @username_1 Raised a PR #65 for adding ES_USERNAME and ES_PASSWORD environment variable to the jaeger-es-index-cleaner job. |
edison23/konachart | 122743088 | Title: relation link-linkedSite doesnt work
Question:
username_0: When try to migrate, exception is thrown:
```
$ php artisan migrate:refresh
Rolled back: 2015_12_04_155401_create_linkedSites_table
Rolled back: 2015_11_17_140726_create_links_table
Rolled back: 2015_11_17_135541_create_aniumus_table
Rolled back: 2015_11_17_135046_create_seasons_table
Rolled back: 2014_10_12_100000_create_password_resets_table
Rolled back: 2014_10_12_000000_create_users_table
[Illuminate\Database\QueryException]
SQLSTATE[HY000]: General error: 1005 Can't create table 'konachart.#sql-506_30' (errno: 150) (SQL: alter table `links` add constraint links_site_id_foreign foreign key (`site_id`) references `linkedSites` (`id`))
[PDOException]
SQLSTATE[HY000]: General error: 1005 Can't create table 'konachart.#sql-506_30' (errno: 150)
```
this commit https://github.com/username_0/konachart/commit/feaeda53fae97be14c477f7a779a05cd22ffb3ed
Answers:
username_1: I am far from being sure, but I am afraid the issue is referencing table `linkedSites` from `2015_11_17_140726_create_links_table.php` which doesn't exist at that point yet, as the `2015_12_04_155401_create_linkedSites_table` migration is to be migrated afterwards. |
viant/toolbox | 645429837 | Title: go tests fail on 32-bit systems
Question:
username_0: two tests pass on 64-bit systems, but fail on 32-bit systems, such as i386, armhf.
error log:
```
=== RUN Test_AsString
Test_AsString: conversion_test.go:143:
Error Trace: conversion_test.go:143
Error: Not equal:
expected: "6273346999"
actual : "6.273346999e+09"
Diff:
--- Expected
+++ Actual
@@ -1 +1 @@
-6273346999
+6.273346999e+09
Test: Test_AsString
--- FAIL: Test_AsString (0.00s)
=== RUN TestNewCollection
--- FAIL: TestNewCollection (0.00s)
panic: runtime error: invalid memory address or nil pointer dereference [recovered]
panic: runtime error: invalid memory address or nil pointer dereference
[signal SIGSEGV: segmentation violation code=0x1 addr=0x0 pc=0x804a78c]
goroutine 21 [running]:
testing.tRunner.func1.1(0x83569a0, 0x863c860)
/usr/lib/go-1.14/src/testing/testing.go:940 +0x27c
testing.tRunner.func1(0x8ce01e0)
/usr/lib/go-1.14/src/testing/testing.go:943 +0x349
panic(0x83569a0, 0x863c860)
/usr/lib/go-1.14/src/runtime/panic.go:969 +0x122
runtime/internal/atomic.Xadd64(0x8c94f84, 0x1, 0x0, 0xffffffff, 0xf)
/usr/lib/go-1.14/src/runtime/internal/atomic/asm_386.s:105 +0xc
github.com/viant/toolbox/data.(*CompactedSlice).Add(0x8c94f60, 0x8c4ce70)
/build/1st/golang-github-viant-toolbox-0.32.2/obj-i686-linux-gnu/src/github.com/viant/toolbox/data/compacted.go:146 +0x72
github.com/viant/toolbox/data.TestNewCollection(0x8ce01e0)
/build/1st/golang-github-viant-toolbox-0.32.2/obj-i686-linux-gnu/src/github.com/viant/toolbox/data/compacted_test.go:14 +0x30d
testing.tRunner(0x8ce01e0, 0x83b8410)
/usr/lib/go-1.14/src/testing/testing.go:991 +0xb4
created by testing.(*T).Run
/usr/lib/go-1.14/src/testing/testing.go:1042 +0x2ad
FAIL github.com/viant/toolbox/data 0.032s
```
Reference:
- https://tests.reproducible-builds.org/debian/rbuild/unstable/i386/golang-github-viant-toolbox_0.32.2-1.rbuild.log.gz
- https://tests.reproducible-builds.org/debian/rbuild/unstable/armhf/golang-github-viant-toolbox_0.32.2-1.rbuild.log.gz
Answers:
username_1: I've patched converter test,
However, with the second issue, it looks like using int64 data type cause panic,
If there is a way to detect test architecture I can disable that test ?
var initSize = len(s.fieldNames)
if initSize < len(data) {
initSize = len(data)
}
line 146# atomic.AddInt64(&s.size, 1)
username_0: for detecting CPU, maybe you can refer:
- https://github.com/klauspost/cpuid
username_0: tried latest 0.32.3 on i386, and there're two fails:
```
# github.com/viant/toolbox/data/udf [github.com/viant/toolbox/data/udf.test]
src/github.com/viant/toolbox/data/udf/conversion_test.go:142:17: constant 6273346999 overflows int
# github.com/viant/toolbox_test [github.com/viant/toolbox.test]
src/github.com/viant/toolbox/converter_test.go:409:9: constant 1668069210749 overflows int
src/github.com/viant/toolbox/converter_test.go:521:8: constant 1668069210749 overflows int
FAIL github.com/viant/toolbox [build failed]
```
another one:
```
=== RUN TestCompactedSlice_SortedRange
--- FAIL: TestCompactedSlice_SortedRange (0.00s)
panic: runtime error: invalid memory address or nil pointer dereference [recovered]
panic: runtime error: invalid memory address or nil pointer dereference
[signal SIGSEGV: segmentation violation code=0x1 addr=0x0 pc=0x804a78c]
goroutine 8 [running]:
testing.tRunner.func1.1(0x83569a0, 0x863c860)
/usr/lib/go-1.14/src/testing/testing.go:940 +0x27c
testing.tRunner.func1(0xa4a6320)
/usr/lib/go-1.14/src/testing/testing.go:943 +0x349
panic(0x83569a0, 0x863c860)
/usr/lib/go-1.14/src/runtime/panic.go:969 +0x122
runtime/internal/atomic.Xadd64(0xa466fe4, 0x1, 0x0, 0xfffffff, 0x0)
/usr/lib/go-1.14/src/runtime/internal/atomic/asm_386.s:105 +0xc
github.com/viant/toolbox/data.(*CompactedSlice).Add(0xa466fc0, 0xa413540)
/build/golang-github-viant-toolbox-0.32.3/obj-i686-linux-gnu/src/github.com/viant/toolbox/data/compacted.go:146 +0x72
github.com/viant/toolbox/data.TestCompactedSlice_SortedRange(0xa4a6320)
/build/golang-github-viant-toolbox-0.32.3/obj-i686-linux-gnu/src/github.com/viant/toolbox/data/compacted_test.go:211 +0x12a6
testing.tRunner(0xa4a6320, 0x83b8404)
/usr/lib/go-1.14/src/testing/testing.go:991 +0xb4
created by testing.(*T).Run
/usr/lib/go-1.14/src/testing/testing.go:1042 +0x2ad
FAIL github.com/viant/toolbox/data 0.145s
FAIL github.com/viant/toolbox/data/udf [build failed]
```
username_1: Updated test with skip for non 64 bit arch
username_0: @username_1 Thanks for your update!
For latest 0.33.0, the result is:
- https://tests.reproducible-builds.org/debian/rbuild/unstable/i386/golang-github-viant-toolbox_0.33.0-1.rbuild.log.gz
- https://tests.reproducible-builds.org/debian/rbuild/unstable/armhf/golang-github-viant-toolbox_0.33.0-1.rbuild.log.gz
```
# github.com/viant/toolbox/data/udf [github.com/viant/toolbox/data/udf.test]
src/github.com/viant/toolbox/data/udf/conversion_test.go:151:17: constant 2323232323223 overflows int
# github.com/viant/toolbox/storage_test [github.com/viant/toolbox/storage.test]
src/github.com/viant/toolbox/storage/file_info_test.go:13:3: constant 2147484141 overflows int
src/github.com/viant/toolbox/storage/file_info_test.go:14:3: constant 2147484159 overflows int
src/github.com/viant/toolbox/storage/file_info_test.go:15:3: constant 2147484128 overflows int
# github.com/viant/toolbox_test [github.com/viant/toolbox.test]
src/github.com/viant/toolbox/converter_test.go:527:8: constant 1668069210749 overflows int
```
another one:
```
=== RUN TestNewCollection
--- FAIL: TestNewCollection (0.00s)
panic: runtime error: invalid memory address or nil pointer dereference [recovered]
panic: runtime error: invalid memory address or nil pointer dereference
[signal SIGSEGV: segmentation violation code=0x1 addr=0x0 pc=0x804a78c]
goroutine 6 [running]:
testing.tRunner.func1.1(0x83594e0, 0x8642880)
/usr/lib/go-1.14/src/testing/testing.go:940 +0x27c
testing.tRunner.func1(0x94ac1e0)
/usr/lib/go-1.14/src/testing/testing.go:943 +0x349
panic(0x83594e0, 0x8642880)
/usr/lib/go-1.14/src/runtime/panic.go:969 +0x122
runtime/internal/atomic.Xadd64(0x9428fb4, 0x1, 0x0, 0xffffffff, 0xffffff)
/usr/lib/go-1.14/src/runtime/internal/atomic/asm_386.s:105 +0xc
github.com/viant/toolbox/data.(*CompactedSlice).Add(0x9428f90, 0x9451e70)
/build/1st/golang-github-viant-toolbox-0.33.0/obj-i686-linux-gnu/src/github.com/viant/toolbox/data/compacted.go:146 +0x72
github.com/viant/toolbox/data.TestNewCollection(0x94ac1e0)
/build/1st/golang-github-viant-toolbox-0.33.0/obj-i686-linux-gnu/src/github.com/viant/toolbox/data/compacted_test.go:17 +0x31b
testing.tRunner(0x94ac1e0, 0x83bbe50)
/usr/lib/go-1.14/src/testing/testing.go:991 +0xb4
created by testing.(*T).Run
/usr/lib/go-1.14/src/testing/testing.go:1042 +0x2ad
FAIL github.com/viant/toolbox/data 0.036s
```
username_1: Looks like 64 detection, I've added to the test does not work,
if cpuid.CPU.Family < 64 {
return
}
could you please print me the following on the machine the test failed ?
fmt.Println("Name:", cpuid.CPU.BrandName)
fmt.Println("PhysicalCores:", cpuid.CPU.PhysicalCores)
fmt.Println("ThreadsPerCore:", cpuid.CPU.ThreadsPerCore)
fmt.Println("LogicalCores:", cpuid.CPU.LogicalCores)
fmt.Println("Family", cpuid.CPU.Family, "Model:", cpuid.CPU.Model)
fmt.Println("Features:", cpuid.CPU.Features)
fmt.Println("Cacheline bytes:", cpuid.CPU.CacheLine)
fmt.Println("L1 Data Cache:", cpuid.CPU.Cache.L1D, "bytes")
fmt.Println("L1 Instruction Cache:", cpuid.CPU.Cache.L1D, "bytes")
fmt.Println("L2 Cache:", cpuid.CPU.Cache.L2, "bytes")
fmt.Println("L3 Cache:", cpuid.CPU.Cache.L3, "bytes")
username_0: For debian, I used the patch:
https://salsa.debian.org/go-team/packages/golang-github-viant-toolbox/-/blob/debian/sid/debian/patches/0001-Skip-some-tests-fail-at-git-pbuilder-environment.patch
This most important part is:
```
- if cpuid.CPU.CacheLine < 64 {
- return
+ if (32 << uintptr(^uintptr(0)>>63)) < 64 {
+ t.Skip()
```
username_0: Let me clarify. I do not mean you should merge this patch to upstream repository.
I just show you how I workaround current errors from reproducible builds.
Thanks for your working!
username_1: Thanks for the suggestion, updated tests
username_1: Have you run test with the latest code ?
username_0: Sorry for the late reply.
The test code still build fails under 32-bit systems, so I have to disable it.
- https://salsa.debian.org/go-team/packages/golang-github-viant-toolbox/-/blob/debian/sid/debian/rules#L17
But I still think some code need to be updated:
- https://salsa.debian.org/go-team/packages/golang-github-viant-toolbox/-/blob/debian/sid/debian/patches/0002-Skip-some-tests-fail-at-git-pbuilder-environment.patch
I'll send you a PR later.
username_1: Removed 32 bit test
username_0: Except PR #39
there're a few other issues.
(since I set ignore all 32-bit go test errors, so currently reproducible builds all "passed")
- https://tests.reproducible-builds.org/debian/rb-pkg/unstable/amd64/golang-github-viant-toolbox.html
All current log based on latest, + PR #39 patch:
- test fails on i386 (x86 32-bit)
```
=== RUN TestConverter_AsInt
TestConverter_AsInt: converter_test.go:702:
Error Trace: converter_test.go:702
Error: Should be true
Test: TestConverter_AsInt
--- FAIL: TestConverter_AsInt (0.00s)
```
https://tests.reproducible-builds.org/debian/rbuild/unstable/i386/golang-github-viant-toolbox_0.33.0-11.rbuild.log.gz
- panic: runtime error on i386 (x86 32-bit) and armhf (arm 32-bit)
```
=== RUN TestLogger
panic: runtime error: invalid memory address or nil pointer dereference
[signal SIGSEGV: segmentation violation code=0x1 addr=0x0 pc=0x804ab9b]
goroutine 165 [running]:
runtime/internal/atomic.Load64(0xa86b2ec, 0xbfc1ad5d, 0xd905ff2)
/usr/lib/go-1.14/src/runtime/internal/atomic/asm_386.s:194 +0xb
github.com/viant/toolbox.(*LogStream).isFrequencyFlushNeeded(0xa86b2c0, 0x0)
/build/1st/golang-github-viant-toolbox-0.33.0/obj-i686-linux-gnu/src/github.com/viant/toolbox/file_logger.go:138 +0x69
github.com/viant/toolbox.(*LogStream).manageWritesInBatch(0xa86b2c0)
/build/1st/golang-github-viant-toolbox-0.33.0/obj-i686-linux-gnu/src/github.com/viant/toolbox/file_logger.go:166 +0x23b
github.com/viant/toolbox.(*FileLogger).NewLogStream.func1(0xa86b2c0)
/build/1st/golang-github-viant-toolbox-0.33.0/obj-i686-linux-gnu/src/github.com/viant/toolbox/file_logger.go:219 +0x21
created by github.com/viant/toolbox.(*FileLogger).NewLogStream
/build/1st/golang-github-viant-toolbox-0.33.0/obj-i686-linux-gnu/src/github.com/viant/toolbox/file_logger.go:218 +0xfe
FAIL github.com/viant/toolbox 0.239s
```
https://tests.reproducible-builds.org/debian/rbuild/unstable/armhf/golang-github-viant-toolbox_0.33.0-11.rbuild.log.gz
- build error on 32-bit systems:
```
# github.com/viant/toolbox/data/udf [github.com/viant/toolbox/data/udf.test]
src/github.com/viant/toolbox/data/udf/conversion_test.go:142:17: constant 2323232323223 overflows int
# github.com/viant/toolbox/storage_test [github.com/viant/toolbox/storage.test]
src/github.com/viant/toolbox/storage/file_info_test.go:13:3: constant 2147484141 overflows int
src/github.com/viant/toolbox/storage/file_info_test.go:14:3: constant 2147484159 overflows int
src/github.com/viant/toolbox/storage/file_info_test.go:15:3: constant 2147484128 overflows int
```
https://tests.reproducible-builds.org/debian/rbuild/unstable/i386/golang-github-viant-toolbox_0.33.0-11.rbuild.log.gz
Thank you! |
confluentinc/kafka-connect-datagen | 998119342 | Title: Setup to pick the same index for different field names (in options)
Question:
username_0: Is there a way to make sure when the data is generated for two different fields using options, the same index from the options is picked for both the inputs for that particular run?
As an example, below are my input fields and I want to ensure that - (latitude[0],longitude[0]) or (latitude[1],longitude[1]) get picked each time and not any random combinations of them.
```
{
"name": "latitude",
"doc": "The Simulated latitude for the location",
"type": {
"type": "string",
"arg.properties": {
"options": [
"34.0522",
"34.1083",
]
}
}
},
{
"name": "longitude",
"doc": "The Simulated longitude for the location",
"type": {
"type": "string",
"arg.properties": {
"options": [
"118.2436",
"117.2941",
]
}
}
}
```
Much appreciated. |
hasadna/Open-Knesset | 157543056 | Title: Bug: Not recieving attendance data, starting at least Janury 2016
Question:
username_0: http://data.obudget.org/queries/538/source
Per queries, seems like we stop recieving attendace data starting January 2016 (seems like we are getting information until December 2015).
Also, would be useful if we could document exactly what sort of information is actually stored (attendance where?)
Answers:
username_1: see also #691
username_2: Not sure this is connected. But I'll check
username_2: @username_0 Can you please check if we have data from January 2016 before 18/1?
@username_1 the way PresenceManager is run from crontab is wrong (and perhaps some bugs in the presence manager also).. I'll triangulate exactly. But I see the crontab was created in 18/1/2016 how did the presenceManager run before that?
username_1: @username_2 I assume someone just ran it manually once in a while..
username_2: @username_1 sorry for the misleading, it is run correctly (someone wrapped a shell script that kills the running thread and starts a new one, and the cron runs the script.. ) Still investigating what is wrong there. But be sure, this is a **very shaky** implementation.. and should be rewritten sometime
username_2: @username_1 lol I found the problem. the task does update presence (in some weird way) but runs in a different folder then the task that parses the presence data and attach it to members, last time someone copied the data to where syncdata expects it to be was December 2015. which explains @username_0 findings . I'll try to think how this can be fixed..
username_0: @username_2, far as I my queries show, I don't get any information after Jan 1, 2016, and... you just just you found the problem :)
username_2: @username_1 BTW it is connected to #691 (it surfaces the problem, since we are processing an old file with lots of mks not relevant) but should be solved separately. I'm on it
username_3: @username_2 @username_1 What's this issue's current status?
(I passed some attendance data from re:dash to <NAME>, and I'm trying to figure out whether the data is reliable).
btw, when I run the following query:
`select distinct attended.id from committees_committeemeeting_mks_attended attended, committees_committeemeeting meeting
where attended.committeemeeting_id = meeting.id
and meeting.date between date '2016-01-01' and date '2016-08-01'`
I get 9,944 distinct attendance IDs, so it's definitely not "no data" :)
username_2: Attendance and committee attendance have nothing in common, completely
different systems
The regular attendance is still open on my table, mostly "solved" (hacked
around), not completely yet
username_3: So basically "מליאה" attendance is currently incorrect, and committee attendance is correct?
And also, can I help somehow to have it done by this Thursday? :)
username_2: Probably not.. I'll try. I can manually load an updated file for now.. If
this would help?
--
<NAME>
twitter:@username_2 <https://twitter.com/username_2>
Facebook <https://www.facebook.com/username_2>
Tech Blog: https://medium.com/@username_2/
*Personal Blog:*degeladom.wordpress.com
Tel:972-54-6734469
username_3: I guess this is better than publishing the current missing data..
But if there are disadvantages to this solution than there's also the option to just not publish מליאה attendance data.. what do you think?
username_1: There are 2 types of attendance data -
1. General attendance in the Knesset, calculated by looking every few
minutes at Knesset data about which MKs are currently inside the Knesset.
2. Committees attendance data - which also includes מליאה which is treated
as another committee.
The committee attendance data should work well, we recently updated and
tested those scripts. The general attendance had some problems, I'm not
sure what's the status of that.
Status: Issue closed
username_2: So yes the previous problem is now fixed and we have data until 15.9
After that we bump into the knesset new firewall |
seylorx1/BugsOfDischordia | 965887059 | Title: Save Lights positioned incorrectly level 2
Question:
username_0: Expected Result:
the light source is positioned at the center of the symbol
Actual Result:
the light source is offset

another light source offset example

Steps to Reproduce:
1. get to level 2
Build:
b.37<issue_closed>
Status: Issue closed |
clulab/eidos | 324079233 | Title: Namespace for JSON-LD
Question:
username_0: This is being moved here to separated it from other JSON-LD issues. More will appear soon.
Answers:
username_0: So, I think this means that we can use the URL (below) on the left in the JSON-LD and it will map to the one on the right which will take the curious to our specification. If our wiki page moves or changes format, the mapping can be changed to match. @username_1, can you make a rewrite rule for .htaccess to do this mapping? If so, we have about 15 of them just now and they will change as we update the format. I'll make a list.
https://w3id.org/wm/cag/Sentence => https://github.com/clulab/eidos/wiki/JSON-LD#user-content-Sentence
username_0: Cool. The present list then continues with the same pattern for these items: Corpus, Document, Word, Dependency, Entity, DirectedRelation, UndirectedRelation, Groundings, Grounding, Provenance, Interval, State, Modifier, and Trigger. If there is a way to match them generically, it could save a lot of work. Something like ^cag/([^/]+)$ https://github.com/clulab/eidos/wiki/JSON-LD#user-content-$1 would be in that direction. I didn't check the details, though.
username_1: Done: e.g., https://w3id.org/wm/cag/Grounding
username_0: Great. Thanks for sorting out the details.
username_1: no problem |
SpeciesFileGroup/taxonworks | 627876291 | Title: Task - Edit taxon name - hyperlink Comprehensive
Question:
username_0: Expected behavior:
1. When I click the Comprehensive link to go to comprehensive digitization, create an OTU, if one does not exist.
2. When I come to Comprehensive from Edit taxon name, I want to see OTU associated with this name in the Quick list in Smart selector for determination. (this could be a newly created OTU, or an older one, if the taxon was created long time ago).
Status: Issue closed
Answers:
username_0: Perfect, thank you. |
InseeFr/Pogues | 1047587451 | Title: Generation error synthesis for the final user
Question:
username_0: ⚠️ This is a stub issue, to be more thoroughly described in a near future.
If a problem happens during a questionnaire generation, most of the time the user is clueless regarding the root cause.
In the main setting (use at Insee), the generation backend for Pogues is [Eno](https://github.com/InseeFr/Eno/), more precisely the [web service wrapper for Eno](https://github.com/InseeFr/Eno-WS). We must provide a way to catch those generation errors and provide a clean report to the user.
It can be done:
- through the Pogues front-end directly before starting the generation process,
- after the generation process by catching a proper error message coming from the Eno API.
Answers:
username_0: Analysis work to be done in a future sprint.
username_0: A spec is needed, the idea is to cactch errors early, before the generation call. |
pulumi/pulumi-azure | 613394370 | Title: Creating a VirtualNetworkGateway errors despite the resource creating correctly
Question:
username_0: Creating a VirtualNetworkGateway takes a long time on Azure but unlike most API timeouts, pulumi doesn't recognise that the resource was accepted and doesn't add it to the stack state.
(Azure docs say up to 45 minutes: https://docs.microsoft.com/en-us/azure/vpn-gateway/create-routebased-vpn-gateway-cli#CreateGateway)
```
Updating (advanced-networking-test):
Type Name Status Info
pulumi:pulumi:Stack advanced-networking-test-advanced-networking-test **failed** 1 error
+ └─ azure:network:VirtualNetworkGateway vnet-gw **creating failed** 1 error
Diagnostics:
pulumi:pulumi:Stack (advanced-networking-test-advanced-networking-test):
error: update failed
azure:network:VirtualNetworkGateway (vnet-gw):
error: Error waiting for completion of AzureRM Virtual Network Gateway "vnet-gw" (Resource Group "test-dev-eu"): Future#WaitForCompletion: context has been cancelled: StatusCode=200 -- Original Error:
context deadline exceeded
Resources:
102 unchanged
~~~
Previewing update (advanced-networking-test):
Type Name Plan
pulumi:pulumi:Stack advanced-networking-test-advanced-networking-test
+ └─ azure:network:VirtualNetworkGateway vnet-gw create
Resources:
+ 1 to create
102 unchanged
```
Answers:
username_1: @username_0 Did you notice how much time passes until it fails?
@username_2 Could this be related to https://github.com/pulumi/pulumi-terraform-bridge/issues/177?
username_0: @username_1 I'm not certain but I did it a couple of times and about twenty minutes seems right.
username_2: Thanks for the info @username_0! This definitely looks related to https://github.com/pulumi/pulumi-terraform-bridge/issues/177, which I'm looking into. In the meanwhile, you can use a [custom timeout](https://www.pulumi.com/docs/intro/concepts/programming-model/#customtimeouts) as a workaround.
Status: Issue closed
username_2: The underlying problem should be fixed in https://github.com/pulumi/pulumi-azure/releases/tag/v3.5.1 so I’m closing out this issue but please feel free to re-open it if you come across this again.
username_3: I got the problem with it using pulumi-azure 3.11.0. It creates gateways quickly, but then error out waiting for completion . I can only setup timeout to short the waiting time.
`error: Error waiting for completion of AzureRM Virtual Network Gateway`
Anyways, creating resources look ok , pulumi can't just verify completion correctly |
eroshenkoam/imagediff-example | 646690642 | Title: Пишем обычный тест
Question:
username_0: Добавляем зависимости:
```
compile('com.codeborne:selenide:4.2.1')
compile('org.junit.jupiter:junit-jupiter-api:5.3.0')
compile('org.junit.jupiter:junit-jupiter-engine:5.3.0')
compile('org.junit.jupiter:junit-jupiter-params:5.3.0')
```
Пишем обычный тест:
```
@BeforeEach
public void setUp() {
WebDriverManager.chromedriver().setup();
Configuration.browser = "chrome";
Configuration.browserSize = "1200x800";
}
@Test
public void testFullPage() throws Exception {
final String request = "Погода в Санкт-Петербурге";
open("https://yandex.ru");
$("#text").sendKeys(request);
$("#text").submit();
$("[data-fast-wzrd='weather']").should(Condition.exist);
}
@AfterEach
public void afterEach() {
WebDriverRunner.closeWebDriver();
}
``` |
jasonmamy/cordova-wheel-selector-plugin | 567226831 | Title: Text not showing up on android 10
Question:
username_0: 
Failed to set text color in Android 10 Devices.
I fork the branch and made one quick fix, please merge that.
https://github.com/username_0/cordova-wheel-selector-plugin/commit/80c1ecb69fc22ea7148f8d6a615eea605fdfe9ae#diff-8bc41c3962ff62f78ac188b0c90cb9c6
Answers:
username_1: Great! Do you think you could create a 'pull request' so I can merge it
seamlessly into the code base?
Thanks, Jason
username_0: The pull request has been submitted. Thanks.
username_1: Thanks, everything is updated (new release 1.1.5) and published to npm.
Status: Issue closed
|
eclipse/microprofile-rest-client | 294525439 | Title: Need to propagate HTTP headers and cookies
Question:
username_0: If microservice A wants to call microservice B via the rest client "remote interface" for B, there should be an option to flow along the HTTP headers and cookies, from the inbound request to A, to the outbound request to B. Automatically forwarding along security credentials, like an Authentication header, or an LTPA SSO cookie, is one key scenario, though there are likely others too. Without this, a secured microservice B will reject the calls from A, if the Auth header, or the SSO cookie, don't flow along on the requests from microservice to microservice.
I was thinking a new CDI annotation, perhaps named @PropagateHeaders, could be defined at the point of injection. It probably needs to take the HttpServletRequest as a param (so it has something to get the inbound headers/cookies from) or maybe use of an @PostContruct annotation.
I don't have all the answers - I just know we need a way to tell the rest client that it needs to not lose the security (or other) headers/cookies that flowed in to microservice A, when it calls microservice B. This is possibly related to whatever resolution occurs for issue #16.
Answers:
username_0: FYI, today my code manually copies these headers/cookies. I have the following method:
//forward headers (including cookies) from inbound request
private static void copyFromRequest(HttpURLConnection conn, HttpServletRequest request) {
Enumeration<String> headers = request.getHeaderNames();
if (headers != null) {
while (headers.hasMoreElements()) {
String headerName = headers.nextElement(); //"Authorization" and "Cookie" are especially important headers
String headerValue = request.getHeader(headerName);
conn.setRequestProperty(headerName, headerValue); //odd it's called request property here, rather than header...
}
}
}
I'd like to switch to using the strongly-typed "remote interface" for my JAX-RS service, but I need it to do something like the above, so that I don't lose security context on microservice to microservice calls.
username_0: Note I had to add an @Context param to my JAX-RS method to get access to the HttpServletRequest, to use when my Portfolio microservice makes a call to the downstream StockQuote microservice:
@GET
@Path("/{owner}")
@Produces("application/json")
public JsonObject getPortfolio(@PathParam("owner") String owner, @Context HttpServletRequest request) throws IOException, SQLException {
username_0: Attaching my Portfolio.java, that shows the ugly manual way I'm making microservice to microservice calls today. I want to replace that with this much cleaner strongly-typed REST client approach instead, once it can make calls without losing the security context. Note I had to zip it, since Git won't let me attach a file with a .java extension.
[Portfolio.zip](https://github.com/eclipse/microprofile-rest-client/files/1696793/Portfolio.zip)
username_1: Can you set these forwardings up using `@HeaderParam` and `@CookieParam` in parent resources that return your resource as a subresource?
I'm able to do this kind of thing with `jersey-proxy-client`.
username_1: So let's say I have a `RootResource` class that models everything available under `/`. I might do:
```java
Client client = ClientBuilder.newClient();
WebTarget root = client.target("{host}").resolveTemplate("host", this.host, false);
rawEndpoint = WebResourceFactory.newResource(RootResource.class, root);
User user = rawEndpoint.users().sessions().post(new Credentials(orgUser, orgPass));
authEndpoint = WebResourceFactory.newResource(AuthResource.class, root);
rootEndpoint = authEndpoint.withAuthToken(user.sessionToken);
```
where `AuthResource` looks like:
```java
public interface AuthResource {
@Path("/")
RootResource withAuthToken(@HeaderParam("X-Auth-Token") final String sessionToken);
}
```
and where `RootResource` is anything you want it to be.
I'm a big believer that a single JAX-RS interface should only model the single path parts under it, returning subresources until you get to terminal HTTP verb actions (making proxy clients model the URL structure, where `/` in URLs become `.()` in fluent API calls).
username_0: Forwarding the Open Tracing headers is another use case here.
@username_6 - this is the Git issue I mentioned on our call earlier.
username_1: @username_0 -- Can you confirm whether Rest Client 1.0 supports the approach I advocate above? I think it should! @username_3 wanted to make sure subresources were well-supported by 1.0. So long as parent resource context is added to the subresource when it is returned, you should be able to do what you want in a very declarative manner.
Rest Client 1.0 spec does show `@HeaderParam` examples, but they're very contrived and encouraging of bad practice, IMHO.
username_1: The down side of my approach is it can't blindly forward all headers. The headers have to be explicitly defined in a parent resource, which is perfect for all the cases I've encountered. I, frankly, don't want to forward every junk header that might be included (I'm looking you, `X-Requested-With`!!!).
In my cases, all the headers I want to setup or forward are well-known, so my approach is more than adequate.
username_1: In fact, I would say blindly forwarding all headers could be dangerous and possibly lead to "header injection" scenarios.
username_0: @username_1 - if I understand your suggestion, it would make the header values "first class" parameters on the call to the REST API/operation, but I'd still have to do something like I originally described to pluck such values off of the incoming request, so that I could forward them as such parameters to the outgoing call. Getting to where that could happen automatically is what I was hoping for. But thanks for the suggestions - I'll give that a try!
username_1: Yes, you would have to pick the specific headers you want to forward off of the incoming `@Context HttpServletRequest request` or `@Context HttpHeaders headers`.
I would make these class-level members instead of method parameters as you show above, that way you don't pollute the public API you can publish to others in an interfaces jar.
(I put my JAX-RS annotations on interfaces only so that I can publish interfaces jars for others to use with their preferred proxy clients.)
username_1: To your point, @username_0, `@HeaderParam` could be used to declaratively pick specific headers out of the context instead of programmatically picking them out. You still would have to programmatically forward the picked-out header to another resource, but it should be a simple pass-thru at that point:
```java
public interface ClientResource {
@Path("/")
RootResource withHeaders(@HeaderParam("X-Random-Header-Forwarded") randomHeader);
}
@Path("random")
@RequestScoped
public class RandomEndpoint {
@HeaderParam("X-Random-Header)"
String randomHeader;
// setup proxy client as appropriate with ClientResource
@GET
public get() {
proxyClient.withHeaders(randomHeader).whateverElse().post(somePostBodyPojoHere);
}
}
```
This would pass incoming `X-Random-Header` as `X-Random-Header-Forwarded` on the outgoing request. The two headers could, of course, have the same name, but I named them differently for illustrative purposes.
username_2: Hi @username_0 - apologies, but I'm struggling with the requirement of this issue. Is the requirement to be able to forward all headers/cookies? Or forward only select headers/cookies?
I agree with @username_1 that blindly forwarding all headers/cookies is dangerous. Would it be sufficient to have an annotation that forwards a set of headers/cookies that is either specified in the annotation or via MP Config?
username_0: Hi Andy. I think if there were an easy way to get the other MicroProfile-related headers to flow, that would be sufficient. Mostly that means the "Authorization" header (where the JWT flows) and the OpenTracing headers. I don't really feel I should have to code up my "remote interface" differently because I want these MicroProfile headers for JWT and/or OpenTracing to flow. Maybe some annotation/CDI thing on the remote interface could help? Happy to have a discussion at your convenience.
username_2: So I think this could be accomplished using `@BeanParam` where the specific bean has fields annotated with `@CookieParam` and `@HeaderParam`, and then the bean would be passed into the client interface method. Even so, I can imagine that it is still not as simple or clean as the `@PropagateHeader` annotation you proposed.
So maybe what we need is something like:
```
@RegisterRestClient
@Propagate(cookies={"MyCookie1"
"MyCookie2"},
headers={"Authorization",
"MyHeader1"})
public interface MyServiceClient { //...
```
Then the MP Rest Client implementation would be responsible for obtaining the cookies and headers from the current incoming request (using JAX-RS) and putting them on the outbound (MP Rest Client) request. In order for this to work the MP Rest Client implementation would need to register a hook with the JAX-RS runtime that it is running in (i.e. a `ContainerRequestFilter`) - currently, JAX-RS doesn't have a spec-defined means to register providers globally - see chkal's comment on JakartaEE JAX-RS [issue 596](https://github.com/eclipse-ee4j/jaxrs-api/issues/596) for more details. Most JAX-RS implementations have their own proprietary mechanism to support this though.
The other thing we talked about on the call today was the notion of automatically setting some headers to be forwarded via MP Config. Then anything declared in the `@Propagate` annotation would add additional headers/cookies to be propagated.
username_3: Adding a `Propagate` would mean that you need client stubs that are aware of how they're being used. E.g. if my microservice is messaging based, then the receipt of a message has no HTTP request in context and as a result wouldn't have headers to send over. Should it fail if `Propagate` is setup? If so, then we need a second interface.
username_0: Any news on this? The only thing I really need is the Authorization header, containing the JWT, to get propagated, so that the thing I'm calling doesn't reject the call for lack of the JWT with the right claims. I think most people using this will need the auth to flow. Everything else can wait till later, but this otherwise really cool feature is severely hampered by not having built-in JWT integration (which is odd, given JWT is one of the MicroProfile specs). I'm fine with making it a param on the Java interface (and maybe such propagation code would only get called if its is defined as needed by the method on the interface). Let me know if I can help. Thanks!
username_2: Hi @username_0 - nothing from a spec perspective, but I coded up a sample that should make propagation possible:
https://github.com/username_2/HttpHeaderPropagation/blob/master/src/io/openliberty/propagate/HeaderPropagationFilter.java
This will register a `ContainerRequestFilter` for incoming JAX-RS request and will store the requested headers in a `ThreadLocal` object. Then when the outbound MP Rest Client call is invoked, it will run through this same class as a `ClientRequestFilter` and will add those headers to the outbound request. Can you give this approach a try? If this works, we may want to document it on a MP Rest Client blog post/doc page/etc. rather than modify the spec. Thanks
username_4: Hi @username_0, I was wondering if you successfully propagated the auth header because I am having a similar issue with that.
username_5: Another use case is the propagation of tracing headers. See also
https://github.com/eclipse/microprofile-opentracing/issues/82
username_6: Just heads up... we are trying to find a solution for https://github.com/eclipse/microprofile-config/issues/382 yesterday and discussed a couple of potential solutions. We will continue the discussion in the coming week...
username_7: Would this propagation work properly on another thread? If I do some async operation and call a remote service, would it still propagate?
username_2: @username_7 my workaround proposal in [this thread](https://github.com/eclipse/microprofile-rest-client/issues/73#issuecomment-387837911) would not work with async methods as-is. It should be possible to make it work using `AsyncInvocationInterceptors` to transfer the thread locals to the async thread though.
The final solution for this issue should work with async methods though - certainly this can be controlled in the MP Rest Client implementation. I can foresee issues with async methods on the server-side, since JAX-RS has no concept of `AsyncInvocationInterceptors`.
username_6: @username_2 do you still need help from MP Config?
username_2: Resolved by PRs #143 and #149.
Status: Issue closed
|
Azure/azure-sdk-for-media-services | 78570976 | Title: sample code generates a 400 bad request
Question:
username_0: The following sample code, used twice in CopyBlobsIntoAnAsset/Program.cs, consistently receives a 400 Bad Request from Azure.
var mediaBlobContainer = cloudBlobClient.GetContainerReference(cloudBlobClient.BaseUri + "mediafiles001");
mediaBlobContainer.CreateIfNotExists();
Still getting familiar with the API--but this is presumably from sending a fully qualified URL instead of a container name, as the following works as expected.
var mediaBlobContainer = cloudBlobClient.GetContainerReference("mediafiles001");
mediaBlobContainer.CreateIfNotExists();
Answers:
username_1: Thanks, this worked but I got stuck in the next step when trying to create a new asset. Have you encountered issues there also?
username_2: Sure, here's my working code on production (var names & message changed):
private IAsset CreateMediaAsset(CloudFile model, string lecId)
{
CloudMediaContext context = new CloudMediaContext(mediaAccountName, mediaAccountKey);
CloudStorageAccount storageAccount = CloudStorageAccount.Parse(storageConnectionString);
CloudBlobClient cloudBlobClient = storageAccount.CreateCloudBlobClient();
CloudBlobContainer mediaBlobContainer = cloudBlobClient.GetContainerReference(storageContainerReference);
mediaBlobContainer.CreateIfNotExists();
// Create a new asset.
IAsset asset = context.Assets.Create("Lec-" + lecId + "-" + Guid.NewGuid(), AssetCreationOptions.None);
IAccessPolicy writePolicy = context.AccessPolicies.Create("writePolicy", TimeSpan.FromMinutes(120), AccessPermissions.Write);
ILocator destinationLocator = context.Locators.CreateLocator(LocatorType.Sas, asset, writePolicy);
// Get the asset container URI and copy blobs from mediaContainer to assetContainer
Uri uploadUri = new Uri(destinationLocator.Path);
string assetContainerName = uploadUri.Segments[1];
CloudBlobContainer assetContainer = cloudBlobClient.GetContainerReference(assetContainerName);
string fileName = HttpUtility.UrlDecode(Path.GetFileName(model.BlockBlob.Uri.AbsoluteUri));
var sourceCloudBlob = mediaBlobContainer.GetBlockBlobReference(fileName);
sourceCloudBlob.FetchAttributes();
if (sourceCloudBlob.Properties.Length > 0)
{
IAssetFile assetFile = asset.AssetFiles.Create(fileName);
var destinationBlob = assetContainer.GetBlockBlobReference(fileName);
destinationBlob.DeleteIfExists();
destinationBlob.StartCopyFromBlob(sourceCloudBlob);
destinationBlob.FetchAttributes();
if (sourceCloudBlob.Properties.Length != destinationBlob.Properties.Length)
model.UploadStatusMessage += "Failed to copy as Media Asset!";
}
// Delete destination locator as we want our assets to access through media services only
destinationLocator.Delete();
writePolicy.Delete();
// Delete temp blob to free up chunked uploaded space
sourceCloudBlob.Delete();
// Refresh the new asset
asset = context.Assets.Where(a => a.Id == asset.Id).FirstOrDefault();
var ismAssetFiles = asset.AssetFiles.ToList().Where(f => f.Name.EndsWith(".mp4", StringComparison.OrdinalIgnoreCase)).ToArray();
if (ismAssetFiles.Count() != 1)
throw new ArgumentException("The asset should have only one, .ism file");
ismAssetFiles.First().IsPrimary = true;
ismAssetFiles.First().Update();
model.UploadStatusMessage += " File uploaded successfully by id: " + asset.Id;
model.AssetId = asset.Id;
return asset;
}
Hope this helps, let me know if you still face issue. |
flyway/flyway-play | 799752208 | Title: `flyway.ignoreMissingMigrations` does not seem to work.
Question:
username_0: Added `db.default.migration.ignoreMissingMigrations=true` with no effect.
I checked `PlayInitializer.scala` and it seems to map only few of the many flyway properties.
Is there another way of setting this configuration?
I added `flyway.conf` to my resources folder with this config but did not help.
Answers:
username_0: Looks like this is fixed in master but we are on play 2.6.x so 5.2 is the highest available version and that still has this issue.
username_0: FWIW this is fixed here:
https://github.com/flyway/flyway-play/commit/cf274a3eb023588ce360fc5c993495a100c1b1bf
Any chance of getting this against play 2.6.x? |
okteto/okteto | 1106768976 | Title: e2e test for `okteto deploy --build`
Question:
username_0: Implement an e2e test for the new `okteto deploy --build` command.
The test should cover:
- `okteto deploy` builds images if they don't exist
- `okteto deploy` doesn't build images if they already exist
- `okteto deploy --build` always builds images even if they already exist<issue_closed>
Status: Issue closed |
CoinAlpha/hummingbot | 1071855239 | Title: Solana Connector
Question:
username_0: We have made substantial progress on building Gateway V2, the new standalone middleware that connects to various blockchains and the DEXs on them. While Gateway won’t be merged to the master branch until Q1 2022, developers can already check out the code on the feature branch: https://github.com/coinalpha/hummingbot/tree/feat/gateway-v2
We're seeing a lot more interest in building connectors to other blockchains like Solana. We have a private Telegram group for developers from the community to discuss how to integrate Hummingbot to Solana.
Email <EMAIL> in order to get access to these private Discord and Telegram groups. |
tt-acm/Spectacles.GrasshopperExporter | 117858248 | Title: Give user warning if Layer name is too long
Question:
username_0: @anagpuyol --
If the layer name supplied is too long to be displayed in the Spectacles Web Browser we should turn the component orange and give the user a warning. OR we should make sure the Browser can support longer names.

 |
chibisov/drf-extensions | 398007859 | Title: Custom HyperlinkedIdentityField for HyperlinkedModelSerializer
Question:
username_0: Hi,
I had difficulty with a HyperlinkedModelSerializer url for a nested viewset, and the best solution I found was this piece of code:
https://github.com/miki725/formslayer/blob/master/formslayer/pdf/relations.py#L7-L46
I would like to know if there is a solution within the library I have missed. If not, addition of a similar class would be nice.
Thanks. |
autofac/Autofac | 156509329 | Title: Using Autofac4 & Asp.net core, how to get IContainer from IServiceProvider instance?
Question:
username_0: Is there anyway to get the container from the IServiceProvider which is the return value of Startup.ConfigureServices.
I need to create a object with some specific parameters, and IServiceProvider don't support this feature.
Thanks,
IL
Answers:
username_1: Instead of trying to find the container (it's bad practice to resolve from the container if avoidable - can lead to inadvertent memory leaks for disposable components) resolve an ILifetimeScope from the service provider and use that.
Status: Issue closed
username_2: @username_1 can you show an example of how can we do that?
username_1: `IServiceProvider.GetService<ILifetimeScope>()` |
Leko/hothouse | 340850270 | Title: `SyntaxError: Unexpected token` is occered at local(mac)
Question:
username_0: SyntaxError: Unexpected token { in JSON at position 222
```
Answers:
username_1: @username_0 Thank you for your reporting.
What version of hothouse are you using?
Currently (`Fri Jul 13 2018 00:22:33 GMT+0900 (JST)`) we published new version to support line json of Yarn outdated.
Please try [email protected].
Status: Issue closed
username_1: I'll close this issue.
If you have any trouble in the latest version, please re-open this issue. |
google/oss-fuzz | 432641428 | Title: Linking errors in some builds (e.g. envoy, openssl)
Question:
username_0: e.g. envoy and openssl build are failing for a week with some linking issues. There was clang roll in Chromium around that time, so I'm going to revert the roll here in OSS-Fuzz and see what happens.
Answers:
username_0: So, envoy build got fixed, but not OpenSSL :/
username_1: I think envoy builds have failed to link since this change was made
username_0: Which change?
It succeeded after "[infra] Revert the recent clang roll to fix some builds (#2315). (#2316)" two times, than failed ones, then recovered, and now is broken with the following regression range:
ChangesChange LIB_FUZZING_ENGINE to -fsanitize=fuzzer in libFuzzer builds (detail / githubweb)[libical] Make sure to do an apt update before apt install (#2319) (detail / githubweb)[libxslt] Start with integration (#2320) (detail / githubweb)[libxslt] Update project.yaml: replace Chromium owner in auto_cc. (detail / githubweb)
--
Change LIB_FUZZING_ENGINE to -fsanitize=fuzzer in libFuzzer builds (detail / githubweb)
[libical] Make sure to do an apt update before apt install (#2319) (detail / githubweb)
[libxslt] Start with integration (#2320) (detail / githubweb)
[libxslt] Update project.yaml: replace Chromium owner in auto_cc. (detail / githubweb)
username_1: The revert. I meant to say that the revert hasn't solved all of envoy's build problems (though it may have solved one of them). The latest builds are failing because of what I think is a linker OOM
Here's an excerpt from the logs:
```
Step #12: clang-9: error: unable to execute command: Killed
Step #12: clang-9: error: linker command failed due to signal (use -v to see invocation)
```
username_1: The last build succeeded but I think the build will fail again randomly if this issue isn't solved (sorry for taking over this issue, this should probably moved).
username_0: Ha, interesting. Let me see if we can bump the RAM on the builder (I'm afraid we can't).
username_0: Yeah, we're using the most powerful machine type available (`N1_HIGHCPU_32`) and the only customizeable thing is the disk size.
username_0: @oliverchang may know if there are any possibilities :)
username_0: @username_2 this workaround has been reverted, looks like envoy is the only project (maybe file is the second one, that's it) that is having troubles with a newer clang version, so I'm afraid this has to be fixed on envoy side, sorry :(
username_1: @username_0 for the record "file" build succeeds now, I'm not sure any other project failed because of this other than envoy.
username_0: Nice, thanks for checking!
username_2: OK, I see the build failures happening since this morning. I'll dig into this as time allows, thanks for the heads up.
Status: Issue closed
username_1: I think this can be closed as recent Tor and envoy builds have been succeeding.
username_3: I now get the same error (just with clang-11) in coverage builds of Qt. Is this something I need to fix in Qt? The comments above look like it's a problem with the platform.
username_0: @username_3 is it still happening for you? On our builder the coverage build is doing fine since June 9th.
username_3: That doesn't necessarily mean the linking would work now. I had to remove the affected Qt submodule from oss-fuzz for other reasons: https://github.com/google/oss-fuzz/pull/3949
I'll try locally whether linking it succeeds now.
username_3: When I revert https://github.com/google/oss-fuzz/pull/3949 I still get the same linker error. |
uqzqiao/M-GCTA | 721536697 | Title: Application to twin cohort?
Question:
username_0: Hi,
I was wondering about the application of m-gcta to twin data (that includes information on maternal genotype). In order for the software to run, would one have to basically exclude one of the twins from each pair so that the related combos were mom-offspring rather than mom-twins. I hope that question makes sense!
Thanks so much!
Jess |
Nolanus/ngx-page-scroll | 849871736 | Title: PeerDependency issue
Question:
username_0: Hello hello,
[email protected] has a peerDependency to [email protected] instead of @7.0.4 !!
Beside not having your last code revision, it also lead to an issue when using current angular version (^11.0.0) and updating with current npm version (7):
because [email protected] wants to install [email protected] which is not compatible (as PeerDependecies) with angular ^11.0.0 ... and make the update fails :/
Answers:
username_0: One way to get around this, is to first update the 'core' package and then update the 'main one' with `--legacy-peer-deps`
```
npm update ngx-page-scroll-core
npm update --legacy-peer-deps ngx-page-scroll
```
Which will put both in version 7.0.4 :)
username_1: Sorry for the inconvenience, fixed with v7.0.5
Status: Issue closed
|
facebook/react-native | 211615188 | Title: RN start command uses so much ram.
Question:
username_0: ### Description
in recent versions of RN 40+, react native start command uses more than 1.5G of RAM.
### Additional Information
* React Native version: 42
* Platform: both
* Operating System: Linux
Answers:
username_1: +1, it would start 4~6 node instance and every one use 300~400MB RAM.
username_2: cc @davidaurelio, any idea on how we can track down the RAM-hungry parts of packager?
Status: Issue closed
username_3: Hi there! This issue is being closed because it has been inactive for a while. Maybe the issue has been fixed in a recent release, or perhaps it is not affecting a lot of people. Either way, we're automatically closing issues after a period of inactivity. Please do not take it personally!
If you think this issue should definitely remain open, please let us know. The following information is helpful when it comes to determining if the issue should be re-opened:
- Does the issue still reproduce on the latest release candidate? Post a comment with the version you tested.
- If so, is there any information missing from the bug report? Post a comment with all the information required by the [issue template](https://github.com/facebook/react-native/blob/master/.github/ISSUE_TEMPLATE.md).
- Is there a pull request that addresses this issue? Post a comment with the PR number so we can follow up.
If you would like to work on a patch to fix the issue, *contributions are very welcome*! Read through the [contribution guide](http://facebook.github.io/react-native/docs/contributing.html), and feel free to hop into [#react-native](https://discordapp.com/invite/0ZcbPKXt5bZjGY5n) if you need help planning your contribution. |
kubernetes-client/csharp | 545267386 | Title: Port-forward strange..
Question:
username_0: There's something not quite working in the [port-forward example](https://github.com/kubernetes-client/csharp/blob/master/examples/portforward/PortForward.cs), does anyone have something that is equivalent of doing this:
`kubectl port-forward jupyter-notebook-pod-123 8888:8888 -n default` ?
When I try to launch a jupyter workbook pod the UI is not fully loaded, ex. you cannot make a new Python3 notebook.
I checked #297, and it seemed to work if I added `listener.Close();` using @gittadesushil suggestion. But then after a short while the connection seems to break down.
**To replicate:**
Use the following yaml-file:
```
apiVersion: apps/v1
kind: Deployment
metadata:
name: jupyter-notebook
labels:
app: jupyter-notebook
spec:
replicas: 1
selector:
matchLabels:
app: jupyter-notebook
template:
metadata:
labels:
app: jupyter-notebook
spec:
containers:
- name: minimal-notebook
image: jupyter/minimal-notebook:latest
ports:
- containerPort: 8888
```
Apply it:
`kubectl apply -f jupyter.yaml`
After fully deployed, check the logs:
`kubectl logs jupyter-notebook-pod-123`
Note down the URI and then do a port-forward:
`kubectl port-forward jupyter-notebook-pod-123 8888:8888 -n default`
This works, but the example not quite. You have to replace the hardcoded port-numbers with 8888 in the example.
Thanks
Answers:
username_1: Have you looked at the errors in the Javascript console? It would be interesting to understand if this is a Web/Browser issue or if this is something in the port forwarding itself.
The fact that it at least partially loads makes me believe that port forwarding is working correctly (or at least partially working)
username_0: Thanks for helping @username_1, I don't really get any special errors related as far as I can see.
This is a screenshot from the output:

This is a screenshot of the same running using kubectl port-forward:

There's some resources it can't find for both, but the first one times out while the second one loads fast.
I'm not sure if this has to do with multi-threading. In the code example it says: _"// Note this is single-threaded, it won't handle concurrent requests well..."_
I'm guessing, that like many other web-apps, this one too is doing various asynchronous calls. And perhaps the example code cannot handle that well? The next million-dollar question is of course how or what steps to take to enable it to do that (if that is the case) 😄 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.