repo_name
stringlengths
4
136
issue_id
stringlengths
5
10
text
stringlengths
37
4.84M
google/filament
585095231
Title: Alternatives to engine.flushAndWait() Question: username_0: I'm currently using Filament to render to shared textures and then sending the result off to another library that ultimately displays those textures on a screen. I'm calling `engine.flushAndWait()` to make sure the texture render has finished, but I would rather do it through some sort of callback. Is that an option in Filament? I guess this is related to fences, which I never really got my head around. Answers: username_1: A fence is what you want except that we don't support GPU fences anymore (in the public API that is). The current fences will only tell you that the render thread has executed all the commands up until the fence. username_2: This is not a supported use case right now. `flushAndWait` will work, however it is very heavy, it will stall both the current thread and the gl thread until the texture is rendered. Alas, this is the only solution right now. username_0: ☹️ thanks for the info anyway. username_2: You could also use the async version of readPixels maybe -- this would stall less.
go-swagger/go-swagger
249864485
Title: Custom template model generation .Package Question: username_0: ## Problem statement Hello, I am trying to create a custom layout for a project with templates. Everything is fine with *custom* operations ```target: "{{ if gt (len .Tags) 0 }}{{ joinFilePath .Package .ServerPackage .APIPackage .Package }}{{ else }}{{ joinFilePath .Package .ServerPackage .Package }}{{ end }}"``` However, I also want to move models to different folders too, I tried using this configuration: ``` target: "{{ joinFilePath .Package .ModelPackage }}" ``` It seems that `.Package` is not defined and everything is moved to `ModelPackage` - that is `models` folder. **Is there a way to define path like this:** `operation/models/model_name` ? ## Environment swagger version: dev go version: 1.8.3 OS: Macos Answers: username_1: how are you trying to do the custom layout? is that following this guide? https://goswagger.io/use/template_layout.html username_0: Yes. I pretty much copied the default config and changed `operations` to what I described. It works for `operations` (it finds and does not overwrite). It does not work with customizing `models`, because it lacks `.Package` property. Is that by design? Can it be added? username_1: I have no objection to somebody adding the functionality. I suspect the code to fix would be around here somewhere: https://github.com/go-swagger/go-swagger/blob/master/generator/shared.go#L390 username_0: I've got into code and found out that model movement is limited by the information that can be provided in `swagger.json` configuration file. Schema does not allow to add information (unless it's a hacky way) about model whereabouts (I assume, that models should be shared). But the #2 issue that i have, is that when I change the directory (with the template) to move operations to a specific folder, `imports` are not correct in `service/restapi/operations/service_api.go` file. They are imported in the default way `project_name/service/restapi/operations/x`. @username_1 do you have any ideas where can I find code for this part? username_1: what is it you're trying to achieve? Because if you just want to group operations, perhaps look into using tags in the swagger spec, it will do what you want username_0: I am trying to achieve this kind of structure: ``` servicex - operations - models ... swagger.json servicey - operations - models ... swagger.json global_swagger.json ``` But it seems I won't be able to do it because the `models` swagger spec part does not contain `Tags` and cannot be added. Because of this, after generating server from `global_swagger.json` you cannot use models from different services, but you can reference operations (operations in swagger spec has `Tags` prop). Status: Issue closed username_1: why not make more than 1 spec and then merge the specs together with the mixin command? username_0: As I stated above, I already do that successfully, however it causes 2 issues: 1) Models cannot be generated to different folders (.Package is not defined while generating with custom template (I assume that it is because `definitions` does not contain `Tags`)) (shown above) 2) The imports in `service/restapi/operations/service_api.go` will point to `default` location instead of the one in custom template, where I have to change auto generated file (which can be ignored by using `skip_exists: true` - but then missing the whole auto generated idea?) username_1: but you just generate 2 applications, is the issue that 1 is using types from the other? username_0: No. The goal is to merge 2 (or more) already established applications into 1 application. P.S. Thank you for answering so quickly username_1: yes but the question is do you need them to be because you want independent code files except that you want the server to present as 1 process with 1 spec. At least if I understand your use case correclty if not I apologize for jumping to conclusions. This you can do by creating your own main file and serving both API's. it's a lot easier to make a middleware that accommodates this pattern username_0: You are correct. It seems that having a middleware and main file might work. I will try doing this. Thank you.
manmal/hn-android
171966471
Title: Material design Question: username_0: This app uses a very custom design. I think it would be better to use the modern, clean and simple material design ;) Answers: username_1: Pretty sure there are materitrash designed apps available already if you want one. https://play.google.com/store/search?q=hacker%20news%20material I'm not this app's dev, and my opinion doesn't matter, but I specifically was interested in this app because it wasn't material. username_0: But it's the only one in f-droid ;) username_2: Hi! Thanks for this report. Unfortunately, currently there are no plans to migrate to a more materially designed version of the app. We wanted to refresh the design a bit (not really materially), but currently we do not have enough time to undertake larger changes in this project, so we just maintain the repository and do critical fixes (like the one last week). Also we are quite happy with the current design. Even though it is outdated, it seems that most of the users still like it. There are plenty of other HN apps and some of them are adopting the material guidelines well. Maybe you can ask the authors to open source them and/or publish them on f-droid. I hope this does not come as a big disappointment to you :) username_0: It's OK ;) Status: Issue closed
cloudhut/charts
1148650319
Title: Reference brokers in different clusters Question: username_0: Is there a way to reference brokers from different kafka clusters? I have 2 kafka clusters in 2 separate k8s clusters. Each cluster has its own set of ssl certificates for tls authentication. I see in the chart, I can only reference brokers and tls once. Status: Issue closed Answers: username_1: Hey, no multi-cluster support is not supported in Kowl. Therefore you are supposed to always setup one Kowl deployment per Kafka cluster.
theCrag/website
198690713
Title: Sort ascents by tick shift value in facet page Question: username_0: This is so we can link from a climbers card on a ranking page to the list of ascents for that particular person, in that particular node, and sorted by difficultly but in the order of the ascent worth the most after taking into consideration the tick shift. Answers: username_1: I have added two new sort by's in the facet so we can align to the rating list for an area * rating * all-time-rating Also I have added two more has assertions so we can get the right list of ascents: * rating * date If you want to link to my rating ascents in Arapiles it would look like: http://dev.thecrag.com/climbing/australia/arapiles/ascents/with-route-gear-style/trad/has/rating+ascent-date/by/username_1/?sortby=rating,desc And if you want to link to my all time rating: http://dev.thecrag.com/climbing/australia/arapiles/ascents/with-route-gear-style/trad/has/ascent-date/by/username_1/?sortby=all-time-rating,desc The facet UI is done for this issue. However @username_0 you may want to play with making this link from the climber card in the ranking lists. username_2: can we make sure they at least show CPR for branding purposes, e.g. "rating (CPR)" if we don’t want to call the climber performance rating because of the length? > username_1: @username_2 I changed the GUI display to show 'Performance rating' rather than 'Rating'. In this instance the rating is associated with an individual ascent, so use of 'Climber' would be incorrect. The Climber Performance Rating is the sum of individual ascent Performance Ratings. Please let me know if you still disagree. Actually the page we you should review with finer detail for branding purposes is http://dev.thecrag.com/climbing/australia/arapiles/climbers/using-stat/sport-rating/ ![screenshot from 2017-01-17 20-12-30](https://cloud.githubusercontent.com/assets/373058/22014729/3a8a9c4a-dcf3-11e6-93ba-796f8e30e483.png) This page shows Climber Performance Rating stats. username_0: @username_1 I'm partially blocked on this, I need more template data. eg something similar to the profile page data for one rating: ![image](https://cloud.githubusercontent.com/assets/187449/22320394/3f078ac8-e3e0-11e6-9d02-a7e298358525.png) We already have the rank and score, but I also want for each climber: * [ ] the grade the rating converts to (scoreGrade ?) * [ ] which grade system (scoreGradeSystem ?) * [ ] it would be nice to have the total ascents use to calculate this score. This isn't as important so leave off if not easy to add (score total ascents) ![image](https://cloud.githubusercontent.com/assets/187449/22320593/ad9a7616-e3e1-11e6-9c50-790b19a98885.png) username_0: Ok so we are now linking from the rankings page into the ascents for that person filtered the right way. eg here: https://dev.thecrag.com/climbing/australia/arapiles/climbers/using-stat/trad-rating/ ![image](https://cloud.githubusercontent.com/assets/187449/22323335/68121fda-e3f6-11e6-81a9-b00cc66a3b43.png) we link into: https://dev.thecrag.com/climbing/australia/arapiles/ascents/with-route-gear-style/trad/by/adam_demmert/?sortby=rating,desc @theCrag/translators note I just added a new lang pack for this: "template.card.climber.rating-ascents-count": "Ascents for rating", @username_1 I haven't used the 'has/rating+ascent-date/' - this seems mostly redundant to me, ie if the ascent doesn't have a date then wouldn't it's rating default to 1900 or something and so have a near zero score. And for the all time rating we'd want to include it anyway. So the output as-is almost seems like a feature to me because it exposes the bad data, eg when you scroll down on your list you can clearly see the cutover point and that the older ascents without dates are worth less. Happy to go either way if others think this is weird ![image](https://cloud.githubusercontent.com/assets/187449/22323504/79c8a018-e3f7-11e6-8525-3f68cee20d11.png) username_0: I'd still like to add the grade conversion, but I think if pushed we could split that off and release as-is. username_1: This is going to sound pathetic, but it is a little more difficult that I have energy to do right now. We know the default context so I have got to write a query that works out the technical grade that should be shown for that context. It is based on RatingSystem.CPRPriorityOrder I will do this after the release when I resolve the profile page context, which is the reverse process. Status: Issue closed username_0: Ok I'm cool with all of that so I'm closing this and I've split off #2552 for later
jupyterhub/traefik-proxy
398235160
Title: Test with the littlest jupyterhub Question: username_0: Perhaps before we go all the way to kubernetes in #26, we should try using this in [the littlest jupyterhub](https://the-littlest-jupyterhub.readthedocs.io/en/latest/). TLJH is a small, single-machine JupyterHub installation that already includes traefik as an edge proxy (pretty much just for letsencrypt support), in addition to configurable-http-proxy. We could use this traefik-proxy to eliminate the extra configurable-http-proxy service and use only traefik. What it will look like: 1. remove configurable-http-proxy 2. traefik is already launched by systemd (not by `TraefikProxy.start()` here, so no 'static configuration' would be loaded from the code in this repo), so static config would have to be created in tljh. 3. probably use the toml implementation in #31. Alternately, we would need to add and run an etcd service Answers: username_1: @username_0, I've added new commits addressing your comments in the PR opened on tljh, if you want to take a look. username_1: @username_0, there's the ```test_long_username``` that fails intermittently in tljh. From the logs it seems that the timeout to wait for a route (20s currently) ends before the route is actually added. Assuming is not something else, I believe it would be good to increase this timeout since we are dealing with I/O operations. Do you agree? username_0: Yeah, go ahead and increase the timeout on the test to see if it helps. username_1: @username_0, thanks a lot for solving the issues with the credentials. Everything seems to work now :confetti_ball: Status: Issue closed
Argonne-National-Laboratory/DSP
644936227
Title: length in line 479 and 453 are defined twice Question: username_0: "length" in line 479 and 453 are defined twice, which I think is unnecessary. https://github.com/Argonne-National-Laboratory/DSP/blob/8a659f3c3ed29c653d69b40f83dc5a772d65ee37/src/Model/TssModel.cpp#L479 Answers: username_0: We can remove line 479. https://github.com/Argonne-National-Laboratory/DSP/blob/8a659f3c3ed29c653d69b40f83dc5a772d65ee37/src/Model/TssModel.cpp#L479 Status: Issue closed username_0: This has been resolved in https://github.com/Argonne-National-Laboratory/DSP/commit/6dc94dd64c25d1c6dc3d6a3c31c465b18294b046.
spencermountain/compromise
1077986128
Title: Lexicon terms beginning with # do not get matched Question: username_0: I have a custom lexicon that links hashtags to tags. This worked in an older version of compromise (v11.14) but does not seem to be working in 13.11.4. ```js // I would expect this to match, but it does not nlp("#GoJetsGo", { "#GoJetsGo": "SportsTeam" }).match("#SportsTeam").text() // '' // this can work, however I would like to make sure only the hashtag is matched nlp("#GoJetsGo GoJetsGo", { "GoJetsGo": "SportsTeam" }).match("#SportsTeam").text() // '#GoJetsGo GoJetsGo' // if # is not the leading character it does work, so seems to only happen when it's leading nlp("Go#JetsGo", { "Go#JetsGo": "SportsTeam" }).match("#SportsTeam").text() // '#Go#JetsGo' ``` This seems like it may be intentional (perhaps the built-in hashTag logic is conflicting?), but I'm having trouble finding anything in the docs that would say so. Answers: username_1: hey @username_0 thanks - this is a good issue. You're right, something is bad. It's started tripping on the TitleCase bit, after the pound symbol. This is bad. I removed an 'i' from a regex a few versions back, and didn't have a test for it. Sorry! here's what i'd do, right now: https://runkit.com/username_1/61b7bb4bd4140000092a6925 I will add a proper fix to v14, which will ship in January. I have been thinking about cleaning this stuff up, the timing is good. Will keep this open, until then. cheers (sorry bout the jets this year) username_0: No problem, I can wait until v14. I appreciate the quick response and the great work you've been doing!
lyucean/spaced_repetition_bot
636979244
Title: [Airbrake] [Production] Typed property srbot\command\Content::$message_id must be int, null used Question: username_0: **Airbrake error:** [#4979](https://username_0-gmail-com.airbrake.io/projects/276279/groups/2764676685674804979) **Airbrake project:** SRB **Error type:** `TypeError` **Error message:** `` Typed property srbot\command\Content::$message_id must be int, null used `` **Where:** `<no information>` **Occurred at:** Jun 11, 2020 12:39:02 UTC **First seen at:** Jun 11, 2020 12:39:02 UTC **Occurrences:** 0 (0 since last deploy on `<no information>`) **Severity:** `error` **URL:** `<no information>` **File:** `/var/www/bot.abxtest.com/command/Content.php` **Backtrace:** ``` /var/www/bot.abxtest.com/command/Content.php:22:in srbot\command\Content->__construct /var/www/bot.abxtest.com/model/Schedule.php:20:in srbot\model\Schedule->check /var/www/bot.abxtest.com/demon.php:9:in ```<issue_closed> Status: Issue closed
hamdle/Drupal_Quest
329176097
Title: Flying Enemy Question: username_0: Implement the flying enemy into the game. This has been reduced down to do just the follow: fly in a figure-eight pattern, jump on top to kill, other hit directions cause damage(kill). Removed: Shooting/dropping a projectile that tracks the player. Bonus Add: Can interact with other enemies, if a bug falls on the top of a flying enemy he takes damage(dies).<issue_closed> Status: Issue closed
squidfunk/mkdocs-material
602202744
Title: Allow arbitrary URI in "socials" Question: username_0: {% include ".icons/" ~ social.icon ~ ".svg" %} </a> {% endfor %} ``` Answers: username_1: @username_2 that's why I suggested a `text` option which could simplify and improve things here if present. username_2: It wouldn't solve the issue. username_2: db7c28b2 adds a `name` attribute to each social link. Furthermore, the domain is now only set when `//` is a substring of the URL, as you suggested. Status: Issue closed username_2: Released as part of 5.1.5. username_0: Working great now. Thanks :smile:
ErikEJ/SqlCeToolbox
219506372
Title: Migrating large SQL DB to SQLite just stops. Question: username_0: Hi @username_1 it's me again :) I have tested to migrate very small SQL DB before and it was working perfect. Now I try to migrate large SQL DB to SQLite in SSMS and Visual studio and the task just stops after "Creating SQLite database... Starting import Scripting server database..." The SQLite DB file is 0kb. Is there any size limitations or something? Answers: username_1: No size limitations, but the process could take a long time for a large database.. - Check your %temp% folder for generated .sql files, to verify if the export is working.. - How large is large? - No errors?? username_0: Ok, now I see that the SQLite file is growing. The db table that I am migrating from has 350 000 rows. If it is possible, where can I see total progress or some indication that it is running? username_1: You could also try the script to file option, and the run the generated scripts using SQLite3.exe to see if any errors occur. username_1: You should see Importing data...{0} in the status bar for each imported 18 MB file username_1: Check you %temp% folder to see how many .sql files were generated username_0: Windows temp folder? Becouse there is no .sql files there. username_1: In %temp% look for the newest files, extension is .sqlce, not .sql, sorry. Status: Issue closed username_1: Closing, as the import is actually running... I have logged: https://github.com/username_1/SqlCeToolbox/issues/334 username_1: I have added improved feedback (progress bar) in the latest daily build - thanks for your valuable feedback!
CharsetMC/Charset
301268052
Title: Unable to make custom barrels with non-vanilla blocks via CraftTweaker Question: username_0: Recipes can be made with CT for barrels with custom textures, however it seems to only work for vanilla blocks. this line works: `recipes.addShapeless(<charset:barrel>.withTag({log: "minecraft:stone;0", slab: "minecraft:stone;0"}),[<minecraft:stone>]);` but this line will output a default oak barrel with the defined tag: `recipes.addShapeless(<charset:barrel>.withTag({log: "quark:iron_plate;0", slab: "quark:iron_plate;0"}),[<quark:iron_plate>]);` Answers: username_1: Try a non-vanilla _block_ in the tag, as the barrel reuses the block's textures - which won't work for items. username_0: The iron plate from quark is a block. username_1: Huh. Sorry, was thinking of IC2. I have a theory - Charset has a builtin "item material" system, and - as the Iron Plate is never registered to it - it might not be willing to use it. This might be a tougher issue to solve than I initially thought - I'll get back to you on that, but it will take some time. username_0: no worries. username_1: ![](https://img.asie.pl/m7c1) The stopgap solution? is implementing hooks for the material registry to Charset's CraftTweaker integration. mods.charset.MaterialRegistry.registerTypes(<quark:iron_plate>, "plate", "block", "iron"); If you need tips on what to use for each block as types, set enableDebugInfo=true in module/lib.cfg and consult the then-created charsetItemMaterials.txt - or study the source code. username_1: It's not pushed yet, though, so it won't work just now - it should be out within 24 hours, however. username_1: I guess it's solved. Status: Issue closed
hasadna/Open-Knesset
46836098
Title: Support a Person's calendar Question: username_0: Folliwng discussion with <NAME>, we want to offer the MKs the ability to publish their shared calendar in Open Knesset. We can assume the MKs are using google calendar and the URL of the calendar will be entered through the admin interface. We need to store the URL and display the calender on the MKs page. <!--- @huboard:{"order":279.0,"milestone_order":7.9375,"custom_state":""} --> Answers: username_1: @OriHoch Could you create api for the mk's calendar similar to this : http://data.obudget.org/queries/528/source#table I could use a filter with date range and mk_id. It will be better to return all the relevant data for each mk in one row. Thanks, Alon fyi @morchickit
othreecodes/Zapier-Twitter-Jobs
350067368
Title: We're #hiring! Read about our latest #job opening here: Front End Developer - https://t.co/HTE5QDeQ4l #WebDesign #Nanuet, NY #CareerArc Question: username_0: We're #hiring! Read about our latest #job opening here: Front End Developer - https://t.co/HTE5QDeQ4l #WebDesign #Nanuet, NY #CareerArc <a href="https://www.careerarc.com" rel="nofollow">CareerArc 2.0</a> https://twitter.com/564183679/status/1029019112226222080 Mon Aug 13 14:57:29 +0000 2018
GoogleChrome/lighthouse
404276892
Title: DevTools Error: Failed to construct 'URL': Invalid URL Question: username_0: **Initial URL**: https://www.magazineluiza.com.br/ **Chrome Version**: 71.0.3578.98 **Error Message**: Failed to construct 'URL': Invalid URL **Stack Trace**: ``` TypeError: Failed to construct 'URL': Invalid URL at new URLShim (chrome-devtools://devtools/remote/serve_file/@15234034d19b85dcd9a03b164ae89d04145d8368/audits2_worker/audits2_worker_module.js:4575:1) at NetworkRequest.onRequestWillBeSent (chrome-devtools://devtools/remote/serve_file/@15234034d19b85dcd9a03b164ae89d04145d8368/audits2_worker/audits2_worker_module.js:1421:67) at NetworkRecorder.onRequestWillBeSent (chrome-devtools://devtools/remote/serve_file/@15234034d19b85dcd9a03b164ae89d04145d8368/audits2_worker/audits2_worker_module.js:1399:151) at NetworkRecorder.dispatch (chrome-devtools://devtools/remote/serve_file/@15234034d19b85dcd9a03b164ae89d04145d8368/audits2_worker/audits2_worker_module.js:1409:66) at devtoolsLog.forEach.message (chrome-devtools://devtools/remote/serve_file/@15234034d19b85dcd9a03b164ae89d04145d8368/audits2_worker/audits2_worker_module.js:1412:126) at Array.forEach (<anonymous>) at Function.recordsFromLogs (chrome-devtools://devtools/remote/serve_file/@15234034d19b85dcd9a03b164ae89d04145d8368/audits2_worker/audits2_worker_module.js:1412:93) at Function.afterPass (chrome-devtools://devtools/remote/serve_file/@15234034d19b85dcd9a03b164ae89d04145d8368/audits2_worker/audits2_worker_module.js:1115:163) ``` Answers: username_1: Duplicate https://github.com/GoogleChrome/lighthouse/issues/6145, fixed in latest LH and shipping in upcoming Chrome. Status: Issue closed
Clancey/simple_auth
384490908
Title: Unable to fetch cached account from Dropbox provider Question: username_0: I have been trying to get this working with Dropbox but I keep running into problems. The one I am stuck on now is that after I have authenticated and allowed access to my app - I keep getting prompted to allow access whenever I make an API call. I believe the problem is with `loadAccountFromCache()`. The method is able to retrieve the credentials from authStorage but it fails when it gets to `getAccountFromMap<T>(data)`. Here is the error I see: ``` I/flutter (13160): NoSuchMethodError: The getter 'iterator' was called on null. I/flutter (13160): Receiver: null I/flutter (13160): Tried calling: iterator ``` I will continue to hunt for why this is happening but I am struggling to understand why it gets tripped up here. I tested it using Google as the provider just to make sure my workflow wasn't missing a step. Answers: username_0: Okay, I think the problem is with the dropbox provider overriding the `getAccountFromAuthCode` method without setting the `scope`. I just set it to an empty list which seems to fix it for me. I can submit another PR unless you think there is a better way to handle this somewhere else. username_1: It requires scopes even if it's empty? There is already a scope property. Right now it is only passed in if the scopes are not empty. For Dropbox just overriding the method and forcing it to use the scopes should be good. Sorry for this. I spent most my time on the core lib. I quickly ported over the providers from my c# version. For the most part they work perfectly. Status: Issue closed
Masuzu/GBFPokerBot
226227904
Title: Trying to renew licence Question: username_0: Hey i tried renewing the licence by paying through card twice. Both times the thing loaded and came up with a green tick, then it just closes and nothing happened. Ive tried once 9 hours ago and now and both times it did not work. Answers: username_1: have tired with difference browser yet? username_2: Hi, You can try the Paypal option which also supports credit cards. No feeds are charged if the transaction failed. Status: Issue closed
infinitered/reactotron
261510125
Title: Convert stateless components to functional component Question: username_0: While working on pull request yesterday for migrating to prop types packages I have noticed that there are few components which are stateless and doesn't use lifecycle hooks. I want to propose that can they be changed to functional component instead. I know its not a big deal but I could submit a PR for that if you like under hacktoberfest tag. Answers: username_1: Absolutely. I learned how to do React JS on this project so it's a bit messy. username_0: sure I will start this weekend username_2: @username_0 @username_1 Would love to do some of these, mind splitting them? username_0: I have submitted the PR for the files which were straight forward to convert to stateless function component. I hope @username_1 will check them if it works for him. @username_2 sure i could definitely use some help. i have not converted the following files to stateless components as they require a little bit more of testing. - Commands/AsyncStorageValuesCommand - Commands/ApiResponseCommand - Commands/SagaTaskCompleteCommand - All files under `Dialogs/` - Foundation/Backups - Foundation/Footer - Foundation/BakupsHeader - Foundation/HelpFeedback - Foundation/HelpHeader - Foundation/NativeHeader - Foundation/NativeOverlayMargins - Foundation/Sidebar - Foundation/Subscriptions - Foundation/SubscriptionHeader - Foundation/TimelineHeader - Foundation/VisualRoot - Shared/CommandToolbar There may be other which i must have missed. For example to remove mobx decorators from classes we can use HOC. i suggest you check out with the docs. username_0: @username_2 please see effected files in the PR to avoid merge conflict username_2: @username_0 I will wait for that PR to be merged before I continue, that way we can completely avoid merge conflicts username_0: sure sounds like a good idea username_2: @username_0 @username_1 just added a new PR updating only the `App/Dialog` so to avoid merge conflicts :) https://github.com/infinitered/reactotron/pull/522 Status: Issue closed username_0: While working on pull request yesterday for migrating to prop types packages I have noticed that there are few components which are stateless and doesn't use lifecycle hooks. I want to propose that can they be changed to functional component instead. I know its not a big deal but I could submit a PR for that if you like under hacktoberfest tag. username_0: @username_2 i will keep this issue open for further discussion of the work you are doing. username_1: I'm happy to close this too if you're done for now. I wanna make sure you get the `hacktoberfest` credits and it's big undertaking to convert them all. username_0: @username_1 sure please go ahead Status: Issue closed username_1: Thank you very much again! username_2: @username_1 I have a friend I work with who is also interested in converting some of the components, I will also continue to work through them and put up PRs when time permits!
business-science/timetk
1079734646
Title: Error in plot_acf_diagnostics() Question: username_0: Hi Matt, I think I found a little bug when I run `plot_acf_diagnostics()` ``` Error in str_c("CCF_", .) : could not find function "str_c" ``` I think the issue is that you missed `stringr::str_c` [here](https://github.com/business-science/timetk/blob/693c329c61e906d7c26bbe0194fa09a956c2a35e/R/diagnostics-tk_acf_diagnostics.R#L194). If you approve I can send a PR fixing this. I'm your student in your time series course and it is really amazing. I'm learning a lot... Thanks :) Answers: username_1: Great find! Sure thing on the PR. Just put it together and I'll approve. And thanks for joining the Time Series Course. I love that course. username_1: Closing this. Should be fixed now. Status: Issue closed
divan/expvarmon
113783321
Title: go get fails - errors Question: username_0: Hi, can you help me with this... "go get github.com/username_1/expvarmon" fails. I am attaching the output of failed installation. http://pastebin.com/cTEyw3y1 Thanks, Tomaz Answers: username_1: Oh, thanks for reporting. It's TermUI changed it's API. Going to fix it. username_1: Should be ok now with pinned dependency via gopkg.in. https://github.com/username_1/expvarmon/commit/383ee494c1dee0a8062137d81e08c6b18c80af13 Status: Issue closed
TomLippincott/haskseg
499107402
Title: Wrong github account listed on Hackage Question: username_0: Hi, Your `package.yaml` file lists https://github.com/githubuser/haskseg#readme as the home page for the project, making it a bit tricky to find your way here :slightly_smiling_face: . Answers: username_1: Thanks! Fixed. Status: Issue closed
wso2/product-apim
756820778
Title: Displaying comments with additional User Information. Question: username_0: ### Improvements Response from Comment REST API can be improved to include additional details of First Name, Last name and Full name. Currently it is only displaying the username in `createdBy` field. **Request :** ` curl -k "https://localhost:9443/api/am/store/v1/apis/e93fb282-b456-48fc-8981-003fb89086ae/comments"` **Response :** ``` { "count": 2, "list": [ { "id": "6315a127-4eb5-4642-953a-1c7ae20ccac8", "content": "This is a comment from admin", "createdTime": 1606821504945, "createdBy": "admin", }, { "id": "79f91673-69e8-480d-9aa2-2094390cf02d", "content": "This is a comment from user saran", "createdTime": 1606821534430, "createdBy": "saran", } ], "pagination": null } ``` ### Approach We can use an optional query parameter property when retrieving comments. This value will be by-default `false`. When this query parameter is provided we can display the additional meta information in the response. **Request :** `curl -k "https://localhost:9443/api/am/store/v1/apis/e93fb282-b456-48fc-8981-003fb89086ae/comments?<param>=true"` **Response:** ``` { "id": "6315a127-4eb5-4642-953a-1c7ae20ccac8", "content": "This is a comment from admin", "createdTime": 1606821504945, "createdBy": "admin", "commenterInfo": { "firstName": "Admin", "LastName": "Admin", "fullName": "Admin Admin" } } ``` ### Affected Product Version: APIM 4.x.x ### Environment details (with versions): OS: Ubuntu 18.04.1 LTS ### Suggested Labels: 4.x.x<issue_closed> Status: Issue closed
LeftistTachyon/GameAI
299134250
Title: ChessAI: can't keep track of King Question: username_0: Exception in thread "main" java.lang.ClassCastException: chess.Knight cannot be cast to chess.King at chess.ChessBoard.pieceFromTo(ChessBoard.java:108) at common.Board.movePiece(Board.java:303) at common.Board.movePiece(Board.java:286) at common.Board.movePiece(Board.java:367) at common.TreeNode.simulate(TreeNode.java:168) at common.TreeNode.selectAction(TreeNode.java:90) at common.GameAIMain.main(GameAIMain.java:20) Answers: username_0: This happens when you eat the king. I think this is a problem with king moves after check Status: Issue closed username_0: Exception in thread "main" java.lang.ClassCastException: chess.Knight cannot be cast to chess.King at chess.ChessBoard.pieceFromTo(ChessBoard.java:108) at common.Board.movePiece(Board.java:303) at common.Board.movePiece(Board.java:286) at common.Board.movePiece(Board.java:367) at common.TreeNode.simulate(TreeNode.java:168) at common.TreeNode.selectAction(TreeNode.java:90) at common.GameAIMain.main(GameAIMain.java:20)
ampproject/amphtml
849708579
Title: Cross-Origin-Resource-Policy "cross-origin" not present in v0.js Question: username_0: Unable to make my website "cross-origin isolated" using COOP and COEP. https://web.dev/coop-coep/ Resource: https://cdn.ampproject.org/v0.js Error: To use this resource from a different origin, the server needs to specify a cross-origin resource policy in the response headers. Missing response header: Cross-Origin-Resource-Policy: cross-origin Answers: username_1: Hi @username_0: Thanks for this report. I'm going to close this as a duplicate of https://github.com/ampproject/amphtml/issues/31579, only because that one was opened first. This should be easy to fix, and I'll provide updates in that issue. Status: Issue closed username_1: This is fixed on our servers.
nwfsc-fram/boatnet
543981206
Title: Fix Build Break Question: username_0: Currently getting the error (Jenkins only?) globals.d.ts(139,11): error TS2320: Interface 'NodeRequire' cannot simultaneously extend types 'Require' and 'RequireFunction'. Named property 'cache' of types 'Require' and 'RequireFunction' are not identical. Investigating. * Updated Dockerfiles to use node:12-alpine * Added @types/node@12 Answers: username_0: Attempting to repro locally, above fixes didn't remedy the issue username_0: Reproduced locally by deleting node_modules and yarn.lock. Investigating further... username_0: I couldn't find the root cause of this, but it happened around 12/23. Luckily it's limited to dev-auth-server, which we don't need to build anyway (we just run yarn serve, so this is OK.) Status: Issue closed
blastrock/pkgj
566010041
Title: Subfolders are ignored Question: username_0: Hardware: PSV Version of firmware: 3.60 Version of PKGj: 0.54 ## Steps to reproduce the issue I am using category lite PSP/Adrenaline plugin to have my PSP/PS1 games in subfolders (so I moved to folders the games I installed before that plugin). Now I want to install DLCs for those PSP games so I select "Show PSP DLCs" and checked on the filter "Installed games only" present in pkgj. Obtained result: I just get a full list of DLCs (filter has been ignored). Expected result: What I expect is to get a list of DLCs for the PSP games I have installed on my console exclusively. Extra info: Another related issue is the expected white circle on some PSP/PS1 games (those installed and moved to subfolders). Please, excuse my bad english language. Thanks in advance for your hard work. Answers: username_1: There has been previous requests for features used in conjunction with gclite (#148, #174). While not out of consideration, it's low priority.
jw3126/UnitfulRecipes.jl
488978643
Title: reg v0.1.1 Question: username_0: `@username_1 register()` Answers: username_1: Registration pull request created: [JuliaRegistries/General/3238](https://github.com/JuliaRegistries/General/pull/3238) After the above pull request is merged, it is recommended that a tag is created on this repository for the registered package version. This will be done automatically if [Julia TagBot](https://github.com/apps/julia-tagbot) is installed, or can be done manually through the github interface, or via: ``` git tag -a v0.1.1 -m "<description of version>" d7d1c530c2b5a60a7b0acad23c25445324d45dbb git push origin v0.1.1 ``` Status: Issue closed
jlippold/tweakCompatible
370353633
Title: `Simple Power Down` not working on iOS 11.3.1 Question: username_0: ``` { "packageId": "com.pineapple.simplepowerdown", "action": "notworking", "userInfo": { "arch32": false, "packageId": "com.pineapple.simplepowerdown", "deviceId": "iPhone10,6", "url": "http://cydia.saurik.com/package/com.pineapple.simplepowerdown/", "iOSVersion": "11.3.1", "packageVersionIndexed": true, "packageName": "Simple Power Down", "category": "Tweaks", "repository": "BigBoss", "name": "Simple Power Down", "installed": "", "packageIndexed": true, "packageStatusExplaination": "This package version has been marked as Not working based on feedback from users in the community. The current positive rating is 0% with 0 working reports.", "id": "com.pineapple.simplepowerdown", "commercial": false, "packageInstalled": false, "tweakCompatVersion": "0.1.0", "shortDescription": "A simple Power Down menu", "latest": "1.0.1", "author": "<NAME>", "packageStatus": "Not working" }, "base64": "<KEY> "chosenStatus": "not working", "notes": "" } ```
Ryochan7/DS4Windows
310266060
Title: [Request] support for nacon Wired Compact Controller Question: username_0: Hello I saw the help from @username_2 and I wanted to ask if its possible to port it to the wired compact controller version. VID/PID: HID\VID_146B&PID_0603&IG_01 [I copied it from hardware manager Idk if its Right....] I hope this is enough to help because the most is already in there and the controller is like a Lite version of the pro it has no gyros in it, no audio and only X-Input greetings marv sry for my bad english Answers: username_1: Just FYI, this situation (not device) is similar to #230 username_0: I would be grateful if you could buy one and figure it out @username_2 username_2: I've ordered one from Amazon and it should be here mid next week. username_2: It's arrived but it doesn't look good I'm afraid. When plugged into a PC it appears as an XInput controller rather than a DInput one and short of using a USB analyser with the controller plugged into a PS4 I'm not sure how I'd figure out a way around that - unfortunately I don't have one. username_3: Thanks for trying I've hit a brick wall trying to make this work, gave in and bought a real ps4 pad as this is properly detected and works fully in DS4 with farcry 5 including the dpad and both triggers. Don't suppose this would be any use https://freeusbanalyzer.com/ or would you need to analyse it in use on a ps4 to see what it was outputting? Status: Issue closed
AuthEceSoftEng/cenote
694973305
Title: Local dev environment failing to write data to cockroach Question: username_0: After experimenting with a number of topologies (changed the code for cockroach handler) the data are not being written at all in cockroach db any more. Tried submitting the old topology (guaranteed to work) as well as replace the kafka and storm distros with new, clean ones, but the issue persists. Need to get deep in the logs to find out what the issue really is, but so far no luck.. Answers: username_0: After consulting @iamnapo about the issue and reviewing the various logs from kafka, storm, zookeeper etc, I could not yet find the cause of the issue. After printing logs in a file, it seems that kafka receives the messages in cenoteIncoming topic, however storm is not able to receive them and process them. ![Screenshot_4](https://user-images.githubusercontent.com/8168416/92473723-a79a5100-f1e3-11ea-857f-da108f1e1904.png) ![Screenshot_5](https://user-images.githubusercontent.com/8168416/92473725-a832e780-f1e3-11ea-8abf-c474b2f1f417.png) ![Screenshot_2](https://user-images.githubusercontent.com/8168416/92473718-a6692400-f1e3-11ea-85c1-b55a35aa8167.png) ![Screenshot_3](https://user-images.githubusercontent.com/8168416/92473719-a701ba80-f1e3-11ea-8e64-027134262edf.png) ![Screenshot_1](https://user-images.githubusercontent.com/8168416/92473714-a5d08d80-f1e3-11ea-9203-ae37b4cc12ee.png) Status: Issue closed username_0: The issue was the values for the bootstrap servers for kafka consumers. These values are retrieved from the dotenv inside cenote-write/src/main/java/com/issel/cenote/WriteTopology.java The right values for local/dev environment should be 'localhost:9092' whereas from the worker logs seems not to be the case (idk why): ``` 2020-09-08 13:34:50.591 o.a.s.u.Utils Thread-19-kafka-spout-executor[18, 18] [ERROR] Async loop died! org.apache.kafka.common.errors.TimeoutException: Timeout expired while fetching topic metadata5:04 PM2020-09-08 13:36:13.948 o.a.k.c.c.ConsumerConfig Thread-19-kafka-spout-executor[18, 18] [INFO] ConsumerConfig values: auto.commit.interval.ms = 5000 auto.offset.reset = earliest bootstrap.servers = [192.168.3.11:9092, 192.168.127.12:9092, 192.168.3.11:9092] ``` Changed the default values inside the java file to be "localhost:9092" and it worked. ![Screenshot_3](https://user-images.githubusercontent.com/8168416/92488237-f3a2c100-f1f6-11ea-8d96-89d5321daf3d.png)
go-pg/pg
180056411
Title: Upgrading from v3 to v5: how to mimic `SSL: true` behaviour with TLSConfig Question: username_0: During the upgrade I noticed that the SSL option has been deprecated in favour of the TLSConfig, I was wondering if you had a sample config that would mimic the now removed feature. Answers: username_1: Previous code used following config: ``` &tls.Config{ InsecureSkipVerify: true, } ``` See https://github.com/go-pg/pg/pull/284/files and linked issue Status: Issue closed
MicrosoftDocs/OfficeDocs-Enterprise
500039157
Title: Step 1 fails; Install-Module doesn't exist on Server 2012R2 Question: username_0: The instructions for Step 1 fail on a brand new install of Server 2012R2; I applied all Windows patches. Installation media was from MSDN, "en_windows_server_2012_r2_with_update_x64_dvd_6052708.iso" I got this: Install-Module : The term 'Install-Module' is not recognized as the name of a cmdlet, function, script file, or operable program. Check the spelling of the name, or if a path was included, verify that the path is correct and try again. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: dbb1fe83-c1b3-3176-4d61-fe4f3ce1d598 * Version Independent ID: c89f295b-4f00-c2c0-c75f-75fa50ecda65 * Content: [Connect to Office 365 PowerShell](https://docs.microsoft.com/en-us/office365/enterprise/powershell/connect-to-office-365-powershell) * Content Source: [Enterprise/powershell/connect-to-office-365-powershell.md](https://github.com/MicrosoftDocs/OfficeDocs-Enterprise/blob/live/Enterprise/powershell/connect-to-office-365-powershell.md) * Service: **o365-administration** * GitHub Login: @username_2 * Microsoft Alias: **josephd** Answers: username_0: I tried the troubleshooting steps from an elevated PowerShell script and got the following error: PS C:\Windows\system32> (Get-Item C:\Windows\System32\WindowsPowerShell\v1.0\Modules\MSOnline\Microsoft.Online.Administr ation.Automation.PSModule.dll).VersionInfo.FileVersion Get-Item : Cannot find path 'C:\Windows\System32\WindowsPowerShell\v1.0\Modules\MSOnline\Microsoft.Online.Administration.Automation.PSModule.dll' because it does not exist. I checked the version of powershell (again, fresh install of 2012R2 from patched media, and all windows patches) PS C:\Windows\system32> Get-Host | Select-Object Version Version ------- 4.0 username_1: @username_0 - Thank you for submitting feedback. I will get this issue over to the Enterprise writing team for investigation. Thank you for reporting and making the docs better. Much appreciated. I made a note to request the team to update this when the work is complete. username_2: @username_0 As per https://docs.microsoft.com/en-us/powershell/scripting/install/installing-windows-powershell?view=powershell-6, please install PowerShell 5.1 from https://www.microsoft.com/en-us/download/details.aspx?id=54616. The try installing the AzureAD module and let use know if it is successful. Thanks username_0: @username_2 That isn't listed as a prerequisite on the referring page (Enterprise/powershell/connect-to-office-365-powershell.md). I followed all of the prereq's and it failed. If Powershell 5.1 is required, it needs to be listed on this page and a step added, before the user is directed to perform an activity that fails. That is the bug. username_0: @username_2 Whoops, I take that back. Server 2012R2 does have that version installed. ``` Windows PowerShell Copyright (C) 2016 Microsoft Corporation. All rights reserved. PS C:\Users\username_0> $PSVersionTable Name Value ---- ----- PSVersion 5.1.14393.3053 PSEdition Desktop PSCompatibleVersions {1.0, 2.0, 3.0, 4.0...} BuildVersion 10.0.14393.3053 CLRVersion 4.0.30319.42000 WSManStackVersion 3.0 PSRemotingProtocolVersion 2.3 SerializationVersion 1.1.0.1 PS C:\Users\username_0> ``` Wouldn't Microsoft double check documentation by doing a fresh install of the products being documented? Status: Issue closed username_2: @username_0 I just tested this on a fresh Windows Server 2012 R2 virtual machine. I added a note in the "What do you need to know before you begin?" section to inform readers that they have to have PowerShell 5.1 or later installed and for listed, older versions of Windows, a link to the 5.1 download. Sorry for the inconvenience. Thanks so much for letting us know.
kasper/phoenix
166861185
Title: Some elements may fail to complete actions with a strict timeout Question: username_0: Some elements (at least windows) may fail to complete actions with a strict timeout (-25200). We could dynamically increase the timeout for these `PHAXUIElement`s and try again. - Version: 2.2.1 - macOS: 10.11.6<issue_closed> Status: Issue closed
networkx/networkx
845012307
Title: Update documentation for geographical_threshold_graph Question: username_0: The documentation state that the return value of `p_dist` must be in the range `[0, 1]`, but the implementation differs from this behavior. The documentation should be updated to remove this discrepancy and better reflect the current implementation. This was originally raised on the mailing list (thanks jimbo): see the [original post](https://groups.google.com/g/networkx-discuss/c/Z_UXaYZcsxw/m/2c377udtAgAJ) for further details. Answers: username_1: Related to the documentation issue there is some counter intuitive behavior with the generator. For a given theta and given node weights, if instead of using 1/r^2 as p_dist, we use 1/r^3, we end up with a lot more edges. <img width="1080" alt="Screen Shot 2021-03-31 at 8 54 21 AM" src="https://user-images.githubusercontent.com/49379192/113147910-45e11900-91ff-11eb-837f-36ade8b9451c.png"> Counter intuitive doesn't mean wrong but I'm not sure this behavior was intended in the implementation. Do you think it may be worth it to try and make the behavior more in line with expectations? username_2: Why do you say unexpected? When you change the p_dist form, you are going to have to adjust the threshold (unless you adjust it implicitly in your p_dist function). I wouldn't know what to expect unless I thought it through. In this case, 1/r**3 is going to be much bigger than 1/r**2 because r is the metric on nodes placed within the unit square. So I would expect many more edges unless theta is similarly divided by some sort of average value for r. Similarly, if you rescale the metric to use cm instead of m, you would have to rescale theta to use the same units if you want to have the same number of edges. I'm guessing an average value of 1/r is about 2, so either use 0.5/r**3 or set theta to 200. That's a guess though -- more complete examination of average 1/r would tell you more. username_1: Yes, of course you are correct. I say unexpected only because normally when I think of going from 1/r^2 to 1/r^3 I think of an effect getting weaker, but here we have the opposite and end up with a lot more connections. Also some of the points within the unit square are separated by more than 1 (consider two points connected by a diagonal across the square), so 1/r^3 is not always going to be bigger than 1/r^2, which adds slightly to the confusion. At any rate, I think it's a great idea to update the docs as described in the original comment on the issue. username_1: Hi guys. I'd like to take a crack at updating these docs over the weekend and hopefully getting this issue closed. Noob question: Just updating the docstrings (in my own forked version and then creating a PR from there) is sufficient, right? username_0: Yes, this is correct. I'd recommend making a new branch on your fork for the work as well, e.g. `git checkout -b doc/update-p_dist-description` or whatever you'd like to name the branch. Then you can push that branch up to your fork and create a PR from there. Status: Issue closed
Wynncraft/Issues
66493697
Title: Guild attacks not working Question: username_0: I don't have any screen shots for this but my guild hasn't been able to attack recently because even though all of us are there, it says that no one is. I don't know if we are doing it wrong but I don't think so because we used to be able to do it fine! Also on a side note I tried to start an attack as chief and captain but It said I couldn't because "I wasn't captain" but that might of been it updating since I had just been promoted like 30 seconds ago. I got demoted again so I haven't been able to get screen shots or try at all since. Answers: username_1: The same thing is happening to me. I hope someone can fix this glitch soon D: Status: Issue closed
ziglang/zig
433873073
Title: avoid linking using --export-all for webassembly Question: username_0: Currently Zig uses `--export-all` to the linker line when targeting WebAssembly. This prevents symbols from being deleted from compiler_rt.a and builtin.a. Instead, only functions explicitly marked `export` in the root source file should be exported. We should be able to create a truly minimal webassembly file if there are no dependencies on compiler_rt functions or builtin.a functions. Related: #1570, #2062<issue_closed> Status: Issue closed
square/leakcanary
185598131
Title: Can not detect the inner class of a strong reference memory leak problem? Question: username_0: Can not detect the inner class of a strong reference memory leak problem? Because i used MAT found the memory leak, but leakcanary showed nomal... Answers: username_1: Could you please give more information and an example of what exactly you are referring to? username_2: Could you please give an example ? username_3: @username_0 could you provide a heap dump or a small example highlighting the issue? username_0: I'm sorry for answer you so late. Before used leakcanary test the memory leak problem, I like use MAT. So I found the MAT show Memory leak when I use hander like this: Handler handler = new Handler{ handlerMessage(){ ... } } but leakcanary don't give me any message, so I want to know if leakcanary can detect the inner class of a strong reference memory leak problem. Sorry for my bad English, I wish you can understand what i mean and thanks for your mention about my question. Thanks very much ! username_4: Can you provide code that exhibits the leak in MAT but not leak canary? username_0: Hi, i'm sorry for that i can't reproduce the bug. But there is a question i can't solve until now. So i need your help. For example, when i use okhttp for network request, i need use callback to deal with the result. the code should be like this: okhttp.builder.xxx.enqueue(new Callback() { @Override public void onFailure(Request request, IOException e) { } @Override public void onResponse(final Response response) throws IOException { } }); If i use the code in activity the Callback is a anonymous inner class, it will lead memory leak, am i right? Can you tell me a good way to solve the problem? By the way, i don't want to use static or weakreference, because i think it's a hard work in a big project. Thanks username_4, you are so famous in Android.^~^ username_5: You can call `Call.cancel` on the OkHttp Call. Also, consider updating to OkHttp 3. username_0: I think there is a risk of memory leaks in all asynchronous callbacks, is there a better way to deal with the risk except cancel the request. Besides, if the request has been sent, the cancel operation can work? username_4: Cancel frees the resources and the strong reference to the callback. If you can't reproduce then I'm closing. Feel free to file a new issue if you can supply a test case that demonstrates it. Status: Issue closed
minio/minio-go
703710807
Title: PresignedPutObject -> Corrupt Image Upload Question: username_0: Hi there, I want to use go to upload images to presigned put object urls in minio. However, after uploading an example.png the file is corrupt / i cannot open it with a image viewer after downloading it again from minio (it says file is text/plain) and I dont know why. Here an example code, you need a running minio instance on 127.0.0.1:9000 with "minioadmin"/"minioadmin" and a example.png in the same folder: ```go package main import ( "bytes" "context" "fmt" "io" "log" "mime/multipart" "net/http" "os" "path/filepath" "time" "github.com/minio/minio-go/v7" "github.com/minio/minio-go/v7/pkg/credentials" ) func multipartUpload(uri string, path string) { file, err := os.Open(path) paramName := "image" fmt.Println("Base path: ", filepath.Base(path)) fmt.Println("err: ", err) defer file.Close() body := &bytes.Buffer{} writer := multipart.NewWriter(body) part, err := writer.CreateFormFile(paramName, file.Name()) // filepath.Base(path)) _, err = io.Copy(part, file) //for key, val := range params { // _ = writer.WriteField(key, val) //} err = writer.Close() req, err := http.NewRequest("PUT", uri, body) req.Header.Add("Content-Type", "image/png") // req.Header.Add("Content-Type", writer.FormDataContentType()) client2 := &http.Client{} resp2, err2 := client2.Do(req) if err2 != nil { log.Fatal(err2) } else { body := &bytes.Buffer{} _, err := body.ReadFrom(resp2.Body) if err != nil { log.Fatal(err) } [Truncated] fmt.Println("Error: ", err) return } fmt.Println("Successfully generated presigned URL", presignedURL) presignedURLstring := presignedURL.String() multipartUpload(presignedURLstring, "./example.png") fmt.Println("upload complete? \n\n") } func main() { fmt.Println("install minio-go: GO111MODULE=on go get github.com/minio/minio-go/v7 ") Upload() } ``` Answers: username_0: leaving out the multipart upload it works: ```go b, err := ioutil.ReadFile("example.png") // just pass the file name if err != nil { fmt.Print(err) } req, err := http.NewRequest("PUT", uri, bytes.NewBuffer(b)) ``` username_1: That is whole point, its not a corruption you can't use multipart/form to upload for a PUT request Status: Issue closed username_0: Thank your for the clarification!
angular/angular
441383685
Title: Mistake in Angular document for Singleton Services Question: username_0: Hi, While going through https://angular.io/guide/singleton-services, i have analyzed that even declaring the root value in providedIn property of @Injectable() cannot make a service singleton because Suppose same service is added as dependency in a feature module and angular Router creates a child injector for it when we navigate to a component in that feature module. child injector creates a local instance of that service. So there will be two instances of same service. So please help me out in how to make it singleton. Answers: username_1: Closing this issue as it does not follow the issue template. Please file a new issue that follows the template, which gives the team the information needed to investigate. Status: Issue closed
jitsi/jitsi-meet-sdk-samples
995005878
Title: iPhone disconnecting after initial connection Question: username_0: We have a iOS mobile app based on the nopodssample to provide video calls for our clients. The app and Jitsi (nopodssample) works fine when the phone is connected via WIFI, but when it is on 4G the connection is made between the 2 participants (1 on iPhone one on PC), no video of the iPhone participant is shown, then after a few seconds the connection is lost. Using the exactly same app, PC software and server hosting but with the iPhone connected to WIFI it works fine. Has anyone any ideas what could be causing this issue with a 4G connection?
chiahsien/CHTCollectionViewWaterfallLayout
42986302
Title: Utilize iOS 8 self-sizing cells to find out cell dimensions Question: username_0: iOS 8 provides a new self-sizing API for `CollectionView` and `CollectionViewCells`. It lets cells determine their own height, based on the content that they're about to load. This dance involves various methods like - `preferredLayoutAttributesFittingAttributes:` (on the cell) - `shouldLayoutAttributesFittingAttributes:` (on the layout) - `invalidationContextForPreferredLayoutAttributes:withOriginalAttributes:` (on the layout) The waterfall layout could make use of this dance and harness the self-sizing height measurement, whilst demanding that width be fixed. Answers: username_1: For my worked approach demoed here https://github.com/honghaoz/Dynamic-Collection-View-Cell-With-Auto-Layout-Demo. It is still produces heavy computations (and CPU spikes) on each `realodData` call. Not sure if caching sizes will help in cases when content is rebuilt after user changes search criteria, sort order etc. username_2: @username_1 You were able to get CHTCollectionViewWaterfallLayout working with cell sizes from auto-layout? Were your cells defined in XIBs or as prototype cells in a storyboard? I've got the latter and can't get things working.
TurkeyTickle/mhw-builder
335216738
Title: Get the slot-item origin of the decoration-slots Question: username_0: Hi, I'm doing an improvement in the details of the skills, but I don't know how to get the item.slot.ItemType to which the decoration.slot belongs, . Thanks in advance. https://github.com/username_0/mhw-builder/commit/2910434ad2cd431cd5a1df42a7c2de3e74cd145a ![sin titulo](https://user-images.githubusercontent.com/10821661/41823805-a257c446-77cb-11e8-82fc-8bc7196a6f91.png) Answers: username_1: I need to refresh my memory on this. I'll get back to you shortly. username_1: DecorationModel has an itemId property, so you should be able to do something like this I think: `const parentItem = _.find(equippedItems, item => item.id === decoration.itemId)` username_0: Thank you very much, I found a different solution, I already have the improvements: https://username_0.github.io/mhw-builder-page/?#v1ii31d12i77i104i147d54i191d54i226l3 username_1: Very nice, looks like a good addition. Feel free to submit another PR. Status: Issue closed
klren0312/ZSpider
544371773
Title: 报错 Question: username_0: npm i 后 npm run dev 报 Error: Cannot find module 'monaco-editor/esm/vs/editor/contrib/gotoSymbol/goToCommands' Answers: username_1: 改成阿里源呢 username_1: ```shell npm install --registry=https://registry.npm.taobao.org ``` Status: Issue closed username_1: # 已修复 monaco-editor与monaco-editor-webpack-plugin版本不对应导致无法安装 ![image](https://user-images.githubusercontent.com/10903843/85199328-37d1a400-b321-11ea-958d-a2c2df4878e9.png) username_1: npm i 后 npm run dev 报 Error: Cannot find module 'monaco-editor/esm/vs/editor/contrib/gotoSymbol/goToCommands' Status: Issue closed
cksource/quail
58355685
Title: Some tests does not have title / description Question: username_0: Tests without title / descr: ```javascript ["aMultimediaTextAlternative", "appletTextEquivalentsGetUpdated", "formAllowsCheckIfIrreversable", "liDontUseImageForBullet", "paragraphIsWrittenClearly", "tableSummaryIsNotTooLong"] ``` More info in [tests.yml](https://github.com/cksource/quail/blob/custom/src/resources/tests.yml).
temporalio/temporal
1121230952
Title: Option not to create DB in auto-setup Question: username_0: **Is your feature request related to a problem? Please describe.** Hello. I use managed DB cloud, which does not allow creating a database using sql. But auto-setup.sh script [tries](https://github.com/temporalio/temporal/blob/master/docker/auto-setup.sh#L180) to create it via sql-tool, which in my case leads to a crash. And I can't use it to apply DB schema. **Describe the solution you'd like** Add option SKIP_DB_CREATE to auto-setup.sh witch will skill `temporal-sql-tool create` comands **Describe alternatives you've considered** The only alternative I can imagine is to write my own script. Which of course doesn't look like a good solution. Answers: username_1: There is SKIP_SCHEMA_SETUP, did you try that? username_0: I need ** to** skip `temporal-sql-tool create` commands (cause db already exists), but run all other (setup-schema and update-schema) to create tables. username_1: @username_0 , do you want to use an existing temporal db instance and have the auto setup to be able to upgrade the schema? Or it it just an empty db that you just need to skip the create step? If so, why is there an empty db, how was it get created? username_0: It is created via database management console (cli or UI). But no sql create database available. Documentation about the cloud https://cloud.yandex.com/en-ru/docs/managed-postgresql/operations/databases#add-db username_0: Thanks a lot! Status: Issue closed
mity/windrawlib
154368532
Title: Cannot build with MinGW Question: username_0: The compilation fails under MingGW (stock install): ``` d:\Work\Code\windrawlib>mingw32-make [ 3%] Building C object src/CMakeFiles/windrawlib.dir/backend-d2d.c.obj [ 7%] Building C object src/CMakeFiles/windrawlib.dir/backend-dwrite.c.obj [ 11%] Building C object src/CMakeFiles/windrawlib.dir/backend-gdix.c.obj [ 14%] Building C object src/CMakeFiles/windrawlib.dir/backend-wic.c.obj [ 18%] Building C object src/CMakeFiles/windrawlib.dir/bitblt.c.obj D:\Work\Code\windrawlib\src\bitblt.c: In function 'wdBitBltImage': D:\Work\Code\windrawlib\src\bitblt.c:52:85: error: macro "ID2D1RenderTarget_CreateBitmapFromWicBitmap" passed 4 arguments, but takes just 3 hr = ID2D1RenderTarget_CreateBitmapFromWicBitmap(c->target, bitmap, NULL, &b); ^ D:\Work\Code\windrawlib\src\bitblt.c:52:14: error: 'ID2D1RenderTarget_CreateBitmapFromWicBitmap' undeclared (first use in this function) hr = ID2D1RenderTarget_CreateBitmapFromWicBitmap(c->target, bitmap, NULL, &b); ^ D:\Work\Code\windrawlib\src\bitblt.c:52:14: note: each undeclared identifier is reported only once for each function it appears in D:\Work\Code\windrawlib\src\bitblt.c:37:27: warning: unused variable 'bitmap' [-Wunused-variable] IWICBitmapSource* bitmap = (IWICBitmapSource*) hImage; ^ src\CMakeFiles\windrawlib.dir\build.make:162: recipe for target 'src/CMakeFiles/windrawlib.dir/bitblt.c.obj' failed mingw32-make[2]: *** [src/CMakeFiles/windrawlib.dir/bitblt.c.obj] Error 1 CMakeFiles\Makefile2:84: recipe for target 'src/CMakeFiles/windrawlib.dir/all' failed mingw32-make[1]: *** [src/CMakeFiles/windrawlib.dir/all] Error 2 Makefile:82: recipe for target 'all' failed mingw32-make: *** [all] Error 2 ``` MinGW/gcc version used: ``` d:\Work\Code\windrawlib>gcc -v Using built-in specs. COLLECT_GCC=gcc COLLECT_LTO_WRAPPER=c:/dev/mingw-w64/x86_64-5.3.0-win32-seh-rt_v4-rev0/mingw64/bin/../libexec/gcc/x86_64-w64-mingw32/5.3.0/lto-wrapper.exe Target: x86_64-w64-mingw32 Configured with: ../../../src/gcc-5.3.0/configure --host=x86_64-w64-mingw32 --build=x86_64-w64-mingw32 --target=x86_64-w64-mingw32 --prefix=/mingw64 - -with-sysroot=/c/mingw530/x86_64-530-win32-seh-rt_v4-rev0/mingw64 --with-gxx-include-dir=/mingw64/x86_64-w64-mingw32/include/c++ --enable-shared --ena ble-static --disable-multilib --enable-languages=c,c++,fortran,lto --enable-libstdcxx-time=yes --enable-threads=win32 --enable-libgomp --enable-libato mic --enable-lto --enable-graphite --enable-checking=release --enable-fully-dynamic-string --enable-version-specific-runtime-libs --disable-isl-versio n-check --disable-libstdcxx-pch --disable-libstdcxx-debug --enable-bootstrap --disable-rpath --disable-win32-registry --disable-nls --disable-werror - -disable-symvers --with-gnu-as --with-gnu-ld --with-arch=nocona --with-tune=core2 --with-libiconv --with-system-zlib --with-gmp=/c/mingw530/prerequisi tes/x86_64-w64-mingw32-static --with-mpfr=/c/mingw530/prerequisites/x86_64-w64-mingw32-static --with-mpc=/c/mingw530/prerequisites/x86_64-w64-mingw32- static --with-isl=/c/mingw530/prerequisites/x86_64-w64-mingw32-static --with-pkgversion='x86_64-win32-seh-rev0, Built by MinGW-W64 project' --with-bug url=http://sourceforge.net/projects/mingw-w64 CFLAGS='-O2 -pipe -I/c/mingw530/x86_64-530-win32-seh-rt_v4-rev0/mingw64/opt/include -I/c/mingw530/prereq uisites/x86_64-zlib-static/include -I/c/mingw530/prerequisites/x86_64-w64-mingw32-static/include' CXXFLAGS='-O2 -pipe -I/c/mingw530/x86_64-530-win32-s eh-rt_v4-rev0/mingw64/opt/include -I/c/mingw530/prerequisites/x86_64-zlib-static/include -I/c/mingw530/prerequisites/x86_64-w64-mingw32-static/include ' CPPFLAGS= LDFLAGS='-pipe -L/c/mingw530/x86_64-530-win32-seh-rt_v4-rev0/mingw64/opt/lib -L/c/mingw530/prerequisites/x86_64-zlib-static/lib -L/c/mingw 530/prerequisites/x86_64-w64-mingw32-static/lib ' Thread model: win32 gcc version 5.3.0 (x86_64-win32-seh-rev0, Built by MinGW-W64 project) ``` Answers: username_1: Ouch. This was a bug in mingw-w64. I already provided patch to them (see https://sourceforge.net/p/mingw-w64/patches/75/) but then forgot about adding a workaround into WinDrawLib, at least until new builds of mingw-w64 are available. I'll try to cook something when I come home tonight. username_0: Looks like they lost the patch somehow then... Thanks for looking into this! username_1: Oops. Overlooked. Recent builds of mingw-w64 already have it fixed. Including mingw-builds x86_64-win32-seh-rev0 you seem to be using. Just tested with the package downloaded here: https://sourceforge.net/projects/mingw-w64/files/Toolchains%20targetting%20Win64/Personal%20Builds/mingw-builds/5.3.0/threads-win32/seh/ Is it possible there is something wrong with your mingw-w64 installation and you are using quite old Win32API headers, e.g. from much older mingw-w64 installations if you have multiple there? username_0: You were right, I forgot to re-run CMake after installing MinGW 5.3.0... Now it compiles flawlessly 👍 My bad... I've just noticed that there's no bezier curve and gradient brush support at all. Are you planning to add those in anytime soon? (Alternatively, I can just add that myself and submit a patch, of course, it doesn't seem too hard) Btw, the things you had to do to make DirectWrite work in C was quite heroic... :) Status: Issue closed username_1: Ok. Closing the issue. I don't plan to work on those features in a foreseeable future as I don't need them. But, of course, feel free to provide patches/pull requests for those. ;-)
openthread/openthread
971614256
Title: ot-daemon in Ubuntu Host, fails to connect to RCP device(nrf528xx) Question: username_0: Hi, I am not able to somehow connect to my RCP device from ot-daemon running in the Host. I have followed following steps. **Environment:** a) HostOS: Ubuntu 18.04 b) RCP Device: nRF52840 USB dongle **Build** a) Built ot-rcp.hex for my nrf52840 USB dongle by following steps given in [this](https://github.com/openthread/ot-nrf528xx/blob/main/src/nrf52840/README.md). b) OT daemon: Built by following [this](https://openthread.io/platforms/co-processor/ot-daemon) c) Run OT daemon in Ubuntu: $./output/posix/bin/ot-daemon 'spinel+hdlc+uart:///dev/ttyACM1?uart-baudrate=115200' $ sudo ./output/posix/bin/ot-ctl **connect session failed: No such file or directory** Note: my RCP device is enumerated in /dev/ttyACM1. Also, I am able to run OTBR docker image on the same RCP device from Ubuntu Host & form a THREAD network with no problem. So, it apparently seems, no issue with the RCP device. Am I missing something basic? Any guidance/insights will be appreciated. Thank You BR, Answers: username_1: Hi @username_0, it looks like the `ot-daemon` didn't start successfully. What's the output of `$./output/posix/bin/ot-daemon 'spinel+hdlc+uart:///dev/ttyACM1?uart-baudrate=115200'`? Did it exit directly? username_0: Hi @username_1 , Actually, output of the command is *Nothing. So, yes, it seems to be fail to launch completely. I am wondering how to get this running.. username_1: Hi @username_0, could you try the guide [here](https://github.com/openthread/openthread/blob/main/src/posix/README.md) instead? And I think you can find some logs from syslog on ubuntu. Could you also try `cat /var/log/syslog | grep ot-daemon` and check the output? username_0: By doing a quick check, it is spitting below error. ./ot-daemon[17887]: Running OPENTHREAD/20200818-01160-gb42261f88; POSIX; Aug 16 2021 15:43:41 ./ot-daemon[17887]: Thread version: 3 ./ot-daemon[17887]: [CRIT]-PLAT----: RCP API Version 3 is not in the supported range [4-4] ./ot-daemon[17887]: [CRIT]-PLAT----: CheckRcpApiVersion() at /<openthread project>/openthread/src/posix/../../src/lib/spinel/radio_spinel_impl.hpp:414: RadioSpinelIncompatible username_2: There was a recent change to the spinel protocol that requires updating the RCP (https://github.com/openthread/openthread/commit/a94e88cd5081148a8bcc99c5123337b94612f995). Can you try building a new RCP with the latest main branch of https://github.com/openthread/ot-nrf528xx ? username_0: Sure, will give it a try ASAP. Thank You~ username_0: After updating new RCP, the ot-daemon seems to be launched successfully. Following are sys logs ./output/posix/bin/ot-daemon[27074]: Running OPENTHREAD/20200818-01168-g3a90ed81f; POSIX; Aug 17 2021 10:39:13 ./output/posix/bin/ot-daemon[27074]: Thread version: 3 ./output/posix/bin/ot-daemon[27074]: Thread interface: wpan0 ./output/posix/bin/ot-daemon[27074]: RCP version: OPENTHREAD/thread-reference-20200818-1160-gb42261f88; NRF52840; Aug 16 2021 21:45:47 **Thank You for your guidance!** However, I ran into a different problem this time. The OTBR docker image which use to run previously has now stopped run on top of this new RCP. Do you see some incompatibility between RCP & host ? avahi-daemon[189]: Registering new address record for 127.0.0.1 on lo.IPv4. otbr-agent[217]: [INFO]-UTILS---: Running 0.3.0-8e0ee63 otbr-agent[217]: [INFO]-UTILS---: Thread version: 1.2.0 otbr-agent[217]: [INFO]-UTILS---: Thread interface: wpan0 otbr-agent[217]: [INFO]-UTILS---: Backbone interface: eth0 otbr-agent[217]: [INFO]-UTILS---: Radio URL: spinel+hdlc+uart:///dev/radio otbr-agent[217]: [INFO]-PLAT----: RCP reset: RESET_POWER_ON otbr-agent[217]: [NOTE]-PLAT----: RCP API Version: 4 otbr-agent[217]: [CRIT]-PLAT----: RCP API Version 4 is not in the supported range [1-3] otbr-agent[217]: [CRIT]-PLAT----: CheckRcpApiVersion() at ../../third_party/openthread/repo/src/lib/spinel/radio_spinel_impl.hpp:414: RadioSpinelIncompatible otbr-web[245]: [INFO]-WEB-----: Running 0.3.0-8e0ee63 otbr-web[245]: [INFO]-WEB-----: Border router web started on wpan0 otbr-web[245]: [ERR ]-WEB-----: OpenThread daemon is not running. username_2: Yes, this is due to the same incompatibility issue, this time the host is too old for the given RCP. See the relevant log output below: ``` otbr-agent[217]: [NOTE]-PLAT----: RCP API Version: 4 otbr-agent[217]: [CRIT]-PLAT----: RCP API Version 4 is not in the supported range [1-3] ``` username_0: After updating the otbr via [this](https://openthread.io/guides/border-router/docker), the issue seems to go away. Thank you very much for sharing insights!! I had a query (off topic). As I understand, OBR can launch, either by directly running docker or by building ot-br-posix repo & executing otbr-agent manually & it has no pre-requirement of ot-daemon running in background, as per my current understanding. Then, why do it sometimes complains of "otbr-web[245]: [ERR ]-WEB-----: OpenThread daemon is not running." ? Am I missing something? And another one:: how does ot-ctl & ot-cli utility applications differ? I observed that, both are working & able to create a Network from command line while otbr-agent is running in background(with my nRF528XX RCP device) ? Is there any recommendation about which one should be used when? username_2: `ot-ctl` is a utility for issuing CLI commands to `ot-daemon`. In this case, `ot-daemon` includes the OpenThread stack, whereas `ot-ctl` is simply a frontend for CLI interactions and connects to `ot-daemon` using a UNIX domain socket. `ot-cli` is a single process that incorporates both the CLI and OpenThread. It does not does not provide a UNIX domain socket. username_0: Thank you for sharing the insights~ username_2: `otbr-agent` does support a D-Bus interface. However, that is not what `ot-ctl` uses. Instead, `ot-ctl` communicates with `otbr-agent` using a UNIX domain socket and generally forwards CLI commands and output verbatim. Status: Issue closed
Azure/azure-cli
802331530
Title: Cannot create Global anycast IP Question: username_0: 1. “az network public-ip create" does not seem to have global tier. This will block customers who want to use multiple Global IPs for the Cross-region LB. 2. The IP created in the backend for Cross-region LB is not of Global tier, which cannot be used for load balancing rule. Answers: username_1: network Status: Issue closed
rancher/rancher
463527888
Title: When nodes deleted from cluster (using kubectl delete node) are stopped and started , they get registered back as worker nodes that are non schedulable to the cluster. Question: username_0: Rancher Server version - Latest build from master - `5144aa85` Steps to reproduce the problem: Create a custom cluster(no cloud provider) with following node configuration with nodes from AWS: 2 control nodes - c1,c2 3 worker nodes - w1, w2,w3 3 etcd nodes - e1, e2, e3 From AWS console , stop one of the worker nodes say w1. This node gets to "Unavailable" state in rancher server UI as expected and it is in "Not Ready" state when monitoring node list from kubectl command. Delete the node from cluster using kubectl command. This results in node getting removed from cluster in rancher server side as expected. Start this worker node from AWS console . The worker node gets registered to the cluster again as "worker" node in rancher server UI. Node list from kubectl command indicate that there is NO role assigned to this node. Also containers dont get deployed on this worker node , since they are non schedulable. <img width="213" alt="Screen Shot 2019-07-02 at 8 09 04 PM" src="https://user-images.githubusercontent.com/4266958/60561617-fe1b1e00-9d08-11e9-9900-f72c009b1e44.png"> Same behavior is seen when a control node / etcd node gets deleted from the custom cluster using `kubectl delete node` command and then it was stopped and started from the AWS console. In all the cases , the nodes get registered back as worker node that is not schedulable. Answers: username_1: @username_0 what do you expect to happen in this case? username_0: @username_1 , Ideally as part of removing a node from the cluster , we should be able to delete all the K8s related components on this node . This way , when the nodes get restarted , there will not be an attempt made to join the cluster back. username_1: Hm. @username_2 what do you think? Is there anything we can do here? username_2: @username_1 I think the best we can do in this case is to instruct the user to cleanup the data on the node before bringing it back. So it gets registered as a new one. username_3: @username_0 when we're deleting the node through kubectl delete, there is no way for us to clean up the node from these components. I don't think there is much we can do in this use case. If you choose to manage your k8s nodes outside using kubectl versus Rancher, we will hit use cases like this. Status: Issue closed username_0: Related issue - https://github.com/rancher/rancher/issues/17893
opencv/cvat
528958170
Title: Tfrecord Dump Error Question: username_0: If you dump annotation for `TFRecord Zip 1.0` then there will be a lot of problem with the dumped file. (`/annotation/tfrecord.py`) 1) First problem is with `label_map.pbtxt`. The problem is `id = 0` is reserved for background. So, it needs to start from 1 NOT 0. This can easily be fixed. 2) Tfrecord contains encoded images so that training can be faster. But if you look at the current file, it does not encode images into tfrecord. If you use this tfrecord in TF Detection API, it will throw an error. Answers: username_1: @username_0 , Thanks for the report. The second problem will be fixed in our new dataset framework (datumaro). Dump annotations is used only to dump annotations (without images). We recently introduced `export as dataset` feature. Now it doesn't support TFRecord format but it will do. The feature will export images + annotations. Please stay tuned. username_2: @username_0, please check if [this](#894) PR satisfies your needs - you can test it by exporting a dataset. Status: Issue closed
nltk/nltk
307754788
Title: How to run Stanford postagger to listen to port 9000 so we can use it from NLTK? Question: username_0: How to run Stanford postagger to listen to port 9000 so we can use it from NLTK, without running the Java jar each time (and leave the loaded model in memory, to save time) Answers: username_1: If I understand you correctly, you need to start CoreNLP server. Refer to https://stanfordnlp.github.io/CoreNLP/cmdline.html Once the server is running, you can tag sequnces https://github.com/nltk/nltk/blob/develop/nltk/parse/corenlp.py#L363 username_1: Also, consider using https://github.com/stanfordnlp/python-stanford-corenlp username_0: @username_1, 1. The words "server", "port" or "9000" do not appear in the first link that you sent. It's not clear how to run a server for POS tagging, and how to specify the tagger model (e.g. english-left3words-distsim.tagger). 2. I am currently using the jar stanford-postagger-3.9.1.jar from [here](https://nlp.stanford.edu/software/tagger.shtml). Is it different from CoreNLP? Can what I use server as server as well? Status: Issue closed username_1: yeah, i meant to link to the corenlp-server page, my mistake
LunaMultiplayer/LunaMultiplayer
333827400
Title: Contract interface is buggy Question: username_0: ------------------------------------------------------------------------------ OS: Client : Windows 10 - Server : Linux Debian ------------------------------------------------------------------------------ KSPVersion: 1.4.3 ------------------------------------------------------------------------------ LMP version (include the build number if you are using a nighly build): ------------------------------------------------------------------------------ MODS (add link to download them): 0.10.46 ------------------------------------------------------------------------------ Expected behaviour: Clean contract interface ------------------------------------------------------------------------------ Current behaviour: Too much contracts available + contract history is too short (only 3) ------------------------------------------------------------------------------ Steps to reproduce: Not sure. Once, I got a contract where every steps are "done" but the contract is still ongoing. So I used KSP debug interface to close contracts. Since then, I often got this contract mess. [Debug.tar.gz](https://github.com/LunaMultiplayer/LunaMultiplayer/files/2116796/Debug.tar.gz) Answers: username_0: ![screenshot1](https://user-images.githubusercontent.com/1453854/41622793-fdf2001e-7410-11e8-9c8d-c6fb39603a81.png) username_1: I can duplicate. Even have multiple contracts for "terminal" contract (major contacts given when leveled up) 2 leave the atmosphere. username_2: We are bit crowded with normal jobs and other more urgent bugs but will have a look once we're more free username_1: No need to rush on my account. Just want to help the environment. Sent from my iPhone Status: Issue closed username_2: Hey @MalteJanz I think I managed to fix this let me know if you've got any idea of what may be causing bugs. @username_1 and @username_0 feel free to test it on nightly and let me know if the problem still persists. Thank you all!
scikit-mobility/scikit-mobility
467688437
Title: Fatal error in Github Desktop on checkout Question: username_0: Trying to checkout the repo on Windows using Github Desktop. I think the '.' directory in the docs is the issue: ![skikit-mobility-checkout-error](https://user-images.githubusercontent.com/590385/61168872-cad05e80-a554-11e9-976e-3e9a5c8f6ba5.PNG) Answers: username_0: As discussed on Skype on Thursday, it might be best to move the documentation to a separate repository. username_1: The issue is now solved. The compiled documentation has been moved on a separate branch. Status: Issue closed
steelbrain/linter
206585659
Title: Consider integrating busy-signal Question: username_0: I know it's planned for v2 but that it has being a long time coming and isn't really clear when it's actually gonna arrive. In the meantime I would really like to have flowtype plugin with would be able to signal if it's still type-checking or if it has type checked code successfully. Existing plugins do not want to add dependency on busy-signal as they expect this to be done by linter instead, so I'm stuck running a forked flowplugin for quite some time now. Thanks for considering. Answers: username_1: It's not just planned it's implemented in v2 and working perfectly I'm sorry you're stuck on a forked flowplugin, the work for the linter isn't much. It's just that I have so many public and private projects, it's been getting hard for me to spare time for linter. But linter is still, by far, my favorite project. I'll keep this open so it's up for grab if someone from the community wants to work on it but linter v2 isn't that far away, linter itself is done completely, only things left are in the ui consumer username_1: [Sneak Peek of the Linter v2](https://www.youtube.com/watch?v=Ek7p49sf8Eo) All main issues resolved, pending on docs and specs username_1: Linter v2 is finally out people. You can read more about it in the [release post](http://username_1.me/2017/03/13/linter-v2-released.html). Status: Issue closed
hyperf/hyperf
948437716
Title: [BUG] [hyperf/async-queue :2.2.*] Interface 'Hyperf\Signal\SignalHandlerInterface' not found Question: username_0: Execute the command and paste the result below. Command: `uname -a && php -v && composer info | grep hyperf && php --ri swoole` ```bash # Paste the result here. Darwin xuzhendeMBP.lan 20.5.0 Darwin Kernel Version 20.5.0: Sat May 8 05:10:33 PDT 2021; root:xnu-7195.121.3~9/RELEASE_X86_64 x86_64 PHP 7.3.29 (cli) (built: Jul 12 2021 11:36:07) ( NTS ) Copyright (c) 1997-2018 The PHP Group Zend Engine v3.3.29, Copyright (c) 1998-2018 Zend Technologies with Zend OPcache v7.3.29, Copyright (c) 1999-2018, by Zend Technologies hyperf/async-queue v2.2.0 A async queue component for hyperf. hyperf/cache v2.2.0 A cache component for hyperf. hyperf/command v2.2.0 Command for hyperf hyperf/config v2.2.0 An independent component that provides configuration container. hyperf/constants v2.2.0 A constants component for hyperf. hyperf/contract v2.2.0 The contracts of Hyperf. hyperf/crontab v2.2.0 A crontab component for Hyperf. hyperf/database v2.2.0 A flexible database library. hyperf/db v2.2.0 hyperf/db-connection v2.2.0 A hyperf db connection handler for hyperf/database. hyperf/devtool v2.2.0 A Devtool for Hyperf. hyperf/di v2.2.0 A DI for Hyperf. hyperf/dispatcher v2.2.0 A HTTP Server for Hyperf. hyperf/engine v1.1.6 hyperf/event v2.2.0 an event manager that implements PSR-14. hyperf/exception-handler v2.2.0 Exception handler for hyperf hyperf/filesystem v2.2.0 flysystem integration for hyperf hyperf/framework v2.2.0 A coroutine framework that focuses on hyperspeed and flexible, specifically use for build microservices and middlewares. hyperf/guzzle v2.2.0 Swoole coroutine handler for guzzle hyperf/http-message v2.2.0 microservice framework base on swoole hyperf/http-server v2.2.0 A HTTP Server for Hyperf. hyperf/ide-helper v2.2.0 IDE help files for Hyperf. hyperf/logger v2.2.0 A logger component for hyperf. hyperf/memory v2.2.0 An independent component that use to operate and manage memory. hyperf/model-listener v2.2.0 A model listener for Hyperf. hyperf/paginator v2.2.0 A paginator component for hyperf. hyperf/pool v2.2.0 An independent universal connection pool component. hyperf/process v2.2.0 A process component for hyperf. hyperf/redis v2.2.0 A redis component for hyperf. hyperf/server v2.2.0 A base server library for Hyperf. hyperf/testing v2.2.0 Testing for hyperf hyperf/utils v2.2.0 A tools package that could help developer solved the problem quickly. hyperf/watcher v2.2.0 Hot reload watcher for Hyperf xiangxin/hyperf-gaode 2.2.0 hyperf gaode swoole Swoole => enabled Author => <NAME> <<EMAIL>> Version => 4.6.7 Built => May 17 2021 10:26:05 coroutine => enabled with boost asm context kqueue => enabled rwlock => enabled sockets => enabled openssl => OpenSSL 1.1.1j 16 Feb 2021 dtls => enabled http2 => enabled json => enabled [Truncated] swoole.enable_library => On => On swoole.enable_preemptive_scheduler => Off => Off swoole.display_errors => On => On swoole.use_shortname => On => On swoole.unixsock_buffer_size => 262144 => 262144 ``` ### Description: 组件:hyperf/async-queue PHP Fatal error: Interface 'Hyperf\Signal\SignalHandlerInterface' not found in /path/vendor/hyperf/async-queue/src/Signal/DriverStopHandler.php on line 20 Fatal error: Interface 'Hyperf\Signal\SignalHandlerInterface' not found in /path/vendor/hyperf/async-queue/src/Signal/DriverStopHandler.php on line 20 ### Steps To Reproduce: 1. 将 hyperf 组件版本都修改为 '2.2.*' 2. composer update -o 3. php bin/hyperf.php start Answers: username_1: 升级到 PHP7.4 版本以上,或者 安装对应的组件 Status: Issue closed username_0: 组件:hyperf/async-queue 缺少组件:hyperf/signal 安装 ```shell composer require hyperf/signal ``` 问题已解决
ps2homebrew/Open-PS2-Loader
994016780
Title: Dont read memory cards [ISSUE] Question: username_0: ### Checks - [ ] I have checked existing [__OPL issues__](https://github.com/ps2homebrew/Open-PS2-Loader/issues) for duplicates and found none - [ ] I am using either OPL [latest stable version](https://github.com/ps2homebrew/Open-PS2-Loader/releases/latest), or [archived version](https://mega.nz/folder/Ndwi1bAK#oLWNhH_g-h0p4BoT4c556A) or [__OPL beta version__](https://github.com/ps2homebrew/open-ps2-loader/releases) ### Describe the issue i have a ps2 model:SCPH-90001 and using the last version of OPL (1.1.0.7) but in this version OPL cant read any memory card, and dont recognize any vmc too, if i try to use a old version, i can read the memory cards but no in this version, if i try to save the settings just frezze in saving screen, i need to shut it off to start, configurate it again and start a game but dont matter wath kind of game i play i cant save the progress of my games. ### Console model 90001 ### OPL version / revision 1.1.1.0.7 ### In which device(s) have you experienced this issue? MX4SIO ### Context and extra information _No response_ Answers: username_1: I am not sure if I understand. You have a problem with a normal memory card & VMC or with MX4SIO? Or maybe the problem starts when you turn on MX4SIO? I have never seen OPL in version `1.1.1.0.7` or `1.1.0.7`. username_2: Same: 1. Saving after MX4SIO is on freezes console 2. Memory card can't be detected when playing from MX4SIO 3. Trying to save when MX4SIO is on not only freezes console but wipes out entire OPL configuration except internet settings 4. Only manual configuration for MX4SIO with integrated in wLE text editor works. username_3: i have the same problem
zikula-modules/Pages
209854341
Title: Cannot install Pages Module under Zikula 1.4.6 - this Errors Question: username_0: Hello Developers, Unfortunately I can not install "Pages" under Zikula 1.4.6. After clicking "Install", I see these errors. `Oops! An Error Occurred The server returned a "500 Internal Server Error". Something is broken. Please let us know what you were doing when this error occurred. We will fix it as soon as possible. Sorry for any inconvenience caused. An exception occurred while executing 'INSERT INTO pages (title, metadescription, metakeywords, urltitle, content, counter, displaywrapper, displaytitle, displaycreated, displayupdated, displaytextinfo, displayprint, language, cr_date, lu_date, obj_status, cr_uid, lu_uid) VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)' with params ["Welcome to Pages content manager", "", "", "welcome-to-pages-content-manager", "This is a demonstration page. You can use Pages to create simple static content pages. It is excellent if you only need basic html for your pages. You can also utilize the Scribite module for WYSIWYG content creation. It is well suited for informational articles, documents and other \"long term\" type content items.<br \/><br \/>Pages is a hookable module which allows you to hook EZComments or other hook providers to extend the capabilities of your module.", 0, 1, 1, 1, 1, 1, 1, "en", null, null, "A", null, null]: SQLSTATE[23000]: Integrity constraint violation: 1048 Column 'cr_date' cannot be null` I have already downloaded the master branch, but the same error. Can you please take a look with the last Zikula version. Thanks in advance. Answers: username_0: This was a hard birth with the installation of Pages Module, but I have successfully implemented it. I had to run again and again in the MySQL database "drop tables pages" and then the installation tried again and again. Now I just have to test the pages. The installation of knowledgebase still remains, which also shows errors. username_1: I wouldn't bother with knowledgebase. are you using MySQL? username_0: OK, So, I made deleting the tables directly using an SQL query in the database. It did not work. Too bad, but maybe someone wants to see the knowledgebase module :smile: username_0: Unfortunately I have problems with the installation of pages this time. These errors are displayed. `An exception occurred while executing 'INSERT INTO pages (title, metadescription, metakeywords, urltitle, content, counter, displaywrapper, displaytitle, displaycreated, displayupdated, displaytextinfo, displayprint, language, cr_date, lu_date, obj_status, cr_uid, lu_uid) VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)' with params ["Welcome to Pages content manager", "", "", "welcome-to-pages-content-manager", "This is a demonstration page. You can use Pages to create simple static content pages. It is excellent if you only need basic html for your pages. You can also utilize the Scribite module for WYSIWYG content creation. It is well suited for informational articles, documents and other \"long term\" type content items.<br \/><br \/>Pages is a hookable module which allows you to hook EZComments or other hook providers to extend the capabilities of your module.", 0, 1, 1, 1, 1, 1, 1, "en", null, null, "A", null, null]: SQLSTATE[23000]: Integrity constraint violation: 1048 Column 'cr_date' cannot be null` This is again the same problem and I try again and again to install. Server is Ubuntu with PHP 5.6 and MySQL 5.6. Module version is the last one that should be compatible with Zikula 1.4. username_2: Closing since this issue is outdated and Pages 4.0.0 provides a complete rewrite. Status: Issue closed
magicblack/maccms10
1084000240
Title: 会员组权限里缺个筛选页权限选项 Question: username_0: 打开页面筛选页面提示权限不足,查看后台网站参数配置----基本设置------筛选页开关 已开启,用户组权限里没有筛选页权限选项可以勾选,其它权限都是勾选开启了,前台无法正常打开筛选页面,重新在给用户组所有权限后更新缓存也一样问题,maccms10 2022.1000.3025版本的 Answers: username_1: 提供一下重现步骤,以及对应的网址路由 username_0: 我的网址是www.5yeb.com 打开名人栏目------点击查看更多,展示网址:https://www.5yeb.com/actorshow/%E4%BA%9A%E6%B4%B2--------.html 然后提示没有权限,是我的模板标签地址错误还是权限问题呢 username_2: 同樣問題, 3005, 3025, 3026我都試過 https://kanba.ga/index.php/vod/type/id/1.html 進去再點 更多 我是跑出 “筛选页功能关闭中” username_0: 你这个是网站参数配置-----性能优化-------筛选页开关 没开启,开启后在看看是不是提示权限问题,我的是开启了提示权限不足 username_2: 解決了,是我的問題 從非官方版本換成官方版本後,沒注意到多了"筛选页开关" 還以為是BUG 謝謝username_1! username_0: 兄弟果然看东西不仔细呀,回复你的不是username_1 囧 username_2: username_0抱歉!真的是眼瞎了! 再次謝謝您! 另外,我試筛选页的手機端和電腦端都沒發生跳其他網站, 不知是哪出問題了呢! Status: Issue closed username_1: 感谢提供的更多信息,已解决提交,将在下个版本里发布更新。
elixir-editors/vim-elixir
292637128
Title: How do you utilize the `mix format` integration? Question: username_0: Hi thanks for all the work on this plugin. I am however confused on how to use the mix formatter integration. Or even the breadth of the integration. Is it supposed to show you formatting errors in the editor as you type (or on save)? Is it supposed to provide a format command? Thanks.<issue_closed> Status: Issue closed
ga-dc/project1-gallery
159703601
Title: Concentration (<NAME>) Question: username_0: Link to repo: https://github.com/username_0/concentration Link to deployed app: http://username_0.github.io/concentration/ Things you'd like specific feedback on: * Is the code somewhat readable and logically organized for other developers to follow? * From a potential employer reviewing this project perspective, is the provided CSS "good enough" and I can focus future effort on additional JS functionality or is it too basic? * When clicking multiple boxes quickly, the timer seems to not reset and a new click nearly immediately hides the selected box. How can I reset the timer so after two selects the timer resets so the next selection will stay listed for two seconds?<issue_closed> Status: Issue closed
wombats-writing-code/fbw-instructor-app
168818341
Title: Check and merge into master Question: username_0: Components have been reorganized with additional authoring functionality to support the workflow of: choosing a directive >> choosing question(s) that address that directive >> specifying the number required. Please help merging reorg-components into master? Things to note: * Tap on a directive to check out EditDirective.js * Comments littered throughout where stubs are. Feel free to delete when done plugging in. Status: Issue closed Answers: username_1: merged in 9a079918ef09908a0a4344d2df84a5851b765a31
mebjas/html5-qrcode
718198149
Title: Compatibility - [Linux] [Firefox] - [aspectRatio] Question: username_0: **Describe the bug** aspectRatio configuration doesn't work under Firefox because this constraint is not yet supported by Firefox, see: https://developer.mozilla.org/en-US/docs/Web/API/MediaTrackConstraints/aspectRatio#Browser_compatibility **Additional context** By allowing the user to pass a full videoTrackConstraint object the problems caused by this can be avoided (since the user can then specify width and height, which imply an aspect ratio). I made these changes and will include the diff here (I'm not very good with Git yet), if you want I'll attempt a pull request as well. ``` --- a/src/html5-qrcode.js +++ b/src/html5-qrcode.js @@ -94,9 +94,18 @@ class Html5Qrcode { * |********************| * |********************| * ---------------------- - * - aspectRatio: Optional, desired aspect ratio for the video feed. - * Ideal aspect ratios are 4:3 or 16:9. Passing very wrong aspect - * ratio could lead to video feed not showing up. + * - videoConstraints: Optional object, can specify the desired width, + * height, aspect ratio, and frame rate for the video feed. + * + * const constraints = constraints = { + * width: { min: 640, ideal: 1920, max: 1920 }, + * height: { min: 400, ideal: 1080 }, + * aspectRatio: 1.777777778, + * frameRate: { max: 30 }, + * }; + * + * Example from: https://developer.mozilla.org/en-US/docs/Web/API/Media_Streams_API/Constraints + * Reference: https://developer.mozilla.org/en-US/docs/Web/API/MediaTrackConstraints#Properties_of_video_tracks * - disableFlip: Optional, if {@code true} flipped QR Code won't be * scanned. Only use this if you are sure the camera cannot give * mirrored feed if you are facing performance constraints. @@ -300,14 +309,11 @@ class Html5Qrcode { } $this._localMediaStream = mediaStream; - if (!config.aspectRatio) { + if (!config.videoConstraints) { setupVideo(); } else { - const constraints = { - aspectRatio : config.aspectRatio - } const track = mediaStream.getVideoTracks()[0]; - track.applyConstraints(constraints) + track.applyConstraints(config.videoConstraints) .then(_ => setupVideo()) .catch(error => { console.log("[Warning] [Html5Qrcode] Constriants could not be satisfied, ignoring constraints", error); ``` Answers: username_1: This feature request looks similar to https://github.com/username_1/html5-qrcode/issues/97 Let me take this on and add support for passing video constraints directly as a hidden API for now. You can follow the conversation in https://github.com/username_1/html5-qrcode/issues/97 to understand why I do not want to make it public yet but you can go ahead and use the latest version of library I publish. If this would solve your requirement please close this issue and track https://github.com/username_1/html5-qrcode/issues/97. Thanks! username_0: You are right, passing the videoConstraints mentioned in #97 would solve this (specifically if width and height are included). I'll close this request, and will track #97 Thanks! Status: Issue closed
mapbox/mapbox-gl-js
917336503
Title: How to combine case expression and property in styling line-color? Question: username_0: Hi, I'm stuck with styling `line-color`. I want to achieve something like this below, but I don't know how to get color combining colorId and stops in case expression ``` 'line-color': ['case', ['==', ['get', 'isOpen'], true], *use color based on property colorId and stops*, '#878787', ], ``` Status: Issue closed Answers: username_1: I believe what you looking for might be [line gradients](https://docs.mapbox.com/mapbox-gl-js/example/line-gradient/). Which generally achieve these stop-based rendering of colors on lines. But unfortunately, you're trying to get the color stops inferred from the data, and we currently do not support data driven line gradients. I'll close as duplicate of https://github.com/mapbox/mapbox-gl-js/issues/8977, but feel free to reopen if I misunderstood your use case.
ericcornelissen/pinned-tabs-for-atom
127900147
Title: Close all tabs but pinned Question: username_0: Can I propose a feature for an command to close all tabs but pinned. Adding it to the contextmenu on the tabs would be great (when at least one tab is pinned). Answers: username_1: I like this idea, I will try to get it to work as soon as possible :smile: username_1: I implemented this into the package. I did not publish it to APM yet because I want to fix #10 first. If you want, you can test its functionality by cloning this repository. Also, the entry for this feature in the context menu is now at kind of a wierd place (it is now second in the list). I would prefer to move it 'Close tabs to the right'. Unfortunately I couldn't get this to work, do you by any chance know how to do this? username_0: Have a read here: https://github.com/atom/atom/issues/6270 Status: Issue closed username_1: Well, that seems clear to me :cry:
rust-lang/rust
361544861
Title: i128 / u128 are not compatible with C's definition. Question: username_0: While fixing various bindgen bugs related to `long double`, `int128`, (https://github.com/rust-lang-nursery/rust-bindgen/issues/1370, etc). I realized that the following Rust program, in my x86_64 Linux machine: ```rust #[repr(C)] struct Foo { f: u128, } fn main() { println!("Align: {}", ::std::mem::align_of::<Foo>()); } ``` Prints `8`. While the following C program: ```c #include <stdio.h> struct foo { unsigned __int128 t; }; int main() { printf("Align: %ld\n", _Alignof(struct foo)); } ``` Prints `16` on the same system. This is pretty unexpected, and means that i128 / u128 are not really usable for FFI / alignment purposes. Answers: username_1: I think the correct fix would be to add `long_double` type to `libc` with the correct alignment. I guess that's probably not hard? I could have a go at that if it sounds like the right thing to do. username_2: This is a bug in llvm/clang, not sure about gcc. There [was an attempt](https://reviews.llvm.org/D28990) to fix this in the past, but it got reverted because of reasons. The sysV __int128 should be 16-byte aligned. username_2: Rather, this is blocked on LLVM data-layouts getting fixed, because rustc expects them to be matched last time I checked. username_0: Ugh, thanks @username_2... I _think_ this is also a problem for `long double`, reading that LLVM patch they'd align it to 64 bits. Sigh. ```c #include <stdio.h> struct foo { long double bar; }; int main() { printf("%ld\n", _Alignof(struct foo)); // 16 } ``` username_3: Copying from the last thread, since thid thread seems to mostly be about long double. In C and C++, on x86_64 Linux, alignof(__int128) is equal to 16. However, in Rust, align_of::<u128>() is equal to 8. C++: https://gcc.godbolt.org/z/YAq1XC Rust: https://gcc.godbolt.org/z/QvZeqK This will cause subtle UB if you ever try to use i128s across FFI boundaries; for example: C++: https://gcc.godbolt.org/z/PrtHlp Rust: https://gcc.godbolt.org/z/_SdVqD username_4: The `improper_ctypes` warning diagnoses usage of these types in C FFI, so at least users are alerted that this does not work. username_4: @centril @username_9 this should be I-unsound. username_5: Why that? Rust itself is sound. The types are not FFI-safe, but FFI needs unsafe code. username_4: Using `i128` and `u128` on FFI is sound. It currently isn't due to a bug on our end, therefore, this is IMO a soundness bug. username_2: This can be generalized for any `T`. The complications arise with getting layout right on both sides of the FFI (more specifically it is *sound* to use `repr(Rust)` type as long as you manage to get layout of `T` right on the other side). Since only a few targets have a defined layout for 128-bit integer types and many don’t even have an implementation defined behaviour, it makes perfect sense to delegate them to the same level of FFI support as we do for other `repr(Rust)` types. Tust to be entirely clear: even if we fixed this particular issue for x86_64 SysV targets, it will only make u128 matching with other x86_64 SysV-abiding artifacts. For most other targets the fact that i/u128 ABI is not specified or is implementation-specific will remain. username_4: That means that the layout of these types is unspecified, which prevents users from using `i128` to interface with C, requiring us to add a `libc::__int128` on those platforms that's incompatible with the Rust type, and requiring users to use that in C FFI. What's the rationale for not providing a layout that's compatible with `__int128` on those platforms? username_4: For all other integer types, we match the C layout of the platform, when the platform has the integer type (the `stdint.h` header is optional, so not all platforms necessarily have it). username_5: Nobody argued that we should have a different layout. I agree using a different layout is a bug. I just don't agree it's a soundness bug. `u128`/`i128` are not FFI-safe when they should be. That's a bug. We also don't *claim* that they are FFI safe, so it's not a soundness bug. Ultimately where to draw the line for soundness bugs can be arbitrary though. username_6: Is anyone working on fixing this? Trying to understand why does clang and rust have different u128 types https://godbolt.org/z/WyzeRz username_4: The reason is that `u128` is not FFI safe, so you can't use it on FFI, and therefore it doesn't matter whether it has the same layout as some other clang type. username_6: @username_4 but why isn't it FFI safe? I would've thought they would use the same llvm type username_4: See: https://github.com/rust-lang/rust/issues/54341#issuecomment-506719475 username_5: So which targets *do* have a defined call ABI for passing 128-bit integers? You said [all the MSVC targets](https://github.com/rust-lang/rust/issues/54341#issuecomment-506728338), at least. I suppose 64bit x86 on Linux also has an ABI for this. What else? It was my understanding that this is [blocked on LLVM bugs](https://github.com/rust-lang/rust/issues/54341#issuecomment-422805235). Not sure if there is an LLVM issue tracking this. But now it sounds more like this is blocked on Rust needing a target-dependent notion of "FFI safe"? username_4: I said that the msvc compiler is the only widely-used C compiler I know that does not have an `__int128` type, but that's not the only C compiler for this target, and other compilers for this target like clang do have a `__int128` type and support using it on the windows msvc targets (https://gcc.godbolt.org/z/vivo9o). I'd suspect this agrees with what GCC does. If all these other compilers for this target agree, then there is an ABI for `__int128` that everybody agrees on for this target. The fact that MSVC doesn't support it just means that MSVC is not able to export symbols or call other symbols that use this type, but that's it. username_4: There might be targets without an ABI spec and/or without C compilers, in which case, we can try to agree with whatever is used on the target, and if that doesn't support 128-bit integers, does it even matter what we do? username_7: Most 32 bit C toolchains don't have `__int128` AFAIK. Definitely not 32 bit Linux. username_3: @username_7 I don't know of any 32-bit C toolchains with `__int128`, fwiw username_5: @username_4 ah sorry, I read your prior post too hastily. Indeed you said everything *except* MSVC. The LLVM bug @username_2 mentioned is about making `i128` 16-aligned everywhere, if I understand correctly. But Clang already does something to make its `__int128` 16-aligned, why can't rustc do the same? It does seem rather odd though that on 32bit, `u64` would be 4-aligned but `u128` would be 16-aligned... actually doesn't that violate the "max alignment" thing C defines? username_5: So also taking into account what @username_7 and @username_3 just wrote, my understanding is there are actually two issues: * The 128-bit integer types are only FFI-safe on 64bit-platforms. So if we allowed them in the lint, we'd need to make it target-dependent, which is a portability hazard. * As a consequence of that, we can basically do whatever we want on 32bit platforms (these are effectively like `repr(Rust)` types). But on 64bit we should do what the platform ABI says, so we need to bump the alignment to 16. Is that an accurate summary? For the second issue, @username_2 tried to make LLVM always do that for `i128`, but that causes problems. However, clang already does "something" to make its `__i128` 16-aligned, can't rustc do the same? username_4: No, most (all?) 32-bit targets support a type that sets the `_Alignof(max_align_t) == 16`, so `__int128` doesn't break that. @username_3 and @username_7 are right. I just skimmed the x86 and riscv 32-bit linux ABIs and they don't mention the `__int128` type in there, while they do so for the 64-bit types. I don't know why this is, but if the C compilers on those targets don't support it, then there is no need to document an ABI for it. username_2: To elaborate on MSVC story: the Microsoft documentation for the x64 ABI does *not* explicitly specify the ABI for `__int128`[1][2]. If one was to infer the ABI for `u128` from the rest of the wording in Microsoft’s documentation then neither clang/llvm nor gcc correctly implement it (and both differ). [1]: https://docs.microsoft.com/en-us/cpp/build/x64-software-conventions?view=vs-2019 [2]: https://docs.microsoft.com/en-us/cpp/build/x64-calling-convention?view=vs-2019 This in my opinion makes the MSVC target the prime example of a 64-bit target where the __int128 is unspecified and therefore not FFI-safe. username_4: Does the implementation of clang/llvm matches that of gcc for the target ? If so, it is possible to make `i128` FFI safe by giving it the same ABI that all compilers of that platform _that support the type_ use. One can't use the type to interface with MSVC, so it doesn't matter that the MSVC ABI doesn't document the type. There is precedent of doing this, for example, it is exactly how we argue that ZSTs are FFI safe, e.g., by arguing that all compilers in a platform treat them in the same way, independently of whether the ABI spec covers them or not. We could provide a stronger _definition_ of what FFI safety means in Rust. For example, we could define that FFI safety means that (1) the platform has an ABI spec, (2) the platform ABI spec provides a precise ABI for the type, (3) and all C compilers of the platform stick to that ABI. I'm quite certain that, under that definition, there aren't any Rust types that can be used in C-FFI, on any platform, and therefore, such a definition wouldn't be any useful. So the question is how much can we relax it to make it useful. If MSVC ever provides `__int128`, and if when it does, it gives it an ABI different than what clang and gcc do, then we'd need to change the ABI on the target. I think that's fine. It means that all Rust libraries using `u128` that were compiled before that change can't interface with MSVC compiled code, but that's something that they couldn't do before anyways because MSVC did not provide a `__int128` type. This combined with the fact that MSVC is one of the platforms that breaks its ABI quite often (although the trend is currently for this to happen less often) makes me comfortable with such a situation. username_2: I did say in the quoted sentence that they differ from each other; at least they did back when I was looking at it. username_4: I'm trying to reproduce without any luck. Do you remember what was different? Did you fill a clang bug report about this? username_2: The only related issue filled by myself I can find is https://gcc.gnu.org/bugzilla/show_bug.cgi?id=78799 username_4: I've found a clang bug about this that was fixed on clang 6 (https://bugs.llvm.org/show_bug.cgi?id=31362), so maybe this is fixed now? --- I'll be fine with having the FFI-safety warning be target dependent, iff we document for the targets for which we issue the warning the reason (e.g. on the msvc targets that Microsoft does not document an ABI for the type, and therefore our implementation could change in the future to match that), and also, if we implement the type on those targets in a "sane" way (e.g. on the `msvc` targets we implement it to do the same thing that gcc and clang do, until there is something better that we can do). username_8: Sorry to be That Girl and bump this ancient topic, but I have spent the last couple weeks getting repeatedly stymied by this issue. Here is a writeup of the situation *as I understand it*, in the hopes that it encourages/helps progress. ---- # Terminology I will be using the same terminology as my article on rust layouts and abis[5]. A type's (target-specific) *layout* is its size and align, and for composites the offsets of its fields (arguably also endianess). If two types agree on layout ("layout-compatible"), then we can use them in FFI *when passed by reference*. A type's (target-specific and calling-convention-specific) *ABI* is its layout and its *kind*[4] (integer, float, composite, vector, ...). Something like the System V ABI will define how things are passed by appealing to this "kind", so getting this right is necessary for FFI. Specifically, the kind of a type (and its fields) defines whether passing it "by value" is done with registers, on that stack, or by-reference. # The Core Problem u128/i128 are not layout-compatible with C(++) __int128 on most platforms (and therefore also not ABI-compatible). This makes them unsuitable for FFI, which is a problem for interoperating with APIs which use them. Rust lowers u128/i128 directly to the llvm i128 primitive type. This makes sense on paper, but unfortunately llvm has a long-standing bug[0] where it aligns i128 on 64-bit platforms to 8, as if it were an emulated integer type (just as u64 on x64 only has align 4). The typical 64-bit ABI of __int128 aligns it to 16. This doesn't effect clang because it manually defines the layout of int128 somehow[2] (I don't know enough about llvm to elaborate on this). # Consequences of the Problem (And How To Hack Around It?) **Only the compiler can define correct definitions of core primitive types, because of ABI "kinds".** **It is impossible for any Rust FFI crate to soundly expose an __int128 type.** **Many aarch64 APIs are missing from crates because platforms use __int128 in the definition of simd registers.**[3] To understand why this is, let's look at trying to provide an __int128-compatible type in user code. For the rest of this comment, I will refer to "passing i128" by value/reference for brevity, but keep in mind that I am actually referring to "passing i128, or a struct containing i128". As a baseline, I believe i128 is *basically* layout-compatible *except* for its alignment. This means if you're Clever and ensure it's properly aligned, you can still pass it by-reference. Some very bespoke code can do this, but just handing out the type for people to do arbitrary FFI with is obviously a Bad Idea. User code has 2 relevant tools to try to build a Better i128: * `repr(align)` - manually (over-)aligns a struct * `repr(transparent)` - gives a wrapper struct the underlying ABI (kind) of the type it wraps, specifically for making newtyped integers have the integer ABI! Since i128's only problem is that it's under-aligned, we can use `repr(align)` to make what I believe is a layout-compatible version: ``` #[repr(C, align(16))] pub struct layout_only_ffi_i128(i128); ``` layout_only_ffi_i128 can be used to interoperate with C-code that use __i128 *if passed by-reference*, because only layout matters there. Unfortunately, this type is not *ABI*-compatible because wrapping an integer in a struct changes its *kind*. But ok, that's what `repr(transparent)` exists to fix: ``` #[repr(transparent, align(16))] pub struct layout_only_ffi_i128(i128); ``` error[[E0692]](https://doc.rust-lang.org/stable/error-index.html#E0692): transparent struct cannot have other repr hints Sadly, this is illegal. This makes a sort of sense: integer types have to be handled in a fairly bespoke way in an ABI, and so the compiler needs to *properly* understand and support them. It doesn't really make sense to let user code just randomly define its own integer types and then also say "and make it have the native platform ABI for this integer type". You can't "do the native thing" for a thing you don't natively understand! [Truncated] ... // Hack for LLVM expectations for ABI on windows. This is used by the // `#[win64_128bit_abi_hack]` attribute recognized above ``` This is beyond the scope of my knowledge. ---- [0]: attempt 1 to fix this in llvm (merged, backed out "temporarily", dead): https://reviews.llvm.org/D28990 [1]: attempt 2 to fix this in llvm (blocked, dead): https://reviews.llvm.org/D86310 [2]: clang __int128 lowering: https://github.com/llvm-mirror/clang/blob/916645c2637f8c80948eedcdea02258d0a6f79f5/lib/AST/ASTContext.cpp#L1734-L1738 [3]: libc blocked on exposing some core aarch64 linux types: https://github.com/rust-lang/libc/issues/2524 [4]: integers have a special "kind" in ABIs, and can't be emulated: https://gankra.github.io/blah/rust-layouts-and-abis/#abi [5]: explanation and definition of layout/abi disinction: https://gankra.github.io/blah/rust-layouts-and-abis/#the-anatomy-of-a-type [6]: C array ABI weirdness: https://godbolt.org/z/4bb8T7shb [7]: mocked out aarch64 linux getcontext: https://play.rust-lang.org/?version=stable&mode=debug&edition=2021&gist=b3f594736d683fc32d8dbd2ac05b1d8d [8]: win64_128bit_abi_hack: https://github.com/rust-lang/compiler-builtins/blob/4f7ca0a343314e7c534be194a06ef78734243884/src/macros.rs#L369 username_9: It's hard to be 100% without looking further, but there are two main, mostly-orthogonal, concerns: * in-memory layout (alignment and field offsets) * this one is pretty easy, "just" make sure to put the correct value in `i128_align` instead of doing this kind of complex calculation: https://github.com/rust-lang/rust/blob/c42d846add941a26bd254911e16f02c4a3f9346f/compiler/rustc_target/src/abi/mod.rs#L137-L142 * changing our alignment of `{i,u}128` will result in correct call ABI behavior when a `struct` containing `{i,u}128` fields is passed via a stack copy (`byval`) or by indirection * however, there is perhaps a risk here if we assume our alignment is the same LLVM would use for that type, without asking LLVM, which could get us in trouble if we use it for e.g. "LLVM `struct` padding elision" in `rustc_codegen_llvm` * by-value call ABI (single `{i,u}128`, `#[repr(transparent)]` wrapper, maybe larger `struct`s?) * IIUC it sounds like LLVM `i128` is fine for anything passed *in registers* (after all, best we could do is explicitly split the 128-bit value into 2 64-bit values or 4 32-bit values, before LLVM) * but what happens for stack passing? (the "copying" kind we never use for Rust call ABIs) * if we end up using `byval` explicitly (e.g. for a larger `struct`), then all LLVM sees is size+alignment, so we should be able to fully control that (this is the "stack copy" case mentioned earlier) * if we just run out of registers but still use plain LLVM scalar arguments, we could be in a situation where LLVM uses less stack alignment than necessary (and this is #65111 territory again *sigh*) * for targets we already use `byval` on, it might be as easy as treating `{i,u}128` as an aggregate and forcing it to go through the `byval` path * for targets we don't already use `byval` on, we probably need to seriously consider making stack-passing *always* explicit on our end, and make skipping `byval` for some cases a `rustc_codegen_llvm` optimization username_2: Clang just `alloca i128, align 16`s their `__int128`s and copies the value over to the allocas. All i128s seem to always go through at least one such alloca. However, when passing `__int128`s by value I believe clang will still might produce stack layout that's not compatible with gcc. E.g. you can see https://godbolt.org/z/v737MzM8Y that one way or another clang will push onto stack a `__int128` aligned to 8 bytes, while GCC goes through the effort to introduce some padding with `sub rsp, 8`s. A fully correct fix would be to fix the `data-layout` string(s). Today we require that our `data-layout` in target specifications matches exactly what LLVM thinks it should be, but that limitation seems somewhat artificial. The only downside is that we'd have to maintain them and maybe stand up some infrastructure to remind us to verify them when LLVM's ideas change. username_8: Ok to help move along concerns of "so what are the actual ABIs?" I have written a tool that can take a function signature and generate C and Rust code on both sides and checks that they FFI'd to eachother properly: https://github.com/username_8/abi-checker Here we can see attempting to FFI with u128/i128 fails on x64 linux with both pass-as-args and pass-in-struct if you perturb the alignment with some extra values inbetween the u128's: https://github.com/username_8/abi-checker/runs/5427444966?check_suite_focus=true#step:4:4653 ([detailed failures](https://github.com/username_8/abi-checker/runs/5427444966?check_suite_focus=true#step:4:2267). Otherwise Rust seems to classify and pass the value correctly, and appears to classify it properly for sys v. x64 macos is similar, but only fails out on the pass-in-struct case: https://github.com/username_8/abi-checker/runs/5427445107?check_suite_focus=true#step:4:4435 I also have windows CI but msvc doesn't define `__uint128_t` (as discussed in this thread) so those two tests just fail to compile altogether. username_8: Ok so I tested out my new tool (which now has even more tests) on aarch64 via: * My coworker running it on an m1 mac * Me running it via termux on an android phone Much to my surprise **both of these platforms completely pass the suite**! Which is to say, u128/i128 in rust are 100% ABI compatible with both major aarch64 platform's `__uint128_t` and `__int128_t` definitions (I pray to bejeezus there isn't some evil happening where `__int128` and `__int128_t` are ever different). Specifically, for whatever reason rustc (or rather, llvm) accurately understands and believes that u128/i128 are aligned to 16 on this platform, and that is indeed what the platform's standard C ABI is too: https://godbolt.org/z/E6Tfjxdfj ----- So actually for the original motivating problem, __int128 being part of aarch64 system ABIs, isn't actually a problem! We can just happily do FFI with it today! Hooray!! username_2: Yes, the data layout of all of the the aarch64 targets supported by rust has the prerequisite `-i128:128` component within it. username_8: Oh nice! This problem is so historically jank that I am a little paranoid about this... it would still be nice to actually confirm FFI works on all our aarch64 platforms. I don't *expect* any platform's C to randomly overwrite the arch default but....... username_8: memory must be aligned on a 16-byte boundary. Unfortunately, at least on x64 linux it looks like clang and gcc disagree on whether an `__attribute__((aligned(16)))` struct is *actually* ABI-compatible with __int128: https://godbolt.org/z/TPYfaf4vz And indeed if I define ``` #[repr(C, align(16))] pub struct my_i128 { low: i64, high: i64, } ``` And try to do the same emulated->native call from Rust to C (x64 linux GCC), the values don't get passed correctly and argument 5 gets corrupted. 😭 username_8: I've added this attempted-emulation to ABI-checker as `sysv_i128_emulation` https://github.com/username_8/abi-checker failing on linux (gcc): https://github.com/username_8/abi-checker/runs/5500611507?check_suite_focus=true#step:4:4639 passing on macos (clang): https://github.com/username_8/abi-checker/runs/5500611774?check_suite_focus=true#step:4:4399 lol computers are so bad username_8: OK SO THESE TYPES ARE JUST COMPLETELY BUSTED AND CURSED FOR FFI literally just on x64 linux, if you ask clang and gcc to do ``` void i128_val_in_0_perturbed_small( uint8_t arg0, __int128_t arg1, __int128_t arg2, __int128_t arg3, __int128_t arg4, float arg5 ); ``` They will disagree on the ABI and fail to properly pass the values once they start needing to be passed in memory. ``` Test ui128::c::c_calls_c::i128_val_in_0_perturbed_small failed! test 53 input 3 field 0 mismatch caller: [82, 0A, 12, E0, 01, 0C, 32, 7A, 42, F1, EA, 23, 4D, 3C, 2B, 0A] callee: [42, F1, EA, 23, 4D, 3C, 2B, 0A, 82, 0A, 12, E0, 01, 0C, 32, 7A] ``` It looks like they're literally pushing them eightbytes in different orders, but I think that's actually an artifact of me using the same value for all the inputs. In other tests it looked like they just disagreed on whether these values needed to be aligned when pushed (gcc wants them aligned, clang wants them packed). So this type is fake on msvc and isn't consistently handled on linux. A pretty fuckin' bad type to use on x64! It's basically *only* usable for by-ref stuff, but we don't currently have a usable notion for that. username_8: Sure am fucking glad I made an objective FFI validator cuz I believed a lot of absolute lies coming into this! username_5: Wow what a disaster.^^ Are there bugreports for GCC/clang? At least one of them is violating the x86_64 Linux ABI but I guess they will disagree on who. ;) username_2: Clang/LLVM is. I tried to fix it in https://reviews.llvm.org/D28990 and then there's a latter attempt still open at https://reviews.llvm.org/D86310. My view is that clang/LLVM needs to stomach that some breakage will happen. Or we'll continue living with clang being wrong indefinitely. username_8: Yeah up until this point I thought clang *did* handle it right because it added some extra metadata on top of llvm. I was surprised to learn that it *still* doesn't do the right thing even with its changes. 😱 username_8: @username_2 can you confirm that I'm understanding you right? That your patches to llvm would have fixed clang as well? This is not the impression I got from earlier discussion in this thread and your PR to llvm. In your PR you note that "this is what clang does" and "for the 64-bit targets 128-bit alignment is necessary for correct FFI with Clang-generated code, which generates code with assumption that i128 is 128-aligned always." This suggests that the llvm patch was *only* trying to make frontends using i128 directly interoperate with clang, but if your patches actually fixed the issue I'm seeing then your patches were actually changing the ABI of clang itself (and making it easier for other front-ends to match that new ABI as well). Which was it? username_2: clang lowers `__int128` to `i128` for arguments that would end up on the stack: ```llvm ; __int128 banana(uint8_t a0, __int128 a, __int128 b, __int128 c, __int128 d); declare dso_local { i64, i64 } @banana( i8 noundef zeroext, i64 noundef, i64 noundef, i64 noundef, i64 noundef, i128 noundef, i128 noundef) #1 ``` So yes, the patch would have corrected the ABI definition for clang as well as for all the other frontends that produce `i128` parameter types in LLVM IR. username_2: Perhaps the most straightforward way to double-check would be to apply this patch locally (I doubt there should be any conflicts) and try running the patched clang against your test suite again. username_8: progress report: just landed conditionally-available typedefs for __int128 and friends in libc on many major aarch64 platforms: https://github.com/rust-lang/libc/pull/2719 So at least on the platform where we have the right def and it's part of like the hardware vendor's ABI we can do FFI. I am sadly forced to agree with @username_2's assessment that the correct fix is *still* to fix upstream LLVM, given that clang's hack only half works and is still busted ABI-wise.
typelevel/cats
163720647
Title: Add Leibniz equality Question: username_0: This might fall a bit outside the standard "only the things you need" toolkit, but having a notion of Leibniz equality makes handling type equalities much easier and is useful for some pure functional idioms. Scalaz's implementation is more powerful to account for subtyping, but a simpler version might be the following ```scala import scala.language.higherKinds import scala.language.implicitConversions /** * `Teq[A, B]` is a witness that the types `A` and `B` are equal. It is more * powerful than the standard `A =:= B` since it offers the ability to * convert types appearing within other structures. */ sealed abstract class Teq[A, B] { def subst[F[_]](fa: F[A]): F[B] final def apply(a: A): B = Teq.witness(this)(a) final def andThen[C](next: B Teq C): A Teq C = next.subst[A Teq ?](this) final def compose[C](prev: C Teq A): C Teq B = prev andThen this final def from: B Teq A = this.subst[? Teq A](Teq.refl) final def lift[F[_]]: F[A] Teq F[B] = subst[Lambda[X => F[A] Teq F[X]]](Teq.refl) } object Teq { /** * The only real value of `Teq` is the statement `A Teq A`. */ implicit def refl[A]: A Teq A = new Teq[A, A] { def subst[F[_]](fa: F[A]): F[A] = fa } /** * A `Teq` immediausername_0y furnishes a coercion function. */ implicit def witness[A, B](t: A Teq B): A => B = t.subst[A => ?](identity) /** * We can convert `A Teq B` to a `A =:= B` witness via substitution. */ implicit def scalaEq[A, B](t: A Teq B): A =:= B = t.subst[A =:= ?](implicitly[A =:= A]) } ``` Answers: username_1: I think both `Leibniz username_0: I'm a little curious about Lizkov and the subtyping parameters around Leibniz. Also, any thoughts on `TEq` as the name? The other option I thought about was `Is`. username_2: I've found these to be useful, :+1: As for naming, I would prefer `Leibniz` and `Liskov` if only to provide access to the general Google-able principle username_0: I suppose I'm not sure how that particular bikeshed is painted in Cats. `A Is B` reads nicely to me and I like the infix, but `Leibniz` is also fine in my opinion. Is it sufficient to just document the technical term? username_0: I'll turn this into a PR to make the discussion a bit more concrete. username_0: @non, to be clear, I think this is a valuable shed to paint in a project like this and I definiusername_0y defer to those familiar with the style and spirit of the project! Status: Issue closed username_3: Added in #1178.
MacsiDigital/laravel-zoom
750866940
Title: Create Registrant Question: username_0: <!-- DO NOT THROW THIS AWAY --> <!-- Fill out the FULL versions with patch versions --> - API Version: 4.0 - Laravel Version: 8.0 - PHP Version: 7 We have this in the documentation $registrant = Zoom::registrant()->make([...]); Can't find this anywhere in the code and can't make a registrant. I was wondering if the documentation was from a previous version and there was a new way to create a registrant. Thanks. Status: Issue closed Answers: username_0: I used Zoom::WebinarRegistrant() instead.
LeetCode-Feedback/LeetCode-Feedback
1008232940
Title: 1293. Shortest Path in a Grid with Obstacles Elimination Question: username_0: <!-- Note - Any content mention below in `<!-- ->` blocks are just comments to help you fill-up the issue. It won't be visible in the actual issue after you click on submit. --> #### Your LeetCode username username_0 #### Category of the bug - [ ] Question - [ ] Solution - [ ] Language - [ X] Missing Test Cases #### Description of the bug <!-- A clear and concise description of what the bug is. --> Currently, DFS + Memoization is accepted as a "solution". This is not correct. #### Code you used for Submit/Run operation <!-- Please make sure you wrap your code with ``` tags. Otherwise we may reject your request. --> ``` class Solution: def shortestPath(self, grid: List[List[int]], k: int) -> int: visited, m, n = set(), len(grid), len(grid[0]) @cache def dfs(i, j, k): key = (i, j) if (k < 0 or i >= m or i < 0 or j >= n or j < 0 or key in visited): return math.inf if (i == m - 1 and j == n - 1): return 0 k -= grid[i][j] visited.add(key) minSteps = min(dfs(i + 1, j, k), dfs(i - 1, j, k), dfs(i, j + 1, k), dfs(i, j - 1 , k)) visited.remove(key) return 1 + minSteps ans = dfs(0, 0, k) return -1 if ans == math.inf else ans ``` #### Language used for code Python #### Expected behavior <!-- A clear and concise description of what you expected to happen in contrast with what actually happened. --> The above solution should not be accepted. #### Screenshots <!-- If applicable, add screenshots to explain your issue. --> <img width="1048" alt="Screen Shot 2021-09-27 at 10 35 00 AM" src="https://user-images.githubusercontent.com/58147810/134929687-0e5748cc-9d1c-440d-af81-251a03788fe0.png"> <img width="1438" alt="Screen Shot 2021-09-27 at 10 35 56 AM" src="https://user-images.githubusercontent.com/58147810/134929863-0e0b5ddc-da5d-4ba5-900c-80d4ecbb1701.png"> #### Additional context <!-- Add any other additional context about the bug. --> This is my sample test case: ``` [[0,0,0,0,0,0,0,0,0,0],[0,1,1,1,1,1,1,1,1,0],[0,1,0,0,0,0,0,0,0,0],[0,1,0,1,0,1,0,0,1,0],[0,1,0,1,0,1,0,0,1,0],[0,1,0,1,0,1,0,0,1,0],[0,1,0,1,0,1,0,0,1,0],[0,1,0,1,0,1,0,0,1,0],[0,1,0,1,0,1,0,0,1,0],[0,1,0,1,0,1,0,0,1,0],[0,1,0,1,0,0,0,0,1,0],[0,1,0,1,1,1,1,1,1,0],[0,0,0,1,0,0,0,0,1,0]] 1 ``` Answers: username_1: Hi @username_0 Thank you for reaching out to us. I've relayed this issue to our team to investigate. Status: Issue closed
qooxdoo/qooxdoo
459814981
Title: New website look and feel - design brief Question: username_0: With the impending 6.x release, we would like to create a new website with a fresh look and feel - we're going to outsource the work to a professional designer on a paid basis, and so we need to create a design brief for him to work from. This issue is to help us form that design brief, and we welcome opinions and thoughts from everybody, not just core team members. The goal at the end of this issue is to have a design which is primarily used for the public website, and serves as promotional brochure website as well as hosting documentation, API viewer, demo apps, etc. We have some design work already done a while ago (generously donated by @cajus) and while it's been successfully used in the implementation of `qx serve`, further work has stalled. The design could form a basis for the new designs, or it might not. Here's a screenshot of this design <img width="400" alt="Screenshot 2019-06-24 at 11 30 26" src="https://user-images.githubusercontent.com/129568/60012336-96fbca80-9673-11e9-933d-d6483f514922.png"> Answers: username_0: Personally, I quite like the design above - it hits a lot of my boxes for new modern look; the logo is OK, and is quite clever in how (once you know that is's for Qooxdoo) it ties in with the name. IMHO what's missing is a style guide for content generally and a design for the home page. For example, would we want a long home page? IMHO long pages require sub navigation so you can see where you are, but this can be designed in nicely and then there's those sexy parralax-type effects. The real question for the home page is not so much the design, but what do we want on it? Here's some ideas off the top of my head: * Build rich, Single Page Applications using OO development, No CSS or HTML required * Wide range of UI components * Powerful Server I/O communication * Cross platform, built in ES6-everywhere (including server and mobile) * 100% Javascript including tooling and user interface testing * Over 10 years of development username_1: I have been working with flutter.dev (for writing a mobile app) over the last week and they have a pretty simlilar 'narrative' to ours they also say you should be able to write your application in a single language ... without css/xml/dom ... pretty interesting to read their stuff :) flutter could often be replaced by qooxdoo directly ... username_0: Looking at the flutter.dev website, it is a nice clean look and reminds me of the google material theme. And thinking about design, it would be very nice if the new website should echo the design ethos of the new UI appearance - which AFAICR we wanted to base on @username_1 's material theme https://github.com/username_1/qx-osparc-theme (is this still the plan?) username_0: What I love about Qooxdoo is that it brought desktop development to the browser - after decades as a C++ and Java developer, being dropped into native javascript was like having a lobotomy. OO class design is a really strong feature, and on the odd occasion that I've looked at other "OO" class libraries they just don't compare. So I guess that one feature I'd like to see promoted is that it is a high quality engineering tool of classes and libraries - and at the other extreme it's easy to get to grips with and well documented. username_2: Hi. By the moment I just have comment about the new logo design, I love that «Q», however IMHO a kind of «</>» could reflect better the code nature than an «X». Of course, I'm guessing that pro designer will to propose some changes, but well, I'm just saying. :) username_3: Maybe mention somehow that developing with Qooxdoo looks more like Class inheritance vs Prototypal. We have Interfaces, Abstract Classes, Mixins. username_0: I've started summarising the brief here: https://github.com/qooxdoo/qooxdoo/wiki/New-Website-Brief username_4: Maybe we can advertise it as the "hidden gem" - since it is almost never in the javascript news, but has been powering desktop-grade applications for over 10 years. We should also add a section on history in the docs that explains the origins as an inhouse-project at 1&1 and its transformation into a foundation, with important milestones. It helps that[ the demos of all releases are still there](https://github.com/qooxdoo/qooxdoo.github.io) and [still work](http://www.qooxdoo.org/0.6.7/apiviewer/index.html) !!! username_0: Also, in no particular order: * Advanced property mechanism with binding and forms. * Baked in support for Promises * Webfonts * Package manager * Annotations * Automatic Memory Management username_5: On the old website we had a section real life examples. Could we get that back on the new page? username_4: @username_5 This could be made part of the docs (easier to update) and then the website links to it. username_0: Nuts and bolts, construction site, scaffolding etc. Please no cute mascot! username_4: Maybe some imagery in this direction? https://ya-webdesign.com/explore/svg-gear-machine/ The "Gear" imagery is already present in the current logo. username_0: Here's some sample designs from Adam (the designer) - this is just for the logos at this stage, the idea being that the logo sets the scene for the design. What do we think? :) [3logoDesigns.pdf](https://github.com/qooxdoo/qooxdoo/files/3526096/3logoDesigns.pdf) username_6: I don't know if you're voting, for me I liked the first one. username_7: The designer must try again. I like the proposo from the start in this thread : https://user-images.githubusercontent.com/129568/60012336-96fbca80-9673-11e9-933d-d6483f514922.png The name space is `qx`, so **Q** with **x** in middle is supper cool. username_0: Here's some design's from @username_8 to consider (thanks very much Luca!) [Proposal-A.pdf](https://github.com/qooxdoo/qooxdoo/files/3526920/Proposal-A.pdf) [Proposal-A1.pdf](https://github.com/qooxdoo/qooxdoo/files/3526921/Proposal-A1.pdf) [Proposal-B.pdf](https://github.com/qooxdoo/qooxdoo/files/3526922/Proposal-B.pdf) username_5: For me it‘s a - the black one username_1: Some things to consider * the logo should work well in black and white too (think font) * if there is a wide variation of width in some design features, these elements will tend to get lost at small sizes which makes to logo hard to use ... (Proposal-A) username_8: If you guys want to, please provide feedback in terms of approach (which option, or a combination, other logos, etc.) and usage, etc. We (Ice Tea Group / Wisej) would be very happy to contribute our graphic designer time to qooxdoo's redesign logo and web site layout. username_4: As to the last three proposals - I don't like the font - it won't stand the test of time. Too much of a "shampoo" font. username_0: Please see attached for updates on the design - the main document is design4-1.pdf, and we asked for some variations on the resolution of logos see what it looks like at different sizes (especially low res) and to look at slightly less humorous/frivolous iconography. [design4-1.pdf](https://github.com/qooxdoo/qooxdoo/files/3656887/design4-1.pdf) [logo-faceVariations.pdf](https://github.com/qooxdoo/qooxdoo/files/3656891/logo-faceVariations.pdf) ![logoResolutions](https://user-images.githubusercontent.com/129568/65674217-179bdf80-e044-11e9-92f5-b8bdc0410c25.jpg) username_0: We're looking for logos of companies that use Qooxdoo, either as developers or as end-users - obviously the better known the name, the more we'd like to list it front and centre on the new home page. We also want to be able to show that Qooxdoo has a long standing following and is used to build substantial apps, so part of the redesign of the website is to show off a page listing projects that have been built using Qooxdoo. If you feel that you would like to list your project or customer, please let us know - you could either list the details here, or email one of the core team. You should make sure that you have the permission of the company to have the logo on our website, and all submissions would be gratefully received. Thanks! username_6: Here at unidev, we're using qooxdoo as a front end for our GPS tracking solution, for mobile and desktop. Here's the logos (horizontal and vertical) in png format, our website: www.unidev.com.tn My best regards, [image: UD logo Horizontal.png] [image: UD logo Upside Down.png] username_5: ![Produktlogo_VZAweb_Logo+Schrift_PNG-big_BHe](https://user-images.githubusercontent.com/1053029/68862846-7994cf00-06ee-11ea-8ec4-aaf4d199ca79.png) ![Firmenlogo_SWK_Logo_mittig_1](https://user-images.githubusercontent.com/1053029/68862854-7d285600-06ee-11ea-865e-50cdf2efc931.jpg) username_9: You may use [Ergobyte](https://www.ergobyte.gr/portal/ergobyte/en/home)'s ![ergobyte](https://user-images.githubusercontent.com/1220774/68868845-e1541580-0700-11ea-9d43-e37a61bc19c4.png) username_5: @username_6 : Images are missing? username_6: They were embedded in the email, now they're attached as .png files. username_5: Seems that attachment of emails arew not delivered. I added them via web interface username_1: ![](https://www.username_1.ch/assets/Customer/Logo/_resampled/CroppedFocusedImageWzI1NiwyMDAsInkiLDI4XQ/logo-litecom.png) username_1: [](!https://www.username_1.ch/assets/Customer/Logo/oe-small.png) username_1: ![](https://www.username_1.ch/assets/Customer/Logo/_resampled/CroppedFocusedImageWzI1NiwyMDAsInkiLDI4XQ/logo-trustx.png) username_1: ![](https://www.upc.ch/etc/designs/lgi-pe-etlem/media/logos/horizon-logo_upc_small_hidpi.png) username_1: ![](https://www.strom.ch/themes/custom/strom/logo.svg) username_0: ![gsa](https://user-images.githubusercontent.com/129568/68870805-646e6f00-06f3-11ea-9d45-0340d85689f8.png) ![intentor](https://user-images.githubusercontent.com/129568/68870806-65070580-06f3-11ea-93b5-5123caa01fa2.png) ![smilevision](https://user-images.githubusercontent.com/129568/68870807-65070580-06f3-11ea-8e11-51b785fba33a.png) ![spar](https://user-images.githubusercontent.com/129568/68870809-65070580-06f3-11ea-84c8-8540af8971dc.png) username_0: I feel a carousel coming on... 😀 username_0: not yet, but it is getting closer. Design mockups are being done now, when they're settled we'll post them here and then get straight on to html/css version for previewing. username_1: ![](https://www.hin.ch/wp-content/uploads/2016/03/HIN_nur_logo_klein.png) username_6: thumbs up username_1: ![](https://www.fmi.ch/img/fmi_logo.svg) username_10: ![plaidcloud-a](https://user-images.githubusercontent.com/165186/68879562-3d9e4180-06cf-11ea-97f6-7d339f9fe346.png) username_5: Would'nt it not be nice to have the URLs to Websites of the logos included as link of the logo? username_6: Here at unidev, we're using qooxdoo as a front end for our GPS tracking solution, for mobile and desktop. Here's the logos (horizontal and vertical) in png format, our website: www.unidev.com.tn ![UD logo Horizontal](https://user-images.githubusercontent.com/2532330/68883104-a528b100-0710-11ea-86d4-2159202e9bbf.png) ![UD logo Upside Down](https://user-images.githubusercontent.com/2532330/68883107-a5c14780-0710-11ea-89a2-17d80738d17c.png) username_2: ![upn_header](https://user-images.githubusercontent.com/579563/68892268-36376200-06e8-11ea-8ef7-f2562e0caf34.png) https://upnfm.edu.hn/ username_4: Hi everyone - it would be great if you could add information on your public-facing or internal apps to #9814 ! Please feel free to advertise your product this way and for qooxdoo it is a way of showcasing its usefulness, which will in turn increase the community! Thank you. username_11: Our website: www.digisoftph.com ![image](https://user-images.githubusercontent.com/2255528/70334821-19bebd80-1881-11ea-9123-e2a60a5510b8.png) username_12: I realized some of the logos are somewhat similar to superuser.com logo: ![superuser](https://cdn.sstatic.net/Sites/superuser/img/apple-touch-icon.png?v=0ad5b7a83e49) Has this been mentioned? Is the logo decided? username_1: ![image](https://user-images.githubusercontent.com/429279/79753938-09cae600-8317-11ea-8d29-c9ef5cc22cea.png) another similar one username_4: Website is launched, any new issue should be created and dealt with here: https://github.com/qooxdoo/website/issues Status: Issue closed
blue-oil/blueoil
518206981
Title: Bump up docker base image to Ubuntu bionic Question: username_0: What prevent it? - DE10-Nano glibc version, will be resolved by new OS image Answers: username_1: This is a sample error log run on `Ubuntu16.04 DE10-Nano` using binary converted by `Ubuntu 18.04 docker base image`. ``` [2019-07-30T06:04:51Z] ---begining of stderr---   | [2019-07-30T06:04:51Z] Traceback (most recent call last):   | [2019-07-30T06:04:51Z] File "testing_code.py", line 24, in <module>   | [2019-07-30T06:04:51Z] print(t.run_library('./lib_arm_fpga.so', './000_images_placeholder:0.npy', './337_output:0.npy'))   | [2019-07-30T06:04:51Z] File "testing_code.py", line 10, in run_library   | [2019-07-30T06:04:51Z] nn.load(library)   | [2019-07-30T06:04:51Z] File "/root/automated_testing/nnlib.py", line 29, in load   | [2019-07-30T06:04:51Z] self.lib = ct.cdll.LoadLibrary(libpath)   | [2019-07-30T06:04:51Z] File "/usr/lib/python2.7/ctypes/__init__.py", line 440, in LoadLibrary   | [2019-07-30T06:04:51Z] return self._dlltype(name)   | [2019-07-30T06:04:51Z] File "/usr/lib/python2.7/ctypes/__init__.py", line 362, in __init__   | [2019-07-30T06:04:51Z] self._handle = _dlopen(self._name, mode)   | [2019-07-30T06:04:51Z] OSError: /lib/arm-linux-gnueabihf/libm.so.6: version `GLIBC_2.27' not found (required by ./lib_arm_fpga.so)   | [2019-07-30T06:04:51Z]   | [2019-07-30T06:04:51Z] ---end of stderr--- ``` Ref https://buildkite.com/blueoil/dlk-test/builds/55#7622fea2-61fc-4f7d-9ac0-49aa5016f5f1
Azure/azure-quickstart-templates
311518699
Title: No VMSS ? Elasticsearch, X-Pack, VM Scale Sets and Managed Disks Question: username_0: Hi, I am confusing why it doesn't exist VMSS in this solution but just 3 master nodes ( as well as other resources related to VM such as NIC, VNET, NSG ) and 1 availability set? I am looking for a real example vmss with Elastic Stack especially ES. ![image](https://user-images.githubusercontent.com/4621560/38355344-e7f14782-38bc-11e8-89e1-032158f47919.png) --------------------MESSAGE FROM ADMIN, DELETE BEFORE SUBMITTING---------------------- Sorry to hear you had a bad experience with one of the templates :worried: But, in case you're just asking a question, we're happy to help. You can also check if the question might already have been asked here https://github.com/Azure/azure-quickstart-templates/issues?utf8=%E2%9C%93&q=is%3Aissue We've created an outline of recommended sections to fill out that will help make this Pull Request awesome! --------------------MESSAGE FROM ADMIN, DELETE BEFORE SUBMITTING---------------------- [Template Name goes here](Template link goes here) ### Issue Details ### Repro steps (*if necessary, delete otherwise*) 1. 2. 3. 4. 5. Answers: username_0: To prepare a deployment of E stack, I developed an ARM template with VMSS ( master node and data node in two vmss ) will update it, when I'll also deploy E stack with my own ARM template
akkeris/cli
281189559
Title: actions output is [object Object] Question: username_0: when running the addons:create command the output of actions is 0=[object Object] 1=[object Object] etc. This can be fixed by just calling stringify on that. Here is my output if it makes things a bit more clear. ``` λ aka addons:create alamo-postgresql:hobby -a yearbookapi-ca-dev-us [1460c82] === Addon alamo-postgresql-language-8328 Provisioned === alamo-postgresql-language-8328 (83194d59-47d9-4829-85f1-e191a7be8d4c) actions 0=[object Object], 1=[object Object], 2=[object Object], 3=[object Object], 4=[object Object], 5=[object Object], 6=[object Object], 7=[object Object], 8=[object Object], 9=[object Object] addon_service id=01bb60d2-f2bb-64c0-4c8b-ead731a690bc, name=alamo-postgresql app id=bb3146df-fc00-4279-8641-d525e6a17d0e, name=yearbookapi-ca-dev-us config_vars DATABASE_URL=postgres://uee3f7246:[email protected]:5432/pga0302163 created_at 2017-12-11T22:21:49.825Z id 83194d59-47d9-4829-85f1-e191a7be8d4c name alamo-postgresql-language-8328 plan id=50660450-61d3-2c13-a3fd-d379997932fb, name=alamo-postgresql:hobby provider_id alamo updated_at 2017-12-11T22:21:49.825Z web_url https://ui.appkit.sh/#/apps/yearbookapi-ca-dev-us ```
gjanders/SplunkVersionControl
812091805
Title: Unable to Connect to Github Repository Question: username_0: I'm not able push the objects into my remote git repository. I have configured SSH on my github account as well as on the splunk server machine. I'm able to clone the repository via the git bash. Below are the configurations: ![image](https://user-images.githubusercontent.com/75374701/108520341-9310c780-72f0-11eb-9271-a3f9de2f1666.png) I'm getting the below error while saving the configurations on the splunk: ![image](https://user-images.githubusercontent.com/75374701/108520434-b176c300-72f0-11eb-821b-f1b92e26e407.png) **Gitusername**: samules8 **Splunk Username**: marlonsamuels **Windows System Username**: Jackson Need more info on the below points: ![image](https://user-images.githubusercontent.com/75374701/108520840-1c27fe80-72f1-11eb-8600-6e663e82ae14.png) Answers: username_1: Is this in Windows or Linux? username_1: And assuming it is windows is Splunk running as the system user? And if so is the SSH key in the system user directory, something like : C:\windows\system32\config\systemprofile\.ssh username_2: This is on Windows system. Yes splunk is running as the sytem user. Below are the files under C:\windows\system32\config ![image](https://user-images.githubusercontent.com/8663552/108585163-d6a91700-736c-11eb-8761-f23053f333ea.png) username_1: Is there a hidden directory with .ssh in there under c:\windows\system32\config\systemprofile\.ssh ? If not than when the system user attempts to run the git clone it does not have access to the required SSH keys. On my test server I put the id.rsa (or your key file name) under: c:\windows\system32\config\systemprofile\.ssh With appropriate permissions, if you can get a command window running as the system user you can also test it that way... username_0: Windows Thanks & Regards, <NAME>, +919505446851 username_0: Hii, As per the documentation available in Splunk base we have followed the below steps. 1. We have crated a GitHub repository. 2. We have generated SSH key for windows server with the help of below link (https://www.atlassian.com/git/tutorials/git-ssh) and add to shh agent to our key. 3. We have added SSH key to GitHub account under SSH keys. 4. We have cloned the git repository by using SSH key into our windows server and it is successfully cloned. 5. After that we have open the DataInputs from splunk web and Selected the splunkversion control backup option and gave the below parameters. Name : splunk_backup srcURL : https://localhost:8089 srcUsername : Splunk Enterprise username srcPassword : <PASSWORD> gitTempDir : local System path where git repository cloned gitRepoURL : Git repository SSH URL 6.After clicking Next button, we are getting the below error **Encountered the following error while trying to save: Failed to validate the git repo URL, stdout of 'b''', stderr of 'b'[email protected]: Permission denied (publickey).\r\nfatal: Could not read from remote repository.\n\nPlease make sure you have the correct access rights\nand the repository exists.\n''** ![image](https://user-images.githubusercontent.com/75374701/108849983-d0cf6200-7608-11eb-9004-1e1d3d60504e.png) So, can you please let us know the above steps and parameters are valid are not to establish the connection, also below .ssh folder created under C:\Users\username_0\.ssh ![image](https://user-images.githubusercontent.com/75374701/108849909-b5fced80-7608-11eb-81db-a38bd2e8d93f.png) So, please help us to fix this. username_1: So your key is under C:\Users\username_0\.ssh But Splunk runs as the system user? Perhaps try copying the SSH key to directory: C:\windows\system32\config\systemprofile\.ssh So in the above directory which you may need to create, add the SSH key. You can also test by opening a command prompt and using git clone but you must open the command prompt as the *system* user, not your current user... Also, this tends to be easier on Linux if you have any Linux machines...(or docker images running splunk/linux) username_0: Hi Thanks for your help. As you said we have copied all the files from C:\Users\username_0\.ssh\* to C:\windows\system32\config\systemprofile\.ssh\* After that we are able save the created input successfully but all the apps are coming to local directory where given under gitTemDir but not able to find in GitHub repo. Please help us ,Is anything we have missed. Do we need to give any git commands under Data Inputs(Splunk Version Control Backup)? username_1: Check the splunk version control backup log. It will be in your _internal index or the logs directory. You might need to set an email address and name for git, I added that into the new version of the app. Let me know if the logs aren't clear username_0: Hii, We have done setup git_name and git_email. After that we seen the logs that shows waring like this, **i="splun_na" git checkout master or git pull failed, stdout is 'b''' stderrout is 'b"error: pathspec 'master' did not match any file(s) known to git\nThere is no tracking information for the current branch.\nPlease specify which branch you want to merge with.\nSee git-pull(1) for details.\n\n git pull <remote> <branch>\n\nIf you wish to set tracking information for this branch you can do so with:\n\n git branch --set-upstream-to=<remote>/<branch> master\n\n"'. Wiping git directory** username_1: I found on Windows you might have to remove the temporary directory manually. Did you initialise the git repository? I e. It contains 1 or more files? username_0: I removed temporary directory. I initialized the git repository and it contains 2 files. username_1: Ok same error? Do you have a master branch in the git repo? username_0: No Error, but still we are unable to find any Splunk app folders under git repo.(for backup) ![image](https://user-images.githubusercontent.com/75374701/108990527-fcac1f80-76bc-11eb-9030-faade4780f0c.png) But Splunk app folders getting into local backup folder. ![image](https://user-images.githubusercontent.com/75374701/108990564-0afa3b80-76bd-11eb-91a5-26bd389e5cde.png) How can take backup into git? username_1: What does the splunkversioncontrol_backup.log say? username_0: Hi , Can you please find the attached file for log Details If it is possible, can you please connect with us for 10 mins so that we will complete this implementation. username_1: I cannot see any files attached, I'm currently on leave and have very little time to investigate the issues on my open source projects
ContinuumIO/anaconda-issues
358404849
Title: R kernel keeps dying in Jupyter notebook Question: username_0: <!-- Thanks for opening an issue! A few things to keep in mind: - This issue tracker is for issues with installing Anaconda or Miniconda or with packages built by Anaconda, Inc. like Anaconda Navigator. It is also a place to request packages or other enhancements of the Anaconda Distribution. For conda issues, you should open an issue at https://github.com/conda/conda/issues For conda-build issues, you should open an issue at https://github.com/conda/conda-build For an issue with a particular conda-forge package, you should open an issue on the corresponding feedstock: https://github.com/conda-forge --> ### Actual Behavior Kernel keeps dying and restarting. <!-- What actually happens? --> R kernel keeps dying in Jupyter notebook every time I open a new kernel. No problem with the python kernel though. ### Expected Behavior Proper functioning of the kernel <!-- What do you think should happen? --> Proper functioning of the kernel ### Steps to Reproduce <!-- Steps to reproduce issue. --> ##### Anaconda or Miniconda version: ##### Operating System: ##### `conda info` <!-- between the ticks below, paste the output of 'conda info' --> ``` ``` ##### `conda list --show-channel-urls` <!-- between the ticks below, paste the output of 'conda list --show-channel-urls' --> ``` ``` Answers: username_1: Closing. Please try to fill out all requested information when you open bug reports. Status: Issue closed
rainlab/userplus-plugin
208281707
Title: "plugins" folder inside the rainlab\userplus plugin folder with copies of both User and Location plugins? Question: username_0: This is a question... I installed the UserPlus plugin yesterday so to check it out, and at first the installation didn't work, but once I installed the Location plugin and tried installing UserPlus once again, it eventually worked. But now one thing strikes me as kind of weird. Inside the rainlab\userplus plugin folder, there's another plugin folder with what apparently seems to be full copies of both the Location and the User plugins. Is it supposed to be like that? If so, why? Answers: username_1: Same here: ![snimek obrazovky 2017-03-26 v 16 41 06](https://cloud.githubusercontent.com/assets/374917/24332215/0bf054a8-1243-11e7-8c34-2c583528806a.png) username_2: Hmm, this might have been a temporary error with the build process on the October marketplace. I'm guessing this plugin probably needs to be rebuilt on the marketplace to clean out those artifacts. @username_3 any input? username_3: Confirmed. This odd nesting exists in the marketplace source version. Looks like composer might be bringing them in as deps. username_3: This is fixed! Status: Issue closed
Matrix20085/cptSmallTools
1048764106
Title: Snapshot fails on boxes that are turned on for configuration Question: username_0: Error: Server task failed: Can not complete this operation in the current state Pretty sure it is because the box is not fully shutdown before that command runs. Need to put in sleep or a loop to check current state then run Answers: username_1: This still looks to be an issue.
flutter/flutter
609291833
Title: [tool_crash] FileSystemException: Failed to set file modification time, OS Error: Operation not permitted, errno = 1 Question: username_0: ## Command ``` flutter run ``` ## Steps to Reproduce 1. ... 2. ... 3. ... ## Logs FileSystemException: Failed to set file modification time, OS Error: Operation not permitted, errno = 1 ``` #0 _File.setLastModifiedSync (dart:io/file_impl.dart:473:7) #1 ForwardingFile.setLastModifiedSync (package:file/src/forwarding/forwarding_file.dart:62:55) #2 ForwardingFile.setLastModifiedSync (package:file/src/forwarding/forwarding_file.dart:62:55) #3 _DefaultPub.get (package:flutter_tools/src/dart/pub.dart:247:25) #4 FlutterCommand.verifyThenRunCommand (package:flutter_tools/src/runner/flutter_command.dart:743:17) <asynchronous suspension> #5 FlutterCommand.run.<anonymous closure> (package:flutter_tools/src/runner/flutter_command.dart:657:33) <asynchronous suspension> #6 FlutterCommand.run.<anonymous closure> (package:flutter_tools/src/runner/flutter_command.dart) #7 AppContext.run.<anonymous closure> (package:flutter_tools/src/base/context.dart:150:29) #8 _rootRun (dart:async/zone.dart:1184:13) #9 _CustomZone.run (dart:async/zone.dart:1077:19) #10 _runZoned (dart:async/zone.dart:1619:10) #11 runZoned (dart:async/zone.dart:1539:10) #12 AppContext.run (package:flutter_tools/src/base/context.dart:149:18) #13 FlutterCommand.run (package:flutter_tools/src/runner/flutter_command.dart:647:20) #14 CommandRunner.runCommand (package:args/command_runner.dart:197:27) #15 FlutterCommandRunner.runCommand.<anonymous closure> (package:flutter_tools/src/runner/flutter_command_runner.dart:339:21) <asynchronous suspension> #16 FlutterCommandRunner.runCommand.<anonymous closure> (package:flutter_tools/src/runner/flutter_command_runner.dart) #17 AppContext.run.<anonymous closure> (package:flutter_tools/src/base/context.dart:150:29) #18 _rootRun (dart:async/zone.dart:1184:13) #19 _CustomZone.run (dart:async/zone.dart:1077:19) #20 _runZoned (dart:async/zone.dart:1619:10) #21 runZoned (dart:async/zone.dart:1539:10) ``` ``` [✓] Flutter (Channel master, 1.18.0-9.0.pre.62, on Mac OS X 10.15.4 19E287, locale pt-BR) • Flutter version 1.18.0-9.0.pre.62 at /Users/devfacil/Desktop/flutter • Framework revision b7586d873a (2 hours ago), 2020-04-29 12:44:01 -0400 • Engine revision 2db3276573 • Dart version 2.9.0 (build 2.9.0-3.0.dev 726d3c7725) [!] Android toolchain - develop for Android devices (Android SDK version 29.0.3) • Android SDK at /Users/devfacil/Library/Android/sdk • Platform android-29, build-tools 29.0.3 • Java binary at: /Applications/Android Studio.app/Contents/jre/jdk/Contents/Home/bin/java • Java version OpenJDK Runtime Environment (build 1.8.0_212-release-1586-b4-5784211) ! Some Android licenses not accepted. To resolve this, run: flutter doctor --android-licenses [✓] Xcode - develop for iOS and macOS (Xcode 11.3.1) • Xcode at /Applications/Xcode.app/Contents/Developer • Xcode 11.3.1, Build version 11C504 [Truncated] geocoder-0.2.1 get_ip-0.4.0 google_maps_flutter-0.5.26+4 google_sign_in-4.4.4 google_sign_in_web-0.9.1 image_picker-0.6.5+2 in_app_purchase-0.3.2+2 move_to_background-1.0.0+1 multi_image_picker-4.6.6+rc.1 onesignal_flutter-2.4.0 path_provider-1.6.7 path_provider_macos-0.0.4+1 shared_preferences-0.5.7 shared_preferences_macos-0.0.1+7 shared_preferences_web-0.1.2+4 sqflite-1.3.0 stripe_payment-1.0.7 video_player-0.10.9 video_player_web-0.1.2+1 webview_flutter-0.3.20+2 Answers: username_1: please make sure that the plugins you are using are compatible, some on your list definitely aren't. could you please run your full `flutter run -d macOS --verbose` and a reproducible minimal code sample Thank you username_2: /cc @jonahwilliams Status: Issue closed
LukeGarrigan/hacktober-race
502524364
Title: Have some way to tell your type speed! Question: username_0: I know that in most type racer games you'd have words per minute but for this, I don't think it'll be all that suitable. I'd like some ideas on this, I've only really got one: - Characters per minute/per second Answers: username_1: I'm working on it. Status: Issue closed
ossrs/srs
1040875401
Title: > Invalid usage, please read https://github.com/ossrs/srs-bench/tree/feature/rtc#publisher-for-live-or-rtc Question: username_0: 您好,我是參照那個文檔做壓測的沒錯,參照Player for Live最後兩行, ./objs/srs_bench -sr webrtc://172.16.31.10/live/testfw?eip=192.168.3.11 -nn [100|10] 後面[100|10]是代表我輸入過的數值有100跟10, 不是所有拉流都失敗, 大概二分之一左右, 這個場景理論上沒限制WebRTC流一定要按照文檔用ffmpeg推流吧? 另外我寫了shell script用for迴圈背景去執行DVR, 同樣的次數連線失敗的量跟上面壓測的量也差不多, 連線成功的確定有紀錄到正確的影像, 還是您是指測試的WebRTC url不能用eip指定對外ip? _Originally posted by @username_0 in https://github.com/ossrs/srs/issues/2696#issuecomment-955882148_<issue_closed> Status: Issue closed
zyedidia/micro
151321864
Title: Repainting console window black/white Question: username_0: after terminating micro any console window running cmd.exe or msys/bash is painted with black backround and white text. The console windows are configured for grey background and white text. See: http://abload.de/img/unbenannteesd9.png Answers: username_1: Quick note to add this bug (ie on exit from micro the screen colours are set to black backround and white text) also occurs when running micro.exe version 1c2b815 on Windows 10 in a PowerShell console window. username_2: This should be fixed now. Just make sure to update your version of tcell. Status: Issue closed
DoctorMcKay/node-steamcommunity
114139343
Title: Suggestion to add new feature for getting average price for an item Question: username_0: If this possible to be implemented , it would be nice to have a feature to get average price of a certain item over a 'n' piriod of time . Becuase I don't think there is much use for the get maden prices becuase the market isn't stable enough. Thanks! Status: Issue closed Answers: username_1: Steam only gives us median prices. You can calculate the average of those medians, but we can't get the raw averages.
tensorflow/tensorflow
910438613
Title: tf.contrib.framework.assign_from_variables in Tensorflow 2 Question: username_0: I'm running an inference model that used tf1, and I've changed the code to tf2 but I can't find a proper tf2 alternative to contrib.framework.assign_from_variables anywhere. I've tried tf.compat.v1 but still the same issue. Answers: username_0: Just found out the solution ; just use tf_slim username_1: @username_0 , Please refer this link for info on contrib.[Link](https://www.tensorflow.org/guide/migrate).It helps.Thanks username_1: @username_0 , Please feel free to move this issue to closed status if the issue has resolved.Thanks!
platformio/platform-atmelsam
399871621
Title: Support for MoteinoM0 & CurrentRanger SAMD21G boards Question: username_0: Please add support for MoteinoM0 and CurrentRanger. Latest definitions for Arduino can be found here: https://github.com/username_0/username_0.github.io/tree/master/MoteinoCore Thanks! Status: Issue closed Answers: username_1: Please re-test with upstream version https://docs.platformio.org/en/latest/platforms/atmelsam.html#upstream username_0: Thank you, how can I do that? username_1: 1. Install PlatformIO IDE => https://platformio.org/platformio-ide 2. Open PlatformIO IDE > PIO Home > Platforms > Click on "Advanced Installlation" > Paste `https://github.com/platformio/platform-atmelsam.git` 3. Close PIO Home tab and open it again, click on "New Project", search for your boards. Does it work? username_0: Oh, I'm sorry, I didn't have a chance to try it out yet. I was just so busy. I trust you that it works, you do great work, thanks for all your support 👍 username_1: Thanks! ;)
rust-vmm/vmm-sys-util
524293526
Title: Block other signals while a registered signal handler is running Question: username_0: Just noticed that `register_signal_handler` from `signal.rs`: https://github.com/rust-vmm/vmm-sys-util/blob/472614c3a21231f6a0c3696fab83fb34d6592ade/src/signal.rs#L191, lacks a way to specify to a signal handler what signals to be blocked while it is running. To be specific, I am referring to a way to initialize the `sa_mask` field of a `struct sigaction`, when calling `register_signal_handler`. An example of my usecase is here [0]. Are you fond of a way to achieve this with the current implementation? If it couldn't be achieved with the current implementation, I can gladly contribute with a PR. [0] https://www.gnu.org/software/libc/manual/html_node/Blocking-for-Handler.html#Blocking-for-Handler. Answers: username_1: To give a bit more context, we are currently working on consuming vmm-sys-util in Firecracker and we need this functionality to consume the signal module as well. Status: Issue closed
katsu-skillup/protospace-35898
941432597
Title: 【依頼】ProtoSpace挙動確認 Question: username_0: # URL https://protospace-35898.herokuapp.com/ Answers: username_1: @username_0 さん、お待たせいたしました。ご提出ありがとうございます‼︎ 問題なく正常に動作することが確認できました! LGTMです✨  おめでとうございます🎉 次はいよいよ最終課題ですね!今まで学習したことを活かして進めていきましょう(^^) 別途、最終課題の案内をチャンネルに送らせて頂きます。 今しばらくお待ちください🍀
shershen08/vue-masonry
249852584
Title: [Vue warn]: Failed to resolve directive: masonry-tile Question: username_0: I've been trying to get vue-masonry to work in various Vuejs projects. I've tried blank projects from the CLI (webpack, pwa, and nuxt) and it seems like it's always failing with this error `[Vue warn]: Failed to resolve directive: masonry-tile` `Failed to resolve directive: masonry` I've followed the steps verbatim and for some reason vue js cannot use the directive... Also, I tried the demo `vue-masonry-demo` and it's also giving me the same issues. Any idea? Can't seem to figure this one out! Answers: username_1: Hey @username_0 I had the same problem, I ended up installing "vue-masonry": "^0.10.5" version and it's working for me. Try npm install [email protected]. Best, Will username_0: Thanks @username_1, that solved my problem as well. @username_3, any ideas on what might be going on with the latest version? Thanks for the plugin btw, very handy. username_2: https://github.com/username_3/vue-masonry/issues/26 Status: Issue closed
benrhughes/todotxt.net
520569775
Title: Minimise to system tray - error "Hotkey already in use" Question: username_0: I pinned to taskbar Todotxt.net and check the option "Minimise to system tray", after minimise and click on pin shortcut show error "Hotkey already in use" **log** [11/10/2019 9:15:28 AM] Error Global HotKey Registered System.ApplicationException: Hotkey already in use at Client.HotKey.RegisterHotKey() at Client.HotKey..ctor(ModifierKeys modifierKeys, Keys key, IntPtr windowHandle) at Client.HotKeyMainWindows..ctor(Window window, ModifierKeys modifierKeys, Keys key)
DataDog/dd-trace-js
353361757
Title: A request to move @airbnb/node-memwatch to devDependencies in package.json Question: username_0: Any chance of `@airbnb/node-memwatch` being moved into a `devDependency` in `package.json`? As it's just increasing CI times for us - especially sad as it is a dev dependency. It also causes errors that get ignored in some environments (not a real bother or your problem). I have resorted to using the `--no-optional` flag for `npm install` and `npm prune` and that does avoid the problem. Using that flag feels a little heavy handed for just your package though. Although this isn't your problem, the errors I get look like: ..\src\memwatch.cc(19): fatal error C1083: Cannot open include file: 'sys/time.h': No such file or directory [...\node_modules\@airbnb\node-memwatch\build\memwatch.vcxproj] ..\src\heapdiff.cc(291): warning C4244: 'argument': conversion from 'unsigned __int64' to 'v8::SnapshotObjectId', possible loss of data [...\node_modules\@airbnb\node-memwatch\build\memwatch.vcxproj] ..\src\heapdiff.cc(303): warning C4244: 'argument': conversion from 'unsigned __int64' to 'v8::SnapshotObjectId', possible loss of data [...\node_modules\@airbnb\node-memwatch\build\memwatch.vcxproj] ..\node_modules\nan\nan_new.h(208): warning C4244: 'argument': conversion from 'unsigned __int64' to 'double', possible loss of data (compiling source file ..\src\heapdiff.cc) [...\node_modules\@airbnb\node-memwatch\build\memwatch.vcxproj] I understand that you might not want dev package restores to fail, but increasing everyone else's bother isn't great either. Anyway, just a humble request :) Answers: username_1: That was actually a mistake on my part. I moved it from `devDependencies` to `optionalDependencies` because it doesn't install on Node <8 but I forgot that `optionalDependencies` are not installed only in dev. I'll simply remove the dependency completely and install it as a build step. Status: Issue closed
yorikvanhavre/barcelona-pavillion
735479610
Title: import ifc to Freecad or Blender Question: username_0: 17:46:14 Opening 'C:/Users/alberts/Downloads/Pabellon de Barcelona.ifc'... Traceback (most recent call last): File "<string>", line 1, in <module> File "C:\Users\alberts\Documents\FreeCAD_0.19.22284_x64_LP_12.1.6_PY3QT5-WinVS2015\Mod\Arch\importIFC.py", line 289, in insert ifcfile = ifcopenshell.open(filename) File "C:\Users\alberts\Documents\FreeCAD_0.19.22284_x64_LP_12.1.6_PY3QT5-WinVS2015\bin\lib\site-packages\ifcopenshell\__init__.py", line 67, in open raise IOError("Unable to open file for reading") <class 'OSError'>: Unable to open file for reading for FreeCAD and Blender: Traceback (most recent call last): File "C:\Users\alberts\AppData\Roaming\Blender Foundation\Blender\2.90\scripts\addons\blenderbim\bim\operator.py", line 131, in execute ifc_importer.execute() File "C:\Users\alberts\AppData\Roaming\Blender Foundation\Blender\2.90\scripts\addons\blenderbim\bim\import_ifc.py", line 347, in execute self.load_file() File "C:\Users\alberts\AppData\Roaming\Blender Foundation\Blender\2.90\scripts\addons\blenderbim\bim\import_ifc.py", line 1393, in load_file self.file = ifcopenshell.open(self.ifc_import_settings.input_file) File "C:\Users\alberts\AppData\Roaming\Blender Foundation\Blender\2.90\scripts\addons\blenderbim\libs\site\packages\ifcopenshell\__init__.py", line 67, in open raise IOError("Unable to open file for reading") OSError: Unable to open file for reading location: <unknown location>:-1
pcm-dpc/COVID-19
734417233
Title: Chiarimento sulla definizione di "isolamento domiciliare" Question: username_0: <!-- Grazie per aver mostrato interesse in questo repo. Da sapere prima di aprire issue: - Questo lavoro è svolto con puro scopo informativo - Saranno accettate issue relativi a: - richieste di informazione - richiesta di dati - suggerimenti di miglioramento dei documenti, metadati del repo, dashboard grafica, ecc Prima di aprire un'issue: - Fate una breve ricerca degli issue già creati (incluso issue già chiusi) per convincersi che la vostra richiesta non è stata ancora fatta - Sarà apprezzato una spiegazione di come la vostra richiesta beneficierà l'interesse pubblico di questo repo - Cancellare il testo sopra --> **Tipo di richiesta**: <!-- eliminare tutti non-relativi --> richiesta di informazione ## Riassunto G<NAME>, ad oggi 2 novembre 2020 risulta che ben 357.288 persone sono in "isolamento domiciliare", ma vorrei avere alcuni chiarimenti, anche perché ho vinto questa posizione pure io, in quanto una mia alunna è stata trovata positiva. Gli "isolati" sono solo i "positivi" asintomatici o lievi o post-sintomatici oppure contemplano anche quelli che, come me, sono stati a contatto con "positivi" e sono ancora in attesa di un tampone? Anzi, in generale, chi come me è in questa posizione deve ritenersi in "quarantena" o in "isolamento fiduciario"? A me risulta che dovrei essere in "isolamento fiduciario", ma in comunicazioni dell'ASL di miei colleghi (a me ancora la ASL non ha mandato nulla) risulta lo status di "quarantena". Spero possiate chiarire questi miei dubbi. Intanto, vi ringrazio infinitamente sia per l'eventuale risposta sia per il lavoro che state svolgendo da mesi. <!-- Scrivi qui il riassunto della richiesta. --> ## Interesse pubblico <!-- Spiegate in che modo questa richiesta beneficierà l'interesse pubblico --> Maggior chiarezza, IMHO Answers: username_1: Buongiorno, la voce "isolamento domiciliare" dovrebbe essere riferita ai casi di positivi (asintomatici o lievi o post-sintomatici). Infatti il valore del "Totale attualmente positivi" (quindi riscontrati con tampone), è dato dalla somma dei "Ricoverati con sintomi" + "Terapia intensiva" + "Isolamento domiciliare". username_2: buongiorno, grazie della spiegazione quindi in siolamento domiciliare ci sono dentro solo i positivi (quindi tutti loro hanno fatto il tampone con esito positivo) e quindi ci sono asintomatici e sintomatici? è possibile inserire due colonne che suddividano i sintomatici dagli asintomatici? ve lo dico perchè ad oggi risulta impossibile fare dei confronti con marzo aprile (perchè a marzo ed aprile facendo molti meno tamponi gli asintomatici positivi non venivano calcolati...) e il dato che guardiamo oggi e che ci fa tanta paura quando cresce in realtà è un dato ben diverso da quello di aprile dove gli asintomatici non venivano intercettati... e secondo me per fare un confronto e delle proiezioni sarebbe più corretto pur mantenendoli nel totale dei positivi suddividere le categorie in sintomatici e asintomatici.... non è logico? ovviamente se possibile!!!! grazie per l'attenzione! elena username_2: buongiorno, grazie della spiegazione quindi in isolamento domiciliare ci sono dentro i positivi asintomatici e sintomatici( tutti loro hanno fatto il tampone con esito positivo) è possibile inserire due colonne che suddividano i sintomatici dagli asintomatici? ve lo dico perchè ad oggi risulta impossibile fare dei confronti sensati con marzo e aprile (perchè a marzo ed aprile facendo molti meno tamponi agli asintomatici, gli asintomatici positivi erano molti meno) e il dato che guardiamo oggi non possiamo compararlo... Oggi il numero dei nuovi casi ci fa tanta paura quando cresce in realtà è un dato ben diverso da quello di aprile dove gli asintomatici non venivano intercettati con regolarità.. e secondo me per fare un confronto e delle proiezioni sarebbe più corretto pur mantenendoli nel totale dei positivi suddividere le categorie in isolamento domiciliare sintomatici e isolamento domiciliare asintomatici.... non è logico? secondo me avrebbe più senso... ovviamente se possibile!!!! gli asintomatici si sono pericolosi per contagio ma non andranno sicuramente a riempire gli ospedali... mentro un sintomatico potrebbe dover andare in ospedale e finire in terapia intensiva... il sintomatico esce dal sistema se muore o se guarisce (magari pasando prima da un ospedale o da una terapia intensiva)... l'asintomatco esce dal sistema senza intasare l'spedale... fanno parte di due "flussi" diversi... grazie per l'attenzione! elena username_3: Se può essere utile le suggerisco, per tentare un confronto con la prima ondata, di far riferimento al rapporto percentuale tra il Delta giornaliero dei tamponi e i nuovi_positivi. Può verificare direttamente il risuoltato su [questo grafico](https://covid19.zappi.me/casiTestati/). username_4: Ma il valore e' cumulato o giornaliero? perche' ci sono tanti valori negativi facendo la differenza col giorno prima. username_2: Ci sono entrambi, sia i totali che sono quelliche vengono dati, sia i delta giornalieri, sono negativi perché diminuiscono i totali.. Inviato da iPhone username_3: Non sono sicuro di interpretare correttamente la tua domanda: **totale_casi**: è un indicatore cumulato: l'ultimo valore pubblicato rappresenta il totale dei casi registrati. **nuovi_positivi**: è un indicatore giornaliero: l'ultimo valore pubblicato rappresenta il valore giornaliero del giorno di pubblicazione. Matematica vorrebbe che la somma di tutti i nuovi_positivi pubblicati fossse uguale all'ultimo totale_casi pubblicato. Così non è perchè Protezione Civile applica dei correttivi su casi passati intervenendo su **totale_casi**. A causa di questi interventi di correzione, può capitare nei periodi in cui il numero di novi_positivi è molto basso, che la differenza tra l'ultimo **totale_casi** e quello del giorno precedente possa risultare negativa. Se questo rappresenta un problema, la soluzione sta nell'utilizzare l'indicatore **nuovi_positivi** invece della differenza tra **totale_casi**. Nel caso chiedessi altro, ignora pure quanto qui scritto. username_4: La mia domanda e' riferita solo al campo "isolamento_domiciliare". La descrizione dati non menziona nulla sul fatto che il dato sia giornaliero o meno. Ho notato che quando il dato e' giornaliero viene menzionato, quindi assumo sia totale. Matematica volel che la differenza del cumulato col giorno precedente non sia mai negativa. username_4: Mi riferisco solo al campo "isolamento_domiciliare", secondo me e' cumulato perche' piu' o meno vedo valori sempre crescenti. username_3: Provo ad esporti la mia versione, prendila in considerazione solo se per qualche motivo ti convince. La definizione di isolamento_domiciliare è: "persone in isolamento domiciliare" Dall'isolamento domiciliare si entra e dopo qualche settimana si esce auspicabilmente guariti. Un indicatore che rappresenta quindi una pila o un buffer, come ricoverati_con_sintomi, terapia_intensiva o anche altri. Un giorno in cui si verificano più casi di uscita dall'isolamento domiciliare che nuovi ingressi, il valore di quell'indicatore si abbasserà. Quel giorno la differenza con il valore del giorno precedente sarà negativa. Forse può essere d'aiuto vederlo graficamente: [Evoluzione epidemia COVID-19 in Italia](https://covid19.zappi.me/andamento/). username_4: Grazie per il plot. Quindi una sorta di numero "netto" di persone in isolamento in totali. Probabilmente aggiungendo decessi e guariti giornalieri si ottiene il totale che dici? Grazie per la intrepretazione username_3: Penso che mancherebbero ancora ricoverati_con_sintomi e terapia_intensiva, oppure sommandoli totale_ospedalizzati. Applico questo raggruppamento gerarchico, non è l'unico, ma è basilare. Vedi se ti può essere utile: ricoverati_con_sintomi + terapia_intensiva = **totale_ospedalizzati** totale_ospedalizzati + isolamento_domiciliare = **totale_positivi** totale_positivi + dimessi_guariti + deceduti = **totale_casi** Forse, anche in questo caso, potrebbe aiutare vedere i numeri su una rappresentazione tabellare ordinata: [IT covid table](https://covid19.zappi.me/table/) username_2: buonasera, qualcuno sa dirmi come mai salta il link ai dati? https://raw.githubusercontent.com/pcm-dpc/COVID-19/master/dati-andamento-nazioale/dpc-covid19-ita-andamento-nazionale.csv da qualche giorno, fino a ieri, saltava solo il collegamento al file csv ma i dati c'erano, ora la pagina va in errore mi da questo messaggio [image: image.png] è possibile ripristinare? grazie! username_3: Hai digitato l'URL in maniera non perfetta: **nazioale** al posto di **nazionale**, correggi e risolverai il problema. username_0: @username_2 Bella domanda! Il comando IMPORTDATA oggi ha ripreso a funzionare, almeno nel mio file, ma in altri file non funziona ancora. Puoi provare ad usare uno script, ma anche quello funziona e non funziona. Non ho capito se il problema è legato a GitHub oppure a qualche aggiornamento lato Google. Come script, puoi usare un codice come questo: ``` //Displays an alert as a Toast message function displayToastAlert(message) { SpreadsheetApp.getActive().toast(message, "⚠️ Alert"); } //Imports Italy CSV file at a URL into the Google Sheet function importCSVItaly() { var csvUrl = "https://raw.githubusercontent.com/pcm-dpc/COVID-19/master/dati-andamento-nazionale/dpc-covid19-ita-andamento-nazionale.csv"; var csvContent = UrlFetchApp.fetch(csvUrl).getContentText(); var csvData = Utilities.parseCsv(csvContent); var sheet = SpreadsheetApp.getActiveSheet(); sheet.getRange(1, 1, csvData.length, csvData[0].length).setValues(csvData); displayToastAlert("The CSV file was successfully imported."); } ``` username_2: ma è proprio l'url che non è raggiungibile... se entri nel link vedi lo stesso errore anche tu? 404: Not Found Il giorno gio 10 dic 2020 alle ore 18:57 <NAME> < username_0: @username_2 L'url [https://raw.githubusercontent.com/pcm-dpc/COVID-19/master/dati-andamento-nazionale/dpc-covid19-ita-andamento-nazionale.csv](https://raw.githubusercontent.com/pcm-dpc/COVID-19/master/dati-andamento-nazionale/dpc-covid19-ita-andamento-nazionale.csv) funziona benissimo. Quello scritto da te, come ti è stato detto sopra, presenta un errore di stampa. username_2: grazie! adesso va.. ho dovuto aggiungere delle colonne ;) Il giorno gio 10 dic 2020 alle ore 21:38 <NAME> <
rickselby/conelanders
164665364
Title: Lean caching Question: username_0: Check that whatever is cached is all that is needed for the view; don't be caching large amounts of data. Question: You can ``load`` things on models; can they be ``unload``ed? Some models are needed for calculations, but not needed for views. Answers: username_0: The driver caches are far too large - this needs addressing. ### Assetto Corsa: - [ ] Championship Cars - [ ] Constructor Standings: Championship - [ ] Constructor Standings: Event - [ ] Constructor Standings: Event Summary - [ ] Driver Standings: Championship - [ ] Driver Standings: Event - [ ] Driver Standings: Event Summary - [ ] Event: Driver IDs - [ ] Results: Fastest Laps - [ ] Results: For Driver - [ ] Results: For Race - [ ] Results: Lap Chart - [ ] Results: Winner - [ ] Team Standings: Championship - [ ] Team Standings: Event - [ ] Team Standings: Event Summary ### Dirt Rally: - [ ] Driver Points: For Event - [ ] Driver Points: For Season - [ ] Driver Points: Overview - [ ] Driver Points: Overall - [ ] Nation Points: For Event - [ ] Nation Points: For Season - [ ] Nation Points: Overview - [ ] Nation Points: Overall - [ ] Nation Points: Details - [ ] Results: Event Results - [ ] Results: Stage Results - [ ] Results: For Driver - [ ] Results: Season Winner - [ ] Results: Event Winner - [ ] Times: For Event - [ ] Times: For Season - [ ] Times: Overall Status: Issue closed
ninenines/ct_helper
336407424
Title: OTP-21 compatibility Question: username_0: Library uses `random` module and `crypto:rand_uniform`. Answers: username_0: Only when you enable warnings as errors ;) username_1: Yeah but I mean when used with erlang.mk as a dep it's not a problem because erlang.mk disables them for deps. Do you run in an issue because of a different environment or scenario? username_0: The infrastructure I work with depends heavily on rebar3. username_1: Then you probably need to add `app:: rebar.config` at the end of the Makefile in ct_helper and that'll disable warnings as errors for rebar. You'll need to rebuild and commit the rebar.config file it generates. PR welcome once you confirm it works. username_0: Thanks, I'll look into it. Just out of curiosity, are you against updating the code? username_1: I have to make it work for 18+. username_1: I've committed a blank rebar.config to fix this. Status: Issue closed
hashbang/docker-slapd
201220159
Title: Automated Testing Question: username_0: To be perfectly honest, I have literally no idea how this will be done. Answers: username_1: Same as for `hashbang.sh` and `provisor`: this goes away in the userdb move, do we really invest in testing it? username_0: I suppose not, but I believe @lrvick wanted a fully testable infrastructure before moving forward with the userdb move.
TransitFeeds/TransitFeeds-Public
228512632
Title: Barrie Transit Url Changes+Additions Question: username_0: http://www.myridebarrie.ca/gtfs/Google_transit.zip http://www.myridebarrie.ca/gtfs/GTFS_ServiceAlerts.pb http://www.myridebarrie.ca/gtfs/GTFS_TripUpdates.pb http://www.myridebarrie.ca/gtfs/GTFS_VehiclePositions.pb Answers: username_1: Thanks, all updated / added: https://transitfeeds.com/admin/providers/provider.php?p=295 Status: Issue closed
ISG-ICS/cloudberry
212296988
Title: Represent Tweets Using Dots on the Map Question: username_0: So far we are using a heat map to show the Tweets on the map. It is a high-level description of the data. Now if we can show the exact location of each Tweet, we can make the connection between the data and the geolocation closer. A good example by MapD: ![screen shot 2017-03-06 at 16 41 41](https://cloud.githubusercontent.com/assets/12385178/23636764/389961c8-028c-11e7-92f1-d80b57434ac0.jpg) However, since there are too much data to load in the browser, we need use some tricks to avoid crashing the browser. We can start implementing this feature in small areas of the map, e.g. city level. Answers: username_1: @username_0 Good luck and have fun! :-) Status: Issue closed
ionic-team/stencil
965498308
Title: Nested slots mis-ordered when not using Shadow DOM Question: username_0: <!-- NOTE: Before submitting an issue, please consult our docs -> https://stenciljs.com/ --> **Stencil version:** <!-- (run `npm list @stencil/core` from a terminal/cmd prompt and paste output below): --> ``` @stencil/[email protected] ``` **I'm submitting a:** <!-- (check one with "x") --> [x] bug report [ ] feature request [ ] support request => Please do not submit support requests here, use one of these channels: https://stencil-worldwide.herokuapp.com/ or https://forum.ionicframework.com/ **Current behavior:** <!-- Describe how the bug manifests. --> When a parent component's template contains a child component which it passes its slot content into, the slot content can end up mis-ordered compared to the child component's template. I believe the child component's props being based on the parent component's state may be related to the issue. For example, in my use case "State: true" should come before "Hello" but it ends up backwards. (code examples below) ``` Hello State: true ``` If the components re-render, or I turn on Shadow DOM, that fixes the issue (turning on Shadow DOM will not work for my real-world use case) **Expected behavior:** <!-- Describe what the behavior would be without the bug. --> I would expect the order of the text above to be swapped: ``` State: true Hello ``` **Steps to reproduce:** <!-- If you are able to illustrate the bug or feature request with an example, please provide steps to reproduce and if possible a demo --> Here is a demo and reproduction repo: - [Demo](https://mystifying-borg-a6eba2.netlify.app/index.html) - [Minimum reproducible test case repo](https://github.com/cloudfour/stencil-nested-slot-bug-repro) **Related code:** Here are two components I used for a minimum reproducible test case: Parent component: ```tsx [Truncated] <div>State: {this.state.toString()}</div> <slot /> </Host> ); } } ``` Here's how I'm using this component: ```html <my-component> <p>Hello</p> </my-component> ``` **Other information:** <!-- List any other information that is relevant to your issue. Stack traces, related issues, suggestions on how to fix, Stack Overflow links, forum links, etc. --> #2232 sounds related, but not quite the same. That issue deals with issues on re-renders. My issue is with the initial render. Answers: username_1: Hi - I'm running into this exact issue as well. Hope the team has time to look at the fixing PR soon!! username_2: Hey there 👋 Thanks for the detailed summary and minimal reproduction! They make verifying issues _so_ easy and are very much appreciated! I've confirmed this is a bug still in Stencil v2.10.0, and am going to label this issue so it gets ingested for the team to look at further. Thanks again! username_0: Awesome, thanks @username_2 !
OrnisOrbit/fido2020
566934778
Title: Prepare data models Question: username_0: - [ ] Setup an ORM library - [ ] Add the DB connection logic - [ ] Define the DB's schema with ORM models - [ ] Define DB's relations - [ ] Add ORM and DB syncing - [ ] Add logic to populate DB with initial data - [ ] Add password hashing with crypt<issue_closed> Status: Issue closed
Ludeon/RimWorld-ru
258178338
Title: Scratch Question: username_0: Судя по характеру получения раны в игре, соответствует не "царапине", а скорее резано-колотой или рваной ране. Чаще всего при описании поврежений от когтей животных встречал просто слово "рана". Из #546.<issue_closed> Status: Issue closed
ilscipio/scipio-erp
399303208
Title: Database Error Question: username_0: I am getting this error while launching for the first time. org.ofbiz.entity.GenericEntityException: org.ofbiz.entity.GenericDataSourceException: Unable to esablish a connection with the database. (Unable to acquire a new connection from the pool) (Unable to esablish a connection with the database. (Unable to acquire a new connection from the pool)) at org.ofbiz.entity.GenericDelegator.findList(GenericDelegator.java:1603) ~[ofbiz-entity.jar:?] at org.ofbiz.entity.util.EntityQuery.query(EntityQuery.java:526) ~[ofbiz-entity.jar:?] at org.ofbiz.entity.util.EntityQuery.queryList(EntityQuery.java:451) ~[ofbiz-entity.jar:?] at org.ofbiz.service.job.JobManager.reloadCrashedJobs(JobManager.java:395) [ofbiz-service.jar:?] at org.ofbiz.service.job.JobPoller$JobManagerPoller.run(JobPoller.java:223) [ofbiz-service.jar:?] at java.lang.Thread.run(Thread.java:748) [?:1.8.0_191] Caused by: org.ofbiz.entity.GenericDataSourceException: Unable to esablish a connection with the database. (Unable to acquire a new connection from the pool) at org.ofbiz.entity.jdbc.SQLProcessor.getConnection(SQLProcessor.java:263) ~[ofbiz-entity.jar:?] at org.ofbiz.entity.jdbc.SQLProcessor.prepareStatement(SQLProcessor.java:366) ~[ofbiz-entity.jar:?] at org.ofbiz.entity.datasource.GenericDAO.selectListIteratorByCondition(GenericDAO.java:763) ~[ofbiz-entity.jar:?] at org.ofbiz.entity.datasource.GenericHelperDAO.findListIteratorByCondition(GenericHelperDAO.java:140) ~[ofbiz-entity.jar:?] at org.ofbiz.entity.GenericDelegator.find(GenericDelegator.java:1547) ~[ofbiz-entity.jar:?] at org.ofbiz.entity.GenericDelegator.findList(GenericDelegator.java:1583) ~[ofbiz-entity.jar:?] ... 5 more Caused by: java.sql.SQLException: Unable to acquire a new connection from the pool at org.apache.commons.dbcp2.managed.ManagedConnection.updateTransactionStatus(ManagedConnection.java:130) ~[commons-dbcp2-2.2.0.jar:2.2.0] at org.apache.commons.dbcp2.managed.ManagedConnection.<init>(ManagedConnection.java:59) ~[commons-dbcp2-2.2.0.jar:2.2.0] at org.apache.commons.dbcp2.managed.ManagedDataSource.getConnection(ManagedDataSource.java:81) ~[commons-dbcp2-2.2.0.jar:2.2.0] at org.ofbiz.entity.connection.DebugManagedDataSource.getConnection(DebugManagedDataSource.java:52) ~[ofbiz-entity.jar:?] at org.ofbiz.entity.connection.DBCPConnectionFactory.getConnection(DBCPConnectionFactory.java:65) ~[ofbiz-entity.jar:?] at org.ofbiz.geronimo.GeronimoTransactionFactory.getConnection(GeronimoTransactionFactory.java:83) ~[ofbiz-geronimo.jar:?] at org.ofbiz.entity.jdbc.SQLProcessor.getConnection(SQLProcessor.java:260) ~[ofbiz-entity.jar:?] at org.ofbiz.entity.jdbc.SQLProcessor.prepareStatement(SQLProcessor.java:366) ~[ofbiz-entity.jar:?] at org.ofbiz.entity.datasource.GenericDAO.selectListIteratorByCondition(GenericDAO.java:763) ~[ofbiz-entity.jar:?] at org.ofbiz.entity.datasource.GenericHelperDAO.findListIteratorByCondition(GenericHelperDAO.java:140) ~[ofbiz-entity.jar:?] at org.ofbiz.entity.GenericDelegator.find(GenericDelegator.java:1547) ~[ofbiz-entity.jar:?] at org.ofbiz.entity.GenericDelegator.findList(GenericDelegator.java:1583) ~[ofbiz-entity.jar:?] ... 5 more Caused by: java.util.NoSuchElementException: Unable to activate object at org.apache.commons.pool2.impl.GenericObjectPool.borrowObject(GenericObjectPool.java:471) ~[commons-pool2-2.5.0.jar:2.5.0] at org.apache.commons.pool2.impl.GenericObjectPool.borrowObject(GenericObjectPool.java:362) ~[commons-pool2-2.5.0.jar:2.5.0] at org.apache.commons.dbcp2.managed.ManagedConnection.updateTransactionStatus(ManagedConnection.java:127) ~[commons-dbcp2-2.2.0.jar:2.2.0] at org.apache.commons.dbcp2.managed.ManagedConnection.<init>(ManagedConnection.java:59) ~[commons-dbcp2-2.2.0.jar:2.2.0] at org.apache.commons.dbcp2.managed.ManagedDataSource.getConnection(ManagedDataSource.java:81) ~[commons-dbcp2-2.2.0.jar:2.2.0] at org.ofbiz.entity.connection.DebugManagedDataSource.getConnection(DebugManagedDataSource.java:52) ~[ofbiz-entity.jar:?] at org.ofbiz.entity.connection.DBCPConnectionFactory.getConnection(DBCPConnectionFactory.java:65) ~[ofbiz-entity.jar:?] at org.ofbiz.geronimo.GeronimoTransactionFactory.getConnection(GeronimoTransactionFactory.java:83) ~[ofbiz-geronimo.jar:?] at org.ofbiz.entity.jdbc.SQLProcessor.getConnection(SQLProcessor.java:260) ~[ofbiz-entity.jar:?] at org.ofbiz.entity.jdbc.SQLProcessor.prepareStatement(SQLProcessor.java:366) ~[ofbiz-entity.jar:?] at org.ofbiz.entity.datasource.GenericDAO.selectListIteratorByCondition(GenericDAO.java:763) ~[ofbiz-entity.jar:?] at org.ofbiz.entity.datasource.GenericHelperDAO.findListIteratorByCondition(GenericHelperDAO.java:140) ~[ofbiz-entity.jar:?] at org.ofbiz.entity.GenericDelegator.find(GenericDelegator.java:1547) ~[ofbiz-entity.jar:?] at org.ofbiz.entity.GenericDelegator.findList(GenericDelegator.java:1583) ~[ofbiz-entity.jar:?] ... 5 more Caused by: java.sql.SQLException: A read-only user or a user in a read-only database is not permitted to disable read-only mode on a connection. at org.apache.derby.impl.jdbc.SQLExceptionFactory.getSQLException(Unknown Source) ~[derby-10.14.2.0.jar:?] at org.apache.derby.impl.jdbc.Util.generateCsSQLException(Unknown Source) ~[derby-10.14.2.0.jar:?] at org.apache.derby.impl.jdbc.TransactionResourceImpl.wrapInSQLException(Unknown Source) ~[derby-10.14.2.0.jar:?] at org.apache.derby.impl.jdbc.TransactionResourceImpl.handleException(Unknown Source) ~[derby-10.14.2.0.jar:?] at org.apache.derby.impl.jdbc.EmbedConnection.handleException(Unknown Source) ~[derby-10.14.2.0.jar:?] at org.apache.derby.impl.jdbc.EmbedConnection.setReadOnly(Unknown Source) ~[derby-10.14.2.0.jar:?] at org.apache.commons.dbcp2.DelegatingConnection.setReadOnly(DelegatingConnection.java:561) ~[commons-dbcp2-2.2.0.jar:2.2.0] at org.apache.commons.dbcp2.PoolableConnectionFactory.activateObject(PoolableConnectionFactory.java:415) ~[commons-dbcp2-2.2.0.jar:2.2.0] at org.apache.commons.pool2.impl.GenericObjectPool.borrowObject(GenericObjectPool.java:462) ~[commons-pool2-2.5.0.jar:2.5.0] at org.apache.commons.pool2.impl.GenericObjectPool.borrowObject(GenericObjectPool.java:362) ~[commons-pool2-2.5.0.jar:2.5.0] at org.apache.commons.dbcp2.managed.ManagedConnection.updateTransactionStatus(ManagedConnection.java:127) ~[commons-dbcp2-2.2.0.jar:2.2.0] at org.apache.commons.dbcp2.managed.ManagedConnection.<init>(ManagedConnection.java:59) ~[commons-dbcp2-2.2.0.jar:2.2.0] at org.apache.commons.dbcp2.managed.ManagedDataSource.getConnection(ManagedDataSource.java:81) ~[commons-dbcp2-2.2.0.jar:2.2.0] at org.ofbiz.entity.connection.DebugManagedDataSource.getConnection(DebugManagedDataSource.java:52) ~[ofbiz-entity.jar:?] at org.ofbiz.entity.connection.DBCPConnectionFactory.getConnection(DBCPConnectionFactory.java:65) ~[ofbiz-entity.jar:?] at org.ofbiz.geronimo.GeronimoTransactionFactory.getConnection(GeronimoTransactionFactory.java:83) ~[ofbiz-geronimo.jar:?] at org.ofbiz.entity.jdbc.SQLProcessor.getConnection(SQLProcessor.java:260) ~[ofbiz-entity.jar:?] at org.ofbiz.entity.jdbc.SQLProcessor.prepareStatement(SQLProcessor.java:366) ~[ofbiz-entity.jar:?] at org.ofbiz.entity.datasource.GenericDAO.selectListIteratorByCondition(GenericDAO.java:763) ~[ofbiz-entity.jar:?] at org.ofbiz.entity.datasource.GenericHelperDAO.findListIteratorByCondition(GenericHelperDAO.java:140) ~[ofbiz-entity.jar:?] at org.ofbiz.entity.GenericDelegator.find(GenericDelegator.java:1547) ~[ofbiz-entity.jar:?] at org.ofbiz.entity.GenericDelegator.findList(GenericDelegator.java:1583) ~[ofbiz-entity.jar:?] ... 5 more Answers: username_1: Hi Vikas, Can you provide more details about how are you trying to connect to the database? Are you using Derby or any other RDBMS? In case you haven't, I'd recommend to take a look at our [setup guide](https://www.scipioerp.com/community/developer/installation-configuration/configuration/). username_2: Hi Vikas, and welcome to the community. Can you check this thread: https://forum.scipioerp.com/t/cant-start-scipio/34/10 I think your error is exactly the same. Thanks & regards, Paul username_0: Thanks for quick response guys. I think it was a localized error, someone messed up the db config in my VM (its a shared VM). I simply redownloaded the repo based on @username_1's comment and voila.. it works! Status: Issue closed