Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 5
112
| repo_url
stringlengths 34
141
| action
stringclasses 3
values | title
stringlengths 1
1k
| labels
stringlengths 4
1.38k
| body
stringlengths 1
262k
| index
stringclasses 16
values | text_combine
stringlengths 96
262k
| label
stringclasses 2
values | text
stringlengths 96
252k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
276,343
| 8,597,573,009
|
IssuesEvent
|
2018-11-15 19:05:58
|
bounswe/bounswe2018group9
|
https://api.github.com/repos/bounswe/bounswe2018group9
|
opened
|
Event Creation Endpoints Update
|
Platform: Backend Priority: High
|
Event creation endpoint has to be improved because we added new fields to the event data model. We need to make sure that before saving an event we have all the necessary fields created in the database.
|
1.0
|
Event Creation Endpoints Update - Event creation endpoint has to be improved because we added new fields to the event data model. We need to make sure that before saving an event we have all the necessary fields created in the database.
|
priority
|
event creation endpoints update event creation endpoint has to be improved because we added new fields to the event data model we need to make sure that before saving an event we have all the necessary fields created in the database
| 1
|
74,419
| 14,247,514,485
|
IssuesEvent
|
2020-11-19 11:33:59
|
Regalis11/Barotrauma
|
https://api.github.com/repos/Regalis11/Barotrauma
|
opened
|
Bots can't find a path through the docking module.
|
Bug Code
|
Ordering a bot from inside the sub to "wait" at the specified location (the pause with a circle to the right), results in them equipping the dive suit correctly, going out of the sub's airlock, but can't manage to path through the docking port module.
_(the bot had a scooter on it to make it more obvious its trying to swim)_
**[Sometimes they also get stuck by trying to swim through the corner of the dugong's airlock on the outside when attempting to path to the same "wait" location]**

|
1.0
|
Bots can't find a path through the docking module. - Ordering a bot from inside the sub to "wait" at the specified location (the pause with a circle to the right), results in them equipping the dive suit correctly, going out of the sub's airlock, but can't manage to path through the docking port module.
_(the bot had a scooter on it to make it more obvious its trying to swim)_
**[Sometimes they also get stuck by trying to swim through the corner of the dugong's airlock on the outside when attempting to path to the same "wait" location]**

|
non_priority
|
bots can t find a path through the docking module ordering a bot from inside the sub to wait at the specified location the pause with a circle to the right results in them equipping the dive suit correctly going out of the sub s airlock but can t manage to path through the docking port module the bot had a scooter on it to make it more obvious its trying to swim
| 0
|
94,924
| 19,607,332,893
|
IssuesEvent
|
2022-01-06 11:08:45
|
creativecommons/vocabulary
|
https://api.github.com/repos/creativecommons/vocabulary
|
opened
|
Split styles and fonts packages into separate projects
|
🟩 priority: low 🚦 status: awaiting triage ✨ goal: improvement 💻 aspect: code
|
## Problem
The structure of the Vocabulary project and lack of developer continuity has led to maintenance issues blocking the release of `styles` and `fonts` packages. We need to publish new versions of `styles` and `fonts` as the creativecommons.org redesign roadmap progresses.
## Description
The Vocabulary project includes transient dependencies preventing us from building and releasing the code, such as those related to `node-sass` and `storybook`. We have tried to reconcile the dependency issues for the past few months but are reaching the limit of our available resources to continue the struggle.
An alternative approach would be to revert the packages to be in separate, simplified projects, containing only dependencies essential to their function. To the extent possible, we should separate dependencies for interactive documentation from core package dependencies to reduce complexity and adverse interactions. This restructuring will improve the immediate maintainability of the projects while leaving open the issue of providing a holistic documentation/development environment.
## Alternatives
Keep the monorepo structure and continue struggling to restore the project to a healthy state. However, the above restructuring would not preclude any further maintenance on this Vocabulary project but would provide some immediate respite and allow us to move forward with the creativecommons.org redesign.
## Additional context
We hope to launch the new creativecommons.org redesign this month.
|
1.0
|
Split styles and fonts packages into separate projects - ## Problem
The structure of the Vocabulary project and lack of developer continuity has led to maintenance issues blocking the release of `styles` and `fonts` packages. We need to publish new versions of `styles` and `fonts` as the creativecommons.org redesign roadmap progresses.
## Description
The Vocabulary project includes transient dependencies preventing us from building and releasing the code, such as those related to `node-sass` and `storybook`. We have tried to reconcile the dependency issues for the past few months but are reaching the limit of our available resources to continue the struggle.
An alternative approach would be to revert the packages to be in separate, simplified projects, containing only dependencies essential to their function. To the extent possible, we should separate dependencies for interactive documentation from core package dependencies to reduce complexity and adverse interactions. This restructuring will improve the immediate maintainability of the projects while leaving open the issue of providing a holistic documentation/development environment.
## Alternatives
Keep the monorepo structure and continue struggling to restore the project to a healthy state. However, the above restructuring would not preclude any further maintenance on this Vocabulary project but would provide some immediate respite and allow us to move forward with the creativecommons.org redesign.
## Additional context
We hope to launch the new creativecommons.org redesign this month.
|
non_priority
|
split styles and fonts packages into separate projects problem the structure of the vocabulary project and lack of developer continuity has led to maintenance issues blocking the release of styles and fonts packages we need to publish new versions of styles and fonts as the creativecommons org redesign roadmap progresses description the vocabulary project includes transient dependencies preventing us from building and releasing the code such as those related to node sass and storybook we have tried to reconcile the dependency issues for the past few months but are reaching the limit of our available resources to continue the struggle an alternative approach would be to revert the packages to be in separate simplified projects containing only dependencies essential to their function to the extent possible we should separate dependencies for interactive documentation from core package dependencies to reduce complexity and adverse interactions this restructuring will improve the immediate maintainability of the projects while leaving open the issue of providing a holistic documentation development environment alternatives keep the monorepo structure and continue struggling to restore the project to a healthy state however the above restructuring would not preclude any further maintenance on this vocabulary project but would provide some immediate respite and allow us to move forward with the creativecommons org redesign additional context we hope to launch the new creativecommons org redesign this month
| 0
|
11,291
| 3,481,926,277
|
IssuesEvent
|
2015-12-29 19:28:59
|
wp-cli/wp-cli
|
https://api.github.com/repos/wp-cli/wp-cli
|
closed
|
`wp help`: don't show GLOBAL PARAMETERS section by default, add `--global-help` switch
|
scope:documentation scope:framework
|
When using `wp help`, the actual command help usually scrolls out of view because there is a relatively long GLOBAL PARAMETERS section at the end (which is generally not very useful).
Maybe this long section could be replaced with a message like:
Use --global-help to show global parameters
and that switch could be added to the `wp help` command.
|
1.0
|
`wp help`: don't show GLOBAL PARAMETERS section by default, add `--global-help` switch - When using `wp help`, the actual command help usually scrolls out of view because there is a relatively long GLOBAL PARAMETERS section at the end (which is generally not very useful).
Maybe this long section could be replaced with a message like:
Use --global-help to show global parameters
and that switch could be added to the `wp help` command.
|
non_priority
|
wp help don t show global parameters section by default add global help switch when using wp help the actual command help usually scrolls out of view because there is a relatively long global parameters section at the end which is generally not very useful maybe this long section could be replaced with a message like use global help to show global parameters and that switch could be added to the wp help command
| 0
|
662,495
| 22,141,543,643
|
IssuesEvent
|
2022-06-03 07:28:17
|
threefoldtech/tfchain_graphql
|
https://api.github.com/repos/threefoldtech/tfchain_graphql
|
closed
|
Public IP's are not synced properly
|
priority_major
|
On `FarmUpdated` the list of ips should be checked with what is saved in the database.
Also `IPsReserved(u64, Vec<PublicIP>)` and `IPsFreed(u64, Vec<Vec<u8>>),` should mutate the public ips
|
1.0
|
Public IP's are not synced properly - On `FarmUpdated` the list of ips should be checked with what is saved in the database.
Also `IPsReserved(u64, Vec<PublicIP>)` and `IPsFreed(u64, Vec<Vec<u8>>),` should mutate the public ips
|
priority
|
public ip s are not synced properly on farmupdated the list of ips should be checked with what is saved in the database also ipsreserved vec and ipsfreed vec should mutate the public ips
| 1
|
172,302
| 14,356,744,092
|
IssuesEvent
|
2020-11-30 12:01:53
|
JetBrains/gradle-intellij-plugin
|
https://api.github.com/repos/JetBrains/gradle-intellij-plugin
|
closed
|
pluginVerifier doesn't fail when compatibility issues are found
|
documentation duplicate
|
In my project, I have:
```properties
ideaVersion=2019.2.4
pluginSinceBuild=192.5728.98
pluginUntilBuild=203.*
pluginVerifierIdeVersions=2019.2.4,2019.3.5,2020.1.4,2020.2.4,203.5981.114
```
When running the pluginVerifier, I get told that the plugin is compatible with 2020.1.4, 2020.2.4, and 203.5981.114; but that it's not compatible with 2019.2.4, and 2019.3.5. Yet, the task completes and I get a `BUILD SUCCESSFUL` result.
I'd expect the build to fail if any compatibility issues are found though.
CC @hsz — you know the project I'm talking about :)
|
1.0
|
pluginVerifier doesn't fail when compatibility issues are found - In my project, I have:
```properties
ideaVersion=2019.2.4
pluginSinceBuild=192.5728.98
pluginUntilBuild=203.*
pluginVerifierIdeVersions=2019.2.4,2019.3.5,2020.1.4,2020.2.4,203.5981.114
```
When running the pluginVerifier, I get told that the plugin is compatible with 2020.1.4, 2020.2.4, and 203.5981.114; but that it's not compatible with 2019.2.4, and 2019.3.5. Yet, the task completes and I get a `BUILD SUCCESSFUL` result.
I'd expect the build to fail if any compatibility issues are found though.
CC @hsz — you know the project I'm talking about :)
|
non_priority
|
pluginverifier doesn t fail when compatibility issues are found in my project i have properties ideaversion pluginsincebuild pluginuntilbuild pluginverifierideversions when running the pluginverifier i get told that the plugin is compatible with and but that it s not compatible with and yet the task completes and i get a build successful result i d expect the build to fail if any compatibility issues are found though cc hsz — you know the project i m talking about
| 0
|
118,237
| 4,733,304,864
|
IssuesEvent
|
2016-10-19 10:43:10
|
thommoboy/There-are-no-brakes
|
https://api.github.com/repos/thommoboy/There-are-no-brakes
|
closed
|
PRE TESTING - Player walk and idle animation doesnt loop correctly
|
bug Priority Low
|
also jittering when player stops walking
|
1.0
|
PRE TESTING - Player walk and idle animation doesnt loop correctly - also jittering when player stops walking
|
priority
|
pre testing player walk and idle animation doesnt loop correctly also jittering when player stops walking
| 1
|
242,579
| 7,844,511,243
|
IssuesEvent
|
2018-06-19 09:51:57
|
xcat2/xcat-core
|
https://api.github.com/repos/xcat2/xcat-core
|
closed
|
[New xCAT prototype]Implement petitboot plugin to deploy OpenPOWER bare metal servers
|
priority:high sprint2 type:feature
|
- [x] Enhance copycd interface to create and destroy osimage.
- [x] Update dhcp interface to support 209 options for petitboot
- [x] Build petitboot configuration file
- [x] Add clean up method to delete configurations during callback(deploy complete) or clean request( -d option).
- [x] Enhance the abstraction interface for pxe and petitboot
|
1.0
|
[New xCAT prototype]Implement petitboot plugin to deploy OpenPOWER bare metal servers - - [x] Enhance copycd interface to create and destroy osimage.
- [x] Update dhcp interface to support 209 options for petitboot
- [x] Build petitboot configuration file
- [x] Add clean up method to delete configurations during callback(deploy complete) or clean request( -d option).
- [x] Enhance the abstraction interface for pxe and petitboot
|
priority
|
implement petitboot plugin to deploy openpower bare metal servers enhance copycd interface to create and destroy osimage update dhcp interface to support options for petitboot build petitboot configuration file add clean up method to delete configurations during callback deploy complete or clean request d option enhance the abstraction interface for pxe and petitboot
| 1
|
12,066
| 3,251,684,182
|
IssuesEvent
|
2015-10-19 11:14:24
|
WormBase/website
|
https://api.github.com/repos/WormBase/website
|
closed
|
dpy-17 search is incorrectly redirected
|
Bug Report HelpDesk Under testing Webteam
|
*Submitted from the feedback form on the WormBase website.*
<p> A recent thread on the forum (http://forums.wormbase.org/index.php?topic=2631) uncovered some issues with the search function to identify balancers that cover a gene of interest. It appears that the data are incomplete and/or the search function not working as expected. For example, searching 'dpy-17' under Rearrangement returns hT2 only. However, the mT1 summary page explicitly lists dpy-17 as a positive marker.</p>
**Reported by:** Haro******** (smit******************)
**Submitted from:** <a target="_blank" href="http://www.wormbase.org//tools/support">/tools/support</a>
**Browser:** Mozilla/5.0 (Macintosh; Intel Mac OS X 10_8_5) AppleWebKit/600.5.17 (KHTML, like Gecko) Version/6.2.5 Safari/537.85.14
|
1.0
|
dpy-17 search is incorrectly redirected -
*Submitted from the feedback form on the WormBase website.*
<p> A recent thread on the forum (http://forums.wormbase.org/index.php?topic=2631) uncovered some issues with the search function to identify balancers that cover a gene of interest. It appears that the data are incomplete and/or the search function not working as expected. For example, searching 'dpy-17' under Rearrangement returns hT2 only. However, the mT1 summary page explicitly lists dpy-17 as a positive marker.</p>
**Reported by:** Haro******** (smit******************)
**Submitted from:** <a target="_blank" href="http://www.wormbase.org//tools/support">/tools/support</a>
**Browser:** Mozilla/5.0 (Macintosh; Intel Mac OS X 10_8_5) AppleWebKit/600.5.17 (KHTML, like Gecko) Version/6.2.5 Safari/537.85.14
|
non_priority
|
dpy search is incorrectly redirected submitted from the feedback form on the wormbase website nbsp nbsp a recent thread on the forum uncovered some issues with the search function to identify balancers that cover a gene of interest it appears that the data are incomplete and or the search function not working as expected for example searching dpy under rearrangement returns only however the summary page explicitly lists dpy as a positive marker reported by haro smit submitted from a target blank href browser mozilla macintosh intel mac os x applewebkit khtml like gecko version safari
| 0
|
61,528
| 15,020,395,477
|
IssuesEvent
|
2021-02-01 14:40:09
|
spacetelescope/romancal
|
https://api.github.com/repos/spacetelescope/romancal
|
closed
|
Mockup level 1 data file
|
build0.1
|
_Issue [RCAL-73](https://jira.stsci.edu/browse/RCAL-73) was created on JIRA by [David Davis](https://jira.stsci.edu/secure/ViewProfile.jspa?name=ddavis):_
We should use the datamodel to mock up a level 1 data file using the JWST NRC data as a source for some of the keywords.
Initially it should be a 3-d cube since nints will be 1 for WIFI data.
Once this is done we should investigate how much effort would be needed to adapt the JWST ramp fit code to work in a 3-d data cube.
|
1.0
|
Mockup level 1 data file - _Issue [RCAL-73](https://jira.stsci.edu/browse/RCAL-73) was created on JIRA by [David Davis](https://jira.stsci.edu/secure/ViewProfile.jspa?name=ddavis):_
We should use the datamodel to mock up a level 1 data file using the JWST NRC data as a source for some of the keywords.
Initially it should be a 3-d cube since nints will be 1 for WIFI data.
Once this is done we should investigate how much effort would be needed to adapt the JWST ramp fit code to work in a 3-d data cube.
|
non_priority
|
mockup level data file issue was created on jira by we should use the datamodel to mock up a level data file using the jwst nrc data as a source for some of the keywords initially it should be a d cube since nints will be for wifi data once this is done we should investigate how much effort would be needed to adapt the jwst ramp fit code to work in a d data cube
| 0
|
262,160
| 8,251,676,224
|
IssuesEvent
|
2018-09-12 08:34:48
|
unitystation/unitystation
|
https://api.github.com/repos/unitystation/unitystation
|
closed
|
Bug: Nuke detonation video does not play in steam builds
|
High Priority bug
|
# Description
As per the title. In the steam builds the Nuke det video fails to play
###TODO:
- Check player logs after the nuke goes off to see if there is an error: https://docs.unity3d.com/Manual/LogFiles.html
- If nothing is found in there, check the Server build logs to see if there was an import problem with the video
|
1.0
|
Bug: Nuke detonation video does not play in steam builds - # Description
As per the title. In the steam builds the Nuke det video fails to play
###TODO:
- Check player logs after the nuke goes off to see if there is an error: https://docs.unity3d.com/Manual/LogFiles.html
- If nothing is found in there, check the Server build logs to see if there was an import problem with the video
|
priority
|
bug nuke detonation video does not play in steam builds description as per the title in the steam builds the nuke det video fails to play todo check player logs after the nuke goes off to see if there is an error if nothing is found in there check the server build logs to see if there was an import problem with the video
| 1
|
52,335
| 13,747,799,875
|
IssuesEvent
|
2020-10-06 08:10:15
|
idmarinas/lotgd-game
|
https://api.github.com/repos/idmarinas/lotgd-game
|
closed
|
CVE-2018-20821 (Medium) detected in multiple libraries - autoclosed
|
bug security vulnerability
|
## CVE-2018-20821 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.14.1.tgz</b></p></summary>
<p>
<details><summary><b>node-sass-4.14.1.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/lotgd-game/package.json</p>
<p>Path to vulnerable library: /lotgd-game/node_modules/node-sass/package.json</p>
<p>
Dependency Hierarchy:
- :x: **node-sass-4.14.1.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/idmarinas/lotgd-game/commit/a68ef33093518f30b58ce6afc7b54944a697f707">a68ef33093518f30b58ce6afc7b54944a697f707</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The parsing component in LibSass through 3.5.5 allows attackers to cause a denial-of-service (uncontrolled recursion in Sass::Parser::parse_css_variable_value in parser.cpp).
<p>Publish Date: 2019-04-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20821>CVE-2018-20821</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20821">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20821</a></p>
<p>Release Date: 2019-04-23</p>
<p>Fix Resolution: LibSass - 3.6.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2018-20821 (Medium) detected in multiple libraries - autoclosed - ## CVE-2018-20821 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.14.1.tgz</b></p></summary>
<p>
<details><summary><b>node-sass-4.14.1.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/lotgd-game/package.json</p>
<p>Path to vulnerable library: /lotgd-game/node_modules/node-sass/package.json</p>
<p>
Dependency Hierarchy:
- :x: **node-sass-4.14.1.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/idmarinas/lotgd-game/commit/a68ef33093518f30b58ce6afc7b54944a697f707">a68ef33093518f30b58ce6afc7b54944a697f707</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The parsing component in LibSass through 3.5.5 allows attackers to cause a denial-of-service (uncontrolled recursion in Sass::Parser::parse_css_variable_value in parser.cpp).
<p>Publish Date: 2019-04-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20821>CVE-2018-20821</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20821">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20821</a></p>
<p>Release Date: 2019-04-23</p>
<p>Fix Resolution: LibSass - 3.6.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_priority
|
cve medium detected in multiple libraries autoclosed cve medium severity vulnerability vulnerable libraries node sass tgz node sass tgz wrapper around libsass library home page a href path to dependency file tmp ws scm lotgd game package json path to vulnerable library lotgd game node modules node sass package json dependency hierarchy x node sass tgz vulnerable library found in head commit a href vulnerability details the parsing component in libsass through allows attackers to cause a denial of service uncontrolled recursion in sass parser parse css variable value in parser cpp publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass step up your open source security game with whitesource
| 0
|
101,437
| 11,235,847,685
|
IssuesEvent
|
2020-01-09 09:16:01
|
xtensor-stack/xtensor
|
https://api.github.com/repos/xtensor-stack/xtensor
|
closed
|
slicing and indexing
|
Documentation Question
|
How to handle multi-dimensional array slicing by using xtensor or some other alternatives:
a.shape (1, 120, 56, 1)
`b = a[0, :, :, 0]` ?
in numpy:
```
a = np.random.randn(1, 120, 56, 1) // a.shape (1, 120, 56, 1)
b = a[0, :, :, 0] // b.shape(120, 56)
```
Could you suggest some example or how to do.
Thank you beforehand for your time
|
1.0
|
slicing and indexing - How to handle multi-dimensional array slicing by using xtensor or some other alternatives:
a.shape (1, 120, 56, 1)
`b = a[0, :, :, 0]` ?
in numpy:
```
a = np.random.randn(1, 120, 56, 1) // a.shape (1, 120, 56, 1)
b = a[0, :, :, 0] // b.shape(120, 56)
```
Could you suggest some example or how to do.
Thank you beforehand for your time
|
non_priority
|
slicing and indexing how to handle multi dimensional array slicing by using xtensor or some other alternatives a shape b a in numpy a np random randn a shape b a b shape could you suggest some example or how to do thank you beforehand for your time
| 0
|
437,040
| 30,583,924,389
|
IssuesEvent
|
2023-07-21 11:56:29
|
orffen/basicfantasyrpg
|
https://api.github.com/repos/orffen/basicfantasyrpg
|
closed
|
README updates/Ability score shortcuts
|
bug documentation
|
- [x] The README says that ability scores can be used by using `@str`, but actually `@str.value` needs to be used. Alternatively, find a way to keep `@str.bonus` but also have `@str` automatically resolve to `@str.value`.
- [x] Change _LICENSE.txt_ to point to latest [LICENSE.txt](https://raw.githubusercontent.com/orffen/basicfantasyrpg/main/LICENSE.txt)) file.
|
1.0
|
README updates/Ability score shortcuts - - [x] The README says that ability scores can be used by using `@str`, but actually `@str.value` needs to be used. Alternatively, find a way to keep `@str.bonus` but also have `@str` automatically resolve to `@str.value`.
- [x] Change _LICENSE.txt_ to point to latest [LICENSE.txt](https://raw.githubusercontent.com/orffen/basicfantasyrpg/main/LICENSE.txt)) file.
|
non_priority
|
readme updates ability score shortcuts the readme says that ability scores can be used by using str but actually str value needs to be used alternatively find a way to keep str bonus but also have str automatically resolve to str value change license txt to point to latest file
| 0
|
303,455
| 9,307,410,749
|
IssuesEvent
|
2019-03-25 12:13:29
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
forum.xda-developers.com - design is broken
|
browser-firefox-tablet priority-important
|
<!-- @browser: Firefox Mobile (Tablet) 67.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Tablet; rv:67.0) Gecko/67.0 Firefox/67.0 -->
<!-- @reported_with: mobile-reporter -->
**URL**: https://forum.xda-developers.com/razer-phone-2/help/pie-volume-buttons-t3908250
**Browser / Version**: Firefox Mobile (Tablet) 67.0
**Operating System**: Android 8.1.0
**Tested Another Browser**: Unknown
**Problem type**: Design is broken
**Description**: screen formatting razer phone 2
**Steps to Reproduce**:
cutting top of screen it's a phone not tablet
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190306095759</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: nightly</li>
</ul>
<p>Console Messages:</p>
<pre>
[u'[JavaScript Error: "ReferenceError: vB_Text_Editor is not defined" {file: "https://forum.xda-developers.com/razer-phone-2/help/pie-volume-buttons-t3908250" line: 1249}]\n@https://forum.xda-developers.com/razer-phone-2/help/pie-volume-buttons-t3908250:1249:1\n', u'[console.log(QuickReply Init) https://forum.xda-developers.com/razer-phone-2/help/pie-volume-buttons-t3908250:1808:21]', u'[JavaScript Warning: "Content Security Policy: Ignoring x-frame-options because of frame-ancestors directive."]', u'[JavaScript Warning: "Content Security Policy: Ignoring x-frame-options because of frame-ancestors directive."]', u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://api.xda-developers.com/user. (Reason: CORS header Access-Control-Allow-Origin missing)."]', u'[console.log(Error during fetch of user data:) https://www.xda-developers.com/dist/js/xda_topnav.js?v=1551904321:13:19767]', u'[console.error([object XMLHttpRequest]) https://www.xda-developers.com/dist/js/xda_topnav.js?v=1551904321:13:19815]', u'[JavaScript Warning: "Use of Mutation Events is deprecated. Use MutationObserver instead." {file: "https://ajax.googleapis.com/ajax/libs/jquery/2.1.1/jquery.min.js" line: 3}]', u'[console.log(DIGI IFRAME:https://cdn.digitru.st/prod/1.5.24/dt.html) https://cdn.digitru.st/prod/1/digitrust.min.js:1:14745]']
</pre>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
forum.xda-developers.com - design is broken - <!-- @browser: Firefox Mobile (Tablet) 67.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Tablet; rv:67.0) Gecko/67.0 Firefox/67.0 -->
<!-- @reported_with: mobile-reporter -->
**URL**: https://forum.xda-developers.com/razer-phone-2/help/pie-volume-buttons-t3908250
**Browser / Version**: Firefox Mobile (Tablet) 67.0
**Operating System**: Android 8.1.0
**Tested Another Browser**: Unknown
**Problem type**: Design is broken
**Description**: screen formatting razer phone 2
**Steps to Reproduce**:
cutting top of screen it's a phone not tablet
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190306095759</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: nightly</li>
</ul>
<p>Console Messages:</p>
<pre>
[u'[JavaScript Error: "ReferenceError: vB_Text_Editor is not defined" {file: "https://forum.xda-developers.com/razer-phone-2/help/pie-volume-buttons-t3908250" line: 1249}]\n@https://forum.xda-developers.com/razer-phone-2/help/pie-volume-buttons-t3908250:1249:1\n', u'[console.log(QuickReply Init) https://forum.xda-developers.com/razer-phone-2/help/pie-volume-buttons-t3908250:1808:21]', u'[JavaScript Warning: "Content Security Policy: Ignoring x-frame-options because of frame-ancestors directive."]', u'[JavaScript Warning: "Content Security Policy: Ignoring x-frame-options because of frame-ancestors directive."]', u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://api.xda-developers.com/user. (Reason: CORS header Access-Control-Allow-Origin missing)."]', u'[console.log(Error during fetch of user data:) https://www.xda-developers.com/dist/js/xda_topnav.js?v=1551904321:13:19767]', u'[console.error([object XMLHttpRequest]) https://www.xda-developers.com/dist/js/xda_topnav.js?v=1551904321:13:19815]', u'[JavaScript Warning: "Use of Mutation Events is deprecated. Use MutationObserver instead." {file: "https://ajax.googleapis.com/ajax/libs/jquery/2.1.1/jquery.min.js" line: 3}]', u'[console.log(DIGI IFRAME:https://cdn.digitru.st/prod/1.5.24/dt.html) https://cdn.digitru.st/prod/1/digitrust.min.js:1:14745]']
</pre>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
priority
|
forum xda developers com design is broken url browser version firefox mobile tablet operating system android tested another browser unknown problem type design is broken description screen formatting razer phone steps to reproduce cutting top of screen it s a phone not tablet browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen true mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel nightly console messages n u u u u u u u u from with ❤️
| 1
|
70,632
| 23,268,678,591
|
IssuesEvent
|
2022-08-04 20:09:32
|
primefaces/primefaces
|
https://api.github.com/repos/primefaces/primefaces
|
closed
|
BlockUI: Incorrect overlay placement
|
:lady_beetle: defect
|
### Describe the bug
The overlay is displayed incorrectly, the size is correct, but position is shifted.
### Reproducer
_No response_
### Expected behavior
It seems it's a mistake to use `.position()` which returns the position of en element relative to its parent. But instead it should be relative to a page `.offset()`, because the overlay element is no more created next to a block element, but instead attached to `@(body)`.
### PrimeFaces edition
Community
### PrimeFaces version
12-RC2
### Theme
_No response_
### JSF implementation
MyFaces
### JSF version
3.0.0
### Browser(s)
Chrome 104
|
1.0
|
BlockUI: Incorrect overlay placement - ### Describe the bug
The overlay is displayed incorrectly, the size is correct, but position is shifted.
### Reproducer
_No response_
### Expected behavior
It seems it's a mistake to use `.position()` which returns the position of en element relative to its parent. But instead it should be relative to a page `.offset()`, because the overlay element is no more created next to a block element, but instead attached to `@(body)`.
### PrimeFaces edition
Community
### PrimeFaces version
12-RC2
### Theme
_No response_
### JSF implementation
MyFaces
### JSF version
3.0.0
### Browser(s)
Chrome 104
|
non_priority
|
blockui incorrect overlay placement describe the bug the overlay is displayed incorrectly the size is correct but position is shifted reproducer no response expected behavior it seems it s a mistake to use position which returns the position of en element relative to its parent but instead it should be relative to a page offset because the overlay element is no more created next to a block element but instead attached to body primefaces edition community primefaces version theme no response jsf implementation myfaces jsf version browser s chrome
| 0
|
78,177
| 3,509,500,412
|
IssuesEvent
|
2016-01-08 23:06:00
|
OregonCore/OregonCore
|
https://api.github.com/repos/OregonCore/OregonCore
|
closed
|
Whispers of the Raven (10607) (BB #947)
|
Category: Quests migrated Priority: Medium Type: Bug
|
This issue was migrated from bitbucket.
**Original Reporter:** bewtihozzy
**Original Date:** 31.05.2015 17:56:23 GMT+0000
**Original Priority:** major
**Original Type:** bug
**Original State:** resolved
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/947
<hr>
Whispers of the Raven (10607)
Kill Grishna arakkoa (22232) to gain Understanding Ravenspeech for 2 minutes. While under the effects of the spell, you can find and receive the prophecy from the four shrines. Return when you have visited all four.
Receive the First Prophecy
Receive the Second Prophecy
Receive the Third Prophecy
Receive the Fourth Prophecy
bug: When you kill the Grishna arakkoa, it doesnt give you the buff.
-> https://youtu.be/SNMBbDA-nR8
Edit: I moved the mob to that spot so he doesnt get attacked by guards xD
|
1.0
|
Whispers of the Raven (10607) (BB #947) - This issue was migrated from bitbucket.
**Original Reporter:** bewtihozzy
**Original Date:** 31.05.2015 17:56:23 GMT+0000
**Original Priority:** major
**Original Type:** bug
**Original State:** resolved
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/947
<hr>
Whispers of the Raven (10607)
Kill Grishna arakkoa (22232) to gain Understanding Ravenspeech for 2 minutes. While under the effects of the spell, you can find and receive the prophecy from the four shrines. Return when you have visited all four.
Receive the First Prophecy
Receive the Second Prophecy
Receive the Third Prophecy
Receive the Fourth Prophecy
bug: When you kill the Grishna arakkoa, it doesnt give you the buff.
-> https://youtu.be/SNMBbDA-nR8
Edit: I moved the mob to that spot so he doesnt get attacked by guards xD
|
priority
|
whispers of the raven bb this issue was migrated from bitbucket original reporter bewtihozzy original date gmt original priority major original type bug original state resolved direct link whispers of the raven kill grishna arakkoa to gain understanding ravenspeech for minutes while under the effects of the spell you can find and receive the prophecy from the four shrines return when you have visited all four receive the first prophecy receive the second prophecy receive the third prophecy receive the fourth prophecy bug when you kill the grishna arakkoa it doesnt give you the buff edit i moved the mob to that spot so he doesnt get attacked by guards xd
| 1
|
285,315
| 31,154,172,986
|
IssuesEvent
|
2023-08-16 12:06:10
|
nidhi7598/linux-4.1.15_CVE-2018-5873
|
https://api.github.com/repos/nidhi7598/linux-4.1.15_CVE-2018-5873
|
opened
|
CVE-2023-3111 (High) detected in linuxlinux-4.1.52
|
Mend: dependency security vulnerability
|
## CVE-2023-3111 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.1.52</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.1.15_CVE-2018-5873/commit/57227113ebd78150a4a62d376636875894cd3d72">57227113ebd78150a4a62d376636875894cd3d72</a></p>
<p>Found in base branch: <b>main</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/btrfs/relocation.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/btrfs/relocation.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
A use after free vulnerability was found in prepare_to_relocate in fs/btrfs/relocation.c in btrfs in the Linux Kernel. This possible flaw can be triggered by calling btrfs_ioctl_balance() before calling btrfs_ioctl_defrag().
<p>Publish Date: 2023-06-05
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-3111>CVE-2023-3111</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-3111">https://www.linuxkernelcves.com/cves/CVE-2023-3111</a></p>
<p>Release Date: 2023-06-05</p>
<p>Fix Resolution: v4.14.318,v4.19.286,v5.4.247,v5.10.184,v5.15.63</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2023-3111 (High) detected in linuxlinux-4.1.52 - ## CVE-2023-3111 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.1.52</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.1.15_CVE-2018-5873/commit/57227113ebd78150a4a62d376636875894cd3d72">57227113ebd78150a4a62d376636875894cd3d72</a></p>
<p>Found in base branch: <b>main</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/btrfs/relocation.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/btrfs/relocation.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
A use after free vulnerability was found in prepare_to_relocate in fs/btrfs/relocation.c in btrfs in the Linux Kernel. This possible flaw can be triggered by calling btrfs_ioctl_balance() before calling btrfs_ioctl_defrag().
<p>Publish Date: 2023-06-05
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-3111>CVE-2023-3111</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-3111">https://www.linuxkernelcves.com/cves/CVE-2023-3111</a></p>
<p>Release Date: 2023-06-05</p>
<p>Fix Resolution: v4.14.318,v4.19.286,v5.4.247,v5.10.184,v5.15.63</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_priority
|
cve high detected in linuxlinux cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch main vulnerable source files fs btrfs relocation c fs btrfs relocation c vulnerability details a use after free vulnerability was found in prepare to relocate in fs btrfs relocation c in btrfs in the linux kernel this possible flaw can be triggered by calling btrfs ioctl balance before calling btrfs ioctl defrag publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
112,016
| 4,500,861,297
|
IssuesEvent
|
2016-09-01 07:11:56
|
rndsolutions/hawkcd
|
https://api.github.com/repos/rndsolutions/hawkcd
|
closed
|
Add option to modify/assign resources for a given job
|
feature high priority
|
We need a button in the UI to enable users to assign resources to jobs.
|
1.0
|
Add option to modify/assign resources for a given job - We need a button in the UI to enable users to assign resources to jobs.
|
priority
|
add option to modify assign resources for a given job we need a button in the ui to enable users to assign resources to jobs
| 1
|
268,905
| 8,415,669,785
|
IssuesEvent
|
2018-10-13 17:02:20
|
react-epic/react-epic
|
https://api.github.com/repos/react-epic/react-epic
|
closed
|
Investigate the way to bundle a Stream
|
high priority
|
Allow to bundle a Stream into a StreamBundle. This is important for implementing such `browserHistoryApi` or `httpEntryPoint` (*) that don't need any alternative library
> (*): When completed it will work the same way with Relay.
Considering using RxJS `Subscriber` or `multicast` operator.
|
1.0
|
Investigate the way to bundle a Stream - Allow to bundle a Stream into a StreamBundle. This is important for implementing such `browserHistoryApi` or `httpEntryPoint` (*) that don't need any alternative library
> (*): When completed it will work the same way with Relay.
Considering using RxJS `Subscriber` or `multicast` operator.
|
priority
|
investigate the way to bundle a stream allow to bundle a stream into a streambundle this is important for implementing such browserhistoryapi or httpentrypoint that don t need any alternative library when completed it will work the same way with relay considering using rxjs subscriber or multicast operator
| 1
|
734,805
| 25,365,079,084
|
IssuesEvent
|
2022-11-21 05:13:51
|
wso2/api-manager
|
https://api.github.com/repos/wso2/api-manager
|
closed
|
[MSSQL] Database deadlock when saving API after adding multiple operation policies
|
Type/Bug Priority/Normal Affected/APIM-4.1.0 Affected/APIM-4.x.x Area/Migration
|
### Description
APIM 4.1.0 runs into a database deadlock scenario when it is setup with MSSQL.
There won't be any errors in the UI or in the console. None of the db commands will work in the server after that point (if you go to the landing page and try to go in to an API, the UI will keep loading infinitely). Only option in this scenario is to restart the server.
We came across this issue while doing some migration testing. However, this issue is not limited to migrated APIs or new APIs in a migrated pack. This can be reproduced in any 4.1.0 pack configured with MSSQL.
### Steps to Reproduce
Take the latest update level of APIM 4.1.0 (or GA)
Setup MSSQL as the database
Create an API
Add about 2 operation policies and save the API. You will be able to save without any issues.
Add another policy to the API and try to save. You will be able to see that UI is loading continuously.
### Affected Component
APIM
### Version
4.1.0
### Environment Details (with versions)
Migration from lower version to 4.1.0
### Relevant Log Output
_No response_
### Related Issues
_No response_
### Suggested Labels
_No response_
|
1.0
|
[MSSQL] Database deadlock when saving API after adding multiple operation policies - ### Description
APIM 4.1.0 runs into a database deadlock scenario when it is setup with MSSQL.
There won't be any errors in the UI or in the console. None of the db commands will work in the server after that point (if you go to the landing page and try to go in to an API, the UI will keep loading infinitely). Only option in this scenario is to restart the server.
We came across this issue while doing some migration testing. However, this issue is not limited to migrated APIs or new APIs in a migrated pack. This can be reproduced in any 4.1.0 pack configured with MSSQL.
### Steps to Reproduce
Take the latest update level of APIM 4.1.0 (or GA)
Setup MSSQL as the database
Create an API
Add about 2 operation policies and save the API. You will be able to save without any issues.
Add another policy to the API and try to save. You will be able to see that UI is loading continuously.
### Affected Component
APIM
### Version
4.1.0
### Environment Details (with versions)
Migration from lower version to 4.1.0
### Relevant Log Output
_No response_
### Related Issues
_No response_
### Suggested Labels
_No response_
|
priority
|
database deadlock when saving api after adding multiple operation policies description apim runs into a database deadlock scenario when it is setup with mssql there won t be any errors in the ui or in the console none of the db commands will work in the server after that point if you go to the landing page and try to go in to an api the ui will keep loading infinitely only option in this scenario is to restart the server we came across this issue while doing some migration testing however this issue is not limited to migrated apis or new apis in a migrated pack this can be reproduced in any pack configured with mssql steps to reproduce take the latest update level of apim or ga setup mssql as the database create an api add about operation policies and save the api you will be able to save without any issues add another policy to the api and try to save you will be able to see that ui is loading continuously affected component apim version environment details with versions migration from lower version to relevant log output no response related issues no response suggested labels no response
| 1
|
583,851
| 17,399,811,796
|
IssuesEvent
|
2021-08-02 17:57:23
|
genesis-community/prometheus-genesis-kit
|
https://api.github.com/repos/genesis-community/prometheus-genesis-kit
|
reopened
|
Check hook + genesis config block
|
PRIORITY: Low-hanging Fruit TRIAGE: Accepted TYPE: Bug
|
Check hook doesn't fail deployment with missing values - continues without failing
Also the new hook doesnt use the `genesis-config-block`
|
1.0
|
Check hook + genesis config block - Check hook doesn't fail deployment with missing values - continues without failing
Also the new hook doesnt use the `genesis-config-block`
|
priority
|
check hook genesis config block check hook doesn t fail deployment with missing values continues without failing also the new hook doesnt use the genesis config block
| 1
|
574,504
| 17,023,827,534
|
IssuesEvent
|
2021-07-03 04:03:28
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
Add ltmgfy support to osqa-based help system
|
Component: website Priority: trivial Resolution: invalid Type: enhancement
|
**[Submitted to the original trac issue database at 10.15am, Thursday, 27th September 2012]**
The help.osm.org system works well, but there is a definite issue with repeat questions, like this one:
https://help.openstreetmap.org/questions/16493/how-to-import-big-data
Sometimes a bit of gentle teasing might be in order, along the lines of
lmgtfy.com/?q=help
|
1.0
|
Add ltmgfy support to osqa-based help system - **[Submitted to the original trac issue database at 10.15am, Thursday, 27th September 2012]**
The help.osm.org system works well, but there is a definite issue with repeat questions, like this one:
https://help.openstreetmap.org/questions/16493/how-to-import-big-data
Sometimes a bit of gentle teasing might be in order, along the lines of
lmgtfy.com/?q=help
|
priority
|
add ltmgfy support to osqa based help system the help osm org system works well but there is a definite issue with repeat questions like this one sometimes a bit of gentle teasing might be in order along the lines of lmgtfy com q help
| 1
|
624,690
| 19,704,422,590
|
IssuesEvent
|
2022-01-12 20:10:43
|
svthalia/concrexit
|
https://api.github.com/repos/svthalia/concrexit
|
opened
|
Better selection of events for the agenda section in the newsletter
|
priority: low feature newsletter
|
### Is your feature request related to a problem? Please describe.
The agenda section in the newsletter should have a more intelligent way of selecting the events that are displayed:
- only the ones for the upcoming 2 weeks
- or the events that will have registration opening in the upcoming week
### Describe the solution you'd like
Changing this
The exact implementation will be thought about by @SnevelRetlaw
### Motivation
### Describe alternatives you've considered
We have talked with Walter
### Additional context
|
1.0
|
Better selection of events for the agenda section in the newsletter - ### Is your feature request related to a problem? Please describe.
The agenda section in the newsletter should have a more intelligent way of selecting the events that are displayed:
- only the ones for the upcoming 2 weeks
- or the events that will have registration opening in the upcoming week
### Describe the solution you'd like
Changing this
The exact implementation will be thought about by @SnevelRetlaw
### Motivation
### Describe alternatives you've considered
We have talked with Walter
### Additional context
|
priority
|
better selection of events for the agenda section in the newsletter is your feature request related to a problem please describe the agenda section in the newsletter should have a more intelligent way of selecting the events that are displayed only the ones for the upcoming weeks or the events that will have registration opening in the upcoming week describe the solution you d like changing this the exact implementation will be thought about by snevelretlaw motivation describe alternatives you ve considered we have talked with walter additional context
| 1
|
645,257
| 20,999,835,054
|
IssuesEvent
|
2022-03-29 16:22:38
|
magento/magento2
|
https://api.github.com/repos/magento/magento2
|
closed
|
Running phpstan twice (using the default Magento phpstan configuration) results in strange error
|
Issue: Confirmed Reproduced on 2.4.x Progress: ready for dev Priority: P3 Reported on 2.4.x Area: Framework
|
### Preconditions (*)
1. clean Magento installation (I've using 2.4-develop, commit 40ceb7019989c949f3e756efe2db673c8e40345c) with composer developer dependencies (which is the default when running `composer install`)
2. macOS 11.6.1
3. Case sensitive filesystem APFS
4. PHP 7.4.25 natively (no virtualisation or docker or whatever)
### Steps to reproduce (*)
1. Analyse the `app/autoload.php` file with phpstan: `./vendor/bin/phpstan analyse --level 1 app/autoload.php`
2. Run the same command a second or third time. Notice that this works fine everytime
3. Now use the default phpstan configuration file from Magento: `./vendor/bin/phpstan analyse --level 1 --configuration ./dev/tests/static/testsuite/Magento/Test/Php/_files/phpstan/phpstan.neon app/autoload.php` - this runs without issues
4. Now run this exact same command a **second** time - now this **fails** with:
```
$ ./vendor/bin/phpstan analyse --level 1 --configuration ./dev/tests/static/testsuite/Magento/Test/Php/_files/phpstan/phpstan.neon app/autoload.php
In File.php line 208:
Failed to open stream hoa://Library/Regex/Grammar.pp.
analyse [-c|--configuration CONFIGURATION] [-l|--level LEVEL] [--no-progress] [--debug] [-a|--autoload-file AUTOLOAD-FILE] [--error-format ERROR-FORMAT] [--generate-baseline [GENERATE-BASELINE]] [--memory-limit MEMORY-LIMIT] [--xdebug] [--fix] [--watch] [--pro] [--] [<paths>...]
```
5. Change level of phpstan's analysis from 1 to 2: `./vendor/bin/phpstan analyse --level 2 --configuration ./dev/tests/static/testsuite/Magento/Test/Php/_files/phpstan/phpstan.neon app/autoload.php` - notice that this works fine again
6. Run the exact same command again a **second** time - notice that this **fails** again
I have no idea what this is, it only fails with Magento's configuration file and fails the **second** time you try to run the command.
Clearing phpstan's cache with `./vendor/bin/phpstan clear-result-cache` doesn't help ...
Removing the entire `vendor` directory and running `composer install` also doesn't help ...
Problem happens on phpstan versions 0.12.x and 1.x
When switching from `2.4-develop` to the `2.4.3` tag and running `composer install`, it does fix the problem, but only for the first run. The **second** run fails again! Switching back to `2.4-develop` and running `composer install` fixes it again but only for the first run.
When I search the web for this error, I end up on https://github.com/wunderio/code-quality/issues/45 but I don't understand much of this.
So not sure if this is a bug in Magento's phpstan configuration file, or in phpstan, or in one of the hoa libraries, it's hard to say ...
But it's very annoying for users of phpstan if they want to run phpstan inside a Magento project and use the default Magento configuration file.
### Expected result (*)
1. No failures
### Actual result (*)
1. Failures, see above
---
Please provide [Severity](https://devdocs.magento.com/guides/v2.4/contributor-guide/contributing.html#backlog) assessment for the Issue as Reporter. This information will help during Confirmation and Issue triage processes.
- [ ] Severity: **S0** _- Affects critical data or functionality and leaves users without workaround._
- [ ] Severity: **S1** _- Affects critical data or functionality and forces users to employ a workaround._
- [ ] Severity: **S2** _- Affects non-critical data or functionality and forces users to employ a workaround._
- [ ] Severity: **S3** _- Affects non-critical data or functionality and does not force users to employ a workaround._
- [ ] Severity: **S4** _- Affects aesthetics, professional look and feel, “quality” or “usability”._
|
1.0
|
Running phpstan twice (using the default Magento phpstan configuration) results in strange error - ### Preconditions (*)
1. clean Magento installation (I've using 2.4-develop, commit 40ceb7019989c949f3e756efe2db673c8e40345c) with composer developer dependencies (which is the default when running `composer install`)
2. macOS 11.6.1
3. Case sensitive filesystem APFS
4. PHP 7.4.25 natively (no virtualisation or docker or whatever)
### Steps to reproduce (*)
1. Analyse the `app/autoload.php` file with phpstan: `./vendor/bin/phpstan analyse --level 1 app/autoload.php`
2. Run the same command a second or third time. Notice that this works fine everytime
3. Now use the default phpstan configuration file from Magento: `./vendor/bin/phpstan analyse --level 1 --configuration ./dev/tests/static/testsuite/Magento/Test/Php/_files/phpstan/phpstan.neon app/autoload.php` - this runs without issues
4. Now run this exact same command a **second** time - now this **fails** with:
```
$ ./vendor/bin/phpstan analyse --level 1 --configuration ./dev/tests/static/testsuite/Magento/Test/Php/_files/phpstan/phpstan.neon app/autoload.php
In File.php line 208:
Failed to open stream hoa://Library/Regex/Grammar.pp.
analyse [-c|--configuration CONFIGURATION] [-l|--level LEVEL] [--no-progress] [--debug] [-a|--autoload-file AUTOLOAD-FILE] [--error-format ERROR-FORMAT] [--generate-baseline [GENERATE-BASELINE]] [--memory-limit MEMORY-LIMIT] [--xdebug] [--fix] [--watch] [--pro] [--] [<paths>...]
```
5. Change level of phpstan's analysis from 1 to 2: `./vendor/bin/phpstan analyse --level 2 --configuration ./dev/tests/static/testsuite/Magento/Test/Php/_files/phpstan/phpstan.neon app/autoload.php` - notice that this works fine again
6. Run the exact same command again a **second** time - notice that this **fails** again
I have no idea what this is, it only fails with Magento's configuration file and fails the **second** time you try to run the command.
Clearing phpstan's cache with `./vendor/bin/phpstan clear-result-cache` doesn't help ...
Removing the entire `vendor` directory and running `composer install` also doesn't help ...
Problem happens on phpstan versions 0.12.x and 1.x
When switching from `2.4-develop` to the `2.4.3` tag and running `composer install`, it does fix the problem, but only for the first run. The **second** run fails again! Switching back to `2.4-develop` and running `composer install` fixes it again but only for the first run.
When I search the web for this error, I end up on https://github.com/wunderio/code-quality/issues/45 but I don't understand much of this.
So not sure if this is a bug in Magento's phpstan configuration file, or in phpstan, or in one of the hoa libraries, it's hard to say ...
But it's very annoying for users of phpstan if they want to run phpstan inside a Magento project and use the default Magento configuration file.
### Expected result (*)
1. No failures
### Actual result (*)
1. Failures, see above
---
Please provide [Severity](https://devdocs.magento.com/guides/v2.4/contributor-guide/contributing.html#backlog) assessment for the Issue as Reporter. This information will help during Confirmation and Issue triage processes.
- [ ] Severity: **S0** _- Affects critical data or functionality and leaves users without workaround._
- [ ] Severity: **S1** _- Affects critical data or functionality and forces users to employ a workaround._
- [ ] Severity: **S2** _- Affects non-critical data or functionality and forces users to employ a workaround._
- [ ] Severity: **S3** _- Affects non-critical data or functionality and does not force users to employ a workaround._
- [ ] Severity: **S4** _- Affects aesthetics, professional look and feel, “quality” or “usability”._
|
priority
|
running phpstan twice using the default magento phpstan configuration results in strange error preconditions clean magento installation i ve using develop commit with composer developer dependencies which is the default when running composer install macos case sensitive filesystem apfs php natively no virtualisation or docker or whatever steps to reproduce analyse the app autoload php file with phpstan vendor bin phpstan analyse level app autoload php run the same command a second or third time notice that this works fine everytime now use the default phpstan configuration file from magento vendor bin phpstan analyse level configuration dev tests static testsuite magento test php files phpstan phpstan neon app autoload php this runs without issues now run this exact same command a second time now this fails with vendor bin phpstan analyse level configuration dev tests static testsuite magento test php files phpstan phpstan neon app autoload php in file php line failed to open stream hoa library regex grammar pp analyse change level of phpstan s analysis from to vendor bin phpstan analyse level configuration dev tests static testsuite magento test php files phpstan phpstan neon app autoload php notice that this works fine again run the exact same command again a second time notice that this fails again i have no idea what this is it only fails with magento s configuration file and fails the second time you try to run the command clearing phpstan s cache with vendor bin phpstan clear result cache doesn t help removing the entire vendor directory and running composer install also doesn t help problem happens on phpstan versions x and x when switching from develop to the tag and running composer install it does fix the problem but only for the first run the second run fails again switching back to develop and running composer install fixes it again but only for the first run when i search the web for this error i end up on but i don t understand much of this so not sure if this is a bug in magento s phpstan configuration file or in phpstan or in one of the hoa libraries it s hard to say but it s very annoying for users of phpstan if they want to run phpstan inside a magento project and use the default magento configuration file expected result no failures actual result failures see above please provide assessment for the issue as reporter this information will help during confirmation and issue triage processes severity affects critical data or functionality and leaves users without workaround severity affects critical data or functionality and forces users to employ a workaround severity affects non critical data or functionality and forces users to employ a workaround severity affects non critical data or functionality and does not force users to employ a workaround severity affects aesthetics professional look and feel “quality” or “usability”
| 1
|
423,784
| 12,302,096,515
|
IssuesEvent
|
2020-05-11 16:24:22
|
internetarchive/openlibrary
|
https://api.github.com/repos/internetarchive/openlibrary
|
closed
|
Sitemap returning 404
|
Lead: @cdrini Priority: 1 Theme: SEO Type: Bug
|
The sitemap looks like it's no longer available
### Evidence / Screenshot (if possible)
Google Search Console:

### Relevant url?
- https://openlibrary.org/static/sitemaps/siteindex.xml.gz
### Details
- **Logged in (Y/N)?** Y
- **Browser type/version?** FF75
- **Operating system?** Win10
- **Environment (prod/dev/local)?** prod
<!-- If not sure, put prod -->
### Proposal & Constraints
<!-- What is the proposed solution / implementation? Is there a precedent of this approach succeeding elsewhere? -->
### Related files
<!-- Files related to this issue; this is super useful for new contributors who might want to help! If you're not sure, leave this blank; a maintainer will add them. -->
### Stakeholders
@mekarpeles
|
1.0
|
Sitemap returning 404 - The sitemap looks like it's no longer available
### Evidence / Screenshot (if possible)
Google Search Console:

### Relevant url?
- https://openlibrary.org/static/sitemaps/siteindex.xml.gz
### Details
- **Logged in (Y/N)?** Y
- **Browser type/version?** FF75
- **Operating system?** Win10
- **Environment (prod/dev/local)?** prod
<!-- If not sure, put prod -->
### Proposal & Constraints
<!-- What is the proposed solution / implementation? Is there a precedent of this approach succeeding elsewhere? -->
### Related files
<!-- Files related to this issue; this is super useful for new contributors who might want to help! If you're not sure, leave this blank; a maintainer will add them. -->
### Stakeholders
@mekarpeles
|
priority
|
sitemap returning the sitemap looks like it s no longer available evidence screenshot if possible google search console relevant url details logged in y n y browser type version operating system environment prod dev local prod proposal constraints related files stakeholders mekarpeles
| 1
|
829,209
| 31,859,029,291
|
IssuesEvent
|
2023-09-15 09:32:59
|
shaka-project/shaka-player
|
https://api.github.com/repos/shaka-project/shaka-player
|
reopened
|
WebVTT wrong writing-mode in nested cues
|
type: bug priority: P2 component: WebVTT
|
**Have you read the [FAQ](https://bit.ly/ShakaFAQ) and checked for duplicate open issues?**
Yes.
**What version of Shaka Player are you using?**
4.3.0
**Can you reproduce the issue with our latest release version?**
Yes.
**Can you reproduce the issue with the latest code from `main`?**
Yes.
**Are you using the demo app or your own custom app?**
My own app.
**If custom app, can you reproduce the issue using our demo app?**
Not tried.
**What browser and OS are you using?**
Chrome 107.0.5304.107 (Build officiel) (64 bits)
Windows Windows 10 Professionnel 21H2
**What did you do?**
I played a video with side-car WebVTT subtitles that have `vertical-lr` writing-mode.
**What did you expect to happen?**
The subtitles should be written vertically.
**What actually happened?**
They're displayed horizontally, as if the writing-mode was `horizontal-tb`.
When I look at the cues just after they're parsed, I see this:

The root cue (or whatever you want to call it) has the correct `vertical-lr` writing-mode, but the nested cue has `horizontal-tb` writing-mode.
I think all nested cues should have the same writing-mode as their root cue.
|
1.0
|
WebVTT wrong writing-mode in nested cues - **Have you read the [FAQ](https://bit.ly/ShakaFAQ) and checked for duplicate open issues?**
Yes.
**What version of Shaka Player are you using?**
4.3.0
**Can you reproduce the issue with our latest release version?**
Yes.
**Can you reproduce the issue with the latest code from `main`?**
Yes.
**Are you using the demo app or your own custom app?**
My own app.
**If custom app, can you reproduce the issue using our demo app?**
Not tried.
**What browser and OS are you using?**
Chrome 107.0.5304.107 (Build officiel) (64 bits)
Windows Windows 10 Professionnel 21H2
**What did you do?**
I played a video with side-car WebVTT subtitles that have `vertical-lr` writing-mode.
**What did you expect to happen?**
The subtitles should be written vertically.
**What actually happened?**
They're displayed horizontally, as if the writing-mode was `horizontal-tb`.
When I look at the cues just after they're parsed, I see this:

The root cue (or whatever you want to call it) has the correct `vertical-lr` writing-mode, but the nested cue has `horizontal-tb` writing-mode.
I think all nested cues should have the same writing-mode as their root cue.
|
priority
|
webvtt wrong writing mode in nested cues have you read the and checked for duplicate open issues yes what version of shaka player are you using can you reproduce the issue with our latest release version yes can you reproduce the issue with the latest code from main yes are you using the demo app or your own custom app my own app if custom app can you reproduce the issue using our demo app not tried what browser and os are you using chrome build officiel bits windows windows professionnel what did you do i played a video with side car webvtt subtitles that have vertical lr writing mode what did you expect to happen the subtitles should be written vertically what actually happened they re displayed horizontally as if the writing mode was horizontal tb when i look at the cues just after they re parsed i see this the root cue or whatever you want to call it has the correct vertical lr writing mode but the nested cue has horizontal tb writing mode i think all nested cues should have the same writing mode as their root cue
| 1
|
27,031
| 7,889,710,075
|
IssuesEvent
|
2018-06-28 05:58:02
|
ShaikASK/Testing
|
https://api.github.com/repos/ShaikASK/Testing
|
opened
|
UAT : Candidate Module : User is able to navigate in "Dashboard" screen without signing the "Offer Letter"
|
Candidate Dashboard Candidate Module Defect P1 Release #3 Build 7
|
Steps :
Sign in as Candidate credentials
Navigate to "Introduction" page
Click on "Dashboard" from side menu
Experienced Behaviour : Observed that user is able to navigate in "Dashboard" screen without signing the "Offer Letter"
Expected Behaviour : Ensure that application should not allow the user to navigate in "Dashboard" screen without signing the "Offer Letter" and "Dashboard" should be disable until user signs the "Offer Letter"
|
1.0
|
UAT : Candidate Module : User is able to navigate in "Dashboard" screen without signing the "Offer Letter" - Steps :
Sign in as Candidate credentials
Navigate to "Introduction" page
Click on "Dashboard" from side menu
Experienced Behaviour : Observed that user is able to navigate in "Dashboard" screen without signing the "Offer Letter"
Expected Behaviour : Ensure that application should not allow the user to navigate in "Dashboard" screen without signing the "Offer Letter" and "Dashboard" should be disable until user signs the "Offer Letter"
|
non_priority
|
uat candidate module user is able to navigate in dashboard screen without signing the offer letter steps sign in as candidate credentials navigate to introduction page click on dashboard from side menu experienced behaviour observed that user is able to navigate in dashboard screen without signing the offer letter expected behaviour ensure that application should not allow the user to navigate in dashboard screen without signing the offer letter and dashboard should be disable until user signs the offer letter
| 0
|
678,893
| 23,214,909,479
|
IssuesEvent
|
2022-08-02 13:22:52
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
www.pandora.com - video or audio doesn't play
|
browser-firefox priority-normal engine-gecko
|
<!-- @browser: Firefox 103.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:103.0) Gecko/20100101 Firefox/103.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/108296 -->
**URL**: https://www.pandora.com
**Browser / Version**: Firefox 103.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Video or audio doesn't play
**Description**: There is no audio
**Steps to Reproduce**:
Appears to be playing but no sound
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
www.pandora.com - video or audio doesn't play - <!-- @browser: Firefox 103.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:103.0) Gecko/20100101 Firefox/103.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/108296 -->
**URL**: https://www.pandora.com
**Browser / Version**: Firefox 103.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Video or audio doesn't play
**Description**: There is no audio
**Steps to Reproduce**:
Appears to be playing but no sound
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
priority
|
video or audio doesn t play url browser version firefox operating system windows tested another browser yes chrome problem type video or audio doesn t play description there is no audio steps to reproduce appears to be playing but no sound browser configuration none from with ❤️
| 1
|
21,455
| 6,153,059,942
|
IssuesEvent
|
2017-06-28 09:01:19
|
src-d/go-kallax
|
https://api.github.com/repos/src-d/go-kallax
|
closed
|
PK issues
|
area/codegen enhancement
|
Hello!
I'd like to use kallax in my projects, however, I have some issues with doing this.
I prefer to have my models and store layer in different packages, I would like not to have kallax to import in the models package. I do this by having a new type in the store package for every model I need, for example:
```go
package models
type Foo struct {
ID int `pk:"autoincr"`
Bar string
Baz string
}
```
```go
package kxstore
import (
kallax "gopkg.in/src-d/go-kallax.v1"
"my-project/models"
)
type Foo struct {
kallax.Model `table:"foos"`
models.Foo
}
```
When running `kallax gen`, i get the following message: `kallax: model Foo has no primary key defined`.
Also, some of my models don't have a primary key, does kallax not support models without ids?
|
1.0
|
PK issues - Hello!
I'd like to use kallax in my projects, however, I have some issues with doing this.
I prefer to have my models and store layer in different packages, I would like not to have kallax to import in the models package. I do this by having a new type in the store package for every model I need, for example:
```go
package models
type Foo struct {
ID int `pk:"autoincr"`
Bar string
Baz string
}
```
```go
package kxstore
import (
kallax "gopkg.in/src-d/go-kallax.v1"
"my-project/models"
)
type Foo struct {
kallax.Model `table:"foos"`
models.Foo
}
```
When running `kallax gen`, i get the following message: `kallax: model Foo has no primary key defined`.
Also, some of my models don't have a primary key, does kallax not support models without ids?
|
non_priority
|
pk issues hello i d like to use kallax in my projects however i have some issues with doing this i prefer to have my models and store layer in different packages i would like not to have kallax to import in the models package i do this by having a new type in the store package for every model i need for example go package models type foo struct id int pk autoincr bar string baz string go package kxstore import kallax gopkg in src d go kallax my project models type foo struct kallax model table foos models foo when running kallax gen i get the following message kallax model foo has no primary key defined also some of my models don t have a primary key does kallax not support models without ids
| 0
|
134,967
| 12,643,144,409
|
IssuesEvent
|
2020-06-16 09:19:26
|
crate/crate-howtos
|
https://api.github.com/repos/crate/crate-howtos
|
opened
|
the status of the cr8 tool is not clear
|
documentation enhancement
|
### Documentation feedback
<!--Please do not edit or remove the following information -->
- Page title: Testing inserts performance
- Page URL: https://crate.io/docs/crate/howtos/en/latest/performance/inserts/testing.html
- Source: https://github.com/crate/crate-howtos/blob/master/docs/performance/inserts/testing.rst
---
<!-- Please add your comments here -->
the status of the cr8 tool is not clear. cr8 is not maintained by cr8, it is maintained by @mfussenegger. we should add a note with the appropriate caveat
|
1.0
|
the status of the cr8 tool is not clear - ### Documentation feedback
<!--Please do not edit or remove the following information -->
- Page title: Testing inserts performance
- Page URL: https://crate.io/docs/crate/howtos/en/latest/performance/inserts/testing.html
- Source: https://github.com/crate/crate-howtos/blob/master/docs/performance/inserts/testing.rst
---
<!-- Please add your comments here -->
the status of the cr8 tool is not clear. cr8 is not maintained by cr8, it is maintained by @mfussenegger. we should add a note with the appropriate caveat
|
non_priority
|
the status of the tool is not clear documentation feedback page title testing inserts performance page url source the status of the tool is not clear is not maintained by it is maintained by mfussenegger we should add a note with the appropriate caveat
| 0
|
107,307
| 11,525,841,505
|
IssuesEvent
|
2020-02-15 11:23:25
|
google/jax
|
https://api.github.com/repos/google/jax
|
closed
|
Autodiff Cookbook Docs Out of Sync
|
documentation
|
In second-to last part of the section [Jacobians and Hessians using jacfwd and jacrev](https://jax.readthedocs.io/en/latest/notebooks/autodiff_cookbook.html#Jacobians-and-Hessians-using-jacfwd-and-jacrev), the readthedocs markdown is not rendering correctly, namely there is a list that is being rendered inline like:
"This shape makes sense: if we start with a function f:Rn→Rm, then at a point x∈Rn we expect to get the shapes * f(x)∈Rm, the value of f at x, * ∂f(x)∈Rm×n, the Jacobian matrix at x, * ∂2f(x)∈Rm×n×n, the Hessian at x,
and so on."
In the checked in ipython notebook, [this same section](https://github.com/google/jax/blob/master/docs/notebooks/autodiff_cookbook.ipynb) is rendered correctly as:
"
This shape makes sense: if we start with a function $f : \mathbb{R}^n \to \mathbb{R}^m$, then at a point $x \in \mathbb{R}^n$ we expect to get the shapes
$f(x) \in \mathbb{R}^m$, the value of $f$ at $x$,
$\partial f(x) \in \mathbb{R}^{m \times n}$, the Jacobian matrix at $x$,
$\partial^2 f(x) \in \mathbb{R}^{m \times n \times n}$, the Hessian at $x$,
and so on.
"
I don't know if readthedocs uses a different markdown parser than ipython or if this is just a case of out of date docs, but figured it would be useful to flag.
|
1.0
|
Autodiff Cookbook Docs Out of Sync - In second-to last part of the section [Jacobians and Hessians using jacfwd and jacrev](https://jax.readthedocs.io/en/latest/notebooks/autodiff_cookbook.html#Jacobians-and-Hessians-using-jacfwd-and-jacrev), the readthedocs markdown is not rendering correctly, namely there is a list that is being rendered inline like:
"This shape makes sense: if we start with a function f:Rn→Rm, then at a point x∈Rn we expect to get the shapes * f(x)∈Rm, the value of f at x, * ∂f(x)∈Rm×n, the Jacobian matrix at x, * ∂2f(x)∈Rm×n×n, the Hessian at x,
and so on."
In the checked in ipython notebook, [this same section](https://github.com/google/jax/blob/master/docs/notebooks/autodiff_cookbook.ipynb) is rendered correctly as:
"
This shape makes sense: if we start with a function $f : \mathbb{R}^n \to \mathbb{R}^m$, then at a point $x \in \mathbb{R}^n$ we expect to get the shapes
$f(x) \in \mathbb{R}^m$, the value of $f$ at $x$,
$\partial f(x) \in \mathbb{R}^{m \times n}$, the Jacobian matrix at $x$,
$\partial^2 f(x) \in \mathbb{R}^{m \times n \times n}$, the Hessian at $x$,
and so on.
"
I don't know if readthedocs uses a different markdown parser than ipython or if this is just a case of out of date docs, but figured it would be useful to flag.
|
non_priority
|
autodiff cookbook docs out of sync in second to last part of the section the readthedocs markdown is not rendering correctly namely there is a list that is being rendered inline like this shape makes sense if we start with a function f rn→rm then at a point x∈rn we expect to get the shapes f x ∈rm the value of f at x ∂f x ∈rm×n the jacobian matrix at x ∂ x ∈rm×n×n the hessian at x and so on in the checked in ipython notebook is rendered correctly as this shape makes sense if we start with a function f mathbb r n to mathbb r m then at a point x in mathbb r n we expect to get the shapes f x in mathbb r m the value of f at x partial f x in mathbb r m times n the jacobian matrix at x partial f x in mathbb r m times n times n the hessian at x and so on i don t know if readthedocs uses a different markdown parser than ipython or if this is just a case of out of date docs but figured it would be useful to flag
| 0
|
412,502
| 12,043,167,049
|
IssuesEvent
|
2020-04-14 11:56:56
|
guidesmiths/cybersecurity-handbook
|
https://api.github.com/repos/guidesmiths/cybersecurity-handbook
|
closed
|
OWASP TOP 10 Refactor
|
content priority-low
|
### TODO:
- [ ] Review tools
- [ ] Add new tools referenced in other sections
- [ ] Add nodejs examples
- [ ] Add videos
- [ ] Improve and curate content
|
1.0
|
OWASP TOP 10 Refactor - ### TODO:
- [ ] Review tools
- [ ] Add new tools referenced in other sections
- [ ] Add nodejs examples
- [ ] Add videos
- [ ] Improve and curate content
|
priority
|
owasp top refactor todo review tools add new tools referenced in other sections add nodejs examples add videos improve and curate content
| 1
|
210,554
| 16,374,832,931
|
IssuesEvent
|
2021-05-15 22:09:04
|
Mustard2/MustardUI
|
https://api.github.com/repos/Mustard2/MustardUI
|
closed
|
Outdated documentation
|
bug documentation
|
The documentation is outdated. Missing stuffs:
- how to add custom properties
- how to add particle physics hair in the UI
|
1.0
|
Outdated documentation - The documentation is outdated. Missing stuffs:
- how to add custom properties
- how to add particle physics hair in the UI
|
non_priority
|
outdated documentation the documentation is outdated missing stuffs how to add custom properties how to add particle physics hair in the ui
| 0
|
49,269
| 6,186,457,099
|
IssuesEvent
|
2017-07-04 02:32:10
|
18F/nsf-sbir
|
https://api.github.com/repos/18F/nsf-sbir
|
closed
|
Landing page - hero image update & caption
|
design feedback - critical Priority
|
I received the updated image from Sigray and saved it here: https://github.com/18F/nsf-sbir/blob/dev/assets/img/hero_sigray_seedling_alone.jpg
"Sigray is revolutionizing laboratory x-ray research equipment with patented innovations to deliver synchrotron performance. Pictured is a microscopic image of the elements within a hyperaccumulating seedling"
Here's their page: http://www.sigray.com/
If you update the caption, please let me know and I'll send it to the company for approval.
|
1.0
|
Landing page - hero image update & caption - I received the updated image from Sigray and saved it here: https://github.com/18F/nsf-sbir/blob/dev/assets/img/hero_sigray_seedling_alone.jpg
"Sigray is revolutionizing laboratory x-ray research equipment with patented innovations to deliver synchrotron performance. Pictured is a microscopic image of the elements within a hyperaccumulating seedling"
Here's their page: http://www.sigray.com/
If you update the caption, please let me know and I'll send it to the company for approval.
|
non_priority
|
landing page hero image update caption i received the updated image from sigray and saved it here sigray is revolutionizing laboratory x ray research equipment with patented innovations to deliver synchrotron performance pictured is a microscopic image of the elements within a hyperaccumulating seedling here s their page if you update the caption please let me know and i ll send it to the company for approval
| 0
|
570,266
| 17,023,076,078
|
IssuesEvent
|
2021-07-03 00:16:39
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
street segments with the same name should be merged into streets
|
Component: api Priority: major Resolution: invalid Type: enhancement
|
**[Submitted to the original trac issue database at 10.36am, Friday, 11th November 2005]**
Not sure if this is a client issue or server issue, but it should definitely be looked into.
Ben's Tiger import script did this for the XML-RPC API, but I'm not sure how streets are handled in the REST API.
|
1.0
|
street segments with the same name should be merged into streets - **[Submitted to the original trac issue database at 10.36am, Friday, 11th November 2005]**
Not sure if this is a client issue or server issue, but it should definitely be looked into.
Ben's Tiger import script did this for the XML-RPC API, but I'm not sure how streets are handled in the REST API.
|
priority
|
street segments with the same name should be merged into streets not sure if this is a client issue or server issue but it should definitely be looked into ben s tiger import script did this for the xml rpc api but i m not sure how streets are handled in the rest api
| 1
|
34,760
| 7,460,080,549
|
IssuesEvent
|
2018-03-30 18:04:12
|
kerdokullamae/test_koik_issued
|
https://api.github.com/repos/kerdokullamae/test_koik_issued
|
closed
|
Lehitse > Ainese märksõnad ei ole sorteeritud tekstiliselt
|
C: AIS P: highest R: duplicate T: defect
|
**Reported by aadikaljuvee on 27 Apr 2016 09:27 UTC**
http://ais2.arhiiv.ee/et/topic/browse#/
(Teistest hiljem lisatud ainesemärksõna)"Fotod" peaks paiknema tähestikuliselt enne "Fotograafe". Aga lehitsemisel, kui ainult 1 kord on vajutatud "näita järgmised 100", siis seda "Fotod" seal pole. "Fotod" jõuab oma õigele kohale alles kõigi märksõnade avamise järel.
|
1.0
|
Lehitse > Ainese märksõnad ei ole sorteeritud tekstiliselt - **Reported by aadikaljuvee on 27 Apr 2016 09:27 UTC**
http://ais2.arhiiv.ee/et/topic/browse#/
(Teistest hiljem lisatud ainesemärksõna)"Fotod" peaks paiknema tähestikuliselt enne "Fotograafe". Aga lehitsemisel, kui ainult 1 kord on vajutatud "näita järgmised 100", siis seda "Fotod" seal pole. "Fotod" jõuab oma õigele kohale alles kõigi märksõnade avamise järel.
|
non_priority
|
lehitse ainese märksõnad ei ole sorteeritud tekstiliselt reported by aadikaljuvee on apr utc teistest hiljem lisatud ainesemärksõna fotod peaks paiknema tähestikuliselt enne fotograafe aga lehitsemisel kui ainult kord on vajutatud näita järgmised siis seda fotod seal pole fotod jõuab oma õigele kohale alles kõigi märksõnade avamise järel
| 0
|
61,290
| 14,965,281,723
|
IssuesEvent
|
2021-01-27 13:10:55
|
eventespresso/barista
|
https://api.github.com/repos/eventespresso/barista
|
closed
|
Prevent Duplicate SASS Imports
|
C: build-process 🔨 D: Packages 📦 P2: HIGH priority 😮 S:1 new 👶🏻 T: bug 🐞
|
Seems to be a fairly common issue for ppl using complex SASS setups. Here's a thread regarding the issue with a possible fix:
https://github.com/webpack-contrib/sass-loader/issues/145
# OOPS
that link above is for webpack 🤦🏻♂️
|
1.0
|
Prevent Duplicate SASS Imports - Seems to be a fairly common issue for ppl using complex SASS setups. Here's a thread regarding the issue with a possible fix:
https://github.com/webpack-contrib/sass-loader/issues/145
# OOPS
that link above is for webpack 🤦🏻♂️
|
non_priority
|
prevent duplicate sass imports seems to be a fairly common issue for ppl using complex sass setups here s a thread regarding the issue with a possible fix oops that link above is for webpack 🤦🏻♂️
| 0
|
813,271
| 30,450,498,316
|
IssuesEvent
|
2023-07-16 08:22:10
|
realpoke/genranks
|
https://api.github.com/repos/realpoke/genranks
|
opened
|
Update my games
|
Type: Enhancement Priority: Medium Status: Available
|
Have an update my games button for users that's logged in, so they don't have to wait for the daily update. But can get their games updated instantly.
|
1.0
|
Update my games - Have an update my games button for users that's logged in, so they don't have to wait for the daily update. But can get their games updated instantly.
|
priority
|
update my games have an update my games button for users that s logged in so they don t have to wait for the daily update but can get their games updated instantly
| 1
|
757,006
| 26,492,421,534
|
IssuesEvent
|
2023-01-18 00:31:06
|
zephyrproject-rtos/zephyr
|
https://api.github.com/repos/zephyrproject-rtos/zephyr
|
closed
|
Settings with FCB backend does not pass test on stm32h743
|
bug priority: low platform: STM32 Stale
|
**Describe the bug**
I tried to run the `zephyr/tests/subsys/settings/fcb/raw/` on a nucleo h743zi2 but I have a test errors, bus faults and stack overflows:
**To Reproduce**
Steps to reproduce the behavior:
1. `west build -b nucleo_h743zi -d build_test/fcb zephyr/tests/subsys/settings/fcb/raw/`
2. `west flash -d build_test/fcb`
3. See error on serial output.
**Expected behavior**
All tests should pass
**Impact**
Settings with FCB backend cannot be used on stm32h7 (NVS is not an option either: #45031)
**Logs and console output**
```
*** Booting Zephyr OS build zephyr-v3.1.0-1919-g8a2eed6cbdbc ***
Running TESTSUITE test_config_fcb
===================================================================
START - test_config_empty_lookups
PASS - test_config_empty_lookups in 0.1 seconds
===================================================================
START - test_config_insert
PASS - test_config_insert in 0.1 seconds
===================================================================
START - test_config_getset_unknown
PASS - test_config_getset_unknown in 0.1 seconds
===================================================================
START - test_config_getset_int
PASS - test_config_getset_int in 0.1 seconds
===================================================================
START - test_config_getset_int64
PASS - test_config_getset_int64 in 0.1 seconds
===================================================================
START - test_config_commit
PASS - test_config_commit in 0.1 seconds
===================================================================
START - test_settings_check_target
Assertion failed at WEST_TOPDIR/zephyr/tests/subsys/settings/fcb/src/settings_test_fcb.c:354: test_settings_check_target: (wbs <= 16 is false)
Flash driver is not compatible with the settings fcb-backend
FAIL - test_settings_check_target in 0.19 seconds
===================================================================
START - test_config_save_fcb_unaligned
E: ***** BUS FAULT *****
E: Precise data bus error
E: BFAR Address: 0x934b2921
E: r0/a1: 0x00000022 r1/a2: 0x00000020 r2/a3: 0x24008870
E: r3/a4: 0x00000002 r12/ip: 0x00000732 r14/lr: 0x08007a1d
E: xpsr: 0x21000000
E: Faulting instruction address (r15/pc): 0x08007fb4
E: >>> ZEPHYR FATAL ERROR 0: CPU exception on CPU 0
E: Current thread: 0x240002c8 (test_config_save_fcb_unaligned)
E: Halting system
```
**Environment (please complete the following information):**
- Linux
- zephyr-sdk-0.14.2
- 8a2eed6cbdbc82550d3b39e8ba808102e82c3fb8
**Additional context**
I fixed the first bus fault by increasing the
buffer used to read/write to flash from 16 bytes to 32 bytes [settings_fcb_h7.diff.txt](https://github.com/zephyrproject-rtos/zephyr/files/9082889/settings_fcb_h7.diff.txt)
Then the first crash appears on `test_config_save_2_fcb`:
```
START - test_config_save_2_fcb
E: ***** MPU FAULT *****
E: Stacking error (context area might be not valid)
E: Data Access Violation
E: MMFAR Address: 0x240086fc
E: r0/a1: 0x5d9e865c r1/a2: 0xc96aa6ea r2/a3: 0xf624b53f
E: r3/a4: 0x9baa9bd7 r12/ip: 0xcf8fb769 r14/lr: 0xcc5f6b2c
E: xpsr: 0x24008600
E: Faulting instruction address (r15/pc): 0xe1af64c5
E: >>> ZEPHYR FATAL ERROR 2: Stack overflow on CPU 0
E: Current thread: 0x240002c8 (test_config_save_2_fcb)
E: Halting system
```
|
1.0
|
Settings with FCB backend does not pass test on stm32h743 - **Describe the bug**
I tried to run the `zephyr/tests/subsys/settings/fcb/raw/` on a nucleo h743zi2 but I have a test errors, bus faults and stack overflows:
**To Reproduce**
Steps to reproduce the behavior:
1. `west build -b nucleo_h743zi -d build_test/fcb zephyr/tests/subsys/settings/fcb/raw/`
2. `west flash -d build_test/fcb`
3. See error on serial output.
**Expected behavior**
All tests should pass
**Impact**
Settings with FCB backend cannot be used on stm32h7 (NVS is not an option either: #45031)
**Logs and console output**
```
*** Booting Zephyr OS build zephyr-v3.1.0-1919-g8a2eed6cbdbc ***
Running TESTSUITE test_config_fcb
===================================================================
START - test_config_empty_lookups
PASS - test_config_empty_lookups in 0.1 seconds
===================================================================
START - test_config_insert
PASS - test_config_insert in 0.1 seconds
===================================================================
START - test_config_getset_unknown
PASS - test_config_getset_unknown in 0.1 seconds
===================================================================
START - test_config_getset_int
PASS - test_config_getset_int in 0.1 seconds
===================================================================
START - test_config_getset_int64
PASS - test_config_getset_int64 in 0.1 seconds
===================================================================
START - test_config_commit
PASS - test_config_commit in 0.1 seconds
===================================================================
START - test_settings_check_target
Assertion failed at WEST_TOPDIR/zephyr/tests/subsys/settings/fcb/src/settings_test_fcb.c:354: test_settings_check_target: (wbs <= 16 is false)
Flash driver is not compatible with the settings fcb-backend
FAIL - test_settings_check_target in 0.19 seconds
===================================================================
START - test_config_save_fcb_unaligned
E: ***** BUS FAULT *****
E: Precise data bus error
E: BFAR Address: 0x934b2921
E: r0/a1: 0x00000022 r1/a2: 0x00000020 r2/a3: 0x24008870
E: r3/a4: 0x00000002 r12/ip: 0x00000732 r14/lr: 0x08007a1d
E: xpsr: 0x21000000
E: Faulting instruction address (r15/pc): 0x08007fb4
E: >>> ZEPHYR FATAL ERROR 0: CPU exception on CPU 0
E: Current thread: 0x240002c8 (test_config_save_fcb_unaligned)
E: Halting system
```
**Environment (please complete the following information):**
- Linux
- zephyr-sdk-0.14.2
- 8a2eed6cbdbc82550d3b39e8ba808102e82c3fb8
**Additional context**
I fixed the first bus fault by increasing the
buffer used to read/write to flash from 16 bytes to 32 bytes [settings_fcb_h7.diff.txt](https://github.com/zephyrproject-rtos/zephyr/files/9082889/settings_fcb_h7.diff.txt)
Then the first crash appears on `test_config_save_2_fcb`:
```
START - test_config_save_2_fcb
E: ***** MPU FAULT *****
E: Stacking error (context area might be not valid)
E: Data Access Violation
E: MMFAR Address: 0x240086fc
E: r0/a1: 0x5d9e865c r1/a2: 0xc96aa6ea r2/a3: 0xf624b53f
E: r3/a4: 0x9baa9bd7 r12/ip: 0xcf8fb769 r14/lr: 0xcc5f6b2c
E: xpsr: 0x24008600
E: Faulting instruction address (r15/pc): 0xe1af64c5
E: >>> ZEPHYR FATAL ERROR 2: Stack overflow on CPU 0
E: Current thread: 0x240002c8 (test_config_save_2_fcb)
E: Halting system
```
|
priority
|
settings with fcb backend does not pass test on describe the bug i tried to run the zephyr tests subsys settings fcb raw on a nucleo but i have a test errors bus faults and stack overflows to reproduce steps to reproduce the behavior west build b nucleo d build test fcb zephyr tests subsys settings fcb raw west flash d build test fcb see error on serial output expected behavior all tests should pass impact settings with fcb backend cannot be used on nvs is not an option either logs and console output booting zephyr os build zephyr running testsuite test config fcb start test config empty lookups pass test config empty lookups in seconds start test config insert pass test config insert in seconds start test config getset unknown pass test config getset unknown in seconds start test config getset int pass test config getset int in seconds start test config getset pass test config getset in seconds start test config commit pass test config commit in seconds start test settings check target assertion failed at west topdir zephyr tests subsys settings fcb src settings test fcb c test settings check target wbs is false flash driver is not compatible with the settings fcb backend fail test settings check target in seconds start test config save fcb unaligned e bus fault e precise data bus error e bfar address e e ip lr e xpsr e faulting instruction address pc e zephyr fatal error cpu exception on cpu e current thread test config save fcb unaligned e halting system environment please complete the following information linux zephyr sdk additional context i fixed the first bus fault by increasing the buffer used to read write to flash from bytes to bytes then the first crash appears on test config save fcb start test config save fcb e mpu fault e stacking error context area might be not valid e data access violation e mmfar address e e ip lr e xpsr e faulting instruction address pc e zephyr fatal error stack overflow on cpu e current thread test config save fcb e halting system
| 1
|
10,298
| 2,622,141,300
|
IssuesEvent
|
2015-03-04 00:02:07
|
byzhang/spserver
|
https://api.github.com/repos/byzhang/spserver
|
opened
|
请问如何发送文件?
|
auto-migrated Priority-Medium Type-Defect
|
```
例子里的是字符串数据,如果发送文件如何做?
```
Original issue reported on code.google.com by `xiejie...@126.com` on 12 Apr 2011 at 4:46
|
1.0
|
请问如何发送文件? - ```
例子里的是字符串数据,如果发送文件如何做?
```
Original issue reported on code.google.com by `xiejie...@126.com` on 12 Apr 2011 at 4:46
|
non_priority
|
请问如何发送文件 例子里的是字符串数据 如果发送文件如何做 original issue reported on code google com by xiejie com on apr at
| 0
|
280,418
| 24,304,004,982
|
IssuesEvent
|
2022-09-29 15:50:40
|
yugabyte/yugabyte-db
|
https://api.github.com/repos/yugabyte/yugabyte-db
|
closed
|
[YSQL] flaky test: org.yb.pgsql.TestPgRegressExtension.testPgRegressExtension
|
kind/bug area/ysql kind/failing-test priority/high
|
Jira Link: [DB-3649](https://yugabyte.atlassian.net/browse/DB-3649)
### Description
https://detective-gcp.dev.yugabyte.com/stability/test?analyze_trends=true&branch=master&build_type=all&class=org.yb.pgsql.TestPgRegressExtension&fail_tag=all&name=testPgRegressExtension&platform=linux
seems flaky in several build types, but clearly broken in ASAN since https://github.com/yugabyte/yugabyte-db/commit/388ae9def67cd91527b5968f58b7fe1e8b6e7766
|
1.0
|
[YSQL] flaky test: org.yb.pgsql.TestPgRegressExtension.testPgRegressExtension - Jira Link: [DB-3649](https://yugabyte.atlassian.net/browse/DB-3649)
### Description
https://detective-gcp.dev.yugabyte.com/stability/test?analyze_trends=true&branch=master&build_type=all&class=org.yb.pgsql.TestPgRegressExtension&fail_tag=all&name=testPgRegressExtension&platform=linux
seems flaky in several build types, but clearly broken in ASAN since https://github.com/yugabyte/yugabyte-db/commit/388ae9def67cd91527b5968f58b7fe1e8b6e7766
|
non_priority
|
flaky test org yb pgsql testpgregressextension testpgregressextension jira link description seems flaky in several build types but clearly broken in asan since
| 0
|
250,509
| 18,892,285,135
|
IssuesEvent
|
2021-11-15 14:27:36
|
boom210232/ku-polls
|
https://api.github.com/repos/boom210232/ku-polls
|
reopened
|
[Info] Readme for instructor about detail.
|
documentation
|
# MY ANNOUNCEMENT
- Iteration2 need to merge again because flake8 style advice something that make some function can't use.
- Builds in travis-ci have been temporarily disabled for private and public repositories due to a negative credit balance.
|
1.0
|
[Info] Readme for instructor about detail. - # MY ANNOUNCEMENT
- Iteration2 need to merge again because flake8 style advice something that make some function can't use.
- Builds in travis-ci have been temporarily disabled for private and public repositories due to a negative credit balance.
|
non_priority
|
readme for instructor about detail my announcement need to merge again because style advice something that make some function can t use builds in travis ci have been temporarily disabled for private and public repositories due to a negative credit balance
| 0
|
474,341
| 13,657,405,795
|
IssuesEvent
|
2020-09-28 05:39:43
|
Azure/autorest.az
|
https://api.github.com/repos/Azure/autorest.az
|
closed
|
Expose key of a named collection as a parameter
|
priority_p2_nice-to-have type_feature-request
|
In the `create_or_update` operation of Blueprint, [resourceGroups](https://github.com/Azure/azure-rest-api-specs/blob/2f48d306ba63a7ea3c19bd122a274dc465a84e4c/specification/blueprint/resource-manager/Microsoft.Blueprint/preview/2018-11-01-preview/blueprintDefinition.json#L847) is added as a [named collection](https://github.com/Azure/azure-rest-api-specs/blob/2f48d306ba63a7ea3c19bd122a274dc465a84e4c/specification/blueprint/resource-manager/Microsoft.Blueprint/preview/2018-11-01-preview/blueprintDefinition.json#L1200) which will be mapped as a dictionary in python. Its key would be used to get/update/delete a resource group artifact in the blueprint.
In Azure Portal, this key is not exposed and will be autogenerated as 'ResourceGroup', 'ResourceGroup2' e.t.c. When users operate on the created resource group artifact on GUI, the underlying key will be used when sending requests to the service.
In Azure CLI, we need to let users be able to name this key so they can later use this key to operate the created resource group. So we want to expose the key parameter.
I'm not sure how common named collection is used in swagger. This could be of low priority if it's not a common use.
|
1.0
|
Expose key of a named collection as a parameter - In the `create_or_update` operation of Blueprint, [resourceGroups](https://github.com/Azure/azure-rest-api-specs/blob/2f48d306ba63a7ea3c19bd122a274dc465a84e4c/specification/blueprint/resource-manager/Microsoft.Blueprint/preview/2018-11-01-preview/blueprintDefinition.json#L847) is added as a [named collection](https://github.com/Azure/azure-rest-api-specs/blob/2f48d306ba63a7ea3c19bd122a274dc465a84e4c/specification/blueprint/resource-manager/Microsoft.Blueprint/preview/2018-11-01-preview/blueprintDefinition.json#L1200) which will be mapped as a dictionary in python. Its key would be used to get/update/delete a resource group artifact in the blueprint.
In Azure Portal, this key is not exposed and will be autogenerated as 'ResourceGroup', 'ResourceGroup2' e.t.c. When users operate on the created resource group artifact on GUI, the underlying key will be used when sending requests to the service.
In Azure CLI, we need to let users be able to name this key so they can later use this key to operate the created resource group. So we want to expose the key parameter.
I'm not sure how common named collection is used in swagger. This could be of low priority if it's not a common use.
|
priority
|
expose key of a named collection as a parameter in the create or update operation of blueprint is added as a which will be mapped as a dictionary in python its key would be used to get update delete a resource group artifact in the blueprint in azure portal this key is not exposed and will be autogenerated as resourcegroup e t c when users operate on the created resource group artifact on gui the underlying key will be used when sending requests to the service in azure cli we need to let users be able to name this key so they can later use this key to operate the created resource group so we want to expose the key parameter i m not sure how common named collection is used in swagger this could be of low priority if it s not a common use
| 1
|
101,330
| 8,785,402,314
|
IssuesEvent
|
2018-12-20 12:53:38
|
elastic/elasticsearch
|
https://api.github.com/repos/elastic/elasticsearch
|
closed
|
[CI][ML] MlDistributedFailureIT.testLoseDedicatedMasterNode randomly fails on feature-jindex-master branch
|
:ml >test-failure v7.0.0
|
This test has been observed to fail occasionally in the `feature-jindex-master` branch. I have not yet managed to reproduce locally. I will shortly be muting the test as we need a green build to merge the branch in master. However, I am raising the issue to ensure we get to the bottom of this failure.
Link to failure (one of them): https://elasticsearch-ci.elastic.co/job/elastic+elasticsearch+pull-request-2/2556/console
Reproduce with:
```
./gradlew :x-pack:plugin:ml:internalClusterTest -Dtests.seed=D2A618A38265651F -Dtests.class=org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT -Dtests.method="testLoseDedicatedMasterNode" -Dtests.security.manager=true -Dtests.locale=be-BY -Dtests.timezone=Asia/Katmandu -Dcompiler.java=11 -Druntime.java=8
```
Failure:
```
> Throwable #1: java.lang.AssertionError
> at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298)
> at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:847)
> at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:821)
> at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.run(MlDistributedFailureIT.java:292)
> at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.testLoseDedicatedMasterNode(MlDistributedFailureIT.java:88)
> at java.lang.Thread.run(Thread.java:748)
> Suppressed: java.lang.AssertionError
> at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298)
> at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835)
> ... 40 more
> Suppressed: java.lang.AssertionError
> at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298)
> at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835)
> ... 40 more
> Suppressed: java.lang.AssertionError
> at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298)
> at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835)
> ... 40 more
> Suppressed: java.lang.AssertionError
> at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298)
> at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835)
> ... 40 more
> Suppressed: java.lang.AssertionError
> at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298)
> at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835)
> ... 40 more
> Suppressed: java.lang.AssertionError
> at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298)
> at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835)
> ... 40 more
> Suppressed: java.lang.AssertionError
> at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298)
> at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835)
> ... 40 more
> Suppressed: java.lang.AssertionError
> at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298)
> at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835)
> ... 40 more
> Suppressed: java.lang.AssertionError
> at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298)
> at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835)
> ... 40 more
> Suppressed: java.lang.AssertionError
> at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298)
> at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835)
> ... 40 more
> Suppressed: java.lang.AssertionError
> at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298)
> at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835)
> ... 40 more
> Suppressed: java.lang.AssertionError
> at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298)
> at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835)
> ... 40 more
> Suppressed: java.lang.AssertionError
> at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298)
> at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835)
> ... 40 moreThrowable #2: java.lang.RuntimeException: Had to resort to force-stopping datafeed, something went wrong?
> at org.elasticsearch.xpack.ml.support.BaseMlIntegTestCase.deleteAllDatafeeds(BaseMlIntegTestCase.java:296)
> at org.elasticsearch.xpack.ml.support.BaseMlIntegTestCase.cleanupWorkaround(BaseMlIntegTestCase.java:209)
> at java.lang.Thread.run(Thread.java:748)
> Caused by: java.util.concurrent.ExecutionException: ElasticsearchStatusException[Cannot stop datafeed [data_feed_id] because the datafeed does not have an assigned node. Use force stop to stop the datafeed]
> at org.elasticsearch.common.util.concurrent.BaseFuture$Sync.getValue(BaseFuture.java:266)
> at org.elasticsearch.common.util.concurrent.BaseFuture$Sync.get(BaseFuture.java:253)
> at org.elasticsearch.common.util.concurrent.BaseFuture.get(BaseFuture.java:87)
> at org.elasticsearch.xpack.ml.support.BaseMlIntegTestCase.deleteAllDatafeeds(BaseMlIntegTestCase.java:284)
> ... 36 more
> Caused by: ElasticsearchStatusException[Cannot stop datafeed [data_feed_id] because the datafeed does not have an assigned node. Use force stop to stop the datafeed]
> at org.elasticsearch.xpack.core.ml.utils.ExceptionsHelper.conflictStatusException(ExceptionsHelper.java:50)
> at org.elasticsearch.xpack.ml.action.TransportStopDatafeedAction.normalStopDatafeed(TransportStopDatafeedAction.java:147)
> at org.elasticsearch.xpack.ml.action.TransportStopDatafeedAction.lambda$doExecute$0(TransportStopDatafeedAction.java:130)
> at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:60)
> at org.elasticsearch.xpack.ml.datafeed.persistence.DatafeedConfigProvider.lambd
```
|
1.0
|
[CI][ML] MlDistributedFailureIT.testLoseDedicatedMasterNode randomly fails on feature-jindex-master branch - This test has been observed to fail occasionally in the `feature-jindex-master` branch. I have not yet managed to reproduce locally. I will shortly be muting the test as we need a green build to merge the branch in master. However, I am raising the issue to ensure we get to the bottom of this failure.
Link to failure (one of them): https://elasticsearch-ci.elastic.co/job/elastic+elasticsearch+pull-request-2/2556/console
Reproduce with:
```
./gradlew :x-pack:plugin:ml:internalClusterTest -Dtests.seed=D2A618A38265651F -Dtests.class=org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT -Dtests.method="testLoseDedicatedMasterNode" -Dtests.security.manager=true -Dtests.locale=be-BY -Dtests.timezone=Asia/Katmandu -Dcompiler.java=11 -Druntime.java=8
```
Failure:
```
> Throwable #1: java.lang.AssertionError
> at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298)
> at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:847)
> at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:821)
> at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.run(MlDistributedFailureIT.java:292)
> at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.testLoseDedicatedMasterNode(MlDistributedFailureIT.java:88)
> at java.lang.Thread.run(Thread.java:748)
> Suppressed: java.lang.AssertionError
> at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298)
> at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835)
> ... 40 more
> Suppressed: java.lang.AssertionError
> at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298)
> at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835)
> ... 40 more
> Suppressed: java.lang.AssertionError
> at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298)
> at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835)
> ... 40 more
> Suppressed: java.lang.AssertionError
> at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298)
> at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835)
> ... 40 more
> Suppressed: java.lang.AssertionError
> at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298)
> at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835)
> ... 40 more
> Suppressed: java.lang.AssertionError
> at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298)
> at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835)
> ... 40 more
> Suppressed: java.lang.AssertionError
> at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298)
> at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835)
> ... 40 more
> Suppressed: java.lang.AssertionError
> at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298)
> at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835)
> ... 40 more
> Suppressed: java.lang.AssertionError
> at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298)
> at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835)
> ... 40 more
> Suppressed: java.lang.AssertionError
> at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298)
> at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835)
> ... 40 more
> Suppressed: java.lang.AssertionError
> at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298)
> at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835)
> ... 40 more
> Suppressed: java.lang.AssertionError
> at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298)
> at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835)
> ... 40 more
> Suppressed: java.lang.AssertionError
> at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298)
> at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835)
> ... 40 moreThrowable #2: java.lang.RuntimeException: Had to resort to force-stopping datafeed, something went wrong?
> at org.elasticsearch.xpack.ml.support.BaseMlIntegTestCase.deleteAllDatafeeds(BaseMlIntegTestCase.java:296)
> at org.elasticsearch.xpack.ml.support.BaseMlIntegTestCase.cleanupWorkaround(BaseMlIntegTestCase.java:209)
> at java.lang.Thread.run(Thread.java:748)
> Caused by: java.util.concurrent.ExecutionException: ElasticsearchStatusException[Cannot stop datafeed [data_feed_id] because the datafeed does not have an assigned node. Use force stop to stop the datafeed]
> at org.elasticsearch.common.util.concurrent.BaseFuture$Sync.getValue(BaseFuture.java:266)
> at org.elasticsearch.common.util.concurrent.BaseFuture$Sync.get(BaseFuture.java:253)
> at org.elasticsearch.common.util.concurrent.BaseFuture.get(BaseFuture.java:87)
> at org.elasticsearch.xpack.ml.support.BaseMlIntegTestCase.deleteAllDatafeeds(BaseMlIntegTestCase.java:284)
> ... 36 more
> Caused by: ElasticsearchStatusException[Cannot stop datafeed [data_feed_id] because the datafeed does not have an assigned node. Use force stop to stop the datafeed]
> at org.elasticsearch.xpack.core.ml.utils.ExceptionsHelper.conflictStatusException(ExceptionsHelper.java:50)
> at org.elasticsearch.xpack.ml.action.TransportStopDatafeedAction.normalStopDatafeed(TransportStopDatafeedAction.java:147)
> at org.elasticsearch.xpack.ml.action.TransportStopDatafeedAction.lambda$doExecute$0(TransportStopDatafeedAction.java:130)
> at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:60)
> at org.elasticsearch.xpack.ml.datafeed.persistence.DatafeedConfigProvider.lambd
```
|
non_priority
|
mldistributedfailureit testlosededicatedmasternode randomly fails on feature jindex master branch this test has been observed to fail occasionally in the feature jindex master branch i have not yet managed to reproduce locally i will shortly be muting the test as we need a green build to merge the branch in master however i am raising the issue to ensure we get to the bottom of this failure link to failure one of them reproduce with gradlew x pack plugin ml internalclustertest dtests seed dtests class org elasticsearch xpack ml integration mldistributedfailureit dtests method testlosededicatedmasternode dtests security manager true dtests locale be by dtests timezone asia katmandu dcompiler java druntime java failure throwable java lang assertionerror at org elasticsearch xpack ml integration mldistributedfailureit lambda run mldistributedfailureit java at org elasticsearch test estestcase assertbusy estestcase java at org elasticsearch test estestcase assertbusy estestcase java at org elasticsearch xpack ml integration mldistributedfailureit run mldistributedfailureit java at org elasticsearch xpack ml integration mldistributedfailureit testlosededicatedmasternode mldistributedfailureit java at java lang thread run thread java suppressed java lang assertionerror at org elasticsearch xpack ml integration mldistributedfailureit lambda run mldistributedfailureit java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror at org elasticsearch xpack ml integration mldistributedfailureit lambda run mldistributedfailureit java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror at org elasticsearch xpack ml integration mldistributedfailureit lambda run mldistributedfailureit java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror at org elasticsearch xpack ml integration mldistributedfailureit lambda run mldistributedfailureit java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror at org elasticsearch xpack ml integration mldistributedfailureit lambda run mldistributedfailureit java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror at org elasticsearch xpack ml integration mldistributedfailureit lambda run mldistributedfailureit java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror at org elasticsearch xpack ml integration mldistributedfailureit lambda run mldistributedfailureit java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror at org elasticsearch xpack ml integration mldistributedfailureit lambda run mldistributedfailureit java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror at org elasticsearch xpack ml integration mldistributedfailureit lambda run mldistributedfailureit java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror at org elasticsearch xpack ml integration mldistributedfailureit lambda run mldistributedfailureit java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror at org elasticsearch xpack ml integration mldistributedfailureit lambda run mldistributedfailureit java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror at org elasticsearch xpack ml integration mldistributedfailureit lambda run mldistributedfailureit java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror at org elasticsearch xpack ml integration mldistributedfailureit lambda run mldistributedfailureit java at org elasticsearch test estestcase assertbusy estestcase java morethrowable java lang runtimeexception had to resort to force stopping datafeed something went wrong at org elasticsearch xpack ml support basemlintegtestcase deletealldatafeeds basemlintegtestcase java at org elasticsearch xpack ml support basemlintegtestcase cleanupworkaround basemlintegtestcase java at java lang thread run thread java caused by java util concurrent executionexception elasticsearchstatusexception because the datafeed does not have an assigned node use force stop to stop the datafeed at org elasticsearch common util concurrent basefuture sync getvalue basefuture java at org elasticsearch common util concurrent basefuture sync get basefuture java at org elasticsearch common util concurrent basefuture get basefuture java at org elasticsearch xpack ml support basemlintegtestcase deletealldatafeeds basemlintegtestcase java more caused by elasticsearchstatusexception because the datafeed does not have an assigned node use force stop to stop the datafeed at org elasticsearch xpack core ml utils exceptionshelper conflictstatusexception exceptionshelper java at org elasticsearch xpack ml action transportstopdatafeedaction normalstopdatafeed transportstopdatafeedaction java at org elasticsearch xpack ml action transportstopdatafeedaction lambda doexecute transportstopdatafeedaction java at org elasticsearch action actionlistener onresponse actionlistener java at org elasticsearch xpack ml datafeed persistence datafeedconfigprovider lambd
| 0
|
264,721
| 8,318,943,668
|
IssuesEvent
|
2018-09-25 15:51:31
|
swarm-robotics/fordyca
|
https://api.github.com/repos/swarm-robotics/fordyca
|
closed
|
task/345-run-exp-with-new-task-allocation-params
|
Priority: Major Status: Available Type: Task
|
Pulling from the matlab function mesh analysis I FINALLY got around to doing
|
1.0
|
task/345-run-exp-with-new-task-allocation-params - Pulling from the matlab function mesh analysis I FINALLY got around to doing
|
priority
|
task run exp with new task allocation params pulling from the matlab function mesh analysis i finally got around to doing
| 1
|
123,841
| 16,542,304,451
|
IssuesEvent
|
2021-05-27 18:28:46
|
elementary/switchboard-plug-onlineaccounts
|
https://api.github.com/repos/elementary/switchboard-plug-onlineaccounts
|
closed
|
Add an IMAP account
|
Needs Design Priority: Wishlist
|
<!--
* Please read and follow these tips: https://elementary.io/docs/code/reference#proposing-design-changes
* Be sure to search open and closed issues for duplicates
-->
We need to be able to add standard IMAP accounts to EDS
<!--Please be sure to preview your issue before saving. Thanks!-->
|
1.0
|
Add an IMAP account - <!--
* Please read and follow these tips: https://elementary.io/docs/code/reference#proposing-design-changes
* Be sure to search open and closed issues for duplicates
-->
We need to be able to add standard IMAP accounts to EDS
<!--Please be sure to preview your issue before saving. Thanks!-->
|
non_priority
|
add an imap account please read and follow these tips be sure to search open and closed issues for duplicates we need to be able to add standard imap accounts to eds
| 0
|
2,216
| 2,588,223,766
|
IssuesEvent
|
2015-02-17 23:29:19
|
rlf/uSkyBlock
|
https://api.github.com/repos/rlf/uSkyBlock
|
closed
|
Support UUID
|
bug feature request ready for test
|
Currently, the uSkyBlock plugin doesn't correctly handle players.
Too many places rely on the username, instead of the UUID.
This should be fixed.
|
1.0
|
Support UUID - Currently, the uSkyBlock plugin doesn't correctly handle players.
Too many places rely on the username, instead of the UUID.
This should be fixed.
|
non_priority
|
support uuid currently the uskyblock plugin doesn t correctly handle players too many places rely on the username instead of the uuid this should be fixed
| 0
|
343,978
| 30,704,847,247
|
IssuesEvent
|
2023-07-27 04:53:34
|
litentry/litentry-parachain
|
https://api.github.com/repos/litentry/litentry-parachain
|
closed
|
DI tests: identity
|
I3-high D5-testing
|
## Context
See #1682
## Task
Implement each of the following test steps using the DI interface.
Refactor and extract common functionality as needed, and establish a pattern for the remaining test suites.
If needed, merge to dev in smaller batches to mitigate conflicts with other developments.
Along the way, consider better options for reducing the combinational explosion :P
### Steps
- [x] `check user sidechain storage before create`
- [x] `Invalid user shielding key`
- [x] `set user shielding key`
- [x] `check user shielding key from sidechain storage after setUserShieldingKey`
- [x] `check idgraph from sidechain storage before linking`
- [x] `link identities`
- [x] `check IDGraph after LinkIdentity`
- [x] `link invalid identities`
- [x] `link identities with wrong signature`
- [x] `deactivate identities`
- [x] `check IDGraph after deactivate identity`
- [x] `deactivate prime identity is disallowed`
- [x] `activate identity`
- [x] `check IDGraph after activate identity`
- [ ] `set error user shielding key`
- [ ] `exceeding IDGraph limit not allowed`
## Due Diligence
- [ ] Create issues for missing test cases from the original II test suite (marked with `TODO` comments in the code)
|
1.0
|
DI tests: identity - ## Context
See #1682
## Task
Implement each of the following test steps using the DI interface.
Refactor and extract common functionality as needed, and establish a pattern for the remaining test suites.
If needed, merge to dev in smaller batches to mitigate conflicts with other developments.
Along the way, consider better options for reducing the combinational explosion :P
### Steps
- [x] `check user sidechain storage before create`
- [x] `Invalid user shielding key`
- [x] `set user shielding key`
- [x] `check user shielding key from sidechain storage after setUserShieldingKey`
- [x] `check idgraph from sidechain storage before linking`
- [x] `link identities`
- [x] `check IDGraph after LinkIdentity`
- [x] `link invalid identities`
- [x] `link identities with wrong signature`
- [x] `deactivate identities`
- [x] `check IDGraph after deactivate identity`
- [x] `deactivate prime identity is disallowed`
- [x] `activate identity`
- [x] `check IDGraph after activate identity`
- [ ] `set error user shielding key`
- [ ] `exceeding IDGraph limit not allowed`
## Due Diligence
- [ ] Create issues for missing test cases from the original II test suite (marked with `TODO` comments in the code)
|
non_priority
|
di tests identity context see task implement each of the following test steps using the di interface refactor and extract common functionality as needed and establish a pattern for the remaining test suites if needed merge to dev in smaller batches to mitigate conflicts with other developments along the way consider better options for reducing the combinational explosion p steps check user sidechain storage before create invalid user shielding key set user shielding key check user shielding key from sidechain storage after setusershieldingkey check idgraph from sidechain storage before linking link identities check idgraph after linkidentity link invalid identities link identities with wrong signature deactivate identities check idgraph after deactivate identity deactivate prime identity is disallowed activate identity check idgraph after activate identity set error user shielding key exceeding idgraph limit not allowed due diligence create issues for missing test cases from the original ii test suite marked with todo comments in the code
| 0
|
252,467
| 19,027,407,132
|
IssuesEvent
|
2021-11-24 06:27:45
|
DocNow/twarc-csv
|
https://api.github.com/repos/DocNow/twarc-csv
|
closed
|
DataFrameConverter to single tweet
|
documentation
|
I have a large set of tweets, and I would like to wrangle and write them to file as I go. DataFrameConverter would be ideal for this, but if I try to pass a tweet to it, I get an error message: `TypeError: process() missing 1 required positional argument: 'objects'`. I followed these instructions:
```
from twarc_csv import DataFrameConverter
json_objects = [...]
df = DataFrameConverter.process(json_objects)
```
passing converter a tweet or a page scraped as described in examples.
What am I doing wrong or can I not use this at all as I would like to?
|
1.0
|
DataFrameConverter to single tweet - I have a large set of tweets, and I would like to wrangle and write them to file as I go. DataFrameConverter would be ideal for this, but if I try to pass a tweet to it, I get an error message: `TypeError: process() missing 1 required positional argument: 'objects'`. I followed these instructions:
```
from twarc_csv import DataFrameConverter
json_objects = [...]
df = DataFrameConverter.process(json_objects)
```
passing converter a tweet or a page scraped as described in examples.
What am I doing wrong or can I not use this at all as I would like to?
|
non_priority
|
dataframeconverter to single tweet i have a large set of tweets and i would like to wrangle and write them to file as i go dataframeconverter would be ideal for this but if i try to pass a tweet to it i get an error message typeerror process missing required positional argument objects i followed these instructions from twarc csv import dataframeconverter json objects df dataframeconverter process json objects passing converter a tweet or a page scraped as described in examples what am i doing wrong or can i not use this at all as i would like to
| 0
|
150,864
| 19,634,070,037
|
IssuesEvent
|
2022-01-08 01:21:55
|
vipinsun/cactus
|
https://api.github.com/repos/vipinsun/cactus
|
opened
|
CVE-2018-25023 (High) detected in smallvec-0.6.14.crate
|
security vulnerability
|
## CVE-2018-25023 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>smallvec-0.6.14.crate</b></p></summary>
<p>'Small vector' optimization: store up to a small number of items on the stack</p>
<p>Library home page: <a href="https://crates.io/api/v1/crates/smallvec/0.6.14/download">https://crates.io/api/v1/crates/smallvec/0.6.14/download</a></p>
<p>
Dependency Hierarchy:
- hashicorp_vault-1.1.0.crate (Root Library)
- reqwest-0.9.24.crate
- hyper-0.12.35.crate
- tokio-reactor-0.1.12.crate
- parking_lot-0.9.0.crate
- parking_lot_core-0.6.2.crate
- :x: **smallvec-0.6.14.crate** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the smallvec crate before 0.6.13 for Rust. It can create an uninitialized value of any type, including a reference type.
<p>Publish Date: 2021-12-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-25023>CVE-2018-25023</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://rustsec.org/advisories/RUSTSEC-2018-0018.html">https://rustsec.org/advisories/RUSTSEC-2018-0018.html</a></p>
<p>Release Date: 2021-12-27</p>
<p>Fix Resolution: smallvec - 0.6.13</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2018-25023 (High) detected in smallvec-0.6.14.crate - ## CVE-2018-25023 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>smallvec-0.6.14.crate</b></p></summary>
<p>'Small vector' optimization: store up to a small number of items on the stack</p>
<p>Library home page: <a href="https://crates.io/api/v1/crates/smallvec/0.6.14/download">https://crates.io/api/v1/crates/smallvec/0.6.14/download</a></p>
<p>
Dependency Hierarchy:
- hashicorp_vault-1.1.0.crate (Root Library)
- reqwest-0.9.24.crate
- hyper-0.12.35.crate
- tokio-reactor-0.1.12.crate
- parking_lot-0.9.0.crate
- parking_lot_core-0.6.2.crate
- :x: **smallvec-0.6.14.crate** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the smallvec crate before 0.6.13 for Rust. It can create an uninitialized value of any type, including a reference type.
<p>Publish Date: 2021-12-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-25023>CVE-2018-25023</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://rustsec.org/advisories/RUSTSEC-2018-0018.html">https://rustsec.org/advisories/RUSTSEC-2018-0018.html</a></p>
<p>Release Date: 2021-12-27</p>
<p>Fix Resolution: smallvec - 0.6.13</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_priority
|
cve high detected in smallvec crate cve high severity vulnerability vulnerable library smallvec crate small vector optimization store up to a small number of items on the stack library home page a href dependency hierarchy hashicorp vault crate root library reqwest crate hyper crate tokio reactor crate parking lot crate parking lot core crate x smallvec crate vulnerable library found in base branch master vulnerability details an issue was discovered in the smallvec crate before for rust it can create an uninitialized value of any type including a reference type publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution smallvec step up your open source security game with whitesource
| 0
|
177,007
| 28,309,841,446
|
IssuesEvent
|
2023-04-10 14:28:04
|
kookmin-sw/capstone-2023-09
|
https://api.github.com/repos/kookmin-sw/capstone-2023-09
|
closed
|
[관리자] 카테고리 관리 화면 개발
|
💄 design
|
## Description
도움받기 도움주기 게시판에서 사용하는 카테고리를 추가및 삭제할수 있는 화면
## To-do
- [ ] 카테고리 조회 화면
- [ ] 카테고리 수정 기능
- [ ] 카테고리 삭제 기능
|
1.0
|
[관리자] 카테고리 관리 화면 개발 - ## Description
도움받기 도움주기 게시판에서 사용하는 카테고리를 추가및 삭제할수 있는 화면
## To-do
- [ ] 카테고리 조회 화면
- [ ] 카테고리 수정 기능
- [ ] 카테고리 삭제 기능
|
non_priority
|
카테고리 관리 화면 개발 description 도움받기 도움주기 게시판에서 사용하는 카테고리를 추가및 삭제할수 있는 화면 to do 카테고리 조회 화면 카테고리 수정 기능 카테고리 삭제 기능
| 0
|
47,388
| 6,050,920,224
|
IssuesEvent
|
2017-06-12 22:15:27
|
Microsoft/TypeScript
|
https://api.github.com/repos/Microsoft/TypeScript
|
closed
|
Design Meeting Notes, 6/2/2017
|
Design Notes
|
# Propagated Inference for Uninstantiated (Free) Type Parameters (#9366)
```ts
function compose<A, B, C>(f: (x: A) => B, g: (y: B) => C): (x: A) => C {
// ...
}
let boxElements: (a: string) => { value: string[] } =
compose(x => [x], y => { value: y });
```
Currently this works.
Now we ask, what if we wanted `boxElements` to operate on *any* type?
```ts
function compose<A, B, C>(f: (x: A) => B, g: (y: B) => C): (x: A) => C {
// ...
}
let boxElements: <T>(a: T) => { value: T[] } =
compose(x => [x], y => { value: y });
```
Currently this doesn't work well - when we get the contextual type for `x`, we get it from the *erased* signature of `<T>(a: T) => { value: T[] }`, which is really just `(a: any) => { value: any[] }`.
We will have a PR that will help fix this.
Q: What about constraints?
A: Constraints will be carried through.
We will still have some problems with compositional patterns *without* explicit types.
For example:
```ts
declare function compose<A, B, C>(f: (x: A) => B, g: (y: B) => C): (x: A) => C;
declare function box<T>(x: T): { value: T };
declare function list<U>(x: U): U[];
compose(list, box);
```
The problem is that TypeScript draws the following inferences:
Type Parameter | Inferences
---------------|-----------
`A` | `U`
`B` | `U[]`, `T`
`C` | `{ value: T }`
The way we walk through arguments (simple and left-to-right) and draw inferences simply isn't sufficient - we need to find a way to *unify* these type parameters.
But it's not just a matter of throwing unification at the type system. So the current question is how do we introduce some unification to the current process. Perhaps it will be a "last resort" process. And this will be exploratory work.
@gcnew did do great work, but pointed out many of the difficulties and roadblocks you can run into.
# Changing default target to ES5
- Very strange that we wouldn't simply move the the higher one?
- Why take a breaking change when we'll need to make another breaking change in the future?
- Are there a lot of people running into this problem?
- Doesn't appear that a lot of people have been complaining about it.
# Pure annotation in downlevel emits
- Currently, Uglify doesn't understand when our class emit.
- It would be great if Uglify could operate on the ES2015 code.
- We just want to tell other tools it's a class - onus of determining side-effects being on TypeScript is probably more questionable than an optimizer doing so.
- Decision: emit JSDoc `@class` comment unconditionally.
# Lib reference directive (#15780)
- `/// <reference lib="name" />`
- Problem: things like corejs on DefinitelyTyped conflict with compiler defaults.
- This means could simply be reduced to `/// <reference lib="es2015.d.ts" />`.
- Q: What about `/// <reference no-default-lib="true"/>`?
- This PR ignores any `/// <reference no-default-lib="true"/>` comments.
- Could also do `/// <reference lib="..." />` in `lib.d.ts` as well.
- Let's do that.
- What about Daniel's lib versioning idea of publishing to `@types`?
- Example: DOM APIs require new logic - suddenly breaks certain builds.
- People would be able to go to a specific version of a lib if they got broken, or just lock down all their dependencies.
- Also might allow us to bring `node.d.ts` in.
|
1.0
|
Design Meeting Notes, 6/2/2017 - # Propagated Inference for Uninstantiated (Free) Type Parameters (#9366)
```ts
function compose<A, B, C>(f: (x: A) => B, g: (y: B) => C): (x: A) => C {
// ...
}
let boxElements: (a: string) => { value: string[] } =
compose(x => [x], y => { value: y });
```
Currently this works.
Now we ask, what if we wanted `boxElements` to operate on *any* type?
```ts
function compose<A, B, C>(f: (x: A) => B, g: (y: B) => C): (x: A) => C {
// ...
}
let boxElements: <T>(a: T) => { value: T[] } =
compose(x => [x], y => { value: y });
```
Currently this doesn't work well - when we get the contextual type for `x`, we get it from the *erased* signature of `<T>(a: T) => { value: T[] }`, which is really just `(a: any) => { value: any[] }`.
We will have a PR that will help fix this.
Q: What about constraints?
A: Constraints will be carried through.
We will still have some problems with compositional patterns *without* explicit types.
For example:
```ts
declare function compose<A, B, C>(f: (x: A) => B, g: (y: B) => C): (x: A) => C;
declare function box<T>(x: T): { value: T };
declare function list<U>(x: U): U[];
compose(list, box);
```
The problem is that TypeScript draws the following inferences:
Type Parameter | Inferences
---------------|-----------
`A` | `U`
`B` | `U[]`, `T`
`C` | `{ value: T }`
The way we walk through arguments (simple and left-to-right) and draw inferences simply isn't sufficient - we need to find a way to *unify* these type parameters.
But it's not just a matter of throwing unification at the type system. So the current question is how do we introduce some unification to the current process. Perhaps it will be a "last resort" process. And this will be exploratory work.
@gcnew did do great work, but pointed out many of the difficulties and roadblocks you can run into.
# Changing default target to ES5
- Very strange that we wouldn't simply move the the higher one?
- Why take a breaking change when we'll need to make another breaking change in the future?
- Are there a lot of people running into this problem?
- Doesn't appear that a lot of people have been complaining about it.
# Pure annotation in downlevel emits
- Currently, Uglify doesn't understand when our class emit.
- It would be great if Uglify could operate on the ES2015 code.
- We just want to tell other tools it's a class - onus of determining side-effects being on TypeScript is probably more questionable than an optimizer doing so.
- Decision: emit JSDoc `@class` comment unconditionally.
# Lib reference directive (#15780)
- `/// <reference lib="name" />`
- Problem: things like corejs on DefinitelyTyped conflict with compiler defaults.
- This means could simply be reduced to `/// <reference lib="es2015.d.ts" />`.
- Q: What about `/// <reference no-default-lib="true"/>`?
- This PR ignores any `/// <reference no-default-lib="true"/>` comments.
- Could also do `/// <reference lib="..." />` in `lib.d.ts` as well.
- Let's do that.
- What about Daniel's lib versioning idea of publishing to `@types`?
- Example: DOM APIs require new logic - suddenly breaks certain builds.
- People would be able to go to a specific version of a lib if they got broken, or just lock down all their dependencies.
- Also might allow us to bring `node.d.ts` in.
|
non_priority
|
design meeting notes propagated inference for uninstantiated free type parameters ts function compose f x a b g y b c x a c let boxelements a string value string compose x y value y currently this works now we ask what if we wanted boxelements to operate on any type ts function compose f x a b g y b c x a c let boxelements a t value t compose x y value y currently this doesn t work well when we get the contextual type for x we get it from the erased signature of a t value t which is really just a any value any we will have a pr that will help fix this q what about constraints a constraints will be carried through we will still have some problems with compositional patterns without explicit types for example ts declare function compose f x a b g y b c x a c declare function box x t value t declare function list x u u compose list box the problem is that typescript draws the following inferences type parameter inferences a u b u t c value t the way we walk through arguments simple and left to right and draw inferences simply isn t sufficient we need to find a way to unify these type parameters but it s not just a matter of throwing unification at the type system so the current question is how do we introduce some unification to the current process perhaps it will be a last resort process and this will be exploratory work gcnew did do great work but pointed out many of the difficulties and roadblocks you can run into changing default target to very strange that we wouldn t simply move the the higher one why take a breaking change when we ll need to make another breaking change in the future are there a lot of people running into this problem doesn t appear that a lot of people have been complaining about it pure annotation in downlevel emits currently uglify doesn t understand when our class emit it would be great if uglify could operate on the code we just want to tell other tools it s a class onus of determining side effects being on typescript is probably more questionable than an optimizer doing so decision emit jsdoc class comment unconditionally lib reference directive problem things like corejs on definitelytyped conflict with compiler defaults this means could simply be reduced to q what about this pr ignores any comments could also do in lib d ts as well let s do that what about daniel s lib versioning idea of publishing to types example dom apis require new logic suddenly breaks certain builds people would be able to go to a specific version of a lib if they got broken or just lock down all their dependencies also might allow us to bring node d ts in
| 0
|
386,903
| 11,452,516,244
|
IssuesEvent
|
2020-02-06 13:50:14
|
ooni/probe
|
https://api.github.com/repos/ooni/probe
|
opened
|
Properly handle SIGTERM signal on windows
|
bug ooni/probe-cli ooni/probe-desktop platform/windows priority/high
|
It has been reported that the SIGTERM signal is not getting trapped on windows and therefore when you try to terminate an OONI Probe run on windows it does not shutdown cleanly as it does on macOS.
We should check to see how we can handle this properly on windows.
|
1.0
|
Properly handle SIGTERM signal on windows - It has been reported that the SIGTERM signal is not getting trapped on windows and therefore when you try to terminate an OONI Probe run on windows it does not shutdown cleanly as it does on macOS.
We should check to see how we can handle this properly on windows.
|
priority
|
properly handle sigterm signal on windows it has been reported that the sigterm signal is not getting trapped on windows and therefore when you try to terminate an ooni probe run on windows it does not shutdown cleanly as it does on macos we should check to see how we can handle this properly on windows
| 1
|
113,082
| 11,787,254,264
|
IssuesEvent
|
2020-03-17 13:44:59
|
sei-nope/inStock-client
|
https://api.github.com/repos/sei-nope/inStock-client
|
closed
|
Update README
|
documentation
|
- An explanation of the what the app does and how it works.
- A link to the other repo
- A link to both deployed sites
- List of technologies used
- List unsolved problems which would be fixed in future iterations.
- Document your planning, process and problem-solving strategy
- Complete the repository Description field and Website field with a meaningful sentence description of the application and link to the live URL
Your front-end repo's README should also have:
- Link to wireframes and user stories
|
1.0
|
Update README - - An explanation of the what the app does and how it works.
- A link to the other repo
- A link to both deployed sites
- List of technologies used
- List unsolved problems which would be fixed in future iterations.
- Document your planning, process and problem-solving strategy
- Complete the repository Description field and Website field with a meaningful sentence description of the application and link to the live URL
Your front-end repo's README should also have:
- Link to wireframes and user stories
|
non_priority
|
update readme an explanation of the what the app does and how it works a link to the other repo a link to both deployed sites list of technologies used list unsolved problems which would be fixed in future iterations document your planning process and problem solving strategy complete the repository description field and website field with a meaningful sentence description of the application and link to the live url your front end repo s readme should also have link to wireframes and user stories
| 0
|
101,210
| 21,629,193,062
|
IssuesEvent
|
2022-05-05 07:54:36
|
withfig/fig
|
https://api.github.com/repos/withfig/fig
|
closed
|
Show custom aliases in autocompletion
|
type:feature-request codebase:autocomplete-app
|
**Is your feature request related to a problem? Please describe.**
Given an alias defined as such:
```
alias gc="git commit -m" # make a commit with a message (gc [message])
```
**Describe the solution you'd like**
* I'd love to see the comment parsed and turned into the suggestion at the bottom of the autocomplete window.
* I'd also love to see all my aliases listed out in the autocomplete window, so when I start to type `g` they are selectable.
**Describe alternatives you've considered**
The only one I can think of is defining the rules around my aliases separately in the fig autocompletion settings. This is not ideal because then I'd be managing them in two spots.
**Additional context**
N/A
|
1.0
|
Show custom aliases in autocompletion - **Is your feature request related to a problem? Please describe.**
Given an alias defined as such:
```
alias gc="git commit -m" # make a commit with a message (gc [message])
```
**Describe the solution you'd like**
* I'd love to see the comment parsed and turned into the suggestion at the bottom of the autocomplete window.
* I'd also love to see all my aliases listed out in the autocomplete window, so when I start to type `g` they are selectable.
**Describe alternatives you've considered**
The only one I can think of is defining the rules around my aliases separately in the fig autocompletion settings. This is not ideal because then I'd be managing them in two spots.
**Additional context**
N/A
|
non_priority
|
show custom aliases in autocompletion is your feature request related to a problem please describe given an alias defined as such alias gc git commit m make a commit with a message gc describe the solution you d like i d love to see the comment parsed and turned into the suggestion at the bottom of the autocomplete window i d also love to see all my aliases listed out in the autocomplete window so when i start to type g they are selectable describe alternatives you ve considered the only one i can think of is defining the rules around my aliases separately in the fig autocompletion settings this is not ideal because then i d be managing them in two spots additional context n a
| 0
|
8,234
| 10,311,685,778
|
IssuesEvent
|
2019-08-29 17:58:40
|
cobalt-org/liquid-rust
|
https://api.github.com/repos/cobalt-org/liquid-rust
|
closed
|
Implement where filter
|
enhancement good first issue std-compatibility
|
Example test:
```rust
assert_eq!([{ v!("a"): v!("ok") }], filters!(where, { v!("a"): v!("ok") }, "a", r#"ok")#);
```
- [Where filters live](https://github.com/cobalt-org/liquid-rust/blob/master/src/filters/mod.rs)
- [Ruby implementation](https://github.com/Shopify/liquid/blob/master/lib/liquid/standardfilters.rb)
- [Relevant tests](https://github.com/cobalt-org/liquid-rust/search?q=liquid+291&unscoped_q=liquid+291)
- Once this is working, ideally the test will start failing. We would just need to remove the `#[should_panic]` on it
- If the test doesn't start failing, please temporarily remove `#[should_panic]`, and find or create an issue for the new failure
|
True
|
Implement where filter - Example test:
```rust
assert_eq!([{ v!("a"): v!("ok") }], filters!(where, { v!("a"): v!("ok") }, "a", r#"ok")#);
```
- [Where filters live](https://github.com/cobalt-org/liquid-rust/blob/master/src/filters/mod.rs)
- [Ruby implementation](https://github.com/Shopify/liquid/blob/master/lib/liquid/standardfilters.rb)
- [Relevant tests](https://github.com/cobalt-org/liquid-rust/search?q=liquid+291&unscoped_q=liquid+291)
- Once this is working, ideally the test will start failing. We would just need to remove the `#[should_panic]` on it
- If the test doesn't start failing, please temporarily remove `#[should_panic]`, and find or create an issue for the new failure
|
non_priority
|
implement where filter example test rust assert eq filters where v a v ok a r ok once this is working ideally the test will start failing we would just need to remove the on it if the test doesn t start failing please temporarily remove and find or create an issue for the new failure
| 0
|
483,342
| 13,923,154,989
|
IssuesEvent
|
2020-10-21 14:06:44
|
zephyrproject-rtos/zephyr
|
https://api.github.com/repos/zephyrproject-rtos/zephyr
|
closed
|
[Coverity CID :214214] Uninitialized pointer read in tests/benchmarks/data_structure_perf/rbtree_perf/src/rbtree_perf.c
|
Coverity bug priority: low
|
Static code scan issues found in file:
https://github.com/zephyrproject-rtos/zephyr/tree/27e1fd69f987d3082229b3368a87d59ffdcab5a4/tests/benchmarks/data_structure_perf/rbtree_perf/src/rbtree_perf.c#L58
Category: Memory - illegal accesses
Function: `test_rbtree_container`
Component: Tests
CID: [214214](https://scan9.coverity.com/reports.htm#v29726/p12996/mergedDefectId=214214)
Details:
```
52 struct rbnode *foreach_node;
53 struct container_node tree_node[10];
54
55 test_tree_l.lessthan_fn = node_lessthan;
56 for (uint32_t i = 0; i < ARRAY_SIZE(tree_node); i++) {
57 tree_node[i].value = i;
>>> CID 214214: Memory - illegal accesses (UNINIT)
>>> Using uninitialized value "test_tree_l.max_depth" when calling "rb_insert".
58 rb_insert(&test_tree_l, &tree_node[i].node);
59 }
60
61 RB_FOR_EACH(&test_tree_l, foreach_node) {
62 zassert_true(CONTAINER_OF(foreach_node, struct container_node,
63 node)->value == count, "RB_FOR_EACH failed");
```
Please fix or provide comments in coverity using the link:
https://scan9.coverity.com/reports.htm#v32951/p12996.
Note: This issue was created automatically. Priority was set based on classification
of the file affected and the impact field in coverity. Assignees were set using the CODEOWNERS file.
|
1.0
|
[Coverity CID :214214] Uninitialized pointer read in tests/benchmarks/data_structure_perf/rbtree_perf/src/rbtree_perf.c -
Static code scan issues found in file:
https://github.com/zephyrproject-rtos/zephyr/tree/27e1fd69f987d3082229b3368a87d59ffdcab5a4/tests/benchmarks/data_structure_perf/rbtree_perf/src/rbtree_perf.c#L58
Category: Memory - illegal accesses
Function: `test_rbtree_container`
Component: Tests
CID: [214214](https://scan9.coverity.com/reports.htm#v29726/p12996/mergedDefectId=214214)
Details:
```
52 struct rbnode *foreach_node;
53 struct container_node tree_node[10];
54
55 test_tree_l.lessthan_fn = node_lessthan;
56 for (uint32_t i = 0; i < ARRAY_SIZE(tree_node); i++) {
57 tree_node[i].value = i;
>>> CID 214214: Memory - illegal accesses (UNINIT)
>>> Using uninitialized value "test_tree_l.max_depth" when calling "rb_insert".
58 rb_insert(&test_tree_l, &tree_node[i].node);
59 }
60
61 RB_FOR_EACH(&test_tree_l, foreach_node) {
62 zassert_true(CONTAINER_OF(foreach_node, struct container_node,
63 node)->value == count, "RB_FOR_EACH failed");
```
Please fix or provide comments in coverity using the link:
https://scan9.coverity.com/reports.htm#v32951/p12996.
Note: This issue was created automatically. Priority was set based on classification
of the file affected and the impact field in coverity. Assignees were set using the CODEOWNERS file.
|
priority
|
uninitialized pointer read in tests benchmarks data structure perf rbtree perf src rbtree perf c static code scan issues found in file category memory illegal accesses function test rbtree container component tests cid details struct rbnode foreach node struct container node tree node test tree l lessthan fn node lessthan for t i i array size tree node i tree node value i cid memory illegal accesses uninit using uninitialized value test tree l max depth when calling rb insert rb insert test tree l tree node node rb for each test tree l foreach node zassert true container of foreach node struct container node node value count rb for each failed please fix or provide comments in coverity using the link note this issue was created automatically priority was set based on classification of the file affected and the impact field in coverity assignees were set using the codeowners file
| 1
|
646,861
| 21,077,065,888
|
IssuesEvent
|
2022-04-02 09:52:48
|
apache/incubator-kyuubi
|
https://api.github.com/repos/apache/incubator-kyuubi
|
closed
|
[Bug] The running query will not update the duration of the page
|
kind:bug priority:major
|
### Code of Conduct
- [X] I agree to follow this project's [Code of Conduct](https://www.apache.org/foundation/policies/conduct)
### Search before asking
- [X] I have searched in the [issues](https://github.com/apache/incubator-kyuubi/issues?q=is%3Aissue) and found no similar issues.
### Describe the bug
The default value of completedTime is 0, and the calculation of duration only considers the case where completedTime is -1.
org.apache.kyuubi.operation.AbstractOperation
```scala
protected var completedTime: Long = _
````
org.apache.kyuubi.engine.spark.events.SparkOperationEvent
```scala
def duration: Long = {
if (completeTime == -1L) {
System.currentTimeMillis - createTime
} else {
completeTime - createTime
}
}
```
### Affects Version(s)
1.5.0
### Kyuubi Server Log Output
_No response_
### Kyuubi Engine Log Output
_No response_
### Kyuubi Server Configurations
_No response_
### Kyuubi Engine Configurations
_No response_
### Additional context
_No response_
### Are you willing to submit PR?
- [X] Yes I am willing to submit a PR!
|
1.0
|
[Bug] The running query will not update the duration of the page - ### Code of Conduct
- [X] I agree to follow this project's [Code of Conduct](https://www.apache.org/foundation/policies/conduct)
### Search before asking
- [X] I have searched in the [issues](https://github.com/apache/incubator-kyuubi/issues?q=is%3Aissue) and found no similar issues.
### Describe the bug
The default value of completedTime is 0, and the calculation of duration only considers the case where completedTime is -1.
org.apache.kyuubi.operation.AbstractOperation
```scala
protected var completedTime: Long = _
````
org.apache.kyuubi.engine.spark.events.SparkOperationEvent
```scala
def duration: Long = {
if (completeTime == -1L) {
System.currentTimeMillis - createTime
} else {
completeTime - createTime
}
}
```
### Affects Version(s)
1.5.0
### Kyuubi Server Log Output
_No response_
### Kyuubi Engine Log Output
_No response_
### Kyuubi Server Configurations
_No response_
### Kyuubi Engine Configurations
_No response_
### Additional context
_No response_
### Are you willing to submit PR?
- [X] Yes I am willing to submit a PR!
|
priority
|
the running query will not update the duration of the page code of conduct i agree to follow this project s search before asking i have searched in the and found no similar issues describe the bug the default value of completedtime is and the calculation of duration only considers the case where completedtime is org apache kyuubi operation abstractoperation scala protected var completedtime long org apache kyuubi engine spark events sparkoperationevent scala def duration long if completetime system currenttimemillis createtime else completetime createtime affects version s kyuubi server log output no response kyuubi engine log output no response kyuubi server configurations no response kyuubi engine configurations no response additional context no response are you willing to submit pr yes i am willing to submit a pr
| 1
|
159,258
| 20,048,345,438
|
IssuesEvent
|
2022-02-03 01:07:30
|
kapseliboi/owner-profiles
|
https://api.github.com/repos/kapseliboi/owner-profiles
|
opened
|
CVE-2018-1107 (Medium) detected in is-my-json-valid-2.17.1.tgz
|
security vulnerability
|
## CVE-2018-1107 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>is-my-json-valid-2.17.1.tgz</b></p></summary>
<p>A JSONSchema validator that uses code generation to be extremely fast</p>
<p>Library home page: <a href="https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.17.1.tgz">https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.17.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/is-my-json-valid/package.json</p>
<p>
Dependency Hierarchy:
- standard-10.0.3.tgz (Root Library)
- eslint-3.19.0.tgz
- :x: **is-my-json-valid-2.17.1.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
It was discovered that the is-my-json-valid JavaScript library used an inefficient regular expression to validate JSON fields defined to have email format. A specially crafted JSON file could cause it to consume an excessive amount of CPU time when validated.
<p>Publish Date: 2021-03-30
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1107>CVE-2018-1107</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-4hpf-3wq7-5rpr">https://github.com/advisories/GHSA-4hpf-3wq7-5rpr</a></p>
<p>Release Date: 2021-03-30</p>
<p>Fix Resolution: is-my-json-valid - 1.4.1, 2.17.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2018-1107 (Medium) detected in is-my-json-valid-2.17.1.tgz - ## CVE-2018-1107 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>is-my-json-valid-2.17.1.tgz</b></p></summary>
<p>A JSONSchema validator that uses code generation to be extremely fast</p>
<p>Library home page: <a href="https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.17.1.tgz">https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.17.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/is-my-json-valid/package.json</p>
<p>
Dependency Hierarchy:
- standard-10.0.3.tgz (Root Library)
- eslint-3.19.0.tgz
- :x: **is-my-json-valid-2.17.1.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
It was discovered that the is-my-json-valid JavaScript library used an inefficient regular expression to validate JSON fields defined to have email format. A specially crafted JSON file could cause it to consume an excessive amount of CPU time when validated.
<p>Publish Date: 2021-03-30
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1107>CVE-2018-1107</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-4hpf-3wq7-5rpr">https://github.com/advisories/GHSA-4hpf-3wq7-5rpr</a></p>
<p>Release Date: 2021-03-30</p>
<p>Fix Resolution: is-my-json-valid - 1.4.1, 2.17.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_priority
|
cve medium detected in is my json valid tgz cve medium severity vulnerability vulnerable library is my json valid tgz a jsonschema validator that uses code generation to be extremely fast library home page a href path to dependency file package json path to vulnerable library node modules is my json valid package json dependency hierarchy standard tgz root library eslint tgz x is my json valid tgz vulnerable library found in base branch master vulnerability details it was discovered that the is my json valid javascript library used an inefficient regular expression to validate json fields defined to have email format a specially crafted json file could cause it to consume an excessive amount of cpu time when validated publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution is my json valid step up your open source security game with whitesource
| 0
|
59,851
| 7,298,719,650
|
IssuesEvent
|
2018-02-26 17:48:31
|
thePlums/web-app
|
https://api.github.com/repos/thePlums/web-app
|
closed
|
Set a remind for individual item
|
design
|
Pretty straight forward, he went to the alarm icon right away and undstood the "Select" instructions. Recommended perhaps the icon is on each item/collection instead.
|
1.0
|
Set a remind for individual item - Pretty straight forward, he went to the alarm icon right away and undstood the "Select" instructions. Recommended perhaps the icon is on each item/collection instead.
|
non_priority
|
set a remind for individual item pretty straight forward he went to the alarm icon right away and undstood the select instructions recommended perhaps the icon is on each item collection instead
| 0
|
242,750
| 7,846,471,130
|
IssuesEvent
|
2018-06-19 15:34:12
|
whole-tale/girder_wholetale
|
https://api.github.com/repos/whole-tale/girder_wholetale
|
closed
|
Allow Tales creation to accept multiple files and folders
|
area/tale-creation enhancement priority/blocker
|
The current endpoint for creating a new tale only allows to add a single folder, following the previous specification.
```
tale {
folderId (string): ID of a data folder used by the Tale , // <--- HERE
imageId (string): ID of a WT Image used by the Tale ,
title (string, optional): Title of the Tale ,
// ... other fileds
}
```
After it was decided to allow multiple files and folders, the endpoint needs to change from a _string_ to an _array_.
|
1.0
|
Allow Tales creation to accept multiple files and folders - The current endpoint for creating a new tale only allows to add a single folder, following the previous specification.
```
tale {
folderId (string): ID of a data folder used by the Tale , // <--- HERE
imageId (string): ID of a WT Image used by the Tale ,
title (string, optional): Title of the Tale ,
// ... other fileds
}
```
After it was decided to allow multiple files and folders, the endpoint needs to change from a _string_ to an _array_.
|
priority
|
allow tales creation to accept multiple files and folders the current endpoint for creating a new tale only allows to add a single folder following the previous specification tale folderid string id of a data folder used by the tale here imageid string id of a wt image used by the tale title string optional title of the tale other fileds after it was decided to allow multiple files and folders the endpoint needs to change from a string to an array
| 1
|
86,308
| 10,481,956,097
|
IssuesEvent
|
2019-09-24 10:49:34
|
lhy0403/Qv2ray
|
https://api.github.com/repos/lhy0403/Qv2ray
|
opened
|
!! Version 2.0 功能建议 !!
|
Documentations help wanted
|
# Version 2.0
Version 2.0 将会包含以下功能更新
## 复杂连接编辑
- [ ] 路由添加
- [ ] 路由编辑
- [ ] 入站编辑
- [ ] 增加更多出站编辑选项
- [ ] 自动启动对应编辑器(简单出站编辑 / 复杂路由编辑)
- [ ] 自动生成 tProxy 的 iptables 命令(应该挺难的)
- [ ] JSON 编辑器
## 下载模块
- [ ] 自动下载 v2ray core
- [ ] 自动下载更新
## v2ray API
- [ ] 流量统计
## 订阅功能
- [ ] 订阅功能
## 分享功能
- [ ] 一键生成 vmess://
- [ ] 生成二维码
--------------
有没有别的建议可以提出来
|
1.0
|
!! Version 2.0 功能建议 !! - # Version 2.0
Version 2.0 将会包含以下功能更新
## 复杂连接编辑
- [ ] 路由添加
- [ ] 路由编辑
- [ ] 入站编辑
- [ ] 增加更多出站编辑选项
- [ ] 自动启动对应编辑器(简单出站编辑 / 复杂路由编辑)
- [ ] 自动生成 tProxy 的 iptables 命令(应该挺难的)
- [ ] JSON 编辑器
## 下载模块
- [ ] 自动下载 v2ray core
- [ ] 自动下载更新
## v2ray API
- [ ] 流量统计
## 订阅功能
- [ ] 订阅功能
## 分享功能
- [ ] 一键生成 vmess://
- [ ] 生成二维码
--------------
有没有别的建议可以提出来
|
non_priority
|
!! version 功能建议 !! version version 将会包含以下功能更新 复杂连接编辑 路由添加 路由编辑 入站编辑 增加更多出站编辑选项 自动启动对应编辑器(简单出站编辑 复杂路由编辑) 自动生成 tproxy 的 iptables 命令(应该挺难的) json 编辑器 下载模块 自动下载 core 自动下载更新 api 流量统计 订阅功能 订阅功能 分享功能 一键生成 vmess 生成二维码 有没有别的建议可以提出来
| 0
|
41,965
| 16,990,874,021
|
IssuesEvent
|
2021-06-30 20:16:52
|
cityofaustin/atd-data-tech
|
https://api.github.com/repos/cityofaustin/atd-data-tech
|
closed
|
TIA Case Loads: Filter out Inactive Status
|
Product: TDS Portal Project: TIA Module Service: Apps Type: Enhancement Workgroup: TDSD
|
On Area Case Load & Reviewer Case Load, filter out cases w/ status of Inactive.
|
1.0
|
TIA Case Loads: Filter out Inactive Status - On Area Case Load & Reviewer Case Load, filter out cases w/ status of Inactive.
|
non_priority
|
tia case loads filter out inactive status on area case load reviewer case load filter out cases w status of inactive
| 0
|
10,236
| 4,729,201,033
|
IssuesEvent
|
2016-10-18 18:02:20
|
blackbaud/sky-pages-cli
|
https://api.github.com/repos/blackbaud/sky-pages-cli
|
closed
|
Allow appBase to be overridden
|
builder
|
In https://github.com/blackbaud/sky-pages-out-skyux2/blob/master/config/webpack/common.webpack.config.js#L17-L20, first check for the `name` property in `sky-pages.json` (without running through replace). If it doesn't exist, use the current behavior.
|
1.0
|
Allow appBase to be overridden - In https://github.com/blackbaud/sky-pages-out-skyux2/blob/master/config/webpack/common.webpack.config.js#L17-L20, first check for the `name` property in `sky-pages.json` (without running through replace). If it doesn't exist, use the current behavior.
|
non_priority
|
allow appbase to be overridden in first check for the name property in sky pages json without running through replace if it doesn t exist use the current behavior
| 0
|
126,734
| 17,970,664,478
|
IssuesEvent
|
2021-09-14 01:15:41
|
Sh2dowFi3nd/Test_2
|
https://api.github.com/repos/Sh2dowFi3nd/Test_2
|
opened
|
CVE-2020-15522 (Medium) detected in bcprov-jdk15on-1.50.jar
|
security vulnerability
|
## CVE-2020-15522 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bcprov-jdk15on-1.50.jar</b></p></summary>
<p>The Bouncy Castle Crypto package is a Java implementation of cryptographic algorithms. This jar contains JCE provider and lightweight API for the Bouncy Castle Cryptography APIs for JDK 1.5 to JDK 1.7.</p>
<p>Library home page: <a href="http://www.bouncycastle.org/java.html">http://www.bouncycastle.org/java.html</a></p>
<p>Path to dependency file: /Test_2/fs-agent-master/fs-agent-master/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.50/bcprov-jdk15on-1.50.jar</p>
<p>
Dependency Hierarchy:
- redline-1.2.1.jar (Root Library)
- bcpg-jdk15on-1.50.jar
- :x: **bcprov-jdk15on-1.50.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Bouncy Castle BC Java before 1.66, BC C# .NET before 1.8.7, BC-FJA before 1.0.1.2, 1.0.2.1, and BC-FNA before 1.0.1.1 have a timing issue within the EC math library that can expose information about the private key when an attacker is able to observe timing information for the generation of multiple deterministic ECDSA signatures.
<p>Publish Date: 2021-05-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15522>CVE-2020-15522</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-15522">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-15522</a></p>
<p>Release Date: 2021-05-20</p>
<p>Fix Resolution: C#- release-1.8.7, Java- 1.66</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-15522 (Medium) detected in bcprov-jdk15on-1.50.jar - ## CVE-2020-15522 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bcprov-jdk15on-1.50.jar</b></p></summary>
<p>The Bouncy Castle Crypto package is a Java implementation of cryptographic algorithms. This jar contains JCE provider and lightweight API for the Bouncy Castle Cryptography APIs for JDK 1.5 to JDK 1.7.</p>
<p>Library home page: <a href="http://www.bouncycastle.org/java.html">http://www.bouncycastle.org/java.html</a></p>
<p>Path to dependency file: /Test_2/fs-agent-master/fs-agent-master/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.50/bcprov-jdk15on-1.50.jar</p>
<p>
Dependency Hierarchy:
- redline-1.2.1.jar (Root Library)
- bcpg-jdk15on-1.50.jar
- :x: **bcprov-jdk15on-1.50.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Bouncy Castle BC Java before 1.66, BC C# .NET before 1.8.7, BC-FJA before 1.0.1.2, 1.0.2.1, and BC-FNA before 1.0.1.1 have a timing issue within the EC math library that can expose information about the private key when an attacker is able to observe timing information for the generation of multiple deterministic ECDSA signatures.
<p>Publish Date: 2021-05-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15522>CVE-2020-15522</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-15522">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-15522</a></p>
<p>Release Date: 2021-05-20</p>
<p>Fix Resolution: C#- release-1.8.7, Java- 1.66</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_priority
|
cve medium detected in bcprov jar cve medium severity vulnerability vulnerable library bcprov jar the bouncy castle crypto package is a java implementation of cryptographic algorithms this jar contains jce provider and lightweight api for the bouncy castle cryptography apis for jdk to jdk library home page a href path to dependency file test fs agent master fs agent master pom xml path to vulnerable library root repository org bouncycastle bcprov bcprov jar dependency hierarchy redline jar root library bcpg jar x bcprov jar vulnerable library vulnerability details bouncy castle bc java before bc c net before bc fja before and bc fna before have a timing issue within the ec math library that can expose information about the private key when an attacker is able to observe timing information for the generation of multiple deterministic ecdsa signatures publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution c release java step up your open source security game with whitesource
| 0
|
258,325
| 22,302,896,782
|
IssuesEvent
|
2022-06-13 10:19:37
|
opencollective/opencollective
|
https://api.github.com/repos/opencollective/opencollective
|
closed
|
Add tests for member invitation mutations
|
complexity → simple api test bounty $100
|
Given that these mutations represent quite important features, we want them to be covered.
Here's the skeleton that we want to implement in `test/server/graphql/v2/mutation/MemberInvitationMutations.test.ts`:
```es6
describe('MemberInvitationMutations', () => {
describe('inviteMember', () => {
it('should create a new member invitation and its related activity', () => {});
it('must be authenticated as an admin of the collective', () => {});
it('can only add with role accountant, admin or member', () => {});
it('can only add with a user account', () => {});
});
describe('editMemberInvitation', () => {
it('should edit the role, description and since and document the changes in an activity', () => {});
it('must be authenticated as an admin of the collective', () => {});
it('can only update role to accountant, admin or member', () => {});
});
describe('replyToMemberInvitation', () => {
it('can accept the invitation and document the changes in an activity', () => {});
it('can decline the invitation and document the changes in an activity', () => {});
it('must be authenticated as the invited user', () => {});
});
});
```
*Tip: run with `npm run test -- --watch test/server/graphql/v2/mutation/MemberInvitationMutations.test.ts`*
|
1.0
|
Add tests for member invitation mutations - Given that these mutations represent quite important features, we want them to be covered.
Here's the skeleton that we want to implement in `test/server/graphql/v2/mutation/MemberInvitationMutations.test.ts`:
```es6
describe('MemberInvitationMutations', () => {
describe('inviteMember', () => {
it('should create a new member invitation and its related activity', () => {});
it('must be authenticated as an admin of the collective', () => {});
it('can only add with role accountant, admin or member', () => {});
it('can only add with a user account', () => {});
});
describe('editMemberInvitation', () => {
it('should edit the role, description and since and document the changes in an activity', () => {});
it('must be authenticated as an admin of the collective', () => {});
it('can only update role to accountant, admin or member', () => {});
});
describe('replyToMemberInvitation', () => {
it('can accept the invitation and document the changes in an activity', () => {});
it('can decline the invitation and document the changes in an activity', () => {});
it('must be authenticated as the invited user', () => {});
});
});
```
*Tip: run with `npm run test -- --watch test/server/graphql/v2/mutation/MemberInvitationMutations.test.ts`*
|
non_priority
|
add tests for member invitation mutations given that these mutations represent quite important features we want them to be covered here s the skeleton that we want to implement in test server graphql mutation memberinvitationmutations test ts describe memberinvitationmutations describe invitemember it should create a new member invitation and its related activity it must be authenticated as an admin of the collective it can only add with role accountant admin or member it can only add with a user account describe editmemberinvitation it should edit the role description and since and document the changes in an activity it must be authenticated as an admin of the collective it can only update role to accountant admin or member describe replytomemberinvitation it can accept the invitation and document the changes in an activity it can decline the invitation and document the changes in an activity it must be authenticated as the invited user tip run with npm run test watch test server graphql mutation memberinvitationmutations test ts
| 0
|
423,278
| 12,293,774,409
|
IssuesEvent
|
2020-05-10 20:33:06
|
hildebro/moneysplitter
|
https://api.github.com/repos/hildebro/moneysplitter
|
closed
|
generic callback query handler
|
high priority up next
|
when i change the pattern on buttons, existing menus for users will break.
add new handler at the end with pattern = ALL, so these deprecated buttons redirect to some info screen with button to go back to menu
|
1.0
|
generic callback query handler - when i change the pattern on buttons, existing menus for users will break.
add new handler at the end with pattern = ALL, so these deprecated buttons redirect to some info screen with button to go back to menu
|
priority
|
generic callback query handler when i change the pattern on buttons existing menus for users will break add new handler at the end with pattern all so these deprecated buttons redirect to some info screen with button to go back to menu
| 1
|
310,502
| 9,514,597,215
|
IssuesEvent
|
2019-04-26 01:16:17
|
grpc/grpc
|
https://api.github.com/repos/grpc/grpc
|
closed
|
Python: c-ares resolver with enabled gevent crashes on Linux
|
kind/bug lang/Python priority/P3
|
<!--
This form is for bug reports and feature requests ONLY!
For general questions and troubleshooting, please ask/look for answers here:
- grpc.io mailing list: https://groups.google.com/forum/#!forum/grpc-io
- StackOverflow, with "grpc" tag: https://stackoverflow.com/questions/tagged/grpc
Issues specific to *grpc-java*, *grpc-go*, *grpc-node*, *grpc-dart*, *grpc-web* should be created in the repository they belong to (e.g. https://github.com/grpc/grpc-LANGUAGE/issues/new)
-->
### What version of gRPC and what language are you using?
gRPC 1.20, Python
### What operating system (Linux, Windows,...) and version?
Linux Ubuntu (Python Docker: 3.7-stretch)
### What runtime / compiler are you using (e.g. python version or version of gcc)
Python 3.7
### What did you do?
I'm trying to make a call to the service using expiremental gevent functionality:
```
from gevent import monkey
monkey.patch_all()
import grpc.experimental.gevent
grpc.experimental.gevent.init_gevent()
import grpc
import helloworld_pb2
import helloworld_pb2_grpc
if __name__ == '__main__':
channel = grpc.insecure_channel('hello.domain.com')
stub = helloworld_pb2_grpc.GreeterStub(channel)
response = stub.SayHello(helloworld_pb2.HelloRequest(name='you'))
```
### What did you expect to see?
Response from service.
### What did you see instead?
Call to service crashes:
```
D0424 13:55:39.475437200 308 dns_resolver_ares.cc:436] Using ares dns resolver
I0424 13:55:39.475501500 308 timer_manager.cc:94] Spawn timer thread
I0424 13:55:39.475555300 308 init.cc:164] grpc_init(void)
I0424 13:55:39.475651200 313 timer_manager.cc:259] timers not checked: expect another thread to
I0424 13:55:39.476036000 313 timer_manager.cc:202] sleep until kicked
I0424 13:55:39.476754800 308 completion_queue.cc:503] grpc_completion_queue_create_internal(completion_type=0, polling_type=0)
I0424 13:55:39.476831400 308 completion_queue.cc:503] grpc_completion_queue_create_internal(completion_type=0, polling_type=0)
I0424 13:55:39.477093700 308 ssl_credentials.cc:128] grpc_ssl_credentials_create(pem_root_certs=(null), pem_key_cert_pair=(nil), verify_options=(nil), reserved=(nil))
I0424 13:55:39.477203900 308 secure_channel_create.cc:198] grpc_secure_channel_create(creds=0x55e705cf5aa0, target=hello.domain.com, args=0x7fc78452c2e0, reserved=(nil))
I0424 13:55:39.478266800 308 client_channel.cc:229] chand=0x55e705fb6fe8: update: state=IDLE error="No Error" picker=0x55e705fe5a50
I0424 13:55:39.478413600 308 connectivity_state.cc:164] SET: 0x55e705fb7090 client_channel: IDLE --> IDLE [helper] error=(nil) "No Error"
I0424 13:55:39.478852900 308 client_channel.cc:488] chand=0x55e705fb6fe8: created resolving_lb_policy=0x55e705c443e0
I0424 13:55:39.479489500 308 channel.cc:285] grpc_channel_get_target(channel=0x55e705fb6f20)
I0424 13:55:39.479648000 308 credentials.cc:43] grpc_channel_credentials_release(creds=0x55e705cf5aa0)
I0424 13:55:39.480383100 308 completion_queue.cc:503] grpc_completion_queue_create_internal(completion_type=0, polling_type=0)
I0424 13:55:39.480459800 308 metadata_array.cc:29] grpc_metadata_array_init(array=0x7fc784535168)
I0424 13:55:39.480989600 308 metadata_array.cc:29] grpc_metadata_array_init(array=0x7fc785170ce8)
I0424 13:55:39.481340900 308 call.cc:1910] grpc_call_start_batch(call=0x55e705fe63e0, ops=0x55e705ea3170, nops=6, tag=0x7fc78452cfc0, reserved=(nil))
I0424 13:55:39.481670300 308 call.cc:1516] ops[0]: SEND_INITIAL_METADATA(nil)
I0424 13:55:39.481909800 308 call.cc:1516] ops[1]: SEND_MESSAGE ptr=0x55e705aea420
I0424 13:55:39.482198900 308 call.cc:1516] ops[2]: SEND_CLOSE_FROM_CLIENT
I0424 13:55:39.482334300 308 call.cc:1516] ops[3]: RECV_INITIAL_METADATA ptr=0x7fc784535168
I0424 13:55:39.482369500 308 call.cc:1516] ops[4]: RECV_MESSAGE ptr=0x7fc7845796a0
I0424 13:55:39.482710500 308 call.cc:1516] ops[5]: RECV_STATUS_ON_CLIENT metadata=0x7fc785170ce8 status=0x7fc785170d00 details=0x7fc785170d08
I0424 13:55:39.483026300 308 call_combiner.cc:116] ==> grpc_call_combiner_start() [0x55e705fe63f0] closure=0x55e705fe72c8 [executing batch] error="No Error"
I0424 13:55:39.483197400 308 call_combiner.cc:125] size: 0 -> 1
I0424 13:55:39.483383900 308 call_combiner.cc:134] EXECUTING IMMEDIATELY
I0424 13:55:39.483639100 308 call.cc:615] OP[client-channel:0x55e705fe6d50]: SEND_INITIAL_METADATA{key=3a 70 61 74 68 ':path' value=2f 49 74 65 6d 73 53 65 72 76 69 63 65 2f 67 65 74 5f 61 6c 6c 5f 69 74 65 6d 73 '/HelloService/SayHello'} SEND_MESSA
GE:flags=0x00000000:len=0 SEND_TRAILING_METADATA{} RECV_INITIAL_METADATA RECV_MESSAGE RECV_TRAILING_METADATA
I0424 13:55:39.484109800 308 client_channel.cc:972] chand=0x55e705fb6fe8 calld=0x55e705fe6d70: adding pending batch at index 0
I0424 13:55:39.484324000 308 client_channel.cc:2814] chand=0x55e705fb6fe8 calld=0x55e705fe6d70: entering client_channel combiner
I0424 13:55:39.484548300 308 client_channel.cc:2686] chand=0x55e705fb6fe8 calld=0x55e705fe6d70: LB pick returned QUEUE (connected_subchannel=(nil), error="No Error")
I0424 13:55:39.484737200 308 client_channel.cc:2553] chand=0x55e705fb6fe8 calld=0x55e705fe6d70: adding to queued picks list
I0424 13:55:39.484910200 308 call_combiner.cc:215] call_combiner=0x55e705fe63f0: setting notify_on_cancel=0x55e705de9518
I0424 13:55:39.485290900 308 resolving_lb_policy.cc:311] resolving_lb=0x55e705c443e0: starting name resolution
I0424 13:55:39.485491600 308 client_channel.cc:229] chand=0x55e705fb6fe8: update: state=CONNECTING error="No Error" picker=0x55e705fe5a30
I0424 13:55:39.486124200 308 connectivity_state.cc:164] SET: 0x55e705fb7090 client_channel: IDLE --> CONNECTING [helper] error=(nil) "No Error"
I0424 13:55:39.486387200 308 client_channel.cc:2686] chand=0x55e705fb6fe8 calld=0x55e705fe6d70: LB pick returned QUEUE (connected_subchannel=(nil), error="No Error")
D0424 13:55:39.486719500 308 dns_resolver_ares.cc:176] (c-ares resolver) resolver:0x55e705c918a0 AresDnsResolver::StartLocked() is called.
D0424 13:55:39.486896200 308 grpc_ares_wrapper.cc:536] (c-ares resolver) request:0x55e705b03950 c-ares grpc_dns_lookup_ares_locked_impl name=hello.domain.com, default_port=https
D0424 13:55:39.487483200 308 grpc_ares_ev_driver.cc:143] (c-ares resolver) request:0x55e705b03950 grpc_ares_ev_driver_create_locked
I0424 13:55:39.487778800 308 socket_utils_common_posix.cc:346] Disabling AF_INET6 sockets because ::1 is not available.
I0424 13:55:39.488282100 308 ev_posix.cc:252] (fd-trace) fd_create(6, c-ares fd: 6, 0)
Fatal Python error: Segmentation fault
Current thread 0x00007fc78e976400 (most recent call first):
File "/usr/local/lib/python3.7/site-packages/grpc/_channel.py", line 549 in _blocking
File "/usr/local/lib/python3.7/site-packages/grpc/_channel.py", line 561 in __call__
File "wsgi.py", line 54 in <module>
Segmentation fault
```
This code perfectly works on windows with enabled gevent and also work when gevent support is disabled on Linux.
|
1.0
|
Python: c-ares resolver with enabled gevent crashes on Linux - <!--
This form is for bug reports and feature requests ONLY!
For general questions and troubleshooting, please ask/look for answers here:
- grpc.io mailing list: https://groups.google.com/forum/#!forum/grpc-io
- StackOverflow, with "grpc" tag: https://stackoverflow.com/questions/tagged/grpc
Issues specific to *grpc-java*, *grpc-go*, *grpc-node*, *grpc-dart*, *grpc-web* should be created in the repository they belong to (e.g. https://github.com/grpc/grpc-LANGUAGE/issues/new)
-->
### What version of gRPC and what language are you using?
gRPC 1.20, Python
### What operating system (Linux, Windows,...) and version?
Linux Ubuntu (Python Docker: 3.7-stretch)
### What runtime / compiler are you using (e.g. python version or version of gcc)
Python 3.7
### What did you do?
I'm trying to make a call to the service using expiremental gevent functionality:
```
from gevent import monkey
monkey.patch_all()
import grpc.experimental.gevent
grpc.experimental.gevent.init_gevent()
import grpc
import helloworld_pb2
import helloworld_pb2_grpc
if __name__ == '__main__':
channel = grpc.insecure_channel('hello.domain.com')
stub = helloworld_pb2_grpc.GreeterStub(channel)
response = stub.SayHello(helloworld_pb2.HelloRequest(name='you'))
```
### What did you expect to see?
Response from service.
### What did you see instead?
Call to service crashes:
```
D0424 13:55:39.475437200 308 dns_resolver_ares.cc:436] Using ares dns resolver
I0424 13:55:39.475501500 308 timer_manager.cc:94] Spawn timer thread
I0424 13:55:39.475555300 308 init.cc:164] grpc_init(void)
I0424 13:55:39.475651200 313 timer_manager.cc:259] timers not checked: expect another thread to
I0424 13:55:39.476036000 313 timer_manager.cc:202] sleep until kicked
I0424 13:55:39.476754800 308 completion_queue.cc:503] grpc_completion_queue_create_internal(completion_type=0, polling_type=0)
I0424 13:55:39.476831400 308 completion_queue.cc:503] grpc_completion_queue_create_internal(completion_type=0, polling_type=0)
I0424 13:55:39.477093700 308 ssl_credentials.cc:128] grpc_ssl_credentials_create(pem_root_certs=(null), pem_key_cert_pair=(nil), verify_options=(nil), reserved=(nil))
I0424 13:55:39.477203900 308 secure_channel_create.cc:198] grpc_secure_channel_create(creds=0x55e705cf5aa0, target=hello.domain.com, args=0x7fc78452c2e0, reserved=(nil))
I0424 13:55:39.478266800 308 client_channel.cc:229] chand=0x55e705fb6fe8: update: state=IDLE error="No Error" picker=0x55e705fe5a50
I0424 13:55:39.478413600 308 connectivity_state.cc:164] SET: 0x55e705fb7090 client_channel: IDLE --> IDLE [helper] error=(nil) "No Error"
I0424 13:55:39.478852900 308 client_channel.cc:488] chand=0x55e705fb6fe8: created resolving_lb_policy=0x55e705c443e0
I0424 13:55:39.479489500 308 channel.cc:285] grpc_channel_get_target(channel=0x55e705fb6f20)
I0424 13:55:39.479648000 308 credentials.cc:43] grpc_channel_credentials_release(creds=0x55e705cf5aa0)
I0424 13:55:39.480383100 308 completion_queue.cc:503] grpc_completion_queue_create_internal(completion_type=0, polling_type=0)
I0424 13:55:39.480459800 308 metadata_array.cc:29] grpc_metadata_array_init(array=0x7fc784535168)
I0424 13:55:39.480989600 308 metadata_array.cc:29] grpc_metadata_array_init(array=0x7fc785170ce8)
I0424 13:55:39.481340900 308 call.cc:1910] grpc_call_start_batch(call=0x55e705fe63e0, ops=0x55e705ea3170, nops=6, tag=0x7fc78452cfc0, reserved=(nil))
I0424 13:55:39.481670300 308 call.cc:1516] ops[0]: SEND_INITIAL_METADATA(nil)
I0424 13:55:39.481909800 308 call.cc:1516] ops[1]: SEND_MESSAGE ptr=0x55e705aea420
I0424 13:55:39.482198900 308 call.cc:1516] ops[2]: SEND_CLOSE_FROM_CLIENT
I0424 13:55:39.482334300 308 call.cc:1516] ops[3]: RECV_INITIAL_METADATA ptr=0x7fc784535168
I0424 13:55:39.482369500 308 call.cc:1516] ops[4]: RECV_MESSAGE ptr=0x7fc7845796a0
I0424 13:55:39.482710500 308 call.cc:1516] ops[5]: RECV_STATUS_ON_CLIENT metadata=0x7fc785170ce8 status=0x7fc785170d00 details=0x7fc785170d08
I0424 13:55:39.483026300 308 call_combiner.cc:116] ==> grpc_call_combiner_start() [0x55e705fe63f0] closure=0x55e705fe72c8 [executing batch] error="No Error"
I0424 13:55:39.483197400 308 call_combiner.cc:125] size: 0 -> 1
I0424 13:55:39.483383900 308 call_combiner.cc:134] EXECUTING IMMEDIATELY
I0424 13:55:39.483639100 308 call.cc:615] OP[client-channel:0x55e705fe6d50]: SEND_INITIAL_METADATA{key=3a 70 61 74 68 ':path' value=2f 49 74 65 6d 73 53 65 72 76 69 63 65 2f 67 65 74 5f 61 6c 6c 5f 69 74 65 6d 73 '/HelloService/SayHello'} SEND_MESSA
GE:flags=0x00000000:len=0 SEND_TRAILING_METADATA{} RECV_INITIAL_METADATA RECV_MESSAGE RECV_TRAILING_METADATA
I0424 13:55:39.484109800 308 client_channel.cc:972] chand=0x55e705fb6fe8 calld=0x55e705fe6d70: adding pending batch at index 0
I0424 13:55:39.484324000 308 client_channel.cc:2814] chand=0x55e705fb6fe8 calld=0x55e705fe6d70: entering client_channel combiner
I0424 13:55:39.484548300 308 client_channel.cc:2686] chand=0x55e705fb6fe8 calld=0x55e705fe6d70: LB pick returned QUEUE (connected_subchannel=(nil), error="No Error")
I0424 13:55:39.484737200 308 client_channel.cc:2553] chand=0x55e705fb6fe8 calld=0x55e705fe6d70: adding to queued picks list
I0424 13:55:39.484910200 308 call_combiner.cc:215] call_combiner=0x55e705fe63f0: setting notify_on_cancel=0x55e705de9518
I0424 13:55:39.485290900 308 resolving_lb_policy.cc:311] resolving_lb=0x55e705c443e0: starting name resolution
I0424 13:55:39.485491600 308 client_channel.cc:229] chand=0x55e705fb6fe8: update: state=CONNECTING error="No Error" picker=0x55e705fe5a30
I0424 13:55:39.486124200 308 connectivity_state.cc:164] SET: 0x55e705fb7090 client_channel: IDLE --> CONNECTING [helper] error=(nil) "No Error"
I0424 13:55:39.486387200 308 client_channel.cc:2686] chand=0x55e705fb6fe8 calld=0x55e705fe6d70: LB pick returned QUEUE (connected_subchannel=(nil), error="No Error")
D0424 13:55:39.486719500 308 dns_resolver_ares.cc:176] (c-ares resolver) resolver:0x55e705c918a0 AresDnsResolver::StartLocked() is called.
D0424 13:55:39.486896200 308 grpc_ares_wrapper.cc:536] (c-ares resolver) request:0x55e705b03950 c-ares grpc_dns_lookup_ares_locked_impl name=hello.domain.com, default_port=https
D0424 13:55:39.487483200 308 grpc_ares_ev_driver.cc:143] (c-ares resolver) request:0x55e705b03950 grpc_ares_ev_driver_create_locked
I0424 13:55:39.487778800 308 socket_utils_common_posix.cc:346] Disabling AF_INET6 sockets because ::1 is not available.
I0424 13:55:39.488282100 308 ev_posix.cc:252] (fd-trace) fd_create(6, c-ares fd: 6, 0)
Fatal Python error: Segmentation fault
Current thread 0x00007fc78e976400 (most recent call first):
File "/usr/local/lib/python3.7/site-packages/grpc/_channel.py", line 549 in _blocking
File "/usr/local/lib/python3.7/site-packages/grpc/_channel.py", line 561 in __call__
File "wsgi.py", line 54 in <module>
Segmentation fault
```
This code perfectly works on windows with enabled gevent and also work when gevent support is disabled on Linux.
|
priority
|
python c ares resolver with enabled gevent crashes on linux this form is for bug reports and feature requests only for general questions and troubleshooting please ask look for answers here grpc io mailing list stackoverflow with grpc tag issues specific to grpc java grpc go grpc node grpc dart grpc web should be created in the repository they belong to e g what version of grpc and what language are you using grpc python what operating system linux windows and version linux ubuntu python docker stretch what runtime compiler are you using e g python version or version of gcc python what did you do i m trying to make a call to the service using expiremental gevent functionality from gevent import monkey monkey patch all import grpc experimental gevent grpc experimental gevent init gevent import grpc import helloworld import helloworld grpc if name main channel grpc insecure channel hello domain com stub helloworld grpc greeterstub channel response stub sayhello helloworld hellorequest name you what did you expect to see response from service what did you see instead call to service crashes dns resolver ares cc using ares dns resolver timer manager cc spawn timer thread init cc grpc init void timer manager cc timers not checked expect another thread to timer manager cc sleep until kicked completion queue cc grpc completion queue create internal completion type polling type completion queue cc grpc completion queue create internal completion type polling type ssl credentials cc grpc ssl credentials create pem root certs null pem key cert pair nil verify options nil reserved nil secure channel create cc grpc secure channel create creds target hello domain com args reserved nil client channel cc chand update state idle error no error picker connectivity state cc set client channel idle idle error nil no error client channel cc chand created resolving lb policy channel cc grpc channel get target channel credentials cc grpc channel credentials release creds completion queue cc grpc completion queue create internal completion type polling type metadata array cc grpc metadata array init array metadata array cc grpc metadata array init array call cc grpc call start batch call ops nops tag reserved nil call cc ops send initial metadata nil call cc ops send message ptr call cc ops send close from client call cc ops recv initial metadata ptr call cc ops recv message ptr call cc ops recv status on client metadata status details call combiner cc grpc call combiner start closure error no error call combiner cc size call combiner cc executing immediately call cc op send initial metadata key path value helloservice sayhello send messa ge flags len send trailing metadata recv initial metadata recv message recv trailing metadata client channel cc chand calld adding pending batch at index client channel cc chand calld entering client channel combiner client channel cc chand calld lb pick returned queue connected subchannel nil error no error client channel cc chand calld adding to queued picks list call combiner cc call combiner setting notify on cancel resolving lb policy cc resolving lb starting name resolution client channel cc chand update state connecting error no error picker connectivity state cc set client channel idle connecting error nil no error client channel cc chand calld lb pick returned queue connected subchannel nil error no error dns resolver ares cc c ares resolver resolver aresdnsresolver startlocked is called grpc ares wrapper cc c ares resolver request c ares grpc dns lookup ares locked impl name hello domain com default port https grpc ares ev driver cc c ares resolver request grpc ares ev driver create locked socket utils common posix cc disabling af sockets because is not available ev posix cc fd trace fd create c ares fd fatal python error segmentation fault current thread most recent call first file usr local lib site packages grpc channel py line in blocking file usr local lib site packages grpc channel py line in call file wsgi py line in segmentation fault this code perfectly works on windows with enabled gevent and also work when gevent support is disabled on linux
| 1
|
214,321
| 24,068,078,079
|
IssuesEvent
|
2022-09-17 19:41:02
|
ofcoursedude/confd
|
https://api.github.com/repos/ofcoursedude/confd
|
opened
|
github.com/golang/net/http2-640f4622ab692b87c2f3a94265e6f579fe38263d: 1 vulnerabilities (highest severity is: 7.5)
|
security vulnerability
|
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>github.com/golang/net/http2-640f4622ab692b87c2f3a94265e6f579fe38263d</b></p></summary>
<p>[mirror] Go supplementary network libraries</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/ofcoursedude/confd/commit/20911c124d56b581ce97f9cf63499b5489cac35a">20911c124d56b581ce97f9cf63499b5489cac35a</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2021-44716](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-44716) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | github.com/golang/net/http2-640f4622ab692b87c2f3a94265e6f579fe38263d | Direct | github.com/golang/net - 491a49abca63de5e07ef554052d180a1b5fe2d70 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-44716</summary>
### Vulnerable Library - <b>github.com/golang/net/http2-640f4622ab692b87c2f3a94265e6f579fe38263d</b></p>
<p>[mirror] Go supplementary network libraries</p>
<p>
Dependency Hierarchy:
- :x: **github.com/golang/net/http2-640f4622ab692b87c2f3a94265e6f579fe38263d** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ofcoursedude/confd/commit/20911c124d56b581ce97f9cf63499b5489cac35a">20911c124d56b581ce97f9cf63499b5489cac35a</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
net/http in Go before 1.16.12 and 1.17.x before 1.17.5 allows uncontrolled memory consumption in the header canonicalization cache via HTTP/2 requests.
<p>Publish Date: 2022-01-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-44716>CVE-2021-44716</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-vc3p-29h2-gpcp">https://github.com/advisories/GHSA-vc3p-29h2-gpcp</a></p>
<p>Release Date: 2022-01-01</p>
<p>Fix Resolution: github.com/golang/net - 491a49abca63de5e07ef554052d180a1b5fe2d70</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
|
True
|
github.com/golang/net/http2-640f4622ab692b87c2f3a94265e6f579fe38263d: 1 vulnerabilities (highest severity is: 7.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>github.com/golang/net/http2-640f4622ab692b87c2f3a94265e6f579fe38263d</b></p></summary>
<p>[mirror] Go supplementary network libraries</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/ofcoursedude/confd/commit/20911c124d56b581ce97f9cf63499b5489cac35a">20911c124d56b581ce97f9cf63499b5489cac35a</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2021-44716](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-44716) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | github.com/golang/net/http2-640f4622ab692b87c2f3a94265e6f579fe38263d | Direct | github.com/golang/net - 491a49abca63de5e07ef554052d180a1b5fe2d70 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-44716</summary>
### Vulnerable Library - <b>github.com/golang/net/http2-640f4622ab692b87c2f3a94265e6f579fe38263d</b></p>
<p>[mirror] Go supplementary network libraries</p>
<p>
Dependency Hierarchy:
- :x: **github.com/golang/net/http2-640f4622ab692b87c2f3a94265e6f579fe38263d** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ofcoursedude/confd/commit/20911c124d56b581ce97f9cf63499b5489cac35a">20911c124d56b581ce97f9cf63499b5489cac35a</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
net/http in Go before 1.16.12 and 1.17.x before 1.17.5 allows uncontrolled memory consumption in the header canonicalization cache via HTTP/2 requests.
<p>Publish Date: 2022-01-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-44716>CVE-2021-44716</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-vc3p-29h2-gpcp">https://github.com/advisories/GHSA-vc3p-29h2-gpcp</a></p>
<p>Release Date: 2022-01-01</p>
<p>Fix Resolution: github.com/golang/net - 491a49abca63de5e07ef554052d180a1b5fe2d70</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
|
non_priority
|
github com golang net vulnerabilities highest severity is vulnerable library github com golang net go supplementary network libraries found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available high github com golang net direct github com golang net details cve vulnerable library github com golang net go supplementary network libraries dependency hierarchy x github com golang net vulnerable library found in head commit a href found in base branch master vulnerability details net http in go before and x before allows uncontrolled memory consumption in the header canonicalization cache via http requests publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution github com golang net step up your open source security game with mend
| 0
|
228,080
| 18,155,351,740
|
IssuesEvent
|
2021-09-27 00:09:05
|
PlaceOS/backoffice
|
https://api.github.com/repos/PlaceOS/backoffice
|
closed
|
Display Settings history
|
priority: low status: requires testing
|
Add new subsections in backoffice that allow the user to see Settings history for the currently selected driver/module/system/zone:
`/api/engine/v2/settings/{settings-id}/history`
|
1.0
|
Display Settings history - Add new subsections in backoffice that allow the user to see Settings history for the currently selected driver/module/system/zone:
`/api/engine/v2/settings/{settings-id}/history`
|
non_priority
|
display settings history add new subsections in backoffice that allow the user to see settings history for the currently selected driver module system zone api engine settings settings id history
| 0
|
87,637
| 15,786,555,767
|
IssuesEvent
|
2021-04-01 17:55:18
|
joshnewton31080/experian-java
|
https://api.github.com/repos/joshnewton31080/experian-java
|
opened
|
CVE-2018-7489 (High) detected in jackson-databind-2.9.2.jar
|
security vulnerability
|
## CVE-2018-7489 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.2.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: experian-java/MavenWorkspace/bis-services-lib/bis-services-base/pom.xml</p>
<p>Path to vulnerable library: canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.2/jackson-databind-2.9.2.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.2.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/joshnewton31080/experian-java/commit/f7845d2a79eba7a55a8254693f8dc9db8198487a">f7845d2a79eba7a55a8254693f8dc9db8198487a</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind before 2.7.9.3, 2.8.x before 2.8.11.1 and 2.9.x before 2.9.5 allows unauthenticated remote code execution because of an incomplete fix for the CVE-2017-7525 deserialization flaw. This is exploitable by sending maliciously crafted JSON input to the readValue method of the ObjectMapper, bypassing a blacklist that is ineffective if the c3p0 libraries are available in the classpath.
<p>Publish Date: 2018-02-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-7489>CVE-2018-7489</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-7489">https://nvd.nist.gov/vuln/detail/CVE-2018-7489</a></p>
<p>Release Date: 2018-02-26</p>
<p>Fix Resolution: 2.8.11.1,2.9.5</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.2","packageFilePaths":["/MavenWorkspace/bis-services-lib/bis-services-base/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.8.11.1,2.9.5"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-7489","vulnerabilityDetails":"FasterXML jackson-databind before 2.7.9.3, 2.8.x before 2.8.11.1 and 2.9.x before 2.9.5 allows unauthenticated remote code execution because of an incomplete fix for the CVE-2017-7525 deserialization flaw. This is exploitable by sending maliciously crafted JSON input to the readValue method of the ObjectMapper, bypassing a blacklist that is ineffective if the c3p0 libraries are available in the classpath.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-7489","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2018-7489 (High) detected in jackson-databind-2.9.2.jar - ## CVE-2018-7489 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.2.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: experian-java/MavenWorkspace/bis-services-lib/bis-services-base/pom.xml</p>
<p>Path to vulnerable library: canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.2/jackson-databind-2.9.2.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.2.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/joshnewton31080/experian-java/commit/f7845d2a79eba7a55a8254693f8dc9db8198487a">f7845d2a79eba7a55a8254693f8dc9db8198487a</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind before 2.7.9.3, 2.8.x before 2.8.11.1 and 2.9.x before 2.9.5 allows unauthenticated remote code execution because of an incomplete fix for the CVE-2017-7525 deserialization flaw. This is exploitable by sending maliciously crafted JSON input to the readValue method of the ObjectMapper, bypassing a blacklist that is ineffective if the c3p0 libraries are available in the classpath.
<p>Publish Date: 2018-02-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-7489>CVE-2018-7489</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-7489">https://nvd.nist.gov/vuln/detail/CVE-2018-7489</a></p>
<p>Release Date: 2018-02-26</p>
<p>Fix Resolution: 2.8.11.1,2.9.5</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.2","packageFilePaths":["/MavenWorkspace/bis-services-lib/bis-services-base/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.8.11.1,2.9.5"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-7489","vulnerabilityDetails":"FasterXML jackson-databind before 2.7.9.3, 2.8.x before 2.8.11.1 and 2.9.x before 2.9.5 allows unauthenticated remote code execution because of an incomplete fix for the CVE-2017-7525 deserialization flaw. This is exploitable by sending maliciously crafted JSON input to the readValue method of the ObjectMapper, bypassing a blacklist that is ineffective if the c3p0 libraries are available in the classpath.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-7489","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
non_priority
|
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file experian java mavenworkspace bis services lib bis services base pom xml path to vulnerable library canner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind before x before and x before allows unauthenticated remote code execution because of an incomplete fix for the cve deserialization flaw this is exploitable by sending maliciously crafted json input to the readvalue method of the objectmapper bypassing a blacklist that is ineffective if the libraries are available in the classpath publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution check this box to open an automated fix pr isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind before x before and x before allows unauthenticated remote code execution because of an incomplete fix for the cve deserialization flaw this is exploitable by sending maliciously crafted json input to the readvalue method of the objectmapper bypassing a blacklist that is ineffective if the libraries are available in the classpath vulnerabilityurl
| 0
|
86,107
| 16,829,079,528
|
IssuesEvent
|
2021-06-17 23:51:43
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
closed
|
[Wasm][AOT] System.Threading.CancellationTokenSource.ExecuteCallbackHandlers fails to compile
|
arch-wasm area-Codegen-AOT-mono
|
<!--This is just a template - feel free to delete any and all of it and replace as appropriate.-->
### Description
using `System.Threading.CancellationTokenSource:ExecuteCallbackHandlers` indirectly fails with:
```
AOT Compilation failed for method 'void System.Threading.CancellationTokenSource:ExecuteCallbackHandlers (bool)'.
```
### Workaround
Exclude `System.Threading.CancellationTokenSource:ExecuteCallbackHandlers` from the AOT profile.
### Configuration
6b4c1ad8b2297bfa0da566f7097ec5ed38fc79a5
### Regression?
No
|
1.0
|
[Wasm][AOT] System.Threading.CancellationTokenSource.ExecuteCallbackHandlers fails to compile - <!--This is just a template - feel free to delete any and all of it and replace as appropriate.-->
### Description
using `System.Threading.CancellationTokenSource:ExecuteCallbackHandlers` indirectly fails with:
```
AOT Compilation failed for method 'void System.Threading.CancellationTokenSource:ExecuteCallbackHandlers (bool)'.
```
### Workaround
Exclude `System.Threading.CancellationTokenSource:ExecuteCallbackHandlers` from the AOT profile.
### Configuration
6b4c1ad8b2297bfa0da566f7097ec5ed38fc79a5
### Regression?
No
|
non_priority
|
system threading cancellationtokensource executecallbackhandlers fails to compile description using system threading cancellationtokensource executecallbackhandlers indirectly fails with aot compilation failed for method void system threading cancellationtokensource executecallbackhandlers bool workaround exclude system threading cancellationtokensource executecallbackhandlers from the aot profile configuration regression no
| 0
|
814,552
| 30,512,060,873
|
IssuesEvent
|
2023-07-18 21:48:04
|
paleobot/pbot-dev
|
https://api.github.com/repos/paleobot/pbot-dev
|
closed
|
Public vs. private groups
|
question high priority Hack-A-Thon
|
Right now Public is listed as an option inside the Group Selection drop down box when adding/editing a group. Should we just remove Public from drop down list. Is is true that once a group goes public the group is no longer valid? Is there a reason researchers would want to keep a group/project live on their workbench once that information has gone public?
Doug to investigate and provide current status to Botany group.
|
1.0
|
Public vs. private groups - Right now Public is listed as an option inside the Group Selection drop down box when adding/editing a group. Should we just remove Public from drop down list. Is is true that once a group goes public the group is no longer valid? Is there a reason researchers would want to keep a group/project live on their workbench once that information has gone public?
Doug to investigate and provide current status to Botany group.
|
priority
|
public vs private groups right now public is listed as an option inside the group selection drop down box when adding editing a group should we just remove public from drop down list is is true that once a group goes public the group is no longer valid is there a reason researchers would want to keep a group project live on their workbench once that information has gone public doug to investigate and provide current status to botany group
| 1
|
699,285
| 24,010,828,727
|
IssuesEvent
|
2022-09-14 18:39:09
|
LimbersMay/WebNotes
|
https://api.github.com/repos/LimbersMay/WebNotes
|
closed
|
Duplicidad de cuenta al iniciar sesión con Google
|
💡Bug :gear:Controller ⏰High priority
|
Al iniciar sesión con google, si el correo ya se encuentra registrado, el servidor vuelve a crear nuevamente una cuenta con el mismo correo en lugar de iniciar sesión con la cuenta existente
|
1.0
|
Duplicidad de cuenta al iniciar sesión con Google - Al iniciar sesión con google, si el correo ya se encuentra registrado, el servidor vuelve a crear nuevamente una cuenta con el mismo correo en lugar de iniciar sesión con la cuenta existente
|
priority
|
duplicidad de cuenta al iniciar sesión con google al iniciar sesión con google si el correo ya se encuentra registrado el servidor vuelve a crear nuevamente una cuenta con el mismo correo en lugar de iniciar sesión con la cuenta existente
| 1
|
225,647
| 24,881,042,881
|
IssuesEvent
|
2022-10-28 01:07:38
|
LingalaShalini/libexpact_Fix
|
https://api.github.com/repos/LingalaShalini/libexpact_Fix
|
opened
|
CVE-2022-43680 (High) detected in libexpatR_2_2_6
|
security vulnerability
|
## CVE-2022-43680 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>libexpatR_2_2_6</b></p></summary>
<p>
<p>:herb: Expat library: Fast streaming XML parser written in C99; in the process of migrating from SourceForge to GitHub</p>
<p>Library home page: <a href=https://github.com/libexpat/libexpat.git>https://github.com/libexpat/libexpat.git</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/expat/lib/xmlparse.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In libexpat through 2.4.9, there is a use-after free caused by overeager destruction of a shared DTD in XML_ExternalEntityParserCreate in out-of-memory situations.
<p>Publish Date: 2022-10-24
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-43680>CVE-2022-43680</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-43680 (High) detected in libexpatR_2_2_6 - ## CVE-2022-43680 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>libexpatR_2_2_6</b></p></summary>
<p>
<p>:herb: Expat library: Fast streaming XML parser written in C99; in the process of migrating from SourceForge to GitHub</p>
<p>Library home page: <a href=https://github.com/libexpat/libexpat.git>https://github.com/libexpat/libexpat.git</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/expat/lib/xmlparse.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In libexpat through 2.4.9, there is a use-after free caused by overeager destruction of a shared DTD in XML_ExternalEntityParserCreate in out-of-memory situations.
<p>Publish Date: 2022-10-24
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-43680>CVE-2022-43680</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_priority
|
cve high detected in libexpatr cve high severity vulnerability vulnerable library libexpatr herb expat library fast streaming xml parser written in in the process of migrating from sourceforge to github library home page a href found in base branch master vulnerable source files expat lib xmlparse c vulnerability details in libexpat through there is a use after free caused by overeager destruction of a shared dtd in xml externalentityparsercreate in out of memory situations publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with mend
| 0
|
610,073
| 18,893,332,223
|
IssuesEvent
|
2021-11-15 15:22:35
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
store.ubi.com - site is not usable
|
status-needsinfo priority-normal type-webrender-enabled browser-fenix engine-gecko
|
<!-- @browser: Firefox Mobile 96.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 11; Mobile; rv:96.0) Gecko/96.0 Firefox/96.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @extra_labels: browser-fenix, type-webrender-enabled -->
**URL**: https://store.ubi.com/it/game?pid=60a4d7030d253c0e6c630899&dwvar_60a4d7030d253c0e6c630899_Platform=switchdig&edition=Standard&source=detail
**Browser / Version**: Firefox Mobile 96.0
**Operating System**: Android 11
**Tested Another Browser**: No
**Problem type**: Site is not usable
**Description**: Browser unsupported
**Steps to Reproduce**:
The page only scrolls when in landscape, in portrait bounces to top very fast. Happens in game pages.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: true</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20211102094141</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2021/11/1d7e5dc4-9b43-4783-9aac-221f7b6853fd)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
store.ubi.com - site is not usable - <!-- @browser: Firefox Mobile 96.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 11; Mobile; rv:96.0) Gecko/96.0 Firefox/96.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @extra_labels: browser-fenix, type-webrender-enabled -->
**URL**: https://store.ubi.com/it/game?pid=60a4d7030d253c0e6c630899&dwvar_60a4d7030d253c0e6c630899_Platform=switchdig&edition=Standard&source=detail
**Browser / Version**: Firefox Mobile 96.0
**Operating System**: Android 11
**Tested Another Browser**: No
**Problem type**: Site is not usable
**Description**: Browser unsupported
**Steps to Reproduce**:
The page only scrolls when in landscape, in portrait bounces to top very fast. Happens in game pages.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: true</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20211102094141</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2021/11/1d7e5dc4-9b43-4783-9aac-221f7b6853fd)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
priority
|
store ubi com site is not usable url browser version firefox mobile operating system android tested another browser no problem type site is not usable description browser unsupported steps to reproduce the page only scrolls when in landscape in portrait bounces to top very fast happens in game pages browser configuration gfx webrender all true gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
| 1
|
352,012
| 32,040,809,867
|
IssuesEvent
|
2023-09-22 19:09:15
|
astropy/specutils
|
https://api.github.com/repos/astropy/specutils
|
opened
|
TST: Ensure numpy dev is actually installed in devdeps
|
testing bug
|
```
Numpy: 1.26.0
Scipy: 1.12.0.dev0+1726.b786038
Matplotlib: 3.9.0.dev238+ge861ea2ff6
```
This is because:
* https://github.com/matplotlib/matplotlib/issues/26847
Two options:
* Separate out numpy-dev and mpl-dev jobs, or
* Force numpy to upgrade again later like https://github.com/spacetelescope/jdaviz/pull/2472
|
1.0
|
TST: Ensure numpy dev is actually installed in devdeps - ```
Numpy: 1.26.0
Scipy: 1.12.0.dev0+1726.b786038
Matplotlib: 3.9.0.dev238+ge861ea2ff6
```
This is because:
* https://github.com/matplotlib/matplotlib/issues/26847
Two options:
* Separate out numpy-dev and mpl-dev jobs, or
* Force numpy to upgrade again later like https://github.com/spacetelescope/jdaviz/pull/2472
|
non_priority
|
tst ensure numpy dev is actually installed in devdeps numpy scipy matplotlib this is because two options separate out numpy dev and mpl dev jobs or force numpy to upgrade again later like
| 0
|
50,021
| 10,441,112,726
|
IssuesEvent
|
2019-09-18 10:05:42
|
tendermint/tendermint
|
https://api.github.com/repos/tendermint/tendermint
|
closed
|
libs/common: Remove TrapSignal()
|
code-hygiene enhancement libs
|
It is quite conventional in the UNIX world to return `128 + signal code` upon exit due to a signal, so I'd suggest `TrapSignal()` call os.Exit(128 + int($signalcode)) instead of `os.Exit(1)`
**Tendermint version**: develop
**ABCI app**: Gaia
**Environment**: `Ubuntu 18.04`
CC'ing @jaekwon, who appears to be the original author of the function.
|
1.0
|
libs/common: Remove TrapSignal() - It is quite conventional in the UNIX world to return `128 + signal code` upon exit due to a signal, so I'd suggest `TrapSignal()` call os.Exit(128 + int($signalcode)) instead of `os.Exit(1)`
**Tendermint version**: develop
**ABCI app**: Gaia
**Environment**: `Ubuntu 18.04`
CC'ing @jaekwon, who appears to be the original author of the function.
|
non_priority
|
libs common remove trapsignal it is quite conventional in the unix world to return signal code upon exit due to a signal so i d suggest trapsignal call os exit int signalcode instead of os exit tendermint version develop abci app gaia environment ubuntu cc ing jaekwon who appears to be the original author of the function
| 0
|
164,646
| 26,000,830,493
|
IssuesEvent
|
2022-12-20 15:11:54
|
bro-n-bro/bro_rating
|
https://api.github.com/repos/bro-n-bro/bro_rating
|
closed
|
Add validator page
|
design feature
|
To avoid unnecessarily leaving the app it's cool to have our own validator page instead of the link to the mintscan
should contain:
all info from the API
link to the mintscan
|
1.0
|
Add validator page - To avoid unnecessarily leaving the app it's cool to have our own validator page instead of the link to the mintscan
should contain:
all info from the API
link to the mintscan
|
non_priority
|
add validator page to avoid unnecessarily leaving the app it s cool to have our own validator page instead of the link to the mintscan should contain all info from the api link to the mintscan
| 0
|
5,939
| 3,700,017,401
|
IssuesEvent
|
2016-02-29 05:18:17
|
jeff1evesque/drupal-demonstration
|
https://api.github.com/repos/jeff1evesque/drupal-demonstration
|
opened
|
Webcompilers not working 'failed'
|
bug build
|
Our webcompilers no longer work, since we replaced our vagrant base box, with a custom created box from a minimal Centos iso:
```bash
[provisioner@drupal-demonstration system]$ sudo systemctl status sass -l
● sass.service - define service to run corresponding bash script to compile source files
Loaded: loaded (/etc/systemd/system/sass.service; static; vendor preset: disabled)
Active: failed (Result: exit-code) since Mon 2016-02-29 00:06:20 EST; 8min ago
Docs: https://github.com/jeff1evesque/drupal-demonstration/issues/248
Process: 31664 ExecStart=/vagrant/puppet/environment/development/scripts/.sass
(code=exited, status=203/EXEC)
Main PID: 31664 (code=exited, status=203/EXEC)
Feb 29 00:06:20 drupal-demonstration.com systemd[1]: Started define service to run corresponding bash script to compile source files.
Feb 29 00:06:20 drupal-demonstration.com systemd[1]: Starting define service to run corresponding bash script to compile source files...
Feb 29 00:06:20 drupal-demonstration.com systemd[1]: sass.service: main process exited, code=exited, status=203/EXEC
Feb 29 00:06:20 drupal-demonstration.com systemd[1]: Unit sass.service entered failed state.
Feb 29 00:06:20 drupal-demonstration.com systemd[1]: sass.service failed.
```
The above traceback persists, regardless if we `sudo systemctl stop sass`, or `sudo systemctl restart sass`, followed by `sudo systemctl start sass`.
**Note:** this generally applied to our other webcompilers (i.e. imagemin, uglifyjs).
|
1.0
|
Webcompilers not working 'failed' - Our webcompilers no longer work, since we replaced our vagrant base box, with a custom created box from a minimal Centos iso:
```bash
[provisioner@drupal-demonstration system]$ sudo systemctl status sass -l
● sass.service - define service to run corresponding bash script to compile source files
Loaded: loaded (/etc/systemd/system/sass.service; static; vendor preset: disabled)
Active: failed (Result: exit-code) since Mon 2016-02-29 00:06:20 EST; 8min ago
Docs: https://github.com/jeff1evesque/drupal-demonstration/issues/248
Process: 31664 ExecStart=/vagrant/puppet/environment/development/scripts/.sass
(code=exited, status=203/EXEC)
Main PID: 31664 (code=exited, status=203/EXEC)
Feb 29 00:06:20 drupal-demonstration.com systemd[1]: Started define service to run corresponding bash script to compile source files.
Feb 29 00:06:20 drupal-demonstration.com systemd[1]: Starting define service to run corresponding bash script to compile source files...
Feb 29 00:06:20 drupal-demonstration.com systemd[1]: sass.service: main process exited, code=exited, status=203/EXEC
Feb 29 00:06:20 drupal-demonstration.com systemd[1]: Unit sass.service entered failed state.
Feb 29 00:06:20 drupal-demonstration.com systemd[1]: sass.service failed.
```
The above traceback persists, regardless if we `sudo systemctl stop sass`, or `sudo systemctl restart sass`, followed by `sudo systemctl start sass`.
**Note:** this generally applied to our other webcompilers (i.e. imagemin, uglifyjs).
|
non_priority
|
webcompilers not working failed our webcompilers no longer work since we replaced our vagrant base box with a custom created box from a minimal centos iso bash sudo systemctl status sass l ● sass service define service to run corresponding bash script to compile source files loaded loaded etc systemd system sass service static vendor preset disabled active failed result exit code since mon est ago docs process execstart vagrant puppet environment development scripts sass code exited status exec main pid code exited status exec feb drupal demonstration com systemd started define service to run corresponding bash script to compile source files feb drupal demonstration com systemd starting define service to run corresponding bash script to compile source files feb drupal demonstration com systemd sass service main process exited code exited status exec feb drupal demonstration com systemd unit sass service entered failed state feb drupal demonstration com systemd sass service failed the above traceback persists regardless if we sudo systemctl stop sass or sudo systemctl restart sass followed by sudo systemctl start sass note this generally applied to our other webcompilers i e imagemin uglifyjs
| 0
|
83,099
| 10,322,633,478
|
IssuesEvent
|
2019-08-31 14:10:24
|
IBM/kui
|
https://api.github.com/repos/IBM/kui
|
closed
|
activation list does not distinguish activations between packages
|
design :art: plugin/openwhisk
|
these two activations of `a` are not the same - one is from a package `p` and another is not in a package... they can't be distinguished until you click on each one.
<img width="299" alt="Screen Shot 2019-08-22 at 6 34 11 PM" src="https://user-images.githubusercontent.com/4959922/63554476-8d73cf00-c50b-11e9-92b3-fe4f7239194c.png">
|
1.0
|
activation list does not distinguish activations between packages - these two activations of `a` are not the same - one is from a package `p` and another is not in a package... they can't be distinguished until you click on each one.
<img width="299" alt="Screen Shot 2019-08-22 at 6 34 11 PM" src="https://user-images.githubusercontent.com/4959922/63554476-8d73cf00-c50b-11e9-92b3-fe4f7239194c.png">
|
non_priority
|
activation list does not distinguish activations between packages these two activations of a are not the same one is from a package p and another is not in a package they can t be distinguished until you click on each one img width alt screen shot at pm src
| 0
|
310,126
| 26,702,662,391
|
IssuesEvent
|
2023-01-27 15:33:32
|
wazuh/wazuh-qa
|
https://api.github.com/repos/wazuh/wazuh-qa
|
closed
|
Research `test_assign_agent_group_with_enrollment`
|
team/qa test/system type/research subteam/qa-main
|
As part of https://github.com/wazuh/wazuh-qa/issues/3776, in this issue, we will investigate the `test_assign_agent_group_with_enrollment` test in particular. We must verify that the test is stable and passes correctly.
|
1.0
|
Research `test_assign_agent_group_with_enrollment` - As part of https://github.com/wazuh/wazuh-qa/issues/3776, in this issue, we will investigate the `test_assign_agent_group_with_enrollment` test in particular. We must verify that the test is stable and passes correctly.
|
non_priority
|
research test assign agent group with enrollment as part of in this issue we will investigate the test assign agent group with enrollment test in particular we must verify that the test is stable and passes correctly
| 0
|
192,055
| 6,846,181,220
|
IssuesEvent
|
2017-11-13 11:02:43
|
Mandiklopper/UBA-HR-Queries-Issues
|
https://api.github.com/repos/Mandiklopper/UBA-HR-Queries-Issues
|
closed
|
Error on Relief Assignment Processed
|
High Priority
|
A relief assignment processed for A22118 with effective date of 8/11 and end date of 17/11 had an error on processing with error message as 'Error creating new contract Selected notch is invalid for the selected grade'
A review of the Employment contract screen reveal the following;
1. The take-on data contract for this employee has no Grade (this is a mandatory field), but has a base notch
2. The new contract that the system tried to create with the relief assignment processed had no grade but had a base notch, nd also no end date
3. The contract for the resumption from relief assignment was not created.
|
1.0
|
Error on Relief Assignment Processed -
A relief assignment processed for A22118 with effective date of 8/11 and end date of 17/11 had an error on processing with error message as 'Error creating new contract Selected notch is invalid for the selected grade'
A review of the Employment contract screen reveal the following;
1. The take-on data contract for this employee has no Grade (this is a mandatory field), but has a base notch
2. The new contract that the system tried to create with the relief assignment processed had no grade but had a base notch, nd also no end date
3. The contract for the resumption from relief assignment was not created.
|
priority
|
error on relief assignment processed a relief assignment processed for with effective date of and end date of had an error on processing with error message as error creating new contract selected notch is invalid for the selected grade a review of the employment contract screen reveal the following the take on data contract for this employee has no grade this is a mandatory field but has a base notch the new contract that the system tried to create with the relief assignment processed had no grade but had a base notch nd also no end date the contract for the resumption from relief assignment was not created
| 1
|
666,940
| 22,392,925,852
|
IssuesEvent
|
2022-06-17 09:29:09
|
darktable-org/darktable
|
https://api.github.com/repos/darktable-org/darktable
|
closed
|
darktable crashes when deleting styles
|
priority: high bug: pending
|
When I import styles and want to delete them again before restarting darktable, darktable crashes. When I import a style I first have to restart darktable again before I can delete that style. I'm using the latest unstable release.
|
1.0
|
darktable crashes when deleting styles - When I import styles and want to delete them again before restarting darktable, darktable crashes. When I import a style I first have to restart darktable again before I can delete that style. I'm using the latest unstable release.
|
priority
|
darktable crashes when deleting styles when i import styles and want to delete them again before restarting darktable darktable crashes when i import a style i first have to restart darktable again before i can delete that style i m using the latest unstable release
| 1
|
53,771
| 13,883,340,613
|
IssuesEvent
|
2020-10-18 11:31:59
|
theWhiteFox/theWhiteFox
|
https://api.github.com/repos/theWhiteFox/theWhiteFox
|
opened
|
CVE-2015-9251 (Medium) detected in multiple libraries
|
security vulnerability
|
## CVE-2015-9251 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-2.1.4.min.js</b>, <b>jquery-1.11.0.min.js</b>, <b>jquery-1.7.1.min.js</b></p></summary>
<p>
<details><summary><b>jquery-2.1.4.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.4/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.4/jquery.min.js</a></p>
<p>Path to dependency file: theWhiteFox/node_modules/js-base64/test-moment/index.html</p>
<p>Path to vulnerable library: theWhiteFox/node_modules/js-base64/test-moment/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-2.1.4.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-1.11.0.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.0/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.0/jquery.min.js</a></p>
<p>Path to dependency file: theWhiteFox/node_modules/jquery-match-height/test/page/test.html</p>
<p>Path to vulnerable library: theWhiteFox/node_modules/jquery-match-height/test/page/./lib/jquery-1.11.0.min.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.11.0.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-1.7.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p>
<p>Path to dependency file: theWhiteFox/node_modules/vm-browserify/example/run/index.html</p>
<p>Path to vulnerable library: theWhiteFox/node_modules/vm-browserify/example/run/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.7.1.min.js** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/theWhiteFox/theWhiteFox/commit/2fd065d37ccb3afb8e9892f3e7d498280ba67038">2fd065d37ccb3afb8e9892f3e7d498280ba67038</a></p>
<p>Found in base branch: <b>gh-pages</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251>CVE-2015-9251</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - v3.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2015-9251 (Medium) detected in multiple libraries - ## CVE-2015-9251 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-2.1.4.min.js</b>, <b>jquery-1.11.0.min.js</b>, <b>jquery-1.7.1.min.js</b></p></summary>
<p>
<details><summary><b>jquery-2.1.4.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.4/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.4/jquery.min.js</a></p>
<p>Path to dependency file: theWhiteFox/node_modules/js-base64/test-moment/index.html</p>
<p>Path to vulnerable library: theWhiteFox/node_modules/js-base64/test-moment/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-2.1.4.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-1.11.0.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.0/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.0/jquery.min.js</a></p>
<p>Path to dependency file: theWhiteFox/node_modules/jquery-match-height/test/page/test.html</p>
<p>Path to vulnerable library: theWhiteFox/node_modules/jquery-match-height/test/page/./lib/jquery-1.11.0.min.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.11.0.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-1.7.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p>
<p>Path to dependency file: theWhiteFox/node_modules/vm-browserify/example/run/index.html</p>
<p>Path to vulnerable library: theWhiteFox/node_modules/vm-browserify/example/run/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.7.1.min.js** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/theWhiteFox/theWhiteFox/commit/2fd065d37ccb3afb8e9892f3e7d498280ba67038">2fd065d37ccb3afb8e9892f3e7d498280ba67038</a></p>
<p>Found in base branch: <b>gh-pages</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251>CVE-2015-9251</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - v3.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_priority
|
cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries jquery min js jquery min js jquery min js jquery min js javascript library for dom operations library home page a href path to dependency file thewhitefox node modules js test moment index html path to vulnerable library thewhitefox node modules js test moment index html dependency hierarchy x jquery min js vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file thewhitefox node modules jquery match height test page test html path to vulnerable library thewhitefox node modules jquery match height test page lib jquery min js dependency hierarchy x jquery min js vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file thewhitefox node modules vm browserify example run index html path to vulnerable library thewhitefox node modules vm browserify example run index html dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch gh pages vulnerability details jquery before is vulnerable to cross site scripting xss attacks when a cross domain ajax request is performed without the datatype option causing text javascript responses to be executed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource
| 0
|
547,168
| 16,039,038,745
|
IssuesEvent
|
2021-04-22 04:29:38
|
KingSupernova31/RulesGuru
|
https://api.github.com/repos/KingSupernova31/RulesGuru
|
opened
|
Question editor doesn't properly support undoing an action
|
bug medium priority
|
The automatic text replacements on the question editor interfere with the browser's native ctrl+z undo and redo functions. They should instead be added into a queue and handled properly.
|
1.0
|
Question editor doesn't properly support undoing an action - The automatic text replacements on the question editor interfere with the browser's native ctrl+z undo and redo functions. They should instead be added into a queue and handled properly.
|
priority
|
question editor doesn t properly support undoing an action the automatic text replacements on the question editor interfere with the browser s native ctrl z undo and redo functions they should instead be added into a queue and handled properly
| 1
|
315,650
| 9,630,300,986
|
IssuesEvent
|
2019-05-15 11:44:24
|
grafana/grafana
|
https://api.github.com/repos/grafana/grafana
|
closed
|
[Bug] day "d" not allowed in alert query condition eg. query(A,1d,now)
|
area/alerting help wanted priority/important-longterm type/bug
|
Please include this information:
- What Grafana version are you using?
4.02
- What datasource are you using?
OpenTSDB
- What OS are you running grafana on?
Windows
- What did you do?
Used a "day" in my alert query period. When I tried to save the dashboard the error "Invalid alert data cannot save dashboard" appears.
I used "query (A, 1d, now)" for the query. Also tried month (M) and year (y) but they also cause same problem. Only hour (h), minute (m) and second (s) work correctly.
|
1.0
|
[Bug] day "d" not allowed in alert query condition eg. query(A,1d,now) - Please include this information:
- What Grafana version are you using?
4.02
- What datasource are you using?
OpenTSDB
- What OS are you running grafana on?
Windows
- What did you do?
Used a "day" in my alert query period. When I tried to save the dashboard the error "Invalid alert data cannot save dashboard" appears.
I used "query (A, 1d, now)" for the query. Also tried month (M) and year (y) but they also cause same problem. Only hour (h), minute (m) and second (s) work correctly.
|
priority
|
day d not allowed in alert query condition eg query a now please include this information what grafana version are you using what datasource are you using opentsdb what os are you running grafana on windows what did you do used a day in my alert query period when i tried to save the dashboard the error invalid alert data cannot save dashboard appears i used query a now for the query also tried month m and year y but they also cause same problem only hour h minute m and second s work correctly
| 1
|
58,493
| 8,272,583,253
|
IssuesEvent
|
2018-09-16 21:45:21
|
stefanwalther/sense-navigation
|
https://api.github.com/repos/stefanwalther/sense-navigation
|
closed
|
v1.x: Contribution documentation + build info
|
documentation
|
- Document how to run/build/document the project based on v1.0
|
1.0
|
v1.x: Contribution documentation + build info - - Document how to run/build/document the project based on v1.0
|
non_priority
|
x contribution documentation build info document how to run build document the project based on
| 0
|
576,309
| 17,083,814,676
|
IssuesEvent
|
2021-07-08 09:12:47
|
aruppi/aruppi-api
|
https://api.github.com/repos/aruppi/aruppi-api
|
opened
|
Aruppi's new own id's system
|
enhancement important priority
|
**Is your feature request related to a problem? Please describe.**
A new id system needs to be implemented for a more optimised search of series information and the collection of information from all sources.
**Describe the solution you'd like**
This new internal function, would serve to have within the Aruppi database an own ID that, when calling it, would group all the id's of that series in the different sources (Animeflv, MonosChinos, JKAnime, TioAnime, etc) and in this way all the resources of the different pages could be obtained in a more unified way.
|
1.0
|
Aruppi's new own id's system - **Is your feature request related to a problem? Please describe.**
A new id system needs to be implemented for a more optimised search of series information and the collection of information from all sources.
**Describe the solution you'd like**
This new internal function, would serve to have within the Aruppi database an own ID that, when calling it, would group all the id's of that series in the different sources (Animeflv, MonosChinos, JKAnime, TioAnime, etc) and in this way all the resources of the different pages could be obtained in a more unified way.
|
priority
|
aruppi s new own id s system is your feature request related to a problem please describe a new id system needs to be implemented for a more optimised search of series information and the collection of information from all sources describe the solution you d like this new internal function would serve to have within the aruppi database an own id that when calling it would group all the id s of that series in the different sources animeflv monoschinos jkanime tioanime etc and in this way all the resources of the different pages could be obtained in a more unified way
| 1
|
137,418
| 11,136,771,678
|
IssuesEvent
|
2019-12-20 17:26:18
|
forseti-security/forseti-security
|
https://api.github.com/repos/forseti-security/forseti-security
|
closed
|
Verify Config Validator Scanner is working
|
module: testing priority: p3 triaged: yes
|
## Story
Add tests for the config validator to check that some sample policies are producing violations. Setup the policy in a way so that they aren't requiring major vulnerabilities (e.g. public CloudSQL).
## Solution
- Enable CV in test fixture
- Clone Policy Library repo and pull in a couple constraints from GCS
- Reset CV service
- Run scan
## Acceptance Criteria
Verify the constraints produce the correct violations.
|
1.0
|
Verify Config Validator Scanner is working - ## Story
Add tests for the config validator to check that some sample policies are producing violations. Setup the policy in a way so that they aren't requiring major vulnerabilities (e.g. public CloudSQL).
## Solution
- Enable CV in test fixture
- Clone Policy Library repo and pull in a couple constraints from GCS
- Reset CV service
- Run scan
## Acceptance Criteria
Verify the constraints produce the correct violations.
|
non_priority
|
verify config validator scanner is working story add tests for the config validator to check that some sample policies are producing violations setup the policy in a way so that they aren t requiring major vulnerabilities e g public cloudsql solution enable cv in test fixture clone policy library repo and pull in a couple constraints from gcs reset cv service run scan acceptance criteria verify the constraints produce the correct violations
| 0
|
138,610
| 12,824,850,698
|
IssuesEvent
|
2020-07-06 14:06:39
|
OlivierLDff/Qaterial
|
https://api.github.com/repos/OlivierLDff/Qaterial
|
closed
|
Icons ListView
|
documentation
|
Example that display all the icons in "qrc:/Qaterial/Icons"
- Display name in Icons.qml as a tooltip
- TextField to search
|
1.0
|
Icons ListView - Example that display all the icons in "qrc:/Qaterial/Icons"
- Display name in Icons.qml as a tooltip
- TextField to search
|
non_priority
|
icons listview example that display all the icons in qrc qaterial icons display name in icons qml as a tooltip textfield to search
| 0
|
550,707
| 16,130,777,593
|
IssuesEvent
|
2021-04-29 04:10:28
|
ucb-rit/coldfront
|
https://api.github.com/repos/ucb-rit/coldfront
|
closed
|
Give immediate feedback if email is already taken during registration
|
enhancement low priority
|
Currently, during registration, if the email address a user has provided is already in use, he/she must fill out the entire form to find out.
The field should make an asynchronous request when it is changed (possibly to the REST API), and give feedback immediately.
|
1.0
|
Give immediate feedback if email is already taken during registration - Currently, during registration, if the email address a user has provided is already in use, he/she must fill out the entire form to find out.
The field should make an asynchronous request when it is changed (possibly to the REST API), and give feedback immediately.
|
priority
|
give immediate feedback if email is already taken during registration currently during registration if the email address a user has provided is already in use he she must fill out the entire form to find out the field should make an asynchronous request when it is changed possibly to the rest api and give feedback immediately
| 1
|
348,827
| 31,722,011,787
|
IssuesEvent
|
2023-09-10 14:15:20
|
spring-projects/spring-framework
|
https://api.github.com/repos/spring-projects/spring-framework
|
opened
|
Introduce `ExecutableInvoker` API for `TestExecutionListener`s
|
in: test type: enhancement
|
## Overview
In order to be able to support parameter injection in `@BeforeTransaction` and `@AfterTransaction` methods (see #30736), we need a generic mechanism for delegating to the underlying testing framework to invoke methods. While we're at it, we might as well do the same for constructors.
In light of that, we should introduce an `ExecutableInvoker` API for `TestExecutionListener`s in the _Spring TestContext Framework_.
The default implementation should simply invoke the constructor/method without arguments.
A JUnit Jupiter specific implementation (registered in the `SpringExtension`) should delegate to the `org.junit.jupiter.api.extension.ExtensionContext.getExecutableInvoker()` mechanism introduced in JUnit Jupiter 5.9. This will allow a `TestExecutionListener` to transparently benefit from registered `ParameterResolver`s in JUnit Jupiter (including the `SpringExtension`) when invoking user constructors and methods, effectively providing support for _parameter injection_ for constructors and methods.
## Related Issues
- https://github.com/junit-team/junit5/issues/2393
- https://github.com/junit-team/junit5/issues/2191
- #30736
|
1.0
|
Introduce `ExecutableInvoker` API for `TestExecutionListener`s - ## Overview
In order to be able to support parameter injection in `@BeforeTransaction` and `@AfterTransaction` methods (see #30736), we need a generic mechanism for delegating to the underlying testing framework to invoke methods. While we're at it, we might as well do the same for constructors.
In light of that, we should introduce an `ExecutableInvoker` API for `TestExecutionListener`s in the _Spring TestContext Framework_.
The default implementation should simply invoke the constructor/method without arguments.
A JUnit Jupiter specific implementation (registered in the `SpringExtension`) should delegate to the `org.junit.jupiter.api.extension.ExtensionContext.getExecutableInvoker()` mechanism introduced in JUnit Jupiter 5.9. This will allow a `TestExecutionListener` to transparently benefit from registered `ParameterResolver`s in JUnit Jupiter (including the `SpringExtension`) when invoking user constructors and methods, effectively providing support for _parameter injection_ for constructors and methods.
## Related Issues
- https://github.com/junit-team/junit5/issues/2393
- https://github.com/junit-team/junit5/issues/2191
- #30736
|
non_priority
|
introduce executableinvoker api for testexecutionlistener s overview in order to be able to support parameter injection in beforetransaction and aftertransaction methods see we need a generic mechanism for delegating to the underlying testing framework to invoke methods while we re at it we might as well do the same for constructors in light of that we should introduce an executableinvoker api for testexecutionlistener s in the spring testcontext framework the default implementation should simply invoke the constructor method without arguments a junit jupiter specific implementation registered in the springextension should delegate to the org junit jupiter api extension extensioncontext getexecutableinvoker mechanism introduced in junit jupiter this will allow a testexecutionlistener to transparently benefit from registered parameterresolver s in junit jupiter including the springextension when invoking user constructors and methods effectively providing support for parameter injection for constructors and methods related issues
| 0
|
626,592
| 19,828,986,444
|
IssuesEvent
|
2022-01-20 09:59:20
|
ContinualAI/avalanche
|
https://api.github.com/repos/ContinualAI/avalanche
|
opened
|
Experiences (and benchmarks) with arbitrary data
|
Feature - Medium Priority Benchmarks
|
For my own research I need to provide additional data in the experiences/benchmark, such as a bunch of pretrained models (one for each experience). Right now, Avalanche doesn't support this use case. I can go around it, but I think it would be useful in other setting to have other kind of information in experiences. For example, in RL experiences have the environment and they do not have the dataset.
For the moment, I'm assigning this to me. It may be a large issue which will not be closed in a single PR.
|
1.0
|
Experiences (and benchmarks) with arbitrary data - For my own research I need to provide additional data in the experiences/benchmark, such as a bunch of pretrained models (one for each experience). Right now, Avalanche doesn't support this use case. I can go around it, but I think it would be useful in other setting to have other kind of information in experiences. For example, in RL experiences have the environment and they do not have the dataset.
For the moment, I'm assigning this to me. It may be a large issue which will not be closed in a single PR.
|
priority
|
experiences and benchmarks with arbitrary data for my own research i need to provide additional data in the experiences benchmark such as a bunch of pretrained models one for each experience right now avalanche doesn t support this use case i can go around it but i think it would be useful in other setting to have other kind of information in experiences for example in rl experiences have the environment and they do not have the dataset for the moment i m assigning this to me it may be a large issue which will not be closed in a single pr
| 1
|
20,322
| 11,430,510,060
|
IssuesEvent
|
2020-02-04 10:13:45
|
Azure/azure-powershell
|
https://api.github.com/repos/Azure/azure-powershell
|
closed
|
Add-AzTableRow - Exception calling "Execute" with "1" argument(s): "Object reference not set to an instance of an object."
|
External Service Attention Storage customer-reported
|
I am running following command to add record to Azure Table Storage:
Add-AzTableRow -table $tableAIR `
-partitionKey $PartitionKey `
-rowKey ([guid]::NewGuid().tostring()) `
-property @{"CreationTime"=$auditgeneralID.CreationTime;"Operation"=$auditgeneralID.Operation;"InvestigationId"=$auditgeneralID.InvestigationId;"InvestigationType"=$auditgeneralID.InvestigationType;"LastUpdateTimeUtc"=$auditgeneralID.LastUpdateTimeUtc;"StartTimeUtc"=$auditgeneralID.StartTimeUtc;"Status"=$auditgeneralID.Status;""=$auditgeneralID.UserIP}
and I am getting the following error:
Exception calling "Execute" with "1" argument(s): "Object reference not set to an instance of an object."
At C:\Program Files\WindowsPowerShell\Modules\AzureRmStorageTable\2.0.2\AzureRmStorageTableCoreHelper.psm1:239 char:10
+ ... return ($Table.Execute([Microsoft.Azure.Cosmos.Table.TableOperat ...
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+ CategoryInfo : NotSpecified: (:) [], MethodInvocationException
+ FullyQualifiedErrorId : StorageException
I have installed on my Windows 10 machine:
AzureRMStorageTable 2.0.2
Az.Storage 1.11.0
|
1.0
|
Add-AzTableRow - Exception calling "Execute" with "1" argument(s): "Object reference not set to an instance of an object." - I am running following command to add record to Azure Table Storage:
Add-AzTableRow -table $tableAIR `
-partitionKey $PartitionKey `
-rowKey ([guid]::NewGuid().tostring()) `
-property @{"CreationTime"=$auditgeneralID.CreationTime;"Operation"=$auditgeneralID.Operation;"InvestigationId"=$auditgeneralID.InvestigationId;"InvestigationType"=$auditgeneralID.InvestigationType;"LastUpdateTimeUtc"=$auditgeneralID.LastUpdateTimeUtc;"StartTimeUtc"=$auditgeneralID.StartTimeUtc;"Status"=$auditgeneralID.Status;""=$auditgeneralID.UserIP}
and I am getting the following error:
Exception calling "Execute" with "1" argument(s): "Object reference not set to an instance of an object."
At C:\Program Files\WindowsPowerShell\Modules\AzureRmStorageTable\2.0.2\AzureRmStorageTableCoreHelper.psm1:239 char:10
+ ... return ($Table.Execute([Microsoft.Azure.Cosmos.Table.TableOperat ...
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+ CategoryInfo : NotSpecified: (:) [], MethodInvocationException
+ FullyQualifiedErrorId : StorageException
I have installed on my Windows 10 machine:
AzureRMStorageTable 2.0.2
Az.Storage 1.11.0
|
non_priority
|
add aztablerow exception calling execute with argument s object reference not set to an instance of an object i am running following command to add record to azure table storage add aztablerow table tableair partitionkey partitionkey rowkey newguid tostring property creationtime auditgeneralid creationtime operation auditgeneralid operation investigationid auditgeneralid investigationid investigationtype auditgeneralid investigationtype lastupdatetimeutc auditgeneralid lastupdatetimeutc starttimeutc auditgeneralid starttimeutc status auditgeneralid status auditgeneralid userip and i am getting the following error exception calling execute with argument s object reference not set to an instance of an object at c program files windowspowershell modules azurermstoragetable azurermstoragetablecorehelper char return table execute microsoft azure cosmos table tableoperat categoryinfo notspecified methodinvocationexception fullyqualifiederrorid storageexception i have installed on my windows machine azurermstoragetable az storage
| 0
|
215,790
| 7,297,790,950
|
IssuesEvent
|
2018-02-26 15:13:41
|
kubernetes/kubernetes
|
https://api.github.com/repos/kubernetes/kubernetes
|
closed
|
Suggestion: Is it possible to get Kubernetes keep its Secrets in HashiCorp Vault
|
area/secret-api kind/feature priority/important-soon sig/auth
|
Is it possible to get Kubernetes keep its Secrets in HashiCorp Vault ?
https://github.com/hashicorp/vault/issues/377
Alex
|
1.0
|
Suggestion: Is it possible to get Kubernetes keep its Secrets in HashiCorp Vault - Is it possible to get Kubernetes keep its Secrets in HashiCorp Vault ?
https://github.com/hashicorp/vault/issues/377
Alex
|
priority
|
suggestion is it possible to get kubernetes keep its secrets in hashicorp vault is it possible to get kubernetes keep its secrets in hashicorp vault alex
| 1
|
187,688
| 6,760,264,511
|
IssuesEvent
|
2017-10-24 20:00:25
|
CCAFS/MARLO
|
https://api.github.com/repos/CCAFS/MARLO
|
closed
|
Create section for setting-up Contact Point Leaders
|
Priority - Medium Type - Enhancement
|
we need to have this section in "CRP Admin".
|
1.0
|
Create section for setting-up Contact Point Leaders - we need to have this section in "CRP Admin".
|
priority
|
create section for setting up contact point leaders we need to have this section in crp admin
| 1
|
605,448
| 18,735,415,392
|
IssuesEvent
|
2021-11-04 06:36:53
|
matrixorigin/matrixone
|
https://api.github.com/repos/matrixorigin/matrixone
|
opened
|
Comparison operators: Incorrect result when one operand is (1+ the maximum boundary value)
|
kind/bug priority/high needs-triage severity/major
|
<!-- Please describe your issue in English. -->
#### Can be reproduced ?
Yes.
#### Steps:
CREATE TABLE t1
(pk INTEGER PRIMARY KEY,
i1 TINYINT,
u1 TINYINT UNSIGNED,
i2 SMALLINT,
u2 SMALLINT UNSIGNED,
i4 INTEGER,
u4 INTEGER UNSIGNED,
i8 BIGINT,
u8 BIGINT UNSIGNED);
INSERT INTO t1 VALUES
(0, -128, 0, -32768, 0, -2147483648, 0, -9223372036854775808, 0),
(1, -1, 0, -1, 0, -1, 0, -1, 0),
(2, 0, 0, 0, 0, 0, 0, 0, 0),
(3, 1, 1, 1, 1, 1, 1, 1, 1),
(4, 127, 255, 32767, 65535, 2147483647, 4294967295,
9223372036854775807, 18446744073709551615);
select * from t1 where i1=128;
select * from t1 where i1>128;
select * from t1 where i1<128;
select * from t1 where i8=9223372036854775808;
select * from t1 where i8>9223372036854775808;
select * from t1 where i8<9223372036854775808;
#### Expected behavior:
mysql> select * from t1 where i1=128;
Empty set (0.00 sec)
mysql> select * from t1 where i1>128;
Empty set (0.00 sec)
mysql> select * from t1 where i1<128;
+----+------+------+--------+-------+-------------+------------+----------------------+----------------------+
| pk | i1 | u1 | i2 | u2 | i4 | u4 | i8 | u8 |
+----+------+------+--------+-------+-------------+------------+----------------------+----------------------+
| 0 | -128 | 0 | -32768 | 0 | -2147483648 | 0 | -9223372036854775808 | 0 |
| 1 | -1 | 0 | -1 | 0 | -1 | 0 | -1 | 0 |
| 2 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
| 3 | 1 | 1 | 1 | 1 | 1 | 1 | 1 | 1 |
| 4 | 127 | 255 | 32767 | 65535 | 2147483647 | 4294967295 | 9223372036854775807 | 18446744073709551615 |
+----+------+------+--------+-------+-------------+------------+----------------------+----------------------+
5 rows in set (0.00 sec)
mysql> select * from t1 where i8=9223372036854775808;
Empty set (0.00 sec)
mysql> select * from t1 where i8>9223372036854775808;
Empty set (0.00 sec)
mysql> select * from t1 where i8<9223372036854775808;
+----+------+------+--------+-------+-------------+------------+----------------------+----------------------+
| pk | i1 | u1 | i2 | u2 | i4 | u4 | i8 | u8 |
+----+------+------+--------+-------+-------------+------------+----------------------+----------------------+
| 0 | -128 | 0 | -32768 | 0 | -2147483648 | 0 | -9223372036854775808 | 0 |
| 1 | -1 | 0 | -1 | 0 | -1 | 0 | -1 | 0 |
| 2 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
| 3 | 1 | 1 | 1 | 1 | 1 | 1 | 1 | 1 |
| 4 | 127 | 255 | 32767 | 65535 | 2147483647 | 4294967295 | 9223372036854775807 | 18446744073709551615 |
+----+------+------+--------+-------+-------------+------------+----------------------+----------------------+
5 rows in set (0.00 sec)
#### Actual behavior:
mysql> select * from t1 where i1=128;
+------+------+------+--------+------+-------------+------+----------------------+------+
| pk | i1 | u1 | i2 | u2 | i4 | u4 | i8 | u8 |
+------+------+------+--------+------+-------------+------+----------------------+------+
| 0 | -128 | 0 | -32768 | 0 | -2147483648 | 0 | -9223372036854775808 | 0 |
+------+------+------+--------+------+-------------+------+----------------------+------+
1 row in set (0.01 sec)
mysql> select * from t1 where i1>128;
+------+------+------+-------+-------+------------+------------+---------------------+----------------------+
| pk | i1 | u1 | i2 | u2 | i4 | u4 | i8 | u8 |
+------+------+------+-------+-------+------------+------------+---------------------+----------------------+
| 1 | -1 | 0 | -1 | 0 | -1 | 0 | -1 | 0 |
| 2 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
| 3 | 1 | 1 | 1 | 1 | 1 | 1 | 1 | 1 |
| 4 | 127 | 255 | 32767 | 65535 | 2147483647 | 4294967295 | 9223372036854775807 | 18446744073709551615 |
+------+------+------+-------+-------+------------+------------+---------------------+----------------------+
4 rows in set (0.00 sec)
mysql> select * from t1 where i1<128;
Empty set (0.01 sec)
mysql> select * from t1 where i8=9223372036854775808;
+------+------+------+--------+------+-------------+------+----------------------+------+
| pk | i1 | u1 | i2 | u2 | i4 | u4 | i8 | u8 |
+------+------+------+--------+------+-------------+------+----------------------+------+
| 0 | -128 | 0 | -32768 | 0 | -2147483648 | 0 | -9223372036854775808 | 0 |
+------+------+------+--------+------+-------------+------+----------------------+------+
1 row in set (0.00 sec)
mysql> select * from t1 where i8>9223372036854775808;
+------+------+------+-------+-------+------------+------------+---------------------+----------------------+
| pk | i1 | u1 | i2 | u2 | i4 | u4 | i8 | u8 |
+------+------+------+-------+-------+------------+------------+---------------------+----------------------+
| 1 | -1 | 0 | -1 | 0 | -1 | 0 | -1 | 0 |
| 2 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
| 3 | 1 | 1 | 1 | 1 | 1 | 1 | 1 | 1 |
| 4 | 127 | 255 | 32767 | 65535 | 2147483647 | 4294967295 | 9223372036854775807 | 18446744073709551615 |
+------+------+------+-------+-------+------------+------------+---------------------+----------------------+
4 rows in set (0.01 sec)
mysql> select * from t1 where i8<9223372036854775808;
Empty set (0.01 sec)
#### Environment:
- Version or commit-id (e.g. v0.1.0 or 8b23a93):commit 693ab24c885ee26df6fe5ad467dd46c7cbdbe950
- Hardware parameters:
- OS type:
- Others:
#### Configuration file:
#### Additional context:
- Error message from client:
- Server log:
- Other information:
|
1.0
|
Comparison operators: Incorrect result when one operand is (1+ the maximum boundary value) - <!-- Please describe your issue in English. -->
#### Can be reproduced ?
Yes.
#### Steps:
CREATE TABLE t1
(pk INTEGER PRIMARY KEY,
i1 TINYINT,
u1 TINYINT UNSIGNED,
i2 SMALLINT,
u2 SMALLINT UNSIGNED,
i4 INTEGER,
u4 INTEGER UNSIGNED,
i8 BIGINT,
u8 BIGINT UNSIGNED);
INSERT INTO t1 VALUES
(0, -128, 0, -32768, 0, -2147483648, 0, -9223372036854775808, 0),
(1, -1, 0, -1, 0, -1, 0, -1, 0),
(2, 0, 0, 0, 0, 0, 0, 0, 0),
(3, 1, 1, 1, 1, 1, 1, 1, 1),
(4, 127, 255, 32767, 65535, 2147483647, 4294967295,
9223372036854775807, 18446744073709551615);
select * from t1 where i1=128;
select * from t1 where i1>128;
select * from t1 where i1<128;
select * from t1 where i8=9223372036854775808;
select * from t1 where i8>9223372036854775808;
select * from t1 where i8<9223372036854775808;
#### Expected behavior:
mysql> select * from t1 where i1=128;
Empty set (0.00 sec)
mysql> select * from t1 where i1>128;
Empty set (0.00 sec)
mysql> select * from t1 where i1<128;
+----+------+------+--------+-------+-------------+------------+----------------------+----------------------+
| pk | i1 | u1 | i2 | u2 | i4 | u4 | i8 | u8 |
+----+------+------+--------+-------+-------------+------------+----------------------+----------------------+
| 0 | -128 | 0 | -32768 | 0 | -2147483648 | 0 | -9223372036854775808 | 0 |
| 1 | -1 | 0 | -1 | 0 | -1 | 0 | -1 | 0 |
| 2 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
| 3 | 1 | 1 | 1 | 1 | 1 | 1 | 1 | 1 |
| 4 | 127 | 255 | 32767 | 65535 | 2147483647 | 4294967295 | 9223372036854775807 | 18446744073709551615 |
+----+------+------+--------+-------+-------------+------------+----------------------+----------------------+
5 rows in set (0.00 sec)
mysql> select * from t1 where i8=9223372036854775808;
Empty set (0.00 sec)
mysql> select * from t1 where i8>9223372036854775808;
Empty set (0.00 sec)
mysql> select * from t1 where i8<9223372036854775808;
+----+------+------+--------+-------+-------------+------------+----------------------+----------------------+
| pk | i1 | u1 | i2 | u2 | i4 | u4 | i8 | u8 |
+----+------+------+--------+-------+-------------+------------+----------------------+----------------------+
| 0 | -128 | 0 | -32768 | 0 | -2147483648 | 0 | -9223372036854775808 | 0 |
| 1 | -1 | 0 | -1 | 0 | -1 | 0 | -1 | 0 |
| 2 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
| 3 | 1 | 1 | 1 | 1 | 1 | 1 | 1 | 1 |
| 4 | 127 | 255 | 32767 | 65535 | 2147483647 | 4294967295 | 9223372036854775807 | 18446744073709551615 |
+----+------+------+--------+-------+-------------+------------+----------------------+----------------------+
5 rows in set (0.00 sec)
#### Actual behavior:
mysql> select * from t1 where i1=128;
+------+------+------+--------+------+-------------+------+----------------------+------+
| pk | i1 | u1 | i2 | u2 | i4 | u4 | i8 | u8 |
+------+------+------+--------+------+-------------+------+----------------------+------+
| 0 | -128 | 0 | -32768 | 0 | -2147483648 | 0 | -9223372036854775808 | 0 |
+------+------+------+--------+------+-------------+------+----------------------+------+
1 row in set (0.01 sec)
mysql> select * from t1 where i1>128;
+------+------+------+-------+-------+------------+------------+---------------------+----------------------+
| pk | i1 | u1 | i2 | u2 | i4 | u4 | i8 | u8 |
+------+------+------+-------+-------+------------+------------+---------------------+----------------------+
| 1 | -1 | 0 | -1 | 0 | -1 | 0 | -1 | 0 |
| 2 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
| 3 | 1 | 1 | 1 | 1 | 1 | 1 | 1 | 1 |
| 4 | 127 | 255 | 32767 | 65535 | 2147483647 | 4294967295 | 9223372036854775807 | 18446744073709551615 |
+------+------+------+-------+-------+------------+------------+---------------------+----------------------+
4 rows in set (0.00 sec)
mysql> select * from t1 where i1<128;
Empty set (0.01 sec)
mysql> select * from t1 where i8=9223372036854775808;
+------+------+------+--------+------+-------------+------+----------------------+------+
| pk | i1 | u1 | i2 | u2 | i4 | u4 | i8 | u8 |
+------+------+------+--------+------+-------------+------+----------------------+------+
| 0 | -128 | 0 | -32768 | 0 | -2147483648 | 0 | -9223372036854775808 | 0 |
+------+------+------+--------+------+-------------+------+----------------------+------+
1 row in set (0.00 sec)
mysql> select * from t1 where i8>9223372036854775808;
+------+------+------+-------+-------+------------+------------+---------------------+----------------------+
| pk | i1 | u1 | i2 | u2 | i4 | u4 | i8 | u8 |
+------+------+------+-------+-------+------------+------------+---------------------+----------------------+
| 1 | -1 | 0 | -1 | 0 | -1 | 0 | -1 | 0 |
| 2 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
| 3 | 1 | 1 | 1 | 1 | 1 | 1 | 1 | 1 |
| 4 | 127 | 255 | 32767 | 65535 | 2147483647 | 4294967295 | 9223372036854775807 | 18446744073709551615 |
+------+------+------+-------+-------+------------+------------+---------------------+----------------------+
4 rows in set (0.01 sec)
mysql> select * from t1 where i8<9223372036854775808;
Empty set (0.01 sec)
#### Environment:
- Version or commit-id (e.g. v0.1.0 or 8b23a93):commit 693ab24c885ee26df6fe5ad467dd46c7cbdbe950
- Hardware parameters:
- OS type:
- Others:
#### Configuration file:
#### Additional context:
- Error message from client:
- Server log:
- Other information:
|
priority
|
comparison operators incorrect result when one operand is the maximum boundary value can be reproduced yes steps create table pk integer primary key tinyint tinyint unsigned smallint smallint unsigned integer integer unsigned bigint bigint unsigned insert into values select from where select from where select from where select from where select from where select from where expected behavior mysql select from where empty set sec mysql select from where empty set sec mysql select from where pk rows in set sec mysql select from where empty set sec mysql select from where empty set sec mysql select from where pk rows in set sec actual behavior mysql select from where pk row in set sec mysql select from where pk rows in set sec mysql select from where empty set sec mysql select from where pk row in set sec mysql select from where pk rows in set sec mysql select from where empty set sec environment version or commit id e g or commit hardware parameters os type others configuration file additional context error message from client server log other information
| 1
|
162,494
| 6,154,355,716
|
IssuesEvent
|
2017-06-28 12:29:18
|
surajssd/kapp
|
https://api.github.com/repos/surajssd/kapp
|
closed
|
Defining applications in a single file
|
priority/high
|
Looking at the examples, the only way I can see for defining multiple microservices, is by putting them in different files.
e.g. - https://github.com/surajssd/opencomposition/tree/master/examples/allnomagic here, db.yaml and web.yaml _have_ to be separate files.
This is fine and good practice under certain scenarios, but I also want to be able to define this in a single file.
|
1.0
|
Defining applications in a single file - Looking at the examples, the only way I can see for defining multiple microservices, is by putting them in different files.
e.g. - https://github.com/surajssd/opencomposition/tree/master/examples/allnomagic here, db.yaml and web.yaml _have_ to be separate files.
This is fine and good practice under certain scenarios, but I also want to be able to define this in a single file.
|
priority
|
defining applications in a single file looking at the examples the only way i can see for defining multiple microservices is by putting them in different files e g here db yaml and web yaml have to be separate files this is fine and good practice under certain scenarios but i also want to be able to define this in a single file
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.