diff --git "a/reduced_main_data.csv" "b/reduced_main_data.csv"
new file mode 100644--- /dev/null
+++ "b/reduced_main_data.csv"
@@ -0,0 +1,84604 @@
+Unnamed: 0,id,type,created_at,repo,repo_url,action,title,labels,body,index,text_combine,label,text,binary_label
+14636,10197331841.0,IssuesEvent,2019-08-12 23:52:34,cityofaustin/atd-data-tech,https://api.github.com/repos/cityofaustin/atd-data-tech,closed,"Meeting ""Discuss/Demo Digital Interview Roll Out"" with HR",Project: Paperless Hiring Service: PM Type: Meeting Workgroup: HR,"Scheduled: Wed, 8/7/2019
+
+Meet with HR to clarify need
+
+Attendees: John, Diana, Adan and Nathan B.",1.0,"Meeting ""Discuss/Demo Digital Interview Roll Out"" with HR - Scheduled: Wed, 8/7/2019
+
+Meet with HR to clarify need
+
+Attendees: John, Diana, Adan and Nathan B.",0,meeting discuss demo digital interview roll out with hr scheduled wed meet with hr to clarify need attendees john diana adan and nathan b ,0
+279374,21157852282.0,IssuesEvent,2022-04-07 06:25:34,AY2122S2-CS2103-F11-2/tp,https://api.github.com/repos/AY2122S2-CS2103-F11-2/tp,closed,[PE-D] [UG] incorrect command name,Documentation response.Accepted,"Should be notes about the sort format
+
+
+
+
+
+
+
+
+-------------
+Labels: `severity.VeryLow` `type.DocumentationBug`
+original: mazx4960/ped#12",1.0,"[PE-D] [UG] incorrect command name - Should be notes about the sort format
+
+
+
+
+
+
+
+
+-------------
+Labels: `severity.VeryLow` `type.DocumentationBug`
+original: mazx4960/ped#12",0, incorrect command name should be notes about the sort format labels severity verylow type documentationbug original ped ,0
+91701,18676614623.0,IssuesEvent,2021-10-31 17:11:33,CiviWiki/OpenCiviWiki,https://api.github.com/repos/CiviWiki/OpenCiviWiki,closed,[BUG]: Major bug fixes through the UI,bug community code quality documentation triage,"### Description
+
+There are multiple routing issues and bugs which I am spotting through testing, this is a self-assigned issue and I'll elaborate on the issues I resolve here soon.
+
+
+[Update 1]:
+The following bugs have been spotted:
+- [x] Reverse name maps for urls
+- [x] Profile doesn't load from the header after login
+- [x] Profile creation keeps loading infinitely
+- [x] HTTP 403 for POST on profile setup
+- [ ] The footer needs to be added to the bottom
+
+",1.0,"[BUG]: Major bug fixes through the UI - ### Description
+
+There are multiple routing issues and bugs which I am spotting through testing, this is a self-assigned issue and I'll elaborate on the issues I resolve here soon.
+
+
+[Update 1]:
+The following bugs have been spotted:
+- [x] Reverse name maps for urls
+- [x] Profile doesn't load from the header after login
+- [x] Profile creation keeps loading infinitely
+- [x] HTTP 403 for POST on profile setup
+- [ ] The footer needs to be added to the bottom
+
+",0, major bug fixes through the ui description there are multiple routing issues and bugs which i am spotting through testing this is a self assigned issue and i ll elaborate on the issues i resolve here soon the following bugs have been spotted reverse name maps for urls profile doesn t load from the header after login profile creation keeps loading infinitely http for post on profile setup the footer needs to be added to the bottom ,0
+502,8648832136.0,IssuesEvent,2018-11-26 17:36:14,OfficeDev/office-ui-fabric-react,https://api.github.com/repos/OfficeDev/office-ui-fabric-react,closed,PeoplePicker: className of inputProperties is not merged,Component: PeoplePicker,"
+
+Properties ""placeholder"" + ""id"" are working correctly! className is not merged in the control :(",1.0,"PeoplePicker: className of inputProperties is not merged -
+
+Properties ""placeholder"" + ""id"" are working correctly! className is not merged in the control :(",1,peoplepicker classname of inputproperties is not merged peoplepicker inputproperties placeholder bitte einen benutzer angeben id pckselecteduser classname testclass properties placeholder id are working correctly classname is not merged in the control ,1
+123997,12223386864.0,IssuesEvent,2020-05-02 17:21:53,flutter/flutter,https://api.github.com/repos/flutter/flutter,closed,pub.dev should expose if a Plugin supports AndroidX,documentation p: third party plugin proposal severe: new feature,"## Use case
+I got an issue related to a plugin updated while it uses AndroidX
+
+## Proposal
+Option 1-I want a warning on breaking changes on the console when a plugin got an update.
+Option 2-When someone makes a new app or plugin it should use AndroidX by default.
+Option 3-When a plugin uses AndroidX there should be an indication on pub.dev website.
+Option 4-Show specification section on pub.dev website like min SDK, iOS version, AndroidX support, etc.",1.0,"pub.dev should expose if a Plugin supports AndroidX - ## Use case
+I got an issue related to a plugin updated while it uses AndroidX
+
+## Proposal
+Option 1-I want a warning on breaking changes on the console when a plugin got an update.
+Option 2-When someone makes a new app or plugin it should use AndroidX by default.
+Option 3-When a plugin uses AndroidX there should be an indication on pub.dev website.
+Option 4-Show specification section on pub.dev website like min SDK, iOS version, AndroidX support, etc.",0,pub dev should expose if a plugin supports androidx use case i got an issue related to a plugin updated while it uses androidx proposal option i want a warning on breaking changes on the console when a plugin got an update option when someone makes a new app or plugin it should use androidx by default option when a plugin uses androidx there should be an indication on pub dev website option show specification section on pub dev website like min sdk ios version androidx support etc ,0
+1055,25253341457.0,IssuesEvent,2022-11-15 16:10:05,codestates-seb/seb40_main_009,https://api.github.com/repos/codestates-seb/seb40_main_009,closed,"뱃지 세부/ 뱃지, 경험치 챌린지와 멤버 연관관계 매핑 후 어떨때 경험치가 쌓일지 기준 마련",🥴도움이 필요! BE김민섭🦧 BE김유현🚌 BE오영운🚌 FE한병주🚌 FE심이서 :people_holding_hands: FE김은비🚌,"1. 수행전
+- 목표기한 : 2022.11.16
+- 필수 달성 항목:
+ - [x] 경험치 쌓는 여러 기준 마련
+
+2. 수행후
+- 달성일자 : 2022.11.16
+- 달성내용 :
+ - [ ] 1씩 증가 -> 참여중인 챌린지에 하루 인증 마쳤을때
+ - [ ] 3씩 증가 -> 참여중인 챌린지를 완료 했을때
+ - [ ] 각 레벨업의 기준은 2의 제곱으로 증가 ex) 1레벨 ->기준 2포인트/ 2레벨 -> 기준 4포인트/ 8레벨 -> 기준 256포인트
+- 차후 개선사항 :
+- 달성내용에 맞게 경험치, 뱃지 로직 변경
+",1.0,"뱃지 세부/ 뱃지, 경험치 챌린지와 멤버 연관관계 매핑 후 어떨때 경험치가 쌓일지 기준 마련 - 1. 수행전
+- 목표기한 : 2022.11.16
+- 필수 달성 항목:
+ - [x] 경험치 쌓는 여러 기준 마련
+
+2. 수행후
+- 달성일자 : 2022.11.16
+- 달성내용 :
+ - [ ] 1씩 증가 -> 참여중인 챌린지에 하루 인증 마쳤을때
+ - [ ] 3씩 증가 -> 참여중인 챌린지를 완료 했을때
+ - [ ] 각 레벨업의 기준은 2의 제곱으로 증가 ex) 1레벨 ->기준 2포인트/ 2레벨 -> 기준 4포인트/ 8레벨 -> 기준 256포인트
+- 차후 개선사항 :
+- 달성내용에 맞게 경험치, 뱃지 로직 변경
+",1,뱃지 세부 뱃지 경험치 챌린지와 멤버 연관관계 매핑 후 어떨때 경험치가 쌓일지 기준 마련 수행전 목표기한 필수 달성 항목 경험치 쌓는 여러 기준 마련 수행후 달성일자 달성내용 증가 참여중인 챌린지에 하루 인증 마쳤을때 증가 참여중인 챌린지를 완료 했을때 각 레벨업의 기준은 제곱으로 증가 ex 기준 기준 기준 차후 개선사항 달성내용에 맞게 경험치 뱃지 로직 변경 ,1
+120069,10092074619.0,IssuesEvent,2019-07-26 15:43:10,ampproject/amp-wp,https://api.github.com/repos/ampproject/amp-wp,opened,E2E Tests: Switching back from code editor,AMP Stories Needs Tests [Integration] Gutenberg [Type] Enhancement,"* [ ] Add a test to verify that switching to the code editor and back does not break the editor, i.e. all story controls re-appear again.",1.0,"E2E Tests: Switching back from code editor - * [ ] Add a test to verify that switching to the code editor and back does not break the editor, i.e. all story controls re-appear again.",0, tests switching back from code editor add a test to verify that switching to the code editor and back does not break the editor i e all story controls re appear again ,0
+218023,7330209319.0,IssuesEvent,2018-03-05 09:08:40,wso2/product-is,https://api.github.com/repos/wso2/product-is,reopened,config.json file path is taken in to a wrong location when you run forget-me tool,Affected/5.5.0-Alpha Priority/Highest Type/Bug,"config.json file path is taken in to a wrong location when you run forget-me tool as per this [1] doc
+
+[1] https://docs.wso2.com/display/ADMIN44x/Removing+References+to+Deleted+User+Identities+in+WSO2+Products#RemovingReferencestoDeletedUserIdentitiesinWSO2Products-MasterConfig
+
+./forget-me -U
+
+./forget-me -U ushani2
+2018-02-19 16:22:38 INFO ForgetMeTool:167 - Generating pseudonym as pseudo name is not provided : 742878f2-ff35-49d3-9c4f-39b6d60c374a
+Exception in thread ""main"" org.wso2.carbon.privacy.forgetme.runtime.ForgetMeExecutionException: Could not read the config files related to : /home/ushani/WSO2_WORK/Products/IS/5.5.0/alpha/identity-anonymization-tool/components/org.wso2.carbon.privacy.forgetme.tool/target/org.wso2.carbon.privacy.forgetme.tool-1.0.3-SNAPSHOT/config.json
+ at org.wso2.carbon.privacy.forgetme.ConfigReader.readSystemConfig(ConfigReader.java:117)
+ at org.wso2.carbon.privacy.forgetme.ForgetMeTool.process(ForgetMeTool.java:185)
+ at org.wso2.carbon.privacy.forgetme.ForgetMeTool.main(ForgetMeTool.java:128)
+Caused by: java.io.FileNotFoundException: /home/ushani/WSO2_WORK/Products/IS/5.5.0/alpha/identity-anonymization-tool/components/org.wso2.carbon.privacy.forgetme.tool/target/org.wso2.carbon.privacy.forgetme.tool-1.0.3-SNAPSHOT/config.json (No such file or directory)
+ at java.io.FileInputStream.open0(Native Method)
+ at java.io.FileInputStream.open(FileInputStream.java:195)
+ at java.io.FileInputStream.(FileInputStream.java:138)
+ at java.io.FileReader.(FileReader.java:72)
+ at org.wso2.carbon.privacy.forgetme.ConfigReader.readSystemConfig(ConfigReader.java:96)
+ ... 2 more
+
+
+config.json is inside conf folder. Below is displayed without /conf
+**home/ushani/WSO2_WORK/Products/IS/5.5.0/alpha/identity-anonymization-tool/components/org.wso2.carbon.privacy.forgetme.tool/target/org.wso2.carbon.privacy.forgetme.tool-1.0.3-SNAPSHOT/config.json**",1.0,"config.json file path is taken in to a wrong location when you run forget-me tool - config.json file path is taken in to a wrong location when you run forget-me tool as per this [1] doc
+
+[1] https://docs.wso2.com/display/ADMIN44x/Removing+References+to+Deleted+User+Identities+in+WSO2+Products#RemovingReferencestoDeletedUserIdentitiesinWSO2Products-MasterConfig
+
+./forget-me -U
+
+./forget-me -U ushani2
+2018-02-19 16:22:38 INFO ForgetMeTool:167 - Generating pseudonym as pseudo name is not provided : 742878f2-ff35-49d3-9c4f-39b6d60c374a
+Exception in thread ""main"" org.wso2.carbon.privacy.forgetme.runtime.ForgetMeExecutionException: Could not read the config files related to : /home/ushani/WSO2_WORK/Products/IS/5.5.0/alpha/identity-anonymization-tool/components/org.wso2.carbon.privacy.forgetme.tool/target/org.wso2.carbon.privacy.forgetme.tool-1.0.3-SNAPSHOT/config.json
+ at org.wso2.carbon.privacy.forgetme.ConfigReader.readSystemConfig(ConfigReader.java:117)
+ at org.wso2.carbon.privacy.forgetme.ForgetMeTool.process(ForgetMeTool.java:185)
+ at org.wso2.carbon.privacy.forgetme.ForgetMeTool.main(ForgetMeTool.java:128)
+Caused by: java.io.FileNotFoundException: /home/ushani/WSO2_WORK/Products/IS/5.5.0/alpha/identity-anonymization-tool/components/org.wso2.carbon.privacy.forgetme.tool/target/org.wso2.carbon.privacy.forgetme.tool-1.0.3-SNAPSHOT/config.json (No such file or directory)
+ at java.io.FileInputStream.open0(Native Method)
+ at java.io.FileInputStream.open(FileInputStream.java:195)
+ at java.io.FileInputStream.(FileInputStream.java:138)
+ at java.io.FileReader.(FileReader.java:72)
+ at org.wso2.carbon.privacy.forgetme.ConfigReader.readSystemConfig(ConfigReader.java:96)
+ ... 2 more
+
+
+config.json is inside conf folder. Below is displayed without /conf
+**home/ushani/WSO2_WORK/Products/IS/5.5.0/alpha/identity-anonymization-tool/components/org.wso2.carbon.privacy.forgetme.tool/target/org.wso2.carbon.privacy.forgetme.tool-1.0.3-SNAPSHOT/config.json**",0,config json file path is taken in to a wrong location when you run forget me tool config json file path is taken in to a wrong location when you run forget me tool as per this doc forget me u forget me u info forgetmetool generating pseudonym as pseudo name is not provided exception in thread main org carbon privacy forgetme runtime forgetmeexecutionexception could not read the config files related to home ushani work products is alpha identity anonymization tool components org carbon privacy forgetme tool target org carbon privacy forgetme tool snapshot config json at org carbon privacy forgetme configreader readsystemconfig configreader java at org carbon privacy forgetme forgetmetool process forgetmetool java at org carbon privacy forgetme forgetmetool main forgetmetool java caused by java io filenotfoundexception home ushani work products is alpha identity anonymization tool components org carbon privacy forgetme tool target org carbon privacy forgetme tool snapshot config json no such file or directory at java io fileinputstream native method at java io fileinputstream open fileinputstream java at java io fileinputstream fileinputstream java at java io filereader filereader java at org carbon privacy forgetme configreader readsystemconfig configreader java more config json is inside conf folder below is displayed without conf home ushani work products is alpha identity anonymization tool components org carbon privacy forgetme tool target org carbon privacy forgetme tool snapshot config json ,0
+374872,26136767399.0,IssuesEvent,2022-12-29 13:07:14,mlkmooeede/mlkmoood,https://api.github.com/repos/mlkmooeede/mlkmoood,closed,__Originally posted by @mlkmooeede in https://github.com/nicklockwood/AutoCoding/issues/38__,bug documentation duplicate enhancement help wanted good first issue invalid question wontfix,"__Originally posted by @mlkmooeede in https://github.com/nicklockwood/AutoCoding/issues/38__
+
+_Originally posted by @mlkmooeede in https://github.com/mlkmooeede/mlkmoood/issues/1_",1.0,"__Originally posted by @mlkmooeede in https://github.com/nicklockwood/AutoCoding/issues/38__ - __Originally posted by @mlkmooeede in https://github.com/nicklockwood/AutoCoding/issues/38__
+
+_Originally posted by @mlkmooeede in https://github.com/mlkmooeede/mlkmoood/issues/1_",0, originally posted by mlkmooeede in originally posted by mlkmooeede in originally posted by mlkmooeede in ,0
+535557,15690152806.0,IssuesEvent,2021-03-25 16:24:53,craftercms/craftercms,https://api.github.com/repos/craftercms/craftercms,closed,[studio-ui] Publish Status is empty when doing a fresh login to Studio,bug priority: medium,"## Describe the bug
+Publishing Status is empty when a user enter to Studio from a clean browser session (Incognito)
+
+## To Reproduce
+Steps to reproduce the behavior:
+1. Start from a fresh Studio deploy
+2. Create a new Site
+3. Go to Site Config - > Publishing -> Publish Status
+4. See error
+
+## Expected behavior
+Status is displayed.
+
+## Screenshots
+https://user-images.githubusercontent.com/6722074/110540326-fcb01480-80eb-11eb-9c15-79d6f0442664.mp4
+
+## Logs
+```
+jQuery.Deferred exception: Cannot read property 'toLowerCase' of null TypeError: Cannot read property 'toLowerCase' of null
+ at renderDashboard (http://localhost:8080/studio/static-assets/scripts/admin.js?version=dc80c7660cf8a025a2a94cda611ea57c2681a997:1056:28)
+ at HTMLDocument. (http://localhost:8080/studio/static-assets/scripts/admin.js?version=dc80c7660cf8a025a2a94cda611ea57c2681a997:998:11)
+ at mightThrow (http://localhost:8080/studio/static-assets/libs/jquery/dist/jquery.js?version=dc80c7660cf8a025a2a94cda611ea57c2681a997:3557:60)
+ at process (http://localhost:8080/studio/static-assets/libs/jquery/dist/jquery.js?version=dc80c7660cf8a025a2a94cda611ea57c2681a997:3625:49)
+```
+
+## Specs
+### Version
+`4.0.0-SNAPSHOT`
+
+### OS
+Any
+
+### Browser
+Any
+
+## Additional context
+N/A
+",1.0,"[studio-ui] Publish Status is empty when doing a fresh login to Studio - ## Describe the bug
+Publishing Status is empty when a user enter to Studio from a clean browser session (Incognito)
+
+## To Reproduce
+Steps to reproduce the behavior:
+1. Start from a fresh Studio deploy
+2. Create a new Site
+3. Go to Site Config - > Publishing -> Publish Status
+4. See error
+
+## Expected behavior
+Status is displayed.
+
+## Screenshots
+https://user-images.githubusercontent.com/6722074/110540326-fcb01480-80eb-11eb-9c15-79d6f0442664.mp4
+
+## Logs
+```
+jQuery.Deferred exception: Cannot read property 'toLowerCase' of null TypeError: Cannot read property 'toLowerCase' of null
+ at renderDashboard (http://localhost:8080/studio/static-assets/scripts/admin.js?version=dc80c7660cf8a025a2a94cda611ea57c2681a997:1056:28)
+ at HTMLDocument. (http://localhost:8080/studio/static-assets/scripts/admin.js?version=dc80c7660cf8a025a2a94cda611ea57c2681a997:998:11)
+ at mightThrow (http://localhost:8080/studio/static-assets/libs/jquery/dist/jquery.js?version=dc80c7660cf8a025a2a94cda611ea57c2681a997:3557:60)
+ at process (http://localhost:8080/studio/static-assets/libs/jquery/dist/jquery.js?version=dc80c7660cf8a025a2a94cda611ea57c2681a997:3625:49)
+```
+
+## Specs
+### Version
+`4.0.0-SNAPSHOT`
+
+### OS
+Any
+
+### Browser
+Any
+
+## Additional context
+N/A
+",0, publish status is empty when doing a fresh login to studio describe the bug publishing status is empty when a user enter to studio from a clean browser session incognito to reproduce steps to reproduce the behavior start from a fresh studio deploy create a new site go to site config publishing publish status see error expected behavior status is displayed screenshots logs jquery deferred exception cannot read property tolowercase of null typeerror cannot read property tolowercase of null at renderdashboard at htmldocument at mightthrow at process specs version snapshot os any browser any additional context n a ,0
+825,15287985558.0,IssuesEvent,2021-02-23 16:21:41,openstates/issues,https://api.github.com/repos/openstates/issues,closed,Massachuettes Legislator Spot Check Issues,component:people-data type:bug,"State: MA
+Short Description: When checking Massachuettes’s legislators, I found an extra legislator on OS and a vacancy listed on the legislator site that is not on OS
+
+Include these 3 sections as appropriate (or note none)
+
+Missing or Incorrect legislators:
+ (include vacancies, mismatched legislators, etc.)
+A vacancy on the legislative site is listed, but OS lists Robert A. DeLeo instead of the vacancy. His service ended in 2020.
+OS Link: https://openstates.org/person/robert-a-deleo-3pojeaJMWvGlYxjpYkSr9j/
+Legislative Link of Robert A. DeLeo: https://malegislature.gov/Legislators/Profile/RAD1
+Legislative Link Showing The Vacancy: https://malegislature.gov/Legislators/Members/House
+
+Photos:
+
+
+
+
+Data Issues:
+N/A
+
+Additional Data:
+N/A
+
+
+",1.0,"Massachuettes Legislator Spot Check Issues - State: MA
+Short Description: When checking Massachuettes’s legislators, I found an extra legislator on OS and a vacancy listed on the legislator site that is not on OS
+
+Include these 3 sections as appropriate (or note none)
+
+Missing or Incorrect legislators:
+ (include vacancies, mismatched legislators, etc.)
+A vacancy on the legislative site is listed, but OS lists Robert A. DeLeo instead of the vacancy. His service ended in 2020.
+OS Link: https://openstates.org/person/robert-a-deleo-3pojeaJMWvGlYxjpYkSr9j/
+Legislative Link of Robert A. DeLeo: https://malegislature.gov/Legislators/Profile/RAD1
+Legislative Link Showing The Vacancy: https://malegislature.gov/Legislators/Members/House
+
+Photos:
+
+
+
+
+Data Issues:
+N/A
+
+Additional Data:
+N/A
+
+
+",1,massachuettes legislator spot check issues state ma short description when checking massachuettes’s legislators i found an extra legislator on os and a vacancy listed on the legislator site that is not on os include these sections as appropriate or note none missing or incorrect legislators include vacancies mismatched legislators etc a vacancy on the legislative site is listed but os lists robert a deleo instead of the vacancy his service ended in os link legislative link of robert a deleo legislative link showing the vacancy photos data issues n a additional data n a ,1
+653,11615910844.0,IssuesEvent,2020-02-26 14:55:11,PopulateTools/gobierto,https://api.github.com/repos/PopulateTools/gobierto,opened,Agendas / Events for a person should not include the events they have been invited to,gobierto-people,"Situation (applies for IBM Notes integration - we should check if in other integrations happens the same)
+
+- In the source calendar, a user creates an event and invites 3 persons. Those 3 has also users in Gobierto
+- Each user has the event imported into their own calendars, listing as invitees the other 3 users
+- In a person's public agenda, their own event is shown but also the other 3 events they have been invited to, so each of the users have the same 4 events in their own agenda
+
+Proposed solution:
+
+- When building the events for a person, ignore the events they have been invited to (just show their own events)
+- We may do this only for IBM Notes for the moment, in case in other integrations the invitees behaviour is not the same. If we detect in the future that the same problem occurs, the change would be easy",1.0,"Agendas / Events for a person should not include the events they have been invited to - Situation (applies for IBM Notes integration - we should check if in other integrations happens the same)
+
+- In the source calendar, a user creates an event and invites 3 persons. Those 3 has also users in Gobierto
+- Each user has the event imported into their own calendars, listing as invitees the other 3 users
+- In a person's public agenda, their own event is shown but also the other 3 events they have been invited to, so each of the users have the same 4 events in their own agenda
+
+Proposed solution:
+
+- When building the events for a person, ignore the events they have been invited to (just show their own events)
+- We may do this only for IBM Notes for the moment, in case in other integrations the invitees behaviour is not the same. If we detect in the future that the same problem occurs, the change would be easy",1,agendas events for a person should not include the events they have been invited to situation applies for ibm notes integration we should check if in other integrations happens the same in the source calendar a user creates an event and invites persons those has also users in gobierto each user has the event imported into their own calendars listing as invitees the other users in a person s public agenda their own event is shown but also the other events they have been invited to so each of the users have the same events in their own agenda proposed solution when building the events for a person ignore the events they have been invited to just show their own events we may do this only for ibm notes for the moment in case in other integrations the invitees behaviour is not the same if we detect in the future that the same problem occurs the change would be easy,1
+331943,24332140617.0,IssuesEvent,2022-09-30 20:29:46,Real-Dev-Squad/website-goals,https://api.github.com/repos/Real-Dev-Squad/website-goals,opened,Add description about the website in repo,beginner friendly documentation good first issue,"1. Add a brief description of what the website does in readme.
+Refer to https://github.com/Real-Dev-Squad/website-dashboard#what-does-dashboard-does
+
+2. Add the deployed link of the website",1.0,"Add description about the website in repo - 1. Add a brief description of what the website does in readme.
+Refer to https://github.com/Real-Dev-Squad/website-dashboard#what-does-dashboard-does
+
+2. Add the deployed link of the website",0,add description about the website in repo add a brief description of what the website does in readme refer to add the deployed link of the website,0
+356257,25176150708.0,IssuesEvent,2022-11-11 09:26:12,RezwanArefin01/pe,https://api.github.com/repos/RezwanArefin01/pe,opened,DG: Inconsistent formatting of code elements,severity.VeryLow type.DocumentationBug,"
+
+In many places of the document, the code elements are not rendered correctly.
+
+
+
+
+",1.0,"DG: Inconsistent formatting of code elements - 
+
+In many places of the document, the code elements are not rendered correctly.
+
+
+
+
+",0,dg inconsistent formatting of code elements in many places of the document the code elements are not rendered correctly ,0
+79723,23028741511.0,IssuesEvent,2022-07-22 11:54:51,pandas-dev/pandas,https://api.github.com/repos/pandas-dev/pandas,opened,Developement environment with docker BUILD: v1.5.0.dev0-1185-gbedd8f0170,Build Needs Triage,"### Installation check
+
+- [X] I have read the [installation guide](https://pandas.pydata.org/pandas-docs/stable/getting_started/install.html#installing-pandas).
+
+
+### Platform
+
+Linux-5.8.0-7630-generic-x86_64-with-glibc2.10
+
+### Installation Method
+
+Other
+
+### pandas Version
+
+v1.5.0.dev0-1185-gbedd8f0170
+
+### Python Version
+
+3.8.13
+
+### Installation Logs
+
+Hi,
+I am trying to create a development environment by using Docker (following these intructions : https://pandas.pydata.org/docs/dev/development/contributing_environment.html#creating-an-environment-using-docker ). However, I have tests that fail due to my environment.
+
+### Installation steps:
+- clone repo:
+```
+$ git clone git@github.com:pandas-dev/pandas.git
+cd pandas
+```
+- Build docker image:
+```
+$ docker build --no-cache --tag pandas-yourname-env .
+```
+- Launch docker image:
+```
+$ docker run -it -w /home/pandas --rm -v $(pwd):/home/pandas pandas-yourname-env
+```
+- build the C extensions
+```
+root@a371cf8a3f11:/home/pandas# python setup.py build_ext -j 4
+```
+- Launch some tests:
+```
+root@a371cf8a3f11:/home/pandas# cd pandas/tests/io
+root@a371cf8a3f11:/home/pandas# pytest test_feather.py
+```
+
+And the I get import errors:
+```python
+ ImportError: pyarrow requires pandas 0.23.0 or above, pandas 0+unknown is installed
+```
+
+The full output is the following:
+
+base) root@a371cf8a3f11:/home/pandas/pandas/tests/io# pytest test_feather.py
+=========================================================================================================== test session starts ============================================================================================================
+platform linux -- Python 3.8.13, pytest-7.1.2, pluggy-1.0.0
+rootdir: /home/pandas, configfile: pyproject.toml
+plugins: xdist-2.5.0, forked-1.4.0, asyncio-0.19.0, cov-3.0.0, cython-0.2.0, anyio-3.6.1, hypothesis-6.47.1
+asyncio: mode=strict
+collected 12 items
+
+test_feather.py .F..F.FFFFFF
+
+================================================================================================================= FAILURES =================================================================================================================
+__________________________________________________________________________________________________________ TestFeather.test_basic __________________________________________________________________________________________________________
+
+self =
+
+> ???
+
+/workspaces/pandas/pandas/tests/io/test_feather.py:93:
+_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+/workspaces/pandas/pandas/tests/io/test_feather.py:42: in check_round_trip
+ ???
+../../io/feather_format.py:93: in to_feather
+ feather.write_feather(df, handles.handle, **kwargs)
+_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+
+df = string int uint float float_with_null bool bool_with_null cat dt dttz dt_with_nul...False c 2013-01-03 2013-01-03 00:00:00-05:00 2013-01-03 2013-01-01 00:00:00.000000002 2013-03 3 days (2, 3]
+dest = <_io.BufferedWriter name='/tmp/8db00a75-b89e-4baf-b8ac-be193f0a3757'>, compression = 'lz4', compression_level = None, chunksize = None, version = 2
+
+ def write_feather(df, dest, compression=None, compression_level=None,
+ chunksize=None, version=2):
+ """"""
+ Write a pandas.DataFrame to Feather format.
+
+ Parameters
+ ----------
+ df : pandas.DataFrame or pyarrow.Table
+ Data to write out as Feather format.
+ dest : str
+ Local destination path.
+ compression : string, default None
+ Can be one of {""zstd"", ""lz4"", ""uncompressed""}. The default of None uses
+ LZ4 for V2 files if it is available, otherwise uncompressed.
+ compression_level : int, default None
+ Use a compression level particular to the chosen compressor. If None
+ use the default compression level
+ chunksize : int, default None
+ For V2 files, the internal maximum size of Arrow RecordBatch chunks
+ when writing the Arrow IPC file format. None means use the default,
+ which is currently 64K
+ version : int, default 2
+ Feather file version. Version 2 is the current. Version 1 is the more
+ limited legacy format
+ """"""
+ if _pandas_api.have_pandas:
+ _check_pandas_version()
+ if (_pandas_api.has_sparse and
+ isinstance(df, _pandas_api.pd.SparseDataFrame)):
+ df = df.to_dense()
+
+ if _pandas_api.is_data_frame(df):
+ # Feather v1 creates a new column in the resultant Table to
+ # store index information if index type is not RangeIndex
+
+ if version == 1:
+ preserve_index = False
+ elif version == 2:
+ preserve_index = None
+ else:
+ raise ValueError(""Version value should either be 1 or 2"")
+
+ table = Table.from_pandas(df, preserve_index=preserve_index)
+
+ if version == 1:
+ # Version 1 does not chunking
+ for i, name in enumerate(table.schema.names):
+ col = table[i]
+ check_chunked_overflow(name, col)
+ else:
+ table = df
+
+ if version == 1:
+ if len(table.column_names) > len(set(table.column_names)):
+ raise ValueError(""cannot serialize duplicate column names"")
+
+ if compression is not None:
+ raise ValueError(""Feather V1 files do not support compression ""
+ ""option"")
+
+ if chunksize is not None:
+ raise ValueError(""Feather V1 files do not support chunksize ""
+ ""option"")
+ else:
+ if compression is None and Codec.is_available('lz4_frame'):
+ compression = 'lz4'
+ elif (compression is not None and
+ compression not in _FEATHER_SUPPORTED_CODECS):
+ raise ValueError('compression=""{}"" not supported, must be '
+ 'one of {}'.format(compression,
+ _FEATHER_SUPPORTED_CODECS))
+
+ try:
+> _feather.write_feather(table, dest, compression=compression,
+ compression_level=compression_level,
+ chunksize=chunksize, version=version)
+E TypeError: Argument 'table' has incorrect type (expected pyarrow.lib.Table, got DataFrame)
+
+/opt/conda/lib/python3.8/site-packages/pyarrow/feather.py:195: TypeError
+______________________________________________________________________________________________________ TestFeather.test_read_columns _______________________________________________________________________________________________________
+
+self =
+
+> ???
+
+/workspaces/pandas/pandas/tests/io/test_feather.py:119:
+_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+/workspaces/pandas/pandas/tests/io/test_feather.py:42: in check_round_trip
+ ???
+../../io/feather_format.py:93: in to_feather
+ feather.write_feather(df, handles.handle, **kwargs)
+_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+
+df = col1 col2 col3 col4
+0 a 1 x 4
+1 b 2 y 5
+2 c 3 z 6, dest = <_io.BufferedWriter name='/tmp/b42159fe-4949-4333-aea4-b370abc0f2f2'>, compression = 'lz4', compression_level = None
+chunksize = None, version = 2
+
+ def write_feather(df, dest, compression=None, compression_level=None,
+ chunksize=None, version=2):
+ """"""
+ Write a pandas.DataFrame to Feather format.
+
+ Parameters
+ ----------
+ df : pandas.DataFrame or pyarrow.Table
+ Data to write out as Feather format.
+ dest : str
+ Local destination path.
+ compression : string, default None
+ Can be one of {""zstd"", ""lz4"", ""uncompressed""}. The default of None uses
+ LZ4 for V2 files if it is available, otherwise uncompressed.
+ compression_level : int, default None
+ Use a compression level particular to the chosen compressor. If None
+ use the default compression level
+ chunksize : int, default None
+ For V2 files, the internal maximum size of Arrow RecordBatch chunks
+ when writing the Arrow IPC file format. None means use the default,
+ which is currently 64K
+ version : int, default 2
+ Feather file version. Version 2 is the current. Version 1 is the more
+ limited legacy format
+ """"""
+ if _pandas_api.have_pandas:
+ _check_pandas_version()
+ if (_pandas_api.has_sparse and
+ isinstance(df, _pandas_api.pd.SparseDataFrame)):
+ df = df.to_dense()
+
+ if _pandas_api.is_data_frame(df):
+ # Feather v1 creates a new column in the resultant Table to
+ # store index information if index type is not RangeIndex
+
+ if version == 1:
+ preserve_index = False
+ elif version == 2:
+ preserve_index = None
+ else:
+ raise ValueError(""Version value should either be 1 or 2"")
+
+ table = Table.from_pandas(df, preserve_index=preserve_index)
+
+ if version == 1:
+ # Version 1 does not chunking
+ for i, name in enumerate(table.schema.names):
+ col = table[i]
+ check_chunked_overflow(name, col)
+ else:
+ table = df
+
+ if version == 1:
+ if len(table.column_names) > len(set(table.column_names)):
+ raise ValueError(""cannot serialize duplicate column names"")
+
+ if compression is not None:
+ raise ValueError(""Feather V1 files do not support compression ""
+ ""option"")
+
+ if chunksize is not None:
+ raise ValueError(""Feather V1 files do not support chunksize ""
+ ""option"")
+ else:
+ if compression is None and Codec.is_available('lz4_frame'):
+ compression = 'lz4'
+ elif (compression is not None and
+ compression not in _FEATHER_SUPPORTED_CODECS):
+ raise ValueError('compression=""{}"" not supported, must be '
+ 'one of {}'.format(compression,
+ _FEATHER_SUPPORTED_CODECS))
+
+ try:
+> _feather.write_feather(table, dest, compression=compression,
+ compression_level=compression_level,
+ chunksize=chunksize, version=version)
+E TypeError: Argument 'table' has incorrect type (expected pyarrow.lib.Table, got DataFrame)
+
+/opt/conda/lib/python3.8/site-packages/pyarrow/feather.py:195: TypeError
+_____________________________________________________________________________________________________ TestFeather.test_rw_use_threads ______________________________________________________________________________________________________
+
+self =
+
+> ???
+
+/workspaces/pandas/pandas/tests/io/test_feather.py:134:
+_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+/workspaces/pandas/pandas/tests/io/test_feather.py:42: in check_round_trip
+ ???
+../../io/feather_format.py:93: in to_feather
+ feather.write_feather(df, handles.handle, **kwargs)
+_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+
+df = A
+0 0
+1 1
+2 2
+3 3
+4 4
+... ...
+99995 99995
+99996 99996
+99997 99997
+99998 99998
+99999 99999
+
+[100000 rows x 1 columns]
+dest = <_io.BufferedWriter name='/tmp/55de5294-4a52-46c4-98ed-28e62742b883'>, compression = 'lz4', compression_level = None, chunksize = None, version = 2
+
+ def write_feather(df, dest, compression=None, compression_level=None,
+ chunksize=None, version=2):
+ """"""
+ Write a pandas.DataFrame to Feather format.
+
+ Parameters
+ ----------
+ df : pandas.DataFrame or pyarrow.Table
+ Data to write out as Feather format.
+ dest : str
+ Local destination path.
+ compression : string, default None
+ Can be one of {""zstd"", ""lz4"", ""uncompressed""}. The default of None uses
+ LZ4 for V2 files if it is available, otherwise uncompressed.
+ compression_level : int, default None
+ Use a compression level particular to the chosen compressor. If None
+ use the default compression level
+ chunksize : int, default None
+ For V2 files, the internal maximum size of Arrow RecordBatch chunks
+ when writing the Arrow IPC file format. None means use the default,
+ which is currently 64K
+ version : int, default 2
+ Feather file version. Version 2 is the current. Version 1 is the more
+ limited legacy format
+ """"""
+ if _pandas_api.have_pandas:
+ _check_pandas_version()
+ if (_pandas_api.has_sparse and
+ isinstance(df, _pandas_api.pd.SparseDataFrame)):
+ df = df.to_dense()
+
+ if _pandas_api.is_data_frame(df):
+ # Feather v1 creates a new column in the resultant Table to
+ # store index information if index type is not RangeIndex
+
+ if version == 1:
+ preserve_index = False
+ elif version == 2:
+ preserve_index = None
+ else:
+ raise ValueError(""Version value should either be 1 or 2"")
+
+ table = Table.from_pandas(df, preserve_index=preserve_index)
+
+ if version == 1:
+ # Version 1 does not chunking
+ for i, name in enumerate(table.schema.names):
+ col = table[i]
+ check_chunked_overflow(name, col)
+ else:
+ table = df
+
+ if version == 1:
+ if len(table.column_names) > len(set(table.column_names)):
+ raise ValueError(""cannot serialize duplicate column names"")
+
+ if compression is not None:
+ raise ValueError(""Feather V1 files do not support compression ""
+ ""option"")
+
+ if chunksize is not None:
+ raise ValueError(""Feather V1 files do not support chunksize ""
+ ""option"")
+ else:
+ if compression is None and Codec.is_available('lz4_frame'):
+ compression = 'lz4'
+ elif (compression is not None and
+ compression not in _FEATHER_SUPPORTED_CODECS):
+ raise ValueError('compression=""{}"" not supported, must be '
+ 'one of {}'.format(compression,
+ _FEATHER_SUPPORTED_CODECS))
+
+ try:
+> _feather.write_feather(table, dest, compression=compression,
+ compression_level=compression_level,
+ chunksize=chunksize, version=version)
+E TypeError: Argument 'table' has incorrect type (expected pyarrow.lib.Table, got DataFrame)
+
+/opt/conda/lib/python3.8/site-packages/pyarrow/feather.py:195: TypeError
+____________________________________________________________________________________________________ TestFeather.test_write_with_index _____________________________________________________________________________________________________
+
+self =
+
+> ???
+
+/workspaces/pandas/pandas/tests/io/test_feather.py:140:
+_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+/workspaces/pandas/pandas/tests/io/test_feather.py:42: in check_round_trip
+ ???
+../../io/feather_format.py:93: in to_feather
+ feather.write_feather(df, handles.handle, **kwargs)
+_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+
+df = A
+0 1
+1 2
+2 3, dest = <_io.BufferedWriter name='/tmp/d4eb66da-385f-4105-b9e1-34b7d0b9fef2'>, compression = 'lz4', compression_level = None, chunksize = None, version = 2
+
+ def write_feather(df, dest, compression=None, compression_level=None,
+ chunksize=None, version=2):
+ """"""
+ Write a pandas.DataFrame to Feather format.
+
+ Parameters
+ ----------
+ df : pandas.DataFrame or pyarrow.Table
+ Data to write out as Feather format.
+ dest : str
+ Local destination path.
+ compression : string, default None
+ Can be one of {""zstd"", ""lz4"", ""uncompressed""}. The default of None uses
+ LZ4 for V2 files if it is available, otherwise uncompressed.
+ compression_level : int, default None
+ Use a compression level particular to the chosen compressor. If None
+ use the default compression level
+ chunksize : int, default None
+ For V2 files, the internal maximum size of Arrow RecordBatch chunks
+ when writing the Arrow IPC file format. None means use the default,
+ which is currently 64K
+ version : int, default 2
+ Feather file version. Version 2 is the current. Version 1 is the more
+ limited legacy format
+ """"""
+ if _pandas_api.have_pandas:
+ _check_pandas_version()
+ if (_pandas_api.has_sparse and
+ isinstance(df, _pandas_api.pd.SparseDataFrame)):
+ df = df.to_dense()
+
+ if _pandas_api.is_data_frame(df):
+ # Feather v1 creates a new column in the resultant Table to
+ # store index information if index type is not RangeIndex
+
+ if version == 1:
+ preserve_index = False
+ elif version == 2:
+ preserve_index = None
+ else:
+ raise ValueError(""Version value should either be 1 or 2"")
+
+ table = Table.from_pandas(df, preserve_index=preserve_index)
+
+ if version == 1:
+ # Version 1 does not chunking
+ for i, name in enumerate(table.schema.names):
+ col = table[i]
+ check_chunked_overflow(name, col)
+ else:
+ table = df
+
+ if version == 1:
+ if len(table.column_names) > len(set(table.column_names)):
+ raise ValueError(""cannot serialize duplicate column names"")
+
+ if compression is not None:
+ raise ValueError(""Feather V1 files do not support compression ""
+ ""option"")
+
+ if chunksize is not None:
+ raise ValueError(""Feather V1 files do not support chunksize ""
+ ""option"")
+ else:
+ if compression is None and Codec.is_available('lz4_frame'):
+ compression = 'lz4'
+ elif (compression is not None and
+ compression not in _FEATHER_SUPPORTED_CODECS):
+ raise ValueError('compression=""{}"" not supported, must be '
+ 'one of {}'.format(compression,
+ _FEATHER_SUPPORTED_CODECS))
+
+ try:
+> _feather.write_feather(table, dest, compression=compression,
+ compression_level=compression_level,
+ chunksize=chunksize, version=version)
+E TypeError: Argument 'table' has incorrect type (expected pyarrow.lib.Table, got DataFrame)
+
+/opt/conda/lib/python3.8/site-packages/pyarrow/feather.py:195: TypeError
+______________________________________________________________________________________________________ TestFeather.test_path_pathlib _______________________________________________________________________________________________________
+
+self =
+
+> ???
+
+/workspaces/pandas/pandas/tests/io/test_feather.py:172:
+_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+../../_testing/_io.py:337: in round_trip_pathlib
+ writer(Path(path))
+../../util/_decorators.py:211: in wrapper
+ return func(*args, **kwargs)
+../../core/frame.py:2738: in to_feather
+ to_feather(self, path, **kwargs)
+../../io/feather_format.py:93: in to_feather
+ feather.write_feather(df, handles.handle, **kwargs)
+_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+
+df = index A B C D
+0 kkdUX5bh57 1.163034 -0.831564 0.177091 -0.509066
+1 Yhda...-0.800718
+28 6aJ2eCX4gj -0.155434 1.034860 1.003312 -0.292484
+29 za32uHrdC0 0.090177 0.904358 -0.082411 -0.502467
+dest = <_io.BufferedWriter name='/tmp/c0f87565-303e-4a06-8ccb-283b1b939c0f___pathlib___'>, compression = 'lz4', compression_level = None, chunksize = None, version = 2
+
+ def write_feather(df, dest, compression=None, compression_level=None,
+ chunksize=None, version=2):
+ """"""
+ Write a pandas.DataFrame to Feather format.
+
+ Parameters
+ ----------
+ df : pandas.DataFrame or pyarrow.Table
+ Data to write out as Feather format.
+ dest : str
+ Local destination path.
+ compression : string, default None
+ Can be one of {""zstd"", ""lz4"", ""uncompressed""}. The default of None uses
+ LZ4 for V2 files if it is available, otherwise uncompressed.
+ compression_level : int, default None
+ Use a compression level particular to the chosen compressor. If None
+ use the default compression level
+ chunksize : int, default None
+ For V2 files, the internal maximum size of Arrow RecordBatch chunks
+ when writing the Arrow IPC file format. None means use the default,
+ which is currently 64K
+ version : int, default 2
+ Feather file version. Version 2 is the current. Version 1 is the more
+ limited legacy format
+ """"""
+ if _pandas_api.have_pandas:
+ _check_pandas_version()
+ if (_pandas_api.has_sparse and
+ isinstance(df, _pandas_api.pd.SparseDataFrame)):
+ df = df.to_dense()
+
+ if _pandas_api.is_data_frame(df):
+ # Feather v1 creates a new column in the resultant Table to
+ # store index information if index type is not RangeIndex
+
+ if version == 1:
+ preserve_index = False
+ elif version == 2:
+ preserve_index = None
+ else:
+ raise ValueError(""Version value should either be 1 or 2"")
+
+ table = Table.from_pandas(df, preserve_index=preserve_index)
+
+ if version == 1:
+ # Version 1 does not chunking
+ for i, name in enumerate(table.schema.names):
+ col = table[i]
+ check_chunked_overflow(name, col)
+ else:
+ table = df
+
+ if version == 1:
+ if len(table.column_names) > len(set(table.column_names)):
+ raise ValueError(""cannot serialize duplicate column names"")
+
+ if compression is not None:
+ raise ValueError(""Feather V1 files do not support compression ""
+ ""option"")
+
+ if chunksize is not None:
+ raise ValueError(""Feather V1 files do not support chunksize ""
+ ""option"")
+ else:
+ if compression is None and Codec.is_available('lz4_frame'):
+ compression = 'lz4'
+ elif (compression is not None and
+ compression not in _FEATHER_SUPPORTED_CODECS):
+ raise ValueError('compression=""{}"" not supported, must be '
+ 'one of {}'.format(compression,
+ _FEATHER_SUPPORTED_CODECS))
+
+ try:
+> _feather.write_feather(table, dest, compression=compression,
+ compression_level=compression_level,
+ chunksize=chunksize, version=version)
+E TypeError: Argument 'table' has incorrect type (expected pyarrow.lib.Table, got DataFrame)
+
+/opt/conda/lib/python3.8/site-packages/pyarrow/feather.py:195: TypeError
+_____________________________________________________________________________________________________ TestFeather.test_path_localpath ______________________________________________________________________________________________________
+
+self =
+
+> ???
+
+/workspaces/pandas/pandas/tests/io/test_feather.py:177:
+_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+../../_testing/_io.py:366: in round_trip_localpath
+ writer(LocalPath(path))
+../../util/_decorators.py:211: in wrapper
+ return func(*args, **kwargs)
+../../core/frame.py:2738: in to_feather
+ to_feather(self, path, **kwargs)
+../../io/feather_format.py:93: in to_feather
+ feather.write_feather(df, handles.handle, **kwargs)
+_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+
+df = index A B C D
+0 rDYGvLSusu 0.547667 -1.406635 1.902400 -0.936238
+1 S87M...-0.578546
+28 ZUqBYwCm4s -1.216595 -1.489303 1.278688 -0.043533
+29 TIlBv32lx9 0.746072 -0.655865 -1.219951 -0.108043
+dest = <_io.BufferedWriter name='/tmp/56d5e30d-d1c9-4244-9cb5-262466b3c51c___localpath___'>, compression = 'lz4', compression_level = None, chunksize = None, version = 2
+
+ def write_feather(df, dest, compression=None, compression_level=None,
+ chunksize=None, version=2):
+ """"""
+ Write a pandas.DataFrame to Feather format.
+
+ Parameters
+ ----------
+ df : pandas.DataFrame or pyarrow.Table
+ Data to write out as Feather format.
+ dest : str
+ Local destination path.
+ compression : string, default None
+ Can be one of {""zstd"", ""lz4"", ""uncompressed""}. The default of None uses
+ LZ4 for V2 files if it is available, otherwise uncompressed.
+ compression_level : int, default None
+ Use a compression level particular to the chosen compressor. If None
+ use the default compression level
+ chunksize : int, default None
+ For V2 files, the internal maximum size of Arrow RecordBatch chunks
+ when writing the Arrow IPC file format. None means use the default,
+ which is currently 64K
+ version : int, default 2
+ Feather file version. Version 2 is the current. Version 1 is the more
+ limited legacy format
+ """"""
+ if _pandas_api.have_pandas:
+ _check_pandas_version()
+ if (_pandas_api.has_sparse and
+ isinstance(df, _pandas_api.pd.SparseDataFrame)):
+ df = df.to_dense()
+
+ if _pandas_api.is_data_frame(df):
+ # Feather v1 creates a new column in the resultant Table to
+ # store index information if index type is not RangeIndex
+
+ if version == 1:
+ preserve_index = False
+ elif version == 2:
+ preserve_index = None
+ else:
+ raise ValueError(""Version value should either be 1 or 2"")
+
+ table = Table.from_pandas(df, preserve_index=preserve_index)
+
+ if version == 1:
+ # Version 1 does not chunking
+ for i, name in enumerate(table.schema.names):
+ col = table[i]
+ check_chunked_overflow(name, col)
+ else:
+ table = df
+
+ if version == 1:
+ if len(table.column_names) > len(set(table.column_names)):
+ raise ValueError(""cannot serialize duplicate column names"")
+
+ if compression is not None:
+ raise ValueError(""Feather V1 files do not support compression ""
+ ""option"")
+
+ if chunksize is not None:
+ raise ValueError(""Feather V1 files do not support chunksize ""
+ ""option"")
+ else:
+ if compression is None and Codec.is_available('lz4_frame'):
+ compression = 'lz4'
+ elif (compression is not None and
+ compression not in _FEATHER_SUPPORTED_CODECS):
+ raise ValueError('compression=""{}"" not supported, must be '
+ 'one of {}'.format(compression,
+ _FEATHER_SUPPORTED_CODECS))
+
+ try:
+> _feather.write_feather(table, dest, compression=compression,
+ compression_level=compression_level,
+ chunksize=chunksize, version=version)
+E TypeError: Argument 'table' has incorrect type (expected pyarrow.lib.Table, got DataFrame)
+
+/opt/conda/lib/python3.8/site-packages/pyarrow/feather.py:195: TypeError
+__________________________________________________________________________________________________ TestFeather.test_passthrough_keywords ___________________________________________________________________________________________________
+
+self =
+
+> ???
+
+/workspaces/pandas/pandas/tests/io/test_feather.py:182:
+_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+/workspaces/pandas/pandas/tests/io/test_feather.py:42: in check_round_trip
+ ???
+../../io/feather_format.py:93: in to_feather
+ feather.write_feather(df, handles.handle, **kwargs)
+/opt/conda/lib/python3.8/site-packages/pyarrow/feather.py:175: in write_feather
+ if len(table.column_names) > len(set(table.column_names)):
+_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+
+self = index A B C D
+0 BxLds1zytg -1.452649 -1.637048 -0.086740 -0.010354
+1 2Bg6... 0.058544
+28 XJsorYKGCU -0.052787 1.624943 -0.367892 -1.531762
+29 dMh4CIFDYz -0.250029 -0.510402 -0.910837 -1.472121
+name = 'column_names'
+
+ def __getattr__(self, name: str):
+ """"""
+ After regular attribute access, try looking up the name
+ This allows simpler access to columns for interactive use.
+ """"""
+ # Note: obj.x will always call obj.__getattribute__('x') prior to
+ # calling obj.__getattr__('x').
+ if (
+ name not in self._internal_names_set
+ and name not in self._metadata
+ and name not in self._accessors
+ and self._info_axis._can_hold_identifiers_and_holds_name(name)
+ ):
+ return self[name]
+> return object.__getattribute__(self, name)
+E AttributeError: 'DataFrame' object has no attribute 'column_names'
+
+../../core/generic.py:5758: AttributeError
+________________________________________________________________________________________________________ TestFeather.test_http_path ________________________________________________________________________________________________________
+
+self = , feather_file = '/home/pandas/pandas/tests/io/data/feather/feather-0_3_1.feather'
+
+> ???
+
+/workspaces/pandas/pandas/tests/io/test_feather.py:198:
+_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+../../io/feather_format.py:132: in read_feather
+ return feather.read_feather(
+/opt/conda/lib/python3.8/site-packages/pyarrow/feather.py:229: in read_feather
+ _check_pandas_version()
+/opt/conda/lib/python3.8/site-packages/pyarrow/feather.py:31: in _check_pandas_version
+ if _pandas_api.loose_version < Version('0.17.0'):
+pyarrow/pandas-shim.pxi:155: in pyarrow.lib._PandasAPIShim.loose_version.__get__
+ ???
+pyarrow/pandas-shim.pxi:106: in pyarrow.lib._PandasAPIShim._check_import
+ ???
+_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+
+> ???
+E ImportError: pyarrow requires pandas 0.23.0 or above, pandas 0+unknown is installed
+
+pyarrow/pandas-shim.pxi:65: ImportError
+============================================================================================================= warnings summary =============================================================================================================
+../../core/construction.py:332
+ /home/pandas/pandas/core/construction.py:332: UserWarning: pyarrow requires pandas 0.23.0 or above, pandas 0+unknown is installed. Therefore, pandas-specific integration is not used.
+ return cls._from_sequence(data, dtype=dtype, copy=copy)
+
+-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html
+-------------------------------------------------------------------------------------- generated xml file: /home/pandas/pandas/tests/io/test-data.xml --------------------------------------------------------------------------------------
+=========================================================================================================== slowest 30 durations ===========================================================================================================
+0.24s call pandas/tests/io/test_feather.py::TestFeather::test_http_path
+0.01s call pandas/tests/io/test_feather.py::TestFeather::test_error
+
+(28 durations < 0.005s hidden. Use -vv to show these durations.)
+========================================================================================================= short test summary info ==========================================================================================================
+FAILED test_feather.py::TestFeather::test_basic - TypeError: Argument 'table' has incorrect type (expected pyarrow.lib.Table, got DataFrame)
+FAILED test_feather.py::TestFeather::test_read_columns - TypeError: Argument 'table' has incorrect type (expected pyarrow.lib.Table, got DataFrame)
+FAILED test_feather.py::TestFeather::test_rw_use_threads - TypeError: Argument 'table' has incorrect type (expected pyarrow.lib.Table, got DataFrame)
+FAILED test_feather.py::TestFeather::test_write_with_index - TypeError: Argument 'table' has incorrect type (expected pyarrow.lib.Table, got DataFrame)
+FAILED test_feather.py::TestFeather::test_path_pathlib - TypeError: Argument 'table' has incorrect type (expected pyarrow.lib.Table, got DataFrame)
+FAILED test_feather.py::TestFeather::test_path_localpath - TypeError: Argument 'table' has incorrect type (expected pyarrow.lib.Table, got DataFrame)
+FAILED test_feather.py::TestFeather::test_passthrough_keywords - AttributeError: 'DataFrame' object has no attribute 'column_names'
+FAILED test_feather.py::TestFeather::test_http_path - ImportError: pyarrow requires pandas 0.23.0 or above, pandas 0+unknown is installed
+================================================================================================== 8 failed, 4 passed, 1 warning in 0.74s ==================================================================================================
+(base) root@a371cf8a3f11:/home/pandas/pandas/tests/io#
+
+
+
+What am I doing wrong?
+",1.0,"Developement environment with docker BUILD: v1.5.0.dev0-1185-gbedd8f0170 - ### Installation check
+
+- [X] I have read the [installation guide](https://pandas.pydata.org/pandas-docs/stable/getting_started/install.html#installing-pandas).
+
+
+### Platform
+
+Linux-5.8.0-7630-generic-x86_64-with-glibc2.10
+
+### Installation Method
+
+Other
+
+### pandas Version
+
+v1.5.0.dev0-1185-gbedd8f0170
+
+### Python Version
+
+3.8.13
+
+### Installation Logs
+
+Hi,
+I am trying to create a development environment by using Docker (following these intructions : https://pandas.pydata.org/docs/dev/development/contributing_environment.html#creating-an-environment-using-docker ). However, I have tests that fail due to my environment.
+
+### Installation steps:
+- clone repo:
+```
+$ git clone git@github.com:pandas-dev/pandas.git
+cd pandas
+```
+- Build docker image:
+```
+$ docker build --no-cache --tag pandas-yourname-env .
+```
+- Launch docker image:
+```
+$ docker run -it -w /home/pandas --rm -v $(pwd):/home/pandas pandas-yourname-env
+```
+- build the C extensions
+```
+root@a371cf8a3f11:/home/pandas# python setup.py build_ext -j 4
+```
+- Launch some tests:
+```
+root@a371cf8a3f11:/home/pandas# cd pandas/tests/io
+root@a371cf8a3f11:/home/pandas# pytest test_feather.py
+```
+
+And the I get import errors:
+```python
+ ImportError: pyarrow requires pandas 0.23.0 or above, pandas 0+unknown is installed
+```
+
+The full output is the following:
+
+base) root@a371cf8a3f11:/home/pandas/pandas/tests/io# pytest test_feather.py
+=========================================================================================================== test session starts ============================================================================================================
+platform linux -- Python 3.8.13, pytest-7.1.2, pluggy-1.0.0
+rootdir: /home/pandas, configfile: pyproject.toml
+plugins: xdist-2.5.0, forked-1.4.0, asyncio-0.19.0, cov-3.0.0, cython-0.2.0, anyio-3.6.1, hypothesis-6.47.1
+asyncio: mode=strict
+collected 12 items
+
+test_feather.py .F..F.FFFFFF
+
+================================================================================================================= FAILURES =================================================================================================================
+__________________________________________________________________________________________________________ TestFeather.test_basic __________________________________________________________________________________________________________
+
+self =
+
+> ???
+
+/workspaces/pandas/pandas/tests/io/test_feather.py:93:
+_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+/workspaces/pandas/pandas/tests/io/test_feather.py:42: in check_round_trip
+ ???
+../../io/feather_format.py:93: in to_feather
+ feather.write_feather(df, handles.handle, **kwargs)
+_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+
+df = string int uint float float_with_null bool bool_with_null cat dt dttz dt_with_nul...False c 2013-01-03 2013-01-03 00:00:00-05:00 2013-01-03 2013-01-01 00:00:00.000000002 2013-03 3 days (2, 3]
+dest = <_io.BufferedWriter name='/tmp/8db00a75-b89e-4baf-b8ac-be193f0a3757'>, compression = 'lz4', compression_level = None, chunksize = None, version = 2
+
+ def write_feather(df, dest, compression=None, compression_level=None,
+ chunksize=None, version=2):
+ """"""
+ Write a pandas.DataFrame to Feather format.
+
+ Parameters
+ ----------
+ df : pandas.DataFrame or pyarrow.Table
+ Data to write out as Feather format.
+ dest : str
+ Local destination path.
+ compression : string, default None
+ Can be one of {""zstd"", ""lz4"", ""uncompressed""}. The default of None uses
+ LZ4 for V2 files if it is available, otherwise uncompressed.
+ compression_level : int, default None
+ Use a compression level particular to the chosen compressor. If None
+ use the default compression level
+ chunksize : int, default None
+ For V2 files, the internal maximum size of Arrow RecordBatch chunks
+ when writing the Arrow IPC file format. None means use the default,
+ which is currently 64K
+ version : int, default 2
+ Feather file version. Version 2 is the current. Version 1 is the more
+ limited legacy format
+ """"""
+ if _pandas_api.have_pandas:
+ _check_pandas_version()
+ if (_pandas_api.has_sparse and
+ isinstance(df, _pandas_api.pd.SparseDataFrame)):
+ df = df.to_dense()
+
+ if _pandas_api.is_data_frame(df):
+ # Feather v1 creates a new column in the resultant Table to
+ # store index information if index type is not RangeIndex
+
+ if version == 1:
+ preserve_index = False
+ elif version == 2:
+ preserve_index = None
+ else:
+ raise ValueError(""Version value should either be 1 or 2"")
+
+ table = Table.from_pandas(df, preserve_index=preserve_index)
+
+ if version == 1:
+ # Version 1 does not chunking
+ for i, name in enumerate(table.schema.names):
+ col = table[i]
+ check_chunked_overflow(name, col)
+ else:
+ table = df
+
+ if version == 1:
+ if len(table.column_names) > len(set(table.column_names)):
+ raise ValueError(""cannot serialize duplicate column names"")
+
+ if compression is not None:
+ raise ValueError(""Feather V1 files do not support compression ""
+ ""option"")
+
+ if chunksize is not None:
+ raise ValueError(""Feather V1 files do not support chunksize ""
+ ""option"")
+ else:
+ if compression is None and Codec.is_available('lz4_frame'):
+ compression = 'lz4'
+ elif (compression is not None and
+ compression not in _FEATHER_SUPPORTED_CODECS):
+ raise ValueError('compression=""{}"" not supported, must be '
+ 'one of {}'.format(compression,
+ _FEATHER_SUPPORTED_CODECS))
+
+ try:
+> _feather.write_feather(table, dest, compression=compression,
+ compression_level=compression_level,
+ chunksize=chunksize, version=version)
+E TypeError: Argument 'table' has incorrect type (expected pyarrow.lib.Table, got DataFrame)
+
+/opt/conda/lib/python3.8/site-packages/pyarrow/feather.py:195: TypeError
+______________________________________________________________________________________________________ TestFeather.test_read_columns _______________________________________________________________________________________________________
+
+self =
+
+> ???
+
+/workspaces/pandas/pandas/tests/io/test_feather.py:119:
+_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+/workspaces/pandas/pandas/tests/io/test_feather.py:42: in check_round_trip
+ ???
+../../io/feather_format.py:93: in to_feather
+ feather.write_feather(df, handles.handle, **kwargs)
+_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+
+df = col1 col2 col3 col4
+0 a 1 x 4
+1 b 2 y 5
+2 c 3 z 6, dest = <_io.BufferedWriter name='/tmp/b42159fe-4949-4333-aea4-b370abc0f2f2'>, compression = 'lz4', compression_level = None
+chunksize = None, version = 2
+
+ def write_feather(df, dest, compression=None, compression_level=None,
+ chunksize=None, version=2):
+ """"""
+ Write a pandas.DataFrame to Feather format.
+
+ Parameters
+ ----------
+ df : pandas.DataFrame or pyarrow.Table
+ Data to write out as Feather format.
+ dest : str
+ Local destination path.
+ compression : string, default None
+ Can be one of {""zstd"", ""lz4"", ""uncompressed""}. The default of None uses
+ LZ4 for V2 files if it is available, otherwise uncompressed.
+ compression_level : int, default None
+ Use a compression level particular to the chosen compressor. If None
+ use the default compression level
+ chunksize : int, default None
+ For V2 files, the internal maximum size of Arrow RecordBatch chunks
+ when writing the Arrow IPC file format. None means use the default,
+ which is currently 64K
+ version : int, default 2
+ Feather file version. Version 2 is the current. Version 1 is the more
+ limited legacy format
+ """"""
+ if _pandas_api.have_pandas:
+ _check_pandas_version()
+ if (_pandas_api.has_sparse and
+ isinstance(df, _pandas_api.pd.SparseDataFrame)):
+ df = df.to_dense()
+
+ if _pandas_api.is_data_frame(df):
+ # Feather v1 creates a new column in the resultant Table to
+ # store index information if index type is not RangeIndex
+
+ if version == 1:
+ preserve_index = False
+ elif version == 2:
+ preserve_index = None
+ else:
+ raise ValueError(""Version value should either be 1 or 2"")
+
+ table = Table.from_pandas(df, preserve_index=preserve_index)
+
+ if version == 1:
+ # Version 1 does not chunking
+ for i, name in enumerate(table.schema.names):
+ col = table[i]
+ check_chunked_overflow(name, col)
+ else:
+ table = df
+
+ if version == 1:
+ if len(table.column_names) > len(set(table.column_names)):
+ raise ValueError(""cannot serialize duplicate column names"")
+
+ if compression is not None:
+ raise ValueError(""Feather V1 files do not support compression ""
+ ""option"")
+
+ if chunksize is not None:
+ raise ValueError(""Feather V1 files do not support chunksize ""
+ ""option"")
+ else:
+ if compression is None and Codec.is_available('lz4_frame'):
+ compression = 'lz4'
+ elif (compression is not None and
+ compression not in _FEATHER_SUPPORTED_CODECS):
+ raise ValueError('compression=""{}"" not supported, must be '
+ 'one of {}'.format(compression,
+ _FEATHER_SUPPORTED_CODECS))
+
+ try:
+> _feather.write_feather(table, dest, compression=compression,
+ compression_level=compression_level,
+ chunksize=chunksize, version=version)
+E TypeError: Argument 'table' has incorrect type (expected pyarrow.lib.Table, got DataFrame)
+
+/opt/conda/lib/python3.8/site-packages/pyarrow/feather.py:195: TypeError
+_____________________________________________________________________________________________________ TestFeather.test_rw_use_threads ______________________________________________________________________________________________________
+
+self =
+
+> ???
+
+/workspaces/pandas/pandas/tests/io/test_feather.py:134:
+_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+/workspaces/pandas/pandas/tests/io/test_feather.py:42: in check_round_trip
+ ???
+../../io/feather_format.py:93: in to_feather
+ feather.write_feather(df, handles.handle, **kwargs)
+_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+
+df = A
+0 0
+1 1
+2 2
+3 3
+4 4
+... ...
+99995 99995
+99996 99996
+99997 99997
+99998 99998
+99999 99999
+
+[100000 rows x 1 columns]
+dest = <_io.BufferedWriter name='/tmp/55de5294-4a52-46c4-98ed-28e62742b883'>, compression = 'lz4', compression_level = None, chunksize = None, version = 2
+
+ def write_feather(df, dest, compression=None, compression_level=None,
+ chunksize=None, version=2):
+ """"""
+ Write a pandas.DataFrame to Feather format.
+
+ Parameters
+ ----------
+ df : pandas.DataFrame or pyarrow.Table
+ Data to write out as Feather format.
+ dest : str
+ Local destination path.
+ compression : string, default None
+ Can be one of {""zstd"", ""lz4"", ""uncompressed""}. The default of None uses
+ LZ4 for V2 files if it is available, otherwise uncompressed.
+ compression_level : int, default None
+ Use a compression level particular to the chosen compressor. If None
+ use the default compression level
+ chunksize : int, default None
+ For V2 files, the internal maximum size of Arrow RecordBatch chunks
+ when writing the Arrow IPC file format. None means use the default,
+ which is currently 64K
+ version : int, default 2
+ Feather file version. Version 2 is the current. Version 1 is the more
+ limited legacy format
+ """"""
+ if _pandas_api.have_pandas:
+ _check_pandas_version()
+ if (_pandas_api.has_sparse and
+ isinstance(df, _pandas_api.pd.SparseDataFrame)):
+ df = df.to_dense()
+
+ if _pandas_api.is_data_frame(df):
+ # Feather v1 creates a new column in the resultant Table to
+ # store index information if index type is not RangeIndex
+
+ if version == 1:
+ preserve_index = False
+ elif version == 2:
+ preserve_index = None
+ else:
+ raise ValueError(""Version value should either be 1 or 2"")
+
+ table = Table.from_pandas(df, preserve_index=preserve_index)
+
+ if version == 1:
+ # Version 1 does not chunking
+ for i, name in enumerate(table.schema.names):
+ col = table[i]
+ check_chunked_overflow(name, col)
+ else:
+ table = df
+
+ if version == 1:
+ if len(table.column_names) > len(set(table.column_names)):
+ raise ValueError(""cannot serialize duplicate column names"")
+
+ if compression is not None:
+ raise ValueError(""Feather V1 files do not support compression ""
+ ""option"")
+
+ if chunksize is not None:
+ raise ValueError(""Feather V1 files do not support chunksize ""
+ ""option"")
+ else:
+ if compression is None and Codec.is_available('lz4_frame'):
+ compression = 'lz4'
+ elif (compression is not None and
+ compression not in _FEATHER_SUPPORTED_CODECS):
+ raise ValueError('compression=""{}"" not supported, must be '
+ 'one of {}'.format(compression,
+ _FEATHER_SUPPORTED_CODECS))
+
+ try:
+> _feather.write_feather(table, dest, compression=compression,
+ compression_level=compression_level,
+ chunksize=chunksize, version=version)
+E TypeError: Argument 'table' has incorrect type (expected pyarrow.lib.Table, got DataFrame)
+
+/opt/conda/lib/python3.8/site-packages/pyarrow/feather.py:195: TypeError
+____________________________________________________________________________________________________ TestFeather.test_write_with_index _____________________________________________________________________________________________________
+
+self =
+
+> ???
+
+/workspaces/pandas/pandas/tests/io/test_feather.py:140:
+_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+/workspaces/pandas/pandas/tests/io/test_feather.py:42: in check_round_trip
+ ???
+../../io/feather_format.py:93: in to_feather
+ feather.write_feather(df, handles.handle, **kwargs)
+_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+
+df = A
+0 1
+1 2
+2 3, dest = <_io.BufferedWriter name='/tmp/d4eb66da-385f-4105-b9e1-34b7d0b9fef2'>, compression = 'lz4', compression_level = None, chunksize = None, version = 2
+
+ def write_feather(df, dest, compression=None, compression_level=None,
+ chunksize=None, version=2):
+ """"""
+ Write a pandas.DataFrame to Feather format.
+
+ Parameters
+ ----------
+ df : pandas.DataFrame or pyarrow.Table
+ Data to write out as Feather format.
+ dest : str
+ Local destination path.
+ compression : string, default None
+ Can be one of {""zstd"", ""lz4"", ""uncompressed""}. The default of None uses
+ LZ4 for V2 files if it is available, otherwise uncompressed.
+ compression_level : int, default None
+ Use a compression level particular to the chosen compressor. If None
+ use the default compression level
+ chunksize : int, default None
+ For V2 files, the internal maximum size of Arrow RecordBatch chunks
+ when writing the Arrow IPC file format. None means use the default,
+ which is currently 64K
+ version : int, default 2
+ Feather file version. Version 2 is the current. Version 1 is the more
+ limited legacy format
+ """"""
+ if _pandas_api.have_pandas:
+ _check_pandas_version()
+ if (_pandas_api.has_sparse and
+ isinstance(df, _pandas_api.pd.SparseDataFrame)):
+ df = df.to_dense()
+
+ if _pandas_api.is_data_frame(df):
+ # Feather v1 creates a new column in the resultant Table to
+ # store index information if index type is not RangeIndex
+
+ if version == 1:
+ preserve_index = False
+ elif version == 2:
+ preserve_index = None
+ else:
+ raise ValueError(""Version value should either be 1 or 2"")
+
+ table = Table.from_pandas(df, preserve_index=preserve_index)
+
+ if version == 1:
+ # Version 1 does not chunking
+ for i, name in enumerate(table.schema.names):
+ col = table[i]
+ check_chunked_overflow(name, col)
+ else:
+ table = df
+
+ if version == 1:
+ if len(table.column_names) > len(set(table.column_names)):
+ raise ValueError(""cannot serialize duplicate column names"")
+
+ if compression is not None:
+ raise ValueError(""Feather V1 files do not support compression ""
+ ""option"")
+
+ if chunksize is not None:
+ raise ValueError(""Feather V1 files do not support chunksize ""
+ ""option"")
+ else:
+ if compression is None and Codec.is_available('lz4_frame'):
+ compression = 'lz4'
+ elif (compression is not None and
+ compression not in _FEATHER_SUPPORTED_CODECS):
+ raise ValueError('compression=""{}"" not supported, must be '
+ 'one of {}'.format(compression,
+ _FEATHER_SUPPORTED_CODECS))
+
+ try:
+> _feather.write_feather(table, dest, compression=compression,
+ compression_level=compression_level,
+ chunksize=chunksize, version=version)
+E TypeError: Argument 'table' has incorrect type (expected pyarrow.lib.Table, got DataFrame)
+
+/opt/conda/lib/python3.8/site-packages/pyarrow/feather.py:195: TypeError
+______________________________________________________________________________________________________ TestFeather.test_path_pathlib _______________________________________________________________________________________________________
+
+self =
+
+> ???
+
+/workspaces/pandas/pandas/tests/io/test_feather.py:172:
+_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+../../_testing/_io.py:337: in round_trip_pathlib
+ writer(Path(path))
+../../util/_decorators.py:211: in wrapper
+ return func(*args, **kwargs)
+../../core/frame.py:2738: in to_feather
+ to_feather(self, path, **kwargs)
+../../io/feather_format.py:93: in to_feather
+ feather.write_feather(df, handles.handle, **kwargs)
+_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+
+df = index A B C D
+0 kkdUX5bh57 1.163034 -0.831564 0.177091 -0.509066
+1 Yhda...-0.800718
+28 6aJ2eCX4gj -0.155434 1.034860 1.003312 -0.292484
+29 za32uHrdC0 0.090177 0.904358 -0.082411 -0.502467
+dest = <_io.BufferedWriter name='/tmp/c0f87565-303e-4a06-8ccb-283b1b939c0f___pathlib___'>, compression = 'lz4', compression_level = None, chunksize = None, version = 2
+
+ def write_feather(df, dest, compression=None, compression_level=None,
+ chunksize=None, version=2):
+ """"""
+ Write a pandas.DataFrame to Feather format.
+
+ Parameters
+ ----------
+ df : pandas.DataFrame or pyarrow.Table
+ Data to write out as Feather format.
+ dest : str
+ Local destination path.
+ compression : string, default None
+ Can be one of {""zstd"", ""lz4"", ""uncompressed""}. The default of None uses
+ LZ4 for V2 files if it is available, otherwise uncompressed.
+ compression_level : int, default None
+ Use a compression level particular to the chosen compressor. If None
+ use the default compression level
+ chunksize : int, default None
+ For V2 files, the internal maximum size of Arrow RecordBatch chunks
+ when writing the Arrow IPC file format. None means use the default,
+ which is currently 64K
+ version : int, default 2
+ Feather file version. Version 2 is the current. Version 1 is the more
+ limited legacy format
+ """"""
+ if _pandas_api.have_pandas:
+ _check_pandas_version()
+ if (_pandas_api.has_sparse and
+ isinstance(df, _pandas_api.pd.SparseDataFrame)):
+ df = df.to_dense()
+
+ if _pandas_api.is_data_frame(df):
+ # Feather v1 creates a new column in the resultant Table to
+ # store index information if index type is not RangeIndex
+
+ if version == 1:
+ preserve_index = False
+ elif version == 2:
+ preserve_index = None
+ else:
+ raise ValueError(""Version value should either be 1 or 2"")
+
+ table = Table.from_pandas(df, preserve_index=preserve_index)
+
+ if version == 1:
+ # Version 1 does not chunking
+ for i, name in enumerate(table.schema.names):
+ col = table[i]
+ check_chunked_overflow(name, col)
+ else:
+ table = df
+
+ if version == 1:
+ if len(table.column_names) > len(set(table.column_names)):
+ raise ValueError(""cannot serialize duplicate column names"")
+
+ if compression is not None:
+ raise ValueError(""Feather V1 files do not support compression ""
+ ""option"")
+
+ if chunksize is not None:
+ raise ValueError(""Feather V1 files do not support chunksize ""
+ ""option"")
+ else:
+ if compression is None and Codec.is_available('lz4_frame'):
+ compression = 'lz4'
+ elif (compression is not None and
+ compression not in _FEATHER_SUPPORTED_CODECS):
+ raise ValueError('compression=""{}"" not supported, must be '
+ 'one of {}'.format(compression,
+ _FEATHER_SUPPORTED_CODECS))
+
+ try:
+> _feather.write_feather(table, dest, compression=compression,
+ compression_level=compression_level,
+ chunksize=chunksize, version=version)
+E TypeError: Argument 'table' has incorrect type (expected pyarrow.lib.Table, got DataFrame)
+
+/opt/conda/lib/python3.8/site-packages/pyarrow/feather.py:195: TypeError
+_____________________________________________________________________________________________________ TestFeather.test_path_localpath ______________________________________________________________________________________________________
+
+self =
+
+> ???
+
+/workspaces/pandas/pandas/tests/io/test_feather.py:177:
+_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+../../_testing/_io.py:366: in round_trip_localpath
+ writer(LocalPath(path))
+../../util/_decorators.py:211: in wrapper
+ return func(*args, **kwargs)
+../../core/frame.py:2738: in to_feather
+ to_feather(self, path, **kwargs)
+../../io/feather_format.py:93: in to_feather
+ feather.write_feather(df, handles.handle, **kwargs)
+_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+
+df = index A B C D
+0 rDYGvLSusu 0.547667 -1.406635 1.902400 -0.936238
+1 S87M...-0.578546
+28 ZUqBYwCm4s -1.216595 -1.489303 1.278688 -0.043533
+29 TIlBv32lx9 0.746072 -0.655865 -1.219951 -0.108043
+dest = <_io.BufferedWriter name='/tmp/56d5e30d-d1c9-4244-9cb5-262466b3c51c___localpath___'>, compression = 'lz4', compression_level = None, chunksize = None, version = 2
+
+ def write_feather(df, dest, compression=None, compression_level=None,
+ chunksize=None, version=2):
+ """"""
+ Write a pandas.DataFrame to Feather format.
+
+ Parameters
+ ----------
+ df : pandas.DataFrame or pyarrow.Table
+ Data to write out as Feather format.
+ dest : str
+ Local destination path.
+ compression : string, default None
+ Can be one of {""zstd"", ""lz4"", ""uncompressed""}. The default of None uses
+ LZ4 for V2 files if it is available, otherwise uncompressed.
+ compression_level : int, default None
+ Use a compression level particular to the chosen compressor. If None
+ use the default compression level
+ chunksize : int, default None
+ For V2 files, the internal maximum size of Arrow RecordBatch chunks
+ when writing the Arrow IPC file format. None means use the default,
+ which is currently 64K
+ version : int, default 2
+ Feather file version. Version 2 is the current. Version 1 is the more
+ limited legacy format
+ """"""
+ if _pandas_api.have_pandas:
+ _check_pandas_version()
+ if (_pandas_api.has_sparse and
+ isinstance(df, _pandas_api.pd.SparseDataFrame)):
+ df = df.to_dense()
+
+ if _pandas_api.is_data_frame(df):
+ # Feather v1 creates a new column in the resultant Table to
+ # store index information if index type is not RangeIndex
+
+ if version == 1:
+ preserve_index = False
+ elif version == 2:
+ preserve_index = None
+ else:
+ raise ValueError(""Version value should either be 1 or 2"")
+
+ table = Table.from_pandas(df, preserve_index=preserve_index)
+
+ if version == 1:
+ # Version 1 does not chunking
+ for i, name in enumerate(table.schema.names):
+ col = table[i]
+ check_chunked_overflow(name, col)
+ else:
+ table = df
+
+ if version == 1:
+ if len(table.column_names) > len(set(table.column_names)):
+ raise ValueError(""cannot serialize duplicate column names"")
+
+ if compression is not None:
+ raise ValueError(""Feather V1 files do not support compression ""
+ ""option"")
+
+ if chunksize is not None:
+ raise ValueError(""Feather V1 files do not support chunksize ""
+ ""option"")
+ else:
+ if compression is None and Codec.is_available('lz4_frame'):
+ compression = 'lz4'
+ elif (compression is not None and
+ compression not in _FEATHER_SUPPORTED_CODECS):
+ raise ValueError('compression=""{}"" not supported, must be '
+ 'one of {}'.format(compression,
+ _FEATHER_SUPPORTED_CODECS))
+
+ try:
+> _feather.write_feather(table, dest, compression=compression,
+ compression_level=compression_level,
+ chunksize=chunksize, version=version)
+E TypeError: Argument 'table' has incorrect type (expected pyarrow.lib.Table, got DataFrame)
+
+/opt/conda/lib/python3.8/site-packages/pyarrow/feather.py:195: TypeError
+__________________________________________________________________________________________________ TestFeather.test_passthrough_keywords ___________________________________________________________________________________________________
+
+self =
+
+> ???
+
+/workspaces/pandas/pandas/tests/io/test_feather.py:182:
+_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+/workspaces/pandas/pandas/tests/io/test_feather.py:42: in check_round_trip
+ ???
+../../io/feather_format.py:93: in to_feather
+ feather.write_feather(df, handles.handle, **kwargs)
+/opt/conda/lib/python3.8/site-packages/pyarrow/feather.py:175: in write_feather
+ if len(table.column_names) > len(set(table.column_names)):
+_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+
+self = index A B C D
+0 BxLds1zytg -1.452649 -1.637048 -0.086740 -0.010354
+1 2Bg6... 0.058544
+28 XJsorYKGCU -0.052787 1.624943 -0.367892 -1.531762
+29 dMh4CIFDYz -0.250029 -0.510402 -0.910837 -1.472121
+name = 'column_names'
+
+ def __getattr__(self, name: str):
+ """"""
+ After regular attribute access, try looking up the name
+ This allows simpler access to columns for interactive use.
+ """"""
+ # Note: obj.x will always call obj.__getattribute__('x') prior to
+ # calling obj.__getattr__('x').
+ if (
+ name not in self._internal_names_set
+ and name not in self._metadata
+ and name not in self._accessors
+ and self._info_axis._can_hold_identifiers_and_holds_name(name)
+ ):
+ return self[name]
+> return object.__getattribute__(self, name)
+E AttributeError: 'DataFrame' object has no attribute 'column_names'
+
+../../core/generic.py:5758: AttributeError
+________________________________________________________________________________________________________ TestFeather.test_http_path ________________________________________________________________________________________________________
+
+self = , feather_file = '/home/pandas/pandas/tests/io/data/feather/feather-0_3_1.feather'
+
+> ???
+
+/workspaces/pandas/pandas/tests/io/test_feather.py:198:
+_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+../../io/feather_format.py:132: in read_feather
+ return feather.read_feather(
+/opt/conda/lib/python3.8/site-packages/pyarrow/feather.py:229: in read_feather
+ _check_pandas_version()
+/opt/conda/lib/python3.8/site-packages/pyarrow/feather.py:31: in _check_pandas_version
+ if _pandas_api.loose_version < Version('0.17.0'):
+pyarrow/pandas-shim.pxi:155: in pyarrow.lib._PandasAPIShim.loose_version.__get__
+ ???
+pyarrow/pandas-shim.pxi:106: in pyarrow.lib._PandasAPIShim._check_import
+ ???
+_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+
+> ???
+E ImportError: pyarrow requires pandas 0.23.0 or above, pandas 0+unknown is installed
+
+pyarrow/pandas-shim.pxi:65: ImportError
+============================================================================================================= warnings summary =============================================================================================================
+../../core/construction.py:332
+ /home/pandas/pandas/core/construction.py:332: UserWarning: pyarrow requires pandas 0.23.0 or above, pandas 0+unknown is installed. Therefore, pandas-specific integration is not used.
+ return cls._from_sequence(data, dtype=dtype, copy=copy)
+
+-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html
+-------------------------------------------------------------------------------------- generated xml file: /home/pandas/pandas/tests/io/test-data.xml --------------------------------------------------------------------------------------
+=========================================================================================================== slowest 30 durations ===========================================================================================================
+0.24s call pandas/tests/io/test_feather.py::TestFeather::test_http_path
+0.01s call pandas/tests/io/test_feather.py::TestFeather::test_error
+
+(28 durations < 0.005s hidden. Use -vv to show these durations.)
+========================================================================================================= short test summary info ==========================================================================================================
+FAILED test_feather.py::TestFeather::test_basic - TypeError: Argument 'table' has incorrect type (expected pyarrow.lib.Table, got DataFrame)
+FAILED test_feather.py::TestFeather::test_read_columns - TypeError: Argument 'table' has incorrect type (expected pyarrow.lib.Table, got DataFrame)
+FAILED test_feather.py::TestFeather::test_rw_use_threads - TypeError: Argument 'table' has incorrect type (expected pyarrow.lib.Table, got DataFrame)
+FAILED test_feather.py::TestFeather::test_write_with_index - TypeError: Argument 'table' has incorrect type (expected pyarrow.lib.Table, got DataFrame)
+FAILED test_feather.py::TestFeather::test_path_pathlib - TypeError: Argument 'table' has incorrect type (expected pyarrow.lib.Table, got DataFrame)
+FAILED test_feather.py::TestFeather::test_path_localpath - TypeError: Argument 'table' has incorrect type (expected pyarrow.lib.Table, got DataFrame)
+FAILED test_feather.py::TestFeather::test_passthrough_keywords - AttributeError: 'DataFrame' object has no attribute 'column_names'
+FAILED test_feather.py::TestFeather::test_http_path - ImportError: pyarrow requires pandas 0.23.0 or above, pandas 0+unknown is installed
+================================================================================================== 8 failed, 4 passed, 1 warning in 0.74s ==================================================================================================
+(base) root@a371cf8a3f11:/home/pandas/pandas/tests/io#
+
+
+
+What am I doing wrong?
+",0,developement environment with docker build installation check i have read the platform linux generic with installation method other pandas version python version installation logs hi i am trying to create a development environment by using docker following these intructions however i have tests that fail due to my environment installation steps clone repo git clone git github com pandas dev pandas git cd pandas build docker image docker build no cache tag pandas yourname env launch docker image docker run it w home pandas rm v pwd home pandas pandas yourname env build the c extensions root home pandas python setup py build ext j launch some tests root home pandas cd pandas tests io root home pandas pytest test feather py and the i get import errors python importerror pyarrow requires pandas or above pandas unknown is installed the full output is the following base root home pandas pandas tests io pytest test feather py test session starts platform linux python pytest pluggy rootdir home pandas configfile pyproject toml plugins xdist forked asyncio cov cython anyio hypothesis asyncio mode strict collected items test feather py f f ffffff failures testfeather test basic self workspaces pandas pandas tests io test feather py workspaces pandas pandas tests io test feather py in check round trip io feather format py in to feather feather write feather df handles handle kwargs df string int uint float float with null bool bool with null cat dt dttz dt with nul false c days dest compression compression level none chunksize none version def write feather df dest compression none compression level none chunksize none version write a pandas dataframe to feather format parameters df pandas dataframe or pyarrow table data to write out as feather format dest str local destination path compression string default none can be one of zstd uncompressed the default of none uses for files if it is available otherwise uncompressed compression level int default none use a compression level particular to the chosen compressor if none use the default compression level chunksize int default none for files the internal maximum size of arrow recordbatch chunks when writing the arrow ipc file format none means use the default which is currently version int default feather file version version is the current version is the more limited legacy format if pandas api have pandas check pandas version if pandas api has sparse and isinstance df pandas api pd sparsedataframe df df to dense if pandas api is data frame df feather creates a new column in the resultant table to store index information if index type is not rangeindex if version preserve index false elif version preserve index none else raise valueerror version value should either be or table table from pandas df preserve index preserve index if version version does not chunking for i name in enumerate table schema names col table check chunked overflow name col else table df if version if len table column names len set table column names raise valueerror cannot serialize duplicate column names if compression is not none raise valueerror feather files do not support compression option if chunksize is not none raise valueerror feather files do not support chunksize option else if compression is none and codec is available frame compression elif compression is not none and compression not in feather supported codecs raise valueerror compression not supported must be one of format compression feather supported codecs try feather write feather table dest compression compression compression level compression level chunksize chunksize version version e typeerror argument table has incorrect type expected pyarrow lib table got dataframe opt conda lib site packages pyarrow feather py typeerror testfeather test read columns self workspaces pandas pandas tests io test feather py workspaces pandas pandas tests io test feather py in check round trip io feather format py in to feather feather write feather df handles handle kwargs df a x b y c z dest compression compression level none chunksize none version def write feather df dest compression none compression level none chunksize none version write a pandas dataframe to feather format parameters df pandas dataframe or pyarrow table data to write out as feather format dest str local destination path compression string default none can be one of zstd uncompressed the default of none uses for files if it is available otherwise uncompressed compression level int default none use a compression level particular to the chosen compressor if none use the default compression level chunksize int default none for files the internal maximum size of arrow recordbatch chunks when writing the arrow ipc file format none means use the default which is currently version int default feather file version version is the current version is the more limited legacy format if pandas api have pandas check pandas version if pandas api has sparse and isinstance df pandas api pd sparsedataframe df df to dense if pandas api is data frame df feather creates a new column in the resultant table to store index information if index type is not rangeindex if version preserve index false elif version preserve index none else raise valueerror version value should either be or table table from pandas df preserve index preserve index if version version does not chunking for i name in enumerate table schema names col table check chunked overflow name col else table df if version if len table column names len set table column names raise valueerror cannot serialize duplicate column names if compression is not none raise valueerror feather files do not support compression option if chunksize is not none raise valueerror feather files do not support chunksize option else if compression is none and codec is available frame compression elif compression is not none and compression not in feather supported codecs raise valueerror compression not supported must be one of format compression feather supported codecs try feather write feather table dest compression compression compression level compression level chunksize chunksize version version e typeerror argument table has incorrect type expected pyarrow lib table got dataframe opt conda lib site packages pyarrow feather py typeerror testfeather test rw use threads self workspaces pandas pandas tests io test feather py workspaces pandas pandas tests io test feather py in check round trip io feather format py in to feather feather write feather df handles handle kwargs df a dest compression compression level none chunksize none version def write feather df dest compression none compression level none chunksize none version write a pandas dataframe to feather format parameters df pandas dataframe or pyarrow table data to write out as feather format dest str local destination path compression string default none can be one of zstd uncompressed the default of none uses for files if it is available otherwise uncompressed compression level int default none use a compression level particular to the chosen compressor if none use the default compression level chunksize int default none for files the internal maximum size of arrow recordbatch chunks when writing the arrow ipc file format none means use the default which is currently version int default feather file version version is the current version is the more limited legacy format if pandas api have pandas check pandas version if pandas api has sparse and isinstance df pandas api pd sparsedataframe df df to dense if pandas api is data frame df feather creates a new column in the resultant table to store index information if index type is not rangeindex if version preserve index false elif version preserve index none else raise valueerror version value should either be or table table from pandas df preserve index preserve index if version version does not chunking for i name in enumerate table schema names col table check chunked overflow name col else table df if version if len table column names len set table column names raise valueerror cannot serialize duplicate column names if compression is not none raise valueerror feather files do not support compression option if chunksize is not none raise valueerror feather files do not support chunksize option else if compression is none and codec is available frame compression elif compression is not none and compression not in feather supported codecs raise valueerror compression not supported must be one of format compression feather supported codecs try feather write feather table dest compression compression compression level compression level chunksize chunksize version version e typeerror argument table has incorrect type expected pyarrow lib table got dataframe opt conda lib site packages pyarrow feather py typeerror testfeather test write with index self workspaces pandas pandas tests io test feather py workspaces pandas pandas tests io test feather py in check round trip io feather format py in to feather feather write feather df handles handle kwargs df a dest compression compression level none chunksize none version def write feather df dest compression none compression level none chunksize none version write a pandas dataframe to feather format parameters df pandas dataframe or pyarrow table data to write out as feather format dest str local destination path compression string default none can be one of zstd uncompressed the default of none uses for files if it is available otherwise uncompressed compression level int default none use a compression level particular to the chosen compressor if none use the default compression level chunksize int default none for files the internal maximum size of arrow recordbatch chunks when writing the arrow ipc file format none means use the default which is currently version int default feather file version version is the current version is the more limited legacy format if pandas api have pandas check pandas version if pandas api has sparse and isinstance df pandas api pd sparsedataframe df df to dense if pandas api is data frame df feather creates a new column in the resultant table to store index information if index type is not rangeindex if version preserve index false elif version preserve index none else raise valueerror version value should either be or table table from pandas df preserve index preserve index if version version does not chunking for i name in enumerate table schema names col table check chunked overflow name col else table df if version if len table column names len set table column names raise valueerror cannot serialize duplicate column names if compression is not none raise valueerror feather files do not support compression option if chunksize is not none raise valueerror feather files do not support chunksize option else if compression is none and codec is available frame compression elif compression is not none and compression not in feather supported codecs raise valueerror compression not supported must be one of format compression feather supported codecs try feather write feather table dest compression compression compression level compression level chunksize chunksize version version e typeerror argument table has incorrect type expected pyarrow lib table got dataframe opt conda lib site packages pyarrow feather py typeerror testfeather test path pathlib self workspaces pandas pandas tests io test feather py testing io py in round trip pathlib writer path path util decorators py in wrapper return func args kwargs core frame py in to feather to feather self path kwargs io feather format py in to feather feather write feather df handles handle kwargs df index a b c d yhda dest compression compression level none chunksize none version def write feather df dest compression none compression level none chunksize none version write a pandas dataframe to feather format parameters df pandas dataframe or pyarrow table data to write out as feather format dest str local destination path compression string default none can be one of zstd uncompressed the default of none uses for files if it is available otherwise uncompressed compression level int default none use a compression level particular to the chosen compressor if none use the default compression level chunksize int default none for files the internal maximum size of arrow recordbatch chunks when writing the arrow ipc file format none means use the default which is currently version int default feather file version version is the current version is the more limited legacy format if pandas api have pandas check pandas version if pandas api has sparse and isinstance df pandas api pd sparsedataframe df df to dense if pandas api is data frame df feather creates a new column in the resultant table to store index information if index type is not rangeindex if version preserve index false elif version preserve index none else raise valueerror version value should either be or table table from pandas df preserve index preserve index if version version does not chunking for i name in enumerate table schema names col table check chunked overflow name col else table df if version if len table column names len set table column names raise valueerror cannot serialize duplicate column names if compression is not none raise valueerror feather files do not support compression option if chunksize is not none raise valueerror feather files do not support chunksize option else if compression is none and codec is available frame compression elif compression is not none and compression not in feather supported codecs raise valueerror compression not supported must be one of format compression feather supported codecs try feather write feather table dest compression compression compression level compression level chunksize chunksize version version e typeerror argument table has incorrect type expected pyarrow lib table got dataframe opt conda lib site packages pyarrow feather py typeerror testfeather test path localpath self workspaces pandas pandas tests io test feather py testing io py in round trip localpath writer localpath path util decorators py in wrapper return func args kwargs core frame py in to feather to feather self path kwargs io feather format py in to feather feather write feather df handles handle kwargs df index a b c d rdygvlsusu dest compression compression level none chunksize none version def write feather df dest compression none compression level none chunksize none version write a pandas dataframe to feather format parameters df pandas dataframe or pyarrow table data to write out as feather format dest str local destination path compression string default none can be one of zstd uncompressed the default of none uses for files if it is available otherwise uncompressed compression level int default none use a compression level particular to the chosen compressor if none use the default compression level chunksize int default none for files the internal maximum size of arrow recordbatch chunks when writing the arrow ipc file format none means use the default which is currently version int default feather file version version is the current version is the more limited legacy format if pandas api have pandas check pandas version if pandas api has sparse and isinstance df pandas api pd sparsedataframe df df to dense if pandas api is data frame df feather creates a new column in the resultant table to store index information if index type is not rangeindex if version preserve index false elif version preserve index none else raise valueerror version value should either be or table table from pandas df preserve index preserve index if version version does not chunking for i name in enumerate table schema names col table check chunked overflow name col else table df if version if len table column names len set table column names raise valueerror cannot serialize duplicate column names if compression is not none raise valueerror feather files do not support compression option if chunksize is not none raise valueerror feather files do not support chunksize option else if compression is none and codec is available frame compression elif compression is not none and compression not in feather supported codecs raise valueerror compression not supported must be one of format compression feather supported codecs try feather write feather table dest compression compression compression level compression level chunksize chunksize version version e typeerror argument table has incorrect type expected pyarrow lib table got dataframe opt conda lib site packages pyarrow feather py typeerror testfeather test passthrough keywords self workspaces pandas pandas tests io test feather py workspaces pandas pandas tests io test feather py in check round trip io feather format py in to feather feather write feather df handles handle kwargs opt conda lib site packages pyarrow feather py in write feather if len table column names len set table column names self index a b c d xjsorykgcu name column names def getattr self name str after regular attribute access try looking up the name this allows simpler access to columns for interactive use note obj x will always call obj getattribute x prior to calling obj getattr x if name not in self internal names set and name not in self metadata and name not in self accessors and self info axis can hold identifiers and holds name name return self return object getattribute self name e attributeerror dataframe object has no attribute column names core generic py attributeerror testfeather test http path self feather file home pandas pandas tests io data feather feather feather workspaces pandas pandas tests io test feather py io feather format py in read feather return feather read feather opt conda lib site packages pyarrow feather py in read feather check pandas version opt conda lib site packages pyarrow feather py in check pandas version if pandas api loose version version pyarrow pandas shim pxi in pyarrow lib pandasapishim loose version get pyarrow pandas shim pxi in pyarrow lib pandasapishim check import e importerror pyarrow requires pandas or above pandas unknown is installed pyarrow pandas shim pxi importerror warnings summary core construction py home pandas pandas core construction py userwarning pyarrow requires pandas or above pandas unknown is installed therefore pandas specific integration is not used return cls from sequence data dtype dtype copy copy docs generated xml file home pandas pandas tests io test data xml slowest durations call pandas tests io test feather py testfeather test http path call pandas tests io test feather py testfeather test error durations hidden use vv to show these durations short test summary info failed test feather py testfeather test basic typeerror argument table has incorrect type expected pyarrow lib table got dataframe failed test feather py testfeather test read columns typeerror argument table has incorrect type expected pyarrow lib table got dataframe failed test feather py testfeather test rw use threads typeerror argument table has incorrect type expected pyarrow lib table got dataframe failed test feather py testfeather test write with index typeerror argument table has incorrect type expected pyarrow lib table got dataframe failed test feather py testfeather test path pathlib typeerror argument table has incorrect type expected pyarrow lib table got dataframe failed test feather py testfeather test path localpath typeerror argument table has incorrect type expected pyarrow lib table got dataframe failed test feather py testfeather test passthrough keywords attributeerror dataframe object has no attribute column names failed test feather py testfeather test http path importerror pyarrow requires pandas or above pandas unknown is installed failed passed warning in base root home pandas pandas tests io what am i doing wrong ,0
+31103,11869993304.0,IssuesEvent,2020-03-26 11:59:16,JMD60260/llsolidaires,https://api.github.com/repos/JMD60260/llsolidaires,opened,CVE-2012-6708 (Medium) detected in jquery-1.7.1.min.js,security vulnerability,"## CVE-2012-6708 - Medium Severity Vulnerability
+ Vulnerable Library - jquery-1.7.1.min.js
Path to dependency file: /tmp/ws-scm/llsolidaires/les-logements-solidaires/node_modules/sockjs/examples/express/index.html
+
Path to vulnerable library: /llsolidaires/les-logements-solidaires/node_modules/sockjs/examples/express/index.html,/llsolidaires/les-logements-solidaires/node_modules/sockjs/examples/multiplex/index.html,/llsolidaires/les-logements-solidaires/node_modules/sockjs/examples/express-3.x/index.html,/llsolidaires/les-logements-solidaires/node_modules/sockjs/examples/hapi/html/index.html,/llsolidaires/les-logements-solidaires/node_modules/sockjs/examples/echo/index.html
+
+jQuery before 1.9.0 is vulnerable to Cross-site Scripting (XSS) attacks. The jQuery(strInput) function does not differentiate selectors from HTML in a reliable fashion. In vulnerable versions, jQuery determined whether the input was HTML by looking for the '<' character anywhere in the string, giving attackers more flexibility when attempting to construct a malicious payload. In fixed versions, jQuery only deems the input to be HTML if it explicitly starts with the '<' character, limiting exploitability only to attackers who can control the beginning of a string, which is far less common.
+
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",True,"CVE-2012-6708 (Medium) detected in jquery-1.7.1.min.js - ## CVE-2012-6708 - Medium Severity Vulnerability
+ Vulnerable Library - jquery-1.7.1.min.js
+
+
Path to dependency file: /tmp/ws-scm/llsolidaires/les-logements-solidaires/node_modules/sockjs/examples/express/index.html
+
Path to vulnerable library: /llsolidaires/les-logements-solidaires/node_modules/sockjs/examples/express/index.html,/llsolidaires/les-logements-solidaires/node_modules/sockjs/examples/multiplex/index.html,/llsolidaires/les-logements-solidaires/node_modules/sockjs/examples/express-3.x/index.html,/llsolidaires/les-logements-solidaires/node_modules/sockjs/examples/hapi/html/index.html,/llsolidaires/les-logements-solidaires/node_modules/sockjs/examples/echo/index.html
+
+jQuery before 1.9.0 is vulnerable to Cross-site Scripting (XSS) attacks. The jQuery(strInput) function does not differentiate selectors from HTML in a reliable fashion. In vulnerable versions, jQuery determined whether the input was HTML by looking for the '<' character anywhere in the string, giving attackers more flexibility when attempting to construct a malicious payload. In fixed versions, jQuery only deems the input to be HTML if it explicitly starts with the '<' character, limiting exploitability only to attackers who can control the beginning of a string, which is far less common.
+
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",0,cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file tmp ws scm llsolidaires les logements solidaires node modules sockjs examples express index html path to vulnerable library llsolidaires les logements solidaires node modules sockjs examples express index html llsolidaires les logements solidaires node modules sockjs examples multiplex index html llsolidaires les logements solidaires node modules sockjs examples express x index html llsolidaires les logements solidaires node modules sockjs examples hapi html index html llsolidaires les logements solidaires node modules sockjs examples echo index html dependency hierarchy x jquery min js vulnerable library found in head commit a href vulnerability details jquery before is vulnerable to cross site scripting xss attacks the jquery strinput function does not differentiate selectors from html in a reliable fashion in vulnerable versions jquery determined whether the input was html by looking for the character anywhere in the string giving attackers more flexibility when attempting to construct a malicious payload in fixed versions jquery only deems the input to be html if it explicitly starts with the character limiting exploitability only to attackers who can control the beginning of a string which is far less common publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource ,0
+3234,2537521843.0,IssuesEvent,2015-01-26 21:10:05,web2py/web2py,https://api.github.com/repos/web2py/web2py,opened,Group of Groups and setting auth.requires_ on the app itself,1 star enhancement imported Priority-Medium,"_From [ramstei..._at_gmail.com](https://code.google.com/u/106807652641341582815/) on July 22, 2014 10:56:56_
+
+In a small organization (+- 100 users) it can be dificult to manage ACLs without being able to add group to group membership.
+
+consider my company
+
+we have this groups in the maintenance department
+
+MMC (20 users)
+ME ( 15 users)
+MI ( 10 users)
+
+MD (maintenance department)
+(all users above)
+
+
+so this group MD is composed of MMC,ME, MI and we dont repeat adding all users to MD, we just add group MMC,ME,MI to MD
+
+
+
+
+When we delete a person from ME, the MD group automaticaly reflects the change.
+
+
+Also in web2py if i want to block user A to access most of the app, i would need to decorate every function to block user A.
+
+If i could decorate the app itself i would only do it once and would not need to repeat myself for every function that should block user A.
+
+_Original issue: http://code.google.com/p/web2py/issues/detail?id=1953_",1.0,"Group of Groups and setting auth.requires_ on the app itself - _From [ramstei..._at_gmail.com](https://code.google.com/u/106807652641341582815/) on July 22, 2014 10:56:56_
+
+In a small organization (+- 100 users) it can be dificult to manage ACLs without being able to add group to group membership.
+
+consider my company
+
+we have this groups in the maintenance department
+
+MMC (20 users)
+ME ( 15 users)
+MI ( 10 users)
+
+MD (maintenance department)
+(all users above)
+
+
+so this group MD is composed of MMC,ME, MI and we dont repeat adding all users to MD, we just add group MMC,ME,MI to MD
+
+
+
+
+When we delete a person from ME, the MD group automaticaly reflects the change.
+
+
+Also in web2py if i want to block user A to access most of the app, i would need to decorate every function to block user A.
+
+If i could decorate the app itself i would only do it once and would not need to repeat myself for every function that should block user A.
+
+_Original issue: http://code.google.com/p/web2py/issues/detail?id=1953_",0,group of groups and setting auth requires on the app itself from on july in a small organization users it can be dificult to manage acls without being able to add group to group membership consider my company we have this groups in the maintenance department mmc users me users mi users md maintenance department all users above so this group md is composed of mmc me mi and we dont repeat adding all users to md we just add group mmc me mi to md when we delete a person from me the md group automaticaly reflects the change also in if i want to block user a to access most of the app i would need to decorate every function to block user a if i could decorate the app itself i would only do it once and would not need to repeat myself for every function that should block user a original issue ,0
+205,4271762323.0,IssuesEvent,2016-07-13 12:26:17,wordpress-mobile/WordPress-Android,https://api.github.com/repos/wordpress-mobile/WordPress-Android,closed,Viewers section in People Management,People Management [Type] Enhancement,"We have decided to separate viewers section from the followers issue #4142.
+",1.0,"Viewers section in People Management - We have decided to separate viewers section from the followers issue #4142.
+",1,viewers section in people management we have decided to separate viewers section from the followers issue ,1
+5756,8208337170.0,IssuesEvent,2018-09-04 00:55:51,ice1000/julia-intellij,https://api.github.com/repos/ice1000/julia-intellij,closed,Can't seem to type in a backslash,IDE: WebStorm bug compatibility,"Hello, I've just begun using Julia, and the Julia plugin with WebStorm.
+
+I was trying to write a string with a ""\n"" in it, but it seems that the IDE won't let me type a ""\\"" when I have the language of the file set to Julia (*.jl).
+
+This only happens when I have this plugin enabled, and only when the language is Julia, so I assume it is this plugin that is preventing it.
+
+",True,"Can't seem to type in a backslash - Hello, I've just begun using Julia, and the Julia plugin with WebStorm.
+
+I was trying to write a string with a ""\n"" in it, but it seems that the IDE won't let me type a ""\\"" when I have the language of the file set to Julia (*.jl).
+
+This only happens when I have this plugin enabled, and only when the language is Julia, so I assume it is this plugin that is preventing it.
+
+",0,can t seem to type in a backslash hello i ve just begun using julia and the julia plugin with webstorm i was trying to write a string with a n in it but it seems that the ide won t let me type a when i have the language of the file set to julia jl this only happens when i have this plugin enabled and only when the language is julia so i assume it is this plugin that is preventing it ,0
+81,3361837518.0,IssuesEvent,2015-11-20 00:35:10,bbergen/tbmd.com,https://api.github.com/repos/bbergen/tbmd.com,closed,Person Block View,People,"Currently tbmd.com/person.php only shows the person's associated image, and no other information (birthdate, bio, etc.).
+
+A block view needs to be created in Person.php to return this html display.
+
+Use Movie::asBlockView() as a template for this issue.
+
+CSS style will be handled in a separate issue, just focus on blocking out the content into appropriately named and ordered divs so that the PersonView class can easily retrieve the block view for a single or multiple people easily. ",1.0,"Person Block View - Currently tbmd.com/person.php only shows the person's associated image, and no other information (birthdate, bio, etc.).
+
+A block view needs to be created in Person.php to return this html display.
+
+Use Movie::asBlockView() as a template for this issue.
+
+CSS style will be handled in a separate issue, just focus on blocking out the content into appropriately named and ordered divs so that the PersonView class can easily retrieve the block view for a single or multiple people easily. ",1,person block view currently tbmd com person php only shows the person s associated image and no other information birthdate bio etc a block view needs to be created in person php to return this html display use movie asblockview as a template for this issue css style will be handled in a separate issue just focus on blocking out the content into appropriately named and ordered divs so that the personview class can easily retrieve the block view for a single or multiple people easily ,1
+1008,24560297503.0,IssuesEvent,2022-10-12 19:35:56,microsoftgraph/microsoft-graph-toolkit,https://api.github.com/repos/microsoftgraph/microsoft-graph-toolkit,closed,People Picker component should be updated to the latest Fluent UI design,feature-request Component: people-picker State: Completed Fluent UI,"
+
+
+
+# Proposal: facilitate the integration of Fluent UI web components into the `mgt-people-picker` control
+
+
+## Description
+
+
+This is an open thread for implementation, discussion and review of the `mgt-people-picker` upgrade to utilize Fluent UI web components.
+
+https://github.com/microsoft/fluentui/tree/master/packages/web-components
+
+## Rationale
+
+Further mgt alignment with the current fluent guidelines
+
+## Preferred Solution
+
+- [x] Phase 1: Suggest controls/UI/text with similar functionality/design that could be replaced with the Fluent UI web component. **any concerns with implementation should be noted**
+- [ ] Phase 2: Review preferred components to be used with the `mgt-people-picker` (ideally with conformation from additional authors).
+- [ ] Phase 3: Develop POC with the `mgt-people-picker` and the new Fluent UI web components
+- [ ] Phase 4: Test against current functionality.
+- [ ] Phase 5: Prepare for merge
+",1.0,"People Picker component should be updated to the latest Fluent UI design -
+
+
+
+# Proposal: facilitate the integration of Fluent UI web components into the `mgt-people-picker` control
+
+
+## Description
+
+
+This is an open thread for implementation, discussion and review of the `mgt-people-picker` upgrade to utilize Fluent UI web components.
+
+https://github.com/microsoft/fluentui/tree/master/packages/web-components
+
+## Rationale
+
+Further mgt alignment with the current fluent guidelines
+
+## Preferred Solution
+
+- [x] Phase 1: Suggest controls/UI/text with similar functionality/design that could be replaced with the Fluent UI web component. **any concerns with implementation should be noted**
+- [ ] Phase 2: Review preferred components to be used with the `mgt-people-picker` (ideally with conformation from additional authors).
+- [ ] Phase 3: Develop POC with the `mgt-people-picker` and the new Fluent UI web components
+- [ ] Phase 4: Test against current functionality.
+- [ ] Phase 5: Prepare for merge
+",1,people picker component should be updated to the latest fluent ui design proposal facilitate the integration of fluent ui web components into the mgt people picker control description this is an open thread for implementation discussion and review of the mgt people picker upgrade to utilize fluent ui web components rationale further mgt alignment with the current fluent guidelines preferred solution phase suggest controls ui text with similar functionality design that could be replaced with the fluent ui web component any concerns with implementation should be noted phase review preferred components to be used with the mgt people picker ideally with conformation from additional authors phase develop poc with the mgt people picker and the new fluent ui web components phase test against current functionality phase prepare for merge ,1
+408105,11941653621.0,IssuesEvent,2020-04-02 18:50:01,LibreTexts/ckeditor-binder-plugin,https://api.github.com/repos/LibreTexts/ckeditor-binder-plugin,closed,define HTML structure,High Priority,"To ensure all we never have to go back and change the HTML structure for our plugin, we need to redefine our HTML structure. The current structure is as below:
+
+```html
+
+
+ print('hello world')
+
+
+ hello world
+
+
+```
+
+The only tag defined by us is the most outer tag.",1.0,"define HTML structure - To ensure all we never have to go back and change the HTML structure for our plugin, we need to redefine our HTML structure. The current structure is as below:
+
+```html
+
+
+ print('hello world')
+
+
+ hello world
+
+
+```
+
+The only tag defined by us is the most outer tag.",0,define html structure to ensure all we never have to go back and change the html structure for our plugin we need to redefine our html structure the current structure is as below html print hello world hello world the only tag defined by us is the most outer tag ,0
+942,20911596553.0,IssuesEvent,2022-03-24 09:49:01,abumunye/orderpin-bugs,https://api.github.com/repos/abumunye/orderpin-bugs,opened,Text not readable due to busy background,Bug Cosmetic RestarauntPeoples,"## Describe the bug
+The tags on some of the images are not readable due to the contrasting background and alternative is to have the text on the carosel at the top highlighted so that whatever the background is it can be seen
+
+## To Reproduce
+Steps to reproduce the behavior:
+1. Go to 'home'
+2. Scroll down to 'various different posts on the top carousel'
+4. See error
+
+## Expected behavior
+You wont be able to see some tags due to the background
+## Screenshots
+If applicable, add screenshots to help explain your problem.
+
+
+## Desktop (please complete the following information):
+ - OS: [Windows]
+ - Browser [Chrome]
+",1.0,"Text not readable due to busy background - ## Describe the bug
+The tags on some of the images are not readable due to the contrasting background and alternative is to have the text on the carosel at the top highlighted so that whatever the background is it can be seen
+
+## To Reproduce
+Steps to reproduce the behavior:
+1. Go to 'home'
+2. Scroll down to 'various different posts on the top carousel'
+4. See error
+
+## Expected behavior
+You wont be able to see some tags due to the background
+## Screenshots
+If applicable, add screenshots to help explain your problem.
+
+
+## Desktop (please complete the following information):
+ - OS: [Windows]
+ - Browser [Chrome]
+",1,text not readable due to busy background describe the bug the tags on some of the images are not readable due to the contrasting background and alternative is to have the text on the carosel at the top highlighted so that whatever the background is it can be seen to reproduce steps to reproduce the behavior go to home scroll down to various different posts on the top carousel see error expected behavior you wont be able to see some tags due to the background screenshots if applicable add screenshots to help explain your problem desktop please complete the following information os browser ,1
+88796,8178293485.0,IssuesEvent,2018-08-28 13:30:18,Microsoft/vscode,https://api.github.com/repos/Microsoft/vscode,closed,Test settings editor,testplan-item,"Testing https://github.com/Microsoft/vscode/issues/55791
+
+Complexity: 4
+
+- [x] Mac @chrmarti
+- [x] Win @isidorn
+- [x] Linux @jrieken
+
+Test everything in the settings editor. Note that there are screen reader fixes still incoming.
+
+Test combinations of `@filter`, search, and clicking on the TOC to filter during search. Test that editing settings works while searching and filtering.
+
+Test that you can open multiple instances of the settings editor, and they don't get out of sync.
+
+Test that you can navigate by keyboard in a sane way.
+
+Test that the new setting to pick your preferred settings editor, `workbench.settings.editor`, works, and it applies to all the different ways to open settings (default keybinding, menu, activity bar, search viewlet, maybe others)
+
+The settings editor loads pages of settings as you scroll. You can see the scrollbar jerk as you scroll to the end. Watch for any issues this might cause.
+
+Test the gear icon context menu. Default keybinding is shift+f9.
+
+Test editing in different scopes in multi and single folder workspaces, and empty windows. Test that you only see settings that are valid for the scope.
+
+Test that the new dropdown control is a good experience and works the way a dropdown control should, and that enumDescriptions are shown and rendered as markdown when they should be.
+
+Watch for descriptions or enumDescriptions that are not rendered as markdown when they should be (setting contributions now have to opt in to be rendered as markdown, and I had to manually find all descriptions that use markdown)",1.0,"Test settings editor - Testing https://github.com/Microsoft/vscode/issues/55791
+
+Complexity: 4
+
+- [x] Mac @chrmarti
+- [x] Win @isidorn
+- [x] Linux @jrieken
+
+Test everything in the settings editor. Note that there are screen reader fixes still incoming.
+
+Test combinations of `@filter`, search, and clicking on the TOC to filter during search. Test that editing settings works while searching and filtering.
+
+Test that you can open multiple instances of the settings editor, and they don't get out of sync.
+
+Test that you can navigate by keyboard in a sane way.
+
+Test that the new setting to pick your preferred settings editor, `workbench.settings.editor`, works, and it applies to all the different ways to open settings (default keybinding, menu, activity bar, search viewlet, maybe others)
+
+The settings editor loads pages of settings as you scroll. You can see the scrollbar jerk as you scroll to the end. Watch for any issues this might cause.
+
+Test the gear icon context menu. Default keybinding is shift+f9.
+
+Test editing in different scopes in multi and single folder workspaces, and empty windows. Test that you only see settings that are valid for the scope.
+
+Test that the new dropdown control is a good experience and works the way a dropdown control should, and that enumDescriptions are shown and rendered as markdown when they should be.
+
+Watch for descriptions or enumDescriptions that are not rendered as markdown when they should be (setting contributions now have to opt in to be rendered as markdown, and I had to manually find all descriptions that use markdown)",0,test settings editor testing complexity mac chrmarti win isidorn linux jrieken test everything in the settings editor note that there are screen reader fixes still incoming test combinations of filter search and clicking on the toc to filter during search test that editing settings works while searching and filtering test that you can open multiple instances of the settings editor and they don t get out of sync test that you can navigate by keyboard in a sane way test that the new setting to pick your preferred settings editor workbench settings editor works and it applies to all the different ways to open settings default keybinding menu activity bar search viewlet maybe others the settings editor loads pages of settings as you scroll you can see the scrollbar jerk as you scroll to the end watch for any issues this might cause test the gear icon context menu default keybinding is shift test editing in different scopes in multi and single folder workspaces and empty windows test that you only see settings that are valid for the scope test that the new dropdown control is a good experience and works the way a dropdown control should and that enumdescriptions are shown and rendered as markdown when they should be watch for descriptions or enumdescriptions that are not rendered as markdown when they should be setting contributions now have to opt in to be rendered as markdown and i had to manually find all descriptions that use markdown ,0
+9521,3051895857.0,IssuesEvent,2015-08-12 11:42:57,Financial-Times/next-front-page,https://api.github.com/repos/Financial-Times/next-front-page,closed,Fetch video data from brightcove,Ready to test,"Submitted https://github.com/Financial-Times/next-brightcove-proxy-api/pull/33 to fetch videos by playlist id.
+
+Blocked until PR is merged",1.0,"Fetch video data from brightcove - Submitted https://github.com/Financial-Times/next-brightcove-proxy-api/pull/33 to fetch videos by playlist id.
+
+Blocked until PR is merged",0,fetch video data from brightcove submitted to fetch videos by playlist id blocked until pr is merged,0
+293012,22042062819.0,IssuesEvent,2022-05-29 14:02:26,lardemua/atom,https://api.github.com/repos/lardemua/atom,closed,LiDAR resolution,documentation,"FYI @miguelriemoliveira
+
+I made measurements for all the LiDARs and all of them were around 10cm
+
+(length in the bottom left corner)
+
+
+
+",1.0,"LiDAR resolution - FYI @miguelriemoliveira
+
+I made measurements for all the LiDARs and all of them were around 10cm
+
+(length in the bottom left corner)
+
+
+
+",0,lidar resolution fyi miguelriemoliveira i made measurements for all the lidars and all of them were around length in the bottom left corner ,0
+891,18249272304.0,IssuesEvent,2021-10-02 00:29:22,planningcenter/developers,https://api.github.com/repos/planningcenter/developers,closed,Date/time Unexpected Behavior,People,"**Affected Product**
+Which product does this bug affect?
+People
+
+**Describe the bug**
+It seems that the URL https://api.planningcenteronline.com/people/v2/people?where[created_at][gt]=2021-09-01 is returning not only people that were created today (09/02/21), but also people that were created yesterday, on 09/01/21. From searching for this issue, I would guess that ""greater than 2021-09-01"" is including all times after 2021-09-01T00:00:00Z, even if the date is still 2021-09-01. If this is the intended behavior, the documentation may need to be updated. When looking at this page in the documentation (https://developer.planning.center/docs/#/overview/dates-times), this text is found toward the bottom of the page:
+
+
+
+Going through the table listed there, these are the results I get when using each operator with the date 2021-09-01:
+
+""="" returns no people at all. Presumably, it is searching for people created at exactly 2021-09-01T00:00:00Z, and (rightly so) not finding any. Given the description listed on the documentation table, I expected to see all profiles created on that day, regardless of the hour and minute they were created.
+
+""[gt]="" returns people created on 09/01/21 as well as people created on 09/02/21. Given the description in the table, I did not expect to see people created on 09/01/21 in the results.
+
+""[gte]="" returns, in this case, exactly the same results as the above query. I assume if there were a profile created at exactly 2021-09-01T00:00:00Z, it would appear in this query, but not the previous one. However, this does appear to be the expected behavior for this query.
+
+""[lt]="" returns people created before 09/01/21. This is the expected behavior.
+
+""[lte]="" returns, in this case, the same results as the previous query. I assume if there were a profile created at exactly 2021-09-01T00:00:00Z, it would appear in this query, but not the previous one. Given the description in the documentation table, I expected to see profiles created at any time on 09/01/21 and earlier.
+
+Lastly, at the bottom of the page is this info, ""gt, gte, lt, and lte can be combined to further filter records. For example ?where[created_at][gte]=2018-02-01&where[created_at][lte]=2018-02-28 will return all plans created during February 2018."" However, this query does not actually return all results for the month of February, as any plans (or people in my case) created on 02/28/18 would not be included, unless they were created at exactly 2018-02-28T00:00:00Z. As far as I can tell, to get all results from February, you would have to do something like:
+
+?where[created_at][gte]=2018-02-01&where[created_at][lt]=2018-03-01
+or
+?where[created_at][gte]=2018-02-01&where[created_at][lte]=2018-02-28T23:59:59Z
+
+**To Reproduce**
+1. Make a GET request to the endpoint listed below, querying using any of the above operators
+2. Review the response and see that the behavior of some operators appears to differ from their documented behavior
+
+**Additional Context:**
+- Endpoint: https://api.planningcenteronline.com/people/v2/people
+- Python 3.7
+- Authentication: Personal Access Token
+
+**Additional context**
+If this is indeed the desired behavior from each operator, I suggest just updating the documentation to reflect that. In my specific case, I am helping a co-worker expand on an existing third-party integration, where it would be less than ideal for them as an end user to have to remember to add a day to their searches each time, but it's not the end of the world if that's the way it has to be. For other cases where I could just rework the code to account for this, it wouldn't be a big deal.
+
+## I have..
+
+- [x ] Reviewed the documentation found at https://developer.planning.center/docs
+- [x ] Searched for previous issues reporting this bug
+- [x ] Removed all private information from this issue (credentials, tokens, emails, phone numbers, etc.)
+- [x] Reviewed my issue for completeness",1.0,"Date/time Unexpected Behavior - **Affected Product**
+Which product does this bug affect?
+People
+
+**Describe the bug**
+It seems that the URL https://api.planningcenteronline.com/people/v2/people?where[created_at][gt]=2021-09-01 is returning not only people that were created today (09/02/21), but also people that were created yesterday, on 09/01/21. From searching for this issue, I would guess that ""greater than 2021-09-01"" is including all times after 2021-09-01T00:00:00Z, even if the date is still 2021-09-01. If this is the intended behavior, the documentation may need to be updated. When looking at this page in the documentation (https://developer.planning.center/docs/#/overview/dates-times), this text is found toward the bottom of the page:
+
+
+
+Going through the table listed there, these are the results I get when using each operator with the date 2021-09-01:
+
+""="" returns no people at all. Presumably, it is searching for people created at exactly 2021-09-01T00:00:00Z, and (rightly so) not finding any. Given the description listed on the documentation table, I expected to see all profiles created on that day, regardless of the hour and minute they were created.
+
+""[gt]="" returns people created on 09/01/21 as well as people created on 09/02/21. Given the description in the table, I did not expect to see people created on 09/01/21 in the results.
+
+""[gte]="" returns, in this case, exactly the same results as the above query. I assume if there were a profile created at exactly 2021-09-01T00:00:00Z, it would appear in this query, but not the previous one. However, this does appear to be the expected behavior for this query.
+
+""[lt]="" returns people created before 09/01/21. This is the expected behavior.
+
+""[lte]="" returns, in this case, the same results as the previous query. I assume if there were a profile created at exactly 2021-09-01T00:00:00Z, it would appear in this query, but not the previous one. Given the description in the documentation table, I expected to see profiles created at any time on 09/01/21 and earlier.
+
+Lastly, at the bottom of the page is this info, ""gt, gte, lt, and lte can be combined to further filter records. For example ?where[created_at][gte]=2018-02-01&where[created_at][lte]=2018-02-28 will return all plans created during February 2018."" However, this query does not actually return all results for the month of February, as any plans (or people in my case) created on 02/28/18 would not be included, unless they were created at exactly 2018-02-28T00:00:00Z. As far as I can tell, to get all results from February, you would have to do something like:
+
+?where[created_at][gte]=2018-02-01&where[created_at][lt]=2018-03-01
+or
+?where[created_at][gte]=2018-02-01&where[created_at][lte]=2018-02-28T23:59:59Z
+
+**To Reproduce**
+1. Make a GET request to the endpoint listed below, querying using any of the above operators
+2. Review the response and see that the behavior of some operators appears to differ from their documented behavior
+
+**Additional Context:**
+- Endpoint: https://api.planningcenteronline.com/people/v2/people
+- Python 3.7
+- Authentication: Personal Access Token
+
+**Additional context**
+If this is indeed the desired behavior from each operator, I suggest just updating the documentation to reflect that. In my specific case, I am helping a co-worker expand on an existing third-party integration, where it would be less than ideal for them as an end user to have to remember to add a day to their searches each time, but it's not the end of the world if that's the way it has to be. For other cases where I could just rework the code to account for this, it wouldn't be a big deal.
+
+## I have..
+
+- [x ] Reviewed the documentation found at https://developer.planning.center/docs
+- [x ] Searched for previous issues reporting this bug
+- [x ] Removed all private information from this issue (credentials, tokens, emails, phone numbers, etc.)
+- [x] Reviewed my issue for completeness",1,date time unexpected behavior affected product which product does this bug affect people describe the bug it seems that the url is returning not only people that were created today but also people that were created yesterday on from searching for this issue i would guess that greater than is including all times after even if the date is still if this is the intended behavior the documentation may need to be updated when looking at this page in the documentation this text is found toward the bottom of the page img width alt screen shot at pm src going through the table listed there these are the results i get when using each operator with the date returns no people at all presumably it is searching for people created at exactly and rightly so not finding any given the description listed on the documentation table i expected to see all profiles created on that day regardless of the hour and minute they were created returns people created on as well as people created on given the description in the table i did not expect to see people created on in the results returns in this case exactly the same results as the above query i assume if there were a profile created at exactly it would appear in this query but not the previous one however this does appear to be the expected behavior for this query returns people created before this is the expected behavior returns in this case the same results as the previous query i assume if there were a profile created at exactly it would appear in this query but not the previous one given the description in the documentation table i expected to see profiles created at any time on and earlier lastly at the bottom of the page is this info gt gte lt and lte can be combined to further filter records for example where where will return all plans created during february however this query does not actually return all results for the month of february as any plans or people in my case created on would not be included unless they were created at exactly as far as i can tell to get all results from february you would have to do something like where where or where where to reproduce make a get request to the endpoint listed below querying using any of the above operators review the response and see that the behavior of some operators appears to differ from their documented behavior additional context endpoint python authentication personal access token additional context if this is indeed the desired behavior from each operator i suggest just updating the documentation to reflect that in my specific case i am helping a co worker expand on an existing third party integration where it would be less than ideal for them as an end user to have to remember to add a day to their searches each time but it s not the end of the world if that s the way it has to be for other cases where i could just rework the code to account for this it wouldn t be a big deal i have reviewed the documentation found at searched for previous issues reporting this bug removed all private information from this issue credentials tokens emails phone numbers etc reviewed my issue for completeness,1
+836,15570351004.0,IssuesEvent,2021-03-17 02:21:31,gather-community/gather,https://api.github.com/repos/gather-community/gather,closed,Birthday calendar export,area:people priority:medium type:enhancement,"_Originally created by **Tom Smyth** at **2016-11-06 15:38**, migrated from [redmine-#5375](https://redmine.sassafras.coop/issues/5375)_",1.0,"Birthday calendar export - _Originally created by **Tom Smyth** at **2016-11-06 15:38**, migrated from [redmine-#5375](https://redmine.sassafras.coop/issues/5375)_",1,birthday calendar export originally created by tom smyth at migrated from ,1
+475109,13686988689.0,IssuesEvent,2020-09-30 09:28:06,gnosis/conditional-tokens-explorer,https://api.github.com/repos/gnosis/conditional-tokens-explorer,closed,Add custom token: Error messages for the contract validations show code/no validation message about unsuccessful attempt of unlocking an account,Medium priority QA Passed bug,"UPD: 17/09/2020
+Validation messages are not displayed at all not on the pop-uo, however, they are displayed in console
+
+
+
+
+Related to #202
+
+1. Error messages for the contract validations show code: see the video:
+https://drive.google.com/file/d/1UWUkgwc-cHCm_I1pQiU-4sijvPLwVBPK/view
+**Expected result:** messages are written in understandable to an end user manner
+
+2. No validation message about unsuccessful attempt of unlocking an account (see the video)
+ https://drive.google.com/file/d/1kiGBF0UK4GaG6wDyXEEYXKZwF65DgOoT/view
+
+**Expected result:** Error message should appear on the window",1.0,"Add custom token: Error messages for the contract validations show code/no validation message about unsuccessful attempt of unlocking an account - UPD: 17/09/2020
+Validation messages are not displayed at all not on the pop-uo, however, they are displayed in console
+
+
+
+
+Related to #202
+
+1. Error messages for the contract validations show code: see the video:
+https://drive.google.com/file/d/1UWUkgwc-cHCm_I1pQiU-4sijvPLwVBPK/view
+**Expected result:** messages are written in understandable to an end user manner
+
+2. No validation message about unsuccessful attempt of unlocking an account (see the video)
+ https://drive.google.com/file/d/1kiGBF0UK4GaG6wDyXEEYXKZwF65DgOoT/view
+
+**Expected result:** Error message should appear on the window",0,add custom token error messages for the contract validations show code no validation message about unsuccessful attempt of unlocking an account upd validation messages are not displayed at all not on the pop uo however they are displayed in console related to error messages for the contract validations show code see the video expected result messages are written in understandable to an end user manner no validation message about unsuccessful attempt of unlocking an account see the video expected result error message should appear on the window,0
+171217,20955498953.0,IssuesEvent,2022-03-27 03:22:48,RG4421/openedr,https://api.github.com/repos/RG4421/openedr,closed,CVE-2019-11358 (Medium) detected in multiple libraries - autoclosed,security vulnerability,"## CVE-2019-11358 - Medium Severity Vulnerability
+ Vulnerable Libraries - jquery-1.11.1.min.js, jquery-3.1.0.min.js, jquery-3.1.0.js, jquery-2.1.1.min.js
+
Path to dependency file: /edrav2/eprj/boost/libs/hof/doc/html/include/boost/hof/partial.html
+
Path to vulnerable library: /edrav2/eprj/boost/libs/hof/doc/html/_static/jquery.js,/edrav2/eprj/boost/libs/hof/doc/html/include/boost/hof/../../../_static/jquery.js,/edrav2/eprj/boost/libs/python/doc/html/numpy/reference/../_static/jquery.js,/edrav2/eprj/boost/libs/python/doc/html/numpy/_static/jquery.js,/edrav2/eprj/boost/libs/python/doc/html/numpy/_static/jquery.js,/edrav2/eprj/boost/libs/hof/doc/html/doc/../_static/jquery.js,/edrav2/eprj/boost/libs/hof/doc/html/_static/jquery.js,/edrav2/eprj/boost/libs/hof/doc/html/doc/src/../../_static/jquery.js,/edrav2/eprj/boost/libs/python/doc/html/numpy/tutorial/../_static/jquery.js
Path to vulnerable library: /edrav2/eprj/boost/libs/hof/doc/html/_static/jquery-3.1.0.js,/edrav2/eprj/boost/libs/python/doc/html/numpy/_static/jquery-3.1.0.js
+
+jQuery before 3.4.0, as used in Drupal, Backdrop CMS, and other products, mishandles jQuery.extend(true, {}, ...) because of Object.prototype pollution. If an unsanitized source object contained an enumerable __proto__ property, it could extend the native Object.prototype.
+
+
Path to dependency file: /edrav2/eprj/boost/libs/hof/doc/html/include/boost/hof/partial.html
+
Path to vulnerable library: /edrav2/eprj/boost/libs/hof/doc/html/_static/jquery.js,/edrav2/eprj/boost/libs/hof/doc/html/include/boost/hof/../../../_static/jquery.js,/edrav2/eprj/boost/libs/python/doc/html/numpy/reference/../_static/jquery.js,/edrav2/eprj/boost/libs/python/doc/html/numpy/_static/jquery.js,/edrav2/eprj/boost/libs/python/doc/html/numpy/_static/jquery.js,/edrav2/eprj/boost/libs/hof/doc/html/doc/../_static/jquery.js,/edrav2/eprj/boost/libs/hof/doc/html/_static/jquery.js,/edrav2/eprj/boost/libs/hof/doc/html/doc/src/../../_static/jquery.js,/edrav2/eprj/boost/libs/python/doc/html/numpy/tutorial/../_static/jquery.js
Path to vulnerable library: /edrav2/eprj/boost/libs/hof/doc/html/_static/jquery-3.1.0.js,/edrav2/eprj/boost/libs/python/doc/html/numpy/_static/jquery-3.1.0.js
+
+jQuery before 3.4.0, as used in Drupal, Backdrop CMS, and other products, mishandles jQuery.extend(true, {}, ...) because of Object.prototype pollution. If an unsanitized source object contained an enumerable __proto__ property, it could extend the native Object.prototype.
+
+
+
+
+
+
+
+",0,cve medium detected in multiple libraries autoclosed cve medium severity vulnerability vulnerable libraries jquery min js jquery min js jquery js jquery min js jquery min js javascript library for dom operations library home page a href path to dependency file eprj boost libs hana benchmark chart html path to vulnerable library eprj boost libs hana benchmark chart html dependency hierarchy x jquery min js vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file eprj boost libs hof doc html include boost hof partial html path to vulnerable library eprj boost libs hof doc html static jquery js eprj boost libs hof doc html include boost hof static jquery js eprj boost libs python doc html numpy reference static jquery js eprj boost libs python doc html numpy static jquery js eprj boost libs python doc html numpy static jquery js eprj boost libs hof doc html doc static jquery js eprj boost libs hof doc html static jquery js eprj boost libs hof doc html doc src static jquery js eprj boost libs python doc html numpy tutorial static jquery js dependency hierarchy x jquery min js vulnerable library jquery js javascript library for dom operations library home page a href path to vulnerable library eprj boost libs hof doc html static jquery js eprj boost libs python doc html numpy static jquery js dependency hierarchy x jquery js vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file eprj jsonrpccpp src examples index html path to vulnerable library eprj jsonrpccpp src examples index html dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch main vulnerability details jquery before as used in drupal backdrop cms and other products mishandles jquery extend true because of object prototype pollution if an unsanitized source object contained an enumerable proto property it could extend the native object prototype publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree jquery isminimumfixversionavailable true minimumfixversion isbinary false packagetype javascript packagename jquery packageversion packagefilepaths istransitivedependency false dependencytree jquery isminimumfixversionavailable true minimumfixversion isbinary false packagetype javascript packagename jquery packageversion packagefilepaths istransitivedependency false dependencytree jquery isminimumfixversionavailable true minimumfixversion isbinary true packagetype javascript packagename jquery packageversion packagefilepaths istransitivedependency false dependencytree jquery isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails jquery before as used in drupal backdrop cms and other products mishandles jquery extend true because of object prototype pollution if an unsanitized source object contained an enumerable proto property it could extend the native object prototype vulnerabilityurl ,0
+32130,4751406554.0,IssuesEvent,2016-10-22 21:29:58,ReactiveX/RxJava,https://api.github.com/repos/ReactiveX/RxJava,closed,FlowableConcatMapEagerTest > testAsynchronousRun FAILED,Test-Failures,"```
+io.reactivex.internal.operators.flowable.FlowableConcatMapEagerTest > testAsynchronousRun FAILED
+ java.lang.AssertionError: Value counts differ; Expected: 2000, Actual: 1184 (latch = 1, values = 1184, errors = 0, completions = 0)
+ at io.reactivex.observers.BaseTestConsumer.fail(BaseTestConsumer.java:132)
+ at io.reactivex.observers.BaseTestConsumer.assertValueCount(BaseTestConsumer.java:389)
+ at io.reactivex.internal.operators.flowable.FlowableConcatMapEagerTest.testAsynchronousRun(FlowableConcatMapEagerTest.java:583)
+```",1.0,"FlowableConcatMapEagerTest > testAsynchronousRun FAILED - ```
+io.reactivex.internal.operators.flowable.FlowableConcatMapEagerTest > testAsynchronousRun FAILED
+ java.lang.AssertionError: Value counts differ; Expected: 2000, Actual: 1184 (latch = 1, values = 1184, errors = 0, completions = 0)
+ at io.reactivex.observers.BaseTestConsumer.fail(BaseTestConsumer.java:132)
+ at io.reactivex.observers.BaseTestConsumer.assertValueCount(BaseTestConsumer.java:389)
+ at io.reactivex.internal.operators.flowable.FlowableConcatMapEagerTest.testAsynchronousRun(FlowableConcatMapEagerTest.java:583)
+```",0,flowableconcatmapeagertest testasynchronousrun failed io reactivex internal operators flowable flowableconcatmapeagertest testasynchronousrun failed java lang assertionerror value counts differ expected actual latch values errors completions at io reactivex observers basetestconsumer fail basetestconsumer java at io reactivex observers basetestconsumer assertvaluecount basetestconsumer java at io reactivex internal operators flowable flowableconcatmapeagertest testasynchronousrun flowableconcatmapeagertest java ,0
+897,18732446312.0,IssuesEvent,2021-11-04 00:13:28,Andidy/engine-v2,https://api.github.com/repos/Andidy/engine-v2,closed,Add component indices to the constructor of struct Entity,new people implmentation,"`struct Entity` has indices into the `std::vector`s in `struct GameState`. These should be populated inside the constructor of `struct Entity`.
+
+Here is an example of the `struct Entity` constructor with the components: cTransform and cGridTransform already filled out.
+
+https://github.com/Andidy/engine-v2/blob/d3f47492e45b14edad01d5dbbcfa1d88cdd094d3/engine-v2/src/entity.h#L24-L29",1.0,"Add component indices to the constructor of struct Entity - `struct Entity` has indices into the `std::vector`s in `struct GameState`. These should be populated inside the constructor of `struct Entity`.
+
+Here is an example of the `struct Entity` constructor with the components: cTransform and cGridTransform already filled out.
+
+https://github.com/Andidy/engine-v2/blob/d3f47492e45b14edad01d5dbbcfa1d88cdd094d3/engine-v2/src/entity.h#L24-L29",1,add component indices to the constructor of struct entity struct entity has indices into the std vector s in struct gamestate these should be populated inside the constructor of struct entity here is an example of the struct entity constructor with the components ctransform and cgridtransform already filled out ,1
+505,2524513055.0,IssuesEvent,2015-01-20 18:15:26,xunit/xunit,https://api.github.com/repos/xunit/xunit,closed,List all dependencies on the wiki or in the source so all projects can be listed,Documentation,"Hi,
+
+I tried to open the solution in VS 2013 and got the following message: ""The project is targeting frameworks that are either not installed or are included as part of future updates to Visual Studio. See http://go.microsoft.com/fwlink/?LinkID=287985 for more information."" This is not really helpful and still don't know what I'm missing on my machine. As work around I simply removed TargetProfile and ProjectTypeGuids from the project files.
+
+Can you list all required dependencies?
+
+The projects that failed to open: xunit.assert, xunit.core, xunit.abstractsion.pcl
+
+Thanks
+
+awel",1.0,"List all dependencies on the wiki or in the source so all projects can be listed - Hi,
+
+I tried to open the solution in VS 2013 and got the following message: ""The project is targeting frameworks that are either not installed or are included as part of future updates to Visual Studio. See http://go.microsoft.com/fwlink/?LinkID=287985 for more information."" This is not really helpful and still don't know what I'm missing on my machine. As work around I simply removed TargetProfile and ProjectTypeGuids from the project files.
+
+Can you list all required dependencies?
+
+The projects that failed to open: xunit.assert, xunit.core, xunit.abstractsion.pcl
+
+Thanks
+
+awel",0,list all dependencies on the wiki or in the source so all projects can be listed hi i tried to open the solution in vs and got the following message the project is targeting frameworks that are either not installed or are included as part of future updates to visual studio see for more information this is not really helpful and still don t know what i m missing on my machine as work around i simply removed targetprofile and projecttypeguids from the project files can you list all required dependencies the projects that failed to open xunit assert xunit core xunit abstractsion pcl thanks awel,0
+15423,10019211847.0,IssuesEvent,2019-07-16 09:35:20,godotengine/godot,https://api.github.com/repos/godotengine/godot,closed,Make 3D viewport locked nodes unselectable as in 2D viewport,enhancement junior job topic:editor usability,"**Godot version:**
+3.2 master (43a9b8c76)
+
+
+**OS/device including version:**
+Any
+
+
+**Issue description:**
+Follow-up to discussion in #26944.
+
+> So there has been some confusion here as the feature partly exists already for children of a locked nodes, which are not selectable.
+>
+> What should be changed is to make things behave like in 2D, where locking also makes the locked node unselectable. The code should be kept as similar as possible as the features and behaviour should be the same for the 2D and 3D viewports. The tooltips in 2D and 3D should be adjusted to mention that they lock the position and prevent selection in the viewport.
+>
+> In a later PR, it could be interesting to implement a picking mode for 2D and 3D which bypasses the selection restriction, so e.g. Ctrl+Click would let you select locked nodes (to ease finding them in the scene tree/viewing their properties/unlocking them).
+
+**Steps to reproduce:**
+- Create 2D scene with a node and a child
+- Lock the parent node, see that neither parent nor child can be selected by picking in the viewport
+- Create 3D scene with a node and a child
+- Lock the parent node, see that the child cannot be selected by picking in the viewport, but the parent can still be",True,"Make 3D viewport locked nodes unselectable as in 2D viewport - **Godot version:**
+3.2 master (43a9b8c76)
+
+
+**OS/device including version:**
+Any
+
+
+**Issue description:**
+Follow-up to discussion in #26944.
+
+> So there has been some confusion here as the feature partly exists already for children of a locked nodes, which are not selectable.
+>
+> What should be changed is to make things behave like in 2D, where locking also makes the locked node unselectable. The code should be kept as similar as possible as the features and behaviour should be the same for the 2D and 3D viewports. The tooltips in 2D and 3D should be adjusted to mention that they lock the position and prevent selection in the viewport.
+>
+> In a later PR, it could be interesting to implement a picking mode for 2D and 3D which bypasses the selection restriction, so e.g. Ctrl+Click would let you select locked nodes (to ease finding them in the scene tree/viewing their properties/unlocking them).
+
+**Steps to reproduce:**
+- Create 2D scene with a node and a child
+- Lock the parent node, see that neither parent nor child can be selected by picking in the viewport
+- Create 3D scene with a node and a child
+- Lock the parent node, see that the child cannot be selected by picking in the viewport, but the parent can still be",0,make viewport locked nodes unselectable as in viewport godot version master os device including version any issue description follow up to discussion in so there has been some confusion here as the feature partly exists already for children of a locked nodes which are not selectable what should be changed is to make things behave like in where locking also makes the locked node unselectable the code should be kept as similar as possible as the features and behaviour should be the same for the and viewports the tooltips in and should be adjusted to mention that they lock the position and prevent selection in the viewport in a later pr it could be interesting to implement a picking mode for and which bypasses the selection restriction so e g ctrl click would let you select locked nodes to ease finding them in the scene tree viewing their properties unlocking them steps to reproduce create scene with a node and a child lock the parent node see that neither parent nor child can be selected by picking in the viewport create scene with a node and a child lock the parent node see that the child cannot be selected by picking in the viewport but the parent can still be,0
+907,18950125347.0,IssuesEvent,2021-11-18 14:25:22,omigroup/omigroup,https://api.github.com/repos/omigroup/omigroup,closed,Align backlog tags with OMI core values,enhancement Make the metaverse more human Consistently deliver value Empower the people making the metaverse Cultivate Resiliency Maintain sustainable innovation,What if we added tags for each of our 5 core values?* That way our tasks / stories / epics can display how that work aligns with one of our core values. I can demonstrate an example:,1.0,Align backlog tags with OMI core values - What if we added tags for each of our 5 core values?* That way our tasks / stories / epics can display how that work aligns with one of our core values. I can demonstrate an example:,1,align backlog tags with omi core values what if we added tags for each of our core values that way our tasks stories epics can display how that work aligns with one of our core values i can demonstrate an example ,1
+59615,14620207741.0,IssuesEvent,2020-12-22 19:17:08,TheSLinux/gs,https://api.github.com/repos/TheSLinux/gs,closed,empty strings $provides and $conflicts makes #makepkg stupid,_backlog _critical _important buildystem clean_up_2020,"Scope:
+- `pacman=4.1.6`
+- `_utils=2.2.0`
+
+Wrong thing:
+
+```
+provides=(""${provides[@]:-}"" ""${PACKAGE_BASE}@lib32"")
+conflicts=(""${conflicts[@]:-}"" ""${PACKAGE_BASE}@lib32"")
+```
+
+If `${provides[@]}` is currently empty, we will see in `pkg/.PKGINFO`
+
+```
+conflict =
+conflict = libdatrie@lib32
+provides =
+provides = libdatrie@lib32
+```
+
+and the final package knows nothing about its dependencies.
+
+Right thing:
+
+```
+provides=(""${PACKAGE_BASE}@lib32"" ""${provides[@]:-}"")
+conflicts=(""${PACKAGE_BASE}@lib32"" ""${conflicts[@]:-}"")
+```
+
+Within this fix, we will have in the `pkg/.PKGINFO`
+
+```
+conflict =
+conflict = libdatrie@lib32
+provides =
+provides = libdatrie@lib32
+```
+
+and the final package has correct dependency information.
+
+This should be a bug in `makepkg` that doesn't check for input.
+",1.0,"empty strings $provides and $conflicts makes #makepkg stupid - Scope:
+- `pacman=4.1.6`
+- `_utils=2.2.0`
+
+Wrong thing:
+
+```
+provides=(""${provides[@]:-}"" ""${PACKAGE_BASE}@lib32"")
+conflicts=(""${conflicts[@]:-}"" ""${PACKAGE_BASE}@lib32"")
+```
+
+If `${provides[@]}` is currently empty, we will see in `pkg/.PKGINFO`
+
+```
+conflict =
+conflict = libdatrie@lib32
+provides =
+provides = libdatrie@lib32
+```
+
+and the final package knows nothing about its dependencies.
+
+Right thing:
+
+```
+provides=(""${PACKAGE_BASE}@lib32"" ""${provides[@]:-}"")
+conflicts=(""${PACKAGE_BASE}@lib32"" ""${conflicts[@]:-}"")
+```
+
+Within this fix, we will have in the `pkg/.PKGINFO`
+
+```
+conflict =
+conflict = libdatrie@lib32
+provides =
+provides = libdatrie@lib32
+```
+
+and the final package has correct dependency information.
+
+This should be a bug in `makepkg` that doesn't check for input.
+",0,empty strings provides and conflicts makes makepkg stupid scope pacman utils wrong thing provides provides package base conflicts conflicts package base if provides is currently empty we will see in pkg pkginfo conflict conflict libdatrie provides provides libdatrie and the final package knows nothing about its dependencies right thing provides package base provides conflicts package base conflicts within this fix we will have in the pkg pkginfo conflict conflict libdatrie provides provides libdatrie and the final package has correct dependency information this should be a bug in makepkg that doesn t check for input ,0
+343,6544132497.0,IssuesEvent,2017-09-03 12:06:13,newtheatre/history-project,https://api.github.com/repos/newtheatre/history-project,closed,Twitter feeds on people records,discussion people,"For someone's bio entry should we embed their twitter feed on the page?
+",1.0,"Twitter feeds on people records - For someone's bio entry should we embed their twitter feed on the page?
+",1,twitter feeds on people records for someone s bio entry should we embed their twitter feed on the page ,1
+577,10243786235.0,IssuesEvent,2019-08-20 08:57:53,ushahidi/tenfour,https://api.github.com/repos/ushahidi/tenfour,opened,Owners can add contacts to their organization from a single button #618,Feature: People P2 - Normal,"Owners & admin can add contacts from a single button that has the following options:
+
+- Owners & Admin can add People manually
+- Owners can add contacts via CSV
+- Owners & Admin can invite email addresses in bulk to join the organization without adding contacts
+- Users invited by email will not appear in the People list or count against the # of users per plan until they’ve added their contacts
+- Owners can set permissions for contacts being added
+- Owners can also invite People to become Users when they are invited
+- Invited user can add their contact information to an org before they are required to create their own account and password",1.0,"Owners can add contacts to their organization from a single button #618 - Owners & admin can add contacts from a single button that has the following options:
+
+- Owners & Admin can add People manually
+- Owners can add contacts via CSV
+- Owners & Admin can invite email addresses in bulk to join the organization without adding contacts
+- Users invited by email will not appear in the People list or count against the # of users per plan until they’ve added their contacts
+- Owners can set permissions for contacts being added
+- Owners can also invite People to become Users when they are invited
+- Invited user can add their contact information to an org before they are required to create their own account and password",1,owners can add contacts to their organization from a single button owners admin can add contacts from a single button that has the following options owners admin can add people manually owners can add contacts via csv owners admin can invite email addresses in bulk to join the organization without adding contacts users invited by email will not appear in the people list or count against the of users per plan until they’ve added their contacts owners can set permissions for contacts being added owners can also invite people to become users when they are invited invited user can add their contact information to an org before they are required to create their own account and password,1
+107766,16762299575.0,IssuesEvent,2021-06-14 01:33:02,rzr/rzr-presentation-gstreamer,https://api.github.com/repos/rzr/rzr-presentation-gstreamer,opened,CVE-2020-36048 (High) detected in engine.io-1.5.4.tgz,security vulnerability,"## CVE-2020-36048 - High Severity Vulnerability
+ Vulnerable Library - engine.io-1.5.4.tgz
+
+
The realtime engine behind Socket.IO. Provides the foundation of a bidirectional connection between client and server
Path to dependency file: rzr-presentation-gstreamer/reveal.js-master/plugin/multiplex/package.json
+
Path to vulnerable library: rzr-presentation-gstreamer/reveal.js-master/node_modules/engine.io/package.json,rzr-presentation-gstreamer/reveal.js-master/node_modules/engine.io/package.json
+
+
+
+
+
+***
+Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",True,"CVE-2020-36048 (High) detected in engine.io-1.5.4.tgz - ## CVE-2020-36048 - High Severity Vulnerability
+ Vulnerable Library - engine.io-1.5.4.tgz
+
+
The realtime engine behind Socket.IO. Provides the foundation of a bidirectional connection between client and server
Path to dependency file: rzr-presentation-gstreamer/reveal.js-master/plugin/multiplex/package.json
+
Path to vulnerable library: rzr-presentation-gstreamer/reveal.js-master/node_modules/engine.io/package.json,rzr-presentation-gstreamer/reveal.js-master/node_modules/engine.io/package.json
+
+
+
+
+
+***
+Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",0,cve high detected in engine io tgz cve high severity vulnerability vulnerable library engine io tgz the realtime engine behind socket io provides the foundation of a bidirectional connection between client and server library home page a href path to dependency file rzr presentation gstreamer reveal js master plugin multiplex package json path to vulnerable library rzr presentation gstreamer reveal js master node modules engine io package json rzr presentation gstreamer reveal js master node modules engine io package json dependency hierarchy socket io tgz root library x engine io tgz vulnerable library found in base branch master vulnerability details engine io before allows attackers to cause a denial of service resource consumption via a post request to the long polling transport publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution engine io step up your open source security game with whitesource ,0
+387121,26713583743.0,IssuesEvent,2023-01-28 07:21:41,extratone/mastodon-ios-apps,https://api.github.com/repos/extratone/mastodon-ios-apps,opened,Tootsuite as Installable PWA,documentation,"With Tootsuite open in Safari, open the share sheet, select `Add to Home Screen`, and this is the result.
+
+https://user-images.githubusercontent.com/43663476/215252875-3a0972bf-5271-48e8-a257-9c89c838dcb1.MOV",1.0,"Tootsuite as Installable PWA - With Tootsuite open in Safari, open the share sheet, select `Add to Home Screen`, and this is the result.
+
+https://user-images.githubusercontent.com/43663476/215252875-3a0972bf-5271-48e8-a257-9c89c838dcb1.MOV",0,tootsuite as installable pwa with tootsuite open in safari open the share sheet select add to home screen and this is the result ,0
+145063,22604199450.0,IssuesEvent,2022-06-29 11:53:12,openshift-pipelines/pipelines-as-code,https://api.github.com/repos/openshift-pipelines/pipelines-as-code,closed,Configuring webhook using pac cli,feature design,"Currently, we have support for configuring GitHub App in the bootstrap command.
+But if user want to configure webhook the process is manual.
+we want to automate the possible steps using the pac cli.
+
+
+
+
+(this is one of the poc done.)
+
+the next things do we want this as a
+- separate command, or
+- include this as a part of `repo create cmd` (favourable as discussed on the slack)
+
+if we add this as a part of `tkn-pac repo create` command
+then it need to know if `GitHub App` is already configured ?
+
+Why?
+- If GitHub App is already configured then repo create cmd doesn't need to provide
+option to configure webhook
+
+How will it know cli know if GH App is already configured?
+(should also work for non admin users)
+
+- May be a configmap with controller url and the provider configured
+- Who will create and who will update it?
+
+
+
+
+",1.0,"Configuring webhook using pac cli - Currently, we have support for configuring GitHub App in the bootstrap command.
+But if user want to configure webhook the process is manual.
+we want to automate the possible steps using the pac cli.
+
+
+
+
+(this is one of the poc done.)
+
+the next things do we want this as a
+- separate command, or
+- include this as a part of `repo create cmd` (favourable as discussed on the slack)
+
+if we add this as a part of `tkn-pac repo create` command
+then it need to know if `GitHub App` is already configured ?
+
+Why?
+- If GitHub App is already configured then repo create cmd doesn't need to provide
+option to configure webhook
+
+How will it know cli know if GH App is already configured?
+(should also work for non admin users)
+
+- May be a configmap with controller url and the provider configured
+- Who will create and who will update it?
+
+
+
+
+",0,configuring webhook using pac cli currently we have support for configuring github app in the bootstrap command but if user want to configure webhook the process is manual we want to automate the possible steps using the pac cli this is one of the poc done the next things do we want this as a separate command or include this as a part of repo create cmd favourable as discussed on the slack if we add this as a part of tkn pac repo create command then it need to know if github app is already configured why if github app is already configured then repo create cmd doesn t need to provide option to configure webhook how will it know cli know if gh app is already configured should also work for non admin users may be a configmap with controller url and the provider configured who will create and who will update it ,0
+568797,16988932350.0,IssuesEvent,2021-06-30 17:42:16,robotframework/robotframework,https://api.github.com/repos/robotframework/robotframework,closed,Libdoc stores data type documentation with extra indentation,bug priority: medium,Libdoc started handling enums and TypedDicts specially in RF 4.0 (#3607 and #3783). Their documentation is got directly by reading their `__doc__` which leaves all indentation in place. This is easy to fix by using [inspect.getdoc](https://python-docs-34-korean.readthedocs.io/en/latest/library/inspect.html#inspect.getdoc) that cleans up the documentation automatically.,1.0,Libdoc stores data type documentation with extra indentation - Libdoc started handling enums and TypedDicts specially in RF 4.0 (#3607 and #3783). Their documentation is got directly by reading their `__doc__` which leaves all indentation in place. This is easy to fix by using [inspect.getdoc](https://python-docs-34-korean.readthedocs.io/en/latest/library/inspect.html#inspect.getdoc) that cleans up the documentation automatically.,0,libdoc stores data type documentation with extra indentation libdoc started handling enums and typeddicts specially in rf and their documentation is got directly by reading their doc which leaves all indentation in place this is easy to fix by using that cleans up the documentation automatically ,0
+803,14935960308.0,IssuesEvent,2021-01-25 12:43:37,googleapis/google-api-python-client,https://api.github.com/repos/googleapis/google-api-python-client,closed,"googleapiclient in python, using updateContact returns 403",api: people type: question,"Hi, I'm trying to get a list of contacts from Google's API, and then update some of them. While running the `updateContact` function I get 403 error `Request person.etag is different than the current person.etag. Clear local cache and get the latest person.""` every time.
+
+#### Environment details
+ Using Jupyter Notebook
+ - OS type and version: ubuntu 18 (wsl)
+ - Python version: Python 3.6.9
+ - pip version: pip 9.0.1
+ - google-api-python-client version: 1.12.8
+
+#### Code example
+
+Im reading the contact's etag using this code:
+```
+service = build('people', 'v1', credentials=creds)
+results = service.people().connections().list(
+ resourceName='people/me',
+ pageSize=100,
+ personFields='names,emailAddresses').execute()
+connections = results.get('connections', [])
+```
+Then for each `contact` in the `connections`, Im trying yo change the contact givenName and familyName:
+```
+service.people().updateContact(resourceName=""contact's resource name"",
+ updatePersonFields=""names"",
+ body={
+ ""etag"": ""contact's etag"",
+ ""names"": [
+ {
+ ""familyName"": ""new family name"",
+ ""givenName"": ""new given name""
+ }
+ ]
+ }
+ ).execute()
+```
+The contact's `etag` and `resourceName` are taken from the connection's list response. for example:
+```
+print(connections[0])
+{'resourceName': 'people/',
+ 'etag': 'etag string',
+ 'names': [{'metadata': {'primary': True,
+ 'source': {'type': 'CONTACT', 'id': 'id number'}},
+ 'displayName': 'x',
+ 'familyName': 'x',
+ 'givenName': 'x',
+ 'displayNameLastFirst': 'x',
+ 'unstructuredName': 'x'}]}
+```
+
+#### Stack trace
+```
+~/.local/lib/python3.6/site-packages/googleapiclient/_helpers.py in positional_wrapper(*args, **kwargs)
+ 132 elif positional_parameters_enforcement == POSITIONAL_WARNING:
+ 133 logger.warning(message)
+--> 134 return wrapped(*args, **kwargs)
+ 135
+ 136 return positional_wrapper
+
+~/.local/lib/python3.6/site-packages/googleapiclient/http.py in execute(self, http, num_retries)
+ 913 callback(resp)
+ 914 if resp.status >= 300:
+--> 915 raise HttpError(resp, content, uri=self.uri)
+ 916 return self.postproc(resp, content)
+ 917
+
+HttpError: :updateContact?updatePersonFields=names&alt=json returned ""Request person.etag is different than the current person.etag. Clear local cache and get the latest person."". Details: ""Request person.etag is different than the current person.etag. Clear local cache and get the latest person."">
+
+```",1.0,"googleapiclient in python, using updateContact returns 403 - Hi, I'm trying to get a list of contacts from Google's API, and then update some of them. While running the `updateContact` function I get 403 error `Request person.etag is different than the current person.etag. Clear local cache and get the latest person.""` every time.
+
+#### Environment details
+ Using Jupyter Notebook
+ - OS type and version: ubuntu 18 (wsl)
+ - Python version: Python 3.6.9
+ - pip version: pip 9.0.1
+ - google-api-python-client version: 1.12.8
+
+#### Code example
+
+Im reading the contact's etag using this code:
+```
+service = build('people', 'v1', credentials=creds)
+results = service.people().connections().list(
+ resourceName='people/me',
+ pageSize=100,
+ personFields='names,emailAddresses').execute()
+connections = results.get('connections', [])
+```
+Then for each `contact` in the `connections`, Im trying yo change the contact givenName and familyName:
+```
+service.people().updateContact(resourceName=""contact's resource name"",
+ updatePersonFields=""names"",
+ body={
+ ""etag"": ""contact's etag"",
+ ""names"": [
+ {
+ ""familyName"": ""new family name"",
+ ""givenName"": ""new given name""
+ }
+ ]
+ }
+ ).execute()
+```
+The contact's `etag` and `resourceName` are taken from the connection's list response. for example:
+```
+print(connections[0])
+{'resourceName': 'people/',
+ 'etag': 'etag string',
+ 'names': [{'metadata': {'primary': True,
+ 'source': {'type': 'CONTACT', 'id': 'id number'}},
+ 'displayName': 'x',
+ 'familyName': 'x',
+ 'givenName': 'x',
+ 'displayNameLastFirst': 'x',
+ 'unstructuredName': 'x'}]}
+```
+
+#### Stack trace
+```
+~/.local/lib/python3.6/site-packages/googleapiclient/_helpers.py in positional_wrapper(*args, **kwargs)
+ 132 elif positional_parameters_enforcement == POSITIONAL_WARNING:
+ 133 logger.warning(message)
+--> 134 return wrapped(*args, **kwargs)
+ 135
+ 136 return positional_wrapper
+
+~/.local/lib/python3.6/site-packages/googleapiclient/http.py in execute(self, http, num_retries)
+ 913 callback(resp)
+ 914 if resp.status >= 300:
+--> 915 raise HttpError(resp, content, uri=self.uri)
+ 916 return self.postproc(resp, content)
+ 917
+
+HttpError: :updateContact?updatePersonFields=names&alt=json returned ""Request person.etag is different than the current person.etag. Clear local cache and get the latest person."". Details: ""Request person.etag is different than the current person.etag. Clear local cache and get the latest person."">
+
+```",1,googleapiclient in python using updatecontact returns hi i m trying to get a list of contacts from google s api and then update some of them while running the updatecontact function i get error request person etag is different than the current person etag clear local cache and get the latest person every time environment details using jupyter notebook os type and version ubuntu wsl python version python pip version pip google api python client version code example im reading the contact s etag using this code service build people credentials creds results service people connections list resourcename people me pagesize personfields names emailaddresses execute connections results get connections then for each contact in the connections im trying yo change the contact givenname and familyname service people updatecontact resourcename contact s resource name updatepersonfields names body etag contact s etag names familyname new family name givenname new given name execute the contact s etag and resourcename are taken from the connection s list response for example print connections resourcename people etag etag string names metadata primary true source type contact id id number displayname x familyname x givenname x displaynamelastfirst x unstructuredname x stack trace local lib site packages googleapiclient helpers py in positional wrapper args kwargs elif positional parameters enforcement positional warning logger warning message return wrapped args kwargs return positional wrapper local lib site packages googleapiclient http py in execute self http num retries callback resp if resp status raise httperror resp content uri self uri return self postproc resp content httperror ,1
+899,18737573023.0,IssuesEvent,2021-11-04 09:40:41,PostHog/posthog,https://api.github.com/repos/PostHog/posthog,closed,"Don't allow identifying ""non-anonomous"" user -> ""non-anonomous"" user",enhancement people data-ingestion,"## Is your feature request related to a problem?
+
+Currently users can end up with massive merged users if they do something like ""posthog.identify(null)"" or something for logged-out users. Worse, it can lead to data corruption if in a B2B product one user logs in as another.
+
+## Describe the solution you'd like
+
+Proposal:
+If user has been ""identified""/is non-anonymous, don't allow them to be merged again into another user.
+
+Posthog-js has this guard: https://github.com/PostHog/posthog-js/blob/cd9ec768d0662a82469d200eb7ff23a9dcda7c98/src/posthog-core.js#L842-L864
+
+However other libraries and integrations (*cough*rudderstack*cough*) does not
+
+## Describe alternatives you've considered
+
+Teaching all users how to use tools perfectly from day one.
+
+## Additional context
+
+Related issue: https://github.com/PostHog/product-internal/issues/132
+
+cc @timgl we discussed this in meeting yesterday
+cc @yakkomajuri for ingestion woes
+cc @EDsCODE who created the linked issue
+
+This was also a pain at other companies with similar apis - teaching users is impossible
+
+#### *Thank you* for your feature request – we love each and every one!
+",1.0,"Don't allow identifying ""non-anonomous"" user -> ""non-anonomous"" user - ## Is your feature request related to a problem?
+
+Currently users can end up with massive merged users if they do something like ""posthog.identify(null)"" or something for logged-out users. Worse, it can lead to data corruption if in a B2B product one user logs in as another.
+
+## Describe the solution you'd like
+
+Proposal:
+If user has been ""identified""/is non-anonymous, don't allow them to be merged again into another user.
+
+Posthog-js has this guard: https://github.com/PostHog/posthog-js/blob/cd9ec768d0662a82469d200eb7ff23a9dcda7c98/src/posthog-core.js#L842-L864
+
+However other libraries and integrations (*cough*rudderstack*cough*) does not
+
+## Describe alternatives you've considered
+
+Teaching all users how to use tools perfectly from day one.
+
+## Additional context
+
+Related issue: https://github.com/PostHog/product-internal/issues/132
+
+cc @timgl we discussed this in meeting yesterday
+cc @yakkomajuri for ingestion woes
+cc @EDsCODE who created the linked issue
+
+This was also a pain at other companies with similar apis - teaching users is impossible
+
+#### *Thank you* for your feature request – we love each and every one!
+",1,don t allow identifying non anonomous user non anonomous user is your feature request related to a problem currently users can end up with massive merged users if they do something like posthog identify null or something for logged out users worse it can lead to data corruption if in a product one user logs in as another describe the solution you d like proposal if user has been identified is non anonymous don t allow them to be merged again into another user posthog js has this guard however other libraries and integrations cough rudderstack cough does not describe alternatives you ve considered teaching all users how to use tools perfectly from day one additional context related issue cc timgl we discussed this in meeting yesterday cc yakkomajuri for ingestion woes cc edscode who created the linked issue this was also a pain at other companies with similar apis teaching users is impossible thank you for your feature request – we love each and every one ,1
+1500,2772247728.0,IssuesEvent,2015-05-02 13:46:54,gregorio-project/gregorio,https://api.github.com/repos/gregorio-project/gregorio,closed,make dist failure,bug build/install,"Trying to build source package for debian, I've encountered this problem on develop branch : if you try `make dist`, here is the result:
+
+ make dist-bzip2 am__post_remove_distdir='@:'
+ make[1]: Entering directory '/root/greg/gregorio-dev'
+ if test -d ""gregorio-4.0.0-rc1""; then find ""gregorio-4.0.0-rc1"" -type d ! -perm -200 -exec chmod u+w {} ';' && rm -rf ""gregorio-4.0.0-rc1"" || { sleep 5 && rm -rf ""gregorio-4.0.0-rc1""; }; else :; fi
+ test -d ""gregorio-4.0.0-rc1"" || mkdir ""gregorio-4.0.0-rc1""
+ (cd src && make top_distdir=../gregorio-4.0.0-rc1 distdir=../gregorio-4.0.0-rc1/src \
+ am__remove_distdir=: am__skip_length_check=: am__skip_mode_fix=: distdir)
+ make[2]: Entering directory '/root/greg/gregorio-dev/src'
+ make[2]: *** No rule to make target 'dump/dump.h', needed by 'distdir'. Arrêt.
+ make[2]: Leaving directory '/root/greg/gregorio-dev/src'
+ Makefile:455: recipe for target 'distdir' failed
+ make[1]: *** [distdir] Error 1
+ make[1]: Leaving directory '/root/greg/gregorio-dev'
+ Makefile:553: recipe for target 'dist' failed
+ make: *** [dist] Error 2
+",1.0,"make dist failure - Trying to build source package for debian, I've encountered this problem on develop branch : if you try `make dist`, here is the result:
+
+ make dist-bzip2 am__post_remove_distdir='@:'
+ make[1]: Entering directory '/root/greg/gregorio-dev'
+ if test -d ""gregorio-4.0.0-rc1""; then find ""gregorio-4.0.0-rc1"" -type d ! -perm -200 -exec chmod u+w {} ';' && rm -rf ""gregorio-4.0.0-rc1"" || { sleep 5 && rm -rf ""gregorio-4.0.0-rc1""; }; else :; fi
+ test -d ""gregorio-4.0.0-rc1"" || mkdir ""gregorio-4.0.0-rc1""
+ (cd src && make top_distdir=../gregorio-4.0.0-rc1 distdir=../gregorio-4.0.0-rc1/src \
+ am__remove_distdir=: am__skip_length_check=: am__skip_mode_fix=: distdir)
+ make[2]: Entering directory '/root/greg/gregorio-dev/src'
+ make[2]: *** No rule to make target 'dump/dump.h', needed by 'distdir'. Arrêt.
+ make[2]: Leaving directory '/root/greg/gregorio-dev/src'
+ Makefile:455: recipe for target 'distdir' failed
+ make[1]: *** [distdir] Error 1
+ make[1]: Leaving directory '/root/greg/gregorio-dev'
+ Makefile:553: recipe for target 'dist' failed
+ make: *** [dist] Error 2
+",0,make dist failure trying to build source package for debian i ve encountered this problem on develop branch if you try make dist here is the result make dist am post remove distdir make entering directory root greg gregorio dev if test d gregorio then find gregorio type d perm exec chmod u w rm rf gregorio sleep rm rf gregorio else fi test d gregorio mkdir gregorio cd src make top distdir gregorio distdir gregorio src am remove distdir am skip length check am skip mode fix distdir make entering directory root greg gregorio dev src make no rule to make target dump dump h needed by distdir arrêt make leaving directory root greg gregorio dev src makefile recipe for target distdir failed make error make leaving directory root greg gregorio dev makefile recipe for target dist failed make error ,0
+544,9621967550.0,IssuesEvent,2019-05-14 11:59:31,newtheatre/history-project,https://api.github.com/repos/newtheatre/history-project,closed,Broken Arthur,bug people report-tool,"End user submitted issue from page: [/people/arthur_mckechnie/](https://history.newtheatre.org.uk/people/arthur_mckechnie/)
+---
+Links from other shows brings me to Arthur but his page only displays Stags to me
+*Zoe Smith*
+",1.0,"Broken Arthur - End user submitted issue from page: [/people/arthur_mckechnie/](https://history.newtheatre.org.uk/people/arthur_mckechnie/)
+---
+Links from other shows brings me to Arthur but his page only displays Stags to me
+*Zoe Smith*
+",1,broken arthur end user submitted issue from page links from other shows brings me to arthur but his page only displays stags to me zoe smith ,1
+496,8556003803.0,IssuesEvent,2018-11-08 11:46:15,folivoraAI/BetterTouchTool,https://api.github.com/repos/folivoraAI/BetterTouchTool,closed,MBP 2016 brightness increasing automatically with extended display,Need other people who can reproduce the issue bug,"
+#### Description of bug/feature request/question:
+Whenever I connect (or disconnect) my MacBook Pro 2016 with Touch Bar to an extended display, its brightness increases one or multiple blocks. This happened after upgrading to High Sierra. The steps to reproduce are:
+Put your brightness to a certain level.
+Connect an extended display (I tested with multiple displays, HDMI to USB-C or VGA to USB-C)
+Check if the brightness increased or not.
+
+After going through multiple rounds with Apple Support (SMC and NVRAM Reset, macOS reinstall), I notice this happens only when BTT was active. It'll be of immense help if you can take a look into this!
+
+#### Affected input device (e.g. MacBook Trackpad, Magic Mouse/Trackpad, Touch Bar, etc.):
+Macbook
+
+#### Device information:
+
+Type of Mac: MacBook Pro 2016 with Touch Bar
+macOS version: 10.13
+BetterTouchTool version: 2.310
+",1.0,"MBP 2016 brightness increasing automatically with extended display -
+#### Description of bug/feature request/question:
+Whenever I connect (or disconnect) my MacBook Pro 2016 with Touch Bar to an extended display, its brightness increases one or multiple blocks. This happened after upgrading to High Sierra. The steps to reproduce are:
+Put your brightness to a certain level.
+Connect an extended display (I tested with multiple displays, HDMI to USB-C or VGA to USB-C)
+Check if the brightness increased or not.
+
+After going through multiple rounds with Apple Support (SMC and NVRAM Reset, macOS reinstall), I notice this happens only when BTT was active. It'll be of immense help if you can take a look into this!
+
+#### Affected input device (e.g. MacBook Trackpad, Magic Mouse/Trackpad, Touch Bar, etc.):
+Macbook
+
+#### Device information:
+
+Type of Mac: MacBook Pro 2016 with Touch Bar
+macOS version: 10.13
+BetterTouchTool version: 2.310
+",1,mbp brightness increasing automatically with extended display description of bug feature request question whenever i connect or disconnect my macbook pro with touch bar to an extended display its brightness increases one or multiple blocks this happened after upgrading to high sierra the steps to reproduce are put your brightness to a certain level connect an extended display i tested with multiple displays hdmi to usb c or vga to usb c check if the brightness increased or not after going through multiple rounds with apple support smc and nvram reset macos reinstall i notice this happens only when btt was active it ll be of immense help if you can take a look into this affected input device e g macbook trackpad magic mouse trackpad touch bar etc macbook device information type of mac macbook pro with touch bar macos version bettertouchtool version ,1
+600490,18297835830.0,IssuesEvent,2021-10-05 22:24:50,usc-isi-i2/kgtk,https://api.github.com/repos/usc-isi-i2/kgtk,closed,`kgtk unique --min-count nnn --max-count mmm`: Filter the Results,enhancement priority 1,Filter the unique results by specifying min and max thresholds. Min defaults to 0. Max defaults to something very large.,1.0,`kgtk unique --min-count nnn --max-count mmm`: Filter the Results - Filter the unique results by specifying min and max thresholds. Min defaults to 0. Max defaults to something very large.,0, kgtk unique min count nnn max count mmm filter the results filter the unique results by specifying min and max thresholds min defaults to max defaults to something very large ,0
+0,2534245333.0,IssuesEvent,2015-01-24 19:14:41,ufvivotech/ufDataQualityImprovement,https://api.github.com/repos/ufvivotech/ufDataQualityImprovement,closed,Develop photo uploader,People Software Dev,"Develop a command line photo loader -- scan a directory for files named with identifiers, match each to VIVO URI, upload to VIVO.
+
+Hmmm. We don't have access to the VIVO file system. ",1.0,"Develop photo uploader - Develop a command line photo loader -- scan a directory for files named with identifiers, match each to VIVO URI, upload to VIVO.
+
+Hmmm. We don't have access to the VIVO file system. ",1,develop photo uploader develop a command line photo loader scan a directory for files named with identifiers match each to vivo uri upload to vivo hmmm we don t have access to the vivo file system ,1
+56111,6960535929.0,IssuesEvent,2017-12-08 04:11:09,mozilla/network-pulse,https://api.github.com/repos/mozilla/network-pulse,reopened,Design profile tabs for favs,design,"We're going to add a User's Favs to their profile.
+
+The group decided that MVP version of this should include tabs so that Favs can be loaded separately to optimize how we engineer and serve this visitors.
+
+For now, tabs will just include Projects and Favs. Possibly extended later.
+
+ - [x] Review @kristinashu's [desktop comp](https://redpen.io/ns0835f2e5d9956858) design-wise. Update as needed.
+ - [x] Design mvp tabs for mobile
+ - [x] Review with engineers
+
+Note: Projects will include both Published and Created as currently implemented. Down the line, I'd like to mix this as a single list in the Projects tab, perhaps adding a simple icon to indicate something is publish (@gvn suggested maybe something like twitter's retweet icon). We won't tackle this right now engineering wise, staying focused on Favs.
+
+cc @gvn @mmmavis for awareness",1.0,"Design profile tabs for favs - We're going to add a User's Favs to their profile.
+
+The group decided that MVP version of this should include tabs so that Favs can be loaded separately to optimize how we engineer and serve this visitors.
+
+For now, tabs will just include Projects and Favs. Possibly extended later.
+
+ - [x] Review @kristinashu's [desktop comp](https://redpen.io/ns0835f2e5d9956858) design-wise. Update as needed.
+ - [x] Design mvp tabs for mobile
+ - [x] Review with engineers
+
+Note: Projects will include both Published and Created as currently implemented. Down the line, I'd like to mix this as a single list in the Projects tab, perhaps adding a simple icon to indicate something is publish (@gvn suggested maybe something like twitter's retweet icon). We won't tackle this right now engineering wise, staying focused on Favs.
+
+cc @gvn @mmmavis for awareness",0,design profile tabs for favs we re going to add a user s favs to their profile the group decided that mvp version of this should include tabs so that favs can be loaded separately to optimize how we engineer and serve this visitors for now tabs will just include projects and favs possibly extended later review kristinashu s design wise update as needed design mvp tabs for mobile review with engineers note projects will include both published and created as currently implemented down the line i d like to mix this as a single list in the projects tab perhaps adding a simple icon to indicate something is publish gvn suggested maybe something like twitter s retweet icon we won t tackle this right now engineering wise staying focused on favs cc gvn mmmavis for awareness,0
+185953,15038505664.0,IssuesEvent,2021-02-02 17:31:22,corona-warn-app/cwa-documentation,https://api.github.com/repos/corona-warn-app/cwa-documentation,closed,Spelling mistake in figure 9 of solution_architecture document,bug documentation mirrored-to-jira,"
+
+## Where to find the issue
+
+solution_architecture.md, figure 9
+
+## Describe the issue
+
+The box is named ""Event Notification Framework"", but should be ""Exposure Notification Framework""
+
+## Suggested change
+
+
+---
+Internal Tracking ID: EXPOSUREAPP-2717",1.0,"Spelling mistake in figure 9 of solution_architecture document -
+
+## Where to find the issue
+
+solution_architecture.md, figure 9
+
+## Describe the issue
+
+The box is named ""Event Notification Framework"", but should be ""Exposure Notification Framework""
+
+## Suggested change
+
+
+---
+Internal Tracking ID: EXPOSUREAPP-2717",0,spelling mistake in figure of solution architecture document thanks for reporting an issue in the documentation 🙌 ❤️ before opening a new issue please make sure that we do not have any duplicates already open you can ensure this by searching the issue list for this repository if there is a duplicate please close your issue and add a comment to the existing issue instead where to find the issue solution architecture md figure describe the issue the box is named event notification framework but should be exposure notification framework suggested change internal tracking id exposureapp ,0
+685,12287486813.0,IssuesEvent,2020-05-09 12:26:07,googleapis/elixir-google-api,https://api.github.com/repos/googleapis/elixir-google-api,opened,Synthesis failed for People,api: people autosynth failure priority: p1 type: bug,"Hello! Autosynth couldn't regenerate People. :broken_heart:
+
+Here's the output from running `synth.py`:
+
+```
+2020-05-09 05:16:57 [INFO] logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/elixir-google-api
+2020-05-09 05:16:57,351 autosynth > logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/elixir-google-api
+Switched to branch 'autosynth-people'
+2020-05-09 05:16:58 [INFO] Running synthtool
+2020-05-09 05:16:58,996 autosynth > Running synthtool
+2020-05-09 05:16:58 [INFO] ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/people/synth.metadata', 'synth.py', '--']
+2020-05-09 05:16:58,996 autosynth > ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/people/synth.metadata', 'synth.py', '--']
+2020-05-09 05:16:59,205 synthtool > Executing /home/kbuilder/.cache/synthtool/elixir-google-api/synth.py.
+On branch autosynth-people
+nothing to commit, working tree clean
+2020-05-09 05:16:59,513 synthtool > Cloning https://github.com/googleapis/elixir-google-api.git.
+2020-05-09 05:16:59,977 synthtool > Running: docker run --rm -v/home/kbuilder/.cache/synthtool/elixir-google-api:/workspace -v/var/run/docker.sock:/var/run/docker.sock -e USER_GROUP=1000:1000 -w /workspace gcr.io/cloud-devrel-public-resources/elixir19 scripts/generate_client.sh People
+2020-05-09 05:17:03,627 synthtool > No files in sources /home/kbuilder/.cache/synthtool/elixir-google-api/clients were copied. Does the source contain files?
+Traceback (most recent call last):
+ File ""/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py"", line 193, in _run_module_as_main
+ ""__main__"", mod_spec)
+ File ""/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py"", line 85, in _run_code
+ exec(code, run_globals)
+ File ""/tmpfs/src/github/synthtool/synthtool/__main__.py"", line 102, in
+ main()
+ File ""/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py"", line 829, in __call__
+ return self.main(*args, **kwargs)
+ File ""/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py"", line 782, in main
+ rv = self.invoke(ctx)
+ File ""/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py"", line 1066, in invoke
+ return ctx.invoke(self.callback, **ctx.params)
+ File ""/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py"", line 610, in invoke
+ return callback(*args, **kwargs)
+ File ""/tmpfs/src/github/synthtool/synthtool/__main__.py"", line 94, in main
+ spec.loader.exec_module(synth_module) # type: ignore
+ File ""/tmpfs/src/github/synthtool/synthtool/metadata.py"", line 180, in __exit__
+ write(self.metadata_file_path)
+ File ""/tmpfs/src/github/synthtool/synthtool/metadata.py"", line 112, in write
+ with open(outfile, ""w"") as fh:
+FileNotFoundError: [Errno 2] No such file or directory: 'clients/people/synth.metadata'
+2020-05-09 05:17:03 [ERROR] Synthesis failed
+2020-05-09 05:17:03,654 autosynth > Synthesis failed
+Traceback (most recent call last):
+ File ""/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py"", line 193, in _run_module_as_main
+ ""__main__"", mod_spec)
+ File ""/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py"", line 85, in _run_code
+ exec(code, run_globals)
+ File ""/tmpfs/src/github/synthtool/autosynth/synth.py"", line 599, in
+ main()
+ File ""/tmpfs/src/github/synthtool/autosynth/synth.py"", line 471, in main
+ return _inner_main(temp_dir)
+ File ""/tmpfs/src/github/synthtool/autosynth/synth.py"", line 549, in _inner_main
+ ).synthesize(base_synth_log_path)
+ File ""/tmpfs/src/github/synthtool/autosynth/synthesizer.py"", line 118, in synthesize
+ synth_proc.check_returncode() # Raise an exception.
+ File ""/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py"", line 389, in check_returncode
+ self.stderr)
+subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/people/synth.metadata', 'synth.py', '--', 'People']' returned non-zero exit status 1.
+
+```
+
+Google internal developers can see the full log [here](https://sponge/11ff3741-9158-4831-8681-fff828f77e1a).
+",1.0,"Synthesis failed for People - Hello! Autosynth couldn't regenerate People. :broken_heart:
+
+Here's the output from running `synth.py`:
+
+```
+2020-05-09 05:16:57 [INFO] logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/elixir-google-api
+2020-05-09 05:16:57,351 autosynth > logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/elixir-google-api
+Switched to branch 'autosynth-people'
+2020-05-09 05:16:58 [INFO] Running synthtool
+2020-05-09 05:16:58,996 autosynth > Running synthtool
+2020-05-09 05:16:58 [INFO] ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/people/synth.metadata', 'synth.py', '--']
+2020-05-09 05:16:58,996 autosynth > ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/people/synth.metadata', 'synth.py', '--']
+2020-05-09 05:16:59,205 synthtool > Executing /home/kbuilder/.cache/synthtool/elixir-google-api/synth.py.
+On branch autosynth-people
+nothing to commit, working tree clean
+2020-05-09 05:16:59,513 synthtool > Cloning https://github.com/googleapis/elixir-google-api.git.
+2020-05-09 05:16:59,977 synthtool > Running: docker run --rm -v/home/kbuilder/.cache/synthtool/elixir-google-api:/workspace -v/var/run/docker.sock:/var/run/docker.sock -e USER_GROUP=1000:1000 -w /workspace gcr.io/cloud-devrel-public-resources/elixir19 scripts/generate_client.sh People
+2020-05-09 05:17:03,627 synthtool > No files in sources /home/kbuilder/.cache/synthtool/elixir-google-api/clients were copied. Does the source contain files?
+Traceback (most recent call last):
+ File ""/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py"", line 193, in _run_module_as_main
+ ""__main__"", mod_spec)
+ File ""/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py"", line 85, in _run_code
+ exec(code, run_globals)
+ File ""/tmpfs/src/github/synthtool/synthtool/__main__.py"", line 102, in
+ main()
+ File ""/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py"", line 829, in __call__
+ return self.main(*args, **kwargs)
+ File ""/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py"", line 782, in main
+ rv = self.invoke(ctx)
+ File ""/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py"", line 1066, in invoke
+ return ctx.invoke(self.callback, **ctx.params)
+ File ""/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py"", line 610, in invoke
+ return callback(*args, **kwargs)
+ File ""/tmpfs/src/github/synthtool/synthtool/__main__.py"", line 94, in main
+ spec.loader.exec_module(synth_module) # type: ignore
+ File ""/tmpfs/src/github/synthtool/synthtool/metadata.py"", line 180, in __exit__
+ write(self.metadata_file_path)
+ File ""/tmpfs/src/github/synthtool/synthtool/metadata.py"", line 112, in write
+ with open(outfile, ""w"") as fh:
+FileNotFoundError: [Errno 2] No such file or directory: 'clients/people/synth.metadata'
+2020-05-09 05:17:03 [ERROR] Synthesis failed
+2020-05-09 05:17:03,654 autosynth > Synthesis failed
+Traceback (most recent call last):
+ File ""/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py"", line 193, in _run_module_as_main
+ ""__main__"", mod_spec)
+ File ""/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py"", line 85, in _run_code
+ exec(code, run_globals)
+ File ""/tmpfs/src/github/synthtool/autosynth/synth.py"", line 599, in
+ main()
+ File ""/tmpfs/src/github/synthtool/autosynth/synth.py"", line 471, in main
+ return _inner_main(temp_dir)
+ File ""/tmpfs/src/github/synthtool/autosynth/synth.py"", line 549, in _inner_main
+ ).synthesize(base_synth_log_path)
+ File ""/tmpfs/src/github/synthtool/autosynth/synthesizer.py"", line 118, in synthesize
+ synth_proc.check_returncode() # Raise an exception.
+ File ""/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py"", line 389, in check_returncode
+ self.stderr)
+subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/people/synth.metadata', 'synth.py', '--', 'People']' returned non-zero exit status 1.
+
+```
+
+Google internal developers can see the full log [here](https://sponge/11ff3741-9158-4831-8681-fff828f77e1a).
+",1,synthesis failed for people hello autosynth couldn t regenerate people broken heart here s the output from running synth py logs will be written to tmpfs src github synthtool logs googleapis elixir google api autosynth logs will be written to tmpfs src github synthtool logs googleapis elixir google api switched to branch autosynth people running synthtool autosynth running synthtool autosynth synthtool executing home kbuilder cache synthtool elixir google api synth py on branch autosynth people nothing to commit working tree clean synthtool cloning synthtool running docker run rm v home kbuilder cache synthtool elixir google api workspace v var run docker sock var run docker sock e user group w workspace gcr io cloud devrel public resources scripts generate client sh people synthtool no files in sources home kbuilder cache synthtool elixir google api clients were copied does the source contain files traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool synthtool main py line in main file tmpfs src github synthtool env lib site packages click core py line in call return self main args kwargs file tmpfs src github synthtool env lib site packages click core py line in main rv self invoke ctx file tmpfs src github synthtool env lib site packages click core py line in invoke return ctx invoke self callback ctx params file tmpfs src github synthtool env lib site packages click core py line in invoke return callback args kwargs file tmpfs src github synthtool synthtool main py line in main spec loader exec module synth module type ignore file tmpfs src github synthtool synthtool metadata py line in exit write self metadata file path file tmpfs src github synthtool synthtool metadata py line in write with open outfile w as fh filenotfounderror no such file or directory clients people synth metadata synthesis failed autosynth synthesis failed traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main synthesize base synth log path file tmpfs src github synthtool autosynth synthesizer py line in synthesize synth proc check returncode raise an exception file home kbuilder pyenv versions lib subprocess py line in check returncode self stderr subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log ,1
+1170,30631443750.0,IssuesEvent,2023-07-24 14:47:06,restincode/restincode,https://api.github.com/repos/restincode/restincode,opened,Chris (The Nightstalker),People Add Person Needs Review,"Please fill out as much information as you can. No fields are required, but the more you can provide the better.
+
+**General Info**
+* First name: Chris
+* Last name:
+* Handle: The Nightstalker
+* Birth Year:
+* Death Year:
+* Link to Obituary:
+* Group Affiliations: Cult of the Dead Cow
+* URL to main photo (or attach to Issue):
+* Description of person and/or activities:
+* Facebook memorial group URL:
+
+**Social Media Links**
+* Twitter:
+* Github:
+* LinkedIn:
+* Facebook:
+* Other:
+
+**Contributions**
+Repeat as many times as needed
+
+* Project name:
+* Project URL:
+* Project Description:
+
+**Photo Gallery**
+* URL(s) to additional photos:
+",1.0,"Chris (The Nightstalker) - Please fill out as much information as you can. No fields are required, but the more you can provide the better.
+
+**General Info**
+* First name: Chris
+* Last name:
+* Handle: The Nightstalker
+* Birth Year:
+* Death Year:
+* Link to Obituary:
+* Group Affiliations: Cult of the Dead Cow
+* URL to main photo (or attach to Issue):
+* Description of person and/or activities:
+* Facebook memorial group URL:
+
+**Social Media Links**
+* Twitter:
+* Github:
+* LinkedIn:
+* Facebook:
+* Other:
+
+**Contributions**
+Repeat as many times as needed
+
+* Project name:
+* Project URL:
+* Project Description:
+
+**Photo Gallery**
+* URL(s) to additional photos:
+",1,chris the nightstalker please fill out as much information as you can no fields are required but the more you can provide the better general info first name chris last name handle the nightstalker birth year death year link to obituary group affiliations cult of the dead cow url to main photo or attach to issue description of person and or activities facebook memorial group url social media links twitter github linkedin facebook other contributions repeat as many times as needed project name project url project description photo gallery url s to additional photos ,1
+344897,10349721527.0,IssuesEvent,2019-09-04 23:38:11,oslc-op/oslc-specs,https://api.github.com/repos/oslc-op/oslc-specs,closed,Extend ReSpec to support oslc:range and oslc:allowedValues for enumerations,Priority: High Status: Deferred Tool: ReSpec Xtra: Jira,"OSLC ResourceShapes defines oslc:allowedValues and oslc:range on oslc:Property which can be used define enumeration types and enumeration literals. An rdfs:Class is defined for the enumeration, and instances of this class are defined to give URIs and labels to possible enumeration instances.
+
+oslc:range can be set to the enumeration class, and oslc:allowValues can be used to list the URIs of the specific enumeration literals.
+
+ReSpec should be modified to generate HTML tables for the enumeration class and the enumeration literals. for oslc:range and oslc:allowedValues.
+
+---
+_Migrated from https://issues.oasis-open.org/browse/OSLCCORE-54 (opened by @jamsden; previously assigned to _**Unknown user**_)_
+",1.0,"Extend ReSpec to support oslc:range and oslc:allowedValues for enumerations - OSLC ResourceShapes defines oslc:allowedValues and oslc:range on oslc:Property which can be used define enumeration types and enumeration literals. An rdfs:Class is defined for the enumeration, and instances of this class are defined to give URIs and labels to possible enumeration instances.
+
+oslc:range can be set to the enumeration class, and oslc:allowValues can be used to list the URIs of the specific enumeration literals.
+
+ReSpec should be modified to generate HTML tables for the enumeration class and the enumeration literals. for oslc:range and oslc:allowedValues.
+
+---
+_Migrated from https://issues.oasis-open.org/browse/OSLCCORE-54 (opened by @jamsden; previously assigned to _**Unknown user**_)_
+",0,extend respec to support oslc range and oslc allowedvalues for enumerations oslc resourceshapes defines oslc allowedvalues and oslc range on oslc property which can be used define enumeration types and enumeration literals an rdfs class is defined for the enumeration and instances of this class are defined to give uris and labels to possible enumeration instances oslc range can be set to the enumeration class and oslc allowvalues can be used to list the uris of the specific enumeration literals respec should be modified to generate html tables for the enumeration class and the enumeration literals for oslc range and oslc allowedvalues migrated from opened by jamsden previously assigned to unknown user ,0
+254185,19189102080.0,IssuesEvent,2021-12-05 17:56:02,numpy/numpy,https://api.github.com/repos/numpy/numpy,opened,DOC: The examples for numpy.mat do not demonstrate numpy.mat,04 - Documentation,"### Issue with current documentation:
+
+On the doc page https://numpy.org/doc/stable/reference/generated/numpy.mat.html, the examples use `np.asmatrix` but not `np.mat`.
+
+### Idea or request for content:
+
+_No response_",1.0,"DOC: The examples for numpy.mat do not demonstrate numpy.mat - ### Issue with current documentation:
+
+On the doc page https://numpy.org/doc/stable/reference/generated/numpy.mat.html, the examples use `np.asmatrix` but not `np.mat`.
+
+### Idea or request for content:
+
+_No response_",0,doc the examples for numpy mat do not demonstrate numpy mat issue with current documentation on the doc page the examples use np asmatrix but not np mat idea or request for content no response ,0
+202952,15308827444.0,IssuesEvent,2021-02-24 23:10:20,fair-software/howfairis,https://api.github.com/repos/fair-software/howfairis,closed,testing: clitests/script.sh better user feedback,testing,"- `clitests/script.sh` print message when sleeping, maybe also use quiet mode
+- `clitests/script.sh` print final message when done
+",1.0,"testing: clitests/script.sh better user feedback - - `clitests/script.sh` print message when sleeping, maybe also use quiet mode
+- `clitests/script.sh` print final message when done
+",0,testing clitests script sh better user feedback clitests script sh print message when sleeping maybe also use quiet mode clitests script sh print final message when done ,0
+47,2660236408.0,IssuesEvent,2015-03-19 04:19:42,fskreuz/Tree-Trail,https://api.github.com/repos/fskreuz/Tree-Trail,closed,Failed to delete badge bug,QA people! Test these!,"Logged in as admin
+Tried to delete a badge by clicking on it then clicking trashcan icon.
+Prompt said ""Failed to delete location""
+
+Kindly fix thanks :)",1.0,"Failed to delete badge bug - Logged in as admin
+Tried to delete a badge by clicking on it then clicking trashcan icon.
+Prompt said ""Failed to delete location""
+
+Kindly fix thanks :)",1,failed to delete badge bug logged in as admin tried to delete a badge by clicking on it then clicking trashcan icon prompt said failed to delete location kindly fix thanks ,1
+458,8353405680.0,IssuesEvent,2018-10-02 09:55:41,allenleein/brains,https://api.github.com/repos/allenleein/brains,closed,Mysterious great white shark lair discovered in Pacific Ocean,People,"Mysterious great white shark lair discovered in Pacific Ocean
+A scientific mission into the secret ocean lair of California’s great white sharks has provided tantalizing clues into a vexing mystery — why the fearsome predators spend winter and spring in what has long appeared to be an empty void in the deep sea.
+
+via Pocket https://ift.tt/2pguv87
+
+September 19, 2018 at 11:16PM",1.0,"Mysterious great white shark lair discovered in Pacific Ocean - Mysterious great white shark lair discovered in Pacific Ocean
+A scientific mission into the secret ocean lair of California’s great white sharks has provided tantalizing clues into a vexing mystery — why the fearsome predators spend winter and spring in what has long appeared to be an empty void in the deep sea.
+
+via Pocket https://ift.tt/2pguv87
+
+September 19, 2018 at 11:16PM",1,mysterious great white shark lair discovered in pacific ocean mysterious great white shark lair discovered in pacific ocean a scientific mission into the secret ocean lair of california rsquo s great white sharks has provided tantalizing clues into a vexing mystery mdash why the fearsome predators spend winter and spring in what has long appeared to be an empty void in the deep sea via pocket september at ,1
+339,6533139451.0,IssuesEvent,2017-08-31 04:09:19,Orbiit/gunn-remake,https://api.github.com/repos/Orbiit/gunn-remake,closed,Design concept,DISCUSSION TIME FEATURE REQUEST JUST A QUESTION REQUIRES MORE PEOPLE,"So I suppose we're moving on to making la ĉefretpaĝo so we know what we actually need to design.
+
+Features the current Gunn ĉefretpaĝo has that I think we should incorporate into our redesign in some way:
+
+- [x] image carousel (see #8) which can be that huge image modern sites have for some reason ([example](https://host.gwiddlefoundation.org.uk/) and [another example](https://themes.materializecss.com/pages/dark-theme))
+- [x] events calendar (it only has to list a few items; there can be a view more option linking to a full-sized calendar)
+- [x] recent news/announcements
+- [x] a very easy-to-access ""Crisis Resources"" link",1.0,"Design concept - So I suppose we're moving on to making la ĉefretpaĝo so we know what we actually need to design.
+
+Features the current Gunn ĉefretpaĝo has that I think we should incorporate into our redesign in some way:
+
+- [x] image carousel (see #8) which can be that huge image modern sites have for some reason ([example](https://host.gwiddlefoundation.org.uk/) and [another example](https://themes.materializecss.com/pages/dark-theme))
+- [x] events calendar (it only has to list a few items; there can be a view more option linking to a full-sized calendar)
+- [x] recent news/announcements
+- [x] a very easy-to-access ""Crisis Resources"" link",1,design concept so i suppose we re moving on to making la ĉefretpaĝo so we know what we actually need to design features the current gunn ĉefretpaĝo has that i think we should incorporate into our redesign in some way image carousel see which can be that huge image modern sites have for some reason and events calendar it only has to list a few items there can be a view more option linking to a full sized calendar recent news announcements a very easy to access crisis resources link,1
+727694,25044289860.0,IssuesEvent,2022-11-05 03:26:13,AY2223S1-CS2113-F11-4/tp,https://api.github.com/repos/AY2223S1-CS2113-F11-4/tp,closed,[TA Code Review] for Ria,type.Task priority.Medium,"@RiaVora
+Good adherence to coding standards and code quality guidelines. Good use of regex to help parse the input. 👍🏽 Here are some of my comments on your code:
+- [x] `Parser.java` has function `patientParser` which is quite long (~40 lines). Consider shortening by extractibg methods out from them
+- [x] I see a lot of deep nested clauses. Try and extract the inner nests into functions
+- [x] Like I mentioned to your teammates, there are many magic literals in the code consider refactoring them to static final
+- [x] In `PatientList.java` lines 49 to 51, please avoid using comments in the code as programmers notes",1.0,"[TA Code Review] for Ria - @RiaVora
+Good adherence to coding standards and code quality guidelines. Good use of regex to help parse the input. 👍🏽 Here are some of my comments on your code:
+- [x] `Parser.java` has function `patientParser` which is quite long (~40 lines). Consider shortening by extractibg methods out from them
+- [x] I see a lot of deep nested clauses. Try and extract the inner nests into functions
+- [x] Like I mentioned to your teammates, there are many magic literals in the code consider refactoring them to static final
+- [x] In `PatientList.java` lines 49 to 51, please avoid using comments in the code as programmers notes",0, for ria riavora good adherence to coding standards and code quality guidelines good use of regex to help parse the input 👍🏽 here are some of my comments on your code parser java has function patientparser which is quite long lines consider shortening by extractibg methods out from them i see a lot of deep nested clauses try and extract the inner nests into functions like i mentioned to your teammates there are many magic literals in the code consider refactoring them to static final in patientlist java lines to please avoid using comments in the code as programmers notes,0
+485936,14001580882.0,IssuesEvent,2020-10-28 13:49:48,AY2021S1-CS2103T-F11-3/tp,https://api.github.com/repos/AY2021S1-CS2103T-F11-3/tp,closed,Shortcut to add tags to assignments,enhancement priority.MEDIUM type.Task,Modify the add command to be able to add tags when adding an assignment,1.0,Shortcut to add tags to assignments - Modify the add command to be able to add tags when adding an assignment,0,shortcut to add tags to assignments modify the add command to be able to add tags when adding an assignment,0
+799,14813293739.0,IssuesEvent,2021-01-14 01:41:03,mnalis/ironseed_fpc,https://api.github.com/repos/mnalis/ironseed_fpc,closed,FTBFS on mipsel,Other_People_Bug bug,"Fails to build from source: https://buildd.debian.org/status/package.php?p=ironseed
+
+```
+(3104) Compiling is.pas
+(3104) Compiling version.pas
+(9009) Assembling version
+(3104) Compiling utils_.pas
+(3104) Compiling _paths_.pas
+_paths_.pas(7,30) Warning: (4110) range check error while evaluating constants (1995272928 must be between 0 and 255)
+_paths_.pas(10) Fatal: (10026) There were 1 errors compiling module, stopping
+Fatal: (1018) Compilation aborted
+Error: /usr/bin/ppcmipsel returned an error exitcode
+make[2]: *** [Makefile:81: is] Error 1
+make[2]: Leaving directory '/<>'
+dh_auto_build: error: make -j1 ""INSTALL=install --strip-program=true"" prefix=/usr libdir=/usr/libexec/ironseed returned exit code 2
+make[1]: *** [debian/rules:14: override_dh_auto_build] Error 25
+make[1]: Leaving directory '/<>'
+make: *** [debian/rules:8: binary-arch] Error 2
+```",1.0,"FTBFS on mipsel - Fails to build from source: https://buildd.debian.org/status/package.php?p=ironseed
+
+```
+(3104) Compiling is.pas
+(3104) Compiling version.pas
+(9009) Assembling version
+(3104) Compiling utils_.pas
+(3104) Compiling _paths_.pas
+_paths_.pas(7,30) Warning: (4110) range check error while evaluating constants (1995272928 must be between 0 and 255)
+_paths_.pas(10) Fatal: (10026) There were 1 errors compiling module, stopping
+Fatal: (1018) Compilation aborted
+Error: /usr/bin/ppcmipsel returned an error exitcode
+make[2]: *** [Makefile:81: is] Error 1
+make[2]: Leaving directory '/<>'
+dh_auto_build: error: make -j1 ""INSTALL=install --strip-program=true"" prefix=/usr libdir=/usr/libexec/ironseed returned exit code 2
+make[1]: *** [debian/rules:14: override_dh_auto_build] Error 25
+make[1]: Leaving directory '/<>'
+make: *** [debian/rules:8: binary-arch] Error 2
+```",1,ftbfs on mipsel fails to build from source compiling is pas compiling version pas assembling version compiling utils pas compiling paths pas paths pas warning range check error while evaluating constants must be between and paths pas fatal there were errors compiling module stopping fatal compilation aborted error usr bin ppcmipsel returned an error exitcode make error make leaving directory dh auto build error make install install strip program true prefix usr libdir usr libexec ironseed returned exit code make error make leaving directory make error ,1
+712477,24496592029.0,IssuesEvent,2022-10-10 09:12:45,DH-IT-Portal-Development/ethics,https://api.github.com/repos/DH-IT-Portal-Development/ethics,opened,Elders goedgekeurde studies - meer documenten aanleveren,priority new-regulation,"In het nieuwe reglement staat (artikel 4.3.): Hiervoor moet een aanvraag gedaan worden in de FETC-GW portal, met geïnformeerde toestemmingsdocumenten die in overeenstemming zijn met voorbeelddocumenten die de FETC-GW op de website en op intranet ter beschikking heeft gesteld.
+
+Dit kan nu nog niet, maar dat moet dus wel. Voorstel om na 'Upload hier je formele toestemming....' een extra regel toe te voegen met meerdere mogelijkheden om documenten te uploaden (acht?)
+
+""Upload hier de documenten die je naar je deelnemers/gatekeepers wilt sturen en die zijn goedgekeurd door de andere ethische toetsingcommissie.""
+
+Tja, hier moeten dan ook bestandsnamen aan gekoppeld worden vrees ik? Het kan gaan om meerdere informatiebrieven, toestemmingsverklaringen en bezwaarformulieren.",1.0,"Elders goedgekeurde studies - meer documenten aanleveren - In het nieuwe reglement staat (artikel 4.3.): Hiervoor moet een aanvraag gedaan worden in de FETC-GW portal, met geïnformeerde toestemmingsdocumenten die in overeenstemming zijn met voorbeelddocumenten die de FETC-GW op de website en op intranet ter beschikking heeft gesteld.
+
+Dit kan nu nog niet, maar dat moet dus wel. Voorstel om na 'Upload hier je formele toestemming....' een extra regel toe te voegen met meerdere mogelijkheden om documenten te uploaden (acht?)
+
+""Upload hier de documenten die je naar je deelnemers/gatekeepers wilt sturen en die zijn goedgekeurd door de andere ethische toetsingcommissie.""
+
+Tja, hier moeten dan ook bestandsnamen aan gekoppeld worden vrees ik? Het kan gaan om meerdere informatiebrieven, toestemmingsverklaringen en bezwaarformulieren.",0,elders goedgekeurde studies meer documenten aanleveren in het nieuwe reglement staat artikel hiervoor moet een aanvraag gedaan worden in de fetc gw portal met geïnformeerde toestemmingsdocumenten die in overeenstemming zijn met voorbeelddocumenten die de fetc gw op de website en op intranet ter beschikking heeft gesteld dit kan nu nog niet maar dat moet dus wel voorstel om na upload hier je formele toestemming een extra regel toe te voegen met meerdere mogelijkheden om documenten te uploaden acht upload hier de documenten die je naar je deelnemers gatekeepers wilt sturen en die zijn goedgekeurd door de andere ethische toetsingcommissie tja hier moeten dan ook bestandsnamen aan gekoppeld worden vrees ik het kan gaan om meerdere informatiebrieven toestemmingsverklaringen en bezwaarformulieren ,0
+790334,27823057938.0,IssuesEvent,2023-03-19 13:08:53,microsoft/PowerToys,https://api.github.com/repos/microsoft/PowerToys,closed,ADMX file update,Issue-Bug Issue-Docs Resolution-Fix-Committed Priority-3 Cost-Small Area-Enterprise,"### Microsoft PowerToys version
+
+0.66.0
+
+### Installation method
+
+GitHub
+
+### Running as admin
+
+Yes
+
+### Area(s) with issue?
+
+General
+
+### Steps to reproduce
+
+ADMX file seems not updated since 0.64.0.
+
+Probably different issue. I am trying to create a configuration profile from this ADMX in Intune and the Administration Template doesn't list-up. Endless ""Loading"" (Chargement) in French.
+
+
+
+
+
+
+### ✔️ Expected Behavior
+
+_No response_
+
+### ❌ Actual Behavior
+
+_No response_
+
+### Other Software
+
+_No response_",1.0,"ADMX file update - ### Microsoft PowerToys version
+
+0.66.0
+
+### Installation method
+
+GitHub
+
+### Running as admin
+
+Yes
+
+### Area(s) with issue?
+
+General
+
+### Steps to reproduce
+
+ADMX file seems not updated since 0.64.0.
+
+Probably different issue. I am trying to create a configuration profile from this ADMX in Intune and the Administration Template doesn't list-up. Endless ""Loading"" (Chargement) in French.
+
+
+
+
+
+
+### ✔️ Expected Behavior
+
+_No response_
+
+### ❌ Actual Behavior
+
+_No response_
+
+### Other Software
+
+_No response_",0,admx file update microsoft powertoys version installation method github running as admin yes area s with issue general steps to reproduce admx file seems not updated since probably different issue i am trying to create a configuration profile from this admx in intune and the administration template doesn t list up endless loading chargement in french ✔️ expected behavior no response ❌ actual behavior no response other software no response ,0
+5115,4791482386.0,IssuesEvent,2016-10-31 12:45:02,Automattic/jetpack,https://api.github.com/repos/Automattic/jetpack,closed,Twenty Sixteen compat file loaded in the dashboard.,Performance Theme Tools [Team] tdiv [Type] Enhancement [Type] Good First Bug,"I have jetpack loaded on a fresh wordpress install using development mode. I ran the audit tool and noticed that a file called `twentysixteen.css`.
+
+Currently this file loads in the WordPress admin, but I can't find any page admin page where it is used.
+
+I also noticed it is loading a non-minified file:
+https://github.com/Automattic/jetpack/blob/a15e5028bc2fd10af3b8740ec69d0a3c0ab6eff9/modules/theme-tools/compat/twentysixteen.php#L15-L28
+
+I suggest we load a minified version, and not in the WordPress Admin. Looks like the same might happen with the `twentyfifteen.css` file.
+",True,"Twenty Sixteen compat file loaded in the dashboard. - I have jetpack loaded on a fresh wordpress install using development mode. I ran the audit tool and noticed that a file called `twentysixteen.css`.
+
+Currently this file loads in the WordPress admin, but I can't find any page admin page where it is used.
+
+I also noticed it is loading a non-minified file:
+https://github.com/Automattic/jetpack/blob/a15e5028bc2fd10af3b8740ec69d0a3c0ab6eff9/modules/theme-tools/compat/twentysixteen.php#L15-L28
+
+I suggest we load a minified version, and not in the WordPress Admin. Looks like the same might happen with the `twentyfifteen.css` file.
+",0,twenty sixteen compat file loaded in the dashboard i have jetpack loaded on a fresh wordpress install using development mode i ran the audit tool and noticed that a file called twentysixteen css currently this file loads in the wordpress admin but i can t find any page admin page where it is used i also noticed it is loading a non minified file i suggest we load a minified version and not in the wordpress admin looks like the same might happen with the twentyfifteen css file ,0
+213159,7246361014.0,IssuesEvent,2018-02-14 21:23:01,psouza4/mediacentermaster,https://api.github.com/repos/psouza4/mediacentermaster,closed,Release Date (XML > NFO),Feature-Movies Fixed / Resolved Player-Kodi Priority-Low Type-FeatureRequest,"Originally reported on Google Code with ID 1413
+
+```
+It would be really nice to add a 'releasedate' field to the XBMC .nfo.
+Since the .xml already has this field as 'ReleaseDate',
+it seems pretty easy to just implement it to the .nfo.
+
+NOTE: I actually want this because I use Plex + XBMC Agent and the only field I'm missing
+is the release date.
+```
+
+Reported by `crazyrabbit0` on 2013-08-20 22:33:59
+",1.0,"Release Date (XML > NFO) - Originally reported on Google Code with ID 1413
+
+```
+It would be really nice to add a 'releasedate' field to the XBMC .nfo.
+Since the .xml already has this field as 'ReleaseDate',
+it seems pretty easy to just implement it to the .nfo.
+
+NOTE: I actually want this because I use Plex + XBMC Agent and the only field I'm missing
+is the release date.
+```
+
+Reported by `crazyrabbit0` on 2013-08-20 22:33:59
+",0,release date xml nfo originally reported on google code with id it would be really nice to add a releasedate field to the xbmc nfo since the xml already has this field as releasedate it seems pretty easy to just implement it to the nfo note i actually want this because i use plex xbmc agent and the only field i m missing is the release date reported by on ,0
+744406,25941545017.0,IssuesEvent,2022-12-16 18:56:35,idom-team/idom,https://api.github.com/repos/idom-team/idom,opened,Look into Signals Pattern,type: feature priority: 3 (low),"### Current Situation
+
+Preact, a fast and lightweight drop-in replacement for React, has a pluggable renderer that allows them to add new state and rendering primitives without a wholesale re-write. Signals are one of the new primitives they created. In this [introductory article](https://preactjs.com/blog/introducing-signals/) they lay out some advantages of signals over hooks.
+
+### Proposed Actions
+
+While signals are an interesting pattern to explore. The more useful action here would be to implement signals for the purpose of developing a similarly pluggable layout rendering pipeline. This would ensure that, as better reactive patterns are developed, IDOM can more easily evolve with them.",1.0,"Look into Signals Pattern - ### Current Situation
+
+Preact, a fast and lightweight drop-in replacement for React, has a pluggable renderer that allows them to add new state and rendering primitives without a wholesale re-write. Signals are one of the new primitives they created. In this [introductory article](https://preactjs.com/blog/introducing-signals/) they lay out some advantages of signals over hooks.
+
+### Proposed Actions
+
+While signals are an interesting pattern to explore. The more useful action here would be to implement signals for the purpose of developing a similarly pluggable layout rendering pipeline. This would ensure that, as better reactive patterns are developed, IDOM can more easily evolve with them.",0,look into signals pattern current situation preact a fast and lightweight drop in replacement for react has a pluggable renderer that allows them to add new state and rendering primitives without a wholesale re write signals are one of the new primitives they created in this they lay out some advantages of signals over hooks proposed actions while signals are an interesting pattern to explore the more useful action here would be to implement signals for the purpose of developing a similarly pluggable layout rendering pipeline this would ensure that as better reactive patterns are developed idom can more easily evolve with them ,0
+151628,13428021460.0,IssuesEvent,2020-09-06 20:18:09,jhipster/generator-jhipster,https://api.github.com/repos/jhipster/generator-jhipster,closed,Broken Links in Showcase of Apps,$$ bug-bounty $$ $100 area: cleanup area: documentation:books:,"
+
+##### **Overview of the issue**
+
+As [mentioned in my email](https://groups.google.com/forum/#!topic/jhipster-dev/ZcoTQagXF0s); most of the webpages in our app showcase seems to have broken links or pointing to GitHub pages.
+
+**1) Devoxx**
+
+Gives a security warning (SSL certificate problems).
+
+@deepu105 : Since you have added this, do you know a different link perhaps? :smile_cat:
+
+**2) Dan Cancro's Great Big Example Application**
+
+This one is a sample application and the link points to the GitHub repo. I vote for removing this since the showcase should probably include real world applications that are made using jhipster? :smile:
+
+**3) SoundXtream Audio Platform**
+
+This one takes us to the GitHub account. And there's a link to application there; but it doesn't seem to work (http://soundxtream.com/).
+
+@xavierpandis: Do you have this app running somewhere? If so update us with the link so we can update it. :smile:
+
+**4) The Rock Bible**
+
+Link points to the GitHub repo.
+
+@jotabono: Is this app deployed? If so please update us with the link. :smile:
+
+**5) Real State Camp**
+
+Link points to the GitHub repo.
+
+@arnaugarcia: Is this app deployed? If so please update us with the link. :smile:
+
+**6) Coincoindrone**
+
+404 error.
+
+@Z3nk: Is this app deployed and running? If so could you please provide us with the link? :smile:
+
+**7) Guc Voile**
+
+Points to a YouTube video.
+
+@GregTres: Do you have the actual url of the application. Maybe we can include that instead? :smile:
+
+##### **Motivation for or Use Case**
+
+Cleanup the showcase section.
+
+cc: @jhipster/developers
+
+- [x] Checking this box is mandatory (this is just to show you read everything)
+
+
+",1.0,"Broken Links in Showcase of Apps -
+
+##### **Overview of the issue**
+
+As [mentioned in my email](https://groups.google.com/forum/#!topic/jhipster-dev/ZcoTQagXF0s); most of the webpages in our app showcase seems to have broken links or pointing to GitHub pages.
+
+**1) Devoxx**
+
+Gives a security warning (SSL certificate problems).
+
+@deepu105 : Since you have added this, do you know a different link perhaps? :smile_cat:
+
+**2) Dan Cancro's Great Big Example Application**
+
+This one is a sample application and the link points to the GitHub repo. I vote for removing this since the showcase should probably include real world applications that are made using jhipster? :smile:
+
+**3) SoundXtream Audio Platform**
+
+This one takes us to the GitHub account. And there's a link to application there; but it doesn't seem to work (http://soundxtream.com/).
+
+@xavierpandis: Do you have this app running somewhere? If so update us with the link so we can update it. :smile:
+
+**4) The Rock Bible**
+
+Link points to the GitHub repo.
+
+@jotabono: Is this app deployed? If so please update us with the link. :smile:
+
+**5) Real State Camp**
+
+Link points to the GitHub repo.
+
+@arnaugarcia: Is this app deployed? If so please update us with the link. :smile:
+
+**6) Coincoindrone**
+
+404 error.
+
+@Z3nk: Is this app deployed and running? If so could you please provide us with the link? :smile:
+
+**7) Guc Voile**
+
+Points to a YouTube video.
+
+@GregTres: Do you have the actual url of the application. Maybe we can include that instead? :smile:
+
+##### **Motivation for or Use Case**
+
+Cleanup the showcase section.
+
+cc: @jhipster/developers
+
+- [x] Checking this box is mandatory (this is just to show you read everything)
+
+
+",0,broken links in showcase of apps please follow the issue template below for bug reports if you have a support request rather than a bug please use with the jhipster tag for bug reports it is mandatory to run the command jhipster info in your project s root folder and paste the result here tickets opened without any of these pieces of information will be closed without any explanation overview of the issue as most of the webpages in our app showcase seems to have broken links or pointing to github pages devoxx gives a security warning ssl certificate problems since you have added this do you know a different link perhaps smile cat dan cancro s great big example application this one is a sample application and the link points to the github repo i vote for removing this since the showcase should probably include real world applications that are made using jhipster smile soundxtream audio platform this one takes us to the github account and there s a link to application there but it doesn t seem to work xavierpandis do you have this app running somewhere if so update us with the link so we can update it smile the rock bible link points to the github repo jotabono is this app deployed if so please update us with the link smile real state camp link points to the github repo arnaugarcia is this app deployed if so please update us with the link smile coincoindrone error is this app deployed and running if so could you please provide us with the link smile guc voile points to a youtube video gregtres do you have the actual url of the application maybe we can include that instead smile motivation for or use case cleanup the showcase section cc jhipster developers checking this box is mandatory this is just to show you read everything love jhipster please consider supporting our collective 👉 ,0
+676050,23114531144.0,IssuesEvent,2022-07-27 15:31:45,apcountryman/picolibrary,https://api.github.com/repos/apcountryman/picolibrary,closed,Add TCP over IP facilities namespace,priority-normal status-awaiting_review type-feature,"Add TCP over IP facilities namespace (`::picolibrary::IP::TCP`).
+- [x] The `TCP` namespace should be defined in the `include/picolibrary/ip/tcp.h`/`source/picolibrary/ip/tcp.cc` header/source file pair",1.0,"Add TCP over IP facilities namespace - Add TCP over IP facilities namespace (`::picolibrary::IP::TCP`).
+- [x] The `TCP` namespace should be defined in the `include/picolibrary/ip/tcp.h`/`source/picolibrary/ip/tcp.cc` header/source file pair",0,add tcp over ip facilities namespace add tcp over ip facilities namespace picolibrary ip tcp the tcp namespace should be defined in the include picolibrary ip tcp h source picolibrary ip tcp cc header source file pair,0
+596597,18107540592.0,IssuesEvent,2021-09-22 21:00:21,googleapis/python-bigquery-pandas,https://api.github.com/repos/googleapis/python-bigquery-pandas,closed,Authenticate with Pandas GBQ for Python BigQuery client,type: question priority: p3 api: bigquery,"I'm trying to list all the tables in a GCP project inside a Jupyter notebook. I can use the information schema tables, but then I have to know all the datasets in advance.
+
+Instead, I can use the [Python client](https://cloud.google.com/bigquery/docs/listing-datasets), but that requires authentication. People who (re-)use the notebook do not have service account keys or local keys from the gcloud SDK, but they do authenticate with their own accounts using Pandas GBQ.
+
+Is there a way I can pass that information on to Python client?",1.0,"Authenticate with Pandas GBQ for Python BigQuery client - I'm trying to list all the tables in a GCP project inside a Jupyter notebook. I can use the information schema tables, but then I have to know all the datasets in advance.
+
+Instead, I can use the [Python client](https://cloud.google.com/bigquery/docs/listing-datasets), but that requires authentication. People who (re-)use the notebook do not have service account keys or local keys from the gcloud SDK, but they do authenticate with their own accounts using Pandas GBQ.
+
+Is there a way I can pass that information on to Python client?",0,authenticate with pandas gbq for python bigquery client i m trying to list all the tables in a gcp project inside a jupyter notebook i can use the information schema tables but then i have to know all the datasets in advance instead i can use the but that requires authentication people who re use the notebook do not have service account keys or local keys from the gcloud sdk but they do authenticate with their own accounts using pandas gbq is there a way i can pass that information on to python client ,0
+67791,13032309461.0,IssuesEvent,2020-07-28 03:54:05,topcoder-platform/community-app,https://api.github.com/repos/topcoder-platform/community-app,opened,Stalled Challenge phase,Beta Env Challenge Listing P4 v5-intgration-sub-code,"When the submission phase ends and the review phase does not open automatically, the challenge is displayed with as submission late phase in V4 and Stalled state in V5.
+
+example:https://beta-community-app.topcoder.com/challenges/ca14c0f3-b29e-4826-a6bb-560d8448cd9b
+https://www.topcoder.com/challenges/30132551
+
+
+
+
+cc @sushilshinde ",1.0,"Stalled Challenge phase - When the submission phase ends and the review phase does not open automatically, the challenge is displayed with as submission late phase in V4 and Stalled state in V5.
+
+example:https://beta-community-app.topcoder.com/challenges/ca14c0f3-b29e-4826-a6bb-560d8448cd9b
+https://www.topcoder.com/challenges/30132551
+
+
+
+
+cc @sushilshinde ",0,stalled challenge phase when the submission phase ends and the review phase does not open automatically the challenge is displayed with as submission late phase in and stalled state in example img width alt screenshot at am src cc sushilshinde ,0
+982,23080670114.0,IssuesEvent,2022-07-26 06:52:03,wordpress-mobile/WordPress-Android,https://api.github.com/repos/wordpress-mobile/WordPress-Android,closed,With some keyboard creates chips on dot,[Type] Bug People Management [Status] Stale,"### Expected behavior
+In the Invite People screen, while adding `Usernames or emails`I would expect to be able to add emails with dot sign without the string being transformed in a chip.
+
+### Actual behavior
+In the Invite People screen, while adding `Usernames or emails`I can have the string validated and being transformed in a chip when using a keyboard different from the Google Gboard standard one. Not sure on other keyboards but I noticed this for example with the MS SwiftKey keyboard.
+
+### Steps to reproduce the behavior
+- Install and enable the MS SwiftKey keyboard
+- Try to add a user email with a dot (like name.surname@domain.com)
+- Check the string is validated and transformed into a chip when pressing dot (so getting a chip with `name.` string)
+
+##### Tested on [Pixel 3 Emu], Android [API 29], WPAndroid [current 16.7 beta]
+
+",1.0,"With some keyboard creates chips on dot - ### Expected behavior
+In the Invite People screen, while adding `Usernames or emails`I would expect to be able to add emails with dot sign without the string being transformed in a chip.
+
+### Actual behavior
+In the Invite People screen, while adding `Usernames or emails`I can have the string validated and being transformed in a chip when using a keyboard different from the Google Gboard standard one. Not sure on other keyboards but I noticed this for example with the MS SwiftKey keyboard.
+
+### Steps to reproduce the behavior
+- Install and enable the MS SwiftKey keyboard
+- Try to add a user email with a dot (like name.surname@domain.com)
+- Check the string is validated and transformed into a chip when pressing dot (so getting a chip with `name.` string)
+
+##### Tested on [Pixel 3 Emu], Android [API 29], WPAndroid [current 16.7 beta]
+
+",1,with some keyboard creates chips on dot expected behavior in the invite people screen while adding usernames or emails i would expect to be able to add emails with dot sign without the string being transformed in a chip actual behavior in the invite people screen while adding usernames or emails i can have the string validated and being transformed in a chip when using a keyboard different from the google gboard standard one not sure on other keyboards but i noticed this for example with the ms swiftkey keyboard steps to reproduce the behavior install and enable the ms swiftkey keyboard try to add a user email with a dot like name surname domain com check the string is validated and transformed into a chip when pressing dot so getting a chip with name string tested on android wpandroid ,1
+679,12223112565.0,IssuesEvent,2020-05-02 16:10:58,ruby-grape/grape,https://api.github.com/repos/ruby-grape/grape,closed,Are we interested in TideLift?,chore needs info people you can help,"Are we interested in offering enterprise support for Grape via TideLift? Anyone with experience using the service here for enterprise support? I already sort of make myself available for anything urgent like security patches, but it hasn't really come up.
+
+https://tidelift.com/lifter/search/rubygems/grape claims there's $144.65/month for Grape from Enterprise subscriptions. That's enough for a once a year dinner/drinks for maintainers and regular contributors, for example in New York. I'm happy to organize after the pandemic.
+
+I would need at least a 👍from anyone with r/w permissions here.
+
+",1.0,"Are we interested in TideLift? - Are we interested in offering enterprise support for Grape via TideLift? Anyone with experience using the service here for enterprise support? I already sort of make myself available for anything urgent like security patches, but it hasn't really come up.
+
+https://tidelift.com/lifter/search/rubygems/grape claims there's $144.65/month for Grape from Enterprise subscriptions. That's enough for a once a year dinner/drinks for maintainers and regular contributors, for example in New York. I'm happy to organize after the pandemic.
+
+I would need at least a 👍from anyone with r/w permissions here.
+
+",1,are we interested in tidelift are we interested in offering enterprise support for grape via tidelift anyone with experience using the service here for enterprise support i already sort of make myself available for anything urgent like security patches but it hasn t really come up claims there s month for grape from enterprise subscriptions that s enough for a once a year dinner drinks for maintainers and regular contributors for example in new york i m happy to organize after the pandemic i would need at least a 👍from anyone with r w permissions here ,1
+285966,8781658853.0,IssuesEvent,2018-12-19 21:13:15,joatuapp/joatu-v2,https://api.github.com/repos/joatuapp/joatu-v2,opened,Integrate Event Creation features from Django into Ruby,Priority RFC enhancement feature help wanted,"The majority of what exists in the Django version of the application for event creation is quite good. We want to transfer over the majority of this feature set.
+
+https://github.com/joatuapp/joatu-django",1.0,"Integrate Event Creation features from Django into Ruby - The majority of what exists in the Django version of the application for event creation is quite good. We want to transfer over the majority of this feature set.
+
+https://github.com/joatuapp/joatu-django",0,integrate event creation features from django into ruby the majority of what exists in the django version of the application for event creation is quite good we want to transfer over the majority of this feature set ,0
+1121,27197378147.0,IssuesEvent,2023-02-20 06:53:59,ianb/llm-garden,https://api.github.com/repos/ianb/llm-garden,opened,Make peoplesim prompts full sentences,peoplesim,"It's not clear when we ask for a Description, what grammatically do we want? We should use the description consistently, and then use any prefix in the prompt. E.g., Label=""Johnny is"" then the description field",1.0,"Make peoplesim prompts full sentences - It's not clear when we ask for a Description, what grammatically do we want? We should use the description consistently, and then use any prefix in the prompt. E.g., Label=""Johnny is"" then the description field",1,make peoplesim prompts full sentences it s not clear when we ask for a description what grammatically do we want we should use the description consistently and then use any prefix in the prompt e g label johnny is then the description field,1
+189,4056151638.0,IssuesEvent,2016-05-24 17:41:35,wordpress-mobile/WordPress-iOS,https://api.github.com/repos/wordpress-mobile/WordPress-iOS,closed,People: Display Site's Readers,People Management [Type] Enhancement,"Building on the people management work done thus far, allow the user to display a site's followers:
+
+
+
+ref: #5030 ",1.0,"People: Display Site's Readers - Building on the people management work done thus far, allow the user to display a site's followers:
+
+
+
+ref: #5030 ",1,people display site s readers building on the people management work done thus far allow the user to display a site s followers img width alt ta src ref ,1
+472,8380197440.0,IssuesEvent,2018-10-07 12:19:03,yoda-pa/yoda,https://api.github.com/repos/yoda-pa/yoda,closed,People manager beautify output ,difficulty: easy hacktoberfest help wanted module: people up-for-grabs,"Current output structure: https://github.com/yoda-pa/yoda#people
+Feature added in: #184
+
+### Things to improve
+- [ ] `yoda people status` command should show a structured output instead of json
+- [ ] update error message when likes not exist: [code](https://github.com/yoda-pa/yoda/blob/master/modules/people.py#L200)
+- [ ] update error message when notes do not exist: [code](https://github.com/yoda-pa/yoda/blob/master/modules/people.py#L226)
+- [ ] update line [52](https://github.com/yoda-pa/yoda/blob/master/modules/people.py#L52) and [249](https://github.com/yoda-pa/yoda/blob/master/modules/people.py#L249)",1.0,"People manager beautify output - Current output structure: https://github.com/yoda-pa/yoda#people
+Feature added in: #184
+
+### Things to improve
+- [ ] `yoda people status` command should show a structured output instead of json
+- [ ] update error message when likes not exist: [code](https://github.com/yoda-pa/yoda/blob/master/modules/people.py#L200)
+- [ ] update error message when notes do not exist: [code](https://github.com/yoda-pa/yoda/blob/master/modules/people.py#L226)
+- [ ] update line [52](https://github.com/yoda-pa/yoda/blob/master/modules/people.py#L52) and [249](https://github.com/yoda-pa/yoda/blob/master/modules/people.py#L249)",1,people manager beautify output current output structure feature added in things to improve yoda people status command should show a structured output instead of json update error message when likes not exist update error message when notes do not exist update line and ,1
+159420,12475097891.0,IssuesEvent,2020-05-29 10:53:57,aliasrobotics/RVD,https://api.github.com/repos/aliasrobotics/RVD,closed,"RVD#2039: Use of possibly insecure function - consider using safer ast., /opt/ros_noetic_ws/src/ros/roslib/src/roslib/msgs.py:582",bandit bug static analysis testing triage,"```yaml
+{
+ ""id"": 2039,
+ ""title"": ""RVD#2039: Use of possibly insecure function - consider using safer ast., /opt/ros_noetic_ws/src/ros/roslib/src/roslib/msgs.py:582"",
+ ""type"": ""bug"",
+ ""description"": ""HIGH confidence of MEDIUM severity bug. Use of possibly insecure function - consider using safer ast.literal_eval. at /opt/ros_noetic_ws/src/ros/roslib/src/roslib/msgs.py:582 See links for more info on the bug."",
+ ""cwe"": ""None"",
+ ""cve"": ""None"",
+ ""keywords"": [
+ ""bandit"",
+ ""bug"",
+ ""static analysis"",
+ ""testing"",
+ ""triage"",
+ ""bug""
+ ],
+ ""system"": """",
+ ""vendor"": null,
+ ""severity"": {
+ ""rvss-score"": 0,
+ ""rvss-vector"": """",
+ ""severity-description"": """",
+ ""cvss-score"": 0,
+ ""cvss-vector"": """"
+ },
+ ""links"": [
+ ""https://github.com/aliasrobotics/RVD/issues/2039"",
+ ""https://bandit.readthedocs.io/en/latest/blacklists/blacklist_calls.html#b307-eval""
+ ],
+ ""flaw"": {
+ ""phase"": ""testing"",
+ ""specificity"": ""subject-specific"",
+ ""architectural-location"": ""application-specific"",
+ ""application"": ""N/A"",
+ ""subsystem"": ""N/A"",
+ ""package"": ""N/A"",
+ ""languages"": ""None"",
+ ""date-detected"": ""2020-05-29 (09:22)"",
+ ""detected-by"": ""Alias Robotics"",
+ ""detected-by-method"": ""testing static"",
+ ""date-reported"": ""2020-05-29 (09:22)"",
+ ""reported-by"": ""Alias Robotics"",
+ ""reported-by-relationship"": ""automatic"",
+ ""issue"": ""https://github.com/aliasrobotics/RVD/issues/2039"",
+ ""reproducibility"": ""always"",
+ ""trace"": ""/opt/ros_noetic_ws/src/ros/roslib/src/roslib/msgs.py:582"",
+ ""reproduction"": ""See artifacts below (if available)"",
+ ""reproduction-image"": """"
+ },
+ ""exploitation"": {
+ ""description"": """",
+ ""exploitation-image"": """",
+ ""exploitation-vector"": """"
+ },
+ ""mitigation"": {
+ ""description"": """",
+ ""pull-request"": """",
+ ""date-mitigation"": """"
+ }
+}
+```",1.0,"RVD#2039: Use of possibly insecure function - consider using safer ast., /opt/ros_noetic_ws/src/ros/roslib/src/roslib/msgs.py:582 - ```yaml
+{
+ ""id"": 2039,
+ ""title"": ""RVD#2039: Use of possibly insecure function - consider using safer ast., /opt/ros_noetic_ws/src/ros/roslib/src/roslib/msgs.py:582"",
+ ""type"": ""bug"",
+ ""description"": ""HIGH confidence of MEDIUM severity bug. Use of possibly insecure function - consider using safer ast.literal_eval. at /opt/ros_noetic_ws/src/ros/roslib/src/roslib/msgs.py:582 See links for more info on the bug."",
+ ""cwe"": ""None"",
+ ""cve"": ""None"",
+ ""keywords"": [
+ ""bandit"",
+ ""bug"",
+ ""static analysis"",
+ ""testing"",
+ ""triage"",
+ ""bug""
+ ],
+ ""system"": """",
+ ""vendor"": null,
+ ""severity"": {
+ ""rvss-score"": 0,
+ ""rvss-vector"": """",
+ ""severity-description"": """",
+ ""cvss-score"": 0,
+ ""cvss-vector"": """"
+ },
+ ""links"": [
+ ""https://github.com/aliasrobotics/RVD/issues/2039"",
+ ""https://bandit.readthedocs.io/en/latest/blacklists/blacklist_calls.html#b307-eval""
+ ],
+ ""flaw"": {
+ ""phase"": ""testing"",
+ ""specificity"": ""subject-specific"",
+ ""architectural-location"": ""application-specific"",
+ ""application"": ""N/A"",
+ ""subsystem"": ""N/A"",
+ ""package"": ""N/A"",
+ ""languages"": ""None"",
+ ""date-detected"": ""2020-05-29 (09:22)"",
+ ""detected-by"": ""Alias Robotics"",
+ ""detected-by-method"": ""testing static"",
+ ""date-reported"": ""2020-05-29 (09:22)"",
+ ""reported-by"": ""Alias Robotics"",
+ ""reported-by-relationship"": ""automatic"",
+ ""issue"": ""https://github.com/aliasrobotics/RVD/issues/2039"",
+ ""reproducibility"": ""always"",
+ ""trace"": ""/opt/ros_noetic_ws/src/ros/roslib/src/roslib/msgs.py:582"",
+ ""reproduction"": ""See artifacts below (if available)"",
+ ""reproduction-image"": """"
+ },
+ ""exploitation"": {
+ ""description"": """",
+ ""exploitation-image"": """",
+ ""exploitation-vector"": """"
+ },
+ ""mitigation"": {
+ ""description"": """",
+ ""pull-request"": """",
+ ""date-mitigation"": """"
+ }
+}
+```",0,rvd use of possibly insecure function consider using safer ast opt ros noetic ws src ros roslib src roslib msgs py yaml id title rvd use of possibly insecure function consider using safer ast opt ros noetic ws src ros roslib src roslib msgs py type bug description high confidence of medium severity bug use of possibly insecure function consider using safer ast literal eval at opt ros noetic ws src ros roslib src roslib msgs py see links for more info on the bug cwe none cve none keywords bandit bug static analysis testing triage bug system vendor null severity rvss score rvss vector severity description cvss score cvss vector links flaw phase testing specificity subject specific architectural location application specific application n a subsystem n a package n a languages none date detected detected by alias robotics detected by method testing static date reported reported by alias robotics reported by relationship automatic issue reproducibility always trace opt ros noetic ws src ros roslib src roslib msgs py reproduction see artifacts below if available reproduction image exploitation description exploitation image exploitation vector mitigation description pull request date mitigation ,0
+1042,25169615698.0,IssuesEvent,2022-11-11 01:11:24,jongfeel/BookReview,https://api.github.com/repos/jongfeel/BookReview,closed,4부 2장 비판을 하면서도 미움받지 않는 법,2022 How to Win Friends & Influence People,"### 2장 비판을 하면서도 미움받지 않는 법
+
+���칙 2: 사람들의 잘못을 간접적으로 지적하라.
+Call attention to people’s mistakes indirectly.",1.0,"4부 2장 비판을 하면서도 미움받지 않는 법 - ### 2장 비판을 하면서도 미움받지 않는 법
+
+규칙 2: 사람들의 잘못을 간접적으로 지적하라.
+Call attention to people’s mistakes indirectly.",1, 비판을 하면서도 미움받지 않는 법 비판을 하면서도 미움받지 않는 법 규칙 사람들의 잘못을 간접적으로 지적하라 call attention to people’s mistakes indirectly ,1
+611,10682541630.0,IssuesEvent,2019-10-22 05:50:35,data2health/website,https://api.github.com/repos/data2health/website,closed,add additional photos to website,enhancement people workgroup,"https://ctsa.ncats.nih.gov/cd2h/workgroup/people-expertise-attribution/
+
+All of the images are very clinical - would be great to get some real-life images of teams working together, a workshop, etc.
+",1.0,"add additional photos to website - https://ctsa.ncats.nih.gov/cd2h/workgroup/people-expertise-attribution/
+
+All of the images are very clinical - would be great to get some real-life images of teams working together, a workshop, etc.
+",1,add additional photos to website all of the images are very clinical would be great to get some real life images of teams working together a workshop etc ,1
+597669,18168587428.0,IssuesEvent,2021-09-27 17:10:45,UniVE-SSV/lisa,https://api.github.com/repos/UniVE-SSV/lisa,opened,[FEATURE REQUEST] Exception/error handlers,enhancement priority-p2,"**Description**
+Most programming languages have constructs to define code blocks to catch and handle runtime errors. There is currently no way to define these in cfgs.
+",1.0,"[FEATURE REQUEST] Exception/error handlers - **Description**
+Most programming languages have constructs to define code blocks to catch and handle runtime errors. There is currently no way to define these in cfgs.
+",0, exception error handlers description most programming languages have constructs to define code blocks to catch and handle runtime errors there is currently no way to define these in cfgs ,0
+156480,24624238208.0,IssuesEvent,2022-10-16 09:57:31,dotnet/efcore,https://api.github.com/repos/dotnet/efcore,closed,Make IMigrationsModelDiffer a provider service,type-enhancement closed-by-design,"This will allow providers to create custom model differs without clobbering each other.
+",1.0,"Make IMigrationsModelDiffer a provider service - This will allow providers to create custom model differs without clobbering each other.
+",0,make imigrationsmodeldiffer a provider service this will allow providers to create custom model differs without clobbering each other ,0
+36955,9933339449.0,IssuesEvent,2019-07-02 12:06:42,jupyterlab/jupyterlab,https://api.github.com/repos/jupyterlab/jupyterlab,closed,Enable Publish from Windows,tag:Build System tag:DevOps type:Maintenance,"We should support publishing from a Windows machine. We ran into https://bugs.python.org/issue31226 when trying to publish 1.0.0.
+
+For dev_mode, we should have a `clean:node` command that removes all node_modules. This should be called before trying to create the `sdist` if on Windows: https://github.com/jupyterlab/jupyterlab/blob/e2fd4c8841a7393f9131fb6b6e8252864a8bd351/buildutils/src/publish.ts#L53.
+
+We can refactor the logic [here](https://github.com/jupyterlab/jupyterlab/blob/master/clean.py#L8) used for `clean:slate` into a node module that also calls `rmdir` in a child_process.
+",1.0,"Enable Publish from Windows - We should support publishing from a Windows machine. We ran into https://bugs.python.org/issue31226 when trying to publish 1.0.0.
+
+For dev_mode, we should have a `clean:node` command that removes all node_modules. This should be called before trying to create the `sdist` if on Windows: https://github.com/jupyterlab/jupyterlab/blob/e2fd4c8841a7393f9131fb6b6e8252864a8bd351/buildutils/src/publish.ts#L53.
+
+We can refactor the logic [here](https://github.com/jupyterlab/jupyterlab/blob/master/clean.py#L8) used for `clean:slate` into a node module that also calls `rmdir` in a child_process.
+",0,enable publish from windows we should support publishing from a windows machine we ran into when trying to publish for dev mode we should have a clean node command that removes all node modules this should be called before trying to create the sdist if on windows we can refactor the logic used for clean slate into a node module that also calls rmdir in a child process ,0
+17140,9631572252.0,IssuesEvent,2019-05-15 14:29:07,EvotecIT/PSWinDocumentation,https://api.github.com/repos/EvotecIT/PSWinDocumentation,closed,Issue: Section Skip Broken or Missing Logic,enhancement performance,"First, love what you've built here, particularly in the latest version with the ability to provide a config file for runtime.
+
+Issue:
+You provide the option to bypass sections via $Document, however code still runs unless all related elements are also bypassed.
+
+Runtime Environment:
+Windows 10 v1803
+PowerShell v5.1
+Module Versions - Latest of all dependent modules, pulled directly from GitHub
+
+Expectation:
+Setting non-primary items without sub-dependencies as $false for inclusion should result in those portions of the data gathering process to be bypassed unless they are a dependency for another item.
+
+Example:
+I set 'SectionDomainOrganizationalUnits' to $true and 'SectionExcelDomainOrganizationalUnitsBasicACL' and 'SectionExcelDomainOrganizationalUnitsExtended' to $false with the expectation that the basic OU information would be collected, but the ACLs would not. Based on what I interpret from the script logic, selecting one will always include all. I can see why you need to get the base org units to support permissions, however there is no reason to require permissions when all I want is a list of org units.
+
+It would be great if:
+a) the $documents schema were arranged to more clearly show dependency relationships, possibly by making child dependencies sub-sections within the config (ie SectionDomainOrganizationalUnits for parent and SectionDomainOrganizationalUnits.SubSectionBaseACL and SubSectionExtACL for depedencies)
+b) when setting a parent item to $false, but setting a child dependent item to $true, prompt the user to tell them of the dependency and ask if they want to go ahead and collect the required parent, or skip the child item
+c) Alternative: code to collect a reduced set of values or data when excluding the parent, but including a child (ie Including SectionExcelDomainUsersNeverExpiring should retrieve only the minimum required attributes rather than always getting all users with all properties, which causes performance problems *see other issue*)",True,"Issue: Section Skip Broken or Missing Logic - First, love what you've built here, particularly in the latest version with the ability to provide a config file for runtime.
+
+Issue:
+You provide the option to bypass sections via $Document, however code still runs unless all related elements are also bypassed.
+
+Runtime Environment:
+Windows 10 v1803
+PowerShell v5.1
+Module Versions - Latest of all dependent modules, pulled directly from GitHub
+
+Expectation:
+Setting non-primary items without sub-dependencies as $false for inclusion should result in those portions of the data gathering process to be bypassed unless they are a dependency for another item.
+
+Example:
+I set 'SectionDomainOrganizationalUnits' to $true and 'SectionExcelDomainOrganizationalUnitsBasicACL' and 'SectionExcelDomainOrganizationalUnitsExtended' to $false with the expectation that the basic OU information would be collected, but the ACLs would not. Based on what I interpret from the script logic, selecting one will always include all. I can see why you need to get the base org units to support permissions, however there is no reason to require permissions when all I want is a list of org units.
+
+It would be great if:
+a) the $documents schema were arranged to more clearly show dependency relationships, possibly by making child dependencies sub-sections within the config (ie SectionDomainOrganizationalUnits for parent and SectionDomainOrganizationalUnits.SubSectionBaseACL and SubSectionExtACL for depedencies)
+b) when setting a parent item to $false, but setting a child dependent item to $true, prompt the user to tell them of the dependency and ask if they want to go ahead and collect the required parent, or skip the child item
+c) Alternative: code to collect a reduced set of values or data when excluding the parent, but including a child (ie Including SectionExcelDomainUsersNeverExpiring should retrieve only the minimum required attributes rather than always getting all users with all properties, which causes performance problems *see other issue*)",0,issue section skip broken or missing logic first love what you ve built here particularly in the latest version with the ability to provide a config file for runtime issue you provide the option to bypass sections via document however code still runs unless all related elements are also bypassed runtime environment windows powershell module versions latest of all dependent modules pulled directly from github expectation setting non primary items without sub dependencies as false for inclusion should result in those portions of the data gathering process to be bypassed unless they are a dependency for another item example i set sectiondomainorganizationalunits to true and sectionexceldomainorganizationalunitsbasicacl and sectionexceldomainorganizationalunitsextended to false with the expectation that the basic ou information would be collected but the acls would not based on what i interpret from the script logic selecting one will always include all i can see why you need to get the base org units to support permissions however there is no reason to require permissions when all i want is a list of org units it would be great if a the documents schema were arranged to more clearly show dependency relationships possibly by making child dependencies sub sections within the config ie sectiondomainorganizationalunits for parent and sectiondomainorganizationalunits subsectionbaseacl and subsectionextacl for depedencies b when setting a parent item to false but setting a child dependent item to true prompt the user to tell them of the dependency and ask if they want to go ahead and collect the required parent or skip the child item c alternative code to collect a reduced set of values or data when excluding the parent but including a child ie including sectionexceldomainusersneverexpiring should retrieve only the minimum required attributes rather than always getting all users with all properties which causes performance problems see other issue ,0
+1105,26943779063.0,IssuesEvent,2023-02-08 05:53:47,ballerina-platform/ballerina-extended-library,https://api.github.com/repos/ballerina-platform/ballerina-extended-library,closed,[Feature]: Employee Custom Screen,Type/New Feature Team/Connector Component/Connector module/peoplehr Resolution/Done,"### Connector Name
+
+module/peoplehr (People HR)
+
+### Suggested feature
+
+We have requirement to get custom screen data of an employee.
+
+This is the API Support from People HR [[1](https://apidocs.peoplehr.com/#tag/Employee-Custom-Screen)]
+Following are the method required.
+
+
+
+[1] [https://apidocs.peoplehr.com/#tag/Employee-Custom-Screen](https://apidocs.peoplehr.com/#tag/Employee-Custom-Screen)
+
+### Related issues
+
+_No response_",1.0,"[Feature]: Employee Custom Screen - ### Connector Name
+
+module/peoplehr (People HR)
+
+### Suggested feature
+
+We have requirement to get custom screen data of an employee.
+
+This is the API Support from People HR [[1](https://apidocs.peoplehr.com/#tag/Employee-Custom-Screen)]
+Following are the method required.
+
+
+
+[1] [https://apidocs.peoplehr.com/#tag/Employee-Custom-Screen](https://apidocs.peoplehr.com/#tag/Employee-Custom-Screen)
+
+### Related issues
+
+_No response_",1, employee custom screen connector name module peoplehr people hr suggested feature we have requirement to get custom screen data of an employee this is the api support from people hr following are the method required img width alt screenshot at pm src related issues no response ,1
+158605,20028801937.0,IssuesEvent,2022-02-02 01:17:19,ibm-skills-network/editor.md,https://api.github.com/repos/ibm-skills-network/editor.md,opened,CVE-2021-41182 (Medium) detected in jquery-ui-1.12.0.min.js,security vulnerability,"## CVE-2021-41182 - Medium Severity Vulnerability
+ Vulnerable Library - jquery-ui-1.12.0.min.js
+
+
A curated set of user interface interactions, effects, widgets, and themes built on top of the jQuery JavaScript Library.
+
+jQuery-UI is the official jQuery user interface library. Prior to version 1.13.0, accepting the value of the `altField` option of the Datepicker widget from untrusted sources may execute untrusted code. The issue is fixed in jQuery UI 1.13.0. Any string value passed to the `altField` option is now treated as a CSS selector. A workaround is to not accept the value of the `altField` option from untrusted sources.
+
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",True,"CVE-2021-41182 (Medium) detected in jquery-ui-1.12.0.min.js - ## CVE-2021-41182 - Medium Severity Vulnerability
+ Vulnerable Library - jquery-ui-1.12.0.min.js
+
+
A curated set of user interface interactions, effects, widgets, and themes built on top of the jQuery JavaScript Library.
+
+jQuery-UI is the official jQuery user interface library. Prior to version 1.13.0, accepting the value of the `altField` option of the Datepicker widget from untrusted sources may execute untrusted code. The issue is fixed in jQuery UI 1.13.0. Any string value passed to the `altField` option is now treated as a CSS selector. A workaround is to not accept the value of the `altField` option from untrusted sources.
+
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",0,cve medium detected in jquery ui min js cve medium severity vulnerability vulnerable library jquery ui min js a curated set of user interface interactions effects widgets and themes built on top of the jquery javascript library library home page a href path to dependency file lib codemirror mode slim index html path to vulnerable library lib codemirror mode slim index html dependency hierarchy x jquery ui min js vulnerable library found in base branch master vulnerability details jquery ui is the official jquery user interface library prior to version accepting the value of the altfield option of the datepicker widget from untrusted sources may execute untrusted code the issue is fixed in jquery ui any string value passed to the altfield option is now treated as a css selector a workaround is to not accept the value of the altfield option from untrusted sources publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery ui step up your open source security game with whitesource ,0
+395465,11686713508.0,IssuesEvent,2020-03-05 11:23:30,rapsaGnauJ/Mega-Man-Space-2,https://api.github.com/repos/rapsaGnauJ/Mega-Man-Space-2,closed,[FEATURE REQUEST] Pause menu,feature low priority request,"**Is your feature request related to a problem? Please describe.**
+A simple whole screen menu. Active afther pressing start on the title screen and when pausing the game (except on a stage, as the weapon change menu opens first and then this one).
+
+**Describe the solution you'd like**
+This menu options include:
+- New game: opens the difficulty selection menu. Only on the title game screen.
+- Resume: go back to the game to the previous menu instead.
+- Configuration: open the settings menu on top of this menu (Issue #15).
+- Exit stage: go to the select stage scene. Only visible if the game is on a stage.
+- Exit the game: closes the game.
+
+**Describe alternatives you've considered**
+
+**Additional context**
+The theme of the menu should be dependent on the screen it is on. On the title screen, there msut be no background. While in-game, the background should be like the weapon change menu background.",1.0,"[FEATURE REQUEST] Pause menu - **Is your feature request related to a problem? Please describe.**
+A simple whole screen menu. Active afther pressing start on the title screen and when pausing the game (except on a stage, as the weapon change menu opens first and then this one).
+
+**Describe the solution you'd like**
+This menu options include:
+- New game: opens the difficulty selection menu. Only on the title game screen.
+- Resume: go back to the game to the previous menu instead.
+- Configuration: open the settings menu on top of this menu (Issue #15).
+- Exit stage: go to the select stage scene. Only visible if the game is on a stage.
+- Exit the game: closes the game.
+
+**Describe alternatives you've considered**
+
+**Additional context**
+The theme of the menu should be dependent on the screen it is on. On the title screen, there msut be no background. While in-game, the background should be like the weapon change menu background.",0, pause menu is your feature request related to a problem please describe a simple whole screen menu active afther pressing start on the title screen and when pausing the game except on a stage as the weapon change menu opens first and then this one describe the solution you d like this menu options include new game opens the difficulty selection menu only on the title game screen resume go back to the game to the previous menu instead configuration open the settings menu on top of this menu issue exit stage go to the select stage scene only visible if the game is on a stage exit the game closes the game describe alternatives you ve considered additional context the theme of the menu should be dependent on the screen it is on on the title screen there msut be no background while in game the background should be like the weapon change menu background ,0
+600,10598053328.0,IssuesEvent,2019-10-10 03:12:40,DimensionDev/Maskbook,https://api.github.com/repos/DimensionDev/Maskbook,closed,Feature: Backup alert,Component: Database Component: People Effort: Moderate Priority: P5 (plan in future) Type: Feature,"
+
+Such a backup alert shall appear if:
+
+- Last saved backup was 168 hours ago; or
+- No backup saved ever
+
+... and:
+
+- The last time the user manually clicked ""close"" was within the last 168 hours.
+
+The alert text shall differ according to the case.
+
+- `Last backup was 7 days ago.`
+- `Create your first backup!`
+
+Notice:
+
+- There can be saved backup (saved) and unsaved backup (user clicked ""cancel"").
+- Regardless of manual closings, the time span for ""X days"" always starts from last saved backup.",1.0,"Feature: Backup alert - 
+
+Such a backup alert shall appear if:
+
+- Last saved backup was 168 hours ago; or
+- No backup saved ever
+
+... and:
+
+- The last time the user manually clicked ""close"" was within the last 168 hours.
+
+The alert text shall differ according to the case.
+
+- `Last backup was 7 days ago.`
+- `Create your first backup!`
+
+Notice:
+
+- There can be saved backup (saved) and unsaved backup (user clicked ""cancel"").
+- Regardless of manual closings, the time span for ""X days"" always starts from last saved backup.",1,feature backup alert such a backup alert shall appear if last saved backup was hours ago or no backup saved ever and the last time the user manually clicked close was within the last hours the alert text shall differ according to the case last backup was days ago create your first backup notice there can be saved backup saved and unsaved backup user clicked cancel regardless of manual closings the time span for x days always starts from last saved backup ,1
+45045,9667772764.0,IssuesEvent,2019-05-21 13:54:33,IBAIT18/group5,https://api.github.com/repos/IBAIT18/group5,closed,GUI mit Datenbank verknüpfen,GUI Hintergrundcode,Eingaben aus der GUI sollen in der Datenbank gespeichert und aus der Datenbank gezogen werden können.,1.0,GUI mit Datenbank verknüpfen - Eingaben aus der GUI sollen in der Datenbank gespeichert und aus der Datenbank gezogen werden können.,0,gui mit datenbank verknüpfen eingaben aus der gui sollen in der datenbank gespeichert und aus der datenbank gezogen werden können ,0
+911,19318143902.0,IssuesEvent,2021-12-14 00:19:19,schmave/demschooltools,https://api.github.com/repos/schmave/demschooltools,closed,Google/FB SSO should ignore case when comparing email addresses; also fix DUMMY_USERNAME handling,bug people easy,"If an admin enters ""foo@Ourschool.com"" but Google tells us that your email address is ""foo@ourschool.com"", then DST won't let you in. That's a bit confusing.
+
+As an additional way to reduce confusion, it should be impossible to create a user ""foo@Ourschool"" when ""foo@ourschool"" already exists. And the error message shown when that happens should be way better than it is now.
+
+Also, if there is a user with DUMMY_USERNAME and their email conflicts, the user you are trying to create should replace the DUMMY_USERNAME one.",1.0,"Google/FB SSO should ignore case when comparing email addresses; also fix DUMMY_USERNAME handling - If an admin enters ""foo@Ourschool.com"" but Google tells us that your email address is ""foo@ourschool.com"", then DST won't let you in. That's a bit confusing.
+
+As an additional way to reduce confusion, it should be impossible to create a user ""foo@Ourschool"" when ""foo@ourschool"" already exists. And the error message shown when that happens should be way better than it is now.
+
+Also, if there is a user with DUMMY_USERNAME and their email conflicts, the user you are trying to create should replace the DUMMY_USERNAME one.",1,google fb sso should ignore case when comparing email addresses also fix dummy username handling if an admin enters foo ourschool com but google tells us that your email address is foo ourschool com then dst won t let you in that s a bit confusing as an additional way to reduce confusion it should be impossible to create a user foo ourschool when foo ourschool already exists and the error message shown when that happens should be way better than it is now also if there is a user with dummy username and their email conflicts the user you are trying to create should replace the dummy username one ,1
+297838,25766892576.0,IssuesEvent,2022-12-09 03:06:25,Siwon-L/SearchBooks,https://api.github.com/repos/Siwon-L/SearchBooks,closed,Domain 객층 구현,feature test,"- [x] Repository 구현
+- [x] Repository Test Case 작성
+- [x] Model Type 구현
+- [x] UseCase 구현
+- [x] UseCase Test Case 작성",1.0,"Domain 객층 구현 - - [x] Repository 구현
+- [x] Repository Test Case 작성
+- [x] Model Type 구현
+- [x] UseCase 구현
+- [x] UseCase Test Case 작성",0,domain 객층 구현 repository 구현 repository test case 작성 model type 구현 usecase 구현 usecase test case 작성,0
+70627,18243181857.0,IssuesEvent,2021-10-01 15:06:14,spack/spack,https://api.github.com/repos/spack/spack,closed,Installation issue: llvm-openmp for M1,macOS build-error arm,"### Steps to reproduce the issue
+
+On the new macOS M1 (aarch64) systems with BigSur, compiling llvm-openmp does not work with the provided Xcode (AppleClang 12.0.5).
+
+```console
+space install llvm-openmp
+```
+
+leads to:
+```
+cd /private/var/folders/wy/w22w0mkx67711r3m3t98_g7m0000gn/T/axel/spack-stage/spack-stage-llvm-openmp-9.0.0-ikuqdzdhbfgvj7lcsubvgb24rkhwq7wc
+ /spack-build-ikuqdzd/runtime/src && /Users/axel/src/spack/lib/spack/env/clang/clang -Domp_EXPORTS -I/private/var/folders/wy/w22w0mkx67711r3
+ m3t98_g7m0000gn/T/axel/spack-stage/spack-stage-llvm-openmp-9.0.0-ikuqdzdhbfgvj7lcsubvgb24rkhwq7wc/spack-build-ikuqdzd/runtime/src -I/var/fo
+ lders/wy/w22w0mkx67711r3m3t98_g7m0000gn/T/axel/spack-stage/spack-stage-llvm-openmp-9.0.0-ikuqdzdhbfgvj7lcsubvgb24rkhwq7wc/spack-src/runtime
+ /src -I/var/folders/wy/w22w0mkx67711r3m3t98_g7m0000gn/T/axel/spack-stage/spack-stage-llvm-openmp-9.0.0-ikuqdzdhbfgvj7lcsubvgb24rkhwq7wc/spa
+ ck-src/runtime/src/i18n -I/var/folders/wy/w22w0mkx67711r3m3t98_g7m0000gn/T/axel/spack-stage/spack-stage-llvm-openmp-9.0.0-ikuqdzdhbfgvj7lcs
+ ubvgb24rkhwq7wc/spack-src/runtime/src/include -I/var/folders/wy/w22w0mkx67711r3m3t98_g7m0000gn/T/axel/spack-stage/spack-stage-llvm-openmp-9
+ .0.0-ikuqdzdhbfgvj7lcsubvgb24rkhwq7wc/spack-src/runtime/src/thirdparty/ittnotify -O2 -g -DNDEBUG -arch arm64 -isysroot /Library/Developer/C
+ ommandLineTools/SDKs/MacOSX11.3.sdk -fPIC -D _GNU_SOURCE -D _REENTRANT -x assembler-with-cpp -MD -MT runtime/src/CMakeFiles/omp.dir/z_Lin
+ ux_asm.S.o -MF CMakeFiles/omp.dir/z_Linux_asm.S.o.d -o CMakeFiles/omp.dir/z_Linux_asm.S.o -c /var/folders/wy/w22w0mkx67711r3m3t98_g7m0000gn
+ /T/axel/spack-stage/spack-stage-llvm-openmp-9.0.0-ikuqdzdhbfgvj7lcsubvgb24rkhwq7wc/spack-src/runtime/src/z_Linux_asm.S
+ >> 221 /var/folders/wy/w22w0mkx67711r3m3t98_g7m0000gn/T/axel/spack-stage/spack-stage-llvm-openmp-9.0.0-ikuqdzdhbfgvj7lcsubvgb24rkhwq7wc/spack-src/
+ runtime/src/z_Linux_asm.S:1546:5: error: unknown directive
+ 222 .size __kmp_unnamed_critical_addr,8
+ 223 ^
+ >> 224 make[2]: *** [runtime/src/CMakeFiles/omp.dir/z_Linux_asm.S.o] Error 1
+ 225 make[2]: *** Waiting for unfinished jobs....
+ >> 226 make[1]: *** [runtime/src/CMakeFiles/omp.dir/all] Error 2
+ >> 227 make: *** [all] Error 2
+```
+
+### Information on your system
+
+* **Spack:** 0.16.3-4508-2b5893c932
+* **Python:** 3.9.7
+* **Platform:** darwin-bigsur-m1
+* **Concretizer:** clingo
+
+`compilers.yaml:`
+```
+compilers:
+- compiler:
+ spec: apple-clang@12.0.5
+ paths:
+ cc: /usr/bin/clang
+ cxx: /usr/bin/clang++
+ f77: /opt/homebrew/Cellar/gcc/11.2.0/bin/gfortran
+ fc: /opt/homebrew/Cellar/gcc/11.2.0/bin/gfortran
+ flags: {}
+ operating_system: bigsur
+ target: aarch64
+ modules: []
+ environment: {}
+ extra_rpaths: []
+```
+
+### Additional information
+
+* [spack-configure-args.txt](https://github.com/spack/spack/files/7248062/spack-configure-args.txt)
+* [spack-build-01-cmake-out.txt](https://github.com/spack/spack/files/7248058/spack-build-01-cmake-out.txt)
+* [spack-build-02-build-out.txt](https://github.com/spack/spack/files/7248059/spack-build-02-build-out.txt)
+* [spack-build-env.txt](https://github.com/spack/spack/files/7248060/spack-build-env.txt)
+* [spack-build-out.txt](https://github.com/spack/spack/files/7248061/spack-build-out.txt)
+
+
+### General information
+
+- [X] I have run `spack debug report` and reported the version of Spack/Python/Platform
+- [X] I have run `spack maintainers ` and **@mentioned** any maintainers
+- [X] I have uploaded the build log and environment files
+- [X] I have searched the issues of this repo and believe this is not a duplicate",1.0,"Installation issue: llvm-openmp for M1 - ### Steps to reproduce the issue
+
+On the new macOS M1 (aarch64) systems with BigSur, compiling llvm-openmp does not work with the provided Xcode (AppleClang 12.0.5).
+
+```console
+space install llvm-openmp
+```
+
+leads to:
+```
+cd /private/var/folders/wy/w22w0mkx67711r3m3t98_g7m0000gn/T/axel/spack-stage/spack-stage-llvm-openmp-9.0.0-ikuqdzdhbfgvj7lcsubvgb24rkhwq7wc
+ /spack-build-ikuqdzd/runtime/src && /Users/axel/src/spack/lib/spack/env/clang/clang -Domp_EXPORTS -I/private/var/folders/wy/w22w0mkx67711r3
+ m3t98_g7m0000gn/T/axel/spack-stage/spack-stage-llvm-openmp-9.0.0-ikuqdzdhbfgvj7lcsubvgb24rkhwq7wc/spack-build-ikuqdzd/runtime/src -I/var/fo
+ lders/wy/w22w0mkx67711r3m3t98_g7m0000gn/T/axel/spack-stage/spack-stage-llvm-openmp-9.0.0-ikuqdzdhbfgvj7lcsubvgb24rkhwq7wc/spack-src/runtime
+ /src -I/var/folders/wy/w22w0mkx67711r3m3t98_g7m0000gn/T/axel/spack-stage/spack-stage-llvm-openmp-9.0.0-ikuqdzdhbfgvj7lcsubvgb24rkhwq7wc/spa
+ ck-src/runtime/src/i18n -I/var/folders/wy/w22w0mkx67711r3m3t98_g7m0000gn/T/axel/spack-stage/spack-stage-llvm-openmp-9.0.0-ikuqdzdhbfgvj7lcs
+ ubvgb24rkhwq7wc/spack-src/runtime/src/include -I/var/folders/wy/w22w0mkx67711r3m3t98_g7m0000gn/T/axel/spack-stage/spack-stage-llvm-openmp-9
+ .0.0-ikuqdzdhbfgvj7lcsubvgb24rkhwq7wc/spack-src/runtime/src/thirdparty/ittnotify -O2 -g -DNDEBUG -arch arm64 -isysroot /Library/Developer/C
+ ommandLineTools/SDKs/MacOSX11.3.sdk -fPIC -D _GNU_SOURCE -D _REENTRANT -x assembler-with-cpp -MD -MT runtime/src/CMakeFiles/omp.dir/z_Lin
+ ux_asm.S.o -MF CMakeFiles/omp.dir/z_Linux_asm.S.o.d -o CMakeFiles/omp.dir/z_Linux_asm.S.o -c /var/folders/wy/w22w0mkx67711r3m3t98_g7m0000gn
+ /T/axel/spack-stage/spack-stage-llvm-openmp-9.0.0-ikuqdzdhbfgvj7lcsubvgb24rkhwq7wc/spack-src/runtime/src/z_Linux_asm.S
+ >> 221 /var/folders/wy/w22w0mkx67711r3m3t98_g7m0000gn/T/axel/spack-stage/spack-stage-llvm-openmp-9.0.0-ikuqdzdhbfgvj7lcsubvgb24rkhwq7wc/spack-src/
+ runtime/src/z_Linux_asm.S:1546:5: error: unknown directive
+ 222 .size __kmp_unnamed_critical_addr,8
+ 223 ^
+ >> 224 make[2]: *** [runtime/src/CMakeFiles/omp.dir/z_Linux_asm.S.o] Error 1
+ 225 make[2]: *** Waiting for unfinished jobs....
+ >> 226 make[1]: *** [runtime/src/CMakeFiles/omp.dir/all] Error 2
+ >> 227 make: *** [all] Error 2
+```
+
+### Information on your system
+
+* **Spack:** 0.16.3-4508-2b5893c932
+* **Python:** 3.9.7
+* **Platform:** darwin-bigsur-m1
+* **Concretizer:** clingo
+
+`compilers.yaml:`
+```
+compilers:
+- compiler:
+ spec: apple-clang@12.0.5
+ paths:
+ cc: /usr/bin/clang
+ cxx: /usr/bin/clang++
+ f77: /opt/homebrew/Cellar/gcc/11.2.0/bin/gfortran
+ fc: /opt/homebrew/Cellar/gcc/11.2.0/bin/gfortran
+ flags: {}
+ operating_system: bigsur
+ target: aarch64
+ modules: []
+ environment: {}
+ extra_rpaths: []
+```
+
+### Additional information
+
+* [spack-configure-args.txt](https://github.com/spack/spack/files/7248062/spack-configure-args.txt)
+* [spack-build-01-cmake-out.txt](https://github.com/spack/spack/files/7248058/spack-build-01-cmake-out.txt)
+* [spack-build-02-build-out.txt](https://github.com/spack/spack/files/7248059/spack-build-02-build-out.txt)
+* [spack-build-env.txt](https://github.com/spack/spack/files/7248060/spack-build-env.txt)
+* [spack-build-out.txt](https://github.com/spack/spack/files/7248061/spack-build-out.txt)
+
+
+### General information
+
+- [X] I have run `spack debug report` and reported the version of Spack/Python/Platform
+- [X] I have run `spack maintainers ` and **@mentioned** any maintainers
+- [X] I have uploaded the build log and environment files
+- [X] I have searched the issues of this repo and believe this is not a duplicate",0,installation issue llvm openmp for steps to reproduce the issue on the new macos systems with bigsur compiling llvm openmp does not work with the provided xcode appleclang console space install llvm openmp leads to cd private var folders wy t axel spack stage spack stage llvm openmp spack build ikuqdzd runtime src users axel src spack lib spack env clang clang domp exports i private var folders wy t axel spack stage spack stage llvm openmp spack build ikuqdzd runtime src i var fo lders wy t axel spack stage spack stage llvm openmp spack src runtime src i var folders wy t axel spack stage spack stage llvm openmp spa ck src runtime src i var folders wy t axel spack stage spack stage llvm openmp spack src runtime src include i var folders wy t axel spack stage spack stage llvm openmp spack src runtime src thirdparty ittnotify g dndebug arch isysroot library developer c ommandlinetools sdks sdk fpic d gnu source d reentrant x assembler with cpp md mt runtime src cmakefiles omp dir z lin ux asm s o mf cmakefiles omp dir z linux asm s o d o cmakefiles omp dir z linux asm s o c var folders wy t axel spack stage spack stage llvm openmp spack src runtime src z linux asm s var folders wy t axel spack stage spack stage llvm openmp spack src runtime src z linux asm s error unknown directive size kmp unnamed critical addr make error make waiting for unfinished jobs make error make error information on your system spack python platform darwin bigsur concretizer clingo compilers yaml compilers compiler spec apple clang paths cc usr bin clang cxx usr bin clang opt homebrew cellar gcc bin gfortran fc opt homebrew cellar gcc bin gfortran flags operating system bigsur target modules environment extra rpaths additional information general information i have run spack debug report and reported the version of spack python platform i have run spack maintainers and mentioned any maintainers i have uploaded the build log and environment files i have searched the issues of this repo and believe this is not a duplicate,0
+67179,8098943525.0,IssuesEvent,2018-08-11 01:07:09,voteview/WebVoteView,https://api.github.com/repos/voteview/WebVoteView,closed,Party colour schemes,design-enhancement,"I assigned colour schemes to parties earlier in development. The colour schemes are: red, orange, yellow, green, teal, blue, purple, pinkpurple, grey, brown
+
+You can see which colour each party is in based on the little left border strips in the party list, or by clicking into the party page and observing the scheme used:
+http://voteview.polisci.ucla.edu/parties/all
+
+I don't have the substantive knowledge to know if my colour choices were at all appropriate. The idea was to ensure that concurrent major parties would use different colours. Beyond that, I didn't really have any priorities.
+
+This issue is to allow people to review my colour choices and suggest if anything should be in a different colour category.
+
+In addition, we should consider reallocating people off yellow; yellow is not recommended for use on a white background. It's the only palette that doesn't come from ColorBrewer. It's generally not recommended. It's bad on low contrast screens, it goes bad with f.lux, etc.
+
+I'll close the issue if no one has any input. In addition, if Erik finds anything out about historical party colours in his research, I'll set them accordingly.
+",1.0,"Party colour schemes - I assigned colour schemes to parties earlier in development. The colour schemes are: red, orange, yellow, green, teal, blue, purple, pinkpurple, grey, brown
+
+You can see which colour each party is in based on the little left border strips in the party list, or by clicking into the party page and observing the scheme used:
+http://voteview.polisci.ucla.edu/parties/all
+
+I don't have the substantive knowledge to know if my colour choices were at all appropriate. The idea was to ensure that concurrent major parties would use different colours. Beyond that, I didn't really have any priorities.
+
+This issue is to allow people to review my colour choices and suggest if anything should be in a different colour category.
+
+In addition, we should consider reallocating people off yellow; yellow is not recommended for use on a white background. It's the only palette that doesn't come from ColorBrewer. It's generally not recommended. It's bad on low contrast screens, it goes bad with f.lux, etc.
+
+I'll close the issue if no one has any input. In addition, if Erik finds anything out about historical party colours in his research, I'll set them accordingly.
+",0,party colour schemes i assigned colour schemes to parties earlier in development the colour schemes are red orange yellow green teal blue purple pinkpurple grey brown you can see which colour each party is in based on the little left border strips in the party list or by clicking into the party page and observing the scheme used i don t have the substantive knowledge to know if my colour choices were at all appropriate the idea was to ensure that concurrent major parties would use different colours beyond that i didn t really have any priorities this issue is to allow people to review my colour choices and suggest if anything should be in a different colour category in addition we should consider reallocating people off yellow yellow is not recommended for use on a white background it s the only palette that doesn t come from colorbrewer it s generally not recommended it s bad on low contrast screens it goes bad with f lux etc i ll close the issue if no one has any input in addition if erik finds anything out about historical party colours in his research i ll set them accordingly ,0
+396,7204667811.0,IssuesEvent,2018-02-06 13:31:03,wordpress-mobile/WordPress-Android,https://api.github.com/repos/wordpress-mobile/WordPress-Android,opened,Invite roles are not working for the first time,People Management [Type] Broken Window,"After a fresh install, it looks like the roles won't be fetched for the first time a user tries to invite (and possibly change) roles. It only shows the follower role as the available option. We should probably refresh the page after roles are fetched or start the fetch early so it's ready for first use.",1.0,"Invite roles are not working for the first time - After a fresh install, it looks like the roles won't be fetched for the first time a user tries to invite (and possibly change) roles. It only shows the follower role as the available option. We should probably refresh the page after roles are fetched or start the fetch early so it's ready for first use.",1,invite roles are not working for the first time after a fresh install it looks like the roles won t be fetched for the first time a user tries to invite and possibly change roles it only shows the follower role as the available option we should probably refresh the page after roles are fetched or start the fetch early so it s ready for first use ,1
+681994,23330152659.0,IssuesEvent,2022-08-09 03:47:57,City-Bureau/city-scrapers-atl,https://api.github.com/repos/City-Bureau/city-scrapers-atl,closed,New Scraper: Clayton County Library Board,priority-unset,"Create a new scraper for Clayton County Library Board
+
+Website: https://claytonpl.org/about-us/library-board/
+Jurisdiction: Clayton County
+Classification: Basic Government
+
+",1.0,"New Scraper: Clayton County Library Board - Create a new scraper for Clayton County Library Board
+
+Website: https://claytonpl.org/about-us/library-board/
+Jurisdiction: Clayton County
+Classification: Basic Government
+
+",0,new scraper clayton county library board create a new scraper for clayton county library board website jurisdiction clayton county classification basic government ,0
+1003,24243239599.0,IssuesEvent,2022-09-27 08:31:39,microsoftgraph/microsoft-graph-toolkit,https://api.github.com/repos/microsoftgraph/microsoft-graph-toolkit,closed,[BUG] mgt-people component: Cannot read properties of null (reading 'displayName'),bug Component: people State: In Review,"
+
+
+**Describe the bug**
+When using mgt-people component by passing a list of user ids:
+```html
+
+```
+
+We can see the below error in console if some users have been removed and their IDs are invalid:
+
+
+
+The code throwing the error is here:
+
+https://github.com/microsoftgraph/microsoft-graph-toolkit/blob/b9c56bcd88a7d752c7e653a80884690d6b2eb2da/packages/mgt-components/src/graph/graph.user.ts#L199
+
+The code gets a user from cache, and the user could be null if it the ID is not valid anymore:
+
+
+
+**To Reproduce**
+Steps to reproduce the behavior:
+1. Pass an ID of a deleted user to mgt-people component
+
+Note that this doesn't guarantee a consistent reproduction. Let's say we pass the id ""invalid-user-id"" to mgt-people.
+
+Sometimes there won't be a cache with `invalid-user-id` as key and `{user: null}` as value in the IndexedDB, and mgt-people won't throw but just simply omitting this user.
+
+Sometimes there will be a cache item of `invalid-user-id` with `{user: null}` showing up in the IndexedDB, like the screenshot above, and then the error will occur.
+
+We also have tried using fallback-details but it doesn't help. As long as there is a `{user: null}` cache for the removed user, this error will happen.
+
+
+
+
+**Expected behavior**
+When there is an invalid user id, the mgt-people component should just ignore the user and not throw any error. If fallback-details is used, use the fallback user. If not, just omit it.
+
+**Environment (please complete the following information):**
+ - OS: Windows 11
+ - Browser: Edge
+ - Framework: Angular
+ - Context: Web
+ - Version: 2.6.0
+ - Provider: Msal2Provider
+",1.0,"[BUG] mgt-people component: Cannot read properties of null (reading 'displayName') -
+
+
+**Describe the bug**
+When using mgt-people component by passing a list of user ids:
+```html
+
+```
+
+We can see the below error in console if some users have been removed and their IDs are invalid:
+
+
+
+The code throwing the error is here:
+
+https://github.com/microsoftgraph/microsoft-graph-toolkit/blob/b9c56bcd88a7d752c7e653a80884690d6b2eb2da/packages/mgt-components/src/graph/graph.user.ts#L199
+
+The code gets a user from cache, and the user could be null if it the ID is not valid anymore:
+
+
+
+**To Reproduce**
+Steps to reproduce the behavior:
+1. Pass an ID of a deleted user to mgt-people component
+
+Note that this doesn't guarantee a consistent reproduction. Let's say we pass the id ""invalid-user-id"" to mgt-people.
+
+Sometimes there won't be a cache with `invalid-user-id` as key and `{user: null}` as value in the IndexedDB, and mgt-people won't throw but just simply omitting this user.
+
+Sometimes there will be a cache item of `invalid-user-id` with `{user: null}` showing up in the IndexedDB, like the screenshot above, and then the error will occur.
+
+We also have tried using fallback-details but it doesn't help. As long as there is a `{user: null}` cache for the removed user, this error will happen.
+
+
+
+
+**Expected behavior**
+When there is an invalid user id, the mgt-people component should just ignore the user and not throw any error. If fallback-details is used, use the fallback user. If not, just omit it.
+
+**Environment (please complete the following information):**
+ - OS: Windows 11
+ - Browser: Edge
+ - Framework: Angular
+ - Context: Web
+ - Version: 2.6.0
+ - Provider: Msal2Provider
+",1, mgt people component cannot read properties of null reading displayname describe the bug when using mgt people component by passing a list of user ids html we can see the below error in console if some users have been removed and their ids are invalid the code throwing the error is here the code gets a user from cache and the user could be null if it the id is not valid anymore to reproduce steps to reproduce the behavior pass an id of a deleted user to mgt people component note that this doesn t guarantee a consistent reproduction let s say we pass the id invalid user id to mgt people sometimes there won t be a cache with invalid user id as key and user null as value in the indexeddb and mgt people won t throw but just simply omitting this user sometimes there will be a cache item of invalid user id with user null showing up in the indexeddb like the screenshot above and then the error will occur we also have tried using fallback details but it doesn t help as long as there is a user null cache for the removed user this error will happen expected behavior when there is an invalid user id the mgt people component should just ignore the user and not throw any error if fallback details is used use the fallback user if not just omit it environment please complete the following information os windows browser edge framework angular context web version provider ,1
+314081,26975145714.0,IssuesEvent,2023-02-09 09:01:19,cockroachdb/cockroach,https://api.github.com/repos/cockroachdb/cockroach,opened,roachtest: scrub/index-only/tpcc/w=100 failed,C-test-failure O-robot O-roachtest branch-master release-blocker,"roachtest.scrub/index-only/tpcc/w=100 [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8641928?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8641928?buildTab=artifacts#/scrub/index-only/tpcc/w=100) on master @ [09188370d82e163ff1d44c62fe611104502c548d](https://github.com/cockroachdb/cockroach/commits/09188370d82e163ff1d44c62fe611104502c548d):
+
+
+```
+test artifacts and logs in: /artifacts/scrub/index-only/tpcc/w=100/run_1
+(monitor.go:127).Wait: monitor failure: monitor command failure: unexpected node event: 4: dead (exit status 7)
+```
+
+
+/cc @cockroachdb/sql-queries
+
+
+[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*scrub/index-only/tpcc/w=100.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
+
+",0,roachtest scrub index only tpcc w failed roachtest scrub index only tpcc w with on master test artifacts and logs in artifacts scrub index only tpcc w run monitor go wait monitor failure monitor command failure unexpected node event dead exit status parameters roachtest cloud gce roachtest cpu roachtest encrypted false roachtest ssd help see see cc cockroachdb sql queries ,0
+273,5328053076.0,IssuesEvent,2017-02-15 10:54:53,PopulateTools/gobierto,https://api.github.com/repos/PopulateTools/gobierto,closed,People/Person: send email through form,gobierto-people,"- [x] Markup
+- [x] Integration
+
+* Captcha: I guess we should use a Captcha since anybody should be able to send an email. Google seems to be working in an invisible captcha: http://www.ghacks.net/2016/12/05/googles-invisible-captchas/ (but is not ready yet: https://www.google.com/recaptcha/intro/comingsoon/invisiblebeta.html). Also: https://github.com/markets/invisible_captcha
+* Destination email: for a further iteration, we may offer a conf option so this emails are received in a different email than the user's one. ",1.0,"People/Person: send email through form - - [x] Markup
+- [x] Integration
+
+* Captcha: I guess we should use a Captcha since anybody should be able to send an email. Google seems to be working in an invisible captcha: http://www.ghacks.net/2016/12/05/googles-invisible-captchas/ (but is not ready yet: https://www.google.com/recaptcha/intro/comingsoon/invisiblebeta.html). Also: https://github.com/markets/invisible_captcha
+* Destination email: for a further iteration, we may offer a conf option so this emails are received in a different email than the user's one. ",1,people person send email through form markup integration captcha i guess we should use a captcha since anybody should be able to send an email google seems to be working in an invisible captcha but is not ready yet also destination email for a further iteration we may offer a conf option so this emails are received in a different email than the user s one ,1
+19702,27341409742.0,IssuesEvent,2023-02-26 20:50:47,ValveSoftware/Proton,https://api.github.com/repos/ValveSoftware/Proton,closed,"Warhammer 40,000: Darktide Closed Beta (2156210)",Game compatibility - Unofficial .NET,"Hello ! the microsoft edge webview2 installer dont work to proton experimental and 7.0.4, and when the launcher run the buttons,play,setting,etc dont appear so from the proton 7.0.4 the buttons appears but the laucher disappear instantly, with 6.0.3 the launcher stay but impossible clicks to the buttons.
+My config : Intel(R) Core(TM) i7-4790K CPU @ 4.00GHz, Nvidia RTX 3050 VENTUS x2, 16 GO DDR3, ARCH LINUX
+Kernel 6.0.1 64 bit.
+
+[steam-2156210.log](https://github.com/ValveSoftware/Proton/files/9791005/steam-2156210.log)
+",True,"Warhammer 40,000: Darktide Closed Beta (2156210) - Hello ! the microsoft edge webview2 installer dont work to proton experimental and 7.0.4, and when the launcher run the buttons,play,setting,etc dont appear so from the proton 7.0.4 the buttons appears but the laucher disappear instantly, with 6.0.3 the launcher stay but impossible clicks to the buttons.
+My config : Intel(R) Core(TM) i7-4790K CPU @ 4.00GHz, Nvidia RTX 3050 VENTUS x2, 16 GO DDR3, ARCH LINUX
+Kernel 6.0.1 64 bit.
+
+[steam-2156210.log](https://github.com/ValveSoftware/Proton/files/9791005/steam-2156210.log)
+",0,warhammer darktide closed beta hello the microsoft edge installer dont work to proton experimental and and when the launcher run the buttons play setting etc dont appear so from the proton the buttons appears but the laucher disappear instantly with the launcher stay but impossible clicks to the buttons my config intel r core tm cpu nvidia rtx ventus go arch linux kernel bit ,0
+155438,13625112646.0,IssuesEvent,2020-09-24 09:02:22,Staff-Alerts/project,https://api.github.com/repos/Staff-Alerts/project,closed,Document team roles and responsibilities,documentation high priority,"**What:** Create a short document showing the team roles and responsibilities
+
+**Why;** So it is clear who leads what during the work.",1.0,"Document team roles and responsibilities - **What:** Create a short document showing the team roles and responsibilities
+
+**Why;** So it is clear who leads what during the work.",0,document team roles and responsibilities what create a short document showing the team roles and responsibilities why so it is clear who leads what during the work ,0
+697,12511669336.0,IssuesEvent,2020-06-02 21:01:21,microsoftgraph/microsoft-graph-toolkit,https://api.github.com/repos/microsoftgraph/microsoft-graph-toolkit,closed,[People-picker] add a defaultSelectedUserIds property for setting initial state,//b Component: people-picker Needs: Triage :mag: Priority: 1 State: Committed feature-request,"## Description
+Add a new property/attribute on `mgt-people-picker` to set initially selected people by user id.
+
+## Rationale
+To select people by user id, the developer needs to use the `selectUsersByIds` method. Selecting people declaratively must be done with the `selectedPeople` property. Keeping track of objects is not as ideal as just keep track of user ids, and calling a method is not ideal. This issue proposes to add a new property to set the initial selection state declaratively via user ids
+
+## Preferred Solution
+Add a new property `defaultSelectedUserIds` (and matching attribute `default-selected-user-ids`) that will set the initial state of the component. Changing this property after the component has loaded will have no effect on the UI, and changes in the selection state will have no effect on the property. The `selectedPeople` property will continue to be used as the true state of the component.
+
+```
+
+```",1.0,"[People-picker] add a defaultSelectedUserIds property for setting initial state - ## Description
+Add a new property/attribute on `mgt-people-picker` to set initially selected people by user id.
+
+## Rationale
+To select people by user id, the developer needs to use the `selectUsersByIds` method. Selecting people declaratively must be done with the `selectedPeople` property. Keeping track of objects is not as ideal as just keep track of user ids, and calling a method is not ideal. This issue proposes to add a new property to set the initial selection state declaratively via user ids
+
+## Preferred Solution
+Add a new property `defaultSelectedUserIds` (and matching attribute `default-selected-user-ids`) that will set the initial state of the component. Changing this property after the component has loaded will have no effect on the UI, and changes in the selection state will have no effect on the property. The `selectedPeople` property will continue to be used as the true state of the component.
+
+```
+
+```",1, add a defaultselecteduserids property for setting initial state description add a new property attribute on mgt people picker to set initially selected people by user id rationale to select people by user id the developer needs to use the selectusersbyids method selecting people declaratively must be done with the selectedpeople property keeping track of objects is not as ideal as just keep track of user ids and calling a method is not ideal this issue proposes to add a new property to set the initial selection state declaratively via user ids preferred solution add a new property defaultselecteduserids and matching attribute default selected user ids that will set the initial state of the component changing this property after the component has loaded will have no effect on the ui and changes in the selection state will have no effect on the property the selectedpeople property will continue to be used as the true state of the component ,1
+717,13064248027.0,IssuesEvent,2020-07-30 17:47:00,openstates/issues,https://api.github.com/repos/openstates/issues,closed,MD legislators need update,component:people-data,"Hello, we need Maryland state legislators to be updated to current representatives.
+
+Thank you for your help in this.",1.0,"MD legislators need update - Hello, we need Maryland state legislators to be updated to current representatives.
+
+Thank you for your help in this.",1,md legislators need update hello we need maryland state legislators to be updated to current representatives thank you for your help in this ,1
+628,11048555051.0,IssuesEvent,2019-12-09 21:24:55,planningcenter/developers,https://api.github.com/repos/planningcenter/developers,closed,400 errors when retrieving people (GET),People,"**Affected Applications**
+Which app(s) does this bug affect?
+
+PCO people API
+
+**Describe the bug**
+
+Since 11th Nov, we have had persistent failures for all requests to the endpoint
+https://api.planningcenteronline.com/people/v2/people?where[...
+
+This is affecting the Pushpay integration for all our clients, and is a blocker for us right now.
+
+Specifically these requests take a while to respond and then fail with a 400 response, prior to the 11th these were returning a 200 response (and was a frequently exercised endpoint).
+
+**To Reproduce**
+I believe you should have everything you need from logs, it will be all the failed non-200 requests from us.
+
+Note: these requests would have stopped a day or so ago, because we have stopped the synchronisation job that was creating these requests to reduce error volumes, but the issue was not resolved at that point.
+
+**Expected behavior**
+A clear and concise description of what you expected to happen.
+
+200 responses to people GET requests.
+
+**Screenshots**
+If applicable, add screenshots to help explain your problem.
+
+**Context (please complete the following information):**
+- Endpoint: https://api.planningcenteronline.com/people/v2/people
+- Language: .Net
+- Authentication: OAuth2
+
+**Additional context**
+Add any other context about the problem here.
+",1.0,"400 errors when retrieving people (GET) - **Affected Applications**
+Which app(s) does this bug affect?
+
+PCO people API
+
+**Describe the bug**
+
+Since 11th Nov, we have had persistent failures for all requests to the endpoint
+https://api.planningcenteronline.com/people/v2/people?where[...
+
+This is affecting the Pushpay integration for all our clients, and is a blocker for us right now.
+
+Specifically these requests take a while to respond and then fail with a 400 response, prior to the 11th these were returning a 200 response (and was a frequently exercised endpoint).
+
+**To Reproduce**
+I believe you should have everything you need from logs, it will be all the failed non-200 requests from us.
+
+Note: these requests would have stopped a day or so ago, because we have stopped the synchronisation job that was creating these requests to reduce error volumes, but the issue was not resolved at that point.
+
+**Expected behavior**
+A clear and concise description of what you expected to happen.
+
+200 responses to people GET requests.
+
+**Screenshots**
+If applicable, add screenshots to help explain your problem.
+
+**Context (please complete the following information):**
+- Endpoint: https://api.planningcenteronline.com/people/v2/people
+- Language: .Net
+- Authentication: OAuth2
+
+**Additional context**
+Add any other context about the problem here.
+",1, errors when retrieving people get affected applications which app s does this bug affect pco people api describe the bug since nov we have had persistent failures for all requests to the endpoint this is affecting the pushpay integration for all our clients and is a blocker for us right now specifically these requests take a while to respond and then fail with a response prior to the these were returning a response and was a frequently exercised endpoint to reproduce i believe you should have everything you need from logs it will be all the failed non requests from us note these requests would have stopped a day or so ago because we have stopped the synchronisation job that was creating these requests to reduce error volumes but the issue was not resolved at that point expected behavior a clear and concise description of what you expected to happen responses to people get requests screenshots if applicable add screenshots to help explain your problem context please complete the following information endpoint language net authentication additional context add any other context about the problem here ,1
+589,10437982884.0,IssuesEvent,2019-09-18 00:10:07,OfficeDev/office-ui-fabric-react,https://api.github.com/repos/OfficeDev/office-ui-fabric-react,closed,PeoplePicker - Korean Characters issue in IE11.,Area: Internationalization Component: PeoplePicker Status: In PR,"### Environment Information
+Windows 10, IE11.
+
+### Please provide a reproduction of the bug in a codepen:
+Since last fix in IE11, it works great in BasePicker.
+
+
+However, in people picker, due to loading list of components and autofilter, before Korean Characters composition is end, characters are being swallowed.
+
+
+Repro Steps:
+1. `npm start`, run `localhost:4322` with IE 11. ( Chrome is not repro. )
+2. Go to PeoplePicker, active input field, to show list of available contacts,
+3. Typing Korean Character as you can faster like below sample.
+A. Test Option 1. : “t” -> “h” -> “f” -> “t” -> “h” -> “f” -> “t” -> “h” -> “f”
+It should display “솔솔솔” but, sometimes display “솔솔ㅗ”
+B. Test Option 2 : “r” -> “h” -> “f” -> “r” -> “h” -> “f” -> “r” -> “h” -> “f”
+C. Test Option 3 : “r” -> “h” -> “c” -> “r” -> “h” -> “c” -> “r” -> “h” -> “c”
+
+#### Actual behavior:
+It should display “솔솔솔” but, sometimes display “솔솔ㅗ”
+
+#### Expected behavior:
+ It should display “솔솔솔”
+
+
+#### Priorities and help requested:
+
+Are you willing to submit a PR to fix? (Yes, No)
+
+Requested priority: (Blocking)
+
+Products/sites affected: Sharepoint and OneDrive
+
+",1.0,"PeoplePicker - Korean Characters issue in IE11. - ### Environment Information
+Windows 10, IE11.
+
+### Please provide a reproduction of the bug in a codepen:
+Since last fix in IE11, it works great in BasePicker.
+
+
+However, in people picker, due to loading list of components and autofilter, before Korean Characters composition is end, characters are being swallowed.
+
+
+Repro Steps:
+1. `npm start`, run `localhost:4322` with IE 11. ( Chrome is not repro. )
+2. Go to PeoplePicker, active input field, to show list of available contacts,
+3. Typing Korean Character as you can faster like below sample.
+A. Test Option 1. : “t” -> “h” -> “f” -> “t” -> “h” -> “f” -> “t” -> “h” -> “f”
+It should display “솔솔솔” but, sometimes display “솔솔ㅗ”
+B. Test Option 2 : “r” -> “h” -> “f” -> “r” -> “h” -> “f” -> “r” -> “h” -> “f”
+C. Test Option 3 : “r” -> “h” -> “c” -> “r” -> “h” -> “c” -> “r” -> “h” -> “c”
+
+#### Actual behavior:
+It should display “솔솔솔” but, sometimes display ��솔솔ㅗ”
+
+#### Expected behavior:
+ It should display “솔솔솔”
+
+
+#### Priorities and help requested:
+
+Are you willing to submit a PR to fix? (Yes, No)
+
+Requested priority: (Blocking)
+
+Products/sites affected: Sharepoint and OneDrive
+
+",1,peoplepicker korean characters issue in environment information windows please provide a reproduction of the bug in a codepen since last fix in it works great in basepicker however in people picker due to loading list of components and autofilter before korean characters composition is end characters are being swallowed repro steps npm start run localhost with ie chrome is not repro go to peoplepicker active input field to show list of available contacts typing korean character as you can faster like below sample a test option “t” “h” “f” “t” “h” “f” “t” “h” “f” it should display “솔솔솔” but sometimes display “솔솔ㅗ” b test option “r” “h” “f” “r” “h” “f” “r” “h” “f” c test option “r” “h” “c” “r” “h” “c” “r” “h” “c” actual behavior it should display “솔솔솔” but sometimes display “솔솔ㅗ” expected behavior it should display “솔솔솔” priorities and help requested are you willing to submit a pr to fix yes no requested priority blocking products sites affected sharepoint and onedrive ,1
+211,4366315315.0,IssuesEvent,2016-08-03 14:01:21,wordpress-mobile/WordPress-Android,https://api.github.com/repos/wordpress-mobile/WordPress-Android,closed,PeopleTable cache order is wrong,People Management [Type] Bug,"### Expected behavior
+When I go into the people page that I visited earlier, I expect the list to be ordered by display name (users) or default (follower, email follower & viewers)
+
+### Actual behavior
+When I visit the people page for the second time, I see the list start with display names starting with `C` and after the network request is completed people with display names starting with `A` and `B` just gets inserted.
+
+### Steps to reproduce the behavior
+* Go into the people page for a site with a lot of users (at least 21). You can change `PeopleUtils.FETCH_LIMIT` to something smaller if you don't have a site that has 21 users.
+* Browse to the end of list, so the network requests are finished and at least the second page is loaded
+* Go back to My Site
+* Go into the People page again and you should see the order is messed up for a second until the network request is finished and people gets inserted all over the place
+
+##### Tested on [device], Android [version]
+It's a DB issue, so it's across all devices and versions.",1.0,"PeopleTable cache order is wrong - ### Expected behavior
+When I go into the people page that I visited earlier, I expect the list to be ordered by display name (users) or default (follower, email follower & viewers)
+
+### Actual behavior
+When I visit the people page for the second time, I see the list start with display names starting with `C` and after the network request is completed people with display names starting with `A` and `B` just gets inserted.
+
+### Steps to reproduce the behavior
+* Go into the people page for a site with a lot of users (at least 21). You can change `PeopleUtils.FETCH_LIMIT` to something smaller if you don't have a site that has 21 users.
+* Browse to the end of list, so the network requests are finished and at least the second page is loaded
+* Go back to My Site
+* Go into the People page again and you should see the order is messed up for a second until the network request is finished and people gets inserted all over the place
+
+##### Tested on [device], Android [version]
+It's a DB issue, so it's across all devices and versions.",1,peopletable cache order is wrong expected behavior when i go into the people page that i visited earlier i expect the list to be ordered by display name users or default follower email follower viewers actual behavior when i visit the people page for the second time i see the list start with display names starting with c and after the network request is completed people with display names starting with a and b just gets inserted steps to reproduce the behavior go into the people page for a site with a lot of users at least you can change peopleutils fetch limit to something smaller if you don t have a site that has users browse to the end of list so the network requests are finished and at least the second page is loaded go back to my site go into the people page again and you should see the order is messed up for a second until the network request is finished and people gets inserted all over the place tested on android it s a db issue so it s across all devices and versions ,1
+24,2651928877.0,IssuesEvent,2015-03-16 14:45:30,fskreuz/Tree-Trail,https://api.github.com/repos/fskreuz/Tree-Trail,closed,Texts from the previous Badge.,QA people! Test these!,Previous information from the previous badge you added are not cleared once you add a new badge. You still have to refresh it to clear all the texts in the textbox in the add badge.,1.0,Texts from the previous Badge. - Previous information from the previous badge you added are not cleared once you add a new badge. You still have to refresh it to clear all the texts in the textbox in the add badge.,1,texts from the previous badge previous information from the previous badge you added are not cleared once you add a new badge you still have to refresh it to clear all the texts in the textbox in the add badge ,1
+1125,27264842438.0,IssuesEvent,2023-02-22 17:14:36,openstates/issues,https://api.github.com/repos/openstates/issues,closed,New HI Committee Scraper,good first issue component:people-data good first scraper,"### We need a new committee scraper written in [spatula](https://jamesturk.github.io/spatula/) for Hawaii.
+
+It should scrape this [webpage of House, Senate, and Special Committees](https://www.capitol.hawaii.gov/comminfolist.aspx) to get:
+- **name**
+ - “Small Business”, “Agriculture”, etc.
+- **chamber**
+ - “upper”, “lower”, or “legislature” (when joint, etc.)
+ - *Note: HI's Special Committees are typically chamber-specific because Hawaii rarely has joint committees.
+ - The Special Committee on State of Hawaiʻi Procurement (SCP) is Senate (`chamber=""upper""`).*
+- **classification**
+ - ex: ""committee"", ""subcommittee""
+- **parent**
+ - *only if it is a subcommittee* - scraper should have a way to determine this
+ - ex: parent=""Natural Resources"" when *sub*committee name=""Forestry""
+- **sources**
+ - each added using `add_source()` method on instance of `ScrapeCommittee` type object
+ - ex: home page for list of committees, specific page for that committee, etc.
+- **members**
+ - - each added using `add_member()`method on instance of `ScrapeCommittee` type object
+ - *Attributes:*
+ - **name**: “Jane Doe”, “John Smith”, etc.
+ - **role** (where applicable): “Chair”, “Ranking Member”, etc.
+
+#### [Here is helpful documentation](https://docs.openstates.org/contributing/writing-a-committee-scraper/) for writing a committee scraper
+
+### Useful scrapers for reference
+An [old non-spatula-based scraper](https://github.com/openstates/openstates-scrapers/blob/main/scrapers/hi/committees.py) that may be helpful in understanding how we have captured the data in the past.
+
+A [spatula-based scraper](https://github.com/openstates/openstates-scrapers/blob/main/scrapers_next/mo/committees.py) that also parses an HTML page for reference.
+- As in the other spatula scraper, you will need to set the session variable separately (rather than depend on the `__init__.py`).
+
+### Other useful resources
+You can reference the [object models for committee scrapers](https://github.com/openstates/openstates-core/blob/main/openstates/models/committees.py) in the Open States Core repo.
+
+Further documentation on running spatula scrapers in the command line can be found [here](https://docs.openstates.org/contributing/scrapers/#running-spatula-scrapers).",1.0,"New HI Committee Scraper - ### We need a new committee scraper written in [spatula](https://jamesturk.github.io/spatula/) for Hawaii.
+
+It should scrape this [webpage of House, Senate, and Special Committees](https://www.capitol.hawaii.gov/comminfolist.aspx) to get:
+- **name**
+ - “Small Business”, “Agriculture”, etc.
+- **chamber**
+ - “upper”, “lower”, or “legislature” (when joint, etc.)
+ - *Note: HI's Special Committees are typically chamber-specific because Hawaii rarely has joint committees.
+ - The Special Committee on State of Hawaiʻi Procurement (SCP) is Senate (`chamber=""upper""`).*
+- **classification**
+ - ex: ""committee"", ""subcommittee""
+- **parent**
+ - *only if it is a subcommittee* - scraper should have a way to determine this
+ - ex: parent=""Natural Resources"" when *sub*committee name=""Forestry""
+- **sources**
+ - each added using `add_source()` method on instance of `ScrapeCommittee` type object
+ - ex: home page for list of committees, specific page for that committee, etc.
+- **members**
+ - - each added using `add_member()`method on instance of `ScrapeCommittee` type object
+ - *Attributes:*
+ - **name**: “Jane Doe”, “John Smith”, etc.
+ - **role** (where applicable): “Chair”, “Ranking Member”, etc.
+
+#### [Here is helpful documentation](https://docs.openstates.org/contributing/writing-a-committee-scraper/) for writing a committee scraper
+
+### Useful scrapers for reference
+An [old non-spatula-based scraper](https://github.com/openstates/openstates-scrapers/blob/main/scrapers/hi/committees.py) that may be helpful in understanding how we have captured the data in the past.
+
+A [spatula-based scraper](https://github.com/openstates/openstates-scrapers/blob/main/scrapers_next/mo/committees.py) that also parses an HTML page for reference.
+- As in the other spatula scraper, you will need to set the session variable separately (rather than depend on the `__init__.py`).
+
+### Other useful resources
+You can reference the [object models for committee scrapers](https://github.com/openstates/openstates-core/blob/main/openstates/models/committees.py) in the Open States Core repo.
+
+Further documentation on running spatula scrapers in the command line can be found [here](https://docs.openstates.org/contributing/scrapers/#running-spatula-scrapers).",1,new hi committee scraper we need a new committee scraper written in for hawaii it should scrape this to get name “small business” “agriculture” etc chamber “upper” “lower” or “legislature” when joint etc note hi s special committees are typically chamber specific because hawaii rarely has joint committees the special committee on state of hawaiʻi procurement scp is senate chamber upper classification ex committee subcommittee parent only if it is a subcommittee scraper should have a way to determine this ex parent natural resources when sub committee name forestry sources each added using add source method on instance of scrapecommittee type object ex home page for list of committees specific page for that committee etc members each added using add member method on instance of scrapecommittee type object attributes name “jane doe” “john smith” etc role where applicable “chair” “ranking member” etc for writing a committee scraper useful scrapers for reference an that may be helpful in understanding how we have captured the data in the past a that also parses an html page for reference as in the other spatula scraper you will need to set the session variable separately rather than depend on the init py other useful resources you can reference the in the open states core repo further documentation on running spatula scrapers in the command line can be found ,1
+649447,21301377756.0,IssuesEvent,2022-04-15 03:57:05,wso2/kubernetes-apim,https://api.github.com/repos/wso2/kubernetes-apim,closed,Ingress API version extensions/v1beta1 not supported in K8s v1.23,duplicate Priority/High,"**Description:**
+$subject
+
+**Suggested Labels:**
+
+
+**Suggested Assignees:**
+
+
+**Affected Product Version:**
+K8s version 1.23
+
+**OS, DB, other environment details and versions:**
+
+**Steps to reproduce:**
+
+
+**Related Issues:**
+",1.0,"Ingress API version extensions/v1beta1 not supported in K8s v1.23 - **Description:**
+$subject
+
+**Suggested Labels:**
+
+
+**Suggested Assignees:**
+
+
+**Affected Product Version:**
+K8s version 1.23
+
+**OS, DB, other environment details and versions:**
+
+**Steps to reproduce:**
+
+
+**Related Issues:**
+",0,ingress api version extensions not supported in description subject suggested labels suggested assignees affected product version version os db other environment details and versions steps to reproduce related issues ,0
+44206,11406403572.0,IssuesEvent,2020-01-31 14:15:33,coin-or-tools/BuildTools,https://api.github.com/repos/coin-or-tools/BuildTools,closed,abs_*_dir variables do assumptions on configure options,bug build system,"Issue created by migration from Trac.
+
+Original creator: @svigerske
+
+Original creation time: 2013-04-22 18:19:10
+
+Assignee: @svigerske
+
+Version: 0.7
+
+The code
+```
+# Stuff for example Makefiles
+if test x$prefix = xNONE; then
+ full_prefix=$ac_default_prefix
+else
+ full_prefix=$prefix
+fi
+full_prefix=`cd $full_prefix ; pwd`
+AC_SUBST(abs_lib_dir)
+abs_lib_dir=$full_prefix/lib
+AC_SUBST(abs_include_dir)
+abs_include_dir=$full_prefix/include
+AC_SUBST(abs_bin_dir)
+abs_bin_dir=$full_prefix/bin
+```
+in the `AC_COIN_INIT_AUTOMAKE` macro in `coin.m4` assume that the user did not specify own values for `--exec-prefix`, `--libdir`, and so on.
+
+However, already some newer 64bit systems reset `$libdir` to `$exec_prefix/lib64`, see also #90.
+
+",1.0,"abs_*_dir variables do assumptions on configure options - Issue created by migration from Trac.
+
+Original creator: @svigerske
+
+Original creation time: 2013-04-22 18:19:10
+
+Assignee: @svigerske
+
+Version: 0.7
+
+The code
+```
+# Stuff for example Makefiles
+if test x$prefix = xNONE; then
+ full_prefix=$ac_default_prefix
+else
+ full_prefix=$prefix
+fi
+full_prefix=`cd $full_prefix ; pwd`
+AC_SUBST(abs_lib_dir)
+abs_lib_dir=$full_prefix/lib
+AC_SUBST(abs_include_dir)
+abs_include_dir=$full_prefix/include
+AC_SUBST(abs_bin_dir)
+abs_bin_dir=$full_prefix/bin
+```
+in the `AC_COIN_INIT_AUTOMAKE` macro in `coin.m4` assume that the user did not specify own values for `--exec-prefix`, `--libdir`, and so on.
+
+However, already some newer 64bit systems reset `$libdir` to `$exec_prefix/lib64`, see also #90.
+
+",0,abs dir variables do assumptions on configure options issue created by migration from trac original creator svigerske original creation time assignee svigerske version the code stuff for example makefiles if test x prefix xnone then full prefix ac default prefix else full prefix prefix fi full prefix cd full prefix pwd ac subst abs lib dir abs lib dir full prefix lib ac subst abs include dir abs include dir full prefix include ac subst abs bin dir abs bin dir full prefix bin in the ac coin init automake macro in coin assume that the user did not specify own values for exec prefix libdir and so on however already some newer systems reset libdir to exec prefix see also ,0
+14268,5619688792.0,IssuesEvent,2017-04-04 02:55:39,Polymer/polymer-cli,https://api.github.com/repos/Polymer/polymer-cli,closed,polymer build --bundled Breaks only one of my pages,Area: Build Priority: High Status: Available Type: Bug,"
+
+
+
+### Description
+
+Building with --bundled breaks only the messages page
+
+
+
+### Versions & Environment
+
+- Polymer CLI: 0.18.0-pre.13
+- node: v7.6.0
+- Operating System: Linux Mint 17 Qiana
+
+#### Steps to Reproduce
+1. Clone my project
+
+```
+git clone https://github.com/jifalops/app-starter.git -b build-tests
+cd app-starter
+bower install
+```
+
+2. Build and test
+
+```
+polymer build
+polymer serve build/bundled-es6
+```
+
+3. Navigate to the `messages` page (/messages). Many console errors appear (see above pic).
+Serving the unbuilt code or the build but unbundled code has no errors.
+
+
+
+
+",1.0,"polymer build --bundled Breaks only one of my pages -
+
+
+
+### Description
+
+Building with --bundled breaks only the messages page
+
+
+
+### Versions & Environment
+
+- Polymer CLI: 0.18.0-pre.13
+- node: v7.6.0
+- Operating System: Linux Mint 17 Qiana
+
+#### Steps to Reproduce
+1. Clone my project
+
+```
+git clone https://github.com/jifalops/app-starter.git -b build-tests
+cd app-starter
+bower install
+```
+
+2. Build and test
+
+```
+polymer build
+polymer serve build/bundled-es6
+```
+
+3. Navigate to the `messages` page (/messages). Many console errors appear (see above pic).
+Serving the unbuilt code or the build but unbundled code has no errors.
+
+
+
+
+",0,polymer build bundled breaks only one of my pages if you are asking a question rather than filing a bug you ll get better results using one of these instead stack overflow polymer slack channel mailing list description building with bundled breaks only the messages page versions environment polymer version will show the version for polymer cli node version will show the version for node polymer cli pre node operating system linux mint qiana steps to reproduce clone my project git clone b build tests cd app starter bower install build and test polymer build polymer serve build bundled navigate to the messages page messages many console errors appear see above pic serving the unbuilt code or the build but unbundled code has no errors example create an application project polymer init application add script tag to index html script src build polymer build ,0
+158469,24846520802.0,IssuesEvent,2022-10-26 16:17:24,ThreeSixtyGiving/grantnav,https://api.github.com/repos/ThreeSixtyGiving/grantnav,closed,As Publisher - Datasets table is missing information from ‘Funders Covered’,bug - not as designed Org pages,"**Describe the bug**
+A clear and concise description of what the bug is.
+
+On the As Publisher section, where it's available, the Datasets table column for Funders Covered is blank. It should contain a list of all the Funders covered in the dataset, as in live GrantNav.
+
+**To Reproduce**
+Steps to reproduce the behavior:
+1. Go to Funder search
+2. Select any funder with a Publisher section, e.g. Sport England
+3. Scroll to As Publisher section to view Datasets table
+
+**Expected behavior**
+A clear and concise description of what you expected to happen.
+
+The Datasets table Funders Covered column should contain a list of all the Funders covered in the dataset, as in live GrantNav.
+
+**Screenshots**
+If applicable, add screenshots to help explain your problem.
+
+Dev GrantNav Publisher section - Sport England:
+
+
+
+Live GrantNav Publisher page - Sport England:
+
+
+
+**Desktop (please complete the following information):**
+ - OS: [e.g. iOS]
+ - Browser [e.g. chrome, safari]
+ - Version [e.g. 22]
+
+**Smartphone (please complete the following information):**
+ - Device: [e.g. iPhone6]
+ - OS: [e.g. iOS8.1]
+ - Browser [e.g. stock browser, safari]
+ - Version [e.g. 22]
+
+**Additional context**
+Add any other context about the problem here.
+",1.0,"As Publisher - Datasets table is missing information from ‘Funders Covered’ - **Describe the bug**
+A clear and concise description of what the bug is.
+
+On the As Publisher section, where it's available, the Datasets table column for Funders Covered is blank. It should contain a list of all the Funders covered in the dataset, as in live GrantNav.
+
+**To Reproduce**
+Steps to reproduce the behavior:
+1. Go to Funder search
+2. Select any funder with a Publisher section, e.g. Sport England
+3. Scroll to As Publisher section to view Datasets table
+
+**Expected behavior**
+A clear and concise description of what you expected to happen.
+
+The Datasets table Funders Covered column should contain a list of all the Funders covered in the dataset, as in live GrantNav.
+
+**Screenshots**
+If applicable, add screenshots to help explain your problem.
+
+Dev GrantNav Publisher section - Sport England:
+
+
+
+Live GrantNav Publisher page - Sport England:
+
+
+
+**Desktop (please complete the following information):**
+ - OS: [e.g. iOS]
+ - Browser [e.g. chrome, safari]
+ - Version [e.g. 22]
+
+**Smartphone (please complete the following information):**
+ - Device: [e.g. iPhone6]
+ - OS: [e.g. iOS8.1]
+ - Browser [e.g. stock browser, safari]
+ - Version [e.g. 22]
+
+**Additional context**
+Add any other context about the problem here.
+",0,as publisher datasets table is missing information from ‘funders covered’ describe the bug a clear and concise description of what the bug is on the as publisher section where it s available the datasets table column for funders covered is blank it should contain a list of all the funders covered in the dataset as in live grantnav to reproduce steps to reproduce the behavior go to funder search select any funder with a publisher section e g sport england scroll to as publisher section to view datasets table expected behavior a clear and concise description of what you expected to happen the datasets table funders covered column should contain a list of all the funders covered in the dataset as in live grantnav screenshots if applicable add screenshots to help explain your problem dev grantnav publisher section sport england live grantnav publisher page sport england desktop please complete the following information os browser version smartphone please complete the following information device os browser version additional context add any other context about the problem here ,0
+267331,23291819617.0,IssuesEvent,2022-08-06 01:11:02,backend-br/vagas,https://api.github.com/repos/backend-br/vagas,closed,[REMOTO-PJ] Back-End Developer @DQRTECH,PJ Pleno Remoto AWS Testes Unitários RabbitMQ Stale,"
+
+
+> Vaga 100% Remota
+
+## Nossa empresa
+
+Nascemos de uma parceria de mais de 10 anos entre seus sócios, somando a paixão por tecnologia à vontade de construir uma empresa diferente, investindo na proximidade com nossos parceiros e clientes. Somos uma consultoria especializada em Arquitetura de TI, Outsourcing de Profissionais e Hunting.
+
+## Descrição da vaga
+
+A DQR TECH está em busca de um profissional para atuar em Squad ágil no setor financeiro.
+
+## Local
+
+100% remoto
+
+## Requisitos
+
+**Obrigatórios:**
+Experiência em projetos de desenvolvimento de software utilizando .NET Core;
+Desenvolvimento de aplicações em nuvem AWS, utilizando os serviços: Lambda, SQS, SNS; DynamoDb, S3, EC2, RDS;
+Experiência em serviços de mensageria (RabbitMQ/Kafka ou semelhante);
+Design Patterns GoF;
+
+**Desejáveis:**
+Experiência em modelagem de dados em DDD;
+Conhecimento em Arquitetura de Microsserviços;
+Execução de testes unitários;
+
+**Diferencial:**
+- Ter atuado em instituições financeiras
+
+## Contratação
+160 horas mês - PJ 15k
+
+## Como se candidatar
+Por favor envie um e-mail para vagas@dqrtech.com.br com seu CV anexado
+Enviar no assunto: Vaga Backend .Net - PJ
+
+## Tempo médio de feedbacks
+Costumamos enviar feedbacks em até 01 dia após cada processo.
+
+#### Alocação
+- Remoto
+
+#### Regime
+- PJ
+
+#### Nível
+- Pleno
+- Sênior
+
+
+",1.0,"[REMOTO-PJ] Back-End Developer @DQRTECH -
+
+
+> Vaga 100% Remota
+
+## Nossa empresa
+
+Nascemos de uma parceria de mais de 10 anos entre seus sócios, somando a paixão por tecnologia à vontade de construir uma empresa diferente, investindo na proximidade com nossos parceiros e clientes. Somos uma consultoria especializada em Arquitetura de TI, Outsourcing de Profissionais e Hunting.
+
+## Descrição da vaga
+
+A DQR TECH está em busca de um profissional para atuar em Squad ágil no setor financeiro.
+
+## Local
+
+100% remoto
+
+## Requisitos
+
+**Obrigatórios:**
+Experiência em projetos de desenvolvimento de software utilizando .NET Core;
+Desenvolvimento de aplicações em nuvem AWS, utilizando os serviços: Lambda, SQS, SNS; DynamoDb, S3, EC2, RDS;
+Experiência em serviços de mensageria (RabbitMQ/Kafka ou semelhante);
+Design Patterns GoF;
+
+**Desejáveis:**
+Experiência em modelagem de dados em DDD;
+Conhecimento em Arquitetura de Microsserviços;
+Execução de testes unitários;
+
+**Diferencial:**
+- Ter atuado em instituições financeiras
+
+## Contratação
+160 horas mês - PJ 15k
+
+## Como se candidatar
+Por favor envie um e-mail para vagas@dqrtech.com.br com seu CV anexado
+Enviar no assunto: Vaga Backend .Net - PJ
+
+## Tempo médio de feedbacks
+Costumamos enviar feedbacks em até 01 dia após cada processo.
+
+#### Alocação
+- Remoto
+
+#### Regime
+- PJ
+
+#### Nível
+- Pleno
+- Sênior
+
+
+",0, back end developer dqrtech caso a vaga for remoto durante a pandemia informar no texto remoto durante o covid por favor só poste se a vaga for para back end não faça distinção de gênero no título da vaga use back end developer ao invés de desenvolvedor back end o exemplo back end developer nome da empresa caso a vaga for remoto durante a pandemia deixar a linha abaixo vaga remota nossa empresa nascemos de uma parceria de mais de anos entre seus sócios somando a paixão por tecnologia à vontade de construir uma empresa diferente investindo na proximidade com nossos parceiros e clientes somos uma consultoria especializada em arquitetura de ti outsourcing de profissionais e hunting descrição da vaga a dqr tech está em busca de um profissional para atuar em squad ágil no setor financeiro local remoto requisitos obrigatórios experiência em projetos de desenvolvimento de software utilizando net core desenvolvimento de aplicações em nuvem aws utilizando os serviços lambda sqs sns dynamodb rds experiência em serviços de mensageria rabbitmq kafka ou semelhante design patterns gof desejáveis experiência em modelagem de dados em ddd conhecimento em arquitetura de microsserviços execução de testes unitários diferencial ter atuado em instituições financeiras contratação horas mês pj como se candidatar por favor envie um e mail para vagas dqrtech com br com seu cv anexado enviar no assunto vaga backend net pj tempo médio de feedbacks costumamos enviar feedbacks em até dia após cada processo alocação remoto regime pj nível pleno sênior ,0
+5977,2610218874.0,IssuesEvent,2015-02-26 19:09:36,chrsmith/somefinders,https://api.github.com/repos/chrsmith/somefinders,opened,бруклин бонс прогресив атак,auto-migrated Priority-Medium Type-Defect,"```
+'''Аполлон Дроздов'''
+День добрый никак не могу найти .бруклин
+бонс прогресив атак. как то выкладывали уже
+
+'''Вацлав Колесников'''
+Качай тут http://bit.ly/17CiD4b
+
+'''Геодар Орехов'''
+Спасибо вроде то но просит телефон вводить
+
+'''Валерий Иванов'''
+Не это не влияет на баланс
+
+'''Аким Никитин'''
+Неа все ок у меня ничего не списало
+
+
+
+
+
+
+Информация о файле: бруклин бонс прогресив
+атак
+Загружен: В этом месяце
+Скачан раз: 135
+Рейтинг: 616
+Средняя скорость скачивания: 1370
+Похожих файлов: 40
+
+```
+
+-----
+Original issue reported on code.google.com by `kondense...@gmail.com` on 17 Dec 2013 at 5:10",1.0,"бруклин бонс прогресив атак - ```
+'''Аполлон Дроздов'''
+День добрый никак не могу найти .бруклин
+бонс прогресив атак. как то выкладывали уже
+
+'''Вацлав Колесников'''
+Качай тут http://bit.ly/17CiD4b
+
+'''Геодар Орехов'''
+Спасибо вроде то но просит телефон вводить
+
+'''Валерий Иванов'''
+Не это не влияет на баланс
+
+'''Аким Никитин'''
+Неа все ок у меня ничего не списало
+
+
+
+
+
+
+Информация о файле: бруклин бонс прогресив
+атак
+Загружен: В этом месяце
+Скачан раз: 135
+Рейтинг: 616
+Средняя скорость скачивания: 1370
+Похожих файлов: 40
+
+```
+
+-----
+Original issue reported on code.google.com by `kondense...@gmail.com` on 17 Dec 2013 at 5:10",0,бруклин бонс прогресив атак аполлон дроздов день добрый никак не могу найти бруклин бонс прогресив атак как то выкладывали уже вацлав колесников качай тут геодар орехов спасибо вроде то но просит телефон вводить валерий иванов не это не влияет на баланс аким никитин неа все ок у меня ничего не списало информация о файле бруклин бонс прогресив атак загружен в этом месяце скачан раз рейтинг средняя скорость скачивания похожих файлов original issue reported on code google com by kondense gmail com on dec at ,0
+382030,11299831631.0,IssuesEvent,2020-01-17 12:11:24,StrangeLoopGames/EcoIssues,https://api.github.com/repos/StrangeLoopGames/EcoIssues,closed,USER ISSUE: can't go under overhangs while diving,Priority: Medium,"When Diving (holding CTRL underwater to stay on the bottom), any overhang, at any height above you, will produce an invisible wall that stops you from going under the overhang. (tested up to height about 10, not certain about overhangs above the water's surface, or overhangs made by constructed blocks) Releasing CTRL while pushing into the invisible wall will let you enter the space as long as it would be walkable above ground (so an opening at least 2 tiles high). Once you are under the overhang, you can then hold down CTRL again to stay on the bottom, or just leave it off and let the ceiling keep you where you want to be. I noticed this while mining underwater iron ore in 0.7.3.3, and later tested with stone & soil in 0.7.4.0.
+
+Expected behavior: Diving should not affect what areas can be entered except insofar as it changes what vertical level you are at underwater
+
+Steps to reproduce:
+- jump in a river
+- hold CTRL to go to the bottom, and keep holding it to stay there
+- find a sheer vertical surface at least 3 tiles high
+- mine out a 2x1 opening at the bottom of the face
+- try to enter the mined-out area, and fail",1.0,"USER ISSUE: can't go under overhangs while diving - When Diving (holding CTRL underwater to stay on the bottom), any overhang, at any height above you, will produce an invisible wall that stops you from going under the overhang. (tested up to height about 10, not certain about overhangs above the water's surface, or overhangs made by constructed blocks) Releasing CTRL while pushing into the invisible wall will let you enter the space as long as it would be walkable above ground (so an opening at least 2 tiles high). Once you are under the overhang, you can then hold down CTRL again to stay on the bottom, or just leave it off and let the ceiling keep you where you want to be. I noticed this while mining underwater iron ore in 0.7.3.3, and later tested with stone & soil in 0.7.4.0.
+
+Expected behavior: Diving should not affect what areas can be entered except insofar as it changes what vertical level you are at underwater
+
+Steps to reproduce:
+- jump in a river
+- hold CTRL to go to the bottom, and keep holding it to stay there
+- find a sheer vertical surface at least 3 tiles high
+- mine out a 2x1 opening at the bottom of the face
+- try to enter the mined-out area, and fail",0,user issue can t go under overhangs while diving when diving holding ctrl underwater to stay on the bottom any overhang at any height above you will produce an invisible wall that stops you from going under the overhang tested up to height about not certain about overhangs above the water s surface or overhangs made by constructed blocks releasing ctrl while pushing into the invisible wall will let you enter the space as long as it would be walkable above ground so an opening at least tiles high once you are under the overhang you can then hold down ctrl again to stay on the bottom or just leave it off and let the ceiling keep you where you want to be i noticed this while mining underwater iron ore in and later tested with stone soil in expected behavior diving should not affect what areas can be entered except insofar as it changes what vertical level you are at underwater steps to reproduce jump in a river hold ctrl to go to the bottom and keep holding it to stay there find a sheer vertical surface at least tiles high mine out a opening at the bottom of the face try to enter the mined out area and fail,0
+371,6891710214.0,IssuesEvent,2017-11-22 18:01:19,PopulateTools/gobierto,https://api.github.com/repos/PopulateTools/gobierto,closed,Don't show blocks/links to sections if there is no content for it,gobierto-people,"- [ ] Filters for Oposition, Directives, Groups - in Home, /personas, /agendas
+- [ ] Blog option in local menu
+- [ ] Blog block in home page
+- [ ] Biography block in /personas/person
+
+
+",1.0,"Don't show blocks/links to sections if there is no content for it - - [ ] Filters for Oposition, Directives, Groups - in Home, /personas, /agendas
+- [ ] Blog option in local menu
+- [ ] Blog block in home page
+- [ ] Biography block in /personas/person
+
+
+",1,don t show blocks links to sections if there is no content for it filters for oposition directives groups in home personas agendas blog option in local menu blog block in home page biography block in personas person ,1
+41,2658170961.0,IssuesEvent,2015-03-18 14:17:46,fskreuz/Tree-Trail,https://api.github.com/repos/fskreuz/Tree-Trail,closed,Add Close functionality to sidebar,QA people! Test these!,"As a guest user, when I click an approved badge and a sidebar appears, I should be able to have an option to close the sidebar para dili samok ig navigate sa map.",1.0,"Add Close functionality to sidebar - As a guest user, when I click an approved badge and a sidebar appears, I should be able to have an option to close the sidebar para dili samok ig navigate sa map.",1,add close functionality to sidebar as a guest user when i click an approved badge and a sidebar appears i should be able to have an option to close the sidebar para dili samok ig navigate sa map ,1
+181,3966838930.0,IssuesEvent,2016-05-03 14:23:46,wordpress-mobile/WordPress-Android,https://api.github.com/repos/wordpress-mobile/WordPress-Android,opened,Use roles from the network request,People Management,"For updating the users in people management, we should use the `/v1.1/sites/$site/roles` to get a possible list of roles instead of using an array resource to be able to support Jetpack sites.",1.0,"Use roles from the network request - For updating the users in people management, we should use the `/v1.1/sites/$site/roles` to get a possible list of roles instead of using an array resource to be able to support Jetpack sites.",1,use roles from the network request for updating the users in people management we should use the sites site roles to get a possible list of roles instead of using an array resource to be able to support jetpack sites ,1
+37,2652594655.0,IssuesEvent,2015-03-16 18:10:03,fskreuz/Tree-Trail,https://api.github.com/repos/fskreuz/Tree-Trail,closed,Update Notification Modal on Admin,QA people! Test these!,"Every an admin updates information from their profile, for example if I want to update my contact number, it pops up a modal stating that the Name has been successfully updated, not the contact number.",1.0,"Update Notification Modal on Admin - Every an admin updates information from their profile, for example if I want to update my contact number, it pops up a modal stating that the Name has been successfully updated, not the contact number.",1,update notification modal on admin every an admin updates information from their profile for example if i want to update my contact number it pops up a modal stating that the name has been successfully updated not the contact number ,1
+561,9994574883.0,IssuesEvent,2019-07-11 18:01:15,mubix/restincode,https://api.github.com/repos/mubix/restincode,opened,Nurse,People,"I don't remember his twitter account but he was an amazing person and friend to all, I could use some help finding links for things about him again.",1.0,"Nurse - I don't remember his twitter account but he was an amazing person and friend to all, I could use some help finding links for things about him again.",1,nurse i don t remember his twitter account but he was an amazing person and friend to all i could use some help finding links for things about him again ,1
+235014,25901420758.0,IssuesEvent,2022-12-15 06:12:43,Trinadh465/linux-3.0.35_CVE-2019-10220,https://api.github.com/repos/Trinadh465/linux-3.0.35_CVE-2019-10220,opened,CVE-2016-3156 (Medium) detected in linuxlinux-3.0.40,security vulnerability,"## CVE-2016-3156 - Medium Severity Vulnerability
+ Vulnerable Library - linuxlinux-3.0.40
+
+
+The IPv4 implementation in the Linux kernel before 4.5.2 mishandles destruction of device objects, which allows guest OS users to cause a denial of service (host OS networking outage) by arranging for a large number of IP addresses.
+
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",True,"CVE-2016-3156 (Medium) detected in linuxlinux-3.0.40 - ## CVE-2016-3156 - Medium Severity Vulnerability
+ Vulnerable Library - linuxlinux-3.0.40
+
+
+The IPv4 implementation in the Linux kernel before 4.5.2 mishandles destruction of device objects, which allows guest OS users to cause a denial of service (host OS networking outage) by arranging for a large number of IP addresses.
+
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",0,cve medium detected in linuxlinux cve medium severity vulnerability vulnerable library linuxlinux apache software foundation asf library home page a href found in head commit a href found in base branch master vulnerable source files net fib frontend c vulnerability details the implementation in the linux kernel before mishandles destruction of device objects which allows guest os users to cause a denial of service host os networking outage by arranging for a large number of ip addresses publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend ,0
+110629,11706773962.0,IssuesEvent,2020-03-08 00:41:35,pyinstaller/pyinstaller,https://api.github.com/repos/pyinstaller/pyinstaller,closed,generate API doc from PyInstaller code,@low / cleanup area:documentation help wanted,"I think it might be helpful to implement generating PyInstaller API doc from the docstrings in the PyInstaller code.
+",1.0,"generate API doc from PyInstaller code - I think it might be helpful to implement generating PyInstaller API doc from the docstrings in the PyInstaller code.
+",0,generate api doc from pyinstaller code i think it might be helpful to implement generating pyinstaller api doc from the docstrings in the pyinstaller code ,0
+880,17273971684.0,IssuesEvent,2021-07-23 01:41:58,hackforla/civic-opportunity,https://api.github.com/repos/hackforla/civic-opportunity,closed,Establish HfLA Affinity Groups,initiative: affinity groups people: 1 role: ux research,"### Dependency
+Jayashree's return
+
+
+### Overview
+Being able to connect with peer groups through shared life experiences can be instrumental to professional confidence, growth and mentorship. HfLA should provide spaces for volunteers to connect with like-minded peers.
+
+### Action Items
+
+- [ ] Create a Google Form Questionnaire to gauge participation interest Affinity Groups
+- [ ] Review questions with HfLA executive leadership
+- [ ] Finalize Form and send out to the network
+- [ ] Create necessary Slack Groups and publicize (We probably don't need to create Communities of Practice for these, but rather just them exist as Slack support channels?)
+
+Examples of potential affinity groups are: Veterans in Tech, Women in Tech, Parents in Tech, LGBTQ in Tech, Latinos in Tech, Black in Tech, etc.
+
+### Resources/Instructions
+[Draft Google Form Questions](https://docs.google.com/document/d/1fUcVXsA7Yw7gs79hMrPuMgL1LIqLoBgAxm7_HixP1Xg/edit)
+[Draft Google Form](https://docs.google.com/forms/d/1aJz9sijWCIVQnz2dbuO5p18WTQltAJTAJqmwUSjecMg/edit)",1.0,"Establish HfLA Affinity Groups - ### Dependency
+Jayashree's return
+
+
+### Overview
+Being able to connect with peer groups through shared life experiences can be instrumental to professional confidence, growth and mentorship. HfLA should provide spaces for volunteers to connect with like-minded peers.
+
+### Action Items
+
+- [ ] Create a Google Form Questionnaire to gauge participation interest Affinity Groups
+- [ ] Review questions with HfLA executive leadership
+- [ ] Finalize Form and send out to the network
+- [ ] Create necessary Slack Groups and publicize (We probably don't need to create Communities of Practice for these, but rather just them exist as Slack support channels?)
+
+Examples of potential affinity groups are: Veterans in Tech, Women in Tech, Parents in Tech, LGBTQ in Tech, Latinos in Tech, Black in Tech, etc.
+
+### Resources/Instructions
+[Draft Google Form Questions](https://docs.google.com/document/d/1fUcVXsA7Yw7gs79hMrPuMgL1LIqLoBgAxm7_HixP1Xg/edit)
+[Draft Google Form](https://docs.google.com/forms/d/1aJz9sijWCIVQnz2dbuO5p18WTQltAJTAJqmwUSjecMg/edit)",1,establish hfla affinity groups dependency jayashree s return overview being able to connect with peer groups through shared life experiences can be instrumental to professional confidence growth and mentorship hfla should provide spaces for volunteers to connect with like minded peers action items create a google form questionnaire to gauge participation interest affinity groups review questions with hfla executive leadership finalize form and send out to the network create necessary slack groups and publicize we probably don t need to create communities of practice for these but rather just them exist as slack support channels examples of potential affinity groups are veterans in tech women in tech parents in tech lgbtq in tech latinos in tech black in tech etc resources instructions ,1
+638012,20693424562.0,IssuesEvent,2022-03-11 04:31:37,AY2122S2-CS2103T-W14-1/tp,https://api.github.com/repos/AY2122S2-CS2103T-W14-1/tp,closed,"As a librarian, I want to be able to add a patron (student or staff) to my database",priority.High type.Story,...so that I can keep track of books borrowed or requested by the patron .,1.0,"As a librarian, I want to be able to add a patron (student or staff) to my database - ...so that I can keep track of books borrowed or requested by the patron .",0,as a librarian i want to be able to add a patron student or staff to my database so that i can keep track of books borrowed or requested by the patron ,0
+15035,3918638044.0,IssuesEvent,2016-04-21 13:17:26,requestly/blog,https://api.github.com/repos/requestly/blog,closed,Documentation: Replace Rule,Documentation,"Following Points to cover
+- Should be used to replace a string in Url with another string.
+- Use Cases
+- Regex Matching
+
+## Use Cases
+- Dropbox Use Case ?dl=0 to ?dl=1
+- Domain Switching (Paste Image of feedback on Chrome Store from Adam B) and how to work with Rest APIs. Replace domain with local servers So no need to change UI code with fake domain.
+- Fix broken/changed Urls
+- Language issue in documentation sites
+
+## Chrome Store Feedbacks:
+
+Adam B
+> Perfect. I use a replace host rule to redirect youtube.com/watch?v= links to gaming.youtube.com/watch?v=
+
+## Regex Matching
+- JS Regex expression to match a certain part of URL and use $s to use the values extracted by group expressions.
+- Example1: Extract query String from Url
+- Reference regex101",1.0,"Documentation: Replace Rule - Following Points to cover
+- Should be used to replace a string in Url with another string.
+- Use Cases
+- Regex Matching
+
+## Use Cases
+- Dropbox Use Case ?dl=0 to ?dl=1
+- Domain Switching (Paste Image of feedback on Chrome Store from Adam B) and how to work with Rest APIs. Replace domain with local servers So no need to change UI code with fake domain.
+- Fix broken/changed Urls
+- Language issue in documentation sites
+
+## Chrome Store Feedbacks:
+
+Adam B
+> Perfect. I use a replace host rule to redirect youtube.com/watch?v= links to gaming.youtube.com/watch?v=
+
+## Regex Matching
+- JS Regex expression to match a certain part of URL and use $s to use the values extracted by group expressions.
+- Example1: Extract query String from Url
+- Reference regex101",0,documentation replace rule following points to cover should be used to replace a string in url with another string use cases regex matching use cases dropbox use case dl to dl domain switching paste image of feedback on chrome store from adam b and how to work with rest apis replace domain with local servers so no need to change ui code with fake domain fix broken changed urls language issue in documentation sites chrome store feedbacks adam b perfect i use a replace host rule to redirect youtube com watch v links to gaming youtube com watch v regex matching js regex expression to match a certain part of url and use s to use the values extracted by group expressions extract query string from url reference ,0
+538,9594869513.0,IssuesEvent,2019-05-09 14:51:42,planningcenter/developers,https://api.github.com/repos/planningcenter/developers,closed,Can no longer send name_suffix with Person POST,People,"**Affected Applications**
+People API
+
+**Describe the bug**
+Suddenly cannot assign a name_suffix while creating person record via the API with included array as worked 6 months ago.
+
+*PLEASE DO NOT INCLUDE YOUR CLIENT ID, APPLICATION ID, OR SECRET IN THIS ISSUE. WE WILL REVOKE YOUR TOKEN IF YOU DO.*
+
+**To Reproduce**
+Steps to reproduce the behavior:
+1. attempt to create a new person WITH a name_suffix using the `pco_api gem
+
+2.)
+```ruby
+PCI.people.v2.people.post(
+ data: {
+ type: ""Person"",
+ attributes: {
+ first_name: ""Test"",
+ last_name: ""User"",
+ gender: ""male"",
+ grade: nil,
+ birthdate: ""1986-08-02"",
+ child: false
+ },
+ included: [
+ {
+ type: ""NameSuffix"",
+ attributes: {
+ value: ""Jr.""
+ }
+ }
+ ]
+ }
+)
+```
+3. Get an error back form API saying that included cannot be at root level of object
+
+**Expected behavior**
+A suffix to be added to the user as it's created. This worked for me in July 2018
+
+**Screenshots**
+If applicable, add screenshots to help explain your problem.
+
+**Context (please complete the following information):**
+- Endpoint: v2/people via ruby gem
+- Language: ruby
+- Authentication: Personal Access Token
+
+**Additional context**
+This was for a small integration I wrote six months ago and tested for about a month. We came back to it to modify and improve a few things and suddenly the name_suffix functionality had stopped working.
+",1.0,"Can no longer send name_suffix with Person POST - **Affected Applications**
+People API
+
+**Describe the bug**
+Suddenly cannot assign a name_suffix while creating person record via the API with included array as worked 6 months ago.
+
+*PLEASE DO NOT INCLUDE YOUR CLIENT ID, APPLICATION ID, OR SECRET IN THIS ISSUE. WE WILL REVOKE YOUR TOKEN IF YOU DO.*
+
+**To Reproduce**
+Steps to reproduce the behavior:
+1. attempt to create a new person WITH a name_suffix using the `pco_api gem
+
+2.)
+```ruby
+PCI.people.v2.people.post(
+ data: {
+ type: ""Person"",
+ attributes: {
+ first_name: ""Test"",
+ last_name: ""User"",
+ gender: ""male"",
+ grade: nil,
+ birthdate: ""1986-08-02"",
+ child: false
+ },
+ included: [
+ {
+ type: ""NameSuffix"",
+ attributes: {
+ value: ""Jr.""
+ }
+ }
+ ]
+ }
+)
+```
+3. Get an error back form API saying that included cannot be at root level of object
+
+**Expected behavior**
+A suffix to be added to the user as it's created. This worked for me in July 2018
+
+**Screenshots**
+If applicable, add screenshots to help explain your problem.
+
+**Context (please complete the following information):**
+- Endpoint: v2/people via ruby gem
+- Language: ruby
+- Authentication: Personal Access Token
+
+**Additional context**
+This was for a small integration I wrote six months ago and tested for about a month. We came back to it to modify and improve a few things and suddenly the name_suffix functionality had stopped working.
+",1,can no longer send name suffix with person post affected applications people api describe the bug suddenly cannot assign a name suffix while creating person record via the api with included array as worked months ago please do not include your client id application id or secret in this issue we will revoke your token if you do to reproduce steps to reproduce the behavior attempt to create a new person with a name suffix using the pco api gem ruby pci people people post data type person attributes first name test last name user gender male grade nil birthdate child false included type namesuffix attributes value jr get an error back form api saying that included cannot be at root level of object expected behavior a suffix to be added to the user as it s created this worked for me in july screenshots if applicable add screenshots to help explain your problem context please complete the following information endpoint people via ruby gem language ruby authentication personal access token additional context this was for a small integration i wrote six months ago and tested for about a month we came back to it to modify and improve a few things and suddenly the name suffix functionality had stopped working ,1
+265,5292550871.0,IssuesEvent,2017-02-09 02:41:05,henrikfroehling/TraktApiSharp,https://api.github.com/repos/henrikfroehling/TraktApiSharp,closed,Remove Images option in TraktExtendedInfo,library optimization trakt-api-episodes trakt-api-movies trakt-api-people trakt-api-seasons trakt-api-shows trakt-api-users v1.0.0 Alpha 2,"As pointed out in issue #32, images are no longer returned by the Trakt API.
+
+Therefore, remove the `Images` option in `TraktExtendedInfo` since it's no longer needed.
+User images (`TraktUser.Images`) are returned, if `TraktExtendedInfo.Full` is enabled.
+
+Also remove all `TraktImages` properties in `TraktShow`, `TraktMovie`, `TraktSeason`, `TraktEpisode` and `TraktPerson`, except in `TraktUser`.",1.0,"Remove Images option in TraktExtendedInfo - As pointed out in issue #32, images are no longer returned by the Trakt API.
+
+Therefore, remove the `Images` option in `TraktExtendedInfo` since it's no longer needed.
+User images (`TraktUser.Images`) are returned, if `TraktExtendedInfo.Full` is enabled.
+
+Also remove all `TraktImages` properties in `TraktShow`, `TraktMovie`, `TraktSeason`, `TraktEpisode` and `TraktPerson`, except in `TraktUser`.",1,remove images option in traktextendedinfo as pointed out in issue images are no longer returned by the trakt api therefore remove the images option in traktextendedinfo since it s no longer needed user images traktuser images are returned if traktextendedinfo full is enabled also remove all traktimages properties in traktshow traktmovie traktseason traktepisode and traktperson except in traktuser ,1
+46159,13055860282.0,IssuesEvent,2020-07-30 02:57:06,icecube-trac/tix2,https://api.github.com/repos/icecube-trac/tix2,opened,"icetray.inspect clashes with ""the real"" python inspect (Trac #660)",IceTray Incomplete Migration Migrated from Trac defect,"Migrated from https://code.icecube.wisc.edu/ticket/660
+```json
+{
+ ""status"": ""closed"",
+ ""changetime"": ""2011-12-15T16:53:37"",
+ ""description"": ""Currently (icerec trunk based latest offline-rc V11-12-00), importing python modules failes, e.g.:\n\n{{{\nIn [1]: from icecube import rootwriter\n---------------------------------------------------------------------------\nAttributeError Traceback (most recent call last)\n\n/home/fabian/Physik/software/icecube/icerec/trunk/src/ in ()\n\n/home/fabian/Physik/software/icecube/icerec/trunk/build-release/lib/icecube/rootwriter/__init__.py in ()\n 6 @icetray.traysegment_inherit(tableio.I3TableWriter,\n 7 removeopts=('TableService',))\n----> 8 def I3ROOTWriter(tray, name, Output=None, **kwargs):\n 9 \""\""\""Tabulate data to a ROOT file.\n 10 \n\n/home/fabian/Physik/software/icecube/icerec/trunk/build-release/lib/icecube/icetray/traysegment.pyc in traysegment_(function)\n 39 \n 40 def traysegment_(function):\n---> 41 func = traysegment(function)\n 42 func.module = parent\n 43 if defaultoverrides != None:\n\n/home/fabian/Physik/software/icecube/icerec/trunk/build-release/lib/icecube/icetray/traysegment.pyc in traysegment(function)\n 18 \""\""\""\n 19 \n---> 20 if inspect.getdoc(function) is None:\n 21 function.__doc__ = \""I3Tray segments should have docstrings. This one doesn't. Fix it.\""\n 22 \n\nAttributeError: 'module' object has no attribute 'getdoc'\n}}}\n\nThe reason is a namespace clash inside icetray.traysegment. I suggest renaming icetray.i3inspect. Alternatively, one could change icetray.traysegment to work without inspect, but this would only be a workaround and we'd likely be hit by this again."",
+ ""reporter"": ""kislat"",
+ ""cc"": """",
+ ""resolution"": ""invalid"",
+ ""_ts"": ""1323968017000000"",
+ ""component"": ""IceTray"",
+ ""summary"": ""icetray.inspect clashes with \""the real\"" python inspect"",
+ ""priority"": ""normal"",
+ ""keywords"": """",
+ ""time"": ""2011-12-12T16:27:53"",
+ ""milestone"": """",
+ ""owner"": """",
+ ""type"": ""defect""
+}
+```
+",1.0,"icetray.inspect clashes with ""the real"" python inspect (Trac #660) - Migrated from https://code.icecube.wisc.edu/ticket/660
+```json
+{
+ ""status"": ""closed"",
+ ""changetime"": ""2011-12-15T16:53:37"",
+ ""description"": ""Currently (icerec trunk based latest offline-rc V11-12-00), importing python modules failes, e.g.:\n\n{{{\nIn [1]: from icecube import rootwriter\n---------------------------------------------------------------------------\nAttributeError Traceback (most recent call last)\n\n/home/fabian/Physik/software/icecube/icerec/trunk/src/ in ()\n\n/home/fabian/Physik/software/icecube/icerec/trunk/build-release/lib/icecube/rootwriter/__init__.py in ()\n 6 @icetray.traysegment_inherit(tableio.I3TableWriter,\n 7 removeopts=('TableService',))\n----> 8 def I3ROOTWriter(tray, name, Output=None, **kwargs):\n 9 \""\""\""Tabulate data to a ROOT file.\n 10 \n\n/home/fabian/Physik/software/icecube/icerec/trunk/build-release/lib/icecube/icetray/traysegment.pyc in traysegment_(function)\n 39 \n 40 def traysegment_(function):\n---> 41 func = traysegment(function)\n 42 func.module = parent\n 43 if defaultoverrides != None:\n\n/home/fabian/Physik/software/icecube/icerec/trunk/build-release/lib/icecube/icetray/traysegment.pyc in traysegment(function)\n 18 \""\""\""\n 19 \n---> 20 if inspect.getdoc(function) is None:\n 21 function.__doc__ = \""I3Tray segments should have docstrings. This one doesn't. Fix it.\""\n 22 \n\nAttributeError: 'module' object has no attribute 'getdoc'\n}}}\n\nThe reason is a namespace clash inside icetray.traysegment. I suggest renaming icetray.i3inspect. Alternatively, one could change icetray.traysegment to work without inspect, but this would only be a workaround and we'd likely be hit by this again."",
+ ""reporter"": ""kislat"",
+ ""cc"": """",
+ ""resolution"": ""invalid"",
+ ""_ts"": ""1323968017000000"",
+ ""component"": ""IceTray"",
+ ""summary"": ""icetray.inspect clashes with \""the real\"" python inspect"",
+ ""priority"": ""normal"",
+ ""keywords"": """",
+ ""time"": ""2011-12-12T16:27:53"",
+ ""milestone"": """",
+ ""owner"": """",
+ ""type"": ""defect""
+}
+```
+",0,icetray inspect clashes with the real python inspect trac migrated from json status closed changetime description currently icerec trunk based latest offline rc importing python modules failes e g n n nin from icecube import rootwriter n nattributeerror traceback most recent call last n n home fabian physik software icecube icerec trunk src in n n home fabian physik software icecube icerec trunk build release lib icecube rootwriter init py in n icetray traysegment inherit tableio n removeopts tableservice n def tray name output none kwargs n tabulate data to a root file n n n home fabian physik software icecube icerec trunk build release lib icecube icetray traysegment pyc in traysegment function n n def traysegment function n func traysegment function n func module parent n if defaultoverrides none n n home fabian physik software icecube icerec trunk build release lib icecube icetray traysegment pyc in traysegment function n n n if inspect getdoc function is none n function doc segments should have docstrings this one doesn t fix it n n nattributeerror module object has no attribute getdoc n n nthe reason is a namespace clash inside icetray traysegment i suggest renaming icetray alternatively one could change icetray traysegment to work without inspect but this would only be a workaround and we d likely be hit by this again reporter kislat cc resolution invalid ts component icetray summary icetray inspect clashes with the real python inspect priority normal keywords time milestone owner type defect ,0
+1187,30848007367.0,IssuesEvent,2023-08-02 14:51:57,culturesofknowledge/emlo-project,https://api.github.com/repos/culturesofknowledge/emlo-project,closed,People search resources not displayed correctly,people search feedback0625,"When I search on a person to check the person record to see what information we have for them, the synonyms and roles are working well (better than EMLO-Edit-OLD, although they’re taking up a lot of space, which is hard when you have a long, long list to skim through/compare)
+
+However, it’s
+a) hard to distinguish at a glance what we have, e.g. In EMLO-Edit-OLD, I can see very clearly the different related resources, e.g.:
+
+Compared to EMLO-Edit-New:
+
+b) I’m not able to click on the Descriptors or the links to open the resource from that place (which I often need to do).
+c) and I am not able to copy for pasting any of the data easily.
+I suggest that we don’t display the urls, but rather stick with what was in EMLO-Edit-Old, and have clickable descriptors. You have done this in locations and repositories, etc. and it works well there.",1.0,"People search resources not displayed correctly - When I search on a person to check the person record to see what information we have for them, the synonyms and roles are working well (better than EMLO-Edit-OLD, although they’re taking up a lot of space, which is hard when you have a long, long list to skim through/compare)
+
+However, it’s
+a) hard to distinguish at a glance what we have, e.g. In EMLO-Edit-OLD, I can see very clearly the different related resources, e.g.:
+
+Compared to EMLO-Edit-New:
+
+b) I’m not able to click on the Descriptors or the links to open the resource from that place (which I often need to do).
+c) and I am not able to copy for pasting any of the data easily.
+I suggest that we don’t display the urls, but rather stick with what was in EMLO-Edit-Old, and have clickable descriptors. You have done this in locations and repositories, etc. and it works well there.",1,people search resources not displayed correctly when i search on a person to check the person record to see what information we have for them the synonyms and roles are working well better than emlo edit old although they’re taking up a lot of space which is hard when you have a long long list to skim through compare however it’s a hard to distinguish at a glance what we have e g in emlo edit old i can see very clearly the different related resources e g compared to emlo edit new b i’m not able to click on the descriptors or the links to open the resource from that place which i often need to do c and i am not able to copy for pasting any of the data easily i suggest that we don’t display the urls but rather stick with what was in emlo edit old and have clickable descriptors you have done this in locations and repositories etc and it works well there ,1
+49257,3001867995.0,IssuesEvent,2015-07-24 14:11:07,jayway/powermock,https://api.github.com/repos/jayway/powermock,closed,ExpectNew probably fails for varargs invocation when var args is not the first parameter,bug imported Milestone-Release1.4 Priority-Medium,"_From [johan.ha...@gmail.com](https://code.google.com/u/105676376875942041029/) on September 11, 2009 16:04:41_
+
+fix this
+
+_Original issue: http://code.google.com/p/powermock/issues/detail?id=163_",1.0,"ExpectNew probably fails for varargs invocation when var args is not the first parameter - _From [johan.ha...@gmail.com](https://code.google.com/u/105676376875942041029/) on September 11, 2009 16:04:41_
+
+fix this
+
+_Original issue: http://code.google.com/p/powermock/issues/detail?id=163_",0,expectnew probably fails for varargs invocation when var args is not the first parameter from on september fix this original issue ,0
+498828,14434357301.0,IssuesEvent,2020-12-07 06:55:42,boostcamp-2020/Project01-C-User-Event-Collector,https://api.github.com/repos/boostcamp-2020/Project01-C-User-Event-Collector,opened,"[WEB] Library 페이지 (Mixtape, Artist, Playlist, Track, Album) 이벤트 로깅",🌟 high-priority 👩💻 Web 👾 frontend,"## Todo
+- Library 페이지 (Mixtape, Artist, Playlist, Track, Album) 이벤트 로깅",1.0,"[WEB] Library 페이지 (Mixtape, Artist, Playlist, Track, Album) 이벤트 로깅 - ## Todo
+- Library 페이지 (Mixtape, Artist, Playlist, Track, Album) 이벤트 로깅",0, library 페이지 mixtape artist playlist track album 이벤트 로깅 todo library 페이지 mixtape artist playlist track album 이벤트 로깅,0
+60,3035224434.0,IssuesEvent,2015-08-06 01:00:06,ufvivotech/ufDataQualityImprovement,https://api.github.com/repos/ufvivotech/ufDataQualityImprovement,opened,Multiple home departments for people,bug Data Mgt People,"People should have single home departments. Looks like people ingest was not updating homedept, but rather adding homedepts when new ones came along. There are now many people wiht multiple homedepts, some with as many as 4.",1.0,"Multiple home departments for people - People should have single home departments. Looks like people ingest was not updating homedept, but rather adding homedepts when new ones came along. There are now many people wiht multiple homedepts, some with as many as 4.",1,multiple home departments for people people should have single home departments looks like people ingest was not updating homedept but rather adding homedepts when new ones came along there are now many people wiht multiple homedepts some with as many as ,1
+529,9303252265.0,IssuesEvent,2019-03-24 16:07:56,JohnnySn0w/MoD,https://api.github.com/repos/JohnnySn0w/MoD,closed,Get access to dynamo web interface on host system,people and learning,"Dynamo has a web interface, but we need to be able to access it on the host, so there should be some way to passthrough that connection.",1.0,"Get access to dynamo web interface on host system - Dynamo has a web interface, but we need to be able to access it on the host, so there should be some way to passthrough that connection.",1,get access to dynamo web interface on host system dynamo has a web interface but we need to be able to access it on the host so there should be some way to passthrough that connection ,1
+242121,18516256147.0,IssuesEvent,2021-10-20 10:24:46,CemrgDevelopers/CemrgApp,https://api.github.com/repos/CemrgDevelopers/CemrgApp,opened,Add description on CemrgScar3D where placeholder variables have been commented,documentation ci/cd,"In [CemrgScar3D](https://github.com/CemrgDevelopers/CemrgApp/blob/development/CemrgApp/Modules/CemrgAppModule/src/CemrgScar3D.cpp), we commented some unused placeholder variables to pass the code-analysis tests. This issue is to remind us to write a more thorough description of why we have those variables there (if/when they're needed)",1.0,"Add description on CemrgScar3D where placeholder variables have been commented - In [CemrgScar3D](https://github.com/CemrgDevelopers/CemrgApp/blob/development/CemrgApp/Modules/CemrgAppModule/src/CemrgScar3D.cpp), we commented some unused placeholder variables to pass the code-analysis tests. This issue is to remind us to write a more thorough description of why we have those variables there (if/when they're needed)",0,add description on where placeholder variables have been commented in we commented some unused placeholder variables to pass the code analysis tests this issue is to remind us to write a more thorough description of why we have those variables there if when they re needed ,0
+264138,28106234361.0,IssuesEvent,2023-03-31 01:05:33,Nivaskumark/CVE-2020-0097-frameworks_base_after,https://api.github.com/repos/Nivaskumark/CVE-2020-0097-frameworks_base_after,reopened,CVE-2020-0203 (High) detected in baseandroid-10.0.0_r14,Mend: dependency security vulnerability,"## CVE-2020-0203 - High Severity Vulnerability
+ Vulnerable Library - baseandroid-10.0.0_r14
+
+
+In freeIsolatedUidLocked of ProcessList.java, there is a possible UID reuse due to improper cleanup. This could lead to local escalation of privilege between constrained processes with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-10Android ID: A-146313311
+
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",True,"CVE-2020-0203 (High) detected in baseandroid-10.0.0_r14 - ## CVE-2020-0203 - High Severity Vulnerability
+ Vulnerable Library - baseandroid-10.0.0_r14
+
+
+In freeIsolatedUidLocked of ProcessList.java, there is a possible UID reuse due to improper cleanup. This could lead to local escalation of privilege between constrained processes with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-10Android ID: A-146313311
+
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",0,cve high detected in baseandroid cve high severity vulnerability vulnerable library baseandroid android framework classes and services library home page a href found in head commit a href found in base branch master vulnerable source files core java com android server am processlist java vulnerability details in freeisolateduidlocked of processlist java there is a possible uid reuse due to improper cleanup this could lead to local escalation of privilege between constrained processes with no additional execution privileges needed user interaction is not needed for exploitation product androidversions android id a publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution android step up your open source security game with mend ,0
+668,11906372960.0,IssuesEvent,2020-03-30 20:14:23,newtheatre/history-project,https://api.github.com/repos/newtheatre/history-project,closed,Rohan Rakhit bio submission,bio people report-tool,"End user submitted issue from page: [/collect/person/](https://history.newtheatre.org.uk/collect/person/)
+---
+# 'Submit an almni bio' form submission
+
+Field | Data
+----- | ----
+Name | Rohan Rakhit
+Grad Year | 2020
+Course | History BA
+
+## Bio1 (Time at theatre)
+
+
+
+
+## Bio2 (Post-graduation)
+
+
+
+Field | Data
+------| ----
+Checked careers |
+Other careers |
+
+## Links
+
+Twitter: @RRakhit
+
+Instagram: rohan.rakhit
+
+## Shows
+
+
+
+## Committees
+
+
+
+## Awards
+
+Was invited to join the NSDF Company Ensemble in 2019 for my role as 'Perseus' in Here, the World Entire.
+
+## Contact Preferences
+
+Are we allowed to facilitate contact to this alumnus? **Yes**
+
+## Attempted File Generation
+
+```
+---
+title: Rohan Rakhit
+course:
+ - History BA
+graduated: 2020
+contact_allowed: true
+careers:
+
+
+links: *fill me out
+Twitter: @RRakhit
+
+Instagram: rohan.rakhit
+award: *fill me out
+Was invited to join the NSDF Company Ensemble in 2019 for my role as 'Perseus' in Here, the World Entire.
+---
+
+
+
+
+
+
+```
+**
+",1.0,"Rohan Rakhit bio submission - End user submitted issue from page: [/collect/person/](https://history.newtheatre.org.uk/collect/person/)
+---
+# 'Submit an almni bio' form submission
+
+Field | Data
+----- | ----
+Name | Rohan Rakhit
+Grad Year | 2020
+Course | History BA
+
+## Bio1 (Time at theatre)
+
+
+
+
+## Bio2 (Post-graduation)
+
+
+
+Field | Data
+------| ----
+Checked careers |
+Other careers |
+
+## Links
+
+Twitter: @RRakhit
+
+Instagram: rohan.rakhit
+
+## Shows
+
+
+
+## Committees
+
+
+
+## Awards
+
+Was invited to join the NSDF Company Ensemble in 2019 for my role as 'Perseus' in Here, the World Entire.
+
+## Contact Preferences
+
+Are we allowed to facilitate contact to this alumnus? **Yes**
+
+## Attempted File Generation
+
+```
+---
+title: Rohan Rakhit
+course:
+ - History BA
+graduated: 2020
+contact_allowed: true
+careers:
+
+
+links: *fill me out
+Twitter: @RRakhit
+
+Instagram: rohan.rakhit
+award: *fill me out
+Was invited to join the NSDF Company Ensemble in 2019 for my role as 'Perseus' in Here, the World Entire.
+---
+
+
+
+
+
+
+```
+**
+",1,rohan rakhit bio submission end user submitted issue from page submit an almni bio form submission field data name rohan rakhit grad year course history ba time at theatre post graduation field data checked careers other careers links twitter rrakhit instagram rohan rakhit shows committees awards was invited to join the nsdf company ensemble in for my role as perseus in here the world entire contact preferences are we allowed to facilitate contact to this alumnus yes attempted file generation title rohan rakhit course history ba graduated contact allowed true careers links fill me out twitter rrakhit instagram rohan rakhit award fill me out was invited to join the nsdf company ensemble in for my role as perseus in here the world entire ,1
+254651,19255179262.0,IssuesEvent,2021-12-09 10:29:53,KDAB/cxx-qt,https://api.github.com/repos/KDAB/cxx-qt,opened,Document enabling LTO,documentation good first issue,Potentially document in the Rust Book how enabling lto in the cargo.toml of the project can reduce the plugin size (eg we went from 5MB → 2MB),1.0,Document enabling LTO - Potentially document in the Rust Book how enabling lto in the cargo.toml of the project can reduce the plugin size (eg we went from 5MB → 2MB),0,document enabling lto potentially document in the rust book how enabling lto in the cargo toml of the project can reduce the plugin size eg we went from → ,0
+34637,30239613628.0,IssuesEvent,2023-07-06 12:42:49,AbsaOSS/atum-service,https://api.github.com/repos/AbsaOSS/atum-service,closed,Auto-assign project to a PR after opening,blocked infrastructure,Add an action to assign PR to the Atum service project upon creation,1.0,Auto-assign project to a PR after opening - Add an action to assign PR to the Atum service project upon creation,0,auto assign project to a pr after opening add an action to assign pr to the atum service project upon creation,0
+898,18732446315.0,IssuesEvent,2021-11-04 00:13:28,Andidy/engine-v2,https://api.github.com/repos/Andidy/engine-v2,closed,"Create ""ToString()"" methods for each component type",new people implmentation,"The `ToString()` method is used in the `WriteEntityToFile()` function that is used to save Entity data to disk. It can later be reloaded by the `ReadEntityFromFile()` function.
+
+Here is an example of the output from `cTransform` and `cGridTransform`:
+`component transform = [ 490.000000, 420.000000 ] ;`
+`component grid_transform = [ 49, 42 ] ;`
+
+Here is an example of the code used by `cGridTransform`:
+https://github.com/Andidy/engine-v2/blob/d3f47492e45b14edad01d5dbbcfa1d88cdd094d3/engine-v2/src/grid_transform.cpp#L3-L11
+
+The space at the end before the semi-colon is intentional.",1.0,"Create ""ToString()"" methods for each component type - The `ToString()` method is used in the `WriteEntityToFile()` function that is used to save Entity data to disk. It can later be reloaded by the `ReadEntityFromFile()` function.
+
+Here is an example of the output from `cTransform` and `cGridTransform`:
+`component transform = [ 490.000000, 420.000000 ] ;`
+`component grid_transform = [ 49, 42 ] ;`
+
+Here is an example of the code used by `cGridTransform`:
+https://github.com/Andidy/engine-v2/blob/d3f47492e45b14edad01d5dbbcfa1d88cdd094d3/engine-v2/src/grid_transform.cpp#L3-L11
+
+The space at the end before the semi-colon is intentional.",1,create tostring methods for each component type the tostring method is used in the writeentitytofile function that is used to save entity data to disk it can later be reloaded by the readentityfromfile function here is an example of the output from ctransform and cgridtransform component transform component grid transform here is an example of the code used by cgridtransform the space at the end before the semi colon is intentional ,1
+168999,26724930596.0,IssuesEvent,2023-01-29 15:48:12,codestates-seb/seb41_main_028,https://api.github.com/repos/codestates-seb/seb41_main_028,opened,"[FE] feat: OAuth 페이지 헤더 숨기기, 회원가입시 나이-성별 추가, OAuth 나이-성별 입력페이지 추가, patch 관련 모듈 수정",FE Feat Design,"## To do List
+
+- [ ] OAuth 페이지 헤더 숨기기
+- [ ] 회원가입시 나이-성별 추가
+- [ ] OAuth 나이-성별 입력페이지 추가
+- [ ] patch 관련 모듈 수정
+
+## Deadline
+
+
+
+## Issue Checklist
+아래 체크리스트를 확인해주세요.
+- [ ] Issue Title을 다음 예시와 같이 작성했습니다. (e.g. [FE] Feat: 회원가입 기능 구현 )
+- [ ] 우측의 Assignees, Labels, Projects, Milestone을 적절하게 선택했습니다.
+",1.0,"[FE] feat: OAuth 페이지 헤더 숨기기, 회원가입시 나이-성별 추가, OAuth 나이-성별 입력페이지 추가, patch 관련 모듈 수정 - ## To do List
+
+- [ ] OAuth 페이지 헤더 숨기기
+- [ ] 회원가입시 나이-성별 추가
+- [ ] OAuth 나이-성별 입력페이지 추가
+- [ ] patch 관련 모듈 수정
+
+## Deadline
+
+
+
+## Issue Checklist
+아래 체크리스트를 확인해주세요.
+- [ ] Issue Title을 다음 예시와 같이 작성했습니다. (e.g. [FE] Feat: 회원가입 기능 구현 )
+- [ ] 우측의 Assignees, Labels, Projects, Milestone을 적절하게 선택했습니다.
+",0, feat oauth 페이지 헤더 숨기기 회원가입시 나이 성별 추가 oauth 나이 성별 입력페이지 추가 patch 관련 모듈 수정 to do list oauth 페이지 헤더 숨기기 회원가입시 나이 성별 추가 oauth 나이 성별 입력페이지 추가 patch 관련 모듈 수정 deadline issue checklist 아래 체크리스트를 확인해주세요 issue title을 다음 예시와 같이 작성했습니다 e g feat 회원가입 기능 구현 우측의 assignees labels projects milestone을 적절하게 선택했습니다 ,0
+922,19577505606.0,IssuesEvent,2022-01-04 16:54:28,openstates/issues,https://api.github.com/repos/openstates/issues,closed,Florida district office addresses for senators need an update,component:people-data type:bug,"**Issue Description:**
+[Aaron Bean](https://openstates.org/person/aaron-bean-7DMnAeW8iVIQHm8uIxCwf1/), [Darry Rouson](https://openstates.org/person/darryl-ervin-rouson-1slOwFmQ6yxQgbgmXjEgYJ/), and [Ben Albritton](https://openstates.org/person/ben-albritton-2oYe6sKJPblPg3Sf9T4gpM/) all have the wrong District Office Address on OpenStates and there are some others as well from a glance but not for all. This problem also seems to be isolated to the Senate as I did not see this problem for the House.
+
+**URL where this issue may be seen (API or OpenStates.org):**
+
+
+**Source URL of correct data if applicable:**
+",1.0,"Florida district office addresses for senators need an update - **Issue Description:**
+[Aaron Bean](https://openstates.org/person/aaron-bean-7DMnAeW8iVIQHm8uIxCwf1/), [Darry Rouson](https://openstates.org/person/darryl-ervin-rouson-1slOwFmQ6yxQgbgmXjEgYJ/), and [Ben Albritton](https://openstates.org/person/ben-albritton-2oYe6sKJPblPg3Sf9T4gpM/) all have the wrong District Office Address on OpenStates and there are some others as well from a glance but not for all. This problem also seems to be isolated to the Senate as I did not see this problem for the House.
+
+**URL where this issue may be seen (API or OpenStates.org):**
+
+
+**Source URL of correct data if applicable:**
+",1,florida district office addresses for senators need an update issue description and all have the wrong district office address on openstates and there are some others as well from a glance but not for all this problem also seems to be isolated to the senate as i did not see this problem for the house url where this issue may be seen api or openstates org source url of correct data if applicable ,1
+443185,12760844128.0,IssuesEvent,2020-06-29 08:45:47,TerriaJS/terriajs,https://api.github.com/repos/TerriaJS/terriajs,closed,"Onboarding v3: ""Terry the trainer"" v1",High priority,"As a new map user,
+I want an always-in-view, step by step guide,
+so I have follow along without relying on my memory
+
+**Design file**: `terry-task-trainer.sketch` in https://drive.google.com/drive/u/0/folders/14c-UE2Cmn3zivZr25Sw8eQv_o_wfc88u
+
+**Design ticket**: https://github.com/TerriaJS/terriajs/issues/4365
+
+**Invision**: https://projects.invisionapp.com/share/X6XK52ZRJ5D
+
+Other notes:
+- very quick inline, ""no smarts"" - aka manual back/next
+
+### DoD:
+- [x] Unit test written (if applicable)
+- [x] Works on mobile (or deliberately disabled for mobile)
+- [x] Code/peer reviewed (if didn't pair)
+- [x] Relevant docs are updated - docs at https://github.com/TerriaJS/terriajs/issues/4421#issue-635052382
+- [x] Previewable/deployed to dev (see separation of deployments)
+- [x] Previewable/deployed to test (see separation of deployments)
+- [ ] (If a map-specific ticket) Issue linked to a release ticket
+
+### Acceptance Criteria:
+- [x] Step by step guide is available as an item in help panel through config
+- [x] Opening an item (e.g. ""adding satellite imagery"" / ""adding data to the map"") from the help pane, loads inline-guide at top
+- [x] Inline-guide has a dropdown to swap between different guides
+- [x] Inline-guide has instructions
+- [x] Inline-guide has back / next
+
+
+",1.0,"Onboarding v3: ""Terry the trainer"" v1 - As a new map user,
+I want an always-in-view, step by step guide,
+so I have follow along without relying on my memory
+
+**Design file**: `terry-task-trainer.sketch` in https://drive.google.com/drive/u/0/folders/14c-UE2Cmn3zivZr25Sw8eQv_o_wfc88u
+
+**Design ticket**: https://github.com/TerriaJS/terriajs/issues/4365
+
+**Invision**: https://projects.invisionapp.com/share/X6XK52ZRJ5D
+
+Other notes:
+- very quick inline, ""no smarts"" - aka manual back/next
+
+### DoD:
+- [x] Unit test written (if applicable)
+- [x] Works on mobile (or deliberately disabled for mobile)
+- [x] Code/peer reviewed (if didn't pair)
+- [x] Relevant docs are updated - docs at https://github.com/TerriaJS/terriajs/issues/4421#issue-635052382
+- [x] Previewable/deployed to dev (see separation of deployments)
+- [x] Previewable/deployed to test (see separation of deployments)
+- [ ] (If a map-specific ticket) Issue linked to a release ticket
+
+### Acceptance Criteria:
+- [x] Step by step guide is available as an item in help panel through config
+- [x] Opening an item (e.g. ""adding satellite imagery"" / ""adding data to the map"") from the help pane, loads inline-guide at top
+- [x] Inline-guide has a dropdown to swap between different guides
+- [x] Inline-guide has instructions
+- [x] Inline-guide has back / next
+
+
+",0,onboarding terry the trainer as a new map user i want an always in view step by step guide so i have follow along without relying on my memory design file terry task trainer sketch in design ticket invision other notes very quick inline no smarts aka manual back next dod unit test written if applicable works on mobile or deliberately disabled for mobile code peer reviewed if didn t pair relevant docs are updated docs at previewable deployed to dev see separation of deployments previewable deployed to test see separation of deployments if a map specific ticket issue linked to a release ticket acceptance criteria step by step guide is available as an item in help panel through config opening an item e g adding satellite imagery adding data to the map from the help pane loads inline guide at top inline guide has a dropdown to swap between different guides inline guide has instructions inline guide has back next ,0
+40866,6875139440.0,IssuesEvent,2017-11-19 10:08:31,react-community/react-navigation,https://api.github.com/repos/react-community/react-navigation,closed,CardStack docs or examples,7 days countdown documentation question,"I searched through the examples and documentation, and can't find any examples of CardStack usage. I'm trying to upgrade the `@shoutem/ui` library to pull the CardStack and Card components from this library, since they were deprecated and moved from NavigationExperimental, but can't find any documentation or examples on which to base it on (which would be OK if the API hadn't changed). Could someone point me to a working example of CardStack?
+
+Thanks!",1.0,"CardStack docs or examples - I searched through the examples and documentation, and can't find any examples of CardStack usage. I'm trying to upgrade the `@shoutem/ui` library to pull the CardStack and Card components from this library, since they were deprecated and moved from NavigationExperimental, but can't find any documentation or examples on which to base it on (which would be OK if the API hadn't changed). Could someone point me to a working example of CardStack?
+
+Thanks!",0,cardstack docs or examples i searched through the examples and documentation and can t find any examples of cardstack usage i m trying to upgrade the shoutem ui library to pull the cardstack and card components from this library since they were deprecated and moved from navigationexperimental but can t find any documentation or examples on which to base it on which would be ok if the api hadn t changed could someone point me to a working example of cardstack thanks ,0
+154543,13552610051.0,IssuesEvent,2020-09-17 12:50:33,airctic/icevision,https://api.github.com/repos/airctic/icevision,opened,Update CONTRIBUTING Guide,documentation enhancement good first issue,"## 📓 Documentation Update
+Add info about how to resolve conflicts in the CONTRIBUTING Guide
+",1.0,"Update CONTRIBUTING Guide - ## 📓 Documentation Update
+Add info about how to resolve conflicts in the CONTRIBUTING Guide
+",0,update contributing guide 📓 documentation update add info about how to resolve conflicts in the contributing guide ,0
+975,22440595828.0,IssuesEvent,2022-06-21 00:46:09,wordpress-mobile/WordPress-iOS,https://api.github.com/repos/wordpress-mobile/WordPress-iOS,closed,WPiOS crashed when clicking on the People tab,[Type] Bug [Type] Crash People Management,"A user reported that when they click on the People link under **CONFIGURE**, their app shuts down immediately. They reinstalled the app but that didn't help.
+
+I found the [Sentry reports](https://sentry.io/organizations/a8c/discover/results/?field=title&field=event.type&field=project&field=user.display&field=timestamp&name=All+Events&project=1438083&query=user.email%3Agregcrandall%40mac.com&sort=-timestamp&statsPeriod=7d&yAxis=count%28%29) with the following error:
+
+```
+NSInvalidArgumentException: no object at index 1 in section at index 0
+```
+
+##### User/device details
+Ticket link: 4663505-zen
+App version: 18.8.0.5
+Device model: iPad Pro 12.9-inch 4th-gen (WiFi) (iPad8,11)
+OS version: iOS 15.2
+Type of site (Simple, Atomic, Jetpack, or self-hosted): Simple",1.0,"WPiOS crashed when clicking on the People tab - A user reported that when they click on the People link under **CONFIGURE**, their app shuts down immediately. They reinstalled the app but that didn't help.
+
+I found the [Sentry reports](https://sentry.io/organizations/a8c/discover/results/?field=title&field=event.type&field=project&field=user.display&field=timestamp&name=All+Events&project=1438083&query=user.email%3Agregcrandall%40mac.com&sort=-timestamp&statsPeriod=7d&yAxis=count%28%29) with the following error:
+
+```
+NSInvalidArgumentException: no object at index 1 in section at index 0
+```
+
+##### User/device details
+Ticket link: 4663505-zen
+App version: 18.8.0.5
+Device model: iPad Pro 12.9-inch 4th-gen (WiFi) (iPad8,11)
+OS version: iOS 15.2
+Type of site (Simple, Atomic, Jetpack, or self-hosted): Simple",1,wpios crashed when clicking on the people tab a user reported that when they click on the people link under configure their app shuts down immediately they reinstalled the app but that didn t help i found the with the following error nsinvalidargumentexception no object at index in section at index user device details ticket link zen app version device model ipad pro inch gen wifi os version ios type of site simple atomic jetpack or self hosted simple,1
+270886,29145867155.0,IssuesEvent,2023-05-18 02:48:15,aayant-mend/WebGoat,https://api.github.com/repos/aayant-mend/WebGoat,opened,spring-boot-starter-security-2.7.1.jar: 2 vulnerabilities (highest severity is: 9.8),Mend: dependency security vulnerability," Vulnerable Library - spring-boot-starter-security-2.7.1.jar
+
+
+
Path to dependency file: /pom.xml
+
Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/security/spring-security-web/5.7.2/spring-security-web-5.7.2.jar
+
+ #### Mend has checked all newer package trees, and you are on the least vulnerable package!
+
+ #### Please note: There might be a version that explicitly solves one or more of the vulnerabilities listed below, but we do not recommend it. For more info about the optional fixes, check the ""Details"" section below.
+
+## Vulnerabilities
+
+| CVE | Severity | CVSS | Dependency | Type | Fixed in (spring-boot-starter-security version) | Fix PR available |
+| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
+| [CVE-2023-20862](https://www.mend.io/vulnerability-database/CVE-2023-20862) | High | 9.8 | spring-security-web-5.7.2.jar | Transitive | N/A* | ❌ |
+| [CVE-2022-31692](https://www.mend.io/vulnerability-database/CVE-2022-31692) | High | 9.8 | spring-security-web-5.7.2.jar | Transitive | N/A* | ❌ |
+
*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the ""Details"" section below to see if there is a version of transitive dependency where vulnerability is fixed.
+
+In Spring Security, versions 5.7.x prior to 5.7.8, versions 5.8.x prior to 5.8.3, and versions 6.0.x prior to 6.0.3, the logout support does not properly clean the security context if using serialized versions. Additionally, it is not possible to explicitly save an empty security context to the HttpSessionSecurityContextRepository. This vulnerability can keep users authenticated even after they performed logout. Users of affected versions should apply the following mitigation. 5.7.x users should upgrade to 5.7.8. 5.8.x users should upgrade to 5.8.3. 6.0.x users should upgrade to 6.0.3.
+
+
+
+Spring Security, versions 5.7 prior to 5.7.5 and 5.6 prior to 5.6.9 could be susceptible to authorization rules bypass via forward or include dispatcher types. Specifically, an application is vulnerable when all of the following are true: The application expects that Spring Security applies security to forward and include dispatcher types. The application uses the AuthorizationFilter either manually or via the authorizeHttpRequests() method. The application configures the FilterChainProxy to apply to forward and/or include requests (e.g. spring.security.filter.dispatcher-types = request, error, async, forward, include). The application may forward or include the request to a higher privilege-secured endpoint.The application configures Spring Security to apply to every dispatcher type via authorizeHttpRequests().shouldFilterAllDispatcherTypes(true)
+
+
+
+ #### Mend has checked all newer package trees, and you are on the least vulnerable package!
+
+ #### Please note: There might be a version that explicitly solves one or more of the vulnerabilities listed below, but we do not recommend it. For more info about the optional fixes, check the ""Details"" section below.
+
+## Vulnerabilities
+
+| CVE | Severity | CVSS | Dependency | Type | Fixed in (spring-boot-starter-security version) | Fix PR available |
+| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
+| [CVE-2023-20862](https://www.mend.io/vulnerability-database/CVE-2023-20862) | High | 9.8 | spring-security-web-5.7.2.jar | Transitive | N/A* | ❌ |
+| [CVE-2022-31692](https://www.mend.io/vulnerability-database/CVE-2022-31692) | High | 9.8 | spring-security-web-5.7.2.jar | Transitive | N/A* | ❌ |
+
*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the ""Details"" section below to see if there is a version of transitive dependency where vulnerability is fixed.
+
+In Spring Security, versions 5.7.x prior to 5.7.8, versions 5.8.x prior to 5.8.3, and versions 6.0.x prior to 6.0.3, the logout support does not properly clean the security context if using serialized versions. Additionally, it is not possible to explicitly save an empty security context to the HttpSessionSecurityContextRepository. This vulnerability can keep users authenticated even after they performed logout. Users of affected versions should apply the following mitigation. 5.7.x users should upgrade to 5.7.8. 5.8.x users should upgrade to 5.8.3. 6.0.x users should upgrade to 6.0.3.
+
+
+
+Spring Security, versions 5.7 prior to 5.7.5 and 5.6 prior to 5.6.9 could be susceptible to authorization rules bypass via forward or include dispatcher types. Specifically, an application is vulnerable when all of the following are true: The application expects that Spring Security applies security to forward and include dispatcher types. The application uses the AuthorizationFilter either manually or via the authorizeHttpRequests() method. The application configures the FilterChainProxy to apply to forward and/or include requests (e.g. spring.security.filter.dispatcher-types = request, error, async, forward, include). The application may forward or include the request to a higher privilege-secured endpoint.The application configures Spring Security to apply to every dispatcher type via authorizeHttpRequests().shouldFilterAllDispatcherTypes(true)
+
+
+
+
+
+
+
+",0,spring boot starter security jar vulnerabilities highest severity is vulnerable library spring boot starter security jar path to dependency file pom xml path to vulnerable library home wss scanner repository org springframework security spring security web spring security web jar found in head commit a href mend has checked all newer package trees and you are on the least vulnerable package please note there might be a version that explicitly solves one or more of the vulnerabilities listed below but we do not recommend it for more info about the optional fixes check the details section below vulnerabilities cve severity cvss dependency type fixed in spring boot starter security version fix pr available high spring security web jar transitive n a high spring security web jar transitive n a for some transitive vulnerabilities there is no version of direct dependency with a fix check the details section below to see if there is a version of transitive dependency where vulnerability is fixed details cve vulnerable library spring security web jar spring security library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository org springframework security spring security web spring security web jar dependency hierarchy spring boot starter security jar root library x spring security web jar vulnerable library found in head commit a href found in base branch main vulnerability details in spring security versions x prior to versions x prior to and versions x prior to the logout support does not properly clean the security context if using serialized versions additionally it is not possible to explicitly save an empty security context to the httpsessionsecuritycontextrepository this vulnerability can keep users authenticated even after they performed logout users of affected versions should apply the following mitigation x users should upgrade to x users should upgrade to x users should upgrade to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org springframework security spring security web cve vulnerable library spring security web jar spring security library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository org springframework security spring security web spring security web jar dependency hierarchy spring boot starter security jar root library x spring security web jar vulnerable library found in head commit a href found in base branch main vulnerability details spring security versions prior to and prior to could be susceptible to authorization rules bypass via forward or include dispatcher types specifically an application is vulnerable when all of the following are true the application expects that spring security applies security to forward and include dispatcher types the application uses the authorizationfilter either manually or via the authorizehttprequests method the application configures the filterchainproxy to apply to forward and or include requests e g spring security filter dispatcher types request error async forward include the application may forward or include the request to a higher privilege secured endpoint the application configures spring security to apply to every dispatcher type via authorizehttprequests shouldfilteralldispatchertypes true publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org springframework security spring security web ,0
+303,5639725396.0,IssuesEvent,2017-04-06 14:55:18,PopulateTools/gobierto,https://api.github.com/repos/PopulateTools/gobierto,opened,People > Add support for boolean settings,enhancement gobierto-people,"Right now the settings related to activation / deactivation of Gobierto People submodules requiere to type 'true' or 'false'. Apart from strings, we should permit boolean settings so we can display this as checkboxes instead of text areas.
+
+
+",1.0,"People > Add support for boolean settings - Right now the settings related to activation / deactivation of Gobierto People submodules requiere to type 'true' or 'false'. Apart from strings, we should permit boolean settings so we can display this as checkboxes instead of text areas.
+
+
+",1,people add support for boolean settings right now the settings related to activation deactivation of gobierto people submodules requiere to type true or false apart from strings we should permit boolean settings so we can display this as checkboxes instead of text areas ,1
+280218,30807616624.0,IssuesEvent,2023-08-01 08:17:10,dotnet/runtime,https://api.github.com/repos/dotnet/runtime,closed,AesCryptoServiceProvider,area-System.Security,"When using AesCryptoServiceProvider for encryption, I found that when setting the Mode property to CipherMode.CFB, the public properties were modified but the protected properties were not when viewing the variables in the Core debugging mode. This raises concerns about potential errors in this encryption algorithm.
+
+Although I am aware that AesCryptoServiceProvider is obsolete in net core, my project needs to be compatible for some time, and the encrypted data generated in net core needs to be decrypted in net framework, so I need to address this issue.
+
+I hope that Microsoft can help fix this problem and ensure that AesCryptoServiceProvider works properly on all platforms. This is essential to ensure the security and reliability of programs using this encryption algorithm.
+```
+
+ public class AES
+ {
+
+ public static string ConvertByteArrayToHexString(byte[] input)
+ {
+ bool flag;
+
+ StringBuilder stringBuilder = new StringBuilder();
+ byte[] bs = input;
+ int i = 0;
+ do
+ {
+ byte b = bs[i];
+ stringBuilder.AppendFormat(""{0:x2}"", b);
+ i++;
+
+ flag = i < (int)bs.Length;
+ }
+ while (flag);
+ return stringBuilder.ToString();
+ }
+
+
+ public static byte[] ConvertHexStringToByteArray(string hexString)
+ {
+ bool flag = (hexString.Length & 1) == 0;
+ if (!flag)
+ {
+ throw new ArgumentOutOfRangeException(""hexString"", hexString, ""hexString must contain an even number of characters."");
+ }
+ byte[] bs1 = new byte[hexString.Length / 2];
+ int i = 0;
+ do
+ {
+ bs1[i / 2] = Byte.Parse(hexString.Substring(i, 2), NumberStyles.HexNumber);
+ i += 2;
+
+ flag = i < hexString.Length;
+ }
+ while (flag);
+ return bs1;
+ }
+
+
+
+ public static string GetEncode(string source, string key)
+ {
+ // char[] c = key.ToCharArray();
+ byte[] keys = Encoding.UTF8.GetBytes(key.Substring(0, 32));
+ byte[] iv = Encoding.UTF8.GetBytes(key.Substring(0, 16));
+
+
+ AesCryptoServiceProvider aes = new AesCryptoServiceProvider();
+ aes.Mode = CipherMode.CFB;
+ aes.Padding = PaddingMode.PKCS7;
+ aes.BlockSize = 128;
+ aes.KeySize = 256;
+ aes.Key = keys;
+ aes.IV = iv;
+
+ byte[] cipherText = null;
+ byte[] rawPlaintext = Encoding.Unicode.GetBytes(source);
+ byte[] cipherText2 = null;
+
+
+ using (MemoryStream ms = new MemoryStream())
+ {
+ using (CryptoStream cs = new CryptoStream(ms, aes.CreateEncryptor(), CryptoStreamMode.Write))
+ {
+ cs.Write(rawPlaintext, 0, rawPlaintext.Length);
+ }
+
+ cipherText = ms.ToArray();
+ }
+
+ string s = ConvertByteArrayToHexString(cipherText);
+
+
+ return s;
+ }
+
+ public static string GetDecode(string source, string key)
+ {
+
+ byte[] keys = Encoding.UTF8.GetBytes(key.Substring(0, 32));
+ byte[] iv = Encoding.UTF8.GetBytes(key.Substring(0, 16));
+
+ AesCryptoServiceProvider aes = new AesCryptoServiceProvider();
+ aes.BlockSize = 128;
+ aes.KeySize = 256;
+ aes.Key = keys;
+ aes.IV = iv;
+ aes.Mode = CipherMode.CFB;
+ aes.Padding = PaddingMode.PKCS7;
+
+ byte[] cipherText = ConvertHexStringToByteArray(source);
+
+ byte[] plainText = null;
+ using (MemoryStream ms = new MemoryStream())
+ {
+ try
+ {
+ using (CryptoStream cs = new CryptoStream(ms, aes.CreateDecryptor(), CryptoStreamMode.Write))
+ {
+ cs.Write(cipherText, 0, cipherText.Length);
+ }
+ }
+ catch
+ {
+ ////無法解密
+ return null;
+ }
+
+ plainText = ms.ToArray();
+ }
+
+ string s = Encoding.Unicode.GetString(plainText);
+ return s;
+ }
+ }
+
+```
+Here is Console For NET7
+```
+Console.WriteLine(""=======================NET7"");
+string key = ""D36C2FFC94649A3EB946152985CC0B7317F89077FD4F38C1"";
+var strSource = ""test123"";
+
+var strEncode = AES.GetEncode(strSource, key);
+
+Console.WriteLine($""Encode:{strEncode}"");
+
+var strDecode = AES.GetDecode(strEncode, key);
+
+Console.WriteLine($""Decode:{strDecode}"");
+
+var strNetFrameworkEncode = ""39000c17c7fb12ee58f20239197605d4"";
+Console.WriteLine($""Encode From NET Framework:{strNetFrameworkEncode}"");
+Console.WriteLine($""Decode From NET Framework:{AES.GetDecode(strNetFrameworkEncode, key)}"");
+Console.WriteLine(""=======================NET7"");
+
+//OutPut:
+//Encode: 39000c17c7fb12ee58f20239197606
+//Decode:test123
+//Encode From NET Framework:39000c17c7fb12ee58f20239197605d4
+//Decode From NET Framework:test123
+```
+Here is Console For NET Framework 4.6.1
+```
+
+internal class Program
+ {
+ private static void Main(string[] args)
+ {
+ Console.WriteLine(""=======================NET FRAMEWORK"");
+ string key = ""D36C2FFC94649A3EB946152985CC0B7317F89077FD4F38C1"";
+ var strSource = ""test123"";
+
+ var strEncode = AES.GetEncode(strSource, key);
+
+ var strDecode = AES.GetDecode(strEncode, key);
+
+ var strNet7Encode = ""39000c17c7fb12ee58f20239197606"";
+
+ Console.WriteLine($""Encode:{strEncode}"");
+ Console.WriteLine($""Decode:{strDecode}"");
+ Console.WriteLine($""Encode From NET 7:{strNet7Encode}"");
+ Console.WriteLine($""Decode From NET 7:{AES.GetDecode(strNet7Encode, key)}"");
+ Console.WriteLine(""=======================NET FRAMEWORK"");
+ //OutPut:
+ //Encode:39000c17c7fb12ee58f20239197605d4
+ //Decode:test123
+ //Encode From NET 7:39000c17c7fb12ee58f20239197606
+ //Decode From NET 7:
+ }
+ }
+```
+",True,"AesCryptoServiceProvider - When using AesCryptoServiceProvider for encryption, I found that when setting the Mode property to CipherMode.CFB, the public properties were modified but the protected properties were not when viewing the variables in the Core debugging mode. This raises concerns about potential errors in this encryption algorithm.
+
+Although I am aware that AesCryptoServiceProvider is obsolete in net core, my project needs to be compatible for some time, and the encrypted data generated in net core needs to be decrypted in net framework, so I need to address this issue.
+
+I hope that Microsoft can help fix this problem and ensure that AesCryptoServiceProvider works properly on all platforms. This is essential to ensure the security and reliability of programs using this encryption algorithm.
+```
+
+ public class AES
+ {
+
+ public static string ConvertByteArrayToHexString(byte[] input)
+ {
+ bool flag;
+
+ StringBuilder stringBuilder = new StringBuilder();
+ byte[] bs = input;
+ int i = 0;
+ do
+ {
+ byte b = bs[i];
+ stringBuilder.AppendFormat(""{0:x2}"", b);
+ i++;
+
+ flag = i < (int)bs.Length;
+ }
+ while (flag);
+ return stringBuilder.ToString();
+ }
+
+
+ public static byte[] ConvertHexStringToByteArray(string hexString)
+ {
+ bool flag = (hexString.Length & 1) == 0;
+ if (!flag)
+ {
+ throw new ArgumentOutOfRangeException(""hexString"", hexString, ""hexString must contain an even number of characters."");
+ }
+ byte[] bs1 = new byte[hexString.Length / 2];
+ int i = 0;
+ do
+ {
+ bs1[i / 2] = Byte.Parse(hexString.Substring(i, 2), NumberStyles.HexNumber);
+ i += 2;
+
+ flag = i < hexString.Length;
+ }
+ while (flag);
+ return bs1;
+ }
+
+
+
+ public static string GetEncode(string source, string key)
+ {
+ // char[] c = key.ToCharArray();
+ byte[] keys = Encoding.UTF8.GetBytes(key.Substring(0, 32));
+ byte[] iv = Encoding.UTF8.GetBytes(key.Substring(0, 16));
+
+
+ AesCryptoServiceProvider aes = new AesCryptoServiceProvider();
+ aes.Mode = CipherMode.CFB;
+ aes.Padding = PaddingMode.PKCS7;
+ aes.BlockSize = 128;
+ aes.KeySize = 256;
+ aes.Key = keys;
+ aes.IV = iv;
+
+ byte[] cipherText = null;
+ byte[] rawPlaintext = Encoding.Unicode.GetBytes(source);
+ byte[] cipherText2 = null;
+
+
+ using (MemoryStream ms = new MemoryStream())
+ {
+ using (CryptoStream cs = new CryptoStream(ms, aes.CreateEncryptor(), CryptoStreamMode.Write))
+ {
+ cs.Write(rawPlaintext, 0, rawPlaintext.Length);
+ }
+
+ cipherText = ms.ToArray();
+ }
+
+ string s = ConvertByteArrayToHexString(cipherText);
+
+
+ return s;
+ }
+
+ public static string GetDecode(string source, string key)
+ {
+
+ byte[] keys = Encoding.UTF8.GetBytes(key.Substring(0, 32));
+ byte[] iv = Encoding.UTF8.GetBytes(key.Substring(0, 16));
+
+ AesCryptoServiceProvider aes = new AesCryptoServiceProvider();
+ aes.BlockSize = 128;
+ aes.KeySize = 256;
+ aes.Key = keys;
+ aes.IV = iv;
+ aes.Mode = CipherMode.CFB;
+ aes.Padding = PaddingMode.PKCS7;
+
+ byte[] cipherText = ConvertHexStringToByteArray(source);
+
+ byte[] plainText = null;
+ using (MemoryStream ms = new MemoryStream())
+ {
+ try
+ {
+ using (CryptoStream cs = new CryptoStream(ms, aes.CreateDecryptor(), CryptoStreamMode.Write))
+ {
+ cs.Write(cipherText, 0, cipherText.Length);
+ }
+ }
+ catch
+ {
+ ////無法解密
+ return null;
+ }
+
+ plainText = ms.ToArray();
+ }
+
+ string s = Encoding.Unicode.GetString(plainText);
+ return s;
+ }
+ }
+
+```
+Here is Console For NET7
+```
+Console.WriteLine(""=======================NET7"");
+string key = ""D36C2FFC94649A3EB946152985CC0B7317F89077FD4F38C1"";
+var strSource = ""test123"";
+
+var strEncode = AES.GetEncode(strSource, key);
+
+Console.WriteLine($""Encode:{strEncode}"");
+
+var strDecode = AES.GetDecode(strEncode, key);
+
+Console.WriteLine($""Decode:{strDecode}"");
+
+var strNetFrameworkEncode = ""39000c17c7fb12ee58f20239197605d4"";
+Console.WriteLine($""Encode From NET Framework:{strNetFrameworkEncode}"");
+Console.WriteLine($""Decode From NET Framework:{AES.GetDecode(strNetFrameworkEncode, key)}"");
+Console.WriteLine(""=======================NET7"");
+
+//OutPut:
+//Encode: 39000c17c7fb12ee58f20239197606
+//Decode:test123
+//Encode From NET Framework:39000c17c7fb12ee58f20239197605d4
+//Decode From NET Framework:test123
+```
+Here is Console For NET Framework 4.6.1
+```
+
+internal class Program
+ {
+ private static void Main(string[] args)
+ {
+ Console.WriteLine(""=======================NET FRAMEWORK"");
+ string key = ""D36C2FFC94649A3EB946152985CC0B7317F89077FD4F38C1"";
+ var strSource = ""test123"";
+
+ var strEncode = AES.GetEncode(strSource, key);
+
+ var strDecode = AES.GetDecode(strEncode, key);
+
+ var strNet7Encode = ""39000c17c7fb12ee58f20239197606"";
+
+ Console.WriteLine($""Encode:{strEncode}"");
+ Console.WriteLine($""Decode:{strDecode}"");
+ Console.WriteLine($""Encode From NET 7:{strNet7Encode}"");
+ Console.WriteLine($""Decode From NET 7:{AES.GetDecode(strNet7Encode, key)}"");
+ Console.WriteLine(""=======================NET FRAMEWORK"");
+ //OutPut:
+ //Encode:39000c17c7fb12ee58f20239197605d4
+ //Decode:test123
+ //Encode From NET 7:39000c17c7fb12ee58f20239197606
+ //Decode From NET 7:
+ }
+ }
+```
+",0,aescryptoserviceprovider when using aescryptoserviceprovider for encryption i found that when setting the mode property to ciphermode cfb the public properties were modified but the protected properties were not when viewing the variables in the core debugging mode this raises concerns about potential errors in this encryption algorithm although i am aware that aescryptoserviceprovider is obsolete in net core my project needs to be compatible for some time and the encrypted data generated in net core needs to be decrypted in net framework so i need to address this issue i hope that microsoft can help fix this problem and ensure that aescryptoserviceprovider works properly on all platforms this is essential to ensure the security and reliability of programs using this encryption algorithm public class aes public static string convertbytearraytohexstring byte input bool flag stringbuilder stringbuilder new stringbuilder byte bs input int i do byte b bs stringbuilder appendformat b i flag i int bs length while flag return stringbuilder tostring public static byte converthexstringtobytearray string hexstring bool flag hexstring length if flag throw new argumentoutofrangeexception hexstring hexstring hexstring must contain an even number of characters byte new byte int i do byte parse hexstring substring i numberstyles hexnumber i flag i hexstring length while flag return public static string getencode string source string key char c key tochararray byte keys encoding getbytes key substring byte iv encoding getbytes key substring aescryptoserviceprovider aes new aescryptoserviceprovider aes mode ciphermode cfb aes padding paddingmode aes blocksize aes keysize aes key keys aes iv iv byte ciphertext null byte rawplaintext encoding unicode getbytes source byte null using memorystream ms new memorystream using cryptostream cs new cryptostream ms aes createencryptor cryptostreammode write cs write rawplaintext rawplaintext length ciphertext ms toarray string s convertbytearraytohexstring ciphertext return s public static string getdecode string source string key byte keys encoding getbytes key substring byte iv encoding getbytes key substring aescryptoserviceprovider aes new aescryptoserviceprovider aes blocksize aes keysize aes key keys aes iv iv aes mode ciphermode cfb aes padding paddingmode byte ciphertext converthexstringtobytearray source byte plaintext null using memorystream ms new memorystream try using cryptostream cs new cryptostream ms aes createdecryptor cryptostreammode write cs write ciphertext ciphertext length catch 無法解密 return null plaintext ms toarray string s encoding unicode getstring plaintext return s here is console for console writeline string key var strsource var strencode aes getencode strsource key console writeline encode strencode var strdecode aes getdecode strencode key console writeline decode strdecode var strnetframeworkencode console writeline encode from net framework strnetframeworkencode console writeline decode from net framework aes getdecode strnetframeworkencode key console writeline output encode decode encode from net framework decode from net framework here is console for net framework internal class program private static void main string args console writeline net framework string key var strsource var strencode aes getencode strsource key var strdecode aes getdecode strencode key var console writeline encode strencode console writeline decode strdecode console writeline encode from net console writeline decode from net aes getdecode key console writeline net framework output encode decode encode from net decode from net ,0
+2835,8378302109.0,IssuesEvent,2018-10-06 12:49:47,ryota-murakami/blog,https://api.github.com/repos/ryota-murakami/blog,closed,JSビルドスタックのマイグレーション,current-scope🔎 re-architecture🚀,"やりたいこと
+- [ ] coffeeやめてes2018で書きたい
+- [ ] JSモジュールをruby gemで管理するのをやめ、npm管理へ移行したい
+- [ ] turbolinksは残した
+
+webpack使いたい訳ではないけどAssetspipelineで要望が実現できなければ検討してみる。",1.0,"JSビルドスタックのマイグレーション - やりたいこと
+- [ ] coffeeやめてes2018で書きたい
+- [ ] JSモジュールをruby gemで管理するのをやめ、npm管理へ移行したい
+- [ ] turbolinksは残した
+
+webpack使いたい訳ではないけどAssetspipelineで要望が実現できなければ検討してみる。",0,jsビルドスタックのマイグレーション やりたいこと jsモジュールをruby gemで管理するのをやめ、npm管理へ移行したい turbolinksは残した webpack使いたい訳ではないけどassetspipelineで要望が実現できなければ検討してみる。,0
+270514,8461326236.0,IssuesEvent,2018-10-22 21:28:00,nprapps/elections18-general,https://api.github.com/repos/nprapps/elections18-general,closed,"Turn off elections staging daemon, to conserve API key usage",effort:light priority:high,No need until soon before election night,1.0,"Turn off elections staging daemon, to conserve API key usage - No need until soon before election night",0,turn off elections staging daemon to conserve api key usage no need until soon before election night,0
+62466,26006260390.0,IssuesEvent,2022-12-20 19:41:41,cityofaustin/atd-data-tech,https://api.github.com/repos/cityofaustin/atd-data-tech,closed,[TDS-Determination] - Enhancement to VOID Determinations,Service: Apps Need: 1-Must Have Type: Enhancement Workgroup: TDS Product: TDS Portal,"### Requirements:
+
+- [x] Add new determination decision Void
+- [x] When saved, update the determination status to Void.
+
+
+
+----
+
+
+
+> What application are you using?
+
+Transportation Development Services (TDS)
+
+> Describe the problem.
+
+Multiple determination worksheet submissions submitted for the same case (address)
+
+> Describe the outcome you'd like to see when this feature is implemented.
+
+Enhancement that would allow me to mark a determination worksheet as duplicate/cancel the review
+
+> Describe any workarounds you currently have in place or alternative solutions you've considered.
+
+Waive TIA and add a note (not most appropriate solution)
+
+> Requested By
+Joan M.
+
+Request ID: DTS22-105630
+
+",1.0,"[TDS-Determination] - Enhancement to VOID Determinations - ### Requirements:
+
+- [x] Add new determination decision Void
+- [x] When saved, update the determination status to Void.
+
+
+
+----
+
+
+
+> What application are you using?
+
+Transportation Development Services (TDS)
+
+> Describe the problem.
+
+Multiple determination worksheet submissions submitted for the same case (address)
+
+> Describe the outcome you'd like to see when this feature is implemented.
+
+Enhancement that would allow me to mark a determination worksheet as duplicate/cancel the review
+
+> Describe any workarounds you currently have in place or alternative solutions you've considered.
+
+Waive TIA and add a note (not most appropriate solution)
+
+> Requested By
+Joan M.
+
+Request ID: DTS22-105630
+
+",0, enhancement to void determinations requirements add new determination decision void when saved update the determination status to void what application are you using transportation development services tds describe the problem multiple determination worksheet submissions submitted for the same case address describe the outcome you d like to see when this feature is implemented enhancement that would allow me to mark a determination worksheet as duplicate cancel the review describe any workarounds you currently have in place or alternative solutions you ve considered waive tia and add a note not most appropriate solution requested by joan m request id ,0
+723,13227582452.0,IssuesEvent,2020-08-18 03:37:51,hackforla/civic-opportunity,https://api.github.com/repos/hackforla/civic-opportunity,opened,Write Copy for the Website,people: 1 role: content strategist/writer role: product management,"### Overview
+Develop Copy for the CoP website.
+
+### Action Items
+Content Strategist/Writer
+Refer to One-Sheet for Project Mission and Purpose
+Develop 1st draft of website Copy
+Review with Bonnie and Product Team
+Finalize Copy
+Work with Design Team to integrate into web design
+
+### Resources/Instructions
+
+",1.0,"Write Copy for the Website - ### Overview
+Develop Copy for the CoP website.
+
+### Action Items
+Content Strategist/Writer
+Refer to One-Sheet for Project Mission and Purpose
+Develop 1st draft of website Copy
+Review with Bonnie and Product Team
+Finalize Copy
+Work with Design Team to integrate into web design
+
+### Resources/Instructions
+
+",1,write copy for the website overview develop copy for the cop website action items content strategist writer refer to one sheet for project mission and purpose develop draft of website copy review with bonnie and product team finalize copy work with design team to integrate into web design resources instructions ,1
+104828,4225890425.0,IssuesEvent,2016-07-02 03:52:05,coreos/bugs,https://api.github.com/repos/coreos/bugs,closed,early-docker.service is broken with Docker 1.11.2,area/usability component/docker kind/regression priority/P0 team/os,"I don't have any details on this, but I was told this fails. We'll need to fix this before we ship Docker 1.11.2.",1.0,"early-docker.service is broken with Docker 1.11.2 - I don't have any details on this, but I was told this fails. We'll need to fix this before we ship Docker 1.11.2.",0,early docker service is broken with docker i don t have any details on this but i was told this fails we ll need to fix this before we ship docker ,0
+510,8789663100.0,IssuesEvent,2018-12-21 05:10:32,JohnnySn0w/MoD,https://api.github.com/repos/JohnnySn0w/MoD,closed,Figure out how to get the discord bot running,people and learning,We need to find out how discord bots are hosted and actually run,1.0,Figure out how to get the discord bot running - We need to find out how discord bots are hosted and actually run,1,figure out how to get the discord bot running we need to find out how discord bots are hosted and actually run,1
+22,2649890762.0,IssuesEvent,2015-03-15 11:56:58,fskreuz/Tree-Trail,https://api.github.com/repos/fskreuz/Tree-Trail,closed,Guest Hopepage,QA people! Test these!,The guest user are redirected to a homepage where logout button and administrator dashboard are present,1.0,Guest Hopepage - The guest user are redirected to a homepage where logout button and administrator dashboard are present,1,guest hopepage the guest user are redirected to a homepage where logout button and administrator dashboard are present,1
+684,12266203557.0,IssuesEvent,2020-05-07 08:33:31,bitprj/bitproject,https://api.github.com/repos/bitprj/bitproject,opened,Add Delegation Guidelines for DevRel,people,"**Objectives**
+People don't know how to delegate tasks when given OKRs. We need to come up with guideliens and training on how to do this in the manager training.
+
+**Collaborators and Tasks**
+@beccatran
+
+**Deadline**
+May 22, 2020",1.0,"Add Delegation Guidelines for DevRel - **Objectives**
+People don't know how to delegate tasks when given OKRs. We need to come up with guideliens and training on how to do this in the manager training.
+
+**Collaborators and Tasks**
+@beccatran
+
+**Deadline**
+May 22, 2020",1,add delegation guidelines for devrel objectives people don t know how to delegate tasks when given okrs we need to come up with guideliens and training on how to do this in the manager training collaborators and tasks beccatran deadline may ,1
+258921,22358557262.0,IssuesEvent,2022-06-15 18:00:25,elastic/kibana,https://api.github.com/repos/elastic/kibana,closed,Failing test: X-Pack Alerting API Integration Tests.x-pack/test/alerting_api_integration/spaces_only/tests/alerting/bulk_edit·ts - alerting api integration spaces only Alerting bulkEdit should return mapped params after bulk edit,blocker failed-test skipped-test Team:ResponseOps v8.3.0,"A test failed on a tracked branch
+
+```
+Error: expected [ { message: 'Saved object [alert/3920aa90-d2d2-11ec-b6f5-efaf8297ed5e] conflict',
+ rule:
+ { id: '3920aa90-d2d2-11ec-b6f5-efaf8297ed5e',
+ name: 'abc' } } ] to have a length of 0 but got 1
+ at Assertion.assert (node_modules/@kbn/expect/expect.js:100:11)
+ at Assertion.length (node_modules/@kbn/expect/expect.js:374:8)
+ at Context. (x-pack/test/alerting_api_integration/spaces_only/tests/alerting/bulk_edit.ts:173:52)
+ at runMicrotasks ()
+ at processTicksAndRejections (node:internal/process/task_queues:96:5)
+ at Object.apply (node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16)
+```
+
+First failure: [CI Build - main](https://buildkite.com/elastic/kibana-on-merge/builds/15986#b065ffbd-c94f-4c4a-be84-c52b64578bf7)
+
+",2.0,"Failing test: X-Pack Alerting API Integration Tests.x-pack/test/alerting_api_integration/spaces_only/tests/alerting/bulk_edit·ts - alerting api integration spaces only Alerting bulkEdit should return mapped params after bulk edit - A test failed on a tracked branch
+
+```
+Error: expected [ { message: 'Saved object [alert/3920aa90-d2d2-11ec-b6f5-efaf8297ed5e] conflict',
+ rule:
+ { id: '3920aa90-d2d2-11ec-b6f5-efaf8297ed5e',
+ name: 'abc' } } ] to have a length of 0 but got 1
+ at Assertion.assert (node_modules/@kbn/expect/expect.js:100:11)
+ at Assertion.length (node_modules/@kbn/expect/expect.js:374:8)
+ at Context. (x-pack/test/alerting_api_integration/spaces_only/tests/alerting/bulk_edit.ts:173:52)
+ at runMicrotasks ()
+ at processTicksAndRejections (node:internal/process/task_queues:96:5)
+ at Object.apply (node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16)
+```
+
+First failure: [CI Build - main](https://buildkite.com/elastic/kibana-on-merge/builds/15986#b065ffbd-c94f-4c4a-be84-c52b64578bf7)
+
+",0,failing test x pack alerting api integration tests x pack test alerting api integration spaces only tests alerting bulk edit·ts alerting api integration spaces only alerting bulkedit should return mapped params after bulk edit a test failed on a tracked branch error expected conflict rule id name abc to have a length of but got at assertion assert node modules kbn expect expect js at assertion length node modules kbn expect expect js at context x pack test alerting api integration spaces only tests alerting bulk edit ts at runmicrotasks at processticksandrejections node internal process task queues at object apply node modules kbn test target node functional test runner lib mocha wrap function js first failure ,0
+172097,14350209572.0,IssuesEvent,2020-11-29 19:53:15,UB-ES-2020-A/Alejandria,https://api.github.com/repos/UB-ES-2020-A/Alejandria,closed,[USLOG6] Page Library User,6h documentation enhancement frontend,"The **User Library Page**, containing all the information requested in the Acceptance Criteria in the US.
+Shows all the information passed by the backend.
+And in general terms, it represents a list of items, **with pagination if possible** and if necessary.
+The list is vertical and at the left side of the item contains the information related to the product , and every item is a button that redirects you to its ""book page"".",1.0,"[USLOG6] Page Library User - The **User Library Page**, containing all the information requested in the Acceptance Criteria in the US.
+Shows all the information passed by the backend.
+And in general terms, it represents a list of items, **with pagination if possible** and if necessary.
+The list is vertical and at the left side of the item contains the information related to the product , and every item is a button that redirects you to its ""book page"".",0, page library user the user library page containing all the information requested in the acceptance criteria in the us shows all the information passed by the backend and in general terms it represents a list of items with pagination if possible and if necessary the list is vertical and at the left side of the item contains the information related to the product and every item is a button that redirects you to its book page ,0
+926,19843346041.0,IssuesEvent,2022-01-21 01:20:57,planningcenter/developers,https://api.github.com/repos/planningcenter/developers,closed,Update a Custom Field associated with a person's profile,People,"**Related Product**
+Which product is this question related to?
+
+People
+
+
+
+**Describe the question**
+I am trying to update a custom field on a person's profile. Where I am getting stuck at is I see how to access the custom field, but I don't see how to update the custom field on a specific persons profile. I am currently using the following URL to access the field_data, but I don't see how to update the field. The field I am trying to ""uncheck"" is the ""Financial/Bills"" (bolded below). I see that it tells me the custom field id under FieldDefinition. Can you help me understand how to uncheck the ""Financial/Bills"" checkbox field on the profile?
+
+URL: api.planningcenteronline.com/people/v2/people//field_data
+
+Results:
+{'links': {'self': 'https://api.planningcenteronline.com/people/v2/people/102421804/field_data'}, 'data': [{'type': 'FieldDatum', 'id': '123979937', 'attributes': {'file': {'url': None}, 'file_content_type': None, 'file_name': None
+, 'file_size': None, 'value': **'Financial/Bills**'}, 'relationships': {'field_definition': {'data': {'type': 'FieldDefinition', 'id': '**516612**'}}, 'customizable': {'data': {'type': 'Person', 'id': '102421804'}}}, 'links': {'self': 'htt
+ps://api.planningcenteronline.com/people/v2/field_data/123979937'}}], 'included': [], 'meta': {'total_count': 1, 'count': 1, 'can_order_by': ['value', 'file', 'file_size', 'file_content_type', 'file_name'], 'can_query_by': ['value'
+, 'file', 'file_size', 'file_content_type', 'file_name', 'field_definition_id'], 'can_include': ['field_definition', 'field_option', 'tab'], 'parent': {'id': '102421804', 'type': 'Person'}}}
+
+
+**What have you tried that worked?**
+
+
+**What have you tried that didn't work?**
+
+
+**Additional context**
+
+
+
+## I have..
+
+- [ ] Reviewed the documentation found at https://developer.planning.center/docs
+- [ ] Searched for previous issues answering this question
+- [ ] Removed all private information from this issue (credentials, tokens, emails, phone numbers, etc.)
+- [x] Reviewed my issue for completeness
+",1.0,"Update a Custom Field associated with a person's profile - **Related Product**
+Which product is this question related to?
+
+People
+
+
+
+**Describe the question**
+I am trying to update a custom field on a person's profile. Where I am getting stuck at is I see how to access the custom field, but I don't see how to update the custom field on a specific persons profile. I am currently using the following URL to access the field_data, but I don't see how to update the field. The field I am trying to ""uncheck"" is the ""Financial/Bills"" (bolded below). I see that it tells me the custom field id under FieldDefinition. Can you help me understand how to uncheck the ""Financial/Bills"" checkbox field on the profile?
+
+URL: api.planningcenteronline.com/people/v2/people//field_data
+
+Results:
+{'links': {'self': 'https://api.planningcenteronline.com/people/v2/people/102421804/field_data'}, 'data': [{'type': 'FieldDatum', 'id': '123979937', 'attributes': {'file': {'url': None}, 'file_content_type': None, 'file_name': None
+, 'file_size': None, 'value': **'Financial/Bills**'}, 'relationships': {'field_definition': {'data': {'type': 'FieldDefinition', 'id': '**516612**'}}, 'customizable': {'data': {'type': 'Person', 'id': '102421804'}}}, 'links': {'self': 'htt
+ps://api.planningcenteronline.com/people/v2/field_data/123979937'}}], 'included': [], 'meta': {'total_count': 1, 'count': 1, 'can_order_by': ['value', 'file', 'file_size', 'file_content_type', 'file_name'], 'can_query_by': ['value'
+, 'file', 'file_size', 'file_content_type', 'file_name', 'field_definition_id'], 'can_include': ['field_definition', 'field_option', 'tab'], 'parent': {'id': '102421804', 'type': 'Person'}}}
+
+
+**What have you tried that worked?**
+
+
+**What have you tried that didn't work?**
+
+
+**Additional context**
+
+
+
+## I have..
+
+- [ ] Reviewed the documentation found at https://developer.planning.center/docs
+- [ ] Searched for previous issues answering this question
+- [ ] Removed all private information from this issue (credentials, tokens, emails, phone numbers, etc.)
+- [x] Reviewed my issue for completeness
+",1,update a custom field associated with a person s profile related product which product is this question related to people describe the question i am trying to update a custom field on a person s profile where i am getting stuck at is i see how to access the custom field but i don t see how to update the custom field on a specific persons profile i am currently using the following url to access the field data but i don t see how to update the field the field i am trying to uncheck is the financial bills bolded below i see that it tells me the custom field id under fielddefinition can you help me understand how to uncheck the financial bills checkbox field on the profile url api planningcenteronline com people people field data results links self data type fielddatum id attributes file url none file content type none file name none file size none value financial bills relationships field definition data type fielddefinition id customizable data type person id links self htt ps api planningcenteronline com people field data included meta total count count can order by can query by value file file size file content type file name field definition id can include parent id type person what have you tried that worked what have you tried that didn t work additional context i have reviewed the documentation found at searched for previous issues answering this question removed all private information from this issue credentials tokens emails phone numbers etc reviewed my issue for completeness ,1
+444619,12815055018.0,IssuesEvent,2020-07-04 23:06:57,Atlantiss/NetherwingBugtracker,https://api.github.com/repos/Atlantiss/NetherwingBugtracker,closed,[Core][Pet] Pet follow speed mimicking the master's speed,- Core Mechanic Hotfix Issue - Priority Pets Status: Confirmed,"Pets on this server will always set their speed to their master's speed regardless of the situation. This is a huge issue in pvp.
+
+Take this video from retail tbc for an example:
+https://youtu.be/4jbSgaPROH0?t=128 - Hunter is hamstrung, sends his pet on the mage and then calls it back, you can see the pet running back at normal speed while the hunter is slowed.
+
+Now how it works on netherwing:
+https://www.twitch.tv/videos/360748087 - notice that the follow speed is reduced however the pursuit speed remain unchanged.
+
+Might be related to issue #2310",1.0,"[Core][Pet] Pet follow speed mimicking the master's speed - Pets on this server will always set their speed to their master's speed regardless of the situation. This is a huge issue in pvp.
+
+Take this video from retail tbc for an example:
+https://youtu.be/4jbSgaPROH0?t=128 - Hunter is hamstrung, sends his pet on the mage and then calls it back, you can see the pet running back at normal speed while the hunter is slowed.
+
+Now how it works on netherwing:
+https://www.twitch.tv/videos/360748087 - notice that the follow speed is reduced however the pursuit speed remain unchanged.
+
+Might be related to issue #2310",0, pet follow speed mimicking the master s speed pets on this server will always set their speed to their master s speed regardless of the situation this is a huge issue in pvp take this video from retail tbc for an example hunter is hamstrung sends his pet on the mage and then calls it back you can see the pet running back at normal speed while the hunter is slowed now how it works on netherwing notice that the follow speed is reduced however the pursuit speed remain unchanged might be related to issue ,0
+473982,13650176964.0,IssuesEvent,2020-09-26 17:51:55,HaxeFoundation/intellij-haxe,https://api.github.com/repos/HaxeFoundation/intellij-haxe,closed,"1.0 unable to compile openfl because of the compilation parameter lime ""run""",Priority 3 Works For Me bug,"1.0 unable to compile openfl because of the compilation parameter lime ""run""",1.0,"1.0 unable to compile openfl because of the compilation parameter lime ""run"" - 1.0 unable to compile openfl because of the compilation parameter lime ""run""",0, unable to compile openfl because of the compilation parameter lime run unable to compile openfl because of the compilation parameter lime run ,0
+1097,26813508288.0,IssuesEvent,2023-02-02 01:11:44,openstates/issues,https://api.github.com/repos/openstates/issues,closed,New CT Committee Scraper,good first issue component:people-data good first scraper,"*Special Note: all committees of the Connecticut General Assembly are joint committees with House and Senate co-chairs and rank-and-file members.*
+
+### We need a new committee scraper written in [spatula](https://jamesturk.github.io/spatula/) for Connecticut.
+
+It should scrape this [webpage of committees](https://www.cga.ct.gov/asp/menu/cgacommittees.asp) to get:
+- **name**
+ - “Small Business”, “Agriculture”, etc.
+- **chamber**
+ - “upper”, “lower”, or “legislature” (when joint, etc.)
+- **classification**
+ - ex: ""committee"", ""subcommittee""
+- **parent**
+ - *only if it is a subcommittee* - scraper should have a way to determine this
+ - ex: parent=""Natural Resources"" when *sub*committee name=""Forestry""
+- **sources**
+ - each added using `add_source()` method on instance of `ScrapeCommittee` type object
+ - ex: home page for list of committees, specific page for that committee, etc.
+- **members**
+ - - each added using `add_member()`method on instance of `ScrapeCommittee` type object
+ - *Attributes:*
+ - **name**: “Jane Doe”, “John Smith”, etc.
+ - **role** (where applicable): “Chair”, “Ranking Member”, etc.
+
+#### [Here is a helpful general reference doc](https://docs.openstates.org/contributing/writing-a-committee-scraper/) for those new to writing scrapers, the Open States project, or the specific task of writing a committee scraper
+
+### Useful scrapers for reference
+A [spatula-based scraper](https://github.com/openstates/openstates-scrapers/blob/main/scrapers_next/mo/committees.py) that also scrapes an HTML List Page (using `HtmlListPage` and `HtmlPage` spatula classes) for reference.
+- As in the other spatula scraper, you will need to set the session variable separately (rather than depend on the `__init__.py`).
+
+### Other useful resources
+You can reference the [object models for committee scrapers](https://github.com/openstates/openstates-core/blob/main/openstates/models/committees.py) in the Open States Core repo.
+
+Further documentation on running spatula scrapers in the command line can be found [here](https://docs.openstates.org/contributing/scrapers/#running-spatula-scrapers).",1.0,"New CT Committee Scraper - *Special Note: all committees of the Connecticut General Assembly are joint committees with House and Senate co-chairs and rank-and-file members.*
+
+### We need a new committee scraper written in [spatula](https://jamesturk.github.io/spatula/) for Connecticut.
+
+It should scrape this [webpage of committees](https://www.cga.ct.gov/asp/menu/cgacommittees.asp) to get:
+- **name**
+ - “Small Business”, “Agriculture”, etc.
+- **chamber**
+ - “upper”, “lower”, or “legislature” (when joint, etc.)
+- **classification**
+ - ex: ""committee"", ""subcommittee""
+- **parent**
+ - *only if it is a subcommittee* - scraper should have a way to determine this
+ - ex: parent=""Natural Resources"" when *sub*committee name=""Forestry""
+- **sources**
+ - each added using `add_source()` method on instance of `ScrapeCommittee` type object
+ - ex: home page for list of committees, specific page for that committee, etc.
+- **members**
+ - - each added using `add_member()`method on instance of `ScrapeCommittee` type object
+ - *Attributes:*
+ - **name**: “Jane Doe”, “John Smith”, etc.
+ - **role** (where applicable): “Chair”, “Ranking Member”, etc.
+
+#### [Here is a helpful general reference doc](https://docs.openstates.org/contributing/writing-a-committee-scraper/) for those new to writing scrapers, the Open States project, or the specific task of writing a committee scraper
+
+### Useful scrapers for reference
+A [spatula-based scraper](https://github.com/openstates/openstates-scrapers/blob/main/scrapers_next/mo/committees.py) that also scrapes an HTML List Page (using `HtmlListPage` and `HtmlPage` spatula classes) for reference.
+- As in the other spatula scraper, you will need to set the session variable separately (rather than depend on the `__init__.py`).
+
+### Other useful resources
+You can reference the [object models for committee scrapers](https://github.com/openstates/openstates-core/blob/main/openstates/models/committees.py) in the Open States Core repo.
+
+Further documentation on running spatula scrapers in the command line can be found [here](https://docs.openstates.org/contributing/scrapers/#running-spatula-scrapers).",1,new ct committee scraper special note all committees of the connecticut general assembly are joint committees with house and senate co chairs and rank and file members we need a new committee scraper written in for connecticut it should scrape this to get name “small business” “agriculture” etc chamber “upper” “lower” or “legislature” when joint etc classification ex committee subcommittee parent only if it is a subcommittee scraper should have a way to determine this ex parent natural resources when sub committee name forestry sources each added using add source method on instance of scrapecommittee type object ex home page for list of committees specific page for that committee etc members each added using add member method on instance of scrapecommittee type object attributes name “jane doe” “john smith” etc role where applicable “chair” “ranking member” etc for those new to writing scrapers the open states project or the specific task of writing a committee scraper useful scrapers for reference a that also scrapes an html list page using htmllistpage and htmlpage spatula classes for reference as in the other spatula scraper you will need to set the session variable separately rather than depend on the init py other useful resources you can reference the in the open states core repo further documentation on running spatula scrapers in the command line can be found ,1
+1142,28967298776.0,IssuesEvent,2023-05-10 08:47:48,Swiss-Polar-Institute/project-application,https://api.github.com/repos/Swiss-Polar-Institute/project-application,closed,group dependent on person position rather than organisation,people,"A person (postition) can add a number of organisations to which they belong, but only one group / lab. It is possible that they belong to a group / lab at each organisation, therefore group / lab would need to be part of the person position organisation, rather than the person position.",1.0,"group dependent on person position rather than organisation - A person (postition) can add a number of organisations to which they belong, but only one group / lab. It is possible that they belong to a group / lab at each organisation, therefore group / lab would need to be part of the person position organisation, rather than the person position.",1,group dependent on person position rather than organisation a person postition can add a number of organisations to which they belong but only one group lab it is possible that they belong to a group lab at each organisation therefore group lab would need to be part of the person position organisation rather than the person position ,1
+248770,7935700246.0,IssuesEvent,2018-07-09 06:45:45,openshift/origin,https://api.github.com/repos/openshift/origin,closed,HTTP proxy implementation does not escape correctly,component/install lifecycle/rotten priority/P2,"$CUSTOMER has to use an HTTP proxy to access the internet. Authentication requires an Active Directory account in the format: DOMAIN\USER or USER@DOMAIN which means the URL to the proxy results in:
+
+```
+http://DOMAIN\USER:password@proxy.local:8080/
+```
+
+or
+
+```
+http://USER@DOMAIN:password@proxy.local:8080/
+```
+
+Both are problematic for various reasons:
+- Docker and Openshift require environment variables in a shell-script, which means the backslash needs to be escaped ! (Which it does not, so we have to do it ourselves)
+ - openshift_http_proxy='http://DOMAIN\USER:password@proxy.local:8080/'
+ - openshift_https_proxy='http://DOMAIN\USER:password@proxy.local:8080/'
+- Openshift provides the proxy settings for pulling sources and pushing images, and this leads to various problems depending on the tools that require these settings, e.g.
+ - github works fine with the backslash as is (so it should not be escaped)
+ - however for npm there is an issue because the backslash is being translated into %5C, and npm is trying to resolve DOMAIN rather than proxy.local.
+
+So the whole http_proxy implementation is very prone to issues specific to how this specific string is being used by individual tools.
+##### Version
+
+Openshift Container Platform 3.3
+##### Steps To Reproduce
+1. Configure the HTTP_PROXY in your Ansible inventory and use an Active Directory style username
+2. Install/re-configure your setup
+3. Watch it fail to build or push (or worse, fail to see any proper logs)
+ ##### Current Result
+
+Failure
+##### Expected Result
+
+Success
+",1.0,"HTTP proxy implementation does not escape correctly - $CUSTOMER has to use an HTTP proxy to access the internet. Authentication requires an Active Directory account in the format: DOMAIN\USER or USER@DOMAIN which means the URL to the proxy results in:
+
+```
+http://DOMAIN\USER:password@proxy.local:8080/
+```
+
+or
+
+```
+http://USER@DOMAIN:password@proxy.local:8080/
+```
+
+Both are problematic for various reasons:
+- Docker and Openshift require environment variables in a shell-script, which means the backslash needs to be escaped ! (Which it does not, so we have to do it ourselves)
+ - openshift_http_proxy='http://DOMAIN\USER:password@proxy.local:8080/'
+ - openshift_https_proxy='http://DOMAIN\USER:password@proxy.local:8080/'
+- Openshift provides the proxy settings for pulling sources and pushing images, and this leads to various problems depending on the tools that require these settings, e.g.
+ - github works fine with the backslash as is (so it should not be escaped)
+ - however for npm there is an issue because the backslash is being translated into %5C, and npm is trying to resolve DOMAIN rather than proxy.local.
+
+So the whole http_proxy implementation is very prone to issues specific to how this specific string is being used by individual tools.
+##### Version
+
+Openshift Container Platform 3.3
+##### Steps To Reproduce
+1. Configure the HTTP_PROXY in your Ansible inventory and use an Active Directory style username
+2. Install/re-configure your setup
+3. Watch it fail to build or push (or worse, fail to see any proper logs)
+ ##### Current Result
+
+Failure
+##### Expected Result
+
+Success
+",0,http proxy implementation does not escape correctly customer has to use an http proxy to access the internet authentication requires an active directory account in the format domain user or user domain which means the url to the proxy results in or both are problematic for various reasons docker and openshift require environment variables in a shell script which means the backslash needs to be escaped which it does not so we have to do it ourselves openshift http proxy openshift https proxy openshift provides the proxy settings for pulling sources and pushing images and this leads to various problems depending on the tools that require these settings e g github works fine with the backslash as is so it should not be escaped however for npm there is an issue because the backslash is being translated into and npm is trying to resolve domain rather than proxy local so the whole http proxy implementation is very prone to issues specific to how this specific string is being used by individual tools version openshift container platform steps to reproduce configure the http proxy in your ansible inventory and use an active directory style username install re configure your setup watch it fail to build or push or worse fail to see any proper logs current result failure expected result success ,0
+943,20914446212.0,IssuesEvent,2022-03-24 12:13:40,AtB-AS/mittatb-app,https://api.github.com/repos/AtB-AS/mittatb-app,opened,Reisesøk: Nedprioritere/filtrere ut gå-reiser,:people_holding_hands: Teaminitiativ,"## Origin
+
+Dette har blitt diskutert på Slack tidligere, både [internt i teamet](https://mittatb.slack.com/archives/CSFC6PG23/p1648121168815099), men også [sammen med EnTur](https://mittatb.slack.com/archives/CHLDG8C30/p1642511653009200), _""spesielt i kontekst av at det var noen steder hvor det var ulogisk at gåturen kom høyere opp enn en mer logisk bussreise.""_
+
+**EnTur nevnte at:**
+> _..et reiseforslag med kun gange i praksis kan tidforskyves frem og tilbake vil dette rent sorteringsmessig alltid komme først. Dette er pga. at det er vanlig å vise slike ren-gange forslag i en annen visning enn sammen med kollektivtransport alternativer og at det er sånn sett er enkelt å plukke ut av lista._
+
+**Det var et svar på denne meldingen fra Gøran:**
+
+> Vi holder nå på å teste vår oppgradering til JP3, og ser at det noen ganger blir foreslått gå-reise som vi ikke synes er logisk at skal komme så høyt prioritert i returnerte reiseforslag. Eksempel i vedlagt bilde hvor en gå-reise på 28 minutter prioriteres over buss-reise på 10 minutter. At gå-reisen kommer med som reiseforslag er greit nok, men at den blir prioritert over en buss-reise som er raskere og er fremme 16 minutter tidligere synes vi er rart.
+> Også vedlagt bilde av at det er gjenskapt i graphql-explorer.
+> Vi vet vi kan stille på `walkReluctance`, men hadde egentlig forventet at selv med defaults ble ikke gå-reisen prioritert så høyt opp.
+
+
+
+
+
+## Motivation
+
+
+
+## Hypotheses and assumptions
+
+Å få gåturen promotert som førstevalget på et reisesøk oppleves som unødvendig for brukeren. Dette er en tilbakemelding vi har fått høre på intercom tidligere, i tillegg til ""her og der"" i andre kanaler. Nå som vi erstatter reisesøkmotoren med otp2, og det ikke lenger er opt-in vil det sannsynligvis være en tilbakemelding som vil dukke opp oftere.
+
+## Proposed solution
+
+Ved å filtrere ut, eller nedprioritere gåturene, vil vi fjerne støy fra de andre relevante reiseforslagene, og sånn sett senke ""cognitive load"" for brukeren, og gi en mer sømløs brukeropplevelse.
+
+**Eksempler fra ulike løsninger fra Ruter og EnTur:**
+
+
+
+
+
+
+
+## Acceptance Criteria
+
+_List of relevant acceptance criteria as a part of a QA flow_
+
+- [] Renders colors as expected in dark and light mode.
+- [] Is translatable to English and Norwegian.
+- [] Does not drain battery
+- [] Works in up to 200% font size
+- [] Is operable and perceivable using screen reader
+
+Beskrivelse av QA prosessen finnes [her](https://github.com/AtB-AS/org/blob/master/guides/quality_assurance.md#qa-in-atb-mobile-application)
+",1.0,"Reisesøk: Nedprioritere/filtrere ut gå-reiser - ## Origin
+
+Dette har blitt diskutert på Slack tidligere, både [internt i teamet](https://mittatb.slack.com/archives/CSFC6PG23/p1648121168815099), men også [sammen med EnTur](https://mittatb.slack.com/archives/CHLDG8C30/p1642511653009200), _""spesielt i kontekst av at det var noen steder hvor det var ulogisk at g��turen kom høyere opp enn en mer logisk bussreise.""_
+
+**EnTur nevnte at:**
+> _..et reiseforslag med kun gange i praksis kan tidforskyves frem og tilbake vil dette rent sorteringsmessig alltid komme først. Dette er pga. at det er vanlig å vise slike ren-gange forslag i en annen visning enn sammen med kollektivtransport alternativer og at det er sånn sett er enkelt å plukke ut av lista._
+
+**Det var et svar på denne meldingen fra Gøran:**
+
+> Vi holder nå på å teste vår oppgradering til JP3, og ser at det noen ganger blir foreslått gå-reise som vi ikke synes er logisk at skal komme så høyt prioritert i returnerte reiseforslag. Eksempel i vedlagt bilde hvor en gå-reise på 28 minutter prioriteres over buss-reise på 10 minutter. At gå-reisen kommer med som reiseforslag er greit nok, men at den blir prioritert over en buss-reise som er raskere og er fremme 16 minutter tidligere synes vi er rart.
+> Også vedlagt bilde av at det er gjenskapt i graphql-explorer.
+> Vi vet vi kan stille på `walkReluctance`, men hadde egentlig forventet at selv med defaults ble ikke gå-reisen prioritert så høyt opp.
+
+
+
+
+
+## Motivation
+
+
+
+## Hypotheses and assumptions
+
+Å få gåturen promotert som førstevalget på et reisesøk oppleves som unødvendig for brukeren. Dette er en tilbakemelding vi har fått høre på intercom tidligere, i tillegg til ""her og der"" i andre kanaler. Nå som vi erstatter reisesøkmotoren med otp2, og det ikke lenger er opt-in vil det sannsynligvis være en tilbakemelding som vil dukke opp oftere.
+
+## Proposed solution
+
+Ved å filtrere ut, eller nedprioritere gåturene, vil vi fjerne støy fra de andre relevante reiseforslagene, og sånn sett senke ""cognitive load"" for brukeren, og gi en mer sømløs brukeropplevelse.
+
+**Eksempler fra ulike løsninger fra Ruter og EnTur:**
+
+
+
+
+
+
+
+## Acceptance Criteria
+
+_List of relevant acceptance criteria as a part of a QA flow_
+
+- [] Renders colors as expected in dark and light mode.
+- [] Is translatable to English and Norwegian.
+- [] Does not drain battery
+- [] Works in up to 200% font size
+- [] Is operable and perceivable using screen reader
+
+Beskrivelse av QA prosessen finnes [her](https://github.com/AtB-AS/org/blob/master/guides/quality_assurance.md#qa-in-atb-mobile-application)
+",1,reisesøk nedprioritere filtrere ut gå reiser origin dette har blitt diskutert på slack tidligere både men også spesielt i kontekst av at det var noen steder hvor det var ulogisk at gåturen kom høyere opp enn en mer logisk bussreise entur nevnte at et reiseforslag med kun gange i praksis kan tidforskyves frem og tilbake vil dette rent sorteringsmessig alltid komme først dette er pga at det er vanlig å vise slike ren gange forslag i en annen visning enn sammen med kollektivtransport alternativer og at det er sånn sett er enkelt å plukke ut av lista det var et svar på denne meldingen fra gøran vi holder nå på å teste vår oppgradering til og ser at det noen ganger blir foreslått gå reise som vi ikke synes er logisk at skal komme så høyt prioritert i returnerte reiseforslag eksempel i vedlagt bilde hvor en gå reise på minutter prioriteres over buss reise på minutter at gå reisen kommer med som reiseforslag er greit nok men at den blir prioritert over en buss reise som er raskere og er fremme minutter tidligere synes vi er rart også vedlagt bilde av at det er gjenskapt i graphql explorer vi vet vi kan stille på walkreluctance men hadde egentlig forventet at selv med defaults ble ikke gå reisen prioritert så høyt opp img width alt screenshot at src motivation hypotheses and assumptions å få gåturen promotert som førstevalget på et reisesøk oppleves som unødvendig for brukeren dette er en tilbakemelding vi har fått høre på intercom tidligere i tillegg til her og der i andre kanaler nå som vi erstatter reisesøkmotoren med og det ikke lenger er opt in vil det sannsynligvis være en tilbakemelding som vil dukke opp oftere proposed solution ved å filtrere ut eller nedprioritere gåturene vil vi fjerne støy fra de andre relevante reiseforslagene og sånn sett senke cognitive load for brukeren og gi en mer sømløs brukeropplevelse eksempler fra ulike løsninger fra ruter og entur acceptance criteria list of relevant acceptance criteria as a part of a qa flow renders colors as expected in dark and light mode is translatable to english and norwegian does not drain battery works in up to font size is operable and perceivable using screen reader beskrivelse av qa prosessen finnes ,1
+206087,15708146328.0,IssuesEvent,2021-03-26 20:02:26,xbmc/kodi-tv,https://api.github.com/repos/xbmc/kodi-tv,closed,Need a Live Preview of Not Yet Published Blog Posts,backend needs-testing,"Right now the only way to see a blog post with all the site CSS applied is to publish it. It would be nice to have a way to see a post in full context before publishing.
+
+There is a way to do that with Netlify, so we need to figure that out.",1.0,"Need a Live Preview of Not Yet Published Blog Posts - Right now the only way to see a blog post with all the site CSS applied is to publish it. It would be nice to have a way to see a post in full context before publishing.
+
+There is a way to do that with Netlify, so we need to figure that out.",0,need a live preview of not yet published blog posts right now the only way to see a blog post with all the site css applied is to publish it it would be nice to have a way to see a post in full context before publishing there is a way to do that with netlify so we need to figure that out ,0
+103,3432005624.0,IssuesEvent,2015-12-11 01:34:02,Princeton-CDH/mapping-expatriate-paris,https://api.github.com/repos/Princeton-CDH/mapping-expatriate-paris,opened,"""Vieizzot"" is misspelling of ""Vieillot""",not in personography two people merged in one card folder,"The cards under ""Vieizzot"" are actually deposit cards for Mlle Vieillot, and should be filed under her name. The person ""Vieizzot"" in the personography should be deleted.",1.0,"""Vieizzot"" is misspelling of ""Vieillot"" - The cards under ""Vieizzot"" are actually deposit cards for Mlle Vieillot, and should be filed under her name. The person ""Vieizzot"" in the personography should be deleted.",1, vieizzot is misspelling of vieillot the cards under vieizzot are actually deposit cards for mlle vieillot and should be filed under her name the person vieizzot in the personography should be deleted ,1
+174248,27606003106.0,IssuesEvent,2023-03-09 13:02:56,Kotlin/kotlinx.coroutines,https://api.github.com/repos/Kotlin/kotlinx.coroutines,closed,Consider deprecating or changing the behaviour of CoroutineContext.isActive,enhancement design breaking change for 1.7,"According to the [doc](https://kotlin.github.io/kotlinx.coroutines/kotlinx-coroutines-core/kotlinx.coroutines/is-active.html), `isActive` has the following property:
+
+>The coroutineContext.isActive expression is a shortcut for coroutineContext[Job]?.isActive == true. See [Job.isActive](https://kotlin.github.io/kotlinx.coroutines/kotlinx-coroutines-core/kotlinx.coroutines/-job/is-active.html).
+
+It means that, if the `Job` is not present, `isActive` always returns `false`.
+
+We have multiple reports that such behaviour can be error-prone when used with non-`kotlinx.coroutines` entry points, such as Ktor and `suspend fun main`, because it is inconsistent with the overall contract:
+
+>(Job) has not been completed and was not cancelled yet
+
+
+`CoroutineContext.isActive` predates both `CoroutineScope` (which should always have a `Job` in it, if it's not `GlobalScope`) and `job` extension, so it may be the case that it can be safely deprecated.
+
+Basically, we have three options:
+
+* Do nothing, left things as is. It doesn't solve the original issue, but also doesn't introduce any potentially breaking changes
+* Deprecate `CoroutineContext.isActive`. Such change has multiple potential downsides
+ * Its only possible replacement is `this.job.isActive`, but this replacement is not equivalent to the original method -- `.job` throws an exception for contexts without a `Job`. An absence of replacement can be too disturbing as [a lot of code](https://grep.app/search?q=context.isActive&filter[lang][0]=Kotlin) rely on a perfectly fine `ctxWithJob.isActive`
+ * Code that relies on `.job.isActive` no longer can be called from such entry points safely
+* Change the default behaviour -- return `true`. It also ""fixes"" such patterns as `GlobalScope.isActive` but basically is a breaking change",1.0,"Consider deprecating or changing the behaviour of CoroutineContext.isActive - According to the [doc](https://kotlin.github.io/kotlinx.coroutines/kotlinx-coroutines-core/kotlinx.coroutines/is-active.html), `isActive` has the following property:
+
+>The coroutineContext.isActive expression is a shortcut for coroutineContext[Job]?.isActive == true. See [Job.isActive](https://kotlin.github.io/kotlinx.coroutines/kotlinx-coroutines-core/kotlinx.coroutines/-job/is-active.html).
+
+It means that, if the `Job` is not present, `isActive` always returns `false`.
+
+We have multiple reports that such behaviour can be error-prone when used with non-`kotlinx.coroutines` entry points, such as Ktor and `suspend fun main`, because it is inconsistent with the overall contract:
+
+>(Job) has not been completed and was not cancelled yet
+
+
+`CoroutineContext.isActive` predates both `CoroutineScope` (which should always have a `Job` in it, if it's not `GlobalScope`) and `job` extension, so it may be the case that it can be safely deprecated.
+
+Basically, we have three options:
+
+* Do nothing, left things as is. It doesn't solve the original issue, but also doesn't introduce any potentially breaking changes
+* Deprecate `CoroutineContext.isActive`. Such change has multiple potential downsides
+ * Its only possible replacement is `this.job.isActive`, but this replacement is not equivalent to the original method -- `.job` throws an exception for contexts without a `Job`. An absence of replacement can be too disturbing as [a lot of code](https://grep.app/search?q=context.isActive&filter[lang][0]=Kotlin) rely on a perfectly fine `ctxWithJob.isActive`
+ * Code that relies on `.job.isActive` no longer can be called from such entry points safely
+* Change the default behaviour -- return `true`. It also ""fixes"" such patterns as `GlobalScope.isActive` but basically is a breaking change",0,consider deprecating or changing the behaviour of coroutinecontext isactive according to the isactive has the following property the coroutinecontext isactive expression is a shortcut for coroutinecontext isactive true see it means that if the job is not present isactive always returns false we have multiple reports that such behaviour can be error prone when used with non kotlinx coroutines entry points such as ktor and suspend fun main because it is inconsistent with the overall contract job has not been completed and was not cancelled yet coroutinecontext isactive predates both coroutinescope which should always have a job in it if it s not globalscope and job extension so it may be the case that it can be safely deprecated basically we have three options do nothing left things as is it doesn t solve the original issue but also doesn t introduce any potentially breaking changes deprecate coroutinecontext isactive such change has multiple potential downsides its only possible replacement is this job isactive but this replacement is not equivalent to the original method job throws an exception for contexts without a job an absence of replacement can be too disturbing as kotlin rely on a perfectly fine ctxwithjob isactive code that relies on job isactive no longer can be called from such entry points safely change the default behaviour return true it also fixes such patterns as globalscope isactive but basically is a breaking change,0
+415002,12122791432.0,IssuesEvent,2020-04-22 11:36:47,onaio/reveal-frontend,https://api.github.com/repos/onaio/reveal-frontend,closed,The Conditional formatting rules percentage table shows double,Priority: High bug,"Login to the system
+Select the IRS reporting
+Select one of the IRS plans
+Observe the Conditional formatting rules percentage table.
+Observe if it shows double data. It should not show two sets of data.
+
+
+",1.0,"The Conditional formatting rules percentage table shows double - Login to the system
+Select the IRS reporting
+Select one of the IRS plans
+Observe the Conditional formatting rules percentage table.
+Observe if it shows double data. It should not show two sets of data.
+
+
+",0,the conditional formatting rules percentage table shows double login to the system select the irs reporting select one of the irs plans observe the conditional formatting rules percentage table observe if it shows double data it should not show two sets of data ,0
+41946,16990537448.0,IssuesEvent,2021-06-30 19:48:32,MicrosoftDocs/azure-docs,https://api.github.com/repos/MicrosoftDocs/azure-docs,closed,Remove --export from kubectl command,Pri2 assigned-to-author container-service/svc doc-bug triaged,"Remove --export from this command as it is deprecated.
+`kubectl get deployment -o=yaml --export > deployments.yaml`
+
+Exact link: https://docs.microsoft.com/en-us/azure/aks/aks-migration#deployment-of-your-cluster-configuration
+
+> $ kubectl get deployment -o=yaml --export > deployments.yaml
+Error: unknown flag: --export
+See 'kubectl get --help' for usage.
+
+
+---
+#### Document Details
+
+⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
+
+* ID: 93700120-128a-abdb-64b6-d6c3704bcde5
+* Version Independent ID: afe2bb28-effb-a0a0-463a-51b1e2e5137d
+* Content: [Migrate to Azure Kubernetes Service (AKS) - Azure Kubernetes Service](https://docs.microsoft.com/en-us/azure/aks/aks-migration)
+* Content Source: [articles/aks/aks-migration.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/aks/aks-migration.md)
+* Service: **container-service**
+* GitHub Login: @mlearned
+* Microsoft Alias: **mlearned**",1.0,"Remove --export from kubectl command - Remove --export from this command as it is deprecated.
+`kubectl get deployment -o=yaml --export > deployments.yaml`
+
+Exact link: https://docs.microsoft.com/en-us/azure/aks/aks-migration#deployment-of-your-cluster-configuration
+
+> $ kubectl get deployment -o=yaml --export > deployments.yaml
+Error: unknown flag: --export
+See 'kubectl get --help' for usage.
+
+
+---
+#### Document Details
+
+⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
+
+* ID: 93700120-128a-abdb-64b6-d6c3704bcde5
+* Version Independent ID: afe2bb28-effb-a0a0-463a-51b1e2e5137d
+* Content: [Migrate to Azure Kubernetes Service (AKS) - Azure Kubernetes Service](https://docs.microsoft.com/en-us/azure/aks/aks-migration)
+* Content Source: [articles/aks/aks-migration.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/aks/aks-migration.md)
+* Service: **container-service**
+* GitHub Login: @mlearned
+* Microsoft Alias: **mlearned**",0,remove export from kubectl command remove export from this command as it is deprecated kubectl get deployment o yaml export deployments yaml exact link kubectl get deployment o yaml export deployments yaml error unknown flag export see kubectl get help for usage document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id abdb version independent id effb content content source service container service github login mlearned microsoft alias mlearned ,0
+275650,8578530970.0,IssuesEvent,2018-11-13 05:35:07,shelljs/shelljs,https://api.github.com/repos/shelljs/shelljs,closed,cp -Ru respects the -R but not the -u,fix low priority,"Love the lib - was looking to use this to reduce copy times on builds, but ran into the following:
+
+### Node version (or tell us if you're using electron or some other framework):
+8.6.0
+### ShellJS version (the most recent version/Github branch you see the bug on):
+shelljs 0.7.8, tested both directly and via shx 0.2.2
+### Operating system:
+Arch Linux (kernel 4.13.12-1)
+### Description of the bug:
+When combining `recursive` `-R` and `update` `-u` options in both `shelljs` and `shx`, ""recursive"" overrides and ignores ""update"" - files are recursively copied regardless of whether the src files are newer than dist.
+
+### Example ShellJS command to reproduce the error:
+Test case (assuming src/static and dist/static have contents)
+```javascript
+const shell = require('shelljs');
+
+// overwrite dist/static, verify that all timestamps in dist are now newer than those in src
+shell.cp('-R', 'src/static', 'dist/');
+console.log(shell.ls('-l', 'src/static').stdout, '\n' + shell.ls('-l', 'dist/static').stdout);
+console.log('----------');
+
+setTimeout(() => {
+ shell.cp('-Ru', 'src/static/', 'dist/');
+ console.log(shell.ls('-l', 'dist/static').stdout);
+
+ // timestamps shouldn't change, but are ~1s newer
+}, 1000);
+```
+
+### Expected result
+Same as
+```bash
+cp -R src/static/ dist/ && ls -l src/static && ls -l dist/static
+echo '\n-----\n'
+sleep 1
+cp -Ru src/static/ dist/ && ls -l dist/static
+```
+
+I think I see where the fix is - just wanted to file this before I forgot in case I don't end up having time in the near future to create a PR.",1.0,"cp -Ru respects the -R but not the -u - Love the lib - was looking to use this to reduce copy times on builds, but ran into the following:
+
+### Node version (or tell us if you're using electron or some other framework):
+8.6.0
+### ShellJS version (the most recent version/Github branch you see the bug on):
+shelljs 0.7.8, tested both directly and via shx 0.2.2
+### Operating system:
+Arch Linux (kernel 4.13.12-1)
+### Description of the bug:
+When combining `recursive` `-R` and `update` `-u` options in both `shelljs` and `shx`, ""recursive"" overrides and ignores ""update"" - files are recursively copied regardless of whether the src files are newer than dist.
+
+### Example ShellJS command to reproduce the error:
+Test case (assuming src/static and dist/static have contents)
+```javascript
+const shell = require('shelljs');
+
+// overwrite dist/static, verify that all timestamps in dist are now newer than those in src
+shell.cp('-R', 'src/static', 'dist/');
+console.log(shell.ls('-l', 'src/static').stdout, '\n' + shell.ls('-l', 'dist/static').stdout);
+console.log('----------');
+
+setTimeout(() => {
+ shell.cp('-Ru', 'src/static/', 'dist/');
+ console.log(shell.ls('-l', 'dist/static').stdout);
+
+ // timestamps shouldn't change, but are ~1s newer
+}, 1000);
+```
+
+### Expected result
+Same as
+```bash
+cp -R src/static/ dist/ && ls -l src/static && ls -l dist/static
+echo '\n-----\n'
+sleep 1
+cp -Ru src/static/ dist/ && ls -l dist/static
+```
+
+I think I see where the fix is - just wanted to file this before I forgot in case I don't end up having time in the near future to create a PR.",0,cp ru respects the r but not the u love the lib was looking to use this to reduce copy times on builds but ran into the following node version or tell us if you re using electron or some other framework shelljs version the most recent version github branch you see the bug on shelljs tested both directly and via shx operating system arch linux kernel description of the bug when combining recursive r and update u options in both shelljs and shx recursive overrides and ignores update files are recursively copied regardless of whether the src files are newer than dist example shelljs command to reproduce the error test case assuming src static and dist static have contents javascript const shell require shelljs overwrite dist static verify that all timestamps in dist are now newer than those in src shell cp r src static dist console log shell ls l src static stdout n shell ls l dist static stdout console log settimeout shell cp ru src static dist console log shell ls l dist static stdout timestamps shouldn t change but are newer expected result same as bash cp r src static dist ls l src static ls l dist static echo n n sleep cp ru src static dist ls l dist static i think i see where the fix is just wanted to file this before i forgot in case i don t end up having time in the near future to create a pr ,0
+158182,12405427030.0,IssuesEvent,2020-05-21 17:15:19,HMIS/LSASampleCode,https://api.github.com/repos/HMIS/LSASampleCode,closed,Test Data - Inventory designated beds issues,Test Kit Data,"InventoryID = 4083317,
+- HouseholdType = 4 (Households with only Children)
+- Designated beds for this Inventory include CHVetBeds, VetBed, CHBeds
+- For HouseholdType 4, we have no way to add this type of designated beds to the inventory.
+
+InventoryID = 4083318,
+- HouseholdType = 3 (Households with at least one adult and one child)
+- Designated beds for this Inventory include CHYouthBedInventory
+- For HouseholdType 3, we have no way to add this type of designated bed to the inventory.
+
+
+",1.0,"Test Data - Inventory designated beds issues - InventoryID = 4083317,
+- HouseholdType = 4 (Households with only Children)
+- Designated beds for this Inventory include CHVetBeds, VetBed, CHBeds
+- For HouseholdType 4, we have no way to add this type of designated beds to the inventory.
+
+InventoryID = 4083318,
+- HouseholdType = 3 (Households with at least one adult and one child)
+- Designated beds for this Inventory include CHYouthBedInventory
+- For HouseholdType 3, we have no way to add this type of designated bed to the inventory.
+
+
+",0,test data inventory designated beds issues inventoryid householdtype households with only children designated beds for this inventory include chvetbeds vetbed chbeds for householdtype we have no way to add this type of designated beds to the inventory inventoryid householdtype households with at least one adult and one child designated beds for this inventory include chyouthbedinventory for householdtype we have no way to add this type of designated bed to the inventory ,0
+105367,4234626319.0,IssuesEvent,2016-07-05 12:42:43,flurinduerst/WPSeed,https://api.github.com/repos/flurinduerst/WPSeed,closed,super duper cleanup,cleanup high-priority,"
+called her to clean lots of stuff like
+* ~~unused css~~
+* ~~assets~~
+* functions",1.0,"super duper cleanup - 
+called her to clean lots of stuff like
+* ~~unused css~~
+* ~~assets~~
+* functions",0,super duper cleanup called her to clean lots of stuff like unused css assets functions,0
+183,3994642760.0,IssuesEvent,2016-05-10 13:10:21,wordpress-mobile/WordPress-iOS,https://api.github.com/repos/wordpress-mobile/WordPress-iOS,closed,People Management: Update Users,People Management [Type] Enhancement,"#### Details:
+We should allow Blog Admins to edit Users. Initially, we need to support *Update Roles*
+
+Ref. #5030
+
+--
+
+
+
+",1.0,"People Management: Update Users - #### Details:
+We should allow Blog Admins to edit Users. Initially, we need to support *Update Roles*
+
+Ref. #5030
+
+--
+
+
+
+",1,people management update users details we should allow blog admins to edit users initially we need to support update roles ref ,1
+1122,27230646310.0,IssuesEvent,2023-02-21 13:00:34,scikit-image/scikit-image,https://api.github.com/repos/scikit-image/scikit-image,closed,Could not build wheels for scikit-image,:people_hugging: Support,"Hello, I have been trying to install this Python module for a number of days now as it is a dependency of img2database. I'm not completely sure what is causing the problem. Hence I have been unable to try a large variety of fixes. I believe the problem might potentially have something to do with pip as the module installed successfully using mini conda however the module I wish to use is only available on pip. Any help would be greatly appreciated.
+The full output is as follows:
+
+
+Click to expand
+
+```
+C:\Windows\System32>pip install scikit-image
+Collecting scikit-image
+ Using cached scikit-image-0.19.3.tar.gz (22.2 MB)
+ Installing build dependencies ... done
+ Getting requirements to build wheel ... done
+ Preparing metadata (pyproject.toml) ... done
+Requirement already satisfied: numpy>=1.17.0 in c:\users\jay\appdata\local\programs\python\python311\lib\site-packages (from scikit-image) (1.24.2)
+Requirement already satisfied: scipy>=1.4.1 in c:\users\jay\appdata\local\programs\python\python311\lib\site-packages (from scikit-image) (1.10.1)
+Requirement already satisfied: networkx>=2.2 in c:\users\jay\appdata\local\programs\python\python311\lib\site-packages (from scikit-image) (3.0)
+Requirement already satisfied: pillow!=7.1.0,!=7.1.1,!=8.3.0,>=6.1.0 in c:\users\jay\appdata\local\programs\python\python311\lib\site-packages (from scikit-image) (9.4.0)
+Collecting imageio>=2.4.1
+ Using cached imageio-2.25.1-py3-none-any.whl (3.4 MB)
+Collecting tifffile>=2019.7.26
+ Using cached tifffile-2023.2.3-py3-none-any.whl (215 kB)
+Collecting PyWavelets>=1.1.1
+ Using cached PyWavelets-1.4.1-cp311-cp311-win_amd64.whl (4.2 MB)
+Requirement already satisfied: packaging>=20.0 in c:\users\jay\appdata\local\programs\python\python311\lib\site-packages (from scikit-image) (23.0)
+Building wheels for collected packages: scikit-image
+ Building wheel for scikit-image (pyproject.toml) ... error
+ error: subprocess-exited-with-error
+
+ × Building wheel for scikit-image (pyproject.toml) did not run successfully.
+ │ exit code: 1
+ ╰─> [626 lines of output]
+ setup.py:9: DeprecationWarning:
+
+ `numpy.distutils` is deprecated since NumPy 1.23.0, as a result
+ of the deprecation of `distutils` itself. It will be removed for
+ Python >= 3.12. For older Python versions it will remain present.
+ It is recommended to use `setuptools < 60.0` for those Python versions.
+ For more details, see:
+ https://numpy.org/devdocs/reference/distutils_status_migration.html
+
+
+ from numpy.distutils.command.build_ext import build_ext as npy_build_ext
+ Partial import of skimage during the build process.
+ Compiling C:\Users\jay\AppData\Local\Temp\pip-install-s2xz2koj\scikit-image_cad4b2fea6c2495988cddf5b95ba2806\skimage\morphology\_skeletonize_3d_cy.pyx because it changed.
+ [1/1] Cythonizing C:\Users\jay\AppData\Local\Temp\pip-install-s2xz2koj\scikit-image_cad4b2fea6c2495988cddf5b95ba2806\skimage\morphology\_skeletonize_3d_cy.pyx
+ running bdist_wheel
+ running build
+ running config_cc
+ INFO: unifing config_cc, config, build_clib, build_ext, build commands --compiler options
+ running config_fc
+ INFO: unifing config_fc, config, build_clib, build_ext, build commands --fcompiler options
+ running build_src
+ INFO: build_src
+ INFO: building extension ""skimage._shared.geometry"" sources
+ INFO: building extension ""skimage._shared.transform"" sources
+ INFO: building extension ""skimage._shared.interpolation"" sources
+ INFO: building extension ""skimage._shared.fast_exp"" sources
+ INFO: building extension ""skimage.draw._draw"" sources
+ INFO: building extension ""skimage.feature._cascade"" sources
+ INFO: building extension ""skimage.feature.corner_cy"" sources
+ INFO: building extension ""skimage.feature.censure_cy"" sources
+ INFO: building extension ""skimage.feature.orb_cy"" sources
+ INFO: building extension ""skimage.feature._texture"" sources
+ INFO: building extension ""skimage.feature._hoghistogram"" sources
+ INFO: building extension ""skimage.feature._haar"" sources
+ INFO: building extension ""skimage.feature._sift"" sources
+ INFO: building extension ""skimage.feature.brief_cy"" sources
+ INFO: building extension ""skimage.feature._hessian_det_appx"" sources
+ INFO: building extension ""skimage.restoration._unwrap_1d"" sources
+ INFO: building extension ""skimage.restoration._unwrap_2d"" sources
+ INFO: building extension ""skimage.restoration._unwrap_3d"" sources
+ INFO: building extension ""skimage.restoration._denoise_cy"" sources
+ INFO: building extension ""skimage.restoration._nl_means_denoising"" sources
+ INFO: building extension ""skimage.restoration._rolling_ball_cy"" sources
+ INFO: building extension ""skimage.restoration._inpaint"" sources
+ INFO: building extension ""skimage.filters.rank.core_cy"" sources
+ INFO: building extension ""skimage.filters.rank.core_cy_3d"" sources
+ INFO: building extension ""skimage.filters._multiotsu"" sources
+ INFO: building extension ""skimage.filters.rank.generic_cy"" sources
+ INFO: building extension ""skimage.filters.rank.percentile_cy"" sources
+ INFO: building extension ""skimage.filters.rank.bilateral_cy"" sources
+ INFO: building extension ""skimage.future.graph._ncut_cy"" sources
+ INFO: building extension ""skimage.graph._spath"" sources
+ INFO: building extension ""skimage.graph._mcp"" sources
+ INFO: building extension ""skimage.graph.heap"" sources
+ INFO: building extension ""skimage.io._plugins._colormixer"" sources
+ INFO: building extension ""skimage.io._plugins._histograms"" sources
+ INFO: building extension ""skimage.measure._ccomp"" sources
+ INFO: building extension ""skimage.measure._find_contours_cy"" sources
+ INFO: building extension ""skimage.measure._moments_cy"" sources
+ INFO: building extension ""skimage.measure._marching_cubes_classic_cy"" sources
+ INFO: building extension ""skimage.measure._marching_cubes_lewiner_cy"" sources
+ INFO: building extension ""skimage.measure._pnpoly"" sources
+ INFO: building extension ""skimage.morphology._skeletonize_cy"" sources
+ INFO: building extension ""skimage.morphology._convex_hull"" sources
+ INFO: building extension ""skimage.morphology._grayreconstruct"" sources
+ INFO: building extension ""skimage.morphology._max_tree"" sources
+ INFO: building extension ""skimage.morphology._skeletonize_3d_cy"" sources
+ INFO: building extension ""skimage.morphology._extrema_cy"" sources
+ INFO: building extension ""skimage.morphology._flood_fill_cy"" sources
+ INFO: building extension ""skimage.transform._hough_transform"" sources
+ INFO: building extension ""skimage.transform._warps_cy"" sources
+ INFO: building extension ""skimage.transform._radon_transform"" sources
+ INFO: building extension ""skimage.util._remap"" sources
+ INFO: building extension ""skimage.segmentation._watershed_cy"" sources
+ INFO: building extension ""skimage.segmentation._felzenszwalb_cy"" sources
+ INFO: building extension ""skimage.segmentation._quickshift_cy"" sources
+ INFO: building extension ""skimage.segmentation._slic"" sources
+ INFO: building data_files sources
+ INFO: build_src: building npy-pkg config files
+ running build_py
+ creating build
+ creating build\lib.win-amd64-3.11
+ creating build\lib.win-amd64-3.11\skimage
+ copying skimage\conftest.py -> build\lib.win-amd64-3.11\skimage
+ copying skimage\setup.py -> build\lib.win-amd64-3.11\skimage
+ copying skimage\_build.py -> build\lib.win-amd64-3.11\skimage
+ copying skimage\__init__.py -> build\lib.win-amd64-3.11\skimage
+ creating build\lib.win-amd64-3.11\doc
+ creating build\lib.win-amd64-3.11\doc\ext
+ copying doc\ext\doi_role.py -> build\lib.win-amd64-3.11\doc\ext
+ copying doc\ext\notebook_doc.py -> build\lib.win-amd64-3.11\doc\ext
+ copying doc\ext\plot2rst.py -> build\lib.win-amd64-3.11\doc\ext
+ copying doc\ext\__init__.py -> build\lib.win-amd64-3.11\doc\ext
+ creating build\lib.win-amd64-3.11\doc\ext\tests
+ copying doc\ext\tests\test_notebook_doc.py -> build\lib.win-amd64-3.11\doc\ext\tests
+ copying doc\ext\tests\__init__.py -> build\lib.win-amd64-3.11\doc\ext\tests
+ creating build\lib.win-amd64-3.11\skimage\color
+ copying skimage\color\adapt_rgb.py -> build\lib.win-amd64-3.11\skimage\color
+ copying skimage\color\colorconv.py -> build\lib.win-amd64-3.11\skimage\color
+ copying skimage\color\colorlabel.py -> build\lib.win-amd64-3.11\skimage\color
+ copying skimage\color\delta_e.py -> build\lib.win-amd64-3.11\skimage\color
+ copying skimage\color\rgb_colors.py -> build\lib.win-amd64-3.11\skimage\color
+ copying skimage\color\__init__.py -> build\lib.win-amd64-3.11\skimage\color
+ creating build\lib.win-amd64-3.11\skimage\data
+ copying skimage\data\setup.py -> build\lib.win-amd64-3.11\skimage\data
+ copying skimage\data\_binary_blobs.py -> build\lib.win-amd64-3.11\skimage\data
+ copying skimage\data\_fetchers.py -> build\lib.win-amd64-3.11\skimage\data
+ copying skimage\data\_registry.py -> build\lib.win-amd64-3.11\skimage\data
+ copying skimage\data\__init__.py -> build\lib.win-amd64-3.11\skimage\data
+ creating build\lib.win-amd64-3.11\skimage\draw
+ copying skimage\draw\draw.py -> build\lib.win-amd64-3.11\skimage\draw
+ copying skimage\draw\draw3d.py -> build\lib.win-amd64-3.11\skimage\draw
+ copying skimage\draw\draw_nd.py -> build\lib.win-amd64-3.11\skimage\draw
+ copying skimage\draw\setup.py -> build\lib.win-amd64-3.11\skimage\draw
+ copying skimage\draw\_polygon2mask.py -> build\lib.win-amd64-3.11\skimage\draw
+ copying skimage\draw\_random_shapes.py -> build\lib.win-amd64-3.11\skimage\draw
+ copying skimage\draw\__init__.py -> build\lib.win-amd64-3.11\skimage\draw
+ creating build\lib.win-amd64-3.11\skimage\exposure
+ copying skimage\exposure\exposure.py -> build\lib.win-amd64-3.11\skimage\exposure
+ copying skimage\exposure\histogram_matching.py -> build\lib.win-amd64-3.11\skimage\exposure
+ copying skimage\exposure\setup.py -> build\lib.win-amd64-3.11\skimage\exposure
+ copying skimage\exposure\_adapthist.py -> build\lib.win-amd64-3.11\skimage\exposure
+ copying skimage\exposure\__init__.py -> build\lib.win-amd64-3.11\skimage\exposure
+ creating build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\blob.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\brief.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\brief_pythran.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\censure.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\corner.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\haar.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\match.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\orb.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\peak.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\setup.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\sift.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\template.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\texture.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\util.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\_basic_features.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\_canny.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\_daisy.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\_hessian_det_appx_pythran.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\_hog.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\_orb_descriptor_positions.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\__init__.py -> build\lib.win-amd64-3.11\skimage\feature
+ creating build\lib.win-amd64-3.11\skimage\filters
+ copying skimage\filters\edges.py -> build\lib.win-amd64-3.11\skimage\filters
+ copying skimage\filters\lpi_filter.py -> build\lib.win-amd64-3.11\skimage\filters
+ copying skimage\filters\ridges.py -> build\lib.win-amd64-3.11\skimage\filters
+ copying skimage\filters\setup.py -> build\lib.win-amd64-3.11\skimage\filters
+ copying skimage\filters\thresholding.py -> build\lib.win-amd64-3.11\skimage\filters
+ copying skimage\filters\_fft_based.py -> build\lib.win-amd64-3.11\skimage\filters
+ copying skimage\filters\_gabor.py -> build\lib.win-amd64-3.11\skimage\filters
+ copying skimage\filters\_gaussian.py -> build\lib.win-amd64-3.11\skimage\filters
+ copying skimage\filters\_median.py -> build\lib.win-amd64-3.11\skimage\filters
+ copying skimage\filters\_rank_order.py -> build\lib.win-amd64-3.11\skimage\filters
+ copying skimage\filters\_sparse.py -> build\lib.win-amd64-3.11\skimage\filters
+ copying skimage\filters\_unsharp_mask.py -> build\lib.win-amd64-3.11\skimage\filters
+ copying skimage\filters\_window.py -> build\lib.win-amd64-3.11\skimage\filters
+ copying skimage\filters\__init__.py -> build\lib.win-amd64-3.11\skimage\filters
+ creating build\lib.win-amd64-3.11\skimage\future
+ copying skimage\future\manual_segmentation.py -> build\lib.win-amd64-3.11\skimage\future
+ copying skimage\future\setup.py -> build\lib.win-amd64-3.11\skimage\future
+ copying skimage\future\trainable_segmentation.py -> build\lib.win-amd64-3.11\skimage\future
+ copying skimage\future\__init__.py -> build\lib.win-amd64-3.11\skimage\future
+ creating build\lib.win-amd64-3.11\skimage\graph
+ copying skimage\graph\mcp.py -> build\lib.win-amd64-3.11\skimage\graph
+ copying skimage\graph\setup.py -> build\lib.win-amd64-3.11\skimage\graph
+ copying skimage\graph\spath.py -> build\lib.win-amd64-3.11\skimage\graph
+ copying skimage\graph\_graph.py -> build\lib.win-amd64-3.11\skimage\graph
+ copying skimage\graph\__init__.py -> build\lib.win-amd64-3.11\skimage\graph
+ creating build\lib.win-amd64-3.11\skimage\io
+ copying skimage\io\collection.py -> build\lib.win-amd64-3.11\skimage\io
+ copying skimage\io\manage_plugins.py -> build\lib.win-amd64-3.11\skimage\io
+ copying skimage\io\setup.py -> build\lib.win-amd64-3.11\skimage\io
+ copying skimage\io\sift.py -> build\lib.win-amd64-3.11\skimage\io
+ copying skimage\io\util.py -> build\lib.win-amd64-3.11\skimage\io
+ copying skimage\io\_image_stack.py -> build\lib.win-amd64-3.11\skimage\io
+ copying skimage\io\_io.py -> build\lib.win-amd64-3.11\skimage\io
+ copying skimage\io\__init__.py -> build\lib.win-amd64-3.11\skimage\io
+ creating build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\block.py -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\entropy.py -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\fit.py -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\pnpoly.py -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\profile.py -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\setup.py -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\_blur_effect.py -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\_find_contours.py -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\_label.py -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\_marching_cubes_classic.py -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\_marching_cubes_lewiner.py -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\_marching_cubes_lewiner_luts.py -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\_moments.py -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\_polygon.py -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\_regionprops.py -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\_regionprops_utils.py -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\__init__.py -> build\lib.win-amd64-3.11\skimage\measure
+ creating build\lib.win-amd64-3.11\skimage\metrics
+ copying skimage\metrics\set_metrics.py -> build\lib.win-amd64-3.11\skimage\metrics
+ copying skimage\metrics\simple_metrics.py -> build\lib.win-amd64-3.11\skimage\metrics
+ copying skimage\metrics\_adapted_rand_error.py -> build\lib.win-amd64-3.11\skimage\metrics
+ copying skimage\metrics\_contingency_table.py -> build\lib.win-amd64-3.11\skimage\metrics
+ copying skimage\metrics\_structural_similarity.py -> build\lib.win-amd64-3.11\skimage\metrics
+ copying skimage\metrics\_variation_of_information.py -> build\lib.win-amd64-3.11\skimage\metrics
+ copying skimage\metrics\__init__.py -> build\lib.win-amd64-3.11\skimage\metrics
+ creating build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\binary.py -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\convex_hull.py -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\extrema.py -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\footprints.py -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\gray.py -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\grayreconstruct.py -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\grey.py -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\greyreconstruct.py -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\max_tree.py -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\misc.py -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\selem.py -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\setup.py -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\_flood_fill.py -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\_skeletonize.py -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\_util.py -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\__init__.py -> build\lib.win-amd64-3.11\skimage\morphology
+ creating build\lib.win-amd64-3.11\skimage\registration
+ copying skimage\registration\_masked_phase_cross_correlation.py -> build\lib.win-amd64-3.11\skimage\registration
+ copying skimage\registration\_optical_flow.py -> build\lib.win-amd64-3.11\skimage\registration
+ copying skimage\registration\_optical_flow_utils.py -> build\lib.win-amd64-3.11\skimage\registration
+ copying skimage\registration\_phase_cross_correlation.py -> build\lib.win-amd64-3.11\skimage\registration
+ copying skimage\registration\__init__.py -> build\lib.win-amd64-3.11\skimage\registration
+ creating build\lib.win-amd64-3.11\skimage\restoration
+ copying skimage\restoration\deconvolution.py -> build\lib.win-amd64-3.11\skimage\restoration
+ copying skimage\restoration\inpaint.py -> build\lib.win-amd64-3.11\skimage\restoration
+ copying skimage\restoration\j_invariant.py -> build\lib.win-amd64-3.11\skimage\restoration
+ copying skimage\restoration\non_local_means.py -> build\lib.win-amd64-3.11\skimage\restoration
+ copying skimage\restoration\rolling_ball.py -> build\lib.win-amd64-3.11\skimage\restoration
+ copying skimage\restoration\setup.py -> build\lib.win-amd64-3.11\skimage\restoration
+ copying skimage\restoration\uft.py -> build\lib.win-amd64-3.11\skimage\restoration
+ copying skimage\restoration\unwrap.py -> build\lib.win-amd64-3.11\skimage\restoration
+ copying skimage\restoration\_cycle_spin.py -> build\lib.win-amd64-3.11\skimage\restoration
+ copying skimage\restoration\_denoise.py -> build\lib.win-amd64-3.11\skimage\restoration
+ copying skimage\restoration\__init__.py -> build\lib.win-amd64-3.11\skimage\restoration
+ creating build\lib.win-amd64-3.11\skimage\scripts
+ copying skimage\scripts\skivi.py -> build\lib.win-amd64-3.11\skimage\scripts
+ copying skimage\scripts\__init__.py -> build\lib.win-amd64-3.11\skimage\scripts
+ creating build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\active_contour_model.py -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\boundaries.py -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\morphsnakes.py -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\random_walker_segmentation.py -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\setup.py -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\slic_superpixels.py -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\_chan_vese.py -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\_clear_border.py -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\_expand_labels.py -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\_felzenszwalb.py -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\_join.py -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\_quickshift.py -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\_watershed.py -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\__init__.py -> build\lib.win-amd64-3.11\skimage\segmentation
+ creating build\lib.win-amd64-3.11\skimage\transform
+ copying skimage\transform\finite_radon_transform.py -> build\lib.win-amd64-3.11\skimage\transform
+ copying skimage\transform\hough_transform.py -> build\lib.win-amd64-3.11\skimage\transform
+ copying skimage\transform\integral.py -> build\lib.win-amd64-3.11\skimage\transform
+ copying skimage\transform\pyramids.py -> build\lib.win-amd64-3.11\skimage\transform
+ copying skimage\transform\radon_transform.py -> build\lib.win-amd64-3.11\skimage\transform
+ copying skimage\transform\setup.py -> build\lib.win-amd64-3.11\skimage\transform
+ copying skimage\transform\_geometric.py -> build\lib.win-amd64-3.11\skimage\transform
+ copying skimage\transform\_warps.py -> build\lib.win-amd64-3.11\skimage\transform
+ copying skimage\transform\__init__.py -> build\lib.win-amd64-3.11\skimage\transform
+ creating build\lib.win-amd64-3.11\skimage\util
+ copying skimage\util\apply_parallel.py -> build\lib.win-amd64-3.11\skimage\util
+ copying skimage\util\arraycrop.py -> build\lib.win-amd64-3.11\skimage\util
+ copying skimage\util\compare.py -> build\lib.win-amd64-3.11\skimage\util
+ copying skimage\util\dtype.py -> build\lib.win-amd64-3.11\skimage\util
+ copying skimage\util\lookfor.py -> build\lib.win-amd64-3.11\skimage\util
+ copying skimage\util\noise.py -> build\lib.win-amd64-3.11\skimage\util
+ copying skimage\util\setup.py -> build\lib.win-amd64-3.11\skimage\util
+ copying skimage\util\shape.py -> build\lib.win-amd64-3.11\skimage\util
+ copying skimage\util\unique.py -> build\lib.win-amd64-3.11\skimage\util
+ copying skimage\util\_invert.py -> build\lib.win-amd64-3.11\skimage\util
+ copying skimage\util\_label.py -> build\lib.win-amd64-3.11\skimage\util
+ copying skimage\util\_map_array.py -> build\lib.win-amd64-3.11\skimage\util
+ copying skimage\util\_montage.py -> build\lib.win-amd64-3.11\skimage\util
+ copying skimage\util\_regular_grid.py -> build\lib.win-amd64-3.11\skimage\util
+ copying skimage\util\__init__.py -> build\lib.win-amd64-3.11\skimage\util
+ creating build\lib.win-amd64-3.11\skimage\viewer
+ copying skimage\viewer\qt.py -> build\lib.win-amd64-3.11\skimage\viewer
+ copying skimage\viewer\__init__.py -> build\lib.win-amd64-3.11\skimage\viewer
+ creating build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\coord.py -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\filters.py -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\lazy.py -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\setup.py -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\tester.py -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\testing.py -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\utils.py -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\version_requirements.py -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\_dependency_checks.py -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\_geometry.py -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\_tempfile.py -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\_warnings.py -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\__init__.py -> build\lib.win-amd64-3.11\skimage\_shared
+ creating build\lib.win-amd64-3.11\skimage\color\tests
+ copying skimage\color\tests\test_adapt_rgb.py -> build\lib.win-amd64-3.11\skimage\color\tests
+ copying skimage\color\tests\test_colorconv.py -> build\lib.win-amd64-3.11\skimage\color\tests
+ copying skimage\color\tests\test_colorlabel.py -> build\lib.win-amd64-3.11\skimage\color\tests
+ copying skimage\color\tests\test_delta_e.py -> build\lib.win-amd64-3.11\skimage\color\tests
+ copying skimage\color\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\color\tests
+ creating build\lib.win-amd64-3.11\skimage\data\tests
+ copying skimage\data\tests\test_data.py -> build\lib.win-amd64-3.11\skimage\data\tests
+ copying skimage\data\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\data\tests
+ creating build\lib.win-amd64-3.11\skimage\draw\tests
+ copying skimage\draw\tests\test_draw.py -> build\lib.win-amd64-3.11\skimage\draw\tests
+ copying skimage\draw\tests\test_draw3d.py -> build\lib.win-amd64-3.11\skimage\draw\tests
+ copying skimage\draw\tests\test_draw_nd.py -> build\lib.win-amd64-3.11\skimage\draw\tests
+ copying skimage\draw\tests\test_polygon2mask.py -> build\lib.win-amd64-3.11\skimage\draw\tests
+ copying skimage\draw\tests\test_random_shapes.py -> build\lib.win-amd64-3.11\skimage\draw\tests
+ copying skimage\draw\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\draw\tests
+ creating build\lib.win-amd64-3.11\skimage\exposure\tests
+ copying skimage\exposure\tests\test_exposure.py -> build\lib.win-amd64-3.11\skimage\exposure\tests
+ copying skimage\exposure\tests\test_histogram_matching.py -> build\lib.win-amd64-3.11\skimage\exposure\tests
+ copying skimage\exposure\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\exposure\tests
+ creating build\lib.win-amd64-3.11\skimage\feature\tests
+ copying skimage\feature\tests\test_basic_features.py -> build\lib.win-amd64-3.11\skimage\feature\tests
+ copying skimage\feature\tests\test_blob.py -> build\lib.win-amd64-3.11\skimage\feature\tests
+ copying skimage\feature\tests\test_brief.py -> build\lib.win-amd64-3.11\skimage\feature\tests
+ copying skimage\feature\tests\test_canny.py -> build\lib.win-amd64-3.11\skimage\feature\tests
+ copying skimage\feature\tests\test_cascade.py -> build\lib.win-amd64-3.11\skimage\feature\tests
+ copying skimage\feature\tests\test_censure.py -> build\lib.win-amd64-3.11\skimage\feature\tests
+ copying skimage\feature\tests\test_corner.py -> build\lib.win-amd64-3.11\skimage\feature\tests
+ copying skimage\feature\tests\test_daisy.py -> build\lib.win-amd64-3.11\skimage\feature\tests
+ copying skimage\feature\tests\test_haar.py -> build\lib.win-amd64-3.11\skimage\feature\tests
+ copying skimage\feature\tests\test_hog.py -> build\lib.win-amd64-3.11\skimage\feature\tests
+ copying skimage\feature\tests\test_match.py -> build\lib.win-amd64-3.11\skimage\feature\tests
+ copying skimage\feature\tests\test_orb.py -> build\lib.win-amd64-3.11\skimage\feature\tests
+ copying skimage\feature\tests\test_peak.py -> build\lib.win-amd64-3.11\skimage\feature\tests
+ copying skimage\feature\tests\test_sift.py -> build\lib.win-amd64-3.11\skimage\feature\tests
+ copying skimage\feature\tests\test_template.py -> build\lib.win-amd64-3.11\skimage\feature\tests
+ copying skimage\feature\tests\test_texture.py -> build\lib.win-amd64-3.11\skimage\feature\tests
+ copying skimage\feature\tests\test_util.py -> build\lib.win-amd64-3.11\skimage\feature\tests
+ copying skimage\feature\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\feature\tests
+ creating build\lib.win-amd64-3.11\skimage\filters\rank
+ copying skimage\filters\rank\bilateral.py -> build\lib.win-amd64-3.11\skimage\filters\rank
+ copying skimage\filters\rank\generic.py -> build\lib.win-amd64-3.11\skimage\filters\rank
+ copying skimage\filters\rank\_percentile.py -> build\lib.win-amd64-3.11\skimage\filters\rank
+ copying skimage\filters\rank\__init__.py -> build\lib.win-amd64-3.11\skimage\filters\rank
+ creating build\lib.win-amd64-3.11\skimage\filters\tests
+ copying skimage\filters\tests\test_correlate.py -> build\lib.win-amd64-3.11\skimage\filters\tests
+ copying skimage\filters\tests\test_edges.py -> build\lib.win-amd64-3.11\skimage\filters\tests
+ copying skimage\filters\tests\test_fft_based.py -> build\lib.win-amd64-3.11\skimage\filters\tests
+ copying skimage\filters\tests\test_gabor.py -> build\lib.win-amd64-3.11\skimage\filters\tests
+ copying skimage\filters\tests\test_gaussian.py -> build\lib.win-amd64-3.11\skimage\filters\tests
+ copying skimage\filters\tests\test_lpi_filter.py -> build\lib.win-amd64-3.11\skimage\filters\tests
+ copying skimage\filters\tests\test_median.py -> build\lib.win-amd64-3.11\skimage\filters\tests
+ copying skimage\filters\tests\test_ridges.py -> build\lib.win-amd64-3.11\skimage\filters\tests
+ copying skimage\filters\tests\test_thresholding.py -> build\lib.win-amd64-3.11\skimage\filters\tests
+ copying skimage\filters\tests\test_unsharp_mask.py -> build\lib.win-amd64-3.11\skimage\filters\tests
+ copying skimage\filters\tests\test_window.py -> build\lib.win-amd64-3.11\skimage\filters\tests
+ copying skimage\filters\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\filters\tests
+ creating build\lib.win-amd64-3.11\skimage\filters\rank\tests
+ copying skimage\filters\rank\tests\test_rank.py -> build\lib.win-amd64-3.11\skimage\filters\rank\tests
+ copying skimage\filters\rank\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\filters\rank\tests
+ creating build\lib.win-amd64-3.11\skimage\future\graph
+ copying skimage\future\graph\graph_cut.py -> build\lib.win-amd64-3.11\skimage\future\graph
+ copying skimage\future\graph\graph_merge.py -> build\lib.win-amd64-3.11\skimage\future\graph
+ copying skimage\future\graph\rag.py -> build\lib.win-amd64-3.11\skimage\future\graph
+ copying skimage\future\graph\setup.py -> build\lib.win-amd64-3.11\skimage\future\graph
+ copying skimage\future\graph\_ncut.py -> build\lib.win-amd64-3.11\skimage\future\graph
+ copying skimage\future\graph\__init__.py -> build\lib.win-amd64-3.11\skimage\future\graph
+ creating build\lib.win-amd64-3.11\skimage\future\tests
+ copying skimage\future\tests\test_trainable_segmentation.py -> build\lib.win-amd64-3.11\skimage\future\tests
+ copying skimage\future\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\future\tests
+ creating build\lib.win-amd64-3.11\skimage\future\graph\tests
+ copying skimage\future\graph\tests\test_rag.py -> build\lib.win-amd64-3.11\skimage\future\graph\tests
+ copying skimage\future\graph\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\future\graph\tests
+ creating build\lib.win-amd64-3.11\skimage\graph\tests
+ copying skimage\graph\tests\test_anisotropy.py -> build\lib.win-amd64-3.11\skimage\graph\tests
+ copying skimage\graph\tests\test_connect.py -> build\lib.win-amd64-3.11\skimage\graph\tests
+ copying skimage\graph\tests\test_flexible.py -> build\lib.win-amd64-3.11\skimage\graph\tests
+ copying skimage\graph\tests\test_heap.py -> build\lib.win-amd64-3.11\skimage\graph\tests
+ copying skimage\graph\tests\test_mcp.py -> build\lib.win-amd64-3.11\skimage\graph\tests
+ copying skimage\graph\tests\test_pixel_graph.py -> build\lib.win-amd64-3.11\skimage\graph\tests
+ copying skimage\graph\tests\test_spath.py -> build\lib.win-amd64-3.11\skimage\graph\tests
+ copying skimage\graph\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\graph\tests
+ creating build\lib.win-amd64-3.11\skimage\io\tests
+ copying skimage\io\tests\test_collection.py -> build\lib.win-amd64-3.11\skimage\io\tests
+ copying skimage\io\tests\test_colormixer.py -> build\lib.win-amd64-3.11\skimage\io\tests
+ copying skimage\io\tests\test_fits.py -> build\lib.win-amd64-3.11\skimage\io\tests
+ copying skimage\io\tests\test_histograms.py -> build\lib.win-amd64-3.11\skimage\io\tests
+ copying skimage\io\tests\test_imageio.py -> build\lib.win-amd64-3.11\skimage\io\tests
+ copying skimage\io\tests\test_imread.py -> build\lib.win-amd64-3.11\skimage\io\tests
+ copying skimage\io\tests\test_io.py -> build\lib.win-amd64-3.11\skimage\io\tests
+ copying skimage\io\tests\test_mpl_imshow.py -> build\lib.win-amd64-3.11\skimage\io\tests
+ copying skimage\io\tests\test_multi_image.py -> build\lib.win-amd64-3.11\skimage\io\tests
+ copying skimage\io\tests\test_pil.py -> build\lib.win-amd64-3.11\skimage\io\tests
+ copying skimage\io\tests\test_plugin.py -> build\lib.win-amd64-3.11\skimage\io\tests
+ copying skimage\io\tests\test_plugin_util.py -> build\lib.win-amd64-3.11\skimage\io\tests
+ copying skimage\io\tests\test_sift.py -> build\lib.win-amd64-3.11\skimage\io\tests
+ copying skimage\io\tests\test_simpleitk.py -> build\lib.win-amd64-3.11\skimage\io\tests
+ copying skimage\io\tests\test_tifffile.py -> build\lib.win-amd64-3.11\skimage\io\tests
+ copying skimage\io\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\io\tests
+ creating build\lib.win-amd64-3.11\skimage\io\_plugins
+ copying skimage\io\_plugins\fits_plugin.py -> build\lib.win-amd64-3.11\skimage\io\_plugins
+ copying skimage\io\_plugins\gdal_plugin.py -> build\lib.win-amd64-3.11\skimage\io\_plugins
+ copying skimage\io\_plugins\gtk_plugin.py -> build\lib.win-amd64-3.11\skimage\io\_plugins
+ copying skimage\io\_plugins\imageio_plugin.py -> build\lib.win-amd64-3.11\skimage\io\_plugins
+ copying skimage\io\_plugins\imread_plugin.py -> build\lib.win-amd64-3.11\skimage\io\_plugins
+ copying skimage\io\_plugins\matplotlib_plugin.py -> build\lib.win-amd64-3.11\skimage\io\_plugins
+ copying skimage\io\_plugins\pil_plugin.py -> build\lib.win-amd64-3.11\skimage\io\_plugins
+ copying skimage\io\_plugins\qt_plugin.py -> build\lib.win-amd64-3.11\skimage\io\_plugins
+ copying skimage\io\_plugins\q_color_mixer.py -> build\lib.win-amd64-3.11\skimage\io\_plugins
+ copying skimage\io\_plugins\q_histogram.py -> build\lib.win-amd64-3.11\skimage\io\_plugins
+ copying skimage\io\_plugins\simpleitk_plugin.py -> build\lib.win-amd64-3.11\skimage\io\_plugins
+ copying skimage\io\_plugins\skivi.py -> build\lib.win-amd64-3.11\skimage\io\_plugins
+ copying skimage\io\_plugins\tifffile_plugin.py -> build\lib.win-amd64-3.11\skimage\io\_plugins
+ copying skimage\io\_plugins\util.py -> build\lib.win-amd64-3.11\skimage\io\_plugins
+ copying skimage\io\_plugins\__init__.py -> build\lib.win-amd64-3.11\skimage\io\_plugins
+ creating build\lib.win-amd64-3.11\skimage\measure\tests
+ copying skimage\measure\tests\test_block.py -> build\lib.win-amd64-3.11\skimage\measure\tests
+ copying skimage\measure\tests\test_blur_effect.py -> build\lib.win-amd64-3.11\skimage\measure\tests
+ copying skimage\measure\tests\test_ccomp.py -> build\lib.win-amd64-3.11\skimage\measure\tests
+ copying skimage\measure\tests\test_entropy.py -> build\lib.win-amd64-3.11\skimage\measure\tests
+ copying skimage\measure\tests\test_find_contours.py -> build\lib.win-amd64-3.11\skimage\measure\tests
+ copying skimage\measure\tests\test_fit.py -> build\lib.win-amd64-3.11\skimage\measure\tests
+ copying skimage\measure\tests\test_label.py -> build\lib.win-amd64-3.11\skimage\measure\tests
+ copying skimage\measure\tests\test_marching_cubes.py -> build\lib.win-amd64-3.11\skimage\measure\tests
+ copying skimage\measure\tests\test_moments.py -> build\lib.win-amd64-3.11\skimage\measure\tests
+ copying skimage\measure\tests\test_pnpoly.py -> build\lib.win-amd64-3.11\skimage\measure\tests
+ copying skimage\measure\tests\test_polygon.py -> build\lib.win-amd64-3.11\skimage\measure\tests
+ copying skimage\measure\tests\test_profile.py -> build\lib.win-amd64-3.11\skimage\measure\tests
+ copying skimage\measure\tests\test_regionprops.py -> build\lib.win-amd64-3.11\skimage\measure\tests
+ copying skimage\measure\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\measure\tests
+ creating build\lib.win-amd64-3.11\skimage\metrics\tests
+ copying skimage\metrics\tests\test_segmentation_metrics.py -> build\lib.win-amd64-3.11\skimage\metrics\tests
+ copying skimage\metrics\tests\test_set_metrics.py -> build\lib.win-amd64-3.11\skimage\metrics\tests
+ copying skimage\metrics\tests\test_simple_metrics.py -> build\lib.win-amd64-3.11\skimage\metrics\tests
+ copying skimage\metrics\tests\test_structural_similarity.py -> build\lib.win-amd64-3.11\skimage\metrics\tests
+ copying skimage\metrics\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\metrics\tests
+ creating build\lib.win-amd64-3.11\skimage\morphology\tests
+ copying skimage\morphology\tests\test_binary.py -> build\lib.win-amd64-3.11\skimage\morphology\tests
+ copying skimage\morphology\tests\test_convex_hull.py -> build\lib.win-amd64-3.11\skimage\morphology\tests
+ copying skimage\morphology\tests\test_extrema.py -> build\lib.win-amd64-3.11\skimage\morphology\tests
+ copying skimage\morphology\tests\test_flood_fill.py -> build\lib.win-amd64-3.11\skimage\morphology\tests
+ copying skimage\morphology\tests\test_footprints.py -> build\lib.win-amd64-3.11\skimage\morphology\tests
+ copying skimage\morphology\tests\test_gray.py -> build\lib.win-amd64-3.11\skimage\morphology\tests
+ copying skimage\morphology\tests\test_max_tree.py -> build\lib.win-amd64-3.11\skimage\morphology\tests
+ copying skimage\morphology\tests\test_misc.py -> build\lib.win-amd64-3.11\skimage\morphology\tests
+ copying skimage\morphology\tests\test_reconstruction.py -> build\lib.win-amd64-3.11\skimage\morphology\tests
+ copying skimage\morphology\tests\test_skeletonize.py -> build\lib.win-amd64-3.11\skimage\morphology\tests
+ copying skimage\morphology\tests\test_skeletonize_3d.py -> build\lib.win-amd64-3.11\skimage\morphology\tests
+ copying skimage\morphology\tests\test_util.py -> build\lib.win-amd64-3.11\skimage\morphology\tests
+ copying skimage\morphology\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\morphology\tests
+ creating build\lib.win-amd64-3.11\skimage\registration\tests
+ copying skimage\registration\tests\test_ilk.py -> build\lib.win-amd64-3.11\skimage\registration\tests
+ copying skimage\registration\tests\test_masked_phase_cross_correlation.py -> build\lib.win-amd64-3.11\skimage\registration\tests
+ copying skimage\registration\tests\test_phase_cross_correlation.py -> build\lib.win-amd64-3.11\skimage\registration\tests
+ copying skimage\registration\tests\test_tvl1.py -> build\lib.win-amd64-3.11\skimage\registration\tests
+ copying skimage\registration\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\registration\tests
+ creating build\lib.win-amd64-3.11\skimage\restoration\tests
+ copying skimage\restoration\tests\test_denoise.py -> build\lib.win-amd64-3.11\skimage\restoration\tests
+ copying skimage\restoration\tests\test_inpaint.py -> build\lib.win-amd64-3.11\skimage\restoration\tests
+ copying skimage\restoration\tests\test_j_invariant.py -> build\lib.win-amd64-3.11\skimage\restoration\tests
+ copying skimage\restoration\tests\test_restoration.py -> build\lib.win-amd64-3.11\skimage\restoration\tests
+ copying skimage\restoration\tests\test_rolling_ball.py -> build\lib.win-amd64-3.11\skimage\restoration\tests
+ copying skimage\restoration\tests\test_unwrap.py -> build\lib.win-amd64-3.11\skimage\restoration\tests
+ copying skimage\restoration\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\restoration\tests
+ creating build\lib.win-amd64-3.11\skimage\segmentation\tests
+ copying skimage\segmentation\tests\test_active_contour_model.py -> build\lib.win-amd64-3.11\skimage\segmentation\tests
+ copying skimage\segmentation\tests\test_boundaries.py -> build\lib.win-amd64-3.11\skimage\segmentation\tests
+ copying skimage\segmentation\tests\test_chan_vese.py -> build\lib.win-amd64-3.11\skimage\segmentation\tests
+ copying skimage\segmentation\tests\test_clear_border.py -> build\lib.win-amd64-3.11\skimage\segmentation\tests
+ copying skimage\segmentation\tests\test_expand_labels.py -> build\lib.win-amd64-3.11\skimage\segmentation\tests
+ copying skimage\segmentation\tests\test_felzenszwalb.py -> build\lib.win-amd64-3.11\skimage\segmentation\tests
+ copying skimage\segmentation\tests\test_join.py -> build\lib.win-amd64-3.11\skimage\segmentation\tests
+ copying skimage\segmentation\tests\test_morphsnakes.py -> build\lib.win-amd64-3.11\skimage\segmentation\tests
+ copying skimage\segmentation\tests\test_quickshift.py -> build\lib.win-amd64-3.11\skimage\segmentation\tests
+ copying skimage\segmentation\tests\test_random_walker.py -> build\lib.win-amd64-3.11\skimage\segmentation\tests
+ copying skimage\segmentation\tests\test_slic.py -> build\lib.win-amd64-3.11\skimage\segmentation\tests
+ copying skimage\segmentation\tests\test_watershed.py -> build\lib.win-amd64-3.11\skimage\segmentation\tests
+ copying skimage\segmentation\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\segmentation\tests
+ creating build\lib.win-amd64-3.11\skimage\transform\tests
+ copying skimage\transform\tests\test_finite_radon_transform.py -> build\lib.win-amd64-3.11\skimage\transform\tests
+ copying skimage\transform\tests\test_geometric.py -> build\lib.win-amd64-3.11\skimage\transform\tests
+ copying skimage\transform\tests\test_hough_transform.py -> build\lib.win-amd64-3.11\skimage\transform\tests
+ copying skimage\transform\tests\test_integral.py -> build\lib.win-amd64-3.11\skimage\transform\tests
+ copying skimage\transform\tests\test_pyramids.py -> build\lib.win-amd64-3.11\skimage\transform\tests
+ copying skimage\transform\tests\test_radon_transform.py -> build\lib.win-amd64-3.11\skimage\transform\tests
+ copying skimage\transform\tests\test_warps.py -> build\lib.win-amd64-3.11\skimage\transform\tests
+ copying skimage\transform\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\transform\tests
+ creating build\lib.win-amd64-3.11\skimage\util\tests
+ copying skimage\util\tests\test_apply_parallel.py -> build\lib.win-amd64-3.11\skimage\util\tests
+ copying skimage\util\tests\test_arraycrop.py -> build\lib.win-amd64-3.11\skimage\util\tests
+ copying skimage\util\tests\test_compare.py -> build\lib.win-amd64-3.11\skimage\util\tests
+ copying skimage\util\tests\test_dtype.py -> build\lib.win-amd64-3.11\skimage\util\tests
+ copying skimage\util\tests\test_invert.py -> build\lib.win-amd64-3.11\skimage\util\tests
+ copying skimage\util\tests\test_labels.py -> build\lib.win-amd64-3.11\skimage\util\tests
+ copying skimage\util\tests\test_map_array.py -> build\lib.win-amd64-3.11\skimage\util\tests
+ copying skimage\util\tests\test_montage.py -> build\lib.win-amd64-3.11\skimage\util\tests
+ copying skimage\util\tests\test_random_noise.py -> build\lib.win-amd64-3.11\skimage\util\tests
+ copying skimage\util\tests\test_regular_grid.py -> build\lib.win-amd64-3.11\skimage\util\tests
+ copying skimage\util\tests\test_shape.py -> build\lib.win-amd64-3.11\skimage\util\tests
+ copying skimage\util\tests\test_unique_rows.py -> build\lib.win-amd64-3.11\skimage\util\tests
+ copying skimage\util\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\util\tests
+ creating build\lib.win-amd64-3.11\skimage\viewer\canvastools
+ copying skimage\viewer\canvastools\base.py -> build\lib.win-amd64-3.11\skimage\viewer\canvastools
+ copying skimage\viewer\canvastools\linetool.py -> build\lib.win-amd64-3.11\skimage\viewer\canvastools
+ copying skimage\viewer\canvastools\painttool.py -> build\lib.win-amd64-3.11\skimage\viewer\canvastools
+ copying skimage\viewer\canvastools\recttool.py -> build\lib.win-amd64-3.11\skimage\viewer\canvastools
+ copying skimage\viewer\canvastools\__init__.py -> build\lib.win-amd64-3.11\skimage\viewer\canvastools
+ creating build\lib.win-amd64-3.11\skimage\viewer\plugins
+ copying skimage\viewer\plugins\base.py -> build\lib.win-amd64-3.11\skimage\viewer\plugins
+ copying skimage\viewer\plugins\canny.py -> build\lib.win-amd64-3.11\skimage\viewer\plugins
+ copying skimage\viewer\plugins\color_histogram.py -> build\lib.win-amd64-3.11\skimage\viewer\plugins
+ copying skimage\viewer\plugins\crop.py -> build\lib.win-amd64-3.11\skimage\viewer\plugins
+ copying skimage\viewer\plugins\labelplugin.py -> build\lib.win-amd64-3.11\skimage\viewer\plugins
+ copying skimage\viewer\plugins\lineprofile.py -> build\lib.win-amd64-3.11\skimage\viewer\plugins
+ copying skimage\viewer\plugins\measure.py -> build\lib.win-amd64-3.11\skimage\viewer\plugins
+ copying skimage\viewer\plugins\overlayplugin.py -> build\lib.win-amd64-3.11\skimage\viewer\plugins
+ copying skimage\viewer\plugins\plotplugin.py -> build\lib.win-amd64-3.11\skimage\viewer\plugins
+ copying skimage\viewer\plugins\__init__.py -> build\lib.win-amd64-3.11\skimage\viewer\plugins
+ creating build\lib.win-amd64-3.11\skimage\viewer\tests
+ copying skimage\viewer\tests\test_plugins.py -> build\lib.win-amd64-3.11\skimage\viewer\tests
+ copying skimage\viewer\tests\test_tools.py -> build\lib.win-amd64-3.11\skimage\viewer\tests
+ copying skimage\viewer\tests\test_utils.py -> build\lib.win-amd64-3.11\skimage\viewer\tests
+ copying skimage\viewer\tests\test_viewer.py -> build\lib.win-amd64-3.11\skimage\viewer\tests
+ copying skimage\viewer\tests\test_widgets.py -> build\lib.win-amd64-3.11\skimage\viewer\tests
+ copying skimage\viewer\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\viewer\tests
+ creating build\lib.win-amd64-3.11\skimage\viewer\utils
+ copying skimage\viewer\utils\canvas.py -> build\lib.win-amd64-3.11\skimage\viewer\utils
+ copying skimage\viewer\utils\core.py -> build\lib.win-amd64-3.11\skimage\viewer\utils
+ copying skimage\viewer\utils\dialogs.py -> build\lib.win-amd64-3.11\skimage\viewer\utils
+ copying skimage\viewer\utils\__init__.py -> build\lib.win-amd64-3.11\skimage\viewer\utils
+ creating build\lib.win-amd64-3.11\skimage\viewer\viewers
+ copying skimage\viewer\viewers\core.py -> build\lib.win-amd64-3.11\skimage\viewer\viewers
+ copying skimage\viewer\viewers\__init__.py -> build\lib.win-amd64-3.11\skimage\viewer\viewers
+ creating build\lib.win-amd64-3.11\skimage\viewer\widgets
+ copying skimage\viewer\widgets\core.py -> build\lib.win-amd64-3.11\skimage\viewer\widgets
+ copying skimage\viewer\widgets\history.py -> build\lib.win-amd64-3.11\skimage\viewer\widgets
+ copying skimage\viewer\widgets\__init__.py -> build\lib.win-amd64-3.11\skimage\viewer\widgets
+ creating build\lib.win-amd64-3.11\skimage\_shared\tests
+ copying skimage\_shared\tests\test_coord.py -> build\lib.win-amd64-3.11\skimage\_shared\tests
+ copying skimage\_shared\tests\test_fast_exp.py -> build\lib.win-amd64-3.11\skimage\_shared\tests
+ copying skimage\_shared\tests\test_geometry.py -> build\lib.win-amd64-3.11\skimage\_shared\tests
+ copying skimage\_shared\tests\test_interpolation.py -> build\lib.win-amd64-3.11\skimage\_shared\tests
+ copying skimage\_shared\tests\test_safe_as_int.py -> build\lib.win-amd64-3.11\skimage\_shared\tests
+ copying skimage\_shared\tests\test_testing.py -> build\lib.win-amd64-3.11\skimage\_shared\tests
+ copying skimage\_shared\tests\test_utils.py -> build\lib.win-amd64-3.11\skimage\_shared\tests
+ copying skimage\_shared\tests\test_version_requirements.py -> build\lib.win-amd64-3.11\skimage\_shared\tests
+ copying skimage\_shared\tests\test_warnings.py -> build\lib.win-amd64-3.11\skimage\_shared\tests
+ copying skimage\_shared\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\_shared\tests
+ copying skimage\draw\_draw.pyx -> build\lib.win-amd64-3.11\skimage\draw
+ copying skimage\feature\censure_cy.pyx -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\corner_cy.pyx -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\orb_cy.pyx -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\_cascade.pyx -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\_haar.pyx -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\_hoghistogram.pyx -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\_sift.pyx -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\_texture.pyx -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\safe_openmp.pxd -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\_haar.pxd -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\_texture.pxd -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\filters\_multiotsu.pyx -> build\lib.win-amd64-3.11\skimage\filters
+ copying skimage\graph\heap.pyx -> build\lib.win-amd64-3.11\skimage\graph
+ copying skimage\graph\_mcp.pyx -> build\lib.win-amd64-3.11\skimage\graph
+ copying skimage\graph\_spath.pyx -> build\lib.win-amd64-3.11\skimage\graph
+ copying skimage\graph\heap.pxd -> build\lib.win-amd64-3.11\skimage\graph
+ copying skimage\graph\_mcp.pxd -> build\lib.win-amd64-3.11\skimage\graph
+ copying skimage\measure\_ccomp.pyx -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\_find_contours_cy.pyx -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\_marching_cubes_classic_cy.pyx -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\_marching_cubes_lewiner_cy.pyx -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\_moments_cy.pyx -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\_pnpoly.pyx -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\_ccomp.pxd -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\morphology\_convex_hull.pyx -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\_extrema_cy.pyx -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\_flood_fill_cy.pyx -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\_grayreconstruct.pyx -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\_max_tree.pyx -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\_skeletonize_3d_cy.pyx -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\_skeletonize_cy.pyx -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\_queue_with_history.pxi -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\restoration\_denoise_cy.pyx -> build\lib.win-amd64-3.11\skimage\restoration
+ copying skimage\restoration\_inpaint.pyx -> build\lib.win-amd64-3.11\skimage\restoration
+ copying skimage\restoration\_nl_means_denoising.pyx -> build\lib.win-amd64-3.11\skimage\restoration
+ copying skimage\restoration\_rolling_ball_cy.pyx -> build\lib.win-amd64-3.11\skimage\restoration
+ copying skimage\restoration\_unwrap_1d.pyx -> build\lib.win-amd64-3.11\skimage\restoration
+ copying skimage\restoration\_unwrap_2d.pyx -> build\lib.win-amd64-3.11\skimage\restoration
+ copying skimage\restoration\_unwrap_3d.pyx -> build\lib.win-amd64-3.11\skimage\restoration
+ copying skimage\segmentation\_felzenszwalb_cy.pyx -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\_quickshift_cy.pyx -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\_slic.pyx -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\_watershed_cy.pyx -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\heap_general.pxi -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\heap_watershed.pxi -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\transform\_hough_transform.pyx -> build\lib.win-amd64-3.11\skimage\transform
+ copying skimage\transform\_radon_transform.pyx -> build\lib.win-amd64-3.11\skimage\transform
+ copying skimage\transform\_warps_cy.pyx -> build\lib.win-amd64-3.11\skimage\transform
+ copying skimage\util\_remap.pyx -> build\lib.win-amd64-3.11\skimage\util
+ copying skimage\_shared\fast_exp.pyx -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\geometry.pyx -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\interpolation.pyx -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\transform.pyx -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\fast_exp.pxd -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\fused_numerics.pxd -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\geometry.pxd -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\interpolation.pxd -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\transform.pxd -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\filters\rank\bilateral_cy.pyx -> build\lib.win-amd64-3.11\skimage\filters\rank
+ copying skimage\filters\rank\core_cy.pyx -> build\lib.win-amd64-3.11\skimage\filters\rank
+ copying skimage\filters\rank\core_cy_3d.pyx -> build\lib.win-amd64-3.11\skimage\filters\rank
+ copying skimage\filters\rank\generic_cy.pyx -> build\lib.win-amd64-3.11\skimage\filters\rank
+ copying skimage\filters\rank\percentile_cy.pyx -> build\lib.win-amd64-3.11\skimage\filters\rank
+ copying skimage\filters\rank\core_cy.pxd -> build\lib.win-amd64-3.11\skimage\filters\rank
+ copying skimage\filters\rank\core_cy_3d.pxd -> build\lib.win-amd64-3.11\skimage\filters\rank
+ copying skimage\future\graph\_ncut_cy.pyx -> build\lib.win-amd64-3.11\skimage\future\graph
+ copying skimage\io\_plugins\_colormixer.pyx -> build\lib.win-amd64-3.11\skimage\io\_plugins
+ copying skimage\io\_plugins\_histograms.pyx -> build\lib.win-amd64-3.11\skimage\io\_plugins
+ running build_ext
+ INFO: No module named 'numpy.distutils._msvccompiler' in numpy.distutils; trying from distutils
+ INFO: customize MSVCCompiler
+ INFO: customize MSVCCompiler using ConditionalOpenMP
+ INFO: CCompilerOpt.cc_test_flags[1077] : testing flags (/O2)
+ error: Microsoft Visual C++ 14.0 or greater is required. Get it with ""Microsoft C++ Build Tools"": https://visualstudio.microsoft.com/visual-cpp-build-tools/
+ INFO: CCompilerOpt.cache_flush[857] : write cache to path -> C:\Users\jay\AppData\Local\Temp\pip-install-s2xz2koj\scikit-image_cad4b2fea6c2495988cddf5b95ba2806\build\temp.win-amd64-3.11\Release\ccompiler_opt_cache_ext.py
+ [end of output]
+
+ note: This error originates from a subprocess, and is likely not a problem with pip.
+ ERROR: Failed building wheel for scikit-image
+Failed to build scikit-image
+ERROR: Could not build wheels for scikit-image, which is required to install pyproject.toml-based projects
+
+C:\Windows\System32>source skimage-dev/bin/activate
+'source' is not recognized as an internal or external command,
+operable program or batch file.
+```
+
+
+",1.0,"Could not build wheels for scikit-image - Hello, I have been trying to install this Python module for a number of days now as it is a dependency of img2database. I'm not completely sure what is causing the problem. Hence I have been unable to try a large variety of fixes. I believe the problem might potentially have something to do with pip as the module installed successfully using mini conda however the module I wish to use is only available on pip. Any help would be greatly appreciated.
+The full output is as follows:
+
+
+Click to expand
+
+```
+C:\Windows\System32>pip install scikit-image
+Collecting scikit-image
+ Using cached scikit-image-0.19.3.tar.gz (22.2 MB)
+ Installing build dependencies ... done
+ Getting requirements to build wheel ... done
+ Preparing metadata (pyproject.toml) ... done
+Requirement already satisfied: numpy>=1.17.0 in c:\users\jay\appdata\local\programs\python\python311\lib\site-packages (from scikit-image) (1.24.2)
+Requirement already satisfied: scipy>=1.4.1 in c:\users\jay\appdata\local\programs\python\python311\lib\site-packages (from scikit-image) (1.10.1)
+Requirement already satisfied: networkx>=2.2 in c:\users\jay\appdata\local\programs\python\python311\lib\site-packages (from scikit-image) (3.0)
+Requirement already satisfied: pillow!=7.1.0,!=7.1.1,!=8.3.0,>=6.1.0 in c:\users\jay\appdata\local\programs\python\python311\lib\site-packages (from scikit-image) (9.4.0)
+Collecting imageio>=2.4.1
+ Using cached imageio-2.25.1-py3-none-any.whl (3.4 MB)
+Collecting tifffile>=2019.7.26
+ Using cached tifffile-2023.2.3-py3-none-any.whl (215 kB)
+Collecting PyWavelets>=1.1.1
+ Using cached PyWavelets-1.4.1-cp311-cp311-win_amd64.whl (4.2 MB)
+Requirement already satisfied: packaging>=20.0 in c:\users\jay\appdata\local\programs\python\python311\lib\site-packages (from scikit-image) (23.0)
+Building wheels for collected packages: scikit-image
+ Building wheel for scikit-image (pyproject.toml) ... error
+ error: subprocess-exited-with-error
+
+ × Building wheel for scikit-image (pyproject.toml) did not run successfully.
+ │ exit code: 1
+ ╰─> [626 lines of output]
+ setup.py:9: DeprecationWarning:
+
+ `numpy.distutils` is deprecated since NumPy 1.23.0, as a result
+ of the deprecation of `distutils` itself. It will be removed for
+ Python >= 3.12. For older Python versions it will remain present.
+ It is recommended to use `setuptools < 60.0` for those Python versions.
+ For more details, see:
+ https://numpy.org/devdocs/reference/distutils_status_migration.html
+
+
+ from numpy.distutils.command.build_ext import build_ext as npy_build_ext
+ Partial import of skimage during the build process.
+ Compiling C:\Users\jay\AppData\Local\Temp\pip-install-s2xz2koj\scikit-image_cad4b2fea6c2495988cddf5b95ba2806\skimage\morphology\_skeletonize_3d_cy.pyx because it changed.
+ [1/1] Cythonizing C:\Users\jay\AppData\Local\Temp\pip-install-s2xz2koj\scikit-image_cad4b2fea6c2495988cddf5b95ba2806\skimage\morphology\_skeletonize_3d_cy.pyx
+ running bdist_wheel
+ running build
+ running config_cc
+ INFO: unifing config_cc, config, build_clib, build_ext, build commands --compiler options
+ running config_fc
+ INFO: unifing config_fc, config, build_clib, build_ext, build commands --fcompiler options
+ running build_src
+ INFO: build_src
+ INFO: building extension ""skimage._shared.geometry"" sources
+ INFO: building extension ""skimage._shared.transform"" sources
+ INFO: building extension ""skimage._shared.interpolation"" sources
+ INFO: building extension ""skimage._shared.fast_exp"" sources
+ INFO: building extension ""skimage.draw._draw"" sources
+ INFO: building extension ""skimage.feature._cascade"" sources
+ INFO: building extension ""skimage.feature.corner_cy"" sources
+ INFO: building extension ""skimage.feature.censure_cy"" sources
+ INFO: building extension ""skimage.feature.orb_cy"" sources
+ INFO: building extension ""skimage.feature._texture"" sources
+ INFO: building extension ""skimage.feature._hoghistogram"" sources
+ INFO: building extension ""skimage.feature._haar"" sources
+ INFO: building extension ""skimage.feature._sift"" sources
+ INFO: building extension ""skimage.feature.brief_cy"" sources
+ INFO: building extension ""skimage.feature._hessian_det_appx"" sources
+ INFO: building extension ""skimage.restoration._unwrap_1d"" sources
+ INFO: building extension ""skimage.restoration._unwrap_2d"" sources
+ INFO: building extension ""skimage.restoration._unwrap_3d"" sources
+ INFO: building extension ""skimage.restoration._denoise_cy"" sources
+ INFO: building extension ""skimage.restoration._nl_means_denoising"" sources
+ INFO: building extension ""skimage.restoration._rolling_ball_cy"" sources
+ INFO: building extension ""skimage.restoration._inpaint"" sources
+ INFO: building extension ""skimage.filters.rank.core_cy"" sources
+ INFO: building extension ""skimage.filters.rank.core_cy_3d"" sources
+ INFO: building extension ""skimage.filters._multiotsu"" sources
+ INFO: building extension ""skimage.filters.rank.generic_cy"" sources
+ INFO: building extension ""skimage.filters.rank.percentile_cy"" sources
+ INFO: building extension ""skimage.filters.rank.bilateral_cy"" sources
+ INFO: building extension ""skimage.future.graph._ncut_cy"" sources
+ INFO: building extension ""skimage.graph._spath"" sources
+ INFO: building extension ""skimage.graph._mcp"" sources
+ INFO: building extension ""skimage.graph.heap"" sources
+ INFO: building extension ""skimage.io._plugins._colormixer"" sources
+ INFO: building extension ""skimage.io._plugins._histograms"" sources
+ INFO: building extension ""skimage.measure._ccomp"" sources
+ INFO: building extension ""skimage.measure._find_contours_cy"" sources
+ INFO: building extension ""skimage.measure._moments_cy"" sources
+ INFO: building extension ""skimage.measure._marching_cubes_classic_cy"" sources
+ INFO: building extension ""skimage.measure._marching_cubes_lewiner_cy"" sources
+ INFO: building extension ""skimage.measure._pnpoly"" sources
+ INFO: building extension ""skimage.morphology._skeletonize_cy"" sources
+ INFO: building extension ""skimage.morphology._convex_hull"" sources
+ INFO: building extension ""skimage.morphology._grayreconstruct"" sources
+ INFO: building extension ""skimage.morphology._max_tree"" sources
+ INFO: building extension ""skimage.morphology._skeletonize_3d_cy"" sources
+ INFO: building extension ""skimage.morphology._extrema_cy"" sources
+ INFO: building extension ""skimage.morphology._flood_fill_cy"" sources
+ INFO: building extension ""skimage.transform._hough_transform"" sources
+ INFO: building extension ""skimage.transform._warps_cy"" sources
+ INFO: building extension ""skimage.transform._radon_transform"" sources
+ INFO: building extension ""skimage.util._remap"" sources
+ INFO: building extension ""skimage.segmentation._watershed_cy"" sources
+ INFO: building extension ""skimage.segmentation._felzenszwalb_cy"" sources
+ INFO: building extension ""skimage.segmentation._quickshift_cy"" sources
+ INFO: building extension ""skimage.segmentation._slic"" sources
+ INFO: building data_files sources
+ INFO: build_src: building npy-pkg config files
+ running build_py
+ creating build
+ creating build\lib.win-amd64-3.11
+ creating build\lib.win-amd64-3.11\skimage
+ copying skimage\conftest.py -> build\lib.win-amd64-3.11\skimage
+ copying skimage\setup.py -> build\lib.win-amd64-3.11\skimage
+ copying skimage\_build.py -> build\lib.win-amd64-3.11\skimage
+ copying skimage\__init__.py -> build\lib.win-amd64-3.11\skimage
+ creating build\lib.win-amd64-3.11\doc
+ creating build\lib.win-amd64-3.11\doc\ext
+ copying doc\ext\doi_role.py -> build\lib.win-amd64-3.11\doc\ext
+ copying doc\ext\notebook_doc.py -> build\lib.win-amd64-3.11\doc\ext
+ copying doc\ext\plot2rst.py -> build\lib.win-amd64-3.11\doc\ext
+ copying doc\ext\__init__.py -> build\lib.win-amd64-3.11\doc\ext
+ creating build\lib.win-amd64-3.11\doc\ext\tests
+ copying doc\ext\tests\test_notebook_doc.py -> build\lib.win-amd64-3.11\doc\ext\tests
+ copying doc\ext\tests\__init__.py -> build\lib.win-amd64-3.11\doc\ext\tests
+ creating build\lib.win-amd64-3.11\skimage\color
+ copying skimage\color\adapt_rgb.py -> build\lib.win-amd64-3.11\skimage\color
+ copying skimage\color\colorconv.py -> build\lib.win-amd64-3.11\skimage\color
+ copying skimage\color\colorlabel.py -> build\lib.win-amd64-3.11\skimage\color
+ copying skimage\color\delta_e.py -> build\lib.win-amd64-3.11\skimage\color
+ copying skimage\color\rgb_colors.py -> build\lib.win-amd64-3.11\skimage\color
+ copying skimage\color\__init__.py -> build\lib.win-amd64-3.11\skimage\color
+ creating build\lib.win-amd64-3.11\skimage\data
+ copying skimage\data\setup.py -> build\lib.win-amd64-3.11\skimage\data
+ copying skimage\data\_binary_blobs.py -> build\lib.win-amd64-3.11\skimage\data
+ copying skimage\data\_fetchers.py -> build\lib.win-amd64-3.11\skimage\data
+ copying skimage\data\_registry.py -> build\lib.win-amd64-3.11\skimage\data
+ copying skimage\data\__init__.py -> build\lib.win-amd64-3.11\skimage\data
+ creating build\lib.win-amd64-3.11\skimage\draw
+ copying skimage\draw\draw.py -> build\lib.win-amd64-3.11\skimage\draw
+ copying skimage\draw\draw3d.py -> build\lib.win-amd64-3.11\skimage\draw
+ copying skimage\draw\draw_nd.py -> build\lib.win-amd64-3.11\skimage\draw
+ copying skimage\draw\setup.py -> build\lib.win-amd64-3.11\skimage\draw
+ copying skimage\draw\_polygon2mask.py -> build\lib.win-amd64-3.11\skimage\draw
+ copying skimage\draw\_random_shapes.py -> build\lib.win-amd64-3.11\skimage\draw
+ copying skimage\draw\__init__.py -> build\lib.win-amd64-3.11\skimage\draw
+ creating build\lib.win-amd64-3.11\skimage\exposure
+ copying skimage\exposure\exposure.py -> build\lib.win-amd64-3.11\skimage\exposure
+ copying skimage\exposure\histogram_matching.py -> build\lib.win-amd64-3.11\skimage\exposure
+ copying skimage\exposure\setup.py -> build\lib.win-amd64-3.11\skimage\exposure
+ copying skimage\exposure\_adapthist.py -> build\lib.win-amd64-3.11\skimage\exposure
+ copying skimage\exposure\__init__.py -> build\lib.win-amd64-3.11\skimage\exposure
+ creating build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\blob.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\brief.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\brief_pythran.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\censure.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\corner.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\haar.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\match.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\orb.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\peak.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\setup.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\sift.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\template.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\texture.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\util.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\_basic_features.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\_canny.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\_daisy.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\_hessian_det_appx_pythran.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\_hog.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\_orb_descriptor_positions.py -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\__init__.py -> build\lib.win-amd64-3.11\skimage\feature
+ creating build\lib.win-amd64-3.11\skimage\filters
+ copying skimage\filters\edges.py -> build\lib.win-amd64-3.11\skimage\filters
+ copying skimage\filters\lpi_filter.py -> build\lib.win-amd64-3.11\skimage\filters
+ copying skimage\filters\ridges.py -> build\lib.win-amd64-3.11\skimage\filters
+ copying skimage\filters\setup.py -> build\lib.win-amd64-3.11\skimage\filters
+ copying skimage\filters\thresholding.py -> build\lib.win-amd64-3.11\skimage\filters
+ copying skimage\filters\_fft_based.py -> build\lib.win-amd64-3.11\skimage\filters
+ copying skimage\filters\_gabor.py -> build\lib.win-amd64-3.11\skimage\filters
+ copying skimage\filters\_gaussian.py -> build\lib.win-amd64-3.11\skimage\filters
+ copying skimage\filters\_median.py -> build\lib.win-amd64-3.11\skimage\filters
+ copying skimage\filters\_rank_order.py -> build\lib.win-amd64-3.11\skimage\filters
+ copying skimage\filters\_sparse.py -> build\lib.win-amd64-3.11\skimage\filters
+ copying skimage\filters\_unsharp_mask.py -> build\lib.win-amd64-3.11\skimage\filters
+ copying skimage\filters\_window.py -> build\lib.win-amd64-3.11\skimage\filters
+ copying skimage\filters\__init__.py -> build\lib.win-amd64-3.11\skimage\filters
+ creating build\lib.win-amd64-3.11\skimage\future
+ copying skimage\future\manual_segmentation.py -> build\lib.win-amd64-3.11\skimage\future
+ copying skimage\future\setup.py -> build\lib.win-amd64-3.11\skimage\future
+ copying skimage\future\trainable_segmentation.py -> build\lib.win-amd64-3.11\skimage\future
+ copying skimage\future\__init__.py -> build\lib.win-amd64-3.11\skimage\future
+ creating build\lib.win-amd64-3.11\skimage\graph
+ copying skimage\graph\mcp.py -> build\lib.win-amd64-3.11\skimage\graph
+ copying skimage\graph\setup.py -> build\lib.win-amd64-3.11\skimage\graph
+ copying skimage\graph\spath.py -> build\lib.win-amd64-3.11\skimage\graph
+ copying skimage\graph\_graph.py -> build\lib.win-amd64-3.11\skimage\graph
+ copying skimage\graph\__init__.py -> build\lib.win-amd64-3.11\skimage\graph
+ creating build\lib.win-amd64-3.11\skimage\io
+ copying skimage\io\collection.py -> build\lib.win-amd64-3.11\skimage\io
+ copying skimage\io\manage_plugins.py -> build\lib.win-amd64-3.11\skimage\io
+ copying skimage\io\setup.py -> build\lib.win-amd64-3.11\skimage\io
+ copying skimage\io\sift.py -> build\lib.win-amd64-3.11\skimage\io
+ copying skimage\io\util.py -> build\lib.win-amd64-3.11\skimage\io
+ copying skimage\io\_image_stack.py -> build\lib.win-amd64-3.11\skimage\io
+ copying skimage\io\_io.py -> build\lib.win-amd64-3.11\skimage\io
+ copying skimage\io\__init__.py -> build\lib.win-amd64-3.11\skimage\io
+ creating build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\block.py -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\entropy.py -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\fit.py -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\pnpoly.py -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\profile.py -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\setup.py -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\_blur_effect.py -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\_find_contours.py -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\_label.py -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\_marching_cubes_classic.py -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\_marching_cubes_lewiner.py -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\_marching_cubes_lewiner_luts.py -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\_moments.py -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\_polygon.py -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\_regionprops.py -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\_regionprops_utils.py -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\__init__.py -> build\lib.win-amd64-3.11\skimage\measure
+ creating build\lib.win-amd64-3.11\skimage\metrics
+ copying skimage\metrics\set_metrics.py -> build\lib.win-amd64-3.11\skimage\metrics
+ copying skimage\metrics\simple_metrics.py -> build\lib.win-amd64-3.11\skimage\metrics
+ copying skimage\metrics\_adapted_rand_error.py -> build\lib.win-amd64-3.11\skimage\metrics
+ copying skimage\metrics\_contingency_table.py -> build\lib.win-amd64-3.11\skimage\metrics
+ copying skimage\metrics\_structural_similarity.py -> build\lib.win-amd64-3.11\skimage\metrics
+ copying skimage\metrics\_variation_of_information.py -> build\lib.win-amd64-3.11\skimage\metrics
+ copying skimage\metrics\__init__.py -> build\lib.win-amd64-3.11\skimage\metrics
+ creating build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\binary.py -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\convex_hull.py -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\extrema.py -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\footprints.py -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\gray.py -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\grayreconstruct.py -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\grey.py -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\greyreconstruct.py -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\max_tree.py -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\misc.py -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\selem.py -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\setup.py -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\_flood_fill.py -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\_skeletonize.py -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\_util.py -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\__init__.py -> build\lib.win-amd64-3.11\skimage\morphology
+ creating build\lib.win-amd64-3.11\skimage\registration
+ copying skimage\registration\_masked_phase_cross_correlation.py -> build\lib.win-amd64-3.11\skimage\registration
+ copying skimage\registration\_optical_flow.py -> build\lib.win-amd64-3.11\skimage\registration
+ copying skimage\registration\_optical_flow_utils.py -> build\lib.win-amd64-3.11\skimage\registration
+ copying skimage\registration\_phase_cross_correlation.py -> build\lib.win-amd64-3.11\skimage\registration
+ copying skimage\registration\__init__.py -> build\lib.win-amd64-3.11\skimage\registration
+ creating build\lib.win-amd64-3.11\skimage\restoration
+ copying skimage\restoration\deconvolution.py -> build\lib.win-amd64-3.11\skimage\restoration
+ copying skimage\restoration\inpaint.py -> build\lib.win-amd64-3.11\skimage\restoration
+ copying skimage\restoration\j_invariant.py -> build\lib.win-amd64-3.11\skimage\restoration
+ copying skimage\restoration\non_local_means.py -> build\lib.win-amd64-3.11\skimage\restoration
+ copying skimage\restoration\rolling_ball.py -> build\lib.win-amd64-3.11\skimage\restoration
+ copying skimage\restoration\setup.py -> build\lib.win-amd64-3.11\skimage\restoration
+ copying skimage\restoration\uft.py -> build\lib.win-amd64-3.11\skimage\restoration
+ copying skimage\restoration\unwrap.py -> build\lib.win-amd64-3.11\skimage\restoration
+ copying skimage\restoration\_cycle_spin.py -> build\lib.win-amd64-3.11\skimage\restoration
+ copying skimage\restoration\_denoise.py -> build\lib.win-amd64-3.11\skimage\restoration
+ copying skimage\restoration\__init__.py -> build\lib.win-amd64-3.11\skimage\restoration
+ creating build\lib.win-amd64-3.11\skimage\scripts
+ copying skimage\scripts\skivi.py -> build\lib.win-amd64-3.11\skimage\scripts
+ copying skimage\scripts\__init__.py -> build\lib.win-amd64-3.11\skimage\scripts
+ creating build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\active_contour_model.py -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\boundaries.py -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\morphsnakes.py -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\random_walker_segmentation.py -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\setup.py -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\slic_superpixels.py -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\_chan_vese.py -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\_clear_border.py -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\_expand_labels.py -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\_felzenszwalb.py -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\_join.py -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\_quickshift.py -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\_watershed.py -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\__init__.py -> build\lib.win-amd64-3.11\skimage\segmentation
+ creating build\lib.win-amd64-3.11\skimage\transform
+ copying skimage\transform\finite_radon_transform.py -> build\lib.win-amd64-3.11\skimage\transform
+ copying skimage\transform\hough_transform.py -> build\lib.win-amd64-3.11\skimage\transform
+ copying skimage\transform\integral.py -> build\lib.win-amd64-3.11\skimage\transform
+ copying skimage\transform\pyramids.py -> build\lib.win-amd64-3.11\skimage\transform
+ copying skimage\transform\radon_transform.py -> build\lib.win-amd64-3.11\skimage\transform
+ copying skimage\transform\setup.py -> build\lib.win-amd64-3.11\skimage\transform
+ copying skimage\transform\_geometric.py -> build\lib.win-amd64-3.11\skimage\transform
+ copying skimage\transform\_warps.py -> build\lib.win-amd64-3.11\skimage\transform
+ copying skimage\transform\__init__.py -> build\lib.win-amd64-3.11\skimage\transform
+ creating build\lib.win-amd64-3.11\skimage\util
+ copying skimage\util\apply_parallel.py -> build\lib.win-amd64-3.11\skimage\util
+ copying skimage\util\arraycrop.py -> build\lib.win-amd64-3.11\skimage\util
+ copying skimage\util\compare.py -> build\lib.win-amd64-3.11\skimage\util
+ copying skimage\util\dtype.py -> build\lib.win-amd64-3.11\skimage\util
+ copying skimage\util\lookfor.py -> build\lib.win-amd64-3.11\skimage\util
+ copying skimage\util\noise.py -> build\lib.win-amd64-3.11\skimage\util
+ copying skimage\util\setup.py -> build\lib.win-amd64-3.11\skimage\util
+ copying skimage\util\shape.py -> build\lib.win-amd64-3.11\skimage\util
+ copying skimage\util\unique.py -> build\lib.win-amd64-3.11\skimage\util
+ copying skimage\util\_invert.py -> build\lib.win-amd64-3.11\skimage\util
+ copying skimage\util\_label.py -> build\lib.win-amd64-3.11\skimage\util
+ copying skimage\util\_map_array.py -> build\lib.win-amd64-3.11\skimage\util
+ copying skimage\util\_montage.py -> build\lib.win-amd64-3.11\skimage\util
+ copying skimage\util\_regular_grid.py -> build\lib.win-amd64-3.11\skimage\util
+ copying skimage\util\__init__.py -> build\lib.win-amd64-3.11\skimage\util
+ creating build\lib.win-amd64-3.11\skimage\viewer
+ copying skimage\viewer\qt.py -> build\lib.win-amd64-3.11\skimage\viewer
+ copying skimage\viewer\__init__.py -> build\lib.win-amd64-3.11\skimage\viewer
+ creating build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\coord.py -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\filters.py -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\lazy.py -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\setup.py -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\tester.py -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\testing.py -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\utils.py -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\version_requirements.py -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\_dependency_checks.py -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\_geometry.py -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\_tempfile.py -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\_warnings.py -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\__init__.py -> build\lib.win-amd64-3.11\skimage\_shared
+ creating build\lib.win-amd64-3.11\skimage\color\tests
+ copying skimage\color\tests\test_adapt_rgb.py -> build\lib.win-amd64-3.11\skimage\color\tests
+ copying skimage\color\tests\test_colorconv.py -> build\lib.win-amd64-3.11\skimage\color\tests
+ copying skimage\color\tests\test_colorlabel.py -> build\lib.win-amd64-3.11\skimage\color\tests
+ copying skimage\color\tests\test_delta_e.py -> build\lib.win-amd64-3.11\skimage\color\tests
+ copying skimage\color\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\color\tests
+ creating build\lib.win-amd64-3.11\skimage\data\tests
+ copying skimage\data\tests\test_data.py -> build\lib.win-amd64-3.11\skimage\data\tests
+ copying skimage\data\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\data\tests
+ creating build\lib.win-amd64-3.11\skimage\draw\tests
+ copying skimage\draw\tests\test_draw.py -> build\lib.win-amd64-3.11\skimage\draw\tests
+ copying skimage\draw\tests\test_draw3d.py -> build\lib.win-amd64-3.11\skimage\draw\tests
+ copying skimage\draw\tests\test_draw_nd.py -> build\lib.win-amd64-3.11\skimage\draw\tests
+ copying skimage\draw\tests\test_polygon2mask.py -> build\lib.win-amd64-3.11\skimage\draw\tests
+ copying skimage\draw\tests\test_random_shapes.py -> build\lib.win-amd64-3.11\skimage\draw\tests
+ copying skimage\draw\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\draw\tests
+ creating build\lib.win-amd64-3.11\skimage\exposure\tests
+ copying skimage\exposure\tests\test_exposure.py -> build\lib.win-amd64-3.11\skimage\exposure\tests
+ copying skimage\exposure\tests\test_histogram_matching.py -> build\lib.win-amd64-3.11\skimage\exposure\tests
+ copying skimage\exposure\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\exposure\tests
+ creating build\lib.win-amd64-3.11\skimage\feature\tests
+ copying skimage\feature\tests\test_basic_features.py -> build\lib.win-amd64-3.11\skimage\feature\tests
+ copying skimage\feature\tests\test_blob.py -> build\lib.win-amd64-3.11\skimage\feature\tests
+ copying skimage\feature\tests\test_brief.py -> build\lib.win-amd64-3.11\skimage\feature\tests
+ copying skimage\feature\tests\test_canny.py -> build\lib.win-amd64-3.11\skimage\feature\tests
+ copying skimage\feature\tests\test_cascade.py -> build\lib.win-amd64-3.11\skimage\feature\tests
+ copying skimage\feature\tests\test_censure.py -> build\lib.win-amd64-3.11\skimage\feature\tests
+ copying skimage\feature\tests\test_corner.py -> build\lib.win-amd64-3.11\skimage\feature\tests
+ copying skimage\feature\tests\test_daisy.py -> build\lib.win-amd64-3.11\skimage\feature\tests
+ copying skimage\feature\tests\test_haar.py -> build\lib.win-amd64-3.11\skimage\feature\tests
+ copying skimage\feature\tests\test_hog.py -> build\lib.win-amd64-3.11\skimage\feature\tests
+ copying skimage\feature\tests\test_match.py -> build\lib.win-amd64-3.11\skimage\feature\tests
+ copying skimage\feature\tests\test_orb.py -> build\lib.win-amd64-3.11\skimage\feature\tests
+ copying skimage\feature\tests\test_peak.py -> build\lib.win-amd64-3.11\skimage\feature\tests
+ copying skimage\feature\tests\test_sift.py -> build\lib.win-amd64-3.11\skimage\feature\tests
+ copying skimage\feature\tests\test_template.py -> build\lib.win-amd64-3.11\skimage\feature\tests
+ copying skimage\feature\tests\test_texture.py -> build\lib.win-amd64-3.11\skimage\feature\tests
+ copying skimage\feature\tests\test_util.py -> build\lib.win-amd64-3.11\skimage\feature\tests
+ copying skimage\feature\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\feature\tests
+ creating build\lib.win-amd64-3.11\skimage\filters\rank
+ copying skimage\filters\rank\bilateral.py -> build\lib.win-amd64-3.11\skimage\filters\rank
+ copying skimage\filters\rank\generic.py -> build\lib.win-amd64-3.11\skimage\filters\rank
+ copying skimage\filters\rank\_percentile.py -> build\lib.win-amd64-3.11\skimage\filters\rank
+ copying skimage\filters\rank\__init__.py -> build\lib.win-amd64-3.11\skimage\filters\rank
+ creating build\lib.win-amd64-3.11\skimage\filters\tests
+ copying skimage\filters\tests\test_correlate.py -> build\lib.win-amd64-3.11\skimage\filters\tests
+ copying skimage\filters\tests\test_edges.py -> build\lib.win-amd64-3.11\skimage\filters\tests
+ copying skimage\filters\tests\test_fft_based.py -> build\lib.win-amd64-3.11\skimage\filters\tests
+ copying skimage\filters\tests\test_gabor.py -> build\lib.win-amd64-3.11\skimage\filters\tests
+ copying skimage\filters\tests\test_gaussian.py -> build\lib.win-amd64-3.11\skimage\filters\tests
+ copying skimage\filters\tests\test_lpi_filter.py -> build\lib.win-amd64-3.11\skimage\filters\tests
+ copying skimage\filters\tests\test_median.py -> build\lib.win-amd64-3.11\skimage\filters\tests
+ copying skimage\filters\tests\test_ridges.py -> build\lib.win-amd64-3.11\skimage\filters\tests
+ copying skimage\filters\tests\test_thresholding.py -> build\lib.win-amd64-3.11\skimage\filters\tests
+ copying skimage\filters\tests\test_unsharp_mask.py -> build\lib.win-amd64-3.11\skimage\filters\tests
+ copying skimage\filters\tests\test_window.py -> build\lib.win-amd64-3.11\skimage\filters\tests
+ copying skimage\filters\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\filters\tests
+ creating build\lib.win-amd64-3.11\skimage\filters\rank\tests
+ copying skimage\filters\rank\tests\test_rank.py -> build\lib.win-amd64-3.11\skimage\filters\rank\tests
+ copying skimage\filters\rank\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\filters\rank\tests
+ creating build\lib.win-amd64-3.11\skimage\future\graph
+ copying skimage\future\graph\graph_cut.py -> build\lib.win-amd64-3.11\skimage\future\graph
+ copying skimage\future\graph\graph_merge.py -> build\lib.win-amd64-3.11\skimage\future\graph
+ copying skimage\future\graph\rag.py -> build\lib.win-amd64-3.11\skimage\future\graph
+ copying skimage\future\graph\setup.py -> build\lib.win-amd64-3.11\skimage\future\graph
+ copying skimage\future\graph\_ncut.py -> build\lib.win-amd64-3.11\skimage\future\graph
+ copying skimage\future\graph\__init__.py -> build\lib.win-amd64-3.11\skimage\future\graph
+ creating build\lib.win-amd64-3.11\skimage\future\tests
+ copying skimage\future\tests\test_trainable_segmentation.py -> build\lib.win-amd64-3.11\skimage\future\tests
+ copying skimage\future\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\future\tests
+ creating build\lib.win-amd64-3.11\skimage\future\graph\tests
+ copying skimage\future\graph\tests\test_rag.py -> build\lib.win-amd64-3.11\skimage\future\graph\tests
+ copying skimage\future\graph\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\future\graph\tests
+ creating build\lib.win-amd64-3.11\skimage\graph\tests
+ copying skimage\graph\tests\test_anisotropy.py -> build\lib.win-amd64-3.11\skimage\graph\tests
+ copying skimage\graph\tests\test_connect.py -> build\lib.win-amd64-3.11\skimage\graph\tests
+ copying skimage\graph\tests\test_flexible.py -> build\lib.win-amd64-3.11\skimage\graph\tests
+ copying skimage\graph\tests\test_heap.py -> build\lib.win-amd64-3.11\skimage\graph\tests
+ copying skimage\graph\tests\test_mcp.py -> build\lib.win-amd64-3.11\skimage\graph\tests
+ copying skimage\graph\tests\test_pixel_graph.py -> build\lib.win-amd64-3.11\skimage\graph\tests
+ copying skimage\graph\tests\test_spath.py -> build\lib.win-amd64-3.11\skimage\graph\tests
+ copying skimage\graph\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\graph\tests
+ creating build\lib.win-amd64-3.11\skimage\io\tests
+ copying skimage\io\tests\test_collection.py -> build\lib.win-amd64-3.11\skimage\io\tests
+ copying skimage\io\tests\test_colormixer.py -> build\lib.win-amd64-3.11\skimage\io\tests
+ copying skimage\io\tests\test_fits.py -> build\lib.win-amd64-3.11\skimage\io\tests
+ copying skimage\io\tests\test_histograms.py -> build\lib.win-amd64-3.11\skimage\io\tests
+ copying skimage\io\tests\test_imageio.py -> build\lib.win-amd64-3.11\skimage\io\tests
+ copying skimage\io\tests\test_imread.py -> build\lib.win-amd64-3.11\skimage\io\tests
+ copying skimage\io\tests\test_io.py -> build\lib.win-amd64-3.11\skimage\io\tests
+ copying skimage\io\tests\test_mpl_imshow.py -> build\lib.win-amd64-3.11\skimage\io\tests
+ copying skimage\io\tests\test_multi_image.py -> build\lib.win-amd64-3.11\skimage\io\tests
+ copying skimage\io\tests\test_pil.py -> build\lib.win-amd64-3.11\skimage\io\tests
+ copying skimage\io\tests\test_plugin.py -> build\lib.win-amd64-3.11\skimage\io\tests
+ copying skimage\io\tests\test_plugin_util.py -> build\lib.win-amd64-3.11\skimage\io\tests
+ copying skimage\io\tests\test_sift.py -> build\lib.win-amd64-3.11\skimage\io\tests
+ copying skimage\io\tests\test_simpleitk.py -> build\lib.win-amd64-3.11\skimage\io\tests
+ copying skimage\io\tests\test_tifffile.py -> build\lib.win-amd64-3.11\skimage\io\tests
+ copying skimage\io\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\io\tests
+ creating build\lib.win-amd64-3.11\skimage\io\_plugins
+ copying skimage\io\_plugins\fits_plugin.py -> build\lib.win-amd64-3.11\skimage\io\_plugins
+ copying skimage\io\_plugins\gdal_plugin.py -> build\lib.win-amd64-3.11\skimage\io\_plugins
+ copying skimage\io\_plugins\gtk_plugin.py -> build\lib.win-amd64-3.11\skimage\io\_plugins
+ copying skimage\io\_plugins\imageio_plugin.py -> build\lib.win-amd64-3.11\skimage\io\_plugins
+ copying skimage\io\_plugins\imread_plugin.py -> build\lib.win-amd64-3.11\skimage\io\_plugins
+ copying skimage\io\_plugins\matplotlib_plugin.py -> build\lib.win-amd64-3.11\skimage\io\_plugins
+ copying skimage\io\_plugins\pil_plugin.py -> build\lib.win-amd64-3.11\skimage\io\_plugins
+ copying skimage\io\_plugins\qt_plugin.py -> build\lib.win-amd64-3.11\skimage\io\_plugins
+ copying skimage\io\_plugins\q_color_mixer.py -> build\lib.win-amd64-3.11\skimage\io\_plugins
+ copying skimage\io\_plugins\q_histogram.py -> build\lib.win-amd64-3.11\skimage\io\_plugins
+ copying skimage\io\_plugins\simpleitk_plugin.py -> build\lib.win-amd64-3.11\skimage\io\_plugins
+ copying skimage\io\_plugins\skivi.py -> build\lib.win-amd64-3.11\skimage\io\_plugins
+ copying skimage\io\_plugins\tifffile_plugin.py -> build\lib.win-amd64-3.11\skimage\io\_plugins
+ copying skimage\io\_plugins\util.py -> build\lib.win-amd64-3.11\skimage\io\_plugins
+ copying skimage\io\_plugins\__init__.py -> build\lib.win-amd64-3.11\skimage\io\_plugins
+ creating build\lib.win-amd64-3.11\skimage\measure\tests
+ copying skimage\measure\tests\test_block.py -> build\lib.win-amd64-3.11\skimage\measure\tests
+ copying skimage\measure\tests\test_blur_effect.py -> build\lib.win-amd64-3.11\skimage\measure\tests
+ copying skimage\measure\tests\test_ccomp.py -> build\lib.win-amd64-3.11\skimage\measure\tests
+ copying skimage\measure\tests\test_entropy.py -> build\lib.win-amd64-3.11\skimage\measure\tests
+ copying skimage\measure\tests\test_find_contours.py -> build\lib.win-amd64-3.11\skimage\measure\tests
+ copying skimage\measure\tests\test_fit.py -> build\lib.win-amd64-3.11\skimage\measure\tests
+ copying skimage\measure\tests\test_label.py -> build\lib.win-amd64-3.11\skimage\measure\tests
+ copying skimage\measure\tests\test_marching_cubes.py -> build\lib.win-amd64-3.11\skimage\measure\tests
+ copying skimage\measure\tests\test_moments.py -> build\lib.win-amd64-3.11\skimage\measure\tests
+ copying skimage\measure\tests\test_pnpoly.py -> build\lib.win-amd64-3.11\skimage\measure\tests
+ copying skimage\measure\tests\test_polygon.py -> build\lib.win-amd64-3.11\skimage\measure\tests
+ copying skimage\measure\tests\test_profile.py -> build\lib.win-amd64-3.11\skimage\measure\tests
+ copying skimage\measure\tests\test_regionprops.py -> build\lib.win-amd64-3.11\skimage\measure\tests
+ copying skimage\measure\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\measure\tests
+ creating build\lib.win-amd64-3.11\skimage\metrics\tests
+ copying skimage\metrics\tests\test_segmentation_metrics.py -> build\lib.win-amd64-3.11\skimage\metrics\tests
+ copying skimage\metrics\tests\test_set_metrics.py -> build\lib.win-amd64-3.11\skimage\metrics\tests
+ copying skimage\metrics\tests\test_simple_metrics.py -> build\lib.win-amd64-3.11\skimage\metrics\tests
+ copying skimage\metrics\tests\test_structural_similarity.py -> build\lib.win-amd64-3.11\skimage\metrics\tests
+ copying skimage\metrics\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\metrics\tests
+ creating build\lib.win-amd64-3.11\skimage\morphology\tests
+ copying skimage\morphology\tests\test_binary.py -> build\lib.win-amd64-3.11\skimage\morphology\tests
+ copying skimage\morphology\tests\test_convex_hull.py -> build\lib.win-amd64-3.11\skimage\morphology\tests
+ copying skimage\morphology\tests\test_extrema.py -> build\lib.win-amd64-3.11\skimage\morphology\tests
+ copying skimage\morphology\tests\test_flood_fill.py -> build\lib.win-amd64-3.11\skimage\morphology\tests
+ copying skimage\morphology\tests\test_footprints.py -> build\lib.win-amd64-3.11\skimage\morphology\tests
+ copying skimage\morphology\tests\test_gray.py -> build\lib.win-amd64-3.11\skimage\morphology\tests
+ copying skimage\morphology\tests\test_max_tree.py -> build\lib.win-amd64-3.11\skimage\morphology\tests
+ copying skimage\morphology\tests\test_misc.py -> build\lib.win-amd64-3.11\skimage\morphology\tests
+ copying skimage\morphology\tests\test_reconstruction.py -> build\lib.win-amd64-3.11\skimage\morphology\tests
+ copying skimage\morphology\tests\test_skeletonize.py -> build\lib.win-amd64-3.11\skimage\morphology\tests
+ copying skimage\morphology\tests\test_skeletonize_3d.py -> build\lib.win-amd64-3.11\skimage\morphology\tests
+ copying skimage\morphology\tests\test_util.py -> build\lib.win-amd64-3.11\skimage\morphology\tests
+ copying skimage\morphology\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\morphology\tests
+ creating build\lib.win-amd64-3.11\skimage\registration\tests
+ copying skimage\registration\tests\test_ilk.py -> build\lib.win-amd64-3.11\skimage\registration\tests
+ copying skimage\registration\tests\test_masked_phase_cross_correlation.py -> build\lib.win-amd64-3.11\skimage\registration\tests
+ copying skimage\registration\tests\test_phase_cross_correlation.py -> build\lib.win-amd64-3.11\skimage\registration\tests
+ copying skimage\registration\tests\test_tvl1.py -> build\lib.win-amd64-3.11\skimage\registration\tests
+ copying skimage\registration\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\registration\tests
+ creating build\lib.win-amd64-3.11\skimage\restoration\tests
+ copying skimage\restoration\tests\test_denoise.py -> build\lib.win-amd64-3.11\skimage\restoration\tests
+ copying skimage\restoration\tests\test_inpaint.py -> build\lib.win-amd64-3.11\skimage\restoration\tests
+ copying skimage\restoration\tests\test_j_invariant.py -> build\lib.win-amd64-3.11\skimage\restoration\tests
+ copying skimage\restoration\tests\test_restoration.py -> build\lib.win-amd64-3.11\skimage\restoration\tests
+ copying skimage\restoration\tests\test_rolling_ball.py -> build\lib.win-amd64-3.11\skimage\restoration\tests
+ copying skimage\restoration\tests\test_unwrap.py -> build\lib.win-amd64-3.11\skimage\restoration\tests
+ copying skimage\restoration\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\restoration\tests
+ creating build\lib.win-amd64-3.11\skimage\segmentation\tests
+ copying skimage\segmentation\tests\test_active_contour_model.py -> build\lib.win-amd64-3.11\skimage\segmentation\tests
+ copying skimage\segmentation\tests\test_boundaries.py -> build\lib.win-amd64-3.11\skimage\segmentation\tests
+ copying skimage\segmentation\tests\test_chan_vese.py -> build\lib.win-amd64-3.11\skimage\segmentation\tests
+ copying skimage\segmentation\tests\test_clear_border.py -> build\lib.win-amd64-3.11\skimage\segmentation\tests
+ copying skimage\segmentation\tests\test_expand_labels.py -> build\lib.win-amd64-3.11\skimage\segmentation\tests
+ copying skimage\segmentation\tests\test_felzenszwalb.py -> build\lib.win-amd64-3.11\skimage\segmentation\tests
+ copying skimage\segmentation\tests\test_join.py -> build\lib.win-amd64-3.11\skimage\segmentation\tests
+ copying skimage\segmentation\tests\test_morphsnakes.py -> build\lib.win-amd64-3.11\skimage\segmentation\tests
+ copying skimage\segmentation\tests\test_quickshift.py -> build\lib.win-amd64-3.11\skimage\segmentation\tests
+ copying skimage\segmentation\tests\test_random_walker.py -> build\lib.win-amd64-3.11\skimage\segmentation\tests
+ copying skimage\segmentation\tests\test_slic.py -> build\lib.win-amd64-3.11\skimage\segmentation\tests
+ copying skimage\segmentation\tests\test_watershed.py -> build\lib.win-amd64-3.11\skimage\segmentation\tests
+ copying skimage\segmentation\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\segmentation\tests
+ creating build\lib.win-amd64-3.11\skimage\transform\tests
+ copying skimage\transform\tests\test_finite_radon_transform.py -> build\lib.win-amd64-3.11\skimage\transform\tests
+ copying skimage\transform\tests\test_geometric.py -> build\lib.win-amd64-3.11\skimage\transform\tests
+ copying skimage\transform\tests\test_hough_transform.py -> build\lib.win-amd64-3.11\skimage\transform\tests
+ copying skimage\transform\tests\test_integral.py -> build\lib.win-amd64-3.11\skimage\transform\tests
+ copying skimage\transform\tests\test_pyramids.py -> build\lib.win-amd64-3.11\skimage\transform\tests
+ copying skimage\transform\tests\test_radon_transform.py -> build\lib.win-amd64-3.11\skimage\transform\tests
+ copying skimage\transform\tests\test_warps.py -> build\lib.win-amd64-3.11\skimage\transform\tests
+ copying skimage\transform\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\transform\tests
+ creating build\lib.win-amd64-3.11\skimage\util\tests
+ copying skimage\util\tests\test_apply_parallel.py -> build\lib.win-amd64-3.11\skimage\util\tests
+ copying skimage\util\tests\test_arraycrop.py -> build\lib.win-amd64-3.11\skimage\util\tests
+ copying skimage\util\tests\test_compare.py -> build\lib.win-amd64-3.11\skimage\util\tests
+ copying skimage\util\tests\test_dtype.py -> build\lib.win-amd64-3.11\skimage\util\tests
+ copying skimage\util\tests\test_invert.py -> build\lib.win-amd64-3.11\skimage\util\tests
+ copying skimage\util\tests\test_labels.py -> build\lib.win-amd64-3.11\skimage\util\tests
+ copying skimage\util\tests\test_map_array.py -> build\lib.win-amd64-3.11\skimage\util\tests
+ copying skimage\util\tests\test_montage.py -> build\lib.win-amd64-3.11\skimage\util\tests
+ copying skimage\util\tests\test_random_noise.py -> build\lib.win-amd64-3.11\skimage\util\tests
+ copying skimage\util\tests\test_regular_grid.py -> build\lib.win-amd64-3.11\skimage\util\tests
+ copying skimage\util\tests\test_shape.py -> build\lib.win-amd64-3.11\skimage\util\tests
+ copying skimage\util\tests\test_unique_rows.py -> build\lib.win-amd64-3.11\skimage\util\tests
+ copying skimage\util\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\util\tests
+ creating build\lib.win-amd64-3.11\skimage\viewer\canvastools
+ copying skimage\viewer\canvastools\base.py -> build\lib.win-amd64-3.11\skimage\viewer\canvastools
+ copying skimage\viewer\canvastools\linetool.py -> build\lib.win-amd64-3.11\skimage\viewer\canvastools
+ copying skimage\viewer\canvastools\painttool.py -> build\lib.win-amd64-3.11\skimage\viewer\canvastools
+ copying skimage\viewer\canvastools\recttool.py -> build\lib.win-amd64-3.11\skimage\viewer\canvastools
+ copying skimage\viewer\canvastools\__init__.py -> build\lib.win-amd64-3.11\skimage\viewer\canvastools
+ creating build\lib.win-amd64-3.11\skimage\viewer\plugins
+ copying skimage\viewer\plugins\base.py -> build\lib.win-amd64-3.11\skimage\viewer\plugins
+ copying skimage\viewer\plugins\canny.py -> build\lib.win-amd64-3.11\skimage\viewer\plugins
+ copying skimage\viewer\plugins\color_histogram.py -> build\lib.win-amd64-3.11\skimage\viewer\plugins
+ copying skimage\viewer\plugins\crop.py -> build\lib.win-amd64-3.11\skimage\viewer\plugins
+ copying skimage\viewer\plugins\labelplugin.py -> build\lib.win-amd64-3.11\skimage\viewer\plugins
+ copying skimage\viewer\plugins\lineprofile.py -> build\lib.win-amd64-3.11\skimage\viewer\plugins
+ copying skimage\viewer\plugins\measure.py -> build\lib.win-amd64-3.11\skimage\viewer\plugins
+ copying skimage\viewer\plugins\overlayplugin.py -> build\lib.win-amd64-3.11\skimage\viewer\plugins
+ copying skimage\viewer\plugins\plotplugin.py -> build\lib.win-amd64-3.11\skimage\viewer\plugins
+ copying skimage\viewer\plugins\__init__.py -> build\lib.win-amd64-3.11\skimage\viewer\plugins
+ creating build\lib.win-amd64-3.11\skimage\viewer\tests
+ copying skimage\viewer\tests\test_plugins.py -> build\lib.win-amd64-3.11\skimage\viewer\tests
+ copying skimage\viewer\tests\test_tools.py -> build\lib.win-amd64-3.11\skimage\viewer\tests
+ copying skimage\viewer\tests\test_utils.py -> build\lib.win-amd64-3.11\skimage\viewer\tests
+ copying skimage\viewer\tests\test_viewer.py -> build\lib.win-amd64-3.11\skimage\viewer\tests
+ copying skimage\viewer\tests\test_widgets.py -> build\lib.win-amd64-3.11\skimage\viewer\tests
+ copying skimage\viewer\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\viewer\tests
+ creating build\lib.win-amd64-3.11\skimage\viewer\utils
+ copying skimage\viewer\utils\canvas.py -> build\lib.win-amd64-3.11\skimage\viewer\utils
+ copying skimage\viewer\utils\core.py -> build\lib.win-amd64-3.11\skimage\viewer\utils
+ copying skimage\viewer\utils\dialogs.py -> build\lib.win-amd64-3.11\skimage\viewer\utils
+ copying skimage\viewer\utils\__init__.py -> build\lib.win-amd64-3.11\skimage\viewer\utils
+ creating build\lib.win-amd64-3.11\skimage\viewer\viewers
+ copying skimage\viewer\viewers\core.py -> build\lib.win-amd64-3.11\skimage\viewer\viewers
+ copying skimage\viewer\viewers\__init__.py -> build\lib.win-amd64-3.11\skimage\viewer\viewers
+ creating build\lib.win-amd64-3.11\skimage\viewer\widgets
+ copying skimage\viewer\widgets\core.py -> build\lib.win-amd64-3.11\skimage\viewer\widgets
+ copying skimage\viewer\widgets\history.py -> build\lib.win-amd64-3.11\skimage\viewer\widgets
+ copying skimage\viewer\widgets\__init__.py -> build\lib.win-amd64-3.11\skimage\viewer\widgets
+ creating build\lib.win-amd64-3.11\skimage\_shared\tests
+ copying skimage\_shared\tests\test_coord.py -> build\lib.win-amd64-3.11\skimage\_shared\tests
+ copying skimage\_shared\tests\test_fast_exp.py -> build\lib.win-amd64-3.11\skimage\_shared\tests
+ copying skimage\_shared\tests\test_geometry.py -> build\lib.win-amd64-3.11\skimage\_shared\tests
+ copying skimage\_shared\tests\test_interpolation.py -> build\lib.win-amd64-3.11\skimage\_shared\tests
+ copying skimage\_shared\tests\test_safe_as_int.py -> build\lib.win-amd64-3.11\skimage\_shared\tests
+ copying skimage\_shared\tests\test_testing.py -> build\lib.win-amd64-3.11\skimage\_shared\tests
+ copying skimage\_shared\tests\test_utils.py -> build\lib.win-amd64-3.11\skimage\_shared\tests
+ copying skimage\_shared\tests\test_version_requirements.py -> build\lib.win-amd64-3.11\skimage\_shared\tests
+ copying skimage\_shared\tests\test_warnings.py -> build\lib.win-amd64-3.11\skimage\_shared\tests
+ copying skimage\_shared\tests\__init__.py -> build\lib.win-amd64-3.11\skimage\_shared\tests
+ copying skimage\draw\_draw.pyx -> build\lib.win-amd64-3.11\skimage\draw
+ copying skimage\feature\censure_cy.pyx -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\corner_cy.pyx -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\orb_cy.pyx -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\_cascade.pyx -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\_haar.pyx -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\_hoghistogram.pyx -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\_sift.pyx -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\_texture.pyx -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\safe_openmp.pxd -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\_haar.pxd -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\feature\_texture.pxd -> build\lib.win-amd64-3.11\skimage\feature
+ copying skimage\filters\_multiotsu.pyx -> build\lib.win-amd64-3.11\skimage\filters
+ copying skimage\graph\heap.pyx -> build\lib.win-amd64-3.11\skimage\graph
+ copying skimage\graph\_mcp.pyx -> build\lib.win-amd64-3.11\skimage\graph
+ copying skimage\graph\_spath.pyx -> build\lib.win-amd64-3.11\skimage\graph
+ copying skimage\graph\heap.pxd -> build\lib.win-amd64-3.11\skimage\graph
+ copying skimage\graph\_mcp.pxd -> build\lib.win-amd64-3.11\skimage\graph
+ copying skimage\measure\_ccomp.pyx -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\_find_contours_cy.pyx -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\_marching_cubes_classic_cy.pyx -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\_marching_cubes_lewiner_cy.pyx -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\_moments_cy.pyx -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\_pnpoly.pyx -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\measure\_ccomp.pxd -> build\lib.win-amd64-3.11\skimage\measure
+ copying skimage\morphology\_convex_hull.pyx -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\_extrema_cy.pyx -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\_flood_fill_cy.pyx -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\_grayreconstruct.pyx -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\_max_tree.pyx -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\_skeletonize_3d_cy.pyx -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\_skeletonize_cy.pyx -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\morphology\_queue_with_history.pxi -> build\lib.win-amd64-3.11\skimage\morphology
+ copying skimage\restoration\_denoise_cy.pyx -> build\lib.win-amd64-3.11\skimage\restoration
+ copying skimage\restoration\_inpaint.pyx -> build\lib.win-amd64-3.11\skimage\restoration
+ copying skimage\restoration\_nl_means_denoising.pyx -> build\lib.win-amd64-3.11\skimage\restoration
+ copying skimage\restoration\_rolling_ball_cy.pyx -> build\lib.win-amd64-3.11\skimage\restoration
+ copying skimage\restoration\_unwrap_1d.pyx -> build\lib.win-amd64-3.11\skimage\restoration
+ copying skimage\restoration\_unwrap_2d.pyx -> build\lib.win-amd64-3.11\skimage\restoration
+ copying skimage\restoration\_unwrap_3d.pyx -> build\lib.win-amd64-3.11\skimage\restoration
+ copying skimage\segmentation\_felzenszwalb_cy.pyx -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\_quickshift_cy.pyx -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\_slic.pyx -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\_watershed_cy.pyx -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\heap_general.pxi -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\segmentation\heap_watershed.pxi -> build\lib.win-amd64-3.11\skimage\segmentation
+ copying skimage\transform\_hough_transform.pyx -> build\lib.win-amd64-3.11\skimage\transform
+ copying skimage\transform\_radon_transform.pyx -> build\lib.win-amd64-3.11\skimage\transform
+ copying skimage\transform\_warps_cy.pyx -> build\lib.win-amd64-3.11\skimage\transform
+ copying skimage\util\_remap.pyx -> build\lib.win-amd64-3.11\skimage\util
+ copying skimage\_shared\fast_exp.pyx -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\geometry.pyx -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\interpolation.pyx -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\transform.pyx -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\fast_exp.pxd -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\fused_numerics.pxd -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\geometry.pxd -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\interpolation.pxd -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\_shared\transform.pxd -> build\lib.win-amd64-3.11\skimage\_shared
+ copying skimage\filters\rank\bilateral_cy.pyx -> build\lib.win-amd64-3.11\skimage\filters\rank
+ copying skimage\filters\rank\core_cy.pyx -> build\lib.win-amd64-3.11\skimage\filters\rank
+ copying skimage\filters\rank\core_cy_3d.pyx -> build\lib.win-amd64-3.11\skimage\filters\rank
+ copying skimage\filters\rank\generic_cy.pyx -> build\lib.win-amd64-3.11\skimage\filters\rank
+ copying skimage\filters\rank\percentile_cy.pyx -> build\lib.win-amd64-3.11\skimage\filters\rank
+ copying skimage\filters\rank\core_cy.pxd -> build\lib.win-amd64-3.11\skimage\filters\rank
+ copying skimage\filters\rank\core_cy_3d.pxd -> build\lib.win-amd64-3.11\skimage\filters\rank
+ copying skimage\future\graph\_ncut_cy.pyx -> build\lib.win-amd64-3.11\skimage\future\graph
+ copying skimage\io\_plugins\_colormixer.pyx -> build\lib.win-amd64-3.11\skimage\io\_plugins
+ copying skimage\io\_plugins\_histograms.pyx -> build\lib.win-amd64-3.11\skimage\io\_plugins
+ running build_ext
+ INFO: No module named 'numpy.distutils._msvccompiler' in numpy.distutils; trying from distutils
+ INFO: customize MSVCCompiler
+ INFO: customize MSVCCompiler using ConditionalOpenMP
+ INFO: CCompilerOpt.cc_test_flags[1077] : testing flags (/O2)
+ error: Microsoft Visual C++ 14.0 or greater is required. Get it with ""Microsoft C++ Build Tools"": https://visualstudio.microsoft.com/visual-cpp-build-tools/
+ INFO: CCompilerOpt.cache_flush[857] : write cache to path -> C:\Users\jay\AppData\Local\Temp\pip-install-s2xz2koj\scikit-image_cad4b2fea6c2495988cddf5b95ba2806\build\temp.win-amd64-3.11\Release\ccompiler_opt_cache_ext.py
+ [end of output]
+
+ note: This error originates from a subprocess, and is likely not a problem with pip.
+ ERROR: Failed building wheel for scikit-image
+Failed to build scikit-image
+ERROR: Could not build wheels for scikit-image, which is required to install pyproject.toml-based projects
+
+C:\Windows\System32>source skimage-dev/bin/activate
+'source' is not recognized as an internal or external command,
+operable program or batch file.
+```
+
+
+",1,could not build wheels for scikit image hello i have been trying to install this python module for a number of days now as it is a dependency of i m not completely sure what is causing the problem hence i have been unable to try a large variety of fixes i believe the problem might potentially have something to do with pip as the module installed successfully using mini conda however the module i wish to use is only available on pip any help would be greatly appreciated the full output is as follows click to expand c windows pip install scikit image collecting scikit image using cached scikit image tar gz mb installing build dependencies done getting requirements to build wheel done preparing metadata pyproject toml done requirement already satisfied numpy in c users jay appdata local programs python lib site packages from scikit image requirement already satisfied scipy in c users jay appdata local programs python lib site packages from scikit image requirement already satisfied networkx in c users jay appdata local programs python lib site packages from scikit image requirement already satisfied pillow in c users jay appdata local programs python lib site packages from scikit image collecting imageio using cached imageio none any whl mb collecting tifffile using cached tifffile none any whl kb collecting pywavelets using cached pywavelets win whl mb requirement already satisfied packaging in c users jay appdata local programs python lib site packages from scikit image building wheels for collected packages scikit image building wheel for scikit image pyproject toml error error subprocess exited with error × building wheel for scikit image pyproject toml did not run successfully │ exit code ╰─ setup py deprecationwarning numpy distutils is deprecated since numpy as a result of the deprecation of distutils itself it will be removed for python for older python versions it will remain present it is recommended to use setuptools for those python versions for more details see from numpy distutils command build ext import build ext as npy build ext partial import of skimage during the build process compiling c users jay appdata local temp pip install scikit image skimage morphology skeletonize cy pyx because it changed cythonizing c users jay appdata local temp pip install scikit image skimage morphology skeletonize cy pyx running bdist wheel running build running config cc info unifing config cc config build clib build ext build commands compiler options running config fc info unifing config fc config build clib build ext build commands fcompiler options running build src info build src info building extension skimage shared geometry sources info building extension skimage shared transform sources info building extension skimage shared interpolation sources info building extension skimage shared fast exp sources info building extension skimage draw draw sources info building extension skimage feature cascade sources info building extension skimage feature corner cy sources info building extension skimage feature censure cy sources info building extension skimage feature orb cy sources info building extension skimage feature texture sources info building extension skimage feature hoghistogram sources info building extension skimage feature haar sources info building extension skimage feature sift sources info building extension skimage feature brief cy sources info building extension skimage feature hessian det appx sources info building extension skimage restoration unwrap sources info building extension skimage restoration unwrap sources info building extension skimage restoration unwrap sources info building extension skimage restoration denoise cy sources info building extension skimage restoration nl means denoising sources info building extension skimage restoration rolling ball cy sources info building extension skimage restoration inpaint sources info building extension skimage filters rank core cy sources info building extension skimage filters rank core cy sources info building extension skimage filters multiotsu sources info building extension skimage filters rank generic cy sources info building extension skimage filters rank percentile cy sources info building extension skimage filters rank bilateral cy sources info building extension skimage future graph ncut cy sources info building extension skimage graph spath sources info building extension skimage graph mcp sources info building extension skimage graph heap sources info building extension skimage io plugins colormixer sources info building extension skimage io plugins histograms sources info building extension skimage measure ccomp sources info building extension skimage measure find contours cy sources info building extension skimage measure moments cy sources info building extension skimage measure marching cubes classic cy sources info building extension skimage measure marching cubes lewiner cy sources info building extension skimage measure pnpoly sources info building extension skimage morphology skeletonize cy sources info building extension skimage morphology convex hull sources info building extension skimage morphology grayreconstruct sources info building extension skimage morphology max tree sources info building extension skimage morphology skeletonize cy sources info building extension skimage morphology extrema cy sources info building extension skimage morphology flood fill cy sources info building extension skimage transform hough transform sources info building extension skimage transform warps cy sources info building extension skimage transform radon transform sources info building extension skimage util remap sources info building extension skimage segmentation watershed cy sources info building extension skimage segmentation felzenszwalb cy sources info building extension skimage segmentation quickshift cy sources info building extension skimage segmentation slic sources info building data files sources info build src building npy pkg config files running build py creating build creating build lib win creating build lib win skimage copying skimage conftest py build lib win skimage copying skimage setup py build lib win skimage copying skimage build py build lib win skimage copying skimage init py build lib win skimage creating build lib win doc creating build lib win doc ext copying doc ext doi role py build lib win doc ext copying doc ext notebook doc py build lib win doc ext copying doc ext py build lib win doc ext copying doc ext init py build lib win doc ext creating build lib win doc ext tests copying doc ext tests test notebook doc py build lib win doc ext tests copying doc ext tests init py build lib win doc ext tests creating build lib win skimage color copying skimage color adapt rgb py build lib win skimage color copying skimage color colorconv py build lib win skimage color copying skimage color colorlabel py build lib win skimage color copying skimage color delta e py build lib win skimage color copying skimage color rgb colors py build lib win skimage color copying skimage color init py build lib win skimage color creating build lib win skimage data copying skimage data setup py build lib win skimage data copying skimage data binary blobs py build lib win skimage data copying skimage data fetchers py build lib win skimage data copying skimage data registry py build lib win skimage data copying skimage data init py build lib win skimage data creating build lib win skimage draw copying skimage draw draw py build lib win skimage draw copying skimage draw py build lib win skimage draw copying skimage draw draw nd py build lib win skimage draw copying skimage draw setup py build lib win skimage draw copying skimage draw py build lib win skimage draw copying skimage draw random shapes py build lib win skimage draw copying skimage draw init py build lib win skimage draw creating build lib win skimage exposure copying skimage exposure exposure py build lib win skimage exposure copying skimage exposure histogram matching py build lib win skimage exposure copying skimage exposure setup py build lib win skimage exposure copying skimage exposure adapthist py build lib win skimage exposure copying skimage exposure init py build lib win skimage exposure creating build lib win skimage feature copying skimage feature blob py build lib win skimage feature copying skimage feature brief py build lib win skimage feature copying skimage feature brief pythran py build lib win skimage feature copying skimage feature censure py build lib win skimage feature copying skimage feature corner py build lib win skimage feature copying skimage feature haar py build lib win skimage feature copying skimage feature match py build lib win skimage feature copying skimage feature orb py build lib win skimage feature copying skimage feature peak py build lib win skimage feature copying skimage feature setup py build lib win skimage feature copying skimage feature sift py build lib win skimage feature copying skimage feature template py build lib win skimage feature copying skimage feature texture py build lib win skimage feature copying skimage feature util py build lib win skimage feature copying skimage feature basic features py build lib win skimage feature copying skimage feature canny py build lib win skimage feature copying skimage feature daisy py build lib win skimage feature copying skimage feature hessian det appx pythran py build lib win skimage feature copying skimage feature hog py build lib win skimage feature copying skimage feature orb descriptor positions py build lib win skimage feature copying skimage feature init py build lib win skimage feature creating build lib win skimage filters copying skimage filters edges py build lib win skimage filters copying skimage filters lpi filter py build lib win skimage filters copying skimage filters ridges py build lib win skimage filters copying skimage filters setup py build lib win skimage filters copying skimage filters thresholding py build lib win skimage filters copying skimage filters fft based py build lib win skimage filters copying skimage filters gabor py build lib win skimage filters copying skimage filters gaussian py build lib win skimage filters copying skimage filters median py build lib win skimage filters copying skimage filters rank order py build lib win skimage filters copying skimage filters sparse py build lib win skimage filters copying skimage filters unsharp mask py build lib win skimage filters copying skimage filters window py build lib win skimage filters copying skimage filters init py build lib win skimage filters creating build lib win skimage future copying skimage future manual segmentation py build lib win skimage future copying skimage future setup py build lib win skimage future copying skimage future trainable segmentation py build lib win skimage future copying skimage future init py build lib win skimage future creating build lib win skimage graph copying skimage graph mcp py build lib win skimage graph copying skimage graph setup py build lib win skimage graph copying skimage graph spath py build lib win skimage graph copying skimage graph graph py build lib win skimage graph copying skimage graph init py build lib win skimage graph creating build lib win skimage io copying skimage io collection py build lib win skimage io copying skimage io manage plugins py build lib win skimage io copying skimage io setup py build lib win skimage io copying skimage io sift py build lib win skimage io copying skimage io util py build lib win skimage io copying skimage io image stack py build lib win skimage io copying skimage io io py build lib win skimage io copying skimage io init py build lib win skimage io creating build lib win skimage measure copying skimage measure block py build lib win skimage measure copying skimage measure entropy py build lib win skimage measure copying skimage measure fit py build lib win skimage measure copying skimage measure pnpoly py build lib win skimage measure copying skimage measure profile py build lib win skimage measure copying skimage measure setup py build lib win skimage measure copying skimage measure blur effect py build lib win skimage measure copying skimage measure find contours py build lib win skimage measure copying skimage measure label py build lib win skimage measure copying skimage measure marching cubes classic py build lib win skimage measure copying skimage measure marching cubes lewiner py build lib win skimage measure copying skimage measure marching cubes lewiner luts py build lib win skimage measure copying skimage measure moments py build lib win skimage measure copying skimage measure polygon py build lib win skimage measure copying skimage measure regionprops py build lib win skimage measure copying skimage measure regionprops utils py build lib win skimage measure copying skimage measure init py build lib win skimage measure creating build lib win skimage metrics copying skimage metrics set metrics py build lib win skimage metrics copying skimage metrics simple metrics py build lib win skimage metrics copying skimage metrics adapted rand error py build lib win skimage metrics copying skimage metrics contingency table py build lib win skimage metrics copying skimage metrics structural similarity py build lib win skimage metrics copying skimage metrics variation of information py build lib win skimage metrics copying skimage metrics init py build lib win skimage metrics creating build lib win skimage morphology copying skimage morphology binary py build lib win skimage morphology copying skimage morphology convex hull py build lib win skimage morphology copying skimage morphology extrema py build lib win skimage morphology copying skimage morphology footprints py build lib win skimage morphology copying skimage morphology gray py build lib win skimage morphology copying skimage morphology grayreconstruct py build lib win skimage morphology copying skimage morphology grey py build lib win skimage morphology copying skimage morphology greyreconstruct py build lib win skimage morphology copying skimage morphology max tree py build lib win skimage morphology copying skimage morphology misc py build lib win skimage morphology copying skimage morphology selem py build lib win skimage morphology copying skimage morphology setup py build lib win skimage morphology copying skimage morphology flood fill py build lib win skimage morphology copying skimage morphology skeletonize py build lib win skimage morphology copying skimage morphology util py build lib win skimage morphology copying skimage morphology init py build lib win skimage morphology creating build lib win skimage registration copying skimage registration masked phase cross correlation py build lib win skimage registration copying skimage registration optical flow py build lib win skimage registration copying skimage registration optical flow utils py build lib win skimage registration copying skimage registration phase cross correlation py build lib win skimage registration copying skimage registration init py build lib win skimage registration creating build lib win skimage restoration copying skimage restoration deconvolution py build lib win skimage restoration copying skimage restoration inpaint py build lib win skimage restoration copying skimage restoration j invariant py build lib win skimage restoration copying skimage restoration non local means py build lib win skimage restoration copying skimage restoration rolling ball py build lib win skimage restoration copying skimage restoration setup py build lib win skimage restoration copying skimage restoration uft py build lib win skimage restoration copying skimage restoration unwrap py build lib win skimage restoration copying skimage restoration cycle spin py build lib win skimage restoration copying skimage restoration denoise py build lib win skimage restoration copying skimage restoration init py build lib win skimage restoration creating build lib win skimage scripts copying skimage scripts skivi py build lib win skimage scripts copying skimage scripts init py build lib win skimage scripts creating build lib win skimage segmentation copying skimage segmentation active contour model py build lib win skimage segmentation copying skimage segmentation boundaries py build lib win skimage segmentation copying skimage segmentation morphsnakes py build lib win skimage segmentation copying skimage segmentation random walker segmentation py build lib win skimage segmentation copying skimage segmentation setup py build lib win skimage segmentation copying skimage segmentation slic superpixels py build lib win skimage segmentation copying skimage segmentation chan vese py build lib win skimage segmentation copying skimage segmentation clear border py build lib win skimage segmentation copying skimage segmentation expand labels py build lib win skimage segmentation copying skimage segmentation felzenszwalb py build lib win skimage segmentation copying skimage segmentation join py build lib win skimage segmentation copying skimage segmentation quickshift py build lib win skimage segmentation copying skimage segmentation watershed py build lib win skimage segmentation copying skimage segmentation init py build lib win skimage segmentation creating build lib win skimage transform copying skimage transform finite radon transform py build lib win skimage transform copying skimage transform hough transform py build lib win skimage transform copying skimage transform integral py build lib win skimage transform copying skimage transform pyramids py build lib win skimage transform copying skimage transform radon transform py build lib win skimage transform copying skimage transform setup py build lib win skimage transform copying skimage transform geometric py build lib win skimage transform copying skimage transform warps py build lib win skimage transform copying skimage transform init py build lib win skimage transform creating build lib win skimage util copying skimage util apply parallel py build lib win skimage util copying skimage util arraycrop py build lib win skimage util copying skimage util compare py build lib win skimage util copying skimage util dtype py build lib win skimage util copying skimage util lookfor py build lib win skimage util copying skimage util noise py build lib win skimage util copying skimage util setup py build lib win skimage util copying skimage util shape py build lib win skimage util copying skimage util unique py build lib win skimage util copying skimage util invert py build lib win skimage util copying skimage util label py build lib win skimage util copying skimage util map array py build lib win skimage util copying skimage util montage py build lib win skimage util copying skimage util regular grid py build lib win skimage util copying skimage util init py build lib win skimage util creating build lib win skimage viewer copying skimage viewer qt py build lib win skimage viewer copying skimage viewer init py build lib win skimage viewer creating build lib win skimage shared copying skimage shared coord py build lib win skimage shared copying skimage shared filters py build lib win skimage shared copying skimage shared lazy py build lib win skimage shared copying skimage shared setup py build lib win skimage shared copying skimage shared tester py build lib win skimage shared copying skimage shared testing py build lib win skimage shared copying skimage shared utils py build lib win skimage shared copying skimage shared version requirements py build lib win skimage shared copying skimage shared dependency checks py build lib win skimage shared copying skimage shared geometry py build lib win skimage shared copying skimage shared tempfile py build lib win skimage shared copying skimage shared warnings py build lib win skimage shared copying skimage shared init py build lib win skimage shared creating build lib win skimage color tests copying skimage color tests test adapt rgb py build lib win skimage color tests copying skimage color tests test colorconv py build lib win skimage color tests copying skimage color tests test colorlabel py build lib win skimage color tests copying skimage color tests test delta e py build lib win skimage color tests copying skimage color tests init py build lib win skimage color tests creating build lib win skimage data tests copying skimage data tests test data py build lib win skimage data tests copying skimage data tests init py build lib win skimage data tests creating build lib win skimage draw tests copying skimage draw tests test draw py build lib win skimage draw tests copying skimage draw tests test py build lib win skimage draw tests copying skimage draw tests test draw nd py build lib win skimage draw tests copying skimage draw tests test py build lib win skimage draw tests copying skimage draw tests test random shapes py build lib win skimage draw tests copying skimage draw tests init py build lib win skimage draw tests creating build lib win skimage exposure tests copying skimage exposure tests test exposure py build lib win skimage exposure tests copying skimage exposure tests test histogram matching py build lib win skimage exposure tests copying skimage exposure tests init py build lib win skimage exposure tests creating build lib win skimage feature tests copying skimage feature tests test basic features py build lib win skimage feature tests copying skimage feature tests test blob py build lib win skimage feature tests copying skimage feature tests test brief py build lib win skimage feature tests copying skimage feature tests test canny py build lib win skimage feature tests copying skimage feature tests test cascade py build lib win skimage feature tests copying skimage feature tests test censure py build lib win skimage feature tests copying skimage feature tests test corner py build lib win skimage feature tests copying skimage feature tests test daisy py build lib win skimage feature tests copying skimage feature tests test haar py build lib win skimage feature tests copying skimage feature tests test hog py build lib win skimage feature tests copying skimage feature tests test match py build lib win skimage feature tests copying skimage feature tests test orb py build lib win skimage feature tests copying skimage feature tests test peak py build lib win skimage feature tests copying skimage feature tests test sift py build lib win skimage feature tests copying skimage feature tests test template py build lib win skimage feature tests copying skimage feature tests test texture py build lib win skimage feature tests copying skimage feature tests test util py build lib win skimage feature tests copying skimage feature tests init py build lib win skimage feature tests creating build lib win skimage filters rank copying skimage filters rank bilateral py build lib win skimage filters rank copying skimage filters rank generic py build lib win skimage filters rank copying skimage filters rank percentile py build lib win skimage filters rank copying skimage filters rank init py build lib win skimage filters rank creating build lib win skimage filters tests copying skimage filters tests test correlate py build lib win skimage filters tests copying skimage filters tests test edges py build lib win skimage filters tests copying skimage filters tests test fft based py build lib win skimage filters tests copying skimage filters tests test gabor py build lib win skimage filters tests copying skimage filters tests test gaussian py build lib win skimage filters tests copying skimage filters tests test lpi filter py build lib win skimage filters tests copying skimage filters tests test median py build lib win skimage filters tests copying skimage filters tests test ridges py build lib win skimage filters tests copying skimage filters tests test thresholding py build lib win skimage filters tests copying skimage filters tests test unsharp mask py build lib win skimage filters tests copying skimage filters tests test window py build lib win skimage filters tests copying skimage filters tests init py build lib win skimage filters tests creating build lib win skimage filters rank tests copying skimage filters rank tests test rank py build lib win skimage filters rank tests copying skimage filters rank tests init py build lib win skimage filters rank tests creating build lib win skimage future graph copying skimage future graph graph cut py build lib win skimage future graph copying skimage future graph graph merge py build lib win skimage future graph copying skimage future graph rag py build lib win skimage future graph copying skimage future graph setup py build lib win skimage future graph copying skimage future graph ncut py build lib win skimage future graph copying skimage future graph init py build lib win skimage future graph creating build lib win skimage future tests copying skimage future tests test trainable segmentation py build lib win skimage future tests copying skimage future tests init py build lib win skimage future tests creating build lib win skimage future graph tests copying skimage future graph tests test rag py build lib win skimage future graph tests copying skimage future graph tests init py build lib win skimage future graph tests creating build lib win skimage graph tests copying skimage graph tests test anisotropy py build lib win skimage graph tests copying skimage graph tests test connect py build lib win skimage graph tests copying skimage graph tests test flexible py build lib win skimage graph tests copying skimage graph tests test heap py build lib win skimage graph tests copying skimage graph tests test mcp py build lib win skimage graph tests copying skimage graph tests test pixel graph py build lib win skimage graph tests copying skimage graph tests test spath py build lib win skimage graph tests copying skimage graph tests init py build lib win skimage graph tests creating build lib win skimage io tests copying skimage io tests test collection py build lib win skimage io tests copying skimage io tests test colormixer py build lib win skimage io tests copying skimage io tests test fits py build lib win skimage io tests copying skimage io tests test histograms py build lib win skimage io tests copying skimage io tests test imageio py build lib win skimage io tests copying skimage io tests test imread py build lib win skimage io tests copying skimage io tests test io py build lib win skimage io tests copying skimage io tests test mpl imshow py build lib win skimage io tests copying skimage io tests test multi image py build lib win skimage io tests copying skimage io tests test pil py build lib win skimage io tests copying skimage io tests test plugin py build lib win skimage io tests copying skimage io tests test plugin util py build lib win skimage io tests copying skimage io tests test sift py build lib win skimage io tests copying skimage io tests test simpleitk py build lib win skimage io tests copying skimage io tests test tifffile py build lib win skimage io tests copying skimage io tests init py build lib win skimage io tests creating build lib win skimage io plugins copying skimage io plugins fits plugin py build lib win skimage io plugins copying skimage io plugins gdal plugin py build lib win skimage io plugins copying skimage io plugins gtk plugin py build lib win skimage io plugins copying skimage io plugins imageio plugin py build lib win skimage io plugins copying skimage io plugins imread plugin py build lib win skimage io plugins copying skimage io plugins matplotlib plugin py build lib win skimage io plugins copying skimage io plugins pil plugin py build lib win skimage io plugins copying skimage io plugins qt plugin py build lib win skimage io plugins copying skimage io plugins q color mixer py build lib win skimage io plugins copying skimage io plugins q histogram py build lib win skimage io plugins copying skimage io plugins simpleitk plugin py build lib win skimage io plugins copying skimage io plugins skivi py build lib win skimage io plugins copying skimage io plugins tifffile plugin py build lib win skimage io plugins copying skimage io plugins util py build lib win skimage io plugins copying skimage io plugins init py build lib win skimage io plugins creating build lib win skimage measure tests copying skimage measure tests test block py build lib win skimage measure tests copying skimage measure tests test blur effect py build lib win skimage measure tests copying skimage measure tests test ccomp py build lib win skimage measure tests copying skimage measure tests test entropy py build lib win skimage measure tests copying skimage measure tests test find contours py build lib win skimage measure tests copying skimage measure tests test fit py build lib win skimage measure tests copying skimage measure tests test label py build lib win skimage measure tests copying skimage measure tests test marching cubes py build lib win skimage measure tests copying skimage measure tests test moments py build lib win skimage measure tests copying skimage measure tests test pnpoly py build lib win skimage measure tests copying skimage measure tests test polygon py build lib win skimage measure tests copying skimage measure tests test profile py build lib win skimage measure tests copying skimage measure tests test regionprops py build lib win skimage measure tests copying skimage measure tests init py build lib win skimage measure tests creating build lib win skimage metrics tests copying skimage metrics tests test segmentation metrics py build lib win skimage metrics tests copying skimage metrics tests test set metrics py build lib win skimage metrics tests copying skimage metrics tests test simple metrics py build lib win skimage metrics tests copying skimage metrics tests test structural similarity py build lib win skimage metrics tests copying skimage metrics tests init py build lib win skimage metrics tests creating build lib win skimage morphology tests copying skimage morphology tests test binary py build lib win skimage morphology tests copying skimage morphology tests test convex hull py build lib win skimage morphology tests copying skimage morphology tests test extrema py build lib win skimage morphology tests copying skimage morphology tests test flood fill py build lib win skimage morphology tests copying skimage morphology tests test footprints py build lib win skimage morphology tests copying skimage morphology tests test gray py build lib win skimage morphology tests copying skimage morphology tests test max tree py build lib win skimage morphology tests copying skimage morphology tests test misc py build lib win skimage morphology tests copying skimage morphology tests test reconstruction py build lib win skimage morphology tests copying skimage morphology tests test skeletonize py build lib win skimage morphology tests copying skimage morphology tests test skeletonize py build lib win skimage morphology tests copying skimage morphology tests test util py build lib win skimage morphology tests copying skimage morphology tests init py build lib win skimage morphology tests creating build lib win skimage registration tests copying skimage registration tests test ilk py build lib win skimage registration tests copying skimage registration tests test masked phase cross correlation py build lib win skimage registration tests copying skimage registration tests test phase cross correlation py build lib win skimage registration tests copying skimage registration tests test py build lib win skimage registration tests copying skimage registration tests init py build lib win skimage registration tests creating build lib win skimage restoration tests copying skimage restoration tests test denoise py build lib win skimage restoration tests copying skimage restoration tests test inpaint py build lib win skimage restoration tests copying skimage restoration tests test j invariant py build lib win skimage restoration tests copying skimage restoration tests test restoration py build lib win skimage restoration tests copying skimage restoration tests test rolling ball py build lib win skimage restoration tests copying skimage restoration tests test unwrap py build lib win skimage restoration tests copying skimage restoration tests init py build lib win skimage restoration tests creating build lib win skimage segmentation tests copying skimage segmentation tests test active contour model py build lib win skimage segmentation tests copying skimage segmentation tests test boundaries py build lib win skimage segmentation tests copying skimage segmentation tests test chan vese py build lib win skimage segmentation tests copying skimage segmentation tests test clear border py build lib win skimage segmentation tests copying skimage segmentation tests test expand labels py build lib win skimage segmentation tests copying skimage segmentation tests test felzenszwalb py build lib win skimage segmentation tests copying skimage segmentation tests test join py build lib win skimage segmentation tests copying skimage segmentation tests test morphsnakes py build lib win skimage segmentation tests copying skimage segmentation tests test quickshift py build lib win skimage segmentation tests copying skimage segmentation tests test random walker py build lib win skimage segmentation tests copying skimage segmentation tests test slic py build lib win skimage segmentation tests copying skimage segmentation tests test watershed py build lib win skimage segmentation tests copying skimage segmentation tests init py build lib win skimage segmentation tests creating build lib win skimage transform tests copying skimage transform tests test finite radon transform py build lib win skimage transform tests copying skimage transform tests test geometric py build lib win skimage transform tests copying skimage transform tests test hough transform py build lib win skimage transform tests copying skimage transform tests test integral py build lib win skimage transform tests copying skimage transform tests test pyramids py build lib win skimage transform tests copying skimage transform tests test radon transform py build lib win skimage transform tests copying skimage transform tests test warps py build lib win skimage transform tests copying skimage transform tests init py build lib win skimage transform tests creating build lib win skimage util tests copying skimage util tests test apply parallel py build lib win skimage util tests copying skimage util tests test arraycrop py build lib win skimage util tests copying skimage util tests test compare py build lib win skimage util tests copying skimage util tests test dtype py build lib win skimage util tests copying skimage util tests test invert py build lib win skimage util tests copying skimage util tests test labels py build lib win skimage util tests copying skimage util tests test map array py build lib win skimage util tests copying skimage util tests test montage py build lib win skimage util tests copying skimage util tests test random noise py build lib win skimage util tests copying skimage util tests test regular grid py build lib win skimage util tests copying skimage util tests test shape py build lib win skimage util tests copying skimage util tests test unique rows py build lib win skimage util tests copying skimage util tests init py build lib win skimage util tests creating build lib win skimage viewer canvastools copying skimage viewer canvastools base py build lib win skimage viewer canvastools copying skimage viewer canvastools linetool py build lib win skimage viewer canvastools copying skimage viewer canvastools painttool py build lib win skimage viewer canvastools copying skimage viewer canvastools recttool py build lib win skimage viewer canvastools copying skimage viewer canvastools init py build lib win skimage viewer canvastools creating build lib win skimage viewer plugins copying skimage viewer plugins base py build lib win skimage viewer plugins copying skimage viewer plugins canny py build lib win skimage viewer plugins copying skimage viewer plugins color histogram py build lib win skimage viewer plugins copying skimage viewer plugins crop py build lib win skimage viewer plugins copying skimage viewer plugins labelplugin py build lib win skimage viewer plugins copying skimage viewer plugins lineprofile py build lib win skimage viewer plugins copying skimage viewer plugins measure py build lib win skimage viewer plugins copying skimage viewer plugins overlayplugin py build lib win skimage viewer plugins copying skimage viewer plugins plotplugin py build lib win skimage viewer plugins copying skimage viewer plugins init py build lib win skimage viewer plugins creating build lib win skimage viewer tests copying skimage viewer tests test plugins py build lib win skimage viewer tests copying skimage viewer tests test tools py build lib win skimage viewer tests copying skimage viewer tests test utils py build lib win skimage viewer tests copying skimage viewer tests test viewer py build lib win skimage viewer tests copying skimage viewer tests test widgets py build lib win skimage viewer tests copying skimage viewer tests init py build lib win skimage viewer tests creating build lib win skimage viewer utils copying skimage viewer utils canvas py build lib win skimage viewer utils copying skimage viewer utils core py build lib win skimage viewer utils copying skimage viewer utils dialogs py build lib win skimage viewer utils copying skimage viewer utils init py build lib win skimage viewer utils creating build lib win skimage viewer viewers copying skimage viewer viewers core py build lib win skimage viewer viewers copying skimage viewer viewers init py build lib win skimage viewer viewers creating build lib win skimage viewer widgets copying skimage viewer widgets core py build lib win skimage viewer widgets copying skimage viewer widgets history py build lib win skimage viewer widgets copying skimage viewer widgets init py build lib win skimage viewer widgets creating build lib win skimage shared tests copying skimage shared tests test coord py build lib win skimage shared tests copying skimage shared tests test fast exp py build lib win skimage shared tests copying skimage shared tests test geometry py build lib win skimage shared tests copying skimage shared tests test interpolation py build lib win skimage shared tests copying skimage shared tests test safe as int py build lib win skimage shared tests copying skimage shared tests test testing py build lib win skimage shared tests copying skimage shared tests test utils py build lib win skimage shared tests copying skimage shared tests test version requirements py build lib win skimage shared tests copying skimage shared tests test warnings py build lib win skimage shared tests copying skimage shared tests init py build lib win skimage shared tests copying skimage draw draw pyx build lib win skimage draw copying skimage feature censure cy pyx build lib win skimage feature copying skimage feature corner cy pyx build lib win skimage feature copying skimage feature orb cy pyx build lib win skimage feature copying skimage feature cascade pyx build lib win skimage feature copying skimage feature haar pyx build lib win skimage feature copying skimage feature hoghistogram pyx build lib win skimage feature copying skimage feature sift pyx build lib win skimage feature copying skimage feature texture pyx build lib win skimage feature copying skimage feature safe openmp pxd build lib win skimage feature copying skimage feature haar pxd build lib win skimage feature copying skimage feature texture pxd build lib win skimage feature copying skimage filters multiotsu pyx build lib win skimage filters copying skimage graph heap pyx build lib win skimage graph copying skimage graph mcp pyx build lib win skimage graph copying skimage graph spath pyx build lib win skimage graph copying skimage graph heap pxd build lib win skimage graph copying skimage graph mcp pxd build lib win skimage graph copying skimage measure ccomp pyx build lib win skimage measure copying skimage measure find contours cy pyx build lib win skimage measure copying skimage measure marching cubes classic cy pyx build lib win skimage measure copying skimage measure marching cubes lewiner cy pyx build lib win skimage measure copying skimage measure moments cy pyx build lib win skimage measure copying skimage measure pnpoly pyx build lib win skimage measure copying skimage measure ccomp pxd build lib win skimage measure copying skimage morphology convex hull pyx build lib win skimage morphology copying skimage morphology extrema cy pyx build lib win skimage morphology copying skimage morphology flood fill cy pyx build lib win skimage morphology copying skimage morphology grayreconstruct pyx build lib win skimage morphology copying skimage morphology max tree pyx build lib win skimage morphology copying skimage morphology skeletonize cy pyx build lib win skimage morphology copying skimage morphology skeletonize cy pyx build lib win skimage morphology copying skimage morphology queue with history pxi build lib win skimage morphology copying skimage restoration denoise cy pyx build lib win skimage restoration copying skimage restoration inpaint pyx build lib win skimage restoration copying skimage restoration nl means denoising pyx build lib win skimage restoration copying skimage restoration rolling ball cy pyx build lib win skimage restoration copying skimage restoration unwrap pyx build lib win skimage restoration copying skimage restoration unwrap pyx build lib win skimage restoration copying skimage restoration unwrap pyx build lib win skimage restoration copying skimage segmentation felzenszwalb cy pyx build lib win skimage segmentation copying skimage segmentation quickshift cy pyx build lib win skimage segmentation copying skimage segmentation slic pyx build lib win skimage segmentation copying skimage segmentation watershed cy pyx build lib win skimage segmentation copying skimage segmentation heap general pxi build lib win skimage segmentation copying skimage segmentation heap watershed pxi build lib win skimage segmentation copying skimage transform hough transform pyx build lib win skimage transform copying skimage transform radon transform pyx build lib win skimage transform copying skimage transform warps cy pyx build lib win skimage transform copying skimage util remap pyx build lib win skimage util copying skimage shared fast exp pyx build lib win skimage shared copying skimage shared geometry pyx build lib win skimage shared copying skimage shared interpolation pyx build lib win skimage shared copying skimage shared transform pyx build lib win skimage shared copying skimage shared fast exp pxd build lib win skimage shared copying skimage shared fused numerics pxd build lib win skimage shared copying skimage shared geometry pxd build lib win skimage shared copying skimage shared interpolation pxd build lib win skimage shared copying skimage shared transform pxd build lib win skimage shared copying skimage filters rank bilateral cy pyx build lib win skimage filters rank copying skimage filters rank core cy pyx build lib win skimage filters rank copying skimage filters rank core cy pyx build lib win skimage filters rank copying skimage filters rank generic cy pyx build lib win skimage filters rank copying skimage filters rank percentile cy pyx build lib win skimage filters rank copying skimage filters rank core cy pxd build lib win skimage filters rank copying skimage filters rank core cy pxd build lib win skimage filters rank copying skimage future graph ncut cy pyx build lib win skimage future graph copying skimage io plugins colormixer pyx build lib win skimage io plugins copying skimage io plugins histograms pyx build lib win skimage io plugins running build ext info no module named numpy distutils msvccompiler in numpy distutils trying from distutils info customize msvccompiler info customize msvccompiler using conditionalopenmp info ccompileropt cc test flags testing flags error microsoft visual c or greater is required get it with microsoft c build tools info ccompileropt cache flush write cache to path c users jay appdata local temp pip install scikit image build temp win release ccompiler opt cache ext py note this error originates from a subprocess and is likely not a problem with pip error failed building wheel for scikit image failed to build scikit image error could not build wheels for scikit image which is required to install pyproject toml based projects c windows source skimage dev bin activate source is not recognized as an internal or external command operable program or batch file ,1
+194,4162516935.0,IssuesEvent,2016-06-17 20:43:40,ccswbs/hjckrrh,https://api.github.com/repos/ccswbs/hjckrrh,closed,CC - Allow Site Managers to add custom panes to a specific node layout (Configure Panelizer/Page Manager),feature: book (BK) feature: Custom Content (C) feature: events (E) feature: faq (F) feature: featured item (FT) feature: general (G) feature: news (N) feature: page (P) feature: people profiles (PP) feature: services (SR) priority: high type: missing requirement,"Allow site managers to add custom panes to a specific node layout.
+
+@brharp has recommended using Panelizer as a solution. https://www.drupal.org/project/panelizer
+
+May require configuring page manager to allow use of panelizer.
+
+Clients who are interested in this:
+- Math & Stats
+- Business
+- SOCS",1.0,"CC - Allow Site Managers to add custom panes to a specific node layout (Configure Panelizer/Page Manager) - Allow site managers to add custom panes to a specific node layout.
+
+@brharp has recommended using Panelizer as a solution. https://www.drupal.org/project/panelizer
+
+May require configuring page manager to allow use of panelizer.
+
+Clients who are interested in this:
+- Math & Stats
+- Business
+- SOCS",1,cc allow site managers to add custom panes to a specific node layout configure panelizer page manager allow site managers to add custom panes to a specific node layout brharp has recommended using panelizer as a solution may require configuring page manager to allow use of panelizer clients who are interested in this math stats business socs,1
+7564,10700086695.0,IssuesEvent,2019-10-23 22:42:29,Botffy/XPLanG,https://api.github.com/repos/Botffy/XPLanG,closed,Scoping refactor,requirement,"Currently we maintain a LeBlanc-Cook symbol table to keep track of lexical scoping. However, the language will actually only have two scopes: local and global. Only functions and procedures will have their own lexical scope. Also, it was decided that name shadowing shall not be allowed.
+
+Accordingly, the current implementation is overly complex, and should be simplified. Greatly.
+
+Throw the LeBlanc-Cook symbol table away entirely, and maintain only two maps: one for the global scope, and one for the currently active local scope, if there is one. When declaring a name, Context should check availabilty in the global scope, then in the local scope. If the name is still free, the name is declared in local scope, if there is one, otherwise in the global scope. All other cases shall result in errors.
+
+This is also a good possibility to clean up Context a bit, especially wrt the declaration functions which are in general inconsistent.",1.0,"Scoping refactor - Currently we maintain a LeBlanc-Cook symbol table to keep track of lexical scoping. However, the language will actually only have two scopes: local and global. Only functions and procedures will have their own lexical scope. Also, it was decided that name shadowing shall not be allowed.
+
+Accordingly, the current implementation is overly complex, and should be simplified. Greatly.
+
+Throw the LeBlanc-Cook symbol table away entirely, and maintain only two maps: one for the global scope, and one for the currently active local scope, if there is one. When declaring a name, Context should check availabilty in the global scope, then in the local scope. If the name is still free, the name is declared in local scope, if there is one, otherwise in the global scope. All other cases shall result in errors.
+
+This is also a good possibility to clean up Context a bit, especially wrt the declaration functions which are in general inconsistent.",0,scoping refactor currently we maintain a leblanc cook symbol table to keep track of lexical scoping however the language will actually only have two scopes local and global only functions and procedures will have their own lexical scope also it was decided that name shadowing shall not be allowed accordingly the current implementation is overly complex and should be simplified greatly throw the leblanc cook symbol table away entirely and maintain only two maps one for the global scope and one for the currently active local scope if there is one when declaring a name context should check availabilty in the global scope then in the local scope if the name is still free the name is declared in local scope if there is one otherwise in the global scope all other cases shall result in errors this is also a good possibility to clean up context a bit especially wrt the declaration functions which are in general inconsistent ,0
+602,10606312912.0,IssuesEvent,2019-10-10 22:54:23,planningcenter/developers,https://api.github.com/repos/planningcenter/developers,closed,Marital Status - Cannot be updated for a person,People,"**Affected Applications**
+People
+
+**Describe the bug**
+We are attempting to update the marital status of a person via the API. The documentation shows we can create new status's but doesn't show how to update the person.
+
+**To Reproduce**
+Steps to reproduce the behavior:
+Attempted a post and a patch request
+
+**Expected behavior**
+To be able to update the person object to have a new martial status.
+
+**Context (please complete the following information):**
+Language: Python or Ruby
+Endpoint: https://api.planningcenteronline.com/people/v2/people/{person_id}/marital_status
+Attempted both a patch and a post
+Returns a 403 Forbidden Error
+
+**Additional context**
+Add any other context about the problem here.
+
+The API documentation shows no ability to address this, but your mobile app, which runs on the API allows for changes. So this is clearly possible. We would love the ability to capture this data to better provide analytics. Thanks.
+",1.0,"Marital Status - Cannot be updated for a person - **Affected Applications**
+People
+
+**Describe the bug**
+We are attempting to update the marital status of a person via the API. The documentation shows we can create new status's but doesn't show how to update the person.
+
+**To Reproduce**
+Steps to reproduce the behavior:
+Attempted a post and a patch request
+
+**Expected behavior**
+To be able to update the person object to have a new martial status.
+
+**Context (please complete the following information):**
+Language: Python or Ruby
+Endpoint: https://api.planningcenteronline.com/people/v2/people/{person_id}/marital_status
+Attempted both a patch and a post
+Returns a 403 Forbidden Error
+
+**Additional context**
+Add any other context about the problem here.
+
+The API documentation shows no ability to address this, but your mobile app, which runs on the API allows for changes. So this is clearly possible. We would love the ability to capture this data to better provide analytics. Thanks.
+",1,marital status cannot be updated for a person affected applications people describe the bug we are attempting to update the marital status of a person via the api the documentation shows we can create new status s but doesn t show how to update the person to reproduce steps to reproduce the behavior attempted a post and a patch request expected behavior to be able to update the person object to have a new martial status context please complete the following information language python or ruby endpoint attempted both a patch and a post returns a forbidden error additional context add any other context about the problem here the api documentation shows no ability to address this but your mobile app which runs on the api allows for changes so this is clearly possible we would love the ability to capture this data to better provide analytics thanks ,1
+194649,14684624690.0,IssuesEvent,2021-01-01 04:04:04,github-vet/rangeloop-pointer-findings,https://api.github.com/repos/github-vet/rangeloop-pointer-findings,closed,itsivareddy/terrafrom-Oci: oci/core_volume_test.go; 16 LoC,fresh small test,"
+Found a possible issue in [itsivareddy/terrafrom-Oci](https://www.github.com/itsivareddy/terrafrom-Oci) at [oci/core_volume_test.go](https://github.com/itsivareddy/terrafrom-Oci/blob/075608a9e201ee0e32484da68d5ba5370dfde1be/oci/core_volume_test.go#L652-L667)
+
+Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message.
+
+> reference to volumeId is reassigned at line 656
+
+[Click here to see the code in its original context.](https://github.com/itsivareddy/terrafrom-Oci/blob/075608a9e201ee0e32484da68d5ba5370dfde1be/oci/core_volume_test.go#L652-L667)
+
+
+Click here to show the 16 line(s) of Go which triggered the analyzer.
+
+```go
+ for _, volumeId := range volumeIds {
+ if ok := SweeperDefaultResourceId[volumeId]; !ok {
+ deleteVolumeRequest := oci_core.DeleteVolumeRequest{}
+
+ deleteVolumeRequest.VolumeId = &volumeId
+
+ deleteVolumeRequest.RequestMetadata.RetryPolicy = getRetryPolicy(true, ""core"")
+ _, error := blockstorageClient.DeleteVolume(context.Background(), deleteVolumeRequest)
+ if error != nil {
+ fmt.Printf(""Error deleting Volume %s %s, It is possible that the resource is already deleted. Please verify manually \n"", volumeId, error)
+ continue
+ }
+ waitTillCondition(testAccProvider, &volumeId, volumeSweepWaitCondition, time.Duration(3*time.Minute),
+ volumeSweepResponseFetchOperation, ""core"", true)
+ }
+ }
+
+```
+
+
+
+
+Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
+See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
+
+commit ID: 075608a9e201ee0e32484da68d5ba5370dfde1be
+",1.0,"itsivareddy/terrafrom-Oci: oci/core_volume_test.go; 16 LoC -
+Found a possible issue in [itsivareddy/terrafrom-Oci](https://www.github.com/itsivareddy/terrafrom-Oci) at [oci/core_volume_test.go](https://github.com/itsivareddy/terrafrom-Oci/blob/075608a9e201ee0e32484da68d5ba5370dfde1be/oci/core_volume_test.go#L652-L667)
+
+Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message.
+
+> reference to volumeId is reassigned at line 656
+
+[Click here to see the code in its original context.](https://github.com/itsivareddy/terrafrom-Oci/blob/075608a9e201ee0e32484da68d5ba5370dfde1be/oci/core_volume_test.go#L652-L667)
+
+
+Click here to show the 16 line(s) of Go which triggered the analyzer.
+
+```go
+ for _, volumeId := range volumeIds {
+ if ok := SweeperDefaultResourceId[volumeId]; !ok {
+ deleteVolumeRequest := oci_core.DeleteVolumeRequest{}
+
+ deleteVolumeRequest.VolumeId = &volumeId
+
+ deleteVolumeRequest.RequestMetadata.RetryPolicy = getRetryPolicy(true, ""core"")
+ _, error := blockstorageClient.DeleteVolume(context.Background(), deleteVolumeRequest)
+ if error != nil {
+ fmt.Printf(""Error deleting Volume %s %s, It is possible that the resource is already deleted. Please verify manually \n"", volumeId, error)
+ continue
+ }
+ waitTillCondition(testAccProvider, &volumeId, volumeSweepWaitCondition, time.Duration(3*time.Minute),
+ volumeSweepResponseFetchOperation, ""core"", true)
+ }
+ }
+
+```
+
+
+
+
+Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
+See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
+
+commit ID: 075608a9e201ee0e32484da68d5ba5370dfde1be
+",0,itsivareddy terrafrom oci oci core volume test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message reference to volumeid is reassigned at line click here to show the line s of go which triggered the analyzer go for volumeid range volumeids if ok sweeperdefaultresourceid ok deletevolumerequest oci core deletevolumerequest deletevolumerequest volumeid volumeid deletevolumerequest requestmetadata retrypolicy getretrypolicy true core error blockstorageclient deletevolume context background deletevolumerequest if error nil fmt printf error deleting volume s s it is possible that the resource is already deleted please verify manually n volumeid error continue waittillcondition testaccprovider volumeid volumesweepwaitcondition time duration time minute volumesweepresponsefetchoperation core true leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id ,0
+878,17117788277.0,IssuesEvent,2021-07-11 18:11:16,Clauvin/Quarter_Circle_Forward_PUNCH_Mod,https://api.github.com/repos/Clauvin/Quarter_Circle_Forward_PUNCH_Mod,opened,Duffel Bag is crashing on zhs (thanks to Steam's Frigid - https://steamcommunity.com/profiles/76561198799518884 - for finding this one),bug localization suggestion from internet people,"As pointed by Steam's Frigid on https://steamcommunity.com/workshop/filedetails/discussion/1848229709/1637543304832643838/ , the description of Duffel Bag was missing a few extra lines that made it crash on the zhs language.",1.0,"Duffel Bag is crashing on zhs (thanks to Steam's Frigid - https://steamcommunity.com/profiles/76561198799518884 - for finding this one) - As pointed by Steam's Frigid on https://steamcommunity.com/workshop/filedetails/discussion/1848229709/1637543304832643838/ , the description of Duffel Bag was missing a few extra lines that made it crash on the zhs language.",1,duffel bag is crashing on zhs thanks to steam s frigid for finding this one as pointed by steam s frigid on the description of duffel bag was missing a few extra lines that made it crash on the zhs language ,1
+15724,9602439096.0,IssuesEvent,2019-05-10 14:36:04,GEWIS/gewisweb,https://api.github.com/repos/GEWIS/gewisweb,closed,Non-user cannot view activity even when allowed,bug security,"PR #859 introduced a requirement in this [commit](https://github.com/GEWIS/gewisweb/commit/2d92f4a950a566f229dbd8e247b2a346a6353a3e#diff-98000bbef9461386cb74746e7b66c51b) for viewing an activity to be done by a user.
+
+While most activities should be invisible (or should they?), this should be handled properly using the permissions via in the acl, as defined in [the module file](https://github.com/GEWIS/gewisweb/blob/master/module/Activity/Module.php). Related is #884, which may be automatically resolved by this issue.",True,"Non-user cannot view activity even when allowed - PR #859 introduced a requirement in this [commit](https://github.com/GEWIS/gewisweb/commit/2d92f4a950a566f229dbd8e247b2a346a6353a3e#diff-98000bbef9461386cb74746e7b66c51b) for viewing an activity to be done by a user.
+
+While most activities should be invisible (or should they?), this should be handled properly using the permissions via in the acl, as defined in [the module file](https://github.com/GEWIS/gewisweb/blob/master/module/Activity/Module.php). Related is #884, which may be automatically resolved by this issue.",0,non user cannot view activity even when allowed pr introduced a requirement in this for viewing an activity to be done by a user while most activities should be invisible or should they this should be handled properly using the permissions via in the acl as defined in related is which may be automatically resolved by this issue ,0
+367423,25738467829.0,IssuesEvent,2022-12-08 03:26:43,pbi-tools/pbi-tools,https://api.github.com/repos/pbi-tools/pbi-tools,closed,Question regarding Power BI Pro Deployments,documentation,"Hi Matthias,
+one of our clients is keen on using pbi-tools for their development/deployments but currently, they are only on power bi pro, thus not able to use xmla deployments. Is there a plan to support also standard [REST](https://learn.microsoft.com/en-us/rest/api/power-bi/imports/post-import) pbix deployments in the future?
+
+I do see, that many cool deploy features like parameterization are not supported, if there is a simple upload of the pbix, but just asking :-)
+
+Cheers, Tarek",1.0,"Question regarding Power BI Pro Deployments - Hi Matthias,
+one of our clients is keen on using pbi-tools for their development/deployments but currently, they are only on power bi pro, thus not able to use xmla deployments. Is there a plan to support also standard [REST](https://learn.microsoft.com/en-us/rest/api/power-bi/imports/post-import) pbix deployments in the future?
+
+I do see, that many cool deploy features like parameterization are not supported, if there is a simple upload of the pbix, but just asking :-)
+
+Cheers, Tarek",0,question regarding power bi pro deployments hi matthias one of our clients is keen on using pbi tools for their development deployments but currently they are only on power bi pro thus not able to use xmla deployments is there a plan to support also standard pbix deployments in the future i do see that many cool deploy features like parameterization are not supported if there is a simple upload of the pbix but just asking cheers tarek,0
+149,3696990381.0,IssuesEvent,2016-02-27 10:37:06,friendlycode/strategic_plan,https://api.github.com/repos/friendlycode/strategic_plan,opened,Activity Level,Membership People & Partnerships,Maintain level of activity of twice monthly meetings with an average of 5 or more participants,1.0,Activity Level - Maintain level of activity of twice monthly meetings with an average of 5 or more participants,1,activity level maintain level of activity of twice monthly meetings with an average of or more participants,1
+1080,26366929987.0,IssuesEvent,2023-01-11 17:16:47,openstates/issues,https://api.github.com/repos/openstates/issues,opened,New HI Committee Scraper,good first issue component:people-data good first scraper,"### We need a new committee scraper written in [spatula](https://jamesturk.github.io/spatula/) for Hawaii.
+
+It should scrape this [webpage](https://www.capitol.hawaii.gov/legislature/committees.aspx?chamber=all) to get:
+- **name**
+ - “Small Business”, “Agriculture”, etc.
+- **chamber**
+ - “upper”, “lower”, or “legislature” (when joint, etc.)
+- **classification**
+ - ex: ""committee"", ""subcommittee""
+- **parent**
+ - *only if it is a subcommittee* - scraper should have a way to determine this
+ - ex: parent=""Natural Resources"" when *sub*committee name=""Forestry""
+- **sources**
+ - each added using `add_source()` method on instance of `ScrapeCommittee` type object
+ - ex: home page for list of committees, specific page for that committee, etc.
+- **members**
+ - - each added using `add_member()`method on instance of `ScrapeCommittee` type object
+ - *Attributes:*
+ - **name**: “Jane Doe”, “John Smith”, etc.
+ - **role** (where applicable): “Chair”, “Ranking Member”, etc.
+
+#### [Here is helpful documentation](https://docs.openstates.org/contributing/writing-a-committee-scraper/) for writing a committee scraper
+
+### Useful scrapers for reference
+An [old non-spatula-based scraper](https://github.com/openstates/openstates-scrapers/blob/main/scrapers/hi/committees.py) that may be helpful in understanding how we have captured the data in the past.
+
+A [spatula-based scraper](https://github.com/openstates/openstates-scrapers/blob/main/scrapers_next/mo/committees.py) that also parses an HTML page for reference.
+- As in the other spatula scraper, you will need to set the session variable separately (rather than depend on the `__init__.py`).
+
+### Other useful resources
+You can reference the [object models for committee scrapers](https://github.com/openstates/openstates-core/blob/main/openstates/models/committees.py) in the Open States Core repo.
+
+Further documentation on running spatula scrapers in the command line can be found [here](https://docs.openstates.org/contributing/scrapers/#running-spatula-scrapers).",1.0,"New HI Committee Scraper - ### We need a new committee scraper written in [spatula](https://jamesturk.github.io/spatula/) for Hawaii.
+
+It should scrape this [webpage](https://www.capitol.hawaii.gov/legislature/committees.aspx?chamber=all) to get:
+- **name**
+ - “Small Business”, “Agriculture”, etc.
+- **chamber**
+ - “upper”, “lower”, or “legislature” (when joint, etc.)
+- **classification**
+ - ex: ""committee"", ""subcommittee""
+- **parent**
+ - *only if it is a subcommittee* - scraper should have a way to determine this
+ - ex: parent=""Natural Resources"" when *sub*committee name=""Forestry""
+- **sources**
+ - each added using `add_source()` method on instance of `ScrapeCommittee` type object
+ - ex: home page for list of committees, specific page for that committee, etc.
+- **members**
+ - - each added using `add_member()`method on instance of `ScrapeCommittee` type object
+ - *Attributes:*
+ - **name**: “Jane Doe”, “John Smith”, etc.
+ - **role** (where applicable): “Chair”, “Ranking Member”, etc.
+
+#### [Here is helpful documentation](https://docs.openstates.org/contributing/writing-a-committee-scraper/) for writing a committee scraper
+
+### Useful scrapers for reference
+An [old non-spatula-based scraper](https://github.com/openstates/openstates-scrapers/blob/main/scrapers/hi/committees.py) that may be helpful in understanding how we have captured the data in the past.
+
+A [spatula-based scraper](https://github.com/openstates/openstates-scrapers/blob/main/scrapers_next/mo/committees.py) that also parses an HTML page for reference.
+- As in the other spatula scraper, you will need to set the session variable separately (rather than depend on the `__init__.py`).
+
+### Other useful resources
+You can reference the [object models for committee scrapers](https://github.com/openstates/openstates-core/blob/main/openstates/models/committees.py) in the Open States Core repo.
+
+Further documentation on running spatula scrapers in the command line can be found [here](https://docs.openstates.org/contributing/scrapers/#running-spatula-scrapers).",1,new hi committee scraper we need a new committee scraper written in for hawaii it should scrape this to get name “small business” “agriculture” etc chamber “upper” “lower” or “legislature” when joint etc classification ex committee subcommittee parent only if it is a subcommittee scraper should have a way to determine this ex parent natural resources when sub committee name forestry sources each added using add source method on instance of scrapecommittee type object ex home page for list of committees specific page for that committee etc members each added using add member method on instance of scrapecommittee type object attributes name “jane doe” “john smith” etc role where applicable “chair” “ranking member” etc for writing a committee scraper useful scrapers for reference an that may be helpful in understanding how we have captured the data in the past a that also parses an html page for reference as in the other spatula scraper you will need to set the session variable separately rather than depend on the init py other useful resources you can reference the in the open states core repo further documentation on running spatula scrapers in the command line can be found ,1
+787499,27719619827.0,IssuesEvent,2023-03-14 19:30:22,DSC-iTC/cPP,https://api.github.com/repos/DSC-iTC/cPP,closed,CCDB AU/001 - Whole document,SD editorial 1+ Priority,"
+
+### Comment:
+There are no page numbers.
+
+### Proposal:
+Please put in page numbers.
+
+### Priority: 2",1.0,"CCDB AU/001 - Whole document -
+
+### Comment:
+There are no page numbers.
+
+### Proposal:
+Please put in page numbers.
+
+### Priority: 2",0,ccdb au whole document comment there are no page numbers proposal please put in page numbers priority ,0
+18455,3690713002.0,IssuesEvent,2016-02-25 21:03:51,cockroachdb/cockroach,https://api.github.com/repos/cockroachdb/cockroach,closed,circleci: failed tests (13427): TestSSLEnforcement ,test-failure,"The following test appears to have failed:
+
+[#13427](https://circleci.com/gh/cockroachdb/cockroach/13427):
+
+```
+E160225 20:40:58.939951 server.go:2159 http: TLS handshake error from 127.0.0.1:60760: tls: first record does not look like a TLS handshake
+E160225 20:40:58.967439 server.go:2159 http: TLS handshake error from 127.0.0.1:60761: tls: first record does not look like a TLS handshake
+I160225 20:40:59.014594 stopper.go:273 draining; tasks left:
+1 server/server.go:450
+I160225 20:40:59.018361 storage/engine/rocksdb.go:137 closing in-memory rocksdb instance
+--- FAIL: TestSSLEnforcement (1.15s)
+ authentication_test.go:148: [28]: expected success=true, got err=Post https://127.0.0.1:37385/sql/Execute: stream error: stream ID 13; STREAM_CLOSED
+=== RUN TestParseStores
+I160225 20:40:59.033331 server/context.go:165 can't read available memory from cgroups (open /sys/fs/cgroup/memory/memory.limit_in_bytes: no such file or directory)
+ setting default rocksdb cache size to 30 GiB (half of system memory)
+I160225 20:40:59.033829 server/context.go:165 can't read available memory from cgroups (open /sys/fs/cgroup/memory/memory.limit_in_bytes: no such file or directory)
+ setting default rocksdb cache size to 30 GiB (half of system memory)
+I160225 20:40:59.034130 server/context.go:225 1 storage engine(s) specified
+I160225 20:40:59.034595 server/context.go:165 can't read available memory from cgroups (open /sys/fs/cgroup/memory/memory.limit_in_bytes: no such file or directory)
+ setting default rocksdb cache size to 30 GiB (half of system memory)
+I160225 20:40:59.034747 server/context.go:225 2 storage engine(s) specified
+--
+W160225 20:41:21.004827 kv/send.go:200 error reply: server/node.go:594: node 1 stopped
+I160225 20:41:21.005037 client/db.go:488 failed batch: failed to send RPC: too many errors encountered (1 of 1 total): server/node.go:594: node 1 stopped
+W160225 20:41:21.005266 ts/db.go:96 error writing time series data: failed to send RPC: too many errors encountered (1 of 1 total): server/node.go:594: node 1 stopped
+I160225 20:41:21.006592 storage/engine/rocksdb.go:137 closing in-memory rocksdb instance
+--- PASS: TestMetricsRecording (1.35s)
+FAIL
+FAIL github.com/cockroachdb/cockroach/server 27.301s
+=== RUN TestMetricsRecorder
+--- PASS: TestMetricsRecorder (0.05s)
+=== RUN TestRuntimeStatRecorder
+I160225 20:41:08.743295 server/status/runtime.go:137 runtime stats: 4 goroutines, 4.86MiB active, 50000000.00cgo/sec, 847490.00/363210.00 %(u/s)time, 6881.69 %gc (4x)
+--- PASS: TestRuntimeStatRecorder (0.00s)
+PASS
+ok github.com/cockroachdb/cockroach/server/status 1.119s
+=== RUN TestSplitOrExpr
+--- PASS: TestSplitOrExpr (0.02s)
+=== RUN TestSplitAndExpr
+
+```
+Please assign, take a look and update the issue accordingly.",1.0,"circleci: failed tests (13427): TestSSLEnforcement - The following test appears to have failed:
+
+[#13427](https://circleci.com/gh/cockroachdb/cockroach/13427):
+
+```
+E160225 20:40:58.939951 server.go:2159 http: TLS handshake error from 127.0.0.1:60760: tls: first record does not look like a TLS handshake
+E160225 20:40:58.967439 server.go:2159 http: TLS handshake error from 127.0.0.1:60761: tls: first record does not look like a TLS handshake
+I160225 20:40:59.014594 stopper.go:273 draining; tasks left:
+1 server/server.go:450
+I160225 20:40:59.018361 storage/engine/rocksdb.go:137 closing in-memory rocksdb instance
+--- FAIL: TestSSLEnforcement (1.15s)
+ authentication_test.go:148: [28]: expected success=true, got err=Post https://127.0.0.1:37385/sql/Execute: stream error: stream ID 13; STREAM_CLOSED
+=== RUN TestParseStores
+I160225 20:40:59.033331 server/context.go:165 can't read available memory from cgroups (open /sys/fs/cgroup/memory/memory.limit_in_bytes: no such file or directory)
+ setting default rocksdb cache size to 30 GiB (half of system memory)
+I160225 20:40:59.033829 server/context.go:165 can't read available memory from cgroups (open /sys/fs/cgroup/memory/memory.limit_in_bytes: no such file or directory)
+ setting default rocksdb cache size to 30 GiB (half of system memory)
+I160225 20:40:59.034130 server/context.go:225 1 storage engine(s) specified
+I160225 20:40:59.034595 server/context.go:165 can't read available memory from cgroups (open /sys/fs/cgroup/memory/memory.limit_in_bytes: no such file or directory)
+ setting default rocksdb cache size to 30 GiB (half of system memory)
+I160225 20:40:59.034747 server/context.go:225 2 storage engine(s) specified
+--
+W160225 20:41:21.004827 kv/send.go:200 error reply: server/node.go:594: node 1 stopped
+I160225 20:41:21.005037 client/db.go:488 failed batch: failed to send RPC: too many errors encountered (1 of 1 total): server/node.go:594: node 1 stopped
+W160225 20:41:21.005266 ts/db.go:96 error writing time series data: failed to send RPC: too many errors encountered (1 of 1 total): server/node.go:594: node 1 stopped
+I160225 20:41:21.006592 storage/engine/rocksdb.go:137 closing in-memory rocksdb instance
+--- PASS: TestMetricsRecording (1.35s)
+FAIL
+FAIL github.com/cockroachdb/cockroach/server 27.301s
+=== RUN TestMetricsRecorder
+--- PASS: TestMetricsRecorder (0.05s)
+=== RUN TestRuntimeStatRecorder
+I160225 20:41:08.743295 server/status/runtime.go:137 runtime stats: 4 goroutines, 4.86MiB active, 50000000.00cgo/sec, 847490.00/363210.00 %(u/s)time, 6881.69 %gc (4x)
+--- PASS: TestRuntimeStatRecorder (0.00s)
+PASS
+ok github.com/cockroachdb/cockroach/server/status 1.119s
+=== RUN TestSplitOrExpr
+--- PASS: TestSplitOrExpr (0.02s)
+=== RUN TestSplitAndExpr
+
+```
+Please assign, take a look and update the issue accordingly.",0,circleci failed tests testsslenforcement the following test appears to have failed server go http tls handshake error from tls first record does not look like a tls handshake server go http tls handshake error from tls first record does not look like a tls handshake stopper go draining tasks left server server go storage engine rocksdb go closing in memory rocksdb instance fail testsslenforcement authentication test go expected success true got err post stream error stream id stream closed run testparsestores server context go can t read available memory from cgroups open sys fs cgroup memory memory limit in bytes no such file or directory setting default rocksdb cache size to gib half of system memory server context go can t read available memory from cgroups open sys fs cgroup memory memory limit in bytes no such file or directory setting default rocksdb cache size to gib half of system memory server context go storage engine s specified server context go can t read available memory from cgroups open sys fs cgroup memory memory limit in bytes no such file or directory setting default rocksdb cache size to gib half of system memory server context go storage engine s specified kv send go error reply server node go node stopped client db go failed batch failed to send rpc too many errors encountered of total server node go node stopped ts db go error writing time series data failed to send rpc too many errors encountered of total server node go node stopped storage engine rocksdb go closing in memory rocksdb instance pass testmetricsrecording fail fail github com cockroachdb cockroach server run testmetricsrecorder pass testmetricsrecorder run testruntimestatrecorder server status runtime go runtime stats goroutines active sec u s time gc pass testruntimestatrecorder pass ok github com cockroachdb cockroach server status run testsplitorexpr pass testsplitorexpr run testsplitandexpr please assign take a look and update the issue accordingly ,0
+220231,7354416680.0,IssuesEvent,2018-03-09 06:44:56,Motoxpro/WorldCupStatsSite,https://api.github.com/repos/Motoxpro/WorldCupStatsSite,closed,Get forecasted weather data,Medium Priority Data Issue MySQL Weather,"Finish function to get forecasted weather from venue date data
+
+Create a table
+
+Add it to DB",1.0,"Get forecasted weather data - Finish function to get forecasted weather from venue date data
+
+Create a table
+
+Add it to DB",0,get forecasted weather data finish function to get forecasted weather from venue date data create a table add it to db,0
+201,4240673890.0,IssuesEvent,2016-07-06 14:10:24,wordpress-mobile/WordPress-Android,https://api.github.com/repos/wordpress-mobile/WordPress-Android,opened,People: Should not be possible to enter more than 10 usernames/emails in invite form,People Management [Type] Bug,"### Expected behavior
+
+When inviting users to join a site, I expect the invite form to limit me to 10 usernames/emails at a time, like in Calypso web:
+
+
+
+After entering 10 usernames/emails in that web form, the ""Usernames or Emails"" field won't accept any more entries.
+
+### Actual behavior
+
+I can enter as many usernames/emails as I want in the ""Usernames or Emails"" field in the invite form.
+
+However, after I hit ""Send"" 10 of the usernames/emails I entered (not necessarily the first 10) disappear from the form, and this message appears: `An error occured while trying to send the invite!`
+
+After that message disappears, this message appears: `Invite sent successfully`
+
+The 10 usernames/emails that disappeared receive an invite, but any remaining usernames/emails in the invite form don't get the invite. It isn't clear from the app messaging what happened and why.
+
+### Steps to reproduce the behavior
+
+1. Go to the site dashboard > People in the app
+2. Tap the plus icon in the top right to invite new users
+3. In the ""Usernames or Emails"" field, enter at least 11 usernames/emails
+4. Tap SEND in the top right to send the invite
+
+##### Tested on Nexus 9, Android 6.0.1, WPAndroid alpha-16",1.0,"People: Should not be possible to enter more than 10 usernames/emails in invite form - ### Expected behavior
+
+When inviting users to join a site, I expect the invite form to limit me to 10 usernames/emails at a time, like in Calypso web:
+
+
+
+After entering 10 usernames/emails in that web form, the ""Usernames or Emails"" field won't accept any more entries.
+
+### Actual behavior
+
+I can enter as many usernames/emails as I want in the ""Usernames or Emails"" field in the invite form.
+
+However, after I hit ""Send"" 10 of the usernames/emails I entered (not necessarily the first 10) disappear from the form, and this message appears: `An error occured while trying to send the invite!`
+
+After that message disappears, this message appears: `Invite sent successfully`
+
+The 10 usernames/emails that disappeared receive an invite, but any remaining usernames/emails in the invite form don't get the invite. It isn't clear from the app messaging what happened and why.
+
+### Steps to reproduce the behavior
+
+1. Go to the site dashboard > People in the app
+2. Tap the plus icon in the top right to invite new users
+3. In the ""Usernames or Emails"" field, enter at least 11 usernames/emails
+4. Tap SEND in the top right to send the invite
+
+##### Tested on Nexus 9, Android 6.0.1, WPAndroid alpha-16",1,people should not be possible to enter more than usernames emails in invite form expected behavior when inviting users to join a site i expect the invite form to limit me to usernames emails at a time like in calypso web after entering usernames emails in that web form the usernames or emails field won t accept any more entries actual behavior i can enter as many usernames emails as i want in the usernames or emails field in the invite form however after i hit send of the usernames emails i entered not necessarily the first disappear from the form and this message appears an error occured while trying to send the invite after that message disappears this message appears invite sent successfully the usernames emails that disappeared receive an invite but any remaining usernames emails in the invite form don t get the invite it isn t clear from the app messaging what happened and why steps to reproduce the behavior go to the site dashboard people in the app tap the plus icon in the top right to invite new users in the usernames or emails field enter at least usernames emails tap send in the top right to send the invite tested on nexus android wpandroid alpha ,1
+374957,26140700157.0,IssuesEvent,2022-12-29 18:00:16,google/jax,https://api.github.com/repos/google/jax,closed,Unclarity documentation tree_flatten.,documentation,"The `is_leaf` option of the tree_flatten function is described as,
+
+> an optionally specified function that will be called at each flattening step. It should return a boolean, which indicates whether the flattening should traverse the current object, or if it should be stopped immediately, with the whole subtree being treated as a leaf.
+
+I just got bitten by this as (to me) it seems to imply `true` should be returned if the object needs to be traversed, and `false` if it is a leaf. I would suggest changing it to something clearer along the lines of:
+
+> an optionally specified function that will be called at each flattening step. It should return a boolean, with true stopping the traversal and the whole subtree being treated as a leaf, and false indicating the flattening should traverse the current object.",1.0,"Unclarity documentation tree_flatten. - The `is_leaf` option of the tree_flatten function is described as,
+
+> an optionally specified function that will be called at each flattening step. It should return a boolean, which indicates whether the flattening should traverse the current object, or if it should be stopped immediately, with the whole subtree being treated as a leaf.
+
+I just got bitten by this as (to me) it seems to imply `true` should be returned if the object needs to be traversed, and `false` if it is a leaf. I would suggest changing it to something clearer along the lines of:
+
+> an optionally specified function that will be called at each flattening step. It should return a boolean, with true stopping the traversal and the whole subtree being treated as a leaf, and false indicating the flattening should traverse the current object.",0,unclarity documentation tree flatten the is leaf option of the tree flatten function is described as an optionally specified function that will be called at each flattening step it should return a boolean which indicates whether the flattening should traverse the current object or if it should be stopped immediately with the whole subtree being treated as a leaf i just got bitten by this as to me it seems to imply true should be returned if the object needs to be traversed and false if it is a leaf i would suggest changing it to something clearer along the lines of an optionally specified function that will be called at each flattening step it should return a boolean with true stopping the traversal and the whole subtree being treated as a leaf and false indicating the flattening should traverse the current object ,0
+69934,9359012585.0,IssuesEvent,2019-04-02 05:09:19,jupyterlab/jupyterlab,https://api.github.com/repos/jupyterlab/jupyterlab,closed,Add workspace clearing trick to faq,good first issue tag:Documentation,@ian-r-rose had a nice trick at https://github.com/jupyterlab/jupyterlab/issues/3872#issuecomment-459366460 - I think it would be great to have it in a faq.,1.0,Add workspace clearing trick to faq - @ian-r-rose had a nice trick at https://github.com/jupyterlab/jupyterlab/issues/3872#issuecomment-459366460 - I think it would be great to have it in a faq.,0,add workspace clearing trick to faq ian r rose had a nice trick at i think it would be great to have it in a faq ,0
+1140,28507102721.0,IssuesEvent,2023-04-18 22:44:09,Clicki-Digital/uptime,https://api.github.com/repos/Clicki-Digital/uptime,closed,⚠️ Bookkeeper to the Smart People has degraded performance,status bookkeeper-to-the-smart-people,"In [`95cabf4`](https://github.com/Clicki-Digital/uptime/commit/95cabf496974af7b6bfbee2adb0fac5f7ac81cd6
+), Bookkeeper to the Smart People (https://bksp.com.au) experienced **degraded performance**:
+- HTTP code: 200
+- Response time: 7502 ms
+",1.0,"⚠️ Bookkeeper to the Smart People has degraded performance - In [`95cabf4`](https://github.com/Clicki-Digital/uptime/commit/95cabf496974af7b6bfbee2adb0fac5f7ac81cd6
+), Bookkeeper to the Smart People (https://bksp.com.au) experienced **degraded performance**:
+- HTTP code: 200
+- Response time: 7502 ms
+",1,⚠️ bookkeeper to the smart people has degraded performance in bookkeeper to the smart people experienced degraded performance http code response time ms ,1
+136634,12728680073.0,IssuesEvent,2020-06-25 03:23:45,PaloAltoNetworks/aws-traffic-mirroring,https://api.github.com/repos/PaloAltoNetworks/aws-traffic-mirroring,opened,Any Update on VPC Mirroring,documentation,@ jigarshah04 Do you have any updated templates od documentation - I have followed the documentation but still coming up short. Any update and insights would be appreciated. ,1.0,Any Update on VPC Mirroring - @ jigarshah04 Do you have any updated templates od documentation - I have followed the documentation but still coming up short. Any update and insights would be appreciated. ,0,any update on vpc mirroring do you have any updated templates od documentation i have followed the documentation but still coming up short any update and insights would be appreciated ,0
+359,6705910101.0,IssuesEvent,2017-10-12 03:28:13,jenntrass/halloween-party,https://api.github.com/repos/jenntrass/halloween-party,closed,Have event invites printed & mailed to attendees. ,logistics People,Send over PSD to Fedex for printing. Mail them to guest list day of receipt. ,1.0,Have event invites printed & mailed to attendees. - Send over PSD to Fedex for printing. Mail them to guest list day of receipt. ,1,have event invites printed mailed to attendees send over psd to fedex for printing mail them to guest list day of receipt ,1
+66007,27300032858.0,IssuesEvent,2023-02-24 00:36:50,Azure/azure-sdk-for-java,https://api.github.com/repos/Azure/azure-sdk-for-java,closed,[QUERY] Cannot perform operation 'sendMessage' on a disposed sender.,Service Bus customer-reported,"Copied from https://github.com/microsoft/azure-spring-boot/issues/956:
+
+In our Spring boot Java app flow, we were not able to send message using ASB sdk to a topic.
+We also did not see any retries while publishing the message.
+
+**Here is the stack trace**
+```console
+Caused by: java.lang.IllegalStateException: Cannot perform operation 'sendMessage' on a disposed sender.
+ at com.azure.messaging.servicebus.ServiceBusSenderAsyncClient.sendInternal(ServiceBusSenderAsyncClient.java:798)
+ at com.azure.messaging.servicebus.ServiceBusSenderAsyncClient.sendMessage(ServiceBusSenderAsyncClient.java:237)
+ at com.azure.messaging.servicebus.ServiceBusSenderClient.sendMessage(ServiceBusSenderClient.java:183)
+ at
+com.fmr.prdc.postprice.recon.infrastructure.adapter.IncompleteTransactionEventPublisherAdapter.publishIncompleteTransactionEvent(IncompleteTransactionEventPublisherAdapter.java:61)
+ at com.fmr.prdc.postprice.recon.core.facade.BaseReconEventHandler.notifyIncompleteTransactionProcessor(BaseReconEventHandler.java:545)
+ at com.fmr.prdc.postprice.recon.core.facade.ContributionActivityDefaultReconHandler.handleReconSummaryEvent(ContributionActivityDefaultReconHandler.java:84)
+ at com.fmr.prdc.postprice.recon.core.facade.LoanRepaymentsDefaultReconHandler.handleReconSummaryEvent(LoanRepaymentsDefaultReconHandler.java:61)
+ at com.fmr.prdc.postprice.recon.application.consumer.ReconSetupListener.processMessage(ReconSetupListener.java:95)
+ at com.fmr.prdc.postprice.recon.application.consumer.ReconSetupListener.handleReconEvent(ReconSetupListener.java:59)
+ at jdk.internal.reflect.GeneratedMethodAccessor226.invoke(Unknown Source)
+ at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
+ at java.base/java.lang.reflect.Method.invoke(Method.java:566)
+ at org.springframework.messaging.handler.invocation.InvocableHandlerMethod.doInvoke(InvocableHandlerMethod.java:169)
+ at org.springframework.messaging.handler.invocation.InvocableHandlerMethod.invoke(InvocableHandlerMethod.java:119)
+ at org.springframework.jms.listener.adapter.MessagingMessageListenerAdapter.invokeHandler(MessagingMessageListenerAdapter.java:114)
+ ... 10 common frames omitted
+```
+
+**Here is the SDK version we use**
+
+```xml
+
+ com.azure
+ azure-messaging-servicebus
+ 7.10.0
+
+```
+
+**JDK version - 11**
+
+**Spring boot version - 2.6.6**
+
+**Here is the code snippet we use to publsh the message**
+
+```java
+package com.fmr.prdc.postprice.recon.infrastructure.adapter;
+
+import com.azure.messaging.servicebus.ServiceBusMessage;
+import com.azure.messaging.servicebus.ServiceBusSenderClient;
+import com.fasterxml.jackson.core.JsonProcessingException;
+import com.fasterxml.jackson.databind.ObjectMapper;
+import com.fmr.prdc.constants.IncompleteTransactionProcessMsdIds;
+import com.fmr.prdc.constants.JmsHeaderConstants;
+import com.fmr.prdc.contracts.IncompleteTransactionEvent;
+import com.fmr.prdc.contracts.RRKEvent;
+import com.fmr.prdc.postprice.recon.core.ports.outgoing.IncompleteTransactionEventPublisher;
+import lombok.RequiredArgsConstructor;
+import lombok.extern.slf4j.Slf4j;
+import org.springframework.stereotype.Component;
+
+import java.util.HashMap;
+import java.util.Map;
+import java.util.UUID;
+
+@Component
+@Slf4j
+@RequiredArgsConstructor
+public class IncompleteTransactionEventPublisherAdapter implements IncompleteTransactionEventPublisher {
+ private final ObjectMapper objectMapper;
+ private final ServiceBusSenderClient incompleteTransactionEventClient;
+ public static final String APPLICATION_JSON_VALUE = ""application/json"";
+
+ @Override
+ public void publishIncompleteTransactionEvent(IncompleteTransactionEvent incompleteTransactionEvent, IncompleteTransactionProcessMsdIds incompleteTransactionProcessMsdIds) {
+
+
+ var trackingId = UUID.randomUUID().toString();
+ var eventAction = incompleteTransactionProcessMsdIds.toString();
+ var eventRoutingType = ""INCP_"" + incompleteTransactionProcessMsdIds.getIncpTxnMsgId();
+ var msdId = incompleteTransactionProcessMsdIds.toString();
+
+ // Adding Headers
+ Map headers = new HashMap<>();
+ headers.putIfAbsent(JmsHeaderConstants.SOURCE_DOMAIN, ""AP140321"");
+ headers.putIfAbsent(JmsHeaderConstants.VERSION, ""0.1"");
+ headers.putIfAbsent(
+ JmsHeaderConstants.EVENT_ACTION, eventAction);
+ headers.putIfAbsent(JmsHeaderConstants.TRACKING_ID, trackingId);
+
+ // Creating New Event
+ var rrkEvent = RRKEvent.builder().headers(headers).payload(incompleteTransactionEvent).build();
+ ServiceBusMessage serviceBusMessage;
+
+ try {
+ serviceBusMessage = new ServiceBusMessage(
+ objectMapper.writeValueAsString(rrkEvent));
+ } catch (JsonProcessingException jsonProcessingException) {
+ log.error(""JsonProcessingException :"", jsonProcessingException);
+ throw new RuntimeException(""Unable to publish incompleteTransactionEvent"");
+ }
+ Map appProperties = serviceBusMessage.getApplicationProperties();
+ appProperties.put(""reprocessEventRoutingType"", eventRoutingType);
+ serviceBusMessage
+ .setContentType(APPLICATION_JSON_VALUE)
+ .setMessageId(msdId);
+ incompleteTransactionEventClient.sendMessage(serviceBusMessage);
+ log.info(""Published {} Event = {} TrackingId = {}"", eventAction, rrkEvent, trackingId);
+
+ }
+}
+```
+",1.0,"[QUERY] Cannot perform operation 'sendMessage' on a disposed sender. - Copied from https://github.com/microsoft/azure-spring-boot/issues/956:
+
+In our Spring boot Java app flow, we were not able to send message using ASB sdk to a topic.
+We also did not see any retries while publishing the message.
+
+**Here is the stack trace**
+```console
+Caused by: java.lang.IllegalStateException: Cannot perform operation 'sendMessage' on a disposed sender.
+ at com.azure.messaging.servicebus.ServiceBusSenderAsyncClient.sendInternal(ServiceBusSenderAsyncClient.java:798)
+ at com.azure.messaging.servicebus.ServiceBusSenderAsyncClient.sendMessage(ServiceBusSenderAsyncClient.java:237)
+ at com.azure.messaging.servicebus.ServiceBusSenderClient.sendMessage(ServiceBusSenderClient.java:183)
+ at
+com.fmr.prdc.postprice.recon.infrastructure.adapter.IncompleteTransactionEventPublisherAdapter.publishIncompleteTransactionEvent(IncompleteTransactionEventPublisherAdapter.java:61)
+ at com.fmr.prdc.postprice.recon.core.facade.BaseReconEventHandler.notifyIncompleteTransactionProcessor(BaseReconEventHandler.java:545)
+ at com.fmr.prdc.postprice.recon.core.facade.ContributionActivityDefaultReconHandler.handleReconSummaryEvent(ContributionActivityDefaultReconHandler.java:84)
+ at com.fmr.prdc.postprice.recon.core.facade.LoanRepaymentsDefaultReconHandler.handleReconSummaryEvent(LoanRepaymentsDefaultReconHandler.java:61)
+ at com.fmr.prdc.postprice.recon.application.consumer.ReconSetupListener.processMessage(ReconSetupListener.java:95)
+ at com.fmr.prdc.postprice.recon.application.consumer.ReconSetupListener.handleReconEvent(ReconSetupListener.java:59)
+ at jdk.internal.reflect.GeneratedMethodAccessor226.invoke(Unknown Source)
+ at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
+ at java.base/java.lang.reflect.Method.invoke(Method.java:566)
+ at org.springframework.messaging.handler.invocation.InvocableHandlerMethod.doInvoke(InvocableHandlerMethod.java:169)
+ at org.springframework.messaging.handler.invocation.InvocableHandlerMethod.invoke(InvocableHandlerMethod.java:119)
+ at org.springframework.jms.listener.adapter.MessagingMessageListenerAdapter.invokeHandler(MessagingMessageListenerAdapter.java:114)
+ ... 10 common frames omitted
+```
+
+**Here is the SDK version we use**
+
+```xml
+
+ com.azure
+ azure-messaging-servicebus
+ 7.10.0
+
+```
+
+**JDK version - 11**
+
+**Spring boot version - 2.6.6**
+
+**Here is the code snippet we use to publsh the message**
+
+```java
+package com.fmr.prdc.postprice.recon.infrastructure.adapter;
+
+import com.azure.messaging.servicebus.ServiceBusMessage;
+import com.azure.messaging.servicebus.ServiceBusSenderClient;
+import com.fasterxml.jackson.core.JsonProcessingException;
+import com.fasterxml.jackson.databind.ObjectMapper;
+import com.fmr.prdc.constants.IncompleteTransactionProcessMsdIds;
+import com.fmr.prdc.constants.JmsHeaderConstants;
+import com.fmr.prdc.contracts.IncompleteTransactionEvent;
+import com.fmr.prdc.contracts.RRKEvent;
+import com.fmr.prdc.postprice.recon.core.ports.outgoing.IncompleteTransactionEventPublisher;
+import lombok.RequiredArgsConstructor;
+import lombok.extern.slf4j.Slf4j;
+import org.springframework.stereotype.Component;
+
+import java.util.HashMap;
+import java.util.Map;
+import java.util.UUID;
+
+@Component
+@Slf4j
+@RequiredArgsConstructor
+public class IncompleteTransactionEventPublisherAdapter implements IncompleteTransactionEventPublisher {
+ private final ObjectMapper objectMapper;
+ private final ServiceBusSenderClient incompleteTransactionEventClient;
+ public static final String APPLICATION_JSON_VALUE = ""application/json"";
+
+ @Override
+ public void publishIncompleteTransactionEvent(IncompleteTransactionEvent incompleteTransactionEvent, IncompleteTransactionProcessMsdIds incompleteTransactionProcessMsdIds) {
+
+
+ var trackingId = UUID.randomUUID().toString();
+ var eventAction = incompleteTransactionProcessMsdIds.toString();
+ var eventRoutingType = ""INCP_"" + incompleteTransactionProcessMsdIds.getIncpTxnMsgId();
+ var msdId = incompleteTransactionProcessMsdIds.toString();
+
+ // Adding Headers
+ Map headers = new HashMap<>();
+ headers.putIfAbsent(JmsHeaderConstants.SOURCE_DOMAIN, ""AP140321"");
+ headers.putIfAbsent(JmsHeaderConstants.VERSION, ""0.1"");
+ headers.putIfAbsent(
+ JmsHeaderConstants.EVENT_ACTION, eventAction);
+ headers.putIfAbsent(JmsHeaderConstants.TRACKING_ID, trackingId);
+
+ // Creating New Event
+ var rrkEvent = RRKEvent.builder().headers(headers).payload(incompleteTransactionEvent).build();
+ ServiceBusMessage serviceBusMessage;
+
+ try {
+ serviceBusMessage = new ServiceBusMessage(
+ objectMapper.writeValueAsString(rrkEvent));
+ } catch (JsonProcessingException jsonProcessingException) {
+ log.error(""JsonProcessingException :"", jsonProcessingException);
+ throw new RuntimeException(""Unable to publish incompleteTransactionEvent"");
+ }
+ Map appProperties = serviceBusMessage.getApplicationProperties();
+ appProperties.put(""reprocessEventRoutingType"", eventRoutingType);
+ serviceBusMessage
+ .setContentType(APPLICATION_JSON_VALUE)
+ .setMessageId(msdId);
+ incompleteTransactionEventClient.sendMessage(serviceBusMessage);
+ log.info(""Published {} Event = {} TrackingId = {}"", eventAction, rrkEvent, trackingId);
+
+ }
+}
+```
+",0, cannot perform operation sendmessage on a disposed sender copied from in our spring boot java app flow we were not able to send message using asb sdk to a topic we also did not see any retries while publishing the message here is the stack trace console caused by java lang illegalstateexception cannot perform operation sendmessage on a disposed sender at com azure messaging servicebus servicebussenderasyncclient sendinternal servicebussenderasyncclient java at com azure messaging servicebus servicebussenderasyncclient sendmessage servicebussenderasyncclient java at com azure messaging servicebus servicebussenderclient sendmessage servicebussenderclient java at com fmr prdc postprice recon infrastructure adapter incompletetransactioneventpublisheradapter publishincompletetransactionevent incompletetransactioneventpublisheradapter java at com fmr prdc postprice recon core facade basereconeventhandler notifyincompletetransactionprocessor basereconeventhandler java at com fmr prdc postprice recon core facade contributionactivitydefaultreconhandler handlereconsummaryevent contributionactivitydefaultreconhandler java at com fmr prdc postprice recon core facade loanrepaymentsdefaultreconhandler handlereconsummaryevent loanrepaymentsdefaultreconhandler java at com fmr prdc postprice recon application consumer reconsetuplistener processmessage reconsetuplistener java at com fmr prdc postprice recon application consumer reconsetuplistener handlereconevent reconsetuplistener java at jdk internal reflect invoke unknown source at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at org springframework messaging handler invocation invocablehandlermethod doinvoke invocablehandlermethod java at org springframework messaging handler invocation invocablehandlermethod invoke invocablehandlermethod java at org springframework jms listener adapter messagingmessagelisteneradapter invokehandler messagingmessagelisteneradapter java common frames omitted here is the sdk version we use xml com azure azure messaging servicebus jdk version spring boot version here is the code snippet we use to publsh the message java package com fmr prdc postprice recon infrastructure adapter import com azure messaging servicebus servicebusmessage import com azure messaging servicebus servicebussenderclient import com fasterxml jackson core jsonprocessingexception import com fasterxml jackson databind objectmapper import com fmr prdc constants incompletetransactionprocessmsdids import com fmr prdc constants jmsheaderconstants import com fmr prdc contracts incompletetransactionevent import com fmr prdc contracts rrkevent import com fmr prdc postprice recon core ports outgoing incompletetransactioneventpublisher import lombok requiredargsconstructor import lombok extern import org springframework stereotype component import java util hashmap import java util map import java util uuid component requiredargsconstructor public class incompletetransactioneventpublisheradapter implements incompletetransactioneventpublisher private final objectmapper objectmapper private final servicebussenderclient incompletetransactioneventclient public static final string application json value application json override public void publishincompletetransactionevent incompletetransactionevent incompletetransactionevent incompletetransactionprocessmsdids incompletetransactionprocessmsdids var trackingid uuid randomuuid tostring var eventaction incompletetransactionprocessmsdids tostring var eventroutingtype incp incompletetransactionprocessmsdids getincptxnmsgid var msdid incompletetransactionprocessmsdids tostring adding headers map headers new hashmap headers putifabsent jmsheaderconstants source domain headers putifabsent jmsheaderconstants version headers putifabsent jmsheaderconstants event action eventaction headers putifabsent jmsheaderconstants tracking id trackingid creating new event var rrkevent rrkevent builder headers headers payload incompletetransactionevent build servicebusmessage servicebusmessage try servicebusmessage new servicebusmessage objectmapper writevalueasstring rrkevent catch jsonprocessingexception jsonprocessingexception log error jsonprocessingexception jsonprocessingexception throw new runtimeexception unable to publish incompletetransactionevent map appproperties servicebusmessage getapplicationproperties appproperties put reprocesseventroutingtype eventroutingtype servicebusmessage setcontenttype application json value setmessageid msdid incompletetransactioneventclient sendmessage servicebusmessage log info published event trackingid eventaction rrkevent trackingid ,0
+283736,21330222544.0,IssuesEvent,2022-04-18 07:16:39,file-share-platform/ws-com-framework,https://api.github.com/repos/file-share-platform/ws-com-framework,closed,Autobuild/deploy documentation,documentation,"## Description
+Add a gh action to automatically build & deploy gh documentation.
+
+## Tasks/General Thoughts
+- [x] Add gh action
+- [x] setup auto-deploy to gh-page
+- [x] add badge to readme
+- [x] Link to docs in repo description",1.0,"Autobuild/deploy documentation - ## Description
+Add a gh action to automatically build & deploy gh documentation.
+
+## Tasks/General Thoughts
+- [x] Add gh action
+- [x] setup auto-deploy to gh-page
+- [x] add badge to readme
+- [x] Link to docs in repo description",0,autobuild deploy documentation description add a gh action to automatically build deploy gh documentation tasks general thoughts add gh action setup auto deploy to gh page add badge to readme link to docs in repo description,0
+99,3410277027.0,IssuesEvent,2015-12-04 19:25:31,bbergen/tbmd.com,https://api.github.com/repos/bbergen/tbmd.com,closed,RSS Automation,Movies People Reviews Wow factor,"As a developer, I would like the RSS feed to be updated automatically when new movies, actors and directors are created.
+
+Acceptance Criteria
+
+* RSS feed updated when new entries are added to the site.",1.0,"RSS Automation - As a developer, I would like the RSS feed to be updated automatically when new movies, actors and directors are created.
+
+Acceptance Criteria
+
+* RSS feed updated when new entries are added to the site.",1,rss automation as a developer i would like the rss feed to be updated automatically when new movies actors and directors are created acceptance criteria rss feed updated when new entries are added to the site ,1
+77542,3506777733.0,IssuesEvent,2016-01-08 09:25:32,pmem/issues,https://api.github.com/repos/pmem/issues,opened,libpmemobj man page: wrong number of parameters in POBJ_FREE's prototype,Exposure: Low Priority: 4 low Type: Bug,"Man page says that POBJ_FREE takes two arguments but it should take TOID *oidp only:
+`POBJ_FREE(PMEMobjpool *pop, TOID *oidp)`
+",1.0,"libpmemobj man page: wrong number of parameters in POBJ_FREE's prototype - Man page says that POBJ_FREE takes two arguments but it should take TOID *oidp only:
+`POBJ_FREE(PMEMobjpool *pop, TOID *oidp)`
+",0,libpmemobj man page wrong number of parameters in pobj free s prototype man page says that pobj free takes two arguments but it should take toid oidp only pobj free pmemobjpool pop toid oidp ,0
+94750,8515211193.0,IssuesEvent,2018-10-31 20:52:31,WebliniaERP/webliniaerp-web,https://api.github.com/repos/WebliniaERP/webliniaerp-web,reopened,Incluir mais opções de filtro de produtos,1-melhoria 2-prioridade alta 3- EL SHADDAI GOURMET test,"Cliente deseja filtrar os produtos por tipo:
+- [x] Tipo de Produto (Normal, Composto, Serviço)
+- [x] Sub-tipo (Matéria Prima)",1.0,"Incluir mais opções de filtro de produtos - Cliente deseja filtrar os produtos por tipo:
+- [x] Tipo de Produto (Normal, Composto, Serviço)
+- [x] Sub-tipo (Matéria Prima)",0,incluir mais opções de filtro de produtos cliente deseja filtrar os produtos por tipo tipo de produto normal composto serviço sub tipo matéria prima ,0
+112316,14238305370.0,IssuesEvent,2020-11-18 18:25:50,1Hive/honey-pot,https://api.github.com/repos/1Hive/honey-pot,closed,Administrative votes,needs design,"Once designs are ready we need to create a dandelion voting interface.
+
+Designs Explorations.
+
+https://www.figma.com/file/jBEao1Z0PovUj40tJJG8pY/Honey-Pot-Site?node-id=12%3A9493",1.0,"Administrative votes - Once designs are ready we need to create a dandelion voting interface.
+
+Designs Explorations.
+
+https://www.figma.com/file/jBEao1Z0PovUj40tJJG8pY/Honey-Pot-Site?node-id=12%3A9493",0,administrative votes once designs are ready we need to create a dandelion voting interface designs explorations ,0
+425312,29351280740.0,IssuesEvent,2023-05-27 00:46:06,ClickHouse/ClickHouse,https://api.github.com/repos/ClickHouse/ClickHouse,closed,Document `max_partition_size_to_drop`,comp-documentation,"Max table size to drop is documented here: https://clickhouse.com/docs/en/operations/server-configuration-parameters/settings#max-table-size-to-drop, but `max_partition_size_to_drop` is missing.",1.0,"Document `max_partition_size_to_drop` - Max table size to drop is documented here: https://clickhouse.com/docs/en/operations/server-configuration-parameters/settings#max-table-size-to-drop, but `max_partition_size_to_drop` is missing.",0,document max partition size to drop max table size to drop is documented here but max partition size to drop is missing ,0
+86021,8016714522.0,IssuesEvent,2018-07-25 14:06:42,ValveSoftware/steam-for-linux,https://api.github.com/repos/ValveSoftware/steam-for-linux,closed,"tf2: if the mvm loot review popup is closed, loot reveal stops",MOVED: Team Fortress 2 Need Retest reviewed,"When finishing a mann up mission, a popup shows the earned loot for all players, and the stuff is revealed one-by-one. If I close the popup (eg. to collect the remaining money), and reopen it, the reveal doesn't continue.
+",1.0,"tf2: if the mvm loot review popup is closed, loot reveal stops - When finishing a mann up mission, a popup shows the earned loot for all players, and the stuff is revealed one-by-one. If I close the popup (eg. to collect the remaining money), and reopen it, the reveal doesn't continue.
+",0, if the mvm loot review popup is closed loot reveal stops when finishing a mann up mission a popup shows the earned loot for all players and the stuff is revealed one by one if i close the popup eg to collect the remaining money and reopen it the reveal doesn t continue ,0
+923,19577579598.0,IssuesEvent,2022-01-04 16:58:08,openstates/issues,https://api.github.com/repos/openstates/issues,closed,SD: all legislators: street and city concatenated without delimiter,component:people-data type:bug,"**Issue Description:**
+For _all_ SD legislators, there is no space or delimiter between street and city. Examples:
+```
+609 Kent St.Harrisburg, SD 57032
+1705 Northview Ln.Aberdeen, SD 57401
+4001 W. 74th St.Sioux Falls, SD 57108
+PO Box 312Vermillion, SD 57069
+1 Wyoming St.Rapid City, SD 57701
+PO Box 84124Sioux Falls, SD 57108
+507 N. Smith St.Clark, SD 57225
+1002 W. 5th St.Miller, SD 57362
+40976 294th St.Tripp, SD 57376
+2722 Railroad CircleAberdeen, SD 57401
+115 Woodland DriveMadison, SD 57042
+34328 106th St.Eureka, SD 57437
+7421 W. Luke Dr.Sioux Falls, SD 57106
+2526 Grandview Dr.Rapid City, SD 57701
+```
+
+**URL where this issue may be seen (API or OpenStates.org):**
+https://data.openstates.org/people/current/ut.csv
+
+**Source URL of correct data if applicable:**
+https://sdlegislature.gov/Legislators/Profile/1794/Detail",1.0,"SD: all legislators: street and city concatenated without delimiter - **Issue Description:**
+For _all_ SD legislators, there is no space or delimiter between street and city. Examples:
+```
+609 Kent St.Harrisburg, SD 57032
+1705 Northview Ln.Aberdeen, SD 57401
+4001 W. 74th St.Sioux Falls, SD 57108
+PO Box 312Vermillion, SD 57069
+1 Wyoming St.Rapid City, SD 57701
+PO Box 84124Sioux Falls, SD 57108
+507 N. Smith St.Clark, SD 57225
+1002 W. 5th St.Miller, SD 57362
+40976 294th St.Tripp, SD 57376
+2722 Railroad CircleAberdeen, SD 57401
+115 Woodland DriveMadison, SD 57042
+34328 106th St.Eureka, SD 57437
+7421 W. Luke Dr.Sioux Falls, SD 57106
+2526 Grandview Dr.Rapid City, SD 57701
+```
+
+**URL where this issue may be seen (API or OpenStates.org):**
+https://data.openstates.org/people/current/ut.csv
+
+**Source URL of correct data if applicable:**
+https://sdlegislature.gov/Legislators/Profile/1794/Detail",1,sd all legislators street and city concatenated without delimiter issue description for all sd legislators there is no space or delimiter between street and city examples kent st harrisburg sd northview ln aberdeen sd w st sioux falls sd po box sd wyoming st rapid city sd po box falls sd n smith st clark sd w st miller sd st tripp sd railroad circleaberdeen sd woodland drivemadison sd st eureka sd w luke dr sioux falls sd grandview dr rapid city sd url where this issue may be seen api or openstates org source url of correct data if applicable ,1
+433300,12505529083.0,IssuesEvent,2020-06-02 10:54:08,Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth,https://api.github.com/repos/Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth,closed,Rework accepting new dynasty members,:exclamation: priority critical :question: suggestion - balance :balance_scale: :question: suggestion - feature :star: :question: suggestion :question:,"
+**Describe your suggestion in full detail below:**
+Rework `WCCOU.200`, an event that allows infertile characters to accept new dynasty members, into a decision where you pay something (gold, prestige, piety etc.).
+
+It would be nice to allow vampires and worgens to share their curse with new dynasty members.
+",1.0,"Rework accepting new dynasty members -
+**Describe your suggestion in full detail below:**
+Rework `WCCOU.200`, an event that allows infertile characters to accept new dynasty members, into a decision where you pay something (gold, prestige, piety etc.).
+
+It would be nice to allow vampires and worgens to share their curse with new dynasty members.
+",0,rework accepting new dynasty members do not remove pre existing lines if you want to suggest a few things open a new issue per every suggestion describe your suggestion in full detail below rework wccou an event that allows infertile characters to accept new dynasty members into a decision where you pay something gold prestige piety etc it would be nice to allow vampires and worgens to share their curse with new dynasty members ,0
+431,7929492767.0,IssuesEvent,2018-07-06 15:14:43,pastorhudson/pcobot,https://api.github.com/repos/pastorhudson/pcobot,closed,Re do Phone & Birthday & Set List as Slack Attachments,People Services enhancement,Slack allows nice formatting so the initial commands need to be updated to make them look nicer. ,1.0,Re do Phone & Birthday & Set List as Slack Attachments - Slack allows nice formatting so the initial commands need to be updated to make them look nicer. ,1,re do phone birthday set list as slack attachments slack allows nice formatting so the initial commands need to be updated to make them look nicer ,1
+516,8863273744.0,IssuesEvent,2019-01-10 09:12:38,University-of-Potsdam-MM/Mobile.UP,https://api.github.com/repos/University-of-Potsdam-MM/Mobile.UP,closed,Keine Umlautunterstützung,People bug to check,"
+
+von @PhilippJakupovic
+
+Der Server mag das Sonderzeichen ""ä"" nicht
+
+@alekiy sagt: ""Der Service wird von PHP auf Java umgestellt, das wird jedoch noch etwas dauern und passiert im Rahmen meines Projekts.""",1.0,"Keine Umlautunterstützung - 
+
+von @PhilippJakupovic
+
+Der Server mag das Sonderzeichen ""ä"" nicht
+
+@alekiy sagt: ""Der Service wird von PHP auf Java umgestellt, das wird jedoch noch etwas dauern und passiert im Rahmen meines Projekts.""",1,keine umlautunterstützung von philippjakupovic der server mag das sonderzeichen ä nicht alekiy sagt der service wird von php auf java umgestellt das wird jedoch noch etwas dauern und passiert im rahmen meines projekts ,1
+1184,30787964517.0,IssuesEvent,2023-07-31 14:22:52,culturesofknowledge/emlo-project,https://api.github.com/repos/culturesofknowledge/emlo-project,opened,People: Do not link names/synonyms in search results,people search feedback0625,"Because the names/synonyms/roles, etc. are all ‘live’ linked and underlined in EMLO-Edit-NEW, it’s hard to select a name to copy/paste for use elsewhere. I need to do this often, e.g. copy/paste ‘Gebwiler, Protasius’ from the record above and drop it somewhere else.",1.0,"People: Do not link names/synonyms in search results - Because the names/synonyms/roles, etc. are all ‘live’ linked and underlined in EMLO-Edit-NEW, it’s hard to select a name to copy/paste for use elsewhere. I need to do this often, e.g. copy/paste ‘Gebwiler, Protasius’ from the record above and drop it somewhere else.",1,people do not link names synonyms in search results because the names synonyms roles etc are all ‘live’ linked and underlined in emlo edit new it’s hard to select a name to copy paste for use elsewhere i need to do this often e g copy paste ‘gebwiler protasius’ from the record above and drop it somewhere else ,1
+1135,27713020911.0,IssuesEvent,2023-03-14 15:18:36,culturesofknowledge/emlo-project,https://api.github.com/repos/culturesofknowledge/emlo-project,closed,Sort by is not remembered when I choose the next page,bug people search AR test,"I am in People search. The sort options are not remembered when I click on another page number
+
+Also, does it sort just the 100 records shown on page or all search options? If it's just what is displayed on page, we may need sort in the search form.",1.0,"Sort by is not remembered when I choose the next page - I am in People search. The sort options are not remembered when I click on another page number
+
+Also, does it sort just the 100 records shown on page or all search options? If it's just what is displayed on page, we may need sort in the search form.",1,sort by is not remembered when i choose the next page i am in people search the sort options are not remembered when i click on another page number also does it sort just the records shown on page or all search options if it s just what is displayed on page we may need sort in the search form ,1
+860,16525213266.0,IssuesEvent,2021-05-26 19:08:52,microsoft/botframework-components,https://api.github.com/repos/microsoft/botframework-components,closed,Enterprise People: Evaluate WorkedWith feature,Area: Generators: Enterprise People Owner: PBX backlog,"Evaluate the ""WorkedWith"" feature to verify if it is possible to achieve with Microsoft Graph, and if not, if there is another approach to consider.
+
+
+WorkedWith | Users wants to find who works with a given user | Who worked with Divya? Who does Jason Bajaj work with most frequently? Who works with Tina Jones
+-- | -- | --
+
+
+",1.0,"Enterprise People: Evaluate WorkedWith feature - Evaluate the ""WorkedWith"" feature to verify if it is possible to achieve with Microsoft Graph, and if not, if there is another approach to consider.
+
+
+WorkedWith | Users wants to find who works with a given user | Who worked with Divya? Who does Jason Bajaj work with most frequently? Who works with Tina Jones
+-- | -- | --
+
+
+",1,enterprise people evaluate workedwith feature evaluate the workedwith feature to verify if it is possible to achieve with microsoft graph and if not if there is another approach to consider workedwith users wants to find who works with a given user who worked with divya who does jason bajaj work with most frequently who works with tina jones ,1
+764160,26787462846.0,IssuesEvent,2023-02-01 04:53:46,internetarchive/openlibrary,https://api.github.com/repos/internetarchive/openlibrary,opened,Extend ILE Toolbar to Bulk Edit (subjects) ,Type: Feature Request Type: Epic Needs: Detail Needs: Breakdown Priority: 2 Affects: Librarians Lead: @mheiman Module: Integrated Librarian Environment (ILE Bar),"
+
+### Describe the problem that you'd like solved
+
+
+
+
+### Proposal & Constraints
+
+
+
+
+### Additional context
+
+
+### Stakeholders
+
+
+
+",1.0,"Extend ILE Toolbar to Bulk Edit (subjects) -
+
+### Describe the problem that you'd like solved
+
+
+
+
+### Proposal & Constraints
+
+
+
+
+### Additional context
+
+
+### Stakeholders
+
+
+
+",0,extend ile toolbar to bulk edit subjects describe the problem that you d like solved proposal constraints additional context stakeholders ,0
+177884,6588187936.0,IssuesEvent,2017-09-14 01:24:43,assyrianic/Tagha-Virtual-Machine,https://api.github.com/repos/assyrianic/Tagha-Virtual-Machine,closed,upgrade SIMD opcodes to use 16 bytes instead of 8 bytes,development on hold enhancement low priority,"allows us to do math with eight ints, 16 shorts, and 32 bytes for binary operations!
+",1.0,"upgrade SIMD opcodes to use 16 bytes instead of 8 bytes - allows us to do math with eight ints, 16 shorts, and 32 bytes for binary operations!
+",0,upgrade simd opcodes to use bytes instead of bytes allows us to do math with eight ints shorts and bytes for binary operations ,0
+2847,5809376072.0,IssuesEvent,2017-05-04 13:17:08,AllenFang/react-bootstrap-table,https://api.github.com/repos/AllenFang/react-bootstrap-table,closed,defaultSortName and defaultSortOrder for remote sorting,help wanted inprocess,"it would be nice to have defaultSortName and defaultSortOrder for remote table too. Now if I set them, I'm also supposed to check if they are set and change the query to fetch data accordingly.
+
+I would expect that, if they are set, also the method onSortChange is called or even better, something more efficient, in order not to do 2 api calls on component mounting. ",1.0,"defaultSortName and defaultSortOrder for remote sorting - it would be nice to have defaultSortName and defaultSortOrder for remote table too. Now if I set them, I'm also supposed to check if they are set and change the query to fetch data accordingly.
+
+I would expect that, if they are set, also the method onSortChange is called or even better, something more efficient, in order not to do 2 api calls on component mounting. ",0,defaultsortname and defaultsortorder for remote sorting it would be nice to have defaultsortname and defaultsortorder for remote table too now if i set them i m also supposed to check if they are set and change the query to fetch data accordingly i would expect that if they are set also the method onsortchange is called or even better something more efficient in order not to do api calls on component mounting ,0
+760,13795315581.0,IssuesEvent,2020-10-09 17:48:02,restincode/restincode,https://api.github.com/repos/restincode/restincode,opened,Steve Steinberg (Frank Drake),Add Person Needs Review People,"Please fill out as much information as you can. No fields are required, but the more you can provide the better.
+
+**General Info**
+* First name: Steve
+* Last name: Steinberg
+* Handle: Frank Drake
+* Birth Year: 1970
+* Death Year: 2020
+* Link to Obituary:
+* Group Affiliations:
+* URL to main photo (or attach to Issue): https://i0.wp.com/boingboing.net/wp-content/uploads/2020/10/screenshot-37.jpg?fit=1&resize=620%2C4000&ssl=1
+* Description of person and/or activities: ""R.I.P. Steve Steinberg aka Frank Drake, publisher of WORM and Intertek back in the day - a true hacker.""
+* Facebook memorial group URL:
+
+**Social Media Links**
+* Twitter:
+* Github:
+* LinkedIn:
+* Facebook:
+* Other:
+
+**Contributions**
+Repeat as many times as needed
+
+* Project name:
+* Project URL:
+* Project Description:
+
+**Photo Gallery**
+* URL(s) to additional photos:
+
+https://boingboing.net/2020/10/09/steve-steinberg-hacker-writer-friend-1970-2020.html
+https://twitter.com/2600/status/1314618873107415040
+",1.0,"Steve Steinberg (Frank Drake) - Please fill out as much information as you can. No fields are required, but the more you can provide the better.
+
+**General Info**
+* First name: Steve
+* Last name: Steinberg
+* Handle: Frank Drake
+* Birth Year: 1970
+* Death Year: 2020
+* Link to Obituary:
+* Group Affiliations:
+* URL to main photo (or attach to Issue): https://i0.wp.com/boingboing.net/wp-content/uploads/2020/10/screenshot-37.jpg?fit=1&resize=620%2C4000&ssl=1
+* Description of person and/or activities: ""R.I.P. Steve Steinberg aka Frank Drake, publisher of WORM and Intertek back in the day - a true hacker.""
+* Facebook memorial group URL:
+
+**Social Media Links**
+* Twitter:
+* Github:
+* LinkedIn:
+* Facebook:
+* Other:
+
+**Contributions**
+Repeat as many times as needed
+
+* Project name:
+* Project URL:
+* Project Description:
+
+**Photo Gallery**
+* URL(s) to additional photos:
+
+https://boingboing.net/2020/10/09/steve-steinberg-hacker-writer-friend-1970-2020.html
+https://twitter.com/2600/status/1314618873107415040
+",1,steve steinberg frank drake please fill out as much information as you can no fields are required but the more you can provide the better general info first name steve last name steinberg handle frank drake birth year death year link to obituary group affiliations url to main photo or attach to issue description of person and or activities r i p steve steinberg aka frank drake publisher of worm and intertek back in the day a true hacker facebook memorial group url social media links twitter github linkedin facebook other contributions repeat as many times as needed project name project url project description photo gallery url s to additional photos ,1
+932,20237227571.0,IssuesEvent,2022-02-14 04:19:23,PopulateTools/gobierto,https://api.github.com/repos/PopulateTools/gobierto,closed,"Calendars / Start day at 7AM, not 00AM",gobierto-people gobierto-calendars,"When embedding an agenda in another site with not much height, this situation can happen (no events appear unless you scroll, because only night hours are visible and not much happens during this hours):
+
+
+
+https://www.alcobendas.org/es/persona/roberto-fraile-herrera
+
+Let's configure the calendar so the day starts at 7AM
+",1.0,"Calendars / Start day at 7AM, not 00AM - When embedding an agenda in another site with not much height, this situation can happen (no events appear unless you scroll, because only night hours are visible and not much happens during this hours):
+
+
+
+https://www.alcobendas.org/es/persona/roberto-fraile-herrera
+
+Let's configure the calendar so the day starts at 7AM
+",1,calendars start day at not when embedding an agenda in another site with not much height this situation can happen no events appear unless you scroll because only night hours are visible and not much happens during this hours img width alt cleanshot at src let s configure the calendar so the day starts at ,1
+248405,26785151836.0,IssuesEvent,2023-02-01 01:44:09,raindigi/site-landing,https://api.github.com/repos/raindigi/site-landing,opened,CVE-2022-0686 (High) detected in url-parse-1.4.4.tgz,security vulnerability,"## CVE-2022-0686 - High Severity Vulnerability
+ Vulnerable Library - url-parse-1.4.4.tgz
+
+
Small footprint URL parser that works seamlessly across Node.js and browser environments
+
+
+
+
+
+***
+Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",True,"CVE-2022-0686 (High) detected in url-parse-1.4.4.tgz - ## CVE-2022-0686 - High Severity Vulnerability
+ Vulnerable Library - url-parse-1.4.4.tgz
+
+
Small footprint URL parser that works seamlessly across Node.js and browser environments
+
+
+
+
+
+***
+Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",0,cve high detected in url parse tgz cve high severity vulnerability vulnerable library url parse tgz small footprint url parser that works seamlessly across node js and browser environments library home page a href path to dependency file site landing package json path to vulnerable library node modules url parse package json dependency hierarchy gatsby tgz root library react dev utils tgz sockjs client tgz x url parse tgz vulnerable library found in base branch master vulnerability details authorization bypass through user controlled key in npm url parse prior to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution url parse direct dependency fix resolution gatsby step up your open source security game with mend ,0
+104455,4211846464.0,IssuesEvent,2016-06-29 14:41:10,bitaverage/bitaverage,https://api.github.com/repos/bitaverage/bitaverage,closed,Don't let error messages be cashed by the browser,low priority,If the error message is cashed the user will see it again for a page that has no errors. I think we can do this with set expires headers.,1.0,Don't let error messages be cashed by the browser - If the error message is cashed the user will see it again for a page that has no errors. I think we can do this with set expires headers.,0,don t let error messages be cashed by the browser if the error message is cashed the user will see it again for a page that has no errors i think we can do this with set expires headers ,0
+93056,10764471666.0,IssuesEvent,2019-11-01 08:23:25,openorclose/ped,https://api.github.com/repos/openorclose/ped,opened,Not stated how to get inputs into program,severity.Medium type.DocumentationBug,"Screenshot shows that I need to use scanner, but not explicitly documented.
+
",1.0,"Not stated how to get inputs into program - Screenshot shows that I need to use scanner, but not explicitly documented.
+
",0,not stated how to get inputs into program screenshot shows that i need to use scanner but not explicitly documented ,0
+541854,15834981322.0,IssuesEvent,2021-04-06 17:28:35,carpentries-incubator/SDC-BIDS-dMRI,https://api.github.com/repos/carpentries-incubator/SDC-BIDS-dMRI,closed,Fix utils.visualizations_utils.py path,high priority status:in progress type:bug,"Noted in #128, the module needs to be packaged or added to the `sys.path` for calls in both the actions workflow, as well as going through the notebooks locally.",1.0,"Fix utils.visualizations_utils.py path - Noted in #128, the module needs to be packaged or added to the `sys.path` for calls in both the actions workflow, as well as going through the notebooks locally.",0,fix utils visualizations utils py path noted in the module needs to be packaged or added to the sys path for calls in both the actions workflow as well as going through the notebooks locally ,0
+32558,4776273896.0,IssuesEvent,2016-10-27 13:18:58,AlexZH1979/tutorials-3,https://api.github.com/repos/AlexZH1979/tutorials-3,opened,Tutorial Page my-tutorial.md Issue,Meta tutorials/test/my-tutorial.md,"Tutorial issue found: [https://github.com/AlexZH1979/tutorials-3/blob/master/tutorials/test/my-tutorial.md](https://github.com/AlexZH1979/tutorials-3/blob/master/tutorials/test/my-tutorial.md) is missing meta or contains invalid front matter. Your tutorial in sapcom was not updated.
+ facked service",1.0,"Tutorial Page my-tutorial.md Issue - Tutorial issue found: [https://github.com/AlexZH1979/tutorials-3/blob/master/tutorials/test/my-tutorial.md](https://github.com/AlexZH1979/tutorials-3/blob/master/tutorials/test/my-tutorial.md) is missing meta or contains invalid front matter. Your tutorial in sapcom was not updated.
+ facked service",0,tutorial page my tutorial md issue tutorial issue found is missing meta or contains invalid front matter your tutorial in sapcom was not updated facked service,0
+70618,30699351367.0,IssuesEvent,2023-07-26 21:27:30,hashicorp/terraform-provider-azurerm,https://api.github.com/repos/hashicorp/terraform-provider-azurerm,closed,azurerm_logic_app_trigger_recurrence: time_zone does not affect start_time,bug service/logic,"
+
+
+
+### Community Note
+
+* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
+* Please do not leave ""+1"" or ""me too"" comments, they generate extra noise for issue followers and do not help prioritize the request
+* If you are interested in working on this issue or have submitted a pull request, please leave a comment
+
+
+
+### Terraform (and AzureRM Provider) Version
+
+
+
+### Affected Resource(s)
+
+
+
+* `azurerm_logic_app_trigger_recurrence`
+
+### Terraform Configuration Files
+
+
+
+```hcl
+resource azurerm_resource_group repro {
+ name = ""startdate-timezone-issue""
+ location = ""westeurope""
+}
+
+resource azurerm_logic_app_workflow stop {
+ name = ""stop-workflow""
+ resource_group_name = azurerm_resource_group.repro.name
+ location = ""westeurope""
+}
+
+resource azurerm_logic_app_trigger_recurrence workweek_stop_trigger {
+ name = ""workweek_stop""
+ logic_app_id = azurerm_logic_app_workflow.stop
+ frequency = ""Week""
+ interval = 1
+ schedule {
+ at_these_hours = [21]
+ at_these_minutes = [00]
+ on_these_days = [
+ ""Monday"",
+ ""Tuesday"",
+ ""Wednesday"",
+ ""Thursday"",
+ ""Friday"",
+ ]
+ }
+ # Valid RFC 3339 Date is not accepted
+ # start_time requires a 'Z' at the end, meaning UTC time.
+ start_time = ""2021-12-19T21:00:00""
+ # Time zone will not affect start_time as start_time is always UTC
+ time_zone = ""W. Europe Standard Time""
+}
+```
+
+### Debug Output
+
+
+
+### Panic Output
+
+
+
+### Expected Behaviour
+
+
+Similar behavior as logic app UI/ARM: start_time to accept `2021-12-19T21:00:00` as a valid RFC 3339 date.
+
+### Actual Behaviour
+
+
+start_time requires a `Z` to be appended. This turns the date into a UTC time, so `time_zone` can't effect `start_date`.
+
+### Steps to Reproduce
+
+
+
+1. `terraform plan`
+
+### Important Factoids
+
+
+[RFC3339](https://datatracker.ietf.org/doc/html/rfc3339) refers to the `Z` as an optional element.
+
+### References
+
+
+
+* #0000
+",1.0,"azurerm_logic_app_trigger_recurrence: time_zone does not affect start_time -
+
+
+
+### Community Note
+
+* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
+* Please do not leave ""+1"" or ""me too"" comments, they generate extra noise for issue followers and do not help prioritize the request
+* If you are interested in working on this issue or have submitted a pull request, please leave a comment
+
+
+
+### Terraform (and AzureRM Provider) Version
+
+
+
+### Affected Resource(s)
+
+
+
+* `azurerm_logic_app_trigger_recurrence`
+
+### Terraform Configuration Files
+
+
+
+```hcl
+resource azurerm_resource_group repro {
+ name = ""startdate-timezone-issue""
+ location = ""westeurope""
+}
+
+resource azurerm_logic_app_workflow stop {
+ name = ""stop-workflow""
+ resource_group_name = azurerm_resource_group.repro.name
+ location = ""westeurope""
+}
+
+resource azurerm_logic_app_trigger_recurrence workweek_stop_trigger {
+ name = ""workweek_stop""
+ logic_app_id = azurerm_logic_app_workflow.stop
+ frequency = ""Week""
+ interval = 1
+ schedule {
+ at_these_hours = [21]
+ at_these_minutes = [00]
+ on_these_days = [
+ ""Monday"",
+ ""Tuesday"",
+ ""Wednesday"",
+ ""Thursday"",
+ ""Friday"",
+ ]
+ }
+ # Valid RFC 3339 Date is not accepted
+ # start_time requires a 'Z' at the end, meaning UTC time.
+ start_time = ""2021-12-19T21:00:00""
+ # Time zone will not affect start_time as start_time is always UTC
+ time_zone = ""W. Europe Standard Time""
+}
+```
+
+### Debug Output
+
+
+
+### Panic Output
+
+
+
+### Expected Behaviour
+
+
+Similar behavior as logic app UI/ARM: start_time to accept `2021-12-19T21:00:00` as a valid RFC 3339 date.
+
+### Actual Behaviour
+
+
+start_time requires a `Z` to be appended. This turns the date into a UTC time, so `time_zone` can't effect `start_date`.
+
+### Steps to Reproduce
+
+
+
+1. `terraform plan`
+
+### Important Factoids
+
+
+[RFC3339](https://datatracker.ietf.org/doc/html/rfc3339) refers to the `Z` as an optional element.
+
+### References
+
+
+
+* #0000
+",0,azurerm logic app trigger recurrence time zone does not affect start time please note the following potential times when an issue might be in terraform core or resource ordering issues and issues issues issues spans resources across multiple providers if you are running into one of these scenarios we recommend opening an issue in the instead community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment terraform and azurerm provider version affected resource s azurerm logic app trigger recurrence terraform configuration files hcl resource azurerm resource group repro name startdate timezone issue location westeurope resource azurerm logic app workflow stop name stop workflow resource group name azurerm resource group repro name location westeurope resource azurerm logic app trigger recurrence workweek stop trigger name workweek stop logic app id azurerm logic app workflow stop frequency week interval schedule at these hours at these minutes on these days monday tuesday wednesday thursday friday valid rfc date is not accepted start time requires a z at the end meaning utc time start time time zone will not affect start time as start time is always utc time zone w europe standard time debug output please provide a link to a github gist containing the complete debug output please do not paste the debug output in the issue just paste a link to the gist to obtain the debug output see the panic output expected behaviour similar behavior as logic app ui arm start time to accept as a valid rfc date actual behaviour start time requires a z to be appended this turns the date into a utc time so time zone can t effect start date steps to reproduce terraform plan important factoids refers to the z as an optional element references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here such as vendor documentation ,0
+243505,7858706530.0,IssuesEvent,2018-06-21 14:36:57,larsiusprime/tdrpg-bugs,https://api.github.com/repos/larsiusprime/tdrpg-bugs,closed,"ET dying doesn't trigger the ending, waits forever",1 Battle DQ1 Priority HIGH in progress,"2.0.8c - steam, win7sp1x64, HD mode@1704x960windowed
+
+
+The game has been continuing several minutes at 16X. Everyone's animating, Zelemir is occasionally making everyone have a Frenzy icon, but nothing more is happening. I had just beaten ET by getting his last bearded self down to 0 HP, and he exploded into psi stars IIRC.
+",1.0,"ET dying doesn't trigger the ending, waits forever - 2.0.8c - steam, win7sp1x64, HD mode@1704x960windowed
+
+
+The game has been continuing several minutes at 16X. Everyone's animating, Zelemir is occasionally making everyone have a Frenzy icon, but nothing more is happening. I had just beaten ET by getting his last bearded self down to 0 HP, and he exploded into psi stars IIRC.
+",0,et dying doesn t trigger the ending waits forever steam hd mode the game has been continuing several minutes at everyone s animating zelemir is occasionally making everyone have a frenzy icon but nothing more is happening i had just beaten et by getting his last bearded self down to hp and he exploded into psi stars iirc ,0
+1180,30726744568.0,IssuesEvent,2023-07-27 20:20:56,metabase/metabase,https://api.github.com/repos/metabase/metabase,closed,Password reset is useless when SSO active and password auth is disabled,Type:Bug Priority:P3 Administration/Auth/SSO Administration/People .Team/PixelPolice :police_officer:,"**Describe the bug**
+We shouldn't offer the possibility of doing password resets if SSO is enabled and password auth is disable, it confuses the admins and also the end users, who end up in a page that doesn't do anything.
+
+**Logs**
+NA
+
+**To Reproduce**
+1) add a user
+2) configure MB with SSO and also disable password auth
+3) reset password for the user you created in the first step
+4) see the error
+
+**Expected behavior**
+Remove the option to reset password when password auth is disabled
+
+**Screenshots**
+
+
+**Information about your Metabase Installation:**
+- Metabase version: 44.1
+- Metabase hosting environment: Docker
+
+**Severity**
+P3
+
+**Additional context**
+NA
+",1.0,"Password reset is useless when SSO active and password auth is disabled - **Describe the bug**
+We shouldn't offer the possibility of doing password resets if SSO is enabled and password auth is disable, it confuses the admins and also the end users, who end up in a page that doesn't do anything.
+
+**Logs**
+NA
+
+**To Reproduce**
+1) add a user
+2) configure MB with SSO and also disable password auth
+3) reset password for the user you created in the first step
+4) see the error
+
+**Expected behavior**
+Remove the option to reset password when password auth is disabled
+
+**Screenshots**
+
+
+**Information about your Metabase Installation:**
+- Metabase version: 44.1
+- Metabase hosting environment: Docker
+
+**Severity**
+P3
+
+**Additional context**
+NA
+",1,password reset is useless when sso active and password auth is disabled describe the bug we shouldn t offer the possibility of doing password resets if sso is enabled and password auth is disable it confuses the admins and also the end users who end up in a page that doesn t do anything logs na to reproduce add a user configure mb with sso and also disable password auth reset password for the user you created in the first step see the error expected behavior remove the option to reset password when password auth is disabled screenshots information about your metabase installation metabase version metabase hosting environment docker severity additional context na ,1
+842,15784913511.0,IssuesEvent,2021-04-01 15:41:25,microsoftgraph/microsoft-graph-toolkit,https://api.github.com/repos/microsoftgraph/microsoft-graph-toolkit,closed,Extend People with support for specifying a Graph query,Component: people State: In Review feature-request good first issue work in progress,"# Proposal: Extend People with support for specifying a Graph query
+
+## Description
+
+Extend People with support for specifying a Graph query
+
+## Rationale
+
+Adding support for specifying Graph query to select people to display would be invaluable for scenarios like showing direct reports or people who match certain criteria.
+
+## Preferred Solution
+
+Introduce new attribute named `people-query` that allows specifying the Graph query. Developers should be able to set this value through code using the `peopleQuery` property.
+
+`peopleQuery` is ignored if `peopleQueries` or `people` are specified.
+
+## Additional Context
+
+/cc: @aycabas",1.0,"Extend People with support for specifying a Graph query - # Proposal: Extend People with support for specifying a Graph query
+
+## Description
+
+Extend People with support for specifying a Graph query
+
+## Rationale
+
+Adding support for specifying Graph query to select people to display would be invaluable for scenarios like showing direct reports or people who match certain criteria.
+
+## Preferred Solution
+
+Introduce new attribute named `people-query` that allows specifying the Graph query. Developers should be able to set this value through code using the `peopleQuery` property.
+
+`peopleQuery` is ignored if `peopleQueries` or `people` are specified.
+
+## Additional Context
+
+/cc: @aycabas",1,extend people with support for specifying a graph query proposal extend people with support for specifying a graph query description extend people with support for specifying a graph query rationale adding support for specifying graph query to select people to display would be invaluable for scenarios like showing direct reports or people who match certain criteria preferred solution introduce new attribute named people query that allows specifying the graph query developers should be able to set this value through code using the peoplequery property peoplequery is ignored if peoplequeries or people are specified additional context cc aycabas,1
+1132,27403372343.0,IssuesEvent,2023-03-01 03:39:07,culturesofknowledge/emlo-project,https://api.github.com/repos/culturesofknowledge/emlo-project,opened,"People search: using the browser back button changes the results, not the form",people search AR test,"If I go back, the search results change to what it was, but not the search form corresponding to the results.
+
+",1.0,"People search: using the browser back button changes the results, not the form - If I go back, the search results change to what it was, but not the search form corresponding to the results.
+
+",1,people search using the browser back button changes the results not the form if i go back the search results change to what it was but not the search form corresponding to the results ,1
+59,3011005842.0,IssuesEvent,2015-07-28 15:48:12,brharp/hjckrrh,https://api.github.com/repos/brharp/hjckrrh,opened,P1 - Basic Page and include all other default pages,feature: events (E) feature: news (N) feature: page (P) feature: people profiles (PP) feature: services (SR) priority: high,"Right side - default to include; Share This, Menus, and News Archives....and every other important widget (ie new archives)
+
+Find Related Services by Keyword in Bottom section (Miranda to try small formatting)
+",1.0,"P1 - Basic Page and include all other default pages - Right side - default to include; Share This, Menus, and News Archives....and every other important widget (ie new archives)
+
+Find Related Services by Keyword in Bottom section (Miranda to try small formatting)
+",1, basic page and include all other default pages right side default to include share this menus and news archives and every other important widget ie new archives find related services by keyword in bottom section miranda to try small formatting ,1
+617,10775650098.0,IssuesEvent,2019-11-03 15:41:08,microsoftgraph/microsoft-graph-toolkit,https://api.github.com/repos/microsoftgraph/microsoft-graph-toolkit,opened,[BUG] people picker shows error while loading results,Area: Components Component: people-picker Needs: Triage :mag: bug,"
+
+
+**Describe the bug**
+When typing in the people picker, the people pickers shows an error ""we didnt' find any matches"" while loading and then shows the results. This is a regression.
+
+**To Reproduce**
+Steps to reproduce the behavior:
+1. Add a people picker on the page
+1. Search for a person
+
+**Expected behavior**
+While the search is loading, the people picker should not show an error message
+
+",1.0,"[BUG] people picker shows error while loading results -
+
+
+**Describe the bug**
+When typing in the people picker, the people pickers shows an error ""we didnt' find any matches"" while loading and then shows the results. This is a regression.
+
+**To Reproduce**
+Steps to reproduce the behavior:
+1. Add a people picker on the page
+1. Search for a person
+
+**Expected behavior**
+While the search is loading, the people picker should not show an error message
+
+",1, people picker shows error while loading results describe the bug when typing in the people picker the people pickers shows an error we didnt find any matches while loading and then shows the results this is a regression to reproduce steps to reproduce the behavior add a people picker on the page search for a person expected behavior while the search is loading the people picker should not show an error message ,1
+191,4106832404.0,IssuesEvent,2016-06-06 10:23:05,auth0/auth0-angular,https://api.github.com/repos/auth0/auth0-angular,closed,signout() race condition,Waiting for people feedback,"Calls to signout() do not reliably clear localStorage.
+
+The method is synchronous but it looks like there is a race condition in the handler this method invokes. This then breaks my standard implementation of route restriction which uses angular stateChange detection and a route restriction flag on ui-router states. Inside the stateChange handler, we check for presence of a token and redirect if necessary, if the signout() call failed to clear the token, a user who has logged out can still access restricted views.
+
+The workaround I'm using now is to clear localStorage on logout using angular $window service, rather than calling signout()",1.0,"signout() race condition - Calls to signout() do not reliably clear localStorage.
+
+The method is synchronous but it looks like there is a race condition in the handler this method invokes. This then breaks my standard implementation of route restriction which uses angular stateChange detection and a route restriction flag on ui-router states. Inside the stateChange handler, we check for presence of a token and redirect if necessary, if the signout() call failed to clear the token, a user who has logged out can still access restricted views.
+
+The workaround I'm using now is to clear localStorage on logout using angular $window service, rather than calling signout()",1,signout race condition calls to signout do not reliably clear localstorage the method is synchronous but it looks like there is a race condition in the handler this method invokes this then breaks my standard implementation of route restriction which uses angular statechange detection and a route restriction flag on ui router states inside the statechange handler we check for presence of a token and redirect if necessary if the signout call failed to clear the token a user who has logged out can still access restricted views the workaround i m using now is to clear localstorage on logout using angular window service rather than calling signout ,1
+682790,23357825713.0,IssuesEvent,2022-08-10 09:00:56,webcompat/web-bugs,https://api.github.com/repos/webcompat/web-bugs,closed,meet.google.com - Bad performance,status-needsinfo browser-firefox priority-critical os-linux engine-gecko,"
+
+
+
+**URL**: https://meet.google.com
+
+**Browser / Version**: Firefox 102.0
+**Operating System**: Linux
+**Tested Another Browser**: Yes Chrome
+
+**Problem type**: Something else
+**Description**: performance is intermittetenly bad
+**Steps to Reproduce**:
+While one a video call with more than a couple people, my entire system will get slow, and the audio from Google Meet gets choppy (both incoming and outgoing).
+
+I've found that my entire CPU is down clocking to less than 1GHz for about 15 seconds out of every two minutes. It comes in bursts, dropping frequency and then recovering, only to do it again a minute or two later.
+
+Chrome does not exhibit this behavior. My laptop has a modern i7 Intel CPU with integrated graphics. Neither the laptop nor CPU are overheating.
+
+
+Browser Configuration
+
+
None
+
+
+
+_From [webcompat.com](https://webcompat.com/) with ❤️_",1.0,"meet.google.com - Bad performance -
+
+
+
+**URL**: https://meet.google.com
+
+**Browser / Version**: Firefox 102.0
+**Operating System**: Linux
+**Tested Another Browser**: Yes Chrome
+
+**Problem type**: Something else
+**Description**: performance is intermittetenly bad
+**Steps to Reproduce**:
+While one a video call with more than a couple people, my entire system will get slow, and the audio from Google Meet gets choppy (both incoming and outgoing).
+
+I've found that my entire CPU is down clocking to less than 1GHz for about 15 seconds out of every two minutes. It comes in bursts, dropping frequency and then recovering, only to do it again a minute or two later.
+
+Chrome does not exhibit this behavior. My laptop has a modern i7 Intel CPU with integrated graphics. Neither the laptop nor CPU are overheating.
+
+
+Browser Configuration
+
+
None
+
+
+
+_From [webcompat.com](https://webcompat.com/) with ❤️_",0,meet google com bad performance url browser version firefox operating system linux tested another browser yes chrome problem type something else description performance is intermittetenly bad steps to reproduce while one a video call with more than a couple people my entire system will get slow and the audio from google meet gets choppy both incoming and outgoing i ve found that my entire cpu is down clocking to less than for about seconds out of every two minutes it comes in bursts dropping frequency and then recovering only to do it again a minute or two later chrome does not exhibit this behavior my laptop has a modern intel cpu with integrated graphics neither the laptop nor cpu are overheating browser configuration none from with ❤️ ,0
+1047,25179552227.0,IssuesEvent,2022-11-11 12:24:28,jongfeel/BookReview,https://api.github.com/repos/jongfeel/BookReview,closed,4부 7장 개에게도 좋은 이름을 붙여 주어라,2022 How to Win Friends & Influence People,"### 7장 개에게도 좋은 이름을 붙여 주어라
+
+오래된 속담이 있다. “개에게 나쁜 이름을 붙여주면 그 개를 매달고 싶어질 것이다.” 반대로 개에게 좋은 이름을 붙여 보라. 그리고 어떤 일이 일어나는지 지켜보라!
+
+부자든 가난한 사람이든, 거지든 도둑이든 거의 모든 사람들은 자신에게 정직한 사람이라는 명성이 주어지면 그 기대에 부응하려 든다.
+
+---
+
+규칙 7: 기꺼이 부응할 만한 평판을 부여하라.
+Give the other person a fine reputation to live up to.",1.0,"4부 7장 개에게도 좋은 이름을 붙여 주어라 - ### 7장 개에게도 좋은 이름을 붙여 주어라
+
+오래된 속담이 있다. “개에게 나쁜 이름을 붙여주면 그 개를 매달고 싶어질 것이다.” 반대로 개에게 좋은 이름을 붙여 보라. 그리고 어떤 일이 일어나는지 지켜보라!
+
+부자든 가난한 사람이든, 거지든 도둑이든 거의 모든 사람들은 자신에게 정직한 사람이라는 명성이 주어지면 그 기대에 부응하려 든다.
+
+---
+
+규칙 7: 기꺼이 부응할 만한 평판을 부여하라.
+Give the other person a fine reputation to live up to.",1, 개에게도 좋은 이름을 붙여 주어라 개에게도 좋은 이름을 붙여 주어라 오래된 속담이 있다 “개에게 나쁜 이름을 붙여주면 그 개를 매달고 싶어질 것이다 ” 반대로 개에게 좋은 이름을 붙여 보라 그리고 어떤 일이 일어나는지 지켜보라 부자든 가난한 사람이든 거지든 도둑이든 거의 모든 사람들은 자신에게 정직한 사람이라는 명성이 주어지면 그 기대에 부응하려 든다 규칙 기꺼이 부응할 만한 평판을 부여하라 give the other person a fine reputation to live up to ,1
+404616,27491587872.0,IssuesEvent,2023-03-04 17:35:36,colinhacks/zod,https://api.github.com/repos/colinhacks/zod,closed,Custom schema example for template literals gives type error,documentation,"In the [Custom schemas](https://github.com/colinhacks/zod#custom-schemas) section of the docs, this example is given for creating a template literal schema:
+```ts
+const px = z.custom<`${number}px`>((val) => /^\d+px$/.test(val));
+px.parse(""100px""); // pass
+px.parse(""100vw""); // fail
+```
+
+As is, this example shows an error on val being passed into test:
+
+> `Argument of type 'unknown' is not assignable to parameter of type 'string'.`
+
+`z.custom()` takes parameter `check?: ((data: unknown) => any) | undefined`, while `test.()` takes parameter `string: string`.
+
+I think ideally, there should be a way to annotate or infer the type of `data`, but that section is pretty empty.
+I found two ways around the type error, one of which would be good to update the docs with:
+```ts
+const pxWithAs = z.custom<`${number}px`>((val) => /^\d+px$/.test(val as string));
+const pxWithTemplate = z.custom<`${number}px`>((val) => /^\d+px$/.test(`${val}`));
+```
+
+[Playground link showing the error and ways around](https://www.typescriptlang.org/play?#code/JYWwDg9gTgLgBAbzgLzgXzgMyhEcDkyEAJvgNwBQFA9NXAKIAeAhuADYCmWOeAFjDDABnAFy0A5sBi8ArgCMAdAGNc1FW2AA7XsyUBrIdSLFqAYiUyhMXAFohS3hxDMhFFZqtwwjOAF4UypbWIAA8AAYAJAiaMiByHFBo3mEAfAAUaQBuzGwAlH4pcNQAegA6xADU3hHUCjAcVlk5ubmU3gpgzFBCHGkARACMAAxD3n2tRXSdQq7tnd29gyOZAO7jZJNYzMBsVLRwACorEHArzACeQnBdEDKaxHDSXAk4UG4QHvDeAOpSvACCV38yECVlw4SiMTiCSSjFSGWyeQKRTKlWqtXqjUR1yuVigWnELUo7k8Pz+BycYDYzHqfgCFjBoUi0Vi8USyXSTSRvkKJXKVUYNTqDRgaWZiLQYSJQA)
+
+I don't like either of these that much, and I'll probably be using `z.templateLiteral()` after #1786 is merged, but the docs example should be updated with whatever method is recommended to avoid the error.",1.0,"Custom schema example for template literals gives type error - In the [Custom schemas](https://github.com/colinhacks/zod#custom-schemas) section of the docs, this example is given for creating a template literal schema:
+```ts
+const px = z.custom<`${number}px`>((val) => /^\d+px$/.test(val));
+px.parse(""100px""); // pass
+px.parse(""100vw""); // fail
+```
+
+As is, this example shows an error on val being passed into test:
+
+> `Argument of type 'unknown' is not assignable to parameter of type 'string'.`
+
+`z.custom()` takes parameter `check?: ((data: unknown) => any) | undefined`, while `test.()` takes parameter `string: string`.
+
+I think ideally, there should be a way to annotate or infer the type of `data`, but that section is pretty empty.
+I found two ways around the type error, one of which would be good to update the docs with:
+```ts
+const pxWithAs = z.custom<`${number}px`>((val) => /^\d+px$/.test(val as string));
+const pxWithTemplate = z.custom<`${number}px`>((val) => /^\d+px$/.test(`${val}`));
+```
+
+[Playground link showing the error and ways around](https://www.typescriptlang.org/play?#code/JYWwDg9gTgLgBAbzgLzgXzgMyhEcDkyEAJvgNwBQFA9NXAKIAeAhuADYCmWOeAFjDDABnAFy0A5sBi8ArgCMAdAGNc1FW2AA7XsyUBrIdSLFqAYiUyhMXAFohS3hxDMhFFZqtwwjOAF4UypbWIAA8AAYAJAiaMiByHFBo3mEAfAAUaQBuzGwAlH4pcNQAegA6xADU3hHUCjAcVlk5ubmU3gpgzFBCHGkARACMAAxD3n2tRXSdQq7tnd29gyOZAO7jZJNYzMBsVLRwACorEHArzACeQnBdEDKaxHDSXAk4UG4QHvDeAOpSvACCV38yECVlw4SiMTiCSSjFSGWyeQKRTKlWqtXqjUR1yuVigWnELUo7k8Pz+BycYDYzHqfgCFjBoUi0Vi8USyXSTSRvkKJXKVUYNTqDRgaWZiLQYSJQA)
+
+I don't like either of these that much, and I'll probably be using `z.templateLiteral()` after #1786 is merged, but the docs example should be updated with whatever method is recommended to avoid the error.",0,custom schema example for template literals gives type error in the section of the docs this example is given for creating a template literal schema ts const px z custom val d px test val px parse pass px parse fail as is this example shows an error on val being passed into test argument of type unknown is not assignable to parameter of type string z custom takes parameter check data unknown any undefined while test takes parameter string string i think ideally there should be a way to annotate or infer the type of data but that section is pretty empty i found two ways around the type error one of which would be good to update the docs with ts const pxwithas z custom val d px test val as string const pxwithtemplate z custom val d px test val i don t like either of these that much and i ll probably be using z templateliteral after is merged but the docs example should be updated with whatever method is recommended to avoid the error ,0
+38026,8402321175.0,IssuesEvent,2018-10-11 06:06:16,dotnet/coreclr,https://api.github.com/repos/dotnet/coreclr,reopened,RyuJIT: By-ref assignment with null leads to runtime crash,area-CodeGen,"The following example crashes the runtime when compiled and run in either debug or release:
+```csharp
+class C0
+{
+}
+
+struct S0
+{
+ public C0 F0;
+ public ulong F4;
+}
+
+class C1
+{
+ public S0 F3;
+}
+
+struct S1
+{
+ public S0 F3;
+}
+
+public class Program
+{
+ static S1 s_38;
+ static C1 s_43;
+
+ public static void Main()
+ {
+ s_38.F3 = s_43.F3;
+ }
+}
+```
+
+In debug (and probably checked) builds, this assertion triggers:
+```
+Assert failure(PID 20848 [0x00005170], Thread: 17420 [0x440c]): Consistency check failed: AV in clr at this callstack:
+------
+CORECLR! JIT_ByRefWriteBarrier + 0x0 (0x00007ffc`5e207d70)
+! + 0x0 (0x00007ffb`fef22611)
+! + 0x0 (0x00007ffb`fee045e0)
+! + 0x0 (0x000000fa`00000005)
+! + 0x0 (0x000000fa`2077d580)
+-----
+.AV on tid=0x440c (17420), cxr=000000FA2077CE00, exr=000000FA2077D2F0
+FAILED: false
+
+CORECLR! CHECK::Trigger + 0x275 (0x00007ffc`5db0cea5)
+CORECLR! CLRVectoredExceptionHandlerPhase3 + 0x332 (0x00007ffc`5dc1bd62)
+CORECLR! CLRVectoredExceptionHandlerPhase2 + 0x8C (0x00007ffc`5dc1b70c)
+CORECLR! CLRVectoredExceptionHandler + 0x275 (0x00007ffc`5dc1b665)
+CORECLR! CLRVectoredExceptionHandlerShim + 0x18D (0x00007ffc`5dc1c04d)
+NTDLL! RtlInitializeCriticalSection + 0x1D8 (0x00007ffc`c5bd5678)
+NTDLL! RtlWalkFrameChain + 0x109A (0x00007ffc`c5b7692a)
+NTDLL! KiUserExceptionDispatcher + 0x2E (0x00007ffc`c5c0dc1e)
+CORECLR! JIT_ByRefWriteBarrier + 0x0 (0x00007ffc`5e207d70)
+! + 0x0 (0x00007ffb`fef22611)
+ File: z:\programming\dotnet\coreclr\src\vm\excep.cpp Line: 7831
+ Image: Z:\Programming\dotnet\coreclr\bin\Product\Windows_NT.x64.Debug\CoreRun.exe
+```
+
+The code generated is:
+```asm
+; Assembly listing for method Program:Main()
+; Emitting BLENDED_CODE for X64 CPU with AVX
+; optimized code
+; rsp based frame
+; partially interruptible
+; Final local variable assignments
+;
+; V00 OutArgs [V00 ] ( 1, 1 ) lclBlk (32) [rsp+0x00]
+; V01 cse0 [V01,T00] ( 5, 5 ) long -> [rsp+0x20]
+;
+; Lcl frame size = 40
+
+G_M5092_IG01:
+ 57 push rdi
+ 56 push rsi
+ 4883EC28 sub rsp, 40
+
+G_M5092_IG02:
+ 48B9E045E3FEFB7F0000 mov rcx, 0x7FFBFEE345E0
+ BA05000000 mov edx, 5
+ E8461E455F call CORINFO_HELP_GETSHARED_NONGCSTATIC_BASE
+ 48B87829001091010000 mov rax, 0x19110002978
+ 488B00 mov rax, gword ptr [rax]
+ 488D7008 lea rsi, bword ptr [rax+8]
+ 48B87029001091010000 mov rax, 0x19110002970
+ 488B00 mov rax, gword ptr [rax]
+ 488D7808 lea rdi, bword ptr [rax+8]
+ E85F572B5F call CORINFO_HELP_ASSIGN_BYREF
+ 48A5 movsq
+
+G_M5092_IG03:
+ 4883C428 add rsp, 40
+ 5E pop rsi
+ 5F pop rdi
+ C3 ret
+
+; Total bytes of code 74, prolog size 6 for method Program:Main()
+;
+```
+
+Question: How will this normally be handled by CoreCLR? Should the runtime turn the AV in `CORINFO_HELP_ASSIGN_BYREF` into a managed NRE, or is the caller supposed to null-check/deref the reference before calling `CORINFO_HELP_ASSIGN_BYREF`?",1.0,"RyuJIT: By-ref assignment with null leads to runtime crash - The following example crashes the runtime when compiled and run in either debug or release:
+```csharp
+class C0
+{
+}
+
+struct S0
+{
+ public C0 F0;
+ public ulong F4;
+}
+
+class C1
+{
+ public S0 F3;
+}
+
+struct S1
+{
+ public S0 F3;
+}
+
+public class Program
+{
+ static S1 s_38;
+ static C1 s_43;
+
+ public static void Main()
+ {
+ s_38.F3 = s_43.F3;
+ }
+}
+```
+
+In debug (and probably checked) builds, this assertion triggers:
+```
+Assert failure(PID 20848 [0x00005170], Thread: 17420 [0x440c]): Consistency check failed: AV in clr at this callstack:
+------
+CORECLR! JIT_ByRefWriteBarrier + 0x0 (0x00007ffc`5e207d70)
+! + 0x0 (0x00007ffb`fef22611)
+! + 0x0 (0x00007ffb`fee045e0)
+! + 0x0 (0x000000fa`00000005)
+! + 0x0 (0x000000fa`2077d580)
+-----
+.AV on tid=0x440c (17420), cxr=000000FA2077CE00, exr=000000FA2077D2F0
+FAILED: false
+
+CORECLR! CHECK::Trigger + 0x275 (0x00007ffc`5db0cea5)
+CORECLR! CLRVectoredExceptionHandlerPhase3 + 0x332 (0x00007ffc`5dc1bd62)
+CORECLR! CLRVectoredExceptionHandlerPhase2 + 0x8C (0x00007ffc`5dc1b70c)
+CORECLR! CLRVectoredExceptionHandler + 0x275 (0x00007ffc`5dc1b665)
+CORECLR! CLRVectoredExceptionHandlerShim + 0x18D (0x00007ffc`5dc1c04d)
+NTDLL! RtlInitializeCriticalSection + 0x1D8 (0x00007ffc`c5bd5678)
+NTDLL! RtlWalkFrameChain + 0x109A (0x00007ffc`c5b7692a)
+NTDLL! KiUserExceptionDispatcher + 0x2E (0x00007ffc`c5c0dc1e)
+CORECLR! JIT_ByRefWriteBarrier + 0x0 (0x00007ffc`5e207d70)
+! + 0x0 (0x00007ffb`fef22611)
+ File: z:\programming\dotnet\coreclr\src\vm\excep.cpp Line: 7831
+ Image: Z:\Programming\dotnet\coreclr\bin\Product\Windows_NT.x64.Debug\CoreRun.exe
+```
+
+The code generated is:
+```asm
+; Assembly listing for method Program:Main()
+; Emitting BLENDED_CODE for X64 CPU with AVX
+; optimized code
+; rsp based frame
+; partially interruptible
+; Final local variable assignments
+;
+; V00 OutArgs [V00 ] ( 1, 1 ) lclBlk (32) [rsp+0x00]
+; V01 cse0 [V01,T00] ( 5, 5 ) long -> [rsp+0x20]
+;
+; Lcl frame size = 40
+
+G_M5092_IG01:
+ 57 push rdi
+ 56 push rsi
+ 4883EC28 sub rsp, 40
+
+G_M5092_IG02:
+ 48B9E045E3FEFB7F0000 mov rcx, 0x7FFBFEE345E0
+ BA05000000 mov edx, 5
+ E8461E455F call CORINFO_HELP_GETSHARED_NONGCSTATIC_BASE
+ 48B87829001091010000 mov rax, 0x19110002978
+ 488B00 mov rax, gword ptr [rax]
+ 488D7008 lea rsi, bword ptr [rax+8]
+ 48B87029001091010000 mov rax, 0x19110002970
+ 488B00 mov rax, gword ptr [rax]
+ 488D7808 lea rdi, bword ptr [rax+8]
+ E85F572B5F call CORINFO_HELP_ASSIGN_BYREF
+ 48A5 movsq
+
+G_M5092_IG03:
+ 4883C428 add rsp, 40
+ 5E pop rsi
+ 5F pop rdi
+ C3 ret
+
+; Total bytes of code 74, prolog size 6 for method Program:Main()
+;
+```
+
+Question: How will this normally be handled by CoreCLR? Should the runtime turn the AV in `CORINFO_HELP_ASSIGN_BYREF` into a managed NRE, or is the caller supposed to null-check/deref the reference before calling `CORINFO_HELP_ASSIGN_BYREF`?",0,ryujit by ref assignment with null leads to runtime crash the following example crashes the runtime when compiled and run in either debug or release csharp class struct public public ulong class public struct public public class program static s static s public static void main s s in debug and probably checked builds this assertion triggers assert failure pid thread consistency check failed av in clr at this callstack coreclr jit byrefwritebarrier av on tid cxr exr failed false coreclr check trigger coreclr coreclr coreclr clrvectoredexceptionhandler coreclr clrvectoredexceptionhandlershim ntdll rtlinitializecriticalsection ntdll rtlwalkframechain ntdll kiuserexceptiondispatcher coreclr jit byrefwritebarrier file z programming dotnet coreclr src vm excep cpp line image z programming dotnet coreclr bin product windows nt debug corerun exe the code generated is asm assembly listing for method program main emitting blended code for cpu with avx optimized code rsp based frame partially interruptible final local variable assignments outargs lclblk long lcl frame size g push rdi push rsi sub rsp g mov rcx mov edx call corinfo help getshared nongcstatic base mov rax mov rax gword ptr lea rsi bword ptr mov rax mov rax gword ptr lea rdi bword ptr call corinfo help assign byref movsq g add rsp pop rsi pop rdi ret total bytes of code prolog size for method program main question how will this normally be handled by coreclr should the runtime turn the av in corinfo help assign byref into a managed nre or is the caller supposed to null check deref the reference before calling corinfo help assign byref ,0
+701523,24100112931.0,IssuesEvent,2022-09-19 23:11:49,googleapis/nodejs-language,https://api.github.com/repos/googleapis/nodejs-language,closed,set endpoint for language api call: should analyze sentiment in text at a specific api endpoint failed,priority: p1 type: bug api: language flakybot: issue,"Note: #616 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
+
+----
+
+commit: b5a177d9c6a82ef012ab18c463c796444a34fe1b
+buildURL: [Build Status](https://source.cloud.google.com/results/invocations/3c6a85b2-e3d6-40b4-a140-7c7c27524e79), [Sponge](http://sponge2/3c6a85b2-e3d6-40b4-a140-7c7c27524e79)
+status: failed
+Test output
msg.replace is not a function
+TypeError: msg.replace is not a function
+ at Object.getMessage (node_modules/chai/lib/chai/utils/getMessage.js:45:6)
+ at Proxy.Assertion.assert (node_modules/chai/lib/chai/assertion.js:139:18)
+ at chai.assert (node_modules/chai/lib/chai/interface/assert.js:36:10)
+ at Context. (test/setEndpoint.test.js:26:5)
+ at processImmediate (internal/timers.js:461:21)
",1.0,"set endpoint for language api call: should analyze sentiment in text at a specific api endpoint failed - Note: #616 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
+
+----
+
+commit: b5a177d9c6a82ef012ab18c463c796444a34fe1b
+buildURL: [Build Status](https://source.cloud.google.com/results/invocations/3c6a85b2-e3d6-40b4-a140-7c7c27524e79), [Sponge](http://sponge2/3c6a85b2-e3d6-40b4-a140-7c7c27524e79)
+status: failed
+Test output
msg.replace is not a function
+TypeError: msg.replace is not a function
+ at Object.getMessage (node_modules/chai/lib/chai/utils/getMessage.js:45:6)
+ at Proxy.Assertion.assert (node_modules/chai/lib/chai/assertion.js:139:18)
+ at chai.assert (node_modules/chai/lib/chai/interface/assert.js:36:10)
+ at Context. (test/setEndpoint.test.js:26:5)
+ at processImmediate (internal/timers.js:461:21)
",0,set endpoint for language api call should analyze sentiment in text at a specific api endpoint failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output msg replace is not a function typeerror msg replace is not a function at object getmessage node modules chai lib chai utils getmessage js at proxy assertion assert node modules chai lib chai assertion js at chai assert node modules chai lib chai interface assert js at context test setendpoint test js at processimmediate internal timers js ,0
+601914,18442400635.0,IssuesEvent,2021-10-14 19:49:25,carbon-design-system/carbon-for-ibm-dotcom-website,https://api.github.com/repos/carbon-design-system/carbon-for-ibm-dotcom-website,closed,Website: Upgrade `gatsby-carbon-theme` dependency to v2.x,Feature request priority: medium dev,"### The problem
+
+Investigate upgrading to `gatsby-carbon-theme` v2.x. In addition to containing component enhancements which we are looking to make use of (#1147) it comes with a variety of quality of life features like improved build and reload times
+
+
+
+### The proposed solution
+
+migration guide https://gatsby.carbondesignsystem.com/guides/v2-migration/
+
+### Examples
+
+_No response_
+
+### Code of conduct
+
+- [X] I agree to follow this project's [Code of conduct](https://github.com/carbon-design-system/carbon-for-ibm-dotcom-website/blob/master/.github/CODE_OF_CONDUCT.md).
+- [X] I checked the [current issues](https://github.com/carbon-design-system/carbon-for-ibm-dotcom-website/issues) for duplicate issues.",1.0,"Website: Upgrade `gatsby-carbon-theme` dependency to v2.x - ### The problem
+
+Investigate upgrading to `gatsby-carbon-theme` v2.x. In addition to containing component enhancements which we are looking to make use of (#1147) it comes with a variety of quality of life features like improved build and reload times
+
+
+
+### The proposed solution
+
+migration guide https://gatsby.carbondesignsystem.com/guides/v2-migration/
+
+### Examples
+
+_No response_
+
+### Code of conduct
+
+- [X] I agree to follow this project's [Code of conduct](https://github.com/carbon-design-system/carbon-for-ibm-dotcom-website/blob/master/.github/CODE_OF_CONDUCT.md).
+- [X] I checked the [current issues](https://github.com/carbon-design-system/carbon-for-ibm-dotcom-website/issues) for duplicate issues.",0,website upgrade gatsby carbon theme dependency to x the problem investigate upgrading to gatsby carbon theme x in addition to containing component enhancements which we are looking to make use of it comes with a variety of quality of life features like improved build and reload times the proposed solution migration guide examples no response code of conduct i agree to follow this project s i checked the for duplicate issues ,0
+144366,5539221629.0,IssuesEvent,2017-03-22 05:34:17,alexrp/alkahest,https://api.github.com/repos/alexrp/alkahest,opened,Use lists instead of arrays in packet structures,Area: Core Library Priority: Normal Type: Enhancement,This makes it easier to modify existing arrays.,1.0,Use lists instead of arrays in packet structures - This makes it easier to modify existing arrays.,0,use lists instead of arrays in packet structures this makes it easier to modify existing arrays ,0
+14,2649586756.0,IssuesEvent,2015-03-15 02:21:05,fskreuz/Tree-Trail,https://api.github.com/repos/fskreuz/Tree-Trail,closed,Display Latitude and Longitude in Badge,QA people! Test these!,"As a user, I should be able to see the coordinates as to where to plant the trees.",1.0,"Display Latitude and Longitude in Badge - As a user, I should be able to see the coordinates as to where to plant the trees.",1,display latitude and longitude in badge as a user i should be able to see the coordinates as to where to plant the trees ,1
+659,11714100169.0,IssuesEvent,2020-03-09 11:38:16,kazuwjnlab/cvpaper,https://api.github.com/repos/kazuwjnlab/cvpaper,opened,[cvpaper] CVPR2019 #824 Learning the Depths of Moving People by Watching Frozen People,deoth estimation motion parallax moving people multi-view stereo,"## \#824 [Learning the Depths of Moving People by Watching Frozen People](http://openaccess.thecvf.com/content_CVPR_2019/papers/Li_Learning_the_Depths_of_Moving_People_by_Watching_Frozen_People_CVPR_2019_paper.pdf)
+Zhengqi Li, Tali Dekel, Forrester Cole, Richard Tucker, Noah Snavely, Ce Liu, William T. Freeman
+### どんな論文か?
+カメラと人物両方が動いている映像からの深度を学習・推定するための効果的なデータセットと予測手法の提案。人のような非剛体は静止することがないため正解深度を算出することは困難であったが、マネキンチャレンジ(人物が静止する)の動画をインターネットより収集すること(!)で解決。
+### 新規性
+動く人物を含む深度推定のためにマネキンチャレンジ動画にSfMとMVSをかけることによってその正確な深度及びカメラ姿勢を取得したこと。推論時、物体が動いている場合にも背景の運動視差を最大限利用しつつ推定を行えるネットワークの提案。
+### 結果
+新しく構築したデータセットにより単眼でも正確な推定を実現しつつ、2フレームからの運動視差を考慮することで人物・非人物領域の両方において大幅に制度を改善。従来のデータセットには含まれない状況を含むインターネット上の様々の動画において従来手法が殆ど失敗する中高品質の推定に成功。
+### その他(なぜ通ったか?等)
+アイデアに注目が行きがちだが正しい深度およびカメラ姿勢が取得できる映像区間を吟味するためのデータセット構築に相当の注意と手間をかけている。既存のデータセットでは実現しえない設定での深度推定を独自の工夫で達成しており見習いたい。
+
+deoth estimation, motion parallax, moving people, multi-view stereo, ",1.0,"[cvpaper] CVPR2019 #824 Learning the Depths of Moving People by Watching Frozen People - ## \#824 [Learning the Depths of Moving People by Watching Frozen People](http://openaccess.thecvf.com/content_CVPR_2019/papers/Li_Learning_the_Depths_of_Moving_People_by_Watching_Frozen_People_CVPR_2019_paper.pdf)
+Zhengqi Li, Tali Dekel, Forrester Cole, Richard Tucker, Noah Snavely, Ce Liu, William T. Freeman
+### どんな論文か?
+カメラと人物両方が動いている映像からの深度を学習・推定するための効果的なデータセットと予測手法の提案。人のような非剛体は静止することがないため正解深度を算出することは困難であったが、マネキンチャレンジ(人物が静止する)の動画をインターネットより収集すること(!)で解決。
+### 新規性
+動く人物を含む深度推定のためにマネキンチャレンジ動画にSfMとMVSをかけることによってその正確な深度及びカメラ姿勢を取得したこと。推論時、物体が動いている場合にも背景の運動視差を最大限利用しつつ推定を行えるネットワークの提案。
+### 結果
+新しく構築したデータセットにより単眼でも正確な推定を実現しつつ、2フレームからの運動視差を考慮することで人物・非人物領域の両方において大幅に制度を改善。従来のデータセットには含まれない状況を含むインターネット上の様々の動画において従来手法が殆ど失敗する中高品質の推定に成功。
+### その他(なぜ通ったか?等)
+アイデアに注目が行きがちだが正しい深度およびカメラ姿勢が取得できる映像区間を吟味するためのデータセット構築に相当の注意と手間をかけている。既存のデータセットでは実現しえない設定での深度推定を独自の工夫で達成しており見習いたい。
+
+deoth estimation, motion parallax, moving people, multi-view stereo, ",1, learning the depths of moving people by watching frozen people zhengqi li tali dekel forrester cole richard tucker noah snavely ce liu william t freeman どんな論文か? カメラと人物両方が動いている映像からの深度を学習・推定するための効果的なデータセットと予測手法の提案。人のような非剛体は静止することがないため正解深度を算出することは困難であったが、マネキンチャレンジ(人物が静止する)の動画をインターネットより収集すること(!)で解決。 新規性 動く人物を含む深度推定のためにマネキンチャレンジ動画にsfmとmvsをかけることによってその正確な深度及びカメラ姿勢を取得したこと。推論時、物体が動いている場合にも背景の運動視差を最大限利用しつつ推定を行えるネットワークの提案。 結果 新しく構築したデータセットにより単眼でも正確な推定を実現しつつ、 ・非人物領域の両方において大幅に制度を改善。従来のデータセットには含まれない状況を含むインターネット上の様々の動画において従来手法が殆ど失敗する中高品質の推定に成功。 その他(なぜ通ったか?等) アイデアに注目が行きがちだが正しい深度およびカメラ姿勢が取得できる映像区間を吟味するためのデータセット構築に相当の注意と手間をかけている。既存のデータセットでは実現しえない設定での深度推定を独自の工夫で達成しており見習いたい。 deoth estimation motion parallax moving people multi view stereo ,1
+486,8443355131.0,IssuesEvent,2018-10-18 15:23:32,openstates/openstates,https://api.github.com/repos/openstates/openstates,closed,"MT: Switch committee scraper to use structured HTML, rather than PDF scrape",people priority.low,"Montana now serves its committees in a clean way, on http://leg.mt.gov/css/Committees/default.asp",1.0,"MT: Switch committee scraper to use structured HTML, rather than PDF scrape - Montana now serves its committees in a clean way, on http://leg.mt.gov/css/Committees/default.asp",1,mt switch committee scraper to use structured html rather than pdf scrape montana now serves its committees in a clean way on ,1
+1104,26908864660.0,IssuesEvent,2023-02-06 21:34:00,Reviewable/Reviewable,https://api.github.com/repos/Reviewable/Reviewable,closed,"""$USER is blocking this review"" should be inspectable",enhancement UI/UX feature:people,"The workflow requested is for when user Alice has commented on a review for Bob (or anything else that might block the review) and then gone on vacation, that Bob or Carol can view the review, inspect why Alice is blocking the review, then unblock the review by dismissing comments, etc.
+
+Currently, they have to look over the whole PR looking for unresolved code comments, file comments, or PR comments and dismiss them.
+
+Cherry on-top: The ""$USER"" in the message should be a link to ""why they block the review"" instead of to github, since users intuitively click that.",1.0,"""$USER is blocking this review"" should be inspectable - The workflow requested is for when user Alice has commented on a review for Bob (or anything else that might block the review) and then gone on vacation, that Bob or Carol can view the review, inspect why Alice is blocking the review, then unblock the review by dismissing comments, etc.
+
+Currently, they have to look over the whole PR looking for unresolved code comments, file comments, or PR comments and dismiss them.
+
+Cherry on-top: The ""$USER"" in the message should be a link to ""why they block the review"" instead of to github, since users intuitively click that.",1, user is blocking this review should be inspectable the workflow requested is for when user alice has commented on a review for bob or anything else that might block the review and then gone on vacation that bob or carol can view the review inspect why alice is blocking the review then unblock the review by dismissing comments etc currently they have to look over the whole pr looking for unresolved code comments file comments or pr comments and dismiss them cherry on top the user in the message should be a link to why they block the review instead of to github since users intuitively click that ,1
+758006,26539988404.0,IssuesEvent,2023-01-19 18:26:07,GoogleCloudPlatform/professional-services-data-validator,https://api.github.com/repos/GoogleCloudPlatform/professional-services-data-validator,closed,Issue while performing schema & data validations for teradata to bq with google-pso-data-validator==2.7.0,type: bug priority: p0,"Hi Team,
+
+We were running a workflow for DVT to perform schema, row and column validations for
+
+Source - Teradata
+Target - Bigquery
+
+It was working fine till end of last month, our requirements.txt for running DVT on Cloud Run and Kubernetes Pod Operator does not mention specific version and now it has started to install google-pso-data-validator==2.7.0 which was released on Jan 6th I believe.
+
+Since this update, we are receiving the below error while executing dvt against source DB which was working fine earlier.
+
+**teradatasql.OperationalError: [Version 17.20.0.10] [Session 0] [Teradata Database] [Error 8019] The parcel stream is invalid.**
+
+On reverting the cli to version 2.6.0, it starts working fine
+
+**pip install google-pso-data-validator==2.6.0**
+
+Full log file - [error.txt](https://github.com/GoogleCloudPlatform/professional-services-data-validator/files/10457674/error.txt)
+
+ data-validation connections add --connection-name BigQuery_CONN BigQuery --project-id pso-e2e-sql
+
+data-validation connections add --connection-name Teradata_CONN Teradata --host 34.66.209.133 --port 1025 --user-name test --password test
+
+data-validation validate schema -sc Teradata_CONN -tc BigQuery_CONN -tbls test.party=pso-e2e-sql.e2e_data.party
+data-validation validate schema -sc Teradata_CONN -tc BigQuery_CONN -tbls test.party=pso-e2e-sql.e2e_data.party
+Traceback (most recent call last):
+ File ""/home/user/atc/professional-services-data-validator/venv/lib/python3.10/site-packages/pandas/io/sql.py"", line 2056, in execute
+ cur.execute(*args, **kwargs)
+ File ""/home/user/atc/professional-services-data-validator/venv/lib/python3.10/site-packages/teradatasql/__init__.py"", line 686, in execute
+ self.executemany (sOperation, None, ignoreErrors)
+ File ""/home/user/atc/professional-services-data-validator/venv/lib/python3.10/site-packages/teradatasql/__init__.py"", line 933, in executemany
+ raise OperationalError (sErr)
+**teradatasql.OperationalError: [Version 17.20.0.10] [Session 0] [Teradata Database] [Error 8019] The parcel stream is invalid.**
+ at gosqldriver/teradatasql.formatError ErrorUtil.go:88
+ at gosqldriver/teradatasql.(*teradataConnection).formatDatabaseError ErrorUtil.go:216
+ at gosqldriver/teradatasql.(*teradataConnection).makeChainedDatabaseError ErrorUtil.go:232
+ at gosqldriver/teradatasql.(*teradataConnection).processErrorParcel TeradataConnection.go:812
+ at gosqldriver/teradatasql.(*TeradataRows).processResponseBundle TeradataRows.go:2221
+ at gosqldriver/teradatasql.(*TeradataRows).executeSQLRequest TeradataRows.go:809
+ at gosqldriver/teradatasql.newTeradataRows TeradataRows.go:668
+ at gosqldriver/teradatasql.(*teradataStatement).QueryContext TeradataStatement.go:122
+ at gosqldriver/teradatasql.(*teradataConnection).QueryContext TeradataConnection.go:1316
+ at database/sql.ctxDriverQuery ctxutil.go:48
+ at database/sql.(*DB).queryDC.func1 sql.go:1759
+ at database/sql.withLock sql.go:3437
+ at database/sql.(*DB).queryDC sql.go:1754
+ at database/sql.(*Conn).QueryContext sql.go:2013
+ at main.goCreateRows goside.go:666
+ at _cgoexp_092f630b41aa_goCreateRows _cgo_gotypes.go:340
+ at runtime.cgocallbackg1 cgocall.go:314
+ at runtime.cgocallbackg cgocall.go:233
+ at runtime.cgocallback asm_amd64.s:971
+ at runtime.goexit asm_amd64.s:1571
+
+
+
+",1.0,"Issue while performing schema & data validations for teradata to bq with google-pso-data-validator==2.7.0 - Hi Team,
+
+We were running a workflow for DVT to perform schema, row and column validations for
+
+Source - Teradata
+Target - Bigquery
+
+It was working fine till end of last month, our requirements.txt for running DVT on Cloud Run and Kubernetes Pod Operator does not mention specific version and now it has started to install google-pso-data-validator==2.7.0 which was released on Jan 6th I believe.
+
+Since this update, we are receiving the below error while executing dvt against source DB which was working fine earlier.
+
+**teradatasql.OperationalError: [Version 17.20.0.10] [Session 0] [Teradata Database] [Error 8019] The parcel stream is invalid.**
+
+On reverting the cli to version 2.6.0, it starts working fine
+
+**pip install google-pso-data-validator==2.6.0**
+
+Full log file - [error.txt](https://github.com/GoogleCloudPlatform/professional-services-data-validator/files/10457674/error.txt)
+
+ data-validation connections add --connection-name BigQuery_CONN BigQuery --project-id pso-e2e-sql
+
+data-validation connections add --connection-name Teradata_CONN Teradata --host 34.66.209.133 --port 1025 --user-name test --password test
+
+data-validation validate schema -sc Teradata_CONN -tc BigQuery_CONN -tbls test.party=pso-e2e-sql.e2e_data.party
+data-validation validate schema -sc Teradata_CONN -tc BigQuery_CONN -tbls test.party=pso-e2e-sql.e2e_data.party
+Traceback (most recent call last):
+ File ""/home/user/atc/professional-services-data-validator/venv/lib/python3.10/site-packages/pandas/io/sql.py"", line 2056, in execute
+ cur.execute(*args, **kwargs)
+ File ""/home/user/atc/professional-services-data-validator/venv/lib/python3.10/site-packages/teradatasql/__init__.py"", line 686, in execute
+ self.executemany (sOperation, None, ignoreErrors)
+ File ""/home/user/atc/professional-services-data-validator/venv/lib/python3.10/site-packages/teradatasql/__init__.py"", line 933, in executemany
+ raise OperationalError (sErr)
+**teradatasql.OperationalError: [Version 17.20.0.10] [Session 0] [Teradata Database] [Error 8019] The parcel stream is invalid.**
+ at gosqldriver/teradatasql.formatError ErrorUtil.go:88
+ at gosqldriver/teradatasql.(*teradataConnection).formatDatabaseError ErrorUtil.go:216
+ at gosqldriver/teradatasql.(*teradataConnection).makeChainedDatabaseError ErrorUtil.go:232
+ at gosqldriver/teradatasql.(*teradataConnection).processErrorParcel TeradataConnection.go:812
+ at gosqldriver/teradatasql.(*TeradataRows).processResponseBundle TeradataRows.go:2221
+ at gosqldriver/teradatasql.(*TeradataRows).executeSQLRequest TeradataRows.go:809
+ at gosqldriver/teradatasql.newTeradataRows TeradataRows.go:668
+ at gosqldriver/teradatasql.(*teradataStatement).QueryContext TeradataStatement.go:122
+ at gosqldriver/teradatasql.(*teradataConnection).QueryContext TeradataConnection.go:1316
+ at database/sql.ctxDriverQuery ctxutil.go:48
+ at database/sql.(*DB).queryDC.func1 sql.go:1759
+ at database/sql.withLock sql.go:3437
+ at database/sql.(*DB).queryDC sql.go:1754
+ at database/sql.(*Conn).QueryContext sql.go:2013
+ at main.goCreateRows goside.go:666
+ at _cgoexp_092f630b41aa_goCreateRows _cgo_gotypes.go:340
+ at runtime.cgocallbackg1 cgocall.go:314
+ at runtime.cgocallbackg cgocall.go:233
+ at runtime.cgocallback asm_amd64.s:971
+ at runtime.goexit asm_amd64.s:1571
+
+
+
+",0,issue while performing schema data validations for teradata to bq with google pso data validator hi team we were running a workflow for dvt to perform schema row and column validations for source teradata target bigquery it was working fine till end of last month our requirements txt for running dvt on cloud run and kubernetes pod operator does not mention specific version and now it has started to install google pso data validator which was released on jan i believe since this update we are receiving the below error while executing dvt against source db which was working fine earlier teradatasql operationalerror the parcel stream is invalid on reverting the cli to version it starts working fine pip install google pso data validator full log file data validation connections add connection name bigquery conn bigquery project id pso sql data validation connections add connection name teradata conn teradata host port user name test password test data validation validate schema sc teradata conn tc bigquery conn tbls test party pso sql data party data validation validate schema sc teradata conn tc bigquery conn tbls test party pso sql data party traceback most recent call last file home user atc professional services data validator venv lib site packages pandas io sql py line in execute cur execute args kwargs file home user atc professional services data validator venv lib site packages teradatasql init py line in execute self executemany soperation none ignoreerrors file home user atc professional services data validator venv lib site packages teradatasql init py line in executemany raise operationalerror serr teradatasql operationalerror the parcel stream is invalid at gosqldriver teradatasql formaterror errorutil go at gosqldriver teradatasql teradataconnection formatdatabaseerror errorutil go at gosqldriver teradatasql teradataconnection makechaineddatabaseerror errorutil go at gosqldriver teradatasql teradataconnection processerrorparcel teradataconnection go at gosqldriver teradatasql teradatarows processresponsebundle teradatarows go at gosqldriver teradatasql teradatarows executesqlrequest teradatarows go at gosqldriver teradatasql newteradatarows teradatarows go at gosqldriver teradatasql teradatastatement querycontext teradatastatement go at gosqldriver teradatasql teradataconnection querycontext teradataconnection go at database sql ctxdriverquery ctxutil go at database sql db querydc sql go at database sql withlock sql go at database sql db querydc sql go at database sql conn querycontext sql go at main gocreaterows goside go at cgoexp gocreaterows cgo gotypes go at runtime cgocall go at runtime cgocallbackg cgocall go at runtime cgocallback asm s at runtime goexit asm s ,0
+222,4586798692.0,IssuesEvent,2016-09-20 00:39:40,elmsln/elmsln,https://api.github.com/repos/elmsln/elmsln,opened,Add banner picture to user profile fields list,people / cpr style guide / theme,"it's becoming common practice with social tools to have a headshot type image (user profile picture) as well as a landscape / inspiration / larger image. Google+ FB and twitter all have them and it would be cool to have support for in our people distribution as well. Especially as we move towards implementation of side nav it would look REALLY nice for the user profile flyout / to style into any area that's related to the user and their stuff.
+
+Throwing this one Brad's way so he can dig into the guts of a new system we haven't done much with yet, see how we export features for things that ride across distributions (as this will) and just get in the front end a bit more.",1.0,"Add banner picture to user profile fields list - it's becoming common practice with social tools to have a headshot type image (user profile picture) as well as a landscape / inspiration / larger image. Google+ FB and twitter all have them and it would be cool to have support for in our people distribution as well. Especially as we move towards implementation of side nav it would look REALLY nice for the user profile flyout / to style into any area that's related to the user and their stuff.
+
+Throwing this one Brad's way so he can dig into the guts of a new system we haven't done much with yet, see how we export features for things that ride across distributions (as this will) and just get in the front end a bit more.",1,add banner picture to user profile fields list it s becoming common practice with social tools to have a headshot type image user profile picture as well as a landscape inspiration larger image google fb and twitter all have them and it would be cool to have support for in our people distribution as well especially as we move towards implementation of side nav it would look really nice for the user profile flyout to style into any area that s related to the user and their stuff throwing this one brad s way so he can dig into the guts of a new system we haven t done much with yet see how we export features for things that ride across distributions as this will and just get in the front end a bit more ,1
+394,7192810822.0,IssuesEvent,2018-02-03 08:50:30,fifafu/BetterTouchTool,https://api.github.com/repos/fifafu/BetterTouchTool,closed,Suddenly BetterTouchTool Stop to work,Need other people who can reproduce the issue,"Hey,
+I am using this in order to copy\paste and more tasks like this like open a new tab, copy custom text and more... everyday it worked, suddenly i have open the app today and nothing is working. i have tried to restart and quit and open, and still does not work",1.0,"Suddenly BetterTouchTool Stop to work - Hey,
+I am using this in order to copy\paste and more tasks like this like open a new tab, copy custom text and more... everyday it worked, suddenly i have open the app today and nothing is working. i have tried to restart and quit and open, and still does not work",1,suddenly bettertouchtool stop to work hey i am using this in order to copy paste and more tasks like this like open a new tab copy custom text and more everyday it worked suddenly i have open the app today and nothing is working i have tried to restart and quit and open and still does not work,1
+1493,6122924471.0,IssuesEvent,2017-06-23 02:02:42,Endogix/WebFormWeaver,https://api.github.com/repos/Endogix/WebFormWeaver,opened,Export form structure,architecture business logic feature presentation logic,"Allow the user to export the form structure when they are finished editing the form, through a button near the bottom.
+
+Exporting will show a modal with the code - as optionally either HTML code, JSON, or XML.
+
+There should also be the option to specify an endpoint that this script can connect to through AJAX to save to a server, with another option to specify a function to call when the connection is complete (most likely to be a redirect function).
+
+## Acceptance criteria
+User should be able to:
+- [ ] Export the form as HTML code
+- [ ] Export the form as a JSON string
+- [ ] Export the form as an XML string
+- [ ] Export the form to an endpoint through AJAX (either JSON or XML)
+- [ ] Configure a function in the options to call when then connection to the endpoint is complete",1.0,"Export form structure - Allow the user to export the form structure when they are finished editing the form, through a button near the bottom.
+
+Exporting will show a modal with the code - as optionally either HTML code, JSON, or XML.
+
+There should also be the option to specify an endpoint that this script can connect to through AJAX to save to a server, with another option to specify a function to call when the connection is complete (most likely to be a redirect function).
+
+## Acceptance criteria
+User should be able to:
+- [ ] Export the form as HTML code
+- [ ] Export the form as a JSON string
+- [ ] Export the form as an XML string
+- [ ] Export the form to an endpoint through AJAX (either JSON or XML)
+- [ ] Configure a function in the options to call when then connection to the endpoint is complete",0,export form structure allow the user to export the form structure when they are finished editing the form through a button near the bottom exporting will show a modal with the code as optionally either html code json or xml there should also be the option to specify an endpoint that this script can connect to through ajax to save to a server with another option to specify a function to call when the connection is complete most likely to be a redirect function acceptance criteria user should be able to export the form as html code export the form as a json string export the form as an xml string export the form to an endpoint through ajax either json or xml configure a function in the options to call when then connection to the endpoint is complete,0
+886,17599816160.0,IssuesEvent,2021-08-17 10:22:21,quat1024/incorporeal-2-forge,https://api.github.com/repos/quat1024/incorporeal-2-forge,closed,MatcherUtils uses reflection,Incorporeal 2 talk with the botania people,Generally serializing and deserializing ICorporeaRequestMatchers is sorta messy and i should talk to botania people about this,1.0,MatcherUtils uses reflection - Generally serializing and deserializing ICorporeaRequestMatchers is sorta messy and i should talk to botania people about this,1,matcherutils uses reflection generally serializing and deserializing icorporearequestmatchers is sorta messy and i should talk to botania people about this,1
+189428,22047037811.0,IssuesEvent,2022-05-30 03:45:48,praneethpanasala/linux,https://api.github.com/repos/praneethpanasala/linux,closed,CVE-2021-3348 (High) detected in linuxlinux-4.19.6 - autoclosed,security vulnerability,"## CVE-2021-3348 - High Severity Vulnerability
+ Vulnerable Library - linuxlinux-4.19.6
+
+
+nbd_add_socket in drivers/block/nbd.c in the Linux kernel through 5.10.12 has an ndb_queue_rq use-after-free that could be triggered by local attackers (with access to the nbd device) via an I/O request at a certain point during device setup, aka CID-b98e762e3d71.
+
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",True,"CVE-2021-3348 (High) detected in linuxlinux-4.19.6 - autoclosed - ## CVE-2021-3348 - High Severity Vulnerability
+ Vulnerable Library - linuxlinux-4.19.6
+
+
+nbd_add_socket in drivers/block/nbd.c in the Linux kernel through 5.10.12 has an ndb_queue_rq use-after-free that could be triggered by local attackers (with access to the nbd device) via an I/O request at a certain point during device setup, aka CID-b98e762e3d71.
+
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",0,cve high detected in linuxlinux autoclosed cve high severity vulnerability vulnerable library linuxlinux apache software foundation asf library home page a href found in head commit a href found in base branch master vulnerable source files drivers block nbd c drivers block nbd c vulnerability details nbd add socket in drivers block nbd c in the linux kernel through has an ndb queue rq use after free that could be triggered by local attackers with access to the nbd device via an i o request at a certain point during device setup aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource ,0
+234024,19091149220.0,IssuesEvent,2021-11-29 12:15:07,weaveworks/flintlock,https://api.github.com/repos/weaveworks/flintlock,closed,Backfill `infrastructure/grpc/server.go` tests,good first issue help wanted area/testing priority/important-soon,"There are none yet, but should be pretty straightforward as everything in there is mostly interface delegation",1.0,"Backfill `infrastructure/grpc/server.go` tests - There are none yet, but should be pretty straightforward as everything in there is mostly interface delegation",0,backfill infrastructure grpc server go tests there are none yet but should be pretty straightforward as everything in there is mostly interface delegation,0
+27507,21799522054.0,IssuesEvent,2022-05-16 02:22:24,OpenLiberty/openliberty.io,https://api.github.com/repos/OpenLiberty/openliberty.io,opened,Provide ability to disable 'run in cloud' link/button for guides without rebuilding site,infrastructure,"If there's an issue with SNL that's impacting the OL guides, it would be good to have a way to remove the link to the SNL version of the guide without needing to rebuild the server.
+
+1. Add 'Heath Check' call/logic to ol.io server-side code (executes once every ~10min)
+ - Preferably, this would just call a (public) health check provided by SNL.
+ - Another option would be to GET some file in GH (guides-common) which would state whether links should be enabled/disabled (enabled by default). We'd use a GH Token to avoid any rate limiting. The value in the file would be set manually by guides team.
+ - Another option would be to create our own health check that queries guide(s) in SNL. This should be avoided.
+2. Expose health check result with our own API that the client will call.
+3. Guide page calls the API in previous step on load to determine whether the button should render or not.
+4. (This should probably be in a separate issue, as I would not consider it MVP) On /guides (and technically on each guide), the 'Run in cloud' tag would also either render or not based on the API in 2.",1.0,"Provide ability to disable 'run in cloud' link/button for guides without rebuilding site - If there's an issue with SNL that's impacting the OL guides, it would be good to have a way to remove the link to the SNL version of the guide without needing to rebuild the server.
+
+1. Add 'Heath Check' call/logic to ol.io server-side code (executes once every ~10min)
+ - Preferably, this would just call a (public) health check provided by SNL.
+ - Another option would be to GET some file in GH (guides-common) which would state whether links should be enabled/disabled (enabled by default). We'd use a GH Token to avoid any rate limiting. The value in the file would be set manually by guides team.
+ - Another option would be to create our own health check that queries guide(s) in SNL. This should be avoided.
+2. Expose health check result with our own API that the client will call.
+3. Guide page calls the API in previous step on load to determine whether the button should render or not.
+4. (This should probably be in a separate issue, as I would not consider it MVP) On /guides (and technically on each guide), the 'Run in cloud' tag would also either render or not based on the API in 2.",0,provide ability to disable run in cloud link button for guides without rebuilding site if there s an issue with snl that s impacting the ol guides it would be good to have a way to remove the link to the snl version of the guide without needing to rebuild the server add heath check call logic to ol io server side code executes once every preferably this would just call a public health check provided by snl another option would be to get some file in gh guides common which would state whether links should be enabled disabled enabled by default we d use a gh token to avoid any rate limiting the value in the file would be set manually by guides team another option would be to create our own health check that queries guide s in snl this should be avoided expose health check result with our own api that the client will call guide page calls the api in previous step on load to determine whether the button should render or not this should probably be in a separate issue as i would not consider it mvp on guides and technically on each guide the run in cloud tag would also either render or not based on the api in ,0
+3108,11868516767.0,IssuesEvent,2020-03-26 09:20:22,chocolatey-community/chocolatey-package-requests,https://api.github.com/repos/chocolatey-community/chocolatey-package-requests,closed,RFM - Software Ideas Viewer,Status: Available For Maintainer(s),"## Current Maintainer
+
+- [x] I am the maintainer of the package and wish to pass it to someone else;
+
+## Checklist
+
+- [x] Issue title starts with 'RFM - '
+
+## Existing Package Details
+
+Package URL: https://chocolatey.org/packages/software-ideas-viewer
+Package source URL: https://github.com/abejenaru/chocolatey-packages/tree/master/automatic/software-ideas-viewer
+",True,"RFM - Software Ideas Viewer - ## Current Maintainer
+
+- [x] I am the maintainer of the package and wish to pass it to someone else;
+
+## Checklist
+
+- [x] Issue title starts with 'RFM - '
+
+## Existing Package Details
+
+Package URL: https://chocolatey.org/packages/software-ideas-viewer
+Package source URL: https://github.com/abejenaru/chocolatey-packages/tree/master/automatic/software-ideas-viewer
+",0,rfm software ideas viewer current maintainer i am the maintainer of the package and wish to pass it to someone else checklist issue title starts with rfm existing package details package url package source url ,0
+1161,30278385782.0,IssuesEvent,2023-07-07 22:20:31,henrikfroehling/Trakt.NET,https://api.github.com/repos/henrikfroehling/Trakt.NET,closed,"Additional user information can be retrieved with `?extended=full,vip`",Library Module Comments Module People Module Seasons Module Shows Module Movies Module Episodes Module Lists Improvement Epic v1.4.0 Target-Branch: release-1.4.0,"- [x] #467
+- [x] #487
+- [x] #488
+- [x] #489
+- [x] #490
+- [x] #491
+- [x] #492
+- [x] #493
+- [x] #494
+- [x] #495
+- [x] #496
+- [x] #497
+- [x] #498
+- [x] #499
+- [x] #500
+- [x] #501",1.0,"Additional user information can be retrieved with `?extended=full,vip` - - [x] #467
+- [x] #487
+- [x] #488
+- [x] #489
+- [x] #490
+- [x] #491
+- [x] #492
+- [x] #493
+- [x] #494
+- [x] #495
+- [x] #496
+- [x] #497
+- [x] #498
+- [x] #499
+- [x] #500
+- [x] #501",1,additional user information can be retrieved with extended full vip ,1
+345,6553161044.0,IssuesEvent,2017-09-05 21:16:33,GDLActivity/GDL-Meetup,https://api.github.com/repos/GDLActivity/GDL-Meetup,closed,Patrocinio para meetup de septiembre,in progress Meetup Reach people,Estamos en el proceso de confirmar a Centraal para el meetup de septiembre.,1.0,Patrocinio para meetup de septiembre - Estamos en el proceso de confirmar a Centraal para el meetup de septiembre.,1,patrocinio para meetup de septiembre estamos en el proceso de confirmar a centraal para el meetup de septiembre ,1
+81484,10143440759.0,IssuesEvent,2019-08-04 12:15:48,bolta/ModularAudioSharp,https://api.github.com/repos/bolta/ModularAudioSharp,closed,Temperament の扱いを再考,Core ModDL design,"Temperament は現在 Node として実装している(Sequencer から ToneUser には音名が渡され、その後 Temperament が音名を周波数に変換する)が、
+周波数への変換は Sequencer 側で行った方がよさそう(Temperament は Node ではない普通のクラスにする)。
+
+理由
+* 現状の(Node としての)Temperament は各 Instrument の内部にあるので、ModDL から音律を設定したいときに面倒になりそう
+* MML でポルタメントを実現するとき、まだ発音していない終点の音名についても周波数を計算する必要があるが、現状では実際に発音しないと計算できない
+* デチューン、ビブラート等がやりにくい
+
+影響範囲
+* Sequencer が音名を与える対象は VarController\ 型の ToneUser だが、これを VarController\ 型の FreqUser に変更する必要がある
+",1.0,"Temperament の扱いを再考 - Temperament は現在 Node として実装している(Sequencer から ToneUser には音名が渡され、その後 Temperament が音名を周波数に変換する)が、
+周波数への変換は Sequencer 側で行った方がよさそう(Temperament は Node ではない普通のクラスにする)。
+
+理由
+* 現状の(Node としての)Temperament は各 Instrument の内部にあるので、ModDL から音律を設定したいときに面倒になりそう
+* MML でポルタメントを実現するとき、まだ発音していない終点の音名についても周波数を計算する必要があるが、現状では実際に発音しないと計算できない
+* デチューン、ビブラート等がやりにくい
+
+影響範囲
+* Sequencer が音名を与える対象は VarController\ 型の ToneUser だが、これを VarController\ 型の FreqUser に変更する必要がある
+",0,temperament の扱いを再考 temperament は現在 node として実装している(sequencer から toneuser には音名が渡され、その後 temperament が音名を周波数に変換する)が、 周波数への変換は sequencer 側で行った方がよさそう(temperament は node ではない普通のクラスにする)。 理由 現状の(node としての)temperament は各 instrument の内部にあるので、moddl から音律を設定したいときに面倒になりそう mml でポルタメントを実現するとき、まだ発音していない終点の音名についても周波数を計算する必要があるが、現状では実際に発音しないと計算できない デチューン、ビブラート等がやりにくい 影響範囲 sequencer が音名を与える対象は varcontroller 型の toneuser だが、これを varcontroller 型の frequser に変更する必要がある ,0
+773,14264612742.0,IssuesEvent,2020-11-20 15:58:59,wordpress-mobile/WordPress-iOS,https://api.github.com/repos/wordpress-mobile/WordPress-iOS,opened,P2: Add invite links,My Site P2 People Management,"TBD.
+
+As a P2 user, I want the ability to invite others to my P2.
+
+(Include People Management refactor here.)
+
+On the web:
+
+
+
+",1.0,"P2: Add invite links - TBD.
+
+As a P2 user, I want the ability to invite others to my P2.
+
+(Include People Management refactor here.)
+
+On the web:
+
+
+
+",1, add invite links tbd as a user i want the ability to invite others to my include people management refactor here on the web img width alt screen shot at am src img width alt screen shot at am src ,1
+9683,13759622301.0,IssuesEvent,2020-10-07 03:34:41,CMPUT301F20T01/Bookmark,https://api.github.com/repos/CMPUT301F20T01/Bookmark,opened,6: Create List of owned books,requirement,"### Task:
+Create a scrollable list of owned books
+This should be placed in its own tab, and should have layout as specified in figma.
+This should be a [viewElement] and should have custom adapters written for it to format.
+
+#### Story Points
+- 3
+
+#### Risk Level
+- medium
+",1.0,"6: Create List of owned books - ### Task:
+Create a scrollable list of owned books
+This should be placed in its own tab, and should have layout as specified in figma.
+This should be a [viewElement] and should have custom adapters written for it to format.
+
+#### Story Points
+- 3
+
+#### Risk Level
+- medium
+",0, create list of owned books task create a scrollable list of owned books this should be placed in its own tab and should have layout as specified in figma this should be a and should have custom adapters written for it to format story points risk level medium ,0
+700,12544853502.0,IssuesEvent,2020-06-05 17:55:15,microsoftgraph/microsoft-graph-toolkit,https://api.github.com/repos/microsoftgraph/microsoft-graph-toolkit,closed,"[BUG] people picker shows list of people when type is set to Group, or when group-id is set",1.3 Component: people-picker Priority: 0 State: In Review bug,"
+
+
+**Describe the bug**
+When I set the group-id and click inside the people picker, it shows my most recent people, even if those people are not members of the group
+
+Likewise, when I set the type to Group, it shows me my list of people, even though only groups are allowed to be selected
+
+**To Reproduce**
+Steps to reproduce the behavior:
+Go to the dev version of storybook and test with the `Group Id` and `Pick Groups` stories.
+
+**Expected behavior**
+At a minimum, the flyout with most recent people should not show when the people picker is filtered to a group, or only groups can be selected.
+
+",1.0,"[BUG] people picker shows list of people when type is set to Group, or when group-id is set -
+
+
+**Describe the bug**
+When I set the group-id and click inside the people picker, it shows my most recent people, even if those people are not members of the group
+
+Likewise, when I set the type to Group, it shows me my list of people, even though only groups are allowed to be selected
+
+**To Reproduce**
+Steps to reproduce the behavior:
+Go to the dev version of storybook and test with the `Group Id` and `Pick Groups` stories.
+
+**Expected behavior**
+At a minimum, the flyout with most recent people should not show when the people picker is filtered to a group, or only groups can be selected.
+
+",1, people picker shows list of people when type is set to group or when group id is set describe the bug when i set the group id and click inside the people picker it shows my most recent people even if those people are not members of the group likewise when i set the type to group it shows me my list of people even though only groups are allowed to be selected to reproduce steps to reproduce the behavior go to the dev version of storybook and test with the group id and pick groups stories expected behavior at a minimum the flyout with most recent people should not show when the people picker is filtered to a group or only groups can be selected ,1
+993,23538275437.0,IssuesEvent,2022-08-20 01:47:36,GoogleCloudPlatform/python-docs-samples,https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples,closed,Import errors in Wildlife Insights -- image-classification,type: question samples api: people-and-planet-ai,"## In which file did you encounter the issue?
+
+python-docs-samples/tree/main/people-and-planet-ai/image-classification/train_model.py
+
+### Did you change the file? If so, how?
+
+[My fork](https://github.com/GoogleCloudPlatform/python-docs-samples/compare/main...pravarmahajan:main)
+
+## Describe the issue
+
+
+
+I am executing the sample notebook [README.ipynb](https://github.com/GoogleCloudPlatform/python-docs-samples/blob/main/people-and-planet-ai/image-classification/README.ipynb) and running into errors. The errors suggest the mapper function doesn't have access to the external modules.
+
+
+
+So I started adding explicit imports inside each of the mapper functions as shown in my forked version. But in the latest error, the mapper function isn't able to locate a function (`url_get`) which is defined in the same script `train_model.py`.
+
+
+A few questions here:
+1. Is this the right way of doing it? Do we need to add explicit imports inside the mapper function for every external module? If so, what should I do about the function defined in the same script `url_get`?
+2. And if so, how are there no import errors for `logging` [here](https://github.com/pravarmahajan/python-docs-samples/blob/5f519b6d637cfb87cf132d2ada4b4dad1f31f12d/people-and-planet-ai/image-classification/train_model.py#L133)? ",1.0,"Import errors in Wildlife Insights -- image-classification - ## In which file did you encounter the issue?
+
+python-docs-samples/tree/main/people-and-planet-ai/image-classification/train_model.py
+
+### Did you change the file? If so, how?
+
+[My fork](https://github.com/GoogleCloudPlatform/python-docs-samples/compare/main...pravarmahajan:main)
+
+## Describe the issue
+
+
+
+I am executing the sample notebook [README.ipynb](https://github.com/GoogleCloudPlatform/python-docs-samples/blob/main/people-and-planet-ai/image-classification/README.ipynb) and running into errors. The errors suggest the mapper function doesn't have access to the external modules.
+
+
+
+So I started adding explicit imports inside each of the mapper functions as shown in my forked version. But in the latest error, the mapper function isn't able to locate a function (`url_get`) which is defined in the same script `train_model.py`.
+
+
+A few questions here:
+1. Is this the right way of doing it? Do we need to add explicit imports inside the mapper function for every external module? If so, what should I do about the function defined in the same script `url_get`?
+2. And if so, how are there no import errors for `logging` [here](https://github.com/pravarmahajan/python-docs-samples/blob/5f519b6d637cfb87cf132d2ada4b4dad1f31f12d/people-and-planet-ai/image-classification/train_model.py#L133)? ",1,import errors in wildlife insights image classification in which file did you encounter the issue python docs samples tree main people and planet ai image classification train model py did you change the file if so how describe the issue please be specific copying and pasting your invocation and the entire output is often helpful i am executing the sample notebook and running into errors the errors suggest the mapper function doesn t have access to the external modules so i started adding explicit imports inside each of the mapper functions as shown in my forked version but in the latest error the mapper function isn t able to locate a function url get which is defined in the same script train model py a few questions here is this the right way of doing it do we need to add explicit imports inside the mapper function for every external module if so what should i do about the function defined in the same script url get and if so how are there no import errors for logging ,1
+87247,8068901542.0,IssuesEvent,2018-08-06 01:56:58,project-koku/koku,https://api.github.com/repos/project-koku/koku,closed,Provide data output in CSV or JSON output,enhancement ready to test,"## User Story
+As a user, I want to be able to download the data used in a chart as a CSV or JSON file so that I can reuse it in external tools (Excel, Jasper Reports, etc).
+
+## Impacts
+API, UI, Docs
+
+## Role
+User
+
+## Assumptions
+The user will get the data in used to comprised the chart from the API in addition to any data they may have “hidden” based on any interactive selections in the UI.
+
+## API Details
+
+Add support for `media_type = 'text/csv'`
+http://www.django-rest-framework.org/api-guide/renderers/#custom-renderers
+https://github.com/mjumbewu/django-rest-framework-csv
+
+
+## UI Details
+https://github.com/project-koku/koku-ui/issues/21
+
+## Acceptance Criteria
+
+- [ ] Verify that a chart data can be downloaded from the user interface dashboard and reflects the same chart data seen in the dashboard.
+",1.0,"Provide data output in CSV or JSON output - ## User Story
+As a user, I want to be able to download the data used in a chart as a CSV or JSON file so that I can reuse it in external tools (Excel, Jasper Reports, etc).
+
+## Impacts
+API, UI, Docs
+
+## Role
+User
+
+## Assumptions
+The user will get the data in used to comprised the chart from the API in addition to any data they may have “hidden” based on any interactive selections in the UI.
+
+## API Details
+
+Add support for `media_type = 'text/csv'`
+http://www.django-rest-framework.org/api-guide/renderers/#custom-renderers
+https://github.com/mjumbewu/django-rest-framework-csv
+
+
+## UI Details
+https://github.com/project-koku/koku-ui/issues/21
+
+## Acceptance Criteria
+
+- [ ] Verify that a chart data can be downloaded from the user interface dashboard and reflects the same chart data seen in the dashboard.
+",0,provide data output in csv or json output user story as a user i want to be able to download the data used in a chart as a csv or json file so that i can reuse it in external tools excel jasper reports etc impacts api ui docs role user assumptions the user will get the data in used to comprised the chart from the api in addition to any data they may have “hidden” based on any interactive selections in the ui api details add support for media type text csv ui details acceptance criteria verify that a chart data can be downloaded from the user interface dashboard and reflects the same chart data seen in the dashboard ,0
+361647,25347332321.0,IssuesEvent,2022-11-19 10:59:46,astrolink/pluto,https://api.github.com/repos/astrolink/pluto,closed,Add semantic versioning of versions to the project,documentation enhancement help wanted,"Documentation
+
+https://semver.org/lang/pt-BR/
+
+If it is possible to add an action that will create the release version on github and distribute it in some channel",1.0,"Add semantic versioning of versions to the project - Documentation
+
+https://semver.org/lang/pt-BR/
+
+If it is possible to add an action that will create the release version on github and distribute it in some channel",0,add semantic versioning of versions to the project documentation if it is possible to add an action that will create the release version on github and distribute it in some channel,0
+675,12201145543.0,IssuesEvent,2020-04-30 06:37:57,bitprj/onboarding,https://api.github.com/repos/bitprj/onboarding,opened,Holly Cao,People_Onboarding,"# Technical Writer Onboarding Tasks
+
+#### Lead Name: @beccatran
+#### Member Name: @HollyCao
+#### Manager Name: @vkxu657
+
+## Lead's Tasks
+
+Finish these within next one week to ensure that the new member is onboarded properly:
+
+**You do not need to complete these following in the date listed, but you need to check on the new member and be prepared to answer their questions**
+
+### Day 0:
+
+- [ ] First email are about welcome to the club and information on Daniel's welcome webinar that is held weekly is sent out (Minh-Tu)
+- [ ] Welcome Webinar (Only President checks this)
+
+**The rest are the manager's responsibilities**
+
+- [ ] Check if they were able to join Slack
+ - [ ] Emoji react to their introduction post and give a warm welcome
+
+### Day 1-2:
+
+- [ ] Send template message with important resources (GitHub, Webinars,
+
+### Day 3:
+
+- [ ] Be there to answer questions for them
+- [ ] Check-in with them on their progress on the Webinars and Hubspot Tasks
+
+### Day 4-5:
+
+- [ ] Manager meets with the new members
+ - [ ] Make sure they know how to write meeting notes in GitHub
+ - [ ] Discuss what the People team does and what the new member's role might be
+ - [ ] Talk about the Handbook
+ - [ ] Talk about the onboarding process, show them where they can access it
+ - [ ] Discuss the Writing Tasks
+ - [ ] Note the feedback that the new member gives
+ - [ ] Tell them the next steps and their next project!
+ - [ ] Congratulate them on becoming a part of Bit!
+
+## Member's Tasks
+
+_Finish these within next one week to ensure that the you are onboarded properly:_
+
+### Check out the following links that you might need access to this week:
+- [ ] [GitHub](github.com/bitprj/marketing)
+ - Our L&D team is completing the webinar and will share it with you within the next day or so
+- [ ] [Airtable](airtable.com)
+ - Where we store all of our data in spreadsheets - you’ll need this for your second issue mentioned in the onboarding checklist
+- [ ] [Handbook](about.bitproject.org)
+ - Documentation on everything we do and where you can find all the webinars
+
+### Day 0:
+
+- [ ] Receive welcome email from Onboarding Team
+- [ ] Join our Slack
+ - [ ] Include in Profile
+ - Name
+ - Profile Picture
+ - In the description list out: School and team (Marketing)
+ - [ ] Send a greeting message through the #welcome Channel
+ - Name
+ - Position
+ - Major
+ - Year
+ - Spirit Animal
+- [ ] DM your Manager on Slack and let them know that you received your onboarding checklist
+- [ ] Locate your Manager's Calendly on their Slack profile and set up a 1:1 meeting for the 3rd or 4th day or onboarding
+
+### Day 1-2:
+
+- [ ] Sign up for Slack, Zoom, and [Airtable](https://airtable.com/tblRN2fngeYXUlGGp)
+- [ ] Have access to Bit Project Google Calendar
+- [ ] Watch [Webinars](about.bitproject.org)
+ - [ ] [Communication in a Remote Workspace](https://www.youtube.com/watch?v=2f9TkttynIk)
+ - [ ] Feedback
+ - [ ] GitHub
+ - [ ] Complete the [Webinar Quiz](https://airtable.com/shr31V9xdVRxOffMA)
+
+### Day 3-4:
+
+- [ ] Complete 1:1 with Technical Writing Lead
+ - [ ] Understand the overview of the onboarding procedure
+ - [ ] Breakdown of tasks and expectations for the first week
+ - [ ] Understand what a technical writer does
+ - [ ] Practice taking notes on Github during the 1:1
+ - [ ] Given Task
+
+### Day 5:
+- [ ] Complete your Writing Tasks:
+
+For ALL files:
+- Copy the test file
+- Push it into the repository
+- Label it yourname_writingtest.md
+
+Checklist for case study file:
+[ ] Grammar, punctuation, spelling check
+[ ] Word count: 500-700, about 1.5 pages in Google Docs if you're unsure
+[ ] Include 2-5 sentence abstract about what you think the content is about
+[ ] List of 3 strengths and 3 weaknesses you noticed when making changes; write brief explanation for all of them
+[ ] Organizational and structural issues; this includes correcting any necessary formatting or citing of sources
+
+Checklist for curriculum file:
+[ ] Grammar, punctuation, spelling check
+[ ] Word count: 650 words or less
+[ ] Include 2-5 sentence abstract about what you think the content is about
+[ ] List of 3 strengths and 3 weaknesses when making changes; write brief explanation for all of them
+[ ] Organizational and structural issues; this includes correcting any necessary formatting or citing of sources
+
+Checklist for About Us website blurb:
+[ ] Grammar, punctuation, spelling check
+[ ] Word count: 60-75
+[ ] Include 2-5 sentence abstract about what you think the content is about/what changes you made at top of file
+[ ] List of 3 strengths and 3 weaknesses you noticed when making changes; write brief explanation for all of them
+[ ] Organizational and structural issues; this includes correcting any necessary formatting or citing of sources
+
+- [ ] Attend Welcome Webinar (within you first week on onboarding)
+",1.0,"Holly Cao - # Technical Writer Onboarding Tasks
+
+#### Lead Name: @beccatran
+#### Member Name: @HollyCao
+#### Manager Name: @vkxu657
+
+## Lead's Tasks
+
+Finish these within next one week to ensure that the new member is onboarded properly:
+
+**You do not need to complete these following in the date listed, but you need to check on the new member and be prepared to answer their questions**
+
+### Day 0:
+
+- [ ] First email are about welcome to the club and information on Daniel's welcome webinar that is held weekly is sent out (Minh-Tu)
+- [ ] Welcome Webinar (Only President checks this)
+
+**The rest are the manager's responsibilities**
+
+- [ ] Check if they were able to join Slack
+ - [ ] Emoji react to their introduction post and give a warm welcome
+
+### Day 1-2:
+
+- [ ] Send template message with important resources (GitHub, Webinars,
+
+### Day 3:
+
+- [ ] Be there to answer questions for them
+- [ ] Check-in with them on their progress on the Webinars and Hubspot Tasks
+
+### Day 4-5:
+
+- [ ] Manager meets with the new members
+ - [ ] Make sure they know how to write meeting notes in GitHub
+ - [ ] Discuss what the People team does and what the new member's role might be
+ - [ ] Talk about the Handbook
+ - [ ] Talk about the onboarding process, show them where they can access it
+ - [ ] Discuss the Writing Tasks
+ - [ ] Note the feedback that the new member gives
+ - [ ] Tell them the next steps and their next project!
+ - [ ] Congratulate them on becoming a part of Bit!
+
+## Member's Tasks
+
+_Finish these within next one week to ensure that the you are onboarded properly:_
+
+### Check out the following links that you might need access to this week:
+- [ ] [GitHub](github.com/bitprj/marketing)
+ - Our L&D team is completing the webinar and will share it with you within the next day or so
+- [ ] [Airtable](airtable.com)
+ - Where we store all of our data in spreadsheets - you’ll need this for your second issue mentioned in the onboarding checklist
+- [ ] [Handbook](about.bitproject.org)
+ - Documentation on everything we do and where you can find all the webinars
+
+### Day 0:
+
+- [ ] Receive welcome email from Onboarding Team
+- [ ] Join our Slack
+ - [ ] Include in Profile
+ - Name
+ - Profile Picture
+ - In the description list out: School and team (Marketing)
+ - [ ] Send a greeting message through the #welcome Channel
+ - Name
+ - Position
+ - Major
+ - Year
+ - Spirit Animal
+- [ ] DM your Manager on Slack and let them know that you received your onboarding checklist
+- [ ] Locate your Manager's Calendly on their Slack profile and set up a 1:1 meeting for the 3rd or 4th day or onboarding
+
+### Day 1-2:
+
+- [ ] Sign up for Slack, Zoom, and [Airtable](https://airtable.com/tblRN2fngeYXUlGGp)
+- [ ] Have access to Bit Project Google Calendar
+- [ ] Watch [Webinars](about.bitproject.org)
+ - [ ] [Communication in a Remote Workspace](https://www.youtube.com/watch?v=2f9TkttynIk)
+ - [ ] Feedback
+ - [ ] GitHub
+ - [ ] Complete the [Webinar Quiz](https://airtable.com/shr31V9xdVRxOffMA)
+
+### Day 3-4:
+
+- [ ] Complete 1:1 with Technical Writing Lead
+ - [ ] Understand the overview of the onboarding procedure
+ - [ ] Breakdown of tasks and expectations for the first week
+ - [ ] Understand what a technical writer does
+ - [ ] Practice taking notes on Github during the 1:1
+ - [ ] Given Task
+
+### Day 5:
+- [ ] Complete your Writing Tasks:
+
+For ALL files:
+- Copy the test file
+- Push it into the repository
+- Label it yourname_writingtest.md
+
+Checklist for case study file:
+[ ] Grammar, punctuation, spelling check
+[ ] Word count: 500-700, about 1.5 pages in Google Docs if you're unsure
+[ ] Include 2-5 sentence abstract about what you think the content is about
+[ ] List of 3 strengths and 3 weaknesses you noticed when making changes; write brief explanation for all of them
+[ ] Organizational and structural issues; this includes correcting any necessary formatting or citing of sources
+
+Checklist for curriculum file:
+[ ] Grammar, punctuation, spelling check
+[ ] Word count: 650 words or less
+[ ] Include 2-5 sentence abstract about what you think the content is about
+[ ] List of 3 strengths and 3 weaknesses when making changes; write brief explanation for all of them
+[ ] Organizational and structural issues; this includes correcting any necessary formatting or citing of sources
+
+Checklist for About Us website blurb:
+[ ] Grammar, punctuation, spelling check
+[ ] Word count: 60-75
+[ ] Include 2-5 sentence abstract about what you think the content is about/what changes you made at top of file
+[ ] List of 3 strengths and 3 weaknesses you noticed when making changes; write brief explanation for all of them
+[ ] Organizational and structural issues; this includes correcting any necessary formatting or citing of sources
+
+- [ ] Attend Welcome Webinar (within you first week on onboarding)
+",1,holly cao technical writer onboarding tasks lead name beccatran member name hollycao manager name lead s tasks finish these within next one week to ensure that the new member is onboarded properly you do not need to complete these following in the date listed but you need to check on the new member and be prepared to answer their questions day first email are about welcome to the club and information on daniel s welcome webinar that is held weekly is sent out minh tu welcome webinar only president checks this the rest are the manager s responsibilities check if they were able to join slack emoji react to their introduction post and give a warm welcome day send template message with important resources github webinars day be there to answer questions for them check in with them on their progress on the webinars and hubspot tasks day manager meets with the new members make sure they know how to write meeting notes in github discuss what the people team does and what the new member s role might be talk about the handbook talk about the onboarding process show them where they can access it discuss the writing tasks note the feedback that the new member gives tell them the next steps and their next project congratulate them on becoming a part of bit member s tasks finish these within next one week to ensure that the you are onboarded properly check out the following links that you might need access to this week github com bitprj marketing our l d team is completing the webinar and will share it with you within the next day or so airtable com where we store all of our data in spreadsheets you’ll need this for your second issue mentioned in the onboarding checklist about bitproject org documentation on everything we do and where you can find all the webinars day receive welcome email from onboarding team join our slack include in profile name profile picture in the description list out school and team marketing send a greeting message through the welcome channel name position major year spirit animal dm your manager on slack and let them know that you received your onboarding checklist locate your manager s calendly on their slack profile and set up a meeting for the or day or onboarding day sign up for slack zoom and have access to bit project google calendar watch about bitproject org feedback github complete the day complete with technical writing lead understand the overview of the onboarding procedure breakdown of tasks and expectations for the first week understand what a technical writer does practice taking notes on github during the given task day complete your writing tasks for all files copy the test file push it into the repository label it yourname writingtest md checklist for case study file grammar punctuation spelling check word count about pages in google docs if you re unsure include sentence abstract about what you think the content is about list of strengths and weaknesses you noticed when making changes write brief explanation for all of them organizational and structural issues this includes correcting any necessary formatting or citing of sources checklist for curriculum file grammar punctuation spelling check word count words or less include sentence abstract about what you think the content is about list of strengths and weaknesses when making changes write brief explanation for all of them organizational and structural issues this includes correcting any necessary formatting or citing of sources checklist for about us website blurb grammar punctuation spelling check word count include sentence abstract about what you think the content is about what changes you made at top of file list of strengths and weaknesses you noticed when making changes write brief explanation for all of them organizational and structural issues this includes correcting any necessary formatting or citing of sources attend welcome webinar within you first week on onboarding ,1
+334122,29820356778.0,IssuesEvent,2023-06-17 01:33:39,unifyai/ivy,https://api.github.com/repos/unifyai/ivy,closed,Fix array.test_array__eq__,Sub Task Failing Test,"| | |
+|---|---|
+|tensorflow|
+|torch|
+|numpy|
+|jax|
+|paddle|
+
+
+
+FAILED ivy_tests/test_ivy/test_misc/test_array.py::test_array__eq__[cpu-ivy.functional.backends.paddle-False-False]
+
+2023-06-04T03:00:49.3030318Z E RuntimeError: (NotFound) The kernel with key (CPU, NCHW, uint8) of kernel `subtract` is not registered.2023-06-04T03:00:49.3031332Z E [Hint: Expected kernel_iter == iter->second.end() && kernel_key.backend() == Backend::CPU != true, but received kernel_iter == iter->second.end() && kernel_key.backend() == Backend::CPU:1 == true:1.] (at /paddle/paddle/phi/core/kernel_factory.cc:147)2023-06-04T03:00:49.3036652Z E ivy.utils.exceptions.IvyBackendException: paddle: equal: RuntimeError: (NotFound) The kernel with key (CPU, NCHW, uint8) of kernel `subtract` is not registered.2023-06-04T03:00:49.3037553Z E [Hint: Expected kernel_iter == iter->second.end() && kernel_key.backend() == Backend::CPU != true, but received kernel_iter == iter->second.end() && kernel_key.backend() == Backend::CPU:1 == true:1.] (at /paddle/paddle/phi/core/kernel_factory.cc:147)2023-06-04T03:00:49.3038223Z E 2023-06-04T03:00:49.3038705Z E Falsifying example: test_array__eq__(2023-06-04T03:00:49.3038988Z E on_device='cpu',2023-06-04T03:00:49.3039550Z E dtype_and_x=(['uint8', 'uint8'],2023-06-04T03:00:49.3039841Z E [array(0, dtype=uint8), array(0, dtype=uint8)]),2023-06-04T03:00:49.3040181Z E ground_truth_backend='tensorflow',2023-06-04T03:00:49.3040893Z E method_name='__eq__',2023-06-04T03:00:49.3041174Z E method_flags=MethodTestFlags(2023-06-04T03:00:49.3041443Z E num_positional_args=0,2023-06-04T03:00:49.3041855Z E as_variable=[False],2023-06-04T03:00:49.3042423Z E native_arrays=[False],2023-06-04T03:00:49.3042658Z E container_flags=[False],2023-06-04T03:00:49.3042872Z E ),2023-06-04T03:00:49.3043319Z E class_name='Array',2023-06-04T03:00:49.3045096Z E init_flags=InitMethodTestFlags(2023-06-04T03:00:49.3045447Z E num_positional_args=0,2023-06-04T03:00:49.3045743Z E as_variable=[False],2023-06-04T03:00:49.3046035Z E native_arrays=[False],2023-06-04T03:00:49.3046299Z E ),2023-06-04T03:00:49.3046537Z E )2023-06-04T03:00:49.3046762Z E 2023-06-04T03:00:49.3047413Z E You can reproduce this example by temporarily adding @reproduce_failure('6.75.9', b'AXicY2dAAQAAeAAI') as a decorator on your test case
+
+",1.0,"Fix array.test_array__eq__ - | | |
+|---|---|
+|tensorflow|
+|torch|
+|numpy|
+|jax|
+|paddle|
+
+
+
+FAILED ivy_tests/test_ivy/test_misc/test_array.py::test_array__eq__[cpu-ivy.functional.backends.paddle-False-False]
+
+2023-06-04T03:00:49.3030318Z E RuntimeError: (NotFound) The kernel with key (CPU, NCHW, uint8) of kernel `subtract` is not registered.2023-06-04T03:00:49.3031332Z E [Hint: Expected kernel_iter == iter->second.end() && kernel_key.backend() == Backend::CPU != true, but received kernel_iter == iter->second.end() && kernel_key.backend() == Backend::CPU:1 == true:1.] (at /paddle/paddle/phi/core/kernel_factory.cc:147)2023-06-04T03:00:49.3036652Z E ivy.utils.exceptions.IvyBackendException: paddle: equal: RuntimeError: (NotFound) The kernel with key (CPU, NCHW, uint8) of kernel `subtract` is not registered.2023-06-04T03:00:49.3037553Z E [Hint: Expected kernel_iter == iter->second.end() && kernel_key.backend() == Backend::CPU != true, but received kernel_iter == iter->second.end() && kernel_key.backend() == Backend::CPU:1 == true:1.] (at /paddle/paddle/phi/core/kernel_factory.cc:147)2023-06-04T03:00:49.3038223Z E 2023-06-04T03:00:49.3038705Z E Falsifying example: test_array__eq__(2023-06-04T03:00:49.3038988Z E on_device='cpu',2023-06-04T03:00:49.3039550Z E dtype_and_x=(['uint8', 'uint8'],2023-06-04T03:00:49.3039841Z E [array(0, dtype=uint8), array(0, dtype=uint8)]),2023-06-04T03:00:49.3040181Z E ground_truth_backend='tensorflow',2023-06-04T03:00:49.3040893Z E method_name='__eq__',2023-06-04T03:00:49.3041174Z E method_flags=MethodTestFlags(2023-06-04T03:00:49.3041443Z E num_positional_args=0,2023-06-04T03:00:49.3041855Z E as_variable=[False],2023-06-04T03:00:49.3042423Z E native_arrays=[False],2023-06-04T03:00:49.3042658Z E container_flags=[False],2023-06-04T03:00:49.3042872Z E ),2023-06-04T03:00:49.3043319Z E class_name='Array',2023-06-04T03:00:49.3045096Z E init_flags=InitMethodTestFlags(2023-06-04T03:00:49.3045447Z E num_positional_args=0,2023-06-04T03:00:49.3045743Z E as_variable=[False],2023-06-04T03:00:49.3046035Z E native_arrays=[False],2023-06-04T03:00:49.3046299Z E ),2023-06-04T03:00:49.3046537Z E )2023-06-04T03:00:49.3046762Z E 2023-06-04T03:00:49.3047413Z E You can reproduce this example by temporarily adding @reproduce_failure('6.75.9', b'AXicY2dAAQAAeAAI') as a decorator on your test case
+
+",0,fix array test array eq tensorflow img src torch img src numpy img src jax img src paddle img src failed ivy tests test ivy test misc test array py test array eq e runtimeerror notfound the kernel with key cpu nchw of kernel subtract is not registered e at paddle paddle phi core kernel factory cc e ivy utils exceptions ivybackendexception paddle equal runtimeerror notfound the kernel with key cpu nchw of kernel subtract is not registered e at paddle paddle phi core kernel factory cc e e falsifying example test array eq e on device cpu e dtype and x e e ground truth backend tensorflow e method name eq e method flags methodtestflags e num positional args e as variable e native arrays e container flags e e class name array e init flags initmethodtestflags e num positional args e as variable e native arrays e e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case ,0
+214264,16578960961.0,IssuesEvent,2021-05-31 09:04:46,hazelcast/hazelcast,https://api.github.com/repos/hazelcast/hazelcast,closed,com.hazelcast.internal.partition.impl.InternalPartitionServiceLiteMemberTest,Source: Internal Team: Core Type: Test-Failure,"_master_ (commit c3c641fc6af0924ccc140f24d98e37f7f8dd66b1)
+
+Failed on IBM JDK 8: http://jenkins.hazelcast.com/view/Official%20Builds/job/Hazelcast-master-IbmJDK8/131/testReport/com.hazelcast.internal.partition.impl/InternalPartitionServiceLiteMemberTest/
+
+Following tests from `com.hazelcast.internal.partition.impl.InternalPartitionServiceLiteMemberTest` failed with the same exception:
+- `test_liteMemberCanTerminate_whenDataMemberExistsInCluster`
+- `test_dataMemberCanShutdownSafely_whenOnlyLiteMemberExistsInCluster`
+- `test_liteMemberCanShutdownSafely_whenDataMemberExistsInCluster`
+- `test_liteMemberCanTerminate_withClusterSize2`
+- `test_dataMemberCanTerminate_whenOnlyLiteMemberExistsInCluster`
+
+Stacktrace (for `test_liteMemberCanTerminate_whenDataMemberExistsInCluster`):
+```
+org.junit.runners.model.TestTimedOutException: test timed out after 120000 milliseconds
+ at java.lang.Thread.startImpl(Native Method)
+ at java.lang.Thread.start(Thread.java:992)
+ at com.hazelcast.spi.impl.operationexecutor.impl.OperationExecutorImpl.startAll(OperationExecutorImpl.java:527)
+ at com.hazelcast.spi.impl.operationexecutor.impl.OperationExecutorImpl.start(OperationExecutorImpl.java:521)
+ at com.hazelcast.spi.impl.operationservice.impl.OperationServiceImpl.start(OperationServiceImpl.java:488)
+ at com.hazelcast.spi.impl.NodeEngineImpl.start(NodeEngineImpl.java:227)
+ at com.hazelcast.instance.impl.Node.start(Node.java:447)
+ at com.hazelcast.instance.impl.HazelcastInstanceImpl.(HazelcastInstanceImpl.java:124)
+ at com.hazelcast.instance.impl.HazelcastInstanceFactory.constructHazelcastInstance(HazelcastInstanceFactory.java:211)
+ at com.hazelcast.instance.impl.HazelcastInstanceFactory.newHazelcastInstance(HazelcastInstanceFactory.java:190)
+ at com.hazelcast.test.TestHazelcastInstanceFactory.newHazelcastInstance(TestHazelcastInstanceFactory.java:224)
+ at com.hazelcast.test.TestHazelcastInstanceFactory.newHazelcastInstance(TestHazelcastInstanceFactory.java:106)
+ at com.hazelcast.internal.partition.impl.InternalPartitionServiceLiteMemberTest.test_liteMemberCanTerminate_whenDataMemberExistsInCluster(InternalPartitionServiceLiteMemberTest.java:330)
+ at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
+ at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:90)
+ at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:55)
+ at java.lang.reflect.Method.invoke(Method.java:508)
+ at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
+ at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
+ at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
+ at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
+ at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:115)
+ at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:107)
+ at java.util.concurrent.FutureTask.run(FutureTask.java:277)
+ at java.lang.Thread.run(Thread.java:822)
+```
+
+Standard output can be downloaded here - https://s3.console.aws.amazon.com/s3/object/j-artifacts?region=us-east-1&prefix=Hazelcast-master-IbmJDK8/131/Hazelcast-master-IbmJDK8-131.tar
+",1.0,"com.hazelcast.internal.partition.impl.InternalPartitionServiceLiteMemberTest - _master_ (commit c3c641fc6af0924ccc140f24d98e37f7f8dd66b1)
+
+Failed on IBM JDK 8: http://jenkins.hazelcast.com/view/Official%20Builds/job/Hazelcast-master-IbmJDK8/131/testReport/com.hazelcast.internal.partition.impl/InternalPartitionServiceLiteMemberTest/
+
+Following tests from `com.hazelcast.internal.partition.impl.InternalPartitionServiceLiteMemberTest` failed with the same exception:
+- `test_liteMemberCanTerminate_whenDataMemberExistsInCluster`
+- `test_dataMemberCanShutdownSafely_whenOnlyLiteMemberExistsInCluster`
+- `test_liteMemberCanShutdownSafely_whenDataMemberExistsInCluster`
+- `test_liteMemberCanTerminate_withClusterSize2`
+- `test_dataMemberCanTerminate_whenOnlyLiteMemberExistsInCluster`
+
+Stacktrace (for `test_liteMemberCanTerminate_whenDataMemberExistsInCluster`):
+```
+org.junit.runners.model.TestTimedOutException: test timed out after 120000 milliseconds
+ at java.lang.Thread.startImpl(Native Method)
+ at java.lang.Thread.start(Thread.java:992)
+ at com.hazelcast.spi.impl.operationexecutor.impl.OperationExecutorImpl.startAll(OperationExecutorImpl.java:527)
+ at com.hazelcast.spi.impl.operationexecutor.impl.OperationExecutorImpl.start(OperationExecutorImpl.java:521)
+ at com.hazelcast.spi.impl.operationservice.impl.OperationServiceImpl.start(OperationServiceImpl.java:488)
+ at com.hazelcast.spi.impl.NodeEngineImpl.start(NodeEngineImpl.java:227)
+ at com.hazelcast.instance.impl.Node.start(Node.java:447)
+ at com.hazelcast.instance.impl.HazelcastInstanceImpl.(HazelcastInstanceImpl.java:124)
+ at com.hazelcast.instance.impl.HazelcastInstanceFactory.constructHazelcastInstance(HazelcastInstanceFactory.java:211)
+ at com.hazelcast.instance.impl.HazelcastInstanceFactory.newHazelcastInstance(HazelcastInstanceFactory.java:190)
+ at com.hazelcast.test.TestHazelcastInstanceFactory.newHazelcastInstance(TestHazelcastInstanceFactory.java:224)
+ at com.hazelcast.test.TestHazelcastInstanceFactory.newHazelcastInstance(TestHazelcastInstanceFactory.java:106)
+ at com.hazelcast.internal.partition.impl.InternalPartitionServiceLiteMemberTest.test_liteMemberCanTerminate_whenDataMemberExistsInCluster(InternalPartitionServiceLiteMemberTest.java:330)
+ at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
+ at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:90)
+ at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:55)
+ at java.lang.reflect.Method.invoke(Method.java:508)
+ at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
+ at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
+ at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
+ at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
+ at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:115)
+ at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:107)
+ at java.util.concurrent.FutureTask.run(FutureTask.java:277)
+ at java.lang.Thread.run(Thread.java:822)
+```
+
+Standard output can be downloaded here - https://s3.console.aws.amazon.com/s3/object/j-artifacts?region=us-east-1&prefix=Hazelcast-master-IbmJDK8/131/Hazelcast-master-IbmJDK8-131.tar
+",0,com hazelcast internal partition impl internalpartitionservicelitemembertest master commit failed on ibm jdk following tests from com hazelcast internal partition impl internalpartitionservicelitemembertest failed with the same exception test litemembercanterminate whendatamemberexistsincluster test datamembercanshutdownsafely whenonlylitememberexistsincluster test litemembercanshutdownsafely whendatamemberexistsincluster test litemembercanterminate test datamembercanterminate whenonlylitememberexistsincluster stacktrace for test litemembercanterminate whendatamemberexistsincluster org junit runners model testtimedoutexception test timed out after milliseconds at java lang thread startimpl native method at java lang thread start thread java at com hazelcast spi impl operationexecutor impl operationexecutorimpl startall operationexecutorimpl java at com hazelcast spi impl operationexecutor impl operationexecutorimpl start operationexecutorimpl java at com hazelcast spi impl operationservice impl operationserviceimpl start operationserviceimpl java at com hazelcast spi impl nodeengineimpl start nodeengineimpl java at com hazelcast instance impl node start node java at com hazelcast instance impl hazelcastinstanceimpl hazelcastinstanceimpl java at com hazelcast instance impl hazelcastinstancefactory constructhazelcastinstance hazelcastinstancefactory java at com hazelcast instance impl hazelcastinstancefactory newhazelcastinstance hazelcastinstancefactory java at com hazelcast test testhazelcastinstancefactory newhazelcastinstance testhazelcastinstancefactory java at com hazelcast test testhazelcastinstancefactory newhazelcastinstance testhazelcastinstancefactory java at com hazelcast internal partition impl internalpartitionservicelitemembertest test litemembercanterminate whendatamemberexistsincluster internalpartitionservicelitemembertest java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at com hazelcast test failontimeoutstatement callablestatement call failontimeoutstatement java at com hazelcast test failontimeoutstatement callablestatement call failontimeoutstatement java at java util concurrent futuretask run futuretask java at java lang thread run thread java standard output can be downloaded here ,0
+940,20822912785.0,IssuesEvent,2022-03-18 17:12:06,FortAwesome/Font-Awesome,https://api.github.com/repos/FortAwesome/Font-Awesome,closed,Icon request: Unisex or male/female,new icon medical / health (category) genders (category) users & people (category) humanitarian (category),"I propose to add a unisex or combined male/female icon. It can be based on the existing male and female icons, combining two halves into a single new icon.
+
+**What can it be used for?**
+It's very useful in cases when indicating something unisex where the specific gender is not relevant.
+
+**Example image**
+",1.0,"Icon request: Unisex or male/female - I propose to add a unisex or combined male/female icon. It can be based on the existing male and female icons, combining two halves into a single new icon.
+
+**What can it be used for?**
+It's very useful in cases when indicating something unisex where the specific gender is not relevant.
+
+**Example image**
+",1,icon request unisex or male female i propose to add a unisex or combined male female icon it can be based on the existing male and female icons combining two halves into a single new icon what can it be used for it s very useful in cases when indicating something unisex where the specific gender is not relevant example image ,1
+93854,15946426168.0,IssuesEvent,2021-04-15 01:02:27,jgeraigery/core,https://api.github.com/repos/jgeraigery/core,opened,CVE-2020-26945 (High) detected in mybatis-3.4.2.jar,security vulnerability,"## CVE-2020-26945 - High Severity Vulnerability
+ Vulnerable Library - mybatis-3.4.2.jar
+
+
The MyBatis SQL mapper framework makes it easier to use a relational database with object-oriented
+ applications. MyBatis couples objects with stored procedures or SQL statements using a XML descriptor or
+ annotations. Simplicity is the biggest advantage of the MyBatis data mapper over object relational mapping
+ tools.
Path to vulnerable library: /home/wss-scanner/.m2/repository/org/mybatis/mybatis/3.4.2/mybatis-3.4.2.jar,/home/wss-scanner/.m2/repository/org/mybatis/mybatis/3.4.2/mybatis-3.4.2.jar,/home/wss-scanner/.m2/repository/org/mybatis/mybatis/3.4.2/mybatis-3.4.2.jar
The MyBatis SQL mapper framework makes it easier to use a relational database with object-oriented
+ applications. MyBatis couples objects with stored procedures or SQL statements using a XML descriptor or
+ annotations. Simplicity is the biggest advantage of the MyBatis data mapper over object relational mapping
+ tools.
Path to vulnerable library: /home/wss-scanner/.m2/repository/org/mybatis/mybatis/3.4.2/mybatis-3.4.2.jar,/home/wss-scanner/.m2/repository/org/mybatis/mybatis/3.4.2/mybatis-3.4.2.jar,/home/wss-scanner/.m2/repository/org/mybatis/mybatis/3.4.2/mybatis-3.4.2.jar
+
+
+
+
+
+",0,cve high detected in mybatis jar cve high severity vulnerability vulnerable library mybatis jar the mybatis sql mapper framework makes it easier to use a relational database with object oriented applications mybatis couples objects with stored procedures or sql statements using a xml descriptor or annotations simplicity is the biggest advantage of the mybatis data mapper over object relational mapping tools library home page a href path to dependency file core nimbus test pom xml path to vulnerable library home wss scanner repository org mybatis mybatis mybatis jar home wss scanner repository org mybatis mybatis mybatis jar home wss scanner repository org mybatis mybatis mybatis jar dependency hierarchy activiti spring boot starter jar root library activiti engine jar x mybatis jar vulnerable library found in base branch master vulnerability details mybatis before mishandles deserialization of object streams publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org mybatis mybatis isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org activiti activiti spring boot starter org activiti activiti engine org mybatis mybatis isminimumfixversionavailable true minimumfixversion org mybatis mybatis basebranches vulnerabilityidentifier cve vulnerabilitydetails mybatis before mishandles deserialization of object streams vulnerabilityurl ,0
+257934,27563834356.0,IssuesEvent,2023-03-08 01:09:50,LynRodWS/alcor,https://api.github.com/repos/LynRodWS/alcor,opened,"CVE-2017-18640 (High) detected in snakeyaml-1.25.jar, snakeyaml-1.23.jar",security vulnerability,"## CVE-2017-18640 - High Severity Vulnerability
+ Vulnerable Libraries - snakeyaml-1.25.jar, snakeyaml-1.23.jar
+
Path to dependency file: /services/route_manager/pom.xml
+
Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.25/snakeyaml-1.25.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.25/snakeyaml-1.25.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.25/snakeyaml-1.25.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.25/snakeyaml-1.25.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.25/snakeyaml-1.25.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.25/snakeyaml-1.25.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.25/snakeyaml-1.25.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.25/snakeyaml-1.25.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.25/snakeyaml-1.25.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.25/snakeyaml-1.25.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.25/snakeyaml-1.25.jar
Path to dependency file: /services/vpc_manager/pom.xml
+
Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.23/snakeyaml-1.23.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.23/snakeyaml-1.23.jar
Path to dependency file: /services/route_manager/pom.xml
+
Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.25/snakeyaml-1.25.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.25/snakeyaml-1.25.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.25/snakeyaml-1.25.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.25/snakeyaml-1.25.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.25/snakeyaml-1.25.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.25/snakeyaml-1.25.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.25/snakeyaml-1.25.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.25/snakeyaml-1.25.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.25/snakeyaml-1.25.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.25/snakeyaml-1.25.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.25/snakeyaml-1.25.jar
Path to dependency file: /services/vpc_manager/pom.xml
+
Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.23/snakeyaml-1.23.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.23/snakeyaml-1.23.jar
Direct dependency fix Resolution (org.springframework.boot:spring-boot-starter-actuator): 2.3.0.RELEASE
Fix Resolution (org.yaml:snakeyaml): 1.26
+
Direct dependency fix Resolution (org.springframework.boot:spring-boot-starter-actuator): 2.3.0.RELEASE
+
+
+
+
+
+***
+:rescue_worker_helmet: Automatic Remediation is available for this issue",0,cve high detected in snakeyaml jar snakeyaml jar cve high severity vulnerability vulnerable libraries snakeyaml jar snakeyaml jar snakeyaml jar yaml parser and emitter for java library home page a href path to dependency file services route manager pom xml path to vulnerable library home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar dependency hierarchy spring boot starter actuator release jar root library spring boot starter release jar x snakeyaml jar vulnerable library snakeyaml jar yaml parser and emitter for java library home page a href path to dependency file services vpc manager pom xml path to vulnerable library home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar dependency hierarchy spring boot starter actuator release jar root library spring boot starter release jar x snakeyaml jar vulnerable library found in base branch master vulnerability details the alias feature in snakeyaml before allows entity expansion during a load operation a related issue to cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org yaml snakeyaml direct dependency fix resolution org springframework boot spring boot starter actuator release fix resolution org yaml snakeyaml direct dependency fix resolution org springframework boot spring boot starter actuator release rescue worker helmet automatic remediation is available for this issue,0
+15,2649587188.0,IssuesEvent,2015-03-15 02:21:52,fskreuz/Tree-Trail,https://api.github.com/repos/fskreuz/Tree-Trail,closed,Add crops in Badge,QA people! Test these!,"As a user, I should be able to input what crops are to be planted in that area.",1.0,"Add crops in Badge - As a user, I should be able to input what crops are to be planted in that area.",1,add crops in badge as a user i should be able to input what crops are to be planted in that area ,1
+94117,10791131315.0,IssuesEvent,2019-11-05 16:07:14,usaybia/usaybia-data,https://api.github.com/repos/usaybia/usaybia-data,closed,Linebreaks for Transkribus,add to documentation question,"Dear all,
+it is somehow not possible to add a baseline in Transkribus on page123 l. 1-5. Unfortunately I dont know why.
+
+Also I want to ask whether to keep everything in one line at page 129 (الكامل at line 2-25) or seperate the poetic meter key in a line on its own. At line 2-16 on the same page is a name in braces (نحم الدين) is this also supposed to be in the line or create a new one (and then mark it as heading)? Same goes for page 127 at line 2-20. And at least I found a strange linebreak at page 130 l. 5-6.
+",1.0,"Linebreaks for Transkribus - Dear all,
+it is somehow not possible to add a baseline in Transkribus on page123 l. 1-5. Unfortunately I dont know why.
+
+Also I want to ask whether to keep everything in one line at page 129 (الكامل at line 2-25) or seperate the poetic meter key in a line on its own. At line 2-16 on the same page is a name in braces (نحم الدين) is this also supposed to be in the line or create a new one (and then mark it as heading)? Same goes for page 127 at line 2-20. And at least I found a strange linebreak at page 130 l. 5-6.
+",0,linebreaks for transkribus dear all it is somehow not possible to add a baseline in transkribus on l unfortunately i dont know why also i want to ask whether to keep everything in one line at page الكامل at line or seperate the poetic meter key in a line on its own at line on the same page is a name in braces نحم الدين is this also supposed to be in the line or create a new one and then mark it as heading same goes for page at line and at least i found a strange linebreak at page l ,0
+105307,16637200761.0,IssuesEvent,2021-06-04 01:34:57,easycv/easycv,https://api.github.com/repos/easycv/easycv,opened,CVE-2021-28678 (High) detected in Pillow-8.1.1-cp37-cp37m-manylinux1_x86_64.whl,security vulnerability,"## CVE-2021-28678 - High Severity Vulnerability
+ Vulnerable Library - Pillow-8.1.1-cp37-cp37m-manylinux1_x86_64.whl
+
+
+
+An issue was discovered in Pillow before 8.2.0. For BLP data, BlpImagePlugin did not properly check that reads (after jumping to file offsets) returned data. This could lead to a DoS where the decoder could be run a large number of times on empty data.
+
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",True,"CVE-2021-28678 (High) detected in Pillow-8.1.1-cp37-cp37m-manylinux1_x86_64.whl - ## CVE-2021-28678 - High Severity Vulnerability
+ Vulnerable Library - Pillow-8.1.1-cp37-cp37m-manylinux1_x86_64.whl
+
+
+
+An issue was discovered in Pillow before 8.2.0. For BLP data, BlpImagePlugin did not properly check that reads (after jumping to file offsets) returned data. This could lead to a DoS where the decoder could be run a large number of times on empty data.
+
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",0,cve high detected in pillow whl cve high severity vulnerability vulnerable library pillow whl python imaging library fork library home page a href path to dependency file easycv path to vulnerable library easycv dependency hierarchy x pillow whl vulnerable library found in base branch master vulnerability details an issue was discovered in pillow before for blp data blpimageplugin did not properly check that reads after jumping to file offsets returned data this could lead to a dos where the decoder could be run a large number of times on empty data publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution pillow step up your open source security game with whitesource ,0
+795,14766938726.0,IssuesEvent,2021-01-10 03:28:09,restincode/restincode,https://api.github.com/repos/restincode/restincode,opened,Yonathan Klijnsma (@ydklijnsma),Add Person Needs Review People,"Please fill out as much information as you can. No fields are required, but the more you can provide the better.
+
+**General Info**
+* First name: Yonathan
+* Last name: Klijnsma
+* Handle:
+* Birth Year:
+* Death Year:
+* Link to Obituary:
+* Group Affiliations:
+* URL to main photo (or attach to Issue):
+* Description of person and/or activities:
+* Facebook memorial group URL:
+
+**Social Media Links**
+* Twitter: @ydklijnsma
+* Github:
+* LinkedIn:
+* Facebook:
+* Other:
+
+**Contributions**
+Repeat as many times as needed
+
+* Project name:
+* Project URL:
+* Project Description:
+
+**Photo Gallery**
+* URL(s) to additional photos:
+",1.0,"Yonathan Klijnsma (@ydklijnsma) - Please fill out as much information as you can. No fields are required, but the more you can provide the better.
+
+**General Info**
+* First name: Yonathan
+* Last name: Klijnsma
+* Handle:
+* Birth Year:
+* Death Year:
+* Link to Obituary:
+* Group Affiliations:
+* URL to main photo (or attach to Issue):
+* Description of person and/or activities:
+* Facebook memorial group URL:
+
+**Social Media Links**
+* Twitter: @ydklijnsma
+* Github:
+* LinkedIn:
+* Facebook:
+* Other:
+
+**Contributions**
+Repeat as many times as needed
+
+* Project name:
+* Project URL:
+* Project Description:
+
+**Photo Gallery**
+* URL(s) to additional photos:
+",1,yonathan klijnsma ydklijnsma please fill out as much information as you can no fields are required but the more you can provide the better general info first name yonathan last name klijnsma handle birth year death year link to obituary group affiliations url to main photo or attach to issue description of person and or activities facebook memorial group url social media links twitter ydklijnsma github linkedin facebook other contributions repeat as many times as needed project name project url project description photo gallery url s to additional photos ,1
+169431,26802382198.0,IssuesEvent,2023-02-01 15:56:05,department-of-veterans-affairs/vets-design-system-documentation,https://api.github.com/repos/department-of-veterans-affairs/vets-design-system-documentation,reopened,"508-defect-2 [COGNITION, KEYBOARD]: Focus MUST not be hidden by the leftnav",accessibility vsp-design-system-team va-sidenav,"# [508-defect-2](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-2)
+
+[This is a duplicate ticket of the VAA issue here](https://github.com/department-of-veterans-affairs/va.gov-team/issues/22824), but I assume it may be easier to fix at the source (hence the duplicate ticket here in DST).
+
+
+
+
+
+## Feedback framework
+
+- **❗️ Must** for if the feedback must be applied
+- **⚠️ Should** if the feedback is best practice
+- **✔️ Consider** for suggestions/enhancements
+
+## Definition of done
+
+1. Review and acknowledge feedback.
+1. Fix and/or document decisions made.
+1. Accessibility specialist will close ticket after reviewing documented decisions / validating fix.
+
+## Point of Contact
+
+
+
+**VFS Point of Contact:** _Josh_
+
+## User Story or Problem Statement
+
+As a keyboard user using magnification or on a smaller viewport, I always want to be able to see and track focus so I can navigate and interact with the page.
+
+## Details
+
+At smaller viewports or when magnified, the in this section menu (leftnav) will become sticky on the page. This results in it occasionally covering focusable elements. [Keyboard users who rely on visible focus to interact with the page will struggle to do so as shown in research done by the gov.uk team.](https://technology.blog.gov.uk/2018/05/21/sticky-elements-functionality-and-accessibility-testing/)
+
+> Keyboard users often navigate web pages using the TAB key to move through focusable elements, such as links and input boxes. The SHIFT + TAB key combination can be used to move through those elements backwards. We found that where the sticky header overlapped the page, navigating backwards through elements often left the focused element obscured by the sticky element.
+
+> This was an accessibility barrier for users. The sticky element was preventing a commonly used navigation technique from working properly, leaving users unable to see or read the element they had focused on.
+
+> Does this mean that sticky functionality should never be implemented? No, but it must be considered and tested carefully, particularly if the sticky element in any way overlaps other content. A safer use would be a sticky element positioned over an empty part of the page, or to have the content scroll in a way that it is never overlapped by the sticky element.
+
+This is a sitewide issue [that has been resolved before in ticket 19779](https://github.com/department-of-veterans-affairs/va.gov-team/issues/19779) by [referencing the in this section menu on Pittsburgh which is not sticky](https://www.va.gov/pittsburgh-health-care/work-with-us/jobs-careers/).
+
+## Acceptance Criteria
+
+- [ ] Focus is not hidden on the page
+
+
+## Environment
+
+* Operating System: Any
+* Browser: Any (reduce viewport to activate sticky element)
+* Screenreading device: Any
+* Server destination: staging
+
+
+## Steps to Recreate
+
+1. Enter `https://staging.va.gov/careers-employment/vocational-rehabilitation/` in browser
+2. Reduce screen width to ~600px
+3. Tab all the way through the page
+4. Shift tab all the way back up the page
+5. Confirm focus is hidden by the sticky profile menu
+
+## Proposed Solution
+
+[This has been resolved before in ticket 19779](https://github.com/department-of-veterans-affairs/va.gov-team/issues/19779) by [referencing the in this section menu on Pittsburgh which is not sticky](https://www.va.gov/pittsburgh-health-care/work-with-us/jobs-careers/).
+
+## WCAG or Vendor Guidance (optional)
+
+* [Gov.uk - Sticky elements functionality and accessibility testing](https://technology.blog.gov.uk/2018/05/21/sticky-elements-functionality-and-accessibility-testing/)
+* [WCAG 2.4.7: Focus Visible](https://www.w3.org/WAI/WCAG21/Understanding/focus-visible.html)
+
+
+## Screenshots or Trace Logs
+Focus gets hidden when tabbing back up the page
+
+https://user-images.githubusercontent.com/14154792/113886654-da62f280-978e-11eb-8226-60eadc7427a9.mov
+
+
+
+https://user-images.githubusercontent.com/14154792/184384970-cf2ed2a4-b4a8-42b8-b534-159c3338ffea.mov
+
+
+
+
+
+
+",1.0,"508-defect-2 [COGNITION, KEYBOARD]: Focus MUST not be hidden by the leftnav - # [508-defect-2](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-2)
+
+[This is a duplicate ticket of the VAA issue here](https://github.com/department-of-veterans-affairs/va.gov-team/issues/22824), but I assume it may be easier to fix at the source (hence the duplicate ticket here in DST).
+
+
+
+
+
+## Feedback framework
+
+- **❗️ Must** for if the feedback must be applied
+- **⚠️ Should** if the feedback is best practice
+- **✔️ Consider** for suggestions/enhancements
+
+## Definition of done
+
+1. Review and acknowledge feedback.
+1. Fix and/or document decisions made.
+1. Accessibility specialist will close ticket after reviewing documented decisions / validating fix.
+
+## Point of Contact
+
+
+
+**VFS Point of Contact:** _Josh_
+
+## User Story or Problem Statement
+
+As a keyboard user using magnification or on a smaller viewport, I always want to be able to see and track focus so I can navigate and interact with the page.
+
+## Details
+
+At smaller viewports or when magnified, the in this section menu (leftnav) will become sticky on the page. This results in it occasionally covering focusable elements. [Keyboard users who rely on visible focus to interact with the page will struggle to do so as shown in research done by the gov.uk team.](https://technology.blog.gov.uk/2018/05/21/sticky-elements-functionality-and-accessibility-testing/)
+
+> Keyboard users often navigate web pages using the TAB key to move through focusable elements, such as links and input boxes. The SHIFT + TAB key combination can be used to move through those elements backwards. We found that where the sticky header overlapped the page, navigating backwards through elements often left the focused element obscured by the sticky element.
+
+> This was an accessibility barrier for users. The sticky element was preventing a commonly used navigation technique from working properly, leaving users unable to see or read the element they had focused on.
+
+> Does this mean that sticky functionality should never be implemented? No, but it must be considered and tested carefully, particularly if the sticky element in any way overlaps other content. A safer use would be a sticky element positioned over an empty part of the page, or to have the content scroll in a way that it is never overlapped by the sticky element.
+
+This is a sitewide issue [that has been resolved before in ticket 19779](https://github.com/department-of-veterans-affairs/va.gov-team/issues/19779) by [referencing the in this section menu on Pittsburgh which is not sticky](https://www.va.gov/pittsburgh-health-care/work-with-us/jobs-careers/).
+
+## Acceptance Criteria
+
+- [ ] Focus is not hidden on the page
+
+
+## Environment
+
+* Operating System: Any
+* Browser: Any (reduce viewport to activate sticky element)
+* Screenreading device: Any
+* Server destination: staging
+
+
+## Steps to Recreate
+
+1. Enter `https://staging.va.gov/careers-employment/vocational-rehabilitation/` in browser
+2. Reduce screen width to ~600px
+3. Tab all the way through the page
+4. Shift tab all the way back up the page
+5. Confirm focus is hidden by the sticky profile menu
+
+## Proposed Solution
+
+[This has been resolved before in ticket 19779](https://github.com/department-of-veterans-affairs/va.gov-team/issues/19779) by [referencing the in this section menu on Pittsburgh which is not sticky](https://www.va.gov/pittsburgh-health-care/work-with-us/jobs-careers/).
+
+## WCAG or Vendor Guidance (optional)
+
+* [Gov.uk - Sticky elements functionality and accessibility testing](https://technology.blog.gov.uk/2018/05/21/sticky-elements-functionality-and-accessibility-testing/)
+* [WCAG 2.4.7: Focus Visible](https://www.w3.org/WAI/WCAG21/Understanding/focus-visible.html)
+
+
+## Screenshots or Trace Logs
+Focus gets hidden when tabbing back up the page
+
+https://user-images.githubusercontent.com/14154792/113886654-da62f280-978e-11eb-8226-60eadc7427a9.mov
+
+
+
+https://user-images.githubusercontent.com/14154792/184384970-cf2ed2a4-b4a8-42b8-b534-159c3338ffea.mov
+
+
+
+
+
+
+",0, defect focus must not be hidden by the leftnav but i assume it may be easier to fix at the source hence the duplicate ticket here in dst enter an issue title using the format brief description of the problem edit buttons need aria label for context add another user link will not receive keyboard focus heading levels should increase by one error messages should be more specific blue button on blue background does not have sufficient contrast ratio feedback framework ❗️ must for if the feedback must be applied ⚠️ should if the feedback is best practice ✔️ consider for suggestions enhancements definition of done review and acknowledge feedback fix and or document decisions made accessibility specialist will close ticket after reviewing documented decisions validating fix point of contact vfs point of contact josh user story or problem statement as a keyboard user using magnification or on a smaller viewport i always want to be able to see and track focus so i can navigate and interact with the page details at smaller viewports or when magnified the in this section menu leftnav will become sticky on the page this results in it occasionally covering focusable elements keyboard users often navigate web pages using the tab key to move through focusable elements such as links and input boxes the shift tab key combination can be used to move through those elements backwards we found that where the sticky header overlapped the page navigating backwards through elements often left the focused element obscured by the sticky element this was an accessibility barrier for users the sticky element was preventing a commonly used navigation technique from working properly leaving users unable to see or read the element they had focused on does this mean that sticky functionality should never be implemented no but it must be considered and tested carefully particularly if the sticky element in any way overlaps other content a safer use would be a sticky element positioned over an empty part of the page or to have the content scroll in a way that it is never overlapped by the sticky element this is a sitewide issue by acceptance criteria focus is not hidden on the page environment operating system any browser any reduce viewport to activate sticky element screenreading device any server destination staging steps to recreate enter in browser reduce screen width to tab all the way through the page shift tab all the way back up the page confirm focus is hidden by the sticky profile menu proposed solution by wcag or vendor guidance optional screenshots or trace logs focus gets hidden when tabbing back up the page ,0
+139277,11255117428.0,IssuesEvent,2020-01-12 06:21:50,supercollider/supercollider,https://api.github.com/repos/supercollider/supercollider,closed,FR: add docs for testsuite,comp: project docs comp: test suite enhancement,"
+
+## Motivation
+
+There's little information currently on the c++ testsuite, AFAICT.
+
+## Description of Proposed Feature
+
+Add documentation about running c++ tests.
+
+## Plan for Implementation
+Put info about running `ctest` on the SC wiki.
+
+What do you think?
+
+",1.0,"FR: add docs for testsuite -
+
+## Motivation
+
+There's little information currently on the c++ testsuite, AFAICT.
+
+## Description of Proposed Feature
+
+Add documentation about running c++ tests.
+
+## Plan for Implementation
+Put info about running `ctest` on the SC wiki.
+
+What do you think?
+
+",0,fr add docs for testsuite motivation there s little information currently on the c testsuite afaict description of proposed feature add documentation about running c tests plan for implementation put info about running ctest on the sc wiki what do you think ,0
+497,8556017916.0,IssuesEvent,2018-11-08 11:49:02,folivoraAI/BetterTouchTool,https://api.github.com/repos/folivoraAI/BetterTouchTool,closed,3-finger click bug?,Need other people who can reproduce the issue,"
+
+I have middle click mapped to a 3 finger click. When I lift one finger up (so 2 fingers are on the trackpad now, but no additional pressure applied), another, 2-finger left click is triggered. I do not have a 2 finger click mapped in BTT, so this is the system click being triggered. I am not sure why this happens, not sure if this is a bug or a setting I have overlooked. Any help is appreciated.
+
+
+
+Macbook Pro (2015) Trackpad
+
+
+#### Device information:
+
+Type of Mac: MacBook Pro
+macOS version: 10.12.3
+BetterTouchTool version: 2.017 (608)
+
+",1.0,"3-finger click bug? -
+
+I have middle click mapped to a 3 finger click. When I lift one finger up (so 2 fingers are on the trackpad now, but no additional pressure applied), another, 2-finger left click is triggered. I do not have a 2 finger click mapped in BTT, so this is the system click being triggered. I am not sure why this happens, not sure if this is a bug or a setting I have overlooked. Any help is appreciated.
+
+
+
+Macbook Pro (2015) Trackpad
+
+
+#### Device information:
+
+Type of Mac: MacBook Pro
+macOS version: 10.12.3
+BetterTouchTool version: 2.017 (608)
+
+",1, finger click bug i have middle click mapped to a finger click when i lift one finger up so fingers are on the trackpad now but no additional pressure applied another finger left click is triggered i do not have a finger click mapped in btt so this is the system click being triggered i am not sure why this happens not sure if this is a bug or a setting i have overlooked any help is appreciated macbook pro trackpad device information type of mac macbook pro macos version bettertouchtool version ,1
+166305,20718424729.0,IssuesEvent,2022-03-13 01:37:58,kapseliboi/crowdfunding-frontend,https://api.github.com/repos/kapseliboi/crowdfunding-frontend,opened,CVE-2021-32803 (High) detected in tar-2.2.1.tgz,security vulnerability,"## CVE-2021-32803 - High Severity Vulnerability
+ Vulnerable Library - tar-2.2.1.tgz
+
+
+
+The npm package ""tar"" (aka node-tar) before versions 6.1.2, 5.0.7, 4.4.15, and 3.2.3 has an arbitrary File Creation/Overwrite vulnerability via insufficient symlink protection. `node-tar` aims to guarantee that any file whose location would be modified by a symbolic link is not extracted. This is, in part, achieved by ensuring that extracted directories are not symlinks. Additionally, in order to prevent unnecessary `stat` calls to determine whether a given path is a directory, paths are cached when directories are created. This logic was insufficient when extracting tar files that contained both a directory and a symlink with the same name as the directory. This order of operations resulted in the directory being created and added to the `node-tar` directory cache. When a directory is present in the directory cache, subsequent calls to mkdir for that directory are skipped. However, this is also where `node-tar` checks for symlinks occur. By first creating a directory, and then replacing that directory with a symlink, it was thus possible to bypass `node-tar` symlink checks on directories, essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location, thus allowing arbitrary file creation and overwrite. This issue was addressed in releases 3.2.3, 4.4.15, 5.0.7 and 6.1.2.
+
+
Direct dependency fix Resolution (nodemon): 1.11.1
+
+
+
+
+
+***
+Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",True,"CVE-2021-32803 (High) detected in tar-2.2.1.tgz - ## CVE-2021-32803 - High Severity Vulnerability
+ Vulnerable Library - tar-2.2.1.tgz
+
+
+
+The npm package ""tar"" (aka node-tar) before versions 6.1.2, 5.0.7, 4.4.15, and 3.2.3 has an arbitrary File Creation/Overwrite vulnerability via insufficient symlink protection. `node-tar` aims to guarantee that any file whose location would be modified by a symbolic link is not extracted. This is, in part, achieved by ensuring that extracted directories are not symlinks. Additionally, in order to prevent unnecessary `stat` calls to determine whether a given path is a directory, paths are cached when directories are created. This logic was insufficient when extracting tar files that contained both a directory and a symlink with the same name as the directory. This order of operations resulted in the directory being created and added to the `node-tar` directory cache. When a directory is present in the directory cache, subsequent calls to mkdir for that directory are skipped. However, this is also where `node-tar` checks for symlinks occur. By first creating a directory, and then replacing that directory with a symlink, it was thus possible to bypass `node-tar` symlink checks on directories, essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location, thus allowing arbitrary file creation and overwrite. This issue was addressed in releases 3.2.3, 4.4.15, 5.0.7 and 6.1.2.
+
+
Direct dependency fix Resolution (nodemon): 1.11.1
+
+
+
+
+
+***
+Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",0,cve high detected in tar tgz cve high severity vulnerability vulnerable library tar tgz tar for node library home page a href dependency hierarchy nodemon tgz root library chokidar tgz fsevents tgz node pre gyp tgz x tar tgz vulnerable library found in base branch master vulnerability details the npm package tar aka node tar before versions and has an arbitrary file creation overwrite vulnerability via insufficient symlink protection node tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted this is in part achieved by ensuring that extracted directories are not symlinks additionally in order to prevent unnecessary stat calls to determine whether a given path is a directory paths are cached when directories are created this logic was insufficient when extracting tar files that contained both a directory and a symlink with the same name as the directory this order of operations resulted in the directory being created and added to the node tar directory cache when a directory is present in the directory cache subsequent calls to mkdir for that directory are skipped however this is also where node tar checks for symlinks occur by first creating a directory and then replacing that directory with a symlink it was thus possible to bypass node tar symlink checks on directories essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location thus allowing arbitrary file creation and overwrite this issue was addressed in releases and publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tar direct dependency fix resolution nodemon step up your open source security game with whitesource ,0
+737,13434111535.0,IssuesEvent,2020-09-07 10:50:39,samvera-labs/samvera-connect,https://api.github.com/repos/samvera-labs/samvera-connect,opened,Case Study: Batch Update - A User Driven Approach: What We Talk About When We Talk About Design Thinking,Devs Managers (general) Managers (repo) Metadata Newcomers Presentation (30 min slot (20+5+5)) UI/UX people,"Karen Shaw karen.shaw@northwestern.edu
+Adam Arling
+
+Over the past year, NUL dev team has implemented and refined a workflow for modular development of repository applications. It starts with addressing a specific user need or problem. Using design-thinking techniques, we next generate visual solutions through rudimentary wire framing, white boarding sessions and architecture discussions. We then move to API design and mocking before starting development with two teams working independently, from the API, outwards.
+
+This presentation will demonstrate the iterative approach in action; using a recent major batch update feature as a case study. We’ll explore how our approach allowed us to stay connected to our users and helped keep our development team in sync. Finally, we’ll reflect on what we’ve found most successful in this approach as well as stumbling blocks we encountered along the way.
+
+",1.0,"Case Study: Batch Update - A User Driven Approach: What We Talk About When We Talk About Design Thinking - Karen Shaw karen.shaw@northwestern.edu
+Adam Arling
+
+Over the past year, NUL dev team has implemented and refined a workflow for modular development of repository applications. It starts with addressing a specific user need or problem. Using design-thinking techniques, we next generate visual solutions through rudimentary wire framing, white boarding sessions and architecture discussions. We then move to API design and mocking before starting development with two teams working independently, from the API, outwards.
+
+This presentation will demonstrate the iterative approach in action; using a recent major batch update feature as a case study. We’ll explore how our approach allowed us to stay connected to our users and helped keep our development team in sync. Finally, we’ll reflect on what we’ve found most successful in this approach as well as stumbling blocks we encountered along the way.
+
+",1,case study batch update a user driven approach what we talk about when we talk about design thinking karen shaw karen shaw northwestern edu adam arling over the past year nul dev team has implemented and refined a workflow for modular development of repository applications it starts with addressing a specific user need or problem using design thinking techniques we next generate visual solutions through rudimentary wire framing white boarding sessions and architecture discussions we then move to api design and mocking before starting development with two teams working independently from the api outwards this presentation will demonstrate the iterative approach in action using a recent major batch update feature as a case study we’ll explore how our approach allowed us to stay connected to our users and helped keep our development team in sync finally we’ll reflect on what we’ve found most successful in this approach as well as stumbling blocks we encountered along the way ,1
+378,7047948227.0,IssuesEvent,2018-01-02 15:43:21,PopulateTools/gobierto,https://api.github.com/repos/PopulateTools/gobierto,closed,Move travels and gifts external URLs to module settings,gobierto-people,"Travel and Gift sections in GobiertoPeople module can be configured to redirecto an external URL (outside Gobierto). This is implemented as an option in `config/application.yml`, and therefore applies to all the sites.
+
+We need that each site can define its own urls, so we need to:
+
+- move these settings to the GobiertoPeople module / Settings page
+- remove the option from the application.yml
+- if the option is set, then redirect to the url, otherwise don't show the link in the home",1.0,"Move travels and gifts external URLs to module settings - Travel and Gift sections in GobiertoPeople module can be configured to redirecto an external URL (outside Gobierto). This is implemented as an option in `config/application.yml`, and therefore applies to all the sites.
+
+We need that each site can define its own urls, so we need to:
+
+- move these settings to the GobiertoPeople module / Settings page
+- remove the option from the application.yml
+- if the option is set, then redirect to the url, otherwise don't show the link in the home",1,move travels and gifts external urls to module settings travel and gift sections in gobiertopeople module can be configured to redirecto an external url outside gobierto this is implemented as an option in config application yml and therefore applies to all the sites we need that each site can define its own urls so we need to move these settings to the gobiertopeople module settings page remove the option from the application yml if the option is set then redirect to the url otherwise don t show the link in the home,1
+696222,23891449127.0,IssuesEvent,2022-09-08 11:49:29,ITI/searcch,https://api.github.com/repos/ITI/searcch,opened,"Move ""Best Practices"" directly to hub",priority: low,"At present, the ""Best Practices"" menu item takes the user to the searcch.cyberexperimentation.org web site.
+
+We should move this information directly to the hub. ",1.0,"Move ""Best Practices"" directly to hub - At present, the ""Best Practices"" menu item takes the user to the searcch.cyberexperimentation.org web site.
+
+We should move this information directly to the hub. ",0,move best practices directly to hub at present the best practices menu item takes the user to the searcch cyberexperimentation org web site we should move this information directly to the hub ,0
+1183,30787820432.0,IssuesEvent,2023-07-31 14:18:23,culturesofknowledge/emlo-project,https://api.github.com/repos/culturesofknowledge/emlo-project,opened,"People: Synonyms: In the person record, these need to display in a list, not in a continuous line",people form feedback0625,"Please, otherwise it’s a) hard to see what’s there, or b) set them easily into alphabetical order (which I’m doing increasingly) with an easy sort / copy / paste using excel.
+EMLO-Edit-NEW:
+
+And EMLO-Edit-OLD:
+
+",1.0,"People: Synonyms: In the person record, these need to display in a list, not in a continuous line - Please, otherwise it’s a) hard to see what’s there, or b) set them easily into alphabetical order (which I’m doing increasingly) with an easy sort / copy / paste using excel.
+EMLO-Edit-NEW:
+
+And EMLO-Edit-OLD:
+
+",1,people synonyms in the person record these need to display in a list not in a continuous line please otherwise it’s a hard to see what’s there or b set them easily into alphabetical order which i’m doing increasingly with an easy sort copy paste using excel emlo edit new and emlo edit old ,1
+15934,28181235401.0,IssuesEvent,2023-04-04 02:44:00,NikkisStickies/EcommerceWebsite,https://api.github.com/repos/NikkisStickies/EcommerceWebsite,opened,Recently Viewed,Functional Requirement,The website should have a feature that displays recently viewed products to improve user experience.,1.0,Recently Viewed - The website should have a feature that displays recently viewed products to improve user experience.,0,recently viewed the website should have a feature that displays recently viewed products to improve user experience ,0
+195,4174220135.0,IssuesEvent,2016-06-21 13:26:38,wordpress-mobile/WordPress-iOS,https://api.github.com/repos/wordpress-mobile/WordPress-iOS,opened,People Management: Viewers,People Management,"Whenever a Blog is set to private, we'll need to:
+
+- [ ] Allow listing Viewers
+- [ ] Sending Invitations for the Viewer Role
+",1.0,"People Management: Viewers - Whenever a Blog is set to private, we'll need to:
+
+- [ ] Allow listing Viewers
+- [ ] Sending Invitations for the Viewer Role
+",1,people management viewers whenever a blog is set to private we ll need to allow listing viewers sending invitations for the viewer role ,1
+853,16110672549.0,IssuesEvent,2021-04-27 20:43:31,restincode/restincode,https://api.github.com/repos/restincode/restincode,opened,Veruus,Add Person Needs Review People,"Please fill out as much information as you can. No fields are required, but the more you can provide the better.
+
+**General Info**
+* First name:
+* Last name:
+* Handle: Veruus
+* Birth Year:
+* Death Year:
+* Link to Obituary:
+* Group Affiliations:
+* URL to main photo (or attach to Issue):
+* Description of person and/or activities:
+* Facebook memorial group URL:
+
+**Social Media Links**
+* Twitter: https://twitter.com/veruus
+* Github:
+* LinkedIn:
+* Facebook:
+* Other:
+
+**Contributions**
+DEFCON goon (swag)
+
+* Project name:
+* Project URL:
+* Project Description:
+
+**Photo Gallery**
+* URL(s) to additional photos:
+",1.0,"Veruus - Please fill out as much information as you can. No fields are required, but the more you can provide the better.
+
+**General Info**
+* First name:
+* Last name:
+* Handle: Veruus
+* Birth Year:
+* Death Year:
+* Link to Obituary:
+* Group Affiliations:
+* URL to main photo (or attach to Issue):
+* Description of person and/or activities:
+* Facebook memorial group URL:
+
+**Social Media Links**
+* Twitter: https://twitter.com/veruus
+* Github:
+* LinkedIn:
+* Facebook:
+* Other:
+
+**Contributions**
+DEFCON goon (swag)
+
+* Project name:
+* Project URL:
+* Project Description:
+
+**Photo Gallery**
+* URL(s) to additional photos:
+",1,veruus please fill out as much information as you can no fields are required but the more you can provide the better general info first name last name handle veruus birth year death year link to obituary group affiliations url to main photo or attach to issue description of person and or activities facebook memorial group url social media links twitter github linkedin facebook other contributions defcon goon swag project name project url project description photo gallery url s to additional photos ,1
+1153,30181451316.0,IssuesEvent,2023-07-04 09:09:42,newtheatre/history-project,https://api.github.com/repos/newtheatre/history-project,closed,Paul Young,report-tool people/bio,"End user submitted issue from page: [/people/paul_young/](https://history.newtheatre.org.uk/people/paul_young/)
+---
+Paul graduated in Theology.
+*Marilyn Bird 1999*
+",1.0,"Paul Young - End user submitted issue from page: [/people/paul_young/](https://history.newtheatre.org.uk/people/paul_young/)
+---
+Paul graduated in Theology.
+*Marilyn Bird 1999*
+",1,paul young end user submitted issue from page paul graduated in theology marilyn bird ,1
+276,5331055452.0,IssuesEvent,2017-02-15 18:32:47,PopulateTools/gobierto,https://api.github.com/repos/PopulateTools/gobierto,closed,Admin: Hide political / executive options depending on the option chosen,enhancement gobierto-people,"As a follow up of #245, in the controls to check if a person is a politic or an executive, add a JS interaction to hide the list of political groups and the options government / opposition when executive is chosen.
+
+There's no need to do a check in the backend, because it's already performed, so it's just a UI issue.
+
+
+",1.0,"Admin: Hide political / executive options depending on the option chosen - As a follow up of #245, in the controls to check if a person is a politic or an executive, add a JS interaction to hide the list of political groups and the options government / opposition when executive is chosen.
+
+There's no need to do a check in the backend, because it's already performed, so it's just a UI issue.
+
+
+",1,admin hide political executive options depending on the option chosen as a follow up of in the controls to check if a person is a politic or an executive add a js interaction to hide the list of political groups and the options government opposition when executive is chosen there s no need to do a check in the backend because it s already performed so it s just a ui issue ,1
+526469,15293763404.0,IssuesEvent,2021-02-24 00:58:34,internetarchive/openlibrary,https://api.github.com/repos/internetarchive/openlibrary,closed,Authors not findable using Search,Module: Authors Module: Solr Needs: Lead Priority: 3 Theme: Search Type: Bug,"None of these authors are findable using search even though (many) author records exist for them.
+
+- United States. Congress. House. Committee on the District of Columbia. Subcommittee on Investigation of Food Storage and Prices
+- United States. Congress. House. Committee on the Pacific Railroad
+- United States. Congress. House. Committee on Transportation and Infrastructure. Subcommittee on Aviation
+
+I thought perhaps it was associated with long authors, but this one is findable even though it's longer than the _Committee on the Pacific Railroad._
+Princeton University. Dept. of Economics and Social Institutions. Industrial Relations Section.
+
+Here's the list of author records for one of the unsearchable names:
+/authors/OL4620383A United States. Congress. House. Committee on Transportation and Infrastructure. Subcommittee on Aviation
+/authors/OL4620614A
+/authors/OL4625592A
+/authors/OL4620175A
+/authors/OL4620217A
+/authors/OL4625064A
+/authors/OL48266A
+/authors/OL4626004A
+/authors/OL4625755A
+/authors/OL4625259A
+/authors/OL4625904A
+/authors/OL4625065A
+/authors/OL4625754A
+/authors/OL4620231A
+/authors/OL4620213A
+/authors/OL4623159A
+/authors/OL4625899A",1.0,"Authors not findable using Search - None of these authors are findable using search even though (many) author records exist for them.
+
+- United States. Congress. House. Committee on the District of Columbia. Subcommittee on Investigation of Food Storage and Prices
+- United States. Congress. House. Committee on the Pacific Railroad
+- United States. Congress. House. Committee on Transportation and Infrastructure. Subcommittee on Aviation
+
+I thought perhaps it was associated with long authors, but this one is findable even though it's longer than the _Committee on the Pacific Railroad._
+Princeton University. Dept. of Economics and Social Institutions. Industrial Relations Section.
+
+Here's the list of author records for one of the unsearchable names:
+/authors/OL4620383A United States. Congress. House. Committee on Transportation and Infrastructure. Subcommittee on Aviation
+/authors/OL4620614A
+/authors/OL4625592A
+/authors/OL4620175A
+/authors/OL4620217A
+/authors/OL4625064A
+/authors/OL48266A
+/authors/OL4626004A
+/authors/OL4625755A
+/authors/OL4625259A
+/authors/OL4625904A
+/authors/OL4625065A
+/authors/OL4625754A
+/authors/OL4620231A
+/authors/OL4620213A
+/authors/OL4623159A
+/authors/OL4625899A",0,authors not findable using search none of these authors are findable using search even though many author records exist for them united states congress house committee on the district of columbia subcommittee on investigation of food storage and prices united states congress house committee on the pacific railroad united states congress house committee on transportation and infrastructure subcommittee on aviation i thought perhaps it was associated with long authors but this one is findable even though it s longer than the committee on the pacific railroad princeton university dept of economics and social institutions industrial relations section here s the list of author records for one of the unsearchable names authors united states congress house committee on transportation and infrastructure subcommittee on aviation authors authors authors authors authors authors authors authors authors authors authors authors authors authors authors authors ,0
+1092,26754508452.0,IssuesEvent,2023-01-30 22:36:46,Reviewable/Reviewable,https://api.github.com/repos/Reviewable/Reviewable,closed,Review participants panel,enhancement feature:people,"We should make a dedicated panel that lists all participants in the review with all aspects of their current state, and allows sorting or filtering the list based on useful attributes then copying out the selected usernames. This panel would also be the central spot for taking review-wide actions on specific participants, such as dismissing their (GitHub) review or switching between author and reviewer roles.
+
+A good starting point would be the participants list at the bottom of the main discussion, which this panel will supersede. We should probably also remove the list of waited-on user from the review status display in the checks dropdown panel, as it's not particularly useful there. However, properties that sit at the intersection of users and files (such as who reviewed which file) probably still belong in the file matrix, not here.
+
+Once designated reviewers (https://github.com/Reviewable/Reviewable/issues/918) is implemented, a useful feature in this panel might be to find the minimum required set of reviewers. This would be an advanced version of the sorting / filtering capability suggested above.
+
+The panel may also need to support other features tagged as [feature:people](https://github.com/Reviewable/Reviewable/issues?q=is%3Aopen+is%3Aissue+label%3Afeature%3Apeople).",1.0,"Review participants panel - We should make a dedicated panel that lists all participants in the review with all aspects of their current state, and allows sorting or filtering the list based on useful attributes then copying out the selected usernames. This panel would also be the central spot for taking review-wide actions on specific participants, such as dismissing their (GitHub) review or switching between author and reviewer roles.
+
+A good starting point would be the participants list at the bottom of the main discussion, which this panel will supersede. We should probably also remove the list of waited-on user from the review status display in the checks dropdown panel, as it's not particularly useful there. However, properties that sit at the intersection of users and files (such as who reviewed which file) probably still belong in the file matrix, not here.
+
+Once designated reviewers (https://github.com/Reviewable/Reviewable/issues/918) is implemented, a useful feature in this panel might be to find the minimum required set of reviewers. This would be an advanced version of the sorting / filtering capability suggested above.
+
+The panel may also need to support other features tagged as [feature:people](https://github.com/Reviewable/Reviewable/issues?q=is%3Aopen+is%3Aissue+label%3Afeature%3Apeople).",1,review participants panel we should make a dedicated panel that lists all participants in the review with all aspects of their current state and allows sorting or filtering the list based on useful attributes then copying out the selected usernames this panel would also be the central spot for taking review wide actions on specific participants such as dismissing their github review or switching between author and reviewer roles a good starting point would be the participants list at the bottom of the main discussion which this panel will supersede we should probably also remove the list of waited on user from the review status display in the checks dropdown panel as it s not particularly useful there however properties that sit at the intersection of users and files such as who reviewed which file probably still belong in the file matrix not here once designated reviewers is implemented a useful feature in this panel might be to find the minimum required set of reviewers this would be an advanced version of the sorting filtering capability suggested above the panel may also need to support other features tagged as ,1
+661,11775551076.0,IssuesEvent,2020-03-16 11:33:23,restincode/restincode,https://api.github.com/repos/restincode/restincode,closed,Péter Szőr,Add Person People,"Péter Szőr (17 July 1970 – 12 November 2013) was a computer virus researcher. He was also hacker, and entrepreneur and succesful author. Originally from Hungary, he worked in Finland and the United States.
+
+",1.0,"Péter Szőr - Péter Szőr (17 July 1970 – 12 November 2013) was a computer virus researcher. He was also hacker, and entrepreneur and succesful author. Originally from Hungary, he worked in Finland and the United States.
+
+",1,péter szőr péter szőr july – november was a computer virus researcher he was also hacker and entrepreneur and succesful author originally from hungary he worked in finland and the united states ,1
+61182,6726910494.0,IssuesEvent,2017-10-17 11:43:32,QubesOS/updates-status,https://api.github.com/repos/QubesOS/updates-status,closed,core-agent-linux v3.2.20 (r3.2),r3.2-fc23-testing r3.2-fc24-testing r3.2-fc25-testing r3.2-fc26-testing r3.2-jessie-testing r3.2-stretch-testing,"Update of core-agent-linux to v3.2.20 for Qubes r3.2, see comments below for details.
+
+Built from: https://github.com/QubesOS/qubes-core-agent-linux/commit/f7bf46f780047fbc57638207a9d250bad530cdcb
+
+[Changes since previous version](https://github.com/QubesOS/qubes-core-agent-linux/compare/v3.2.19...v3.2.20):
+QubesOS/qubes-core-agent-linux@f7bf46f version 3.2.20
+QubesOS/qubes-core-agent-linux@60b6f13 dnf-qubes-hooks: handle newer DNF >= 2.x
+
+Referenced issues:
+
+
+If you're release manager, you can issue GPG-inline signed command:
+
+* `Upload core-agent-linux f7bf46f780047fbc57638207a9d250bad530cdcb r3.2 current repo` (available 7 days from now)
+* `Upload core-agent-linux f7bf46f780047fbc57638207a9d250bad530cdcb r3.2 current (dists) repo`, you can choose subset of distributions, like `vm-fc24 vm-fc25` (available 7 days from now)
+* `Upload core-agent-linux f7bf46f780047fbc57638207a9d250bad530cdcb r3.2 security-testing repo`
+
+Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it).
+",6.0,"core-agent-linux v3.2.20 (r3.2) - Update of core-agent-linux to v3.2.20 for Qubes r3.2, see comments below for details.
+
+Built from: https://github.com/QubesOS/qubes-core-agent-linux/commit/f7bf46f780047fbc57638207a9d250bad530cdcb
+
+[Changes since previous version](https://github.com/QubesOS/qubes-core-agent-linux/compare/v3.2.19...v3.2.20):
+QubesOS/qubes-core-agent-linux@f7bf46f version 3.2.20
+QubesOS/qubes-core-agent-linux@60b6f13 dnf-qubes-hooks: handle newer DNF >= 2.x
+
+Referenced issues:
+
+
+If you're release manager, you can issue GPG-inline signed command:
+
+* `Upload core-agent-linux f7bf46f780047fbc57638207a9d250bad530cdcb r3.2 current repo` (available 7 days from now)
+* `Upload core-agent-linux f7bf46f780047fbc57638207a9d250bad530cdcb r3.2 current (dists) repo`, you can choose subset of distributions, like `vm-fc24 vm-fc25` (available 7 days from now)
+* `Upload core-agent-linux f7bf46f780047fbc57638207a9d250bad530cdcb r3.2 security-testing repo`
+
+Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it).
+",0,core agent linux update of core agent linux to for qubes see comments below for details built from qubesos qubes core agent linux version qubesos qubes core agent linux dnf qubes hooks handle newer dnf x referenced issues if you re release manager you can issue gpg inline signed command upload core agent linux current repo available days from now upload core agent linux current dists repo you can choose subset of distributions like vm vm available days from now upload core agent linux security testing repo above commands will work only if packages in current testing repository were built from given commit i e no new version superseded it ,0
+408,7481193841.0,IssuesEvent,2018-04-04 19:51:06,RowdyRuffBoysINC/RRB-client,https://api.github.com/repos/RowdyRuffBoysINC/RRB-client,closed,No components should be directly setting state or keeping their own internal state. Redux store is only source of truth.,May require 2 people Refactor Self Assign,This one may be a 2 person endeavor as its probably the most involved refactor.,1.0,No components should be directly setting state or keeping their own internal state. Redux store is only source of truth. - This one may be a 2 person endeavor as its probably the most involved refactor.,1,no components should be directly setting state or keeping their own internal state redux store is only source of truth this one may be a person endeavor as its probably the most involved refactor ,1
+91486,26400969267.0,IssuesEvent,2023-01-13 01:09:22,opensearch-project/OpenSearch-Dashboards,https://api.github.com/repos/opensearch-project/OpenSearch-Dashboards,closed,[VizBuilder] Increase test coverage,vis builder v2.5.0,"Parent task to fill in the gaps from the low initial coverage that the VisBuilder has.
+
+## Sub tasks:
+
+- [x] Add unit tests to functional code where it makes sense
+- [x] Investigate and document functional testing flow to see if they can be made easier to write and debug
+- [x] Add missing functional test workflows
+- [x] Add Vis Builder FTR tests to OpenSearch Dashboards Repo (set the appropriate env flag)
+
+## Missing unit tests
+
+- [x] `src/plugins/wizard/public/application/utils/get_top_nav_config.tsx`
+- [x] `src/plugins/wizard/public/application/utils/validate_schema_state.ts`
+- [x] `src/plugins/wizard/public/application/components/data_tab/utils/get_available_fields.ts`
+
+## Missing functional tests
+
+- [ ] Save workflow
+- [x] View on dashboards
+- [x] Edit from dashboards
+- [x] Create from dashboards:
+- [x] Edit from Visualization listing
+- [x] Metric tests
+- [x] Bar chart tests
+- [x] Line chart tests
+- [x] Area Chart tests
+- [x] Drag and Drop test
+- [x] Change datasource",1.0,"[VizBuilder] Increase test coverage - Parent task to fill in the gaps from the low initial coverage that the VisBuilder has.
+
+## Sub tasks:
+
+- [x] Add unit tests to functional code where it makes sense
+- [x] Investigate and document functional testing flow to see if they can be made easier to write and debug
+- [x] Add missing functional test workflows
+- [x] Add Vis Builder FTR tests to OpenSearch Dashboards Repo (set the appropriate env flag)
+
+## Missing unit tests
+
+- [x] `src/plugins/wizard/public/application/utils/get_top_nav_config.tsx`
+- [x] `src/plugins/wizard/public/application/utils/validate_schema_state.ts`
+- [x] `src/plugins/wizard/public/application/components/data_tab/utils/get_available_fields.ts`
+
+## Missing functional tests
+
+- [ ] Save workflow
+- [x] View on dashboards
+- [x] Edit from dashboards
+- [x] Create from dashboards:
+- [x] Edit from Visualization listing
+- [x] Metric tests
+- [x] Bar chart tests
+- [x] Line chart tests
+- [x] Area Chart tests
+- [x] Drag and Drop test
+- [x] Change datasource",0, increase test coverage parent task to fill in the gaps from the low initial coverage that the visbuilder has sub tasks add unit tests to functional code where it makes sense investigate and document functional testing flow to see if they can be made easier to write and debug add missing functional test workflows add vis builder ftr tests to opensearch dashboards repo set the appropriate env flag missing unit tests src plugins wizard public application utils get top nav config tsx src plugins wizard public application utils validate schema state ts src plugins wizard public application components data tab utils get available fields ts missing functional tests save workflow view on dashboards edit from dashboards create from dashboards edit from visualization listing metric tests bar chart tests line chart tests area chart tests drag and drop test change datasource,0
+22833,15554117217.0,IssuesEvent,2021-03-16 03:03:33,astropy/photutils,https://api.github.com/repos/astropy/photutils,closed,MNT: Rename default branch from master to main,infrastructure,"Please rename your default branch from `master` to `main` as part of astropy/astropy-project#151 , preferably by 2021-03-22. Also a friendly reminder to check documentation, workflows, etc., and update them accordingly. Please don't forget to communicate this change to your users and stakeholders. To summarize:
+
+- [x] Rename branch from `master` to `main`, preferably by 2021-03-22.
+- [x] Update documentation, workflows, etc., accordingly. -- See #1178
+- [ ] Communicate this change to your users and stakeholders.
+
+Once this is taken care of, you may close this issue.
+
+*This is an automated issue. If this is opened in error, please let @pllim know!*",1.0,"MNT: Rename default branch from master to main - Please rename your default branch from `master` to `main` as part of astropy/astropy-project#151 , preferably by 2021-03-22. Also a friendly reminder to check documentation, workflows, etc., and update them accordingly. Please don't forget to communicate this change to your users and stakeholders. To summarize:
+
+- [x] Rename branch from `master` to `main`, preferably by 2021-03-22.
+- [x] Update documentation, workflows, etc., accordingly. -- See #1178
+- [ ] Communicate this change to your users and stakeholders.
+
+Once this is taken care of, you may close this issue.
+
+*This is an automated issue. If this is opened in error, please let @pllim know!*",0,mnt rename default branch from master to main please rename your default branch from master to main as part of astropy astropy project preferably by also a friendly reminder to check documentation workflows etc and update them accordingly please don t forget to communicate this change to your users and stakeholders to summarize rename branch from master to main preferably by update documentation workflows etc accordingly see communicate this change to your users and stakeholders once this is taken care of you may close this issue this is an automated issue if this is opened in error please let pllim know ,0
+1000,24093593086.0,IssuesEvent,2022-09-19 16:41:11,scikit-image/scikit-image,https://api.github.com/repos/scikit-image/scikit-image,closed,`convex_hull_image` returns different results between 0.18.0 and 0.19.3,:people_hugging: Support :pray: Feature request,"## Description
+
+
+
+Hi guys, my team and I recently decided to migrate to `scikit-image==0.19.3` due to a bigger migration to Python 3.10. Our algorithms rely heavily on `scikit-image` (0.14.2).
+
+During the migration process, we encountered a difference in the `convex_hull_image` function. Seems like the underlying calculation was changed. I found this [issue](https://github.com/scikit-image/scikit-image/issues/3892) ([relevant PR](https://github.com/scikit-image/scikit-image/pull/5029)), and I saw that in the newer versions, the condition that checks whether a point is in a polygon was changed from this:
+
+ Pre 0.18.1
+```
+min <= point < max
+```
+
+ Post 0.18.1
+```
+min <= point <= max
+```
+
+This is a bit problematic for us, because it changes the results of our algorithms pretty dramatically. I tried to look at the code a bit and I saw that `point_in_polygon` now returns 4 possible values:
+```
+0: outside, 1: inside, 2: vertex; 3: edge
+```
+
+However, the function that `convex_hull_image` uses casts the array of numbers into a boolean mask. So for an array like this:
+```
+[0, 1, 2, 3]
+```
+It'd return this:
+```
+[False, True, True, True]
+```
+
+This implicitly includes edge/vertex points inside the polygon. This is a problem because our algorithms were not trained with this kind of shapes.
+
+Is there a way to change this behavior outside of `scikit-image`? I couldn't find any flag that controls this behavior. The most feasible option that we thought about is to fork `scikit-image` and modify the `convex_hull_image` -> `grid_points_in_poly` -> `_grid_points_in_poly` to not cast the labels output to boolean and instead control this behavior with a flag.
+
+However, we prefer not to fork this project and change its internals.
+Any suggestions would be highly appreciated!
+
+## Way to reproduce
+Run `convex_hull_image` in 0.14.2 and in 0.19.3 and witness the results. The mask in 0.19.3 will be slightly larger, due to the inclusion of edge/vertex points.
+
+
+## Version information
+```python
+# Paste the output of the following python commands
+from __future__ import print_function
+import sys; print(sys.version)
+import platform; print(platform.platform())
+import skimage; print(f'scikit-image version: {skimage.__version__}')
+import numpy; print(f'numpy version: {numpy.__version__}')
+```
+
+#### Output
+```python
+3.10.5 (main, Aug 14 2022, 07:33:04) [GCC 9.4.0]
+Linux-5.15.0-1015-aws-x86_64-with-glibc2.31
+scikit-image version: 0.19.3
+numpy version: 1.23.1
+```
+",1.0,"`convex_hull_image` returns different results between 0.18.0 and 0.19.3 - ## Description
+
+
+
+Hi guys, my team and I recently decided to migrate to `scikit-image==0.19.3` due to a bigger migration to Python 3.10. Our algorithms rely heavily on `scikit-image` (0.14.2).
+
+During the migration process, we encountered a difference in the `convex_hull_image` function. Seems like the underlying calculation was changed. I found this [issue](https://github.com/scikit-image/scikit-image/issues/3892) ([relevant PR](https://github.com/scikit-image/scikit-image/pull/5029)), and I saw that in the newer versions, the condition that checks whether a point is in a polygon was changed from this:
+
+ Pre 0.18.1
+```
+min <= point < max
+```
+
+ Post 0.18.1
+```
+min <= point <= max
+```
+
+This is a bit problematic for us, because it changes the results of our algorithms pretty dramatically. I tried to look at the code a bit and I saw that `point_in_polygon` now returns 4 possible values:
+```
+0: outside, 1: inside, 2: vertex; 3: edge
+```
+
+However, the function that `convex_hull_image` uses casts the array of numbers into a boolean mask. So for an array like this:
+```
+[0, 1, 2, 3]
+```
+It'd return this:
+```
+[False, True, True, True]
+```
+
+This implicitly includes edge/vertex points inside the polygon. This is a problem because our algorithms were not trained with this kind of shapes.
+
+Is there a way to change this behavior outside of `scikit-image`? I couldn't find any flag that controls this behavior. The most feasible option that we thought about is to fork `scikit-image` and modify the `convex_hull_image` -> `grid_points_in_poly` -> `_grid_points_in_poly` to not cast the labels output to boolean and instead control this behavior with a flag.
+
+However, we prefer not to fork this project and change its internals.
+Any suggestions would be highly appreciated!
+
+## Way to reproduce
+Run `convex_hull_image` in 0.14.2 and in 0.19.3 and witness the results. The mask in 0.19.3 will be slightly larger, due to the inclusion of edge/vertex points.
+
+
+## Version information
+```python
+# Paste the output of the following python commands
+from __future__ import print_function
+import sys; print(sys.version)
+import platform; print(platform.platform())
+import skimage; print(f'scikit-image version: {skimage.__version__}')
+import numpy; print(f'numpy version: {numpy.__version__}')
+```
+
+#### Output
+```python
+3.10.5 (main, Aug 14 2022, 07:33:04) [GCC 9.4.0]
+Linux-5.15.0-1015-aws-x86_64-with-glibc2.31
+scikit-image version: 0.19.3
+numpy version: 1.23.1
+```
+",1, convex hull image returns different results between and description note for guidance on how to use scikit image please post instead on hi guys my team and i recently decided to migrate to scikit image due to a bigger migration to python our algorithms rely heavily on scikit image during the migration process we encountered a difference in the convex hull image function seems like the underlying calculation was changed i found this and i saw that in the newer versions the condition that checks whether a point is in a polygon was changed from this pre min point max post min point max this is a bit problematic for us because it changes the results of our algorithms pretty dramatically i tried to look at the code a bit and i saw that point in polygon now returns possible values outside inside vertex edge however the function that convex hull image uses casts the array of numbers into a boolean mask so for an array like this it d return this this implicitly includes edge vertex points inside the polygon this is a problem because our algorithms were not trained with this kind of shapes is there a way to change this behavior outside of scikit image i couldn t find any flag that controls this behavior the most feasible option that we thought about is to fork scikit image and modify the convex hull image grid points in poly grid points in poly to not cast the labels output to boolean and instead control this behavior with a flag however we prefer not to fork this project and change its internals any suggestions would be highly appreciated way to reproduce run convex hull image in and in and witness the results the mask in will be slightly larger due to the inclusion of edge vertex points version information python paste the output of the following python commands from future import print function import sys print sys version import platform print platform platform import skimage print f scikit image version skimage version import numpy print f numpy version numpy version output python main aug linux aws with scikit image version numpy version ,1
+274,5329642674.0,IssuesEvent,2017-02-15 15:20:35,PopulateTools/gobierto,https://api.github.com/repos/PopulateTools/gobierto,opened,Structure for statements in Dival,gobierto-people,"ref: http://www.transparencia.gva.es/documents/162282364/162502126/Anexo+III-Registro+Altos+Cargos/91fe6dbd-bb01-4e2f-8eda-520e7a23b8b6;jsessionid=8A90CD9EEF2E77E9A074D9EBAF9065EA
+
+####
+
+Activitats públiques exercides durant els dos anys anteriors a la presa de possessió
+Actividades públicas ejercidas durante los dos años anteriores a la toma de posesión
+Entitat, organisme, empresa o societat
+Entidad, organismo, empresa o sociedad
+Activitat realitzada
+Actividad desempeñada
+Data d'inici
+Fecha de inicio
+Data fi
+Fecha fin
+[IGNORAR?] Feu constar si heu sol·licitat passar a la situació de serveis especials o equivalent, o a la que corresponga
+[IGNORAR?] Haga constar si ha solicitado el pase a la situación de servicios especiales o equivalente, o a la que corresponda
+
+Activitats privades remunerades exercides durant els dos anys anteriors a la presa de possessió
+Actividades privadas remuneradas ejercidas durante los dos años anteriores a la toma de posesión
+Activitat
+Actividad
+Empresa o entitat
+Empresa o entidad
+Autònom / Autònoma
+Autónomo / Autónoma
+Data d'inici
+Fecha de inicio
+Data fi
+Fecha fin
+
+####
+
+Càrrecs públics, participació en òrgans col·legiats i activitats en representació de la generalitat
+Cargos públicos, participación en órganos colegiados y actividades en representación de la generalitat
+Entitat
+Entidad
+Càrrec ocupat
+Cargo desempeñado
+Data de nomenament
+Fecha de nombramiento
+
+Activitats privades remunerades exercides durant els dos anys anteriors a la presa de possessió
+Actividades privadas remuneradas ejercidas durante los dos años anteriores a la toma de posesión
+Activitat
+Actividad
+Entidad
+Entidad
+Autònom / Autònoma
+Autónomo / Autónoma
+Data d'inici
+Fecha de inicio
+Data fi
+Fecha fin
+
+Càrrecs públics, participació en òrgans col·legiats i activitats en representació de la generalitat
+Cargos públicos, participación en órganos colegiados y actividades en representación de la generalitat
+Entitat
+Entidad
+Càrrec ocupat
+Cargo desempeñado
+Data de nomenament
+Fecha de nombramiento
+
+Altres activitats i carrecs no recollits al apartat anterior
+Otras actividades y cargos no recogidas en el apartado anterior
+Entitat
+Entidad
+Càrrec ocupat
+Cargo desempeñado
+Data de nomenament
+Fecha de nombramiento
+
+
+####
+
+Activitat a què va a dedicar-se al cessament
+Actividad a la que va a dedicarse al cese
+Activitat
+Actividad
+Empresa o entitat
+Empresa o entidad
+Autònom / Autònoma
+Autónomo / Autónoma
+Data d'inici
+Fecha de inicio
+
+
+####
+
+Béns immobles urbans i rústics
+Bienes inmuebles urbanos y rústicos
+Clau
+Clave
+Tipus
+Tipo
+% en cas de Domini ple
+% en caso de Dominio pleno
+Província
+Provincia
+Valor cadastral
+Valor catastral
+
+Béns i drets de naturalesa no immobiliària
+Bienes y derechos de naturaleza no inmobiliaria
+Descripció
+Descripción
+Valor (euros)
+Valor (euros)
+
+Passiu
+Pasivo
+Descripció
+Descripción
+Valor (euros)
+Valor (euros)
+
+
+####
+
+Sous y honoraris anuals, aranzels i altres retribucions
+Sueldos y honorarios anuales, aranceles y otras retribuciones
+Concepte
+Concepto
+Euros
+Euros
+
+Dividends i participació en beneficis de societats, comunitats o entitats de qualsevol classe
+Dividendos y participación en beneficios de sociedades, comunidades o entidades de cualquier clase.
+Concepte
+Concepto
+Euros
+Euros
+
+Interessos o rendiments de comptes, dipòsits i actius financers.
+Intereses o rendimientos de cuentas, depósitos y activos financieros.
+Concepte
+Concepto
+Euros
+Euros
+
+Altres rendes o percepcions de qualsevol classe
+Otras rentas o percepciones de cualquier clase.
+Concepte
+Concepto
+Euros
+Euros
+",1.0,"Structure for statements in Dival - ref: http://www.transparencia.gva.es/documents/162282364/162502126/Anexo+III-Registro+Altos+Cargos/91fe6dbd-bb01-4e2f-8eda-520e7a23b8b6;jsessionid=8A90CD9EEF2E77E9A074D9EBAF9065EA
+
+####
+
+Activitats públiques exercides durant els dos anys anteriors a la presa de possessió
+Actividades públicas ejercidas durante los dos años anteriores a la toma de posesión
+Entitat, organisme, empresa o societat
+Entidad, organismo, empresa o sociedad
+Activitat realitzada
+Actividad desempeñada
+Data d'inici
+Fecha de inicio
+Data fi
+Fecha fin
+[IGNORAR?] Feu constar si heu sol·licitat passar a la situació de serveis especials o equivalent, o a la que corresponga
+[IGNORAR?] Haga constar si ha solicitado el pase a la situación de servicios especiales o equivalente, o a la que corresponda
+
+Activitats privades remunerades exercides durant els dos anys anteriors a la presa de possessió
+Actividades privadas remuneradas ejercidas durante los dos años anteriores a la toma de posesión
+Activitat
+Actividad
+Empresa o entitat
+Empresa o entidad
+Autònom / Autònoma
+Autónomo / Autónoma
+Data d'inici
+Fecha de inicio
+Data fi
+Fecha fin
+
+####
+
+Càrrecs públics, participació en òrgans col·legiats i activitats en representació de la generalitat
+Cargos públicos, participación en órganos colegiados y actividades en representación de la generalitat
+Entitat
+Entidad
+Càrrec ocupat
+Cargo desempeñado
+Data de nomenament
+Fecha de nombramiento
+
+Activitats privades remunerades exercides durant els dos anys anteriors a la presa de possessió
+Actividades privadas remuneradas ejercidas durante los dos años anteriores a la toma de posesión
+Activitat
+Actividad
+Entidad
+Entidad
+Autònom / Autònoma
+Autónomo / Autónoma
+Data d'inici
+Fecha de inicio
+Data fi
+Fecha fin
+
+Càrrecs públics, participació en òrgans col·legiats i activitats en representació de la generalitat
+Cargos públicos, participación en órganos colegiados y actividades en representación de la generalitat
+Entitat
+Entidad
+Càrrec ocupat
+Cargo desempeñado
+Data de nomenament
+Fecha de nombramiento
+
+Altres activitats i carrecs no recollits al apartat anterior
+Otras actividades y cargos no recogidas en el apartado anterior
+Entitat
+Entidad
+Càrrec ocupat
+Cargo desempeñado
+Data de nomenament
+Fecha de nombramiento
+
+
+####
+
+Activitat a què va a dedicar-se al cessament
+Actividad a la que va a dedicarse al cese
+Activitat
+Actividad
+Empresa o entitat
+Empresa o entidad
+Autònom / Autònoma
+Autónomo / Autónoma
+Data d'inici
+Fecha de inicio
+
+
+####
+
+Béns immobles urbans i rústics
+Bienes inmuebles urbanos y rústicos
+Clau
+Clave
+Tipus
+Tipo
+% en cas de Domini ple
+% en caso de Dominio pleno
+Província
+Provincia
+Valor cadastral
+Valor catastral
+
+Béns i drets de naturalesa no immobiliària
+Bienes y derechos de naturaleza no inmobiliaria
+Descripció
+Descripción
+Valor (euros)
+Valor (euros)
+
+Passiu
+Pasivo
+Descripció
+Descripción
+Valor (euros)
+Valor (euros)
+
+
+####
+
+Sous y honoraris anuals, aranzels i altres retribucions
+Sueldos y honorarios anuales, aranceles y otras retribuciones
+Concepte
+Concepto
+Euros
+Euros
+
+Dividends i participació en beneficis de societats, comunitats o entitats de qualsevol classe
+Dividendos y participación en beneficios de sociedades, comunidades o entidades de cualquier clase.
+Concepte
+Concepto
+Euros
+Euros
+
+Interessos o rendiments de comptes, dipòsits i actius financers.
+Intereses o rendimientos de cuentas, depósitos y activos financieros.
+Concepte
+Concepto
+Euros
+Euros
+
+Altres rendes o percepcions de qualsevol classe
+Otras rentas o percepciones de cualquier clase.
+Concepte
+Concepto
+Euros
+Euros
+",1,structure for statements in dival ref activitats públiques exercides durant els dos anys anteriors a la presa de possessió actividades públicas ejercidas durante los dos años anteriores a la toma de posesión entitat organisme empresa o societat entidad organismo empresa o sociedad activitat realitzada actividad desempeñada data d inici fecha de inicio data fi fecha fin feu constar si heu sol·licitat passar a la situació de serveis especials o equivalent o a la que corresponga haga constar si ha solicitado el pase a la situación de servicios especiales o equivalente o a la que corresponda activitats privades remunerades exercides durant els dos anys anteriors a la presa de possessió actividades privadas remuneradas ejercidas durante los dos años anteriores a la toma de posesión activitat actividad empresa o entitat empresa o entidad autònom autònoma autónomo autónoma data d inici fecha de inicio data fi fecha fin càrrecs públics participació en òrgans col·legiats i activitats en representació de la generalitat cargos públicos participación en órganos colegiados y actividades en representación de la generalitat entitat entidad càrrec ocupat cargo desempeñado data de nomenament fecha de nombramiento activitats privades remunerades exercides durant els dos anys anteriors a la presa de possessió actividades privadas remuneradas ejercidas durante los dos años anteriores a la toma de posesión activitat actividad entidad entidad autònom autònoma autónomo autónoma data d inici fecha de inicio data fi fecha fin càrrecs públics participació en òrgans col·legiats i activitats en representació de la generalitat cargos públicos participación en órganos colegiados y actividades en representación de la generalitat entitat entidad càrrec ocupat cargo desempeñado data de nomenament fecha de nombramiento altres activitats i carrecs no recollits al apartat anterior otras actividades y cargos no recogidas en el apartado anterior entitat entidad càrrec ocupat cargo desempeñado data de nomenament fecha de nombramiento activitat a què va a dedicar se al cessament actividad a la que va a dedicarse al cese activitat actividad empresa o entitat empresa o entidad autònom autònoma autónomo autónoma data d inici fecha de inicio béns immobles urbans i rústics bienes inmuebles urbanos y rústicos clau clave tipus tipo en cas de domini ple en caso de dominio pleno província provincia valor cadastral valor catastral béns i drets de naturalesa no immobiliària bienes y derechos de naturaleza no inmobiliaria descripció descripción valor euros valor euros passiu pasivo descripció descripción valor euros valor euros sous y honoraris anuals aranzels i altres retribucions sueldos y honorarios anuales aranceles y otras retribuciones concepte concepto euros euros dividends i participació en beneficis de societats comunitats o entitats de qualsevol classe dividendos y participación en beneficios de sociedades comunidades o entidades de cualquier clase concepte concepto euros euros interessos o rendiments de comptes dipòsits i actius financers intereses o rendimientos de cuentas depósitos y activos financieros concepte concepto euros euros altres rendes o percepcions de qualsevol classe otras rentas o percepciones de cualquier clase concepte concepto euros euros ,1
+287679,31856283342.0,IssuesEvent,2023-09-15 07:42:04,Trinadh465/linux-4.1.15_CVE-2023-26607,https://api.github.com/repos/Trinadh465/linux-4.1.15_CVE-2023-26607,opened,CVE-2017-18344 (Medium) detected in linuxlinux-4.6,Mend: dependency security vulnerability,"## CVE-2017-18344 - Medium Severity Vulnerability
+ Vulnerable Library - linuxlinux-4.6
+
+
+The timer_create syscall implementation in kernel/time/posix-timers.c in the Linux kernel before 4.14.8 doesn't properly validate the sigevent->sigev_notify field, which leads to out-of-bounds access in the show_timer function (called when /proc/$PID/timers is read). This allows userspace applications to read arbitrary kernel memory (on a kernel built with CONFIG_POSIX_TIMERS and CONFIG_CHECKPOINT_RESTORE).
+
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",True,"CVE-2017-18344 (Medium) detected in linuxlinux-4.6 - ## CVE-2017-18344 - Medium Severity Vulnerability
+ Vulnerable Library - linuxlinux-4.6
+
+
+The timer_create syscall implementation in kernel/time/posix-timers.c in the Linux kernel before 4.14.8 doesn't properly validate the sigevent->sigev_notify field, which leads to out-of-bounds access in the show_timer function (called when /proc/$PID/timers is read). This allows userspace applications to read arbitrary kernel memory (on a kernel built with CONFIG_POSIX_TIMERS and CONFIG_CHECKPOINT_RESTORE).
+
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",0,cve medium detected in linuxlinux cve medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch main vulnerable source files kernel time posix timers c kernel time posix timers c vulnerability details the timer create syscall implementation in kernel time posix timers c in the linux kernel before doesn t properly validate the sigevent sigev notify field which leads to out of bounds access in the show timer function called when proc pid timers is read this allows userspace applications to read arbitrary kernel memory on a kernel built with config posix timers and config checkpoint restore publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend ,0
+595270,18063308462.0,IssuesEvent,2021-09-20 16:07:02,dotnet/machinelearning-modelbuilder,https://api.github.com/repos/dotnet/machinelearning-modelbuilder,closed,Consume page code not selectable,Priority:0 Stale,"1. I am able to hit the copy button, but I can't select in the box itself to highlight and copy the code
+
+
+
+2. Also, when I hit the copy button, there is no feedback at all that I have copied the code snippet. When I hover, there is a blue highlight which I don't think is supposed to be there? @romalpani
+
+
+
+3. Also, can we add color to the code snippet? Is it possible to hardcode color values since we know exactly what we'll be generating? I think it takes away from the modern design-- there's something just off about it.",1.0,"Consume page code not selectable - 1. I am able to hit the copy button, but I can't select in the box itself to highlight and copy the code
+
+
+
+2. Also, when I hit the copy button, there is no feedback at all that I have copied the code snippet. When I hover, there is a blue highlight which I don't think is supposed to be there? @romalpani
+
+
+
+3. Also, can we add color to the code snippet? Is it possible to hardcode color values since we know exactly what we'll be generating? I think it takes away from the modern design-- there's something just off about it.",0,consume page code not selectable i am able to hit the copy button but i can t select in the box itself to highlight and copy the code also when i hit the copy button there is no feedback at all that i have copied the code snippet when i hover there is a blue highlight which i don t think is supposed to be there romalpani also can we add color to the code snippet is it possible to hardcode color values since we know exactly what we ll be generating i think it takes away from the modern design there s something just off about it ,0
+685678,23465129665.0,IssuesEvent,2022-08-16 16:04:29,linkerd/linkerd2,https://api.github.com/repos/linkerd/linkerd2,closed,multicluster: Restrict access to the proxy's admin server on gateways,priority/P1 area/multicluster area/policy,"Multicluster gateways expose two ports: the proxy's admin server and the proxy's inbound port.
+
+The inbound port explicitly requires authentication (in code), so there's no sense in describing additional authorization policies by default (though users are welcome to restrict access to specific clients).
+
+The proxy's admin server is exposed, however, so that other clusters (service mirrors) can hit the proxy's readiness endpoint to determine its health. This means that, currently, all proxy admin endpoints are exposed from gateways (including `/metrics` and `/env.json`).
+
+We should restrict the proxy admin endpoint so that only authenticated clients can reach it. The simplest approach is probably to ship the gateway with an annotation that sets the default policy to 'all-authenticated' so that the gateway only permits authenticated clients. There are probably other approaches, too...",1.0,"multicluster: Restrict access to the proxy's admin server on gateways - Multicluster gateways expose two ports: the proxy's admin server and the proxy's inbound port.
+
+The inbound port explicitly requires authentication (in code), so there's no sense in describing additional authorization policies by default (though users are welcome to restrict access to specific clients).
+
+The proxy's admin server is exposed, however, so that other clusters (service mirrors) can hit the proxy's readiness endpoint to determine its health. This means that, currently, all proxy admin endpoints are exposed from gateways (including `/metrics` and `/env.json`).
+
+We should restrict the proxy admin endpoint so that only authenticated clients can reach it. The simplest approach is probably to ship the gateway with an annotation that sets the default policy to 'all-authenticated' so that the gateway only permits authenticated clients. There are probably other approaches, too...",0,multicluster restrict access to the proxy s admin server on gateways multicluster gateways expose two ports the proxy s admin server and the proxy s inbound port the inbound port explicitly requires authentication in code so there s no sense in describing additional authorization policies by default though users are welcome to restrict access to specific clients the proxy s admin server is exposed however so that other clusters service mirrors can hit the proxy s readiness endpoint to determine its health this means that currently all proxy admin endpoints are exposed from gateways including metrics and env json we should restrict the proxy admin endpoint so that only authenticated clients can reach it the simplest approach is probably to ship the gateway with an annotation that sets the default policy to all authenticated so that the gateway only permits authenticated clients there are probably other approaches too ,0
+562923,16672724575.0,IssuesEvent,2021-06-07 12:57:24,webcompat/web-bugs,https://api.github.com/repos/webcompat/web-bugs,closed,www.xnxx.com - see bug description,browser-firefox-ios os-ios priority-important,"
+
+
+
+
+**URL**: https://www.xnxx.com/video-ywp8bb1/cute_big_boobed_19-year-old_loves_anal._pussy_munching_butt_fucking_and_atm_cum_facial
+
+**Browser / Version**: Firefox iOS 33.1
+**Operating System**: iOS 14.4.2
+**Tested Another Browser**: Yes Safari
+
+**Problem type**: Something else
+**Description**: This site contains viruses and bots the contain illegal camera control without permission
+**Steps to Reproduce**:
+I visited the site and while watching a couple videos noticed that the screen kept getting bigger and smaller and modifications were being made on my the site that were being shared and mirrored to these users
+
+ View the screenshot
+ Screenshot removed
+
+
+
+Browser Configuration
+
+
None
+
+
+
+_From [webcompat.com](https://webcompat.com/) with ❤️_",1.0,"www.xnxx.com - see bug description -
+
+
+
+
+**URL**: https://www.xnxx.com/video-ywp8bb1/cute_big_boobed_19-year-old_loves_anal._pussy_munching_butt_fucking_and_atm_cum_facial
+
+**Browser / Version**: Firefox iOS 33.1
+**Operating System**: iOS 14.4.2
+**Tested Another Browser**: Yes Safari
+
+**Problem type**: Something else
+**Description**: This site contains viruses and bots the contain illegal camera control without permission
+**Steps to Reproduce**:
+I visited the site and while watching a couple videos noticed that the screen kept getting bigger and smaller and modifications were being made on my the site that were being shared and mirrored to these users
+
+ View the screenshot
+ Screenshot removed
+
+
+
+Browser Configuration
+
+
None
+
+
+
+_From [webcompat.com](https://webcompat.com/) with ❤️_",0, see bug description url browser version firefox ios operating system ios tested another browser yes safari problem type something else description this site contains viruses and bots the contain illegal camera control without permission steps to reproduce i visited the site and while watching a couple videos noticed that the screen kept getting bigger and smaller and modifications were being made on my the site that were being shared and mirrored to these users view the screenshot screenshot removed browser configuration none from with ❤️ ,0
+84876,3681267707.0,IssuesEvent,2016-02-24 02:07:01,coreos/bugs,https://api.github.com/repos/coreos/bugs,closed,segfault in libc on 962.0.0,area/usability component/other kind/regression priority/P0 team/os,"Steps to reproduce:
+```
+git clone https://github.com/coreos/kubernetes
+cd kubernetes
+git checkout coreos-hyperkube-v1.1.7
+
+sudo docker run --rm --net=host \
+-v /usr/bin/docker:/usr/bin/docker \
+-v /lib64:/lib64 \
+-v ${PWD}:/opt/kubernetes \
+quay.io/peanutbutter/debian \
+/bin/bash -c \
+""cd /opt/kubernetes/coreos && make clean && yes | make container TAG=v1.1.7""
+```
+
+Then I see this from journalctl:
+
+```
+Feb 22 20:27:33 pb-build-alpha-2.c.coreos-gce-testing.internal kernel: bash[1447]: segfault at 100000018 ip 00007fa374884bc7 sp 00007fffe9fa72a0 error 4 in libc-2.19.so[7fa374863
+Feb 22 20:27:34 pb-build-alpha-2.c.coreos-gce-testing.internal systemd-coredump[1453]: Process 1447 (bash) of user 0 dumped core.
+```
+
+Same thing happens with the equivalent rkt command:
+
+```
+sudo rkt run \
+--volume opt-kubernetes,kind=host,source=${PWD} \
+--volume bin-docker,kind=host,source=/usr/bin/docker \
+--volume lib64,kind=host,source=/lib64 \
+--mount volume=opt-kubernetes,target=/opt/kubernetes \
+--mount volume=bin-docker,target=/usr/bin/docker \
+--mount volume=lib64,target=/lib64 \
+--trust-keys-from-https \
+--stage1-path=/usr/share/rkt/stage1-fly.aci \
+quay.io/peanutbutter/debian:latest \
+--exec=/bin/bash -- -c \
+""cd /opt/kubernetes/coreos && make clean && yes | make container TAG=v1.1.7""
+```
+
+Trying this on stable v835.13.0 with the docker flavor of this I also get a coredump but the message about libc is missing from the logs.
+
+Sorry this isn't a minimal reproduction, the fault happens before the stuff i'm building even runs, but I was having trouble pairing this down.
+
+For some context on what I'm doing: I'm using a debian image that has build-essential installed to build hyperkube (requires make). This test builds kubernetes and runs some tests inside docker images. So the build also needs host access to the docker client .",1.0,"segfault in libc on 962.0.0 - Steps to reproduce:
+```
+git clone https://github.com/coreos/kubernetes
+cd kubernetes
+git checkout coreos-hyperkube-v1.1.7
+
+sudo docker run --rm --net=host \
+-v /usr/bin/docker:/usr/bin/docker \
+-v /lib64:/lib64 \
+-v ${PWD}:/opt/kubernetes \
+quay.io/peanutbutter/debian \
+/bin/bash -c \
+""cd /opt/kubernetes/coreos && make clean && yes | make container TAG=v1.1.7""
+```
+
+Then I see this from journalctl:
+
+```
+Feb 22 20:27:33 pb-build-alpha-2.c.coreos-gce-testing.internal kernel: bash[1447]: segfault at 100000018 ip 00007fa374884bc7 sp 00007fffe9fa72a0 error 4 in libc-2.19.so[7fa374863
+Feb 22 20:27:34 pb-build-alpha-2.c.coreos-gce-testing.internal systemd-coredump[1453]: Process 1447 (bash) of user 0 dumped core.
+```
+
+Same thing happens with the equivalent rkt command:
+
+```
+sudo rkt run \
+--volume opt-kubernetes,kind=host,source=${PWD} \
+--volume bin-docker,kind=host,source=/usr/bin/docker \
+--volume lib64,kind=host,source=/lib64 \
+--mount volume=opt-kubernetes,target=/opt/kubernetes \
+--mount volume=bin-docker,target=/usr/bin/docker \
+--mount volume=lib64,target=/lib64 \
+--trust-keys-from-https \
+--stage1-path=/usr/share/rkt/stage1-fly.aci \
+quay.io/peanutbutter/debian:latest \
+--exec=/bin/bash -- -c \
+""cd /opt/kubernetes/coreos && make clean && yes | make container TAG=v1.1.7""
+```
+
+Trying this on stable v835.13.0 with the docker flavor of this I also get a coredump but the message about libc is missing from the logs.
+
+Sorry this isn't a minimal reproduction, the fault happens before the stuff i'm building even runs, but I was having trouble pairing this down.
+
+For some context on what I'm doing: I'm using a debian image that has build-essential installed to build hyperkube (requires make). This test builds kubernetes and runs some tests inside docker images. So the build also needs host access to the docker client .",0,segfault in libc on steps to reproduce git clone cd kubernetes git checkout coreos hyperkube sudo docker run rm net host v usr bin docker usr bin docker v v pwd opt kubernetes quay io peanutbutter debian bin bash c cd opt kubernetes coreos make clean yes make container tag then i see this from journalctl feb pb build alpha c coreos gce testing internal kernel bash segfault at ip sp error in libc so feb pb build alpha c coreos gce testing internal systemd coredump process bash of user dumped core same thing happens with the equivalent rkt command sudo rkt run volume opt kubernetes kind host source pwd volume bin docker kind host source usr bin docker volume kind host source mount volume opt kubernetes target opt kubernetes mount volume bin docker target usr bin docker mount volume target trust keys from https path usr share rkt fly aci quay io peanutbutter debian latest exec bin bash c cd opt kubernetes coreos make clean yes make container tag trying this on stable with the docker flavor of this i also get a coredump but the message about libc is missing from the logs sorry this isn t a minimal reproduction the fault happens before the stuff i m building even runs but i was having trouble pairing this down for some context on what i m doing i m using a debian image that has build essential installed to build hyperkube requires make this test builds kubernetes and runs some tests inside docker images so the build also needs host access to the docker client ,0
+557201,16503451994.0,IssuesEvent,2021-05-25 16:28:33,webcompat/web-bugs,https://api.github.com/repos/webcompat/web-bugs,closed,www.beautiful.ai - Firefox does not support Some playback features,browser-firefox engine-gecko priority-normal severity-minor,"
+
+
+
+**URL**: https://www.beautiful.ai/player/-L7ziqaFNjyqjpQZlOO0/Welcome-Packet/23
+
+**Browser / Version**: Firefox 63.0
+**Operating System**: Windows 7
+**Tested Another Browser**: Yes
+
+**Problem type**: Something else
+**Description**: Firefox compatibility notice trigerred
+**Steps to Reproduce**:
+1. Visit the site
+[](https://webcompat.com/uploads/2018/11/376b7c10-4b68-464a-b73b-cf1d2585b1bf.jpg)
+
+
+Browser Configuration
+
+
None
+
+
+
+
+_From [webcompat.com](https://webcompat.com/) with ❤️_",1.0,"www.beautiful.ai - Firefox does not support Some playback features -
+
+
+
+**URL**: https://www.beautiful.ai/player/-L7ziqaFNjyqjpQZlOO0/Welcome-Packet/23
+
+**Browser / Version**: Firefox 63.0
+**Operating System**: Windows 7
+**Tested Another Browser**: Yes
+
+**Problem type**: Something else
+**Description**: Firefox compatibility notice trigerred
+**Steps to Reproduce**:
+1. Visit the site
+[](https://webcompat.com/uploads/2018/11/376b7c10-4b68-464a-b73b-cf1d2585b1bf.jpg)
+
+
+Browser Configuration
+
+
None
+
+
+
+
+_From [webcompat.com](https://webcompat.com/) with ❤️_",0, firefox does not support some playback features url browser version firefox operating system windows tested another browser yes problem type something else description firefox compatibility notice trigerred steps to reproduce visit the site browser configuration none from with ❤️ ,0
+257,5233826819.0,IssuesEvent,2017-01-30 14:06:24,PopulateTools/gobierto,https://api.github.com/repos/PopulateTools/gobierto,opened,Give people an order attribute,gobierto-people,"People should have an order attribute that should be used in the public UI.
+
+Administrators should be able to sort people in the UI.",1.0,"Give people an order attribute - People should have an order attribute that should be used in the public UI.
+
+Administrators should be able to sort people in the UI.",1,give people an order attribute people should have an order attribute that should be used in the public ui administrators should be able to sort people in the ui ,1
+217,4445230808.0,IssuesEvent,2016-08-20 00:10:03,monarch-initiative/monarch-app,https://api.github.com/repos/monarch-initiative/monarch-app,opened,Collapse duplicate PMIDs when pulling from multiple sources,literature and people,"Probably the result of overlapping annotations from mulitple sources but the list of PMIDs has lots of duplicates. Eg here https://monarchinitiative.org/gene/NCBIGene%3A4591
+
+
+
+I don't know which is the best place to uniquify the list of of PMIDs (SciGraph or App) but it should be done. This is low priority.
+",1.0,"Collapse duplicate PMIDs when pulling from multiple sources - Probably the result of overlapping annotations from mulitple sources but the list of PMIDs has lots of duplicates. Eg here https://monarchinitiative.org/gene/NCBIGene%3A4591
+
+
+
+I don't know which is the best place to uniquify the list of of PMIDs (SciGraph or App) but it should be done. This is low priority.
+",1,collapse duplicate pmids when pulling from multiple sources probably the result of overlapping annotations from mulitple sources but the list of pmids has lots of duplicates eg here img width alt screen shot at pm src i don t know which is the best place to uniquify the list of of pmids scigraph or app but it should be done this is low priority ,1
+23739,16550483081.0,IssuesEvent,2021-05-28 08:00:46,google/site-kit-wp,https://api.github.com/repos/google/site-kit-wp,opened,Extend issue templates with config,P2 Type: Infrastructure,"## Feature Description
+
+Currently we have a few issue templates for creating a new issue for a bug or feature request, but GitHub also supports adding additional links via a [configuration file](https://docs.github.com/en/communities/using-templates-to-encourage-useful-issues-and-pull-requests/configuring-issue-templates-for-your-repository).
+
+E.g. https://github.com/ampproject/amp-wp/issues/new/choose
+
+We should leverage this to add useful links in a similar way, as well as maybe add a new template or two.
+
+---------------
+
+_Do not alter or remove anything below. The following sections will be managed by moderators only._
+
+## Acceptance criteria
+
+*
+
+## Implementation Brief
+
+*
+
+### Test Coverage
+
+*
+
+### Visual Regression Changes
+
+*
+
+## QA Brief
+
+*
+
+## Changelog entry
+
+*
+",1.0,"Extend issue templates with config - ## Feature Description
+
+Currently we have a few issue templates for creating a new issue for a bug or feature request, but GitHub also supports adding additional links via a [configuration file](https://docs.github.com/en/communities/using-templates-to-encourage-useful-issues-and-pull-requests/configuring-issue-templates-for-your-repository).
+
+E.g. https://github.com/ampproject/amp-wp/issues/new/choose
+
+We should leverage this to add useful links in a similar way, as well as maybe add a new template or two.
+
+---------------
+
+_Do not alter or remove anything below. The following sections will be managed by moderators only._
+
+## Acceptance criteria
+
+*
+
+## Implementation Brief
+
+*
+
+### Test Coverage
+
+*
+
+### Visual Regression Changes
+
+*
+
+## QA Brief
+
+*
+
+## Changelog entry
+
+*
+",0,extend issue templates with config feature description currently we have a few issue templates for creating a new issue for a bug or feature request but github also supports adding additional links via a e g we should leverage this to add useful links in a similar way as well as maybe add a new template or two do not alter or remove anything below the following sections will be managed by moderators only acceptance criteria implementation brief test coverage visual regression changes qa brief changelog entry ,0
+218027,16919428706.0,IssuesEvent,2021-06-25 01:43:40,lightningnetwork/lnd,https://api.github.com/repos/lightningnetwork/lnd,closed,itests: shutdown failure for on chain fund recovery test,P2 itests regression,"I've seen this pop up somewhat recently in the itests which causes them to take even longer than usual, as the entire test is halted when a node fails to shutdown properly:
+```
+=== RUN TestLightningNetworkDaemon/55-of-79/btcd/channel_backup_restore/restore_unconfirmed_channel_RPC
+ test_harness.go:88: Failed: (onchain fund recovery): exited with error:
+ *errors.errorString unable to shutdown Carol: process did not exit
+ /home/travis/gopath/src/github.com/lightningnetwork/lnd/lntest/itest/lnd_test.go:598 (0xf229c5)
+ shutdownAndAssert: t.Fatalf(""unable to shutdown %v: %v"", node.Name(), err)
+ /home/travis/gopath/src/github.com/lightningnetwork/lnd/lntest/itest/lnd_test.go:971 (0xf881c5)
+ testOnchainFundRecovery.func1: shutdownAndAssert(net, t, node)
+ /home/travis/gopath/src/github.com/lightningnetwork/lnd/lntest/itest/lnd_test.go:1094 (0xf23f50)
+ testOnchainFundRecovery: restoreCheckBalance(finalBalance, 6, 20, promptChangeAddr)
+ /home/travis/gopath/src/github.com/lightningnetwork/lnd/lntest/itest/test_harness.go:112 (0xee9e8e)
+ (*harnessTest).RunTestCase: testCase.test(h.lndHarness, h)
+ /home/travis/gopath/src/github.com/lightningnetwork/lnd/lntest/itest/lnd_test.go:13878 (0xf99fb0)
+ TestLightningNetworkDaemon.func4: ht.RunTestCase(testCase)
+ /home/travis/.gimme/versions/go1.16.3.linux.amd64/src/testing/testing.go:1193 (0x51c72f)
+ tRunner: fn(t)
+ /home/travis/.gimme/versions/go1.16.3.linux.amd64/src/runtime/asm_amd64.s:1371 (0x471c01)
+```
+
+This may be related to the recent changes where we modify our shutdown behavior if we're mid rescan. ",1.0,"itests: shutdown failure for on chain fund recovery test - I've seen this pop up somewhat recently in the itests which causes them to take even longer than usual, as the entire test is halted when a node fails to shutdown properly:
+```
+=== RUN TestLightningNetworkDaemon/55-of-79/btcd/channel_backup_restore/restore_unconfirmed_channel_RPC
+ test_harness.go:88: Failed: (onchain fund recovery): exited with error:
+ *errors.errorString unable to shutdown Carol: process did not exit
+ /home/travis/gopath/src/github.com/lightningnetwork/lnd/lntest/itest/lnd_test.go:598 (0xf229c5)
+ shutdownAndAssert: t.Fatalf(""unable to shutdown %v: %v"", node.Name(), err)
+ /home/travis/gopath/src/github.com/lightningnetwork/lnd/lntest/itest/lnd_test.go:971 (0xf881c5)
+ testOnchainFundRecovery.func1: shutdownAndAssert(net, t, node)
+ /home/travis/gopath/src/github.com/lightningnetwork/lnd/lntest/itest/lnd_test.go:1094 (0xf23f50)
+ testOnchainFundRecovery: restoreCheckBalance(finalBalance, 6, 20, promptChangeAddr)
+ /home/travis/gopath/src/github.com/lightningnetwork/lnd/lntest/itest/test_harness.go:112 (0xee9e8e)
+ (*harnessTest).RunTestCase: testCase.test(h.lndHarness, h)
+ /home/travis/gopath/src/github.com/lightningnetwork/lnd/lntest/itest/lnd_test.go:13878 (0xf99fb0)
+ TestLightningNetworkDaemon.func4: ht.RunTestCase(testCase)
+ /home/travis/.gimme/versions/go1.16.3.linux.amd64/src/testing/testing.go:1193 (0x51c72f)
+ tRunner: fn(t)
+ /home/travis/.gimme/versions/go1.16.3.linux.amd64/src/runtime/asm_amd64.s:1371 (0x471c01)
+```
+
+This may be related to the recent changes where we modify our shutdown behavior if we're mid rescan. ",0,itests shutdown failure for on chain fund recovery test i ve seen this pop up somewhat recently in the itests which causes them to take even longer than usual as the entire test is halted when a node fails to shutdown properly run testlightningnetworkdaemon of btcd channel backup restore restore unconfirmed channel rpc test harness go failed onchain fund recovery exited with error errors errorstring unable to shutdown carol process did not exit home travis gopath src github com lightningnetwork lnd lntest itest lnd test go shutdownandassert t fatalf unable to shutdown v v node name err home travis gopath src github com lightningnetwork lnd lntest itest lnd test go testonchainfundrecovery shutdownandassert net t node home travis gopath src github com lightningnetwork lnd lntest itest lnd test go testonchainfundrecovery restorecheckbalance finalbalance promptchangeaddr home travis gopath src github com lightningnetwork lnd lntest itest test harness go harnesstest runtestcase testcase test h lndharness h home travis gopath src github com lightningnetwork lnd lntest itest lnd test go testlightningnetworkdaemon ht runtestcase testcase home travis gimme versions linux src testing testing go trunner fn t home travis gimme versions linux src runtime asm s this may be related to the recent changes where we modify our shutdown behavior if we re mid rescan ,0
+199,4231627286.0,IssuesEvent,2016-07-04 16:57:41,wordpress-mobile/WordPress-Android,https://api.github.com/repos/wordpress-mobile/WordPress-Android,opened,People: Viewer on a private site can't be invited as a member of the site,People Management [Type] Bug,"### Expected behavior
+
+If a user is a Viewer on a private site, it should be possible to invite them to become a member (Contributor, Author, Editor, or Administrator) on the same site.
+
+This is possible in Calypso web. If a user is a Viewer on a private site, and you try to invite them as a Viewer again, Calypso shows an error:
+
+
+
+But if you select a different role for that Viewer, the invite works:
+
+
+
+
+### Actual behavior
+
+If a user is a Viewer on a private site, you are blocked from inviting them to become a member (Contributor, Author, Editor, or Administrator) on the same site.
+
+
+
+
+### Steps to reproduce the behavior
+
+1. Select a private site and add a user as a Viewer on that site
+2. In the app, go to the site dashboard > People and tap the plus icon to invite a new member
+3. Enter the username or email of the Viewer on the site
+4. Tap in another field or tap Send
+
+Result: An error appears in red: ""There's already a member with the username 'USERNAME'""
+
+##### Tested on Nexus 9, Android 6.0.1, WPAndroid alpha-16",1.0,"People: Viewer on a private site can't be invited as a member of the site - ### Expected behavior
+
+If a user is a Viewer on a private site, it should be possible to invite them to become a member (Contributor, Author, Editor, or Administrator) on the same site.
+
+This is possible in Calypso web. If a user is a Viewer on a private site, and you try to invite them as a Viewer again, Calypso shows an error:
+
+
+
+But if you select a different role for that Viewer, the invite works:
+
+
+
+
+### Actual behavior
+
+If a user is a Viewer on a private site, you are blocked from inviting them to become a member (Contributor, Author, Editor, or Administrator) on the same site.
+
+
+
+
+### Steps to reproduce the behavior
+
+1. Select a private site and add a user as a Viewer on that site
+2. In the app, go to the site dashboard > People and tap the plus icon to invite a new member
+3. Enter the username or email of the Viewer on the site
+4. Tap in another field or tap Send
+
+Result: An error appears in red: ""There's already a member with the username 'USERNAME'""
+
+##### Tested on Nexus 9, Android 6.0.1, WPAndroid alpha-16",1,people viewer on a private site can t be invited as a member of the site expected behavior if a user is a viewer on a private site it should be possible to invite them to become a member contributor author editor or administrator on the same site this is possible in calypso web if a user is a viewer on a private site and you try to invite them as a viewer again calypso shows an error but if you select a different role for that viewer the invite works actual behavior if a user is a viewer on a private site you are blocked from inviting them to become a member contributor author editor or administrator on the same site steps to reproduce the behavior select a private site and add a user as a viewer on that site in the app go to the site dashboard people and tap the plus icon to invite a new member enter the username or email of the viewer on the site tap in another field or tap send result an error appears in red there s already a member with the username username tested on nexus android wpandroid alpha ,1
+631419,20151609434.0,IssuesEvent,2022-02-09 12:57:54,ita-social-projects/horondi_client_fe,https://api.github.com/repos/ita-social-projects/horondi_client_fe,closed,"(SP:1)[Register page] The label ‘by clicking Register, you agree to our Terms’ is not located below ‘Password’ input field.",bug UI priority: medium severity: minor,"OS: Windows 10
+Browser: Google Chrome 90.0.4430.212
+Reproducible: always.
+Build found: last commit [5e32bd7](https://github.com/ita-social-projects/horondi_client_fe/commit/5e32bd772d80792c0e6139d99c9cac788dc3a690)
+
+Steps to reproduce:
+
+1. Go to https://horondi-front-staging.azurewebsites.net/register.
+2. Select English as site language.
+3. Verify that the label ‘by clicking Register, you agree to our Terms’ is located below ‘Password’ input field.
+
+**Actual result**
+The label ‘by clicking Register, you agree to our Terms’ is not located below ‘Password’ input field.
+
+
+
+**Expected result**
+The label ‘by clicking Register, you agree to our Terms’ must be located below ‘Password’ input field.
+
+
+
+Change request [User story #35] UI controls. English L10N. #769
+",1.0,"(SP:1)[Register page] The label ‘by clicking Register, you agree to our Terms’ is not located below ‘Password’ input field. - OS: Windows 10
+Browser: Google Chrome 90.0.4430.212
+Reproducible: always.
+Build found: last commit [5e32bd7](https://github.com/ita-social-projects/horondi_client_fe/commit/5e32bd772d80792c0e6139d99c9cac788dc3a690)
+
+Steps to reproduce:
+
+1. Go to https://horondi-front-staging.azurewebsites.net/register.
+2. Select English as site language.
+3. Verify that the label ‘by clicking Register, you agree to our Terms’ is located below ‘Password’ input field.
+
+**Actual result**
+The label ‘by clicking Register, you agree to our Terms’ is not located below ‘Password’ input field.
+
+
+
+**Expected result**
+The label ‘by clicking Register, you agree to our Terms’ must be located below ‘Password’ input field.
+
+
+
+Change request [User story #35] UI controls. English L10N. #769
+",0, sp the label ‘by clicking register you agree to our terms’ is not located below ‘password’ input field os windows browser google chrome reproducible always build found last commit steps to reproduce go to select english as site language verify that the label ‘by clicking register you agree to our terms’ is located below ‘password’ input field actual result the label ‘by clicking register you agree to our terms’ is not located below ‘password’ input field expected result the label ‘by clicking register you agree to our terms’ must be located below ‘password’ input field change request ui controls english ,0
+204842,15558267715.0,IssuesEvent,2021-03-16 10:05:35,eclipse/che,https://api.github.com/repos/eclipse/che,opened,Stabilize E2E tests related to unexpected fail when start test workspace,area/qe e2e-test/failure kind/task severity/P1 status/in-progress team/che-qe,"### Is your task related to a problem? Please describe.
+Several last launches on Che nightly jobs show unexpected fails of E2E tests related to start test workspace. This problem is observing to 'GitSsh flow', ""OpenshiftConnector', 'Happy Path' ( here on Che single-host)
+
+### Describe the solution you'd like
+We need to investigate a reason of failing E2E test and stabilize it.
+
+### Additional context
+
+
+
+
+",1.0,"Stabilize E2E tests related to unexpected fail when start test workspace - ### Is your task related to a problem? Please describe.
+Several last launches on Che nightly jobs show unexpected fails of E2E tests related to start test workspace. This problem is observing to 'GitSsh flow', ""OpenshiftConnector', 'Happy Path' ( here on Che single-host)
+
+### Describe the solution you'd like
+We need to investigate a reason of failing E2E test and stabilize it.
+
+### Additional context
+
+
+
+
+",0,stabilize tests related to unexpected fail when start test workspace is your task related to a problem please describe several last launches on che nightly jobs show unexpected fails of tests related to start test workspace this problem is observing to gitssh flow openshiftconnector happy path here on che single host describe the solution you d like we need to investigate a reason of failing test and stabilize it additional context ,0
+131823,18250414830.0,IssuesEvent,2021-10-02 05:14:17,tqa236/csharp_exercism,https://api.github.com/repos/tqa236/csharp_exercism,opened,CVE-2019-0548 (High) detected in multiple libraries,security vulnerability,"## CVE-2019-0548 - High Severity Vulnerability
+ Vulnerable Libraries - microsoft.netcore.app.2.1.0.nupkg, microsoft.netcore.dotnethostpolicy.2.1.0.nupkg, microsoft.netcore.dotnethostresolver.2.1.0.nupkg, microsoft.netcore.dotnetapphost.2.1.0.nupkg
+
+
+microsoft.netcore.app.2.1.0.nupkg
+
+
A set of .NET API's that are included in the default .NET Core application model.
+caa7b7e2bad98e56a...
Path to dependency file: csharp_exercism/rna-transcription/RnaTranscription.csproj
+
Path to vulnerable library: /home/wss-scanner/.nuget/packages/microsoft.netcore.dotnethostresolver/2.1.0/microsoft.netcore.dotnethostresolver.2.1.0.nupkg
+
+A denial of service vulnerability exists when ASP.NET Core improperly handles web requests, aka ""ASP.NET Core Denial of Service Vulnerability."" This affects ASP.NET Core 2.2, ASP.NET Core 2.1. This CVE ID is unique from CVE-2019-0564.
+
+
Path to dependency file: csharp_exercism/rna-transcription/RnaTranscription.csproj
+
Path to vulnerable library: /home/wss-scanner/.nuget/packages/microsoft.netcore.dotnethostresolver/2.1.0/microsoft.netcore.dotnethostresolver.2.1.0.nupkg
+
+A denial of service vulnerability exists when ASP.NET Core improperly handles web requests, aka ""ASP.NET Core Denial of Service Vulnerability."" This affects ASP.NET Core 2.2, ASP.NET Core 2.1. This CVE ID is unique from CVE-2019-0564.
+
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",0,cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries microsoft netcore app nupkg microsoft netcore dotnethostpolicy nupkg microsoft netcore dotnethostresolver nupkg microsoft netcore dotnetapphost nupkg microsoft netcore app nupkg a set of net api s that are included in the default net core application model library home page a href path to dependency file csharp exercism rna transcription rnatranscription csproj path to vulnerable library ckages microsoft netcore app microsoft netcore app nupkg dependency hierarchy x microsoft netcore app nupkg vulnerable library microsoft netcore dotnethostpolicy nupkg provides a coreclr hosting policy implementation configuration settings assembly paths and assem library home page a href path to dependency file csharp exercism rna transcription rnatranscription csproj path to vulnerable library home wss scanner nuget packages microsoft netcore dotnethostpolicy microsoft netcore dotnethostpolicy nupkg dependency hierarchy microsoft netcore app nupkg root library x microsoft netcore dotnethostpolicy nupkg vulnerable library microsoft netcore dotnethostresolver nupkg provides an implementation of framework resolution strategy used by microsoft netcore dotnethost ca library home page a href path to dependency file csharp exercism rna transcription rnatranscription csproj path to vulnerable library home wss scanner nuget packages microsoft netcore dotnethostresolver microsoft netcore dotnethostresolver nupkg dependency hierarchy microsoft netcore app nupkg root library microsoft netcore dotnethostpolicy nupkg x microsoft netcore dotnethostresolver nupkg vulnerable library microsoft netcore dotnetapphost nupkg provides the net core app bootstrapper intended for use in the application directory library home page a href path to dependency file csharp exercism rna transcription rnatranscription csproj path to vulnerable library home wss scanner nuget packages microsoft netcore dotnetapphost microsoft netcore dotnetapphost nupkg dependency hierarchy microsoft netcore app nupkg root library microsoft netcore dotnethostpolicy nupkg microsoft netcore dotnethostresolver nupkg x microsoft netcore dotnetapphost nupkg vulnerable library found in head commit a href found in base branch master vulnerability details a denial of service vulnerability exists when asp net core improperly handles web requests aka asp net core denial of service vulnerability this affects asp net core asp net core this cve id is unique from cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution microsoft aspnetcore signalr microsoft aspnetcore server kestrel transport sockets microsoft aspnetcore server iis microsoft aspnetcore server iisintegration microsoft aspnetcore server kestrel core step up your open source security game with whitesource ,0
+105206,16624726019.0,IssuesEvent,2021-06-03 08:08:10,opfab/operatorfabric-core,https://api.github.com/repos/opfab/operatorfabric-core,closed,CVE-2021-20328 (Medium) detected in mongodb-driver-sync-4.0.4.jar - autoclosed,security vulnerability,"## CVE-2021-20328 - Medium Severity Vulnerability
+ Vulnerable Library - mongodb-driver-sync-4.0.4.jar
+
+
+
+Specific versions of the Java driver that support client-side field level encryption (CSFLE) fail to perform correct host name verification on the KMS server’s certificate. This vulnerability in combination with a privileged network position active MITM attack could result in interception of traffic between the Java driver and the KMS service rendering Field Level Encryption ineffective. This issue was discovered during internal testing and affects all versions of the Java driver that support CSFLE. The Java async, Scala, and reactive streams drivers are not impacted. This vulnerability does not impact driver traffic payloads with CSFLE-supported key services originating from applications residing inside the AWS, GCP, and Azure network fabrics due to compensating controls in these environments. This issue does not impact driver workloads that don’t use Field Level Encryption.
+
+
+
+Specific versions of the Java driver that support client-side field level encryption (CSFLE) fail to perform correct host name verification on the KMS server’s certificate. This vulnerability in combination with a privileged network position active MITM attack could result in interception of traffic between the Java driver and the KMS service rendering Field Level Encryption ineffective. This issue was discovered during internal testing and affects all versions of the Java driver that support CSFLE. The Java async, Scala, and reactive streams drivers are not impacted. This vulnerability does not impact driver traffic payloads with CSFLE-supported key services originating from applications residing inside the AWS, GCP, and Azure network fabrics due to compensating controls in these environments. This issue does not impact driver workloads that don’t use Field Level Encryption.
+
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",0,cve medium detected in mongodb driver sync jar autoclosed cve medium severity vulnerability vulnerable library mongodb driver sync jar the mongodb synchronous driver library home page a href path to dependency file operatorfabric core tools spring spring mongo utilities build gradle path to vulnerable library tmp ws ua gwaqkz downloadresource aicikj mongodb driver sync jar dependency hierarchy spring boot starter data mongodb root library x mongodb driver sync jar vulnerable library found in head commit a href found in base branch develop vulnerability details specific versions of the java driver that support client side field level encryption csfle fail to perform correct host name verification on the kms server’s certificate this vulnerability in combination with a privileged network position active mitm attack could result in interception of traffic between the java driver and the kms service rendering field level encryption ineffective this issue was discovered during internal testing and affects all versions of the java driver that support csfle the java async scala and reactive streams drivers are not impacted this vulnerability does not impact driver traffic payloads with csfle supported key services originating from applications residing inside the aws gcp and azure network fabrics due to compensating controls in these environments this issue does not impact driver workloads that don’t use field level encryption publish date url a href cvss score details base score metrics exploitability metrics attack vector adjacent attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org mongodb mongodb driver sync org mongodb mongodb driver legacy org mongodb mongodb driver org mongodb mongo java driver step up your open source security game with whitesource ,0
+345296,24852255444.0,IssuesEvent,2022-10-26 21:13:28,pypa/pipenv,https://api.github.com/repos/pypa/pipenv,closed,Adding Table of Contents To Readme File,Type: Documentation :book:,Adding Table of Contents to Readme file will increase the accessibility and readability of the document.,1.0,Adding Table of Contents To Readme File - Adding Table of Contents to Readme file will increase the accessibility and readability of the document.,0,adding table of contents to readme file adding table of contents to readme file will increase the accessibility and readability of the document ,0
+132824,28364383221.0,IssuesEvent,2023-04-12 13:00:55,apache/daffodil-vscode,https://api.github.com/repos/apache/daffodil-vscode,closed,Nested element and sequence tags return incorrect element completion items,bug code completion,"When an element tag is nested in another element tag, and the cursor is placed before the closing tag of the initial element tag, and control space is press, the incorrect items are suggested. The logic in-correctly assumes the closing tag for the nested (inner) element tag is the closing tag for the outer element. The same situation occurs with nested sequence tags.",1.0,"Nested element and sequence tags return incorrect element completion items - When an element tag is nested in another element tag, and the cursor is placed before the closing tag of the initial element tag, and control space is press, the incorrect items are suggested. The logic in-correctly assumes the closing tag for the nested (inner) element tag is the closing tag for the outer element. The same situation occurs with nested sequence tags.",0,nested element and sequence tags return incorrect element completion items when an element tag is nested in another element tag and the cursor is placed before the closing tag of the initial element tag and control space is press the incorrect items are suggested the logic in correctly assumes the closing tag for the nested inner element tag is the closing tag for the outer element the same situation occurs with nested sequence tags ,0
+348355,24912356128.0,IssuesEvent,2022-10-30 01:40:32,AY2223S1-CS2103-F13-3/tp,https://api.github.com/repos/AY2223S1-CS2103-F13-3/tp,closed,[PE-D][Tester A] findmeetings feature not working properly,DocumentationBug In the works,"User guide says that the AFTER_DATE should come first followed by the BEFORE_DATE.
+
+Here I have entered the dates of the incorrect format but the error thrown is simply Unknown Command. Perhaps a more insightful error can be thrown if I have typed the incorrect format of the dates.
+
+Furthermore I have typed out the command as specified in the example, yet I still get an error for some reason. Perhaps theres some bug with the feature itself?
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+-------------
+Labels: `severity.High` `type.FunctionalityBug`
+original: JordanChua/ped#7",1.0,"[PE-D][Tester A] findmeetings feature not working properly - User guide says that the AFTER_DATE should come first followed by the BEFORE_DATE.
+
+Here I have entered the dates of the incorrect format but the error thrown is simply Unknown Command. Perhaps a more insightful error can be thrown if I have typed the incorrect format of the dates.
+
+Furthermore I have typed out the command as specified in the example, yet I still get an error for some reason. Perhaps theres some bug with the feature itself?
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+-------------
+Labels: `severity.High` `type.FunctionalityBug`
+original: JordanChua/ped#7",0, findmeetings feature not working properly user guide says that the after date should come first followed by the before date here i have entered the dates of the incorrect format but the error thrown is simply unknown command perhaps a more insightful error can be thrown if i have typed the incorrect format of the dates furthermore i have typed out the command as specified in the example yet i still get an error for some reason perhaps theres some bug with the feature itself labels severity high type functionalitybug original jordanchua ped ,0
+149145,19566137717.0,IssuesEvent,2022-01-04 00:48:05,dotnet/runtime,https://api.github.com/repos/dotnet/runtime,opened,LocalCertificateSelectionCallback passed to LocalCertificateSelectionCallback is always null on Windows,area-System.Net.Security os-windows,"I bump to it while working on test for #52499. While the trusted CA list _is_ populated on Windows the `emoteCertificate` is not.
+We try but the call we use always fails with `SEC_E_INVALID_HANDLE`. I confirmed with experts that the call cannot be used until handshake is done. We use same method and credentials to get the certificate after completed handshake and it works as expected. e.g. there is nothing wrong with the method or the credential handle.
+
+It seems like there may be way how to get the certificate but it will need some exploration and more testing.
+
+",True,"LocalCertificateSelectionCallback passed to LocalCertificateSelectionCallback is always null on Windows - I bump to it while working on test for #52499. While the trusted CA list _is_ populated on Windows the `emoteCertificate` is not.
+We try but the call we use always fails with `SEC_E_INVALID_HANDLE`. I confirmed with experts that the call cannot be used until handshake is done. We use same method and credentials to get the certificate after completed handshake and it works as expected. e.g. there is nothing wrong with the method or the credential handle.
+
+It seems like there may be way how to get the certificate but it will need some exploration and more testing.
+
+",0,localcertificateselectioncallback passed to localcertificateselectioncallback is always null on windows i bump to it while working on test for while the trusted ca list is populated on windows the emotecertificate is not we try but the call we use always fails with sec e invalid handle i confirmed with experts that the call cannot be used until handshake is done we use same method and credentials to get the certificate after completed handshake and it works as expected e g there is nothing wrong with the method or the credential handle it seems like there may be way how to get the certificate but it will need some exploration and more testing ,0
+264927,20049095903.0,IssuesEvent,2022-02-03 02:34:36,OrchardCMS/OrchardCore,https://api.github.com/repos/OrchardCMS/OrchardCore,reopened,Current widgets / Types do Not support Metrics / Graphs,documentation,"### Which component Strategy for new developers & Students to build metrics based components
+
+There's some confusion in the Developers & students community. they do not know, **which path to take to build transactional components/functionality**, for e.g. if a developer wants to create a component with a` master-details table` for customer shipping/orders tables Or, **GRAPH** for e.g. Number of `returns per product D3 / eCharts Graph`, and a table below with details on selected row-item in the graph... which one / do they use a module, do they use a widget, or content type its very fuzzy on which one to pick??
+
+Next, if a developer wants to scaffold Orchard Form, or custom Widgets , module, for e.g. in the last .net dev meetup many mentioned, they don't know how to create a simple pie chart in orchard for e.g. there are several use cases like a customer returns per product.
+
+I posted a help request in the Lombiq CVS2022 extension, and proposed using Extension, they gave up after trying, first because of the compatibility issue, which is being fixed, but also some functionality issue. So, if it has more comments/picture on what the code structure/components its scaffolding, step 1, 2 etc. in comments form then developers would be able to use it to easily create Module or Widget.
+
+They key thing to understand is the flexibility in your scaffolding / visual extension is perhaps a good entry point as a visual tool in VS to help them create the components they want.
+
+If you notice in the admin dashboard there is no single graph there?
+
+Can you create a core component, with a graph and table to show number of failed attempts, IP aaddress (kinda like WordFence or loginizer in WP), this would be a good reference app. ",1.0,"Current widgets / Types do Not support Metrics / Graphs - ### Which component Strategy for new developers & Students to build metrics based components
+
+There's some confusion in the Developers & students community. they do not know, **which path to take to build transactional components/functionality**, for e.g. if a developer wants to create a component with a` master-details table` for customer shipping/orders tables Or, **GRAPH** for e.g. Number of `returns per product D3 / eCharts Graph`, and a table below with details on selected row-item in the graph... which one / do they use a module, do they use a widget, or content type its very fuzzy on which one to pick??
+
+Next, if a developer wants to scaffold Orchard Form, or custom Widgets , module, for e.g. in the last .net dev meetup many mentioned, they don't know how to create a simple pie chart in orchard for e.g. there are several use cases like a customer returns per product.
+
+I posted a help request in the Lombiq CVS2022 extension, and proposed using Extension, they gave up after trying, first because of the compatibility issue, which is being fixed, but also some functionality issue. So, if it has more comments/picture on what the code structure/components its scaffolding, step 1, 2 etc. in comments form then developers would be able to use it to easily create Module or Widget.
+
+They key thing to understand is the flexibility in your scaffolding / visual extension is perhaps a good entry point as a visual tool in VS to help them create the components they want.
+
+If you notice in the admin dashboard there is no single graph there?
+
+Can you create a core component, with a graph and table to show number of failed attempts, IP aaddress (kinda like WordFence or loginizer in WP), this would be a good reference app. ",0,current widgets types do not support metrics graphs which component strategy for new developers students to build metrics based components there s some confusion in the developers students community they do not know which path to take to build transactional components functionality for e g if a developer wants to create a component with a master details table for customer shipping orders tables or graph for e g number of returns per product echarts graph and a table below with details on selected row item in the graph which one do they use a module do they use a widget or content type its very fuzzy on which one to pick next if a developer wants to scaffold orchard form or custom widgets module for e g in the last net dev meetup many mentioned they don t know how to create a simple pie chart in orchard for e g there are several use cases like a customer returns per product i posted a help request in the lombiq extension and proposed using extension they gave up after trying first because of the compatibility issue which is being fixed but also some functionality issue so if it has more comments picture on what the code structure components its scaffolding step etc in comments form then developers would be able to use it to easily create module or widget they key thing to understand is the flexibility in your scaffolding visual extension is perhaps a good entry point as a visual tool in vs to help them create the components they want if you notice in the admin dashboard there is no single graph there can you create a core component with a graph and table to show number of failed attempts ip aaddress kinda like wordfence or loginizer in wp this would be a good reference app ,0
+3,2534255072.0,IssuesEvent,2015-01-24 19:27:28,ufvivotech/ufDataQualityImprovement,https://api.github.com/repos/ufvivotech/ufDataQualityImprovement,closed,VIVO Data Project,CTSI People,"Work with Tom Pearson. Money from SP. Hire students. Assign project manager.
+
+Goal: Update all HSC faculty and faculty on NIH grants across UF.
+
+Educational background, overview, awards, previous positions/grants/papers. Suitable for biosketch.
+
+This is a one time effort leading to self maintenance.",1.0,"VIVO Data Project - Work with Tom Pearson. Money from SP. Hire students. Assign project manager.
+
+Goal: Update all HSC faculty and faculty on NIH grants across UF.
+
+Educational background, overview, awards, previous positions/grants/papers. Suitable for biosketch.
+
+This is a one time effort leading to self maintenance.",1,vivo data project work with tom pearson money from sp hire students assign project manager goal update all hsc faculty and faculty on nih grants across uf educational background overview awards previous positions grants papers suitable for biosketch this is a one time effort leading to self maintenance ,1
+352,6673432014.0,IssuesEvent,2017-10-04 15:03:12,PopulateTools/gobierto,https://api.github.com/repos/PopulateTools/gobierto,opened,Allow an admin user to edit only specific People,gobierto-admin gobierto-people,"- [ ] Markup
+- [ ] Implementation (define tasks)",1.0,"Allow an admin user to edit only specific People - - [ ] Markup
+- [ ] Implementation (define tasks)",1,allow an admin user to edit only specific people markup implementation define tasks ,1
+301,5629470166.0,IssuesEvent,2017-04-05 09:33:43,wordpress-mobile/WordPress-iOS,https://api.github.com/repos/wordpress-mobile/WordPress-iOS,opened,People: Add filter for email followers,People Management [Pri] Low [Type] Enhancement,"### Expected behavior
+
+In the People section, I expect to see Users (members of the site), Followers (WordPress.com followers), and Email Followers (non-WordPress.com email subscribers), similar to what appears at https://wordpress.com/people in Calypso.
+
+### Actual behavior
+
+The app only has sections for Users and Followers. Email followers only appear in the Followers module in Stats, not in the People section.
+
+##### Tested on iPhone 6, iOS 10.3, WPiOS 7.3.0.20170327",1.0,"People: Add filter for email followers - ### Expected behavior
+
+In the People section, I expect to see Users (members of the site), Followers (WordPress.com followers), and Email Followers (non-WordPress.com email subscribers), similar to what appears at https://wordpress.com/people in Calypso.
+
+### Actual behavior
+
+The app only has sections for Users and Followers. Email followers only appear in the Followers module in Stats, not in the People section.
+
+##### Tested on iPhone 6, iOS 10.3, WPiOS 7.3.0.20170327",1,people add filter for email followers expected behavior in the people section i expect to see users members of the site followers wordpress com followers and email followers non wordpress com email subscribers similar to what appears at in calypso actual behavior the app only has sections for users and followers email followers only appear in the followers module in stats not in the people section tested on iphone ios wpios ,1
+295478,22218204290.0,IssuesEvent,2022-06-08 05:26:23,OpenMined/PySyft,https://api.github.com/repos/OpenMined/PySyft,closed,Audit vertical-learning tutorials,Good first issue :mortar_board: Type: Documentation :books: Documentation: docstring,"## Context
+`docs_team` is trying to push some good practices about documentation (inline docstring, tutorials, examples and Readmes)
+
+one of the tasks that is not done is to check tutorials in order to know what of them work without problem and what of them need to be updated.
+
+## Output
+The output of this issue is just the status:
+* Bad/Red means that does not work at all.
+* Needs to updated/yellow means that the tutorial works but needs some clarifications.
+* Good/Green means that the tutorials works as expected.
+
+This is a great oportunity to get involved into pysyft!",2.0,"Audit vertical-learning tutorials - ## Context
+`docs_team` is trying to push some good practices about documentation (inline docstring, tutorials, examples and Readmes)
+
+one of the tasks that is not done is to check tutorials in order to know what of them work without problem and what of them need to be updated.
+
+## Output
+The output of this issue is just the status:
+* Bad/Red means that does not work at all.
+* Needs to updated/yellow means that the tutorial works but needs some clarifications.
+* Good/Green means that the tutorials works as expected.
+
+This is a great oportunity to get involved into pysyft!",0,audit vertical learning tutorials context docs team is trying to push some good practices about documentation inline docstring tutorials examples and readmes one of the tasks that is not done is to check tutorials in order to know what of them work without problem and what of them need to be updated output the output of this issue is just the status bad red means that does not work at all needs to updated yellow means that the tutorial works but needs some clarifications good green means that the tutorials works as expected this is a great oportunity to get involved into pysyft ,0
+372385,11013572624.0,IssuesEvent,2019-12-04 20:46:10,acidanthera/bugtracker,https://api.github.com/repos/acidanthera/bugtracker,closed,10.15 beta 4 AppleIntelMCEReporter,bug priority:normal project:cpuf,"Is there away we can get a new patch added to Lilu to patch AppleIntelMCEReporter.kext. Apple has introduced a new controller inside the driver called AppleIntelMCEInteruptController which has included the iMacPro1,1, MacPro6,1 and assume the new MacPro7,1 board-ids. This is causing kernel panic for Ivy Bridge-EP CPUs.
+10.15 beta 3 AppleIntelMCEReporter.kext Info.plist
+https://imgur.com/S9De4FP
+10.15 beta 4 AppleIntelMCEReporter.kext Ino.plist
+https://imgur.com/ACDbmQf",1.0,"10.15 beta 4 AppleIntelMCEReporter - Is there away we can get a new patch added to Lilu to patch AppleIntelMCEReporter.kext. Apple has introduced a new controller inside the driver called AppleIntelMCEInteruptController which has included the iMacPro1,1, MacPro6,1 and assume the new MacPro7,1 board-ids. This is causing kernel panic for Ivy Bridge-EP CPUs.
+10.15 beta 3 AppleIntelMCEReporter.kext Info.plist
+https://imgur.com/S9De4FP
+10.15 beta 4 AppleIntelMCEReporter.kext Ino.plist
+https://imgur.com/ACDbmQf",0, beta appleintelmcereporter is there away we can get a new patch added to lilu to patch appleintelmcereporter kext apple has introduced a new controller inside the driver called appleintelmceinteruptcontroller which has included the and assume the new board ids this is causing kernel panic for ivy bridge ep cpus beta appleintelmcereporter kext info plist beta appleintelmcereporter kext ino plist ,0
+40086,9841894983.0,IssuesEvent,2019-06-18 08:02:51,zotonic/zotonic,https://api.github.com/repos/zotonic/zotonic,closed,Create ACL role button no longer works,admin-ui core defect,"Clicking the ""Create ACL Role"" button does nothing.
+
+
+
+```erlang
+2019-06-05 11:25:39.034 [error] <0.15866.135>@z_transport:incoming_msgs:202 Stack: [{erlang,list_to_integer,[""acl_role""],[]},{z_convert,to_integer,1,[{file,""src/z_convert.erl""},{line,120}]},{action_admin_dialog_new_rsc,maybe_add_prop,3,[{file,""modules/mod_admin/actions/action_admin_dialog_new_rsc.erl""},{line,228}]},{lists,foldl,3,[{file,""lists.erl""},{line,1263}]},{action_admin_dialog_new_rsc,get_base_props,2,[{file,""modules/mod_admin/actions/action_admin_dialog_new_rsc.erl""},{line,208}]},{action_admin_dialog_new_rsc,event,2,[{file,""modules/mod_admin/actions/action_admin_dialog_new_rsc.erl""},{line,94}]},{z_transport,incoming_with_session,2,[{file,""src/support/z_transport.erl""},{line,252}]},{z_transport,incoming_msgs,2,[{file,""src/support/z_transport.erl""},{line,193}]}]
+```
+
+I'm not sure if it is a bug in `mod_acl_simple_roles` or in the implementation of the new dialog which adds resources. It could be that `m_rsc:rid/2` is needed instead of `list_to_integer`",1.0,"Create ACL role button no longer works - Clicking the ""Create ACL Role"" button does nothing.
+
+
+
+```erlang
+2019-06-05 11:25:39.034 [error] <0.15866.135>@z_transport:incoming_msgs:202 Stack: [{erlang,list_to_integer,[""acl_role""],[]},{z_convert,to_integer,1,[{file,""src/z_convert.erl""},{line,120}]},{action_admin_dialog_new_rsc,maybe_add_prop,3,[{file,""modules/mod_admin/actions/action_admin_dialog_new_rsc.erl""},{line,228}]},{lists,foldl,3,[{file,""lists.erl""},{line,1263}]},{action_admin_dialog_new_rsc,get_base_props,2,[{file,""modules/mod_admin/actions/action_admin_dialog_new_rsc.erl""},{line,208}]},{action_admin_dialog_new_rsc,event,2,[{file,""modules/mod_admin/actions/action_admin_dialog_new_rsc.erl""},{line,94}]},{z_transport,incoming_with_session,2,[{file,""src/support/z_transport.erl""},{line,252}]},{z_transport,incoming_msgs,2,[{file,""src/support/z_transport.erl""},{line,193}]}]
+```
+
+I'm not sure if it is a bug in `mod_acl_simple_roles` or in the implementation of the new dialog which adds resources. It could be that `m_rsc:rid/2` is needed instead of `list_to_integer`",0,create acl role button no longer works clicking the create acl role button does nothing erlang z transport incoming msgs stack z convert to integer action admin dialog new rsc maybe add prop lists foldl action admin dialog new rsc get base props action admin dialog new rsc event z transport incoming with session z transport incoming msgs i m not sure if it is a bug in mod acl simple roles or in the implementation of the new dialog which adds resources it could be that m rsc rid is needed instead of list to integer ,0
+1020,24968679611.0,IssuesEvent,2022-11-01 21:58:50,omigroup/omigroup,https://api.github.com/repos/omigroup/omigroup,closed,Let's talk about virtual World location exif data for screenshots captured inside of Virtual Worlds.,enhancement Make the metaverse more human Consistently deliver value Empower the people making the metaverse Maintain sustainable innovation,"Let's also discuss the Privacy aspects: like how do we keep that from being used maliciously?
+
+Is this a project worth pursuing?
+
+https://www.exif.org/Exif2-2.PDF",1.0,"Let's talk about virtual World location exif data for screenshots captured inside of Virtual Worlds. - Let's also discuss the Privacy aspects: like how do we keep that from being used maliciously?
+
+Is this a project worth pursuing?
+
+https://www.exif.org/Exif2-2.PDF",1,let s talk about virtual world location exif data for screenshots captured inside of virtual worlds let s also discuss the privacy aspects like how do we keep that from being used maliciously is this a project worth pursuing ,1
+366,6801220171.0,IssuesEvent,2017-11-02 16:10:11,planningcenter/developers,https://api.github.com/repos/planningcenter/developers,closed,People workflows - snooze endpoint,People question,"### Detailed Description of the Problem/Question
+I'm trying to snooze a card in a workflow. It's unclear to me by the documentation if I should specify a duration or if it will use some kind of default. Either way, whether I pass data to the POST or not, it will not snooze the card.
+
+Could you please help me understand how to use the snooze endpoint or determine if it is not working properly? Thank you!
+
+##### Steps to reproduce:
+Here the code I'm using:
+_PCO_BASE_URL = 'https://api.planningcenteronline.com/'
+_APPLICATION_ID = os.environ.get('PLANNINGCENTERAPPLICATIONID')
+_SECRET = os.environ.get('PLANNINGCENTERSECRET')
+
+url = '{api_url}people/v2/workflows/47512/cards/1017083/snooze'.format(api_url=_PCO_BASE_URL)
+
+d = {""data"": {""type"": ""WorkflowCard"", ""attributes"": {""snooze_until"": ""2017-11-25T12:00:00Z""}}}
+
+r = requests.post(url, data=d, auth=(_APPLICATION_ID, _SECRET))
+results = r.json()
+print(results)
+
+
+##### API endpoint I'm using:
+https://api.planningcenteronline.com/people/v2/workflows/1/cards/1/snooze
+
+
+##### Programming language I'm using:
+Python
+
+
+##### Authentication method I'm using:
+
+e.g. OAuth 2, Personal Access Token, Browser Session (Cookie)
+",1.0,"People workflows - snooze endpoint - ### Detailed Description of the Problem/Question
+I'm trying to snooze a card in a workflow. It's unclear to me by the documentation if I should specify a duration or if it will use some kind of default. Either way, whether I pass data to the POST or not, it will not snooze the card.
+
+Could you please help me understand how to use the snooze endpoint or determine if it is not working properly? Thank you!
+
+##### Steps to reproduce:
+Here the code I'm using:
+_PCO_BASE_URL = 'https://api.planningcenteronline.com/'
+_APPLICATION_ID = os.environ.get('PLANNINGCENTERAPPLICATIONID')
+_SECRET = os.environ.get('PLANNINGCENTERSECRET')
+
+url = '{api_url}people/v2/workflows/47512/cards/1017083/snooze'.format(api_url=_PCO_BASE_URL)
+
+d = {""data"": {""type"": ""WorkflowCard"", ""attributes"": {""snooze_until"": ""2017-11-25T12:00:00Z""}}}
+
+r = requests.post(url, data=d, auth=(_APPLICATION_ID, _SECRET))
+results = r.json()
+print(results)
+
+
+##### API endpoint I'm using:
+https://api.planningcenteronline.com/people/v2/workflows/1/cards/1/snooze
+
+
+##### Programming language I'm using:
+Python
+
+
+##### Authentication method I'm using:
+
+e.g. OAuth 2, Personal Access Token, Browser Session (Cookie)
+",1,people workflows snooze endpoint detailed description of the problem question i m trying to snooze a card in a workflow it s unclear to me by the documentation if i should specify a duration or if it will use some kind of default either way whether i pass data to the post or not it will not snooze the card could you please help me understand how to use the snooze endpoint or determine if it is not working properly thank you steps to reproduce here the code i m using pco base url application id os environ get planningcenterapplicationid secret os environ get planningcentersecret url api url people workflows cards snooze format api url pco base url d data type workflowcard attributes snooze until r requests post url data d auth application id secret results r json print results api endpoint i m using programming language i m using python authentication method i m using e g oauth personal access token browser session cookie ,1
+68298,7092353603.0,IssuesEvent,2018-01-12 16:16:27,openshift/origin,https://api.github.com/repos/openshift/origin,closed,flake: jenkins-plugin test imagestream SCM DSL,area/tests kind/test-flake priority/P1,"```/go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/image_ecosystem/jenkins_plugin.go:668
+Expected error:
+ <*errors.errorString | 0xc420828030>: {
+ s: ""timed out while waiting of an image stream tag extended-test-jenkins-plugin-6r9td-fkh6n/localjenkins:develop"",
+ }
+ timed out while waiting of an image stream tag extended-test-jenkins-plugin-6r9td-fkh6n/localjenkins:develop
+not to have occurred
+/go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/image_ecosystem/jenkins_plugin.go:659",2.0,"flake: jenkins-plugin test imagestream SCM DSL - ```/go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/image_ecosystem/jenkins_plugin.go:668
+Expected error:
+ <*errors.errorString | 0xc420828030>: {
+ s: ""timed out while waiting of an image stream tag extended-test-jenkins-plugin-6r9td-fkh6n/localjenkins:develop"",
+ }
+ timed out while waiting of an image stream tag extended-test-jenkins-plugin-6r9td-fkh6n/localjenkins:develop
+not to have occurred
+/go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/image_ecosystem/jenkins_plugin.go:659",0,flake jenkins plugin test imagestream scm dsl go src github com openshift origin output local go src github com openshift origin test extended image ecosystem jenkins plugin go expected error s timed out while waiting of an image stream tag extended test jenkins plugin localjenkins develop timed out while waiting of an image stream tag extended test jenkins plugin localjenkins develop not to have occurred go src github com openshift origin output local go src github com openshift origin test extended image ecosystem jenkins plugin go ,0
+406,7465738164.0,IssuesEvent,2018-04-02 06:49:04,awesome-jobs/vietnam,https://api.github.com/repos/awesome-jobs/vietnam,closed,LARION - Senior BA cum Scrum Master - Saigon - FT,Full-time People Skill Saigon,"## Location
+* Hall 5, Quang Trung Software City, Tan Chanh Hiep Ward, District 12, HCMC
+
+## Salary Expectation
+* $1200 ~ $1500
+
+## Requirements
+* At least 4 years of business analysis related experience in software development projects
+* At least 2 years of working experience in Scrum Master or Project Manager
+* Strong communications, mentoring, and negotiation skills with stakeholders
+* Use of English proficiency (Read and understand specialized documents)
+* Excellent knowledge at software engineering and SDLC
+* Strong at the tools, techniques, processes required to capture requirements at generic and detailed levels. Those are for example: UML, OpenOffice, Mindmap, Trello, Basecamp, Pivotal Tracker, Mavenlink, Jira, Mantis, Redmine
+
+## Skills
+* Business Analyst
+* Scrum Master
+
+### Must Have
+* Have a deep know-how (at expert level) on at least 1 business domain (e.g. Healthcare, Finance, Security, Insurance...) for at least 3 continuous years in projects
+* Ability to write documents URD, SRS
+* Able to perform check (test) the business requirements
+
+### Responsibilities
+* Gather business and system requirements from client
+* Assist client to define requirements, suggest ideas, and provide solutions
+* Communicate with customers to clarify any unclear information about the business requirements
+* Analyze and specify requirements, synthesize information provided by relevant stakeholders
+* Design application mock-up / wire-frame using prototyping tools and model the requirements by using charting tools
+* Work with Quality Controller (QC) to verify that all system requirements, including change requests, are implemented and traceable
+* Take main BA role in big/complex projects (team size > 15)
+
+
+### Why it would be awesome to work with us
+* Attractive salary and benefits (13th salary, distinguised employee of the quater and year, seniority award...)
+* Performance appraisal twice a year
+* Healthcare and accident insurance
+* Various training on best practices and soft skills
+* Teambuilding activities in every summer, company trip, big annual year-end party every year, etc.
+Fitness & sports activities: football, tennis, table-tennis, badminton...
+* Commitment to community development: charity every quarter, blood donation, public seminars, career orientation talks…
+* Free in-house entertainment facilities (foodball, table tennis, boxing, coffee...)
+* Support for personal loans such as: home loan, vehicle loan, tuition fee...
+
+### Interesting Engineering Projects
+Our wide range of services includes:
+* Ruby on Rails Web Application Development
+* Big Data – Data Analytics especially in Healthcare
+* Supply Chain Management Application Development
+* Social E-commerce App Development
+* Securities Trading Solutions Development
+* SurroundCore Banking Solutions Development
+
+### Contact
+Email: recruitment@elarion.com",1.0,"LARION - Senior BA cum Scrum Master - Saigon - FT - ## Location
+* Hall 5, Quang Trung Software City, Tan Chanh Hiep Ward, District 12, HCMC
+
+## Salary Expectation
+* $1200 ~ $1500
+
+## Requirements
+* At least 4 years of business analysis related experience in software development projects
+* At least 2 years of working experience in Scrum Master or Project Manager
+* Strong communications, mentoring, and negotiation skills with stakeholders
+* Use of English proficiency (Read and understand specialized documents)
+* Excellent knowledge at software engineering and SDLC
+* Strong at the tools, techniques, processes required to capture requirements at generic and detailed levels. Those are for example: UML, OpenOffice, Mindmap, Trello, Basecamp, Pivotal Tracker, Mavenlink, Jira, Mantis, Redmine
+
+## Skills
+* Business Analyst
+* Scrum Master
+
+### Must Have
+* Have a deep know-how (at expert level) on at least 1 business domain (e.g. Healthcare, Finance, Security, Insurance...) for at least 3 continuous years in projects
+* Ability to write documents URD, SRS
+* Able to perform check (test) the business requirements
+
+### Responsibilities
+* Gather business and system requirements from client
+* Assist client to define requirements, suggest ideas, and provide solutions
+* Communicate with customers to clarify any unclear information about the business requirements
+* Analyze and specify requirements, synthesize information provided by relevant stakeholders
+* Design application mock-up / wire-frame using prototyping tools and model the requirements by using charting tools
+* Work with Quality Controller (QC) to verify that all system requirements, including change requests, are implemented and traceable
+* Take main BA role in big/complex projects (team size > 15)
+
+
+### Why it would be awesome to work with us
+* Attractive salary and benefits (13th salary, distinguised employee of the quater and year, seniority award...)
+* Performance appraisal twice a year
+* Healthcare and accident insurance
+* Various training on best practices and soft skills
+* Teambuilding activities in every summer, company trip, big annual year-end party every year, etc.
+Fitness & sports activities: football, tennis, table-tennis, badminton...
+* Commitment to community development: charity every quarter, blood donation, public seminars, career orientation talks…
+* Free in-house entertainment facilities (foodball, table tennis, boxing, coffee...)
+* Support for personal loans such as: home loan, vehicle loan, tuition fee...
+
+### Interesting Engineering Projects
+Our wide range of services includes:
+* Ruby on Rails Web Application Development
+* Big Data – Data Analytics especially in Healthcare
+* Supply Chain Management Application Development
+* Social E-commerce App Development
+* Securities Trading Solutions Development
+* SurroundCore Banking Solutions Development
+
+### Contact
+Email: recruitment@elarion.com",1,larion senior ba cum scrum master saigon ft location hall quang trung software city tan chanh hiep ward district hcmc salary expectation requirements at least years of business analysis related experience in software development projects at least years of working experience in scrum master or project manager strong communications mentoring and negotiation skills with stakeholders use of english proficiency read and understand specialized documents excellent knowledge at software engineering and sdlc strong at the tools techniques processes required to capture requirements at generic and detailed levels those are for example uml openoffice mindmap trello basecamp pivotal tracker mavenlink jira mantis redmine skills business analyst scrum master must have have a deep know how at expert level on at least business domain e g healthcare finance security insurance for at least continuous years in projects ability to write documents urd srs able to perform check test the business requirements responsibilities gather business and system requirements from client assist client to define requirements suggest ideas and provide solutions communicate with customers to clarify any unclear information about the business requirements analyze and specify requirements synthesize information provided by relevant stakeholders design application mock up wire frame using prototyping tools and model the requirements by using charting tools work with quality controller qc to verify that all system requirements including change requests are implemented and traceable take main ba role in big complex projects team size why it would be awesome to work with us attractive salary and benefits salary distinguised employee of the quater and year seniority award performance appraisal twice a year healthcare and accident insurance various training on best practices and soft skills teambuilding activities in every summer company trip big annual year end party every year etc fitness sports activities football tennis table tennis badminton commitment to community development charity every quarter blood donation public seminars career orientation talks… free in house entertainment facilities foodball table tennis boxing coffee support for personal loans such as home loan vehicle loan tuition fee interesting engineering projects our wide range of services includes ruby on rails web application development big data – data analytics especially in healthcare supply chain management application development social e commerce app development securities trading solutions development surroundcore banking solutions development contact email recruitment elarion com,1
+622,10916568412.0,IssuesEvent,2019-11-21 13:37:20,Crevette/Test,https://api.github.com/repos/Crevette/Test,opened,Migration ECM Prod-US - 17.2,DB - peopleask Deployment Env - prod-us,Could you deploy the 17.2 version of the SQL for ECM on Prod-US. Thank you !,1.0,Migration ECM Prod-US - 17.2 - Could you deploy the 17.2 version of the SQL for ECM on Prod-US. Thank you !,1,migration ecm prod us could you deploy the version of the sql for ecm on prod us thank you ,1
+2166,4306295710.0,IssuesEvent,2016-07-21 02:12:09,IBM-Bluemix/logistics-wizard,https://api.github.com/repos/IBM-Bluemix/logistics-wizard,opened,Refactor Controller utils.py files,back end controller-service task,There are too many `utils.py` files in the Controller service. Consolidate the functions scattered across these utility files and place them where they make better sense.,1.0,Refactor Controller utils.py files - There are too many `utils.py` files in the Controller service. Consolidate the functions scattered across these utility files and place them where they make better sense.,0,refactor controller utils py files there are too many utils py files in the controller service consolidate the functions scattered across these utility files and place them where they make better sense ,0
+477467,13763033725.0,IssuesEvent,2020-10-07 09:57:37,logseq/logseq,https://api.github.com/repos/logseq/logseq,closed,Cmd/Ctrl + Enter or just enter to commit,feature request priority-A,"**Is your feature request related to a problem? Please describe.**
+I love the keyboard first approach of vim and other editors and I find it disruptive whenever I have to switch from my keyboard to mouse, especially for one off button clicks or something.
+
+**Describe the solution you'd like**
+When commit dialogue is open, allow ctrl + enter to push the commit; I think this is a pretty common git shortcut in all the other git clients I've used and I find myself doing it out of habit a lot
+
+And instead of adding it to the shortcuts, you can just display the shortcut in light type underneath the button if you decide to do ctrl + enter, if not enter is native and common enough to not display anything. Basically a way to commit or sync without leaving keyboard would be nice, since I sync a lot!
+
+**Describe alternatives you've considered**
+none",1.0,"Cmd/Ctrl + Enter or just enter to commit - **Is your feature request related to a problem? Please describe.**
+I love the keyboard first approach of vim and other editors and I find it disruptive whenever I have to switch from my keyboard to mouse, especially for one off button clicks or something.
+
+**Describe the solution you'd like**
+When commit dialogue is open, allow ctrl + enter to push the commit; I think this is a pretty common git shortcut in all the other git clients I've used and I find myself doing it out of habit a lot
+
+And instead of adding it to the shortcuts, you can just display the shortcut in light type underneath the button if you decide to do ctrl + enter, if not enter is native and common enough to not display anything. Basically a way to commit or sync without leaving keyboard would be nice, since I sync a lot!
+
+**Describe alternatives you've considered**
+none",0,cmd ctrl enter or just enter to commit is your feature request related to a problem please describe i love the keyboard first approach of vim and other editors and i find it disruptive whenever i have to switch from my keyboard to mouse especially for one off button clicks or something describe the solution you d like when commit dialogue is open allow ctrl enter to push the commit i think this is a pretty common git shortcut in all the other git clients i ve used and i find myself doing it out of habit a lot and instead of adding it to the shortcuts you can just display the shortcut in light type underneath the button if you decide to do ctrl enter if not enter is native and common enough to not display anything basically a way to commit or sync without leaving keyboard would be nice since i sync a lot describe alternatives you ve considered none,0
+545,9659732494.0,IssuesEvent,2019-05-20 14:04:25,planningcenter/developers,https://api.github.com/repos/planningcenter/developers,closed,"If 2+ address/phone fields in form, only 2 of the form fields get returned.",People,"**Describe the bug**
+When creating a form, the builder allows users to add as many address and phone number fields as they want, but when the form is submitted, the form fields endpoint only returns a maximum of 2 of those address/phone number fields instead of returning all of them. Additionally, the form submission values endpoint DOES return all values submitted for the multiple fields, including an associated form field ID, but again, that form field ID is not present in the response from the form field endpoint.
+
+**To Reproduce**
+Steps to reproduce the behavior:
+1. Create a form with 3 or more phone number or address fields.
+2. Submit a test response to the form with all fields filled out.
+3. Make a call to the form's field endpoint
+4. Make a call t the form's submission values endpoint for your test submission above.
+5. Compare the form field ID fields returned with the associated submission values to all the form fields returned from step 3. In my case, I wasn't able to find a matching form field ID on a form submission value.
+
+**Expected behavior**
+Either for the front end form builder to not allow users to create forms with more than 2 phone number or address fields or for the form fields API endpoint to return data about ALL fields in the form.
+
+**Screenshots**
+If applicable, add screenshots to help explain your problem.
+
+**Context (please complete the following information):**
+- Endpoint: [https://api.planningcenteronline.com/people/v2/forms/42527/fields, https://api.planningcenteronline.com/people/v2/forms/42527/form_submissions/461596/form_submission_value]
+
+**Additional context**
+- Form fields response - https://gist.github.com/michellechu77/d09199503efe4900eeb930e2820483d0
+- Form submission values response - https://gist.github.com/michellechu77/18d5627b2613fc8a5b44717f9367b0a7
+
+For example, 268488 is return as a form field ID in the form submission values response, but not in the form fields response.
+",1.0,"If 2+ address/phone fields in form, only 2 of the form fields get returned. - **Describe the bug**
+When creating a form, the builder allows users to add as many address and phone number fields as they want, but when the form is submitted, the form fields endpoint only returns a maximum of 2 of those address/phone number fields instead of returning all of them. Additionally, the form submission values endpoint DOES return all values submitted for the multiple fields, including an associated form field ID, but again, that form field ID is not present in the response from the form field endpoint.
+
+**To Reproduce**
+Steps to reproduce the behavior:
+1. Create a form with 3 or more phone number or address fields.
+2. Submit a test response to the form with all fields filled out.
+3. Make a call to the form's field endpoint
+4. Make a call t the form's submission values endpoint for your test submission above.
+5. Compare the form field ID fields returned with the associated submission values to all the form fields returned from step 3. In my case, I wasn't able to find a matching form field ID on a form submission value.
+
+**Expected behavior**
+Either for the front end form builder to not allow users to create forms with more than 2 phone number or address fields or for the form fields API endpoint to return data about ALL fields in the form.
+
+**Screenshots**
+If applicable, add screenshots to help explain your problem.
+
+**Context (please complete the following information):**
+- Endpoint: [https://api.planningcenteronline.com/people/v2/forms/42527/fields, https://api.planningcenteronline.com/people/v2/forms/42527/form_submissions/461596/form_submission_value]
+
+**Additional context**
+- Form fields response - https://gist.github.com/michellechu77/d09199503efe4900eeb930e2820483d0
+- Form submission values response - https://gist.github.com/michellechu77/18d5627b2613fc8a5b44717f9367b0a7
+
+For example, 268488 is return as a form field ID in the form submission values response, but not in the form fields response.
+",1,if address phone fields in form only of the form fields get returned describe the bug when creating a form the builder allows users to add as many address and phone number fields as they want but when the form is submitted the form fields endpoint only returns a maximum of of those address phone number fields instead of returning all of them additionally the form submission values endpoint does return all values submitted for the multiple fields including an associated form field id but again that form field id is not present in the response from the form field endpoint to reproduce steps to reproduce the behavior create a form with or more phone number or address fields submit a test response to the form with all fields filled out make a call to the form s field endpoint make a call t the form s submission values endpoint for your test submission above compare the form field id fields returned with the associated submission values to all the form fields returned from step in my case i wasn t able to find a matching form field id on a form submission value expected behavior either for the front end form builder to not allow users to create forms with more than phone number or address fields or for the form fields api endpoint to return data about all fields in the form screenshots if applicable add screenshots to help explain your problem context please complete the following information endpoint additional context form fields response form submission values response for example is return as a form field id in the form submission values response but not in the form fields response ,1
+47303,2974714078.0,IssuesEvent,2015-07-15 03:35:02,cjfields/redmine-test,https://api.github.com/repos/cjfields/redmine-test,opened,Bio::Tools:pSW stop codon bug,Category: bioperl-ext Priority: Normal Status: New Tracker: Migrate,"---
+
+
+Author Name: **Prachi Shah** (Prachi Shah)
+Original Redmine Issue: 2069, https://redmine.open-bio.org/issues/2069
+Original Date: 2006-08-10
+Original Assignee: Bioperl Guts
+
+---
+
+I am trying to align very similar protein sequences with the
+Bio::Tools::pSW modules but running into a bug. One of the two sequences is extended with gaps so that an Amino acid residue matches the stop codon (*). The alignment should match the two sequences (because they are the same) up until the stop codon is encountered in the new sequence. Instead, it artificially extends the old sequence and matches the Alanine with the stop codon.
+
+
+ Here is an example set of two sequences I am trying to align:
+
+>orf19.6264.3
+MSNYLNLAQFSGVTDRFNLERIKSDFSSVQSTISKLRPPQEFFDFRRLSKPANFGEIQQRVGYNLGYFSANYITIVLGLSIYALITNFLLLFVTIFVLGGIYGINKLNGEDLVLPVGRFNTSQLYTGLLIVAVPLGFLASPISTMMWLIGSSGVTVGAHAALMEKPIETVFEEEV*V
+>orf19.6264.3_old
+MSNYLNLAQFSGVTDRFNLERIKSDFSSVQSTISKLRPPQEFFDFRRLSKPANFGEIQQRVGYNLGYFSANYITIVLGLSIYALITNFLLLFVTIFVLGGIYGINKLNGEDLVLPVGRFNTSQLYTGLLIVAVPLGFLASPISTMMWLIGSSGVTVGAHAALMEKPIETVFEEEV
+
+and below is the part of code that generates the alignments --
+
+################
+my $new_translatedSeqObj = Bio::Seq->new(-display_id => $gene,
+ -seq => $new_translatedSeq);
+
+my $old_translatedSeqObj = Bio::Seq->new(-display_id => $gene. ""_old"",
+ -seq => $old_translatedSeq);
+
+1. do alignments
+my $align_factory = new Bio::Tools::pSW( '-matrix' =>'blosum62.bla',
+ '-gap' => 12,
+ '-ext' => 2 );
+
+my $aln = $align_factory->pairwise_alignment( $old_translatedSeqObj,
+$new_translatedSeqObj );
+
+my $alnout = new Bio::AlignIO(-format => 'clustalw',
+ -fh => \*STDOUT);
+
+##################
+
+The alignment --
+
+CLUSTAL W(1.81) multiple sequence alignment
+
+
+orf19.6264.3_old/1-162
+MSNYLNLAQFSGVTDRFNLERIKSDFSSVQSTISKLRPPQEFFDFRRLSKPANFGEIQQR
+orf19.6264.3/1-177
+MSNYLNLAQFSGVTDRFNLERIKSDFSSVQSTISKLRPPQEFFDFRRLSKPANFGEIQQR
+
+************************************************************
+
+
+orf19.6264.3_old/1-162
+VGYNLGYFSANYITIVLGLSIYALITNFLLLFVTIFVLGGIYGINKLNGEDLVLPVGRFN
+orf19.6264.3/1-177
+VGYNLGYFSANYITIVLGLSIYALITNFLLLFVTIFVLGGIYGINKLNGEDLVLPVGRFN
+
+************************************************************
+
+
+orf19.6264.3_old/1-162 TSQLYTGLLIVAVPLGFLASPISTMMWLIGSSGVTVGAHA---------------AL
+orf19.6264.3/1-177 TSQLYTGLLIVAVPLGFLASPISTMMWLIGSSGVTVGAHAALMEKPIETVFEEEV*V
+ **************************************** :
+
+
+",1.0,"Bio::Tools:pSW stop codon bug - ---
+
+
+Author Name: **Prachi Shah** (Prachi Shah)
+Original Redmine Issue: 2069, https://redmine.open-bio.org/issues/2069
+Original Date: 2006-08-10
+Original Assignee: Bioperl Guts
+
+---
+
+I am trying to align very similar protein sequences with the
+Bio::Tools::pSW modules but running into a bug. One of the two sequences is extended with gaps so that an Amino acid residue matches the stop codon (*). The alignment should match the two sequences (because they are the same) up until the stop codon is encountered in the new sequence. Instead, it artificially extends the old sequence and matches the Alanine with the stop codon.
+
+
+ Here is an example set of two sequences I am trying to align:
+
+>orf19.6264.3
+MSNYLNLAQFSGVTDRFNLERIKSDFSSVQSTISKLRPPQEFFDFRRLSKPANFGEIQQRVGYNLGYFSANYITIVLGLSIYALITNFLLLFVTIFVLGGIYGINKLNGEDLVLPVGRFNTSQLYTGLLIVAVPLGFLASPISTMMWLIGSSGVTVGAHAALMEKPIETVFEEEV*V
+>orf19.6264.3_old
+MSNYLNLAQFSGVTDRFNLERIKSDFSSVQSTISKLRPPQEFFDFRRLSKPANFGEIQQRVGYNLGYFSANYITIVLGLSIYALITNFLLLFVTIFVLGGIYGINKLNGEDLVLPVGRFNTSQLYTGLLIVAVPLGFLASPISTMMWLIGSSGVTVGAHAALMEKPIETVFEEEV
+
+and below is the part of code that generates the alignments --
+
+################
+my $new_translatedSeqObj = Bio::Seq->new(-display_id => $gene,
+ -seq => $new_translatedSeq);
+
+my $old_translatedSeqObj = Bio::Seq->new(-display_id => $gene. ""_old"",
+ -seq => $old_translatedSeq);
+
+1. do alignments
+my $align_factory = new Bio::Tools::pSW( '-matrix' =>'blosum62.bla',
+ '-gap' => 12,
+ '-ext' => 2 );
+
+my $aln = $align_factory->pairwise_alignment( $old_translatedSeqObj,
+$new_translatedSeqObj );
+
+my $alnout = new Bio::AlignIO(-format => 'clustalw',
+ -fh => \*STDOUT);
+
+##################
+
+The alignment --
+
+CLUSTAL W(1.81) multiple sequence alignment
+
+
+orf19.6264.3_old/1-162
+MSNYLNLAQFSGVTDRFNLERIKSDFSSVQSTISKLRPPQEFFDFRRLSKPANFGEIQQR
+orf19.6264.3/1-177
+MSNYLNLAQFSGVTDRFNLERIKSDFSSVQSTISKLRPPQEFFDFRRLSKPANFGEIQQR
+
+************************************************************
+
+
+orf19.6264.3_old/1-162
+VGYNLGYFSANYITIVLGLSIYALITNFLLLFVTIFVLGGIYGINKLNGEDLVLPVGRFN
+orf19.6264.3/1-177
+VGYNLGYFSANYITIVLGLSIYALITNFLLLFVTIFVLGGIYGINKLNGEDLVLPVGRFN
+
+************************************************************
+
+
+orf19.6264.3_old/1-162 TSQLYTGLLIVAVPLGFLASPISTMMWLIGSSGVTVGAHA---------------AL
+orf19.6264.3/1-177 TSQLYTGLLIVAVPLGFLASPISTMMWLIGSSGVTVGAHAALMEKPIETVFEEEV*V
+ **************************************** :
+
+
+",0,bio tools psw stop codon bug author name prachi shah prachi shah original redmine issue original date original assignee bioperl guts i am trying to align very similar protein sequences with the bio tools psw modules but running into a bug one of the two sequences is extended with gaps so that an amino acid residue matches the stop codon the alignment should match the two sequences because they are the same up until the stop codon is encountered in the new sequence instead it artificially extends the old sequence and matches the alanine with the stop codon here is an example set of two sequences i am trying to align msnylnlaqfsgvtdrfnleriksdfssvqstisklrppqeffdfrrlskpanfgeiqqrvgynlgyfsanyitivlglsiyalitnflllfvtifvlggiyginklngedlvlpvgrfntsqlytgllivavplgflaspistmmwligssgvtvgahaalmekpietvfeeev v old msnylnlaqfsgvtdrfnleriksdfssvqstisklrppqeffdfrrlskpanfgeiqqrvgynlgyfsanyitivlglsiyalitnflllfvtifvlggiyginklngedlvlpvgrfntsqlytgllivavplgflaspistmmwligssgvtvgahaalmekpietvfeeev and below is the part of code that generates the alignments my new translatedseqobj bio seq new display id gene seq new translatedseq my old translatedseqobj bio seq new display id gene old seq old translatedseq do alignments my align factory new bio tools psw matrix bla gap ext my aln align factory pairwise alignment old translatedseqobj new translatedseqobj my alnout new bio alignio format clustalw fh stdout the alignment clustal w multiple sequence alignment old msnylnlaqfsgvtdrfnleriksdfssvqstisklrppqeffdfrrlskpanfgeiqqr msnylnlaqfsgvtdrfnleriksdfssvqstisklrppqeffdfrrlskpanfgeiqqr old vgynlgyfsanyitivlglsiyalitnflllfvtifvlggiyginklngedlvlpvgrfn vgynlgyfsanyitivlglsiyalitnflllfvtifvlggiyginklngedlvlpvgrfn old tsqlytgllivavplgflaspistmmwligssgvtvgaha al tsqlytgllivavplgflaspistmmwligssgvtvgahaalmekpietvfeeev v ,0
+203947,7078347245.0,IssuesEvent,2018-01-10 03:11:35,TMats/survey,https://api.github.com/repos/TMats/survey,opened,Vision-Based Multi-Task Manipulation for Inexpensive Robots Using End-To-End Learning from Demonstration,Priority: High robotics,"https://arxiv.org/abs/1707.02920
+- Rouhollah Rahmatizadeh, Pooya Abolghasemi, Ladislau Bölöni, Sergey Levine
+- Submitted on 10 Jul 2017",1.0,"Vision-Based Multi-Task Manipulation for Inexpensive Robots Using End-To-End Learning from Demonstration - https://arxiv.org/abs/1707.02920
+- Rouhollah Rahmatizadeh, Pooya Abolghasemi, Ladislau Bölöni, Sergey Levine
+- Submitted on 10 Jul 2017",0,vision based multi task manipulation for inexpensive robots using end to end learning from demonstration rouhollah rahmatizadeh pooya abolghasemi ladislau bölöni sergey levine submitted on jul ,0
+496518,14349099057.0,IssuesEvent,2020-11-29 15:14:44,Kekmech/mpeix-common,https://api.github.com/repos/Kekmech/mpeix-common,closed,Портировать CI/CD на Portainer API,Priority: High Type: Enhancement,Сейчас в Github Actions мы через Ansible деплоим Docker контейнер с приложением. Так как перешли на Docker Swarm данный подход больше не будет работать. Так же данный подход является не самым лучшим решением и был выбран лишь для написания прототипа CI/CD. Необходимо переписать старое решение на update приложения через Portainer API,1.0,Портировать CI/CD на Portainer API - Сейчас в Github Actions мы через Ansible деплоим Docker контейнер с приложением. Так как перешли на Docker Swarm данный подход больше не будет работать. Так же данный подход является не самым лучшим решением и был выбран лишь для написания прототипа CI/CD. Необходимо переписать старое решение на update приложения через Portainer API,0,портировать ci cd на portainer api сейчас в github actions мы через ansible деплоим docker контейнер с приложением так как перешли на docker swarm данный подход больше не будет работать так же данный подход является не самым лучшим решением и был выбран лишь для написания прототипа ci cd необходимо переписать старое решение на update приложения через portainer api,0
+988,23416081222.0,IssuesEvent,2022-08-13 01:45:24,GoogleCloudPlatform/python-docs-samples,https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples,closed,Issue in Global Fishing Watch -- timeseries-classification,needs more info type: question samples api: people-and-planet-ai,"## In which file did you encounter the issue?
+
+https://github.com/GoogleCloudPlatform/python-docs-samples/tree/main/people-and-planet-ai/timeseries-classification
+```
+from tensorflow import keras
+
+model_dir = f'{storage_path}/training/model'
+model = keras.models.load_model(model_dir)
+
+model.summary()
+keras.utils.plot_model(model, show_shapes=True)
+
+```
+
+
+
+
+ 4 model = keras.models.load_model(model_dir)
+ 5
+ 6 model.summary()
+
+1 frames
+[/usr/local/lib/python3.7/dist-packages/keras/saving/save.py](https://localhost:8080/#) in load_model(filepath, custom_objects, compile, options)
+ 202 if isinstance(filepath_str, str):
+ 203 if not tf.io.gfile.exists(filepath_str):
+--> 204 raise IOError(f'No file or directory found at {filepath_str}')
+ 205
+ 206 if tf.io.gfile.isdir(filepath_str):
+
+OSError: No file or directory found at gs://fishingbucket/samples/global-fishing-watch/training/model. -->
+",1.0,"Issue in Global Fishing Watch -- timeseries-classification - ## In which file did you encounter the issue?
+
+https://github.com/GoogleCloudPlatform/python-docs-samples/tree/main/people-and-planet-ai/timeseries-classification
+```
+from tensorflow import keras
+
+model_dir = f'{storage_path}/training/model'
+model = keras.models.load_model(model_dir)
+
+model.summary()
+keras.utils.plot_model(model, show_shapes=True)
+
+```
+
+
+
+
+ 4 model = keras.models.load_model(model_dir)
+ 5
+ 6 model.summary()
+
+1 frames
+[/usr/local/lib/python3.7/dist-packages/keras/saving/save.py](https://localhost:8080/#) in load_model(filepath, custom_objects, compile, options)
+ 202 if isinstance(filepath_str, str):
+ 203 if not tf.io.gfile.exists(filepath_str):
+--> 204 raise IOError(f'No file or directory found at {filepath_str}')
+ 205
+ 206 if tf.io.gfile.isdir(filepath_str):
+
+OSError: No file or directory found at gs://fishingbucket/samples/global-fishing-watch/training/model. -->
+",1,issue in global fishing watch timeseries classification in which file did you encounter the issue from tensorflow import keras model dir f storage path training model model keras models load model model dir model summary keras utils plot model model show shapes true oserror traceback most recent call last in model dir f storage path training model model keras models load model model dir model summary frames in load model filepath custom objects compile options if isinstance filepath str str if not tf io gfile exists filepath str raise ioerror f no file or directory found at filepath str if tf io gfile isdir filepath str oserror no file or directory found at gs fishingbucket samples global fishing watch training model ,1
+678,12219884832.0,IssuesEvent,2020-05-01 23:08:02,microsoftgraph/microsoft-graph-toolkit,https://api.github.com/repos/microsoftgraph/microsoft-graph-toolkit,closed,[People picker] should be able to search in /users and /groups,Component: people-picker Priority: 1 State: In Review State: Started feature-request,"## Description
+People picker should be able to pick any user in the organization or a group. It should also be configurable by the developer to allow for only people or only groups.
+
+## Rationale
+The people picker can only search for people using the people api. However, the people api does not expose all the users or groups in the organizations, only the ones related to the current signed in user.
+
+## Preferred Solution
+
+The people picker should first try to use the `/people` api to ensure first results are relevant to the user. If the `/people` api does not return enough values, then the picker should use the `/users` or `/groups` api to get additional values.
+
+Add two properties to the people picker to configure what type of person to search for and what type of group to search for:
+
+```html
+
+```
+
+## Additional Context
+
+",1.0,"[People picker] should be able to search in /users and /groups - ## Description
+People picker should be able to pick any user in the organization or a group. It should also be configurable by the developer to allow for only people or only groups.
+
+## Rationale
+The people picker can only search for people using the people api. However, the people api does not expose all the users or groups in the organizations, only the ones related to the current signed in user.
+
+## Preferred Solution
+
+The people picker should first try to use the `/people` api to ensure first results are relevant to the user. If the `/people` api does not return enough values, then the picker should use the `/users` or `/groups` api to get additional values.
+
+Add two properties to the people picker to configure what type of person to search for and what type of group to search for:
+
+```html
+
+```
+
+## Additional Context
+
+",1, should be able to search in users and groups description people picker should be able to pick any user in the organization or a group it should also be configurable by the developer to allow for only people or only groups rationale the people picker can only search for people using the people api however the people api does not expose all the users or groups in the organizations only the ones related to the current signed in user preferred solution the people picker should first try to use the people api to ensure first results are relevant to the user if the people api does not return enough values then the picker should use the users or groups api to get additional values add two properties to the people picker to configure what type of person to search for and what type of group to search for html additional context ,1
+366428,10820664131.0,IssuesEvent,2019-11-08 16:51:23,AY1920S1-CS2113T-W17-4/main,https://api.github.com/repos/AY1920S1-CS2113T-W17-4/main,closed,"As a Computing student, I can add my estimated time taken to complete a task",priority.Low type.Story,"so that I know how much free time I would have.
+",1.0,"As a Computing student, I can add my estimated time taken to complete a task - so that I know how much free time I would have.
+",0,as a computing student i can add my estimated time taken to complete a task so that i know how much free time i would have ,0
+10663,8670447834.0,IssuesEvent,2018-11-29 16:31:37,DestinyItemManager/DIM,https://api.github.com/repos/DestinyItemManager/DIM,closed,Save Data (loadouts and tags) gone (Beta) from 2 weeks,Bug Infrastructure: Sync,"I lost all my loadout from the last 2 week. I‘m also missing the 200-300 tags on my items i made in that time.
+Cant restore the data form this time periode (google drive)
+Anybody got an idea? Could it be that not all backup data is displayed?
+
+https://i.imgur.com/3fRKc25.png",1.0,"Save Data (loadouts and tags) gone (Beta) from 2 weeks - I lost all my loadout from the last 2 week. I‘m also missing the 200-300 tags on my items i made in that time.
+Cant restore the data form this time periode (google drive)
+Anybody got an idea? Could it be that not all backup data is displayed?
+
+https://i.imgur.com/3fRKc25.png",0,save data loadouts and tags gone beta from weeks i lost all my loadout from the last week i‘m also missing the tags on my items i made in that time cant restore the data form this time periode google drive anybody got an idea could it be that not all backup data is displayed ,0
+17581,24247910784.0,IssuesEvent,2022-09-27 12:08:33,ZsoltMolnarrr/BetterCombat,https://api.github.com/repos/ZsoltMolnarrr/BetterCombat,reopened,Incompatibility with PlayerRevive,enhancement incompatibility,"PlayerRevive_FORGE_v2.0.10_mc1.19.2.jar (https://www.curseforge.com/minecraft/mc-mods/playerrevive)
+bettercombat-forge-1.4.1+1.19.jar
+
+PlayerRevive mode makes player down for certain amount of seconds when hp reach to zero. When player down, they crawl slowly and can't attack before someone revive them or die bleeding. But with Better Combat mode, player is still able to attack when downed.
+
+According to PlayerRevive Mode developer, PlayerRevive cancels the ClickInputEvent when a player is downed([source](https://github.com/CreativeMD/PlayerRevive/blob/1.18/src/main/java/team/creative/playerrevive/client/ReviveEventClient.java#L62)). So somehow this is being ignored with Better Combat.",True,"Incompatibility with PlayerRevive - PlayerRevive_FORGE_v2.0.10_mc1.19.2.jar (https://www.curseforge.com/minecraft/mc-mods/playerrevive)
+bettercombat-forge-1.4.1+1.19.jar
+
+PlayerRevive mode makes player down for certain amount of seconds when hp reach to zero. When player down, they crawl slowly and can't attack before someone revive them or die bleeding. But with Better Combat mode, player is still able to attack when downed.
+
+According to PlayerRevive Mode developer, PlayerRevive cancels the ClickInputEvent when a player is downed([source](https://github.com/CreativeMD/PlayerRevive/blob/1.18/src/main/java/team/creative/playerrevive/client/ReviveEventClient.java#L62)). So somehow this is being ignored with Better Combat.",0,incompatibility with playerrevive playerrevive forge jar bettercombat forge jar playerrevive mode makes player down for certain amount of seconds when hp reach to zero when player down they crawl slowly and can t attack before someone revive them or die bleeding but with better combat mode player is still able to attack when downed according to playerrevive mode developer playerrevive cancels the clickinputevent when a player is downed so somehow this is being ignored with better combat ,0
+505389,14632319305.0,IssuesEvent,2020-12-23 22:04:44,teamforus/general,https://api.github.com/repos/teamforus/general,closed,Provider overview page,Approval: Granted Impact: Significant Priority: Must have Status: Not Planned Type: Change request Urgency: Low - When planning allows project-100,"Learn more about change requests here: https://bit.ly/39CWeEE
+
+### Requested by:
+Jamal
+
+### Change description
+As a provider I would like an overview of what's expected of me after completing the signupflow.
+
+This is a proposed solution to https://github.com/teamforus/general/issues/624
+
+Another important reason: right now provider's have a conceptual ""gap"" after sign in, it's very unclear that you should start adding some offers in order to start making some profits. This page will hopefully help with that problem as well.
+
+Figma: https://www.figma.com/file/hJS4uLSGgA1cMdyVtqDinG/?node-id=1%3A80
+
+## Proposal:
+
+
+
+",1.0,"Provider overview page - Learn more about change requests here: https://bit.ly/39CWeEE
+
+### Requested by:
+Jamal
+
+### Change description
+As a provider I would like an overview of what's expected of me after completing the signupflow.
+
+This is a proposed solution to https://github.com/teamforus/general/issues/624
+
+Another important reason: right now provider's have a conceptual ""gap"" after sign in, it's very unclear that you should start adding some offers in order to start making some profits. This page will hopefully help with that problem as well.
+
+Figma: https://www.figma.com/file/hJS4uLSGgA1cMdyVtqDinG/?node-id=1%3A80
+
+## Proposal:
+
+
+
+",0,provider overview page learn more about change requests here requested by jamal change description as a provider i would like an overview of what s expected of me after completing the signupflow this is a proposed solution to another important reason right now provider s have a conceptual gap after sign in it s very unclear that you should start adding some offers in order to start making some profits this page will hopefully help with that problem as well figma proposal ,0
+84,3362033537.0,IssuesEvent,2015-11-20 01:55:47,bbergen/tbmd.com,https://api.github.com/repos/bbergen/tbmd.com,closed,people.php styling,People Style,"Similar to movie block views, people block views will need styling to lay out the content in a nice fashion. This should be located in a new and separate style sheet (people.css) similar to how movie.css was done. Thus only pages that require the styling need to link to it. ",1.0,"people.php styling - Similar to movie block views, people block views will need styling to lay out the content in a nice fashion. This should be located in a new and separate style sheet (people.css) similar to how movie.css was done. Thus only pages that require the styling need to link to it. ",1,people php styling similar to movie block views people block views will need styling to lay out the content in a nice fashion this should be located in a new and separate style sheet people css similar to how movie css was done thus only pages that require the styling need to link to it ,1
+1062,25291163469.0,IssuesEvent,2022-11-17 00:25:24,jongfeel/BookReview,https://api.github.com/repos/jongfeel/BookReview,closed,6부 6장 행복하길 원한다면 이것을 게을리하지 마라,2022 How to Win Friends & Influence People,"### 6장 행복하길 원한다면 이것을 게을리하지 마라
+
+“놀라운 사실을 말하자면, 사실상 비열하고, 모욕적이고, 우리에게 상처를 주는 말을 하는 유일한 사람들은 바로 우리의 가족들이다.”
+
+“예의란 망가진 문에는 관심을 두지 않고 문 뒤 정원에 핀 꽃에 관심을 갖는 마음 자세이다.”
+
+기름이 없으면 차가 움직일 수 없듯이 예의야말로 결혼을 유지하는 비결이다.
+
+---
+
+비결 6: 예의를 차려라.
+Be courteous.",1.0,"6부 6장 행복하길 원한다면 이것을 게을리하지 마라 - ### 6장 행복하길 원한다면 이것을 게을리하지 마라
+
+“놀라운 사실을 말하자면, 사실상 비열하고, 모욕적이고, 우리에게 상처를 주는 말을 하는 유일한 사람들은 바로 우리의 가족들이다.”
+
+“예의란 망가진 문에는 관심을 두지 않고 문 뒤 정원에 핀 꽃에 관심을 갖는 마음 자세이다.”
+
+기름이 없으면 차가 움직일 수 없듯이 예의야말로 결혼을 유지하는 비결이다.
+
+---
+
+비결 6: 예의를 차려라.
+Be courteous.",1, 행복하길 원한다면 이것을 게을리하지 마라 행복하길 원한다면 이것을 게을리하지 마라 “놀라운 사실을 말하자면 사실상 비열하고 모욕적이고 우리에게 상처를 주는 말을 하는 유일한 사람들은 바로 우리의 가족들이다 ” “예의란 망가진 문에는 관심을 두지 않고 문 뒤 정원에 핀 꽃에 관심을 갖는 마음 자세이다 ” 기름이 없으면 차가 움직일 수 없듯이 예의야말로 결혼을 유지하는 비결이다 비결 예의를 차려라 be courteous ,1
+1923,2645590424.0,IssuesEvent,2015-03-13 00:06:21,pvlib/pvlib-python,https://api.github.com/repos/pvlib/pvlib-python,closed,Add What's New section to docs,documentation,"A good pattern seems to be to make a separate file for each release.
+
+Possibly helpful links
+
+https://github.com/pydata/pandas/tree/master/doc/source
+
+https://github.com/ipython/ipython/tree/master/docs/source
+
+https://github.com/statsmodels/statsmodels/tree/master/docs/source",1.0,"Add What's New section to docs - A good pattern seems to be to make a separate file for each release.
+
+Possibly helpful links
+
+https://github.com/pydata/pandas/tree/master/doc/source
+
+https://github.com/ipython/ipython/tree/master/docs/source
+
+https://github.com/statsmodels/statsmodels/tree/master/docs/source",0,add what s new section to docs a good pattern seems to be to make a separate file for each release possibly helpful links ,0
+128307,10524579732.0,IssuesEvent,2019-09-30 13:34:09,hazelcast/hazelcast-jet,https://api.github.com/repos/hazelcast/hazelcast-jet,closed,com.hazelcast.jet.core.JobTest.when_joinFromClientTimesOut_then_futureShouldNotBeCompletedEarly,test-failure,"_master_ (commit f8fd5970adfbfcf47944e306bbe569924ace8c27)
+
+Failed on OracleJDK 11: http://jenkins.hazelcast.com/job/jet-oss-master-oracle-jdk11/111/testReport/junit/com.hazelcast.jet.core/JobTest/when_joinFromClientTimesOut_then_futureShouldNotBeCompletedEarly/
+
+Stacktrace:
+```
+com.hazelcast.spi.exception.TargetNotMemberException: Target '[127.0.0.1]:5701' is not a member.
+ at com.hazelcast.client.spi.impl.SmartClientInvocationService.invokeOnTarget(SmartClientInvocationService.java:66)
+ at com.hazelcast.client.spi.impl.ClientInvocation.invokeOnSelection(ClientInvocation.java:165)
+ at com.hazelcast.client.spi.impl.ClientInvocation.retry(ClientInvocation.java:194)
+ at com.hazelcast.client.spi.impl.ClientInvocation.run(ClientInvocation.java:179)
+ at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
+ at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
+ at java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304)
+ at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
+ at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
+ at java.base/java.lang.Thread.run(Thread.java:834)
+ at com.hazelcast.util.executor.HazelcastManagedThread.executeRun(HazelcastManagedThread.java:64)
+ at com.hazelcast.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:80)
+ at ------ submitted from ------.(Unknown Source)
+ at com.hazelcast.client.spi.impl.ClientInvocationFuture.resolveAndThrowIfException(ClientInvocationFuture.java:96)
+ at com.hazelcast.client.spi.impl.ClientInvocationFuture.resolveAndThrowIfException(ClientInvocationFuture.java:33)
+ at com.hazelcast.spi.impl.AbstractInvocationFuture.get(AbstractInvocationFuture.java:163)
+ at com.hazelcast.jet.impl.ClientJobProxy$CancellableFuture.get(ClientJobProxy.java:214)
+ at com.hazelcast.jet.impl.ClientJobProxy$CancellableFuture.get(ClientJobProxy.java:189)
+ at com.hazelcast.jet.impl.AbstractJobProxy.terminate(AbstractJobProxy.java:158)
+ at com.hazelcast.jet.impl.AbstractJobProxy.cancel(AbstractJobProxy.java:141)
+ at com.hazelcast.jet.core.JobTest.when_joinFromClientTimesOut_then_futureShouldNotBeCompletedEarly(JobTest.java:779)
+ at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
+ at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
+ at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
+ at java.base/java.lang.reflect.Method.invoke(Method.java:566)
+ at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
+ at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
+ at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
+ at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
+ at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:106)
+ at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:98)
+ at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
+ at java.base/java.lang.Thread.run(Thread.java:834)
+```
+
+Standard output:
+```
+Hiccups measured while running test 'when_joinFromClientTimesOut_then_futureShouldNotBeCompletedEarly(com.hazelcast.jet.core.JobTest):'
+19:40:50, accumulated pauses: 320 ms, max pause: 6 ms, pauses over 1000 ms: 0
+19:40:55, accumulated pauses: 299 ms, max pause: 226 ms, pauses over 1000 ms: 0
+```
+",1.0,"com.hazelcast.jet.core.JobTest.when_joinFromClientTimesOut_then_futureShouldNotBeCompletedEarly - _master_ (commit f8fd5970adfbfcf47944e306bbe569924ace8c27)
+
+Failed on OracleJDK 11: http://jenkins.hazelcast.com/job/jet-oss-master-oracle-jdk11/111/testReport/junit/com.hazelcast.jet.core/JobTest/when_joinFromClientTimesOut_then_futureShouldNotBeCompletedEarly/
+
+Stacktrace:
+```
+com.hazelcast.spi.exception.TargetNotMemberException: Target '[127.0.0.1]:5701' is not a member.
+ at com.hazelcast.client.spi.impl.SmartClientInvocationService.invokeOnTarget(SmartClientInvocationService.java:66)
+ at com.hazelcast.client.spi.impl.ClientInvocation.invokeOnSelection(ClientInvocation.java:165)
+ at com.hazelcast.client.spi.impl.ClientInvocation.retry(ClientInvocation.java:194)
+ at com.hazelcast.client.spi.impl.ClientInvocation.run(ClientInvocation.java:179)
+ at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
+ at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
+ at java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304)
+ at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
+ at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
+ at java.base/java.lang.Thread.run(Thread.java:834)
+ at com.hazelcast.util.executor.HazelcastManagedThread.executeRun(HazelcastManagedThread.java:64)
+ at com.hazelcast.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:80)
+ at ------ submitted from ------.(Unknown Source)
+ at com.hazelcast.client.spi.impl.ClientInvocationFuture.resolveAndThrowIfException(ClientInvocationFuture.java:96)
+ at com.hazelcast.client.spi.impl.ClientInvocationFuture.resolveAndThrowIfException(ClientInvocationFuture.java:33)
+ at com.hazelcast.spi.impl.AbstractInvocationFuture.get(AbstractInvocationFuture.java:163)
+ at com.hazelcast.jet.impl.ClientJobProxy$CancellableFuture.get(ClientJobProxy.java:214)
+ at com.hazelcast.jet.impl.ClientJobProxy$CancellableFuture.get(ClientJobProxy.java:189)
+ at com.hazelcast.jet.impl.AbstractJobProxy.terminate(AbstractJobProxy.java:158)
+ at com.hazelcast.jet.impl.AbstractJobProxy.cancel(AbstractJobProxy.java:141)
+ at com.hazelcast.jet.core.JobTest.when_joinFromClientTimesOut_then_futureShouldNotBeCompletedEarly(JobTest.java:779)
+ at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
+ at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
+ at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
+ at java.base/java.lang.reflect.Method.invoke(Method.java:566)
+ at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
+ at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
+ at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
+ at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
+ at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:106)
+ at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:98)
+ at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
+ at java.base/java.lang.Thread.run(Thread.java:834)
+```
+
+Standard output:
+```
+Hiccups measured while running test 'when_joinFromClientTimesOut_then_futureShouldNotBeCompletedEarly(com.hazelcast.jet.core.JobTest):'
+19:40:50, accumulated pauses: 320 ms, max pause: 6 ms, pauses over 1000 ms: 0
+19:40:55, accumulated pauses: 299 ms, max pause: 226 ms, pauses over 1000 ms: 0
+```
+",0,com hazelcast jet core jobtest when joinfromclienttimesout then futureshouldnotbecompletedearly master commit failed on oraclejdk stacktrace com hazelcast spi exception targetnotmemberexception target is not a member at com hazelcast client spi impl smartclientinvocationservice invokeontarget smartclientinvocationservice java at com hazelcast client spi impl clientinvocation invokeonselection clientinvocation java at com hazelcast client spi impl clientinvocation retry clientinvocation java at com hazelcast client spi impl clientinvocation run clientinvocation java at java base java util concurrent executors runnableadapter call executors java at java base java util concurrent futuretask run futuretask java at java base java util concurrent scheduledthreadpoolexecutor scheduledfuturetask run scheduledthreadpoolexecutor java at java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java base java lang thread run thread java at com hazelcast util executor hazelcastmanagedthread executerun hazelcastmanagedthread java at com hazelcast util executor hazelcastmanagedthread run hazelcastmanagedthread java at submitted from unknown source at com hazelcast client spi impl clientinvocationfuture resolveandthrowifexception clientinvocationfuture java at com hazelcast client spi impl clientinvocationfuture resolveandthrowifexception clientinvocationfuture java at com hazelcast spi impl abstractinvocationfuture get abstractinvocationfuture java at com hazelcast jet impl clientjobproxy cancellablefuture get clientjobproxy java at com hazelcast jet impl clientjobproxy cancellablefuture get clientjobproxy java at com hazelcast jet impl abstractjobproxy terminate abstractjobproxy java at com hazelcast jet impl abstractjobproxy cancel abstractjobproxy java at com hazelcast jet core jobtest when joinfromclienttimesout then futureshouldnotbecompletedearly jobtest java at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at com hazelcast test failontimeoutstatement callablestatement call failontimeoutstatement java at com hazelcast test failontimeoutstatement callablestatement call failontimeoutstatement java at java base java util concurrent futuretask run futuretask java at java base java lang thread run thread java standard output hiccups measured while running test when joinfromclienttimesout then futureshouldnotbecompletedearly com hazelcast jet core jobtest accumulated pauses ms max pause ms pauses over ms accumulated pauses ms max pause ms pauses over ms ,0
+30576,14610911894.0,IssuesEvent,2020-12-22 01:47:35,sergiorribeiro/webmetry,https://api.github.com/repos/sergiorribeiro/webmetry,opened,[Controller] Controller/admin/dashboards/show,prodops transaction-performance,"The transaction **`Controller/admin/dashboards/show` (Controller)** violated a performance threshold.
+
+## Violations:
+- [2020-12-22] Maximum execution duration during the current week was exceeded. Duration: **`1.2 min`**. Limit: **`30 s`**.
+
+## Weekly transaction performance:
+
+### Evolution graph (percentile 95):
+
+```
+[2020-12-15] ~ [2020-12-22] 🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦⬜️⬜️⬜️⬜️⬜️ 15.3 s
+[2020-12-08] ~ [2020-12-15] 🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦 19.5 s
+[2020-12-01] ~ [2020-12-08] 🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦⬜️⬜️⬜️⬜️⬜️⬜️⬜️ 12.8 s
+[2020-11-24] ~ [2020-12-01] 🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦⬜️⬜️⬜️⬜️⬜️⬜️⬜️⬜️ 12.3 s
+
+```
+### Weekly indicators:
+
+| Indicator | Week -3 | Week -2 | Week -1 | Week 0 |
+|-|-|-|-|-|
+| Above average hits | 3 | 9 | 11 | 9 |
+| Max | 43.8 s | 58.9 s | 1.2 min | 1.2 min |
+| Average | 1.4 s | 1.8 s | 2.6 s | 2.1 s |
+| Percentile 95 | 12.3 s | 12.8 s | 19.5 s | 15.3 s |
+
+### Month totals:
+
+| Indicator | Value |
+|-|-|
+| Max | 1.2 min |
+| Average | 2.0 s |
+| Percentile 95 | 14.8 s |
+
+",True,"[Controller] Controller/admin/dashboards/show - The transaction **`Controller/admin/dashboards/show` (Controller)** violated a performance threshold.
+
+## Violations:
+- [2020-12-22] Maximum execution duration during the current week was exceeded. Duration: **`1.2 min`**. Limit: **`30 s`**.
+
+## Weekly transaction performance:
+
+### Evolution graph (percentile 95):
+
+```
+[2020-12-15] ~ [2020-12-22] 🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦⬜️⬜️⬜️⬜️⬜️ 15.3 s
+[2020-12-08] ~ [2020-12-15] 🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦 19.5 s
+[2020-12-01] ~ [2020-12-08] 🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦⬜️⬜️⬜️⬜️⬜️⬜️⬜️ 12.8 s
+[2020-11-24] ~ [2020-12-01] 🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦⬜️⬜️⬜️⬜️⬜️⬜️⬜️⬜️ 12.3 s
+
+```
+### Weekly indicators:
+
+| Indicator | Week -3 | Week -2 | Week -1 | Week 0 |
+|-|-|-|-|-|
+| Above average hits | 3 | 9 | 11 | 9 |
+| Max | 43.8 s | 58.9 s | 1.2 min | 1.2 min |
+| Average | 1.4 s | 1.8 s | 2.6 s | 2.1 s |
+| Percentile 95 | 12.3 s | 12.8 s | 19.5 s | 15.3 s |
+
+### Month totals:
+
+| Indicator | Value |
+|-|-|
+| Max | 1.2 min |
+| Average | 2.0 s |
+| Percentile 95 | 14.8 s |
+
+",0, controller admin dashboards show the transaction controller admin dashboards show controller violated a performance threshold violations maximum execution duration during the current week was exceeded duration min limit s weekly transaction performance evolution graph percentile 🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦⬜️⬜️⬜️⬜️⬜️ s 🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦 s 🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦⬜️⬜️⬜️⬜️⬜️⬜️⬜️ s 🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦⬜️⬜️⬜️⬜️⬜️⬜️⬜️⬜️ s weekly indicators indicator week week week week above average hits max s s min min average s s s s percentile s s s s month totals indicator value max min average s percentile s ,0
+275728,23933814456.0,IssuesEvent,2022-09-10 23:53:40,kubernetes/test-infra,https://api.github.com/repos/kubernetes/test-infra,closed,"error unmarshaling config/jobs/kubernetes-sigs/sig-windows/soak-tests.yaml: error unmarshaling JSON: while decoding JSON: json: unknown field \""always_run\""",kind/bug sig/testing,"### What happened:
+***
+
+Prow's [checkconfig](https://github.com/kubernetes/test-infra/tree/master/prow/cmd/checkconfig) tool is responsible for presubmit validation of ProwJob configuration. It has an optional check for unknown fields in ProwJob config that can be enabled with `--warnings=unknown-fields-all`.
+
+As a result it is possible to add configuration that is not recognized by Prow, causing it to be silently ignored. Common examples of this are typos in a field name and incorrect indentation.
+
+It would be great to enable this check across the repo, but before we can we must address all existing unknown fields.
+
+ `--warnings=unknown-fields-all` [was added to an optional job and merged](https://github.com/kubernetes/test-infra/pull/27391)
+
+***
+[checkconfig](https://github.com/kubernetes/test-infra/tree/master/prow/cmd/checkconfig) with --warnings=unknown-fields-all run locally with the following-
+
+`docker run -i --rm -v ""${PWD}:${PWD}"" -w ""${PWD}"" gcr.io/k8s-prow/checkconfig:v20220905-8bcebc6376 --config-path=config/prow/config.yaml --job-config-path=config/jobs --plugin-config=config/prow/plugins.yaml --strict --warnings=mismatched-tide-lenient --warnings=tide-strict-branch --warnings=needs-ok-to-test --warnings=validate-owners --warnings=missing-trigger --warnings=validate-urls --warnings=unknown-fields --warnings=duplicate-job-refs --warnings=unknown-fields-all`
+
+returns error:
+
+`{""component"":""checkconfig"",""file"":""k8s.io/test-infra/prow/cmd/checkconfig/main.go:91"",""func"":""main.reportWarning"",""level"":""warning"",""msg"":""error unmarshaling config/jobs/kubernetes-sigs/sig-windows/soak-tests.yaml: error unmarshaling JSON: while decoding JSON: json: unknown field \""always_run\"""",""severity"":""warning"",""time"":""2022-09-08T18:58:25Z""}
+`
+
+
+***
+
+### What you expected to happen:
+***
+Output:
+`{""component"":""checkconfig"",""file"":""k8s.io/test-infra/prow/cmd/checkconfig/main.go:252"",""func"":""main.main"",""level"":""info"",""msg"":""checkconfig passes without any error!"",""severity"":""info"",""time"":""2022-09-08T17:44:36Z""}`
+***
+
+### How to reproduce it (as minimally and precisely as possible):
+***
+run the following in the root locally
+
+`docker run -i --rm -v ""${PWD}:${PWD}"" -w ""${PWD}"" gcr.io/k8s-prow/checkconfig:v20220905-8bcebc6376 --config-path=config/prow/config.yaml --job-config-path=config/jobs --plugin-config=config/prow/plugins.yaml --strict --warnings=mismatched-tide-lenient --warnings=tide-strict-branch --warnings=needs-ok-to-test --warnings=validate-owners --warnings=missing-trigger --warnings=validate-urls --warnings=unknown-fields --warnings=duplicate-job-refs --warnings=unknown-fields-all`
+
+***
+
+### Proposed Fix:
+
+Issue occurs when `periodics:` is passed either `always_run:` or `optional:` ... the errors resolve themselves if the following diff is applied
+***
+```diff
+
+index b45dd90a45..5e626166fa 100644
+--- a/config/jobs/kubernetes-sigs/sig-windows/soak-tests.yaml
++++ b/config/jobs/kubernetes-sigs/sig-windows/soak-tests.yaml
+@@ -86,8 +86,6 @@ periodics:
+ - interval: 24h
+ name: periodic-soak-tests-capz-windows-2019
+ decorate: true
+- always_run: false
+- optional: true
+ decoration_config:
+ timeout: 8h
+ path_alias: k8s.io/perf-tests
+
+
+```
+
+***
+
+/sig testing
+
+",1.0,"error unmarshaling config/jobs/kubernetes-sigs/sig-windows/soak-tests.yaml: error unmarshaling JSON: while decoding JSON: json: unknown field \""always_run\"" - ### What happened:
+***
+
+Prow's [checkconfig](https://github.com/kubernetes/test-infra/tree/master/prow/cmd/checkconfig) tool is responsible for presubmit validation of ProwJob configuration. It has an optional check for unknown fields in ProwJob config that can be enabled with `--warnings=unknown-fields-all`.
+
+As a result it is possible to add configuration that is not recognized by Prow, causing it to be silently ignored. Common examples of this are typos in a field name and incorrect indentation.
+
+It would be great to enable this check across the repo, but before we can we must address all existing unknown fields.
+
+ `--warnings=unknown-fields-all` [was added to an optional job and merged](https://github.com/kubernetes/test-infra/pull/27391)
+
+***
+[checkconfig](https://github.com/kubernetes/test-infra/tree/master/prow/cmd/checkconfig) with --warnings=unknown-fields-all run locally with the following-
+
+`docker run -i --rm -v ""${PWD}:${PWD}"" -w ""${PWD}"" gcr.io/k8s-prow/checkconfig:v20220905-8bcebc6376 --config-path=config/prow/config.yaml --job-config-path=config/jobs --plugin-config=config/prow/plugins.yaml --strict --warnings=mismatched-tide-lenient --warnings=tide-strict-branch --warnings=needs-ok-to-test --warnings=validate-owners --warnings=missing-trigger --warnings=validate-urls --warnings=unknown-fields --warnings=duplicate-job-refs --warnings=unknown-fields-all`
+
+returns error:
+
+`{""component"":""checkconfig"",""file"":""k8s.io/test-infra/prow/cmd/checkconfig/main.go:91"",""func"":""main.reportWarning"",""level"":""warning"",""msg"":""error unmarshaling config/jobs/kubernetes-sigs/sig-windows/soak-tests.yaml: error unmarshaling JSON: while decoding JSON: json: unknown field \""always_run\"""",""severity"":""warning"",""time"":""2022-09-08T18:58:25Z""}
+`
+
+
+***
+
+### What you expected to happen:
+***
+Output:
+`{""component"":""checkconfig"",""file"":""k8s.io/test-infra/prow/cmd/checkconfig/main.go:252"",""func"":""main.main"",""level"":""info"",""msg"":""checkconfig passes without any error!"",""severity"":""info"",""time"":""2022-09-08T17:44:36Z""}`
+***
+
+### How to reproduce it (as minimally and precisely as possible):
+***
+run the following in the root locally
+
+`docker run -i --rm -v ""${PWD}:${PWD}"" -w ""${PWD}"" gcr.io/k8s-prow/checkconfig:v20220905-8bcebc6376 --config-path=config/prow/config.yaml --job-config-path=config/jobs --plugin-config=config/prow/plugins.yaml --strict --warnings=mismatched-tide-lenient --warnings=tide-strict-branch --warnings=needs-ok-to-test --warnings=validate-owners --warnings=missing-trigger --warnings=validate-urls --warnings=unknown-fields --warnings=duplicate-job-refs --warnings=unknown-fields-all`
+
+***
+
+### Proposed Fix:
+
+Issue occurs when `periodics:` is passed either `always_run:` or `optional:` ... the errors resolve themselves if the following diff is applied
+***
+```diff
+
+index b45dd90a45..5e626166fa 100644
+--- a/config/jobs/kubernetes-sigs/sig-windows/soak-tests.yaml
++++ b/config/jobs/kubernetes-sigs/sig-windows/soak-tests.yaml
+@@ -86,8 +86,6 @@ periodics:
+ - interval: 24h
+ name: periodic-soak-tests-capz-windows-2019
+ decorate: true
+- always_run: false
+- optional: true
+ decoration_config:
+ timeout: 8h
+ path_alias: k8s.io/perf-tests
+
+
+```
+
+***
+
+/sig testing
+
+",0,error unmarshaling config jobs kubernetes sigs sig windows soak tests yaml error unmarshaling json while decoding json json unknown field always run what happened prow s tool is responsible for presubmit validation of prowjob configuration it has an optional check for unknown fields in prowjob config that can be enabled with warnings unknown fields all as a result it is possible to add configuration that is not recognized by prow causing it to be silently ignored common examples of this are typos in a field name and incorrect indentation it would be great to enable this check across the repo but before we can we must address all existing unknown fields warnings unknown fields all with warnings unknown fields all run locally with the following docker run i rm v pwd pwd w pwd gcr io prow checkconfig config path config prow config yaml job config path config jobs plugin config config prow plugins yaml strict warnings mismatched tide lenient warnings tide strict branch warnings needs ok to test warnings validate owners warnings missing trigger warnings validate urls warnings unknown fields warnings duplicate job refs warnings unknown fields all returns error component checkconfig file io test infra prow cmd checkconfig main go func main reportwarning level warning msg error unmarshaling config jobs kubernetes sigs sig windows soak tests yaml error unmarshaling json while decoding json json unknown field always run severity warning time what you expected to happen output component checkconfig file io test infra prow cmd checkconfig main go func main main level info msg checkconfig passes without any error severity info time how to reproduce it as minimally and precisely as possible run the following in the root locally docker run i rm v pwd pwd w pwd gcr io prow checkconfig config path config prow config yaml job config path config jobs plugin config config prow plugins yaml strict warnings mismatched tide lenient warnings tide strict branch warnings needs ok to test warnings validate owners warnings missing trigger warnings validate urls warnings unknown fields warnings duplicate job refs warnings unknown fields all proposed fix issue occurs when periodics is passed either always run or optional the errors resolve themselves if the following diff is applied diff index a config jobs kubernetes sigs sig windows soak tests yaml b config jobs kubernetes sigs sig windows soak tests yaml periodics interval name periodic soak tests capz windows decorate true always run false optional true decoration config timeout path alias io perf tests sig testing ,0
+89504,25819711712.0,IssuesEvent,2022-12-12 08:40:18,gitpod-io/gitpod,https://api.github.com/repos/gitpod-io/gitpod,closed,Prebuild triggered on project creation does not get reported as usage,type: bug feature: prebuilds feature: teams and projects,"Unlike prebuilds triggered by a commit webhook, the prebuild triggered after new project creation is not recorded in d_b_usage.
+
+**NOTE** This may be moot if we stop auto-triggering prebuilds on new projects. (#15281)
+
+Example workspace instance below - looks like the startedTime is missing
+
+```json
+{
+ ""id"": ""26a24057-cd7b-4e03-bdcb-28190a69fa3c"",
+ ""workspaceId"": ""jldec-teamwowproject-l9n8921gofu"",
+ ""creationTime"": ""2022-12-10T10:41:17.513Z"",
+ ""startedTime"": """",
+ ""stoppedTime"": ""2022-12-10T10:43:14.344Z"",
+ ""lastHeartbeat"": """",
+ ""ideUrl"": ""https://jldec-teamwowproject-l9n8921gofu.ws-eu78.gitpod.io"",
+ ""status_old"": null,
+ ""workspaceImage"": ""eu.gcr.io/gitpod-dev/workspace-images:3537a65c39f3373206fe0e6bdf38ae4bc221068731c57ddce7a798aa4a91b15a"",
+ ""region"": ""eu78"",
+ ""deployedTime"": """",
+ ""workspaceBaseImage"": """",
+ ""_lastModified"": ""2022-12-10 10:43:14.349798 UTC"",
+ ""status"": ""{\""repo\"": {\""branch\"": \""main\"", \""latestCommit\"": \""e7e873a35f6264ccbb27f0a929596f48d81b5e9c\"", \""totalUntrackedFiles\"": 0, \""totalUncommitedFiles\"": 0, \""totalUnpushedCommits\"": 0}, \""phase\"": \""stopped\"", \""nodeIp\"": \""10.10.0.29\"", \""message\"": \""\"", \""podName\"": \""prebuild-26a24057-cd7b-4e03-bdcb-28190a69fa3c\"", \""timeout\"": \""30m0s\"", \""version\"": 109488963190786, \""nodeName\"": \""headless-ws-eu78-pool-z4xg\"", \""conditions\"": {\""failed\"": \""\"", \""timeout\"": \""\"", \""deployed\"": false, \""pullingImages\"": false, \""stoppedByRequest\"": false, \""headlessTaskFailed\"": \""\""}, \""ownerToken\"": \""xxx\"", \""exposedPorts\"": []}"",
+ ""phase"": ""stopped"",
+ ""deleted"": ""0"",
+ ""phasePersisted"": ""stopped"",
+ ""configuration"": ""{\""ideImage\"":\""eu.gcr.io/gitpod-core-dev/build/ide/code:commit-8335b0de46d748b9d12119bc7cbdf8554a9e121c\"",\""ideImageLayers\"":[],\""supervisorImage\"":\""eu.gcr.io/gitpod-core-dev/build/supervisor:commit-478a75e744a642d9b764de37cfae655bc8b29dd5\"",\""ideConfig\"":{\""useLatest\"":false},\""featureFlags\"":[\""workspace_psi\"",\""workspace_class_limiting\""]}"",
+ ""stoppingTime"": ""2022-12-10T10:43:11.505Z"",
+ ""imageBuildInfo"": null,
+ ""workspaceClass"": ""g1-standard"",
+ ""usageAttributionId"": ""team:b3ddef51-43cf-4fb3-ad0b-3edbd4e57c96""
+}
+```
+
+",1.0,"Prebuild triggered on project creation does not get reported as usage - Unlike prebuilds triggered by a commit webhook, the prebuild triggered after new project creation is not recorded in d_b_usage.
+
+**NOTE** This may be moot if we stop auto-triggering prebuilds on new projects. (#15281)
+
+Example workspace instance below - looks like the startedTime is missing
+
+```json
+{
+ ""id"": ""26a24057-cd7b-4e03-bdcb-28190a69fa3c"",
+ ""workspaceId"": ""jldec-teamwowproject-l9n8921gofu"",
+ ""creationTime"": ""2022-12-10T10:41:17.513Z"",
+ ""startedTime"": """",
+ ""stoppedTime"": ""2022-12-10T10:43:14.344Z"",
+ ""lastHeartbeat"": """",
+ ""ideUrl"": ""https://jldec-teamwowproject-l9n8921gofu.ws-eu78.gitpod.io"",
+ ""status_old"": null,
+ ""workspaceImage"": ""eu.gcr.io/gitpod-dev/workspace-images:3537a65c39f3373206fe0e6bdf38ae4bc221068731c57ddce7a798aa4a91b15a"",
+ ""region"": ""eu78"",
+ ""deployedTime"": """",
+ ""workspaceBaseImage"": """",
+ ""_lastModified"": ""2022-12-10 10:43:14.349798 UTC"",
+ ""status"": ""{\""repo\"": {\""branch\"": \""main\"", \""latestCommit\"": \""e7e873a35f6264ccbb27f0a929596f48d81b5e9c\"", \""totalUntrackedFiles\"": 0, \""totalUncommitedFiles\"": 0, \""totalUnpushedCommits\"": 0}, \""phase\"": \""stopped\"", \""nodeIp\"": \""10.10.0.29\"", \""message\"": \""\"", \""podName\"": \""prebuild-26a24057-cd7b-4e03-bdcb-28190a69fa3c\"", \""timeout\"": \""30m0s\"", \""version\"": 109488963190786, \""nodeName\"": \""headless-ws-eu78-pool-z4xg\"", \""conditions\"": {\""failed\"": \""\"", \""timeout\"": \""\"", \""deployed\"": false, \""pullingImages\"": false, \""stoppedByRequest\"": false, \""headlessTaskFailed\"": \""\""}, \""ownerToken\"": \""xxx\"", \""exposedPorts\"": []}"",
+ ""phase"": ""stopped"",
+ ""deleted"": ""0"",
+ ""phasePersisted"": ""stopped"",
+ ""configuration"": ""{\""ideImage\"":\""eu.gcr.io/gitpod-core-dev/build/ide/code:commit-8335b0de46d748b9d12119bc7cbdf8554a9e121c\"",\""ideImageLayers\"":[],\""supervisorImage\"":\""eu.gcr.io/gitpod-core-dev/build/supervisor:commit-478a75e744a642d9b764de37cfae655bc8b29dd5\"",\""ideConfig\"":{\""useLatest\"":false},\""featureFlags\"":[\""workspace_psi\"",\""workspace_class_limiting\""]}"",
+ ""stoppingTime"": ""2022-12-10T10:43:11.505Z"",
+ ""imageBuildInfo"": null,
+ ""workspaceClass"": ""g1-standard"",
+ ""usageAttributionId"": ""team:b3ddef51-43cf-4fb3-ad0b-3edbd4e57c96""
+}
+```
+
+",0,prebuild triggered on project creation does not get reported as usage unlike prebuilds triggered by a commit webhook the prebuild triggered after new project creation is not recorded in d b usage note this may be moot if we stop auto triggering prebuilds on new projects example workspace instance below looks like the startedtime is missing json id bdcb workspaceid jldec teamwowproject creationtime startedtime stoppedtime lastheartbeat ideurl status old null workspaceimage eu gcr io gitpod dev workspace images region deployedtime workspacebaseimage lastmodified utc status repo branch main latestcommit totaluntrackedfiles totaluncommitedfiles totalunpushedcommits phase stopped nodeip message podname prebuild bdcb timeout version nodename headless ws pool conditions failed timeout deployed false pullingimages false stoppedbyrequest false headlesstaskfailed ownertoken xxx exposedports phase stopped deleted phasepersisted stopped configuration ideimage eu gcr io gitpod core dev build ide code commit ideimagelayers supervisorimage eu gcr io gitpod core dev build supervisor commit ideconfig uselatest false featureflags stoppingtime imagebuildinfo null workspaceclass standard usageattributionid team ,0
+524,2509578930.0,IssuesEvent,2015-01-13 14:57:48,aspnet/Diagnostics,https://api.github.com/repos/aspnet/Diagnostics,closed,Special rendering for compilation errors,1 - Ready enhancement needs design,"The error page should have special rendering for compilation errors.
+
+
+",1.0,"Special rendering for compilation errors - The error page should have special rendering for compilation errors.
+
+
+",0,special rendering for compilation errors the error page should have special rendering for compilation errors huboard order custom state ,0
+610,10682533483.0,IssuesEvent,2019-10-22 05:48:38,data2health/website,https://api.github.com/repos/data2health/website,closed,Add link outs to each person's name,about us page high priority (easy fix) people workgroup,"https://ctsa.ncats.nih.gov/cd2h/about-us/
+https://ctsa.ncats.nih.gov/cd2h/workgroup/people-expertise-attribution/
+
+Add link outs on each person’s name, like to their institutional/personal webpages on the about us page and on the people workgroup page
+
+related to #9 ",1.0,"Add link outs to each person's name - https://ctsa.ncats.nih.gov/cd2h/about-us/
+https://ctsa.ncats.nih.gov/cd2h/workgroup/people-expertise-attribution/
+
+Add link outs on each person’s name, like to their institutional/personal webpages on the about us page and on the people workgroup page
+
+related to #9 ",1,add link outs to each person s name add link outs on each person’s name like to their institutional personal webpages on the about us page and on the people workgroup page related to ,1
+279170,24204340923.0,IssuesEvent,2022-09-25 02:03:10,WordPress/gutenberg,https://api.github.com/repos/WordPress/gutenberg,closed,[Flaky Test] Shows the quick inserter when the block contains non-navigation specific blocks,[Status] In Progress [Status] Stale [Block] Navigation [Type] Flaky Test,"
+**Flaky test detected. This is an auto-generated issue by GitHub Actions. Please do NOT edit this manually.**
+
+## Test title
+Shows the quick inserter when the block contains non-navigation specific blocks
+
+## Test path
+`specs/editor/blocks/navigation.test.js`
+
+## Errors
+
+ Test passed after 1 failed attempt on try/cover-min-height-placeholder-resizer.
+
+ Test passed after 1 failed attempt on update/secure-settings-to-use-value-func.
+
+ Test passed after 1 failed attempt on fix/ux-when-link-ui-has-empty-href.
+
+ Test passed after 1 failed attempt on mukeshpanchal27-patch-32434.
+
+ Test passed after 1 failed attempt on rn/add/modified-style-hook.
+
+ Test passed after 1 failed attempt on update/secure-styles-duotone.
+
+ Test passed after 1 failed attempt on rnmobile/embed-block-paste-url.
+
+ Test passed after 1 failed attempt on add/writing-flow-in-between-click-redirect.
+
+ Test passed after 1 failed attempt on add/comment-avatar.
+
+ Test passed after 1 failed attempt on trunk.
+
+ Test passed after 1 failed attempt on mobile/issue/3055-show-disabled-block-reason.
+
+ Test passed after 1 failed attempt on add/drag-and-inserter-motion-effects.
+
+ Test passed after 1 failed attempt on feature/test-richtext.
+
+ Test passed after 1 failed attempt on update/react-native-instructions.
+
+ Test passed after 1 failed attempt on polish/card.
+
+ Test passed after 1 failed attempt on try/add-breadcrumbs-block.
+
+ Test passed after 1 failed attempt on refactor/components-toggle-group-control-option-button.
+
+ Test passed after 1 failed attempt on try/font-size-control-changes.
+
+ Test passed after 1 failed attempt on update/font-family-block-supports-use-classes.
+
+ Test passed after 1 failed attempt on trunk.
+
+ Test passed after 1 failed attempt on feat/components-custom-select-control-followup-docs.
+
+ Test passed after 1 failed attempt on fix/cover-opacity-bug.
+
+ Test passed after 1 failed attempt on docs/add-block-json-schema.
+
+ Test passed after 1 failed attempt on trunk.
+
+ Test passed after 1 failed attempt on rnmobile/fix/e2e-ios-tests.
+
+ Test passed after 1 failed attempt on try/list-view-window.
+
+ Test passed after 1 failed attempt on rnmobile/try/creating-context-constants-for-gallery-block.
+
+ Test passed after 1 failed attempt on try/combine-block-supports-style-engine-stores.
+
+ Test passed after 1 failed attempt on trunk.
+
+ Test passed after 1 failed attempt on remove/lodash-blocks-api-templates.
+
+
+
+ Test passed after 1 failed attempt on remove/lodash-blocks-api-serializer.
+
+
+```
+ ● Navigation › Shows the quick inserter when the block contains non-navigation specific blocks
+
+ TimeoutError: waiting for function failed: timeout 30000ms exceeded
+
+ at new WaitTask (../../node_modules/puppeteer-core/src/common/DOMWorld.ts:813:28)
+ at DOMWorld.waitForFunction (../../node_modules/puppeteer-core/src/common/DOMWorld.ts:728:22)
+ at Frame.waitForFunction (../../node_modules/puppeteer-core/src/common/FrameManager.ts:1351:28)
+ at Page.waitForFunction (../../node_modules/puppeteer-core/src/common/Page.ts:3346:29)
+ at waitForInserterCloseAndContentFocus (../e2e-test-utils/build/@wordpress/e2e-test-utils/src/inserter.js:92:8)
+ at insertBlock (../e2e-test-utils/build/@wordpress/e2e-test-utils/src/inserter.js:171:8)
+ at Object. (specs/editor/blocks/navigation.test.js:930:3)
+
+```
+
+
+",1.0,"[Flaky Test] Shows the quick inserter when the block contains non-navigation specific blocks -
+**Flaky test detected. This is an auto-generated issue by GitHub Actions. Please do NOT edit this manually.**
+
+## Test title
+Shows the quick inserter when the block contains non-navigation specific blocks
+
+## Test path
+`specs/editor/blocks/navigation.test.js`
+
+## Errors
+
+ Test passed after 1 failed attempt on try/cover-min-height-placeholder-resizer.
+
+ Test passed after 1 failed attempt on update/secure-settings-to-use-value-func.
+
+ Test passed after 1 failed attempt on fix/ux-when-link-ui-has-empty-href.
+
+ Test passed after 1 failed attempt on mukeshpanchal27-patch-32434.
+
+ Test passed after 1 failed attempt on rn/add/modified-style-hook.
+
+ Test passed after 1 failed attempt on update/secure-styles-duotone.
+
+ Test passed after 1 failed attempt on rnmobile/embed-block-paste-url.
+
+ Test passed after 1 failed attempt on add/writing-flow-in-between-click-redirect.
+
+ Test passed after 1 failed attempt on add/comment-avatar.
+
+ Test passed after 1 failed attempt on trunk.
+
+ Test passed after 1 failed attempt on mobile/issue/3055-show-disabled-block-reason.
+
+ Test passed after 1 failed attempt on add/drag-and-inserter-motion-effects.
+
+ Test passed after 1 failed attempt on feature/test-richtext.
+
+ Test passed after 1 failed attempt on update/react-native-instructions.
+
+ Test passed after 1 failed attempt on polish/card.
+
+ Test passed after 1 failed attempt on try/add-breadcrumbs-block.
+
+ Test passed after 1 failed attempt on refactor/components-toggle-group-control-option-button.
+
+ Test passed after 1 failed attempt on try/font-size-control-changes.
+
+ Test passed after 1 failed attempt on update/font-family-block-supports-use-classes.
+
+ Test passed after 1 failed attempt on trunk.
+
+ Test passed after 1 failed attempt on feat/components-custom-select-control-followup-docs.
+
+ Test passed after 1 failed attempt on fix/cover-opacity-bug.
+
+ Test passed after 1 failed attempt on docs/add-block-json-schema.
+
+ Test passed after 1 failed attempt on trunk.
+
+ Test passed after 1 failed attempt on rnmobile/fix/e2e-ios-tests.
+
+ Test passed after 1 failed attempt on try/list-view-window.
+
+ Test passed after 1 failed attempt on rnmobile/try/creating-context-constants-for-gallery-block.
+
+ Test passed after 1 failed attempt on try/combine-block-supports-style-engine-stores.
+
+ Test passed after 1 failed attempt on trunk.
+
+ Test passed after 1 failed attempt on remove/lodash-blocks-api-templates.
+
+
+
+ Test passed after 1 failed attempt on remove/lodash-blocks-api-serializer.
+
+
+```
+ ● Navigation › Shows the quick inserter when the block contains non-navigation specific blocks
+
+ TimeoutError: waiting for function failed: timeout 30000ms exceeded
+
+ at new WaitTask (../../node_modules/puppeteer-core/src/common/DOMWorld.ts:813:28)
+ at DOMWorld.waitForFunction (../../node_modules/puppeteer-core/src/common/DOMWorld.ts:728:22)
+ at Frame.waitForFunction (../../node_modules/puppeteer-core/src/common/FrameManager.ts:1351:28)
+ at Page.waitForFunction (../../node_modules/puppeteer-core/src/common/Page.ts:3346:29)
+ at waitForInserterCloseAndContentFocus (../e2e-test-utils/build/@wordpress/e2e-test-utils/src/inserter.js:92:8)
+ at insertBlock (../e2e-test-utils/build/@wordpress/e2e-test-utils/src/inserter.js:171:8)
+ at Object. (specs/editor/blocks/navigation.test.js:930:3)
+
+```
+
+
+",0, shows the quick inserter when the block contains non navigation specific blocks flaky test detected this is an auto generated issue by github actions please do not edit this manually test title shows the quick inserter when the block contains non navigation specific blocks test path specs editor blocks navigation test js errors test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on a href ● navigation › shows the quick inserter when the block contains non navigation specific blocks timeouterror waiting for function failed timeout exceeded at new waittask node modules puppeteer core src common domworld ts at domworld waitforfunction node modules puppeteer core src common domworld ts at frame waitforfunction node modules puppeteer core src common framemanager ts at page waitforfunction node modules puppeteer core src common page ts at waitforinsertercloseandcontentfocus test utils build wordpress test utils src inserter js at insertblock test utils build wordpress test utils src inserter js at object specs editor blocks navigation test js ,0
+278097,21058059650.0,IssuesEvent,2022-04-01 06:40:59,jiale-c/ped,https://api.github.com/repos/jiale-c/ped,opened,Choice of words,severity.Low type.DocumentationBug,"sentences such as the Note in the find section is hard for non technical people to understand. "" consecutive whitespaces in the query string "". query string should be defined.
+
+
+
+",1.0,"Choice of words - sentences such as the Note in the find section is hard for non technical people to understand. "" consecutive whitespaces in the query string "". query string should be defined.
+
+
+
+",0,choice of words sentences such as the note in the find section is hard for non technical people to understand consecutive whitespaces in the query string query string should be defined ,0
+92963,26824488180.0,IssuesEvent,2023-02-02 11:53:26,NixOS/nixpkgs,https://api.github.com/repos/NixOS/nixpkgs,closed,Unable to build gamescope,0.kind: build failure,"### Steps To Reproduce
+Steps to reproduce the behavior:
+1. Attempt to install gamescope as seen here: https://github.com/Krutonium/My_Unified_NixOS_Config/blob/adadd3a3ec12201a08bf1f5c6a5ce2a0a17b72f9/packages/steam.nix
+
+### Build log
+
+[notbuilding.txt](https://github.com/NixOS/nixpkgs/files/9478436/notbuilding.txt)
+
+
+
+### Additional context
+It seems to be unable to locate either cmake or pkgconfig.
+
+### Notify maintainers
+
+@nrdxp
+
+### Metadata
+Please run `nix-shell -p nix-info --run ""nix-info -m""` and paste the result.
+
+```console
+[user@system:~]$ nix-shell -p nix-info --run ""nix-info -m""
+ - system: `""x86_64-linux""`
+ - host os: `Linux 5.15.63, NixOS, 22.05 (Quokka), 22.05.20220901.b82ccaf`
+ - multi-user?: `yes`
+ - sandbox: `yes`
+ - version: `nix-env (Nix) 2.8.1`
+ - channels(root): `""nixos""`
+ - nixpkgs: `/nix/var/nix/profiles/per-user/root/channels/nixos`
+
+
+```
+",1.0,"Unable to build gamescope - ### Steps To Reproduce
+Steps to reproduce the behavior:
+1. Attempt to install gamescope as seen here: https://github.com/Krutonium/My_Unified_NixOS_Config/blob/adadd3a3ec12201a08bf1f5c6a5ce2a0a17b72f9/packages/steam.nix
+
+### Build log
+
+[notbuilding.txt](https://github.com/NixOS/nixpkgs/files/9478436/notbuilding.txt)
+
+
+
+### Additional context
+It seems to be unable to locate either cmake or pkgconfig.
+
+### Notify maintainers
+
+@nrdxp
+
+### Metadata
+Please run `nix-shell -p nix-info --run ""nix-info -m""` and paste the result.
+
+```console
+[user@system:~]$ nix-shell -p nix-info --run ""nix-info -m""
+ - system: `""x86_64-linux""`
+ - host os: `Linux 5.15.63, NixOS, 22.05 (Quokka), 22.05.20220901.b82ccaf`
+ - multi-user?: `yes`
+ - sandbox: `yes`
+ - version: `nix-env (Nix) 2.8.1`
+ - channels(root): `""nixos""`
+ - nixpkgs: `/nix/var/nix/profiles/per-user/root/channels/nixos`
+
+
+```
+",0,unable to build gamescope steps to reproduce steps to reproduce the behavior attempt to install gamescope as seen here build log additional context it seems to be unable to locate either cmake or pkgconfig notify maintainers please people who are in the meta maintainers list of the offending package or module if in doubt check git blame for whoever last touched something nrdxp metadata please run nix shell p nix info run nix info m and paste the result console nix shell p nix info run nix info m system linux host os linux nixos quokka multi user yes sandbox yes version nix env nix channels root nixos nixpkgs nix var nix profiles per user root channels nixos ,0
+45474,2933864423.0,IssuesEvent,2015-06-30 02:55:16,tokenly/swapbot,https://api.github.com/repos/tokenly/swapbot,closed,Prevent duplicate XChain sends,bug high priority,"If XChain receives a send attempt, but times out, it might return an error to swapbot but go ahead and send the token anyway.
+
+Since swapbot doesn't think the token sent, it tries and sends again at the next opportunity. This can result in multiple tokens being sent instead of just one.
+
+Instead, we need to move the swap into a permanent failure state so that this error can be dealt with manually.
+
+",1.0,"Prevent duplicate XChain sends - If XChain receives a send attempt, but times out, it might return an error to swapbot but go ahead and send the token anyway.
+
+Since swapbot doesn't think the token sent, it tries and sends again at the next opportunity. This can result in multiple tokens being sent instead of just one.
+
+Instead, we need to move the swap into a permanent failure state so that this error can be dealt with manually.
+
+",0,prevent duplicate xchain sends if xchain receives a send attempt but times out it might return an error to swapbot but go ahead and send the token anyway since swapbot doesn t think the token sent it tries and sends again at the next opportunity this can result in multiple tokens being sent instead of just one instead we need to move the swap into a permanent failure state so that this error can be dealt with manually ,0
+279309,21155316243.0,IssuesEvent,2022-04-07 02:08:00,ICEI-PUC-Minas-PMV-ADS/pmv-ads-2022-1-e3-proj-mov-t1-aluguel-de-bicicleta,https://api.github.com/repos/ICEI-PUC-Minas-PMV-ADS/pmv-ads-2022-1-e3-proj-mov-t1-aluguel-de-bicicleta,opened,[GERENCIAR USUÁRIOS - CADASTRAR],documentation enhancement,"- [ ] DEV - Desenvolvimento da funcionalidade
+- [ ] DOC011 - Registro de teste de usabilidade
+- [ ] DOC009 - Registro de teste de software
+- [ ] DOC007 - Documentar Funcionalidade",1.0,"[GERENCIAR USUÁRIOS - CADASTRAR] - - [ ] DEV - Desenvolvimento da funcionalidade
+- [ ] DOC011 - Registro de teste de usabilidade
+- [ ] DOC009 - Registro de teste de software
+- [ ] DOC007 - Documentar Funcionalidade",0, dev desenvolvimento da funcionalidade registro de teste de usabilidade registro de teste de software documentar funcionalidade,0
+1119,27123958625.0,IssuesEvent,2023-02-16 02:34:46,restincode/restincode,https://api.github.com/repos/restincode/restincode,opened,Bruce Esquibel (Dr. Ripco),People Add Person Needs Review,"Please fill out as much information as you can. No fields are required, but the more you can provide the better.
+
+**General Info**
+* First name: Bruce
+* Last name: Esquibel
+* Handle: Dr. Ripco
+* Birth Year:
+* Death Year: 2023
+* Link to Obituary:
+* Group Affiliations:
+* URL to main photo (or attach to Issue):
+* Description of person and/or activities:
+* Facebook memorial group URL:
+
+**Social Media Links**
+* Twitter:
+* Github:
+* LinkedIn:
+* Facebook:
+* Other:
+
+**Contributions**
+Repeat as many times as needed
+
+* Project name:
+* Project URL:
+* Project Description:
+
+**Photo Gallery**
+* URL(s) to additional photos:
+
+https://twitter.com/textfiles/status/1625878219114127362 / https://archive.ph/cfrlm
+https://twitter.com/DethVeggie/status/1626016692613812224 / https://archive.ph/n1rxl",1.0,"Bruce Esquibel (Dr. Ripco) - Please fill out as much information as you can. No fields are required, but the more you can provide the better.
+
+**General Info**
+* First name: Bruce
+* Last name: Esquibel
+* Handle: Dr. Ripco
+* Birth Year:
+* Death Year: 2023
+* Link to Obituary:
+* Group Affiliations:
+* URL to main photo (or attach to Issue):
+* Description of person and/or activities:
+* Facebook memorial group URL:
+
+**Social Media Links**
+* Twitter:
+* Github:
+* LinkedIn:
+* Facebook:
+* Other:
+
+**Contributions**
+Repeat as many times as needed
+
+* Project name:
+* Project URL:
+* Project Description:
+
+**Photo Gallery**
+* URL(s) to additional photos:
+
+https://twitter.com/textfiles/status/1625878219114127362 / https://archive.ph/cfrlm
+https://twitter.com/DethVeggie/status/1626016692613812224 / https://archive.ph/n1rxl",1,bruce esquibel dr ripco please fill out as much information as you can no fields are required but the more you can provide the better general info first name bruce last name esquibel handle dr ripco birth year death year link to obituary group affiliations url to main photo or attach to issue description of person and or activities facebook memorial group url social media links twitter github linkedin facebook other contributions repeat as many times as needed project name project url project description photo gallery url s to additional photos ,1
+204182,15422013930.0,IssuesEvent,2021-03-05 13:51:25,Plant-for-the-Planet-org/treemapper,https://api.github.com/repos/Plant-for-the-Planet-org/treemapper,closed,SyntaxError in [native code]:0,1.0.0 bug bugsnag testflight,"## Error in Tree Mapper
+
+**SyntaxError** in **[native code]:0**
+JSON Parse error: Unexpected identifier ""undefined""
+
+[View on Bugsnag](https://app.bugsnag.com/plant-for-the-planet/tree-mapper/errors/602292c2258b4800171b36b7?event_id=6037920b0072eff327440000&i=gh&m=ci)
+
+## Stacktrace
+
+ [native code]:0 - parse
+
+[View full stacktrace](https://app.bugsnag.com/plant-for-the-planet/tree-mapper/errors/602292c2258b4800171b36b7?event_id=6037920b0072eff327440000&i=gh&m=ci)
+
+*Created automatically via Bugsnag*",1.0,"SyntaxError in [native code]:0 - ## Error in Tree Mapper
+
+**SyntaxError** in **[native code]:0**
+JSON Parse error: Unexpected identifier ""undefined""
+
+[View on Bugsnag](https://app.bugsnag.com/plant-for-the-planet/tree-mapper/errors/602292c2258b4800171b36b7?event_id=6037920b0072eff327440000&i=gh&m=ci)
+
+## Stacktrace
+
+ [native code]:0 - parse
+
+[View full stacktrace](https://app.bugsnag.com/plant-for-the-planet/tree-mapper/errors/602292c2258b4800171b36b7?event_id=6037920b0072eff327440000&i=gh&m=ci)
+
+*Created automatically via Bugsnag*",0,syntaxerror in error in tree mapper syntaxerror in json parse error unexpected identifier undefined stacktrace parse created automatically via bugsnag ,0
+48,2660239186.0,IssuesEvent,2015-03-19 04:22:12,fskreuz/Tree-Trail,https://api.github.com/repos/fskreuz/Tree-Trail,closed,View Pending Badges Bug,QA people! Test these!,"Pre: Login as admin
+1. Go to Contacts page
+2. Click dropdown sa Welcome
+3. When View Pending Badges clicked, not redirecting to map and showing pending badges.
+
+Pre: Login as admin
+1. Go to About Page
+2. Click dropdown sa Welcome
+3. When View Pending Badges clicked, not redirecting to map and showing pending badges.
+
+Pre: Login as admin
+1. Go to About Page
+2. Click dropdown sa Welcome
+3. When View Pending Badges clicked, not redirecting to map and showing pending badges.
+
+Pre: Login as admin
+1. Go to View Feedbacks
+2. Click dropdown sa Welcome
+3. When View Pending Badges clicked, not redirecting to map and showing pending badges.
+
+Pre: Login as admin
+1. Go to Manage Announcements
+2. Click dropdown sa Welcome
+3. When View Pending Badges clicked, not redirecting to map and showing pending badges.
+
+Pre: Login as admin
+1. Go to Contact Page
+2. Click dropdown sa Welcome
+3. When View Pending Badges clicked, not redirecting to map and showing pending badges.
+
+Pre: Login as admin
+1. Go to Announcements Page
+2. Click dropdown sa Welcome
+3. When View Pending Badges clicked, not redirecting to map and showing pending badges.",1.0,"View Pending Badges Bug - Pre: Login as admin
+1. Go to Contacts page
+2. Click dropdown sa Welcome
+3. When View Pending Badges clicked, not redirecting to map and showing pending badges.
+
+Pre: Login as admin
+1. Go to About Page
+2. Click dropdown sa Welcome
+3. When View Pending Badges clicked, not redirecting to map and showing pending badges.
+
+Pre: Login as admin
+1. Go to About Page
+2. Click dropdown sa Welcome
+3. When View Pending Badges clicked, not redirecting to map and showing pending badges.
+
+Pre: Login as admin
+1. Go to View Feedbacks
+2. Click dropdown sa Welcome
+3. When View Pending Badges clicked, not redirecting to map and showing pending badges.
+
+Pre: Login as admin
+1. Go to Manage Announcements
+2. Click dropdown sa Welcome
+3. When View Pending Badges clicked, not redirecting to map and showing pending badges.
+
+Pre: Login as admin
+1. Go to Contact Page
+2. Click dropdown sa Welcome
+3. When View Pending Badges clicked, not redirecting to map and showing pending badges.
+
+Pre: Login as admin
+1. Go to Announcements Page
+2. Click dropdown sa Welcome
+3. When View Pending Badges clicked, not redirecting to map and showing pending badges.",1,view pending badges bug pre login as admin go to contacts page click dropdown sa welcome when view pending badges clicked not redirecting to map and showing pending badges pre login as admin go to about page click dropdown sa welcome when view pending badges clicked not redirecting to map and showing pending badges pre login as admin go to about page click dropdown sa welcome when view pending badges clicked not redirecting to map and showing pending badges pre login as admin go to view feedbacks click dropdown sa welcome when view pending badges clicked not redirecting to map and showing pending badges pre login as admin go to manage announcements click dropdown sa welcome when view pending badges clicked not redirecting to map and showing pending badges pre login as admin go to contact page click dropdown sa welcome when view pending badges clicked not redirecting to map and showing pending badges pre login as admin go to announcements page click dropdown sa welcome when view pending badges clicked not redirecting to map and showing pending badges ,1
+371449,10966842567.0,IssuesEvent,2019-11-28 08:12:52,matteopaoli/WeirdGame,https://api.github.com/repos/matteopaoli/WeirdGame,closed,Implement Foundation Sites SCSS,High Priority frontend,"Add Zurb Foundation using scss
+
+DOCS: https://foundation.zurb.com/sites/docs/sass.html
+
+Please avoid to import the whole framework. Only the grid system (xy-grid) is needed for now (check in the link above 'Adjusting CSS Output')",1.0,"Implement Foundation Sites SCSS - Add Zurb Foundation using scss
+
+DOCS: https://foundation.zurb.com/sites/docs/sass.html
+
+Please avoid to import the whole framework. Only the grid system (xy-grid) is needed for now (check in the link above 'Adjusting CSS Output')",0,implement foundation sites scss add zurb foundation using scss docs please avoid to import the whole framework only the grid system xy grid is needed for now check in the link above adjusting css output ,0
+334644,29931301979.0,IssuesEvent,2023-06-22 09:39:44,pytorch/pytorch,https://api.github.com/repos/pytorch/pytorch,opened,DISABLED test_benchmark_choice_fail_in_subproc (__main__.TestDoBench),triaged module: flaky-tests skipped module: inductor,"Platforms: rocm
+
+This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/failure/test_benchmark_choice_fail_in_subproc) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/14459220551).
+
+Over the past 72 hours, it has flakily failed in 9 workflow(s).
+
+**Debugging instructions (after clicking on the recent samples link):**
+To find relevant log snippets:
+1. Click on the workflow logs linked above
+2. Grep for `test_benchmark_choice_fail_in_subproc`
+
+
+Test file path: `inductor/test_max_autotune.py`",1.0,"DISABLED test_benchmark_choice_fail_in_subproc (__main__.TestDoBench) - Platforms: rocm
+
+This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/failure/test_benchmark_choice_fail_in_subproc) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/14459220551).
+
+Over the past 72 hours, it has flakily failed in 9 workflow(s).
+
+**Debugging instructions (after clicking on the recent samples link):**
+To find relevant log snippets:
+1. Click on the workflow logs linked above
+2. Grep for `test_benchmark_choice_fail_in_subproc`
+
+
+Test file path: `inductor/test_max_autotune.py`",0,disabled test benchmark choice fail in subproc main testdobench platforms rocm this test was disabled because it is failing in ci see and the most recent trunk over the past hours it has flakily failed in workflow s debugging instructions after clicking on the recent samples link to find relevant log snippets click on the workflow logs linked above grep for test benchmark choice fail in subproc test file path inductor test max autotune py ,0
+72018,7275700549.0,IssuesEvent,2018-02-21 14:23:53,Microsoft/vscode,https://api.github.com/repos/Microsoft/vscode,opened,Test: centered editor layout,testplan-item,"Refs: https://github.com/Microsoft/vscode/issues/15684
+
+- [ ] any os
+- [ ] any os
+
+Complexity: 3
+
+This milestone we have added centered editor layout. Verify:
+* You can enter to this layout via command palette, view menu, by entering zen mode
+* You can use the vertical sashes to control the size of the editor
+* Sashes are paired unless you hold the `alt` key while dragging them
+* Editors are automatically not centered if more than one is shown side-by-side (also for diff viewer)
+* Double click on sash returns them to default golden ratio
+* Changing sash size and centered mode are preserved across vscode sessions
+* Setting `zenMode.centerLayout` behaves as expected
+
+fyi @SrTobi i",1.0,"Test: centered editor layout - Refs: https://github.com/Microsoft/vscode/issues/15684
+
+- [ ] any os
+- [ ] any os
+
+Complexity: 3
+
+This milestone we have added centered editor layout. Verify:
+* You can enter to this layout via command palette, view menu, by entering zen mode
+* You can use the vertical sashes to control the size of the editor
+* Sashes are paired unless you hold the `alt` key while dragging them
+* Editors are automatically not centered if more than one is shown side-by-side (also for diff viewer)
+* Double click on sash returns them to default golden ratio
+* Changing sash size and centered mode are preserved across vscode sessions
+* Setting `zenMode.centerLayout` behaves as expected
+
+fyi @SrTobi i",0,test centered editor layout refs any os any os complexity this milestone we have added centered editor layout verify you can enter to this layout via command palette view menu by entering zen mode you can use the vertical sashes to control the size of the editor sashes are paired unless you hold the alt key while dragging them editors are automatically not centered if more than one is shown side by side also for diff viewer double click on sash returns them to default golden ratio changing sash size and centered mode are preserved across vscode sessions setting zenmode centerlayout behaves as expected fyi srtobi i,0
+837,15689175029.0,IssuesEvent,2021-03-25 15:24:37,PostHog/posthog,https://api.github.com/repos/PostHog/posthog,closed,Show cohorts person belongs to in user page,cohorts enhancement people,"## Is your feature request related to a problem?
+
+I'm browsing user data in posthog and I'd like to figure out whether a user has correctly been added to a cohort
+
+## Describe the solution you'd like
+
+Show tab ""cohorts"" next to ""properties"" on people page.
+
+## Describe alternatives you've considered
+
+Going through all cohorts.
+
+## Additional context
+
+
+
+#### *Thank you* for your feature request – we love each and every one!
+",1.0,"Show cohorts person belongs to in user page - ## Is your feature request related to a problem?
+
+I'm browsing user data in posthog and I'd like to figure out whether a user has correctly been added to a cohort
+
+## Describe the solution you'd like
+
+Show tab ""cohorts"" next to ""properties"" on people page.
+
+## Describe alternatives you've considered
+
+Going through all cohorts.
+
+## Additional context
+
+
+
+#### *Thank you* for your feature request – we love each and every one!
+",1,show cohorts person belongs to in user page is your feature request related to a problem i m browsing user data in posthog and i d like to figure out whether a user has correctly been added to a cohort describe the solution you d like show tab cohorts next to properties on people page describe alternatives you ve considered going through all cohorts additional context thank you for your feature request – we love each and every one ,1
+43,2658520867.0,IssuesEvent,2015-03-18 16:01:23,fskreuz/Tree-Trail,https://api.github.com/repos/fskreuz/Tree-Trail,closed,Note the default password in Superadmin,QA people! Test these!,The superadmin should see a note that the default password is [what ever it is] everytime superadmin adds a user.,1.0,Note the default password in Superadmin - The superadmin should see a note that the default password is [what ever it is] everytime superadmin adds a user.,1,note the default password in superadmin the superadmin should see a note that the default password is everytime superadmin adds a user ,1
+414,7703837749.0,IssuesEvent,2018-05-21 09:53:35,DrewAPicture/ensemble,https://api.github.com/repos/DrewAPicture/ensemble,opened,Term query support for Unit Directors,::People ::Taxonomy Enhancement,"Like contests and venues, Unit Directors (users) are connected with various taxonomies, but most notably the Units taxonomy. It would be nice if there were a way to bridge between directors and the terms assigned to them, even if it means going outside of the core API pathways to accomplish it.",1.0,"Term query support for Unit Directors - Like contests and venues, Unit Directors (users) are connected with various taxonomies, but most notably the Units taxonomy. It would be nice if there were a way to bridge between directors and the terms assigned to them, even if it means going outside of the core API pathways to accomplish it.",1,term query support for unit directors like contests and venues unit directors users are connected with various taxonomies but most notably the units taxonomy it would be nice if there were a way to bridge between directors and the terms assigned to them even if it means going outside of the core api pathways to accomplish it ,1
+40907,10220881027.0,IssuesEvent,2019-08-15 22:55:46,idaholab/moose,https://api.github.com/repos/idaholab/moose,opened,Compiler warnings about MOOSE dummies,C: MOOSE P: minor T: defect,"## Bug Description
+Every Registry.h macro invocation generates an 11 line long compiler warning for me, e.g.:
+
+```
+/home/roystgnr/git/moose-fresh/framework/build/header_symlinks/Registry.h:47:28: warning: ‘dummyvar_for_registering_obj_SmoothMeshGenerator19’ defined but not used [-Wunused-variable]
+ static char combineNames(dummyvar_for_registering_obj_##classname, __LINE__) = \
+ ^
+/home/roystgnr/git/moose-fresh/framework/build/header_symlinks/Registry.h:17:29: note: in definition of macro ‘combineNames1’
+ #define combineNames1(X, Y) X##Y
+ ^
+/home/roystgnr/git/moose-fresh/framework/build/header_symlinks/Registry.h:47:15: note: in expansion of macro ‘combineNames’
+ static char combineNames(dummyvar_for_registering_obj_##classname, __LINE__) = \
+ ^
+/home/roystgnr/git/moose-fresh/framework/src/meshgenerators/SmoothMeshGenerator.C:19:1: note: in expansion of macro ‘registerMooseObject’
+ registerMooseObject(""MooseApp"", SmoothMeshGenerator);
+```
+
+## Steps to Reproduce
+Build MOOSE with the -Wunused-variable compiler flag or with one of the many flags that incorporates it.
+
+## Impact
+This is just an annoyance, but it's a voluminous one: tens of thousands of lines of warning text in a typical MOOSE build.
+
+What's the point of the static char here?",1.0,"Compiler warnings about MOOSE dummies - ## Bug Description
+Every Registry.h macro invocation generates an 11 line long compiler warning for me, e.g.:
+
+```
+/home/roystgnr/git/moose-fresh/framework/build/header_symlinks/Registry.h:47:28: warning: ‘dummyvar_for_registering_obj_SmoothMeshGenerator19’ defined but not used [-Wunused-variable]
+ static char combineNames(dummyvar_for_registering_obj_##classname, __LINE__) = \
+ ^
+/home/roystgnr/git/moose-fresh/framework/build/header_symlinks/Registry.h:17:29: note: in definition of macro ‘combineNames1’
+ #define combineNames1(X, Y) X##Y
+ ^
+/home/roystgnr/git/moose-fresh/framework/build/header_symlinks/Registry.h:47:15: note: in expansion of macro ‘combineNames’
+ static char combineNames(dummyvar_for_registering_obj_##classname, __LINE__) = \
+ ^
+/home/roystgnr/git/moose-fresh/framework/src/meshgenerators/SmoothMeshGenerator.C:19:1: note: in expansion of macro ‘registerMooseObject’
+ registerMooseObject(""MooseApp"", SmoothMeshGenerator);
+```
+
+## Steps to Reproduce
+Build MOOSE with the -Wunused-variable compiler flag or with one of the many flags that incorporates it.
+
+## Impact
+This is just an annoyance, but it's a voluminous one: tens of thousands of lines of warning text in a typical MOOSE build.
+
+What's the point of the static char here?",0,compiler warnings about moose dummies bug description every registry h macro invocation generates an line long compiler warning for me e g home roystgnr git moose fresh framework build header symlinks registry h warning ‘dummyvar for registering obj ’ defined but not used static char combinenames dummyvar for registering obj classname line home roystgnr git moose fresh framework build header symlinks registry h note in definition of macro ‘ ’ define x y x y home roystgnr git moose fresh framework build header symlinks registry h note in expansion of macro ‘combinenames’ static char combinenames dummyvar for registering obj classname line home roystgnr git moose fresh framework src meshgenerators smoothmeshgenerator c note in expansion of macro ‘registermooseobject’ registermooseobject mooseapp smoothmeshgenerator steps to reproduce build moose with the wunused variable compiler flag or with one of the many flags that incorporates it impact this is just an annoyance but it s a voluminous one tens of thousands of lines of warning text in a typical moose build what s the point of the static char here ,0
+634860,20374345605.0,IssuesEvent,2022-02-21 14:16:26,teamforus/general,https://api.github.com/repos/teamforus/general,closed,Make fund formula user configurable,Priority: Should have Epic Scope: Small project-31 project-139 Todo: Move to Productboard,"Learn more about change requests here: https://bit.ly/39CWeEE
+
+### Requested by:
+-
+
+### Change description
+Right now we set the fund formula directly in the database (for each child_nth you get €300 worth of budget) - this should be configurable by the user.
+
+Question: can we add it to the fund settings, so that it doesn't need to be changed in service window, and it is more transparant to the sponsor what is done?
+
+@maxvisser for me this is a bit of a blackbox, could you mock something up in figma?
+
+## Figma
+https://www.figma.com/file/99YwnGU4PrgP1zCFudAxmz/Configurable-fund-formula-s?node-id=0%3A1",1.0,"Make fund formula user configurable - Learn more about change requests here: https://bit.ly/39CWeEE
+
+### Requested by:
+-
+
+### Change description
+Right now we set the fund formula directly in the database (for each child_nth you get €300 worth of budget) - this should be configurable by the user.
+
+Question: can we add it to the fund settings, so that it doesn't need to be changed in service window, and it is more transparant to the sponsor what is done?
+
+@maxvisser for me this is a bit of a blackbox, could you mock something up in figma?
+
+## Figma
+https://www.figma.com/file/99YwnGU4PrgP1zCFudAxmz/Configurable-fund-formula-s?node-id=0%3A1",0,make fund formula user configurable learn more about change requests here requested by change description right now we set the fund formula directly in the database for each child nth you get € worth of budget this should be configurable by the user question can we add it to the fund settings so that it doesn t need to be changed in service window and it is more transparant to the sponsor what is done maxvisser for me this is a bit of a blackbox could you mock something up in figma figma ,0
+548,9661247285.0,IssuesEvent,2019-05-20 17:30:21,planningcenter/developers,https://api.github.com/repos/planningcenter/developers,closed,Household Wholesale Edit,People,"**Related Applications**
+People
+**Is your feature request related to a problem? Please describe.**
+When updating a person from a household, each person must be updated for a 'wholesale' edit. Planning Center has wholesale functionality when changing home contact information of a person.
+**Describe the solution you'd like**
+A flag be sent with a person indicating their home contact information should be updated throughout the household. A household id could also be sent along.
+**Describe alternatives you've considered**
+The household could be updated on our side, and then each person's update could be sent to Planning Center.
+**Additional context**
+This functionality exists within Planning Center already. Just hoping to branch it out to the API as well.
+
+",1.0,"Household Wholesale Edit - **Related Applications**
+People
+**Is your feature request related to a problem? Please describe.**
+When updating a person from a household, each person must be updated for a 'wholesale' edit. Planning Center has wholesale functionality when changing home contact information of a person.
+**Describe the solution you'd like**
+A flag be sent with a person indicating their home contact information should be updated throughout the household. A household id could also be sent along.
+**Describe alternatives you've considered**
+The household could be updated on our side, and then each person's update could be sent to Planning Center.
+**Additional context**
+This functionality exists within Planning Center already. Just hoping to branch it out to the API as well.
+
+",1,household wholesale edit related applications people is your feature request related to a problem please describe when updating a person from a household each person must be updated for a wholesale edit planning center has wholesale functionality when changing home contact information of a person describe the solution you d like a flag be sent with a person indicating their home contact information should be updated throughout the household a household id could also be sent along describe alternatives you ve considered the household could be updated on our side and then each person s update could be sent to planning center additional context this functionality exists within planning center already just hoping to branch it out to the api as well ,1
+921,19577505467.0,IssuesEvent,2022-01-04 16:54:27,openstates/issues,https://api.github.com/repos/openstates/issues,closed,CT Legislators: addresses need to be updated,component:people-data type:bug,"State: CT
+Short Description: One vacancy unlisted, issues with incorrect office numbers and phone numbers, however the big issue stems with the way that the CT general assembly site operates (Additional Data). Because the site itself doesn't have pages specifically for legislators and instead outsources to their personal sites or their pages on CT dems/reps sites, information acquisition is less uniform.
+
+**Missing or Incorrect legislators:**
+
+updated: fixed
+
+**Data Issues:**
+
+Many Legislators have the wrong office number/no office number listed in address. Occurs fairly commonly but here are a couple of examples:
+-Alex Kasser (wrong office number listed): https://openstates.org/person/alex-kasser-2LeifQ7tIcgP9cjALFHPV7/, http://www.senatedems.ct.gov/Kasser-contact
+-Mae Flexer (wrong office number listed): https://openstates.org/person/mae-flexer-3dNpiKCcpGQA1Npd3eK4Hi/ , http://www.senatedems.ct.gov/Flexer-contact
+-Rick Lopes (no office number listed): https://openstates.org/person/rick-lopes-5m615OSi3qJt5Tj5AeiaCh/ , http://www.senatedems.ct.gov/lopes-contact
+-Norman Needleman (wrong office number listed): https://openstates.org/person/norman-needleman-VGrjGptCOu7fc4UDdIGAh/ , http://www.senatedems.ct.gov/needleman-contact
+
+Some legislators have the wrong phone number listed (doesn't appear to be systemic):
+-Catherine F. Abercrombie: https://openstates.org/person/catherine-f-abercrombie-44UMlw4hDG9STN1IPVOmce/ , http://www.housedems.ct.gov/Abercrombie
+
+ Some legislators don't have a number listed at all (doesn't appear to be systemic):
+-Amy Morrin Bello: https://openstates.org/person/amy-morrin-bello-7lNycyX1nIFArL15X8KdSo/ , https://www.housedems.ct.gov/morrinbello
+
+**Additional Data:**
+
+CT general assembly lists info for legislative and press aides for Senators (open states doesn't). The phone number listed for most Senators is that of their aides (their numbers aren't listed on CT general assembly), however Open States doesn't specifically clarify this (ex. Alex Kasser: https://openstates.org/person/alex-kasser-2LeifQ7tIcgP9cjALFHPV7/)
+
+IMPORTANT TO NOTE: It seems the CT general assembly site isn't perfectly functioning! Many of the district offices for legislators aren't listed on CT general assembly but are on Open states (ex. Derek Slap: https://openstates.org/person/derek-slap-4y0JUdYxvhVL5aMq6Y0hAE/ and Catherine A. Olsten: https://openstates.org/person/catherine-a-osten-5JlJpA7QTKRx0KSPnjOYO6/) and some senators don't even have pages that are linked to their name in the member list (ex. Kevin D. Witkos and Gary A. Winfield). Furthermore, CT general assembly doesn't have their own pages for legislators, rather links to their information on either CT senate dems/republicans. ",1.0,"CT Legislators: addresses need to be updated - State: CT
+Short Description: One vacancy unlisted, issues with incorrect office numbers and phone numbers, however the big issue stems with the way that the CT general assembly site operates (Additional Data). Because the site itself doesn't have pages specifically for legislators and instead outsources to their personal sites or their pages on CT dems/reps sites, information acquisition is less uniform.
+
+**Missing or Incorrect legislators:**
+
+updated: fixed
+
+**Data Issues:**
+
+Many Legislators have the wrong office number/no office number listed in address. Occurs fairly commonly but here are a couple of examples:
+-Alex Kasser (wrong office number listed): https://openstates.org/person/alex-kasser-2LeifQ7tIcgP9cjALFHPV7/, http://www.senatedems.ct.gov/Kasser-contact
+-Mae Flexer (wrong office number listed): https://openstates.org/person/mae-flexer-3dNpiKCcpGQA1Npd3eK4Hi/ , http://www.senatedems.ct.gov/Flexer-contact
+-Rick Lopes (no office number listed): https://openstates.org/person/rick-lopes-5m615OSi3qJt5Tj5AeiaCh/ , http://www.senatedems.ct.gov/lopes-contact
+-Norman Needleman (wrong office number listed): https://openstates.org/person/norman-needleman-VGrjGptCOu7fc4UDdIGAh/ , http://www.senatedems.ct.gov/needleman-contact
+
+Some legislators have the wrong phone number listed (doesn't appear to be systemic):
+-Catherine F. Abercrombie: https://openstates.org/person/catherine-f-abercrombie-44UMlw4hDG9STN1IPVOmce/ , http://www.housedems.ct.gov/Abercrombie
+
+ Some legislators don't have a number listed at all (doesn't appear to be systemic):
+-Amy Morrin Bello: https://openstates.org/person/amy-morrin-bello-7lNycyX1nIFArL15X8KdSo/ , https://www.housedems.ct.gov/morrinbello
+
+**Additional Data:**
+
+CT general assembly lists info for legislative and press aides for Senators (open states doesn't). The phone number listed for most Senators is that of their aides (their numbers aren't listed on CT general assembly), however Open States doesn't specifically clarify this (ex. Alex Kasser: https://openstates.org/person/alex-kasser-2LeifQ7tIcgP9cjALFHPV7/)
+
+IMPORTANT TO NOTE: It seems the CT general assembly site isn't perfectly functioning! Many of the district offices for legislators aren't listed on CT general assembly but are on Open states (ex. Derek Slap: https://openstates.org/person/derek-slap-4y0JUdYxvhVL5aMq6Y0hAE/ and Catherine A. Olsten: https://openstates.org/person/catherine-a-osten-5JlJpA7QTKRx0KSPnjOYO6/) and some senators don't even have pages that are linked to their name in the member list (ex. Kevin D. Witkos and Gary A. Winfield). Furthermore, CT general assembly doesn't have their own pages for legislators, rather links to their information on either CT senate dems/republicans. ",1,ct legislators addresses need to be updated state ct short description one vacancy unlisted issues with incorrect office numbers and phone numbers however the big issue stems with the way that the ct general assembly site operates additional data because the site itself doesn t have pages specifically for legislators and instead outsources to their personal sites or their pages on ct dems reps sites information acquisition is less uniform missing or incorrect legislators updated fixed data issues many legislators have the wrong office number no office number listed in address occurs fairly commonly but here are a couple of examples alex kasser wrong office number listed mae flexer wrong office number listed rick lopes no office number listed norman needleman wrong office number listed some legislators have the wrong phone number listed doesn t appear to be systemic catherine f abercrombie some legislators don t have a number listed at all doesn t appear to be systemic amy morrin bello additional data ct general assembly lists info for legislative and press aides for senators open states doesn t the phone number listed for most senators is that of their aides their numbers aren t listed on ct general assembly however open states doesn t specifically clarify this ex alex kasser important to note it seems the ct general assembly site isn t perfectly functioning many of the district offices for legislators aren t listed on ct general assembly but are on open states ex derek slap and catherine a olsten and some senators don t even have pages that are linked to their name in the member list ex kevin d witkos and gary a winfield furthermore ct general assembly doesn t have their own pages for legislators rather links to their information on either ct senate dems republicans ,1
+802,14915975265.0,IssuesEvent,2021-01-22 17:29:44,quasarframework/quasar,https://api.github.com/repos/quasarframework/quasar,closed,resumable uploader protocol ,:people_holding_hands: team-pick feature request,I will be great if QUploader can resume and pause uploading file or possibility of using it with other api like `tus.io`,1.0,resumable uploader protocol - I will be great if QUploader can resume and pause uploading file or possibility of using it with other api like `tus.io`,1,resumable uploader protocol i will be great if quploader can resume and pause uploading file or possibility of using it with other api like tus io ,1
+109211,13753914219.0,IssuesEvent,2020-10-06 16:13:33,elastic/kibana,https://api.github.com/repos/elastic/kibana,closed,[Onboarding] Kibana landing page,Team:Core UI design loe:multi-week,"There is always something new in Kibana and customers are very often unclear what these new offerings can do for them. Further, customers coming from Cloud express an intent for their cluster type (e.g. Observability, Ent Search, etc.) and would benefit from landing on an appropriately themed page.
+
+In order to provide a better onboarding experience, we would like to provide landing pages for each solution.
+
+The new Kibana design will also group products by solution, of which, one will likely contain all the analytical or 'traditional' Kibana apps such as Discover, Visualize, and Dashboard.
+
+This issue covers the landing page for the Kibana 'solution', which covers the following products:
+- Discover, Visualize (incl. Lens), Dashboard, Canvas, Maps, ML, and Graph.
+
+### Design
+
+Technically, there are two pages that make up the landing 'page' experience:
+1. A quick 'get started' version when no data is present
+2. A more full, guided overview page
+
+##### Get started (no data)
+
+
+##### Overview page (has data)
+
+
+#### Mockups and prototype
+- [Figma mockups](https://www.figma.com/file/537SsoT2eCzuW8MTDJps2w/In-Progress-Home-and-Kibana-Landing-Pages-MVP?node-id=0%3A1)
+- [Figma clickable prototype](https://www.figma.com/proto/537SsoT2eCzuW8MTDJps2w/In-Progress-Home-and-Kibana-Landing-Pages-MVP?node-id=4377%3A119179&viewport=4642%2C-940%2C0.25&scaling=min-zoom)
+",1.0,"[Onboarding] Kibana landing page - There is always something new in Kibana and customers are very often unclear what these new offerings can do for them. Further, customers coming from Cloud express an intent for their cluster type (e.g. Observability, Ent Search, etc.) and would benefit from landing on an appropriately themed page.
+
+In order to provide a better onboarding experience, we would like to provide landing pages for each solution.
+
+The new Kibana design will also group products by solution, of which, one will likely contain all the analytical or 'traditional' Kibana apps such as Discover, Visualize, and Dashboard.
+
+This issue covers the landing page for the Kibana 'solution', which covers the following products:
+- Discover, Visualize (incl. Lens), Dashboard, Canvas, Maps, ML, and Graph.
+
+### Design
+
+Technically, there are two pages that make up the landing 'page' experience:
+1. A quick 'get started' version when no data is present
+2. A more full, guided overview page
+
+##### Get started (no data)
+
+
+##### Overview page (has data)
+
+
+#### Mockups and prototype
+- [Figma mockups](https://www.figma.com/file/537SsoT2eCzuW8MTDJps2w/In-Progress-Home-and-Kibana-Landing-Pages-MVP?node-id=0%3A1)
+- [Figma clickable prototype](https://www.figma.com/proto/537SsoT2eCzuW8MTDJps2w/In-Progress-Home-and-Kibana-Landing-Pages-MVP?node-id=4377%3A119179&viewport=4642%2C-940%2C0.25&scaling=min-zoom)
+",0, kibana landing page there is always something new in kibana and customers are very often unclear what these new offerings can do for them further customers coming from cloud express an intent for their cluster type e g observability ent search etc and would benefit from landing on an appropriately themed page in order to provide a better onboarding experience we would like to provide landing pages for each solution the new kibana design will also group products by solution of which one will likely contain all the analytical or traditional kibana apps such as discover visualize and dashboard this issue covers the landing page for the kibana solution which covers the following products discover visualize incl lens dashboard canvas maps ml and graph design technically there are two pages that make up the landing page experience a quick get started version when no data is present a more full guided overview page get started no data img width alt no data src overview page has data img width alt learning progress src mockups and prototype ,0
+1124,27264842025.0,IssuesEvent,2023-02-22 17:14:35,openstates/issues,https://api.github.com/repos/openstates/issues,closed,New NM Committee Scraper,good first issue component:people-data good first scraper,"### We need a new committee scraper written in [spatula](https://jamesturk.github.io/spatula/) for New Mexico.
+
+It should scrape these pages for the [Senate committees](https://www.nmlegis.gov/Committee/Senate_Standing), [House committees](https://www.nmlegis.gov/Committee/House_Standing), and [Joint Interim committees](https://www.nmlegis.gov/Committee/Interim) to get:
+- **name**
+ - “Small Business”, “Agriculture”, etc.
+- **chamber**
+ - “upper”, “lower”, or “legislature” (when joint, etc.)
+- **classification**
+ - ex: ""committee"", ""subcommittee""
+- **parent**
+ - *only if it is a subcommittee* - scraper should have a way to determine this
+ - ex: parent=""Natural Resources"" when *sub*committee name=""Forestry""
+- **sources**
+ - each added using `add_source()` method on instance of `ScrapeCommittee` type object
+ - ex: home page for list of committees, specific page for that committee, etc.
+- **members**
+ - - each added using `add_member()`method on instance of `ScrapeCommittee` type object
+ - *Attributes:*
+ - **name**: “Jane Doe”, “John Smith”, etc.
+ - **role** (where applicable): “Chair”, “Ranking Member”, etc.
+
+#### [Here is helpful documentation](https://docs.openstates.org/contributing/writing-a-committee-scraper/) for writing a committee scraper
+
+### Useful scrapers for reference
+An [old non-spatula-based scraper](https://github.com/openstates/openstates-scrapers/blob/main/scrapers/nm/committees.py) that may be helpful in understanding how we have captured the data in the past.
+
+A [spatula-based scraper](https://github.com/openstates/openstates-scrapers/blob/main/scrapers_next/mo/committees.py) that also parses HTML pages, for reference.
+- As in the other spatula scraper, you will need to set the session variable separately (rather than depend on the `__init__.py`).
+
+### Other useful resources
+You can reference the [object models for committee scrapers](https://github.com/openstates/openstates-core/blob/main/openstates/models/committees.py) in the Open States Core repo.
+
+Further documentation on running spatula scrapers in the command line can be found [here](https://docs.openstates.org/contributing/scrapers/#running-spatula-scrapers).",1.0,"New NM Committee Scraper - ### We need a new committee scraper written in [spatula](https://jamesturk.github.io/spatula/) for New Mexico.
+
+It should scrape these pages for the [Senate committees](https://www.nmlegis.gov/Committee/Senate_Standing), [House committees](https://www.nmlegis.gov/Committee/House_Standing), and [Joint Interim committees](https://www.nmlegis.gov/Committee/Interim) to get:
+- **name**
+ - “Small Business”, “Agriculture”, etc.
+- **chamber**
+ - “upper”, “lower”, or “legislature” (when joint, etc.)
+- **classification**
+ - ex: ""committee"", ""subcommittee""
+- **parent**
+ - *only if it is a subcommittee* - scraper should have a way to determine this
+ - ex: parent=""Natural Resources"" when *sub*committee name=""Forestry""
+- **sources**
+ - each added using `add_source()` method on instance of `ScrapeCommittee` type object
+ - ex: home page for list of committees, specific page for that committee, etc.
+- **members**
+ - - each added using `add_member()`method on instance of `ScrapeCommittee` type object
+ - *Attributes:*
+ - **name**: “Jane Doe”, “John Smith”, etc.
+ - **role** (where applicable): “Chair”, “Ranking Member”, etc.
+
+#### [Here is helpful documentation](https://docs.openstates.org/contributing/writing-a-committee-scraper/) for writing a committee scraper
+
+### Useful scrapers for reference
+An [old non-spatula-based scraper](https://github.com/openstates/openstates-scrapers/blob/main/scrapers/nm/committees.py) that may be helpful in understanding how we have captured the data in the past.
+
+A [spatula-based scraper](https://github.com/openstates/openstates-scrapers/blob/main/scrapers_next/mo/committees.py) that also parses HTML pages, for reference.
+- As in the other spatula scraper, you will need to set the session variable separately (rather than depend on the `__init__.py`).
+
+### Other useful resources
+You can reference the [object models for committee scrapers](https://github.com/openstates/openstates-core/blob/main/openstates/models/committees.py) in the Open States Core repo.
+
+Further documentation on running spatula scrapers in the command line can be found [here](https://docs.openstates.org/contributing/scrapers/#running-spatula-scrapers).",1,new nm committee scraper we need a new committee scraper written in for new mexico it should scrape these pages for the and to get name “small business” “agriculture” etc chamber “upper” “lower” or “legislature” when joint etc classification ex committee subcommittee parent only if it is a subcommittee scraper should have a way to determine this ex parent natural resources when sub committee name forestry sources each added using add source method on instance of scrapecommittee type object ex home page for list of committees specific page for that committee etc members each added using add member method on instance of scrapecommittee type object attributes name “jane doe” “john smith” etc role where applicable “chair” “ranking member” etc for writing a committee scraper useful scrapers for reference an that may be helpful in understanding how we have captured the data in the past a that also parses html pages for reference as in the other spatula scraper you will need to set the session variable separately rather than depend on the init py other useful resources you can reference the in the open states core repo further documentation on running spatula scrapers in the command line can be found ,1
+45,2658711484.0,IssuesEvent,2015-03-18 16:59:01,fskreuz/Tree-Trail,https://api.github.com/repos/fskreuz/Tree-Trail,closed,Resize Picture in badges (if naa),QA people! Test these!,Resize picture in badges to a minimum size like 200 x 150. Keep aspect ratio. Basta placeholder for resized picture is 200 x 150 ang size.,1.0,Resize Picture in badges (if naa) - Resize picture in badges to a minimum size like 200 x 150. Keep aspect ratio. Basta placeholder for resized picture is 200 x 150 ang size.,1,resize picture in badges if naa resize picture in badges to a minimum size like x keep aspect ratio basta placeholder for resized picture is x ang size ,1
+150504,23675962460.0,IssuesEvent,2022-08-28 04:28:08,microsoft/pyright,https://api.github.com/repos/microsoft/pyright,closed,Reflected operand methods not recognized in numeric subclasses,as designed,"**Describe the bug**
+Pyright does not recognize reflected (swapped) operand methods in numeric subclasses.
+
+**Code**
+```python
+from __future__ import annotations
+
+
+class intSubclass(int):
+ def __add__(self, __x: int) -> intSubclass:
+ return intSubclass(super().__add__(__x))
+
+ def __radd__(self, __x: int) -> intSubclass:
+ return intSubclass(super().__radd__(__x))
+
+
+int_var: intSubclass = intSubclass(123)
+
+int_var = int_var + 123
+print(""int_var __add__ "", type(int_var))
+# int_var __add__
+
+# Expression of type ""int"" cannot be assigned to declared type ""intSubclass""
+# ""int"" is incompatible with ""intSubclass""
+int_var = 123 + int_var
+print(""int_var __radd__"", type(int_var))
+# int_var __radd__
+
+
+class floatSubclass(float):
+ def __add__(self, __x: float) -> floatSubclass:
+ return floatSubclass(super().__add__(__x))
+
+ def __radd__(self, __x: float) -> floatSubclass:
+ return floatSubclass(super().__radd__(__x))
+
+
+float_var: floatSubclass = floatSubclass(1.23)
+
+float_var = float_var + 1.23
+print(""float_var __add__ "", type(float_var))
+# float_var __add__
+
+# Expression of type ""float"" cannot be assigned to declared type ""floatSubclass""
+# ""float"" is incompatible with ""floatSubclass""
+float_var = 1.23 + float_var
+print(""float_var __radd__"", type(float_var))
+# float_var __radd__
+```
+
+**VS Code extension**
+Pyright 1.1.266 through Pylance 2022.8.30
+
+**Additional context**
+Python 3.10.5
+",1.0,"Reflected operand methods not recognized in numeric subclasses - **Describe the bug**
+Pyright does not recognize reflected (swapped) operand methods in numeric subclasses.
+
+**Code**
+```python
+from __future__ import annotations
+
+
+class intSubclass(int):
+ def __add__(self, __x: int) -> intSubclass:
+ return intSubclass(super().__add__(__x))
+
+ def __radd__(self, __x: int) -> intSubclass:
+ return intSubclass(super().__radd__(__x))
+
+
+int_var: intSubclass = intSubclass(123)
+
+int_var = int_var + 123
+print(""int_var __add__ "", type(int_var))
+# int_var __add__
+
+# Expression of type ""int"" cannot be assigned to declared type ""intSubclass""
+# ""int"" is incompatible with ""intSubclass""
+int_var = 123 + int_var
+print(""int_var __radd__"", type(int_var))
+# int_var __radd__
+
+
+class floatSubclass(float):
+ def __add__(self, __x: float) -> floatSubclass:
+ return floatSubclass(super().__add__(__x))
+
+ def __radd__(self, __x: float) -> floatSubclass:
+ return floatSubclass(super().__radd__(__x))
+
+
+float_var: floatSubclass = floatSubclass(1.23)
+
+float_var = float_var + 1.23
+print(""float_var __add__ "", type(float_var))
+# float_var __add__
+
+# Expression of type ""float"" cannot be assigned to declared type ""floatSubclass""
+# ""float"" is incompatible with ""floatSubclass""
+float_var = 1.23 + float_var
+print(""float_var __radd__"", type(float_var))
+# float_var __radd__
+```
+
+**VS Code extension**
+Pyright 1.1.266 through Pylance 2022.8.30
+
+**Additional context**
+Python 3.10.5
+",0,reflected operand methods not recognized in numeric subclasses describe the bug pyright does not recognize reflected swapped operand methods in numeric subclasses code python from future import annotations class intsubclass int def add self x int intsubclass return intsubclass super add x def radd self x int intsubclass return intsubclass super radd x int var intsubclass intsubclass int var int var print int var add type int var int var add expression of type int cannot be assigned to declared type intsubclass int is incompatible with intsubclass int var int var print int var radd type int var int var radd class floatsubclass float def add self x float floatsubclass return floatsubclass super add x def radd self x float floatsubclass return floatsubclass super radd x float var floatsubclass floatsubclass float var float var print float var add type float var float var add expression of type float cannot be assigned to declared type floatsubclass float is incompatible with floatsubclass float var float var print float var radd type float var float var radd vs code extension pyright through pylance additional context python ,0
+136317,12706535782.0,IssuesEvent,2020-06-23 07:22:21,RedHatInsights/insights-operator-controller,https://api.github.com/repos/RedHatInsights/insights-operator-controller,closed,Add links to GitHub pages as a badge into README.md,documentation enhancement,Add links to GitHub pages as a badge into README.md,1.0,Add links to GitHub pages as a badge into README.md - Add links to GitHub pages as a badge into README.md,0,add links to github pages as a badge into readme md add links to github pages as a badge into readme md,0
+1034,25115490080.0,IssuesEvent,2022-11-09 01:20:23,jongfeel/BookReview,https://api.github.com/repos/jongfeel/BookReview,closed,3부 7장 협조를 얻는 방법,2022 How to Win Friends & Influence People,"### 7장 협조를 얻는 방법
+
+당신은 다른 사람이 건네주는 생각을 그대로 받아들이기보다는 스스로 발견한 생각을 더 믿지 않는가? 만일 그렇다면 당신의 의견을 다른 사람에게 강요하는 것은 잘못된 판단이 아닐까? 제안을 하고, 다른 사람으로 하여금 철저하게 생각해 보도록 만드는 게 더 현명한 일이 아닐까?
+
+---
+
+규칙 7: 다른 사람으로 하여금 스스로 생각해 냈다고 여기도록 만들어라.
+Let the other fellow feel that the idea is his.",1.0,"3부 7장 협조를 얻는 방법 - ### 7장 협조를 얻는 방법
+
+당신은 다른 사람이 건네주는 생각을 그대로 받아들이기보다는 스스로 발견한 생각을 더 믿지 않는가? 만일 그렇다면 당신의 의견을 다른 사람에게 강요하는 것은 잘못된 판단이 아닐까? 제안을 하고, 다른 사람으로 하여금 철저하게 생각해 보도록 만드는 게 더 현명한 일이 아닐까?
+
+---
+
+규칙 7: 다른 사람으로 하여금 스스로 생각해 냈다고 여기도록 만들어라.
+Let the other fellow feel that the idea is his.",1, 협조를 얻는 방법 협조를 얻는 방법 당신은 다른 사람이 건네주는 생각을 그대로 받아들이기보다는 스스로 발견한 생각을 더 믿지 않는가 만일 그렇다면 당신의 의견을 다른 사람에게 강요하는 것은 잘못된 판단이 아닐까 제안을 하고 다른 사람으로 하여금 철저하게 생각해 보도록 만드는 게 더 현명한 일이 아닐까 규칙 다른 사람으로 하여금 스스로 생각해 냈다고 여기도록 만들어라 let the other fellow feel that the idea is his ,1
+70846,9450287176.0,IssuesEvent,2019-04-16 05:56:22,paul-buerkner/brms,https://api.github.com/repos/paul-buerkner/brms,closed,Help files for fitted and predict,documentation,"In the help file for fitted, first paragraph, it says ""By definition, these predictions have smaller variance than the response predictions performed by the predict method. This is because the measurement error is not incorporated."" Is ""measurement error"" correct? I thought it was the residual error that is not included (perhaps I am misunderstanding). There is a similar statement in predict.",1.0,"Help files for fitted and predict - In the help file for fitted, first paragraph, it says ""By definition, these predictions have smaller variance than the response predictions performed by the predict method. This is because the measurement error is not incorporated."" Is ""measurement error"" correct? I thought it was the residual error that is not included (perhaps I am misunderstanding). There is a similar statement in predict.",0,help files for fitted and predict in the help file for fitted first paragraph it says by definition these predictions have smaller variance than the response predictions performed by the predict method this is because the measurement error is not incorporated is measurement error correct i thought it was the residual error that is not included perhaps i am misunderstanding there is a similar statement in predict ,0
+148,3696988800.0,IssuesEvent,2016-02-27 10:35:34,friendlycode/strategic_plan,https://api.github.com/repos/friendlycode/strategic_plan,opened,Diversity,Membership People & Partnerships,Increase percent of membership who identifies with minority and underrepresented groups in 2016,1.0,Diversity - Increase percent of membership who identifies with minority and underrepresented groups in 2016,1,diversity increase percent of membership who identifies with minority and underrepresented groups in ,1
+188,4020461574.0,IssuesEvent,2016-05-16 18:30:03,wordpress-mobile/WordPress-iOS,https://api.github.com/repos/wordpress-mobile/WordPress-iOS,closed,People Management: Delete Users,People Management [Type] Enhancement,"#### Details:
+Admins should be allowed to Delete Users. Let's replicate Calypso's UX:
+
+- On delete, Posts should be reassigned to another user (to be picked).
+- Calypso's User deletion can be found [here](https://github.com/Automattic/wp-calypso/blob/12a27beda288923d5186807f1157f04389f54c06/client/my-sites/people/delete-user/index.jsx), and the actual call, [here](https://github.com/Automattic/wp-calypso/blob/12a27beda288923d5186807f1157f04389f54c06/client/lib/users/actions.js#L59)
+- A user cannot delete himself (sure!)
+
+",1.0,"People Management: Delete Users - #### Details:
+Admins should be allowed to Delete Users. Let's replicate Calypso's UX:
+
+- On delete, Posts should be reassigned to another user (to be picked).
+- Calypso's User deletion can be found [here](https://github.com/Automattic/wp-calypso/blob/12a27beda288923d5186807f1157f04389f54c06/client/my-sites/people/delete-user/index.jsx), and the actual call, [here](https://github.com/Automattic/wp-calypso/blob/12a27beda288923d5186807f1157f04389f54c06/client/lib/users/actions.js#L59)
+- A user cannot delete himself (sure!)
+
+",1,people management delete users details admins should be allowed to delete users let s replicate calypso s ux on delete posts should be reassigned to another user to be picked calypso s user deletion can be found and the actual call a user cannot delete himself sure ,1
+38015,15428598073.0,IssuesEvent,2021-03-06 00:13:58,MicrosoftDocs/azure-docs,https://api.github.com/repos/MicrosoftDocs/azure-docs,closed,azure-preview should be aks-preview,Pri2 container-service/svc cxp docs-experience triaged,"
+In the requirements section, you mention that azure-preview extension is required. I guess you actually mean aks-preview.
+
+
+---
+#### Document Details
+
+⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
+
+* ID: c3c187c0-27f1-11c9-41af-6c65ca36e77e
+* Version Independent ID: 249a570b-90ba-4a11-4467-320f0369ec25
+* Content: [Use Azure Active Directory pod-managed identities in Azure Kubernetes Service (Preview) - Azure Kubernetes Service](https://docs.microsoft.com/en-us/azure/aks/use-azure-ad-pod-identity)
+* Content Source: [articles/aks/use-azure-ad-pod-identity.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/aks/use-azure-ad-pod-identity.md)
+* Service: **container-service**
+* GitHub Login: @mlearned
+* Microsoft Alias: **mlearned**",1.0,"azure-preview should be aks-preview -
+In the requirements section, you mention that azure-preview extension is required. I guess you actually mean aks-preview.
+
+
+---
+#### Document Details
+
+⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
+
+* ID: c3c187c0-27f1-11c9-41af-6c65ca36e77e
+* Version Independent ID: 249a570b-90ba-4a11-4467-320f0369ec25
+* Content: [Use Azure Active Directory pod-managed identities in Azure Kubernetes Service (Preview) - Azure Kubernetes Service](https://docs.microsoft.com/en-us/azure/aks/use-azure-ad-pod-identity)
+* Content Source: [articles/aks/use-azure-ad-pod-identity.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/aks/use-azure-ad-pod-identity.md)
+* Service: **container-service**
+* GitHub Login: @mlearned
+* Microsoft Alias: **mlearned**",0,azure preview should be aks preview in the requirements section you mention that azure preview extension is required i guess you actually mean aks preview document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service container service github login mlearned microsoft alias mlearned ,0
+140566,32027593508.0,IssuesEvent,2023-09-22 09:55:35,llvm/llvm-project,https://api.github.com/repos/llvm/llvm-project,opened,Model catch object write in MIR,llvm:codegen platform:windows,"To capture the idea from https://github.com/llvm/llvm-project/pull/66988#discussion_r1333561055:
+
+Currently, the potential write to the WinEH catch object at the point of the invoke is not properly modeled in MIR. We could fix this by adding a reference to all possible catch object frame indices to the corresponding call instructions. This would make stack coloring correct out of the box.",1.0,"Model catch object write in MIR - To capture the idea from https://github.com/llvm/llvm-project/pull/66988#discussion_r1333561055:
+
+Currently, the potential write to the WinEH catch object at the point of the invoke is not properly modeled in MIR. We could fix this by adding a reference to all possible catch object frame indices to the corresponding call instructions. This would make stack coloring correct out of the box.",0,model catch object write in mir to capture the idea from currently the potential write to the wineh catch object at the point of the invoke is not properly modeled in mir we could fix this by adding a reference to all possible catch object frame indices to the corresponding call instructions this would make stack coloring correct out of the box ,0
+11879,4316889916.0,IssuesEvent,2016-07-23 00:27:18,Idrinth/github-auto-release,https://api.github.com/repos/Idrinth/github-auto-release,closed,split the script into class-files,CodeQuality,"The script currently combines a bit of procedural code as well as multiple classes in a single file.
+For easier usage in projects that would include this as a library this needs to be split into parts that work with the usual autoloaders.",1.0,"split the script into class-files - The script currently combines a bit of procedural code as well as multiple classes in a single file.
+For easier usage in projects that would include this as a library this needs to be split into parts that work with the usual autoloaders.",0,split the script into class files the script currently combines a bit of procedural code as well as multiple classes in a single file for easier usage in projects that would include this as a library this needs to be split into parts that work with the usual autoloaders ,0
+292484,8958725883.0,IssuesEvent,2019-01-27 16:37:20,NullCGT/SpliceHack,https://api.github.com/repos/NullCGT/SpliceHack,closed,The monster zaps a knife! The fAâ╛Ç hits you!,bug high priority,"A few different people have reported this. Monsters apparently will sometimes zap knives which shoot poison gas. This is by far the strangest bug that I have ever encountered, and I have absolutely no idea what is causing it.",1.0,"The monster zaps a knife! The fAâ╛Ç hits you! - A few different people have reported this. Monsters apparently will sometimes zap knives which shoot poison gas. This is by far the strangest bug that I have ever encountered, and I have absolutely no idea what is causing it.",0,the monster zaps a knife the faâ╛ç hits you a few different people have reported this monsters apparently will sometimes zap knives which shoot poison gas this is by far the strangest bug that i have ever encountered and i have absolutely no idea what is causing it ,0
+16815,2615124356.0,IssuesEvent,2015-03-01 05:52:12,chrsmith/google-api-java-client,https://api.github.com/repos/chrsmith/google-api-java-client,opened,Sample for adding Calendar by URL,auto-migrated Priority-Medium Type-Sample,"```
+Which Google API and version (e.g. Google Calendar Data API version 2)?
+Google Calendar API v3
+
+What Authentation (e.g. OAuth, OAuth 2, ClientLogin)?
+OAuth 2
+
+Java environment (e.g. Java 6, Android 2.3, App Engine)?
+Java 6, Android 2.3
+
+
+Please provide any additional information below.
+Please include a sample on how to add a calendar to a user's calendar list by
+URL. For example if the URL is a .ics file hosted on a server.
+Thank you.
+```
+
+Original issue reported on code.google.com by `harvey.d...@gmail.com` on 18 Jan 2012 at 10:37",1.0,"Sample for adding Calendar by URL - ```
+Which Google API and version (e.g. Google Calendar Data API version 2)?
+Google Calendar API v3
+
+What Authentation (e.g. OAuth, OAuth 2, ClientLogin)?
+OAuth 2
+
+Java environment (e.g. Java 6, Android 2.3, App Engine)?
+Java 6, Android 2.3
+
+
+Please provide any additional information below.
+Please include a sample on how to add a calendar to a user's calendar list by
+URL. For example if the URL is a .ics file hosted on a server.
+Thank you.
+```
+
+Original issue reported on code.google.com by `harvey.d...@gmail.com` on 18 Jan 2012 at 10:37",0,sample for adding calendar by url which google api and version e g google calendar data api version google calendar api what authentation e g oauth oauth clientlogin oauth java environment e g java android app engine java android please provide any additional information below please include a sample on how to add a calendar to a user s calendar list by url for example if the url is a ics file hosted on a server thank you original issue reported on code google com by harvey d gmail com on jan at ,0
+924,19589071712.0,IssuesEvent,2022-01-05 10:44:30,metabase/metabase,https://api.github.com/repos/metabase/metabase,closed,Frontend load issue: Adding members to groups,Type:Bug Priority:P3 .Frontend Administration/People,"**Describe the bug**
+When adding members to a group you need to go to a separate page and then return before it loads the change in the number of members.
+
+**Logs**
+Nothing in the logs
+
+**To Reproduce**
+Steps to reproduce the behavior:
+1. Go to '/admin/people/groups'
+2. Click on one of your groups
+3. Click 'Add members'
+4. Add someone to the group
+5. Click 'Groups'
+6. Error: In the members column, Groups is still listing the previous number of people in the group
+7. If you click 'People' and click back to 'Group' the page has corrected itself
+
+**Expected behavior**
+I expected to see the correct number of people in the group the first time I went onto the group page.
+
+**Screenshots**
+Screenshot the first time I click on Groups:
+
+
+Screenshot the second time I click on Groups:
+
+
+**Information about your Metabase Installation:**
+```
+{
+ ""browser-info"": {
+ ""language"": ""en-US"",
+ ""platform"": ""MacIntel"",
+ ""userAgent"": ""Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_4) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/83.0.4103.61 Safari/537.36"",
+ ""vendor"": ""Google Inc.""
+ },
+ ""system-info"": {
+ ""file.encoding"": ""UTF-8"",
+ ""java.runtime.name"": ""OpenJDK Runtime Environment"",
+ ""java.runtime.version"": ""1.8.0_242-heroku-b08"",
+ ""java.vendor"": ""Oracle Corporation"",
+ ""java.vendor.url"": ""http://java.oracle.com/"",
+ ""java.version"": ""1.8.0_242-heroku"",
+ ""java.vm.name"": ""OpenJDK 64-Bit Server VM"",
+ ""java.vm.version"": ""25.242-b08"",
+ ""os.name"": ""Linux"",
+ ""os.version"": ""4.4.0-1067-aws"",
+ ""user.language"": ""en"",
+ ""user.timezone"": ""Etc/UTC""
+ },
+ ""metabase-info"": {
+ ""databases"": [
+ ""h2""
+ ],
+ ""hosting-env"": ""heroku"",
+ ""application-database"": ""postgres"",
+ ""application-database-details"": {
+ ""database"": {
+ ""name"": ""PostgreSQL"",
+ ""version"": ""12.3 (Ubuntu 12.3-1.pgdg16.04+1)""
+ },
+ ""jdbc-driver"": {
+ ""name"": ""PostgreSQL JDBC Driver"",
+ ""version"": ""42.2.8""
+ }
+ },
+ ""run-mode"": ""prod"",
+ ""version"": {
+ ""tag"": ""v0.35.3"",
+ ""date"": ""2020-04-21"",
+ ""branch"": ""release-0.35.x"",
+ ""hash"": ""1d424cb""
+ },
+ ""settings"": {
+ ""report-timezone"": null
+ }
+ }
+}
+```
+
+**Severity**
+Minor
+",1.0,"Frontend load issue: Adding members to groups - **Describe the bug**
+When adding members to a group you need to go to a separate page and then return before it loads the change in the number of members.
+
+**Logs**
+Nothing in the logs
+
+**To Reproduce**
+Steps to reproduce the behavior:
+1. Go to '/admin/people/groups'
+2. Click on one of your groups
+3. Click 'Add members'
+4. Add someone to the group
+5. Click 'Groups'
+6. Error: In the members column, Groups is still listing the previous number of people in the group
+7. If you click 'People' and click back to 'Group' the page has corrected itself
+
+**Expected behavior**
+I expected to see the correct number of people in the group the first time I went onto the group page.
+
+**Screenshots**
+Screenshot the first time I click on Groups:
+
+
+Screenshot the second time I click on Groups:
+
+
+**Information about your Metabase Installation:**
+```
+{
+ ""browser-info"": {
+ ""language"": ""en-US"",
+ ""platform"": ""MacIntel"",
+ ""userAgent"": ""Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_4) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/83.0.4103.61 Safari/537.36"",
+ ""vendor"": ""Google Inc.""
+ },
+ ""system-info"": {
+ ""file.encoding"": ""UTF-8"",
+ ""java.runtime.name"": ""OpenJDK Runtime Environment"",
+ ""java.runtime.version"": ""1.8.0_242-heroku-b08"",
+ ""java.vendor"": ""Oracle Corporation"",
+ ""java.vendor.url"": ""http://java.oracle.com/"",
+ ""java.version"": ""1.8.0_242-heroku"",
+ ""java.vm.name"": ""OpenJDK 64-Bit Server VM"",
+ ""java.vm.version"": ""25.242-b08"",
+ ""os.name"": ""Linux"",
+ ""os.version"": ""4.4.0-1067-aws"",
+ ""user.language"": ""en"",
+ ""user.timezone"": ""Etc/UTC""
+ },
+ ""metabase-info"": {
+ ""databases"": [
+ ""h2""
+ ],
+ ""hosting-env"": ""heroku"",
+ ""application-database"": ""postgres"",
+ ""application-database-details"": {
+ ""database"": {
+ ""name"": ""PostgreSQL"",
+ ""version"": ""12.3 (Ubuntu 12.3-1.pgdg16.04+1)""
+ },
+ ""jdbc-driver"": {
+ ""name"": ""PostgreSQL JDBC Driver"",
+ ""version"": ""42.2.8""
+ }
+ },
+ ""run-mode"": ""prod"",
+ ""version"": {
+ ""tag"": ""v0.35.3"",
+ ""date"": ""2020-04-21"",
+ ""branch"": ""release-0.35.x"",
+ ""hash"": ""1d424cb""
+ },
+ ""settings"": {
+ ""report-timezone"": null
+ }
+ }
+}
+```
+
+**Severity**
+Minor
+",1,frontend load issue adding members to groups describe the bug when adding members to a group you need to go to a separate page and then return before it loads the change in the number of members logs nothing in the logs to reproduce steps to reproduce the behavior go to admin people groups click on one of your groups click add members add someone to the group click groups error in the members column groups is still listing the previous number of people in the group if you click people and click back to group the page has corrected itself expected behavior i expected to see the correct number of people in the group the first time i went onto the group page screenshots screenshot the first time i click on groups img width alt screen shot at pm src screenshot the second time i click on groups img width alt screen shot at pm src information about your metabase installation browser info language en us platform macintel useragent mozilla macintosh intel mac os x applewebkit khtml like gecko chrome safari vendor google inc system info file encoding utf java runtime name openjdk runtime environment java runtime version heroku java vendor oracle corporation java vendor url java version heroku java vm name openjdk bit server vm java vm version os name linux os version aws user language en user timezone etc utc metabase info databases hosting env heroku application database postgres application database details database name postgresql version ubuntu jdbc driver name postgresql jdbc driver version run mode prod version tag date branch release x hash settings report timezone null severity minor ,1
+20739,4627876511.0,IssuesEvent,2016-09-28 00:57:26,gavinsimpson/cocorresp,https://api.github.com/repos/gavinsimpson/cocorresp,opened,Remove all instances of \donttest in examples,documentation,"Need to remove all instances of `\donttest` from examples. In some cases this was there to stop checking of `summary()` output from Co-CA models. As I'm going to fix those to be much less verbose (#5) there should be no reason to not test example code.
+
+The following examples need to be fixed:
+
+ * [ ] `coca.Rd`",1.0,"Remove all instances of \donttest in examples - Need to remove all instances of `\donttest` from examples. In some cases this was there to stop checking of `summary()` output from Co-CA models. As I'm going to fix those to be much less verbose (#5) there should be no reason to not test example code.
+
+The following examples need to be fixed:
+
+ * [ ] `coca.Rd`",0,remove all instances of donttest in examples need to remove all instances of donttest from examples in some cases this was there to stop checking of summary output from co ca models as i m going to fix those to be much less verbose there should be no reason to not test example code the following examples need to be fixed coca rd ,0
+608,10682460287.0,IssuesEvent,2019-10-22 05:29:49,data2health/website,https://api.github.com/repos/data2health/website,closed,the project proposals (linked from the projects) seem out of date,people workgroup,"see: https://ctsa.ncats.nih.gov/cd2h/a-computable-representation-of-contributions/
+
+This lists some expected outputs, which may not be current
+It has a section for the GitHub repo, but nothing is linked",1.0,"the project proposals (linked from the projects) seem out of date - see: https://ctsa.ncats.nih.gov/cd2h/a-computable-representation-of-contributions/
+
+This lists some expected outputs, which may not be current
+It has a section for the GitHub repo, but nothing is linked",1,the project proposals linked from the projects seem out of date see this lists some expected outputs which may not be current it has a section for the github repo but nothing is linked,1
+2222,4351792507.0,IssuesEvent,2016-08-01 01:48:19,AtlasOfLivingAustralia/bie-index,https://api.github.com/repos/AtlasOfLivingAustralia/bie-index,opened,Search results are missing image URLs,webservices,"Previous BIE version of search.json supplied a number of image URL fields (smallImageUrl, thumbnailUrl, etc - need to check bie-service for full list). Current version has a single `image` field which is a UUID. The ALA pages assemble a URL but this 3rd users have no way of knowing how to do this without looking at our code (bad).
+
+So the old fields need to be put in the new service as complete URLs.",1.0,"Search results are missing image URLs - Previous BIE version of search.json supplied a number of image URL fields (smallImageUrl, thumbnailUrl, etc - need to check bie-service for full list). Current version has a single `image` field which is a UUID. The ALA pages assemble a URL but this 3rd users have no way of knowing how to do this without looking at our code (bad).
+
+So the old fields need to be put in the new service as complete URLs.",0,search results are missing image urls previous bie version of search json supplied a number of image url fields smallimageurl thumbnailurl etc need to check bie service for full list current version has a single image field which is a uuid the ala pages assemble a url but this users have no way of knowing how to do this without looking at our code bad so the old fields need to be put in the new service as complete urls ,0
+156471,24624223925.0,IssuesEvent,2022-10-16 09:55:08,dotnet/efcore,https://api.github.com/repos/dotnet/efcore,closed,How to avoid complex insert queries,closed-by-design,"I can't figure out how to do simple inserts without having the result merged and returned back.
+
+```
+using (var db = new MyDb())
+{
+ db.Something.Add(foo);
+ db.Something.Add(bar);
+ db.SaveChanges();
+}
+```
+
+When db is disposed after SaveChanges I see no reason to
+- INSERT into @toInsert1
+- MERGE into Something
+- OUTPUT into @inserted1
+- SELECT @inserted1 back to client
+
+Am I missing something?
+",1.0,"How to avoid complex insert queries - I can't figure out how to do simple inserts without having the result merged and returned back.
+
+```
+using (var db = new MyDb())
+{
+ db.Something.Add(foo);
+ db.Something.Add(bar);
+ db.SaveChanges();
+}
+```
+
+When db is disposed after SaveChanges I see no reason to
+- INSERT into @toInsert1
+- MERGE into Something
+- OUTPUT into @inserted1
+- SELECT @inserted1 back to client
+
+Am I missing something?
+",0,how to avoid complex insert queries i can t figure out how to do simple inserts without having the result merged and returned back using var db new mydb db something add foo db something add bar db savechanges when db is disposed after savechanges i see no reason to insert into merge into something output into select back to client am i missing something ,0
+532135,15530212162.0,IssuesEvent,2021-03-13 18:12:17,passwall/passwall-extension,https://api.github.com/repos/passwall/passwall-extension,closed,Alert box on delete action needed?,Priority: Medium,"When user click on delete button on details, we should show an alert with context ""Are you sure?"" or something like this.",1.0,"Alert box on delete action needed? - When user click on delete button on details, we should show an alert with context ""Are you sure?"" or something like this.",0,alert box on delete action needed when user click on delete button on details we should show an alert with context are you sure or something like this ,0
+333,6344836374.0,IssuesEvent,2017-07-27 20:50:30,wordpress-mobile/WordPress-Android,https://api.github.com/repos/wordpress-mobile/WordPress-Android,closed,Use roles from the network request,Jetpack People Management [Type] Enhancement,"For updating the users in people management, we should use the `/v1.1/sites/$site/roles` to get a possible list of roles instead of using an array resource to be able to support Jetpack sites.
+",1.0,"Use roles from the network request - For updating the users in people management, we should use the `/v1.1/sites/$site/roles` to get a possible list of roles instead of using an array resource to be able to support Jetpack sites.
+",1,use roles from the network request for updating the users in people management we should use the sites site roles to get a possible list of roles instead of using an array resource to be able to support jetpack sites ,1
+399069,11742666823.0,IssuesEvent,2020-03-12 01:38:39,thaliawww/concrexit,https://api.github.com/repos/thaliawww/concrexit,closed,Pizza orders reachable for people not in the organising member group,bug pizzas priority: medium,"In GitLab by @se-bastiaan on May 13, 2019, 13:17
+
+### One-sentence description
+
+Pizza orders reachable for people not in the organising member group
+
+### Current behaviour / Reproducing the bug
+
+1. Make a pizza event and add some orders
+2. Your user should not be a superuser
+3. Try to mark the orders as paid in the frontend
+4. Clicking doesn't do anything
+
+### Expected behaviour
+
+1. Make a pizza event and add some orders
+2. Your user should not be a superuser
+3. Try to mark the orders as paid in the frontend
+4. The orders page is permission denied",1.0,"Pizza orders reachable for people not in the organising member group - In GitLab by @se-bastiaan on May 13, 2019, 13:17
+
+### One-sentence description
+
+Pizza orders reachable for people not in the organising member group
+
+### Current behaviour / Reproducing the bug
+
+1. Make a pizza event and add some orders
+2. Your user should not be a superuser
+3. Try to mark the orders as paid in the frontend
+4. Clicking doesn't do anything
+
+### Expected behaviour
+
+1. Make a pizza event and add some orders
+2. Your user should not be a superuser
+3. Try to mark the orders as paid in the frontend
+4. The orders page is permission denied",0,pizza orders reachable for people not in the organising member group in gitlab by se bastiaan on may one sentence description pizza orders reachable for people not in the organising member group current behaviour reproducing the bug make a pizza event and add some orders your user should not be a superuser try to mark the orders as paid in the frontend clicking doesn t do anything expected behaviour make a pizza event and add some orders your user should not be a superuser try to mark the orders as paid in the frontend the orders page is permission denied,0
+55601,3073814259.0,IssuesEvent,2015-08-20 00:46:17,RobotiumTech/robotium,https://api.github.com/repos/RobotiumTech/robotium,closed,WebElement createWebElementAndSetLocation ArrayIndexOutOfBoundsException,bug imported Priority-Medium,"_From [casey.d....@gmail.com](https://code.google.com/u/115120640709842292399/) on August 24, 2013 17:53:03_
+
+What steps will reproduce the problem? 1. While using solo.waitForText(""example"") in some circumstances, such as waiting for text with this library http://www.androidviews.net/2012/11/changelog-library/ What is the expected output? What do you see instead? Will find the text, but throws ArrayIndexOutOfBoundsException at
+
+elements = data[9].split(""\\#\\$"");
+
+in WebElementCreator.createWebElementAndSetLocation What version of the product are you using? On what operating system? Robotium 4.2.1-SNAPSHOT, Android 4.3 and Android 2.3. Please provide any additional information below. This problem started when I switched to the source code version 4.2.1, it was working fine with the prebuilt version 4.2.
+
+My solution was to create the elemments varaible before the try - catch and put the elements = data[9].split(""\\#\\$""); within the try - catch.
+
+Example code:
+
+private WebElement createWebElementAndSetLocation(String information, WebView webView){
+ String[] data = information.split("";,"");
+ int x = 0;
+ int y = 0;
+ int width = 0;
+ int height = 0;
+ Hashtable attributes = new Hashtable();
+ String[] elements = {};
+ try{
+ x = Math.round(Float.valueOf(data[5]));
+ y = Math.round(Float.valueOf(data[6]));
+ width = Math.round(Float.valueOf(data[7]));
+ height = Math.round(Float.valueOf(data[8]));
+ elements = data[9].split(""\\#\\$"");
+
+ }catch(Exception ignored){}
+
+ for (int index = 0; index \< elements.length; index++){
+ String[] element = elements[index].split(""::"");
+ if (element.length > 1) {
+ attributes.put(element[0], element[1]);
+ } else {
+ attributes.put(element[0], element[0]);
+ }
+ }
+
+ WebElement webElement = null;
+
+ try{
+ webElement = new WebElement(data[0], data[1], data[2], data[3], data[4], attributes);
+ setLocation(webElement, webView, x, y, width, height);
+ }catch(Exception ignored) {}
+
+ return webElement;
+ }
+
+_Original issue: http://code.google.com/p/robotium/issues/detail?id=504_",1.0,"WebElement createWebElementAndSetLocation ArrayIndexOutOfBoundsException - _From [casey.d....@gmail.com](https://code.google.com/u/115120640709842292399/) on August 24, 2013 17:53:03_
+
+What steps will reproduce the problem? 1. While using solo.waitForText(""example"") in some circumstances, such as waiting for text with this library http://www.androidviews.net/2012/11/changelog-library/ What is the expected output? What do you see instead? Will find the text, but throws ArrayIndexOutOfBoundsException at
+
+elements = data[9].split(""\\#\\$"");
+
+in WebElementCreator.createWebElementAndSetLocation What version of the product are you using? On what operating system? Robotium 4.2.1-SNAPSHOT, Android 4.3 and Android 2.3. Please provide any additional information below. This problem started when I switched to the source code version 4.2.1, it was working fine with the prebuilt version 4.2.
+
+My solution was to create the elemments varaible before the try - catch and put the elements = data[9].split(""\\#\\$""); within the try - catch.
+
+Example code:
+
+private WebElement createWebElementAndSetLocation(String information, WebView webView){
+ String[] data = information.split("";,"");
+ int x = 0;
+ int y = 0;
+ int width = 0;
+ int height = 0;
+ Hashtable attributes = new Hashtable();
+ String[] elements = {};
+ try{
+ x = Math.round(Float.valueOf(data[5]));
+ y = Math.round(Float.valueOf(data[6]));
+ width = Math.round(Float.valueOf(data[7]));
+ height = Math.round(Float.valueOf(data[8]));
+ elements = data[9].split(""\\#\\$"");
+
+ }catch(Exception ignored){}
+
+ for (int index = 0; index \< elements.length; index++){
+ String[] element = elements[index].split(""::"");
+ if (element.length > 1) {
+ attributes.put(element[0], element[1]);
+ } else {
+ attributes.put(element[0], element[0]);
+ }
+ }
+
+ WebElement webElement = null;
+
+ try{
+ webElement = new WebElement(data[0], data[1], data[2], data[3], data[4], attributes);
+ setLocation(webElement, webView, x, y, width, height);
+ }catch(Exception ignored) {}
+
+ return webElement;
+ }
+
+_Original issue: http://code.google.com/p/robotium/issues/detail?id=504_",0,webelement createwebelementandsetlocation arrayindexoutofboundsexception from on august what steps will reproduce the problem while using solo waitfortext example in some circumstances such as waiting for text with this library what is the expected output what do you see instead will find the text but throws arrayindexoutofboundsexception at elements data split in webelementcreator createwebelementandsetlocation what version of the product are you using on what operating system robotium snapshot android and android please provide any additional information below this problem started when i switched to the source code version it was working fine with the prebuilt version my solution was to create the elemments varaible before the try catch and put the elements data split within the try catch example code private webelement createwebelementandsetlocation string information webview webview string data information split int x int y int width int height hashtable attributes new hashtable string elements try x math round float valueof data y math round float valueof data width math round float valueof data height math round float valueof data elements data split catch exception ignored for int index index elements length index string element elements split if element length attributes put element element else attributes put element element webelement webelement null try webelement new webelement data data data data data attributes setlocation webelement webview x y width height catch exception ignored return webelement original issue ,0
+245349,20763949437.0,IssuesEvent,2022-03-15 18:46:08,sourcegraph/sec-pr-audit-trail,https://api.github.com/repos/sourcegraph/sec-pr-audit-trail,closed,"sourcegraph/deploy-sourcegraph#4096: ""Update Sourcegraph Docker insiders images to 3d926ff""",exception/review exception/test-plan sourcegraph/deploy-sourcegraph,"https://github.com/sourcegraph/deploy-sourcegraph/pull/4096 ""Update Sourcegraph Docker insiders images to 3d926ff"" **has no test plan** and **was not reviewed**.
+
+Learn more about test plans in our [testing guidelines](https://docs.sourcegraph.com/dev/background-information/testing_principles#test-plans).
+
+@caugustus-sourcegraph please comment in this issue with an explanation for this exception and close this issue.",1.0,"sourcegraph/deploy-sourcegraph#4096: ""Update Sourcegraph Docker insiders images to 3d926ff"" - https://github.com/sourcegraph/deploy-sourcegraph/pull/4096 ""Update Sourcegraph Docker insiders images to 3d926ff"" **has no test plan** and **was not reviewed**.
+
+Learn more about test plans in our [testing guidelines](https://docs.sourcegraph.com/dev/background-information/testing_principles#test-plans).
+
+@caugustus-sourcegraph please comment in this issue with an explanation for this exception and close this issue.",0,sourcegraph deploy sourcegraph update sourcegraph docker insiders images to update sourcegraph docker insiders images to has no test plan and was not reviewed learn more about test plans in our caugustus sourcegraph please comment in this issue with an explanation for this exception and close this issue ,0
+130454,10616491325.0,IssuesEvent,2019-10-12 12:09:48,taskcluster/taskcluster,https://api.github.com/repos/taskcluster/taskcluster,closed,smoketest - purge a cache,smoketest,"Add a new test to the `yarn smoketest` by adding a module to `infrastructure/tooling/src/smoketest/checks/`. In this case, it would call the purge-cache service to create a purge, then pretend to be a worker and see that the purge was properly recorded.
+
+See #1218.",1.0,"smoketest - purge a cache - Add a new test to the `yarn smoketest` by adding a module to `infrastructure/tooling/src/smoketest/checks/`. In this case, it would call the purge-cache service to create a purge, then pretend to be a worker and see that the purge was properly recorded.
+
+See #1218.",0,smoketest purge a cache add a new test to the yarn smoketest by adding a module to infrastructure tooling src smoketest checks in this case it would call the purge cache service to create a purge then pretend to be a worker and see that the purge was properly recorded see ,0
+1181,30787333367.0,IssuesEvent,2023-07-31 14:02:11,culturesofknowledge/emlo-project,https://api.github.com/repos/culturesofknowledge/emlo-project,opened,People search and location search starts with,people search location search feedback0625,"Searching for people starting with ‘M’, it is calling on all names in the synonyms as well as the master name, so I am getting:
+
+
+
+Starts with search should be just the primary name, because often we need to edit by begins with a certain letter as it’s the easiest way to work though a new batch of people records.
+This should be the same in locations starts with search as well.",1.0,"People search and location search starts with - Searching for people starting with ‘M’, it is calling on all names in the synonyms as well as the master name, so I am getting:
+
+
+
+Starts with search should be just the primary name, because often we need to edit by begins with a certain letter as it’s the easiest way to work though a new batch of people records.
+This should be the same in locations starts with search as well.",1,people search and location search starts with searching for people starting with ‘m’ it is calling on all names in the synonyms as well as the master name so i am getting starts with search should be just the primary name because often we need to edit by begins with a certain letter as it’s the easiest way to work though a new batch of people records this should be the same in locations starts with search as well ,1
+736,13434072613.0,IssuesEvent,2020-09-07 10:46:56,samvera-labs/samvera-connect,https://api.github.com/repos/samvera-labs/samvera-connect,opened,Front End New Additions 2020,Administrators Devs Managers (general) Newcomers Presentation (30 min slot (20+5+5)) UI/UX people,"Adam Joseph Arling adam.arling@northwestern.edu
+Divya Katpally
+
+With a recent front-end team expansion, our workgroup’s efforts can further harness the power of UI component-driven development. We’ve experimented, tested, and implemented a few new technologies into our development approach, and would like to share our experiences. We’ll give not-super-techy-at-all, overviews of: Packaging components / component libraries for NPM with Rollup & Styleguidist (ex: openseadragon-react-viewer). Integration testing w/ Cypress. Unit testing (in line with how components are realistically used) with Jest/@testing-library. GraphQL w/ Apollo Client & React and how it’s a game-changer for architecting single page applications. Consistent, configurable styling approaches with Bulma CSS, Styled Components and @emotion. This presentation will aim to connect popular front-end jargon with what these things actually do, and why it matters or is worth knowing about. Open to all.
+
+",1.0,"Front End New Additions 2020 - Adam Joseph Arling adam.arling@northwestern.edu
+Divya Katpally
+
+With a recent front-end team expansion, our workgroup’s efforts can further harness the power of UI component-driven development. We’ve experimented, tested, and implemented a few new technologies into our development approach, and would like to share our experiences. We’ll give not-super-techy-at-all, overviews of: Packaging components / component libraries for NPM with Rollup & Styleguidist (ex: openseadragon-react-viewer). Integration testing w/ Cypress. Unit testing (in line with how components are realistically used) with Jest/@testing-library. GraphQL w/ Apollo Client & React and how it’s a game-changer for architecting single page applications. Consistent, configurable styling approaches with Bulma CSS, Styled Components and @emotion. This presentation will aim to connect popular front-end jargon with what these things actually do, and why it matters or is worth knowing about. Open to all.
+
+",1,front end new additions adam joseph arling adam arling northwestern edu divya katpally with a recent front end team expansion our workgroup’s efforts can further harness the power of ui component driven development we’ve experimented tested and implemented a few new technologies into our development approach and would like to share our experiences we’ll give not super techy at all overviews of packaging components component libraries for npm with rollup styleguidist ex openseadragon react viewer integration testing w cypress unit testing in line with how components are realistically used with jest testing library graphql w apollo client react and how it’s a game changer for architecting single page applications consistent configurable styling approaches with bulma css styled components and emotion this presentation will aim to connect popular front end jargon with what these things actually do and why it matters or is worth knowing about open to all ,1
+120038,10090919785.0,IssuesEvent,2019-07-26 13:02:00,elastic/kibana,https://api.github.com/repos/elastic/kibana,closed,Failing test: Chrome UI Functional Tests.test/functional/apps/dashboard/dashboard_time·js - dashboard app using legacy data dashboard time time changes preserved during navigation,Feature:Dashboard Team:KibanaApp failed-test,"A test failed on a tracked branch
+```
+
+ Error: retry.try timeout: TimeoutError: Waiting for element to be located By(css selector, [data-test-subj~=""superDatePickerendDatePopoverButton""])
+Wait timed out after 10016ms
+ at node_modules/selenium-webdriver/lib/webdriver.js:841:17
+ at process._tickCallback (internal/process/next_tick.js:68:7)
+ at lastError (test/common/services/retry/retry_for_success.ts:28:9)
+ at onFailure (test/common/services/retry/retry_for_success.ts:68:13)
+
+```
+First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+master/JOB=kibana-ciGroup5,node=linux-immutable/1143/)
+
+",1.0,"Failing test: Chrome UI Functional Tests.test/functional/apps/dashboard/dashboard_time·js - dashboard app using legacy data dashboard time time changes preserved during navigation - A test failed on a tracked branch
+```
+
+ Error: retry.try timeout: TimeoutError: Waiting for element to be located By(css selector, [data-test-subj~=""superDatePickerendDatePopoverButton""])
+Wait timed out after 10016ms
+ at node_modules/selenium-webdriver/lib/webdriver.js:841:17
+ at process._tickCallback (internal/process/next_tick.js:68:7)
+ at lastError (test/common/services/retry/retry_for_success.ts:28:9)
+ at onFailure (test/common/services/retry/retry_for_success.ts:68:13)
+
+```
+First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+master/JOB=kibana-ciGroup5,node=linux-immutable/1143/)
+
+",0,failing test chrome ui functional tests test functional apps dashboard dashboard time·js dashboard app using legacy data dashboard time time changes preserved during navigation a test failed on a tracked branch error retry try timeout timeouterror waiting for element to be located by css selector wait timed out after at node modules selenium webdriver lib webdriver js at process tickcallback internal process next tick js at lasterror test common services retry retry for success ts at onfailure test common services retry retry for success ts first failure ,0
+141237,18956969274.0,IssuesEvent,2021-11-18 21:33:05,CliffCrerar/aws-amplift-starter,https://api.github.com/repos/CliffCrerar/aws-amplift-starter,opened,CVE-2020-28498 (Medium) detected in elliptic-6.5.0.tgz,security vulnerability,"## CVE-2020-28498 - Medium Severity Vulnerability
+ Vulnerable Library - elliptic-6.5.0.tgz
+
+
+
+The package elliptic before 6.5.4 are vulnerable to Cryptographic Issues via the secp256k1 implementation in elliptic/ec/key.js. There is no check to confirm that the public key point passed into the derive function actually exists on the secp256k1 curve. This results in the potential for the private key used in this implementation to be revealed after a number of ECDH operations are performed.
+
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",True,"CVE-2020-28498 (Medium) detected in elliptic-6.5.0.tgz - ## CVE-2020-28498 - Medium Severity Vulnerability
+ Vulnerable Library - elliptic-6.5.0.tgz
+
+
+
+The package elliptic before 6.5.4 are vulnerable to Cryptographic Issues via the secp256k1 implementation in elliptic/ec/key.js. There is no check to confirm that the public key point passed into the derive function actually exists on the secp256k1 curve. This results in the potential for the private key used in this implementation to be revealed after a number of ECDH operations are performed.
+
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",0,cve medium detected in elliptic tgz cve medium severity vulnerability vulnerable library elliptic tgz ec cryptography library home page a href path to dependency file aws amplift starter package json path to vulnerable library aws amplift starter node modules elliptic package json dependency hierarchy webpack tgz root library node libs browser tgz crypto browserify tgz create ecdh tgz x elliptic tgz vulnerable library found in head commit a href found in base branch master vulnerability details the package elliptic before are vulnerable to cryptographic issues via the implementation in elliptic ec key js there is no check to confirm that the public key point passed into the derive function actually exists on the curve this results in the potential for the private key used in this implementation to be revealed after a number of ecdh operations are performed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope changed impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource ,0
+100811,8755618856.0,IssuesEvent,2018-12-14 15:22:06,SME-Issues/issues,https://api.github.com/repos/SME-Issues/issues,closed,Intent Errors (5004) - 14/12/2018,NLP Api PETEDEV pulse_tests,"|Expression|Result|
+|---|---|
+| _show me the invoices I haven't yet paid_After pre-procesing_show me the invoices I have not yet paid_ |
expected intent to be `query_invoices` but found `query_payment`
Expected to find one of entities {bill_from, pay_to} but found only {paid_yet, dt, pay_type, text, date_in, date_mod, datetime}.
|
+| _show me the Q3 bills_ | expected intent to be `query_invoices` but found `query_payment` |
+",1.0,"Intent Errors (5004) - 14/12/2018 - |Expression|Result|
+|---|---|
+| _show me the invoices I haven't yet paid_After pre-procesing_show me the invoices I have not yet paid_ |
expected intent to be `query_invoices` but found `query_payment`
Expected to find one of entities {bill_from, pay_to} but found only {paid_yet, dt, pay_type, text, date_in, date_mod, datetime}.
|
+| _show me the Q3 bills_ | expected intent to be `query_invoices` but found `query_payment` |
+",0,intent errors expression result show me the invoices i haven t yet paid after pre procesing show me the invoices i have not yet paid expected intent to be query invoices but found query payment expected to find one of entities bill from pay to but found only paid yet dt pay type text date in date mod datetime show me the bills expected intent to be query invoices but found query payment ,0
+475762,13725452903.0,IssuesEvent,2020-10-03 18:34:17,la-silvestri/dash-to-taskbar,https://api.github.com/repos/la-silvestri/dash-to-taskbar,opened,Open a Task Icon context menu not synchronize their hover correctly,bug good first issue low priority quick,"
+After open context menu, two task icons shows hover.",1.0,"Open a Task Icon context menu not synchronize their hover correctly - 
+After open context menu, two task icons shows hover.",0,open a task icon context menu not synchronize their hover correctly after open context menu two task icons shows hover ,0
+398,7268823196.0,IssuesEvent,2018-02-20 11:28:33,Open-NC/Are-We-Represented-Project-Center,https://api.github.com/repos/Open-NC/Are-We-Represented-Project-Center,opened,Identify & research potential uses and users of data and/or website,Multiple people needed User Research,"The initial stage of the project is to prepare data on the demographics of the community, voters, candidates, and elected commissioners. Before we can go any further, we need to understand who might use this data and how precisely they want to use it (use cases). These use cases will guide the development of any website or application we might undertake.
+
+Potential user groups include:
+
+- Get-out-the-vote (GOTV) organizations like the League of Women Voters
+- Advocacy groups like NAACP
+- Journalists
+- Political scientists
+
+The task here is to find specific people or organizations, reach out and find out what, if any interest they have in using the data and how we can support that use.
+
+This task will spawn multiple specific user research tasks for specific users and uses.",1.0,"Identify & research potential uses and users of data and/or website - The initial stage of the project is to prepare data on the demographics of the community, voters, candidates, and elected commissioners. Before we can go any further, we need to understand who might use this data and how precisely they want to use it (use cases). These use cases will guide the development of any website or application we might undertake.
+
+Potential user groups include:
+
+- Get-out-the-vote (GOTV) organizations like the League of Women Voters
+- Advocacy groups like NAACP
+- Journalists
+- Political scientists
+
+The task here is to find specific people or organizations, reach out and find out what, if any interest they have in using the data and how we can support that use.
+
+This task will spawn multiple specific user research tasks for specific users and uses.",1,identify research potential uses and users of data and or website the initial stage of the project is to prepare data on the demographics of the community voters candidates and elected commissioners before we can go any further we need to understand who might use this data and how precisely they want to use it use cases these use cases will guide the development of any website or application we might undertake potential user groups include get out the vote gotv organizations like the league of women voters advocacy groups like naacp journalists political scientists the task here is to find specific people or organizations reach out and find out what if any interest they have in using the data and how we can support that use this task will spawn multiple specific user research tasks for specific users and uses ,1
+9,2649574164.0,IssuesEvent,2015-03-15 01:59:01,fskreuz/Tree-Trail,https://api.github.com/repos/fskreuz/Tree-Trail,closed,UX Improvement for Add Badge,QA people! Test these!,"When Add Badge mode is clicked or if user is in Add Badge mode, Add Badge link should be highlighted and mouse cursor should change to indicate that the user can now add badge anywhere on the map.
+
+On the side window that will show when a location on the map is clicked, a ""Cancel"" button should also be present. When clicked, side bar slides out of view.",1.0,"UX Improvement for Add Badge - When Add Badge mode is clicked or if user is in Add Badge mode, Add Badge link should be highlighted and mouse cursor should change to indicate that the user can now add badge anywhere on the map.
+
+On the side window that will show when a location on the map is clicked, a ""Cancel"" button should also be present. When clicked, side bar slides out of view.",1,ux improvement for add badge when add badge mode is clicked or if user is in add badge mode add badge link should be highlighted and mouse cursor should change to indicate that the user can now add badge anywhere on the map on the side window that will show when a location on the map is clicked a cancel button should also be present when clicked side bar slides out of view ,1
+553,9780194918.0,IssuesEvent,2019-06-07 16:13:56,3box/ops,https://api.github.com/repos/3box/ops,closed,Enroll in COBRA,Ops People,"Cannot begin healthcare until we've had 6 weeks of payroll, so we do need to enroll in COBRA. Will discuss offline reimbursement.
+
+- [x] Danny
+- [ ] Sena
+- [ ] Zach",1.0,"Enroll in COBRA - Cannot begin healthcare until we've had 6 weeks of payroll, so we do need to enroll in COBRA. Will discuss offline reimbursement.
+
+- [x] Danny
+- [ ] Sena
+- [ ] Zach",1,enroll in cobra cannot begin healthcare until we ve had weeks of payroll so we do need to enroll in cobra will discuss offline reimbursement danny sena zach,1
+813,15164450041.0,IssuesEvent,2021-02-12 13:45:37,planningcenter/developers,https://api.github.com/repos/planningcenter/developers,closed,/people/v2/people API incomplete response for include=name_suffix,People question,"**Affected Product**
+Which product does this bug affect?
+People
+
+**Describe the bug**
+Calling https://api.planningcenteronline.com/people/v2/people
+ with: X-PCO-API-Version: 2020-07-22
+ and include: emails,phone_numbers,name_prefix,name_suffix
+
+returns incomplete name_suffix responses:
+{'attributes': {'value': 'Sr.'},
+'id': '',
+'links': {'self': 'https://api.planningcenteronline.com/people/v2/name_suffixes/'},
+'type': 'NameSuffix'}
+
+each of which are missing:
+'relationships': {'person': {'data': {'id': '', 'type': 'Person'}}},
+
+In fact, all items returned that reference the same suffix are indistinguishable because none identify the person they apply to.
+
+**To Reproduce**
+See above.
+
+**Expected behavior**
+As with other include items, the response should include a relationship to a person.
+
+**Screenshots**
+n/a
+
+**Additional Context:**
+
+- Endpoint: https://api.planningcenteronline.com/people/v2/people
+- Language: Python 3
+- Authentication: OAuth2
+
+**Additional context**
+
+
+## I have..
+
+- [x] Reviewed the documentation found at https://developer.planning.center/docs
+- [x] Searched for previous issues reporting this bug
+- [x] Removed all private information from this issue (credentials, tokens, emails, phone numbers, etc.)
+- [x] Reviewed my issue for completeness",1.0,"/people/v2/people API incomplete response for include=name_suffix - **Affected Product**
+Which product does this bug affect?
+People
+
+**Describe the bug**
+Calling https://api.planningcenteronline.com/people/v2/people
+ with: X-PCO-API-Version: 2020-07-22
+ and include: emails,phone_numbers,name_prefix,name_suffix
+
+returns incomplete name_suffix responses:
+{'attributes': {'value': 'Sr.'},
+'id': '',
+'links': {'self': 'https://api.planningcenteronline.com/people/v2/name_suffixes/'},
+'type': 'NameSuffix'}
+
+each of which are missing:
+'relationships': {'person': {'data': {'id': '', 'type': 'Person'}}},
+
+In fact, all items returned that reference the same suffix are indistinguishable because none identify the person they apply to.
+
+**To Reproduce**
+See above.
+
+**Expected behavior**
+As with other include items, the response should include a relationship to a person.
+
+**Screenshots**
+n/a
+
+**Additional Context:**
+
+- Endpoint: https://api.planningcenteronline.com/people/v2/people
+- Language: Python 3
+- Authentication: OAuth2
+
+**Additional context**
+
+
+## I have..
+
+- [x] Reviewed the documentation found at https://developer.planning.center/docs
+- [x] Searched for previous issues reporting this bug
+- [x] Removed all private information from this issue (credentials, tokens, emails, phone numbers, etc.)
+- [x] Reviewed my issue for completeness",1, people people api incomplete response for include name suffix affected product which product does this bug affect people describe the bug calling with x pco api version and include emails phone numbers name prefix name suffix returns incomplete name suffix responses attributes value sr id links self type namesuffix each of which are missing relationships person data id type person in fact all items returned that reference the same suffix are indistinguishable because none identify the person they apply to to reproduce see above expected behavior as with other include items the response should include a relationship to a person screenshots n a additional context endpoint language python authentication additional context i have reviewed the documentation found at searched for previous issues reporting this bug removed all private information from this issue credentials tokens emails phone numbers etc reviewed my issue for completeness,1
+1086,26548872953.0,IssuesEvent,2023-01-20 04:51:41,santaclaraca/upptime,https://api.github.com/repos/santaclaraca/upptime,closed,🛑 PeopleSoft is down,status people-soft,"In [`f3da0d4`](https://github.com/santaclaraca/upptime/commit/f3da0d40db1101d3b7fbe21ddee0daafe977d132
+), PeopleSoft ($SITE_PEOPLESOFT) was **down**:
+- HTTP code: 0
+- Response time: 0 ms
+",1.0,"🛑 PeopleSoft is down - In [`f3da0d4`](https://github.com/santaclaraca/upptime/commit/f3da0d40db1101d3b7fbe21ddee0daafe977d132
+), PeopleSoft ($SITE_PEOPLESOFT) was **down**:
+- HTTP code: 0
+- Response time: 0 ms
+",1,🛑 peoplesoft is down in peoplesoft site peoplesoft was down http code response time ms ,1
+633,11211833193.0,IssuesEvent,2020-01-06 16:15:09,planningcenter/developers,https://api.github.com/repos/planningcenter/developers,closed,500 response creating user with Post to People,People,"**Related Applications**
+People
+
+**Describe the question**
+Trying to invoke the 'Creating' method via a POST with respect to this documentation:
+https://developer.planning.center/docs/#/apps/people/2019-10-10/vertices/person
+
+Intend to generate a new user of People as a new contact (they don't need login credentials).
+
+**What have you tried that didn't work?**
+Using Basic HTTP authentication (linked to my administrator account) I attempt to POST via cURL on my personal computer.
+
+```bash
+curl -v -u $ID:$SECRET https://api.planningcenteronline.com/people/v2/people -X POST -d @sample.json
+```
+Where $ID and $SECRET are my API ID and password respectively, and `sample.json` is the following:
+```json
+[{
+""type"": ""Person"",
+""id"": ""1"",
+""attributes"": {
+ ""first_name"": ""APITest"",
+ ""last_name"": ""Test""
+},
+""relationships"": {
+ ""primary_campus"": {
+ ""data"": {
+ ""type"": ""PrimaryCampus"",
+ ""id"": ""1""
+ }
+ }
+}
+}]
+```
+I receive a 500 response with the note: ""An internal server error occurred on our end and we've been notified. Please contact support if the issue persists.""
+
+
+## I have..
+
+- [X] Reviewed the documentation found at https://developer.planning.center/docs
+- [X] Searched for previous issues answering this question
+- [X] Removed all private information from this issue (credentials, tokens, emails, phone numbers, etc.)
+- [x] Reviewed my issue for completeness
+",1.0,"500 response creating user with Post to People - **Related Applications**
+People
+
+**Describe the question**
+Trying to invoke the 'Creating' method via a POST with respect to this documentation:
+https://developer.planning.center/docs/#/apps/people/2019-10-10/vertices/person
+
+Intend to generate a new user of People as a new contact (they don't need login credentials).
+
+**What have you tried that didn't work?**
+Using Basic HTTP authentication (linked to my administrator account) I attempt to POST via cURL on my personal computer.
+
+```bash
+curl -v -u $ID:$SECRET https://api.planningcenteronline.com/people/v2/people -X POST -d @sample.json
+```
+Where $ID and $SECRET are my API ID and password respectively, and `sample.json` is the following:
+```json
+[{
+""type"": ""Person"",
+""id"": ""1"",
+""attributes"": {
+ ""first_name"": ""APITest"",
+ ""last_name"": ""Test""
+},
+""relationships"": {
+ ""primary_campus"": {
+ ""data"": {
+ ""type"": ""PrimaryCampus"",
+ ""id"": ""1""
+ }
+ }
+}
+}]
+```
+I receive a 500 response with the note: ""An internal server error occurred on our end and we've been notified. Please contact support if the issue persists.""
+
+
+## I have..
+
+- [X] Reviewed the documentation found at https://developer.planning.center/docs
+- [X] Searched for previous issues answering this question
+- [X] Removed all private information from this issue (credentials, tokens, emails, phone numbers, etc.)
+- [x] Reviewed my issue for completeness
+",1, response creating user with post to people related applications people describe the question trying to invoke the creating method via a post with respect to this documentation intend to generate a new user of people as a new contact they don t need login credentials what have you tried that didn t work using basic http authentication linked to my administrator account i attempt to post via curl on my personal computer bash curl v u id secret x post d sample json where id and secret are my api id and password respectively and sample json is the following json type person id attributes first name apitest last name test relationships primary campus data type primarycampus id i receive a response with the note an internal server error occurred on our end and we ve been notified please contact support if the issue persists i have reviewed the documentation found at searched for previous issues answering this question removed all private information from this issue credentials tokens emails phone numbers etc reviewed my issue for completeness ,1
+139387,20831813079.0,IssuesEvent,2022-03-19 15:15:53,gitpod-io/gitpod,https://api.github.com/repos/gitpod-io/gitpod,closed,Allow users to transfer projects to another namespace (team or personal account),type: feature request meta: stale component: dashboard needs visual design feature: teams and projects team: webapp,"### Problem to solve
+
+Currently, users can add or remove projects form a team or their personal account. The addition and removal process contains some friction that could be avoided when a user simply wants to transfer a project to another namespace (team or personal account).
+
+
+### Proposal
+
+Given that project settings start to build up with incremental settings (https://github.com/gitpod-io/gitpod/issues/6727) , project-level environment variables (https://github.com/gitpod-io/gitpod/issues/4456), etc. it makes sense to introduce another setting for transfering a project to another namespace.
+
+### Designs
+
+TBD (To be discussed)
+
+This could introduce a small section under _General_ project settings with one action button that triggers a modal to select a namespace to transfer the proejct to. 💭 ",1.0,"Allow users to transfer projects to another namespace (team or personal account) - ### Problem to solve
+
+Currently, users can add or remove projects form a team or their personal account. The addition and removal process contains some friction that could be avoided when a user simply wants to transfer a project to another namespace (team or personal account).
+
+
+### Proposal
+
+Given that project settings start to build up with incremental settings (https://github.com/gitpod-io/gitpod/issues/6727) , project-level environment variables (https://github.com/gitpod-io/gitpod/issues/4456), etc. it makes sense to introduce another setting for transfering a project to another namespace.
+
+### Designs
+
+TBD (To be discussed)
+
+This could introduce a small section under _General_ project settings with one action button that triggers a modal to select a namespace to transfer the proejct to. 💭 ",0,allow users to transfer projects to another namespace team or personal account problem to solve currently users can add or remove projects form a team or their personal account the addition and removal process contains some friction that could be avoided when a user simply wants to transfer a project to another namespace team or personal account proposal given that project settings start to build up with incremental settings project level environment variables etc it makes sense to introduce another setting for transfering a project to another namespace designs tbd to be discussed this could introduce a small section under general project settings with one action button that triggers a modal to select a namespace to transfer the proejct to 💭 ,0
+162622,6156664171.0,IssuesEvent,2017-06-28 17:13:12,Cadasta/cadasta-platform,https://api.github.com/repos/Cadasta/cadasta-platform,closed,No search results for party-related information in demo/platform,bug priority: high search,"### Steps to reproduce the error
+
+1. Try to execute any search query in demo or production with a keyword for a party: examples: ""David"", ""IN"" (for individual), etc.
+2. No results are returned
+
+However in staging this is working as expected. And location and relationship related queries are also working fine in demo and platform.
+
+Can @seav or @amplifi take a look?
+
+",1.0,"No search results for party-related information in demo/platform - ### Steps to reproduce the error
+
+1. Try to execute any search query in demo or production with a keyword for a party: examples: ""David"", ""IN"" (for individual), etc.
+2. No results are returned
+
+However in staging this is working as expected. And location and relationship related queries are also working fine in demo and platform.
+
+Can @seav or @amplifi take a look?
+
+",0,no search results for party related information in demo platform steps to reproduce the error try to execute any search query in demo or production with a keyword for a party examples david in for individual etc no results are returned however in staging this is working as expected and location and relationship related queries are also working fine in demo and platform can seav or amplifi take a look ,0
+41709,10572905152.0,IssuesEvent,2019-10-07 10:39:44,primefaces/primeng,https://api.github.com/repos/primefaces/primeng,closed,Turbotable - Multisort with null values doesn't quite work correctly,defect,"**I'm submitting a ...** (check one with ""x"")
+```
+[x ] bug report => Search github for a similar issue or PR before submitting
+[ ] feature request => Please check if request is not on the roadmap already https://github.com/primefaces/primeng/wiki/Roadmap
+[ ] support request => Please do not submit support request here, instead see http://forum.primefaces.org/viewforum.php?f=35
+```
+
+**Current behavior**
+When multisort is enabled, columns that are sorted with null values return in random order when sorting the column multiple times
+
+**Expected behavior**
+To sort like single sort mode does
+
+I believe this bug is because you're missing an 'else' here: https://github.com/primefaces/primeng/blob/master/src/app/components/table/table.ts#L771
+
+```TSX
+ if (value1 == null && value2 != null)
+ result = -1;
+ else if (value1 != null && value2 == null)
+ result = 1;
+ else if (value1 == null && value2 == null)
+ result = 0;
+ if (typeof value1 == 'string' || value1 instanceof String) {
+ if (value1.localeCompare && (value1 != value2)) {
+ return (multiSortMeta[index].order * value1.localeCompare(value2));
+ }
+ }
+ else {
+ result = (value1 < value2) ? -1 : 1;
+ }
+```
+
+Should be
+
+```TSX
+ if (value1 == null && value2 != null)
+ result = -1;
+ else if (value1 != null && value2 == null)
+ result = 1;
+ else if (value1 == null && value2 == null)
+ result = 0;
+ else if (typeof value1 == 'string' || value1 instanceof String) {
+ if (value1.localeCompare && (value1 != value2)) {
+ return (multiSortMeta[index].order * value1.localeCompare(value2));
+ }
+ }
+ else {
+ result = (value1 < value2) ? -1 : 1;
+ }
+```
+Like in singlesort. Otherwise the null checking is pretty much ignored.
+
+",1.0,"Turbotable - Multisort with null values doesn't quite work correctly - **I'm submitting a ...** (check one with ""x"")
+```
+[x ] bug report => Search github for a similar issue or PR before submitting
+[ ] feature request => Please check if request is not on the roadmap already https://github.com/primefaces/primeng/wiki/Roadmap
+[ ] support request => Please do not submit support request here, instead see http://forum.primefaces.org/viewforum.php?f=35
+```
+
+**Current behavior**
+When multisort is enabled, columns that are sorted with null values return in random order when sorting the column multiple times
+
+**Expected behavior**
+To sort like single sort mode does
+
+I believe this bug is because you're missing an 'else' here: https://github.com/primefaces/primeng/blob/master/src/app/components/table/table.ts#L771
+
+```TSX
+ if (value1 == null && value2 != null)
+ result = -1;
+ else if (value1 != null && value2 == null)
+ result = 1;
+ else if (value1 == null && value2 == null)
+ result = 0;
+ if (typeof value1 == 'string' || value1 instanceof String) {
+ if (value1.localeCompare && (value1 != value2)) {
+ return (multiSortMeta[index].order * value1.localeCompare(value2));
+ }
+ }
+ else {
+ result = (value1 < value2) ? -1 : 1;
+ }
+```
+
+Should be
+
+```TSX
+ if (value1 == null && value2 != null)
+ result = -1;
+ else if (value1 != null && value2 == null)
+ result = 1;
+ else if (value1 == null && value2 == null)
+ result = 0;
+ else if (typeof value1 == 'string' || value1 instanceof String) {
+ if (value1.localeCompare && (value1 != value2)) {
+ return (multiSortMeta[index].order * value1.localeCompare(value2));
+ }
+ }
+ else {
+ result = (value1 < value2) ? -1 : 1;
+ }
+```
+Like in singlesort. Otherwise the null checking is pretty much ignored.
+
+",0,turbotable multisort with null values doesn t quite work correctly i m submitting a check one with x bug report search github for a similar issue or pr before submitting feature request please check if request is not on the roadmap already support request please do not submit support request here instead see current behavior when multisort is enabled columns that are sorted with null values return in random order when sorting the column multiple times expected behavior to sort like single sort mode does i believe this bug is because you re missing an else here tsx if null null result else if null null result else if null null result if typeof string instanceof string if localecompare return multisortmeta order localecompare else result should be tsx if null null result else if null null result else if null null result else if typeof string instanceof string if localecompare return multisortmeta order localecompare else result like in singlesort otherwise the null checking is pretty much ignored ,0
+36933,8195695525.0,IssuesEvent,2018-08-31 07:13:16,phetsims/coulombs-law,https://api.github.com/repos/phetsims/coulombs-law,opened,Inconsistent screen names,dev:code-review,"As noted in #63:
+
+> Do filenames use an appropriate prefix?
+
+CLMacroScreen.js and CLAtomicScreen.js are inconsistent with the rest of the file names, which use the extend name of the sim (i.e. CoulombsLawGlobalOptionsNode.js). These file names should be changed @mbarlow12.",1.0,"Inconsistent screen names - As noted in #63:
+
+> Do filenames use an appropriate prefix?
+
+CLMacroScreen.js and CLAtomicScreen.js are inconsistent with the rest of the file names, which use the extend name of the sim (i.e. CoulombsLawGlobalOptionsNode.js). These file names should be changed @mbarlow12.",0,inconsistent screen names as noted in do filenames use an appropriate prefix clmacroscreen js and clatomicscreen js are inconsistent with the rest of the file names which use the extend name of the sim i e coulombslawglobaloptionsnode js these file names should be changed ,0
+57066,3081235925.0,IssuesEvent,2015-08-22 14:24:55,bitfighter/bitfighter,https://api.github.com/repos/bitfighter/bitfighter,closed,Game Type sorting in Editor Game Parameters is wonky,019 bug imported Priority-Medium,"_From [buckyballreaction](https://code.google.com/u/buckyballreaction/) on November 13, 2013 15:39:35_
+
+What steps will reproduce the problem? 1. Edit a level
+2. Go to game parameters
+3. Cycle through the game types
+
+
+Game Types should cycle in a sorted method, but do not.
+
+_Original issue: http://code.google.com/p/bitfighter/issues/detail?id=297_",1.0,"Game Type sorting in Editor Game Parameters is wonky - _From [buckyballreaction](https://code.google.com/u/buckyballreaction/) on November 13, 2013 15:39:35_
+
+What steps will reproduce the problem? 1. Edit a level
+2. Go to game parameters
+3. Cycle through the game types
+
+
+Game Types should cycle in a sorted method, but do not.
+
+_Original issue: http://code.google.com/p/bitfighter/issues/detail?id=297_",0,game type sorting in editor game parameters is wonky from on november what steps will reproduce the problem edit a level go to game parameters cycle through the game types game types should cycle in a sorted method but do not original issue ,0
+103882,8952883725.0,IssuesEvent,2019-01-25 17:46:28,kubernetes/kubernetes,https://api.github.com/repos/kubernetes/kubernetes,closed,Resource Quota test with uninitialized pod fails,kind/failing-test sig/api-machinery sig/testing,"
+
+**Which jobs are failing**:
+
+Unknown.
+
+**Which test(s) are failing**:
+
+ [Feature:Initializers] should create a ResourceQuota and capture the life of an uninitialized pod.
+
+**Since when has it been failing**:
+
+Unknown.
+
+**Testgrid link**:
+
+Unknown
+
+**Reason for failure**:
+
+``INFO: expect err to be timeout error, got Pod ""too-large-pod"" is invalid: metadata.initializers.pending[0].name: Invalid value: ""unhandled"": should be a domain with at least three segments separated by dots``
+
+Even after that particular issue is fixed, the test continues to fail at the step when creating a Pod with an Initializer which exceeds the Resource Quota:
+
+``INFO: expect err to be timeout error, got pods ""too-large-pod"" is forbidden: exceeded quota: t
+est-quota, requested: cpu=1100m, used: cpu=0, limited: cpu=1``
+
+Additionally, in my case, the 30 seconds wait for the Resource Quota to be updated is not always enough, which causes the test to sometimes timeout.
+
+**Anything else we need to know**:
+
+/sig testing
+",2.0,"Resource Quota test with uninitialized pod fails -
+
+**Which jobs are failing**:
+
+Unknown.
+
+**Which test(s) are failing**:
+
+ [Feature:Initializers] should create a ResourceQuota and capture the life of an uninitialized pod.
+
+**Since when has it been failing**:
+
+Unknown.
+
+**Testgrid link**:
+
+Unknown
+
+**Reason for failure**:
+
+``INFO: expect err to be timeout error, got Pod ""too-large-pod"" is invalid: metadata.initializers.pending[0].name: Invalid value: ""unhandled"": should be a domain with at least three segments separated by dots``
+
+Even after that particular issue is fixed, the test continues to fail at the step when creating a Pod with an Initializer which exceeds the Resource Quota:
+
+``INFO: expect err to be timeout error, got pods ""too-large-pod"" is forbidden: exceeded quota: t
+est-quota, requested: cpu=1100m, used: cpu=0, limited: cpu=1``
+
+Additionally, in my case, the 30 seconds wait for the Resource Quota to be updated is not always enough, which causes the test to sometimes timeout.
+
+**Anything else we need to know**:
+
+/sig testing
+",0,resource quota test with uninitialized pod fails which jobs are failing unknown which test s are failing should create a resourcequota and capture the life of an uninitialized pod since when has it been failing unknown testgrid link unknown reason for failure info expect err to be timeout error got pod too large pod is invalid metadata initializers pending name invalid value unhandled should be a domain with at least three segments separated by dots even after that particular issue is fixed the test continues to fail at the step when creating a pod with an initializer which exceeds the resource quota info expect err to be timeout error got pods too large pod is forbidden exceeded quota t est quota requested cpu used cpu limited cpu additionally in my case the seconds wait for the resource quota to be updated is not always enough which causes the test to sometimes timeout anything else we need to know sig testing ,0
+720839,24808299177.0,IssuesEvent,2022-10-25 07:22:32,AY2223S1-CS2103T-T13-4/tp,https://api.github.com/repos/AY2223S1-CS2103T-T13-4/tp,closed,"As an expert user, I can tag tasks",type.Story priority.Medium,... so that the tasks can be easily categorized and found using the search keyword feature,1.0,"As an expert user, I can tag tasks - ... so that the tasks can be easily categorized and found using the search keyword feature",0,as an expert user i can tag tasks so that the tasks can be easily categorized and found using the search keyword feature,0
+1048,25179661767.0,IssuesEvent,2022-11-11 12:30:01,jongfeel/BookReview,https://api.github.com/repos/jongfeel/BookReview,closed,4부 8장 고치기 쉬운 잘못처럼 보이게 하라,2022 How to Win Friends & Influence People,"### 8장 고치기 쉬운 잘못처럼 보이게 하라
+
+격려를 아끼지 않고, 그 일이 하기 쉬운 일처럼 여겨지게 만들어라. 그가 그 일을 할 수 있는 능력이 있음을 당신이 믿고 있다고 느끼게 만들어라. 그가 스스로를 아직 개발되지는 않았지만 타고난 재주가 있는 사람이라고 느끼게 만들어라.
+
+---
+
+규칙 8: 격려하라. 고쳐 주고 싶은 잘못은 고치기 쉬운 잘못처럼 보이게 하라. 다른 사람이 해 주었으면 하는 일은 쉬운 일처럼 보이게 만들어라.
+Use encouragement. Make the fault you want to correct seem easy to correct; make the thing you want the other person to do seem easy to do.",1.0,"4부 8장 고치기 쉬운 잘못처럼 보이게 하라 - ### 8장 고치기 쉬운 잘못처럼 보이게 하라
+
+격려를 아끼지 않고, 그 일이 하기 쉬운 일처럼 여겨지게 만들어라. 그가 그 일을 할 수 있는 능력이 있음을 당신이 믿고 있다고 느끼게 만들어라. 그가 스스로를 아직 개발되지는 않았지만 타고난 재주가 있는 사람이라고 느끼게 만들어라.
+
+---
+
+규칙 8: 격려하라. 고쳐 주고 싶은 잘못은 고치기 쉬운 잘못처럼 보이게 하라. 다른 사람이 해 주었으면 하는 일은 쉬운 일처럼 보이게 만들어라.
+Use encouragement. Make the fault you want to correct seem easy to correct; make the thing you want the other person to do seem easy to do.",1, 고치기 쉬운 잘못처럼 보이게 하라 ���치기 쉬운 잘못처럼 보이게 하라 격려를 아끼지 않고 그 일이 하기 쉬운 일처럼 여겨지게 만들어라 그가 그 일을 할 수 있는 능력이 있음을 당신이 믿고 있다고 느끼게 만들어라 그가 스스로를 아직 개발되지는 않았지만 타고난 재주가 있는 사람이라고 느끼게 만들어라 규칙 격려하라 고쳐 주고 싶은 잘못은 고치기 쉬운 잘못처럼 보이게 하라 다른 사람이 해 주었으면 하는 일은 쉬운 일처럼 보이게 만들어라 use encouragement make the fault you want to correct seem easy to correct make the thing you want the other person to do seem easy to do ,1
+686,12289704975.0,IssuesEvent,2020-05-09 23:00:02,microsoft/fluentui,https://api.github.com/repos/microsoft/fluentui,closed,Getting error while using people picker control,Component: PeoplePicker Component: Pickers Needs: Actionable Feedback :female_detective: Status: No Recent Activity,"I am not able to resolve bellow issue which are relate to people picker control.
+
+onResolveSuggestions={onFilterChanged}
+onEmptyInputFocus={returnMostRecentlyUsed}
+
+
+
+
+A callback for what should happen when a person types text into the input. Returns the already selected items so the resolver can filter them out. If used in conjunction with resolveDelay this will only kick off after the delay throttle.
+
+Type '(filterText: string, currentPersonas: IPersonaProps[], limitResults?: number | undefined) => IPersonaProps[] | Promise' is not assignable to type '(filter: string, selectedItems?: IPersonaProps[] | undefined) => IPersonaProps[] | PromiseLike'.
+ Types of parameters 'currentPersonas' and 'selectedItems' are incompatible.
+ Type 'IPersonaProps[] | undefined' is not assignable to type 'IPersonaProps[]'.
+ Type 'undefined' is not assignable to type 'IPersonaProps[]'.ts(2322)
+BasePicker.types.d.ts(53, 5): The expected type comes from property 'onResolveSuggestions' which is declared here on type 'IntrinsicAttributes & IPeoplePickerProps & { children?: ReactNode; }'
+
+",1.0,"Getting error while using people picker control - I am not able to resolve bellow issue which are relate to people picker control.
+
+onResolveSuggestions={onFilterChanged}
+onEmptyInputFocus={returnMostRecentlyUsed}
+
+
+
+
+A callback for what should happen when a person types text into the input. Returns the already selected items so the resolver can filter them out. If used in conjunction with resolveDelay this will only kick off after the delay throttle.
+
+Type '(filterText: string, currentPersonas: IPersonaProps[], limitResults?: number | undefined) => IPersonaProps[] | Promise' is not assignable to type '(filter: string, selectedItems?: IPersonaProps[] | undefined) => IPersonaProps[] | PromiseLike'.
+ Types of parameters 'currentPersonas' and 'selectedItems' are incompatible.
+ Type 'IPersonaProps[] | undefined' is not assignable to type 'IPersonaProps[]'.
+ Type 'undefined' is not assignable to type 'IPersonaProps[]'.ts(2322)
+BasePicker.types.d.ts(53, 5): The expected type comes from property 'onResolveSuggestions' which is declared here on type 'IntrinsicAttributes & IPeoplePickerProps & { children?: ReactNode; }'
+
+",1,getting error while using people picker control i am not able to resolve bellow issue which are relate to people picker control onresolvesuggestions onfilterchanged onemptyinputfocus returnmostrecentlyused a callback for what should happen when a person types text into the input returns the already selected items so the resolver can filter them out if used in conjunction with resolvedelay this will only kick off after the delay throttle type filtertext string currentpersonas ipersonaprops limitresults number undefined ipersonaprops promise is not assignable to type filter string selecteditems ipersonaprops undefined ipersonaprops promiselike types of parameters currentpersonas and selecteditems are incompatible type ipersonaprops undefined is not assignable to type ipersonaprops type undefined is not assignable to type ipersonaprops ts basepicker types d ts the expected type comes from property onresolvesuggestions which is declared here on type intrinsicattributes ipeoplepickerprops children reactnode ,1
+58,2989649489.0,IssuesEvent,2015-07-21 01:32:58,schmave/demschooltools,https://api.github.com/repos/schmave/demschooltools,closed,Crash when trying to make a new tasklist with empty title,bug people, at controllers.Settings.newTaskList(Settings.java:89),1.0,Crash when trying to make a new tasklist with empty title - at controllers.Settings.newTaskList(Settings.java:89),1,crash when trying to make a new tasklist with empty title at controllers settings newtasklist settings java ,1
+358268,25184921242.0,IssuesEvent,2022-11-11 16:58:46,Prairie-Shell-Studios/Yeddie-Summit,https://api.github.com/repos/Prairie-Shell-Studios/Yeddie-Summit,closed,Dev Task: Snowball Collision Splitting,documentation enhancement,"# Description
+For issue #5, the snowball should split into 2 half-sized snowballs upon collision with a prop.
+
+Create a script that removes the existing snowball GO and replaces it with 2 other snowball GOs at half the scale.
+## Notes
+- Complete after issue #4.
+- The 2 new smaller GOs should have a force applied to them in the direction of the collision point.
+- The original GO should be placed back in the object pool.
+- The script should check if half the current scale is above the minimum scale and only spawn 2 new ones if it is.
+- Only split (or destroy) the GO if it is above a certain scale.",1.0,"Dev Task: Snowball Collision Splitting - # Description
+For issue #5, the snowball should split into 2 half-sized snowballs upon collision with a prop.
+
+Create a script that removes the existing snowball GO and replaces it with 2 other snowball GOs at half the scale.
+## Notes
+- Complete after issue #4.
+- The 2 new smaller GOs should have a force applied to them in the direction of the collision point.
+- The original GO should be placed back in the object pool.
+- The script should check if half the current scale is above the minimum scale and only spawn 2 new ones if it is.
+- Only split (or destroy) the GO if it is above a certain scale.",0,dev task snowball collision splitting description for issue the snowball should split into half sized snowballs upon collision with a prop create a script that removes the existing snowball go and replaces it with other snowball gos at half the scale notes complete after issue the new smaller gos should have a force applied to them in the direction of the collision point the original go should be placed back in the object pool the script should check if half the current scale is above the minimum scale and only spawn new ones if it is only split or destroy the go if it is above a certain scale ,0
+946,20987979518.0,IssuesEvent,2022-03-29 06:27:31,AtB-AS/mittatb-app,https://api.github.com/repos/AtB-AS/mittatb-app,closed,Oppdatere informasjon rundt togbillett i app ,:shirt: S :iphone: 1.18 :people_holding_hands: Teaminitiativ,"## Origin
+
+
+
+Meldingen sier at man må kjøpe en annen billett, og som ikke er tilgjengelig i appen. Nå som samarbeidsbilletten er ute, må meldingen endres.
+
+Siden det er komplisert å evaluere soner og hvilke operatører som inngår i samarbeidsbilletten gjør vi her en inkrementell tilnærming, og utsetter den perfekte løsningen til litt senere.
+
+Per i dag viser vi Gul melding hvis reisen inneholder en annen operatør enn AtB.
+Endringen som skal gjøres i denne saken er å legge til tekst hvis Tog er involvert i reisen.
+
+ToDo:
+- [x] Logikk for når den gule meldingen skal vises forblir som før
+- [x] Sjekk om resien innbefatter Tog, i så fall skal denne teksten legges til:
+ - Med enkelt- og periodebillett for sone A fra AtB kan du reise med tog i sonen.
+ - With a single or periodic ticket for zone A from AtB, you can travel by train in the zone.
+
+## Implmentasjon
+
+
+
+Skrev om den biten som tester kort ventetid, for å fjerne en god porsjon spaghetti.
+
+
+## Test
+Siden det er gjort en del omskrivinger på biten som sjekker kort ventetid, er det lagt til enhetstester som tester dette.
+Testene kjøres ved `yarn test`.
+Ellers kan både reisesøk og avganger testes ganske greit, der utvidet gul melding skal vises hvis reisen inneholder tog.
+
+
+",1.0,"Oppdatere informasjon rundt togbillett i app - ## Origin
+
+
+
+Meldingen sier at man må kjøpe en annen billett, og som ikke er tilgjengelig i appen. Nå som samarbeidsbilletten er ute, må meldingen endres.
+
+Siden det er komplisert å evaluere soner og hvilke operatører som inngår i samarbeidsbilletten gjør vi her en inkrementell tilnærming, og utsetter den perfekte løsningen til litt senere.
+
+Per i dag viser vi Gul melding hvis reisen inneholder en annen operatør enn AtB.
+Endringen som skal gjøres i denne saken er å legge til tekst hvis Tog er involvert i reisen.
+
+ToDo:
+- [x] Logikk for når den gule meldingen skal vises forblir som før
+- [x] Sjekk om resien innbefatter Tog, i så fall skal denne teksten legges til:
+ - Med enkelt- og periodebillett for sone A fra AtB kan du reise med tog i sonen.
+ - With a single or periodic ticket for zone A from AtB, you can travel by train in the zone.
+
+## Implmentasjon
+
+
+
+Skrev om den biten som tester kort ventetid, for å fjerne en god porsjon spaghetti.
+
+
+## Test
+Siden det er gjort en del omskrivinger på biten som sjekker kort ventetid, er det lagt til enhetstester som tester dette.
+Testene kjøres ved `yarn test`.
+Ellers kan både reisesøk og avganger testes ganske greit, der utvidet gul melding skal vises hvis reisen inneholder tog.
+
+
+",1,oppdatere informasjon rundt togbillett i app origin meldingen sier at man må kjøpe en annen billett og som ikke er tilgjengelig i appen nå som samarbeidsbilletten er ute må meldingen endres siden det er komplisert å evaluere soner og hvilke operatører som inngår i samarbeidsbilletten gjør vi her en inkrementell tilnærming og utsetter den perfekte løsningen til litt senere per i dag viser vi gul melding hvis reisen inneholder en annen operatør enn atb endringen som skal gjøres i denne saken er å legge til tekst hvis tog er involvert i reisen todo logikk for når den gule meldingen skal vises forblir som før sjekk om resien innbefatter tog i så fall skal denne teksten legges til med enkelt og periodebillett for sone a fra atb kan du reise med tog i sonen with a single or periodic ticket for zone a from atb you can travel by train in the zone implmentasjon skrev om den biten som tester kort ventetid for å fjerne en god porsjon spaghetti test siden det er gjort en del omskrivinger på biten som sjekker kort ventetid er det lagt til enhetstester som tester dette testene kjøres ved yarn test ellers kan både reisesøk og avganger testes ganske greit der utvidet gul melding skal vises hvis reisen inneholder tog ,1
+794,14689469020.0,IssuesEvent,2021-01-02 09:56:48,freeyourscience/freeyourscience,https://api.github.com/repos/freeyourscience/freeyourscience,closed,Fix search behaviour,helpPeopleUnderstandOaPathways,"- [x] autodetect orcid
+- [x] autodetect DOI
+- [x] unify paper and author endpoint to /search
+
+## Refactor
+
+- [x] names of variables (profile...)",1.0,"Fix search behaviour - - [x] autodetect orcid
+- [x] autodetect DOI
+- [x] unify paper and author endpoint to /search
+
+## Refactor
+
+- [x] names of variables (profile...)",1,fix search behaviour autodetect orcid autodetect doi unify paper and author endpoint to search refactor names of variables profile ,1
+970,21945064161.0,IssuesEvent,2022-05-23 23:00:01,microsoft/fluentui,https://api.github.com/repos/microsoft/fluentui,closed,[PeoplePicker] - NVDA Screen reader doesn't read when an email has been removed (FluentUI v8),Area: Accessibility Status: No Recent Activity Needs: Author Feedback Component: PeoplePicker Area: NVDA,"### Environment Information
+
+- **Package version(s)**: 8.41.2
+- **Browser and OS versions**: Microsoft Edge
+
+### Describe the issue:
+
+When using NVDA screen reader and you use keyboard navigation to go to the ""remove"" button to remove an email from the people picker. If you listen to the screen reader, it never reads that the email has been removed. What's happening is that the people picker is taking focus of the input box to quickly after you click the ""remove"" button which results the screen reader to read ""People picker combo box collapsed has auto complete editable blank"" instead of ""removed (email)"". I did notice that the screen reader does notice the ""removed (email)"" but doesn't have enough time to read it due to it taking focus of the input box.
+
+
+### Please provide a reproduction of the issue in a codepen: https://fabricweb.z5.web.core.windows.net/oufr/8.50.0/#/examples/peoplepicker
+
+#### Actual behavior:
+
+NVDA Screen reader doesn't read to the user ""removed (email)"" instead reads ""People picker combo box collapsed has auto complete editable blank""
+
+#### Expected behavior:
+
+NVDA Screen reader to say to the user ""removed (email)""
+
+### Documentation describing expected behavior
+
+https://www.w3.org/WAI/WCAG21/Understanding/info-and-relationships.html
+",1.0,"[PeoplePicker] - NVDA Screen reader doesn't read when an email has been removed (FluentUI v8) - ### Environment Information
+
+- **Package version(s)**: 8.41.2
+- **Browser and OS versions**: Microsoft Edge
+
+### Describe the issue:
+
+When using NVDA screen reader and you use keyboard navigation to go to the ""remove"" button to remove an email from the people picker. If you listen to the screen reader, it never reads that the email has been removed. What's happening is that the people picker is taking focus of the input box to quickly after you click the ""remove"" button which results the screen reader to read ""People picker combo box collapsed has auto complete editable blank"" instead of ""removed (email)"". I did notice that the screen reader does notice the ""removed (email)"" but doesn't have enough time to read it due to it taking focus of the input box.
+
+
+### Please provide a reproduction of the issue in a codepen: https://fabricweb.z5.web.core.windows.net/oufr/8.50.0/#/examples/peoplepicker
+
+#### Actual behavior:
+
+NVDA Screen reader doesn't read to the user ""removed (email)"" instead reads ""People picker combo box collapsed has auto complete editable blank""
+
+#### Expected behavior:
+
+NVDA Screen reader to say to the user ""removed (email)""
+
+### Documentation describing expected behavior
+
+https://www.w3.org/WAI/WCAG21/Understanding/info-and-relationships.html
+",1, nvda screen reader doesn t read when an email has been removed fluentui environment information package version s browser and os versions microsoft edge describe the issue when using nvda screen reader and you use keyboard navigation to go to the remove button to remove an email from the people picker if you listen to the screen reader it never reads that the email has been removed what s happening is that the people picker is taking focus of the input box to quickly after you click the remove button which results the screen reader to read people picker combo box collapsed has auto complete editable blank instead of removed email i did notice that the screen reader does notice the removed email but doesn t have enough time to read it due to it taking focus of the input box please provide a reproduction of the issue in a codepen actual behavior nvda screen reader doesn t read to the user removed email instead reads people picker combo box collapsed has auto complete editable blank expected behavior nvda screen reader to say to the user removed email documentation describing expected behavior ,1
+400538,27289352491.0,IssuesEvent,2023-02-23 15:34:39,mindsdb/mindsdb,https://api.github.com/repos/mindsdb/mindsdb,opened,[Docs]: ,documentation,"### Short description of what should be added or improved
+
+## Instructions
+
+Here are step by step instructions
+1. Go to ``mindsdb/integration/handler/ckan-handler/README.md``
+2. Delete line number 27, preview should look like this :
+```python
+ckan.action.datastore_search_sql(sql='SELECT * FROM ""resource_id""')
+```
+instead of:
+```python
+...
+ckan.action.datastore_search_sql(sql='SELECT * FROM ""resource_id""')
+```
+3. Save changes and create a PR
+
+## Tip for first contributors
+Do sign the CLA when asked to, by commenting :
+```
+I have read the CLA Document and I hereby sign the CLA
+```
+
+### Video or screenshots
+
+_No response_
+
+### Anything else?
+
+_No response_",1.0,"[Docs]: - ### Short description of what should be added or improved
+
+## Instructions
+
+Here are step by step instructions
+1. Go to ``mindsdb/integration/handler/ckan-handler/README.md``
+2. Delete line number 27, preview should look like this :
+```python
+ckan.action.datastore_search_sql(sql='SELECT * FROM ""resource_id""')
+```
+instead of:
+```python
+...
+ckan.action.datastore_search_sql(sql='SELECT * FROM ""resource_id""')
+```
+3. Save changes and create a PR
+
+## Tip for first contributors
+Do sign the CLA when asked to, by commenting :
+```
+I have read the CLA Document and I hereby sign the CLA
+```
+
+### Video or screenshots
+
+_No response_
+
+### Anything else?
+
+_No response_",0, short description of what should be added or improved instructions here are step by step instructions go to mindsdb integration handler ckan handler readme md delete line number preview should look like this python ckan action datastore search sql sql select from resource id instead of python ckan action datastore search sql sql select from resource id save changes and create a pr tip for first contributors do sign the cla when asked to by commenting i have read the cla document and i hereby sign the cla video or screenshots no response anything else no response ,0
+80904,30591916530.0,IssuesEvent,2023-07-21 17:51:01,department-of-veterans-affairs/va.gov-team,https://api.github.com/repos/department-of-veterans-affairs/va.gov-team,closed,"[Content organization] Your page doesn't follow heading hierarchy. Headings aren't properly nested, and headings and/or labels aren't relevant to the page. (02.04.1)",content design accessibility ia a11y-defect-2 collab-cycle-feedback Staging CCIssue02.04 CC-Dashboard MHV-Secure-Messaging,"### General Information
+
+#### VFS team name
+MHV
+
+#### VFS product name
+My HealtheVet
+
+#### VFS feature name
+Secure Messaging
+
+#### Point of Contact/Reviewers
+Brian DeConinck - @briandeconinck - Accessibility
+
+*For more information on how to interpret this ticket, please refer to the [Anatomy of a Staging Review issue ticket](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/Anatomy-of-a-Staging-Review-Issue-ticket.2060320997.html) guidance on Platform Website.
+
+---
+
+### Platform Issue
+Your page doesn't follow heading hierarchy. Headings aren't properly nested, and headings and/or labels aren't relevant to the page.
+
+### Issue Details
+On the [/secure-messages/compose](https://staging.va.gov/my-health/secure-messages/compose) interstitial page, the content starting with ""Your care team"" and ending with the ""Continue to start message"" link is wrapped in a div with `role=""heading"" aria-level=""2""`. That's indicating to screen reader users that most of the content of the page should be interpreted as a single H2 heading --- which I don't think is the intention here.
+
+### Link, screenshot or steps to recreate
+As currently coded:
+
+```
+
+
Only use messages for non-urgent needs
+
+
Your care team may take up to 3 business days to reply.
+
If you need help sooner, use one of these urgent communication options:
+
+
If you’re in crisis or having thoughts of suicide, connect with our Veterans Crisis Line. We offer confidential support anytime, day or night.
+
If you think your life or health is in danger, call or go to the nearest emergency room.
+
+```
+
+### VA.gov Experience Standard
+[Category Number 02, Issue Number 04](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/VA.gov-experience-standards.1683980311.html)
+
+### Other References
+WCAG SC 1.3.1_A
+WCAG SC 2.4.6_AA
+
+### Platform Recommendation
+I don't think the heading role or aria-level get you anything here, and if it was intentional I'd be curious to understand the thought process. My recommendation would be to remove those attributes entirely and look for another solution to whatever the intention here was.
+
+---
+
+### VFS Guidance
+
+- Close the ticket when the issue has been resolved or validated by your Product Owner
+- If your team has additional questions or needs Platform help validating the issue, please comment on the ticket
+- Some feedback provided may be out of scope for your iteration of the product, however, Platform's OCTO leadership has stated that all identified issues need to be documented and it is still your responsibility to resolve the issue.
+- If you do not believe that this Staging Review issue ticket is the responsibility of your team, comment below providing an explanation and who you believe is responsible. Please tag the Point of Contact/Reviewers. Governance team will research and will follow up.",1.0,"[Content organization] Your page doesn't follow heading hierarchy. Headings aren't properly nested, and headings and/or labels aren't relevant to the page. (02.04.1) - ### General Information
+
+#### VFS team name
+MHV
+
+#### VFS product name
+My HealtheVet
+
+#### VFS feature name
+Secure Messaging
+
+#### Point of Contact/Reviewers
+Brian DeConinck - @briandeconinck - Accessibility
+
+*For more information on how to interpret this ticket, please refer to the [Anatomy of a Staging Review issue ticket](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/Anatomy-of-a-Staging-Review-Issue-ticket.2060320997.html) guidance on Platform Website.
+
+---
+
+### Platform Issue
+Your page doesn't follow heading hierarchy. Headings aren't properly nested, and headings and/or labels aren't relevant to the page.
+
+### Issue Details
+On the [/secure-messages/compose](https://staging.va.gov/my-health/secure-messages/compose) interstitial page, the content starting with ""Your care team"" and ending with the ""Continue to start message"" link is wrapped in a div with `role=""heading"" aria-level=""2""`. That's indicating to screen reader users that most of the content of the page should be interpreted as a single H2 heading --- which I don't think is the intention here.
+
+### Link, screenshot or steps to recreate
+As currently coded:
+
+```
+
+
Only use messages for non-urgent needs
+
+
Your care team may take up to 3 business days to reply.
+
If you need help sooner, use one of these urgent communication options:
+
+
If you’re in crisis or having thoughts of suicide, connect with our Veterans Crisis Line. We offer confidential support anytime, day or night.
+
If you think your life or health is in danger, call or go to the nearest emergency room.
+
+```
+
+### VA.gov Experience Standard
+[Category Number 02, Issue Number 04](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/VA.gov-experience-standards.1683980311.html)
+
+### Other References
+WCAG SC 1.3.1_A
+WCAG SC 2.4.6_AA
+
+### Platform Recommendation
+I don't think the heading role or aria-level get you anything here, and if it was intentional I'd be curious to understand the thought process. My recommendation would be to remove those attributes entirely and look for another solution to whatever the intention here was.
+
+---
+
+### VFS Guidance
+
+- Close the ticket when the issue has been resolved or validated by your Product Owner
+- If your team has additional questions or needs Platform help validating the issue, please comment on the ticket
+- Some feedback provided may be out of scope for your iteration of the product, however, Platform's OCTO leadership has stated that all identified issues need to be documented and it is still your responsibility to resolve the issue.
+- If you do not believe that this Staging Review issue ticket is the responsibility of your team, comment below providing an explanation and who you believe is responsible. Please tag the Point of Contact/Reviewers. Governance team will research and will follow up.",0, your page doesn t follow heading hierarchy headings aren t properly nested and headings and or labels aren t relevant to the page general information vfs team name mhv vfs product name my healthevet vfs feature name secure messaging point of contact reviewers brian deconinck briandeconinck accessibility for more information on how to interpret this ticket please refer to the guidance on platform website platform issue your page doesn t follow heading hierarchy headings aren t properly nested and headings and or labels aren t relevant to the page issue details on the interstitial page the content starting with your care team and ending with the continue to start message link is wrapped in a div with role heading aria level that s indicating to screen reader users that most of the content of the page should be interpreted as a single heading which i don t think is the intention here link screenshot or steps to recreate as currently coded only use messages for non urgent needs your care team may take up to business days to reply if you need help sooner use one of these urgent communication options if you’re in crisis or having thoughts of suicide connect with our veterans crisis line we offer confidential support anytime day or night if you think your life or health is in danger call or go to the nearest emergency room continue to start message va gov experience standard other references wcag sc a wcag sc aa platform recommendation i don t think the heading role or aria level get you anything here and if it was intentional i d be curious to understand the thought process my recommendation would be to remove those attributes entirely and look for another solution to whatever the intention here was vfs guidance close the ticket when the issue has been resolved or validated by your product owner if your team has additional questions or needs platform help validating the issue please comment on the ticket some feedback provided may be out of scope for your iteration of the product however platform s octo leadership has stated that all identified issues need to be documented and it is still your responsibility to resolve the issue if you do not believe that this staging review issue ticket is the responsibility of your team comment below providing an explanation and who you believe is responsible please tag the point of contact reviewers governance team will research and will follow up ,0
+239601,7799873036.0,IssuesEvent,2018-06-09 01:30:38,tine20/Tine-2.0-Open-Source-Groupware-and-CRM,https://api.github.com/repos/tine20/Tine-2.0-Open-Source-Groupware-and-CRM,closed,"0005720:
+'my contacts' favorite does not work correctly",Addressbook Bug Mantis high priority,"**Reported by pschuele on 15 Feb 2012 14:41**
+
+**Version:** Milan (2012-03) Beta 4
+
+'my contacts' favorite does not work correctly
+
+it looks like the server does not send the correct node info, the users displayname is shown in the filter instead of the node/path ...
+",1.0,"0005720:
+'my contacts' favorite does not work correctly - **Reported by pschuele on 15 Feb 2012 14:41**
+
+**Version:** Milan (2012-03) Beta 4
+
+'my contacts' favorite does not work correctly
+
+it looks like the server does not send the correct node info, the users displayname is shown in the filter instead of the node/path ...
+",0, my contacts favorite does not work correctly reported by pschuele on feb version milan beta my contacts favorite does not work correctly it looks like the server does not send the correct node info the users displayname is shown in the filter instead of the node path ,0
+273033,29800381833.0,IssuesEvent,2023-06-16 07:38:37,billmcchesney1/foxtrot,https://api.github.com/repos/billmcchesney1/foxtrot,closed,CVE-2019-10247 (Medium) detected in jetty-server-9.4.11.v20180605.jar - autoclosed,Mend: dependency security vulnerability,"## CVE-2019-10247 - Medium Severity Vulnerability
+ Vulnerable Library - jetty-server-9.4.11.v20180605.jar
+
+
+
+In Eclipse Jetty version 7.x, 8.x, 9.2.27 and older, 9.3.26 and older, and 9.4.16 and older, the server running on any OS and Jetty version combination will reveal the configured fully qualified directory base resource location on the output of the 404 error for not finding a Context that matches the requested path. The default server behavior on jetty-distribution and jetty-home will include at the end of the Handler tree a DefaultHandler, which is responsible for reporting this 404 error, it presents the various configured contexts as HTML for users to click through to. This produced HTML includes output that contains the configured fully qualified directory base resource location for each context.
+
+
+
+In Eclipse Jetty version 7.x, 8.x, 9.2.27 and older, 9.3.26 and older, and 9.4.16 and older, the server running on any OS and Jetty version combination will reveal the configured fully qualified directory base resource location on the output of the 404 error for not finding a Context that matches the requested path. The default server behavior on jetty-distribution and jetty-home will include at the end of the Handler tree a DefaultHandler, which is responsible for reporting this 404 error, it presents the various configured contexts as HTML for users to click through to. This produced HTML includes output that contains the configured fully qualified directory base resource location for each context.
+
+
Direct dependency fix Resolution (com.smoketurner:dropwizard-swagger): 2.0.0-1
+
+
+
+
+
+***
+
+- [ ] Check this box to open an automated fix PR
+
+",0,cve medium detected in jetty server jar autoclosed cve medium severity vulnerability vulnerable library jetty server jar the core jetty server artifact library home page a href path to dependency file foxtrot common pom xml path to vulnerable library home wss scanner repository org eclipse jetty jetty server jetty server jar dependency hierarchy dropwizard swagger jar root library dropwizard core jar dropwizard jersey jar x jetty server jar vulnerable library found in head commit a href found in base branch master vulnerability details in eclipse jetty version x x and older and older and and older the server running on any os and jetty version combination will reveal the configured fully qualified directory base resource location on the output of the error for not finding a context that matches the requested path the default server behavior on jetty distribution and jetty home will include at the end of the handler tree a defaulthandler which is responsible for reporting this error it presents the various configured contexts as html for users to click through to this produced html includes output that contains the configured fully qualified directory base resource location for each context publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org eclipse jetty jetty server direct dependency fix resolution com smoketurner dropwizard swagger check this box to open an automated fix pr ,0
+791,14632210898.0,IssuesEvent,2020-12-23 21:47:06,planningcenter/developers,https://api.github.com/repos/planningcenter/developers,closed,Add Household Photo to People API,People feature request,"**Related Product**
+Which product is this feature related to?
+People
+
+**Is your feature request related to a problem? Please describe.**
+I would like Household Photo to be added to the Household link in the Planning Center People API.
+
+**Describe the solution you'd like**
+Add Household Photo to the Household link in the Planning Center People API.
+
+
+## I have..
+
+- [x] Reviewed the documentation found at https://developer.planning.center/docs
+- [x] Searched for previous issues asking for this feature request
+- [x] Removed all private information from this issue (credentials, tokens, emails, phone numbers, etc.)
+- [x] Reviewed my issue for completeness",1.0,"Add Household Photo to People API - **Related Product**
+Which product is this feature related to?
+People
+
+**Is your feature request related to a problem? Please describe.**
+I would like Household Photo to be added to the Household link in the Planning Center People API.
+
+**Describe the solution you'd like**
+Add Household Photo to the Household link in the Planning Center People API.
+
+
+## I have..
+
+- [x] Reviewed the documentation found at https://developer.planning.center/docs
+- [x] Searched for previous issues asking for this feature request
+- [x] Removed all private information from this issue (credentials, tokens, emails, phone numbers, etc.)
+- [x] Reviewed my issue for completeness",1,add household photo to people api related product which product is this feature related to people is your feature request related to a problem please describe i would like household photo to be added to the household link in the planning center people api describe the solution you d like add household photo to the household link in the planning center people api i have reviewed the documentation found at searched for previous issues asking for this feature request removed all private information from this issue credentials tokens emails phone numbers etc reviewed my issue for completeness,1
+581360,17291914550.0,IssuesEvent,2021-07-24 23:49:05,myConsciousness/twitter-bot-j,https://api.github.com/repos/myConsciousness/twitter-bot-j,opened,一定期間経過後のMongoDBのコレクションを一括削除する機能を追加,Priority: low Type: new feature,"# Add New Feature
+
+## 1. Feature details
+
+現在日から一定期間が経過したMongoDBのコレクションを一括削除する機能を追加する。
+コレクションを一括削除する対象ドキュメントと経過日数に関してはデータベースで管理し、
+データ駆動が可能なように設計する。
+
+## 2. Why it is necessary
+
+時間経過に比例して増大する不要なドキュメントのコレクション容量を減らし、
+サーバーのリソースを開放する。
+
+## 3. How to implement
+
+現在日から一定期間が経過したMongoDBのコレクションを一括削除する機能を追加する。
+コレクションを一括削除する対象ドキュメントと経過日数に関してはデータベースで管理し、
+データ駆動が可能なように設計する。
+
+実行タイミングは一日の総タスクが完了した24時頃とする。
+今回の対応に関してレポートの送信機能は現状考慮しない。
+
+## 4. References
+",1.0,"一定期間経過後のMongoDBのコレクションを一括削除する機能を追加 - # Add New Feature
+
+## 1. Feature details
+
+現在日から一定期間が経過したMongoDBのコレクションを一括削除する機能を追加する。
+コレクションを一括削除する対象ドキュメントと経過日数に関してはデータベースで管理し、
+データ駆動が可能なように設計する。
+
+## 2. Why it is necessary
+
+時間経過に比例して増大する不要なドキュメントのコレクション容量を減らし、
+サーバーのリソースを開放する。
+
+## 3. How to implement
+
+現在日から一定期間が経過したMongoDBのコレクションを一括削除する機能を追加する。
+コレクションを一括削除する対象ドキュメントと経過日数に関してはデータベースで管理し、
+データ駆動が可能なように設計する。
+
+実行タイミングは一日の総タスクが完了した24時頃とする。
+今回の対応に関してレポートの送信機能は現状考慮しない。
+
+## 4. References
+",0,一定期間経過後のmongodbのコレクションを一括削除する機能を追加 add new feature feature details 現在日から一定期間が経過したmongodbのコレクションを一括削除する機能を追加する。 コレクションを一括削除する対象ドキュメントと経過日数に関してはデータベースで管理し、 データ駆動が可能なように設計する。 why it is necessary 時間経過に比例して増大する不要なドキュメントのコレ��ション容量を減らし、 サーバーのリソースを開放する。 how to implement 現在日から一定期間が経過したmongodbのコレクションを一括削除する機能を追加する。 コレクションを一括削除する対象ドキュメントと経過日数に関してはデータベースで管理し、 データ駆動が可能なように設計する。 。 今回の対応に関してレポートの送信機能は現状考慮しない。 references ,0
+212,4386167304.0,IssuesEvent,2016-08-08 11:48:32,wordpress-mobile/WordPress-Android,https://api.github.com/repos/wordpress-mobile/WordPress-Android,closed,Follower & Viewers list is jumpy when you first visit the page,People Management [Type] Bug,"We currently keep only the first page of the users/followers/viewers to avoid syncing issues. In order to do that, we make a query to the DB to remove every user/follower/viewer that's not in the first page. When we define this ""first page"" we need to be careful and make sure the sorting is exactly the same as the remote responses, otherwise users will get added in the first page and there is jumping around.
+
+The users endpoint supports sorting, so we can be consistent within the app. However, follower & viewers endpoint doesn't have a sorting and only using a default sorting on the server side. The current implementation doesn't actually remove the correct set of people from DB, because the ""first page"" definition is wrong.",1.0,"Follower & Viewers list is jumpy when you first visit the page - We currently keep only the first page of the users/followers/viewers to avoid syncing issues. In order to do that, we make a query to the DB to remove every user/follower/viewer that's not in the first page. When we define this ""first page"" we need to be careful and make sure the sorting is exactly the same as the remote responses, otherwise users will get added in the first page and there is jumping around.
+
+The users endpoint supports sorting, so we can be consistent within the app. However, follower & viewers endpoint doesn't have a sorting and only using a default sorting on the server side. The current implementation doesn't actually remove the correct set of people from DB, because the ""first page"" definition is wrong.",1,follower viewers list is jumpy when you first visit the page we currently keep only the first page of the users followers viewers to avoid syncing issues in order to do that we make a query to the db to remove every user follower viewer that s not in the first page when we define this first page we need to be careful and make sure the sorting is exactly the same as the remote responses otherwise users will get added in the first page and there is jumping around the users endpoint supports sorting so we can be consistent within the app however follower viewers endpoint doesn t have a sorting and only using a default sorting on the server side the current implementation doesn t actually remove the correct set of people from db because the first page definition is wrong ,1
+21,2649873855.0,IssuesEvent,2015-03-15 11:29:09,fskreuz/Tree-Trail,https://api.github.com/repos/fskreuz/Tree-Trail,closed,Add pictures in Badge,QA people! Test these!,"As a user, I should be able to upload photos about the place being pinned.",1.0,"Add pictures in Badge - As a user, I should be able to upload photos about the place being pinned.",1,add pictures in badge as a user i should be able to upload photos about the place being pinned ,1
+1175,30668830312.0,IssuesEvent,2023-07-25 20:31:33,restincode/restincode,https://api.github.com/repos/restincode/restincode,opened,Kevin Riggins,People Add Person Needs Review,"Please fill out as much information as you can. No fields are required, but the more you can provide the better.
+
+**General Info**
+* First name: Kevin
+* Last name: Riggins
+* Handle: kriggins
+* Birth Year:
+* Death Year:
+* Link to Obituary:
+* Group Affiliations:
+* URL to main photo (or attach to Issue):
+* Description of person and/or activities:
+* Facebook memorial group URL:
+
+**Social Media Links**
+* Twitter: @kriggins
+* Github:
+* LinkedIn:
+* Facebook:
+* Other:
+
+**Contributions**
+Repeat as many times as needed
+
+* Project name:
+* Project URL:
+* Project Description:
+
+**Photo Gallery**
+* URL(s) to additional photos:
+",1.0,"Kevin Riggins - Please fill out as much information as you can. No fields are required, but the more you can provide the better.
+
+**General Info**
+* First name: Kevin
+* Last name: Riggins
+* Handle: kriggins
+* Birth Year:
+* Death Year:
+* Link to Obituary:
+* Group Affiliations:
+* URL to main photo (or attach to Issue):
+* Description of person and/or activities:
+* Facebook memorial group URL:
+
+**Social Media Links**
+* Twitter: @kriggins
+* Github:
+* LinkedIn:
+* Facebook:
+* Other:
+
+**Contributions**
+Repeat as many times as needed
+
+* Project name:
+* Project URL:
+* Project Description:
+
+**Photo Gallery**
+* URL(s) to additional photos:
+",1,kevin riggins please fill out as much information as you can no fields are required but the more you can provide the better general info first name kevin last name riggins handle kriggins birth year death year link to obituary group affiliations url to main photo or attach to issue description of person and or activities facebook memorial group url social media links twitter kriggins github linkedin facebook other contributions repeat as many times as needed project name project url project description photo gallery url s to additional photos ,1
+106573,11492524166.0,IssuesEvent,2020-02-11 21:09:54,BIAPT/Scripts,https://api.github.com/repos/BIAPT/Scripts,opened,Define the features inside the excell database,documentation,Don sent an excel file and we need to define the feature we are using over there so that everyone can understand what is hapening.,1.0,Define the features inside the excell database - Don sent an excel file and we need to define the feature we are using over there so that everyone can understand what is hapening.,0,define the features inside the excell database don sent an excel file and we need to define the feature we are using over there so that everyone can understand what is hapening ,0
+11668,7350209188.0,IssuesEvent,2018-03-08 13:37:27,Welthungerhilfe/ChildGrowthMonitor,https://api.github.com/repos/Welthungerhilfe/ChildGrowthMonitor,opened,positive feedback,enhancement scanner app usability,"successfully finishing a scanning step should give audiovisual feedback
+
+- measurement accepted sound
+- screen briefly turns green and shows big check mark",True,"positive feedback - successfully finishing a scanning step should give audiovisual feedback
+
+- measurement accepted sound
+- screen briefly turns green and shows big check mark",0,positive feedback successfully finishing a scanning step should give audiovisual feedback measurement accepted sound screen briefly turns green and shows big check mark,0
+92290,15856971157.0,IssuesEvent,2021-04-08 03:37:49,hiptest/cucumber-electron,https://api.github.com/repos/hiptest/cucumber-electron,opened,CVE-2020-28500 (Medium) detected in lodash-4.17.20.tgz,security vulnerability,"## CVE-2020-28500 - Medium Severity Vulnerability
+ Vulnerable Library - lodash-4.17.20.tgz
+
+
+
+Lodash versions prior to 4.17.21 are vulnerable to Regular Expression Denial of Service (ReDoS) via the toNumber, trim and trimEnd functions.
+ WhiteSource Note: After conducting further research, WhiteSource has determined that CVE-2020-28500 only affects environments with versions 4.0.0 to 4.17.20 of Lodash.
+
+
+
+Lodash versions prior to 4.17.21 are vulnerable to Regular Expression Denial of Service (ReDoS) via the toNumber, trim and trimEnd functions.
+ WhiteSource Note: After conducting further research, WhiteSource has determined that CVE-2020-28500 only affects environments with versions 4.0.0 to 4.17.20 of Lodash.
+
+
+
+
+
+
+
+",0,cve medium detected in lodash tgz cve medium severity vulnerability vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file cucumber electron package json path to vulnerable library cucumber electron node modules lodash package json dependency hierarchy cucumber tgz root library x lodash tgz vulnerable library found in base branch master vulnerability details lodash versions prior to are vulnerable to regular expression denial of service redos via the tonumber trim and trimend functions whitesource note after conducting further research whitesource has determined that cve only affects environments with versions to of lodash publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lodash isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree cucumber lodash isminimumfixversionavailable true minimumfixversion lodash basebranches vulnerabilityidentifier cve vulnerabilitydetails lodash versions prior to are vulnerable to regular expression denial of service redos via the tonumber trim and trimend functions n whitesource note after conducting further research whitesource has determined that cve only affects environments with versions to of lodash vulnerabilityurl ,0
+949,20993006267.0,IssuesEvent,2022-03-29 11:04:02,AtB-AS/mittatb-app,https://api.github.com/repos/AtB-AS/mittatb-app,closed,Avganger v2 - Favoritter,PI: Avganger fase 2 :people_holding_hands: Teaminitiativ :shirt: L,"### Skisser
+
+
+Krever oppdatering av BFF:
+
+- Nytt endepunkt: Utlisting av informasjon om ett eller flere stopp til visning i favoritt-tab (**potensielt samme endepunkt som i #2227**)
+- Mulighet for filtrering av avganger i quay-departures og stop-departures endepunktene.
+
+Kan løses i tre deler: #2427, #2428 og #2429
+
+**Favoritt-tab:**
+
+
+**Favoritter:**
+
+
+",1.0,"Avganger v2 - Favoritter - ### Skisser
+
+
+Krever oppdatering av BFF:
+
+- Nytt endepunkt: Utlisting av informasjon om ett eller flere stopp til visning i favoritt-tab (**potensielt samme endepunkt som i #2227**)
+- Mulighet for filtrering av avganger i quay-departures og stop-departures endepunktene.
+
+Kan løses i tre deler: #2427, #2428 og #2429
+
+**Favoritt-tab:**
+
+
+**Favoritter:**
+
+
+",1,avganger favoritter skisser krever oppdatering av bff nytt endepunkt utlisting av informasjon om ett eller flere stopp til visning i favoritt tab potensielt samme endepunkt som i mulighet for filtrering av avganger i quay departures og stop departures endepunktene kan løses i tre deler og favoritt tab favoritter ,1
+56,2883311827.0,IssuesEvent,2015-06-11 11:23:06,University-of-Potsdam-MM/UP.App,https://api.github.com/repos/University-of-Potsdam-MM/UP.App,closed,Personensuche liefert einelementiges Ergebnis,bug People to check,"Wenn ich nach ""alex"" suche kommt genau ein leeres Ergebnis zurück. Ich vermute das ist ein Fehler.
+",1.0,"Personensuche liefert einelementiges Ergebnis - Wenn ich nach ""alex"" suche kommt genau ein leeres Ergebnis zurück. Ich vermute das ist ein Fehler.
+",1,personensuche liefert einelementiges ergebnis wenn ich nach alex suche kommt genau ein leeres ergebnis zurück ich vermute das ist ein fehler ,1
+27566,2694187828.0,IssuesEvent,2015-04-01 18:50:29,empirical-org/Empirical-Core,https://api.github.com/repos/empirical-org/Empirical-Core,closed,Add new Concepts + Explore Rails Admin for admin interface,Priority: ★,"New activities cannot be added to the app until Marcello manually assigns a value to them.
+
+We could build an interface for this page. However, it is kind of a pain to build a new page for each field.
+
+@wlaurance Found out a tool called Rails Admin which automatically builds an interface based on the structure of the rails database. We might want to consider moving to this interface rather than maintaining our current interface.
+
+https://github.com/sferik/rails_admin
+
+Here is a link to the demo: http://rails-admin-tb.herokuapp.com/",1.0,"Add new Concepts + Explore Rails Admin for admin interface - New activities cannot be added to the app until Marcello manually assigns a value to them.
+
+We could build an interface for this page. However, it is kind of a pain to build a new page for each field.
+
+@wlaurance Found out a tool called Rails Admin which automatically builds an interface based on the structure of the rails database. We might want to consider moving to this interface rather than maintaining our current interface.
+
+https://github.com/sferik/rails_admin
+
+Here is a link to the demo: http://rails-admin-tb.herokuapp.com/",0,add new concepts explore rails admin for admin interface new activities cannot be added to the app until marcello manually assigns a value to them we could build an interface for this page however it is kind of a pain to build a new page for each field wlaurance found out a tool called rails admin which automatically builds an interface based on the structure of the rails database we might want to consider moving to this interface rather than maintaining our current interface here is a link to the demo ,0
+67,3129648603.0,IssuesEvent,2015-09-09 03:08:50,phetsims/website,https://api.github.com/repos/phetsims/website,closed,Team page minor tweak,enhancement People page,"@aaronsamuel137 Dave Underwood pointed out one minor tweak to the Team page. I asked @amanda-phet and she agreed.
+
+The spacing for the text to the right of Kathy's photo is a little tight. Here is more what it should look like:
+
+
+
+Please adjust when time allows (should be a quick fix).
+",1.0,"Team page minor tweak - @aaronsamuel137 Dave Underwood pointed out one minor tweak to the Team page. I asked @amanda-phet and she agreed.
+
+The spacing for the text to the right of Kathy's photo is a little tight. Here is more what it should look like:
+
+
+
+Please adjust when time allows (should be a quick fix).
+",1,team page minor tweak dave underwood pointed out one minor tweak to the team page i asked amanda phet and she agreed the spacing for the text to the right of kathy s photo is a little tight here is more what it should look like please adjust when time allows should be a quick fix ,1
+451,8273663550.0,IssuesEvent,2018-09-17 07:03:11,softwareplanet/leadboard,https://api.github.com/repos/softwareplanet/leadboard,reopened,User can see all people on people page (12),epic: people story,"## Criteria
+- [x] User can see people button on leadbord
+- [x] User can see table with all people on people page
+## Tasks
+- [ ] Add people option in contacts dropdown on leadbord
+- [ ] Create table component
+",1.0,"User can see all people on people page (12) - ## Criteria
+- [x] User can see people button on leadbord
+- [x] User can see table with all people on people page
+## Tasks
+- [ ] Add people option in contacts dropdown on leadbord
+- [ ] Create table component
+",1,user can see all people on people page criteria user can see people button on leadbord user can see table with all people on people page tasks add people option in contacts dropdown on leadbord create table component ,1
+188629,15166411264.0,IssuesEvent,2021-02-12 16:19:58,microcks/microcks,https://api.github.com/repos/microcks/microcks,closed,Update the Installing on Minikube with Operator section,component/documentation kind/task,Installation of Operator has been drastically simplified thanks to @hguerrero push up on microcks/microcks-ansible-operator#22. We should update the documentation accordingly.,1.0,Update the Installing on Minikube with Operator section - Installation of Operator has been drastically simplified thanks to @hguerrero push up on microcks/microcks-ansible-operator#22. We should update the documentation accordingly.,0,update the installing on minikube with operator section installation of operator has been drastically simplified thanks to hguerrero push up on microcks microcks ansible operator we should update the documentation accordingly ,0
+7174,9457884616.0,IssuesEvent,2019-04-17 02:32:33,evhub/coconut,https://api.github.com/repos/evhub/coconut,closed,Problems installing coconut on python 3.7,compatibility,"When installing coconut in python 3.7 using pip:
+
+`pip install coconut[all]`
+
+cPyparsing is failed to build, it doesn't support python 3.7 yet...
+
+I then tried to install coconut using conda. Installation succeeded, but couldn't launch jupyter console using
+
+`coconut --jupyter console`
+
+Error message as follow:
+
+`ImportError: cannot import name 'create_prompt_application' from 'prompt_toolkit.shortcuts'`
+
+After some Googling I found that the jupyter console didn't support prompt_toolkit 2.0 yet, see:
+
+[https://github.com/ipython/ipython/issues/11261](https://github.com/ipython/ipython/issues/11261)
+
+and it could be solved by downgrading prompt_toolkit.
+
+`conda install prompt_toolkit=1.0.15`
+
+Everything is fine afterwards.",True,"Problems installing coconut on python 3.7 - When installing coconut in python 3.7 using pip:
+
+`pip install coconut[all]`
+
+cPyparsing is failed to build, it doesn't support python 3.7 yet...
+
+I then tried to install coconut using conda. Installation succeeded, but couldn't launch jupyter console using
+
+`coconut --jupyter console`
+
+Error message as follow:
+
+`ImportError: cannot import name 'create_prompt_application' from 'prompt_toolkit.shortcuts'`
+
+After some Googling I found that the jupyter console didn't support prompt_toolkit 2.0 yet, see:
+
+[https://github.com/ipython/ipython/issues/11261](https://github.com/ipython/ipython/issues/11261)
+
+and it could be solved by downgrading prompt_toolkit.
+
+`conda install prompt_toolkit=1.0.15`
+
+Everything is fine afterwards.",0,problems installing coconut on python when installing coconut in python using pip pip install coconut cpyparsing is failed to build it doesn t support python yet i then tried to install coconut using conda installation succeeded but couldn t launch jupyter console using coconut jupyter console error message as follow importerror cannot import name create prompt application from prompt toolkit shortcuts after some googling i found that the jupyter console didn t support prompt toolkit yet see and it could be solved by downgrading prompt toolkit conda install prompt toolkit everything is fine afterwards ,0
+96805,12158188369.0,IssuesEvent,2020-04-26 02:25:20,agda/agda-stdlib,https://api.github.com/repos/agda/agda-stdlib,closed,"Remove records for IsCommutative, Commutative ?",library-design status: wontfix,"lib-1.3 has 10 parasitic records of (Is)Commutative,
+10 levels in the bundle hierarchy for algebra.
+Code and effort are spent in filling instances of these records, in standard library as well as in applied programs.
+
+This can be improved by defining Commutative via Σ:
+```
+module _ {α α=}
+ where
+ CommutativeSemigroup : Set _
+ CommutativeSemigroup =
+ Σ (Semigroup α α=) (\H → let open Semigroup H in Commutative _≈_ _∙_)
+ ...
+ CommutativeRing : Set _
+ CommutativeRing =
+ Σ (Ring α α=) (\R → let open Ring R in Commutative _≈_ _*_)
+```
+These are five declarations.
+Usage:
+```
+foo : {α α=} → CommutativeRing α α= → Foo
+foo (R , *-comm) =
+ < evaluate with using (R : Ring) and *-comm >
+```
+For example, this allows us to remove 13 record instances of IsCommutative and Commutative from Data.Integer.Properties.
+
+A small drawback: for example,
+instead of ``CommutativeRing.1# C`` one will need to write ``Ring.1# (proj₁ C)``,
+or ``let (R, *comm) = C in Ring.1# R``
+(and ``*comm`` will be sometimes used ...).
+Similarly it is with ``open Commutative``
+(if only Σ does not export from its first argument record).
+
+?
+ ",1.0,"Remove records for IsCommutative, Commutative ? - lib-1.3 has 10 parasitic records of (Is)Commutative,
+10 levels in the bundle hierarchy for algebra.
+Code and effort are spent in filling instances of these records, in standard library as well as in applied programs.
+
+This can be improved by defining Commutative via Σ:
+```
+module _ {α α=}
+ where
+ CommutativeSemigroup : Set _
+ CommutativeSemigroup =
+ Σ (Semigroup α α=) (\H → let open Semigroup H in Commutative _≈_ _∙_)
+ ...
+ CommutativeRing : Set _
+ CommutativeRing =
+ Σ (Ring α α=) (\R → let open Ring R in Commutative _≈_ _*_)
+```
+These are five declarations.
+Usage:
+```
+foo : {α α=} → CommutativeRing α α= → Foo
+foo (R , *-comm) =
+ < evaluate with using (R : Ring) and *-comm >
+```
+For example, this allows us to remove 13 record instances of IsCommutative and Commutative from Data.Integer.Properties.
+
+A small drawback: for example,
+instead of ``CommutativeRing.1# C`` one will need to write ``Ring.1# (proj₁ C)``,
+or ``let (R, *comm) = C in Ring.1# R``
+(and ``*comm`` will be sometimes used ...).
+Similarly it is with ``open Commutative``
+(if only Σ does not export from its first argument record).
+
+?
+ ",0,remove records for iscommutative commutative lib has parasitic records of is commutative levels in the bundle hierarchy for algebra code and effort are spent in filling instances of these records in standard library as well as in applied programs this can be improved by defining commutative via σ module α α where commutativesemigroup set commutativesemigroup σ semigroup α α h → let open semigroup h in commutative ≈ ∙ commutativering set commutativering σ ring α α r → let open ring r in commutative ≈ these are five declarations usage foo α α → commutativering α α → foo foo r comm for example this allows us to remove record instances of iscommutative and commutative from data integer properties a small drawback for example instead of commutativering c one will need to write ring proj₁ c or let r comm c in ring r and comm will be sometimes used similarly it is with open commutative if only σ does not export from its first argument record ,0
+108317,9301031748.0,IssuesEvent,2019-03-23 18:28:28,adalsa91/vedetra-server,https://api.github.com/repos/adalsa91/vedetra-server,closed,Error al insertar datos de prueba en bbdd de tests,bug test,Ala lanzar los tests se produce un error debido a que se intenta insertar los datos de prueba en la base de datos antes de haber creado la estructura.,1.0,Error al insertar datos de prueba en bbdd de tests - Ala lanzar los tests se produce un error debido a que se intenta insertar los datos de prueba en la base de datos antes de haber creado la estructura.,0,error al insertar datos de prueba en bbdd de tests ala lanzar los tests se produce un error debido a que se intenta insertar los datos de prueba en la base de datos antes de haber creado la estructura ,0
+29671,5809397001.0,IssuesEvent,2017-05-04 13:20:03,OpenMS/OpenMS,https://api.github.com/repos/OpenMS/OpenMS,closed,TOPPAS/ExecutePipeline: problems with long paths,defect major TOPPAS,"TOPPAS and ExecutePipeline have a limit to the lengths of paths that they can handle (255 characters) and will quit with an error if a path gets too long. There are several problems with this:
+1. The limitation to 260 characters comes from Windows and does not exist on Mac/Linux. (There, the length of a file name is limited to 256 characters, but paths can get much longer.) Therefore, it would be helpful if the length limit would only be enforced on Windows.
+2. The limit applies to files created by TOPPAS/ExecutePipeline, but no attempt is made to create shorter directory or file names if the limit would be reached.
+3. Most annoyingly, the file names created during the initial ""dry run"" - which can then lead to abortion of the workflow - can be much longer than the file names used during actual processing:
+> Error: Unexpected internal error (the file '/hps/nobackup/research/proteomics/sanger-data/results/CHPP_TESTIS_Tricine_1006/TOPPAS_out/043-PeptideIndexer-out/CHPP_TESTIS_Tricine_1006.mzML.idXML_to_CHPP_TESTIS_Tricine_1006.mzML.idXML_merged.idXML.idXML.unknown.idXML.idXML.idXML.idXML.idXML.idXML.idXML' is too long (256 chars) and exceeds the allowed limit of 255. Use shorter filenames and/or less sub-directories.)
+
+The actual name of the output file, if it could be created, would be simply ""CHPP_TESTIS_Tricine_1006.idXML"", not ""CHPP_TESTIS_Tricine_1006.mzML.idXML_to_CHPP_TESTIS_Tricine_1006.mzML.idXML_merged.idXML.idXML.unknown.idXML.idXML.idXML.idXML.idXML.idXML.idXML"".",1.0,"TOPPAS/ExecutePipeline: problems with long paths - TOPPAS and ExecutePipeline have a limit to the lengths of paths that they can handle (255 characters) and will quit with an error if a path gets too long. There are several problems with this:
+1. The limitation to 260 characters comes from Windows and does not exist on Mac/Linux. (There, the length of a file name is limited to 256 characters, but paths can get much longer.) Therefore, it would be helpful if the length limit would only be enforced on Windows.
+2. The limit applies to files created by TOPPAS/ExecutePipeline, but no attempt is made to create shorter directory or file names if the limit would be reached.
+3. Most annoyingly, the file names created during the initial ""dry run"" - which can then lead to abortion of the workflow - can be much longer than the file names used during actual processing:
+> Error: Unexpected internal error (the file '/hps/nobackup/research/proteomics/sanger-data/results/CHPP_TESTIS_Tricine_1006/TOPPAS_out/043-PeptideIndexer-out/CHPP_TESTIS_Tricine_1006.mzML.idXML_to_CHPP_TESTIS_Tricine_1006.mzML.idXML_merged.idXML.idXML.unknown.idXML.idXML.idXML.idXML.idXML.idXML.idXML' is too long (256 chars) and exceeds the allowed limit of 255. Use shorter filenames and/or less sub-directories.)
+
+The actual name of the output file, if it could be created, would be simply ""CHPP_TESTIS_Tricine_1006.idXML"", not ""CHPP_TESTIS_Tricine_1006.mzML.idXML_to_CHPP_TESTIS_Tricine_1006.mzML.idXML_merged.idXML.idXML.unknown.idXML.idXML.idXML.idXML.idXML.idXML.idXML"".",0,toppas executepipeline problems with long paths toppas and executepipeline have a limit to the lengths of paths that they can handle characters and will quit with an error if a path gets too long there are several problems with this the limitation to characters comes from windows and does not exist on mac linux there the length of a file name is limited to characters but paths can get much longer therefore it would be helpful if the length limit would only be enforced on windows the limit applies to files created by toppas executepipeline but no attempt is made to create shorter directory or file names if the limit would be reached most annoyingly the file names created during the initial dry run which can then lead to abortion of the workflow can be much longer than the file names used during actual processing error unexpected internal error the file hps nobackup research proteomics sanger data results chpp testis tricine toppas out peptideindexer out chpp testis tricine mzml idxml to chpp testis tricine mzml idxml merged idxml idxml unknown idxml idxml idxml idxml idxml idxml idxml is too long chars and exceeds the allowed limit of use shorter filenames and or less sub directories the actual name of the output file if it could be created would be simply chpp testis tricine idxml not chpp testis tricine mzml idxml to chpp testis tricine mzml idxml merged idxml idxml unknown idxml idxml idxml idxml idxml idxml idxml ,0
+326,5923142695.0,IssuesEvent,2017-05-23 07:02:05,awesome-jobs/vietnam,https://api.github.com/repos/awesome-jobs/vietnam,closed,Ematic Solutions - Customer Success Engineer - HCMC - FT,Full-time People Skill Saigon,"Do you want to join a recently funded, fast-growing software start-up? Work with international team members that are passionate about digital marketing? Learn a ton and have fun doing it? Ematic Solutions just might be the place for you! We’re a SaaS provider offering an Email Intelligence Platform that helps digital marketers nimbly build and deploy sophisticated and high-ROI email marketing programs. Our Email Intelligence Platform seamlessly integrates with a company’s Email Service Provider (ESP) and leverages a powerful suite of plug-and-play apps. Our apps have been proven to massively increase conversions by driving database growth and improving customer engagement. Our goal is to make the digital marketer smarter, and you’ll play a role in that relationship. The Ematic Solutions team is a small group of smart people from different backgrounds, based around the world, and share one passion: helping customers achieve better results. At lunch we find ourselves talking shop, and actually enjoying it. We’re a nimble, flexible and fun group, where your good ideas can make a difference.
+## Location
+- Saigon, Vietnam
+## Salary Expectation
+- Competitive
+## Requirements
+- Impeccable written and verbal communication skills in both Vietnamese and English; and experience with multicultural environments is a plus
+- Problem Solving skills
+- Data Structure and optimizations experience and skills
+- Comfortable with taking on leadership roles and responsibilities
+- Strong SQL, HTML5 and PHP knowledge is strongly preferred
+- Strong MailChimp API knowledge is a plus
+- 0-2 years of relevant experience
+- Detail oriented and analytical
+- Must be willing to work flexible hours to adapt to customer requests
+- Strong team player, but still a self-starter
+- Thrives in a multitasking environment and can adjust priorities on-the-fly
+- You are tech-savvy - learning new tools does not scare you!
+- You’re driven, and eager to learn, adapt and perfect your work
+- You want to help and serve our customers: they win, so you win
+### Responsibilities
+
+Great power comes with great responsibilities!
+### You’ll be responsible for
+- Attending customer meetings, interfacing with the client.
+- Preparing data specification for customers
+- Helping with data integration, cleaning data, porting, lists consolidation, etc.
+- Preparing API working payload examples for the customers
+- Work with customers to establish and achieve critical goals and other KPIs
+- Help customers understand MailChimp basics (e.g. lists consolidation), and answering data/technical-related questions
+- Work to identify and/or develop up-sell opportunities
+- Advocate customer needs/issues cross-departmentally.
+### Contact
+
+Nick Ippel, Vietnam Country Manager, nick.ippel@ematicsolutions.com | 012165180498
+
+
+",1.0,"Ematic Solutions - Customer Success Engineer - HCMC - FT - Do you want to join a recently funded, fast-growing software start-up? Work with international team members that are passionate about digital marketing? Learn a ton and have fun doing it? Ematic Solutions just might be the place for you! We’re a SaaS provider offering an Email Intelligence Platform that helps digital marketers nimbly build and deploy sophisticated and high-ROI email marketing programs. Our Email Intelligence Platform seamlessly integrates with a company’s Email Service Provider (ESP) and leverages a powerful suite of plug-and-play apps. Our apps have been proven to massively increase conversions by driving database growth and improving customer engagement. Our goal is to make the digital marketer smarter, and you’ll play a role in that relationship. The Ematic Solutions team is a small group of smart people from different backgrounds, based around the world, and share one passion: helping customers achieve better results. At lunch we find ourselves talking shop, and actually enjoying it. We’re a nimble, flexible and fun group, where your good ideas can make a difference.
+## Location
+- Saigon, Vietnam
+## Salary Expectation
+- Competitive
+## Requirements
+- Impeccable written and verbal communication skills in both Vietnamese and English; and experience with multicultural environments is a plus
+- Problem Solving skills
+- Data Structure and optimizations experience and skills
+- Comfortable with taking on leadership roles and responsibilities
+- Strong SQL, HTML5 and PHP knowledge is strongly preferred
+- Strong MailChimp API knowledge is a plus
+- 0-2 years of relevant experience
+- Detail oriented and analytical
+- Must be willing to work flexible hours to adapt to customer requests
+- Strong team player, but still a self-starter
+- Thrives in a multitasking environment and can adjust priorities on-the-fly
+- You are tech-savvy - learning new tools does not scare you!
+- You’re driven, and eager to learn, adapt and perfect your work
+- You want to help and serve our customers: they win, so you win
+### Responsibilities
+
+Great power comes with great responsibilities!
+### You’ll be responsible for
+- Attending customer meetings, interfacing with the client.
+- Preparing data specification for customers
+- Helping with data integration, cleaning data, porting, lists consolidation, etc.
+- Preparing API working payload examples for the customers
+- Work with customers to establish and achieve critical goals and other KPIs
+- Help customers understand MailChimp basics (e.g. lists consolidation), and answering data/technical-related questions
+- Work to identify and/or develop up-sell opportunities
+- Advocate customer needs/issues cross-departmentally.
+### Contact
+
+Nick Ippel, Vietnam Country Manager, nick.ippel@ematicsolutions.com | 012165180498
+
+
+",1,ematic solutions customer success engineer hcmc ft do you want to join a recently funded fast growing software start up work with international team members that are passionate about digital marketing learn a ton and have fun doing it ematic solutions just might be the place for you we’re a saas provider offering an email intelligence platform that helps digital marketers nimbly build and deploy sophisticated and high roi email marketing programs our email intelligence platform seamlessly integrates with a company’s email service provider esp and leverages a powerful suite of plug and play apps our apps have been proven to massively increase conversions by driving database growth and improving customer engagement our goal is to make the digital marketer smarter and you’ll play a role in that relationship the ematic solutions team is a small group of smart people from different backgrounds based around the world and share one passion helping customers achieve better results at lunch we find ourselves talking shop and actually enjoying it we’re a nimble flexible and fun group where your good ideas can make a difference location saigon vietnam salary expectation competitive requirements impeccable written and verbal communication skills in both vietnamese and english and experience with multicultural environments is a plus problem solving skills data structure and optimizations experience and skills comfortable with taking on leadership roles and responsibilities strong sql and php knowledge is strongly preferred strong mailchimp api knowledge is a plus years of relevant experience detail oriented and analytical must be willing to work flexible hours to adapt to customer requests strong team player but still a self starter thrives in a multitasking environment and can adjust priorities on the fly you are tech savvy learning new tools does not scare you you’re driven and eager to learn adapt and perfect your work you want to help and serve our customers they win so you win responsibilities great power comes with great responsibilities you’ll be responsible for attending customer meetings interfacing with the client preparing data specification for customers helping with data integration cleaning data porting lists consolidation etc preparing api working payload examples for the customers work with customers to establish and achieve critical goals and other kpis help customers understand mailchimp basics e g lists consolidation and answering data technical related questions work to identify and or develop up sell opportunities advocate customer needs issues cross departmentally contact nick ippel vietnam country manager nick ippel ematicsolutions com ,1
+1041,25169525861.0,IssuesEvent,2022-11-11 01:03:34,jongfeel/BookReview,https://api.github.com/repos/jongfeel/BookReview,closed,4부 1장 비판을 해야만 한다면 이렇게 시작하라,2022 How to Win Friends & Influence People,"### 1장 비판을 해야만 한다면 이렇게 시작하라
+
+규칙 1: 칭찬과 진심에서 우러나온 감사로 대화를 시작하라.
+Begin with praise and honest appreciation.",1.0,"4부 1장 비판을 해야만 한다면 이렇게 시작하라 - ### 1장 비판을 해야만 한다면 이렇게 시작하라
+
+규칙 1: 칭찬과 진심에서 우러나온 감사로 대화를 시작하라.
+Begin with praise and honest appreciation.",1, 비판을 해야만 한다면 이렇게 시작하라 비판을 해야만 한다면 이렇게 시작하라 규칙 칭찬과 진심에서 우러나온 감사로 대화를 시작하라 begin with praise and honest appreciation ,1
+221,4539729841.0,IssuesEvent,2016-09-09 12:16:55,wordpress-mobile/WordPress-Android,https://api.github.com/repos/wordpress-mobile/WordPress-Android,closed,People: User role from Team shows up next to non-user viewer,People Management [Type] Bug,"### Expected behavior
+When a user is going through the Team, Followers, Email Followers, and Viewers, the correct user role (or lack of) should show up next to the username.
+
+### Actual behavior
+Going from Team > Followers > Viewers in rapid succession results in the user role of someone in Team showing up as the user role next to the Viewer's username.
+
+
+
+
+### Steps to reproduce the behavior
+1) Go to Site dashboard > People.
+2) Add viewer to site.
+3) Start with Team selected, then select Followers (or Email Followers), then Viewers.
+
+Result: A user role of a user in Team will show up as the user role of the viewer, despite the viewer not having a user role.
+
+You must go from Team > Followers or Email Followers > Viewers to replicate the issue, going from Team > Viewers will result in seeing no user role next to the viewer, which is correct.
+
+##### Tested on Samsung Galaxy S6 Edge Plus, Android 6.0.1, WPAndroid alpha-20
+",1.0,"People: User role from Team shows up next to non-user viewer - ### Expected behavior
+When a user is going through the Team, Followers, Email Followers, and Viewers, the correct user role (or lack of) should show up next to the username.
+
+### Actual behavior
+Going from Team > Followers > Viewers in rapid succession results in the user role of someone in Team showing up as the user role next to the Viewer's username.
+
+
+
+
+### Steps to reproduce the behavior
+1) Go to Site dashboard > People.
+2) Add viewer to site.
+3) Start with Team selected, then select Followers (or Email Followers), then Viewers.
+
+Result: A user role of a user in Team will show up as the user role of the viewer, despite the viewer not having a user role.
+
+You must go from Team > Followers or Email Followers > Viewers to replicate the issue, going from Team > Viewers will result in seeing no user role next to the viewer, which is correct.
+
+##### Tested on Samsung Galaxy S6 Edge Plus, Android 6.0.1, WPAndroid alpha-20
+",1,people user role from team shows up next to non user viewer expected behavior when a user is going through the team followers email followers and viewers the correct user role or lack of should show up next to the username actual behavior going from team followers viewers in rapid succession results in the user role of someone in team showing up as the user role next to the viewer s username steps to reproduce the behavior go to site dashboard people add viewer to site start with team selected then select followers or email followers then viewers result a user role of a user in team will show up as the user role of the viewer despite the viewer not having a user role you must go from team followers or email followers viewers to replicate the issue going from team viewers will result in seeing no user role next to the viewer which is correct tested on samsung galaxy edge plus android wpandroid alpha ,1
+947,20992555360.0,IssuesEvent,2022-03-29 10:38:09,AtB-AS/mittatb-app,https://api.github.com/repos/AtB-AS/mittatb-app,closed,Nullstilling av reisesøk,:iphone: 1.17 :people_holding_hands: Teaminitiativ,"## Origin
+
+Etter oppsøkende brukertest på gata (aka geriljatest) så lærte vi om ønsket om å nullstille et reisesøk, enten fordi du søkte feil, eller om du skal gjøre et nytt søk.
+
+**Også nevnt på intercom:**
+
+> _Når jeg har gjort et reisesøk og ønsket å gjøre et nytt hadde det vært enkelt å kunne trykke på reisesøkikonet nede til venstre for så å komme tilbake til åpningssiden med reisesøk - denne er mye lettere å taste inn fra enn å måtte flytte fingrene opp til toppen av skjermen. Den gir meg også direkte tilgang til mine favoritter_
+
+## Motivation
+
+Etter å ha utført et reisesøk har man behov for å kunne nullstille dette. I dag skjer dette via tap på AtB-logoen men dette er ikke et etablert pattern. Konvensjonen er at man nullstiller ved å trykke på hovedelementet i navigasjonen på nytt.
+
+Sånn sett vil det fungere på samme måte som Avganger gjør i dag (ny visning). Hvis du er inne på én spesifikk holdeplass kan du alltids trykke på Avganger fra menylinja for å komme deg tilbake og se hvilke andre holdeplasser som er i nærheten.
+
+## Proposed solution
+
+- Når man har utført et reisesøk og man står på reisesøk-fanen kan man trykke på ""Reisesøk"" i hovednavigasjonen for å nullstille søket (_se illustrasjonen under_)
+- Nullstill søk fjernes fra AtB-logo
+
+
+
+
+
+## Acceptance Criteria
+
+_List of relevant acceptance criteria as a part of a QA flow_
+
+- [] Renders colors as expected in dark and light mode.
+- [] Is translatable to English and Norwegian.
+- [] Does not drain battery
+- [] Works in up to 200% font size
+- [] Is operable and perceivable using screen reader
+
+Beskrivelse av QA prosessen finnes [her](https://github.com/AtB-AS/org/blob/master/guides/quality_assurance.md#qa-in-atb-mobile-application)
+",1.0,"Nullstilling av reisesøk - ## Origin
+
+Etter oppsøkende brukertest på gata (aka geriljatest) så lærte vi om ønsket om å nullstille et reisesøk, enten fordi du søkte feil, eller om du skal gjøre et nytt søk.
+
+**Også nevnt på intercom:**
+
+> _Når jeg har gjort et reisesøk og ønsket å gjøre et nytt hadde det vært enkelt å kunne trykke på reisesøkikonet nede til venstre for så å komme tilbake til åpningssiden med reisesøk - denne er mye lettere å taste inn fra enn å måtte flytte fingrene opp til toppen av skjermen. Den gir meg også direkte tilgang til mine favoritter_
+
+## Motivation
+
+Etter å ha utført et reisesøk har man behov for å kunne nullstille dette. I dag skjer dette via tap på AtB-logoen men dette er ikke et etablert pattern. Konvensjonen er at man nullstiller ved å trykke på hovedelementet i navigasjonen på nytt.
+
+Sånn sett vil det fungere på samme måte som Avganger gjør i dag (ny visning). Hvis du er inne på én spesifikk holdeplass kan du alltids trykke på Avganger fra menylinja for å komme deg tilbake og se hvilke andre holdeplasser som er i nærheten.
+
+## Proposed solution
+
+- Når man har utført et reisesøk og man står på reisesøk-fanen kan man trykke på ""Reisesøk"" i hovednavigasjonen for å nullstille søket (_se illustrasjonen under_)
+- Nullstill søk fjernes fra AtB-logo
+
+
+
+
+
+## Acceptance Criteria
+
+_List of relevant acceptance criteria as a part of a QA flow_
+
+- [] Renders colors as expected in dark and light mode.
+- [] Is translatable to English and Norwegian.
+- [] Does not drain battery
+- [] Works in up to 200% font size
+- [] Is operable and perceivable using screen reader
+
+Beskrivelse av QA prosessen finnes [her](https://github.com/AtB-AS/org/blob/master/guides/quality_assurance.md#qa-in-atb-mobile-application)
+",1,nullstilling av reisesøk origin etter oppsøkende brukertest på gata aka geriljatest så lærte vi om ønsket om å nullstille et reisesøk enten fordi du søkte feil eller om du skal gjøre et nytt søk også nevnt på intercom når jeg har gjort et reisesøk og ønsket å gjøre et nytt hadde det vært enkelt å kunne trykke på reisesøkikonet nede til venstre for så å komme tilbake til åpningssiden med reisesøk denne er mye lettere å taste inn fra enn å måtte flytte fingrene opp til toppen av skjermen den gir meg også direkte tilgang til mine favoritter motivation etter å ha utført et reisesøk har man behov for å kunne nullstille dette i dag skjer dette via tap på atb logoen men dette er ikke et etablert pattern konvensjonen er at man nullstiller ved å trykke på hovedelementet i navigasjonen på nytt sånn sett vil det fungere på samme måte som avganger gjør i dag ny visning hvis du er inne på én spesifikk holdeplass kan du alltids trykke på avganger fra menylinja for å komme deg tilbake og se hvilke andre holdeplasser som er i nærheten proposed solution når man har utført et reisesøk og man står på reisesøk fanen kan man trykke på reisesøk i hovednavigasjonen for å nullstille søket se illustrasjonen under nullstill søk fjernes fra atb logo img width alt image src acceptance criteria list of relevant acceptance criteria as a part of a qa flow renders colors as expected in dark and light mode is translatable to english and norwegian does not drain battery works in up to font size is operable and perceivable using screen reader beskrivelse av qa prosessen finnes ,1
+49,2660242732.0,IssuesEvent,2015-03-19 04:25:33,fskreuz/Tree-Trail,https://api.github.com/repos/fskreuz/Tree-Trail,closed,Rejecting Badge Bug,QA people! Test these!,"using superadmin/admin account:
+step 1: click unapproved badge
+step 2: click reject
+2 pop-ups will appear.
+1. are you sure you want to reject location?
+after clicking ok
+2. location has been deleted
+and a check box on whether to prevent the page from adding dialogs
+
+after checking the box, and clicked ok.
+Rejecting badge would not function anymore",1.0,"Rejecting Badge Bug - using superadmin/admin account:
+step 1: click unapproved badge
+step 2: click reject
+2 pop-ups will appear.
+1. are you sure you want to reject location?
+after clicking ok
+2. location has been deleted
+and a check box on whether to prevent the page from adding dialogs
+
+after checking the box, and clicked ok.
+Rejecting badge would not function anymore",1,rejecting badge bug using superadmin admin account step click unapproved badge step click reject pop ups will appear are you sure you want to reject location after clicking ok location has been deleted and a check box on whether to prevent the page from adding dialogs after checking the box and clicked ok rejecting badge would not function anymore,1
+828,15288217430.0,IssuesEvent,2021-02-23 16:37:14,openstates/issues,https://api.github.com/repos/openstates/issues,closed,Maryland legislator issues,component:people-data type:bug,"**Short Description:** There were two legislators missing and a large address formatting error.
+
+**Missing or Incorrect legislators:**
+Reid J Novotny (House District 9A) is missing from OpenStates [https://msa.maryland.gov/msa/mdmanual/06hse/html/msa18290.html]
+
+Michael A Jackson (Senate, District 27) missing from Openstates [https://msa.maryland.gov/msa/mdmanual/05sen/html/msa14132.html]
+
+**Data Issues:**
+Every legislator, there were addresses that were either incomplete or formatted incorrectly.
+Examples:
+Adrienne Jones
+Alonzo T Washington
+Adelaide Eckardt
+Arthur Ellis
+
+
+
+",1.0,"Maryland legislator issues - **Short Description:** There were two legislators missing and a large address formatting error.
+
+**Missing or Incorrect legislators:**
+Reid J Novotny (House District 9A) is missing from OpenStates [https://msa.maryland.gov/msa/mdmanual/06hse/html/msa18290.html]
+
+Michael A Jackson (Senate, District 27) missing from Openstates [https://msa.maryland.gov/msa/mdmanual/05sen/html/msa14132.html]
+
+**Data Issues:**
+Every legislator, there were addresses that were either incomplete or formatted incorrectly.
+Examples:
+Adrienne Jones
+Alonzo T Washington
+Adelaide Eckardt
+Arthur Ellis
+
+
+
+",1,maryland legislator issues short description there were two legislators missing and a large address formatting error missing or incorrect legislators reid j novotny house district is missing from openstates michael a jackson senate district missing from openstates data issues every legislator there were addresses that were either incomplete or formatted incorrectly examples adrienne jones alonzo t washington adelaide eckardt arthur ellis ,1
+550,9669911569.0,IssuesEvent,2019-05-21 18:33:26,openopps/openopps-platform,https://api.github.com/repos/openopps/openopps-platform,closed,admin export users & tasks need tests,Admin People,"We used to have a very basic test of user export: https://github.com/18F/openopps-platform/blob/v0.9.5/test/api/sails/admin.test.js#L244
+
+Needs to be upgraded and use new fixtures.
+I could only find export user test, would be good to add tasks also
+",1.0,"admin export users & tasks need tests - We used to have a very basic test of user export: https://github.com/18F/openopps-platform/blob/v0.9.5/test/api/sails/admin.test.js#L244
+
+Needs to be upgraded and use new fixtures.
+I could only find export user test, would be good to add tasks also
+",1,admin export users tasks need tests we used to have a very basic test of user export needs to be upgraded and use new fixtures i could only find export user test would be good to add tasks also ,1
+951,21009838660.0,IssuesEvent,2022-03-30 04:47:48,metabase/metabase,https://api.github.com/repos/metabase/metabase,closed,"""Invite someone"" button overlaps activate/deactivate radio buttons on small screen widths",Type:Bug Priority:P2 Client:Mobile Web .Reproduced Administration/People,"**To Reproduce**
+Steps to reproduce the behavior:
+1. Go to Admin -> People
+2. Reduce width of screen
+
+
+
+:arrow_down: Please click the :+1: reaction instead of leaving a `+1` or `update?` comment",1.0,"""Invite someone"" button overlaps activate/deactivate radio buttons on small screen widths - **To Reproduce**
+Steps to reproduce the behavior:
+1. Go to Admin -> People
+2. Reduce width of screen
+
+
+
+:arrow_down: Please click the :+1: reaction instead of leaving a `+1` or `update?` comment",1, invite someone button overlaps activate deactivate radio buttons on small screen widths to reproduce steps to reproduce the behavior go to admin people reduce width of screen img width alt image src arrow down please click the reaction instead of leaving a or update comment,1
+453,3385358186.0,IssuesEvent,2015-11-27 11:01:40,openETCS/toolchain,https://api.github.com/repos/openETCS/toolchain,closed,Review of tracability Architecture (ends 12-Nov-2015),US-Traceabiliy-Architecture,"Here my comments on the document linked to #504
+
+- § 1.1 and Fig 2:
+ - in the figure are mixed functionnal, HW, procedural,... requirements, at the top level (for example from User stories or Cenelec Standard) and all seems to be derived up to SW level (I understand that only specification and design of SW appear on the figure, not the Validation). But I think that lots of the initial requirements can not be derived on Sw, but on other activities (quality or project plan, Validation,...) or subsystems (HW, API,...); How it is plan to take into account these exported requirements ?
+
+>> Agree. ""Derive"" is not the right general term for all the arrows. Changed figure 1and used ""transform"" instead of ""derive"" and better explained that initial requirements are transformed to subsystem and then HW or SW or data or procedures. I improved fig 2 with better alignement on EN 50128:2011 and used only term ""input for"" for relations between artefacts at this stage of the document.
+
+>> V&V not shown at this stage of the document. Added as a note.
+ - some non-functional requirements can be introduced (or derived from Cenelec standards) in openETCS quality or project plans.
+>> Yes. Do you think we need to show quality and project plans for this document? will those artefacts be >>traced to requirements?
+ - in the fig 2 it seems there is a direct traceability between SRS and Cenelec (orange arrow): I am not agree.
+>> Removed. I removed initial arrows coming from ISO 15288 vision and focused now on OpenETCS >>only. ISO15288 was just a way to introduce engineering levels and help me understanding scope of >>different requirements and models by asking partners the position in those levels.
+
+in the current state of SRS it is difficult to explicitly defined a traceability between this document and stakeholders requirements. I consider more the SRS in midway between stakeholders requirement and a real System specification, I will put it in parallel of Cenelec and User stories.
+>> OK. Done.
+ - I think validation are missing in fig 1 and 2: lots of requirements can not be derived up to SW code only, but will be link to the different test or V&V phases.
+
+>> OK. Which openETCS document can I read to add missing information?
+
+- §1.2 and Fig4 , It is necessary to clarify the data dictionary model and how it is defined (textual, SysML, Scade ?) as a Scade representation of it is one of the SW model.
+
+>> OK. ToDo.
+
+-§2.2.1:
+ - Please give clearly definition of the mining of the different arrows (for example ""refines"" seems to correspond to a SysML definition which is very different from a classical formal definition of ""refines"").
+ - why ""Documentation"" is an activity ?
+ - why ""V&V"" do not ""use"" the requirement database ?
+ - meaning of the arrows are not clear for me, so I do not understand why there are no linked between System model and requirement database or functional model and requirement data base. The figure need some comments as it is not self-sufficient for those who are not used of these notations.
+
+>> perfectly agree. I had almost same remarks than you when reading this figure the first time and I did >>not dare to remove it until now because it was not mine and because I thought it was ""validated"" after >>a previous review. As soon as I can express the traceability process through other diagrams easier to >>understand I will remove this initial figure.
+
+- §2.2.2: This means we consider only functional requirements. User stories, SRS, API or Cenelec are far to contain only functional requirements.
+
+>> yes because I wanted to focus on Functional formal model that seemed to be ""functional"". But I >>understand that this model is also behavioral and that we target an executable model, so containing >>non functional requirements. Will update this scenario with other non functional requirements taken >>into account.
+
+- Fig 7 : I do not think that the ""openETCS system designer"" is in charge of all the actions. Typically ""trace model element to SRS"" is made by SW designer, ""Create verification view"" by a verificator....
+>> OK. This was a ""generic"" term used to simplify diagram (showing several actors would make it too >>large). I will use a more generic term and will precise the different possible roles according to activities.
+
+
+- §1 and 2 : Maybe it will be nice to have a look on QA plan (WP1 https://github.com/openETCS/governance/blob/master/QA%20Plan/D1.3.1_QA_Plan.pdf), definition plan (WP2 https://github.com/openETCS/requirements) and safety plan (WP4 https://github.com/openETCS/validation/tree/master/Reports/D4.2) to have a better view of what would be expected at the beginning of the project.
+>> OK. thanks for the reference.
+
+- §3 Ok for me.
+
+-§4.2.3, for the moment the tool is Scade studio (v16.2)
+
+>> mistake. fixed.
+
+- §5, in the view of the openETCS toolchain, totally open, I am agree with the left branch (ProR linked to papyrus). However in practice the sysML model has been made with Scade system which contains an old version of papyrus not really compatible with the one in openETCS toolchain. In this case I'am not sure that ProR can be used at system level (which do not allow us to have an open-source tool for traceability !)
+
+>> OK. will take that into account.
+
+
+- § 5.1.2: How is identify the first sentence ""If the establishment....."" ? Are we sure that we shall always share such a requirement in different sub requirements with different Id ? Are we not going to lost information (for example in this case that ALL the sequence of actions shall be made in a given order) ?
+
+>> This is initial text (I did not change that assuming that it was validated). I'll look at your point.
+
+- §5, 6 and 7: Three solutions are proposed:
+ -why ? maybe an introduction in the document is missing to explain its contents and why 3 solutions are proposed
+
+>> Well: that might be a question of document organization. First version of document mentioned 1 first >> solution and I understood that this traceability solution was far from being perfect. So I have decided >> to investigate on possible improvements through alternate solutions.
+>> If this document reflects what IS DONE in the project, then I must focus on the reality only and >>perhaps conclude the document with ""current limits"". In that case I can create another document that >>would be ""proposals for improvements of traceability support by the tool chain"".
+
+ - some parts of some solutions are already implemented or largely analyzed (eg. link between ProR and payprus, use of genDoc...) other seems just propositions. It will be nice to have a clear view of what exists and can be used right now, and other elements.
+
+>> OK. I will distinguish between existing (tested) solutions and ideas for improvements.
+
+
+
+To continue depending updating and comments.",1.0,"Review of tracability Architecture (ends 12-Nov-2015) - Here my comments on the document linked to #504
+
+- § 1.1 and Fig 2:
+ - in the figure are mixed functionnal, HW, procedural,... requirements, at the top level (for example from User stories or Cenelec Standard) and all seems to be derived up to SW level (I understand that only specification and design of SW appear on the figure, not the Validation). But I think that lots of the initial requirements can not be derived on Sw, but on other activities (quality or project plan, Validation,...) or subsystems (HW, API,...); How it is plan to take into account these exported requirements ?
+
+>> Agree. ""Derive"" is not the right general term for all the arrows. Changed figure 1and used ""transform"" instead of ""derive"" and better explained that initial requirements are transformed to subsystem and then HW or SW or data or procedures. I improved fig 2 with better alignement on EN 50128:2011 and used only term ""input for"" for relations between artefacts at this stage of the document.
+
+>> V&V not shown at this stage of the document. Added as a note.
+ - some non-functional requirements can be introduced (or derived from Cenelec standards) in openETCS quality or project plans.
+>> Yes. Do you think we need to show quality and project plans for this document? will those artefacts be >>traced to requirements?
+ - in the fig 2 it seems there is a direct traceability between SRS and Cenelec (orange arrow): I am not agree.
+>> Removed. I removed initial arrows coming from ISO 15288 vision and focused now on OpenETCS >>only. ISO15288 was just a way to introduce engineering levels and help me understanding scope of >>different requirements and models by asking partners the position in those levels.
+
+in the current state of SRS it is difficult to explicitly defined a traceability between this document and stakeholders requirements. I consider more the SRS in midway between stakeholders requirement and a real System specification, I will put it in parallel of Cenelec and User stories.
+>> OK. Done.
+ - I think validation are missing in fig 1 and 2: lots of requirements can not be derived up to SW code only, but will be link to the different test or V&V phases.
+
+>> OK. Which openETCS document can I read to add missing information?
+
+- §1.2 and Fig4 , It is necessary to clarify the data dictionary model and how it is defined (textual, SysML, Scade ?) as a Scade representation of it is one of the SW model.
+
+>> OK. ToDo.
+
+-§2.2.1:
+ - Please give clearly definition of the mining of the different arrows (for example ""refines"" seems to correspond to a SysML definition which is very different from a classical formal definition of ""refines"").
+ - why ""Documentation"" is an activity ?
+ - why ""V&V"" do not ""use"" the requirement database ?
+ - meaning of the arrows are not clear for me, so I do not understand why there are no linked between System model and requirement database or functional model and requirement data base. The figure need some comments as it is not self-sufficient for those who are not used of these notations.
+
+>> perfectly agree. I had almost same remarks than you when reading this figure the first time and I did >>not dare to remove it until now because it was not mine and because I thought it was ""validated"" after >>a previous review. As soon as I can express the traceability process through other diagrams easier to >>understand I will remove this initial figure.
+
+- §2.2.2: This means we consider only functional requirements. User stories, SRS, API or Cenelec are far to contain only functional requirements.
+
+>> yes because I wanted to focus on Functional formal model that seemed to be ""functional"". But I >>understand that this model is also behavioral and that we target an executable model, so containing >>non functional requirements. Will update this scenario with other non functional requirements taken >>into account.
+
+- Fig 7 : I do not think that the ""openETCS system designer"" is in charge of all the actions. Typically ""trace model element to SRS"" is made by SW designer, ""Create verification view"" by a verificator....
+>> OK. This was a ""generic"" term used to simplify diagram (showing several actors would make it too >>large). I will use a more generic term and will precise the different possible roles according to activities.
+
+
+- §1 and 2 : Maybe it will be nice to have a look on QA plan (WP1 https://github.com/openETCS/governance/blob/master/QA%20Plan/D1.3.1_QA_Plan.pdf), definition plan (WP2 https://github.com/openETCS/requirements) and safety plan (WP4 https://github.com/openETCS/validation/tree/master/Reports/D4.2) to have a better view of what would be expected at the beginning of the project.
+>> OK. thanks for the reference.
+
+- §3 Ok for me.
+
+-§4.2.3, for the moment the tool is Scade studio (v16.2)
+
+>> mistake. fixed.
+
+- §5, in the view of the openETCS toolchain, totally open, I am agree with the left branch (ProR linked to papyrus). However in practice the sysML model has been made with Scade system which contains an old version of papyrus not really compatible with the one in openETCS toolchain. In this case I'am not sure that ProR can be used at system level (which do not allow us to have an open-source tool for traceability !)
+
+>> OK. will take that into account.
+
+
+- § 5.1.2: How is identify the first sentence ""If the establishment....."" ? Are we sure that we shall always share such a requirement in different sub requirements with different Id ? Are we not going to lost information (for example in this case that ALL the sequence of actions shall be made in a given order) ?
+
+>> This is initial text (I did not change that assuming that it was validated). I'll look at your point.
+
+- §5, 6 and 7: Three solutions are proposed:
+ -why ? maybe an introduction in the document is missing to explain its contents and why 3 solutions are proposed
+
+>> Well: that might be a question of document organization. First version of document mentioned 1 first >> solution and I understood that this traceability solution was far from being perfect. So I have decided >> to investigate on possible improvements through alternate solutions.
+>> If this document reflects what IS DONE in the project, then I must focus on the reality only and >>perhaps conclude the document with ""current limits"". In that case I can create another document that >>would be ""proposals for improvements of traceability support by the tool chain"".
+
+ - some parts of some solutions are already implemented or largely analyzed (eg. link between ProR and payprus, use of genDoc...) other seems just propositions. It will be nice to have a clear view of what exists and can be used right now, and other elements.
+
+>> OK. I will distinguish between existing (tested) solutions and ideas for improvements.
+
+
+
+To continue depending updating and comments.",0,review of tracability architecture ends nov here my comments on the document linked to § and fig in the figure are mixed functionnal hw procedural requirements at the top level for example from user stories or cenelec standard and all seems to be derived up to sw level i understand that only specification and design of sw appear on the figure not the validation but i think that lots of the initial requirements can not be derived on sw but on other activities quality or project plan validation or subsystems hw api how it is plan to take into account these exported requirements agree derive is not the right general term for all the arrows changed figure used transform instead of derive and better explained that initial requirements are transformed to subsystem and then hw or sw or data or procedures i improved fig with better alignement on en and used only term input for for relations between artefacts at this stage of the document v v not shown at this stage of the document added as a note some non functional requirements can be introduced or derived from cenelec standards in openetcs quality or project plans yes do you think we need to show quality and project plans for this document will those artefacts be traced to requirements in the fig it seems there is a direct traceability between srs and cenelec orange arrow i am not agree removed i removed initial arrows coming from iso vision and focused now on openetcs only was just a way to introduce engineering levels and help me understanding scope of different requirements and models by asking partners the position in those levels in the current state of srs it is difficult to explicitly defined a traceability between this document and stakeholders requirements i consider more the srs in midway between stakeholders requirement and a real system specification i will put it in parallel of cenelec and user stories ok done i think validation are missing in fig and lots of requirements can not be derived up to sw code only but will be link to the different test or v v phases ok which openetcs document can i read to add missing information § and it is necessary to clarify the data dictionary model and how it is defined textual sysml scade as a scade representation of it is one of the sw model ok todo § please give clearly definition of the mining of the different arrows for example refines seems to correspond to a sysml definition which is very different from a classical formal definition of refines why documentation is an activity why v v do not use the requirement database meaning of the arrows are not clear for me so i do not understand why there are no linked between system model and requirement database or functional model and requirement data base the figure need some comments as it is not self sufficient for those who are not used of these notations perfectly agree i had almost same remarks than you when reading this figure the first time and i did not dare to remove it until now because it was not mine and because i thought it was validated after a previous review as soon as i can express the traceability process through other diagrams easier to understand i will remove this initial figure § this means we consider only functional requirements user stories srs api or cenelec are far to contain only functional requirements yes because i wanted to focus on functional formal model that seemed to be functional but i understand that this model is also behavioral and that we target an executable model so containing non functional requirements will update this scenario with other non functional requirements taken into account fig i do not think that the openetcs system designer is in charge of all the actions typically trace model element to srs is made by sw designer create verification view by a verificator ok this was a generic term used to simplify diagram showing several actors would make it too large i will use a more generic term and will precise the different possible roles according to activities § and maybe it will be nice to have a look on qa plan definition plan and safety plan to have a better view of what would be expected at the beginning of the project ok thanks for the reference § ok for me § for the moment the tool is scade studio mistake fixed § in the view of the openetcs toolchain totally open i am agree with the left branch pror linked to papyrus however in practice the sysml model has been made with scade system which contains an old version of papyrus not really compatible with the one in openetcs toolchain in this case i am not sure that pror can be used at system level which do not allow us to have an open source tool for traceability ok will take that into account § how is identify the first sentence if the establishment are we sure that we shall always share such a requirement in different sub requirements with different id are we not going to lost information for example in this case that all the sequence of actions shall be made in a given order this is initial text i did not change that assuming that it was validated i ll look at your point § and three solutions are proposed why maybe an introduction in the document is missing to explain its contents and why solutions are proposed well that might be a question of document organization first version of document mentioned first solution and i understood that this traceability solution was far from being perfect so i have decided to investigate on possible improvements through alternate solutions if this document reflects what is done in the project then i must focus on the reality only and perhaps conclude the document with current limits in that case i can create another document that would be proposals for improvements of traceability support by the tool chain some parts of some solutions are already implemented or largely analyzed eg link between pror and payprus use of gendoc other seems just propositions it will be nice to have a clear view of what exists and can be used right now and other elements ok i will distinguish between existing tested solutions and ideas for improvements to continue depending updating and comments ,0
+425154,12336465371.0,IssuesEvent,2020-05-14 13:38:02,googleapis/nodejs-translate,https://api.github.com/repos/googleapis/nodejs-translate,opened,Synthesis failed for nodejs-translate,autosynth failure priority: p1 type: bug,"Hello! Autosynth couldn't regenerate nodejs-translate. :broken_heart:
+
+Here's the output from running `synth.py`:
+
+```
+2020-05-14 06:37:08,347 autosynth [INFO] > logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/nodejs-translate
+2020-05-14 06:37:08,350 synthtool [ERROR] > Failed executing git clone --single-branch https://github.com/googleapis/nodejs-translate.git /home/kbuilder/.cache/synthtool/nodejs-translate:
+
+fatal: could not create work tree dir '/home/kbuilder/.cache/synthtool/nodejs-translate': No space left on device
+
+Traceback (most recent call last):
+ File ""/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py"", line 193, in _run_module_as_main
+ ""__main__"", mod_spec)
+ File ""/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py"", line 85, in _run_code
+ exec(code, run_globals)
+ File ""/tmpfs/src/github/synthtool/autosynth/synth.py"", line 600, in
+ main()
+ File ""/tmpfs/src/github/synthtool/autosynth/synth.py"", line 472, in main
+ return _inner_main(temp_dir)
+ File ""/tmpfs/src/github/synthtool/autosynth/synth.py"", line 519, in _inner_main
+ working_repo_path = synthtool_git.clone(f""https://github.com/{args.repository}.git"")
+ File ""/tmpfs/src/github/synthtool/synthtool/sources/git.py"", line 83, in clone
+ shell.run(cmd, check=True)
+ File ""/tmpfs/src/github/synthtool/synthtool/shell.py"", line 39, in run
+ raise exc
+ File ""/tmpfs/src/github/synthtool/synthtool/shell.py"", line 33, in run
+ encoding=""utf-8"",
+ File ""/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py"", line 438, in run
+ output=stdout, stderr=stderr)
+subprocess.CalledProcessError: Command '['git', 'clone', '--single-branch', 'https://github.com/googleapis/nodejs-translate.git', PosixPath('/home/kbuilder/.cache/synthtool/nodejs-translate')]' returned non-zero exit status 128.
+
+```
+
+Google internal developers can see the full log [here](https://sponge/ea51e8ba-def4-460f-9acd-a09a559fce5d).
+",1.0,"Synthesis failed for nodejs-translate - Hello! Autosynth couldn't regenerate nodejs-translate. :broken_heart:
+
+Here's the output from running `synth.py`:
+
+```
+2020-05-14 06:37:08,347 autosynth [INFO] > logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/nodejs-translate
+2020-05-14 06:37:08,350 synthtool [ERROR] > Failed executing git clone --single-branch https://github.com/googleapis/nodejs-translate.git /home/kbuilder/.cache/synthtool/nodejs-translate:
+
+fatal: could not create work tree dir '/home/kbuilder/.cache/synthtool/nodejs-translate': No space left on device
+
+Traceback (most recent call last):
+ File ""/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py"", line 193, in _run_module_as_main
+ ""__main__"", mod_spec)
+ File ""/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py"", line 85, in _run_code
+ exec(code, run_globals)
+ File ""/tmpfs/src/github/synthtool/autosynth/synth.py"", line 600, in
+ main()
+ File ""/tmpfs/src/github/synthtool/autosynth/synth.py"", line 472, in main
+ return _inner_main(temp_dir)
+ File ""/tmpfs/src/github/synthtool/autosynth/synth.py"", line 519, in _inner_main
+ working_repo_path = synthtool_git.clone(f""https://github.com/{args.repository}.git"")
+ File ""/tmpfs/src/github/synthtool/synthtool/sources/git.py"", line 83, in clone
+ shell.run(cmd, check=True)
+ File ""/tmpfs/src/github/synthtool/synthtool/shell.py"", line 39, in run
+ raise exc
+ File ""/tmpfs/src/github/synthtool/synthtool/shell.py"", line 33, in run
+ encoding=""utf-8"",
+ File ""/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py"", line 438, in run
+ output=stdout, stderr=stderr)
+subprocess.CalledProcessError: Command '['git', 'clone', '--single-branch', 'https://github.com/googleapis/nodejs-translate.git', PosixPath('/home/kbuilder/.cache/synthtool/nodejs-translate')]' returned non-zero exit status 128.
+
+```
+
+Google internal developers can see the full log [here](https://sponge/ea51e8ba-def4-460f-9acd-a09a559fce5d).
+",0,synthesis failed for nodejs translate hello autosynth couldn t regenerate nodejs translate broken heart here s the output from running synth py autosynth logs will be written to tmpfs src github synthtool logs googleapis nodejs translate synthtool failed executing git clone single branch home kbuilder cache synthtool nodejs translate fatal could not create work tree dir home kbuilder cache synthtool nodejs translate no space left on device traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main working repo path synthtool git clone f file tmpfs src github synthtool synthtool sources git py line in clone shell run cmd check true file tmpfs src github synthtool synthtool shell py line in run raise exc file tmpfs src github synthtool synthtool shell py line in run encoding utf file home kbuilder pyenv versions lib subprocess py line in run output stdout stderr stderr subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log ,0
+1040,25127608841.0,IssuesEvent,2022-11-09 12:58:22,jongfeel/BookReview,https://api.github.com/repos/jongfeel/BookReview,closed,3부 사람들을 설득하는 12가지 방법,2022 How to Win Friends & Influence People,"- [x] #214
+- [x] #216
+- [x] #217
+- [x] #218
+- [x] #219
+- [x] #220
+- [x] #221
+- [x] #222
+- [x] #223
+- [x] #224
+- [x] #225
+- [x] #226",1.0,"3부 사람들을 설득하는 12가지 방법 - - [x] #214
+- [x] #216
+- [x] #217
+- [x] #218
+- [x] #219
+- [x] #220
+- [x] #221
+- [x] #222
+- [x] #223
+- [x] #224
+- [x] #225
+- [x] #226",1, 사람들을 설득하는 방법 ,1
+276181,30350456028.0,IssuesEvent,2023-07-11 18:32:08,mailpile/Mailpile,https://api.github.com/repos/mailpile/Mailpile,closed,BCC of encrypted messages should send/encrypt separately,Privacy / Security Mailpile-v1-is-Obsolete,"The user expectation when using BCC is that the people in the To/CC lines are unaware of the BCC'ed recipients.
+
+PGP may violate this expectation, by including the key IDs of the BCC'ed folks. We should fix this by making sure BCC'ed folks get a separate copy of the message that is encrypted only to them.
+
+Relates to #733.
+",True,"BCC of encrypted messages should send/encrypt separately - The user expectation when using BCC is that the people in the To/CC lines are unaware of the BCC'ed recipients.
+
+PGP may violate this expectation, by including the key IDs of the BCC'ed folks. We should fix this by making sure BCC'ed folks get a separate copy of the message that is encrypted only to them.
+
+Relates to #733.
+",0,bcc of encrypted messages should send encrypt separately the user expectation when using bcc is that the people in the to cc lines are unaware of the bcc ed recipients pgp may violate this expectation by including the key ids of the bcc ed folks we should fix this by making sure bcc ed folks get a separate copy of the message that is encrypted only to them relates to ,0
+10953,8229272255.0,IssuesEvent,2018-09-07 08:52:02,dotnet/corefx,https://api.github.com/repos/dotnet/corefx,closed,new X509Certificate2 throws exception while loading certificate in Docker container,area-System.Security,"SDK: Microsoft.AspNetCore.App 2.1.1
+
+The following code throws an exception when it runs in a docker container
+
+` var certificatePath = Path.Combine(env.ContentRootPath, ""TestCertificate.pfx"");
+
+ Console.WriteLine($""Certificate file exists: {File.Exists(certificatePath)}"");
+
+ var certificate =new X509Certificate2(certificatePath, ""Password"");`
+
+The following exception occcurs:
+
+Object was not found
+
+ at Internal.Cryptography.Pal.CertificatePal.FilterPFXStore(Byte[] rawData, SafePasswordHandle password, PfxCertStoreFlags pfxCertStoreFlags)
+ at Internal.Cryptography.Pal.CertificatePal.FromBlobOrFile(Byte[] rawData, String fileName, SafePasswordHandle password, X509KeyStorageFlags keyStorageFlags)
+ at System.Security.Cryptography.X509Certificates.X509Certificate..ctor(String fileName, String password, X509KeyStorageFlags keyStorageFlags)
+ at System.Security.Cryptography.X509Certificates.X509Certificate2..ctor(String fileName, String password)
+
+The certificate file exists in the Docker container. Running the web application outside the Docker container throws no exception.
+
+Example project can be find here: https://github.com/MarcelWouters/DockerLoadCertificate
+
+",True,"new X509Certificate2 throws exception while loading certificate in Docker container - SDK: Microsoft.AspNetCore.App 2.1.1
+
+The following code throws an exception when it runs in a docker container
+
+` var certificatePath = Path.Combine(env.ContentRootPath, ""TestCertificate.pfx"");
+
+ Console.WriteLine($""Certificate file exists: {File.Exists(certificatePath)}"");
+
+ var certificate =new X509Certificate2(certificatePath, ""Password"");`
+
+The following exception occcurs:
+
+Object was not found
+
+ at Internal.Cryptography.Pal.CertificatePal.FilterPFXStore(Byte[] rawData, SafePasswordHandle password, PfxCertStoreFlags pfxCertStoreFlags)
+ at Internal.Cryptography.Pal.CertificatePal.FromBlobOrFile(Byte[] rawData, String fileName, SafePasswordHandle password, X509KeyStorageFlags keyStorageFlags)
+ at System.Security.Cryptography.X509Certificates.X509Certificate..ctor(String fileName, String password, X509KeyStorageFlags keyStorageFlags)
+ at System.Security.Cryptography.X509Certificates.X509Certificate2..ctor(String fileName, String password)
+
+The certificate file exists in the Docker container. Running the web application outside the Docker container throws no exception.
+
+Example project can be find here: https://github.com/MarcelWouters/DockerLoadCertificate
+
+",0,new throws exception while loading certificate in docker container sdk microsoft aspnetcore app the following code throws an exception when it runs in a docker container var certificatepath path combine env contentrootpath testcertificate pfx console writeline certificate file exists file exists certificatepath var certificate new certificatepath password the following exception occcurs object was not found at internal cryptography pal certificatepal filterpfxstore byte rawdata safepasswordhandle password pfxcertstoreflags pfxcertstoreflags at internal cryptography pal certificatepal frombloborfile byte rawdata string filename safepasswordhandle password keystorageflags at system security cryptography ctor string filename string password keystorageflags at system security cryptography ctor string filename string password the certificate file exists in the docker container running the web application outside the docker container throws no exception example project can be find here ,0
+751,13522719638.0,IssuesEvent,2020-09-15 08:57:07,elifesciences/schematron-gitbook,https://api.github.com/repos/elifesciences/schematron-gitbook,closed,Create Deceased status page,people,"- [x] Create page
+- [x] Add description of what is being covered
+- [x] Add description of checks required on this content, compiled from current protocol documents
+- [x] Add examples covering correct and incorrect scenarios, as required to support checks
+- [x] Populate with schematron messages
+- [x] Provide explanations for how to fix schematron errors and warnings
+- [x] Review house style document to ensure all requirements are covered by GitBook and schematron
+- [x] Review schematron messages related to this topic to ensure they are user friendly
+- [x] Becky approved page
+- [x] Melissa approved page
+- [x] Fred approved page
+- [x] James approved page
+- [x] Naushin approved page
+- [x] Share page with Exeter and invite feedback
+- [x] Revise with Exeter's suggestions
+
+
+### Definition of done
+- [x] Page completed
+- [x] Team reviewed and approved page?
+- [ ] Assign Fred to update the Schematron
+",1.0,"Create Deceased status page - - [x] Create page
+- [x] Add description of what is being covered
+- [x] Add description of checks required on this content, compiled from current protocol documents
+- [x] Add examples covering correct and incorrect scenarios, as required to support checks
+- [x] Populate with schematron messages
+- [x] Provide explanations for how to fix schematron errors and warnings
+- [x] Review house style document to ensure all requirements are covered by GitBook and schematron
+- [x] Review schematron messages related to this topic to ensure they are user friendly
+- [x] Becky approved page
+- [x] Melissa approved page
+- [x] Fred approved page
+- [x] James approved page
+- [x] Naushin approved page
+- [x] Share page with Exeter and invite feedback
+- [x] Revise with Exeter's suggestions
+
+
+### Definition of done
+- [x] Page completed
+- [x] Team reviewed and approved page?
+- [ ] Assign Fred to update the Schematron
+",1,create deceased status page create page add description of what is being covered add description of checks required on this content compiled from current protocol documents add examples covering correct and incorrect scenarios as required to support checks populate with schematron messages provide explanations for how to fix schematron errors and warnings review house style document to ensure all requirements are covered by gitbook and schematron review schematron messages related to this topic to ensure they are user friendly becky approved page melissa approved page fred approved page james approved page naushin approved page share page with exeter and invite feedback revise with exeter s suggestions definition of done page completed team reviewed and approved page assign fred to update the schematron ,1
+382695,26510697555.0,IssuesEvent,2023-01-18 16:52:59,facebook/docusaurus,https://api.github.com/repos/facebook/docusaurus,closed,404 error page after click `remark plugins` and `rehype plugins`,documentation,"### Have you read the Contributing Guidelines on issues?
+
+- [X] I have read the [Contributing Guidelines on issues](https://github.com/facebook/docusaurus/blob/main/CONTRIBUTING.md#reporting-new-issues).
+
+### Prerequisites
+
+- [X] I'm using the latest version of Docusaurus.
+- [ ] I have tried the `npm run clear` or `yarn clear` command.
+- [ ] I have tried `rm -rf node_modules yarn.lock package-lock.json` and re-installing packages.
+- [ ] I have tried creating a repro with https://new.docusaurus.io.
+- [ ] I have read the console error message carefully (if applicable).
+
+### Description
+
+On the `docs/markdown-feature/plugins` (MDX Plugins) page, both `remark plugins` and `rehype plugins` links lead to 404 error page.
+
+
+
+### Reproducible demo
+
+_No response_
+
+### Steps to reproduce
+
+- go to docs/markdown-feature/plugins
+- click remark/rehype plugins links
+
+### Expected behavior
+
+it should lead to the correct page about remark and rehype plugin pages
+
+### Actual behavior
+
+both links leads to 404 error page
+
+### Your environment
+
+- Public source code:
+- Public site URL:
+- Docusaurus version used:
+- Environment name and version (e.g. Chrome 89, Node.js 16.4):
+- Operating system and version (e.g. Ubuntu 20.04.2 LTS):
+
+
+### Self-service
+
+- [ ] I'd be willing to fix this bug myself.",1.0,"404 error page after click `remark plugins` and `rehype plugins` - ### Have you read the Contributing Guidelines on issues?
+
+- [X] I have read the [Contributing Guidelines on issues](https://github.com/facebook/docusaurus/blob/main/CONTRIBUTING.md#reporting-new-issues).
+
+### Prerequisites
+
+- [X] I'm using the latest version of Docusaurus.
+- [ ] I have tried the `npm run clear` or `yarn clear` command.
+- [ ] I have tried `rm -rf node_modules yarn.lock package-lock.json` and re-installing packages.
+- [ ] I have tried creating a repro with https://new.docusaurus.io.
+- [ ] I have read the console error message carefully (if applicable).
+
+### Description
+
+On the `docs/markdown-feature/plugins` (MDX Plugins) page, both `remark plugins` and `rehype plugins` links lead to 404 error page.
+
+
+
+### Reproducible demo
+
+_No response_
+
+### Steps to reproduce
+
+- go to docs/markdown-feature/plugins
+- click remark/rehype plugins links
+
+### Expected behavior
+
+it should lead to the correct page about remark and rehype plugin pages
+
+### Actual behavior
+
+both links leads to 404 error page
+
+### Your environment
+
+- Public source code:
+- Public site URL:
+- Docusaurus version used:
+- Environment name and version (e.g. Chrome 89, Node.js 16.4):
+- Operating system and version (e.g. Ubuntu 20.04.2 LTS):
+
+
+### Self-service
+
+- [ ] I'd be willing to fix this bug myself.",0, error page after click remark plugins and rehype plugins have you read the contributing guidelines on issues i have read the prerequisites i m using the latest version of docusaurus i have tried the npm run clear or yarn clear command i have tried rm rf node modules yarn lock package lock json and re installing packages i have tried creating a repro with i have read the console error message carefully if applicable description on the docs markdown feature plugins mdx plugins page both remark plugins and rehype plugins links lead to error page reproducible demo no response steps to reproduce go to docs markdown feature plugins click remark rehype plugins links expected behavior it should lead to the correct page about remark and rehype plugin pages actual behavior both links leads to error page your environment public source code public site url docusaurus version used environment name and version e g chrome node js operating system and version e g ubuntu lts self service i d be willing to fix this bug myself ,0
+822,15287979174.0,IssuesEvent,2021-02-23 16:21:15,openstates/issues,https://api.github.com/repos/openstates/issues,closed,UT Legislator Spot Check Issues,component:people-data type:bug,"State: Utah
+
+Short Description: Many senators missing additional phone numbers and missing address information on general assembly site, discrepancy in phone numbers listed.
+
+**Missing or Incorrect legislators:**
+n/a
+
+**Data Issues:**
+General Assembly doesn't list a full address but OS does (seems systemic, I just can't verify if the OS addresses are accurate because I don't know where the software is pulling them from, only an issue for senators) :
+-Luz Escamilla: https://openstates.org/person/luz-escamilla-3D6Ps1Nc79GQy5HRt1EQ88/ , https://senate.utah.gov/luz-escamilla
+-Jani Iwamoto: https://senate.utah.gov/jani-iwamoto
+
+Only lists mobile number, not home/work number as well (seems systemic, only an issue for senators):
+-Todd Weiler: https://senate.utah.gov/todd-weiler , https://openstates.org/person/todd-d-weiler-3KDxU9BmAikCSSK8DxNaWn/
+
+Phone number listed in open states is different than general assembly pages:
+-Joel Ferry: https://house.utah.gov/rep/FERRYJ/ , https://openstates.org/person/joel-ferry-2zqWdtTtjGbLb0oMkr4ntM/
+-Brad R. Wilson: https://house.utah.gov/rep/WILSOBR/ , https://openstates.org/person/brad-r-wilson-3T8x5FjzRXtbR0scXJYRJ2/
+
+Wrong address listed on Open States:
+-Joel K. Briscoe: https://house.utah.gov/rep/BRISCJK/ , https://openstates.org/person/joel-k-briscoe-36waS8WrKvQoJbQ0znuZVa/
+
+**Additional Data:**
+n/a
+
+",1.0,"UT Legislator Spot Check Issues - State: Utah
+
+Short Description: Many senators missing additional phone numbers and missing address information on general assembly site, discrepancy in phone numbers listed.
+
+**Missing or Incorrect legislators:**
+n/a
+
+**Data Issues:**
+General Assembly doesn't list a full address but OS does (seems systemic, I just can't verify if the OS addresses are accurate because I don't know where the software is pulling them from, only an issue for senators) :
+-Luz Escamilla: https://openstates.org/person/luz-escamilla-3D6Ps1Nc79GQy5HRt1EQ88/ , https://senate.utah.gov/luz-escamilla
+-Jani Iwamoto: https://senate.utah.gov/jani-iwamoto
+
+Only lists mobile number, not home/work number as well (seems systemic, only an issue for senators):
+-Todd Weiler: https://senate.utah.gov/todd-weiler , https://openstates.org/person/todd-d-weiler-3KDxU9BmAikCSSK8DxNaWn/
+
+Phone number listed in open states is different than general assembly pages:
+-Joel Ferry: https://house.utah.gov/rep/FERRYJ/ , https://openstates.org/person/joel-ferry-2zqWdtTtjGbLb0oMkr4ntM/
+-Brad R. Wilson: https://house.utah.gov/rep/WILSOBR/ , https://openstates.org/person/brad-r-wilson-3T8x5FjzRXtbR0scXJYRJ2/
+
+Wrong address listed on Open States:
+-Joel K. Briscoe: https://house.utah.gov/rep/BRISCJK/ , https://openstates.org/person/joel-k-briscoe-36waS8WrKvQoJbQ0znuZVa/
+
+**Additional Data:**
+n/a
+
+",1,ut legislator spot check issues state utah short description many senators missing additional phone numbers and missing address information on general assembly site discrepancy in phone numbers listed missing or incorrect legislators n a data issues general assembly doesn t list a full address but os does seems systemic i just can t verify if the os addresses are accurate because i don t know where the software is pulling them from only an issue for senators luz escamilla jani iwamoto only lists mobile number not home work number as well seems systemic only an issue for senators todd weiler phone number listed in open states is different than general assembly pages joel ferry brad r wilson wrong address listed on open states joel k briscoe additional data n a ,1
+743,13436054128.0,IssuesEvent,2020-09-07 13:50:00,samvera-labs/samvera-connect,https://api.github.com/repos/samvera-labs/samvera-connect,opened,Ye Olde Tech Call,DevOps Developers Lightning talk Managers (general) Managers (repo) Metadata Newcomers Sysadmins UI/UX people,"Jeremy Friesen jfriesen@nd.edu
+
+A quick introduction to the Samvera Tech call to address what is it, who is it for, and how can I engage.
+
+",1.0,"Ye Olde Tech Call - Jeremy Friesen jfriesen@nd.edu
+
+A quick introduction to the Samvera Tech call to address what is it, who is it for, and how can I engage.
+
+",1,ye olde tech call jeremy friesen jfriesen nd edu a quick introduction to the samvera tech call to address what is it who is it for and how can i engage ,1
+450,8251856058.0,IssuesEvent,2018-09-12 09:05:36,newtheatre/history-project,https://api.github.com/repos/newtheatre/history-project,closed,Graduated person appearing as student,bug people report-tool,"End user submitted issue from page: [/people/ramsay_gilderdale/](https://history.newtheatre.org.uk/people/ramsay_gilderdale/)
+---
+He appears as a student when he quite clearly isn't
+*NP*
+",1.0,"Graduated person appearing as student - End user submitted issue from page: [/people/ramsay_gilderdale/](https://history.newtheatre.org.uk/people/ramsay_gilderdale/)
+---
+He appears as a student when he quite clearly isn't
+*NP*
+",1,graduated person appearing as student end user submitted issue from page he appears as a student when he quite clearly isn t np ,1
+233174,17856382145.0,IssuesEvent,2021-09-05 05:13:29,Justin62628/Squirrel-RIFE,https://api.github.com/repos/Justin62628/Squirrel-RIFE,closed,社区版教程提醒,documentation help wanted UI,"考虑到上steam后用户很少看根目录,因此将教程放在根目录等用户自行阅读并不是个很好的选项了
+建议在初次启动应用时弹窗提醒阅读教程
+在主界面显示教程的链接,点击即可阅读(只需本地word/PDF链接均可),用户添加视频后再自行隐去
+
+",1.0,"社区版教程提醒 - 考虑到上steam后用户很少看根目录,因此将教程放在根目录等用户自行阅读并不是个很好的选项了
+建议在初次启动应用时弹窗提醒阅读教程
+在主界面显示教程的链接,点击即可阅读(只需本地word/PDF链接均可),用户添加视频后再自行隐去
+
+",0,社区版教程提醒 考虑到上steam后用户很少看根目录,因此将教程放在根目录等用户自行阅读并不是个很好的选项了 建议在初次启动应用时弹窗提醒阅读教程 在主界面显示教程的链接,点击即可阅读(只需本地word pdf链接均可),用户添加视频后再自行隐去 ,0
+997,23748117525.0,IssuesEvent,2022-08-31 17:52:38,microsoft/fluentui,https://api.github.com/repos/microsoft/fluentui,closed,People picker: contrast ratio is less than 3:1 for the close button when focus on people picker item,Type: Bug :bug: Component: PeoplePicker Status: In PR,"### Library
+
+React / v8 (@fluentui/react)
+
+### System Info
+
+```shell
+Win11 Pro 10.0.22000 Build 22000
+URL: https://developer.microsoft.com/en-us/fluentui#/controls/web/peoplepicker(Fluent UI React 8.93.0)
+Microsoft Edge Version 104.0.1293.70 (Official build) (64-bit)
+```
+
+
+### Are you reporting Accessibility issue?
+
+yes
+
+### Reproduction
+
+https://developer.microsoft.com/en-us/fluentui#/controls/web/peoplepicker
+
+### Bug Description
+
+## Actual Behavior
+Color contrast is less than 3:1 for the close button when focus on people picker item
+
+
+## Expected Behavior
+The close button when focus on people picker item has a contrast ratio of at least 3:1 against adjacent color(s).
+
+MAS Reference:
+[MAS 1.4.11 - Non-text Contrast](https://microsoft.sharepoint.com/:w:/r/teams/msenable/_layouts/15/Doc.aspx?sourcedoc=%7b6CA3BF14-2635-434A-832C-28DD9A18B9FC%7d&file=MAS%201.4.11%20%E2%80%93%20Non-text%20Contrast.docx&action=default&mobileredirect=true)
+
+
+
+
+### Logs
+
+_No response_
+
+### Requested priority
+
+Normal
+
+### Products/sites affected
+
+https://df.compliance.microsoft.com/auditlogsearch
+
+### Are you willing to submit a PR to fix?
+
+no
+
+### Validations
+
+- [X] Check that there isn't already an issue that reports the same bug to avoid creating a duplicate.
+- [X] The provided reproduction is a minimal reproducible example of the bug.",1.0,"People picker: contrast ratio is less than 3:1 for the close button when focus on people picker item - ### Library
+
+React / v8 (@fluentui/react)
+
+### System Info
+
+```shell
+Win11 Pro 10.0.22000 Build 22000
+URL: https://developer.microsoft.com/en-us/fluentui#/controls/web/peoplepicker(Fluent UI React 8.93.0)
+Microsoft Edge Version 104.0.1293.70 (Official build) (64-bit)
+```
+
+
+### Are you reporting Accessibility issue?
+
+yes
+
+### Reproduction
+
+https://developer.microsoft.com/en-us/fluentui#/controls/web/peoplepicker
+
+### Bug Description
+
+## Actual Behavior
+Color contrast is less than 3:1 for the close button when focus on people picker item
+
+
+## Expected Behavior
+The close button when focus on people picker item has a contrast ratio of at least 3:1 against adjacent color(s).
+
+MAS Reference:
+[MAS 1.4.11 - Non-text Contrast](https://microsoft.sharepoint.com/:w:/r/teams/msenable/_layouts/15/Doc.aspx?sourcedoc=%7b6CA3BF14-2635-434A-832C-28DD9A18B9FC%7d&file=MAS%201.4.11%20%E2%80%93%20Non-text%20Contrast.docx&action=default&mobileredirect=true)
+
+
+
+
+### Logs
+
+_No response_
+
+### Requested priority
+
+Normal
+
+### Products/sites affected
+
+https://df.compliance.microsoft.com/auditlogsearch
+
+### Are you willing to submit a PR to fix?
+
+no
+
+### Validations
+
+- [X] Check that there isn't already an issue that reports the same bug to avoid creating a duplicate.
+- [X] The provided reproduction is a minimal reproducible example of the bug.",1,people picker contrast ratio is less than for the close button when focus on people picker item library react fluentui react system info shell pro build url ui react microsoft edge version official build bit are you reporting accessibility issue yes reproduction bug description actual behavior color contrast is less than for the close button when focus on people picker item expected behavior the close button when focus on people picker item has a contrast ratio of at least against adjacent color s mas reference logs no response requested priority normal products sites affected are you willing to submit a pr to fix no validations check that there isn t already an issue that reports the same bug to avoid creating a duplicate the provided reproduction is a minimal reproducible example of the bug ,1
+704081,24184929835.0,IssuesEvent,2022-09-23 12:30:16,metabase/metabase,https://api.github.com/repos/metabase/metabase,closed,Loading spinners in filters don't appear anymore,Type:Bug Priority:P2 Querying/Parameters & Variables .Frontend .Reproduced .Regression,"**Describe the bug**
+In pre-44, we had a loading spinner to indicate that the filter values were being loaded
+
+
+In 44, we don't show anything, and this is concerning for users with huge app db's where the queries take a long time
+
+
+**Logs**
+NA
+
+**To Reproduce**
+1) create a question in 43.4 (just select people table), add it to a dashboard, then add a location filter to the dashboard, connect it to the question and save it
+2) throttle the browser to slow 3g, see the spinner
+3) upgrade the instance to 44
+4) do the same and see that the spinner dissapeared
+
+**Expected behavior**
+bring back the spinners when things take time!
+
+**Screenshots**
+Above
+
+**Information about your Metabase Installation:**
+- Your browser and the version: Brave latest
+- Your operating system: Pop 22.04
+- Your databases: Postgres 14
+- Metabase version: 44
+- Metabase hosting environment: Docker
+- Metabase internal database: Postgres 14
+
+**Severity**
+P2 (since it impacts our bigger customers)
+
+**Additional context**
+NA
+",1.0,"Loading spinners in filters don't appear anymore - **Describe the bug**
+In pre-44, we had a loading spinner to indicate that the filter values were being loaded
+
+
+In 44, we don't show anything, and this is concerning for users with huge app db's where the queries take a long time
+
+
+**Logs**
+NA
+
+**To Reproduce**
+1) create a question in 43.4 (just select people table), add it to a dashboard, then add a location filter to the dashboard, connect it to the question and save it
+2) throttle the browser to slow 3g, see the spinner
+3) upgrade the instance to 44
+4) do the same and see that the spinner dissapeared
+
+**Expected behavior**
+bring back the spinners when things take time!
+
+**Screenshots**
+Above
+
+**Information about your Metabase Installation:**
+- Your browser and the version: Brave latest
+- Your operating system: Pop 22.04
+- Your databases: Postgres 14
+- Metabase version: 44
+- Metabase hosting environment: Docker
+- Metabase internal database: Postgres 14
+
+**Severity**
+P2 (since it impacts our bigger customers)
+
+**Additional context**
+NA
+",0,loading spinners in filters don t appear anymore describe the bug in pre we had a loading spinner to indicate that the filter values were being loaded in we don t show anything and this is concerning for users with huge app db s where the queries take a long time logs na to reproduce create a question in just select people table add it to a dashboard then add a location filter to the dashboard connect it to the question and save it throttle the browser to slow see the spinner upgrade the instance to do the same and see that the spinner dissapeared expected behavior bring back the spinners when things take time screenshots above information about your metabase installation your browser and the version brave latest your operating system pop your databases postgres metabase version metabase hosting environment docker metabase internal database postgres severity since it impacts our bigger customers additional context na ,0
+509,8745013487.0,IssuesEvent,2018-12-13 00:37:24,cloudamatic/mu,https://api.github.com/repos/cloudamatic/mu,closed,the latest Windows bootstrap idiosyncrasies,bug in progress workaround needed for other peoples' bugs,"- Initial knife bootstrap appears to hang forever. If you RDP into an in-progress node exhibiting this behavior, you'll see a ruby.exe process associated with a Chef run, sitting idle. If you kill it, subsequent bootstrap calls seem to proceed normally. It's unclear if this process is one that we're initiating as part of bootstrapping, or coming from somewhere else (inherited from the AMI, for example).
+
+- Windows nodes still routinely get wedged in the middle of a reboot, sometimes badly enough for AutoScale to terminate them. This behavior is nondeterministic. All of our Mu-side ""connect and do something"" loops for Windows have some logic to force reboots along when they suspect this condition. It can, however, occur when MommaCat isn't actively involved, e.g. when userdata reboots to apply updates, and so there's nobody to jump in and save the node from itself. We may have to have a general-purpose `isWindowsHosed?` method that prods Mu-tagged Windows machines with both RDP and WinRM, and if nothing answers for an extended period of time gives them a kick in the ribs.
+
+- Current builds eventually reach a state where `--no-fips` is required to run normally:
+```
+Jan 11 13:34:56 - groomers/chef - chef-client.bat : C:/opscode/chef/embedded/lib/ruby/gems/2.3.0/gems/chef-config-12.21.14/lib/chef-config/config.rb:1084:in `fips_mode=': Turning on FIPS mode failed: fingerprint does not match (OpenSSL::OpenSSLError)
+ + CategoryInfo : NotSpecified: (C:/opscode/chef...::OpenSSLError):String) [], RemoteException
+ + FullyQualifiedErrorId : NativeCommandError
+```
+
+Cursory research suggests that this may be an OpenSSL library problem.
+
+",1.0,"the latest Windows bootstrap idiosyncrasies - - Initial knife bootstrap appears to hang forever. If you RDP into an in-progress node exhibiting this behavior, you'll see a ruby.exe process associated with a Chef run, sitting idle. If you kill it, subsequent bootstrap calls seem to proceed normally. It's unclear if this process is one that we're initiating as part of bootstrapping, or coming from somewhere else (inherited from the AMI, for example).
+
+- Windows nodes still routinely get wedged in the middle of a reboot, sometimes badly enough for AutoScale to terminate them. This behavior is nondeterministic. All of our Mu-side ""connect and do something"" loops for Windows have some logic to force reboots along when they suspect this condition. It can, however, occur when MommaCat isn't actively involved, e.g. when userdata reboots to apply updates, and so there's nobody to jump in and save the node from itself. We may have to have a general-purpose `isWindowsHosed?` method that prods Mu-tagged Windows machines with both RDP and WinRM, and if nothing answers for an extended period of time gives them a kick in the ribs.
+
+- Current builds eventually reach a state where `--no-fips` is required to run normally:
+```
+Jan 11 13:34:56 - groomers/chef - chef-client.bat : C:/opscode/chef/embedded/lib/ruby/gems/2.3.0/gems/chef-config-12.21.14/lib/chef-config/config.rb:1084:in `fips_mode=': Turning on FIPS mode failed: fingerprint does not match (OpenSSL::OpenSSLError)
+ + CategoryInfo : NotSpecified: (C:/opscode/chef...::OpenSSLError):String) [], RemoteException
+ + FullyQualifiedErrorId : NativeCommandError
+```
+
+Cursory research suggests that this may be an OpenSSL library problem.
+
+",1,the latest windows bootstrap idiosyncrasies initial knife bootstrap appears to hang forever if you rdp into an in progress node exhibiting this behavior you ll see a ruby exe process associated with a chef run sitting idle if you kill it subsequent bootstrap calls seem to proceed normally it s unclear if this process is one that we re initiating as part of bootstrapping or coming from somewhere else inherited from the ami for example windows nodes still routinely get wedged in the middle of a reboot sometimes badly enough for autoscale to terminate them this behavior is nondeterministic all of our mu side connect and do something loops for windows have some logic to force reboots along when they suspect this condition it can however occur when mommacat isn t actively involved e g when userdata reboots to apply updates and so there s nobody to jump in and save the node from itself we may have to have a general purpose iswindowshosed method that prods mu tagged windows machines with both rdp and winrm and if nothing answers for an extended period of time gives them a kick in the ribs current builds eventually reach a state where no fips is required to run normally jan groomers chef chef client bat c opscode chef embedded lib ruby gems gems chef config lib chef config config rb in fips mode turning on fips mode failed fingerprint does not match openssl opensslerror categoryinfo notspecified c opscode chef opensslerror string remoteexception fullyqualifiederrorid nativecommanderror cursory research suggests that this may be an openssl library problem ,1
+271,5315392461.0,IssuesEvent,2017-02-13 17:12:18,PopulateTools/gobierto,https://api.github.com/repos/PopulateTools/gobierto,closed,Mark days with events in the calendar,enhancement gobierto-people,The days that have events should be higlighted and differentiated from the days without events.,1.0,Mark days with events in the calendar - The days that have events should be higlighted and differentiated from the days without events.,1,mark days with events in the calendar the days that have events should be higlighted and differentiated from the days without events ,1
+581,10326646780.0,IssuesEvent,2019-09-02 03:11:19,DimensionDev/Maskbook,https://api.github.com/repos/DimensionDev/Maskbook,opened,[Roadmap] auto share post to friends part 1,Component: Compose box Component: Database Component: People Type: Feature Type: UI,"1. [ ] Recognize who are friends
+2. [ ] Impl virtual friend user group ( #34 )
+3. [ ] Add friends to user group (need 1 and 2)
+4. [ ] Add all people to selection by default (UI)
+5. [ ] Allow user group as selection target (UI, need 2)
+6. [ ] Auto share old posts to new user in the group (DB, need 2, 3)
+",1.0,"[Roadmap] auto share post to friends part 1 - 1. [ ] Recognize who are friends
+2. [ ] Impl virtual friend user group ( #34 )
+3. [ ] Add friends to user group (need 1 and 2)
+4. [ ] Add all people to selection by default (UI)
+5. [ ] Allow user group as selection target (UI, need 2)
+6. [ ] Auto share old posts to new user in the group (DB, need 2, 3)
+",1, auto share post to friends part recognize who are friends impl virtual friend user group add friends to user group need and add all people to selection by default ui allow user group as selection target ui need auto share old posts to new user in the group db need ,1
+8,2649568397.0,IssuesEvent,2015-03-15 01:48:38,fskreuz/Tree-Trail,https://api.github.com/repos/fskreuz/Tree-Trail,closed,No more Register link,QA people! Test these!,"When clicking ""Login"" button, there is a link at the bottom right of the form which says ""Register"".
+
+This is not needed anymore. Only the super admin user can add new admin accounts when he is logged into the system.
+
+No registration of users is required. Only creation of users by the super admin user once he is logged in.
+
+Assigned to: fskreuz and chuakarlo",1.0,"No more Register link - When clicking ""Login"" button, there is a link at the bottom right of the form which says ""Register"".
+
+This is not needed anymore. Only the super admin user can add new admin accounts when he is logged into the system.
+
+No registration of users is required. Only creation of users by the super admin user once he is logged in.
+
+Assigned to: fskreuz and chuakarlo",1,no more register link when clicking login button there is a link at the bottom right of the form which says register this is not needed anymore only the super admin user can add new admin accounts when he is logged into the system no registration of users is required only creation of users by the super admin user once he is logged in assigned to fskreuz and chuakarlo,1
+309,5716473803.0,IssuesEvent,2017-04-19 15:12:49,Princeton-CDH/mapping-expatriate-paris,https://api.github.com/repos/Princeton-CDH/mapping-expatriate-paris,reopened,Denis//Denis-Graterolle cards issues,has unknown dates not in personography two people merged in one card folder,"issue one:
+on the Denis card file there are three different people:
+
+1) Madame Maurice (s1 - s4)
+2) Pierre (s5- s6)
+3) a carte postale (unclear to whom it belongs :: s7-s8)
+
+issue two:
+also, the Denis-Graterolle card file has a Madame Maurice Denis-Graterolle
+the main Denis on the Denis card file is also a Madame Maurice
+are they the same person :: & should they be merged in the personography?
+",1.0,"Denis//Denis-Graterolle cards issues - issue one:
+on the Denis card file there are three different people:
+
+1) Madame Maurice (s1 - s4)
+2) Pierre (s5- s6)
+3) a carte postale (unclear to whom it belongs :: s7-s8)
+
+issue two:
+also, the Denis-Graterolle card file has a Madame Maurice Denis-Graterolle
+the main Denis on the Denis card file is also a Madame Maurice
+are they the same person :: & should they be merged in the personography?
+",1,denis denis graterolle cards issues issue one on the denis card file there are three different people madame maurice pierre a carte postale unclear to whom it belongs issue two also the denis graterolle card file has a madame maurice denis graterolle the main denis on the denis card file is also a madame maurice are they the same person should they be merged in the personography ,1
+32888,13941605853.0,IssuesEvent,2020-10-22 19:40:29,Azure/azure-sdk-for-js,https://api.github.com/repos/Azure/azure-sdk-for-js,closed,[Service Bus] Swap the order of messages & enqueuedTime in scheduled_messages method,Client Service Bus good first issue help wanted,"The `scheduled_messages()` method in the Service Bus package takes the time first and the messages later. It is more intuitive to swap this order because the message(s) is the main entity here. This also makes us consistent with the Service Bus packages in other languages
+
+PRs are welcome to add this feature.
+Code Pointers:
+- [Set up environment](https://github.com/Azure/azure-sdk-for-js/blob/master/CONTRIBUTING.md#setting-up-your-environment) and run `rush build`
+- Open the `sdk/servicebus/service-bus` folder in your editor/IDE, find the `scheduled_messages()` method and swap the order
+
+",1.0,"[Service Bus] Swap the order of messages & enqueuedTime in scheduled_messages method - The `scheduled_messages()` method in the Service Bus package takes the time first and the messages later. It is more intuitive to swap this order because the message(s) is the main entity here. This also makes us consistent with the Service Bus packages in other languages
+
+PRs are welcome to add this feature.
+Code Pointers:
+- [Set up environment](https://github.com/Azure/azure-sdk-for-js/blob/master/CONTRIBUTING.md#setting-up-your-environment) and run `rush build`
+- Open the `sdk/servicebus/service-bus` folder in your editor/IDE, find the `scheduled_messages()` method and swap the order
+
+",0, swap the order of messages enqueuedtime in scheduled messages method the scheduled messages method in the service bus package takes the time first and the messages later it is more intuitive to swap this order because the message s is the main entity here this also makes us consistent with the service bus packages in other languages prs are welcome to add this feature code pointers and run rush build open the sdk servicebus service bus folder in your editor ide find the scheduled messages method and swap the order ,0
+857,16524755992.0,IssuesEvent,2021-05-26 18:32:47,microsoft/botframework-components,https://api.github.com/repos/microsoft/botframework-components,closed,"People skill - if someone simply asks ""Who reports to"" the bot should follow up and ask for a name",Area: Generators: Enterprise People Owner: PBX P1 backlog bug,"### [Github issues](https://github.com/Microsoft/botframework-skills/issues) should be used for bugs and feature requests. Use [Stack Overflow](https://stackoverflow.com/questions/tagged/botframework) for general ""how-to"" questions.
+
+## Version
+4.13
+
+## Describe the bug
+When someone simply asks a question without an entity, it should trigger a new flow to ask for the name to search.
+
+## To Reproduce
+1. Ask ""who reports to""
+
+## Expected behavior
+Trigger a flow to disambiguate who the person is about.
+
+## Screenshots
+If applicable, add screenshots to help explain your problem.
+
+## Additional context
+Add any other context about the problem here.
+",1.0,"People skill - if someone simply asks ""Who reports to"" the bot should follow up and ask for a name - ### [Github issues](https://github.com/Microsoft/botframework-skills/issues) should be used for bugs and feature requests. Use [Stack Overflow](https://stackoverflow.com/questions/tagged/botframework) for general ""how-to"" questions.
+
+## Version
+4.13
+
+## Describe the bug
+When someone simply asks a question without an entity, it should trigger a new flow to ask for the name to search.
+
+## To Reproduce
+1. Ask ""who reports to""
+
+## Expected behavior
+Trigger a flow to disambiguate who the person is about.
+
+## Screenshots
+If applicable, add screenshots to help explain your problem.
+
+## Additional context
+Add any other context about the problem here.
+",1,people skill if someone simply asks who reports to the bot should follow up and ask for a name should be used for bugs and feature requests use for general how to questions version describe the bug when someone simply asks a question without an entity it should trigger a new flow to ask for the name to search to reproduce ask who reports to expected behavior trigger a flow to disambiguate who the person is about screenshots if applicable add screenshots to help explain your problem additional context add any other context about the problem here ,1
+72978,31814935105.0,IssuesEvent,2023-09-13 19:41:48,MicrosoftDocs/azure-docs,https://api.github.com/repos/MicrosoftDocs/azure-docs,closed,Sample csharp code doesn't compile as-is,triaged in-progress assigned-to-author doc-bug Pri2 azure-communication-services/svc data/subsvc,"The sample is currently in doc won't compile. I had to make some changes.
+
+The command below didn't work. I needed to add a `--prerelease` and it seems OK.
+```script
+dotnet add package Azure.Monitor.OpenTelemetry.Exporter
+```
+
+I had to manually add a few `using` statements.
+
+```csharp
+using System.Diagnostics;
+using OpenTelemetry.Resources;
+using OpenTelemetry;
+
+```
+
+---
+#### Document Details
+
+⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
+
+* ID: 087817c3-e26e-2a9b-baf7-db6ad20141a1
+* Version Independent ID: 9409baaf-bbb6-d920-d957-fa8eed5dcd50
+* Content: [Quickstart - Exporting SDK telemetry data to Application Insights - An Azure Communication Services quickstart](https://docs.microsoft.com/en-us/azure/communication-services/quickstarts/telemetry-application-insights?pivots=programming-language-csharp)
+* Content Source: [articles/communication-services/quickstarts/telemetry-application-insights.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/communication-services/quickstarts/telemetry-application-insights.md)
+* Service: **azure-communication-services**
+* Sub-service: **data**
+* GitHub Login: @peiliu
+* Microsoft Alias: **peiliu**",1.0,"Sample csharp code doesn't compile as-is - The sample is currently in doc won't compile. I had to make some changes.
+
+The command below didn't work. I needed to add a `--prerelease` and it seems OK.
+```script
+dotnet add package Azure.Monitor.OpenTelemetry.Exporter
+```
+
+I had to manually add a few `using` statements.
+
+```csharp
+using System.Diagnostics;
+using OpenTelemetry.Resources;
+using OpenTelemetry;
+
+```
+
+---
+#### Document Details
+
+⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
+
+* ID: 087817c3-e26e-2a9b-baf7-db6ad20141a1
+* Version Independent ID: 9409baaf-bbb6-d920-d957-fa8eed5dcd50
+* Content: [Quickstart - Exporting SDK telemetry data to Application Insights - An Azure Communication Services quickstart](https://docs.microsoft.com/en-us/azure/communication-services/quickstarts/telemetry-application-insights?pivots=programming-language-csharp)
+* Content Source: [articles/communication-services/quickstarts/telemetry-application-insights.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/communication-services/quickstarts/telemetry-application-insights.md)
+* Service: **azure-communication-services**
+* Sub-service: **data**
+* GitHub Login: @peiliu
+* Microsoft Alias: **peiliu**",0,sample csharp code doesn t compile as is the sample is currently in doc won t compile i had to make some changes the command below didn t work i needed to add a prerelease and it seems ok script dotnet add package azure monitor opentelemetry exporter i had to manually add a few using statements csharp using system diagnostics using opentelemetry resources using opentelemetry document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service azure communication services sub service data github login peiliu microsoft alias peiliu ,0
+688,12393967766.0,IssuesEvent,2020-05-20 16:12:14,googleapis/elixir-google-api,https://api.github.com/repos/googleapis/elixir-google-api,closed,Synthesis failed for People,api: people autosynth failure priority: p1 type: bug,"Hello! Autosynth couldn't regenerate People. :broken_heart:
+
+Here's the output from running `synth.py`:
+
+```
+ng: failed to remove deps/parse_trans/ebin/parse_trans.app: Permission denied
+warning: failed to remove deps/parse_trans/ebin/parse_trans_mod.beam: Permission denied
+warning: failed to remove deps/parse_trans/ebin/parse_trans_codegen.beam: Permission denied
+warning: failed to remove deps/parse_trans/ebin/ct_expand.beam: Permission denied
+warning: failed to remove deps/parse_trans/ebin/parse_trans.beam: Permission denied
+warning: failed to remove deps/parse_trans/ebin/exprecs.beam: Permission denied
+warning: failed to remove deps/parse_trans/ebin/parse_trans_pp.beam: Permission denied
+warning: failed to remove deps/parse_trans/.rebar3/erlcinfo: Permission denied
+warning: failed to remove deps/parse_trans/hex_metadata.config: Permission denied
+warning: failed to remove deps/parse_trans/README.md: Permission denied
+warning: failed to remove deps/parse_trans/rebar.config: Permission denied
+warning: failed to remove deps/parse_trans/include/codegen.hrl: Permission denied
+warning: failed to remove deps/parse_trans/include/exprecs.hrl: Permission denied
+warning: failed to remove deps/parse_trans/.fetch: Permission denied
+warning: failed to remove deps/parse_trans/.hex: Permission denied
+warning: failed to remove deps/idna/LICENSE: Permission denied
+warning: failed to remove deps/idna/rebar.lock: Permission denied
+warning: failed to remove deps/idna/src/idna.erl: Permission denied
+warning: failed to remove deps/idna/src/idna_logger.hrl: Permission denied
+warning: failed to remove deps/idna/src/idna_ucs.erl: Permission denied
+warning: failed to remove deps/idna/src/punycode.erl: Permission denied
+warning: failed to remove deps/idna/src/idna_table.erl: Permission denied
+warning: failed to remove deps/idna/src/idna_context.erl: Permission denied
+warning: failed to remove deps/idna/src/idna.app.src: Permission denied
+warning: failed to remove deps/idna/src/idna_mapping.erl: Permission denied
+warning: failed to remove deps/idna/src/idna_data.erl: Permission denied
+warning: failed to remove deps/idna/src/idna_bidi.erl: Permission denied
+warning: failed to remove deps/idna/ebin/idna_mapping.beam: Permission denied
+warning: failed to remove deps/idna/ebin/idna_context.beam: Permission denied
+warning: failed to remove deps/idna/ebin/idna_bidi.beam: Permission denied
+warning: failed to remove deps/idna/ebin/punycode.beam: Permission denied
+warning: failed to remove deps/idna/ebin/idna_table.beam: Permission denied
+warning: failed to remove deps/idna/ebin/idna_data.beam: Permission denied
+warning: failed to remove deps/idna/ebin/idna_ucs.beam: Permission denied
+warning: failed to remove deps/idna/ebin/idna.app: Permission denied
+warning: failed to remove deps/idna/ebin/idna.beam: Permission denied
+warning: failed to remove deps/idna/.rebar3/erlcinfo: Permission denied
+warning: failed to remove deps/idna/hex_metadata.config: Permission denied
+warning: failed to remove deps/idna/README.md: Permission denied
+warning: failed to remove deps/idna/rebar.config: Permission denied
+warning: failed to remove deps/idna/.fetch: Permission denied
+warning: failed to remove deps/idna/rebar.config.script: Permission denied
+warning: failed to remove deps/idna/.hex: Permission denied
+warning: failed to remove deps/hackney/MAINTAINERS: Permission denied
+warning: failed to remove deps/hackney/LICENSE: Permission denied
+warning: failed to remove deps/hackney/rebar.lock: Permission denied
+warning: failed to remove deps/hackney/src/hackney_ssl.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_response.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_tcp.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_http.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_cookie.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_url.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_headers.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney.app.src: Permission denied
+warning: failed to remove deps/hackney/src/hackney_pool_handler.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_trace.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_multipart.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_headers_new.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_http_connect.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_util.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_socks5.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_request.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_app.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_internal.hrl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_date.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_manager.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_connect.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_bstr.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_sup.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_local_tcp.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_stream.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_pool.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_metrics.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_methods.hrl: Permission denied
+warning: failed to remove deps/hackney/NOTICE: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_pool.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_trace.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_pool_handler.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_headers.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_url.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_manager.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_metrics.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_stream.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_sup.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_multipart.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_http.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_socks5.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_app.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_http_connect.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_response.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney.app: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_headers_new.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_cookie.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_request.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_util.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_connect.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_date.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_ssl.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_bstr.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_tcp.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_local_tcp.beam: Permission denied
+warning: failed to remove deps/hackney/.rebar3/erlcinfo: Permission denied
+warning: failed to remove deps/hackney/hex_metadata.config: Permission denied
+warning: failed to remove deps/hackney/README.md: Permission denied
+warning: failed to remove deps/hackney/rebar.config: Permission denied
+warning: failed to remove deps/hackney/include/hackney.hrl: Permission denied
+warning: failed to remove deps/hackney/include/hackney_lib.hrl: Permission denied
+warning: failed to remove deps/hackney/.fetch: Permission denied
+warning: failed to remove deps/hackney/.hex: Permission denied
+warning: failed to remove deps/hackney/NEWS.md: Permission denied
+Removing __pycache__/
+Removing specifications/gdd/People-v1.json
+Traceback (most recent call last):
+ File ""/tmpfs/src/github/synthtool/autosynth/synth.py"", line 559, in _inner_main
+ sys.exit(EXIT_CODE_SKIPPED)
+SystemExit: 28
+
+During handling of the above exception, another exception occurred:
+
+Traceback (most recent call last):
+ File ""/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py"", line 193, in _run_module_as_main
+ ""__main__"", mod_spec)
+ File ""/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py"", line 85, in _run_code
+ exec(code, run_globals)
+ File ""/tmpfs/src/github/synthtool/autosynth/synth.py"", line 615, in
+ main()
+ File ""/tmpfs/src/github/synthtool/autosynth/synth.py"", line 476, in main
+ return _inner_main(temp_dir)
+ File ""/tmpfs/src/github/synthtool/autosynth/synth.py"", line 611, in _inner_main
+ executor.check_call([""git"", ""clean"", ""-fdx""], cwd=working_repo_path)
+ File ""/tmpfs/src/github/synthtool/autosynth/executor.py"", line 29, in check_call
+ subprocess.check_call(command, **args)
+ File ""/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py"", line 311, in check_call
+ raise CalledProcessError(retcode, cmd)
+subprocess.CalledProcessError: Command '['git', 'clean', '-fdx']' returned non-zero exit status 1.
+
+```
+
+Google internal developers can see the full log [here](http://sponge/c19bc74d-4d42-4305-94c3-323bf90f3ebc).
+",1.0,"Synthesis failed for People - Hello! Autosynth couldn't regenerate People. :broken_heart:
+
+Here's the output from running `synth.py`:
+
+```
+ng: failed to remove deps/parse_trans/ebin/parse_trans.app: Permission denied
+warning: failed to remove deps/parse_trans/ebin/parse_trans_mod.beam: Permission denied
+warning: failed to remove deps/parse_trans/ebin/parse_trans_codegen.beam: Permission denied
+warning: failed to remove deps/parse_trans/ebin/ct_expand.beam: Permission denied
+warning: failed to remove deps/parse_trans/ebin/parse_trans.beam: Permission denied
+warning: failed to remove deps/parse_trans/ebin/exprecs.beam: Permission denied
+warning: failed to remove deps/parse_trans/ebin/parse_trans_pp.beam: Permission denied
+warning: failed to remove deps/parse_trans/.rebar3/erlcinfo: Permission denied
+warning: failed to remove deps/parse_trans/hex_metadata.config: Permission denied
+warning: failed to remove deps/parse_trans/README.md: Permission denied
+warning: failed to remove deps/parse_trans/rebar.config: Permission denied
+warning: failed to remove deps/parse_trans/include/codegen.hrl: Permission denied
+warning: failed to remove deps/parse_trans/include/exprecs.hrl: Permission denied
+warning: failed to remove deps/parse_trans/.fetch: Permission denied
+warning: failed to remove deps/parse_trans/.hex: Permission denied
+warning: failed to remove deps/idna/LICENSE: Permission denied
+warning: failed to remove deps/idna/rebar.lock: Permission denied
+warning: failed to remove deps/idna/src/idna.erl: Permission denied
+warning: failed to remove deps/idna/src/idna_logger.hrl: Permission denied
+warning: failed to remove deps/idna/src/idna_ucs.erl: Permission denied
+warning: failed to remove deps/idna/src/punycode.erl: Permission denied
+warning: failed to remove deps/idna/src/idna_table.erl: Permission denied
+warning: failed to remove deps/idna/src/idna_context.erl: Permission denied
+warning: failed to remove deps/idna/src/idna.app.src: Permission denied
+warning: failed to remove deps/idna/src/idna_mapping.erl: Permission denied
+warning: failed to remove deps/idna/src/idna_data.erl: Permission denied
+warning: failed to remove deps/idna/src/idna_bidi.erl: Permission denied
+warning: failed to remove deps/idna/ebin/idna_mapping.beam: Permission denied
+warning: failed to remove deps/idna/ebin/idna_context.beam: Permission denied
+warning: failed to remove deps/idna/ebin/idna_bidi.beam: Permission denied
+warning: failed to remove deps/idna/ebin/punycode.beam: Permission denied
+warning: failed to remove deps/idna/ebin/idna_table.beam: Permission denied
+warning: failed to remove deps/idna/ebin/idna_data.beam: Permission denied
+warning: failed to remove deps/idna/ebin/idna_ucs.beam: Permission denied
+warning: failed to remove deps/idna/ebin/idna.app: Permission denied
+warning: failed to remove deps/idna/ebin/idna.beam: Permission denied
+warning: failed to remove deps/idna/.rebar3/erlcinfo: Permission denied
+warning: failed to remove deps/idna/hex_metadata.config: Permission denied
+warning: failed to remove deps/idna/README.md: Permission denied
+warning: failed to remove deps/idna/rebar.config: Permission denied
+warning: failed to remove deps/idna/.fetch: Permission denied
+warning: failed to remove deps/idna/rebar.config.script: Permission denied
+warning: failed to remove deps/idna/.hex: Permission denied
+warning: failed to remove deps/hackney/MAINTAINERS: Permission denied
+warning: failed to remove deps/hackney/LICENSE: Permission denied
+warning: failed to remove deps/hackney/rebar.lock: Permission denied
+warning: failed to remove deps/hackney/src/hackney_ssl.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_response.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_tcp.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_http.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_cookie.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_url.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_headers.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney.app.src: Permission denied
+warning: failed to remove deps/hackney/src/hackney_pool_handler.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_trace.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_multipart.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_headers_new.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_http_connect.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_util.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_socks5.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_request.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_app.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_internal.hrl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_date.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_manager.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_connect.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_bstr.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_sup.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_local_tcp.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_stream.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_pool.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_metrics.erl: Permission denied
+warning: failed to remove deps/hackney/src/hackney_methods.hrl: Permission denied
+warning: failed to remove deps/hackney/NOTICE: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_pool.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_trace.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_pool_handler.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_headers.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_url.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_manager.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_metrics.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_stream.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_sup.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_multipart.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_http.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_socks5.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_app.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_http_connect.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_response.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney.app: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_headers_new.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_cookie.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_request.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_util.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_connect.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_date.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_ssl.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_bstr.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_tcp.beam: Permission denied
+warning: failed to remove deps/hackney/ebin/hackney_local_tcp.beam: Permission denied
+warning: failed to remove deps/hackney/.rebar3/erlcinfo: Permission denied
+warning: failed to remove deps/hackney/hex_metadata.config: Permission denied
+warning: failed to remove deps/hackney/README.md: Permission denied
+warning: failed to remove deps/hackney/rebar.config: Permission denied
+warning: failed to remove deps/hackney/include/hackney.hrl: Permission denied
+warning: failed to remove deps/hackney/include/hackney_lib.hrl: Permission denied
+warning: failed to remove deps/hackney/.fetch: Permission denied
+warning: failed to remove deps/hackney/.hex: Permission denied
+warning: failed to remove deps/hackney/NEWS.md: Permission denied
+Removing __pycache__/
+Removing specifications/gdd/People-v1.json
+Traceback (most recent call last):
+ File ""/tmpfs/src/github/synthtool/autosynth/synth.py"", line 559, in _inner_main
+ sys.exit(EXIT_CODE_SKIPPED)
+SystemExit: 28
+
+During handling of the above exception, another exception occurred:
+
+Traceback (most recent call last):
+ File ""/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py"", line 193, in _run_module_as_main
+ ""__main__"", mod_spec)
+ File ""/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py"", line 85, in _run_code
+ exec(code, run_globals)
+ File ""/tmpfs/src/github/synthtool/autosynth/synth.py"", line 615, in
+ main()
+ File ""/tmpfs/src/github/synthtool/autosynth/synth.py"", line 476, in main
+ return _inner_main(temp_dir)
+ File ""/tmpfs/src/github/synthtool/autosynth/synth.py"", line 611, in _inner_main
+ executor.check_call([""git"", ""clean"", ""-fdx""], cwd=working_repo_path)
+ File ""/tmpfs/src/github/synthtool/autosynth/executor.py"", line 29, in check_call
+ subprocess.check_call(command, **args)
+ File ""/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py"", line 311, in check_call
+ raise CalledProcessError(retcode, cmd)
+subprocess.CalledProcessError: Command '['git', 'clean', '-fdx']' returned non-zero exit status 1.
+
+```
+
+Google internal developers can see the full log [here](http://sponge/c19bc74d-4d42-4305-94c3-323bf90f3ebc).
+",1,synthesis failed for people hello autosynth couldn t regenerate people broken heart here s the output from running synth py ng failed to remove deps parse trans ebin parse trans app permission denied warning failed to remove deps parse trans ebin parse trans mod beam permission denied warning failed to remove deps parse trans ebin parse trans codegen beam permission denied warning failed to remove deps parse trans ebin ct expand beam permission denied warning failed to remove deps parse trans ebin parse trans beam permission denied warning failed to remove deps parse trans ebin exprecs beam permission denied warning failed to remove deps parse trans ebin parse trans pp beam permission denied warning failed to remove deps parse trans erlcinfo permission denied warning failed to remove deps parse trans hex metadata config permission denied warning failed to remove deps parse trans readme md permission denied warning failed to remove deps parse trans rebar config permission denied warning failed to remove deps parse trans include codegen hrl permission denied warning failed to remove deps parse trans include exprecs hrl permission denied warning failed to remove deps parse trans fetch permission denied warning failed to remove deps parse trans hex permission denied warning failed to remove deps idna license permission denied warning failed to remove deps idna rebar lock permission denied warning failed to remove deps idna src idna erl permission denied warning failed to remove deps idna src idna logger hrl permission denied warning failed to remove deps idna src idna ucs erl permission denied warning failed to remove deps idna src punycode erl permission denied warning failed to remove deps idna src idna table erl permission denied warning failed to remove deps idna src idna context erl permission denied warning failed to remove deps idna src idna app src permission denied warning failed to remove deps idna src idna mapping erl permission denied warning failed to remove deps idna src idna data erl permission denied warning failed to remove deps idna src idna bidi erl permission denied warning failed to remove deps idna ebin idna mapping beam permission denied warning failed to remove deps idna ebin idna context beam permission denied warning failed to remove deps idna ebin idna bidi beam permission denied warning failed to remove deps idna ebin punycode beam permission denied warning failed to remove deps idna ebin idna table beam permission denied warning failed to remove deps idna ebin idna data beam permission denied warning failed to remove deps idna ebin idna ucs beam permission denied warning failed to remove deps idna ebin idna app permission denied warning failed to remove deps idna ebin idna beam permission denied warning failed to remove deps idna erlcinfo permission denied warning failed to remove deps idna hex metadata config permission denied warning failed to remove deps idna readme md permission denied warning failed to remove deps idna rebar config permission denied warning failed to remove deps idna fetch permission denied warning failed to remove deps idna rebar config script permission denied warning failed to remove deps idna hex permission denied warning failed to remove deps hackney maintainers permission denied warning failed to remove deps hackney license permission denied warning failed to remove deps hackney rebar lock permission denied warning failed to remove deps hackney src hackney ssl erl permission denied warning failed to remove deps hackney src hackney response erl permission denied warning failed to remove deps hackney src hackney tcp erl permission denied warning failed to remove deps hackney src hackney http erl permission denied warning failed to remove deps hackney src hackney cookie erl permission denied warning failed to remove deps hackney src hackney url erl permission denied warning failed to remove deps hackney src hackney headers erl permission denied warning failed to remove deps hackney src hackney app src permission denied warning failed to remove deps hackney src hackney pool handler erl permission denied warning failed to remove deps hackney src hackney trace erl permission denied warning failed to remove deps hackney src hackney multipart erl permission denied warning failed to remove deps hackney src hackney headers new erl permission denied warning failed to remove deps hackney src hackney http connect erl permission denied warning failed to remove deps hackney src hackney util erl permission denied warning failed to remove deps hackney src hackney erl permission denied warning failed to remove deps hackney src hackney request erl permission denied warning failed to remove deps hackney src hackney app erl permission denied warning failed to remove deps hackney src hackney internal hrl permission denied warning failed to remove deps hackney src hackney date erl permission denied warning failed to remove deps hackney src hackney manager erl permission denied warning failed to remove deps hackney src hackney connect erl permission denied warning failed to remove deps hackney src hackney bstr erl permission denied warning failed to remove deps hackney src hackney sup erl permission denied warning failed to remove deps hackney src hackney erl permission denied warning failed to remove deps hackney src hackney local tcp erl permission denied warning failed to remove deps hackney src hackney stream erl permission denied warning failed to remove deps hackney src hackney pool erl permission denied warning failed to remove deps hackney src hackney metrics erl permission denied warning failed to remove deps hackney src hackney methods hrl permission denied warning failed to remove deps hackney notice permission denied warning failed to remove deps hackney ebin hackney pool beam permission denied warning failed to remove deps hackney ebin hackney trace beam permission denied warning failed to remove deps hackney ebin hackney pool handler beam permission denied warning failed to remove deps hackney ebin hackney beam permission denied warning failed to remove deps hackney ebin hackney headers beam permission denied warning failed to remove deps hackney ebin hackney url beam permission denied warning failed to remove deps hackney ebin hackney manager beam permission denied warning failed to remove deps hackney ebin hackney metrics beam permission denied warning failed to remove deps hackney ebin hackney stream beam permission denied warning failed to remove deps hackney ebin hackney sup beam permission denied warning failed to remove deps hackney ebin hackney multipart beam permission denied warning failed to remove deps hackney ebin hackney http beam permission denied warning failed to remove deps hackney ebin hackney beam permission denied warning failed to remove deps hackney ebin hackney app beam permission denied warning failed to remove deps hackney ebin hackney http connect beam permission denied warning failed to remove deps hackney ebin hackney response beam permission denied warning failed to remove deps hackney ebin hackney app permission denied warning failed to remove deps hackney ebin hackney headers new beam permission denied warning failed to remove deps hackney ebin hackney cookie beam permission denied warning failed to remove deps hackney ebin hackney request beam permission denied warning failed to remove deps hackney ebin hackney util beam permission denied warning failed to remove deps hackney ebin hackney connect beam permission denied warning failed to remove deps hackney ebin hackney date beam permission denied warning failed to remove deps hackney ebin hackney ssl beam permission denied warning failed to remove deps hackney ebin hackney bstr beam permission denied warning failed to remove deps hackney ebin hackney tcp beam permission denied warning failed to remove deps hackney ebin hackney local tcp beam permission denied warning failed to remove deps hackney erlcinfo permission denied warning failed to remove deps hackney hex metadata config permission denied warning failed to remove deps hackney readme md permission denied warning failed to remove deps hackney rebar config permission denied warning failed to remove deps hackney include hackney hrl permission denied warning failed to remove deps hackney include hackney lib hrl permission denied warning failed to remove deps hackney fetch permission denied warning failed to remove deps hackney hex permission denied warning failed to remove deps hackney news md permission denied removing pycache removing specifications gdd people json traceback most recent call last file tmpfs src github synthtool autosynth synth py line in inner main sys exit exit code skipped systemexit during handling of the above exception another exception occurred traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main executor check call cwd working repo path file tmpfs src github synthtool autosynth executor py line in check call subprocess check call command args file home kbuilder pyenv versions lib subprocess py line in check call raise calledprocesserror retcode cmd subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log ,1
+973,22283527824.0,IssuesEvent,2022-06-11 08:38:04,metabase/metabase,https://api.github.com/repos/metabase/metabase,reopened,Changing account password to a non-complex password misleads the user with a success message even though the process fails,Type:Bug Priority:P2 .Frontend Administration/People,"**Describe the bug**
+Attempting to change account password to a common one will notify the user that password change has succeeded when in fact it hasn't due to lack of password complexity.
+
+**Logs**
+See Screenshots.
+
+**To Reproduce**
+Steps to reproduce the behavior:
+1. Go to 'Account Settings'
+2. Enter a common password to 'Create a Password' fields
+3. Click 'Save'
+
+**Expected behavior**
+When password is too common, it should notify the user that the password is too common and the process has failed.
+
+**Screenshots**
+
+
+**Information about your Metabase Installation:**
+
+- Your browser and the version: Chrome 101.0
+- Your operating system: Ubuntu 20.04
+- Metabase version: 0.43.2
+- Metabase hosting environment: Docker
+- Metabase internal database: Postgres
+
+**Severity**
+Not critical but it causes confusion when new users try to change their temporary password.
+",1.0,"Changing account password to a non-complex password misleads the user with a success message even though the process fails - **Describe the bug**
+Attempting to change account password to a common one will notify the user that password change has succeeded when in fact it hasn't due to lack of password complexity.
+
+**Logs**
+See Screenshots.
+
+**To Reproduce**
+Steps to reproduce the behavior:
+1. Go to 'Account Settings'
+2. Enter a common password to 'Create a Password' fields
+3. Click 'Save'
+
+**Expected behavior**
+When password is too common, it should notify the user that the password is too common and the process has failed.
+
+**Screenshots**
+
+
+**Information about your Metabase Installation:**
+
+- Your browser and the version: Chrome 101.0
+- Your operating system: Ubuntu 20.04
+- Metabase version: 0.43.2
+- Metabase hosting environment: Docker
+- Metabase internal database: Postgres
+
+**Severity**
+Not critical but it causes confusion when new users try to change their temporary password.
+",1,changing account password to a non complex password misleads the user with a success message even though the process fails describe the bug attempting to change account password to a common one will notify the user that password change has succeeded when in fact it hasn t due to lack of password complexity logs see screenshots to reproduce steps to reproduce the behavior go to account settings enter a common password to create a password fields click save expected behavior when password is too common it should notify the user that the password is too common and the process has failed screenshots information about your metabase installation your browser and the version chrome your operating system ubuntu metabase version metabase hosting environment docker metabase internal database postgres severity not critical but it causes confusion when new users try to change their temporary password ,1
+21614,7048520151.0,IssuesEvent,2018-01-02 18:01:11,moby/moby,https://api.github.com/repos/moby/moby,closed,Since 17.12 docker CE - HEALTHCHECK no longer inherits WORKDIR (PWD==/) while in 17.09 HEALTHCHECK it was inherited from WORKDIR (PWD==WORKDIR).,area/builder version/17.09,"**Description**
+Since 17.12 docker CE - HEALTHCHECK no longer inherits WORKDIR (PWD==/) while in 17.09 HEALTHCHECK it was inherited from WORKDIR (PWD==WORKDIR) in Dockerfile. Which might break containers which were expecting Healthcheck to execute commands while being in WORKDIR
+
+1. Build simple docker image, simple Dockerfile:
+```
+FROM busybox
+WORKDIR /root
+
+HEALTHCHECK CMD env
+```
+
+```
+docker build .
+Sending build context to Docker daemon 2.048kB
+Step 1/3 : FROM busybox
+ ---> 6ad733544a63
+Step 2/3 : WORKDIR /root
+ ---> Using cache
+ ---> d2147f9d3cd8
+Step 3/3 : HEALTHCHECK CMD env
+ ---> Running in b465b176cf95
+Removing intermediate container b465b176cf95
+ ---> 12344e707666
+Successfully built 12344e707666
+```
+
+2. Start docker container in 17.12:
+```
+docker run -ti --rm --name busy docker_build_id
+~ # pwd
+/root
+
+```
+3. Check for healthcheck output:
+```
+[
+ {
+ ""Id"": ""af70cb836acbfad56e804e4b3d1210d0e97a25d145a7a19cd02eb5488a6e6323"",
+ ""Created"": ""2018-01-02T15:42:26.283916609Z"",
+ ""Path"": ""sh"",
+ ""Args"": [],
+ ""State"": {
+ ""Status"": ""running"",
+ ""Running"": true,
+ ""Paused"": false,
+ ""Restarting"": false,
+ ""OOMKilled"": false,
+ ""Dead"": false,
+ ""Pid"": 25724,
+ ""ExitCode"": 0,
+ ""Error"": """",
+ ""StartedAt"": ""2018-01-02T15:42:26.654086Z"",
+ ""FinishedAt"": ""0001-01-01T00:00:00Z"",
+ ""Health"": {
+ ""Status"": ""healthy"",
+ ""FailingStreak"": 0,
+ ""Log"": [
+ {
+ ""Start"": ""2018-01-02T17:42:56.654412529+02:00"",
+ ""End"": ""2018-01-02T17:42:56.700667346+02:00"",
+ ""ExitCode"": 0,
+ ""Output"": ""HOSTNAME=af70cb836acb\nSHLVL=1\nHOME=/root\nPATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin\nPWD=/\n""
+ },
+
+```
+4. PWD is ""/"" not WORKDIR which was ""/root""
+
+Same workflow executed on 17.09 yields:
+```
+[
+ {
+ ""Id"": ""3fc05daafccb7884926756f68c04aef64944cdc0be2ad88fcbc18c30c94353f3"",
+ ""Created"": ""2018-01-02T17:38:53.011840641Z"",
+ ""Path"": ""sh"",
+ ""Args"": [],
+ ""State"": {
+ ""Status"": ""running"",
+ ""Running"": true,
+ ""Paused"": false,
+ ""Restarting"": false,
+ ""OOMKilled"": false,
+ ""Dead"": false,
+ ""Pid"": 27262,
+ ""ExitCode"": 0,
+ ""Error"": """",
+ ""StartedAt"": ""2018-01-02T17:38:53.314761685Z"",
+ ""FinishedAt"": ""0001-01-01T00:00:00Z"",
+ ""Health"": {
+ ""Status"": ""healthy"",
+ ""FailingStreak"": 0,
+ ""Log"": [
+ {
+ ""Start"": ""2018-01-02T19:39:23.315050864+02:00"",
+ ""End"": ""2018-01-02T19:39:23.355276833+02:00"",
+ ""ExitCode"": 0,
+ ""Output"": ""HOSTNAME=3fc05daafccb\nSHLVL=1\nHOME=/root\nPATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin\nPWD=/root\n""
+ }
+ ]
+ }
+```
+
+**Describe the results you received:**
+WORKDIR is no longer set for healtchecks which might break assumptions which worked in previous versions. Release notes is not mentioning such breaking changes.
+
+**Describe the results you expected:**
+Release notes should mention breaking change or healthcheck should have same WORKDIR as exec
+
+**Additional information you deem important (e.g. issue happens only occasionally):**
+Probably caused by: https://github.com/moby/moby/issues/33753
+**Output of `docker version`:**
+
+```
+Client:
+ Version: 17.09.0-ce
+ API version: 1.32
+ Go version: go1.8.3
+ Git commit: afdb6d4
+ Built: Tue Sep 26 22:41:23 2017
+ OS/Arch: linux/amd64
+
+Server:
+ Version: 17.09.0-ce
+ API version: 1.32 (minimum version 1.12)
+ Go version: go1.8.3
+ Git commit: afdb6d4
+ Built: Tue Sep 26 22:42:49 2017
+ OS/Arch: linux/amd64
+ Experimental: false
+
+```
+and
+```
+Client:
+ Version: 17.12.0-ce
+ API version: 1.35
+ Go version: go1.9.2
+ Git commit: c97c6d6
+ Built: Wed Dec 27 20:10:14 2017
+ OS/Arch: linux/amd64
+
+Server:
+ Engine:
+ Version: 17.12.0-ce
+ API version: 1.35 (minimum version 1.12)
+ Go version: go1.9.2
+ Git commit: c97c6d6
+ Built: Wed Dec 27 20:12:46 2017
+ OS/Arch: linux/amd64
+ Experimental: false
+```
+**Output of `docker info`:**
+N/A
+
+**Additional environment details (AWS, VirtualBox, physical, etc.):**
+Physical (affects AWS based machines also)",1.0,"Since 17.12 docker CE - HEALTHCHECK no longer inherits WORKDIR (PWD==/) while in 17.09 HEALTHCHECK it was inherited from WORKDIR (PWD==WORKDIR). - **Description**
+Since 17.12 docker CE - HEALTHCHECK no longer inherits WORKDIR (PWD==/) while in 17.09 HEALTHCHECK it was inherited from WORKDIR (PWD==WORKDIR) in Dockerfile. Which might break containers which were expecting Healthcheck to execute commands while being in WORKDIR
+
+1. Build simple docker image, simple Dockerfile:
+```
+FROM busybox
+WORKDIR /root
+
+HEALTHCHECK CMD env
+```
+
+```
+docker build .
+Sending build context to Docker daemon 2.048kB
+Step 1/3 : FROM busybox
+ ---> 6ad733544a63
+Step 2/3 : WORKDIR /root
+ ---> Using cache
+ ---> d2147f9d3cd8
+Step 3/3 : HEALTHCHECK CMD env
+ ---> Running in b465b176cf95
+Removing intermediate container b465b176cf95
+ ---> 12344e707666
+Successfully built 12344e707666
+```
+
+2. Start docker container in 17.12:
+```
+docker run -ti --rm --name busy docker_build_id
+~ # pwd
+/root
+
+```
+3. Check for healthcheck output:
+```
+[
+ {
+ ""Id"": ""af70cb836acbfad56e804e4b3d1210d0e97a25d145a7a19cd02eb5488a6e6323"",
+ ""Created"": ""2018-01-02T15:42:26.283916609Z"",
+ ""Path"": ""sh"",
+ ""Args"": [],
+ ""State"": {
+ ""Status"": ""running"",
+ ""Running"": true,
+ ""Paused"": false,
+ ""Restarting"": false,
+ ""OOMKilled"": false,
+ ""Dead"": false,
+ ""Pid"": 25724,
+ ""ExitCode"": 0,
+ ""Error"": """",
+ ""StartedAt"": ""2018-01-02T15:42:26.654086Z"",
+ ""FinishedAt"": ""0001-01-01T00:00:00Z"",
+ ""Health"": {
+ ""Status"": ""healthy"",
+ ""FailingStreak"": 0,
+ ""Log"": [
+ {
+ ""Start"": ""2018-01-02T17:42:56.654412529+02:00"",
+ ""End"": ""2018-01-02T17:42:56.700667346+02:00"",
+ ""ExitCode"": 0,
+ ""Output"": ""HOSTNAME=af70cb836acb\nSHLVL=1\nHOME=/root\nPATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin\nPWD=/\n""
+ },
+
+```
+4. PWD is ""/"" not WORKDIR which was ""/root""
+
+Same workflow executed on 17.09 yields:
+```
+[
+ {
+ ""Id"": ""3fc05daafccb7884926756f68c04aef64944cdc0be2ad88fcbc18c30c94353f3"",
+ ""Created"": ""2018-01-02T17:38:53.011840641Z"",
+ ""Path"": ""sh"",
+ ""Args"": [],
+ ""State"": {
+ ""Status"": ""running"",
+ ""Running"": true,
+ ""Paused"": false,
+ ""Restarting"": false,
+ ""OOMKilled"": false,
+ ""Dead"": false,
+ ""Pid"": 27262,
+ ""ExitCode"": 0,
+ ""Error"": """",
+ ""StartedAt"": ""2018-01-02T17:38:53.314761685Z"",
+ ""FinishedAt"": ""0001-01-01T00:00:00Z"",
+ ""Health"": {
+ ""Status"": ""healthy"",
+ ""FailingStreak"": 0,
+ ""Log"": [
+ {
+ ""Start"": ""2018-01-02T19:39:23.315050864+02:00"",
+ ""End"": ""2018-01-02T19:39:23.355276833+02:00"",
+ ""ExitCode"": 0,
+ ""Output"": ""HOSTNAME=3fc05daafccb\nSHLVL=1\nHOME=/root\nPATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin\nPWD=/root\n""
+ }
+ ]
+ }
+```
+
+**Describe the results you received:**
+WORKDIR is no longer set for healtchecks which might break assumptions which worked in previous versions. Release notes is not mentioning such breaking changes.
+
+**Describe the results you expected:**
+Release notes should mention breaking change or healthcheck should have same WORKDIR as exec
+
+**Additional information you deem important (e.g. issue happens only occasionally):**
+Probably caused by: https://github.com/moby/moby/issues/33753
+**Output of `docker version`:**
+
+```
+Client:
+ Version: 17.09.0-ce
+ API version: 1.32
+ Go version: go1.8.3
+ Git commit: afdb6d4
+ Built: Tue Sep 26 22:41:23 2017
+ OS/Arch: linux/amd64
+
+Server:
+ Version: 17.09.0-ce
+ API version: 1.32 (minimum version 1.12)
+ Go version: go1.8.3
+ Git commit: afdb6d4
+ Built: Tue Sep 26 22:42:49 2017
+ OS/Arch: linux/amd64
+ Experimental: false
+
+```
+and
+```
+Client:
+ Version: 17.12.0-ce
+ API version: 1.35
+ Go version: go1.9.2
+ Git commit: c97c6d6
+ Built: Wed Dec 27 20:10:14 2017
+ OS/Arch: linux/amd64
+
+Server:
+ Engine:
+ Version: 17.12.0-ce
+ API version: 1.35 (minimum version 1.12)
+ Go version: go1.9.2
+ Git commit: c97c6d6
+ Built: Wed Dec 27 20:12:46 2017
+ OS/Arch: linux/amd64
+ Experimental: false
+```
+**Output of `docker info`:**
+N/A
+
+**Additional environment details (AWS, VirtualBox, physical, etc.):**
+Physical (affects AWS based machines also)",0,since docker ce healthcheck no longer inherits workdir pwd while in healthcheck it was inherited from workdir pwd workdir description since docker ce healthcheck no longer inherits workdir pwd while in healthcheck it was inherited from workdir pwd workdir in dockerfile which might break containers which were expecting healthcheck to execute commands while being in workdir build simple docker image simple dockerfile from busybox workdir root healthcheck cmd env docker build sending build context to docker daemon step from busybox step workdir root using cache step healthcheck cmd env running in removing intermediate container successfully built start docker container in docker run ti rm name busy docker build id pwd root check for healthcheck output id created path sh args state status running running true paused false restarting false oomkilled false dead false pid exitcode error startedat finishedat health status healthy failingstreak log start end exitcode output hostname nshlvl nhome root npath usr local sbin usr local bin usr sbin usr bin sbin bin npwd n pwd is not workdir which was root same workflow executed on yields id created path sh args state status running running true paused false restarting false oomkilled false dead false pid exitcode error startedat finishedat health status healthy failingstreak log start end exitcode output hostname nshlvl nhome root npath usr local sbin usr local bin usr sbin usr bin sbin bin npwd root n describe the results you received workdir is no longer set for healtchecks which might break assumptions which worked in previous versions release notes is not mentioning such breaking changes describe the results you expected release notes should mention breaking change or healthcheck should have same workdir as exec additional information you deem important e g issue happens only occasionally probably caused by output of docker version client version ce api version go version git commit built tue sep os arch linux server version ce api version minimum version go version git commit built tue sep os arch linux experimental false and client version ce api version go version git commit built wed dec os arch linux server engine version ce api version minimum version go version git commit built wed dec os arch linux experimental false output of docker info n a additional environment details aws virtualbox physical etc physical affects aws based machines also ,0
+690,12407045783.0,IssuesEvent,2020-05-21 20:18:41,PointCloudLibrary/pcl,https://api.github.com/repos/PointCloudLibrary/pcl,closed,Investigate disabled test_people on Ubuntu CI,module: ci module: people module: test platform: linux status: stale,"A comment to discuss later, just noticed this in the CI log
+```
+-- The following subsystems will be built:
+-- common
+-- octree
+-- io
+-- kdtree
+-- geometry
+-- search
+-- visualization
+-- surface
+-- sample_consensus
+-- filters
+-- 2d
+-- features
+-- simulation
+-- tracking
+-- ml
+-- segmentation
+-- registration
+-- recognition
+-- outofcore
+-- keypoints
+-- stereo
+-- examples
+-- global_tests
+-- tests_2d
+-- tests_common
+-- tests_features
+-- tests_filters
+-- tests_geometry
+-- tests_io
+-- tests_kdtree
+-- tests_keypoints
+-- tests_octree
+-- tests_outofcore
+-- tests_recognition
+-- tests_registration
+-- tests_search
+-- tests_surface
+-- tests_segmentation
+-- tests_sample_consensus
+-- tests_visualization
+-- tools
+-- apps
+ building:
+ |_ in_hand_scanner
+ |_ 3d_rec_framework
+ |_ cloud_composer
+ |_ point_cloud_editor
+ |_ modeler
+-- people
+-- The following subsystems will not be built:
+-- tests_people: Requires people.
+```
+
+Probably unrelated to this PR.
+
+_Originally posted by @SergioRAgostinho in https://github.com/PointCloudLibrary/pcl/pull/3068#issuecomment-490086357_",1.0,"Investigate disabled test_people on Ubuntu CI - A comment to discuss later, just noticed this in the CI log
+```
+-- The following subsystems will be built:
+-- common
+-- octree
+-- io
+-- kdtree
+-- geometry
+-- search
+-- visualization
+-- surface
+-- sample_consensus
+-- filters
+-- 2d
+-- features
+-- simulation
+-- tracking
+-- ml
+-- segmentation
+-- registration
+-- recognition
+-- outofcore
+-- keypoints
+-- stereo
+-- examples
+-- global_tests
+-- tests_2d
+-- tests_common
+-- tests_features
+-- tests_filters
+-- tests_geometry
+-- tests_io
+-- tests_kdtree
+-- tests_keypoints
+-- tests_octree
+-- tests_outofcore
+-- tests_recognition
+-- tests_registration
+-- tests_search
+-- tests_surface
+-- tests_segmentation
+-- tests_sample_consensus
+-- tests_visualization
+-- tools
+-- apps
+ building:
+ |_ in_hand_scanner
+ |_ 3d_rec_framework
+ |_ cloud_composer
+ |_ point_cloud_editor
+ |_ modeler
+-- people
+-- The following subsystems will not be built:
+-- tests_people: Requires people.
+```
+
+Probably unrelated to this PR.
+
+_Originally posted by @SergioRAgostinho in https://github.com/PointCloudLibrary/pcl/pull/3068#issuecomment-490086357_",1,investigate disabled test people on ubuntu ci a comment to discuss later just noticed this in the ci log the following subsystems will be built common octree io kdtree geometry search visualization surface sample consensus filters features simulation tracking ml segmentation registration recognition outofcore keypoints stereo examples global tests tests tests common tests features tests filters tests geometry tests io tests kdtree tests keypoints tests octree tests outofcore tests recognition tests registration tests search tests surface tests segmentation tests sample consensus tests visualization tools apps building in hand scanner rec framework cloud composer point cloud editor modeler people the following subsystems will not be built tests people requires people probably unrelated to this pr originally posted by sergioragostinho in ,1
+1106,26996594134.0,IssuesEvent,2023-02-10 01:56:18,wordpress-mobile/WordPress-iOS,https://api.github.com/repos/wordpress-mobile/WordPress-iOS,closed,Crash when navigating to People screen,[Type] Bug [Pri] High People Management Core Data,"### Expected behavior
+
+The app should load the People screen.
+
+### Actual behavior
+
+The app crashes when loading the People screen.
+
+### Steps to reproduce the behavior
+
+1. Log into the app (I used the WP app)
+2. Navigate to the People screen
+3. Notice the app crashes
+
+Relevant error:
+```
+*** Terminating app due to uncaught exception 'NSInvalidArgumentException', reason: 'Illegal attempt to establish a relationship 'blog' between objects in different contexts (source = (entity: Role; id: 0x1198fd460
+```
+
+##### Tested on iPhone 11, iOS 16.3, WPiOS `trunk` a2e6168e448852b0099faea5a01f35249a0ae2e4
+",1.0,"Crash when navigating to People screen - ### Expected behavior
+
+The app should load the People screen.
+
+### Actual behavior
+
+The app crashes when loading the People screen.
+
+### Steps to reproduce the behavior
+
+1. Log into the app (I used the WP app)
+2. Navigate to the People screen
+3. Notice the app crashes
+
+Relevant error:
+```
+*** Terminating app due to uncaught exception 'NSInvalidArgumentException', reason: 'Illegal attempt to establish a relationship 'blog' between objects in different contexts (source = (entity: Role; id: 0x1198fd460
+```
+
+##### Tested on iPhone 11, iOS 16.3, WPiOS `trunk` a2e6168e448852b0099faea5a01f35249a0ae2e4
+",1,crash when navigating to people screen expected behavior the app should load the people screen actual behavior the app crashes when loading the people screen steps to reproduce the behavior log into the app i used the wp app navigate to the people screen notice the app crashes relevant error terminating app due to uncaught exception nsinvalidargumentexception reason illegal attempt to establish a relationship blog between objects in different contexts source entity role id tested on iphone ios wpios trunk ,1
+192,4118638159.0,IssuesEvent,2016-06-08 12:20:26,wordpress-mobile/WordPress-iOS,https://api.github.com/repos/wordpress-mobile/WordPress-iOS,closed,People Management: Infinite Scroll,People Management [Type] Enhancement,"#### Details:
+Current People Management implementation will sync up to the first 50 Users / Followers.
+Implement proper **Infinite Scroll** support.
+
+Ref. #5030
+",1.0,"People Management: Infinite Scroll - #### Details:
+Current People Management implementation will sync up to the first 50 Users / Followers.
+Implement proper **Infinite Scroll** support.
+
+Ref. #5030
+",1,people management infinite scroll details current people management implementation will sync up to the first users followers implement proper infinite scroll support ref ,1
+249,5144063262.0,IssuesEvent,2017-01-12 17:34:07,PopulateTools/gobierto,https://api.github.com/repos/PopulateTools/gobierto,closed,People module bootstrap,acceptance gobierto-admin gobierto-people,"- [x] Set up the new Gobierto module
+- [x] Dynamic content block feature design and implementation
+- [x] Feature implementation
+ - [x] People (w/ Dynamic content extension)
+ - [x] Events
+ - [x] Statements (w/ Dynamic content extension)
+ - [x] Posts",1.0,"People module bootstrap - - [x] Set up the new Gobierto module
+- [x] Dynamic content block feature design and implementation
+- [x] Feature implementation
+ - [x] People (w/ Dynamic content extension)
+ - [x] Events
+ - [x] Statements (w/ Dynamic content extension)
+ - [x] Posts",1,people module bootstrap set up the new gobierto module dynamic content block feature design and implementation feature implementation people w dynamic content extension events statements w dynamic content extension posts,1
+952,21030150597.0,IssuesEvent,2022-03-31 00:05:56,planningcenter/developers,https://api.github.com/repos/planningcenter/developers,closed,"Please, just a boolean will do for webhook fire.",People,"**Related Product**
+Which product is this feature related to?
+People, Webhooks
+
+**Is your feature request related to a problem? Please describe.**
+Somewhat, or maybe so.
+
+**Describe the solution you'd like**
+Currently, when you edit a custom field from a Person, it will fire a webhook with an event of `*.people.updated` ""if field datum already exists `(PUT)`"" and or `created` ""If it blank `(POST)`"". We do have monitoring in-place, for example a Persons Name was already added to known list, and we mark mark its custom field some data, after sending a post, put or delete, a webhook is fired, without knowing it was due to field_datum change/delete.
+
+I would like that something to the attributes or relation or anything that can determine that this webhook was received due to field_datum change/put or delete `field_datum: 'delete/post/put'`. Just so we can ignore it even before it passes through. Just a single attribute like that will suffice. PLEASE!!!
+
+**Describe alternatives you've considered**
+-
+
+**Additional context**
+- Not possible to have a screenshot, but I know PCO dev's for API knows what I'm talking about.
+
+## I have..
+
+- [x] Reviewed the documentation found at https://developer.planning.center/docs
+- [x] Searched for previous issues asking for this feature request
+- [x] Removed all private information from this issue (credentials, tokens, emails, phone numbers, etc.)
+- [x] Reviewed my issue for completeness",1.0,"Please, just a boolean will do for webhook fire. - **Related Product**
+Which product is this feature related to?
+People, Webhooks
+
+**Is your feature request related to a problem? Please describe.**
+Somewhat, or maybe so.
+
+**Describe the solution you'd like**
+Currently, when you edit a custom field from a Person, it will fire a webhook with an event of `*.people.updated` ""if field datum already exists `(PUT)`"" and or `created` ""If it blank `(POST)`"". We do have monitoring in-place, for example a Persons Name was already added to known list, and we mark mark its custom field some data, after sending a post, put or delete, a webhook is fired, without knowing it was due to field_datum change/delete.
+
+I would like that something to the attributes or relation or anything that can determine that this webhook was received due to field_datum change/put or delete `field_datum: 'delete/post/put'`. Just so we can ignore it even before it passes through. Just a single attribute like that will suffice. PLEASE!!!
+
+**Describe alternatives you've considered**
+-
+
+**Additional context**
+- Not possible to have a screenshot, but I know PCO dev's for API knows what I'm talking about.
+
+## I have..
+
+- [x] Reviewed the documentation found at https://developer.planning.center/docs
+- [x] Searched for previous issues asking for this feature request
+- [x] Removed all private information from this issue (credentials, tokens, emails, phone numbers, etc.)
+- [x] Reviewed my issue for completeness",1,please just a boolean will do for webhook fire related product which product is this feature related to people webhooks is your feature request related to a problem please describe somewhat or maybe so describe the solution you d like currently when you edit a custom field from a person it will fire a webhook with an event of people updated if field datum already exists put and or created if it blank post we do have monitoring in place for example a persons name was already added to known list and we mark mark its custom field some data after sending a post put or delete a webhook is fired without knowing it was due to field datum change delete i would like that something to the attributes or relation or anything that can determine that this webhook was received due to field datum change put or delete field datum delete post put just so we can ignore it even before it passes through just a single attribute like that will suffice please describe alternatives you ve considered additional context not possible to have a screenshot but i know pco dev s for api knows what i m talking about i have reviewed the documentation found at searched for previous issues asking for this feature request removed all private information from this issue credentials tokens emails phone numbers etc reviewed my issue for completeness,1
+379,7048245161.0,IssuesEvent,2018-01-02 16:52:29,training-center/study-groups,https://api.github.com/repos/training-center/study-groups,closed,Grupo de Estudos JavaScript,bringing-people,"Criando esse issue para interessados em criar um grupo de estudos de JavaScript
+
+Lembrando: É necessário no mínimo um organizador + 4 membros, e esses membros serão responsáveis por manter o grupo se reunindo, definindo pauta e planos de estudos.
+
+(Nota: eu **não** tenho interesse de ser organizador nem membro)",1.0,"Grupo de Estudos JavaScript - Criando esse issue para interessados em criar um grupo de estudos de JavaScript
+
+Lembrando: É necessário no mínimo um organizador + 4 membros, e esses membros serão responsáveis por manter o grupo se reunindo, definindo pauta e planos de estudos.
+
+(Nota: eu **não** tenho interesse de ser organizador nem membro)",1,grupo de estudos javascript criando esse issue para interessados em criar um grupo de estudos de javascript lembrando é necessário no mínimo um organizador membros e esses membros serão responsáveis por manter o grupo se reunindo definindo pauta e planos de estudos nota eu não tenho interesse de ser organizador nem membro ,1
+234686,19247989552.0,IssuesEvent,2021-12-09 00:03:19,atlasmap/atlasmap,https://api.github.com/repos/atlasmap/atlasmap,closed,[TASK] Let chromatic ignore NodesArc story,group/ui group/tests,"The moving svg is detected as a change again, got a reply from chromatic team that we have to add `disable` flag into MDX file if there's MDX file.
+",1.0,"[TASK] Let chromatic ignore NodesArc story - The moving svg is detected as a change again, got a reply from chromatic team that we have to add `disable` flag into MDX file if there's MDX file.
+",0, let chromatic ignore nodesarc story the moving svg is detected as a change again got a reply from chromatic team that we have to add disable flag into mdx file if there s mdx file ,0
+177130,6574639592.0,IssuesEvent,2017-09-11 13:36:35,mantidproject/mantid,https://api.github.com/repos/mantidproject/mantid,closed,Zooming into peaks in the SplatterplotView makes rendering very slow,Group: Vates Priority: High,"When loading an event workspace into the splatter plot view the data is displayed as PointGaussian. When zooming into a peak region the VSI freezes and cannot cope with updating the rendered image in a reasonable time (takes seconds between frame updates). When switching from PointGaussian to Points it is fine.
+
+### Expected behavior
+
+Frame update with little lag when zooming into a peak.
+
+### Actual behavior
+
+Frame updates with lags on the order of seconds when zooming into a peak.
+
+### Steps to reproduce the behavior
+
+1. Load a MDEventWorkspace into the splatter plot view of the VSI
+2. Zoom into a peak region
+ * Confirm that it gets laggy
+
+It can be hard to zoom nicely into a peak
+
+### Platforms affected
+All",1.0,"Zooming into peaks in the SplatterplotView makes rendering very slow - When loading an event workspace into the splatter plot view the data is displayed as PointGaussian. When zooming into a peak region the VSI freezes and cannot cope with updating the rendered image in a reasonable time (takes seconds between frame updates). When switching from PointGaussian to Points it is fine.
+
+### Expected behavior
+
+Frame update with little lag when zooming into a peak.
+
+### Actual behavior
+
+Frame updates with lags on the order of seconds when zooming into a peak.
+
+### Steps to reproduce the behavior
+
+1. Load a MDEventWorkspace into the splatter plot view of the VSI
+2. Zoom into a peak region
+ * Confirm that it gets laggy
+
+It can be hard to zoom nicely into a peak
+
+### Platforms affected
+All",0,zooming into peaks in the splatterplotview makes rendering very slow when loading an event workspace into the splatter plot view the data is displayed as pointgaussian when zooming into a peak region the vsi freezes and cannot cope with updating the rendered image in a reasonable time takes seconds between frame updates when switching from pointgaussian to points it is fine expected behavior frame update with little lag when zooming into a peak actual behavior frame updates with lags on the order of seconds when zooming into a peak steps to reproduce the behavior load a mdeventworkspace into the splatter plot view of the vsi zoom into a peak region confirm that it gets laggy it can be hard to zoom nicely into a peak platforms affected all,0
+148360,13235555736.0,IssuesEvent,2020-08-18 18:15:12,kubernetes-sigs/cluster-api,https://api.github.com/repos/kubernetes-sigs/cluster-api,closed,Document clusterctl version support,area/clusterctl kind/documentation,"The support matrix (added in https://github.com/kubernetes-sigs/cluster-api/pull/3031) needs details for clusterctl.
+
+/kind documentation",1.0,"Document clusterctl version support - The support matrix (added in https://github.com/kubernetes-sigs/cluster-api/pull/3031) needs details for clusterctl.
+
+/kind documentation",0,document clusterctl version support the support matrix added in needs details for clusterctl kind documentation,0
+275044,8570831897.0,IssuesEvent,2018-11-11 23:44:01,StrangeLoopGames/EcoIssues,https://api.github.com/repos/StrangeLoopGames/EcoIssues,opened,QoL-Fixes for Player Currencies,D3nnis3n's List High Priority,"I got a lot of opinions on player credits and propose the following QoL Fixes for this matter
+
+- Allow the world leader to set a default currency for the world that changes the default currency from player credits to the chosen one of any new placed objects that use credits. (Shop, Crafting Tables, as well as new contracts)
+- Allow players to either hide or remove player credits as a server setting in general. On big servers with many players it's hard to scroll through a laggy mess of credits that are not used at all in order to choose the currency you want to use.
+- Sort the drop down menus for currencies to have backed currencies on TOP instead of buttom and sort the backed currencies for how much they are used.",1.0,"QoL-Fixes for Player Currencies - I got a lot of opinions on player credits and propose the following QoL Fixes for this matter
+
+- Allow the world leader to set a default currency for the world that changes the default currency from player credits to the chosen one of any new placed objects that use credits. (Shop, Crafting Tables, as well as new contracts)
+- Allow players to either hide or remove player credits as a server setting in general. On big servers with many players it's hard to scroll through a laggy mess of credits that are not used at all in order to choose the currency you want to use.
+- Sort the drop down menus for currencies to have backed currencies on TOP instead of buttom and sort the backed currencies for how much they are used.",0,qol fixes for player currencies i got a lot of opinions on player credits and propose the following qol fixes for this matter allow the world leader to set a default currency for the world that changes the default currency from player credits to the chosen one of any new placed objects that use credits shop crafting tables as well as new contracts allow players to either hide or remove player credits as a server setting in general on big servers with many players it s hard to scroll through a laggy mess of credits that are not used at all in order to choose the currency you want to use sort the drop down menus for currencies to have backed currencies on top instead of buttom and sort the backed currencies for how much they are used ,0
+50,2774202506.0,IssuesEvent,2015-05-04 06:26:19,punongbayan-araullo/tickets,https://api.github.com/repos/punongbayan-araullo/tickets,opened,"Clicking ""Shift Change"" redirects to the ""Time-in correction application"" page",priority - high status - accepted system - peoplecore type - bug,"Clicking ""Shift Change"" redirects to the ""Time-in correction application"" page",1.0,"Clicking ""Shift Change"" redirects to the ""Time-in correction application"" page - Clicking ""Shift Change"" redirects to the ""Time-in correction application"" page",1,clicking shift change redirects to the time in correction application page clicking shift change redirects to the time in correction application page,1
+300287,22666271320.0,IssuesEvent,2022-07-03 00:03:52,StoneCypher/fsl,https://api.github.com/repos/StoneCypher/fsl,opened,Code quality report,Chore Automation Underway Documentation,"Derive metrics from lines of code, test counts; put report in markdown; echo to console
+
+Related to #736 ",1.0,"Code quality report - Derive metrics from lines of code, test counts; put report in markdown; echo to console
+
+Related to #736 ",0,code quality report derive metrics from lines of code test counts put report in markdown echo to console related to ,0
+59816,14476398430.0,IssuesEvent,2020-12-10 04:01:35,imatlin/Sonar-Plugin,https://api.github.com/repos/imatlin/Sonar-Plugin,opened,CVE-2019-11358 (Medium) detected in jquery-3.3.1.min.js,security vulnerability,"## CVE-2019-11358 - Medium Severity Vulnerability
+ Vulnerable Library - jquery-3.3.1.min.js
+
+
+
+jQuery before 3.4.0, as used in Drupal, Backdrop CMS, and other products, mishandles jQuery.extend(true, {}, ...) because of Object.prototype pollution. If an unsanitized source object contained an enumerable __proto__ property, it could extend the native Object.prototype.
+
+
+
+jQuery before 3.4.0, as used in Drupal, Backdrop CMS, and other products, mishandles jQuery.extend(true, {}, ...) because of Object.prototype pollution. If an unsanitized source object contained an enumerable __proto__ property, it could extend the native Object.prototype.
+
+
+
+
+
+
+
+",0,cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to vulnerable library sonar plugin src main resources static jquery min js dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch master vulnerability details jquery before as used in drupal backdrop cms and other products mishandles jquery extend true because of object prototype pollution if an unsanitized source object contained an enumerable proto property it could extend the native object prototype publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails jquery before as used in drupal backdrop cms and other products mishandles jquery extend true because of object prototype pollution if an unsanitized source object contained an enumerable proto property it could extend the native object prototype vulnerabilityurl ,0
+1173,30644672182.0,IssuesEvent,2023-07-25 02:54:09,restincode/restincode,https://api.github.com/repos/restincode/restincode,opened,Maria Moraru,People Add Person Needs Review,"Please fill out as much information as you can. No fields are required, but the more you can provide the better.
+
+**General Info**
+* First name:
+* Last name:
+* Handle:
+* Birth Year:
+* Death Year:
+* Link to Obituary:
+* Group Affiliations:
+* URL to main photo (or attach to Issue):
+* Description of person and/or activities:
+* Facebook memorial group URL:
+
+**Social Media Links**
+* Twitter:
+* Github:
+* LinkedIn:
+* Facebook:
+* Other:
+
+**Contributions**
+Repeat as many times as needed
+
+* Project name:
+* Project URL:
+* Project Description:
+
+**Photo Gallery**
+* URL(s) to additional photos:
+",1.0,"Maria Moraru - Please fill out as much information as you can. No fields are required, but the more you can provide the better.
+
+**General Info**
+* First name:
+* Last name:
+* Handle:
+* Birth Year:
+* Death Year:
+* Link to Obituary:
+* Group Affiliations:
+* URL to main photo (or attach to Issue):
+* Description of person and/or activities:
+* Facebook memorial group URL:
+
+**Social Media Links**
+* Twitter:
+* Github:
+* LinkedIn:
+* Facebook:
+* Other:
+
+**Contributions**
+Repeat as many times as needed
+
+* Project name:
+* Project URL:
+* Project Description:
+
+**Photo Gallery**
+* URL(s) to additional photos:
+",1,maria moraru please fill out as much information as you can no fields are required but the more you can provide the better general info first name last name handle birth year death year link to obituary group affiliations url to main photo or attach to issue description of person and or activities facebook memorial group url social media links twitter github linkedin facebook other contributions repeat as many times as needed project name project url project description photo gallery url s to additional photos ,1
+396077,27100380685.0,IssuesEvent,2023-02-15 08:09:33,zonemaster/zonemaster,https://api.github.com/repos/zonemaster/zonemaster,closed,Translation: improve how a new language is created and added,A-Translation A-Documentation,"Currently, the [translation documentation](https://github.com/zonemaster/zonemaster/blob/master/docs/internal-documentation/maintenance/Instructions-for-translators.md) suggests to copy/paste an existing .po file to add a new language:
+
+> Go to the share directory and use an existing PO file, say sv.po, and make a copy of that to the new file name. Update it and ""add"" it to git before working on it.
+
+Doing so, all msgstr will be already populated. It would be nice to provide a cleaner way to create a new .po file.
+
+Each components already have a `extract-pot` recipe. This POT file could be used when creating a new language.
+
+Suggestions:
+* update the documentation instruction using the `extract-pot` recipe
+* create a new recipe that will create the new `xx.po` file for the new language",1.0,"Translation: improve how a new language is created and added - Currently, the [translation documentation](https://github.com/zonemaster/zonemaster/blob/master/docs/internal-documentation/maintenance/Instructions-for-translators.md) suggests to copy/paste an existing .po file to add a new language:
+
+> Go to the share directory and use an existing PO file, say sv.po, and make a copy of that to the new file name. Update it and ""add"" it to git before working on it.
+
+Doing so, all msgstr will be already populated. It would be nice to provide a cleaner way to create a new .po file.
+
+Each components already have a `extract-pot` recipe. This POT file could be used when creating a new language.
+
+Suggestions:
+* update the documentation instruction using the `extract-pot` recipe
+* create a new recipe that will create the new `xx.po` file for the new language",0,translation improve how a new language is created and added currently the suggests to copy paste an existing po file to add a new language go to the share directory and use an existing po file say sv po and make a copy of that to the new file name update it and add it to git before working on it doing so all msgstr will be already populated it would be nice to provide a cleaner way to create a new po file each components already have a extract pot recipe this pot file could be used when creating a new language suggestions update the documentation instruction using the extract pot recipe create a new recipe that will create the new xx po file for the new language,0
+804,15022247821.0,IssuesEvent,2021-02-01 16:43:57,microsoftgraph/microsoft-graph-toolkit,https://api.github.com/repos/microsoftgraph/microsoft-graph-toolkit,closed,Microsoft Graph Toolkit People Picker Read-only Property,Component: people-picker State: In Review feature-request good first issue,"I am using Microsoft Graph Toolkit People Picker Component as illustrated in the documentation [https://docs.microsoft.com/en-us/graph/toolkit/components/people-picker][1].
+
+All the functionality is working fine but unable to set the people picker to disabled/read-only. I have tried `disabled=""true"" but no luck. Anybody came across similar scenario?",1.0,"Microsoft Graph Toolkit People Picker Read-only Property - I am using Microsoft Graph Toolkit People Picker Component as illustrated in the documentation [https://docs.microsoft.com/en-us/graph/toolkit/components/people-picker][1].
+
+All the functionality is working fine but unable to set the people picker to disabled/read-only. I have tried `disabled=""true"" but no luck. Anybody came across similar scenario?",1,microsoft graph toolkit people picker read only property i am using microsoft graph toolkit people picker component as illustrated in the documentation all the functionality is working fine but unable to set the people picker to disabled read only i have tried disabled true but no luck anybody came across similar scenario ,1
+525,9106957786.0,IssuesEvent,2019-02-21 02:07:06,jonfroehlich/makeabilitylabwebsite,https://api.github.com/repos/jonfroehlich/makeabilitylabwebsite,closed,Reorganize the order on the people page,People Page,"As of right now, the people.html page has sections that are structured in the order listed below.
+current team -> alumni -> current collaborators -> past collaborators
+
+Jon and I thought it would be better to have it organized in this order:
+current team -> current collaborators -> alumni -> past collaborators
+
+Essentially, flip the current collaborators and alumni sections.
+
+",1.0,"Reorganize the order on the people page - As of right now, the people.html page has sections that are structured in the order listed below.
+current team -> alumni -> current collaborators -> past collaborators
+
+Jon and I thought it would be better to have it organized in this order:
+current team -> current collaborators -> alumni -> past collaborators
+
+Essentially, flip the current collaborators and alumni sections.
+
+",1,reorganize the order on the people page as of right now the people html page has sections that are structured in the order listed below current team alumni current collaborators past collaborators jon and i thought it would be better to have it organized in this order current team current collaborators alumni past collaborators essentially flip the current collaborators and alumni sections ,1
+10,2649583942.0,IssuesEvent,2015-03-15 02:15:57,fskreuz/Tree-Trail,https://api.github.com/repos/fskreuz/Tree-Trail,closed,[back-end] Create database structures,QA people! Test these!,"# Acceptance criteria
+
+- [x] Must adhere to entity specifications as stated in the wiki: https://github.com/fskreuz/Tree-Trail/wiki
+- [ ] Changes must be exported via sql dump (bzip compression preferred) and included in the repo
+- [x] Must be in MySQL
+- [ ] Add import and export procedures in the wiki",1.0,"[back-end] Create database structures - # Acceptance criteria
+
+- [x] Must adhere to entity specifications as stated in the wiki: https://github.com/fskreuz/Tree-Trail/wiki
+- [ ] Changes must be exported via sql dump (bzip compression preferred) and included in the repo
+- [x] Must be in MySQL
+- [ ] Add import and export procedures in the wiki",1, create database structures acceptance criteria must adhere to entity specifications as stated in the wiki changes must be exported via sql dump bzip compression preferred and included in the repo must be in mysql add import and export procedures in the wiki,1
+621240,19581239528.0,IssuesEvent,2022-01-04 21:39:10,VoltanFr/memcheck,https://api.github.com/repos/VoltanFr/memcheck,opened,Among the 28 cards of your deck PPP...: don't display PPP if single deck,priority-low complexity-low,This is just useless info,1.0,Among the 28 cards of your deck PPP...: don't display PPP if single deck - This is just useless info,0,among the cards of your deck ppp don t display ppp if single deck this is just useless info,0
+421577,12258672465.0,IssuesEvent,2020-05-06 15:26:44,grpc/grpc,https://api.github.com/repos/grpc/grpc,closed,C# server currently doesn't allow listening on UDS sockets,disposition/help wanted kind/enhancement lang/C# priority/P2,"C core supports UDS sockets and so would C#, but currently listening on a UDS socket doesn't work for C# servers because of a small API glitch.
+
+- while the server can accept a UDS address just fine (e.g. ""unix:///path/to/my/uds_test_socket"") but `ServerPort` always requires to pass a port and concatenates the address and the port, thus creating a broken UDS address.
+
+- creating a client for a UDS socket works just fine for a UDS socket: e.g. `var channel = new Channel(""unix:///path/to/my/uds_test_socket"", ChannelCredentials.Insecure);`
+
+https://github.com/grpc/grpc/blob/39e982a263e5c48a650990743ed398c1c76db1ac/src/csharp/Grpc.Core/Server.cs#L277
+
+workaround:
+`new ServerPort(""unix:///path/to/my/uds_test_socket"", 1234, ServerCredentials.Insecure)` actually creates a UDS socket named ""/path/to/my/uds_test_socket:1234"" and the client can connect to it (by providing the ""port number"" that will become suffix of the UDS socket name). Obviously, the name of the file is wrong, but it allows to make a successful connection.
+
+",1.0,"C# server currently doesn't allow listening on UDS sockets - C core supports UDS sockets and so would C#, but currently listening on a UDS socket doesn't work for C# servers because of a small API glitch.
+
+- while the server can accept a UDS address just fine (e.g. ""unix:///path/to/my/uds_test_socket"") but `ServerPort` always requires to pass a port and concatenates the address and the port, thus creating a broken UDS address.
+
+- creating a client for a UDS socket works just fine for a UDS socket: e.g. `var channel = new Channel(""unix:///path/to/my/uds_test_socket"", ChannelCredentials.Insecure);`
+
+https://github.com/grpc/grpc/blob/39e982a263e5c48a650990743ed398c1c76db1ac/src/csharp/Grpc.Core/Server.cs#L277
+
+workaround:
+`new ServerPort(""unix:///path/to/my/uds_test_socket"", 1234, ServerCredentials.Insecure)` actually creates a UDS socket named ""/path/to/my/uds_test_socket:1234"" and the client can connect to it (by providing the ""port number"" that will become suffix of the UDS socket name). Obviously, the name of the file is wrong, but it allows to make a successful connection.
+
+",0,c server currently doesn t allow listening on uds sockets c core supports uds sockets and so would c but currently listening on a uds socket doesn t work for c servers because of a small api glitch while the server can accept a uds address just fine e g unix path to my uds test socket but serverport always requires to pass a port and concatenates the address and the port thus creating a broken uds address creating a client for a uds socket works just fine for a uds socket e g var channel new channel unix path to my uds test socket channelcredentials insecure workaround new serverport unix path to my uds test socket servercredentials insecure actually creates a uds socket named path to my uds test socket and the client can connect to it by providing the port number that will become suffix of the uds socket name obviously the name of the file is wrong but it allows to make a successful connection ,0
+61,3054894320.0,IssuesEvent,2015-08-13 07:43:17,punongbayan-araullo/tickets,https://api.github.com/repos/punongbayan-araullo/tickets,closed,Allow PCG to post HRAN without affecting payroll processing,priority - high status - accepted system - peoplecore type - bug,Allow PCG to post HRAN without affecting payroll processing,1.0,Allow PCG to post HRAN without affecting payroll processing - Allow PCG to post HRAN without affecting payroll processing,1,allow pcg to post hran without affecting payroll processing allow pcg to post hran without affecting payroll processing,1
+629555,20036148123.0,IssuesEvent,2022-02-02 12:07:26,o3de/o3de,https://api.github.com/repos/o3de/o3de,opened,Lua Editor: Prev Document Tab option from the View menu is misspelled,kind/bug needs-triage sig/content priority/minor,"**Describe the bug**
+In the Lua Editor View menu the ""Prev Document Tab"" option is misspelled, it should be called ""Previous Document Tab"". It should also be noted that none of the other options in the Lua Editor View menu are shortened.
+
+**Steps to reproduce**
+Steps to reproduce the behavior:
+1. Go to Tools > Lua Editor.
+2. Open the View dropdown menu and check the ""Prev Document Tab"" option.
+
+**Expected behavior**
+The option is called ""Previous Document Tab"".
+
+**Actual behavior**
+The option is called ""Prev Document Tab"".
+
+**Screenshot**
+
+
+**Found in Branch**
+Development (d6c419d)
+
+**Desktop**
+ - Device: PC
+ - OS: Linux
+ - Version: Ubuntu 20.04.3 LTS
+ - CPU AMD Ryzen 5 3600
+ - GPU NVIDIA GeForce RTX 2060 SUPER
+ - Memory 16GB",1.0,"Lua Editor: Prev Document Tab option from the View menu is misspelled - **Describe the bug**
+In the Lua Editor View menu the ""Prev Document Tab"" option is misspelled, it should be called ""Previous Document Tab"". It should also be noted that none of the other options in the Lua Editor View menu are shortened.
+
+**Steps to reproduce**
+Steps to reproduce the behavior:
+1. Go to Tools > Lua Editor.
+2. Open the View dropdown menu and check the ""Prev Document Tab"" option.
+
+**Expected behavior**
+The option is called ""Previous Document Tab"".
+
+**Actual behavior**
+The option is called ""Prev Document Tab"".
+
+**Screenshot**
+
+
+**Found in Branch**
+Development (d6c419d)
+
+**Desktop**
+ - Device: PC
+ - OS: Linux
+ - Version: Ubuntu 20.04.3 LTS
+ - CPU AMD Ryzen 5 3600
+ - GPU NVIDIA GeForce RTX 2060 SUPER
+ - Memory 16GB",0,lua editor prev document tab option from the view menu is misspelled describe the bug in the lua editor view menu the prev document tab option is misspelled it should be called previous document tab it should also be noted that none of the other options in the lua editor view menu are shortened steps to reproduce steps to reproduce the behavior go to tools lua editor open the view dropdown menu and check the prev document tab option expected behavior the option is called previous document tab actual behavior the option is called prev document tab screenshot found in branch development desktop device pc os linux version ubuntu lts cpu amd ryzen gpu nvidia geforce rtx super memory ,0
+243,4972614208.0,IssuesEvent,2016-12-05 22:05:03,wordpress-mobile/WordPress-iOS,https://api.github.com/repos/wordpress-mobile/WordPress-iOS,closed,People: Include username or display name in messaging about the user,People Management [Type] Bug,"### Expected behavior
+
+I expect messaging (buttons and errors) in People management to refer to the user I am working with, rather than a generic ""User."" For example, here are error messages from Calypso:
+- User's role can't be changed: ""There was an error updating @[username]""
+- To remove a user: ""Remove [Display Name]""
+### Actual behavior
+
+The messaging in the app refers to the user as ""User"" instead of by username/display name:
+- User's role can't be changed: ""Sorry! Something went wrong while updating the User's Role""
+- To remove a user: ""Remove User""
+
+I'd suggest changing the app messaging to match Calypso.
+### Steps to reproduce the behavior
+1. Set up a site with two users: the site owner and a second account you are using in the app
+2. Open the app (logged in to the second account)
+3. Go to My Site > People
+4. Select the site owner in the list of users
+5. Try to change the user's role. Result: Error message appears ""Sorry! Something went wrong while updating the User's Role""
+6. Look for the option to remove the user from the site. The link is a generic ""Remove User"" instead of referring to the display name of the user being removed.
+##### Tested on iPhone 6, iOS 9.3.2, WPiOS 6.4.0.20160707.1
+",1.0,"People: Include username or display name in messaging about the user - ### Expected behavior
+
+I expect messaging (buttons and errors) in People management to refer to the user I am working with, rather than a generic ""User."" For example, here are error messages from Calypso:
+- User's role can't be changed: ""There was an error updating @[username]""
+- To remove a user: ""Remove [Display Name]""
+### Actual behavior
+
+The messaging in the app refers to the user as ""User"" instead of by username/display name:
+- User's role can't be changed: ""Sorry! Something went wrong while updating the User's Role""
+- To remove a user: ""Remove User""
+
+I'd suggest changing the app messaging to match Calypso.
+### Steps to reproduce the behavior
+1. Set up a site with two users: the site owner and a second account you are using in the app
+2. Open the app (logged in to the second account)
+3. Go to My Site > People
+4. Select the site owner in the list of users
+5. Try to change the user's role. Result: Error message appears ""Sorry! Something went wrong while updating the User's Role""
+6. Look for the option to remove the user from the site. The link is a generic ""Remove User"" instead of referring to the display name of the user being removed.
+##### Tested on iPhone 6, iOS 9.3.2, WPiOS 6.4.0.20160707.1
+",1,people include username or display name in messaging about the user expected behavior i expect messaging buttons and errors in people management to refer to the user i am working with rather than a generic user for example here are error messages from calypso user s role can t be changed there was an error updating to remove a user remove actual behavior the messaging in the app refers to the user as user instead of by username display name user s role can t be changed sorry something went wrong while updating the user s role to remove a user remove user i d suggest changing the app messaging to match calypso steps to reproduce the behavior set up a site with two users the site owner and a second account you are using in the app open the app logged in to the second account go to my site people select the site owner in the list of users try to change the user s role result error message appears sorry something went wrong while updating the user s role look for the option to remove the user from the site the link is a generic remove user instead of referring to the display name of the user being removed tested on iphone ios wpios ,1
+120232,15712637879.0,IssuesEvent,2021-03-27 12:59:53,Lightcaster-Studios/Beamerman,https://api.github.com/repos/Lightcaster-Studios/Beamerman,closed,Laser SFX (Charging) Syncing,Feature.Design Priority.Medium,Laser charging SFX not in sync with the actual firing of the laser.,1.0,Laser SFX (Charging) Syncing - Laser charging SFX not in sync with the actual firing of the laser.,0,laser sfx charging syncing laser charging sfx not in sync with the actual firing of the laser ,0
+53358,22762163330.0,IssuesEvent,2022-07-07 22:28:40,cityofaustin/atd-data-tech,https://api.github.com/repos/cityofaustin/atd-data-tech,closed,Deploy Hasura graphql-engine 2.7.0 to production,Service: Dev Need: 1-Must Have Product: Moped Type: DevOps,"With Hasura's `graphql-engine` having run in staging for an acceptable amount of time and passed any tests required of it, a plan to install this same version in production is required. The process should be similar to the deployment onto staging, and I will propose a punch list of deployment steps in reply to this issue. ",1.0,"Deploy Hasura graphql-engine 2.7.0 to production - With Hasura's `graphql-engine` having run in staging for an acceptable amount of time and passed any tests required of it, a plan to install this same version in production is required. The process should be similar to the deployment onto staging, and I will propose a punch list of deployment steps in reply to this issue. ",0,deploy hasura graphql engine to production with hasura s graphql engine having run in staging for an acceptable amount of time and passed any tests required of it a plan to install this same version in production is required the process should be similar to the deployment onto staging and i will propose a punch list of deployment steps in reply to this issue ,0
+356274,25176153417.0,IssuesEvent,2022-11-11 09:26:19,Brandon-OS/pe,https://api.github.com/repos/Brandon-OS/pe,opened,Wrong step for Find Command in DG,type.DocumentationBug severity.VeryLow,"
+
+Should be find command instead of the add command.
+
+
+
+",1.0,"Wrong step for Find Command in DG - 
+
+Should be find command instead of the add command.
+
+
+
+",0,wrong step for find command in dg should be find command instead of the add command ,0
+8833,3793434912.0,IssuesEvent,2016-03-22 13:57:42,pgmasters/backrest,https://api.github.com/repos/pgmasters/backrest,closed,Backup format regression in v0.90,bug (code) High Priority,"A change to the repository format was accidentally introduced in 0.90 which means the on-disk backup was no longer a valid PostgreSQL cluster when the backup contained tablespaces. This only affected users who directly copied the backups to restore PostgreSQL clusters rather than using the restore command. However, the fix breaks compatibility with older backups that contain tablespaces no matter how they are being restored (pgBackRest will throw errors and refuse to restore). New full backups should be taken immediately after installing version 0.91 for any clusters that contain tablespaces. If older backups need to be restored then use a version of pgBackRest that matches the backup version.",1.0,"Backup format regression in v0.90 - A change to the repository format was accidentally introduced in 0.90 which means the on-disk backup was no longer a valid PostgreSQL cluster when the backup contained tablespaces. This only affected users who directly copied the backups to restore PostgreSQL clusters rather than using the restore command. However, the fix breaks compatibility with older backups that contain tablespaces no matter how they are being restored (pgBackRest will throw errors and refuse to restore). New full backups should be taken immediately after installing version 0.91 for any clusters that contain tablespaces. If older backups need to be restored then use a version of pgBackRest that matches the backup version.",0,backup format regression in a change to the repository format was accidentally introduced in which means the on disk backup was no longer a valid postgresql cluster when the backup contained tablespaces this only affected users who directly copied the backups to restore postgresql clusters rather than using the restore command however the fix breaks compatibility with older backups that contain tablespaces no matter how they are being restored pgbackrest will throw errors and refuse to restore new full backups should be taken immediately after installing version for any clusters that contain tablespaces if older backups need to be restored then use a version of pgbackrest that matches the backup version ,0
+995,23626249694.0,IssuesEvent,2022-08-25 04:25:24,orangemn6/uptime-monitor,https://api.github.com/repos/orangemn6/uptime-monitor,opened,🛑 The Epic Proxy that too many people use is down,status the-epic-proxy-that-too-many-people-use,"In [`4d00536`](https://github.com/orangemn6/uptime-monitor/commit/4d0053601c00d5c459326412ae259238fe96b8af
+), The Epic Proxy that too many people use (https://68656e647279646564.cf/) was **down**:
+- HTTP code: 429
+- Response time: 4493 ms
+",1.0,"🛑 The Epic Proxy that too many people use is down - In [`4d00536`](https://github.com/orangemn6/uptime-monitor/commit/4d0053601c00d5c459326412ae259238fe96b8af
+), The Epic Proxy that too many people use (https://68656e647279646564.cf/) was **down**:
+- HTTP code: 429
+- Response time: 4493 ms
+",1,🛑 the epic proxy that too many people use is down in the epic proxy that too many people use was down http code response time ms ,1
+771,14114562190.0,IssuesEvent,2020-11-07 16:34:53,peopledoc/procrastinate,https://api.github.com/repos/peopledoc/procrastinate,closed,Django migrations: generate on the fly via import hooks,Issue appropriate for: People up for a challenge Issue contains: Some Django Issue contains: Some Python Issue type: Refactor,"Currently, Django migrations are generated once via makemigrations, because I believed they had to be stored on disk, and generation at release time was complicated and generation at migration time was likely to fail if the process that runs migration can’t write to the disk.
+
+@MarkusH came up with a very interesting idea:
+Actually there is one abstraction layer between Django migrations and reading the disk that we can tap into: django uses the Python import mechanizm that exposes all kinds of hooks. If we can generate the migration modules in memory and convince Python that our migrations top-level module contains our in-memory modules, then we’ll be much better off !
+
+Time to experiment !",1.0,"Django migrations: generate on the fly via import hooks - Currently, Django migrations are generated once via makemigrations, because I believed they had to be stored on disk, and generation at release time was complicated and generation at migration time was likely to fail if the process that runs migration can’t write to the disk.
+
+@MarkusH came up with a very interesting idea:
+Actually there is one abstraction layer between Django migrations and reading the disk that we can tap into: django uses the Python import mechanizm that exposes all kinds of hooks. If we can generate the migration modules in memory and convince Python that our migrations top-level module contains our in-memory modules, then we’ll be much better off !
+
+Time to experiment !",1,django migrations generate on the fly via import hooks currently django migrations are generated once via makemigrations because i believed they had to be stored on disk and generation at release time was complicated and generation at migration time was likely to fail if the process that runs migration can’t write to the disk markush came up with a very interesting idea actually there is one abstraction layer between django migrations and reading the disk that we can tap into django uses the python import mechanizm that exposes all kinds of hooks if we can generate the migration modules in memory and convince python that our migrations top level module contains our in memory modules then we’ll be much better off time to experiment ,1
+310,5716512852.0,IssuesEvent,2017-04-19 15:17:20,Princeton-CDH/mapping-expatriate-paris,https://api.github.com/repos/Princeton-CDH/mapping-expatriate-paris,reopened,"""Richard Wallace"" is two people",has unknown dates not in personography two people merged in one card folder,"The ""Richard Wallace"" card is separated into ""Mrs. Richard Wallace"" (s1 and s5) and ""Mr. Richard Wallace"" (s3).
+",1.0,"""Richard Wallace"" is two people - The ""Richard Wallace"" card is separated into ""Mrs. Richard Wallace"" (s1 and s5) and ""Mr. Richard Wallace"" (s3).
+",1, richard wallace is two people the richard wallace card is separated into mrs richard wallace and and mr richard wallace ,1
+362152,10723525891.0,IssuesEvent,2019-10-27 19:22:51,Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth,https://api.github.com/repos/Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth,opened,Cenarion Enclave without Reaper's Due DLC,:beetle: bug :beetle: :grey_exclamation: priority medium,"**Mod Version**
+Master Branch
+
+**What expansions do you have installed?**
+No Reaper's Due
+
+**Please explain your issue in as much detail as possible:**
+You can build Cenarion Enclave without Reaper's Due DLC
+Nothing will happen, tho
+
+**Steps to reproduce the issue:**
+Build Cenarion Enclave without Reaper's Due DLC
+
+**Upload an attachment below: .zip of your save, or screenshots:**
+
+
+Click to expand
+
+
+
+
+",1.0,"Cenarion Enclave without Reaper's Due DLC - **Mod Version**
+Master Branch
+
+**What expansions do you have installed?**
+No Reaper's Due
+
+**Please explain your issue in as much detail as possible:**
+You can build Cenarion Enclave without Reaper's Due DLC
+Nothing will happen, tho
+
+**Steps to reproduce the issue:**
+Build Cenarion Enclave without Reaper's Due DLC
+
+**Upload an attachment below: .zip of your save, or screenshots:**
+
+
+Click to expand
+
+
+
+
+",0,cenarion enclave without reaper s due dlc mod version master branch what expansions do you have installed no reaper s due please explain your issue in as much detail as possible you can build cenarion enclave without reaper s due dlc nothing will happen tho steps to reproduce the issue build cenarion enclave without reaper s due dlc upload an attachment below zip of your save or screenshots click to expand ,0
+127339,17258162115.0,IssuesEvent,2021-07-22 00:58:46,ithaka/pharos,https://api.github.com/repos/ithaka/pharos,opened,Tokens: Sizing,design tokens feature request good first issue,"**The problem**
+
+Some components need to have an explicit height or width defined, and it may be beneficial for those values to be tokenized instead of hard coded.
+
+**The solution**
+
+Creating a new set of ""sizing"" tokens that can be used to size components. An analysis should be included as part of this that determines which things should be included (or not), and what values to settle out to (we don't want a token for every possible size).
+
+**Additional information**
+
+Some folks have been inclined to use the existing spacing tokens to represent a size, but these are really separate purposes—one is about how much room to take up, while the other is how much space to place between things.
+
+",1.0,"Tokens: Sizing - **The problem**
+
+Some components need to have an explicit height or width defined, and it may be beneficial for those values to be tokenized instead of hard coded.
+
+**The solution**
+
+Creating a new set of ""sizing"" tokens that can be used to size components. An analysis should be included as part of this that determines which things should be included (or not), and what values to settle out to (we don't want a token for every possible size).
+
+**Additional information**
+
+Some folks have been inclined to use the existing spacing tokens to represent a size, but these are really separate purposes—one is about how much room to take up, while the other is how much space to place between things.
+
+",0,tokens sizing the problem some components need to have an explicit height or width defined and it may be beneficial for those values to be tokenized instead of hard coded the solution creating a new set of sizing tokens that can be used to size components an analysis should be included as part of this that determines which things should be included or not and what values to settle out to we don t want a token for every possible size additional information some folks have been inclined to use the existing spacing tokens to represent a size but these are really separate purposes—one is about how much room to take up while the other is how much space to place between things ,0
+826,15288040778.0,IssuesEvent,2021-02-23 16:25:23,openstates/issues,https://api.github.com/repos/openstates/issues,closed,Wisconsin Legislator Spot Check Issues,component:people-data type:bug,"State: WI
+Short Description: When checking Wisconsin’s legislators, I found two vacancies and noticed that secondary phone numbers were not captured.
+
+Include these 3 sections as appropriate (or note none)
+
+Missing or Incorrect legislators:
+ (include vacancies, mismatched legislators, etc.)
+The legislative site has two vacancies, but OS shows them as John Nygren (District 89) and Scott L. Fitzgerald (District 13)
+
+
+
+
+Data Issues:
+N/A
+
+Additional Data:
+(note missing contact information, social media, photos, biographical information, etc. with 1-2 examples)
+(no need to mention: campaign information, committees, sponsored bills, other related information)
+Quite a few legislators have a secondary phone number listed:
+Steve Doyle https://docs.legis.wisconsin.gov/2021/legislators/assembly/2133
+Deb A. Andraca https://docs.legis.wisconsin.gov/2021/legislators/assembly/2266",1.0,"Wisconsin Legislator Spot Check Issues - State: WI
+Short Description: When checking Wisconsin’s legislators, I found two vacancies and noticed that secondary phone numbers were not captured.
+
+Include these 3 sections as appropriate (or note none)
+
+Missing or Incorrect legislators:
+ (include vacancies, mismatched legislators, etc.)
+The legislative site has two vacancies, but OS shows them as John Nygren (District 89) and Scott L. Fitzgerald (District 13)
+
+
+
+
+Data Issues:
+N/A
+
+Additional Data:
+(note missing contact information, social media, photos, biographical information, etc. with 1-2 examples)
+(no need to mention: campaign information, committees, sponsored bills, other related information)
+Quite a few legislators have a secondary phone number listed:
+Steve Doyle https://docs.legis.wisconsin.gov/2021/legislators/assembly/2133
+Deb A. Andraca https://docs.legis.wisconsin.gov/2021/legislators/assembly/2266",1,wisconsin legislator spot check issues state wi short description when checking wisconsin’s legislators i found two vacancies and noticed that secondary phone numbers were not captured include these sections as appropriate or note none missing or incorrect legislators include vacancies mismatched legislators etc the legislative site has two vacancies but os shows them as john nygren district and scott l fitzgerald district data issues n a additional data note missing contact information social media photos biographical information etc with examples no need to mention campaign information committees sponsored bills other related information quite a few legislators have a secondary phone number listed steve doyle deb a andraca ,1
+989,23437208679.0,IssuesEvent,2022-08-15 11:15:49,metabase/metabase,https://api.github.com/repos/metabase/metabase,reopened,Sandboxed group managers can't see other users in the People tab,Type:Bug Priority:P2 Administration/Permissions .Backend .Reproduced .Regression Administration/Data Sandboxes Administration/People,"**Describe the bug**
+A group manager that has sandboxed access to any database can't see other users in the People tab in the Admin panel.
+
+As @calherries found:
+
+One strange thing I noticed with the query is that if a group manager has sandboxed permissions, they can't see any other users in the group. Perhaps we should make a separate issue for this?
+
+Below is an example query that runs to load users in a group on this page. Note the clause core_user.id = 272. That restriction is included in the query if the user has any sandboxed permissions on their account.
+
+```
+SELECT ""core_user"".""id"", ""core_user"".""email"", ""core_user"".""first_name"", ""core_user"".""last_name"", ""core_user"".""is_superuser"", ""core_user"".""last_login""
+FROM ""core_user"" RIGHT JOIN ""permissions_group_membership"" ON ""core_user"".""id"" = ""permissions_group_membership"".""user_id""
+WHERE (""is_active"" = TRUE AND ""core_user"".""id"" = 272 AND ""permissions_group_membership"".""group_id"" = 162)
+ORDER BY ""core_user"".""is_superuser"" DESC, ""is_group_manager"" DESC, lower(""last_name"") ASC, lower(""first_name"")
+```
+
+**To Reproduce**
+Steps to reproduce the behavior:
+1. As an admin, go to the People tab, create a group, add at least 2 users, promote one of them to group manager and restrict their access to any database to a sandbox.
+2. Log in as the group manager and access the people tab. You can't see the other group members besides yourself.
+
+**Expected behavior**
+The group manager should see every other instance user in the People/People tab and the users in their group in the People/Groups tab.
+
+**Severity**
+This makes group managers useless for sandboxed managers. But there shouldn't be a lot of sandboxed group managers.
+
+**Additional context**
+Related to #23246
+
+Not sure why we are restricting the people tab in the admin panel for sandboxed users, it may be wise to investigate why we are doing it.",1.0,"Sandboxed group managers can't see other users in the People tab - **Describe the bug**
+A group manager that has sandboxed access to any database can't see other users in the People tab in the Admin panel.
+
+As @calherries found:
+
+One strange thing I noticed with the query is that if a group manager has sandboxed permissions, they can't see any other users in the group. Perhaps we should make a separate issue for this?
+
+Below is an example query that runs to load users in a group on this page. Note the clause core_user.id = 272. That restriction is included in the query if the user has any sandboxed permissions on their account.
+
+```
+SELECT ""core_user"".""id"", ""core_user"".""email"", ""core_user"".""first_name"", ""core_user"".""last_name"", ""core_user"".""is_superuser"", ""core_user"".""last_login""
+FROM ""core_user"" RIGHT JOIN ""permissions_group_membership"" ON ""core_user"".""id"" = ""permissions_group_membership"".""user_id""
+WHERE (""is_active"" = TRUE AND ""core_user"".""id"" = 272 AND ""permissions_group_membership"".""group_id"" = 162)
+ORDER BY ""core_user"".""is_superuser"" DESC, ""is_group_manager"" DESC, lower(""last_name"") ASC, lower(""first_name"")
+```
+
+**To Reproduce**
+Steps to reproduce the behavior:
+1. As an admin, go to the People tab, create a group, add at least 2 users, promote one of them to group manager and restrict their access to any database to a sandbox.
+2. Log in as the group manager and access the people tab. You can't see the other group members besides yourself.
+
+**Expected behavior**
+The group manager should see every other instance user in the People/People tab and the users in their group in the People/Groups tab.
+
+**Severity**
+This makes group managers useless for sandboxed managers. But there shouldn't be a lot of sandboxed group managers.
+
+**Additional context**
+Related to #23246
+
+Not sure why we are restricting the people tab in the admin panel for sandboxed users, it may be wise to investigate why we are doing it.",1,sandboxed group managers can t see other users in the people tab describe the bug a group manager that has sandboxed access to any database can t see other users in the people tab in the admin panel as calherries found one strange thing i noticed with the query is that if a group manager has sandboxed permissions they can t see any other users in the group perhaps we should make a separate issue for this below is an example query that runs to load users in a group on this page note the clause core user id that restriction is included in the query if the user has any sandboxed permissions on their account select core user id core user email core user first name core user last name core user is superuser core user last login from core user right join permissions group membership on core user id permissions group membership user id where is active true and core user id and permissions group membership group id order by core user is superuser desc is group manager desc lower last name asc lower first name to reproduce steps to reproduce the behavior as an admin go to the people tab create a group add at least users promote one of them to group manager and restrict their access to any database to a sandbox log in as the group manager and access the people tab you can t see the other group members besides yourself expected behavior the group manager should see every other instance user in the people people tab and the users in their group in the people groups tab severity this makes group managers useless for sandboxed managers but there shouldn t be a lot of sandboxed group managers additional context related to not sure why we are restricting the people tab in the admin panel for sandboxed users it may be wise to investigate why we are doing it ,1
+291,5558214150.0,IssuesEvent,2017-03-24 14:14:19,fifafu/BetterTouchTool,https://api.github.com/repos/fifafu/BetterTouchTool,closed,Move window to next monitor doesn't work,bug Need other people who can reproduce the issue,"I have the following configured as global shortcuts.
+
+
+
+It doesn't work. I am currently using the Sierra public beta, but it hasn't worked in a very long time. It used to work a very long time ago (I couldn't say exactly when, but it's been a few OS X versions), but it doesn't now. ",1.0,"Move window to next monitor doesn't work - I have the following configured as global shortcuts.
+
+
+
+It doesn't work. I am currently using the Sierra public beta, but it hasn't worked in a very long time. It used to work a very long time ago (I couldn't say exactly when, but it's been a few OS X versions), but it doesn't now. ",1,move window to next monitor doesn t work i have the following configured as global shortcuts img width alt screenshot src it doesn t work i am currently using the sierra public beta but it hasn t worked in a very long time it used to work a very long time ago i couldn t say exactly when but it s been a few os x versions but it doesn t now ,1
+213,4423930731.0,IssuesEvent,2016-08-16 10:26:22,wordpress-mobile/WordPress-Android,https://api.github.com/repos/wordpress-mobile/WordPress-Android,opened,Move roles string array to `strings.xml`,i18n People Management,"We have a `roles.xml` string resource with `translatable=""true""` flag, however the `strings.xml` file is the only resource that's translated. Also, more importantly, once the roles values are translated it can not be used as REST API parameters.",1.0,"Move roles string array to `strings.xml` - We have a `roles.xml` string resource with `translatable=""true""` flag, however the `strings.xml` file is the only resource that's translated. Also, more importantly, once the roles values are translated it can not be used as REST API parameters.",1,move roles string array to strings xml we have a roles xml string resource with translatable true flag however the strings xml file is the only resource that s translated also more importantly once the roles values are translated it can not be used as rest api parameters ,1
+77421,14816419587.0,IssuesEvent,2021-01-14 09:00:24,jitsucom/eventnative,https://api.github.com/repos/jitsucom/eventnative,opened,Support new mapping format in ENManager,Feature code backend,"### Problem
+
+Since EventNative has a new mapping rules style, ENHelper should use it while generating config.
+Read more about [new mapping rules style](https://docs.eventnative.org/configuration-1/configuration/schema-and-mappings).
+
+### Solution
+
+Update config_mapper.go
+",1.0,"Support new mapping format in ENManager - ### Problem
+
+Since EventNative has a new mapping rules style, ENHelper should use it while generating config.
+Read more about [new mapping rules style](https://docs.eventnative.org/configuration-1/configuration/schema-and-mappings).
+
+### Solution
+
+Update config_mapper.go
+",0,support new mapping format in enmanager problem since eventnative has a new mapping rules style enhelper should use it while generating config read more about solution update config mapper go ,0
+369,6877610393.0,IssuesEvent,2017-11-20 08:47:22,PopulateTools/gobierto,https://api.github.com/repos/PopulateTools/gobierto,opened,"Missing admin_preview_token in ""view event"" links",bug gobierto-people,"I got [this exception](https://rollbar.com/Populate/gobierto/items/476/) after clicking on the ""view event"" link in the admin of [this event]((http://madrid.gobify.net/agendas/rafa-garcia/2017-11-22-nuevo-evento-24h-format)).
+
+It happend because the person was drafted and the `preview token` wasn't being included in the generated link.",1.0,"Missing admin_preview_token in ""view event"" links - I got [this exception](https://rollbar.com/Populate/gobierto/items/476/) after clicking on the ""view event"" link in the admin of [this event]((http://madrid.gobify.net/agendas/rafa-garcia/2017-11-22-nuevo-evento-24h-format)).
+
+It happend because the person was drafted and the `preview token` wasn't being included in the generated link.",1,missing admin preview token in view event links i got after clicking on the view event link in the admin of it happend because the person was drafted and the preview token wasn t being included in the generated link ,1
+244813,26476068993.0,IssuesEvent,2023-01-17 11:13:03,BogdanOrg/WebGoat,https://api.github.com/repos/BogdanOrg/WebGoat,closed,"CVE-2018-14042 (Medium) detected in bootstrap-3.3.7.jar, bootstrap-3.1.1.min.js - autoclosed",security vulnerability,"## CVE-2018-14042 - Medium Severity Vulnerability
+ Vulnerable Libraries - bootstrap-3.3.7.jar, bootstrap-3.1.1.min.js
+
+For more information on CVSS3 Scores, click here.
+
+
+
+ Suggested Fix
+
+
+
Type: Upgrade version
+
Release Date: 2018-07-13
+
Fix Resolution: 3.4.0
+
+
+
+
+
+***
+
+- [ ] Check this box to open an automated fix PR
+
+",0,cve medium detected in bootstrap jar bootstrap min js autoclosed cve medium severity vulnerability vulnerable libraries bootstrap jar bootstrap min js bootstrap jar webjar for bootstrap library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository org webjars bootstrap bootstrap jar dependency hierarchy x bootstrap jar vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to vulnerable library src main resources webgoat static js libs bootstrap min js src main resources lessons challenges js bootstrap min js dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch develop vulnerability details in bootstrap before xss is possible in the data container property of tooltip publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution check this box to open an automated fix pr ,0
+682036,23330158637.0,IssuesEvent,2022-08-09 03:48:36,City-Bureau/city-scrapers-atl,https://api.github.com/repos/City-Bureau/city-scrapers-atl,closed,New Scraper: Clayton County MARTA Citizens Advisory Board,priority-low,"Create a new scraper for Clayton County MARTA Citizens Advisory Board
+
+Website: https://www.claytoncountyga.gov/government/board-of-commissioners/county-boards/clayton-marta-citizens-advisory-board
+Jurisdiction: Clayton County
+Classification: Transportation
+The Clayton MARTA Citizens Advisory Board will act as a liaison between the Clayton County Board of Commissioners and MARTA; and provide input in reference to community transit needs, function as a sounding board for project teams, provide feedback on study direction and analysis and provide an outside opinion on proposed funding, operation and marketing. (Created by Resolution 2021-40) (Each member of Clayton County Board of Commissioners appoints 2 members to this 15 member Board). (Resolution 2021-40)
+
+
+",1.0,"New Scraper: Clayton County MARTA Citizens Advisory Board - Create a new scraper for Clayton County MARTA Citizens Advisory Board
+
+Website: https://www.claytoncountyga.gov/government/board-of-commissioners/county-boards/clayton-marta-citizens-advisory-board
+Jurisdiction: Clayton County
+Classification: Transportation
+The Clayton MARTA Citizens Advisory Board will act as a liaison between the Clayton County Board of Commissioners and MARTA; and provide input in reference to community transit needs, function as a sounding board for project teams, provide feedback on study direction and analysis and provide an outside opinion on proposed funding, operation and marketing. (Created by Resolution 2021-40) (Each member of Clayton County Board of Commissioners appoints 2 members to this 15 member Board). (Resolution 2021-40)
+
+
+",0,new scraper clayton county marta citizens advisory board create a new scraper for clayton county marta citizens advisory board website jurisdiction clayton county classification transportation the clayton marta citizens advisory board will act as a liaison between the clayton county board of commissioners and marta and provide input in reference to community transit needs function as a sounding board for project teams provide feedback on study direction and analysis and provide an outside opinion on proposed funding operation and marketing created by resolution each member of clayton county board of commissioners appoints members to this member board resolution ,0
+664915,22292450441.0,IssuesEvent,2022-06-12 15:08:56,kir-dev/cmsch,https://api.github.com/repos/kir-dev/cmsch,opened,Component: Signup component,FRONTEND priority: high,"- We should discuss the details in a meetings
+- depends on: #272 , #271 , #273 ",1.0,"Component: Signup component - - We should discuss the details in a meetings
+- depends on: #272 , #271 , #273 ",0,component signup component we should discuss the details in a meetings depends on ,0
+655,11660562102.0,IssuesEvent,2020-03-03 03:46:14,microsoftgraph/microsoft-graph-toolkit,https://api.github.com/repos/microsoftgraph/microsoft-graph-toolkit,opened,[BUG] mgt-people does not update when user-ids changes,Area: Components Component: people Priority: 0 State: Committed bug,"
+
+
+**Describe the bug**
+[Originally raised on StackOverflow](https://stackoverflow.com/questions/60389563/mgt-people-doesnt-refresh-in-child-component-when-updating-user-ids-from-parent?noredirect=1#comment106857165_60389563)
+
+Changing the `user-ids` attribute (or the `userIds` property) on mgt-people has no effect on the component.
+
+**To Reproduce**
+Steps to reproduce the behavior:
+
+```html
+
+
+
+
+```
+
+**Expected behavior**
+When the `user-ids` attribute is changed, the people component should update and render the appropriate people (if valid user-id)
+
+",1.0,"[BUG] mgt-people does not update when user-ids changes -
+
+
+**Describe the bug**
+[Originally raised on StackOverflow](https://stackoverflow.com/questions/60389563/mgt-people-doesnt-refresh-in-child-component-when-updating-user-ids-from-parent?noredirect=1#comment106857165_60389563)
+
+Changing the `user-ids` attribute (or the `userIds` property) on mgt-people has no effect on the component.
+
+**To Reproduce**
+Steps to reproduce the behavior:
+
+```html
+
+
+
+
+```
+
+**Expected behavior**
+When the `user-ids` attribute is changed, the people component should update and render the appropriate people (if valid user-id)
+
+",1, mgt people does not update when user ids changes describe the bug changing the user ids attribute or the userids property on mgt people has no effect on the component to reproduce steps to reproduce the behavior html let picker document getelementbyid picker let people document getelementbyid people picker addeventlistener selectionchanged e people setattribute user ids picker selectedpeople map p p id expected behavior when the user ids attribute is changed the people component should update and render the appropriate people if valid user id ,1
+351610,25033773794.0,IssuesEvent,2022-11-04 14:27:39,NethermindEth/juno,https://api.github.com/repos/NethermindEth/juno,closed,Make staging environment for docs,documentation devops,"RIght now, you can push changes to docs and break them in main.",1.0,"Make staging environment for docs - RIght now, you can push changes to docs and break them in main.",0,make staging environment for docs right now you can push changes to docs and break them in main ,0
+318667,9695962473.0,IssuesEvent,2019-05-25 02:28:05,etternagame/etterna,https://api.github.com/repos/etternagame/etterna,closed,Changing devices (plugging or removing) causes game stutter,Priority: Medium Type: Maintenance,"This is due to ""Polling for Joystick Changes"" usually, which lasts for 60 seconds.
+Something should be done to either stop this or subdue the stuttering.
+
+Relevant section of code (good old 10+ year old code): https://github.com/etternagame/etterna/blob/develop/src/arch/InputHandler/InputHandler_DirectInput.cpp#L841
+
+Possibly related: #484 ",1.0,"Changing devices (plugging or removing) causes game stutter - This is due to ""Polling for Joystick Changes"" usually, which lasts for 60 seconds.
+Something should be done to either stop this or subdue the stuttering.
+
+Relevant section of code (good old 10+ year old code): https://github.com/etternagame/etterna/blob/develop/src/arch/InputHandler/InputHandler_DirectInput.cpp#L841
+
+Possibly related: #484 ",0,changing devices plugging or removing causes game stutter this is due to polling for joystick changes usually which lasts for seconds something should be done to either stop this or subdue the stuttering relevant section of code good old year old code possibly related ,0
+665,11884954505.0,IssuesEvent,2020-03-27 18:37:58,microsoftgraph/microsoft-graph-toolkit,https://api.github.com/repos/microsoftgraph/microsoft-graph-toolkit,closed,[BUG] people-picker flyout in mgt-tasks opens of screen,Component: people-picker Component: tasks Priority: 0 State: In Review bug,"
+
+
+**Describe the bug**
+When using the people picker inside the mgt-tasks component, when starting to type, the flyout opens to the right of screen.
+
+**To Reproduce**
+Steps to reproduce the behavior:
+1. Add mgt-tasks to the page and make sure it is set to planner
+2. On a task, click on the assigned people and start typing to assign more people
+3. Notice the suggested people flyout will not open under the people picker, but instead open off screen
+
+**Expected behavior**
+The people picker should open under (or over) the people picker
+
+**Screenshots**
+
+
+
+**Environment (please complete the following information):**
+ - OS: [e.g. iOS] Windows
+ - Browser [e.g. edge, chrome, safari] new Edge
+ - Framework [e.g. react, vue, none] none - also validated in Storybook
+ - Context [e.g. Microsoft Teams, SharePoint, Office Add-ins, Web]: web
+ - Version [e.g. 0.1]: latest
+",1.0,"[BUG] people-picker flyout in mgt-tasks opens of screen -
+
+
+**Describe the bug**
+When using the people picker inside the mgt-tasks component, when starting to type, the flyout opens to the right of screen.
+
+**To Reproduce**
+Steps to reproduce the behavior:
+1. Add mgt-tasks to the page and make sure it is set to planner
+2. On a task, click on the assigned people and start typing to assign more people
+3. Notice the suggested people flyout will not open under the people picker, but instead open off screen
+
+**Expected behavior**
+The people picker should open under (or over) the people picker
+
+**Screenshots**
+
+
+
+**Environment (please complete the following information):**
+ - OS: [e.g. iOS] Windows
+ - Browser [e.g. edge, chrome, safari] new Edge
+ - Framework [e.g. react, vue, none] none - also validated in Storybook
+ - Context [e.g. Microsoft Teams, SharePoint, Office Add-ins, Web]: web
+ - Version [e.g. 0.1]: latest
+",1, people picker flyout in mgt tasks opens of screen describe the bug when using the people picker inside the mgt tasks component when starting to type the flyout opens to the right of screen to reproduce steps to reproduce the behavior add mgt tasks to the page and make sure it is set to planner on a task click on the assigned people and start typing to assign more people notice the suggested people flyout will not open under the people picker but instead open off screen expected behavior the people picker should open under or over the people picker screenshots environment please complete the following information os windows browser new edge framework none also validated in storybook context web version latest ,1
+781,14405899338.0,IssuesEvent,2020-12-03 19:24:25,mozilla-services/updatebot,https://api.github.com/repos/mozilla-services/updatebot,closed,Get the verify-updatebot added to the mercurial update process,Easy requires-talking-to-people,"(Mozilla bug, filing for tracking.)",1.0,"Get the verify-updatebot added to the mercurial update process - (Mozilla bug, filing for tracking.)",1,get the verify updatebot added to the mercurial update process mozilla bug filing for tracking ,1
+78861,15086993685.0,IssuesEvent,2021-02-05 21:18:16,eneko/Blog,https://api.github.com/repos/eneko/Blog,opened,Dealing with merge conflicts in Xcode project.pbxproj file,Xcode,"## What is a merge conflict?
+
+Merge conflicts can be pretty scary. It is relatively easy to lose a good amount of work, if the merge is resolved without care.
+
+```diff
+<<<<<<< HEAD
+[...]
+=======
+[...]
+>>>>>>> origin/develop
+```
+
+## How to resolve merge conflicts
+
+When merging _other_ branch into _our_ branch, we can solve each individual conflict in many different ways:
+
+1. Accept incoming changes from the _other_ branch
+2. Accept changes from _our_ branch_ (essentially, preserving our version)
+3. Accept both changes
+4. Reject both changes
+5. Replace with something else
+
+## Resolving merge conflicts in Xcode `project.pbxproj` file
+
+### Files added in both branches
+
+Git often complains when files are added to an Xcode project in two different branches, and it will indicate this as a merge conflict.
+
+```diff
+<<<<<<< HEAD
+ 9CD4B56C24F06F0E00331CA4 /* PushNotificationsRegistrant.swift in Sources */ = {isa = PBXBuildFile; fileRef = 9CD4B56B24F06F0E00331CA4 /* PushNotificationsRegistrant.swift */; };
+=======
+ 9CD1F3ED25A38652008C6E92 /* ShareFeedbackLogger.swift in Sources */ = {isa = PBXBuildFile; fileRef = 9CD1F3EC25A38652008C6E92 /* ShareFeedbackLogger.swift */; };
+ 9CD1F3FA25A3867A008C6E92 /* FeatureLoggerType.swift in Sources */ = {isa = PBXBuildFile; fileRef = 9CD1F3F925A3867A008C6E92 /* FeatureLoggerType.swift */; };
+>>>>>>> origin/develop
+```
+
+The resolution in this case is pretty easy: accept both changes.
+
+```diff
+ 9CD4B56C24F06F0E00331CA4 /* PushNotificationsRegistrant.swift in Sources */ = {isa = PBXBuildFile; fileRef = 9CD4B56B24F06F0E00331CA4 /* PushNotificationsRegistrant.swift */; };
+ 9CD1F3ED25A38652008C6E92 /* ShareFeedbackLogger.swift in Sources */ = {isa = PBXBuildFile; fileRef = 9CD1F3EC25A38652008C6E92 /* ShareFeedbackLogger.swift */; };
+ 9CD1F3FA25A3867A008C6E92 /* FeatureLoggerType.swift in Sources */ = {isa = PBXBuildFile; fileRef = 9CD1F3F925A3867A008C6E92 /* FeatureLoggerType.swift */; };
+```
+",1.0,"Dealing with merge conflicts in Xcode project.pbxproj file - ## What is a merge conflict?
+
+Merge conflicts can be pretty scary. It is relatively easy to lose a good amount of work, if the merge is resolved without care.
+
+```diff
+<<<<<<< HEAD
+[...]
+=======
+[...]
+>>>>>>> origin/develop
+```
+
+## How to resolve merge conflicts
+
+When merging _other_ branch into _our_ branch, we can solve each individual conflict in many different ways:
+
+1. Accept incoming changes from the _other_ branch
+2. Accept changes from _our_ branch_ (essentially, preserving our version)
+3. Accept both changes
+4. Reject both changes
+5. Replace with something else
+
+## Resolving merge conflicts in Xcode `project.pbxproj` file
+
+### Files added in both branches
+
+Git often complains when files are added to an Xcode project in two different branches, and it will indicate this as a merge conflict.
+
+```diff
+<<<<<<< HEAD
+ 9CD4B56C24F06F0E00331CA4 /* PushNotificationsRegistrant.swift in Sources */ = {isa = PBXBuildFile; fileRef = 9CD4B56B24F06F0E00331CA4 /* PushNotificationsRegistrant.swift */; };
+=======
+ 9CD1F3ED25A38652008C6E92 /* ShareFeedbackLogger.swift in Sources */ = {isa = PBXBuildFile; fileRef = 9CD1F3EC25A38652008C6E92 /* ShareFeedbackLogger.swift */; };
+ 9CD1F3FA25A3867A008C6E92 /* FeatureLoggerType.swift in Sources */ = {isa = PBXBuildFile; fileRef = 9CD1F3F925A3867A008C6E92 /* FeatureLoggerType.swift */; };
+>>>>>>> origin/develop
+```
+
+The resolution in this case is pretty easy: accept both changes.
+
+```diff
+ 9CD4B56C24F06F0E00331CA4 /* PushNotificationsRegistrant.swift in Sources */ = {isa = PBXBuildFile; fileRef = 9CD4B56B24F06F0E00331CA4 /* PushNotificationsRegistrant.swift */; };
+ 9CD1F3ED25A38652008C6E92 /* ShareFeedbackLogger.swift in Sources */ = {isa = PBXBuildFile; fileRef = 9CD1F3EC25A38652008C6E92 /* ShareFeedbackLogger.swift */; };
+ 9CD1F3FA25A3867A008C6E92 /* FeatureLoggerType.swift in Sources */ = {isa = PBXBuildFile; fileRef = 9CD1F3F925A3867A008C6E92 /* FeatureLoggerType.swift */; };
+```
+",0,dealing with merge conflicts in xcode project pbxproj file what is a merge conflict merge conflicts can be pretty scary it is relatively easy to lose a good amount of work if the merge is resolved without care diff head origin develop how to resolve merge conflicts when merging other branch into our branch we can solve each individual conflict in many different ways accept incoming changes from the other branch accept changes from our branch essentially preserving our version accept both changes reject both changes replace with something else resolving merge conflicts in xcode project pbxproj file files added in both branches git often complains when files are added to an xcode project in two different branches and it will indicate this as a merge conflict diff head pushnotificationsregistrant swift in sources isa pbxbuildfile fileref pushnotificationsregistrant swift sharefeedbacklogger swift in sources isa pbxbuildfile fileref sharefeedbacklogger swift featureloggertype swift in sources isa pbxbuildfile fileref featureloggertype swift origin develop the resolution in this case is pretty easy accept both changes diff pushnotificationsregistrant swift in sources isa pbxbuildfile fileref pushnotificationsregistrant swift sharefeedbacklogger swift in sources isa pbxbuildfile fileref sharefeedbacklogger swift featureloggertype swift in sources isa pbxbuildfile fileref featureloggertype swift ,0
+994,23598079288.0,IssuesEvent,2022-08-23 21:25:54,omigroup/omigroup,https://api.github.com/repos/omigroup/omigroup,closed,Mapping the Metaverse - a Cross-Industry Leadership from OMI,proposal Make the metaverse more human Empower the people making the metaverse Cultivate Resiliency Call for Participation,"# Metaverse Mapping
+
+## Contributors
+
+- Evo, Co-Chair
+- Article by Doug Thompson: https://outofscope.bureauofbrightideas.com/a-map-of-the-metaverse/
+
+## Summary
+
+Dynamic map of the metaverse industry with a clear assessment toward interoperability
+
+## Example Use Cases
+
+Dynamic maps exist of other fields, such as the cloud industry:
+
+## Implementation
+
+1. Begin conceptualizing how to pull 2D field maps into a 3D industry map (needs assessment, iconography, how to map?)
+2. Address best use of dynamic database/dataviz tools native for 3D web
+3. Set dates to hack together a mapping process (proposed November 12-Thansgiving week)
+4. Let the dust settle, fix issues, edit and bring to community (Dec 1-5)
+5. Open for public participation, include public engagement as part of OMI end of year community membership drive
+",1.0,"Mapping the Metaverse - a Cross-Industry Leadership from OMI - # Metaverse Mapping
+
+## Contributors
+
+- Evo, Co-Chair
+- Article by Doug Thompson: https://outofscope.bureauofbrightideas.com/a-map-of-the-metaverse/
+
+## Summary
+
+Dynamic map of the metaverse industry with a clear assessment toward interoperability
+
+## Example Use Cases
+
+Dynamic maps exist of other fields, such as the cloud industry:
+
+## Implementation
+
+1. Begin conceptualizing how to pull 2D field maps into a 3D industry map (needs assessment, iconography, how to map?)
+2. Address best use of dynamic database/dataviz tools native for 3D web
+3. Set dates to hack together a mapping process (proposed November 12-Thansgiving week)
+4. Let the dust settle, fix issues, edit and bring to community (Dec 1-5)
+5. Open for public participation, include public engagement as part of OMI end of year community membership drive
+",1,mapping the metaverse a cross industry leadership from omi metaverse mapping contributors evo co chair article by doug thompson summary dynamic map of the metaverse industry with a clear assessment toward interoperability example use cases dynamic maps exist of other fields such as the cloud industry implementation begin conceptualizing how to pull field maps into a industry map needs assessment iconography how to map address best use of dynamic database dataviz tools native for web set dates to hack together a mapping process proposed november thansgiving week let the dust settle fix issues edit and bring to community dec open for public participation include public engagement as part of omi end of year community membership drive ,1
+69600,13298449727.0,IssuesEvent,2020-08-25 08:17:03,MeAmAnUsername/pie,https://api.github.com/repos/MeAmAnUsername/pie,opened,Generated code layout,Component: code generation Priority: low Status: specified Type: bug,"Generated code currently has a weird layout where blocks are indented to the curly brace, e.g.
+```
+if (true) {
+ return 5;
+} else {
+ return 7;
+}
+```
+The generated code should look like handwritten code, e.g. tab size 2 or 4, indent one tab per level.",1.0,"Generated code layout - Generated code currently has a weird layout where blocks are indented to the curly brace, e.g.
+```
+if (true) {
+ return 5;
+} else {
+ return 7;
+}
+```
+The generated code should look like handwritten code, e.g. tab size 2 or 4, indent one tab per level.",0,generated code layout generated code currently has a weird layout where blocks are indented to the curly brace e g if true return else return the generated code should look like handwritten code e g tab size or indent one tab per level ,0
+172,3796488300.0,IssuesEvent,2016-03-23 00:48:41,elmsln/elmsln,https://api.github.com/repos/elmsln/elmsln,closed,Adding Students to Offerings/Classes,duplicate people / cpr,"I have been unable to find out how to attach a student to a course. If I'm not mistaken the finest resolution of a course is an offering. This would be the scheduled class complete with instructor and allocated resources. I have added a person to the system with the role of student but I have been unable to associate this ""person"" with an offering.",1.0,"Adding Students to Offerings/Classes - I have been unable to find out how to attach a student to a course. If I'm not mistaken the finest resolution of a course is an offering. This would be the scheduled class complete with instructor and allocated resources. I have added a person to the system with the role of student but I have been unable to associate this ""person"" with an offering.",1,adding students to offerings classes i have been unable to find out how to attach a student to a course if i m not mistaken the finest resolution of a course is an offering this would be the scheduled class complete with instructor and allocated resources i have added a person to the system with the role of student but i have been unable to associate this person with an offering ,1
+6,2603742365.0,IssuesEvent,2015-02-24 17:41:32,fskreuz/Tree-Trail,https://api.github.com/repos/fskreuz/Tree-Trail,closed,[back-end] Create RESTful API,QA people! Test these! ToDo,"# Acceptance Criteria
+
+As a front-end developer, I should be able to use JavaScript to call the server using an API provided by the server. This API should allow me to manipulate data without the need for the page to reload. This is primarily the case for the maps.
+
+- [ ] Must adhere to entity specifications as stated in the wiki: https://github.com/fskreuz/Tree-Trail/wiki
+- [ ] Endpoints (Urls), payloads and returns must be documented in the wiki
+- [ ] May accept `Content-Type` of `application/json` where needed
+- [ ] Return values must be in JSON format, and supplied the correct headers
+- [ ] Use HTTP status codes where appropriate. https://en.wikipedia.org/wiki/List_of_HTTP_status_codes
+
+---
+
+Reading materials and tools:
+
+- https://github.com/chriskacerguis/codeigniter-restserver",1.0,"[back-end] Create RESTful API - # Acceptance Criteria
+
+As a front-end developer, I should be able to use JavaScript to call the server using an API provided by the server. This API should allow me to manipulate data without the need for the page to reload. This is primarily the case for the maps.
+
+- [ ] Must adhere to entity specifications as stated in the wiki: https://github.com/fskreuz/Tree-Trail/wiki
+- [ ] Endpoints (Urls), payloads and returns must be documented in the wiki
+- [ ] May accept `Content-Type` of `application/json` where needed
+- [ ] Return values must be in JSON format, and supplied the correct headers
+- [ ] Use HTTP status codes where appropriate. https://en.wikipedia.org/wiki/List_of_HTTP_status_codes
+
+---
+
+Reading materials and tools:
+
+- https://github.com/chriskacerguis/codeigniter-restserver",1, create restful api acceptance criteria as a front end developer i should be able to use javascript to call the server using an api provided by the server this api should allow me to manipulate data without the need for the page to reload this is primarily the case for the maps must adhere to entity specifications as stated in the wiki endpoints urls payloads and returns must be documented in the wiki may accept content type of application json where needed return values must be in json format and supplied the correct headers use http status codes where appropriate reading materials and tools ,1
+192373,22215940326.0,IssuesEvent,2022-06-08 01:39:18,panasalap/linux-4.1.15,https://api.github.com/repos/panasalap/linux-4.1.15,reopened,CVE-2021-3348 (High) detected in linux179e72b561d3d331c850e1a5779688d7a7de5246,security vulnerability,"## CVE-2021-3348 - High Severity Vulnerability
+ Vulnerable Library - linux179e72b561d3d331c850e1a5779688d7a7de5246
+
+
+nbd_add_socket in drivers/block/nbd.c in the Linux kernel through 5.10.12 has an ndb_queue_rq use-after-free that could be triggered by local attackers (with access to the nbd device) via an I/O request at a certain point during device setup, aka CID-b98e762e3d71.
+
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",True,"CVE-2021-3348 (High) detected in linux179e72b561d3d331c850e1a5779688d7a7de5246 - ## CVE-2021-3348 - High Severity Vulnerability
+ Vulnerable Library - linux179e72b561d3d331c850e1a5779688d7a7de5246
+
+
+nbd_add_socket in drivers/block/nbd.c in the Linux kernel through 5.10.12 has an ndb_queue_rq use-after-free that could be triggered by local attackers (with access to the nbd device) via an I/O request at a certain point during device setup, aka CID-b98e762e3d71.
+
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",0,cve high detected in cve high severity vulnerability vulnerable library linux kernel stable tree mirror library home page a href found in head commit a href found in base branch master vulnerable source files drivers block nbd c drivers block nbd c vulnerability details nbd add socket in drivers block nbd c in the linux kernel through has an ndb queue rq use after free that could be triggered by local attackers with access to the nbd device via an i o request at a certain point during device setup aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource ,0
+32686,8921078995.0,IssuesEvent,2019-01-21 09:08:32,neovim/neovim,https://api.github.com/repos/neovim/neovim,closed,"(clang >= 6.0 bug) Annoying warnings for isnan(), fpclassify(), et al",blocked:external build help wanted,"- `nvim --version`: ef4feab0e75be
+- Vim (version: 8.0.1565) behaves differently? No warnings, but I didn't check if the same functions were used
+- Operating system/version: arch linux
+- Terminal name/version: pangoterm
+- `$TERM`: xterm
+
+### Steps to reproduce using `nvim -u NORC`
+```
+rm -rf build && CMAKE_EXTRA_FLAGS=""-DCMAKE_C_COMPILER=clang -DCLANG_ASAN_UBSAN=1"" make -j4
+```
+
+### Actual behaviour
+```
+[197/284] Building C object src/nvim/CMakeFiles/nvim.dir/eval/encode.c.o
+In file included from ../src/nvim/eval/encode.c:455:
+../src/nvim/eval/typval_encode.c.h:330:7: warning: implicit conversion loses floating-point precision: 'const float_T' (aka 'const double') to 'float' [-Wconv
+ersion]
+ TYPVAL_ENCODE_CONV_FLOAT(tv, tv->vval.v_float);
+ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+../src/nvim/eval/encode.c:330:26: note: expanded from macro 'TYPVAL_ENCODE_CONV_FLOAT'
+ switch (fpclassify(flt_)) { \
+ ~~~~~~~~~~~^~~~~
+/usr/include/math.h:415:56: note: expanded from macro 'fpclassify'
+# define fpclassify(x) __MATH_TG ((x), __fpclassify, (x))
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~
+/usr/include/math.h:370:16: note: expanded from macro '__MATH_TG'
+ ? FUNC ## f ARGS \
+ ~~~~~~~~~ ^~~~
+In file included from ../src/nvim/eval/encode.c:455:
+../src/nvim/eval/typval_encode.c.h:491:13: warning: implicit conversion loses floating-point precision: 'const float_T' (aka 'const double') to 'float' [-Wcon
+version]
+ TYPVAL_ENCODE_CONV_FLOAT(tv, val_di->di_tv.vval.v_float);
+ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+../src/nvim/eval/encode.c:330:26: note: expanded from macro 'TYPVAL_ENCODE_CONV_FLOAT'
+ switch (fpclassify(flt_)) { \
+ ~~~~~~~~~~~^~~~~
+/usr/include/math.h:415:56: note: expanded from macro 'fpclassify'
+# define fpclassify(x) __MATH_TG ((x), __fpclassify, (x))
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~
+/usr/include/math.h:370:16: note: expanded from macro '__MATH_TG'
+ ? FUNC ## f ARGS \
+ ~~~~~~~~~ ^~~~
+In file included from ../src/nvim/eval/encode.c:493:
+../src/nvim/eval/typval_encode.c.h:330:7: warning: implicit conversion loses floating-point precision: 'const float_T' (aka 'const double') to 'float' [-Wconv
+ersion]
+ TYPVAL_ENCODE_CONV_FLOAT(tv, tv->vval.v_float);
+ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+../src/nvim/eval/encode.c:330:26: note: expanded from macro 'TYPVAL_ENCODE_CONV_FLOAT'
+ switch (fpclassify(flt_)) { \
+ ~~~~~~~~~~~^~~~~
+/usr/include/math.h:415:56: note: expanded from macro 'fpclassify'
+# define fpclassify(x) __MATH_TG ((x), __fpclassify, (x))
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~
+/usr/include/math.h:370:16: note: expanded from macro '__MATH_TG'
+ ? FUNC ## f ARGS \
+ ~~~~~~~~~ ^~~~
+In file included from ../src/nvim/eval/encode.c:493:
+../src/nvim/eval/typval_encode.c.h:491:13: warning: implicit conversion loses floating-point precision: 'const float_T' (aka 'const double') to 'float' [-Wcon
+version]
+ TYPVAL_ENCODE_CONV_FLOAT(tv, val_di->di_tv.vval.v_float);
+ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+../src/nvim/eval/encode.c:330:26: note: expanded from macro 'TYPVAL_ENCODE_CONV_FLOAT'
+ switch (fpclassify(flt_)) { \
+ ~~~~~~~~~~~^~~~~
+/usr/include/math.h:415:56: note: expanded from macro 'fpclassify'
+# define fpclassify(x) __MATH_TG ((x), __fpclassify, (x))
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~
+/usr/include/math.h:370:16: note: expanded from macro '__MATH_TG'
+ ? FUNC ## f ARGS \
+ ~~~~~~~~~ ^~~~
+In file included from ../src/nvim/eval/encode.c:762:
+../src/nvim/eval/typval_encode.c.h:330:7: warning: implicit conversion loses floating-point precision: 'const float_T' (aka 'const double') to 'float' [-Wconv
+ersion]
+ TYPVAL_ENCODE_CONV_FLOAT(tv, tv->vval.v_float);
+ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+../src/nvim/eval/encode.c:534:26: note: expanded from macro 'TYPVAL_ENCODE_CONV_FLOAT'
+ switch (fpclassify(flt_)) { \
+ ~~~~~~~~~~~^~~~~
+/usr/include/math.h:415:56: note: expanded from macro 'fpclassify'
+# define fpclassify(x) __MATH_TG ((x), __fpclassify, (x))
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~
+/usr/include/math.h:370:16: note: expanded from macro '__MATH_TG'
+ ? FUNC ## f ARGS \
+ ~~~~~~~~~ ^~~~
+In file included from ../src/nvim/eval/encode.c:762:
+../src/nvim/eval/typval_encode.c.h:491:13: warning: implicit conversion loses floating-point precision: 'const float_T' (aka 'const double') to 'float' [-Wcon
+version]
+ TYPVAL_ENCODE_CONV_FLOAT(tv, val_di->di_tv.vval.v_float);
+ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+../src/nvim/eval/encode.c:534:26: note: expanded from macro 'TYPVAL_ENCODE_CONV_FLOAT'
+ switch (fpclassify(flt_)) { \
+ ~~~~~~~~~~~^~~~~
+/usr/include/math.h:415:56: note: expanded from macro 'fpclassify'
+# define fpclassify(x) __MATH_TG ((x), __fpclassify, (x))
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~
+/usr/include/math.h:370:16: note: expanded from macro '__MATH_TG'
+ ? FUNC ## f ARGS \
+ ~~~~~~~~~ ^~~~
+6 warnings generated.
+[261/284] Building C object src/nvim/CMakeFiles/nvim.dir/strings.c.o
+../src/nvim/strings.c:1223:23: warning: implicit conversion loses floating-point precision: 'double' to 'float' [-Wconversion]
+ if (isinf((double)f)
+ ~~~~~~^~~~~~~~~~
+/usr/include/math.h:472:46: note: expanded from macro 'isinf'
+# define isinf(x) __MATH_TG ((x), __isinf, (x))
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~^~~
+/usr/include/math.h:370:16: note: expanded from macro '__MATH_TG'
+ ? FUNC ## f ARGS \
+ ~~~~~~~~~ ^~~~
+../src/nvim/strings.c:1230:30: warning: implicit conversion loses floating-point precision: 'double' to 'float' [-Wconversion]
+ } else if (isnan(f)) {
+ ~~~~~~^~
+/usr/include/math.h:455:46: note: expanded from macro 'isnan'
+# define isnan(x) __MATH_TG ((x), __isnan, (x))
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~^~~
+/usr/include/math.h:370:16: note: expanded from macro '__MATH_TG'
+ ? FUNC ## f ARGS \
+ ~~~~~~~~~ ^~~~
+2 warnings generated.
+```
+### Expected behaviour
+
+No warnings. `maths.h` uses gcc magic to avoid this warning when compiling with gcc 4.4+, which doesn't work with clang (unless C11 is active, but using a different C version than the project's chosen one is probably a bad idea).
+
+Anyone thinking of a better workaround than moving float code to a special `-Wno-conversion` c file?
+",1.0,"(clang >= 6.0 bug) Annoying warnings for isnan(), fpclassify(), et al - - `nvim --version`: ef4feab0e75be
+- Vim (version: 8.0.1565) behaves differently? No warnings, but I didn't check if the same functions were used
+- Operating system/version: arch linux
+- Terminal name/version: pangoterm
+- `$TERM`: xterm
+
+### Steps to reproduce using `nvim -u NORC`
+```
+rm -rf build && CMAKE_EXTRA_FLAGS=""-DCMAKE_C_COMPILER=clang -DCLANG_ASAN_UBSAN=1"" make -j4
+```
+
+### Actual behaviour
+```
+[197/284] Building C object src/nvim/CMakeFiles/nvim.dir/eval/encode.c.o
+In file included from ../src/nvim/eval/encode.c:455:
+../src/nvim/eval/typval_encode.c.h:330:7: warning: implicit conversion loses floating-point precision: 'const float_T' (aka 'const double') to 'float' [-Wconv
+ersion]
+ TYPVAL_ENCODE_CONV_FLOAT(tv, tv->vval.v_float);
+ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+../src/nvim/eval/encode.c:330:26: note: expanded from macro 'TYPVAL_ENCODE_CONV_FLOAT'
+ switch (fpclassify(flt_)) { \
+ ~~~~~~~~~~~^~~~~
+/usr/include/math.h:415:56: note: expanded from macro 'fpclassify'
+# define fpclassify(x) __MATH_TG ((x), __fpclassify, (x))
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~
+/usr/include/math.h:370:16: note: expanded from macro '__MATH_TG'
+ ? FUNC ## f ARGS \
+ ~~~~~~~~~ ^~~~
+In file included from ../src/nvim/eval/encode.c:455:
+../src/nvim/eval/typval_encode.c.h:491:13: warning: implicit conversion loses floating-point precision: 'const float_T' (aka 'const double') to 'float' [-Wcon
+version]
+ TYPVAL_ENCODE_CONV_FLOAT(tv, val_di->di_tv.vval.v_float);
+ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+../src/nvim/eval/encode.c:330:26: note: expanded from macro 'TYPVAL_ENCODE_CONV_FLOAT'
+ switch (fpclassify(flt_)) { \
+ ~~~~~~~~~~~^~~~~
+/usr/include/math.h:415:56: note: expanded from macro 'fpclassify'
+# define fpclassify(x) __MATH_TG ((x), __fpclassify, (x))
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~
+/usr/include/math.h:370:16: note: expanded from macro '__MATH_TG'
+ ? FUNC ## f ARGS \
+ ~~~~~~~~~ ^~~~
+In file included from ../src/nvim/eval/encode.c:493:
+../src/nvim/eval/typval_encode.c.h:330:7: warning: implicit conversion loses floating-point precision: 'const float_T' (aka 'const double') to 'float' [-Wconv
+ersion]
+ TYPVAL_ENCODE_CONV_FLOAT(tv, tv->vval.v_float);
+ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+../src/nvim/eval/encode.c:330:26: note: expanded from macro 'TYPVAL_ENCODE_CONV_FLOAT'
+ switch (fpclassify(flt_)) { \
+ ~~~~~~~~~~~^~~~~
+/usr/include/math.h:415:56: note: expanded from macro 'fpclassify'
+# define fpclassify(x) __MATH_TG ((x), __fpclassify, (x))
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~
+/usr/include/math.h:370:16: note: expanded from macro '__MATH_TG'
+ ? FUNC ## f ARGS \
+ ~~~~~~~~~ ^~~~
+In file included from ../src/nvim/eval/encode.c:493:
+../src/nvim/eval/typval_encode.c.h:491:13: warning: implicit conversion loses floating-point precision: 'const float_T' (aka 'const double') to 'float' [-Wcon
+version]
+ TYPVAL_ENCODE_CONV_FLOAT(tv, val_di->di_tv.vval.v_float);
+ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+../src/nvim/eval/encode.c:330:26: note: expanded from macro 'TYPVAL_ENCODE_CONV_FLOAT'
+ switch (fpclassify(flt_)) { \
+ ~~~~~~~~~~~^~~~~
+/usr/include/math.h:415:56: note: expanded from macro 'fpclassify'
+# define fpclassify(x) __MATH_TG ((x), __fpclassify, (x))
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~
+/usr/include/math.h:370:16: note: expanded from macro '__MATH_TG'
+ ? FUNC ## f ARGS \
+ ~~~~~~~~~ ^~~~
+In file included from ../src/nvim/eval/encode.c:762:
+../src/nvim/eval/typval_encode.c.h:330:7: warning: implicit conversion loses floating-point precision: 'const float_T' (aka 'const double') to 'float' [-Wconv
+ersion]
+ TYPVAL_ENCODE_CONV_FLOAT(tv, tv->vval.v_float);
+ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+../src/nvim/eval/encode.c:534:26: note: expanded from macro 'TYPVAL_ENCODE_CONV_FLOAT'
+ switch (fpclassify(flt_)) { \
+ ~~~~~~~~~~~^~~~~
+/usr/include/math.h:415:56: note: expanded from macro 'fpclassify'
+# define fpclassify(x) __MATH_TG ((x), __fpclassify, (x))
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~
+/usr/include/math.h:370:16: note: expanded from macro '__MATH_TG'
+ ? FUNC ## f ARGS \
+ ~~~~~~~~~ ^~~~
+In file included from ../src/nvim/eval/encode.c:762:
+../src/nvim/eval/typval_encode.c.h:491:13: warning: implicit conversion loses floating-point precision: 'const float_T' (aka 'const double') to 'float' [-Wcon
+version]
+ TYPVAL_ENCODE_CONV_FLOAT(tv, val_di->di_tv.vval.v_float);
+ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+../src/nvim/eval/encode.c:534:26: note: expanded from macro 'TYPVAL_ENCODE_CONV_FLOAT'
+ switch (fpclassify(flt_)) { \
+ ~~~~~~~~~~~^~~~~
+/usr/include/math.h:415:56: note: expanded from macro 'fpclassify'
+# define fpclassify(x) __MATH_TG ((x), __fpclassify, (x))
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~
+/usr/include/math.h:370:16: note: expanded from macro '__MATH_TG'
+ ? FUNC ## f ARGS \
+ ~~~~~~~~~ ^~~~
+6 warnings generated.
+[261/284] Building C object src/nvim/CMakeFiles/nvim.dir/strings.c.o
+../src/nvim/strings.c:1223:23: warning: implicit conversion loses floating-point precision: 'double' to 'float' [-Wconversion]
+ if (isinf((double)f)
+ ~~~~~~^~~~~~~~~~
+/usr/include/math.h:472:46: note: expanded from macro 'isinf'
+# define isinf(x) __MATH_TG ((x), __isinf, (x))
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~^~~
+/usr/include/math.h:370:16: note: expanded from macro '__MATH_TG'
+ ? FUNC ## f ARGS \
+ ~~~~~~~~~ ^~~~
+../src/nvim/strings.c:1230:30: warning: implicit conversion loses floating-point precision: 'double' to 'float' [-Wconversion]
+ } else if (isnan(f)) {
+ ~~~~~~^~
+/usr/include/math.h:455:46: note: expanded from macro 'isnan'
+# define isnan(x) __MATH_TG ((x), __isnan, (x))
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~^~~
+/usr/include/math.h:370:16: note: expanded from macro '__MATH_TG'
+ ? FUNC ## f ARGS \
+ ~~~~~~~~~ ^~~~
+2 warnings generated.
+```
+### Expected behaviour
+
+No warnings. `maths.h` uses gcc magic to avoid this warning when compiling with gcc 4.4+, which doesn't work with clang (unless C11 is active, but using a different C version than the project's chosen one is probably a bad idea).
+
+Anyone thinking of a better workaround than moving float code to a special `-Wno-conversion` c file?
+",0, clang bug annoying warnings for isnan fpclassify et al nvim version vim version behaves differently no warnings but i didn t check if the same functions were used operating system version arch linux terminal name version pangoterm term xterm steps to reproduce using nvim u norc rm rf build cmake extra flags dcmake c compiler clang dclang asan ubsan make actual behaviour building c object src nvim cmakefiles nvim dir eval encode c o in file included from src nvim eval encode c src nvim eval typval encode c h warning implicit conversion loses floating point precision const float t aka const double to float wconv ersion typval encode conv float tv tv vval v float src nvim eval encode c note expanded from macro typval encode conv float switch fpclassify flt usr include math h note expanded from macro fpclassify define fpclassify x math tg x fpclassify x usr include math h note expanded from macro math tg func f args in file included from src nvim eval encode c src nvim eval typval encode c h warning implicit conversion loses floating point precision const float t aka const double to float wcon version typval encode conv float tv val di di tv vval v float src nvim eval encode c note expanded from macro typval encode conv float switch fpclassify flt usr include math h note expanded from macro fpclassify define fpclassify x math tg x fpclassify x usr include math h note expanded from macro math tg func f args in file included from src nvim eval encode c src nvim eval typval encode c h warning implicit conversion loses floating point precision const float t aka const double to float wconv ersion typval encode conv float tv tv vval v float src nvim eval encode c note expanded from macro typval encode conv float switch fpclassify flt usr include math h note expanded from macro fpclassify define fpclassify x math tg x fpclassify x usr include math h note expanded from macro math tg func f args in file included from src nvim eval encode c src nvim eval typval encode c h warning implicit conversion loses floating point precision const float t aka const double to float wcon version typval encode conv float tv val di di tv vval v float src nvim eval encode c note expanded from macro typval encode conv float switch fpclassify flt usr include math h note expanded from macro fpclassify define fpclassify x math tg x fpclassify x usr include math h note expanded from macro math tg func f args in file included from src nvim eval encode c src nvim eval typval encode c h warning implicit conversion loses floating point precision const float t aka const double to float wconv ersion typval encode conv float tv tv vval v float src nvim eval encode c note expanded from macro typval encode conv float switch fpclassify flt usr include math h note expanded from macro fpclassify define fpclassify x math tg x fpclassify x usr include math h note expanded from macro math tg func f args in file included from src nvim eval encode c src nvim eval typval encode c h warning implicit conversion loses floating point precision const float t aka const double to float wcon version typval encode conv float tv val di di tv vval v float src nvim eval encode c note expanded from macro typval encode conv float switch fpclassify flt usr include math h note expanded from macro fpclassify define fpclassify x math tg x fpclassify x usr include math h note expanded from macro math tg func f args warnings generated building c object src nvim cmakefiles nvim dir strings c o src nvim strings c warning implicit conversion loses floating point precision double to float if isinf double f usr include math h note expanded from macro isinf define isinf x math tg x isinf x usr include math h note expanded from macro math tg func f args src nvim strings c warning implicit conversion loses floating point precision double to float else if isnan f usr include math h note expanded from macro isnan define isnan x math tg x isnan x usr include math h note expanded from macro math tg func f args warnings generated expected behaviour no warnings maths h uses gcc magic to avoid this warning when compiling with gcc which doesn t work with clang unless is active but using a different c version than the project s chosen one is probably a bad idea anyone thinking of a better workaround than moving float code to a special wno conversion c file ,0
+37822,15379186743.0,IssuesEvent,2021-03-02 19:19:25,cityofaustin/atd-data-tech,https://api.github.com/repos/cityofaustin/atd-data-tech,closed,[RPP Back Office] Clean up Day Pass field,Product: Residential Parking Permit Digitization Service: Apps Type: Data Workgroup: PE,Clean up task - update the Number of Permits and Passes field with the correct value on existing applications where Day Passes were issued,1.0,[RPP Back Office] Clean up Day Pass field - Clean up task - update the Number of Permits and Passes field with the correct value on existing applications where Day Passes were issued,0, clean up day pass field clean up task update the number of permits and passes field with the correct value on existing applications where day passes were issued,0
+94,3387103297.0,IssuesEvent,2015-11-28 03:01:35,bbergen/tbmd.com,https://api.github.com/repos/bbergen/tbmd.com,closed,Edit Entity,Movies People,"We should give logged in users the ability to edit existing content on the site.
+
+Stub for now. ",1.0,"Edit Entity - We should give logged in users the ability to edit existing content on the site.
+
+Stub for now. ",1,edit entity we should give logged in users the ability to edit existing content on the site stub for now ,1
+704456,24197738556.0,IssuesEvent,2022-09-24 05:25:19,AY2223S1-CS2103-F13-2/tp,https://api.github.com/repos/AY2223S1-CS2103-F13-2/tp,opened,"As a user, I can copy a surveyee's information with shortcut key.",type.Story priority.Medium,So that I can extract important information from the surveyee.,1.0,"As a user, I can copy a surveyee's information with shortcut key. - So that I can extract important information from the surveyee.",0,as a user i can copy a surveyee s information with shortcut key so that i can extract important information from the surveyee ,0
+638,11431436429.0,IssuesEvent,2020-02-04 12:08:48,techtolentino/randytolentinocom,https://api.github.com/repos/techtolentino/randytolentinocom,opened,The Unoffical Leader,article people team,"One can be exhibit leadership qualities without having the title
+
+it's about the behavioral patterns and qualities of the individual
+
+Companies need to spot these people and train them, because if they
+already have a following, you, as a business, want to inherit that following.
+
+Think of the upcoming generation of employees, consider how they communicate
+how they behave. Think about social media and how it has _influencers_... then tie that back to your organization.
+
+Who are your influencers, how do you, as a corporation, take these influencers
+and have them help your business.
+
+If there's one thing social media has taught us, is that people follow people they
+relate to or aspire to be.
+
+Look around you. Think about the people in your area that are the influencers
+on your team. If they're inspiring others, if they're solving problems,
+then you might be looking at your
+next set of leaders.
+
+
+Qualities
+- caring for people in a holistic way, and understanding how that relates to the business
+- they listen for needs across the team, growth wise, daily... it's the act of listening
+- people go to them for advice
+- they advocate for others, they help others shine, speak,
+- transparency of failure, how they handle it, and how they show they've learned from it
+- the art of feedback: seek, acknowledge, and learn from it",1.0,"The Unoffical Leader - One can be exhibit leadership qualities without having the title
+
+it's about the behavioral patterns and qualities of the individual
+
+Companies need to spot these people and train them, because if they
+already have a following, you, as a business, want to inherit that following.
+
+Think of the upcoming generation of employees, consider how they communicate
+how they behave. Think about social media and how it has _influencers_... then tie that back to your organization.
+
+Who are your influencers, how do you, as a corporation, take these influencers
+and have them help your business.
+
+If there's one thing social media has taught us, is that people follow people they
+relate to or aspire to be.
+
+Look around you. Think about the people in your area that are the influencers
+on your team. If they're inspiring others, if they're solving problems,
+then you might be looking at your
+next set of leaders.
+
+
+Qualities
+- caring for people in a holistic way, and understanding how that relates to the business
+- they listen for needs across the team, growth wise, daily... it's the act of listening
+- people go to them for advice
+- they advocate for others, they help others shine, speak,
+- transparency of failure, how they handle it, and how they show they've learned from it
+- the art of feedback: seek, acknowledge, and learn from it",1,the unoffical leader one can be exhibit leadership qualities without having the title it s about the behavioral patterns and qualities of the individual companies need to spot these people and train them because if they already have a following you as a business want to inherit that following think of the upcoming generation of employees consider how they communicate how they behave think about social media and how it has influencers then tie that back to your organization who are your influencers how do you as a corporation take these influencers and have them help your business if there s one thing social media has taught us is that people follow people they relate to or aspire to be look around you think about the people in your area that are the influencers on your team if they re inspiring others if they re solving problems then you might be looking at your next set of leaders qualities caring for people in a holistic way and understanding how that relates to the business they listen for needs across the team growth wise daily it s the act of listening people go to them for advice they advocate for others they help others shine speak transparency of failure how they handle it and how they show they ve learned from it the art of feedback seek acknowledge and learn from it,1
+86164,24774850547.0,IssuesEvent,2022-10-23 15:55:34,atsams-rs/atsamx7x-rust,https://api.github.com/repos/atsams-rs/atsamx7x-rust,opened,Introduce feature permutations in CI,enhancement build system,"Raised as result of discussion in #38.
+
+Consider [`cargo-hack`] to support feature permutations in CI.",1.0,"Introduce feature permutations in CI - Raised as result of discussion in #38.
+
+Consider [`cargo-hack`] to support feature permutations in CI.",0,introduce feature permutations in ci raised as result of discussion in consider to support feature permutations in ci ,0
+56429,6519495092.0,IssuesEvent,2017-08-28 13:04:40,Microsoft/vsts-tasks,https://api.github.com/repos/Microsoft/vsts-tasks,closed,Visual Studio Test Agent Deployment - Version 2.* credentials do not match version 1.*,Area: Test,"I use version 1.* of ""Visual Studio Test Agent Deployment"" task, configured as such:
+
+
+I have a single test afterwards (Run functional tests), which asserts that currently executing user (Environment.UserName) is equal to value of $(TestUsername). This test passes in version 1.*.
+
+Switching to version 2.* of ""Visual Studio Test Agent Deployment"" task, configured as such:
+
+
+The same test now fails, because currently executing user (Environment.UserName) is equal to value of $(AdminUsername).",1.0,"Visual Studio Test Agent Deployment - Version 2.* credentials do not match version 1.* - I use version 1.* of ""Visual Studio Test Agent Deployment"" task, configured as such:
+
+
+I have a single test afterwards (Run functional tests), which asserts that currently executing user (Environment.UserName) is equal to value of $(TestUsername). This test passes in version 1.*.
+
+Switching to version 2.* of ""Visual Studio Test Agent Deployment"" task, configured as such:
+
+
+The same test now fails, because currently executing user (Environment.UserName) is equal to value of $(AdminUsername).",0,visual studio test agent deployment version credentials do not match version i use version of visual studio test agent deployment task configured as such i have a single test afterwards run functional tests which asserts that currently executing user environment username is equal to value of testusername this test passes in version switching to version of visual studio test agent deployment task configured as such the same test now fails because currently executing user environment username is equal to value of adminusername ,0
+86657,17033882719.0,IssuesEvent,2021-07-05 02:32:11,nothings/stb,https://api.github.com/repos/nothings/stb,closed,"stb_vorbis.c: Dead assignment, initialization, Assigned value is garbage or undefined",1 stb_vorbis code quality,"Hi , scan-build reported on lines
+
+Dead store | Dead initialization | stb_vorbis.c | start_decoder | 3429 |
+-- | -- | -- | -- | -- |
+Dead store | Dead assignment | stb_vorbis.c | stb_vorbis_stream_length_in_samples | 4497 |
+Dead store | Dead assignment | stb_vorbis.c | stb_vorbis_get_samples_short | 4830 |
+Dead store | Dead assignment | stb_vorbis.c | stb_vorbis_get_samples_short_interleaved | 4810 |
+Logic error | Assigned value is garbage or undefined | stb_vorbis.c | get_seek_page_info | 4171 |
+Logic error | Result of operation is garbage or undefined | stb_vorbis.c | get_seek_page_info | 4164 |
+
+
+
+
+
+
+
+
+
+
+",1.0,"stb_vorbis.c: Dead assignment, initialization, Assigned value is garbage or undefined - Hi , scan-build reported on lines
+
+Dead store | Dead initialization | stb_vorbis.c | start_decoder | 3429 |
+-- | -- | -- | -- | -- |
+Dead store | Dead assignment | stb_vorbis.c | stb_vorbis_stream_length_in_samples | 4497 |
+Dead store | Dead assignment | stb_vorbis.c | stb_vorbis_get_samples_short | 4830 |
+Dead store | Dead assignment | stb_vorbis.c | stb_vorbis_get_samples_short_interleaved | 4810 |
+Logic error | Assigned value is garbage or undefined | stb_vorbis.c | get_seek_page_info | 4171 |
+Logic error | Result of operation is garbage or undefined | stb_vorbis.c | get_seek_page_info | 4164 |
+
+
+
+
+
+
+
+
+
+
+",0,stb vorbis c dead assignment initialization assigned value is garbage or undefined hi scan build reported on lines dead store dead initialization stb vorbis c start decoder dead store dead assignment stb vorbis c stb vorbis stream length in samples dead store dead assignment stb vorbis c stb vorbis get samples short dead store dead assignment stb vorbis c stb vorbis get samples short interleaved logic error assigned value is garbage or undefined stb vorbis c get seek page info logic error result of operation is garbage or undefined stb vorbis c get seek page info ,0
+34643,16623505766.0,IssuesEvent,2021-06-03 06:34:39,angular/angular-cli,https://api.github.com/repos/angular/angular-cli,closed,Upgrading from v9 to v11 ng build/serve are very slow #20227,need: more info severity4: memory/performance,"
+
+# 🐞 Bug report
+
+### Command (mark with an `x`)
+
+
+
+
+- [ ] new
+- [x] build
+- [ ] serve
+- [ ] test
+- [ ] e2e
+- [ ] generate
+- [ ] add
+- [ ] update
+- [ ] lint
+- [ ] extract-i18n
+- [ ] run
+- [ ] config
+- [ ] help
+- [ ] version
+- [ ] doc
+
+### Is this a regression?
+
+
+ Maybe? Hard to tell if we just aren't doing something right currently.
+### Description
+
+We were unable to upgrade from angular 9 to 11 last year. After doing the migration the builds are taking around an hour and ng serve is taking around 10minutes to start and then a few minutes to recompile.
+
+## 🔬 Minimal Reproduction
+
+The project is quite large. After the discussion here: https://github.com/angular/angular-cli/issues/20227 we had lot of failures of generating the .json files with NG_BUILD_PROFILING set to 1. Eventually one of our developers was able too after cutting out a large portion of our code base, but it just mimics how we're building stuff elsewhere.
+
+## 🔥 Exception or Error
+
+(copied from the last bug) I saw this issue raised: #17557 that was able to spit out the webpack timings but I tried the same command and didn't get anything. Here are some logs from that command though:
+
+Hash: 10062e72e655a942dd34
+Version: webpack 4.44.2
+Time: 708716ms
+Built at: 03/09/2021 10:15:12 AM
+
+LOG from webpack.buildChunkGraph.visitModules
+prepare: 157.7756ms
+visiting: 11.9087ms
+calculating available modules: 0.9064ms
+merging available modules: 0.1087ms
+visiting: 47.7026ms
+calculating available modules: 1.9921ms
+merging available modules: 0.023ms
+visiting: 1.39ms
+
+Initial Chunk Files | Names | Size
+main-es5.3ac0413d5fda208dcf49.js | main | 3.60 MB
+main-es2015.3ac0413d5fda208dcf49.js | main | 3.32 MB
+polyfills-es5.1577f06bbd7cd498d2aa.js | polyfills-es5 | 131.77 kB
+styles.088de8f15a462993e614.css | styles | 108.92 kB
+polyfills-es2015.603bb5ba1ae3ee6b7388.js | polyfills | 45.98 kB
+runtime-es2015.12e587ff0e65ab518ba8.js | runtime | 3.40 kB
+runtime-es5.12e587ff0e65ab518ba8.js | runtime | 3.40 kB
+
+## 🌍 Your Environment
+
+
+
+**Anything else relevant?**
+
+Here's the speed-measure-plugin.json:
+
+```
+{
+ ""misc"": {
+ ""compileTime"": 485627
+ },
+ ""plugins"": {
+ ""AngularCompilerPlugin"": 253168,
+ ""BuildOptimizerWebpackPlugin"": 61121,
+ ""ContextReplacementPlugin"": 79,
+ ""CopyPlugin"": 67,
+ ""MiniCssExtractPlugin"": 29,
+ ""ProfilingPlugin"": 70949,
+ ""ProgressPlugin"": 11394,
+ ""CircularDependencyPlugin"": 6080,
+ ""LicenseWebpackPlugin"": 9014,
+ ""AnyComponentStyleBudgetChecker"": 241,
+ ""SuppressExtractedTextChunksWebpackPlugin"": 125,
+ ""NgBuildAnalyticsPlugin"": 5058,
+ ""HashedModuleIdsPlugin"": 631,
+ ""OptimizeCssWebpackPlugin"": 5030,
+ ""TerserPlugin"": 44879,
+ ""BundleBudgetPlugin"": 970
+ },
+ ""loaders"": {
+ ""build"": [
+ {
+ ""averages"": {
+ ""dataPoints"": 1,
+ ""median"": 202470,
+ ""mean"": 202470,
+ ""range"": {
+ ""start"": 202470,
+ ""end"": 202470
+ }
+ },
+ ""activeTime"": 202470,
+ ""loaders"": [
+ ""C:\\git\\bpi\\ng\\BPI.UI\\node_modules\\mini-css-extract-plugin\\dist\\loader.js"",
+ ""C:\\git\\bpi\\ng\\BPI.UI\\node_modules\\css-loader\\dist\\cjs.js"",
+ ""C:\\git\\bpi\\ng\\BPI.UI\\node_modules\\postcss-loader\\src\\index.js"",
+ ""C:\\git\\bpi\\ng\\BPI.UI\\node_modules\\sass-loader\\dist\\cjs.js""
+ ],
+ ""subLoadersTime"": {}
+ },
+ {
+ ""averages"": {
+ ""dataPoints"": 1,
+ ""median"": 194400,
+ ""mean"": 194400,
+ ""range"": {
+ ""start"": 194400,
+ ""end"": 194400
+ }
+ },
+ ""activeTime"": 194400,
+ ""loaders"": [
+ ""C:\\git\\bpi\\ng\\BPI.UI\\node_modules\\css-loader\\dist\\cjs.js"",
+ ""C:\\git\\bpi\\ng\\BPI.UI\\node_modules\\postcss-loader\\src\\index.js"",
+ ""C:\\git\\bpi\\ng\\BPI.UI\\node_modules\\sass-loader\\dist\\cjs.js""
+ ],
+ ""subLoadersTime"": {}
+ },
+ {
+ ""averages"": {
+ ""dataPoints"": 405,
+ ""median"": 4328,
+ ""mean"": 5250,
+ ""range"": {
+ ""start"": 363,
+ ""end"": 13806
+ },
+ ""variance"": 12656309
+ },
+ ""activeTime"": 13806,
+ ""loaders"": [
+ ""C:\\git\\bpi\\ng\\BPI.UI\\node_modules\\raw-loader\\dist\\cjs.js"",
+ ""C:\\git\\bpi\\ng\\BPI.UI\\node_modules\\postcss-loader\\src\\index.js"",
+ ""C:\\git\\bpi\\ng\\BPI.UI\\node_modules\\sass-loader\\dist\\cjs.js""
+ ],
+ ""subLoadersTime"": {}
+ },
+ {
+ ""averages"": {
+ ""dataPoints"": 1212,
+ ""median"": 5,
+ ""mean"": 26,
+ ""range"": {
+ ""start"": 2,
+ ""end"": 6484
+ },
+ ""variance"": 39836
+ },
+ ""activeTime"": 32036,
+ ""loaders"": [
+ ""C:\\git\\bpi\\ng\\BPI.UI\\node_modules\\@angular-devkit\\build-optimizer\\src\\build-optimizer\\webpack-loader.js""
+ ],
+ ""subLoadersTime"": {}
+ },
+ {
+ ""averages"": {
+ ""dataPoints"": 1304,
+ ""median"": 31,
+ ""mean"": 17,
+ ""range"": {
+ ""start"": 3,
+ ""end"": 644
+ },
+ ""variance"": 639
+ },
+ ""activeTime"": 22403,
+ ""loaders"": [
+ ""C:\\git\\bpi\\ng\\BPI.UI\\node_modules\\@angular-devkit\\build-optimizer\\src\\build-optimizer\\webpack-loader.js"",
+ ""C:\\git\\bpi\\ng\\BPI.UI\\node_modules\\@ngtools\\webpack\\src\\index.js""
+ ],
+ ""subLoadersTime"": {}
+ },
+ {
+ ""averages"": {
+ ""dataPoints"": 3,
+ ""median"": 7,
+ ""mean"": 15,
+ ""range"": {
+ ""start"": 7,
+ ""end"": 29
+ },
+ ""variance"": 148
+ },
+ ""activeTime"": 45,
+ ""loaders"": [
+ ""modules with no loaders""
+ ],
+ ""subLoadersTime"": {}
+ },
+ {
+ ""averages"": {
+ ""dataPoints"": 73,
+ ""median"": 8,
+ ""mean"": 10,
+ ""range"": {
+ ""start"": 7,
+ ""end"": 68
+ },
+ ""variance"": 51
+ },
+ ""activeTime"": 718,
+ ""loaders"": [
+ ""C:\\git\\bpi\\ng\\BPI.UI\\node_modules\\svg-inline-loader\\index.js""
+ ],
+ ""subLoadersTime"": {}
+ }
+ ]
+ }
+}
+
+```
+
+
+
+
+",0,upgrading from to ng build serve are very slow 🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅 oh hi there 😄 to expedite issue processing please search open and closed issues before submitting a new one existing issues often contain information about workarounds resolution or progress updates 🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅 🐞 bug report command mark with an x new build serve test generate add update lint extract run config help version doc is this a regression maybe hard to tell if we just aren t doing something right currently description we were unable to upgrade from angular to last year after doing the migration the builds are taking around an hour and ng serve is taking around to start and then a few minutes to recompile 🔬 minimal reproduction the project is quite large after the discussion here we had lot of failures of generating the json files with ng build profiling set to eventually one of our developers was able too after cutting out a large portion of our code base but it just mimics how we re building stuff elsewhere 🔥 exception or error copied from the last bug i saw this issue raised that was able to spit out the webpack timings but i tried the same command and didn t get anything here are some logs from that command though hash version webpack time built at am log from webpack buildchunkgraph visitmodules prepare visiting calculating available modules merging available modules visiting calculating available modules merging available modules visiting initial chunk files names size main js main mb main js main mb polyfills js polyfills kb styles css styles kb polyfills js polyfills kb runtime js runtime kb runtime js runtime kb 🌍 your environment angular cli node os angular animations common compiler compiler cli core forms language service localize platform browser platform browser dynamic router ivy workspace yes package version angular devkit architect cli only angular devkit build angular angular devkit core cli only angular devkit schematics angular cdk angular cli angular flex layout beta angular material schematics angular schematics update ng packagr rxjs typescript anything else relevant here s the speed measure plugin json misc compiletime plugins angularcompilerplugin buildoptimizerwebpackplugin contextreplacementplugin copyplugin minicssextractplugin profilingplugin progressplugin circulardependencyplugin licensewebpackplugin anycomponentstylebudgetchecker suppressextractedtextchunkswebpackplugin ngbuildanalyticsplugin hashedmoduleidsplugin optimizecsswebpackplugin terserplugin bundlebudgetplugin loaders build averages datapoints median mean range start end activetime loaders c git bpi ng bpi ui node modules mini css extract plugin dist loader js c git bpi ng bpi ui node modules css loader dist cjs js c git bpi ng bpi ui node modules postcss loader src index js c git bpi ng bpi ui node modules sass loader dist cjs js subloaderstime averages datapoints median mean range start end activetime loaders c git bpi ng bpi ui node modules css loader dist cjs js c git bpi ng bpi ui node modules postcss loader src index js c git bpi ng bpi ui node modules sass loader dist cjs js subloaderstime averages datapoints median mean range start end variance activetime loaders c git bpi ng bpi ui node modules raw loader dist cjs js c git bpi ng bpi ui node modules postcss loader src index js c git bpi ng bpi ui node modules sass loader dist cjs js subloaderstime averages datapoints median mean range start end variance activetime loaders c git bpi ng bpi ui node modules angular devkit build optimizer src build optimizer webpack loader js subloaderstime averages datapoints median mean range start end variance activetime loaders c git bpi ng bpi ui node modules angular devkit build optimizer src build optimizer webpack loader js c git bpi ng bpi ui node modules ngtools webpack src index js subloaderstime averages datapoints median mean range start end variance activetime loaders modules with no loaders subloaderstime averages datapoints median mean range start end variance activetime loaders c git bpi ng bpi ui node modules svg inline loader index js subloaderstime ,0
+1084,26398267628.0,IssuesEvent,2023-01-12 21:41:03,openstates/issues,https://api.github.com/repos/openstates/issues,opened,New SC Committee Scraper,good first issue component:people-data good first scraper,"### We need a new committee scraper written in [spatula](https://jamesturk.github.io/spatula/) for South Carolina.
+
+*Special note about SC joint committees PDF page: scraper should a text content check (or other approach) built into it to verify if the PDF was published during current session)*
+
+It should scrape this [webpage of Senate Committees](https://www.scstatehouse.gov/committee.php?chamber=S), this [webpage of House Committees](https://www.scstatehouse.gov/committee.php?chamber=H), and this [PDF webpage of Joint Committees](https://www.scstatehouse.gov/publicationspage/JtSpecCommList.pdf) to get:
+- **name**
+ - “Small Business”, “Agriculture”, etc.
+- **chamber**
+ - “upper”, “lower”, or “legislature” (which, in NJ's case, are those on Joint or Other Committees pages)
+- **classification**
+ - ex: ""committee"", ""subcommittee""
+- **parent**
+ - *only if it is a subcommittee* - scraper should have a way to determine this
+ - ex: parent=""Natural Resources"" when *sub*committee name=""Forestry""
+- **sources**
+ - each added using `add_source()` method on instance of `ScrapeCommittee` type object
+ - ex: home page for list of committees, specific page for that committee, etc.
+- **members**
+ - - each added using `add_member()`method on instance of `ScrapeCommittee` type object
+ - *Attributes:*
+ - **name**: “Jane Doe”, “John Smith”, etc.
+ - **role** (where applicable): “Chair”, “Ranking Member”, etc.
+
+#### [Here is a helpful general reference doc](https://docs.openstates.org/contributing/writing-a-committee-scraper/) for those new to writing scrapers, the Open States project, or the specific task of writing a committee scraper
+
+### Useful scrapers for reference
+A [spatula-based scraper](https://github.com/openstates/openstates-scrapers/blob/main/scrapers_next/mo/committees.py) that also scrapes an HTML List Page (using `HtmlListPage` and `HtmlPage` spatula classes) for reference.
+- As in the other spatula scraper, you will need to set the session variable separately (rather than depend on the `__init__.py`).
+
+### Other useful resources
+You can reference the [object models for committee scrapers](https://github.com/openstates/openstates-core/blob/main/openstates/models/committees.py) in the Open States Core repo.
+
+Further documentation on running spatula scrapers in the command line can be found [here](https://docs.openstates.org/contributing/scrapers/#running-spatula-scrapers).",1.0,"New SC Committee Scraper - ### We need a new committee scraper written in [spatula](https://jamesturk.github.io/spatula/) for South Carolina.
+
+*Special note about SC joint committees PDF page: scraper should a text content check (or other approach) built into it to verify if the PDF was published during current session)*
+
+It should scrape this [webpage of Senate Committees](https://www.scstatehouse.gov/committee.php?chamber=S), this [webpage of House Committees](https://www.scstatehouse.gov/committee.php?chamber=H), and this [PDF webpage of Joint Committees](https://www.scstatehouse.gov/publicationspage/JtSpecCommList.pdf) to get:
+- **name**
+ - “Small Business”, “Agriculture”, etc.
+- **chamber**
+ - “upper”, “lower”, or “legislature” (which, in NJ's case, are those on Joint or Other Committees pages)
+- **classification**
+ - ex: ""committee"", ""subcommittee""
+- **parent**
+ - *only if it is a subcommittee* - scraper should have a way to determine this
+ - ex: parent=""Natural Resources"" when *sub*committee name=""Forestry""
+- **sources**
+ - each added using `add_source()` method on instance of `ScrapeCommittee` type object
+ - ex: home page for list of committees, specific page for that committee, etc.
+- **members**
+ - - each added using `add_member()`method on instance of `ScrapeCommittee` type object
+ - *Attributes:*
+ - **name**: “Jane Doe”, “John Smith”, etc.
+ - **role** (where applicable): “Chair”, “Ranking Member”, etc.
+
+#### [Here is a helpful general reference doc](https://docs.openstates.org/contributing/writing-a-committee-scraper/) for those new to writing scrapers, the Open States project, or the specific task of writing a committee scraper
+
+### Useful scrapers for reference
+A [spatula-based scraper](https://github.com/openstates/openstates-scrapers/blob/main/scrapers_next/mo/committees.py) that also scrapes an HTML List Page (using `HtmlListPage` and `HtmlPage` spatula classes) for reference.
+- As in the other spatula scraper, you will need to set the session variable separately (rather than depend on the `__init__.py`).
+
+### Other useful resources
+You can reference the [object models for committee scrapers](https://github.com/openstates/openstates-core/blob/main/openstates/models/committees.py) in the Open States Core repo.
+
+Further documentation on running spatula scrapers in the command line can be found [here](https://docs.openstates.org/contributing/scrapers/#running-spatula-scrapers).",1,new sc committee scraper we need a new committee scraper written in for south carolina special note about sc joint committees pdf page scraper should a text content check or other approach built into it to verify if the pdf was published during current session it should scrape this this and this to get name “small business” “agriculture” etc chamber “upper” “lower” or “legislature” which in nj s case are those on joint or other committees pages classification ex committee subcommittee parent only if it is a subcommittee scraper should have a way to determine this ex parent natural resources when sub committee name forestry sources each added using add source method on instance of scrapecommittee type object ex home page for list of committees specific page for that committee etc members each added using add member method on instance of scrapecommittee type object attributes name “jane doe” “john smith” etc role where applicable “chair” “ranking member” etc for those new to writing scrapers the open states project or the specific task of writing a committee scraper useful scrapers for reference a that also scrapes an html list page using htmllistpage and htmlpage spatula classes for reference as in the other spatula scraper you will need to set the session variable separately rather than depend on the init py other useful resources you can reference the in the open states core repo further documentation on running spatula scrapers in the command line can be found ,1
+968,21803436232.0,IssuesEvent,2022-05-16 08:08:49,microsoftgraph/microsoft-graph-toolkit,https://api.github.com/repos/microsoftgraph/microsoft-graph-toolkit,closed,Graph user-ids property does not restrict ability to just select only those users named [BUG],bug Area: Components State: In Review Component: people-picker,"
+
+
+**Describe the bug**
+
+If I try to restrict the people picker to a list of ID's, in this case email addresses, the functionality does not work. I can see only the 2 addresses on the list BUT have the ability to type any other users
+
+
+
+**To Reproduce**
+Follow below
+
+For example the page:
+https://mgt.dev/?path=/story/components-mgt-people-picker-properties--picker-user-ids&source=docs
+
+
+
+
+restricts to 3 users in the demo. I can though type ""megan"" and it finds megan, not what was expected.
+
+
+**Expected behavior**
+when typing ""megan"" in either of the above it returns no matches as I have asked not to find her.
+
+**Screenshots**
+If applicable, add screenshots to help explain your problem.
+
+**Environment (please complete the following information):**
+ - OS: Windows 10
+ - Browser Chrome, Edge tested
+ - MS Graph
+ - Version: current on webpage
+
+**Additional context**
+Add any other context about the problem here.
+",1.0,"Graph user-ids property does not restrict ability to just select only those users named [BUG] -
+
+
+**Describe the bug**
+
+If I try to restrict the people picker to a list of ID's, in this case email addresses, the functionality does not work. I can see only the 2 addresses on the list BUT have the ability to type any other users
+
+
+
+**To Reproduce**
+Follow below
+
+For example the page:
+https://mgt.dev/?path=/story/components-mgt-people-picker-properties--picker-user-ids&source=docs
+
+
+
+
+restricts to 3 users in the demo. I can though type ""megan"" and it finds megan, not what was expected.
+
+
+**Expected behavior**
+when typing ""megan"" in either of the above it returns no matches as I have asked not to find her.
+
+**Screenshots**
+If applicable, add screenshots to help explain your problem.
+
+**Environment (please complete the following information):**
+ - OS: Windows 10
+ - Browser Chrome, Edge tested
+ - MS Graph
+ - Version: current on webpage
+
+**Additional context**
+Add any other context about the problem here.
+",1,graph user ids property does not restrict ability to just select only those users named describe the bug if i try to restrict the people picker to a list of id s in this case email addresses the functionality does not work i can see only the addresses on the list but have the ability to type any other users mgt people picker user ids emilyb onmicrosoft com lynner onmicrosoft com to reproduce follow below for example the page mgt people picker user ids ceea restricts to users in the demo i can though type megan and it finds megan not what was expected expected behavior when typing megan in either of the above it returns no matches as i have asked not to find her screenshots if applicable add screenshots to help explain your problem environment please complete the following information os windows browser chrome edge tested ms graph version current on webpage additional context add any other context about the problem here ,1
+370,6882977760.0,IssuesEvent,2017-11-21 07:26:40,PopulateTools/gobierto,https://api.github.com/repos/PopulateTools/gobierto,closed,Google Calendar: Import all received events (unless private),bug gobierto-people,"Import all received events in an API request to the correponding person calendar collection, no matter if the event comes from a shared calendar, the person is assisting or not or is organizer.",1.0,"Google Calendar: Import all received events (unless private) - Import all received events in an API request to the correponding person calendar collection, no matter if the event comes from a shared calendar, the person is assisting or not or is organizer.",1,google calendar import all received events unless private import all received events in an api request to the correponding person calendar collection no matter if the event comes from a shared calendar the person is assisting or not or is organizer ,1
+1155,30181451729.0,IssuesEvent,2023-07-04 09:09:42,newtheatre/history-project,https://api.github.com/repos/newtheatre/history-project,closed,Stage name,people report-tool,"End user submitted issue from page: [/people/sophie_walton/](https://history.newtheatre.org.uk/people/sophie_walton/)
+---
+Believe she’s now going by Pippa Walton for the acting stuff
+*Ben Woodford*
+",1.0,"Stage name - End user submitted issue from page: [/people/sophie_walton/](https://history.newtheatre.org.uk/people/sophie_walton/)
+---
+Believe she’s now going by Pippa Walton for the acting stuff
+*Ben Woodford*
+",1,stage name end user submitted issue from page believe she’s now going by pippa walton for the acting stuff ben woodford ,1
+7808,5221526864.0,IssuesEvent,2017-01-27 02:02:16,kubernetes/dashboard,https://api.github.com/repos/kubernetes/dashboard,opened,"Supporting ""Apps""",area/usability area/ux kind/feature priority/P1,"There has been some discussion about how to best support ""Applications"". Many users will want to view deployed objects as an ""application"" and may not care or even be aware of each of the individual objects (Deployments, Services, ConfigMaps, ThirdPartyResources, etc.) deployed in the cluster.
+
+We will need to figure out what an ""app"" is. It implies some kind of ownership relationship between an ""app"" and the individual objects deployed in the cluster. Helm does something similar with objects deployed as part of a helm chart. This may be a good place to start.
+
+Some other UIs like OpenShift do some grouping of objects into projects and applications.
+
+Related issues/PRs: #1504 #1570 #1332
+
+cc @erictune @davidopp ",True,"Supporting ""Apps"" - There has been some discussion about how to best support ""Applications"". Many users will want to view deployed objects as an ""application"" and may not care or even be aware of each of the individual objects (Deployments, Services, ConfigMaps, ThirdPartyResources, etc.) deployed in the cluster.
+
+We will need to figure out what an ""app"" is. It implies some kind of ownership relationship between an ""app"" and the individual objects deployed in the cluster. Helm does something similar with objects deployed as part of a helm chart. This may be a good place to start.
+
+Some other UIs like OpenShift do some grouping of objects into projects and applications.
+
+Related issues/PRs: #1504 #1570 #1332
+
+cc @erictune @davidopp ",0,supporting apps there has been some discussion about how to best support applications many users will want to view deployed objects as an application and may not care or even be aware of each of the individual objects deployments services configmaps thirdpartyresources etc deployed in the cluster we will need to figure out what an app is it implies some kind of ownership relationship between an app and the individual objects deployed in the cluster helm does something similar with objects deployed as part of a helm chart this may be a good place to start some other uis like openshift do some grouping of objects into projects and applications related issues prs cc erictune davidopp ,0
+40207,5190886596.0,IssuesEvent,2017-01-21 14:40:41,aborruso/blackboard,https://api.github.com/repos/aborruso/blackboard,closed, Segnalazione di prova,design,"tel: ''
+email: g.dicarlo@provincia.teramo.it
+descrizione: ' Segnalazione di prova'
+indirizzo: ''
+lat: '42.501721'
+lon: '13.66018",1.0," Segnalazione di prova - tel: ''
+email: g.dicarlo@provincia.teramo.it
+descrizione: ' Segnalazione di prova'
+indirizzo: ''
+lat: '42.501721'
+lon: '13.66018",0, segnalazione di prova tel email g dicarlo provincia teramo it descrizione segnalazione di prova indirizzo lat lon ,0
+92561,26713946634.0,IssuesEvent,2023-01-28 08:30:53,sandboxie-plus/Sandboxie,https://api.github.com/repos/sandboxie-plus/Sandboxie,closed,[1.7.0] SandMan crashes when hovering over Run menu if there are any custom Run Menu entries are defined and OptaneShellExt is enabled.,fixed in next build,"### Describe what you noticed and did
+
+1. Install `Intel Optane Pinning Shell Extension`. (Comes with chipset driver? Intel RST 17.11.0.1000)
+2. Restart SandMan.exe.
+3. Add a custom Run Menu entry. (SandMan > Sandbox Options > General Options > Run Menu)
+4. Hover over the `Run` menu. (SanMan > Right click on a box to open sandbox context menu > Run)
+5. SandMan.exe crashes.
+6. Disabling the `Intel.Optane.Pinning.ShellExt.OptaneIconOverlay` fixes the issue. (with NirSoft ShellExView)
+
+###
+pw is `sandboxie`
+[pw-is-sandboxie.zip](https://github.com/sandboxie-plus/Sandboxie/files/10520761/pw-is-sandboxie.zip)
+
+
+### How often did you encounter it so far?
+
+Every time
+
+### Affected program
+
+Not relevant
+
+### Download link
+
+It is in the zip file
+
+### Where is the program located?
+
+Not relevant to my request.
+
+### Expected behavior
+
+The app shouldn't crash.
+
+### What is your Windows edition and version?
+
+Windows 10 Pro 22H2 64-bit
+
+### In which Windows account you have this problem?
+
+User account with UAC protection set to Always notify.
+
+### Please mention any installed security software
+
+Microsoft Windows Defender
+
+### What version of Sandboxie are you running?
+
+Sandboxie-Plus 1.7.0 64-bit
+
+### Is it a new installation of Sandboxie?
+
+I recently did a new clean installation.
+
+### Is it a regression?
+
+Since Sandboxie Plus v1.6.7 64-bit
+
+### In which sandbox type you have this problem?
+
+Not relevant to my request.
+
+### Can you reproduce this problem on a new empty sandbox?
+
+Not relevant to my request.
+
+### Did you previously enable some security policy settings outside Sandboxie?
+
+_No response_
+
+### Crash dump
+
+It is in the zip file
+
+### Trace log
+
+_No response_
+
+### Sandboxie.ini configuration
+
+```shell
+[Global Settings]
+RunCommand=Command Prompt|""C:\Windows\System32\cmd.exe""
+RunCommand=PowerShell|""C:\Windows\System32\WindowsPowerShell\v1.0\powershell.exe""
+RunCommand=PowerShell 7|""C:\Program Files\PowerShell\7\pwsh.exe""
+```
+",1.0,"[1.7.0] SandMan crashes when hovering over Run menu if there are any custom Run Menu entries are defined and OptaneShellExt is enabled. - ### Describe what you noticed and did
+
+1. Install `Intel Optane Pinning Shell Extension`. (Comes with chipset driver? Intel RST 17.11.0.1000)
+2. Restart SandMan.exe.
+3. Add a custom Run Menu entry. (SandMan > Sandbox Options > General Options > Run Menu)
+4. Hover over the `Run` menu. (SanMan > Right click on a box to open sandbox context menu > Run)
+5. SandMan.exe crashes.
+6. Disabling the `Intel.Optane.Pinning.ShellExt.OptaneIconOverlay` fixes the issue. (with NirSoft ShellExView)
+
+###
+pw is `sandboxie`
+[pw-is-sandboxie.zip](https://github.com/sandboxie-plus/Sandboxie/files/10520761/pw-is-sandboxie.zip)
+
+
+### How often did you encounter it so far?
+
+Every time
+
+### Affected program
+
+Not relevant
+
+### Download link
+
+It is in the zip file
+
+### Where is the program located?
+
+Not relevant to my request.
+
+### Expected behavior
+
+The app shouldn't crash.
+
+### What is your Windows edition and version?
+
+Windows 10 Pro 22H2 64-bit
+
+### In which Windows account you have this problem?
+
+User account with UAC protection set to Always notify.
+
+### Please mention any installed security software
+
+Microsoft Windows Defender
+
+### What version of Sandboxie are you running?
+
+Sandboxie-Plus 1.7.0 64-bit
+
+### Is it a new installation of Sandboxie?
+
+I recently did a new clean installation.
+
+### Is it a regression?
+
+Since Sandboxie Plus v1.6.7 64-bit
+
+### In which sandbox type you have this problem?
+
+Not relevant to my request.
+
+### Can you reproduce this problem on a new empty sandbox?
+
+Not relevant to my request.
+
+### Did you previously enable some security policy settings outside Sandboxie?
+
+_No response_
+
+### Crash dump
+
+It is in the zip file
+
+### Trace log
+
+_No response_
+
+### Sandboxie.ini configuration
+
+```shell
+[Global Settings]
+RunCommand=Command Prompt|""C:\Windows\System32\cmd.exe""
+RunCommand=PowerShell|""C:\Windows\System32\WindowsPowerShell\v1.0\powershell.exe""
+RunCommand=PowerShell 7|""C:\Program Files\PowerShell\7\pwsh.exe""
+```
+",0, sandman crashes when hovering over run menu if there are any custom run menu entries are defined and optaneshellext is enabled describe what you noticed and did install intel optane pinning shell extension comes with chipset driver intel rst restart sandman exe add a custom run menu entry sandman sandbox options general options run menu hover over the run menu sanman right click on a box to open sandbox context menu run sandman exe crashes disabling the intel optane pinning shellext optaneiconoverlay fixes the issue with nirsoft shellexview pw is sandboxie how often did you encounter it so far every time affected program not relevant download link it is in the zip file where is the program located not relevant to my request expected behavior the app shouldn t crash what is your windows edition and version windows pro bit in which windows account you have this problem user account with uac protection set to always notify please mention any installed security software microsoft windows defender what version of sandboxie are you running sandboxie plus bit is it a new installation of sandboxie i recently did a new clean installation is it a regression since sandboxie plus bit in which sandbox type you have this problem not relevant to my request can you reproduce this problem on a new empty sandbox not relevant to my request did you previously enable some security policy settings outside sandboxie no response crash dump it is in the zip file trace log no response sandboxie ini configuration shell runcommand command prompt c windows cmd exe runcommand powershell c windows windowspowershell powershell exe runcommand powershell c program files powershell pwsh exe ,0
+1063,25291345448.0,IssuesEvent,2022-11-17 00:39:25,jongfeel/BookReview,https://api.github.com/repos/jongfeel/BookReview,closed,6부 7장 '결혼에 무지한 사람'이 되지 마라,2022 How to Win Friends & Influence People,"### 7장 '결혼에 무지한 사람'이 되지 마라
+
+결혼에 무지한 사람들!
+
+“결혼이라는 대단히 어려운 조정과 적응 과정을 그저 운에 맡기고 있다는 점을 감안해 보면, 이혼률이 16퍼센트밖에 안 된다는 사실이 놀라울 지경이다. 놀라울 정도로 많은 남편과 아내들이 사실상 결혼 생활을 하는 게 아니라 이혼하지 않은 상태를 유지하고 있을 따름이다. 이들은 지옥과 같은 삶을 살고 있다.”
+
+---
+
+비결 7: 결혼의 성적 측면에 관한 좋은 책을 읽어라.
+Read a good book on the sexual side of marriage.",1.0,"6부 7장 '결혼에 무지한 사람'이 되지 마라 - ### 7장 '결혼에 무지한 사람'이 되지 마라
+
+결혼에 무지한 사람들!
+
+“결혼이라는 대단히 어려운 조정과 적응 과정을 그저 운에 맡기고 있다는 점을 감안해 보면, 이혼률이 16퍼센트밖에 안 된다는 사실이 놀라울 지경이다. 놀라울 정도로 많은 남편과 아내들이 사실상 결혼 생활을 하는 게 아니라 이혼하지 않은 상태를 유지하고 있을 따름이다. 이들은 지옥과 같은 삶을 살고 있다.”
+
+---
+
+비결 7: 결혼의 성적 측면에 관한 좋은 책을 읽어라.
+Read a good book on the sexual side of marriage.",1, 결혼에 무지한 사람 이 되지 마라 결혼에 무지한 사람 이 되지 마라 결혼에 무지한 사람들 “결혼이라는 대단히 어려운 조정과 적응 과정을 그저 운에 맡기고 있다는 점을 감안해 보면 이혼률이 안 된다는 사실이 놀라울 지경이다 놀라울 정도로 많은 남편과 아내들이 사실상 결혼 생활을 하는 게 아니라 이혼하지 않은 상태를 유지하고 있을 따름이다 이들은 지옥과 같은 삶을 살고 있다 ” 비결 결혼의 성적 측면에 관한 좋은 책을 읽어라 read a good book on the sexual side of marriage ,1
+769,14100993120.0,IssuesEvent,2020-11-06 05:47:28,planningcenter/developers,https://api.github.com/repos/planningcenter/developers,closed,Creating/Updating Emails,People,"**Related Product**
+Which product is this question related to?
+People API
+
+**Describe the question**
+I'm fairly new to API's so I hope I'm not missing anything.
+I'm able to GET, POST, and PATCH most user data, except for emails. I've checked the [documentations ](https://developer.planning.center/docs/#/apps/people/2020-07-22/vertices/email) and I don't know what I am missing.
+GET emails works fine. Somehow POST and PATCH has issues.
+
+**What have you tried that worked?**
+GET with endpoint ""/people/v2/people/123456/emails"" works fine.
+
+**What have you tried that didn't work?**
+POST from said endpoint brings an error.
+PATCH from ""people/v2/emails/123456"" brings up a different error.
+
+**Additional context**
+
+
+## I have..
+
+- [x] Reviewed the documentation found at https://developer.planning.center/docs
+- [x] Searched for previous issues answering this question
+- [x] Removed all private information from this issue (credentials, tokens, emails, phone numbers, etc.)
+- [x] Reviewed my issue for completeness
+",1.0,"Creating/Updating Emails - **Related Product**
+Which product is this question related to?
+People API
+
+**Describe the question**
+I'm fairly new to API's so I hope I'm not missing anything.
+I'm able to GET, POST, and PATCH most user data, except for emails. I've checked the [documentations ](https://developer.planning.center/docs/#/apps/people/2020-07-22/vertices/email) and I don't know what I am missing.
+GET emails works fine. Somehow POST and PATCH has issues.
+
+**What have you tried that worked?**
+GET with endpoint ""/people/v2/people/123456/emails"" works fine.
+
+**What have you tried that didn't work?**
+POST from said endpoint brings an error.
+PATCH from ""people/v2/emails/123456"" brings up a different error.
+
+**Additional context**
+
+
+## I have..
+
+- [x] Reviewed the documentation found at https://developer.planning.center/docs
+- [x] Searched for previous issues answering this question
+- [x] Removed all private information from this issue (credentials, tokens, emails, phone numbers, etc.)
+- [x] Reviewed my issue for completeness
+",1,creating updating emails related product which product is this question related to people api describe the question i m fairly new to api s so i hope i m not missing anything i m able to get post and patch most user data except for emails i ve checked the and i don t know what i am missing get emails works fine somehow post and patch has issues what have you tried that worked get with endpoint people people emails works fine what have you tried that didn t work post from said endpoint brings an error patch from people emails brings up a different error additional context i have reviewed the documentation found at searched for previous issues answering this question removed all private information from this issue credentials tokens emails phone numbers etc reviewed my issue for completeness ,1
+279217,30702472550.0,IssuesEvent,2023-07-27 01:33:06,nidhi7598/linux-3.0.35_CVE-2018-13405,https://api.github.com/repos/nidhi7598/linux-3.0.35_CVE-2018-13405,closed,CVE-2023-3268 (High) detected in linux-stable-rtv3.8.6 - autoclosed,Mend: dependency security vulnerability,"## CVE-2023-3268 - High Severity Vulnerability
+ Vulnerable Library - linux-stable-rtv3.8.6
+
+
+An out of bounds (OOB) memory access flaw was found in the Linux kernel in relay_file_read_start_pos in kernel/relay.c in the relayfs. This flaw could allow a local attacker to crash the system or leak kernel internal information.
+
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",True,"CVE-2023-3268 (High) detected in linux-stable-rtv3.8.6 - autoclosed - ## CVE-2023-3268 - High Severity Vulnerability
+ Vulnerable Library - linux-stable-rtv3.8.6
+
+
+An out of bounds (OOB) memory access flaw was found in the Linux kernel in relay_file_read_start_pos in kernel/relay.c in the relayfs. This flaw could allow a local attacker to crash the system or leak kernel internal information.
+
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",0,cve high detected in linux stable autoclosed cve high severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files kernel relay c vulnerability details an out of bounds oob memory access flaw was found in the linux kernel in relay file read start pos in kernel relay c in the relayfs this flaw could allow a local attacker to crash the system or leak kernel internal information publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend ,0
+337,6517106777.0,IssuesEvent,2017-08-27 18:45:52,Velosify/gunn-remake,https://api.github.com/repos/Velosify/gunn-remake,opened,Design concept,DISCUSSION TIME FEATURE REQUEST JUST A QUESTION REQUIRES MORE PEOPLE,"So I suppose we're moving on to making la ĉefretpaĝo so we know what we actually need to design.
+
+Features the current Gunn ĉefretpaĝo has that I think we should incorporate into our redesign in some way:
+
+- [ ] image carousel (see #8) which can be that huge image modern sites have for some reason ([example]())
+- [ ] events calendar (it only has to list a few items; there can be a view more option linking to a full-sized calendar)
+- [ ] recent news/announcements
+- [ ] a very easy-to-access ""Crisis Resources"" link",1.0,"Design concept - So I suppose we're moving on to making la ĉefretpaĝo so we know what we actually need to design.
+
+Features the current Gunn ĉefretpaĝo has that I think we should incorporate into our redesign in some way:
+
+- [ ] image carousel (see #8) which can be that huge image modern sites have for some reason ([example]())
+- [ ] events calendar (it only has to list a few items; there can be a view more option linking to a full-sized calendar)
+- [ ] recent news/announcements
+- [ ] a very easy-to-access ""Crisis Resources"" link",1,design concept so i suppose we re moving on to making la ĉefretpaĝo so we know what we actually need to design features the current gunn ĉefretpaĝo has that i think we should incorporate into our redesign in some way image carousel see which can be that huge image modern sites have for some reason events calendar it only has to list a few items there can be a view more option linking to a full sized calendar recent news announcements a very easy to access crisis resources link,1
+965,21575728253.0,IssuesEvent,2022-05-02 13:33:50,AtB-AS/mittatb-app,https://api.github.com/repos/AtB-AS/mittatb-app,opened,"Bytte ut ikon for ""logg ut""-handling under Min profil",:people_holding_hands: Teaminitiativ,"## Origin
+
+I sammenheng med en implementasjon hos Nordland, som baserer seg på vårt grunnlag, kom det tilbakemelding som var overførbar for vår del også, hvor ikonbruken ikke stemmer overens med handlingen.
+
+
+
+## Motivation
+
+Samkjøre designet mellom web og mobil, i tillegg til å tilrettelegge for intuitiv bruk av ikoner.
+
+## Proposed solution
+
+- [ ] Sørg for at ikonet er hvitt på mørk bakgrunn, og svart på lys bakgrunn
+- [ ] Bytt ut ikonet som er brukt i sammenheng med ""Logg ut / Sign out"" fra pila som er der nå, til [""Log out""-ikonet](https://www.figma.com/file/e8MWjmPKrisUDbT1zc0fvv/Icons-1.0?node-id=1141%3A29) som blant annet er brukt i nettbutikken tidligere:
+
+
+
+## Acceptance Criteria
+
+_List of relevant acceptance criteria as a part of a QA flow_
+
+- [] Renders colors as expected in dark and light mode.
+- [] Is translatable to English and Norwegian.
+- [] Does not drain battery
+- [] Works in up to 200% font size
+- [] Is operable and perceivable using screen reader
+
+Beskrivelse av QA prosessen finnes [her](https://github.com/AtB-AS/org/blob/master/guides/quality_assurance.md#qa-in-atb-mobile-application)
+",1.0,"Bytte ut ikon for ""logg ut""-handling under Min profil - ## Origin
+
+I sammenheng med en implementasjon hos Nordland, som baserer seg på vårt grunnlag, kom det tilbakemelding som var overførbar for vår del også, hvor ikonbruken ikke stemmer overens med handlingen.
+
+
+
+## Motivation
+
+Samkjøre designet mellom web og mobil, i tillegg til å tilrettelegge for intuitiv bruk av ikoner.
+
+## Proposed solution
+
+- [ ] Sørg for at ikonet er hvitt på mørk bakgrunn, og svart på lys bakgrunn
+- [ ] Bytt ut ikonet som er brukt i sammenheng med ""Logg ut / Sign out"" fra pila som er der nå, til [""Log out""-ikonet](https://www.figma.com/file/e8MWjmPKrisUDbT1zc0fvv/Icons-1.0?node-id=1141%3A29) som blant annet er brukt i nettbutikken tidligere:
+
+
+
+## Acceptance Criteria
+
+_List of relevant acceptance criteria as a part of a QA flow_
+
+- [] Renders colors as expected in dark and light mode.
+- [] Is translatable to English and Norwegian.
+- [] Does not drain battery
+- [] Works in up to 200% font size
+- [] Is operable and perceivable using screen reader
+
+Beskrivelse av QA prosessen finnes [her](https://github.com/AtB-AS/org/blob/master/guides/quality_assurance.md#qa-in-atb-mobile-application)
+",1,bytte ut ikon for logg ut handling under min profil origin i sammenheng med en implementasjon hos nordland som baserer seg på vårt grunnlag kom det tilbakemelding som var overførbar for vår del også hvor ikonbruken ikke stemmer overens med handlingen motivation samkjøre designet mellom web og mobil i tillegg til å tilrettelegge for intuitiv bruk av ikoner proposed solution sørg for at ikonet er hvitt på mørk bakgrunn og svart på lys bakgrunn bytt ut ikonet som er brukt i sammenheng med logg ut sign out fra pila som er der nå til som blant annet er brukt i nettbutikken tidligere acceptance criteria list of relevant acceptance criteria as a part of a qa flow renders colors as expected in dark and light mode is translatable to english and norwegian does not drain battery works in up to font size is operable and perceivable using screen reader beskrivelse av qa prosessen finnes ,1
+492,8452237831.0,IssuesEvent,2018-10-20 01:18:19,openstates/openstates,https://api.github.com/repos/openstates/openstates,closed,Wyoming Data Dupes / Out of date records,people,"State: Wyoming
+
+There are a handful of duplicate records in Wyoming Data and an extraneous out-of-date record as well.
+
+[R.] Ray Peterson: WYL000025 / WYL000171
+[James Lee / Jim] Anderson: WYL000093 / WYL000168
+[Timothy / Tim] Hallinan: WYL000159 / WYL000170
+[Michael / Mike] Madden: WYL000068 / WYL000167
+[James / Jim] Byrd: WYL000041 / WYL000169
+
+Out of date Record:
+Lars Lone (Resigned): WYL000138
+",1.0,"Wyoming Data Dupes / Out of date records - State: Wyoming
+
+There are a handful of duplicate records in Wyoming Data and an extraneous out-of-date record as well.
+
+[R.] Ray Peterson: WYL000025 / WYL000171
+[James Lee / Jim] Anderson: WYL000093 / WYL000168
+[Timothy / Tim] Hallinan: WYL000159 / WYL000170
+[Michael / Mike] Madden: WYL000068 / WYL000167
+[James / Jim] Byrd: WYL000041 / WYL000169
+
+Out of date Record:
+Lars Lone (Resigned): WYL000138
+",1,wyoming data dupes out of date records state wyoming there are a handful of duplicate records in wyoming data and an extraneous out of date record as well ray peterson anderson hallinan madden byrd out of date record lars lone resigned ,1
+1134,27713018403.0,IssuesEvent,2023-03-14 15:18:31,culturesofknowledge/emlo-project,https://api.github.com/repos/culturesofknowledge/emlo-project,closed,People search: Choosing page 2 looses my search,bug people search AR test,"I searched for `{Born between 1467 and 1598.}` in the people search form.
+
+When I click on page 2 in the results, this is forgotten.",1.0,"People search: Choosing page 2 looses my search - I searched for `{Born between 1467 and 1598.}` in the people search form.
+
+When I click on page 2 in the results, this is forgotten.",1,people search choosing page looses my search i searched for born between and in the people search form when i click on page in the results this is forgotten ,1
+82,3361943887.0,IssuesEvent,2015-11-20 01:17:23,bbergen/tbmd.com,https://api.github.com/repos/bbergen/tbmd.com,opened,Client Side File Size Validation,bug Movies People,"Currently the website will let you try to upload very large images for people and movies.
+
+If the file size is greater that that allowed in php.ini however, an exception is thrown and obviously the image is not uploaded properly.
+
+Acceptance Criteria:
+
+* Client Side validation (in javascript) is done on the Movie and Person forms that check for a maximum file size.
+* The Uploader validates the file has actually been uploaded before acting on it to prevent the exception. ",1.0,"Client Side File Size Validation - Currently the website will let you try to upload very large images for people and movies.
+
+If the file size is greater that that allowed in php.ini however, an exception is thrown and obviously the image is not uploaded properly.
+
+Acceptance Criteria:
+
+* Client Side validation (in javascript) is done on the Movie and Person forms that check for a maximum file size.
+* The Uploader validates the file has actually been uploaded before acting on it to prevent the exception. ",1,client side file size validation currently the website will let you try to upload very large images for people and movies if the file size is greater that that allowed in php ini however an exception is thrown and obviously the image is not uploaded properly acceptance criteria client side validation in javascript is done on the movie and person forms that check for a maximum file size the uploader validates the file has actually been uploaded before acting on it to prevent the exception ,1
+841,15766517636.0,IssuesEvent,2021-03-31 15:08:50,openstates/issues,https://api.github.com/repos/openstates/issues,closed,Montana Legislators need a full update,component:people-data type:bug,"State: MT
+Short Description: When checking Montana’s legislators, I found several legislators with incorrectly linked addresses, mismatched legislators, and missing contact information
+
+Include these 3 sections as appropriate (or note none)
+
+Missing or Incorrect legislators:
+ (include vacancies, mismatched legislators, etc.)
+There are legislators that are listed on OS, that is not on the legislator website. I’ve noticed that these legislators are done with their service. They are the following:
+ [1] ""Alan Doane"" ""Alan Redfield"" ""Albert Olszewski""
+ [4] ""Barbara Bessette"" ""Bradley Maxon Hamlett"" ""Bridget Smith""
+ [7] ""Bruce Grubbs"" ""Casey Schreiner"" ""David Dunn""
+[10] ""Debo Powers"" ""Dee L Brown"" ""Dick Barrett""
+[13] ""Forrest Mandeville"" ""Fred Thomas"" ""Frederick (Eric) Moore""
+[16] ""Gene Vuckovich"" ""Gordon Pierson"" ""Greg Devries""
+[19] ""Jacob Bachmeier"" ""Jade Bahr"" ""Jasmine Krotkov""
+[22] ""Jennifer Fielder"" ""Joel Krautter"" ""Jon Sesso""
+[25] ""Kerry White"" ""Kimberly Dudik"" ""Margaret (Margie) Macdonald""
+[28] ""Marilyn Ryan"" ""Mary McNally"" ""Mike Phillips""
+[31] ""Nancy Ballance"" ""Peggy Webb"" ""Rae Peppers""
+[34] ""Ray Shaw"" ""Rodney Garcia"" ""Roger Webb""
+[37] ""Scott Sales"" ""Sharon Stewart-Peregoy"" ""Sue Malek""
+[40] ""Tom Burnett"" ""Tom Richmond"" ""Tom Winter""
+[43] ""Tom Woods"" ""Zach Brown""
+
+Example Tom Woods:
+
+
+
+
+There are also legislators that are listed on the legislative website, that is not on OS. I’ve noticed that these legislators just started their service in 2021.
+
+""Marta Bertoglio"" ""Michele Binkley"" ""Ellie Boldman""
+ [4] ""Alice Buckley"" ""Jennifer Carlson"" ""Paul Fielder""
+ [7] ""Mike Fox"" ""Tom France"" ""Gregory Frazer""
+[10] ""Chris Friedel"" ""Steven Galloway"" ""Jane Gillette""
+[13] ""Steve Gist"" ""Donavon Hawk"" ""Ed Hill""
+[16] ""Jedediah Hinkle"" ""Caleb Hinkle"" ""Scot Kerns""
+[19] ""Kelly Kortum"" ""Brandon Ler"" ""Marty Malone""
+[22] ""Ron Marshall"" ""Tom Mcgillvray"" ""Mary Mcnally""
+[25] ""Braxton Mitchell"" ""Brad Molnar"" ""Fiona Nave""
+[28] ""Sara Novak"" ""Shannon O'brien"" ""Jimmy Patelis""
+[31] ""Bob Phalen"" ""Brian Putnam"" ""Amy Regier""
+[34] ""Linda Reksten"" ""Jerry Schillinger"" ""Kerri Seekins-Crowe""
+[37] ""Ed Stafman"" ""Sharon Stewart Peregoy"" ""Mallerie Stromswold""
+[40] ""Mark Thane"" ""Jeremy Trebas"" ""Kenneth Walsh""
+[43] ""Rynalea Whiteman Pena"" ""Kathy Whitman""
+
+Example Kathy Whitman:
+
+
+
+
+Legislative Website Used To Search: https://leg.mt.gov/legislator-information/?session_select=116
+OS Link Used To Search: https://openstates.org/mt/legislators/
+
+
+Data Issues:
+(issues with data that is present, give at least 2 examples if possible, but no need to list all of them if issue looks systemic)
+For some legislators, the link on OS does not show the same address listed on OS. The address listed on OS is in a different link
+
+Christopher Pope
+OS Link: https://openstates.org/person/christopher-pope-3ijCIAT5LpYCe6JGb6q5PS/
+The link for the address that shows on OS: https://uat.leg.mt.gov/legislator-information/roster/individual/5124.
+The actual link on OS: https://leg.mt.gov/legislator-information/roster/individual/5124
+Photos:
+
+
+
+
+
+Shane Morigeau
+OS Link: https://openstates.org/person/shane-morigeau-40zxzjDD1nUC6kL0x8EwZs/
+The link for the address that shows on OS: https://uat.leg.mt.gov/legislator-information/roster/individual/9550.
+The actual link on OS: https://leg.mt.gov/legislator-information/roster/individual/4837
+Photos:
+
+
+
+
+Additional Data:
+(note missing contact information, social media, photos, biographical information, etc. with 1-2 examples)
+(no need to mention: campaign information, committees, sponsored bills, other related information)
+Quite a few legislators have secondary phone numbers listed:
+Becky Beard https://leg.mt.gov/legislator-information/roster/individual/5139
+Willis Curdy https://leg.mt.gov/legislator-information/roster/individual/5157",1.0,"Montana Legislators need a full update - State: MT
+Short Description: When checking Montana’s legislators, I found several legislators with incorrectly linked addresses, mismatched legislators, and missing contact information
+
+Include these 3 sections as appropriate (or note none)
+
+Missing or Incorrect legislators:
+ (include vacancies, mismatched legislators, etc.)
+There are legislators that are listed on OS, that is not on the legislator website. I’ve noticed that these legislators are done with their service. They are the following:
+ [1] ""Alan Doane"" ""Alan Redfield"" ""Albert Olszewski""
+ [4] ""Barbara Bessette"" ""Bradley Maxon Hamlett"" ""Bridget Smith""
+ [7] ""Bruce Grubbs"" ""Casey Schreiner"" ""David Dunn""
+[10] ""Debo Powers"" ""Dee L Brown"" ""Dick Barrett""
+[13] ""Forrest Mandeville"" ""Fred Thomas"" ""Frederick (Eric) Moore""
+[16] ""Gene Vuckovich"" ""Gordon Pierson"" ""Greg Devries""
+[19] ""Jacob Bachmeier"" ""Jade Bahr"" ""Jasmine Krotkov""
+[22] ""Jennifer Fielder"" ""Joel Krautter"" ""Jon Sesso""
+[25] ""Kerry White"" ""Kimberly Dudik"" ""Margaret (Margie) Macdonald""
+[28] ""Marilyn Ryan"" ""Mary McNally"" ""Mike Phillips""
+[31] ""Nancy Ballance"" ""Peggy Webb"" ""Rae Peppers""
+[34] ""Ray Shaw"" ""Rodney Garcia"" ""Roger Webb""
+[37] ""Scott Sales"" ""Sharon Stewart-Peregoy"" ""Sue Malek""
+[40] ""Tom Burnett"" ""Tom Richmond"" ""Tom Winter""
+[43] ""Tom Woods"" ""Zach Brown""
+
+Example Tom Woods:
+
+
+
+
+There are also legislators that are listed on the legislative website, that is not on OS. I’ve noticed that these legislators just started their service in 2021.
+
+""Marta Bertoglio"" ""Michele Binkley"" ""Ellie Boldman""
+ [4] ""Alice Buckley"" ""Jennifer Carlson"" ""Paul Fielder""
+ [7] ""Mike Fox"" ""Tom France"" ""Gregory Frazer""
+[10] ""Chris Friedel"" ""Steven Galloway"" ""Jane Gillette""
+[13] ""Steve Gist"" ""Donavon Hawk"" ""Ed Hill""
+[16] ""Jedediah Hinkle"" ""Caleb Hinkle"" ""Scot Kerns""
+[19] ""Kelly Kortum"" ""Brandon Ler"" ""Marty Malone""
+[22] ""Ron Marshall"" ""Tom Mcgillvray"" ""Mary Mcnally""
+[25] ""Braxton Mitchell"" ""Brad Molnar"" ""Fiona Nave""
+[28] ""Sara Novak"" ""Shannon O'brien"" ""Jimmy Patelis""
+[31] ""Bob Phalen"" ""Brian Putnam"" ""Amy Regier""
+[34] ""Linda Reksten"" ""Jerry Schillinger"" ""Kerri Seekins-Crowe""
+[37] ""Ed Stafman"" ""Sharon Stewart Peregoy"" ""Mallerie Stromswold""
+[40] ""Mark Thane"" ""Jeremy Trebas"" ""Kenneth Walsh""
+[43] ""Rynalea Whiteman Pena"" ""Kathy Whitman""
+
+Example Kathy Whitman:
+
+
+
+
+Legislative Website Used To Search: https://leg.mt.gov/legislator-information/?session_select=116
+OS Link Used To Search: https://openstates.org/mt/legislators/
+
+
+Data Issues:
+(issues with data that is present, give at least 2 examples if possible, but no need to list all of them if issue looks systemic)
+For some legislators, the link on OS does not show the same address listed on OS. The address listed on OS is in a different link
+
+Christopher Pope
+OS Link: https://openstates.org/person/christopher-pope-3ijCIAT5LpYCe6JGb6q5PS/
+The link for the address that shows on OS: https://uat.leg.mt.gov/legislator-information/roster/individual/5124.
+The actual link on OS: https://leg.mt.gov/legislator-information/roster/individual/5124
+Photos:
+
+
+
+
+
+Shane Morigeau
+OS Link: https://openstates.org/person/shane-morigeau-40zxzjDD1nUC6kL0x8EwZs/
+The link for the address that shows on OS: https://uat.leg.mt.gov/legislator-information/roster/individual/9550.
+The actual link on OS: https://leg.mt.gov/legislator-information/roster/individual/4837
+Photos:
+
+
+
+
+Additional Data:
+(note missing contact information, social media, photos, biographical information, etc. with 1-2 examples)
+(no need to mention: campaign information, committees, sponsored bills, other related information)
+Quite a few legislators have secondary phone numbers listed:
+Becky Beard https://leg.mt.gov/legislator-information/roster/individual/5139
+Willis Curdy https://leg.mt.gov/legislator-information/roster/individual/5157",1,montana legislators need a full update state mt short description when checking montana’s legislators i found several legislators with incorrectly linked addresses mismatched legislators and missing contact information include these sections as appropriate or note none missing or incorrect legislators include vacancies mismatched legislators etc there are legislators that are listed on os that is not on the legislator website i’ve noticed that these legislators are done with their service they are the following alan doane alan redfield albert olszewski barbara bessette bradley maxon hamlett bridget smith bruce grubbs casey schreiner david dunn debo powers dee l brown dick barrett forrest mandeville fred thomas frederick eric moore gene vuckovich gordon pierson greg devries jacob bachmeier jade bahr jasmine krotkov jennifer fielder joel krautter jon sesso kerry white kimberly dudik margaret margie macdonald marilyn ryan mary mcnally mike phillips nancy ballance peggy webb rae peppers ray shaw rodney garcia roger webb scott sales sharon stewart peregoy sue malek tom burnett tom richmond tom winter tom woods zach brown example tom woods there are also legislators that are listed on the legislative website that is not on os i’ve noticed that these legislators just started their service in marta bertoglio michele binkley ellie boldman alice buckley jennifer carlson paul fielder mike fox tom france gregory frazer chris friedel steven galloway jane gillette steve gist donavon hawk ed hill jedediah hinkle caleb hinkle scot kerns kelly kortum brandon ler marty malone ron marshall tom mcgillvray mary mcnally braxton mitchell brad molnar fiona nave sara novak shannon o brien jimmy patelis bob phalen brian putnam amy regier linda reksten jerry schillinger kerri seekins crowe ed stafman sharon stewart peregoy mallerie stromswold mark thane jeremy trebas kenneth walsh rynalea whiteman pena kathy whitman example kathy whitman legislative website used to search os link used to search data issues issues with data that is present give at least examples if possible but no need to list all of them if issue looks systemic for some legislators the link on os does not show the same address listed on os the address listed on os is in a different link christopher pope os link the link for the address that shows on os the actual link on os photos shane morigeau os link the link for the address that shows on os the actual link on os photos additional data note missing contact information social media photos biographical information etc with examples no need to mention campaign information committees sponsored bills other related information quite a few legislators have secondary phone numbers listed becky beard willis curdy ,1
+1168,30631051372.0,IssuesEvent,2023-07-24 14:33:54,restincode/restincode,https://api.github.com/repos/restincode/restincode,opened,Michael Park (Hosaka),People Add Person Needs Review,"Please fill out as much information as you can. No fields are required, but the more you can provide the better.
+
+**General Info**
+* First name: Michael
+* Last name: Park
+* Handle: Hosaka
+* Birth Year:
+* Death Year: 2023
+* Link to Obituary:
+* Group Affiliations: r00t
+* URL to main photo (or attach to Issue):
+* Description of person and/or activities:
+* Facebook memorial group URL:
+
+**Social Media Links**
+* Twitter:
+* Github:
+* LinkedIn:
+* Facebook:
+* Other:
+
+**Contributions**
+Repeat as many times as needed
+
+* Project name:
+* Project URL:
+* Project Description:
+
+**Photo Gallery**
+* URL(s) to additional photos:
+",1.0,"Michael Park (Hosaka) - Please fill out as much information as you can. No fields are required, but the more you can provide the better.
+
+**General Info**
+* First name: Michael
+* Last name: Park
+* Handle: Hosaka
+* Birth Year:
+* Death Year: 2023
+* Link to Obituary:
+* Group Affiliations: r00t
+* URL to main photo (or attach to Issue):
+* Description of person and/or activities:
+* Facebook memorial group URL:
+
+**Social Media Links**
+* Twitter:
+* Github:
+* LinkedIn:
+* Facebook:
+* Other:
+
+**Contributions**
+Repeat as many times as needed
+
+* Project name:
+* Project URL:
+* Project Description:
+
+**Photo Gallery**
+* URL(s) to additional photos:
+",1,michael park hosaka please fill out as much information as you can no fields are required but the more you can provide the better general info first name michael last name park handle hosaka birth year death year link to obituary group affiliations url to main photo or attach to issue description of person and or activities facebook memorial group url social media links twitter github linkedin facebook other contributions repeat as many times as needed project name project url project description photo gallery url s to additional photos ,1
+262918,27989511462.0,IssuesEvent,2023-03-27 01:37:46,Satheesh575555/frameworks_base_AOSP10_r33_CVE-2020-0401,https://api.github.com/repos/Satheesh575555/frameworks_base_AOSP10_r33_CVE-2020-0401,opened,CVE-2023-21026 (Medium) detected in baseandroid-10.0.0_r34,Mend: dependency security vulnerability,"## CVE-2023-21026 - Medium Severity Vulnerability
+ Vulnerable Library - baseandroid-10.0.0_r34
+
+
+In updateInputChannel of WindowManagerService.java, there is a possible way to set a touchable region beyond its own SurfaceControl due to a logic error in the code. This could lead to local denial of service with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-13Android ID: A-254681548
+
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",True,"CVE-2023-21026 (Medium) detected in baseandroid-10.0.0_r34 - ## CVE-2023-21026 - Medium Severity Vulnerability
+ Vulnerable Library - baseandroid-10.0.0_r34
+
+
+In updateInputChannel of WindowManagerService.java, there is a possible way to set a touchable region beyond its own SurfaceControl due to a logic error in the code. This could lead to local denial of service with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-13Android ID: A-254681548
+
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",0,cve medium detected in baseandroid cve medium severity vulnerability vulnerable library baseandroid android framework classes and services library home page a href found in head commit a href found in base branch master vulnerable source files services core java com android server wm windowmanagerservice java vulnerability details in updateinputchannel of windowmanagerservice java there is a possible way to set a touchable region beyond its own surfacecontrol due to a logic error in the code this could lead to local denial of service with no additional execution privileges needed user interaction is not needed for exploitation product androidversions android id a publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution android step up your open source security game with mend ,0
+299372,22601499536.0,IssuesEvent,2022-06-29 09:31:20,notDmDrl/pull_down_button,https://api.github.com/repos/notDmDrl/pull_down_button,closed,"If `PullDownMenuItem`'s `onTap` is a `showDialog` function, nothing happens when clicking it.",documentation enhancement,"```dart
+PullDownButton(
+ itemBuilder: (ctx) => [
+ PullDownMenuItem(
+ title: 'Compress',
+ icon: CupertinoIcons.archivebox,
+ onTap: () => showDialog(context: context, builder: (context) => AlertDialog(content: Container())),
+ ),
+ ],
+ position: PullDownMenuPosition.under,
+ buttonBuilder: (context, showMenu) => CupertinoButton(
+ onPressed: showMenu,
+ padding: EdgeInsets.zero,
+ child: Icon(CupertinoIcons.ellipsis_circle),
+ ),
+ ),
+```",1.0,"If `PullDownMenuItem`'s `onTap` is a `showDialog` function, nothing happens when clicking it. - ```dart
+PullDownButton(
+ itemBuilder: (ctx) => [
+ PullDownMenuItem(
+ title: 'Compress',
+ icon: CupertinoIcons.archivebox,
+ onTap: () => showDialog(context: context, builder: (context) => AlertDialog(content: Container())),
+ ),
+ ],
+ position: PullDownMenuPosition.under,
+ buttonBuilder: (context, showMenu) => CupertinoButton(
+ onPressed: showMenu,
+ padding: EdgeInsets.zero,
+ child: Icon(CupertinoIcons.ellipsis_circle),
+ ),
+ ),
+```",0,if pulldownmenuitem s ontap is a showdialog function nothing happens when clicking it dart pulldownbutton itembuilder ctx pulldownmenuitem title compress icon cupertinoicons archivebox ontap showdialog context context builder context alertdialog content container position pulldownmenuposition under buttonbuilder context showmenu cupertinobutton onpressed showmenu padding edgeinsets zero child icon cupertinoicons ellipsis circle ,0
+434,7978331508.0,IssuesEvent,2018-07-17 17:58:59,planningcenter/developers,https://api.github.com/repos/planningcenter/developers,closed,Impossible to filter person_mergers endpoint by date (which makes it rather difficult to use),People enhancement,"### Detailed Description of the Problem/Question
+
+The data available/queryable at the person_mergers endpoint is not very useful because in order to discern when new merges have occurred for a church one would need to repeatedly download the entire list of merge records (by paging the person_mergers endpoint) and diff the new list from the previous list (wasteful for obvious reasons.) since there is no way to filter the historical records by date.
+
+Would you guys be interested in adding queryable created_at/updated_at information to the person_merger endpoint to remedy this situation? This would be super helpful!
+
+
+
+
+
+##### Steps to reproduce:
+Run a GET request against people/v2/person_mergers
+
+
+##### API endpoint I'm using:
+people/v2/person_mergers
+
+
+##### Programming language I'm using:
+C#
+
+
+##### Authentication method I'm using:
+Personal Access Token
+e.g. OAuth 2, Personal Access Token, Browser Session (Cookie)
+",1.0,"Impossible to filter person_mergers endpoint by date (which makes it rather difficult to use) - ### Detailed Description of the Problem/Question
+
+The data available/queryable at the person_mergers endpoint is not very useful because in order to discern when new merges have occurred for a church one would need to repeatedly download the entire list of merge records (by paging the person_mergers endpoint) and diff the new list from the previous list (wasteful for obvious reasons.) since there is no way to filter the historical records by date.
+
+Would you guys be interested in adding queryable created_at/updated_at information to the person_merger endpoint to remedy this situation? This would be super helpful!
+
+
+
+
+
+##### Steps to reproduce:
+Run a GET request against people/v2/person_mergers
+
+
+##### API endpoint I'm using:
+people/v2/person_mergers
+
+
+##### Programming language I'm using:
+C#
+
+
+##### Authentication method I'm using:
+Personal Access Token
+e.g. OAuth 2, Personal Access Token, Browser Session (Cookie)
+",1,impossible to filter person mergers endpoint by date which makes it rather difficult to use detailed description of the problem question the data available queryable at the person mergers endpoint is not very useful because in order to discern when new merges have occurred for a church one would need to repeatedly download the entire list of merge records by paging the person mergers endpoint and diff the new list from the previous list wasteful for obvious reasons since there is no way to filter the historical records by date would you guys be interested in adding queryable created at updated at information to the person merger endpoint to remedy this situation this would be super helpful steps to reproduce run a get request against people person mergers api endpoint i m using people person mergers programming language i m using c authentication method i m using personal access token e g oauth personal access token browser session cookie ,1
+414930,12121569710.0,IssuesEvent,2020-04-22 09:30:56,wso2/product-microgateway,https://api.github.com/repos/wso2/product-microgateway,closed,Redeclared symbol error when APIs with basic auth secured endpoints,Priority/Normal Type/Bug,"### Description:
+
+
+### Steps to reproduce:
+init a project
+Copy 2 API definitions which has production/sandbox endpoints secured with basic auth to the project.
+build the project.
+
+Error:
+```
+Compiling source
+ wso2/dev:3.1.0
+error: wso2/dev:3.1.0::Basic_Auth__2_0_0.bal:12:32: redeclared symbol 'prod_OutboundBasicAuthProvider'
+error: wso2/dev:3.1.0::Basic_Auth__2_0_0.bal:17:23: redeclared symbol '_prodOutboundBasicAuthHandler'
+error: wso2/dev:3.1.0::Basic_Auth__2_0_0.bal:41:32: redeclared symbol 'sand_OutboundBasicAuthProvider'
+error: wso2/dev:3.1.0::Basic_Auth__2_0_0.bal:46:23: redeclared symbol '_sandOutboundBasicAuthHandler'
+```
+
+caused by lines - variable naming in https://github.com/wso2/product-microgateway/blob/60288b120106d07dec210c170ca7a8b02791ad65/components/micro-gateway-cli/src/main/resources/templates/basicAuth.mustache#L3, https://github.com/wso2/product-microgateway/blob/60288b120106d07dec210c170ca7a8b02791ad65/components/micro-gateway-cli/src/main/resources/templates/basicAuthHandler.mustache#L2
+### Affected Product Version:
+
+3.1.0
+
+### Environment details (with versions):
+- OS:
+- Client:
+- Env (Docker/K8s):
+
+---
+### Optional Fields
+#### Related Issues:
+
+
+#### Suggested Labels:
+
+
+#### Suggested Assignees:
+
+",1.0,"Redeclared symbol error when APIs with basic auth secured endpoints - ### Description:
+
+
+### Steps to reproduce:
+init a project
+Copy 2 API definitions which has production/sandbox endpoints secured with basic auth to the project.
+build the project.
+
+Error:
+```
+Compiling source
+ wso2/dev:3.1.0
+error: wso2/dev:3.1.0::Basic_Auth__2_0_0.bal:12:32: redeclared symbol 'prod_OutboundBasicAuthProvider'
+error: wso2/dev:3.1.0::Basic_Auth__2_0_0.bal:17:23: redeclared symbol '_prodOutboundBasicAuthHandler'
+error: wso2/dev:3.1.0::Basic_Auth__2_0_0.bal:41:32: redeclared symbol 'sand_OutboundBasicAuthProvider'
+error: wso2/dev:3.1.0::Basic_Auth__2_0_0.bal:46:23: redeclared symbol '_sandOutboundBasicAuthHandler'
+```
+
+caused by lines - variable naming in https://github.com/wso2/product-microgateway/blob/60288b120106d07dec210c170ca7a8b02791ad65/components/micro-gateway-cli/src/main/resources/templates/basicAuth.mustache#L3, https://github.com/wso2/product-microgateway/blob/60288b120106d07dec210c170ca7a8b02791ad65/components/micro-gateway-cli/src/main/resources/templates/basicAuthHandler.mustache#L2
+### Affected Product Version:
+
+3.1.0
+
+### Environment details (with versions):
+- OS:
+- Client:
+- Env (Docker/K8s):
+
+---
+### Optional Fields
+#### Related Issues:
+
+
+#### Suggested Labels:
+
+
+#### Suggested Assignees:
+
+",0,redeclared symbol error when apis with basic auth secured endpoints description steps to reproduce init a project copy api definitions which has production sandbox endpoints secured with basic auth to the project build the project error compiling source dev error dev basic auth bal redeclared symbol prod outboundbasicauthprovider error dev basic auth bal redeclared symbol prodoutboundbasicauthhandler error dev basic auth bal redeclared symbol sand outboundbasicauthprovider error dev basic auth bal redeclared symbol sandoutboundbasicauthhandler caused by lines variable naming in affected product version environment details with versions os client env docker optional fields related issues suggested labels suggested assignees ,0
+866,16601196791.0,IssuesEvent,2021-06-01 19:43:04,openstates/issues,https://api.github.com/repos/openstates/issues,closed,Nevada Dead Links,component:people-data type:bug,"**All the links for all members (senate and house) lead to an error page that says ""Sorry, an error occurred while processing your request**
+
+
+**https://openstates.org/person/chris-brooks-1Vijktsqg9OHK2VXUeQrAt/
+https://openstates.org/person/dallas-harris-5D1vWpMHqCzF6cyC48jcfG/
+https://openstates.org/person/heidi-kasama-2xS7mD45WRBqKIwBAXtuRC/**
+
+
+**Source URL of correct data if applicable:**
+",1.0,"Nevada Dead Links - **All the links for all members (senate and house) lead to an error page that says ""Sorry, an error occurred while processing your request**
+
+
+**https://openstates.org/person/chris-brooks-1Vijktsqg9OHK2VXUeQrAt/
+https://openstates.org/person/dallas-harris-5D1vWpMHqCzF6cyC48jcfG/
+https://openstates.org/person/heidi-kasama-2xS7mD45WRBqKIwBAXtuRC/**
+
+
+**Source URL of correct data if applicable:**
+",1,nevada dead links all the links for all members senate and house lead to an error page that says sorry an error occurred while processing your request source url of correct data if applicable ,1
+111023,4448681194.0,IssuesEvent,2016-08-22 01:36:16,docker/docker,https://api.github.com/repos/docker/docker,closed,Cannot add an additional swarm manager,area/swarm priority/P2 version/1.12,"
+
+**Output of `docker version`:**
+
+```
+Client:
+ Version: 1.12.0-rc2
+ API version: 1.24
+ Go version: go1.6.2
+ Git commit: 906eacd
+ Built: Fri Jun 17 20:35:33 2016
+ OS/Arch: linux/amd64
+ Experimental: true
+
+Server:
+ Version: 1.12.0-rc2
+ API version: 1.24
+ Go version: go1.6.2
+ Git commit: 906eacd
+ Built: Fri Jun 17 20:35:33 2016
+ OS/Arch: linux/amd64
+ Experimental: true
+```
+
+
+**Output of `docker info`:**
+
+```
+Containers: 0
+ Running: 0
+ Paused: 0
+ Stopped: 0
+Images: 0
+Server Version: 1.12.0-rc2
+Storage Driver: aufs
+ Root Dir: /mnt/sda1/var/lib/docker/aufs
+ Backing Filesystem: extfs
+ Dirs: 0
+ Dirperm1 Supported: true
+Logging Driver: json-file
+Cgroup Driver: cgroupfs
+Plugins:
+ Volume: local
+ Network: host bridge overlay null
+Swarm: inactive
+Runtimes: default
+Default Runtime: default
+Security Options: seccomp
+Kernel Version: 4.4.13-boot2docker
+Operating System: Boot2Docker 1.12.0-rc2 (TCL 7.1); HEAD : 52952ef - Fri Jun 17 21:01:09 UTC 2016
+OSType: linux
+Architecture: x86_64
+CPUs: 1
+Total Memory: 995.9 MiB
+Name: manager1
+ID: XP2B:EU32:5OYY:NBOD:EVKD:K344:3DG7:3YSE:E6UI:Q3Y6:IWOR:MCBS
+Docker Root Dir: /mnt/sda1/var/lib/docker
+Debug Mode (client): false
+Debug Mode (server): true
+ File Descriptors: 17
+ Goroutines: 23
+ System Time: 2016-06-24T10:09:37.185763563Z
+ EventsListeners: 0
+Registry: https://index.docker.io/v1/
+Labels:
+ provider=virtualbox
+Experimental: true
+Insecure Registries:
+ 127.0.0.0/8
+```
+
+**Additional environment details (AWS, VirtualBox, physical, etc.):**
+
+3 virtualbox machines with docker 1.12-dev created by docker-machine
+
+**Steps to reproduce the issue:**
+
+1. Create three machines.
+2. Create a swarm cluster and add an additional swarm manager
+
+ ```
+docker-machine ssh manager1 'docker swarm init'
+docker-machine ssh node1 'docker swarm join 192.168.99.106:2377'
+docker-machine ssh manager2 'docker swarm join --manager 192.168.99.106:2377'
+```
+3. Run `docker-machine ssh manager1 'docker node ls'` to list the nodes
+
+**Describe the results you received:**
+
+The status of new manager is ""Unknown""
+
+```
+ID NAME MEMBERSHIP STATUS AVAILABILITY MANAGER STATUS
+10ztstjk5dt6yb5ikm1yqrben Accepted Unknown Active Reachable
+2x91zrw9n5rsvavisiezb1flp node1 Accepted Ready Active
+d728vzs91jfn4ctw8cx9yjs1h * manager1 Accepted Ready Active Leader
+```
+
+**Describe the results you expected:**
+
+All nodes are ready.
+
+
+**Additional information you deem important (e.g. issue happens only occasionally):**
+",1.0,"Cannot add an additional swarm manager -
+
+**Output of `docker version`:**
+
+```
+Client:
+ Version: 1.12.0-rc2
+ API version: 1.24
+ Go version: go1.6.2
+ Git commit: 906eacd
+ Built: Fri Jun 17 20:35:33 2016
+ OS/Arch: linux/amd64
+ Experimental: true
+
+Server:
+ Version: 1.12.0-rc2
+ API version: 1.24
+ Go version: go1.6.2
+ Git commit: 906eacd
+ Built: Fri Jun 17 20:35:33 2016
+ OS/Arch: linux/amd64
+ Experimental: true
+```
+
+
+**Output of `docker info`:**
+
+```
+Containers: 0
+ Running: 0
+ Paused: 0
+ Stopped: 0
+Images: 0
+Server Version: 1.12.0-rc2
+Storage Driver: aufs
+ Root Dir: /mnt/sda1/var/lib/docker/aufs
+ Backing Filesystem: extfs
+ Dirs: 0
+ Dirperm1 Supported: true
+Logging Driver: json-file
+Cgroup Driver: cgroupfs
+Plugins:
+ Volume: local
+ Network: host bridge overlay null
+Swarm: inactive
+Runtimes: default
+Default Runtime: default
+Security Options: seccomp
+Kernel Version: 4.4.13-boot2docker
+Operating System: Boot2Docker 1.12.0-rc2 (TCL 7.1); HEAD : 52952ef - Fri Jun 17 21:01:09 UTC 2016
+OSType: linux
+Architecture: x86_64
+CPUs: 1
+Total Memory: 995.9 MiB
+Name: manager1
+ID: XP2B:EU32:5OYY:NBOD:EVKD:K344:3DG7:3YSE:E6UI:Q3Y6:IWOR:MCBS
+Docker Root Dir: /mnt/sda1/var/lib/docker
+Debug Mode (client): false
+Debug Mode (server): true
+ File Descriptors: 17
+ Goroutines: 23
+ System Time: 2016-06-24T10:09:37.185763563Z
+ EventsListeners: 0
+Registry: https://index.docker.io/v1/
+Labels:
+ provider=virtualbox
+Experimental: true
+Insecure Registries:
+ 127.0.0.0/8
+```
+
+**Additional environment details (AWS, VirtualBox, physical, etc.):**
+
+3 virtualbox machines with docker 1.12-dev created by docker-machine
+
+**Steps to reproduce the issue:**
+
+1. Create three machines.
+2. Create a swarm cluster and add an additional swarm manager
+
+ ```
+docker-machine ssh manager1 'docker swarm init'
+docker-machine ssh node1 'docker swarm join 192.168.99.106:2377'
+docker-machine ssh manager2 'docker swarm join --manager 192.168.99.106:2377'
+```
+3. Run `docker-machine ssh manager1 'docker node ls'` to list the nodes
+
+**Describe the results you received:**
+
+The status of new manager is ""Unknown""
+
+```
+ID NAME MEMBERSHIP STATUS AVAILABILITY MANAGER STATUS
+10ztstjk5dt6yb5ikm1yqrben Accepted Unknown Active Reachable
+2x91zrw9n5rsvavisiezb1flp node1 Accepted Ready Active
+d728vzs91jfn4ctw8cx9yjs1h * manager1 Accepted Ready Active Leader
+```
+
+**Describe the results you expected:**
+
+All nodes are ready.
+
+
+**Additional information you deem important (e.g. issue happens only occasionally):**
+",0,cannot add an additional swarm manager if you are reporting a new issue make sure that we do not have any duplicates already open you can ensure this by searching the issue list for this repository if there is a duplicate please close your issue and add a comment to the existing issue instead if you suspect your issue is a bug please edit your issue description to include the bug report information shown below if you fail to provide this information within days we cannot debug your issue and will close it we will however reopen it if you later provide the information for more information about reporting issues see bug report information use the commands below to provide key information from your environment you do not have to include this information if this is a feature request output of docker version client version api version go version git commit built fri jun os arch linux experimental true server version api version go version git commit built fri jun os arch linux experimental true output of docker info containers running paused stopped images server version storage driver aufs root dir mnt var lib docker aufs backing filesystem extfs dirs supported true logging driver json file cgroup driver cgroupfs plugins volume local network host bridge overlay null swarm inactive runtimes default default runtime default security options seccomp kernel version operating system tcl head fri jun utc ostype linux architecture cpus total memory mib name id nbod evkd iwor mcbs docker root dir mnt var lib docker debug mode client false debug mode server true file descriptors goroutines system time eventslisteners registry labels provider virtualbox experimental true insecure registries additional environment details aws virtualbox physical etc virtualbox machines with docker dev created by docker machine steps to reproduce the issue create three machines create a swarm cluster and add an additional swarm manager docker machine ssh docker swarm init docker machine ssh docker swarm join docker machine ssh docker swarm join manager run docker machine ssh docker node ls to list the nodes describe the results you received the status of new manager is unknown id name membership status availability manager status accepted unknown active reachable accepted ready active accepted ready active leader describe the results you expected all nodes are ready additional information you deem important e g issue happens only occasionally ,0
+385987,26661928620.0,IssuesEvent,2023-01-25 22:00:29,reframe-hpc/reframe,https://api.github.com/repos/reframe-hpc/reframe,closed,Missing line break in reframe.core.systems.SystemPartition in the docco,bug invalid prio: low documentation,"There is a missing line break between
+""property name"" and ""property prepare_cmds""
+under reframe.core.systems.SystemPartition in the documentation",1.0,"Missing line break in reframe.core.systems.SystemPartition in the docco - There is a missing line break between
+""property name"" and ""property prepare_cmds""
+under reframe.core.systems.SystemPartition in the documentation",0,missing line break in reframe core systems systempartition in the docco there is a missing line break between property name and property prepare cmds under reframe core systems systempartition in the documentation,0
+885,17594872981.0,IssuesEvent,2021-08-17 02:39:32,microsoftgraph/microsoft-graph-toolkit,https://api.github.com/repos/microsoftgraph/microsoft-graph-toolkit,closed,[BUG] Changing templateContext prop of People component doesn't cause template re-render,bug Component: people State: In Review,"**Describe the bug**
+This bug is related to #1226. When templateContext prop of People component changes, the expected behaviour is that the template would be re-rendered, just like it is when other props (eg. userIds) change.
+
+Please see example code below for demonstration:
+
+```
+import React, { useState } from 'react';
+import { Person, People, MgtTemplateProps } from '@microsoft/mgt-react';
+
+//template should re-render when MGT People component props change,
+//but it doesn't re-render when templateContext (one of its props) changes...
+const PersonTemplate = ({ dataContext }: MgtTemplateProps) => {
+ console.log('Template render');
+ const { person } = dataContext;
+ return ;
+};
+
+export default function Main() {
+ console.log('Main render');
+
+ const [userId, setUserId] = useState(0);
+ const [number, setNumber] = useState(0);
+
+ const users = [
+ { id: 0, email: 'user1@a.com' },
+ { id: 1, email: 'user2@a.com' }
+ ];
+
+ return (
+
+
+
+
+
+
+
+ );
+}
+```
+
+**To Reproduce**
+Use above script in react-app sample - click the buttons to change userIds / templateContext props and check console for re-render print.
+
+**Expected behavior**
+Template component is re-rendered when templateContext of parent MGT component changes.
+
+**Environment (please complete the following information):**
+ - OS: Windows 10
+ - Browser: Chromium
+ - Framework: React
+ - Version: v2.3.0-preview.07b40fa (i.e. post commit that fixed issue #1226)
+ - Provider: Msal2Provider
+ - Provider Config:
+
+```
+import React from 'react';
+import ReactDOM from 'react-dom';
+import App from './App';
+import { Providers } from '@microsoft/mgt-element';
+import { Msal2Provider, Msal2Config } from '@microsoft/mgt-msal2-provider';
+
+const config: Msal2Config = {
+ clientId: 'xxxxxxxx',
+ scopes: ['User.Read', 'Group.Read.All'],
+ redirectUri: 'http://localhost:3000',
+ authority: 'https://login.microsoftonline.com/xxxxxxxxx',
+ options: {
+ auth: {
+ clientId: 'xxxxxxxxxx'
+ },
+ cache: {
+ cacheLocation: 'localStorage',
+ storeAuthStateInCookie: true
+ }
+ }
+};
+
+Providers.globalProvider = new Msal2Provider(config);
+
+ReactDOM.render(, document.getElementById('root'));
+```
+
+**Additional context**
+Nil
+",1.0,"[BUG] Changing templateContext prop of People component doesn't cause template re-render - **Describe the bug**
+This bug is related to #1226. When templateContext prop of People component changes, the expected behaviour is that the template would be re-rendered, just like it is when other props (eg. userIds) change.
+
+Please see example code below for demonstration:
+
+```
+import React, { useState } from 'react';
+import { Person, People, MgtTemplateProps } from '@microsoft/mgt-react';
+
+//template should re-render when MGT People component props change,
+//but it doesn't re-render when templateContext (one of its props) changes...
+const PersonTemplate = ({ dataContext }: MgtTemplateProps) => {
+ console.log('Template render');
+ const { person } = dataContext;
+ return ;
+};
+
+export default function Main() {
+ console.log('Main render');
+
+ const [userId, setUserId] = useState(0);
+ const [number, setNumber] = useState(0);
+
+ const users = [
+ { id: 0, email: 'user1@a.com' },
+ { id: 1, email: 'user2@a.com' }
+ ];
+
+ return (
+
+
+
+
+
+
+
+ );
+}
+```
+
+**To Reproduce**
+Use above script in react-app sample - click the buttons to change userIds / templateContext props and check console for re-render print.
+
+**Expected behavior**
+Template component is re-rendered when templateContext of parent MGT component changes.
+
+**Environment (please complete the following information):**
+ - OS: Windows 10
+ - Browser: Chromium
+ - Framework: React
+ - Version: v2.3.0-preview.07b40fa (i.e. post commit that fixed issue #1226)
+ - Provider: Msal2Provider
+ - Provider Config:
+
+```
+import React from 'react';
+import ReactDOM from 'react-dom';
+import App from './App';
+import { Providers } from '@microsoft/mgt-element';
+import { Msal2Provider, Msal2Config } from '@microsoft/mgt-msal2-provider';
+
+const config: Msal2Config = {
+ clientId: 'xxxxxxxx',
+ scopes: ['User.Read', 'Group.Read.All'],
+ redirectUri: 'http://localhost:3000',
+ authority: 'https://login.microsoftonline.com/xxxxxxxxx',
+ options: {
+ auth: {
+ clientId: 'xxxxxxxxxx'
+ },
+ cache: {
+ cacheLocation: 'localStorage',
+ storeAuthStateInCookie: true
+ }
+ }
+};
+
+Providers.globalProvider = new Msal2Provider(config);
+
+ReactDOM.render(, document.getElementById('root'));
+```
+
+**Additional context**
+Nil
+",1, changing templatecontext prop of people component doesn t cause template re render describe the bug this bug is related to when templatecontext prop of people component changes the expected behaviour is that the template would be re rendered just like it is when other props eg userids change please see example code below for demonstration import react usestate from react import person people mgttemplateprops from microsoft mgt react template should re render when mgt people component props change but it doesn t re render when templatecontext one of its props changes const persontemplate datacontext mgttemplateprops console log template render const person datacontext return export default function main console log main render const usestate const usestate const users id email a com id email a com return setuserid userid change userids setnumber number change templatecontext number number to reproduce use above script in react app sample click the buttons to change userids templatecontext props and check console for re render print expected behavior template component is re rendered when templatecontext of parent mgt component changes environment please complete the following information os windows browser chromium framework react version preview i e post commit that fixed issue provider provider config import react from react import reactdom from react dom import app from app import providers from microsoft mgt element import from microsoft mgt provider const config clientid xxxxxxxx scopes redirecturi authority options auth clientid xxxxxxxxxx cache cachelocation localstorage storeauthstateincookie true providers globalprovider new config reactdom render document getelementbyid root additional context nil ,1
+36809,6551689398.0,IssuesEvent,2017-09-05 15:32:23,coreos/tectonic-installer,https://api.github.com/repos/coreos/tectonic-installer,closed,docs request: AWS BYO VPC - ELB Integration,kind/documentation,"## Versions
+Any
+
+## What happened?
+When users deploy Tectonic into their own VPC and attempt to create a Kubernetes Service with type Load Balancer, ELB gets created however does not get attached to all subnets.
+
+In order to fully utilize AWS ELB / Kubernetes integration, users must tag their AWS subnets with; `kubernetes.io/cluster/my-cluster-name = owned` tag.
+
+Kindly requesting the @coreos/team-docs team to assist with documentation. Happy to submit a PR if you can provide an appropriate page/location.
+
+",1.0,"docs request: AWS BYO VPC - ELB Integration - ## Versions
+Any
+
+## What happened?
+When users deploy Tectonic into their own VPC and attempt to create a Kubernetes Service with type Load Balancer, ELB gets created however does not get attached to all subnets.
+
+In order to fully utilize AWS ELB / Kubernetes integration, users must tag their AWS subnets with; `kubernetes.io/cluster/my-cluster-name = owned` tag.
+
+Kindly requesting the @coreos/team-docs team to assist with documentation. Happy to submit a PR if you can provide an appropriate page/location.
+
+",0,docs request aws byo vpc elb integration versions any what happened when users deploy tectonic into their own vpc and attempt to create a kubernetes service with type load balancer elb gets created however does not get attached to all subnets in order to fully utilize aws elb kubernetes integration users must tag their aws subnets with kubernetes io cluster my cluster name owned tag kindly requesting the coreos team docs team to assist with documentation happy to submit a pr if you can provide an appropriate page location ,0
+50688,10546987413.0,IssuesEvent,2019-10-02 23:12:36,MicrosoftDocs/visualstudio-docs,https://api.github.com/repos/MicrosoftDocs/visualstudio-docs,closed,unclear documentation,Pri2 area - C++ doc-bug visual-studio-windows/prod vs-ide-code-analysis/tech,"Please define what slicing is.
+
+---
+#### Document Details
+
+⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
+
+* ID: be6a0e19-d61f-fefb-4a83-afa87219e168
+* Version Independent ID: 206fffd7-3651-5783-49e0-c586a9b8b992
+* Content: [C26437 - Visual Studio](https://docs.microsoft.com/en-us/visualstudio/code-quality/c26437?f1url=https%3A%2F%2Fmsdn.microsoft.com%2Fquery%2Fdev16.query%3FappId%3DDev16IDEF1%26l%3DEN-US%26k%3Dk(C26437)%26rd%3Dtrue&view=vs-2019#feedback)
+* Content Source: [docs/code-quality/C26437.md](https://github.com/MicrosoftDocs/visualstudio-docs/blob/master/docs/code-quality/C26437.md)
+* Product: **visual-studio-windows**
+* Technology: **vs-ide-code-analysis**
+* GitHub Login: @mikeblome
+* Microsoft Alias: **mblome**",1.0,"unclear documentation - Please define what slicing is.
+
+---
+#### Document Details
+
+⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
+
+* ID: be6a0e19-d61f-fefb-4a83-afa87219e168
+* Version Independent ID: 206fffd7-3651-5783-49e0-c586a9b8b992
+* Content: [C26437 - Visual Studio](https://docs.microsoft.com/en-us/visualstudio/code-quality/c26437?f1url=https%3A%2F%2Fmsdn.microsoft.com%2Fquery%2Fdev16.query%3FappId%3DDev16IDEF1%26l%3DEN-US%26k%3Dk(C26437)%26rd%3Dtrue&view=vs-2019#feedback)
+* Content Source: [docs/code-quality/C26437.md](https://github.com/MicrosoftDocs/visualstudio-docs/blob/master/docs/code-quality/C26437.md)
+* Product: **visual-studio-windows**
+* Technology: **vs-ide-code-analysis**
+* GitHub Login: @mikeblome
+* Microsoft Alias: **mblome**",0,unclear documentation please define what slicing is document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id fefb version independent id content content source product visual studio windows technology vs ide code analysis github login mikeblome microsoft alias mblome ,0
+1116,27089394060.0,IssuesEvent,2023-02-14 19:39:17,openstates/issues,https://api.github.com/repos/openstates/issues,closed,New ID Committee Scraper,good first issue component:people-data good first scraper,"### We need a new committee scraper written in [spatula](https://jamesturk.github.io/spatula/) for Idaho.
+
+It should scrape this [webpage of Senate Committees](https://legislature.idaho.gov/committees/senatecommittees/), this [webpage of House Committees](https://legislature.idaho.gov/committees/housecommittees/), this [webpage of Joint Committees](https://legislature.idaho.gov/committees/jointcommittees/), this [webpage of Interim/Special Committees](https://legislature.idaho.gov/committees/interimcommittees/), and this [specific webpage for Legislative Council (a unique management committee)](https://legislature.idaho.gov/legcouncil/) to get:
+- **name**
+ - “Small Business”, “Agriculture”, etc.
+- **chamber**
+ - “upper”, “lower”, or “legislature” (when joint, special/interim, legislative council)
+- **classification**
+ - ex: ""committee"", ""subcommittee""
+- **parent**
+ - *only if it is a subcommittee* - scraper should have a way to determine this
+ - ex: parent=""Natural Resources"" when *sub*committee name=""Forestry""
+- **sources**
+ - each added using `add_source()` method on instance of `ScrapeCommittee` type object
+ - ex: home page for list of committees, specific page for that committee, etc.
+- **members**
+ - - each added using `add_member()`method on instance of `ScrapeCommittee` type object
+ - *Attributes:*
+ - **name**: “Jane Doe”, “John Smith”, etc.
+ - **role** (where applicable): “Chair”, “Ranking Member”, etc.
+
+#### [Here is a helpful general reference doc](https://docs.openstates.org/contributing/writing-a-committee-scraper/) for those new to writing scrapers, the Open States project, or the specific task of writing a committee scraper
+
+### Useful scrapers for reference
+An [old non-spatula-based scraper](https://github.com/openstates/openstates-scrapers/blob/main/scrapers/id/committees.py) that may be helpful in understanding how we have captured the data in the past.
+
+A [spatula-based scraper](https://github.com/openstates/openstates-scrapers/blob/main/scrapers_next/mo/committees.py) that also scrapes an HTML List Page (using `HtmlListPage` and `HtmlPage` spatula classes) for reference.
+- As in the other spatula scraper, you will need to set the session variable separately (rather than depend on the `__init__.py`).
+
+### Other useful resources
+You can reference the [object models for committee scrapers](https://github.com/openstates/openstates-core/blob/main/openstates/models/committees.py) in the Open States Core repo.
+
+Further documentation on running spatula scrapers in the command line can be found [here](https://docs.openstates.org/contributing/scrapers/#running-spatula-scrapers).",1.0,"New ID Committee Scraper - ### We need a new committee scraper written in [spatula](https://jamesturk.github.io/spatula/) for Idaho.
+
+It should scrape this [webpage of Senate Committees](https://legislature.idaho.gov/committees/senatecommittees/), this [webpage of House Committees](https://legislature.idaho.gov/committees/housecommittees/), this [webpage of Joint Committees](https://legislature.idaho.gov/committees/jointcommittees/), this [webpage of Interim/Special Committees](https://legislature.idaho.gov/committees/interimcommittees/), and this [specific webpage for Legislative Council (a unique management committee)](https://legislature.idaho.gov/legcouncil/) to get:
+- **name**
+ - “Small Business”, “Agriculture”, etc.
+- **chamber**
+ - “upper”, “lower”, or “legislature” (when joint, special/interim, legislative council)
+- **classification**
+ - ex: ""committee"", ""subcommittee""
+- **parent**
+ - *only if it is a subcommittee* - scraper should have a way to determine this
+ - ex: parent=""Natural Resources"" when *sub*committee name=""Forestry""
+- **sources**
+ - each added using `add_source()` method on instance of `ScrapeCommittee` type object
+ - ex: home page for list of committees, specific page for that committee, etc.
+- **members**
+ - - each added using `add_member()`method on instance of `ScrapeCommittee` type object
+ - *Attributes:*
+ - **name**: “Jane Doe”, “John Smith”, etc.
+ - **role** (where applicable): “Chair”, “Ranking Member”, etc.
+
+#### [Here is a helpful general reference doc](https://docs.openstates.org/contributing/writing-a-committee-scraper/) for those new to writing scrapers, the Open States project, or the specific task of writing a committee scraper
+
+### Useful scrapers for reference
+An [old non-spatula-based scraper](https://github.com/openstates/openstates-scrapers/blob/main/scrapers/id/committees.py) that may be helpful in understanding how we have captured the data in the past.
+
+A [spatula-based scraper](https://github.com/openstates/openstates-scrapers/blob/main/scrapers_next/mo/committees.py) that also scrapes an HTML List Page (using `HtmlListPage` and `HtmlPage` spatula classes) for reference.
+- As in the other spatula scraper, you will need to set the session variable separately (rather than depend on the `__init__.py`).
+
+### Other useful resources
+You can reference the [object models for committee scrapers](https://github.com/openstates/openstates-core/blob/main/openstates/models/committees.py) in the Open States Core repo.
+
+Further documentation on running spatula scrapers in the command line can be found [here](https://docs.openstates.org/contributing/scrapers/#running-spatula-scrapers).",1,new id committee scraper we need a new committee scraper written in for idaho it should scrape this this this this and this to get name “small business” “agriculture” etc chamber “upper” “lower” or “legislature” when joint special interim legislative council classification ex committee subcommittee parent only if it is a subcommittee scraper should have a way to determine this ex parent natural resources when sub committee name forestry sources each added using add source method on instance of scrapecommittee type object ex home page for list of committees specific page for that committee etc members each added using add member method on instance of scrapecommittee type object attributes name “jane doe” “john smith” etc role where applicable “chair” “ranking member” etc for those new to writing scrapers the open states project or the specific task of writing a committee scraper useful scrapers for reference an that may be helpful in understanding how we have captured the data in the past a that also scrapes an html list page using htmllistpage and htmlpage spatula classes for reference as in the other spatula scraper you will need to set the session variable separately rather than depend on the init py other useful resources you can reference the in the open states core repo further documentation on running spatula scrapers in the command line can be found ,1
+845,15838563647.0,IssuesEvent,2021-04-06 22:45:10,microsoft/fluentui,https://api.github.com/repos/microsoft/fluentui,opened,Pickers that open on click do not open on alt + down,Area: Accessibility Component: PeoplePicker Component: Pickers,"### Environment Information
+
+- **Package version(s)**: latest/8.9.0
+
+### Describe the issue:
+
+Although not all of the pickers show a suggestions list without typing, some (e.g. the PeoplePicker) do. If they do, the same list that shows up on click should also be available with the standard alt + down keyboard shortcut.",1.0,"Pickers that open on click do not open on alt + down - ### Environment Information
+
+- **Package version(s)**: latest/8.9.0
+
+### Describe the issue:
+
+Although not all of the pickers show a suggestions list without typing, some (e.g. the PeoplePicker) do. If they do, the same list that shows up on click should also be available with the standard alt + down keyboard shortcut.",1,pickers that open on click do not open on alt down environment information package version s latest describe the issue although not all of the pickers show a suggestions list without typing some e g the peoplepicker do if they do the same list that shows up on click should also be available with the standard alt down keyboard shortcut ,1
+300090,9206126888.0,IssuesEvent,2019-03-08 12:49:06,threefoldtech/0-templates,https://api.github.com/repos/threefoldtech/0-templates,closed,Issue in reverse proxy template,priority_major type_question,"- had to comment that line to make it to work [link](https://github.com/threefoldtech/0-templates/blob/development/templates/reverse_proxy/reverse_proxy.py#L27)
+- getting this error when i try to create a [reverse_proxy](https://github.com/threefoldtech/0-templates/tree/development/templates/reverse_proxy)
+
+
+",1.0,"Issue in reverse proxy template - - had to comment that line to make it to work [link](https://github.com/threefoldtech/0-templates/blob/development/templates/reverse_proxy/reverse_proxy.py#L27)
+- getting this error when i try to create a [reverse_proxy](https://github.com/threefoldtech/0-templates/tree/development/templates/reverse_proxy)
+
+
+",0,issue in reverse proxy template had to comment that line to make it to work getting this error when i try to create a ,0
+5590,3251785631.0,IssuesEvent,2015-10-19 11:52:14,robertocarroll/ideas,https://api.github.com/repos/robertocarroll/ideas,opened,Stories around photos,code writing,"Writing around photos, Barthes, see also http://www.thehypertext.com/2015/04/11/word-camera/
+
+1. Found photos from parents' album
+2. Process them with Clarifai API
+3. Use tags as inspiration to write - short things -
+4. Put the writing back with the images ",1.0,"Stories around photos - Writing around photos, Barthes, see also http://www.thehypertext.com/2015/04/11/word-camera/
+
+1. Found photos from parents' album
+2. Process them with Clarifai API
+3. Use tags as inspiration to write - short things -
+4. Put the writing back with the images ",0,stories around photos writing around photos barthes see also found photos from parents album process them with clarifai api use tags as inspiration to write short things put the writing back with the images ,0
+217080,16679106039.0,IssuesEvent,2021-06-07 20:22:16,DhenSouza/cursomc,https://api.github.com/repos/DhenSouza/cursomc,closed,"Ajustar o endPoint categorias id, e tratamento de exceções personalizadas.",documentation enhancement,"Adicionado ResourceExceptionHandler personalização de error;
+Adicionado StandardError;",1.0,"Ajustar o endPoint categorias id, e tratamento de exceções personalizadas. - Adicionado ResourceExceptionHandler personalização de error;
+Adicionado StandardError;",0,ajustar o endpoint categorias id e tratamento de exceções personalizadas adicionado resourceexceptionhandler personalização de error adicionado standarderror ,0
+51857,6552785374.0,IssuesEvent,2017-09-05 19:45:22,NuGet/Home,https://api.github.com/repos/NuGet/Home,closed,No Way to Control Order of Sources with nuget sources,ClosedAs:ByDesign,"I have a network path where I have cached numerous *.nupkg files (from the public gallery) for other developers in my department with slow Internet but fast LAN access. I'd like to add this source at the top of the list so that during package restore, packages will be taken from the cache first (if available).
+
+The command-line tool does not seem to offer any way to accomplish this. Sources are always added at the bottom of the list. I thought about removing the main ""nuget.org"" source and adding it back, but this seems to be a special entry that is listed twice, once with protocolVersion=""3"", and again, there doesn't seem to be a way to add that from the command line.
+
+I would be satisfied with a ""nuget sources move"" command with a -Location parameter with options of ""top"" or ""bottom"" (also with a -Name parameter), or the ""nuget sources add"" command could have the -Location parameter (with default value of ""bottom""). I guess the location could also be a 0-based index if you want better control, but the ability to insert at the top is the basic requirement.
+
+A unit test for this could:
+1. Start with the default config.
+2. Run a command or sequence of commands that puts a new source at the top without interfering with the existing ""nuget.org"" source. For example, nuget.exe sources add -Name ""Public NuGet Cache"" -Source ""\server\path\to\nuget\cache"" -Location top
+3. Compare nuget.config to a sample config file in the desired state.
+
+Thanks!
+",1.0,"No Way to Control Order of Sources with nuget sources - I have a network path where I have cached numerous *.nupkg files (from the public gallery) for other developers in my department with slow Internet but fast LAN access. I'd like to add this source at the top of the list so that during package restore, packages will be taken from the cache first (if available).
+
+The command-line tool does not seem to offer any way to accomplish this. Sources are always added at the bottom of the list. I thought about removing the main ""nuget.org"" source and adding it back, but this seems to be a special entry that is listed twice, once with protocolVersion=""3"", and again, there doesn't seem to be a way to add that from the command line.
+
+I would be satisfied with a ""nuget sources move"" command with a -Location parameter with options of ""top"" or ""bottom"" (also with a -Name parameter), or the ""nuget sources add"" command could have the -Location parameter (with default value of ""bottom""). I guess the location could also be a 0-based index if you want better control, but the ability to insert at the top is the basic requirement.
+
+A unit test for this could:
+1. Start with the default config.
+2. Run a command or sequence of commands that puts a new source at the top without interfering with the existing ""nuget.org"" source. For example, nuget.exe sources add -Name ""Public NuGet Cache"" -Source ""\server\path\to\nuget\cache"" -Location top
+3. Compare nuget.config to a sample config file in the desired state.
+
+Thanks!
+",0,no way to control order of sources with nuget sources i have a network path where i have cached numerous nupkg files from the public gallery for other developers in my department with slow internet but fast lan access i d like to add this source at the top of the list so that during package restore packages will be taken from the cache first if available the command line tool does not seem to offer any way to accomplish this sources are always added at the bottom of the list i thought about removing the main nuget org source and adding it back but this seems to be a special entry that is listed twice once with protocolversion and again there doesn t seem to be a way to add that from the command line i would be satisfied with a nuget sources move command with a location parameter with options of top or bottom also with a name parameter or the nuget sources add command could have the location parameter with default value of bottom i guess the location could also be a based index if you want better control but the ability to insert at the top is the basic requirement a unit test for this could start with the default config run a command or sequence of commands that puts a new source at the top without interfering with the existing nuget org source for example nuget exe sources add name public nuget cache source server path to nuget cache location top compare nuget config to a sample config file in the desired state thanks ,0
+306094,9380601949.0,IssuesEvent,2019-04-04 17:30:05,citizenlabsgr/adopt-a-drain,https://api.github.com/repos/citizenlabsgr/adopt-a-drain,closed,Adopt a Drain Logo is too big in Outlook,Priority-1 review,"The logo in the automated email shows up way too large, but only in Outlook. Photo now shows up fine. Reference email that Joel sent.",1.0,"Adopt a Drain Logo is too big in Outlook - The logo in the automated email shows up way too large, but only in Outlook. Photo now shows up fine. Reference email that Joel sent.",0,adopt a drain logo is too big in outlook the logo in the automated email shows up way too large but only in outlook photo now shows up fine reference email that joel sent ,0
+240,4953910930.0,IssuesEvent,2016-12-01 16:14:15,wordpress-mobile/WordPress-Android,https://api.github.com/repos/wordpress-mobile/WordPress-Android,closed,Denying access to self-hosted WP-Site. ,Media People Management,"### Expected behavior
+Denies access to other author's media gallerys. Prohibits use of other's media
+
+### Actual behavior
+Permits access to and use of media in other author's media libaries even though permission has been restricted for the internet site at www.___.com
+
+### Steps to reproduce the behavior
+None, no error message.
+
+##### Tested on [device], Android [version]
+Samsung Galaxy Note II, Android 4.4.2.
+
+##### Detailed description
+
+I've setup a self-hosted multi-author website that would deny authors to view or use other author's media (photos, video, sounds). This is done via the plugin WPFront User Role Pro, because WP 4.6.1 appears not to have a similar limitation for the author roles. Using the WP App in my smart phone, I find that the restrictions are circumvented and the authors can use any media of other author's libraries. I only know this of the Android-Version so far.
+
+Is there a way to prevent the Wordpress APPs (regardless of OS or version) from accessing the site at all?
+
+Thanks for the help",1.0,"Denying access to self-hosted WP-Site. - ### Expected behavior
+Denies access to other author's media gallerys. Prohibits use of other's media
+
+### Actual behavior
+Permits access to and use of media in other author's media libaries even though permission has been restricted for the internet site at www.___.com
+
+### Steps to reproduce the behavior
+None, no error message.
+
+##### Tested on [device], Android [version]
+Samsung Galaxy Note II, Android 4.4.2.
+
+##### Detailed description
+
+I've setup a self-hosted multi-author website that would deny authors to view or use other author's media (photos, video, sounds). This is done via the plugin WPFront User Role Pro, because WP 4.6.1 appears not to have a similar limitation for the author roles. Using the WP App in my smart phone, I find that the restrictions are circumvented and the authors can use any media of other author's libraries. I only know this of the Android-Version so far.
+
+Is there a way to prevent the Wordpress APPs (regardless of OS or version) from accessing the site at all?
+
+Thanks for the help",1,denying access to self hosted wp site expected behavior denies access to other author s media gallerys prohibits use of other s media actual behavior permits access to and use of media in other author s media libaries even though permission has been restricted for the internet site at steps to reproduce the behavior none no error message tested on android samsung galaxy note ii android detailed description i ve setup a self hosted multi author website that would deny authors to view or use other author s media photos video sounds this is done via the plugin wpfront user role pro because wp appears not to have a similar limitation for the author roles using the wp app in my smart phone i find that the restrictions are circumvented and the authors can use any media of other author s libraries i only know this of the android version so far is there a way to prevent the wordpress apps regardless of os or version from accessing the site at all thanks for the help,1
+55487,30770540212.0,IssuesEvent,2023-07-30 21:15:22,sqlfluff/sqlfluff,https://api.github.com/repos/sqlfluff/sqlfluff,closed,Initial SQLFluff invocation loads many unnecessary plugins,bug performance,"### Search before asking
+
+- [X] I searched the [issues](https://github.com/sqlfluff/sqlfluff/issues) and found no similar issues.
+
+
+### What Happened
+
+I am using sqlfluff.parse to perform validations on LLM generated SQL statements. Seems like on the first time it is called it is taking too long to run.
+
+
+### Expected Behaviour
+
+sqlfluff.parse should not take over 2 seconds on first invocation
+
+### Observed Behaviour
+
+The first time parse is called, it takes quite long (2-3 seconds) while following executions are much much shorter (around 1/10 of second).
+This is obviously better than all invocation taking that long, but I wonder if there is a way to reduce the time of the first call as well, or maybe even know what's causing it to take so long.
+I further tested this behavior by adding sqlfluff.parse("""") after module load, which took the initial long time and afterwards later invocations were fast again.
+
+### How to reproduce
+
+Wrap each invocation of sqlfluff.parse with something simple like:
+start_time = timeit.default_timer()
+sqlfluff.parse("""")
+end_time = timeit.default_timer()
+print(f""1 - The function took {end_time - start_time} seconds to complete."")
+
+start_time = timeit.default_timer()
+sqlfluff.parse("""")
+end_time = timeit.default_timer()
+print(f""2 - The function took {end_time - start_time} seconds to complete."")
+
+And we'll be able to see that the second time works much faster
+
+### Dialect
+
+ansi
+
+### Version
+
+sqlfluff, version 2.1.1
+Python 3.9.16
+
+### Configuration
+
+Using default
+
+### Are you willing to work on and submit a PR to address the issue?
+
+- [ ] Yes I am willing to submit a PR!
+
+### Code of Conduct
+
+- [X] I agree to follow this project's [Code of Conduct](https://github.com/sqlfluff/sqlfluff/blob/main/CODE_OF_CONDUCT.md)
+",True,"Initial SQLFluff invocation loads many unnecessary plugins - ### Search before asking
+
+- [X] I searched the [issues](https://github.com/sqlfluff/sqlfluff/issues) and found no similar issues.
+
+
+### What Happened
+
+I am using sqlfluff.parse to perform validations on LLM generated SQL statements. Seems like on the first time it is called it is taking too long to run.
+
+
+### Expected Behaviour
+
+sqlfluff.parse should not take over 2 seconds on first invocation
+
+### Observed Behaviour
+
+The first time parse is called, it takes quite long (2-3 seconds) while following executions are much much shorter (around 1/10 of second).
+This is obviously better than all invocation taking that long, but I wonder if there is a way to reduce the time of the first call as well, or maybe even know what's causing it to take so long.
+I further tested this behavior by adding sqlfluff.parse("""") after module load, which took the initial long time and afterwards later invocations were fast again.
+
+### How to reproduce
+
+Wrap each invocation of sqlfluff.parse with something simple like:
+start_time = timeit.default_timer()
+sqlfluff.parse("""")
+end_time = timeit.default_timer()
+print(f""1 - The function took {end_time - start_time} seconds to complete."")
+
+start_time = timeit.default_timer()
+sqlfluff.parse("""")
+end_time = timeit.default_timer()
+print(f""2 - The function took {end_time - start_time} seconds to complete."")
+
+And we'll be able to see that the second time works much faster
+
+### Dialect
+
+ansi
+
+### Version
+
+sqlfluff, version 2.1.1
+Python 3.9.16
+
+### Configuration
+
+Using default
+
+### Are you willing to work on and submit a PR to address the issue?
+
+- [ ] Yes I am willing to submit a PR!
+
+### Code of Conduct
+
+- [X] I agree to follow this project's [Code of Conduct](https://github.com/sqlfluff/sqlfluff/blob/main/CODE_OF_CONDUCT.md)
+",0,initial sqlfluff invocation loads many unnecessary plugins search before asking i searched the and found no similar issues what happened i am using sqlfluff parse to perform validations on llm generated sql statements seems like on the first time it is called it is taking too long to run expected behaviour sqlfluff parse should not take over seconds on first invocation observed behaviour the first time parse is called it takes quite long seconds while following executions are much much shorter around of second this is obviously better than all invocation taking that long but i wonder if there is a way to reduce the time of the first call as well or maybe even know what s causing it to take so long i further tested this behavior by adding sqlfluff parse after module load which took the initial long time and afterwards later invocations were fast again how to reproduce wrap each invocation of sqlfluff parse with something simple like start time timeit default timer sqlfluff parse end time timeit default timer print f the function took end time start time seconds to complete start time timeit default timer sqlfluff parse end time timeit default timer print f the function took end time start time seconds to complete and we ll be able to see that the second time works much faster dialect ansi version sqlfluff version python configuration using default are you willing to work on and submit a pr to address the issue yes i am willing to submit a pr code of conduct i agree to follow this project s ,0
+643,11543310698.0,IssuesEvent,2020-02-18 09:22:29,michelmelo/instagram-private-api,https://api.github.com/repos/michelmelo/instagram-private-api,reopened,get all feeds,discoverPeople people question,"how I can get the suggested posts, I mean the post that appears when I click on the search icon. and not timeline",2.0,"get all feeds - how I can get the suggested posts, I mean the post that appears when I click on the search icon. and not timeline",1,get all feeds how i can get the suggested posts i mean the post that appears when i click on the search icon and not timeline,1
+432,7931616780.0,IssuesEvent,2018-07-07 02:28:58,CodexioDevCamp-TeamIII/ConferenceScheduler,https://api.github.com/repos/CodexioDevCamp-TeamIII/ConferenceScheduler,closed,fix register html,Bug Good for new people,Register html should show username instead of emil and the constrains for the email must be removed.,1.0,fix register html - Register html should show username instead of emil and the constrains for the email must be removed.,1,fix register html register html should show username instead of emil and the constrains for the email must be removed ,1
+119533,4771582506.0,IssuesEvent,2016-10-26 18:24:53,CS2103AUG2016-T09-C1/main,https://api.github.com/repos/CS2103AUG2016-T09-C1/main,closed,As a user I can select my storage location for my data file.,priority.medium type.story,...so that I can access task data from multiple computers.,1.0,As a user I can select my storage location for my data file. - ...so that I can access task data from multiple computers.,0,as a user i can select my storage location for my data file so that i can access task data from multiple computers ,0
+122870,17771603433.0,IssuesEvent,2021-08-30 14:15:29,Tim-sandbox/barista,https://api.github.com/repos/Tim-sandbox/barista,opened,CVE-2020-28498 (Medium) detected in elliptic-6.5.3.tgz,security vulnerability,"## CVE-2020-28498 - Medium Severity Vulnerability
+ Vulnerable Library - elliptic-6.5.3.tgz
+
+
+
+The package elliptic before 6.5.4 are vulnerable to Cryptographic Issues via the secp256k1 implementation in elliptic/ec/key.js. There is no check to confirm that the public key point passed into the derive function actually exists on the secp256k1 curve. This results in the potential for the private key used in this implementation to be revealed after a number of ECDH operations are performed.
+
+
+
+The package elliptic before 6.5.4 are vulnerable to Cryptographic Issues via the secp256k1 implementation in elliptic/ec/key.js. There is no check to confirm that the public key point passed into the derive function actually exists on the secp256k1 curve. This results in the potential for the private key used in this implementation to be revealed after a number of ECDH operations are performed.
+
+
+
+
+
+
+
+",0,cve medium detected in elliptic tgz cve medium severity vulnerability vulnerable library elliptic tgz ec cryptography library home page a href path to dependency file barista barista web package json path to vulnerable library barista barista web node modules elliptic package json dependency hierarchy build angular tgz root library webpack tgz node libs browser tgz crypto browserify tgz browserify sign tgz x elliptic tgz vulnerable library found in head commit a href found in base branch master vulnerability details the package elliptic before are vulnerable to cryptographic issues via the implementation in elliptic ec key js there is no check to confirm that the public key point passed into the derive function actually exists on the curve this results in the potential for the private key used in this implementation to be revealed after a number of ecdh operations are performed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope changed impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree angular devkit build angular webpack node libs browser crypto browserify browserify sign elliptic isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails the package elliptic before are vulnerable to cryptographic issues via the implementation in elliptic ec key js there is no check to confirm that the public key point passed into the derive function actually exists on the curve this results in the potential for the private key used in this implementation to be revealed after a number of ecdh operations are performed vulnerabilityurl ,0
+123523,4864382965.0,IssuesEvent,2016-11-14 17:50:30,emfoundation/ce100-app,https://api.github.com/repos/emfoundation/ce100-app,opened,Primary users can delete challenges once they've been archived,primary-user priority-3,"Deleting a challenge removes it from the database.
+
+Relevant screens:
+* [delete from archived challenges listing view](https://zpl.io/2eeRHw)
+* [user needs to confirm that the challenge can be deleted](https://zpl.io/15ejog)
+* [final confirmation that challenge has been deleted](https://zpl.io/1mcNH0); this could also be done inline (like on other pages)",1.0,"Primary users can delete challenges once they've been archived - Deleting a challenge removes it from the database.
+
+Relevant screens:
+* [delete from archived challenges listing view](https://zpl.io/2eeRHw)
+* [user needs to confirm that the challenge can be deleted](https://zpl.io/15ejog)
+* [final confirmation that challenge has been deleted](https://zpl.io/1mcNH0); this could also be done inline (like on other pages)",0,primary users can delete challenges once they ve been archived deleting a challenge removes it from the database relevant screens this could also be done inline like on other pages ,0
+151561,12043317311.0,IssuesEvent,2020-04-14 12:12:58,hail-is/hail,https://api.github.com/repos/hail-is/hail,closed,missing TableIR tests,testing,"- [ ] TableRowsTable
+- [ ] TableUnion
+- [ ] TableJoin
+- [ ] TableRange
+- [ ] TableMapGlobals
+- [ ] TableExplode
+- [ ] TableKeyBy
+- [ ] TableMapRows
+- [ ] TableUnkey
+- [ ] TableOrderBy
+- [ ] TableRead
+- [ ] TableImport
+- [ ] TableEntriesTable
+- [ ] TableAggregateByKey
+- [ ] MatrixColsTable
+- [ ] TableDistinct
+- [x] TableParallelize
+- [ ] TableRepartition
+",1.0,"missing TableIR tests - - [ ] TableRowsTable
+- [ ] TableUnion
+- [ ] TableJoin
+- [ ] TableRange
+- [ ] TableMapGlobals
+- [ ] TableExplode
+- [ ] TableKeyBy
+- [ ] TableMapRows
+- [ ] TableUnkey
+- [ ] TableOrderBy
+- [ ] TableRead
+- [ ] TableImport
+- [ ] TableEntriesTable
+- [ ] TableAggregateByKey
+- [ ] MatrixColsTable
+- [ ] TableDistinct
+- [x] TableParallelize
+- [ ] TableRepartition
+",0,missing tableir tests tablerowstable tableunion tablejoin tablerange tablemapglobals tableexplode tablekeyby tablemaprows tableunkey tableorderby tableread tableimport tableentriestable tableaggregatebykey matrixcolstable tabledistinct tableparallelize tablerepartition ,0
+13153,3315342813.0,IssuesEvent,2015-11-06 11:28:42,SPW-DIG/metawal-core-geonetwork,https://api.github.com/repos/SPW-DIG/metawal-core-geonetwork,closed,Modifier la définition du gestionnaire de données,Env test - OK Env valid - OK,"Nouvelle définition : L'organisme ou la personne qui veille à garantir l’acquisition, la qualité, l’intégrité, le stockage, la mise à jour, la documentation et la diffusion de la ressource.",1.0,"Modifier la définition du gestionnaire de données - Nouvelle définition : L'organisme ou la personne qui veille à garantir l’acquisition, la qualité, l’intégrité, le stockage, la mise à jour, la documentation et la diffusion de la ressource.",0,modifier la définition du gestionnaire de données nouvelle définition l organisme ou la personne qui veille à garantir l’acquisition la qualité l’intégrité le stockage la mise à jour la documentation et la diffusion de la ressource ,0
+35495,6475278598.0,IssuesEvent,2017-08-17 20:01:30,cerner/clara-rules,https://api.github.com/repos/cerner/clara-rules,closed,Inconsistent naming on Fact Type Customization example,documentation,"The last example given at [Fact Type Customization](http://www.clara-rules.org/docs/fact_type_customization/) is a bit confusing because of that.
+
+At the last paragraph:
+
+> Clara will then call the ancestors-fn with an argument of :precise-temperature-fn.
+
+should be
+
+Clara will then call the ancestors-fn with an argument of **:precise-temperature-reading**.
+
+, and
+
+> In this case, since :temperature-reading is in the set of ancestors of :precise-temperatures-fn it is considered…
+
+should be
+
+In this case, since :temperature-reading is in the set of ancestors of **:precise-temperatures-reading** it is considered…",1.0,"Inconsistent naming on Fact Type Customization example - The last example given at [Fact Type Customization](http://www.clara-rules.org/docs/fact_type_customization/) is a bit confusing because of that.
+
+At the last paragraph:
+
+> Clara will then call the ancestors-fn with an argument of :precise-temperature-fn.
+
+should be
+
+Clara will then call the ancestors-fn with an argument of **:precise-temperature-reading**.
+
+, and
+
+> In this case, since :temperature-reading is in the set of ancestors of :precise-temperatures-fn it is considered…
+
+should be
+
+In this case, since :temperature-reading is in the set of ancestors of **:precise-temperatures-reading** it is considered…",0,inconsistent naming on fact type customization example the last example given at is a bit confusing because of that at the last paragraph clara will then call the ancestors fn with an argument of precise temperature fn should be clara will then call the ancestors fn with an argument of precise temperature reading and in this case since temperature reading is in the set of ancestors of precise temperatures fn it is considered… should be in this case since temperature reading is in the set of ancestors of precise temperatures reading it is considered…,0
+216,4424711635.0,IssuesEvent,2016-08-16 13:28:30,befair/soulShape,https://api.github.com/repos/befair/soulShape,closed,[POPOLI] una pagina per ogni popolo,area:indigenouspeople priority:P0,"Per ogni popolo si crea una pagina Wordpress descrittiva, compilata da Pierpaolo. Il link verrà associato ad un gruppo. Vedi issue #12 ",1.0,"[POPOLI] una pagina per ogni popolo - Per ogni popolo si crea una pagina Wordpress descrittiva, compilata da Pierpaolo. Il link verrà associato ad un gruppo. Vedi issue #12 ",1, una pagina per ogni popolo per ogni popolo si crea una pagina wordpress descrittiva compilata da pierpaolo il link verrà associato ad un gruppo vedi issue ,1
+10642,8130220610.0,IssuesEvent,2018-08-17 17:42:01,csterwa/scs-combined-board,https://api.github.com/repos/csterwa/scs-combined-board,closed,Compatibility with Spring Cloud Gateway & WebFlux,security,"Because SC Gateway is based on Spring WebFlux, but our connectors depend on Spring Security for OAuth (S2OAuth), which depends on Spring MVC, it has been found that SCS client apps are not compatible with Spring Cloud Gateway and, more specifically, Spring WebFlux.
+
+Given that the Spring Framework is promoting reactive programming heavily and given that Spring Cloud Gateway is replacing Spring Cloud Netflix Zuul as the gateway of choice, it is likely that SCS customers will want to use WebFlux (either for API purposes or using `WebClient` instead of `RestTemplate`...or both). Therefore, we should aim to support WebFlux as an option in SCS client applications.
+
+Note, the incompatibility appears to come from our reliance on S2OAuth for security between clients and the services. And since S2OAuth does not support WebFlux, we may should consider moving toward Spring Security 5's client-side OAuth support, working with the Spring Security team to ensure that our needs are met for client-side applications. Also note that this is largely a starters/connectors issue and there is no anticipated impact to the broker or backing apps.",True,"Compatibility with Spring Cloud Gateway & WebFlux - Because SC Gateway is based on Spring WebFlux, but our connectors depend on Spring Security for OAuth (S2OAuth), which depends on Spring MVC, it has been found that SCS client apps are not compatible with Spring Cloud Gateway and, more specifically, Spring WebFlux.
+
+Given that the Spring Framework is promoting reactive programming heavily and given that Spring Cloud Gateway is replacing Spring Cloud Netflix Zuul as the gateway of choice, it is likely that SCS customers will want to use WebFlux (either for API purposes or using `WebClient` instead of `RestTemplate`...or both). Therefore, we should aim to support WebFlux as an option in SCS client applications.
+
+Note, the incompatibility appears to come from our reliance on S2OAuth for security between clients and the services. And since S2OAuth does not support WebFlux, we may should consider moving toward Spring Security 5's client-side OAuth support, working with the Spring Security team to ensure that our needs are met for client-side applications. Also note that this is largely a starters/connectors issue and there is no anticipated impact to the broker or backing apps.",0,compatibility with spring cloud gateway webflux because sc gateway is based on spring webflux but our connectors depend on spring security for oauth which depends on spring mvc it has been found that scs client apps are not compatible with spring cloud gateway and more specifically spring webflux given that the spring framework is promoting reactive programming heavily and given that spring cloud gateway is replacing spring cloud netflix zuul as the gateway of choice it is likely that scs customers will want to use webflux either for api purposes or using webclient instead of resttemplate or both therefore we should aim to support webflux as an option in scs client applications note the incompatibility appears to come from our reliance on for security between clients and the services and since does not support webflux we may should consider moving toward spring security s client side oauth support working with the spring security team to ensure that our needs are met for client side applications also note that this is largely a starters connectors issue and there is no anticipated impact to the broker or backing apps ,0
+178607,21509439115.0,IssuesEvent,2022-04-28 01:41:22,Baneeishaque/ask-med-pharma_Wordpress,https://api.github.com/repos/Baneeishaque/ask-med-pharma_Wordpress,closed,WS-2019-0184 (High) detected in lodash.merge-3.3.2.tgz - autoclosed,security vulnerability,"## WS-2019-0184 - High Severity Vulnerability
+ Vulnerable Library - lodash.merge-3.3.2.tgz
+
+
The modern build of lodash���s `_.merge` as a module.
Path to dependency file: /tmp/ws-scm/ask-med-pharma_Wordpress/wp-content/plugins/the-events-calendar/common/vendor/faction23/a11y-dialog/package.json
+
Path to vulnerable library: /tmp/ws-scm/ask-med-pharma_Wordpress/wp-content/plugins/the-events-calendar/common/vendor/faction23/a11y-dialog/node_modules/lodash.merge/package.json
+
+lodash.merge before 4.6.1 is vulnerable to Prototype Pollution. The function merge() may allow a malicious user to modify the prototype of Object via __proto__ causing the addition or modification of an existing property that will exist on all objects.
+
+
Path to dependency file: /tmp/ws-scm/ask-med-pharma_Wordpress/wp-content/plugins/the-events-calendar/common/vendor/faction23/a11y-dialog/package.json
+
Path to vulnerable library: /tmp/ws-scm/ask-med-pharma_Wordpress/wp-content/plugins/the-events-calendar/common/vendor/faction23/a11y-dialog/node_modules/lodash.merge/package.json
+
+lodash.merge before 4.6.1 is vulnerable to Prototype Pollution. The function merge() may allow a malicious user to modify the prototype of Object via __proto__ causing the addition or modification of an existing property that will exist on all objects.
+
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",0,ws high detected in lodash merge tgz autoclosed ws high severity vulnerability vulnerable library lodash merge tgz the modern build of lodash���s merge as a module library home page a href path to dependency file tmp ws scm ask med pharma wordpress wp content plugins the events calendar common vendor dialog package json path to vulnerable library tmp ws scm ask med pharma wordpress wp content plugins the events calendar common vendor dialog node modules lodash merge package json dependency hierarchy semistandard tgz root library eslint tgz x lodash merge tgz vulnerable library found in head commit a href vulnerability details lodash merge before is vulnerable to prototype pollution the function merge may allow a malicious user to modify the prototype of object via proto causing the addition or modification of an existing property that will exist on all objects publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource ,0
+146147,11727650115.0,IssuesEvent,2020-03-10 16:15:50,mautic/mautic,https://api.github.com/repos/mautic/mautic,closed,The getLeadIpLogs function causes a 'Incorrect DATETIME value' error,Bug Ready To Test,"| Q | A
+| --- | ---
+| Mautic version | v2.15.2-beta
+| PHP version | PHP 7.2.19-1+ubuntu18.04.1+deb.sury.org+1
+| MySQL version | Ver 8.0.16 for Linux on x86_64 (MySQL Community Server - GPL)
+| Browser | n/a
+
+## Bug Description
+
+This is a bit of a weird one and I'm not sure how to replicate the issue as it only happens on our production server, but I've identified where the problem is and how to fix it.
+
+In https://github.com/mautic/mautic/blob/staging/app/bundles/CoreBundle/Entity/AuditLogRepository.php we have the function `getLeadIpLogs`
+
+Which contains the following line:
+
+```php
+$sqb->expr()->gte('l.date_added', $sqb->expr()->literal($dt->getUtcTimestamp()))
+```
+
+When executed, I get the error detailed at the bottom of this issue.
+
+Long story short, checking for a timestamp is causing this error.
+
+## (Possible) Solution
+
+By changing the above line to:
+
+```php
+$sqb->expr()->gte('l.date_added', $sqb->expr()->literal($lead->getDateAdded()->format('Y-m-d H:i:s')))
+```
+
+Everything works as expected. So simply put, the problem seems to be that using a timestamp instead of a datetime string is causing a problem.
+
+### Log errors
+
+```
+[2019-06-17 14:38:13] mautic.CRITICAL:
+
+Uncaught PHP Exception Doctrine\DBAL\Exception\DriverException:
+""An exception occurred while executing 'SELECT ip.date_added, ip.ip_address, ip.lead_id, ip.id FROM (SELECT MAX(l.date_added) as date_added, MIN(l.id) as id, l.ip_address, l.object_id as lead_id FROM audit_log l WHERE (l.bundle = 'lead') AND (l.object = 'lead') AND (l.action = 'ipadded') AND ((l.object_id = 71) AND (l.date_added >= '1560775838')) GROUP BY l.ip_address) ip ORDER BY ip.date_added DESC LIMIT 25': SQLSTATE[HY000]: General error: 1525 Incorrect DATETIME value: '1560775838'"" at /vendor/doctrine/dbal/lib/Doctrine/DBAL/Driver/AbstractMySQLDriver.php line 115 {""exception"":""[object] (Doctrine\\DBAL\\Exception\\DriverException(code: 0): An exception occurred while executing 'SELECT ip.date_added, ip.ip_address, ip.lead_id, ip.id FROM (SELECT MAX(l.date_added) as date_added, MIN(l.id) as id, l.ip_address, l.object_id as lead_id FROM audit_log l WHERE (l.bundle = 'lead') AND (l.object = 'lead') AND (l.action = 'ipadded') AND ((l.object_id = 71) AND (l.date_added >= '1560775838')) GROUP BY l.ip_address) ip ORDER BY ip.date_added DESC LIMIT 25': SQLSTATE[HY000]: General error: 1525 Incorrect DATETIME value: '1560775838'
+at /vendor/doctrine/dbal/lib/Doctrine/DBAL/Driver/AbstractMySQLDriver.php:115
+Doctrine\\DBAL\\Driver\\PDOException(code: HY000): SQLSTATE[HY000]: General error: 1525 Incorrect DATETIME value: '1560775838' at/vendor/doctrine/dbal/lib/Doctrine/DBAL/Driver/PDOConnection.php:106, PDOException(code: HY000): SQLSTATE[HY000]: General error: 1525 Incorrect DATETIME value: '1560775838' at /vendor/doctrine/dbal/lib/Doctrine/DBAL/Driver/PDOConnection.php:104)""} []
+
+```",1.0,"The getLeadIpLogs function causes a 'Incorrect DATETIME value' error - | Q | A
+| --- | ---
+| Mautic version | v2.15.2-beta
+| PHP version | PHP 7.2.19-1+ubuntu18.04.1+deb.sury.org+1
+| MySQL version | Ver 8.0.16 for Linux on x86_64 (MySQL Community Server - GPL)
+| Browser | n/a
+
+## Bug Description
+
+This is a bit of a weird one and I'm not sure how to replicate the issue as it only happens on our production server, but I've identified where the problem is and how to fix it.
+
+In https://github.com/mautic/mautic/blob/staging/app/bundles/CoreBundle/Entity/AuditLogRepository.php we have the function `getLeadIpLogs`
+
+Which contains the following line:
+
+```php
+$sqb->expr()->gte('l.date_added', $sqb->expr()->literal($dt->getUtcTimestamp()))
+```
+
+When executed, I get the error detailed at the bottom of this issue.
+
+Long story short, checking for a timestamp is causing this error.
+
+## (Possible) Solution
+
+By changing the above line to:
+
+```php
+$sqb->expr()->gte('l.date_added', $sqb->expr()->literal($lead->getDateAdded()->format('Y-m-d H:i:s')))
+```
+
+Everything works as expected. So simply put, the problem seems to be that using a timestamp instead of a datetime string is causing a problem.
+
+### Log errors
+
+```
+[2019-06-17 14:38:13] mautic.CRITICAL:
+
+Uncaught PHP Exception Doctrine\DBAL\Exception\DriverException:
+""An exception occurred while executing 'SELECT ip.date_added, ip.ip_address, ip.lead_id, ip.id FROM (SELECT MAX(l.date_added) as date_added, MIN(l.id) as id, l.ip_address, l.object_id as lead_id FROM audit_log l WHERE (l.bundle = 'lead') AND (l.object = 'lead') AND (l.action = 'ipadded') AND ((l.object_id = 71) AND (l.date_added >= '1560775838')) GROUP BY l.ip_address) ip ORDER BY ip.date_added DESC LIMIT 25': SQLSTATE[HY000]: General error: 1525 Incorrect DATETIME value: '1560775838'"" at /vendor/doctrine/dbal/lib/Doctrine/DBAL/Driver/AbstractMySQLDriver.php line 115 {""exception"":""[object] (Doctrine\\DBAL\\Exception\\DriverException(code: 0): An exception occurred while executing 'SELECT ip.date_added, ip.ip_address, ip.lead_id, ip.id FROM (SELECT MAX(l.date_added) as date_added, MIN(l.id) as id, l.ip_address, l.object_id as lead_id FROM audit_log l WHERE (l.bundle = 'lead') AND (l.object = 'lead') AND (l.action = 'ipadded') AND ((l.object_id = 71) AND (l.date_added >= '1560775838')) GROUP BY l.ip_address) ip ORDER BY ip.date_added DESC LIMIT 25': SQLSTATE[HY000]: General error: 1525 Incorrect DATETIME value: '1560775838'
+at /vendor/doctrine/dbal/lib/Doctrine/DBAL/Driver/AbstractMySQLDriver.php:115
+Doctrine\\DBAL\\Driver\\PDOException(code: HY000): SQLSTATE[HY000]: General error: 1525 Incorrect DATETIME value: '1560775838' at/vendor/doctrine/dbal/lib/Doctrine/DBAL/Driver/PDOConnection.php:106, PDOException(code: HY000): SQLSTATE[HY000]: General error: 1525 Incorrect DATETIME value: '1560775838' at /vendor/doctrine/dbal/lib/Doctrine/DBAL/Driver/PDOConnection.php:104)""} []
+
+```",0,the getleadiplogs function causes a incorrect datetime value error q a mautic version beta php version php deb sury org mysql version ver for linux on mysql community server gpl browser n a bug description this is a bit of a weird one and i m not sure how to replicate the issue as it only happens on our production server but i ve identified where the problem is and how to fix it in we have the function getleadiplogs which contains the following line php sqb expr gte l date added sqb expr literal dt getutctimestamp when executed i get the error detailed at the bottom of this issue long story short checking for a timestamp is causing this error possible solution by changing the above line to php sqb expr gte l date added sqb expr literal lead getdateadded format y m d h i s everything works as expected so simply put the problem seems to be that using a timestamp instead of a datetime string is causing a problem log errors mautic critical uncaught php exception doctrine dbal exception driverexception an exception occurred while executing select ip date added ip ip address ip lead id ip id from select max l date added as date added min l id as id l ip address l object id as lead id from audit log l where l bundle lead and l object lead and l action ipadded and l object id and l date added group by l ip address ip order by ip date added desc limit sqlstate general error incorrect datetime value at vendor doctrine dbal lib doctrine dbal driver abstractmysqldriver php line exception doctrine dbal exception driverexception code an exception occurred while executing select ip date added ip ip address ip lead id ip id from select max l date added as date added min l id as id l ip address l object id as lead id from audit log l where l bundle lead and l object lead and l action ipadded and l object id and l date added group by l ip address ip order by ip date added desc limit sqlstate general error incorrect datetime value at vendor doctrine dbal lib doctrine dbal driver abstractmysqldriver php doctrine dbal driver pdoexception code sqlstate general error incorrect datetime value at vendor doctrine dbal lib doctrine dbal driver pdoconnection php pdoexception code sqlstate general error incorrect datetime value at vendor doctrine dbal lib doctrine dbal driver pdoconnection php ,0
+699,2582311636.0,IssuesEvent,2015-02-15 03:02:33,cakephp/cakephp,https://api.github.com/repos/cakephp/cakephp,closed,"""created' and ""modified"" receive unexpected datetime value.",Defect On hold ORM,"After upgrading to latest 3.0, tried to save a model that worked before the upgrade:
+Error: SQLSTATE[22007]: Invalid datetime format: 1292 Incorrect datetime value: '2014-12-09T18:13:54Z' for column 'created' at row 1
+
+Both ""created"" and ""modified"" fields are set to type datetime.",1.0,"""created' and ""modified"" receive unexpected datetime value. - After upgrading to latest 3.0, tried to save a model that worked before the upgrade:
+Error: SQLSTATE[22007]: Invalid datetime format: 1292 Incorrect datetime value: '2014-12-09T18:13:54Z' for column 'created' at row 1
+
+Both ""created"" and ""modified"" fields are set to type datetime.",0, created and modified receive unexpected datetime value after upgrading to latest tried to save a model that worked before the upgrade error sqlstate invalid datetime format incorrect datetime value for column created at row both created and modified fields are set to type datetime ,0
+42983,5511371803.0,IssuesEvent,2017-03-17 04:28:15,jelmer/xandikos,https://api.github.com/repos/jelmer/xandikos,opened,list subcollections,bug client-compat vdirsyncer-tests,"Any store-based collections currently only return object members, and not subcollections.",1.0,"list subcollections - Any store-based collections currently only return object members, and not subcollections.",0,list subcollections any store based collections currently only return object members and not subcollections ,0
+391,2496218966.0,IssuesEvent,2015-01-06 17:54:07,nlbdev/nordic-epub3-dtbook-migrator,https://api.github.com/repos/nlbdev/nordic-epub3-dtbook-migrator,opened,"Check that ""mailto:"" links are handled correctly when converting to DTBook",0 - Low priority enhancement epub3-to-dtbook,"`external=""true""` should be added to links when converting mailto addresses.
+
+By default we don't even keep the links after conversion, so this is a pretty low-priority issue. But it's worth getting into the code so that a possible future ""generic DTBook"" fork can support this properly.",1.0,"Check that ""mailto:"" links are handled correctly when converting to DTBook - `external=""true""` should be added to links when converting mailto addresses.
+
+By default we don't even keep the links after conversion, so this is a pretty low-priority issue. But it's worth getting into the code so that a possible future ""generic DTBook"" fork can support this properly.",0,check that mailto links are handled correctly when converting to dtbook external true should be added to links when converting mailto addresses by default we don t even keep the links after conversion so this is a pretty low priority issue but it s worth getting into the code so that a possible future generic dtbook fork can support this properly ,0
+747,13465143769.0,IssuesEvent,2020-09-09 20:22:13,microsoftgraph/microsoft-graph-toolkit,https://api.github.com/repos/microsoftgraph/microsoft-graph-toolkit,closed,Add dark theme to mgt-people-picker ,Area: Components Component: people-picker State: In Review State: Started feature-request,"# Proposal: Add dark theme to mgt-people-picker
+
+## Description
+0. Add common colors and mixins base on design
+1. Revamp mgt-people-picker css classes and structure
+2. Add scss variables and mixins to support dark theme
+
+## Rationale
+MGT components right now support light theme. Developer community has asked for dark theme, Microsoft Teams theme, and custom themes.
+
+## Preferred Solution
+Enable developers to configure mgt-people-picker to show dark theme by adding attribute theme=""dark"".
+Enable developers to configure mgt-people-picker to show custom theme by adding attribute theme=""my-theme"" (attribute value in discussion)
+
+Design by @tramadon: https://www.figma.com/file/tVj3DubCqeip6mhw14Up0j/Microsoft-Graph-Toolkit-(main)?node-id=1111%3A19711",1.0,"Add dark theme to mgt-people-picker - # Proposal: Add dark theme to mgt-people-picker
+
+## Description
+0. Add common colors and mixins base on design
+1. Revamp mgt-people-picker css classes and structure
+2. Add scss variables and mixins to support dark theme
+
+## Rationale
+MGT components right now support light theme. Developer community has asked for dark theme, Microsoft Teams theme, and custom themes.
+
+## Preferred Solution
+Enable developers to configure mgt-people-picker to show dark theme by adding attribute theme=""dark"".
+Enable developers to configure mgt-people-picker to show custom theme by adding attribute theme=""my-theme"" (attribute value in discussion)
+
+Design by @tramadon: https://www.figma.com/file/tVj3DubCqeip6mhw14Up0j/Microsoft-Graph-Toolkit-(main)?node-id=1111%3A19711",1,add dark theme to mgt people picker proposal add dark theme to mgt people picker description add common colors and mixins base on design revamp mgt people picker css classes and structure add scss variables and mixins to support dark theme rationale mgt components right now support light theme developer community has asked for dark theme microsoft teams theme and custom themes preferred solution enable developers to configure mgt people picker to show dark theme by adding attribute theme dark enable developers to configure mgt people picker to show custom theme by adding attribute theme my theme attribute value in discussion design by tramadon ,1
+149754,11914033357.0,IssuesEvent,2020-03-31 13:00:20,pingcap/tidb,https://api.github.com/repos/pingcap/tidb,opened,store_test.go:testStoreSuite.TestOracleChangeByFailpoint failed,component/test,"store_test.go:testStoreSuite.TestOracleChangeByFailpoint
+
+Latest failed builds:
+https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_check_2/29453/display/redirect
+https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_unit_test/29714/display/redirect
+https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_unit_test/29897/display/redirect
+https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_unit_test/29791/display/redirect
+https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_unit_test/29722/display/redirect
+https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_unit_test/29764/display/redirect
+",1.0,"store_test.go:testStoreSuite.TestOracleChangeByFailpoint failed - store_test.go:testStoreSuite.TestOracleChangeByFailpoint
+
+Latest failed builds:
+https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_check_2/29453/display/redirect
+https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_unit_test/29714/display/redirect
+https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_unit_test/29897/display/redirect
+https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_unit_test/29791/display/redirect
+https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_unit_test/29722/display/redirect
+https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_unit_test/29764/display/redirect
+",0,store test go teststoresuite testoraclechangebyfailpoint failed store test go teststoresuite testoraclechangebyfailpoint latest failed builds ,0
+1117,27089394271.0,IssuesEvent,2023-02-14 19:39:17,openstates/issues,https://api.github.com/repos/openstates/issues,closed,New DE Committee Scraper,good first issue component:people-data good first scraper,"### We need a new committee scraper written in [spatula](https://jamesturk.github.io/spatula/) for Delaware.
+
+It should scrape this [webpage of Senate Committees](https://legis.delaware.gov/Committees/Senate), this [webpage of House Committees](https://legis.delaware.gov/Committees/House), and this [webpage of Joint Committees](https://legis.delaware.gov/Committees/Joint) to get:
+- **name**
+ - “Small Business”, “Agriculture”, etc.
+- **chamber**
+ - “upper”, “lower”, or “legislature” (when joint, etc.)
+- **classification**
+ - ex: ""committee"", ""subcommittee""
+- **parent**
+ - *only if it is a subcommittee* - scraper should have a way to determine this
+ - ex: parent=""Natural Resources"" when *sub*committee name=""Forestry""
+- **sources**
+ - each added using `add_source()` method on instance of `ScrapeCommittee` type object
+ - ex: home page for list of committees, specific page for that committee, etc.
+- **members**
+ - - each added using `add_member()`method on instance of `ScrapeCommittee` type object
+ - *Attributes:*
+ - **name**: “Jane Doe”, “John Smith”, etc.
+ - **role** (where applicable): “Chair”, “Ranking Member”, etc.
+
+#### [Here is a helpful general reference doc](https://docs.openstates.org/contributing/writing-a-committee-scraper/) for those new to writing scrapers, the Open States project, or the specific task of writing a committee scraper
+
+### Useful scrapers for reference
+An [old non-spatula-based scraper](https://github.com/openstates/openstates-scrapers/blob/main/scrapers/de/committees.py) that may be helpful in understanding how we have captured the data in the past.
+
+A [spatula-based scraper](https://github.com/openstates/openstates-scrapers/blob/main/scrapers_next/mo/committees.py) that also scrapes an HTML List Page (using `HtmlListPage` and `HtmlPage` spatula classes) for reference.
+- As in the other spatula scraper, you will need to set the session variable separately (rather than depend on the `__init__.py`).
+
+### Other useful resources
+You can reference the [object models for committee scrapers](https://github.com/openstates/openstates-core/blob/main/openstates/models/committees.py) in the Open States Core repo.
+
+Further documentation on running spatula scrapers in the command line can be found [here](https://docs.openstates.org/contributing/scrapers/#running-spatula-scrapers).",1.0,"New DE Committee Scraper - ### We need a new committee scraper written in [spatula](https://jamesturk.github.io/spatula/) for Delaware.
+
+It should scrape this [webpage of Senate Committees](https://legis.delaware.gov/Committees/Senate), this [webpage of House Committees](https://legis.delaware.gov/Committees/House), and this [webpage of Joint Committees](https://legis.delaware.gov/Committees/Joint) to get:
+- **name**
+ - “Small Business”, “Agriculture”, etc.
+- **chamber**
+ - “upper”, “lower”, or “legislature” (when joint, etc.)
+- **classification**
+ - ex: ""committee"", ""subcommittee""
+- **parent**
+ - *only if it is a subcommittee* - scraper should have a way to determine this
+ - ex: parent=""Natural Resources"" when *sub*committee name=""Forestry""
+- **sources**
+ - each added using `add_source()` method on instance of `ScrapeCommittee` type object
+ - ex: home page for list of committees, specific page for that committee, etc.
+- **members**
+ - - each added using `add_member()`method on instance of `ScrapeCommittee` type object
+ - *Attributes:*
+ - **name**: “Jane Doe”, “John Smith”, etc.
+ - **role** (where applicable): “Chair”, “Ranking Member”, etc.
+
+#### [Here is a helpful general reference doc](https://docs.openstates.org/contributing/writing-a-committee-scraper/) for those new to writing scrapers, the Open States project, or the specific task of writing a committee scraper
+
+### Useful scrapers for reference
+An [old non-spatula-based scraper](https://github.com/openstates/openstates-scrapers/blob/main/scrapers/de/committees.py) that may be helpful in understanding how we have captured the data in the past.
+
+A [spatula-based scraper](https://github.com/openstates/openstates-scrapers/blob/main/scrapers_next/mo/committees.py) that also scrapes an HTML List Page (using `HtmlListPage` and `HtmlPage` spatula classes) for reference.
+- As in the other spatula scraper, you will need to set the session variable separately (rather than depend on the `__init__.py`).
+
+### Other useful resources
+You can reference the [object models for committee scrapers](https://github.com/openstates/openstates-core/blob/main/openstates/models/committees.py) in the Open States Core repo.
+
+Further documentation on running spatula scrapers in the command line can be found [here](https://docs.openstates.org/contributing/scrapers/#running-spatula-scrapers).",1,new de committee scraper we need a new committee scraper written in for delaware it should scrape this this and this to get name “small business” “agriculture” etc chamber “upper” “lower” or “legislature” when joint etc classification ex committee subcommittee parent only if it is a subcommittee scraper should have a way to determine this ex parent natural resources when sub committee name forestry sources each added using add source method on instance of scrapecommittee type object ex home page for list of committees specific page for that committee etc members each added using add member method on instance of scrapecommittee type object attributes name “jane doe” “john smith” etc role where applicable “chair” “ranking member” etc for those new to writing scrapers the open states project or the specific task of writing a committee scraper useful scrapers for reference an that may be helpful in understanding how we have captured the data in the past a that also scrapes an html list page using htmllistpage and htmlpage spatula classes for reference as in the other spatula scraper you will need to set the session variable separately rather than depend on the init py other useful resources you can reference the in the open states core repo further documentation on running spatula scrapers in the command line can be found ,1
+66601,12805844517.0,IssuesEvent,2020-07-03 08:20:06,danglotb/skillful_network,https://api.github.com/repos/danglotb/skillful_network,closed,%3.2 FEAT: Refonte register,code enhancement,"* Tempory code becomes current password
+* Add template form registration-confirmation
+* Add new feature in registration-confirmation to get more informations directly: firstName, lastName, and role.",1.0,"%3.2 FEAT: Refonte register - * Tempory code becomes current password
+* Add template form registration-confirmation
+* Add new feature in registration-confirmation to get more informations directly: firstName, lastName, and role.",0, feat refonte register tempory code becomes current password add template form registration confirmation add new feature in registration confirmation to get more informations directly firstname lastname and role ,0
+409,7609550177.0,IssuesEvent,2018-05-01 01:14:18,schmave/demschooltools,https://api.github.com/repos/schmave/demschooltools,opened,Add stats on the tag history page,people,"Show total count of people ever with this tag.
+Show table/graph of monthly/yearly count of people with this tag.",1.0,"Add stats on the tag history page - Show total count of people ever with this tag.
+Show table/graph of monthly/yearly count of people with this tag.",1,add stats on the tag history page show total count of people ever with this tag show table graph of monthly yearly count of people with this tag ,1
+188040,15114472669.0,IssuesEvent,2021-02-09 02:00:41,golang/go,https://api.github.com/repos/golang/go,closed,runtime/metrics: typos in documentation,Documentation NeedsFix,"There are some typos in the documentation of the new ""runtime/metrics"" package
+
+* `encouranged` should be `encouraged`
+* `""kind.""` should be `""kind"".`
+* `the a tag` should be `the tag`",1.0,"runtime/metrics: typos in documentation - There are some typos in the documentation of the new ""runtime/metrics"" package
+
+* `encouranged` should be `encouraged`
+* `""kind.""` should be `""kind"".`
+* `the a tag` should be `the tag`",0,runtime metrics typos in documentation there are some typos in the documentation of the new runtime metrics package encouranged should be encouraged kind should be kind the a tag should be the tag ,0
+1176,30668987538.0,IssuesEvent,2023-07-25 20:39:10,restincode/restincode,https://api.github.com/repos/restincode/restincode,opened,Jerome Heckenkamp (Sk8),People Add Person Needs Review,"Please fill out as much information as you can. No fields are required, but the more you can provide the better.
+
+**General Info**
+* First name: Jerome
+* Last name: Heckenkamp
+* Handle: SK8. MagicFX
+* Birth Year: 1979
+* Death Year: 2016
+* Link to Obituary: https://www.legacy.com/funeral-homes/obituaries/name/jerome-heckenkamp-obituary?pid=178007168&v=batesville
+* Group Affiliations:
+* URL to main photo (or attach to Issue):
+* Description of person and/or activities:
+* Facebook memorial group URL: https://www.jeromeheckenkamp.com/
+
+**Social Media Links**
+* Twitter:
+* Github:
+* LinkedIn:
+* Facebook:
+* Other:
+
+**Contributions**
+Repeat as many times as needed
+
+* Project name:
+* Project URL:
+* Project Description:
+
+**Photo Gallery**
+* URL(s) to additional photos:
+
+",1.0,"Jerome Heckenkamp (Sk8) - Please fill out as much information as you can. No fields are required, but the more you can provide the better.
+
+**General Info**
+* First name: Jerome
+* Last name: Heckenkamp
+* Handle: SK8. MagicFX
+* Birth Year: 1979
+* Death Year: 2016
+* Link to Obituary: https://www.legacy.com/funeral-homes/obituaries/name/jerome-heckenkamp-obituary?pid=178007168&v=batesville
+* Group Affiliations:
+* URL to main photo (or attach to Issue):
+* Description of person and/or activities:
+* Facebook memorial group URL: https://www.jeromeheckenkamp.com/
+
+**Social Media Links**
+* Twitter:
+* Github:
+* LinkedIn:
+* Facebook:
+* Other:
+
+**Contributions**
+Repeat as many times as needed
+
+* Project name:
+* Project URL:
+* Project Description:
+
+**Photo Gallery**
+* URL(s) to additional photos:
+
+",1,jerome heckenkamp please fill out as much information as you can no fields are required but the more you can provide the better general info first name jerome last name heckenkamp handle magicfx birth year death year link to obituary group affiliations url to main photo or attach to issue description of person and or activities facebook memorial group url social media links twitter github linkedin facebook other contributions repeat as many times as needed project name project url project description photo gallery url s to additional photos ,1
+603,10606799827.0,IssuesEvent,2019-10-11 00:57:48,AY1920S1-CS2103T-T13-3/main,https://api.github.com/repos/AY1920S1-CS2103T-T13-3/main,closed,"As a user, I want to have information of the person I need to reimburse",Integration People UseCase status.Ongoing type.Story,so that I can easily find information to contact the person for reimbursement,1.0,"As a user, I want to have information of the person I need to reimburse - so that I can easily find information to contact the person for reimbursement",1,as a user i want to have information of the person i need to reimburse so that i can easily find information to contact the person for reimbursement,1
+1545,2608987514.0,IssuesEvent,2015-02-26 11:30:17,llaville/php-reflect,https://api.github.com/repos/llaville/php-reflect,closed,Roadmap to version 3.0,documentation,"If you want to have more feature than currently existing in branch 2, don't hesitate to open a request here.
+If it's accepted, it will be included in this roadmap.
+
+# Next major release
+
+## alpha-1
+*due date 2014-12-22*
+
+- [x] improve analysis speed. See issue #9
+- [x] reduce memory consumption. See issue #9
+- [x] ability to provide a file or a directory in command line, without to specify it in `phpreflect.json` config file
+- [x] use Symfony Component 2.5 or better
+- [x] replace Symfony Console `TableHelper`, by `Table` helper (Ready for Symfony 3)
+- [x] replace Symfony Console `ProgressHelper`, by `ProgressBar` helper (Ready for Symfony 3)
+- [x] remove old component `ProviderManager`, and used instead directly a Symfony `Finder` instance
+- [x] ability to disable all plugins with option `--no-plugins` (ala Composer)
+- [x] add a `PluginManager` to handle all plugins with configuration and the programmation
+- [x] Similar API with other SAPI than the CLI version
+- [x] Write CLI result to file or URL (E.g: FTP)
+
+## alpha-2
+*due date 2015-01-05*
+
+Stabilize and synchronize with CompatInfo 4.0 alpha-2
+
+## alpha-3
+*due date 2015-01-15*
+
+Last alpha version, before stabilization of final API 3.0
+- [x] checking pre-conditions before traversing AST
+
+## stable
+*in 2015 but no due date*
+
+- [x] full documentation available
+ - [x] a user guide
+ - [x] a developer guide
+ - [x] a migration guide
+- [x] full unit tests passing",1.0,"Roadmap to version 3.0 - If you want to have more feature than currently existing in branch 2, don't hesitate to open a request here.
+If it's accepted, it will be included in this roadmap.
+
+# Next major release
+
+## alpha-1
+*due date 2014-12-22*
+
+- [x] improve analysis speed. See issue #9
+- [x] reduce memory consumption. See issue #9
+- [x] ability to provide a file or a directory in command line, without to specify it in `phpreflect.json` config file
+- [x] use Symfony Component 2.5 or better
+- [x] replace Symfony Console `TableHelper`, by `Table` helper (Ready for Symfony 3)
+- [x] replace Symfony Console `ProgressHelper`, by `ProgressBar` helper (Ready for Symfony 3)
+- [x] remove old component `ProviderManager`, and used instead directly a Symfony `Finder` instance
+- [x] ability to disable all plugins with option `--no-plugins` (ala Composer)
+- [x] add a `PluginManager` to handle all plugins with configuration and the programmation
+- [x] Similar API with other SAPI than the CLI version
+- [x] Write CLI result to file or URL (E.g: FTP)
+
+## alpha-2
+*due date 2015-01-05*
+
+Stabilize and synchronize with CompatInfo 4.0 alpha-2
+
+## alpha-3
+*due date 2015-01-15*
+
+Last alpha version, before stabilization of final API 3.0
+- [x] checking pre-conditions before traversing AST
+
+## stable
+*in 2015 but no due date*
+
+- [x] full documentation available
+ - [x] a user guide
+ - [x] a developer guide
+ - [x] a migration guide
+- [x] full unit tests passing",0,roadmap to version if you want to have more feature than currently existing in branch don t hesitate to open a request here if it s accepted it will be included in this roadmap next major release alpha due date improve analysis speed see issue reduce memory consumption see issue ability to provide a file or a directory in command line without to specify it in phpreflect json config file use symfony component or better replace symfony console tablehelper by table helper ready for symfony replace symfony console progresshelper by progressbar helper ready for symfony remove old component providermanager and used instead directly a symfony finder instance ability to disable all plugins with option no plugins ala composer add a pluginmanager to handle all plugins with configuration and the programmation similar api with other sapi than the cli version write cli result to file or url e g ftp alpha due date stabilize and synchronize with compatinfo alpha alpha due date last alpha version before stabilization of final api checking pre conditions before traversing ast stable in but no due date full documentation available a user guide a developer guide a migration guide full unit tests passing,0
+830,15363554639.0,IssuesEvent,2021-03-01 20:54:44,microsoft/fluentui,https://api.github.com/repos/microsoft/fluentui,closed,Accessibility Issue with PeoplePicker component,Area: Accessibility Area: NVDA Component: PeoplePicker Fluent UI react Needs: Investigation Status: In PR,"PWD Impact:
+
+Visually challenged users who rely on screen readers are unable to use People picker, as NVDA is not navigating to the People picker field when navigated using form mode(f) or browse mode(down arrow key).
+
+
+
+Test Environment:
+
+OS : Win 10 2004 (OS Build 19041.329)
+
+Browser: New Edge 83.0.478.58
+
+URL: https://developer.microsoft.com/en-us/fluentui#/controls/web/peoplepicker
+
+Tools used: NVDA 2020.1
+
+
+
+Repro Steps:
+
+Step 1: Enable screen reader NVDA and open the above URL in New Edge browser.
+
+Step 2: People picker screen will be displayed.
+
+Step 3: Now navigate through the page in form mode(f) or browse mode(down arrow ) and verify whether NVDA is navigating to people picker field.
+
+
+
+Actual Result:
+
+NVDA is not navigating to the people picker field when navigated using form mode(f) or browse mode(down arrow key). NVDA is navigating only to the label and not moving to the edit field. (Only the ListPeoplePicker component can get navigated and open the combobox when using NVDA form mode.)
+
+
+
+Expected Result:
+
+NVDA should navigate to the people picker field when navigated using form mode(f) or browse mode(down arrow key).
+
+
+MAS References:
+
+MAS 1.3.2 - Meaningful Sequence
+
+
+
+WCAG Reference:
+
+https://www.w3.org/WAI/WCAG21/Understanding/meaningful-sequence.html
+
+",1.0,"Accessibility Issue with PeoplePicker component - PWD Impact:
+
+Visually challenged users who rely on screen readers are unable to use People picker, as NVDA is not navigating to the People picker field when navigated using form mode(f) or browse mode(down arrow key).
+
+
+
+Test Environment:
+
+OS : Win 10 2004 (OS Build 19041.329)
+
+Browser: New Edge 83.0.478.58
+
+URL: https://developer.microsoft.com/en-us/fluentui#/controls/web/peoplepicker
+
+Tools used: NVDA 2020.1
+
+
+
+Repro Steps:
+
+Step 1: Enable screen reader NVDA and open the above URL in New Edge browser.
+
+Step 2: People picker screen will be displayed.
+
+Step 3: Now navigate through the page in form mode(f) or browse mode(down arrow ) and verify whether NVDA is navigating to people picker field.
+
+
+
+Actual Result:
+
+NVDA is not navigating to the people picker field when navigated using form mode(f) or browse mode(down arrow key). NVDA is navigating only to the label and not moving to the edit field. (Only the ListPeoplePicker component can get navigated and open the combobox when using NVDA form mode.)
+
+
+
+Expected Result:
+
+NVDA should navigate to the people picker field when navigated using form mode(f) or browse mode(down arrow key).
+
+
+MAS References:
+
+MAS 1.3.2 - Meaningful Sequence
+
+
+
+WCAG Reference:
+
+https://www.w3.org/WAI/WCAG21/Understanding/meaningful-sequence.html
+
+",1,accessibility issue with peoplepicker component pwd impact visually challenged users who rely on screen readers are unable to use people picker as nvda is not navigating to the people picker field when navigated using form mode f or browse mode down arrow key test environment os win os build browser new edge url tools used nvda repro steps step enable screen reader nvda and open the above url in new edge browser step people picker screen will be displayed step now navigate through the page in form mode f or browse mode down arrow and verify whether nvda is navigating to people picker field actual result nvda is not navigating to the people picker field when navigated using form mode f or browse mode down arrow key nvda is navigating only to the label and not moving to the edit field only the listpeoplepicker component can get navigated and open the combobox when using nvda form mode expected result nvda should navigate to the people picker field when navigated using form mode f or browse mode down arrow key mas references mas meaningful sequence wcag reference ,1
+599353,18271857692.0,IssuesEvent,2021-10-04 14:33:32,medialab/portic-storymaps-2021,https://api.github.com/repos/medialab/portic-storymaps-2021,opened,"Viz ""partie-2-carte-direction-bateaux-de-la-rochelle""",enhancement priority : medium,"Guillaume:
+
+> Souci avec ce graphique : l’importance de la Traite est minimisée, car les deux flèches vers la côte de l’Or et la côte de l’Angola sont superposées. Suggestion : pointer la flèche vers le golfe de Guinée plus à l’Ouest, du côté du Ghana.
+En plus, la Côte de l’Or, c’est effectivement le Ghana......
+
+Guillaume:
+
+En fait, je ne comprends pas ce graphique...
+1) Quelle règle détermine la couleur des flux si ni le port de départ ni celui d’arrivée sont La Rochelle ?
+2) Quel est l’univers des relevés de Navigo 1787 ? Que les congés français je suppose ?
+3) Lorsque je passe le curseur sur les flux jaunes vers Saint-Pierre et Miquelon, j’obtiens les ports de Saint-Martin-de-Ré ou les Sables-d’Olonnes.
+
+Réponse Silvia:
+
+1) le port d'attache = La Rochelle (j'ai reformulé pour que ce soit plus explicite)
+2) j'ai précisé
+3) oui car tu as des navires appartenant à des négociants de La Rochelle qui partent de ces autres ports (on ne représente que les navires dont le port d'attache est = La Rochelle
+
+Réponse Guillaume:
+
+1) et 2) c’est clair maintenant..
+3) ce n’est pas clair. Voir email avec capture d’écran.
+
+
+
+
+Réponse Silvia:
+
+Je pense que le toponyme St Martin est juste à la gauche du point concerné et donc à sa place et pas à Terre-Neuve. Ou alors je n'ai pas compris? Le souci de cette carte est qu'il est difficile de distingue rou se termine une carte et où commence l'autre. Au pire on regarde ça tous devant un écran le 14?
+
+Réponse Guillaume:
+
+Oui, mais le problème vient de la position de la souris à ce moment-là. Si on passe la souris sur la flèche trajet Saint-Pierre et Miquelon — New York dans la carte du bas à droite, c’est « Les Sables d’Olonne—Rochefort » qui apparaît. Pour avoir Saint-Pierre et Miquelon — New York, il faut avoir la souris sur les Grands Lacs (ou l’Ohio… ces coins-là).
+",1.0,"Viz ""partie-2-carte-direction-bateaux-de-la-rochelle"" - Guillaume:
+
+> Souci avec ce graphique : l’importance de la Traite est minimisée, car les deux flèches vers la côte de l’Or et la côte de l’Angola sont superposées. Suggestion : pointer la flèche vers le golfe de Guinée plus à l’Ouest, du côté du Ghana.
+En plus, la Côte de l’Or, c’est effectivement le Ghana......
+
+Guillaume:
+
+En fait, je ne comprends pas ce graphique...
+1) Quelle règle détermine la couleur des flux si ni le port de départ ni celui d’arrivée sont La Rochelle ?
+2) Quel est l’univers des relevés de Navigo 1787 ? Que les congés français je suppose ?
+3) Lorsque je passe le curseur sur les flux jaunes vers Saint-Pierre et Miquelon, j’obtiens les ports de Saint-Martin-de-Ré ou les Sables-d’Olonnes.
+
+Réponse Silvia:
+
+1) le port d'attache = La Rochelle (j'ai reformulé pour que ce soit plus explicite)
+2) j'ai précisé
+3) oui car tu as des navires appartenant à des négociants de La Rochelle qui partent de ces autres ports (on ne représente que les navires dont le port d'attache est = La Rochelle
+
+Réponse Guillaume:
+
+1) et 2) c’est clair maintenant..
+3) ce n’est pas clair. Voir email avec capture d’écran.
+
+
+
+
+Réponse Silvia:
+
+Je pense que le toponyme St Martin est juste à la gauche du point concerné et donc à sa place et pas à Terre-Neuve. Ou alors je n'ai pas compris? Le souci de cette carte est qu'il est difficile de distingue rou se termine une carte et où commence l'autre. Au pire on regarde ça tous devant un écran le 14?
+
+Réponse Guillaume:
+
+Oui, mais le problème vient de la position de la souris à ce moment-là. Si on passe la souris sur la flèche trajet Saint-Pierre et Miquelon — New York dans la carte du bas à droite, c’est « Les Sables d’Olonne—Rochefort » qui apparaît. Pour avoir Saint-Pierre et Miquelon — New York, il faut avoir la souris sur les Grands Lacs (ou l’Ohio… ces coins-là).
+",0,viz partie carte direction bateaux de la rochelle guillaume souci avec ce graphique l’importance de la traite est minimisée car les deux flèches vers la côte de l’or et la côte de l’angola sont superposées suggestion pointer la flèche vers le golfe de guinée plus à l’ouest du côté du ghana en plus la côte de l’or c’est effectivement le ghana guillaume en fait je ne comprends pas ce graphique quelle règle détermine la couleur des flux si ni le port de départ ni celui d’arrivée sont la rochelle quel est l’univers des relevés de navigo que les congés français je suppose lorsque je passe le curseur sur les flux jaunes vers saint pierre et miquelon j’obtiens les ports de saint martin de ré ou les sables d’olonnes réponse silvia le port d attache la rochelle j ai reformulé pour que ce soit plus explicite j ai précisé oui car tu as des navires appartenant à des négociants de la rochelle qui partent de ces autres ports on ne représente que les navires dont le port d attache est la rochelle réponse guillaume et c’est clair maintenant ce n’est pas clair voir email avec capture d’écran réponse silvia je pense que le toponyme st martin est juste à la gauche du point concerné et donc à sa place et pas à terre neuve ou alors je n ai pas compris le souci de cette carte est qu il est difficile de distingue rou se termine une carte et où commence l autre au pire on regarde ça tous devant un écran le réponse guillaume oui mais le problème vient de la position de la souris à ce moment là si on passe la souris sur la flèche trajet saint pierre et miquelon — new york dans la carte du bas à droite c’est « les sables d’olonne—rochefort » qui apparaît pour avoir saint pierre et miquelon — new york il faut avoir la souris sur les grands lacs ou l’ohio… ces coins là ,0
+455,8309525163.0,IssuesEvent,2018-09-24 07:06:28,softwareplanet/leadboard,https://api.github.com/repos/softwareplanet/leadboard,closed,User can see all people on people page (10),epic: people story,"## Criteria
+- [x] User can see contacts dropdown on leadbord
+- [x] User can see table with all organisations on people page
+- [x] Table should has styles:
+
+- [x] User can see info from custom fields for peoples in the table
+- [x] User can see organization connected with this person
+- [x] User can see amount of open and closed leads, owner and next activity date
+
+
+## Tasks
+- [x] Add styles to table (5)
+- [x] Disable edit possibility of table (1)
+- [x] Make table reusable (1)
+- [ ] Add tests (2)
+- [x] Click on a row will open the contact page (1)",1.0,"User can see all people on people page (10) - ## Criteria
+- [x] User can see contacts dropdown on leadbord
+- [x] User can see table with all organisations on people page
+- [x] Table should has styles:
+
+- [x] User can see info from custom fields for peoples in the table
+- [x] User can see organization connected with this person
+- [x] User can see amount of open and closed leads, owner and next activity date
+
+
+## Tasks
+- [x] Add styles to table (5)
+- [x] Disable edit possibility of table (1)
+- [x] Make table reusable (1)
+- [ ] Add tests (2)
+- [x] Click on a row will open the contact page (1)",1,user can see all people on people page criteria user can see contacts dropdown on leadbord user can see table with all organisations on people page table should has styles user can see info from custom fields for peoples in the table user can see organization connected with this person user can see amount of open and closed leads owner and next activity date tasks add styles to table disable edit possibility of table make table reusable add tests click on a row will open the contact page ,1
+1169,30631427854.0,IssuesEvent,2023-07-24 14:46:33,restincode/restincode,https://api.github.com/repos/restincode/restincode,opened,Richard Batka (Master Chemist),People Add Person Needs Review,"Please fill out as much information as you can. No fields are required, but the more you can provide the better.
+
+**General Info**
+* First name: Richard
+* Last name: Batka
+* Handle: Master Chemist
+* Birth Year:
+* Death Year: 2023
+* Link to Obituary:
+* Group Affiliations:
+* URL to main photo (or attach to Issue):
+* Description of person and/or activities:
+* Facebook memorial group URL:
+
+**Social Media Links**
+* Twitter:
+* Github:
+* LinkedIn:
+* Facebook:
+* Other:
+
+**Contributions**
+Repeat as many times as needed
+
+* Project name:
+* Project URL:
+* Project Description:
+
+**Photo Gallery**
+* URL(s) to additional photos:
+",1.0,"Richard Batka (Master Chemist) - Please fill out as much information as you can. No fields are required, but the more you can provide the better.
+
+**General Info**
+* First name: Richard
+* Last name: Batka
+* Handle: Master Chemist
+* Birth Year:
+* Death Year: 2023
+* Link to Obituary:
+* Group Affiliations:
+* URL to main photo (or attach to Issue):
+* Description of person and/or activities:
+* Facebook memorial group URL:
+
+**Social Media Links**
+* Twitter:
+* Github:
+* LinkedIn:
+* Facebook:
+* Other:
+
+**Contributions**
+Repeat as many times as needed
+
+* Project name:
+* Project URL:
+* Project Description:
+
+**Photo Gallery**
+* URL(s) to additional photos:
+",1,richard batka master chemist please fill out as much information as you can no fields are required but the more you can provide the better general info first name richard last name batka handle master chemist birth year death year link to obituary group affiliations url to main photo or attach to issue description of person and or activities facebook memorial group url social media links twitter github linkedin facebook other contributions repeat as many times as needed project name project url project description photo gallery url s to additional photos ,1
+247,5099072126.0,IssuesEvent,2017-01-04 06:01:29,PopulateTools/gobierto,https://api.github.com/repos/PopulateTools/gobierto,closed,Decide WYSIWYG for People Module bio + posts,gobierto-admin gobierto-people in progress,"re #177 - Any preference @ferblape @danguita ?
+
+Requirements:
+
+- should allow file uploading, for images, and attachments
+",1.0,"Decide WYSIWYG for People Module bio + posts - re #177 - Any preference @ferblape @danguita ?
+
+Requirements:
+
+- should allow file uploading, for images, and attachments
+",1,decide wysiwyg for people module bio posts re any preference ferblape danguita requirements should allow file uploading for images and attachments ,1
+204798,7091015600.0,IssuesEvent,2018-01-12 11:14:17,akvo/akvo-rsr,https://api.github.com/repos/akvo/akvo-rsr,closed,Newly added attachments not displayed until page is reloaded,Bug Priority: Medium Type: Results framework,"- Create a new update
+- Add an attachment
+- Saved update doesn't show the attachment
+- Reload the page, attachment is shown",1.0,"Newly added attachments not displayed until page is reloaded - - Create a new update
+- Add an attachment
+- Saved update doesn't show the attachment
+- Reload the page, attachment is shown",0,newly added attachments not displayed until page is reloaded create a new update add an attachment saved update doesn t show the attachment reload the page attachment is shown,0
+7378,6877508055.0,IssuesEvent,2017-11-20 08:20:03,wso2/carbon-uuf,https://api.github.com/repos/wso2/carbon-uuf,closed,Improve 'sendToClient' server-side JS function to output strictly JSON,Security,"Ether we load it as separate json file with json content type
+OR
+We encode it and using HTML encoding and parse using JSON.parse
+
+This will enhance the security of SendToClient
+
+see https://www.owasp.org/index.php/XSS_(Cross_Site_Scripting)_Prevention_Cheat_Sheet#RULE_.233.1_-_HTML_escape_JSON_values_in_an_HTML_context_and_read_the_data_with_JSON.parse",True,"Improve 'sendToClient' server-side JS function to output strictly JSON - Ether we load it as separate json file with json content type
+OR
+We encode it and using HTML encoding and parse using JSON.parse
+
+This will enhance the security of SendToClient
+
+see https://www.owasp.org/index.php/XSS_(Cross_Site_Scripting)_Prevention_Cheat_Sheet#RULE_.233.1_-_HTML_escape_JSON_values_in_an_HTML_context_and_read_the_data_with_JSON.parse",0,improve sendtoclient server side js function to output strictly json ether we load it as separate json file with json content type or we encode it and using html encoding and parse using json parse this will enhance the security of sendtoclient see ,0
+420,7804378675.0,IssuesEvent,2018-06-11 07:12:37,Brexit-Tears/iRekonJS,https://api.github.com/repos/Brexit-Tears/iRekonJS,opened,🇬🇧Break Travis Integration 🇬🇧,the people will it,"We made the decision to try out TravisCI... It's clearly not working.
+I think it's time to exit this integration.",1.0,"🇬🇧Break Travis Integration 🇬🇧 - We made the decision to try out TravisCI... It's clearly not working.
+I think it's time to exit this integration.",1,🇬🇧break travis integration 🇬🇧 we made the decision to try out travisci it s clearly not working i think it s time to exit this integration ,1
+465552,13387911081.0,IssuesEvent,2020-09-02 16:37:48,openforis/collect-earth-online,https://api.github.com/repos/openforis/collect-earth-online,closed,SRS Mismatch Between Uploaded Shapefile and Project Map,Bug Fix High Priority,"In project creation, uploading shapefiles appears to require that they be in a Lat/Long form (WGS84 works). This applies both to plot locations and to point locations.
+
+However, the map displayed in the data collection interface appears to use Web Mercator, so circular points that were uploaded as a shapefile tend to become ellipses when displayed. ",1.0,"SRS Mismatch Between Uploaded Shapefile and Project Map - In project creation, uploading shapefiles appears to require that they be in a Lat/Long form (WGS84 works). This applies both to plot locations and to point locations.
+
+However, the map displayed in the data collection interface appears to use Web Mercator, so circular points that were uploaded as a shapefile tend to become ellipses when displayed. ",0,srs mismatch between uploaded shapefile and project map in project creation uploading shapefiles appears to require that they be in a lat long form works this applies both to plot locations and to point locations however the map displayed in the data collection interface appears to use web mercator so circular points that were uploaded as a shapefile tend to become ellipses when displayed ,0
+491703,14169500512.0,IssuesEvent,2020-11-12 13:19:22,CESNET/perun-web-apps,https://api.github.com/repos/CESNET/perun-web-apps,closed,"Atributes list - too long row, because there are too many items in map/array",High priority small Issue,"
+
+Atributes list row can be really way too long if the map or array is too long.
+When there is 5 items its ok, but if it is bigger than that the table start to look very cluttered.
+
+I would like when there are more than 5 items in array/map in one row of the table to hide that other items of map/array and show button ""show more"" or ""..."". When user click on that button it will show him all the items in the array/map. The button ""show less"" would be nice too(when the items are expanded).
+
+Something like that already exist in the microsoft teams. You can inspire there. Dont forget about good UX.
+",1.0,"Atributes list - too long row, because there are too many items in map/array - 
+
+Atributes list row can be really way too long if the map or array is too long.
+When there is 5 items its ok, but if it is bigger than that the table start to look very cluttered.
+
+I would like when there are more than 5 items in array/map in one row of the table to hide that other items of map/array and show button ""show more"" or ""..."". When user click on that button it will show him all the items in the array/map. The button ""show less"" would be nice too(when the items are expanded).
+
+Something like that already exist in the microsoft teams. You can inspire there. Dont forget about good UX.
+",0,atributes list too long row because there are too many items in map array atributes list row can be really way too long if the map or array is too long when there is items its ok but if it is bigger than that the table start to look very cluttered i would like when there are more than items in array map in one row of the table to hide that other items of map array and show button show more or when user click on that button it will show him all the items in the array map the button show less would be nice too when the items are expanded something like that already exist in the microsoft teams you can inspire there dont forget about good ux ,0
+473,8443322489.0,IssuesEvent,2018-10-18 15:18:51,openstates/openstates,https://api.github.com/repos/openstates/openstates,closed,NH: duplicate legislators in API V2,people,"State: NH
+
+I am not at all confident that this is a scraper issue, or that it's limited to NH, or that it's not just my ignorance of how to handle this.
+
+This is similar to the duplicate legislators that occur in API V1, but AFAICT there's no all_ids field or other feature to mitigate it.
+
+Using API V2, one can ask for the votes on a bill:
+
+```gql
+{
+ bill(id: ""ocd-bill/8aab955f-6ad9-4f06-a8e2-64ed156940eb"") {
+ id
+ identifier
+ votes {
+ edges {
+ node {
+ id
+ votes {
+ option
+ voter {
+ id
+ }
+ voterName
+} } } } } }
+```
+
+The second individual vote reads:
+
+```json
+{
+ ""option"": ""yes"",
+ ""voter"": {
+ ""id"": ""ocd-person/51dfef5c-362f-46ff-bfa5-3e58998cb2dd""
+ },
+ ""voterName"": ""Chuck W Morse""
+}
+```
+The problem is that AFAICT, this voter ID isn't listed anywhere one might get all the legislators from.
+
+If, for example, one asks for all the legislators who ever served in the NH Senate:
+
+```gql
+{
+ people(first: 100, everMemberOf: ""ocd-organization/c95932a0-2ac1-4f14-a0c4-ade8568daac6"") {
+ edges {
+ node {
+ id
+ name
+ identifiers {
+ identifier
+ scheme
+} } } } }
+```
+
+... Chuck Morse is listed with a different id:
+
+```json
+{
+ ""id"": ""ocd-person/2d25d1fd-dfdb-44f0-86a9-874ec1dc3f81"",
+ ""name"": ""Chuck W Morse"",
+ ""identifiers"": []
+}
+
+```
+It is possible to do a single Person query for the `voter.id` `ocd-person/51dfef5c-362f-46ff-bfa5-3e58998cb2dd`, but then (AFAICT) the app must discover that it's a duplicate on its own, using something like pupa's pseudo-identifiers.
+",1.0,"NH: duplicate legislators in API V2 - State: NH
+
+I am not at all confident that this is a scraper issue, or that it's limited to NH, or that it's not just my ignorance of how to handle this.
+
+This is similar to the duplicate legislators that occur in API V1, but AFAICT there's no all_ids field or other feature to mitigate it.
+
+Using API V2, one can ask for the votes on a bill:
+
+```gql
+{
+ bill(id: ""ocd-bill/8aab955f-6ad9-4f06-a8e2-64ed156940eb"") {
+ id
+ identifier
+ votes {
+ edges {
+ node {
+ id
+ votes {
+ option
+ voter {
+ id
+ }
+ voterName
+} } } } } }
+```
+
+The second individual vote reads:
+
+```json
+{
+ ""option"": ""yes"",
+ ""voter"": {
+ ""id"": ""ocd-person/51dfef5c-362f-46ff-bfa5-3e58998cb2dd""
+ },
+ ""voterName"": ""Chuck W Morse""
+}
+```
+The problem is that AFAICT, this voter ID isn't listed anywhere one might get all the legislators from.
+
+If, for example, one asks for all the legislators who ever served in the NH Senate:
+
+```gql
+{
+ people(first: 100, everMemberOf: ""ocd-organization/c95932a0-2ac1-4f14-a0c4-ade8568daac6"") {
+ edges {
+ node {
+ id
+ name
+ identifiers {
+ identifier
+ scheme
+} } } } }
+```
+
+... Chuck Morse is listed with a different id:
+
+```json
+{
+ ""id"": ""ocd-person/2d25d1fd-dfdb-44f0-86a9-874ec1dc3f81"",
+ ""name"": ""Chuck W Morse"",
+ ""identifiers"": []
+}
+
+```
+It is possible to do a single Person query for the `voter.id` `ocd-person/51dfef5c-362f-46ff-bfa5-3e58998cb2dd`, but then (AFAICT) the app must discover that it's a duplicate on its own, using something like pupa's pseudo-identifiers.
+",1,nh duplicate legislators in api state nh i am not at all confident that this is a scraper issue or that it s limited to nh or that it s not just my ignorance of how to handle this this is similar to the duplicate legislators that occur in api but afaict there s no all ids field or other feature to mitigate it using api one can ask for the votes on a bill gql bill id ocd bill id identifier votes edges node id votes option voter id votername the second individual vote reads json option yes voter id ocd person votername chuck w morse the problem is that afaict this voter id isn t listed anywhere one might get all the legislators from if for example one asks for all the legislators who ever served in the nh senate gql people first evermemberof ocd organization edges node id name identifiers identifier scheme chuck morse is listed with a different id json id ocd person dfdb name chuck w morse identifiers it is possible to do a single person query for the voter id ocd person but then afaict the app must discover that it s a duplicate on its own using something like pupa s pseudo identifiers ,1
+521,9022501371.0,IssuesEvent,2019-02-07 01:55:08,3box/ops,https://api.github.com/repos/3box/ops,closed,Draft agreements for possible contractors ,Ops People,Work with HR team to get contracts ready for possible candidates so we can move quickly. ,1.0,Draft agreements for possible contractors - Work with HR team to get contracts ready for possible candidates so we can move quickly. ,1,draft agreements for possible contractors work with hr team to get contracts ready for possible candidates so we can move quickly ,1
+644,11571344858.0,IssuesEvent,2020-02-20 21:21:13,OfficeDev/office-ui-fabric-react,https://api.github.com/repos/OfficeDev/office-ui-fabric-react,closed,"[Accessibility][Keyboard navigation] After selecting last contact from 'Process selection contact' list, keyboard focus gets trapped somewhere and become invisible.",Area: Accessibility Component: PeoplePicker Priority 3: Fit and finish Resolution: Can't Repro,Breakout of #6259,1.0,"[Accessibility][Keyboard navigation] After selecting last contact from 'Process selection contact' list, keyboard focus gets trapped somewhere and become invisible. - Breakout of #6259",1, after selecting last contact from process selection contact list keyboard focus gets trapped somewhere and become invisible breakout of ,1
+210820,7194955354.0,IssuesEvent,2018-02-04 12:02:35,Bios-Marcel/ServerBrowser,https://api.github.com/repos/Bios-Marcel/ServerBrowser,closed,Ensure validity of language files ( completion / escaping)) ,Priority 1 enhancement,"In order to assure that everything is correct, the project is in need of tests which assure that every file is complete, at least in comparison to the English file, which is the default.
+
+In addition to completeness i could test or even correct the escaping of special characters. ",1.0,"Ensure validity of language files ( completion / escaping)) - In order to assure that everything is correct, the project is in need of tests which assure that every file is complete, at least in comparison to the English file, which is the default.
+
+In addition to completeness i could test or even correct the escaping of special characters. ",0,ensure validity of language files completion escaping in order to assure that everything is correct the project is in need of tests which assure that every file is complete at least in comparison to the english file which is the default in addition to completeness i could test or even correct the escaping of special characters ,0
+71,3200962834.0,IssuesEvent,2015-10-02 01:18:44,newtheatre/history-project,https://api.github.com/repos/newtheatre/history-project,closed,People's names on a person list should link to their person entry (if they have one),enhancement people technical,"These lists:
+
+
+
+the names should link to the people's person entries (alumni bio) if one exists.
+
+Potential issue is when people have the same name. Will have to work out an ID system for these cases.",1.0,"People's names on a person list should link to their person entry (if they have one) - These lists:
+
+
+
+the names should link to the people's person entries (alumni bio) if one exists.
+
+Potential issue is when people have the same name. Will have to work out an ID system for these cases.",1,people s names on a person list should link to their person entry if they have one these lists the names should link to the people s person entries alumni bio if one exists potential issue is when people have the same name will have to work out an id system for these cases ,1
+267145,28500994022.0,IssuesEvent,2023-04-18 17:19:11,Dima2021/mcwtest,https://api.github.com/repos/Dima2021/mcwtest,opened,mongodb-2.2.36.tgz: 3 vulnerabilities (highest severity is: 9.8),Mend: dependency security vulnerability," Vulnerable Library - mongodb-2.2.36.tgz
+
+
+
+All versions of bson before 1.1.4 are vulnerable to Deserialization of Untrusted Data. The package will ignore an unknown value for an object's _bsotype, leading to cases where an object is serialized as a document rather than the intended BSON type.
+
+
+
+In 'node-mongodb-native', versions prior to v3.1.13 are vulnerable against DOS as a result of a potential crash when a collection name is invalid and the DB doesn't exist.
+
+
+
+Incorrect parsing of certain JSON input may result in js-bson not correctly serializing BSON. This may cause unexpected application behaviour including data disclosure. This issue affects: MongoDB Inc. js-bson library version 1.1.3 and prior to.
+
+
+
+All versions of bson before 1.1.4 are vulnerable to Deserialization of Untrusted Data. The package will ignore an unknown value for an object's _bsotype, leading to cases where an object is serialized as a document rather than the intended BSON type.
+
+
+
+In 'node-mongodb-native', versions prior to v3.1.13 are vulnerable against DOS as a result of a potential crash when a collection name is invalid and the DB doesn't exist.
+
+
+
+Incorrect parsing of certain JSON input may result in js-bson not correctly serializing BSON. This may cause unexpected application behaviour including data disclosure. This issue affects: MongoDB Inc. js-bson library version 1.1.3 and prior to.
+
+
+
+
+
+
+
+
+:rescue_worker_helmet: Automatic Remediation is available for this issue
+
+
+***
+
+
:rescue_worker_helmet: Automatic Remediation is available for this issue.
",0,mongodb tgz vulnerabilities highest severity is vulnerable library mongodb tgz the official mongodb driver for node js library home page a href path to dependency file package json path to vulnerable library node modules mongodb package json found in head commit a href vulnerabilities cve severity cvss dependency type fixed in mongodb version remediation available high bson tgz transitive medium mongodb tgz direct medium bson tgz transitive details cve vulnerable library bson tgz a bson parser for node js and the browser library home page a href path to dependency file package json path to vulnerable library node modules bson package json dependency hierarchy mongodb tgz root library mongodb core tgz x bson tgz vulnerable library found in head commit a href found in base branch main vulnerability details all versions of bson before are vulnerable to deserialization of untrusted data the package will ignore an unknown value for an object s bsotype leading to cases where an object is serialized as a document rather than the intended bson type publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution bson direct dependency fix resolution mongodb rescue worker helmet automatic remediation is available for this issue ws vulnerable library mongodb tgz the official mongodb driver for node js library home page a href path to dependency file package json path to vulnerable library node modules mongodb package json dependency hierarchy x mongodb tgz vulnerable library found in head commit a href found in base branch main vulnerability details in node mongodb native versions prior to are vulnerable against dos as a result of a potential crash when a collection name is invalid and the db doesn t exist publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rescue worker helmet automatic remediation is available for this issue cve vulnerable library bson tgz a bson parser for node js and the browser library home page a href path to dependency file package json path to vulnerable library node modules bson package json dependency hierarchy mongodb tgz root library mongodb core tgz x bson tgz vulnerable library found in head commit a href found in base branch main vulnerability details incorrect parsing of certain json input may result in js bson not correctly serializing bson this may cause unexpected application behaviour including data disclosure this issue affects mongodb inc js bson library version and prior to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bson direct dependency fix resolution mongodb rescue worker helmet automatic remediation is available for this issue rescue worker helmet automatic remediation is available for this issue ,0
+295,5613589726.0,IssuesEvent,2017-04-03 09:42:37,PopulateTools/gobierto,https://api.github.com/repos/PopulateTools/gobierto,closed,"If no future events, show a notice and past events",gobierto-people,"If we don't have future events but past ones for an specific view:
+
+* show a notice:
+ No hay eventos futuros, échale un vistazo a los pasados
+ There are no future events in the calendar, take a look at the past ones
+
+* show past events",1.0,"If no future events, show a notice and past events - If we don't have future events but past ones for an specific view:
+
+* show a notice:
+ No hay eventos futuros, échale un vistazo a los pasados
+ There are no future events in the calendar, take a look at the past ones
+
+* show past events",1,if no future events show a notice and past events if we don t have future events but past ones for an specific view show a notice no hay eventos futuros échale un vistazo a los pasados there are no future events in the calendar take a look at the past ones show past events,1
+881,17362323931.0,IssuesEvent,2021-07-29 23:00:01,microsoft/fluentui,https://api.github.com/repos/microsoft/fluentui,closed,From peoplepicker which user is removed on click of remove (X) icon after selection of user,Component: PeoplePicker Needs: Author Feedback Status: No Recent Activity,"
+
+### Environment Information
+
+- **Package version(s)**: (fill this out)
+- **Browser and OS versions**: (fill this out if relevant)
+
+### Please provide a reproduction of the bug in a codepen:
+
+
+
+#### Actual behavior:
+From peoplepicker which user is removed on click of remove (X) icon after selection of user
+
+
+#### Expected behavior:
+
+On click of remove button need to get the user details
+### Priorities and help requested:
+
+Are you willing to submit a PR to fix? (Yes, No)
+
+Requested priority: (Blocking, High, Normal, Low)
+
+Products/sites affected: (if applicable)
+",1.0,"From peoplepicker which user is removed on click of remove (X) icon after selection of user -
+
+### Environment Information
+
+- **Package version(s)**: (fill this out)
+- **Browser and OS versions**: (fill this out if relevant)
+
+### Please provide a reproduction of the bug in a codepen:
+
+
+
+#### Actual behavior:
+From peoplepicker which user is removed on click of remove (X) icon after selection of user
+
+
+#### Expected behavior:
+
+On click of remove button need to get the user details
+### Priorities and help requested:
+
+Are you willing to submit a PR to fix? (Yes, No)
+
+Requested priority: (Blocking, High, Normal, Low)
+
+Products/sites affected: (if applicable)
+",1,from peoplepicker which user is removed on click of remove x icon after selection of user thanks for contacting us we re here to help before you report an issue check if it s been reported before search search by area or component note that if you do not provide enough information to reproduce the issue we may not be able to take action on your report environment information package version s fill this out browser and os versions fill this out if relevant please provide a reproduction of the bug in a codepen providing an isolated reproduction of the bug in a codepen makes it much easier for us to help you here are some ways to get started go to for a starter codepen you can also use the export to codepen feature for the various components in our documentation site see for a variety of examples alternatively you can also use to get permanent repro links if the repro occurs with an example a permanent link is preferable to use the website as the website can change actual behavior from peoplepicker which user is removed on click of remove x icon after selection of user expected behavior on click of remove button need to get the user details priorities and help requested are you willing to submit a pr to fix yes no requested priority blocking high normal low products sites affected if applicable ,1
+61683,25600401650.0,IssuesEvent,2022-12-01 19:39:44,hashicorp/terraform-provider-aws,https://api.github.com/repos/hashicorp/terraform-provider-aws,closed,[Bug]: EKS cluster ID is not always cluster name,bug service/eks,"### Terraform Core Version
+
+1.3.3
+
+### AWS Provider Version
+
+4.37.0
+
+### Affected Resource(s)
+
+- `aws_eks_cluster`
+
+### Expected Behavior
+
+The computed output for `id` should be pulled from the `describe-cluster` API when present, otherwise fallback to current behavior of using the cluster name
+
+Specifically, if you provision a local cluster on AWS Outposts, the cluster ID will NOT match the cluster name and instead look like a UUID/GUID
+
+### Actual Behavior
+
+All clusters created, the computed ID is set to the cluster name per https://github.com/hashicorp/terraform-provider-aws/blob/990f9aecb9a49fafff5f28f28e7737b5eab7d744/internal/service/eks/cluster.go#L345
+
+### Relevant Error/Panic Output Snippet
+
+_No response_
+
+### Terraform Configuration Files
+
+An example can be found here https://github.com/bryantbiggs/terraform-aws-eks/tree/refactor/v19/examples/outposts
+
+### Steps to Reproduce
+
+## AWS CLI
+Commercial regions with EKS cluster provisioned on AWS:
+```sh
+aws eks describe-cluster --name --region --query 'cluster.id'
+```
+
+Will return `null`
+
+Local cluster provisioned on AWS Outposts (both control plane and data plane):
+```sh
+aws eks describe-cluster --name --region --query 'cluster.id'
+```
+
+Should return a value that looks like a UUID/GUID value
+
+## Terraform
+
+Minimal reproduction using and updated version of [this test case](https://github.com/hashicorp/terraform-provider-aws/blob/990f9aecb9a49fafff5f28f28e7737b5eab7d744/internal/service/eks/cluster_test.go#L1045). Note, this will result in an `Unauthorized` error since the `--cluster-id` passed is the cluster name. Once corrected in the provider, this should apply successfully without modification
+
+```hcl
+data ""aws_iam_role"" ""test"" {
+ name = ""AmazonEKSLocalOutpostClusterRole""
+}
+data ""aws_outposts_outposts"" ""test"" {}
+data ""aws_subnets"" test {
+ filter {
+ name = ""outpost-arn""
+ values = [tolist(data.aws_outposts_outposts.test.arns)[0]]
+ }
+}
+resource ""aws_eks_cluster"" ""test"" {
+ name = %[1]q
+ role_arn = data.aws_iam_role.test.arn
+ outpost_config {
+ control_plane_instance_type = ""m5d.large""
+ outpost_arns = [tolist(data.aws_outposts_outposts.test.arns)[0]]
+ }
+ vpc_config {
+ endpoint_private_access = true
+ endpoint_public_access = false
+ subnet_ids = [tolist(data.aws_subnets.test.ids)[0]]
+ }
+}
+
+data ""aws_region"" ""current"" {}
+
+provider ""kubernetes"" {
+ host = aws_eks_cluster.test.endpoint
+ cluster_ca_certificate = base64decode(aws_eks_cluster.this[0].certificate_authority[0].data)
+
+ exec {
+ api_version = ""client.authentication.k8s.io/v1beta1""
+ command = ""aws""
+ args = [""eks"", ""get-token"", ""--cluster-id"", aws_eks_cluster.this.id, ""--region"", data.aws_region.current.name]
+ }
+}
+
+resource ""kubernetes_storage_class_v1"" ""this"" {
+ metadata {
+ name = ""ebs-sc""
+ annotations = {
+ ""storageclass.kubernetes.io/is-default-class"" = ""true""
+ }
+ }
+
+ storage_provisioner = ""ebs.csi.aws.com""
+ volume_binding_mode = ""WaitForFirstConsumer""
+ allow_volume_expansion = true
+
+ parameters = {
+ type = ""gp2""
+ encrypted = ""true""
+ }
+}
+```
+
+### Debug Output
+
+_No response_
+
+### Panic Output
+
+_No response_
+
+### Important Factoids
+
+_No response_
+
+### References
+
+_No response_
+
+### Would you like to implement a fix?
+
+Yes",1.0,"[Bug]: EKS cluster ID is not always cluster name - ### Terraform Core Version
+
+1.3.3
+
+### AWS Provider Version
+
+4.37.0
+
+### Affected Resource(s)
+
+- `aws_eks_cluster`
+
+### Expected Behavior
+
+The computed output for `id` should be pulled from the `describe-cluster` API when present, otherwise fallback to current behavior of using the cluster name
+
+Specifically, if you provision a local cluster on AWS Outposts, the cluster ID will NOT match the cluster name and instead look like a UUID/GUID
+
+### Actual Behavior
+
+All clusters created, the computed ID is set to the cluster name per https://github.com/hashicorp/terraform-provider-aws/blob/990f9aecb9a49fafff5f28f28e7737b5eab7d744/internal/service/eks/cluster.go#L345
+
+### Relevant Error/Panic Output Snippet
+
+_No response_
+
+### Terraform Configuration Files
+
+An example can be found here https://github.com/bryantbiggs/terraform-aws-eks/tree/refactor/v19/examples/outposts
+
+### Steps to Reproduce
+
+## AWS CLI
+Commercial regions with EKS cluster provisioned on AWS:
+```sh
+aws eks describe-cluster --name --region --query 'cluster.id'
+```
+
+Will return `null`
+
+Local cluster provisioned on AWS Outposts (both control plane and data plane):
+```sh
+aws eks describe-cluster --name --region --query 'cluster.id'
+```
+
+Should return a value that looks like a UUID/GUID value
+
+## Terraform
+
+Minimal reproduction using and updated version of [this test case](https://github.com/hashicorp/terraform-provider-aws/blob/990f9aecb9a49fafff5f28f28e7737b5eab7d744/internal/service/eks/cluster_test.go#L1045). Note, this will result in an `Unauthorized` error since the `--cluster-id` passed is the cluster name. Once corrected in the provider, this should apply successfully without modification
+
+```hcl
+data ""aws_iam_role"" ""test"" {
+ name = ""AmazonEKSLocalOutpostClusterRole""
+}
+data ""aws_outposts_outposts"" ""test"" {}
+data ""aws_subnets"" test {
+ filter {
+ name = ""outpost-arn""
+ values = [tolist(data.aws_outposts_outposts.test.arns)[0]]
+ }
+}
+resource ""aws_eks_cluster"" ""test"" {
+ name = %[1]q
+ role_arn = data.aws_iam_role.test.arn
+ outpost_config {
+ control_plane_instance_type = ""m5d.large""
+ outpost_arns = [tolist(data.aws_outposts_outposts.test.arns)[0]]
+ }
+ vpc_config {
+ endpoint_private_access = true
+ endpoint_public_access = false
+ subnet_ids = [tolist(data.aws_subnets.test.ids)[0]]
+ }
+}
+
+data ""aws_region"" ""current"" {}
+
+provider ""kubernetes"" {
+ host = aws_eks_cluster.test.endpoint
+ cluster_ca_certificate = base64decode(aws_eks_cluster.this[0].certificate_authority[0].data)
+
+ exec {
+ api_version = ""client.authentication.k8s.io/v1beta1""
+ command = ""aws""
+ args = [""eks"", ""get-token"", ""--cluster-id"", aws_eks_cluster.this.id, ""--region"", data.aws_region.current.name]
+ }
+}
+
+resource ""kubernetes_storage_class_v1"" ""this"" {
+ metadata {
+ name = ""ebs-sc""
+ annotations = {
+ ""storageclass.kubernetes.io/is-default-class"" = ""true""
+ }
+ }
+
+ storage_provisioner = ""ebs.csi.aws.com""
+ volume_binding_mode = ""WaitForFirstConsumer""
+ allow_volume_expansion = true
+
+ parameters = {
+ type = ""gp2""
+ encrypted = ""true""
+ }
+}
+```
+
+### Debug Output
+
+_No response_
+
+### Panic Output
+
+_No response_
+
+### Important Factoids
+
+_No response_
+
+### References
+
+_No response_
+
+### Would you like to implement a fix?
+
+Yes",0, eks cluster id is not always cluster name terraform core version aws provider version affected resource s aws eks cluster expected behavior the computed output for id should be pulled from the describe cluster api when present otherwise fallback to current behavior of using the cluster name specifically if you provision a local cluster on aws outposts the cluster id will not match the cluster name and instead look like a uuid guid actual behavior all clusters created the computed id is set to the cluster name per relevant error panic output snippet no response terraform configuration files an example can be found here steps to reproduce aws cli commercial regions with eks cluster provisioned on aws sh aws eks describe cluster name region query cluster id will return null local cluster provisioned on aws outposts both control plane and data plane sh aws eks describe cluster name region query cluster id should return a value that looks like a uuid guid value terraform minimal reproduction using and updated version of note this will result in an unauthorized error since the cluster id passed is the cluster name once corrected in the provider this should apply successfully without modification hcl data aws iam role test name amazonekslocaloutpostclusterrole data aws outposts outposts test data aws subnets test filter name outpost arn values resource aws eks cluster test name q role arn data aws iam role test arn outpost config control plane instance type large outpost arns vpc config endpoint private access true endpoint public access false subnet ids data aws region current provider kubernetes host aws eks cluster test endpoint cluster ca certificate aws eks cluster this certificate authority data exec api version client authentication io command aws args resource kubernetes storage class this metadata name ebs sc annotations storageclass kubernetes io is default class true storage provisioner ebs csi aws com volume binding mode waitforfirstconsumer allow volume expansion true parameters type encrypted true debug output no response panic output no response important factoids no response references no response would you like to implement a fix yes,0
+58164,8231492587.0,IssuesEvent,2018-09-07 16:03:09,RainbowMiner/RainbowMiner,https://api.github.com/repos/RainbowMiner/RainbowMiner,closed,Issue: Miner.config.txt out of order,documentation enhancement,"Would be nice to see the organiztaion better. Ocprofiles are sometimes above and also below MSIprofiles.
+
+Also, will any future updates over write this file since you are pushing a preset file? That means if I were to forget to backup my file and an update pushes, I would loose all my profile settings and be working on either your settings which would not match mine or completely erase my work with the customconfigreset.bat you created.
+
+",1.0,"Issue: Miner.config.txt out of order - Would be nice to see the organiztaion better. Ocprofiles are sometimes above and also below MSIprofiles.
+
+Also, will any future updates over write this file since you are pushing a preset file? That means if I were to forget to backup my file and an update pushes, I would loose all my profile settings and be working on either your settings which would not match mine or completely erase my work with the customconfigreset.bat you created.
+
+",0,issue miner config txt out of order would be nice to see the organiztaion better ocprofiles are sometimes above and also below msiprofiles also will any future updates over write this file since you are pushing a preset file that means if i were to forget to backup my file and an update pushes i would loose all my profile settings and be working on either your settings which would not match mine or completely erase my work with the customconfigreset bat you created ,0
+985,23225226795.0,IssuesEvent,2022-08-02 22:54:02,planningcenter/developers,https://api.github.com/repos/planningcenter/developers,closed,Send texts using third party SMS-provider?,People,"Hi!
+
+Not sure if this is the correct place to ask, but I was sent here by Planning center support :)
+
+Planning center is currently not supporting sending texts in Europe. Is it possible to set up an API integration between Planning center and our local SMS-provider?
+
+Thanks!
+
+BR Henrik",1.0,"Send texts using third party SMS-provider? - Hi!
+
+Not sure if this is the correct place to ask, but I was sent here by Planning center support :)
+
+Planning center is currently not supporting sending texts in Europe. Is it possible to set up an API integration between Planning center and our local SMS-provider?
+
+Thanks!
+
+BR Henrik",1,send texts using third party sms provider hi not sure if this is the correct place to ask but i was sent here by planning center support planning center is currently not supporting sending texts in europe is it possible to set up an api integration between planning center and our local sms provider thanks br henrik,1
+1069,25796393863.0,IssuesEvent,2022-12-10 15:57:44,newtheatre/history-project,https://api.github.com/repos/newtheatre/history-project,closed,Ronan Sullivan bio submission,report-tool people/bio,"End user submitted issue from page: [/collect/person/](https://history.newtheatre.org.uk/collect/person/)
+---
+# 'Submit an alumni bio' form submission
+
+Field | Data
+----- | ----
+Name | Ronan Sullivan
+Grad Year | 2021
+Course | English
+
+## Bio1 (Time at theatre)
+
+
+
+
+## Bio2 (Post-graduation)
+
+
+
+Field | Data
+------| ----
+Checked careers |
+Other careers |
+
+## Links
+
+
+
+## Shows
+
+
+
+## Committees
+
+
+
+## Awards
+
+
+
+## Contact Preferences
+
+Are we allowed to facilitate contact to this alumnus? **Yes**
+
+## Attempted File Generation
+
+```
+---
+title: Ronan Sullivan
+course:
+ - English
+graduated: 2021
+contact_allowed: true
+careers:
+
+
+links: *fill me out
+
+award: *fill me out
+
+---
+
+
+
+
+
+
+```
+**
+",1.0,"Ronan Sullivan bio submission - End user submitted issue from page: [/collect/person/](https://history.newtheatre.org.uk/collect/person/)
+---
+# 'Submit an alumni bio' form submission
+
+Field | Data
+----- | ----
+Name | Ronan Sullivan
+Grad Year | 2021
+Course | English
+
+## Bio1 (Time at theatre)
+
+
+
+
+## Bio2 (Post-graduation)
+
+
+
+Field | Data
+------| ----
+Checked careers |
+Other careers |
+
+## Links
+
+
+
+## Shows
+
+
+
+## Committees
+
+
+
+## Awards
+
+
+
+## Contact Preferences
+
+Are we allowed to facilitate contact to this alumnus? **Yes**
+
+## Attempted File Generation
+
+```
+---
+title: Ronan Sullivan
+course:
+ - English
+graduated: 2021
+contact_allowed: true
+careers:
+
+
+links: *fill me out
+
+award: *fill me out
+
+---
+
+
+
+
+
+
+```
+**
+",1,ronan sullivan bio submission end user submitted issue from page submit an alumni bio form submission field data name ronan sullivan grad year course english time at theatre post graduation field data checked careers other careers links shows committees awards contact preferences are we allowed to facilitate contact to this alumnus yes attempted file generation title ronan sullivan course english graduated contact allowed true careers links fill me out award fill me out ,1
+450294,13001601573.0,IssuesEvent,2020-07-24 00:14:56,kubernetes/website,https://api.github.com/repos/kubernetes/website,closed,Add kubernetes.io to the HSTS preload list,lifecycle/rotten priority/backlog,"Since this is a Google project, the website should be hosted using HTTPS (already the case! :-) and preloaded for HSTS.
+
+This mostly involves sending a header like `Strict-Transport-Security: max-age=63072000; includeSubDomains; preload` and getting redirects correct.
+See https://hstspreload.org/ for details.
+
+Note that the requirements for preloading require fixing #2371 in this case.",1.0,"Add kubernetes.io to the HSTS preload list - Since this is a Google project, the website should be hosted using HTTPS (already the case! :-) and preloaded for HSTS.
+
+This mostly involves sending a header like `Strict-Transport-Security: max-age=63072000; includeSubDomains; preload` and getting redirects correct.
+See https://hstspreload.org/ for details.
+
+Note that the requirements for preloading require fixing #2371 in this case.",0,add kubernetes io to the hsts preload list since this is a google project the website should be hosted using https already the case and preloaded for hsts this mostly involves sending a header like strict transport security max age includesubdomains preload and getting redirects correct see for details note that the requirements for preloading require fixing in this case ,0
+441,8102858916.0,IssuesEvent,2018-08-13 04:59:44,allenleein/brains,https://api.github.com/repos/allenleein/brains,reopened,The Liquid Advantage in Our Era,Closed New $ Power People ⚡️Output-observables,"# Macro Business Opportunities
+
+## 世界的經濟結構,正往「高流動性與不可複製的優勢」演化贏家。
+
+### Before:
+
+最大、穩定的商業成長機會,都在發展中國家的在地經濟( 大的固定市場,很難輕易的讓優勢流動)、和「低不可複製性、低流動性」的跨國商業。競爭的關鍵��是必須仰賴技術+便宜人力+政商關係和高成長的市場經濟動能,來轉取利潤。
+
+現在這樣的商業模式,已經變成國家拓展軍事力量的手段:一帶一路。但宏觀的不確定因素,讓這樣的商業模式在現代變的比以前更難獲利,也更不穩定。因為資本的流動更快,更投機、更集中。
+
+### Now:
+
+由於過去資本經濟發展架構已經趨向穩定,全球化後的現代商業最大的機會,在於創造「高/低技術,高流動性、高不可複製性」的商業模式。
+
+像是 Facebook, Google, Airbnb, Uber....直播,eSports..。也可以說是「高不可複製性、高流動性」的商業機會,其優勢可以在不同的市場流動。可以因為龐大資本的擴充,而不等比例地把優勢擴大的商業模式。
+
+# Macro Problems
+
+#### Politics
+
+We are quickly moving away from a slowly decentralized global politics to a more centralized ecosystems. (Syria, ISIS, China, Trump…)
+
+China’s ruling Communist Party has proposed scrapping term limits for the country’s president earlier today, appearing to lay the groundwork for party leader Xi Jinping to rule as president beyond 2023. The planned constitutional change allows Xi to hold power indefinitely. As you can see, Xi has already grasped more power more quickly than his two recent predecessors.
+
+The world must now come to terms with China’s most powerful leader in decades, perhaps for decades to come. And China is accumulating ever greater international influence, increasingly pitting its own political system against western democracies, and pursuing the so-called “Chinese Dream” to build a more prosperous and glorious future.
+
+#### Capitalism
+
+Both capitalism and communism have decentralized voluntaryism, but both ideals corrupt in practice to centralized rent-seeking and use of lethal force. The world was never so much about ideal distribution of resources, as it was about the centralized use of force vs a decentralized voluntary trade and resilience to such force.
+
+Because of this condition, It’s possible that human can create a totally different power (Economy/Politics/Gov) structure based on future of blockchain technology. Its all about people who want to have more power to control Privacy / Tax / Control / Macro Economy: rates, currency, politics / Regulation and identity.
+
+I believe that the future of Internet and blockchain tech is the best chance of decentralized power we’ve had in past 1000 years since Song dynasty.
+
+## Gap
+
+The growing inequality in our society is producing such a disconnection between those who rule and those who are ruled that they might as well live on separate planets.
+
+Not just that, but because wealth equates to power and since there just isn’t anywhere when the very wealthy get to speak to most of the rest of us, the notion that ‘democracy’ should be pursued or encouraged or supported is something the elites are increasingly less likely to feel even makes sense.
+
+## 公共論辯的式微
+
+相對於 19 世紀
+
+- 現代人、資本的流動性大幅提高,對地區的忠誠與貢獻度大幅減少
+
+- 新財富的增速和流動性遠超越古代,但階級流動停滯
+
+- 城市的崛起,貧富差距擴大
+
+- 新貴族沒有長遠的地域性傳承,倫理觀的墮落
+
+## Edu
+
+The things that’s not taught In our current edu system is iteration.
+
+Iteration is everything outside of school. You do something and then you launch it and then you are on to making a better version of that thing pretty quickly. In school we do something, hand it in and then it’s done. We never get to **revisit** things.",1.0,"The Liquid Advantage in Our Era - # Macro Business Opportunities
+
+## 世界的經濟結構,正往「高流動性與不可複製的優勢」演化贏家。
+
+### Before:
+
+最大、穩定的商業成長機會,都在發展中國家的在地經濟( 大的固定市場,很難輕易的讓優勢流動)、和「低不可複製性、低流動性」的跨國商業。競爭的關鍵,是必須仰賴技術+便宜人力+政商關係和高成長的市場經濟動能,來轉取利潤。
+
+現在這樣的商業模式,已經變成國家拓展軍事力量的手段:一帶一路。但宏觀的不確定因素,讓這樣的商業模式在現代變的比以前更難獲利,也更不穩定。因為資本的流動更快,更投機、更集中。
+
+### Now:
+
+由於過去資本經濟發展架構已經趨向穩定,全球化後的現代商業最大的機會,在於創造「高/低技術,高流動性、高不可複製性」的商業模式。
+
+像是 Facebook, Google, Airbnb, Uber....直播,eSports..。也可以說是「高不可複製性、高流動性」的商業機會,其優勢可以在不同的市場流動。可以因為龐大資本的擴充,而不等比例地把優勢擴大的商業模式。
+
+# Macro Problems
+
+#### Politics
+
+We are quickly moving away from a slowly decentralized global politics to a more centralized ecosystems. (Syria, ISIS, China, Trump…)
+
+China’s ruling Communist Party has proposed scrapping term limits for the country’s president earlier today, appearing to lay the groundwork for party leader Xi Jinping to rule as president beyond 2023. The planned constitutional change allows Xi to hold power indefinitely. As you can see, Xi has already grasped more power more quickly than his two recent predecessors.
+
+The world must now come to terms with China’s most powerful leader in decades, perhaps for decades to come. And China is accumulating ever greater international influence, increasingly pitting its own political system against western democracies, and pursuing the so-called “Chinese Dream” to build a more prosperous and glorious future.
+
+#### Capitalism
+
+Both capitalism and communism have decentralized voluntaryism, but both ideals corrupt in practice to centralized rent-seeking and use of lethal force. The world was never so much about ideal distribution of resources, as it was about the centralized use of force vs a decentralized voluntary trade and resilience to such force.
+
+Because of this condition, It’s possible that human can create a totally different power (Economy/Politics/Gov) structure based on future of blockchain technology. Its all about people who want to have more power to control Privacy / Tax / Control / Macro Economy: rates, currency, politics / Regulation and identity.
+
+I believe that the future of Internet and blockchain tech is the best chance of decentralized power we’ve had in past 1000 years since Song dynasty.
+
+## Gap
+
+The growing inequality in our society is producing such a disconnection between those who rule and those who are ruled that they might as well live on separate planets.
+
+Not just that, but because wealth equates to power and since there just isn’t anywhere when the very wealthy get to speak to most of the rest of us, the notion that ‘democracy’ should be pursued or encouraged or supported is something the elites are increasingly less likely to feel even makes sense.
+
+## 公共論辯的式微
+
+相對於 19 世紀
+
+- 現代人、資本的流動性大幅提高,對地區的忠誠與貢獻度大幅減少
+
+- 新財富的增速和流動性遠超越古代,但階級流動停滯
+
+- 城市的崛起,貧富差距擴大
+
+- 新貴族沒有長遠的地域性傳承,倫理觀的墮落
+
+## Edu
+
+The things that’s not taught In our current edu system is iteration.
+
+Iteration is everything outside of school. You do something and then you launch it and then you are on to making a better version of that thing pretty quickly. In school we do something, hand it in and then it’s done. We never get to **revisit** things.",1,the liquid advantage in our era macro business opportunities 世界的經濟結構,正往「高流動性與不可複製的優勢」演化贏家。 before: 最大、穩定的商業成長機會,都在發展中國家的在地經濟 大的固定市場,很難輕易的讓優勢流動 、和「低不可複製性、低流動性」的跨國商業。競爭的關鍵,是必須仰賴技術+便宜人力+政商關係和高成長的市場經濟動能,來轉取利潤。 現在這樣的商業模式,已經變成國家拓展軍事力量的手段:一帶一路。但宏觀的不確定因素,讓這樣的商業模式在現代變的比以前更難獲利,也更不穩定。因為資本的流動更快,更投機、更集中。 now: 由於過去資本經濟發展架構已經趨向穩定,全球化後的現代商業最大的機會,在於創造「高 低技術,高流動性、高不可複製性」的商業模式。 像是 facebook google airbnb uber 直播 esports 。也可以說是「高不可複製性、高流動性」的商業機會,其優勢可以在不同的市場流動。可以因為龐大資本的擴充,而不等比例地把優勢擴大的商業模式。 macro problems politics we are quickly moving away from a slowly decentralized global politics to a more centralized ecosystems syria isis china trump… china’s ruling communist party has proposed scrapping term limits for the country’s president earlier today appearing to lay the groundwork for party leader xi jinping to rule as president beyond the planned constitutional change allows xi to hold power indefinitely as you can see xi has already grasped more power more quickly than his two recent predecessors the world must now come to terms with china’s most powerful leader in decades perhaps for decades to come and china is accumulating ever greater international influence increasingly pitting its own political system against western democracies and pursuing the so called “chinese dream” to build a more prosperous and glorious future capitalism both capitalism and communism have decentralized voluntaryism but both ideals corrupt in practice to centralized rent seeking and use of lethal force the world was never so much about ideal distribution of resources as it was about the centralized use of force vs a decentralized voluntary trade and resilience to such force because of this condition it’s possible that human can create a totally different power economy politics gov structure based on future of blockchain technology its all about people who want to have more power to control privacy tax control macro economy rates currency politics regulation and identity i believe that the future of internet and blockchain tech is the best chance of decentralized power we’ve had in past years since song dynasty gap the growing inequality in our society is producing such a disconnection between those who rule and those who are ruled that they might as well live on separate planets not just that but because wealth equates to power and since there just isn’t anywhere when the very wealthy get to speak to most of the rest of us the notion that ‘democracy’ should be pursued or encouraged or supported is something the elites are increasingly less likely to feel even makes sense 公共論辯的式微 相對於 世紀 現代人、資本的流動性大幅提高,對地區的忠誠與貢獻度大幅減少 新財富的增速和流動性遠超越古代,但階級流動停滯 城市的崛起,貧富差距擴大 新貴族沒有長遠的地域性傳承,倫理觀的墮落 edu the things that’s not taught in our current edu system is iteration iteration is everything outside of school you do something and then you launch it and then you are on to making a better version of that thing pretty quickly in school we do something hand it in and then it’s done we never get to revisit things ,1
+450887,13021146584.0,IssuesEvent,2020-07-27 05:32:36,webcompat/web-bugs,https://api.github.com/repos/webcompat/web-bugs,closed,www.hepsiburada.com - site is not usable,browser-firefox engine-gecko ml-needsdiagnosis-false ml-probability-high priority-important,"
+
+
+
+
+**URL**: https://www.hepsiburada.com/
+
+**Browser / Version**: Firefox 78.0
+**Operating System**: Windows 10
+**Tested Another Browser**: No
+
+**Problem type**: Site is not usable
+**Description**: Page not loading correctly
+**Steps to Reproduce**:
+.......................................
+
+
+Browser Configuration
+
+
None
+
+
+
+_From [webcompat.com](https://webcompat.com/) with ❤️_",1.0,"www.hepsiburada.com - site is not usable -
+
+
+
+
+**URL**: https://www.hepsiburada.com/
+
+**Browser / Version**: Firefox 78.0
+**Operating System**: Windows 10
+**Tested Another Browser**: No
+
+**Problem type**: Site is not usable
+**Description**: Page not loading correctly
+**Steps to Reproduce**:
+.......................................
+
+
+Browser Configuration
+
+
None
+
+
+
+_From [webcompat.com](https://webcompat.com/) with ❤️_",0, site is not usable url browser version firefox operating system windows tested another browser no problem type site is not usable description page not loading correctly steps to reproduce browser configuration none from with ❤️ ,0
+31,2652099371.0,IssuesEvent,2015-03-16 15:36:26,fskreuz/Tree-Trail,https://api.github.com/repos/fskreuz/Tree-Trail,closed,Create Contacts Page,QA people! Test these!,"As a user, I should be able to view a list of contact person with connections about tree planting and other involved parties",1.0,"Create Contacts Page - As a user, I should be able to view a list of contact person with connections about tree planting and other involved parties",1,create contacts page as a user i should be able to view a list of contact person with connections about tree planting and other involved parties,1
+69438,17669242720.0,IssuesEvent,2021-08-23 01:57:03,apache/shardingsphere,https://api.github.com/repos/apache/shardingsphere,closed,Build Failure,type: question type: build,"## Bug Report
+
+**For English only**, other languages will not accept.
+
+Before report a bug, make sure you have:
+
+- Searched open and closed [GitHub issues](https://github.com/apache/shardingsphere/issues).
+- Read documentation: [ShardingSphere Doc](https://shardingsphere.apache.org/document/current/en/overview).
+
+Please pay attention on issues you submitted, because we maybe need more details.
+If no response anymore and we cannot reproduce it on current information, we will **close it**.
+
+Please answer these questions before submitting your issue. Thanks!
+
+### Which version of ShardingSphere did you use?
+
+### Which project did you use? ShardingSphere-JDBC or ShardingSphere-Proxy?
+
+### Expected behavior
+Build should be successful
+### Actual behavior
+On running `mvn clean install`
+
+
+On running maven with -X switch
+
+
+
+On running `mvn -Dmaven.javadoc.skip=true -Djacoco.skip=true -DskipITs -DskipTests clean install`
+
+
+
+### Reason analyze (If you can)
+
+### Steps to reproduce the behavior, such as: SQL to execute, sharding rule configuration, when exception occur etc.
+Run `mvn clean install`
+Run maven with -X switch
+Run `mvn -Dmaven.javadoc.skip=true -Djacoco.skip=true -DskipITs -DskipTests clean install`
+### Example codes for reproduce this issue (such as a github link).
+",1.0,"Build Failure - ## Bug Report
+
+**For English only**, other languages will not accept.
+
+Before report a bug, make sure you have:
+
+- Searched open and closed [GitHub issues](https://github.com/apache/shardingsphere/issues).
+- Read documentation: [ShardingSphere Doc](https://shardingsphere.apache.org/document/current/en/overview).
+
+Please pay attention on issues you submitted, because we maybe need more details.
+If no response anymore and we cannot reproduce it on current information, we will **close it**.
+
+Please answer these questions before submitting your issue. Thanks!
+
+### Which version of ShardingSphere did you use?
+
+### Which project did you use? ShardingSphere-JDBC or ShardingSphere-Proxy?
+
+### Expected behavior
+Build should be successful
+### Actual behavior
+On running `mvn clean install`
+
+
+On running maven with -X switch
+
+
+
+On running `mvn -Dmaven.javadoc.skip=true -Djacoco.skip=true -DskipITs -DskipTests clean install`
+
+
+
+### Reason analyze (If you can)
+
+### Steps to reproduce the behavior, such as: SQL to execute, sharding rule configuration, when exception occur etc.
+Run `mvn clean install`
+Run maven with -X switch
+Run `mvn -Dmaven.javadoc.skip=true -Djacoco.skip=true -DskipITs -DskipTests clean install`
+### Example codes for reproduce this issue (such as a github link).
+",0,build failure bug report for english only other languages will not accept before report a bug make sure you have searched open and closed read documentation please pay attention on issues you submitted because we maybe need more details if no response anymore and we cannot reproduce it on current information we will close it please answer these questions before submitting your issue thanks which version of shardingsphere did you use which project did you use shardingsphere jdbc or shardingsphere proxy expected behavior build should be successful actual behavior on running mvn clean install on running maven with x switch on running mvn dmaven javadoc skip true djacoco skip true dskipits dskiptests clean install reason analyze if you can steps to reproduce the behavior such as sql to execute sharding rule configuration when exception occur etc run mvn clean install run maven with x switch run mvn dmaven javadoc skip true djacoco skip true dskipits dskiptests clean install example codes for reproduce this issue such as a github link ,0
+459,8353407267.0,IssuesEvent,2018-10-02 09:55:59,allenleein/brains,https://api.github.com/repos/allenleein/brains,closed,Convicted of drug trafficking Bronwyn Atherton arrives home in Sydney after eight years in a Peruvian jail,Functions People,"Convicted of drug trafficking, Bronwyn Atherton arrives home in Sydney after eight years in a Peruvian jail
+The 32-year-old from Ulladulla on the NSW South Coast who greets her family and friends with joy and laughter looks like she's just returned from a gap year in Europe. In reality, she has landed in her homeland after more than eight years in a Peruvian prison.
+
+via Pocket https://ift.tt/2x2EzEc
+
+September 14, 2018 at 05:11PM",1.0,"Convicted of drug trafficking Bronwyn Atherton arrives home in Sydney after eight years in a Peruvian jail - Convicted of drug trafficking, Bronwyn Atherton arrives home in Sydney after eight years in a Peruvian jail
+The 32-year-old from Ulladulla on the NSW South Coast who greets her family and friends with joy and laughter looks like she's just returned from a gap year in Europe. In reality, she has landed in her homeland after more than eight years in a Peruvian prison.
+
+via Pocket https://ift.tt/2x2EzEc
+
+September 14, 2018 at 05:11PM",1,convicted of drug trafficking bronwyn atherton arrives home in sydney after eight years in a peruvian jail convicted of drug trafficking bronwyn atherton arrives home in sydney after eight years in a peruvian jail the year old from ulladulla on the nsw south coast who greets her family and friends with joy and laughter looks like she s just returned from a gap year in europe in reality she has landed in her homeland after more than eight years in a peruvian prison via pocket september at ,1
+916,19565062825.0,IssuesEvent,2022-01-03 22:23:33,makeabilitylab/makeabilitylabwebsite,https://api.github.com/repos/makeabilitylab/makeabilitylabwebsite,closed,People page throwing Server Error 500,bug FixedNeedsToBeTestedOnTestServer People Page,"The https://makeabilitylab.cs.washington.edu/people/ page is just showing `Server Error (500)`. I don't see anything particularly interesting in the browser console either.
+
+I'm using latest Google Chrome on Linux.
+
+I tried to add a new member to the lab with a start date in the future, then I checked the page and saw this error. But I didn't check the page _before_ adding that member, so I have no idea if the data I added was related.",1.0,"People page throwing Server Error 500 - The https://makeabilitylab.cs.washington.edu/people/ page is just showing `Server Error (500)`. I don't see anything particularly interesting in the browser console either.
+
+I'm using latest Google Chrome on Linux.
+
+I tried to add a new member to the lab with a start date in the future, then I checked the page and saw this error. But I didn't check the page _before_ adding that member, so I have no idea if the data I added was related.",1,people page throwing server error the page is just showing server error i don t see anything particularly interesting in the browser console either i m using latest google chrome on linux i tried to add a new member to the lab with a start date in the future then i checked the page and saw this error but i didn t check the page before adding that member so i have no idea if the data i added was related ,1
+226405,24947090468.0,IssuesEvent,2022-11-01 01:50:33,Trinadh465/device_renesas_kernel_AOSP10_r33_CVE-2021-33034,https://api.github.com/repos/Trinadh465/device_renesas_kernel_AOSP10_r33_CVE-2021-33034,closed,CVE-2021-45480 (Medium) detected in linuxlinux-4.19.239 - autoclosed,security vulnerability,"## CVE-2021-45480 - Medium Severity Vulnerability
+ Vulnerable Library - linuxlinux-4.19.239
+
+
+An issue was discovered in the Linux kernel before 5.15.11. There is a memory leak in the __rds_conn_create() function in net/rds/connection.c in a certain combination of circumstances.
+
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",True,"CVE-2021-45480 (Medium) detected in linuxlinux-4.19.239 - autoclosed - ## CVE-2021-45480 - Medium Severity Vulnerability
+ Vulnerable Library - linuxlinux-4.19.239
+
+
+An issue was discovered in the Linux kernel before 5.15.11. There is a memory leak in the __rds_conn_create() function in net/rds/connection.c in a certain combination of circumstances.
+
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",0,cve medium detected in linuxlinux autoclosed cve medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details an issue was discovered in the linux kernel before there is a memory leak in the rds conn create function in net rds connection c in a certain combination of circumstances publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend ,0
+363902,10756634284.0,IssuesEvent,2019-10-31 11:38:32,OpenNebula/one,https://api.github.com/repos/OpenNebula/one,opened,Deprecate driver configuration files in XML,Category: Core & System Category: vCenter Priority: High Status: Accepted Type: Feature,"**Description**
+This issue is to migrate the XML configuration file in a common one to ease upgrades
+
+
+
+
+
+
+
+
+## Progress Status
+- [ ] Branch created
+- [ ] Code committed to development branch
+- [ ] Testing - QA
+- [ ] Documentation
+- [ ] Release notes - resolved issues, compatibility, known issues
+- [ ] Code committed to upstream release/hotfix branches
+- [ ] Documentation committed to upstream release/hotfix branches
+",1.0,"Deprecate driver configuration files in XML - **Description**
+This issue is to migrate the XML configuration file in a common one to ease upgrades
+
+
+
+
+
+
+
+
+## Progress Status
+- [ ] Branch created
+- [ ] Code committed to development branch
+- [ ] Testing - QA
+- [ ] Documentation
+- [ ] Release notes - resolved issues, compatibility, known issues
+- [ ] Code committed to upstream release/hotfix branches
+- [ ] Documentation committed to upstream release/hotfix branches
+",0,deprecate driver configuration files in xml description this issue is to migrate the xml configuration file in a common one to ease upgrades progress status branch created code committed to development branch testing qa documentation release notes resolved issues compatibility known issues code committed to upstream release hotfix branches documentation committed to upstream release hotfix branches ,0
+260415,27783147943.0,IssuesEvent,2023-03-16 23:17:03,dotnet/aspnetcore,https://api.github.com/repos/dotnet/aspnetcore,closed,.net 7 JwtBearerDefaults.AuthenticationScheme not work with Entity framework core Identity,area-security area-identity Needs: Attention :wave:,"### Is there an existing issue for this?
+
+- [X] I have searched the existing issues
+
+### Describe the bug
+
+# Asp.NetCoreApi_Identity_bug
+bug in asp.net core WebApi with Entity framework core
+when test with ``` JwtBearerDefaults.AuthenticationScheme ``` get this
+
+
+when test with
+```
+options =>
+{
+ options.DefaultAuthenticateScheme = JwtBearerDefaults.AuthenticationScheme;
+ options.DefaultChallengeScheme = JwtBearerDefaults.AuthenticationScheme;
+}
+```
+work normal
+
+
+and if test with ``` JwtBearerDefaults.AuthenticationScheme ``` but remove this
+```
+builder.Services.AddIdentity()
+ .AddEntityFrameworkStores()
+ .AddDefaultTokenProviders();
+```
+work normal
+
+
+i create repo with code i test https://github.com/amrmostafa800/Asp.NetCoreApi_Identity_bug
+
+
+### Expected Behavior
+
+_No response_
+
+### Steps To Reproduce
+
+_No response_
+
+### Exceptions (if any)
+
+_No response_
+
+### .NET Version
+
+7.0.200
+
+### Anything else?
+
+i found issue similar to it but it closed https://github.com/dotnet/aspnetcore/issues/46774",True,".net 7 JwtBearerDefaults.AuthenticationScheme not work with Entity framework core Identity - ### Is there an existing issue for this?
+
+- [X] I have searched the existing issues
+
+### Describe the bug
+
+# Asp.NetCoreApi_Identity_bug
+bug in asp.net core WebApi with Entity framework core
+when test with ``` JwtBearerDefaults.AuthenticationScheme ``` get this
+
+
+when test with
+```
+options =>
+{
+ options.DefaultAuthenticateScheme = JwtBearerDefaults.AuthenticationScheme;
+ options.DefaultChallengeScheme = JwtBearerDefaults.AuthenticationScheme;
+}
+```
+work normal
+
+
+and if test with ``` JwtBearerDefaults.AuthenticationScheme ``` but remove this
+```
+builder.Services.AddIdentity()
+ .AddEntityFrameworkStores()
+ .AddDefaultTokenProviders();
+```
+work normal
+
+
+i create repo with code i test https://github.com/amrmostafa800/Asp.NetCoreApi_Identity_bug
+
+
+### Expected Behavior
+
+_No response_
+
+### Steps To Reproduce
+
+_No response_
+
+### Exceptions (if any)
+
+_No response_
+
+### .NET Version
+
+7.0.200
+
+### Anything else?
+
+i found issue similar to it but it closed https://github.com/dotnet/aspnetcore/issues/46774",0, net jwtbearerdefaults authenticationscheme not work with entity framework core identity is there an existing issue for this i have searched the existing issues describe the bug asp netcoreapi identity bug bug in asp net core webapi with entity framework core when test with jwtbearerdefaults authenticationscheme get this when test with options options defaultauthenticatescheme jwtbearerdefaults authenticationscheme options defaultchallengescheme jwtbearerdefaults authenticationscheme work normal and if test with jwtbearerdefaults authenticationscheme but remove this builder services addidentity addentityframeworkstores adddefaulttokenproviders work normal i create repo with code i test expected behavior no response steps to reproduce no response exceptions if any no response net version anything else i found issue similar to it but it closed ,0
+55934,11489339330.0,IssuesEvent,2020-02-11 15:21:40,jOOQ/jOOQ,https://api.github.com/repos/jOOQ/jOOQ,closed,Generated constraint names should not be table prefixed in meta data in DB2 and PostgreSQL,C: Code Generation E: All Editions P: Medium R: Fixed T: Defect T: Incompatible change,"A long time ago, we fixed a problem in jOOQ where some dialects allow per-table namespacing of constraints, whereas other dialects implement per-schema namespacing of contraints: https://github.com/jOOQ/jOOQ/issues/986
+
+That fix was incorrect for DB2, where the generated constraint name has a prefix in the meta model, not only in generated code. A later fix disambiguated things in generator strategies, because it is only a problem for generated code, not for the meta model.
+
+This fix will not be backported to 3.12 to prevent any regressions in client code that may have come to depend on the wrong implementation.",1.0,"Generated constraint names should not be table prefixed in meta data in DB2 and PostgreSQL - A long time ago, we fixed a problem in jOOQ where some dialects allow per-table namespacing of constraints, whereas other dialects implement per-schema namespacing of contraints: https://github.com/jOOQ/jOOQ/issues/986
+
+That fix was incorrect for DB2, where the generated constraint name has a prefix in the meta model, not only in generated code. A later fix disambiguated things in generator strategies, because it is only a problem for generated code, not for the meta model.
+
+This fix will not be backported to 3.12 to prevent any regressions in client code that may have come to depend on the wrong implementation.",0,generated constraint names should not be table prefixed in meta data in and postgresql a long time ago we fixed a problem in jooq where some dialects allow per table namespacing of constraints whereas other dialects implement per schema namespacing of contraints that fix was incorrect for where the generated constraint name has a prefix in the meta model not only in generated code a later fix disambiguated things in generator strategies because it is only a problem for generated code not for the meta model this fix will not be backported to to prevent any regressions in client code that may have come to depend on the wrong implementation ,0
+44212,17926860890.0,IssuesEvent,2021-09-10 01:32:44,MicrosoftDocs/azure-docs,https://api.github.com/repos/MicrosoftDocs/azure-docs,closed,Regional Vnet on a flex stamp,app-service/svc triaged assigned-to-author doc-enhancement Pri2,"Feedback regarding this part
+""The feature is available only from newer Azure App Service scale units that support PremiumV2 App Service plans. Note that this does not mean your app must run on a PremiumV2 pricing tier, only that it must run on an App Service Plan where the PremiumV2 option is available (which implies that it is a newer scale unit where this VNet integration feature is then also available).""
+
+its kind of misleading, if your app is running on a flex stamp, its required to run on a premiumV2 to use the regional vnet, i think its worth clarifying this.
+
+---
+#### Document Details
+
+⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
+
+* ID: a7a98803-1438-b1b5-f543-7dd88bc4294e
+* Version Independent ID: 37ff1d0f-ed8e-5e4d-1f4c-1b9f6cffb938
+* Content: [Integrate app with Azure Virtual Network - Azure App Service](https://docs.microsoft.com/en-us/azure/app-service/web-sites-integrate-with-vnet?source=docs#regional-vnet-integration)
+* Content Source: [articles/app-service/web-sites-integrate-with-vnet.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/app-service/web-sites-integrate-with-vnet.md)
+* Service: **app-service**
+* GitHub Login: @ccompy
+* Microsoft Alias: **ccompy**",1.0,"Regional Vnet on a flex stamp - Feedback regarding this part
+""The feature is available only from newer Azure App Service scale units that support PremiumV2 App Service plans. Note that this does not mean your app must run on a PremiumV2 pricing tier, only that it must run on an App Service Plan where the PremiumV2 option is available (which implies that it is a newer scale unit where this VNet integration feature is then also available).""
+
+its kind of misleading, if your app is running on a flex stamp, its required to run on a premiumV2 to use the regional vnet, i think its worth clarifying this.
+
+---
+#### Document Details
+
+⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
+
+* ID: a7a98803-1438-b1b5-f543-7dd88bc4294e
+* Version Independent ID: 37ff1d0f-ed8e-5e4d-1f4c-1b9f6cffb938
+* Content: [Integrate app with Azure Virtual Network - Azure App Service](https://docs.microsoft.com/en-us/azure/app-service/web-sites-integrate-with-vnet?source=docs#regional-vnet-integration)
+* Content Source: [articles/app-service/web-sites-integrate-with-vnet.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/app-service/web-sites-integrate-with-vnet.md)
+* Service: **app-service**
+* GitHub Login: @ccompy
+* Microsoft Alias: **ccompy**",0,regional vnet on a flex stamp feedback regarding this part the feature is available only from newer azure app service scale units that support app service plans note that this does not mean your app must run on a pricing tier only that it must run on an app service plan where the option is available which implies that it is a newer scale unit where this vnet integration feature is then also available its kind of misleading if your app is running on a flex stamp its required to run on a to use the regional vnet i think its worth clarifying this document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service app service github login ccompy microsoft alias ccompy ,0
+393660,11623046635.0,IssuesEvent,2020-02-27 08:06:59,gambitph/Stackable,https://api.github.com/repos/gambitph/Stackable,closed,Some Blocks doesn't support grid column in IE11,[browser] ie11 [version] V1 bug high priority,"
+
+**Describe the bug**
+Some blocks that have a number of columns doesn't show in columns at IE11. Instead, they show up vertically.
+
+**To Reproduce**
+Steps to reproduce the behavior:
+1. Add 2 or more columns of these blocks: Pricing Box, Number Box, Image Box, Testimonial or Team Member
+2. View in IE11, they'll show up vertically.
+
+**Expected behavior**
+They should show up in columns accordingly.
+
+**Screenshots**
+
+
+**Desktop (please complete the following information):**
+ - OS: Windows
+ - Browser: IE11",1.0,"Some Blocks doesn't support grid column in IE11 -
+
+**Describe the bug**
+Some blocks that have a number of columns doesn't show in columns at IE11. Instead, they show up vertically.
+
+**To Reproduce**
+Steps to reproduce the behavior:
+1. Add 2 or more columns of these blocks: Pricing Box, Number Box, Image Box, Testimonial or Team Member
+2. View in IE11, they'll show up vertically.
+
+**Expected behavior**
+They should show up in columns accordingly.
+
+**Screenshots**
+
+
+**Desktop (please complete the following information):**
+ - OS: Windows
+ - Browser: IE11",0,some blocks doesn t support grid column in before posting make sure that you are running the latest version of stackable and you have searched whether your issue has already been reported describe the bug some blocks that have a number of columns doesn t show in columns at instead they show up vertically to reproduce steps to reproduce the behavior add or more columns of these blocks pricing box number box image box testimonial or team member view in they ll show up vertically expected behavior they should show up in columns accordingly screenshots desktop please complete the following information os windows browser ,0
+367,6822925179.0,IssuesEvent,2017-11-07 21:48:26,PopulateTools/gobierto,https://api.github.com/repos/PopulateTools/gobierto,closed,Fix calendar integration picker on Firefox,bug gobierto-people,"There's an issue with Firefox that provokes the calendar integration picker to only display the current integration, instead of all of them:
+
+
+",1.0,"Fix calendar integration picker on Firefox - There's an issue with Firefox that provokes the calendar integration picker to only display the current integration, instead of all of them:
+
+
+",1,fix calendar integration picker on firefox there s an issue with firefox that provokes the calendar integration picker to only display the current integration instead of all of them ,1
+278,5345463110.0,IssuesEvent,2017-02-17 17:01:07,PopulateTools/gobierto,https://api.github.com/repos/PopulateTools/gobierto,closed,Share buttons integration + OG metatags,gobierto-people,"- [x] Metatags: description
+- [x] OG tags for TW + FB; title, img (default: logo), description
+- [x] Share buttons: with the page title + url",1.0,"Share buttons integration + OG metatags - - [x] Metatags: description
+- [x] OG tags for TW + FB; title, img (default: logo), description
+- [x] Share buttons: with the page title + url",1,share buttons integration og metatags metatags description og tags for tw fb title img default logo description share buttons with the page title url,1
+699,12534534647.0,IssuesEvent,2020-06-04 19:36:55,microsoftgraph/microsoft-graph-toolkit,https://api.github.com/repos/microsoftgraph/microsoft-graph-toolkit,closed,[BUG] People Picker using moving cursor with keyboard while searching breaks search,1.3 Component: people-picker Priority: 0 State: In Review bug,"
+
+
+**Describe the bug**
+When typing in the people picker and moving the cursor with the left or right arrow, it gets stuck on the loading animation
+
+**To Reproduce**
+1. Use the people picker
+2. Search for ""meg"" and notice the flyout lists Megan
+3. Move the cursor with the left arrow keyboard button and notice the flyout gets stuck in loading state
+
+
+**Expected behavior**
+Moving the cursor should not have any effect on the search results
+
+",1.0,"[BUG] People Picker using moving cursor with keyboard while searching breaks search -
+
+
+**Describe the bug**
+When typing in the people picker and moving the cursor with the left or right arrow, it gets stuck on the loading animation
+
+**To Reproduce**
+1. Use the people picker
+2. Search for ""meg"" and notice the flyout lists Megan
+3. Move the cursor with the left arrow keyboard button and notice the flyout gets stuck in loading state
+
+
+**Expected behavior**
+Moving the cursor should not have any effect on the search results
+
+",1, people picker using moving cursor with keyboard while searching breaks search describe the bug when typing in the people picker and moving the cursor with the left or right arrow it gets stuck on the loading animation to reproduce use the people picker search for meg and notice the flyout lists megan move the cursor with the left arrow keyboard button and notice the flyout gets stuck in loading state expected behavior moving the cursor should not have any effect on the search results ,1
+15517,2858829459.0,IssuesEvent,2015-06-03 06:40:54,hazelcast/hazelcast,https://api.github.com/repos/hazelcast/hazelcast,opened,[TEST-FAILURE] ClientNearCacheTest.putToCacheAndUpdateFromOtherNodeThenGetUpdatedFromClientNearCacheWithBinaryInMemoryFormat,Team: Core Type: Defect,https://hazelcast-l337.ci.cloudbees.com/job/Hazelcast-3.x/3268/com.hazelcast$hazelcast-client-new/testReport/com.hazelcast.client.cache.nearcache/ClientNearCacheTest/putToCacheAndUpdateFromOtherNodeThenGetUpdatedFromClientNearCacheWithBinaryInMemoryFormat/,1.0,[TEST-FAILURE] ClientNearCacheTest.putToCacheAndUpdateFromOtherNodeThenGetUpdatedFromClientNearCacheWithBinaryInMemoryFormat - https://hazelcast-l337.ci.cloudbees.com/job/Hazelcast-3.x/3268/com.hazelcast$hazelcast-client-new/testReport/com.hazelcast.client.cache.nearcache/ClientNearCacheTest/putToCacheAndUpdateFromOtherNodeThenGetUpdatedFromClientNearCacheWithBinaryInMemoryFormat/,0, clientnearcachetest puttocacheandupdatefromothernodethengetupdatedfromclientnearcachewithbinaryinmemoryformat ,0
+109425,4387217278.0,IssuesEvent,2016-08-08 15:11:08,a8cteam51/smittenkitchen,https://api.github.com/repos/a8cteam51/smittenkitchen,closed,Improve print behaviour,enhancement high-priority,"1. There are two print templates for recipes, the one from a ""Print"" link in the recipe card (leads to an unformatted page) and the one in the print icon (leads to a formatted page). The first I think should be the same as the second.
+
+2. The second readers are unhappy with because they want recipes to print (except when they're very long) on one page again. Is it possible to use a smaller font size and tighter line spacing so that most recipes will fit on one or at most two pages? There were also a few requests for a darker font color so it's easier to read.
+
+3. The first print link, the in the recipe card, doesn't print from phones or tablets.
+
+4. Finally, it would be great to have the site's logo back on top. Recipe printouts tend to get distributed far and wide and I'd love if we can make them as professionally reflective of the site they came from as possible. Thank you.",1.0,"Improve print behaviour - 1. There are two print templates for recipes, the one from a ""Print"" link in the recipe card (leads to an unformatted page) and the one in the print icon (leads to a formatted page). The first I think should be the same as the second.
+
+2. The second readers are unhappy with because they want recipes to print (except when they're very long) on one page again. Is it possible to use a smaller font size and tighter line spacing so that most recipes will fit on one or at most two pages? There were also a few requests for a darker font color so it's easier to read.
+
+3. The first print link, the in the recipe card, doesn't print from phones or tablets.
+
+4. Finally, it would be great to have the site's logo back on top. Recipe printouts tend to get distributed far and wide and I'd love if we can make them as professionally reflective of the site they came from as possible. Thank you.",0,improve print behaviour there are two print templates for recipes the one from a print link in the recipe card leads to an unformatted page and the one in the print icon leads to a formatted page the first i think should be the same as the second the second readers are unhappy with because they want recipes to print except when they re very long on one page again is it possible to use a smaller font size and tighter line spacing so that most recipes will fit on one or at most two pages there were also a few requests for a darker font color so it s easier to read the first print link the in the recipe card doesn t print from phones or tablets finally it would be great to have the site s logo back on top recipe printouts tend to get distributed far and wide and i d love if we can make them as professionally reflective of the site they came from as possible thank you ,0
+547,9661204142.0,IssuesEvent,2019-05-20 17:22:59,planningcenter/developers,https://api.github.com/repos/planningcenter/developers,closed,Inconsistent response for workflow_cards,People,"**Affected Applications**
+People
+
+**Describe the bug**
+My end goal is to create a report for our office staff listing every workflow card that is overdue for **all assignees** with all relevant information related to the workflow and the card in the report (I'll be exporting to csv or xlsx).
+
+I am, however, running into a situation where the web interface is showing active cards assigned to an individual, but the API response does not provide card data for that individual. (works for some and not for others... see below).
+
+Here are the steps that I'm going through to pull the cards for the assignees:
+
+1. I get a list of the Workflows https://api.planningcenteronline.com/people/v2/workflows?per_page=200
+2. I filter where total_ready_and_snoozed_card_count is greater than 0
+3. From there I get a list of all the steps (for each workflow): https://api.planningcenteronline.com/people/v2/workflows/$workflowID/steps
+4. I then get the assignee from each step: https://api.planningcenteronline.com/people/v2/workflows/$WorkFlowID/steps/$stepID/assignee_summaries
+5. At that point, I take each assignee ID and pull the workflow cards assigned to the person: https://api.planningcenteronline.com/people/v2/people/' + $personID + '/workflow_cards
+
+Issue: The response should include ""data"" for all users that have cards assigned, For **some** users it works as designed. For other users, it returns only ""links"" and ""meta"", with meta showing no cards. ""Data"" is null.
+
+Specific example of one nonworking user: User has 2 cards assigned and active, (ID:32251118.
+
+- Web interface link: https://people.planningcenteronline.com/workflows/106177/steps/278694/assignees/32251118/ready
+
+Here is the response I get from querying her card assignments through the API:
+
+```powershell
+PS> invoke-RestMethod -URI ('https://api.planningcenteronline.com/people/v2/people/32251118/workflow_cards') -Credential $credential
+
+links data included meta
+----- ---- -------- ----
+@{self=https://api.planningcenteronline.com/people/v2/people/32251118/workflow_cards} {} {} @{total_count=0;...
+```
+
+As you can see, ""data"" is null.
+
+On the same query, using person ID: 32476237, data is populated with the workflow cards. Web interface link: https://people.planningcenteronline.com/workflows/106177/steps/278694/assignees/32476237/ready
+
+```powershell
+PS C:\Repositories> invoke-RestMethod -URI ('https://api.planningcenteronline.com/people/v2/people/32476237/workflow_cards') -Credential $credential
+
+links data
+----- ----
+@{self=https://api.planningcenteronline.com/people/v2/people/32476237/workflow_cards} {@{type=WorkflowCard; id=4232506; attributes=; relationships=...
+```
+
+Expected behavior:
+Be able to list the workflow_cards for every individual that has the card assigned through the API
+
+**Context (please complete the following information):**
+- Endpoint: https://api.planningcenteronline.com/people/v2
+- Language: PowerShell
+- Authentication: Personal Access Token
+",1.0,"Inconsistent response for workflow_cards - **Affected Applications**
+People
+
+**Describe the bug**
+My end goal is to create a report for our office staff listing every workflow card that is overdue for **all assignees** with all relevant information related to the workflow and the card in the report (I'll be exporting to csv or xlsx).
+
+I am, however, running into a situation where the web interface is showing active cards assigned to an individual, but the API response does not provide card data for that individual. (works for some and not for others... see below).
+
+Here are the steps that I'm going through to pull the cards for the assignees:
+
+1. I get a list of the Workflows https://api.planningcenteronline.com/people/v2/workflows?per_page=200
+2. I filter where total_ready_and_snoozed_card_count is greater than 0
+3. From there I get a list of all the steps (for each workflow): https://api.planningcenteronline.com/people/v2/workflows/$workflowID/steps
+4. I then get the assignee from each step: https://api.planningcenteronline.com/people/v2/workflows/$WorkFlowID/steps/$stepID/assignee_summaries
+5. At that point, I take each assignee ID and pull the workflow cards assigned to the person: https://api.planningcenteronline.com/people/v2/people/' + $personID + '/workflow_cards
+
+Issue: The response should include ""data"" for all users that have cards assigned, For **some** users it works as designed. For other users, it returns only ""links"" and ""meta"", with meta showing no cards. ""Data"" is null.
+
+Specific example of one nonworking user: User has 2 cards assigned and active, (ID:32251118.
+
+- Web interface link: https://people.planningcenteronline.com/workflows/106177/steps/278694/assignees/32251118/ready
+
+Here is the response I get from querying her card assignments through the API:
+
+```powershell
+PS> invoke-RestMethod -URI ('https://api.planningcenteronline.com/people/v2/people/32251118/workflow_cards') -Credential $credential
+
+links data included meta
+----- ---- -------- ----
+@{self=https://api.planningcenteronline.com/people/v2/people/32251118/workflow_cards} {} {} @{total_count=0;...
+```
+
+As you can see, ""data"" is null.
+
+On the same query, using person ID: 32476237, data is populated with the workflow cards. Web interface link: https://people.planningcenteronline.com/workflows/106177/steps/278694/assignees/32476237/ready
+
+```powershell
+PS C:\Repositories> invoke-RestMethod -URI ('https://api.planningcenteronline.com/people/v2/people/32476237/workflow_cards') -Credential $credential
+
+links data
+----- ----
+@{self=https://api.planningcenteronline.com/people/v2/people/32476237/workflow_cards} {@{type=WorkflowCard; id=4232506; attributes=; relationships=...
+```
+
+Expected behavior:
+Be able to list the workflow_cards for every individual that has the card assigned through the API
+
+**Context (please complete the following information):**
+- Endpoint: https://api.planningcenteronline.com/people/v2
+- Language: PowerShell
+- Authentication: Personal Access Token
+",1,inconsistent response for workflow cards affected applications people describe the bug my end goal is to create a report for our office staff listing every workflow card that is overdue for all assignees with all relevant information related to the workflow and the card in the report i ll be exporting to csv or xlsx i am however running into a situation where the web interface is showing active cards assigned to an individual but the api response does not provide card data for that individual works for some and not for others see below here are the steps that i m going through to pull the cards for the assignees i get a list of the workflows i filter where total ready and snoozed card count is greater than from there i get a list of all the steps for each workflow i then get the assignee from each step at that point i take each assignee id and pull the workflow cards assigned to the person personid workflow cards issue the response should include data for all users that have cards assigned for some users it works as designed for other users it returns only links and meta with meta showing no cards data is null specific example of one nonworking user user has cards assigned and active id web interface link here is the response i get from querying her card assignments through the api powershell ps invoke restmethod uri credential credential links data included meta self total count as you can see data is null on the same query using person id data is populated with the workflow cards web interface link powershell ps c repositories invoke restmethod uri credential credential links data self type workflowcard id attributes relationships expected behavior be able to list the workflow cards for every individual that has the card assigned through the api context please complete the following information endpoint language powershell authentication personal access token ,1
+31208,5918275590.0,IssuesEvent,2017-05-22 15:02:41,openfin/fin-hypergrid,https://api.github.com/repos/openfin/fin-hypergrid,closed,DataModel Abstraction: Example sorting and filtering,Documentation,"Hi, the performance of hypergrid is great! However, I am a little lost with the documentation.
+
+What are the steps to enable sorting and filtering, for the version 1.3.0 of hypergrid (using hyper-analytics tree view data source)?
+
+---------------
+
+From master (1.3.0), I am trying to use the tree-view add-on demo (I see that in develop, this is removed). On chrome the sorter does not sort the treeview columns (expected sorting at each depth), and the Filter does not seem to work. For IE11 there is a failure on rendering the headers.
+
+I am using the sample add-ons:
+- grouped headers
+- tree view (with the old drilldown and rowById which I seems will be deprecated/merged into hypergrid)
+
+I see hyper-analytics 0.12.0 allows adding a DataSource filter and sorter to the pipeline, but I am not familiar enough with the HyperGrid stack to know which steps I am missing, or what would replace the filter/sorter controller, and or plugins, in order to make the filter and sorter work with the Tree View",1.0,"DataModel Abstraction: Example sorting and filtering - Hi, the performance of hypergrid is great! However, I am a little lost with the documentation.
+
+What are the steps to enable sorting and filtering, for the version 1.3.0 of hypergrid (using hyper-analytics tree view data source)?
+
+---------------
+
+From master (1.3.0), I am trying to use the tree-view add-on demo (I see that in develop, this is removed). On chrome the sorter does not sort the treeview columns (expected sorting at each depth), and the Filter does not seem to work. For IE11 there is a failure on rendering the headers.
+
+I am using the sample add-ons:
+- grouped headers
+- tree view (with the old drilldown and rowById which I seems will be deprecated/merged into hypergrid)
+
+I see hyper-analytics 0.12.0 allows adding a DataSource filter and sorter to the pipeline, but I am not familiar enough with the HyperGrid stack to know which steps I am missing, or what would replace the filter/sorter controller, and or plugins, in order to make the filter and sorter work with the Tree View",0,datamodel abstraction example sorting and filtering hi the performance of hypergrid is great however i am a little lost with the documentation what are the steps to enable sorting and filtering for the version of hypergrid using hyper analytics tree view data source from master i am trying to use the tree view add on demo i see that in develop this is removed on chrome the sorter does not sort the treeview columns expected sorting at each depth and the filter does not seem to work for there is a failure on rendering the headers i am using the sample add ons grouped headers tree view with the old drilldown and rowbyid which i seems will be deprecated merged into hypergrid i see hyper analytics allows adding a datasource filter and sorter to the pipeline but i am not familiar enough with the hypergrid stack to know which steps i am missing or what would replace the filter sorter controller and or plugins in order to make the filter and sorter work with the tree view,0
+823861,31070759210.0,IssuesEvent,2023-08-12 00:04:35,ubiquibot/staging,https://api.github.com/repos/ubiquibot/staging,closed,"Disable Bounties for ""Parent Issues""",Time: <1 Hour Priority: 0 (Normal),"QA for [268](Disable Bounties for ""Parent Issues"")
+
+- [x] [1](https://github.com/ubiquibot/staging/issues/9)
+- [x] [2](https://github.com/ubiquibot/staging/issues/14)
+- [ ] #85
+- [ ] 4",1.0,"Disable Bounties for ""Parent Issues"" - QA for [268](Disable Bounties for ""Parent Issues"")
+
+- [x] [1](https://github.com/ubiquibot/staging/issues/9)
+- [x] [2](https://github.com/ubiquibot/staging/issues/14)
+- [ ] #85
+- [ ] 4",0,disable bounties for parent issues qa for disable bounties for parent issues ,0
+270936,8474744799.0,IssuesEvent,2018-10-24 16:58:36,brainvisa/testbidon,https://api.github.com/repos/brainvisa/testbidon,closed,Add soma-pipeline to soma weak dependencies for build-config,Priority: Normal Status: Closed Tracker: Feature,"---
+
+
+Author Name: **Souedet, Nicolas** (Souedet, Nicolas)
+Original Redmine Issue: 1319, https://bioproj.extra.cea.fr/redmine/issues/1319
+Original Date: 2010-05-18
+Original Assignee: Souedet, Nicolas
+
+---
+
+None
+
+
+",1.0,"Add soma-pipeline to soma weak dependencies for build-config - ---
+
+
+Author Name: **Souedet, Nicolas** (Souedet, Nicolas)
+Original Redmine Issue: 1319, https://bioproj.extra.cea.fr/redmine/issues/1319
+Original Date: 2010-05-18
+Original Assignee: Souedet, Nicolas
+
+---
+
+None
+
+
+",0,add soma pipeline to soma weak dependencies for build config author name souedet nicolas souedet nicolas original redmine issue original date original assignee souedet nicolas none ,0
+108885,9334891083.0,IssuesEvent,2019-03-28 17:17:43,SNLComputation/Albany,https://api.github.com/repos/SNLComputation/Albany,closed,Compilation errors due to libma.a,SCOREC Testing question,"It seems in the past few days, compilations errors due to routines in the libma.a libray have crept in:
+```
+[ 97%] Built target BifurcationTest
+Scanning dependencies of target MeshComponents
+
+[ 97%] Building CXX object src/LCM/CMakeFiles/MeshComponents.dir/test/utils/MeshComponents.cpp.o
+/.../test/TrilinosInstall/lib/libma.a(maShape.cc.o): In function `ma::fixLargeAngleTets(ma::Adapt*)':
+
+maShape.cc:(.text+0x1476): undefined reference to `ma::SingleSplitCollapse::SingleSplitCollapse(ma::Adapt*)'
+```
+Does anyone know what is this library so we can assign this issue to the right person?
+",1.0,"Compilation errors due to libma.a - It seems in the past few days, compilations errors due to routines in the libma.a libray have crept in:
+```
+[ 97%] Built target BifurcationTest
+Scanning dependencies of target MeshComponents
+
+[ 97%] Building CXX object src/LCM/CMakeFiles/MeshComponents.dir/test/utils/MeshComponents.cpp.o
+/.../test/TrilinosInstall/lib/libma.a(maShape.cc.o): In function `ma::fixLargeAngleTets(ma::Adapt*)':
+
+maShape.cc:(.text+0x1476): undefined reference to `ma::SingleSplitCollapse::SingleSplitCollapse(ma::Adapt*)'
+```
+Does anyone know what is this library so we can assign this issue to the right person?
+",0,compilation errors due to libma a it seems in the past few days compilations errors due to routines in the libma a libray have crept in built target bifurcationtest scanning dependencies of target meshcomponents building cxx object src lcm cmakefiles meshcomponents dir test utils meshcomponents cpp o test trilinosinstall lib libma a mashape cc o in function ma fixlargeangletets ma adapt mashape cc text undefined reference to ma singlesplitcollapse singlesplitcollapse ma adapt does anyone know what is this library so we can assign this issue to the right person ,0
+196,4184548432.0,IssuesEvent,2016-06-23 07:36:06,awesome-jobs/vietnam,https://api.github.com/repos/awesome-jobs/vietnam,closed,8 Bit Rockstars - Project Manager - Saigon - FT,Full-time People Skill Saigon,"
+# 8 BIT ROCKSTARS
+Floor 8th, Vincom B, 72 Le Thanh Ton, Ben Nghe Ward, Dist.1, HCMC
+http://8bitrockr.com/
+hr@8bitrockr.com
+
+Email CV to **_hr@8bitrockr.com_**
+
+## Salary
+
+**_Up to 3,000 USD_**
+
+## Job Description
+
+We are looking for an **experienced Project manager**, who will work with us on building and scaling our startup teams.
+
+**You will have responsibility for:**
+
+- Managing full life-cycle customer projects
+- Developing strong relationships with customers, acting as the primary point of contact as well as managing stakeholder expectations
+- You are concerned about client satisfaction and always seek for solution improvement.
+- Motivating and leading the engineers with agile methods
+
+## Requirement
+
+- A degree would be nice, however attitude and experience matters more.
+- Good communication skills, written and spoken in English.
+- Strong technical and project management skills.
+- Negotiation skill.
+- 2+ years' Project Management or technical leadership experience.
+- Experience Web and Software Development.
+- Agile methods.
+- Interested in travelling and visiting our customers on-site.
+
+## Benefits:
+
+- Open and honest culture where people are valued, treated fairly, trusted and empowered to do great things
+- Attractive, **competitive salary** and incentives
+- Year-end **bonus** (13th month salary ++)
+- 40 hour work weeks (9:00-18:00, Mon-Fri)
+- **15 days** of vacation per year
+- **Probation** with **100% of salary**
+- Performance and **salary review** every **6 months**
+- **International Medical Insurance** will be paid by the company
+- **English Class** with foreigner 3 times a week.
+- New **Technologies Club** every week
+- Quarterly **rockstar Hackathon** with the Best prizes
+- Short and long-term **international training** / work
+- **Company trip** (local or internationally) every year
+- **Team building events**, party monthly and much more
+- **Entertainment Zone** for all members with full media in the company (game, music, TV…).
+- Enjoy **Relax and Coffee Bar** in the company.
+
+Drop your CV to: **_hr@8bitrockr.com_** and we will be in touch with you shortly!",1.0,"8 Bit Rockstars - Project Manager - Saigon - FT - 
+# 8 BIT ROCKSTARS
+Floor 8th, Vincom B, 72 Le Thanh Ton, Ben Nghe Ward, Dist.1, HCMC
+http://8bitrockr.com/
+hr@8bitrockr.com
+
+Email CV to **_hr@8bitrockr.com_**
+
+## Salary
+
+**_Up to 3,000 USD_**
+
+## Job Description
+
+We are looking for an **experienced Project manager**, who will work with us on building and scaling our startup teams.
+
+**You will have responsibility for:**
+
+- Managing full life-cycle customer projects
+- Developing strong relationships with customers, acting as the primary point of contact as well as managing stakeholder expectations
+- You are concerned about client satisfaction and always seek for solution improvement.
+- Motivating and leading the engineers with agile methods
+
+## Requirement
+
+- A degree would be nice, however attitude and experience matters more.
+- Good communication skills, written and spoken in English.
+- Strong technical and project management skills.
+- Negotiation skill.
+- 2+ years' Project Management or technical leadership experience.
+- Experience Web and Software Development.
+- Agile methods.
+- Interested in travelling and visiting our customers on-site.
+
+## Benefits:
+
+- Open and honest culture where people are valued, treated fairly, trusted and empowered to do great things
+- Attractive, **competitive salary** and incentives
+- Year-end **bonus** (13th month salary ++)
+- 40 hour work weeks (9:00-18:00, Mon-Fri)
+- **15 days** of vacation per year
+- **Probation** with **100% of salary**
+- Performance and **salary review** every **6 months**
+- **International Medical Insurance** will be paid by the company
+- **English Class** with foreigner 3 times a week.
+- New **Technologies Club** every week
+- Quarterly **rockstar Hackathon** with the Best prizes
+- Short and long-term **international training** / work
+- **Company trip** (local or internationally) every year
+- **Team building events**, party monthly and much more
+- **Entertainment Zone** for all members with full media in the company (game, music, TV…).
+- Enjoy **Relax and Coffee Bar** in the company.
+
+Drop your CV to: **_hr@8bitrockr.com_** and we will be in touch with you shortly!",1, bit rockstars project manager saigon ft bit rockstars floor vincom b le thanh ton ben nghe ward dist hcmc hr com email cv to hr com salary up to usd job description we are looking for an experienced project manager who will work with us on building and scaling our startup teams you will have responsibility for managing full life cycle customer projects developing strong relationships with customers acting as the primary point of contact as well as managing stakeholder expectations you are concerned about client satisfaction and always seek for solution improvement motivating and leading the engineers with agile methods requirement a degree would be nice however attitude and experience matters more good communication skills written and spoken in english strong technical and project management skills negotiation skill years project management or technical leadership experience experience web and software development agile methods interested in travelling and visiting our customers on site benefits open and honest culture where people are valued treated fairly trusted and empowered to do great things attractive competitive salary and incentives year end bonus month salary hour work weeks mon fri days of vacation per year probation with of salary performance and salary review every months international medical insurance will be paid by the company english class with foreigner times a week new technologies club every week quarterly rockstar hackathon with the best prizes short and long term international training work company trip local or internationally every year team building events party monthly and much more entertainment zone for all members with full media in the company game music tv… enjoy relax and coffee bar in the company drop your cv to hr com and we will be in touch with you shortly ,1
+865,16543407228.0,IssuesEvent,2021-05-27 20:00:12,microsoftgraph/microsoft-graph-toolkit,https://api.github.com/repos/microsoftgraph/microsoft-graph-toolkit,closed,Cannot enter the lowercase letter 'i' in PeoplePicker inside the SPO Document Library.,Component: people-picker Needs: Author Feedback no-recent-activity,"For the PeoplePicker from '@microsoft/mgt-react' inside the SPO document library, I cannot type the lowercase letter 'i' anywhere in the input, but I can press all other keys.
+
+When I attempt to call the events for onInput() or onKeyPress() from the PeoplePicker, they're never called for that key press, so I can't capture the event.
+
+**Describe the bug**
+A clear and concise description of the bug
+
+**To Reproduce**
+Steps to reproduce the behavior:
+1. Go to '...'
+2. Click on '....'
+3. Scroll down to '....'
+4. See error
+
+
+**Expected behavior**
+A clear and concise description of what you expected to happen.
+
+**Screenshots**
+If applicable, add screenshots to help explain your problem.
+
+**Environment (please complete the following information):**
+ - OS: [e.g. iOS]
+ - Browser [e.g. edge, chrome, safari]
+ - Framework [e.g. react, vue, none]
+ - Context [e.g. Microsoft Teams, SharePoint, Office Add-ins, Web]
+ - Version [e.g. 0.1]
+ - Provider [e.g. Msal2Provider] (bonus if you share the provider config)
+
+**Additional context**
+Add any other context about the problem here.",1.0,"Cannot enter the lowercase letter 'i' in PeoplePicker inside the SPO Document Library. - For the PeoplePicker from '@microsoft/mgt-react' inside the SPO document library, I cannot type the lowercase letter 'i' anywhere in the input, but I can press all other keys.
+
+When I attempt to call the events for onInput() or onKeyPress() from the PeoplePicker, they're never called for that key press, so I can't capture the event.
+
+**Describe the bug**
+A clear and concise description of the bug
+
+**To Reproduce**
+Steps to reproduce the behavior:
+1. Go to '...'
+2. Click on '....'
+3. Scroll down to '....'
+4. See error
+
+
+**Expected behavior**
+A clear and concise description of what you expected to happen.
+
+**Screenshots**
+If applicable, add screenshots to help explain your problem.
+
+**Environment (please complete the following information):**
+ - OS: [e.g. iOS]
+ - Browser [e.g. edge, chrome, safari]
+ - Framework [e.g. react, vue, none]
+ - Context [e.g. Microsoft Teams, SharePoint, Office Add-ins, Web]
+ - Version [e.g. 0.1]
+ - Provider [e.g. Msal2Provider] (bonus if you share the provider config)
+
+**Additional context**
+Add any other context about the problem here.",1,cannot enter the lowercase letter i in peoplepicker inside the spo document library for the peoplepicker from microsoft mgt react inside the spo document library i cannot type the lowercase letter i anywhere in the input but i can press all other keys when i attempt to call the events for oninput or onkeypress from the peoplepicker they re never called for that key press so i can t capture the event describe the bug a clear and concise description of the bug to reproduce steps to reproduce the behavior go to click on scroll down to see error expected behavior a clear and concise description of what you expected to happen screenshots if applicable add screenshots to help explain your problem environment please complete the following information os browser framework context version provider bonus if you share the provider config additional context add any other context about the problem here ,1
+443554,30923434345.0,IssuesEvent,2023-08-06 07:22:46,PyPSA/pypsa-eur,https://api.github.com/repos/PyPSA/pypsa-eur,closed,Document existing and planned validation,help wanted documentation,"PyPSA-Eur(-Sec) can be validated against historical data about the energy system to see if it reproduces energy balances, energy prices, import/export flows, curtailment, redispatch and various other energy system statistics. Several validations have already been done, and some are planned. The different validations should be collected here, and then added to a new page ""Validation"" in the [documentation for PyPSA-Eur](https://pypsa-eur.readthedocs.io/en/latest/). Since PyPSA-Eur-Sec will be [merged into PyPSA-Eur](https://github.com/PyPSA/pypsa-eur/issues/463), we include issues affected other energy sectors too.
+
+## Existing validation
+- [PyPSA-Eur paper](https://arxiv.org/abs/1806.01613) includes validation of circuit lengths, network topology and matching of load to grid (looking at load shedding)
+- @martacki [paper validating historical curtailment in German transmission network](https://arxiv.org/abs/2009.03801)
+- validation of historical redispatch in Germany by ETH Zürich group (in Willem Laumen Master thesis)
+- some validation of curtailment in [DLR paper](https://arxiv.org/abs/2208.04716) by @FabianHofmann and @p-glaum
+- [validation by Freiburg group](https://freidok.uni-freiburg.de/data/227624) by Freddie, Mirko and Anke, for import/export flows and load distribution
+- @matteodefelice [modelling of European system with ENTSO-E data (no PyPSA-Eur)](https://www.matteodefelice.name/post/pypsa-entsoe/), including price time series where peaks are missed (probably because of missing unit commitment)
+
+## Ongoing and planned validation
+- [Validation of power system against 2019](https://github.com/PyPSA/pypsa-eur/issues/363) - should include electricity price time series with linearised unit commitment, ramping, CO2 certificate price time series from historical, heat-led CHP and updated varied efficiencies of gas plants
+- [Improved differentiation of vehicle types in transport model in PyPSA-Eur-Sec](https://github.com/PyPSA/pypsa-eur/issues/491)
+- [Improved spatial distribution of district heating, particularly in Germany](https://github.com/PyPSA/pypsa-eur/issues/490)
+- Improved development of RE over 2020-2050 (e.g. going from existing power plants to those with higher hub height, smaller generator to rotor ratio and larger size for wind, and axis tracking for PV with lower AC-DC ratio)
+- Comparison of network impedances with [Core FBMC model](https://www.jao.eu/core-fb-mc), see @fneum [Core TSO project](https://github.com/fneum/core-tso-data)
+- Redispatch again following [SciGRID redispatch example](https://pypsa.readthedocs.io/en/latest/examples/scigrid-redispatch.html)
+- Comparison with [CORE grid model](https://www.jao.eu/static-grid-model) and OSM from PyPSA-Earth
+- [TYNDP projects PR](https://github.com/PyPSA/pypsa-eur/pull/626) from @grecht",1.0,"Document existing and planned validation - PyPSA-Eur(-Sec) can be validated against historical data about the energy system to see if it reproduces energy balances, energy prices, import/export flows, curtailment, redispatch and various other energy system statistics. Several validations have already been done, and some are planned. The different validations should be collected here, and then added to a new page ""Validation"" in the [documentation for PyPSA-Eur](https://pypsa-eur.readthedocs.io/en/latest/). Since PyPSA-Eur-Sec will be [merged into PyPSA-Eur](https://github.com/PyPSA/pypsa-eur/issues/463), we include issues affected other energy sectors too.
+
+## Existing validation
+- [PyPSA-Eur paper](https://arxiv.org/abs/1806.01613) includes validation of circuit lengths, network topology and matching of load to grid (looking at load shedding)
+- @martacki [paper validating historical curtailment in German transmission network](https://arxiv.org/abs/2009.03801)
+- validation of historical redispatch in Germany by ETH Zürich group (in Willem Laumen Master thesis)
+- some validation of curtailment in [DLR paper](https://arxiv.org/abs/2208.04716) by @FabianHofmann and @p-glaum
+- [validation by Freiburg group](https://freidok.uni-freiburg.de/data/227624) by Freddie, Mirko and Anke, for import/export flows and load distribution
+- @matteodefelice [modelling of European system with ENTSO-E data (no PyPSA-Eur)](https://www.matteodefelice.name/post/pypsa-entsoe/), including price time series where peaks are missed (probably because of missing unit commitment)
+
+## Ongoing and planned validation
+- [Validation of power system against 2019](https://github.com/PyPSA/pypsa-eur/issues/363) - should include electricity price time series with linearised unit commitment, ramping, CO2 certificate price time series from historical, heat-led CHP and updated varied efficiencies of gas plants
+- [Improved differentiation of vehicle types in transport model in PyPSA-Eur-Sec](https://github.com/PyPSA/pypsa-eur/issues/491)
+- [Improved spatial distribution of district heating, particularly in Germany](https://github.com/PyPSA/pypsa-eur/issues/490)
+- Improved development of RE over 2020-2050 (e.g. going from existing power plants to those with higher hub height, smaller generator to rotor ratio and larger size for wind, and axis tracking for PV with lower AC-DC ratio)
+- Comparison of network impedances with [Core FBMC model](https://www.jao.eu/core-fb-mc), see @fneum [Core TSO project](https://github.com/fneum/core-tso-data)
+- Redispatch again following [SciGRID redispatch example](https://pypsa.readthedocs.io/en/latest/examples/scigrid-redispatch.html)
+- Comparison with [CORE grid model](https://www.jao.eu/static-grid-model) and OSM from PyPSA-Earth
+- [TYNDP projects PR](https://github.com/PyPSA/pypsa-eur/pull/626) from @grecht",0,document existing and planned validation pypsa eur sec can be validated against historical data about the energy system to see if it reproduces energy balances energy prices import export flows curtailment redispatch and various other energy system statistics several validations have already been done and some are planned the different validations should be collected here and then added to a new page validation in the since pypsa eur sec will be we include issues affected other energy sectors too existing validation includes validation of circuit lengths network topology and matching of load to grid looking at load shedding martacki validation of historical redispatch in germany by eth zürich group in willem laumen master thesis some validation of curtailment in by fabianhofmann and p glaum by freddie mirko and anke for import export flows and load distribution matteodefelice including price time series where peaks are missed probably because of missing unit commitment ongoing and planned validation should include electricity price time series with linearised unit commitment ramping certificate price time series from historical heat led chp and updated varied efficiencies of gas plants improved development of re over e g going from existing power plants to those with higher hub height smaller generator to rotor ratio and larger size for wind and axis tracking for pv with lower ac dc ratio comparison of network impedances with see fneum redispatch again following comparison with and osm from pypsa earth from grecht,0
+283,5438821763.0,IssuesEvent,2017-03-06 11:35:16,PopulateTools/gobierto,https://api.github.com/repos/PopulateTools/gobierto,closed,Structure for statements in Dival,gobierto-people,"ref: http://www.transparencia.gva.es/documents/162282364/162502126/Anexo+III-Registro+Altos+Cargos/91fe6dbd-bb01-4e2f-8eda-520e7a23b8b6;jsessionid=8A90CD9EEF2E77E9A074D9EBAF9065EA
+
+(first line is the name of the block)
+
+####
+
+Activitats públiques exercides durant els dos anys anteriors a la presa de possessió
+Actividades públicas ejercidas durante los dos años anteriores a la toma de posesión
+Entitat, organisme, empresa o societat
+Entidad, organismo, empresa o sociedad
+Activitat realitzada
+Actividad desempeñada
+Data d'inici
+Fecha de inicio
+Data fi
+Fecha fin
+[IGNORAR?] Feu constar si heu sol·licitat passar a la situació de serveis especials o equivalent, o a la que corresponga
+[IGNORAR?] Haga constar si ha solicitado el pase a la situación de servicios especiales o equivalente, o a la que corresponda
+
+Activitats privades remunerades exercides durant els dos anys anteriors a la presa de possessió
+Actividades privadas remuneradas ejercidas durante los dos años anteriores a la toma de posesión
+Activitat
+Actividad
+Empresa o entitat
+Empresa o entidad
+Autònom / Autònoma
+Autónomo / Autónoma
+Data d'inici
+Fecha de inicio
+Data fi
+Fecha fin
+
+####
+
+Càrrecs públics, participació en òrgans col·legiats i activitats en representació de la generalitat
+Cargos públicos, participación en órganos colegiados y actividades en representación de la generalitat
+Entitat
+Entidad
+Càrrec ocupat
+Cargo desempeñado
+Data de nomenament
+Fecha de nombramiento
+
+Activitats privades remunerades exercides durant els dos anys anteriors a la presa de possessió
+Actividades privadas remuneradas ejercidas durante los dos años anteriores a la toma de posesión
+Activitat
+Actividad
+Entidad
+Entidad
+Autònom / Autònoma
+Autónomo / Autónoma
+Data d'inici
+Fecha de inicio
+Data fi
+Fecha fin
+
+Càrrecs públics, participació en òrgans col·legiats i activitats en representació de la generalitat
+Cargos públicos, participación en órganos colegiados y actividades en representación de la generalitat
+Entitat
+Entidad
+Càrrec ocupat
+Cargo desempeñado
+Data de nomenament
+Fecha de nombramiento
+
+Altres activitats i carrecs no recollits al apartat anterior
+Otras actividades y cargos no recogidas en el apartado anterior
+Entitat
+Entidad
+Càrrec ocupat
+Cargo desempeñado
+Data de nomenament
+Fecha de nombramiento
+
+
+####
+
+Activitat a què va a dedicar-se al cessament
+Actividad a la que va a dedicarse al cese
+Activitat
+Actividad
+Empresa o entitat
+Empresa o entidad
+Autònom / Autònoma
+Autónomo / Autónoma
+Data d'inici
+Fecha de inicio
+
+
+####
+
+Béns immobles urbans i rústics
+Bienes inmuebles urbanos y rústicos
+Clau
+Clave
+Tipus
+Tipo
+% en cas de Domini ple
+% en caso de Dominio pleno
+Província
+Provincia
+Valor cadastral
+Valor catastral
+
+Béns i drets de naturalesa no immobiliària
+Bienes y derechos de naturaleza no inmobiliaria
+Descripció
+Descripción
+Valor (euros)
+Valor (euros)
+
+Passiu
+Pasivo
+Descripció
+Descripción
+Valor (euros)
+Valor (euros)
+
+
+####
+
+Sous y honoraris anuals, aranzels i altres retribucions
+Sueldos y honorarios anuales, aranceles y otras retribuciones
+Concepte
+Concepto
+Euros
+Euros
+
+Dividends i participació en beneficis de societats, comunitats o entitats de qualsevol classe
+Dividendos y participación en beneficios de sociedades, comunidades o entidades de cualquier clase.
+Concepte
+Concepto
+Euros
+Euros
+
+Interessos o rendiments de comptes, dipòsits i actius financers.
+Intereses o rendimientos de cuentas, depósitos y activos financieros.
+Concepte
+Concepto
+Euros
+Euros
+
+Altres rendes o percepcions de qualsevol classe
+Otras rentas o percepciones de cualquier clase.
+Concepte
+Concepto
+Euros
+Euros
+",1.0,"Structure for statements in Dival - ref: http://www.transparencia.gva.es/documents/162282364/162502126/Anexo+III-Registro+Altos+Cargos/91fe6dbd-bb01-4e2f-8eda-520e7a23b8b6;jsessionid=8A90CD9EEF2E77E9A074D9EBAF9065EA
+
+(first line is the name of the block)
+
+####
+
+Activitats públiques exercides durant els dos anys anteriors a la presa de possessió
+Actividades públicas ejercidas durante los dos años anteriores a la toma de posesión
+Entitat, organisme, empresa o societat
+Entidad, organismo, empresa o sociedad
+Activitat realitzada
+Actividad desempeñada
+Data d'inici
+Fecha de inicio
+Data fi
+Fecha fin
+[IGNORAR?] Feu constar si heu sol·licitat passar a la situació de serveis especials o equivalent, o a la que corresponga
+[IGNORAR?] Haga constar si ha solicitado el pase a la situación de servicios especiales o equivalente, o a la que corresponda
+
+Activitats privades remunerades exercides durant els dos anys anteriors a la presa de possessió
+Actividades privadas remuneradas ejercidas durante los dos años anteriores a la toma de posesión
+Activitat
+Actividad
+Empresa o entitat
+Empresa o entidad
+Autònom / Autònoma
+Autónomo / Autónoma
+Data d'inici
+Fecha de inicio
+Data fi
+Fecha fin
+
+####
+
+Càrrecs públics, participació en òrgans col·legiats i activitats en representació de la generalitat
+Cargos públicos, participación en órganos colegiados y actividades en representación de la generalitat
+Entitat
+Entidad
+Càrrec ocupat
+Cargo desempeñado
+Data de nomenament
+Fecha de nombramiento
+
+Activitats privades remunerades exercides durant els dos anys anteriors a la presa de possessió
+Actividades privadas remuneradas ejercidas durante los dos años anteriores a la toma de posesión
+Activitat
+Actividad
+Entidad
+Entidad
+Autònom / Autònoma
+Autónomo / Autónoma
+Data d'inici
+Fecha de inicio
+Data fi
+Fecha fin
+
+Càrrecs públics, participació en òrgans col·legiats i activitats en representació de la generalitat
+Cargos públicos, participación en órganos colegiados y actividades en representación de la generalitat
+Entitat
+Entidad
+Càrrec ocupat
+Cargo desempeñado
+Data de nomenament
+Fecha de nombramiento
+
+Altres activitats i carrecs no recollits al apartat anterior
+Otras actividades y cargos no recogidas en el apartado anterior
+Entitat
+Entidad
+Càrrec ocupat
+Cargo desempeñado
+Data de nomenament
+Fecha de nombramiento
+
+
+####
+
+Activitat a què va a dedicar-se al cessament
+Actividad a la que va a dedicarse al cese
+Activitat
+Actividad
+Empresa o entitat
+Empresa o entidad
+Autònom / Autònoma
+Autónomo / Autónoma
+Data d'inici
+Fecha de inicio
+
+
+####
+
+Béns immobles urbans i rústics
+Bienes inmuebles urbanos y rústicos
+Clau
+Clave
+Tipus
+Tipo
+% en cas de Domini ple
+% en caso de Dominio pleno
+Província
+Provincia
+Valor cadastral
+Valor catastral
+
+Béns i drets de naturalesa no immobiliària
+Bienes y derechos de naturaleza no inmobiliaria
+Descripció
+Descripción
+Valor (euros)
+Valor (euros)
+
+Passiu
+Pasivo
+Descripció
+Descripción
+Valor (euros)
+Valor (euros)
+
+
+####
+
+Sous y honoraris anuals, aranzels i altres retribucions
+Sueldos y honorarios anuales, aranceles y otras retribuciones
+Concepte
+Concepto
+Euros
+Euros
+
+Dividends i participació en beneficis de societats, comunitats o entitats de qualsevol classe
+Dividendos y participación en beneficios de sociedades, comunidades o entidades de cualquier clase.
+Concepte
+Concepto
+Euros
+Euros
+
+Interessos o rendiments de comptes, dipòsits i actius financers.
+Intereses o rendimientos de cuentas, depósitos y activos financieros.
+Concepte
+Concepto
+Euros
+Euros
+
+Altres rendes o percepcions de qualsevol classe
+Otras rentas o percepciones de cualquier clase.
+Concepte
+Concepto
+Euros
+Euros
+",1,structure for statements in dival ref first line is the name of the block activitats públiques exercides durant els dos anys anteriors a la presa de possessió actividades públicas ejercidas durante los dos años anteriores a la toma de posesión entitat organisme empresa o societat entidad organismo empresa o sociedad activitat realitzada actividad desempeñada data d inici fecha de inicio data fi fecha fin feu constar si heu sol·licitat passar a la situació de serveis especials o equivalent o a la que corresponga haga constar si ha solicitado el pase a la situación de servicios especiales o equivalente o a la que corresponda activitats privades remunerades exercides durant els dos anys anteriors a la presa de possessió actividades privadas remuneradas ejercidas durante los dos años anteriores a la toma de posesión activitat actividad empresa o entitat empresa o entidad autònom autònoma autónomo autónoma data d inici fecha de inicio data fi fecha fin càrrecs públics participació en òrgans col·legiats i activitats en representació de la generalitat cargos públicos participación en órganos colegiados y actividades en representación de la generalitat entitat entidad càrrec ocupat cargo desempeñado data de nomenament fecha de nombramiento activitats privades remunerades exercides durant els dos anys anteriors a la presa de possessió actividades privadas remuneradas ejercidas durante los dos años anteriores a la toma de posesión activitat actividad entidad entidad autònom autònoma autónomo autónoma data d inici fecha de inicio data fi fecha fin càrrecs públics participació en òrgans col·legiats i activitats en representació de la generalitat cargos públicos participación en órganos colegiados y actividades en representación de la generalitat entitat entidad càrrec ocupat cargo desempeñado data de nomenament fecha de nombramiento altres activitats i carrecs no recollits al apartat anterior otras actividades y cargos no recogidas en el apartado anterior entitat entidad càrrec ocupat cargo desempeñado data de nomenament fecha de nombramiento activitat a què va a dedicar se al cessament actividad a la que va a dedicarse al cese activitat actividad empresa o entitat empresa o entidad autònom autònoma autónomo autónoma data d inici fecha de inicio béns immobles urbans i rústics bienes inmuebles urbanos y rústicos clau clave tipus tipo en cas de domini ple en caso de dominio pleno província provincia valor cadastral valor catastral béns i drets de naturalesa no immobiliària bienes y derechos de naturaleza no inmobiliaria descripció descripción valor euros valor euros passiu pasivo descripció descripción valor euros valor euros sous y honoraris anuals aranzels i altres retribucions sueldos y honorarios anuales aranceles y otras retribuciones concepte concepto euros euros dividends i participació en beneficis de societats comunitats o entitats de qualsevol classe dividendos y participación en beneficios de sociedades comunidades o entidades de cualquier clase concepte concepto euros euros interessos o rendiments de comptes dipòsits i actius financers intereses o rendimientos de cuentas depósitos y activos financieros concepte concepto euros euros altres rendes o percepcions de qualsevol classe otras rentas o percepciones de cualquier clase concepte concepto euros euros ,1
+123030,10244843299.0,IssuesEvent,2019-08-20 11:27:07,mapbox/mapbox-gl-native,https://api.github.com/repos/mapbox/mapbox-gl-native,closed,[android] android-gnustl-arm-v7 can timeout and fail waiting for Firebase,Android bug tests,"The `android-gnustl-arm-v7` job c[an hit the 20 minute CircleCI timeout and fail](https://circleci.com/gh/mapbox/mapbox-gl-native/311274), even though [the Firebase device test may ultimately succeed](https://console.firebase.google.com/u/0/project/android-gl-native/testlab/histories/bh.1470d2ab45df6f1f/matrices/8232742097943354519) in “12 minutes, 15 seconds”.
+
+/cc @mapbox/maps-android
+
+",1.0,"[android] android-gnustl-arm-v7 can timeout and fail waiting for Firebase - The `android-gnustl-arm-v7` job c[an hit the 20 minute CircleCI timeout and fail](https://circleci.com/gh/mapbox/mapbox-gl-native/311274), even though [the Firebase device test may ultimately succeed](https://console.firebase.google.com/u/0/project/android-gl-native/testlab/histories/bh.1470d2ab45df6f1f/matrices/8232742097943354519) in “12 minutes, 15 seconds”.
+
+/cc @mapbox/maps-android
+
+",0, android gnustl arm can timeout and fail waiting for firebase the android gnustl arm job c even though in “ minutes seconds” cc mapbox maps android ,0
+39049,6718483930.0,IssuesEvent,2017-10-15 13:28:07,perl5-metaconfig/metaconfig,https://api.github.com/repos/perl5-metaconfig/metaconfig,opened,Workflow,Documentation,"• Do we need to document the workflow to the group?
+• Do we want a Label that members cat put with an issue like ""Taken"" or ""Work in progress"" so it is easier to pick issues to work on?",1.0,"Workflow - • Do we need to document the workflow to the group?
+• Do we want a Label that members cat put with an issue like ""Taken"" or ""Work in progress"" so it is easier to pick issues to work on?",0,workflow • do we need to document the workflow to the group • do we want a label that members cat put with an issue like taken or work in progress so it is easier to pick issues to work on ,0
+2741,12520026956.0,IssuesEvent,2020-06-03 15:15:54,wazuh/wazuh-qa,https://api.github.com/repos/wazuh/wazuh-qa,opened,Vuln detector: Test extra fields in Redhat feeds,automation core/vuln detector,"Hello team,
+
+The purpose of this issue is to develop a test suite to check the behavior of vulnerability detector against a Redhat feed in which there are extra fields of all types.
+
+Parent issue #742
+
+Best regards.",1.0,"Vuln detector: Test extra fields in Redhat feeds - Hello team,
+
+The purpose of this issue is to develop a test suite to check the behavior of vulnerability detector against a Redhat feed in which there are extra fields of all types.
+
+Parent issue #742
+
+Best regards.",0,vuln detector test extra fields in redhat feeds hello team the purpose of this issue is to develop a test suite to check the behavior of vulnerability detector against a redhat feed in which there are extra fields of all types parent issue best regards ,0
+41627,5345946375.0,IssuesEvent,2017-02-17 18:19:21,elastic/kibana,https://api.github.com/repos/elastic/kibana,opened,"Index creation UX: ""Missing index"" error reporting",:Design :Management,"_Taken from https://github.com/elastic/kibana/pull/9991#issuecomment-274556775:_
+
+
+
+The message ""Unable to fetch mapping. Do you have indices matching the pattern?"" is too disconnected from the source of the error (the bad index pattern field). I think the solution is:
+
+* This button should just become disabled without changing text.
+* The index pattern field should be outlined in red.
+* There should be an error message below it that states, ""Do you have indices matching this pattern? We couldn't fetch the mapping."" (And what does ""couldn't fetch the mapping mean""?"" Doesn't it make more sense to say, ""couldn't find any matching indices?"")
+",1.0,"Index creation UX: ""Missing index"" error reporting - _Taken from https://github.com/elastic/kibana/pull/9991#issuecomment-274556775:_
+
+
+
+The message ""Unable to fetch mapping. Do you have indices matching the pattern?"" is too disconnected from the source of the error (the bad index pattern field). I think the solution is:
+
+* This button should just become disabled without changing text.
+* The index pattern field should be outlined in red.
+* There should be an error message below it that states, ""Do you have indices matching this pattern? We couldn't fetch the mapping."" (And what does ""couldn't fetch the mapping mean""?"" Doesn't it make more sense to say, ""couldn't find any matching indices?"")
+",0,index creation ux missing index error reporting taken from the message unable to fetch mapping do you have indices matching the pattern is too disconnected from the source of the error the bad index pattern field i think the solution is this button should just become disabled without changing text the index pattern field should be outlined in red there should be an error message below it that states do you have indices matching this pattern we couldn t fetch the mapping and what does couldn t fetch the mapping mean doesn t it make more sense to say couldn t find any matching indices ,0
+204512,15499159504.0,IssuesEvent,2021-03-11 07:34:09,alexbprofit/testcase,https://api.github.com/repos/alexbprofit/testcase,closed,Керування ролями,testcase,"Ціль:
+Як адміністратор я маю можливість змінювати ролі користувачів на сайті.
+Головна роль
+Адміністратор
+Передумови:
+Присутнє з'єднання з інтернетом. Користувач виконав вхід на сайт як адміністратор.
+Успішний сценарій:
+1. Система пропонує інтерфейс для перегляду користувачів та їх ролей ( у перегляді немає адміністратора).
+2. Адміністратор натискає змінити роль навпроти певного користувача.
+3. Система пропонує можливі варіанти ролей ( випадаючий список).
+4. Адміністратор вибирає роль.
+5. система фіксує зміни.",1.0,"Керування ролями - Ціль:
+Як адміністратор я маю можливість змінювати ролі користувачів на сайті.
+Головна роль
+Адміністратор
+Передумови:
+Присутнє з'єднання з інтернетом. Користувач виконав вхід на сайт як адміністратор.
+Успішний сценарій:
+1. Система пропонує інтерфейс для перегляду користувачів та їх ролей ( у перегляді немає адміністратора).
+2. Адміністратор натискає змінити роль навпроти певного користувача.
+3. Система пропонує можливі варіанти ролей ( випадаючий список).
+4. Адміністратор вибирає роль.
+5. система фіксує зміни.",0,керування ролями ціль як адміністратор я маю можливість змінювати ролі користувачів на сайті головна роль адміністратор передумови присутнє з єднання з інтернетом користувач виконав вхід на сайт як адміністратор успішний сценарій система пропонує інтерфейс для перегляду користувачів та їх ролей у перегляді немає адміністратора адміністратор натискає змінити роль навпроти певного користувача система пропонує можливі варіанти ролей випадаючий список адміністратор вибирає роль система фіксує зміни ,0
+58887,7188731234.0,IssuesEvent,2018-02-02 11:12:53,nerdalize/nerd,https://api.github.com/repos/nerdalize/nerd,opened,The Windows installer could add a shortcut to the Desktop,Design Required,"Currently, after installation the Nerd cli is only available through the user opening a terminal but it would probably be possible to add a shortcut to the users desktop that opens a terminal with the nerd cli open. ",1.0,"The Windows installer could add a shortcut to the Desktop - Currently, after installation the Nerd cli is only available through the user opening a terminal but it would probably be possible to add a shortcut to the users desktop that opens a terminal with the nerd cli open. ",0,the windows installer could add a shortcut to the desktop currently after installation the nerd cli is only available through the user opening a terminal but it would probably be possible to add a shortcut to the users desktop that opens a terminal with the nerd cli open ,0
+98580,20763171954.0,IssuesEvent,2022-03-15 18:00:48,rust-analyzer/rust-analyzer,https://api.github.com/repos/rust-analyzer/rust-analyzer,closed,[VSCode] Allow customisation of inlay hint styling,E-has-instructions A-vscode S-actionable A-inlay-hints,"I personally don't like the new smaller inlay hints, and I think it would be great to be able to customise it.
+
+It was introduced here: #6394
+Discussion of customising was held here: #6380 ",1.0,"[VSCode] Allow customisation of inlay hint styling - I personally don't like the new smaller inlay hints, and I think it would be great to be able to customise it.
+
+It was introduced here: #6394
+Discussion of customising was held here: #6380 ",0, allow customisation of inlay hint styling i personally don t like the new smaller inlay hints and i think it would be great to be able to customise it it was introduced here discussion of customising was held here ,0
+3297,12689281438.0,IssuesEvent,2020-06-21 04:58:18,diofant/diofant,https://api.github.com/repos/diofant/diofant,opened,Make .diff() call convention for high derivatives - consistent across the codebase,core help wanted maintainability polys,"Right now, `Poly.diff()` and `Expr.diff()` - differ (e.g. see sympy/sympy#19590). They shouldn't. Perhaps, Poly's syntax
+`Poly(x**2).diff((x, 2))` for higher derivatives must be chosen, as it seems simple.
+",True,"Make .diff() call convention for high derivatives - consistent across the codebase - Right now, `Poly.diff()` and `Expr.diff()` - differ (e.g. see sympy/sympy#19590). They shouldn't. Perhaps, Poly's syntax
+`Poly(x**2).diff((x, 2))` for higher derivatives must be chosen, as it seems simple.
+",0,make diff call convention for high derivatives consistent across the codebase right now poly diff and expr diff differ e g see sympy sympy they shouldn t perhaps poly s syntax poly x diff x for higher derivatives must be chosen as it seems simple ,0
+602268,18459435325.0,IssuesEvent,2021-10-15 21:36:46,DSpace/dspace-angular,https://api.github.com/repos/DSpace/dspace-angular,closed,Automated accessibility (a11y) testing via GitHub CI,accessibility high priority,"While Angular itself provides some basic, automated accessibility features (like [ARIA-enabled components](https://angular.io/features)), we should be performing some automated accessibility checks to ensure our new UI meets basic accessibility guidelines.
+
+Ideally, we should be striving for [WCAG 2.0 guidelines](https://www.w3.org/TR/WCAG20/). We should minimally aim for Level A compliance. But, I'd ask any accessibility experts here to chime in and recommend a compliance level. https://www.w3.org/WAI/WCAG20/quickref/ (_Note: some of the AA Level requirements are quite content specific, and may be hard to generalize for all DSpace sites_)
+
+Digging around, here's a list of some tools we may be able to use to automate some minimal accessibility testing:
+
+* Pa11y: http://pa11y.org/ (_This seems to be the most promising to me_)
+ * Enabling in Travis: http://cruft.io/posts/automated-accessibility-testing-node-travis-ci-pa11y/
+* Axe: https://www.axe-core.org/
+ * A basic example: https://www.deque.com/blog/accessibility-testing-axe-webdriverjs/
+* grunt-accessibility: https://www.npmjs.com/package/grunt-accessibility
+* HTML CodeSniffer: https://squizlabs.github.io/HTML_CodeSniffer/
+ * Looks to be usable to parse any HTML code for web standards (and used by Pa11y). Can also be called via CLI using PhantomJS or similar, see README at https://github.com/squizlabs/HTML_CodeSniffer/
+
+I'd encourage others to share tools/ideas/best practices on automating some level of a11y testing (even though I realize a11y testing cannot always be fully automated). I fully admit I'm not an a11y expert, so suggestions/improvements are welcome!
+",1.0,"Automated accessibility (a11y) testing via GitHub CI - While Angular itself provides some basic, automated accessibility features (like [ARIA-enabled components](https://angular.io/features)), we should be performing some automated accessibility checks to ensure our new UI meets basic accessibility guidelines.
+
+Ideally, we should be striving for [WCAG 2.0 guidelines](https://www.w3.org/TR/WCAG20/). We should minimally aim for Level A compliance. But, I'd ask any accessibility experts here to chime in and recommend a compliance level. https://www.w3.org/WAI/WCAG20/quickref/ (_Note: some of the AA Level requirements are quite content specific, and may be hard to generalize for all DSpace sites_)
+
+Digging around, here's a list of some tools we may be able to use to automate some minimal accessibility testing:
+
+* Pa11y: http://pa11y.org/ (_This seems to be the most promising to me_)
+ * Enabling in Travis: http://cruft.io/posts/automated-accessibility-testing-node-travis-ci-pa11y/
+* Axe: https://www.axe-core.org/
+ * A basic example: https://www.deque.com/blog/accessibility-testing-axe-webdriverjs/
+* grunt-accessibility: https://www.npmjs.com/package/grunt-accessibility
+* HTML CodeSniffer: https://squizlabs.github.io/HTML_CodeSniffer/
+ * Looks to be usable to parse any HTML code for web standards (and used by Pa11y). Can also be called via CLI using PhantomJS or similar, see README at https://github.com/squizlabs/HTML_CodeSniffer/
+
+I'd encourage others to share tools/ideas/best practices on automating some level of a11y testing (even though I realize a11y testing cannot always be fully automated). I fully admit I'm not an a11y expert, so suggestions/improvements are welcome!
+",0,automated accessibility testing via github ci while angular itself provides some basic automated accessibility features like we should be performing some automated accessibility checks to ensure our new ui meets basic accessibility guidelines ideally we should be striving for we should minimally aim for level a compliance but i d ask any accessibility experts here to chime in and recommend a compliance level note some of the aa level requirements are quite content specific and may be hard to generalize for all dspace sites digging around here s a list of some tools we may be able to use to automate some minimal accessibility testing this seems to be the most promising to me enabling in travis axe a basic example grunt accessibility html codesniffer looks to be usable to parse any html code for web standards and used by can also be called via cli using phantomjs or similar see readme at i d encourage others to share tools ideas best practices on automating some level of testing even though i realize testing cannot always be fully automated i fully admit i m not an expert so suggestions improvements are welcome ,0
+1158,30218052981.0,IssuesEvent,2023-07-05 17:02:35,metabase/metabase,https://api.github.com/repos/metabase/metabase,opened,Enable anonymizing archived users,Type:New Feature .Needs Triage Administration/Auth/SSO Administration/People,"**Is your feature request related to a problem? Please describe.**
+For compliance reasons, sometimes it is necessary to anonymize users (right to be forgotten).
+As of today, admins can change a user email or name, but they can't do it if the user props come from an IdP.
+
+**Describe the solution you'd like**
+An API endpoint to anonymize users.
+A button in the UI that calls this endpoint for a given archived user.
+
+**Describe alternatives you've considered**
+Manually anonymizing the user in the AppDB
+
+**How important is this feature to you?**
+There are workarounds, but it is important for compliance reasons
+",1.0,"Enable anonymizing archived users - **Is your feature request related to a problem? Please describe.**
+For compliance reasons, sometimes it is necessary to anonymize users (right to be forgotten).
+As of today, admins can change a user email or name, but they can't do it if the user props come from an IdP.
+
+**Describe the solution you'd like**
+An API endpoint to anonymize users.
+A button in the UI that calls this endpoint for a given archived user.
+
+**Describe alternatives you've considered**
+Manually anonymizing the user in the AppDB
+
+**How important is this feature to you?**
+There are workarounds, but it is important for compliance reasons
+",1,enable anonymizing archived users is your feature request related to a problem please describe for compliance reasons sometimes it is necessary to anonymize users right to be forgotten as of today admins can change a user email or name but they can t do it if the user props come from an idp describe the solution you d like an api endpoint to anonymize users a button in the ui that calls this endpoint for a given archived user describe alternatives you ve considered manually anonymizing the user in the appdb how important is this feature to you there are workarounds but it is important for compliance reasons ,1
+103070,16601971553.0,IssuesEvent,2021-06-01 20:51:26,samq-ghdemo/SEARCH-NCJIS-nibrs,https://api.github.com/repos/samq-ghdemo/SEARCH-NCJIS-nibrs,opened,CVE-2020-36186 (High) detected in multiple libraries,security vulnerability,"## CVE-2020-36186 - High Severity Vulnerability
+ Vulnerable Libraries - jackson-databind-2.9.8.jar, jackson-databind-2.9.5.jar, jackson-databind-2.8.10.jar, jackson-databind-2.9.6.jar
+
+
+jackson-databind-2.9.8.jar
+
+
General data-binding functionality for Jackson: works on core streaming API
Path to dependency file: SEARCH-NCJIS-nibrs/tools/nibrs-flatfile/pom.xml
+
Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar
Path to dependency file: SEARCH-NCJIS-nibrs/tools/nibrs-fbi-service/pom.xml
+
Path to vulnerable library: SEARCH-NCJIS-nibrs/tools/nibrs-fbi-service/target/nibrs-fbi-service-1.0.0/WEB-INF/lib/jackson-databind-2.8.10.jar,canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.10/jackson-databind-2.8.10.jar
Path to dependency file: SEARCH-NCJIS-nibrs/tools/nibrs-validation/pom.xml
+
Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,SEARCH-NCJIS-nibrs/web/nibrs-web/target/nibrs-web/WEB-INF/lib/jackson-databind-2.9.6.jar,canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar
+
+FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp.datasources.PerUserPoolDataSource.
+
+
Path to dependency file: SEARCH-NCJIS-nibrs/tools/nibrs-flatfile/pom.xml
+
Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar
Path to dependency file: SEARCH-NCJIS-nibrs/tools/nibrs-fbi-service/pom.xml
+
Path to vulnerable library: SEARCH-NCJIS-nibrs/tools/nibrs-fbi-service/target/nibrs-fbi-service-1.0.0/WEB-INF/lib/jackson-databind-2.8.10.jar,canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.10/jackson-databind-2.8.10.jar
Path to dependency file: SEARCH-NCJIS-nibrs/tools/nibrs-validation/pom.xml
+
Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,SEARCH-NCJIS-nibrs/web/nibrs-web/target/nibrs-web/WEB-INF/lib/jackson-databind-2.9.6.jar,canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar
+
+FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp.datasources.PerUserPoolDataSource.
+
+
+
+
+
+
+
+",0,cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file search ncjis nibrs tools nibrs summary report common pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter web release jar root library spring boot starter json release jar x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file search ncjis nibrs tools nibrs flatfile pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy tika parsers jar root library x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file search ncjis nibrs tools nibrs fbi service pom xml path to vulnerable library search ncjis nibrs tools nibrs fbi service target nibrs fbi service web inf lib jackson databind jar canner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file search ncjis nibrs tools nibrs validation pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar search ncjis nibrs web nibrs web target nibrs web web inf lib jackson databind jar canner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache tomcat dbcp dbcp datasources peruserpooldatasource publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org springframework boot spring boot starter web release org springframework boot spring boot starter json release com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree org apache tika tika parsers com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache tomcat dbcp dbcp datasources peruserpooldatasource vulnerabilityurl ,0
+231,4728971279.0,IssuesEvent,2016-10-18 17:23:48,training-center/mentoria,https://api.github.com/repos/training-center/mentoria,closed,Quero ser um mentor de desenvolvimento FullStack,css front end javascript mentor candidate monster others ruby this people are crazy work in progress,"Olá!
+
+Meu nome é Guilherme, tenho 24 anos e trabalho com desenvolvimento web profissionalmente há 7 anos, mas já estou vagando no mundo de desenvolvimento há uns 12 anos. Já passei por várias empresas, pequenas e grandes, essas foram as últimas: Gauge, Embraer, R7, Emprego Ligado e The Next Web (TwitterCounter).
+
+Já trabalhei com todo tipo de tecnologia possível, mas tenho especialidade em JavaScript (vanilla + jQuery, Backbonejs, React + Redux), CSS (SASS, BEM), Ruby (Rails), Linux (AWS, Linux Servers) entre mais algumas coisas por aí.
+
+Sou natural de São José dos Campos. Morei 3-4 anos em São Paulo antes de me mudar para Amsterdam - Países Baixos, onde eu trabalho e moro hoje!
+
+Quero ajudar esse tipo de pupilo (sendo em inglês ou não — preferência para pessoas que querem fazer isso em inglês):
+- Quero aprender mais sobre FrontEnd
+- Quero aprender mais sobre JavaScript
+- Quero aprender mais DevOps
+- Quero aprender mais Ruby on Rails
+- Quero dicas de carreira (Internacional ou não)
+
+E sim...estou de acordo com o código de conduta.
+
+**Twitter:** http://twitter.com/guiiipontes
+**Linkedin:** http://linkedin.com/in/guilhermepontes",1.0,"Quero ser um mentor de desenvolvimento FullStack - Olá!
+
+Meu nome é Guilherme, tenho 24 anos e trabalho com desenvolvimento web profissionalmente há 7 anos, mas já estou vagando no mundo de desenvolvimento há uns 12 anos. Já passei por várias empresas, pequenas e grandes, essas foram as últimas: Gauge, Embraer, R7, Emprego Ligado e The Next Web (TwitterCounter).
+
+Já trabalhei com todo tipo de tecnologia possível, mas tenho especialidade em JavaScript (vanilla + jQuery, Backbonejs, React + Redux), CSS (SASS, BEM), Ruby (Rails), Linux (AWS, Linux Servers) entre mais algumas coisas por aí.
+
+Sou natural de São José dos Campos. Morei 3-4 anos em São Paulo antes de me mudar para Amsterdam - Países Baixos, onde eu trabalho e moro hoje!
+
+Quero ajudar esse tipo de pupilo (sendo em inglês ou não — preferência para pessoas que querem fazer isso em inglês):
+- Quero aprender mais sobre FrontEnd
+- Quero aprender mais sobre JavaScript
+- Quero aprender mais DevOps
+- Quero aprender mais Ruby on Rails
+- Quero dicas de carreira (Internacional ou não)
+
+E sim...estou de acordo com o código de conduta.
+
+**Twitter:** http://twitter.com/guiiipontes
+**Linkedin:** http://linkedin.com/in/guilhermepontes",1,quero ser um mentor de desenvolvimento fullstack olá meu nome é guilherme tenho anos e trabalho com desenvolvimento web profissionalmente há anos mas já estou vagando no mundo de desenvolvimento há uns anos já passei por várias empresas pequenas e grandes essas foram as últimas gauge embraer emprego ligado e the next web twittercounter já trabalhei com todo tipo de tecnologia possível mas tenho especialidade em javascript vanilla jquery backbonejs react redux css sass bem ruby rails linux aws linux servers entre mais algumas coisas por aí sou natural de são josé dos campos morei anos em são paulo antes de me mudar para amsterdam países baixos onde eu trabalho e moro hoje quero ajudar esse tipo de pupilo sendo em inglês ou não — preferência para pessoas que querem fazer isso em inglês quero aprender mais sobre frontend quero aprender mais sobre javascript quero aprender mais devops quero aprender mais ruby on rails quero dicas de carreira internacional ou não e sim estou de acordo com o código de conduta twitter linkedin ,1
+626043,19783656512.0,IssuesEvent,2022-01-18 02:13:48,Baystation12/Baystation12,https://api.github.com/repos/Baystation12/Baystation12,closed,HasProximity() is broken,Priority: High ⚠,"
+#### Description of issue
+
+HasProximity() was broken by #10275 and never fixed.
+#### Difference between expected and actual behavior
+
+Most notably has affected vine entangling. A work around was proposed by #12572 and included in #13428. Still, if it isn't going to be fixed, it should be removed altogether so that whatever relies on it can be updated.
+",1.0,"HasProximity() is broken -
+#### Description of issue
+
+HasProximity() was broken by #10275 and never fixed.
+#### Difference between expected and actual behavior
+
+Most notably has affected vine entangling. A work around was proposed by #12572 and included in #13428. Still, if it isn't going to be fixed, it should be removed altogether so that whatever relies on it can be updated.
+",0,hasproximity is broken if a specific field doesn t apply remove it anything inside tags like these is a comment and will not be displayed in the final issue be careful not to write inside them joke or spammed issues can and will result in punishment put your answers on the blank lines below the headers the lines with four s don t edit them or delete them it s part of the formatting description of issue hasproximity was broken by and never fixed difference between expected and actual behavior most notably has affected vine entangling a work around was proposed by and included in still if it isn t going to be fixed it should be removed altogether so that whatever relies on it can be updated ,0
+594794,18054147561.0,IssuesEvent,2021-09-20 05:06:08,AY2122S1-CS2103-T14-3/tp,https://api.github.com/repos/AY2122S1-CS2103-T14-3/tp,opened,Command History,type.Story priority.Medium,"As a regular user, I can retrieve my old commands that I've previously used so that I can reuse them immediately to repeat a command",1.0,"Command History - As a regular user, I can retrieve my old commands that I've previously used so that I can reuse them immediately to repeat a command",0,command history as a regular user i can retrieve my old commands that i ve previously used so that i can reuse them immediately to repeat a command,0
+39811,8688868337.0,IssuesEvent,2018-12-03 17:07:53,CCBlueX/LiquidBounce1.8-Issues,https://api.github.com/repos/CCBlueX/LiquidBounce1.8-Issues,closed,improve aac 1.9.10 fly,Bypass Recode,"aac 1.9.10 fly ist relative schlecht im vergleich zu anderen clients (alpha centaurie, eazy usw.).",1.0,"improve aac 1.9.10 fly - aac 1.9.10 fly ist relative schlecht im vergleich zu anderen clients (alpha centaurie, eazy usw.).",0,improve aac fly aac fly ist relative schlecht im vergleich zu anderen clients alpha centaurie eazy usw ,0
+915,19441256996.0,IssuesEvent,2021-12-22 01:23:23,PostHog/posthog,https://api.github.com/repos/PostHog/posthog,closed,Retention line graph modal white screen,bug retention people insights,"## Bug description
+
+*Please describe.*
+- trying to click on a datapoint for retention line graph doesn't work. It whitescreens
+*If this affects the front-end, screenshots would be of great help.*
+
+## Expected behavior
+The modal should appear
+
+
+## How to reproduce
+
+1. Go to retention
+2. Make the chart type line graph
+3. click on a datapoint
+
+## Environment
+
+- [ ] PostHog Cloud
+- [ ] self-hosted PostHog (ClickHouse-based), version/commit: _please provide_
+- [ ] self-hosted PostHog (Postgres-based, legacy), version/commit: _please provide_
+
+## Additional context
+
+
+
+#### *Thank you* for your bug report – we love squashing them!
+",1.0,"Retention line graph modal white screen - ## Bug description
+
+*Please describe.*
+- trying to click on a datapoint for retention line graph doesn't work. It whitescreens
+*If this affects the front-end, screenshots would be of great help.*
+
+## Expected behavior
+The modal should appear
+
+
+## How to reproduce
+
+1. Go to retention
+2. Make the chart type line graph
+3. click on a datapoint
+
+## Environment
+
+- [ ] PostHog Cloud
+- [ ] self-hosted PostHog (ClickHouse-based), version/commit: _please provide_
+- [ ] self-hosted PostHog (Postgres-based, legacy), version/commit: _please provide_
+
+## Additional context
+
+
+
+#### *Thank you* for your bug report – we love squashing them!
+",1,retention line graph modal white screen bug description please describe trying to click on a datapoint for retention line graph doesn t work it whitescreens if this affects the front end screenshots would be of great help expected behavior the modal should appear how to reproduce go to retention make the chart type line graph click on a datapoint environment posthog cloud self hosted posthog clickhouse based version commit please provide self hosted posthog postgres based legacy version commit please provide additional context thank you for your bug report – we love squashing them ,1
+1073,26247855784.0,IssuesEvent,2023-01-05 16:37:52,PrjAdv/prjuptime,https://api.github.com/repos/PrjAdv/prjuptime,closed,🛑 VERY INUTIL PEOPLE is down,status very-inutil-people,"In [`cb96c66`](https://github.com/PrjAdv/prjuptime/commit/cb96c665c14178656267cf82ed7fce651fd83b6a
+), VERY INUTIL PEOPLE (https://veryinutilpeople.it) was **down**:
+- HTTP code: 0
+- Response time: 0 ms
+",1.0,"🛑 VERY INUTIL PEOPLE is down - In [`cb96c66`](https://github.com/PrjAdv/prjuptime/commit/cb96c665c14178656267cf82ed7fce651fd83b6a
+), VERY INUTIL PEOPLE (https://veryinutilpeople.it) was **down**:
+- HTTP code: 0
+- Response time: 0 ms
+",1,🛑 very inutil people is down in very inutil people was down http code response time ms ,1
+90579,26141177299.0,IssuesEvent,2022-12-29 18:42:24,NoahAGonzales/Visual-Physical-Converter,https://api.github.com/repos/NoahAGonzales/Visual-Physical-Converter,opened,Add support for transparency,building,"Transparent images don't appear without a base. In fact, they don't have a object created for them at all - all that is shown in the preview is the base. Add support for transparency so that transparent images are generated without a base in areas that are transparent.",1.0,"Add support for transparency - Transparent images don't appear without a base. In fact, they don't have a object created for them at all - all that is shown in the preview is the base. Add support for transparency so that transparent images are generated without a base in areas that are transparent.",0,add support for transparency transparent images don t appear without a base in fact they don t have a object created for them at all all that is shown in the preview is the base add support for transparency so that transparent images are generated without a base in areas that are transparent ,0
+389055,26795909053.0,IssuesEvent,2023-02-01 11:53:07,nuxt/nuxt,https://api.github.com/repos/nuxt/nuxt,closed,[Docs] add runtimeConfig / appConfig explainer,documentation 3.x,"At the moment, docs lack an explainer of the differences between the `appConfig` and `runtimeConfig` concepts, as they both allow to use environment variables.
+
+https://v3.nuxtjs.org/guide/going-further/runtime-config
+https://v3.nuxtjs.org/guide/directory-structure/app.config",1.0,"[Docs] add runtimeConfig / appConfig explainer - At the moment, docs lack an explainer of the differences between the `appConfig` and `runtimeConfig` concepts, as they both allow to use environment variables.
+
+https://v3.nuxtjs.org/guide/going-further/runtime-config
+https://v3.nuxtjs.org/guide/directory-structure/app.config",0, add runtimeconfig appconfig explainer at the moment docs lack an explainer of the differences between the appconfig and runtimeconfig concepts as they both allow to use environment variables ,0
+150041,13308217502.0,IssuesEvent,2020-08-26 00:18:09,JJguri/bestiapop,https://api.github.com/repos/JJguri/bestiapop,closed,Running with `-a download-and-convert-to-met` doesn't work,documentation enhancement,"Hi @jjguri, just found a minor oversight in the help info generated by running with the `-h` flag. It says that there's an option to download and convert to met format in one go with `-a download-and-convert-to-met`:
+
+```
+ -a {download-silo-file,convert-nc4-to-met,convert-nc4-to-csv,generate-met-file}, --action {download-silo-file,convert-nc4-to-met,convert-nc4-to-csv,generate-met-file}
+ The type of operation to want to perform: download-silo-file (it will only download a particular SILO file from S3 to your
+ local disk), convert-nc4-to-met (it will only convert a local or S3 file from NC4 format to MET), convert-nc4-to-csv(it will
+ only convert a local or S3 file from NC4 format to CSV), download-and-convert-to-met (combines the first two actions)
+```
+
+However if I try to use this option, I get the following error:
+
+```sh
+bestiapop.py: error: argument -a/--action: invalid choice: 'download-and-convert-to-met' (choose from 'download-silo-file', 'convert-nc4-to-met', 'convert-nc4-to-csv', 'generate-met-file')
+```",1.0,"Running with `-a download-and-convert-to-met` doesn't work - Hi @jjguri, just found a minor oversight in the help info generated by running with the `-h` flag. It says that there's an option to download and convert to met format in one go with `-a download-and-convert-to-met`:
+
+```
+ -a {download-silo-file,convert-nc4-to-met,convert-nc4-to-csv,generate-met-file}, --action {download-silo-file,convert-nc4-to-met,convert-nc4-to-csv,generate-met-file}
+ The type of operation to want to perform: download-silo-file (it will only download a particular SILO file from S3 to your
+ local disk), convert-nc4-to-met (it will only convert a local or S3 file from NC4 format to MET), convert-nc4-to-csv(it will
+ only convert a local or S3 file from NC4 format to CSV), download-and-convert-to-met (combines the first two actions)
+```
+
+However if I try to use this option, I get the following error:
+
+```sh
+bestiapop.py: error: argument -a/--action: invalid choice: 'download-and-convert-to-met' (choose from 'download-silo-file', 'convert-nc4-to-met', 'convert-nc4-to-csv', 'generate-met-file')
+```",0,running with a download and convert to met doesn t work hi jjguri just found a minor oversight in the help info generated by running with the h flag it says that there s an option to download and convert to met format in one go with a download and convert to met a download silo file convert to met convert to csv generate met file action download silo file convert to met convert to csv generate met file the type of operation to want to perform download silo file it will only download a particular silo file from to your local disk convert to met it will only convert a local or file from format to met convert to csv it will only convert a local or file from format to csv download and convert to met combines the first two actions however if i try to use this option i get the following error sh bestiapop py error argument a action invalid choice download and convert to met choose from download silo file convert to met convert to csv generate met file ,0
+314,5724320590.0,IssuesEvent,2017-04-20 14:18:21,PopulateTools/gobierto,https://api.github.com/repos/PopulateTools/gobierto,closed,Associate a Lotus Notes calendar to a GobiertoPerson,gobierto-people,"Follows #591
+
+Allow admins to introduce the URL of the Lotus Notes events calendar of a given GobiertoPerson, when that Calendar integration is enabled.
+
+1. Add a Configuration option in the Agendas tab of every person
+2. In the Configuration page, depending on the module enabled, add a form to introduce the URL of the calendar integration",1.0,"Associate a Lotus Notes calendar to a GobiertoPerson - Follows #591
+
+Allow admins to introduce the URL of the Lotus Notes events calendar of a given GobiertoPerson, when that Calendar integration is enabled.
+
+1. Add a Configuration option in the Agendas tab of every person
+2. In the Configuration page, depending on the module enabled, add a form to introduce the URL of the calendar integration",1,associate a lotus notes calendar to a gobiertoperson follows allow admins to introduce the url of the lotus notes events calendar of a given gobiertoperson when that calendar integration is enabled add a configuration option in the agendas tab of every person in the configuration page depending on the module enabled add a form to introduce the url of the calendar integration,1
+45733,13046186493.0,IssuesEvent,2020-07-29 08:34:59,ryanhsu828/hijack-main,https://api.github.com/repos/ryanhsu828/hijack-main,closed,What software do I use to create the iPhone App?,Priority-Medium Type-Defect auto-migrated,"```
+Documentation seems to be a bit light for the HiJack. What SDK is used to
+create the iPhone App?
+```
+
+Original issue reported on code.google.com by `bearlaug...@gmail.com` on 21 Jul 2013 at 1:42
+",1.0,"What software do I use to create the iPhone App? - ```
+Documentation seems to be a bit light for the HiJack. What SDK is used to
+create the iPhone App?
+```
+
+Original issue reported on code.google.com by `bearlaug...@gmail.com` on 21 Jul 2013 at 1:42
+",0,what software do i use to create the iphone app documentation seems to be a bit light for the hijack what sdk is used to create the iphone app original issue reported on code google com by bearlaug gmail com on jul at ,0
+956,21104662000.0,IssuesEvent,2022-04-04 17:31:13,abumunye/orderpin-bugs,https://api.github.com/repos/abumunye/orderpin-bugs,closed,No logo on newsletter email,Bug Cosmetic RestarauntPeoples,"## Describe the bug
+When a subscription is made and the email is received, there is no logo that can be seen but a replacement shotcut and the word logo
+
+## To Reproduce
+Steps to reproduce the behavior:
+1. Go to 'Home'
+2. Click on 'newsletter and subscribe'
+3. See error
+
+## Expected behavior
+There will be no logo
+
+## Screenshots
+
+
+
+## Desktop (please complete the following information):
+ - OS: [Windows]
+ - Browser [Chrome]
+
+
+",1.0,"No logo on newsletter email - ## Describe the bug
+When a subscription is made and the email is received, there is no logo that can be seen but a replacement shotcut and the word logo
+
+## To Reproduce
+Steps to reproduce the behavior:
+1. Go to 'Home'
+2. Click on 'newsletter and subscribe'
+3. See error
+
+## Expected behavior
+There will be no logo
+
+## Screenshots
+
+
+
+## Desktop (please complete the following information):
+ - OS: [Windows]
+ - Browser [Chrome]
+
+
+",1,no logo on newsletter email describe the bug when a subscription is made and the email is received there is no logo that can be seen but a replacement shotcut and the word logo to reproduce steps to reproduce the behavior go to home click on newsletter and subscribe see error expected behavior there will be no logo screenshots desktop please complete the following information os browser ,1
+412,7696145516.0,IssuesEvent,2018-05-18 14:29:03,SmartDataAnalytics/sda.tech,https://api.github.com/repos/SmartDataAnalytics/sda.tech,closed,FEEDBACK: List of all SDA members,People duplicate enhancement survey,People page should display all the members of the sda group by the default.,1.0,FEEDBACK: List of all SDA members - People page should display all the members of the sda group by the default.,1,feedback list of all sda members people page should display all the members of the sda group by the default ,1
+134942,18518217171.0,IssuesEvent,2021-10-20 12:35:38,Seagate/cortx-s3server,https://api.github.com/repos/Seagate/cortx-s3server,closed,CVE-2015-3255 (Medium) detected in policykit-1_0.105-20ubuntu0.18.04.5_amd64.deb,security vulnerability,"## CVE-2015-3255 - Medium Severity Vulnerability
+ Vulnerable Library - policykit-1_0.105-20ubuntu0.18.04.5_amd64.deb
+
+
framework for managing administrative policies and privileges
+
+The polkit_backend_action_pool_init function in polkitbackend/polkitbackendactionpool.c in PolicyKit (aka polkit) before 0.113 might allow local users to gain privileges via duplicate action IDs in action descriptions.
+
+
+
+The polkit_backend_action_pool_init function in polkitbackend/polkitbackendactionpool.c in PolicyKit (aka polkit) before 0.113 might allow local users to gain privileges via duplicate action IDs in action descriptions.
+
+
+
+
+
+
+
+",0,cve medium detected in policykit deb cve medium severity vulnerability vulnerable library policykit deb framework for managing administrative policies and privileges library home page a href dependency hierarchy x policykit deb vulnerable library found in head commit a href found in base branch main vulnerability details the polkit backend action pool init function in polkitbackend polkitbackendactionpool c in policykit aka polkit before might allow local users to gain privileges via duplicate action ids in action descriptions publish date url a href cvss score details base score metrics exploitability metrics attack vector n a attack complexity n a privileges required n a user interaction n a scope n a impact metrics confidentiality impact n a integrity impact n a availability impact n a for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree policykit isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails the polkit backend action pool init function in polkitbackend polkitbackendactionpool c in policykit aka polkit before might allow local users to gain privileges via duplicate action ids in action descriptions vulnerabilityurl ,0
+310099,9485775478.0,IssuesEvent,2019-04-22 11:38:33,FACK1/GBMQ,https://api.github.com/repos/FACK1/GBMQ,closed,Comparison,Front end Student T2h T4h priority-4 technical,"Page component to display the student result as a comparison between the first and second quiz.
+
+ - [ ] Create page component.
+ - [ ] Send request to the server `GET: /comparison`.
+ - [ ] Handle server response, it should return score, percentage, and rank for each quiz (first and second quizzes).
+ - [ ] Display the comparison result in the page.",1.0,"Comparison - Page component to display the student result as a comparison between the first and second quiz.
+
+ - [ ] Create page component.
+ - [ ] Send request to the server `GET: /comparison`.
+ - [ ] Handle server response, it should return score, percentage, and rank for each quiz (first and second quizzes).
+ - [ ] Display the comparison result in the page.",0,comparison page component to display the student result as a comparison between the first and second quiz create page component send request to the server get comparison handle server response it should return score percentage and rank for each quiz first and second quizzes display the comparison result in the page ,0
+1059,25262833632.0,IssuesEvent,2022-11-16 00:47:13,jongfeel/BookReview,https://api.github.com/repos/jongfeel/BookReview,closed,6부 3장 이혼하고 싶다면 이렇게 하라,2022 How to Win Friends & Influence People,"### 3장 이혼하고 싶다면 이렇게 하라
+
+많은 낭만적인 꿈들이 이혼이라는 바위 앞에 산산이 부서지는 가장 커다란 이유는 비판이다. 아무런 의미도 없고, 그저 다른 사람의 마음만 아프게 만드는 비판 말이다.
+
+---
+
+비결 3: 비판하지 마라.
+Don’t criticize.",1.0,"6부 3장 이혼하고 싶다면 이렇게 하라 - ### 3장 이혼하고 싶다면 이렇게 하라
+
+많은 낭만적인 꿈들이 이혼이라는 바위 앞에 산산이 부서지는 가장 커다란 이유는 비판이다. 아무런 의미도 없고, 그저 다른 사람의 마음만 아프게 만드는 비판 말이다.
+
+---
+
+비결 3: 비판하지 마라.
+Don’t criticize.",1, 이혼하고 싶다면 이렇게 하라 이혼하고 싶다면 이렇게 하라 많은 낭만적인 꿈들이 이혼이라는 바위 앞에 산산이 부서지는 가장 커다란 이유는 비판이다 아무런 의미도 없고 그저 다른 사람의 마음만 아프게 만드는 비판 말이다 비결 비판하지 마라 don’t criticize ,1
+193528,6885703193.0,IssuesEvent,2017-11-21 16:52:54,webcompat/web-bugs,https://api.github.com/repos/webcompat/web-bugs,closed,portal.mypearson.com - site is not usable,browser-firefox-mobile priority-normal,"
+
+
+
+**URL**: https://portal.mypearson.com/login
+
+**Browser / Version**: Firefox Mobile 59.0
+**Operating System**: Android 6.0.1
+**Tested Another Browser**: Yes
+
+**Problem type**: Site is not usable
+**Description**: Websites a pos and doesn't support ff mobile
+**Steps to Reproduce**:
+Went to do maths, couldn't without using Chrome.
+
+
+
+_From [webcompat.com](https://webcompat.com/) with ❤️_",1.0,"portal.mypearson.com - site is not usable -
+
+
+
+**URL**: https://portal.mypearson.com/login
+
+**Browser / Version**: Firefox Mobile 59.0
+**Operating System**: Android 6.0.1
+**Tested Another Browser**: Yes
+
+**Problem type**: Site is not usable
+**Description**: Websites a pos and doesn't support ff mobile
+**Steps to Reproduce**:
+Went to do maths, couldn't without using Chrome.
+
+
+
+_From [webcompat.com](https://webcompat.com/) with ❤️_",0,portal mypearson com site is not usable url browser version firefox mobile operating system android tested another browser yes problem type site is not usable description websites a pos and doesn t support ff mobile steps to reproduce went to do maths couldn t without using chrome from with ❤️ ,0
+17892,10149757169.0,IssuesEvent,2019-08-05 15:54:30,fecgov/openFEC,https://api.github.com/repos/fecgov/openFEC,opened,[Snyk: High] CRLF injection found in urllib3 [due: 9/5/2019],Security: high,"## Summary ##
+
+found in `requirements-ci.txt`:
+
+```
+High severity vulnerability found in urllib3
+Description: CRLF injection
+Info: https://snyk.io/vuln/SNYK-PYTHON-URLLIB3-174323
+Introduced through: codecov@2.0.9
+From: codecov@2.0.9 > requests@2.21.0 > urllib3@1.24.1
+```
+
+## Overview ##
+`urllib3` is an HTTP library with thread-safe connection pooling, file post, and more.
+
+Affected versions of this package are vulnerable to CRLF injection. Attacker who has the control of the requesting address parameter, could manipulate an HTTP header and attack an internal service.
+
+## Remediation ##
+Upgrade `urllib3` to version 1.24.3 or higher.
+",True,"[Snyk: High] CRLF injection found in urllib3 [due: 9/5/2019] - ## Summary ##
+
+found in `requirements-ci.txt`:
+
+```
+High severity vulnerability found in urllib3
+Description: CRLF injection
+Info: https://snyk.io/vuln/SNYK-PYTHON-URLLIB3-174323
+Introduced through: codecov@2.0.9
+From: codecov@2.0.9 > requests@2.21.0 > urllib3@1.24.1
+```
+
+## Overview ##
+`urllib3` is an HTTP library with thread-safe connection pooling, file post, and more.
+
+Affected versions of this package are vulnerable to CRLF injection. Attacker who has the control of the requesting address parameter, could manipulate an HTTP header and attack an internal service.
+
+## Remediation ##
+Upgrade `urllib3` to version 1.24.3 or higher.
+",0, crlf injection found in summary found in requirements ci txt high severity vulnerability found in description crlf injection info introduced through codecov from codecov requests overview is an http library with thread safe connection pooling file post and more affected versions of this package are vulnerable to crlf injection attacker who has the control of the requesting address parameter could manipulate an http header and attack an internal service remediation upgrade to version or higher ,0
+9740,11795013514.0,IssuesEvent,2020-03-18 08:03:44,jenkinsci/configuration-as-code-plugin,https://api.github.com/repos/jenkinsci/configuration-as-code-plugin,closed,Usuck,plugin-compatibility,"[jenkins-jira]: https://issues.jenkins-ci.org
+[dashboard]: https://issues.jenkins.io/secure/Dashboard.jspa?selectPageId=18341
+[contributing]: ../blob/master/docs/CONTRIBUTING.md
+[compatibility]: ../blob/master/docs/COMPATIBILITY.md
+
+### Your checklist for this issue
+
+🚨 Please review the [guidelines for contributing][contributing] to this repository.
+💡 To better understand plugin compatibility issues, you can [read more here][compatibility]
+
+
+
+- [ ] Create an issue on [issues.jenkins-ci.org][jenkins-jira], set the component to the plugin you are reporting it against
+
+- [ ] Before creating an issue on [Jenkins JIRA][jenkins-jira], check for [an existing one via dashboard][dashboard]
+
+- [x] Link to [Jenkins JIRA issue][jenkins-jira]
+
+- [x] Ensure [Jenkins JIRA issue][jenkins-jira] has the label `jcasc-compatibility`
+
+- [x] Link to plugin's GitHub repository
+
+u
+
+### Description
+
+Please describe your issue here.
+I'm being stopped",True,"Usuck - [jenkins-jira]: https://issues.jenkins-ci.org
+[dashboard]: https://issues.jenkins.io/secure/Dashboard.jspa?selectPageId=18341
+[contributing]: ../blob/master/docs/CONTRIBUTING.md
+[compatibility]: ../blob/master/docs/COMPATIBILITY.md
+
+### Your checklist for this issue
+
+🚨 Please review the [guidelines for contributing][contributing] to this repository.
+💡 To better understand plugin compatibility issues, you can [read more here][compatibility]
+
+
+
+- [ ] Create an issue on [issues.jenkins-ci.org][jenkins-jira], set the component to the plugin you are reporting it against
+
+- [ ] Before creating an issue on [Jenkins JIRA][jenkins-jira], check for [an existing one via dashboard][dashboard]
+
+- [x] Link to [Jenkins JIRA issue][jenkins-jira]
+
+- [x] Ensure [Jenkins JIRA issue][jenkins-jira] has the label `jcasc-compatibility`
+
+- [x] Link to plugin's GitHub repository
+
+u
+
+### Description
+
+Please describe your issue here.
+I'm being stopped",0,usuck blob master docs contributing md blob master docs compatibility md your checklist for this issue 🚨 please review the to this repository 💡 to better understand plugin compatibility issues you can here is a link to get you started with creating the issue over at jenkins jira create an issue on set the component to the plugin you are reporting it against before creating an issue on check for link to ensure has the label jcasc compatibility link to plugin s github repository put an x into the to show you have filled the information below describe your issue below u description please describe your issue here i m being stopped,0
+844,15820366995.0,IssuesEvent,2021-04-05 18:51:57,restincode/restincode,https://api.github.com/repos/restincode/restincode,opened,Chris Scott (Cyb3r_Assassin),Add Person Needs Review People,"Please fill out as much information as you can. No fields are required, but the more you can provide the better.
+
+**General Info**
+* First name: Chris
+* Last name: Scott
+* Handle: Cyb3r_Assassin
+* Birth Year:
+* Death Year: 2021
+* Link to Obituary:
+* Group Affiliations:
+* URL to main photo (or attach to Issue):
+* Description of person and/or activities:
+* Facebook memorial group URL:
+
+**Social Media Links**
+* Twitter:
+* Github:
+* LinkedIn:
+* Facebook:
+* Other:
+
+**Contributions**
+Repeat as many times as needed
+
+* Project name:
+* Project URL:
+* Project Description:
+
+**Photo Gallery**
+* URL(s) to additional photos:
+
+https://twitter.com/brentwdesign/status/1379081573409173508
+",1.0,"Chris Scott (Cyb3r_Assassin) - Please fill out as much information as you can. No fields are required, but the more you can provide the better.
+
+**General Info**
+* First name: Chris
+* Last name: Scott
+* Handle: Cyb3r_Assassin
+* Birth Year:
+* Death Year: 2021
+* Link to Obituary:
+* Group Affiliations:
+* URL to main photo (or attach to Issue):
+* Description of person and/or activities:
+* Facebook memorial group URL:
+
+**Social Media Links**
+* Twitter:
+* Github:
+* LinkedIn:
+* Facebook:
+* Other:
+
+**Contributions**
+Repeat as many times as needed
+
+* Project name:
+* Project URL:
+* Project Description:
+
+**Photo Gallery**
+* URL(s) to additional photos:
+
+https://twitter.com/brentwdesign/status/1379081573409173508
+",1,chris scott assassin please fill out as much information as you can no fields are required but the more you can provide the better general info first name chris last name scott handle assassin birth year death year link to obituary group affiliations url to main photo or attach to issue description of person and or activities facebook memorial group url social media links twitter github linkedin facebook other contributions repeat as many times as needed project name project url project description photo gallery url s to additional photos ,1
+318933,9725397158.0,IssuesEvent,2019-05-30 08:35:28,StrangeLoopGames/EcoIssues,https://api.github.com/repos/StrangeLoopGames/EcoIssues,closed,[7.6.3] Blask furnace exhaust pipe graphical bug,Low Priority,"See video, top of Exhaust of blast furnace is have a graphical error.
+https://youtu.be/GZGgblM1XTo",1.0,"[7.6.3] Blask furnace exhaust pipe graphical bug - See video, top of Exhaust of blast furnace is have a graphical error.
+https://youtu.be/GZGgblM1XTo",0, blask furnace exhaust pipe graphical bug see video top of exhaust of blast furnace is have a graphical error ,0
+264772,23138010895.0,IssuesEvent,2022-07-28 15:45:41,nicolargo/glances,https://api.github.com/repos/nicolargo/glances,closed,Config to disable all plugins by default (or enable an exclusive list,enhancement needs test,"Thank you for making glances! It's awesome, and I've been looking into using it in combination with Home Assistant to monitor system stats on my network.
+
+**Is your feature request related to a problem? Please describe.**
+
+When deploying glances as a JSON API only, I'd like to be able to limit the stats exposed -- monitoring simple stats like CPU and memory are very useful to me, but I would prefer to not expose things like the process list and software versions to the network, since they could aid automated vulnerability finders.
+
+**Describe the solution you'd like**
+
+It'd be really useful to either be able to disable all plugins by default. For example, support for an argument like `--disable-plugin=*` or `--disable-plugin=all` would work well for this. Alternatively, it would be useful to be able to specify a list of plugins like `--plugins=cpu,mem` and have those be the only plugins loaded.
+
+**Describe alternatives you've considered**
+
+Currently, the only way I've found to accomplish this is listing the entire set of available plugins via:
+
+```
+--disable-plugin=alert,amps,cloud,connections,core,cpu,diskio,docker,folders,fs,gpu,help,ip,irq,load,mem,memswap,network,now,percpu,ports,processcount,processlist,psutilversion,quicklook,raid,sensors,smart,system,uptime,wifi
+```
+
+This isn't a total solution, because it's possible a future glances update will introduce a new plugin which is default-on.",1.0,"Config to disable all plugins by default (or enable an exclusive list - Thank you for making glances! It's awesome, and I've been looking into using it in combination with Home Assistant to monitor system stats on my network.
+
+**Is your feature request related to a problem? Please describe.**
+
+When deploying glances as a JSON API only, I'd like to be able to limit the stats exposed -- monitoring simple stats like CPU and memory are very useful to me, but I would prefer to not expose things like the process list and software versions to the network, since they could aid automated vulnerability finders.
+
+**Describe the solution you'd like**
+
+It'd be really useful to either be able to disable all plugins by default. For example, support for an argument like `--disable-plugin=*` or `--disable-plugin=all` would work well for this. Alternatively, it would be useful to be able to specify a list of plugins like `--plugins=cpu,mem` and have those be the only plugins loaded.
+
+**Describe alternatives you've considered**
+
+Currently, the only way I've found to accomplish this is listing the entire set of available plugins via:
+
+```
+--disable-plugin=alert,amps,cloud,connections,core,cpu,diskio,docker,folders,fs,gpu,help,ip,irq,load,mem,memswap,network,now,percpu,ports,processcount,processlist,psutilversion,quicklook,raid,sensors,smart,system,uptime,wifi
+```
+
+This isn't a total solution, because it's possible a future glances update will introduce a new plugin which is default-on.",0,config to disable all plugins by default or enable an exclusive list thank you for making glances it s awesome and i ve been looking into using it in combination with home assistant to monitor system stats on my network is your feature request related to a problem please describe when deploying glances as a json api only i d like to be able to limit the stats exposed monitoring simple stats like cpu and memory are very useful to me but i would prefer to not expose things like the process list and software versions to the network since they could aid automated vulnerability finders describe the solution you d like it d be really useful to either be able to disable all plugins by default for example support for an argument like disable plugin or disable plugin all would work well for this alternatively it would be useful to be able to specify a list of plugins like plugins cpu mem and have those be the only plugins loaded describe alternatives you ve considered currently the only way i ve found to accomplish this is listing the entire set of available plugins via disable plugin alert amps cloud connections core cpu diskio docker folders fs gpu help ip irq load mem memswap network now percpu ports processcount processlist psutilversion quicklook raid sensors smart system uptime wifi this isn t a total solution because it s possible a future glances update will introduce a new plugin which is default on ,0
+262034,22787500350.0,IssuesEvent,2022-07-09 13:38:26,WTA-Stats/application,https://api.github.com/repos/WTA-Stats/application,closed,[App-115]: add Testing Multi Env Encrypted Secure Credentials,Configuration Testing Credentials,add Testing Multi Env Encrypted Secure Credentials,1.0,[App-115]: add Testing Multi Env Encrypted Secure Credentials - add Testing Multi Env Encrypted Secure Credentials,0, add testing multi env encrypted secure credentials add testing multi env encrypted secure credentials,0
+25421,11171418689.0,IssuesEvent,2019-12-28 19:35:09,eldorplus/conception-website,https://api.github.com/repos/eldorplus/conception-website,opened,WS-2019-0183 (Medium) detected in lodash.defaultsdeep-4.3.2.tgz,security vulnerability,"## WS-2019-0183 - Medium Severity Vulnerability
+ Vulnerable Library - lodash.defaultsdeep-4.3.2.tgz
+
+
The lodash method `_.defaultsDeep` exported as a module.
+
+lodash.defaultsdeep before 4.6.1 is vulnerable to prototype pollution. The function mergeWith() may allow a malicious user to modify the prototype of Object via {constructor: {prototype: {...}}} causing the addition or modification of an existing property that will exist on all objects.
+
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",True,"WS-2019-0183 (Medium) detected in lodash.defaultsdeep-4.3.2.tgz - ## WS-2019-0183 - Medium Severity Vulnerability
+ Vulnerable Library - lodash.defaultsdeep-4.3.2.tgz
+
+
The lodash method `_.defaultsDeep` exported as a module.
+
+lodash.defaultsdeep before 4.6.1 is vulnerable to prototype pollution. The function mergeWith() may allow a malicious user to modify the prototype of Object via {constructor: {prototype: {...}}} causing the addition or modification of an existing property that will exist on all objects.
+
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",0,ws medium detected in lodash defaultsdeep tgz ws medium severity vulnerability vulnerable library lodash defaultsdeep tgz the lodash method defaultsdeep exported as a module library home page a href path to dependency file tmp ws scm conception website package json path to vulnerable library tmp ws scm conception website node modules lodash defaultsdeep package json dependency hierarchy nightwatch tgz root library x lodash defaultsdeep tgz vulnerable library found in head commit a href vulnerability details lodash defaultsdeep before is vulnerable to prototype pollution the function mergewith may allow a malicious user to modify the prototype of object via constructor prototype causing the addition or modification of an existing property that will exist on all objects publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource ,0
+622462,19636116609.0,IssuesEvent,2022-01-08 10:02:37,rancher/k3d,https://api.github.com/repos/rancher/k3d,closed,[Feature] Volume Shortcuts,enhancement priority/high scope/cli scope/package component/main difficulty/low,"Introducing two things:
+- automatic volume creation for `k3d-` prefixed volumes using the `--volume` flag -> k3d will handle creation + deletion
+- magic mount destination keywords: e.g. `--volume /some/path:k3s-storage` will internally be handled as `/some/path:/var/lib/rancher/k3s/storage`
+
+_Originally posted by @iwilltry42 in https://github.com/rancher/k3d/issues/865#issuecomment-990932170_
+
+EDIT: actually, bullet point 1 is already a docker feature: if the source part of the volume mount is not a path on the local filesystem and not an existing named volume, docker will create it (k3d will still print a warning though).",1.0,"[Feature] Volume Shortcuts - Introducing two things:
+- automatic volume creation for `k3d-` prefixed volumes using the `--volume` flag -> k3d will handle creation + deletion
+- magic mount destination keywords: e.g. `--volume /some/path:k3s-storage` will internally be handled as `/some/path:/var/lib/rancher/k3s/storage`
+
+_Originally posted by @iwilltry42 in https://github.com/rancher/k3d/issues/865#issuecomment-990932170_
+
+EDIT: actually, bullet point 1 is already a docker feature: if the source part of the volume mount is not a path on the local filesystem and not an existing named volume, docker will create it (k3d will still print a warning though).",0, volume shortcuts introducing two things automatic volume creation for prefixed volumes using the volume flag will handle creation deletion magic mount destination keywords e g volume some path storage will internally be handled as some path var lib rancher storage originally posted by in edit actually bullet point is already a docker feature if the source part of the volume mount is not a path on the local filesystem and not an existing named volume docker will create it will still print a warning though ,0
+313149,23458759212.0,IssuesEvent,2022-08-16 11:17:48,commercetools/merchant-center-application-kit,https://api.github.com/repos/commercetools/merchant-center-application-kit,closed,[docs] Request deployment example for [Azure (Docker =>Azure Container Registry =>Azure App Service)],✍️ Type: Documentation 📞 Type: Customer request ⌛ Status: Awaiting for feedback,"**Where would you like to deploy your Custom Application?**
+List one of the cloud providers that you would be interested to use, that are not listed in the examples.
+**_### [Azure (Docker =>Azure Container Registry =>Azure App Service)]_**
+
+**Additional context**
+Add any other context or information about the requested example.
+
+1. To migrate custom application from Project level to Organization level
+ a. Followed the documentation and successfully updated the configuration and the development changes
+ b. For deployment overridden the output directory with: public.
+2. We are deploying the application using [Azure (Docker =>Azure Container Registry =>Azure App Service)]
+3. For docker image we are referring to base image ""FROM eu.gcr.io/ct-images/mc-http-server:v16.15.8""
+
+**Error details:**
+
+1. Error: Could not find ""dist/assets"" folder. Did you run `yarn build` before starting the server?
+
+2. Error: should NOT have additional properties: oAuthScopes
+ at validateConfig (/app/node_modules/@commercetools-frontend/application-config/build/validate-config.js:53:11)
+ at processConfig (/app/node_modules/@commercetools-frontend/application-config/build/process-config.js:64:31)
+ at compileHtml (/app/node_modules/@commercetools-frontend/mc-html-template/build/compile-html.js:52:29)
+ at start (/app/bin/start-mc.js:88:26)
+ at Object. (/app/bin/start-mc.js:110:1)
+ at Module._compile (internal/modules/cjs/loader.js:1256:30)
+ at Object.Module._extensions..js (internal/modules/cjs/loader.js:1277:10)
+ at Module.load (internal/modules/cjs/loader.js:1105:32)
+ at Function.Module._load (internal/modules/cjs/loader.js:967:14)
+ at Function.executeUserEntryPoint [as runMain] (internal/modules/run_main.js:60:12)
+
+
+
+
+**Packages Used for Migration/Development:**
+
+ ""@commercetools-frontend/actions-global"": ""^21.3.4"",
+ ""@commercetools-frontend/application-components"": ""^21.3.5"",
+ ""@commercetools-frontend/application-shell"": ""^21.3.5"",
+ ""@commercetools-frontend/assets"": ""^21.0.0"",
+ ""@commercetools-frontend/constants"": ""^21.3.4"",
+ ""@commercetools-frontend/i18n"": ""^21.3.4"",
+ ""@commercetools-frontend/permissions"": ""^21.3.4"",
+ ""@commercetools-frontend/ui-kit"": ""^14.0.6"",
+ ""@commercetools-uikit/flat-button"": ""^14.0.6"",
+ ""@commercetools-uikit/icons"": ""^14.0.1"",
+ ""@commercetools-uikit/link"": ""^14.0.6"",
+ ""@commercetools-uikit/loading-spinner"": ""^14.0.6"",
+ ""@commercetools-uikit/spacings"": ""^14.0.6"",
+ ""@commercetools-uikit/text"": ""^14.0.1"",
+ ""@commercetools-frontend/jest-preset-mc-app"": ""^21.3.4"",
+ ""@commercetools-frontend/mc-dev-authentication"": ""^21.0.0"",
+ ""@commercetools-frontend/mc-scripts"": ""^21.3.4"",
+",1.0,"[docs] Request deployment example for [Azure (Docker =>Azure Container Registry =>Azure App Service)] - **Where would you like to deploy your Custom Application?**
+List one of the cloud providers that you would be interested to use, that are not listed in the examples.
+**_### [Azure (Docker =>Azure Container Registry =>Azure App Service)]_**
+
+**Additional context**
+Add any other context or information about the requested example.
+
+1. To migrate custom application from Project level to Organization level
+ a. Followed the documentation and successfully updated the configuration and the development changes
+ b. For deployment overridden the output directory with: public.
+2. We are deploying the application using [Azure (Docker =>Azure Container Registry =>Azure App Service)]
+3. For docker image we are referring to base image ""FROM eu.gcr.io/ct-images/mc-http-server:v16.15.8""
+
+**Error details:**
+
+1. Error: Could not find ""dist/assets"" folder. Did you run `yarn build` before starting the server?
+
+2. Error: should NOT have additional properties: oAuthScopes
+ at validateConfig (/app/node_modules/@commercetools-frontend/application-config/build/validate-config.js:53:11)
+ at processConfig (/app/node_modules/@commercetools-frontend/application-config/build/process-config.js:64:31)
+ at compileHtml (/app/node_modules/@commercetools-frontend/mc-html-template/build/compile-html.js:52:29)
+ at start (/app/bin/start-mc.js:88:26)
+ at Object. (/app/bin/start-mc.js:110:1)
+ at Module._compile (internal/modules/cjs/loader.js:1256:30)
+ at Object.Module._extensions..js (internal/modules/cjs/loader.js:1277:10)
+ at Module.load (internal/modules/cjs/loader.js:1105:32)
+ at Function.Module._load (internal/modules/cjs/loader.js:967:14)
+ at Function.executeUserEntryPoint [as runMain] (internal/modules/run_main.js:60:12)
+
+
+
+
+**Packages Used for Migration/Development:**
+
+ ""@commercetools-frontend/actions-global"": ""^21.3.4"",
+ ""@commercetools-frontend/application-components"": ""^21.3.5"",
+ ""@commercetools-frontend/application-shell"": ""^21.3.5"",
+ ""@commercetools-frontend/assets"": ""^21.0.0"",
+ ""@commercetools-frontend/constants"": ""^21.3.4"",
+ ""@commercetools-frontend/i18n"": ""^21.3.4"",
+ ""@commercetools-frontend/permissions"": ""^21.3.4"",
+ ""@commercetools-frontend/ui-kit"": ""^14.0.6"",
+ ""@commercetools-uikit/flat-button"": ""^14.0.6"",
+ ""@commercetools-uikit/icons"": ""^14.0.1"",
+ ""@commercetools-uikit/link"": ""^14.0.6"",
+ ""@commercetools-uikit/loading-spinner"": ""^14.0.6"",
+ ""@commercetools-uikit/spacings"": ""^14.0.6"",
+ ""@commercetools-uikit/text"": ""^14.0.1"",
+ ""@commercetools-frontend/jest-preset-mc-app"": ""^21.3.4"",
+ ""@commercetools-frontend/mc-dev-authentication"": ""^21.0.0"",
+ ""@commercetools-frontend/mc-scripts"": ""^21.3.4"",
+",0, request deployment example for where would you like to deploy your custom application list one of the cloud providers that you would be interested to use that are not listed in the examples additional context add any other context or information about the requested example to migrate custom application from project level to organization level a followed the documentation and successfully updated the configuration and the development changes b for deployment overridden the output directory with public we are deploying the application using for docker image we are referring to base image from eu gcr io ct images mc http server error details error could not find dist assets folder did you run yarn build before starting the server error should not have additional properties oauthscopes at validateconfig app node modules commercetools frontend application config build validate config js at processconfig app node modules commercetools frontend application config build process config js at compilehtml app node modules commercetools frontend mc html template build compile html js at start app bin start mc js at object app bin start mc js at module compile internal modules cjs loader js at object module extensions js internal modules cjs loader js at module load internal modules cjs loader js at function module load internal modules cjs loader js at function executeuserentrypoint internal modules run main js packages used for migration development commercetools frontend actions global commercetools frontend application components commercetools frontend application shell commercetools frontend assets commercetools frontend constants commercetools frontend commercetools frontend permissions commercetools frontend ui kit commercetools uikit flat button commercetools uikit icons commercetools uikit link commercetools uikit loading spinner commercetools uikit spacings commercetools uikit text commercetools frontend jest preset mc app commercetools frontend mc dev authentication commercetools frontend mc scripts ,0
+252025,27225917774.0,IssuesEvent,2023-02-21 09:42:09,uriel-naor/Test,https://api.github.com/repos/uriel-naor/Test,closed,ejs-locals-1.0.2.tgz: 1 vulnerabilities (highest severity is: 9.8) - autoclosed,Mend: dependency security vulnerability," Vulnerable Library - ejs-locals-1.0.2.tgz
+
+
+
Path to dependency file: /package.json
+
Path to vulnerable library: /node_modules/ejs-locals/node_modules/ejs/package.json
+
+ #### Mend has checked all newer package trees, and you are on the least vulnerable package!
+
+ #### Please note: There might be a version that explicitly solves one or more of the vulnerabilities listed below, but we do not recommend it. For more info about the optional fixes, check the section “Details” below.
+
+## Vulnerabilities
+
+| CVE | Severity | CVSS | Dependency | Type | Fixed in (ejs-locals version) | Fix PR available |
+| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
+| [CVE-2017-1000228](https://www.mend.io/vulnerability-database/CVE-2017-1000228) | High | 9.8 | ejs-0.8.8.tgz | Transitive | N/A* | ❌ |
+
*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section ""Details"" below to see if there is a version of transitive dependency where vulnerability is fixed.
+
+ #### Mend has checked all newer package trees, and you are on the least vulnerable package!
+
+ #### Please note: There might be a version that explicitly solves one or more of the vulnerabilities listed below, but we do not recommend it. For more info about the optional fixes, check the section “Details” below.
+
+## Vulnerabilities
+
+| CVE | Severity | CVSS | Dependency | Type | Fixed in (ejs-locals version) | Fix PR available |
+| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
+| [CVE-2017-1000228](https://www.mend.io/vulnerability-database/CVE-2017-1000228) | High | 9.8 | ejs-0.8.8.tgz | Transitive | N/A* | ❌ |
+
*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section ""Details"" below to see if there is a version of transitive dependency where vulnerability is fixed.
+
+
+
+
+
+",0,ejs locals tgz vulnerabilities highest severity is autoclosed vulnerable library ejs locals tgz path to dependency file package json path to vulnerable library node modules ejs locals node modules ejs package json found in head commit a href mend has checked all newer package trees and you are on the least vulnerable package please note there might be a version that explicitly solves one or more of the vulnerabilities listed below but we do not recommend it for more info about the optional fixes check the section “details” below vulnerabilities cve severity cvss dependency type fixed in ejs locals version fix pr available high ejs tgz transitive n a for some transitive vulnerabilities there is no version of direct dependency with a fix check the section details below to see if there is a version of transitive dependency where vulnerability is fixed details cve vulnerable library ejs tgz embedded javascript templates library home page a href path to dependency file package json path to vulnerable library node modules ejs locals node modules ejs package json dependency hierarchy ejs locals tgz root library x ejs tgz vulnerable library found in head commit a href found in base branch main vulnerability details nodejs ejs versions older than is vulnerable to remote code execution due to weak input validation in ejs renderfile function publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution ,0
+601,10605333551.0,IssuesEvent,2019-10-10 20:13:45,microsoftgraph/microsoft-graph-toolkit,https://api.github.com/repos/microsoftgraph/microsoft-graph-toolkit,closed,[BUG] attendee images in mgt-agenda are misaligned ,Area: Components Component: agenda Component: people Hacktoberfest In-PR :shipit: Needs: Triage :mag: bug good first issue in progress,"
+
+
+**Describe the bug**
+The attendee photos in the mgt-agenda component are not aligned when some attendees don't have an image and some do.
+
+**To Reproduce**
+Steps to reproduce the behavior:
+1. Use the `mgt-agenda` and ensure there are events with a combination of attendees that have and don't have profile image
+
+
+**Expected behavior**
+All images should be aligned
+
+**Screenshots**
+
+
+
+**Environment (please complete the following information):**
+ - OS: [e.g. iOS] Windows
+ - Browser [e.g. edge, chrome, safari] edge beta
+ - Framework [e.g. react, vue, none] none
+ - Context [e.g. Microsoft Teams, SharePoint, Office Add-ins, Web] web app
+ - Version [e.g. 0.1] 1.0.0
+
+",1.0,"[BUG] attendee images in mgt-agenda are misaligned -
+
+
+**Describe the bug**
+The attendee photos in the mgt-agenda component are not aligned when some attendees don't have an image and some do.
+
+**To Reproduce**
+Steps to reproduce the behavior:
+1. Use the `mgt-agenda` and ensure there are events with a combination of attendees that have and don't have profile image
+
+
+**Expected behavior**
+All images should be aligned
+
+**Screenshots**
+
+
+
+**Environment (please complete the following information):**
+ - OS: [e.g. iOS] Windows
+ - Browser [e.g. edge, chrome, safari] edge beta
+ - Framework [e.g. react, vue, none] none
+ - Context [e.g. Microsoft Teams, SharePoint, Office Add-ins, Web] web app
+ - Version [e.g. 0.1] 1.0.0
+
+",1, attendee images in mgt agenda are misaligned describe the bug the attendee photos in the mgt agenda component are not aligned when some attendees don t have an image and some do to reproduce steps to reproduce the behavior use the mgt agenda and ensure there are events with a combination of attendees that have and don t have profile image expected behavior all images should be aligned screenshots environment please complete the following information os windows browser edge beta framework none context web app version ,1
+46382,5806264794.0,IssuesEvent,2017-05-04 01:47:54,NSW-OEH-EMS-KST/grid-garage-3,https://api.github.com/repos/NSW-OEH-EMS-KST/grid-garage-3,closed,raster - clip,requires testing,"When you specify a feature to clip with but DO NOT tick the 'Use Features for clipping (optional) ' the tools still clips to the feature i.e. it behaves as if the box IS ticked (see screen grab). See latest run of tool in 'Results' window.
+",1.0,"raster - clip - When you specify a feature to clip with but DO NOT tick the 'Use Features for clipping (optional) ' the tools still clips to the feature i.e. it behaves as if the box IS ticked (see screen grab). See latest run of tool in 'Results' window.
+",0,raster clip when you specify a feature to clip with but do not tick the use features for clipping optional the tools still clips to the feature i e it behaves as if the box is ticked see screen grab see latest run of tool in results window ,0
+298973,9204035594.0,IssuesEvent,2019-03-08 05:34:53,webcompat/web-bugs,https://api.github.com/repos/webcompat/web-bugs,closed,www.google.com - video or audio doesn't play,browser-firefox-mobile priority-critical,"
+
+
+
+**URL**: https://www.google.com/search?q=beeg&ie=utf-8&oe=utf-8&client=firefox-b-m
+
+**Browser / Version**: Firefox Mobile 66.0
+**Operating System**: Android 8.0.0
+**Tested Another Browser**: No
+
+**Problem type**: Video or audio doesn't play
+**Description**: DON'T open web
+**Steps to Reproduce**:
+
+
+
+Browser Configuration
+
+
mixed active content blocked: false
image.mem.shared: true
buildID: 20190304101322
tracking content blocked: false
gfx.webrender.blob-images: true
hasTouchScreen: true
mixed passive content blocked: false
gfx.webrender.enabled: false
gfx.webrender.all: false
channel: beta
+
+
+
+
+_From [webcompat.com](https://webcompat.com/) with ❤️_",1.0,"www.google.com - video or audio doesn't play -
+
+
+
+**URL**: https://www.google.com/search?q=beeg&ie=utf-8&oe=utf-8&client=firefox-b-m
+
+**Browser / Version**: Firefox Mobile 66.0
+**Operating System**: Android 8.0.0
+**Tested Another Browser**: No
+
+**Problem type**: Video or audio doesn't play
+**Description**: DON'T open web
+**Steps to Reproduce**:
+
+
+
+Browser Configuration
+
+
mixed active content blocked: false
image.mem.shared: true
buildID: 20190304101322
tracking content blocked: false
gfx.webrender.blob-images: true
hasTouchScreen: true
mixed passive content blocked: false
gfx.webrender.enabled: false
gfx.webrender.all: false
channel: beta
+
+
+
+
+_From [webcompat.com](https://webcompat.com/) with ❤️_",0, video or audio doesn t play url browser version firefox mobile operating system android tested another browser no problem type video or audio doesn t play description don t open web steps to reproduce browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen true mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel beta from with ❤️ ,0
+189919,22047152685.0,IssuesEvent,2022-05-30 04:00:21,nanopathi/linux-4.19.72_CVE-2021-32399,https://api.github.com/repos/nanopathi/linux-4.19.72_CVE-2021-32399,closed,CVE-2019-19052 (High) detected in linuxlinux-4.19.236 - autoclosed,security vulnerability,"## CVE-2019-19052 - High Severity Vulnerability
+ Vulnerable Library - linuxlinux-4.19.236
+
+
+A memory leak in the gs_can_open() function in drivers/net/can/usb/gs_usb.c in the Linux kernel before 5.3.11 allows attackers to cause a denial of service (memory consumption) by triggering usb_submit_urb() failures, aka CID-fb5be6a7b486.
+
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",True,"CVE-2019-19052 (High) detected in linuxlinux-4.19.236 - autoclosed - ## CVE-2019-19052 - High Severity Vulnerability
+ Vulnerable Library - linuxlinux-4.19.236
+
+
+A memory leak in the gs_can_open() function in drivers/net/can/usb/gs_usb.c in the Linux kernel before 5.3.11 allows attackers to cause a denial of service (memory consumption) by triggering usb_submit_urb() failures, aka CID-fb5be6a7b486.
+
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",0,cve high detected in linuxlinux autoclosed cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files drivers net can usb gs usb c drivers net can usb gs usb c vulnerability details a memory leak in the gs can open function in drivers net can usb gs usb c in the linux kernel before allows attackers to cause a denial of service memory consumption by triggering usb submit urb failures aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource ,0
+169319,20834553188.0,IssuesEvent,2022-03-20 01:08:18,kapseliboi/coronavirus-dashboard,https://api.github.com/repos/kapseliboi/coronavirus-dashboard,opened,CVE-2021-23648 (Medium) detected in sanitize-url-5.0.2.tgz,security vulnerability,"## CVE-2021-23648 - Medium Severity Vulnerability
+ Vulnerable Library - sanitize-url-5.0.2.tgz
+
+
+
+The package @braintree/sanitize-url before 6.0.0 are vulnerable to Cross-site Scripting (XSS) due to improper sanitization in sanitizeUrl function.
+
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",True,"CVE-2021-23648 (Medium) detected in sanitize-url-5.0.2.tgz - ## CVE-2021-23648 - Medium Severity Vulnerability
+ Vulnerable Library - sanitize-url-5.0.2.tgz
+
+
+
+The package @braintree/sanitize-url before 6.0.0 are vulnerable to Cross-site Scripting (XSS) due to improper sanitization in sanitizeUrl function.
+
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",0,cve medium detected in sanitize url tgz cve medium severity vulnerability vulnerable library sanitize url tgz a url sanitizer library home page a href path to dependency file package json path to vulnerable library node modules braintree sanitize url package json dependency hierarchy swagger ui react tgz root library x sanitize url tgz vulnerable library found in base branch development vulnerability details the package braintree sanitize url before are vulnerable to cross site scripting xss due to improper sanitization in sanitizeurl function publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution braintree sanitize url step up your open source security game with whitesource ,0
+346422,24886833206.0,IssuesEvent,2022-10-28 08:31:04,CedricChia123/ped,https://api.github.com/repos/CedricChia123/ped,opened,Case sensitive status application,severity.Low type.DocumentationBug,"When editing the status of an application using mark INDEX s/APPLICATION_STATUS, inputing an application status of ""rEjECted"" is still accepted. Perhaps can be specified in the UG that the command is case insensitive.
+
+
+
+",1.0,"Case sensitive status application - When editing the status of an application using mark INDEX s/APPLICATION_STATUS, inputing an application status of ""rEjECted"" is still accepted. Perhaps can be specified in the UG that the command is case insensitive.
+
+
+
+",0,case sensitive status application when editing the status of an application using mark index s application status inputing an application status of rejected is still accepted perhaps can be specified in the ug that the command is case insensitive ,0
+839,15731278015.0,IssuesEvent,2021-03-29 16:52:55,openstates/issues,https://api.github.com/repos/openstates/issues,closed,"SD: old, invalid source URLs for one legislator",component:people-data type:bug,"**Issue Description:**
+For `ocd-person/e40b342d-1732-4d1d-aa23-6d904e871e98` (Wayne H. Steinhauer), the source URLs listed are
+
+```
+http://legis.sd.gov/Legislators/Legislators/MemberCommittees.aspx?Member=1069&Session=2016
+http://legis.sd.gov/Legislators/Legislators/MemberDetail.aspx?Member=1069&Session=2016
+```
+
+`legis.sd.gov` no longer exists. All other source correctly use `sdlegislature.gov`.",1.0,"SD: old, invalid source URLs for one legislator - **Issue Description:**
+For `ocd-person/e40b342d-1732-4d1d-aa23-6d904e871e98` (Wayne H. Steinhauer), the source URLs listed are
+
+```
+http://legis.sd.gov/Legislators/Legislators/MemberCommittees.aspx?Member=1069&Session=2016
+http://legis.sd.gov/Legislators/Legislators/MemberDetail.aspx?Member=1069&Session=2016
+```
+
+`legis.sd.gov` no longer exists. All other source correctly use `sdlegislature.gov`.",1,sd old invalid source urls for one legislator issue description for ocd person wayne h steinhauer the source urls listed are legis sd gov no longer exists all other source correctly use sdlegislature gov ,1
+284322,30913618492.0,IssuesEvent,2023-08-05 02:24:51,Satheesh575555/linux-4.1.15_CVE-2022-45934,https://api.github.com/repos/Satheesh575555/linux-4.1.15_CVE-2022-45934,reopened,CVE-2022-28356 (Medium) detected in linuxlinux-4.6,Mend: dependency security vulnerability,"## CVE-2022-28356 - Medium Severity Vulnerability
+ Vulnerable Library - linuxlinux-4.6
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",True,"CVE-2022-28356 (Medium) detected in linuxlinux-4.6 - ## CVE-2022-28356 - Medium Severity Vulnerability
+ Vulnerable Library - linuxlinux-4.6
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",0,cve medium detected in linuxlinux cve medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files net llc af llc c net llc af llc c vulnerability details in the linux kernel before a refcount leak bug was found in net llc af llc c publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend ,0
+612,10703339306.0,IssuesEvent,2019-10-24 09:22:08,3box/ops,https://api.github.com/repos/3box/ops,closed,RB onboarding plan ,Ops People,"Update https://www.notion.so/threebox/Rachel-Black-2dc04ffdff654f34bec391fa4afdeb51
+
+- [ ] 10-45-90 day plan
+- [ ] ramp up materials
+- [ ] kickoff meetings ",1.0,"RB onboarding plan - Update https://www.notion.so/threebox/Rachel-Black-2dc04ffdff654f34bec391fa4afdeb51
+
+- [ ] 10-45-90 day plan
+- [ ] ramp up materials
+- [ ] kickoff meetings ",1,rb onboarding plan update day plan ramp up materials kickoff meetings ,1
+468,8353448879.0,IssuesEvent,2018-10-02 10:03:38,allenleein/brains,https://api.github.com/repos/allenleein/brains,closed,How Trumps Election Shook Obama: What if We Were Wrong?,People,"How Trump’s Election Shook Obama: ‘What if We Were Wrong?’
+WASHINGTON — Riding in a motorcade in Lima, Peru, shortly after the 2016 election, President Barack Obama was struggling to understand Donald J. Trump’s victory.
+
+via Pocket https://ift.tt/2JhsNyD
+
+May 31, 2018 at 01:35PM",1.0,"How Trumps Election Shook Obama: What if We Were Wrong? - How Trump’s Election Shook Obama: ‘What if We Were Wrong?’
+WASHINGTON — Riding in a motorcade in Lima, Peru, shortly after the 2016 election, President Barack Obama was struggling to understand Donald J. Trump’s victory.
+
+via Pocket https://ift.tt/2JhsNyD
+
+May 31, 2018 at 01:35PM",1,how trumps election shook obama what if we were wrong how trump rsquo s election shook obama lsquo what if we were wrong rsquo washington mdash riding in a motorcade in lima peru shortly after the election president barack obama was struggling to understand donald j trump rsquo s victory via pocket may at ,1
+1072,26206298712.0,IssuesEvent,2023-01-03 23:06:50,openstates/issues,https://api.github.com/repos/openstates/issues,opened,New VT Committee Scraper,good first issue component:people-data good first scraper,"We need a new committee scraper written in [spatula](https://jamesturk.github.io/spatula/) for Vermont. The scraper should scrape this [webpage](https://legislature.vermont.gov/committee/loadList/2024/) to get the committee name, chamber, and members (including roles where applicable). There is an [old scraper](https://github.com/openstates/openstates-scrapers/blob/main/scrapers/vt/committees.py) that may be helpful in how we have captured the data in the past, along with a [spatula scraper](https://github.com/openstates/openstates-scrapers/blob/main/scrapers_next/az/committees.py) that also parses a JSON page for reference",1.0,"New VT Committee Scraper - We need a new committee scraper written in [spatula](https://jamesturk.github.io/spatula/) for Vermont. The scraper should scrape this [webpage](https://legislature.vermont.gov/committee/loadList/2024/) to get the committee name, chamber, and members (including roles where applicable). There is an [old scraper](https://github.com/openstates/openstates-scrapers/blob/main/scrapers/vt/committees.py) that may be helpful in how we have captured the data in the past, along with a [spatula scraper](https://github.com/openstates/openstates-scrapers/blob/main/scrapers_next/az/committees.py) that also parses a JSON page for reference",1,new vt committee scraper we need a new committee scraper written in for vermont the scraper should scrape this to get the committee name chamber and members including roles where applicable there is an that may be helpful in how we have captured the data in the past along with a that also parses a json page for reference,1
+134,3588476772.0,IssuesEvent,2016-01-31 01:43:03,newtheatre/history-project,https://api.github.com/repos/newtheatre/history-project,opened,Twitter feeds on people records,discussion people,For someone's bio entry should we embed their twitter feed on the page?,1.0,Twitter feeds on people records - For someone's bio entry should we embed their twitter feed on the page?,1,twitter feeds on people records for someone s bio entry should we embed their twitter feed on the page ,1
+1,2534249440.0,IssuesEvent,2015-01-24 19:19:57,ufvivotech/ufDataQualityImprovement,https://api.github.com/repos/ufvivotech/ufDataQualityImprovement,closed,Develop ingest for medinfo directory,People Software Dev,"The medinfo directory has education, board certification and other data of interest to VIVO. It is available here: https://find.medinfo.ufl.edu/remote_superfind1.php?p=xxx&id=g9595 (firewall protected, reguires HSC VPN).
+
+An ingest will be written to consume this data and update VIVO appropriately.",1.0,"Develop ingest for medinfo directory - The medinfo directory has education, board certification and other data of interest to VIVO. It is available here: https://find.medinfo.ufl.edu/remote_superfind1.php?p=xxx&id=g9595 (firewall protected, reguires HSC VPN).
+
+An ingest will be written to consume this data and update VIVO appropriately.",1,develop ingest for medinfo directory the medinfo directory has education board certification and other data of interest to vivo it is available here firewall protected reguires hsc vpn an ingest will be written to consume this data and update vivo appropriately ,1
+169947,26879082736.0,IssuesEvent,2023-02-05 12:14:44,microsoft/vscode,https://api.github.com/repos/microsoft/vscode,closed,Consider consolidating `Open folder...` and `Open file...` into one,*as-designed,"
+
+
+
+
+
+
+
+They both open a similar type of window with title `Open file` or `Open folder`. Instead we could have the title of the window say `Open file or folder`.",1.0,"Consider consolidating `Open folder...` and `Open file...` into one -
+
+
+
+
+
+
+
+They both open a similar type of window with title `Open file` or `Open folder`. Instead we could have the title of the window say `Open file or folder`.",0,consider consolidating open folder and open file into one they both open a similar type of window with title open file or open folder instead we could have the title of the window say open file or folder ,0
+149214,5713897545.0,IssuesEvent,2017-04-19 09:01:04,HPI-SWA-Lab/BP2016H1,https://api.github.com/repos/HPI-SWA-Lab/BP2016H1,opened,Make sure users can only edit their own data,bug priority high,We (A and J) ran into some issues concerning proper combination of Authorization for partial GET requests and only allowing write access for the user who created the resource. Make sure this works somehow.,1.0,Make sure users can only edit their own data - We (A and J) ran into some issues concerning proper combination of Authorization for partial GET requests and only allowing write access for the user who created the resource. Make sure this works somehow.,0,make sure users can only edit their own data we a and j ran into some issues concerning proper combination of authorization for partial get requests and only allowing write access for the user who created the resource make sure this works somehow ,0
+180447,6649935795.0,IssuesEvent,2017-09-28 14:46:53,siteorigin/so-widgets-bundle,https://api.github.com/repos/siteorigin/so-widgets-bundle,closed,Video Player Widget External oEmbed has issues with disabling related videos,bug priority-3,"As per this [support thread](https://siteorigin.com/thread/remove-related-videos-from-embedded-youtube-video/). With WordPress oEmbed you should be able to simply at ?ref=0 ([source](http://www.wpbeginner.com/plugins/how-to-turn-off-related-videos-when-embedding-youtube-videos-in-wordpress/)) to a YouTube link and have it disable related videos. However when this is present the video player is no longer present on the page.
+",1.0,"Video Player Widget External oEmbed has issues with disabling related videos - As per this [support thread](https://siteorigin.com/thread/remove-related-videos-from-embedded-youtube-video/). With WordPress oEmbed you should be able to simply at ?ref=0 ([source](http://www.wpbeginner.com/plugins/how-to-turn-off-related-videos-when-embedding-youtube-videos-in-wordpress/)) to a YouTube link and have it disable related videos. However when this is present the video player is no longer present on the page.
+",0,video player widget external oembed has issues with disabling related videos as per this with wordpress oembed you should be able to simply at ref to a youtube link and have it disable related videos however when this is present the video player is no longer present on the page ,0
+16075,21417490471.0,IssuesEvent,2022-04-22 12:26:34,WebberZone/contextual-related-posts,https://api.github.com/repos/WebberZone/contextual-related-posts,opened,Deprecation notice for get_crp_posts_id,compatibility,Add a deprecation notice for the function with the alternative to use `get_crp_posts`,True,Deprecation notice for get_crp_posts_id - Add a deprecation notice for the function with the alternative to use `get_crp_posts`,0,deprecation notice for get crp posts id add a deprecation notice for the function with the alternative to use get crp posts ,0
+637,11402720381.0,IssuesEvent,2020-01-31 04:27:32,OfficeDev/office-ui-fabric-react,https://api.github.com/repos/OfficeDev/office-ui-fabric-react,closed,how to access input DOM in people picker? ,Component: PeoplePicker Type: Question ❔,"how to access input DOM in peoplepicker? I need to clear what user has entered and change the behaviour of people-picker when user past or type in the people-picker. can you please share some code in react ( not typescript)
+",1.0,"how to access input DOM in people picker? - how to access input DOM in peoplepicker? I need to clear what user has entered and change the behaviour of people-picker when user past or type in the people-picker. can you please share some code in react ( not typescript)
+",1,how to access input dom in people picker how to access input dom in peoplepicker i need to clear what user has entered and change the behaviour of people picker when user past or type in the people picker can you please share some code in react not typescript ,1
+415568,12131182033.0,IssuesEvent,2020-04-23 03:53:17,buddyboss/buddyboss-platform,https://api.github.com/repos/buddyboss/buddyboss-platform,closed,Commenting disabled in activity stream for event published via events calendar pro,Has-PR bug component: activity priority: medium,"An option to comment on the activities generated for events via events calendar pro is not available.
+
+Steps to reproduce the behavior:
+
+-Go to activity settings page in the backend and enable activities for custom post type events
+-Click on ""activity page""
+-Scroll down to any of the events activities
+-See no commenting available, only Like or Delete ",1.0,"Commenting disabled in activity stream for event published via events calendar pro - An option to comment on the activities generated for events via events calendar pro is not available.
+
+Steps to reproduce the behavior:
+
+-Go to activity settings page in the backend and enable activities for custom post type events
+-Click on ""activity page""
+-Scroll down to any of the events activities
+-See no commenting available, only Like or Delete ",0,commenting disabled in activity stream for event published via events calendar pro an option to comment on the activities generated for events via events calendar pro is not available steps to reproduce the behavior go to activity settings page in the backend and enable activities for custom post type events click on activity page scroll down to any of the events activities see no commenting available only like or delete ,0
+20828,3421540277.0,IssuesEvent,2015-12-08 19:11:34,dart-lang/sdk,https://api.github.com/repos/dart-lang/sdk,closed,Security: Avoid bad range check pattern,accepted area-vm priority-unassigned security Type-Defect,"The following are two examples of code that uses a bad range check pattern:
+
+- DART_EXPORT Dart_Handle Dart_ListGetAsBytes(Dart_Handle list,
+ intptr_t offset,
+ uint8_t\* native_array,
+ intptr_t length) {
+ Isolate\* isolate = Isolate::Current();
+ DARTSCOPE(isolate);
+ const Object& obj = Object::Handle(Api::UnwrapHandle(list));
+ if (obj.IsArray()) {
+ Array& array_obj = Array::Handle();
+ array_obj ^= obj.raw();
+ if ((offset + length) <= array_obj.Length()) {
+ Object& element = Object::Handle();
+
+- byte_array.cc:
+ static void RangeCheck(const ByteArray& array, const Smi& index,
+ intptr_t num_bytes) {
+ if ((index.Value() < 0) || ((index.Value() + num_bytes) > array.Length())) {",1.0,"Security: Avoid bad range check pattern - The following are two examples of code that uses a bad range check pattern:
+
+- DART_EXPORT Dart_Handle Dart_ListGetAsBytes(Dart_Handle list,
+ intptr_t offset,
+ uint8_t\* native_array,
+ intptr_t length) {
+ Isolate\* isolate = Isolate::Current();
+ DARTSCOPE(isolate);
+ const Object& obj = Object::Handle(Api::UnwrapHandle(list));
+ if (obj.IsArray()) {
+ Array& array_obj = Array::Handle();
+ array_obj ^= obj.raw();
+ if ((offset + length) <= array_obj.Length()) {
+ Object& element = Object::Handle();
+
+- byte_array.cc:
+ static void RangeCheck(const ByteArray& array, const Smi& index,
+ intptr_t num_bytes) {
+ if ((index.Value() < 0) || ((index.Value() + num_bytes) > array.Length())) {",0,security avoid bad range check pattern the following are two examples of code that uses a bad range check pattern dart export dart handle dart listgetasbytes dart handle list nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp intptr t offset nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp t native array nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp intptr t length nbsp nbsp isolate isolate isolate current nbsp nbsp dartscope isolate nbsp nbsp const object amp obj object handle api unwraphandle list nbsp nbsp if obj isarray nbsp nbsp nbsp nbsp array amp array obj array handle nbsp nbsp nbsp nbsp array obj obj raw nbsp nbsp nbsp nbsp if offset length lt array obj length nbsp nbsp nbsp nbsp nbsp nbsp object amp element object handle byte array cc nbsp nbsp static void rangecheck const bytearray amp array const smi amp index nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp intptr t num bytes nbsp nbsp if index value lt index value num bytes gt array length ,0
+395211,11672637730.0,IssuesEvent,2020-03-04 07:11:14,AugurProject/augur,https://api.github.com/repos/AugurProject/augur,closed,disputing and reporting right hand panels should only switch to mobile components at mobile break.,Add post v2 launch Priority: Medium,"Looks like it's just the My available rep balance and current dispute window components that are switching to mobile components too early. My available REP balance shouldn't be collapsable until mobile.
+
+Build:
+
+
+
+Design:
+
+https://www.figma.com/file/aAzKHh4cA6OT2t7WFv2BQ7fB/Reporting-and-Disputing?node-id=2192%3A0
+
+",1.0,"disputing and reporting right hand panels should only switch to mobile components at mobile break. - Looks like it's just the My available rep balance and current dispute window components that are switching to mobile components too early. My available REP balance shouldn't be collapsable until mobile.
+
+Build:
+
+
+
+Design:
+
+https://www.figma.com/file/aAzKHh4cA6OT2t7WFv2BQ7fB/Reporting-and-Disputing?node-id=2192%3A0
+
+",0,disputing and reporting right hand panels should only switch to mobile components at mobile break looks like it s just the my available rep balance and current dispute window components that are switching to mobile components too early my available rep balance shouldn t be collapsable until mobile build design ,0
+604,10607357470.0,IssuesEvent,2019-10-11 03:25:38,ItsExtra/m4ftm,https://api.github.com/repos/ItsExtra/m4ftm,closed,reddit posts: A website to educate cis gay men about hooking up with trans men: What do you think?,:coffee: People to communicate with :coffee:,"
+Posted to [/r/ftm](https://www.reddit.com/r/ftm/comments/ddqv5w/a_website_to_educate_cis_gay_men_about_hooking_up/) then to [r/gaytransguys](https://www.reddit.com/r/gaytransguys/comments/ddvv0j/a_website_to_educate_cis_gay_men_about_hooking_up/)
+
+> ## A website to educate cis gay men about hooking up with trans men: What do you think?
+> Hey guys! This is primarily for those of you who are sexually/romantically interested in cis gay guys (not necessarily exclusively) though comments from anyone are also welcome.
+>
+> I've been working on a project which is basically aimed to correct misinformation about trans men specifically in the cis gay community. I think it is really sad that lots of cis gay guys are missing out on great sex and other fun they could be having with trans guys. (This is not at all self interested, only a community service!! haha)
+>
+> - Do you know of anything like this that already exists? Writing, videos, sites, socials etc?
+> - Would such a resource be of any use to you?
+> - What do you think would make such a thing helpful? Anything you think should be avoided?
+> - Would you be interested in collaborating in any way? Or giving feedback on work done?
+>
+> If everything goes as planned this will become an open source type community-driven project.
+>
+> I'm trying to keep this post brief but I can say more about the specifics if anyone's interested. For a random sampling of questions/misinformation you can search for ftm on /r/askgaybros.",1.0,"reddit posts: A website to educate cis gay men about hooking up with trans men: What do you think? -
+Posted to [/r/ftm](https://www.reddit.com/r/ftm/comments/ddqv5w/a_website_to_educate_cis_gay_men_about_hooking_up/) then to [r/gaytransguys](https://www.reddit.com/r/gaytransguys/comments/ddvv0j/a_website_to_educate_cis_gay_men_about_hooking_up/)
+
+> ## A website to educate cis gay men about hooking up with trans men: What do you think?
+> Hey guys! This is primarily for those of you who are sexually/romantically interested in cis gay guys (not necessarily exclusively) though comments from anyone are also welcome.
+>
+> I've been working on a project which is basically aimed to correct misinformation about trans men specifically in the cis gay community. I think it is really sad that lots of cis gay guys are missing out on great sex and other fun they could be having with trans guys. (This is not at all self interested, only a community service!! haha)
+>
+> - Do you know of anything like this that already exists? Writing, videos, sites, socials etc?
+> - Would such a resource be of any use to you?
+> - What do you think would make such a thing helpful? Anything you think should be avoided?
+> - Would you be interested in collaborating in any way? Or giving feedback on work done?
+>
+> If everything goes as planned this will become an open source type community-driven project.
+>
+> I'm trying to keep this post brief but I can say more about the specifics if anyone's interested. For a random sampling of questions/misinformation you can search for ftm on /r/askgaybros.",1,reddit posts a website to educate cis gay men about hooking up with trans men what do you think posted to then to a website to educate cis gay men about hooking up with trans men what do you think hey guys this is primarily for those of you who are sexually romantically interested in cis gay guys not necessarily exclusively though comments from anyone are also welcome i ve been working on a project which is basically aimed to correct misinformation about trans men specifically in the cis gay community i think it is really sad that lots of cis gay guys are missing out on great sex and other fun they could be having with trans guys this is not at all self interested only a community service haha do you know of anything like this that already exists writing videos sites socials etc would such a resource be of any use to you what do you think would make such a thing helpful anything you think should be avoided would you be interested in collaborating in any way or giving feedback on work done if everything goes as planned this will become an open source type community driven project i m trying to keep this post brief but i can say more about the specifics if anyone s interested for a random sampling of questions misinformation you can search for ftm on r askgaybros ,1
+1139,28351811731.0,IssuesEvent,2023-04-12 03:22:44,scikit-image/scikit-image,https://api.github.com/repos/scikit-image/scikit-image,closed,Numpy array for list of AffineTransform not working since Numpy '1.24.2',:people_hugging: Support,"Since I updated Numpy, I can't compute the cumsum of a set of AffineTransfrorms anymore.
+
+For example:
+
+` trafos = [transform.AffineTransform(translation=(0, 0)),
+ transform.AffineTransform(translation=(1, 1))]
+ newts = np.cumsum(np.asarray(trafos))`
+
+worked with older versions of numpy, but now throws the following error.
+
+`ValueError: setting an array element with a sequence. The requested array has an inhomogeneous shape after 1 dimensions. The detected shape was (2,) + inhomogeneous part.`
+
+Is this a numpy or skimage issue?",1.0,"Numpy array for list of AffineTransform not working since Numpy '1.24.2' - Since I updated Numpy, I can't compute the cumsum of a set of AffineTransfrorms anymore.
+
+For example:
+
+` trafos = [transform.AffineTransform(translation=(0, 0)),
+ transform.AffineTransform(translation=(1, 1))]
+ newts = np.cumsum(np.asarray(trafos))`
+
+worked with older versions of numpy, but now throws the following error.
+
+`ValueError: setting an array element with a sequence. The requested array has an inhomogeneous shape after 1 dimensions. The detected shape was (2,) + inhomogeneous part.`
+
+Is this a numpy or skimage issue?",1,numpy array for list of affinetransform not working since numpy since i updated numpy i can t compute the cumsum of a set of affinetransfrorms anymore for example trafos transform affinetransform translation transform affinetransform translation newts np cumsum np asarray trafos worked with older versions of numpy but now throws the following error valueerror setting an array element with a sequence the requested array has an inhomogeneous shape after dimensions the detected shape was inhomogeneous part is this a numpy or skimage issue ,1
+110545,13913415099.0,IssuesEvent,2020-10-20 20:24:36,department-of-veterans-affairs/va.gov-team,https://api.github.com/repos/department-of-veterans-affairs/va.gov-team,closed,[Design] Healthcare: Refill and track your prescriptions ,design my-va-dashboard my-va-phase-2 vsa-authenticated-exp,"## Background
+See [Design plan](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/products/identity-personalization/logged-in-homepage/2.0-redesign/product/LIH-outline-and-timeline.md)
+
+Note: Shown conditionally if someone has prescriptions
+
+## Tasks
+- [ ] Wireframes for all states
+- [ ] Visual design exploration
+- [ ] High fidelity mockups
+- [ ] Appropriate stakeholder reviews and approvals have been completed
+",1.0,"[Design] Healthcare: Refill and track your prescriptions - ## Background
+See [Design plan](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/products/identity-personalization/logged-in-homepage/2.0-redesign/product/LIH-outline-and-timeline.md)
+
+Note: Shown conditionally if someone has prescriptions
+
+## Tasks
+- [ ] Wireframes for all states
+- [ ] Visual design exploration
+- [ ] High fidelity mockups
+- [ ] Appropriate stakeholder reviews and approvals have been completed
+",0, healthcare refill and track your prescriptions background see note shown conditionally if someone has prescriptions tasks wireframes for all states visual design exploration high fidelity mockups appropriate stakeholder reviews and approvals have been completed ,0
+489,8443369294.0,IssuesEvent,2018-10-18 15:25:31,openstates/openstates,https://api.github.com/repos/openstates/openstates,closed,Some VT reps not coming up correctly,manual people,"1. Christopher Chris Mattos house VT shows up inactive instead of Ronald E. Hubert who is supposed to be inactive but is active
+
+ 2. David Soucy senate VT shows up inactive instead of Kevin Mullin who is supposed to be inactive but is active
+
+3. James Jim Harrison house VT shows up inactive instead of Job Tate who is supposed to be inactive but is active
+
+",1.0,"Some VT reps not coming up correctly - 1. Christopher Chris Mattos house VT shows up inactive instead of Ronald E. Hubert who is supposed to be inactive but is active
+
+ 2. David Soucy senate VT shows up inactive instead of Kevin Mullin who is supposed to be inactive but is active
+
+3. James Jim Harrison house VT shows up inactive instead of Job Tate who is supposed to be inactive but is active
+
+",1,some vt reps not coming up correctly christopher chris mattos house vt shows up inactive instead of ronald e hubert who is supposed to be inactive but is active david soucy senate vt shows up inactive instead of kevin mullin who is supposed to be inactive but is active james jim harrison house vt shows up inactive instead of job tate who is supposed to be inactive but is active ,1
+118703,15358857239.0,IssuesEvent,2021-03-01 15:14:18,jameszangari/idm372,https://api.github.com/repos/jameszangari/idm372,opened,Chat requests for all new incoming messages,bug design enhancement,"
+",1.0,"Chat requests for all new incoming messages -
+",0,chat requests for all new incoming messages img width alt screen shot at am src ,0
+682850,23359439091.0,IssuesEvent,2022-08-10 10:19:35,zuri-training/Col_Films_Proj_Team_113,https://api.github.com/repos/zuri-training/Col_Films_Proj_Team_113,closed,Sign up (creators),frontend priority feature,"Kindly implement sign up (creators) for phone and email. And upload document form.
+
+https://www.figma.com/file/HmFzbwjxQiDFUxPYxsp8Q6/Col-films-designs-(Copy)?node-id=1%3A2",1.0,"Sign up (creators) - Kindly implement sign up (creators) for phone and email. And upload document form.
+
+https://www.figma.com/file/HmFzbwjxQiDFUxPYxsp8Q6/Col-films-designs-(Copy)?node-id=1%3A2",0,sign up creators kindly implement sign up creators for phone and email and upload document form ,0
+566174,16813943232.0,IssuesEvent,2021-06-17 04:00:43,oppia/oppia-android,https://api.github.com/repos/oppia/oppia-android,closed,Create model layer for checkpointing,Priority: Essential Type: Task Where: Exploration player Where: Topics/Stories,"Create new and modify existing protobufs to implement lightweight checkpointing.
+Target | Date
+--|--
+**PR creation:** | 2021-06-08
+**PR completion:** | 2021-06-11",1.0,"Create model layer for checkpointing - Create new and modify existing protobufs to implement lightweight checkpointing.
+Target | Date
+--|--
+**PR creation:** | 2021-06-08
+**PR completion:** | 2021-06-11",0,create model layer for checkpointing create new and modify existing protobufs to implement lightweight checkpointing target date pr creation pr completion ,0
+584,10374343836.0,IssuesEvent,2019-09-09 09:24:07,ushahidi/tenfour,https://api.github.com/repos/ushahidi/tenfour,opened,Create a location based group when someone add a location on their profile,Feature: Check-In Feature: People P2 - Normal,"From tenfour-archive created by [Erioldoesdesign](https://github.com/Erioldoesdesign): ushahidi/tenfour-archive#1569
+
+As a user, I was groups to be created automatically when an individual team member registered/sets-up on TenFour and adds in their location on their profile.
+
+If a user adds in 'Bristol UK' into their profile then a location specific group is created for 'Bristol UK'.
+
+Any new team members who add 'Bristol UK' as their location on the profile should then be added to that same 'Bristol UK' group.
+
+The work on Alerts #1104 depends on this to function.
+
+",1.0,"Create a location based group when someone add a location on their profile - From tenfour-archive created by [Erioldoesdesign](https://github.com/Erioldoesdesign): ushahidi/tenfour-archive#1569
+
+As a user, I was groups to be created automatically when an individual team member registered/sets-up on TenFour and adds in their location on their profile.
+
+If a user adds in 'Bristol UK' into their profile then a location specific group is created for 'Bristol UK'.
+
+Any new team members who add 'Bristol UK' as their location on the profile should then be added to that same 'Bristol UK' group.
+
+The work on Alerts #1104 depends on this to function.
+
+",1,create a location based group when someone add a location on their profile from tenfour archive created by ushahidi tenfour archive as a user i was groups to be created automatically when an individual team member registered sets up on tenfour and adds in their location on their profile if a user adds in bristol uk into their profile then a location specific group is created for bristol uk any new team members who add bristol uk as their location on the profile should then be added to that same bristol uk group the work on alerts depends on this to function ,1
+747820,26099916462.0,IssuesEvent,2022-12-27 05:06:12,Digital-Will-Inc/wortal-sdk-unity,https://api.github.com/repos/Digital-Will-Inc/wortal-sdk-unity,closed,Fix compress fallback logging in installer,Priority: Low Type: Bug,"### Description
+
+
+- Type: Bug
+- ID: 864djz9rp
+- Priority: Low",0,fix compress fallback logging in installer description type bug id priority low,0
+1174,30644716350.0,IssuesEvent,2023-07-25 02:57:31,restincode/restincode,https://api.github.com/repos/restincode/restincode,opened,Carol Anthony,People Add Person Needs Review,"Carol
+Anthony
+
+ISACA
+MTA
+BANK OF AMERICA
+PHILLIPS
+PROTIVITI
+
+* Facebook:
+* Other:
+
+**Contributions**
+Repeat as many times as needed
+
+* Project name:
+* Project URL:
+* Project Description:
+
+**Photo Gallery**
+* URL(s) to additional photos:
+",1.0,"Carol Anthony - Carol
+Anthony
+
+ISACA
+MTA
+BANK OF AMERICA
+PHILLIPS
+PROTIVITI
+
+* Facebook:
+* Other:
+
+**Contributions**
+Repeat as many times as needed
+
+* Project name:
+* Project URL:
+* Project Description:
+
+**Photo Gallery**
+* URL(s) to additional photos:
+",1,carol anthony carol anthony isaca mta bank of america phillips protiviti facebook other contributions repeat as many times as needed project name project url project description photo gallery url s to additional photos ,1
+332486,10096494102.0,IssuesEvent,2019-07-27 18:37:18,kubernetes/kubernetes,https://api.github.com/repos/kubernetes/kubernetes,closed,Refactor reconciler code,area/kubelet lifecycle/rotten priority/backlog sig/node,"Refactor reconciler code in /pkg/kubelet/volumemanager/reconciler/reconciler.go
+",1.0,"Refactor reconciler code - Refactor reconciler code in /pkg/kubelet/volumemanager/reconciler/reconciler.go
+",0,refactor reconciler code refactor reconciler code in pkg kubelet volumemanager reconciler reconciler go ,0
+703900,24177258836.0,IssuesEvent,2022-09-23 04:13:38,kubermatic/operating-system-manager,https://api.github.com/repos/kubermatic/operating-system-manager,closed,Support Ubuntu 22.04 in OSM,kind/feature priority/low sig/cluster-management,"Description of the feature you would like to add / User story
+As Ubuntu 22.04 is the new LTS Version of Ubuntu we should also support this as an OS.
+KKP and KubeOne rely on the support in OSM, to make this a supported OS in those projects we need the suppport in OSM.
+
+Solution details
+Support Ubuntu 22.04 as OS in operating-system-manager
+
+Related issues:
+https://github.com/kubermatic/kubeone/issues/2223
+https://github.com/kubermatic/kubermatic/issues/10669
+https://github.com/kubermatic/machine-controller/issues/1394",1.0,"Support Ubuntu 22.04 in OSM - Description of the feature you would like to add / User story
+As Ubuntu 22.04 is the new LTS Version of Ubuntu we should also support this as an OS.
+KKP and KubeOne rely on the support in OSM, to make this a supported OS in those projects we need the suppport in OSM.
+
+Solution details
+Support Ubuntu 22.04 as OS in operating-system-manager
+
+Related issues:
+https://github.com/kubermatic/kubeone/issues/2223
+https://github.com/kubermatic/kubermatic/issues/10669
+https://github.com/kubermatic/machine-controller/issues/1394",0,support ubuntu in osm description of the feature you would like to add user story as ubuntu is the new lts version of ubuntu we should also support this as an os kkp and kubeone rely on the support in osm to make this a supported os in those projects we need the suppport in osm solution details support ubuntu as os in operating system manager related issues ,0
+186751,15083158078.0,IssuesEvent,2021-02-05 15:28:56,gianlucadetommaso/volatile,https://api.github.com/repos/gianlucadetommaso/volatile,closed,Do you have plan to provide a interact jupyter notebook to dive into the model construction ?,documentation,"The tensorflow official website provide a example about multilevel modeling
+https://www.tensorflow.org/probability/examples/Multilevel_Modeling_Primer
+it interpret its components by visualize them separately.
+I think you should also provide a correspondence about your model.
+Let it to validate your model construction and change to adapt their own applications.
+Even, if you have a guide about the feature selection and aggregate method
+in your model with visualize is more bravo.",1.0,"Do you have plan to provide a interact jupyter notebook to dive into the model construction ? - The tensorflow official website provide a example about multilevel modeling
+https://www.tensorflow.org/probability/examples/Multilevel_Modeling_Primer
+it interpret its components by visualize them separately.
+I think you should also provide a correspondence about your model.
+Let it to validate your model construction and change to adapt their own applications.
+Even, if you have a guide about the feature selection and aggregate method
+in your model with visualize is more bravo.",0,do you have plan to provide a interact jupyter notebook to dive into the model construction the tensorflow official website provide a example about multilevel modeling it interpret its components by visualize them separately i think you should also provide a correspondence about your model let it to validate your model construction and change to adapt their own applications even if you have a guide about the feature selection and aggregate method in your model with visualize is more bravo ,0
+683,12266157701.0,IssuesEvent,2020-05-07 08:28:52,bitprj/bitproject,https://api.github.com/repos/bitprj/bitproject,opened,Manager Procedures for Mid-Week Syncs on Partners/Devs ,people,"**Objectives**
+No one is keeping account of people on Github. To change that we need to develop procedures and guides for managers/directors to check in mid week for each of their team members **if**
+there is no updates on their issues.
+
+**Collaborators and Tasks**
+@beccatran - Create a Play by Play procedure for managers to follow to ensure that people are in sync with their OKRs.
+
+For example,
+**Devrel**
+- Every Wed
+If there has been NO Updates on the assigned issue, send slack message
+""Hey There [insertnamehere]! I hope you are doing okay. I noticed that there hasn't been any movement on your assigned github issue. Just wanted to check in if there were any blocks that I can help resolve.
+
+- Every Friday (given meeting is on Every Saturday)
+If there has been NO Updates on the assigned issue, send slack message
+
+""Hey There [insertnamehere]! I hope you are doing okay. I noticed that you haven't finished your assigned github issue. Just wanted to check in if there were any blocks that I can help resolve. I would love to set up a quick sync to make sure you are ready for the meeting.
+
+**and set up quick 15 min call**
+
+
+**Deadline**
+Fri, 05/15/20",1.0,"Manager Procedures for Mid-Week Syncs on Partners/Devs - **Objectives**
+No one is keeping account of people on Github. To change that we need to develop procedures and guides for managers/directors to check in mid week for each of their team members **if**
+there is no updates on their issues.
+
+**Collaborators and Tasks**
+@beccatran - Create a Play by Play procedure for managers to follow to ensure that people are in sync with their OKRs.
+
+For example,
+**Devrel**
+- Every Wed
+If there has been NO Updates on the assigned issue, send slack message
+""Hey There [insertnamehere]! I hope you are doing okay. I noticed that there hasn't been any movement on your assigned github issue. Just wanted to check in if there were any blocks that I can help resolve.
+
+- Every Friday (given meeting is on Every Saturday)
+If there has been NO Updates on the assigned issue, send slack message
+
+""Hey There [insertnamehere]! I hope you are doing okay. I noticed that you haven't finished your assigned github issue. Just wanted to check in if there were any blocks that I can help resolve. I would love to set up a quick sync to make sure you are ready for the meeting.
+
+**and set up quick 15 min call**
+
+
+**Deadline**
+Fri, 05/15/20",1,manager procedures for mid week syncs on partners devs objectives no one is keeping account of people on github to change that we need to develop procedures and guides for managers directors to check in mid week for each of their team members if there is no updates on their issues collaborators and tasks beccatran create a play by play procedure for managers to follow to ensure that people are in sync with their okrs for example devrel every wed if there has been no updates on the assigned issue send slack message hey there i hope you are doing okay i noticed that there hasn t been any movement on your assigned github issue just wanted to check in if there were any blocks that i can help resolve every friday given meeting is on every saturday if there has been no updates on the assigned issue send slack message hey there i hope you are doing okay i noticed that you haven t finished your assigned github issue just wanted to check in if there were any blocks that i can help resolve i would love to set up a quick sync to make sure you are ready for the meeting and set up quick min call deadline fri ,1
+7134,16659847456.0,IssuesEvent,2021-06-06 06:16:29,OndrejSzekely/metron,https://api.github.com/repos/OndrejSzekely/metron,opened,Add Hydra config framework,Metron architecture,"Moving from [Dynaconf](https://dynaconf.readthedocs.io/en/docs_223/) configuration framework to Facebook's [Hydra](https://hydra.cc/) config tool. This allows to have much broader configuration options, which more fit into our scenario. ",1.0,"Add Hydra config framework - Moving from [Dynaconf](https://dynaconf.readthedocs.io/en/docs_223/) configuration framework to Facebook's [Hydra](https://hydra.cc/) config tool. This allows to have much broader configuration options, which more fit into our scenario. ",0,add hydra config framework moving from configuration framework to facebook s config tool this allows to have much broader configuration options which more fit into our scenario ,0
+10596,2956223986.0,IssuesEvent,2015-07-08 09:52:31,ThibaultLatrille/ControverSciences,https://api.github.com/repos/ThibaultLatrille/ControverSciences,opened,La référence Effet hépatoprotecteur de la substance homéopathique Lycopodium clavatum (Lyc). n'a aucun lien vers d'autres références,*** important design,"Il faudrait mettre des guillemets pour éviter qu´un ""."" n´apparaisse au milieu de la phrase :
+
+La référence ""Effet hépatoprotecteur de la substance homéopathique Lycopodium clavatum (Lyc)."" n'a aucun lien vers d'autres références
+ Sur la page https://www.controversciences.org/reference_edges?reference_id=7&timeline_id=17
+ Par : F. Giry
+ Navigateur : chrome modern windows webkit",1.0,"La référence Effet hépatoprotecteur de la substance homéopathique Lycopodium clavatum (Lyc). n'a aucun lien vers d'autres références - Il faudrait mettre des guillemets pour éviter qu´un ""."" n´apparaisse au milieu de la phrase :
+
+La référence ""Effet hépatoprotecteur de la substance homéopathique Lycopodium clavatum (Lyc)."" n'a aucun lien vers d'autres références
+ Sur la page https://www.controversciences.org/reference_edges?reference_id=7&timeline_id=17
+ Par : F. Giry
+ Navigateur : chrome modern windows webkit",0,la référence effet hépatoprotecteur de la substance homéopathique lycopodium clavatum lyc n a aucun lien vers d autres références il faudrait mettre des guillemets pour éviter qu´un n´apparaisse au milieu de la phrase la référence effet hépatoprotecteur de la substance homéopathique lycopodium clavatum lyc n a aucun lien vers d autres références sur la page par f giry navigateur chrome modern windows webkit,0
+313,5723617538.0,IssuesEvent,2017-04-20 12:44:18,PopulateTools/gobierto,https://api.github.com/repos/PopulateTools/gobierto,opened,Add location to IBM Notes imported events,gobierto-people,IBM Notes events do have a location attribute. It should be created and synched as a GobiertoPersonEvent instance. For the moment we just need to store the **name** of the place.,1.0,Add location to IBM Notes imported events - IBM Notes events do have a location attribute. It should be created and synched as a GobiertoPersonEvent instance. For the moment we just need to store the **name** of the place.,1,add location to ibm notes imported events ibm notes events do have a location attribute it should be created and synched as a gobiertopersonevent instance for the moment we just need to store the name of the place ,1
+575,10130756480.0,IssuesEvent,2019-08-01 17:46:17,openopps/openopps-platform,https://api.github.com/repos/openopps/openopps-platform,closed,"As an administrator, I want the People search to be weighted to better allow searching on multiple fields.",Enhancement People,"People search searches on name, title, agency, and location. Upgrade to work better for multiple search fields. Determine appropriate weighting if multiple search options are entered (and search vs or?)
+
+Entered per 3/6/18 bug bash",1.0,"As an administrator, I want the People search to be weighted to better allow searching on multiple fields. - People search searches on name, title, agency, and location. Upgrade to work better for multiple search fields. Determine appropriate weighting if multiple search options are entered (and search vs or?)
+
+Entered per 3/6/18 bug bash",1,as an administrator i want the people search to be weighted to better allow searching on multiple fields people search searches on name title agency and location upgrade to work better for multiple search fields determine appropriate weighting if multiple search options are entered and search vs or entered per bug bash,1
+102,3422094464.0,IssuesEvent,2015-12-08 21:32:31,faithmade/faith-builder,https://api.github.com/repos/faithmade/faith-builder,closed,What is Sort 'low to high' and 'high to low' mean?,Location Module People Module Question Sermons Module,"http://lift.d.pr/1lEam/4dU4HI3g
+
+This will need to be labeled more intuitively or educate the customer ",1.0,"What is Sort 'low to high' and 'high to low' mean? - http://lift.d.pr/1lEam/4dU4HI3g
+
+This will need to be labeled more intuitively or educate the customer ",1,what is sort low to high and high to low mean this will need to be labeled more intuitively or educate the customer ,1
+430,7926427119.0,IssuesEvent,2018-07-06 02:03:25,SmartDataAnalytics/sda.tech,https://api.github.com/repos/SmartDataAnalytics/sda.tech,closed,Missing values for Social profiles,People,"As some of the members may not have all the social profiles defined to be shown on the profile page, would be great to make it more dynamic which lists only those which are listed on the Person card/entry.
+ - http://sda.tech/Person/GezimSejdiu/ -- if you click on the facebook icon it will open an instance of the same page, but would have been better if we hide it and list other profiles which are linked on my profile (e.g. [Twitter](https://github.com/SmartDataAnalytics/sda.tech/blob/develop/rdf-data/sda.tech.ttl#L348)).
+",1.0,"Missing values for Social profiles - As some of the members may not have all the social profiles defined to be shown on the profile page, would be great to make it more dynamic which lists only those which are listed on the Person card/entry.
+ - http://sda.tech/Person/GezimSejdiu/ -- if you click on the facebook icon it will open an instance of the same page, but would have been better if we hide it and list other profiles which are linked on my profile (e.g. [Twitter](https://github.com/SmartDataAnalytics/sda.tech/blob/develop/rdf-data/sda.tech.ttl#L348)).
+",1,missing values for social profiles as some of the members may not have all the social profiles defined to be shown on the profile page would be great to make it more dynamic which lists only those which are listed on the person card entry if you click on the facebook icon it will open an instance of the same page but would have been better if we hide it and list other profiles which are linked on my profile e g ,1
+17796,3641276588.0,IssuesEvent,2016-02-13 14:20:30,elastic/elasticsearch,https://api.github.com/repos/elastic/elasticsearch,closed,Can not run Tika StandAlone runner,:Plugin Mapper Attachment test,"In mapper-attachments plugin, I created originally a [Stand Alone command line tool](https://github.com/elastic/elasticsearch/blob/master/plugins/mapper-attachments/src/test/java/org/elasticsearch/mapper/attachments/StandaloneRunner.java) which helps me to extract data from binary files.
+It helps to understand what Tika actually extracts before sending the text to elasticsearch. Basically it helps to debug when users report issues.
+
+In master branch, when I run the StandAlone tool (which is a `main()` - not a test) from IntelliJ, I'm getting this error:
+
+```
+Exception in thread ""main"" java.lang.IllegalStateException: running tests but failed to invoke RandomizedContext#getRandom
+ at org.elasticsearch.common.Randomness.get(Randomness.java:105)
+ at org.elasticsearch.node.internal.InternalSettingsPreparer.randomNodeName(InternalSettingsPreparer.java:194)
+ at org.elasticsearch.node.internal.InternalSettingsPreparer.finalizeSettings(InternalSettingsPreparer.java:167)
+ at org.elasticsearch.node.internal.InternalSettingsPreparer.prepareEnvironment(InternalSettingsPreparer.java:106)
+ at org.elasticsearch.common.cli.CliTool.(CliTool.java:100)
+ at org.elasticsearch.common.cli.CliTool.(CliTool.java:91)
+ at org.elasticsearch.mapper.attachments.StandaloneRunner.(StandaloneRunner.java:171)
+ at org.elasticsearch.mapper.attachments.StandaloneRunner.main(StandaloneRunner.java:176)
+ at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
+ at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
+ at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
+ at java.lang.reflect.Method.invoke(Method.java:497)
+ at com.intellij.rt.execution.application.AppMain.main(AppMain.java:144)
+Caused by: java.lang.reflect.InvocationTargetException
+ at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
+ at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
+ at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
+ at java.lang.reflect.Method.invoke(Method.java:497)
+ at org.elasticsearch.common.Randomness.get(Randomness.java:101)
+ ... 12 more
+Caused by: java.lang.IllegalStateException: No context information for thread: Thread[id=1, name=main, state=RUNNABLE, group=main]. Is this thread running under a class com.carrotsearch.randomizedtesting.RandomizedRunner runner context? Add @RunWith(class com.carrotsearch.randomizedtesting.RandomizedRunner.class) to your test class. Make sure your code accesses random contexts within @BeforeClass and @AfterClass boundary (for example, static test class initializers are not permitted to access random contexts).
+ at com.carrotsearch.randomizedtesting.RandomizedContext.context(RandomizedContext.java:244)
+ at com.carrotsearch.randomizedtesting.RandomizedContext.current(RandomizedContext.java:151)
+ ... 17 more
+```
+
+I'm not expecting anything from the test framework here so I'm a bit surprise about the result.
+
+I know that a `main` should not be in our `src/test` dir but I don't really want to commit this class in `src/main`.
+
+Should I simply remove that class because we can't use it anymore?
+Knowing that this plugin will probably become deprecated by the node-ingest one...",1.0,"Can not run Tika StandAlone runner - In mapper-attachments plugin, I created originally a [Stand Alone command line tool](https://github.com/elastic/elasticsearch/blob/master/plugins/mapper-attachments/src/test/java/org/elasticsearch/mapper/attachments/StandaloneRunner.java) which helps me to extract data from binary files.
+It helps to understand what Tika actually extracts before sending the text to elasticsearch. Basically it helps to debug when users report issues.
+
+In master branch, when I run the StandAlone tool (which is a `main()` - not a test) from IntelliJ, I'm getting this error:
+
+```
+Exception in thread ""main"" java.lang.IllegalStateException: running tests but failed to invoke RandomizedContext#getRandom
+ at org.elasticsearch.common.Randomness.get(Randomness.java:105)
+ at org.elasticsearch.node.internal.InternalSettingsPreparer.randomNodeName(InternalSettingsPreparer.java:194)
+ at org.elasticsearch.node.internal.InternalSettingsPreparer.finalizeSettings(InternalSettingsPreparer.java:167)
+ at org.elasticsearch.node.internal.InternalSettingsPreparer.prepareEnvironment(InternalSettingsPreparer.java:106)
+ at org.elasticsearch.common.cli.CliTool.(CliTool.java:100)
+ at org.elasticsearch.common.cli.CliTool.(CliTool.java:91)
+ at org.elasticsearch.mapper.attachments.StandaloneRunner.(StandaloneRunner.java:171)
+ at org.elasticsearch.mapper.attachments.StandaloneRunner.main(StandaloneRunner.java:176)
+ at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
+ at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
+ at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
+ at java.lang.reflect.Method.invoke(Method.java:497)
+ at com.intellij.rt.execution.application.AppMain.main(AppMain.java:144)
+Caused by: java.lang.reflect.InvocationTargetException
+ at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
+ at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
+ at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
+ at java.lang.reflect.Method.invoke(Method.java:497)
+ at org.elasticsearch.common.Randomness.get(Randomness.java:101)
+ ... 12 more
+Caused by: java.lang.IllegalStateException: No context information for thread: Thread[id=1, name=main, state=RUNNABLE, group=main]. Is this thread running under a class com.carrotsearch.randomizedtesting.RandomizedRunner runner context? Add @RunWith(class com.carrotsearch.randomizedtesting.RandomizedRunner.class) to your test class. Make sure your code accesses random contexts within @BeforeClass and @AfterClass boundary (for example, static test class initializers are not permitted to access random contexts).
+ at com.carrotsearch.randomizedtesting.RandomizedContext.context(RandomizedContext.java:244)
+ at com.carrotsearch.randomizedtesting.RandomizedContext.current(RandomizedContext.java:151)
+ ... 17 more
+```
+
+I'm not expecting anything from the test framework here so I'm a bit surprise about the result.
+
+I know that a `main` should not be in our `src/test` dir but I don't really want to commit this class in `src/main`.
+
+Should I simply remove that class because we can't use it anymore?
+Knowing that this plugin will probably become deprecated by the node-ingest one...",0,can not run tika standalone runner in mapper attachments plugin i created originally a which helps me to extract data from binary files it helps to understand what tika actually extracts before sending the text to elasticsearch basically it helps to debug when users report issues in master branch when i run the standalone tool which is a main not a test from intellij i m getting this error exception in thread main java lang illegalstateexception running tests but failed to invoke randomizedcontext getrandom at org elasticsearch common randomness get randomness java at org elasticsearch node internal internalsettingspreparer randomnodename internalsettingspreparer java at org elasticsearch node internal internalsettingspreparer finalizesettings internalsettingspreparer java at org elasticsearch node internal internalsettingspreparer prepareenvironment internalsettingspreparer java at org elasticsearch common cli clitool clitool java at org elasticsearch common cli clitool clitool java at org elasticsearch mapper attachments standalonerunner standalonerunner java at org elasticsearch mapper attachments standalonerunner main standalonerunner java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at com intellij rt execution application appmain main appmain java caused by java lang reflect invocationtargetexception at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org elasticsearch common randomness get randomness java more caused by java lang illegalstateexception no context information for thread thread is this thread running under a class com carrotsearch randomizedtesting randomizedrunner runner context add runwith class com carrotsearch randomizedtesting randomizedrunner class to your test class make sure your code accesses random contexts within beforeclass and afterclass boundary for example static test class initializers are not permitted to access random contexts at com carrotsearch randomizedtesting randomizedcontext context randomizedcontext java at com carrotsearch randomizedtesting randomizedcontext current randomizedcontext java more i m not expecting anything from the test framework here so i m a bit surprise about the result i know that a main should not be in our src test dir but i don t really want to commit this class in src main should i simply remove that class because we can t use it anymore knowing that this plugin will probably become deprecated by the node ingest one ,0
+296,5613622444.0,IssuesEvent,2017-04-03 09:48:17,PopulateTools/gobierto,https://api.github.com/repos/PopulateTools/gobierto,closed,Events political group filters,gobierto-people,"There are a few issues related with events political groups filters:
+
+1. calendar events should be filterd by the political group. Example: in a site without directives, if I filter in the events page the calendar shows past and future events but the list of events in the main column is empty.
+
+2. when I'm in the filter ""Past Events"" and I click into another political group filter this filter should be maintained. Right now, it changes to ""Future events""
+
+3. people list should be filterd by the political group. Right now in that list I can see everybody in the list
+
+
+
+",1.0,"Events political group filters - There are a few issues related with events political groups filters:
+
+1. calendar events should be filterd by the political group. Example: in a site without directives, if I filter in the events page the calendar shows past and future events but the list of events in the main column is empty.
+
+2. when I'm in the filter ""Past Events"" and I click into another political group filter this filter should be maintained. Right now, it changes to ""Future events""
+
+3. people list should be filterd by the political group. Right now in that list I can see everybody in the list
+
+
+
+",1,events political group filters there are a few issues related with events political groups filters calendar events should be filterd by the political group example in a site without directives if i filter in the events page the calendar shows past and future events but the list of events in the main column is empty when i m in the filter past events and i click into another political group filter this filter should be maintained right now it changes to future events people list should be filterd by the political group right now in that list i can see everybody in the list ,1
+127997,5042293966.0,IssuesEvent,2016-12-19 13:31:24,koding/koding,https://api.github.com/repos/koding/koding,opened,go: fix flag handling,A-Bug Priority-Low,"Currently there is a number of packages, that register flags globally. It makes kd and klient to inhering flags from non-related packages, like github.com/koding/runner or terraformer.
+
+Example output from instrumented `flag.Var` method call for klient binary:
+
+```
+./klient
+Registering ""test.bench"" with (*flag.stringValue)(0xc42010d6f0) from: [flag.(*FlagSet).Var flag.(*FlagSet).StringVar flag.(*FlagSet).String flag.String testing.init go-getter.init module.init terraform.init terraformer.init session.init]
+Registering ""test.benchtime"" with (*flag.durationValue)(0xc42010d7c0) from: [flag.(*FlagSet).Var flag.(*FlagSet).DurationVar flag.(*FlagSet).Duration flag.Duration testing.init go-getter.init module.init terraform.init terraformer.init session.init]
+Registering ""test.benchmem"" with (*flag.boolValue)(0xc42010d7ca) from: [flag.(*FlagSet).Var flag.(*FlagSet).BoolVar flag.(*FlagSet).Bool flag.Bool testing.init go-getter.init module.init terraform.init terraformer.init session.init]
+Registering ""test.short"" with (*flag.boolValue)(0xc42010d935) from: [flag.(*FlagSet).Var flag.(*FlagSet).BoolVar flag.(*FlagSet).Bool flag.Bool testing.init go-getter.init module.init terraform.init terraformer.init session.init]
+Registering ""test.outputdir"" with (*flag.stringValue)(0xc42010d9f0) from: [flag.(*FlagSet).Var flag.(*FlagSet).StringVar flag.(*FlagSet).String flag.String testing.init go-getter.init module.init terraform.init terraformer.init session.init]
+Registering ""test.v"" with (*flag.boolValue)(0xc42010d93c) from: [flag.(*FlagSet).Var flag.(*FlagSet).BoolVar flag.(*FlagSet).Bool flag.Bool testing.init go-getter.init module.init terraform.init terraformer.init session.init]
+Registering ""test.count"" with (*flag.uintValue)(0xc42010db78) from: [flag.(*FlagSet).Var flag.(*FlagSet).UintVar flag.(*FlagSet).Uint flag.Uint testing.init go-getter.init module.init terraform.init terraformer.init session.init]
+Registering ""test.coverprofile"" with (*flag.stringValue)(0xc42010dc40) from: [flag.(*FlagSet).Var flag.(*FlagSet).StringVar flag.(*FlagSet).String flag.String testing.init go-getter.init module.init terraform.init terraformer.init session.init]
+Registering ""test.run"" with (*flag.stringValue)(0xc42010dd10) from: [flag.(*FlagSet).Var flag.(*FlagSet).StringVar flag.(*FlagSet).String flag.String testing.init go-getter.init module.init terraform.init terraformer.init session.init]
+Registering ""test.memprofile"" with (*flag.stringValue)(0xc42010dde0) from: [flag.(*FlagSet).Var flag.(*FlagSet).StringVar flag.(*FlagSet).String flag.String testing.init go-getter.init module.init terraform.init terraformer.init session.init]
+Registering ""test.memprofilerate"" with (*flag.intValue)(0xc42010deb0) from: [flag.(*FlagSet).Var flag.(*FlagSet).IntVar flag.(*FlagSet).Int flag.Int testing.init go-getter.init module.init terraform.init terraformer.init session.init]
+Registering ""test.cpuprofile"" with (*flag.stringValue)(0xc42010df80) from: [flag.(*FlagSet).Var flag.(*FlagSet).StringVar flag.(*FlagSet).String flag.String testing.init go-getter.init module.init terraform.init terraformer.init session.init]
+Registering ""test.blockprofile"" with (*flag.stringValue)(0xc42014e050) from: [flag.(*FlagSet).Var flag.(*FlagSet).StringVar flag.(*FlagSet).String flag.String testing.init go-getter.init module.init terraform.init terraformer.init session.init]
+Registering ""test.blockprofilerate"" with (*flag.intValue)(0xc42010df78) from: [flag.(*FlagSet).Var flag.(*FlagSet).IntVar flag.(*FlagSet).Int flag.Int testing.init go-getter.init module.init terraform.init terraformer.init session.init]
+Registering ""test.trace"" with (*flag.stringValue)(0xc42014e1e0) from: [flag.(*FlagSet).Var flag.(*FlagSet).StringVar flag.(*FlagSet).String flag.String testing.init go-getter.init module.init terraform.init terraformer.init session.init]
+Registering ""test.timeout"" with (*flag.durationValue)(0xc42014e2b0) from: [flag.(*FlagSet).Var flag.(*FlagSet).DurationVar flag.(*FlagSet).Duration flag.Duration testing.init go-getter.init module.init terraform.init terraformer.init session.init]
+Registering ""test.cpu"" with (*flag.stringValue)(0xc42014e370) from: [flag.(*FlagSet).Var flag.(*FlagSet).StringVar flag.(*FlagSet).String flag.String testing.init go-getter.init module.init terraform.init terraformer.init session.init]
+Registering ""test.parallel"" with (*flag.intValue)(0xc42014e2b8) from: [flag.(*FlagSet).Var flag.(*FlagSet).IntVar flag.(*FlagSet).Int flag.Int testing.init go-getter.init module.init terraform.init terraformer.init session.init]
+Registering ""c"" with (*flag.stringValue)(0xc420215ea0) from: [flag.(*FlagSet).Var flag.(*FlagSet).StringVar flag.(*FlagSet).String flag.String runner.init config.init models.init api.init client.init stack.init]
+Registering ""r"" with (*flag.stringValue)(0xc420215f90) from: [flag.(*FlagSet).Var flag.(*FlagSet).StringVar flag.(*FlagSet).String flag.String runner.init config.init models.init api.init client.init stack.init]
+Registering ""d"" with (*flag.boolValue)(0xc42020a200) from: [flag.(*FlagSet).Var flag.(*FlagSet).BoolVar flag.(*FlagSet).Bool flag.Bool runner.init config.init models.init api.init client.init stack.init]
+Registering ""v"" with (*flag.intValue)(0xc42020a208) from: [flag.(*FlagSet).Var flag.(*FlagSet).IntVar flag.(*FlagSet).Int flag.Int runner.init config.init models.init api.init client.init stack.init]
+Registering ""outputMetrics"" with (*flag.boolValue)(0xc42020a569) from: [flag.(*FlagSet).Var flag.(*FlagSet).BoolVar flag.(*FlagSet).Bool flag.Bool runner.init config.init models.init api.init client.init stack.init]
+Registering ""kite-init"" with (*flag.boolValue)(0xc42020a620) from: [flag.(*FlagSet).Var flag.(*FlagSet).BoolVar flag.(*FlagSet).Bool flag.Bool runner.init config.init models.init api.init client.init stack.init]
+Registering ""kite-local"" with (*flag.boolValue)(0xc42020a627) from: [flag.(*FlagSet).Var flag.(*FlagSet).BoolVar flag.(*FlagSet).Bool flag.Bool runner.init config.init models.init api.init client.init stack.init]
+Registering ""kite-proxy"" with (*flag.boolValue)(0xc42020a62e) from: [flag.(*FlagSet).Var flag.(*FlagSet).BoolVar flag.(*FlagSet).Bool flag.Bool runner.init config.init models.init api.init client.init stack.init]
+Registering ""kite-kontrol-url"" with (*flag.stringValue)(0xc42020a910) from: [flag.(*FlagSet).Var flag.(*FlagSet).StringVar flag.(*FlagSet).String flag.String runner.init config.init models.init api.init client.init stack.init]
+Registering ""host"" with (*flag.stringValue)(0xc42020aab0) from: [flag.(*FlagSet).Var flag.(*FlagSet).StringVar flag.(*FlagSet).String flag.String runner.init config.init models.init api.init client.init stack.init]
+Registering ""port"" with (*flag.stringValue)(0xc42020aba0) from: [flag.(*FlagSet).Var flag.(*FlagSet).StringVar flag.(*FlagSet).String flag.String runner.init config.init models.init api.init client.init stack.init]
+Registering ""convey-json"" with (*flag.boolValue)(0x16ccd02) from: [flag.(*FlagSet).Var flag.BoolVar convey.declareFlags convey.init.1 convey.init models.init api.init client.init stack.init apiutil.init]
+Registering ""convey-silent"" with (*flag.boolValue)(0x16ccd03) from: [flag.(*FlagSet).Var flag.BoolVar convey.declareFlags convey.init.1 convey.init models.init api.init client.init stack.init apiutil.init]
+Registering ""convey-story"" with (*flag.boolValue)(0x16ccd04) from: [flag.(*FlagSet).Var flag.BoolVar convey.declareFlags convey.init.1 convey.init models.init api.init client.init stack.init apiutil.init]
+Registering ""httptest.serve"" with (*flag.stringValue)(0xc42051cb60) from: [flag.(*FlagSet).Var flag.(*FlagSet).StringVar flag.(*FlagSet).String flag.String httptest.init tlsproxy.init tunnel.init app.init main.init runtime.main]
+Registering ""ip"" with (*flag.stringValue)(0xc420539710) from: [flag.(*FlagSet).Var flag.(*FlagSet).StringVar flag.(*FlagSet).String main.init runtime.main runtime.goexit]
+Registering ""port"" with (*flag.intValue)(0xc4205397d0) from: [flag.(*FlagSet).Var flag.(*FlagSet).IntVar flag.(*FlagSet).Int main.init runtime.main runtime.goexit]
+./klient flag redefined: port
+panic: ./klient flag redefined: port
+
+goroutine 1 [running]:
+panic(0xc7cf40, 0xc4205278a0)
+ /Users/rjeczalik/go/src/runtime/panic.go:500 +0x1a1
+flag.(*FlagSet).Var(0xc42001c300, 0x166b280, 0xc420527800, 0xf11c0e, 0x4, 0xf25b3f, 0x13)
+ /Users/rjeczalik/go/src/flag/flag.go:814 +0x543
+flag.(*FlagSet).IntVar(0xc42001c300, 0xc420527800, 0xf11c0e, 0x4, 0xddd5, 0xf25b3f, 0x13)
+ /Users/rjeczalik/go/src/flag/flag.go:593 +0x71
+flag.(*FlagSet).Int(0xc42001c300, 0xf11c0e, 0x4, 0xddd5, 0xf25b3f, 0x13, 0xc4205277c0)
+ /Users/rjeczalik/go/src/flag/flag.go:606 +0x80
+flag.Int(0xf11c0e, 0x4, 0xddd5, 0xf25b3f, 0x13, 0x10)
+ /Users/rjeczalik/go/src/flag/flag.go:613 +0x5f
+main.init()
+ /Users/rjeczalik/src/github.com/koding/koding/go/src/koding/klient/main.go:28 +0x12a
+
+```",1.0,"go: fix flag handling - Currently there is a number of packages, that register flags globally. It makes kd and klient to inhering flags from non-related packages, like github.com/koding/runner or terraformer.
+
+Example output from instrumented `flag.Var` method call for klient binary:
+
+```
+./klient
+Registering ""test.bench"" with (*flag.stringValue)(0xc42010d6f0) from: [flag.(*FlagSet).Var flag.(*FlagSet).StringVar flag.(*FlagSet).String flag.String testing.init go-getter.init module.init terraform.init terraformer.init session.init]
+Registering ""test.benchtime"" with (*flag.durationValue)(0xc42010d7c0) from: [flag.(*FlagSet).Var flag.(*FlagSet).DurationVar flag.(*FlagSet).Duration flag.Duration testing.init go-getter.init module.init terraform.init terraformer.init session.init]
+Registering ""test.benchmem"" with (*flag.boolValue)(0xc42010d7ca) from: [flag.(*FlagSet).Var flag.(*FlagSet).BoolVar flag.(*FlagSet).Bool flag.Bool testing.init go-getter.init module.init terraform.init terraformer.init session.init]
+Registering ""test.short"" with (*flag.boolValue)(0xc42010d935) from: [flag.(*FlagSet).Var flag.(*FlagSet).BoolVar flag.(*FlagSet).Bool flag.Bool testing.init go-getter.init module.init terraform.init terraformer.init session.init]
+Registering ""test.outputdir"" with (*flag.stringValue)(0xc42010d9f0) from: [flag.(*FlagSet).Var flag.(*FlagSet).StringVar flag.(*FlagSet).String flag.String testing.init go-getter.init module.init terraform.init terraformer.init session.init]
+Registering ""test.v"" with (*flag.boolValue)(0xc42010d93c) from: [flag.(*FlagSet).Var flag.(*FlagSet).BoolVar flag.(*FlagSet).Bool flag.Bool testing.init go-getter.init module.init terraform.init terraformer.init session.init]
+Registering ""test.count"" with (*flag.uintValue)(0xc42010db78) from: [flag.(*FlagSet).Var flag.(*FlagSet).UintVar flag.(*FlagSet).Uint flag.Uint testing.init go-getter.init module.init terraform.init terraformer.init session.init]
+Registering ""test.coverprofile"" with (*flag.stringValue)(0xc42010dc40) from: [flag.(*FlagSet).Var flag.(*FlagSet).StringVar flag.(*FlagSet).String flag.String testing.init go-getter.init module.init terraform.init terraformer.init session.init]
+Registering ""test.run"" with (*flag.stringValue)(0xc42010dd10) from: [flag.(*FlagSet).Var flag.(*FlagSet).StringVar flag.(*FlagSet).String flag.String testing.init go-getter.init module.init terraform.init terraformer.init session.init]
+Registering ""test.memprofile"" with (*flag.stringValue)(0xc42010dde0) from: [flag.(*FlagSet).Var flag.(*FlagSet).StringVar flag.(*FlagSet).String flag.String testing.init go-getter.init module.init terraform.init terraformer.init session.init]
+Registering ""test.memprofilerate"" with (*flag.intValue)(0xc42010deb0) from: [flag.(*FlagSet).Var flag.(*FlagSet).IntVar flag.(*FlagSet).Int flag.Int testing.init go-getter.init module.init terraform.init terraformer.init session.init]
+Registering ""test.cpuprofile"" with (*flag.stringValue)(0xc42010df80) from: [flag.(*FlagSet).Var flag.(*FlagSet).StringVar flag.(*FlagSet).String flag.String testing.init go-getter.init module.init terraform.init terraformer.init session.init]
+Registering ""test.blockprofile"" with (*flag.stringValue)(0xc42014e050) from: [flag.(*FlagSet).Var flag.(*FlagSet).StringVar flag.(*FlagSet).String flag.String testing.init go-getter.init module.init terraform.init terraformer.init session.init]
+Registering ""test.blockprofilerate"" with (*flag.intValue)(0xc42010df78) from: [flag.(*FlagSet).Var flag.(*FlagSet).IntVar flag.(*FlagSet).Int flag.Int testing.init go-getter.init module.init terraform.init terraformer.init session.init]
+Registering ""test.trace"" with (*flag.stringValue)(0xc42014e1e0) from: [flag.(*FlagSet).Var flag.(*FlagSet).StringVar flag.(*FlagSet).String flag.String testing.init go-getter.init module.init terraform.init terraformer.init session.init]
+Registering ""test.timeout"" with (*flag.durationValue)(0xc42014e2b0) from: [flag.(*FlagSet).Var flag.(*FlagSet).DurationVar flag.(*FlagSet).Duration flag.Duration testing.init go-getter.init module.init terraform.init terraformer.init session.init]
+Registering ""test.cpu"" with (*flag.stringValue)(0xc42014e370) from: [flag.(*FlagSet).Var flag.(*FlagSet).StringVar flag.(*FlagSet).String flag.String testing.init go-getter.init module.init terraform.init terraformer.init session.init]
+Registering ""test.parallel"" with (*flag.intValue)(0xc42014e2b8) from: [flag.(*FlagSet).Var flag.(*FlagSet).IntVar flag.(*FlagSet).Int flag.Int testing.init go-getter.init module.init terraform.init terraformer.init session.init]
+Registering ""c"" with (*flag.stringValue)(0xc420215ea0) from: [flag.(*FlagSet).Var flag.(*FlagSet).StringVar flag.(*FlagSet).String flag.String runner.init config.init models.init api.init client.init stack.init]
+Registering ""r"" with (*flag.stringValue)(0xc420215f90) from: [flag.(*FlagSet).Var flag.(*FlagSet).StringVar flag.(*FlagSet).String flag.String runner.init config.init models.init api.init client.init stack.init]
+Registering ""d"" with (*flag.boolValue)(0xc42020a200) from: [flag.(*FlagSet).Var flag.(*FlagSet).BoolVar flag.(*FlagSet).Bool flag.Bool runner.init config.init models.init api.init client.init stack.init]
+Registering ""v"" with (*flag.intValue)(0xc42020a208) from: [flag.(*FlagSet).Var flag.(*FlagSet).IntVar flag.(*FlagSet).Int flag.Int runner.init config.init models.init api.init client.init stack.init]
+Registering ""outputMetrics"" with (*flag.boolValue)(0xc42020a569) from: [flag.(*FlagSet).Var flag.(*FlagSet).BoolVar flag.(*FlagSet).Bool flag.Bool runner.init config.init models.init api.init client.init stack.init]
+Registering ""kite-init"" with (*flag.boolValue)(0xc42020a620) from: [flag.(*FlagSet).Var flag.(*FlagSet).BoolVar flag.(*FlagSet).Bool flag.Bool runner.init config.init models.init api.init client.init stack.init]
+Registering ""kite-local"" with (*flag.boolValue)(0xc42020a627) from: [flag.(*FlagSet).Var flag.(*FlagSet).BoolVar flag.(*FlagSet).Bool flag.Bool runner.init config.init models.init api.init client.init stack.init]
+Registering ""kite-proxy"" with (*flag.boolValue)(0xc42020a62e) from: [flag.(*FlagSet).Var flag.(*FlagSet).BoolVar flag.(*FlagSet).Bool flag.Bool runner.init config.init models.init api.init client.init stack.init]
+Registering ""kite-kontrol-url"" with (*flag.stringValue)(0xc42020a910) from: [flag.(*FlagSet).Var flag.(*FlagSet).StringVar flag.(*FlagSet).String flag.String runner.init config.init models.init api.init client.init stack.init]
+Registering ""host"" with (*flag.stringValue)(0xc42020aab0) from: [flag.(*FlagSet).Var flag.(*FlagSet).StringVar flag.(*FlagSet).String flag.String runner.init config.init models.init api.init client.init stack.init]
+Registering ""port"" with (*flag.stringValue)(0xc42020aba0) from: [flag.(*FlagSet).Var flag.(*FlagSet).StringVar flag.(*FlagSet).String flag.String runner.init config.init models.init api.init client.init stack.init]
+Registering ""convey-json"" with (*flag.boolValue)(0x16ccd02) from: [flag.(*FlagSet).Var flag.BoolVar convey.declareFlags convey.init.1 convey.init models.init api.init client.init stack.init apiutil.init]
+Registering ""convey-silent"" with (*flag.boolValue)(0x16ccd03) from: [flag.(*FlagSet).Var flag.BoolVar convey.declareFlags convey.init.1 convey.init models.init api.init client.init stack.init apiutil.init]
+Registering ""convey-story"" with (*flag.boolValue)(0x16ccd04) from: [flag.(*FlagSet).Var flag.BoolVar convey.declareFlags convey.init.1 convey.init models.init api.init client.init stack.init apiutil.init]
+Registering ""httptest.serve"" with (*flag.stringValue)(0xc42051cb60) from: [flag.(*FlagSet).Var flag.(*FlagSet).StringVar flag.(*FlagSet).String flag.String httptest.init tlsproxy.init tunnel.init app.init main.init runtime.main]
+Registering ""ip"" with (*flag.stringValue)(0xc420539710) from: [flag.(*FlagSet).Var flag.(*FlagSet).StringVar flag.(*FlagSet).String main.init runtime.main runtime.goexit]
+Registering ""port"" with (*flag.intValue)(0xc4205397d0) from: [flag.(*FlagSet).Var flag.(*FlagSet).IntVar flag.(*FlagSet).Int main.init runtime.main runtime.goexit]
+./klient flag redefined: port
+panic: ./klient flag redefined: port
+
+goroutine 1 [running]:
+panic(0xc7cf40, 0xc4205278a0)
+ /Users/rjeczalik/go/src/runtime/panic.go:500 +0x1a1
+flag.(*FlagSet).Var(0xc42001c300, 0x166b280, 0xc420527800, 0xf11c0e, 0x4, 0xf25b3f, 0x13)
+ /Users/rjeczalik/go/src/flag/flag.go:814 +0x543
+flag.(*FlagSet).IntVar(0xc42001c300, 0xc420527800, 0xf11c0e, 0x4, 0xddd5, 0xf25b3f, 0x13)
+ /Users/rjeczalik/go/src/flag/flag.go:593 +0x71
+flag.(*FlagSet).Int(0xc42001c300, 0xf11c0e, 0x4, 0xddd5, 0xf25b3f, 0x13, 0xc4205277c0)
+ /Users/rjeczalik/go/src/flag/flag.go:606 +0x80
+flag.Int(0xf11c0e, 0x4, 0xddd5, 0xf25b3f, 0x13, 0x10)
+ /Users/rjeczalik/go/src/flag/flag.go:613 +0x5f
+main.init()
+ /Users/rjeczalik/src/github.com/koding/koding/go/src/koding/klient/main.go:28 +0x12a
+
+```",0,go fix flag handling currently there is a number of packages that register flags globally it makes kd and klient to inhering flags from non related packages like github com koding runner or terraformer example output from instrumented flag var method call for klient binary klient registering test bench with flag stringvalue from registering test benchtime with flag durationvalue from registering test benchmem with flag boolvalue from registering test short with flag boolvalue from registering test outputdir with flag stringvalue from registering test v with flag boolvalue from registering test count with flag uintvalue from registering test coverprofile with flag stringvalue from registering test run with flag stringvalue from registering test memprofile with flag stringvalue from registering test memprofilerate with flag intvalue from registering test cpuprofile with flag stringvalue from registering test blockprofile with flag stringvalue from registering test blockprofilerate with flag intvalue from registering test trace with flag stringvalue from registering test timeout with flag durationvalue from registering test cpu with flag stringvalue from registering test parallel with flag intvalue from registering c with flag stringvalue from registering r with flag stringvalue from registering d with flag boolvalue from registering v with flag intvalue from registering outputmetrics with flag boolvalue from registering kite init with flag boolvalue from registering kite local with flag boolvalue from registering kite proxy with flag boolvalue from registering kite kontrol url with flag stringvalue from registering host with flag stringvalue from registering port with flag stringvalue from registering convey json with flag boolvalue from registering convey silent with flag boolvalue from registering convey story with flag boolvalue from registering httptest serve with flag stringvalue from registering ip with flag stringvalue from registering port with flag intvalue from klient flag redefined port panic klient flag redefined port goroutine panic users rjeczalik go src runtime panic go flag flagset var users rjeczalik go src flag flag go flag flagset intvar users rjeczalik go src flag flag go flag flagset int users rjeczalik go src flag flag go flag int users rjeczalik go src flag flag go main init users rjeczalik src github com koding koding go src koding klient main go ,0
+1010,3293361970.0,IssuesEvent,2015-10-30 18:34:56,rancher/rancher,https://api.github.com/repos/rancher/rancher,closed,Changing URL doesn't work with existing DATA folder,area/catalog-service kind/bug status/resolved status/to-test,"If you run rancher-catalog-service with one catalogUrl and then change it to another, it doesn't clean up the old data or pull down the new repo.",1.0,"Changing URL doesn't work with existing DATA folder - If you run rancher-catalog-service with one catalogUrl and then change it to another, it doesn't clean up the old data or pull down the new repo.",0,changing url doesn t work with existing data folder if you run rancher catalog service with one catalogurl and then change it to another it doesn t clean up the old data or pull down the new repo ,0
+814,15189135779.0,IssuesEvent,2021-02-15 16:00:25,LiamWellacott/CDT2019-ERL,https://api.github.com/repos/LiamWellacott/CDT2019-ERL,closed,Emergency stop state (Manipulation and State Machine),Arm Trajectory Control NLP Object Perception People Perception State Machine,"Establish what triggers cause the arms to intisialise an emergency state
+
+Force over threshold,
+Vocal command
+Physical command
+Virtual/Controller command
+
+Determine what the emergency state is based on whether there is an object in hand, and what that object is.",1.0,"Emergency stop state (Manipulation and State Machine) - Establish what triggers cause the arms to intisialise an emergency state
+
+Force over threshold,
+Vocal command
+Physical command
+Virtual/Controller command
+
+Determine what the emergency state is based on whether there is an object in hand, and what that object is.",1,emergency stop state manipulation and state machine establish what triggers cause the arms to intisialise an emergency state force over threshold vocal command physical command virtual controller command determine what the emergency state is based on whether there is an object in hand and what that object is ,1
+299,5628290486.0,IssuesEvent,2017-04-05 05:48:13,PopulateTools/gobierto,https://api.github.com/repos/PopulateTools/gobierto,closed,Agendas: fix calendar navigation arrows,bug gobierto-people,"Calendar component navigation arrows stop working after having clicked on an event link inside the calendar.
+
+
+",1.0,"Agendas: fix calendar navigation arrows - Calendar component navigation arrows stop working after having clicked on an event link inside the calendar.
+
+
+",1,agendas fix calendar navigation arrows calendar component navigation arrows stop working after having clicked on an event link inside the calendar ,1
+294544,25380478470.0,IssuesEvent,2022-11-21 17:08:19,dask/dask,https://api.github.com/repos/dask/dask,closed,`test_from_dict_backends` failure on gpuCI,tests gpu,"See https://github.com/dask/dask/pull/9675#pullrequestreview-1186408344
+
+```python
+10:10:22 ________________________ test_from_dict_backends[cudf] _________________________
+10:10:22 [gw2] linux -- Python 3.9.13 /opt/conda/envs/dask/bin/python3.9
+10:10:22
+10:10:22 backend = 'cudf'
+10:10:22
+10:10:22 @pytest.mark.gpu
+10:10:22 @pytest.mark.parametrize(""backend"", [""pandas"", ""cudf""])
+10:10:22 def test_from_dict_backends(backend):
+10:10:22 _lib = pytest.importorskip(backend)
+10:10:22 with config.set({""dataframe.backend"": backend}):
+10:10:22 data = {""a"": [1, 2, 3, 4], ""B"": [10, 11, 12, 13]}
+10:10:22 expected = _lib.DataFrame(data)
+10:10:22
+10:10:22 # Check dd.from_dict API
+10:10:22 got = dd.from_dict(data, npartitions=2)
+10:10:22 assert_eq(expected, got)
+10:10:22
+10:10:22 # Check from_dict classmethod
+10:10:22 > got_classmethod = got.from_dict(data, npartitions=2)
+10:10:22
+10:10:22 dask/dataframe/io/tests/test_io.py:988:
+10:10:22 _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+10:10:22 dask/dataframe/core.py:6054: in from_dict
+10:10:22 return from_dict(
+10:10:22 _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+10:10:22
+10:10:22 args = ({'B': [10, 11, 12, 13], 'a': [1, 2, 3, 4]}, 2)
+10:10:22 kwargs = {'columns': None, 'constructor': , 'dtype': None, 'orient': 'columns'}
+10:10:22
+10:10:22 @wraps(fn)
+10:10:22 def wrapper(*args, **kwargs):
+10:10:22 > return getattr(self, dispatch_name)(*args, **kwargs)
+10:10:22 E TypeError: from_dict() got an unexpected keyword argument 'constructor'
+10:10:22
+10:10:22 dask/backends.py:122: TypeError
+```
+
+cc @rjzamora @galipremsagar @charlesbluca ",1.0,"`test_from_dict_backends` failure on gpuCI - See https://github.com/dask/dask/pull/9675#pullrequestreview-1186408344
+
+```python
+10:10:22 ________________________ test_from_dict_backends[cudf] _________________________
+10:10:22 [gw2] linux -- Python 3.9.13 /opt/conda/envs/dask/bin/python3.9
+10:10:22
+10:10:22 backend = 'cudf'
+10:10:22
+10:10:22 @pytest.mark.gpu
+10:10:22 @pytest.mark.parametrize(""backend"", [""pandas"", ""cudf""])
+10:10:22 def test_from_dict_backends(backend):
+10:10:22 _lib = pytest.importorskip(backend)
+10:10:22 with config.set({""dataframe.backend"": backend}):
+10:10:22 data = {""a"": [1, 2, 3, 4], ""B"": [10, 11, 12, 13]}
+10:10:22 expected = _lib.DataFrame(data)
+10:10:22
+10:10:22 # Check dd.from_dict API
+10:10:22 got = dd.from_dict(data, npartitions=2)
+10:10:22 assert_eq(expected, got)
+10:10:22
+10:10:22 # Check from_dict classmethod
+10:10:22 > got_classmethod = got.from_dict(data, npartitions=2)
+10:10:22
+10:10:22 dask/dataframe/io/tests/test_io.py:988:
+10:10:22 _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+10:10:22 dask/dataframe/core.py:6054: in from_dict
+10:10:22 return from_dict(
+10:10:22 _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+10:10:22
+10:10:22 args = ({'B': [10, 11, 12, 13], 'a': [1, 2, 3, 4]}, 2)
+10:10:22 kwargs = {'columns': None, 'constructor': , 'dtype': None, 'orient': 'columns'}
+10:10:22
+10:10:22 @wraps(fn)
+10:10:22 def wrapper(*args, **kwargs):
+10:10:22 > return getattr(self, dispatch_name)(*args, **kwargs)
+10:10:22 E TypeError: from_dict() got an unexpected keyword argument 'constructor'
+10:10:22
+10:10:22 dask/backends.py:122: TypeError
+```
+
+cc @rjzamora @galipremsagar @charlesbluca ",0, test from dict backends failure on gpuci see python test from dict backends linux python opt conda envs dask bin backend cudf pytest mark gpu pytest mark parametrize backend def test from dict backends backend lib pytest importorskip backend with config set dataframe backend backend data a b expected lib dataframe data check dd from dict api got dd from dict data npartitions assert eq expected got check from dict classmethod got classmethod got from dict data npartitions dask dataframe io tests test io py dask dataframe core py in from dict return from dict args b a kwargs columns none constructor dtype none orient columns wraps fn def wrapper args kwargs return getattr self dispatch name args kwargs e typeerror from dict got an unexpected keyword argument constructor dask backends py typeerror cc rjzamora galipremsagar charlesbluca ,0
+21229,4697373796.0,IssuesEvent,2016-10-12 09:08:43,caskroom/homebrew-cask,https://api.github.com/repos/caskroom/homebrew-cask,opened,Document binary stanza,documentation,"A contributor [had trouble](https://github.com/fulldecent/corelocationcli/issues/14#issue-182396130) finding documentation of the `binary` stanza. Given that we are deprecating [homebrew-binary](https://github.com/Homebrew/homebrew-binary), we should make it easy for contributors to find examples of `binary` Casks.
+
+---
+
+- [ ] Create `doc/stanzas/binary.md`
+- [ ] Add `binary` Cask examples to [doc/development/adding_a_cask.md](https://github.com/caskroom/homebrew-cask/blob/cc0fdc77c153681e47257c7621ded95a05331f3f/doc/development/adding_a_cask.md)
+- [ ] Add `binary` Cask examples to [doc/cask_language_reference/readme.md](https://github.com/caskroom/homebrew-cask/blob/cc0fdc77c153681e47257c7621ded95a05331f3f/doc/cask_language_reference/readme.md)
+- [ ] Add `binary` Cask examples to [CONTRIBUTING.md](https://github.com/caskroom/homebrew-cask/blob/master/CONTRIBUTING.md)",1.0,"Document binary stanza - A contributor [had trouble](https://github.com/fulldecent/corelocationcli/issues/14#issue-182396130) finding documentation of the `binary` stanza. Given that we are deprecating [homebrew-binary](https://github.com/Homebrew/homebrew-binary), we should make it easy for contributors to find examples of `binary` Casks.
+
+---
+
+- [ ] Create `doc/stanzas/binary.md`
+- [ ] Add `binary` Cask examples to [doc/development/adding_a_cask.md](https://github.com/caskroom/homebrew-cask/blob/cc0fdc77c153681e47257c7621ded95a05331f3f/doc/development/adding_a_cask.md)
+- [ ] Add `binary` Cask examples to [doc/cask_language_reference/readme.md](https://github.com/caskroom/homebrew-cask/blob/cc0fdc77c153681e47257c7621ded95a05331f3f/doc/cask_language_reference/readme.md)
+- [ ] Add `binary` Cask examples to [CONTRIBUTING.md](https://github.com/caskroom/homebrew-cask/blob/master/CONTRIBUTING.md)",0,document binary stanza a contributor finding documentation of the binary stanza given that we are deprecating we should make it easy for contributors to find examples of binary casks create doc stanzas binary md add binary cask examples to add binary cask examples to add binary cask examples to ,0
+12240,3593006089.0,IssuesEvent,2016-02-01 18:04:03,whatwg/streams,https://api.github.com/repos/whatwg/streams,opened,Define valid queuing strategy,documentation queuing strategies,"As an author- and other-spec-facing requirement, we should define valid queuing strategies. I think the requirements are:
+
+- have size() methods and highWaterMark properties.
+- size() is pure (gives the same result for the same input every time)
+- size() never returns NaN, +Infinity, or negative
+- highWaterMark is not NaN and not < 0
+- (maybe) highWaterMark does not change over time?
+- (maybe) size never throws an exception?",1.0,"Define valid queuing strategy - As an author- and other-spec-facing requirement, we should define valid queuing strategies. I think the requirements are:
+
+- have size() methods and highWaterMark properties.
+- size() is pure (gives the same result for the same input every time)
+- size() never returns NaN, +Infinity, or negative
+- highWaterMark is not NaN and not < 0
+- (maybe) highWaterMark does not change over time?
+- (maybe) size never throws an exception?",0,define valid queuing strategy as an author and other spec facing requirement we should define valid queuing strategies i think the requirements are have size methods and highwatermark properties size is pure gives the same result for the same input every time size never returns nan infinity or negative highwatermark is not nan and not maybe highwatermark does not change over time maybe size never throws an exception ,0
+247478,18857814706.0,IssuesEvent,2021-11-12 09:03:38,clementkfj/pe,https://api.github.com/repos/clementkfj/pe,opened,Use 4 instead of iv,severity.Low type.DocumentationBug,"Should state 4 instead of iv, in case user does not understand roman numerals.
+
+
+
+
+
+",1.0,"Use 4 instead of iv - Should state 4 instead of iv, in case user does not understand roman numerals.
+
+
+
+
+
+",0,use instead of iv should state instead of iv in case user does not understand roman numerals ,0
+15294,19523806334.0,IssuesEvent,2021-12-30 01:30:50,tsunamods-codes/7th-Heaven,https://api.github.com/repos/tsunamods-codes/7th-Heaven,closed,SFX settings audio.dat and vgmstream,good first issue Compatibility,"As I understand it now mods that require vgmstream can automatically set it when the developer configures it in their mod, meaning there is no longer a need for a visible UI setting.
+
+In addition 7h honours what is set in the ffnx.toml if one already exists, so I suggest two things.
+
+1 - The sfx setting is removed from 7H_GameDriver_UI.xml, as it can only cause confusion and misconfiguration at this point.
+
+2 - In the same update the 7h installer will replace the ffnx.toml, I did check and by default sfx is actually set to audio.dat as trueodin said. So the issue is coming from when people had previously turned it on.",True,"SFX settings audio.dat and vgmstream - As I understand it now mods that require vgmstream can automatically set it when the developer configures it in their mod, meaning there is no longer a need for a visible UI setting.
+
+In addition 7h honours what is set in the ffnx.toml if one already exists, so I suggest two things.
+
+1 - The sfx setting is removed from 7H_GameDriver_UI.xml, as it can only cause confusion and misconfiguration at this point.
+
+2 - In the same update the 7h installer will replace the ffnx.toml, I did check and by default sfx is actually set to audio.dat as trueodin said. So the issue is coming from when people had previously turned it on.",0,sfx settings audio dat and vgmstream as i understand it now mods that require vgmstream can automatically set it when the developer configures it in their mod meaning there is no longer a need for a visible ui setting in addition honours what is set in the ffnx toml if one already exists so i suggest two things the sfx setting is removed from gamedriver ui xml as it can only cause confusion and misconfiguration at this point in the same update the installer will replace the ffnx toml i did check and by default sfx is actually set to audio dat as trueodin said so the issue is coming from when people had previously turned it on ,0
+135865,18722156811.0,IssuesEvent,2021-11-03 13:01:19,KDWSS/dd-trace-java,https://api.github.com/repos/KDWSS/dd-trace-java,opened,CVE-2019-11808 (Low) detected in ratpack-groovy-1.5.0.jar,security vulnerability,"## CVE-2019-11808 - Low Severity Vulnerability
+ Vulnerable Library - ratpack-groovy-1.5.0.jar
+
+
Ratpack integration for the Groovy programmming language - http://groovy-lang.org
Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/ratpack-1.5/ratpack-1.5.gradle
+
Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/io.ratpack/ratpack-groovy/1.5.0/eab32e0a233cdeeaff73054d8722a1ce2555c6df/ratpack-groovy-1.5.0.jar
+
+Ratpack versions before 1.6.1 generate a session ID using a cryptographically weak PRNG in the JDK's ThreadLocalRandom. This means that if an attacker can determine a small window for the server start time and obtain a session ID value, they can theoretically determine the sequence of session IDs.
+
+
Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/ratpack-1.5/ratpack-1.5.gradle
+
Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/io.ratpack/ratpack-groovy/1.5.0/eab32e0a233cdeeaff73054d8722a1ce2555c6df/ratpack-groovy-1.5.0.jar
+
+Ratpack versions before 1.6.1 generate a session ID using a cryptographically weak PRNG in the JDK's ThreadLocalRandom. This means that if an attacker can determine a small window for the server start time and obtain a session ID value, they can theoretically determine the sequence of session IDs.
+
+
Path to vulnerable library: /frontend/node_modules/webpack/node_modules/loader-utils/package.json,/frontend/node_modules/worker-plugin/node_modules/loader-utils/package.json,/frontend/node_modules/babel-loader/node_modules/loader-utils/package.json
+
+A Regular expression denial of service (ReDoS) flaw was found in Function interpolateName in interpolateName.js in webpack loader-utils 2.0.0 via the resourcePath variable in interpolateName.js.
+
+
+For more information on CVSS3 Scores, click here.
+
+
+
+
+***
+Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",True,"CVE-2022-37599 (Medium) detected in multiple libraries - ## CVE-2022-37599 - Medium Severity Vulnerability
+ Vulnerable Libraries - loader-utils-1.2.3.tgz, loader-utils-1.4.0.tgz, loader-utils-2.0.0.tgz
+
Path to vulnerable library: /frontend/node_modules/webpack/node_modules/loader-utils/package.json,/frontend/node_modules/worker-plugin/node_modules/loader-utils/package.json,/frontend/node_modules/babel-loader/node_modules/loader-utils/package.json
+
+A Regular expression denial of service (ReDoS) flaw was found in Function interpolateName in interpolateName.js in webpack loader-utils 2.0.0 via the resourcePath variable in interpolateName.js.
+
+
+For more information on CVSS3 Scores, click here.
+
+
+
+
+***
+Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",0,cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries loader utils tgz loader utils tgz loader utils tgz loader utils tgz utils for webpack loaders library home page a href path to dependency file frontend package json path to vulnerable library frontend node modules resolve url loader node modules loader utils package json dependency hierarchy build angular tgz root library resolve url loader tgz x loader utils tgz vulnerable library loader utils tgz utils for webpack loaders library home page a href path to dependency file frontend package json path to vulnerable library frontend node modules webpack node modules loader utils package json frontend node modules worker plugin node modules loader utils package json frontend node modules babel loader node modules loader utils package json dependency hierarchy build angular tgz root library worker plugin tgz x loader utils tgz vulnerable library loader utils tgz utils for webpack loaders library home page a href path to dependency file frontend package json path to vulnerable library frontend node modules loader utils package json dependency hierarchy build angular tgz root library x loader utils tgz vulnerable library found in base branch master vulnerability details a regular expression denial of service redos flaw was found in function interpolatename in interpolatename js in webpack loader utils via the resourcepath variable in interpolatename js publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with mend ,0
+1029,25085137798.0,IssuesEvent,2022-11-07 23:02:01,jongfeel/BookReview,https://api.github.com/repos/jongfeel/BookReview,closed,3부 2장 적을 만드는 확실한 방법과 그 예방법,2022 How to Win Friends & Influence People,"### 3부 2장 적을 만드는 확실한 방법과 그 예방법
+
+당신이 틀릴 수도 있다는 것을 인정하면 어떠한 어려운 상황도 마주치지 않게 될 것이다. 모든 논쟁은 중단될 것이고, 다른 사람들로 하여금 당신처럼 공정하고 열린 마음, 넓은 마음을 가지게 할 수 있을 것이다. 다른 사람들로 하여금 그들 역시 틀릴 수도 있다는 사실을 인정하게 만들 수 있을 것이다.
+
+---
+
+규칙 2: 다른 사람의 의견을 존중하라. 절대로 그 사람이 틀렸다고 이야기하지 마라.
+Show respect for the other man’s opinions. Never tell a man he is wrong.",1.0,"3부 2장 적을 만드는 확실한 방법과 그 예방법 - ### 3부 2장 적을 만드는 확실한 방법과 그 예방법
+
+당신이 틀릴 수도 있다는 것을 인정하면 어떠한 어려운 상황도 마주치지 않게 될 것이다. 모든 논쟁은 중단될 것이고, 다른 사람들로 하여금 당신처럼 공정하고 열린 마음, 넓은 마음을 가지게 할 수 있을 것이다. 다른 사람들로 하여금 그들 역시 틀릴 수도 있다는 사실을 인정하게 만들 수 있을 것이다.
+
+---
+
+규칙 2: 다른 사람의 의견을 존중하라. 절대로 그 사람이 틀렸다고 이야기하지 마라.
+Show respect for the other man’s opinions. Never tell a man he is wrong.",1, 적을 만드는 확실한 방법과 그 예방법 적을 만드는 확실한 방법과 그 예방법 당신이 틀릴 수도 있다는 것을 인정하면 어떠한 어려운 상황도 마주치지 않게 될 것이다 모든 논쟁은 중단될 것이고 다른 사람들로 하여금 당신처럼 공정하고 열린 마음 넓은 마음을 가지게 할 수 있을 것이다 다른 사람들로 하여금 그들 역시 틀릴 수도 있다는 사실을 인정하게 만들 수 있을 것이다 규칙 다른 사람의 의견을 존중하라 절대로 그 사람이 틀렸다고 이야기하지 마라 show respect for the other man’s opinions never tell a man he is wrong ,1
+508,8745013393.0,IssuesEvent,2018-12-13 00:37:22,cloudamatic/mu,https://api.github.com/repos/cloudamatic/mu,closed,init script funk on RHEL/CentOS 7,bug in progress workaround needed for other peoples' bugs,"Nagios and 389DS are both known to behave badly, in the sense that issuing a service stop/start/restart does not work reliably. Further examination is needed to figure out the specifics.
+
+Nagios' problem may have more to do with the community cookbook than the package itself.",1.0,"init script funk on RHEL/CentOS 7 - Nagios and 389DS are both known to behave badly, in the sense that issuing a service stop/start/restart does not work reliably. Further examination is needed to figure out the specifics.
+
+Nagios' problem may have more to do with the community cookbook than the package itself.",1,init script funk on rhel centos nagios and are both known to behave badly in the sense that issuing a service stop start restart does not work reliably further examination is needed to figure out the specifics nagios problem may have more to do with the community cookbook than the package itself ,1
+4906,2566012118.0,IssuesEvent,2015-02-08 01:45:20,GoogleCloudPlatform/kubernetes,https://api.github.com/repos/GoogleCloudPlatform/kubernetes,closed,"Reduce vebosity level of ""Failed to get any services: Key not found""",kind/bug priority/P2,"On the minions I get the following error with `sudo journalctl -r`
+
+```
+Aug 12 10:24:53 kubernetes-minion-1 kube-proxy[24148]: E0812 10:24:53.156229 24148 etcd.go:80] Failed to get any services: 100: Key not found (/registry/services) [15]
+Aug 12 10:24:53 kubernetes-minion-1 kube-proxy[24148]: E0812 10:24:53.156216 24148 etcd.go:120] Failed to get the key registry/services: 100: Key not found (/registry/services) [15]
+Aug 12 10:24:53 kubernetes-minion-1 kube-proxy[24148]: I0812 10:24:53.156186 24148 logs.go:39] etcd DEBUG: [recv.success. http://10.245.1.2:4001/v2/keys/registry/services/specs?consistent=true&recursive=false&sorted=true]
+Aug 12 10:24:53 kubernetes-minion-1 kube-proxy[24148]: I0812 10:24:53.156164 24148 logs.go:39] etcd DEBUG: [recv.response.from http://10.245.1.2:4001/v2/keys/registry/services/specs?consistent=true&recursive=false&sorted=true]
+Aug 12 10:24:53 kubernetes-minion-1 kube-proxy[24148]: I0812 10:24:53.155197 24148 logs.go:39] etcd DEBUG: [send.request.to http://10.245.1.2:4001/v2/keys/registry/services/specs?consistent=true&recursive=false&sorted=true | method GET]
+Aug 12 10:24:53 kubernetes-minion-1 kube-proxy[24148]: I0812 10:24:53.155191 24148 logs.go:39] etcd DEBUG: [Connecting to etcd: attempt 1 for keys/registry/services/specs?consistent=true&recursive=false&sorted=true]
+Aug 12 10:24:53 kubernetes-minion-1 kube-proxy[24148]: I0812 10:24:53.155158 24148 logs.go:39] etcd DEBUG: get [registry/services/specs http://10.245.1.2:4001] [%!s(MISSING)]
+```
+
+I checked the result manually with curl:
+```
+$ curl ""http://10.245.1.2:4001/v2/keys/registry/services/specs?consistent=true&recursive=false&sorted=true""
+{""errorCode"":100,""message"":""Key not found"",""cause"":""/registry/services"",""index"":15}
+```
+
+How to reproduce:
+* after destroy the cluster: `vagrant up`
+* create a new replicationController with: `cluster/kubecfg.sh -p 8080:80 run dockerfile/nginx 3 myNginx`
+
+I can list the new replicationController and pods but the containers don't exist on the minions.
+
+I also tried to do `vagrant provision`, I got the same result.
+
+I'm on c6dcfd5 and did a fresh `vagrant up` today after destroyed my previous cluster.
+
+**Edit**: after a few vagrant halt / up the containers are started to work but the error message mentioned above stayed.",1.0,"Reduce vebosity level of ""Failed to get any services: Key not found"" - On the minions I get the following error with `sudo journalctl -r`
+
+```
+Aug 12 10:24:53 kubernetes-minion-1 kube-proxy[24148]: E0812 10:24:53.156229 24148 etcd.go:80] Failed to get any services: 100: Key not found (/registry/services) [15]
+Aug 12 10:24:53 kubernetes-minion-1 kube-proxy[24148]: E0812 10:24:53.156216 24148 etcd.go:120] Failed to get the key registry/services: 100: Key not found (/registry/services) [15]
+Aug 12 10:24:53 kubernetes-minion-1 kube-proxy[24148]: I0812 10:24:53.156186 24148 logs.go:39] etcd DEBUG: [recv.success. http://10.245.1.2:4001/v2/keys/registry/services/specs?consistent=true&recursive=false&sorted=true]
+Aug 12 10:24:53 kubernetes-minion-1 kube-proxy[24148]: I0812 10:24:53.156164 24148 logs.go:39] etcd DEBUG: [recv.response.from http://10.245.1.2:4001/v2/keys/registry/services/specs?consistent=true&recursive=false&sorted=true]
+Aug 12 10:24:53 kubernetes-minion-1 kube-proxy[24148]: I0812 10:24:53.155197 24148 logs.go:39] etcd DEBUG: [send.request.to http://10.245.1.2:4001/v2/keys/registry/services/specs?consistent=true&recursive=false&sorted=true | method GET]
+Aug 12 10:24:53 kubernetes-minion-1 kube-proxy[24148]: I0812 10:24:53.155191 24148 logs.go:39] etcd DEBUG: [Connecting to etcd: attempt 1 for keys/registry/services/specs?consistent=true&recursive=false&sorted=true]
+Aug 12 10:24:53 kubernetes-minion-1 kube-proxy[24148]: I0812 10:24:53.155158 24148 logs.go:39] etcd DEBUG: get [registry/services/specs http://10.245.1.2:4001] [%!s(MISSING)]
+```
+
+I checked the result manually with curl:
+```
+$ curl ""http://10.245.1.2:4001/v2/keys/registry/services/specs?consistent=true&recursive=false&sorted=true""
+{""errorCode"":100,""message"":""Key not found"",""cause"":""/registry/services"",""index"":15}
+```
+
+How to reproduce:
+* after destroy the cluster: `vagrant up`
+* create a new replicationController with: `cluster/kubecfg.sh -p 8080:80 run dockerfile/nginx 3 myNginx`
+
+I can list the new replicationController and pods but the containers don't exist on the minions.
+
+I also tried to do `vagrant provision`, I got the same result.
+
+I'm on c6dcfd5 and did a fresh `vagrant up` today after destroyed my previous cluster.
+
+**Edit**: after a few vagrant halt / up the containers are started to work but the error message mentioned above stayed.",0,reduce vebosity level of failed to get any services key not found on the minions i get the following error with sudo journalctl r aug kubernetes minion kube proxy etcd go failed to get any services key not found registry services aug kubernetes minion kube proxy etcd go failed to get the key registry services key not found registry services aug kubernetes minion kube proxy logs go etcd debug aug kubernetes minion kube proxy logs go etcd debug aug kubernetes minion kube proxy logs go etcd debug aug kubernetes minion kube proxy logs go etcd debug aug kubernetes minion kube proxy logs go etcd debug get i checked the result manually with curl curl errorcode message key not found cause registry services index how to reproduce after destroy the cluster vagrant up create a new replicationcontroller with cluster kubecfg sh p run dockerfile nginx mynginx i can list the new replicationcontroller and pods but the containers don t exist on the minions i also tried to do vagrant provision i got the same result i m on and did a fresh vagrant up today after destroyed my previous cluster edit after a few vagrant halt up the containers are started to work but the error message mentioned above stayed ,0
+449,8237150458.0,IssuesEvent,2018-09-10 00:53:33,planningcenter/developers,https://api.github.com/repos/planningcenter/developers,closed,Having issues updating an avatar after uploading a new file.,People question,"**Affected Applications**
+people
+
+**Describe the bug**
+I am writing a javascript with a node/express api layer. I am able to get the file to server, then upload it to https://upload.planningcenteronline.com/v2/files. I get the following response from the upload:
+
+{""data"":[{""type"":""File"",""attributes"":{""source_ip"":""174.105.160.85"",""md5"":""d56c8a33d952a2e4a7062825aa451e9e"",""content_type"":""image/jpeg"",""file_size"":523602,""name"":""828ee46323feb4b93441470542a90298"",""expires_at"":""2018-08-29T06:06:09Z""},""id"":""us3-b9f2365e-ad82-48c9-ae1c-1f81e7800d64""}]}
+
+In turn I submit a patch request with the following:
+{""method"":""PATCH"",""headers"":{""Authorization"":""Basic XXXXXXXXXXXXX_REMOVED_XXXXXXXXX""},""uri"":""https://api.planningcenteronline.com/people/v2/people/25591436"",""json"":true,""body"":{""data"":{""type"":""Person"",""id"":""25591436"",""attributes"":{""avatar"":""us3-b9f2365e-ad82-48c9-ae1c-1f81e7800d64""}}}}
+
+The request takes quite a while then returns with ""Failed to read file input: timeout"" and the avatar is not changed in planning center.
+
+*PLEASE DO NOT INCLUDE YOUR CLIENT ID, APPLICATION ID, OR SECRET IN THIS ISSUE. WE WILL REVOKE YOUR TOKEN IF YOU DO.*
+
+**To Reproduce**
+Steps to reproduce the behavior:
+1. upload a file using the api
+2. use the api to update the persons avatar using the files id
+
+**Expected behavior**
+Avatar should be updated
+
+**Screenshots**
+If applicable, add screenshots to help explain your problem.
+
+**Context (please complete the following information):**
+- Endpoint: https://api.planningcenteronline.com/people/v2/people/
+- Language: javascript
+- Authentication: Personal Access Token
+
+**Additional context**
+Add any other context about the problem here.
+",1.0,"Having issues updating an avatar after uploading a new file. - **Affected Applications**
+people
+
+**Describe the bug**
+I am writing a javascript with a node/express api layer. I am able to get the file to server, then upload it to https://upload.planningcenteronline.com/v2/files. I get the following response from the upload:
+
+{""data"":[{""type"":""File"",""attributes"":{""source_ip"":""174.105.160.85"",""md5"":""d56c8a33d952a2e4a7062825aa451e9e"",""content_type"":""image/jpeg"",""file_size"":523602,""name"":""828ee46323feb4b93441470542a90298"",""expires_at"":""2018-08-29T06:06:09Z""},""id"":""us3-b9f2365e-ad82-48c9-ae1c-1f81e7800d64""}]}
+
+In turn I submit a patch request with the following:
+{""method"":""PATCH"",""headers"":{""Authorization"":""Basic XXXXXXXXXXXXX_REMOVED_XXXXXXXXX""},""uri"":""https://api.planningcenteronline.com/people/v2/people/25591436"",""json"":true,""body"":{""data"":{""type"":""Person"",""id"":""25591436"",""attributes"":{""avatar"":""us3-b9f2365e-ad82-48c9-ae1c-1f81e7800d64""}}}}
+
+The request takes quite a while then returns with ""Failed to read file input: timeout"" and the avatar is not changed in planning center.
+
+*PLEASE DO NOT INCLUDE YOUR CLIENT ID, APPLICATION ID, OR SECRET IN THIS ISSUE. WE WILL REVOKE YOUR TOKEN IF YOU DO.*
+
+**To Reproduce**
+Steps to reproduce the behavior:
+1. upload a file using the api
+2. use the api to update the persons avatar using the files id
+
+**Expected behavior**
+Avatar should be updated
+
+**Screenshots**
+If applicable, add screenshots to help explain your problem.
+
+**Context (please complete the following information):**
+- Endpoint: https://api.planningcenteronline.com/people/v2/people/
+- Language: javascript
+- Authentication: Personal Access Token
+
+**Additional context**
+Add any other context about the problem here.
+",1,having issues updating an avatar after uploading a new file affected applications people describe the bug i am writing a javascript with a node express api layer i am able to get the file to server then upload it to i get the following response from the upload data in turn i submit a patch request with the following method patch headers authorization basic xxxxxxxxxxxxx removed xxxxxxxxx uri the request takes quite a while then returns with failed to read file input timeout and the avatar is not changed in planning center please do not include your client id application id or secret in this issue we will revoke your token if you do to reproduce steps to reproduce the behavior upload a file using the api use the api to update the persons avatar using the files id expected behavior avatar should be updated screenshots if applicable add screenshots to help explain your problem context please complete the following information endpoint language javascript authentication personal access token additional context add any other context about the problem here ,1
+792431,27959993578.0,IssuesEvent,2023-03-24 15:00:38,asastats/channel,https://api.github.com/repos/asastats/channel,closed,NFTs listed on Rand no longer show in Stats,bug high priority addressed,"Bug description:
+
+I noticed that an NFT I listed on Rand wasn't showing up in my view.
+
+I updated the price from its initial listing so to make sure that wasn't the issue, I listed a new NFT without changing the price but it too does not show up. The caveat to this is that it did show up for a minute or two just after listing the item. Refreshing after a few minutes and that same listing no longer shows up in Stats.
+
+I suspect the change from https://github.com/asastats/channel/issues/500 might be the culprit.
+",1.0,"NFTs listed on Rand no longer show in Stats - Bug description:
+
+I noticed that an NFT I listed on Rand wasn't showing up in my view.
+
+I updated the price from its initial listing so to make sure that wasn't the issue, I listed a new NFT without changing the price but it too does not show up. The caveat to this is that it did show up for a minute or two just after listing the item. Refreshing after a few minutes and that same listing no longer shows up in Stats.
+
+I suspect the change from https://github.com/asastats/channel/issues/500 might be the culprit.
+",0,nfts listed on rand no longer show in stats bug description i noticed that an nft i listed on rand wasn t showing up in my view i updated the price from its initial listing so to make sure that wasn t the issue i listed a new nft without changing the price but it too does not show up the caveat to this is that it did show up for a minute or two just after listing the item refreshing after a few minutes and that same listing no longer shows up in stats i suspect the change from might be the culprit ,0
+316509,27167019538.0,IssuesEvent,2023-02-17 16:06:08,unifyai/ivy,https://api.github.com/repos/unifyai/ivy,reopened,Fix nn.test_tensorflow_conv3d_transpose,TensorFlow Frontend Sub Task Failing Test,"| | |
+|---|---|
+|tensorflow|
+|torch|
+|numpy|
+|jax|
+
+
+
+Not found
+
+Not found
+
+
+Not found
+
+Not found
+
+
+Not found
+
+Not found
+
+
+Not found
+
+Not found
+
+",1.0,"Fix nn.test_tensorflow_conv3d_transpose - | | |
+|---|---|
+|tensorflow|
+|torch|
+|numpy|
+|jax|
+
+
+
+Not found
+
+Not found
+
+
+Not found
+
+Not found
+
+
+Not found
+
+Not found
+
+
+Not found
+
+Not found
+
+",0,fix nn test tensorflow transpose tensorflow img src torch img src numpy img src jax img src not found not found not found not found not found not found not found not found ,0
+725651,24969852108.0,IssuesEvent,2022-11-01 23:24:06,brave/brave-browser,https://api.github.com/repos/brave/brave-browser,closed,Crash from Google Play Console inside RetentionNotificationPublisher,crash priority/P2 QA/No release-notes/exclude OS/Android,"```
+Exception java.lang.RuntimeException: Unable to start receiver org.chromium.chrome.browser.notifications.retention.RetentionNotificationPublisher: java.lang.NullPointerException: Attempt to invoke virtual method 'android.content.pm.PackageManager android.content.Context.getPackageManager()' on a null object reference
+ at android.app.ActivityThread.handleReceiver (ActivityThread.java:4345)
+ at android.app.ActivityThread.access$1700 (ActivityThread.java:286)
+ at android.app.ActivityThread$H.handleMessage (ActivityThread.java:2153)
+ at android.os.Handler.dispatchMessage (Handler.java:106)
+ at android.os.Looper.loop (Looper.java:254)
+ at android.app.ActivityThread.main (ActivityThread.java:8219)
+ at java.lang.reflect.Method.invoke (Method.java)
+ at com.android.internal.os.RuntimeInit$MethodAndArgsCaller.run (RuntimeInit.java:612)
+ at com.android.internal.os.ZygoteInit.main (ZygoteInit.java:1006)
+Caused by java.lang.NullPointerException: Attempt to invoke virtual method 'android.content.pm.PackageManager android.content.Context.getPackageManager()' on a null object reference
+ at org.chromium.chrome.browser.download.home.list.holder.CardDividerTopViewHolder.b (CardDividerTopViewHolder.java:14)
+ at org.chromium.chrome.browser.notifications.retention.RetentionNotificationPublisher.onReceive (RetentionNotificationPublisher.java:139)
+ at android.app.ActivityThread.handleReceiver (ActivityThread.java:4329)
+```
+
+I wasn't able to replicate it, but it looks like it cannot execute a retention notification for some reason. It's good to just catch a `NullPointerException` to prevent a whole browser crash.",1.0,"Crash from Google Play Console inside RetentionNotificationPublisher - ```
+Exception java.lang.RuntimeException: Unable to start receiver org.chromium.chrome.browser.notifications.retention.RetentionNotificationPublisher: java.lang.NullPointerException: Attempt to invoke virtual method 'android.content.pm.PackageManager android.content.Context.getPackageManager()' on a null object reference
+ at android.app.ActivityThread.handleReceiver (ActivityThread.java:4345)
+ at android.app.ActivityThread.access$1700 (ActivityThread.java:286)
+ at android.app.ActivityThread$H.handleMessage (ActivityThread.java:2153)
+ at android.os.Handler.dispatchMessage (Handler.java:106)
+ at android.os.Looper.loop (Looper.java:254)
+ at android.app.ActivityThread.main (ActivityThread.java:8219)
+ at java.lang.reflect.Method.invoke (Method.java)
+ at com.android.internal.os.RuntimeInit$MethodAndArgsCaller.run (RuntimeInit.java:612)
+ at com.android.internal.os.ZygoteInit.main (ZygoteInit.java:1006)
+Caused by java.lang.NullPointerException: Attempt to invoke virtual method 'android.content.pm.PackageManager android.content.Context.getPackageManager()' on a null object reference
+ at org.chromium.chrome.browser.download.home.list.holder.CardDividerTopViewHolder.b (CardDividerTopViewHolder.java:14)
+ at org.chromium.chrome.browser.notifications.retention.RetentionNotificationPublisher.onReceive (RetentionNotificationPublisher.java:139)
+ at android.app.ActivityThread.handleReceiver (ActivityThread.java:4329)
+```
+
+I wasn't able to replicate it, but it looks like it cannot execute a retention notification for some reason. It's good to just catch a `NullPointerException` to prevent a whole browser crash.",0,crash from google play console inside retentionnotificationpublisher exception java lang runtimeexception unable to start receiver org chromium chrome browser notifications retention retentionnotificationpublisher java lang nullpointerexception attempt to invoke virtual method android content pm packagemanager android content context getpackagemanager on a null object reference at android app activitythread handlereceiver activitythread java at android app activitythread access activitythread java at android app activitythread h handlemessage activitythread java at android os handler dispatchmessage handler java at android os looper loop looper java at android app activitythread main activitythread java at java lang reflect method invoke method java at com android internal os runtimeinit methodandargscaller run runtimeinit java at com android internal os zygoteinit main zygoteinit java caused by java lang nullpointerexception attempt to invoke virtual method android content pm packagemanager android content context getpackagemanager on a null object reference at org chromium chrome browser download home list holder carddividertopviewholder b carddividertopviewholder java at org chromium chrome browser notifications retention retentionnotificationpublisher onreceive retentionnotificationpublisher java at android app activitythread handlereceiver activitythread java i wasn t able to replicate it but it looks like it cannot execute a retention notification for some reason it s good to just catch a nullpointerexception to prevent a whole browser crash ,0
+234202,25806038750.0,IssuesEvent,2022-12-11 12:20:09,SmartBear/readyapi-swagger-assertion-plugin,https://api.github.com/repos/SmartBear/readyapi-swagger-assertion-plugin,closed,WS-2021-0170 (High) detected in spring-core-4.1.6.RELEASE.jar - autoclosed,security vulnerability,"## WS-2021-0170 - High Severity Vulnerability
+ Vulnerable Library - spring-core-4.1.6.RELEASE.jar
+
+
+
+In spring-core, versions v4.0.0.RC1 through v4.1.8.RELEASE, and v4.2.0.RC1 through v4.2.2.RELEASE, allow arbitrary code execution due to the `SerializableTypeWrapper` class, which allows invocation of any method on the Java classpath through its `MethodInvokeTypeProvider` method. `SerializableTypeWrapper`, implementing the `Serializable` interface, can be included in a maliciously crafted serialized object and be used to eventually invoke `Runtime.getRuntime.exec()`, if `commons-collections` is included in the classpath.
+
+
+
+In spring-core, versions v4.0.0.RC1 through v4.1.8.RELEASE, and v4.2.0.RC1 through v4.2.2.RELEASE, allow arbitrary code execution due to the `SerializableTypeWrapper` class, which allows invocation of any method on the Java classpath through its `MethodInvokeTypeProvider` method. `SerializableTypeWrapper`, implementing the `Serializable` interface, can be included in a maliciously crafted serialized object and be used to eventually invoke `Runtime.getRuntime.exec()`, if `commons-collections` is included in the classpath.
+
+
+
+
+
+
+",0,ws high detected in spring core release jar autoclosed ws high severity vulnerability vulnerable library spring core release jar spring core library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository org springframework spring core release spring core release jar dependency hierarchy ready api soapui pro jar root library ready api db jar spring jdbc release jar x spring core release jar vulnerable library found in base branch master vulnerability details in spring core versions through release and through release allow arbitrary code execution due to the serializabletypewrapper class which allows invocation of any method on the java classpath through its methodinvoketypeprovider method serializabletypewrapper implementing the serializable interface can be included in a maliciously crafted serialized object and be used to eventually invoke runtime getruntime exec if commons collections is included in the classpath publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope changed impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution org springframework spring core release release ,0
+488,8443357669.0,IssuesEvent,2018-10-18 15:23:53,openstates/openstates,https://api.github.com/repos/openstates/openstates,closed,PR: Update committee PDF parsing,people priority.low,"PR committee parsing fails currently. Although their PDFs also aren't up to date for the current session.
+
+Eg, scraping fails for http://senado.pr.gov/comisiones/Pages/ComposicionComisionesPermanentes.aspx",1.0,"PR: Update committee PDF parsing - PR committee parsing fails currently. Although their PDFs also aren't up to date for the current session.
+
+Eg, scraping fails for http://senado.pr.gov/comisiones/Pages/ComposicionComisionesPermanentes.aspx",1,pr update committee pdf parsing pr committee parsing fails currently although their pdfs also aren t up to date for the current session eg scraping fails for ,1
+43876,13040709778.0,IssuesEvent,2020-07-28 19:00:30,LevyForchh/fuzzbench,https://api.github.com/repos/LevyForchh/fuzzbench,opened,CVE-2020-1747 (High) detected in PyYAML-5.3.tar.gz,security vulnerability,"## CVE-2020-1747 - High Severity Vulnerability
+ Vulnerable Library - PyYAML-5.3.tar.gz
+
+
+
+A vulnerability was discovered in the PyYAML library in versions before 5.3.1, where it is susceptible to arbitrary code execution when it processes untrusted YAML files through the full_load method or with the FullLoader loader. Applications that use the library to process untrusted input may be vulnerable to this flaw. An attacker could use this flaw to execute arbitrary code on the system by abusing the python/object/new constructor.
+
+
+
+A vulnerability was discovered in the PyYAML library in versions before 5.3.1, where it is susceptible to arbitrary code execution when it processes untrusted YAML files through the full_load method or with the FullLoader loader. Applications that use the library to process untrusted input may be vulnerable to this flaw. An attacker could use this flaw to execute arbitrary code on the system by abusing the python/object/new constructor.
+
+
+
+
+
+
+
+",0,cve high detected in pyyaml tar gz cve high severity vulnerability vulnerable library pyyaml tar gz yaml parser and emitter for python library home page a href path to dependency file tmp ws scm fuzzbench requirements txt path to vulnerable library tmp ws scm fuzzbench requirements txt tmp ws scm fuzzbench docker benchmark runner requirements txt dependency hierarchy pytype tar gz root library x pyyaml tar gz vulnerable library found in head commit a href vulnerability details a vulnerability was discovered in the pyyaml library in versions before where it is susceptible to arbitrary code execution when it processes untrusted yaml files through the full load method or with the fullloader loader applications that use the library to process untrusted input may be vulnerable to this flaw an attacker could use this flaw to execute arbitrary code on the system by abusing the python object new constructor publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails a vulnerability was discovered in the pyyaml library in versions before where it is susceptible to arbitrary code execution when it processes untrusted yaml files through the full load method or with the fullloader loader applications that use the library to process untrusted input may be vulnerable to this flaw an attacker could use this flaw to execute arbitrary code on the system by abusing the python object new constructor vulnerabilityurl ,0
+104299,16613588391.0,IssuesEvent,2021-06-02 14:16:52,Thanraj/linux-4.1.15,https://api.github.com/repos/Thanraj/linux-4.1.15,opened,CVE-2016-4557 (High) detected in linux-stable-rtv4.1.33,security vulnerability,"## CVE-2016-4557 - High Severity Vulnerability
+ Vulnerable Library - linux-stable-rtv4.1.33
+
+
+The replace_map_fd_with_map_ptr function in kernel/bpf/verifier.c in the Linux kernel before 4.5.5 does not properly maintain an fd data structure, which allows local users to gain privileges or cause a denial of service (use-after-free) via crafted BPF instructions that reference an incorrect file descriptor.
+
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",True,"CVE-2016-4557 (High) detected in linux-stable-rtv4.1.33 - ## CVE-2016-4557 - High Severity Vulnerability
+ Vulnerable Library - linux-stable-rtv4.1.33
+
+
+The replace_map_fd_with_map_ptr function in kernel/bpf/verifier.c in the Linux kernel before 4.5.5 does not properly maintain an fd data structure, which allows local users to gain privileges or cause a denial of service (use-after-free) via crafted BPF instructions that reference an incorrect file descriptor.
+
+
+
+
+
+```
+
+## Give your JavaScript Slot Machine some Stylish Images
+
+现在给我们的老虎机加点图片。
+
+我们已经为你准备好了图片`images`,我们可以通过不同的索引来获取每个图片。
+
+现在让我们设置第一个老虎机根据随机数来显示一张图片:
+
+`$($('.slot')[0]).html('');`
+
+任务:设置所有的老虎机根据随机数来显示对应的图片,最后点击RUN。
+
+```
+
+
+
+
+
+
+
FCC Slot Machine
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+```
+
+很遗憾的是。。
+
+墙外的 FCC 已经没了老虎机这个项目了。。
+
+不知道去哪了。。",0,freecodecamp basic javascript freecodecamp basic javascript 写在前面: 我曾经在刚接触前端的时候刷过这一套题,不过当时只是为了速度和闯关的快感看着答案写,结果 javascript 只是学了个半斤八两。所以不要求快,多看看 mdn 和 javascript 高级程序设计吧。要沉下心来学习不要浮躁。要不然永远只能成为前端爱好者。 comment your javascript code 注释的代码块在javascript之中是不会运行的。注释是一个非常好的方式让你自己以及其他人明白这段代码是怎么运行的。 javascript中的注释方式有以下两种: 使用 来告诉javascript来忽略当前行的代码 this is an in line comment 你也可以使用多行注释来注释你的代码,以 开始,用 来结束,就像下面这样: this is a multi line comment 最佳实践 你应该给你写的代码添加注释,来让你的代码看起来更加地清晰易懂。良好的注释能够清晰地传达你写的代码的意图—对于那些读你的代码的人来说 以及 未来你看到它的时候,还能理解这段代码的意图。 任务 尝试创建这两种类型的注释。 this is an in line comment this is an in line comment declare javascript variables 在计算机科学中 data 数据 就是一切,因为它对于计算机的意义重大。javascript提供七种不同的data types 数据类型 ,它们是 undefined (未定义) null (空) boolean (布尔型) string (字符串) symbol 符号 number (数字) and object (对象)。 举个例子 计算机能够分辨不同的数字 例如数字 和 strings , 和 dog 或 cats 都是字母的集合。 计算机能够精确地操作数字, 但是对于字符串却无能为力。 variables (变量)允许计算机以一种动态的形式来存储和操作数据,通过操作指向数据的指针而不是数据本身来避免了内存泄露,以上的七种数据类型都可以存储到一个变量(variable)中。 variables 非常类似于你在数学中使用的x y变量 这意味着它们都是以一个简单命名的名字来代替我们赋值给它的数据。计算机中的 variables (变量)与数学中的变量不同的是,计算机可以在不同的时间存储不同类型的变量。 通过在变量的前面使用关键字 var ,我们告诉 javascript 来创建或者 declare(声明)一个变量 就像这样: var ourname 上面代码的意思是创建一个名为 ourname 的 variable (变量),在javascript中我们使用分号来结束一段声明。 variable (变量)的名字可以由数字、字母、 或者 组成,但是不能包含空格或者以数字为首。 任务 使用 var 关键字来创建一个名为 myname 的变量。 提示 如果遇到困难了,请看下 ourname 的例子是怎么写的。 举例 var ourname 请在这条注释以下定义 myname 变量 var myname storing values with the equal operator 在javascript中,你可以通过assignment 分配 操作符把一个值存储到变量中。 myvariable 把 number 数字 赋给变量 myvariable 。 赋值过程是从右到左进行的。所有 操作符右边的值都会被赋到左边的变量。 myvar mynum myvar 数值 被赋给变量 myvar 中, 然后变量 myvar 又赋给变量 mynum ,这样子 mynum 变量中的值也是 了。 任务 把数值 赋给变量 a 。 把变量 a 中的内容赋给变量 b 。 初始化变量 var a var b 请把你的代码写在这条注释以下 a b a initializing variables with the equal operator 通常地我们会在initialize开始声明变量的时候就会给变量赋一个初始值。 var myvar 创建一个名为 myvar 的变量并指定一个初始值 。 任务 通过关键字 var 定义一个变量 a 并且给它一个初始值 。 举例 var ourvar 请把你的代码写在这条注释以下 var a understanding uninitialized variables 当 javascript 中的变量被声明的时候,程序内部会给它一个初始值 undefined 。当你对一个值为 undefined 的变量进行运算操作的时候,算出来的结果将会是 nan , nan 的意思是 not a number 。当你用一个没有 定义 的变量来做字符串连接操作的时候,它会如实的输出 undefined 。 任务 a 、 b 、 c ,并且分别给他们赋值: 、 、 i am a ,这样它们就不会是 undefined (未定义的了)。 initialize these three variables var a var b var c i am a do not change code below this line a a b b c c string understanding case sensitivity in variables 在 javascript 中所有的变量都是大小写敏感的。这意味着你要区别对待大写字母和小写字母。 myvar 与 myvar 和 myvar 是截然不同的变量。这就有可能导致多个截然不同的变量却有着有相似的名字。正是由于以上原因所以强烈地建议你 不要 使用这一特性。(以免给自己带来麻烦) 最佳实践 使用 驼峰命名法 来书写一个 javascript 变量,在 驼峰命名法 中,变量名的第一个单词的首写字母小写,后面的单词的第一个字母大写。 初始化变量 var studlycapvar var propercamelcase var titlecaseover 给变量赋值 studlycapvar propercamelcase a string titlecaseover add two numbers with javascript 现在让我们来尝试在javascript中做加法运算。 javascript 中使用 号来让两个数字执行加法运算。 举例 myvar 等于 任务 改变数字 让变量 sum 的值为 var sum subtract one number from another with javascript 我们也可以在 javascript 中执行减法运算。 javascript 中使用 来做减法运算。 举例 myvar 等于 任务 改变数字 让变量 difference 的值为 。 var difference multiply two numbers with javascript 我们也可在 javascript 中使用乘法运算。 javascript 使用这个 符号来让两个数字相乘。 举例 myvar 把 赋值给 myvar 任务 改变数值 来让变量 product 的值等于 。 var product divide one number by another with javascript 我们可以在 javascript 中做除法运算。 javascript 中使用 符号做除法运算。 举例 myvar 把 赋值给 myvar 任务 改变数值 来让变量 quotient 的值等于 。 var quotient increment a number with javascript 使用 ,我们可以很容易地对变量进行自增或者 运算。 i 等效于 i i 提示 i 这种写法,省去了书写 符号的必要。 任务 重写代码,使用 来对变量 myvar 进行自增操作。 var myvar 请只修改这条注释以下的代码 myvar decrement a number with javascript 使用自减符号 ,你可以很方便地对一个变量执行 自减 或者减一操作。 i 等效于 i i 提示 i 这种写法,省去了书写等号的必要。 任务 重写代码,使用 符号对 myvar 执行自减操作。 var myvar 请只修改这条注释以下的代码 myvar create decimal numbers with javascript 我们也可以把小数存储到变量中。小数也被称作 浮点数 。 提示 不是所有的实数都可以用 浮点数 来表示。因为可能存在四舍五入的错误, 任务 创建一个变量 mydecimal 并给它复制一个浮点数。 e g 。 var ourdecimal 请把你的代码写在这条注释以下 var mydecimal multiply two decimals with javascript 在 javascript 中,你也可以用小数进行计算,就像整数一样。 让我们把两个小数相乘,并得到它们相乘的结果(product)。 任务 改变 的数值让变量 product 的值等于 。 var product divide one decimal by another with javascript 现在让我们来用一个小数做除法操作。 任务 改变数值 的值让变量 quotient 的值等于 var quotient finding a remainder in javascript 现在我们用 运算符来取余。 举例 用法 在数学中,看一个数是奇数还是偶数, 。 is 奇数 is 偶数 任务 使用 操作符, ,并把余数赋给变量remainder。 请只修改这条注释以下的代码 var remainder assignment with plus equals 在编程当中,通常通过赋值来修改变量的内容。请记住,先计算 右边,然后把计算出来的结果赋给左边。 myvar myvar 以上是最常见的运算赋值语句,先运算、再赋值。 还有一类操作符是一步到位既做运算也赋值的。 这类操作符的其中一种就是 运算符。 myvar 也是把数值 加到变量 myvar 上。 任务 使用 操作符实现同样的效果。 var a var b var c 请只修改这条注释以下的代码 a b c assignment with minus equals 与 操作符类似, 操作符用来对一个变量进行减法赋值操作。 myvar myvar 将会从变量 myvar 中减去数值 。也可以写成这种形式: myvar 任务 使用 操作符实现同样的效果。 var a var b var c 请只修改这条注释以下的代码 a b c assignment with times equals 操作符是让变量与一个数相乘并赋值。 myvar myvar 将会把变量 myvar 与数值 相乘。也可以写作这样的形式 myvar 任务 使用 操作符实现同样的效果。 var a var b var c 请只修改这条注释以下的代码 a b c assignment with divided by equals 操作符是让变量与另一个数相除并赋值。 myvar myvar 会把变量 myvar 的值除于 。等价于 myvar 任务 使用 操作符实现同样的效果。 var a var b var c 请只修改这条注释以下的代码 a b c convert celsius to fahrenheit 为了测试你的学习效果,我们来做一个摄氏度转华氏度的小应用。 从 celsius 摄氏度转换为 fahrenheit 华氏度的算法是: , 。 创建一个变量 fahrenheit ,然后计算出摄氏度对应的华氏度。 function convert celsius 请把你的代码写在这条注释以下 fahrenheit celsius 请把你的代码写在这条注释以上 return fahrenheit convert 你可以修改这一行来测试你的代码 declare string variables 先前我们使用过的代码: var myname your name your name 被称作 字符串。 字符串是用单或双引号包裹起来的一连串的零个或多个字符。 任务 创建两个新的 字符串 变量: myfirstname 和 mylastname 分别为它们赋上你的姓和名的值。 举例 var firstname alan var lastname turing 请把你的代码写在这条注释以下 var myfirstname frankie var mylastname tang escaping literal quotes in strings 当你定义一个字符串必须要用单引号或双引号来包裹它。那么当你需要在字符串中使用一个 或者 时该怎么办呢 在 javascript 中,你可以通过在引号前面使用 反斜杠 来转义引号。 var samplestr alan said peter is learning javascript 这标志着提醒 javascript 单引号或双引号并不是字符串的结尾,而是出现在字符串内的字符。所以,如果你要打印字符串到控制台,你将得到: alan said peter is learning javascript 任务 使用 反斜杠 将一个字符串赋值给变量 mystr ,以便如果你要打印到控制台,你会看到: i am a double quoted string inside double quotes var mystr i am a double quoted string inside double quotes 请修改这一行 quoting strings with single quotes 在 javascript 中的 字符串 要用单引号或双引号来包裹它,只要你在开始和结束都使用相同类型的引号,单引号和双引号的功能在javascript中是相同的。 this string has double quotes in it 当我们需要在字符串中使用与开头结尾相同的引号时,我们需要对引号进行 转义 。如果你有很多双引号的字符串,使用转义字符可能导致难以阅读。这时候可以使用单引号。 this string has double quotes in it and probably lots of them 任务 更改mystr字符串的双引号为单引号,并移除转义符号。 var mystr link escape sequences in strings 字符串中的转义序列 引号不是字符串中唯一的可以被转义字符。下面是常见的转义序列列表 code output 单引号 双引号 反斜杠符 n 换行符 r 回车符 t 制表符 b 退格符 f 换页符 注意,如果你想要显示一个反斜杠就必须要转义它。 任务 按照下面的顺序,并且用空格来分割: 反斜杠 制表符 退格符 回车符 换行符 并把它赋值给变量 mystr var mystr t b r n 请修改这一行 concatenating strings with plus operator 在 javascript 中,当 操作符与 字符串 一起使用的时候,它被称作 连接 操作符。你可以通过和其他字符串连接 来创建一个新的字符串。 举个例子 my name is alan i concatenate 注意 当心空格。连接操作不会添加两个字符串之外的空格,所以想加上空格的话,你需要自己在字符串里面添加。 任务 使用 操作,把字符串 this is the start 和 this is the end 连接起来并赋值给变量 mystr 。 举例 var ourstr i come first i come second 请只修改这条注释以下的代码 var mystr this is the start this is the end concatenating strings with the plus equals operator 我们还可以使用 运算符来 连接 字符串到现有字符串的结尾。对于那些非常长的字符串来说,这一操作是非常有用的。 注意 当心空格。连接操作不会添加两个字符串外面的空格,所以如果想要加上空格的话,你需要自己在字符串里面添加。 任务 通过使用 操作符来连接这两个字符串 this is the first sentence 和 this is the second sentence 并赋给变量 mystr 。 举例 var ourstr i come first ourstr i come second 请只修改这条注释以下的代码 var mystr this is the first sentence mystr this is the second sentence constructing strings with variables 有时候你需要创建一个填字风格的字符串。 通过使用连接运算符 ,你可以插入一个或多个变量来组成一个字符串。 任务 把你的名字赋值给变量 myname ,然后把变量 myname 插入到字符串 my name is 和 and i am swell 之间,并把连接后的结果赋值给变量 mystr 。 举例 var ourname free code camp var ourstr hello our name is ourname how are you 请只修改这条注释以下的代码 var myname tang kalun var mystr my name is myname and i am swell appending variables to strings 我们不仅可以创建出多行的字符串,还可以使用加等号 运算符来追加变量到字符串上。 任务 设置变量 someadjective 的值,并使用 运算符把它追加到变量 mystr 上。 举例 var anadjective awesome var ourstr free code camp is ourstr anadjective 请只修改这条注释以下的代码 var someadjective a happy thing var mystr learning to code is mystr someadjective find the length of a string 你可以通过在字符串变量或字符串后面写上 length 来获得字符串变量 字符串 值的长度。 alan peter length 例如,我们创建了一个变量 var firstname charles ,我们就可以通过使用 firstname length 来获得 charles 字符串的长度。 任务 使用 length 属性来获得变量 lastname 的长度,并把它赋值给变量 lastnamelength 。 举例 var firstnamelength var firstname ada firstnamelength firstname length 初始化变量 var lastnamelength var lastname lovelace 请只修改这条注释以下的代码 lastnamelength lastname length use bracket notation to find the first character in a string 叫中括号, 叫大括号, 叫小括号。 javascript中只有字符串类型,没有字符类型。那么如何获取到字符串中的某个字符呢? 我们通过 来获得对应的字符。 大多数现代编程语言,如javascript, 。 ,这被称为 基于零 的索引。 例如 在单词 charles c ,所以在 var firstname charles 中,你可以使用 firstname 来获得第一个位置上的字符。 任务 使用 来得到变量 lastname 中的第一个字符,并赋给变量 firstletteroflastname 。 提示 如果你遇到困难了,不妨看看变量 firstletteroffirstname 是如何赋值的。 举例 var firstletteroffirstname var firstname ada firstletteroffirstname firstname 初始化变量 var firstletteroflastname var lastname lovelace 请只修改这条注释以下的代码 firstletteroflastname lastname understand string immutability 理解字符串的不可变性!当你搞懂不可变性后immutable js对于你就是小菜一碟了。 在 javascript 中, 字符串 的值是 不可变的,这意味着一旦字符串被创建就不能被改变。 例如,下面的代码: var mystr bob mystr j 是不会把变量 mystr 的值改变成 job 的,因为变量 mystr 是不可变的。注意,这 并不 意味着 mystr 永远不能被改变,只是字符串字面量 string literal 的各个字符不能被改变。改变 mystr 中的唯一方法是重新给它赋一个值,就像这样: var mystr bob mystr job 任务 把 mystr 的值改为 hello world 。 初始化变量 var mystr jello world 请只修改这条注释以下的代码 原来是mystr h console log mystr jello world mystr hello world 请修改这一行 use bracket notation to find the nth character in a string 你也可以使用 来获得一个字符串中的其他位置的字符。 请记住,程序是从 开始计数,所以获取第一个字符实际上是 。 任务 让我们使用 ,把 lastname 变量的第三个字符赋值给 thirdletteroflastname 。 提示 如果你遇到困难了,看看 secondletteroffirstname 变量是如何做的。 举例 var firstname ada var secondletteroffirstname firstname 初始化变量 var lastname lovelace 请只修改这条注释以下的代码 var thirdletteroflastname lastname use bracket notation to find the last character in a string 为了得到一个字符串的最后一个字符,你可以用 。 例如,在 var firstname charles 中,你可以这样操作 firstname 来得到字符串的最后的一个字符。 任务 使用 来取得 lastname 变量中的最后一个字符。 提示 如果你遇到困难了,不妨看看在 lastletteroffirstname 变量上是怎么做的。 举例 var firstname ada var lastletteroffirstname firstname 初始化变量 var lastname lovelace 请只修改这条注释以下的代码 var lastletteroflastname lastname use bracket notation to find the nthtolast character in a string 我们既可以获取字符串的最后一个字符,也可以用获取字符串的倒数第n个字符。 例如,你可以这样 firstname 操作来获得 var firstname charles 字符串中的倒数第三个字符。 任务 使用 来获得 lastname 字符串中的倒数第二个字符。 提示 如果你遇到困难了,不妨看看 thirdtolastletteroffirstname 变量是如何做到的。 举例 var firstname ada var thirdtolastletteroffirstname firstname 初始化变量 var lastname lovelace 请只修改这条注释以下的代码 var secondtolastletteroflastname lastname word blanks 填词造句 现在,我们来用字符串的相关知识实现一个造句函数。 通过使用提供的变量参数:名词 mynoun 、形容词 myadjective 、动词 myverb 、副词 myadverb ,来创建一个新的句子 result 。 请注意,在英文中,句中的单词是必须用空格来分隔的 举个例子,如果名词为 dog ,形容词为 big ,动词为 run ,副词为 quickly ,那么函数返回值为 dog big run quickly 就是没问题的 此外,为了句子通顺,你可以在包含所有传入单词的前提下自己添加一些其他单词。对于上面的例子,函数返回值为 that big brown dog just run quickly 也是没问题的 function wordblanks mynoun myadjective myverb myadverb var result 请把你的代码写在这条注释以下 result mynoun myadjective myverb myadverb 请把你的代码写在这条注释以上 return result wordblanks dog big ran quickly 你可以修改这一行来测试你的代码 store multiple values in one variable using javascript arrays 使用 数组 ,我们可以在一个地方存储多个数据。 你以左方括号 结束定义,并把每个条目之间用逗号隔开,就像这样: var sandwich 。 任务 创建一个包含 字符串 和 数字 的数组 myarray 。 提示 如果你遇到困难,请参考文本编辑器中的示例代码。 举例 var array 请只修改这条注释以下的代码 var myarray nest one array within another array 你也可以在数组中包含其他数组,就像这样 。这被称为一个 多维数组。 任务 创建一个名为 myarray 的多维数组。 举例 var ourarray 请只修改这条注释以下的代码 var myarray access array data with indexes 我们可以像操作字符串一样通过数组索引 来访问数组中的数据。 数组索引的使用与字符串索引一样,不同的是,通过字符串的索引得到的是一个字符,通过数组索引得到的是一个条目。与字符串类似,数组也是 基于零 的索引,因此数组的第一个元素的索引是 。 例如 var array array 等于 var data array 等于 任务 创建一个名为 mydata 的变量,并把 myarray 的第一个索引上的值赋给它。 举例 var ourarray var ourdata ourarray ourdata 的值为 初始化变量 var myarray 请把你的代码写在这条注释以下 var mydata myarray modify array data with indexes 与字符串的数据不可变不同,数组的数据是可变的,并且可以自由地改变。 例如 var ourarray ourarray ourarray等于 任务 修改数组 myarray 。 举例 var ourarray ourarray ourarray 的值为 初始化变量 var myarray 请把你的代码写在这条注释以下 myarray access multidimensional arrays with indexes 可以把 多维 数组看作成是一个 数组中的数组 。当使用 去访问数组的时候,第一个 访问的是第n个子数组,第二个 访问的是第n个子数组的第n个元素。 例如 var arr arr 等于 arr 等于 arr 等于 任务 使用恰当的 访问 myarray ,使得 mydata 的值为 初始化变量 var myarray 请只修改这条注释以下的代码 var mydata myarray manipulate arrays with push 一个简单的方法将数据追加到一个数组的末尾是通过 push 函数。 push 接受把一个或多个参数,并把它“推”入到数组的末尾。 var arr arr push 现在arr的值为 任务 把 “推”入到 myarray 变量的末尾。 举例 var ourarray ourarray push 经过 push 操作后,ourarray 的值为 初始化变量 var myarray 请把你的代码写在这条注释以下 myarray push manipulate arrays with pop 改变数组中数据的另一种方法是用 pop 函数。 pop 函数用来“抛出”一个数组末尾的值。我们可以把这个“抛出”的值赋给一个变量存储起来。 数组中任何类型的条目(数值,字符串,甚至是数组)可以被“抛出来” 。 举个例子 对于这段代码 var onedown pop 现在 onedown 的值为 ,数组变成了 。 任务 使用 pop 函数移除 myarray 中的最后一条,并且把“抛出”的值赋给 removedfrommyarray 。 举例 var ourarray var removedfromourarray ourarray pop 经过 pop 操作之后,removedfromourarray 的值为 ourarray 的值为 初始化变量 var myarray 请只修改这条注释以下的代码 var removedfrommyarray myarray pop manipulate arrays with shift pop 函数用来移出数组中最后一个元素。如果想要移出第一个元素要怎么办呢? 这就是 shift 的用武之地。它的工作原理就像 pop ,但它移除的是第一个元素,而不是最后一个。 任务 使用 shift 函数移出 myarray 中的第一项,并把“移出”的值赋给 removedfrommyarray 。 举例 var ourarray removedfromourarray ourarray shift 经过 shift 操作后,removedfromourarray 的值为 stimpson ,ourarray 的值为 初始化变量 var myarray 请只修改这条注释以下的代码 var removedfrommyarray myarray shift manipulate arrays with unshift 你不仅可以 shift (移出)数组中的第一个元素,你也可以 unshift (移入)一个元素到数组的头部。 unshift 函数用起来就像 push 函数一样 但不是在数组的末尾添加元素,而是在数组的头部添加元素。 任务 使用 unshift 函数把 加入到 myarray 的头部。 举例 var ourarray ourarray shift 经过 shift 操作后,ourarray 的值为 ourarray unshift happy 经过 unshift 操作后,ourarray 的值为 初始化变量 var myarray myarray shift 请把你的代码写在这条注释以下 myarray unshift shopping list 购物清单 创建一个名叫 mylist 的购物清单,清单的数据格式就是多维数组。 每个子数组中的第一个元素应该是购买的物品名称,第二个元素应该是物品的数量,类似于: 任务: 。 var mylist write reusable javascript with functions 在 javascript 中,我们可以把代码的重复部分抽取出来,放到一个函数(functions)中。 这是一个函数(function)的例子: function functionname console log hello world 你可以通过函数名称 functionname 加上后面的小括号来调用这个函数(function),就像这样: functionname 每次调用函数时它会打印出消息的“hello world”到开发的控制台上。所有的大括号之间的代码将在每次函数调用时执行。 任务 创建一个名为 myfunction 的函数,这个函数可以打印“hi world”到开发控制台上。 调用这个函数。 举例 function ourfunction console log heyya world ourfunction 请把你的代码写在这条注释以下 function myfunction console log hi world myfunction passing values to functions with arguments 函数的参数 parameters 在���数中充当占位符 也叫形参 的作用,参数可以为一个或多个。调用一个函数时所传入的参数为实参,实参决定着形参真正的值。简单理解:形参即形式、实参即内容。 这是带有两个参数的函数, 和 : function testfun console log 接着我们调用 testfun : testfun hello world 我们传递了两个参数, hello 和 world 。在函数内部, 等于“hello”, 等于“world”。请注意, testfun 函数可以多次调用,每次调用时传递的参数会决定形参的实际值。 任务 创建一个名为 myfunction 的函数,它可以接收两个参数,计算参数的和,将结果输出到控制台。 调用这个函数。 举例 function ourfunction a b console log a b ourfunction 输出 请把你的代码写在这条注释以下 function myfunction a b console log a b myfunction global scope and functions 在 javascript 中, 作用域 涉及到变量的作用范围。在函数外定义的变量具有 全局 作用域。这意味着,具有全局作用域的变量可以在代码的任何地方被调用。 这些没有使用 var 关键字定义的变量,会被自动创建在全局作用域中,形成全局变量。当在代码其他地方无意间定义了一个变量,刚好变量名与全局变量相同,这时会产生意想不到的后果。因此你应该总是使用var关键字来声明你的变量。 任务 在函数外声明一个 全局 变量 myglobal ,并给它一个初始值 在函数 的内部, 不 使用 var 关键字来声明 oopsglobal ,并赋值为 。 请在这里定义变量 var myglobal function 请在这里把 赋值给 oopsglobal oopsglobal 请只修改这条注释以上的代码 function var output if typeof myglobal undefined output myglobal myglobal if typeof oopsglobal undefined output oopsglobal oopsglobal console log output local scope and functions 在一个函数内声明的变量,以及该函数的参数都是局部变量,意味着它们只在该函数内可见。 这是在函数 mytest 内声明局部变量 loc 的最佳例子: function mytest var loc foo console log loc mytest foo console log loc undefined 在函数外, loc 是未定义的。 任务 在函数 myfunction 内部声明一个局部变量 myvar ,并删除外部console log。 function myfunction use strict var myvar use strict console log myvar myfunction 请先运行这段代码,并在左边的输出区域或浏览器的控制台中查看输出 由于 myvar 在 myfunction 内外均没有定义,因此才会有报错 console log myvar 现在,在 myfunction 中定义 myvar,并删掉 myfunction 外面的 console log 那一行 global vs local scope in functions 一个程序中有可能具有相同名称的 局部 变量 和 全局 变量。在这种情况下, 局部 变量将会优先于 全局 变量。 下面为例: var somevar hat function myfun var somevar head return somevar 函数 myfun 将会返回 head ,因为 局部变量 优先级更高。 任务 给 myfunction 添加一个局部变量来覆盖 outerwear 的值为 sweater 。 setup var outerwear t shirt function myfunction only change code below this line var outerwear sweater only change code above this line return outerwear myfunction return a value from a function with return 我们可以把数据通过函数的 参数 来传入函数,也可以使用 return 语句把数据从一个函数中传出来。 例如 function plusthree num return num var answer plusthree plusthree 带有一个为 num 的 参数 并且返回(returns)一个等于 num 的值。 任务 创建一个函数 timesfive 接受一个参数 把它乘于 之后并返回(returns)。 举例 function minusseven num return num only change code below this line function timesfive num return num var answer timesfive var answer timesfive var answer timesfive assignment with a returned value 如果你还记得我们在这一节 的讨论,赋值之前,先完成等号右边的操作。这意味着我们可把一个函数的返回值,赋值给一个变量。 假设我们预先定义的函数 sum 其功能就是将两个数字相加,那么: oursum sum 将调用 sum 函数,返回 return 了一个数值 ,然后把它赋值给了 oursum 变量。 任务 调用 process 函数并给参数一个值 ,然后把返回的值赋值给变量 processed 。 举例 var changed function change num return num changed change setup var processed function process num return num only change code below this line processed process stand in line 在计算机科学中 队列(queue)是一个抽象的数据结构,队列中的条目都是有秩序的。新的条目会被加到 队列 的末尾,旧的条目会从 队列 的头部被移出。 写一个函数 queue ,用一个数组 arr 和一个数字 item 作为参数。数字 item 添加到数组的结尾,然后移出数组的第一个元素,最后队列函数应该返回被删除的元素。 function queue arr item 请把你的代码写在这里 arr push item item arr shift return item 请修改这一行 初始化测试数据 var testarr 控制台输出 console log before json stringify testarr console log queue testarr 你可以修改这一行来测试你的代码 console log after json stringify testarr understanding boolean values 另一种数据类型是布尔(boolean)。 布尔 值要么是 true 要么是 false 。它非常像电路开关, true 是“开”, false 是“关”。这两种状态是互斥的。 注意 boolean 值绝不会写作被引号包裹起来的形式。 字符串 的 true 和 false 不是 布尔值 ,在 javascript 中也没有特殊含义。 任务 修改 welcometobooleans 函数,让它返回 true 而不是 false 。 function welcometobooleans only change code below this line return true change this line only change code above this line use conditional logic with if statements if 语句用于在代码中做条件判断。关键字 if 告诉 javascript 在小括号中的条件为真的情况下去执行定义在大括号里面的代码。这种条件被称为 boolean 条件,因为他们只可能是 true (真)或 false (假)。 当条件的计算结果为 true ,程序执行大括号内的语句。当布尔条件的计算结果为 false ,大括号内的代码将不会执行。 伪代码 if 条件为真 语句被执行 示例 function test mycondition if mycondition return it was true return it was false test true 返回 it was true test false 返回 it was false 当 test 被调用,并且传递进来的参数值为 true , if 语句会计算 mycondition 的结果,看它是真还是假。如果条件为 true ,函数会返回 it was true 。当 test 被调用,并且传递进来的参数值为 false , mycondition 不 为 true ,并且不执行大括号后面的语句,函数返回 it was false 。 任务 在函数内部创建一个 if 语句,如果该参数 wasthattrue 值为 true ,返回 that was true ,否则,并返回 that was false 。 举例 function ourfunction isittrue if isittrue return yes it s true return no it s false setup function myfunction wasthattrue if wasthattrue return that was true only change code below this line return that was false only change code above this line change this value to test myfunction true comparison with the equality operator 在 javascript 中,有很多 相互比较的操作。所有这些操作符都返回一个 true 或 false 值。 最基本的运算符是相等运算符: 。相等运算符比较两个值,如果它们是同等,返回 true ,如果它们不等,返回 false 。值得注意的是相等运算符不同于赋值运算符( ),赋值运算符是把等号右边的值赋给左边的变量。 function equalitytest myval if myval return equal return not equal 如果 myval 等于 ,相等运算符会返回 true ,因此大括号里面的代码会被执行,函数将返回 equal 。否则,函数返回 not equal 。 在 javascript 中,为了让两个不同的 数据类型 (例如 数字 和 字符串 )的值可以作比较,它必须把一种类型转换为另一种类型。然而一旦这样做,它可以像下面这样来比较: true false true true 任务 把 相等运算符 添加到指定的行,这样当 val 的值为 的时候,函数会返回 equal 。 setup function mytest val if val change this line return equal return not equal change this value to test mytest comparison with the strict equality operator 严格相等运算符( )是相对于相等操作符( )的一种操作符。与相等操作符不同的是,它会同时比较元素的值和 数据类型 。 举个例子 true false 是一个 数字 类型的,而 是一个 字符 类型的, 。 任务 在 if 语句值使用严格相等运算符,这样当 val ,函数会返回 equal 。 setup function mytest val if val change this line return equal return not equal change this value to test mytest comparison with the inequality operator 不相等运算符( )与相等运算符是相反的。这意味着不相等运算符中,如果“不为真”并且返回 false 的地方,在相等运算符中会返回 true , 反之亦然 。与相等运算符类似,不相等运算符在比较的时候也会转换值的数据类型。 例如 true false false true false false false 任务 在 if 语句中,添加不相等运算符 ,这样函数在当 val 不等于 的时候,会返回 not equal 。 setup function mytest val if val change this line return not equal return equal change this value to test mytest comparison with the strict inequality operator 严格不相等运算符( )与全等运算符是相反的。这意味着严格不相等并返回 false 的地方,用严格相等运算符会返回 true , 反之亦然 。严格相等运算符不会转换值的数据类型。 例如 false true true 任务 在 if 语句中,添加严格不相等运算符 ,这样如果 val 与 严格不相等的时候,函数会返回 not equal 。 setup function mytest val only change code below this line if val only change code above this line return not equal return equal change this value to test mytest comparison with the greater than operator 使用大于运算符( )来比较两个数字。如果大于运算符左边的数字大于右边的数字,将会返回 true 。否则,它返回 false 。 与相等运算符一样,大于运算符在比较的时候,会转换值的数据类型。 例如 true true false false 任务 添加 大于 运算符到指定的行,使得返回的语句是有意义的。 function mytest val if val change this line return over if val change this line return over return or under change this value to test mytest comparison with the greater than or equal to operator 使用 大于等于 运算符( )来比较两个数字的大小。如果大于等于运算符左边的数字比右边的数字大或者相等,它会返回 true 。否则,它会返回 false 。 与相等运算符相似, 大于等于 运算符在比较的时候会转换值的数据类型。 例如 true true false false 任务 添加 大于等于 运算符到指定行,使得函数的返回语句有意义。 function mytest val if val change this line return or over if val change this line return or over return or under change this value to test mytest comparison with the less than operator 使用 小于 运算符( )比较两个数字的大小。如果小于运算符左边的数字比右边的数字小,它会返回 true 。否则,他会返回 false 。与相等运算符类似,小于 运算符在做比较的时候会转换值的数据类型。 例如 true true false false false 任务 添加 小于 运算符到指定行,使得函数的返回语句有意义。 function mytest val if val change this line return under if val change this line return under return or over change this value to test mytest comparison with the less than or equal to operator 使用 小于等于 运算符( )比较两个数字的大小。如果在小于等于运算符,左边的数字小于或者等于右边的数字,它会返回 true 。如果在小于等于运算符,左边的数字大于或者等于右边的数字,它会返回 false 。与相等运算符类型, 小于等于 运算符会转换数据类型。 例如 true true true false false 任务 添加 小于等于 运算符到指定行,使得函数的返回语句有意义。 function mytest val if val change this line return smaller than or equal to if val change this line return smaller than or equal to return or more change this value to test mytest comparisons with the logical and operator 有时你需要在一次判断中做多个操作。当且仅当运算符的左边和右边都是 true ,逻辑与 运算符( )才会返回 true 。 同样的效果可以通过if语句的嵌套来实现: if num if num return yes return no 只有当 num ( )才会返回 yes 。相同的逻辑可被写为: if num num return yes return no 任务 结合两个if语句为一个语句,如果 val 小于或等于 并且大于或等于 ,返回 yes 。否则,将返回 no 。 function mytest val only change code below this line if val return yes only change code above this line return no change this value to test mytest comparisons with the logical or operator 如果任何一个操作数是 true ,逻辑或 运算符 返回 true 。反之,返回 false 。 举个例子: if num return no if num return no return yes 只有当 num ,函数返回 yes 。相同的逻辑可以简写成: if num num return no return yes 任务 结合两个if语句为一个语句,如果 val ,返回 outside 。反之,返回 inside 。 function mytest val only change code below this line if val return outside only change code above this line return inside change this value to test mytest introducing else statements 当 if 语句的条件为真,大括号里的代码执行,那如果条件为假呢? 正常情况下什么也不会发生。 写一个 else 语句,当条件为假时执行相应的代码。 if num return bigger than else return or less 任务 结合多个 if 语句为一个 if else 语句。 function mytest val var result only change code below this line if val result bigger than else result or smaller only change code above this line return result change this value to test mytest introducing else if statements 如果你有多个条件语句,你可以通过 else if 语句把 if 语句链起来。 if num return bigger than else if num return smaller than else return between and 任务 使用 else if 实现同样的效果。 function mytest val if val return greater than else if val return smaller than else return between and change this value to test mytest logical order in if else statements if 、 else if 语句中代码的执行顺序是很重要的。 在条件判断语句中,代码的执行顺序是从上到下,所以你需要考虑清楚先执行哪一句,后执行哪一句。 这有两个例子。 第一个例子: function foo x if x return less than one else if x return less than two else return greater than or equal to two 第二个例子更改了代码的执行顺序: function bar x if x return less than two else if x return less than one else return greater than or equal to two 这两个函数看起来几乎一模一样,我们传一个值进去看看它们有什么区别。 foo less than one bar less than two 任务 更改函数的逻辑顺序以便通过所有的测试用例。 function mytest val if val return less than else if val return less than else return greater than or equal to change this value to test mytest chaining if else statements if else 语句串联在一起可以实现复杂的逻辑,这是多个 if else if 语句串联在一起的伪代码: if else if else if else statementn 任务 把 if else if 语句串联起来实现下面的逻辑: num return tiny num return small num return medium num return large num return huge function mytest num only change code below this line if num return tiny else if num return small else if num return medium else if num return large else if num return huge return change me only change code above this line change this value to test mytest golf code 在高尔夫 golf 游戏中,每个洞都有自己的标准杆数 par ,代表着距离。根据你把球打进洞所挥杆的次数 strokes ,可以计算出你的高尔夫水平。 ,分别是标准杆数 par 和 挥杆次数 strokes ,根据下面的表格返回正确的水平段位。 strokes return hole in one par eagle par birdie par par par bogey par double bogey par go home par 和 strokes 必须是数字而且是正数。 function golfscore par strokes only change code below this line if strokes return hole in one else if strokes par return eagle else if strokes par return birdie else if strokes par return par else if strokes par return bogey else if strokes par return double bogey else if strokes par return go home return change me only change code above this line change these values to test golfscore selecting from many options with switch statements 如果你有非常多的选项需要选择,可以使用switch语句。根据不同的参数值会匹配上不同的case分支,语句会从第一个匹配的case分支开始执行,直到碰到break就结束。 这是一个伪代码案例: switch num case break case break case valuen statementn break 测试 case 值使用严格相等运算符进行比较,break关键字告诉javascript停止执行语句。如果没有break关键字,下一个语句会继续执行。 任务 写一个测试 val 的switch语句,并且根据下面的条件来设置不同的 answer : alpha beta gamma delta function mytest val var answer only change code below this line switch val case answer alpha break case answer beta break case answer gamma break case answer delta break only change code above this line return answer change this value to test mytest adding a default option in switch statements 在 switch 语句中你可能无法用case来指定所有情况,这时你可以添加default语句。当再也找不到case匹配的时候default语句会执行,非常类似于if else组合中的else语句。 default 语句应该是最后一个case。 switch num case break case break default defaultstatement 任务 写一个根据下面的条件来设置 answer 的switch语句: a apple b bird c cat default stuff function mytest val var answer only change code below this line switch val case a answer apple break case b answer bird break case c answer cat break default answer stuff break only change code above this line return answer change this value to test mytest multiple identical options in switch statements 如果 switch 语句中的 case 分支的 break 语句漏掉了,后面的 case 语句会一直执行直到遇到 break 。如果你有多个输入值和输出值一样,可以试试下面的 switch 语句: switch val case case case result or break case result alone 、 、 。 任务 写一个根据下面的范围来设置 answer 的switch语句: low mid high 提示 你需要为每一个包含数字的范围准备一个 answer 语句。 function mytest val var answer only change code below this line switch val case case case answer low break case case case answer mid break case case case answer high break only change code above this line return answer change this value to test mytest replacing if else chains with switch 如果你有多个选项需要选择, switch 语句写起来会比多个串联的 if if else 语句容易些,譬如 if val answer a else if val answer b else answer c 可以被下面替代: switch val case answer a break case answer b break default answer c 任务 把串联的 if if else 语句改成 switch 语句。 function mytest val var answer only change code below this line switch val case bob answer marley break case answer the answer break case answer there is no break case answer missed me by this much break case answer ate nine break only change code above this line return answer change this value to test mytest returning boolean values from functions 你可能会回想起 ,所有的比较操作符返回的都是一个boolean值,要么是 true 要么是 false 。 使用 if else 语句来做比较然后返回 true 或 false 已经成为大家的共识 function isequal a b if a b return true else return false 因为 总是返回 true 或 false ,所以我们可以直接返回比较的结果: function isequal a b return a b 任务 移除 isless 函数的 if else 语句但不影响函数的功能。 function isless a b fix this code return a b change these values to test isless return early pattern for functions 当代码执行到return语句时,函数返回一个结果就结束运行了,return后面的语句根本不会执行。 举例 function myfun console log hello return world console log byebye myfun 上面的代码输出 hello 到控制台、返回 world ,但没有输出 byebye ,因为函数遇到return语句就退出了。 任务 修改函数 abtest 当 a 或 b ,函数立即返回一个 undefined 并退出。 提示 记住 setup function abtest a b only change code below this line if a b return undefined only change code above this line return math round math pow math sqrt a math sqrt b change values below to test your code abtest counting cards ,玩家可以通过计算牌桌上已经发放的卡牌的高低值来让自己在游戏中保持优势,这就叫 根据下面的表格,每张卡牌都分配了一个值。 ,那么玩家应该追加赌注。反之,追加少许赌注甚至不追加赌注。 count change cards j q k a 你需要写一个函数, ,它根据参数 card 的值来递增或递减变量 count ,函数返回一个由当前 count 和 bet count 或 hold count 拼接的字符串。注意 count 和 bet 或 hold 应该用空格分开。 例如: hold bet 提示 、 、 ,count值不变,那我们就可以忽略这种情况。 var count function cc card only change code below this line switch card case case case case case count break case case j case q case k case a count break case case case count break if count return count bet else return count hold return change me only change code above this line add remove calls to test your function 提示 only the last will display cc cc cc cc k cc a build javascript objects 你之前可能听说过对象 object 。 对象和数组很相似,数组是通过索引来访问和修改数据,对象是通过属性来访问和修改数据的。 这是一个示例对象: var cat name whiskers legs tails enemies 对象适合用来存储结构化数据,就和真实世界的对象一模一样,比如一只猫。 任务 创建一个叫做 mydog 的对象,它里面有这些属性: 名称 name 、 腿 legs 尾巴 tails 、 朋友 friends 。 你可以设置对象属性为任何你想要的值,��要 name 是字符串、 legs 和 tails 是数字、 friends 是数组。 举例 var ourdog name camper legs tails friends only change code below this line var mydog name wangzhe legs tails friends accessing objects properties with the dot operator 有两种方式访问对象属性,一个是点操作符 ,一个是中括号操作符 。 当你知道属性的名称的时候,使用点操作符。 这是一个使用点操作符读取对象属性的例子: var myobj var myobj var myobj 任务 通过点操作符读取对象 testobj ,把 hat 的属性值赋给变量 hatvalue ,把 shirt 的属性值赋给 shirtvalue 。 setup var testobj hat ballcap shirt jersey shoes cleats only change code below this line var hatvalue testobj hat change this line var shirtvalue testobj shirt change this line accessing objects properties with bracket notation 第二种访问对象的方式就是中括号操作符 ,如果你想访问的属性的名称有一个空格,这时你只能使用中括号操作符 。 这是一个使用中括号操作符 读取对象属性的例子: var myobj space name kirk more space spock myobj kirk myobj spock 提示:属性名称中如果有空格,必须把属性名称用单引号或双引号包裹起来。 任务 用中括号操作符读取对象 testobj 的属性 an entree 值和属性 the drink 值。 setup var testobj an entree hamburger my side veggies the drink water only change code below this line var entreevalue testobj change this line var drinkvalue testobj change this line accessing objects properties with variables 中括号操作符的另一个使用方式是用变量来访问一个属性。当你需要遍历对象的属性列表或查表时,这种方式极为有用。 这有一个使用变量来访问属性的例子: var someprop propname var myobj propname some value myobj some value 还有更多: var mydog hunter var dogs fido mutt hunter doberman snoopie beagle var breed dogs console log breed doberman 提示:当我们通过变量名访问属性的时候,不需要给变量名包裹引号。因为实际上我们使用的是变量的值,而不是变量的名称。 任务 使用变量 playernumber ,通过中括号操作符找到 testobj 中 playernumber 为 的值。 setup var testobj namath montana unitas only change code below this line var playernumber change this line var player testobj change this line updating object properties 当你创建了一个对象后,你可以用点操作符或中括号操作符来更新对象的属性。 举个例子,让我们看看 ourdog var ourdog name camper legs tails friends 让我们更改它的名称为 happy camper ,这有两种方式来更新对象的 name 属性: ourdog name happy camper ourdog happy camper 任务 更新 mydog 对象的 name 属性,让它的名字从 coder 变成 happy coder 。 举例 var ourdog name camper legs tails friends ourdog name happy camper setup var mydog name coder legs tails friends only change code below this line mydog name happy coder add new properties to a javascript object 你也可以像更改属性一样给对象添加属性。 看看我们是如何给 ourdog 添加 bark 属性: ourdog bark bow wow 或者 ourdog bow wow 任务 给 mydog 添加一个 bark 属性,设置它的值为狗的声音,例如: woof 。 举例 var ourdog name camper legs tails friends ourdog bark bow wow setup var mydog name happy coder legs tails friends only change code below this line mydog bark woof delete properties from a javascript object 我们同样可以删除对象的属性,例如: delete ourdog bark 任务 删除 mydog 对象的 tails 属性。 举例 var ourdog name camper legs tails friends bark bow wow delete ourdog bark setup var mydog name happy coder legs tails friends bark woof only change code below this line delete mydog tails using objects for lookups 对象和字典一样,可以用来存储键 值对。如果你的数据跟对象一样,你可以用对象来查找你想要的值,而不是使用switch或if else语句。当你知道你的输入数据在某个范围时,这种查找方式极为有效。 这是简单的反向字母表: var alpha z y x w c b a alpha y alpha c var value alpha y 任务 把switch语句转化为一个叫做lookup的对象。 setup function phoneticlookup val var result only change code below this line var lookup alpha adams bravo boston charlie chicago delta denver echo easy foxtrot frank undefined only change code above this line return lookup change this value to test phoneticlookup charlie testing objects for properties 有时检查一个对象属性是否存在是非常有用的,我们可以用 hasownproperty propname 方法来检查对象是否有该属性。如果有返回 true ,反之返回 false 。 举例 var myobj top hat bottom pants myobj hasownproperty top true myobj hasownproperty middle false 任务 修改函数 checkobj 检查 myobj 是否有 checkprop 属性,如果属性存在,返回属性对应的值,如果不存在,返回 not found 。 注意:如果你需要通过变量来访问对象的属性值,请用中括号操作符,点操作符不支持变量。 setup var myobj gift pony pet kitten bed sleigh function checkobj checkprop your code here if myobj hasownproperty checkprop return myobj else return not found test your code by modifying these values checkobj gift introducing javascript object notation json javascript object notation 简称 json ,它使用javascript对象的格式来存储数据。json是灵活的,因为它允许 数据结构 是 字符串,数字,布尔值,字符串,和 对象 的任意组合。 这里是一个json对象的示例: var ourmusic artist daft punk title homework release year formats cd cassette lp gold true 这是一个对象数组,并且对象有各种关于专辑的 详细信息。它也有一个嵌套的 formats 的数组。附加专辑记录可以被添加到数组的最上层。 提示 数组中有多个 json 对象的时候,对象与对象之间要用逗号隔开。 任务 添加一个新专辑到 mymusic 的json对象。添加 artist 和 title 字符串, release year 数字和 formats 字符串数组。 var mymusic artist billy joel title piano man release year formats cs lp gold true 这里很重要,要有逗号隔开。 add record here artist daft punk title homework release year formats cd cassette lp gold true accessing nested objects in json setup var mystorage car inside glove box maps passenger seat crumbs outside trunk jack only change code below this line var gloveboxcontents mystorage car inside change this line accessing nested arrays in json 正如我们在前面的例子所见,json对象可以嵌套对象和数组。与访问嵌套对象一样,用中括号操作符同样可以访问嵌套数组。 下面是如何访问嵌套数组的例子: var ourpets cats meowzer fluffy kit cat dogs spot bowser frankie ourpets cats fluffy ourpets dogs spot 任务 使用点操作符和中括号操作符来检索变量 myplants 的第二棵树。 setup var myplants type flowers list rose tulip dandelion type trees list fir pine birch only change code below this line var secondtree myplants list change this line record collection 右边有一个json对象,代表着你的专辑集。每一张专辑由一个唯一的id标识,并具有多种属性。但并非所有的专辑都有完整的信息。 写一个函数,它有个三个参数, id 、 prop 、 value 。 如果 value 而且 prop tracks , collectioncopy value 。 如果 value 而且 prop tracks , collectioncopy push value 。 如果 value , delete collectioncopy 。 记住:函数返回的永远是整个对象。 提示 使用中括号操作符来 初始化变量 var collection album slippery when wet artist bon jovi tracks let it rock you give love a bad name album artist prince tracks little red corvette artist robert palmer tracks album abba gold 深拷贝 collection,用于测试 var collectioncopy json parse json stringify collection 请只修改这条注释以下的代码 function update id prop value if prop tracks value collection push value else if value collection value else delete collection return collection 你可以修改这一行来测试你的代码 update artist abba iterate with javascript for loops 一个条件语句只能执行一次代码,而一个循环语句可以多次执行代码。 javascript 中最常见的循环就是“ for循环 ”。 for循环中的三个表达式用分号隔开: for 初始化 语句只会在执行循环开始之前执行一次。它通常用于定义和设置你的循环变量。 条件判断 语句会在每一轮循环的开始执行,只要条件判断为 true 就会继续执行循环。当条件为 false 的时候,循环将停止执行。这意味着,如果条件在一开始就为 false ,这个循环将不会执行。 计数器 是在每一轮循环结束时执行,通常用于递增或递减。 在下面的例子中,先初始化 i ,条件 i 为真,进入第一次循环,执行大括号里的代码,第一次循环结束。递增 i 的值,条件判断,就这样依次执行下去,直到条件判断为假,整个循环结束。 var ourarray for var i i i ourarray push i 最终 ourarray 的值为 任务 使用 for 循环把从 到 添加进 myarray 中。 for循环就是if条件语句的进化版。 举例 var ourarray for var i i i ourarray push i setup var myarray for var i i i myarray push i only change code below this line iterate odd numbers with a for loop for循环可以按照我们指定的顺序来迭代,通过更改我们的 计数器 ,我们可以按照偶数顺序来迭代。 初始化 i ,当 i 的时候继续循环。 i 让 i 。 var ourarray for var i i i ourarray push i 循环结束后, ourarray 的值为 。 改变 计数器 ,这样我们可以用奇数来数。 任务 写一个 for 循环, myarray 。 举例 var ourarray for var i i i ourarray push i setup var myarray for var i i i myarray push i only change code below this line count backwards with a for loop for循环也可以逆向迭代,只要我们定义好合适的条件。 为了能够从后往前两两倒数,我们需要改变我们的 初始化 , 条件判断 和 计数器 。 我们让 i ,并且当 i 的时候才继续循环。我们使用 i 来让 i 每次循环递减 。 var ourarray for var i i i ourarray push i 循环结束后, ourarray 的值为 。 让我们改变 初始化 和 计数器 ,这样我们就可以按照奇数从后往前两两倒着数。 任务 使用一个 for 循环, myarray 。 举例 var ourarray for var i i i ourarray push i setup var myarray for var i i i myarray push i only change code below this line iterate through an array with a for loop 迭代输出一个数组的每个元素是 javascript 中的常见需求, for 循环可以做到这一点。 下面的代码将输出数组 arr 的每个元素到控制台: var arr for var i i arr length i console log arr 记住数组的索引从零开始的,这意味着数组的最后一个元素的下标是:数组的长度 。我们这个循环的 条件 是 i arr length ,当 i 的值为 长度 的时候循环就停止了。 任务 声明并初始化一个变量 total 为 。使用 for 循环,使得 total 的值为 myarr 的数组中的每个元素的值的总和。 举例 var ourarr var ourtotal for var i i ourarr length i ourtotal ourarr setup var myarr var total for var i i myarr length i total myarr only change code below this line nesting for loops 如果你有一个二维数组,可以使用相同的逻辑,先遍历外面的数组,再遍历里面的子数组。下面是一个例子: var arr for var i i arr length i for var j j arr length j console log arr 一次输出 arr 中的每个子元素。提示,对于内部循环,我们可以通过 arr 的 length 来获得子数组的长度,因为 arr 的本身就是一个数组。 任务 修改函数 multiplyall ,获得 arr 内部数组的每个数字相乘的结果 product 。 function multiplyall arr var product only change code below this line for var i i arr length i for var j j arr length j product arr only change code above this line return product modify values below to test your code multiplyall iterate with javascript while loops 另一种类型的 javascript 循环被称为 while循环 ,因为它规定,当(while)条件为真,循环才会执行,反之不执行。 var ourarray var i while i ourarray push i i 任务 通过一个 while 循环,把从 到 的值添加到 myarray 中。 setup var myarray var i while i myarray push i i only change code below this line profile lookup 我们有一个对象数组,里面存储着通讯录。 函数 lookup 有两个预定义参数: firstname 值和 prop 属性 。 函数将会检查通讯录中是否存在一个与传入的 firstname 相同的联系人。如果存在,那么还需要检查对应的联系人中是否存在 prop 属性。 如果它们都存在,函数返回 prop 属性对应的值。 如果 firstname 值不存在,返回 no such contact 。 如果 prop 属性不存在,返回 no such property 。 setup var contacts firstname akira lastname laine number likes firstname harry lastname potter number likes firstname sherlock lastname holmes number likes firstname kristian lastname vos number unknown likes function lookup firstname prop only change code below this line for i i contacts length i if contacts firstname firstname if contacts hasownproperty prop return contacts return no such property return no such contact only change code above this line change these values to test your function lookup akira likes generate random fractions with javascript 计算机的行为只有两种:确定性和随机性。当你一步步地闯关来到这里就是确定行为,当你随意点了个链接就来到这里就是随机行为。 而随机数最适合用来创建这种随机行为。 math random 之间的随机小数,因此 math random 。 提示 随后的函数都会在 return 执行前调用,所以我们可以直接返回 math random 的值。 任务 更改 myfunction 来生成一个随机数取代 。 function myfunction only change code below this line var myfunction math random return myfunction only change code above this line generate random whole numbers with javascript 生成随机小数很棒,但随机数更有用的地方在于生成随机整数。 用 math random 生成一个随机小数。 把这个随机小数乘以 。 用 math floor 向下取整 获得它最近的整数。 记住 math random 永远不会返回 。同时因为我们是在用 math floor 向下取整,所以最终我们获得的结果不可能有 。 。 把操作连缀起来,代码类似于下面: math floor math random 我们先调用 math random , ,然后把上一步的结果传给 math floor ,最终通过向下取整获得最近的整数。 var math floor math random function myfunction only change code below this line return math floor math random generate random whole numbers within a range ,现在我们要生成的随机数是在两个指定的数之间。 我们需要定义一个最小值和一个最大值。 下面是我们将要使用的方法,仔细看看并尝试理解这行代码到底在干嘛: math floor math random max min min 任务 创建一个叫 randomrange 的函数,参数为mymin和mymax,返回一个在 mymin 包括mymin 和 mymax 包括mymax 之间的随机数。 举例 function ourfunction ourmin ourmax return math floor math random ourmax ourmin ourmin ourfunction only change code below this line function randomrange mymin mymax return math floor math random mymax mymin mymin change this line change these values to test your function var myrandom randomrange sift through text with regular expressions regular expressions 正则表达式被用来根据某种匹配模式来寻找 strings 中的某些单词。 举例:如果我们想要找到字符串 the dog chased the cat 中单词 the ,我们可以使用下面的正则表达式 the gi 我们可以把这个正则表达式分成几段: 是这个正则表达式的头部 the 是我们想要匹配的模式 是这个正则表达式的尾部 g 代表着 global 全局 ,意味着返回所有的匹配而不仅仅是第一个。 i 代表着忽略大小写,意思是当我们寻找匹配的字符串的时候忽略掉字母的大小写。 任务 用全局、忽略大小写的模式选取字符串 teststring 中所有的单词 and 。 你可以尝试把 替换成 and 。 setup var teststring ada lovelace and charles babbage designed the first computer and the software that would have run on it 举例 var expressiontogetsoftware software gi var softwarecount teststring match expressiontogetsoftware length only change code below this line var expression and gi change this line only change code above this line this code counts the matches of expression in teststring var andcount teststring match expression length find numbers with regular expressions 我们可以在正则表达式中使用特殊选择器来选取特殊类型的值。 特殊选择器中的一种就是数字选择器 d ,意思是被用来获取一个字符串的数字。 在javascript中 数字选择器类似于 d g 。 在选择器后面添加一个加号标记 ,例如: d g ,它允许这个正则表达式匹配一个或更多数字。 尾部的 g 是 global 的简写,意思是允许这个正则表达式 找到所有的匹配而不是仅仅找到第一个匹配。 任务 用 d 选择器来选取字符串中的所有数字。 setup var teststring there are cats but dogs only change code below this line var expression d g change this line only change code above this line this code counts the matches of expression in teststring var digitcount teststring match expression length find whitespace with regular expressions 我们也可以使用正则表达式选择器 s 来选择一个字符串中的空白。 空白字符有 空格符 、 r 回车符 、 n 换行符 、 t 制表符 和 f 换页符 。 空白正则表达式类似于: s g 任务 用 s 选取句子中的所有空白字符。 setup var teststring how many spaces are there in this sentence only change code below this line var expression s g change this line only change code above this line this code counts the matches of expression in teststring var spacecount teststring match expression length invert regular expression matches with javascript 你可以用正则表达式选择器的大写版本 来转化任何匹配。 举个例子: s 匹配任何空白字符, s 匹配任何非空白字符。 任务 用 s g 来匹配字符串 teststring 中的所有非空白字符。 setup var teststring how many non space characters are there in this sentence only change code below this line var expression s g change this line only change code above this line this code counts the matches of expression in teststring var nonspacecount teststring match expression length create a javascript slot machine 现在把我们之前的所学的知识点结合起来完成一个老虎机游戏。 , 。 分别用 slotone 、 slottwo 、 slotthree 。 用我们之前的所学来生成 math floor math random function runslots var slotone var slottwo var slotthree var images only change code below this line var slotone math floor math random var slottwo math floor math random var slotthree math floor math random only change code above this line if slotone undefined slottwo undefined slotthree undefined logger html slotone slottwo slotthree logger append not a win return document ready function go click function runslots fcc slot machine go container background color height width margin auto border radius header border solid fff border radius height margin auto background color header height margin auto header font size margin padding color fff text align center slots display flex background color border radius border solid fff slot flex auto background white height margin border solid border radius go width color fff background color border solid fff border radius box sizing none outline none important foot height background color border solid fff logger color white margin outset webkit box shadow rgba moz box shadow rgba box shadow rgba inset webkit box shadow inset rgba moz box shadow inset rgba box shadow inset rgba add your javascript slot machine slots ,我们得去检查随机数是否全部相等的情况。 如果全部相等,我们应该提示用户他们赢了,并返回中奖号码,否则我们应该返回null。 null 是javascript中的一种数据类型,意味着空。 ,判定用户赢。让我们创建一个 if statement ,用多个条件按顺序来检查它们是否相等。类似于: if slotone slottwo slottwo slotthree return slotone else ,我们把 it s a win 追加到class logger 的html中。 function runslots var slotone var slottwo var slotthree var images slotone math floor math random slottwo math floor math random slotthree math floor math random 请把你的代码写在这条注释以下 if slotone slottwo slottwo slotthree slotthree slotone return null 请把你的代码写在这条注释以上 if slotone undefined slottwo undefined slotthree undefined logger html slotone slottwo slotthree logger append it s a win return document ready function go click function runslots fcc slot machine go container background color height width margin auto border radius header border solid fff border radius height margin auto background color header height margin auto header font size margin padding color fff text align center slots display flex background color border radius border solid fff slot flex auto background white height margin border solid border radius go width color fff background color border solid fff border radius box sizing none outline none important foot height background color border solid fff logger color white margin outset webkit box shadow rgba moz box shadow rgba box shadow rgba inset webkit box shadow inset rgba moz box shadow inset rgba box shadow inset rgba bring your javascript slot machine to life 让我们用 jquery 选择器 slot 获得所有老虎机。 一旦获取到所有老虎机,我们可以通过中括号操作符获取到每一个老虎机: slot html slotone jquery将会获取到第一个老虎机,并更新它的html为正确的数字。 任务:分别更新每个老虎机上的html为对应的数字。 function runslots var slotone var slottwo var slotthree var images slotone math floor math random slottwo math floor math random slotthree math floor math random only change code below this line slot html slotone slot html slottwo slot html slotthree only change code above this line if slotone slottwo slottwo slotthree logger html slotone slottwo slotthree it s a win return null if slotone undefined slottwo undefined slotthree undefined logger html slotone slottwo slotthree logger append not a win return document ready function go click function runslots fcc slot machine go container background color height width margin auto border radius header border solid fff border radius height margin auto background color header height margin auto header font size margin padding color fff text align center slots display flex background color border radius border solid fff slot flex auto background white height margin border solid border radius text align center padding top go width color fff background color border solid fff border radius box sizing none outline none important foot height background color border solid fff logger color white margin outset webkit box shadow rgba moz box shadow rgba box shadow rgba inset webkit box shadow inset rgba moz box shadow inset rgba box shadow inset rgba give your javascript slot machine some stylish images 现在给我们的老虎机加点图片。 我们已经为你准备好了图片 images ,我们可以通过不同的索引来获取每个图片。 现在让我们设置第一个老虎机根据随机数来显示一张图片: slot html 任务:设置所有的老虎机根据随机数来显示对应的图片,最后点击run。 function runslots var slotone var slottwo var slotthree var images slotone math floor math random slottwo math floor math random slotthree math floor math random only change code below this line slot html slot html slot html only change code above this line if slotone slottwo slottwo slotthree logger html slotone slottwo slotthree it s a win return null if slotone undefined slottwo undefined slotthree undefined logger html slotone slottwo slotthree logger append not a win return document ready function go click function runslots fcc slot machine go slot img margin important height width container background color height width margin auto border radius header border solid fff border radius height margin auto background color header height margin auto header font size margin padding color fff text align center slots display flex background color border radius border solid fff slot flex auto background white height width margin border solid border radius text align center go width color fff background color border solid fff border radius box sizing none outline none important foot height background color border solid fff logger color white margin outset webkit box shadow rgba moz box shadow rgba box shadow rgba inset webkit box shadow inset rgba moz box shadow inset rgba box shadow inset rgba 很遗憾的是。。 墙外的 fcc 已经没了老虎机这个项目了。。 不知道去哪了。。,0
+870,16601329239.0,IssuesEvent,2021-06-01 19:54:32,neumaticc/unrestricted-host-status,https://api.github.com/repos/neumaticc/unrestricted-host-status,closed,🛑 hates-people.xyz is down,hates-people-xyz status,"In [`9f82c2e`](https://github.com/neumaticc/unrestricted-host-status/commit/9f82c2e72fc2b0190a174e56c0267fe6149e58ec
+), hates-people.xyz (https://hates-people.xyz) was **down**:
+- HTTP code: 0
+- Response time: 0 ms
+",1.0,"🛑 hates-people.xyz is down - In [`9f82c2e`](https://github.com/neumaticc/unrestricted-host-status/commit/9f82c2e72fc2b0190a174e56c0267fe6149e58ec
+), hates-people.xyz (https://hates-people.xyz) was **down**:
+- HTTP code: 0
+- Response time: 0 ms
+",1,🛑 hates people xyz is down in hates people xyz was down http code response time ms ,1
+19054,11137261850.0,IssuesEvent,2019-12-20 18:50:55,PATRIC3/patric3_website,https://api.github.com/repos/PATRIC3/patric3_website,opened,Assembly failure on pacbio,Service: Assembly,"Job 362439 failed with canu not generating contigs for SRR2064843
+
+```
+$VAR2 = {
+ 'recipe' => 'canu',
+ 'paired_end_libs' => [],
+ 'pilon_iter' => 2,
+ 'trim' => 'false',
+ 'debug_level' => 0,
+ 'min_contig_len' => 300,
+ 'output_path' => '/nconrad@patricbrc.org/home/CGA/.Mycobacterium tuberculosis 18b test',
+ 'single_end_libs' => [],
+ 'output_file' => 'assembly',
+ 'genome_size' => 5000000,
+ 'min_contig_cov' => 90,
+ 'srr_ids' => [
+ 'SRR2064843'
+ ],
+ 'racon_iter' => 2
+ };
+```",1.0,"Assembly failure on pacbio - Job 362439 failed with canu not generating contigs for SRR2064843
+
+```
+$VAR2 = {
+ 'recipe' => 'canu',
+ 'paired_end_libs' => [],
+ 'pilon_iter' => 2,
+ 'trim' => 'false',
+ 'debug_level' => 0,
+ 'min_contig_len' => 300,
+ 'output_path' => '/nconrad@patricbrc.org/home/CGA/.Mycobacterium tuberculosis 18b test',
+ 'single_end_libs' => [],
+ 'output_file' => 'assembly',
+ 'genome_size' => 5000000,
+ 'min_contig_cov' => 90,
+ 'srr_ids' => [
+ 'SRR2064843'
+ ],
+ 'racon_iter' => 2
+ };
+```",0,assembly failure on pacbio job failed with canu not generating contigs for recipe canu paired end libs pilon iter trim false debug level min contig len output path nconrad patricbrc org home cga mycobacterium tuberculosis test single end libs output file assembly genome size min contig cov srr ids racon iter ,0
+568,10024367307.0,IssuesEvent,2019-07-16 21:40:31,planningcenter/developers,https://api.github.com/repos/planningcenter/developers,closed,Webhook not firing on People list creation,People Webhooks bug,"**Affected Applications**
+People
+
+**Describe the bug**
+When subscribed to the people.v2.events.list_result.created event (created via UI) I would expect there to be a webhook event fired with the list information when a list is created. What actually happens is no webhook at all is fired.
+
+Similarly, for people.v2.events.list_result.destroyed, no webhook is fired when a list is deleted.
+
+**To Reproduce**
+Steps to reproduce the behavior:
+1. Create a subscription to both the list_result.created and events.list_result.destroyed events [via UI](https://api.planningcenteronline.com/webhooks#/)
+2. [Create a list](https://people.planningcenteronline.com/lists)
+3. Expect to see an event fired on the subscription page (https://api.planningcenteronline.com/webhooks#/subscriptions/XXXX)
+4. See nothing
+
+**Expected behavior**
+I would expect a webhook to be fired on list creation, and ideally on list refresh.
+
+**Context (please complete the following information):**
+- Endpoint: UI (links above)
+- Language: N/A
+- Authentication: N/A
+
+**Additional context**
+Please let me know if these webhooks are not meant to be used as I'm describing.",1.0,"Webhook not firing on People list creation - **Affected Applications**
+People
+
+**Describe the bug**
+When subscribed to the people.v2.events.list_result.created event (created via UI) I would expect there to be a webhook event fired with the list information when a list is created. What actually happens is no webhook at all is fired.
+
+Similarly, for people.v2.events.list_result.destroyed, no webhook is fired when a list is deleted.
+
+**To Reproduce**
+Steps to reproduce the behavior:
+1. Create a subscription to both the list_result.created and events.list_result.destroyed events [via UI](https://api.planningcenteronline.com/webhooks#/)
+2. [Create a list](https://people.planningcenteronline.com/lists)
+3. Expect to see an event fired on the subscription page (https://api.planningcenteronline.com/webhooks#/subscriptions/XXXX)
+4. See nothing
+
+**Expected behavior**
+I would expect a webhook to be fired on list creation, and ideally on list refresh.
+
+**Context (please complete the following information):**
+- Endpoint: UI (links above)
+- Language: N/A
+- Authentication: N/A
+
+**Additional context**
+Please let me know if these webhooks are not meant to be used as I'm describing.",1,webhook not firing on people list creation affected applications people describe the bug when subscribed to the people events list result created event created via ui i would expect there to be a webhook event fired with the list information when a list is created what actually happens is no webhook at all is fired similarly for people events list result destroyed no webhook is fired when a list is deleted to reproduce steps to reproduce the behavior create a subscription to both the list result created and events list result destroyed events expect to see an event fired on the subscription page see nothing expected behavior i would expect a webhook to be fired on list creation and ideally on list refresh context please complete the following information endpoint ui links above language n a authentication n a additional context please let me know if these webhooks are not meant to be used as i m describing ,1
+153550,19708283701.0,IssuesEvent,2022-01-13 01:18:39,Rossb0b/web-agency,https://api.github.com/repos/Rossb0b/web-agency,opened,CVE-2020-28469 (High) detected in glob-parent-3.1.0.tgz,security vulnerability,"## CVE-2020-28469 - High Severity Vulnerability
+ Vulnerable Library - glob-parent-3.1.0.tgz
+
+
Strips glob magic from a string to provide the parent directory path
+
+
+
+
+
+***
+Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",True,"CVE-2020-28469 (High) detected in glob-parent-3.1.0.tgz - ## CVE-2020-28469 - High Severity Vulnerability
+ Vulnerable Library - glob-parent-3.1.0.tgz
+
+
Strips glob magic from a string to provide the parent directory path
+
+
+
+
+
+***
+Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",0,cve high detected in glob parent tgz cve high severity vulnerability vulnerable library glob parent tgz strips glob magic from a string to provide the parent directory path library home page a href path to dependency file web agency package json path to vulnerable library node modules glob parent package json node modules glob parent package json dependency hierarchy webpack encore tgz root library webpack dev server tgz chokidar tgz x glob parent tgz vulnerable library vulnerability details this affects the package glob parent before the enclosure regex used to check for strings ending in enclosure containing path separator publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution glob parent step up your open source security game with whitesource ,0
+261,5256057666.0,IssuesEvent,2017-02-02 16:58:05,ccswbs/hjckrrh,https://api.github.com/repos/ccswbs/hjckrrh,opened,PP1 - Allow Site Managers to customize People Profiles listing page without overriding the ug_profile feature,feature: people profiles (PP) priority: high type: enhancement request,"Allow Site Managers to customize People Profiles listing page without overriding the ug_profile feature.
+
+**Requirements for listing page**
+- Allow site manager to select size of displayed profile images (eg. thumbnail, medium)
+- Allow site manager to choose *not* to display the profile image on the listing page
+- If profile images are displayed, configure the display so profiles with images show the image and profiles without images do not show the image (see Issue #471)
+- Once site managers begin to make changes to the people profiles listing page, ensure the ug_profile feature is not put into a state of override by these changes.
+- Ability to make these changes should be available to Site manager role only (not authors or editors).
+
+Note: This issue can be modified once we learn more about what is possible given the above constraints.",1.0,"PP1 - Allow Site Managers to customize People Profiles listing page without overriding the ug_profile feature - Allow Site Managers to customize People Profiles listing page without overriding the ug_profile feature.
+
+**Requirements for listing page**
+- Allow site manager to select size of displayed profile images (eg. thumbnail, medium)
+- Allow site manager to choose *not* to display the profile image on the listing page
+- If profile images are displayed, configure the display so profiles with images show the image and profiles without images do not show the image (see Issue #471)
+- Once site managers begin to make changes to the people profiles listing page, ensure the ug_profile feature is not put into a state of override by these changes.
+- Ability to make these changes should be available to Site manager role only (not authors or editors).
+
+Note: This issue can be modified once we learn more about what is possible given the above constraints.",1, allow site managers to customize people profiles listing page without overriding the ug profile feature allow site managers to customize people profiles listing page without overriding the ug profile feature requirements for listing page allow site manager to select size of displayed profile images eg thumbnail medium allow site manager to choose not to display the profile image on the listing page if profile images are displayed configure the display so profiles with images show the image and profiles without images do not show the image see issue once site managers begin to make changes to the people profiles listing page ensure the ug profile feature is not put into a state of override by these changes ability to make these changes should be available to site manager role only not authors or editors note this issue can be modified once we learn more about what is possible given the above constraints ,1
+78704,15052488055.0,IssuesEvent,2021-02-03 15:15:16,mozilla-mobile/android-components,https://api.github.com/repos/mozilla-mobile/android-components,closed,Create a concept to lazily inflate views needed by features,⌨️ code 🌟 feature 🔬 Research,"We've seen [two][0] [cases][1] so far where we want to lazy inflate views that features need. However, it our current feature APIs, we accept a View interface that expects the instance to always be there.
+
+For the Find In Page example, we were able to mitigate this by initializing the feature _after_ the view was inflated and creating the [`InflationAwareFeature`][2].
+
+For `ReaderViewControlsBar`, we always need the feature running (so that we know when a page is readerable™), so we can't use the above abstraction for all features.
+
+What we need is an abstraction that covers both these cases.
+
+[0]: https://github.com/mozilla-mobile/fenix/issues/4307
+[1]: https://github.com/mozilla-mobile/fenix/issues/4309
+[2]: https://github.com/mozilla-mobile/fenix/blob/master/app/src/main/java/org/mozilla/fenix/components/InflationAwareFeature.kt
+
+
+",1.0,"Create a concept to lazily inflate views needed by features - We've seen [two][0] [cases][1] so far where we want to lazy inflate views that features need. However, it our current feature APIs, we accept a View interface that expects the instance to always be there.
+
+For the Find In Page example, we were able to mitigate this by initializing the feature _after_ the view was inflated and creating the [`InflationAwareFeature`][2].
+
+For `ReaderViewControlsBar`, we always need the feature running (so that we know when a page is readerable™), so we can't use the above abstraction for all features.
+
+What we need is an abstraction that covers both these cases.
+
+[0]: https://github.com/mozilla-mobile/fenix/issues/4307
+[1]: https://github.com/mozilla-mobile/fenix/issues/4309
+[2]: https://github.com/mozilla-mobile/fenix/blob/master/app/src/main/java/org/mozilla/fenix/components/InflationAwareFeature.kt
+
+
+",0,create a concept to lazily inflate views needed by features we ve seen so far where we want to lazy inflate views that features need however it our current feature apis we accept a view interface that expects the instance to always be there for the find in page example we were able to mitigate this by initializing the feature after the view was inflated and creating the for readerviewcontrolsbar we always need the feature running so that we know when a page is readerable™ so we can t use the above abstraction for all features what we need is an abstraction that covers both these cases ,0
+287148,24811365525.0,IssuesEvent,2022-10-25 09:40:22,HSLdevcom/jore4,https://api.github.com/repos/HSLdevcom/jore4,closed,Improve testing utilities and add stop tests,frontend Routes and lines testing,"Improve map testing utilities to make it easier to implement map tests.
+Add stop tests.
+
+",1.0,"Improve testing utilities and add stop tests - Improve map testing utilities to make it easier to implement map tests.
+Add stop tests.
+
+",0,improve testing utilities and add stop tests improve map testing utilities to make it easier to implement map tests add stop tests ,0
+738,13434205102.0,IssuesEvent,2020-09-07 10:59:44,samvera-labs/samvera-connect,https://api.github.com/repos/samvera-labs/samvera-connect,opened,Building a GLAM ecosystem: Human and machine collaborations for digitized collections ,Administrators Managers (general) Metadata Presentation (30 min slot (20+5+5)) UI/UX people,"Abigail E. Shelton ashelto3@nd.edu
+Rob Fox (rfox2@nd.edu)
+
+The University of Notre Dame has taken a modular approach to building a new digital collections platform-integrating existing applications and connecting the people that manage and use them across the library, archives, and art museum. We began with two assumptions: one size would not fit all for our campus archives, library, and museum; and community needs above all. In this presentation, team members will talk about the past two years of experimentation, development, and conversation around how to connect our community to our cultural heritage collections through multiple integrations, both human and technological. At a high-level, we’ll discuss our technical architecture that uses legacy applications like ArchivesSpace, an aging Fedora repository, and a decades-old museum database together with the IIIF framework and open-source GatsbyJS. And perhaps more importantly, we’ll outline the cross-departmental team structure that has developers talking to museum curators, library cataloguers, archivists, and everyone in between.
+
+",1.0,"Building a GLAM ecosystem: Human and machine collaborations for digitized collections - Abigail E. Shelton ashelto3@nd.edu
+Rob Fox (rfox2@nd.edu)
+
+The University of Notre Dame has taken a modular approach to building a new digital collections platform-integrating existing applications and connecting the people that manage and use them across the library, archives, and art museum. We began with two assumptions: one size would not fit all for our campus archives, library, and museum; and community needs above all. In this presentation, team members will talk about the past two years of experimentation, development, and conversation around how to connect our community to our cultural heritage collections through multiple integrations, both human and technological. At a high-level, we’ll discuss our technical architecture that uses legacy applications like ArchivesSpace, an aging Fedora repository, and a decades-old museum database together with the IIIF framework and open-source GatsbyJS. And perhaps more importantly, we’ll outline the cross-departmental team structure that has developers talking to museum curators, library cataloguers, archivists, and everyone in between.
+
+",1,building a glam ecosystem human and machine collaborations for digitized collections abigail e shelton nd edu rob fox nd edu the university of notre dame has taken a modular approach to building a new digital collections platform integrating existing applications and connecting the people that manage and use them across the library archives and art museum we began with two assumptions one size would not fit all for our campus archives library and museum and community needs above all in this presentation team members will talk about the past two years of experimentation development and conversation around how to connect our community to our cultural heritage collections through multiple integrations both human and technological at a high level we’ll discuss our technical architecture that uses legacy applications like archivesspace an aging fedora repository and a decades old museum database together with the iiif framework and open source gatsbyjs and perhaps more importantly we’ll outline the cross departmental team structure that has developers talking to museum curators library cataloguers archivists and everyone in between ,1
+198,4228960881.0,IssuesEvent,2016-07-04 04:08:53,elmsln/elmsln,https://api.github.com/repos/elmsln/elmsln,opened,CPR user preferences network transaction,media / elmsmedia people / cpr,"- Someone sets profile preferences in people.elmsln.local for how they want to get media / other preferences
+- media system asks people who needs special media considerations by querying `user.json?field_media_preference=local`
+- this returns json of everyone that currently requires local media, then it renders the video appropriately (if a local alternative exists)
+- This list gets cached as 1 call, if someone updates their preference we pin-prick and reseed the cache entry in media
+
+We could also pull all user data as a single call per section if we did something like...
+`node.json?field_section_id=master_sing100&type=section&xml-out&deep-load-refs=user`
+This would give you account info of everyone taking sing100's master section. This could be useful for getting profile / name data about all these people but I think I might want to sync this data local to the user in the sites they navigate (haven't decided on this yet, kind of a big decision).
+
+Another easier call would be for accessibility preferences which we could add additional ones to and would be similar to the media delivery example",1.0,"CPR user preferences network transaction - - Someone sets profile preferences in people.elmsln.local for how they want to get media / other preferences
+- media system asks people who needs special media considerations by querying `user.json?field_media_preference=local`
+- this returns json of everyone that currently requires local media, then it renders the video appropriately (if a local alternative exists)
+- This list gets cached as 1 call, if someone updates their preference we pin-prick and reseed the cache entry in media
+
+We could also pull all user data as a single call per section if we did something like...
+`node.json?field_section_id=master_sing100&type=section&xml-out&deep-load-refs=user`
+This would give you account info of everyone taking sing100's master section. This could be useful for getting profile / name data about all these people but I think I might want to sync this data local to the user in the sites they navigate (haven't decided on this yet, kind of a big decision).
+
+Another easier call would be for accessibility preferences which we could add additional ones to and would be similar to the media delivery example",1,cpr user preferences network transaction someone sets profile preferences in people elmsln local for how they want to get media other preferences media system asks people who needs special media considerations by querying user json field media preference local this returns json of everyone that currently requires local media then it renders the video appropriately if a local alternative exists this list gets cached as call if someone updates their preference we pin prick and reseed the cache entry in media we could also pull all user data as a single call per section if we did something like node json field section id master type section xml out deep load refs user this would give you account info of everyone taking s master section this could be useful for getting profile name data about all these people but i think i might want to sync this data local to the user in the sites they navigate haven t decided on this yet kind of a big decision another easier call would be for accessibility preferences which we could add additional ones to and would be similar to the media delivery example,1
+209234,16188194563.0,IssuesEvent,2021-05-04 02:24:02,connorkuehl/tftp,https://api.github.com/repos/connorkuehl/tftp,closed,Add documentation for using Wireshark to capture TFTP packets,Hacktoberfest documentation good first issue,"Commit this as **Documentation/wireshark.md**.
+
+Wireshark is a helpful tool for debugging network protocols.
+
+For example, I can capture the packets sent during the unit tests by opening Wireshark, selecting the loopback interface, and then running `cargo test`. Then stop capture.",1.0,"Add documentation for using Wireshark to capture TFTP packets - Commit this as **Documentation/wireshark.md**.
+
+Wireshark is a helpful tool for debugging network protocols.
+
+For example, I can capture the packets sent during the unit tests by opening Wireshark, selecting the loopback interface, and then running `cargo test`. Then stop capture.",0,add documentation for using wireshark to capture tftp packets commit this as documentation wireshark md wireshark is a helpful tool for debugging network protocols for example i can capture the packets sent during the unit tests by opening wireshark selecting the loopback interface and then running cargo test then stop capture ,0
+44174,12025547610.0,IssuesEvent,2020-04-12 09:49:08,tiangolo/jbrout,https://api.github.com/repos/tiangolo/jbrout,closed,unit tests need to be run from manatlan's machine?,Priority-Low Type-Defect auto-migrated,"```
+What steps will reproduce the problem?
+1. `cd unittests`
+2. `python runtests.py`
+3.
+
+What is the expected output? What do you see instead?
+Expected output should be something about how the unit tests completed
+successfully.
+
+Instead I see this:
+(cut)
+...
+--- Tests tests_dbtags.py
+Traceback (most recent call last):
+ File ""runtests.py"", line 29, in
+ execfile( ""../unittests/""+i )
+ File ""../unittests/tests_dbtags.py"", line 62, in
+ dbt.save()
+ File ""/home/conrad/data/documents/projects/photos/jbrout-read-only/jbrout/jbrout/db.py"", line 962, in save
+ fid = open(self.file,""w"")
+IOError: [Errno 2] No such file or directory:
+'/home/manatlan/db_jbrout_tags.xml'
+
+
+
+Since it's looking for '/home/manatlan/db_jbrout_tags.xml', the success of the
+test is based on the developer's environment. Unit tests should be able to
+succeed without being run from a specific developer's machine.
+```
+
+Original issue reported on code.google.com by `conrad.p...@gmail.com` on 24 Aug 2011 at 1:16
+",1.0,"unit tests need to be run from manatlan's machine? - ```
+What steps will reproduce the problem?
+1. `cd unittests`
+2. `python runtests.py`
+3.
+
+What is the expected output? What do you see instead?
+Expected output should be something about how the unit tests completed
+successfully.
+
+Instead I see this:
+(cut)
+...
+--- Tests tests_dbtags.py
+Traceback (most recent call last):
+ File ""runtests.py"", line 29, in
+ execfile( ""../unittests/""+i )
+ File ""../unittests/tests_dbtags.py"", line 62, in
+ dbt.save()
+ File ""/home/conrad/data/documents/projects/photos/jbrout-read-only/jbrout/jbrout/db.py"", line 962, in save
+ fid = open(self.file,""w"")
+IOError: [Errno 2] No such file or directory:
+'/home/manatlan/db_jbrout_tags.xml'
+
+
+
+Since it's looking for '/home/manatlan/db_jbrout_tags.xml', the success of the
+test is based on the developer's environment. Unit tests should be able to
+succeed without being run from a specific developer's machine.
+```
+
+Original issue reported on code.google.com by `conrad.p...@gmail.com` on 24 Aug 2011 at 1:16
+",0,unit tests need to be run from manatlan s machine what steps will reproduce the problem cd unittests python runtests py what is the expected output what do you see instead expected output should be something about how the unit tests completed successfully instead i see this cut tests tests dbtags py traceback most recent call last file runtests py line in execfile unittests i file unittests tests dbtags py line in dbt save file home conrad data documents projects photos jbrout read only jbrout jbrout db py line in save fid open self file w ioerror no such file or directory home manatlan db jbrout tags xml since it s looking for home manatlan db jbrout tags xml the success of the test is based on the developer s environment unit tests should be able to succeed without being run from a specific developer s machine original issue reported on code google com by conrad p gmail com on aug at ,0
+1011,24800335407.0,IssuesEvent,2022-10-24 21:04:37,planningcenter/developers,https://api.github.com/repos/planningcenter/developers,closed,"""Starred"" attribute does not update in Webhook data",help People,"**Affected Product**
+Which product does this bug affect?
+People/Webhooks
+
+**Describe the bug**
+I am building an API to interface with the ""Lists"" portion of PCO. I am pulling data from the Webhooks. When I ""star"" a list on PCO, it is not updated on the data that I receive from the Webhook. However it is shown to be updated on the API explorer here:
+https://api.planningcenteronline.com/explorer/people/v2/lists
+It also appears accurately when I pull the list on Postman. For that reason I assume it is a webhook issue.
+
+I also updated things such as auto_refresh and total_people. Both of those updated as expected while “starred” remained the same.
+
+
+
+
+**To Reproduce**
+To reproduce this step, simply
+1. Create a webhook that listens for when a list is refreshed
+2. Star a list and then refresh the list.
+3. Collect the data that the webhook sends.
+4. Verify that the ""starred"" attribute has not become true.
+
+**Expected behavior**
+I expect that when a list is starred, the webhook will send the ""starred"" attribute as ""true"".
+
+**Screenshots**
+
+
+
+
+**Additional Context:**
+
+- Endpoint: https://api.planningcenteronline.com/people/v2/lists/[list_id]
+- Language: Node JS
+- Authentication: Basic Auth
+
+**Additional context**
+
+
+## I have..
+
+- [ x] Reviewed the documentation found at https://developer.planning.center/docs
+- [x ] Searched for previous issues reporting this bug
+- [x ] Removed all private information from this issue (credentials, tokens, emails, phone numbers, etc.)
+- [x] Reviewed my issue for completeness",1.0,"""Starred"" attribute does not update in Webhook data - **Affected Product**
+Which product does this bug affect?
+People/Webhooks
+
+**Describe the bug**
+I am building an API to interface with the ""Lists"" portion of PCO. I am pulling data from the Webhooks. When I ""star"" a list on PCO, it is not updated on the data that I receive from the Webhook. However it is shown to be updated on the API explorer here:
+https://api.planningcenteronline.com/explorer/people/v2/lists
+It also appears accurately when I pull the list on Postman. For that reason I assume it is a webhook issue.
+
+I also updated things such as auto_refresh and total_people. Both of those updated as expected while “starred” remained the same.
+
+
+
+
+**To Reproduce**
+To reproduce this step, simply
+1. Create a webhook that listens for when a list is refreshed
+2. Star a list and then refresh the list.
+3. Collect the data that the webhook sends.
+4. Verify that the ""starred"" attribute has not become true.
+
+**Expected behavior**
+I expect that when a list is starred, the webhook will send the ""starred"" attribute as ""true"".
+
+**Screenshots**
+
+
+
+
+**Additional Context:**
+
+- Endpoint: https://api.planningcenteronline.com/people/v2/lists/[list_id]
+- Language: Node JS
+- Authentication: Basic Auth
+
+**Additional context**
+
+
+## I have..
+
+- [ x] Reviewed the documentation found at https://developer.planning.center/docs
+- [x ] Searched for previous issues reporting this bug
+- [x ] Removed all private information from this issue (credentials, tokens, emails, phone numbers, etc.)
+- [x] Reviewed my issue for completeness",1, starred attribute does not update in webhook data affected product which product does this bug affect people webhooks describe the bug i am building an api to interface with the lists portion of pco i am pulling data from the webhooks when i star a list on pco it is not updated on the data that i receive from the webhook however it is shown to be updated on the api explorer here it also appears accurately when i pull the list on postman for that reason i assume it is a webhook issue i also updated things such as auto refresh and total people both of those updated as expected while “starred” remained the same to reproduce to reproduce this step simply create a webhook that listens for when a list is refreshed star a list and then refresh the list collect the data that the webhook sends verify that the starred attribute has not become true expected behavior i expect that when a list is starred the webhook will send the starred attribute as true screenshots img width alt screen shot at am src img width alt screen shot at am src additional context endpoint language node js authentication basic auth additional context i have reviewed the documentation found at searched for previous issues reporting this bug removed all private information from this issue credentials tokens emails phone numbers etc reviewed my issue for completeness,1
+1102,26861827337.0,IssuesEvent,2023-02-03 19:08:35,microsoft/fluentui,https://api.github.com/repos/microsoft/fluentui,opened,PeoplePicker Convergence,Component: PeoplePicker Type: Epic Fluent UI react-components (v9),"
+
+💡 When you create a PR for any of the checklist items, add a link to this Epic under the PR's **Related Issues** section.
+
+## Preparation
+
+- [ ] [Open UI Research](https://github.com/microsoft/fluentui/wiki/Component-Implementation-Guide#open-ui-research)
+ - [link to https://open-ui.org/]
+- [ ] [Open GitHub issues related to component](https://github.com/microsoft/fluentui/wiki/Component-Implementation-Guide#find-open-issues-on-github)
+ - [link to each issue]
+- [ ] [Create react-\* package and component from template](https://github.com/microsoft/fluentui/wiki/Component-Implementation-Guide#component-package)
+ - [link to package: https://github.com/microsoft/fluentui/tree/master/packages/react-components/react-(your-component)]
+- [ ] (Optional) [Draft implementation](https://github.com/microsoft/fluentui/wiki/Component-Implementation-Guide#draft-implementation)
+ - [link to draft implementation, if applicable]
+- [ ] [Component Spec authored](https://github.com/microsoft/fluentui/wiki/Component-Implementation-Guide#component-spec) and [reviewed](https://github.com/microsoft/fluentui/wiki/Component-Implementation-Guide#spec-review)
+
+## Implementation
+
+- [ ] [Component implementation](https://github.com/microsoft/fluentui/wiki/Component-Implementation-Guide#implementation)
+- [ ] Initial conformance and unit tests (validate basic functionality)
+- [ ] [Initial documentation](https://github.com/microsoft/fluentui/wiki/Component-Implementation-Guide#documentation)
+ - [ ] [Storybook stories](https://github.com/microsoft/fluentui/wiki/Component-Implementation-Guide#storybook-stories)
+ - [ ] README.md covering basic usage
+ - [ ] MIGRATION.md guide (include v8 and v0)
+- [ ] [Component released as unstable](https://github.com/microsoft/fluentui/wiki/Component-Implementation-Guide#unstable-release) from `@fluentui/react-components/unstable`
+
+## Validation
+
+- [ ] [Add tests](https://github.com/microsoft/fluentui/wiki/Component-Implementation-Guide#tests)
+ - [ ] Unit and conformance tests
+ - [ ] VR tests
+ - [ ] Bundle size fixtures
+ - [ ] Performance test scenario
+ - [ ] Accessibility behavior tests
+ - [ ] Create an issue and run [manual accessibility tests](https://github.com/microsoft/fluentui/wiki/Manual-Accessibility-Review-Checklist): [link to issue]
+- [ ] [Validate with partners](https://github.com/microsoft/fluentui/wiki/Component-Implementation-Guide#validation)
+- [ ] [Run a bug bash with other FUI crews](https://github.com/microsoft/fluentui/wiki/Component-Implementation-Guide#bug-bash)
+- [ ] [Finalize documentation](https://github.com/microsoft/fluentui/wiki/Component-Implementation-Guide#finalize-documentation)
+ - [ ] Review and add any missing storybook stories
+ - [ ] Finalize migration guide
+- [ ] [Component released as stable](https://github.com/microsoft/fluentui/wiki/Component-Implementation-Guide#stable-release) from `@fluentui/react-components`
+ - [ ] Ensure exports are removed from from `@fluentui/react-components/unstable`
+ - [ ] In package.json: Remove the alpha/beta tag from the version number in package.json
+ - [ ] In package.json: Change beachball's `disallowedChangeTypes` to `""major"", ""prerelease""`
+",1.0,"PeoplePicker Convergence -
+
+💡 When you create a PR for any of the checklist items, add a link to this Epic under the PR's **Related Issues** section.
+
+## Preparation
+
+- [ ] [Open UI Research](https://github.com/microsoft/fluentui/wiki/Component-Implementation-Guide#open-ui-research)
+ - [link to https://open-ui.org/]
+- [ ] [Open GitHub issues related to component](https://github.com/microsoft/fluentui/wiki/Component-Implementation-Guide#find-open-issues-on-github)
+ - [link to each issue]
+- [ ] [Create react-\* package and component from template](https://github.com/microsoft/fluentui/wiki/Component-Implementation-Guide#component-package)
+ - [link to package: https://github.com/microsoft/fluentui/tree/master/packages/react-components/react-(your-component)]
+- [ ] (Optional) [Draft implementation](https://github.com/microsoft/fluentui/wiki/Component-Implementation-Guide#draft-implementation)
+ - [link to draft implementation, if applicable]
+- [ ] [Component Spec authored](https://github.com/microsoft/fluentui/wiki/Component-Implementation-Guide#component-spec) and [reviewed](https://github.com/microsoft/fluentui/wiki/Component-Implementation-Guide#spec-review)
+
+## Implementation
+
+- [ ] [Component implementation](https://github.com/microsoft/fluentui/wiki/Component-Implementation-Guide#implementation)
+- [ ] Initial conformance and unit tests (validate basic functionality)
+- [ ] [Initial documentation](https://github.com/microsoft/fluentui/wiki/Component-Implementation-Guide#documentation)
+ - [ ] [Storybook stories](https://github.com/microsoft/fluentui/wiki/Component-Implementation-Guide#storybook-stories)
+ - [ ] README.md covering basic usage
+ - [ ] MIGRATION.md guide (include v8 and v0)
+- [ ] [Component released as unstable](https://github.com/microsoft/fluentui/wiki/Component-Implementation-Guide#unstable-release) from `@fluentui/react-components/unstable`
+
+## Validation
+
+- [ ] [Add tests](https://github.com/microsoft/fluentui/wiki/Component-Implementation-Guide#tests)
+ - [ ] Unit and conformance tests
+ - [ ] VR tests
+ - [ ] Bundle size fixtures
+ - [ ] Performance test scenario
+ - [ ] Accessibility behavior tests
+ - [ ] Create an issue and run [manual accessibility tests](https://github.com/microsoft/fluentui/wiki/Manual-Accessibility-Review-Checklist): [link to issue]
+- [ ] [Validate with partners](https://github.com/microsoft/fluentui/wiki/Component-Implementation-Guide#validation)
+- [ ] [Run a bug bash with other FUI crews](https://github.com/microsoft/fluentui/wiki/Component-Implementation-Guide#bug-bash)
+- [ ] [Finalize documentation](https://github.com/microsoft/fluentui/wiki/Component-Implementation-Guide#finalize-documentation)
+ - [ ] Review and add any missing storybook stories
+ - [ ] Finalize migration guide
+- [ ] [Component released as stable](https://github.com/microsoft/fluentui/wiki/Component-Implementation-Guide#stable-release) from `@fluentui/react-components`
+ - [ ] Ensure exports are removed from from `@fluentui/react-components/unstable`
+ - [ ] In package.json: Remove the alpha/beta tag from the version number in package.json
+ - [ ] In package.json: Change beachball's `disallowedChangeTypes` to `""major"", ""prerelease""`
+",1,peoplepicker convergence these issues are used by core contributors to track the list of items that should be completed as part of creating a component more info can be found here 💡 when you create a pr for any of the checklist items add a link to this epic under the pr s related issues section preparation optional and implementation initial conformance and unit tests validate basic functionality readme md covering basic usage migration md guide include and from fluentui react components unstable validation unit and conformance tests vr tests bundle size fixtures performance test scenario accessibility behavior tests create an issue and run review and add any missing storybook stories finalize migration guide from fluentui react components ensure exports are removed from from fluentui react components unstable in package json remove the alpha beta tag from the version number in package json in package json change beachball s disallowedchangetypes to major prerelease ,1
+231808,7643665371.0,IssuesEvent,2018-05-08 13:25:16,kcgrimes/grimes-simple-revive,https://api.github.com/repos/kcgrimes/grimes-simple-revive,opened,Inconsistent AI behavior with squad leader changes,Priority: Medium Status: Pending Type: Bug,"Per Rockapes, ref https://forums.bohemia.net/forums/topic/167673-grimes-simple-revive-script/?do=findComment&comment=3289054
+
+Since installing v0.9 in the missions I play after I go down the AI start following the other player in my squad even though I still show as the squad leader and have all the men under me. They still follow my orders but wont move in formation unless the other player does.
+
+They still follow my orders when I am healed but when I tell them to regroup or change formation it is always based on the other player in the squad instead of me, other than that everything else is fine.
+
+Regroup command doesn't seem to fix but again, will test further.",1.0,"Inconsistent AI behavior with squad leader changes - Per Rockapes, ref https://forums.bohemia.net/forums/topic/167673-grimes-simple-revive-script/?do=findComment&comment=3289054
+
+Since installing v0.9 in the missions I play after I go down the AI start following the other player in my squad even though I still show as the squad leader and have all the men under me. They still follow my orders but wont move in formation unless the other player does.
+
+They still follow my orders when I am healed but when I tell them to regroup or change formation it is always based on the other player in the squad instead of me, other than that everything else is fine.
+
+Regroup command doesn't seem to fix but again, will test further.",0,inconsistent ai behavior with squad leader changes per rockapes ref since installing in the missions i play after i go down the ai start following the other player in my squad even though i still show as the squad leader and have all the men under me they still follow my orders but wont move in formation unless the other player does they still follow my orders when i am healed but when i tell them to regroup or change formation it is always based on the other player in the squad instead of me other than that everything else is fine regroup command doesn t seem to fix but again will test further ,0
+936,20617530575.0,IssuesEvent,2022-03-07 14:34:43,FortAwesome/Font-Awesome,https://api.github.com/repos/FortAwesome/Font-Awesome,closed,Icon request: icon-pinball,new icon Available in FA Pro tabletop gaming (category) users & people (category),"### What can it be used for?
+
+This icon can be used for arcades/gaming/pinball (a quick growing e-sport)
+
+### Example image (optional)
+
+https://cdn.dribbble.com/users/35567/screenshots/593438/doc6.jpg
+
+### Icon request checklist
+
+- [x ] This is a single icon or matched pair (Ex: `lock` / `unlock`)
+- [ x] The title starts with ""Icon request: "" and is followed by the requested icon name (Ex: `Icon request: magical-unicorn`)
+- [x ] I have [searched for existing issues](https://github.com/FortAwesome/Font-Awesome/issues) and to the best of my knowledge this is not a duplicate
+- [ x] I have included at least one example image if it could be helpful (optional)
+- [x] The request is for a concrete object (it's harder to make an icon to represent happiness, it's easier to make a smiley face. ☺)
+",1.0,"Icon request: icon-pinball - ### What can it be used for?
+
+This icon can be used for arcades/gaming/pinball (a quick growing e-sport)
+
+### Example image (optional)
+
+https://cdn.dribbble.com/users/35567/screenshots/593438/doc6.jpg
+
+### Icon request checklist
+
+- [x ] This is a single icon or matched pair (Ex: `lock` / `unlock`)
+- [ x] The title starts with ""Icon request: "" and is followed by the requested icon name (Ex: `Icon request: magical-unicorn`)
+- [x ] I have [searched for existing issues](https://github.com/FortAwesome/Font-Awesome/issues) and to the best of my knowledge this is not a duplicate
+- [ x] I have included at least one example image if it could be helpful (optional)
+- [x] The request is for a concrete object (it's harder to make an icon to represent happiness, it's easier to make a smiley face. ☺)
+",1,icon request icon pinball what can it be used for this icon can be used for arcades gaming pinball a quick growing e sport example image optional icon request checklist this is a single icon or matched pair ex lock unlock the title starts with icon request and is followed by the requested icon name ex icon request magical unicorn i have and to the best of my knowledge this is not a duplicate i have included at least one example image if it could be helpful optional the request is for a concrete object it s harder to make an icon to represent happiness it s easier to make a smiley face ☺ ,1
+721,13218751307.0,IssuesEvent,2020-08-17 09:16:52,Swiss-Polar-Institute/project-application,https://api.github.com/repos/Swiss-Polar-Institute/project-application,closed,Add grantee / applicant address,before next call opens people proposal,"The address of a grantee or applicant is used for communication, such as sending the decision letter or grant agreement. Currently this is not part of the application form and therefore can be very tricky to find.
+
+The applicant's address (that should be used for correspondance) could be included in the application form so it can be used to send the initial decision letter.
+
+Adding it as part of the person position for the applicant would mean that it is possible for it to appear with the proposal and project, but could also change (as it is not with the physical person) as the person moves to different institutions.
+
+Having the address as a text box is an option. Structuring the full address is tricky to account for variation in all possible countries. Validation for all parts of an address is tricky as well because of variation between countries. It might be useful to have separate fields for the postcode, city and country.
+
+A country list could come from the ISO list of [countries](https://www.iso.org/iso-3166-country-codes.html) (to avoid duplications). ",1.0,"Add grantee / applicant address - The address of a grantee or applicant is used for communication, such as sending the decision letter or grant agreement. Currently this is not part of the application form and therefore can be very tricky to find.
+
+The applicant's address (that should be used for correspondance) could be included in the application form so it can be used to send the initial decision letter.
+
+Adding it as part of the person position for the applicant would mean that it is possible for it to appear with the proposal and project, but could also change (as it is not with the physical person) as the person moves to different institutions.
+
+Having the address as a text box is an option. Structuring the full address is tricky to account for variation in all possible countries. Validation for all parts of an address is tricky as well because of variation between countries. It might be useful to have separate fields for the postcode, city and country.
+
+A country list could come from the ISO list of [countries](https://www.iso.org/iso-3166-country-codes.html) (to avoid duplications). ",1,add grantee applicant address the address of a grantee or applicant is used for communication such as sending the decision letter or grant agreement currently this is not part of the application form and therefore can be very tricky to find the applicant s address that should be used for correspondance could be included in the application form so it can be used to send the initial decision letter adding it as part of the person position for the applicant would mean that it is possible for it to appear with the proposal and project but could also change as it is not with the physical person as the person moves to different institutions having the address as a text box is an option structuring the full address is tricky to account for variation in all possible countries validation for all parts of an address is tricky as well because of variation between countries it might be useful to have separate fields for the postcode city and country a country list could come from the iso list of to avoid duplications ,1
+183359,21721714526.0,IssuesEvent,2022-05-11 01:19:09,L-47/unstoppable-chat,https://api.github.com/repos/L-47/unstoppable-chat,opened,CVE-2021-42581 (Medium) detected in ramda-0.26.1.tgz,security vulnerability,"## CVE-2021-42581 - Medium Severity Vulnerability
+ Vulnerable Library - ramda-0.26.1.tgz
+
+
A practical functional library for JavaScript programmers.
+
+Prototype poisoning in function mapObjIndexed in Ramda 0.27.0 and earlier allows attackers to compromise integrity or availability of application via supplying a crafted object (that contains an own property ""__proto__"") as an argument to the function.
+
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",True,"CVE-2021-42581 (Medium) detected in ramda-0.26.1.tgz - ## CVE-2021-42581 - Medium Severity Vulnerability
+ Vulnerable Library - ramda-0.26.1.tgz
+
+
A practical functional library for JavaScript programmers.
+
+Prototype poisoning in function mapObjIndexed in Ramda 0.27.0 and earlier allows attackers to compromise integrity or availability of application via supplying a crafted object (that contains an own property ""__proto__"") as an argument to the function.
+
+
+
+
+
+
+
+***
+Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)",0,cve medium detected in ramda tgz cve medium severity vulnerability vulnerable library ramda tgz a practical functional library for javascript programmers library home page a href path to dependency file package json path to vulnerable library node modules ramda package json dependency hierarchy gun tgz root library emailjs tgz emailjs mime codec tgz x ramda tgz vulnerable library found in base branch master vulnerability details prototype poisoning in function mapobjindexed in ramda and earlier allows attackers to compromise integrity or availability of application via supplying a crafted object that contains an own property proto as an argument to the function publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution ramda step up your open source security game with whitesource ,0
+1096,26813503368.0,IssuesEvent,2023-02-02 01:11:20,openstates/issues,https://api.github.com/repos/openstates/issues,closed,New OH Committee Scraper,good first issue component:people-data good first scraper,"### We need a new committee scraper written in [spatula](https://jamesturk.github.io/spatula/) for Ohio.
+
+It should scrape this [webpage for Senate Committees](https://www.legislature.ohio.gov/committees/senate-committees) and [this webpage for House and Joint Committees](https://ohiohouse.gov/committees) to get:
+- **name**
+ - “Small Business”, “Agriculture”, etc.
+- **chamber**
+ - “upper”, “lower”, or “legislature” (when joint, etc.)
+- **classification**
+ - ex: ""committee"", ""subcommittee""
+- **parent**
+ - *only if it is a subcommittee* - scraper should have a way to determine this
+ - ex: parent=""Natural Resources"" when *sub*committee name=""Forestry""
+- **sources**
+ - each added using `add_source()` method on instance of `ScrapeCommittee` type object
+ - ex: home page for list of committees, specific page for that committee, etc.
+- **members**
+ - - each added using `add_member()`method on instance of `ScrapeCommittee` type object
+ - *Attributes:*
+ - **name**: “Jane Doe”, “John Smith”, etc.
+ - **role** (where applicable): “Chair”, “Ranking Member”, etc.
+
+#### [Here is helpful documentation](https://docs.openstates.org/contributing/writing-a-committee-scraper/) for writing a committee scraper
+
+### Useful scrapers for reference
+A [spatula-based scraper](https://github.com/openstates/openstates-scrapers/blob/main/scrapers_next/mo/committees.py) that also scrapes an HTML List Page (using `HtmlListPage` and `HtmlPage` spatula classes) for reference.
+- As in the other spatula scraper, you will need to set the session variable separately (rather than depend on the `__init__.py`).
+
+### Other useful resources
+You can reference the [object models for committee scrapers](https://github.com/openstates/openstates-core/blob/main/openstates/models/committees.py) in the Open States Core repo.
+
+Further documentation on running spatula scrapers in the command line can be found [here](https://docs.openstates.org/contributing/scrapers/#running-spatula-scrapers).",1.0,"New OH Committee Scraper - ### We need a new committee scraper written in [spatula](https://jamesturk.github.io/spatula/) for Ohio.
+
+It should scrape this [webpage for Senate Committees](https://www.legislature.ohio.gov/committees/senate-committees) and [this webpage for House and Joint Committees](https://ohiohouse.gov/committees) to get:
+- **name**
+ - “Small Business”, “Agriculture”, etc.
+- **chamber**
+ - “upper”, “lower”, or “legislature” (when joint, etc.)
+- **classification**
+ - ex: ""committee"", ""subcommittee""
+- **parent**
+ - *only if it is a subcommittee* - scraper should have a way to determine this
+ - ex: parent=""Natural Resources"" when *sub*committee name=""Forestry""
+- **sources**
+ - each added using `add_source()` method on instance of `ScrapeCommittee` type object
+ - ex: home page for list of committees, specific page for that committee, etc.
+- **members**
+ - - each added using `add_member()`method on instance of `ScrapeCommittee` type object
+ - *Attributes:*
+ - **name**: “Jane Doe”, “John Smith”, etc.
+ - **role** (where applicable): “Chair”, “Ranking Member”, etc.
+
+#### [Here is helpful documentation](https://docs.openstates.org/contributing/writing-a-committee-scraper/) for writing a committee scraper
+
+### Useful scrapers for reference
+A [spatula-based scraper](https://github.com/openstates/openstates-scrapers/blob/main/scrapers_next/mo/committees.py) that also scrapes an HTML List Page (using `HtmlListPage` and `HtmlPage` spatula classes) for reference.
+- As in the other spatula scraper, you will need to set the session variable separately (rather than depend on the `__init__.py`).
+
+### Other useful resources
+You can reference the [object models for committee scrapers](https://github.com/openstates/openstates-core/blob/main/openstates/models/committees.py) in the Open States Core repo.
+
+Further documentation on running spatula scrapers in the command line can be found [here](https://docs.openstates.org/contributing/scrapers/#running-spatula-scrapers).",1,new oh committee scraper we need a new committee scraper written in for ohio it should scrape this and to get name “small business” “agriculture” etc chamber “upper” “lower” or “legislature” when joint etc classification ex committee subcommittee parent only if it is a subcommittee scraper should have a way to determine this ex parent natural resources when sub committee name forestry sources each added using add source method on instance of scrapecommittee type object ex home page for list of committees specific page for that committee etc members each added using add member method on instance of scrapecommittee type object attributes name “jane doe” “john smith” etc role where applicable “chair” “ranking member” etc for writing a committee scraper useful scrapers for reference a that also scrapes an html list page using htmllistpage and htmlpage spatula classes for reference as in the other spatula scraper you will need to set the session variable separately rather than depend on the init py other useful resources you can reference the in the open states core repo further documentation on running spatula scrapers in the command line can be found ,1
+13675,4757260662.0,IssuesEvent,2016-10-24 16:05:58,joomla/joomla-cms,https://api.github.com/repos/joomla/joomla-cms,opened,Wrong template on ALL front end modals,No Code Attached Yet,"Currently wherever we have a modal window in the front end (there are a lot of them) then the url used includes **&tmpl=component**
+
+The problem is that this will always load the component.php of the default template NOT the component.php of the current template
+
+From what I can see the url should be **tmpl=component&templateStyle=(currentstyle)**
+
+I have no idea how to fix this code as in the php parts its within my skillset but for the JS i dont have a clue
+
+### Steps to reproduce the issue
+Assign protostar as the default template
+Assign beez as the template for a single article
+Edit that single article and try any modal eg the ext-article plugin or any of the media buttons
+View the source code of that modal and you will see that the template being used in the iframe is the protostar template and not the current template - beez
+
+
+
+### Expected result
+The modal uses the currently defined template
+
+
+### Actual result
+The modal always uses the default template
+
+
+### Additional comments
+The same code (tmpl=component) is used in the admin as well but you can only ever have one active template in the admin so this wasnt noticed and so when the code was copy/pasted to the frontend views it was not noticed.
+",1.0,"Wrong template on ALL front end modals - Currently wherever we have a modal window in the front end (there are a lot of them) then the url used includes **&tmpl=component**
+
+The problem is that this will always load the component.php of the default template NOT the component.php of the current template
+
+From what I can see the url should be **tmpl=component&templateStyle=(currentstyle)**
+
+I have no idea how to fix this code as in the php parts its within my skillset but for the JS i dont have a clue
+
+### Steps to reproduce the issue
+Assign protostar as the default template
+Assign beez as the template for a single article
+Edit that single article and try any modal eg the ext-article plugin or any of the media buttons
+View the source code of that modal and you will see that the template being used in the iframe is the protostar template and not the current template - beez
+
+
+
+### Expected result
+The modal uses the currently defined template
+
+
+### Actual result
+The modal always uses the default template
+
+
+### Additional comments
+The same code (tmpl=component) is used in the admin as well but you can only ever have one active template in the admin so this wasnt noticed and so when the code was copy/pasted to the frontend views it was not noticed.
+",0,wrong template on all front end modals currently wherever we have a modal window in the front end there are a lot of them then the url used includes tmpl component the problem is that this will always load the component php of the default template not the component php of the current template from what i can see the url should be tmpl component templatestyle currentstyle i have no idea how to fix this code as in the php parts its within my skillset but for the js i dont have a clue steps to reproduce the issue assign protostar as the default template assign beez as the template for a single article edit that single article and try any modal eg the ext article plugin or any of the media buttons view the source code of that modal and you will see that the template being used in the iframe is the protostar template and not the current template beez expected result the modal uses the currently defined template actual result the modal always uses the default template additional comments the same code tmpl component is used in the admin as well but you can only ever have one active template in the admin so this wasnt noticed and so when the code was copy pasted to the frontend views it was not noticed ,0
+422972,28490567669.0,IssuesEvent,2023-04-18 10:55:58,Mastercard/community-pass-react-native-wrapper,https://api.github.com/repos/Mastercard/community-pass-react-native-wrapper,closed,Add Community Pass Program Space APIs,documentation enhancement,"- [x] Implement getDataSchema()
+- [x] Implement add getWriteProgramSpaceActivityIntent()
+- [x] Implement getReadProgramSpaceActivityIntent()
+- [x] Implement exchangeKey()
+- [x] Test Program Space APIs
+- [x] Update documentation (examples and implementation steps)",1.0,"Add Community Pass Program Space APIs - - [x] Implement getDataSchema()
+- [x] Implement add getWriteProgramSpaceActivityIntent()
+- [x] Implement getReadProgramSpaceActivityIntent()
+- [x] Implement exchangeKey()
+- [x] Test Program Space APIs
+- [x] Update documentation (examples and implementation steps)",0,add community pass program space apis implement getdataschema implement add getwriteprogramspaceactivityintent implement getreadprogramspaceactivityintent implement exchangekey test program space apis update documentation examples and implementation steps ,0
+1166,30612973968.0,IssuesEvent,2023-07-23 21:01:00,restincode/restincode,https://api.github.com/repos/restincode/restincode,opened,Kevin David Mitnick (Condor),People Add Person Needs Review,"Please fill out as much information as you can. No fields are required, but the more you can provide the better.
+
+**General Info**
+* First name: Kevin
+* Last name: Mitnick
+* Handle:
+* Birth Year: 1963
+* Death Year: 2023
+* Link to Obituary: https://www.dignitymemorial.com/obituaries/las-vegas-nv/kevin-mitnick-11371668
+* Group Affiliations:
+* URL to main photo (or attach to Issue):
+* Description of person and/or activities:
+* Facebook memorial group URL:
+
+**Social Media Links**
+* Twitter: https://twitter.com/kevinmitnick?lang=en
+* Github:
+* LinkedIn:
+* Facebook:
+* Other: https://en.wikipedia.org/wiki/Kevin_Mitnick
+
+**Contributions**
+Repeat as many times as needed
+
+* Project name:
+* Project URL:
+* Project Description:
+
+**Photo Gallery**
+* URL(s) to additional photos:
+
+https://www.cnn.com/2023/07/21/us/computer-hacker-kevin-mitnick-obit/index.html
+https://www.bbc.com/news/world-us-canada-66263235
+https://apnews.com/article/mitnick-hacker-ghost-wires-cybersecurity-social-engineering-5648301b615635cb4c781f0c220681d9",1.0,"Kevin David Mitnick (Condor) - Please fill out as much information as you can. No fields are required, but the more you can provide the better.
+
+**General Info**
+* First name: Kevin
+* Last name: Mitnick
+* Handle:
+* Birth Year: 1963
+* Death Year: 2023
+* Link to Obituary: https://www.dignitymemorial.com/obituaries/las-vegas-nv/kevin-mitnick-11371668
+* Group Affiliations:
+* URL to main photo (or attach to Issue):
+* Description of person and/or activities:
+* Facebook memorial group URL:
+
+**Social Media Links**
+* Twitter: https://twitter.com/kevinmitnick?lang=en
+* Github:
+* LinkedIn:
+* Facebook:
+* Other: https://en.wikipedia.org/wiki/Kevin_Mitnick
+
+**Contributions**
+Repeat as many times as needed
+
+* Project name:
+* Project URL:
+* Project Description:
+
+**Photo Gallery**
+* URL(s) to additional photos:
+
+https://www.cnn.com/2023/07/21/us/computer-hacker-kevin-mitnick-obit/index.html
+https://www.bbc.com/news/world-us-canada-66263235
+https://apnews.com/article/mitnick-hacker-ghost-wires-cybersecurity-social-engineering-5648301b615635cb4c781f0c220681d9",1,kevin david mitnick condor please fill out as much information as you can no fields are required but the more you can provide the better general info first name kevin last name mitnick handle birth year death year link to obituary group affiliations url to main photo or attach to issue description of person and or activities facebook memorial group url social media links twitter github linkedin facebook other contributions repeat as many times as needed project name project url project description photo gallery url s to additional photos ,1
+253864,21710256955.0,IssuesEvent,2022-05-10 13:21:08,damccorm/test-migration-target,https://api.github.com/repos/damccorm/test-migration-target,opened,beam_PostCommit_XVR_Flink flaky: Connection refused,bug test-failures P1 flake,"https://ci-beam.apache.org/job/beam_PostCommit_XVR_Flink/4980/
+
+The following tests failed:
+
+org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$CoGroupByKeyTest.test
+org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$CoGroupByKeyTest.test
+org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$CombineGloballyTest.test
+org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$CombineGloballyTest.test
+org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$CombinePerKeyTest.test
+org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$CombinePerKeyTest.test
+org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$FlattenTest.test
+org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$FlattenTest.test
+org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$GroupByKeyTest.test
+org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$GroupByKeyTest.test
+org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$MultiInputOutputWithSideInputTest.test
+org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$MultiInputOutputWithSideInputTest.test
+org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$PartitionTest.test
+org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$PartitionTest.test
+org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$SingleInputOutputTest.test
+org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$SingleInputOutputTest.test
+org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$PythonDependenciesTest.test
+
+error/stack trace:
+org.apache.beam.vendor.grpc.v1p43p2.io.grpc.StatusRuntimeException: UNAVAILABLE: io exception
+ at org.apache.beam.vendor.grpc.v1p43p2.io.grpc.stub.ClientCalls.toStatusRuntimeException(ClientCalls.java:262)
+ at org.apache.beam.vendor.grpc.v1p43p2.io.grpc.stub.ClientCalls.getUnchecked(ClientCalls.java:243)
+ at org.apache.beam.vendor.grpc.v1p43p2.io.grpc.stub.ClientCalls.blockingUnaryCall(ClientCalls.java:156)
+ at org.apache.beam.model.expansion.v1.ExpansionServiceGrpc$ExpansionServiceBlockingStub.expand(ExpansionServiceGrpc.java:169)
+ at org.apache.beam.runners.core.construction.DefaultExpansionServiceClientFactory$1.expand(DefaultExpansionServiceClientFactory.java:63)
+ at org.apache.beam.runners.core.construction.External$ExpandableTransform.expand(External.java:218)
+ at org.apache.beam.sdk.Pipeline.applyInternal(Pipeline.java:548)
+ at org.apache.beam.sdk.Pipeline.applyTransform(Pipeline.java:482)
+ at org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple.apply(KeyedPCollectionTuple.java:108)
+ at org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$ValidateRunnerXlangTestBase.coGroupByKeyTest(ValidateRunnerXlangTest.java:210)
+ at org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$CoGroupByKeyTest.test(ValidateRunnerXlangTest.java:345)
+ at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
+ at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
+ at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
+ at java.lang.reflect.Method.invoke(Method.java:498)
+ at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
+ at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
+ at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
+ at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
+ at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
+ at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
+ at org.apache.beam.sdk.testing.TestPipeline$1.evaluate(TestPipeline.java:323)
+ at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
+ at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
+ at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366)
+ at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
+ at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63)
+ at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
+ at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79)
+ at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
+ at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
+ at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293)
+ at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
+ at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
+ at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
+ at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.runTestClass(JUnitTestClassExecutor.java:110)
+ at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.execute(JUnitTestClassExecutor.java:58)
+ at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.execute(JUnitTestClassExecutor.java:38)
+ at org.gradle.api.internal.tasks.testing.junit.AbstractJUnitTestClassProcessor.processTestClass(AbstractJUnitTestClassProcessor.java:62)
+ at org.gradle.api.internal.tasks.testing.SuiteTestClassProcessor.processTestClass(SuiteTestClassProcessor.java:51)
+ at sun.reflect.GeneratedMethodAccessor10.invoke(Unknown Source)
+ at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
+ at java.lang.reflect.Method.invoke(Method.java:498)
+ at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:36)
+ at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:24)
+ at org.gradle.internal.dispatch.ContextClassLoaderDispatch.dispatch(ContextClassLoaderDispatch.java:33)
+ at org.gradle.internal.dispatch.ProxyDispatchAdapter$DispatchingInvocationHandler.invoke(ProxyDispatchAdapter.java:94)
+ at com.sun.proxy.$Proxy2.processTestClass(Unknown Source)
+ at org.gradle.api.internal.tasks.testing.worker.TestWorker$2.run(TestWorker.java:176)
+ at org.gradle.api.internal.tasks.testing.worker.TestWorker.executeAndMaintainThreadName(TestWorker.java:129)
+ at org.gradle.api.internal.tasks.testing.worker.TestWorker.execute(TestWorker.java:100)
+ at org.gradle.api.internal.tasks.testing.worker.TestWorker.execute(TestWorker.java:60)
+ at org.gradle.process.internal.worker.child.ActionExecutionWorker.execute(ActionExecutionWorker.java:56)
+ at org.gradle.process.internal.worker.child.SystemApplicationClassLoaderWorker.call(SystemApplicationClassLoaderWorker.java:133)
+ at org.gradle.process.internal.worker.child.SystemApplicationClassLoaderWorker.call(SystemApplicationClassLoaderWorker.java:71)
+ at worker.org.gradle.process.internal.worker.GradleWorkerMain.run(GradleWorkerMain.java:69)
+ at worker.org.gradle.process.internal.worker.GradleWorkerMain.main(GradleWorkerMain.java:74)
+Caused by: org.apache.beam.vendor.grpc.v1p43p2.io.netty.channel.AbstractChannel$AnnotatedConnectException: finishConnect(..) failed: Connection refused: localhost/127.0.0.1:45899
+Caused by: java.net.ConnectException: finishConnect(..) failed: Connection refused
+ at org.apache.beam.vendor.grpc.v1p43p2.io.netty.channel.unix.Errors.newConnectException0(Errors.java:155)
+ at org.apache.beam.vendor.grpc.v1p43p2.io.netty.channel.unix.Errors.handleConnectErrno(Errors.java:128)
+ at org.apache.beam.vendor.grpc.v1p43p2.io.netty.channel.unix.Socket.finishConnect(Socket.java:278)
+ at org.apache.beam.vendor.grpc.v1p43p2.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.doFinishConnect(AbstractEpollChannel.java:710)
+ at org.apache.beam.vendor.grpc.v1p43p2.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.finishConnect(AbstractEpollChannel.java:687)
+ at org.apache.beam.vendor.grpc.v1p43p2.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.epollOutReady(AbstractEpollChannel.java:567)
+ at org.apache.beam.vendor.grpc.v1p43p2.io.netty.channel.epoll.EpollEventLoop.processReady(EpollEventLoop.java:470)
+ at org.apache.beam.vendor.grpc.v1p43p2.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:378)
+ at org.apache.beam.vendor.grpc.v1p43p2.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989)
+ at org.apache.beam.vendor.grpc.v1p43p2.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
+ at org.apache.beam.vendor.grpc.v1p43p2.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
+ at java.lang.Thread.run(Thread.java:748)
+
+
+Imported from Jira [BEAM-13809](https://issues.apache.org/jira/browse/BEAM-13809). Original Jira may contain additional context.
+Reported by: ibzib.",1.0,"beam_PostCommit_XVR_Flink flaky: Connection refused - https://ci-beam.apache.org/job/beam_PostCommit_XVR_Flink/4980/
+
+The following tests failed:
+
+org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$CoGroupByKeyTest.test
+org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$CoGroupByKeyTest.test
+org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$CombineGloballyTest.test
+org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$CombineGloballyTest.test
+org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$CombinePerKeyTest.test
+org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$CombinePerKeyTest.test
+org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$FlattenTest.test
+org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$FlattenTest.test
+org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$GroupByKeyTest.test
+org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$GroupByKeyTest.test
+org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$MultiInputOutputWithSideInputTest.test
+org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$MultiInputOutputWithSideInputTest.test
+org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$PartitionTest.test
+org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$PartitionTest.test
+org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$SingleInputOutputTest.test
+org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$SingleInputOutputTest.test
+org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$PythonDependenciesTest.test
+
+error/stack trace:
+org.apache.beam.vendor.grpc.v1p43p2.io.grpc.StatusRuntimeException: UNAVAILABLE: io exception
+ at org.apache.beam.vendor.grpc.v1p43p2.io.grpc.stub.ClientCalls.toStatusRuntimeException(ClientCalls.java:262)
+ at org.apache.beam.vendor.grpc.v1p43p2.io.grpc.stub.ClientCalls.getUnchecked(ClientCalls.java:243)
+ at org.apache.beam.vendor.grpc.v1p43p2.io.grpc.stub.ClientCalls.blockingUnaryCall(ClientCalls.java:156)
+ at org.apache.beam.model.expansion.v1.ExpansionServiceGrpc$ExpansionServiceBlockingStub.expand(ExpansionServiceGrpc.java:169)
+ at org.apache.beam.runners.core.construction.DefaultExpansionServiceClientFactory$1.expand(DefaultExpansionServiceClientFactory.java:63)
+ at org.apache.beam.runners.core.construction.External$ExpandableTransform.expand(External.java:218)
+ at org.apache.beam.sdk.Pipeline.applyInternal(Pipeline.java:548)
+ at org.apache.beam.sdk.Pipeline.applyTransform(Pipeline.java:482)
+ at org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple.apply(KeyedPCollectionTuple.java:108)
+ at org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$ValidateRunnerXlangTestBase.coGroupByKeyTest(ValidateRunnerXlangTest.java:210)
+ at org.apache.beam.runners.core.construction.ValidateRunnerXlangTest$CoGroupByKeyTest.test(ValidateRunnerXlangTest.java:345)
+ at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
+ at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
+ at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
+ at java.lang.reflect.Method.invoke(Method.java:498)
+ at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
+ at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
+ at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
+ at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
+ at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
+ at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
+ at org.apache.beam.sdk.testing.TestPipeline$1.evaluate(TestPipeline.java:323)
+ at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
+ at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
+ at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366)
+ at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
+ at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63)
+ at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
+ at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79)
+ at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
+ at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
+ at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293)
+ at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
+ at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
+ at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
+ at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.runTestClass(JUnitTestClassExecutor.java:110)
+ at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.execute(JUnitTestClassExecutor.java:58)
+ at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.execute(JUnitTestClassExecutor.java:38)
+ at org.gradle.api.internal.tasks.testing.junit.AbstractJUnitTestClassProcessor.processTestClass(AbstractJUnitTestClassProcessor.java:62)
+ at org.gradle.api.internal.tasks.testing.SuiteTestClassProcessor.processTestClass(SuiteTestClassProcessor.java:51)
+ at sun.reflect.GeneratedMethodAccessor10.invoke(Unknown Source)
+ at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
+ at java.lang.reflect.Method.invoke(Method.java:498)
+ at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:36)
+ at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:24)
+ at org.gradle.internal.dispatch.ContextClassLoaderDispatch.dispatch(ContextClassLoaderDispatch.java:33)
+ at org.gradle.internal.dispatch.ProxyDispatchAdapter$DispatchingInvocationHandler.invoke(ProxyDispatchAdapter.java:94)
+ at com.sun.proxy.$Proxy2.processTestClass(Unknown Source)
+ at org.gradle.api.internal.tasks.testing.worker.TestWorker$2.run(TestWorker.java:176)
+ at org.gradle.api.internal.tasks.testing.worker.TestWorker.executeAndMaintainThreadName(TestWorker.java:129)
+ at org.gradle.api.internal.tasks.testing.worker.TestWorker.execute(TestWorker.java:100)
+ at org.gradle.api.internal.tasks.testing.worker.TestWorker.execute(TestWorker.java:60)
+ at org.gradle.process.internal.worker.child.ActionExecutionWorker.execute(ActionExecutionWorker.java:56)
+ at org.gradle.process.internal.worker.child.SystemApplicationClassLoaderWorker.call(SystemApplicationClassLoaderWorker.java:133)
+ at org.gradle.process.internal.worker.child.SystemApplicationClassLoaderWorker.call(SystemApplicationClassLoaderWorker.java:71)
+ at worker.org.gradle.process.internal.worker.GradleWorkerMain.run(GradleWorkerMain.java:69)
+ at worker.org.gradle.process.internal.worker.GradleWorkerMain.main(GradleWorkerMain.java:74)
+Caused by: org.apache.beam.vendor.grpc.v1p43p2.io.netty.channel.AbstractChannel$AnnotatedConnectException: finishConnect(..) failed: Connection refused: localhost/127.0.0.1:45899
+Caused by: java.net.ConnectException: finishConnect(..) failed: Connection refused
+ at org.apache.beam.vendor.grpc.v1p43p2.io.netty.channel.unix.Errors.newConnectException0(Errors.java:155)
+ at org.apache.beam.vendor.grpc.v1p43p2.io.netty.channel.unix.Errors.handleConnectErrno(Errors.java:128)
+ at org.apache.beam.vendor.grpc.v1p43p2.io.netty.channel.unix.Socket.finishConnect(Socket.java:278)
+ at org.apache.beam.vendor.grpc.v1p43p2.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.doFinishConnect(AbstractEpollChannel.java:710)
+ at org.apache.beam.vendor.grpc.v1p43p2.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.finishConnect(AbstractEpollChannel.java:687)
+ at org.apache.beam.vendor.grpc.v1p43p2.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.epollOutReady(AbstractEpollChannel.java:567)
+ at org.apache.beam.vendor.grpc.v1p43p2.io.netty.channel.epoll.EpollEventLoop.processReady(EpollEventLoop.java:470)
+ at org.apache.beam.vendor.grpc.v1p43p2.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:378)
+ at org.apache.beam.vendor.grpc.v1p43p2.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989)
+ at org.apache.beam.vendor.grpc.v1p43p2.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
+ at org.apache.beam.vendor.grpc.v1p43p2.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
+ at java.lang.Thread.run(Thread.java:748)
+
+
+Imported from Jira [BEAM-13809](https://issues.apache.org/jira/browse/BEAM-13809). Original Jira may contain additional context.
+Reported by: ibzib.",0,beam postcommit xvr flink flaky connection refused the following tests failed org apache beam runners core construction validaterunnerxlangtest cogroupbykeytest test org apache beam runners core construction validaterunnerxlangtest cogroupbykeytest test org apache beam runners core construction validaterunnerxlangtest combinegloballytest test org apache beam runners core construction validaterunnerxlangtest combinegloballytest test org apache beam runners core construction validaterunnerxlangtest combineperkeytest test org apache beam runners core construction validaterunnerxlangtest combineperkeytest test org apache beam runners core construction validaterunnerxlangtest flattentest test org apache beam runners core construction validaterunnerxlangtest flattentest test org apache beam runners core construction validaterunnerxlangtest groupbykeytest test org apache beam runners core construction validaterunnerxlangtest groupbykeytest test org apache beam runners core construction validaterunnerxlangtest multiinputoutputwithsideinputtest test org apache beam runners core construction validaterunnerxlangtest multiinputoutputwithsideinputtest test org apache beam runners core construction validaterunnerxlangtest partitiontest test org apache beam runners core construction validaterunnerxlangtest partitiontest test org apache beam runners core construction validaterunnerxlangtest singleinputoutputtest test org apache beam runners core construction validaterunnerxlangtest singleinputoutputtest test org apache beam runners core construction validaterunnerxlangtest pythondependenciestest test error stack trace org apache beam vendor grpc io grpc statusruntimeexception unavailable io exception at org apache beam vendor grpc io grpc stub clientcalls tostatusruntimeexception clientcalls java at org apache beam vendor grpc io grpc stub clientcalls getunchecked clientcalls java at org apache beam vendor grpc io grpc stub clientcalls blockingunarycall clientcalls java at org apache beam model expansion expansionservicegrpc expansionserviceblockingstub expand expansionservicegrpc java at org apache beam runners core construction defaultexpansionserviceclientfactory expand defaultexpansionserviceclientfactory java at org apache beam runners core construction external expandabletransform expand external java at org apache beam sdk pipeline applyinternal pipeline java at org apache beam sdk pipeline applytransform pipeline java at org apache beam sdk transforms join keyedpcollectiontuple apply keyedpcollectiontuple java at org apache beam runners core construction validaterunnerxlangtest validaterunnerxlangtestbase cogroupbykeytest validaterunnerxlangtest java at org apache beam runners core construction validaterunnerxlangtest cogroupbykeytest test validaterunnerxlangtest java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org junit internal runners statements runbefores evaluate runbefores java at org junit internal runners statements runafters evaluate runafters java at org apache beam sdk testing testpipeline evaluate testpipeline java at org junit runners parentrunner evaluate parentrunner java at org junit runners evaluate java at org junit runners parentrunner runleaf parentrunner java at org junit runners runchild java at org junit runners runchild java at org junit runners parentrunner run parentrunner java at org junit runners parentrunner schedule parentrunner java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit internal runners statements runbefores evaluate runbefores java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org gradle api internal tasks testing junit junittestclassexecutor runtestclass junittestclassexecutor java at org gradle api internal tasks testing junit junittestclassexecutor execute junittestclassexecutor java at org gradle api internal tasks testing junit junittestclassexecutor execute junittestclassexecutor java at org gradle api internal tasks testing junit abstractjunittestclassprocessor processtestclass abstractjunittestclassprocessor java at org gradle api internal tasks testing suitetestclassprocessor processtestclass suitetestclassprocessor java at sun reflect invoke unknown source at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org gradle internal dispatch reflectiondispatch dispatch reflectiondispatch java at org gradle internal dispatch reflectiondispatch dispatch reflectiondispatch java at org gradle internal dispatch contextclassloaderdispatch dispatch contextclassloaderdispatch java at org gradle internal dispatch proxydispatchadapter dispatchinginvocationhandler invoke proxydispatchadapter java at com sun proxy processtestclass unknown source at org gradle api internal tasks testing worker testworker run testworker java at org gradle api internal tasks testing worker testworker executeandmaintainthreadname testworker java at org gradle api internal tasks testing worker testworker execute testworker java at org gradle api internal tasks testing worker testworker execute testworker java at org gradle process internal worker child actionexecutionworker execute actionexecutionworker java at org gradle process internal worker child systemapplicationclassloaderworker call systemapplicationclassloaderworker java at org gradle process internal worker child systemapplicationclassloaderworker call systemapplicationclassloaderworker java at worker org gradle process internal worker gradleworkermain run gradleworkermain java at worker org gradle process internal worker gradleworkermain main gradleworkermain java caused by org apache beam vendor grpc io netty channel abstractchannel annotatedconnectexception finishconnect failed connection refused localhost caused by java net connectexception finishconnect failed connection refused at org apache beam vendor grpc io netty channel unix errors errors java at org apache beam vendor grpc io netty channel unix errors handleconnecterrno errors java at org apache beam vendor grpc io netty channel unix socket finishconnect socket java at org apache beam vendor grpc io netty channel epoll abstractepollchannel abstractepollunsafe dofinishconnect abstractepollchannel java at org apache beam vendor grpc io netty channel epoll abstractepollchannel abstractepollunsafe finishconnect abstractepollchannel java at org apache beam vendor grpc io netty channel epoll abstractepollchannel abstractepollunsafe epolloutready abstractepollchannel java at org apache beam vendor grpc io netty channel epoll epolleventloop processready epolleventloop java at org apache beam vendor grpc io netty channel epoll epolleventloop run epolleventloop java at org apache beam vendor grpc io netty util concurrent singlethreadeventexecutor run singlethreadeventexecutor java at org apache beam vendor grpc io netty util internal threadexecutormap run threadexecutormap java at org apache beam vendor grpc io netty util concurrent fastthreadlocalrunnable run fastthreadlocalrunnable java at java lang thread run thread java imported from jira original jira may contain additional context reported by ibzib ,0
+6393,14498679951.0,IssuesEvent,2020-12-11 15:48:09,ratchetphp/Ratchet,https://api.github.com/repos/ratchetphp/Ratchet,opened,Roadmap,architecture docs enhancement feature,"## v0.5
+
+Updates to the next release of Ratchet will be made against the [v0.5 branch](https://github.com/ratchetphp/Ratchet/tree/v0.5). This version will add some functionality, including a transition period, while keeping backwards compatibility. Key features for this version include:
+
+- WebSocket deflate support. Off by default. A new optional parameter to be added to `WsServer` to enable compression.
+- `ConnectionInterface` will implement [PSR-11's `ContainerInterface`](https://www.php-fig.org/psr/psr-11/). Properties from Components will be accessible via `$conn->get('HTTP.request')` as well as the current magic methodical way of `$conn->HTTP->request`.
+- Update dependencies to work with all React 1.0 libraries. We will support a range of what's supported now (0.x versions) up to 1.0. A couple of their APIs have changed in 1.0 so this may be a BC break for some people if they're also using React in their projects, hence maintaining support for the old version as well
+- Add TLS support to the App Facade (#848)
+- Consider adopting [PSR-12](https://www.php-fig.org/psr/psr-12/) in the form of a pre-commit hook or GitHub action to auto-format so the code base is consistent without having to think about it
+
+## v0.6/v1.0
+
+This version will not include any new features but have backwards compatibility breaks from old code.
+
+- Remove the magic accessors from `ConnectionInterface`. All properties set by Components are to be access via `ContainerInterface` methods. This will be a syntactic BC break but not an architectural one.
+- New version of PHP requirement (discussions to be had around which version this should be)
+- Transition return type declarations on all methods from Docblocks to language
+- Session and WAMP components will be moved to their own repositories
+- Drop support for pre 1.0 version of React dependencies
+- Determine optimal target version of Symfony libraries",1.0,"Roadmap - ## v0.5
+
+Updates to the next release of Ratchet will be made against the [v0.5 branch](https://github.com/ratchetphp/Ratchet/tree/v0.5). This version will add some functionality, including a transition period, while keeping backwards compatibility. Key features for this version include:
+
+- WebSocket deflate support. Off by default. A new optional parameter to be added to `WsServer` to enable compression.
+- `ConnectionInterface` will implement [PSR-11's `ContainerInterface`](https://www.php-fig.org/psr/psr-11/). Properties from Components will be accessible via `$conn->get('HTTP.request')` as well as the current magic methodical way of `$conn->HTTP->request`.
+- Update dependencies to work with all React 1.0 libraries. We will support a range of what's supported now (0.x versions) up to 1.0. A couple of their APIs have changed in 1.0 so this may be a BC break for some people if they're also using React in their projects, hence maintaining support for the old version as well
+- Add TLS support to the App Facade (#848)
+- Consider adopting [PSR-12](https://www.php-fig.org/psr/psr-12/) in the form of a pre-commit hook or GitHub action to auto-format so the code base is consistent without having to think about it
+
+## v0.6/v1.0
+
+This version will not include any new features but have backwards compatibility breaks from old code.
+
+- Remove the magic accessors from `ConnectionInterface`. All properties set by Components are to be access via `ContainerInterface` methods. This will be a syntactic BC break but not an architectural one.
+- New version of PHP requirement (discussions to be had around which version this should be)
+- Transition return type declarations on all methods from Docblocks to language
+- Session and WAMP components will be moved to their own repositories
+- Drop support for pre 1.0 version of React dependencies
+- Determine optimal target version of Symfony libraries",0,roadmap updates to the next release of ratchet will be made against the this version will add some functionality including a transition period while keeping backwards compatibility key features for this version include websocket deflate support off by default a new optional parameter to be added to wsserver to enable compression connectioninterface will implement properties from components will be accessible via conn get http request as well as the current magic methodical way of conn http request update dependencies to work with all react libraries we will support a range of what s supported now x versions up to a couple of their apis have changed in so this may be a bc break for some people if they re also using react in their projects hence maintaining support for the old version as well add tls support to the app facade consider adopting in the form of a pre commit hook or github action to auto format so the code base is consistent without having to think about it this version will not include any new features but have backwards compatibility breaks from old code remove the magic accessors from connectioninterface all properties set by components are to be access via containerinterface methods this will be a syntactic bc break but not an architectural one new version of php requirement discussions to be had around which version this should be transition return type declarations on all methods from docblocks to language session and wamp components will be moved to their own repositories drop support for pre version of react dependencies determine optimal target version of symfony libraries,0
+720450,24793283693.0,IssuesEvent,2022-10-24 15:13:59,AY2223S1-CS2103T-T09-4/tp,https://api.github.com/repos/AY2223S1-CS2103T-T09-4/tp,closed,"As a tutor, I can find out who are the students having lessons on a particular day ",type.Story priority.MEDIUM type.Task.Find,... so that I can know and plan my schedule ahead,1.0,"As a tutor, I can find out who are the students having lessons on a particular day - ... so that I can know and plan my schedule ahead",0,as a tutor i can find out who are the students having lessons on a particular day so that i can know and plan my schedule ahead,0
+348018,10437566242.0,IssuesEvent,2019-09-17 22:29:57,funretro/distributed,https://api.github.com/repos/funretro/distributed,closed,Timer for writing cards and voting,idea - not priority,"It could have a timer for the whole board. Everyone should be able to increase or decrease time and stop and start the timer.
+
+Voting and creating cards should only be available while timer is running.
+
+Once timer is over the cards are updated and everyone see the result.
+
+It's like a automated facilitator.
+",1.0,"Timer for writing cards and voting - It could have a timer for the whole board. Everyone should be able to increase or decrease time and stop and start the timer.
+
+Voting and creating cards should only be available while timer is running.
+
+Once timer is over the cards are updated and everyone see the result.
+
+It's like a automated facilitator.
+",0,timer for writing cards and voting it could have a timer for the whole board everyone should be able to increase or decrease time and stop and start the timer voting and creating cards should only be available while timer is running once timer is over the cards are updated and everyone see the result it s like a automated facilitator ,0
+8514,11952470740.0,IssuesEvent,2020-04-03 19:25:13,CMPUT301W20T13/SkipTheGas,https://api.github.com/repos/CMPUT301W20T13/SkipTheGas,closed,US 01.07.01 [R] - Confirm Request Completion,MEDIUM priority initial requirement rider,"**Partial User Story:**
+As a rider, I want to confirm the completion of a request and enable payment
+
+**Rationale:**
+- confirming completion of the ride notifies the system that the rider has reached their destination
+- this also triggers the payment process since the request has been successfully completed
+- let's the driver know that his service is complete and he can receive payment
+
+**Story points:**
+5
+
+**Risk-Level:**
+Low
+
+**Clarifications:**
+(_post any questions or concerns regarding this backlog item here_)",1.0,"US 01.07.01 [R] - Confirm Request Completion - **Partial User Story:**
+As a rider, I want to confirm the completion of a request and enable payment
+
+**Rationale:**
+- confirming completion of the ride notifies the system that the rider has reached their destination
+- this also triggers the payment process since the request has been successfully completed
+- let's the driver know that his service is complete and he can receive payment
+
+**Story points:**
+5
+
+**Risk-Level:**
+Low
+
+**Clarifications:**
+(_post any questions or concerns regarding this backlog item here_)",0,us confirm request completion partial user story as a rider i want to confirm the completion of a request and enable payment rationale confirming completion of the ride notifies the system that the rider has reached their destination this also triggers the payment process since the request has been successfully completed let s the driver know that his service is complete and he can receive payment story points risk level low clarifications post any questions or concerns regarding this backlog item here ,0
+5834,2610216309.0,IssuesEvent,2015-02-26 19:08:57,chrsmith/somefinders,https://api.github.com/repos/chrsmith/somefinders,opened,ответы на государственный экзамен по менеджменту для студентов сга 521500 080500.62,auto-migrated Priority-Medium Type-Defect,"```
+'''Аскольд Капустин'''
+День добрый никак не могу найти .ответы на
+государственный экзамен по менеджменту
+для студентов сга 521500 080500.62. где то видел уже
+
+'''Боеслав Симонов'''
+Вот хороший сайт где можно скачать
+http://bit.ly/1ixXesg
+
+'''Андрон Данилов'''
+Просит ввести номер мобилы!Не опасно ли это?
+
+'''Вениамин Дроздов'''
+Неа все ок у меня ничего не списало
+
+'''Викентий Никитин'''
+Не это не влияет на баланс
+
+
+
+
+
+
+Информация о файле: ответы на
+государственный экзамен по менеджменту
+для студентов сга 521500 080500.62
+Загружен: В этом месяце
+Скачан раз: 1377
+Рейтинг: 453
+Средняя скорость скачивания: 647
+Похожих файлов: 35
+
+```
+
+-----
+Original issue reported on code.google.com by `kondense...@gmail.com` on 16 Dec 2013 at 6:52",1.0,"ответы на государственный экзамен по менеджменту для студентов сга 521500 080500.62 - ```
+'''Аскольд Капустин'''
+День добрый никак не могу найти .ответы на
+государственный экзамен по менеджменту
+для студентов сга 521500 080500.62. где то видел уже
+
+'''Боеслав Симонов'''
+Вот хороший сайт где можно скачать
+http://bit.ly/1ixXesg
+
+'''Андрон Данилов'''
+Просит ввести номер мобилы!Не опасно ли это?
+
+'''Вениамин Дроздов'''
+Неа все ок у меня ничего не списало
+
+'''Викентий Никитин'''
+Не это не влияет на баланс
+
+
+
+
+
+
+Информация о файле: ответы на
+государственный экзамен по менеджменту
+для студентов сга 521500 080500.62
+Загружен: В этом месяце
+Скачан раз: 1377
+Рейтинг: 453
+Средняя скорость скачивания: 647
+Похожих файлов: 35
+
+```
+
+-----
+Original issue reported on code.google.com by `kondense...@gmail.com` on 16 Dec 2013 at 6:52",0,ответы на государственный экзамен по менеджменту для студентов сга аскольд капустин день добрый никак не могу найти ответы на государственный экзамен по менеджменту для студентов сга где то видел уже боеслав симонов вот хороший сайт где можно скачать андрон данилов просит ввести номер мобилы не опасно ли это вениамин дроздов неа все ок у меня ничего не списало викентий никитин не это не влияет на баланс информация о файле ответы на государственный экзамен по менеджменту для студентов сга загружен в этом месяце скачан раз рейтинг средняя скорость скачивания похожих файлов original issue reported on code google com by kondense gmail com on dec at ,0
+207,4312381020.0,IssuesEvent,2016-07-22 05:05:13,wordpress-mobile/WordPress-Android,https://api.github.com/repos/wordpress-mobile/WordPress-Android,closed,People Management: Long usernames do not have X mark near them.,People Management [Type] Bug,"### Expected behavior
+Long usernames have X mark near them that let's you remove them.
+
+### Actual behavior
+Long usernames do not X mark near them.
+
+
+### Steps to reproduce the behavior
+Add the long username to invite text field.
+",1.0,"People Management: Long usernames do not have X mark near them. - ### Expected behavior
+Long usernames have X mark near them that let's you remove them.
+
+### Actual behavior
+Long usernames do not X mark near them.
+
+
+### Steps to reproduce the behavior
+Add the long username to invite text field.
+",1,people management long usernames do not have x mark near them expected behavior long usernames have x mark near them that let s you remove them actual behavior long usernames do not x mark near them img width alt xmark src steps to reproduce the behavior add the long username to invite text field ,1
+859,16525213215.0,IssuesEvent,2021-05-26 19:08:52,microsoft/botframework-components,https://api.github.com/repos/microsoft/botframework-components,closed,Enterprise People: Evaluate MessageAbout feature,Area: Generators: Enterprise People Owner: PBX backlog,"Evaluate if the MessageAbout feature can be completed using Microsoft Graph, and if not, what are the available options.
+
+MessageAbout | User wants to know who send messages about a topic | Who did I message about dyson television last week?
+-- | -- | --
+
+
+",1.0,"Enterprise People: Evaluate MessageAbout feature - Evaluate if the MessageAbout feature can be completed using Microsoft Graph, and if not, what are the available options.
+
+MessageAbout | User wants to know who send messages about a topic | Who did I message about dyson television last week?
+-- | -- | --
+
+
+",1,enterprise people evaluate messageabout feature evaluate if the messageabout feature can be completed using microsoft graph and if not what are the available options messageabout user wants to know who send messages about a topic who did i message about dyson television last week ,1
+18109,12796173145.0,IssuesEvent,2020-07-02 09:58:48,blyndusk/PHRH-PWA,https://api.github.com/repos/blyndusk/PHRH-PWA,closed,"[OPS] - Update workflows & actions ( merge master, push develop )",Infrastructure enhancement,"- CREATE an action for each merge to master -> put in production -> ansible
+- CEATE an action for each push into develop -> lint, build, test",1.0,"[OPS] - Update workflows & actions ( merge master, push develop ) - - CREATE an action for each merge to master -> put in production -> ansible
+- CEATE an action for each push into develop -> lint, build, test",0, update workflows actions merge master push develop create an action for each merge to master put in production ansible ceate an action for each push into develop lint build test,0
+97560,11017305422.0,IssuesEvent,2019-12-05 08:05:52,agriculture-x-iot/Agrimaster_app,https://api.github.com/repos/agriculture-x-iot/Agrimaster_app,closed,flutterのバージョンアップによる問題,documentation,"### flutterのバージョンアップ方法
+
+- `flutter channel dev`
+
+- `flutter upgrade`
+
+`flutter doctor`を実行して以下が出ればひとまずOK!
+(`flutter upgrade`をするとついでに`flutter doctor`してくれる)
+
+`[√] Flutter (Channel dev, v1.12.16, on Microsoft Windows [Version 10.0.18362.476], locale ja-JP)`
+
+### アップグレードに伴う問題を解決
+
+アプリ内で以下を実行
+
+- `flutter clean`
+
+- `flutter packages get`
+
+これでもエラーがあったら
+
+**pubspec.yml**に以下を追記
+
+> module:
+> androidX: true
+
+**android/gradle.properties**にも追記
+
+> android.useAndroidX=true
+> android.enableJetifier=true",1.0,"flutterのバージョンアップによる問題 - ### flutterのバージョンアップ方法
+
+- `flutter channel dev`
+
+- `flutter upgrade`
+
+`flutter doctor`を実行して以下が出ればひとまずOK!
+(`flutter upgrade`をするとついでに`flutter doctor`してくれる)
+
+`[√] Flutter (Channel dev, v1.12.16, on Microsoft Windows [Version 10.0.18362.476], locale ja-JP)`
+
+### アップグレードに伴う問題を解決
+
+アプリ内で以下を実行
+
+- `flutter clean`
+
+- `flutter packages get`
+
+これでもエラーがあったら
+
+**pubspec.yml**に以下を追記
+
+> module:
+> androidX: true
+
+**android/gradle.properties**にも追記
+
+> android.useAndroidX=true
+> android.enableJetifier=true",0,flutterのバージョンアップによる問題 flutterのバージョンアップ方法 flutter channel dev flutter upgrade flutter doctor を実行して以下が出ればひとまずok! flutter upgrade をするとついでに flutter doctor してくれる flutter channel dev on microsoft windows locale ja jp アップグレードに伴う問題を解決 アプリ内で以下を実行 flutter clean flutter packages get これでもエラーがあったら pubspec yml に以下を追記 module androidx true android gradle properties にも追記 android useandroidx true android enablejetifier true,0
+227,4634653841.0,IssuesEvent,2016-09-29 02:22:18,newtheatre/history-project,https://api.github.com/repos/newtheatre/history-project,opened,Amalgamate similar courses for sorting and filtering of people,people,"People's courses tend to be quite specific:
+
+
+
+Careers also may have discrepancies between similar fields. It would be nice to group these similar fields and courses together for the people index.
+
+We could implement a `sort_career` attribute to specify a less specific sorting parameter: Mechnical Engineering, MEng Mechanical Engineering, Mechanical Engineering MSc would all become 'Mechanical Engineering'. Or perhaps even just 'Engineering'. But this would be more work than the alternative:
+
+Dynamically set a generic course by stripping the MScs, BAs *&c*, then matching what's left with a lookup table.
+
+I don't have enough data to do careers yet, we shall investigate them when more bios come in.",1.0,"Amalgamate similar courses for sorting and filtering of people - People's courses tend to be quite specific:
+
+
+
+Careers also may have discrepancies between similar fields. It would be nice to group these similar fields and courses together for the people index.
+
+We could implement a `sort_career` attribute to specify a less specific sorting parameter: Mechnical Engineering, MEng Mechanical Engineering, Mechanical Engineering MSc would all become 'Mechanical Engineering'. Or perhaps even just 'Engineering'. But this would be more work than the alternative:
+
+Dynamically set a generic course by stripping the MScs, BAs *&c*, then matching what's left with a lookup table.
+
+I don't have enough data to do careers yet, we shall investigate them when more bios come in.",1,amalgamate similar courses for sorting and filtering of people people s courses tend to be quite specific careers also may have discrepancies between similar fields it would be nice to group these similar fields and courses together for the people index we could implement a sort career attribute to specify a less specific sorting parameter mechnical engineering meng mechanical engineering mechanical engineering msc would all become mechanical engineering or perhaps even just engineering but this would be more work than the alternative dynamically set a generic course by stripping the mscs bas c then matching what s left with a lookup table i don t have enough data to do careers yet we shall investigate them when more bios come in ,1
+364708,10772459554.0,IssuesEvent,2019-11-02 14:54:27,grpc/grpc,https://api.github.com/repos/grpc/grpc,opened,Can't build C++ helloworld example,kind/bug priority/P2,"
+
+### What version of gRPC and what language are you using?
+
+C++; latest version of gRPC
+
+
+### What operating system (Linux, Windows,...) and version?
+Linux (64-bit)
+
+
+### What runtime / compiler are you using (e.g. python version or version of gcc)
+g++
+
+
+### What did you do?
+```
+$ cd grpc/examples/cpp/helloworld
+$ make
+```
+
+### What did you expect to see?
+I expected a clean build so I can run the server/client
+
+### What did you see instead?
+```
+protoc -I ../../protos --cpp_out=. ../../protos/helloworld.proto
+g++ -std=c++11 `pkg-config --cflags protobuf grpc` -c -o helloworld.pb.o helloworld.pb.cc
+Package grpc was not found in the pkg-config search path.
+Perhaps you should add the directory containing `grpc.pc'
+to the PKG_CONFIG_PATH environment variable
+Package 'grpc', required by 'virtual:world', not found
+protoc -I ../../protos --grpc_out=. --plugin=protoc-gen-grpc=`which grpc_cpp_plugin` ../../protos/helloworld.proto
+g++ -std=c++11 `pkg-config --cflags protobuf grpc` -c -o helloworld.grpc.pb.o helloworld.grpc.pb.cc
+Package grpc was not found in the pkg-config search path.
+Perhaps you should add the directory containing `grpc.pc'
+to the PKG_CONFIG_PATH environment variable
+Package 'grpc', required by 'virtual:world', not found
+g++ -std=c++11 `pkg-config --cflags protobuf grpc` -c -o greeter_client.o greeter_client.cc
+Package grpc was not found in the pkg-config search path.
+Perhaps you should add the directory containing `grpc.pc'
+to the PKG_CONFIG_PATH environment variable
+Package 'grpc', required by 'virtual:world', not found
+g++ helloworld.pb.o helloworld.grpc.pb.o greeter_client.o -L/usr/local/lib `pkg-config --libs protobuf grpc++` -pthread -Wl,--no-as-needed -lgrpc++_reflection -Wl,--as-needed -ldl -o greeter_client
+Package grpc++ was not found in the pkg-config search path.
+Perhaps you should add the directory containing `grpc++.pc'
+to the PKG_CONFIG_PATH environment variable
+Package 'grpc++', required by 'virtual:world', not found
+/usr/bin/ld: helloworld.pb.o: undefined reference to symbol '_ZTIN6google8protobuf7MessageE'
+/usr/bin/ld: /usr/lib/libprotobuf.so.18: error adding symbols: DSO missing from command line
+collect2: error: ld returned 1 exit status
+make: *** [Makefile:44: greeter_client] Error 1
+```
+",1.0,"Can't build C++ helloworld example -
+
+### What version of gRPC and what language are you using?
+
+C++; latest version of gRPC
+
+
+### What operating system (Linux, Windows,...) and version?
+Linux (64-bit)
+
+
+### What runtime / compiler are you using (e.g. python version or version of gcc)
+g++
+
+
+### What did you do?
+```
+$ cd grpc/examples/cpp/helloworld
+$ make
+```
+
+### What did you expect to see?
+I expected a clean build so I can run the server/client
+
+### What did you see instead?
+```
+protoc -I ../../protos --cpp_out=. ../../protos/helloworld.proto
+g++ -std=c++11 `pkg-config --cflags protobuf grpc` -c -o helloworld.pb.o helloworld.pb.cc
+Package grpc was not found in the pkg-config search path.
+Perhaps you should add the directory containing `grpc.pc'
+to the PKG_CONFIG_PATH environment variable
+Package 'grpc', required by 'virtual:world', not found
+protoc -I ../../protos --grpc_out=. --plugin=protoc-gen-grpc=`which grpc_cpp_plugin` ../../protos/helloworld.proto
+g++ -std=c++11 `pkg-config --cflags protobuf grpc` -c -o helloworld.grpc.pb.o helloworld.grpc.pb.cc
+Package grpc was not found in the pkg-config search path.
+Perhaps you should add the directory containing `grpc.pc'
+to the PKG_CONFIG_PATH environment variable
+Package 'grpc', required by 'virtual:world', not found
+g++ -std=c++11 `pkg-config --cflags protobuf grpc` -c -o greeter_client.o greeter_client.cc
+Package grpc was not found in the pkg-config search path.
+Perhaps you should add the directory containing `grpc.pc'
+to the PKG_CONFIG_PATH environment variable
+Package 'grpc', required by 'virtual:world', not found
+g++ helloworld.pb.o helloworld.grpc.pb.o greeter_client.o -L/usr/local/lib `pkg-config --libs protobuf grpc++` -pthread -Wl,--no-as-needed -lgrpc++_reflection -Wl,--as-needed -ldl -o greeter_client
+Package grpc++ was not found in the pkg-config search path.
+Perhaps you should add the directory containing `grpc++.pc'
+to the PKG_CONFIG_PATH environment variable
+Package 'grpc++', required by 'virtual:world', not found
+/usr/bin/ld: helloworld.pb.o: undefined reference to symbol '_ZTIN6google8protobuf7MessageE'
+/usr/bin/ld: /usr/lib/libprotobuf.so.18: error adding symbols: DSO missing from command line
+collect2: error: ld returned 1 exit status
+make: *** [Makefile:44: greeter_client] Error 1
+```
+",0,can t build c helloworld example what version of grpc and what language are you using c latest version of grpc what operating system linux windows and version linux bit what runtime compiler are you using e g python version or version of gcc g what did you do cd grpc examples cpp helloworld make what did you expect to see i expected a clean build so i can run the server client what did you see instead protoc i protos cpp out protos helloworld proto g std c pkg config cflags protobuf grpc c o helloworld pb o helloworld pb cc package grpc was not found in the pkg config search path perhaps you should add the directory containing grpc pc to the pkg config path environment variable package grpc required by virtual world not found protoc i protos grpc out plugin protoc gen grpc which grpc cpp plugin protos helloworld proto g std c pkg config cflags protobuf grpc c o helloworld grpc pb o helloworld grpc pb cc package grpc was not found in the pkg config search path perhaps you should add the directory containing grpc pc to the pkg config path environment variable package grpc required by virtual world not found g std c pkg config cflags protobuf grpc c o greeter client o greeter client cc package grpc was not found in the pkg config search path perhaps you should add the directory containing grpc pc to the pkg config path environment variable package grpc required by virtual world not found g helloworld pb o helloworld grpc pb o greeter client o l usr local lib pkg config libs protobuf grpc pthread wl no as needed lgrpc reflection wl as needed ldl o greeter client package grpc was not found in the pkg config search path perhaps you should add the directory containing grpc pc to the pkg config path environment variable package grpc required by virtual world not found usr bin ld helloworld pb o undefined reference to symbol usr bin ld usr lib libprotobuf so error adding symbols dso missing from command line error ld returned exit status make error ,0
+902,18817327890.0,IssuesEvent,2021-11-10 01:47:24,Another-Glass/Assignment3_Team,https://api.github.com/repos/Another-Glass/Assignment3_Team,closed,[Post] 사용자 회원가입 로직 구성,:fire: Feat :people_holding_hands: All,"# Description
+
+회원가입 로직 구현
+
+## Progress
+
+- [x] 회원가입 로직 설정 (비밀번호 해시화를 위한 salt, encrypt 설정)
+- [x] 회원가입 로직 계층 분리
+
+### Related content
+",1.0,"[Post] 사용자 회원가입 로직 구성 - # Description
+
+회원가입 로직 구현
+
+## Progress
+
+- [x] 회원가입 로직 설정 (비밀번호 해시화를 위한 salt, encrypt 설정)
+- [x] 회원가입 로직 계층 분리
+
+### Related content
+",1, 사용자 회원가입 로직 구성 description 회원가입 로직 구현 progress 회원가입 로직 설정 비밀번호 해시화를 위한 salt encrypt 설정 회원가입 로직 계층 분리 related content ,1
+200,4235848353.0,IssuesEvent,2016-07-05 16:25:45,wordpress-mobile/WordPress-Android,https://api.github.com/repos/wordpress-mobile/WordPress-Android,opened,People: Backspace v. tapping to remove a username/email address from an invitation,People Management [Type] Bug,"### Expected behavior
+
+When inviting a new user, after entering a username or email address in the invitation form I expect to be able to backspace or tap the ""x"" to remove them. (Both methods work in Calypso web for removing an invited user from the invitation form.)
+
+### Actual behavior
+
+I can only tap the ""x"" to remove the username or email address I entered in the invitation. Although I have a cursor (for entering more in the username/email field) I can't backspace to remove a previous entry in that field:
+
+
+
+
+### Steps to reproduce the behavior
+
+1. In the app, go to the site dashboard > People
+2. Tap the plus icon to invite a new user
+3. Enter a username or email address in the ""Usernames or Emails"" field
+4. Tap enter or tap outside the ""Usernames or Emails"" field to complete the entry (an ""x"" will appear next to the username/email you entered)
+5. Backspace to try to remove that username/email. Result: The cursor can't backspace over that username/email.
+6. Tap the ""x"" to try to remove that username/email. Result: the username/email is removed.
+
+##### Tested on Nexus 9, Android 6.0.1, WPAndroid alpha-16
+",1.0,"People: Backspace v. tapping to remove a username/email address from an invitation - ### Expected behavior
+
+When inviting a new user, after entering a username or email address in the invitation form I expect to be able to backspace or tap the ""x"" to remove them. (Both methods work in Calypso web for removing an invited user from the invitation form.)
+
+### Actual behavior
+
+I can only tap the ""x"" to remove the username or email address I entered in the invitation. Although I have a cursor (for entering more in the username/email field) I can't backspace to remove a previous entry in that field:
+
+
+
+
+### Steps to reproduce the behavior
+
+1. In the app, go to the site dashboard > People
+2. Tap the plus icon to invite a new user
+3. Enter a username or email address in the ""Usernames or Emails"" field
+4. Tap enter or tap outside the ""Usernames or Emails"" field to complete the entry (an ""x"" will appear next to the username/email you entered)
+5. Backspace to try to remove that username/email. Result: The cursor can't backspace over that username/email.
+6. Tap the ""x"" to try to remove that username/email. Result: the username/email is removed.
+
+##### Tested on Nexus 9, Android 6.0.1, WPAndroid alpha-16
+",1,people backspace v tapping to remove a username email address from an invitation expected behavior when inviting a new user after entering a username or email address in the invitation form i expect to be able to backspace or tap the x to remove them both methods work in calypso web for removing an invited user from the invitation form actual behavior i can only tap the x to remove the username or email address i entered in the invitation although i have a cursor for entering more in the username email field i can t backspace to remove a previous entry in that field steps to reproduce the behavior in the app go to the site dashboard people tap the plus icon to invite a new user enter a username or email address in the usernames or emails field tap enter or tap outside the usernames or emails field to complete the entry an x will appear next to the username email you entered backspace to try to remove that username email result the cursor can t backspace over that username email tap the x to try to remove that username email result the username email is removed tested on nexus android wpandroid alpha ,1
+107647,16761694748.0,IssuesEvent,2021-06-13 22:59:33,gms-ws-demo/nibrs,https://api.github.com/repos/gms-ws-demo/nibrs,opened,CVE-2019-16276 (High) detected in gccgcc-9.3.0,security vulnerability,"## CVE-2019-16276 - High Severity Vulnerability
+ Vulnerable Library - gccgcc-9.3.0
+
+
+
+
+
+",0,cve high detected in gccgcc cve high severity vulnerability vulnerable library gccgcc gnu distributions library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details go before and x before allow http request smuggling publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution ,0
+182317,30830325970.0,IssuesEvent,2023-08-02 00:49:24,elementor/elementor,https://api.github.com/repos/elementor/elementor,closed,✅ 🐞 Bug Report: Video lightbox gets the wrong aspect ratio when more then one video widget is used on the same page - [ED-10146],bug component/lightbox status/merged component/video type/design 🚀 shipped mod* product/3.15,"### Prerequisites
+
+- [X] I have searched for similar issues in both open and closed tickets and cannot find a duplicate.
+- [X] The issue still exists against the latest stable version of Elementor.
+
+### Description
+
+```html
+
+
+
+```
+
+The problem is caused by the .elementor-aspect-ratio-class.
+
+Each dialog widget should have only one .elementor-aspect-ratio class to set the aspect ratio. But in some cases there are two or more of them, so they overwrite each other. This leads to an incorrect aspect ratio of the lightbox video for some videos. In fact, the videos that are clicked second or third get the aspect ratio of the video that was clicked first. It shouldn't be like this!
+
+### Steps to reproduce
+
+build it
+
+1. Place more than one video widget on the page
+2. Enable the lightbox option for each video widget
+3. Give each video widget a different aspect ratio
+
+try it
+
+1. Click on the first video: The aspect ratio of the lightbox video is correct
+2. Click on the second video: The lightbox video's aspect ratio is incorrect, it will get the aspect ratio of the video that was clicked first
+
+### Isolating the problem
+
+- [ ] This bug happens with only Elementor plugin active (and Elementor Pro).
+- [ ] This bug happens with a Blank WordPress theme active ([Hello theme](https://wordpress.org/themes/hello-elementor/)).
+- [X] I can reproduce this bug consistently following the steps above.
+
+### System Info
+
+
+System Info
+
+````txt
+
+
+
+== Server Environment ==
+ Operating System: Linux
+ Software: nginx/1.22.0
+ MySQL version: mariadb.org binary distribution v10.4.28
+ PHP Version: 7.4.33
+ PHP Memory Limit: 128M
+ PHP Max Input Vars: 10000
+ PHP Max Post Size: 2000M
+ GD Installed: Yes
+ ZIP Installed: Yes
+ Write Permissions: All right
+ Elementor Library: Connected
+
+== WordPress Environment ==
+ Version: 6.1.1
+ Site URL: https://mirkoappel.de
+ Home URL: https://mirkoappel.de
+ WP Multisite: No
+ Max Upload Size: 2 GB
+ Memory limit: 128M
+ Max Memory limit: 128M
+ Permalink Structure: /news/%postname%/
+ Language: de-DE
+ Timezone: Europe/Berlin
+ Admin Email: [REDACTED]
+ Debug Mode: Inactive
+
+== Theme ==
+ Name: feather
+ Version: 1.0
+ Author: Mirko Appel
+ Child Theme: No
+
+== User ==
+ Role: administrator
+ WP Profile lang: de_DE
+ User Agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/111.0.0.0 Safari/537.36
+
+== Active Plugins ==
+ Admin Columns
+ Version: 4.5.5
+ Author: AdminColumns.com
+
+ Admin Menu Editor
+ Version: 1.10.4
+ Author: Janis Elsts
+
+ Advanced Custom Fields PRO
+ Version: 5.12.2
+ Author: Delicious Brains
+
+ Code Snippets
+ Version: 3.2.2
+ Author: Code Snippets Pro
+
+ Disable Gutenberg
+ Version: 2.9
+ Author: Jeff Starr
+
+ DynamicConditions
+ Version: 1.6.0
+ Author: RTO GmbH
+
+ Ele Custom Skin
+ Version: 3.1.7
+ Author: Dudaster.com
+
+ Elementor
+ Version: 3.11.4
+ Author: Elementor.com
+
+ Elementor Pro
+ Version: 3.11.5
+ Author: Elementor.com
+
+ Extras for Elementor
+ Version: 2.2.51
+ Author: Namogo
+
+ Intuitive Custom Post Order
+ Version: 3.1.4.1
+ Author: hijiri
+
+ Page scroll to id
+ Version: 1.7.7
+ Author: malihu
+
+ Post Types Unlimited
+ Version: 1.1
+ Author: WPExplorer
+
+ Regenerate Thumbnails
+ Version: 3.1.5
+ Author: Alex Mills (Viper007Bond)
+
+ WebP Express
+ Version: 0.25.5
+ Author: Bjørn Rosell
+
+ Yoast SEO
+ Version: 20.2.1
+ Author: Team Yoast
+
+
+== Eigenschaften ==
+ Custom Fonts: 0
+ Custom Icons: 0
+
+== Integrations ==
+
+
+
+== Elementor-Experimente ==
+ Optimierte DOM Ausgabe: Standardmäßig aktiviert
+ Verbessertes Laden von Assets: Aktiv
+ Verbessertes Laden von CSS: Aktiv
+ Inline-Schriftarten-Symbole: Standardmäßig deaktiviert
+ Verbesserungen der Zugänglichkeit: Standardmäßig aktiviert
+ Zusätzliche benutzerdefinierte Breakpoints: Inaktiv
+ admin_menu_rearrangement: Standardmäßig deaktiviert
+ Flexbox Container: Aktiv
+ Upgrade Swiper Library: Standardmäßig deaktiviert
+ Default to New Theme Builder: Standardmäßig aktiviert
+ Editor V2: Standardmäßig deaktiviert
+ Startseiten: Inaktiv
+ Nested Elements: Standardmäßig deaktiviert
+ Lazy Load Background Images: Standardmäßig deaktiviert
+ Als Standard speichern: Standardmäßig aktiviert
+ Page Transitions: Inaktiv
+ Notes: Standardmäßig aktiviert
+ Loop: Standardmäßig aktiviert
+ Form Submissions: Standardmäßig aktiviert
+ Scroll Snap: Inaktiv
+ Mega Menu: Standardmäßig deaktiviert
+
+
+== Protokoll ==
+
+Log: showing 16 of 162023-02-28 09:20:50 [info] Elementor data updater process has been completed. [array (
+ 'plugin' => 'Elementor Pro',
+ 'from' => '3.7.3',
+ 'to' => '3.11.3',
+)]
+2023-03-08 18:47:17 [info] elementor::elementor_updater Started
+2023-03-08 18:47:17 [info] Elementor/Upgrades - _on_each_version Start
+2023-03-08 18:47:17 [info] Elementor/Upgrades - _on_each_version Finished
+2023-03-08 18:47:17 [info] Elementor data updater process has been completed. [array (
+ 'plugin' => 'Elementor',
+ 'from' => '3.11.2',
+ 'to' => '3.11.3',
+)]
+2023-03-08 18:47:17 [info] Elementor data updater process has been queued. [array (
+ 'plugin' => 'Elementor',
+ 'from' => '3.11.2',
+ 'to' => '3.11.3',
+)]
+2023-03-12 17:32:58 [info] Elementor data updater process has been queued. [array (
+ 'plugin' => 'Elementor',
+ 'from' => '3.11.3',
+ 'to' => '3.11.4',
+)]
+2023-03-12 17:32:59 [info] elementor::elementor_updater Started
+2023-03-12 17:32:59 [info] Elementor/Upgrades - _on_each_version Finished
+2023-03-12 17:32:59 [info] Elementor/Upgrades - _on_each_version Start
+2023-03-12 17:32:59 [info] Elementor data updater process has been completed. [array (
+ 'plugin' => 'Elementor',
+ 'from' => '3.11.3',
+ 'to' => '3.11.4',
+)]
+2023-03-14 00:01:45 [info] elementor-pro::elementor_pro_updater Started
+2023-03-14 00:01:45 [info] Elementor Pro/Upgrades - _on_each_version Start
+2023-03-14 00:01:45 [info] Elementor Pro/Upgrades - _on_each_version Finished
+2023-03-14 00:01:45 [info] Elementor data updater process has been completed. [array (
+ 'plugin' => 'Elementor Pro',
+ 'from' => '3.11.3',
+ 'to' => '3.11.5',
+)]
+2023-03-14 00:01:45 [info] Elementor data updater process has been queued. [array (
+ 'plugin' => 'Elementor Pro',
+ 'from' => '3.11.3',
+ 'to' => '3.11.5',
+)]
+
+PHP: showing 10 of 10PHP: 2023-02-28 09:27:09 [notice X 57][/home/wp/disk/wordpress/wp-content/plugins/elementor/core/common/modules/ajax/module.php::175] Undefined index: data [array (
+ 'trace' => '
+#0: Elementor\Core\Logger\Manager -> shutdown()
+',
+)]
+PHP: 2023-02-28 09:29:53 [notice X 7][/home/wp/disk/wordpress/wp-content/plugins/elementor-pro/modules/theme-builder/classes/conditions-manager.php::168] Undefined index: editor_post_id [array (
+ 'trace' => '
+#0: Elementor\Core\Logger\Manager -> shutdown()
+',
+)]
+PHP: 2023-02-28 09:34:56 [notice X 47][/home/wp/disk/wordpress/wp-content/plugins/elementor/app/modules/import-export/module.php::629] Trying to get property 'label' of non-object [array (
+ 'trace' => '
+#0: Elementor\Core\Logger\Manager -> shutdown()
+',
+)]
+PHP: 2023-02-28 09:34:59 [notice X 69][/home/wp/disk/wordpress/wp-content/plugins/elementor-pro/core/app/modules/site-editor/data/endpoints/templates.php::150] Undefined index: condition_type [array (
+ 'trace' => '
+#0: /home/wp/disk/wordpress/wp-content/plugins/elementor-pro/core/app/modules/site-editor/data/endpoints/templates.php(150): Elementor\Core\Logger\Manager -> rest_error_handler()
+#1: ElementorPro\Core\App\Modules\SiteEditor\Data\Endpoints\Templates -> normalize_template_json_item()
+#2: /home/wp/disk/wordpress/wp-content/plugins/elementor-pro/core/app/modules/site-editor/data/endpoints/templates.php(120): class type array_map()
+#3: /home/wp/disk/wordpress/wp-content/plugins/elementor-pro/core/app/modules/site-editor/data/endpoints/templates.php(59): ElementorPro\Core\App\Modules\SiteEditor\Data\Endpoints\Templates -> normalize_templates_json()
+#4: /home/wp/disk/wordpress/wp-content/plugins/elementor/data/base/endpoint.php(158): ElementorPro\Core\App\Modules\SiteEditor\Data\Endpoints\Templates -> get_items()
+',
+)]
+PHP: 2023-02-28 23:24:47 [notice X 17][/home/wp/disk/wordpress/wp-content/plugins/elementor-pro/modules/posts/skins/skin-content-base.php::264] Trying to get property 'ID' of non-object [array (
+ 'trace' => '
+#0: Elementor\Core\Logger\Manager -> shutdown()
+',
+)]
+PHP: 2023-03-02 15:36:28 [notice X 4][/home/wp/disk/wordpress/wp-content/plugins/elementor/includes/frontend.php::1436] Trying to get property 'post_excerpt' of non-object [array (
+ 'trace' => '
+#0: Elementor\Core\Logger\Manager -> shutdown()
+',
+)]
+PHP: 2023-03-04 00:49:57 [error X 1][/home/wp/disk/wordpress/wp-content/plugins/elementor-pro/modules/theme-builder/classes/conditions-manager.php::295] Uncaught Error: Call to a member function update_meta() on bool in /home/wp/disk/wordpress/wp-content/plugins/elementor-pro/modules/theme-builder/classes/conditions-manager.php:295
+Stack trace:
+#0 /home/wp/disk/wordpress/wp-content/plugins/elementor-pro/modules/theme-builder/classes/conditions-manager.php(168): ElementorPro\Modules\ThemeBuilder\Classes\Conditions_Manager->save_conditions()
+#1 [internal function]: ElementorPro\Modules\ThemeBuilder\Classes\Conditions_Manager->ajax_save_theme_template_conditions()
+#2 /home/wp/disk/wordpress/wp-content/plugins/elementor/core/common/modules/ajax/module.php(175): call_user_func()
+#3 /home/wp/disk/wordpress/wp-includes/class-wp-hook.php(308): Elementor\Core\Common\Modules\Ajax\Module->handle_ajax_request()
+#4 /home/wp/disk/wordpress/wp-includes/class-wp-hook.php(332): WP_Hook->apply_filters()
+#5 /home/wp/disk/wordpress/wp-includes/plugin.php(517): WP_Hook->do_action()
+#6 /home/wp/disk/wordpress/wp-admin/admin-ajax.php(188): do_action()
+#7 {main}
+ thrown [array (
+ 'trace' => '
+#0: Elementor\Core\Logger\Manager -> shutdown()
+',
+)]
+PHP: 2023-03-07 11:32:45 [notice X 94][/home/wp/disk/wordpress/wp-content/plugins/elementor/includes/controls/groups/image-size.php::110] Trying to access array offset on value of type null [array (
+ 'trace' => '
+#0: Elementor\Core\Logger\Manager -> shutdown()
+',
+)]
+PHP: 2023-03-12 17:33:03 [warning X 1][/home/wp/disk/wordpress/wp-content/plugins/elementor/core/files/manager.php::111] unlink(/home/wp/disk/wordpress/wp-content/uploads/elementor/css/post-17558.css): No such file or directory [array (
+ 'trace' => '
+#0: Elementor\Core\Logger\Manager -> shutdown()
+',
+)]
+PHP: 2023-03-12 17:42:53 [notice X 23][/home/wp/disk/wordpress/wp-content/plugins/elementor-pro/modules/dynamic-tags/acf/tags/acf-text.php::33] Undefined offset: 1 [array (
+ 'trace' => '
+#0: Elementor\Core\Logger\Manager -> shutdown()
+',
+)]
+
+JS: showing 16 of 16JS: 2023-02-28 08:27:15 [error X 7][https://mirkoappel.de/wp-content/plugins/elementor/assets/js/editor.min.js?ver=3.11.2:3:1087291] Cannot read properties of null (reading 'getBoundingClientRect')
+JS: 2023-02-28 09:10:26 [error X 598][https://mirkoappel.de/wp-includes/js/jquery/jquery.min.js?ver=3.6.1:2:51824] Cannot read properties of undefined (reading 'replace')
+JS: 2023-03-02 10:37:36 [error X 3][https://mirkoappel.de/wp-content/plugins/elementor/assets/lib/backbone/backbone.marionette.min.js?ver=2.4.5.e1:24:19952] View (cid: ""view356"") has already been destroyed and cannot be used.
+JS: 2023-03-02 10:44:18 [error X 7][https://mirkoappel.de/wp-admin/post.php?post=16474&action=elementor#:1:1] Identifier '$dcWidgetIdInput' has already been declared
+JS: 2023-03-02 10:48:31 [error X 1][https://mirkoappel.de/wp-content/plugins/elementor/assets/lib/backbone/backbone.marionette.min.js?ver=2.4.5.e1:24:19952] View (cid: ""view27440"") has already been destroyed and cannot be used.
+JS: 2023-03-02 11:40:47 [error X 115][https://mirkoappel.de/wp-content/plugins/elementor-pro/assets/lib/sticky/jquery.sticky.min.js?ver=3.11.3:1:1162] Failed to execute 'getComputedStyle' on 'Window': parameter 1 is not of type 'Element'.
+JS: 2023-03-02 14:05:47 [error X 2][https://mirkoappel.de/wp-content/plugins/elementor/assets/js/editor.min.js?ver=3.11.2:3:931389] Cannot read properties of undefined (reading 'settings')
+JS: 2023-03-03 12:08:55 [error X 2][https://mirkoappel.de/wp-content/plugins/elementor/assets/js/editor.min.js?ver=3.11.2:3:907724] Cannot read properties of undefined (reading 'route')
+JS: 2023-03-04 08:30:07 [error X 4][https://mirkoappel.de/wp-content/plugins/elementor/assets/js/editor.min.js?ver=3.11.2:3:656177] elementorFrontend.elements.window.jQuery is not a function
+JS: 2023-03-04 08:30:07 [error X 4][https://mirkoappel.de/wp-content/plugins/elementor-extras/assets/js/frontend.min.js?ver=2.2.51:1:4462] Cannot read properties of undefined (reading '$previewContents')
+JS: 2023-03-04 23:40:33 [error X 1][https://mirkoappel.de/wp-content/plugins/elementor/assets/js/editor.min.js?ver=3.11.2:3:676196] Cannot read properties of undefined (reading 'get')
+JS: 2023-03-06 10:38:43 [error X 11][https://mirkoappel.de/wp-content/plugins/elementor/assets/lib/backbone/backbone.marionette.min.js?ver=2.4.5.e1:24:19952] View (cid: ""view284"") has already been destroyed and cannot be used.
+JS: 2023-03-06 10:41:09 [error X 814][https://mirkoappel.de/wp-content/plugins/elementor-pro/assets/js/frontend.min.js?ver=3.11.3:2:10935] Failed to execute 'observe' on 'IntersectionObserver': parameter 1 is not of type 'Element'.
+JS: 2023-03-07 16:31:27 [error X 4][https://mirkoappel.de/wp-content/plugins/elementor/assets/js/editor.min.js?ver=3.11.2:3:791668] Cannot read properties of undefined (reading 'localized')
+JS: 2023-03-13 18:01:30 [error X 1][https://mirkoappel.de/wp-content/plugins/elementor/assets/js/editor-loader-v1.min.js?ver=3.11.4:2:37] Cannot read properties of undefined (reading 'start')
+JS: 2023-03-13 18:01:30 [error X 1][https://mirkoappel.de/wp-content/plugins/elementor/assets/js/container-converter.min.js?ver=3.11.4:2:5351] elementor is not defined
+
+
+
+== Elementor - Compatibility Tag ==
+
+ Elementor Pro: Die Kompatibilität ist nicht angegeben
+ Extras for Elementor: Die Kompatibilität ist nicht angegeben
+
+== Elementor Pro - Compatibility Tag ==
+
+
+
+
+````
+
+
+",1.0,"✅ 🐞 Bug Report: Video lightbox gets the wrong aspect ratio when more then one video widget is used on the same page - [ED-10146] - ### Prerequisites
+
+- [X] I have searched for similar issues in both open and closed tickets and cannot find a duplicate.
+- [X] The issue still exists against the latest stable version of Elementor.
+
+### Description
+
+```html
+
+
+
+```
+
+The problem is caused by the .elementor-aspect-ratio-class.
+
+Each dialog widget should have only one .elementor-aspect-ratio class to set the aspect ratio. But in some cases there are two or more of them, so they overwrite each other. This leads to an incorrect aspect ratio of the lightbox video for some videos. In fact, the videos that are clicked second or third get the aspect ratio of the video that was clicked first. It shouldn't be like this!
+
+### Steps to reproduce
+
+build it
+
+1. Place more than one video widget on the page
+2. Enable the lightbox option for each video widget
+3. Give each video widget a different aspect ratio
+
+try it
+
+1. Click on the first video: The aspect ratio of the lightbox video is correct
+2. Click on the second video: The lightbox video's aspect ratio is incorrect, it will get the aspect ratio of the video that was clicked first
+
+### Isolating the problem
+
+- [ ] This bug happens with only Elementor plugin active (and Elementor Pro).
+- [ ] This bug happens with a Blank WordPress theme active ([Hello theme](https://wordpress.org/themes/hello-elementor/)).
+- [X] I can reproduce this bug consistently following the steps above.
+
+### System Info
+
+
+System Info
+
+````txt
+
+
+
+== Server Environment ==
+ Operating System: Linux
+ Software: nginx/1.22.0
+ MySQL version: mariadb.org binary distribution v10.4.28
+ PHP Version: 7.4.33
+ PHP Memory Limit: 128M
+ PHP Max Input Vars: 10000
+ PHP Max Post Size: 2000M
+ GD Installed: Yes
+ ZIP Installed: Yes
+ Write Permissions: All right
+ Elementor Library: Connected
+
+== WordPress Environment ==
+ Version: 6.1.1
+ Site URL: https://mirkoappel.de
+ Home URL: https://mirkoappel.de
+ WP Multisite: No
+ Max Upload Size: 2 GB
+ Memory limit: 128M
+ Max Memory limit: 128M
+ Permalink Structure: /news/%postname%/
+ Language: de-DE
+ Timezone: Europe/Berlin
+ Admin Email: [REDACTED]
+ Debug Mode: Inactive
+
+== Theme ==
+ Name: feather
+ Version: 1.0
+ Author: Mirko Appel
+ Child Theme: No
+
+== User ==
+ Role: administrator
+ WP Profile lang: de_DE
+ User Agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/111.0.0.0 Safari/537.36
+
+== Active Plugins ==
+ Admin Columns
+ Version: 4.5.5
+ Author: AdminColumns.com
+
+ Admin Menu Editor
+ Version: 1.10.4
+ Author: Janis Elsts
+
+ Advanced Custom Fields PRO
+ Version: 5.12.2
+ Author: Delicious Brains
+
+ Code Snippets
+ Version: 3.2.2
+ Author: Code Snippets Pro
+
+ Disable Gutenberg
+ Version: 2.9
+ Author: Jeff Starr
+
+ DynamicConditions
+ Version: 1.6.0
+ Author: RTO GmbH
+
+ Ele Custom Skin
+ Version: 3.1.7
+ Author: Dudaster.com
+
+ Elementor
+ Version: 3.11.4
+ Author: Elementor.com
+
+ Elementor Pro
+ Version: 3.11.5
+ Author: Elementor.com
+
+ Extras for Elementor
+ Version: 2.2.51
+ Author: Namogo
+
+ Intuitive Custom Post Order
+ Version: 3.1.4.1
+ Author: hijiri
+
+ Page scroll to id
+ Version: 1.7.7
+ Author: malihu
+
+ Post Types Unlimited
+ Version: 1.1
+ Author: WPExplorer
+
+ Regenerate Thumbnails
+ Version: 3.1.5
+ Author: Alex Mills (Viper007Bond)
+
+ WebP Express
+ Version: 0.25.5
+ Author: Bjørn Rosell
+
+ Yoast SEO
+ Version: 20.2.1
+ Author: Team Yoast
+
+
+== Eigenschaften ==
+ Custom Fonts: 0
+ Custom Icons: 0
+
+== Integrations ==
+
+
+
+== Elementor-Experimente ==
+ Optimierte DOM Ausgabe: Standardmäßig aktiviert
+ Verbessertes Laden von Assets: Aktiv
+ Verbessertes Laden von CSS: Aktiv
+ Inline-Schriftarten-Symbole: Standardmäßig deaktiviert
+ Verbesserungen der Zugänglichkeit: Standardmäßig aktiviert
+ Zusätzliche benutzerdefinierte Breakpoints: Inaktiv
+ admin_menu_rearrangement: Standardmäßig deaktiviert
+ Flexbox Container: Aktiv
+ Upgrade Swiper Library: Standardmäßig deaktiviert
+ Default to New Theme Builder: Standardmäßig aktiviert
+ Editor V2: Standardmäßig deaktiviert
+ Startseiten: Inaktiv
+ Nested Elements: Standardmäßig deaktiviert
+ Lazy Load Background Images: Standardmäßig deaktiviert
+ Als Standard speichern: Standardmäßig aktiviert
+ Page Transitions: Inaktiv
+ Notes: Standardmäßig aktiviert
+ Loop: Standardmäßig aktiviert
+ Form Submissions: Standardmäßig aktiviert
+ Scroll Snap: Inaktiv
+ Mega Menu: Standardmäßig deaktiviert
+
+
+== Protokoll ==
+
+Log: showing 16 of 162023-02-28 09:20:50 [info] Elementor data updater process has been completed. [array (
+ 'plugin' => 'Elementor Pro',
+ 'from' => '3.7.3',
+ 'to' => '3.11.3',
+)]
+2023-03-08 18:47:17 [info] elementor::elementor_updater Started
+2023-03-08 18:47:17 [info] Elementor/Upgrades - _on_each_version Start
+2023-03-08 18:47:17 [info] Elementor/Upgrades - _on_each_version Finished
+2023-03-08 18:47:17 [info] Elementor data updater process has been completed. [array (
+ 'plugin' => 'Elementor',
+ 'from' => '3.11.2',
+ 'to' => '3.11.3',
+)]
+2023-03-08 18:47:17 [info] Elementor data updater process has been queued. [array (
+ 'plugin' => 'Elementor',
+ 'from' => '3.11.2',
+ 'to' => '3.11.3',
+)]
+2023-03-12 17:32:58 [info] Elementor data updater process has been queued. [array (
+ 'plugin' => 'Elementor',
+ 'from' => '3.11.3',
+ 'to' => '3.11.4',
+)]
+2023-03-12 17:32:59 [info] elementor::elementor_updater Started
+2023-03-12 17:32:59 [info] Elementor/Upgrades - _on_each_version Finished
+2023-03-12 17:32:59 [info] Elementor/Upgrades - _on_each_version Start
+2023-03-12 17:32:59 [info] Elementor data updater process has been completed. [array (
+ 'plugin' => 'Elementor',
+ 'from' => '3.11.3',
+ 'to' => '3.11.4',
+)]
+2023-03-14 00:01:45 [info] elementor-pro::elementor_pro_updater Started
+2023-03-14 00:01:45 [info] Elementor Pro/Upgrades - _on_each_version Start
+2023-03-14 00:01:45 [info] Elementor Pro/Upgrades - _on_each_version Finished
+2023-03-14 00:01:45 [info] Elementor data updater process has been completed. [array (
+ 'plugin' => 'Elementor Pro',
+ 'from' => '3.11.3',
+ 'to' => '3.11.5',
+)]
+2023-03-14 00:01:45 [info] Elementor data updater process has been queued. [array (
+ 'plugin' => 'Elementor Pro',
+ 'from' => '3.11.3',
+ 'to' => '3.11.5',
+)]
+
+PHP: showing 10 of 10PHP: 2023-02-28 09:27:09 [notice X 57][/home/wp/disk/wordpress/wp-content/plugins/elementor/core/common/modules/ajax/module.php::175] Undefined index: data [array (
+ 'trace' => '
+#0: Elementor\Core\Logger\Manager -> shutdown()
+',
+)]
+PHP: 2023-02-28 09:29:53 [notice X 7][/home/wp/disk/wordpress/wp-content/plugins/elementor-pro/modules/theme-builder/classes/conditions-manager.php::168] Undefined index: editor_post_id [array (
+ 'trace' => '
+#0: Elementor\Core\Logger\Manager -> shutdown()
+',
+)]
+PHP: 2023-02-28 09:34:56 [notice X 47][/home/wp/disk/wordpress/wp-content/plugins/elementor/app/modules/import-export/module.php::629] Trying to get property 'label' of non-object [array (
+ 'trace' => '
+#0: Elementor\Core\Logger\Manager -> shutdown()
+',
+)]
+PHP: 2023-02-28 09:34:59 [notice X 69][/home/wp/disk/wordpress/wp-content/plugins/elementor-pro/core/app/modules/site-editor/data/endpoints/templates.php::150] Undefined index: condition_type [array (
+ 'trace' => '
+#0: /home/wp/disk/wordpress/wp-content/plugins/elementor-pro/core/app/modules/site-editor/data/endpoints/templates.php(150): Elementor\Core\Logger\Manager -> rest_error_handler()
+#1: ElementorPro\Core\App\Modules\SiteEditor\Data\Endpoints\Templates -> normalize_template_json_item()
+#2: /home/wp/disk/wordpress/wp-content/plugins/elementor-pro/core/app/modules/site-editor/data/endpoints/templates.php(120): class type array_map()
+#3: /home/wp/disk/wordpress/wp-content/plugins/elementor-pro/core/app/modules/site-editor/data/endpoints/templates.php(59): ElementorPro\Core\App\Modules\SiteEditor\Data\Endpoints\Templates -> normalize_templates_json()
+#4: /home/wp/disk/wordpress/wp-content/plugins/elementor/data/base/endpoint.php(158): ElementorPro\Core\App\Modules\SiteEditor\Data\Endpoints\Templates -> get_items()
+',
+)]
+PHP: 2023-02-28 23:24:47 [notice X 17][/home/wp/disk/wordpress/wp-content/plugins/elementor-pro/modules/posts/skins/skin-content-base.php::264] Trying to get property 'ID' of non-object [array (
+ 'trace' => '
+#0: Elementor\Core\Logger\Manager -> shutdown()
+',
+)]
+PHP: 2023-03-02 15:36:28 [notice X 4][/home/wp/disk/wordpress/wp-content/plugins/elementor/includes/frontend.php::1436] Trying to get property 'post_excerpt' of non-object [array (
+ 'trace' => '
+#0: Elementor\Core\Logger\Manager -> shutdown()
+',
+)]
+PHP: 2023-03-04 00:49:57 [error X 1][/home/wp/disk/wordpress/wp-content/plugins/elementor-pro/modules/theme-builder/classes/conditions-manager.php::295] Uncaught Error: Call to a member function update_meta() on bool in /home/wp/disk/wordpress/wp-content/plugins/elementor-pro/modules/theme-builder/classes/conditions-manager.php:295
+Stack trace:
+#0 /home/wp/disk/wordpress/wp-content/plugins/elementor-pro/modules/theme-builder/classes/conditions-manager.php(168): ElementorPro\Modules\ThemeBuilder\Classes\Conditions_Manager->save_conditions()
+#1 [internal function]: ElementorPro\Modules\ThemeBuilder\Classes\Conditions_Manager->ajax_save_theme_template_conditions()
+#2 /home/wp/disk/wordpress/wp-content/plugins/elementor/core/common/modules/ajax/module.php(175): call_user_func()
+#3 /home/wp/disk/wordpress/wp-includes/class-wp-hook.php(308): Elementor\Core\Common\Modules\Ajax\Module->handle_ajax_request()
+#4 /home/wp/disk/wordpress/wp-includes/class-wp-hook.php(332): WP_Hook->apply_filters()
+#5 /home/wp/disk/wordpress/wp-includes/plugin.php(517): WP_Hook->do_action()
+#6 /home/wp/disk/wordpress/wp-admin/admin-ajax.php(188): do_action()
+#7 {main}
+ thrown [array (
+ 'trace' => '
+#0: Elementor\Core\Logger\Manager -> shutdown()
+',
+)]
+PHP: 2023-03-07 11:32:45 [notice X 94][/home/wp/disk/wordpress/wp-content/plugins/elementor/includes/controls/groups/image-size.php::110] Trying to access array offset on value of type null [array (
+ 'trace' => '
+#0: Elementor\Core\Logger\Manager -> shutdown()
+',
+)]
+PHP: 2023-03-12 17:33:03 [warning X 1][/home/wp/disk/wordpress/wp-content/plugins/elementor/core/files/manager.php::111] unlink(/home/wp/disk/wordpress/wp-content/uploads/elementor/css/post-17558.css): No such file or directory [array (
+ 'trace' => '
+#0: Elementor\Core\Logger\Manager -> shutdown()
+',
+)]
+PHP: 2023-03-12 17:42:53 [notice X 23][/home/wp/disk/wordpress/wp-content/plugins/elementor-pro/modules/dynamic-tags/acf/tags/acf-text.php::33] Undefined offset: 1 [array (
+ 'trace' => '
+#0: Elementor\Core\Logger\Manager -> shutdown()
+',
+)]
+
+JS: showing 16 of 16JS: 2023-02-28 08:27:15 [error X 7][https://mirkoappel.de/wp-content/plugins/elementor/assets/js/editor.min.js?ver=3.11.2:3:1087291] Cannot read properties of null (reading 'getBoundingClientRect')
+JS: 2023-02-28 09:10:26 [error X 598][https://mirkoappel.de/wp-includes/js/jquery/jquery.min.js?ver=3.6.1:2:51824] Cannot read properties of undefined (reading 'replace')
+JS: 2023-03-02 10:37:36 [error X 3][https://mirkoappel.de/wp-content/plugins/elementor/assets/lib/backbone/backbone.marionette.min.js?ver=2.4.5.e1:24:19952] View (cid: ""view356"") has already been destroyed and cannot be used.
+JS: 2023-03-02 10:44:18 [error X 7][https://mirkoappel.de/wp-admin/post.php?post=16474&action=elementor#:1:1] Identifier '$dcWidgetIdInput' has already been declared
+JS: 2023-03-02 10:48:31 [error X 1][https://mirkoappel.de/wp-content/plugins/elementor/assets/lib/backbone/backbone.marionette.min.js?ver=2.4.5.e1:24:19952] View (cid: ""view27440"") has already been destroyed and cannot be used.
+JS: 2023-03-02 11:40:47 [error X 115][https://mirkoappel.de/wp-content/plugins/elementor-pro/assets/lib/sticky/jquery.sticky.min.js?ver=3.11.3:1:1162] Failed to execute 'getComputedStyle' on 'Window': parameter 1 is not of type 'Element'.
+JS: 2023-03-02 14:05:47 [error X 2][https://mirkoappel.de/wp-content/plugins/elementor/assets/js/editor.min.js?ver=3.11.2:3:931389] Cannot read properties of undefined (reading 'settings')
+JS: 2023-03-03 12:08:55 [error X 2][https://mirkoappel.de/wp-content/plugins/elementor/assets/js/editor.min.js?ver=3.11.2:3:907724] Cannot read properties of undefined (reading 'route')
+JS: 2023-03-04 08:30:07 [error X 4][https://mirkoappel.de/wp-content/plugins/elementor/assets/js/editor.min.js?ver=3.11.2:3:656177] elementorFrontend.elements.window.jQuery is not a function
+JS: 2023-03-04 08:30:07 [error X 4][https://mirkoappel.de/wp-content/plugins/elementor-extras/assets/js/frontend.min.js?ver=2.2.51:1:4462] Cannot read properties of undefined (reading '$previewContents')
+JS: 2023-03-04 23:40:33 [error X 1][https://mirkoappel.de/wp-content/plugins/elementor/assets/js/editor.min.js?ver=3.11.2:3:676196] Cannot read properties of undefined (reading 'get')
+JS: 2023-03-06 10:38:43 [error X 11][https://mirkoappel.de/wp-content/plugins/elementor/assets/lib/backbone/backbone.marionette.min.js?ver=2.4.5.e1:24:19952] View (cid: ""view284"") has already been destroyed and cannot be used.
+JS: 2023-03-06 10:41:09 [error X 814][https://mirkoappel.de/wp-content/plugins/elementor-pro/assets/js/frontend.min.js?ver=3.11.3:2:10935] Failed to execute 'observe' on 'IntersectionObserver': parameter 1 is not of type 'Element'.
+JS: 2023-03-07 16:31:27 [error X 4][https://mirkoappel.de/wp-content/plugins/elementor/assets/js/editor.min.js?ver=3.11.2:3:791668] Cannot read properties of undefined (reading 'localized')
+JS: 2023-03-13 18:01:30 [error X 1][https://mirkoappel.de/wp-content/plugins/elementor/assets/js/editor-loader-v1.min.js?ver=3.11.4:2:37] Cannot read properties of undefined (reading 'start')
+JS: 2023-03-13 18:01:30 [error X 1][https://mirkoappel.de/wp-content/plugins/elementor/assets/js/container-converter.min.js?ver=3.11.4:2:5351] elementor is not defined
+
+
+
+== Elementor - Compatibility Tag ==
+
+ Elementor Pro: Die Kompatibilität ist nicht angegeben
+ Extras for Elementor: Die Kompatibilität ist nicht angegeben
+
+== Elementor Pro - Compatibility Tag ==
+
+
+
+
+````
+
+
+",0,✅ 🐞 bug report video lightbox gets the wrong aspect ratio when more then one video widget is used on the same page prerequisites i have searched for similar issues in both open and closed tickets and cannot find a duplicate the issue still exists against the latest stable version of elementor description html the problem is caused by the elementor aspect ratio class each dialog widget should have only one elementor aspect ratio class to set the aspect ratio but in some cases there are two or more of them so they overwrite each other this leads to an incorrect aspect ratio of the lightbox video for some videos in fact the videos that are clicked second or third get the aspect ratio of the video that was clicked first it shouldn t be like this steps to reproduce build it place more than one video widget on the page enable the lightbox option for each video widget give each video widget a different aspect ratio try it click on the first video the aspect ratio of the lightbox video is correct click on the second video the lightbox video s aspect ratio is incorrect it will get the aspect ratio of the video that was clicked first isolating the problem this bug happens with only elementor plugin active and elementor pro this bug happens with a blank wordpress theme active i can reproduce this bug consistently following the steps above system info system info txt server environment operating system linux software nginx mysql version mariadb org binary distribution php version php memory limit php max input vars php max post size gd installed yes zip installed yes write permissions all right elementor library connected wordpress environment version site url home url wp multisite no max upload size gb memory limit max memory limit permalink structure news postname language de de timezone europe berlin admin email debug mode inactive theme name feather version author mirko appel child theme no user role administrator wp profile lang de de user agent mozilla windows nt applewebkit khtml like gecko chrome safari active plugins admin columns version author admincolumns com admin menu editor version author janis elsts advanced custom fields pro version author delicious brains code snippets version author code snippets pro disable gutenberg version author jeff starr dynamicconditions version author rto gmbh ele custom skin version author dudaster com elementor version author elementor com elementor pro version author elementor com extras for elementor version author namogo intuitive custom post order version author hijiri page scroll to id version author malihu post types unlimited version author wpexplorer regenerate thumbnails version author alex mills webp express version author bjørn rosell yoast seo version author team yoast eigenschaften custom fonts custom icons integrations elementor experimente optimierte dom ausgabe standardmäßig aktiviert verbessertes laden von assets aktiv verbessertes laden von css aktiv inline schriftarten symbole standardmäßig deaktiviert verbesserungen der zugänglichkeit standardmäßig aktiviert zusätzliche benutzerdefinierte breakpoints inaktiv admin menu rearrangement standardmäßig deaktiviert flexbox container aktiv upgrade swiper library standardmäßig deaktiviert default to new theme builder standardmäßig aktiviert editor standardmäßig deaktiviert startseiten inaktiv nested elements standardmäßig deaktiviert lazy load background images standardmäßig deaktiviert als standard speichern standardmäßig aktiviert page transitions inaktiv notes standardmäßig aktiviert loop standardmäßig aktiviert form submissions standardmäßig aktiviert scroll snap inaktiv mega menu standardmäßig deaktiviert protokoll log showing of elementor data updater process has been completed array plugin elementor pro from to elementor elementor updater started elementor upgrades on each version start elementor upgrades on each version finished elementor data updater process has been completed array plugin elementor from to elementor data updater process has been queued array plugin elementor from to elementor data updater process has been queued array plugin elementor from to elementor elementor updater started elementor upgrades on each version finished elementor upgrades on each version start elementor data updater process has been completed array plugin elementor from to elementor pro elementor pro updater started elementor pro upgrades on each version start elementor pro upgrades on each version finished elementor data updater process has been completed array plugin elementor pro from to elementor data updater process has been queued array plugin elementor pro from to php showing of undefined index data array trace elementor core logger manager shutdown php undefined index editor post id array trace elementor core logger manager shutdown php trying to get property label of non object array trace elementor core logger manager shutdown php undefined index condition type array trace home wp disk wordpress wp content plugins elementor pro core app modules site editor data endpoints templates php elementor core logger manager rest error handler elementorpro core app modules siteeditor data endpoints templates normalize template json item home wp disk wordpress wp content plugins elementor pro core app modules site editor data endpoints templates php class type array map home wp disk wordpress wp content plugins elementor pro core app modules site editor data endpoints templates php elementorpro core app modules siteeditor data endpoints templates normalize templates json home wp disk wordpress wp content plugins elementor data base endpoint php elementorpro core app modules siteeditor data endpoints templates get items php trying to get property id of non object array trace elementor core logger manager shutdown php trying to get property post excerpt of non object array trace elementor core logger manager shutdown php uncaught error call to a member function update meta on bool in home wp disk wordpress wp content plugins elementor pro modules theme builder classes conditions manager php stack trace home wp disk wordpress wp content plugins elementor pro modules theme builder classes conditions manager php elementorpro modules themebuilder classes conditions manager save conditions elementorpro modules themebuilder classes conditions manager ajax save theme template conditions home wp disk wordpress wp content plugins elementor core common modules ajax module php call user func home wp disk wordpress wp includes class wp hook php elementor core common modules ajax module handle ajax request home wp disk wordpress wp includes class wp hook php wp hook apply filters home wp disk wordpress wp includes plugin php wp hook do action home wp disk wordpress wp admin admin ajax php do action main thrown array trace elementor core logger manager shutdown php trying to access array offset on value of type null array trace elementor core logger manager shutdown php unlink home wp disk wordpress wp content uploads elementor css post css no such file or directory array trace elementor core logger manager shutdown php undefined offset array trace elementor core logger manager shutdown js showing of cannot read properties of null reading getboundingclientrect js cannot read properties of undefined reading replace js view cid has already been destroyed and cannot be used js identifier dcwidgetidinput has already been declared js view cid has already been destroyed and cannot be used js failed to execute getcomputedstyle on window parameter is not of type element js cannot read properties of undefined reading settings js cannot read properties of undefined reading route js elementorfrontend elements window jquery is not a function js cannot read properties of undefined reading previewcontents js cannot read properties of undefined reading get js view cid has already been destroyed and cannot be used js failed to execute observe on intersectionobserver parameter is not of type element js cannot read properties of undefined reading localized js cannot read properties of undefined reading start js elementor is not defined elementor compatibility tag elementor pro die kompatibilität ist nicht angegeben extras for elementor die kompatibilität ist nicht angegeben elementor pro compatibility tag ,0
+83,3361957784.0,IssuesEvent,2015-11-20 01:23:12,bbergen/tbmd.com,https://api.github.com/repos/bbergen/tbmd.com,opened,Add Actor to Movie,Movies People,"On the Person page, there should be the ability to add the currently viewed actor to a movie that exists in the database.
+
+On the Movie page, there should be the ability to add an actor that exists in the database to the currently viewed movie.
+
+Only eligible additions should be shown in any lists of additions (Add Actor list should only show a list of actors to choose from that are not already in the movie)
+
+These edits should only be allowed if the user is logged in.",1.0,"Add Actor to Movie - On the Person page, there should be the ability to add the currently viewed actor to a movie that exists in the database.
+
+On the Movie page, there should be the ability to add an actor that exists in the database to the currently viewed movie.
+
+Only eligible additions should be shown in any lists of additions (Add Actor list should only show a list of actors to choose from that are not already in the movie)
+
+These edits should only be allowed if the user is logged in.",1,add actor to movie on the person page there should be the ability to add the currently viewed actor to a movie that exists in the database on the movie page there should be the ability to add an actor that exists in the database to the currently viewed movie only eligible additions should be shown in any lists of additions add actor list should only show a list of actors to choose from that are not already in the movie these edits should only be allowed if the user is logged in ,1
+304,5683784944.0,IssuesEvent,2017-04-13 13:36:08,ccswbs/hjckrrh,https://api.github.com/repos/ccswbs/hjckrrh,closed,PP1 - Add a tag argument to the /people page associated with the PP1 view,feature: people profiles (PP),"**Request:** Add a tag argument to the /people page associated with the PP1 view. Instead of ```/people``` it would then be ```/people/!tag```.
+
+The reason this would be needed would be to:
+- allow us to override the title of any pages filtered by category or keyword.
+- swap out any pages using the PP5 view with the PP1 view.
+
+This came up as a result of COA requesting to have the Search for people by role filter on their department people profile pages. Since these pages were using the PP5 view, in order to override the title and have it show "" Faculty and Staff"", the page layout would need access to the tag through the URL.
+
+When fulfilling this request, please consider possibility of removing PP5 view and the related page /people/tag/% and instead relying solely on the PP1 view and its associated page. Note that some clients might be using the PP5 view, so you would need to ensure that they are not affected if view is removed. One option is to keep both URLs working, but instead having the /people/tag/% url be associated with the PP1 view instead (perhaps see if you can associate multiple URLs with one page)...?",1.0,"PP1 - Add a tag argument to the /people page associated with the PP1 view - **Request:** Add a tag argument to the /people page associated with the PP1 view. Instead of ```/people``` it would then be ```/people/!tag```.
+
+The reason this would be needed would be to:
+- allow us to override the title of any pages filtered by category or keyword.
+- swap out any pages using the PP5 view with the PP1 view.
+
+This came up as a result of COA requesting to have the Search for people by role filter on their department people profile pages. Since these pages were using the PP5 view, in order to override the title and have it show "" Faculty and Staff"", the page layout would need access to the tag through the URL.
+
+When fulfilling this request, please consider possibility of removing PP5 view and the related page /people/tag/% and instead relying solely on the PP1 view and its associated page. Note that some clients might be using the PP5 view, so you would need to ensure that they are not affected if view is removed. One option is to keep both URLs working, but instead having the /people/tag/% url be associated with the PP1 view instead (perhaps see if you can associate multiple URLs with one page)...?",1, add a tag argument to the people page associated with the view request add a tag argument to the people page associated with the view instead of people it would then be people tag the reason this would be needed would be to allow us to override the title of any pages filtered by category or keyword swap out any pages using the view with the view this came up as a result of coa requesting to have the search for people by role filter on their department people profile pages since these pages were using the view in order to override the title and have it show faculty and staff the page layout would need access to the tag through the url when fulfilling this request please consider possibility of removing view and the related page people tag and instead relying solely on the view and its associated page note that some clients might be using the view so you would need to ensure that they are not affected if view is removed one option is to keep both urls working but instead having the people tag url be associated with the view instead perhaps see if you can associate multiple urls with one page ,1
+702575,24126723556.0,IssuesEvent,2022-09-21 01:38:39,CanberraOceanRacingClub/namadgi3,https://api.github.com/repos/CanberraOceanRacingClub/namadgi3,opened,Replace transom bumper rubber,Priority_2 EMA,"On the B46 the bumper rubber extends from the sheerline around the transom at the topsides down to the water line. It has been subject to impact damage and mould deterioration over time. It needs to be replaced at the next opportunity. The part will need to be ordered from Bavaria.
+
+* remove the old rubber
+* clean f/glass surfaces
+* rebed new rubber",1.0,"Replace transom bumper rubber - On the B46 the bumper rubber extends from the sheerline around the transom at the topsides down to the water line. It has been subject to impact damage and mould deterioration over time. It needs to be replaced at the next opportunity. The part will need to be ordered from Bavaria.
+
+* remove the old rubber
+* clean f/glass surfaces
+* rebed new rubber",0,replace transom bumper rubber on the the bumper rubber extends from the sheerline around the transom at the topsides down to the water line it has been subject to impact damage and mould deterioration over time it needs to be replaced at the next opportunity the part will need to be ordered from bavaria remove the old rubber clean f glass surfaces rebed new rubber,0
+1186,30799547955.0,IssuesEvent,2023-07-31 23:24:12,henrikfroehling/Trakt.NET,https://api.github.com/repos/henrikfroehling/Trakt.NET,closed,"Add method overloads in modules to be more flexible with ""id or slug"" parameters",Library Module People Module Recommendations Module Seasons Module Shows Module Movies Module Episodes Module Users Module Lists Improvement v1.4.0 Target-Branch: release-1.4.0,"New API:
+```diff
+namespace TraktNet.Modules
+{
+ public class TraktEpisodesModule
+ {
++ public Task> GetEpisodeAsync(uint traktShowID, uint seasonNumber, uint episodeNumber);
++ public Task> GetEpisodeAsync(ITraktShowIds showIDs, uint seasonNumber, uint episodeNumber);
++ public Task> GetEpisodeAsync(ITraktShow show, uint seasonNumber, uint episodeNumber);
+
++ public Task> GetEpisodeCommentsAsync(uint traktShowID, uint seasonNumber, uint episodeNumber);
++ public Task> GetEpisodeCommentsAsync(ITraktShowIds showIDs, uint seasonNumber, uint episodeNumber);
++ public Task> GetEpisodeCommentsAsync(ITraktShow show, uint seasonNumber, uint episodeNumber);
+
++ public Task> GetEpisodeListsAsync(uint traktShowID, uint seasonNumber, uint episodeNumber);
++ public Task> GetEpisodeListsAsync(ITraktShowIds showIDs, uint seasonNumber, uint episodeNumber);
++ public Task> GetEpisodeListsAsync(ITraktShow show, uint seasonNumber, uint episodeNumber);
+
++ public Task> GetEpisodePeopleAsync(uint traktShowID, uint seasonNumber, uint episodeNumber);
++ public Task> GetEpisodePeopleAsync(ITraktShowIds showIDs, uint seasonNumber, uint episodeNumber);
++ public Task> GetEpisodePeopleAsync(ITraktShow show, uint seasonNumber, uint episodeNumber);
+
++ public Task> GetEpisodeRatingsAsync(uint traktShowID, uint seasonNumber, uint episodeNumber);
++ public Task> GetEpisodeRatingsAsync(ITraktShowIds showIDs, uint seasonNumber, uint episodeNumber);
++ public Task> GetEpisodeRatingsAsync(ITraktShow show, uint seasonNumber, uint episodeNumber);
+
++ public Task> GetEpisodeStatisticsAsync(uint traktShowID, uint seasonNumber, uint episodeNumber);
++ public Task> GetEpisodeStatisticsAsync(ITraktShowIds showIDs, uint seasonNumber, uint episodeNumber);
++ public Task> GetEpisodeStatisticsAsync(ITraktShow show, uint seasonNumber, uint episodeNumber);
+
++ public Task> GetEpisodeTranslationsAsync(uint traktShowID, uint seasonNumber, uint episodeNumber);
++ public Task> GetEpisodeTranslationsAsync(ITraktShowIds showIDs, uint seasonNumber, uint episodeNumber);
++ public Task> GetEpisodeTranslationsAsync(ITraktShow show, uint seasonNumber, uint episodeNumber);
+
++ public Task> GetEpisodeWatchingUsersAsync(uint traktShowID, uint seasonNumber, uint episodeNumber);
++ public Task> GetEpisodeWatchingUsersAsync(ITraktShowIds showIDs, uint seasonNumber, uint episodeNumber);
++ public Task> GetEpisodeWatchingUsersAsync(ITraktShow show, uint seasonNumber, uint episodeNumber);
+ }
+
+ public class TraktListsModule
+ {
++ public Task> GetListAsync(uint traktListID);
++ public Task> GetListAsync(ITraktListIds listIDs);
+
++ public Task> GetListItemsAsync(uint traktListID);
++ public Task> GetListItemsAsync(ITraktListIds listIDs);
++ public Task> GetListItemsAsync(ITraktList list);
+
++ public Task> GetListLikesAsync(uint traktListID);
++ public Task> GetListLikesAsync(ITraktListIds listIDs);
++ public Task> GetListLikesAsync(ITraktList list);
+
++ public Task> GetListCommentsAsync(uint traktListID);
++ public Task> GetListCommentsAsync(ITraktListIds listIDs);
++ public Task> GetListCommentsAsync(ITraktList list);
+
++ public Task LikeListAsync(uint traktListID);
++ public Task LikeListAsync(ITraktListIds listIDs);
++ public Task LikeListAsync(ITraktList list);
+
++ public Task UnlikeListAsync(uint traktListID);
++ public Task UnlikeListAsync(ITraktListIds listIDs);
++ public Task UnlikeListAsync(ITraktList list);
+ }
+
+ public class TraktMoviesModule
+ {
++ public Task> GetMovieAsync(uint traktMovieID);
++ public Task> GetMovieAsync(ITraktMovieIds movieIDs);
+
++ public Task> GetMovieAliasesAsync(uint traktMovieID);
++ public Task> GetMovieAliasesAsync(ITraktMovieIds movieIDs);
+
++ public Task> GetMovieReleasesAsync(uint traktMovieID);
++ public Task> GetMovieReleasesAsync(ITraktMovieIds movieIDs);
+
++ public Task> GetMovieTranslationsAsync(uint traktMovieID);
++ public Task> GetMovieTranslationsAsync(ITraktMovieIds movieIDs);
+
++ public Task> GetMovieCommentsAsync(uint traktMovieID);
++ public Task> GetMovieCommentsAsync(ITraktMovieIds movieIDs);
+
++ public Task> GetMovieListsAsync(uint traktMovieID);
++ public Task> GetMovieListsAsync(ITraktMovieIds movieIDs);
+
++ public Task> GetMoviePeopleAsync(uint traktMovieID);
++ public Task> GetMoviePeopleAsync(ITraktMovieIds movieIDs);
+
++ public Task> GetMovieRatingsAsync(uint traktMovieID);
++ public Task> GetMovieRatingsAsync(ITraktMovieIds movieIDs);
+
++ public Task> GetMovieRelatedMoviesAsync(uint traktMovieID);
++ public Task> GetMovieRelatedMoviesAsync(ITraktMovieIds movieIDs);
+
++ public Task> GetMovieStatisticsAsync(uint traktMovieID);
++ public Task> GetMovieStatisticsAsync(ITraktMovieIds movieIDs);
+
++ public Task> GetMovieWatchingUsersAsync(uint traktMovieID);
++ public Task> GetMovieWatchingUsersAsync(ITraktMovieIds movieIDs);
+
++ public Task> GetMovieStudiosAsync(uint traktMovieID);
++ public Task> GetMovieStudiosAsync(ITraktMovieIds movieIDs);
+ }
+
+ public class TraktPeopleModule
+ {
++ public Task> GetPersonAsync(uint traktPersonID);
++ public Task> GetPersonAsync(ITraktPersonIds personIDs);
+
++ public Task> GetPersonMovieCreditsAsync(uint traktPersonID);
++ public Task> GetPersonMovieCreditsAsync(ITraktPersonIds personIDs);
+
++ public Task> GetPersonShowCreditsAsync(uint traktPersonID);
++ public Task> GetPersonShowCreditsAsync(ITraktPersonIds personIDs);
+
++ public Task> GetPersonListsAsync(uint traktPersonID);
++ public Task> GetPersonListsAsync(ITraktPersonIds personIDs);
+ }
+
+ public class TraktRecommendationsModule
+ {
++ public Task HideMovieRecommendationAsync(uint traktMovieID);
++ public Task HideMovieRecommendationAsync(ITraktMovieIds movieIDs);
+
++ public Task HideShowRecommendationAsync(uint traktShowID);
++ public Task HideShowRecommendationAsync(ITraktShowIds showIDs);
+ }
+
+ public class TraktSeasonsModule
+ {
++ public Task> GetAllSeasonsAsync(uint traktShowID);
++ public Task> GetAllSeasonsAsync(ITraktShowIds showIDs);
+
++ public Task> GetSeasonAsync(uint traktShowID, uint seasonNumber);
++ public Task> GetSeasonAsync(ITraktShowIds showIDs, uint seasonNumber);
+
++ public Task> GetSeasonCommentsAsync(uint traktShowID, uint seasonNumber);
++ public Task> GetSeasonCommentsAsync(ITraktShowIds showIDs, uint seasonNumber);
+
++ public Task> GetSeasonListsAsync(uint traktShowID, uint seasonNumber);
++ public Task> GetSeasonListsAsync(ITraktShowIds showIDs, uint seasonNumber);
+
++ public Task> GetSeasonPeopleAsync(uint traktShowID, uint seasonNumber);
++ public Task> GetSeasonPeopleAsync(ITraktShowIds showIDs, uint seasonNumber);
+
++ public Task> GetSeasonRatingsAsync(uint traktShowID, uint seasonNumber);
++ public Task> GetSeasonRatingsAsync(ITraktShowIds showIDs, uint seasonNumber);
+
++ public Task> GetSeasonStatisticsAsync(uint traktShowID, uint seasonNumber);
++ public Task> GetSeasonStatisticsAsync(ITraktShowIds showIDs, uint seasonNumber);
+
++ public Task> GetSeasonTranslationsAsync(uint traktShowID, uint seasonNumber);
++ public Task> GetSeasonTranslationsAsync(ITraktShowIds showIDs, uint seasonNumber);
+
++ public Task> GetSeasonWatchingUsersAsync(uint traktShowID, uint seasonNumber);
++ public Task> GetSeasonWatchingUsersAsync(ITraktShowIds showIDs, uint seasonNumber);
+ }
+
+ public class TraktShowsModule
+ {
++ public Task> GetShowAsync(uint traktShowID);
++ public Task> GetShowAsync(ITraktShowIds showIDs);
+
++ public Task> GetShowAliasesAsync(uint traktShowID);
++ public Task> GetShowAliasesAsync(ITraktShowIds showIDs);
+
++ public Task> GetShowTranslationsAsync(uint traktShowID);
++ public Task> GetShowTranslationsAsync(ITraktShowIds showIDs);
+
++ public Task> GetShowCommentsAsync(uint traktShowID);
++ public Task> GetShowCommentsAsync(ITraktShowIds showIDs);
+
++ public Task> GetShowListsAsync(uint traktShowID);
++ public Task> GetShowListsAsync(ITraktShowIds showIDs);
+
++ public Task> GetShowPeopleAsync(uint traktShowID);
++ public Task> GetShowPeopleAsync(ITraktShowIds showIDs);
+
++ public Task> GetShowRatingsAsync(uint traktShowID);
++ public Task> GetShowRatingsAsync(ITraktShowIds showIDs);
+
++ public Task> GetShowRelatedShowsAsync(uint traktShowID);
++ public Task> GetShowRelatedShowsAsync(ITraktShowIds showIDs);
+
++ public Task> GetShowStatisticsAsync(uint traktShowID);
++ public Task> GetShowStatisticsAsync(ITraktShowIds showIDs);
+
++ public Task> GetShowWatchingUsersAsync(uint traktShowID);
++ public Task> GetShowWatchingUsersAsync(ITraktShowIds showIDs);
+
++ public Task> GetShowCollectionProgressAsync(uint traktShowID);
++ public Task> GetShowCollectionProgressAsync(ITraktShowIds showIDs);
+
++ public Task> GetShowWatchedProgressAsync(uint traktShowID);
++ public Task> GetShowWatchedProgressAsync(ITraktShowIds showIDs);
+
++ public Task> ResetShowWatchedProgressAsync(uint traktShowID);
++ public Task> ResetShowWatchedProgressAsync(ITraktShowIds showIDs);
+
++ public Task