source
stringclasses
1 value
text
stringlengths
152
659k
filtering_features
stringlengths
402
437
source_other
stringlengths
440
819k
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: [PrintAsObjc] Import/fwd-declare the unbridged type for known types. username_0: For non-Clang types that conform to `_ObjectiveCBridgeable` but have type mappings defined in `getKnownTypeInfo`, always use the Clang type in the mapping to determine which module to import. `CGFloat` is a rare case of an `_ObjectiveCBridgeable` type that is a native Swift `struct` declared in an overlay instead of extending the original Clang decl with the conformance. Since `hasClangNode` returns false in this case, it fell through to the rest of the logic for `getObjCBridgedClass` and would forward declare `@class NSNumber` instead of `@import CoreGraphics`. Thus, in some cases (e.g., the header only defines APIs that use `CGFloat`), the header would be invalid. Fixes SR-14266. <issue_comment>username_1: cc @xedin, who last worked on `CGFloat`-related stuff, in case it's of interest <issue_comment>username_0: @swift-ci please test <issue_comment>username_0: @swift-ci Please test macOS platform <issue_comment>username_0: @swift-ci Please smoke test Linux platform <issue_comment>username_0: @swift-ci Please smoke test and merge <issue_comment>username_0: @swift-ci Please smoke test macOS platform <issue_comment>username_0: @swift-ci Please smoke test macOS platform
{'fraction_non_alphanumeric': 0.06809487375669472, 'fraction_numerical': 0.009946442234123947, 'mean_word_length': 5.8125, 'pattern_counts': {'":': 0, '<': 9, '<?xml version=': 0, '>': 9, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '3234684', 'n_tokens_mistral': 379, 'n_tokens_neox': 363, 'n_words': 167}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Can't print PDF username_0: I'am using Ionic. with native ionic cordova-plugin-printer plugin, I can't print pdf file. I gave pdf uri as parameter to print() function. I also checked issues about this topic. and @username_1 said pdf file print feature will be added. Is it added? I'm currently use npm cordova-plugin-printer 0.8.0 <issue_comment>username_1: Can you please provide some code snippet and are you using plugin wrapper provided by Ionic? <issue_comment>username_0: that is ionic projects package.json i am using ```json { "scripts": { "clean": "ionic-app-scripts clean", "build": "ionic-app-scripts build --prod", "lint": "ionic-app-scripts lint", "ionic:build": "ionic-app-scripts build --prod", "ionic:serve": "ionic-app-scripts serve" }, "dependencies": { "@angular/animations": "^5.2.11", "@angular/common": "^5.2.11", "@angular/compiler": "^5.2.11", "@angular/compiler-cli": "^5.2.11", "@angular/core": "^5.2.11", "@angular/forms": "^5.2.11", "@angular/http": "^5.2.11", "@angular/platform-browser": "^5.2.11", "@angular/platform-browser-dynamic": "^5.2.11", "@auth0/angular-jwt": "1.2.0", "@ionic-native/barcode-scanner": "^4.20.0", "@ionic-native/camera": "^4.20.0", "@ionic-native/core": "^4.20.0", "@ionic-native/file": "^4.20.0", "@ionic-native/file-transfer": "^4.20.0", "@ionic-native/keyboard": "^4.20.0", "@ionic-native/market": "^4.20.0", "@ionic-native/network": "4.9.0", "@ionic-native/printer": "^4.20.0", "@ionic-native/screen-orientationa": "4.9.0", "@ionic-native/splash-screen": "^4.20.0", "@ionic-native/sqlite": "^4.20.0", "@ionic-native/status-bar": "4.9.0", "@ionic/pro": "1.0.20", "@ionic/storage": "^2.2.0", "@mobiscroll/angular-lite": "^4.5.2", "@ngx-translate/core": "8.0.0", "@ngx-translate/http-loader": "^2.0.0", "@types/node": "^10.12.27", "ajv": "^6.9.2", "chart.js": "^2.7.3", "cordova-android": "7.1.4", "cordova-android-support-gradle-release": "^2.0.1", "cordova-browser": "5.0.4", "cordova-ios": "4.5.5", "cordova-plugin-camera": "^4.0.3", "cordova-plugin-device": "^2.0.2", "cordova-plugin-file": "^6.0.1", "cordova-plugin-ionic-webview": "^2.4.0", "cordova-plugin-market": "^1.2.0", "cordova-plugin-network-information": "^2.0.1", "cordova-plugin-printer": "^0.8.0", "cordova-plugin-screen-orientation": "^3.0.1", "cordova-plugin-splashscreen": "^5.0.2", "cordova-plugin-statusbar": "^2.4.2", "cordova-plugin-whitelist": "^1.3.3", "cordova-sqlite-storage": "^2.6.0", "cordova-windows": "^6.0.1", "es6-promise-plugin": "^4.2.2", "ionic-angular": "^3.9.4", "ionic-plugin-keyboard": "^2.2.1", "ionicons": "3.0.0", "ng-circle-progress": "1.0.0", "ng2-validation": "^4.2.0", "node-gyp": "^3.7.0", "phonegap-plugin-barcodescanner": "^8.0.1", "rxjs": "^5.5.12", "sw-toolbox": "3.6.0", "zone.js": "0.8.26" }, "devDependencies": { "@ionic/app-scripts": "^3.2.2", "typescript": "~2.6.2" }, "description": "...", "config": { "ionic_webpack": "./src/config/webpack.config.js", "ionic_copy": "./scripts/copy-mobiscroll-css.js" }, [Truncated] this.printerAvailable = true; }, () => { this.printerAvailable = false; }); }, () => { this.printerAvailable = false; }) } print() { try { this.printer.print('file://' + 'http://www.africau.edu/images/default/sample.pdf'); } catch(err) { this.toastService.publishTranslateMessage('Unsuccessful'); } } ``` And when I run this neither it does nothing nor it raises error even try on real devices. <issue_comment>username_2: I'm having the same problem did you find a solution? @username_0 ? <issue_comment>username_0: unfortunately @username_2. waiting for response from @username_1 <issue_comment>username_3: look like it stop at nativeApi function, and i think this can run with cordova android platform 8.0.0 up. i can not run with cordova android 7.1.2 via cli 8.1.1 phonegap build <issue_comment>username_4: Am still waiting for this issue to be fixed. I saw the code, and what I can understand is its trying to get the mime type in case of "base64:" which is not working properly. See what I think I should send while passing pdf data as base 64 to the plugin is "base64://application/pdf;base64,mydata". Which seems obscure. Basically what should be passed is "data:application/pdf;base64,mydata". Please solve this issue as I need to use this plugin to print pdf using base64 I get from a service. Also I can create a file as storing file is restricted due to reasons which I can't disclose. @username_1 Please consider this as pdf is very prominent way to print files. <issue_comment>username_4: For time being, I am using this plugin https://github.com/sarahgoldman/cordova-print-pdf-plugin and it seems to work flawlessly. <issue_comment>username_5: @username_4 Did you able to print pdf file? Did you get any other solution other than this https://github.com/sarahgoldman/cordova-print-pdf-plugin? @username_1 Please solve this issue. <issue_comment>username_4: Actually I am still using "cordova-print-pdf-plugin" as my requirements were to only print base64 pdf so I did not use this plugin. Yes @username_5 I was able to print pdf using this other plugin. <issue_comment>username_5: @username_4 If possible can you post some code that how you implemented "cordova-print-pdf-plugin". In my case, "cordova-print-pdf-plugin" plugin is not working. If possible please help me out to solve this. <issue_comment>username_4: @username_5 I don't think this is the correct place for this discussion. Still below is how I am doing it. I have cross platform application so 1. For Desktop: I am using "[print.min.js](https://printjs.crabbly.com/)" `printJS({ "printable": app.getSharedData().consentFormBase64, "type": "pdf", "base64": true })` 2. For Android: I am using this "[cordova-print-pdf-plugin](https://github.com/sarahgoldman/cordova-print-pdf-plugin)". ` var options = { "data": myBase64, // (omit "data:application/pdf;base64," part from base64 string as per their github page) "success": successCallback, "error": failureCallback }; plugins.PrintPDF.print(options); ` Its pretty simple right. If you are failing with this code then I think you are not passing base64 in the right format or your base64 is corrupted/incomplete. <issue_comment>username_5: @username_4 Got it and Thanks for the response. <issue_comment>username_6: Just stumbled on this issue, and want to add results of iOS testing: `window.cordova.plugins.printer.print('https://samples.com/pdf.js')` doesn't work. `window.cordova.plugins.printer.print('base64://Ges4GscWe...')` works great. to convert remote file to base64 you can use this code: ``` const base64 = await downloadAsBase64('https://samples.com/pdf.js') const content = 'base64://' + base64.split(',')[1] async function downloadAsBase64(url) { const response = await fetch(url) const blob = await response.blob() return new Promise((resolve, reject) => { const reader = new FileReader() reader.onload = function() { resolve(this.result) } reader.onerror = reject reader.readAsDataURL(blob) }) } ```
{'fraction_non_alphanumeric': 0.1539240506329114, 'fraction_numerical': 0.03721518987341772, 'mean_word_length': 2.7534441805225653, 'pattern_counts': {'":': 79, '<': 15, '<?xml version=': 0, '>': 18, 'https://': 6, 'lorem ipsum': 0, 'www.': 1, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 2, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '18419643', 'n_tokens_mistral': 2698, 'n_tokens_neox': 2554, 'n_words': 705}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: outdated method to acquire releases username_0: ## Description update method to get new release for emotibit SW ## Type of correction - [ ] Permission Issue(Any page you do not have access to - [ ] Content Problem - [ ] Inappropriate content - [x] Outdated/Deprecated content ## Link - Add Link to the page where the problem exits [here](https://github.com/EmotiBit/EmotiBit_Docs/blob/master/Getting_Started/EmotiBit_Getting_Started.md#get-the-software---download-the-oscilloscope) <issue_comment>username_0: done.<issue_closed>
{'fraction_non_alphanumeric': 0.09773123909249563, 'fraction_numerical': 0.0034904013961605585, 'mean_word_length': 4.917525773195877, 'pattern_counts': {'":': 0, '<': 4, '<?xml version=': 0, '>': 4, 'https://': 1, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 1, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '2544479', 'n_tokens_mistral': 182, 'n_tokens_neox': 170, 'n_words': 49}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Error: "expected unqualified-id before numeric constant" username_0: I am simply running the example code for fingerprint and am getting this error without having made any changes. What could I be doing wrong?<issue_closed> <issue_comment>username_1: ? please reopen with compiler / IDE details! you may want to try the latest version we just committed. it fixes many bugs - thx!
{'fraction_non_alphanumeric': 0.055288461538461536, 'fraction_numerical': 0.004807692307692308, 'mean_word_length': 5.318181818181818, 'pattern_counts': {'":': 0, '<': 4, '<?xml version=': 0, '>': 4, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '23983613', 'n_tokens_mistral': 108, 'n_tokens_neox': 101, 'n_words': 56}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: remove dep 'component-emitter' username_0: <issue_comment>username_1: @username_0 I had to roll back this change because `sc-channel` is a server-side dependency of socketcluster on both the client and server-side and it inherits from `component-emitter` (since EventEmitter cannot be used on the client-side). Is there a technical reason why you want to remove component-emitter as a dependency? <issue_comment>username_0: https://github.com/SocketCluster/socketcluster-server/blob/b8b0bed91af387d829c98a2bdbc985d658baaaa9/scserver.js#L4 <issue_comment>username_1: @username_0 Right now the philosophy is that front-end and isomorphic modules inherit from the `component-emitter` Emitter (regardless of if they are running on the client or server) but pure server-side modules inherit from EventEmitter. The server-side SCSocket is not strictly an isomorphic module but because it shares a lot of functionality with the client-side SCSocket (which is isomorphic), I thought it would be best if they are consistent. That said, I think this the way things are right now is definitely confusing... Maybe I should split up the `sc-channel` module into two separate ones; client and server versions; that way the SCChannel doesn't need to be isomorphic and that lets us be more consistent (use EventEmitter everywhere on the backend and Emitter everywhere on the frontend). Note that this is still slightly confusing because client-side != frontend (e.g. you can run socketcluster-client inside Node.js).
{'fraction_non_alphanumeric': 0.05372168284789644, 'fraction_numerical': 0.017475728155339806, 'mean_word_length': 5.233870967741935, 'pattern_counts': {'":': 0, '<': 5, '<?xml version=': 0, '>': 5, 'https://': 1, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '10401632', 'n_tokens_mistral': 407, 'n_tokens_neox': 385, 'n_words': 202}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Trying to install Python ds214play username_0: Hi All, Just had my new 214play delivered and I'm trying to install Python. I have added the SynoCommunity as a source of packages, but when I try to install Python I get the message "failed to install 'Python'. This package is not published by Synology Inc, Any ideas on what I'm doing wrong? Thanks <issue_comment>username_1: Package Center settings > Set trust level to the second option. It's currently set to the first option, and that limits you to installing only Synology packages. <issue_comment>username_0: Legend! Thanks so much for the swift response! As you replied so quickly and obviously know you're stuff, maybe you could help me with another problem I'm having. I'm migrating everything from my old WD MybookLive (adding remote folder in file station) but I'm only getting 3.5MB/s speeds. Anything I can do to speed it up? Thanks again,<issue_closed> <issue_comment>username_1: Well, communication speed depends on a lot of things: the connection between the two (wired vs wireless, 10/100/1000mbps?), the protocol used to communicate (SMB, NFS, other), and whether or not you're maxing out the performance of either device (with regards to CPU etc). It's possible that this is the best you can get. If you're not familiar with how to investigate the above suggestions, I'd say leave it be for now, and start experimenting with it later :)
{'fraction_non_alphanumeric': 0.05065023956194387, 'fraction_numerical': 0.014373716632443531, 'mean_word_length': 4.580152671755725, 'pattern_counts': {'":': 0, '<': 6, '<?xml version=': 0, '>': 7, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '25935044', 'n_tokens_mistral': 400, 'n_tokens_neox': 361, 'n_words': 226}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Immutable interceptors username_0: Hi @username_1 and @username_2, As I briefly mentioned in one of our discussion, method `WithInterceptor` should not modify the upstream chain of interceptors. Please review the proposed solution. <issue_comment>username_1: Great PR .. I'll have a look! <issue_comment>username_1: Thanks @username_0, good catch! <issue_comment>username_2: Hey, @username_0, it's nice, very nice! I like it! <issue_comment>username_1: :+1: <issue_comment>username_2: @username_0, do you know about a good guideline for fluent API? I've found only some basic and simple. <issue_comment>username_0: Hey @username_2, unfortunately I don't know any specific guidelines. But I think we just all intuitively know how fluent APIs are to look and to behave (type safe, immutable, etc). The particular implementation then would simply follow our expectations, and (as you can see) it doesn't have to be complex at all. I'm glad that you liked my solution. :bow: <issue_comment>username_1: This article by Martin Fowler may be helpful: http://martinfowler.com/bliki/FluentInterface.html. One of the best examples of a fluent API is LINQ. Just use something like dotPeek to look at System.Linq.Enumerable. I like Alex's PR to show the benefits of immutability, with the first commit containing just a failing test. I used Git checkout SHA to go to that commit and see the problem with a mutable InterceptorPool, which is that you were not able to get the same result from the first pool after calling the extension method again. <issue_comment>username_2: I know this article from Martin Fowler. It's the introdution to fluent API, but I meant some recommendation about the implementation :-) For example the immutability is something I didn't think about, when I implemented the interceptors here in TE. So I am glad, that @username_0 mentioned that in a comment earlier and for this PR :+1: <issue_comment>username_1: Maybe one of us should write a blog post? :smile: <issue_comment>username_2: :-)
{'fraction_non_alphanumeric': 0.05951219512195122, 'fraction_numerical': 0.00975609756097561, 'mean_word_length': 5.196374622356496, 'pattern_counts': {'":': 0, '<': 12, '<?xml version=': 0, '>': 12, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '4819195', 'n_tokens_mistral': 568, 'n_tokens_neox': 542, 'n_words': 294}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: feat(config): allow to override config via env variables username_0: **Description** Allow to override some configurations with env variables **Changes** * feat(config): allow to override config via env variables 🚀 PR created with [fotingo](https://github.com/username_0/fotingo)
{'fraction_non_alphanumeric': 0.10062893081761007, 'fraction_numerical': 0.006289308176100629, 'mean_word_length': 6.4186046511627906, 'pattern_counts': {'":': 0, '<': 2, '<?xml version=': 0, '>': 2, 'https://': 1, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '2591091', 'n_tokens_mistral': 85, 'n_tokens_neox': 86, 'n_words': 33}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: dataloader which stratifies train-validation split by adata.obs columns username_0: Would be great to have train validations split balanced by cell labels or batch - which is useful when there are rare populations. This can be addressed by adding a 'stratify' option for dataloader and dataslitter like in `sklearn.model_selection.train_test_split(..., stratify=x)`. <issue_comment>username_1: What do you think @galenxing @romain-lopez ? The data splitters for training can have an `stratify_by_obs_key: str` argument. <issue_comment>username_1: @watiss we can also implement internally as an option if it works
{'fraction_non_alphanumeric': 0.059907834101382486, 'fraction_numerical': 0.004608294930875576, 'mean_word_length': 5.653061224489796, 'pattern_counts': {'":': 0, '<': 4, '<?xml version=': 0, '>': 4, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '7441727', 'n_tokens_mistral': 185, 'n_tokens_neox': 180, 'n_words': 80}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: fix: rename conflicting template project name username_0: Just renames the project name in the basic template package.json <issue_comment>username_1: @username_0 Thanks so much for your initiative, Could you change the example name as well, for something like `basic_example`, maybe? <issue_comment>username_1: Could you fix this conflict name too? https://travis-ci.org/username_1/fastexpress/builds/600365244#L334 <issue_comment>username_0: @username_1 For sure, I'll rename the examples and look at this other issue <issue_comment>username_0: Just closed trying to rename the branch, but it didn't go as planned T-T <issue_comment>username_1: No problem @username_0, That is learning process <issue_comment>username_0: Renamed all the examples names using as pattern the auth example name <issue_comment>username_1: Oh, great, Thanks for your help!
{'fraction_non_alphanumeric': 0.06524184476940383, 'fraction_numerical': 0.02699662542182227, 'mean_word_length': 6.5423728813559325, 'pattern_counts': {'":': 0, '<': 9, '<?xml version=': 0, '>': 9, 'https://': 1, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '15666735', 'n_tokens_mistral': 258, 'n_tokens_neox': 244, 'n_words': 105}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: [WIP] Prepare multi module build username_0: This is a follow-up to https://github.com/JabRef/jabref/pull/3704. This PR separates logic and UI at more places. It cannot easily be based on `maintable-beta` (https://github.com/JabRef/jabref/pull/3621). I propose following procedure: 1. Get this ready to be merged into master 2. Merge master into `maintable-beta` (https://github.com/JabRef/jabref/pull/3621) For the first step, just the following is open: - [ ] The OO preferences should be fixed (reading/writing not fully supported yet). All other features should work. Details at https://github.com/username_0/jabref/issues/304 <issue_comment>username_0: OO fixed in 5c8e3e9. Tested in the UI. So, ready for review 😇 <issue_comment>username_0: Since @Siedlerchr gave his "OK" and @tobiasdiez reviewed the intermediate steps, I'll fix tghe tests and merge to keep things moving.
{'fraction_non_alphanumeric': 0.09051724137931035, 'fraction_numerical': 0.02693965517241379, 'mean_word_length': 4.401162790697675, 'pattern_counts': {'":': 0, '<': 4, '<?xml version=': 0, '>': 4, 'https://': 4, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '5297700', 'n_tokens_mistral': 308, 'n_tokens_neox': 283, 'n_words': 114}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: API username_0: I'm thinking of possible additions to this great application and I was wondering if any API exists. Basically for what's in my mind I'd like to start/stop any task or BREAK by API calls, so that I can add an external UI or device. Query of existing tasks would be a plus but no use for this right now. <issue_comment>username_1: There was a DBUS interface at one point (#200 ). Currently there isn't but I very much would welcome re-implementing it. <issue_comment>username_0: No at the moment, closing for now<issue_closed>
{'fraction_non_alphanumeric': 0.05217391304347826, 'fraction_numerical': 0.010434782608695653, 'mean_word_length': 4.818181818181818, 'pattern_counts': {'":': 0, '<': 5, '<?xml version=': 0, '>': 5, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '8437854', 'n_tokens_mistral': 158, 'n_tokens_neox': 148, 'n_words': 91}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: New version: Bioinformatics v0.1.1 username_0: - Registering package: Bioinformatics - Repository: https://github.com/mrtkp9993/Bioinformatics.jl - Created by: @mrtkp9993 - Version: v0.1.1 - Commit: 74a05f21e9569a220fdaa69a5f5a6443e5b65039 - Reviewed by: @mrtkp9993 - Reference: https://github.com/mrtkp9993/Bioinformatics.jl/commit/74a05f21e9569a220fdaa69a5f5a6443e5b65039#commitcomment-34743574 <!-- aa9d7facc4d4bafb5421b09a92ec4a13d01e00e6295f6813f6fdc201ab225ae4367151866664ed8f67be310c01ab70e4e99a643b1f8a5348e68783785b1127c890348b9c81a91f2c8d6f87991fd034cb43adf126f6578b81c511194ae52b2eb34928152015280240aa69b69fb5a401b2b760cfa3eccfbf857f76765d7cad449b265fc7708f491e918faf25c5c4807be06f6b7fd5eeef34a94bc51284a0308675e657ff3dffc2b796c6b3136fe61e4158b12cb1fe0855cf5cb4cab2e772a4a5fd6aac284efae3c9bb9b6d22e6dde9f73a -->
{'fraction_non_alphanumeric': 0.060676779463243874, 'fraction_numerical': 0.39556592765460913, 'mean_word_length': 12.619047619047619, 'pattern_counts': {'":': 0, '<': 3, '<?xml version=': 0, '>': 3, 'https://': 2, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '27906809', 'n_tokens_mistral': 608, 'n_tokens_neox': 430, 'n_words': 24}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Multicast Listening on Unikernel? username_0: Hi all, I'm not sure this is an issue. There is the possibility to listen on some multicast address when using mirage-tcpip inside an unikernel? I see some ~group parameter in `dhcp_ipv4_stack` but I'm not sure is used for multicast groups. I'm trying to modify the simple TCP example to be able to listen also to UDP multicast. config.ml: ``` open Mirage let main = foreign ~packages: [package "duration"; package "fmt"] "Unikernel.Main" (stackv4 @-> console @-> job) let stack = dhcp_ipv4_stack ~group:"239.0.0.5" default_network let () = register "services" [ main $ stack $ default_console] ``` unikernel.ml: ``` open Lwt.Infix module Main (S: Mirage_types_lwt.STACKV4) (C: Mirage_types_lwt.CONSOLE) = struct let report_and_close flow pp e message = let ip, port = S.TCPV4.dst flow in Logs.warn (fun m -> m "closing connection from %a:%d due to error %a while %s" Ipaddr.V4.pp_hum ip port pp e message); S.TCPV4.close flow let task duration marker = ignore @@ OS.Time.sleep_ns (Duration.of_sec duration); Lwt.return @@ marker ^ marker let listener console udp port ~src ~dst ~src_port buf = let s = Format.asprintf "Data from %s to %s" (Ipaddr.V4.to_string src) (Ipaddr.V4.to_string dst) in ignore @@ C.log console s; let src' = (Ipaddr.V4 dst), port in let dst' = (Ipaddr.V4 src), src_port in S.UDPV4.write ~src_port:port ~dst:src ~dst_port:src_port udp buf >>= function | Ok () -> Lwt.return_unit | Error e -> Lwt.return_unit let rec echo flow = S.TCPV4.read flow >>= function | Error e -> report_and_close flow S.TCPV4.pp_error e "reading in Echo" | Ok `Eof -> report_and_close flow Fmt.string "end of file" "reading in Echo" | Ok (`Data buf) -> Lwt.choose[(task 1 "+");(task 2 "=")] >>= ( fun x -> let b = Cstruct.of_string x in S.TCPV4.write flow b >>= function | Ok () -> echo flow | Error e -> report_and_close flow S.TCPV4.pp_write_error e "writing in Echo") let start s c = S.listen_udpv4 s ~port:7 (listener c (S.udpv4 s) 7); S.listen_tcpv4 s ~port:7 echo; S.listen s end ``` These is a way to do that? I need a new `direct stack` with UDP? <issue_comment>username_1: Dear @username_0, thanks for your report. The `"group"` argument of `dhcp_ipv4_stack` is just a name (used for command-line / boot arguments), and has no relation with multicast. The mirage-tcpip stack does not have appropriate multicast support. From the Static_ipv4 module on the receiving side: ```OCaml let of_interest ip = Ipaddr.V4.(compare ip (Prefix.address t.cidr) = 0 || compare ip broadcast = 0 || compare ip (Prefix.broadcast t.cidr) = 0) in if not (of_interest packet.dst) then begin Log.debug (fun m -> m "dropping IP fragment not for us or broadcast %a" Ipv4_packet.pp packet); Lwt.return_unit ``` This means that all packets which are not broadcast or unicast to the configured IP address are dropped. If you're interested in implementing multicast into tcpip, please let us know and we can help where which changes would be necessary.
{'fraction_non_alphanumeric': 0.09527997654646732, 'fraction_numerical': 0.011726766344180592, 'mean_word_length': 2.591578947368421, 'pattern_counts': {'":': 0, '<': 3, '<?xml version=': 0, '>': 23, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 1, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '15346836', 'n_tokens_mistral': 1182, 'n_tokens_neox': 1092, 'n_words': 411}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Long username_0: recordalerting-1542103259-2723947750748304947 - 3605583 [0] - recordalerting-Repeat-1h0m0s (START) [116] - GetRecords-false-false (START_FUNCTION) [125] - GetRecords-Search-3549 (MARKER) [161] - GetRecords (END_FUNCTION) [3605583] - recordalerting-Repeat-1h0m0s (END)<issue_closed>
{'fraction_non_alphanumeric': 0.13213213213213212, 'fraction_numerical': 0.1921921921921922, 'mean_word_length': 6.9523809523809526, 'pattern_counts': {'":': 0, '<': 3, '<?xml version=': 0, '>': 3, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '27518466', 'n_tokens_mistral': 166, 'n_tokens_neox': 131, 'n_words': 20}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: fix tbl_ddl view username_0: Hello, I used this view earlier today and found a bug with the created ddl statements for some of our tables. Since we have lots of clients and some of them have reserved words for column names, for example: from, table, order, group, etc. If the source table has a reserved word for a column name the output ddl would fail. This fix allows you to create those fields. Thank in advance, Erez <issue_comment>username_1: Thanks so much - good addition.
{'fraction_non_alphanumeric': 0.04364326375711575, 'fraction_numerical': 0.003795066413662239, 'mean_word_length': 4.076923076923077, 'pattern_counts': {'":': 0, '<': 3, '<?xml version=': 0, '>': 3, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '10884538', 'n_tokens_mistral': 142, 'n_tokens_neox': 141, 'n_words': 84}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Add button for theme isn't aligned in search results if name is longer thant one line username_0: STR: 1) Open a theme results page with themes that have long names. [This one should do](https://addons.mozilla.org/en-US/firefox/search/?q=madonna&cat=themes). 2) Hover over a search result that has a name that spans more than one line. Result: The Add button isn't aligned with the bottom left of the theme preview, and overlaps with the theme name. This can make you accidentally install a theme when you just wanted to go to its page. <issue_comment>username_1: Issue is also reproducing for long usernames. Please see the screencast: http://screencast.com/t/2ghyJuxal<issue_closed> <issue_comment>username_1: Verified as fixed on AMO-dev FF44(Win 7). Postfix screenshot: ![2016-02-24_1847](https://cloud.githubusercontent.com/assets/15685960/13293133/17a2f9bc-db27-11e5-9b18-5817d8aef8be.png)
{'fraction_non_alphanumeric': 0.08297872340425531, 'fraction_numerical': 0.05851063829787234, 'mean_word_length': 4.918238993710692, 'pattern_counts': {'":': 0, '<': 5, '<?xml version=': 0, '>': 5, 'https://': 2, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '1210274', 'n_tokens_mistral': 329, 'n_tokens_neox': 293, 'n_words': 112}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Update two tests for changes in rustc output username_0: Required for landing https://github.com/rust-lang/rust/pull/63250. (I'm not sure what is the process for landing synchronous changes like this.) <issue_comment>username_1: r? @username_4 (rust_highfive has picked a reviewer for you, use r? to override) <issue_comment>username_2: The easiest thing here is probably to relax the assertion by changing `module` to `[..]`, and that should pass on both I think? <issue_comment>username_0: Updated with `[..]` instead of `crate`. <issue_comment>username_2: @username_3: r+ <issue_comment>username_3: :pushpin: Commit 695931608ab076fa31f49bd7c65418bb19555eeb has been approved by `username_2` <!-- @username_3 r=username_2 695931608ab076fa31f49bd7c65418bb19555eeb --> <!-- homu: {"type":"Approved","sha":"695931608ab076fa31f49bd7c65418bb19555eeb","approver":"username_2"} --> <issue_comment>username_3: :hourglass: Testing commit 695931608ab076fa31f49bd7c65418bb19555eeb with merge e853aa976543168fbb6bfcc983c35c3facca9840... <!-- homu: {"type":"BuildStarted","head_sha":"695931608ab076fa31f49bd7c65418bb19555eeb","merge_sha":"e853aa976543168fbb6bfcc983c35c3facca9840"} --> <issue_comment>username_3: :sunny: Test successful - [checks-azure](https://dev.azure.com/rust-lang/473adcf1-6656-4c7e-9337-c393e7e0431c/_build/results?buildId=5518) Approved by: username_2 Pushing e853aa976543168fbb6bfcc983c35c3facca9840 to master... <!-- homu: {"type":"BuildCompleted","approved_by":"username_2","base_ref":"master","builders":{"checks-azure":"https://dev.azure.com/rust-lang/473adcf1-6656-4c7e-9337-c393e7e0431c/_build/results?buildId=5518"},"merge_sha":"e853aa976543168fbb6bfcc983c35c3facca9840"} --> <issue_comment>username_4: @username_0 Feel free to update the cargo submodule in your PR if you think it will get merged within the next week or so. Or if you'd prefer, I can post an update PR, but you'll have to wait for it to make its way through the username_3 queue. Just let me know.
{'fraction_non_alphanumeric': 0.11883629191321499, 'fraction_numerical': 0.14940828402366865, 'mean_word_length': 7.051587301587301, 'pattern_counts': {'":': 12, '<': 14, '<?xml version=': 0, '>': 14, 'https://': 3, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '15530397', 'n_tokens_mistral': 876, 'n_tokens_neox': 722, 'n_words': 165}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Delete username_0: # TODO + [x] Delete actually deletes locally @username_0 + [x] Sync deals with delete properly @username_0 + [x] Deletes locally that which was deleted elsewhere @username_0 + [x] If the file has local edits, recovers them to a `recovered` folder @username_0 + [x] Integration tests @username_0 + [ ] CLI @steverusso + [x] Android @SmailBarkouch + [x] iOS @raayan (in bad modal for now, will revamp in #391) + [x] macOS @raayan (in bad modal for now, will revamp in #391) + [x] Linux @steverusso + [ ] Windows @username_1 + [x] update kotlin tests @SmailBarkouch + [x] update swift tests @raayan + [ ] update c# tests @jhecker22 (or @username_1) # Changes to Usage Service + `file_id`: `Uuid`, not `String` + Calculate usage based on what is sent to `s3` not based on `.content` Closes #153 Closes #186 <issue_comment>username_1: there's a problem with the CLI on windows: ``` C:\Users\tvand\code\lockbook\clients\cli>lockbook remove travis5/doc2.txt C:\Users\tvand\code\lockbook\clients\cli>lockbook list Last successful sync: just now C:\Users\tvand\code\lockbook\clients\cli>lockbook sync Pushing: doc2.txt Skipped: Unexpected("Deserialize(\n Error(\"EOF while parsing a value\", line: 1, column: 0),\n)") Pushing: doc2.txt Skipped: Unexpected("Deserialize(\n Error(\"EOF while parsing a value\", line: 1, column: 0),\n)") Pushing: doc2.txt Skipped: Unexpected("Deserialize(\n Error(\"EOF while parsing a value\", line: 1, column: 0),\n)") Pushing: doc2.txt Skipped: Unexpected("Deserialize(\n Error(\"EOF while parsing a value\", line: 1, column: 0),\n)") Pushing: doc2.txt Skipped: Unexpected("Deserialize(\n Error(\"EOF while parsing a value\", line: 1, column: 0),\n)") Pushing: doc2.txt Skipped: Unexpected("Deserialize(\n Error(\"EOF while parsing a value\", line: 1, column: 0),\n)") Pushing: doc2.txt Skipped: Unexpected("Deserialize(\n Error(\"EOF while parsing a value\", line: 1, column: 0),\n)") Pushing: doc2.txt Skipped: Unexpected("Deserialize(\n Error(\"EOF while parsing a value\", line: 1, column: 0),\n)") Pushing: doc2.txt Skipped: Unexpected("Deserialize(\n Error(\"EOF while parsing a value\", line: 1, column: 0),\n)") Pushing: doc2.txt Skipped: Unexpected("Deserialize(\n Error(\"EOF while parsing a value\", line: 1, column: 0),\n)") Pushing: doc2.txt Skipped: Unexpected("Deserialize(\n Error(\"EOF while parsing a value\", line: 1, column: 0),\n)") Pushing: doc2.txt Skipped: Unexpected("Deserialize(\n Error(\"EOF while parsing a value\", line: 1, column: 0),\n)") Pushing: doc2.txt Skipped: Unexpected("Deserialize(\n Error(\"EOF while parsing a value\", line: 1, column: 0),\n)") Pushing: doc2.txt Skipped: Unexpected("Deserialize(\n Error(\"EOF while parsing a value\", line: 1, column: 0),\n)") ``` <issue_comment>username_1: also the `remove` description has a TODO and I recommend renaming the command to `delete` (it's gonna get real ambiguous when we have `archive`) ``` C:\Users\tvand\code\lockbook\clients\cli>lockbook lockbook 0.2.18 A secure and intuitive notebook. USAGE: lockbook <SUBCOMMAND> FLAGS: -h, --help Prints help information -V, --version Prints version information SUBCOMMANDS: backup Backup your Lockbook files and structure to the current directory copy Bring a file from your computer into a target destination inside your Lockbook. If your Lockbook target destination is a Folder, the file name will be taken from the file-system file edit Open a document for editing export-private-key Export your private key get-usage Calculate how much space your Lockbook is occupying help Prints this message or the help of the given subcommand(s) import-private-key Import an existing Lockbook list List all your paths list-all List all your files (the things you can filter for rename and move) list-docs List only documents (the things you can filter for edit) list-folders List all your folders (the things you can filter for the start of new) move Move a specified file such that it has the target parent (list-all for first parameter list-folders for second parameter) new Create a new document or folder new-account Create a new Lockbook account print Print the contents of a file remove Move a file to trash TODO rename Rename a file at a path to a target value status What operations a sync would perform sync Get updates, push changes whoami Display Lockbook username ```
{'fraction_non_alphanumeric': 0.09365225390984361, 'fraction_numerical': 0.01343146274149034, 'mean_word_length': 2.0781426953567386, 'pattern_counts': {'":': 0, '<': 5, '<?xml version=': 0, '>': 9, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 14, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '22673678', 'n_tokens_mistral': 1568, 'n_tokens_neox': 1484, 'n_words': 569}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Is there has plan to support new device types? username_0: Like tvoc, PM, pressure, brightness and other sensors? <issue_comment>username_1: You can easily create any accessory you want. The accessory type is just a wrapper around a specific service. There is currently no plans to predefine those accessory types.<issue_closed>
{'fraction_non_alphanumeric': 0.0521978021978022, 'fraction_numerical': 0.005494505494505495, 'mean_word_length': 5.886792452830188, 'pattern_counts': {'":': 0, '<': 4, '<?xml version=': 0, '>': 4, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '2864604', 'n_tokens_mistral': 92, 'n_tokens_neox': 89, 'n_words': 49}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Bluetooth speaker authorization problem username_0: hi I ran into a problem when I set the CONFIG_SPIRAM_MALLOC_ALWAYSINTERNAL parameter less than 128 bytes (64 and below), then when the esp_a2d_source_connect function is called, the application hangs and restarts. If it is equal to 128 or more, there is no problem. The problem arises only in the process of authorizing a new speaker. If you connect to a previously authorized one, then there is no error when CONFIG_SPIRAM_MALLOC_ALWAYSINTERNAL = 8<issue_closed> <issue_comment>username_1: @username_0 We do not recommend set CONFIG_SPIRAM_MALLOC_ALWAYSINTERNAL too low, the behavior is not predictable.
{'fraction_non_alphanumeric': 0.034532374100719423, 'fraction_numerical': 0.01870503597122302, 'mean_word_length': 5.891089108910891, 'pattern_counts': {'":': 0, '<': 4, '<?xml version=': 0, '>': 4, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '2584235', 'n_tokens_mistral': 202, 'n_tokens_neox': 183, 'n_words': 92}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: [Suggestion] Startup flag for new computers username_0: Hi there, I think Tron is a really great script to use, and it's something I like to run on new PCs to make sure they are a bit cleaner than what would come out of the box, I'm also using it with my job to remove some things from school laptops such as the Xbox apps (because it's completely unnecessary to have) One thing which I think would be a great addition is a flag for new PCs so Tron skips some things such as Anti-Malware & Anti-Virus scans, debloating 3rd part apps, cleaning cache, creating a system restore point, ect. Basically anything that's not going to be needed on a new PC <issue_comment>username_1: Agreed. Tron is a great way of getting rid of the garbage on new systems, especially OEM systems with all their bloatware. <issue_comment>username_0: That's for the suggestion, but I feel like that's not as helpful as a flag for a new computer as that'll still leave out windows optimizations and anti spyware patches, which is why I felt like it would be helpful as a flag <issue_comment>username_2: In that case, running the Stage 2 and Stage 4 scripts should do everything you're looking for (bloat removal, telemetry disabling, etc). Tron doesn't really do any optimizations to Windows outside of a couple _very_ minor tweaks; the real meat of it is in Stages 2 and 4.
{'fraction_non_alphanumeric': 0.03751803751803752, 'fraction_numerical': 0.006493506493506494, 'mean_word_length': 4.592741935483871, 'pattern_counts': {'":': 0, '<': 5, '<?xml version=': 0, '>': 5, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '9170302', 'n_tokens_mistral': 368, 'n_tokens_neox': 345, 'n_words': 237}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: refactor sources to source pkg username_0: This PR refactor different sources to `sources` package in case we will add more sources in the future. /cc @davidopp @dchen1107 @username_2 @username_1 <issue_comment>username_1: I can get to this post v1.8 In general, I think this is a good refactor <issue_comment>username_0: @username_1 Thanks for the reply. Hope you guys take a look at this once you have time. <issue_comment>username_0: @username_1 Done. PTAL. <issue_comment>username_1: Merge bot has gone insane, manually merged. <issue_comment>username_2: @username_1 @username_0 Sorry about the merge bot spams. It should stop now. I've disabled it.
{'fraction_non_alphanumeric': 0.06493506493506493, 'fraction_numerical': 0.025974025974025976, 'mean_word_length': 5.36697247706422, 'pattern_counts': {'":': 0, '<': 7, '<?xml version=': 0, '>': 7, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '12548181', 'n_tokens_mistral': 213, 'n_tokens_neox': 206, 'n_words': 93}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Unable to generate @public or @private username_0: I haven't been able to get `@public` or `@private` options generated using the following config: ```vim let g:jsdoc_allow_input_prompt = 1 let g:jsdoc_input_description = 1 let g:jsdoc_access_descriptions = 2 let g:jsdoc_underscore_private = 1 let g:jsdoc_enable_es6 = 1 ``` Is there supposed to be an additional prompt when using the `:JsDoc` command? <issue_comment>username_1: Sorry for late response. Add following settings to my .vimrc. ```viml let g:jsdoc_enable_es6 = 1 let g:jsdoc_allow_input_prompt = 1 let g:jsdoc_input_description = 1 let g:jsdoc_access_descriptions = 2 let g:jsdoc_underscore_private = 1 ``` ```javascript function _foo() { } ``` Type `:JsDoc` and generate like followings. ```javascript /** * _foo * * @private */ function _foo() { } ``` Could you show me steps to reproduce? <issue_comment>username_1: Not reproduced. If you reproduced, reopen this issue.<issue_closed>
{'fraction_non_alphanumeric': 0.0970873786407767, 'fraction_numerical': 0.014563106796116505, 'mean_word_length': 4.603260869565218, 'pattern_counts': {'":': 0, '<': 5, '<?xml version=': 0, '>': 5, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '17222604', 'n_tokens_mistral': 380, 'n_tokens_neox': 348, 'n_words': 109}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: vdirsyncer fails to sync username_0: ## Issue description `vdirsyncer sync` fails with the vdirsyncer version from nixos-unstable. it works if I install vdirsyncer with pip. ``` $ vdirsyncer -vdebug sync debug: Using 1 maximal workers. debug: Fetching value for password.fetch with command strategy. Syncing contacts/32 error: Unknown error occurred for contacts/32: Server did not return etag. error: Use `-vdebug` to see the full traceback. debug: File "/nix/store/y9qgx26b2w2lzqhqb7sd9cb6cz8dj7pa-vdirsyncer-unstable-2018-08-05/lib/python3.7/site-packages/vdirsyncer/cli/tasks.py", line 69, in sync_collection debug: partial_sync=pair.partial_sync debug: File "/nix/store/y9qgx26b2w2lzqhqb7sd9cb6cz8dj7pa-vdirsyncer-unstable-2018-08-05/lib/python3.7/site-packages/vdirsyncer/sync/__init__.py", line 143, in sync debug: b_nonempty = b_info.prepare_new_status() debug: File "/nix/store/y9qgx26b2w2lzqhqb7sd9cb6cz8dj7pa-vdirsyncer-unstable-2018-08-05/lib/python3.7/site-packages/vdirsyncer/sync/__init__.py", line 60, in prepare_new_status debug: if prefetch else ()): debug: File "/nix/store/y9qgx26b2w2lzqhqb7sd9cb6cz8dj7pa-vdirsyncer-unstable-2018-08-05/lib/python3.7/site-packages/vdirsyncer/storage/base.py", line 149, in get_multi debug: item, etag = self.get(href) debug: File "/nix/store/y9qgx26b2w2lzqhqb7sd9cb6cz8dj7pa-vdirsyncer-unstable-2018-08-05/lib/python3.7/site-packages/vdirsyncer/storage/_rust.py", line 42, in get debug: native.check_error(e) debug: File "/nix/store/y9qgx26b2w2lzqhqb7sd9cb6cz8dj7pa-vdirsyncer-unstable-2018-08-05/lib/python3.7/site-packages/vdirsyncer/native.py", line 35, in check_error debug: errors.check_exception(e[0]) debug: File "/nix/store/4xg6a6pqm9v3b1vfmjncqy0w18iqfrhj-python3.7-shippai-0.3.2/lib/python3.7/site-packages/shippai/__init__.py", line 150, in check_exception debug: raise exc error: 1 out of 2 tasks failed. ``` ## Technical details Output of `nix-shell -p nix-info --run "nix-info -m"`: ``` - system: `"x86_64-linux"` - host os: `Linux 4.19.30, NixOS, 19.09pre173349.07b42ccf2de (Loris)` - multi-user?: `yes` - sandbox: `yes` - version: `nix-env (Nix) 2.2` - channels(root): `"nixos-19.09pre173349.07b42ccf2de"` - channels(****): `"home-manager"` - nixpkgs: `/nix/var/nix/profiles/per-user/root/channels/nixos` ``` <issue_comment>username_1: Which `vdirsyncer` do you get via `pip`? Is it the old 0.16 branch or the new 0.17 one (which was mostly rewritten in rust)? We have the latest git version of 0.17 in nixpkgs. Unfortunately, [vdirsyncer is currently looking for a maintainer](https://github.com/pimutils/vdirsyncer/issues/790) so it might be some time before this bug is fixed. <issue_comment>username_0: `--version` outputs this: ``` $ ~/.local/bin/vdirsyncer --version vdirsyncer, version 0.16.7 ``` <issue_comment>username_2: Same problem here. Was there any reason to go for an alpha release except for the issue referenced in #33050? If not, a downgrade to 0.16.7 should be fine, since the original bug is still open and probably won't be fixed for a while without anyone maintaining vdirsyncer. <issue_comment>username_1: There are [a few bugs that](https://github.com/pimutils/vdirsyncer/issues/740#issuecomment-395047933) [are fixed](https://github.com/pimutils/vdirsyncer/issues/745#issuecomment-396934915) [in master](https://github.com/pimutils/vdirsyncer/issues/761#issuecomment-408588673), so I'd rather not downgrade. I'd be happy with a `vdirsyncerStable` package for the 0.16 branch though. <issue_comment>username_2: That'd be great! I haven't looked into packaging yet, so it would take some time if I tried. Would you mind helping out? <issue_comment>username_0: I just opened a PR. It’s my first time contributing to `nixpkgs`, so I hope it’s up to snuff.<issue_closed>
{'fraction_non_alphanumeric': 0.1167864476386037, 'fraction_numerical': 0.07007186858316222, 'mean_word_length': 4.535511363636363, 'pattern_counts': {'":': 0, '<': 9, '<?xml version=': 0, '>': 9, 'https://': 4, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '23230872', 'n_tokens_mistral': 1626, 'n_tokens_neox': 1460, 'n_words': 355}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: How do I run srrg2_proslam on my data? username_0: I was able to build and run `srrg2_executor` on kitty 00 following the instructions in https://github.com/srrg-sapienza/srrg2_executor/blob/master/srrg2_executor/readme.md and https://github.com/srrg-sapienza/srrg2_proslam/blob/master/README.md. What is the easiest way to run `srrg2_proslam` on my data? I'd like to start with RGBD camera and later on add IMU (orientation, rotational speeds, linear accelerations), altimeter and single point lidars looking up and down for drone or wheel odometry for skid steering robot (=traveled distance very reliable, orientation very unreliable). <issue_comment>username_1: Hello, To run `srrg2_proslam` on your data you should: - have the recordings in a bag or BOSS format (converting tools BOSS <-> bag are located in `srrg2_core_ros`) - generate a configuration file (maybe copying one of the config file located in this repo) - modify the config file (topics in particular) either from `srrg2_config_visualizer`, `srrg2_executor` or manually with a text editor Note that IMU is still unsupported in the `srrg2` framework. If you want to contribute, you should create an IMU-ready factor in the `srrg2_solver`, and add `srrg2_slam_interfaces` instantiations of the "slices" relative to the IMU (AlignerSlice, RawDataPreprocessor ...). For, this last step, we are setting-up some tutorials in the [wiki of `srrg2_slam_interfaces`](https://github.com/srrg-sapienza/srrg2_slam_interfaces/wiki). Hope this helps. Cheers
{'fraction_non_alphanumeric': 0.07115384615384615, 'fraction_numerical': 0.01217948717948718, 'mean_word_length': 5.073929961089494, 'pattern_counts': {'":': 0, '<': 4, '<?xml version=': 0, '>': 4, 'https://': 3, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '25478483', 'n_tokens_mistral': 511, 'n_tokens_neox': 491, 'n_words': 195}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Add Linters username_0: <!--- Provide a general summary of the issue in the Title above --> ## Detailed Description Add the same linters used in other OCF repos, and fix any issues ## Context Its good to keep the code style and such the same throughout our codebase ## Possible Implementation Any of the .github folders in the main OCF repos should work <issue_comment>username_0: This has been done for awhile now. Still need unit tests though<issue_closed>
{'fraction_non_alphanumeric': 0.055445544554455446, 'fraction_numerical': 0.0039603960396039604, 'mean_word_length': 4.270833333333333, 'pattern_counts': {'":': 0, '<': 5, '<?xml version=': 0, '>': 5, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '1442764', 'n_tokens_mistral': 143, 'n_tokens_neox': 127, 'n_words': 72}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: upgraded es-abstract to version 1.9.0 username_0: <issue_comment>username_1: This doesn't actually change anything? <issue_comment>username_0: Please could you restate what you want so I get clear on what exactly I'm up on. Thank you Though I did upgrade es-abstract to the specified version <issue_comment>username_1: It looks like es-abstract is already at v1.9.0 on master, which would explain why there are no changes in the "files changed" tab: https://github.com/username_1/first-pr/blob/gh-pages/yarn.lock#L469
{'fraction_non_alphanumeric': 0.07719928186714542, 'fraction_numerical': 0.025134649910233394, 'mean_word_length': 5.131868131868132, 'pattern_counts': {'":': 0, '<': 5, '<?xml version=': 0, '>': 5, 'https://': 1, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '6992075', 'n_tokens_mistral': 172, 'n_tokens_neox': 162, 'n_words': 67}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Detect symlink-usability by trying to create a link username_0: Actually symlinks are supported on windows; it's just a bit weird that you have to assign permission to the user. This patch detects if the user has the ability to create symlinks and then conditionally activates their usage in straight. The rant about MS Windows would have to fall victim to this patch. <issue_comment>username_1: This patch requires several disk IO operations at startup to support an infrequently exercised use case only relevant on Windows. I would prefer to avoid this, especially as the workaround (change the value of `straight-use-symlinks` in your init-file) is trivial, intuitive, and well-documented. We could save this data in the build cache, but then the user would have to manually clear the build cache after updating permissions, which is undesirable. We could instead check whether symlinks are supported at the time we try to create one, but this has the disadvantage that we can't tell whether to enable `straight-symlink-emulation-mode` at startup. How about a combination of both, where the value is cached but we recompute it the first time we try to create a symlink in any given session? If this seems like premature optimization, it may be relevant to note that `straight.el` is highly optimized enough that the disk IO for loading 100 packages (without changes, using live modification detection) is approximately only reading one file from disk, listing one directory, and checking for the existence of 100 other directories. See also https://github.com/username_1/straight.el/issues/9#issuecomment-337435499 for how expensive disk IO can be. <issue_comment>username_0: Thank you for the response. I wasn't aware of the performance considerations involved. Setting the variable manually is absolutely no problem, of course. The solution you described is maybe a bit too complex, considering that Windows is an infrequently used host system. But I'll let you decide on this as I don't know the code. Anyway, I find it confusing that the docstring of straight-use-symlinks says that on Windows "you will have to use copying". This is not true. Instead, I think the docu could mention that, on Windows, the user needs to grant link-creation permissions and then set the variable manually. I think this makes it a bit easier for newcomers to get that setting right. What do you think? <issue_comment>username_1: This sounds fine to me, and it's probably the best solution—trying to do the right thing is great, but there is a point when it just adds needless complexity. I would gladly accept a change to the documentation which makes it clear that symlinks don't work by default, but that you can make them work via the permissions change that you mentioned. I recall reading somewhere that you also have to disable UAC in order for symlinks to work on Windows 10—is that true? If so, it should be documented. One reason that I didn't document the possibility of symlinks on Windows is that I am not familiar with the exact circumstances under which they can be used, so I appreciate your help. <issue_comment>username_0: On MS Windows the situation is quite diverse - I'm not sure if we have to document the entire mess in the docstring. - Symlinks are not available before Windows Vista and on FAT (ok...) - With UAC disabled it suffices to set the appropriate security policies - With UAC enabled: - users _in the administrators group_ need to run apps in elevated mode to create symlinks (see [here](https://community.perforce.com/s/article/3472) and [here](https://stackoverflow.com/questions/29063916/win32api-symlink-creation-issue-with-uac-enabled#29065060) and, for an official reference, section Access Token Changes [in this document](https://msdn.microsoft.com/en-us/library/bb530410.aspx#vistauac_topic8) - users not in the admin group can create symlinks if the security policies are set accordingly - [Windows Creators Update](https://blogs.windows.com/buildingapps/2016/12/02/symlinks-windows-10/) seems to support symlink creation out of the box I've drafted a reduced docstring to give hints for to the most important steps. Do you think it's enough information? <issue_comment>username_0: I've updated the docu. I decided to put the part on setting up MS Windows mainly into the User Manual and simply link to it from the docstring.
{'fraction_non_alphanumeric': 0.045280468574003155, 'fraction_numerical': 0.014417661635503492, 'mean_word_length': 4.4950495049504955, 'pattern_counts': {'":': 0, '<': 7, '<?xml version=': 0, '>': 7, 'https://': 5, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '18449218', 'n_tokens_mistral': 1148, 'n_tokens_neox': 1064, 'n_words': 657}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: fix(kv): ignore not found error when removing org dep resources username_0: Closes https://github.com/influxdata/influxdb/issues/14876 Issue only happens in cloud, if an org dependent resource doesn't exist, such an error shouldn't be a stopper to prevent the whole remove process. <!-- Checkboxes below this note can be erased if not applicable to your Pull Request. --> - [ ] [CHANGELOG.md](https://github.com/influxdata/influxdb/blob/master/CHANGELOG.md) updated with a link to the PR (not the Issue) - [ ] [Well-formatted commit messages](https://www.conventionalcommits.org/en/v1.0.0-beta.3/) - [ ] Rebased/mergeable - [x] Tests pass - [ ] http/swagger.yml updated (if modified Go structs or API) - [x] Documentation updated or issue created (provide link to issue/pr) - [ ] Signed [CLA](https://influxdata.com/community/cla/) (if not already signed) <issue_comment>username_1: Can you expand on this? What changed to make it so that org dep dependencies aren't being created? <issue_comment>username_0: the cloud onboarding process didn't add the the org dependent bucket resource mapping. So when deleting the urm, the bucket urm couldn't be found. We will have another pr to figure out the onboarding issue for the bucket urm. But for existed urms in cloud, we shouldn't block them from deleting.
{'fraction_non_alphanumeric': 0.09124356144223694, 'fraction_numerical': 0.008830022075055188, 'mean_word_length': 4.210727969348659, 'pattern_counts': {'":': 0, '<': 5, '<?xml version=': 0, '>': 5, 'https://': 4, 'lorem ipsum': 0, 'www.': 1, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '18541118', 'n_tokens_mistral': 408, 'n_tokens_neox': 389, 'n_words': 166}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: (aws-rds): unsupported SqlServerEngineVersion value username_0: ### What is the problem? Looks like the `SqlServerEngineVersion.VER_14_00_3192_2_V1` is no longer supported? I discovered this while trying to run the [`s3-instance`](https://github.com/aws/aws-cdk/blob/v1.130.0/packages/@aws-cdk/aws-rds/test/integ.instance-s3.ts) integration test. To get passed the problem in our pipeline I switched the test to use `SqlServerEngineVersion.VER_14`, but we should still figure out what's happening here since it seems we are currently offering unsupported version strings. ### Reproduction Steps Run this integ test: https://github.com/aws/aws-cdk/blob/v1.130.0/packages/@aws-cdk/aws-rds/test/integ.instance-s3.ts Make sure to use the `1.130.0` tag and not the latest one, since the test was already changed. ### What did you expect to happen? Expected the integ test to successfully deploy. ### What actually happened? Deployment failure: ```console aws-cdk-rds-instance-s3-integ | 20/34 | 1:18:47 PM | CREATE_FAILED | AWS::RDS::DBInstance | Database (DatabaseB269D8BB) Cannot find version 14.00.3192.2.v1 for sqlserver-se (Service: AmazonRDS; Status Code: 400; Error Code: InvalidParameterCombination; Request ID: 6278edb3-d5c6-4424-9f5c-10ab853d4810; Proxy: null) new DatabaseInstance (/Users/epolon/dev/src/github.com/aws/aws-cdk/packages/@aws-cdk/aws-rds/lib/instance.js:386:26) \_ Object.<anonymous> (/Users/epolon/dev/src/github.com/aws/aws-cdk/packages/@aws-cdk/aws-rds/test/integ.instance-s3.js:12:1) \_ Module._compile (internal/modules/cjs/loader.js:1138:30) \_ Object.Module._extensions..js (internal/modules/cjs/loader.js:1158:10) \_ Module.load (internal/modules/cjs/loader.js:986:32) \_ Function.Module._load (internal/modules/cjs/loader.js:879:14) \_ Function.executeUserEntryPoint [as runMain] (internal/modules/run_main.js:71:12) \_ internal/main/run_main_module.js:17:47 ``` ### CDK CLI Version 1.130.0 ### Framework Version 1.130.0 ### Node.js Version v12.17.0 ### OS Mac OS Mojave ### Language Typescript, Python, .NET, Java, Go ### Language Version ALL ### Other information This is what I could find for a list of supported SQLServer versions in RDS: https://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/CHAP_SQLServer.html#SQLServer.Concepts.General.VersionSupport <issue_comment>username_1: Maybe the version was deprecated by RDS...? We for sure deployed it at some point in the past 🙂. We need to deprecate this field.
{'fraction_non_alphanumeric': 0.11843107387661843, 'fraction_numerical': 0.05140898705255141, 'mean_word_length': 3.7418772563176894, 'pattern_counts': {'":': 0, '<': 4, '<?xml version=': 0, '>': 4, 'https://': 3, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '22538858', 'n_tokens_mistral': 969, 'n_tokens_neox': 869, 'n_words': 220}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Remove darkmode toggle and make it its own theme username_0: Darkmode causes issues for using presets that don't have a color mode of `dark`. This is overly opinionated for the blog theme. Adding a secondary theme that sits on top of the blog theme to make this optional behavior would be great. There are multiple ways to do this/make it most useable. Open to ideas. <issue_comment>username_1: Curious what @johno's thoughts are here, but I do like the idea of using horizontal theme composition for something like this, where `gatsby-theme-blog` is the main one you'd use to set up a blog, and `gatsby-theme-blog-dark-mode` is an optional addon that enables the feature. The `gatsby-config` could look something like the following: ```js plugins: [ 'gatsby-theme-blog', 'gatsby-theme-blog-dark-mode', ] ``` <issue_comment>username_0: Horizontal theme composition now implemented. Open to feedback on the warn/error decision and the phrasing of the CLI message.
{'fraction_non_alphanumeric': 0.06521739130434782, 'fraction_numerical': 0.0029644268774703555, 'mean_word_length': 4.505434782608695, 'pattern_counts': {'":': 0, '<': 4, '<?xml version=': 0, '>': 4, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '18133974', 'n_tokens_mistral': 282, 'n_tokens_neox': 270, 'n_words': 143}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Essence Pouches - incorrect inv space username_0: ## 🐛 Bug Report RC trips only providing +89 suddenly after previously providing +216 for having all pouches. ![e1](https://user-images.githubusercontent.com/30452230/121098056-6b5e5480-c838-11eb-8e45-892ac93674ee.png) ![e2](https://user-images.githubusercontent.com/30452230/121098060-6c8f8180-c838-11eb-878a-a369fe72e164.png) <issue_comment>username_0: Req 105 RC<issue_closed>
{'fraction_non_alphanumeric': 0.11991434689507495, 'fraction_numerical': 0.19271948608137046, 'mean_word_length': 6.3125, 'pattern_counts': {'":': 0, '<': 4, '<?xml version=': 0, '>': 4, 'https://': 2, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '5671769', 'n_tokens_mistral': 225, 'n_tokens_neox': 178, 'n_words': 30}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Key Rotation Transform username_0: Looking into using (or documenting) AWS::SecretsManager-2020-07-23 secret rotation. ## Resources AWS::SecretsManager::RotationSchedule HostedRotationLambda https://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/aws-properties-secretsmanager-rotationschedule-hostedrotationlambda.html For the @AWSCloudFormer nerds out there, there's a new hosted template transform that snuck in last week. It generates a Lambda function to rotate secrets for RDS / Redshift / DocDB. Protip: You can use multiple template transforms as a list 💡 https://twitter.com/iann0036/status/1287596970916233217<issue_closed>
{'fraction_non_alphanumeric': 0.07725947521865889, 'fraction_numerical': 0.04664723032069971, 'mean_word_length': 5.87, 'pattern_counts': {'":': 0, '<': 3, '<?xml version=': 0, '>': 3, 'https://': 2, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '8772118', 'n_tokens_mistral': 226, 'n_tokens_neox': 191, 'n_words': 58}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Salesforce.com compatibility: "[NoErrorObjectAvailable] Script error." username_0: When the React-Sight extension is enabled, the Salesforce.com Lightning platform encounters a generic error: "[NoErrorObjectAvailable] Script error.". This does not happen when using the Classic Platform. ![error](https://user-images.githubusercontent.com/4954762/31541084-7350c2b2-b005-11e7-9870-d7557e19df24.png) ![extensions](https://user-images.githubusercontent.com/4954762/31541091-776c86f6-b005-11e7-8f50-4f1c51b1d7b2.png) **Note:** Using only the "React Developer Tools" extension does not give any conflict with Salesforce Lightning. <issue_comment>username_1: We will investigate this. We are working on an update to provide better compatibility <issue_comment>username_2: Seeing another error while using SLDS - an internal version that isn’t public yet, so I can’t help with “details”, but, yeah, this is an issue I am seeing also. :( <issue_comment>username_3: I am also facing "[NoErrorObjectAvailable] Script error." even no React-Sight chrome extension is enabled. <issue_comment>username_4: I'm also getting this error. When I try to update the data dynamically ``` @track tasks = []; // it contains task records handleChange(event) { let response = event.detail; let rec = this.tasks[response.index]; // checked null values rec[response.realApiName] = response.value; // Error on this Line if we comment this part then it works } ``` <issue_comment>username_5: I am also facing "[NoErrorObjectAvailable] Script error." even no React-Sight chrome extension is enabled. It occured in my Sandbox after I performed a Match production and sandbox licenses without a sandbox refresh. <issue_comment>username_6: Same issue with Grammarly for Chrome <issue_comment>username_7: I also encountered this same issue in Chrome only - working fine in Firefox. I do not have Grammarly in either
{'fraction_non_alphanumeric': 0.08179012345679013, 'fraction_numerical': 0.043209876543209874, 'mean_word_length': 5.116352201257862, 'pattern_counts': {'":': 0, '<': 9, '<?xml version=': 0, '>': 9, 'https://': 2, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 1, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '19341590', 'n_tokens_mistral': 594, 'n_tokens_neox': 543, 'n_words': 216}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Add keep_slow option to Buffer username_0: keep_slow: bool, optional if true then we keep values in the slow dict rather than remove them. This can improve performance for repeated storage, but takes up more space. <issue_comment>username_1: Should this be closed as very stale?
{'fraction_non_alphanumeric': 0.04411764705882353, 'fraction_numerical': 0.0058823529411764705, 'mean_word_length': 3.671232876712329, 'pattern_counts': {'":': 0, '<': 3, '<?xml version=': 0, '>': 3, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '21581317', 'n_tokens_mistral': 89, 'n_tokens_neox': 83, 'n_words': 44}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: devcontainer.json context support for docker compose username_0: It seems, that `docker-compose up` command is run inside **devcontainer** folder. This leads to the problems with volumes, defined in docker-compose file. For example, I use volume to persist postgres data through containers lifetime like this: ``` volumes: - ./docker/data/postgres:/var/lib/postgres/data ``` Running `docker-compose up` inside root of the project will create `./docker/data/postgres` folder. BUT: Addressing this docker compose like this in `./.devcontainer/devcontainer.json`: ``` "dockerComposeFile": [ "../docker-compose.yml" ], ``` Will generally resolve it right, but the folder will be created here: `./.devcontainer/docker/data/postgres` So the solution probably can be in specifying **context** for running **docker-compose up** <issue_comment>username_1: Relative paths should be relative to the compose file (https://docs.docker.com/compose/compose-file/). Could you post a minimal working example showing the problem? (devcontainer.json and docker-compose.yml, a Dockerfile if needed) <issue_comment>username_0: This is a link to the repo, is kept as minimum as possible: https://github.com/username_0/compose_bug <issue_comment>username_1: It looks like Docker Compose will use the first docker-compose file for resolving relative paths. When you use `docker-compose -f .devcontainer/docker-compose.extend.yml -f docker-compose.yml` from the command line you will see the same behavior. Try reversing the order of the two docker-compose files in the devcontainer.json. HTH.<issue_closed> <issue_comment>username_0: @username_1 Yep, thanx, this seems to fix the problem. May be it can be specified in documentation, that the order of the files is important? Somewhere here: https://code.visualstudio.com/docs/remote/containers#_extending-your-docker-compose-file-for-development <issue_comment>username_1: /cc @username_2 (see suggestion above on documentation for the order of docker-compose files) <issue_comment>username_2: @username_1 @username_0 Yeah we can document that, but the bigger consideration is the relative path. I think is the only real fix to this. This is effectively what is happening now in the extend case: ```bash cd .devcontainer && docker-compose -f ../docker-compose.yml -f docker-compose.extend.yml up ``` Build has a context that make things work, but for this volume scenario, the only way I can see to fix this is to support `context` property in devcontainer.json for dockerComposeFile like we support for the other cases. <issue_comment>username_2: @username_1 @username_0 Oh wait - Didn't see that the files were reversed in the devcontainer.json [here](https://github.com/username_0/compose_bug). <issue_comment>username_0: @username_2 I'm not sure, but it seems that `docker-compose` resolves the working directory, to which the file paths are relative, as the directory where the first `docker-compose.yml` file is located. So, even if the `docker-compose` command will be called with absolute path to the `docker-compose.yml` file from any working directory - actual working directory will be resolved right. But the problem appears when the `docker-compose` files are more then one. Second and next are called in the context of the first `docker-compose.yml` file, and that's why order is so important. Thank you for updating the docs!
{'fraction_non_alphanumeric': 0.07912025135675521, 'fraction_numerical': 0.005141388174807198, 'mean_word_length': 4.322188449848024, 'pattern_counts': {'":': 1, '<': 11, '<?xml version=': 0, '>': 11, 'https://': 4, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '4192947', 'n_tokens_mistral': 1035, 'n_tokens_neox': 941, 'n_words': 423}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Export zbctl status as json username_0: **Is your feature request related to a problem? Please describe.** Currently we print the topology of the cluster via `zbctl status` and it looks like this: ``` [zell scripts/ ns:zell-chaos]$ k exec -it zell-chaos-zeebe-gateway-5577d6958-r6bfp -- zbctl status --insecure Cluster size: 3 Partitions count: 3 Replication factor: 3 Gateway version: 0.26.0-SNAPSHOT Brokers: Broker 0 - zell-chaos-zeebe-0.zell-chaos-zeebe.zell-chaos.svc.cluster.local:26501 Version: 0.26.0-SNAPSHOT Partition 1 : Follower, Healthy Partition 2 : Follower, Healthy Partition 3 : Follower, Healthy Broker 1 - zell-chaos-zeebe-1.zell-chaos-zeebe.zell-chaos.svc.cluster.local:26501 Version: 0.26.0-SNAPSHOT Partition 1 : Follower, Healthy Partition 2 : Follower, Healthy Partition 3 : Follower, Healthy Broker 2 - zell-chaos-zeebe-2.zell-chaos-zeebe.zell-chaos.svc.cluster.local:26501 Version: 0.26.0-SNAPSHOT Partition 1 : Leader, Healthy Partition 2 : Leader, Healthy Partition 3 : Leader, Healthy ``` This is totally fine for a human beeing, but not if you what to process it. **Describe the solution you'd like** It would be cool if we had a flag or something where we can tell `zbctl` to print the topology as json. **Describe alternatives you've considered** idk - self parsing **Additional context** I need this quite often in building new chaos experiments where I want to know who is the leader for a certain partition or who is taking part of the partition. For our benchmarks it currently works kind of if the partitions are well distributed, like three nodes and three partitions, then you can do something like: ```sh state=$1 # example Leader partition=${2:-3} # For cluster size 3 and replication factor 3 # we know the following partition matrix # partition \ node 0 1 2 # 1 L F F # 2 F L F # 3 F F L # etc. # This means broker 1, 2 or 3 participates on partition 3 # BE AWARE the topology above is just an example and the leader can every node participating node. index=$(($(echo "$topology" \ | grep "Partition $partition" \ | grep -n "$state" -m 1 \ | sed 's/\([0-9]*\).*/\1/') - 1)) pod=$(echo "$pod" | sed 's/\(.*\)\([0-9]\)$/\1/') pod="$pod$index" ``` but this doesn't work if you have not well distributed partitions like 5 nodes and 3 partition etc. <issue_comment>username_0: blocks https://github.com/zeebe-io/zeebe-cluster-testbench/issues/154 <issue_comment>username_1: @username_2 Hi! If you are not already done with this issue, I can take it :) <issue_comment>username_2: Hi @username_1. I thought I could resolve this quickly, but ran into an issue. Feel free to take over. I added a testcase to `clients/go/cmd/zbctl/main_test.go` with a new golden file, and everything seemed to go smoothly until I noticed that the enum values (both partition role and health) are not converted to json properly using the current marshaller. This has to do with the protobuf files. I recommend you try the same, run into the issue and then take a look at `github.com/golang/protobuf/jsonpb` and `google.golang.org/protobuf/runtime/protoiface` to replace the `printJSON` function implementation in `clients/go/cmd/zbctl/internal/commands/root.go`. Let me know if you have any questions and good luck 🍀
{'fraction_non_alphanumeric': 0.08953089244851259, 'fraction_numerical': 0.02631578947368421, 'mean_word_length': 3.168057210965435, 'pattern_counts': {'":': 0, '<': 5, '<?xml version=': 0, '>': 5, 'https://': 1, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '4445201', 'n_tokens_mistral': 1199, 'n_tokens_neox': 1067, 'n_words': 442}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Only works for Apple Music tracks in library username_0: I don't know if this is intended or not or a limitation of the Music/iTunes app, but this is only detecting played songs that are added to my Apple Music library (and not things from random playlists/albums not added to my library). I think this should be possible since last.fm scribblers are able to read tracks played not in the user's library? But I may be wrong. <issue_comment>username_1: Anyt updates on this? <issue_comment>username_2: What do you mean by "things from random playlists/albums not added to my library"? The application should detect everything that gets played within the iTunes/Music app. <issue_comment>username_1: If your listening to apple music from a station (i.e https://music.apple.com/us/station/hip-hop/ra.985488923) It doesnt detect it on the application that your listening to music but if you listen to a song you purchased it detects it <issue_comment>username_2: Ah, I see. I'll have to look into that, if it's even possible to work with those kinds of tracks, because I remember that also happened to me when listening to song previews in the store. <issue_comment>username_0: To clarify, when you have added Apple Music tracks to your library and plays them, it works fine. But if you are using Apple Music to stream music not added to your library (like albums, playlists, or radio stations as mentioned above), then the app does not show the tracks. I'm realizing this application may be intended to only work with locally-stored iTunes tracks and not with the Apple Music service though. <issue_comment>username_2: You're right, but I also assumed it would also work with streamed music, but since it doesn't it must mean that those use a different protocol or format, I'll have to see if it can be worked with. <issue_comment>username_3: This happens to me also. <issue_comment>username_2: The issue has been fixed in the [latest release](https://github.com/username_2/iTunes-Discord-RP/releases/latest), thanks goes to @cawc for finding the fix for it.<issue_closed>
{'fraction_non_alphanumeric': 0.051330798479087454, 'fraction_numerical': 0.00903041825095057, 'mean_word_length': 4.896358543417367, 'pattern_counts': {'":': 0, '<': 11, '<?xml version=': 0, '>': 11, 'https://': 2, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '18950740', 'n_tokens_mistral': 552, 'n_tokens_neox': 526, 'n_words': 320}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Virtualservice created by notebook controller defaults to kubeflow/kubeflow-gateway username_0: /kind bug I have installed kubeflow over an existing k8 cluster with OIDC protocol enabled security. As a part of security the istio-ingress gateway is setup in istio-system and not in kubeflow. So i had to change manifest/<app>/overlays/istio.yaml to point to my hostname and istio-system/kubeflow (the gateway). But i observed when i create jupyter notebook in kfam controlled namespace, the virtual service routes are created properly but the hosts and the gateway are incorrectly created. This causes the notebook to show that my notebook url is not found ``` NAME GATEWAYS HOSTS AGE notebook-anup-t2 [kubeflow/kubeflow-gateway] [*] 17h ``` Note i cannot edit this gateway because it get back to original, seems like its controlled. As a work around i created another dummy gateway and inject it with correct gateway and host and then the notebook starts working. Time being its a hack but i need a long term solution ``` NAME GATEWAYS HOSTS AGE notebook-anup-t2 [kubeflow/kubeflow-gateway] [*] 17h =hack=> notebook-anup-t2-tst [istio-system/kubeflow-gateway] [kf.my.domain] 18h ``` Environment - kfctl version: (use `kfctl version`): 0.6.2 - Kubernetes platform: (e.g. `minikube`): Kops / aws ec2's - Kubernetes version: (use `kubectl version`): ` Client Version: version.Info{Major:"1", Minor:"15", GitVersion:"v1.15.3", GitCommit:"2d3c76f9091b6bec110a5e63777c332469e0cba2", GitTreeState:"clean", BuildDate:"2019-08-19T12:36:28Z", GoVersion:"go1.12.9", Compiler:"gc", Platform:"darwin/amd64"} Server Version: version.Info{Major:"1", Minor:"14", GitVersion:"v1.14.6", GitCommit:"96fac5cd13a5dc064f7d9f4f23030a6aeface6cc", GitTreeState:"clean", BuildDate:"2019-08-19T11:05:16Z", GoVersion:"go1.12.9", Compiler:"gc", Platform:"linux/amd64"} ` - OS (e.g. from `/etc/os-release`): linux <issue_comment>username_1: @username_0 think is this a duplicate of #3945 or am I mistaken? <issue_comment>username_2: @username_1 @jeremievallee if this the same as 3945, should we close one of these issues?
{'fraction_non_alphanumeric': 0.09264644786040714, 'fraction_numerical': 0.052347320315745745, 'mean_word_length': 3.020033388981636, 'pattern_counts': {'":': 0, '<': 5, '<?xml version=': 0, '>': 6, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 4, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '19192400', 'n_tokens_mistral': 830, 'n_tokens_neox': 742, 'n_words': 244}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Error Installation username_0: {"success":false,"error":{"code":11000,"index":0,"errmsg":"E11000 duplicate key error collection: trudesk.counters index: _id_ dup key: { : \"tickets\" }","op":{"_id":"tickets","next":1001,"__v":0}}} any idea ??? <issue_comment>username_1: This happened because the install process was either interrupted. It could have lost connection to the MongoDB server during the install. If you reboot the server and continues to run in install mode. You can try to run the install wizard as an existing MongoDB database, if that fails you will need to empty the database and start the install wizard again.<issue_closed>
{'fraction_non_alphanumeric': 0.10735294117647058, 'fraction_numerical': 0.026470588235294117, 'mean_word_length': 6.24468085106383, 'pattern_counts': {'":': 9, '<': 4, '<?xml version=': 0, '>': 4, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '9350811', 'n_tokens_mistral': 193, 'n_tokens_neox': 174, 'n_words': 81}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: AccessViolationException username_0: ##### Brief Description I'm getting an AccessViolationException when trying to generate the interop classes for [FBX SDK 2019](http://download.autodesk.com/us/fbx/2019/2019.0/fbx20190_fbxsdk_vs2015_win.exe). The error is thrown when reaching this CppSharp method: `CppSharp::Parser::ParserResult^ CppSharp::Parser::ClangParser::ParseHeader(CppSharp::Parser::CppParserOptions^ Opts) { auto __arg0 = (::CppSharp::CppParser::CppParserOptions*)Opts->NativePtr; auto __ret = ::CppSharp::CppParser::ClangParser::ParseHeader(__arg0); if (__ret == nullptr) return nullptr; return (__ret == nullptr) ? nullptr : gcnew CppSharp::Parser::ParserResult((::CppSharp::CppParser::ParserResult*)__ret); }` Specifically, this line: `auto __arg0 = (::CppSharp::CppParser::CppParserOptions*)Opts->NativePtr;` I'm looking for a way to compile LLVM with debug symbols, so I could check what's happening internally, but I can't find a LLVM debug source package on CppSharp repo, and couldn't compile LLVM debug from sources either. OS: Windows / OS X / Linux (include version and/or distro) Windows ##### Used headers `fbxsdh.h` ##### Used settings ` var options = driver.Options; options.Verbose = true; var module = options.AddModule("FBXSDK"); module.IncludeDirs.Add("C:\\Program Files\\Autodesk\\FBX\\FBX SDK\\2019.0\\include"); module.IncludeDirs.Add("C:\\Program Files (x86)\\Microsoft Visual Studio\\2017\\Professional\\VC\\Tools\\MSVC\\14.14.26428\\include"); module.LibraryDirs.Add("C:\\Program Files\\Autodesk\\FBX\\FBX SDK\\2019.0\\lib\\vs2015\\x86\\release"); module.Libraries.Add("libfbxsdk.lib"); ` Target: MSVC/GCC/Clang MSVC Other settings ##### Stack trace or incompilable generated code ` C:\Program Files (x86)\Windows Kits\10\include\10.0.17134.0 Parsing libraries... Parsed 'libfbxsdk.lib' Parsing code... Parsed 'fbxsdk/core/base/fbxarray.h' Parsed 'fbxsdk/core/base/fbxbitset.h' Parsed 'fbxsdk/core/base/fbxcharptrset.h' Parsed 'fbxsdk/core/base/fbxcontainerallocators.h' Parsed 'fbxsdk/core/base/fbxdynamicarray.h' Parsed 'fbxsdk/core/base/fbxstatus.h' Parsed 'fbxsdk/core/base/fbxfile.h' Parsed 'fbxsdk/core/base/fbxhashmap.h' Parsed 'fbxsdk/core/base/fbxintrusivelist.h' Parsed 'fbxsdk/core/base/fbxmap.h' Parsed 'fbxsdk/core/base/fbxmemorypool.h' Parsed 'fbxsdk/core/base/fbxpair.h' Parsed 'fbxsdk/core/base/fbxset.h' Parsed 'fbxsdk/core/base/fbxstring.h' Parsed 'fbxsdk/core/base/fbxstringlist.h' Parsed 'fbxsdk/core/base/fbxtime.h' Parsed 'fbxsdk/core/base/fbxtimecode.h' Parsed 'fbxsdk/core/base/fbxutils.h' Parsed 'fbxsdk/core/math/fbxmath.h' Parsed 'fbxsdk/core/math/fbxdualquaternion.h' Parsed 'fbxsdk/core/math/fbxmatrix.h' Parsed 'fbxsdk/core/math/fbxquaternion.h' Parsed 'fbxsdk/core/math/fbxvector2.h' Parsed 'fbxsdk/core/math/fbxvector4.h' Parsed 'fbxsdk/core/fbxclassid.h' Parsed 'fbxsdk/core/fbxconnectionpoint.h' Parsed 'fbxsdk/core/fbxdatatypes.h' ` <issue_comment>username_1: What issue did you get when compiling LLVM from source in debug mode? We provide some build scripts which you can try to use: https://github.com/mono/CppSharp/blob/master/docs/BuildingLLVM.md#compiling-using-the-build-script If you hit any issue, please me know. To force debug LLVM you can add `debug = true` in the Lua build script: https://github.com/mono/CppSharp/blob/master/build/scripts/LLVM.lua#L142 <issue_comment>username_0: I'm facing some issues when trying to compile LLVM from source. First, if I use the provided scripts to download LLVM and Clang, I get the right files downloaded, but 7Zip don't extract the Tarballs automatically. If I extract the Tarballs at the "Scripts" folder, the build_llvm command will fail, so I have to extract them at "deps/llvm" and "deps/clang" respectively, this way the compilation proccess will start just fine, but at the end of the compilation it will throw the message: `ninja -C "C:/Repos/CppSharp/deps/llvm/llvm-bfd4fd-windows-vs2017-x86-Debug" clang-headers ninja: Entering directory `C:/Repos/CppSharp/deps/llvm/llvm-bfd4fd-windows-vs2017-x86-Debug' ninja: error: unknown target 'clang-headers'` <issue_comment>username_0: CppSharp.CppParser.dll!clang::Parser::LateTemplateParserCallback(void * P, clang::LateParsedTemplate & LPT)Linha 1356 C++ CppSharp.CppParser.dll!clang::Sema::InstantiateFunctionDefinition(clang::SourceLocation PointOfInstantiation, clang::FunctionDecl * Function, bool Recursive, bool DefinitionRequired, bool AtEndOfTU)Linha 3852 C++ CppSharp.CppParser.dll!clang::Sema::PerformPendingInstantiations(bool LocalOnly)Linha 5180 C++ CppSharp.CppParser.dll!clang::Sema::GlobalEagerInstantiationScope::perform()Linha 7628 C++ CppSharp.CppParser.dll!clang::Sema::InstantiateFunctionDefinition(clang::SourceLocation PointOfInstantiation, clang::FunctionDecl * Function, bool Recursive, bool DefinitionRequired, bool AtEndOfTU)Linha 3973 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::MarkValidity(CppSharp::CppParser::AST::Function * F)Linha 3065 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::WalkFunction(const clang::FunctionDecl * FD, CppSharp::CppParser::AST::Function * F, bool IsDependent)Linha 3217 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::WalkMethodCXX(const clang::CXXMethodDecl * MD)Linha 1787 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::WalkRecord(const clang::RecordDecl * Record, CppSharp::CppParser::AST::Class * RC)Linha 987 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::WalkRecordCXX(const clang::CXXRecordDecl * Record, CppSharp::CppParser::AST::Class * RC)Linha 1048 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::WalkClassTemplateSpecialization(const clang::ClassTemplateSpecializationDecl * CTS)Linha 1196 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::WalkDeclaration(const clang::Decl * D)Linha 3749 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::WalkType(clang::QualType QualType, const clang::TypeLoc * TL, bool DesugarType)Linha 2297 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::GetQualifiedType(const clang::QualType & qual, const clang::TypeLoc * TL)Linha 575 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::WalkType(clang::QualType QualType, const clang::TypeLoc * TL, bool DesugarType)Linha 2479 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::GetQualifiedType(const clang::QualType & qual, const clang::TypeLoc * TL)Linha 575 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::WalkDeclaration(const clang::Decl * D)Linha 3854 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::WalkType(clang::QualType QualType, const clang::TypeLoc * TL, bool DesugarType)Linha 2258 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::GetQualifiedType(const clang::QualType & qual, const clang::TypeLoc * TL)Linha 575 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::WalkFieldCXX(const clang::FieldDecl * FD, CppSharp::CppParser::AST::Class * Class)Linha 1827 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::ReadClassLayout(CppSharp::CppParser::AST::Class * Class, const clang::RecordDecl * RD, clang::CharUnits Offset, bool IncludeVirtualBases)Linha 164 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::WalkRecord(const clang::RecordDecl * Record, CppSharp::CppParser::AST::Class * RC)Linha 949 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::WalkRecordCXX(const clang::CXXRecordDecl * Record, CppSharp::CppParser::AST::Class * RC)Linha 1048 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::WalkClassTemplateSpecialization(const clang::ClassTemplateSpecializationDecl * CTS)Linha 1196 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::WalkDeclaration(const clang::Decl * D)Linha 3749 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::WalkType(clang::QualType QualType, const clang::TypeLoc * TL, bool DesugarType)Linha 2297 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::GetQualifiedType(const clang::QualType & qual, const clang::TypeLoc * TL)Linha 575 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::WalkType(clang::QualType QualType, const clang::TypeLoc * TL, bool DesugarType)Linha 2479 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::GetQualifiedType(const clang::QualType & qual, const clang::TypeLoc * TL)Linha 575 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::WalkType(clang::QualType QualType, const clang::TypeLoc * TL, bool DesugarType)Linha 2683 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::GetQualifiedType(const clang::QualType & qual, const clang::TypeLoc * TL)Linha 575 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::WalkParameter(const clang::ParmVarDecl * PVD, const clang::SourceLocation & ParamStartLoc)Linha 2956 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::WalkType(clang::QualType QualType, const clang::TypeLoc * TL, bool DesugarType)Linha 2422 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::GetQualifiedType(const clang::QualType & qual, const clang::TypeLoc * TL)Linha 575 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::WalkFunction(const clang::FunctionDecl * FD, CppSharp::CppParser::AST::Function * F, bool IsDependent)Linha 3128 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::WalkFunction(const clang::FunctionDecl * FD, bool IsDependent, bool AddToNamespace)Linha 3243 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::WalkDeclaration(const clang::Decl * D)Linha 3824 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::WalkDeclarationDef(clang::Decl * D)Linha 3703 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::GetDeclarationFromFriend(clang::NamedDecl * FriendDecl)Linha 4436 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::WalkFriend(const clang::FriendDecl * FD)Linha 3352 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::WalkDeclaration(const clang::Decl * D)Linha 3930 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::WalkRecord(const clang::RecordDecl * Record, CppSharp::CppParser::AST::Class * RC)Linha 1020 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::WalkRecordCXX(const clang::CXXRecordDecl * Record, CppSharp::CppParser::AST::Class * RC)Linha 1048 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::WalkRecordCXX(const clang::CXXRecordDecl * Record)Linha 826 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::WalkDeclaration(const clang::Decl * D)Linha 3735 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::EnsureCompleteRecord(const clang::RecordDecl * Record, CppSharp::CppParser::AST::DeclarationContext * NS, CppSharp::CppParser::AST::Class * RC)Linha 743 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::GetRecord(const clang::RecordDecl * Record, bool & Process)Linha 782 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::WalkRecordCXX(const clang::CXXRecordDecl * Record)Linha 819 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::WalkDeclaration(const clang::Decl * D)Linha 3735 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::WalkDeclarationDef(clang::Decl * D)Linha 3703 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::WalkDeclaration(const clang::Decl * D)Linha 3904 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::WalkDeclarationDef(clang::Decl * D)Linha 3703 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::WalkAST()Linha 3289 C++ CppSharp.CppParser.dll!CppSharp::CppParser::Parser::ParseHeader(const std::vector<std::basic_string<char,std::char_traits<char>,std::allocator<char> >,std::allocator<std::basic_string<char,std::char_traits<char>,std::allocator<char> > > > & SourceFiles)Linha 4152 C++ CppSharp.CppParser.dll!CppSharp::CppParser::ClangParser::ParseHeader(CppSharp::CppParser::CppParserOptions * Opts)Linha 4371 C++ [Transição Gerenciado para Nativo] CppSharp.Parser.CLI.dll!CppSharp::Parser::ClangParser::ParseHeader(CppSharp::Parser::CppParserOptions^ Opts)Linha 771 C++ CppSharp.Parser.dll!CppSharp.ClangParser.ParseSourceFiles(System.Collections.Generic.IEnumerable<string> files, CppSharp.Parser.ParserOptions options)Linha 56 C# CppSharp.Parser.dll!CppSharp.ClangParser.ParseSourceFile(string file, CppSharp.Parser.ParserOptions options)Linha 43 C# CppSharp.Generator.dll!CppSharp.Driver.ParseCode()Linha 146 C# CppSharp.Generator.dll!CppSharp.ConsoleDriver.Run(CppSharp.ILibrary library)Linha 425 C# FBXSDKBindings.exe!FBXSDKBindings.Program.Main(string[] args)Linha 10 C# [Transição Nativo para Gerenciado] mscoreei.dll!00007ffa639a7b2d() Desconhecido mscoree.dll!00007ffa63a4a4cc() Desconhecido kernel32.dll!00007ffa90313034() Desconhecido ntdll.dll!00007ffa92201461() Desconhecido ` There I get the error: ` "Exception thrown: read access violation. this->Actions was 0xFFFFFFFFFFFFFFB7. ` <issue_comment>username_2: @username_0 you've most probably hit https://github.com/mono/CppSharp/issues/1123. But to be sure, could you reduce the test case? <issue_comment>username_1: Can you check if this is still an issue with master? <issue_comment>username_0: Hi @username_1 I will be back to my office in three days, so I will check your solution. Thanks! <issue_comment>username_0: Hi @username_1. The Master version have fixed the issue. There are some issues related to the FBX library code itself, but I'll mark this one as resolved. Ty!<issue_closed> <issue_comment>username_0: ##### Brief Description I'm getting an AccessViolationException when trying to generate the interop classes for [FBX SDK 2019](http://download.autodesk.com/us/fbx/2019/2019.0/fbx20190_fbxsdk_vs2015_win.exe). The error is thrown when reaching this CppSharp method: ``` CppSharp::Parser::ParserResult^ CppSharp::Parser::ClangParser::ParseHeader(CppSharp::Parser::CppParserOptions^ Opts) { auto __arg0 = (::CppSharp::CppParser::CppParserOptions*)Opts->NativePtr; auto __ret = ::CppSharp::CppParser::ClangParser::ParseHeader(__arg0); if (__ret == nullptr) return nullptr; return (__ret == nullptr) ? nullptr : gcnew CppSharp::Parser::ParserResult((::CppSharp::CppParser::ParserResult*)__ret); } ``` Specifically, this line: `auto __arg0 = (::CppSharp::CppParser::CppParserOptions*)Opts->NativePtr;` I'm looking for a way to compile LLVM with debug symbols, so I could check what's happening internally, but I can't find a LLVM debug source package on CppSharp repo, and couldn't compile LLVM debug from sources either. OS: Windows / OS X / Linux (include version and/or distro) Windows ##### Used headers `fbxsdk.h` ##### Used settings ``` var options = driver.Options; options.Verbose = true; var module = options.AddModule("FBXSDK"); module.IncludeDirs.Add("C:\\Program Files\\Autodesk\\FBX\\FBX SDK\\2019.0\\include"); module.IncludeDirs.Add("C:\\Program Files (x86)\\Microsoft Visual Studio\\2017\\Professional\\VC\\Tools\\MSVC\\14.14.26428\\include"); module.LibraryDirs.Add("C:\\Program Files\\Autodesk\\FBX\\FBX SDK\\2019.0\\lib\\vs2015\\x86\\release"); module.Libraries.Add("libfbxsdk.lib"); ``` Target: MSVC/GCC/Clang MSVC Other settings ##### Stack trace or incompilable generated code ``` C:\Program Files (x86)\Windows Kits\10\include\10.0.17134.0 Parsing libraries... Parsed 'libfbxsdk.lib' Parsing code... Parsed 'fbxsdk/core/base/fbxarray.h' Parsed 'fbxsdk/core/base/fbxbitset.h' Parsed 'fbxsdk/core/base/fbxcharptrset.h' Parsed 'fbxsdk/core/base/fbxcontainerallocators.h' Parsed 'fbxsdk/core/base/fbxdynamicarray.h' Parsed 'fbxsdk/core/base/fbxstatus.h' Parsed 'fbxsdk/core/base/fbxfile.h' Parsed 'fbxsdk/core/base/fbxhashmap.h' Parsed 'fbxsdk/core/base/fbxintrusivelist.h' Parsed 'fbxsdk/core/base/fbxmap.h' Parsed 'fbxsdk/core/base/fbxmemorypool.h' Parsed 'fbxsdk/core/base/fbxpair.h' Parsed 'fbxsdk/core/base/fbxset.h' Parsed 'fbxsdk/core/base/fbxstring.h' Parsed 'fbxsdk/core/base/fbxstringlist.h' Parsed 'fbxsdk/core/base/fbxtime.h' Parsed 'fbxsdk/core/base/fbxtimecode.h' Parsed 'fbxsdk/core/base/fbxutils.h' Parsed 'fbxsdk/core/math/fbxmath.h' Parsed 'fbxsdk/core/math/fbxdualquaternion.h' Parsed 'fbxsdk/core/math/fbxmatrix.h' Parsed 'fbxsdk/core/math/fbxquaternion.h' Parsed 'fbxsdk/core/math/fbxvector2.h' Parsed 'fbxsdk/core/math/fbxvector4.h' Parsed 'fbxsdk/core/fbxclassid.h' Parsed 'fbxsdk/core/fbxconnectionpoint.h' Parsed 'fbxsdk/core/fbxdatatypes.h' ``` <issue_comment>username_0: Hi @username_1! The latest commits fixes the issue. There are some other issues related to the FBX library itself, but it has nothing to do with the memory access exception. How can I proceed with BountySource?<issue_closed> <issue_comment>username_1: ##### Brief Description I'm getting an AccessViolationException when trying to generate the interop classes for [FBX SDK 2019](http://download.autodesk.com/us/fbx/2019/2019.0/fbx20190_fbxsdk_vs2015_win.exe). The error is thrown when reaching this CppSharp method: ``` CppSharp::Parser::ParserResult^ CppSharp::Parser::ClangParser::ParseHeader(CppSharp::Parser::CppParserOptions^ Opts) { auto __arg0 = (::CppSharp::CppParser::CppParserOptions*)Opts->NativePtr; auto __ret = ::CppSharp::CppParser::ClangParser::ParseHeader(__arg0); if (__ret == nullptr) return nullptr; return (__ret == nullptr) ? nullptr : gcnew CppSharp::Parser::ParserResult((::CppSharp::CppParser::ParserResult*)__ret); } ``` Specifically, this line: `auto __arg0 = (::CppSharp::CppParser::CppParserOptions*)Opts->NativePtr;` I'm looking for a way to compile LLVM with debug symbols, so I could check what's happening internally, but I can't find a LLVM debug source package on CppSharp repo, and couldn't compile LLVM debug from sources either. OS: Windows / OS X / Linux (include version and/or distro) Windows ##### Used headers `fbxsdk.h` ##### Used settings ``` var options = driver.Options; options.Verbose = true; var module = options.AddModule("FBXSDK"); module.IncludeDirs.Add("C:\\Program Files\\Autodesk\\FBX\\FBX SDK\\2019.0\\include"); module.IncludeDirs.Add("C:\\Program Files (x86)\\Microsoft Visual Studio\\2017\\Professional\\VC\\Tools\\MSVC\\14.14.26428\\include"); module.LibraryDirs.Add("C:\\Program Files\\Autodesk\\FBX\\FBX SDK\\2019.0\\lib\\vs2015\\x86\\release"); module.Libraries.Add("libfbxsdk.lib"); ``` Target: MSVC/GCC/Clang MSVC Other settings ##### Stack trace or incompilable generated code ``` C:\Program Files (x86)\Windows Kits\10\include\10.0.17134.0 Parsing libraries... Parsed 'libfbxsdk.lib' Parsing code... Parsed 'fbxsdk/core/base/fbxarray.h' Parsed 'fbxsdk/core/base/fbxbitset.h' Parsed 'fbxsdk/core/base/fbxcharptrset.h' Parsed 'fbxsdk/core/base/fbxcontainerallocators.h' Parsed 'fbxsdk/core/base/fbxdynamicarray.h' Parsed 'fbxsdk/core/base/fbxstatus.h' Parsed 'fbxsdk/core/base/fbxfile.h' Parsed 'fbxsdk/core/base/fbxhashmap.h' Parsed 'fbxsdk/core/base/fbxintrusivelist.h' Parsed 'fbxsdk/core/base/fbxmap.h' Parsed 'fbxsdk/core/base/fbxmemorypool.h' Parsed 'fbxsdk/core/base/fbxpair.h' Parsed 'fbxsdk/core/base/fbxset.h' Parsed 'fbxsdk/core/base/fbxstring.h' Parsed 'fbxsdk/core/base/fbxstringlist.h' Parsed 'fbxsdk/core/base/fbxtime.h' Parsed 'fbxsdk/core/base/fbxtimecode.h' Parsed 'fbxsdk/core/base/fbxutils.h' Parsed 'fbxsdk/core/math/fbxmath.h' Parsed 'fbxsdk/core/math/fbxdualquaternion.h' Parsed 'fbxsdk/core/math/fbxmatrix.h' Parsed 'fbxsdk/core/math/fbxquaternion.h' Parsed 'fbxsdk/core/math/fbxvector2.h' Parsed 'fbxsdk/core/math/fbxvector4.h' Parsed 'fbxsdk/core/fbxclassid.h' Parsed 'fbxsdk/core/fbxconnectionpoint.h' Parsed 'fbxsdk/core/fbxdatatypes.h' ``` <issue_comment>username_1: I see this on BountySource: `You will be able to claim this bounty as soon as the original issue is marked as closed.` Can you try closing it? <issue_comment>username_0: Done!<issue_closed>
{'fraction_non_alphanumeric': 0.1334128641017994, 'fraction_numerical': 0.02594691321204891, 'mean_word_length': 6.039538138558433, 'pattern_counts': {'":': 0, '<': 26, '<?xml version=': 0, '>': 33, 'https://': 3, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 1, 'word_list_counts': {'cursed_substrings.json': 5, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '11732776', 'n_tokens_mistral': 7800, 'n_tokens_neox': 7332, 'n_words': 1323}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Symfony\Component\ErrorHandler\Error\FatalError Trait 'App\Http\Controllers\Countries' not found username_0: Laravel 8 is giving me the following error: ``` Symfony\Component\ErrorHandler\Error\FatalError Trait 'App\Http\Controllers\Countries' not found ``` Here's my controller's code: ``` namespace App\Http\Controllers; use Validator; use App\Models\Customer; use Illuminate\Http\Request; use Illuminate\Support\Facades\Auth; class CustomerController extends Controller { use Countries; /** * Display a listing of the resource. * * @return \Illuminate\Http\Response */ public function index() { $customers = Customer::where('user_id', Auth::id())->paginate(9); return view('customers')->with(['customers' => $customers]); } ``` The error point to "use Countries".<issue_closed> <issue_comment>username_1: The error is right to point there. In that position ```use Countries``` is trying to load a trait called Countries, but there isn't one. If you move the line outside of the class to where your other use statements it will instead load the Countries Laravel facade which is what this package provides.
{'fraction_non_alphanumeric': 0.1016949152542373, 'fraction_numerical': 0.003228410008071025, 'mean_word_length': 3.732824427480916, 'pattern_counts': {'":': 0, '<': 4, '<?xml version=': 0, '>': 7, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '22877777', 'n_tokens_mistral': 362, 'n_tokens_neox': 352, 'n_words': 117}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Module 1.1 Assignment username_0: Given what you have learned about Git, why would you want to use Git in your classroom? <issue_comment>username_1: 👋 Hi @username_0 you are missing your answer for the question above. Please post it here in this thread. Also note that the Campus Advisor program is no longer required for administrators at GitHub Campus Program partner schools, in case you'd like to skip submitting this issue
{'fraction_non_alphanumeric': 0.03870967741935484, 'fraction_numerical': 0.010752688172043012, 'mean_word_length': 5.131578947368421, 'pattern_counts': {'":': 0, '<': 3, '<?xml version=': 0, '>': 3, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '586547', 'n_tokens_mistral': 122, 'n_tokens_neox': 110, 'n_words': 70}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Spin up Linux EC2 for testing in StreamStats Sub-Account username_0: We're looking to spin up an EC2 instance that will be used to run the StreamStats integration test. There is little expected 'load' on this instance, the test is written in Python 2.7 and has few dependencies. <issue_comment>username_1: Sure. A few questions tho... 3. Will this need any ESRI products, either Server or Desktop or both? 2. Will the server need to run 24/7, or during business hours, or only on demand? 4. How resource-intensive will the integration test be (sustained high usage, or sporadic/burst-like usage) This also might be a great opportunity to explore containerization, since this seems to be a very focused use with few dependencies. Or possibly even Lambda, but only if the test duration is short (on the order of seconds). @username_2 let's work on this together when you're next available. <issue_comment>username_2: I am available after my exam today. So 4ish. <issue_comment>username_1: Let's hold off until we get more details from the SS team. <issue_comment>username_2: This is still in holding correct? <issue_comment>username_1: Yes, until @username_0 or someone on the SS team can answer my previous questions above. <issue_comment>username_0: Wow, missed the response on this, sorry fellas. Answers: (Interesting numbering scheme) 3. No, but the tests do leverage ArcPy. 4. On demand. We plan to write a script to fire up the server, run the tests, output the results, and shut it down. 5. I'll ping Jeremy on this. <issue_comment>username_3: 3) the scripts do not leverage arcpy. they are a simply calling REST services 4) yes on demand, they could run for quite a few hrs (maybe even days) depending on the number of sites we have them runing 5) probably sustained high usage, But as I mentioned in 3) they shouldn't be resources intensive (well maybe memory). It really is just a script that will loop over a n number of sites, hits other REST endpoints, then compares the response against local values stored in files. <issue_comment>username_1: lol, just noticed the numbering. I think I had a longer list then cut it down without renumbering :) So this server could potentially run for a few days, but it would only be used occasionally? Sounds like on-demand to me. @username_3 As for instance type and sizing choices, you say that the scripts aren't resource intensive (maybe RAM), but what will user traffic look like? Will there be a lot of hits (and will they be steady or intermittent), or will there be few hits? <issue_comment>username_0: Very little no traffic. only a handful of individuals will have the ability to kick off a test. <issue_comment>username_3: right, I would say only one or 2 users that will kick off the script that then load tests the services on our test/production machines <issue_comment>username_3: maybe we just start with a smaller/medium instance, and scale if/when needed <issue_comment>username_1: Yes, this definitely sounds like a T series instance (burst-able). Thanks! <issue_comment>username_2: So here is what I did. I built this guy out with Python and bunch helpful server components. I added a ultra lightweight GUI (LXDE-core w/tightvncserver). This is to make it easier to test server tasks on a local port like you have on a windows system. It's at least as fast as a standard Win OS and we can tweak it to make it faster.<issue_closed> <issue_comment>username_2: Up and running.
{'fraction_non_alphanumeric': 0.04908675799086758, 'fraction_numerical': 0.010559360730593607, 'mean_word_length': 4.5023547880690735, 'pattern_counts': {'":': 0, '<': 17, '<?xml version=': 0, '>': 17, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '6848940', 'n_tokens_mistral': 936, 'n_tokens_neox': 889, 'n_words': 558}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: [Terrain] Choosing the collider detail level in terrain colliders. username_0: This request came from this topic: http://urho3d.prophpbb.com/topic758.html and the IRC discussion that came with it. When using a detailed enough terrain (513x513 heightmap in my case) and big colliders: Something like the collider being 1:50 to the terrain, not that big. Bullet has to process a lot of collisions making even the most simple scenes unplayable with modern hardware. I solved the issue by using an invisible terrain on top of the detailed one with a smaller 65x65 texture, but this is nothing more than a workaround for what I think it's a missing feature in terrain CollisionShape. You can check my post on the forum for a detailed description on the issue and even a video of it. I think there should be a way to choose the detail level of a terrain collider based on the mipmaps of the heightmap image. So basically the collider could be defined by a lower detail version of the terrain shape to avoid performance issues.<issue_closed>
{'fraction_non_alphanumeric': 0.030612244897959183, 'fraction_numerical': 0.016697588126159554, 'mean_word_length': 4.477157360406092, 'pattern_counts': {'":': 0, '<': 3, '<?xml version=': 0, '>': 3, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '29242640', 'n_tokens_mistral': 274, 'n_tokens_neox': 253, 'n_words': 174}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Issue with loading a custom driver username_0: I'm trying to port the opengloves driver to linux. Source code of the linux port is liste here https://github.com/username_0/opengloves-driver/tree/feat-linux However for some odd reason it won't be loaded by SteamVR instead I get an error in the log: ``` Unable to load driver shared library openglove from /home/feldi/.local/share/Steam/steamapps/common/SteamVR/drivers/openglove/bin/linux64/driver_openglove.so. Error Code (1) Sun Jul 25 2021 18:59:45.403715 - ASSERT: "DriverManager openglove load error 1" at /data/src/common/vrcommon/drivermanager.cpp:404. Sun Jul 25 2021 18:59:45.435416 - Unable to load driver openglove because of error VRInitError_Init_FileNotFound(103). Skipping. ``` When debugging the vrserver using the method in #1570 gives a sigtrap and the following stack trace: ``` (gdb) i stack -- #0 0x00005555556a99d9 in CAssert::AssertMsgImpl(char const*, unsigned int, bool, unsigned int, ...) () #1 0x00005555557ac1bf in CDriverManager::LoadDriver(unsigned int, CDriverManager::HmdDriverContext, vr::IVRDriverContext*) () #2 0x00005555556476b9 in CVRServer::LoadDriver(unsigned int, vr::EVRInitError*, int*) () #3 0x000055555564c178 in CVRServer::LoadDrivers(vr::EVRInitError*, int*) () #4 0x000055555564c42d in CVRServer::HandleEarlyLoad() () #5 0x000055555566b8a4 in RealMain(unsigned int, VRServerArguments_t&) () #6 0x00005555555d235d in main () ``` I tried the sample driver and copied into the drivers folder for SteamVR and had no issues with it. When I disabled all the sources in this portion of the cmake file https://github.com/username_0/opengloves-driver/blob/feat-linux/driver.cmake#L2-L34 and just loading the ``DeviceFactory.cpp`` while commenting out the ``DeviceProvider`` instantiation and return it will compile and load with the expected with a ``VRInitError_Init_InterfaceNotFound(105)`` error (which is value set in HmdDriverFactory when the interfaceName does not match). I don't know why this won't load, just disabling sources without changing any of the libraries in cmake will allow it to load so I don't think it's a library dependency issue. Also if I just comment out the ``DeviceProvider`` instantiation and return in ``DeviceFactory.cpp`` but keep the Source files in the cmake (so they will get compiled) will also cause SteamVR to not load the driver. So I don't know what is causing this issue.
{'fraction_non_alphanumeric': 0.08192090395480225, 'fraction_numerical': 0.06900726392251816, 'mean_word_length': 4.98792270531401, 'pattern_counts': {'":': 0, '<': 2, '<?xml version=': 0, '>': 2, 'https://': 2, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 1, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '17862949', 'n_tokens_mistral': 866, 'n_tokens_neox': 742, 'n_words': 295}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: truffle test does not work username_0: Running `truffle test` in a fresh checkout of master results in the following output. ``` Using network 'test'. Error: ENOENT: no such file or directory, stat '/home/piers/projects/truffle/cheshire/migrations' ``` If I create the migrations directory and try again it results in the following output. ``` Using network 'test'. ReferenceError: afterAll is not defined at Suite.describe (/home/piers/projects/truffle/cheshire/test/kitty.js:17:3) at Object.create (/home/piers/programs/node-v8.9.4-linux-x64/lib/node_modules/truffle/node_modules/mocha/lib/interfaces/common.js:112:19) at context.describe.context.context (/home/piers/programs/node-v8.9.4-linux-x64/lib/node_modules/truffle/node_modules/mocha/lib/interfaces/bdd.js:44:27) at Object.<anonymous> (/home/piers/projects/truffle/cheshire/test/kitty.js:6:1) at Module._compile (module.js:643:30) at Object.Module._extensions..js (module.js:654:10) at Module.load (module.js:556:32) at tryModuleLoad (module.js:499:12) at Function.Module._load (module.js:491:3) at Module.require (module.js:587:17) at require (internal/module.js:11:18) at /home/piers/programs/node-v8.9.4-linux-x64/lib/node_modules/truffle/node_modules/mocha/lib/mocha.js:231:27 at Array.forEach (<anonymous>) at Mocha.loadFiles (/home/piers/programs/node-v8.9.4-linux-x64/lib/node_modules/truffle/node_modules/mocha/lib/mocha.js:228:14) at Mocha.run (/home/piers/programs/node-v8.9.4-linux-x64/lib/node_modules/truffle/node_modules/mocha/lib/mocha.js:536:10) at /home/piers/programs/node-v8.9.4-linux-x64/lib/node_modules/truffle/build/webpack:/~/truffle-core/lib/test.js:125:1 at <anonymous> ```<issue_closed> <issue_comment>username_1: @username_0 Thank you for reporting this! There were some files/directories in the Cheshire project inconsistent with the layout expected by the `truffle` command. Pull the latest master or run `truffle unbox endless-nameless-inc/cheshire` for the fix.
{'fraction_non_alphanumeric': 0.1295238095238095, 'fraction_numerical': 0.047142857142857146, 'mean_word_length': 4.002380952380952, 'pattern_counts': {'":': 0, '<': 7, '<?xml version=': 0, '>': 7, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '7498506', 'n_tokens_mistral': 823, 'n_tokens_neox': 752, 'n_words': 141}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Change password from LDAP username_0: Currently AuthLDAP simply hides the change password option from a LDAP user profile. Would it be too hard to make AuthLDAP change the user password on LDAP? A bonus would be to implement "forgot my password" for LDAP users also. <issue_comment>username_1: Not really an easy one as there is no default way to change a users password in LDAP. Well, of course you can use ```ldap_mod_replace``` to change an attribute and a password is nothing else than an attribute. But how is the password stored in the LDAP? Encrypted I suppose, but how? And is there possibly more than one field to update using different encryptions? In all setups that I have seen so far there was a central piece of software (mostly a website) where the user could change it's password for the LDAP. We could add a config-option that holds a link to such a password-change site and display that link instead of the password-fields in the backend. But I think everything else wouldn't really work smoothly. But I'm open for suggestions there :wink: <issue_comment>username_0: I was not aware this would be hard... but see from "man ldappasswd" which is part of openldap-clients: "ldappasswd is a tool to set the password of an LDAP user. ldappasswd uses the LDAPv3 Password Modify (RFC 3062) extended operation." So it seems there is a standard password change operation. Maybe it's not supported by some products like MSAD. Maybe not supported by PHP LDAP client library. I guess it would be possible to support a few common cases like OpenLDAP and FreeIPA with inetOrgPerson / posixUserAccount, Samba3 PDC + OpenLDAP or MSAD. Those looks like are well documented. Anyway a google search for "php ldap change password" shows promising results. <issue_comment>username_2: I consider this outside the scope of the functionality provided by this plugin. Changing the LDAP password should be handled via another plugin or an external application. <issue_comment>username_1: The scope of this plugin is to authenticate against an LDAP, not to change the LDAP. That's the scope of dedicated tools like phpmyldap or gosa. Why? ```ldappasswd``` is a tool that can be used to change the password on the commandline. But from a webserver running PHP there is only a slight possibility to use the ```exec```-functions to use that commandline-tool to change the password. That might not be possible as ```exec``` is often either completely blocked or at least very restricted. And it only works together with OpenLDAP-backends whereas a great part of LDAP-Servers we authenticate against are ADs. So the only other option would be to use a function provided by the PHP-library which is only the ```ldap_mod_replace``` to change ay attribute-value. There is **no** way of directly changing the password. And last but not least, on changing the password one would have to check al the possible password-policies that can be set, different encryption/hashing-algoriithms, different attributes to store differently hashed stuff in. Thats *so* many variables that the possibility that the organization already thought of that and has an already implemented way to change a password is so high, that it doesn't make sense to add this to the plugin. As already states, adding a link to such a password-changing website would be possible and I would add that. Everything else is in my eyes out of scope. <issue_comment>username_0: Are username_2 and username_1 the same person? I understand a feature may not be a interesting for the developer, but asking won't hurt. ;-) And a little exchange of ideas can help the developer make its mind or help someone else who would be interested in implementing the feature. Yes it could be done by an external program. But the user experience would not be good, if wordpress is intended to be the "face" of the organization. Not to mention installing, integrating, configuring the external program. Some directories (ex: OpenLDAP) provide no web interface for an end-user, only CLI tools for administrators. Others (ex: FreeIPA) provide a web interface which is useful only for internal users like sysadmins, not for end-users, specially those outside the organization. So IMHO it would be nice having AuthLDAP handing this. But I respect other's opinions and I apologize if this is not the proper forum for discussing such feature viability, even if it's for the sake of someone else picking up the idea. <issue_comment>username_2: LOL! I (username_2) am just a very satisfied user of this plugin. Andreas is a plugin developer/programmer. My statement regarding your request, while short (I'm responding at work!!), should not be regarded as dismissive. I simply believe that when one uses LDAP, there are likely multiple applications which use it for authentication, and adding one more for LDAP password/profile management is not taboo. :) Adding this functionality to this plugin is no easy task and can introduce other config options (like enforcing password complexity/rules) which can make code maintenance burdensome. <issue_comment>username_1: HI @username_0 . No, me and @username_2 are not the same person :wink: And I can understand your idea very well. Having "Once face to the customer" is a great idea! But for me the scope of a plugin should do one thing but that as good as possible. This plugin tries to do authentication a user against an LDAP as good as possible. Changing a users password in LDAP is a completely different story, another scope. Therefore I'd say it's a completely different plugin **if** you want to do that in wordpress. I know Sysadmins that would go berserk on the sheer idea of that. It's a matter of security they do not want to be given out of hands. And as it's a matter of company-wide security I can only advise anyone to consult a security expert on that one. I'm not that person. Therefore I don't want to (and won't) mess with that. Admins can use and setup different GUIs to allow users to change passwords. One of the most well known ones is the change password-function on your computers user-preferences. In most enterprise environments changing the password on your machine changes the password in the LDAP and triggeres everything necessary. Alternate options would be WebInterfaces like [GOsa](https://oss.gonicus.de/labs/gosa/) or [phpLdapAdmin](http://phpldapadmin.sourceforge.net/wiki/index.php/Main_Page). A third alternative would be to create a webapplication solely for changing passwords. But I do not know one out-of-the-box solution for that. PS: There was surely no offence meant by any comments I made. Cheers! And thank you for giving me the opportunity to overthink my opinions on that matter. Very much apprechiated. <issue_comment>username_1: @username_2 Thanks for the kind words :smile: <issue_comment>username_0: Hi username_2, when you told it was outside the scope you told with such an authority it looked like an "official" statement by the developer. So far I'm also very satisfied with the plugin, it's the best I could find. :-) I'll research what self-service end-user options I can find and maybe provide feedback on this github issue. Unfortunately phpLdapAdmin and GOsa look like sysadmin tools, not something I'd expose to end users. My users are external to my organization, so the regular OS change password will no work. There'll be other web apps besides wordpress using the same LDAP directory for users, password and roles, but something has to provide those users the ability to change and reset their password... besides setting it on first access (registration). So far I found no wordpress plugin for that. :-( I really wanted to avoid extensive custom development or having to make two different apps look the same... but do not want to start wordpress without ldap, it'll be harder to integrate later (when the second web app goes live) than now! <issue_comment>username_3: This works with OpenLDAP (using MD5 hashes)... ```php function authLdap_passwordReset($user, $password) { $ldap_rdn = 'cn=admin,dc=company,dc=com'; $ldap_pw = 'adminpw'; $ldap_orgunit = 'ou=people,dc=company,dc=com'; $ldap_conn = ldap_connect("ldap://id.company.com", 389); if ($ldap_conn) { ldap_set_option($ldap_conn, LDAP_OPT_PROTOCOL_VERSION, 3); $ldap_bind = ldap_bind($ldap_conn, $ldap_rdn, $ldap_pw); if ($ldap_bind) { // Check to see if the user exists $searched = ldap_search($ldap_conn, $ldap_orgunit, 'cn=' . $user->user_login, array('cn')); if (!$searched) { authLdap_debug("User does not exist: " . $user . " in LDAP server"); ldap_close($ldap_conn); return null; } $entries = ldap_get_entries($ldap_conn, $searched); if ($entries['count'] != 1) { authLdap_debug("User does not exist, or more than 1 user with this login: " . $user . " in LDAP server"); ldap_close($ldap_conn); return null; } // Overwrite the user's password $ldap_entry = array(); $ldap_entry['userPassword'] = '{MD5}' . base64_encode(pack('H*',md5($password))); $added = ldap_modify($ldap_conn, "cn=" . $user->user_login . ',' . $ldap_orgunit, $ldap_entry); if (!$added) { authLdap_debug("Cannot reset password for user " . $user->user_login . " in LDAP server"); ldap_close($ldap_conn); return null; } } else { authLdap_debug("Cannot bind to LDAP server with admin account"); ldap_close($ldap_conn); return null; } } else { authLdap_debug("Cannot connect to LDAP server with admin account"); ldap_close($ldap_conn); return null; } ldap_close($ldap_conn); return $data; } add_action('password_reset', 'authLdap_passwordReset', 10, 2); ``` See also https://github.com/username_1/authLdap/issues/65 <issue_comment>username_1: Hi @username_3. Thanks for the contribution of a working password-reset script. But as stated earlier changing passwords is Out-Of-Scope for the authLdap-plugin. Changing passwords in an LDAP is a complex task involving password-policies, possibly multiple password fields, different password-hashing algorithms etc. It's not complex from the programming POV but from the underlying workflow POV. And that's why I won't implement it in this plugin. But I'll leave this as a reference for users that want to implement it on their own! Therefore **Thanks for your time and effort for this contribution**! <issue_comment>username_3: @username_1 no probs... hope this helps someone else. <issue_comment>username_1: I hope so too! That's why I want it to stay open :wink: I will later link it also from the README to make it accessible more easily! <issue_comment>username_4: Its a pity this is out of scope since from a user perspective it's less obvious why this should be the case. I'm not arguing against the developer logic though, including it could increase the development size many times over. @username_3 Have you considered wrapping this code into its own plugin. That way the code can be made easily available to those of us it would help. <issue_comment>username_1: @username_4 This depends on whether the wordpress-instance is the ONLY connection of the users to the LDAP. Because usually there are a lot of services that are affected by changing the password like email, computer-logins, login to different other company services… What I could think of would be to have a hook that rewrites the link to the password-change so that it redirects the user to an interface where the LDAP-Password can be changed… <issue_comment>username_4: In another world, other apps than wordpress (including nextcloud) have got round this by flagging individual users as being provided by LDAP so that non-ldap users aren't affected. If it depends the configuration of the wordpress instance then its a matter for configuration options and not scope. But that's as argumentative as I'm going to be on the subject. :-) Yes other Wordpress plugins have just given a different link like you suggest. <issue_comment>username_1: authLDAP actually flags users as LDAP-users and hides the password-change-button in the interface. Doesn't work though for the lost-password link as the user usually isn't logged in at that moment… So you can only change the "lost-password"-link for all users… <issue_comment>username_5: Hello, yes never change something in a application and hope it will work elsewhere :) The idea to have a hook to be able to send the user to the password changing interface could be a nice idea. Cheers <issue_comment>username_4: @username_5 Yes I agree *hope* is bad currency in software. Standards on the other hand... https://tools.ietf.org/html/rfc3112 <issue_comment>username_6: Hello all, I was wondering if there's been any movement on this. I'm happy to contribute to a programmer who wants to make this feature, as authLDAP is almost perfect otherwise. Even if it's technically 'adjacent', this really is an important part of this type of plugin to create an ldap system that works and is as transparent to users as possible. For our instance, we use multiple sites with openLDAP as the backend for auth, and our WP site as the main user-facing site / hub. I expect others have this configuration too. Modifying other attributes would be fantastic too tarek : ) <issue_comment>username_1: Changing the LDAP password or any other attribute inside LDAP is **no** in the scope of this plugin. This plugin merely handles authentication against an LDAP backend. Nothing more but also nothing less! Whoever wants to create a plugin that handles password-changes or merging more attribute feel free to do so and I'm more than happy to help when questions pop up. But not within this plugin. <issue_comment>username_4: I actually looked into doing just that and discovered Wordpres's own code makes it impossible because one of the pluggable functions is circumvented where it shouldn't be. I opened a ticket against the issue on wordpress. If I ever get the time I'll write a pull request that fixes it. But for now, its possible to write something which catches the lost password reset, but not the user resetting their own password in their profile config. <issue_comment>username_7: That would be really great and a lot easier to maintain than the "change password feature". You could use something like: https://github.com/ltb-project/self-service-password
{'fraction_non_alphanumeric': 0.05415186649688359, 'fraction_numerical': 0.004222237115474834, 'mean_word_length': 3.77504, 'pattern_counts': {'":': 0, '<': 25, '<?xml version=': 0, '>': 28, 'https://': 4, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '11684835', 'n_tokens_mistral': 3913, 'n_tokens_neox': 3733, 'n_words': 2125}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Prevent Filter Removal from Grid Reload on New Row Insert username_0: <!--- Please attempt to fit your report into 1 of 3 categories and include the following information: 1. Bug Reports: A) Expected Behavior B) Actual Behavior C) Steps to Reproduce 2. Feature Request: A) Desired Feature B) Justification for Addition 3. Help Request/Question: A) Clearly state your problem B) Include your code C) If Applicable, attach a JSFiddle or CodePen --> <!--- Begin Report Below This Line --> I'm using ajax calls to fetch data from the server backend. After applying filters to the table and then switching over to insert, the insert happens and then the grid reloads. When the grid reloads, it drops all of the previously applied filters. `url/api?name=jane&pageIndex=1&pageSize=15` becomes `url/api?pageIndex=1&pageSize=15` Is there a way to keep the filters applied after a row is inserted? ``` onItemInserted: function(args) { //var filter = $("#jsGrid").jsGrid("getFilter"); can't pass filters to refresh //var $this = this; didn't work to override default behavior //$this.refresh(); }, controller: { loadData: function(filter) { return $.ajax({ type: "GET", url: "/url/api", data: filter }); } } ``` I played around with overriding the default behavior but it didn't work either. ``` jsGrid.loadStrategies.PageLoadingStrategy.prototype.onItemInserted = function(args) { var grid = this._grid; grid.refresh(); } ``` <issue_comment>username_1: Are you fetching data by ajax call inside the `loadData([filter])` method or custom method? if you using loadData method then you can apply filter inside the loaadData method <issue_comment>username_0: Yes I am passing filter through loadData. For some reason when the insert command is issued the filter wasn't being passed through. So I had to do a workaround. I'm not sure if it's a bug or something in my implementation. ``` loadData: function(filter) { $.extend(filter, { count: $("#jsGrid").jsGrid("_itemsCount"), csrf_token: validCSRFToken }); return $.ajax({ type: "GET", url: "url/api", data: filter }); } ``` <issue_comment>username_1: try like this ``` loadData: function(filter) { var d = $.Deferred(); $.extend(filter, { count: $("#jsGrid").jsGrid("_itemsCount"), csrf_token: validCSRFToken }); $.ajax({ type: "GET", url: "url/api", data: filter, dataType: "json" }).done(function(result) { // client-side filtering result = $.grep(result, function(item) { return item.name=== filter.name; }); d.resolve(result); }) return d.promise(); } ``` <issue_comment>username_0: I'd make a codepen with my entire code if I could but I don't believe anything in my code is changing any of the default behavior of jsGrid. I do believe that the search call isn't preserving the filter but instead clearing it. That's why I had to explicitly tell search to preserve the filter in my altered function above.
{'fraction_non_alphanumeric': 0.09978832778953735, 'fraction_numerical': 0.004838221953432114, 'mean_word_length': 2.69608938547486, 'pattern_counts': {'":': 0, '<': 8, '<?xml version=': 0, '>': 8, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '13770452', 'n_tokens_mistral': 981, 'n_tokens_neox': 909, 'n_words': 368}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Auto ml username_0: a few outstanding issues: 1) tests that create models take really long, and I have .skip in front of those tests currently, please advise! 2) Running predict and list with pre-existing model and dataset 3) issues with docs- I'm not sure how to generate docs as the new files are in an automl folder, please advise! 4) notes: deployment should work as normal- npm link ../ in samples folder- may need to npm install in automl folder as well if AutomlDataset/Predict/Model files cannot be called. Fixes #<issue_number_goes_here> (it's a good idea to open an issue first for discussion) - [ ] Tests and linter pass - [ ] Code coverage does not decrease (if any source code was changed) - [ ] Appropriate docs were updated (if necessary) <issue_comment>username_1: /cc @username_2 – from Chris: _"tests that create models take really long, and I have .skip in front of those tests currently, please advise!"_ How did you test these / how do you recommend testing them? If you have chance, could you TAL <issue_comment>username_1: /cc @ace-n @jmdobry – from Chris: _"issues with docs- I'm not sure how to generate docs as the new files are in an automl folder, please advise!"_ Cloud Vision has some sub-products, namely Product Search ([wip PR](https://github.com/googleapis/nodejs-vision/pull/136)) and AutoML (this PR). We'd like to put these into `./samples/automl/` and `./samples/product-search` respectively. Is there a way to configure repo-tools (`.cloud-repo-tools.json`) so that READMEs will be generated in each of these dirs (by calling scripts we list out) If y'all have any guidance, it would be much appreciated! <issue_comment>username_2: please skip tests that create models. they are flaky and long. <issue_comment>username_0: dumb question: I don't see my tests (or any tests from the samples folder) in any of the circleCI details, but i see them when I run npm test from samples folder- any idea why these aren't showing up in details tab for circleCI? Also there's a number of unhandled application warnings in details of circleCI tests that I want to dig into.. Had similiar issue not seeing samples test at automl/natural language here: https://github.com/googleapis/nodejs-language/pull/100 <issue_comment>username_1: Please add a link to the canonical samples to the PR description, thanks! I am trying to understand the unique usage of "`projectId`" in this PR and I need to see the canonical versions for comparison. It's not possible to perform sample reviews without the canonical samples to compare to. <issue_comment>username_1: **Closing** – I will re-send from an `automl` branch on this repo _(so we can edit it; this PR sent from a user's personal fork)_
{'fraction_non_alphanumeric': 0.0664259927797834, 'fraction_numerical': 0.006498194945848376, 'mean_word_length': 4.3494208494208495, 'pattern_counts': {'":': 0, '<': 9, '<?xml version=': 0, '>': 9, 'https://': 2, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '13108702', 'n_tokens_mistral': 760, 'n_tokens_neox': 724, 'n_words': 418}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Run COCO Dataset in e2e_faster_rcnn_R-101-FPN_1x.yaml shows an error username_0: I used e2e_faster_rcnn_R-101-FPN_1x.yaml to run a model with class of 2(person and background) in coco, I only modified NUM_CLASS & NUM_GPU(=1), but I met a problem shows up. What can I do to fix this? ![screenshot from 2018-03-28 10-48-38](https://user-images.githubusercontent.com/28921771/38006011-dbe7751c-3275-11e8-94e3-fed3ea1bc89e.png) ![screenshot from 2018-03-28 10-47-50](https://user-images.githubusercontent.com/28921771/38006019-e359bc92-3275-11e8-80ea-2d8af8abfe08.png)<issue_closed> <issue_comment>username_1: Hi @username_0, the shapes of the output blobs for the classification ('cls_score') and the bounding box ('bbox_pred') heads are class-dependent. Hence, to use a pre-trained Faster R-CNN model with a different number of classes you need to address this (e.g. when loading the weights). Alternatively, you could consider using one of the pretrained keypoint models (w/o the keypoint head) as a person detector. See also #156 for how to extract the per-category detection results.
{'fraction_non_alphanumeric': 0.0898576512455516, 'fraction_numerical': 0.10231316725978648, 'mean_word_length': 5.081081081081081, 'pattern_counts': {'":': 0, '<': 4, '<?xml version=': 0, '>': 4, 'https://': 2, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '25627368', 'n_tokens_mistral': 449, 'n_tokens_neox': 377, 'n_words': 127}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: chore: Run GFPGAN in docker container username_0: **Why we need it:** * Maintainers can develop in a container * Potential users can re-train GFPGAN in a containerized environment * Others can try it in an isolated environment by pulling the image and running a container only **Issue:** https://github.com/TencentARC/GFPGAN/issues/102 **Remarks for your reviewer:** I used my personal dockerhub account to store the docker image. After approving and before merging GFPGAN project can create a free dockerhub account and use it instead. <issue_comment>username_1: hey @username_0, I can't verify this MR... The build in your project is failing and I can't run the examples... ```console GFPGAN_1 | File "/usr/local/lib/python3.8/dist-packages/torch/utils/cpp_extension.py", line 1436, in _write_ninja_file_and_build_library GFPGAN_1 | _write_ninja_file_to_build_library( GFPGAN_1 | File "/usr/local/lib/python3.8/dist-packages/torch/utils/cpp_extension.py", line 1834, in _write_ninja_file_to_build_library GFPGAN_1 | cuda_flags = common_cflags + COMMON_NVCC_FLAGS + _get_cuda_arch_flags() GFPGAN_1 | File "/usr/local/lib/python3.8/dist-packages/torch/utils/cpp_extension.py", line 1606, in _get_cuda_arch_flags GFPGAN_1 | arch_list[-1] += '+PTX' GFPGAN_1 | IndexError: list index out of range gfpgan_GFPGAN_1 exited with code 1 ``` <issue_comment>username_1: # Running Docker-compose Build  And Original Image * Same error while running in a regular machine... ```console $ docker run -ti -v $PWD/inputs:/app/inputs -v $PWD/results:/app/results -v $PWD/experiments:/app/exps mostafaelmenbawy/gfpgan:latest python3 inference_gfpgan.py --model_path /app/exps/GFPGANv1.pth --test_path /app/inputs/whole_imgs --save_root /apps/results --arch original --channel 1 No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' Traceback (most recent call last): File "inference_gfpgan.py", line 7, in <module> from basicsr.utils import imwrite File "/usr/local/lib/python3.6/dist-packages/basicsr/__init__.py", line 3, in <module> from .archs import * File "/usr/local/lib/python3.6/dist-packages/basicsr/archs/__init__.py", line 16, in <module> _arch_modules = [importlib.import_module(f'basicsr.archs.{file_name}') for file_name in arch_filenames] File "/usr/local/lib/python3.6/dist-packages/basicsr/archs/__init__.py", line 16, in <listcomp> _arch_modules = [importlib.import_module(f'basicsr.archs.{file_name}') for file_name in arch_filenames] File "/usr/lib/python3.6/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "/usr/local/lib/python3.6/dist-packages/basicsr/archs/rrdbnet_arch.py", line 6, in <module> from .arch_util import default_init_weights, make_layer, pixel_unshuffle File "/usr/local/lib/python3.6/dist-packages/basicsr/archs/arch_util.py", line 13, in <module> from basicsr.ops.dcn import ModulatedDeformConvPack, modulated_deform_conv File "/usr/local/lib/python3.6/dist-packages/basicsr/ops/dcn/__init__.py", line 1, in <module> from .deform_conv import (DeformConv, DeformConvPack, ModulatedDeformConv, ModulatedDeformConvPack, deform_conv, File "/usr/local/lib/python3.6/dist-packages/basicsr/ops/dcn/deform_conv.py", line 19, in <module> os.path.join(module_path, 'src', 'deform_conv_cuda_kernel.cu'), File "/usr/local/lib/python3.6/dist-packages/torch/utils/cpp_extension.py", line 1136, in load keep_intermediates=keep_intermediates) File "/usr/local/lib/python3.6/dist-packages/torch/utils/cpp_extension.py", line 1347, in _jit_compile is_standalone=is_standalone) File "/usr/local/lib/python3.6/dist-packages/torch/utils/cpp_extension.py", line 1445, in _write_ninja_file_and_build_library is_standalone=is_standalone) File "/usr/local/lib/python3.6/dist-packages/torch/utils/cpp_extension.py", line 1834, in _write_ninja_file_to_build_library cuda_flags = common_cflags + COMMON_NVCC_FLAGS + _get_cuda_arch_flags() File "/usr/local/lib/python3.6/dist-packages/torch/utils/cpp_extension.py", line 1606, in _get_cuda_arch_flags arch_list[-1] += '+PTX' IndexError: list index out of range ``` <issue_comment>username_0: I fixed the pipeline and the error. The problem with the error was that the docker image was meant to run on GPUs only that's what I used the `BASICSR_JIT=True` env variable during build time. Now I removed it during building the image to give the freedom of running the image on CPU or on GPU by adding the flag again during run time as described in the README.md
{'fraction_non_alphanumeric': 0.10497119692767229, 'fraction_numerical': 0.021548965222957114, 'mean_word_length': 4.425925925925926, 'pattern_counts': {'":': 0, '<': 13, '<?xml version=': 0, '>': 13, 'https://': 1, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '21738297', 'n_tokens_mistral': 1719, 'n_tokens_neox': 1608, 'n_words': 407}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Install nodejs/npm from nodesource username_0: Close #316 <issue_comment>username_1: nice! I'm gonna restart the failing travis build because it was a timeout error...let's see how that looks. <issue_comment>username_1: travis is happy so I'm gonna merge this in 12 hours unless anybody disagrees! <issue_comment>username_1: woot, thanks @username_0 !!
{'fraction_non_alphanumeric': 0.07692307692307693, 'fraction_numerical': 0.02564102564102564, 'mean_word_length': 6.109090909090909, 'pattern_counts': {'":': 0, '<': 5, '<?xml version=': 0, '>': 5, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '29971362', 'n_tokens_mistral': 122, 'n_tokens_neox': 111, 'n_words': 48}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: BTable select event with action buttons username_0: Latest Buefy, OSX, Chrome I am using the select event to navigate to a record. I also have an action button in the row. When clicking the action button, both, the rowClick and delete is called **b-table** ``` <b-table @select="rowClick"> <b-table-column custom-key="actions" class="has-text-right"> <button @click="delete(props.row.id, props.index)">Delete</button> </b-table-column> </b-table> ``` **methods** ``` /** * table row selected event */ rowClick(item,event) { this.$router.push({ path: '/processes/' + item.id }) }, /** * delete action */ delete(id, index) { } ``` What happens - both are called What I expect - rowClick not to be called when clicking in a button <issue_comment>username_1: @username_0 Did you try `stop` modifier to button click event ? For example: `<button @click.stop="delete(props.row.id, props.index)">Delete</button>`<issue_closed>
{'fraction_non_alphanumeric': 0.13008130081300814, 'fraction_numerical': 0.0027100271002710027, 'mean_word_length': 2.585760517799353, 'pattern_counts': {'":': 0, '<': 12, '<?xml version=': 0, '>': 12, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '1650513', 'n_tokens_mistral': 350, 'n_tokens_neox': 325, 'n_words': 103}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Clear a Users Warns username_0: **What is the title of your proposal?** Clear a Users Warns **Is it a suggestion for the website or for the bot?** ✔️ Bot **Explain your proposal more accurately (It's best to give as much information as possible, so that we can implement the proposal better)** Be able to clear a users warnings with a command. It can be done either by the number of the warn or clear all warns. **Why should we add this feature?** It'll make a complete warn system. <issue_comment>username_1: Related to #31
{'fraction_non_alphanumeric': 0.06782608695652174, 'fraction_numerical': 0.006956521739130435, 'mean_word_length': 4.142857142857143, 'pattern_counts': {'":': 0, '<': 3, '<?xml version=': 0, '>': 3, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '691771', 'n_tokens_mistral': 168, 'n_tokens_neox': 157, 'n_words': 93}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Improved usage wiki username_0: Hi, can you please improve the usage wiki by adding high resolution images? They're too small to read. Is there any other info source for usage? Can't find any. Can't wait to use it in my projects! <issue_comment>username_1: You can open them in a new tab view in full resolution ;) However, I definitely need documentation refactoring, thanks <issue_comment>username_2: Could you also do one for Batch requests? [https://parse.com/docs/rest/guide/#objects-batch](https://parse.com/docs/rest/guide/#objects-batch) <issue_comment>username_3: Yeah, I definitely find the usage wiki to be out of sync with the plugin. I can't reproduce any of the examples in the wiki. The diagrams show elements that don't appear to exist in the blueprint graph editor. I can't find the SET, Bind Event to OnRequestComplete, Bind Event to OnRequestFail, Process URL, etc. Hopefully I'm just missing some obvious detail, but at this point I'm stuck. <issue_comment>username_4: @username_3 I'm a bit late to the party but for others with the same problem. You won't be able to get those unless: A) You are pulling your pin off a JSON Request Variable (pure reference will not work) ![image](https://cloud.githubusercontent.com/assets/10377276/22813102/d7dfd64c-eefd-11e6-8d6d-df4afce8feaf.png) B) You deselect context sensitive in the top right of the node search box ![image](https://cloud.githubusercontent.com/assets/10377276/22813021/5142b44c-eefd-11e6-9dd3-1d1e38f957ee.png)<issue_closed>
{'fraction_non_alphanumeric': 0.07741522712731926, 'fraction_numerical': 0.04222648752399232, 'mean_word_length': 4.75, 'pattern_counts': {'":': 0, '<': 7, '<?xml version=': 0, '>': 7, 'https://': 4, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '16185411', 'n_tokens_mistral': 515, 'n_tokens_neox': 460, 'n_words': 197}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: [ta edition / blaster] Image preview does not work username_0: It works for png on imgur, but strange not displaying for jpg or for another reasons... browser cache cleaning not helps. need to check and fix that issue. ![2021-05-23_00-20](https://user-images.githubusercontent.com/76843085/119241145-0e3a8180-bb5d-11eb-989f-f45d79cb2ee8.png) ![2021-05-23_00-21](https://user-images.githubusercontent.com/76843085/119241146-0ed31800-bb5d-11eb-8042-b88ce0f915bd.png) ![2021-05-23_00-24](https://user-images.githubusercontent.com/76843085/119241196-58bbfe00-bb5d-11eb-963f-c7f61950e22f.png) <issue_comment>username_0: The issue related to the Chromium security policy. In that example, remote image has http protocol, and the node running under the https. So FF thrown a warning notice and allow to receive remote content by unsecured connection, but Chromium decline that request by error. Solution is available, but I'm not sure that fix should be included to the twister-html, for user's security reasons.<issue_closed>
{'fraction_non_alphanumeric': 0.09193245778611632, 'fraction_numerical': 0.1350844277673546, 'mean_word_length': 4.99438202247191, 'pattern_counts': {'":': 0, '<': 4, '<?xml version=': 0, '>': 4, 'https://': 3, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '27067402', 'n_tokens_mistral': 430, 'n_tokens_neox': 349, 'n_words': 106}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Last row keeps refreshing username_0: I don't want to add extra dummy lines at the end of the data table, but in this case, the last row keeps refreshing. How can I resolve this? Thank in advance. <issue_comment>username_0: ![image](https://user-images.githubusercontent.com/59682514/72232185-6b820280-35fa-11ea-9cd4-d722a631aa82.png) Add a screenshot to display the issue. <issue_comment>username_1: That is if you are using the estimate row count. You can set that to false. I am open to PRs too. <issue_comment>username_0: @username_1 Are you referring to the rowCountApproximate property? I already set it as false, but the result is the same. <issue_comment>username_1: Weird, I’ll have to see what’s causing it though. How many total items do you have? And how many per page? <issue_comment>username_0: I have 12 items in total and 10 per page. <issue_comment>username_0: Hi @username_1, any update? <issue_comment>username_1: Try with the flutter what and see if it is still an issue. If it is then it is the underling material widget. https://youtu.be/ktTajqbhIcY
{'fraction_non_alphanumeric': 0.06955736224028906, 'fraction_numerical': 0.04607046070460705, 'mean_word_length': 5.054644808743169, 'pattern_counts': {'":': 0, '<': 9, '<?xml version=': 0, '>': 9, 'https://': 2, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '11779869', 'n_tokens_mistral': 363, 'n_tokens_neox': 336, 'n_words': 154}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: fix #116 by replacing :touch with :create action username_0: These changes were introduced by #91 (among others). I've split that PR into smaller chunks in hope for easier merge. <issue_comment>username_1: Really :+1: for this! The failing tests seem to be a common problem of all PRs of late, `ChefSpec::Error::CommandNotStubbed:`, not a problem with this specific change. <issue_comment>username_2: I can take a look into the failed tests separately. Can we remove the action statement since the file resource uses create by default? It's pretty confusing looking at it with the guard in there. <issue_comment>username_2: Please rebase! <issue_comment>username_0: #116 has been fixed by #121, so I think this pull request can be closed.
{'fraction_non_alphanumeric': 0.06451612903225806, 'fraction_numerical': 0.02193548387096774, 'mean_word_length': 5.466666666666667, 'pattern_counts': {'":': 0, '<': 6, '<?xml version=': 0, '>': 6, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '13711742', 'n_tokens_mistral': 213, 'n_tokens_neox': 202, 'n_words': 112}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: can't take effect with aliyun rocketmq username_0: When i try to communicate with aliyun rocketmq using rocketmq-client-go sdk, i always get errors: ``` WARN[0000] query topic route from server error underlayError="topic not exist" WARN[0000] queryTopicRouteInfoFromServer return nil topic=saas-stat-test WARN[0000] query topic route from server error underlayError="topic not exist" WARN[0000] queryTopicRouteInfoFromServer return nil topic=saas-stat-test WARN[0000] query topic route from server error underlayError="topic not exist" WARN[0000] queryTopicRouteInfoFromServer return nil topic=saas-stat-test WARN[0000] query topic route from server error underlayError="topic not exist" WARN[0000] queryTopicRouteInfoFromServer return nil topic=saas-stat-test WARN[0000] query topic route from server error underlayError="topic not exist" WARN[0000] queryTopicRouteInfoFromServer return nil topic=saas-stat-test WARN[0000] query topic route from server error underlayError="topic not exist" WARN[0000] queryTopicRouteInfoFromServer return nil topic=saas-stat-test send message error: the topic=saas-stat-test route info not found ``` Following is my test code snippets, I'm sure all the configs such as endpoints,top and so on are correct. So what's the problem? ``` package main import ( "context" "fmt" "os" rocketmq "github.com/apache/rocketmq-client-go/v2" "github.com/apache/rocketmq-client-go/v2/primitive" "github.com/apache/rocketmq-client-go/v2/producer" ) func main() { p, err := rocketmq.NewProducer( producer.WithNameServer([]string{"my correct endpoint"}), producer.WithRetry(2), producer.WithCredentials(primitive.Credentials{ AccessKey: "my correct key", SecretKey: "my correct secret", }), ) if err != nil { fmt.Println("init producer error: " + err.Error()) os.Exit(0) } err = p.Start() if err != nil { fmt.Printf("start producer error: %s", err.Error()) os.Exit(1) } for i := 0; i < 1; i++ { res, err := p.SendSync( context.Background(), primitive.NewMessage( "saas-stat-test", []byte("Hello RocketMQ Go Client!"), ), ) if err != nil { fmt.Printf("send message error: %s\n", err) } else { fmt.Printf("send message success: result=%s\n", res.String()) } } err = p.Shutdown() if err != nil { fmt.Printf("shutdown producer error: %s", err.Error()) } } ``` <issue_comment>username_1: @username_3 Is there aliyun rocketmq test env for our dev, I thought I should check that way <issue_comment>username_1: @username_0 Is that topic exist? <issue_comment>username_0: Absolutely. <issue_comment>username_2: @username_0 There is some advice for you may useful: 1. Can the network of your go-client instance establish with Aliyun RocketMQ instance? 2. Is AK/SK correct? <issue_comment>username_0: @username_2 I'm sure the network works and all mq configs are correct. <issue_comment>username_2: use this option when create producer/consumer instance: https://github.com/apache/rocketmq-client-go/blob/6b7f96c1a217e2ce9915d2fbbbdcb5c527a05f81/consumer/option.go#L187, the argument is '实例ID' in the pic: ![image](https://user-images.githubusercontent.com/9766821/81780180-afddca80-9528-11ea-8b79-713dac33b602.png) refer from: https://help.aliyun.com/document_detail/146599.html?spm=a2c4g.11186623.6.544.1cd777fcnOrpPh <issue_comment>username_3: package main import ( "context" "fmt" "github.com/apache/rocketmq-client-go/v2" "github.com/apache/rocketmq-client-go/v2/primitive" "github.com/apache/rocketmq-client-go/v2/producer" "math/rand" "os" "time" ) func main() { var endPoint = []string{"http://MQ_INST_XXXXXXX.mq-internet-access.mq-internet.aliyuncs.com:80"} var credentials = primitive.Credentials{ AccessKey: “AK”, SecretKey: “SK”, } p, err := rocketmq.NewProducer( producer.WithNameServer(endPoint), //producer.WithNsResovler(primitive.NewPassthroughResolver(endPoint)), producer.WithGroupName("GID_XXXXXXXXX”), **producer.WithNamespace("MQ_INST_XXXXXXXX”),** producer.WithCredentials(credentials), producer.WithTrace(&primitive.TraceConfig{ GroupName: "GID_XXXXXXX”, Access: primitive.Cloud, NamesrvAddrs: endPoint, Credentials: credentials, }), ) if err != nil { fmt.Println("init producer error: " + err.Error()) os.Exit(0) } err = p.Start() if err != nil { fmt.Printf("start producer error: %s", err.Error()) os.Exit(1) } for i := 0; i < 3; i++ { res, err := p.SendSync(context.Background(), primitive.NewMessage(“XXXXXXXXXX”, []byte("Hello RocketMQ Go Client!"))) if err != nil { fmt.Printf("send message error: %s\n", err) } else { fmt.Printf("send message success: result=%s\n", res.String()) } time.Sleep(200 * time.Millisecond) } time.Sleep(1000 * time.Millisecond) err = p.Shutdown() if err != nil { fmt.Printf("shutdown producer error: %s", err.Error()) } } func getRandomString(l int) string { str := "0123456789abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ" bytes := []byte(str) result := []byte{} r := rand.New(rand.NewSource(time.Now().UnixNano())) for i := 0; i < l; i++ { result = append(result, bytes[r.Intn(len(bytes))]) } return string(result) } <issue_comment>username_0: @username_3 Cool, it works. It can produce and consume.👍👍👍<issue_closed> <issue_comment>username_4: @username_3 just solved this problem, but anyone explanation why this work? what's the meaning for this option "Namespace". thanks
{'fraction_non_alphanumeric': 0.12539404553415062, 'fraction_numerical': 0.03222416812609457, 'mean_word_length': 3.914802065404475, 'pattern_counts': {'":': 0, '<': 15, '<?xml version=': 0, '>': 12, 'https://': 3, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 5, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '24358846', 'n_tokens_mistral': 2164, 'n_tokens_neox': 1955, 'n_words': 473}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: [Consignments] Sets the default option when you first select a category username_0: Fixes https://github.com/artsy/collector-experience/issues/894 When you start selecting the categories, painting is now set in the state without you doing anything as the onselect trigger on happens when you change selection. <issue_comment>username_1: 👍 awesome thanks!
{'fraction_non_alphanumeric': 0.061224489795918366, 'fraction_numerical': 0.012755102040816327, 'mean_word_length': 5.775862068965517, 'pattern_counts': {'":': 0, '<': 3, '<?xml version=': 0, '>': 3, 'https://': 1, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '19856016', 'n_tokens_mistral': 103, 'n_tokens_neox': 98, 'n_words': 46}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Frame drops in eCAL Recorder username_0: **Describe the bug** From time to time, frame drops in Recorder can occur. This seems to coincide with hdf5 file splits. ![image](https://user-images.githubusercontent.com/49187426/148375609-f57ef026-43b1-4f4a-9d30-36bae4cd1ca7.png) The image shows per channel all consecutive frame drops. Different **To Reproduce** Create multiple measurements, with split size, and tasks which are publishing at high frequency. **Expected behavior** Ideally, a measurement should not have any message drops **Additional context** We are quite positive, that aquiring the lock in the callback https://github.com/continental/ecal/blob/dc8ae834abee4619d3454246c23f0d1980e75ba8/app/rec/rec_client_core/src/ecal_rec_impl.cpp#L711 can be "expensive", because https://github.com/continental/ecal/blob/dc8ae834abee4619d3454246c23f0d1980e75ba8/app/rec/rec_client_core/src/ecal_rec_impl.cpp#L762-L766 can potentially be blocking until a new file is opened and all messages received in the meantime will be lost. It migth be helpfule to use a `shared` mutex to the `recorder_state` for all functions not modifiying the recorders state.<issue_closed>
{'fraction_non_alphanumeric': 0.07510204081632653, 'fraction_numerical': 0.0783673469387755, 'mean_word_length': 5.255102040816326, 'pattern_counts': {'":': 0, '<': 3, '<?xml version=': 0, '>': 3, 'https://': 3, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '310184', 'n_tokens_mistral': 448, 'n_tokens_neox': 389, 'n_words': 125}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Add IM/Ember Read/Writer Interface and attribute read for IM username_0: Summary of Changes: -- Add initial IM read attribute implementation and ember read/write interface so that read client can send IM read request with attribute path, interaction model engine receive this request and ask reporting engine generate the corresponding interested attributes and send it back via read hander. -- Add unit and cirque integration test
{'fraction_non_alphanumeric': 0.03368421052631579, 'fraction_numerical': 0.002105263157894737, 'mean_word_length': 4.734939759036145, 'pattern_counts': {'":': 0, '<': 2, '<?xml version=': 0, '>': 2, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '30526954', 'n_tokens_mistral': 109, 'n_tokens_neox': 107, 'n_words': 64}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Performance regression for GET requests in 0.3+ username_0: In the version 0.3 some CTEs were introduced to simplify the query structure. But this generates a barrier for the PostgreSQL optimizer and prevents some nodes from collapsing in the query plan tree. To demonstrate I'll use a query that can be generated using the postgrest_test database. First we need some test data for the problems only manifest in large relations: ```sql insert into test.simple_pk select s::text, 'testing for k = ' || s from generate_series(1, 100000) s; ``` Then we generate a query with an ORDER BY: ```sql WITH pg_source AS ( SELECT test.simple_pk.* FROM test.simple_pk ORDER BY k ) SELECT (SELECT pg_catalog.count(1) FROM (SELECT * FROM pg_source) a ), pg_catalog.count(t), null, array_to_json(array_agg(row_to_json(t)))::character varying FROM (SELECT * FROM pg_source LIMIT 1 OFFSET 0) t ``` The above query is generated by a GET request using some filters and a limit 1. The generated plan (with a total cost of 7116 blks in my test env) is: <img width="525" alt="screen shot 2015-12-08 at 5 49 33 pm" src="https://cloud.githubusercontent.com/assets/20662/11671149/2619f4ae-9dd4-11e5-8d86-5141a7d48bbd.png"> --- ## Possible solution After trying several different formats and different queries, the best compromise for all different scenarios was the query bellow: ```sql SELECT (SELECT pg_catalog.count(1) FROM ( SELECT test.simple_pk.* FROM test.simple_pk) a ), pg_catalog.count(t), null, array_to_json(array_agg(row_to_json(t)))::character varying FROM ( SELECT test.simple_pk.* FROM test.simple_pk ORDER BY k LIMIT 1 OFFSET 0) t; ``` This query does not use CTEs and also drops the **ORDER BY** clause in the subquery used for counting. The generated plan is: <img width="538" alt="screen shot 2015-12-08 at 5 56 34 pm" src="https://cloud.githubusercontent.com/assets/20662/11671266/0bbfabac-9dd5-11e5-9c5f-2559c73b5b99.png"> The cost is 1985 in my test env. Much cheaper than the original. Even if alternatives with CTEs can perform quite well in this test environment, when I take the queries to my production database all hell breaks loose unless I use the non-CTE version. I'm going to post another comment here with tests for more complex queries. <issue_comment>username_1: Any way to write a test for this? <issue_comment>username_0: @username_1 I was still investigating, and I found another solution, so we can keep the CTEs! I isolated the problem a bit further, and it's connected with the cost of counting records in the CTE. Using the same example above I've created another query: ```sql WITH pg_source AS ( SELECT test.simple_pk.* FROM test.simple_pk ORDER BY k ), pg_source_count AS ( SELECT count(1) FROM test.simple_pk ) SELECT (SELECT count(1) FROM pg_source_count), pg_catalog.count(t), null, array_to_json(array_agg(row_to_json(t)))::character varying FROM (SELECT * FROM pg_source LIMIT 1 OFFSET 0) t ``` The generated plan is: <img width="526" alt="screen shot 2015-12-08 at 9 34 40 pm" src="https://cloud.githubusercontent.com/assets/20662/11674947/8eb2eb0a-9df3-11e5-9a82-008ed15639f2.png"> The estimated cost in my test env was 6841 blks, but the speed was better than all alternatives. Besides, this query structure works vowed well for my production environment (I'm always skeptical of test data). It seems that this solution gives us the best of both worlds keeping the CTEs and the speed for the price of having an extra CTE for counting. Tomorrow I'll try to create a spec for this. <issue_comment>username_1: Since there is still a need for a separate "count" query, with or without cte is all the same from the code point of view (for the "GET" request). So maybe investigate it a bit more to make a decision which way to go on this. As for the insert/update requests, probably for now it's better to keep it as it is and in most cases the inserted records will be less then 1k so there will not be much of the impact of having it in a single cte <issue_comment>username_0: So I've made some more tests with the same query to check how the cost and execution time changes according to the number of row in the source. I have all steps documented but to spare you guys of details I'm posting just my charted results. First the estimated cost: <img width="587" alt="screen shot 2015-12-09 at 5 24 01 pm" src="https://cloud.githubusercontent.com/assets/20662/11700762/bafce2e6-9e99-11e5-9702-f1667c64457e.png"> And the time: <img width="595" alt="screen shot 2015-12-09 at 5 24 07 pm" src="https://cloud.githubusercontent.com/assets/20662/11700765/c3324532-9e99-11e5-985d-b060ec0a479c.png"> The disparity between estimated cost persists even when I change server, filtering, ordering, use nested data or set a different random page cost. The difference is too big for the estimate to be accurate. My guess is that the CTEs are interfering with the estimates. I honestly believe the two queries are equivalent in terms of execution cost. <issue_comment>username_0: Oh, regarding the spec, I gave up on it. It would be too slow to create all the needed test data. But I have a script that I've used for the tests above and for the test with nested columns, so I can send you these if needed. <issue_comment>username_1: no need right now, i think i will go with inline, it has the same speed and it has also the better estimation, why risk it :) <issue_comment>username_0: :beers: Just an example of a nested query using the inline method: ```sql SELECT (SELECT pg_catalog.count(1) FROM (SELECT test.clients.id, ( SELECT array_to_json(array_agg(row_to_json(p1))) FROM ( SELECT test.projects.id, ( SELECT array_to_json(array_agg(row_to_json(tasks))) FROM (SELECT test.tasks.id, test.tasks.name FROM test.tasks WHERE test.tasks.project_id = test.projects.id AND test.tasks.name LIKE 'Design%'::UNKNOWN) tasks ) AS tasks FROM test.projects WHERE test.projects.client_id = test.clients.id) p1 ) AS projects FROM test.clients) s), pg_catalog.count(t), NULL, array_to_json(array_agg(row_to_json(t)))::character varying FROM ( SELECT test.clients.id, ( SELECT array_to_json(array_agg(row_to_json(projects))) FROM ( SELECT test.projects.id, ( SELECT array_to_json(array_agg(row_to_json(tasks))) FROM (SELECT test.tasks.id, test.tasks.name FROM test.tasks WHERE test.tasks.project_id = test.projects.id AND test.tasks.name LIKE 'Design%'::UNKNOWN) tasks ) AS tasks FROM test.projects WHERE test.projects.client_id = test.clients.id) projects ) AS projects FROM test.clients ORDER BY test.clients.id ASC LIMIT 10 OFFSET 0) t ``` <issue_comment>username_1: The count query will be much simpler since it does not need to go deep and embed related items. <issue_comment>username_1: @username_0 who are you creating all those nice graphs and how are you progressively generating test data? Can you do the same for related tables (clients>projects>tasks) and test the performance of the query generated by ```/projects?select=id,clients{id},tasks{id}&id=in.1,100,400,22``` i am interested to see the impact of the number of rows (in all 3 tables) on the query that extracts the data and embeds related tables PS: the #414 is ready for merging i think<issue_closed>
{'fraction_non_alphanumeric': 0.07164711881693014, 'fraction_numerical': 0.03977562468128506, 'mean_word_length': 3.2822052401746724, 'pattern_counts': {'":': 0, '<': 17, '<?xml version=': 0, '>': 19, 'https://': 5, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '11186765', 'n_tokens_mistral': 2575, 'n_tokens_neox': 2300, 'n_words': 980}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: sqlalchemy connection expiry and session handling bugs username_0: sessions are not being managed cleanly, likely due to the connection-validation query on start-up. also, the mysql connections expire server-side after about 6 hours and so they need to be cycled in the connection pool of the app.
{'fraction_non_alphanumeric': 0.03880597014925373, 'fraction_numerical': 0.005970149253731343, 'mean_word_length': 5.109090909090909, 'pattern_counts': {'":': 0, '<': 2, '<?xml version=': 0, '>': 2, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '30097766', 'n_tokens_mistral': 84, 'n_tokens_neox': 78, 'n_words': 47}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Provide a default service username_0: `service` (which maps to `kSecAttrService`) isn't relevant to many developers, who store only one username/password pair, for example. I propose giving `service` a default value of the bundle identifier when unspecified: var service = NSBundle.mainBundle().infoDictionary![kCFBundleIdentifierKey] as String This would make the interface to Locksmith even cleaner. A few caveats, so don't merge it in right away: - The project won't build (`info.plist` is missing) - I haven't tested this out yet - There seem to be two copies of `Locksmith.swift` and `LocksmithRequest.swift` (one pair in `Locksmith/` and one in `Pod/Classes/`) Just opening this PR for initial code review and thoughts. <issue_comment>username_1: This sounds like a great idea. As you mentioned, I have a couple of things to clean up, then I'll merge it into the 1.1.1 branch. Thanks! >
{'fraction_non_alphanumeric': 0.07291666666666667, 'fraction_numerical': 0.005208333333333333, 'mean_word_length': 4.13903743315508, 'pattern_counts': {'":': 0, '<': 3, '<?xml version=': 0, '>': 4, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '17350078', 'n_tokens_mistral': 283, 'n_tokens_neox': 260, 'n_words': 130}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: 'no widget at all' focusnode strange behavior username_0: Your 'no widget at all' sample used to create an overlay on the center of the screen, but now it is on the top left: Sample code: ``` import 'package:flutter/material.dart'; import 'package:onboarding_overlay/onboarding_overlay.dart'; void main() { runApp(App()); } final GlobalKey<ScaffoldState> scaffoldKey = GlobalKey<ScaffoldState>(); class App extends StatefulWidget { final GlobalKey<OnboardingState> onboardingKey = GlobalKey<OnboardingState>(); @override _AppState createState() => _AppState(); } class _AppState extends State<App> { late List<FocusNode> focusNodes; @override void initState() { super.initState(); focusNodes = List<FocusNode>.generate( 18, (int i) => FocusNode(debugLabel: 'Onboarding Focus Node $i'), growable: false, ); } @override Widget build(BuildContext context) => MaterialApp( home: Onboarding( key: widget.onboardingKey, autoSizeTexts: true, steps: <OnboardingStep>[ OnboardingStep( focusNode: focusNodes[0], title: "Or no widget at all! You're all done!", bodyText: "Or no widget at all! You're all done!", margin: EdgeInsets.zero, labelBoxPadding: const EdgeInsets.all(8.0), shape: const CircleBorder(), fullscreen: false, overlayShape: const CircleBorder(), ), ], child: Home( focusNodes: focusNodes, ), ), ); } class Home extends StatefulWidget { const Home({ Key? key, required this.focusNodes, }) : super(key: key); final List<FocusNode> focusNodes; @override _HomeState createState() => _HomeState(); } class _HomeState extends State<Home> { @override Widget build(BuildContext context) { return Scaffold( key: scaffoldKey, body: const Center(child: Text('Oi')), floatingActionButton: Padding( padding: const EdgeInsets.only(left: 32), child: FloatingActionButton( focusNode: widget.focusNodes[1], onPressed: () { final OnboardingState? onboarding = Onboarding.of(context); if (onboarding != null) { onboarding.show(); } }, child: const Icon(Icons.add), ), ), ); } } ``` ![Screenshot_1636501091](https://user-images.githubusercontent.com/57464965/141022903-3836ec6f-c456-4fd0-954a-a236b2add370.png) <issue_comment>username_0: `onboarding_overlay: ^2.3.3` [√] Flutter (Channel master, 2.6.0-12.0.pre.665, on Microsoft Windows [versão 10.0.22000.282], locale pt-BR) [!] Android toolchain - develop for Android devices (Android SDK version 31.0.0) ! Some Android licenses not accepted. To resolve this, run: flutter doctor --android-licenses [√] Chrome - develop for the web [√] Android Studio (version 2020.3) [√] VS Code (version 1.62.1) [√] Connected device (3 available) ! Doctor found issues in 1 category.<issue_closed>
{'fraction_non_alphanumeric': 0.11035548686244204, 'fraction_numerical': 0.02936630602782071, 'mean_word_length': 2.475832438238453, 'pattern_counts': {'":': 0, '<': 14, '<?xml version=': 0, '>': 18, 'https://': 1, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '14656945', 'n_tokens_mistral': 1062, 'n_tokens_neox': 953, 'n_words': 255}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: [BUG] unable to provision VMs on elite1 and then on any node username_0: **Describe the bug** I created a 3 node cluster but it looks like the initial node was excluded from being able to create virtual machines. Then after I changed the labels to match all 3 nodes (changed the labels on elite1) I cannot get any VMs to provision at all. **To Reproduce** Try creating any VM. **Log** Events: Type Reason Age From Message ---- ------ ---- ---- ------- Warning FailedScheduling 21m default-scheduler 0/3 nodes are available: 3 pod has unbound immediate PersistentVolumeClaims. Normal Scheduled 21m default-scheduler Successfully assigned default/virt-launcher-opensuse-1-sp9gm to elite3 Warning FailedScheduling 21m default-scheduler 0/3 nodes are available: 3 pod has unbound immediate PersistentVolumeClaims. Warning FailedAttachVolume 19m attachdetach-controller AttachVolume.Attach failed for volume "pvc-8e2174f3-28fd-4de7-8bb6-85ab61110f07" : attachdetachment timeout for volume pvc-8e2174f3-28fd-4de7-8bb6-85ab61110f07 Warning FailedMount 19m kubelet Unable to attach or mount volumes: unmounted volumes=[disk-0], unattached volumes=[ephemeral-disks disk-0 container-disks hotplug-disks libvirt-runtime sockets]: timed out waiting for the condition Warning FailedMount 16m kubelet Unable to attach or mount volumes: unmounted volumes=[disk-0], unattached volumes=[sockets ephemeral-disks container-disks disk-0 hotplug-disks libvirt-runtime]: timed out waiting for the condition Warning FailedMount 12m kubelet Unable to attach or mount volumes: unmounted volumes=[disk-0], unattached volumes=[container-disks hotplug-disks libvirt-runtime sockets ephemeral-disks disk-0]: timed out waiting for the condition Warning FailedMount 10m kubelet Unable to attach or mount volumes: unmounted volumes=[disk-0], unattached volumes=[hotplug-disks libvirt-runtime sockets disk-0 ephemeral-disks container-disks]: timed out waiting for the condition Warning FailedMount 7m48s kubelet Unable to attach or mount volumes: unmounted volumes=[disk-0], unattached volumes=[container-disks hotplug-disks disk-0 libvirt-runtime sockets ephemeral-disks]: timed out waiting for the condition Warning FailedMount 5m30s kubelet Unable to attach or mount volumes: unmounted volumes=[disk-0], unattached volumes=[sockets ephemeral-disks container-disks hotplug-disks disk-0 libvirt-runtime]: timed out waiting for the condition Warning FailedMount 3m16s (x2 over 14m) kubelet Unable to attach or mount volumes: unmounted volumes=[disk-0], unattached volumes=[ephemeral-disks container-disks hotplug-disks libvirt-runtime sockets disk-0]: timed out waiting for the condition Warning FailedMount 60s kubelet Unable to attach or mount volumes: unmounted volumes=[disk-0], unattached volumes=[disk-0 hotplug-disks libvirt-runtime sockets ephemeral-disks container-disks]: timed out waiting for the condition Warning FailedAttachVolume 46s (x16 over 19m) attachdetach-controller AttachVolume.Attach failed for volume "pvc-8e2174f3-28fd-4de7-8bb6-85ab61110f07" : rpc error: code = DeadlineExceeded desc = Failed to attach volume pvc-8e2174f3-28fd-4de7-8bb6-85ab61110f07 to node elite3 **Environment:** - Harvester ISO version: 0.2 - Installation Mode: ISO - Underlying Infrastructure: Bare-metal HP EliteDesk Mini PC [harvester-supportbundle_0ce2549a-46c3-4a70-8fc6-0c636e3ab7bb_2021-05-03T20-39-10Z.zip](https://github.com/harvester/harvester/files/6417584/harvester-supportbundle_0ce2549a-46c3-4a70-8fc6-0c636e3ab7bb_2021-05-03T20-39-10Z.zip) <issue_comment>username_1: from the elite1 node: ``` status: allocatable: devices.kubevirt.io/kvm: "0" ``` Can you 1. Run `lscpu` to check if hardware virtualization is supported. 2. Run `lsmod|grep kvm` to check if the KVM module is loaded. <issue_comment>username_0: `abonilla@Alejandros-MacBook-Pro Downloads % ssh rancher@elite1 lscpu | grep vmx Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc art arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc cpuid aperfmperf pni pclmulqdq dtes64 monitor ds_cpl vmx smx est tm2 ssse3 sdbg fma cx16 xtpr pdcm pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm 3dnowprefetch cpuid_fault epb invpcid_single pti ssbd ibrs ibpb stibp tpr_shadow vnmi flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 hle avx2 smep bmi2 erms invpcid rtm mpx rdseed adx smap clflushopt intel_pt xsaveopt xsavec xgetbv1 xsaves dtherm ida arat pln pts hwp hwp_notify hwp_act_window hwp_epp md_clear flush_l1d abonilla@Alejandros-MacBook-Pro Downloads % ssh rancher@elite1 lsmod | grep kvm kvm 663552 0 abonilla@Alejandros-MacBook-Pro Downloads % ssh rancher@elite2 lscpu | grep vmx Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc art arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc cpuid aperfmperf pni pclmulqdq dtes64 monitor ds_cpl vmx smx est tm2 ssse3 sdbg fma cx16 xtpr pdcm pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm 3dnowprefetch cpuid_fault epb invpcid_single pti ssbd ibrs ibpb stibp tpr_shadow vnmi flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 hle avx2 smep bmi2 erms invpcid rtm mpx rdseed adx smap clflushopt intel_pt xsaveopt xsavec xgetbv1 xsaves dtherm ida arat pln pts hwp hwp_notify hwp_act_window hwp_epp md_clear flush_l1d abonilla@Alejandros-MacBook-Pro Downloads % ssh rancher@elite2 lsmod | grep kvm kvm_intel 282624 8 kvm 663552 1 kvm_intel abonilla@Alejandros-MacBook-Pro Downloads % ssh rancher@elite3 lscpu | grep vmx Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc art arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc cpuid aperfmperf pni pclmulqdq dtes64 monitor ds_cpl vmx smx est tm2 ssse3 sdbg fma cx16 xtpr pdcm pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm 3dnowprefetch cpuid_fault epb invpcid_single pti ssbd ibrs ibpb stibp tpr_shadow vnmi flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 hle avx2 smep bmi2 erms invpcid rtm mpx rdseed adx smap clflushopt intel_pt xsaveopt xsavec xgetbv1 xsaves dtherm ida arat pln pts hwp hwp_notify hwp_act_window hwp_epp md_clear flush_l1d abonilla@Alejandros-MacBook-Pro Downloads % ssh rancher@elite3 lsmod | grep kvm kvm_intel 282624 16 kvm 663552 1 kvm_intel` elite1 [~]$ sudo modprobe kvm_intel modprobe: ERROR: could not insert 'kvm_intel': Not supported I went into the BIOS and vtx wasn't selected - so really the way to tell if the system is correctly setup is via trying to load the kernel module kvm_intel - as lscpu/cpuinfo and kvm will load if the CPU supports it. I later rebooted and I was able to provision VMs on top of that specific node too. I did manually add the labels the other nodes had - but this was before I changed the BIOS settings - unsure if that is required or it would have been taken care of by itself. Please close! <issue_comment>username_1: Thanks for the update<issue_closed>
{'fraction_non_alphanumeric': 0.04458911681611596, 'fraction_numerical': 0.0427465913278467, 'mean_word_length': 3.1435114503816792, 'pattern_counts': {'":': 0, '<': 6, '<?xml version=': 0, '>': 6, 'https://': 1, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 12, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '9906683', 'n_tokens_mistral': 2925, 'n_tokens_neox': 2628, 'n_words': 937}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Fixed a bug in `TensorFunctionClassifier::shouldBePartitioned()` username_0: , which returned true for `_hostOp<A>(_:)` defined in the tensorflow stdlib, when we are compiling a user module. The bug was because we were evaluating `if (fn->getInlineStrategy() == NoInline)` (it is true for `_hostOp<A>(_:)`) before evaluating `if (isAvailableExternally(fn->getLinkage()))` (also true for `_hostOp<A>(_:)`). But for any externally defined functions, they should not be partitioned when compiling the user module. Rearranged the code and added comments to improve its readability. Now we first handle cases where we return false, and then the cases where we return true. <issue_comment>username_0: @swift-ci please test tensorflow
{'fraction_non_alphanumeric': 0.09196891191709844, 'fraction_numerical': 0.0025906735751295338, 'mean_word_length': 5.721739130434782, 'pattern_counts': {'":': 0, '<': 6, '<?xml version=': 0, '>': 8, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '5003279', 'n_tokens_mistral': 228, 'n_tokens_neox': 216, 'n_words': 95}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Quick feature: reload song instead of restart song in song menu username_0: Hi, In the menu when double-tapping on a song, can you please make it default to "reload song" instead of "restart song" ? When editing a song from outside stepmania (.sm), it would be really nice to reload the song and update the cache from disk as well. Curently, I have to go back to main menu, then options, then reload songs, then it lost the song I was working on, then I have to find it back in the wheel, then I can restart it. Would save a lot of pain if one could just reload song (from disk!) directly from the song menu when within the song! Thank you <issue_comment>username_1: It is possible this feature already exists, but I am not sure what you are requesting based on what you wrote above. If possible, can you clarify what you mean by * double-tapping * reload song * restart song * editing a song from outside stepmania (.sm) * Curently It is unclear where in StepMania you'd like this reloading feature to be accessible. Are you playing StepMania? Are you writing stepcharts in StepMania? Are you developing for StepMania? More context will help. <issue_comment>username_2: The reload song shortcut is pressing Ctrl+Shift+R when hovering the song in the song select. <issue_comment>username_1: Yes, that is the keyboard shortcut in most themes to reload a song from disk, creating new cache files in the process, while on ScreenSelectMusic. <issue_comment>username_0: Thanks for the quick reply. Double-tapping: Hitting ESC twice while a song is playing (when hitting this key once, the word "double-tap [ ] to access menu" appears. When hitting it twice,this brings a mini-menu for that song. In it, I see at the end of the options: - Forfeit - Restart Song I'm talking about the restart song option, and I was thinking, if for any reason it is needed, can you ALSO add Reload song in this mini menu? I understand there might be shortcut keys, but I'd like to know this is accessible by the UI as well. I don't want to have to exit the song. I would like to be able to hit "Reload Song" from the menu: - Restart Song - Reload song (from disk) - Forfeit And the Reload song would effectively call a reload of the song from disk and update the cache as well. <issue_comment>username_0: Are you playing StepMania? YES Are you writing stepcharts in StepMania? YES Are you developing for StepMania? NOT YET, but I think this is super interesting<issue_closed> <issue_comment>username_0: Nevermind, seems like Restart Song actually does a reload of the song from disk. Somehow it did not work when I had tried.
{'fraction_non_alphanumeric': 0.044642857142857144, 'fraction_numerical': 0.0026041666666666665, 'mean_word_length': 4.054511278195489, 'pattern_counts': {'":': 0, '<': 9, '<?xml version=': 0, '>': 9, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '21045900', 'n_tokens_mistral': 744, 'n_tokens_neox': 700, 'n_words': 441}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: (PRESTASHOP)Status: Disconnected, Reason: The access token that you've provided is not valid, check your credentials or endpoint. username_0: I keep getting this error when i save my configuration with the Prestashop plugin. ![image](https://cloud.githubusercontent.com/assets/5989118/20249335/b5226a20-a9f7-11e6-9582-acf1aa838458.png) I know I have entered the right credentials. <issue_comment>username_1: Hello @username_0 Please contact with our devteam.username_1. An issue is not related directly to PHP-SDK.<issue_closed>
{'fraction_non_alphanumeric': 0.078397212543554, 'fraction_numerical': 0.07142857142857142, 'mean_word_length': 5.46067415730337, 'pattern_counts': {'":': 0, '<': 4, '<?xml version=': 0, '>': 4, 'https://': 1, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '26313705', 'n_tokens_mistral': 197, 'n_tokens_neox': 172, 'n_words': 58}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Enable click-and-drug plotLines on graph username_0: When clicking on a graph, the plotLine should move at the underneath date on all graphs and time slider on the map <issue_comment>username_0: Solved with https://github.com/informatici/survethi-monitoring-tool/commit/24ffddfd87ce6a48791bbb0986d6cfecc7a4cb97<issue_closed>
{'fraction_non_alphanumeric': 0.06685236768802229, 'fraction_numerical': 0.0584958217270195, 'mean_word_length': 7.181818181818182, 'pattern_counts': {'":': 0, '<': 4, '<?xml version=': 0, '>': 4, 'https://': 1, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '16318740', 'n_tokens_mistral': 126, 'n_tokens_neox': 114, 'n_words': 33}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: False positive for regex.Match.__getitem__ username_0: ## Environment data VS Code w/ Pylance v2022.2.3, Python v2022.0.1814523869 - Language Server version: 2022.2.3 (pyright 9b0ea9eb) - OS and version: Manjaro Linux - Python version: 3.10.2 ## Expected behaviour Passes type checking ## Actual behaviour `"__getitem__" method not defined on type "Match[str]"` ## Logs Nothing relevant in the trace output. ## Code Snippet / Additional information ```python import regex if m := regex.match(r'(.)', 'abc'): print(m[1]) ``` virtualenv with `regex==2022.1.18` <issue_comment>username_0: Whoops, turns out I already reported this: https://github.com/mrabarnett/mrab-regex/issues/423 And the solution is to install the `types-regex` module.<issue_closed>
{'fraction_non_alphanumeric': 0.11374407582938388, 'fraction_numerical': 0.05568720379146919, 'mean_word_length': 3.642857142857143, 'pattern_counts': {'":': 0, '<': 4, '<?xml version=': 0, '>': 4, 'https://': 1, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '14566757', 'n_tokens_mistral': 327, 'n_tokens_neox': 278, 'n_words': 83}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Docs Issue: create_db.sql file not found. username_0: When clicking on 'create_db.sql' mentioned in the readme file returns 404 Not found. <issue_comment>username_1: what README docs are you referring to? <issue_comment>username_1: ah I see. Yes, the README is out of date, and should be updated. Meanwhile the best way to build a cluster is either to use the terrraform scripts if you are running on AWS, or follow [these instructions](https://docs.jambonz.org/installing/#otherwise) if you are building on your own infrastructure <issue_comment>username_0: Ok noted thx, ran into other issues with TerraForm, already reported it.
{'fraction_non_alphanumeric': 0.06398809523809523, 'fraction_numerical': 0.010416666666666666, 'mean_word_length': 5.174311926605505, 'pattern_counts': {'":': 0, '<': 5, '<?xml version=': 0, '>': 5, 'https://': 1, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '14617673', 'n_tokens_mistral': 195, 'n_tokens_neox': 183, 'n_words': 90}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Type promotion for unary "universal functions (ufuncs)" username_0: NumPy has the concept of a "[universal function](https://docs.scipy.org/doc/numpy/reference/ufuncs.html) (ufunc)", a class of element-wise functions. This post is concerned with PyTorch's implementation of type promotion for unary universal functions like `torch.sin` and `torch.abs`. It describes PyTorch's and NumPy's current unary ufunc type promotion behavior and makes three proposals for how PyTorch can change to be more compatible with NumPy. First, let's define the unary ufuncs in PyTorch (separated into two groups that will be explained shortly): - **Type preserving unary ufuncs**: abs, conj, bitwise_not, clamp, mul, reciprocal, neg, round, square - **Floating unary ufuncs**: angle, cos, cosh, acos, sin, sinh, asin, tan, tanh, atan, ceil, floor, digamma, lgamma, mvlgamma, polygamma, erf, erfc, erfinv, exp, expm1, frac, log, log10, log1p, log2, sqrt, rsqrt, sigmoid, sign, trunc When describing "type promotion" for these functions we're looking at two things: - the computation type - the output type And how these are affected by: - the input type - the dtype kwarg's (optional) value - the out kwarg's (optional) tensor ## The default case (no kwargs) With no kwargs, "type preserving unary ufuncs" like `np.abs` perform their computation in the input type and set their output type to the same. Most unary ufuncs, however, are "floating unary ufuncs." Let's look at NumPy's cosine as an example. With no kwargs, `np.cos` performs its computation in and outputs in the default scalar type when given integer inputs: ``` np.cos((1, 2, 3)) : array([ 0.54030231, -0.41614684, -0.9899925 ]) a = np.array((1, 2, 3), dtype=np.float16) np.cos(a) : array([ 0.5405, -0.4163, -0.99 ], dtype=float16) ``` This behavior is natural since the result of the cosine operation may not be expressible as an integer. In PyTorch today, on the other hand, `torch.cos` simply fails on integer inputs: ``` t = torch.tensor((1, 2, 3)) torch.cos(t) : RuntimeError: cos_vml_cpu not implemented for 'Long' ``` **Proposal 1**: PyTorch adopts the same floating unary ufunc (no kwargs) behavior as NumPy. ## Setting the dtype kwarg The dtype kwarg determines the type of the computation. For example: ``` a = (1.1, 1.2, 1.3) np.cos(a) : array([0.45359612, 0.36235775, 0.26749883]) np.cos(a).astype(np.float16) : array([0.4536, 0.3623, 0.2676], dtype=float16) np.cos(a, dtype=np.float16) :array([0.4539, 0.362 , 0.2676], dtype=float16) ``` Note the difference between computing in np.float16 and computing in np.double and then casting the result to np.float16. This behavior is easier to see with a binary ufunc like add: ``` np.add(1.7, .5, dtype=np.long, casting='unsafe') : 1 ``` Both inputs are cast to long before being added. (Note the casting rule is set to unsafe to permit this.) This is not to say that every function supports every computation type. For example: ``` np.sin(np.pi, dtype=np.long) : TypeError: No loop matching the specified signature and casting was found for ufunc sin ``` Unfortunately many of PyTorch's unary ufuncs are missing the dtype keyword. **Proposal 2**: PyTorch adds the dtype keyword to the unary ufuncs and adopts the same behavior as NumPy. ## Setting the out kwarg In NumPy, setting the out kwarg, as might be expected, determines a unary ufunc's out type. For example, [Truncated] b = np.array((107076232663.,), dtype=np.float32) out = np.empty(1, dtype=np.float64) np.divide(a, b, out=out) == np.divide(a,b) : True ``` **Proposal 3**: PyTorch adopts the same out kwarg type promotion behavior as NumPy. ## Summary This issue discussed unary ufunc type promotion in PyTorch and three places where we're incompatible with NumPy's behavior: - PyTorch doesn't have a notion of "floating unary ufuncs" that perform their computation in floating point by default. - PyTorch's unary ufuncs are often missing the dtype kwarg. - PyTorch's handling of the out kwarg can change the type of the computation performed, unlike NumPy's behavior where it simply specifies the output type. I propose PyTorch adopt NumPy's type promotion behavior for its unary ufuncs. Natural follow-ups to this issue would be to review type promotion for the binary ufuncs (like add) and reductions (like mean, std, var, and sum). In particular, the binary ufuncs suffer from these same issues, although with the binary ufuncs we have the additional complication of selecting a "common input type" from the possibly heterogeneous types of the input tensors. Generalizing this proposal wouldn't change that logic directly, but it would stop tensors specified in the out kwarg from participating in determining the common input type. cc @username_1, @gchanan, @ngimel <issue_comment>username_1: A few comments/thoughts: * I see a plan to support python3-like division where `int_tensor / int_tensor => floating_tensor`, so when we make changes to how we think about type preserving vs floating ufuncs, we should think about how that applies beyond just the unary ones. * Proposal1: nit: technically there are (or will be) cases where a function can be either a "Type preserving unary ufunc" or "Floating unary ufuncs" depending on context, e.g. we will expect `torch.abs(complex_tensor)` to return a floating tensor, even though `abs(int)->int`. I'm not sure what the full specification of categories of ufuncs is. * Proposal2: Would we want to wait to offer the dtype arg until/unless we also add more complete casting rules? Right now we don't support `casting='unsafe'` * Proposal3: What do the use-cases here look like and what specifically makes numpy preferable? Will it make sense across all types of ops? Some other examples for (3): * 1f / 3f with out=double in numpy is: 0.3333333432674408 and in pytorch is: 0.3333333333333333. This is consistent with C++ (for `double third = f1 / f3`). * True + True with out=int or float in numpy is 1 and in pytorch is 2. This could be inconsistent with how we handle boolTensor.sum(). * Currently `torch.div(torch.tensor(5), torch.tensor(2))` returns 2 but `torch.div(torch.tensor(5), torch.tensor(2), out=floating_tensor)` returns 2.5. If we just ignore the out dtype we'll get the `2` which is inconsistent with numpy, unless we first change the behavior of integer division. If we add the dtype argument from Proposal2 (and specify dtype=out.dtype) I assume the expectation is we'd get the original pytorch result in these examples? <issue_comment>username_0: We don't always compute in the out type! For example: ``` t = torch.tensor((.0000001,), dtype=torch.double).cuda() t_half = t.to(torch.half) out = torch.empty((1,), dtype=torch.half).cuda() print(torch.add(t, t, out=out)) print(torch.add(t, t).to(torch.float16)) print(torch.add(t_half, t_half)) ``` Will print: ``` tensor([1.7881e-07], device='cuda:0', dtype=torch.float16) tensor([1.7881e-07], device='cuda:0', dtype=torch.float16) tensor([2.3842e-07], device='cuda:0', dtype=torch.float16) ``` So we can see here that we computed in double since double is higher in our type hierarchy than half and then we cast the result to half. Setting dtype=torch.float16 would change the computation to occur in half. We can see this in the corresponding NumPy: ``` a = np.array((.0000001,), dtype=np.double) a_half = a.astype(np.half) out = np.empty((1,), dtype=np.half) print(np.add(a, a, dtype=np.float16, out=out)) print(np.add(a, a).astype(np.float16)) print(np.add(a_half, a_half)) ``` Which gives: ``` [2.384185791015625e-07] [1.7881393432617188e-07] [2.384185791015625e-07] ``` Getting back to Proposal #3, I think the behavior is consistent with C++ and NumPy, and those are tough conventions to deviate from -- I think we'd want a compelling reason to do so. I also think the behavior is intuitive and simple to describe, whereas today PyTorch sometimes computes in the out type and sometimes doesn't depending on where the out type's dtype is in the type hierarchy compared to the input dtypes. So we're actually already consistent with NumPy's out behavior when the input dtypes are "above" the output's dtype in our type hierarchy, but we're inconsistent when they're "below" it. I'm sure there are users who want that behavior, but I don't think we see evidence that this behavior is expected or so popular that it's worth deviating from convention. <issue_comment>username_1: It does seem low-priority. I guess I just want to make sure that if we're not adding it, we're highlighting examples of the proposal that don't require it (which I think we have now). <issue_comment>username_0: Aha! My mistake. Yep, you are correct, correct, correct, and correct. <issue_comment>username_0: Closing this issue because we now have floating unary elementwise ops and they handle the out= argument in a NumPy-compatible fashion.<issue_closed>
{'fraction_non_alphanumeric': 0.09088903595852636, 'fraction_numerical': 0.03595852636223252, 'mean_word_length': 4.314771395076201, 'pattern_counts': {'":': 0, '<': 8, '<?xml version=': 0, '>': 10, 'https://': 1, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 2, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '23828816', 'n_tokens_mistral': 3007, 'n_tokens_neox': 2725, 'n_words': 1234}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Django 2.1 & 2.2 compatibility? username_0: Is this package compatible with Django 2.1 & 2.2? Anyone using it in production? If so, we should add it to tox.ini and travis.yml <issue_comment>username_1: This package does nto work with Django 3. results in `TemplateSyntaxError at / 'staticfiles' is not a registered tag library` <issue_comment>username_2: Not working for me in Django 2.2, any tips? I followed changes above. User get logged out after time regardless of activity and the popup warning never shows. <issue_comment>username_3: @username_2 do you get any JavaScript errors in your browser debug console? Did you included the `session_security/all.html` template? And did you loaded jQuery in any way? ```html {# Load SessionSecurity javascript 'class', jquery should be loaded - by you - at this point #} <script type="text/javascript" src="{% static 'session_security/script.js' %}"></script> ```
{'fraction_non_alphanumeric': 0.08376421923474664, 'fraction_numerical': 0.016546018614270942, 'mean_word_length': 4.563218390804598, 'pattern_counts': {'":': 0, '<': 7, '<?xml version=': 0, '>': 7, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '1235481', 'n_tokens_mistral': 287, 'n_tokens_neox': 261, 'n_words': 125}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Updated README.md - Linux and MacOS instructions username_0: Added instructions on how to use and install F# on Linux and MacOS <issue_comment>username_1: Please reference the Issue so that Issue is closed when PR is merged! Kindly Google how to do that and ping me in case you have doubts!
{'fraction_non_alphanumeric': 0.04294478527607362, 'fraction_numerical': 0.006134969325153374, 'mean_word_length': 5.169811320754717, 'pattern_counts': {'":': 0, '<': 3, '<?xml version=': 0, '>': 3, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '16149008', 'n_tokens_mistral': 86, 'n_tokens_neox': 83, 'n_words': 49}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: RecyclerBinderUpdateCallback not work as expected. username_0: - [x] I have searched [existing issues](https://github.com/facebook/litho/issues) and this is not a duplicate ## Issues and Steps to Reproduce I'd like to bind my Recycler by using `DiffUtil` and `RecyclerBinderUpdateCallback` this way: Define a custome component called `PlaygroundComponentSpec` and it will display a list: ```java [PlaygroundComponentSpec.java] @LayoutSpec public class PlaygroundComponentSpec { @OnCreateLayout static ComponentLayout onCreateLayout( ComponentContext c) { final RecyclerBinder recyclerBinder = new RecyclerBinder( c, 4.0f, new LinearLayoutInfo(c, OrientationHelper.VERTICAL, false)); addDataToBinder(recyclerBinder, c); return Column.create(c) .backgroundColor(Color.WHITE) .child( Recycler.create(c) .binder(recyclerBinder) .build()) .build(); } ... } ``` Then populate the binder with list items: ```java [PlaygroundComponentSpec.java] private static void addDataToBinder(RecyclerBinder recyclerBinder, final ComponentContext c) { final List<String> items = Arrays.asList( "Item 1", "Item 2", "Item 3"); final DiffUtil.DiffResult diffResult = DiffUtil.calculateDiff(new DiffUtil.Callback() { @Override public int getOldListSize() { return 0; } @Override public int getNewListSize() { return items.size(); } @Override public boolean areItemsTheSame(int oldItemPosition, int newItemPosition) { return false; } @Override public boolean areContentsTheSame(int oldItemPosition, int newItemPosition) { return false; } }); final RecyclerBinderUpdateCallback<String> updateCallback = RecyclerBinderUpdateCallback.acquire( 0, items, new RecyclerBinderUpdateCallback.ComponentRenderer<String>() { @Override public ComponentInfo render(String item, int idx) { return ComponentInfo.create() .component( Text.create(c) .textSizeSp(20) .text(item) .build()) .build(); } }, recyclerBinder); diffResult.dispatchUpdatesTo(updateCallback); } ``` ## Expected Behavior Recycler should display a list of 3 Texts view. ## Link to Code https://gist.github.com/username_0/19f57d709727257189491447fc87c61a <issue_comment>username_1: cc @username_2 <issue_comment>username_2: Really sorry for the late reply but I've been out of the office. The problem here is probably that you are allocating the RecyclerBinder inside onCreateLayout. I would suggest keeping it in your Fragment and pass it down as prop. Also you want to invoke addDataToBinder from your fragment. onCreateLayout will run every time the system needs to compute a new layout.<issue_closed> <issue_comment>username_0: Thank you for your reveal of "onCreateLayout will run every time the system needs to compute a new layout". But the code is only for demo purpose. After I take a deeper look at source code of the `RecyclerBinderUpdateCallback.java` I figured out that there is a method `applyChangeset()` and it must be called right after `diffResult.dispatchUpdatesTo(updateCallback)` like below: ```java diffResult.dispatchUpdatesTo(updateCallback); updateCallback.applyChangeset(); ``` It's a little bit strange but it worked.
{'fraction_non_alphanumeric': 0.07587548638132295, 'fraction_numerical': 0.011673151750972763, 'mean_word_length': 2.9944506104328523, 'pattern_counts': {'":': 0, '<': 9, '<?xml version=': 0, '>': 9, 'https://': 2, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '18644848', 'n_tokens_mistral': 1057, 'n_tokens_neox': 965, 'n_words': 320}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Adding reading listed in next episode username_0: Not sure I got the links right but I was trying to add reading the callout since it is listed in the readings in the next episode but not here in the homework. <issue_comment>username_0: In the other time reading a callout is listed as homework it is listed in that episode. Should we move this call out to the previous lesson?
{'fraction_non_alphanumeric': 0.029055690072639227, 'fraction_numerical': 0.004842615012106538, 'mean_word_length': 4.75, 'pattern_counts': {'":': 0, '<': 3, '<?xml version=': 0, '>': 3, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '24496086', 'n_tokens_mistral': 101, 'n_tokens_neox': 98, 'n_words': 69}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Typescript Error username_0: I am following the README file in the repo. I am just starting out using the library. Here is the error that I am getting after creating pages/_app.ts: ` Type 'boolean' is not assignable to type 'ReactElement<any, any> | null'.ts(2322) index.d.ts(544, 15): The expected type comes from the return type of this signature. var pageProps: any.` Here's my pages/_app.ts content: ``` import React, { FC } from 'react'; import { AppProps } from 'next/app'; import { wrapper } from '../components/store'; const WrappedApp: FC<AppProps> = ({ Component, pageProps }) => ( <Component {...pageProps} /> ); export default wrapper.withRedux(WrappedApp); ```<issue_closed> <issue_comment>username_0: I found a solution. I wasn't doing the right thing. The following code worked for me: ``` import '../styles/globals.css'; import type { AppProps } from 'next/app'; import { wrapper } from '../components/store'; function WrappedApp({ Component, pageProps }: AppProps) { return <Component {...pageProps} />; } export default wrapper.withRedux(WrappedApp); ```
{'fraction_non_alphanumeric': 0.12756849315068494, 'fraction_numerical': 0.009417808219178082, 'mean_word_length': 3.676, 'pattern_counts': {'":': 0, '<': 8, '<?xml version=': 0, '>': 9, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '28023527', 'n_tokens_mistral': 362, 'n_tokens_neox': 342, 'n_words': 125}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Broken Github Icon in Footer of All Dataverse Guides username_0: In all of our Sphinx guides the GitHub icon is broken at the footer. ![image](https://cloud.githubusercontent.com/assets/5013837/6948066/5429e014-d879-11e4-9ba7-44b4ff405822.png) @username_1 do you know where / how this can be fixed? <issue_comment>username_1: Also, the copyright is outdated. <issue_comment>username_1: Fixed broken image URLs for footer and favicon in User Guides. Updated copyright year. <issue_comment>username_2: Can see fix here: http://guides.dataverse.org/en/4.0.1/user/ Closing ticket.<issue_closed>
{'fraction_non_alphanumeric': 0.08990536277602523, 'fraction_numerical': 0.07255520504731862, 'mean_word_length': 5.105769230769231, 'pattern_counts': {'":': 0, '<': 6, '<?xml version=': 0, '>': 6, 'https://': 1, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '29750372', 'n_tokens_mistral': 235, 'n_tokens_neox': 205, 'n_words': 65}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: TypeError: node.err is not a function in publish node username_0: Hi, there is a problem in the publish node. Errror message: ``` 16 Aug 08:22:13 - [info] [nats-streaming-publish:e7f7705c.c135e] publish failed: Error: stan: publish ack timeout 16 Aug 08:22:13 - [red] Uncaught Exception: 16 Aug 08:22:13 - TypeError: node.err is not a function at Object.ah (/data/node_modules/node-red-contrib-natsstreaming/nats-streaming-publish.js:67:26) at Timeout.ack.t.timers.setTimeout (/data/node_modules/node-nats-streaming/lib/stan.js:606:15) at ontimeout (timers.js:498:11) at tryOnTimeout (timers.js:323:5) at Timer.listOnTimeout (timers.js:290:5) npm ERR! code ELIFECYCLE npm ERR! errno 1 ``` Background: I tried to send 2000 msg/second to the nats streaming server. The msg `publish failed: Error: stan: publish ack timeout` it's ok for me, but node-red stops with Exit status 1 Solution: I comment out the line 67. Thanks <issue_comment>username_1: Hi, yes this line should be: node.error('problem while publishing message', + err); I pushed an update to github. Can you test if this fixes the error for you? <issue_comment>username_0: Hi, the test is done. It works. Thanks <issue_comment>username_1: You are welcome.<issue_closed>
{'fraction_non_alphanumeric': 0.09969558599695585, 'fraction_numerical': 0.0502283105022831, 'mean_word_length': 4.1976284584980235, 'pattern_counts': {'":': 0, '<': 6, '<?xml version=': 0, '>': 6, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '27382750', 'n_tokens_mistral': 492, 'n_tokens_neox': 441, 'n_words': 153}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Fix typos in xdoc constructors xdoc username_0: Fix typos in xdoc constructors xdoc <issue_comment>username_1: I don't know if this acl2s stuff is just a redo of changes already made. Can you chime in, @pmanolios ? <issue_comment>username_1: Ah, based on the commit authors, the acl2s stuff is only showing up because acl2-testing-kestrel doesn't yet have the acl2 commits. Merging.
{'fraction_non_alphanumeric': 0.057279236276849645, 'fraction_numerical': 0.016706443914081145, 'mean_word_length': 5.363636363636363, 'pattern_counts': {'":': 0, '<': 4, '<?xml version=': 0, '>': 4, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '23623832', 'n_tokens_mistral': 136, 'n_tokens_neox': 133, 'n_words': 58}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Minor CPU optimizations for listStatus username_0: During listStatus, couple of expensive methods showed up in profiler. Patch addresses those minor CPU optimizations codepaths. With the patch, following methods do not show up anymore as expensive methods. 1. SAXParser initializations. In SAXParser init, this ended up scanning for jar file entries. <img width="101" alt="saxparser_screenshot" src="https://cloud.githubusercontent.com/assets/323339/25956781/45301efe-368a-11e7-8000-33e38424e7e8.png"> 2. DateFormat showed up. CPU usage can be reduced with JODA time. <issue_comment>username_1: Hi @username_0 , we really try to limit our dependencies when at all possible. Do you have exact perf numbers for the improvement using joda-time? Most likely we cannot accept this part of the change. <issue_comment>username_0: I was using profiler to check the issues. If joda-time dependency is an issue, I can make the the formatter part of the threadlocal itself. Will post the revised PR. <issue_comment>username_0: Uploaded revised PR. <issue_comment>username_1: Also please change the PR to commit into the dev branch not master. <issue_comment>username_0: Changed the PR to commit to dev branch.
{'fraction_non_alphanumeric': 0.05806451612903226, 'fraction_numerical': 0.04032258064516129, 'mean_word_length': 5.205, 'pattern_counts': {'":': 0, '<': 8, '<?xml version=': 0, '>': 8, 'https://': 1, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '1724900', 'n_tokens_mistral': 364, 'n_tokens_neox': 338, 'n_words': 161}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Initial Cinematic Dialog System username_0: ### tl;dr: Major new file is DialogSystem.js. Other changes are minor housekeeping changes and improvements. ## Simplified data transformation to get text on the screen. ![DialogSystem](https://user-images.githubusercontent.com/15080861/56757957-50a8ee80-674a-11e9-82ad-010f96e5e6d5.png) ## Features: - `/cinematic` route now redirects to `/` if not an admin. - Positioning text bubble - A simple template language leveraging a [super light library](https://blueimp.github.io/JavaScript-Templates/) that we've [forked and simplified](https://github.com/codecombat/JavaScript-Templates/tree/remove-load). - Localization added to schema and dialog bubbles correctly handle language. - Left and Right svg speaker bubbles. (svg programmer art) - Some light tests on the most important pieces. (text transpilation and transformation) - Clear all current text bubbles command. - All systems are now iterated for `parseDialogNode` per dialogNode. - All systems are now iterated for `parseSetupShot` per setupShot node. - i18n changes added to cinematic schema. - moved some logic into more rational places. Commands are now in a single file, and systems handle creating their own commands. <issue_comment>username_1: Added some questions and suggestions. Otherwise looks good! <issue_comment>username_1: I generally feel that the dialogSystem file is too big, and can probably be refactored. <issue_comment>username_0: @username_1 I addressed your concerns. Some of them are addressed in the next PR with other changes. Generally made the DialogSystem file smaller, which will encourage the files to stay smaller moving forward. That was an excellent suggestion.
{'fraction_non_alphanumeric': 0.06850084222346996, 'fraction_numerical': 0.024705221785513758, 'mean_word_length': 4.416413373860182, 'pattern_counts': {'":': 0, '<': 5, '<?xml version=': 0, '>': 5, 'https://': 3, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '21984535', 'n_tokens_mistral': 503, 'n_tokens_neox': 464, 'n_words': 204}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Feed a real file path to the cpplint, not a temporary one username_0: Cpplint relies on finding the corresponding header file in the same directory for some checks. Feeding a temporary file causes this check to fail incorrectly. For example it will report error like "Found C++ system header after other header. Should be tmp165kvk.h, [...]" when cc file contains #include <string> line. Also it will report failure like "Add #include <string> for string" when checking cc file and that include is in only added in corresponding header file. None of this will be a problem when linting the actual file. Although it will limit the type of background checks that can be performed - see https://github.com/SublimeLinter/SublimeLinter/pull/332 - I still think that correctness of the linting is more important to fix.<issue_closed>
{'fraction_non_alphanumeric': 0.051605504587155966, 'fraction_numerical': 0.008027522935779817, 'mean_word_length': 4.388888888888889, 'pattern_counts': {'":': 0, '<': 5, '<?xml version=': 0, '>': 5, 'https://': 1, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '22758638', 'n_tokens_mistral': 226, 'n_tokens_neox': 213, 'n_words': 129}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Native library not loaded on Linux username_0: Hi I'm having some serious issues with this package. I added the package to my project and copied the native .so file to the root of my ASP.NET Core project. Now when I start the application and try to convert HTML to PDF I get the following error: ``` (Exception from HRESULT: 0x8007007E)) ---> System.DllNotFoundException: Unable to load DLL 'libwkhtmltox': The specified module could not be found. ``` I've tried setting up `LD_LIBRARY_PATH` to my app folder. I even copied the .so file to `/usr/lib` but it doesn't work at all. Any ideas what to do here? <issue_comment>username_1: Hi, I tried to run example DinkToPdf.TestConsoleApp on Ubuntu 16.04 LTS without problems. ![image](https://user-images.githubusercontent.com/13537823/28706210-89e8aed2-7373-11e7-8fea-6a341ad45c8f.png) Make sure that libwkhtmltox.so is placed in root folder (where your main project .dll is) of your project and is named correctly (libwkhtmltox). <issue_comment>username_1: Any update? <issue_comment>username_0: I have tried it again. I should add that I am trying it on the `microsoft/dotnet:1.1.2-runtime` image. So that might complicate things also. Have you tried using the library in that way? <issue_comment>username_1: Is it possible to share your image? I never tested it in that way, but will happily take a look into it. <issue_comment>username_2: Try to add the files: libwkhtmltox.dll libwkhtmltox.dylib libwkhtmltox.so on the root folder. <issue_comment>username_1: If you are running on Linux copying libwkhtmltox.so is enough. <issue_comment>username_2: I deployed my application on Linux and it works. But I installed libgdiplus as I needed it in Captcha component which I use in my application. I don't know if wkhtmltopdf depends on it or not. Try to install it. Hope this fix your problem. sudo apt-get install libgdiplus <issue_comment>username_1: After install of libgdiplus in `microsoft/dotnet:1.1.2-runtime` image libwkhtmltox library is found. ![image](https://user-images.githubusercontent.com/13537823/28770178-84a12fcc-75de-11e7-9e77-d2b293c36c71.png) Dockerfile: ``` FROM microsoft/dotnet:1.1.2-runtime RUN ["apt-get", "update"] RUN ["apt-get", "-y", "install", "libgdiplus"] WORKDIR /app COPY / . ENTRYPOINT ["dotnet", "DinkToPfd.TestConsoleApp.dll"]``` Thanks @username_2 for pointing this out.<issue_closed> <issue_comment>username_1: Hi I'm having some serious issues with this package. I added the package to my project and copied the native .so file to the root of my ASP.NET Core project. Now when I start the application and try to convert HTML to PDF I get the following error: ``` (Exception from HRESULT: 0x8007007E)) ---> System.DllNotFoundException: Unable to load DLL 'libwkhtmltox': The specified module could not be found. ``` I've tried setting up `LD_LIBRARY_PATH` to my app folder. I even copied the .so file to `/usr/lib` but it doesn't work at all. Any ideas what to do here? <issue_comment>username_0: I added libgdiplus to the image and it works! Thanks for the help :-)<issue_closed> <issue_comment>username_3: In case someone else runs into this problem when running tests (inside a Docker container or in a Linux host), adding `libgdiplus` to the image didn't do the trick for me. I checked `libwkhtmltox.so` with `ldd` and `libgdiplus` is not found as a dependency: ``` # ldd libwkhtmltox.so linux-vdso.so.1 (0x00007ffc29de1000) libXrender.so.1 => /usr/lib/x86_64-linux-gnu/libXrender.so.1 (0x00007f628c25d000) libfontconfig.so.1 => /usr/lib/x86_64-linux-gnu/libfontconfig.so.1 (0x00007f628c01f000) libfreetype.so.6 => /usr/lib/x86_64-linux-gnu/libfreetype.so.6 (0x00007f628bd70000) libXext.so.6 => /usr/lib/x86_64-linux-gnu/libXext.so.6 (0x00007f628bb5e000) libX11.so.6 => /usr/lib/x86_64-linux-gnu/libX11.so.6 (0x00007f628b81e000) libz.so.1 => /lib/x86_64-linux-gnu/libz.so.1 (0x00007f628b604000) libdl.so.2 => /lib/x86_64-linux-gnu/libdl.so.2 (0x00007f628b400000) librt.so.1 => /lib/x86_64-linux-gnu/librt.so.1 (0x00007f628b1f8000) libpthread.so.0 => /lib/x86_64-linux-gnu/libpthread.so.0 (0x00007f628afdb000) libstdc++.so.6 => /usr/lib/x86_64-linux-gnu/libstdc++.so.6 (0x00007f628ac59000) libm.so.6 => /lib/x86_64-linux-gnu/libm.so.6 (0x00007f628a955000) libgcc_s.so.1 => /lib/x86_64-linux-gnu/libgcc_s.so.1 (0x00007f628a73e000) libc.so.6 => /lib/x86_64-linux-gnu/libc.so.6 (0x00007f628a39f000) /lib64/ld-linux-x86-64.so.2 (0x00007f628f1a2000) libexpat.so.1 => /lib/x86_64-linux-gnu/libexpat.so.1 (0x00007f628a175000) libpng16.so.16 => /usr/lib/x86_64-linux-gnu/libpng16.so.16 (0x00007f6289f42000) libxcb.so.1 => /usr/lib/x86_64-linux-gnu/libxcb.so.1 (0x00007f6289d1a000) libXau.so.6 => /usr/lib/x86_64-linux-gnu/libXau.so.6 (0x00007f6289b16000) libXdmcp.so.6 => /usr/lib/x86_64-linux-gnu/libXdmcp.so.6 (0x00007f6289910000) libbsd.so.0 => /lib/x86_64-linux-gnu/libbsd.so.0 (0x00007f62896fa000) ``` What worked for me, for running tests, was to copy `libwkhtmltox.so` to `/usr/lib` directory. <issue_comment>username_4: Hi @username_3 How do I copy libwkhtmltox.so to /usr/lib in DockerFile? <issue_comment>username_3: Hi @username_4, here's the RUN command I used: ``` # Download libwkhtmltox and install it's dependencies RUN apt-get update \ && apt-get install -y --no-install-recommends \ zlib1g \ fontconfig \ libfreetype6 \ libx11-6 \ libxext6 \ libxrender1 \ && curl -o /usr/lib/libwkhtmltox.so \ --location \ https://github.com/username_1/DinkToPdf/raw/v1.0.8/v0.12.4/64%20bit/libwkhtmltox.so ``` Results may vary depending on the source image you're using. In my case I'm using `microsoft/dotnet:2.0-sdk-stretch`. <issue_comment>username_4: Hi @username_3, I´m using microsoft/aspnetcore-build:2.0. I turned his command on, but continued with error. ``` run cd /usr/lib && ls 2018-05-15T19:59:14.5356050Z apt 2018-05-15T19:59:14.5375180Z binfmt.d 2018-05-15T19:59:14.5391880Z coreutils 2018-05-15T19:59:14.5406080Z dpkg 2018-05-15T19:59:14.5421160Z gcc 2018-05-15T19:59:14.5435550Z gnupg 2018-05-15T19:59:14.5449470Z locale 2018-05-15T19:59:14.5464690Z mime 2018-05-15T19:59:14.5481610Z modules-load.d 2018-05-15T19:59:14.5495690Z os-release 2018-05-15T19:59:14.5509430Z perl5 2018-05-15T19:59:14.5523650Z python2.7 2018-05-15T19:59:14.5537710Z python3 2018-05-15T19:59:14.5552810Z sasl2 2018-05-15T19:59:14.5566130Z ssl 2018-05-15T19:59:14.5581820Z sysctl.d 2018-05-15T19:59:14.5596730Z systemd 2018-05-15T19:59:14.5611020Z tar 2018-05-15T19:59:14.5624310Z tc 2018-05-15T19:59:14.5638380Z tmpfiles.d 2018-05-15T19:59:14.5652160Z x86_64-linux-gnu ``` It was to have the packages that I had installed? <issue_comment>username_3: @username_4, if you can share your Dockerfile I review it and maybe give you some guidance. <issue_comment>username_3: @username_4, I believe the problem is that the second stage of your build (`runtime`) is not running the below command again, so `libwkhtmltox` and it's dependencies are left in the first stage (`build`). ``` RUN apt-get update \ && apt-get install -y --no-install-recommends \ zlib1g \ fontconfig \ libfreetype6 \ libx11-6 \ libxext6 \ libxrender1 \ && curl -o /usr/lib/libwkhtmltox.so \ --location \ https://github.com/username_1/DinkToPdf/raw/v1.0.8/v0.12.4/64%20bit/libwkhtmltox.so ``` <issue_comment>username_4: After 4 days it worked. Thank you <issue_comment>username_5: Guys, I faced with the same issue. Can't load library on Linux docker image. I've an .Net Core 2.0 API project and I add a docker support via Visual Studio. Then, I added libwkhtmltox library and it's dependencies to the image. Currently, my dockerfile looks like `FROM microsoft/aspnetcore:2.0 AS base # Install libwkhtmltox dependencies RUN apt-get update \ && apt-get install -y --no-install-recommends \ zlib1g \ fontconfig \ libfreetype6 \ libx11-6 \ libxext6 \ libxrender1 \ && curl -o /usr/lib/libwkhtmltox.so \ --location \ https://github.com/username_1/DinkToPdf/raw/v1.0.8/v0.12.4/64%20bit/libwkhtmltox.so WORKDIR /app EXPOSE 80 FROM microsoft/aspnetcore-build:2.0 AS build WORKDIR /src COPY API/API.csproj API/ RUN dotnet restore API/API.csproj COPY . . WORKDIR /src/API RUN dotnet build API.csproj -c Release -o /app FROM build AS publish RUN dotnet publish API.csproj -c Release -o /app FROM base AS final WORKDIR /app COPY --from=publish /app . ENTRYPOINT ["dotnet", "API.dll"]` I placed the library files to /app/wkhtmltox folder. Also, thanks to [this](https://github.com/username_1/DinkToPdf/issues/5) I can run the project from Visual Studio on Windows. But, docker fails. I tried to use /app/wkhtmltox/libwkhtmltox path and also tried /usr/lib/libwkhtmltox and I still get the error " System.DllNotFoundException: Unable to load DLL '/app/wkhtmltox/libwkhtmltox': The specified module or one of its dependencies could not be found." Could you please help to figure out with that? Thanks. <issue_comment>username_3: @username_5, conceptually the same problem @username_4 had, each new build stage can only get artifacts from the previous stage. More information [here](https://docs.docker.com/develop/develop-images/multistage-build/#before-multi-stage-builds). You should put the RUN step which installs libwkhtmltox and it's dependencies in each build stage that will need it. In our case I'd guess it would only be the last one, `final`. <issue_comment>username_5: @username_3 Thank you for your response. Actually, installing dependencies only in the final was my first idea. And unfortunately it does not work for me. As I understand, you mean this one should work. I provide only last, final image ```docker FROM base AS final # Install libwkhtmltox dependencies RUN apt-get update \ && apt-get install -y --no-install-recommends \ zlib1g \ fontconfig \ libfreetype6 \ libx11-6 \ libxext6 \ libxrender1 \ && curl -o /usr/lib/libwkhtmltox.so \ --location \ https://github.com/username_1/DinkToPdf/raw/v1.0.8/v0.12.4/64%20bit/libwkhtmltox.so WORKDIR /app COPY --from=publish /app . ENTRYPOINT ["dotnet", "API.dll"] ``` The application tries to load the library from '/app/wkhtmltox/libwkhtmltox', but fails. <issue_comment>username_3: @username_5, the library is being downloaded by `curl` to `/usr/lib/libwkhtmltox.so`, so the application will never find it in `/app/wkhtmltox/libwkhtmltox` <issue_comment>username_5: @username_3 Thank you for your help, I just forgot to answer that it is ok and it works. Currently, requirenmets for pdf files were changed (need to support different headers for pages) and now I use .net core node services and nodejs html-pdf utility. I was wonder how fast it works. <issue_comment>username_6: this helped to me <issue_comment>username_7: And how do you using it in the code? I am running .NET Core 2.0 and i registered converter as: `services.AddSingleton(typeof(IConverter), new SynchronizedConverter(new PdfTools()));` and i am using it as _converter.Convert(MyDoc), injected as IConverter from constructor. I tried to install all the suggestions from your comments, still unable to load dll in docker container. <issue_comment>username_8: Hi, I am using Ubuntu 18.04 LTS. Copied the library libwkhtmltox in the root folder of the MAIN project. In the image you can see that it is found there. Any idea ? ![image](https://user-images.githubusercontent.com/42044983/61211650-6eaa3d80-a700-11e9-9785-8c06d512c9ee.png) <issue_comment>username_9: Helo Guys, <issue_comment>username_10: For those still researching, my problem was that I was including all 3 libraries (.dll / .so / .dylib) in my .Net Core 5 WebAPI distributable on the AWS EC2 instance. Deleting 2 of them, just leaving the .so, and installing wkpdftohtml on the instance, solved the problem. <issue_comment>username_11: thanks man, you save my day :) <issue_comment>username_12: can you show me?
{'fraction_non_alphanumeric': 0.10083218873717378, 'fraction_numerical': 0.0938030217338612, 'mean_word_length': 3.599777034559643, 'pattern_counts': {'":': 0, '<': 33, '<?xml version=': 0, '>': 54, 'https://': 9, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 2, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '7406920', 'n_tokens_mistral': 5242, 'n_tokens_neox': 4392, 'n_words': 1313}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Deprecate multiple test case classes in a single file username_0: This is only true for files included directly as in: ```bash $ phpunit path/to/FooTest.php ``` but it is still possible to have multiple and randomly named test classes in that `FooTest.php` when invoked via ```bash $ phpunit path/to/ ``` This is because invoking with a directory (or suite from `config.xml`) the `StandardSuiteLoader` is not used. Proposal: I would create a PR that deprecates this behavior and prints out warnings if the file has more than one test case class in it or the test case class is not named like the filename as it is already behaving when invoking with a filename. Very curious on your opinion. /Flo <issue_comment>username_1: Go ahead. Thanks!<issue_closed>
{'fraction_non_alphanumeric': 0.05947955390334572, 'fraction_numerical': 0.0024783147459727386, 'mean_word_length': 4.315789473684211, 'pattern_counts': {'":': 0, '<': 4, '<?xml version=': 0, '>': 4, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 1}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '20312590', 'n_tokens_mistral': 231, 'n_tokens_neox': 214, 'n_words': 119}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: How to use this?can you make more useful demos? username_0: <issue_comment>username_1: Have you not seen the demos folder? - https://github.com/username_3/vue-smooth-dnd/tree/master/demo/pages You can run these locally and play with them, or view online: - https://username_3.github.io/vue-smooth-dnd 15 demos to play with is great. <issue_comment>username_2: First let me thank you for sharing this. It really is a great package. From my perspective, I think what he means is that it is a bit difficult to figure out what to do to create a simple DnD (for me, between two columns) from the demo code. It took me a while to decipher your examples to understand what the bare minimum was to get it to work due to the complexity of them and lack of comments within the code. The code under Usage in the read.me was missing the get_child_payload() so that coupled with looking at the utils file, I finally figured it out. While I understand the need to highlight all the features and you're obviously quite skilled in your coding, but maybe a really dumbed down version with comments for us newbies would be helpful. :) Again though, thanks for taking the time to provide this. Really nice job. 👍 <issue_comment>username_3: Thanks for your comment. Happy to hear that you liked the package. There are two functions used in all demo files namely applyDrag and generateItems. I moved them and imported from a utility module not to rewrite them for all demo pages. I think they are pretty easy to track. You wont probably need generateItems in your apps since it just produces a random dummy array of objects. You may want to use to applyDrag function in same cases. But it's up to you how to manipulate you component state. Anyway if I find more time in the future I may improve demo app even possible to migrate them to codesandbox.io to make them easy to play with. Cheers!<issue_closed>
{'fraction_non_alphanumeric': 0.040679711637487126, 'fraction_numerical': 0.004119464469618949, 'mean_word_length': 4.0863874345549736, 'pattern_counts': {'":': 0, '<': 6, '<?xml version=': 0, '>': 6, 'https://': 2, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '12481351', 'n_tokens_mistral': 502, 'n_tokens_neox': 481, 'n_words': 316}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Segmentation Settings UI - Phase 2 #1458: Add segment part 2 (jump to frame and visibility toggle) username_0: #1458 - hide/show button for each segment on the segmentation. - Clicking on the left of the panel should jump to the ticket. ### PR Checklist - [x] Brief description of changes - [x] Links to any relevant issues - [ ] Required status checks are passing - [ ] User cases if changes impact the user's experience - [ ] `@mention` a maintainer to request a review <!-- Links --> <!-- prettier-ignore-start --> [blog]: https://circleci.com/blog/triggering-trusted-ci-jobs-on-untrusted-forks/ [script]: https://github.com/jklukas/git-push-fork-to-upstream-branch <!-- prettier-ignore-end --> <issue_comment>username_1: Hey, the jump to segment works really well on [this PET example]( https://deploy-preview-1643--ohif.netlify.app/pwa/viewer/1.3.6.1.4.1.14519.5.2.1.2744.7002.373729467545468642229382466905). But on [this prostate example](https://deploy-preview-1643--ohif.netlify.app/pwa/viewer/1.3.6.1.4.1.14519.5.2.1.7311.5101.170561193612723093192571245493) it doesn't seem to work, scroll away from the segments and try to jump to them. Is it having an issue because its out of plane? Cornerstone is 2D so I doubt this is the case, could you look into this? Also the second consequtive time you click on a segment on the sidepanel it doesn't jump to the segment. Follow these steps: -> click on a segment in the side panel, the cornerstone window will jump to this segment. -> scroll away so the segment is out of view. -> click on the segment in the side panel again, the segment doesn't jump. We want to always jump, even if the segment is "active", as you may want to recentre it. <issue_comment>username_0: @username_1 About the consecutive time, you click the seg to jump after moving the stack, it's because of a known bug in OHIF, it happens in RT and measurements as well, the viewport specific data is set but if the user moves the stack, the state continues the same so it doesn't cause a rerender in the second click. I had a solution merged in react-cornerstone-viewport that would call a callback after the user stops moving the stack (debouncing) so we could update the state. I'll check with @dannyrb about it. About the prostate example, the segmentsOnLabelmap2D are all the same which causes the closest number to be 9 for all segments. Maybe something wrong with the way I'm getting the frameindex? Checking other comments now. <issue_comment>username_1: Ahh interesting, this probably means inside the SEG they are encoding each frame for each segment even if its empty. This is slightly annoying, checking each frame for occupancy would be slow, which is why we do the `segmentsOnLabelmap2D` thing in the first place. I liked this approach of checking which frames are encoded, as It'll also make the VTK integration easier as we just need to find the central frame, and then search only that frame for voxels to set the crosshairs for all viewports. TLDR: This is a "poorly encoded", but completely valid SEG. Check if this affects any of the IDC MVP datasets, if not a better way of finding the correct slice may be out of scope for this ticket. We could do one search through each frame that claims it contains the segment and then cache the central frame that actually does, but this could be very slow for large CTs if the data is encoded similarly to that prostate SEG. cc @username_2 @username_3 for opinions on the above. <issue_comment>username_2: ❤️ the progress on this! Regarding empty frames, couldn't we just record in some kind of meta-metadata while decoding the SEG? <issue_comment>username_3: I wouldn't call it "poorly encoded". Keeping empty frames has its advantages, since there are systems that cannot deal with skipped empty frames. Also, when a SEG with empty frames is converted to a NIfTI or like, it will have the geometry which is matching the geometry of the image, which is really important for those AI researchers that ignore origin/spacing/orientation information (and there are such AI researchers). TLDR: OHIF Viewer should be robust enough to handle this kind of SEGs. <issue_comment>username_3: As aside, I was not aware of existence of SEGs in the prostateX collection - I am glad you came across it! <issue_comment>username_3: I tried and failed to figure out how the jump to segment works on that PET example ... :-\ Can you elaborate? <issue_comment>username_0: @username_3 you can try this flow again, I just pushed the fix, a check was missing in jumpToImage command. This specific PET just screamed the need for it. <issue_comment>username_3: @username_0 thank you, it works for me now - very nice! <issue_comment>username_3: @username_0 what is the source of the SEG for the ProstateX collection? According to the information on the TCIA page for that collection, the only modality it has is MR (which is consistent with what I thought, and that's why I was surprised when you mentioned SEG in the context of that collection). <issue_comment>username_1: Hey, yeah we could do that, currently dcmjs just looks which frames are present and flags these as occupied, not if they are empty. We can iterate through each frame and check its actually occupied by a single positive bit, and only add these ones to the list. <issue_comment>username_1: I asked Erik about this, he doesn't remember when he uploaded these, but thinks the SEGs were made later and uploaded as tests, i.e. they aren't on TCIA. <issue_comment>username_1: To dcmjs.
{'fraction_non_alphanumeric': 0.05824723904524403, 'fraction_numerical': 0.026006412540078374, 'mean_word_length': 4.199074074074074, 'pattern_counts': {'":': 0, '<': 18, '<?xml version=': 0, '>': 21, 'https://': 4, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 4, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '21190596', 'n_tokens_mistral': 1635, 'n_tokens_neox': 1491, 'n_words': 842}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Continuous Integration for MacOS username_0: The cirrus-ci has also some support for MacOS, but I'm not sure about availability for open source, and I lack experience with MacOS. If cirrus fails, maybe there's some other CI service where it may be set up with reasonably small effort?<issue_closed>
{'fraction_non_alphanumeric': 0.047619047619047616, 'fraction_numerical': 0.002976190476190476, 'mean_word_length': 5.3584905660377355, 'pattern_counts': {'":': 0, '<': 3, '<?xml version=': 0, '>': 3, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '12905757', 'n_tokens_mistral': 90, 'n_tokens_neox': 83, 'n_words': 48}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Batch scan username_0: Changed the way that the system reads from C-Series modules to using the scan engine batch APIs. This increases performance insofar as the user configures the system to read/write to all the channels of a module. <issue_comment>username_1: Bleep bloop! LabVIEW Diff Robot here with some diffs served up hot for your pull request. Notice something funny? Help fix me on [my GitHub repo.](https://github.com/LabVIEW-DCAF/buildsystem) ### Scan Engine runtime.lvclass--check_variable_type.vi.png: ![capture](https://raw.githubusercontent.com/username_1/diff-pics/master/LabVIEW-DCAF/Scan-Engine-Module/PR-56/2018-10-24/15%3A09%3A08/Scan%20Engine%20runtime.lvclass--check_variable_type.vi.png) ### Scan Engine runtime.lvclass--create_module_map.vi.png: ![capture](https://raw.githubusercontent.com/username_1/diff-pics/master/LabVIEW-DCAF/Scan-Engine-Module/PR-56/2018-10-24/15%3A09%3A08/Scan%20Engine%20runtime.lvclass--create_module_map.vi.png) ### Scan Engine runtime.lvclass--populate_index.vi.png: ![capture](https://raw.githubusercontent.com/username_1/diff-pics/master/LabVIEW-DCAF/Scan-Engine-Module/PR-56/2018-10-24/15%3A09%3A08/Scan%20Engine%20runtime.lvclass--populate_index.vi.png) ### Scan Engine runtime.lvclass--populate_io_index_array.vi.png: ![capture](https://raw.githubusercontent.com/username_1/diff-pics/master/LabVIEW-DCAF/Scan-Engine-Module/PR-56/2018-10-24/15%3A09%3A08/Scan%20Engine%20runtime.lvclass--populate_io_index_array.vi.png) ### Scan Engine runtime.lvclass--populate_map.vi.png: ![capture](https://raw.githubusercontent.com/username_1/diff-pics/master/LabVIEW-DCAF/Scan-Engine-Module/PR-56/2018-10-24/15%3A09%3A08/Scan%20Engine%20runtime.lvclass--populate_map.vi.png) ### Scan Engine runtime.lvclass--populate_module.vi.png: ![capture](https://raw.githubusercontent.com/username_1/diff-pics/master/LabVIEW-DCAF/Scan-Engine-Module/PR-56/2018-10-24/15%3A09%3A08/Scan%20Engine%20runtime.lvclass--populate_module.vi.png)
{'fraction_non_alphanumeric': 0.14484126984126985, 'fraction_numerical': 0.06944444444444445, 'mean_word_length': 5.678807947019868, 'pattern_counts': {'":': 0, '<': 3, '<?xml version=': 0, '>': 3, 'https://': 7, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '26511200', 'n_tokens_mistral': 815, 'n_tokens_neox': 720, 'n_words': 91}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Order gets stuck in pending state intermittently username_0: Order gets stuck in pending state. This happens rarely but not sure of the cause for the same. Seems like first request for certificate on a k8s cluster gets stuck in pending state. Not sure if this is right but just an observation. Deleting the order gets the order in valid state. Can anyone help me with this? <issue_comment>username_1: Have you tried this guide: https://cert-manager.io/docs/faq/troubleshooting/ ?
{'fraction_non_alphanumeric': 0.05058365758754864, 'fraction_numerical': 0.005836575875486381, 'mean_word_length': 4.8522727272727275, 'pattern_counts': {'":': 0, '<': 3, '<?xml version=': 0, '>': 3, 'https://': 1, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '10965234', 'n_tokens_mistral': 132, 'n_tokens_neox': 128, 'n_words': 74}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Is the PuOr scheme backwards? username_0: `interpolatePuOr` and `schemePuOr` ramp from orange to purple. This behavior is different from all the other examples, where the order of colors matches the order of color names in the exported property name. In addition, ColorBrewer also ramps from purple (a default for low values) to orange (a default for high values). Not sure if this was intentional or not, so thought it worthwhile to add an issue. <issue_comment>username_1: Looks like a mistake to me! Want to send a PR? <issue_comment>username_0: Opened a PR, closing this issue.<issue_closed>
{'fraction_non_alphanumeric': 0.05396825396825397, 'fraction_numerical': 0.004761904761904762, 'mean_word_length': 5.373737373737374, 'pattern_counts': {'":': 0, '<': 5, '<?xml version=': 0, '>': 5, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '6620403', 'n_tokens_mistral': 167, 'n_tokens_neox': 161, 'n_words': 94}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: upgrade consul to 0.7.1 to support snapshots username_0: This adds the support for command "consul snapshot save ..." and "consul snapshot restore ..." I have another PR incoming that implements bbr using this mechanism. Please upload https://releases.hashicorp.com/consul/0.7.1/consul_0.7.1_linux_amd64.zip in your S3 store so that this release can use it before you accept this PR. <issue_comment>username_1: Closing pull request since consul was updated 1.9.3
{'fraction_non_alphanumeric': 0.07569721115537849, 'fraction_numerical': 0.03386454183266932, 'mean_word_length': 4.467391304347826, 'pattern_counts': {'":': 0, '<': 3, '<?xml version=': 0, '>': 3, 'https://': 1, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '18351185', 'n_tokens_mistral': 153, 'n_tokens_neox': 143, 'n_words': 61}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Add uxr_run_session_until_confirm_delivery_one_stream API username_0: Should we flash only one stream with this new API? TODO: - [ ] change all `flash` to `flush` <issue_comment>username_1: Build status: * Linux [![Build Status](http://jenkins.eprosima.com:8080/buildStatus/icon?job=Micro-XRCE-DDS-Client%20Manual%20Linux&build=479)](http://jenkins.eprosima.com:8080/job/Micro-XRCE-DDS-Client%20Manual%20Linux/479/) * Windows [![Build Status](http://jenkins.eprosima.com:8080/buildStatus/icon?job=Micro-XRCE-DDS-Client%20Manual%20Windows&build=495)](http://jenkins.eprosima.com:8080/job/Micro-XRCE-DDS-Client%20Manual%20Windows/495/) <issue_comment>username_1: Build status: * Linux [![Build Status](http://jenkins.eprosima.com:8080/buildStatus/icon?job=Micro-XRCE-DDS-Client%20Manual%20Linux&build=481)](http://jenkins.eprosima.com:8080/job/Micro-XRCE-DDS-Client%20Manual%20Linux/481/) * Windows [![Build Status](http://jenkins.eprosima.com:8080/buildStatus/icon?job=Micro-XRCE-DDS-Client%20Manual%20Windows&build=497)](http://jenkins.eprosima.com:8080/job/Micro-XRCE-DDS-Client%20Manual%20Windows/497/) <issue_comment>username_1: Build status: * Linux [![Build Status](http://jenkins.eprosima.com:8080/buildStatus/icon?job=Micro-XRCE-DDS-Client%20Manual%20Linux&build=482)](http://jenkins.eprosima.com:8080/job/Micro-XRCE-DDS-Client%20Manual%20Linux/482/) * Windows [![Build Status](http://jenkins.eprosima.com:8080/buildStatus/icon?job=Micro-XRCE-DDS-Client%20Manual%20Windows&build=498)](http://jenkins.eprosima.com:8080/job/Micro-XRCE-DDS-Client%20Manual%20Windows/498/) <issue_comment>username_1: Build status: * Linux [![Build Status](http://jenkins.eprosima.com:8080/buildStatus/icon?job=Micro-XRCE-DDS-Client%20Manual%20Linux&build=483)](http://jenkins.eprosima.com:8080/job/Micro-XRCE-DDS-Client%20Manual%20Linux/483/) * Windows [![Build Status](http://jenkins.eprosima.com:8080/buildStatus/icon?job=Micro-XRCE-DDS-Client%20Manual%20Windows&build=499)](http://jenkins.eprosima.com:8080/job/Micro-XRCE-DDS-Client%20Manual%20Windows/499/) <issue_comment>username_1: Build status: * Linux [![Build Status](http://jenkins.eprosima.com:8080/buildStatus/icon?job=Micro-XRCE-DDS-Client%20Manual%20Linux&build=484)](http://jenkins.eprosima.com:8080/job/Micro-XRCE-DDS-Client%20Manual%20Linux/484/) * Windows [![Build Status](http://jenkins.eprosima.com:8080/buildStatus/icon?job=Micro-XRCE-DDS-Client%20Manual%20Windows&build=500)](http://jenkins.eprosima.com:8080/job/Micro-XRCE-DDS-Client%20Manual%20Windows/500/) <issue_comment>username_0: Testing, do not merge
{'fraction_non_alphanumeric': 0.17619961612284069, 'fraction_numerical': 0.08714011516314779, 'mean_word_length': 8.017301038062284, 'pattern_counts': {'":': 0, '<': 8, '<?xml version=': 0, '>': 8, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '12508416', 'n_tokens_mistral': 1159, 'n_tokens_neox': 976, 'n_words': 71}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: remove gradle deprecation warning username_0: Deprecated Gradle features were used in this build, making it incompatible with Gradle 5.0.<issue_closed> <issue_comment>username_0: somehow it's gone :-) <issue_comment>username_0: Deprecated Gradle features were used in this build, making it incompatible with Gradle 5.0. <issue_comment>username_0: The following annotation processors were detected on the compile classpath: 'org.apache.logging.log4j.core.config.plugins.processor.PluginProcessor'. Detecting annotation processors on the compile classpath is deprecated and Gradle 5.0 will ignore them. Please add them to the annotation processor path instead. If you did not intend to use annotation processors, you can use the '-proc:none' compiler argument to ignore them. BUILD SUCCESSFUL in 3s <issue_comment>username_0: fixed by moving log4j into annotationProcesssor group within build.gradle<issue_closed>
{'fraction_non_alphanumeric': 0.055612770339855816, 'fraction_numerical': 0.014418125643666324, 'mean_word_length': 5.311688311688312, 'pattern_counts': {'":': 0, '<': 8, '<?xml version=': 0, '>': 8, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '12699950', 'n_tokens_mistral': 267, 'n_tokens_neox': 241, 'n_words': 108}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Users admin management front-end username_0: <issue_comment>username_1: Requested features: - Table of all the users that have signed up - Filter by those who have verified their email - Show the users auth level, name and email in each row - Can click on a row/button to edit a specific user, change their email verified status + auth level <issue_comment>username_0: All of the requested features are now implemented. Closing.<issue_closed>
{'fraction_non_alphanumeric': 0.05371900826446281, 'fraction_numerical': 0.006198347107438017, 'mean_word_length': 4.773809523809524, 'pattern_counts': {'":': 0, '<': 5, '<?xml version=': 0, '>': 5, 'https://': 0, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '2134929', 'n_tokens_mistral': 127, 'n_tokens_neox': 125, 'n_words': 64}
starcoder-github-issues-filtered-structured
<issue_start><issue_comment>Title: Package git-unix.1.11.4 username_0: ### `git-unix.1.11.4` Unix backend for the Git protocol(s) The library comes with a command-line tool called `ogit` which shares a similar interface with `git`, but where all operations are mapped to the API exposed `ocaml-git` (and hence using only OCaml code). --- * Homepage: https://github.com/mirage/ocaml-git * Source repo: https://github.com/mirage/ocaml-git.git * Bug tracker: https://github.com/mirage/ocaml-git/issues --- --- ### 1.11.4 (2018-01-03) - support cohttp 1.0 (#249, @username_0) :camel: Pull-request generated by opam-publish v0.3.5 <issue_comment>username_1: ##### :white_check_mark: All lint checks passed <small>c64879a2dd12811112bef56dfe8b3e864fc40c71</small> * These packages passed lint tests: `git-unix.1.11.4` --- ##### :white_check_mark: Installability check (8120 &rarr; 8121) * new installable packages (1): `git-unix.1.11.4` <issue_comment>username_0: This needs #11153 to be merged first
{'fraction_non_alphanumeric': 0.14285714285714285, 'fraction_numerical': 0.07837301587301587, 'mean_word_length': 4.367021276595745, 'pattern_counts': {'":': 0, '<': 6, '<?xml version=': 0, '>': 6, 'https://': 3, 'lorem ipsum': 0, 'www.': 0, 'xml': 0}, 'pii_count': 0, 'substrings_counts': 0, 'word_list_counts': {'cursed_substrings.json': 0, 'profanity_word_list.json': 0, 'sexual_word_list.json': 0, 'zh_pornsignals.json': 0}}
{'dir': 'github-issues-filtered-structured', 'id': '198669', 'n_tokens_mistral': 408, 'n_tokens_neox': 366, 'n_words': 97}