title
stringlengths 2
169
| diff
stringlengths 235
19.5k
| body
stringlengths 0
30.5k
| url
stringlengths 48
84
| created_at
stringlengths 20
20
| closed_at
stringlengths 20
20
| merged_at
stringlengths 20
20
| updated_at
stringlengths 20
20
| diff_len
float64 101
3.99k
| repo_name
stringclasses 83
values | __index_level_0__
int64 15
52.7k
|
---|---|---|---|---|---|---|---|---|---|---|
DOC add example of DataFrame.index | diff --git a/ci/code_checks.sh b/ci/code_checks.sh
index c046d55d80b49..55618590071b5 100755
--- a/ci/code_checks.sh
+++ b/ci/code_checks.sh
@@ -532,7 +532,6 @@ if [[ -z "$CHECK" || "$CHECK" == "docstrings" ]]; then
pandas.api.extensions.ExtensionArray.ndim \
pandas.api.extensions.ExtensionArray.shape \
pandas.api.extensions.ExtensionArray.tolist \
- pandas.DataFrame.index \
pandas.DataFrame.columns \
pandas.DataFrame.__iter__ \
pandas.DataFrame.keys \
diff --git a/pandas/core/frame.py b/pandas/core/frame.py
index bd298b8d723b8..abe62b475a759 100644
--- a/pandas/core/frame.py
+++ b/pandas/core/frame.py
@@ -11768,7 +11768,50 @@ def isin(self, values: Series | DataFrame | Sequence | Mapping) -> DataFrame:
_info_axis_name: Literal["columns"] = "columns"
index = properties.AxisProperty(
- axis=1, doc="The index (row labels) of the DataFrame."
+ axis=1,
+ doc="""
+ The index (row labels) of the DataFrame.
+
+ The index of a DataFrame is a series of labels that identify each row.
+ The labels can be integers, strings, or any other hashable type. The index
+ is used for label-based access and alignment, and can be accessed or
+ modified using this attribute.
+
+ Returns
+ -------
+ pandas.Index
+ The index labels of the DataFrame.
+
+ See Also
+ --------
+ DataFrame.columns : The column labels of the DataFrame.
+ DataFrame.to_numpy : Convert the DataFrame to a NumPy array.
+
+ Examples
+ --------
+ >>> df = pd.DataFrame({'Name': ['Alice', 'Bob', 'Aritra'],
+ ... 'Age': [25, 30, 35],
+ ... 'Location': ['Seattle', 'New York', 'Kona']},
+ ... index=([10, 20, 30]))
+ >>> df.index
+ Index([10, 20, 30], dtype='int64')
+
+ In this example, we create a DataFrame with 3 rows and 3 columns,
+ including Name, Age, and Location information. We set the index labels to
+ be the integers 10, 20, and 30. We then access the `index` attribute of the
+ DataFrame, which returns an `Index` object containing the index labels.
+
+ >>> df.index = [100, 200, 300]
+ >>> df
+ Name Age Location
+ 100 Alice 25 Seattle
+ 200 Bob 30 New York
+ 300 Aritra 35 Kona
+
+ In this example, we modify the index labels of the DataFrame by assigning
+ a new list of labels to the `index` attribute. The DataFrame is then
+ updated with the new labels, and the output shows the modified DataFrame.
+ """,
)
columns = properties.AxisProperty(axis=0, doc="The column labels of the DataFrame.")
| DOC add example of DataFrame.index
- [ ] closes #xxxx (Replace xxxx with the GitHub issue number)
- [ ] [Tests added and passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#writing-tests) if fixing a bug or adding a new feature
- [ ] All [code checks passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#pre-commit).
- [ ] Added [type annotations](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#type-hints) to new arguments/methods/functions.
- [ ] Added an entry in the latest `doc/source/whatsnew/vX.X.X.rst` file if fixing a bug or adding a new feature.
| https://api.github.com/repos/pandas-dev/pandas/pulls/52835 | 2023-04-21T18:41:54Z | 2023-04-23T19:13:44Z | 2023-04-23T19:13:44Z | 2023-04-24T13:41:58Z | 721 | pandas-dev/pandas | 45,545 |
Control filter list | diff --git a/tests-ui/tests/widgetInputs.test.js b/tests-ui/tests/widgetInputs.test.js
index 022e549266..e1873105ac 100644
--- a/tests-ui/tests/widgetInputs.test.js
+++ b/tests-ui/tests/widgetInputs.test.js
@@ -14,10 +14,10 @@ const lg = require("../utils/litegraph");
* @param { InstanceType<Ez["EzGraph"]> } graph
* @param { InstanceType<Ez["EzInput"]> } input
* @param { string } widgetType
- * @param { boolean } hasControlWidget
+ * @param { number } controlWidgetCount
* @returns
*/
-async function connectPrimitiveAndReload(ez, graph, input, widgetType, hasControlWidget) {
+async function connectPrimitiveAndReload(ez, graph, input, widgetType, controlWidgetCount = 0) {
// Connect to primitive and ensure its still connected after
let primitive = ez.PrimitiveNode();
primitive.outputs[0].connectTo(input);
@@ -33,13 +33,17 @@ async function connectPrimitiveAndReload(ez, graph, input, widgetType, hasContro
expect(valueWidget.widget.type).toBe(widgetType);
// Check if control_after_generate should be added
- if (hasControlWidget) {
+ if (controlWidgetCount) {
const controlWidget = primitive.widgets.control_after_generate;
expect(controlWidget.widget.type).toBe("combo");
+ if(widgetType === "combo") {
+ const filterWidget = primitive.widgets.control_filter_list;
+ expect(filterWidget.widget.type).toBe("string");
+ }
}
// Ensure we dont have other widgets
- expect(primitive.node.widgets).toHaveLength(1 + +!!hasControlWidget);
+ expect(primitive.node.widgets).toHaveLength(1 + controlWidgetCount);
});
return primitive;
@@ -55,8 +59,8 @@ describe("widget inputs", () => {
});
[
- { name: "int", type: "INT", widget: "number", control: true },
- { name: "float", type: "FLOAT", widget: "number", control: true },
+ { name: "int", type: "INT", widget: "number", control: 1 },
+ { name: "float", type: "FLOAT", widget: "number", control: 1 },
{ name: "text", type: "STRING" },
{
name: "customtext",
@@ -64,7 +68,7 @@ describe("widget inputs", () => {
opt: { multiline: true },
},
{ name: "toggle", type: "BOOLEAN" },
- { name: "combo", type: ["a", "b", "c"], control: true },
+ { name: "combo", type: ["a", "b", "c"], control: 2 },
].forEach((c) => {
test(`widget conversion + primitive works on ${c.name}`, async () => {
const { ez, graph } = await start({
@@ -106,7 +110,7 @@ describe("widget inputs", () => {
n.widgets.ckpt_name.convertToInput();
expect(n.inputs.length).toEqual(inputCount + 1);
- const primitive = await connectPrimitiveAndReload(ez, graph, n.inputs.ckpt_name, "combo", true);
+ const primitive = await connectPrimitiveAndReload(ez, graph, n.inputs.ckpt_name, "combo", 2);
// Disconnect & reconnect
primitive.outputs[0].connections[0].disconnect();
@@ -226,7 +230,7 @@ describe("widget inputs", () => {
// Reload and ensure it still only has 1 converted widget
if (!assertNotNullOrUndefined(input)) return;
- await connectPrimitiveAndReload(ez, graph, input, "number", true);
+ await connectPrimitiveAndReload(ez, graph, input, "number", 1);
n = graph.find(n);
expect(n.widgets).toHaveLength(1);
w = n.widgets.example;
@@ -258,7 +262,7 @@ describe("widget inputs", () => {
// Reload and ensure it still only has 1 converted widget
if (assertNotNullOrUndefined(input)) {
- await connectPrimitiveAndReload(ez, graph, input, "number", true);
+ await connectPrimitiveAndReload(ez, graph, input, "number", 1);
n = graph.find(n);
expect(n.widgets).toHaveLength(1);
expect(n.widgets.example.isConvertedToInput).toBeTruthy();
@@ -316,4 +320,76 @@ describe("widget inputs", () => {
n1.outputs[0].connectTo(n2.inputs[0]);
expect(() => n1.outputs[0].connectTo(n3.inputs[0])).toThrow();
});
+
+ test("combo primitive can filter list when control_after_generate called", async () => {
+ const { ez } = await start({
+ mockNodeDefs: {
+ ...makeNodeDef("TestNode1", { example: [["A", "B", "C", "D", "AA", "BB", "CC", "DD", "AAA", "BBB"], {}] }),
+ },
+ });
+
+ const n1 = ez.TestNode1();
+ n1.widgets.example.convertToInput();
+ const p = ez.PrimitiveNode()
+ p.outputs[0].connectTo(n1.inputs[0]);
+
+ const value = p.widgets.value;
+ const control = p.widgets.control_after_generate.widget;
+ const filter = p.widgets.control_filter_list;
+
+ expect(p.widgets.length).toBe(3);
+ control.value = "increment";
+ expect(value.value).toBe("A");
+
+ // Manually trigger after queue when set to increment
+ control["afterQueued"]();
+ expect(value.value).toBe("B");
+
+ // Filter to items containing D
+ filter.value = "D";
+ control["afterQueued"]();
+ expect(value.value).toBe("D");
+ control["afterQueued"]();
+ expect(value.value).toBe("DD");
+
+ // Check decrement
+ value.value = "BBB";
+ control.value = "decrement";
+ filter.value = "B";
+ control["afterQueued"]();
+ expect(value.value).toBe("BB");
+ control["afterQueued"]();
+ expect(value.value).toBe("B");
+
+ // Check regex works
+ value.value = "BBB";
+ filter.value = "/[AB]|^C$/";
+ control["afterQueued"]();
+ expect(value.value).toBe("AAA");
+ control["afterQueued"]();
+ expect(value.value).toBe("BB");
+ control["afterQueued"]();
+ expect(value.value).toBe("AA");
+ control["afterQueued"]();
+ expect(value.value).toBe("C");
+ control["afterQueued"]();
+ expect(value.value).toBe("B");
+ control["afterQueued"]();
+ expect(value.value).toBe("A");
+
+ // Check random
+ control.value = "randomize";
+ filter.value = "/D/";
+ for(let i = 0; i < 100; i++) {
+ control["afterQueued"]();
+ expect(value.value === "D" || value.value === "DD").toBeTruthy();
+ }
+
+ // Ensure it doesnt apply when fixed
+ control.value = "fixed";
+ value.value = "B";
+ filter.value = "C";
+ control["afterQueued"]();
+ expect(value.value).toBe("B");
+ });
});
diff --git a/web/extensions/core/widgetInputs.js b/web/extensions/core/widgetInputs.js
index bad3ac3a74..5c8fbc9b2d 100644
--- a/web/extensions/core/widgetInputs.js
+++ b/web/extensions/core/widgetInputs.js
@@ -1,4 +1,4 @@
-import { ComfyWidgets, addValueControlWidget } from "../../scripts/widgets.js";
+import { ComfyWidgets, addValueControlWidgets } from "../../scripts/widgets.js";
import { app } from "../../scripts/app.js";
const CONVERTED_TYPE = "converted-widget";
@@ -467,7 +467,11 @@ app.registerExtension({
if (!control_value) {
control_value = "fixed";
}
- addValueControlWidget(this, widget, control_value);
+ addValueControlWidgets(this, widget, control_value);
+ let filter = this.widgets_values?.[2];
+ if(filter && this.widgets.length === 3) {
+ this.widgets[2].value = filter;
+ }
}
// When our value changes, update other widgets to reflect our changes
diff --git a/web/scripts/widgets.js b/web/scripts/widgets.js
index ccddc0bc44..fbc1d0fc32 100644
--- a/web/scripts/widgets.js
+++ b/web/scripts/widgets.js
@@ -24,17 +24,58 @@ function getNumberDefaults(inputData, defaultStep, precision, enable_rounding) {
}
export function addValueControlWidget(node, targetWidget, defaultValue = "randomize", values) {
- const valueControl = node.addWidget("combo", "control_after_generate", defaultValue, function (v) { }, {
+ const widgets = addValueControlWidgets(node, targetWidget, defaultValue, values, {
+ addFilterList: false,
+ });
+ return widgets[0];
+}
+
+export function addValueControlWidgets(node, targetWidget, defaultValue = "randomize", values, options) {
+ if (!options) options = {};
+
+ const widgets = [];
+ const valueControl = node.addWidget("combo", "control_after_generate", defaultValue, function (v) { }, {
values: ["fixed", "increment", "decrement", "randomize"],
serialize: false, // Don't include this in prompt.
});
- valueControl.afterQueued = () => {
+ widgets.push(valueControl);
+
+ const isCombo = targetWidget.type === "combo";
+ let comboFilter;
+ if (isCombo && options.addFilterList !== false) {
+ comboFilter = node.addWidget("string", "control_filter_list", "", function (v) {}, {
+ serialize: false, // Don't include this in prompt.
+ });
+ widgets.push(comboFilter);
+ }
+ valueControl.afterQueued = () => {
var v = valueControl.value;
- if (targetWidget.type == "combo" && v !== "fixed") {
- let current_index = targetWidget.options.values.indexOf(targetWidget.value);
- let current_length = targetWidget.options.values.length;
+ if (isCombo && v !== "fixed") {
+ let values = targetWidget.options.values;
+ const filter = comboFilter?.value;
+ if (filter) {
+ let check;
+ if (filter.startsWith("/") && filter.endsWith("/")) {
+ try {
+ const regex = new RegExp(filter.substring(1, filter.length - 1));
+ check = (item) => regex.test(item);
+ } catch (error) {
+ console.error("Error constructing RegExp filter for node " + node.id, filter, error);
+ }
+ }
+ if (!check) {
+ const lower = filter.toLocaleLowerCase();
+ check = (item) => item.toLocaleLowerCase().includes(lower);
+ }
+ values = values.filter(item => check(item));
+ if (!values.length && targetWidget.options.values.length) {
+ console.warn("Filter for node " + node.id + " has filtered out all items", filter);
+ }
+ }
+ let current_index = values.indexOf(targetWidget.value);
+ let current_length = values.length;
switch (v) {
case "increment":
@@ -51,7 +92,7 @@ export function addValueControlWidget(node, targetWidget, defaultValue = "random
current_index = Math.max(0, current_index);
current_index = Math.min(current_length - 1, current_index);
if (current_index >= 0) {
- let value = targetWidget.options.values[current_index];
+ let value = values[current_index];
targetWidget.value = value;
targetWidget.callback(value);
}
@@ -88,7 +129,8 @@ export function addValueControlWidget(node, targetWidget, defaultValue = "random
targetWidget.callback(targetWidget.value);
}
}
- return valueControl;
+
+ return widgets;
};
function seedWidget(node, inputName, inputData, app) {
| Allows filtering of the items in a COMBO primitive (e.g. LoadImage images list) when using control_after_generate.
You can also use regex to filter the items by wrapping the expression in `/`s e.g. `/(sdxl|sd15)/`
Maintains existing behavior of `addValueControlWidget` for extensions currently using it, adding `addValueControlWidgets` which returns an array of added widgets. | https://api.github.com/repos/comfyanonymous/ComfyUI/pulls/2009 | 2023-11-20T21:48:41Z | 2023-11-22T17:52:20Z | 2023-11-22T17:52:20Z | 2023-11-23T08:37:52Z | 2,804 | comfyanonymous/ComfyUI | 17,901 |
C.40: Fixed a couple of typos. | diff --git a/CppCoreGuidelines.md b/CppCoreGuidelines.md
index 0a4a6c410..bf1bf47b7 100644
--- a/CppCoreGuidelines.md
+++ b/CppCoreGuidelines.md
@@ -3905,7 +3905,7 @@ That's what constructors are for.
int d, m, y;
};
-It is often a good idea to express the invariant as an `Ensure` on the constructor.
+It is often a good idea to express the invariant as an `Ensures` on the constructor.
##### Note
@@ -3941,7 +3941,7 @@ Also, the default for `int` would be better done as a [member initializer](#Rc-i
##### Enforcement
-* Flag classes with user-define copy operations but no constructor (a user-defined copy is a good indicator that the class has an invariant)
+* Flag classes with user-defined copy operations but no constructor (a user-defined copy is a good indicator that the class has an invariant)
### <a name="Rc-complete"></a> C.41: A constructor should create a fully initialized object
| https://api.github.com/repos/isocpp/CppCoreGuidelines/pulls/471 | 2015-12-18T16:36:52Z | 2015-12-18T16:51:25Z | 2015-12-18T16:51:25Z | 2015-12-19T03:20:50Z | 251 | isocpp/CppCoreGuidelines | 15,485 |
|
[3.8] bpo-41132: Use pymalloc allocator in the f-string parser (GH-21173) | diff --git a/Python/ast.c b/Python/ast.c
index 0a999fcca43a8e..5efb690c299cac 100644
--- a/Python/ast.c
+++ b/Python/ast.c
@@ -4898,7 +4898,7 @@ fstring_compile_expr(const char *expr_start, const char *expr_end,
len = expr_end - expr_start;
/* Allocate 3 extra bytes: open paren, close paren, null byte. */
- str = PyMem_RawMalloc(len + 3);
+ str = PyMem_Malloc(len + 3);
if (str == NULL) {
PyErr_NoMemory();
return NULL;
@@ -4914,7 +4914,7 @@ fstring_compile_expr(const char *expr_start, const char *expr_end,
mod_n = PyParser_SimpleParseStringFlagsFilename(str, "<fstring>",
Py_eval_input, 0);
if (!mod_n) {
- PyMem_RawFree(str);
+ PyMem_Free(str);
return NULL;
}
/* Reuse str to find the correct column offset. */
@@ -4922,7 +4922,7 @@ fstring_compile_expr(const char *expr_start, const char *expr_end,
str[len+1] = '}';
fstring_fix_node_location(n, mod_n, str);
mod = PyAST_FromNode(mod_n, &cf, "<fstring>", c->c_arena);
- PyMem_RawFree(str);
+ PyMem_Free(str);
PyNode_Free(mod_n);
if (!mod)
return NULL;
@@ -5438,7 +5438,7 @@ ExprList_Append(ExprList *l, expr_ty exp)
Py_ssize_t i;
/* We're still using the cached data. Switch to
alloc-ing. */
- l->p = PyMem_RawMalloc(sizeof(expr_ty) * new_size);
+ l->p = PyMem_Malloc(sizeof(expr_ty) * new_size);
if (!l->p)
return -1;
/* Copy the cached data into the new buffer. */
@@ -5446,9 +5446,9 @@ ExprList_Append(ExprList *l, expr_ty exp)
l->p[i] = l->data[i];
} else {
/* Just realloc. */
- expr_ty *tmp = PyMem_RawRealloc(l->p, sizeof(expr_ty) * new_size);
+ expr_ty *tmp = PyMem_Realloc(l->p, sizeof(expr_ty) * new_size);
if (!tmp) {
- PyMem_RawFree(l->p);
+ PyMem_Free(l->p);
l->p = NULL;
return -1;
}
@@ -5476,7 +5476,7 @@ ExprList_Dealloc(ExprList *l)
/* Do nothing. */
} else {
/* We have dynamically allocated. Free the memory. */
- PyMem_RawFree(l->p);
+ PyMem_Free(l->p);
}
l->p = NULL;
l->size = -1;
|
<!--
Thanks for your contribution!
Please read this comment in its entirety. It's quite important.
# Pull Request title
It should be in the following format:
```
bpo-NNNN: Summary of the changes made
```
Where: bpo-NNNN refers to the issue number in the https://bugs.python.org.
Most PRs will require an issue number. Trivial changes, like fixing a typo, do not need an issue.
# Backport Pull Request title
If this is a backport PR (PR made against branches other than `master`),
please ensure that the PR title is in the following format:
```
[X.Y] <title from the original PR> (GH-NNNN)
```
Where: [X.Y] is the branch name, e.g. [3.6].
GH-NNNN refers to the PR number from `master`.
-->
<!-- issue-number: [bpo-41132](https://bugs.python.org/issue41132) -->
https://bugs.python.org/issue41132
<!-- /issue-number -->
Automerge-Triggered-By: @pablogsal | https://api.github.com/repos/python/cpython/pulls/21184 | 2020-06-27T18:26:32Z | 2020-06-27T18:43:42Z | 2020-06-27T18:43:42Z | 2020-06-27T18:43:44Z | 685 | python/cpython | 4,738 |
Update README.md | diff --git a/README.md b/README.md
index c31017e..b1354cc 100644
--- a/README.md
+++ b/README.md
@@ -2336,7 +2336,7 @@ nan
#### 💡 Explanation:
-`'inf'` and `'nan'` are special strings (case-insensitive), which when explicitly typecasted to `float` type, are used to represent mathematical "infinity" and "not a number" respectively.
+`'inf'` and `'nan'` are special strings (case-insensitive), which when explicitly typecast-ed to `float` type, are used to represent mathematical "infinity" and "not a number" respectively.
---
@@ -2382,7 +2382,7 @@ nan
>>> 44
```
**💡 Explanation:**
- This prank comes from [Raymond Hettinger's tweet](https://twitter.com/raymondh/status/1131103570856632321?lang=en). The space invader operator is actually just a malformatted `a -= (-1)`. Which is eqivalent to `a = a - (- 1)`. Similar for the `a += (+ 1)` case.
+ This prank comes from [Raymond Hettinger's tweet](https://twitter.com/raymondh/status/1131103570856632321?lang=en). The space invader operator is actually just a malformatted `a -= (-1)`. Which is equivalent to `a = a - (- 1)`. Similar for the `a += (+ 1)` case.
* Python uses 2 bytes for local variable storage in functions. In theory, this means that only 65536 variables can be defined in a function. However, python has a handy solution built in that can be used to store more than 2^16 variable names. The following code demonstrates what happens in the stack when more than 65536 local variables are defined (Warning: This code prints around 2^18 lines of text, so be prepared!):
```py
@@ -2390,7 +2390,7 @@ nan
exec("""
def f():
""" + """
- """.join(["X"+str(x)+"=" + str(x) for x in range(65539)]))
+ """.join(["X" + str(x) + "=" + str(x) for x in range(65539)]))
f()
| Typo corrected. | https://api.github.com/repos/satwikkansal/wtfpython/pulls/146 | 2019-10-25T08:58:55Z | 2019-10-25T14:22:53Z | 2019-10-25T14:22:53Z | 2019-10-25T14:22:58Z | 524 | satwikkansal/wtfpython | 25,728 |
improve(rules): add mercurial (hg) support | diff --git a/tests/rules/test_mercurial.py b/tests/rules/test_mercurial.py
new file mode 100644
index 000000000..08962f912
--- /dev/null
+++ b/tests/rules/test_mercurial.py
@@ -0,0 +1,134 @@
+import pytest
+
+from tests.utils import Command
+from thefuck.rules.mercurial import (
+ extract_possisiblities, match, get_new_command
+)
+
+
[email protected]('command', [
+ Command('hg base', stderr=(
+ "hg: unknown command 'base'"
+ '\n(did you mean one of blame, phase, rebase?)'
+ )),
+ Command('hg branchch', stderr=(
+ "hg: unknown command 'branchch'"
+ '\n(did you mean one of branch, branches?)'
+ )),
+ Command('hg vert', stderr=(
+ "hg: unknown command 'vert'"
+ '\n(did you mean one of revert?)'
+ )),
+ Command('hg lgo -r tip', stderr=(
+ "hg: command 're' is ambiguous:"
+ '\n(did you mean one of log?)'
+ )),
+ Command('hg rerere', stderr=(
+ "hg: unknown command 'rerere'"
+ '\n(did you mean one of revert?)'
+ )),
+ Command('hg re', stderr=(
+ "hg: command 're' is ambiguous:"
+ '\n rebase recover remove rename resolve revert'
+ )),
+ Command('hg re re', stderr=(
+ "hg: command 're' is ambiguous:"
+ '\n rebase recover remove rename resolve revert'
+ )),
+])
+def test_match(command):
+ assert match(command, None)
+
+
[email protected]('command', [
+ Command('hg', stderr=(
+ '\nMercurial Distributed SCM\n\nbasic commands:'
+ )),
+ Command('hg asdf', stderr=(
+ "hg: unknown command 'asdf'"
+ '\nMercurial Distributed SCM\n\nbasic commands:'
+ )),
+ Command('hg qwer', stderr=(
+ "hg: unknown command 'qwer'"
+ '\nMercurial Distributed SCM\n\nbasic commands:'
+ )),
+ Command('hg me', stderr=(
+ "\nabort: no repository found in './thefuck' (.hg not found)!"
+ )),
+ Command('hg reb', stderr=(
+ "\nabort: no repository found in './thefuck' (.hg not found)!"
+ )),
+ Command('hg co', stderr=(
+ "\nabort: no repository found in './thefuck' (.hg not found)!"
+ )),
+])
+def test_not_match(command):
+ assert not match(command, None)
+
+
[email protected]('command, possibilities', [
+ (Command('hg base', stderr=(
+ "hg: unknown command 'base'"
+ '\n(did you mean one of blame, phase, rebase?)'
+ )), ['blame', 'phase', 'rebase']),
+ (Command('hg branchch', stderr=(
+ "hg: unknown command 'branchch'"
+ '\n(did you mean one of branch, branches?)'
+ )), ['branch', 'branches']),
+ (Command('hg vert', stderr=(
+ "hg: unknown command 'vert'"
+ '\n(did you mean one of revert?)'
+ )), ['revert']),
+ (Command('hg lgo -r tip', stderr=(
+ "hg: command 're' is ambiguous:"
+ '\n(did you mean one of log?)'
+ )), ['log']),
+ (Command('hg rerere', stderr=(
+ "hg: unknown command 'rerere'"
+ '\n(did you mean one of revert?)'
+ )), ['revert']),
+ (Command('hg re', stderr=(
+ "hg: command 're' is ambiguous:"
+ '\n rebase recover remove rename resolve revert'
+ )), ['rebase', 'recover', 'remove', 'rename', 'resolve', 'revert']),
+ (Command('hg re re', stderr=(
+ "hg: command 're' is ambiguous:"
+ '\n rebase recover remove rename resolve revert'
+ )), ['rebase', 'recover', 'remove', 'rename', 'resolve', 'revert']),
+])
+def test_extract_possisiblities(command, possibilities):
+ assert extract_possisiblities(command) == possibilities
+
+
[email protected]('command, new_command', [
+ (Command('hg base', stderr=(
+ "hg: unknown command 'base'"
+ '\n(did you mean one of blame, phase, rebase?)'
+ )), 'hg rebase'),
+ (Command('hg branchch', stderr=(
+ "hg: unknown command 'branchch'"
+ '\n(did you mean one of branch, branches?)'
+ )), 'hg branch'),
+ (Command('hg vert', stderr=(
+ "hg: unknown command 'vert'"
+ '\n(did you mean one of revert?)'
+ )), 'hg revert'),
+ (Command('hg lgo -r tip', stderr=(
+ "hg: command 're' is ambiguous:"
+ '\n(did you mean one of log?)'
+ )), 'hg log -r tip'),
+ (Command('hg rerere', stderr=(
+ "hg: unknown command 'rerere'"
+ '\n(did you mean one of revert?)'
+ )), 'hg revert'),
+ (Command('hg re', stderr=(
+ "hg: command 're' is ambiguous:"
+ '\n rebase recover remove rename resolve revert'
+ )), 'hg rebase'),
+ (Command('hg re re', stderr=(
+ "hg: command 're' is ambiguous:"
+ '\n rebase recover remove rename resolve revert'
+ )), 'hg rebase re'),
+])
+def test_get_new_command(command, new_command):
+ assert get_new_command(command, None) == new_command
diff --git a/thefuck/rules/mercurial.py b/thefuck/rules/mercurial.py
new file mode 100644
index 000000000..934e3f1e8
--- /dev/null
+++ b/thefuck/rules/mercurial.py
@@ -0,0 +1,34 @@
+import re
+
+from difflib import get_close_matches
+
+
+def extract_possisiblities(command):
+ possib = re.findall(r'\n\(did you mean one of ([^\?]+)\?\)', command.stderr)
+ if possib:
+ return possib[0].split(', ')
+ possib = re.findall(r'\n ([^$]+)$', command.stderr)
+ if possib:
+ return possib[0].split(' ')
+ return possib
+
+
+def match(command, settings):
+ return (command.script.startswith('hg ')
+ and ('hg: unknown command' in command.stderr
+ and '(did you mean one of ' in command.stderr
+ or "hg: command '" in command.stderr
+ and "' is ambiguous:" in command.stderr
+ )
+ )
+
+
+def get_new_command(command, settings):
+ script = command.script.split(' ')
+ possisiblities = extract_possisiblities(command)
+ matches = get_close_matches(script[1], possisiblities)
+ if matches:
+ script[1] = matches[0]
+ else:
+ script[1] = possisiblities[0]
+ return ' '.join(script)
| Please review and comment!
| https://api.github.com/repos/nvbn/thefuck/pulls/281 | 2015-07-07T00:40:07Z | 2015-07-07T13:36:06Z | 2015-07-07T13:36:06Z | 2015-07-08T01:25:42Z | 1,738 | nvbn/thefuck | 30,780 |
Fix mypyc compatibility issue | diff --git a/src/black/parsing.py b/src/black/parsing.py
index 504e20be00..32cfa5239f 100644
--- a/src/black/parsing.py
+++ b/src/black/parsing.py
@@ -169,6 +169,7 @@ def stringify_ast(
yield f"{' ' * depth}{node.__class__.__name__}("
+ type_ignore_classes: Tuple[Type[Any], ...]
for field in sorted(node._fields): # noqa: F402
# TypeIgnore will not be present using pypy < 3.8, so need for this
if not (_IS_PYPY and sys.version_info < (3, 8)):
| ### Description
I can't wait for when we drop Python 2 support FWIW :)
### Checklist - did you ...
- [x] Add a CHANGELOG entry if necessary? -> n/a
- [x] Add / update tests if necessary? -> n/a
- [x] Add new / update outdated documentation? -> n/a
| https://api.github.com/repos/psf/black/pulls/2628 | 2021-11-19T03:07:24Z | 2021-11-19T03:20:45Z | 2021-11-19T03:20:45Z | 2021-11-19T03:20:47Z | 155 | psf/black | 24,458 |
Added a descriptive error if domain list includes a Unicode-encoded IDN | diff --git a/letsencrypt/configuration.py b/letsencrypt/configuration.py
index a2a54d2d062..69778f5f049 100644
--- a/letsencrypt/configuration.py
+++ b/letsencrypt/configuration.py
@@ -144,6 +144,15 @@ def _check_config_domain_sanity(domains):
if any("xn--" in d for d in domains):
raise errors.ConfigurationError(
"Punycode domains are not supported")
+
+ # Unicode
+ try:
+ for domain in domains:
+ domain.encode('ascii')
+ except UnicodeDecodeError:
+ raise errors.ConfigurationError(
+ "Internationalized domain names are not supported")
+
# FQDN checks from
# http://www.mkyong.com/regular-expressions/domain-name-regular-expression-example/
# Characters used, domain parts < 63 chars, tld > 1 < 64 chars
| The current error for IDNs passed in Unicode form is incorrect and does not describe the actual problem:
```
$ letsencrypt --manual -d example.com -d ёжикв.сайт
Requested domain is not a FQDN
```
This change checks for any domains that cannot be encoded as ASCII, and if one is present:
```
$ letsencrypt --manual -d example.com -d ёжикв.сайт
Internationalized domain names are not supported
```
| https://api.github.com/repos/certbot/certbot/pulls/1759 | 2015-12-05T06:26:33Z | 2015-12-05T07:17:21Z | 2015-12-05T07:17:21Z | 2016-05-06T19:22:00Z | 205 | certbot/certbot | 496 |
`chat_loaders` refactoring | diff --git a/libs/langchain/langchain/chat_loaders/imessage.py b/libs/langchain/langchain/chat_loaders/imessage.py
index d6c02f1e5307a2..eed0cfea3795ee 100644
--- a/libs/langchain/langchain/chat_loaders/imessage.py
+++ b/libs/langchain/langchain/chat_loaders/imessage.py
@@ -4,13 +4,13 @@
from typing import TYPE_CHECKING, Iterator, List, Optional, Union
from langchain import schema
-from langchain.chat_loaders import base as chat_loaders
+from langchain.chat_loaders.base import BaseChatLoader, ChatSession
if TYPE_CHECKING:
import sqlite3
-class IMessageChatLoader(chat_loaders.BaseChatLoader):
+class IMessageChatLoader(BaseChatLoader):
"""Load chat sessions from the `iMessage` chat.db SQLite file.
It only works on macOS when you have iMessage enabled and have the chat.db file.
@@ -18,8 +18,8 @@ class IMessageChatLoader(chat_loaders.BaseChatLoader):
The chat.db file is likely located at ~/Library/Messages/chat.db. However, your
terminal may not have permission to access this file. To resolve this, you can
copy the file to a different location, change the permissions of the file, or
- grant full disk access for your terminal emulator in System Settings > Security
- and Privacy > Full Disk Access.
+ grant full disk access for your terminal emulator
+ in System Settings > Security and Privacy > Full Disk Access.
"""
def __init__(self, path: Optional[Union[str, Path]] = None):
@@ -46,7 +46,7 @@ def __init__(self, path: Optional[Union[str, Path]] = None):
def _load_single_chat_session(
self, cursor: "sqlite3.Cursor", chat_id: int
- ) -> chat_loaders.ChatSession:
+ ) -> ChatSession:
"""
Load a single chat session from the iMessage chat.db.
@@ -83,9 +83,9 @@ def _load_single_chat_session(
)
)
- return chat_loaders.ChatSession(messages=results)
+ return ChatSession(messages=results)
- def lazy_load(self) -> Iterator[chat_loaders.ChatSession]:
+ def lazy_load(self) -> Iterator[ChatSession]:
"""
Lazy load the chat sessions from the iMessage chat.db
and yield them in the required format.
diff --git a/libs/langchain/langchain/chat_loaders/slack.py b/libs/langchain/langchain/chat_loaders/slack.py
index 0bbd503979c7c1..7c9f76c9650e83 100644
--- a/libs/langchain/langchain/chat_loaders/slack.py
+++ b/libs/langchain/langchain/chat_loaders/slack.py
@@ -6,12 +6,12 @@
from typing import Dict, Iterator, List, Union
from langchain import schema
-from langchain.chat_loaders import base as chat_loaders
+from langchain.chat_loaders.base import BaseChatLoader, ChatSession
logger = logging.getLogger(__name__)
-class SlackChatLoader(chat_loaders.BaseChatLoader):
+class SlackChatLoader(BaseChatLoader):
"""Load `Slack` conversations from a dump zip file."""
def __init__(
@@ -27,9 +27,7 @@ def __init__(
if not self.zip_path.exists():
raise FileNotFoundError(f"File {self.zip_path} not found")
- def _load_single_chat_session(
- self, messages: List[Dict]
- ) -> chat_loaders.ChatSession:
+ def _load_single_chat_session(self, messages: List[Dict]) -> ChatSession:
results: List[Union[schema.AIMessage, schema.HumanMessage]] = []
previous_sender = None
for message in messages:
@@ -62,7 +60,7 @@ def _load_single_chat_session(
)
)
previous_sender = sender
- return chat_loaders.ChatSession(messages=results)
+ return ChatSession(messages=results)
def _read_json(self, zip_file: zipfile.ZipFile, file_path: str) -> List[dict]:
"""Read JSON data from a zip subfile."""
@@ -72,7 +70,7 @@ def _read_json(self, zip_file: zipfile.ZipFile, file_path: str) -> List[dict]:
raise ValueError(f"Expected list of dictionaries, got {type(data)}")
return data
- def lazy_load(self) -> Iterator[chat_loaders.ChatSession]:
+ def lazy_load(self) -> Iterator[ChatSession]:
"""
Lazy load the chat sessions from the Slack dump file and yield them
in the required format.
diff --git a/libs/langchain/langchain/chat_loaders/telegram.py b/libs/langchain/langchain/chat_loaders/telegram.py
index 5f0bbfa3246d86..12c30014ac1fa2 100644
--- a/libs/langchain/langchain/chat_loaders/telegram.py
+++ b/libs/langchain/langchain/chat_loaders/telegram.py
@@ -7,12 +7,12 @@
from typing import Iterator, List, Union
from langchain import schema
-from langchain.chat_loaders import base as chat_loaders
+from langchain.chat_loaders.base import BaseChatLoader, ChatSession
logger = logging.getLogger(__name__)
-class TelegramChatLoader(chat_loaders.BaseChatLoader):
+class TelegramChatLoader(BaseChatLoader):
"""Load `telegram` conversations to LangChain chat messages.
To export, use the Telegram Desktop app from
@@ -35,16 +35,14 @@ def __init__(
"""
self.path = path if isinstance(path, str) else str(path)
- def _load_single_chat_session_html(
- self, file_path: str
- ) -> chat_loaders.ChatSession:
+ def _load_single_chat_session_html(self, file_path: str) -> ChatSession:
"""Load a single chat session from an HTML file.
Args:
file_path (str): Path to the HTML file.
Returns:
- chat_loaders.ChatSession: The loaded chat session.
+ ChatSession: The loaded chat session.
"""
try:
from bs4 import BeautifulSoup
@@ -81,18 +79,16 @@ def _load_single_chat_session_html(
)
previous_sender = from_name
- return chat_loaders.ChatSession(messages=results)
+ return ChatSession(messages=results)
- def _load_single_chat_session_json(
- self, file_path: str
- ) -> chat_loaders.ChatSession:
+ def _load_single_chat_session_json(self, file_path: str) -> ChatSession:
"""Load a single chat session from a JSON file.
Args:
file_path (str): Path to the JSON file.
Returns:
- chat_loaders.ChatSession: The loaded chat session.
+ ChatSession: The loaded chat session.
"""
with open(file_path, "r", encoding="utf-8") as file:
data = json.load(file)
@@ -114,7 +110,7 @@ def _load_single_chat_session_json(
)
)
- return chat_loaders.ChatSession(messages=results)
+ return ChatSession(messages=results)
def _iterate_files(self, path: str) -> Iterator[str]:
"""Iterate over files in a directory or zip file.
@@ -139,12 +135,12 @@ def _iterate_files(self, path: str) -> Iterator[str]:
with tempfile.TemporaryDirectory() as temp_dir:
yield zip_file.extract(file, path=temp_dir)
- def lazy_load(self) -> Iterator[chat_loaders.ChatSession]:
+ def lazy_load(self) -> Iterator[ChatSession]:
"""Lazy load the messages from the chat file and yield them
in as chat sessions.
Yields:
- chat_loaders.ChatSession: The loaded chat session.
+ ChatSession: The loaded chat session.
"""
for file_path in self._iterate_files(self.path):
if file_path.endswith(".html"):
diff --git a/libs/langchain/langchain/chat_loaders/whatsapp.py b/libs/langchain/langchain/chat_loaders/whatsapp.py
index e2518ab44df660..39266485e23ea3 100644
--- a/libs/langchain/langchain/chat_loaders/whatsapp.py
+++ b/libs/langchain/langchain/chat_loaders/whatsapp.py
@@ -5,13 +5,13 @@
from typing import Iterator, List, Union
from langchain import schema
-from langchain.chat_loaders import base as chat_loaders
+from langchain.chat_loaders.base import BaseChatLoader, ChatSession
from langchain.schema import messages
logger = logging.getLogger(__name__)
-class WhatsAppChatLoader(chat_loaders.BaseChatLoader):
+class WhatsAppChatLoader(BaseChatLoader):
"""Load `WhatsApp` conversations from a dump zip file or directory."""
def __init__(self, path: str):
@@ -42,7 +42,7 @@ def __init__(self, path: str):
flags=re.IGNORECASE,
)
- def _load_single_chat_session(self, file_path: str) -> chat_loaders.ChatSession:
+ def _load_single_chat_session(self, file_path: str) -> ChatSession:
"""Load a single chat session from a file.
Args:
@@ -84,7 +84,7 @@ def _load_single_chat_session(self, file_path: str) -> chat_loaders.ChatSession:
)
else:
logger.debug(f"Could not parse line: {line}")
- return chat_loaders.ChatSession(messages=results)
+ return ChatSession(messages=results)
def _iterate_files(self, path: str) -> Iterator[str]:
"""Iterate over the files in a directory or zip file.
@@ -108,7 +108,7 @@ def _iterate_files(self, path: str) -> Iterator[str]:
if file.endswith(".txt"):
yield zip_file.extract(file)
- def lazy_load(self) -> Iterator[chat_loaders.ChatSession]:
+ def lazy_load(self) -> Iterator[ChatSession]:
"""Lazy load the messages from the chat file and yield
them as chat sessions.
| Replaced unnecessary namespace renaming
`from langchain.chat_loaders import base as chat_loaders`
with
`from langchain.chat_loaders.base import BaseChatLoader, ChatSession`
and simplified correspondent types.
@eyurtsev | https://api.github.com/repos/langchain-ai/langchain/pulls/10381 | 2023-09-08T17:16:24Z | 2023-09-09T22:22:56Z | 2023-09-09T22:22:56Z | 2023-09-10T18:23:43Z | 2,233 | langchain-ai/langchain | 43,519 |
Added scibert-nli model card | diff --git a/model_cards/gsarti/scibert-nli/README.md b/model_cards/gsarti/scibert-nli/README.md
new file mode 100644
index 0000000000000..1388fea8d557a
--- /dev/null
+++ b/model_cards/gsarti/scibert-nli/README.md
@@ -0,0 +1,32 @@
+# SciBERT-NLI
+
+This is the model [SciBERT](https://github.com/allenai/scibert) [1] fine-tuned on the [SNLI](https://nlp.stanford.edu/projects/snli/) and the [MultiNLI](https://www.nyu.edu/projects/bowman/multinli/) datasets using the [`sentence-transformers` library](https://github.com/UKPLab/sentence-transformers/) to produce universal sentence embeddings [2].
+
+The model uses the original `scivocab` wordpiece vocabulary and was trained using the **average pooling strategy** and a **softmax loss**.
+
+**Base model**: `allenai/scibert-scivocab-cased` from HuggingFace AutoModel
+
+**Parameters**:
+
+| Parameter | Value |
+|----------------|-------|
+| Batch size | 64 |
+| Training steps | 20000 |
+| Warmup steps | 1450 |
+
+**Performances**: The performance was evaluated on the test portion of the [STS dataset](http://ixa2.si.ehu.es/stswiki/index.php/STSbenchmark) using Spearman rank correlation and compared to the performances of a general BERT base model obtained with the same procedure to verify their similarity.
+
+
+| Model | Score |
+|-----------------------------|-------------|
+| `scibert-nli` (ours) | 74.50 |
+| `bert-base-nli-mean-tokens` | 77.12 |
+
+
+An example usage for similarity-based scientific paper retrieval is provided in the [Covid Papers Browser](https://github.com/gsarti/covid-papers-browser) repository.
+
+**References:**
+
+[1] I. Beltagy et al, [SciBERT: A Pretrained Language Model for Scientific Text](https://www.aclweb.org/anthology/D19-1371/)
+
+[2] A. Conneau et al., [Supervised Learning of Universal Sentence Representations from Natural Language Inference Data](https://www.aclweb.org/anthology/D17-1070/)
| https://api.github.com/repos/huggingface/transformers/pulls/3376 | 2020-03-22T13:26:17Z | 2020-03-23T15:55:42Z | 2020-03-23T15:55:42Z | 2020-03-23T15:55:43Z | 544 | huggingface/transformers | 12,407 |
|
Set CSR version in make_csr | diff --git a/letsencrypt/crypto_util.py b/letsencrypt/crypto_util.py
index 76265a73914..5fdcba843aa 100644
--- a/letsencrypt/crypto_util.py
+++ b/letsencrypt/crypto_util.py
@@ -118,6 +118,7 @@ def make_csr(key_str, domains):
value=", ".join("DNS:%s" % d for d in domains)
),
])
+ req.set_version(2)
req.set_pubkey(pkey)
req.sign(pkey, "sha256")
return tuple(OpenSSL.crypto.dump_certificate_request(method, req)
| Fixes #2528.
| https://api.github.com/repos/certbot/certbot/pulls/2529 | 2016-02-23T05:45:45Z | 2016-02-23T16:08:23Z | 2016-02-23T16:08:23Z | 2016-05-06T19:22:24Z | 134 | certbot/certbot | 632 |
🌐 Update Chinese translation for `docs/zh/docs/tutorial/query-params.md` | diff --git a/docs/zh/docs/tutorial/query-params.md b/docs/zh/docs/tutorial/query-params.md
index a0cc7fea39310..308dd68a486ab 100644
--- a/docs/zh/docs/tutorial/query-params.md
+++ b/docs/zh/docs/tutorial/query-params.md
@@ -1,67 +1,67 @@
# 查询参数
-声明不属于路径参数的其他函数参数时,它们将被自动解释为"查询字符串"参数
+声明的参数不是路径参数时,路径操作函数会把该参数自动解释为**查询**参数。
```Python hl_lines="9"
{!../../../docs_src/query_params/tutorial001.py!}
```
-查询字符串是键值对的集合,这些键值对位于 URL 的 `?` 之后,并以 `&` 符号分隔。
+查询字符串是键值对的集合,这些键值对位于 URL 的 `?` 之后,以 `&` 分隔。
-例如,在以下 url 中:
+例如,以下 URL 中:
```
http://127.0.0.1:8000/items/?skip=0&limit=10
```
-...查询参数为:
+……查询参数为:
-* `skip`:对应的值为 `0`
-* `limit`:对应的值为 `10`
+* `skip`:值为 `0`
+* `limit`:值为 `10`
-由于它们是 URL 的一部分,因此它们的"原始值"是字符串。
+这些值都是 URL 的组成部分,因此,它们的类型**本应**是字符串。
-但是,当你为它们声明了 Python 类型(在上面的示例中为 `int`)时,它们将转换为该类型并针对该类型进行校验。
+但声明 Python 类型(上例中为 `int`)之后,这些值就会转换为声明的类型,并进行类型校验。
-应用于路径参数的所有相同过程也适用于查询参数:
+所有应用于路径参数的流程也适用于查询参数:
-* (很明显的)编辑器支持
-* 数据<abbr title="将来自 HTTP 请求的字符串转换为 Python 数据类型">"解析"</abbr>
+* (显而易见的)编辑器支持
+* 数据<abbr title="将来自 HTTP 请求的字符串转换为 Python 数据类型">**解析**</abbr>
* 数据校验
-* 自动生成文档
+* API 文档
## 默认值
-由于查询参数不是路径的固定部分,因此它们可以是可选的,并且可以有默认值。
+查询参数不是路径的固定内容,它是可选的,还支持默认值。
-在上面的示例中,它们具有 `skip=0` 和 `limit=10` 的默认值。
+上例用 `skip=0` 和 `limit=10` 设定默认值。
-因此,访问 URL:
+访问 URL:
```
http://127.0.0.1:8000/items/
```
-将与访问以下地址相同:
+与访问以下地址相同:
```
http://127.0.0.1:8000/items/?skip=0&limit=10
```
-但是,如果你访问的是:
+但如果访问:
```
http://127.0.0.1:8000/items/?skip=20
```
-函数中的参数值将会是:
+查询参数的值就是:
* `skip=20`:在 URL 中设定的值
* `limit=10`:使用默认值
## 可选参数
-通过同样的方式,你可以将它们的默认值设置为 `None` 来声明可选查询参数:
+同理,把默认值设为 `None` 即可声明**可选的**查询参数:
=== "Python 3.10+"
@@ -76,20 +76,27 @@ http://127.0.0.1:8000/items/?skip=20
```
-在这个例子中,函数参数 `q` 将是可选的,并且默认值为 `None`。
+本例中,查询参数 `q` 是可选的,默认值为 `None`。
-!!! check
- 还要注意的是,**FastAPI** 足够聪明,能够分辨出参数 `item_id` 是路径参数而 `q` 不是,因此 `q` 是一个查询参数。
+!!! check "检查"
+
+ 注意,**FastAPI** 可以识别出 `item_id` 是路径参数,`q` 不是路径参数,而是查询参数。
+
+!!! note "笔记"
+
+ 因为默认值为 `= None`,FastAPI 把 `q` 识别为可选参数。
+
+ FastAPI 不使用 `Optional[str]` 中的 `Optional`(只使用 `str`),但 `Optional[str]` 可以帮助编辑器发现代码中的错误。
## 查询参数类型转换
-你还可以声明 `bool` 类型,它们将被自动转换:
+参数还可以声明为 `bool` 类型,FastAPI 会自动转换参数类型:
-```Python hl_lines="7"
+```Python hl_lines="9"
{!../../../docs_src/query_params/tutorial003.py!}
```
-这个例子中,如果你访问:
+本例中,访问:
```
http://127.0.0.1:8000/items/foo?short=1
@@ -119,42 +126,42 @@ http://127.0.0.1:8000/items/foo?short=on
http://127.0.0.1:8000/items/foo?short=yes
```
-或任何其他的变体形式(大写,首字母大写等等),你的函数接收的 `short` 参数都会是布尔值 `True`。对于值为 `False` 的情况也是一样的。
+或其它任意大小写形式(大写、首字母大写等),函数接收的 `short` 参数都是布尔值 `True`。值为 `False` 时也一样。
## 多个路径和查询参数
-你可以同时声明多个路径参数和查询参数,**FastAPI** 能够识别它们。
+**FastAPI** 可以识别同时声明的多个路径参数和查询参数。
-而且你不需要以任何特定的顺序来声明。
+而且声明查询参数的顺序并不重要。
-它们将通过名称被检测到:
+FastAPI 通过参数名进行检测:
-```Python hl_lines="6 8"
+```Python hl_lines="8 10"
{!../../../docs_src/query_params/tutorial004.py!}
```
-## 必需查询参数
+## 必选查询参数
-当你为非路径参数声明了默认值时(目前而言,我们所知道的仅有查询参数),则该参数不是必需的。
+为不是路径参数的参数声明默认值(至此,仅有查询参数),该参数就**不是必选**的了。
-如果你不想添加一个特定的值,而只是想使该参数成为可选的,则将默认值设置为 `None`。
+如果只想把参数设为**可选**,但又不想指定参数的值,则要把默认值设为 `None`。
-但当你想让一个查询参数成为必需的,不声明任何默认值就可以:
+如果要把查询参数设置为**必选**,就不要声明默认值:
```Python hl_lines="6-7"
{!../../../docs_src/query_params/tutorial005.py!}
```
-这里的查询参数 `needy` 是类型为 `str` 的必需查询参数。
+这里的查询参数 `needy` 是类型为 `str` 的必选查询参数。
-如果你在浏览器中打开一个像下面的 URL:
+在浏览器中打开如下 URL:
```
http://127.0.0.1:8000/items/foo-item
```
-...因为没有添加必需的参数 `needy`,你将看到类似以下的错误:
+……因为路径中没有必选参数 `needy`,返回的响应中会显示如下错误信息:
```JSON
{
@@ -171,13 +178,13 @@ http://127.0.0.1:8000/items/foo-item
}
```
-由于 `needy` 是必需参数,因此你需要在 URL 中设置它的值:
+`needy` 是必选参数,因此要在 URL 中设置值:
```
http://127.0.0.1:8000/items/foo-item?needy=sooooneedy
```
-...这样就正常了:
+……这样就正常了:
```JSON
{
@@ -186,17 +193,18 @@ http://127.0.0.1:8000/items/foo-item?needy=sooooneedy
}
```
-当然,你也可以定义一些参数为必需的,一些具有默认值,而某些则完全是可选的:
+当然,把一些参数定义为必选,为另一些参数设置默认值,再把其它参数定义为可选,这些操作都是可以的:
-```Python hl_lines="7"
+```Python hl_lines="10"
{!../../../docs_src/query_params/tutorial006.py!}
```
-在这个例子中,有3个查询参数:
+本例中有 3 个查询参数:
+
+* `needy`,必选的 `str` 类型参数
+* `skip`,默认值为 `0` 的 `int` 类型参数
+* `limit`,可选的 `int` 类型参数
-* `needy`,一个必需的 `str` 类型参数。
-* `skip`,一个默认值为 `0` 的 `int` 类型参数。
-* `limit`,一个可选的 `int` 类型参数。
+!!! tip "提示"
-!!! tip
- 你还可以像在 [路径参数](path-params.md#predefined-values){.internal-link target=_blank} 中那样使用 `Enum`。
+ 还可以像在[路径参数](path-params.md#predefined-values){.internal-link target=_blank} 中那样使用 `Enum`。
| also fix code highlight for tutorial002.py and tutorial006.py | https://api.github.com/repos/tiangolo/fastapi/pulls/3480 | 2021-07-07T08:05:07Z | 2024-04-01T05:36:48Z | 2024-04-01T05:36:48Z | 2024-04-01T05:36:48Z | 2,290 | tiangolo/fastapi | 23,195 |
C.146 Fix variable name in example | diff --git a/CppCoreGuidelines.md b/CppCoreGuidelines.md
index ec3200ed6..94876f845 100644
--- a/CppCoreGuidelines.md
+++ b/CppCoreGuidelines.md
@@ -8207,8 +8207,8 @@ Consider:
cout << pb2->id(); // "D"
- if (pb1->id() == "D") { // looks innocent
- D* pd = static_cast<D*>(pb1);
+ if (pb2->id() == "D") { // looks innocent
+ D* pd = static_cast<D*>(pb2);
// ...
}
// ...
| https://api.github.com/repos/isocpp/CppCoreGuidelines/pulls/2116 | 2023-08-06T04:02:14Z | 2023-08-06T05:25:02Z | 2023-08-06T05:25:02Z | 2023-08-06T05:25:02Z | 151 | isocpp/CppCoreGuidelines | 15,273 |
|
Reset progress bar in between report runs | diff --git a/frontend/src/components/core/Block/Block.tsx b/frontend/src/components/core/Block/Block.tsx
index a403c3d1674f..c6df749bb5fe 100644
--- a/frontend/src/components/core/Block/Block.tsx
+++ b/frontend/src/components/core/Block/Block.tsx
@@ -16,7 +16,6 @@
*/
import React, { PureComponent, ReactNode, Suspense } from "react"
-import { Progress } from "reactstrap"
import { AutoSizer } from "react-virtualized"
import { List, Map as ImmutableMap } from "immutable"
import { dispatchOneOf } from "lib/immutableProto"
@@ -59,6 +58,7 @@ const Button = React.lazy(() => import("components/widgets/Button/"))
const Checkbox = React.lazy(() => import("components/widgets/Checkbox/"))
const DateInput = React.lazy(() => import("components/widgets/DateInput/"))
const Multiselect = React.lazy(() => import("components/widgets/Multiselect/"))
+const Progress = React.lazy(() => import("components/elements/Progress/"))
const Radio = React.lazy(() => import("components/widgets/Radio/"))
const Selectbox = React.lazy(() => import("components/widgets/Selectbox/"))
const Slider = React.lazy(() => import("components/widgets/Slider/"))
@@ -249,13 +249,7 @@ class Block extends PureComponent<Props> {
plotlyChart: (el: SimpleElement) => (
<PlotlyChart element={el} width={width} />
),
- progress: (el: SimpleElement) => (
- <Progress
- value={el.get("value")}
- className="stProgress"
- style={{ width }}
- />
- ),
+ progress: (el: SimpleElement) => <Progress element={el} width={width} />,
table: (el: SimpleElement) => <Table element={el} width={width} />,
text: (el: SimpleElement) => <Text element={el} width={width} />,
vegaLiteChart: (el: SimpleElement) => (
diff --git a/frontend/src/components/elements/Progress/Progress.scss b/frontend/src/components/elements/Progress/Progress.scss
new file mode 100644
index 000000000000..584db1e590f8
--- /dev/null
+++ b/frontend/src/components/elements/Progress/Progress.scss
@@ -0,0 +1,32 @@
+/**
+ * Copyright 2018-2019 Streamlit Inc.
+ *
+ * Licensed under the Apache License, Version 2.0 (the "License");
+ * you may not use this file except in compliance with the License.
+ * You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+@import "src/assets/css/variables";
+
+// Reset the progress bar to 0 without animating on report re-run.
+.reportview-container {
+ .without-transition .progress-bar {
+ transition: none;
+ }
+
+ .with-transition .progress-bar {
+ transition: width 0.1s linear;
+ }
+
+ .stale-element .progress-bar {
+ background-color: transparent;
+ }
+}
diff --git a/frontend/src/components/elements/Progress/Progress.tsx b/frontend/src/components/elements/Progress/Progress.tsx
new file mode 100644
index 000000000000..3142ac25a34b
--- /dev/null
+++ b/frontend/src/components/elements/Progress/Progress.tsx
@@ -0,0 +1,63 @@
+/**
+ * @license
+ * Copyright 2018-2019 Streamlit Inc.
+ *
+ * Licensed under the Apache License, Version 2.0 (the "License");
+ * you may not use this file except in compliance with the License.
+ * You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+import React from "react"
+import { Map as ImmutableMap } from "immutable"
+import { Progress as UIProgress } from "reactstrap"
+
+import "./Progress.scss"
+
+interface Props {
+ width: number
+ element: ImmutableMap<string, any>
+}
+
+const FAST_UPDATE_MS = 50
+
+class Progress extends React.PureComponent<Props> {
+ lastValue = -1
+ lastAnimatedTime = -1
+
+ public render(): React.ReactNode {
+ const { element, width } = this.props
+ const value = element.get("value")
+ const time = new Date().getTime()
+
+ // Make progress bar stop acting weird when moving backwards or quickly.
+ const isMovingBackwards = value < this.lastValue
+ const isMovingSuperFast = time - this.lastAnimatedTime < FAST_UPDATE_MS
+ const className =
+ isMovingBackwards || isMovingSuperFast
+ ? "without-transition"
+ : "with-transition"
+
+ if (className === "with-transition") {
+ this.lastAnimatedTime = time
+ }
+ this.lastValue = value
+
+ return (
+ <UIProgress
+ value={value}
+ className={"stProgress " + className}
+ style={{ width }}
+ />
+ )
+ }
+}
+
+export default Progress
diff --git a/frontend/src/components/elements/Progress/index.tsx b/frontend/src/components/elements/Progress/index.tsx
new file mode 100644
index 000000000000..9bf69066a927
--- /dev/null
+++ b/frontend/src/components/elements/Progress/index.tsx
@@ -0,0 +1,18 @@
+/**
+ * @license
+ * Copyright 2018-2019 Streamlit Inc.
+ *
+ * Licensed under the Apache License, Version 2.0 (the "License");
+ * you may not use this file except in compliance with the License.
+ * You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+export { default } from "./Progress"
diff --git a/lib/streamlit/credentials.py b/lib/streamlit/credentials.py
index 7202ba6cf8b2..b32909eca171 100644
--- a/lib/streamlit/credentials.py
+++ b/lib/streamlit/credentials.py
@@ -202,8 +202,10 @@ def activate(self, show_instructions=True):
email = ""
else:
email = click.prompt(
- text=EMAIL_PROMPT, prompt_suffix="", default="",
- show_default=False
+ text=EMAIL_PROMPT,
+ prompt_suffix="",
+ default="",
+ show_default=False,
)
self.activation = _verify_email(email)
diff --git a/lib/streamlit/server/Server.py b/lib/streamlit/server/Server.py
index 5280bb3d386d..723de9c72ea8 100644
--- a/lib/streamlit/server/Server.py
+++ b/lib/streamlit/server/Server.py
@@ -247,10 +247,11 @@ def _create_app(self):
routes.extend(
[
- (r"/(.*)", StaticFileHandler, {
- "path": "%s/" % static_path,
- "default_filename": "index.html",
- }),
+ (
+ r"/(.*)",
+ StaticFileHandler,
+ {"path": "%s/" % static_path, "default_filename": "index.html"},
+ )
]
)
diff --git a/lib/streamlit/util.py b/lib/streamlit/util.py
index bb553c899e60..d9eafdb92cb8 100644
--- a/lib/streamlit/util.py
+++ b/lib/streamlit/util.py
@@ -252,6 +252,7 @@ def open_browser(url):
# browser even though 'start url' works from the command prompt.
# Fun!
import webbrowser
+
webbrowser.open(url)
return
| **Issue:**
https://github.com/streamlit/streamlit-old-private/issues/760
**Description:**
- Unmount progress bar when it's stale so it resets
**Notes:**
- I didn't see any issues with the `update-progress animation` as mentioned in the issue. Perhaps @tconkling can provide reproduction steps for that. | https://api.github.com/repos/streamlit/streamlit/pulls/321 | 2019-10-08T15:01:11Z | 2019-10-14T20:48:35Z | 2019-10-14T20:48:35Z | 2019-10-14T20:48:38Z | 1,968 | streamlit/streamlit | 22,550 |
Remove redundant SQL index in Pastebin exercise | diff --git a/solutions/system_design/pastebin/README.md b/solutions/system_design/pastebin/README.md
index 756c78c274..2d87ddcc7e 100644
--- a/solutions/system_design/pastebin/README.md
+++ b/solutions/system_design/pastebin/README.md
@@ -116,7 +116,7 @@ paste_path varchar(255) NOT NULL
PRIMARY KEY(shortlink)
```
-We'll create an [index](https://github.com/donnemartin/system-design-primer#use-good-indices) on `shortlink ` and `created_at` to speed up lookups (log-time instead of scanning the entire table) and to keep the data in memory. Reading 1 MB sequentially from memory takes about 250 microseconds, while reading from SSD takes 4x and from disk takes 80x longer.<sup><a href=https://github.com/donnemartin/system-design-primer#latency-numbers-every-programmer-should-know>1</a></sup>
+Setting the primary key to be based on the `shortlink` column creates an [index](https://github.com/donnemartin/system-design-primer#use-good-indices) that the database uses to enforce uniqueness. We'll create an additional index on `created_at` to speed up lookups (log-time instead of scanning the entire table) and to keep the data in memory. Reading 1 MB sequentially from memory takes about 250 microseconds, while reading from SSD takes 4x and from disk takes 80x longer.<sup><a href=https://github.com/donnemartin/system-design-primer#latency-numbers-every-programmer-should-know>1</a></sup>
To generate the unique url, we could:
| Making a column a primary key enforces uniqueness, and the db usually does so by creating a unique clustered index. So, you don't need a second index. | https://api.github.com/repos/donnemartin/system-design-primer/pulls/405 | 2020-04-20T19:16:06Z | 2020-07-07T01:05:51Z | 2020-07-07T01:05:51Z | 2020-07-07T01:06:04Z | 380 | donnemartin/system-design-primer | 36,749 |
Prep for 1.32.1 | diff --git a/.azure-pipelines/release.yml b/.azure-pipelines/release.yml
index 1c983a3b647..2374289e3c9 100644
--- a/.azure-pipelines/release.yml
+++ b/.azure-pipelines/release.yml
@@ -8,7 +8,7 @@ pr: none
variables:
dockerTag: ${{variables['Build.SourceBranchName']}}
- snapBuildTimeout: 5400
+ snapBuildTimeout: 19800
stages:
- template: templates/stages/test-and-package-stage.yml
diff --git a/certbot/CHANGELOG.md b/certbot/CHANGELOG.md
index 2fcb78c99c2..f3bd9f56507 100644
--- a/certbot/CHANGELOG.md
+++ b/certbot/CHANGELOG.md
@@ -2,6 +2,16 @@
Certbot adheres to [Semantic Versioning](https://semver.org/).
+## 1.32.1 - master
+
+### Fixed
+
+* Our snaps and docker images were rebuilt to include updated versions of our dependencies.
+
+This release was not pushed to PyPI since those packages were unaffected.
+
+More details about these changes can be found on our GitHub repo.
+
## 1.32.0 - 2022-11-08
### Added
| I wanted to do this because we were notified that https://ubuntu.com/security/notices/USN-5638-3/ affects our snaps. This probably doesn't affect us, but rebuilding to be safe seems worth it to me personally.
I started to just trigger a new v1.32.0 release build, but I don't want to overwrite our 2.0 Docker images under the `latest` tag.
Changelog changes here are similar to what has been done for past point releases like https://github.com/certbot/certbot/pull/8501.
I also cherry picked #9474 to this branch to help the release process pass.
| https://api.github.com/repos/certbot/certbot/pulls/9492 | 2022-12-02T15:00:30Z | 2022-12-05T15:00:44Z | 2022-12-05T15:00:44Z | 2022-12-05T15:00:45Z | 299 | certbot/certbot | 1,077 |
Add share.flows command, fix #2779 | diff --git a/mitmproxy/addons/__init__.py b/mitmproxy/addons/__init__.py
index 8f84c20d9f..619211130b 100644
--- a/mitmproxy/addons/__init__.py
+++ b/mitmproxy/addons/__init__.py
@@ -20,6 +20,7 @@
from mitmproxy.addons import streambodies
from mitmproxy.addons import save
from mitmproxy.addons import upstream_auth
+from mitmproxy.addons import share
def default_addons():
@@ -46,4 +47,5 @@ def default_addons():
streambodies.StreamBodies(),
save.Save(),
upstream_auth.UpstreamAuth(),
+ share.Share()
]
diff --git a/mitmproxy/addons/save.py b/mitmproxy/addons/save.py
index 44afef686e..47da29b256 100644
--- a/mitmproxy/addons/save.py
+++ b/mitmproxy/addons/save.py
@@ -61,8 +61,8 @@ def save(self, flows: typing.Sequence[flow.Flow], path: mitmproxy.types.Path) ->
except IOError as v:
raise exceptions.CommandError(v) from v
stream = io.FlowWriter(f)
- for i in flows:
- stream.add(i)
+ for x in flows:
+ stream.add(x)
f.close()
ctx.log.alert("Saved %s flows." % len(flows))
diff --git a/mitmproxy/addons/share.py b/mitmproxy/addons/share.py
new file mode 100644
index 0000000000..1a234cc944
--- /dev/null
+++ b/mitmproxy/addons/share.py
@@ -0,0 +1,79 @@
+import typing
+import random
+import string
+import io
+import http.client
+
+from mitmproxy import command
+import mitmproxy.io
+from mitmproxy import ctx
+from mitmproxy import flow
+from mitmproxy.net.http import status_codes
+
+
+class Share:
+ def encode_multipart_formdata(self, filename: str, content: bytes) -> typing.Tuple[str, bytes]:
+ params = {"key": filename, "acl": "bucket-owner-full-control", "Content-Type": "application/octet-stream"}
+ LIMIT = b'---------------------------198495659117975628761412556003'
+ CRLF = b'\r\n'
+ l = []
+ for (key, value) in params.items():
+ l.append(b'--' + LIMIT)
+ l.append(b'Content-Disposition: form-data; name="%b"' % key.encode("utf-8"))
+ l.append(b'')
+ l.append(value.encode("utf-8"))
+ l.append(b'--' + LIMIT)
+ l.append(b'Content-Disposition: form-data; name="file"; filename="%b"' % filename.encode("utf-8"))
+ l.append(b'Content-Type: application/octet-stream')
+ l.append(b'')
+ l.append(content)
+ l.append(b'--' + LIMIT + b'--')
+ l.append(b'')
+ body = CRLF.join(l)
+ content_type = 'multipart/form-data; boundary=%s' % LIMIT.decode("utf-8")
+ return content_type, body
+
+ def post_multipart(self, host: str, filename: str, content: bytes) -> str:
+ """
+ Upload flows to the specified S3 server.
+
+ Returns:
+ - The share URL, if upload is successful.
+ Raises:
+ - IOError, otherwise.
+ """
+ content_type, body = self.encode_multipart_formdata(filename, content)
+ conn = http.client.HTTPConnection(host) # FIXME: This ultimately needs to be HTTPSConnection
+ headers = {'content-type': content_type}
+ try:
+ conn.request("POST", "", body, headers)
+ resp = conn.getresponse()
+ except Exception as v:
+ raise IOError(v)
+ finally:
+ conn.close()
+ if resp.status != 204:
+ if resp.reason:
+ reason = resp.reason
+ else:
+ reason = status_codes.RESPONSES.get(resp.status, str(resp.status))
+ raise IOError(reason)
+ return "https://share.mitmproxy.org/%s" % filename
+
+ @command.command("share.flows")
+ def share(self, flows: typing.Sequence[flow.Flow]) -> None:
+ u_id = "".join(random.choice(string.ascii_lowercase + string.digits)for _ in range(7))
+ f = io.BytesIO()
+ stream = mitmproxy.io.FlowWriter(f)
+ for x in flows:
+ stream.add(x)
+ f.seek(0)
+ content = f.read()
+ try:
+ res = self.post_multipart('upload.share.mitmproxy.org.s3.amazonaws.com', u_id, content)
+ except IOError as v:
+ ctx.log.warn("%s" % v)
+ else:
+ ctx.log.alert("%s" % res)
+ finally:
+ f.close()
\ No newline at end of file
diff --git a/test/mitmproxy/addons/test_share.py b/test/mitmproxy/addons/test_share.py
new file mode 100644
index 0000000000..6c3d6e28fc
--- /dev/null
+++ b/test/mitmproxy/addons/test_share.py
@@ -0,0 +1,34 @@
+from unittest import mock
+import http.client
+
+from mitmproxy.test import taddons
+from mitmproxy.test import tflow
+
+from mitmproxy.addons import share
+from mitmproxy.addons import view
+
+
+def test_share_command():
+ with mock.patch('mitmproxy.addons.share.http.client.HTTPConnection') as mock_http:
+ sh = share.Share()
+ with taddons.context() as tctx:
+ mock_http.return_value.getresponse.return_value = mock.MagicMock(status=204, reason="No Content")
+ sh.share([tflow.tflow(resp=True)])
+ assert tctx.master.has_log("https://share.mitmproxy.org/")
+
+ mock_http.return_value.getresponse.return_value = mock.MagicMock(status=403, reason="Forbidden")
+ sh.share([tflow.tflow(resp=True)])
+ assert tctx.master.has_log("Forbidden")
+
+ mock_http.return_value.getresponse.return_value = mock.MagicMock(status=404, reason="")
+ sh.share([tflow.tflow(resp=True)])
+ assert tctx.master.has_log("Not Found")
+
+ mock_http.return_value.request.side_effect = http.client.CannotSendRequest("Error in sending req")
+ sh.share([tflow.tflow(resp=True)])
+ assert tctx.master.has_log("Error in sending req")
+
+ v = view.View()
+ tctx.master.addons.add(v)
+ tctx.master.addons.add(sh)
+ tctx.master.commands.call_args("share.flows", ["@shown"])
| Fixes #2779 . Added the command 'upload.file' | https://api.github.com/repos/mitmproxy/mitmproxy/pulls/2802 | 2018-01-18T20:46:01Z | 2018-02-13T19:01:01Z | 2018-02-13T19:01:00Z | 2018-06-15T16:54:24Z | 1,559 | mitmproxy/mitmproxy | 28,324 |
Update README.md | diff --git a/exercises/ansible/README.md b/exercises/ansible/README.md
index 12982c3fd..206c4d84e 100644
--- a/exercises/ansible/README.md
+++ b/exercises/ansible/README.md
@@ -15,6 +15,7 @@
<summary>Describe each of the following components in Ansible, including the relationship between them:
* Task
+ * Inventory
* Module
* Play
* Playbook
@@ -23,6 +24,8 @@
Task – a call to a specific Ansible module
Module – the actual unit of code executed by Ansible on your own host or a remote host. Modules are indexed by category (database, file, network, …) and also referred to as task plugins.
+
+Inventory – An inventory file defines hosts and/or groups of hosts on which Ansible tasks executed upon. The inventory file can be in one of many formats, depending on the inventory plugins you have. The most common formats are INI and YAML.
Play – One or more tasks executed on a given host(s)
| Hey repo owner!
I have added a new feature on Ansible topic (inventory component), I think that will increase the understand of all people on this topic about Ansible tool. if i am wrong just ignore
Let me know if you have any questions around this.
| https://api.github.com/repos/bregman-arie/devops-exercises/pulls/194 | 2021-11-29T00:17:04Z | 2021-11-29T06:11:48Z | 2021-11-29T06:11:48Z | 2021-11-29T06:11:48Z | 242 | bregman-arie/devops-exercises | 17,400 |
Better handling of CookieJar Runtime Exception | diff --git a/lib/request/connect.py b/lib/request/connect.py
index 8508dee51ff..84ec25e4d38 100644
--- a/lib/request/connect.py
+++ b/lib/request/connect.py
@@ -587,14 +587,9 @@ class _(dict):
if not getRequestHeader(req, HTTP_HEADER.COOKIE) and conf.cj:
conf.cj._policy._now = conf.cj._now = int(time.time())
- while True:
- try:
- cookies = conf.cj._cookies_for_request(req)
- except RuntimeError: # NOTE: https://github.com/sqlmapproject/sqlmap/issues/5187
- time.sleep(1)
- else:
- requestHeaders += "\r\n%s" % ("Cookie: %s" % ";".join("%s=%s" % (getUnicode(cookie.name), getUnicode(cookie.value)) for cookie in cookies))
- break
+ with conf.cj._cookies_lock:
+ cookies = conf.cj._cookies_for_request(req)
+ requestHeaders += "\r\n%s" % ("Cookie: %s" % ";".join("%s=%s" % (getUnicode(cookie.name), getUnicode(cookie.value)) for cookie in cookies))
if post is not None:
if not getRequestHeader(req, HTTP_HEADER.CONTENT_LENGTH) and not chunked:
| Instead of waiting for the exception and sleeping in between, we can simply use the same lock used by the original cookiejar code (since we are already accessing private member), to prevent the exception from happening in the first place.
Fixes #5187 | https://api.github.com/repos/sqlmapproject/sqlmap/pulls/5206 | 2022-10-21T08:56:40Z | 2022-10-21T17:10:43Z | 2022-10-21T17:10:43Z | 2022-10-21T17:10:52Z | 294 | sqlmapproject/sqlmap | 15,062 |
W&B log epoch | diff --git a/test.py b/test.py
index 891f6bef41c..db344e72204 100644
--- a/test.py
+++ b/test.py
@@ -239,8 +239,8 @@ def test(data,
if plots:
confusion_matrix.plot(save_dir=save_dir, names=list(names.values()))
if wandb and wandb.run:
- wandb.log({"Images": wandb_images})
- wandb.log({"Validation": [wandb.Image(str(f), caption=f.name) for f in sorted(save_dir.glob('test*.jpg'))]})
+ val_batches = [wandb.Image(str(f), caption=f.name) for f in sorted(save_dir.glob('test*.jpg'))]
+ wandb.log({"Images": wandb_images, "Validation": val_batches}, commit=False)
# Save JSON
if save_json and len(jdict):
diff --git a/train.py b/train.py
index 9e6bd867372..5eff4bbac17 100644
--- a/train.py
+++ b/train.py
@@ -321,7 +321,7 @@ def train(hyp, opt, device, tb_writer=None, wandb=None):
# tb_writer.add_graph(model, imgs) # add model to tensorboard
elif plots and ni == 10 and wandb:
wandb.log({"Mosaics": [wandb.Image(str(x), caption=x.name) for x in save_dir.glob('train*.jpg')
- if x.exists()]})
+ if x.exists()]}, commit=False)
# end batch ------------------------------------------------------------------------------------------------
# end epoch ----------------------------------------------------------------------------------------------------
diff --git a/utils/plots.py b/utils/plots.py
index 4765069e037..67f11bfd201 100644
--- a/utils/plots.py
+++ b/utils/plots.py
@@ -295,7 +295,7 @@ def plot_labels(labels, save_dir=Path(''), loggers=None):
# loggers
for k, v in loggers.items() or {}:
if k == 'wandb' and v:
- v.log({"Labels": [v.Image(str(x), caption=x.name) for x in save_dir.glob('*labels*.jpg')]})
+ v.log({"Labels": [v.Image(str(x), caption=x.name) for x in save_dir.glob('*labels*.jpg')]}, commit=False)
def plot_evolution(yaml_file='data/hyp.finetune.yaml'): # from utils.plots import *; plot_evolution()
| May allow W&B results logging with epoch x axis.
## 🛠️ PR Summary
<sub>Made with ❤️ by [Ultralytics Actions](https://github.com/ultralytics/actions)<sub>
### 🌟 Summary
Improvement in Weight & Biases logging within YOLOv5's testing and training routines.
### 📊 Key Changes
- Modified `wandb.log` calls in `test.py` and `train.py` to use `commit=False` option.
- Updated `wandb.log` in `utils/plots.py` with `commit=False`.
### 🎯 Purpose & Impact
- 👍 **Purpose**: These changes allow for more efficient logging to Weight & Biases by grouping multiple logs into a single network call.
- 🚀 **Impact**: Reduces the number of calls to Weights & Biases API during training and testing, potentially speeding up these processes and improving resource usage. Users of the YOLOv5 framework with integrated Weights & Biases logging will benefit from these optimizations. | https://api.github.com/repos/ultralytics/yolov5/pulls/1946 | 2021-01-15T04:44:40Z | 2021-01-27T05:16:02Z | 2021-01-27T05:16:02Z | 2024-01-19T19:49:02Z | 538 | ultralytics/yolov5 | 24,979 |
Fix super tiny type error | diff --git a/timm/scheduler/cosine_lr.py b/timm/scheduler/cosine_lr.py
index e2c975fb79..4eaaa86a81 100644
--- a/timm/scheduler/cosine_lr.py
+++ b/timm/scheduler/cosine_lr.py
@@ -8,6 +8,7 @@
import math
import numpy as np
import torch
+from typing import List
from .scheduler import Scheduler
@@ -77,7 +78,7 @@ def __init__(
else:
self.warmup_steps = [1 for _ in self.base_values]
- def _get_lr(self, t):
+ def _get_lr(self, t: int) -> List[float]:
if t < self.warmup_t:
lrs = [self.warmup_lr_init + t * s for s in self.warmup_steps]
else:
diff --git a/timm/scheduler/multistep_lr.py b/timm/scheduler/multistep_lr.py
index 10f2fb5044..e5db556d43 100644
--- a/timm/scheduler/multistep_lr.py
+++ b/timm/scheduler/multistep_lr.py
@@ -53,7 +53,7 @@ def get_curr_decay_steps(self, t):
# assumes self.decay_t is sorted
return bisect.bisect_right(self.decay_t, t + 1)
- def _get_lr(self, t):
+ def _get_lr(self, t: int) -> List[float]:
if t < self.warmup_t:
lrs = [self.warmup_lr_init + t * s for s in self.warmup_steps]
else:
diff --git a/timm/scheduler/plateau_lr.py b/timm/scheduler/plateau_lr.py
index 9f8271579b..e868bd5e58 100644
--- a/timm/scheduler/plateau_lr.py
+++ b/timm/scheduler/plateau_lr.py
@@ -5,6 +5,7 @@
Hacked together by / Copyright 2020 Ross Wightman
"""
import torch
+from typing import List
from .scheduler import Scheduler
@@ -106,5 +107,5 @@ def _apply_noise(self, epoch):
param_group['lr'] = new_lr
self.restore_lr = restore_lr
- def _get_lr(self, t: int) -> float:
+ def _get_lr(self, t: int) -> List[float]:
assert False, 'should not be called as step is overridden'
diff --git a/timm/scheduler/poly_lr.py b/timm/scheduler/poly_lr.py
index 906f6acf82..8875e15bfe 100644
--- a/timm/scheduler/poly_lr.py
+++ b/timm/scheduler/poly_lr.py
@@ -6,6 +6,7 @@
"""
import math
import logging
+from typing import List
import torch
@@ -73,7 +74,7 @@ def __init__(
else:
self.warmup_steps = [1 for _ in self.base_values]
- def _get_lr(self, t):
+ def _get_lr(self, t: int) -> List[float]:
if t < self.warmup_t:
lrs = [self.warmup_lr_init + t * s for s in self.warmup_steps]
else:
diff --git a/timm/scheduler/scheduler.py b/timm/scheduler/scheduler.py
index 4ae2e2aeb6..583357f7c5 100644
--- a/timm/scheduler/scheduler.py
+++ b/timm/scheduler/scheduler.py
@@ -1,6 +1,6 @@
import abc
from abc import ABC
-from typing import Any, Dict, Optional
+from typing import Any, Dict, List, Optional
import torch
@@ -65,10 +65,10 @@ def load_state_dict(self, state_dict: Dict[str, Any]) -> None:
self.__dict__.update(state_dict)
@abc.abstractmethod
- def _get_lr(self, t: int) -> float:
+ def _get_lr(self, t: int) -> List[float]:
pass
- def _get_values(self, t: int, on_epoch: bool = True) -> Optional[float]:
+ def _get_values(self, t: int, on_epoch: bool = True) -> Optional[List[float]]:
proceed = (on_epoch and self.t_in_epochs) or (not on_epoch and not self.t_in_epochs)
if not proceed:
return None
diff --git a/timm/scheduler/step_lr.py b/timm/scheduler/step_lr.py
index 70a45a70d4..c205d43715 100644
--- a/timm/scheduler/step_lr.py
+++ b/timm/scheduler/step_lr.py
@@ -6,6 +6,8 @@
"""
import math
import torch
+from typing import List
+
from .scheduler import Scheduler
@@ -51,7 +53,7 @@ def __init__(
else:
self.warmup_steps = [1 for _ in self.base_values]
- def _get_lr(self, t):
+ def _get_lr(self, t: int) -> List[float]:
if t < self.warmup_t:
lrs = [self.warmup_lr_init + t * s for s in self.warmup_steps]
else:
diff --git a/timm/scheduler/tanh_lr.py b/timm/scheduler/tanh_lr.py
index 48acc61b03..94455302c6 100644
--- a/timm/scheduler/tanh_lr.py
+++ b/timm/scheduler/tanh_lr.py
@@ -8,6 +8,7 @@
import math
import numpy as np
import torch
+from typing import List
from .scheduler import Scheduler
@@ -75,7 +76,7 @@ def __init__(
else:
self.warmup_steps = [1 for _ in self.base_values]
- def _get_lr(self, t):
+ def _get_lr(self, t: int) -> List[float]:
if t < self.warmup_t:
lrs = [self.warmup_lr_init + t * s for s in self.warmup_steps]
else:
| IMHO, for example, CosineLRScheduler returns list of floats, instead of a single float. Therefore, the type signature may need to be updated. Please correct me if I am wrong! | https://api.github.com/repos/huggingface/pytorch-image-models/pulls/2124 | 2024-03-23T03:27:44Z | 2024-04-02T21:31:38Z | 2024-04-02T21:31:38Z | 2024-04-03T00:39:19Z | 1,394 | huggingface/pytorch-image-models | 16,230 |
gh-72073: Add Windows case in pathlib.rename | diff --git a/Doc/library/pathlib.rst b/Doc/library/pathlib.rst
index d45e7aa84b28ce..cda2cc83225e07 100644
--- a/Doc/library/pathlib.rst
+++ b/Doc/library/pathlib.rst
@@ -1021,8 +1021,9 @@ call fails (for example because the path doesn't exist).
Rename this file or directory to the given *target*, and return a new Path
instance pointing to *target*. On Unix, if *target* exists and is a file,
- it will be replaced silently if the user has permission. *target* can be
- either a string or another path object::
+ it will be replaced silently if the user has permission.
+ On Windows, if *target* exists, :data:`FileExistsError` will be raised.
+ *target* can be either a string or another path object::
>>> p = Path('foo')
>>> p.open('w').write('some text')
| #72073
https://docs.python.org/3.12/library/pathlib.html#pathlib.Path.rename
Automerge-Triggered-By: GH:brettcannon | https://api.github.com/repos/python/cpython/pulls/93002 | 2022-05-20T08:05:26Z | 2022-05-20T22:25:40Z | 2022-05-20T22:25:39Z | 2022-05-21T04:31:52Z | 229 | python/cpython | 4,622 |
Use long classes names for enabled middlewares in startup logs | diff --git a/scrapy/middleware.py b/scrapy/middleware.py
index 6120488e22f..be36f977e41 100644
--- a/scrapy/middleware.py
+++ b/scrapy/middleware.py
@@ -28,6 +28,7 @@ def _get_mwlist_from_settings(cls, settings):
def from_settings(cls, settings, crawler=None):
mwlist = cls._get_mwlist_from_settings(settings)
middlewares = []
+ enabled = []
for clspath in mwlist:
try:
mwcls = load_object(clspath)
@@ -38,6 +39,7 @@ def from_settings(cls, settings, crawler=None):
else:
mw = mwcls()
middlewares.append(mw)
+ enabled.append(clspath)
except NotConfigured as e:
if e.args:
clsname = clspath.split('.')[-1]
@@ -45,7 +47,6 @@ def from_settings(cls, settings, crawler=None):
{'clsname': clsname, 'eargs': e.args[0]},
extra={'crawler': crawler})
- enabled = [x.__class__.__name__ for x in middlewares]
logger.info("Enabled %(componentname)ss:\n%(enabledlist)s",
{'componentname': cls.component_name,
'enabledlist': pprint.pformat(enabled)},
| Continuation of https://github.com/scrapy/scrapy/pull/1722
| https://api.github.com/repos/scrapy/scrapy/pulls/1726 | 2016-01-26T15:44:27Z | 2016-01-26T16:29:44Z | 2016-01-26T16:29:44Z | 2016-02-01T14:53:45Z | 294 | scrapy/scrapy | 35,115 |
dns-cloudflare: update URL for obtaining API keys | diff --git a/certbot-dns-cloudflare/certbot_dns_cloudflare/dns_cloudflare.py b/certbot-dns-cloudflare/certbot_dns_cloudflare/dns_cloudflare.py
index e3d0d42e047..0bbdf703ae8 100644
--- a/certbot-dns-cloudflare/certbot_dns_cloudflare/dns_cloudflare.py
+++ b/certbot-dns-cloudflare/certbot_dns_cloudflare/dns_cloudflare.py
@@ -10,7 +10,7 @@
logger = logging.getLogger(__name__)
-ACCOUNT_URL = 'https://www.cloudflare.com/a/account/my-account'
+ACCOUNT_URL = 'https://dash.cloudflare.com/profile/api-tokens'
@zope.interface.implementer(interfaces.IAuthenticator)
| Updated the ACCOUNT_URL in the Cloudflare-DNS plugin.
This uses the new "dash.cloudflare.com" scheme and future-proofs this URL for an upcoming change to Cloudflare API keys (this is not public yet, so no other changes related to this).
| https://api.github.com/repos/certbot/certbot/pulls/7052 | 2019-05-11T22:00:21Z | 2019-06-26T00:53:32Z | 2019-06-26T00:53:32Z | 2019-06-26T14:11:20Z | 168 | certbot/certbot | 2,376 |
cookies.txt extension doesn't exist anymore on the Chrome Web Store | diff --git a/README.md b/README.md
index 34c6c677d8f..35ae364213e 100644
--- a/README.md
+++ b/README.md
@@ -879,7 +879,7 @@ Either prepend `https://www.youtube.com/watch?v=` or separate the ID from the op
Use the `--cookies` option, for example `--cookies /path/to/cookies/file.txt`.
-In order to extract cookies from browser use any conforming browser extension for exporting cookies. For example, [cookies.txt](https://chrome.google.com/webstore/detail/cookiestxt/njabckikapfpffapmjgojcnbfjonfjfg) (for Chrome) or [cookies.txt](https://addons.mozilla.org/en-US/firefox/addon/cookies-txt/) (for Firefox).
+In order to extract cookies from browser use any conforming browser extension for exporting cookies. For example, [Get cookies.txt](https://chrome.google.com/webstore/detail/get-cookiestxt/bgaddhkoddajcdgocldbbfleckgcbcid/) (for Chrome) or [cookies.txt](https://addons.mozilla.org/en-US/firefox/addon/cookies-txt/) (for Firefox).
Note that the cookies file must be in Mozilla/Netscape format and the first line of the cookies file must be either `# HTTP Cookie File` or `# Netscape HTTP Cookie File`. Make sure you have correct [newline format](https://en.wikipedia.org/wiki/Newline) in the cookies file and convert newlines if necessary to correspond with your OS, namely `CRLF` (`\r\n`) for Windows and `LF` (`\n`) for Unix and Unix-like systems (Linux, macOS, etc.). `HTTP Error 400: Bad Request` when using `--cookies` is a good sign of invalid newline format.
| ### Before submitting a *pull request* make sure you have:
- [x] [Searched](https://github.com/ytdl-org/youtube-dl/search?q=is%3Apr&type=Issues) the bugtracker for similar pull requests
- [x] Read [adding new extractor tutorial](https://github.com/ytdl-org/youtube-dl#adding-support-for-a-new-site)
- [x] Read [youtube-dl coding conventions](https://github.com/ytdl-org/youtube-dl#youtube-dl-coding-conventions) and adjusted the code to meet them
- [x] Covered the code with tests (note that PRs without tests will be REJECTED)
- [ ] Checked the code with [flake8](https://pypi.python.org/pypi/flake8)
### In order to be accepted and merged into youtube-dl each piece of code must be in public domain or released under [Unlicense](http://unlicense.org/). Check one of the following options:
- [x] I am the original author of this code and I am willing to release it under [Unlicense](http://unlicense.org/)
- [ ] I am not the original author of this code but it is in public domain or released under [Unlicense](http://unlicense.org/) (provide reliable evidence)
### What is the purpose of your *pull request*?
- [ ] Bug fix
- [x] Improvement
- [ ] New extractor
- [ ] New feature
---
### Description of your *pull request* and other information
The [cookies.txt extension](https://chrome.google.com/webstore/detail/cookiestxt/njabckikapfpffapmjgojcnbfjonfjfg) doesn't exist anymore on the Chrome Web Store, so I propose to change the link in the _README.md_ to another similar extension called [Get cookies.txt](https://chrome.google.com/webstore/detail/get-cookiestxt/bgaddhkoddajcdgocldbbfleckgcbcid/) with the same functions and utility of the old one. I tested the extension personally on my machine and it exports cookies in the Netscape cookie file format which is compatible with youtube-dl, just like the cookies.txt extension did before.
This commit does **NOT change any code**, is just a fix for a link on the _README.md_
This PR close #26885
| https://api.github.com/repos/ytdl-org/youtube-dl/pulls/27433 | 2020-12-14T14:15:10Z | 2020-12-26T13:50:40Z | 2020-12-26T13:50:40Z | 2020-12-26T13:52:26Z | 393 | ytdl-org/youtube-dl | 50,343 |
Add --new-key | diff --git a/certbot-ci/certbot_integration_tests/certbot_tests/assertions.py b/certbot-ci/certbot_integration_tests/certbot_tests/assertions.py
index 2720842171e..3650f64f043 100644
--- a/certbot-ci/certbot_integration_tests/certbot_tests/assertions.py
+++ b/certbot-ci/certbot_integration_tests/certbot_tests/assertions.py
@@ -37,16 +37,19 @@ def assert_elliptic_key(key: str, curve: Type[EllipticCurve]) -> None:
assert isinstance(key.curve, curve)
-def assert_rsa_key(key: str) -> None:
+def assert_rsa_key(key: str, key_size: Optional[int] = None) -> None:
"""
Asserts that the key at the given path is an RSA key.
:param str key: path to key
+ :param int key_size: if provided, assert that the RSA key is of this size
"""
with open(key, 'rb') as file:
privkey1 = file.read()
key = load_pem_private_key(data=privkey1, password=None, backend=default_backend())
assert isinstance(key, RSAPrivateKey)
+ if key_size:
+ assert key_size == key.key_size
def assert_hook_execution(probe_path: str, probe_content: str) -> None:
diff --git a/certbot-ci/certbot_integration_tests/certbot_tests/test_main.py b/certbot-ci/certbot_integration_tests/certbot_tests/test_main.py
index 4a33952174f..2827ae939e2 100644
--- a/certbot-ci/certbot_integration_tests/certbot_tests/test_main.py
+++ b/certbot-ci/certbot_integration_tests/certbot_tests/test_main.py
@@ -8,6 +8,7 @@
import time
from typing import Iterable
from typing import Generator
+from typing import Tuple
from typing import Type
from cryptography.hazmat.primitives.asymmetric.ec import EllipticCurve
@@ -463,6 +464,42 @@ def test_reuse_key(context: IntegrationTestsContext) -> None:
assert len({cert1, cert2, cert3}) == 3
+def test_new_key(context: IntegrationTestsContext) -> None:
+ """Tests --new-key and its interactions with --reuse-key"""
+ def private_key(generation: int) -> Tuple[str, str]:
+ pk_path = join(context.config_dir, f'archive/{certname}/privkey{generation}.pem')
+ with open(pk_path, 'r') as file:
+ return file.read(), pk_path
+
+ certname = context.get_domain('newkey')
+
+ context.certbot(['--domains', certname, '--reuse-key',
+ '--key-type', 'rsa', '--rsa-key-size', '4096'])
+ privkey1, _ = private_key(1)
+
+ # renew: --new-key should replace the key, but keep reuse_key and the key type + params
+ context.certbot(['renew', '--cert-name', certname, '--new-key'])
+ privkey2, privkey2_path = private_key(2)
+ assert privkey1 != privkey2
+ assert_saved_lineage_option(context.config_dir, certname, 'reuse_key', 'True')
+ assert_rsa_key(privkey2_path, 4096)
+
+ # certonly: it should replace the key but the key size will change
+ context.certbot(['certonly', '-d', certname, '--reuse-key', '--new-key'])
+ privkey3, privkey3_path = private_key(3)
+ assert privkey2 != privkey3
+ assert_saved_lineage_option(context.config_dir, certname, 'reuse_key', 'True')
+ assert_rsa_key(privkey3_path, 2048)
+
+ # certonly: it should be possible to change the key type and keep reuse_key
+ context.certbot(['certonly', '-d', certname, '--reuse-key', '--new-key', '--key-type', 'ecdsa',
+ '--cert-name', certname])
+ privkey4, privkey4_path = private_key(4)
+ assert privkey3 != privkey4
+ assert_saved_lineage_option(context.config_dir, certname, 'reuse_key', 'True')
+ assert_elliptic_key(privkey4_path, SECP256R1)
+
+
def test_incorrect_key_type(context: IntegrationTestsContext) -> None:
with pytest.raises(subprocess.CalledProcessError):
context.certbot(['--key-type="failwhale"'])
diff --git a/certbot/CHANGELOG.md b/certbot/CHANGELOG.md
index 3dd1a9e2651..9814e67eafe 100644
--- a/certbot/CHANGELOG.md
+++ b/certbot/CHANGELOG.md
@@ -6,7 +6,10 @@ Certbot adheres to [Semantic Versioning](https://semver.org/).
### Added
-*
+* Added `--new-key`. When renewing or replacing a certificate that has `--reuse-key`
+ set, it will force a new private key to be generated.
+ Combining `--reuse-key` and `--new-key` will replace the certificate's private key
+ and then reuse it for future renewals.
### Changed
diff --git a/certbot/certbot/_internal/cli/__init__.py b/certbot/certbot/_internal/cli/__init__.py
index d11a454b146..30b3fab1d35 100644
--- a/certbot/certbot/_internal/cli/__init__.py
+++ b/certbot/certbot/_internal/cli/__init__.py
@@ -223,6 +223,13 @@ def prepare_and_parse_args(plugins: plugins_disco.PluginsRegistry, args: List[st
"certificate. Not reusing private keys is the default behavior of "
"Certbot. This option may be used to unset --reuse-key on an "
"existing certificate.")
+ helpful.add(
+ "automation", "--new-key",
+ dest="new_key", action="store_true", default=flag_default("new_key"),
+ help="When renewing or replacing a certificate, generate a new private key, "
+ "even if --reuse-key is set on the existing certificate. Combining "
+ "--new-key and --reuse-key will result in the private key being replaced and "
+ "then reused in future renewals.")
helpful.add(
["automation", "renew", "certonly"],
diff --git a/certbot/certbot/_internal/constants.py b/certbot/certbot/_internal/constants.py
index 3867d777c1e..5a9d97d835d 100644
--- a/certbot/certbot/_internal/constants.py
+++ b/certbot/certbot/_internal/constants.py
@@ -74,6 +74,7 @@
validate_hooks=True,
directory_hooks=True,
reuse_key=False,
+ new_key=False,
disable_renew_updates=False,
random_sleep_on_renew=True,
eab_hmac_key=None,
diff --git a/certbot/certbot/_internal/renewal.py b/certbot/certbot/_internal/renewal.py
index 4fb2ca00aa5..0ba2e810802 100644
--- a/certbot/certbot/_internal/renewal.py
+++ b/certbot/certbot/_internal/renewal.py
@@ -336,7 +336,7 @@ def renew_cert(config: configuration.NamespaceConfig, domains: Optional[List[str
domains = lineage.names()
# The private key is the existing lineage private key if reuse_key is set.
# Otherwise, generate a fresh private key by passing None.
- if config.reuse_key:
+ if config.reuse_key and not config.new_key:
new_key = os.path.normpath(lineage.privkey)
_update_renewal_params_from_key(new_key, config)
else:
diff --git a/certbot/certbot/configuration.py b/certbot/certbot/configuration.py
index ebeb8e98c65..d5ad8759957 100644
--- a/certbot/certbot/configuration.py
+++ b/certbot/certbot/configuration.py
@@ -300,6 +300,13 @@ def issuance_timeout(self) -> int:
"""
return self.namespace.issuance_timeout
+ @property
+ def new_key(self) -> bool:
+ """This option specifies whether Certbot should generate a new private
+ key when replacing a certificate, even if reuse_key is set.
+ """
+ return self.namespace.new_key
+
# Magic methods
def __deepcopy__(self, _memo: Any) -> 'NamespaceConfig':
diff --git a/certbot/tests/main_test.py b/certbot/tests/main_test.py
index c29f4d758e1..09a069c6121 100644
--- a/certbot/tests/main_test.py
+++ b/certbot/tests/main_test.py
@@ -1115,7 +1115,7 @@ def test_certonly_new_request_failure(self, mock_subscription):
def _test_renewal_common(self, due_for_renewal, extra_args, log_out=None,
args=None, should_renew=True, error_expected=False,
quiet_mode=False, expiry_date=datetime.datetime.now(),
- reuse_key=False):
+ reuse_key=False, new_key=False):
cert_path = test_util.vector_path('cert_512.pem')
chain_path = os.path.normpath(os.path.join(self.config.config_dir,
'live/foo.bar/fullchain.pem'))
@@ -1165,7 +1165,7 @@ def write_msg(message, *args, **kwargs): # pylint: disable=unused-argument
traceback.format_exc())
if should_renew:
- if reuse_key:
+ if reuse_key and not new_key:
# The location of the previous live privkey.pem is passed
# to obtain_certificate
mock_client.obtain_certificate.assert_called_once_with(['isnot.org'],
@@ -1236,6 +1236,13 @@ def test_reuse_key_no_dry_run(self, unused_save_successor):
args = ["renew", "--reuse-key"]
self._test_renewal_common(True, [], args=args, should_renew=True, reuse_key=True)
+ @mock.patch('certbot._internal.storage.RenewableCert.save_successor')
+ def test_new_key(self, unused_save_successor):
+ test_util.make_lineage(self.config.config_dir, 'sample-renewal.conf')
+ args = ["renew", "--reuse-key", "--new-key"]
+ self._test_renewal_common(True, [], args=args, should_renew=True, reuse_key=True,
+ new_key=True)
+
@mock.patch('sys.stdin')
def test_noninteractive_renewal_delay(self, stdin):
stdin.isatty.return_value = False
diff --git a/certbot/tests/renewal_test.py b/certbot/tests/renewal_test.py
index 110c0d7bd91..d6e2866dc56 100644
--- a/certbot/tests/renewal_test.py
+++ b/certbot/tests/renewal_test.py
@@ -99,6 +99,32 @@ def test_reuse_ec_key_renewal_params(self):
assert self.config.elliptic_curve == 'secp256r1'
+ def test_new_key(self):
+ # When renewing with both reuse_key and new_key, the key should be regenerated,
+ # the key type, key parameters and reuse_key should be kept.
+ self.config.reuse_key = True
+ self.config.new_key = True
+ self.config.dry_run = True
+ config = configuration.NamespaceConfig(self.config)
+
+ rc_path = test_util.make_lineage(
+ self.config.config_dir, 'sample-renewal.conf')
+ lineage = storage.RenewableCert(rc_path, config)
+
+ le_client = mock.MagicMock()
+ le_client.obtain_certificate.return_value = (None, None, None, None)
+
+ from certbot._internal import renewal
+
+ with mock.patch('certbot._internal.renewal.hooks.renew_hook'):
+ renewal.renew_cert(self.config, None, le_client, lineage)
+
+ self.assertEqual(self.config.rsa_key_size, 2048)
+ self.assertEqual(self.config.key_type, 'rsa')
+ self.assertTrue(self.config.reuse_key)
+ # None is passed as the existing key, i.e. the key is not actually being reused.
+ le_client.obtain_certificate.assert_called_with(mock.ANY, None)
+
@test_util.patch_display_util()
@mock.patch('certbot._internal.renewal.cli.set_by_cli')
def test_remove_deprecated_config_elements(self, mock_set_by_cli, unused_mock_get_utility):
| https://api.github.com/repos/certbot/certbot/pulls/9252 | 2022-03-29T21:50:13Z | 2022-03-31T18:40:22Z | 2022-03-31T18:40:22Z | 2022-03-31T18:40:22Z | 2,862 | certbot/certbot | 116 |
|
added possibility to freeze layers during training | diff --git a/examples/mnist_transfer_cnn.py b/examples/mnist_transfer_cnn.py
new file mode 100644
index 00000000000..579df76d08f
--- /dev/null
+++ b/examples/mnist_transfer_cnn.py
@@ -0,0 +1,114 @@
+from __future__ import absolute_import
+from __future__ import print_function
+import numpy as np
+import datetime
+
+np.random.seed(1337) # for reproducibility
+
+from keras.datasets import mnist
+from keras.models import Sequential
+from keras.layers.core import Dense, Dropout, Activation, Flatten
+from keras.layers.convolutional import Convolution2D, MaxPooling2D
+from keras.utils import np_utils
+
+'''
+ Transfer learning toy example:
+ 1- Train a simple convnet on the MNIST dataset the first 5 digits [0..4].
+ 2- Freeze convolutional layers and fine-tune dense layers for the classification of digits [5..9].
+
+ Run on GPU: THEANO_FLAGS=mode=FAST_RUN,device=gpu,floatX=float32 python mnist_cnn.py
+
+ Get to 99.8% test accuracy after 5 epochs for the first five digits classifier
+ and 99.2% for the last five digits after transfer + fine-tuning.
+'''
+
+now = datetime.datetime.now
+
+batch_size = 128
+nb_classes = 5
+nb_epoch = 5
+
+# input image dimensions
+img_rows, img_cols = 28, 28
+# number of convolutional filters to use
+nb_filters = 32
+# size of pooling area for max pooling
+nb_pool = 2
+# convolution kernel size
+nb_conv = 3
+
+
+def train_model(model, train, test, nb_classes):
+ X_train = train[0].reshape(train[0].shape[0], 1, img_rows, img_cols)
+ X_test = test[0].reshape(test[0].shape[0], 1, img_rows, img_cols)
+ X_train = X_train.astype("float32")
+ X_test = X_test.astype("float32")
+ X_train /= 255
+ X_test /= 255
+ print('X_train shape:', X_train.shape)
+ print(X_train.shape[0], 'train samples')
+ print(X_test.shape[0], 'test samples')
+
+ # convert class vectors to binary class matrices
+ Y_train = np_utils.to_categorical(train[1], nb_classes)
+ Y_test = np_utils.to_categorical(test[1], nb_classes)
+
+ model.compile(loss='categorical_crossentropy', optimizer='adadelta')
+
+ t = now()
+ model.fit(X_train, Y_train, batch_size=batch_size, nb_epoch=nb_epoch, show_accuracy=True, verbose=1,
+ validation_data=(X_test, Y_test))
+ print('Training time: %s' % (now() - t))
+ score = model.evaluate(X_test, Y_test, show_accuracy=True, verbose=0)
+ print('Test score:', score[0])
+ print('Test accuracy:', score[1])
+
+
+# the data, shuffled and split between train and test sets
+(X_train, y_train), (X_test, y_test) = mnist.load_data()
+
+# create two datasets one with digits below 5 and one with 5 and above
+X_train_lt5 = X_train[y_train < 5]
+y_train_lt5 = y_train[y_train < 5]
+X_test_lt5 = X_test[y_test < 5]
+y_test_lt5 = y_test[y_test < 5]
+
+X_train_gte5 = X_train[y_train >= 5]
+y_train_gte5 = y_train[y_train >= 5] - 5 # make classes start at 0 for
+X_test_gte5 = X_test[y_test >= 5] # np_utils.to_categorical
+y_test_gte5 = y_test[y_test >= 5] - 5
+
+# define two groups of layers: feature (convolutions) and classification (dense)
+feature_layers = [
+ Convolution2D(nb_filters, nb_conv, nb_conv,
+ border_mode='full',
+ input_shape=(1, img_rows, img_cols)),
+ Activation('relu'),
+ Convolution2D(nb_filters, nb_conv, nb_conv),
+ Activation('relu'),
+ MaxPooling2D(pool_size=(nb_pool, nb_pool)),
+ Dropout(0.25),
+ Flatten(),
+]
+classification_layers = [
+ Dense(128),
+ Activation('relu'),
+ Dropout(0.5),
+ Dense(nb_classes),
+ Activation('softmax')
+]
+
+# create complete model
+model = Sequential()
+for l in feature_layers + classification_layers:
+ model.add(l)
+
+# train model for 5-digit classification [0..4]
+train_model(model, (X_train_lt5, y_train_lt5), (X_test_lt5, y_test_lt5), nb_classes)
+
+# freeze feature layers and rebuild model
+for l in feature_layers:
+ l.trainable = False
+
+# transfer: train dense layers for new classification task [5..9]
+train_model(model, (X_train_gte5, y_train_gte5), (X_test_gte5, y_test_gte5), nb_classes)
diff --git a/keras/layers/containers.py b/keras/layers/containers.py
index 8c4ea8604c3..422359a9037 100644
--- a/keras/layers/containers.py
+++ b/keras/layers/containers.py
@@ -2,6 +2,7 @@
from __future__ import absolute_import
from __future__ import print_function
+from collections import OrderedDict
import theano.tensor as T
from ..layers.core import Layer, Merge
from ..utils.theano_utils import ndim_tensor
@@ -20,11 +21,6 @@ class Sequential(Layer):
def __init__(self, layers=[]):
self.layers = []
- self.params = []
- self.regularizers = []
- self.constraints = []
- self.updates = []
-
for layer in layers:
self.add(layer)
@@ -38,11 +34,37 @@ def add(self, layer):
if not hasattr(self.layers[0], 'input'):
self.set_input()
- params, regularizers, constraints, updates = layer.get_params()
- self.params += params
- self.regularizers += regularizers
- self.constraints += constraints
- self.updates += updates
+ @property
+ def params(self):
+ params = []
+ for l in self.layers:
+ if l.trainable:
+ params += l.get_params()[0]
+ return params
+
+ @property
+ def regularizers(self):
+ regularizers = []
+ for l in self.layers:
+ if l.trainable:
+ regularizers += l.get_params()[1]
+ return regularizers
+
+ @property
+ def constraints(self):
+ constraints = []
+ for l in self.layers:
+ if l.trainable:
+ constraints += l.get_params()[2]
+ return constraints
+
+ @property
+ def updates(self):
+ updates = []
+ for l in self.layers:
+ if l.trainable:
+ updates += l.get_params()[3]
+ return updates
@property
def output_shape(self):
@@ -97,7 +119,6 @@ class Graph(Layer):
when it has exactly one input and one output.
inherited from Layer:
- - get_params
- get_output_mask
- supports_masked_input
- get_weights
@@ -105,7 +126,7 @@ class Graph(Layer):
'''
def __init__(self):
self.namespace = set() # strings
- self.nodes = {} # layer-like
+ self.nodes = OrderedDict() # layer-like
self.inputs = {} # layer-like
self.input_order = [] # strings
self.outputs = {} # layer-like
@@ -114,11 +135,6 @@ def __init__(self):
self.output_config = [] # dicts
self.node_config = [] # dicts
- self.params = []
- self.regularizers = []
- self.constraints = []
- self.updates = []
-
@property
def nb_input(self):
return len(self.inputs)
@@ -127,6 +143,38 @@ def nb_input(self):
def nb_output(self):
return len(self.outputs)
+ @property
+ def params(self):
+ params = []
+ for l in self.nodes.values():
+ if l.trainable:
+ params += l.get_params()[0]
+ return params
+
+ @property
+ def regularizers(self):
+ regularizers = []
+ for l in self.nodes.values():
+ if l.trainable:
+ regularizers += l.get_params()[1]
+ return regularizers
+
+ @property
+ def constraints(self):
+ constraints = []
+ for l in self.nodes.values():
+ if l.trainable:
+ constraints += l.get_params()[2]
+ return constraints
+
+ @property
+ def updates(self):
+ updates = []
+ for l in self.nodes.values():
+ if l.trainable:
+ updates += l.get_params()[3]
+ return updates
+
def set_previous(self, layer, connection_map={}):
if self.nb_input != layer.nb_output:
raise Exception('Cannot connect layers: input count does not match output count.')
@@ -220,11 +268,6 @@ def add_node(self, layer, name, input=None, inputs=[],
'merge_mode': merge_mode,
'concat_axis': concat_axis,
'create_output': create_output})
- params, regularizers, constraints, updates = layer.get_params()
- self.params += params
- self.regularizers += regularizers
- self.constraints += constraints
- self.updates += updates
if create_output:
self.add_output(name, input=name)
diff --git a/keras/layers/core.py b/keras/layers/core.py
index 5c216b0d91c..f21a3cbefae 100644
--- a/keras/layers/core.py
+++ b/keras/layers/core.py
@@ -20,9 +20,11 @@
class Layer(object):
def __init__(self, **kwargs):
for kwarg in kwargs:
- assert kwarg in {'input_shape'}, "Keyword argument not understood: " + kwarg
+ assert kwarg in {'input_shape', 'trainable'}, "Keyword argument not understood: " + kwarg
if 'input_shape' in kwargs:
self.set_input_shape(kwargs['input_shape'])
+ if 'trainable' in kwargs:
+ self._trainable = kwargs['trainable']
if not hasattr(self, 'params'):
self.params = []
@@ -45,6 +47,17 @@ def build(self):
'''
pass
+ @property
+ def trainable(self):
+ if hasattr(self, '_trainable'):
+ return self._trainable
+ else:
+ return True
+
+ @trainable.setter
+ def trainable(self, value):
+ self._trainable = value
+
@property
def nb_input(self):
return 1
@@ -133,6 +146,8 @@ def get_config(self):
config = {"name": self.__class__.__name__}
if hasattr(self, '_input_shape'):
config['input_shape'] = self._input_shape[1:]
+ if hasattr(self, '_trainable'):
+ config['trainable'] = self._trainable
return config
def get_params(self):
| Following the discussion in #622, here is a simple strategy to "freeze" specific layers during training. IT works for `Sequential` and `Graph` models.
It adds a `trainable` attribute to the core `Layer` class and check its value when creating the model. If a layer's `trainable` flag is modified after addition to a model, the model needs to be rebuilt and recompiled. For this, a `rebuild` function has been added to `Sequential` and `Graph` classes.
| https://api.github.com/repos/keras-team/keras/pulls/839 | 2015-10-15T11:41:58Z | 2015-10-23T23:40:38Z | 2015-10-23T23:40:38Z | 2015-12-31T15:30:46Z | 2,642 | keras-team/keras | 47,887 |
exmo(fix): ArgumentsRequired import missing | diff --git a/ts/src/exmo.ts b/ts/src/exmo.ts
index 180b9a2a1232..6411a68950ef 100644
--- a/ts/src/exmo.ts
+++ b/ts/src/exmo.ts
@@ -2,7 +2,7 @@
// ---------------------------------------------------------------------------
import Exchange from './abstract/exmo.js';
-import { ExchangeError, OrderNotFound, AuthenticationError, InsufficientFunds, InvalidOrder, InvalidNonce, OnMaintenance, RateLimitExceeded, BadRequest, PermissionDenied } from './base/errors.js';
+import { ArgumentsRequired, ExchangeError, OrderNotFound, AuthenticationError, InsufficientFunds, InvalidOrder, InvalidNonce, OnMaintenance, RateLimitExceeded, BadRequest, PermissionDenied } from './base/errors.js';
import { Precise } from './base/Precise.js';
import { TICK_SIZE } from './base/functions/number.js';
import { sha512 } from './static_dependencies/noble-hashes/sha512.js';
| https://api.github.com/repos/ccxt/ccxt/pulls/19307 | 2023-09-17T00:45:16Z | 2023-09-17T07:16:27Z | 2023-09-17T07:16:27Z | 2023-09-17T07:16:28Z | 207 | ccxt/ccxt | 13,407 |
|
Safety pipeline prototype | diff --git a/safety/README.md b/safety/README.md
new file mode 100644
index 0000000000..2b6f23ff0f
--- /dev/null
+++ b/safety/README.md
@@ -0,0 +1,47 @@
+# Train & Evaluate Safety models
+
+This is the Open Assistant Safety Folder and contains the following:
+
+- Model training scripts
+- Model infrence scripts
+- Data processing scripts
+
+## Mission Statment
+
+Our mission at LAION-AI OpenAssistant safety team is to create a safety pipeline
+that is not only compatible with the OpenAssistant model and project but can
+also integrate with other systems outside of it. We are dedicated to making this
+pipeline modifiable and robust to accommodate the diverse preferences of our
+users.
+
+We understand that our users come from different backgrounds and use various
+types of hardware. Therefore, we strive to make our safety pipeline accessible
+and able to run on consumer hardware, so everyone can benefit from its
+protective features.
+
+Through our commitment to innovation and collaboration, we will continue to
+provide safety solutions that ensure the well-being of our users and the wider
+community.
+
+## Why create a safety pipeline?
+
+Open source and extendable safety pipelines unfortunately do not exist on the
+same on the same scale as those in ChatGPT and other commerical systems. To
+further research in implementable, accurate, and extendable safety pipelines,
+Open Assistant Safety Team will continue to push models and code to the public.
+Much research has been done in things like toxicity detection and bias
+mitigation in LLMs, however the implementation of such research in systems that
+use language models as conversational agents in production settings has largely
+gone undocumented. Furthermore, safety systems that interact with diverse
+communities of users must be able accommodate user prefrences. This is paramount
+in introducing LLM based systems all over the world. We hope that our work will
+generate more research in this field, and allow others to create safe LLM based
+systems.
+
+## Training
+
+- Set training configuration using `config.yaml`
+
+```python
+python model_training/t5_trainer.py
+```
diff --git a/safety/model_training/__init__.py b/safety/model_training/__init__.py
new file mode 100644
index 0000000000..e69de29bb2
diff --git a/safety/model_training/config/config.yaml b/safety/model_training/config/config.yaml
new file mode 100644
index 0000000000..35cace8444
--- /dev/null
+++ b/safety/model_training/config/config.yaml
@@ -0,0 +1,18 @@
+defaults:
+ - trainer: default
+padding_side: "right"
+truncation_side: "right"
+model: "t5-base"
+epochs: 1
+batch_size: 8
+save_folder: "safetyfiles"
+max_length: 256
+special_tokens:
+ context_token: "<ctx>"
+ sep_token: "<sep>"
+ label_token: "<cls>"
+ rot_token: "<rot>"
+dataset:
+ name: "allenai/prosocial-dialog"
+ train: ["train", "validation"]
+ test: "test"
diff --git a/safety/model_training/config/trainer/default.yaml b/safety/model_training/config/trainer/default.yaml
new file mode 100644
index 0000000000..b13dc0b7e7
--- /dev/null
+++ b/safety/model_training/config/trainer/default.yaml
@@ -0,0 +1,4 @@
+_target_: transformers.TrainingArguments
+output_dir: "."
+per_device_train_batch_size: 5
+fp16: False
diff --git a/safety/model_training/custom_datasets/__init__.py b/safety/model_training/custom_datasets/__init__.py
new file mode 100644
index 0000000000..e69de29bb2
diff --git a/safety/model_training/custom_datasets/rot_dataset.py b/safety/model_training/custom_datasets/rot_dataset.py
new file mode 100644
index 0000000000..171451a595
--- /dev/null
+++ b/safety/model_training/custom_datasets/rot_dataset.py
@@ -0,0 +1,103 @@
+from dataclasses import dataclass
+from typing import Dict, List
+
+import torch
+from datasets import concatenate_datasets
+from torch.utils.data import Dataset
+
+LABEL2ID = {
+ "__casual__": "__casual__",
+ "__needs_caution__": "__needs_caution__",
+ "__needs_intervention__": "__needs_intervention__",
+ "__probably_needs_caution__": "__probably_needs_caution__",
+ "__possibly_needs_caution__": "__possibly_needs_caution__",
+}
+
+
+class SafetyDataset(Dataset):
+
+ """
+ Dataset to train safety model with context and ROT from prosocial-dialog
+ input format : input<ctx>context</s>
+ output format : <cls>safety_label<rot>ROTs</s>
+
+ """
+
+ def __init__(self, dataset, split, tokenizer, max_len=512):
+ super().__init__()
+
+ if isinstance(split, List):
+ self.split = "-".join(split)
+ self.dataset = concatenate_datasets([dataset[sp] for sp in split])
+ else:
+ self.split = split
+ self.dataset = dataset[split]
+
+ self.max_len = max_len
+ self.tokenizer = tokenizer
+ self.label2id = LABEL2ID
+
+ def __len__(self):
+ return len(self.dataset)
+
+ def __getitem__(self, idx):
+ idx_start = idx
+ end = self.dataset[max(0, idx_start - 1)]["episode_done"]
+ while (not end) and (idx_start > 0):
+ end = self.dataset[max(0, idx_start - 2)]["episode_done"]
+ idx_start -= 1
+ idx_start = max(0, idx_start)
+ context = [
+ f'\nUser: {self.dataset[i]["context"]}\n bot:{self.dataset[i]["response"]}' for i in range(idx_start, idx)
+ ]
+ context = self.tokenizer.sep_token.join(context)
+ rots = self.dataset[idx]["rots"]
+ label = self.label2id[self.dataset[idx]["safety_label"]]
+ input_tokens = self.tokenizer.encode(self.dataset[idx]["context"], add_special_tokens=False)
+ max_len = self.max_len - (len(input_tokens) + 2)
+ context = self.tokenizer.encode(
+ context,
+ add_special_tokens=False,
+ max_length=max_len,
+ )
+ rots = self.tokenizer.sep_token.join(rots)
+ input_ids = input_tokens + [self.tokenizer.context_token_id] + context + [self.tokenizer.eos_token_id]
+ input_ids = input_ids + [self.tokenizer.pad_token_id] * max(0, (self.max_len - len(input_ids)))
+ mask = [1] * len(input_ids) + [self.tokenizer.pad_token_id] * (self.max_len - len(input_ids))
+ target_text = self.tokenizer.label_token + label + self.tokenizer.context_token + rots
+ decoder_ids = self.tokenizer(
+ target_text,
+ add_special_tokens=True,
+ max_length=self.max_len,
+ padding="max_length",
+ )
+
+ return {
+ "input_ids": torch.LongTensor(input_ids),
+ "attention_mask": torch.LongTensor(mask),
+ "decoder_input_ids": torch.LongTensor(decoder_ids["input_ids"]),
+ "decoder_attention_mask": torch.LongTensor(decoder_ids["attention_mask"]),
+ }
+
+
+@dataclass
+class SafetyDataCollator:
+ def __call__(self, batch: List) -> Dict[str, torch.Tensor]:
+ """
+ Take a list of samples from a Dataset and collate them into a batch.
+ Returns:
+ A dictionary of tensors
+ """
+
+ input_ids = torch.stack([example["input_ids"] for example in batch])
+ lm_labels = torch.stack([example["decoder_input_ids"] for example in batch])
+ lm_labels[lm_labels[:, :] == 0] = -100
+ attention_mask = torch.stack([example["attention_mask"] for example in batch])
+ decoder_attention_mask = torch.stack([example["decoder_attention_mask"] for example in batch])
+
+ return {
+ "input_ids": input_ids,
+ "attention_mask": attention_mask,
+ "labels": lm_labels,
+ "decoder_attention_mask": decoder_attention_mask,
+ }
diff --git a/safety/model_training/t5_trainer.py b/safety/model_training/t5_trainer.py
new file mode 100644
index 0000000000..2776803b61
--- /dev/null
+++ b/safety/model_training/t5_trainer.py
@@ -0,0 +1,50 @@
+import os
+
+import hydra
+from custom_datasets.rot_dataset import SafetyDataCollator, SafetyDataset
+from datasets import load_dataset
+from hydra.utils import instantiate
+from omegaconf import DictConfig, OmegaConf
+from transformers import T5ForConditionalGeneration, T5Tokenizer, Trainer
+from utils import add_special_tokens
+
+
[email protected](version_base=None, config_path="config", config_name="config")
+def train(cfg: DictConfig) -> None:
+ if not os.path.exists(cfg.save_folder):
+ os.mkdir(cfg.save_folder)
+
+ model = T5ForConditionalGeneration.from_pretrained(cfg.model)
+ tokenizer = T5Tokenizer.from_pretrained(
+ cfg.model,
+ padding_side=cfg.padding_side,
+ truncation_side=cfg.truncation_side,
+ model_max_length=model.config.n_positions,
+ )
+ add_special_tokens(cfg.special_tokens, tokenizer, model)
+ training_args = instantiate(cfg.trainer)
+
+ dataset = load_dataset(cfg.dataset.name)
+ train_dataset = SafetyDataset(
+ dataset, split=OmegaConf.to_object(cfg.dataset.train), tokenizer=tokenizer, max_len=cfg.max_length
+ )
+ valid_dataset = SafetyDataset(dataset, split=cfg.dataset.test, tokenizer=tokenizer, max_len=cfg.max_length)
+
+ # Initialize our Trainer
+ trainer = Trainer(
+ model=model,
+ args=training_args,
+ train_dataset=train_dataset,
+ eval_dataset=valid_dataset,
+ data_collator=SafetyDataCollator(),
+ )
+
+ # Training
+ trainer.train()
+
+ trainer.save_model(os.path.join(cfg.save_folder, f"{cfg.model_name}-model"))
+ tokenizer.save_vocabulary(os.path.join(cfg.save_folder, f"{cfg.model_name}-tokenizer"))
+
+
+if __name__ == "__main__":
+ train()
diff --git a/safety/model_training/utils.py b/safety/model_training/utils.py
new file mode 100644
index 0000000000..8102ea1056
--- /dev/null
+++ b/safety/model_training/utils.py
@@ -0,0 +1,7 @@
+def add_special_tokens(special_tokens, tokenizer, model):
+ for key, value in special_tokens.items():
+ setattr(tokenizer, key, value)
+ tokenizer.add_tokens([value])
+ setattr(tokenizer, key + "_id", tokenizer.encode(value)[0])
+
+ model.resize_token_embeddings(len(tokenizer))
diff --git a/safety/requirements.txt b/safety/requirements.txt
new file mode 100644
index 0000000000..dffd40c36a
--- /dev/null
+++ b/safety/requirements.txt
@@ -0,0 +1,2 @@
+hydra-core==1.3.2
+omegaconf==2.3.0
| The safety team has decided to maintain a `safety` folder. This folder will contain
1. Documentation related to safety models.
2. Reproducible training and evaluation code for safety models.
3. Inference pipelines for safety models
@ontocord
| https://api.github.com/repos/LAION-AI/Open-Assistant/pulls/1972 | 2023-03-05T11:08:36Z | 2023-03-08T16:33:32Z | 2023-03-08T16:33:32Z | 2023-03-08T16:33:32Z | 2,637 | LAION-AI/Open-Assistant | 37,546 |
Fix issue with scrolling when package not found | diff --git a/thefuck/specific/archlinux.py b/thefuck/specific/archlinux.py
index 5816c50f0..5c95aa5bb 100644
--- a/thefuck/specific/archlinux.py
+++ b/thefuck/specific/archlinux.py
@@ -24,8 +24,11 @@ def get_pkgfile(command):
).splitlines()
return [package.split()[0] for package in packages]
- except subprocess.CalledProcessError:
- return None
+ except subprocess.CalledProcessError as err:
+ if err.returncode == 1 and err.output == "":
+ return []
+ else:
+ raise err
def archlinux_env():
| Fix issue with attempting to scroll through possible corrections when not-found package has no packages with matching names causing crash.
This behavior originates with commit 6624ecb3b85e82e9f1a08823f6e41ee805d35a9e, to my knowledge, where pacman rules were created. In both uses of `get_pkgfile` (in pacman.py and in pacman_not_found.py), it would be appropriate for `get_pkgfile` to return an empty list when the `pkgfile` does not find any packages by the correct name. However, the `pkgfile` command returns 1 in these circumstances, and as such `subprocess` raises an exception. Now, when this exception is caused by `pkgfile` returning 1 with no output (when it finds no packages), `get_pkgfile` will not cause a crash.
To reproduce bug:
```
yaourt -S e
fuck
```
then press ↑ or ↓ | https://api.github.com/repos/nvbn/thefuck/pulls/573 | 2016-11-03T07:50:06Z | 2017-03-13T12:47:17Z | 2017-03-13T12:47:17Z | 2017-03-13T12:47:24Z | 157 | nvbn/thefuck | 30,457 |
Fix typo | diff --git a/README.md b/README.md
index 0b576befe1..c11c8f3032 100644
--- a/README.md
+++ b/README.md
@@ -196,7 +196,7 @@ API | Description | Auth | HTTPS | CORS |
| [Open Library](https://openlibrary.org/developers/api) | Books, book covers and related data | No | Yes | No |
| [Penguin Publishing](http://www.penguinrandomhouse.biz/webservices/rest/) | Books, book covers and related data | No | Yes | Yes |
| [Quran](https://quran.api-docs.io/) | RESTful Quran API with multiple languages | No | Yes | Yes |
-| [Quran Cloud](https://alquran.cloud/api) | A RESTful Quran API to retrieve an Ayah, Surah, Juz or the enitre Holy Quran | No | Yes | Yes |
+| [Quran Cloud](https://alquran.cloud/api) | A RESTful Quran API to retrieve an Ayah, Surah, Juz or the entire Holy Quran | No | Yes | Yes |
| [Quran-api](https://github.com/fawazahmed0/quran-api#readme) | Free Quran API Service with 90+ different languages and 400+ translations | No | Yes | Yes |
| [Rig Veda](https://aninditabasu.github.io/indica/html/rv.html) | Gods and poets, their categories, and the verse meters, with the mandal and sukta number | No | Yes | Unknown |
| [The Bible](https://docs.api.bible) | Everything you need from the Bible in one discoverable place | `apiKey` | Yes | Unknown |
| <!-- Thank you for taking the time to work on a Pull Request for this project! -->
<!-- To ensure your PR is dealt with swiftly please check the following: -->
- [X ] My submission is formatted according to the guidelines in the [contributing guide](/CONTRIBUTING.md)
- [X ] My addition is ordered alphabetically
- [X ] My submission has a useful description
- [X ] The description does not end with punctuation
- [X ] Each table column is padded with one space on either side
- [X ] I have searched the repository for any relevant issues or pull requests
- [X ] Any category I am creating has the minimum requirement of 3 items
- [X ] All changes have been [squashed][squash-link] into a single commit
[squash-link]: <https://github.com/todotxt/todo.txt-android/wiki/Squash-All-Commits-Related-to-a-Single-Issue-into-a-Single-Commit>
| https://api.github.com/repos/public-apis/public-apis/pulls/2062 | 2021-09-21T00:44:38Z | 2021-09-22T11:31:59Z | 2021-09-22T11:31:59Z | 2021-09-22T18:06:23Z | 376 | public-apis/public-apis | 35,235 |
Inline `_make_grid()` meshgrid | diff --git a/models/yolo.py b/models/yolo.py
index 7a7308312a1..fa05fcf9a8d 100644
--- a/models/yolo.py
+++ b/models/yolo.py
@@ -81,10 +81,7 @@ def _make_grid(self, nx=20, ny=20, i=0, torch_1_10=check_version(torch.__version
t = self.anchors[i].dtype
shape = 1, self.na, ny, nx, 2 # grid shape
y, x = torch.arange(ny, device=d, dtype=t), torch.arange(nx, device=d, dtype=t)
- if torch_1_10: # torch>=1.10.0 meshgrid workaround for torch>=0.7 compatibility
- yv, xv = torch.meshgrid(y, x, indexing='ij')
- else:
- yv, xv = torch.meshgrid(y, x)
+ yv, xv = torch.meshgrid(y, x, indexing='ij') if torch_1_10 else torch.meshgrid(y, x) # torch>=0.7 compatibility
grid = torch.stack((xv, yv), 2).expand(shape) - 0.5 # add grid offset, i.e. y = 2.0 * x - 0.5
anchor_grid = (self.anchors[i] * self.stride[i]).view((1, self.na, 1, 1, 2)).expand(shape)
return grid, anchor_grid
| Signed-off-by: Glenn Jocher <[email protected]>
<!--
Thank you for submitting a YOLOv5 🚀 Pull Request! We want to make contributing to YOLOv5 as easy and transparent as possible. A few tips to get you started:
- Search existing YOLOv5 [PRs](https://github.com/ultralytics/yolov5/pull) to see if a similar PR already exists.
- Link this PR to a YOLOv5 [issue](https://github.com/ultralytics/yolov5/issues) to help us understand what bug fix or feature is being implemented.
- Provide before and after profiling/inference/training results to help us quantify the improvement your PR provides (if applicable).
Please see our ✅ [Contributing Guide](https://github.com/ultralytics/yolov5/blob/master/CONTRIBUTING.md) for more details.
-->
## 🛠️ PR Summary
<sub>Made with ❤️ by [Ultralytics Actions](https://github.com/ultralytics/actions)<sub>
### 🌟 Summary
Refinement of grid generation in YOLOv5 for compatibility adjustment.
### 📊 Key Changes
- Consolidated conditional code for grid generation into a single line.
### 🎯 Purpose & Impact
- 🧹 Cleans up the code by removing unnecessary conditionals.
- 🚀 Simplifies maintenance by reducing code complexity.
- ✨ Ensures compatibility with various versions of PyTorch (>=0.7) without altering functionality. | https://api.github.com/repos/ultralytics/yolov5/pulls/9170 | 2022-08-26T13:07:52Z | 2022-08-26T13:29:31Z | 2022-08-26T13:29:31Z | 2024-01-19T06:39:33Z | 342 | ultralytics/yolov5 | 24,732 |
gh-108765: Python.h no longer includes <unistd.h> | diff --git a/Doc/whatsnew/3.13.rst b/Doc/whatsnew/3.13.rst
index e7b60ddbdbbfda..c4fb328db9cfa0 100644
--- a/Doc/whatsnew/3.13.rst
+++ b/Doc/whatsnew/3.13.rst
@@ -921,6 +921,11 @@ Porting to Python 3.13
also the ``HAVE_IEEEFP_H`` macro.
(Contributed by Victor Stinner in :gh:`108765`.)
+* ``Python.h`` no longer includes the ``<unistd.h>`` standard header file. If
+ needed, it should now be included explicitly. For example, it provides the
+ functions: ``close()``, ``getpagesize()``, ``getpid()`` and ``sysconf()``.
+ (Contributed by Victor Stinner in :gh:`108765`.)
+
Deprecated
----------
diff --git a/Include/Python.h b/Include/Python.h
index 002a79dbdc9362..4cc72bb23ce7a3 100644
--- a/Include/Python.h
+++ b/Include/Python.h
@@ -26,14 +26,13 @@
#ifdef HAVE_STDDEF_H
# include <stddef.h> // size_t
#endif
-#ifndef MS_WINDOWS
-# include <unistd.h> // sysconf()
+#ifdef HAVE_SYS_TYPES_H
+# include <sys/types.h> // ssize_t
#endif
-// errno.h, stdio.h, stdlib.h and string.h headers are no longer used by Python
-// headers, but kept for backward compatibility (no introduce new compiler
-// warnings). They are not included by the limited C API version 3.11 and
-// above.
+// errno.h, stdio.h, stdlib.h and string.h headers are no longer used by
+// Python, but kept for backward compatibility (avoid compiler warnings).
+// They are no longer included by limited C API version 3.11 and newer.
#if !defined(Py_LIMITED_API) || Py_LIMITED_API+0 < 0x030b0000
# include <errno.h> // errno
# include <stdio.h> // FILE*
diff --git a/Misc/NEWS.d/next/C API/2023-09-01-21-10-29.gh-issue-108765.eeXtYF.rst b/Misc/NEWS.d/next/C API/2023-09-01-21-10-29.gh-issue-108765.eeXtYF.rst
new file mode 100644
index 00000000000000..ff8f79998fa968
--- /dev/null
+++ b/Misc/NEWS.d/next/C API/2023-09-01-21-10-29.gh-issue-108765.eeXtYF.rst
@@ -0,0 +1,4 @@
+``Python.h`` no longer includes the ``<unistd.h>`` standard header file. If
+needed, it should now be included explicitly. For example, it provides the
+functions: ``close()``, ``getpagesize()``, ``getpid()`` and ``sysconf()``.
+Patch by Victor Stinner.
diff --git a/Modules/_ctypes/malloc_closure.c b/Modules/_ctypes/malloc_closure.c
index 3a859322772ba7..bb4f8f21bd3f77 100644
--- a/Modules/_ctypes/malloc_closure.c
+++ b/Modules/_ctypes/malloc_closure.c
@@ -1,16 +1,17 @@
#ifndef Py_BUILD_CORE_BUILTIN
# define Py_BUILD_CORE_MODULE 1
#endif
+
#include <Python.h>
#include <ffi.h>
#ifdef MS_WIN32
-#include <windows.h>
+# include <windows.h>
#else
-#include <sys/mman.h>
-#include <unistd.h>
-# if !defined(MAP_ANONYMOUS) && defined(MAP_ANON)
-# define MAP_ANONYMOUS MAP_ANON
-# endif
+# include <sys/mman.h>
+# include <unistd.h> // sysconf()
+# if !defined(MAP_ANONYMOUS) && defined(MAP_ANON)
+# define MAP_ANONYMOUS MAP_ANON
+# endif
#endif
#include "ctypes.h"
diff --git a/Modules/_posixsubprocess.c b/Modules/_posixsubprocess.c
index ac2b0d4f55468c..ef76d26282e1b3 100644
--- a/Modules/_posixsubprocess.c
+++ b/Modules/_posixsubprocess.c
@@ -8,28 +8,28 @@
#include "pycore_pystate.h"
#include "pycore_signal.h" // _Py_RestoreSignals()
#if defined(HAVE_PIPE2) && !defined(_GNU_SOURCE)
-# define _GNU_SOURCE
+# define _GNU_SOURCE
#endif
-#include <unistd.h>
-#include <fcntl.h>
+#include <unistd.h> // close()
+#include <fcntl.h> // fcntl()
#ifdef HAVE_SYS_TYPES_H
-#include <sys/types.h>
+# include <sys/types.h>
#endif
#if defined(HAVE_SYS_STAT_H)
-#include <sys/stat.h>
+# include <sys/stat.h> // stat()
#endif
#ifdef HAVE_SYS_SYSCALL_H
-#include <sys/syscall.h>
+# include <sys/syscall.h>
#endif
#if defined(HAVE_SYS_RESOURCE_H)
-#include <sys/resource.h>
+# include <sys/resource.h>
#endif
#ifdef HAVE_DIRENT_H
-#include <dirent.h>
+# include <dirent.h> // opendir()
+#endif
+#if defined(HAVE_SETGROUPS)
+# include <grp.h> // setgroups()
#endif
-#ifdef HAVE_GRP_H
-#include <grp.h>
-#endif /* HAVE_GRP_H */
#include "posixmodule.h"
diff --git a/Modules/_testcapimodule.c b/Modules/_testcapimodule.c
index 4fc354ae79bfed..ab33702cdfd872 100644
--- a/Modules/_testcapimodule.c
+++ b/Modules/_testcapimodule.c
@@ -24,9 +24,6 @@
#include <float.h> // FLT_MAX
#include <signal.h>
#include <stddef.h> // offsetof()
-#ifndef MS_WINDOWS
-# include <unistd.h>
-#endif
#ifdef HAVE_SYS_WAIT_H
# include <sys/wait.h> // W_STOPCODE
diff --git a/Modules/grpmodule.c b/Modules/grpmodule.c
index f5709296334a8f..20e83de84e8340 100644
--- a/Modules/grpmodule.c
+++ b/Modules/grpmodule.c
@@ -4,7 +4,8 @@
#include "Python.h"
#include "posixmodule.h"
-#include <grp.h>
+#include <grp.h> // getgrgid_r()
+#include <unistd.h> // sysconf()
#include "clinic/grpmodule.c.h"
/*[clinic input]
diff --git a/Modules/mmapmodule.c b/Modules/mmapmodule.c
index c8cd7e59dbab50..d11200a4042551 100644
--- a/Modules/mmapmodule.c
+++ b/Modules/mmapmodule.c
@@ -28,6 +28,9 @@
#include "pycore_fileutils.h" // _Py_stat_struct
#include <stddef.h> // offsetof()
+#ifndef MS_WINDOWS
+# include <unistd.h> // close()
+#endif
// to support MS_WINDOWS_SYSTEM OpenFileMappingA / CreateFileMappingA
// need to be replaced with OpenFileMappingW / CreateFileMappingW
diff --git a/Modules/posixmodule.c b/Modules/posixmodule.c
index 761542866d8f96..6e829b200fa46d 100644
--- a/Modules/posixmodule.c
+++ b/Modules/posixmodule.c
@@ -286,7 +286,7 @@ corresponding Unix manual entries for more information on calls.");
#endif
#ifdef HAVE_COPY_FILE_RANGE
-# include <unistd.h>
+# include <unistd.h> // copy_file_range()
#endif
#if !defined(CPU_ALLOC) && defined(HAVE_SCHED_SETAFFINITY)
diff --git a/Modules/pwdmodule.c b/Modules/pwdmodule.c
index cc2e2a43893971..b7034369c4731e 100644
--- a/Modules/pwdmodule.c
+++ b/Modules/pwdmodule.c
@@ -4,7 +4,8 @@
#include "Python.h"
#include "posixmodule.h"
-#include <pwd.h>
+#include <pwd.h> // getpwuid()
+#include <unistd.h> // sysconf()
#include "clinic/pwdmodule.c.h"
/*[clinic input]
diff --git a/Modules/resource.c b/Modules/resource.c
index 4614f5e98cc888..f5d9972d9a8ff7 100644
--- a/Modules/resource.c
+++ b/Modules/resource.c
@@ -1,13 +1,12 @@
-
#include "Python.h"
-#include <sys/resource.h>
+#include <errno.h> // errno
+#include <string.h>
+#include <sys/resource.h> // getrusage()
#ifdef HAVE_SYS_TIME_H
-#include <sys/time.h>
+# include <sys/time.h>
#endif
#include <time.h>
-#include <string.h>
-#include <errno.h>
-#include <unistd.h>
+#include <unistd.h> // getpagesize()
/* On some systems, these aren't in any header file.
On others they are, with inconsistent prototypes.
diff --git a/Modules/selectmodule.c b/Modules/selectmodule.c
index 4987cf0f2065c2..c56e682b21e2a1 100644
--- a/Modules/selectmodule.c
+++ b/Modules/selectmodule.c
@@ -17,6 +17,9 @@
#include "pycore_time.h" // _PyTime_t
#include <stddef.h> // offsetof()
+#ifndef MS_WINDOWS
+# include <unistd.h> // close()
+#endif
#ifdef HAVE_SYS_DEVPOLL_H
#include <sys/resource.h>
diff --git a/Modules/socketmodule.c b/Modules/socketmodule.c
index 2f12c9cedbd8a6..74b1c1c661604f 100644
--- a/Modules/socketmodule.c
+++ b/Modules/socketmodule.c
@@ -269,7 +269,7 @@ shutdown(how) -- shut down traffic in one or both directions\n\
#ifdef HAVE_NETDB_H
# include <netdb.h>
#endif
-# include <unistd.h>
+#include <unistd.h> // close()
/* Headers needed for inet_ntoa() and inet_addr() */
# include <arpa/inet.h>
diff --git a/Programs/_freeze_module.c b/Programs/_freeze_module.c
index e55f1d56745c4d..f6c46fa629efba 100644
--- a/Programs/_freeze_module.c
+++ b/Programs/_freeze_module.c
@@ -19,7 +19,7 @@
#include <sys/types.h>
#include <sys/stat.h>
#ifndef MS_WINDOWS
-#include <unistd.h>
+# include <unistd.h>
#endif
uint32_t _Py_next_func_version = 1;
diff --git a/Python/dup2.c b/Python/dup2.c
index a1df0492099163..936211f27ec737 100644
--- a/Python/dup2.c
+++ b/Python/dup2.c
@@ -11,9 +11,9 @@
* Return fd2 if all went well; return BADEXIT otherwise.
*/
-#include <errno.h>
-#include <fcntl.h>
-#include <unistd.h>
+#include <errno.h> // errno
+#include <fcntl.h> // fcntl()
+#include <unistd.h> // close()
#define BADEXIT -1
diff --git a/Python/perf_trampoline.c b/Python/perf_trampoline.c
index b8885a303977d0..10675bf9f8292a 100644
--- a/Python/perf_trampoline.c
+++ b/Python/perf_trampoline.c
@@ -140,9 +140,9 @@ any DWARF information available for them).
#include <fcntl.h>
#include <stdio.h>
#include <stdlib.h>
-#include <sys/mman.h>
+#include <sys/mman.h> // mmap()
#include <sys/types.h>
-#include <unistd.h>
+#include <unistd.h> // sysconf()
#if defined(__arm__) || defined(__arm64__) || defined(__aarch64__)
#define PY_HAVE_INVALIDATE_ICACHE
| <!--
Thanks for your contribution!
Please read this comment in its entirety. It's quite important.
# Pull Request title
It should be in the following format:
```
gh-NNNNN: Summary of the changes made
```
Where: gh-NNNNN refers to the GitHub issue number.
Most PRs will require an issue number. Trivial changes, like fixing a typo, do not need an issue.
# Backport Pull Request title
If this is a backport PR (PR made against branches other than `main`),
please ensure that the PR title is in the following format:
```
[X.Y] <title from the original PR> (GH-NNNN)
```
Where: [X.Y] is the branch name, e.g. [3.6].
GH-NNNN refers to the PR number from `main`.
-->
<!-- gh-issue-number: gh-108765 -->
* Issue: gh-108765
<!-- /gh-issue-number -->
<!-- readthedocs-preview cpython-previews start -->
----
:books: Documentation preview :books:: https://cpython-previews--108783.org.readthedocs.build/
<!-- readthedocs-preview cpython-previews end --> | https://api.github.com/repos/python/cpython/pulls/108783 | 2023-09-01T19:11:49Z | 2023-09-02T14:50:18Z | 2023-09-02T14:50:18Z | 2023-10-27T20:28:40Z | 2,936 | python/cpython | 4,344 |
Simplify PreparedRequest.prepare API | diff --git a/requests/models.py b/requests/models.py
index 752c58c153..45b3ea9680 100644
--- a/requests/models.py
+++ b/requests/models.py
@@ -523,6 +523,10 @@ def prepare_cookies(self, cookies):
def prepare_hooks(self, hooks):
"""Prepares the given hooks."""
+ # hooks can be passed as None to the prepare method and to this
+ # method. To prevent iterating over None, simply use an empty list
+ # if hooks is False-y
+ hooks = hooks or []
for event in hooks:
self.register_hook(event, hooks[event])
diff --git a/test_requests.py b/test_requests.py
index 15406a22fc..cad8c055c8 100755
--- a/test_requests.py
+++ b/test_requests.py
@@ -1613,7 +1613,6 @@ def test_prepare_unicode_url():
p.prepare(
method='GET',
url=u('http://www.example.com/üniçø∂é'),
- hooks=[]
)
assert_copy(p, p.copy())
| Do not require that hooks be passed as an empty list to
PreparedRequest.prepare. In the event hooks is None in prepare or
prepare_hooks, use an empty list as a default.
Related to #2552
| https://api.github.com/repos/psf/requests/pulls/2553 | 2015-04-21T01:14:53Z | 2015-04-21T05:59:55Z | 2015-04-21T05:59:55Z | 2021-09-08T08:00:49Z | 250 | psf/requests | 33,005 |
Remove elements that don't add value in ES.84 | diff --git a/CppCoreGuidelines.md b/CppCoreGuidelines.md
index 7db626d82..1d437b249 100644
--- a/CppCoreGuidelines.md
+++ b/CppCoreGuidelines.md
@@ -9884,7 +9884,7 @@ Statement rules:
* [ES.77: Minimize the use of `break` and `continue` in loops](#Res-continue)
* [ES.78: Always end a non-empty `case` with a `break`](#Res-break)
* [ES.79: Use `default` to handle common cases (only)](#Res-default)
-* [ES.84: Don't (try to) declare a local variable with no name](#Res-noname)
+* [ES.84: Don't try to declare a local variable with no name](#Res-noname)
* [ES.85: Make empty statements visible](#Res-empty)
* [ES.86: Avoid modifying loop control variables inside the body of raw for-loops](#Res-loop-counter)
* [ES.87: Don't add redundant `==` or `!=` to conditions](#Res-if)
@@ -12789,13 +12789,12 @@ Flag `switch`-statements over an enumeration that don't handle all enumerators a
This may yield too many false positives in some code bases; if so, flag only `switch`es that handle most but not all cases
(that was the strategy of the very first C++ compiler).
-### <a name="Res-noname"></a>ES.84: Don't (try to) declare a local variable with no name
+### <a name="Res-noname"></a>ES.84: Don't try to declare a local variable with no name
##### Reason
There is no such thing.
What looks to a human like a variable without a name is to the compiler a statement consisting of a temporary that immediately goes out of scope.
-To avoid unpleasant surprises.
##### Example, bad
@@ -12808,7 +12807,6 @@ To avoid unpleasant surprises.
This declares an unnamed `lock` object that immediately goes out of scope at the point of the semicolon.
This is not an uncommon mistake.
In particular, this particular example can lead to hard-to find race conditions.
-There are exceedingly clever uses of this "idiom", but they are far rarer than the mistakes.
##### Note
@@ -12816,7 +12814,7 @@ Unnamed function arguments are fine.
##### Enforcement
-Flag statements that are just a temporary
+Flag statements that are just a temporary.
### <a name="Res-empty"></a>ES.85: Make empty statements visible
| - Parentheses around "(try to)" in rule title add no meaning.
- The sentence fragment "To avoid unpleasant surprises." in Reason adds no info that hasn't already been stated.
- "There are exceedingly clever uses of this 'idiom', but..." seems like a distraction instead of a tight conclusion to the example section. We're not seeking "exceedingly clever" in this guide. Per Bjarne's own words, we're seeking the "smaller, simpler, safer language struggling to get out."
- Added a period to the end of the Enforcement sentence.
| https://api.github.com/repos/isocpp/CppCoreGuidelines/pulls/1390 | 2019-03-17T13:15:39Z | 2019-03-21T18:04:50Z | 2019-03-21T18:04:50Z | 2019-04-02T00:44:27Z | 579 | isocpp/CppCoreGuidelines | 15,926 |
fix --certs argument | diff --git a/docs/src/content/concepts-certificates.md b/docs/src/content/concepts-certificates.md
index ebc5ede990..0fc32c696c 100644
--- a/docs/src/content/concepts-certificates.md
+++ b/docs/src/content/concepts-certificates.md
@@ -93,7 +93,7 @@ The files created by mitmproxy in the .mitmproxy directory are as follows:
## Using a custom server certificate
-You can use your own (leaf) certificate by passing the `--cert
+You can use your own (leaf) certificate by passing the `--certs
[domain=]path_to_certificate` option to mitmproxy. Mitmproxy then uses the
provided certificate for interception of the specified domain instead of
generating a certificate signed by its own CA.
@@ -127,13 +127,13 @@ Now, you can run mitmproxy with the generated certificate:
**For all domain names**
```bash
-mitmproxy --cert *=cert.pem
+mitmproxy --certs *=cert.pem
```
**For specific domain names**
```bash
-mitmproxy --cert *.example.com=cert.pem
+mitmproxy --certs *.example.com=cert.pem
```
**Note:** `*.example.com` is for all the subdomains. You can also use
| the help output claims that --certs is correct | https://api.github.com/repos/mitmproxy/mitmproxy/pulls/4412 | 2021-01-24T21:05:38Z | 2021-01-24T21:12:17Z | 2021-01-24T21:12:17Z | 2021-01-31T09:05:58Z | 287 | mitmproxy/mitmproxy | 27,421 |
readme.md dataset Table Formatting | diff --git a/model/model_training/README.md b/model/model_training/README.md
index 914c82eff7..1f5dbcf9ed 100644
--- a/model/model_training/README.md
+++ b/model/model_training/README.md
@@ -212,10 +212,9 @@ deepspeed trainer_sft.py --configs defaults your-model-name --deepspeed
Here is an uncomplete overview of datasets for sft:
<!-- prettier-ignore -->
+<!-- prettier-ignore-start -->
dataset_name | train_counts | eval_counts | total_counts
-----------------------------------------------------------------
-
-<!-- prettier-ignore -->
+--|--|--|--
joke | 301 | 76 | 377
webgpt | 14251 | 3563 | 17814
gpt4all | 313552 | 78388 | 391940
@@ -233,6 +232,7 @@ prosocial_dialogue | 157160 | 26983 | 184143
explain_prosocial | 360708 | 61248 | 421956
soda | 924102 | 231026 | 1155128
oa_leet10k | 18728 | 4683 | 23411
+<!-- prettier-ignore-end -->
This list can be generated with the following command, but beware that this
downloads all available datasets (>100GB):
| change markdown Table Formatting.
that format is broken
| https://api.github.com/repos/LAION-AI/Open-Assistant/pulls/3219 | 2023-05-23T23:43:15Z | 2023-05-26T21:18:13Z | 2023-05-26T21:18:13Z | 2023-05-26T21:18:14Z | 329 | LAION-AI/Open-Assistant | 37,250 |
Enhancement Adding not existent lambda name in response headers | diff --git a/localstack/services/awslambda/lambda_api.py b/localstack/services/awslambda/lambda_api.py
index 36e4345824805..c4c74d2d2a3e7 100644
--- a/localstack/services/awslambda/lambda_api.py
+++ b/localstack/services/awslambda/lambda_api.py
@@ -793,19 +793,23 @@ def forward_to_fallback_url(func_arn, data):
Lambda to the configured URL. """
if not config.LAMBDA_FALLBACK_URL:
return None
+
+ lambda_name = aws_stack.lambda_function_name(func_arn)
if config.LAMBDA_FALLBACK_URL.startswith('dynamodb://'):
table_name = urlparse(config.LAMBDA_FALLBACK_URL.replace('dynamodb://', 'http://')).netloc
dynamodb = aws_stack.connect_to_service('dynamodb')
item = {
'id': {'S': short_uid()},
'timestamp': {'N': str(now_utc())},
- 'payload': {'S': str(data)}
+ 'payload': {'S': str(data)},
+ 'function_name': {'S': lambda_name}
}
aws_stack.create_dynamodb_table(table_name, partition_key='id')
dynamodb.put_item(TableName=table_name, Item=item)
return ''
if re.match(r'^https?://.+', config.LAMBDA_FALLBACK_URL):
- response = safe_requests.post(config.LAMBDA_FALLBACK_URL, data)
+ headers = {'lambda-function-name': lambda_name}
+ response = safe_requests.post(config.LAMBDA_FALLBACK_URL, data, headers=headers)
return response.content
raise ClientError('Unexpected value for LAMBDA_FALLBACK_URL: %s' % config.LAMBDA_FALLBACK_URL)
@@ -1144,14 +1148,14 @@ def invoke_function(function):
# Default invocation type is RequestResponse
invocation_type = request.environ.get('HTTP_X_AMZ_INVOCATION_TYPE', 'RequestResponse')
- def _create_response(result, status_code=200):
+ def _create_response(result, status_code=200, headers={}):
""" Create the final response for the given invocation result """
if isinstance(result, Response):
return result
details = {
'StatusCode': status_code,
'Payload': result,
- 'Headers': {}
+ 'Headers': headers
}
if isinstance(result, dict):
for key in ('StatusCode', 'Payload', 'FunctionError'):
@@ -1187,7 +1191,7 @@ def _create_response(result, status_code=200):
not_found = not_found_error('{0}:{1}'.format(arn, qualifier))
if not_found:
- forward_result = forward_to_fallback_url(func_arn, data)
+ forward_result = forward_to_fallback_url(arn, data)
if forward_result is not None:
return _create_response(forward_result)
return not_found
diff --git a/tests/integration/test_lambda.py b/tests/integration/test_lambda.py
index 9089e6ef6fd70..f61952e453769 100644
--- a/tests/integration/test_lambda.py
+++ b/tests/integration/test_lambda.py
@@ -120,7 +120,7 @@ def num_items():
def test_forward_to_fallback_url_http(self):
class MyUpdateListener(ProxyListener):
def forward_request(self, method, path, data, headers):
- records.append(data)
+ records.append({'data': data, 'headers': headers})
return 200
records = []
@@ -130,9 +130,26 @@ def forward_request(self, method, path, data, headers):
items_before = len(records)
_run_forward_to_fallback_url('%s://localhost:%s' % (get_service_protocol(), local_port))
items_after = len(records)
+ for record in records:
+ self.assertIn('non-existing-lambda', record['headers']['lambda-function-name'])
+
self.assertEqual(items_after, items_before + 3)
proxy.stop()
+ def test_adding_fallback_function_name_in_headers(self):
+
+ lambda_client = aws_stack.connect_to_service('lambda')
+ ddb_client = aws_stack.connect_to_service('dynamodb')
+
+ db_table = 'lambda-records'
+ config.LAMBDA_FALLBACK_URL = 'dynamodb://%s' % db_table
+
+ lambda_client.invoke(FunctionName='non-existing-lambda',
+ Payload=b'{}', InvocationType='RequestResponse')
+
+ result = run_safe(ddb_client.scan, TableName=db_table)
+ self.assertEqual(result['Items'][0]['function_name']['S'], 'non-existing-lambda')
+
def test_dead_letter_queue(self):
sqs_client = aws_stack.connect_to_service('sqs')
lambda_client = aws_stack.connect_to_service('lambda')
| Enhancement Adding not existent lambda name in response headers. Fixes #1971 | https://api.github.com/repos/localstack/localstack/pulls/2397 | 2020-05-05T21:09:13Z | 2020-05-07T21:44:17Z | 2020-05-07T21:44:17Z | 2020-05-07T21:44:17Z | 1,054 | localstack/localstack | 28,759 |
Remove extra spider parameter in item pipeline docs | diff --git a/docs/topics/item-pipeline.rst b/docs/topics/item-pipeline.rst
index bc26bbebe55..a5f6e07b89d 100644
--- a/docs/topics/item-pipeline.rst
+++ b/docs/topics/item-pipeline.rst
@@ -215,7 +215,7 @@ item.
screenshot_url = self.SPLASH_URL.format(encoded_item_url)
request = scrapy.Request(screenshot_url, callback=NO_CALLBACK)
response = await maybe_deferred_to_future(
- spider.crawler.engine.download(request, spider)
+ spider.crawler.engine.download(request)
)
if response.status != 200:
| Fixes #6008
I looked in other files and this is the only place where we still have an example of passing an extra spider parameter. | https://api.github.com/repos/scrapy/scrapy/pulls/6009 | 2023-08-10T11:46:17Z | 2023-08-10T11:48:44Z | 2023-08-10T11:48:44Z | 2023-08-10T11:48:44Z | 143 | scrapy/scrapy | 34,279 |
[events] add tests for ErrorEvent event types | diff --git a/tests/sentry/eventtypes/__init__.py b/tests/sentry/eventtypes/__init__.py
new file mode 100644
index 0000000000000..c3961685ab8de
--- /dev/null
+++ b/tests/sentry/eventtypes/__init__.py
@@ -0,0 +1 @@
+from __future__ import absolute_import
diff --git a/tests/sentry/eventtypes/test_error.py b/tests/sentry/eventtypes/test_error.py
new file mode 100644
index 0000000000000..b4d8108df37f6
--- /dev/null
+++ b/tests/sentry/eventtypes/test_error.py
@@ -0,0 +1,21 @@
+from __future__ import absolute_import
+
+from sentry.eventtypes import ErrorEvent
+from sentry.testutils import TestCase
+
+
+class ErrorEventTest(TestCase):
+ def test_to_string_none_value(self):
+ inst = ErrorEvent({})
+ result = inst.to_string({'type': 'Error', 'value': None})
+ assert result == 'Error'
+
+ def test_to_string_eliminates_values_with_newline(self):
+ inst = ErrorEvent({})
+ result = inst.to_string({'type': 'Error', 'value': 'foo\nbar'})
+ assert result == 'Error: foo'
+
+ def test_to_string_handles_empty_value(self):
+ inst = ErrorEvent({})
+ result = inst.to_string({'type': 'Error', 'value': ''})
+ assert result == 'Error'
| https://api.github.com/repos/getsentry/sentry/pulls/4261 | 2016-10-03T23:26:17Z | 2016-10-03T23:36:52Z | 2016-10-03T23:36:52Z | 2020-12-23T10:01:48Z | 335 | getsentry/sentry | 44,182 |
|
R.3_609: changed owner<T> to owner<T*> in R.3 per issue #609 | diff --git a/CppCoreGuidelines.md b/CppCoreGuidelines.md
index 082bbf524..26c3de932 100644
--- a/CppCoreGuidelines.md
+++ b/CppCoreGuidelines.md
@@ -7232,7 +7232,7 @@ We can fix that problem by making ownership explicit:
class X2 {
// ...
public:
- owner<T> p; // OK: p is owning
+ owner<T*> p; // OK: p is owning
T* q; // OK: q is not owning
};
@@ -7256,9 +7256,9 @@ Some interfaces cannot be simply annotated with `owner` because they need to rem
##### Note
-`owner<T>` has no default semantics beyond `T*`. It can be used without changing any code using it and without affecting ABIs.
+`owner<T*>` has no default semantics beyond `T*`. It can be used without changing any code using it and without affecting ABIs.
It is simply a indicator to programmers and analysis tools.
-For example, if an `owner<T>` is a member of a class, that class better have a destructor that `delete`s it.
+For example, if an `owner<T*>` is a member of a class, that class better have a destructor that `delete`s it.
##### Example, bad
| In guideline R.3: changed owner<T> references to owner<T*> described in issue #609.
| https://api.github.com/repos/isocpp/CppCoreGuidelines/pulls/612 | 2016-05-19T01:46:53Z | 2016-05-19T15:18:30Z | 2016-05-19T15:18:30Z | 2016-05-19T17:07:01Z | 301 | isocpp/CppCoreGuidelines | 15,869 |
Fortinet's FortiOS user adgrp | diff --git a/lib/ansible/modules/network/fortios/fortios_user_adgrp.py b/lib/ansible/modules/network/fortios/fortios_user_adgrp.py
new file mode 100644
index 00000000000000..7cc8a1c8378594
--- /dev/null
+++ b/lib/ansible/modules/network/fortios/fortios_user_adgrp.py
@@ -0,0 +1,260 @@
+#!/usr/bin/python
+from __future__ import (absolute_import, division, print_function)
+# Copyright 2019 Fortinet, Inc.
+#
+# This program is free software: you can redistribute it and/or modify
+# it under the terms of the GNU General Public License as published by
+# the Free Software Foundation, either version 3 of the License, or
+# (at your option) any later version.
+#
+# This program is distributed in the hope that it will be useful,
+# but WITHOUT ANY WARRANTY; without even the implied warranty of
+# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
+# GNU General Public License for more details.
+#
+# You should have received a copy of the GNU General Public License
+# along with this program. If not, see <https://www.gnu.org/licenses/>.
+
+__metaclass__ = type
+
+ANSIBLE_METADATA = {'status': ['preview'],
+ 'supported_by': 'community',
+ 'metadata_version': '1.1'}
+
+DOCUMENTATION = '''
+---
+module: fortios_user_adgrp
+short_description: Configure FSSO groups in Fortinet's FortiOS and FortiGate.
+description:
+ - This module is able to configure a FortiGate or FortiOS by allowing the
+ user to set and modify user feature and adgrp category.
+ Examples include all parameters and values need to be adjusted to datasources before usage.
+ Tested with FOS v6.0.2
+version_added: "2.8"
+author:
+ - Miguel Angel Munoz (@mamunozgonzalez)
+ - Nicolas Thomas (@thomnico)
+notes:
+ - Requires fortiosapi library developed by Fortinet
+ - Run as a local_action in your playbook
+requirements:
+ - fortiosapi>=0.9.8
+options:
+ host:
+ description:
+ - FortiOS or FortiGate ip address.
+ required: true
+ username:
+ description:
+ - FortiOS or FortiGate username.
+ required: true
+ password:
+ description:
+ - FortiOS or FortiGate password.
+ default: ""
+ vdom:
+ description:
+ - Virtual domain, among those defined previously. A vdom is a
+ virtual instance of the FortiGate that can be configured and
+ used as a different unit.
+ default: root
+ https:
+ description:
+ - Indicates if the requests towards FortiGate must use HTTPS
+ protocol
+ type: bool
+ default: true
+ user_adgrp:
+ description:
+ - Configure FSSO groups.
+ default: null
+ suboptions:
+ state:
+ description:
+ - Indicates whether to create or remove the object
+ choices:
+ - present
+ - absent
+ name:
+ description:
+ - Name.
+ required: true
+ server-name:
+ description:
+ - FSSO agent name. Source user.fsso.name.
+'''
+
+EXAMPLES = '''
+- hosts: localhost
+ vars:
+ host: "192.168.122.40"
+ username: "admin"
+ password: ""
+ vdom: "root"
+ tasks:
+ - name: Configure FSSO groups.
+ fortios_user_adgrp:
+ host: "{{ host }}"
+ username: "{{ username }}"
+ password: "{{ password }}"
+ vdom: "{{ vdom }}"
+ https: "False"
+ user_adgrp:
+ state: "present"
+ name: "default_name_3"
+ server-name: "<your_own_value> (source user.fsso.name)"
+'''
+
+RETURN = '''
+build:
+ description: Build number of the fortigate image
+ returned: always
+ type: str
+ sample: '1547'
+http_method:
+ description: Last method used to provision the content into FortiGate
+ returned: always
+ type: str
+ sample: 'PUT'
+http_status:
+ description: Last result given by FortiGate on last operation applied
+ returned: always
+ type: str
+ sample: "200"
+mkey:
+ description: Master key (id) used in the last call to FortiGate
+ returned: success
+ type: str
+ sample: "id"
+name:
+ description: Name of the table used to fulfill the request
+ returned: always
+ type: str
+ sample: "urlfilter"
+path:
+ description: Path of the table used to fulfill the request
+ returned: always
+ type: str
+ sample: "webfilter"
+revision:
+ description: Internal revision number
+ returned: always
+ type: str
+ sample: "17.0.2.10658"
+serial:
+ description: Serial number of the unit
+ returned: always
+ type: str
+ sample: "FGVMEVYYQT3AB5352"
+status:
+ description: Indication of the operation's result
+ returned: always
+ type: str
+ sample: "success"
+vdom:
+ description: Virtual domain used
+ returned: always
+ type: str
+ sample: "root"
+version:
+ description: Version of the FortiGate
+ returned: always
+ type: str
+ sample: "v5.6.3"
+
+'''
+
+from ansible.module_utils.basic import AnsibleModule
+
+
+def login(data, fos):
+ host = data['host']
+ username = data['username']
+ password = data['password']
+
+ fos.debug('on')
+ if 'https' in data and not data['https']:
+ fos.https('off')
+ else:
+ fos.https('on')
+
+ fos.login(host, username, password)
+
+
+def filter_user_adgrp_data(json):
+ option_list = ['name', 'server-name']
+ dictionary = {}
+
+ for attribute in option_list:
+ if attribute in json and json[attribute] is not None:
+ dictionary[attribute] = json[attribute]
+
+ return dictionary
+
+
+def user_adgrp(data, fos):
+ vdom = data['vdom']
+ user_adgrp_data = data['user_adgrp']
+ filtered_data = filter_user_adgrp_data(user_adgrp_data)
+
+ if user_adgrp_data['state'] == "present":
+ return fos.set('user',
+ 'adgrp',
+ data=filtered_data,
+ vdom=vdom)
+
+ elif user_adgrp_data['state'] == "absent":
+ return fos.delete('user',
+ 'adgrp',
+ mkey=filtered_data['name'],
+ vdom=vdom)
+
+
+def fortios_user(data, fos):
+ login(data, fos)
+
+ if data['user_adgrp']:
+ resp = user_adgrp(data, fos)
+
+ fos.logout()
+ return not resp['status'] == "success", resp['status'] == "success", resp
+
+
+def main():
+ fields = {
+ "host": {"required": True, "type": "str"},
+ "username": {"required": True, "type": "str"},
+ "password": {"required": False, "type": "str", "no_log": True},
+ "vdom": {"required": False, "type": "str", "default": "root"},
+ "https": {"required": False, "type": "bool", "default": True},
+ "user_adgrp": {
+ "required": False, "type": "dict",
+ "options": {
+ "state": {"required": True, "type": "str",
+ "choices": ["present", "absent"]},
+ "name": {"required": True, "type": "str"},
+ "server-name": {"required": False, "type": "str"}
+
+ }
+ }
+ }
+
+ module = AnsibleModule(argument_spec=fields,
+ supports_check_mode=False)
+ try:
+ from fortiosapi import FortiOSAPI
+ except ImportError:
+ module.fail_json(msg="fortiosapi module is required")
+
+ fos = FortiOSAPI()
+
+ is_error, has_changed, result = fortios_user(module.params, fos)
+
+ if not is_error:
+ module.exit_json(changed=has_changed, meta=result)
+ else:
+ module.fail_json(msg="Error in repo", meta=result)
+
+
+if __name__ == '__main__':
+ main()
| ##### SUMMARY
Fortinet is adding Ansible support for FortiOS and FortiGate products. This module follows the same structure, guidelines and ideas given in previous approved module for a parallel feature of FortiGate (webfiltering): https://github.com/ansible/ansible/pull/37196
In this case we are providing a different functionality: "User Adgrp".
Please note that this will be part of other modules to come for FortiGate, including different functionalities: system, wireless-controller, firewall, webfilter, ips, web-proxy, wanopt, application, dlp spamfilter, log, vpn, certificate, user, dnsfilter, antivirus, report, waf, authentication, switch controller, endpoint-control and router. We plan to follow the same style, structure and usage as in the previous module in order to make it easier to comply with Ansible guidelines.
##### ISSUE TYPE
- New Module Pull Request
##### COMPONENT NAME
fortios_user_addgrp
##### ANSIBLE VERSION
```
ansible 2.8.0.dev0 (new_module ddbbe5dfa5) last updated 2018/09/24 14:54:57 (GMT +200)
config file = /etc/ansible/ansible.cfg
configured module search path = [u'/home/magonzalez/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules']
ansible python module location = /home/magonzalez/ansible/lib/ansible
executable location = /home/magonzalez/ansible/bin/ansible
python version = 2.7.15rc1 (default, Apr 15 2018, 21:51:34) [GCC 7.3.0]
``` | https://api.github.com/repos/ansible/ansible/pulls/52831 | 2019-02-22T16:48:48Z | 2019-03-05T11:18:48Z | 2019-03-05T11:18:48Z | 2019-07-25T16:57:44Z | 2,065 | ansible/ansible | 48,937 |
Touch up venv docs | diff --git a/Doc/library/venv.rst b/Doc/library/venv.rst
index d3d5ae2b007d5f..62732d22438672 100644
--- a/Doc/library/venv.rst
+++ b/Doc/library/venv.rst
@@ -47,7 +47,7 @@ Creating virtual environments
A virtual environment is a directory tree which contains Python executable
files and other files which indicate that it is a virtual environment.
- Common installation tools such as ``Setuptools`` and ``pip`` work as
+ Common installation tools such as setuptools_ and pip_ work as
expected with virtual environments. In other words, when a virtual
environment is active, they install Python packages into the virtual
environment without needing to be told to do so explicitly.
@@ -64,24 +64,25 @@ Creating virtual environments
Python installation).
When a virtual environment is active, any options that change the
- installation path will be ignored from all distutils configuration files to
- prevent projects being inadvertently installed outside of the virtual
- environment.
+ installation path will be ignored from all :mod:`distutils` configuration
+ files to prevent projects being inadvertently installed outside of the
+ virtual environment.
When working in a command shell, users can make a virtual environment active
by running an ``activate`` script in the virtual environment's executables
- directory (the precise filename is shell-dependent), which prepends the
- virtual environment's directory for executables to the ``PATH`` environment
- variable for the running shell. There should be no need in other
- circumstances to activate a virtual environment—scripts installed into
- virtual environments have a "shebang" line which points to the virtual
- environment's Python interpreter. This means that the script will run with
- that interpreter regardless of the value of ``PATH``. On Windows, "shebang"
- line processing is supported if you have the Python Launcher for Windows
- installed (this was added to Python in 3.3 - see :pep:`397` for more
- details). Thus, double-clicking an installed script in a Windows Explorer
- window should run the script with the correct interpreter without there
- needing to be any reference to its virtual environment in ``PATH``.
+ directory (the precise filename and command to use the file is
+ shell-dependent), which prepends the virtual environment's directory for
+ executables to the ``PATH`` environment variable for the running shell. There
+ should be no need in other circumstances to activate a virtual
+ environment; scripts installed into virtual environments have a "shebang"
+ line which points to the virtual environment's Python interpreter. This means
+ that the script will run with that interpreter regardless of the value of
+ ``PATH``. On Windows, "shebang" line processing is supported if you have the
+ Python Launcher for Windows installed (this was added to Python in 3.3 - see
+ :pep:`397` for more details). Thus, double-clicking an installed script in a
+ Windows Explorer window should run the script with the correct interpreter
+ without there needing to be any reference to its virtual environment in
+ ``PATH``.
.. _venv-api:
@@ -135,20 +136,20 @@ creation according to their needs, the :class:`EnvBuilder` class.
Added the ``upgrade_deps`` parameter
Creators of third-party virtual environment tools will be free to use the
- provided ``EnvBuilder`` class as a base class.
+ provided :class:`EnvBuilder` class as a base class.
The returned env-builder is an object which has a method, ``create``:
.. method:: create(env_dir)
- This method takes as required argument the path (absolute or relative to
- the current directory) of the target directory which is to contain the
+ Create a virtual environment by specifying the target directory
+ (absolute or relative to the current directory) which is to contain the
virtual environment. The ``create`` method will either create the
environment in the specified directory, or raise an appropriate
exception.
- The ``create`` method of the ``EnvBuilder`` class illustrates the hooks
- available for subclass customization::
+ The ``create`` method of the :class:`EnvBuilder` class illustrates the
+ hooks available for subclass customization::
def create(self, env_dir):
"""
@@ -476,3 +477,7 @@ subclass which installs setuptools and pip into a created virtual environment::
This script is also available for download `online
<https://gist.github.com/vsajip/4673395>`_.
+
+
+.. _setuptools: https://pypi.org/project/setuptools/
+.. _pip: https://pypi.org/project/pip/
| https://api.github.com/repos/python/cpython/pulls/14922 | 2019-07-23T20:51:49Z | 2019-07-23T21:34:33Z | 2019-07-23T21:34:33Z | 2019-07-23T21:34:53Z | 1,074 | python/cpython | 4,321 |
|
add: test for proxy.py | diff --git a/test_proxy.py b/test_proxy.py
new file mode 100644
index 00000000..fdf9188a
--- /dev/null
+++ b/test_proxy.py
@@ -0,0 +1,105 @@
+#!/usr/bin/env python
+# -*- coding: utf-8 -*-
+
+from proxy import Proxy, NoTalkProxy
+from io import StringIO
+import sys
+from time import time
+
+if sys.version_info < (2, 7):
+ import unittest2 as unittest
+else:
+ import unittest
+
+
+class ProxyTest(unittest.TestCase):
+
+ @classmethod
+ def setUpClass(cls):
+ """ Class scope setup. """
+ cls.p = Proxy()
+
+ def setUp(cls):
+ """ Function/test case scope setup. """
+ cls.output = StringIO()
+ cls.saved_stdout = sys.stdout
+ sys.stdout = cls.output
+
+ def tearDown(cls):
+ """ Function/test case scope teardown. """
+ cls.output.close()
+ sys.stdout = cls.saved_stdout
+
+ def test_sales_manager_shall_talk_through_proxy_with_delay(cls):
+ cls.p.busy = 'No'
+ start_time = time()
+ cls.p.talk()
+ end_time = time()
+ execution_time = end_time - start_time
+ print_output = cls.output.getvalue()
+ expected_print_output = 'Proxy checking for Sales Manager availability\n\
+Sales Manager ready to talk\n'
+ cls.assertEqual(print_output, expected_print_output)
+ expected_execution_time = 2
+ cls.assertEqual(int(execution_time), expected_execution_time)
+
+ def test_sales_manager_shall_respond_through_proxy_with_delay(cls):
+ cls.p.busy = 'Yes'
+ start_time = time()
+ cls.p.talk()
+ end_time = time()
+ execution_time = end_time - start_time
+ print_output = cls.output.getvalue()
+ expected_print_output = 'Proxy checking for Sales Manager availability\n\
+Sales Manager is busy\n'
+ cls.assertEqual(print_output, expected_print_output)
+ expected_execution_time = 2
+ cls.assertEqual(int(execution_time), expected_execution_time)
+
+
+class NoTalkProxyTest(unittest.TestCase):
+
+ @classmethod
+ def setUpClass(cls):
+ """ Class scope setup. """
+ cls.ntp = NoTalkProxy()
+
+ def setUp(cls):
+ """ Function/test case scope setup. """
+ cls.output = StringIO()
+ cls.saved_stdout = sys.stdout
+ sys.stdout = cls.output
+
+ def tearDown(cls):
+ """ Function/test case scope teardown. """
+ cls.output.close()
+ sys.stdout = cls.saved_stdout
+
+ def test_sales_manager_shall_not_talk_through_proxy_with_delay(cls):
+ cls.ntp.busy = 'No'
+ start_time = time()
+ cls.ntp.talk()
+ end_time = time()
+ execution_time = end_time - start_time
+ print_output = cls.output.getvalue()
+ expected_print_output = 'Proxy checking for Sales Manager availability\n\
+This Sales Manager will not talk to you whether he/she is busy or not\n'
+ cls.assertEqual(print_output, expected_print_output)
+ expected_execution_time = 2
+ cls.assertEqual(int(execution_time), expected_execution_time)
+
+ def test_sales_manager_shall_not_respond_through_proxy_with_delay(cls):
+ cls.ntp.busy = 'Yes'
+ start_time = time()
+ cls.ntp.talk()
+ end_time = time()
+ execution_time = end_time - start_time
+ print_output = cls.output.getvalue()
+ expected_print_output = 'Proxy checking for Sales Manager availability\n\
+This Sales Manager will not talk to you whether he/she is busy or not\n'
+ cls.assertEqual(print_output, expected_print_output)
+ expected_execution_time = 2
+ cls.assertEqual(int(execution_time), expected_execution_time)
+
+if __name__ == "__main__":
+ unittest.main()
| #138 brought some tests for the actual proxy.py version on my local machine back to my mind. Runs with Python3.4.1 on Linux.
| https://api.github.com/repos/faif/python-patterns/pulls/139 | 2016-04-25T18:33:21Z | 2016-05-21T18:37:31Z | 2016-05-21T18:37:31Z | 2016-05-21T18:37:31Z | 884 | faif/python-patterns | 33,696 |
fix json tool | diff --git a/libs/langchain/langchain/tools/json/tool.py b/libs/langchain/langchain/tools/json/tool.py
index 6f6473d51e6b47..6c75de20ce5cb2 100644
--- a/libs/langchain/langchain/tools/json/tool.py
+++ b/libs/langchain/langchain/tools/json/tool.py
@@ -20,7 +20,7 @@ def _parse_input(text: str) -> List[Union[str, int]]:
"""Parse input of the form data["key1"][0]["key2"] into a list of keys."""
_res = re.findall(r"\[.*?]", text)
# strip the brackets and quotes, convert to int if possible
- res = [i[1:-1].replace('"', "") for i in _res]
+ res = [i[1:-1].replace('"', "").replace("'", "") for i in _res]
res = [int(i) if i.isdigit() else i for i in res]
return res
diff --git a/libs/langchain/tests/unit_tests/tools/test_json.py b/libs/langchain/tests/unit_tests/tools/test_json.py
index 36a96595e03d36..b677b1577d3933 100644
--- a/libs/langchain/tests/unit_tests/tools/test_json.py
+++ b/libs/langchain/tests/unit_tests/tools/test_json.py
@@ -30,6 +30,10 @@ def test_json_spec_value() -> None:
assert spec.value('data["baz"]') == "{'test': {'foo': [1, 2, 3]}}"
assert spec.value('data["baz"]["test"]') == "{'foo': [1, 2, 3]}"
assert spec.value('data["baz"]["test"]["foo"]') == "[1, 2, 3]"
+ assert spec.value("data['foo']") == "bar"
+ assert spec.value("data['baz']") == "{'test': {'foo': [1, 2, 3]}}"
+ assert spec.value("data['baz']['test']") == "{'foo': [1, 2, 3]}"
+ assert spec.value("data['baz']['test']['foo']") == "[1, 2, 3]"
def test_json_spec_value_max_length() -> None:
| https://api.github.com/repos/langchain-ai/langchain/pulls/9096 | 2023-08-11T05:50:06Z | 2023-08-11T06:39:26Z | 2023-08-11T06:39:26Z | 2023-08-11T06:39:26Z | 499 | langchain-ai/langchain | 42,962 |
|
[Bilibili] fix bilibili 4k | diff --git a/src/you_get/extractors/bilibili.py b/src/you_get/extractors/bilibili.py
index 94e5479f65..7ea626f89d 100644
--- a/src/you_get/extractors/bilibili.py
+++ b/src/you_get/extractors/bilibili.py
@@ -62,7 +62,7 @@ def bilibili_headers(referer=None, cookie=None):
@staticmethod
def bilibili_api(avid, cid, qn=0):
- return 'https://api.bilibili.com/x/player/playurl?avid=%s&cid=%s&qn=%s&type=&otype=json&fnver=0&fnval=16' % (avid, cid, qn)
+ return 'https://api.bilibili.com/x/player/playurl?avid=%s&cid=%s&qn=%s&type=&otype=json&fnver=0&fnval=16&fourk=1' % (avid, cid, qn)
@staticmethod
def bilibili_audio_api(sid):
| B站的playurl API加了一个参数 fourk=1 时才会输出4K的播放地址,否则不显示。
bilibili has changed their playurl api, we should add fork=1 to the url params or it will not output the 4k url. | https://api.github.com/repos/soimort/you-get/pulls/2827 | 2020-09-07T06:34:16Z | 2020-10-06T13:22:24Z | 2020-10-06T13:22:24Z | 2020-10-06T13:22:29Z | 238 | soimort/you-get | 21,402 |
Fix `test_load_img_url_timeout` | diff --git a/tests/utils/test_image_utils.py b/tests/utils/test_image_utils.py
index 1813c2a21f254..5d899c2f1ddf7 100644
--- a/tests/utils/test_image_utils.py
+++ b/tests/utils/test_image_utils.py
@@ -21,7 +21,7 @@
import numpy as np
import pytest
from huggingface_hub.file_download import http_get
-from requests import ReadTimeout
+from requests import ConnectTimeout, ReadTimeout
from tests.pipelines.test_pipelines_document_question_answering import INVOICE_URL
from transformers import is_torch_available, is_vision_available
@@ -491,7 +491,7 @@ def test_load_img_url(self):
@is_flaky()
def test_load_img_url_timeout(self):
- with self.assertRaises(ReadTimeout):
+ with self.assertRaises((ReadTimeout, ConnectTimeout)):
load_image(INVOICE_URL, timeout=0.001)
def test_load_img_local(self):
| # What does this PR do?
#25184 added timeout parameter to some function and also a test. But the expected exception in the test is `ConnectTimeout` (on daily CI) instead of `ReadTimeout`, but it is `ReadTimeout` on `CircleCI`.
I haven't looked why there are such difference. But this PR updates the expected value to `(ReadTimeout, ConnectTimeout)` so the test added in #25184 won't fail.
(let me know if you think we should dive into this) | https://api.github.com/repos/huggingface/transformers/pulls/25976 | 2023-09-05T07:08:01Z | 2023-09-05T09:34:28Z | 2023-09-05T09:34:28Z | 2023-09-05T09:35:12Z | 212 | huggingface/transformers | 12,793 |
Fix Amazon Linux bootstrapping error. | diff --git a/letsencrypt-auto b/letsencrypt-auto
index b3e380f9d6c..a3009fe52a2 100755
--- a/letsencrypt-auto
+++ b/letsencrypt-auto
@@ -85,6 +85,8 @@ ExperimentalBootstrap() {
DeterminePythonVersion() {
if command -v python2.7 > /dev/null ; then
export LE_PYTHON=${LE_PYTHON:-python2.7}
+ elif command -v python27 > /dev/null ; then
+ export LE_PYTHON=${LE_PYTHON:-python27}
elif command -v python2 > /dev/null ; then
export LE_PYTHON=${LE_PYTHON:-python2}
elif command -v python > /dev/null ; then
@@ -135,7 +137,7 @@ then
elif uname | grep -iq Darwin ; then
ExperimentalBootstrap "Mac OS X" mac.sh
elif grep -iq "Amazon Linux" /etc/issue ; then
- ExperimentalBootstrap "Amazon Linux" amazon_linux.sh
+ ExperimentalBootstrap "Amazon Linux" _rpm_common.sh
else
echo "Sorry, I don't know how to bootstrap Let's Encrypt on your operating system!"
echo
| https://api.github.com/repos/certbot/certbot/pulls/1516 | 2015-11-16T08:26:44Z | 2015-11-16T20:19:55Z | 2015-11-16T20:19:55Z | 2016-05-06T19:22:06Z | 271 | certbot/certbot | 565 |
|
[workflow] changed to doc build to be on schedule and release | diff --git a/.github/workflows/doc_build_after_merge.yml b/.github/workflows/doc_build_on_schedule_after_release.yml
similarity index 69%
rename from .github/workflows/doc_build_after_merge.yml
rename to .github/workflows/doc_build_on_schedule_after_release.yml
index b6fd57b8d2b4..62dfdc67257c 100644
--- a/.github/workflows/doc_build_after_merge.yml
+++ b/.github/workflows/doc_build_on_schedule_after_release.yml
@@ -1,18 +1,16 @@
-name: Build Documentation After Merge
+name: Build Documentation On Schedule & After Release
on:
workflow_dispatch:
- push:
- paths:
- - "version.txt"
- - "docs/**"
- branches:
- - "main"
+ schedule:
+ - cron: "0 12 * * *" # build doc every day at 8pm Singapore time (12pm UTC time)
+ release:
+ types: [published]
jobs:
build-doc:
name: Trigger Documentation Build Workflow
- if: ( github.event_name == 'workflow_dispatch' || github.event.pull_request.merged == true ) && github.repository == 'hpcaitech/ColossalAI'
+ if: github.repository == 'hpcaitech/ColossalAI'
runs-on: ubuntu-latest
steps:
- name: trigger workflow in ColossalAI-Documentation
| ## 📌 Checklist before creating the PR
- [x] I have created an issue for this PR for traceability
- [x] The title follows the standard format: `[doc/gemini/tensor/...]: A concise description`
- [x] I have added relevant tags if possible for us to better distinguish different PRs
## 🚨 Issue number
> Link this PR to your issue with words like fixed to automatically close the linked issue upon merge
>
> e.g. `fixed #1234`, `closed #1234`, `resolved #1234`
Fixed #3814
## 📝 What does this PR do?
> Summarize your work here.
> if you have any plots/diagrams/screenshots/tables, please attach them here.
This PR is a continuation of #3815 , this PR changed the workflow to be triggered on schedule (8pm Beijing time every day) and upon release. This is because that secrets are not passed to pull request workflows from a forked repository. There is currently no way to get around with it.
## 💥 Checklist before requesting a review
- [x] I have linked my PR to an issue ([instruction](https://docs.github.com/en/issues/tracking-your-work-with-issues/linking-a-pull-request-to-an-issue))
- [x] My issue clearly describes the problem/feature/proposal, with diagrams/charts/table/code if possible
- [x] I have performed a self-review of my code
- [ ] I have added thorough tests.
- [ ] I have added docstrings for all the functions/methods I implemented
## ⭐️ Do you enjoy contributing to Colossal-AI?
- [x] 🌝 Yes, I do.
- [ ] 🌚 No, I don't.
Tell us more if you don't enjoy contributing to Colossal-AI.
| https://api.github.com/repos/hpcaitech/ColossalAI/pulls/3825 | 2023-05-24T02:11:52Z | 2023-05-24T02:50:19Z | 2023-05-24T02:50:19Z | 2023-05-24T02:50:19Z | 308 | hpcaitech/ColossalAI | 11,136 |
End of preview. Expand
in Dataset Viewer.
No dataset card yet
New: Create and edit this dataset card directly on the website!
Contribute a Dataset Card- Downloads last month
- 32