Search is not available for this dataset
commit_message
stringlengths 9
4.28k
| sha
stringlengths 40
40
| type
stringclasses 10
values | commit_url
stringlengths 78
90
| masked_commit_message
stringlengths 2
4.27k
| author_email
stringclasses 8
values | git_diff
stringlengths 2
37.4M
|
---|---|---|---|---|---|---|
ci: disable doc publishing until geo blog can be fixed | 989ad4f1e7a11b06af7e0e3ef840022687fbd7af | ci | https://github.com/ibis-project/ibis/commit/989ad4f1e7a11b06af7e0e3ef840022687fbd7af | disable doc publishing until geo blog can be fixed | {"ibis-docs-main.yml": "@@ -51,7 +51,8 @@ jobs:\n - name: verify internal links\n run: nix develop --ignore-environment '.#links' -c just checklinks --offline --no-progress\n \n- - name: build and push quarto docs\n- run: nix develop --ignore-environment --keep NETLIFY_AUTH_TOKEN -c just docs-deploy\n- env:\n- NETLIFY_AUTH_TOKEN: ${{ secrets.NETLIFY_AUTH_TOKEN }}\n+ # TODO: re-enable when geo blog is fixed (to_array)\n+ # - name: build and push quarto docs\n+ # run: nix develop --ignore-environment --keep NETLIFY_AUTH_TOKEN -c just docs-deploy\n+ # env:\n+ # NETLIFY_AUTH_TOKEN: ${{ secrets.NETLIFY_AUTH_TOKEN }}\n"} |
|
test: remove unused `spread_type` function | 527b7501e7b458d092baf18b6605f8a7f4595036 | test | https://github.com/ibis-project/ibis/commit/527b7501e7b458d092baf18b6605f8a7f4595036 | remove unused `spread_type` function | {"datatypes.py": "@@ -65,21 +65,3 @@ class BigQuerySchema(SchemaMapper):\n @classmethod\n def to_ibis(cls, fields: list[bq.SchemaField]) -> sch.Schema:\n return sch.Schema({f.name: cls._dtype_from_bigquery_field(f) for f in fields})\n-\n-\n-# TODO(kszucs): we can eliminate this function by making dt.DataType traversible\n-# using ibis.common.graph.Node, similarly to how we traverse ops.Node instances:\n-# node.find(types)\n-def spread_type(dt: dt.DataType):\n- \"\"\"Returns a generator that contains all the types in the given type.\n-\n- For complex types like set and array, it returns the types of the elements.\n- \"\"\"\n- if dt.is_array():\n- yield from spread_type(dt.value_type)\n- elif dt.is_struct():\n- for type_ in dt.types:\n- yield from spread_type(type_)\n- elif dt.is_map():\n- raise NotImplementedError(\"Maps are not supported in BigQuery\")\n- yield dt\n", "test_datatypes.py": "@@ -5,10 +5,7 @@ import sqlglot as sg\n from pytest import param\n \n import ibis.expr.datatypes as dt\n-from ibis.backends.bigquery.datatypes import (\n- BigQueryType,\n- spread_type,\n-)\n+from ibis.backends.bigquery.datatypes import BigQueryType\n \n \n @pytest.mark.parametrize(\n@@ -79,31 +76,6 @@ def test_simple_failure_mode(datatype):\n BigQueryType.to_string(datatype)\n \n \[email protected](\n- (\"type_\", \"expected\"),\n- [\n- param(\n- dt.int64,\n- [dt.int64],\n- ),\n- param(\n- dt.Array(dt.int64),\n- [dt.int64, dt.Array(value_type=dt.int64)],\n- ),\n- param(\n- dt.Struct.from_tuples([(\"a\", dt.Array(dt.int64))]),\n- [\n- dt.int64,\n- dt.Array(value_type=dt.int64),\n- dt.Struct.from_tuples([(\"a\", dt.Array(value_type=dt.int64))]),\n- ],\n- ),\n- ],\n-)\n-def test_spread_type(type_, expected):\n- assert list(spread_type(type_)) == expected\n-\n-\n def test_struct_type():\n dtype = dt.Array(dt.int64)\n parsed_type = sg.parse_one(\"BIGINT[]\", into=sg.exp.DataType, read=\"duckdb\")\n"} |
|
chore(engine): removed fallbacks for rAF, it's useless | d6151fe959532afc5228aa63fd963406d9de777d | chore | https://github.com/tsparticles/tsparticles/commit/d6151fe959532afc5228aa63fd963406d9de777d | removed fallbacks for rAF, it's useless | {"Container.ts": "@@ -1,4 +1,3 @@\n-import { animate, cancelAnimation, isFunction } from \"../Utils/Utils\";\n import { Canvas } from \"./Canvas\";\n import type { ClickMode } from \"../Enums/Modes/ClickMode\";\n import type { Engine } from \"../engine\";\n@@ -17,6 +16,7 @@ import { Particles } from \"./Particles\";\n import { Retina } from \"./Retina\";\n import type { Vector } from \"./Utils/Vector\";\n import { getRangeValue } from \"../Utils/NumberUtils\";\n+import { isFunction } from \"../Utils/Utils\";\n import { loadOptions } from \"../Utils/OptionsUtils\";\n \n /**\n@@ -400,7 +400,7 @@ export class Container {\n \n let refreshTime = force;\n \n- this._drawAnimationFrame = animate()(async (timestamp) => {\n+ this._drawAnimationFrame = requestAnimationFrame(async (timestamp) => {\n if (refreshTime) {\n this.lastFrameTime = undefined;\n \n@@ -562,7 +562,7 @@ export class Container {\n }\n \n if (this._drawAnimationFrame !== undefined) {\n- cancelAnimation()(this._drawAnimationFrame);\n+ cancelAnimationFrame(this._drawAnimationFrame);\n \n delete this._drawAnimationFrame;\n }\n", "Utils.ts": "@@ -169,26 +169,6 @@ export function safeMatchMedia(query: string): MediaQueryList | undefined {\n return matchMedia(query);\n }\n \n-/**\n- * Calls the requestAnimationFrame function or a polyfill\n- * @returns the animation callback id, so it can be canceled\n- */\n-export function animate(): (callback: FrameRequestCallback) => number {\n- return isSsr()\n- ? (callback: FrameRequestCallback): number => setTimeout(callback)\n- : (callback: FrameRequestCallback): number => (requestAnimationFrame || setTimeout)(callback);\n-}\n-\n-/**\n- * Cancels the requestAnimationFrame function or a polyfill\n- * @returns the animation cancelling function\n- */\n-export function cancelAnimation(): (handle: number) => void {\n- return isSsr()\n- ? (handle: number): void => clearTimeout(handle)\n- : (handle: number): void => (cancelAnimationFrame || clearTimeout)(handle);\n-}\n-\n /**\n * Checks if a value is equal to the destination, if same type, or is in the provided array\n * @param value - the value to check\n"} |
|
docs: fix broken links from Semrush report (#7025) | 39eebf622666fdf220f889850fe7e4981cd90d08 | docs | https://github.com/wzhiqing/cube/commit/39eebf622666fdf220f889850fe7e4981cd90d08 | fix broken links from Semrush report (#7025) | {"AlertBox.tsx": "@@ -1,13 +1,13 @@\n-import React from 'react';\n-import classes from './AlertBox.module.css';\n-import classnames from 'classnames/bind';\n+import React from \"react\";\n+import classes from \"./AlertBox.module.css\";\n+import classnames from \"classnames/bind\";\n const cn = classnames.bind(classes);\n \n export enum AlertBoxTypes {\n- DANGER = 'danger',\n- INFO = 'info',\n- SUCCESS = 'success',\n- WARNING = 'warning',\n+ DANGER = \"danger\",\n+ INFO = \"info\",\n+ SUCCESS = \"success\",\n+ WARNING = \"warning\",\n }\n \n declare const TypeToEmoji: {\n@@ -19,55 +19,64 @@ declare const TypeToEmoji: {\n type CalloutType = keyof typeof TypeToEmoji;\n \n export type AlertBoxProps = {\n- children: string;\n+ children: React.ReactNode;\n heading?: string;\n type: AlertBoxTypes;\n-}\n+};\n \n const typeMapping: Record<AlertBoxTypes, CalloutType> = {\n- 'danger': 'error',\n- info: 'info',\n- warning: 'warning',\n- success: 'default',\n-}\n+ danger: \"error\",\n+ info: \"info\",\n+ warning: \"warning\",\n+ success: \"default\",\n+};\n \n const iconMapping: Record<string, any> = {\n- 'danger': '\ud83d\udeab',\n- info: '\u2139\ufe0f',\n- warning: '\u26a0\ufe0f',\n- success: '\u2705',\n+ danger: \"\ud83d\udeab\",\n+ info: \"\u2139\ufe0f\",\n+ warning: \"\u26a0\ufe0f\",\n+ success: \"\u2705\",\n };\n \n export const AlertBox = ({ children, heading, type }: AlertBoxProps) => {\n- const header = heading\n- ? (\n- <div className={classes.AlertBox__header}>\n- <span className={cn('AlertBox__HeaderIcon')}>{iconMapping[type]}</span>\n- {heading}\n- </div>\n- )\n- : null;\n+ const header = heading ? (\n+ <div className={classes.AlertBox__header}>\n+ <span className={cn(\"AlertBox__HeaderIcon\")}>{iconMapping[type]}</span>\n+ {heading}\n+ </div>\n+ ) : null;\n \n return (\n- <div className={cn('AlertBox__Wrapper', `AlertBox__Wrapper--${typeMapping[type]}`)}>\n+ <div\n+ className={cn(\n+ \"AlertBox__Wrapper\",\n+ `AlertBox__Wrapper--${typeMapping[type]}`\n+ )}\n+ >\n {header}\n- <div className={classes.AlertBox__content}>\n- {children}\n- </div>\n+ <div className={classes.AlertBox__content}>{children}</div>\n </div>\n- )\n-}\n+ );\n+};\n \n-export type AlertBoxSubclass = Omit<AlertBoxProps, 'type'>;\n+export type AlertBoxSubclass = Omit<AlertBoxProps, \"type\">;\n \n export type DangerBoxProps = AlertBoxSubclass;\n-export const DangerBox = (props: DangerBoxProps) => <AlertBox type={AlertBoxTypes.DANGER} {...props} />;\n+export const DangerBox = (props: DangerBoxProps) => (\n+ <AlertBox type={AlertBoxTypes.DANGER} {...props} />\n+);\n \n export type InfoBoxProps = AlertBoxSubclass;\n-export const InfoBox = (props: InfoBoxProps) => <AlertBox type={AlertBoxTypes.INFO} {...props} />;\n+export const InfoBox = (props: InfoBoxProps) => (\n+ <AlertBox type={AlertBoxTypes.INFO} {...props} />\n+);\n \n export type SuccessBoxProps = AlertBoxSubclass;\n-export const SuccessBox = (props: SuccessBoxProps) => <AlertBox type={AlertBoxTypes.SUCCESS} {...props} />;\n+export const SuccessBox = (props: SuccessBoxProps) => (\n+ <AlertBox type={AlertBoxTypes.SUCCESS} {...props} />\n+);\n \n export type WarningBoxProps = AlertBoxSubclass;\n-export const WarningBox = (props: WarningBoxProps) => <AlertBox type={AlertBoxTypes.WARNING} {...props} />;\n+export const WarningBox = (props: WarningBoxProps) => (\n+ <AlertBox type={AlertBoxTypes.WARNING} {...props} />\n+);\n", "CommunitySupportedDriver.tsx": "@@ -0,0 +1,20 @@\n+import { WarningBox } from \"@/components/mdx/AlertBox/AlertBox\";\n+import { Link } from \"@/components/overrides/Anchor/Link\";\n+\n+export interface CommunitySupportedDriverProps {\n+ dataSource: string;\n+}\n+\n+export const CommunitySupportedDriver = ({\n+ dataSource,\n+}: CommunitySupportedDriverProps) => {\n+ return (\n+ <WarningBox>\n+ The driver for {dataSource} is{\" \"}\n+ <Link href=\"/product/configuration/data-sources#driver-support\">\n+ community-supported\n+ </Link>{\" \"}\n+ and is not supported by Cube or the vendor.\n+ </WarningBox>\n+ );\n+};\n", "index.ts": "@@ -26,6 +26,7 @@ import { Table } from '@/components/overrides/Table/Table';\n import { Td } from '@/components/overrides/Table/Td';\n import { Th } from '@/components/overrides/Table/Th';\n import { Tr } from '@/components/overrides/Table/Tr';\n+import { CommunitySupportedDriver } from '@/components/mdx/Banners/CommunitySupportedDriver';\n \n export const components = {\n ...Buttons,\n@@ -54,6 +55,8 @@ export const components = {\n Diagram,\n YouTubeVideo,\n \n+ CommunitySupportedDriver,\n+\n // Overrides\n h1: H1,\n a: Link,\n", "real-time-data-fetch.mdx": "@@ -108,9 +108,9 @@ const Chart = ({ query }) => {\n ## Refresh Rate\n \n As in the case of a regular data fetch, real-time data fetch obeys\n-[`refresh_key` refresh rules](caching#refresh-keys). In order to provide a\n-desired refresh rate, `refresh_key` should reflect the rate of change of the\n-underlying data set; the querying time should also be much less than the desired\n-refresh rate. Please use the\n+[`refresh_key` refresh rules](/product/caching#refresh-keys). In order to\n+provide a desired refresh rate, `refresh_key` should reflect the rate of change\n+of the underlying data set; the querying time should also be much less than the\n+desired refresh rate. Please use the\n [`every`](/product/data-modeling/reference/cube#refresh_key) parameter to adjust\n the refresh interval.\n", "_meta.js": "@@ -7,7 +7,7 @@ module.exports = {\n \"elasticsearch\": \"Elasticsearch\",\n \"firebolt\": \"Firebolt\",\n \"google-bigquery\": \"Google BigQuery\",\n- \"hive\": \"Hive\",\n+ \"hive\": \"Hive / SparkSQL\",\n \"ksqldb\": \"ksqlDB\",\n \"materialize\": \"Materialize\",\n \"mongodb\": \"MongoDB\",\n", "druid.mdx": "@@ -5,11 +5,7 @@ redirect_from:\n \n # Druid\n \n-<WarningBox>\n- The driver for Druid is{\" \"}\n- <a href=\"../databases#driver-support\">community-supported</a> and is not\n- supported by Cube or the vendor.\n-</WarningBox>\n+<CommunitySupportedDriver dataSource=\"Druid\" />\n \n ## Prerequisites\n \n", "elasticsearch.mdx": "@@ -5,11 +5,7 @@ redirect_from:\n \n # Elasticsearch\n \n-<WarningBox>\n- The driver for Elasticsearch is{\" \"}\n- <a href=\"../databases#driver-support\">community-supported</a> and is not\n- supported by Cube or the vendor.\n-</WarningBox>\n+<CommunitySupportedDriver dataSource=\"Elasticsearch\" />\n \n ## Prerequisites\n \n", "hive.mdx": "@@ -3,13 +3,9 @@ redirect_from:\n - /config/databases/hive-sparksql\n ---\n \n-# Hive\n+# Hive / SparkSQL\n \n-<WarningBox>\n- The driver for Hive/SparkSQL is{\" \"}\n- <a href=\"../databases#driver-support\">community-supported</a> and is not\n- supported by Cube or the vendor.\n-</WarningBox>\n+<CommunitySupportedDriver dataSource=\"Hive / SparkSQL\" />\n \n ## Prerequisites\n \n", "mongodb.mdx": "@@ -5,11 +5,7 @@ redirect_from:\n \n # MongoDB\n \n-<WarningBox>\n- The driver for MongoDB is{\" \"}\n- <a href=\"../databases#driver-support\">community-supported</a> and is not\n- supported by Cube or the vendor.\n-</WarningBox>\n+<CommunitySupportedDriver dataSource=\"MongoDB\" />\n \n ## Prerequisites\n \n", "oracle.mdx": "@@ -5,11 +5,7 @@ redirect_from:\n \n # Oracle\n \n-<WarningBox>\n- The driver for Oracle is{\" \"}\n- <a href=\"../databases#driver-support\">community-supported</a> and is not\n- supported by Cube or the vendor.\n-</WarningBox>\n+<CommunitySupportedDriver dataSource=\"Oracle\" />\n \n ## Prerequisites\n \n", "sqlite.mdx": "@@ -5,11 +5,7 @@ redirect_from:\n \n # SQLite\n \n-<WarningBox>\n- The driver for SQLite is{\" \"}\n- <a href=\"../databases#driver-support\">community-supported</a> and is not\n- supported by Cube or the vendor.\n-</WarningBox>\n+<CommunitySupportedDriver dataSource=\"SQLite\" />\n \n ## Prerequisites\n \n", "visualization-tools.mdx": "@@ -135,17 +135,17 @@ Cube provides integration libraries for popular front-end frameworks:\n \n <Grid imageSize={[56, 56]}>\n <GridItem\n- url=\"../frontend-introduction/react\"\n+ url=\"/product/apis-integrations/javascript-sdk/react\"\n imageUrl=\"https://static.cube.dev/icons/react.svg\"\n title=\"React\"\n />\n <GridItem\n- url=\"../frontend-introduction/vue\"\n+ url=\"/product/apis-integrations/javascript-sdk/vue\"\n imageUrl=\"https://static.cube.dev/icons/vue.svg\"\n title=\"Vue\"\n />\n <GridItem\n- url=\"../frontend-introduction/angular\"\n+ url=\"/product/apis-integrations/javascript-sdk/angular\"\n imageUrl=\"https://static.cube.dev/icons/angular.svg\"\n title=\"Angular\"\n />\n@@ -159,17 +159,17 @@ out REST and GraphQL APIs.\n \n <Grid imageSize={[56, 56]}>\n <GridItem\n- url=\"../backend/sql\"\n+ url=\"/product/apis-integrations/sql-api\"\n imageUrl=\"https://raw.githubusercontent.com/cube-js/cube.js/master/docs/static/icons/sql.svg\"\n title=\"SQL API\"\n />\n <GridItem\n- url=\"../rest-api\"\n+ url=\"/product/apis-integrations/rest-api\"\n imageUrl=\"https://raw.githubusercontent.com/cube-js/cube.js/master/docs/static/icons/rest.svg\"\n title=\"REST API\"\n />\n <GridItem\n- url=\"../backend/graphql\"\n+ url=\"/product/apis-integrations/graphql-api\"\n imageUrl=\"https://raw.githubusercontent.com/cube-js/cube.js/master/docs/static/icons/graphql.svg\"\n title=\"GraphQL API\"\n />\n", "observable.mdx": "@@ -211,4 +211,4 @@ You can also create a visualization of the executed REST API request.\n \n [ref-getting-started]: /product/getting-started/cloud\n [ref-sql-api]: /product/apis-integrations/sql-api\n-[ref-rest-api]: /backend/rest-api\n+[ref-rest-api]: /product/apis-integrations/rest-api\n", "concepts.mdx": "@@ -536,7 +536,8 @@ Pre-Aggregations][ref-caching-preaggs-intro].\n /product/data-modeling/reference/joins#relationship\n [ref-schema-ref-sql]: /product/data-modeling/reference/cube#sql\n [ref-schema-ref-sql-table]: /product/data-modeling/reference/cube#sql_table\n-[ref-tutorial-incremental-preagg]: /incremental-pre-aggregations\n+[ref-tutorial-incremental-preagg]:\n+ /product/data-modeling/reference/pre-aggregations#incremental\n [self-dimensions]: #dimensions\n [self-measures]: #measures\n [wiki-olap]: https://en.wikipedia.org/wiki/Online_analytical_processing\n", "learn-more.mdx": "@@ -24,7 +24,7 @@ Cube can be queried in a variety of ways. Explore how to use\n \n ## Caching\n \n-Learn more about the [two-level cache](/docs/caching) and how\n+Learn more about the [two-level cache](/product/caching) and how\n [pre-aggregations help speed up queries](/product/caching/getting-started-pre-aggregations).\n For a deeper dive, take a look at the\n [related recipes](/guides/recipes/overview#recipes-query-acceleration).\n", "integrations.mdx": "@@ -38,12 +38,12 @@ following guides and configuration examples to get tool-specific instructions:\n \n <Grid imageSize={[56, 56]}>\n <GridItem\n- url=\"datadog\"\n+ url=\"integrations/datadog\"\n imageUrl=\"https://static.cube.dev/icons/datadog.svg\"\n title=\"Datadog\"\n />\n <GridItem\n- url=\"grafana-cloud\"\n+ url=\"integrations/grafana-cloud\"\n imageUrl=\"https://static.cube.dev/icons/grafana.svg\"\n title=\"Grafana Cloud\"\n />\n", "index.d.ts": "@@ -116,12 +116,12 @@ declare module '@cubejs-client/react' {\n \n type QueryRendererProps = {\n /**\n- * Analytic query. [Learn more about it's format](query-format)\n+ * Analytic query. [Learn more about it's format](/product/apis-integrations/rest-api/query-format)\n */\n query: Query | Query[];\n queries?: { [key: string]: Query };\n /**\n- * Indicates whether the generated by `Cube.js` SQL Code should be requested. See [rest-api#sql](rest-api#api-reference-v-1-sql). When set to `only` then only the request to [/v1/sql](rest-api#api-reference-v-1-sql) will be performed. When set to `true` the sql request will be performed along with the query request. Will not be performed if set to `false`\n+ * Indicates whether the generated by `Cube.js` SQL Code should be requested. See [rest-api#sql](/reference/rest-api#v1sql). When set to `only` then only the request to [/v1/sql](/reference/rest-api#v1sql) will be performed. When set to `true` the sql request will be performed along with the query request. Will not be performed if set to `false`\n */\n loadSql?: 'only' | boolean;\n /**\n@@ -459,7 +459,7 @@ declare module '@cubejs-client/react' {\n */\n skip?: boolean;\n /**\n- * Use continuous fetch behavior. See [Real-Time Data Fetch](real-time-data-fetch)\n+ * Use continuous fetch behavior. See [Real-Time Data Fetch](/product/apis-integrations/rest-api/real-time-data-fetch)\n */\n subscribe?: boolean;\n /**\n"} |
|
feat: octal Debug representation of `tree::EntryMode`.
This makes it easier to reason about. | cd61c25369d3e39b6160bac4b332b177dabddf4b | feat | https://github.com/Byron/gitoxide/commit/cd61c25369d3e39b6160bac4b332b177dabddf4b | octal Debug representation of `tree::EntryMode`.
This makes it easier to reason about. | {"tree_with_rewrites.rs": "@@ -14,25 +14,19 @@ fn empty_to_new_tree_without_rename_tracking() -> crate::Result {\n Addition {\n location: \"a\",\n relation: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n },\n Addition {\n location: \"b\",\n relation: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n },\n Addition {\n location: \"d\",\n relation: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n },\n Addition {\n@@ -42,9 +36,7 @@ fn empty_to_new_tree_without_rename_tracking() -> crate::Result {\n 1,\n ),\n ),\n- entry_mode: EntryMode(\n- 16384,\n- ),\n+ entry_mode: EntryMode(0o40000),\n id: Sha1(587ff082e0b98914788500eae5dd6a33f04883c9),\n },\n Addition {\n@@ -54,9 +46,7 @@ fn empty_to_new_tree_without_rename_tracking() -> crate::Result {\n 1,\n ),\n ),\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n },\n ]\n@@ -76,25 +66,19 @@ fn empty_to_new_tree_without_rename_tracking() -> crate::Result {\n Addition {\n location: \"a\",\n relation: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n },\n Addition {\n location: \"b\",\n relation: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n },\n Addition {\n location: \"d\",\n relation: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n },\n Addition {\n@@ -104,9 +88,7 @@ fn empty_to_new_tree_without_rename_tracking() -> crate::Result {\n 1,\n ),\n ),\n- entry_mode: EntryMode(\n- 16384,\n- ),\n+ entry_mode: EntryMode(0o40000),\n id: Sha1(587ff082e0b98914788500eae5dd6a33f04883c9),\n },\n Addition {\n@@ -116,9 +98,7 @@ fn empty_to_new_tree_without_rename_tracking() -> crate::Result {\n 1,\n ),\n ),\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n },\n ]\n@@ -153,35 +133,23 @@ fn changes_against_modified_tree_with_filename_tracking() -> crate::Result {\n [\n Modification {\n location: \"a\",\n- previous_entry_mode: EntryMode(\n- 33188,\n- ),\n+ previous_entry_mode: EntryMode(0o100644),\n previous_id: Sha1(78981922613b2afb6025042ff6bd878ac1994e85),\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(b4f17b61de71d9b2e54ac9e62b1629ae2d97a6a7),\n },\n Modification {\n location: \"dir\",\n- previous_entry_mode: EntryMode(\n- 16384,\n- ),\n+ previous_entry_mode: EntryMode(0o40000),\n previous_id: Sha1(e5c63aefe4327cb1c780c71966b678ce8e4225da),\n- entry_mode: EntryMode(\n- 16384,\n- ),\n+ entry_mode: EntryMode(0o40000),\n id: Sha1(c7ac5f82f536976f3561c9999b5f11e5893358be),\n },\n Modification {\n location: \"dir/c\",\n- previous_entry_mode: EntryMode(\n- 33188,\n- ),\n+ previous_entry_mode: EntryMode(0o100644),\n previous_id: Sha1(6695780ceb14b05e076a99bbd2babf34723b3464),\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(40006fcef15a8853a1b7ae186d93b7d680fd29cf),\n },\n ]\n@@ -198,35 +166,23 @@ fn changes_against_modified_tree_with_filename_tracking() -> crate::Result {\n [\n Modification {\n location: \"a\",\n- previous_entry_mode: EntryMode(\n- 33188,\n- ),\n+ previous_entry_mode: EntryMode(0o100644),\n previous_id: Sha1(78981922613b2afb6025042ff6bd878ac1994e85),\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(b4f17b61de71d9b2e54ac9e62b1629ae2d97a6a7),\n },\n Modification {\n location: \"dir\",\n- previous_entry_mode: EntryMode(\n- 16384,\n- ),\n+ previous_entry_mode: EntryMode(0o40000),\n previous_id: Sha1(e5c63aefe4327cb1c780c71966b678ce8e4225da),\n- entry_mode: EntryMode(\n- 16384,\n- ),\n+ entry_mode: EntryMode(0o40000),\n id: Sha1(c7ac5f82f536976f3561c9999b5f11e5893358be),\n },\n Modification {\n location: \"c\",\n- previous_entry_mode: EntryMode(\n- 33188,\n- ),\n+ previous_entry_mode: EntryMode(0o100644),\n previous_id: Sha1(6695780ceb14b05e076a99bbd2babf34723b3464),\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(40006fcef15a8853a1b7ae186d93b7d680fd29cf),\n },\n ]\n@@ -340,40 +296,28 @@ fn rename_by_similarity() -> crate::Result {\n [\n Modification {\n location: \"b\",\n- previous_entry_mode: EntryMode(\n- 33188,\n- ),\n+ previous_entry_mode: EntryMode(0o100644),\n previous_id: Sha1(61780798228d17af2d34fce4cfbdf35556832472),\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(54781fa52cf133fa9d0bf59cfe2ef2621b5ad29f),\n },\n Modification {\n location: \"dir\",\n- previous_entry_mode: EntryMode(\n- 16384,\n- ),\n+ previous_entry_mode: EntryMode(0o40000),\n previous_id: Sha1(d1622e275dbb2cb3215a0bdcd2fc77273891f360),\n- entry_mode: EntryMode(\n- 16384,\n- ),\n+ entry_mode: EntryMode(0o40000),\n id: Sha1(6602e61ea053525e4907e155c0b3da3a269e1385),\n },\n Deletion {\n location: \"dir/c\",\n relation: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(40006fcef15a8853a1b7ae186d93b7d680fd29cf),\n },\n Addition {\n location: \"dir/c-moved\",\n relation: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(f01e8ddf5adc56985b9a1cda6d7c7ef9e3abe034),\n },\n ]\n@@ -404,31 +348,21 @@ fn rename_by_similarity() -> crate::Result {\n [\n Modification {\n location: \"b\",\n- previous_entry_mode: EntryMode(\n- 33188,\n- ),\n+ previous_entry_mode: EntryMode(0o100644),\n previous_id: Sha1(61780798228d17af2d34fce4cfbdf35556832472),\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(54781fa52cf133fa9d0bf59cfe2ef2621b5ad29f),\n },\n Modification {\n location: \"dir\",\n- previous_entry_mode: EntryMode(\n- 16384,\n- ),\n+ previous_entry_mode: EntryMode(0o40000),\n previous_id: Sha1(d1622e275dbb2cb3215a0bdcd2fc77273891f360),\n- entry_mode: EntryMode(\n- 16384,\n- ),\n+ entry_mode: EntryMode(0o40000),\n id: Sha1(6602e61ea053525e4907e155c0b3da3a269e1385),\n },\n Rewrite {\n source_location: \"dir/c\",\n- source_entry_mode: EntryMode(\n- 33188,\n- ),\n+ source_entry_mode: EntryMode(0o100644),\n source_relation: None,\n source_id: Sha1(40006fcef15a8853a1b7ae186d93b7d680fd29cf),\n diff: Some(\n@@ -440,9 +374,7 @@ fn rename_by_similarity() -> crate::Result {\n similarity: 0.65,\n },\n ),\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(f01e8ddf5adc56985b9a1cda6d7c7ef9e3abe034),\n location: \"dir/c-moved\",\n relation: None,\n@@ -508,26 +440,18 @@ fn copies_by_identity() -> crate::Result {\n [\n Modification {\n location: \"dir\",\n- previous_entry_mode: EntryMode(\n- 16384,\n- ),\n+ previous_entry_mode: EntryMode(0o40000),\n previous_id: Sha1(6602e61ea053525e4907e155c0b3da3a269e1385),\n- entry_mode: EntryMode(\n- 16384,\n- ),\n+ entry_mode: EntryMode(0o40000),\n id: Sha1(f01fd5b4d733a4ae749cbb58a828cdb3f342f298),\n },\n Rewrite {\n source_location: \"base\",\n- source_entry_mode: EntryMode(\n- 33188,\n- ),\n+ source_entry_mode: EntryMode(0o100644),\n source_relation: None,\n source_id: Sha1(f00c965d8307308469e537302baa73048488f162),\n diff: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(f00c965d8307308469e537302baa73048488f162),\n location: \"c1\",\n relation: None,\n@@ -535,15 +459,11 @@ fn copies_by_identity() -> crate::Result {\n },\n Rewrite {\n source_location: \"base\",\n- source_entry_mode: EntryMode(\n- 33188,\n- ),\n+ source_entry_mode: EntryMode(0o100644),\n source_relation: None,\n source_id: Sha1(f00c965d8307308469e537302baa73048488f162),\n diff: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(f00c965d8307308469e537302baa73048488f162),\n location: \"c2\",\n relation: None,\n@@ -551,15 +471,11 @@ fn copies_by_identity() -> crate::Result {\n },\n Rewrite {\n source_location: \"base\",\n- source_entry_mode: EntryMode(\n- 33188,\n- ),\n+ source_entry_mode: EntryMode(0o100644),\n source_relation: None,\n source_id: Sha1(f00c965d8307308469e537302baa73048488f162),\n diff: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(f00c965d8307308469e537302baa73048488f162),\n location: \"dir/c3\",\n relation: None,\n@@ -592,26 +508,18 @@ fn copies_by_similarity() -> crate::Result {\n [\n Modification {\n location: \"dir\",\n- previous_entry_mode: EntryMode(\n- 16384,\n- ),\n+ previous_entry_mode: EntryMode(0o40000),\n previous_id: Sha1(f01fd5b4d733a4ae749cbb58a828cdb3f342f298),\n- entry_mode: EntryMode(\n- 16384,\n- ),\n+ entry_mode: EntryMode(0o40000),\n id: Sha1(1d7e20e07562a54af0408fd2669b0c56a6faa6f0),\n },\n Rewrite {\n source_location: \"base\",\n- source_entry_mode: EntryMode(\n- 33188,\n- ),\n+ source_entry_mode: EntryMode(0o100644),\n source_relation: None,\n source_id: Sha1(3bb459b831ea471b9cd1cbb7c6d54a74251a711b),\n diff: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(3bb459b831ea471b9cd1cbb7c6d54a74251a711b),\n location: \"c4\",\n relation: None,\n@@ -619,9 +527,7 @@ fn copies_by_similarity() -> crate::Result {\n },\n Rewrite {\n source_location: \"base\",\n- source_entry_mode: EntryMode(\n- 33188,\n- ),\n+ source_entry_mode: EntryMode(0o100644),\n source_relation: None,\n source_id: Sha1(3bb459b831ea471b9cd1cbb7c6d54a74251a711b),\n diff: Some(\n@@ -633,9 +539,7 @@ fn copies_by_similarity() -> crate::Result {\n similarity: 0.8888889,\n },\n ),\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(08fe19ca4d2f79624f35333157d610811efc1aed),\n location: \"c5\",\n relation: None,\n@@ -643,9 +547,7 @@ fn copies_by_similarity() -> crate::Result {\n },\n Rewrite {\n source_location: \"base\",\n- source_entry_mode: EntryMode(\n- 33188,\n- ),\n+ source_entry_mode: EntryMode(0o100644),\n source_relation: None,\n source_id: Sha1(3bb459b831ea471b9cd1cbb7c6d54a74251a711b),\n diff: Some(\n@@ -657,9 +559,7 @@ fn copies_by_similarity() -> crate::Result {\n similarity: 0.8888889,\n },\n ),\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(cf7a729ca69bfabd0995fc9b083e86a18215bd91),\n location: \"dir/c6\",\n relation: None,\n@@ -729,15 +629,11 @@ fn copies_in_entire_tree_by_similarity() -> crate::Result {\n [\n Rewrite {\n source_location: \"base\",\n- source_entry_mode: EntryMode(\n- 33188,\n- ),\n+ source_entry_mode: EntryMode(0o100644),\n source_relation: None,\n source_id: Sha1(3bb459b831ea471b9cd1cbb7c6d54a74251a711b),\n diff: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(3bb459b831ea471b9cd1cbb7c6d54a74251a711b),\n location: \"c6\",\n relation: None,\n@@ -745,15 +641,11 @@ fn copies_in_entire_tree_by_similarity() -> crate::Result {\n },\n Rewrite {\n source_location: \"dir/c6\",\n- source_entry_mode: EntryMode(\n- 33188,\n- ),\n+ source_entry_mode: EntryMode(0o100644),\n source_relation: None,\n source_id: Sha1(cf7a729ca69bfabd0995fc9b083e86a18215bd91),\n diff: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(cf7a729ca69bfabd0995fc9b083e86a18215bd91),\n location: \"c7\",\n relation: None,\n@@ -761,9 +653,7 @@ fn copies_in_entire_tree_by_similarity() -> crate::Result {\n },\n Rewrite {\n source_location: \"c5\",\n- source_entry_mode: EntryMode(\n- 33188,\n- ),\n+ source_entry_mode: EntryMode(0o100644),\n source_relation: None,\n source_id: Sha1(08fe19ca4d2f79624f35333157d610811efc1aed),\n diff: Some(\n@@ -775,9 +665,7 @@ fn copies_in_entire_tree_by_similarity() -> crate::Result {\n similarity: 0.75,\n },\n ),\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(97b3d1a5707f8a11fa5fa8bc6c3bd7b3965601fd),\n location: \"newly-added\",\n relation: None,\n@@ -785,13 +673,9 @@ fn copies_in_entire_tree_by_similarity() -> crate::Result {\n },\n Modification {\n location: \"b\",\n- previous_entry_mode: EntryMode(\n- 33188,\n- ),\n+ previous_entry_mode: EntryMode(0o100644),\n previous_id: Sha1(54781fa52cf133fa9d0bf59cfe2ef2621b5ad29f),\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(f198d0640214092732566fb00543163845c8252c),\n },\n ]\n@@ -828,15 +712,11 @@ fn copies_in_entire_tree_by_similarity_with_limit() -> crate::Result {\n [\n Rewrite {\n source_location: \"base\",\n- source_entry_mode: EntryMode(\n- 33188,\n- ),\n+ source_entry_mode: EntryMode(0o100644),\n source_relation: None,\n source_id: Sha1(3bb459b831ea471b9cd1cbb7c6d54a74251a711b),\n diff: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(3bb459b831ea471b9cd1cbb7c6d54a74251a711b),\n location: \"c6\",\n relation: None,\n@@ -844,15 +724,11 @@ fn copies_in_entire_tree_by_similarity_with_limit() -> crate::Result {\n },\n Rewrite {\n source_location: \"dir/c6\",\n- source_entry_mode: EntryMode(\n- 33188,\n- ),\n+ source_entry_mode: EntryMode(0o100644),\n source_relation: None,\n source_id: Sha1(cf7a729ca69bfabd0995fc9b083e86a18215bd91),\n diff: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(cf7a729ca69bfabd0995fc9b083e86a18215bd91),\n location: \"c7\",\n relation: None,\n@@ -860,21 +736,15 @@ fn copies_in_entire_tree_by_similarity_with_limit() -> crate::Result {\n },\n Modification {\n location: \"b\",\n- previous_entry_mode: EntryMode(\n- 33188,\n- ),\n+ previous_entry_mode: EntryMode(0o100644),\n previous_id: Sha1(54781fa52cf133fa9d0bf59cfe2ef2621b5ad29f),\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(f198d0640214092732566fb00543163845c8252c),\n },\n Addition {\n location: \"newly-added\",\n relation: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(97b3d1a5707f8a11fa5fa8bc6c3bd7b3965601fd),\n },\n ]\n@@ -910,26 +780,18 @@ fn copies_by_similarity_with_limit() -> crate::Result {\n [\n Modification {\n location: \"dir\",\n- previous_entry_mode: EntryMode(\n- 16384,\n- ),\n+ previous_entry_mode: EntryMode(0o40000),\n previous_id: Sha1(f01fd5b4d733a4ae749cbb58a828cdb3f342f298),\n- entry_mode: EntryMode(\n- 16384,\n- ),\n+ entry_mode: EntryMode(0o40000),\n id: Sha1(1d7e20e07562a54af0408fd2669b0c56a6faa6f0),\n },\n Rewrite {\n source_location: \"base\",\n- source_entry_mode: EntryMode(\n- 33188,\n- ),\n+ source_entry_mode: EntryMode(0o100644),\n source_relation: None,\n source_id: Sha1(3bb459b831ea471b9cd1cbb7c6d54a74251a711b),\n diff: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(3bb459b831ea471b9cd1cbb7c6d54a74251a711b),\n location: \"c4\",\n relation: None,\n@@ -938,17 +800,13 @@ fn copies_by_similarity_with_limit() -> crate::Result {\n Addition {\n location: \"c5\",\n relation: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(08fe19ca4d2f79624f35333157d610811efc1aed),\n },\n Addition {\n location: \"dir/c6\",\n relation: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(cf7a729ca69bfabd0995fc9b083e86a18215bd91),\n },\n ]\n@@ -984,15 +842,11 @@ fn realistic_renames_by_identity() -> crate::Result {\n [\n Rewrite {\n source_location: \"git-index/src/file.rs\",\n- source_entry_mode: EntryMode(\n- 33188,\n- ),\n+ source_entry_mode: EntryMode(0o100644),\n source_relation: None,\n source_id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n diff: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n location: \"git-index/src/file/mod.rs\",\n relation: None,\n@@ -1001,20 +855,14 @@ fn realistic_renames_by_identity() -> crate::Result {\n Addition {\n location: \"git-index/tests/index/file/access.rs\",\n relation: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n },\n Modification {\n location: \"git-index/tests/index/file/mod.rs\",\n- previous_entry_mode: EntryMode(\n- 33188,\n- ),\n+ previous_entry_mode: EntryMode(0o100644),\n previous_id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(8ba3a16384aacc37d01564b28401755ce8053f51),\n },\n ]\n@@ -1070,36 +918,26 @@ fn realistic_renames_disabled() -> crate::Result {\n Deletion {\n location: \"git-index/src/file.rs\",\n relation: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n },\n Addition {\n location: \"git-index/src/file/mod.rs\",\n relation: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n },\n Addition {\n location: \"git-index/tests/index/file/access.rs\",\n relation: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n },\n Modification {\n location: \"git-index/tests/index/file/mod.rs\",\n- previous_entry_mode: EntryMode(\n- 33188,\n- ),\n+ previous_entry_mode: EntryMode(0o100644),\n previous_id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(8ba3a16384aacc37d01564b28401755ce8053f51),\n },\n ]\n@@ -1161,9 +999,7 @@ fn realistic_renames_disabled_2() -> crate::Result {\n 1,\n ),\n ),\n- entry_mode: EntryMode(\n- 16384,\n- ),\n+ entry_mode: EntryMode(0o40000),\n id: Sha1(0026010e87631065a2739f627622feb14f903fd4),\n },\n Addition {\n@@ -1173,9 +1009,7 @@ fn realistic_renames_disabled_2() -> crate::Result {\n 2,\n ),\n ),\n- entry_mode: EntryMode(\n- 16384,\n- ),\n+ entry_mode: EntryMode(0o40000),\n id: Sha1(0026010e87631065a2739f627622feb14f903fd4),\n },\n Deletion {\n@@ -1185,9 +1019,7 @@ fn realistic_renames_disabled_2() -> crate::Result {\n 1,\n ),\n ),\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n },\n Deletion {\n@@ -1197,9 +1029,7 @@ fn realistic_renames_disabled_2() -> crate::Result {\n 1,\n ),\n ),\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n },\n Addition {\n@@ -1209,9 +1039,7 @@ fn realistic_renames_disabled_2() -> crate::Result {\n 2,\n ),\n ),\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n },\n Addition {\n@@ -1221,9 +1049,7 @@ fn realistic_renames_disabled_2() -> crate::Result {\n 2,\n ),\n ),\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n },\n Deletion {\n@@ -1233,9 +1059,7 @@ fn realistic_renames_disabled_2() -> crate::Result {\n 1,\n ),\n ),\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n },\n Deletion {\n@@ -1245,9 +1069,7 @@ fn realistic_renames_disabled_2() -> crate::Result {\n 1,\n ),\n ),\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n },\n Deletion {\n@@ -1257,9 +1079,7 @@ fn realistic_renames_disabled_2() -> crate::Result {\n 1,\n ),\n ),\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n },\n Deletion {\n@@ -1269,9 +1089,7 @@ fn realistic_renames_disabled_2() -> crate::Result {\n 1,\n ),\n ),\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n },\n Deletion {\n@@ -1281,9 +1099,7 @@ fn realistic_renames_disabled_2() -> crate::Result {\n 1,\n ),\n ),\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n },\n Addition {\n@@ -1293,9 +1109,7 @@ fn realistic_renames_disabled_2() -> crate::Result {\n 2,\n ),\n ),\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n },\n Addition {\n@@ -1305,9 +1119,7 @@ fn realistic_renames_disabled_2() -> crate::Result {\n 2,\n ),\n ),\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n },\n Addition {\n@@ -1317,9 +1129,7 @@ fn realistic_renames_disabled_2() -> crate::Result {\n 2,\n ),\n ),\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n },\n Addition {\n@@ -1329,9 +1139,7 @@ fn realistic_renames_disabled_2() -> crate::Result {\n 2,\n ),\n ),\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n },\n Addition {\n@@ -1341,9 +1149,7 @@ fn realistic_renames_disabled_2() -> crate::Result {\n 2,\n ),\n ),\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n },\n Deletion {\n@@ -1353,9 +1159,7 @@ fn realistic_renames_disabled_2() -> crate::Result {\n 1,\n ),\n ),\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n },\n Addition {\n@@ -1365,9 +1169,7 @@ fn realistic_renames_disabled_2() -> crate::Result {\n 2,\n ),\n ),\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n },\n ]\n@@ -1456,33 +1258,25 @@ fn realistic_renames_disabled_3() -> crate::Result {\n Addition {\n location: \"src/ein.rs\",\n relation: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n },\n Addition {\n location: \"src/gix.rs\",\n relation: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n },\n Deletion {\n location: \"src/plumbing-cli.rs\",\n relation: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n },\n Deletion {\n location: \"src/porcelain-cli.rs\",\n relation: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n },\n ]\n@@ -1539,15 +1333,11 @@ fn realistic_renames_by_identity_3() -> crate::Result {\n [\n Rewrite {\n source_location: \"src/plumbing-cli.rs\",\n- source_entry_mode: EntryMode(\n- 33188,\n- ),\n+ source_entry_mode: EntryMode(0o100644),\n source_relation: None,\n source_id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n diff: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n location: \"src/ein.rs\",\n relation: None,\n@@ -1555,15 +1345,11 @@ fn realistic_renames_by_identity_3() -> crate::Result {\n },\n Rewrite {\n source_location: \"src/porcelain-cli.rs\",\n- source_entry_mode: EntryMode(\n- 33188,\n- ),\n+ source_entry_mode: EntryMode(0o100644),\n source_relation: None,\n source_id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n diff: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n location: \"src/gix.rs\",\n relation: None,\n@@ -1629,9 +1415,7 @@ fn realistic_renames_2() -> crate::Result {\n [\n Rewrite {\n source_location: \"git-sec\",\n- source_entry_mode: EntryMode(\n- 16384,\n- ),\n+ source_entry_mode: EntryMode(0o40000),\n source_relation: Some(\n Parent(\n 1,\n@@ -1639,9 +1423,7 @@ fn realistic_renames_2() -> crate::Result {\n ),\n source_id: Sha1(0026010e87631065a2739f627622feb14f903fd4),\n diff: None,\n- entry_mode: EntryMode(\n- 16384,\n- ),\n+ entry_mode: EntryMode(0o40000),\n id: Sha1(0026010e87631065a2739f627622feb14f903fd4),\n location: \"gix-sec\",\n relation: Some(\n@@ -1653,9 +1435,7 @@ fn realistic_renames_2() -> crate::Result {\n },\n Rewrite {\n source_location: \"git-sec/CHANGELOG.md\",\n- source_entry_mode: EntryMode(\n- 33188,\n- ),\n+ source_entry_mode: EntryMode(0o100644),\n source_relation: Some(\n ChildOfParent(\n 1,\n@@ -1663,9 +1443,7 @@ fn realistic_renames_2() -> crate::Result {\n ),\n source_id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n diff: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n location: \"gix-sec/CHANGELOG.md\",\n relation: Some(\n@@ -1677,9 +1455,7 @@ fn realistic_renames_2() -> crate::Result {\n },\n Rewrite {\n source_location: \"git-sec/Cargo.toml\",\n- source_entry_mode: EntryMode(\n- 33188,\n- ),\n+ source_entry_mode: EntryMode(0o100644),\n source_relation: Some(\n ChildOfParent(\n 1,\n@@ -1687,9 +1463,7 @@ fn realistic_renames_2() -> crate::Result {\n ),\n source_id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n diff: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n location: \"gix-sec/Cargo.toml\",\n relation: Some(\n@@ -1701,9 +1475,7 @@ fn realistic_renames_2() -> crate::Result {\n },\n Rewrite {\n source_location: \"git-sec/src/identity.rs\",\n- source_entry_mode: EntryMode(\n- 33188,\n- ),\n+ source_entry_mode: EntryMode(0o100644),\n source_relation: Some(\n ChildOfParent(\n 1,\n@@ -1711,9 +1483,7 @@ fn realistic_renames_2() -> crate::Result {\n ),\n source_id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n diff: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n location: \"gix-sec/src/identity.rs\",\n relation: Some(\n@@ -1725,9 +1495,7 @@ fn realistic_renames_2() -> crate::Result {\n },\n Rewrite {\n source_location: \"git-sec/src/lib.rs\",\n- source_entry_mode: EntryMode(\n- 33188,\n- ),\n+ source_entry_mode: EntryMode(0o100644),\n source_relation: Some(\n ChildOfParent(\n 1,\n@@ -1735,9 +1503,7 @@ fn realistic_renames_2() -> crate::Result {\n ),\n source_id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n diff: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n location: \"gix-sec/src/lib.rs\",\n relation: Some(\n@@ -1749,9 +1515,7 @@ fn realistic_renames_2() -> crate::Result {\n },\n Rewrite {\n source_location: \"git-sec/src/permission.rs\",\n- source_entry_mode: EntryMode(\n- 33188,\n- ),\n+ source_entry_mode: EntryMode(0o100644),\n source_relation: Some(\n ChildOfParent(\n 1,\n@@ -1759,9 +1523,7 @@ fn realistic_renames_2() -> crate::Result {\n ),\n source_id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n diff: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n location: \"gix-sec/src/permission.rs\",\n relation: Some(\n@@ -1773,9 +1535,7 @@ fn realistic_renames_2() -> crate::Result {\n },\n Rewrite {\n source_location: \"git-sec/src/trust.rs\",\n- source_entry_mode: EntryMode(\n- 33188,\n- ),\n+ source_entry_mode: EntryMode(0o100644),\n source_relation: Some(\n ChildOfParent(\n 1,\n@@ -1783,9 +1543,7 @@ fn realistic_renames_2() -> crate::Result {\n ),\n source_id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n diff: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n location: \"gix-sec/src/trust.rs\",\n relation: Some(\n@@ -1797,9 +1555,7 @@ fn realistic_renames_2() -> crate::Result {\n },\n Rewrite {\n source_location: \"git-sec/tests/sec.rs\",\n- source_entry_mode: EntryMode(\n- 33188,\n- ),\n+ source_entry_mode: EntryMode(0o100644),\n source_relation: Some(\n ChildOfParent(\n 1,\n@@ -1807,9 +1563,7 @@ fn realistic_renames_2() -> crate::Result {\n ),\n source_id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n diff: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n location: \"gix-sec/tests/sec.rs\",\n relation: Some(\n@@ -1821,9 +1575,7 @@ fn realistic_renames_2() -> crate::Result {\n },\n Rewrite {\n source_location: \"git-sec/tests/identity/mod.rs\",\n- source_entry_mode: EntryMode(\n- 33188,\n- ),\n+ source_entry_mode: EntryMode(0o100644),\n source_relation: Some(\n ChildOfParent(\n 1,\n@@ -1831,9 +1583,7 @@ fn realistic_renames_2() -> crate::Result {\n ),\n source_id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n diff: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(e69de29bb2d1d6434b8b29ae775ad8c2e48c5391),\n location: \"gix-sec/tests/identity/mod.rs\",\n relation: Some(\n@@ -1927,9 +1677,7 @@ fn realistic_renames_3_without_identity() -> crate::Result {\n [\n Rewrite {\n source_location: \"src/plumbing/options.rs\",\n- source_entry_mode: EntryMode(\n- 33188,\n- ),\n+ source_entry_mode: EntryMode(0o100644),\n source_relation: Some(\n ChildOfParent(\n 2,\n@@ -1937,9 +1685,7 @@ fn realistic_renames_3_without_identity() -> crate::Result {\n ),\n source_id: Sha1(00750edc07d6415dcc07ae0351e9397b0222b7ba),\n diff: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(00750edc07d6415dcc07ae0351e9397b0222b7ba),\n location: \"src/plumbing-renamed/options/mod.rs\",\n relation: Some(\n@@ -1951,9 +1697,7 @@ fn realistic_renames_3_without_identity() -> crate::Result {\n },\n Rewrite {\n source_location: \"src/plumbing/mod.rs\",\n- source_entry_mode: EntryMode(\n- 33188,\n- ),\n+ source_entry_mode: EntryMode(0o100644),\n source_relation: Some(\n ChildOfParent(\n 2,\n@@ -1961,9 +1705,7 @@ fn realistic_renames_3_without_identity() -> crate::Result {\n ),\n source_id: Sha1(0cfbf08886fca9a91cb753ec8734c84fcbe52c9f),\n diff: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(0cfbf08886fca9a91cb753ec8734c84fcbe52c9f),\n location: \"src/plumbing-renamed/mod.rs\",\n relation: Some(\n@@ -1975,9 +1717,7 @@ fn realistic_renames_3_without_identity() -> crate::Result {\n },\n Rewrite {\n source_location: \"src/plumbing/main.rs\",\n- source_entry_mode: EntryMode(\n- 33188,\n- ),\n+ source_entry_mode: EntryMode(0o100644),\n source_relation: Some(\n ChildOfParent(\n 2,\n@@ -1985,9 +1725,7 @@ fn realistic_renames_3_without_identity() -> crate::Result {\n ),\n source_id: Sha1(d00491fd7e5bb6fa28c517a0bb32b8b506539d4d),\n diff: None,\n- entry_mode: EntryMode(\n- 33188,\n- ),\n+ entry_mode: EntryMode(0o100644),\n id: Sha1(d00491fd7e5bb6fa28c517a0bb32b8b506539d4d),\n location: \"src/plumbing-renamed/main.rs\",\n relation: Some(\n@@ -1999,9 +1737,7 @@ fn realistic_renames_3_without_identity() -> crate::Result {\n },\n Rewrite {\n source_location: \"src/plumbing\",\n- source_entry_mode: EntryMode(\n- 16384,\n- ),\n+ source_entry_mode: EntryMode(0o40000),\n source_relation: Some(\n Parent(\n 2,\n@@ -2009,9 +1745,7 @@ fn realistic_renames_3_without_identity() -> crate::Result {\n ),\n source_id: Sha1(b9d41dcdbd92fcab2fb6594d04f2ad99b3472621),\n diff: None,\n- entry_mode: EntryMode(\n- 16384,\n- ),\n+ entry_mode: EntryMode(0o40000),\n id: Sha1(202702465d7bb291153629dc2e8b353afe9cbdae),\n location: \"src/plumbing-renamed\",\n relation: Some(\n", "mod.rs": "@@ -42,10 +42,16 @@ pub struct Editor<'a> {\n ///\n /// Note that even though it can be created from any `u16`, it should be preferable to\n /// create it by converting [`EntryKind`] into `EntryMode`.\n-#[derive(Clone, Copy, PartialEq, Eq, Debug, Ord, PartialOrd, Hash)]\n+#[derive(Clone, Copy, PartialEq, Eq, Ord, PartialOrd, Hash)]\n #[cfg_attr(feature = \"serde\", derive(serde::Serialize, serde::Deserialize))]\n pub struct EntryMode(pub u16);\n \n+impl std::fmt::Debug for EntryMode {\n+ fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {\n+ write!(f, \"EntryMode({:#o})\", self.0)\n+ }\n+}\n+\n /// A discretized version of ideal and valid values for entry modes.\n ///\n /// Note that even though it can represent every valid [mode](EntryMode), it might\n"} |
|
chore: mark query condition keys ('id:in' etc) as deprecated | 7248762178d3c0386e78a72fd5e4cfa9701de882 | chore | https://github.com/mikro-orm/mikro-orm/commit/7248762178d3c0386e78a72fd5e4cfa9701de882 | mark query condition keys ('id:in' etc) as deprecated | {"query-conditions.md": "@@ -35,6 +35,8 @@ const res = await orm.em.find(Author, {\n \n Another way to do this by including the operator in your keys:\n \n+> This approach is deprecated and will be removed in future versions.\n+\n ```typescript\n const res = await orm.em.find(Author, { $and: [\n { 'id:in': [1, 2, 7] },\n"} |
|
build(docker): simplify risingwave docker setup (#8126)
Remove risingwave-specific minio service in favor of existing minio
service. | f2ff173c1467c5921edfb0ac9790ff8b0340bfc9 | build | https://github.com/rohankumardubey/ibis/commit/f2ff173c1467c5921edfb0ac9790ff8b0340bfc9 | simplify risingwave docker setup (#8126)
Remove risingwave-specific minio service in favor of existing minio
service. | {"compose.yaml": "@@ -104,9 +104,10 @@ services:\n retries: 20\n test:\n - CMD-SHELL\n- - mc ready data && mc mb --ignore-existing data/trino\n+ - mc ready data && mc mb --ignore-existing data/trino data/risingwave\n networks:\n - trino\n+ - risingwave\n volumes:\n - $PWD/docker/minio/config.json:/.mc/config.json:ro\n \n@@ -537,74 +538,26 @@ services:\n networks:\n - impala\n \n- risingwave-minio:\n- image: \"quay.io/minio/minio:latest\"\n- command:\n- - server\n- - \"--address\"\n- - \"0.0.0.0:9301\"\n- - \"--console-address\"\n- - \"0.0.0.0:9400\"\n- - /data\n- expose:\n- - \"9301\"\n- - \"9400\"\n- ports:\n- - \"9301:9301\"\n- - \"9400:9400\"\n- depends_on: []\n- volumes:\n- - \"risingwave-minio:/data\"\n- entrypoint: /bin/sh -c \"set -e; mkdir -p \\\"/data/hummock001\\\"; /usr/bin/docker-entrypoint.sh \\\"$$0\\\" \\\"$$@\\\" \"\n- environment:\n- MINIO_CI_CD: \"1\"\n- MINIO_ROOT_PASSWORD: hummockadmin\n- MINIO_ROOT_USER: hummockadmin\n- MINIO_DOMAIN: \"risingwave-minio\"\n- container_name: risingwave-minio\n- healthcheck:\n- test:\n- - CMD-SHELL\n- - bash -c 'printf \\\"GET / HTTP/1.1\\n\\n\\\" > /dev/tcp/127.0.0.1/9301; exit $$?;'\n- interval: 5s\n- timeout: 5s\n- retries: 20\n- restart: always\n- networks:\n- - risingwave\n-\n risingwave:\n image: ghcr.io/risingwavelabs/risingwave:nightly-20240122\n command: \"standalone --meta-opts=\\\" \\\n --advertise-addr 0.0.0.0:5690 \\\n --backend mem \\\n- --state-store hummock+minio://hummockadmin:hummockadmin@risingwave-minio:9301/hummock001 \\\n- --data-directory hummock_001 \\\n- --config-path /risingwave.toml\\\" \\\n- --compute-opts=\\\" \\\n- --config-path /risingwave.toml \\\n- --advertise-addr 0.0.0.0:5688 \\\n- --role both \\\" \\\n- --frontend-opts=\\\" \\\n- --config-path /risingwave.toml \\\n- --listen-addr 0.0.0.0:4566 \\\n- --advertise-addr 0.0.0.0:4566 \\\" \\\n- --compactor-opts=\\\" \\\n- --advertise-addr 0.0.0.0:6660 \\\"\"\n- expose:\n- - \"4566\"\n+ --state-store hummock+minio://accesskey:secretkey@minio:9000/risingwave \\\n+ --data-directory hummock_001\\\" \\\n+ --compute-opts=\\\"--advertise-addr 0.0.0.0:5688 --role both\\\" \\\n+ --frontend-opts=\\\"--listen-addr 0.0.0.0:4566 --advertise-addr 0.0.0.0:4566\\\" \\\n+ --compactor-opts=\\\"--advertise-addr 0.0.0.0:6660\\\"\"\n ports:\n- - \"4566:4566\"\n+ - 4566:4566\n depends_on:\n- - risingwave-minio\n+ minio:\n+ condition: service_healthy\n volumes:\n- - \"./docker/risingwave/risingwave.toml:/risingwave.toml\"\n - risingwave:/data\n environment:\n RUST_BACKTRACE: \"1\"\n- # If ENABLE_TELEMETRY is not set, telemetry will start by default\n- ENABLE_TELEMETRY: ${ENABLE_TELEMETRY:-true}\n- container_name: risingwave\n+ ENABLE_TELEMETRY: \"false\"\n healthcheck:\n test:\n - CMD-SHELL\n@@ -612,10 +565,9 @@ services:\n - bash -c 'printf \\\"GET / HTTP/1.1\\n\\n\\\" > /dev/tcp/127.0.0.1/5688; exit $$?;'\n - bash -c 'printf \\\"GET / HTTP/1.1\\n\\n\\\" > /dev/tcp/127.0.0.1/4566; exit $$?;'\n - bash -c 'printf \\\"GET / HTTP/1.1\\n\\n\\\" > /dev/tcp/127.0.0.1/5690; exit $$?;'\n- interval: 5s\n- timeout: 5s\n+ interval: 1s\n retries: 20\n- restart: always\n+ restart: on-failure\n networks:\n - risingwave\n \n@@ -646,5 +598,4 @@ volumes:\n postgres:\n exasol:\n impala:\n- risingwave-minio:\n risingwave:\n", "risingwave.toml": "@@ -1,2 +0,0 @@\n-# RisingWave config file to be mounted into the Docker containers.\n-# See https://github.com/risingwavelabs/risingwave/blob/main/src/config/example.toml for example\n", "test_json.py": "@@ -41,8 +41,7 @@ pytestmark = [\n reason=\"https://github.com/ibis-project/ibis/pull/6920#discussion_r1373212503\",\n )\n @pytest.mark.broken(\n- [\"risingwave\"],\n- reason=\"TODO(Kexiang): order mismatch in array\",\n+ [\"risingwave\"], reason=\"TODO(Kexiang): order mismatch in array\", strict=False\n )\n def test_json_getitem(json_t, expr_fn, expected):\n expr = expr_fn(json_t)\n"} |
|
test: dont export entities from single file tests | c49db6414b6b6416c16d0d0590e43bbf1162f0a7 | test | https://github.com/mikro-orm/mikro-orm/commit/c49db6414b6b6416c16d0d0590e43bbf1162f0a7 | dont export entities from single file tests | {"custom-pivot-entity-auto-discovery.sqlite.test.ts": "@@ -13,7 +13,7 @@ import {\n import { SqliteDriver } from '@mikro-orm/sqlite';\n \n @Entity()\n-export class Order {\n+class Order {\n \n @PrimaryKey()\n id!: number;\n@@ -33,7 +33,7 @@ export class Order {\n }\n \n @Entity()\n-export class Product {\n+class Product {\n \n @PrimaryKey()\n id!: number;\n@@ -55,7 +55,7 @@ export class Product {\n }\n \n @Entity()\n-export class OrderItem {\n+class OrderItem {\n \n [OptionalProps]?: 'amount';\n \n", "GH725.test.ts": "@@ -1,5 +1,4 @@\n import { EntitySchema, MikroORM, sql, Type, ValidationError } from '@mikro-orm/core';\n-import type { AbstractSqlDriver } from '@mikro-orm/knex';\n import { SqliteDriver } from '@mikro-orm/sqlite';\n import { PostgreSqlDriver } from '@mikro-orm/postgresql';\n \n@@ -105,13 +104,12 @@ export const TestSchema2 = new EntitySchema<Test2>({\n describe('GH issue 725', () => {\n \n test('mapping values from returning statement to custom types', async () => {\n- const orm = await MikroORM.init<AbstractSqlDriver>({\n+ const orm = await MikroORM.init({\n entities: [TestSchema],\n- dbName: `mikro_orm_test_gh_725`,\n+ dbName: 'mikro_orm_test_gh_725',\n driver: PostgreSqlDriver,\n });\n await orm.schema.ensureDatabase();\n- await orm.schema.execute('CREATE EXTENSION IF NOT EXISTS \"uuid-ossp\"');\n await orm.schema.dropSchema();\n await orm.schema.createSchema();\n \n@@ -142,7 +140,7 @@ describe('GH issue 725', () => {\n });\n \n test('validation when trying to persist not discovered entity', async () => {\n- const orm = await MikroORM.init<AbstractSqlDriver>({\n+ const orm = await MikroORM.init({\n entities: [TestSchema2],\n dbName: `:memory:`,\n driver: SqliteDriver,\n", "GH4242.test.ts": "@@ -47,7 +47,6 @@ beforeAll(async () => {\n });\n \n await orm.schema.ensureDatabase();\n- await orm.schema.execute('CREATE EXTENSION IF NOT EXISTS \"uuid-ossp\"');\n await orm.schema.refreshDatabase();\n });\n \n", "GH1003.test.ts": "@@ -2,7 +2,7 @@ import { BaseEntity, Collection, MikroORM, Entity, ManyToOne, OneToMany, Primary\n import type { Ref } from '@mikro-orm/sqlite';\n \n @Entity()\n-export class Parent extends BaseEntity {\n+class Parent extends BaseEntity {\n \n @PrimaryKey()\n id!: string;\n@@ -13,7 +13,7 @@ export class Parent extends BaseEntity {\n }\n \n @Entity()\n-export class Child extends BaseEntity {\n+class Child extends BaseEntity {\n \n @PrimaryKey()\n id!: string;\n", "GH1009.test.ts": "@@ -1,7 +1,7 @@\n import { Collection, Entity, ManyToOne, MikroORM, OneToMany, PrimaryKey, Property } from '@mikro-orm/sqlite';\n \n @Entity({ tableName: 'brands' })\n-export class Brand {\n+class Brand {\n \n @PrimaryKey()\n id!: number;\n@@ -12,7 +12,7 @@ export class Brand {\n }\n \n @Entity({ tableName: 'brand_site_restrictions' })\n-export class BrandSiteRestriction {\n+class BrandSiteRestriction {\n \n @PrimaryKey()\n id!: number;\n@@ -26,7 +26,7 @@ export class BrandSiteRestriction {\n }\n \n @Entity({ tableName: 'placements' })\n-export class Placement {\n+class Placement {\n \n @PrimaryKey()\n id!: number;\n@@ -40,7 +40,7 @@ export class Placement {\n }\n \n @Entity({ tableName: 'publishers' })\n-export class Publisher {\n+class Publisher {\n \n @OneToMany({ entity: () => Site, mappedBy: 'publisher' })\n sites = new Collection<Site>(this);\n@@ -51,7 +51,7 @@ export class Publisher {\n }\n \n @Entity({ tableName: 'sites' })\n-export class Site {\n+class Site {\n \n @ManyToOne({ entity: () => Publisher, nullable: true })\n publisher?: Publisher;\n", "GH1041.test.ts": "@@ -2,7 +2,7 @@ import { Collection, Entity, LoadStrategy, ManyToMany, MikroORM, PopulateHint, P\n import { mockLogger } from '../helpers';\n \n @Entity()\n-export class App {\n+class App {\n \n @PrimaryKey()\n id!: number;\n@@ -16,7 +16,7 @@ export class App {\n }\n \n @Entity()\n-export class User {\n+class User {\n \n @PrimaryKey()\n id!: number;\n", "GH1115.test.ts": "@@ -1,7 +1,7 @@\n import { Entity, ManyToOne, MikroORM, PrimaryKey, Property } from '@mikro-orm/sqlite';\n \n @Entity()\n-export class B {\n+class B {\n \n @PrimaryKey()\n id!: number;\n@@ -12,7 +12,7 @@ export class B {\n }\n \n @Entity()\n-export class A {\n+class A {\n \n @PrimaryKey()\n id!: number;\n", "GH1171.test.ts": "@@ -2,7 +2,7 @@ import { Entity, MikroORM, OneToOne, PrimaryKey, Property } from '@mikro-orm/sql\n import { v4 } from 'uuid';\n \n @Entity()\n-export class B {\n+class B {\n \n @PrimaryKey()\n id: string = v4();\n@@ -13,7 +13,7 @@ export class B {\n }\n \n @Entity()\n-export class A {\n+class A {\n \n @PrimaryKey()\n id!: string;\n", "GH1395.test.ts": "@@ -6,7 +6,7 @@ export interface EmailMessageTest {\n }\n \n @Entity()\n-export class TestTemplate {\n+class TestTemplate {\n \n @PrimaryKey()\n _id!: ObjectId;\n", "GH1616.test.ts": "@@ -1,7 +1,7 @@\n import { Embeddable, Embedded, Entity, MikroORM, PrimaryKey, Property } from '@mikro-orm/sqlite';\n \n @Embeddable()\n-export class D {\n+class D {\n \n @Property({ type: 'boolean', nullable: true })\n test?: boolean = false;\n@@ -9,7 +9,7 @@ export class D {\n }\n \n @Embeddable()\n-export class C {\n+class C {\n \n @Embedded(() => D, { object: true, nullable: false })\n d!: D;\n@@ -17,7 +17,7 @@ export class C {\n }\n \n @Embeddable()\n-export class B {\n+class B {\n \n @Embedded(() => C, { object: true, nullable: false })\n c!: C;\n@@ -28,7 +28,7 @@ export class B {\n }\n \n @Entity()\n-export class A {\n+class A {\n \n @PrimaryKey()\n id!: number;\n", "GH1626.test.ts": "@@ -6,7 +6,7 @@ import {\n Property,\n } from '@mikro-orm/sqlite';\n import { mockLogger } from '../helpers';\n-export class NativeBigIntType extends BigIntType {\n+class NativeBigIntType extends BigIntType {\n \n override convertToJSValue(value: any): any {\n if (!value) {\n@@ -19,7 +19,7 @@ export class NativeBigIntType extends BigIntType {\n }\n \n @Entity()\n-export class Author {\n+class Author {\n \n @PrimaryKey({ type: NativeBigIntType, comment: 'PK' })\n id!: bigint;\n", "GH1704.test.ts": "@@ -2,7 +2,7 @@ import { Entity, PrimaryKey, Property, OneToOne, MikroORM } from '@mikro-orm/sql\n import { mockLogger } from '../helpers';\n \n @Entity()\n-export class Profile {\n+class Profile {\n \n @PrimaryKey()\n id!: number;\n", "GH1721.test.ts": "@@ -2,7 +2,7 @@ import { Entity, MikroORM, PrimaryKey, Property, Type } from '@mikro-orm/sqlite'\n import { Guid } from 'guid-typescript';\n import { mockLogger } from '../helpers';\n \n-export class GuidType extends Type<Guid | undefined, string | undefined> {\n+class GuidType extends Type<Guid | undefined, string | undefined> {\n \n override convertToDatabaseValue(value: Guid | undefined): string | undefined {\n if (!value) {\n@@ -27,7 +27,7 @@ export class GuidType extends Type<Guid | undefined, string | undefined> {\n }\n \n @Entity()\n-export class Couch {\n+class Couch {\n \n @PrimaryKey({ type: GuidType })\n id!: Guid;\n", "GH1902.test.ts": "@@ -14,7 +14,7 @@ import {\n } from '@mikro-orm/sqlite';\n \n @Entity({ tableName: 'users' })\n-export class UserEntity {\n+class UserEntity {\n \n @PrimaryKey({ type: 'number' })\n id!: number;\n@@ -32,7 +32,7 @@ export class UserEntity {\n }\n \n @Entity({ tableName: 'tenants' })\n-export class TenantEntity {\n+class TenantEntity {\n \n [OptionalProps]?: 'isEnabled';\n \n", "GH1910.test.ts": "@@ -2,7 +2,7 @@ import type { EntityManager } from '@mikro-orm/postgresql';\n import { Entity, MikroORM, PrimaryKey, Property } from '@mikro-orm/postgresql';\n \n @Entity()\n-export class A {\n+class A {\n \n @PrimaryKey({ type: 'number' })\n id!: number;\n", "GH1927.test.ts": "@@ -2,7 +2,7 @@ import { Collection, Entity, ManyToOne, MikroORM, OneToMany, PrimaryKey, Propert\n import { mockLogger } from '../helpers';\n \n @Entity()\n-export class Author {\n+class Author {\n \n @PrimaryKey()\n id!: number;\n@@ -20,7 +20,7 @@ export class Author {\n }\n \n @Entity()\n-export class Book {\n+class Book {\n \n @PrimaryKey()\n id!: number;\n", "GH2273.test.ts": "@@ -1,7 +1,7 @@\n import { Entity, LoadStrategy, MikroORM, OneToOne, PrimaryKey, Property } from '@mikro-orm/sqlite';\n \n @Entity()\n-export class Checkout {\n+class Checkout {\n \n @PrimaryKey()\n id!: number;\n@@ -14,7 +14,7 @@ export class Checkout {\n }\n \n @Entity()\n-export class Discount {\n+class Discount {\n \n @PrimaryKey()\n id!: number;\n@@ -35,7 +35,7 @@ export class Discount {\n }\n \n @Entity()\n-export class Checkout2 {\n+class Checkout2 {\n \n @PrimaryKey()\n id!: number;\n@@ -49,7 +49,7 @@ export class Checkout2 {\n }\n \n @Entity()\n-export class Discount2 {\n+class Discount2 {\n \n @PrimaryKey()\n id!: number;\n", "GH228.test.ts": "@@ -2,7 +2,7 @@ import { Entity, ManyToOne, MikroORM, PrimaryKey, Property } from '@mikro-orm/sq\n import { mockLogger } from '../helpers';\n \n @Entity()\n-export class B {\n+class B {\n \n @PrimaryKey({ type: 'number' })\n id!: number;\n", "GH2379.test.ts": "@@ -2,7 +2,7 @@ import { Collection, Entity, Ref, ManyToOne, MikroORM, OneToMany, OptionalProps,\n import { performance } from 'perf_hooks';\n \n @Entity()\n-export class VendorBuyerRelationship {\n+class VendorBuyerRelationship {\n \n [OptionalProps]?: 'created';\n \n@@ -24,7 +24,7 @@ export class VendorBuyerRelationship {\n }\n \n @Entity()\n-export class Member {\n+class Member {\n \n [OptionalProps]?: 'created';\n \n@@ -49,7 +49,7 @@ export class Member {\n }\n \n @Entity()\n-export class Job {\n+class Job {\n \n [OptionalProps]?: 'rejected';\n \n", "GH2395.test.ts": "@@ -1,7 +1,7 @@\n import { Cascade, Collection, Entity, Ref, ManyToOne, MikroORM, OneToMany, PrimaryKey } from '@mikro-orm/sqlite';\n \n @Entity()\n-export class Parent {\n+class Parent {\n \n @PrimaryKey()\n id!: number;\n@@ -18,7 +18,7 @@ export class Parent {\n }\n \n @Entity()\n-export class Child {\n+class Child {\n \n @PrimaryKey()\n id!: number;\n@@ -29,7 +29,7 @@ export class Child {\n }\n \n @Entity()\n-export class Child2 {\n+class Child2 {\n \n @PrimaryKey()\n id!: number;\n", "GH2406.test.ts": "@@ -1,7 +1,7 @@\n import { Collection, Entity, Ref, ManyToOne, MikroORM, OneToMany, PrimaryKey } from '@mikro-orm/sqlite';\n \n @Entity({ forceConstructor: true })\n-export class Parent {\n+class Parent {\n \n @PrimaryKey()\n id!: number;\n@@ -12,7 +12,7 @@ export class Parent {\n }\n \n @Entity({ forceConstructor: true })\n-export class Child {\n+class Child {\n \n @PrimaryKey()\n id!: number;\n", "GH2583.test.ts": "@@ -7,7 +7,7 @@ export enum WithEnumArrayValue {\n }\n \n @Entity()\n-export class WithEnumArray {\n+class WithEnumArray {\n \n @PrimaryKey()\n id!: number;\n", "GH2675.test.ts": "@@ -1,7 +1,7 @@\n import { Entity, LoadStrategy, ManyToOne, MikroORM, PrimaryKey, wrap } from '@mikro-orm/postgresql';\n \n @Entity()\n-export class A {\n+class A {\n \n @PrimaryKey()\n id!: number;\n@@ -9,7 +9,7 @@ export class A {\n }\n \n @Entity()\n-export class B {\n+class B {\n \n @PrimaryKey()\n id!: number;\n", "GH2774.test.ts": "@@ -1,7 +1,7 @@\n import { Embeddable, Embedded, Entity, MikroORM, PrimaryKey, Property } from '@mikro-orm/sqlite';\n \n @Embeddable()\n-export class Nested {\n+class Nested {\n \n @Property({ nullable: true })\n value: string | null = null;\n@@ -9,7 +9,7 @@ export class Nested {\n }\n \n @Embeddable()\n-export class Name {\n+class Name {\n \n @Property({ nullable: true })\n value: string | null = null;\n@@ -20,7 +20,7 @@ export class Name {\n }\n \n @Entity()\n-export class User {\n+class User {\n \n @PrimaryKey()\n id!: number;\n", "GH2781.test.ts": "@@ -1,7 +1,7 @@\n import { Entity, ManyToOne, MikroORM, PrimaryKey, Property } from '@mikro-orm/postgresql';\n \n @Entity()\n-export class Address {\n+class Address {\n \n @PrimaryKey()\n id!: number;\n@@ -22,7 +22,7 @@ export class Address {\n }\n \n @Entity()\n-export class Customer {\n+class Customer {\n \n @PrimaryKey()\n id!: number;\n", "GH2784.test.ts": "@@ -1,7 +1,7 @@\n import { Entity, MikroORM, PrimaryKey, Property } from '@mikro-orm/postgresql';\n \n @Entity()\n-export class Address {\n+class Address {\n \n @PrimaryKey()\n id!: number;\n", "GH2815.test.ts": "@@ -1,7 +1,7 @@\n import { Entity, MikroORM, OneToOne, PrimaryKey } from '@mikro-orm/sqlite';\n \n @Entity()\n-export class Position {\n+class Position {\n \n @PrimaryKey()\n id!: number;\n@@ -12,7 +12,7 @@ export class Position {\n }\n \n @Entity()\n-export class Leg {\n+class Leg {\n \n @PrimaryKey()\n id!: number;\n@@ -23,7 +23,7 @@ export class Leg {\n }\n \n @Entity()\n-export class Position2 {\n+class Position2 {\n \n @PrimaryKey()\n id!: number;\n@@ -34,7 +34,7 @@ export class Position2 {\n }\n \n @Entity()\n-export class Leg2 {\n+class Leg2 {\n \n @PrimaryKey()\n id!: number;\n", "GH2821.test.ts": "@@ -1,7 +1,7 @@\n import { Entity, MikroORM, OneToOne, PrimaryKey } from '@mikro-orm/sqlite';\n \n @Entity()\n-export class Position {\n+class Position {\n \n @PrimaryKey()\n id!: number;\n@@ -15,7 +15,7 @@ export class Position {\n }\n \n @Entity()\n-export class Leg {\n+class Leg {\n \n @PrimaryKey()\n id!: number;\n", "GH2882.test.ts": "@@ -1,7 +1,7 @@\n import { Collection, Entity, ManyToOne, MikroORM, OneToMany, PrimaryKey, wrap } from '@mikro-orm/sqlite';\n \n @Entity()\n-export class Parent {\n+class Parent {\n \n @PrimaryKey()\n id!: number;\n", "GH2974.test.ts": "@@ -1,7 +1,7 @@\n import { Collection, Entity, ManyToOne, MikroORM, OneToMany, PrimaryKey, Property, wrap } from '@mikro-orm/better-sqlite';\n \n @Entity()\n-export class SomeMany {\n+class SomeMany {\n \n @PrimaryKey()\n id!: number;\n@@ -15,7 +15,7 @@ export class SomeMany {\n }\n \n @Entity()\n-export class Test {\n+class Test {\n \n @PrimaryKey()\n id!: number;\n", "GH302.test.ts": "@@ -1,7 +1,7 @@\n import { Entity, Ref, MikroORM, PrimaryKey, Property, Reference, ManyToOne, OneToMany, Collection } from '@mikro-orm/sqlite';\n \n @Entity()\n-export class A {\n+class A {\n \n @PrimaryKey({ type: 'number' })\n id: number;\n@@ -20,7 +20,7 @@ export class A {\n }\n \n @Entity()\n-export class B {\n+class B {\n \n @PrimaryKey({ type: 'number' })\n id!: number;\n", "GH3026.test.ts": "@@ -2,7 +2,7 @@ import { Collection, Entity, ManyToOne, MikroORM, OneToMany, PrimaryKey, Propert\n import { mockLogger } from '../helpers';\n \n @Entity()\n-export class Ingredient {\n+class Ingredient {\n \n @PrimaryKey()\n id!: number;\n@@ -16,7 +16,7 @@ export class Ingredient {\n }\n \n @Entity()\n-export class Recipe {\n+class Recipe {\n \n @PrimaryKey()\n id!: number;\n@@ -30,7 +30,7 @@ export class Recipe {\n }\n \n @Entity()\n-export class RecipeIngredient {\n+class RecipeIngredient {\n \n @PrimaryKey()\n id!: number;\n", "GH3240.test.ts": "@@ -3,7 +3,7 @@ import { Collection, Entity, ManyToMany, MikroORM, PrimaryKey, Property } from '\n type SquadType = 'GROUND' | 'AIR';\n \n @Entity()\n-export class Soldier {\n+class Soldier {\n \n @PrimaryKey()\n id!: number;\n@@ -20,7 +20,7 @@ export class Soldier {\n }\n \n @Entity()\n-export class Squad {\n+class Squad {\n \n @PrimaryKey()\n id!: number;\n", "GH3287.test.ts": "@@ -1,7 +1,7 @@\n import { Collection, Entity, LoadStrategy, ManyToMany, MikroORM, PrimaryKey } from '@mikro-orm/better-sqlite';\n \n @Entity()\n-export class Group {\n+class Group {\n \n @PrimaryKey()\n id!: number;\n@@ -15,7 +15,7 @@ export class Group {\n }\n \n @Entity()\n-export class Participant {\n+class Participant {\n \n @PrimaryKey()\n id!: number;\n", "GH3490.test.ts": "@@ -1,7 +1,7 @@\n import { Collection, Entity, ManyToOne, MikroORM, OneToMany, PrimaryKey } from '@mikro-orm/sqlite';\n \n @Entity()\n-export class Contract {\n+class Contract {\n \n @PrimaryKey()\n id!: number;\n@@ -12,7 +12,7 @@ export class Contract {\n }\n \n @Entity()\n-export class Customer {\n+class Customer {\n \n @PrimaryKey()\n id!: number;\n", "GH3548.test.ts": "@@ -1,7 +1,7 @@\n import { MikroORM, ObjectId, Entity, PrimaryKey, Property, OneToOne } from '@mikro-orm/mongodb';\n \n @Entity()\n-export class Author {\n+class Author {\n \n @PrimaryKey()\n _id!: ObjectId;\n@@ -15,7 +15,7 @@ export class Author {\n }\n \n @Entity()\n-export class AuthorDetail {\n+class AuthorDetail {\n \n @PrimaryKey()\n _id!: ObjectId;\n", "GH3696.test.ts": "@@ -2,7 +2,7 @@ import { FullTextType, MikroORM, Collection, Entity, Index, ManyToMany, PrimaryK\n \n @Entity()\n @Unique({ properties: ['name'] })\n-export class Artist {\n+class Artist {\n \n @PrimaryKey()\n id!: number;\n@@ -23,7 +23,7 @@ export class Artist {\n }\n \n @Entity()\n-export class Song {\n+class Song {\n \n @PrimaryKey()\n id!: number;\n", "GH3738.test.ts": "@@ -12,7 +12,7 @@ import {\n import { randomUUID } from 'crypto';\n \n @Entity()\n-export class Question {\n+class Question {\n \n [OptionalProps]?: 'createdAt';\n \n@@ -31,7 +31,7 @@ export class Question {\n }\n \n @Entity()\n-export class Answer {\n+class Answer {\n \n [OptionalProps]?: 'createdAt' | 'question';\n \n", "GH3844.test.ts": "@@ -2,7 +2,7 @@ import { Entity, PrimaryKey, Property, OneToOne, Ref, ref } from '@mikro-orm/cor\n import { MikroORM } from '@mikro-orm/sqlite';\n \n @Entity()\n-export class GamePoolEntity {\n+class GamePoolEntity {\n \n @PrimaryKey()\n contract_address!: string;\n@@ -36,7 +36,7 @@ export class GamePoolEntity {\n }\n \n @Entity()\n-export class GamePoolScannerEntity {\n+class GamePoolScannerEntity {\n \n @OneToOne(() => GamePoolEntity, e => e.scanner, {\n primary: true,\n", "GH4295.test.ts": "@@ -14,7 +14,7 @@ class RunScheduleEntity {\n }\n \n @Entity()\n-export class AEntity {\n+class AEntity {\n \n @PrimaryKey()\n id!: number;\n", "GH4343.test.ts": "@@ -3,7 +3,7 @@ import { Entity, ManyToOne, PrimaryKey, Property, ref, Ref } from '@mikro-orm/co\n import { v4 } from 'uuid';\n \n @Entity()\n-export class LocalizedString {\n+class LocalizedString {\n \n @PrimaryKey({ type: 'uuid' })\n id = v4();\n@@ -21,7 +21,7 @@ export class LocalizedString {\n }\n \n @Entity()\n-export class Book {\n+class Book {\n \n @PrimaryKey({ type: 'uuid' })\n id = v4();\n", "GH4533.test.ts": "@@ -15,7 +15,7 @@ import { SqliteDriver } from '@mikro-orm/sqlite';\n import { mockLogger } from '../helpers';\n \n @Entity({ tableName: 'core_users' })\n-export class User {\n+class User {\n \n @PrimaryKey()\n id!: number;\n@@ -33,7 +33,7 @@ export class User {\n }\n \n @Entity({ tableName: 'core_roles' })\n-export class Role {\n+class Role {\n \n @PrimaryKey()\n id!: number;\n", "GH4973.test.ts": "@@ -2,7 +2,7 @@ import { Collection, Entity, OneToMany, MikroORM, PrimaryKey, Property, ManyToOn\n import { mockLogger } from '../helpers';\n \n @Entity()\n-export class User {\n+class User {\n \n @PrimaryKey()\n id!: number;\n@@ -13,7 +13,7 @@ export class User {\n }\n \n @Entity()\n-export class Book {\n+class Book {\n \n @PrimaryKey()\n id!: number;\n", "GH557.test.ts": "@@ -1,7 +1,7 @@\n import { MikroORM, Entity, ManyToOne, OneToOne, PrimaryKey, Property } from '@mikro-orm/sqlite';\n \n @Entity()\n-export class Rate {\n+class Rate {\n \n @PrimaryKey()\n id!: number;\n@@ -22,7 +22,7 @@ export class Rate {\n }\n \n @Entity()\n-export class Application {\n+class Application {\n \n @PrimaryKey()\n id!: number;\n", "GH572.test.ts": "@@ -2,7 +2,7 @@ import { Entity, Ref, MikroORM, OneToOne, PrimaryKey, Property, QueryOrder } fro\n import { mockLogger } from '../helpers';\n \n @Entity()\n-export class A {\n+class A {\n \n @PrimaryKey()\n id!: number;\n@@ -13,7 +13,7 @@ export class A {\n }\n \n @Entity()\n-export class B {\n+class B {\n \n @PrimaryKey()\n id!: number;\n", "GH755.test.ts": "@@ -1,6 +1,6 @@\n import { EntitySchema, MikroORM } from '@mikro-orm/sqlite';\n \n-export class Test {\n+class Test {\n \n id!: string;\n createdAt!: Date;\n", "GH811.test.ts": "@@ -2,7 +2,7 @@ import { Entity, helper, MikroORM, OneToOne, PrimaryKey, Property } from '@mikro\n import { v4 } from 'uuid';\n \n @Entity()\n-export class Address {\n+class Address {\n \n @PrimaryKey({ type: 'uuid' })\n id = v4();\n@@ -13,7 +13,7 @@ export class Address {\n }\n \n @Entity()\n-export class Contact {\n+class Contact {\n \n @PrimaryKey({ type: 'uuid' })\n id = v4();\n@@ -27,7 +27,7 @@ export class Contact {\n }\n \n @Entity()\n-export class Employee {\n+class Employee {\n \n @PrimaryKey({ type: 'uuid' })\n id = v4();\n", "sqlite-constraints.test.ts": "@@ -2,9 +2,8 @@ import { Entity, type EntityManager, ManyToOne, MikroORM, PrimaryKey, Property,\n import { SqliteDriver } from '@mikro-orm/sqlite';\n import { BetterSqliteDriver } from '@mikro-orm/better-sqlite';\n \n-\n @Entity()\n-export class Author {\n+class Author {\n \n @PrimaryKey({ type: 'string' })\n id!: string;\n@@ -14,9 +13,8 @@ export class Author {\n \n }\n \n-\n @Entity()\n-export class Book {\n+class Book {\n \n @PrimaryKey({ type: 'string' })\n id!: string;\n@@ -29,7 +27,6 @@ export class Book {\n \n }\n \n-\n async function createEntities(em: EntityManager) {\n const author = new Author();\n author.id = '1';\n@@ -44,7 +41,6 @@ async function createEntities(em: EntityManager) {\n return author;\n }\n \n-\n describe('sqlite driver', () => {\n \n let orm: MikroORM<SqliteDriver>;\n@@ -70,7 +66,6 @@ describe('sqlite driver', () => {\n });\n });\n \n-\n describe('better-sqlite driver', () => {\n \n let orm: MikroORM<BetterSqliteDriver>;\n@@ -95,4 +90,3 @@ describe('better-sqlite driver', () => {\n }\n });\n });\n-\n"} |
|
chore: impl some error conversions | ed0f8e1d57380fe5b76248bf8dd88973898718c4 | chore | https://github.com/erg-lang/erg/commit/ed0f8e1d57380fe5b76248bf8dd88973898718c4 | impl some error conversions | {"mod.rs": "@@ -186,6 +186,12 @@ impl From<ParserRunnerError> for CompileError {\n }\n }\n \n+impl From<CompileError> for ParserRunnerErrors {\n+ fn from(err: CompileError) -> Self {\n+ Self::new(vec![err.into()])\n+ }\n+}\n+\n impl From<CompileError> for ParserRunnerError {\n fn from(err: CompileError) -> Self {\n Self {\n", "error.rs": "@@ -609,6 +609,12 @@ impl ParserRunnerError {\n }\n }\n \n+impl From<ParserRunnerError> for LexError {\n+ fn from(err: ParserRunnerError) -> Self {\n+ Self::new(err.core)\n+ }\n+}\n+\n #[derive(Debug)]\n pub struct ParserRunnerErrors(Vec<ParserRunnerError>);\n \n@@ -618,6 +624,12 @@ impl_stream!(ParserRunnerErrors, ParserRunnerError);\n \n impl MultiErrorDisplay<ParserRunnerError> for ParserRunnerErrors {}\n \n+impl From<ParserRunnerErrors> for LexErrors {\n+ fn from(errs: ParserRunnerErrors) -> Self {\n+ Self(errs.0.into_iter().map(LexError::from).collect())\n+ }\n+}\n+\n impl fmt::Display for ParserRunnerErrors {\n fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {\n self.fmt_all(f)\n"} |
|
chore(deps): relock | 94959b143e68b92360441c7383e1930ff986e5e5 | chore | https://github.com/rohankumardubey/ibis/commit/94959b143e68b92360441c7383e1930ff986e5e5 | relock | {"poetry.lock": "@@ -4855,83 +4855,6 @@ dev = [\"coverage[toml] (==5.0.4)\", \"cryptography (>=3.4.0)\", \"pre-commit\", \"pyte\n docs = [\"sphinx (>=4.5.0,<5.0.0)\", \"sphinx-rtd-theme\", \"zope.interface\"]\n tests = [\"coverage[toml] (==5.0.4)\", \"pytest (>=6.0.0,<7.0.0)\"]\n \n-[[package]]\n-name = \"pymssql\"\n-version = \"2.2.11\"\n-description = \"DB-API interface to Microsoft SQL Server for Python. (new Cython-based version)\"\n-optional = true\n-python-versions = \"*\"\n-files = [\n- {file = \"pymssql-2.2.11-cp310-cp310-macosx_11_0_x86_64.whl\", hash = \"sha256:692ab328ac290bd2031bc4dd6deae32665dfffda1b12aaa92928d3ebc667d5ad\"},\n- {file = \"pymssql-2.2.11-cp310-cp310-manylinux_2_12_i686.manylinux2010_i686.whl\", hash = \"sha256:723a4612421027a01b51e42e786678a18c4a27613a3ccecf331c026e0cc41353\"},\n- {file = \"pymssql-2.2.11-cp310-cp310-manylinux_2_12_x86_64.manylinux2010_x86_64.whl\", hash = \"sha256:34ab2373ca607174ad7244cfe955c07b6bc77a1e21d3c3143dbe934dec82c3a4\"},\n- {file = \"pymssql-2.2.11-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl\", hash = \"sha256:1bc0ba19b4426c57509f065a03748d9ac230f1543ecdac57175e6ebd213a7bc0\"},\n- {file = \"pymssql-2.2.11-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl\", hash = \"sha256:e8d9d42a50f6e8e6b356e4e8b2fa1da725344ec0be6f8a6107b7196e5bd74906\"},\n- {file = \"pymssql-2.2.11-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl\", hash = \"sha256:aec64022a2419fad9f496f8e310522635e39d092970e1d55375ea0be86725174\"},\n- {file = \"pymssql-2.2.11-cp310-cp310-manylinux_2_28_x86_64.whl\", hash = \"sha256:c389c8041c94d4058827faf5735df5f8e4c1c1eebdd051859536dc393925a667\"},\n- {file = \"pymssql-2.2.11-cp310-cp310-win32.whl\", hash = \"sha256:6452326cecd4dcee359a6f8878b827118a8c8523cd24de5b3a971a7a172e4275\"},\n- {file = \"pymssql-2.2.11-cp310-cp310-win_amd64.whl\", hash = \"sha256:c1bde266dbc91b100abd0311102a6585df09cc963599421cc12fd6b4cfa8e3d3\"},\n- {file = \"pymssql-2.2.11-cp311-cp311-macosx_10_9_universal2.whl\", hash = \"sha256:6ddaf0597138179517bdbf5b5aa3caffee65987316dc906359a5d0801d0847ee\"},\n- {file = \"pymssql-2.2.11-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl\", hash = \"sha256:0c26af25991715431559cb5b37f243b8ff676540f504ed0317774dfc71827af1\"},\n- {file = \"pymssql-2.2.11-cp311-cp311-manylinux_2_17_i686.manylinux2014_i686.whl\", hash = \"sha256:410e8c40b7c1b421e750cf80ccf2da8d802ed815575758ac9a78c5f6cd995723\"},\n- {file = \"pymssql-2.2.11-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl\", hash = \"sha256:fa1767239ed45e1fa91d82fc0c63305750530787cd64089cabbe183eb538a35b\"},\n- {file = \"pymssql-2.2.11-cp311-cp311-manylinux_2_28_x86_64.whl\", hash = \"sha256:9a644e4158fed30ae9f3846f2f1c74d36fa1610eb552de35b7f611d063fa3c85\"},\n- {file = \"pymssql-2.2.11-cp311-cp311-win32.whl\", hash = \"sha256:1956c111debe67f69a9c839b33ce420f0e8def1ef5ff9831c03d8ac840f82376\"},\n- {file = \"pymssql-2.2.11-cp311-cp311-win_amd64.whl\", hash = \"sha256:0bdd1fb49b0e331e47e83f39d4af784c857e230bfc73519654bab29285c51c63\"},\n- {file = \"pymssql-2.2.11-cp312-cp312-macosx_10_9_universal2.whl\", hash = \"sha256:2609bbd3b715822bb4fa6d457b2985d32ad6ab9580fdb61ae6e0eee251791d24\"},\n- {file = \"pymssql-2.2.11-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl\", hash = \"sha256:c382aea9adaaee189f352d7a493e3f76c13f9337ec2b6aa40e76b114fa13ebac\"},\n- {file = \"pymssql-2.2.11-cp312-cp312-manylinux_2_17_i686.manylinux2014_i686.whl\", hash = \"sha256:5928324a09de7466368c15ece1de4ab5ea968d24943ceade758836f9fc7149f5\"},\n- {file = \"pymssql-2.2.11-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl\", hash = \"sha256:ee8b10f797d0bfec626b803891cf9e98480ee11f2e8459a7616cdb7e4e4bf2de\"},\n- {file = \"pymssql-2.2.11-cp312-cp312-manylinux_2_28_x86_64.whl\", hash = \"sha256:1d5aa1a090b17f4ba75ffac3bb371f6c8c869692b653689396f9b470fde06981\"},\n- {file = \"pymssql-2.2.11-cp312-cp312-win32.whl\", hash = \"sha256:1f7ba71cf81af65c005173f279928bf86700d295f97e4965e169b5764bc6c4f2\"},\n- {file = \"pymssql-2.2.11-cp312-cp312-win_amd64.whl\", hash = \"sha256:a0ebb0e40c93f8f1e40aad80f512ae4aa89cb1ec8a96964b9afedcff1d5813fd\"},\n- {file = \"pymssql-2.2.11-cp36-cp36m-macosx_10_14_x86_64.whl\", hash = \"sha256:e0ed115902956efaca9d9a20fa9b2b604e3e11d640416ca74900d215cdcbf3ab\"},\n- {file = \"pymssql-2.2.11-cp36-cp36m-manylinux_2_12_i686.manylinux2010_i686.whl\", hash = \"sha256:1a75afa17746972bb61120fb6ea907657fc1ab68250bbbd8b21a00d0720ed0f4\"},\n- {file = \"pymssql-2.2.11-cp36-cp36m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl\", hash = \"sha256:d2ae69d8e46637a203cfb48e05439fc9e2ff7646fa1f5396aa3577ce52810031\"},\n- {file = \"pymssql-2.2.11-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl\", hash = \"sha256:f13710240457ace5b8c9cca7f4971504656f5703b702895a86386e87c7103801\"},\n- {file = \"pymssql-2.2.11-cp36-cp36m-manylinux_2_17_i686.manylinux2014_i686.whl\", hash = \"sha256:d7234b0f61dd9ccb2304171b5fd7ed9db133b4ea7c835c9942c9dc5bfc00c1cb\"},\n- {file = \"pymssql-2.2.11-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl\", hash = \"sha256:0dcd76a8cc757c7cfe2d235f232a20d74ac8cebf9feabcdcbda5ef33157d14b1\"},\n- {file = \"pymssql-2.2.11-cp36-cp36m-manylinux_2_28_x86_64.whl\", hash = \"sha256:84aff3235ad1289c4079c548cfcdf7eaaf2475b9f81557351deb42e8f45a9c2d\"},\n- {file = \"pymssql-2.2.11-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.whl\", hash = \"sha256:5b081aa7b02911e3f299f7d1f68ce8ca585a5119d44601bf4483da0aae8c2181\"},\n- {file = \"pymssql-2.2.11-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.whl\", hash = \"sha256:d315f08c106c884d6b42f70c9518e765a5bc23f6d3a587346bc4e6f198768c7a\"},\n- {file = \"pymssql-2.2.11-cp36-cp36m-win32.whl\", hash = \"sha256:c8b35b3d5e326729e5edb73d593103d2dbfb474bd36ee95b4e85e1f8271ba98a\"},\n- {file = \"pymssql-2.2.11-cp36-cp36m-win_amd64.whl\", hash = \"sha256:139c5032e0a2765764987803f1266132fcc5da572848ccc4d29cebba794a4260\"},\n- {file = \"pymssql-2.2.11-cp37-cp37m-macosx_11_0_x86_64.whl\", hash = \"sha256:7bac28aed1d625a002e0289e0c18d1808cecbdc12e2a1a3927dbbaff66e5fff3\"},\n- {file = \"pymssql-2.2.11-cp37-cp37m-manylinux_2_12_i686.manylinux2010_i686.whl\", hash = \"sha256:4eeaacc1dbbc678f4e80c6fd6fc279468021fdf2e486adc8631ec0de6b6c0e62\"},\n- {file = \"pymssql-2.2.11-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl\", hash = \"sha256:428e32e53c554798bc2d0682a169fcb681df6b68544c4aedd1186018ea7e0447\"},\n- {file = \"pymssql-2.2.11-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl\", hash = \"sha256:2b621c5e32136dabc2fea25696beab0647ec336d25c04ab6d8eb8c8ee92f0e52\"},\n- {file = \"pymssql-2.2.11-cp37-cp37m-manylinux_2_17_i686.manylinux2014_i686.whl\", hash = \"sha256:658c85474ea01ca3a30de769df06f46681e882524b05c6994cd6fd985c485f27\"},\n- {file = \"pymssql-2.2.11-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl\", hash = \"sha256:070181361ab94bdaeb14b591a35d853f327bc90c660b04047d474274fbb80357\"},\n- {file = \"pymssql-2.2.11-cp37-cp37m-manylinux_2_28_x86_64.whl\", hash = \"sha256:492e49616b58b2d6caf4a2598cb344572870171a7b65ba1ac61a5e248b6a8e1c\"},\n- {file = \"pymssql-2.2.11-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.whl\", hash = \"sha256:803122aec31fbd52f5d65ef3b30b3bd2dc7b2a9e3a8223d16078a25805155c45\"},\n- {file = \"pymssql-2.2.11-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl\", hash = \"sha256:09075e129655ab1178d2d60efb9b3fbf5cdb6da2338ecdb3a92c53a4ad7efa0c\"},\n- {file = \"pymssql-2.2.11-cp37-cp37m-win32.whl\", hash = \"sha256:b4a8377527702d746c490c2ce67d17f1c351d182b49b82fae6e67ae206bf9663\"},\n- {file = \"pymssql-2.2.11-cp37-cp37m-win_amd64.whl\", hash = \"sha256:167313d91606dc7a3c05b2ad60491a138b7408a8779599ab6430a48a67f133f0\"},\n- {file = \"pymssql-2.2.11-cp38-cp38-macosx_11_0_x86_64.whl\", hash = \"sha256:8d418f4dca245421242ed9df59d3bcda0cd081650df6deb1bef7f157b6a6f9dd\"},\n- {file = \"pymssql-2.2.11-cp38-cp38-manylinux_2_12_i686.manylinux2010_i686.whl\", hash = \"sha256:f0c44169df8d23c7ce172bd90ef5deb44caf19f15990e4db266e3193071988a4\"},\n- {file = \"pymssql-2.2.11-cp38-cp38-manylinux_2_12_x86_64.manylinux2010_x86_64.whl\", hash = \"sha256:b78032e45ea33c55d430b93e55370b900479ea324fae5d5d32486cc0fdc0fedd\"},\n- {file = \"pymssql-2.2.11-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl\", hash = \"sha256:984d99ee6a2579f86c536b1b0354ad3dc9701e98a4b3953f1301b4695477cd2f\"},\n- {file = \"pymssql-2.2.11-cp38-cp38-manylinux_2_17_i686.manylinux2014_i686.whl\", hash = \"sha256:287c8f79a7eca0c6787405797bac0f7c502d9be151f3f823aae12042235f8426\"},\n- {file = \"pymssql-2.2.11-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl\", hash = \"sha256:85ea4ea296afcae34bc61e4e0ef2f503270fd4bb097b308a07a9194f1f063aa1\"},\n- {file = \"pymssql-2.2.11-cp38-cp38-manylinux_2_28_x86_64.whl\", hash = \"sha256:a114633fa02b7eb5bc63520bf07954106c0ed0ce032449c871abb8b8c435a872\"},\n- {file = \"pymssql-2.2.11-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.whl\", hash = \"sha256:7332db36a537cbc16640a0c3473a2e419aa5bc1f9953cada3212e7b2587de658\"},\n- {file = \"pymssql-2.2.11-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.whl\", hash = \"sha256:cd7292d872948c1f67c8cc12158f2c8ed9873d54368139ce1f67b2262ac34029\"},\n- {file = \"pymssql-2.2.11-cp38-cp38-win32.whl\", hash = \"sha256:fbca115e11685b5891755cc22b3db4348071b8d100a41e1ce93526d9c3dbf2d5\"},\n- {file = \"pymssql-2.2.11-cp38-cp38-win_amd64.whl\", hash = \"sha256:452b88a4ceca7efb934b5babb365851a3c52e723642092ebc92777397c2cacdb\"},\n- {file = \"pymssql-2.2.11-cp39-cp39-macosx_11_0_x86_64.whl\", hash = \"sha256:001242cedc73587cbb10aec4069de50febbff3c4c50f9908a215476496b3beab\"},\n- {file = \"pymssql-2.2.11-cp39-cp39-manylinux_2_12_i686.manylinux2010_i686.whl\", hash = \"sha256:da492482b923b9cc9ad37f0f5592c776279299db2a89c0b7fc931aaefec652d4\"},\n- {file = \"pymssql-2.2.11-cp39-cp39-manylinux_2_12_x86_64.manylinux2010_x86_64.whl\", hash = \"sha256:139a833e6e72a624e4f2cde803a34a616d5661dd9a5b2ae0402d9d8a597b2f1f\"},\n- {file = \"pymssql-2.2.11-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl\", hash = \"sha256:e57fbfad252434d64bdf4b6a935e4241616a4cf8df7af58b9772cd91fce9309a\"},\n- {file = \"pymssql-2.2.11-cp39-cp39-manylinux_2_17_i686.manylinux2014_i686.whl\", hash = \"sha256:a5308507c2c4e94ede7e5b164870c1ba2be55abab6daf795b5529e2da4e838b6\"},\n- {file = \"pymssql-2.2.11-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl\", hash = \"sha256:bdca43c42d5f370358535b2107140ed550d74f9ef0fc95d2d7fa8c4e40ee48c2\"},\n- {file = \"pymssql-2.2.11-cp39-cp39-manylinux_2_28_x86_64.whl\", hash = \"sha256:fe0cc975aac87b364fdb55cb89642435c3e859dcd99d7260f48af94111ba2673\"},\n- {file = \"pymssql-2.2.11-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.whl\", hash = \"sha256:4551f50c8a3b6ffbd71f794ee1c0c0134134c5d6414302c2fa28b67fe4470d07\"},\n- {file = \"pymssql-2.2.11-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.whl\", hash = \"sha256:ae9818df40588d5a49e7476f05e31cc83dea630d607178d66762ca8cf32e9f77\"},\n- {file = \"pymssql-2.2.11-cp39-cp39-win32.whl\", hash = \"sha256:15257c7bd89c0283f70d6eaafd9b872201818572b8ba1e8576408ae23ef50c7c\"},\n- {file = \"pymssql-2.2.11-cp39-cp39-win_amd64.whl\", hash = \"sha256:65bb674c0ba35379bf93d1b2cf06fdc5e7ec56e1d0e9de525bdcf977190b2865\"},\n- {file = \"pymssql-2.2.11.tar.gz\", hash = \"sha256:15815bf1ff9edb475ec4ef567f23e23c4e828ce119ff5bf98a072b66b8d0ac1b\"},\n-]\n-\n [[package]]\n name = \"pymysql\"\n version = \"1.1.0\"\n@@ -7416,7 +7339,7 @@ cffi = {version = \">=1.11\", markers = \"platform_python_implementation == \\\"PyPy\\\n cffi = [\"cffi (>=1.11)\"]\n \n [extras]\n-all = [\"black\", \"clickhouse-connect\", \"dask\", \"datafusion\", \"db-dtypes\", \"deltalake\", \"duckdb\", \"duckdb-engine\", \"fsspec\", \"geoalchemy2\", \"geopandas\", \"google-cloud-bigquery\", \"google-cloud-bigquery-storage\", \"graphviz\", \"impyla\", \"oracledb\", \"packaging\", \"polars\", \"psycopg2\", \"pydata-google-auth\", \"pydruid\", \"pymssql\", \"pymysql\", \"pyspark\", \"regex\", \"requests\", \"shapely\", \"snowflake-connector-python\", \"snowflake-sqlalchemy\", \"sqlalchemy\", \"sqlalchemy-exasol\", \"sqlalchemy-views\", \"trino\"]\n+all = [\"black\", \"clickhouse-connect\", \"dask\", \"datafusion\", \"db-dtypes\", \"deltalake\", \"duckdb\", \"duckdb-engine\", \"geoalchemy2\", \"geopandas\", \"google-cloud-bigquery\", \"google-cloud-bigquery-storage\", \"graphviz\", \"impyla\", \"oracledb\", \"packaging\", \"polars\", \"psycopg2\", \"pydata-google-auth\", \"pydruid\", \"pymysql\", \"pyodbc\", \"pyspark\", \"regex\", \"shapely\", \"snowflake-connector-python\", \"snowflake-sqlalchemy\", \"sqlalchemy\", \"sqlalchemy-exasol\", \"sqlalchemy-views\", \"trino\"]\n bigquery = [\"db-dtypes\", \"google-cloud-bigquery\", \"google-cloud-bigquery-storage\", \"pydata-google-auth\"]\n clickhouse = [\"clickhouse-connect\", \"sqlalchemy\"]\n dask = [\"dask\", \"regex\"]\n@@ -7428,8 +7351,8 @@ duckdb = [\"duckdb\", \"duckdb-engine\", \"sqlalchemy\", \"sqlalchemy-views\"]\n exasol = [\"sqlalchemy\", \"sqlalchemy-exasol\", \"sqlalchemy-views\"]\n flink = []\n geospatial = [\"geoalchemy2\", \"geopandas\", \"shapely\"]\n-impala = [\"fsspec\", \"impyla\", \"requests\", \"sqlalchemy\"]\n-mssql = [\"pymssql\", \"sqlalchemy\", \"sqlalchemy-views\"]\n+impala = [\"impyla\", \"sqlalchemy\"]\n+mssql = [\"pyodbc\", \"sqlalchemy\", \"sqlalchemy-views\"]\n mysql = [\"pymysql\", \"sqlalchemy\", \"sqlalchemy-views\"]\n oracle = [\"oracledb\", \"packaging\", \"sqlalchemy\", \"sqlalchemy-views\"]\n pandas = [\"regex\"]\n@@ -7444,4 +7367,4 @@ visualization = [\"graphviz\"]\n [metadata]\n lock-version = \"2.0\"\n python-versions = \"^3.9\"\n-content-hash = \"e33849b55adc9ca33aa5b98b94dbeca72c6cef7e7150890fe3c3adba206b3892\"\n+content-hash = \"7cdedb3e9657196bfe4485e8cdb35998c826cca681595b38f39e1ba253c2886c\"\n"} |
|
refactor: make EntityProperty interface generic (use keyof T on name) | 8c9ee4d0c15200b4f2a2a93abc6885caf3e6f419 | refactor | https://github.com/mikro-orm/mikro-orm/commit/8c9ee4d0c15200b4f2a2a93abc6885caf3e6f419 | make EntityProperty interface generic (use keyof T on name) | {"Entity.ts": "@@ -53,10 +53,10 @@ export type EntityName<T extends IEntityType<T>> = string | EntityClass<T>;\n \n export type EntityData<T extends IEntityType<T>> = { [P in keyof T]?: T[P] | IPrimaryKey; } & Record<string, any>;\n \n-export interface EntityProperty {\n- name: string;\n+export interface EntityProperty<T extends IEntityType<T> = any> {\n+ name: string & keyof T;\n fk: string;\n- entity: () => EntityName<IEntity>;\n+ entity: () => EntityName<T>;\n type: string;\n primary: boolean;\n length?: any;\n@@ -84,7 +84,7 @@ export interface EntityMetadata<T extends IEntityType<T> = any> {\n path: string;\n primaryKey: keyof T & string;\n serializedPrimaryKey: keyof T & string;\n- properties: { [K in keyof T & string]: EntityProperty };\n+ properties: { [K in keyof T & string]: EntityProperty<T> };\n customRepository: () => { new (em: EntityManager, entityName: EntityName<T>): EntityRepository<T> };\n hooks: Record<string, string[]>;\n prototype: EntityClass<T> & IEntity;\n"} |
|
fix: infinite recursion bugs | c31e93052b3d3390d53340590e78b24e786a4efb | fix | https://github.com/erg-lang/erg/commit/c31e93052b3d3390d53340590e78b24e786a4efb | infinite recursion bugs | {"eval.rs": "@@ -1337,6 +1337,8 @@ impl Context {\n }\n TyParam::FreeVar(fv) if fv.is_linked() => self.convert_tp_into_type(fv.crack().clone()),\n TyParam::Type(t) => Ok(t.as_ref().clone()),\n+ TyParam::Mono(name) => Ok(Type::Mono(name)),\n+ // TyParam::Erased(_t) => Ok(Type::Obj),\n TyParam::Value(v) => self.convert_value_into_type(v).map_err(TyParam::Value),\n // TODO: Dict, Set\n other => Err(other),\n@@ -1672,7 +1674,7 @@ impl Context {\n line!() as usize,\n ().loc(),\n self.caused_by(),\n- &tp.qual_name().unwrap_or(\"_\".into()),\n+ &tp.to_string(),\n )\n })?;\n if qt.is_generalized() {\n", "inquire.rs": "@@ -4,14 +4,15 @@ use std::path::{Path, PathBuf};\n \n use erg_common::config::Input;\n use erg_common::consts::{ERG_MODE, PYTHON_MODE};\n-use erg_common::dict;\n use erg_common::error::{ErrorCore, Location, SubMessage};\n use erg_common::levenshtein;\n use erg_common::set::Set;\n use erg_common::traits::{Locational, NoTypeDisplay, Stream};\n use erg_common::triple::Triple;\n use erg_common::Str;\n-use erg_common::{fmt_option, fmt_slice, log, option_enum_unwrap, set, switch_lang};\n+use erg_common::{\n+ dict, fmt_option, fmt_slice, get_hash, log, option_enum_unwrap, set, switch_lang,\n+};\n \n use erg_parser::ast::{self, Identifier, VarName};\n use erg_parser::token::Token;\n@@ -1024,20 +1025,23 @@ impl Context {\n let coerced = self\n .coerce(obj.t(), &())\n .map_err(|mut errs| errs.remove(0))?;\n- if &coerced == obj.ref_t() {\n- Err(TyCheckError::no_attr_error(\n- self.cfg.input.clone(),\n- line!() as usize,\n- attr_name.loc(),\n- namespace.name.to_string(),\n- obj.ref_t(),\n- attr_name.inspect(),\n- self.get_similar_attr(obj.ref_t(), attr_name.inspect()),\n- ))\n- } else {\n+ if &coerced != obj.ref_t() {\n+ let hash = get_hash(obj.ref_t());\n obj.ref_t().coerce();\n- self.search_method_info(obj, attr_name, pos_args, kw_args, input, namespace)\n+ if get_hash(obj.ref_t()) != hash {\n+ return self\n+ .search_method_info(obj, attr_name, pos_args, kw_args, input, namespace);\n+ }\n }\n+ Err(TyCheckError::no_attr_error(\n+ self.cfg.input.clone(),\n+ line!() as usize,\n+ attr_name.loc(),\n+ namespace.name.to_string(),\n+ obj.ref_t(),\n+ attr_name.inspect(),\n+ self.get_similar_attr(obj.ref_t(), attr_name.inspect()),\n+ ))\n }\n \n fn validate_visibility(\n@@ -1263,12 +1267,13 @@ impl Context {\n return Err(self.not_callable_error(obj, attr_name, instance, None));\n }\n if sub != Never {\n+ let hash = get_hash(instance);\n instance.coerce();\n if instance.is_quantified_subr() {\n let instance = self.instantiate(instance.clone(), obj)?;\n self.substitute_call(obj, attr_name, &instance, pos_args, kw_args)?;\n return Ok(SubstituteResult::Coerced(instance));\n- } else {\n+ } else if get_hash(instance) != hash {\n return self\n .substitute_call(obj, attr_name, instance, pos_args, kw_args);\n }\n"} |
|
fix: Parse refs from bytes, not from String.
The latter can cause issues around illformed UTF-8 which wouldn't
bother git either.
This comes at the expense of not parsing line by line anymore, but
instead reading as fast as possible and parsing afterwards.
Performance wise I think it doesn't matter, but it will cause
more memory to be used. If this ever becomes a problem,
for example during pushes where we are stuck with V1, we can consider
implementing our own streaming appreach that works with packet lines
instead - they are just not exposed here even though they could. | 806b8c2ef392137f3a6ebd0f28da2a3a07a9f3eb | fix | https://github.com/Byron/gitoxide/commit/806b8c2ef392137f3a6ebd0f28da2a3a07a9f3eb | Parse refs from bytes, not from String.
The latter can cause issues around illformed UTF-8 which wouldn't
bother git either.
This comes at the expense of not parsing line by line anymore, but
instead reading as fast as possible and parsing afterwards.
Performance wise I think it doesn't matter, but it will cause
more memory to be used. If this ever becomes a problem,
for example during pushes where we are stuck with V1, we can consider
implementing our own streaming appreach that works with packet lines
instead - they are just not exposed here even though they could. | {"Cargo.toml": "@@ -46,7 +46,7 @@ git-credentials = { version = \"^0.7.0\", path = \"../git-credentials\" }\n \n thiserror = \"1.0.32\"\n serde = { version = \"1.0.114\", optional = true, default-features = false, features = [\"derive\"]}\n-bstr = { version = \"1.0.1\", default-features = false, features = [\"std\"] }\n+bstr = { version = \"1.0.1\", default-features = false, features = [\"std\", \"unicode\"] }\n nom = { version = \"7\", default-features = false, features = [\"std\"]}\n btoi = \"0.4.2\"\n \n", "tests.rs": "@@ -15,6 +15,8 @@ unborn refs/heads/symbolic symref-target:refs/heads/target\n \"\n .as_bytes();\n \n+ #[cfg(feature = \"blocking-client\")]\n+ let input = &mut Fixture(input);\n let out = refs::from_v2_refs(input).await.expect(\"no failure on valid input\");\n \n assert_eq!(\n@@ -56,6 +58,7 @@ unborn refs/heads/symbolic symref-target:refs/heads/target\n \n #[maybe_async::test(feature = \"blocking-client\", async(feature = \"async-client\", async_std::test))]\n async fn extract_references_from_v1_refs() {\n+ #[cfg_attr(feature = \"blocking-client\", allow(unused_mut))]\n let input = &mut \"73a6868963993a3328e7d8fe94e5a6ac5078a944 HEAD\n 21c9b7500cb144b3169a6537961ec2b9e865be81 MISSING_NAMESPACE_TARGET\n 73a6868963993a3328e7d8fe94e5a6ac5078a944 refs/heads/main\n@@ -63,6 +66,8 @@ async fn extract_references_from_v1_refs() {\n dce0ea858eef7ff61ad345cc5cdac62203fb3c10 refs/tags/git-commitgraph-v0.0.0\n 21c9b7500cb144b3169a6537961ec2b9e865be81 refs/tags/git-commitgraph-v0.0.0^{}\"\n .as_bytes();\n+ #[cfg(feature = \"blocking-client\")]\n+ let input = &mut Fixture(input);\n let out = refs::from_v1_refs_received_as_part_of_handshake_and_capabilities(\n input,\n Capabilities::from_bytes(b\"\\0symref=HEAD:refs/heads/main symref=MISSING_NAMESPACE_TARGET:(null)\")\n@@ -106,7 +111,7 @@ fn extract_symbolic_references_from_capabilities() -> Result<(), client::Error>\n let caps = client::Capabilities::from_bytes(\n b\"\\0unrelated symref=HEAD:refs/heads/main symref=ANOTHER:refs/heads/foo symref=MISSING_NAMESPACE_TARGET:(null) agent=git/2.28.0\",\n )?\n- .0;\n+ .0;\n let out = refs::shared::from_capabilities(caps.iter()).expect(\"a working example\");\n \n assert_eq!(\n@@ -128,3 +133,38 @@ fn extract_symbolic_references_from_capabilities() -> Result<(), client::Error>\n );\n Ok(())\n }\n+\n+#[cfg(feature = \"blocking-client\")]\n+struct Fixture<'a>(&'a [u8]);\n+\n+#[cfg(feature = \"blocking-client\")]\n+impl<'a> std::io::Read for Fixture<'a> {\n+ fn read(&mut self, buf: &mut [u8]) -> std::io::Result<usize> {\n+ self.0.read(buf)\n+ }\n+}\n+\n+#[cfg(feature = \"blocking-client\")]\n+impl<'a> std::io::BufRead for Fixture<'a> {\n+ fn fill_buf(&mut self) -> std::io::Result<&[u8]> {\n+ self.0.fill_buf()\n+ }\n+\n+ fn consume(&mut self, amt: usize) {\n+ self.0.consume(amt)\n+ }\n+}\n+\n+#[cfg(feature = \"blocking-client\")]\n+impl<'a> git_transport::client::ReadlineBufRead for Fixture<'a> {\n+ fn readline(\n+ &mut self,\n+ ) -> Option<std::io::Result<Result<git_packetline::PacketLineRef<'_>, git_packetline::decode::Error>>> {\n+ use bstr::{BStr, ByteSlice};\n+ let bytes: &BStr = self.0.into();\n+ let mut lines = bytes.lines();\n+ let res = lines.next()?;\n+ self.0 = lines.as_bytes();\n+ Some(Ok(Ok(git_packetline::PacketLineRef::Data(res))))\n+ }\n+}\n", "arguments.rs": "@@ -1,333 +0,0 @@\n-use bstr::ByteSlice;\n-use git_transport::Protocol;\n-\n-use crate::fetch;\n-\n-fn arguments_v1(features: impl IntoIterator<Item = &'static str>) -> fetch::Arguments {\n- fetch::Arguments::new(Protocol::V1, features.into_iter().map(|n| (n, None)).collect())\n-}\n-\n-fn arguments_v2(features: impl IntoIterator<Item = &'static str>) -> fetch::Arguments {\n- fetch::Arguments::new(Protocol::V2, features.into_iter().map(|n| (n, None)).collect())\n-}\n-\n-struct Transport<T> {\n- inner: T,\n- stateful: bool,\n-}\n-\n-#[cfg(feature = \"blocking-client\")]\n-mod impls {\n- use std::borrow::Cow;\n-\n- use bstr::BStr;\n- use git_transport::{\n- client,\n- client::{Error, MessageKind, RequestWriter, SetServiceResponse, WriteMode},\n- Protocol, Service,\n- };\n-\n- use crate::fetch::tests::arguments::Transport;\n-\n- impl<T: client::TransportWithoutIO> client::TransportWithoutIO for Transport<T> {\n- fn set_identity(&mut self, identity: client::Account) -> Result<(), Error> {\n- self.inner.set_identity(identity)\n- }\n-\n- fn request(&mut self, write_mode: WriteMode, on_into_read: MessageKind) -> Result<RequestWriter<'_>, Error> {\n- self.inner.request(write_mode, on_into_read)\n- }\n-\n- fn to_url(&self) -> Cow<'_, BStr> {\n- self.inner.to_url()\n- }\n-\n- fn supported_protocol_versions(&self) -> &[Protocol] {\n- self.inner.supported_protocol_versions()\n- }\n-\n- fn connection_persists_across_multiple_requests(&self) -> bool {\n- self.stateful\n- }\n-\n- fn configure(\n- &mut self,\n- config: &dyn std::any::Any,\n- ) -> Result<(), Box<dyn std::error::Error + Send + Sync + 'static>> {\n- self.inner.configure(config)\n- }\n- }\n-\n- impl<T: client::Transport> client::Transport for Transport<T> {\n- fn handshake<'a>(\n- &mut self,\n- service: Service,\n- extra_parameters: &'a [(&'a str, Option<&'a str>)],\n- ) -> Result<SetServiceResponse<'_>, Error> {\n- self.inner.handshake(service, extra_parameters)\n- }\n- }\n-}\n-\n-#[cfg(feature = \"async-client\")]\n-mod impls {\n- use std::borrow::Cow;\n-\n- use async_trait::async_trait;\n- use bstr::BStr;\n- use git_transport::{\n- client,\n- client::{Error, MessageKind, RequestWriter, SetServiceResponse, WriteMode},\n- Protocol, Service,\n- };\n-\n- use crate::fetch::tests::arguments::Transport;\n- impl<T: client::TransportWithoutIO + Send> client::TransportWithoutIO for Transport<T> {\n- fn set_identity(&mut self, identity: client::Account) -> Result<(), Error> {\n- self.inner.set_identity(identity)\n- }\n-\n- fn request(&mut self, write_mode: WriteMode, on_into_read: MessageKind) -> Result<RequestWriter<'_>, Error> {\n- self.inner.request(write_mode, on_into_read)\n- }\n-\n- fn to_url(&self) -> Cow<'_, BStr> {\n- self.inner.to_url()\n- }\n-\n- fn supported_protocol_versions(&self) -> &[Protocol] {\n- self.inner.supported_protocol_versions()\n- }\n-\n- fn connection_persists_across_multiple_requests(&self) -> bool {\n- self.stateful\n- }\n-\n- fn configure(\n- &mut self,\n- config: &dyn std::any::Any,\n- ) -> Result<(), Box<dyn std::error::Error + Send + Sync + 'static>> {\n- self.inner.configure(config)\n- }\n- }\n-\n- #[async_trait(?Send)]\n- impl<T: client::Transport + Send> client::Transport for Transport<T> {\n- async fn handshake<'a>(\n- &mut self,\n- service: Service,\n- extra_parameters: &'a [(&'a str, Option<&'a str>)],\n- ) -> Result<SetServiceResponse<'_>, Error> {\n- self.inner.handshake(service, extra_parameters).await\n- }\n- }\n-}\n-\n-fn transport(\n- out: &mut Vec<u8>,\n- stateful: bool,\n-) -> Transport<git_transport::client::git::Connection<&'static [u8], &mut Vec<u8>>> {\n- Transport {\n- inner: git_transport::client::git::Connection::new(\n- &[],\n- out,\n- Protocol::V1, // does not matter\n- b\"does/not/matter\".as_bstr().to_owned(),\n- None::<(&str, _)>,\n- git_transport::client::git::ConnectMode::Process, // avoid header to be sent\n- ),\n- stateful,\n- }\n-}\n-\n-fn id(hex: &str) -> git_hash::ObjectId {\n- git_hash::ObjectId::from_hex(hex.as_bytes()).expect(\"expect valid hex id\")\n-}\n-\n-mod v1 {\n- use bstr::ByteSlice;\n-\n- use crate::fetch::tests::arguments::{arguments_v1, id, transport};\n-\n- #[maybe_async::test(feature = \"blocking-client\", async(feature = \"async-client\", async_std::test))]\n- async fn haves_and_wants_for_clone() {\n- let mut out = Vec::new();\n- let mut t = transport(&mut out, true);\n- let mut arguments = arguments_v1([\"feature-a\", \"feature-b\"].iter().cloned());\n-\n- arguments.want(id(\"7b333369de1221f9bfbbe03a3a13e9a09bc1c907\"));\n- arguments.want(id(\"ff333369de1221f9bfbbe03a3a13e9a09bc1ffff\"));\n- arguments.send(&mut t, true).await.expect(\"sending to buffer to work\");\n- assert_eq!(\n- out.as_bstr(),\n- b\"0046want 7b333369de1221f9bfbbe03a3a13e9a09bc1c907 feature-a feature-b\n-0032want ff333369de1221f9bfbbe03a3a13e9a09bc1ffff\n-00000009done\n-\"\n- .as_bstr()\n- );\n- }\n-\n- #[maybe_async::test(feature = \"blocking-client\", async(feature = \"async-client\", async_std::test))]\n- async fn haves_and_wants_for_fetch_stateless() {\n- let mut out = Vec::new();\n- let mut t = transport(&mut out, false);\n- let mut arguments = arguments_v1([\"feature-a\", \"shallow\", \"deepen-since\", \"deepen-not\"].iter().copied());\n-\n- arguments.deepen(1);\n- arguments.shallow(id(\"7b333369de1221f9bfbbe03a3a13e9a09bc1c9ff\"));\n- arguments.want(id(\"7b333369de1221f9bfbbe03a3a13e9a09bc1c907\"));\n- arguments.deepen_since(12345);\n- arguments.deepen_not(\"refs/heads/main\".into());\n- arguments.have(id(\"0000000000000000000000000000000000000000\"));\n- arguments.send(&mut t, false).await.expect(\"sending to buffer to work\");\n-\n- arguments.have(id(\"1111111111111111111111111111111111111111\"));\n- arguments.send(&mut t, true).await.expect(\"sending to buffer to work\");\n- assert_eq!(\n- out.as_bstr(),\n- b\"005cwant 7b333369de1221f9bfbbe03a3a13e9a09bc1c907 feature-a shallow deepen-since deepen-not\n-0035shallow 7b333369de1221f9bfbbe03a3a13e9a09bc1c9ff\n-000ddeepen 1\n-0017deepen-since 12345\n-001fdeepen-not refs/heads/main\n-00000032have 0000000000000000000000000000000000000000\n-0000005cwant 7b333369de1221f9bfbbe03a3a13e9a09bc1c907 feature-a shallow deepen-since deepen-not\n-0035shallow 7b333369de1221f9bfbbe03a3a13e9a09bc1c9ff\n-000ddeepen 1\n-0017deepen-since 12345\n-001fdeepen-not refs/heads/main\n-00000032have 1111111111111111111111111111111111111111\n-0009done\n-\"\n- .as_bstr()\n- );\n- }\n-\n- #[maybe_async::test(feature = \"blocking-client\", async(feature = \"async-client\", async_std::test))]\n- async fn haves_and_wants_for_fetch_stateful() {\n- let mut out = Vec::new();\n- let mut t = transport(&mut out, true);\n- let mut arguments = arguments_v1([\"feature-a\", \"shallow\"].iter().copied());\n-\n- arguments.deepen(1);\n- arguments.want(id(\"7b333369de1221f9bfbbe03a3a13e9a09bc1c907\"));\n- arguments.have(id(\"0000000000000000000000000000000000000000\"));\n- arguments.send(&mut t, false).await.expect(\"sending to buffer to work\");\n-\n- arguments.have(id(\"1111111111111111111111111111111111111111\"));\n- arguments.send(&mut t, true).await.expect(\"sending to buffer to work\");\n- assert_eq!(\n- out.as_bstr(),\n- b\"0044want 7b333369de1221f9bfbbe03a3a13e9a09bc1c907 feature-a shallow\n-000ddeepen 1\n-00000032have 0000000000000000000000000000000000000000\n-00000032have 1111111111111111111111111111111111111111\n-0009done\n-\"\n- .as_bstr()\n- );\n- }\n-}\n-\n-mod v2 {\n- use bstr::ByteSlice;\n-\n- use crate::fetch::tests::arguments::{arguments_v2, id, transport};\n-\n- #[maybe_async::test(feature = \"blocking-client\", async(feature = \"async-client\", async_std::test))]\n- async fn haves_and_wants_for_clone_stateful() {\n- let mut out = Vec::new();\n- let mut t = transport(&mut out, true);\n- let mut arguments = arguments_v2([\"feature-a\", \"shallow\"].iter().copied());\n-\n- arguments.deepen(1);\n- arguments.deepen_relative();\n- arguments.want(id(\"7b333369de1221f9bfbbe03a3a13e9a09bc1c907\"));\n- arguments.want(id(\"ff333369de1221f9bfbbe03a3a13e9a09bc1ffff\"));\n- arguments.send(&mut t, true).await.expect(\"sending to buffer to work\");\n- assert_eq!(\n- out.as_bstr(),\n- b\"0012command=fetch\n-0001000ethin-pack\n-0010include-tag\n-000eofs-delta\n-000ddeepen 1\n-0014deepen-relative\n-0032want 7b333369de1221f9bfbbe03a3a13e9a09bc1c907\n-0032want ff333369de1221f9bfbbe03a3a13e9a09bc1ffff\n-0009done\n-0000\"\n- .as_bstr(),\n- \"we filter features/capabilities without value as these apparently shouldn't be listed (remote dies otherwise)\"\n- );\n- }\n-\n- #[maybe_async::test(feature = \"blocking-client\", async(feature = \"async-client\", async_std::test))]\n- async fn haves_and_wants_for_fetch_stateless_and_stateful() {\n- for is_stateful in &[false, true] {\n- let mut out = Vec::new();\n- let mut t = transport(&mut out, *is_stateful);\n- let mut arguments = arguments_v2(Some(\"shallow\"));\n-\n- arguments.deepen(1);\n- arguments.deepen_since(12345);\n- arguments.shallow(id(\"7b333369de1221f9bfbbe03a3a13e9a09bc1c9ff\"));\n- arguments.want(id(\"7b333369de1221f9bfbbe03a3a13e9a09bc1c907\"));\n- arguments.deepen_not(\"refs/heads/main\".into());\n- arguments.have(id(\"0000000000000000000000000000000000000000\"));\n- arguments.send(&mut t, false).await.expect(\"sending to buffer to work\");\n-\n- arguments.have(id(\"1111111111111111111111111111111111111111\"));\n- arguments.send(&mut t, true).await.expect(\"sending to buffer to work\");\n- assert_eq!(\n- out.as_bstr(),\n- b\"0012command=fetch\n-0001000ethin-pack\n-0010include-tag\n-000eofs-delta\n-000ddeepen 1\n-0017deepen-since 12345\n-0035shallow 7b333369de1221f9bfbbe03a3a13e9a09bc1c9ff\n-0032want 7b333369de1221f9bfbbe03a3a13e9a09bc1c907\n-001fdeepen-not refs/heads/main\n-0032have 0000000000000000000000000000000000000000\n-00000012command=fetch\n-0001000ethin-pack\n-0010include-tag\n-000eofs-delta\n-000ddeepen 1\n-0017deepen-since 12345\n-0035shallow 7b333369de1221f9bfbbe03a3a13e9a09bc1c9ff\n-0032want 7b333369de1221f9bfbbe03a3a13e9a09bc1c907\n-001fdeepen-not refs/heads/main\n-0032have 1111111111111111111111111111111111111111\n-0009done\n-0000\"\n- .as_bstr(),\n- \"V2 is stateless by default, so it repeats all but 'haves' in each request\"\n- );\n- }\n- }\n-\n- #[maybe_async::test(feature = \"blocking-client\", async(feature = \"async-client\", async_std::test))]\n- async fn ref_in_want() {\n- let mut out = Vec::new();\n- let mut t = transport(&mut out, false);\n- let mut arguments = arguments_v2([\"ref-in-want\"].iter().copied());\n-\n- arguments.want_ref(b\"refs/heads/main\".as_bstr());\n- arguments.send(&mut t, true).await.expect(\"sending to buffer to work\");\n- assert_eq!(\n- out.as_bstr(),\n- b\"0012command=fetch\n-0001000ethin-pack\n-0010include-tag\n-000eofs-delta\n-001dwant-ref refs/heads/main\n-0009done\n-0000\"\n- .as_bstr()\n- )\n- }\n-}\n", "mod.rs": "@@ -13,17 +13,19 @@ pub mod parse {\n #[error(transparent)]\n Io(#[from] std::io::Error),\n #[error(transparent)]\n+ DecodePacketline(#[from] git_transport::packetline::decode::Error),\n+ #[error(transparent)]\n Id(#[from] git_hash::decode::Error),\n #[error(\"{symref:?} could not be parsed. A symref is expected to look like <NAME>:<target>.\")]\n MalformedSymref { symref: BString },\n #[error(\"{0:?} could not be parsed. A V1 ref line should be '<hex-hash> <path>'.\")]\n- MalformedV1RefLine(String),\n+ MalformedV1RefLine(BString),\n #[error(\n \"{0:?} could not be parsed. A V2 ref line should be '<hex-hash> <path>[ (peeled|symref-target):<value>'.\"\n )]\n- MalformedV2RefLine(String),\n+ MalformedV2RefLine(BString),\n #[error(\"The ref attribute {attribute:?} is unknown. Found in line {line:?}\")]\n- UnkownAttribute { attribute: String, line: String },\n+ UnkownAttribute { attribute: BString, line: BString },\n #[error(\"{message}\")]\n InvariantViolation { message: &'static str },\n }\n@@ -65,3 +67,6 @@ pub use async_io::{from_v1_refs_received_as_part_of_handshake_and_capabilities,\n mod blocking_io;\n #[cfg(feature = \"blocking-client\")]\n pub use blocking_io::{from_v1_refs_received_as_part_of_handshake_and_capabilities, from_v2_refs};\n+\n+#[cfg(test)]\n+mod tests;\n", "async_io.rs": "@@ -1,19 +1,17 @@\n use futures_io::AsyncBufRead;\n-use futures_lite::AsyncBufReadExt;\n+use futures_lite::AsyncReadExt;\n \n use crate::handshake::{refs, refs::parse::Error, Ref};\n+use bstr::ByteSlice;\n \n /// Parse refs from the given input line by line. Protocol V2 is required for this to succeed.\n pub async fn from_v2_refs(in_refs: &mut (dyn AsyncBufRead + Unpin)) -> Result<Vec<Ref>, Error> {\n let mut out_refs = Vec::new();\n- let mut line = String::new();\n- loop {\n- line.clear();\n- let bytes_read = in_refs.read_line(&mut line).await?;\n- if bytes_read == 0 {\n- break;\n- }\n- out_refs.push(refs::shared::parse_v2(&line)?);\n+ let mut buf = Vec::new();\n+\n+ in_refs.read_to_end(&mut buf).await?;\n+ for line in ByteSlice::lines(buf.as_slice()) {\n+ out_refs.push(refs::shared::parse_v2(line.into())?);\n }\n Ok(out_refs)\n }\n@@ -32,14 +30,11 @@ pub async fn from_v1_refs_received_as_part_of_handshake_and_capabilities<'a>(\n ) -> Result<Vec<Ref>, refs::parse::Error> {\n let mut out_refs = refs::shared::from_capabilities(capabilities)?;\n let number_of_possible_symbolic_refs_for_lookup = out_refs.len();\n- let mut line = String::new();\n- loop {\n- line.clear();\n- let bytes_read = in_refs.read_line(&mut line).await?;\n- if bytes_read == 0 {\n- break;\n- }\n- refs::shared::parse_v1(number_of_possible_symbolic_refs_for_lookup, &mut out_refs, &line)?;\n+\n+ let mut buf = Vec::new();\n+ in_refs.read_to_end(&mut buf).await?;\n+ for line in buf.as_slice().lines() {\n+ refs::shared::parse_v1(number_of_possible_symbolic_refs_for_lookup, &mut out_refs, line.into())?;\n }\n Ok(out_refs.into_iter().map(Into::into).collect())\n }\n", "blocking_io.rs": "@@ -1,18 +1,10 @@\n-use std::io;\n-\n use crate::handshake::{refs, refs::parse::Error, Ref};\n \n /// Parse refs from the given input line by line. Protocol V2 is required for this to succeed.\n-pub fn from_v2_refs(in_refs: &mut dyn io::BufRead) -> Result<Vec<Ref>, Error> {\n+pub fn from_v2_refs(in_refs: &mut dyn git_transport::client::ReadlineBufRead) -> Result<Vec<Ref>, Error> {\n let mut out_refs = Vec::new();\n- let mut line = String::new();\n- loop {\n- line.clear();\n- let bytes_read = in_refs.read_line(&mut line)?;\n- if bytes_read == 0 {\n- break;\n- }\n- out_refs.push(refs::shared::parse_v2(&line)?);\n+ while let Some(line) = in_refs.readline().transpose()?.transpose()?.and_then(|l| l.as_bstr()) {\n+ out_refs.push(refs::shared::parse_v2(line)?);\n }\n Ok(out_refs)\n }\n@@ -26,19 +18,14 @@ pub fn from_v2_refs(in_refs: &mut dyn io::BufRead) -> Result<Vec<Ref>, Error> {\n /// Symbolic refs are shoe-horned into server capabilities whereas refs (without symbolic ones) are sent automatically as\n /// part of the handshake. Both symbolic and peeled refs need to be combined to fit into the [`Ref`] type provided here.\n pub fn from_v1_refs_received_as_part_of_handshake_and_capabilities<'a>(\n- in_refs: &mut dyn io::BufRead,\n+ in_refs: &mut dyn git_transport::client::ReadlineBufRead,\n capabilities: impl Iterator<Item = git_transport::client::capabilities::Capability<'a>>,\n ) -> Result<Vec<Ref>, Error> {\n let mut out_refs = refs::shared::from_capabilities(capabilities)?;\n let number_of_possible_symbolic_refs_for_lookup = out_refs.len();\n- let mut line = String::new();\n- loop {\n- line.clear();\n- let bytes_read = in_refs.read_line(&mut line)?;\n- if bytes_read == 0 {\n- break;\n- }\n- refs::shared::parse_v1(number_of_possible_symbolic_refs_for_lookup, &mut out_refs, &line)?;\n+\n+ while let Some(line) = in_refs.readline().transpose()?.transpose()?.and_then(|l| l.as_bstr()) {\n+ refs::shared::parse_v1(number_of_possible_symbolic_refs_for_lookup, &mut out_refs, line)?;\n }\n Ok(out_refs.into_iter().map(Into::into).collect())\n }\n", "shared.rs": "@@ -1,4 +1,4 @@\n-use bstr::{BString, ByteSlice};\n+use bstr::{BStr, BString, ByteSlice};\n \n use crate::handshake::{refs::parse::Error, Ref};\n \n@@ -70,7 +70,7 @@ impl InternalRef {\n _ => None,\n }\n }\n- fn lookup_symbol_has_path(&self, predicate_path: &str) -> bool {\n+ fn lookup_symbol_has_path(&self, predicate_path: &BStr) -> bool {\n matches!(self, InternalRef::SymbolicForLookup { path, .. } if path == predicate_path)\n }\n }\n@@ -109,19 +109,19 @@ pub(crate) fn from_capabilities<'a>(\n pub(in crate::handshake::refs) fn parse_v1(\n num_initial_out_refs: usize,\n out_refs: &mut Vec<InternalRef>,\n- line: &str,\n+ line: &BStr,\n ) -> Result<(), Error> {\n let trimmed = line.trim_end();\n let (hex_hash, path) = trimmed.split_at(\n trimmed\n- .find(' ')\n- .ok_or_else(|| Error::MalformedV1RefLine(trimmed.to_owned()))?,\n+ .find(b\" \")\n+ .ok_or_else(|| Error::MalformedV1RefLine(trimmed.to_owned().into()))?,\n );\n let path = &path[1..];\n if path.is_empty() {\n- return Err(Error::MalformedV1RefLine(trimmed.to_owned()));\n+ return Err(Error::MalformedV1RefLine(trimmed.to_owned().into()));\n }\n- match path.strip_suffix(\"^{}\") {\n+ match path.strip_suffix(b\"^{}\") {\n Some(stripped) => {\n let (previous_path, tag) =\n out_refs\n@@ -146,7 +146,7 @@ pub(in crate::handshake::refs) fn parse_v1(\n match out_refs\n .iter()\n .take(num_initial_out_refs)\n- .position(|r| r.lookup_symbol_has_path(path))\n+ .position(|r| r.lookup_symbol_has_path(path.into()))\n {\n Some(position) => match out_refs.swap_remove(position) {\n InternalRef::SymbolicForLookup { path: _, target } => out_refs.push(InternalRef::Symbolic {\n@@ -166,36 +166,36 @@ pub(in crate::handshake::refs) fn parse_v1(\n Ok(())\n }\n \n-pub(in crate::handshake::refs) fn parse_v2(line: &str) -> Result<Ref, Error> {\n+pub(in crate::handshake::refs) fn parse_v2(line: &BStr) -> Result<Ref, Error> {\n let trimmed = line.trim_end();\n- let mut tokens = trimmed.splitn(3, ' ');\n+ let mut tokens = trimmed.splitn(3, |b| *b == b' ');\n match (tokens.next(), tokens.next()) {\n (Some(hex_hash), Some(path)) => {\n- let id = if hex_hash == \"unborn\" {\n+ let id = if hex_hash == b\"unborn\" {\n None\n } else {\n Some(git_hash::ObjectId::from_hex(hex_hash.as_bytes())?)\n };\n if path.is_empty() {\n- return Err(Error::MalformedV2RefLine(trimmed.to_owned()));\n+ return Err(Error::MalformedV2RefLine(trimmed.to_owned().into()));\n }\n Ok(if let Some(attribute) = tokens.next() {\n- let mut tokens = attribute.splitn(2, ':');\n+ let mut tokens = attribute.splitn(2, |b| *b == b':');\n match (tokens.next(), tokens.next()) {\n (Some(attribute), Some(value)) => {\n if value.is_empty() {\n- return Err(Error::MalformedV2RefLine(trimmed.to_owned()));\n+ return Err(Error::MalformedV2RefLine(trimmed.to_owned().into()));\n }\n match attribute {\n- \"peeled\" => Ref::Peeled {\n+ b\"peeled\" => Ref::Peeled {\n full_ref_name: path.into(),\n object: git_hash::ObjectId::from_hex(value.as_bytes())?,\n tag: id.ok_or(Error::InvariantViolation {\n message: \"got 'unborn' as tag target\",\n })?,\n },\n- \"symref-target\" => match value {\n- \"(null)\" => Ref::Direct {\n+ b\"symref-target\" => match value {\n+ b\"(null)\" => Ref::Direct {\n full_ref_name: path.into(),\n object: id.ok_or(Error::InvariantViolation {\n message: \"got 'unborn' while (null) was a symref target\",\n@@ -215,13 +215,13 @@ pub(in crate::handshake::refs) fn parse_v2(line: &str) -> Result<Ref, Error> {\n },\n _ => {\n return Err(Error::UnkownAttribute {\n- attribute: attribute.to_owned(),\n- line: trimmed.to_owned(),\n+ attribute: attribute.to_owned().into(),\n+ line: trimmed.to_owned().into(),\n })\n }\n }\n }\n- _ => return Err(Error::MalformedV2RefLine(trimmed.to_owned())),\n+ _ => return Err(Error::MalformedV2RefLine(trimmed.to_owned().into())),\n }\n } else {\n Ref::Direct {\n@@ -232,6 +232,6 @@ pub(in crate::handshake::refs) fn parse_v2(line: &str) -> Result<Ref, Error> {\n }\n })\n }\n- _ => Err(Error::MalformedV2RefLine(trimmed.to_owned())),\n+ _ => Err(Error::MalformedV2RefLine(trimmed.to_owned().into())),\n }\n }\n"} |
|
test: add "describeMethods" scope | 761adace7c9680c7e16a0f69096cb3b4f66d7410 | test | https://github.com/pmndrs/react-spring/commit/761adace7c9680c7e16a0f69096cb3b4f66d7410 | add "describeMethods" scope | {"SpringValue.test.ts": "@@ -15,41 +15,7 @@ describe('SpringValue', () => {\n })\n \n describeProps()\n-\n- describe('\"set\" method', () => {\n- it('stops the active animation', async () => {\n- const spring = new SpringValue(0)\n- const promise = spring.start(1)\n-\n- await advanceUntilValue(spring, 0.5)\n- spring.set(2)\n-\n- expect(spring.idle).toBeTruthy()\n- expect(await promise).toMatchObject({\n- finished: false,\n- value: 2,\n- })\n- })\n-\n- describe('when a new value is passed', () => {\n- it('calls the \"onChange\" prop', () => {\n- const onChange = jest.fn()\n- const spring = new SpringValue(0, { onChange })\n- spring.set(1)\n- expect(onChange).toBeCalledWith(1, spring)\n- })\n- it.todo('wraps the \"onChange\" call with \"batchedUpdates\"')\n- })\n-\n- describe('when the current value is passed', () => {\n- it('skips the \"onChange\" call', () => {\n- const onChange = jest.fn()\n- const spring = new SpringValue(0, { onChange })\n- spring.set(0)\n- expect(onChange).not.toBeCalled()\n- })\n- })\n- })\n+ describeMethods()\n \n describeTarget('another SpringValue', from => {\n const node = new SpringValue(from)\n@@ -128,6 +94,43 @@ function describeConfigProp() {\n })\n }\n \n+function describeMethods() {\n+ describe('\"set\" method', () => {\n+ it('stops the active animation', async () => {\n+ const spring = new SpringValue(0)\n+ const promise = spring.start(1)\n+\n+ await advanceUntilValue(spring, 0.5)\n+ spring.set(2)\n+\n+ expect(spring.idle).toBeTruthy()\n+ expect(await promise).toMatchObject({\n+ finished: false,\n+ value: 2,\n+ })\n+ })\n+\n+ describe('when a new value is passed', () => {\n+ it('calls the \"onChange\" prop', () => {\n+ const onChange = jest.fn()\n+ const spring = new SpringValue(0, { onChange })\n+ spring.set(1)\n+ expect(onChange).toBeCalledWith(1, spring)\n+ })\n+ it.todo('wraps the \"onChange\" call with \"batchedUpdates\"')\n+ })\n+\n+ describe('when the current value is passed', () => {\n+ it('skips the \"onChange\" call', () => {\n+ const onChange = jest.fn()\n+ const spring = new SpringValue(0, { onChange })\n+ spring.set(0)\n+ expect(onChange).not.toBeCalled()\n+ })\n+ })\n+ })\n+}\n+\n /** The minimum requirements for testing a dynamic target */\n type OpaqueTarget = {\n node: FrameValue\n"} |
|
test(benchmarks): add `to_pyarrow` benchmark for duckdb | a80cac77f749a03d04c5f37edc152ce15ea0c43e | test | https://github.com/rohankumardubey/ibis/commit/a80cac77f749a03d04c5f37edc152ce15ea0c43e | add `to_pyarrow` benchmark for duckdb | {"test_benchmarks.py": "@@ -753,3 +753,59 @@ def test_parse_many_duckdb_types(benchmark):\n \n types = [\"VARCHAR\", \"INTEGER\", \"DOUBLE\", \"BIGINT\"] * 1000\n benchmark(parse_many, types)\n+\n+\[email protected](scope=\"session\")\n+def sql() -> str:\n+ return \"\"\"\n+ SELECT t1.id as t1_id, x, t2.id as t2_id, y\n+ FROM t1 INNER JOIN t2\n+ ON t1.id = t2.id\n+ \"\"\"\n+\n+\[email protected](scope=\"session\")\n+def ddb(tmp_path_factory):\n+ duckdb = pytest.importorskip(\"duckdb\")\n+\n+ N = 20_000_000\n+\n+ con = duckdb.connect()\n+\n+ path = str(tmp_path_factory.mktemp(\"duckdb\") / \"data.ddb\")\n+ sql = (\n+ lambda var, table, n=N: f\"\"\"\n+ CREATE TABLE {table} AS\n+ SELECT ROW_NUMBER() OVER () AS id, {var}\n+ FROM (\n+ SELECT {var}\n+ FROM RANGE({n}) _ ({var})\n+ ORDER BY RANDOM()\n+ )\n+ \"\"\"\n+ )\n+\n+ with duckdb.connect(path) as con:\n+ con.execute(sql(\"x\", table=\"t1\"))\n+ con.execute(sql(\"y\", table=\"t2\"))\n+ return path\n+\n+\n+def test_duckdb_to_pyarrow(benchmark, sql, ddb) -> None:\n+ # yes, we're benchmarking duckdb here, not ibis\n+ #\n+ # we do this to get a baseline for comparison\n+ duckdb = pytest.importorskip(\"duckdb\")\n+ con = duckdb.connect(ddb, read_only=True)\n+\n+ benchmark(lambda sql: con.sql(sql).to_arrow_table(), sql)\n+\n+\n+def test_ibis_duckdb_to_pyarrow(benchmark, sql, ddb) -> None:\n+ pytest.importorskip(\"duckdb\")\n+ pytest.importorskip(\"duckdb_engine\")\n+\n+ con = ibis.duckdb.connect(ddb, read_only=True)\n+\n+ expr = con.sql(sql)\n+ benchmark(expr.to_pyarrow)\n"} |
|
ci: reenable doctest builds (#9353) | 4769ee5ecb8f10f17d10749a0228b6eb4d094a86 | ci | https://github.com/ibis-project/ibis/commit/4769ee5ecb8f10f17d10749a0228b6eb4d094a86 | reenable doctest builds (#9353) | {"ibis-main.yml": "@@ -128,8 +128,6 @@ jobs:\n run: poetry run python -c 'import shapely.geometry, duckdb'\n \n test_doctests:\n- # FIXME(kszucs): re-enable this build\n- if: false\n name: Doctests\n runs-on: ubuntu-latest\n steps:\n@@ -147,7 +145,7 @@ jobs:\n uses: actions/setup-python@v5\n id: install_python\n with:\n- python-version: \"3.12\"\n+ python-version: \"3.10\"\n \n - name: install poetry\n run: pip install 'poetry==1.8.3'\n"} |
|
fix: `where -> were` typo fix. (#560) | 0eca94d84bd82f2083b41acdb316edce54365f11 | fix | https://github.com/Byron/gitoxide/commit/0eca94d84bd82f2083b41acdb316edce54365f11 | `where -> were` typo fix. (#560) | {"write.rs": "@@ -327,7 +327,7 @@ impl section::Segment {\n if *conventional_count == 1 { \"was\" } else { \"were\" }\n )?;\n if unique_issues.is_empty() {\n- writeln!(out, \" - 0 issues like '(#ID)' where seen in commit messages\")?;\n+ writeln!(out, \" - 0 issues like '(#ID)' were seen in commit messages\")?;\n } else {\n writeln!(\n out,\n", "CHANGELOG.md": "@@ -13,7 +13,7 @@ and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0\n \n - 1 commit contributed to the release.\n - 0 commits were understood as [conventional](https://www.conventionalcommits.org).\n- - 0 issues like '(#ID)' where seen in commit messages\n+ - 0 issues like '(#ID)' were seen in commit messages\n \n ### Commit Details\n \n"} |
|
chore: update dependencies | 2ee33a66feacd52b3fa651a1dfbd32b0412949ab | chore | https://github.com/mikro-orm/mikro-orm/commit/2ee33a66feacd52b3fa651a1dfbd32b0412949ab | update dependencies | {"package.json": "@@ -1,6 +1,6 @@\n {\n \"name\": \"mikro-orm\",\n- \"version\": \"2.0.0-alpha.13\",\n+ \"version\": \"2.0.0-rc\",\n \"description\": \"Simple typescript ORM for node.js based on data-mapper, unit-of-work and identity-map patterns. Supports MongoDB, MySQL and SQLite databases as well as usage with vanilla JS.\",\n \"main\": \"dist/index.js\",\n \"typings\": \"dist/index.d.ts\",\n@@ -76,7 +76,7 @@\n \"fast-deep-equal\": \"^2.0.1\",\n \"globby\": \"^9.1.0\",\n \"node-request-context\": \"^1.0.5\",\n- \"ts-morph\": \"^1.2.0\",\n+ \"ts-morph\": \"^1.3.0\",\n \"typescript\": \"^3.3.3\",\n \"uuid\": \"^3.3.2\"\n },\n@@ -91,15 +91,15 @@\n \"@types/clone\": \"^0.1.30\",\n \"@types/globby\": \"^8.0.0\",\n \"@types/jest\": \"^24.0.9\",\n- \"@types/mongodb\": \"^3.1.19\",\n+ \"@types/mongodb\": \"^3.1.20\",\n \"@types/mysql2\": \"types/mysql2\",\n- \"@types/node\": \"^11.9.6\",\n+ \"@types/node\": \"^11.10.4\",\n \"@types/uuid\": \"^3.4.4\",\n \"codacy-coverage\": \"^3.4.0\",\n \"coveralls\": \"^3.0.3\",\n \"husky\": \"^1.3.1\",\n \"jest\": \"^24.1.0\",\n- \"lint-staged\": \"^8.1.4\",\n+ \"lint-staged\": \"^8.1.5\",\n \"mongodb\": \"^3.1.13\",\n \"mysql2\": \"^1.6.5\",\n \"rimraf\": \"^2.6.3\",\n@@ -107,6 +107,6 @@\n \"sqlite\": \"^3.0.2\",\n \"ts-jest\": \"^24.0.0\",\n \"ts-node\": \"^8.0.2\",\n- \"tslint\": \"^5.13.0\"\n+ \"tslint\": \"^5.13.1\"\n }\n }\n"} |
|
chore: replace `quick-error` with `thiserror`
This increases the compile time of the crate alone if there is no proc-macro
in the dependency tree, but will ever so slightly improve compile times for `gix`
as a whole. | cce96ee1382d3d56d77820a2aba6e2d17b52f91c | chore | https://github.com/Byron/gitoxide/commit/cce96ee1382d3d56d77820a2aba6e2d17b52f91c | replace `quick-error` with `thiserror`
This increases the compile time of the crate alone if there is no proc-macro
in the dependency tree, but will ever so slightly improve compile times for `gix`
as a whole. | {"Cargo.lock": "@@ -1622,9 +1622,9 @@ dependencies = [\n \"once_cell\",\n \"parking_lot 0.12.1\",\n \"prodash\",\n- \"quick-error 2.0.1\",\n \"sha1\",\n \"sha1_smol\",\n+ \"thiserror\",\n \"walkdir\",\n ]\n \n", "Cargo.toml": "@@ -43,7 +43,7 @@ crc32 = [\"crc32fast\"]\n ## and reduced performance is acceptable. **zlib-stock** can be used if dynamic linking of an external zlib library is desired or if cmake is not available.\n ## Note that a competitive Zlib implementation is critical to `gitoxide's` object database performance.\n ## Additional backends are supported, each of which overriding the default Rust backend.\n-zlib = [\"flate2\", \"flate2/rust_backend\", \"quick-error\"]\n+zlib = [\"flate2\", \"flate2/rust_backend\", \"thiserror\"]\n ## Use zlib-ng (libz-ng-sys) with native API (no compat mode) that can co-exist with system libz.\n zlib-ng= [\"flate2/zlib-ng\"]\n ## Use a C-based backend which can compress and decompress significantly faster than the other options.\n@@ -125,7 +125,7 @@ bytes = { version = \"1.0.0\", optional = true }\n \n # zlib module\n flate2 = { version = \"1.0.17\", optional = true, default-features = false }\n-quick-error = { version = \"2.0.0\", optional = true }\n+thiserror = { version = \"1.0.38\", optional = true }\n \n ## If enabled, OnceCell will be made available for interior mutability either in sync or unsync forms.\n once_cell = { version = \"1.13.0\", optional = true }\n", "mod.rs": "@@ -2,24 +2,16 @@ pub use flate2::{Decompress, Status};\n \n /// non-streaming interfaces for decompression\n pub mod inflate {\n- use quick_error::quick_error;\n- quick_error! {\n- /// The error returned by various [Inflate methods][super::Inflate]\n- #[allow(missing_docs)]\n- #[derive(Debug)]\n- pub enum Error {\n- WriteInflated(err: std::io::Error) {\n- display(\"Could not write all bytes when decompressing content\")\n- from()\n- }\n- Inflate(err: flate2::DecompressError) {\n- display(\"Could not decode zip stream, status was '{:?}'\", err)\n- from()\n- }\n- Status(status: flate2::Status) {\n- display(\"The zlib status indicated an error, status was '{:?}'\", status)\n- }\n- }\n+ /// The error returned by various [Inflate methods][super::Inflate]\n+ #[derive(Debug, thiserror::Error)]\n+ #[allow(missing_docs)]\n+ pub enum Error {\n+ #[error(\"Could not write all bytes when decompressing content\")]\n+ WriteInflated(#[from] std::io::Error),\n+ #[error(\"Could not decode zip stream, status was '{0:?}'\")]\n+ Inflate(#[from] flate2::DecompressError),\n+ #[error(\"The zlib status indicated an error, status was '{0:?}'\")]\n+ Status(flate2::Status),\n }\n }\n \n"} |
|
build: updated versions | 0b9cb35626036ccc4d41909c1d6c0e43c5f15c60 | build | https://github.com/tsparticles/tsparticles/commit/0b9cb35626036ccc4d41909c1d6c0e43c5f15c60 | updated versions | {"package.dist.json": "@@ -99,7 +99,7 @@\n \"./package.json\": \"./package.json\"\n },\n \"dependencies\": {\n- \"@tsparticles/engine\": \"^\"\n+ \"@tsparticles/engine\": \"^3.5.0\"\n },\n \"publishConfig\": {\n \"access\": \"public\"\n"} |
|
fix: test failure | 3df896e1c699dfcf6f206081c1f8c2b12b8f1a84 | fix | https://github.com/erg-lang/erg/commit/3df896e1c699dfcf6f206081c1f8c2b12b8f1a84 | test failure | {"test.rs": "@@ -184,7 +184,7 @@ fn test_tolerant_completion() -> Result<(), Box<dyn std::error::Error>> {\n let resp = client.request_completion(uri.raw(), 2, 10, \".\")?;\n if let Some(CompletionResponse::Array(items)) = resp {\n assert!(items.len() >= 10);\n- assert!(items.iter().any(|item| item.label == \"tqdm\"));\n+ assert!(items.iter().any(|item| item.label == \"pi\"));\n Ok(())\n } else {\n Err(format!(\"not items: {resp:?}\").into())\n", "tolerant_completion.er": "@@ -1,6 +1,6 @@\n-tqdm = pyimport \"tqdm\"\n+math = pyimport \"math\"\n \n-f _: tqdm\n+f _: math\n i = 1\n s = \"a\"\n g() = None + i s i\n", "build.rs": "@@ -31,6 +31,12 @@ fn main() -> std::io::Result<()> {\n copy_dir(&erg_path, \"lib\").unwrap_or_else(|_| {\n eprintln!(\"failed to copy the std library to {erg_path}\");\n });\n+ let pkgs_path = path::Path::new(&erg_path).join(\"lib\").join(\"pkgs\");\n+ if !pkgs_path.exists() {\n+ fs::create_dir(&pkgs_path).unwrap_or_else(|_| {\n+ eprintln!(\"failed to create the directory: {}\", pkgs_path.display());\n+ });\n+ }\n Ok(())\n }\n \n"} |
|
build: try fixing publish issues | e14dc55a9fdf5368faaaf8ab2c01012eda8c2a39 | build | https://github.com/tsparticles/tsparticles/commit/e14dc55a9fdf5368faaaf8ab2c01012eda8c2a39 | try fixing publish issues | {"package.json": "@@ -19,6 +19,9 @@\n \"ini\": \"^2.0.0\",\n \"lerna\": \"^4.0.0\"\n },\n+ \"resolutions\": {\n+ \"npm-packlist\": \"1.1.12\"\n+ },\n \"husky\": {\n \"hooks\": {\n \"commit-msg\": \"commitlint -E HUSKY_GIT_PARAMS\"\n"} |
|
fix: make `_stopAnimation` clear keys from the prop cache
This ensures that future `_diff` calls return true, which is required for starting animations. | 0e7d65d367fc3d6af7555e079f6306591972d0d7 | fix | https://github.com/pmndrs/react-spring/commit/0e7d65d367fc3d6af7555e079f6306591972d0d7 | make `_stopAnimation` clear keys from the prop cache
This ensures that future `_diff` calls return true, which is required for starting animations. | {"Controller.ts": "@@ -573,11 +573,20 @@ class Controller<State extends object = any> {\n animatedValues = toArray(animated.getPayload() as any)\n }\n \n+ // Replace the animation config with a lighter object\n this.animations[key] = { key, animated, animatedValues } as any\n+\n+ // Tell the frameloop: \"these animations are done\"\n animatedValues.forEach(v => (v.done = true))\n \n- // Prevent delayed updates to this key.\n+ // Prevent delayed updates to this key\n this.timestamps['to.' + key] = now()\n+ this.timestamps['from.' + key] = now()\n+\n+ // Clear this key from the prop cache, so future diffs are guaranteed\n+ const { to, from } = this.props\n+ if (is.obj(to)) delete to[key]\n+ if (from) delete from[key]\n }\n }\n \n"} |
End of preview. Expand
in Dataset Viewer.
No dataset card yet
- Downloads last month
- 10