data-lake
Here are 158 public repositories matching this topic...
Code of Conduct
- I agree to follow this project's Code of Conduct
Search before asking
- I have searched in the issues and found no similar issues.
Describe the subtask
TPCDSTable table should Implement SupportsMetadataColumns to expose meta columns for tables.
-
Updated
Apr 22, 2022 - Java
-
Updated
Mar 9, 2020 - Python
-
Updated
Mar 5, 2020 - Python
-
Updated
Apr 12, 2022 - Java
-
Updated
Nov 5, 2020 - Jupyter Notebook
-
Updated
Apr 15, 2022 - Python
-
Updated
Apr 6, 2022 - JavaScript
When an Item in the queue is added with incorrect type for the corresponding Data Mapper, the Job fails during planning, without any information about which datamapper/queue item id is involved.
Let's take a Data Mapper with a identifier of type int for instance. If we add foo to the deletion queue, the find will fail with a log like this:
{
"EventData": {
"Error": "ValueError
-
Updated
Mar 15, 2022 - Jupyter Notebook
Is your feature request related to a problem? Please describe.
Executing all tests takes already about 30mins. We should try to optimize that.
Describe the solution you'd like
Much time is taken by preparing input data by writing test data to DataObjects (Csv or Hive). This could be significantly reduced by creating a custom DataObject where a DataFrame can be set as input data, which
-
Updated
May 13, 2020 - HTML
-
Updated
Jul 12, 2021 - Scala
-
Updated
Jul 26, 2020 - Scala
-
Updated
Aug 5, 2020
-
Updated
Sep 26, 2020 - Python
-
Updated
Mar 29, 2022 - PLpgSQL
If ZeeQS imports records from a Zeebe cluster with multiple partitions then it can happen that variable updates, element instance transitions, and message correlations are not persisted.
The problem is caused by the importer. It uses the record position as ID for the entities. But the positions are not unique across multiple partitions.
Related issue: https://github.com/camunda-community-hub
-
Updated
Apr 22, 2022 - Go
-
Updated
Apr 20, 2021 - Python
-
Updated
Jan 15, 2020 - Shell
-
Updated
Mar 29, 2022 - Python
-
Updated
Mar 4, 2021 - Jupyter Notebook
Improve this page
Add a description, image, and links to the data-lake topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the data-lake topic, visit your repo's landing page and select "manage topics."
On more advanced versions of LakeFS (probably > = v1.0.0), we would like to remove the logic that tries to fill the generation field in DB when loading old dumps. It means we will no longer support loading dump that made with a version lower than v0.61.0.