dask
Here are 259 public repositories matching this topic...
Is your feature request related to a problem? Please describe.
Sometimes you want to check that data values are present in another array, but only up to a certain tolerance.
Describe the solution you'd like
da.isin(test_values, tolerance=1e-6), where the tolerance argument is optional.
Not sure what the implementation should be but there are two vectorized [suggestions here](http
Support Series.median()
The stumpy.snippets feature is now completed in #283 which follows this work:
We have a rough notebook t
-
Updated
Sep 8, 2021 - Python
What happened:
When creating a LocalCluster object the comm is started on a random high port, even if there are no other clusters running.
What you expected to happen:
Should use port 8786.
Minimal Complete Verifiable Example:
$ conda create -n dask-lc-test -c conda-forge -y python=3.8 ipython dask distributed
$ conda activate dask-lc-testThe `d
-
Updated
Sep 29, 2021 - Python
-
Updated
Sep 19, 2021 - Python
Describe the bug
According to the multiscene documentation, the property all_same_area does:
Determine if all contained Scenes have the same ‘area’.
However, I have created a multiscene where all scenes have the same area (they just differ between datasets), yet the property returns Fa
-
Updated
Sep 24, 2021 - Python
-
Updated
Aug 9, 2021 - Python
-
Updated
Jul 21, 2021 - Python
Code Sample, a minimal, complete, and verifiable piece of code
from pyresample.boundary import Boundary
b = Boundary(my_lons, my_lats)
print(b.contour_poly.area())Problem description
The above code doesn't fail if the provided lons/lats are 2D (not sure on 3D+), but the class and all functions/utilities underneath it assume 1D arrays. The end results are incor
-
Updated
Sep 19, 2021 - Python
The ML implementation is still a bit experimental - we can improve on this:
-
SHOW MODELSandDESCRIBE MODEL - Hyperparameter optimizations, AutoML-like behaviour
- @romainr brought up the idea of exporting models (#191, still missing: onnx - see discussion in the PR by @rajagurunath)
- and some more showcases and examples
from dask_jobqueue import SLURMCluster
cluster = SLURMCluster(cores=1, memory='1GB')
print(cluster.job_script()) #!/usr/bin/env bash
#SBATCH -J dask-worker
#SBATCH -n 1
#SBATCH --cpus-per-task=1
#SBATCH --mem=954M
#SBATCH -t 00:30:00
/home/lesteve/miniconda3/bin/python -m distributed.cli.dask_worker tcp://192.168.0.11:44065 --nthreads 1 --memory-limit 1000.00MB -
Problem description
Reading a dataset with eager's read functionality raises a ValueError when providing columns.
Example code (ideally copy-pastable)
import pandas as pd
from tempfile import TemporaryDirectory
from functools import partial
from storefact import get_store_from_url
from kartothek.io.eager import store_dataframes_as_dataset, read_dataset_as_dataIs your feature request related to a problem? Please describe.
asv performance now depends on local machine performance
Describe the solution you'd like
Independent of my laptop
https://labs.quansight.org/blog/2021/08/github-actions-benchmarks/
Describe alternatives you've considered
keep as is
-
Updated
Sep 29, 2021 - Vue
Currently all of the metrics computed are independent of a target variable or column, but if lens.summarise took the name of a column as the target variable, the output of some metrics could be more interpretable even if the target variable is not used in any kind of predictive modelling.
A good example of this could be PCA (see #14), which could plot the different categories of the target va
In determining the correct reader for the file provided we currently have two options (as of #224).
- Providing
readerparam toAICSImage(i.e.img = AICSImage("s3://some-file.ext", reader=readers.lif_reader.LifReader) - Not providing a reader, and AICSImage looping over all
SUPPORTED_READERS.
Option 1 is the fastest + safest method for loading a file into AICSImage (without using
Passing resampling
Without thinking I put resampling="bilinear" and got an error when I called .compute()
Traceback (most recent call last):
File "carajas.py", line 92, in <module>
band_medianNP = band_median.compute()
File "/home/ubuntu/anaconda3/envs/richard/lib/python3.8/site-packages/xarray/core/dataarray.py", line 899, in compute
return new.load(**kwargs)
File "/home/ubuntu/anaco-
Updated
Sep 30, 2021 - JavaScript
-
Updated
Apr 25, 2018 - Python
-
Updated
Jul 3, 2018 - Python
Improve this page
Add a description, image, and links to the dask topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the dask topic, visit your repo's landing page and select "manage topics."
I'm hoping to get an idea of the memory size of a dask.dataframe once I call .compute() on it
My current approach is