Skip to content
#

dask

Here are 264 public repositories matching this topic...

xarray
rabernat
rabernat commented Sep 24, 2021

Noting down a comment by @DanJonesOcean on Twitter: https://twitter.com/DanJonesOcean/status/1441392596362874882

In general, having more examples on each xarray page (like the one below) would be good. Then they would come up quickly in function searches:

http://xarray.pydata.org/en/stable/generated/xarray.Dataset.merge.html#xarray.Dataset.merge

Our API docs are generated by the func

jacobtomlinson
jacobtomlinson commented Jan 14, 2021

What happened:

When creating a LocalCluster object the comm is started on a random high port, even if there are no other clusters running.

What you expected to happen:

Should use port 8786.

Minimal Complete Verifiable Example:

$ conda create -n dask-lc-test -c conda-forge -y python=3.8 ipython dask distributed
$ conda activate dask-lc-test

The `d

djhoese
djhoese commented Feb 22, 2021

Code Sample, a minimal, complete, and verifiable piece of code

from pyresample.boundary import Boundary
b = Boundary(my_lons, my_lats)
print(b.contour_poly.area())

Problem description

The above code doesn't fail if the provided lons/lats are 2D (not sure on 3D+), but the class and all functions/utilities underneath it assume 1D arrays. The end results are incor

nils-braun
nils-braun commented Feb 5, 2021

The ML implementation is still a bit experimental - we can improve on this:

  • SHOW MODELS and DESCRIBE MODEL
  • Hyperparameter optimizations, AutoML-like behaviour
  • @romainr brought up the idea of exporting models (#191, still missing: onnx - see discussion in the PR by @rajagurunath)
  • and some more showcases and examples
lesteve
lesteve commented May 19, 2020
from dask_jobqueue import SLURMCluster 
cluster = SLURMCluster(cores=1, memory='1GB') 
print(cluster.job_script()) 
#!/usr/bin/env bash

#SBATCH -J dask-worker
#SBATCH -n 1
#SBATCH --cpus-per-task=1
#SBATCH --mem=954M
#SBATCH -t 00:30:00

/home/lesteve/miniconda3/bin/python -m distributed.cli.dask_worker tcp://192.168.0.11:44065 --nthreads 1 --memory-limit 1000.00MB -
NeroCorleone
NeroCorleone commented Aug 11, 2020

Problem description

Reading a dataset with eager's read functionality raises a ValueError when providing columns.

Example code (ideally copy-pastable)

import pandas as pd

from tempfile import TemporaryDirectory
from functools import partial
from storefact import get_store_from_url

from kartothek.io.eager import store_dataframes_as_dataset, read_dataset_as_data
climpred
JacksonMaxfield
JacksonMaxfield commented Apr 6, 2021

In determining the correct reader for the file provided we currently have two options (as of #224).

  1. Providing reader param to AICSImage (i.e. img = AICSImage("s3://some-file.ext", reader=readers.lif_reader.LifReader)
  2. Not providing a reader, and AICSImage looping over all SUPPORTED_READERS.

Option 1 is the fastest + safest method for loading a file into AICSImage (without using

zblz
zblz commented Aug 15, 2017

Currently all of the metrics computed are independent of a target variable or column, but if lens.summarise took the name of a column as the target variable, the output of some metrics could be more interpretable even if the target variable is not used in any kind of predictive modelling.

A good example of this could be PCA (see #14), which could plot the different categories of the target va

RichardScottOZ
RichardScottOZ commented Mar 25, 2021

Without thinking I put resampling="bilinear" and got an error when I called .compute()

Traceback (most recent call last):
  File "carajas.py", line 92, in <module>
    band_medianNP = band_median.compute()
  File "/home/ubuntu/anaconda3/envs/richard/lib/python3.8/site-packages/xarray/core/dataarray.py", line 899, in compute
    return new.load(**kwargs)
  File "/home/ubuntu/anaco

Improve this page

Add a description, image, and links to the dask topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the dask topic, visit your repo's landing page and select "manage topics."

Learn more