Skip to main content
Filter by
Sorted by
Tagged with
0 votes
0 answers
23 views

The code running on the dask worker calls asyncio.run() and proceeds to exectue a series of async calls (on the worker running event_loop) that gather data, and then run a small computation. This ...
Dirich's user avatar
  • 442
0 votes
1 answer
247 views

I've set up a kubernetes cluster with GKE and installed the dask-kubernetes-operator. When i try to start the cluster like this cluster: KubeCluster = KubeCluster(custom_cluster_spec="cluster....
TuemmlerTanne11's user avatar
0 votes
1 answer
78 views

I am using dask distributed package to create a EC2/ECS cluster, I want to read the ML models within the workers, something like def read_model(model_path): model = pickle.load(model_path) ...
himanshu jain's user avatar
0 votes
1 answer
306 views

I am trying to deploy the latest version of 'daskhub' in a GKE cluster (v1.21.12-gke.1700), but getting the below error with 'traefik' helm upgrade --wait --install --dry-run --debug --render-...
SKS's user avatar
  • 171
0 votes
1 answer
229 views

I've deployed dask helm chart on gke, can access to the cluster with distributed.Client. Now I need to connect to dask cluster with dask_kubernetes.HelmCluster, but it raises this exception. Code ...
Vin B.'s user avatar
  • 41
2 votes
1 answer
442 views

We have been running DASK clusters on Kubernetes for some time. Up to now, we have been using CPUs for processing and, of course, system memory for storing our Dataframe of around 1,5 TB (per DASK ...
honor's user avatar
  • 8,238
0 votes
1 answer
444 views

I wanted to know if there's a function in dask.distributed that removes the files uploaded to the cluster using the client.upload_file()? Basically, the opposite of the upload_file() function. best ...
Rehan Rajput's user avatar
0 votes
1 answer
613 views

I'm trying to get up and running with dask on kubernetes. Below is effectively a hello world for dask-kubernetes, but I'm stuck on the error below. main.py: import os from dask_kubernetes import ...
Cliff's user avatar
  • 175
2 votes
1 answer
683 views

I have a simple embarrassingly parallel program that I am successfully running locally on Dask. Yay! Now I want to move it to a cluster and crank up the problem size. In this case, I am using GCP. I ...
adonoho's user avatar
  • 4,329
2 votes
1 answer
458 views

Why does kubectl run dask --image daskdev/dask fail? # starting the container with docker to make sure it basically works ➜ ~ docker run --rm -it --entrypoint bash daskdev/dask:latest (base) root@...
Raffael's user avatar
  • 20.1k
3 votes
0 answers
189 views

We deployed Dask Gateway on Kubernetes on Google Cloud Platform. We are currently using an internal TCP load balancer to expose the traefik proxy for security purposes. Our users are able to create a ...
Riley Hun's user avatar
  • 2,831
2 votes
1 answer
2k views

I'm a bit confused on how exactly I'm supposed to connect to a deployed Dask cluster created via Dask-helm chart from an external service. I deployed a Dask cluster as explained here After a ...
maverick's user avatar
  • 2,385
0 votes
1 answer
322 views

I'm using a vanilla Dask-Kubernetes setup with two workers and one scheduler to iterate over the lines of some JSON file (and apply some functions which don't appear here for simplicity). I see only ...
deeplook's user avatar
2 votes
0 answers
176 views

I am trying to deploy Dask Gateway integrated with JupyterHub which is the reason I decided to give DaskHub Chart a try. After following the instructions on https://docs.dask.org/en/latest/setup/...
Ramon Ramirez-Linan's user avatar
1 vote
0 answers
127 views

I am running Dask on an eight-node Kubernetes cluster with my manifest specifying one scheduler replica and eight worker replicas. My code is processing 80 files of about equal size, and I wanted to ...
user655321's user avatar
  • 1,792
0 votes
1 answer
584 views

Using the code bellow it is possible to create a dask kubernetes cluster in azure aks. It uses a remote scheduler (dask.config.set({"kubernetes.scheduler-service-type": "LoadBalancer&...
Nuno Silva's user avatar
2 votes
1 answer
147 views

I have one kubernetes cluster in region us-east1 where dask-scheduler was started and i want to start another cluster in region us-west1 where would like run dask-workers. As I understand connection ...
Habibutsu's user avatar
  • 632
1 vote
1 answer
2k views

I have some long-running code (~5-10 minute processing) that I'm trying to run as a Dask Future. It's a series of several discrete steps that I can either run as one function: result : Future = ...
user655321's user avatar
  • 1,792
0 votes
1 answer
495 views

I created my own very slightly modified Dockerfile based on the dask-docker Dockerfile that installs adlfs and copies one of my custom libraries into the container in order to make it available to all ...
user655321's user avatar
  • 1,792
0 votes
1 answer
113 views

I have a dask expression as follows where I'm trying to run a sqlalchemy query in a distributed way. However, it references a .pem key file that's inputted in the connect_args parameter. How do I ...
Riley Hun's user avatar
  • 2,831
1 vote
1 answer
336 views

I am learning dask-kubernetes on GKE. I stumbled across an asyncio error (ERROR:asyncio:Task exception was never retrieved). See steps below for the issue. However, additional guidance on using ...
Ray Bell's user avatar
  • 1,628
0 votes
1 answer
372 views

I wrote a program with dask and delayed and now I want to run it on several machines in the cloud. But there's one thing I don't understand - how does dask run the code on multiple machines in the ...
Alexis Travers's user avatar
1 vote
1 answer
556 views

I used the following code to create a cluster from dask_kubernetes import KubeCluster cluster = KubeCluster.from_yaml('worker.yaml') cluster.adapt(minimum=1, maximum=10) with the following yaml code (...
Paeng G's user avatar
  • 11
3 votes
0 answers
967 views

Our company is currently leveraging prefect.io for data workflows (ELT, report generation, ML, etc). We have just started adding the ability to do parallel task execution, which is powered by Dask. ...
braunk's user avatar
  • 31
0 votes
1 answer
258 views

Noob here. I want to have a Dask install with a worker pool that can grow and shrink based on current demands. I followed the instructions in zero to jupyterhub to install on GKE, and then went ...
Patrick Mineault's user avatar