gpu
Here are 2,066 public repositories matching this topic...
-
Updated
Feb 24, 2021 - Jupyter Notebook
-
Updated
Jan 4, 2021 - Makefile
At this moment relu_layer op doesn't allow threshold configuration, and legacy RELU op allows that.
We should add configuration option to relu_layer.
-
Updated
Feb 18, 2021 - JavaScript
-
Updated
Mar 2, 2021 - Python
-
Updated
Feb 5, 2021 - Python
Problem: the approximate method can still be slow for many trees
catboost version: master
Operating System: ubuntu 18.04
CPU: i9
GPU: RTX2080
Would be good to be able to specify how many trees to use for shapley. The model.predict and prediction_type versions allow this. lgbm/xgb allow this.
-
Updated
Feb 17, 2021 - Python
-
Updated
Mar 2, 2021 - Jupyter Notebook
-
Updated
Feb 26, 2021 - Python
As seen in openwall/john#4530 (comment):
Benchmarking: sspr-opencl, NetIQ SSPR / Adobe AEM [MD5/SHA1/SHA2 OpenCL]... Warning: binary() returned misaligned pointer
DONE
This is because opencl_sspr_fmt_plug.c wrongly has:
#define BINARY_ALIGN MEM_ALIGN_WORDwhereas the code only guarantees alignment appropriate for
Hi ,
I have tried out both loss.backward() and model_engine.backward(loss) for my code. There are several subtle differences that I have observed , for one retain_graph = True does not work for model_engine.backward(loss) . This is creating a problem since buffers are not being retained every time I run the code for some reason.
Please look into this if you could.
-
Updated
Mar 1, 2021 - C++
-
Updated
Mar 2, 2021 - C++
-
Updated
Apr 24, 2020 - Jsonnet
-
Updated
Jun 13, 2020 - HTML
Currently, aggregation APIs (groupby, reductions, rolling, etc.) are scattered around in multiple files and there are inconsistencies between the directory structures in cpp/include/, cpp/src/, cpp/tests/, and cpp/benchmarks/. For example:
cpp/include/:
- include/cudf/aggregation.hpp
- include/cudf/groupby.hpp
- include/cudf/rolling.hpp
- ....
cpp/src/:
- src/aggregati
Current implementation of join can be improved by performing the operation in a single call to the backend kernel instead of multiple calls.
This is a fairly easy kernel and may be a good issue for someone getting to know CUDA/ArrayFire internals. Ping me if you want additional info.
We would like to forward a particular 'key' column which is part of the features to appear alongside the predictions - this is to be able to identify to which set of features a particular prediction belongs to. Here is an example of predictions output using the tensorflow.contrib.estimator.multi_class_head:
{"classes": ["0", "1", "2", "3", "4", "5", "6", "7", "8", "9"],
"scores": [0.068196
PR NVIDIA/cub#218 fixes this CUB's radix sort. We should:
- Check whether Thrust's other backends handle this case correctly.
- Provide a guarantee of this in the stable_sort documentation.
- Add regression tests to enforce this on all backends.
-
Updated
Dec 17, 2020 - CMake
-
Updated
Mar 1, 2021 - C++
-
Updated
Mar 1, 2021 - Jupyter Notebook
Improve this page
Add a description, image, and links to the gpu topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the gpu topic, visit your repo's landing page and select "manage topics."
According to
https://github.com/pytorch/pytorch/blob/64847c7f0b3559c6edc40f001619b80c7dc68ef7/c10/util/Exception.h#L479
all usages of AT_ERROR("msg") should be replace with
TORCH_CHECK(false, "msg")There are currently 29 instances of AT_ERROR being used in c10 codebase: