-
Updated
Apr 5, 2022 - Makefile
CUDA
CUDA® is a parallel computing platform and programming model developed by NVIDIA for general computing on graphical processing units (GPUs). With CUDA, developers are able to dramatically speed up computing applications by harnessing the power of GPUs.
Here are 3,478 public repositories matching this topic...
-
Updated
Apr 13, 2022 - Shell
-
Updated
Apr 13, 2022 - C++
Problem:
_catboost.pyx in _catboost._set_features_order_data_pd_data_frame()
_catboost.pyx in _catboost.get_cat_factor_bytes_representation()
CatBoostError: Invalid type for cat_feature[non-default value idx=1,feature_idx=336]=2.0 : cat_features must be integer or string, real number values and NaN values should be converted to string.
Could you also print a feature name, not o
-
Updated
Apr 14, 2022 - Cuda
Description
https://numpy.org/doc/stable/reference/generated/numpy.corrcoef.html
https://docs.cupy.dev/en/stable/reference/generated/cupy.corrcoef.html
Seems args are different
Additional Information
dtype argument added in NumPy version 1.20.
-
Updated
Mar 8, 2022 - Python
-
Updated
Apr 5, 2022 - Go
Recently in Morpheus we encountered a bug where get_current_device_resource was undefined in a place we were not explicitly using it. Most public-facing libcudf APIs provide a memory_resource* as a default argument by calling get_current_device_resource, defined in rmm/mr/per_device_resource.hpp, however in some places this header is not included which requires the caller of libcudf APIs t
请问可以直接training tmfile出来吗? 因为tengine-convert-tool covert 会有error
tengine-lite library version: 1.4-dev
Get input tensor failed

或是有例子能training出下面tmfile 呢?
':
/usr/local/cuda/bin/../targets/x86_64-linux/include/thrust/detail/complex/catrigf.h:170:36: error: implic
Current implementation of join can be improved by performing the operation in a single call to the backend kernel instead of multiple calls.
This is a fairly easy kernel and may be a good issue for someone getting to know CUDA/ArrayFire internals. Ping me if you want additional info.
你好,请问怎么装载 ONNX 模型,目前只看到 Oneflow->ONNX 工具,没有找到 ONNX->Oneflow 工具。
-
Updated
Apr 14, 2022 - C++
Describe the bug
We should raise better error messages in the scenario when users pass stuff like pandas.Series/list etc to the vectorizer.
Steps/Code to reproduce bug
import cudf
import pandas
from cuml.feature_extraction.text import TfidfVectorizer
vec = TfidfVectorizer()
text_s = pandas.Series(["apple", "is", "great"])
vec.fit_transform(text_s)Hey everyone!
mapd-core-cpu is already available on conda-forge (https://anaconda.org/conda-forge/omniscidb-cpu)
now we should add some instructions on the documentation.
at this moment it is available for linux and osx.
some additional information about the configuration:
- for now, always install
omniscidb-cpuinside a conda environment (also it is a good practice), eg:
-
Updated
Apr 13, 2022 - C++
-
Updated
Apr 14, 2022 - Cuda
-
Updated
Apr 7, 2022 - C
In order to test manually altered IR, it would be nice to have a --skip-compilation flag for futhark test, just like we do for futhark bench.
-
Updated
Apr 8, 2022 - C++
-
Updated
Sep 11, 2018 - C++
-
Updated
Apr 14, 2022 - Python
-
Updated
Jan 12, 2022 - Python
Created by Nvidia
Released June 23, 2007
- Website
- developer.nvidia.com/cuda-zone
- Wikipedia
- Wikipedia
I see comments suggesting adding this to understand how loops are being handled by numba, and in the their own FAQ (https://numba.pydata.org/numba-doc/latest/user/faq.html)
You would then create your njit function and run it, and I believe the idea is that it prints debug information about whether