![]() ![]() ![]() Import numpy as np import pandas as pd import csv import math, itertools import os import subprocess import matplotlib.pyplot as plt from multiprocessing import Pool, Manager from collections import Counter from stop_words import get_stop_words import natsort from natsort import natsorted from scipy import spatial from scipy.stats import pearsonr, spearmanr from sklearn.svm import SVR, LinearSVR from sklearn.externals import joblib from sklearn.model_selection import cross_val_predict, cross_val_score from sklearn.model_selection import train_test_split, StratifiedKFold from trics import r2_score, f1_score from trics import classification_report, precision_recall_fscore_support from sklearn.feature_extraction.text import CountVectorizer from imblearn.over_sampling import RandomOverSampler from keras.preprocessing import sequence from _utils import to_categorical from keras.models import Sequential, load_model, model_from_json from keras.layers import Dense, Activation, Embedding, Bidirectional, Dropout, LSTM from keras.regularizers import l2 import keras.backend as K from theano import function import warningsġ9 //If true, when there is a gpu malloc or free error, we print the size of allocated memory on the device.Ģ2 //If true, we fill with NAN allocated device memory.Ģ5 //If true, we print out when we free a device pointer, uninitialize aĢ6 //CudaNdarray, or allocate a device pointerĢ9 //If true, we do error checking at the start of functions, to make sure thereģ0 //is not a pre-existing error when the function is called.ģ1 //You probably need to set the environment variableģ2 //CUDA_LAUNCH_BLOCKING=1, and/or modify the CNDA_THREAD_SYNCģ3 //preprocessor macro in cuda_ndarray.cuhĤ8 CudaNdarray_Dimshuffle(PyObject* _unused, PyObject* args) Ĥ9 static PyObject *CudaNdarray_get_shape(CudaNdarray *self, void *closure) ĥ4 * In the test program I'm using, the _outstanding_mallocs decreases with every call.ĥ5 * This suggests there are more free() calls being made than alloc(), but I can't figure out why.Ħ9 table_struct _alloc_size_table ħ4 return device_malloc(size, VERBOSE_DEVICE_MALLOC) ħ9 int initCnmem(int card_number_provided, int card_nb, size_t mem) Ĥ77 decl_k_elemwise_unary_rowmajor(k_elemwise_unary_rowmajor_exp, unary_exp)Ĥ83 //DON'T use directly(if their is other CudaNdarray that point to it, it will cause problem)! use Py_DECREF() insteadĤ85 CudaNdarray_dealloc(CudaNdarray* self)Ĥ89 printf("WARNING:CudaNdarray_dealloc called when there is still active reference to it. From 0.3.1 onward, cuda capability 5.0 will not be included in the pre-packaged binary release (so all torch.cuda related stuff will not work). ASUS GTX 960 DirectCU III OC: 1228 MHz: 1291 MHz: 1753 MHz: ASUS GTX 960 Mini OC: 1190 MHz: 1253 MHz: 1753 MHz: 170 mm/6.7 inches: ASUS STRIX GTX 960 DirectCU II OC: 1228 MHz: 1291 MHz: 1800 MHz: 215 mm/8.5 inches: ASUS STRIX GTX 960 DirectCU II OC 4 GB: 1228 MHz: 1291 MHz: 1753 MHz: 4 GB, 215 mm/8.5 inches: ASUS STRIX GTX 960 DirectCU II OC SP: 1253 MHz: 1317 MHz: 1800 MHz: 215 mm/8. ![]()
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |