Skip to content
Projects
Groups
Snippets
Help
Loading...
Sign in
Toggle navigation
D
dlib
Project
Project
Details
Activity
Cycle Analytics
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
Issues
0
Issues
0
List
Board
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Charts
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Charts
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
钟尚武
dlib
Commits
fa0e559f
Commit
fa0e559f
authored
Oct 17, 2015
by
Davis King
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
cleaned up the cuda bindings a bit
parent
f9c918eb
Show whitespace changes
Inline
Side-by-side
Showing
6 changed files
with
369 additions
and
40 deletions
+369
-40
CMakeLists.txt
dlib/CMakeLists.txt
+5
-1
cublas_api.cpp
dlib/dnn/cublas_api.cpp
+56
-0
cublas_api.h
dlib/dnn/cublas_api.h
+6
-12
cudnn_api.cpp
dlib/dnn/cudnn_api.cpp
+294
-0
cudnn_api.h
dlib/dnn/cudnn_api.h
+5
-23
tensor.h
dlib/dnn/tensor.h
+3
-4
No files found.
dlib/CMakeLists.txt
View file @
fa0e559f
...
@@ -434,7 +434,11 @@ if (NOT TARGET dlib)
...
@@ -434,7 +434,11 @@ if (NOT TARGET dlib)
if
(
CUDA_FOUND AND cudnn AND cudnn_include AND COMPILER_CAN_DO_CPP_11
)
if
(
CUDA_FOUND AND cudnn AND cudnn_include AND COMPILER_CAN_DO_CPP_11
)
message
(
STATUS
"Found cuDNN: "
${
cudnn
}
)
message
(
STATUS
"Found cuDNN: "
${
cudnn
}
)
set
(
source_files
${
source_files
}
dnn/cuda.cu
)
set
(
source_files
${
source_files
}
dnn/cuda.cu
dnn/cudnn_api.cpp
dnn/cublas_api.cpp
)
set
(
dlib_needed_libraries
${
dlib_needed_libraries
}
${
CUDA_CUBLAS_LIBRARIES
}
${
cudnn
}
)
set
(
dlib_needed_libraries
${
dlib_needed_libraries
}
${
CUDA_CUBLAS_LIBRARIES
}
${
cudnn
}
)
include_directories
(
${
cudnn_include
}
)
include_directories
(
${
cudnn_include
}
)
else
()
else
()
...
...
dlib/dnn/cublas_api.cpp
0 → 100644
View file @
fa0e559f
// Copyright (C) 2015 Davis E. King (davis@dlib.net)
// License: Boost Software License See LICENSE.txt for the full license.
#ifndef DLIB_DNN_CuBLAS_CPP_
#define DLIB_DNN_CuBLAS_CPP_
#ifdef DLIB_USE_CUDA
#include "cublas_api.h"
#include <cublas_v2.h>
namespace
dlib
{
namespace
cuda
{
// -----------------------------------------------------------------------------------
cublas_context
::
cublas_context
()
{
// TODO
}
cublas_context
::
~
cublas_context
()
{
// TODO
}
// -----------------------------------------------------------------------------------
void
gemm
(
cublas_context
&
context
,
float
beta
,
tensor
&
dest
,
float
alpha
,
const
tensor
&
lhs
,
bool
trans_lhs
,
const
tensor
&
rhs
,
bool
trans_rhs
)
{
}
// ------------------------------------------------------------------------------------
}
}
#endif // DLIB_USE_CUDA
#endif // DLIB_DNN_CuBLAS_CPP_
dlib/dnn/cublas_api.h
View file @
fa0e559f
...
@@ -6,6 +6,7 @@
...
@@ -6,6 +6,7 @@
#ifdef DLIB_USE_CUDA
#ifdef DLIB_USE_CUDA
#include "tensor.h"
#include "tensor.h"
#include "../error.h"
namespace
dlib
namespace
dlib
{
{
...
@@ -28,18 +29,11 @@ namespace dlib
...
@@ -28,18 +29,11 @@ namespace dlib
cublas_context
(
const
cublas_context
&
)
=
delete
;
cublas_context
(
const
cublas_context
&
)
=
delete
;
cublas_context
&
operator
=
(
const
cublas_context
&
)
=
delete
;
cublas_context
&
operator
=
(
const
cublas_context
&
)
=
delete
;
// but is movable
// but is movable
cublas_context
(
c
onst
c
ublas_context
&&
)
=
default
;
cublas_context
(
cublas_context
&&
)
=
default
;
cublas_context
&
operator
=
(
c
onst
c
ublas_context
&&
)
=
default
;
cublas_context
&
operator
=
(
cublas_context
&&
)
=
default
;
cublas_context
()
cublas_context
();
{
~
cublas_context
();
// TODO
}
~
cublas_context
()
{
// TODO
}
const
void
*
get_handle
(
const
void
*
get_handle
(
)
const
{
return
handle
;
}
)
const
{
return
handle
;
}
...
@@ -56,7 +50,7 @@ namespace dlib
...
@@ -56,7 +50,7 @@ namespace dlib
float
beta
,
float
beta
,
tensor
&
dest
,
tensor
&
dest
,
float
alpha
,
float
alpha
,
const
tensor
&
lhs
const
tensor
&
lhs
,
bool
trans_lhs
,
bool
trans_lhs
,
const
tensor
&
rhs
,
const
tensor
&
rhs
,
bool
trans_rhs
bool
trans_rhs
...
...
dlib/dnn/cudnn_api.cpp
0 → 100644
View file @
fa0e559f
// Copyright (C) 2015 Davis E. King (davis@dlib.net)
// License: Boost Software License See LICENSE.txt for the full license.
#ifndef DLIB_DNN_CuDNN_CPP_
#define DLIB_DNN_CuDNN_CPP_
#ifdef DLIB_USE_CUDA
#include "cudnn_api.h"
#include "tensor.h"
#include <cudnn.h>
namespace
dlib
{
namespace
cuda
{
// TODO, make into a macro that prints more information like the line number, etc.
static
void
check
(
cudnnStatus_t
s
)
{
switch
(
s
)
{
case
CUDNN_STATUS_SUCCESS
:
return
;
case
CUDNN_STATUS_NOT_INITIALIZED
:
throw
cudnn_error
(
"CUDA Runtime API initialization failed."
);
case
CUDNN_STATUS_ALLOC_FAILED
:
throw
cudnn_error
(
"CUDA Resources could not be allocated."
);
default
:
throw
cudnn_error
(
"A call to cuDNN failed."
);
}
}
// ------------------------------------------------------------------------------------
cudnn_context
::
cudnn_context
()
:
handle
(
nullptr
)
{
cudnnHandle_t
h
;
check
(
cudnnCreate
(
&
h
));
handle
=
h
;
}
cudnn_context
::~
cudnn_context
()
{
if
(
handle
)
{
cudnnDestroy
((
cudnnHandle_t
)
handle
);
handle
=
nullptr
;
}
}
// ------------------------------------------------------------------------------------
tensor_descriptor
::
tensor_descriptor
()
:
handle
(
nullptr
)
{
cudnnTensorDescriptor_t
h
;
check
(
cudnnCreateTensorDescriptor
(
&
h
));
handle
=
h
;
}
tensor_descriptor
::~
tensor_descriptor
()
{
if
(
handle
)
{
cudnnDestroyTensorDescriptor
((
cudnnTensorDescriptor_t
)
handle
);
handle
=
nullptr
;
}
}
void
tensor_descriptor
::
set_size
(
int
n
,
int
nr
,
int
nc
,
int
k
)
{
check
(
cudnnSetTensor4dDescriptor
((
cudnnTensorDescriptor_t
)
handle
,
CUDNN_TENSOR_NHWC
,
CUDNN_DATA_FLOAT
,
n
,
k
,
nr
,
nc
));
}
void
tensor_descriptor
::
get_size
(
int
&
n
,
int
&
nr
,
int
&
nc
,
int
&
k
)
const
{
int
nStride
,
cStride
,
hStride
,
wStride
;
cudnnDataType_t
datatype
;
check
(
cudnnGetTensor4dDescriptor
((
cudnnTensorDescriptor_t
)
handle
,
&
datatype
,
&
n
,
&
k
,
&
nr
,
&
nc
,
&
nStride
,
&
cStride
,
&
hStride
,
&
wStride
));
}
// ------------------------------------------------------------------------------------
void
add
(
cudnn_context
&
context
,
float
beta
,
tensor
&
dest
,
float
alpha
,
const
tensor
&
src
)
{
}
void
set_tensor
(
cudnn_context
&
context
,
tensor
&
t
,
float
value
)
{
}
void
scale_tensor
(
cudnn_context
&
context
,
tensor
&
t
,
float
value
)
{
}
// ------------------------------------------------------------------------------------
// ------------------------------------------------------------------------------------
conv
::
conv
(
cudnn_context
&
context
,
const
tensor
&
data
,
const
tensor
&
filters
,
int
stride_y
,
int
stride_x
)
{
}
void
conv
::
operator
()
(
resizable_tensor
&
output
,
const
tensor
&
data
,
const
tensor
&
filters
)
{
}
void
conv
::
get_gradient_for_data
(
const
tensor
&
gradient_input
,
const
tensor
&
filters
,
tensor
&
data_gradient
)
{
}
void
conv
::
get_gradient_for_filters
(
const
tensor
&
gradient_input
,
const
tensor
&
data
,
tensor
&
filters_gradient
)
{
}
// ------------------------------------------------------------------------------------
// ------------------------------------------------------------------------------------
void
soft_max
(
cudnn_context
&
context
,
resizable_tensor
&
dest
,
const
tensor
&
src
)
{
}
void
soft_max_gradient
(
cudnn_context
&
context
,
tensor
&
grad
,
const
tensor
&
src
,
const
tensor
&
gradient_input
)
{
}
// ------------------------------------------------------------------------------------
// ------------------------------------------------------------------------------------
max_pool
::
max_pool
(
cudnn_context
&
context
,
int
window_height
,
int
window_width
,
int
stride_y
,
int
stride_x
)
{
}
max_pool
::~
max_pool
(
)
{
}
void
max_pool
::
operator
()
(
resizable_tensor
&
dest
,
const
tensor
&
src
)
{
}
void
max_pool
::
get_gradient
(
const
tensor
&
gradient_input
,
const
tensor
&
src
,
tensor
&
grad
)
{
}
// ------------------------------------------------------------------------------------
// ------------------------------------------------------------------------------------
void
sigmoid
(
cudnn_context
&
context
,
resizable_tensor
&
dest
,
const
tensor
&
src
)
{
}
void
sigmoid_gradient
(
cudnn_context
&
context
,
tensor
&
grad
,
const
tensor
&
src
,
const
tensor
&
gradient_input
)
{
}
// ------------------------------------------------------------------------------------
void
relu
(
cudnn_context
&
context
,
resizable_tensor
&
dest
,
const
tensor
&
src
)
{
}
void
relu_gradient
(
cudnn_context
&
context
,
tensor
&
grad
,
const
tensor
&
src
,
const
tensor
&
gradient_input
)
{
}
// ------------------------------------------------------------------------------------
void
tanh
(
cudnn_context
&
context
,
resizable_tensor
&
dest
,
const
tensor
&
src
)
{
}
void
tanh_gradient
(
cudnn_context
&
context
,
tensor
&
grad
,
const
tensor
&
src
,
const
tensor
&
gradient_input
)
{
}
// ------------------------------------------------------------------------------------
}
}
#endif // DLIB_USE_CUDA
#endif // DLIB_DNN_CuDNN_CPP_
dlib/dnn/cudnn_api.h
View file @
fa0e559f
...
@@ -5,6 +5,7 @@
...
@@ -5,6 +5,7 @@
#ifdef DLIB_USE_CUDA
#ifdef DLIB_USE_CUDA
#include "../error.h"
namespace
dlib
namespace
dlib
{
{
...
@@ -33,17 +34,8 @@ namespace dlib
...
@@ -33,17 +34,8 @@ namespace dlib
cudnn_context
(
cudnn_context
&&
)
=
default
;
cudnn_context
(
cudnn_context
&&
)
=
default
;
cudnn_context
&
operator
=
(
cudnn_context
&&
)
=
default
;
cudnn_context
&
operator
=
(
cudnn_context
&&
)
=
default
;
cudnn_context
()
cudnn_context
();
{
~
cudnn_context
();
// TODO
// cudnnCreate()
}
~
cudnn_context
()
{
// TODO
// cudnnDestroy()
}
const
void
*
get_handle
(
const
void
*
get_handle
(
)
const
{
return
handle
;
}
)
const
{
return
handle
;
}
...
@@ -70,18 +62,8 @@ namespace dlib
...
@@ -70,18 +62,8 @@ namespace dlib
tensor_descriptor
(
tensor_descriptor
&&
)
=
default
;
tensor_descriptor
(
tensor_descriptor
&&
)
=
default
;
tensor_descriptor
&
operator
=
(
tensor_descriptor
&&
)
=
default
;
tensor_descriptor
&
operator
=
(
tensor_descriptor
&&
)
=
default
;
tensor_descriptor
()
tensor_descriptor
();
{
~
tensor_descriptor
();
// cudnnCreateTensorDescriptor()
// Also call cudnnSetTensor4dDescriptor() somewhere. Probably in a set()
// method.
}
~
tensor_descriptor
()
{
// cudnnDestroyTensorDescriptor()
}
void
set_size
(
void
set_size
(
int
n
,
int
n
,
...
...
dlib/dnn/tensor.h
View file @
fa0e559f
...
@@ -6,7 +6,7 @@
...
@@ -6,7 +6,7 @@
#include <memory>
#include <memory>
#include <cstring>
#include <cstring>
#include "../matrix.h"
#include "../matrix.h"
#include "cudnn.h"
#include "cudnn
_api
.h"
namespace
dlib
namespace
dlib
{
{
...
@@ -338,7 +338,7 @@ namespace dlib
...
@@ -338,7 +338,7 @@ namespace dlib
// ----------------------------------------------------------------------------------------
// ----------------------------------------------------------------------------------------
const
matrix_op
<
op_pointer_to_mat
<
float
>
>
mat
(
inline
const
matrix_op
<
op_pointer_to_mat
<
float
>
>
mat
(
const
tensor
&
t
,
const
tensor
&
t
,
long
nr
,
long
nr
,
long
nc
long
nc
...
@@ -360,7 +360,7 @@ namespace dlib
...
@@ -360,7 +360,7 @@ namespace dlib
return
matrix_op
<
op
>
(
op
(
t
.
host
(),
nr
,
nc
));
return
matrix_op
<
op
>
(
op
(
t
.
host
(),
nr
,
nc
));
}
}
const
matrix_op
<
op_pointer_to_mat
<
float
>
>
mat
(
inline
const
matrix_op
<
op_pointer_to_mat
<
float
>
>
mat
(
const
tensor
&
t
const
tensor
&
t
)
)
{
{
...
@@ -486,7 +486,6 @@ namespace dlib
...
@@ -486,7 +486,6 @@ namespace dlib
data
.
set_size
(
m_n
*
m_nr
*
m_nc
*
m_k
);
data
.
set_size
(
m_n
*
m_nr
*
m_nc
*
m_k
);
#ifdef DLIB_USE_CUDA
#ifdef DLIB_USE_CUDA
cudnn_descriptor
.
set_size
(
m_n
,
m_nr
,
m_nc
,
m_k
);
cudnn_descriptor
.
set_size
(
m_n
,
m_nr
,
m_nc
,
m_k
);
#endif
#endif
}
}
};
};
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment