Efficient and expressive arrayed vector math library with multi-threading and CUDA support in Common Lisp.
(defkernel mandelbrot (xs)
(labels ((aux (x y a b m)
(if (< m 100)
(let ((x1 (- (* x x) (* y y) a))
(y1 (- (* 2.0 x y) b)))
(if (> (+ (* x1 x1) (* y1 y1)) 4.0)
m
(aux x1 y1 a b (+ m 1))))
0)))
(let ((a (/ (float (- (mod i 2048) 512)) 1024.0))
(b (/ (float (- (/ i 2048) 1024)) 1024.0)))
(setf (aref xs i) (aux 0.0 0.0 a b 1)))))
(defun draw-mandelbrot (pathname xs)
(with-open-file (out pathname :direction :output
:if-does-not-exist :create
:if-exists :supersede)
(write-line "P2" out)
(write-line "2048 2048" out)
(write-line "255" out)
(dotimes (i (* 2048 2048))
(princ (min 255 (* 8 (array-aref xs i))) out)
(terpri out))))
(defun main (&optional dev-id)
(with-cuda (dev-id)
(with-arrays ((xs int (* 2048 2048)))
(mandelbrot xs)
(draw-mandelbrot #P"./mandelbrot.pgm" xs))))
AVM's kernel functions run almost as fast as equivalent C/C++ codes with SBCL Common Lisp compiler. And we can easily make them run in parallel with just specifying the number of threads we use. Here shows a benchmark of computing 2048x2048 Mandelbrot set and 32768 bodies N-body simulation.
Additionally, AVM provides Nvidia CUDA support so we can enormously accelerate computing kernel functions with GPUs.
These benchmarks are measured on the following environment:
- gcc 4.8.4
- SBCL 1.3.3
- Intel Celeron CPU G3900 @ 2.80GHz
- Nvidia GeForce GTX 750 Ti
- Linux ubuntu 4.2.0-41-generic x86_64
To be described.
To be described.
DEFKERNEL name args body
Defines a kernel function. A defined kernel function is callable as if it is an ordinal Common Lisp function except that it may accept :size
keyword parameter that specifies array size to be processed. If it is not given, the size of the first array in arguments is used. The :size
keyword parameter is for the case that you want to pass arrays with various sizes and use the size of array that does not appear first in arguments. You may use built-in variables i
and n
in kernel functions, which contain the index and the size of array the kernel function processes respectively.
Example:
(defkernel fill-one (xs a)
(setf (aref xs i) 1))
(with-array (xs int 1000)
(fill-one xs))
Delays AVM kernel function compilation to runtime if not nil
. Otherwise, AVM kernel functions are compiled at compile time. Delaying compilation to runtime is useful, at least on SBCL, for debuging AVM kernel definitions because it makes CL debugger show backtrace when they have some errors.
Example:
(setf *compile-on-runtime* t)
(defkernel some-error ()
(+ 1 1.0)) ; AVM compile error showing backtrace
DEFKERNEL-MACRO name args &body body => name
Defines name
as a macro by associating a macro function with that name
in the global environment.
Examples:
(defkernel-macro mac1 (a b)
`(+ ,a (* ,b 3)))
(defkernel fill-with-mac1 (xs a b)
(setf (aref xs i) (mac1 a b)))
EXPAND-MACRO
EXPAND-MACRO-1
To be described.
DEFKERNEL-GROBAL
To be described.
DEFKERNEL-CONSTANT
To be described.
DEFKERNEL-SYMBOL-MACRO
To be described.
ALLOC-ARRAY type size => array
Allocates an AVM array with specified type
and size
. type
is a AVM type that means the array's base type and size
is a positive integer that indicates the size of the array. For now, AVM supports one-dimensional array only. If CUDA is not available, memory area for CUDA is not allocated. alloc-array
ed array should be freed with free-array
. For convenience, with-array
macro and with-arrays
macro are provided.
Example:
(with-cuda (0)
(let ((xs (alloc-array 'int 100)))
(unwind-protect
(do-something-with xs)
(free-array xs))))
FREE-ARRAY array => No value
Frees the given array
. Does nothing if array
is already freed.
Example:
(with-cuda (0)
(let ((xs (alloc-array 'int 100)))
(unwind-protect
(do-something-with xs)
(free-array xs))))
WITH-ARRAY (var type size) body
WITH-ARRAYS ({(var type size)}*) body
Binds var
to an AVM array allocated using alloc-array
applied to the given type
and size
during body
. The array is freed using free-alloc
when with-array
exists. with-arrays
is a plural form of with-array
.
Example:
(with-cuda (0)
(with-array (xs int 100)
(do-something-with xs)))
ARRAY-SIZE array => size
Returns the size of given array.
Example:
(with-array (xs int 100)
(array-size xs))
=> 100
ARRAY-AREF array index => element
SETF (ARRAY-AREF array index) new-element
Accesses array
's element specified by index
. When array
's base type is of vector types, array-aref
and setf array-aref
handles values whose size is its base type's size. For example, array-aref
and setf array-aref
for an array of base type float4
handles values of type (values single-float single-float single-float single-float)
.
If CUDA is available, data on device memory is lazily synchronized before accessing elements of array
, it is only if data on device memory is updated so data on host memory is out-of-date.
Example:
(with-array (xs int2 100)
(initialize-with-ones xs)
(array-aref xs 0))
=> (values 1 1)
(with-array (xs int2 100)
(setf (array-aref xs 0) (values 1 1))
(array-aref xs 0))
=> (values 1 1)
(with-cuda (0)
(with-array (xs int2 100)
(fill-ones xs) ; Ones filled on GPU.
(array-aref xs 0))) ; Lazily synchronized to host.
=> (values 1 1)
WITH-CUDA (dev-id) body
Initializes CUDA and keeps a CUDA context during body
. dev-id
is passed to cl-cuda:get-cuda-device
function to get device handler. If dev-id
is nil
, AVM uses Lisp backend so WITH-CUDA
has no effect. In other words, you can switch if you use CUDA or not by dev-id
.
Example:
(defkernel do-something (xs)
...)
(with-cuda (0)
(with-array (xs int 100)
(do-something xs))) ; Processed on GPU.
(with-cuda (nil)
(with-array (xs int 100)
(do-something xs))) ; Processed on CPU.
In with-cuda
macro's context, specifies if CUDA is used or not. The default value in with-cuda
context is t
. For detail, see CUDA state.
Example:
(with-cuda (0)
(with-array (xs int 100)
(do-something xs) ; Processed on GPU.
(let ((*use-cuda-p* nil))
(do-something xs)))) ; Processed on CPU.
SYNCHRONIZE
Explicitly synchronizes CUDA context with cl-cuda:synchronize-context
. This function is useful in case that you want to get timing data of CUDA kernel launching with time
macro because CUDA kernel functions are executed asynchronously so it passes through time
form in a moment without it. It does nothing if CUDA is not available.
Example:
(with-cuda (0)
(time
(progn
(do-something)
(synchronize))))
To be described.
To be described.
Integer type int
and its derived vector types.
Single precision floating point type float
and its derived vector types.
Double precision floating point type double
and its derived vector types.
THE type form => result
the
specifies the value returned by form
is of the type specified by type
.
Example:
(flet ((identity (x)
(the int x)))
(identity 42))
=> 42
IF test-form then-form else-form => result
if
allows the evaluation of a form to be dependent on a single test-form
. First test-form
is evaluated. If the result is true
, then then-form
is selected; otherwise else-form
is selected. Whichever form is selected is then evaluated and its result value returns.
Example:
(let ((a 1))
(if (= a 1)
42
0))
=> 42
LET ({(var value)}*) body => result
let
declares new variable bindings and set corresponding value
s to them and evaluates body
form that uses these bindings and returns its result. let
performs the bindings in parallel. For sequentially, use let*
kernel macro instead.
Example:
(let ((x 42))
x)
=> 42
FLET ({(name args local-form)}*) body => result
flet
defines locally named functions and evaluates its body
with these definition bindings. Any number of such local functions can be defined. The scope of the name binding encompasses only the body. You may use built-in variables i
and n
in local functions as well as global kernel functions.
Example:
(flet ((aux (x) (+ x 1)))
(aux 10))
=> 11
LABELS ({(name args local-form)}*) body => result
labels
is equivalent to flet
except that the scope of the defined function names for labels
encompasses the function definition themselves as well as the body.
Examples:
(labels ((aux (x y)
(if (= x 0)
y
(aux (- x 1) (+ y 1)))))
(aux 10 0))
=> 10
SETF place value => result
setf
changes the value of place
to be value
and returns value
as its result. Accessor forms are acceptable as place
s.
Example:
(setf (aref xs i) (+ (aref xs i) (int2 1 1)))
=> (int2 1 1)
INT2 form form => result
INT3 form form form => result
INT4 form form form form => result
int2
, int3
and int4
provide constructors for int
derived vector types. Each parameter form should have type of int
.
Example:
(int2 0 0) => (int2 0 0)
(int3 0 0 0) => (int3 0 0 0)
(int4 0 0 0 0) => (int4 0 0 0 0)
FLOAT2 form form => result
FLOAT3 form form form => result
FLOAT4 form form form form => result
float2
, float3
and flaot4
provide constructors for float
derived vector types. Each parameter form should have type of float
.
Example:
(float2 0.0 0.0) => (float2 0.0 0.0)
(float3 0.0 0.0 0.0) => (float3 0.0 0.0 0.0)
(float4 0.0 0.0 0.0 0.0) => (float4 0.0 0.0 0.0 0.0)
DOUBLE2 form form => result
DOUBLE3 form form form => result
DOUBLE4 form form form form => result
double2
, double3
and double4
provide constructors for double
derived vector types. Each parameter form should have type of double
.
Example:
(double2 0.0d0 0.0d0) => (double2 0.0d0 0.0d0)
(double3 0.0d0 0.0d0 0.0d0) => (double3 0.0d0 0.0d0 0.0d0)
(double4 0.0d0 0.0d0 0.0d0 0.0d0) => (double4 0.0d0 0.0d0 0.0d0 0.0d0)
INT2-X form => result
INT2-Y form => result
INT3-X form => result
INT3-Y form => result
INT3-Z form => result
INT4-X form => result
INT4-Y form => result
INT4-Z form => result
INT4-W form => result
Accesses each component of int
derived vector types. The type of form
should be of each accessor's corresponding vector type. The type of result is int
. You can read its value as well as destructively set it with SETF form.
Example:
(int2-x (int2 1 2)) => 1
(let ((x (int2 1 2)))
(setf (int2-x x) 3)
x)
=> (int2 3 2)
FLOAT2-X form => result
FLOAT2-Y form => result
FLOAT3-X form => result
FLOAT3-Y form => result
FLOAT3-Z form => result
FLOAT4-X form => result
FLOAT4-Y form => result
FLOAT4-Z form => result
FLOAT4-W form => result
Accesses each component of float
derived vector types. The type of form
should be of each accessor's corresponding vector type. The type of result is float
. You can read its value as well as destructively set it with SETF form.
Example:
(float2-x (float2 1.0 2.0)) => 1.0
(let ((x (float2 1.0 2.0)))
(setf (float2-x x) 3.0)
x)
=> (float2 3.0 2.0)
[Accessor] double2-x, double2-y, double3-x, double3-y, double3-z, double4-x, double4-y, double4-z, double4-w
DOUBLE2-X form => result
DOUBLE2-Y form => result
DOUBLE3-X form => result
DOUBLE3-Y form => result
DOUBLE3-Z form => result
DOUBLE4-X form => result
DOUBLE4-Y form => result
DOUBLE4-Z form => result
DOUBLE4-W form => result
Accesses each component of double
derived vector types. The type of form
should be of each accessor's corresponding vector type. The type of result is double
. You can read its value as well as destructively set it with SETF form.
Example:
(double2-x (double2 1.0d0 2.0d0)) => 1.0d0
(let ((x (double2 1.0d0 2.0d0)))
(setf (double2-x x) 3.0d0)
x)
=> (double2 3.0d0 2.0d0)
AREF array index => result
Accesses the array
element specified by the index
. The type of array
is an array type with base type of int
, float
, double
and their derived vector types. The type of index
is int
, and the type of result
is the base type. You can read its value as well as destructively set it with SETF form.
Example:
(aref xs 0) => 1
(setf (aref xs 0) 1) => 1
(aref ys 0) => (int2 1 1)
(setf (aref ys 0) (int2 2 2)) => 2
i
has the index of current AVM thread, n
has the number of all AVM threads in a calling AVM kernel function. The both have type of int
. You may use them in any kernel functions including ones called by another.
Examples:
(defkernel fill-zeros (xs)
(set (aref xs i) 0))
These functions provide arithmetic operations. +
and -
accept scalar type and vector type arithmetic. *
and /
accept scalar type arithmetic. mod
accepts int
type operands. +
, -
, *
and /
may take more than two arguments.
(defkernel add2 (x y)
(the int2 (+ x y)))
(defkernel add3 (x y z)
(the int2 (+ x y z)))
These functions provide vector algebraic operations. *.
scales a vector type value by a scalar type value. .*
does the same thing, letting a scalar value to scale a vector value. /.
divides a vector type value by a scalar type value. *.
and /.
may take more than two arguments, repeatedly applying scalar values.
(defkernel scale (x a b)
(the float3 (*. x a b))) ; x is of float3 and a, b are of float
(defkernel scaled (a x)
(the float3 (.* a x))) ; a is of float and x is of float3
These functions provide comparison operations for scalar type values.
(defkernel one-or-two (a b)
(if (< (the float a) b)
1
2))
RSQRT x => result
These built-in functions provide mathematical functions.
PROGN
To be described.
LET*
To be described.
The array in AVM is an abstract data structure that consists of a memory area for computation on Lisp and another for computation on Nvidia CUDA.
- Memory area for computation on Lisp
- Tuple array
- Memory area for computation on CUDA
- Host memory
- Device memory
Eacn AVM thread of kernel functions reads from and writes to arrays. Arrays are passed to kernel fucntions on launching them.
alloc-array
allocates an array and it should be freed with free-array
. For convenience, with-array
macro and with-arrays
macro are provided. array-aref
accessor is used to read and write a value to an array. A value of arrays whose base type is a vector type is accessed via values.
AVM has the following three states relevant to CUDA:
- Not Available
- Available
- Used
And here shows CUDA state's transfer diagram:
WITH-CUDA (nil) WITH-CUDA (N)
+-----+ +--------------------------------------+
| | | |
| | | *use-cuda-p* nil |
| | | <------------- v
+-> Not Available Available Used
------------->
*use-cuda-p* not nil
The initial state is "Not Available". When CUDA state is "Not Available", AVM does not use CUDA. When AVM has this state is that actually CUDA is not available on your machine, out of with-cuda
context or in with-cuda
context with its dev-id
parameter nil
.
When CUDA state is "Available", AVM is ready to use CUDA with initializing it and creating CUDA context as well as allocating device memory on alloc-array
ing, though kernel functions are actually not executed on CUDA in this state. When AVM has this state is that CUDA is available on your machine within with-cuda
context with its dev-id
parameter an integer that indicates which GPU you use and *use-cuda-p*
special variable is set to nil
.
When CUDA state is "Used", AVM is ready to use CUDA as well as when CUDA state is "Available" and kernel functions are actually executed on CUDA. When AVM has this state is same as when CUDA state is "Available" except that *use-cuda-p*
special variable is set to not nil
, which is the default value of that in with-cuda
context.
AVM's arrays have the following state variables:
- CUDA availability
- CUDA availability on allocation
- Freed or not
CUDA availability is if CUDA is available or not when array operations are called. CUDA availability on allocation is if CUDA was available or not when arrays are alloc-array
ed. Free or not is if arrays are already free-array
ed or not.
How array operations work is dependent of these state variables. For detail, see each array operation's API description.
To be described.
- Masayuki Takagi ([email protected])
Copyright (c) 2016 Masayuki Takagi ([email protected])
Licensed under the MIT License.