Miscellaneous
Testing the ASTRA installation
astra.test()
astra_test
This will run quick tests of basic CPU/GPU functionality, and report on the results. As a part of this it will report if GPU functionality is available.
Working with large data
In case the data size is so large that it doesn’t fit into the GPU memory, ASTRA will automatically split and process the data in subsets. This functionality is only available for FP3D, BP3D and FDK algorithms.
WARNING! At the moment, if the input/output data is a linked GPU tensor, the automatic splitting will not work correctly.
WARNING! Other GPU libraries, such as PyTorch, often allocate more GPU memory then
they actually need to speed up computations. This significantly reduces the memory
available to ASTRA, and usually it’s a good idea to shrink the memory pool on the
external library side (e.g. with torch.cuda.empty_cache) before calling ASTRA if you
get out-of-memory errors.
Choosing the GPU to use
On systems equipped with several GPUs, you can specify which GPU will be used by ASTRA with:
astra.set_gpu_index(index)
astra_mex('set_gpu_index', index);
WARNING! In multithreading contexts (Python), the GPU index will be set globally,
so it can’t be used to reliably restrict a thread to a given GPU. Instead, you can avoid
calling astra.set_gpu_index whatsoever, and instead set the desired GPU with an
external library where the device context is thread-local, e.g. using
torch.cuda.set_device.
Using several GPUs cooperatively
ASTRA can utilize several GPUs simultaneously for a single algorithm to speed up the computation. To do that, you can define the desired set of GPUs to be used:
astra.set_gpu_index([index1, index2, ...])
astra_mex('set_gpu_index', [index1 index2 ...]);
WARNING! At the moment, only FP3D, BP3D and FDK algorithms support this functionality. For the rest, the first GPU in the specified index list will be used as the fallback.
Multithreading
In Python, Astra supports concurrent execution of its algorithms in multiple threads. This functionality is useful to process different inputs on different GPUs simultaneously. Another use case is processing data blocks on the same GPU in parallel, which is useful when a single ASTRA call under-utilizes the GPU. For instance, here is how one can compute a fan-parallel projection using multithreading:
import astra
import numpy as np
from concurrent.futures import ThreadPoolExecutor
N = 512
vol_geom_full = astra.create_vol_geom(N, N, N)
vol_geom_slice = astra.create_vol_geom(N, N)
angles = np.linspace(0, 2*np.pi, N, endpoint=False)
proj_geom_slice = astra.create_proj_geom('fanflat', 1.0, N, angles, N, N)
projector = astra.create_projector('cuda', proj_geom_slice, vol_geom_slice)
phantom_id, phantom_data = astra.data3d.shepp_logan(vol_geom_full)
def forward_project_slice(vol_slice):
slice_proj_id, slice_proj_data = astra.create_sino(vol_slice, projector)
astra.data2d.delete(slice_proj_id)
return slice_proj_data
with ThreadPoolExecutor(max_workers=16) as executor:
projections = list(executor.map(forward_project_slice, phantom_data))
projections = np.stack(projections, axis=0)
MATLAB doesn’t support executing external libraries in multithreaded environments.
WARNING! No special care is taken about race conditions, so the user has to ensure that the outputs of algorithms are not accessed simultaneously.
WARNING! MATLAB doesn’t support executing external libraries in multithreaded environments, so the only option is much less lightweight process-based concurrent execution. The simplest approach is to just start several copies of MATLAB in batch mode.
Masks
Various reconstruction algorithms support projection data and reconstruction volume masks. These behave as follows.
The projection data elements corresponding to locations with SinogramMask value 0.0 will be ignored during the reconstruction. Similarly, the reconstruction data elements corresponding to locations with ReconstructionMask value 0.0 will be ignored during the reconstruction, and their values will be preserved (mostly, see notes below).
The algorithm will behave as if the rows and columns corresponding to the masked voxels and projection data elements have been removed from the projection matrix entirely. In other words, it will iteratively try to match the projection of the non-masked voxels to the non-masked projection data elements.
WARNING! MinConstraint/MaxConstraint will affect even masked voxels.
WARNING! FP and BP algorithms (CPU versions) overwrite the output, so the values outside the sinogram/reconstruction masks, respectively, will be set to zero instead of being ignored.
ASTRA configuration structure
cfg = astra.astra_dict('NAME')
cfg = astra_struct('NAME');
This is the basic script to create a configuration struct for many astra objects. The returned struct is usually filled with more options after creating it, and then passed to astra functions such as
id = astra.algorithm.create(cfg)
id = astra.projector.create(cfg)
id = astra_mex_algorithm('create', cfg);
id = astra_mex_projector('create', cfg);
The most common usage is for creating algorithm configuration structs. See the pages for individual algorithms for the options they support.
Projection matrix objects
Matrix objects can be created by the ASTRA toolbox to obtain explicit weight
matrices (see here), or you can define them yourself for
use with the sparse_matrix projection geometry. Matrix objects can be
manipulated using the following commands:
create
id = astra.matrix.create(S)
id = astra_mex_matrix('create', S);
get
Create an ASTRA sparse matrix object from a Python sparse matrix of type scipy.sparse.csr_matrix or a MATLAB sparse matrix.
S = astra.matrix.get(id)
S = astra_mex_matrix('get', id);
Return an ASTRA sparse matrix object as a Python sparse matrix of type scipy.sparse.csr_matrix or a MATLAB sparse matrix.
get_size
s = astra.matrix.get_size(id)
s = astra_mex_matrix('get_size', id);
Get the size (rows,columns) of the sparse matrix object.
store
astra.matrix.store(id, S)
astra_mex_matrix('store', id, S);
Store a new Python or MATLAB sparse matrix in an ASTRA sparse matrix object.
NB: This does not re-allocate memory: the number of rows and non-zero entries may not be larger than they were when the object was first created.
delete
astra.matrix.delete(id)
astra.matrix.delete([id1, id2, ...])
astra_mex_matrix('delete', id)
Free a single sparse matrix.
clear
astra.matrix.clear()
astra_mex_matrix('clear')
Free all sparse matrices.
info
astra.matrix.info()
astra_mex_matrix('info')
Print basic information about all allocated sparse matrix objects.