torch.Storage¶
A torch.Storage
is a contiguous, one-dimensional array of a single
data type.
Every torch.Tensor
has a corresponding storage of the same data type.
-
class
torch.
DoubleStorage
(*args, **kwargs)[source]¶ -
bfloat16
()¶ Casts this storage to bfloat16 type
-
bool
()¶ Casts this storage to bool type
-
byte
()¶ Casts this storage to byte type
-
char
()¶ Casts this storage to char type
-
clone
()¶ Returns a copy of this storage
-
complex_double
()¶ Casts this storage to complex double type
-
complex_float
()¶ Casts this storage to complex float type
-
copy_
(source, non_blocking=None)¶
-
cpu
()¶ Returns a CPU copy of this storage if it’s not already on the CPU
-
cuda
(device=None, non_blocking=False, **kwargs)¶
-
data_ptr
()¶
-
property
device
¶
-
double
()¶ Casts this storage to double type
-
element_size
()¶
-
fill_
(value)¶
-
float
()¶ Casts this storage to float type
-
classmethod
from_buffer
(*args, **kwargs)¶
-
classmethod
from_file
(filename, shared, size)¶
-
get_device
()¶
-
half
()¶ Casts this storage to half type
-
int
()¶ Casts this storage to int type
-
property
is_cuda
¶
-
is_pinned
()¶
-
is_sparse
= False¶
-
long
()¶ Casts this storage to long type
-
nbytes
()¶
-
pickle_storage_type
()¶
-
pin_memory
()¶ Coppies the storage to pinned memory, if it’s not already pinned.
-
resize_
(size)¶
Moves the storage to shared memory.
This is a no-op for storages already in shared memory and for CUDA storages, which do not need to be moved for sharing across processes. Storages in shared memory cannot be resized.
Returns: self
-
short
()¶ Casts this storage to short type
-
size
()¶
-
tolist
()¶ Returns a list containing the elements of this storage
-
type
(dtype=None, non_blocking=False)¶
-
-
class
torch.
FloatStorage
(*args, **kwargs)[source]¶ -
bfloat16
()¶ Casts this storage to bfloat16 type
-
bool
()¶ Casts this storage to bool type
-
byte
()¶ Casts this storage to byte type
-
char
()¶ Casts this storage to char type
-
clone
()¶ Returns a copy of this storage
-
complex_double
()¶ Casts this storage to complex double type
-
complex_float
()¶ Casts this storage to complex float type
-
copy_
(source, non_blocking=None)¶
-
cpu
()¶ Returns a CPU copy of this storage if it’s not already on the CPU
-
cuda
(device=None, non_blocking=False, **kwargs)¶
-
data_ptr
()¶
-
property
device
¶
-
double
()¶ Casts this storage to double type
-
element_size
()¶
-
fill_
(value)¶
-
float
()¶ Casts this storage to float type
-
classmethod
from_buffer
(*args, **kwargs)¶
-
classmethod
from_file
(filename, shared, size)¶
-
get_device
()¶
-
half
()¶ Casts this storage to half type
-
int
()¶ Casts this storage to int type
-
property
is_cuda
¶
-
is_pinned
()¶
-
is_sparse
= False¶
-
long
()¶ Casts this storage to long type
-
nbytes
()¶
-
pickle_storage_type
()¶
-
pin_memory
()¶ Coppies the storage to pinned memory, if it’s not already pinned.
-
resize_
(size)¶
Moves the storage to shared memory.
This is a no-op for storages already in shared memory and for CUDA storages, which do not need to be moved for sharing across processes. Storages in shared memory cannot be resized.
Returns: self
-
short
()¶ Casts this storage to short type
-
size
()¶
-
tolist
()¶ Returns a list containing the elements of this storage
-
type
(dtype=None, non_blocking=False)¶
-
-
class
torch.
HalfStorage
(*args, **kwargs)[source]¶ -
bfloat16
()¶ Casts this storage to bfloat16 type
-
bool
()¶ Casts this storage to bool type
-
byte
()¶ Casts this storage to byte type
-
char
()¶ Casts this storage to char type
-
clone
()¶ Returns a copy of this storage
-
complex_double
()¶ Casts this storage to complex double type
-
complex_float
()¶ Casts this storage to complex float type
-
copy_
(source, non_blocking=None)¶
-
cpu
()¶ Returns a CPU copy of this storage if it’s not already on the CPU
-
cuda
(device=None, non_blocking=False, **kwargs)¶
-
data_ptr
()¶
-
property
device
¶
-
double
()¶ Casts this storage to double type
-
element_size
()¶
-
fill_
(value)¶
-
float
()¶ Casts this storage to float type
-
classmethod
from_buffer
(*args, **kwargs)¶
-
classmethod
from_file
(filename, shared, size)¶
-
get_device
()¶
-
half
()¶ Casts this storage to half type
-
int
()¶ Casts this storage to int type
-
property
is_cuda
¶
-
is_pinned
()¶
-
is_sparse
= False¶
-
long
()¶ Casts this storage to long type
-
nbytes
()¶
-
pickle_storage_type
()¶
-
pin_memory
()¶ Coppies the storage to pinned memory, if it’s not already pinned.
-
resize_
(size)¶
Moves the storage to shared memory.
This is a no-op for storages already in shared memory and for CUDA storages, which do not need to be moved for sharing across processes. Storages in shared memory cannot be resized.
Returns: self
-
short
()¶ Casts this storage to short type
-
size
()¶
-
tolist
()¶ Returns a list containing the elements of this storage
-
type
(dtype=None, non_blocking=False)¶
-
-
class
torch.
LongStorage
(*args, **kwargs)[source]¶ -
bfloat16
()¶ Casts this storage to bfloat16 type
-
bool
()¶ Casts this storage to bool type
-
byte
()¶ Casts this storage to byte type
-
char
()¶ Casts this storage to char type
-
clone
()¶ Returns a copy of this storage
-
complex_double
()¶ Casts this storage to complex double type
-
complex_float
()¶ Casts this storage to complex float type
-
copy_
(source, non_blocking=None)¶
-
cpu
()¶ Returns a CPU copy of this storage if it’s not already on the CPU
-
cuda
(device=None, non_blocking=False, **kwargs)¶
-
data_ptr
()¶
-
property
device
¶
-
double
()¶ Casts this storage to double type
-
element_size
()¶
-
fill_
(value)¶
-
float
()¶ Casts this storage to float type
-
classmethod
from_buffer
(*args, **kwargs)¶
-
classmethod
from_file
(filename, shared, size)¶
-
get_device
()¶
-
half
()¶ Casts this storage to half type
-
int
()¶ Casts this storage to int type
-
property
is_cuda
¶
-
is_pinned
()¶
-
is_sparse
= False¶
-
long
()¶ Casts this storage to long type
-
nbytes
()¶
-
pickle_storage_type
()¶
-
pin_memory
()¶ Coppies the storage to pinned memory, if it’s not already pinned.
-
resize_
(size)¶
Moves the storage to shared memory.
This is a no-op for storages already in shared memory and for CUDA storages, which do not need to be moved for sharing across processes. Storages in shared memory cannot be resized.
Returns: self
-
short
()¶ Casts this storage to short type
-
size
()¶
-
tolist
()¶ Returns a list containing the elements of this storage
-
type
(dtype=None, non_blocking=False)¶
-
-
class
torch.
IntStorage
(*args, **kwargs)[source]¶ -
bfloat16
()¶ Casts this storage to bfloat16 type
-
bool
()¶ Casts this storage to bool type
-
byte
()¶ Casts this storage to byte type
-
char
()¶ Casts this storage to char type
-
clone
()¶ Returns a copy of this storage
-
complex_double
()¶ Casts this storage to complex double type
-
complex_float
()¶ Casts this storage to complex float type
-
copy_
(source, non_blocking=None)¶
-
cpu
()¶ Returns a CPU copy of this storage if it’s not already on the CPU
-
cuda
(device=None, non_blocking=False, **kwargs)¶
-
data_ptr
()¶
-
property
device
¶
-
double
()¶ Casts this storage to double type
-
element_size
()¶
-
fill_
(value)¶
-
float
()¶ Casts this storage to float type
-
classmethod
from_buffer
(*args, **kwargs)¶
-
classmethod
from_file
(filename, shared, size)¶
-
get_device
()¶
-
half
()¶ Casts this storage to half type
-
int
()¶ Casts this storage to int type
-
property
is_cuda
¶
-
is_pinned
()¶
-
is_sparse
= False¶
-
long
()¶ Casts this storage to long type
-
nbytes
()¶
-
pickle_storage_type
()¶
-
pin_memory
()¶ Coppies the storage to pinned memory, if it’s not already pinned.
-
resize_
(size)¶
Moves the storage to shared memory.
This is a no-op for storages already in shared memory and for CUDA storages, which do not need to be moved for sharing across processes. Storages in shared memory cannot be resized.
Returns: self
-
short
()¶ Casts this storage to short type
-
size
()¶
-
tolist
()¶ Returns a list containing the elements of this storage
-
type
(dtype=None, non_blocking=False)¶
-
-
class
torch.
ShortStorage
(*args, **kwargs)[source]¶ -
bfloat16
()¶ Casts this storage to bfloat16 type
-
bool
()¶ Casts this storage to bool type
-
byte
()¶ Casts this storage to byte type
-
char
()¶ Casts this storage to char type
-
clone
()¶ Returns a copy of this storage
-
complex_double
()¶ Casts this storage to complex double type
-
complex_float
()¶ Casts this storage to complex float type
-
copy_
(source, non_blocking=None)¶
-
cpu
()¶ Returns a CPU copy of this storage if it’s not already on the CPU
-
cuda
(device=None, non_blocking=False, **kwargs)¶
-
data_ptr
()¶
-
property
device
¶
-
double
()¶ Casts this storage to double type
-
element_size
()¶
-
fill_
(value)¶
-
float
()¶ Casts this storage to float type
-
classmethod
from_buffer
(*args, **kwargs)¶
-
classmethod
from_file
(filename, shared, size)¶
-
get_device
()¶
-
half
()¶ Casts this storage to half type
-
int
()¶ Casts this storage to int type
-
property
is_cuda
¶
-
is_pinned
()¶
-
is_sparse
= False¶
-
long
()¶ Casts this storage to long type
-
nbytes
()¶
-
pickle_storage_type
()¶
-
pin_memory
()¶ Coppies the storage to pinned memory, if it’s not already pinned.
-
resize_
(size)¶
Moves the storage to shared memory.
This is a no-op for storages already in shared memory and for CUDA storages, which do not need to be moved for sharing across processes. Storages in shared memory cannot be resized.
Returns: self
-
short
()¶ Casts this storage to short type
-
size
()¶
-
tolist
()¶ Returns a list containing the elements of this storage
-
type
(dtype=None, non_blocking=False)¶
-
-
class
torch.
CharStorage
(*args, **kwargs)[source]¶ -
bfloat16
()¶ Casts this storage to bfloat16 type
-
bool
()¶ Casts this storage to bool type
-
byte
()¶ Casts this storage to byte type
-
char
()¶ Casts this storage to char type
-
clone
()¶ Returns a copy of this storage
-
complex_double
()¶ Casts this storage to complex double type
-
complex_float
()¶ Casts this storage to complex float type
-
copy_
(source, non_blocking=None)¶
-
cpu
()¶ Returns a CPU copy of this storage if it’s not already on the CPU
-
cuda
(device=None, non_blocking=False, **kwargs)¶
-
data_ptr
()¶
-
property
device
¶
-
double
()¶ Casts this storage to double type
-
element_size
()¶
-
fill_
(value)¶
-
float
()¶ Casts this storage to float type
-
classmethod
from_buffer
(*args, **kwargs)¶
-
classmethod
from_file
(filename, shared, size)¶
-
get_device
()¶
-
half
()¶ Casts this storage to half type
-
int
()¶ Casts this storage to int type
-
property
is_cuda
¶
-
is_pinned
()¶
-
is_sparse
= False¶
-
long
()¶ Casts this storage to long type
-
nbytes
()¶
-
pickle_storage_type
()¶
-
pin_memory
()¶ Coppies the storage to pinned memory, if it’s not already pinned.
-
resize_
(size)¶
Moves the storage to shared memory.
This is a no-op for storages already in shared memory and for CUDA storages, which do not need to be moved for sharing across processes. Storages in shared memory cannot be resized.
Returns: self
-
short
()¶ Casts this storage to short type
-
size
()¶
-
tolist
()¶ Returns a list containing the elements of this storage
-
type
(dtype=None, non_blocking=False)¶
-
-
class
torch.
ByteStorage
(*args, **kwargs)[source]¶ -
bfloat16
()¶ Casts this storage to bfloat16 type
-
bool
()¶ Casts this storage to bool type
-
byte
()¶ Casts this storage to byte type
-
char
()¶ Casts this storage to char type
-
clone
()¶ Returns a copy of this storage
-
complex_double
()¶ Casts this storage to complex double type
-
complex_float
()¶ Casts this storage to complex float type
-
copy_
(source, non_blocking=None)¶
-
cpu
()¶ Returns a CPU copy of this storage if it’s not already on the CPU
-
cuda
(device=None, non_blocking=False, **kwargs)¶
-
data_ptr
()¶
-
property
device
¶
-
double
()¶ Casts this storage to double type
-
element_size
()¶
-
fill_
(value)¶
-
float
()¶ Casts this storage to float type
-
classmethod
from_buffer
(*args, **kwargs)¶
-
classmethod
from_file
(filename, shared, size)¶
-
get_device
()¶
-
half
()¶ Casts this storage to half type
-
int
()¶ Casts this storage to int type
-
property
is_cuda
¶
-
is_pinned
()¶
-
is_sparse
= False¶
-
long
()¶ Casts this storage to long type
-
nbytes
()¶
-
pickle_storage_type
()¶
-
pin_memory
()¶ Coppies the storage to pinned memory, if it’s not already pinned.
-
resize_
(size)¶
Moves the storage to shared memory.
This is a no-op for storages already in shared memory and for CUDA storages, which do not need to be moved for sharing across processes. Storages in shared memory cannot be resized.
Returns: self
-
short
()¶ Casts this storage to short type
-
size
()¶
-
tolist
()¶ Returns a list containing the elements of this storage
-
type
(dtype=None, non_blocking=False)¶
-
-
class
torch.
BoolStorage
(*args, **kwargs)[source]¶ -
bfloat16
()¶ Casts this storage to bfloat16 type
-
bool
()¶ Casts this storage to bool type
-
byte
()¶ Casts this storage to byte type
-
char
()¶ Casts this storage to char type
-
clone
()¶ Returns a copy of this storage
-
complex_double
()¶ Casts this storage to complex double type
-
complex_float
()¶ Casts this storage to complex float type
-
copy_
(source, non_blocking=None)¶
-
cpu
()¶ Returns a CPU copy of this storage if it’s not already on the CPU
-
cuda
(device=None, non_blocking=False, **kwargs)¶
-
data_ptr
()¶
-
property
device
¶
-
double
()¶ Casts this storage to double type
-
element_size
()¶
-
fill_
(value)¶
-
float
()¶ Casts this storage to float type
-
classmethod
from_buffer
(*args, **kwargs)¶
-
classmethod
from_file
(filename, shared, size)¶
-
get_device
()¶
-
half
()¶ Casts this storage to half type
-
int
()¶ Casts this storage to int type
-
property
is_cuda
¶
-
is_pinned
()¶
-
is_sparse
= False¶
-
long
()¶ Casts this storage to long type
-
nbytes
()¶
-
pickle_storage_type
()¶
-
pin_memory
()¶ Coppies the storage to pinned memory, if it’s not already pinned.
-
resize_
(size)¶
Moves the storage to shared memory.
This is a no-op for storages already in shared memory and for CUDA storages, which do not need to be moved for sharing across processes. Storages in shared memory cannot be resized.
Returns: self
-
short
()¶ Casts this storage to short type
-
size
()¶
-
tolist
()¶ Returns a list containing the elements of this storage
-
type
(dtype=None, non_blocking=False)¶
-
-
class
torch.
BFloat16Storage
(*args, **kwargs)[source]¶ -
bfloat16
()¶ Casts this storage to bfloat16 type
-
bool
()¶ Casts this storage to bool type
-
byte
()¶ Casts this storage to byte type
-
char
()¶ Casts this storage to char type
-
clone
()¶ Returns a copy of this storage
-
complex_double
()¶ Casts this storage to complex double type
-
complex_float
()¶ Casts this storage to complex float type
-
copy_
(source, non_blocking=None)¶
-
cpu
()¶ Returns a CPU copy of this storage if it’s not already on the CPU
-
cuda
(device=None, non_blocking=False, **kwargs)¶
-
data_ptr
()¶
-
property
device
¶
-
double
()¶ Casts this storage to double type
-
element_size
()¶
-
fill_
(value)¶
-
float
()¶ Casts this storage to float type
-
classmethod
from_buffer
(*args, **kwargs)¶
-
classmethod
from_file
(filename, shared, size)¶
-
get_device
()¶
-
half
()¶ Casts this storage to half type
-
int
()¶ Casts this storage to int type
-
property
is_cuda
¶
-
is_pinned
()¶
-
is_sparse
= False¶
-
long
()¶ Casts this storage to long type
-
nbytes
()¶
-
pickle_storage_type
()¶
-
pin_memory
()¶ Coppies the storage to pinned memory, if it’s not already pinned.
-
resize_
(size)¶
Moves the storage to shared memory.
This is a no-op for storages already in shared memory and for CUDA storages, which do not need to be moved for sharing across processes. Storages in shared memory cannot be resized.
Returns: self
-
short
()¶ Casts this storage to short type
-
size
()¶
-
tolist
()¶ Returns a list containing the elements of this storage
-
type
(dtype=None, non_blocking=False)¶
-
-
class
torch.
ComplexDoubleStorage
(*args, **kwargs)[source]¶ -
bfloat16
()¶ Casts this storage to bfloat16 type
-
bool
()¶ Casts this storage to bool type
-
byte
()¶ Casts this storage to byte type
-
char
()¶ Casts this storage to char type
-
clone
()¶ Returns a copy of this storage
-
complex_double
()¶ Casts this storage to complex double type
-
complex_float
()¶ Casts this storage to complex float type
-
copy_
(source, non_blocking=None)¶
-
cpu
()¶ Returns a CPU copy of this storage if it’s not already on the CPU
-
cuda
(device=None, non_blocking=False, **kwargs)¶
-
data_ptr
()¶
-
property
device
¶
-
double
()¶ Casts this storage to double type
-
element_size
()¶
-
fill_
(value)¶
-
float
()¶ Casts this storage to float type
-
classmethod
from_buffer
(*args, **kwargs)¶
-
classmethod
from_file
(filename, shared, size)¶
-
get_device
()¶
-
half
()¶ Casts this storage to half type
-
int
()¶ Casts this storage to int type
-
property
is_cuda
¶
-
is_pinned
()¶
-
is_sparse
= False¶
-
long
()¶ Casts this storage to long type
-
nbytes
()¶
-
pickle_storage_type
()¶
-
pin_memory
()¶ Coppies the storage to pinned memory, if it’s not already pinned.
-
resize_
(size)¶
Moves the storage to shared memory.
This is a no-op for storages already in shared memory and for CUDA storages, which do not need to be moved for sharing across processes. Storages in shared memory cannot be resized.
Returns: self
-
short
()¶ Casts this storage to short type
-
size
()¶
-
tolist
()¶ Returns a list containing the elements of this storage
-
type
(dtype=None, non_blocking=False)¶
-
-
class
torch.
ComplexFloatStorage
(*args, **kwargs)[source]¶ -
bfloat16
()¶ Casts this storage to bfloat16 type
-
bool
()¶ Casts this storage to bool type
-
byte
()¶ Casts this storage to byte type
-
char
()¶ Casts this storage to char type
-
clone
()¶ Returns a copy of this storage
-
complex_double
()¶ Casts this storage to complex double type
-
complex_float
()¶ Casts this storage to complex float type
-
copy_
(source, non_blocking=None)¶
-
cpu
()¶ Returns a CPU copy of this storage if it’s not already on the CPU
-
cuda
(device=None, non_blocking=False, **kwargs)¶
-
data_ptr
()¶
-
property
device
¶
-
double
()¶ Casts this storage to double type
-
element_size
()¶
-
fill_
(value)¶
-
float
()¶ Casts this storage to float type
-
classmethod
from_buffer
(*args, **kwargs)¶
-
classmethod
from_file
(filename, shared, size)¶
-
get_device
()¶
-
half
()¶ Casts this storage to half type
-
int
()¶ Casts this storage to int type
-
property
is_cuda
¶
-
is_pinned
()¶
-
is_sparse
= False¶
-
long
()¶ Casts this storage to long type
-
nbytes
()¶
-
pickle_storage_type
()¶
-
pin_memory
()¶ Coppies the storage to pinned memory, if it’s not already pinned.
-
resize_
(size)¶
Moves the storage to shared memory.
This is a no-op for storages already in shared memory and for CUDA storages, which do not need to be moved for sharing across processes. Storages in shared memory cannot be resized.
Returns: self
-
short
()¶ Casts this storage to short type
-
size
()¶
-
tolist
()¶ Returns a list containing the elements of this storage
-
type
(dtype=None, non_blocking=False)¶
-
-
class
torch.
QUInt8Storage
(*args, **kwargs)[source]¶ -
bfloat16
()¶ Casts this storage to bfloat16 type
-
bool
()¶ Casts this storage to bool type
-
byte
()¶ Casts this storage to byte type
-
char
()¶ Casts this storage to char type
-
clone
()¶ Returns a copy of this storage
-
complex_double
()¶ Casts this storage to complex double type
-
complex_float
()¶ Casts this storage to complex float type
-
copy_
(source, non_blocking=None)¶
-
cpu
()¶ Returns a CPU copy of this storage if it’s not already on the CPU
-
cuda
(device=None, non_blocking=False, **kwargs)¶
-
data_ptr
()¶
-
property
device
¶
-
double
()¶ Casts this storage to double type
-
element_size
()¶
-
fill_
(value)¶
-
float
()¶ Casts this storage to float type
-
classmethod
from_buffer
(*args, **kwargs)¶
-
classmethod
from_file
(filename, shared, size)¶
-
get_device
()¶
-
half
()¶ Casts this storage to half type
-
int
()¶ Casts this storage to int type
-
property
is_cuda
¶
-
is_pinned
()¶
-
is_sparse
= False¶
-
long
()¶ Casts this storage to long type
-
nbytes
()¶
-
pickle_storage_type
()¶
-
pin_memory
()¶ Coppies the storage to pinned memory, if it’s not already pinned.
-
resize_
(size)¶
Moves the storage to shared memory.
This is a no-op for storages already in shared memory and for CUDA storages, which do not need to be moved for sharing across processes. Storages in shared memory cannot be resized.
Returns: self
-
short
()¶ Casts this storage to short type
-
size
()¶
-
tolist
()¶ Returns a list containing the elements of this storage
-
type
(dtype=None, non_blocking=False)¶
-
-
class
torch.
QInt8Storage
(*args, **kwargs)[source]¶ -
bfloat16
()¶ Casts this storage to bfloat16 type
-
bool
()¶ Casts this storage to bool type
-
byte
()¶ Casts this storage to byte type
-
char
()¶ Casts this storage to char type
-
clone
()¶ Returns a copy of this storage
-
complex_double
()¶ Casts this storage to complex double type
-
complex_float
()¶ Casts this storage to complex float type
-
copy_
(source, non_blocking=None)¶
-
cpu
()¶ Returns a CPU copy of this storage if it’s not already on the CPU
-
cuda
(device=None, non_blocking=False, **kwargs)¶
-
data_ptr
()¶
-
property
device
¶
-
double
()¶ Casts this storage to double type
-
element_size
()¶
-
fill_
(value)¶
-
float
()¶ Casts this storage to float type
-
classmethod
from_buffer
(*args, **kwargs)¶
-
classmethod
from_file
(filename, shared, size)¶
-
get_device
()¶
-
half
()¶ Casts this storage to half type
-
int
()¶ Casts this storage to int type
-
property
is_cuda
¶
-
is_pinned
()¶
-
is_sparse
= False¶
-
long
()¶ Casts this storage to long type
-
nbytes
()¶
-
pickle_storage_type
()¶
-
pin_memory
()¶ Coppies the storage to pinned memory, if it’s not already pinned.
-
resize_
(size)¶
Moves the storage to shared memory.
This is a no-op for storages already in shared memory and for CUDA storages, which do not need to be moved for sharing across processes. Storages in shared memory cannot be resized.
Returns: self
-
short
()¶ Casts this storage to short type
-
size
()¶
-
tolist
()¶ Returns a list containing the elements of this storage
-
type
(dtype=None, non_blocking=False)¶
-
-
class
torch.
QInt32Storage
(*args, **kwargs)[source]¶ -
bfloat16
()¶ Casts this storage to bfloat16 type
-
bool
()¶ Casts this storage to bool type
-
byte
()¶ Casts this storage to byte type
-
char
()¶ Casts this storage to char type
-
clone
()¶ Returns a copy of this storage
-
complex_double
()¶ Casts this storage to complex double type
-
complex_float
()¶ Casts this storage to complex float type
-
copy_
(source, non_blocking=None)¶
-
cpu
()¶ Returns a CPU copy of this storage if it’s not already on the CPU
-
cuda
(device=None, non_blocking=False, **kwargs)¶
-
data_ptr
()¶
-
property
device
¶
-
double
()¶ Casts this storage to double type
-
element_size
()¶
-
fill_
(value)¶
-
float
()¶ Casts this storage to float type
-
classmethod
from_buffer
(*args, **kwargs)¶
-
classmethod
from_file
(filename, shared, size)¶
-
get_device
()¶
-
half
()¶ Casts this storage to half type
-
int
()¶ Casts this storage to int type
-
property
is_cuda
¶
-
is_pinned
()¶
-
is_sparse
= False¶
-
long
()¶ Casts this storage to long type
-
nbytes
()¶
-
pickle_storage_type
()¶
-
pin_memory
()¶ Coppies the storage to pinned memory, if it’s not already pinned.
-
resize_
(size)¶
Moves the storage to shared memory.
This is a no-op for storages already in shared memory and for CUDA storages, which do not need to be moved for sharing across processes. Storages in shared memory cannot be resized.
Returns: self
-
short
()¶ Casts this storage to short type
-
size
()¶
-
tolist
()¶ Returns a list containing the elements of this storage
-
type
(dtype=None, non_blocking=False)¶
-
-
class
torch.
QUInt4x2Storage
(*args, **kwargs)[source]¶ -
bfloat16
()¶ Casts this storage to bfloat16 type
-
bool
()¶ Casts this storage to bool type
-
byte
()¶ Casts this storage to byte type
-
char
()¶ Casts this storage to char type
-
clone
()¶ Returns a copy of this storage
-
complex_double
()¶ Casts this storage to complex double type
-
complex_float
()¶ Casts this storage to complex float type
-
copy_
(source, non_blocking=None)¶
-
cpu
()¶ Returns a CPU copy of this storage if it’s not already on the CPU
-
cuda
(device=None, non_blocking=False, **kwargs)¶
-
data_ptr
()¶
-
property
device
¶
-
double
()¶ Casts this storage to double type
-
element_size
()¶
-
fill_
(value)¶
-
float
()¶ Casts this storage to float type
-
classmethod
from_buffer
(*args, **kwargs)¶
-
classmethod
from_file
(filename, shared, size)¶
-
get_device
()¶
-
half
()¶ Casts this storage to half type
-
int
()¶ Casts this storage to int type
-
property
is_cuda
¶
-
is_pinned
()¶
-
is_sparse
= False¶
-
long
()¶ Casts this storage to long type
-
nbytes
()¶
-
pickle_storage_type
()¶
-
pin_memory
()¶ Coppies the storage to pinned memory, if it’s not already pinned.
-
resize_
(size)¶
Moves the storage to shared memory.
This is a no-op for storages already in shared memory and for CUDA storages, which do not need to be moved for sharing across processes. Storages in shared memory cannot be resized.
Returns: self
-
short
()¶ Casts this storage to short type
-
size
()¶
-
tolist
()¶ Returns a list containing the elements of this storage
-
type
(dtype=None, non_blocking=False)¶
-