In this module we build a small Tensor
in C, along the lines of torch.Tensor
or numpy.ndarray
. The current code implements a simple 1-dimensional float tensor that we can access and slice. We get to see that the tensor object maintains both a Storage
that holds the 1-dimensional data as it is in physical memory, and a View
over that memory that has some start, end, and stride. This allows us to efficiently slice into a Tensor without creating any additional memory, because the Storage
is re-used, while the View
is updated to reflect the new start, end, and stride. We then get to see how we can wrap our C tensor into a Python module, just like PyTorch and numpy do.
The source code of the 1D Tensor is in tensor1d.h and tensor1d.c. You can compile and run this simply as:
gcc -Wall -O3 tensor1d.c -o tensor1d
./tensor1d
The code contains both the Tensor
class, and also a short int main
that just has a toy example. We can now wrap up this C code into a Python module so we can access it there. For that, compile it as a shared library:
gcc -O3 -shared -fPIC -o libtensor1d.so tensor1d.c
This writes a libtensor1d.so
shared library that we can load from Python using the cffi library, which you can see in the tensor1d.py file. We can then use this in Python simply like:
import tensor1d
# 1D tensor of [0, 1, 2, ..., 19]
t = tensor1d.arange(20)
# getitem / setitem functionality
print(t[3]) # prints 3.0
t[-1] = 100 # sets the last element to 100.0
# slicing, prints [5, 7, 9, 11, 13]
print(t[5:15:2])
# slice of a slice works ok! prints [9, 11, 13]
# (note how the end range is oob and gets cropped)
print(t[5:15:2][2:7])
Finally the tests use pytest and can be found in test_tensor1d.py. You can run this as pytest test_tensor1d.py
.
It is well worth understanding this topic because you can get fairly fancy with torch tensors and you have to be careful and aware of the memory underlying your code, when we're creating new storage or just a new view, functions that may or may not only accept "contiguous" tensors. Another pitfall is when you e.g. create a small slice of a big tensor, assuming that somehow the big tensor will be garbage collected, but in reality the big tensor will still be around because the small slice is just a view over the big tensor's storage. The same would be true of our own tensor here.
Actual production-grade tensors like torch.Tensor
have a lot more functionality we won't cover. You can have different dtype
not just float, different device
, different layout
, and tensors can be quantized, encrypted, etc etc.
TODOs:
- bring our own implementation closer to
torch.Tensor
- implement a few simple ops like add, multiply, etc.
- make tests better
- implement 2D tensor, where we have to start worrying about 2D shapes/strides
- implement broadcasting for 2D tensor
Good related resources:
MIT