Skip to content

Nvdiffrast - Modular Primitives for High-Performance Differentiable Rendering

License

Notifications You must be signed in to change notification settings

flynnamy/nvdiffrast

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Nvdiffrast – Modular Primitives for High-Performance Differentiable Rendering

Teaser image

Modular Primitives for High-Performance Differentiable Rendering
Samuli Laine, Janne Hellsten, Tero Karras, Yeongho Seol, Jaakko Lehtinen, Timo Aila
http://arxiv.org/abs/2011.03277

Nvdiffrast is a PyTorch/TensorFlow library that provides high-performance primitive operations for rasterization-based differentiable rendering. Please refer to ☞☞ nvdiffrast documentation ☜☜ for more information.

Licenses

Copyright © 2020, NVIDIA Corporation. All rights reserved.

This work is made available under the Nvidia Source Code License.

For business inquiries, please contact [email protected]

We do not currently accept outside code contributions in the form of pull requests.

GLEW library redistributed under the Modified BSD License, the Mesa 3-D License (MIT) and the Khronos License (MIT). Environment map stored as part of samples/data/envphong.npz is derived from a Wave Engine sample material originally shared under MIT License. Mesh and texture stored as part of samples/data/earth.npz are derived from 3D Earth Photorealistic 2K model originally made available under TurboSquid 3D Model License.

Citation

@article{Laine2020diffrast,
  title   = {Modular Primitives for High-Performance Differentiable Rendering},
  author  = {Samuli Laine and Janne Hellsten and Tero Karras and Yeongho Seol and Jaakko Lehtinen and Timo Aila},
  journal = {ACM Transactions on Graphics},
  year    = {2020},
  volume  = {39},
  number  = {6}
}

About

Nvdiffrast - Modular Primitives for High-Performance Differentiable Rendering

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Languages

  • C 78.2%
  • Cuda 9.5%
  • C++ 8.9%
  • Python 3.2%
  • Other 0.2%