Partio alternatives and similar libraries
Based on the "Graphics" category.
Alternatively, view Partio alternatives based on common mentions on social networks and blogs.
-
bgfx
Cross-platform, graphics API agnostic, "Bring Your Own Engine/Framework" style rendering library. -
Atomic Game Engine
DISCONTINUED. The Atomic Game Engine is a multi-platform 2D and 3D engine with a consistent API in C++, C#, JavaScript, and TypeScript -
DirectXTK
The DirectX Tool Kit (aka DirectXTK) is a collection of helper classes for writing DirectX 11.x code in C++ -
Horde3D
Horde3D is a small 3D rendering and animation engine. It is written in an effort to create an engine being as lightweight and conceptually clean as possible. -
Visualization Library
Visualization Library is a C++ middleware for high-performance 2D and 3D graphics applications based on OpenGL 1.x-4.x supporting Windows, Linux and Mac OS X.
CodeRabbit: AI Code Reviews for Developers
* Code Quality Rankings and insights are calculated and provided by Lumnify.
They vary from L1 to L5 with "L5" being the highest.
Do you think we are missing an alternative of Partio or a related project?
README
Partio - A library for particle IO and manipulation
This is the initial source code release of partio a tool we used for particle reading/writing. It started out as an abstraction for the commonalities in particle models (i.e. accessing many attributes associated with an index or entity).
Super impatient building guide
# Install Location ~ adjust accordingly
prefix=$HOME/local
git clone https://github.com/wdas/partio.git
cd partio
make -j prefix=$prefix install
Getting Started
CMake is used to build the project, but we provide a top-level Makefile for convenience that takes care of all the steps.
See the Makefile for the user-tweakable variables and corresponding cmake options.
The typical usage for an installation into /usr/local
with a temporary staging directory of /tmp/stage
is:
make DESTDIR=/tmp/stage prefix=/usr/local install
Source code overview
src/
lib/ Library code (public API in root)
lib/core Core library (KDtree traversal, data representations)
lib/io Input/Output (Different file formats)
py/ SWIG based python bindings
doc/ Doxygen documentation and (the start of) a manual
tests/ Start of regression tests (I need more)
tools/ Useful tools
partconvert <input format> <output format>
partinfo <particle file>
partview <particle file>
Class Model
The goal of the library is to abstract the particle interface from the data representation. That is why Partio represents particles using three classes that inherit and provide more functionality
ParticlesInfo - Information about # of particles and attributes ParticlesData - Read only access to all particle data ParticlesDataMutable - Read/write access to all particle data
The functions used to get particle access are these:
readHeaders()
returns ParticlesInfo
reads only the minimum data necessary to get number of particles and
attributes
readCached()
returns ParticlesData
For multiple users in different threads using the same particle file
ParticlesData
create() and read()
returns ParticlesDataMutable
allows read/write access
Behind the scenes you could implement these classes however you like. Headers only representation is called core/ParticleHeader.{h,cpp}. Simple non-interleaved attributes is core/ParticleSimple.{h,cpp}.
Attribute Data Model
All particles have the same data attributes. They have the model that they are of three basic types with a count of how many scalar values they have.
VECTOR[3]
FLOAT[d]
INT[d]
VECTOR[3] and FLOAT[3] have the same data representations.
VECTOR[4] is invalid however FLOAT[4] is valid as is FLOAT[1...infinity]
This seems to encompass the most common file formats for particles
Iterating
There are multiple ways to access data in the API. Here are some tips
Use SIMD functions when possible prefer dataAsFloat(),data(arrayOfIndices) as opposed to data(int singleIndex) which accesses multiple pieces of data at once
Cache ParticleAttributes for quick access instead of calling attributeInfo() over a loop of particles
Use iterators to do linear operations over all particles They are much more optimized than both data() and the dataAsFloat or
Backends
Behind the scenes there are SimpleParticles, ParticleHeaders, and SimpleParticlesInterleaved. In the future I would like to write a disk-based cached back end that can dynamically only load the data that is necessary. create(), read() and readCached could be augmented to create different structures in these cases.
Readers/Writers
New readers and writers can be added in the io/ directory. You simply need to implement the interface ParticlesInfo, ParticlesData and ParticlesDataMutable (or as many as you need). Editing the io/readers.h to add prototypes and io/ParticleIO.cpp to add file extension bindings should be easy.
Building the python Package for PyPi
To the partio for python and publish it to we have to build it using docker and upload it to PyPi.
# build the docker
docker build -t partio:latest .
# run the build
docker run --rm -v $(pwd):/io partio:latest
# use twine to upload to pypi
twine upload dist/*
- Andrew Selle, Walt Disney Animation Studios