frugallydeep alternatives and similar libraries
Based on the "Artificial Intelligence" category.
Alternatively, view frugallydeep alternatives based on common mentions on social networks and blogs.

Eclipse Deeplearning4J
Suite of tools for deploying and training deep learning models using the JVM. Highlights include model import for keras, tensorflow, and onnx/pytorch, a modular and tiny c++ library for running math code and a java based math library on top of the core c++ library. Also includes samediff: a pytorch/tensorflow like library for running deep learning using automatic differentiation. 
Modern C++ framework for Symbolic Regression
Modern C++ framework for symbolic regression that uses genetic programming to explore a hypothesis space of possible mathematical expression. 
Evolving Objects
A templatebased, ANSIC++ evolutionary computation library which helps you to write your own stochastic optimization algorithms insanely fast. [LGPL]
InfluxDB  Power RealTime Data Analytics at Scale
* Code Quality Rankings and insights are calculated and provided by Lumnify.
They vary from L1 to L5 with "L5" being the highest.
Do you think we are missing an alternative of frugallydeep or a related project?
README
[logo](logo/fdeep.png)
frugallydeep
Use Keras models in C++ with ease
Table of contents
Introduction
Would you like to build/train a model using Keras/Python? And would you like to run the prediction (forward pass) on your model in C++ without linking your application against TensorFlow? Then frugallydeep is exactly for you.
frugallydeep
 is a small headeronly library written in modern and pure C++.
 is very easy to integrate and use.
 depends only on FunctionalPlus, Eigen and json  also headeronly libraries.
 supports inference (
model.predict
) not only for sequential models but also for computational graphs with a more complex topology, created with the functional API.  reimplements a (small) subset of TensorFlow, i.e., the operations needed to support prediction.
 results in a much smaller binary size than linking against TensorFlow.
 works outofthebox also when compiled into a 32bit executable. (Of course, 64 bit is fine too.)
 avoids temporarily allocating (potentially large chunks of) additional RAM during convolutions (by not materializing the im2col input matrix).
 utterly ignores even the most powerful GPU in your system and uses only one CPU core per prediction. ;)
 but is quite fast on one CPU core compared to TensorFlow, and you can run multiple predictions in parallel, thus utilizing as many CPUs as you like to improve the overall prediction throughput of your application/pipeline.
Supported layer types
Layer types typically used in image recognition/generation are supported, making many popular model architectures possible (see Performance section).
Add
,Concatenate
,Subtract
,Multiply
,Average
,Maximum
AveragePooling1D/2D
,GlobalAveragePooling1D/2D
Bidirectional
,TimeDistributed
,GRU
,LSTM
,CuDNNGRU
,CuDNNLSTM
Conv1D/2D
,SeparableConv2D
,DepthwiseConv2D
Cropping1D/2D
,ZeroPadding1D/2D
BatchNormalization
,Dense
,Flatten
,Normalization
Dropout
,AlphaDropout
,GaussianDropout
,GaussianNoise
SpatialDropout1D
,SpatialDropout2D
,SpatialDropout3D
RandomContrast
,RandomFlip
,RandomHeight
RandomRotation
,RandomTranslation
,RandomWidth
,RandomZoom
MaxPooling1D/2D
,GlobalMaxPooling1D/2D
ELU
,LeakyReLU
,ReLU
,SeLU
,PReLU
Sigmoid
,Softmax
,Softplus
,Tanh
Exponential
,GELU
,Softsign
,Rescaling
UpSampling1D/2D
Reshape
,Permute
,RepeatVector
Embedding
Also supported
 multiple inputs and outputs
 nested models
 residual connections
 shared layers
 variable input shapes
 arbitrary complex model architectures / computational graphs
 custom layers (by passing custom factory functions to
load_model
)
Currently not supported are the following:
ActivityRegularization
, AdditiveAttention
, Attention
, AveragePooling3D
,
CategoryEncoding
, CenterCrop
, Conv2DTranspose
([why](FAQ.md#whyareconv2dtransposelayersnotsupported)),
Conv3D
, ConvLSTM1D
, ConvLSTM2D
, Cropping3D
, Discretization
,
Dot
, GRUCell
, Hashing
,
IntegerLookup
, Lambda
([why](FAQ.md#whyarelambdalayersnotsupported)),
LayerNormalization
, LocallyConnected1D
, LocallyConnected2D
,
LSTMCell
, Masking
, MaxPooling3D
, Minimum
, MultiHeadAttention
,
RepeatVector
, Resizing
, RNN
, SimpleRNN
,
SimpleRNNCell
, StackedRNNCells
, StringLookup
, TextVectorization
,
ThresholdedReLU
, UnitNormalization
, Upsampling3D
, temporal
models
Usage
1) Use Keras/Python to build (model.compile(...)
), train (model.fit(...)
) and test (model.evaluate(...)
) your model as usual. Then save it to a single HDF5 file using model.save('....h5', include_optimizer=False)
. The image_data_format
in your model must be channels_last
, which is the default when using the TensorFlow backend. Models created with a different image_data_format
and other backends are not supported.
2) Now convert it to the frugallydeep file format with keras_export/convert_model.py
3) Finally load it in C++ (fdeep::load_model(...)
) and use model.predict(...)
to invoke a forward pass with your data.
The following minimal example shows the full workflow:
# create_model.py
import numpy as np
from tensorflow.keras.layers import Input, Dense
from tensorflow.keras.models import Model
inputs = Input(shape=(4,))
x = Dense(5, activation='relu')(inputs)
predictions = Dense(3, activation='softmax')(x)
model = Model(inputs=inputs, outputs=predictions)
model.compile(loss='categorical_crossentropy', optimizer='nadam')
model.fit(
np.asarray([[1, 2, 3, 4], [2, 3, 4, 5]]),
np.asarray([[1, 0, 0], [0, 0, 1]]), epochs=10)
model.save('keras_model.h5', include_optimizer=False)
python3 keras_export/convert_model.py keras_model.h5 fdeep_model.json
// main.cpp
#include <fdeep/fdeep.hpp>
int main()
{
const auto model = fdeep::load_model("fdeep_model.json");
const auto result = model.predict(
{fdeep::tensor(fdeep::tensor_shape(static_cast<std::size_t>(4)),
std::vector<float>{1, 2, 3, 4})});
std::cout << fdeep::show_tensors(result) << std::endl;
}
When using convert_model.py
a test case (input and corresponding output values) is generated automatically and saved along with your model. fdeep::load_model
runs this test to make sure the results of a forward pass in frugallydeep are the same as in Keras.
For more integration examples please have a look at the [FAQ](FAQ.md).
Performance
Below you can find the average durations of multiple consecutive forward passes for some popular models ran on a single core of an Intel Core i56600 CPU @ 3.30GHz. frugallydeep and TensorFlow were compiled (GCC ver. 7.1) with g++ O3 march=native
. The processes were started with CUDA_VISIBLE_DEVICES='' taskset cpulist 1 ...
to disable the GPU and to only allow usage of one CPU.
(see used [Dockerfile
](test/Dockerfile))
Model  Keras + TF  frugallydeep 

DenseNet121 
0.12 s  0.31 s 
DenseNet169 
0.14 s  0.38 s 
DenseNet201 
0.18 s  0.50 s 
EfficientNetB0 
0.12 s  0.11 s 
EfficientNetB1 
0.10 s  0.19 s 
EfficientNetB2 
0.12 s  0.25 s 
EfficientNetB3 
0.19 s  0.49 s 
EfficientNetB4 
0.36 s  1.12 s 
EfficientNetB5 
0.79 s  2.29 s 
EfficientNetB6 
1.34 s  4.52 s 
EfficientNetB7 
2.47 s  8.16 s 
EfficientNetV2B0 
0.07 s  0.09 s 
EfficientNetV2B1 
0.09 s  0.13 s 
EfficientNetV2B2 
0.11 s  0.18 s 
EfficientNetV2B3 
0.15 s  0.30 s 
EfficientNetV2L 
1.70 s  4.56 s 
EfficientNetV2M 
0.84 s  2.12 s 
EfficientNetV2S 
0.33 s  0.72 s 
InceptionV3 
0.17 s  0.29 s 
MobileNet 
0.05 s  0.07 s 
MobileNetV2 
0.05 s  0.08 s 
NASNetLarge 
0.85 s  2.35 s 
NASNetMobile 
0.09 s  0.14 s 
ResNet101 
0.23 s  0.41 s 
ResNet101V2 
0.21 s  0.36 s 
ResNet152 
0.32 s  0.59 s 
ResNet152V2 
0.30 s  0.54 s 
ResNet50 
0.14 s  0.25 s 
ResNet50V2 
0.12 s  0.20 s 
VGG16 
0.40 s  0.48 s 
VGG19 
0.49 s  0.59 s 
Xception 
0.25 s  0.55 s 
Requirements and Installation
 A C++14compatible compiler: Compilers from these versions on are fine: GCC 4.9, Clang 3.7 (libc++ 3.7) and Visual C++ 2015
 Python 3.7 or higher
 TensorFlow 2.10.0 and Keras 2.10.0 (These are the tested versions, but somewhat older ones might work too.)
Guides for different ways to install frugallydeep can be found in [INSTALL.md
](INSTALL.md).
FAQ
See [FAQ.md
](FAQ.md)
Disclaimer
The API of this library still might change in the future. If you have any suggestions, find errors, or want to give general feedback/criticism, I'd [love to hear from you](issues). Of course, [contributions](pulls) are also very welcome.
License
Distributed under the MIT License.
(See accompanying file [LICENSE
](LICENSE) or at
https://opensource.org/licenses/MIT)
*Note that all licence references and agreements mentioned in the frugallydeep README section above
are relevant to that project's source code only.