Guide To THiNC: A Refreshing Functional Take On Deep Learning

THiNC is a lightweight DL framework that makes model composition facile. It’s various enticing advantages like Shape inference, concise model representation, effortless debugging and awesome config system, makes this a recommendable choice of framework.

THiNC is a deep learning framework that makes composing, configuring and deploying models easy. It provides a flexible yet simple approach to modelling by providing low-level abstractions of the training loop, evaluation loop etc. Moreover, it plays well with major deep learning frameworks like TensorFlow and PyTorch. The functional programming API of THiNC is fairly simple and elegant. It’s light weighted API makes THiNC a good option for quick prototyping and deployment of machine learning models.

Core Design Idea

Modularity is the most important aspect of any deep learning framework. The ability to build larger functions from smaller component functions is imperative to build DL Models. To provide this functionality is not as trivial as it seems. Each block of a DL Model should be capable of running a forward pass of activations and a backward pass of gradients. Several Frameworks choose several ways to deal with this requirement. THiNC uses a callback mechanism to solve this model composition problem. Each layer in the forward pass returns a backward callback function along with the activation result. The following example will make things clear.

Example: Let’s say we want a reduce sum layer followed by a RELU layer in a model. THiNC builds layers similar to the following functions.

AIM Daily XO

Join our editors every weekday evening as they steer you through the most significant news of the day, introduce you to fresh perspectives, and provide unexpected moments of joy
Your newsletter subscriptions are subject to AIM Privacy Policy and Terms and Conditions.
def reduce_sum(X: Floats3d) -> Tuple[Floats2d, Callable[[Floats2d], Floats3d]]:

    Y = X.sum(axis=1)

    X_shape = X.shape


    def backprop_reduce_sum(dY: Floats2d) -> Floats3d:

        dX = zeros(X_shape)

        dX += dY.reshape((dY.shape[0], 1, dY.shape[1]))

        return dX


    return Y, backprop_reduce_sum


def relu(inputs: Floats2d) -> Tuple[Floats2d, Callable[[Floats2d], Floats2d]]:

    mask = inputs >= 0

    def backprop_relu(d_outputs: Floats2d) -> Floats2d:

        return d_outputs * mask

    return inputs * mask, backprop_relu

Many layers of DL models require parameters to work. But passing in these parameters individually to the forward function along with input is extremely difficult to manage. It becomes quite messy with more number of parameters and layers. To avoid this parameters are encapsulated in a Model Class and are passed as input to the forward function. Unlike TensorFlow or PyTorch’s model class, this class won’t become different for different kinds of layers through inheritance. This model is the same for almost all types of layers, only the forward function which takes the model as input will differentiate between the layers.


As per the official Documentation THiNC offers the following features and abilities:

Download our Mobile App

  • Type-checking of your model definitions
  • Integrating your favourite DL frameworks seamlessly.
  • Functional Programming API.
  • Operator Overloading
  • Integrated Config system for exposing hyperparameters 
  • Choosing a GPU or CPU backend.
  • Supporting variable-length sequences
  • Low Abstraction Training Loop.

Let’s get started with THiNC.


 !pip install THiNC --pre
 #To use GPU we need to install cupy
 !pip install "cupy-cuda101" 

Classifier Model

To Build a model all the classes and functions must be imported from thin.api .Let’s build a simple FCNN with dropout to recognize the handwritten digits of MNIST dataset.

 from thinc.api import chain, Relu, Softmax
 n_hidden1, n_hidden2, n_hidden3 = 64,32,10
 dropout = 0.2
 model = chain(
     Relu(nO=n_hidden1, dropout=dropout),
     Relu(nO=n_hidden2, dropout=dropout), 
     Relu(nO=n_hidden3, dropout=dropout), 

Operator Overloading

Operator overloading can be used to build complex models in a concise manner.This FCNN model can be built in a single line by using operator overloading.

 with Model.define_operators({">>": chain}):
     model = Relu(nO=n_hidden1, dropout=dropout) >> Relu(nO=n_hidden2, dropout=dropout) >> Relu(nO=n_hidden3, dropout=dropout) >> Softmax() 

Model Initialization

Input and output shapes, along with all the missing shape information in the model can be inferred automatically by initializing the model with sample inputs.

 model.initialize(X=train_X[:5], Y=train_Y[:5])
 nI = model.get_dim("nI")
 nO = model.get_dim("nO")
 print(f"Initialized model with input dimension nI={nI} and output dimension nO={nO}") 

Training the Model

Next, we need to build a training loop. THiNC provides low-level abstractions for batching the data and shuffling it. These help in building custom training loops. The following lines are the backbone of training.

 for X, Y in tqdm(batches, leave=False):
             Yh, backprop = model.begin_update(X)
             backprop(Yh - Y)

We can add all sorts of functionality and build a generic training loop on top of these essential lines.

 def train_model(data, model, optimizer, n_iter, batch_size):
     (train_X, train_Y), (dev_X, dev_Y) = data
     indices = model.ops.xp.arange(train_X.shape[0], dtype="i")
     for i in range(n_iter):
         batches = model.ops.multibatch(batch_size, train_X, train_Y, shuffle=True)
         for X, Y in tqdm(batches, leave=False):
             Yh, backprop = model.begin_update(X)
             backprop(Yh - Y)
         # Evaluate and print progress
         correct = 0
         total = 0
         for X, Y in model.ops.multibatch(batch_size, dev_X, dev_Y):
             Yh = model.predict(X)
             correct += (Yh.argmax(axis=1) == Y.argmax(axis=1)).sum()
             total += Yh.shape[0]
         score = correct / total
         print(f" {i} {float(score):.3f}")
   Now we can define an optimizer and run the training.
 from THiNC.api import Adam, fix_random_seed
 optimizer = Adam(0.001)
 batch_size = 128
 n_iter = 10
 train_model(((train_X, train_Y), (dev_X, dev_Y)), model, optimizer, n_iter, batch_size)
 # We got 96.2% Validation Accuracy 

Compatibility with Other Frameworks

THiNC provides awesome wrappers to integrate with Tensorflow, PyTorch and MXNet.These wrappers help a lot when it comes to porting code amongst different frameworks. If some functionality you need is ready in a different framework, then you can use those special layers by wrapping them using the THiNC wrappers for that particular framework.

Example: Let’s implement the same FCNN as above using layers from both TensorFlow and PyTorch.

 from thinc.api import PyTorchWrapper, TensorFlowWrapper, chain, Linear, Adam
 from tensorflow.keras.layers import Dense, Dropout
 from tensorflow.keras.models import Sequential
 tf_model = Sequential()
 tf_model.add(Dense(64, activation="relu", input_shape=(784,)))
 import torch
 import torch.nn
 import torch.nn.functional as F
 class PyTorchModel(torch.nn.Module):
     def __init__(self, nO, nI, dropout):
         super(PyTorchModel, self).__init__()
         self.dropout1 = torch.nn.Dropout2d(dropout)
         self.fc1 = torch.nn.Linear(nI, nO,)
     def forward(self, x):
         x = F.relu(x)
         x = self.dropout1(x)
         x = self.fc1(x)
         x = F.relu(x)
         return x
 model = chain(
     PyTorchWrapper(PyTorchModel( 32, 64, dropout)),
     Relu(nO=10, dropout=dropout), 

THiNC is not yet completely developed and it is not advisable to use layers from multiple frameworks in production. These frameworks tend to hog the available GPU.

Config System

Configuration is an important aspect of product development. You need models to be manageable. We might often need to expose lots of model components as hyperparameters.

THiNC provides a great config system to do this.We can define trees of hyperparameters using simple json like structures. Then we can resolve these configurations and get the functions to run.Let’s create a config for the classifier model above.

 from thinc.api import Config, registry
 CONFIG = """
 n_hidden1 = 64
 n_hidden2 = 32
 n_hidden3 = 10
 dropout = 0.2
 @layers = "chain.v1"
 @layers = "Relu.v1"
 nO = ${hyper_params:n_hidden1}
 dropout = ${hyper_params:dropout}
 @layers = "Relu.v1"
 nO = ${hyper_params:n_hidden2}
 dropout = ${hyper_params:dropout}
 @layers = "Relu.v1"
 nO = ${hyper_params:n_hidden3}
 dropout = ${hyper_params:dropout}
 @layers = "Softmax.v1"
 @optimizers = "Adam.v1"
 @schedules = "warmup_linear.v1"
 initial_rate = 2e-5
 warmup_steps = 1000
 total_steps = 10000
 n_iter = 10
 batch_size = 128
 config = Config().from_str(CONFIG)

Here each block is separated by [<block name>].THiNC registry contains the definitions of standard functions we can get these definitions from the registry using 

  @<registry name> =”<function name string>”

Once we have the function in the succeeding lines we can add argument values to the function.

<argument-name>= value

The hierarchy of functions can be represented using the “.” operator



Once we have the config we can resolve it and train the model.

 loaded_config = registry.resolve(config)
 model = loaded_config["model"]
 optimizer = loaded_config["optimizer"]
 n_iter = loaded_config["training"]["n_iter"]
 batch_size = loaded_config["training"]["batch_size"]
 model.initialize(X=train_X[:5], Y=train_Y[:5])
 train_model(((train_X, train_Y), (dev_X, dev_Y)), model, optimizer, n_iter, batch_size) 

This unique config system allows us to expose lots of model details to the user without having to write complex code.


Although THiNC is a good approach to solve deep learning modelling problems, it introduces its own set of problems. Following are some of the disadvantages of THiNC

No Implicit Backward Pass of gradients – As discussed in the design idea section of the post, we need every block to return the activation output and a callback function for backpropagation. Unlike other frameworks THiNC doesn’t use reverse model auto differentiation, we need to explicitly code the backward loss function.

THiNC’s capability to integrate various frameworks is rudimentary. Its low-level abstraction makes it easier to make mistakes. 


THiNC is a lightweight DL framework that makes model composition facile. It’s various enticing advantages like Shape inference, concise model representation, effortless debugging and awesome config system, makes this a recommendable choice of framework.


Sign up for The Deep Learning Podcast

by Vijayalakshmi Anandan

The Deep Learning Curve is a technology-based podcast hosted by Vijayalakshmi Anandan - Video Presenter and Podcaster at Analytics India Magazine. This podcast is the narrator's journey of curiosity and discovery in the world of technology.

Pavan Kandru
AI enthusiast with a flair for NLP. I love playing with exotic data.

Our Upcoming Events

24th Mar, 2023 | Webinar
Women-in-Tech: Are you ready for the Techade

27-28th Apr, 2023 I Bangalore
Data Engineering Summit (DES) 2023

23 Jun, 2023 | Bangalore
MachineCon India 2023 [AI100 Awards]

21 Jul, 2023 | New York
MachineCon USA 2023 [AI100 Awards]

3 Ways to Join our Community

Telegram group

Discover special offers, top stories, upcoming events, and more.

Discord Server

Stay Connected with a larger ecosystem of data science and ML Professionals

Subscribe to our Daily newsletter

Get our daily awesome stories & videos in your inbox

Is Foxconn Conning India?

Most recently, Foxconn found itself embroiled in controversy when both Telangana and Karnataka governments simultaneously claimed Foxconn to have signed up for big investments in their respective states