MITB Banner
Picture of Pavan Kandru

Pavan Kandru

AI enthusiast with a flair for NLP. I love playing with exotic data.

What is Haystack for Neural Question Answering

Haystack is a python framework for developing End to End question answering systems. It provides a flexible way to use the latest NLP models to solve several QA tasks in real-world settings with huge data collections.

Guide to Perceiver: A Scalable Transformer-based Model

Perceiver is a transformer-based model that uses both cross attention and self-attention layers to generate representations of multimodal data. A latent array is used to extract information from the input byte array using top-down or feedback processing

Guide To THiNC: A Refreshing Functional Take On Deep Learning

THiNC is a lightweight DL framework that makes model composition facile. It’s various enticing advantages like Shape inference, concise model representation, effortless debugging and awesome config system, makes this a recommendable choice of framework.

How To Detect Objects With Detection Transformers?

DETR(Detection Transformer) is an end to end object detection model that does object classification and localization i.e boundary box detection. It is a simple encoder-decoderTransformer with a novel loss function that allows us to formulate the complex object detection problem as a set prediction problem.

What is Apple’s Quant for Neural Networks Quantization

Quantization is the process of mapping the high precision values (a large set of possible values) to low precision values(a smaller set of possible values). Quantization can be done on both weights and activations of a model.

What is Transformer XL?

Transformer XL is a Transformer model that allows us to model long range dependencies while not disrupting the temporal coherence.

Guide to Bayesian Optimization Using BoTorch

BoTorch is a library built on top of PyTorch for Bayesian Optimization. It combines Monte-Carlo (MC) acquisition functions, a novel sample average approximation optimization approach, auto-differentiation, and variance reduction techniques.

Guide to XLNet for Language Understanding

XLnet is an extension of the Transformer-XL model. It learns bidirectional contexts using an autoregressive method. Let’s first understand the shortcomings of the BERT model so that we can better understand the XLNet Architecture. Let’s see how BERT learns from data.

Subscribe to our Newsletter

Subscribe to Our Newsletter

The Belamy, our weekly Newsletter is a rage. Just enter your email below.