Supervised training without using label + quantization #773
Description
Hi! I wanted to reduce the size of the pre-model (.bin or .vec file, several GB for these files), I started to look at the quantization method proposed by FastText. I tried to quantize a .bin file, however it was not possible because it only works for supervised models. Therefore I use the supervised training method with cc.en.300.vec file in input and quantize the model obtained from the training. The size was greatly reduced (from 4,4 GB to ~300MB) and it seems to work like the fullsized "raw" model. The quantized model is working, I'm able to retrieve logical results from a user input.
But how is it possible to use the .vec file to train through the supervised method without using label ?
I thought : supervised method = text classification and unsupervised method = word representation
Can someone enlighten me ?