Skip to content
This repository was archived by the owner on Mar 19, 2024. It is now read-only.
This repository was archived by the owner on Mar 19, 2024. It is now read-only.

Supervised training without using label + quantization #773

Open
@userFT

Description

@userFT

Hi! I wanted to reduce the size of the pre-model (.bin or .vec file, several GB for these files), I started to look at the quantization method proposed by FastText. I tried to quantize a .bin file, however it was not possible because it only works for supervised models. Therefore I use the supervised training method with cc.en.300.vec file in input and quantize the model obtained from the training. The size was greatly reduced (from 4,4 GB to ~300MB) and it seems to work like the fullsized "raw" model. The quantized model is working, I'm able to retrieve logical results from a user input.

But how is it possible to use the .vec file to train through the supervised method without using label ?
I thought : supervised method = text classification and unsupervised method = word representation

Can someone enlighten me ?

Metadata

Metadata

Assignees

No one assigned

    Labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions