Programming Every Example: Lifting Pre-training Data Quality like Experts at Scale
Paper
•
2409.17115
•
Published
•
63
This is quantized version of gair-prox/web-doc-refining-lm created using llama.cpp
Web-doc-refining-lm is an adapted 0.3B-ProX model, fine-tuned for document level refining via program generation.
@article{zhou2024programming,
title={Programming Every Example: Lifting Pre-training Data Quality like Experts at Scale},
author={Zhou, Fan and Wang, Zengzhi and Liu, Qian and Li, Junlong and Liu, Pengfei},
journal={arXiv preprint arXiv:2409.17115},
year={2024}
}
2-bit
3-bit
4-bit
5-bit
6-bit
8-bit
Base model
gair-prox/RedPJ-ProX-0.3B