Data distillation for #LLM: gives same performances with 0.1% of data,
but on a limited focused evaluation afaik: does this approach give
models as generic as with lots of data??
Get help from #LLM in your command-line: gorilla-cli
A very nice project from Berkeley Univ!
52 cognitive #bias in French
Jeu de cartes sous licence CC-BY de 52 biais cognitifs en français.
New #PhD opportunity in Nancy France on continual learning of large language models:
Don't hesitate to contact us for further details!
Translating to English before #nlp processing works better than multi-lingual processing; self-translating also works better, although a bit less good, but the delta might decrease with scale.
And if you're looking for a good open translation #llm, NLLB-200 is recommended by the authors:
Also, SpQR improves over QLoRA with good scaling laws: