Blog

In awe at the scale of these tensors – a gentle introduction to Unit-Scaled Maximal Update Parametrization
Together with Graphcore, we recently developed u-μP as a new paradigm to parametrize neural networks in terms of width and depth. Our approach combines μP, developed by G. Yang et. al., with Unit Scaling, a concept introduced by Graphcore.

Words don’t come easy (… to LLMs): Universal Text-Encoding for dynamic, multi-lingual alphabets revolutionizing efficiency and effectiveness for LLM training and inference
The remarkable advancements of Large Language Models (LLMs) frequently capture attention as they become valuable collaborators in daily situations, all while progressing towards breakthroughs beyond simple language completion.

Introducing Pharia-1-LLM: transparent and compliant
We are pleased to announce our new foundation model family that includes Pharia-1-LLM-7B-control and Pharia-1-LLM-7B-control-aligned, now publicly available under the Open Aleph License, which explicitly allows for non-commercial research and educational use.

Open-sourcing Codebase Scaling for Non-commercial Research
Aleph Alpha's model training codebase Scaling is publicly available under the Open Aleph License, which explicitly allows for non-commercial research and educational use. Scaling was used to develop our concurrently released new models Pharia-1-LLM-control and Pharia-1-LLM-control-aligned.

Quality Diversity through AI Feedback
Language models carry implicit distributional biases based on their training data, which can reinforce existing norms. In this work, we take one step towards addressing the challenge of unwanted biases by enabling language models to return outputs with a broader spectrum of attribute traits, specified by a user. This is achieved by asking language models to evaluate and modify their outputs.

Luminous Performance Benchmarks
The research compares Luminous to the models from GPT-3 and ChatGPT developer OpenAI, among others. The scientific comparison included tasks related to text classification, evaluation, and generation, as well as answering questions about specific text contents. The result is impressive – with Luminous, a European AI language model is, for the first time, on par with the world's leading AI language models, while being much more efficient.