Aleph Alpha Research

Pioneering Sovereign, Human-Centric AI

From innovation to real-world AI our mission-driven research builds scalable, interpretable, and customizable systems for diverse domains

Research for Real-World Impact

Our areas of research are defined based on our mission around sovereignty, human empowerment and responsibility. We extend what’s possible on a fundamental level, adding to the state-of-the-art to enable the most complex and critical AI use-cases.

Capable AI

We develop AI built for production environments that balance the needs of raw performance, scalability, and robustness – thus meeting the standards of industry leaders beyond proof-of-concept implementations.

We innovate foundational model designs and training approaches with the goal of continuously raising the bar on innovation or sovereignty. We care about compliant, efficient models focused on a wide range of languages and are working to advance foundation model architectures combined with integration patterns beyond chatbots.

Accessible

We make AI accessible for every organization even in data-scarce contexts such as low-resource languages, multimodality and specialised enterprise knowledge.

We develop methods to improve transparency, fairness and efficiency across languages and domains, allowing for better performance when less data is available or compute budget is restricted.

Transparent

Transparency is at the core of our mission. We ensure sovereignty for our customers by openly sharing how our technology works, from in-depth development publications – such as papers and blog posts – to the source-available approach of our stack.

Without transparency, organizations struggle to assess AI suitability and compliance, particularly in regulated environments.

Trustworthy

The black-box nature of GenAI makes it difficult for humans to take responsibility, especially in regulated sectors where a thorough understanding of model behavior is critical. We develop methods for inspecting, understanding, and validating responses, putting the human at the center.

Controllable

We design AI systems with safety and control at their core, empowering users to tailor outputs to their unique needs and preferences in real time.

Rejecting a one size fits all approach, we prioritize diversity of options in AI outputs, enabling seamless customization for each use-case and user. Our customers are in control of their distinct values, beliefs, cultural nuances, and stylistic choices – all without requiring extensive model retraining.

Interactive

We design AI systems not only for benchmarks but with human compatibility at their core, ensuring they align with users’ needs and expectations for seamless integration.

Through human-centred research, our technology adapts to real-world workflows and improves clarity and usability with innovative human-machine interaction patterns tailored to tasks beyond chatbots and questions that don’t have easy answers.

Innovation Highlights

T-Free via HAT

AI models typically rely on a limited set of subwords to process text. This rigid tokenization limits adaptability and efficiency. Our Hierarchical Autoregressive Transformer (HAT) changes the game. By combining character-level flexibility with word-level efficiency, it eliminates fixed vocabularies. This approach increases efficiency making AI more resilient to typos, and more adaptable to new languages or domains.

u-μP

Training AI models at scale is challenging. Hyperparameters that work for small models often fail at larger sizes, and low-precision training introduces instability. u-µP solves both problems. By combining Maximal Update Parametrization (µP) with Unit Scaling, it enables models to train efficiently at any size while keeping hyperparameters stable and transferable. This approach simplifies tuning, accelerates training, and allows models to work seamlessly in FP8 without complex rescaling. With u-µP, AI development becomes faster, more predictable, and more scalable.

T-Free via Trigrams

Tokenizers are essential for encoding text in LLMs, but they face significant limitations, including computational overhead, inefficient vocabulary use, and bias toward reference corpora. Trigram addresses these issues by embedding words through sparse activation patterns over character triplets (“trigrams”), eliminating the need for a reference corpus. This approach exploits morphological similarities and allows for strong compression of embedding layers, achieving competitive downstream performance with over 85% parameter reduction.

AtMan

Every result or decision of an AI system is based on patterns learned during training and knowledge accessed during inference. AtMan makes these patterns visible and usable. We can show the developer or user how evidence has caused AI outputs, which relationship needs to be added to improve AI systems, and where potentially contradicting observations may hint at a complex context. This transparency of the multidimensional context is crucial to optimize prompts and systems, guide UX and workflows, and for humans to take responsibility.

DTM

LLM’s growing size presents deployment challenges. Our innovative Divergent Token Metrics (DTMs) offer a solution by providing a nuanced evaluation of compressed LLMs, overcoming the limitations of traditional metrics. DTMs focus on token divergences, offering insights into the intricacies of model compression, especially when assessing individual components. This innovation ensures efficient deployment while maintaining high performance, making DTMs an essential tool for quantizing LLMs in practical applications.

MultiFusion

Text-to-image AI models are powerful but often limited – stuck in a single language and reliant on only text inputs. MultiFusion changes that. By seamlessly integrating pre-trained models, it enables AI to understand multilingual and multimodal prompts, combining text and images in any order. MultiFusion transfers capabilities from existing individual modules, making image generation more expressive, intuitive, and accessible. With this approach, users can create rich, detailed visuals using any mix of words and images, across multiple languages, for a truly global and flexible creative experience.

Magma

Vision-language modeling has traditionally relied on tailored pretraining objectives and labeled data, limiting its scalability and flexibility. MAGMA introduces a new approach by augmenting generative language models with visual inputs through adapter-based finetuning. This method keeps the language model’s weights unchanged, preserving its encyclopedic knowledge and in-context learning abilities. By leveraging a simple next-token prediction objective, MAGMA achieves state-of-the-art results on the OKVQA benchmark and performs competitively across various VL tasks, all while using significantly less data than existing methods.

"The goal is not to replace people, but to empower them by making design choices that give humans control over technology."

Ben Shneiderman, Human-Centered AI

Research Partnerships and Collaborations

We are partnering with established research organizations to develop cutting-edge AI innovations

In our collaboration with ETH Zurich we are hosting PhD and Post Doc positions focused on agentic systems.

Together with the TU Darmstadt we have established Lab 1141, a team of PhDs and Post Docs focused on making AI explainable, interpretable, and thereby ultimately understandable.

Engineers and researchers from Graphcore and Aleph Alpha Research worked together to co-optimize their respective technologies for pre-training, fine-tuning, and inference of next generation multi-modal language and vision models.

Join Our Open-Source Community

We welcome every contribution, value collaboration, and embrace constructive feedback. Join us to enhance our models, share ideas, and help shape the future of AI together.

Access our model card and share your feedback. Collaborate on improving model performance and creating new applications.

Explore our code, contribute to ongoing projects, or create your own forks. Join our GitHub community to help us build robust, transparent AI solutions.