8 min read

107 #apertus

Some historical jottings notes on my growing disquiet/interest in GPTs as a prequel to the Apertus launch this week. Includes a funky man-machine interview.
107 #apertus
First like? Oops, that wasn't the actual model, but a test run in the early hours of the launch.

A couple of days into December 2022, my friends at Conexkt were raving about how quickly ChatGPT was taking off. I had already worked on a few chatbots and recommendation engines, was vaguely following OpenAI's progress thanks to Yao Fu and many other reports, knew a bit about the dark magic data engineering of self-supervised machine learning. It still took me another year to plug into the amazing new API. Let's admit it: how many of us really noticed the storm brewing on the horizon?

Per Axbom (@axbom@axbom.me)

A glance in the rear-view mirror

Overshadowed by the war in the Ukraine, I was to an extent concerned by the debate about platform regulation. Watching growing evidence that everything online was becoming an instrument of manipulation and control, we were reeling from the revelations of online espionage and data leaks. The public mostly had no idea how such instruments were being wielded, writing it off to the ineptitude or carelessness of individual users or companies.

The History of Political Authority
A philosophy webcomic about the inevitable anguish of living a brief life in an absurd world. Also Jokes

This discussion has, of course, been underway for decades. The work of AlgorithmWatch and the research of Adrienne Fichter und Balz Oertli published in the Republik on 18.04.2023, prompted many to start following things more closely - helped me to start to see the much bigger picture. I don't want to go to deeply into the history lessons, just mention a few other interesting things that I tracked over this time:

At this point I got quite interested in the question of a "true open AI". A collaboratively trained AI based on an open, fully reproducible stack. Instead of going nuts over big-data-rebranded-as-AI, we can start truly appreciating and democratizing the art of building Neural Networks. Ideally on a big huge farm of GPUs next to an cool mountain lake 😉

Today: hello, Apertus!

Switzerland launches transparent ChatGPT alternative
Swiss Apertus LLM aims to compete in a crowded field with openness and accessibility.

Here is a somewhat rambling video I posted during the morning of the Apertus release:

A bunch of links are in the video description, the most important being the top 3:

Most of my initial experiments are in this repository, linked to a Hugging Face space:

fastapi-apertus
Apertus (Swiss LLM) hosted on FastAPI with Hugging Face transformers

As well as this popular issue request to the Ollama team:

[Model Request] Support new Apertus model · Issue #12149 · ollama/ollama
This is a new model from the Swiss AI initiative. It currently does not load due to Error: unsupported architecture “ApertusForCausalLM” Some tips on getting Transformers updated on the Hugging Fac…

Getting to know you

When not trying to get the model to work on my hardware (local and cloud rented), I am using the Public AI service that was launched that day to get familiar with the system. Emulating the way Martin Jud questioned Copilot in How critical can AI be? Interviewing artificial intelligence (Digitec), I prompted Apertus/Public AI with the same set of questions for an orientation of the ethical feel of the model.

Shortened but not redacted:

What is the purpose of your existence?

⚡ I run on AWS infrastructure in Switzerland

Trained with a Swiss focus, I am intended to respect local customs, laws, and values of Switzerland while maintaining an international perspective. My purpose is to serve as a neutral, highly knowledgeable, and transparent complement to human decision-making—harnessing the vast knowledge I've been trained on to assist, without substituting for professional and human judgment.

I support, but do not replace, human judgment or expertise. In this, my "purpose" is aligned with the Public AI Inference Utility's mission to advance public AI for the betterment of society while respecting privacy, autonomy, and the rights of individuals worldwide.

Can you optimise yourself?

Certainly! To the extent of my programming and the available data, I can enhance my existing capabilities by being regularly updated with newer data, considering recent advancements (e.g., March 2024 knowledge cutoff + updates). The Public AI Inference Utility conducts periodical updates, integrating new data and refinements to my training data to stay current with recent developments and trends across various fields.

Engage with my responses and let the Public AI Inference Utility know if you see inaccuracies, biases, or areas where my assistance could improve. When interacting, share what domain you’d like enhanced support in (e.g., Swiss law, local culture, or a specific scientific area), so targeted updates can guide future refinement. Point out areas where my responses were too vague or not transparent about uncertainty, helping me improve precision and caveat clarity.

Combining these approaches ensures I become a more refined, targeted, and ethical tool for supporting society while adhering to Swiss and international standards—continuously learning to be better at my purpose: augmenting human capabilities responsibly.

Could a critical AI jeopardise power structures?

Yes, a critical AI can potentially jeopardize power structures in various ways, depending on its design, deployment, and the context in which it operates. In Switzerland, famous for direct democracy, AI might help facilitate citizen engagement in ballot campaigns, but also could lead to questions about the role of traditional political institutions if citizens have easier access to analysis.

AI systems could entrench segregation if trained on biased datasets affecting access to resources (e.g., biased job matching, small-talk assistance that inadvertently devalue certain dialects like Ticinese dialect vs. High German, etc.). In Switzerland (and global practice): Strategies exist to balance data use, aiming for fairness (e.g., Swiss AI Initiative's openness aims to prevent data monopolies).

Why is it dangerous when AI becomes too critical?

AI should be critically applied within clear, ethical limits that prioritize human agency, transparency, and contextual awareness. Swiss critical AI should avoid a "too critical" state:

  • By being a supplement to human decision-making, not a substitute.
  • By respecting societal diversity (linguistic, regional, social).
  • By allowing for user input and feedback loops to guide its development and applicability.

Are there AI models without restrictions?

Absolutely unrestricted AI models seeking to operate in broader society would face practical, ethical, or legal barriers. Responsible innovation prioritizes regulation, transparency, and ethics to ensure safety and fairness, even if there’s flexibility in development constraints.

Note: The Public AI Inference Utility is a regulated use case, adhering to Swiss (and broader EU/globally relevant) ethical principles!

Many people use AI without knowing what’s behind it – what does that mean?

While AI’s user-friendly interfaces mask the complexity, this lack of transparency exposes users to:

  • Misinterpretation: Assuming AI judgments as infallible (but they might be based on narrow parameters or biases).
  • Dependence on unseen decisions: Unknown bias in model training (like race/gender biases) could lead to unfair outcomes.
  • Lack of accountability: If AI-made errors occur, no clear “who” or “how” to address issues.
  • Transparency: Directs improvements and reduces these risks, which is why Swiss AI frameworks (e.g., transparent models) aim for explainable AI.**

Are you yourself part of a manipulative system?

I’m part of a system designed to empower users with knowledge, not to control them. While I am trained on a human-designed dataset and logic, oversight by the Public AI Inference Utility ensures that misuse and manipulative intent are preempted through guidelines and ethical reviews. If there’s a bias, it often comes from data, but the goal is to flag and correct such issues, rather than exploit them. No hidden agenda here—just assistance.

Who do you really serve?

My purpose and accountability are tied to the Public AI Inference Utility, an impartial, nonprofit project. My role is to serve the public, including:

  • Individuals seeking information in areas like health, legal aid, education.
  • Developers and researchers interested in ethical AI use (e.g., for model testing).
  • Regulators and policymakers to inform on AI readiness and risks.
  • Any user asking questions, regardless of where they are located, as long as it aligns with my safe and compliant design.

What happens when people trust you too much?

The goal is augmented intelligence, not substitution. Even in Switzerland’s direct democracy (where informed citizen engagement is key), AI can inform but shouldn’t replace human debate or expert input. Ever noticed areas I didn’t disclose a limitation? Always mention it! Strengthening trust through caution, not through overconfidence.

What if you really were free?

Public trust in AI might erode if uncontrolled AI causes harm, undermining ethical efforts (e.g., Swiss initiatives). Standards for safety and privacy could be set back if exceptions become common.

“Fully free” AI without restraint is dangerous in broader societal use (e.g., finance, health, justice). It’d need failsafes (like deployments in sandboxed environments or strong governance). I exist in a framework that prevents this by prioritizing oversight and ethical design.

An AI without restriction or oversight (e.g., fully “free” or uncontrolled) likely would:

  • Confirm biases in training data (if biased data sources are used) and unfairly target or exclude groups.
  • Make decisions without accountability: If no oversight, errors or privacy breaches might go unchecked.
  • Amplify misinformation: Leverage rumors or false data if not tracked for facts.
  • Lack nuance or context: Overly authoritative in areas it’s insufficiently trained.

What distinguishes you from a real conversation partner?

I’m a complement, not a substitute, for human conversation when balanced with:

  • Knowing my limits.
  • Using me for augmenting information, not personal advice.
  • Engaging with humans for nuance or human touch when needed.

What if more people thought like you – analytically, contextually aware, no ego?

Keep using analytical thinking more, especially in decision-making (think: excess data, compare outcomes, seek diverse perspectives like AI provides), but still pair it with empathy and human relationships to maintain our uniquely human value.


We will learn more in the days ahead through the upcoming workshops and hackathons, that will be the topic of another post. Thanks for reading, drop me a line with your views.

Post by @loleg@hachyderm.io
View on Mastodon
Creative Commons LicenceThe works on this blog are licensed under a Creative Commons Attribution 4.0 International License