Attached: 1 image
**Eine Hausaufgabe**
Eine Freundin schickte mir das Bild im Anhang zu. Das war eine Aufgabe für die Tochter in der 7. Klasse des lokalen Gymnasiums:
**Absolutismus vor der Haustür**
PRÄSENTATION:
Informiere dich über die Porzellandmanufaktur Kupfermühle Flensburg (1687), die Fayence Manufaktur Schleswig (1755) und die Fayence Manufaktur Rendsburg (1764).
Wähle eine Manufaktur aus und erstelle eine kleine Collage, die die Geschichte, die Produkte und Arbeitsweise sowie die Bedeutung der Manufaktur für die Region darstellt.
Was erinnert heutzutage noch an die Manufakturen aus dem 17./18. Jahrhundert?
Die Tochter kam zu Ihr, da sie nix zu Porzellanherstellung in Flensburg finden konnte. Und sie sollte sich doch darüber informieren. Das kann ich sehr gut nachvollziehen, denn Porzellan in Flensburg im 17. Jahrhundert wäre eine echte Sensation von Weltrang:
Am 15. Januar 1708 gelang es dem Apotheker Johann Friedrich Böttger und Ehrenfried Walther von Tschirnhaus in der Jungfernbastei der Festung Dresden, das erste europäische Hartporzellan zu erzeugen.
Quelle: [https://de.wikipedia.org/wiki/Porzellan#Die_Erfindung_des_europ%C3%A4ischen_Hartporzellans](https://de.wikipedia.org/wiki/Porzellan#Die_Erfindung_des_europ%C3%A4ischen_Hartporzellans)
Fayencen, ja die kann es geben, aber kein Porzellan ist zu der Zeit dokumentiert. Fayence ist Ton (porös) mit Glasur (dann wasserdicht), 2x gebrannt (erst der Ton und dann die Glasur) und dickwandig. Porzellan ist eine besondere Mischung von Kaolin, Feldspat und Quarz, mit einem 1300ºC Brennprozeß, gesintert, ohne Glasur wasserdicht und kann so dünn geformt werden, daß es durchscheinend ist.
Die Manufakturen in Rendsburg und Schleswig erzeugen Fayencen, an der Grenze zur Entdeckung der Porzellanherstellung, haben meines Wissens aber auch nie Porzellan hergestellt.
Schon von der Logik her, macht es wenig Sinn, dass sich eine Kupfermühle mit dem Thema beschäftigt. Der Overlap zwischen beiden Produktionstechniken ist eher gering. Auch die [Geschichte der Mühle](https://de.wikipedia.org/wiki/Kupfermühle) gibt wenig Anlass an eine solche Verwendung zu glauben.
Die Frage macht wenig Sinn und ich fragte mich an der Stelle, ob da nicht eine Halluzination in der Aufgabe gelandet ist. Der Schreibfehler im Text spricht eher dagegen.
Interesant ist in dem Kontext, dass Schülerinnen und Schüler durchaus Texte mit Bezug auf eine "Porzellanmanufaktur Kupfermühle Flensburg" abgaben und diese auch Bilder von dort erzeugten keramischen Produkten enthielten. Hier steht durchaus zu erwarten, dass diese Antworten mit der KI erzeugt wurden und dort Halluzinationen einfach zu weiteren Halluzinationen geführt haben.
Mein Text hier soll kein Dissen der Lehrerin sein. Ich fände es toll, wenn man genau daraus ein "teachable moment" macht, der vor allem auf zwei Punkte abzielt:
- Es ist absolut legitim die Annahmen einer Aufgabe in Frage zu stellen.
- Eine KI kann falsche Ausgaben erzeugen. Die Wahrscheinlichkeit dafür steigt enorm an, wenn man falsche Annahmen in diese hineinsteckt.
From Bytes to Biases: Investigating the Cultural Self-Perception of Large Language Models - Wolfgang Messner, Tatum Greene, Josephine Matalone, 2025
Large language models (LLMs) are able to engage in natural-sounding conversations with humans, showcasing unprecedented capabilities for information retrieval a...
If AI is so good at coding … where are the open source contributions?
You can hardly get online these days without hearing some AI booster talk about how AI coding is going to replace human programmers. AI code is absolutely up to production quality! Also, you’re all…
Kritische Sicherheitslücke in Microsoft 365 Copilot zeigt Risiko von KI-Agenten
Der KI-Agent von M365 konnte per E-Mail und ohne Mausklick zur Freigabe sensibler Informationen verführt werden. Microsoft hat die Lücke jetzt geschlossen.
🧠🤖 Atari 2600 Pulls Off the Upset!!
It started as a simple experiment: pit ChatGPT against the Atari 2600’s chess engine (via Stella emulator) and see what happens. I figured it would be a lighthearted stroll down retro memory lane.
What actually happened?
ChatGPT got absolutely wrecked on the beginner level. This was after a conversation we had regarding the history of AI in Chess which led to it volunteering to play Atari Chess. It wanted to find out how quickly it could beat a game that only thinks 1-2 moves ahead on a 1.19 MHz CPU.
Despite being given a baseline board layout to identify pieces, ChatGPT confused rooks for bishops, missed pawn forks, and repeatedly lost track of where pieces were — first blaming the Atari icons as too abstract to recognize, then faring no better even after switching to standard chess notation. It made enough blunders to get laughed out of a 3rd grade chess club
Meanwhile, Atari’s humble 8-bit engine just did its thing. No language model. No flash. Just brute-force board evaluation and 1977 stubbornness.
For 90 minutes, I had to stop it from making awful moves and correct its board awareness multiple times per turn. It kept promising it would improve “if we just started over.” Eventually, even ChatGPT knew it was beat — and conceded with its head hung low.
👉 Have you played Atari today? ChatGPT wishes it hadn't.
#AI #Chess #Atari2600 #ChatGPT #RetroGaming #HumblingExperience #OpenAI OpenAI ChatGPT Sam Altman Logan Kilpatrick Brian Madden
Christopher Kyba 🇨🇦🇪🇺 (@skyglowberlin@fediscience.org)
"Information" from #ChatGPT when prompted to explain why Berlin is brighter than Chicago in one browser window, compared to when promoted to explain why Chicago is brighter than Berlin in a different browser's window.
This is why I hate when journalists write about "hallucination".
Every single text it generates is just a made up plausible combination of letters. It's just a variant of the idea of a thousand monkeys typing on a thousand typewriters that's more efficient than the monkeys at generating plausible text.
Generative AI runs on gambling addiction — just one more prompt, bro!
You’ll have noticed how previously normal people start acting like addicts to their favourite generative AI and shout at you like you’re trying to take their cocaine away. Matthias Döpm…
Meta sichert sich langfristig Atomstrom für Rechenzentren
Ein fast 40 Jahre altes Atomkraftwerk in den USA bekommt durch den KI-Boom ein neues Leben: Der Meta-Konzern hat sich mit dem Betreiber auf einen langfristigen Versorgungsvertrag geeinigt.
Kris on Recommenders, AI and TikTok (@isotopp@infosec.exchange)
Attached: 1 image
The EU is currently congratulating itself because it managed to get a hashtag banned on TikTok in relatively little time.
---
LLMs encode the meanings of terms as vectors along many semantic dimensions in a semantic space ("latent space"). A concept, then, is a position in that space with a certain diameter — a kind of fuzziness or vagueness.
When I type something into ChatGPT or a recommender system, the input is broken down into tokens, and these tokens are mapped to such vectors.
“I want pizza” becomes:
```
["I", "want", "pizza", "."]
```
The tokens are then internally mapped to embeddings, like:
```
“cat” → [0.24, -1.12, 0.58, …]
“dog” → [0.22, -1.09, 0.60, …]
```
That is, a list of numbers (often normalized between -1 and 1). But usually there are far more dimensions than shown here — an embedding typically has thousands of dimensions.
The latent space — the semantic space — is self-organizing. That happens during training. We don’t know what each dimension in the space represents.
The encoding has meaning. When look at the vectors for "man" and "woman" and for "king" and "queen", we can substract "man" from "woman" and "king" from "queen" and compare the difference vectors. They are almost, but not quite the same – because the difference between these words to us is almost, but not quite the same, in meaning.
LLMs use these embeddings and their internal model to “compute the next output token.”
Recommender systems use such embeddings to compare vectors and find things that are similar to the thing we already have.
So a recommender learns everything that’s relevant to a user, and a modern recommender represents the user through a collection of vectors:
"Interested in travel, digital policy, databases, bikes."
These are all concepts that may also be near other concepts in the space.
At the same time, the recommender classifies content in the same space, and can find content that lies close to one of the user’s sub-interests — or content that’s new, but still compatible.
A modern recommender separates a user’s interests into distinct areas and can decide what the user is interested in right now — meaning, which of the various user interests is currently active. Then, this time, it might only serve database content, and next time only bike content.
A modern recommender will also deliberately serve content that almost — but not quite — matches the user’s interests, to test how wide the bubble is around the center of that interest vector. So a bike session might also include urbanism, city development, and other nearby topics, and the recommender will watch carefully to see what kind of response that triggers — refining its recommendations based on that feedback.
A modern recommender will also know where the available content clusters are and prioritize content that is both relevant to the user and performs well or has current production capacity. In other words, where user interest and available content overlap well.
And a modern recommender will reevaluate every twenty minutes (“Pomodoro”, or “method shift” in educational theory) and attempt to shift the theme — to test whether another known interest can be reactivated.
That’s how TikTok works.
You can ban a hashtag on TikTok (“#skinnytok”).
But as long as related concepts are marketable and socially accepted — or even demanded — that won’t prevent anything.
As soon as you browse categories like “model,” “weight loss,” “fitness,” or “slim,” TikTok will slowly and systematically pull you into the same region, and the end result will be the same.
The actual language, the meaning, is encoded in the tokens of the latent space of the model, not in the words that are used (or prohibited).
And the content density in the models coordinate system will gently push things into certain clusters. If you feed the system with the right interests, you will always drift – relatively quickly even – into the same neighborhood and then learn their current slang to get there with a single word.
No matter what the word actually is.
A similar example, using GenAI instead of a recommender:
"Draw a superheroine, an Amazon warrior that can fly and deflect bullets, running over a battlefield in the first world war."
These 21 words do not say "Wonder Woman", they do not even go near comics, DC, or similar things.
Yet they draw a thousand-dimensional hyberbubble in latent space, the totality of knowledge known to ChatGPT, and the end result leaves just one choice – produce this blatant copyright violation.
I can trigger content with intent, not even going near the keywords that would be associated with it.
This is how jailbreaks work in LLMs, and that is also how you jailbreak Tiktok bans.
Leonard Lin and Adam Lensenmayer have been working on [Shisa](https://shisa.ai/) for a while. They describe their latest release as "Japan's Highest Performing LLM". Shisa V2 405B is the highest-performing …
Poisoning training data: Russian propaganda for AI models
Die Analyse stammt von NewsGuard, einem privaten US-Unternehmen, das die Vertrauenswürdigkeit von Online-Medien bewertet. NewsGuard steht selbst in der Kritik, etwa wegen mangelnder Transparenz. Dennoch ist die Untersuchung relevant.
Denn sie arbeitet systematisch, benennt konkrete Beispiele und testet eine breite Palette an Chatbots unter kontrollierten Bedingungen. Die Ergebnisse lassen sich nachvollziehen.