Digital authenticity: how to spot AI-generated content

In today’s digital age, we often question whether we can trust images, videos, or texts. Tracing the source of information is becoming more and more difficult. Generative AI accelerates the creation of content at an incredible pace. Images and audio files that once required a skilled artist can now be generated by AI models in a matter of seconds. Models like OpenAI’s Sora can even produce high-quality videos!

This technology offers both opportunities and risks. On the one hand, it speeds up creative processes, but on the other hand, it can be misused for malicious purposes, such as phishing attacks or creating deceptively real deepfake videos. So how can we ensure that the content shared online is genuine?

Digital watermarks: invisible protection for content

Digital watermarks are one solution that helps verify the origin of images, videos, or audio files. These patterns are invisible to the human eye but can be detected by algorithms even after minor changes, like compressing or cropping an image, and are difficult to remove. They are primarily used to protect copyright.

However, applying watermarks to text is way more difficult because text has less redundancy than pixels in images. A related method is to insert small but visible errors into the original content. Google Maps, for instance, uses this method with fictional streets – if these streets appear in a copy, it signals copyright infringement.

Digital signatures: security through cryptography

Digital signatures are based on asymmetric cryptography. This means that the content is signed with a private key that only the creator possesses. Others can verify the authenticity of the content using a public key. Even the smallest alteration to the content invalidates the signature, making it nearly impossible to forge. Digital signatures already ensure transparency in online communication, for example through the HTTPS protocol for secure browsing.

In a world where all digital content would be protected by signatures, the origin and authenticity of any piece of media could be easily verified. For example, you could confirm who took a photo, when, and where. An initiative pushing this forward is the Coalition for Content Provenance and Authenticity (C2PA), which is developing technical standards to apply digital signatures to media and document its origin. Unlike watermarks, signatures are not permanently embedded into the content itself and can be removed without altering the material. In an ideal scenario, everyone would use digital signatures – then, missing signatures would raise doubts about the trustworthiness of the content.

GenAI detectors: AI vs. AI

GenAI detectors provide another way to recognize generated content. AI models are algorithms that leave behind certain patterns, such as specific wording or sentence structures. Other AI models can detect these. Tools like GPTZero can already identify with high accuracy whether a text originates from a generative AI model like ChatGPT or Gemini. While these detectors are not perfect yet, they provide an initial indication.

What does this mean for users?

Of all the options, digital signatures offer the strongest protection because they work across all types of content and are based on cryptographic methods. It will be interesting to see if projects like C2PA can establish trusted standards. Still, different measures may be needed depending on the purpose of ensuring the trustworthiness of digital content.
In addition to technological solutions, critical thinking remains one of the best tools for navigating the information age. The amount of available information is constantly growing; therefore, it is important to critically question, verify, and be aware of the capabilities of generative AI models.

For a more comprehensive article, check out the CONTACT Research Blog.

Digitalization for the High Seas

The sun is shining in Hamburg, and the mild autumn air is in motion – even though I am perfectly equipped for rainy weather. In early October, shipbuilders from around the world gather in a conference hotel near the harbor for the CADMATIC Digital Wave Forum. The user meeting invites participants to experience CADMATIC’s CAD application for shipbuilding firsthand and to learn about current trends, product innovations, and new developments. The highlight: CADMATIC Wave, an integrated CAD/PLM solution specifically designed for shipbuilding and jointly developed by CADMATIC and CONTACT.

Model visualization simplifies data retrieval and collaboration

After our first coffee, we slowly make our way into the conference hall. The morning is filled with numbers and facts around CADMATIC’s digitalization strategy. In the afternoon, our Managing Director Maximilian Zachries presents CADMATIC Wave to the 200 participants. As we demonstrate the first functionalities of the integrated Product Data Management (PDM), some attendees quickly pull out their phones to snap a photo of the feature. I am somewhat excited – now it’s official. Now we also need the data model. And that isn’t quite so simple.

Cadmatic's Atte Peltola introduces the audience to Cadmatic Wave

CADMATIC’s Atte Peltola presents CADMATIC Wave. (© CADMATIC)

The resounding call for a data model for shipbuilding carries me through the three days in Hamburg. During my conversations with industry colleagues, it becomes evident that the information required and generated in the shipbuilding process must be able to be mapped within the model. Model-centric is the magic word: the ship’s geometry is visualized, including equipment, fittings, and logistics. Information can then be retrieved and added via the specific parts of the model. Model visualizations provide a shared and intuitive view of the ship for all involved trades, significantly simplifying information retrieval. This enhances the efficiency of engineering activities and collaboration, also with partners.

Basing a data model on ship geometry is challenging

Engaged in a discussion with a research associate from the Norwegian University of Science and Technology (NTNU), we stumble upon a question: Is the geometry model even suitable for generating a generic product structure for data storage in the PDM? After all, as a placeholder in a data model, there are quite a few locations in such a ship. And let me put it this way: data models are typically organized along the processes in product creation, not the geometry of a ship model. I am curious to see how we will solve this challenge in CADMATIC Wave.

The evening event takes place on the Cap San Diego, a museum ship in the Hamburg harbor. The rustic flair of a ship’s belly and the lavish buffet create a cozy atmosphere for lively conversations. We talk about life in Finland and Norway and the difference between information and data management. The evening ends stormy and rainy, and I finally put my rain gear to good use and return to the hotel dry and warm.

SEUS brings European shipbuilding to a new efficiency level

At the CADMATIC Digital Wave Forum, I also meet my consortium partners from the Smart European Shipbuilding (SEUS) project for the first time. Among them are representatives from NTNU and CADMATIC, as well as employees from two shipyards, the Norwegian Ulstein Group and the Spanish Astilleros Gondan SA. SEUS is an EU-funded research project with the goal of developing an integrated CAD and PLM solution for shipbuilding. This endeavor goes way beyond the functionalities we develop in CADMATIC Wave. For instance, we aim to incorporate knowledge management and utilize AI for searching within product data.

In this context, the broad positioning of our research department, CONTACT Research, works to our advantage. Our focus areas include not only Digital Lifecycle Management, where we conduct research on digitalization strategies for various industries, but also Artificial Intelligence. The AI product data search we aim to implement in SEUS allows us to bring our self-declared motto to life: “Bringing artificial intelligence into the engineering domains.”

As three days in Hamburg come to an end, three strong impressions remain:

  1. It is necessary to design an abstract data model for shipbuilding. One that contains the modules of a ship and yet can be customized to fit the specific needs of any shipbuilder. This data model must be closely linked to the development process.
  2. Personal exchange and meeting each other face to face have been an enriching experience for me in this new work area. This positive feeling motivates me for my future work in the SEUS project.
  3. In Hamburg, rain gear is a must.

Big, bigger, giant. The rise of giant AI models

The evolution of language models in the field of NLP (Natural Language Processing) has led to huge leaps in the accuracy of these models for specific tasks, especially since 2019, but also in the number and scope of the capabilities themselves. As an example, the GPT-2 and GPT-3 language models released with much media hype by OpenAI are now available for commercial use and have amazing capabilities both in type, scope, and accuracy, which I will discuss in another blog post. This was achieved in the case of GPT-3 by training using a model with 750 billion parameters on a data set of 570 GB. These are jaw-dropping values.

The larger the models, the higher the cost

However, the costs of training these models are also gigantic: Taking only the stated compute costs 1 for a complete training run, the total amount for training GPT-3 is 10 million USD 2, 3. In addition, there are further costs for pre-testing, storage, commodity costs for deployment, etc., which are likely to be in a similar amount. Over the past few years, the trend of building larger and larger models has been consistent, adding about an order of magnitude each year, i.e., the models are 10x larger than the year before.

Size of NLP models from 2018-2022. Parameter sizes are plotted logarithmically in units of billions. The red line represents the average growth: approx. 10-20 times larger models per year 2.

The next model of OpenAI GPT-4 is supposed to have about 100 trillion parameters (100 x 1012 ). For comparison, the human brain has about 100 billion neurons (100 x 109) which is 1000 times less. The theoretical basis for this gigantism is based on studies which show a clear scaling behavior between model size and performance 4. According to these studies, the so-called loss – a measure for the error of the predictions of the models – decreases by 1, if the model becomes 10 times larger. However, this only works if the computing power and the amount of training are also scaled upwards.

In addition to the enormous amounts of energy required to calculate these models and the associated CO2 footprint, which is assuming worrying proportions, there are direct economic consequences: Apparently, not only smaller companies cannot afford the cost of training such models, but also larger corporations are likely to balk at costs of $10 million, or $100 million or more in the future. Not to mention the necessary infrastructure and staffing for such an endeavor.

Monopoly position of the big players

This has a direct impact on availability: while the smaller models are now open source until the end of 2019 and can be freely accessed via specialized providers, this no longer applies to the larger models from around the end of 2020 (the appearance of GPT-2). OpenAI, for example, offers a commercialized API and only grants access through an approval process. On the one hand, this is convenient for developing applications with these NLP models, as the work of hosting and administration is eliminated; on the other hand, the barrier to entry for competitors in this market is so steep that essentially the super-big AI companies participate there: Google with OpenAI, Microsoft with Deepmind, and Alibaba.

The consequences of these monopoly positions of the leading AI companies are, as with every monopoly, pricing models without alternatives and rigid business practices. However, the capabilities of the current large language models such as GPT-3 and Megatron Turing NLG are already so impressive that it is foreseeable that in 10 years every company will probably need access to the current models for the most varied applications. Another problem is that the origin of the models from the American or Chinese area brings a large bias into the models, which on the one hand is clearly expressed in the fact that English or Chinese is the language with which the models work best. On the other hand, the training datasets that come from these cultural areas bring with them the very cultural tendencies from these spaces, so it is to be expected that other regions of the world will be underrepresented and continue to fall behind..

What can be done?

In my opinion, it is important to keep a careful eye on the development and to be more active in shaping the development of AI in the European area. In any case, a greater effort is needed to avoid dependence on monopolized AI providers in the long term. It is perhaps conceivable to involve national computing centers or research alliances that, united with companies, train and commercialize their own models and form a counterweight to American or Chinese companies. The next 10 years will be decisive here.

1 See here in section D as well as compute costs per GPU e.g. on Google Cloud approx. 1USD/hour for an NVIDIA V100
2 Calculation approach: V100 = 7 TFLOPs = 7 10^12 / s, 3.14 10^23 flops => 3.14 10^23/7×10^12 / 3600 = 10^7 hours = 10 million USD, details of the calculation and research of the parameters here.
3 see also here for comparison graph with older data.
4 see arxiv and Deepmind