Main Topic 3 –  Identification of AI generated content

19 Jun 2024 12:15h - 13:15h

Table of contents

Disclaimer: This is not an official record of the session. The DiploAI system automatically generates these resources from the audiovisual recording. Resources are presented in their original format, as provided by the AI (e.g. including any spelling mistakes). The accuracy of these resources cannot be guaranteed.

Full session report

Experts debate the detection and impact of AI-generated content in society

In a dynamic session focused on the challenges of detecting AI-generated content, experts from various fields engaged in a comprehensive discussion that delved into the limitations of AI detection systems, the societal implications of AI, and the need for robust regulatory frameworks.

The session opened with a live demonstration that revealed the limitations of current AI detection systems. These systems were shown to be easily fooled, which raised serious concerns about their reliability. The demonstration set the stage for a broader conversation on the implications of AI-generated content and the challenges it poses.

Professor Paulius Pakutinskas, working with UNESCO, underscored the importance of addressing the social issues related to AI. He highlighted the misconception that AI-generated content could always be recognized by other AI systems, a notion that was debunked by the initial demonstration. Pakutinskas pointed to a potential technological arms race, with technologies evolving to create content on one side and to detect or disguise it on the other. He emphasized the role of regulation, such as the Artificial Intelligence Act, which requires content labelling and watermarking. Pakutinskas also stressed the need for society to adapt to an increasing volume of AI-generated content, predicting that it could constitute up to 90% by 2026.

Dr Laurens Naudts, from the AI Media and Democracy Lab at the University of Amsterdam, provided a legal perspective, discussing the impact of synthetic content on democratic values and the EU’s ability to protect citizens through regulation. He questioned the effectiveness of transparency as a policy solution, noting that malicious actors are unlikely to adhere to transparency when their intent is to spread misinformation. Naudts suggested that regulators should focus on the values at risk and the tools available to empower citizens against threats posed by AI-generated content.

Audience interventions added depth to the discussion, with one participant warning against the reliance on AI detectors, citing personal tests that showed their unreliability. He cautioned against research that might overstate the prevalence of AI-generated content, potentially leading to further issues. Another participant from the Swiss IGF highlighted the role of media education and the dangers of contributing to destabilisation by overemphasising fake content.

A significant concern raised during the session was the performance of AI systems with non-English and minority languages. The experts acknowledged the risk of cultural imperialism through technology and called for more inclusive AI models that respect and benefit local communities and cultures.

In their closing remarks, the speakers reiterated the importance of maintaining trust in traditional news media as a marker of quality information. They cautioned against overemphasising risks that could erode trust in democratic institutions and advocated for directing citizens to authentic and reliable information. They also discussed the need for a deeper understanding of the responsibilities associated with the power held by technology.

The session concluded with a consensus on the need for a multifaceted approach involving technical solutions, regulatory frameworks, and societal engagement to address the complex issues surrounding AI-generated content. Key observations included the recognition of the need for change within educational systems to prepare for a future where AI-generated content is prevalent, the importance of critical thinking, and the potential for AI to contribute positively to content creation, provided that harmful content is regulated. The discussion underscored the evolving role of media in verifying information and guiding the public towards trustworthy sources, as well as the importance of not exacerbating mistrust in the information space.

Session transcript

Aldan Creo:
Great. Hello. How are you, everyone? Well, it’s a pleasure to be able to have this session. I hope we’ll make it very interesting. But before we start, I’d just love to walk you through what are we going to do in the next 45 minutes. So if we get my screen here, I’ll explain to you how we’re going to structure the session. So that’s one of our participants. But what I’d love to be able to have here is – great. Here we are. Yes. So this is where we’re going to do the session. First, we’re going to have a live demonstration of AI detection systems. So we’ll see two systems in five minutes at the start. And then we’ll move on to some opening remarks by key participants. Three minutes each, three participants. Then we’ll have an open discussion. So that’s where you get to feed in with your ideas. 20 minutes, interventions of one to two minutes. And finally, we’ll close the session with some closing remarks. So that’s how it’s going to go. look like, but let’s start with the first of the points, which is the live demonstration of an AI detection system. And here, what you can see, I’m going to show you. So I’m going to, this is just like a tool to rewrite text in a way that they are not detected anymore. So I’m going to go to an AI text detector. So here I just pasted some text, this has been generated. So you’ll see now that when I run the detector, I will get an output and it will say this text here has been generated by AI, et cetera. Let’s just give it a second to calculate the probabilities. But when we get the output here, what you’ll see is that we’ll get this output in a second where it says that it is AI generated. Let’s just give it a second to think. And in the meanwhile, I’ll show you what we can do about this. We have a text here, but the thing, the problem is that it’s actually very easy to rewrite the text. This is another version of the text. You see that the system told us it’s AI generated, but now let’s rewrite the text, let’s change it. And let’s instead just paste a different version of the text and run the detector again. So what can we see here when we get the output? Well, what we will see is that the prediction changes. We are going to see an example of how can we go from something that is generated and classified as such to something that has been generated. but it’s not classified correctly. So the idea here is we can trick detectors of AI-generated text to get a prediction that has been written by a human. So reliability comes into question here. That’s an issue, but we are going to see another example of how this can become an issue. Here we went from something that was classified correctly to something that wasn’t classified correctly. We are going to do the same now, but now the prediction will be I just took an image from Eurodig here. This is a real image, so this is human. And we see that it’s classified as human. That’s good. That’s a good prediction. The problem is that I can do something that’s called glazing. Glazing means, well, it’s a technique that you can use to change images. And if I change the image and glaze the image, I actually get a different prediction. I can take an image, and I can make it look artificial. That’s also a problem. So this is where the reliability of systems comes into question. I hope that this is a bit informative on, well, what’s the landscape right now? How does it look like? And this is where we actually go into our key participants to get their thoughts, ideas, and opening remarks on the issue. So I’ll invite them one by one. First, we’re going to have Professor Paulius Pakutinskas. Sorry, I’m a bit bad with other languages. But you’ll get three minutes to do a short introduction about yourself and your thoughts about the issue. So go for it.

Paulius Pakutinskas:
OK. OK, so I’m Paulius Pakutinskas. I’m Professor. in law. So, I work with UNESCO. I’m UNESCO Chair on Artificial Intelligence and Emerging Technologies and Innovations for Society. So, that means I work on social issues related to artificial intelligence. And what we were discussing last sessions very correlated to this session. And I think it’s very good just to repeat what was said in the very, very end of last session on usage of AI and so on. So, here we go more a bit on technological side. And there was very good illustration that sometimes we can’t recognize AI in this way what we’d like to have. Because there was an illusion that everything what is created by AI could be recognized by AI. But you can see that it’s not really true. And I think there will be race of technologies. Some technologies will create some content. Doesn’t matter is it text or image or video. And some of them will try to detect and some will try to change text to do to be not detectable. But we need to understand that the life is not so simple. We just can’t sit on technologies. So, when technologies are not so good at this time stage, we have society and we have regulation. So, we have regulation, some requirements set in Artificial Intelligence Act on labeling, on watermarking of content. It will be challenging. But laws doesn’t care how you will do it. You need just to do it. So, that’s very important point. But the third one is the society. We need to be more adapted. We need to understand that there will be huge amount of AI generated content. There are some predictions or ideas that it will be 90% in next or 2026. 90% of content generated by AI. And we do not need to afraid of it. Some of them is very good. But the issue is when we talk about misimation, with some falsification of existing content, that is a topic which need to be regulated. So there are, let’s say, three pillows here. So technical possibilities, the society as such, as humans, we need to have critical thinking anyway. That’s very, very important for us. Not just to rely. And then we have regulation. So thank you.

Aldan Creo:
Great, thanks a lot for that, well, opening remark. We’ll now go to Dr. Naudts. So he’s online with us now. Please feel free to introduce yourself, a little bit what you do, et cetera.

Laurens Naudts:
Perhaps first a technical question. Does everybody hear me correctly and clearly?

Aldan Creo:
Yes. Yes.

Laurens Naudts:
Okay, so thank you for the introduction. My name is Laurens Naudts. I’m a postdoctoral researcher at the AI Media and Democracy Lab, an Institute of Information Law at the University of Amsterdam. And my research looks into how, well, my perspective is legal. And I try to investigate how within an era characterized by an increased presence of synthetic content, this might exert pressures on democratic values and kind of complementing the previous talk, looking into the EU’s regulations ability to protect citizens. Now, I will leave my closing statement for some positive outlooks and perhaps address some of the risks now. And we do see that with the emergence of general purpose models due to their scalability. as well as the democratization of these tools amongst a wider public and ease of use. There is indeed a risk that people’s exposure to artificial content and actors, whether they are static images, like the Pope wearing Balenciaga, or sustained conversations with bots on social media, people are increasingly unable to distinguish what exactly is true, what is false, what is authentic and inauthentic. And if we look at the conditions of what we typically associate with an inclusive democratic model, being able to meaningfully inform political and social opinions and being able to participate with your citizens, this inability to make this distinction kind of erodes those values of inclusivity and participation. And the AI Act recognizes these risks. Now, what is the policy solution? It’s label and transparency. And one of those, of course, within academia, one of the fallacies that we typically tackle is the fallacy of transparency. Now, transparency is, of course, a necessary component to protect citizens. But is it a sufficient condition? Because malicious actors are unlikely to be transparent if their purpose is to spread misinformation. And as mentioned earlier, artificial content, for example, news articles, co-edited, translated, summarized by AI can still be truthful content. So I think that from a regulatory perspective, we also need to kind of take a step back and see, OK, what values are exactly threatened? And what tools are available to empower citizens against those threats? And this, once we are able to define this, we can actually kind of double back. OK, what information do we need? So perhaps as an addendum to your title, it’s also identification as well as communication of AI content. But for what purpose?

Aldan Creo:
Great, thanks a lot. Well, I have some sort of bad news to give to you, which is that one of our speakers has had some logistical issues. So he’s still trying to arrive. So what we’ll do is we’ll go straight to the interventions open to everyone and hopefully he will arrive during the session. All right? So we’ll open the floor now, and I’d love to start by asking you in general do you have any comments to make or any interventions you’d like to have? Feel free to go for that and we’ll take it from there.

Audience:
Yon Erkut, Yon Erkut, University of Geneva. Thanks a lot for these demonstrations, and I would emphasize how very cautious we need to be with this kind of AI detectors. They don’t work. Don’t use them. Just ignore them. When you see a test of AI detectors, I took the one that came out first with a hundred percent reliability. I just asked JetGPT to generate a short text. It was detected as human. Then I took the first part of the US Constitution. It was detected as AI-generated. And then I took the first part of the speech of Dr. Martin Luther King, I have a dream, and it was also detected as AI-generated. No uncertainty. So just ignore them, and we are in danger of running into the same thing that we ran with the bots, the botometer discussion. There will be probably some research saying so much content is AI generated and it’s bad content or whatever, but don’t trust this. There are no reliable means to detect AI-generated content yet. We might have to introduce watermarketing, et cetera, but don’t trust those systems, they are scan.

Aldan Creo:
Thanks a lot, thanks a lot for your input. Actually, that brings up a very interesting point and I’d love to go to you, Mr. Naudts, for a question on that, which is that sometimes AI-generated content detectors can be used to punish people. So, for example, think of a high school student writes an essay and then it’s detected as generated. What should we do when it comes to equality, discrimination, what should be the approach that we take regarding content detectors and punishment of people in general?

Laurens Naudts:
I could, I mean, that’s a very difficult question to provide a concrete singular answer. Two, I think in general, I will go about it in two ways. One, the problem that you described is a problem that we have seen with artificial technologies in general, whether it is automated decision-making in welfare, that they disproportionately disadvantage marginalized communities and here potentially unfairly exclude people from getting a particular opportunity. This is here the case as well. What I think is perhaps important is first of all, that when we introduce technology and I second the intervention earlier that we would first need to establish the efficacy. Do they actually succeed in what they pretend to do or pretend to do, or are they AI snake oil as some might claim? And second, to be reflective also of the powers that one has when developing these tools. So if you say, okay, we can detect false content or AI content, or let’s say misinformation, what are the definition of those concepts? Because in defining what perhaps might be a punishable offense, you determine the conditions that will give somebody an opportunity, yes or no, or deprive somebody of an opportunity. That’s one way of looking at it, to be more cognizant of one’s own responsibilities, powers, whether it is individually, as a designer, or institutionally, or state-wise, as somebody who will use these technologies. And then a point that I couldn’t make earlier, is we also need to be careful on who can occupy the space in determining which tools will be used for AI content detection, and who will determine those conditions. If you look at a risk of misinformation, we know, for example, that state institutions might weaponize the notion of misinformation to silence political dissidents. Private parties have economic interests, standardization bodies, they’re often democratically opaque, or do not include civil society. So on the one hand, be reflective of one’s responsibilities, and what you might deprive of people when a technology is incorporated. And second, who can then occupy that space in making these conditions, and do they have any form of legitimacy?

Aldan Creo:
That’s great. You probably want to intervene.

Paulius Pakutinskas:
Yeah, I’ve seen pain in your question, because the same pain is for all universities and schools, because we have some specific ways of teaching and examining. The thesis is one of the forms. I had the same headache in the very beginning of ChatGPT, when it popped up. And I’ve seen works, what were really done by AI. And I tested a lot of these tools. And I have the same conclusion. It doesn’t work at all. And it’s very easy to falsify it. And we did another experiment. We asked one professor to write a text. She was typing just here. And we gave her the best of these AI solutions, detection solutions. And it was written that it is algorithm-generated text. So you can see that there are both mistakes possible from one to another side, especially when your language is learned, not native language. So you use some forms, very clear forms, and so on and so on. So that is not reliable. Maybe it’s more reliable when we talk about video, images, and so on and so on. But without marking of this content, we will not succeed. And we will not have very, very good results. But there was a good question raised, why we need to know it. Because in quite short time period, most of content will be somehow in some percentage generated by AI, because a lot of tools will be used. So then we need to find some specific rules how to detect what is harmful. Because if it’s not harmful, we have a lot of good ways to use. It’s very efficient, very, very fast. We will have a lot of good content. So that’s not a threat in general. So we need to find what issues we’d like to solve.

Aldan Creo:
Do you have any interventions from the online audience, maybe? Yes?

Audience:
Thank you so much. Jacques Beglinger, coordinator of the Swiss IGF, and I may just share with you one of the findings of our session on the same topic two weeks ago, and I think the most startling was, well, we concur with what Jörn just explained, don’t trust whatever mechanism. So it’s not a question of whether what you see or hear is directly fake or not, but it’s the source that counts. And we had an input from a reputable local newspaper who explained to us how even they have problems finding out, but they then, whenever they are unsure, as a reputable source, they do further digging. So in the end, it’s education, education on the media level, but also in the population. Lies were always there, and there’s nothing new. Just right now, it’s just crying out, crying for the wolf all over. And this was also the last finding, finding the hole in the messages of Swiss IGF, that too much just crying foul play may, in the end, add to more destabilization than is ever wanted. So even media need to cut back a little bit on telling what went wrong once again, and all these beautiful fake pictures, and just decrying them as false, in the end, adds to destabilization and not to better education.

Paulius Pakutinskas:
Maybe, yeah, maybe I can add a bit of work with fake reality and fake things like disinformation and so on. It’s a very simple and easy situation when we have black and white, so it’s just lie. But AI is capable to do it very soft way. So you just can’t understand that just very few details were changed, and that’s semi-truth. And you can go round and round and round, and after some time you will see that you are in absolutely artificial environment without any possibilities to understand that it was faken. Just step by step, there are really good examples we had here in Lithuania, in other countries, when there are some artificial media sources which are backed up by other sources, and when you’re double-checking, that’s right, because other sources are saying the same. And you can just make absolutely unreal reality. That’s the danger.

Aldan Creo:
I see we have a question from the online audience, yes.

Audience:
Thank you once more, and thanks for this interesting discourse. Probably, I might have to go back to the issue that I raised about the treaties and the multilateral instruments, and what we intend doing with them. And I want to link it with this discourse that we’re in, because the issue of deepfakes is a serious one. However, I think we are privileged to have many of us from higher education institutions, the ability of us, some of us are involved in supervision of masters and some doctorate students, and some involved in dealing with assignments. and these assignments are put in what is called Turnitin, which is an AI. When that comes out, we make some judgments. The question is, with all the discourse that we have been having now, truth about the truth, can we rely on such instruments? Remember, this instrument has the capability of making someone who has submitted a dissertation not to be able to proceed because of similarities. Can we then argue that those similarities are real similarities? That’s one. The second one, probably talking from the African context, what we have discovered is, if you look at generative AI and the LLMs, that informs whatever you are looking for. If the population regarding that information, I’ll talk about the cultural aspect of Africa, you will then find out that the response that you get is not helpful. I just want to use to be that diplomatic. Because the AI dips from LLMs that are available in its ecosystem. If the ecosystem itself is such that the information that is provided there, it’s not going to say, I can’t find it. Most importantly, it will just give you something. That takes me to my question there. Noting that we have young people and we have the ability, like we said yesterday, to define the world we want. Could it not be that is the time that we allow our young engineers, mushrooming engineers, mushrooming programmers to begin to determine? capabilities, speaking the military way, that would localize the LLMs. I know AI allows that, but allow for the development, for example, if you are looking at the University of Wales, development of LLMs around their instructional offerings. And such that then we are able to slowly develop the basis for our AI to be helpful, because there is a helpfulness within AI, but we need to manage the risk in terms of everything that is said. What is the take around that? Given that we have young brains that are around us, what’s their take in terms of all this? Because it’s, strategically speaking, it’s creating an opportunity to move into an environment where they develop capabilities, where they develop programs and codes that take us to another strategy in terms of everything.

Aldan Creo:
Thanks a lot. Thanks a lot for that very insightful comment. I’d love to ask you to try to stick to the two minutes possible. And having said that, I’d love to, well, just ask a question on an issue related to equality and inclusion, which is that, you know, LLMs, they are trained mainly in English, etc., a very specific dataset. So they provide much worse results when we go to languages like Lithuanian, for example, or, I don’t know, Spanish, etc. Detectors also fail much more when we have other languages that are not so common. So when we talk about languages that are in a minority, that brings a lot of interesting questions. And I’d love to hear your thoughts in general, but let’s start with Mr. Naudts and then I’ll go with Professor Pakutinskas.

Laurens Naudts:
I mean, first of all, thank you for the previous intervention. And I think it is a question that probably requires a more layered and nuanced answer that I’m able to provide in this particular talk, but feel free to reach out afterwards. But I think that especially when it comes to the creation of AI tools, it has been dominated primarily by a select few economic actors, which for the training of their LLM often depend upon extractive and exploitative data practices. We have seen open AI training their system, using people within Kenya for a low sum to filter out toxic content from their models. Without these models actually being to any benefit to them, because ultimately they function best for an English-speaking Western audience. So when it comes to the problem of how, and I think there should be more initiatives to respectfully engage with local communities, countries to develop AI models and systems for a purpose that benefits the communities to which they communicate or reach out in dialogue respectfully, such as that the ultimate product, an LLM that attacks content, is not extractive, nor to their detriment. So that would be a typical start. So I fully agree with this point. And I would welcome more initiatives in this space, because currently the economic realities and practices do not correspond with this ambition.

Aldan Creo:
Great, thanks a lot. Professor, your turn.

Paulius Pakutinskas:
So when we talk about small languages, small cultures, it’s a topic, a big topic. When we talk about official languages, that’s one topic. But we have countries where there are 200 of dialects, or dialects are like languages, like separate languages. So I think that is a matter of time. So most of the issues will be solved in some time period. Here in Lithuania we have some program. So it was financed by government, some grants were granted. In some countries people are working on these topics too. I think in some time period we will solve it. But it’s a wider question. It’s related to cultures, because the culture that’s called programmed in big LLMs are based on, let’s say, Anglo-Saxon and British culture. And it’s much more difficult to change, because it’s in the… But I think we need to change our understanding of AI and usage of AI in general. As I said, there is technical level, there is human level, and regulatory. For example, I am forcing my students to use AI tools. Any and all tools. And then we discuss how it works. And it’s not so easy. It’s not so easy. If somebody thinks that I will just ask to present an essay of 25 pages on a specific topic, no, you will receive trash. So it will be really, really poor essay. So you need to understand the topic, you need to work with topic, and then you can reach. And I’m not so naive that young generation will not use this powerful tool. They will use. So why to lie for ourselves that they will not use? The issue is to change our understanding and how we teach and how we examine. That’s very difficult for us, I understand, but you know, that is a change. So we need to change in ourself, in our educational systems, how we test, how we interview, how we interact with people by using both human brains and artificial intelligence.

Aldan Creo:
Great, I’d love to close with the audience. If someone has one last comment, potentially two, depending on how long it is, feel free to go ahead.

Audience:
Well, hi, my name is Diego. I’m from the Youth League program, and I’d like to share a concern and a question, if I may. So we talked about detecting texts generated by large language models, and I think we shared the notion that right now it’s extremely difficult, but it was said that perhaps for deepfakes or audiovisual material, it might be easier, but I’m not sure exactly about how easy it would be, because by design, deepfakes, this kind of technology works through what it’s called the adversarial network. That is, these AIs are trained to beat all the AIs that are supposed to detect fake content. By design, they’re supposed to evade this kind of detection. So right now, the technology is quite young, so it could be feasible, but in the long term, I don’t see how this could be applied. So I do not have the authority to assert this. with absolute certainty. I am no expert, so do point out if there’s something, if there’s a catch that’s beyond my understanding. But my question would be, should we abandon trust in audiovisual information altogether? And if so, what possible contingencies should it be to get trustful information, perhaps from the cryptographic field? Thank you.

Aldan Creo:
Great, thanks a lot for your intervention. We need to go to the closing remarks, but maybe if you want, you could also try during your two-minute closing remarks to address a little bit of the question that was just made. So we’ll start with Mr. Naudts, it’s online, and then we’ll move to the room here.

Laurens Naudts:
Yeah, thank you. And I think that as part of my closing remarks, I wanted to actually highlight one of the points just raised. I think that we should not abandon all trust and hope, and maybe then, as one point earlier made, we should also rely on perhaps legacy markers of quality, that they are traditional news media who have been a trustworthy source of information to kind of depend on. And this then brings me to the other point that I earlier made, and watch out who occupies a given space and can claim sudden legitimacy over a traditional field that they might have had no expertise of beyond. But the problem is, we should not overemphasize the risks and erode trust within people’s information space, because if every person departs from a position of lack of trust, yeah, then the democratic institutions themselves become completely undermined. So we should, in concert, also focus and channel attention to how we can direct citizens to authentic information that is trustworthy, that is reliable. and do also a positive effort in that particular space. That’s the first one and I hope that addresses the last point made. The second point, and it’s related, but it then also brings back the theme that was earlier discussed about inclusion and equality. I think we should also recognize that if we allow technology to become an important part of our society, we should also recognize how technology holds power. That power can be regulatory in who can regulate AI, but also economic who has access to the resources. To understand what type of responsibilities are associated with each power and how that might shape our future living environment. I think once we do that, we will be able to perhaps more targetly intervene to realize equality, also on a global level, rather than have a new form of cultural imperialism through technology.

Aldan Creo:
Thanks a lot.

Paulius Pakutinskas:
I think I will stick on the same what I said in the very beginning. First is human, so we need to change ourself. We need to improve our critical thinking. It’s very difficult. Always thought that if something is on video, so it’s truth. It’s not so anymore. We had such changes. We had changes with photos, you remember, before these revision programs, so we trusted that if it’s photo, it looks like true. We need to change our thinking here and we will do it. I think there are other measures like technical, so here will be a race. Who will win, we will see. But anyway, there will be big ideas to catch these fake images or texts. but maybe it’s successful or less successful. And another big part is regulation. And we have Artificial Intelligence Act where most of such topics can go to high penalties. Like for this level, up to 15 million or 3% of income. So, and there are other possible ways like regulation of propaganda and misinformation and so on, so on. So there are other tools. They are complicated. There are no just very simple ways how to do it. We need to do it in quite agile way. Another issue is that sometimes we can over-regulate and it could be not so good too. So we have a lot of new laws and regulations in EU level. And we can’t imagine how it will work, how they will interrelate, how market will take it, how business will take it, how citizens will take it. So it’s challenging time, but it’s interesting.

Aldan Creo:
Thanks a lot. We need to close the session sadly. It’s been really insightful to hear all your thoughts and comments. We’ve touched a lot of topics really from regulation, discrimination, the reliability of such systems, minorities, education, misinformation. And there’s a lot more that could be said. Sadly, we don’t have time, right? But I’d love to thank you all, the audience, the key participants. And also I’d love to ask you to please excuse the participant that couldn’t make it today just due to traffic, et cetera. Sadly, Mr. Katkus would have loved to be here with us. he couldn’t, but I’d love to extend the thanks to really everyone that participated today and also the organizing team and technical people. Thanks a lot. We’re closing the session. Thank you.

AC

Aldan Creo

Speech speed

151 words per minute

Speech length

1487 words

Speech time

590 secs

A

Audience

Speech speed

135 words per minute

Speech length

1213 words

Speech time

538 secs

LN

Laurens Naudts

Speech speed

155 words per minute

Speech length

1454 words

Speech time

563 secs

PP

Paulius Pakutinskas

Speech speed

148 words per minute

Speech length

1608 words

Speech time

654 secs