Skip to content

Book Review: The Tech Coup – How to Save Democracy from Silicon Valley (2024)

Schaake, Marietje (2024). The Tech Coup – How to Save Democracy from Silicon Valley. Princeton University Press, 336 pages. By Melanie Kolbe-Guyot It is safe to say that probably no other book should more be on your reading list this year than Marietje Schaake’s 2024 “The Tech Coup – How to Save Democracy from Silicon Valley”. (…)

The Less People Know About AI, the More They Like It

Do you believe in magic? Recent research reveals a ‘lower literacy-higher receptivity’ link, suggesting that the less people understand AI, the more they perceive its human-like capabilities as magical, making them more likely to use it. This quirky perspective on how people perceive AI could reveal new ways to communicate its benefits and risks to (…)

OpenAI used this subreddit to test AI persuasion

Interesting work from OpenAI, who are testing how good their models are at convincing people to change their minds. Currently, they’re running the tests only internally on pre-selected human testers. But who knows where this will eventually be used, and whether in the open or hidden? For that matter, what about the LLM-generated messages Meta (…)

Meta, X sign up to Euro Commish code of conduct on hate speech

While the ‘Code of conduct on countering illegal hate speech online’ that the European Commission included into the Digital Services Act (DSA) is work in progress, the fact that even companies such as Meta and X feel compelled to sign shows that regulations are far from the toothless tigers that they are often made out (…)

Roundtable of Visionaries on the Impact of AI on the Software Development Life Cycle

Friday, February 7th, 2025, 14h-17h, BC 410, EPFL Introduction Artificial Intelligence has the potential to revolutionize also software development and IT in general. To explore the implications of AI on these domains, we organize a roundtable discussion. The objective of this roundtable is to gather insights from visionaries and experts to understand the impact of (…)

Enshittification isn’t caused by venture capital

Here is an article, in Cory Doctorow’s signature style, discussing social networks and what drives them and what makes people leave or stay. I like specifically how he dissects the way the once-good services these platforms used to provide got untethered from the profits their creators and CEOs were chasing over the years. Towards the (…)

No, Trump didn’t make $50 billion from his memecoin

The awesome Molly White throws light upon how to calculate the market cap of a crypto coin. I still think that decentralized systems like blockchains are very useful in some cases. However, the run for the coin with the most money seems very sad to me, and not just because of all the investors who (…)

More Speech and Fewer Mistakes

Meta lays out in this blog post their rationale behind axing third-party fact checking and sweeping changes in content moderation on Facebook, Instagram and Threads. It is important to read this (or watch Mark Zuckerberg’s video) with recent company history in mind: Facebook’s failure to properly moderate content helped fan the flames in the Rohingya (…)

Get a PET for X-Mas

Privacy Enhancing Technologies, or PETs for short, is an umbrella term for a wide range of technologies and tools designed to protect our privacy online. You may not realize it, but you probably already use PETs on a daily basis. Some common examples [1] include HTTPS, securing connections between you and websites End-to-end encryption, ensuring (…)

Study suggests X turned right just in time for election season

This article discusses a study suggesting algorithmic bias favoring Republican-leaning content, and its owner Elon Musk’s posts in particular, on the social media platform X. The study further claims that this bias dates back to when Musk officially started supporting Donald Trump. While it is of course impossible to prove these allegations without access to (…)

Factory Update, Fall 2024

Welcome to the Factory Update for Fall 2024. Twice a year we take the time to present some of the projects we see coming out of our affiliated labs and give you a short summary of what we’ve been doing the past 12 months. Please also give us a short feedback on what you most (…)

FBI, CISA, and NSA reveal most exploited vulnerabilities of 2023

Interesting to see that 12 out of the 15 top vulnerabilities published by CISA, America’s cyber defense agency, are from 2023. Log4j2 from 2021 is also still in the list! So make sure that your systems are up-to-date with regard to these vulnerabilities, even if it’s not 0-days anymore.Bleeping Computer

C4DT Deepfakes Hands-on Workshop (for C4DT Partners only)

Following on the heels of our conference on “Deepfakes, Distrust and Disinformation: The Impact of AI on Elections and Public Perception”, which was held on October 1st 2024, C4DT proposes to shift the spotlight to the strategic and operational implications of deepfakes and disinformation for organizations. For our C4DT partners we are also hosting a hands-on workshop aimed at engineers, software developers and cybersecurity experts on Tuesday, 26th of November, which will allow the participants to develop skills and expertise in identifying and combating cyberattacks through deepfakes.

C4DT Roundtable on Deepfakes (for C4DT Partners only)

Following on the heels of our conference on “Deepfakes, Distrust and Disinformation: The Impact of AI on Elections and Public Perception”, which was held on October 1st 2024, C4DT proposes to shift the spotlight to the strategic and operational implications of deepfakes and disinformation for organizations. For our C4DT partners we are hosting a high-level roundtable for executives, senior managers and project managers on Tuesday, 19th of November, during which strategies to address the challenges posed by deepfakes, and collaboration opportunities and projects to counter them will be discussed.

Anomaly detection in dynamic networks

The temporal evolution of the structure of dynamic networks carries critical information about the development of complex systems in various applications, from biology to social networks. While this topic is of importance, the literature in network science, graph theory, or network machine learning, still lacks of relevant models for dynamic networks, proper metrics for comparing network structures, as well as scalable algorithms for anomaly detection. This project exactly aims at bridging these gaps.

ANEMONE: Analysis and improvement of LLM robustness

Large Language Models (LLMs) have gained widespread adoption for their ability to generate coherent text, and perform complex tasks. However, concerns around their safety such as biases, misinformation, and user data privacy have emerged. Using LLMs to automatically perform red-teaming has become a growing area of research. In this project, we aim to use techniques like prompt engineering or adversarial paraphrasing to force the victim LLM to generate drastically different, often undesirable responses.