Responsible Technology: Examining ChatGPT in Context
Category Machine Learning Tuesday - June 6 2023, 16:23 UTC - 1 year ago ChatGPT is a search engine that has been met with strong reactions due to how it interacts with users and the data it collects. Brit Ross Winthereik suggest that to assess responsible technologies, we need to understand it in context, analyze its effects, experiment with its capabilities, and check that it's values align with the same values of the organization, school, sector, or country.
The ChatGPT search engine has been met with reactions ranging from great enthusiasm to deep concern. At DTU, Professor Brit Ross Winthereik encourages us to always approach new technology openly and analytically to assess how it may be used responsibly.Why do you think people react so strongly to ChatGPT? .
I think one reason is that ChatGPT interacts with us in a different way than the search engines we know by personalizing its response in impeccable sentences that can be mistaken for reasoning. Some see it as controversial because it uses data that may be freely available on the Internet but is not intended for corporate profit-making. The company behind ChatGPT has broken a social contract by exploiting something in the public domain for its own gain. At the same time, user data is collected in a different way than in the search engines we already know, because, as something new, you give something of yourself in the form of your attachment to ideas or thoughts that you share precisely because of the way ChatGPT interacts with you. We also don't know if it's designed to affect the user in specific ways—for example if it tends to be racist and misogynistic, just like many other machine learning-based software systems, because they magnify predominant elements in the data on which they are trained.
Instead of focusing blindly on individual products, you encourage us to generally demand technology to be responsible. But how do you define 'responsible technology'? .
Whether a technology is responsible depends entirely on the interaction between what it's created for and how it's used. Does it meet the goals set, or does it work completely differently in practice? A manufacturer may aim to create responsible technology, but you cannot unilaterally label technologies as either 'responsible' or 'irresponsible'. This makes it very important to closely monitor specific effects, so we, for example, discover if technologies have inappropriate consequences, despite good intentions.
How do you assess whether a technology is responsible? The first step is to describe the technology in context. How is it part of the bigger context of infrastructures, business models, and sets of cultural values? An irresponsible approach would be to say, "Oh, now there's a technology that's going to revolutionize the world, so we'd better react," and then rush to use it or ban it without examining it further. Step number two is to analyze its effects: What happens in practice? Is the technology delivering on its promises? If not, what else does it do? And what requirements must be met for it to be a good technology? The third step is to experiment with the technology. What is the limit of its capabilities? Does it exclude someone or something? It's also important to assess whether the values that the technology represents align with the values of the organization, school, sector, or country in question. If not, you can say no and choose not to promote it or impose strict restrictions on its use.
In terms of public digital infrastructure—which is my field of research—it's about ensuring that the technology supports the society we want.
Share