Digital enterprises have become increasingly monopolistic, individualistic beyond narcissistic to the point of lamentably solipsistic, devastatingly careless of human potential, and environmentally wasteful.


Digital enterprises have become increasingly monopolistic, individualistic beyond narcissistic to the point of lamentably solipsistic, devastatingly careless of human potential, and environmentally wasteful.
We identified five key issues to prioritize when addressing gender equality: Reaching critical mass; Improving lived and living experience (maintaining critical mass); Selection and resource-allocation criteria; The misconception of merit in academia; Beyond gender equality.

If I knew what made for a successful academic career, I would have bottled and sold it a long time ago. There probably is no silver bullet, and with the organizational/individual duality, it is probably impossible to solve all of the problems or optimize all of the criteria, at all of the different levels, for all of the different people, all at the same time. Focus on those arenas that you can affect and inspire.

Generative artificial intelligence (AI) is rapidly transforming people’s access to and attitudes toward knowledge. It is an extremely powerful technology, but this transformation presents numerous social, environmental, political, and educational considerations.

Generation Z and Millennials face different types of insider threat, in three different dimensions of space: to resources, to citizenship, and to boundaries.

Technology has always been about more than simply a route to increased productivity and economic growth; technology also provides the opportunity to enhance, enrich, and empower—basically, to improve shared qualitative values or people’s quality of life (however that is measured). On the flip side, technology also provides the opportunity to develop and project organizational control, which itself can be weaponized to quantitatively determine human value as an asset to that organization, or to reinforce asymmetric power relationships.

How can local (grassroots) contributive justice be used as a driving force for the common good?

All the deep philosophical questions, starts the joke, were asked by the classical Greeks, and everything since then has been footnotes and comments in the margins, finishes the punchline.

The term “modern indentured servitude” did not originate with this workshop, but we hope that this special issue has highlighted many of the different shapes and processes it can take, some more insidious than others. We would like to think that, if each paper could talk, they would get up one after the other and say, “No, I’m Spartacus.” In these dark times, each of us needs the courage to be Spartacus.

It would be good if whenever a client connected to an http server, or indeed any app connected with a central server, the server responded with a corresponding acknowledgment of data, along the lines of “Before we begin our session this morning, I would like to acknowledge the traditional owner of the data which is being transferred, and respect rights to privacy, identity, location, attention and personhood.”

Functional democratic governance has five fundamental preconditions: civic dignity, confluent values, epistemic diversity, accessible education, and legitimate consent.

If it were possible to formulate laws involving vague predicates and adjudicate them, what would be the implications of such minimalist formulations for soft laws and even for “hard” laws? The possible implications are threefold: 1) does possibility imply desirability; 2) does possibility imply infallibility; and 3) does possibility imply accountability? The answer advanced here, to all three questions, is “no.”

Systems can be designed using methodologies like value-sensitive design, and operationalized, to produce socio-technical solutions to support or complement policies that address environmental sustainability, social justice, or public health. Such systems are then deployed in order to promote the public interest or enable users to act (individually and at scale) in a way that is in the public interest toward individual and communal empowerment.

Just as the “autonomous” in lethal autonomous weapons allows the military to dissemble over responsibility for their effects, there are civilian companies leveraging “AI” to exert control without responsibility.
And so we arrive at “trustworthy AI” because, of course, we are building systems that people should trust and if they don’t it’s their fault, so how can we make them do that, right? Or, we’ve built this amazing “AI” system that can drive your car for you but don’t blame us when it crashes because you should have been paying attention. Or, we built it, sure, but then it learned stuff and it’s not under our control anymore—the world is a complex place.

Understanding the societal trajectory induced by AI, and anticipating its directions so that we might apply it for achieving equity, is a sociological, ethical, legal, cultural, generational, educational, and political problem.

We can perhaps accept Weil’s starting premise of obligations as fundamental concepts, based on which we can also reasonably accept her assertion that “obligations … all stem, without exception, from the vital needs of the human being.”

With techno-feudalism, what is paid and permitted in a digital space is decided by asymmetric power, not mutual consent. Political approval for funding priorities, education programs and regulation all favor Big Tech.

Some collective behavior that supports sustainability entails some individual inconvenience: many small acts of environmental kindness require some thought, effort, or consideration.

Mega-platforms have, with the addition of one extra ingredient, combined lock-in and loyalty to create a grave, and perhaps unexpected, consequence. The extra ingredient is psychology; and the unexpected consequence is what might be called digital dependence.

Democracy itself is under (yet another) threat from deepfake videos … deepfake videos could be used to create compromising material of politicians: for example, the digitally-altered video2 of U.S. House of Representatives speaker Nancy Pelosi appearing to slur drunkenly was viewed millions of times and tweeted by the U.S. President, and although the video is demonstrably a hoax, the tweet remains undeleted.