- May 2023
-
dig.watch dig.watch
-
So disclosure of the data that’s used to train AI, disclosure of the model and how it performs and making sure that there’s continuous governance over these models.
Q: What are the main aspects of AI transparency?
-
the I P C A UN body
The closest analogy is with IPCC
-
So we have existing regulatory authorities in place who have been clear that they have the ability to regulate in their respective domains. A lot of the issues we’re talking about today span multiple domains, elections, and the like.
How to use existing governance and regulatory agencies?
-
Guardrails need to be in place.
Guardrails are increasing in 'lingo intensity'
-
Different rules for different risks.
Good slogan
-
the conception of the EU AI Act is very consistent with this concept of precision regulation where you’re regulating the use of the technology in context.
EU AI Act uses precise regulation of regulation AI in specific contexts.
-
we need to give policy makers and the world as a whole the tools to say, here’s the values and implement them.
Use SDGs as guardrails for AI.
-
that interaction with the world is very important.
-
constitutional AI
New concept?
-
a reasonable care standard.
Another vague concept. What is 'reasonable'? There will be a lot of job for AI-powered lawyers.
-
Thank you, Mr. Chairman and Senator Hawley for having this. I’m trying to find out how it is different than social media and learn from the mistakes we made with social media. The idea of not suing social media companies is to allow the internet to flourish. Because if I slander you you can sue me. If you’re a billboard company and you put up the slander, can you sue the billboard company? We said no. Basically, section 230 is being used by social media companies to high, to avoid liability for activity that other people generate. When they refuse to comply with their terms of use, a mother calls up the company and says, this app is being used to bully my child to death. You promise, in the terms of use, she would prevent bullying. And she calls three times, she gets no response, the child kills herself and they can’t sue. Do you all agree we don’t wanna do that again?
How to avoid repeating with AI governance what happened with Seciton 230 and social media governance?
-
And so the quality of the sort of overall news market is going to decline as we have more generated content by systems that aren’t actually reliable in the content they’re generated.
Risks for newsmarket.
-
the current version of GPT-4 ended to training in 2021.
2021 starts to being 'safety net' for OpenAI
-
And other countries are doing this, Australia and the like. And so my question is, when we already have a study by Northwestern predicting that one-third of the US newspapers are that roughly existed, two decades are gonna go, are gonna be gone by 2025, unless you start compensating for everything from book movies, books. Yes. but also news content. We’re gonna lose any realistic content producers. And so I’d like your response to that. And of course, there is an exemption for copyright in section two 30. But I think asking little newspapers to go out and sue all the time just can’t be the answer. They’re not gonna be able to keep up.
Q: How to protect newspapers and content producers?
-
When some of those fake ads. So that’s number one. Number two is the impact on intellectual property.
Two concers: fake adds and IPRs.
-
Sen. Marsha Blackburn (R-TN):
It is probably the most practical approach to AI governance. Senator from Tennessee asked many questions on the protection of copyright of musicians. Is Nashville endangered. The more we anchor AI governance questions into practical concerns of citizens, communities, and companies - the better AI governance we will have.
-
OpenAI Jukebox
Diplo team shoudl follow on this development.
-
And I wanna come to you on music and content creation, because we’ve got a lot of songwriters and artists. Yeah. And a, I think we have the best creative community on the face of the Earth. They’re in Tennessee, and they should be able to decide if their copyrighted songs and images are going to be used to train these models. And I’m concerned about OpenAI’s jukebox. It offers some renditions in the style of Garth Brooks, which suggests that OpenAI is trained on Garth Brooks songs. I went in this weekend and I said, write me a song that sounds like Garth Brooks. And it gave me a different version of Simple Man. So it’s interesting that it would do that. But you’re training it on these copyrighted songs, these mini files, these sound technologies. So as you do this, who owns the rights to that AI generated material and using your technology, could I remake a song, insert content from my favorite artist, and then own the creative right to that song?
Bring intellectual property into debate
-
that people own their virtual you.
People can own it only with 'bottom-up AI'
-
We’ve done it before with the IAEA.
Now IAEA comes as analogy, probably driven by nuclear power?
-
When you think about the energy costs alone, just for training these systems, it would not be a good model if every country has its own policies and each, for each jurisdiction, every company has to train another model.
It is naive view because AI is shaped by ethics and ethics is very 'local'. Yes, there are some global ethical principles: protect human life and dignity. But many other ethical rules are very 'local'.
-
need a cabinet level organization within the United States in order to address this.
Who can govern AI?
-
So we think that AI should be regulated at the point of risk, essentially, and that’s the point at which technology meets society.
Nice 'meeting' language
-
I can’t recall when we’ve had people representing large corporations or private sector entities come before us and plead with us to regulate them. In fact, many people in the Senate have base their careers on the opposite that the economy will thrive if government gets the hell out of the way. And what I’m hearing instead today is that ‘stop me before I innovate again’ message. And I’m just curious as to how we’re going to achieve this
Great point and strategic shift. It is 'Frankenstein moment'. Companies realised that they created something they cannot control.
-
an enterprise technology company, not consumer focused. S
It is an interesting distinction. However, technology developed by IBM will be used for consumer services.
-
hyper targeting of advertising is definitely going to come.
-
And we probably need scientists in there doing analysis in order to understand what the political influences of, for example, of these systems might be.
Markus tries to make case for 'scientists'. But, frankly speaking, how scientists can decide if AI should rely on book written in favour of republicans or democrats or, even more as AI develops with more sophistication, what 'weight' they should give to one or another source.
It is VERY dangerous to place ethical and political decisions in hands of scientists. It is also unfair towards them.
-
we don’t know what it’s trained on.
It is a good point. OpenAI is not transparent on datasets used for training GPT. But, the problem is that even if they inform us, the question will be who decided what datasets should be used for training.
-
If these large language models can, even now, based on the information we put into them quite accurately predict public opinion, you know, ahead of time. I mean, predict, it’s before you even ask the public these questions, what will happen when entities, whether it’s corporate entities or whether it’s governmental entities, or whether it’s campaigns or whether it’s foreign actors, take this survey information, these predictions about public opinion and then fine tune strategies to elicit certain responses, certain behavioral responses.
this is what worries politicians - how to win elections? They like 'to see' (use AI for their needs) but 'not to be seen' (use by somebody else. The main problem with political elites worldwide is that they may win elections with use of AI (or not), but the humanity is sliding into 'knowledge slavery' by AI.
-
large language models can indeed predict public opinion.
They can as they, for example, predict continuation of this debate in the political space.
-
so-called artificial general intelligence really will replace a large fraction of human jobs.
It is a good point. There won't be more work.
-
And the real question is over what time scale? Is it gonna be 10 years? Is it gonna be a hundred years?
It is a crucial question. One generation will be 'thrown under the bus' in transition. Generation of age 25-50 should 'fasten seat-belts'. They were educated in the 'old system' while they have to work in a very uncertain new economy.
-
that scientists be part of that process.
What should scientist do specifically? Can scientist judge if something is true? Who are scientists (e.g. do we refer to IT specialists)?
-
So I think the most important thing that we could be doing and can, and should be doing now, is to prepare the workforce of today and the workforce of tomorrow for partnering with AI technologies and using them. And we’ve been very involved for, for years now in doing that in focusing on skills-based hiring in educating for the skills of the future. Our skills build platform has 7 million learners and over a thousand courses worldwide focused on skills. And we’ve pledged to train 30 million individuals by 2030 in the skills that are needed for society today.
It is probably the only thing to do. But the problem remains that even re-skilling want be sufficient if we will need less human labour.
-
And so you see already people that are using GPT-4 to do their job much more efficiently by helping them with tasks. Now, GPT-4 will I think entirely automate away some jobs, and it will create new ones that we believe will be much better. This happens again, my understanding of the history of technology is one long technological revolution, not a bunch of different ones put together, but this has been continually happening. We, as our quality of life raises and as machines and tools that we create can help us live better lives the bar raises for what we do and, and our human ability and what we spend our time going after goes after more ambitious, more satisfying projects. So there will be an impact on jobs. We try to be very clear about that, and I think it will require partnership between the industry and government, but mostly action by government to figure out how we want to mitigate that. But I’m very optimistic about how great the jobs of the future will be.
Fair statement. There is a bit naive view that we will increase happiness as we work less in some part of the work. But, so far, digital revolution has proven opposite. With Gig economy people work more and more. There is only sharp increase in inequality as capital becomes more relevant than labour.
-
not a creature,
God point on avoiding anthropomorphism.
-
this technology is in its early stages
As with Google and other tech companies, it is likely to remain in permanent 'beta version'.
-
I think that’s a great idea.
To be cynical - it is a 'great idea' because it won't work in practice, but there is pretention that we are doing something.
-
The National Institutes of Standards and technology actually already has an AI accuracy test,
It would be interesting to see how it works in practice. How can you judge accuracy if AI is about probability. It is not about certainty which is the first building block for accuracy.
-
Some of us might characterize it more like a bomb in a China shop, not a bull.
Q: Is AI bull or bomb in a China ship?
-
Ultimately, we may need something like cern Global, international and neutral, but focused on AI safety rather than high energy physics.
He probably thought of analogy with IPCC as supervisory space. But CERN could play role as place for research on AI and processing huge amount of data.
-
But we also need independent scientists, not just so that we scientists can have a voice, but so that we can participate directly in addressing the problems in evaluating solutions.
An important stakeholder.
-
The sums of money at stake are mind boggling. Emissions drift, OpenAI’s original mission statement proclaimed our goal is to advance AI in the way that most is most likely to benefit humanity as a whole, unconstrained by a need to generate financial return. Seven years later, they’re largely beholden to Microsoft, embroiled in part an epic battle of search engines that routinely make things up.
Why we should not trust AI companies?
-
We want, for example, for our systems to be transparent, to protect our privacy, to be free of bias and above all else to be safe. But current systems are not in line with these values. Current systems are not transparent. They do not adequately protect our privacy, and they continue to perpetuate bias, and even their makers don’t entirely understand how they work. Most of all, we cannot remotely guarantee that they’re safe. And hope here is not enough. The big tech company’s preferred plan boils down to trust us. But why should we?
What is the current situation in AI industry?
-
We all more or less agrees on the values we would like for our AI systems to honor.
Are we? Maybe in the USA, but not globally. Consult the work of Moral Machine which shows that different cultural contexts imply whom we would save in trolley experiment: young - elderly, man - women, rich - poor. See more: https://www.moralmachine.net/
-
A law professor, for example, was accused by a chatbot of sexual harassment untrue. And it pointed to a Washington Post article that didn’t even exist. The more that that happens, the more that anybody can deny anything. As one prominent lawyer told me on Friday, defendants are starting to claim that plaintiffs are making up legitimate evidence. These sorts of allegations undermine the abilities of juries to decide what or who to believe and contribute to the undermining of democracy. Poor medical advice could have serious consequences to an open source large language model recently seems to have played a role in a person’s decision to take their own life. The large language model asked the human, if you wanted to die, why didn’t you do it earlier? And then followed up with, were you thinking of me? When you overdosed without ever referring the patient to the human health?
Examples of risk narrative
-
What criminals are gonna do here is to create counterfeit people.
Risks narrative
-
Choices about data sets that AI companies use will have enormous unseen influence. Those who choose the data will make the rules shaping society in subtle but powerful ways.
What about each of us choosing datasets? AI has to be bottom-up.
-
Fundamentally, these new systems are going to be destabilizing. They can and will create persuasive lies at a scale humanity has never seen before. Outsiders will use them to affect our elections, insiders to manipulate our markets and our political systems. Democracy itself is threatened. Chatbots will also clandestinely shape our opinions, potentially exceeding what social media can do.
Risks narrative
-
guardrails
Guardrails are emerging lingo in AI governance.
-
First, different rules for different risks. The strongest regulation should be applied to use cases with the greatest risks to people and society. Second, clearly defining risks. There must be clear guidance on AI uses or categories of AI supported activity that are inherently high risk. This common definition is key to enabling a clear understanding of what regulatory requirements will apply in different use cases and contexts. Third, be transparent. So AI shouldn’t be hidden. Consumers should know when they’re interacting with an AI system and that they have recourse to engage with a real person should they so desire. No person anywhere should be tricked into interacting with an AI system. And finally, showing the impact. For higher risk use cases, companies should be required to conduct impact assessments that show how their systems perform against tests for bias and other ways that they could potentially impact the public. And to attest that they’ve done so by following risk-based use case-specific approach.
Q: What are 4 elements of precision regulation as proposed by IBM?
-
a precision regulation
Language
Precision regulation is another concept to follow.
-
a threshold of capabilities
What is 'a threashold'. As always devil is in detail.
-
We believe that the benefits of the tools we have deployed so far vastly outweigh the risks,
Balancing narrative Opportunities 80 - Risks 20
-
be My Eyes, used our new multimodal technology in GPT-4 to help visually impaired individuals navigate their environment.
Optimistic narrative
-
We think it can be a printing press moment.
Paradigm shift narrative
-
But the basic question we face is whether or not this issue of AI is a quantitative change in technology or a qualitative change.
Critical question. It is quantiative shift which will evolve into qualitative one.
-
We had four bills initially that were considered by this committee and what may be history in the making. We passed all four bills with unanimous roll calls, unanimous roll calls. I can’t remember another time when we’ve done that.
Child safety online is one of the rare issues that unite all political forces worldwide. Will it be extended to AI?
-
will we strike that balance between technological innovation and our ethical and moral responsibility to humanity, to liberty, to the freedom of this country?
Balance narrative Choice narrative
-
I was reminded of the psychologist and writer Carl Jung, who said at the beginning of the last century that our ability for technological innovation, our capacity for technological revolution, had far outpaced our ethical and moral ability to apply and harness the technology we developed.
A good reminder of Jung's work. It is on the line of Frankenstein's warnings of Mary Shelly.
||Jovan||
-
is it gonna be more like the atom bomb, huge technological breakthrough, but the consequences severe, terrible, continue to haunt us to this day
Analogy - atomic bomb
-
Is it gonna be like the printing press that diffused knowledge, power, and learning widely across the landscape that empowered, ordinary, everyday individuals that led to greater flourishing, that led above all two greater liberty?
Analogy with Printing press
-
We should not repeat our past mistakes, for example, Section 230
Acknowledging msitake with Section 230.
-
scorecards and nutrition labels
Can this analogy work?
-
known risks
The real problem is in 'known'. We can deal with known knowns. (un)known unknowns are major problem.
-
transparency
AI principles
-
Now we have the obligation to do it on AI before the threats and the risks become real. Sensible safeguards are not in opposition to innovation.
-
we may need something like CERN, global, international, and neutral, but focused on AI safety, rather than high-energy physics.
CERN for AI
-
Chatbots can clandestinely shape our opinions, in subtle yet potent ways, potentially exceeding what social media can do. Choices about datasets may have enormous, unseen influence.
-
I call datocracy, the opposite of democracy:
-
These guardrails should be matched with meaningful steps by the business community to do their part.
-
an AI Ethics Board
-
a lead AI ethics official
-
on regulatory guardrails
-
A risk based approach ensures that guardrails for AI apply to any application, even as this new, potentially unforeseen developments in the technology occur, and that those responsible for causing harm are held to account.
-
a “precision regulation” approach to artificial intelligence. This means establishing rules to govern the deployment of AI in specific use-cases, not regulating the technology itself.
This is a new concept a 'precision regulation'
-
international cooperation on AI safety, including examining potential intergovernmental oversight mechanisms and standard-setting.
Call for intergovernmental oversight
-
safety standards, evaluation requirements, disclosure practices, and external validation mechanisms for AI systems subject to license or registration.
-
a combination of licensing or registration requirements
-
adhere to an appropriate set of safety requirements,
-
regulation of AI is essential,
-
to minimize any harmful effects for workers and businesses.
How?
-
the Alignment Research Center (ARC)
-
a Cybersecurity Working Group
-
a Cybersecurity Grant Program
-
novel security controls to help protect core model intellectual property.
Security and intellectual property.
-
We will continue to explore partnerships with industry and researchers, as well as with governments, that encompass the full disinformation lifecycle.
-
takes a whole-of-society approach
also widening responsibility to whole society.
-
we recognize that there is more work to do to educate users about the limitations of AI tools, and to reduce the likelihood of inaccuracy.
-
Thorn’s Safer37 service
-
to generate hateful, harassing, violent or adult content, among other categories,
Prohibited content.
-
Any ChatGPT user can opt-out of having their conversations be used to improve our models.33 Users can delete their accounts,34 delete specific conversations from the history sidebar, and disable their chat history at any time.35
Fair points if it is the case in practice.
-
via our API
What about other uses?
-
for the purposes of advertising, promoting our services, or selling data to third parties
What about other purposes
-
Iterative deployment
'Iterative deployment' since to be the keyword. Can 'agile approach' be applied to policy and law. Is it transferable from technology sector?
-
people and our institutions need time to update and adjust to increasingly capable AI, and that everyone who is affected by this technology should have a significant say in how AI develops further.
Is this 'passing responsibility' for products to citizens and government?
-
unsafe content
What is 'unsafe content'?
-
including but not limited to, generation of violent content, malware, fraudulent activity, high-volume political campaigning, and many other unwelcome areas.
This could be part of content which is prohibited.
-
disallowed content
Who decides what is 'disallowed content'? Is there any list of this type of content provided by OpenAI?
-
These opportunities are why former U.S. Treasury Secretary Lawrence Summers has said that AI tools such as ChatGPT might be as impactful as the printing press, electricity, or even the wheel or fire.
It is a good example of using authority argument. This quote about digital as printing press, electricity, wheel or fire is probably the most frequently mentioned in any discussion on impact of digital technology on society.
In this case Lawrence Summers is quoted because of his authority in the US political/academic establishment.
-
Microsoft is an important investor in OpenAI, and we value their unique alignment with our values and long-term vision, including their shared commitment to building AI systems and products that are trustworthy and safe.
Why would Microsoft invest if they do not have any privileged position in OpenAI. For example, do they have privileged access to GPT?
-
-
-
“What I keep emphasizing to people is to just start using this,” says Mollick. As workers get increasing fluent, he adds, they can find themselves ahead of the curve, and at a distinct advantage in the workplace. Workers resistant to AI could be seen as unwilling or incapable of adapting, says Frey. “I think workers that don't work with AI are going to find their skills [become] obsolete quite rapidly. So, therefore, it's imperative to work with AI to stay employed, stay productive and have up to date skills.”
-
increasing the demand for jobs including data analysts and scientists who work with the technology to create best practices in the workplace.
-
might be able to identify a confirmation bias in their work, meaning they look for evidence to support an outcome they already believe exists
-
functions as a sounding board – a tool to bounce ideas off,
-
In his own field of academia, for instance, he’s seen it test for counterarguments to a thesis, and write an abstract for research. “You can ask it to generate a tweet to promote your paper,” he adds. “There are tremendous possibilities.” For knowledge workers, this could mean creating an outline for a blog and a social media post to go with it, distil complex topics for a target audience
AI for knowledge workers
-
-
www.reddit.com www.reddit.com
-
Stability AI announced StableLM - their Language Models.
Open source AI
-
Open Assistant - just wow - is an open source Chat AI.
Open source AI
-
-
lmsys.org lmsys.org
-
How To Evaluate a Chatbot?
It is critical issue - evaluation.
-
To ensure data quality, we convert the HTML back to markdown and filter out some inappropriate or low-quality samples.
||JovanNj||||anjadjATdiplomacy.edu|| Is this a solution for our markup with TExtus?
-
Vicuna is created by fine-tuning a LLaMA base model using approximately 70K user-shared conversations gathered from ShareGPT.com with public APIs.
-
Inspired by the Meta LLaMA and Stanford Alpaca project, we introduce Vicuna-13B, an open-source chatbot backed by an enhanced dataset and an easy-to-use, scalable infrastructure.
-
-
www.reddit.com www.reddit.com
-
Cerebras (not to be confused with our own Cerebra) trains the GPT-3 architecture using the optimal compute schedule implied by Chinchilla, and the optimal scaling implied by μ-parameterization. This outperforms existing GPT-3 clones by a wide margin, and represents the first confirmed use of μ-parameterization “in the wild”. These models are trained from scratch, meaning the community is no longer dependent on LLaMA.
||JovanNj||||anjadjATdiplomacy.edu|| It is a very interesting development. They do not use any more even LLaMA. How is it possible?
-
LoRa and ControlNet (not to mention in- and outpainting), there’s a clear benefit to letting people go wild with your tech.
-
allowing the model to better understand and use context.
-
We explore two other obvious sources of basis dependency in a Transformer: Layer normalization, and finite-precision floating-point calculations. We confidently rule these out as being the source of the observed basis-alignment.
-
Alpaca AI is open source and around the same performance as gpt3. https://github.com/tatsu-lab/stanford_alpaca
||JovanNj||||anjadjATdiplomacy.edu|| Is Alpaca AI useful model?
-
Look into Modular, they have an interesting platform for AI development.
-
GPT 4 cost well over $100 million to train alone, $700k to run per day.
-
-
bair.berkeley.edu bair.berkeley.edu
-
we focus on collecting a small high-quality dataset.
-
built using significant amounts of human annotation.
-
Our results suggest that learning from high-quality datasets can mitigate some of the shortcomings of smaller models, maybe even matching the capabilities of large closed-source models in the future.
Chance for smaller models ||sorina||||VladaR||||JovanNj||||anjadjATdiplomacy.edu||
-
will the future see increasingly more consolidation around a handful of closed-source models, or the growth of open models with smaller architectures that approach the performance of their larger but closed-source cousins?
-
This suggests that in the future, highly capable LLMs will be largely controlled by a small number of organizations, and both users and researchers will pay to interact with these models without direct access to modify and improve them on their own.
-
the community should put more effort into curating high-quality datasets
-
it suggests that models that are small enough to be run locally can capture much of the performance of their larger cousins if trained on carefully sourced data
-
large closed-source models to smaller public models
-
-
www.semianalysis.com www.semianalysis.com
-
Google "We Have No Moat, And Neither Does OpenAI"
||sorina||||VladaR|| Small data models are becoming reality. The key is to have high quality data which we can provide with textus.
-
Open source alternatives can and will eventually eclipse them unless they change their stance. In this respect, at least, we can make the first move.
-
This necessarily means relinquishing some control over our models. But this compromise is inevitable. We cannot hope to both drive innovation and control it.
-
But this control is a fiction. Anyone seeking to use LLMs for unsanctioned purposes can simply take their pick of the freely available models.
-
By owning the platform where innovation happens, Google cements itself as a thought leader and direction-setter, earning the ability to shape the narrative on ideas that are larger than itself.
-
These models are used and created by people who are deeply immersed in their particular subgenre, lending a depth of knowledge and empathy we cannot hope to match.
-
But holding on to a competitive advantage in technology becomes even harder now that cutting edge research in LLMs is affordable.
-
he best are already largely indistinguishable from ChatGPT.
-
This means that as new and better datasets and tasks become available, the model can be cheaply kept up to date, without ever having to pay the cost of a full run.
-
Being able to personalize a language model in a few hours on consumer hardware is a big deal, particularly for aspirations that involve incorporating new and diverse knowledge in near real-time.
-
called low rank adaptation, or LoRA
||JovanNj||||anjadjATdiplomacy.edu|| Is this something we should use?
-
The barrier to entry for training and experimentation has dropped from the total output of a major research organization to one person, an evening, and a beefy laptop.
-
Giant models are slowing us down.
-
Open-source models are faster, more customizable, more private, and pound-for-pound more capable.
-
Scalable Personal AI: You can finetune a personalized AI on your laptop in an evening.
||JovanNj||||anjadjATdiplomacy.edu|| Is it possible to have personalised AI in an evening.
-
-
-
he world and preparing for change rather than trying to roll history back
-
The author, Ganesh Sitaraman, calls for an American “grand strategy of resilience.”(link is external)
-
it will be the American responsibility to listen more and collaborate more seriously to ensure broadly supported recommendations.
-
its resilience and adaptability.
-
If we had looked forward decades ago to understand and accept what was going to be happening today in front of our eyes, we would have had time to better manage the adjustment and implement the remedies.
-
“Tragedy and Mobilization”, the world confronts the cumulative effects of unmanaged climate change.
-
In “Separate Silos” the world fails to manage a co-existence model and the global order devolves into regional power blocs - the U.S, the European Union, Japan, Korea, Australia, Russia, China, India and some rising states - focused on self-sufficiency.
-
The “Competitive Coexistence” scenario is less dangerous primarily because both the U.S. and China make economic growth a priority and to some extent achieve co-dependency on maintaining a stable global order.
-
“A World Adrift” scenario, the international system breaks down as the rules and institutions of today’s structures are little used by the major powers, regional states and non-state actors.
-
The scale of transnational challenges, and the emerging implications of fragmentation, are exceeding the capacity(link is external) of existing systems and structures . . .”
-
Go global.
-
Bring in more participants from outside the government in working group formats - private enterprise, institutions, and other stakeholders to start discussions earlier than they might otherwise and to speed up the policy formation process.
-
Re-establish the diplomacy and science career track at the State Department
-
Tighten the bond between science and diplomacy.
-
the research and policy dimension within the U.S. government five to ten years ahead.
-
Now soft power has a new role to play, not merely as a cultural tool, but as a science and technology avenue of influence.
Science as part of soft power influence.
-
the ability to guide outcomes with culture, the sciences and by the power of our example – has receded.
-
Overemphasis on the bilateral model of American diplomacy does not provide the best process for dealing with modern large scale over-the-horizon issues.
-
new ideas are often confronted by old thinking, passive resistance and a wait-it-out state of mind.
-
has established a policy ideas channel to inspire new views from within and outside the State Department to challenge “groupthink(link is external)
-
“Bringing America’s Multilateral Diplomacy into the 21st Century(link is external)”,
-
The UN is the depository of more than 560 multilateral treaties(link is external)
-
The Law of the Sea Treaty of 1982
-
The Outer Space Treaty of 1967
-
Nuclear and non-nuclear weapons limitation treaties
-
The Antarctic Treaty of 1959
-
None is perfect; all treaties have to keep up with the times in order to survive, and all treaties leave some gaps to be solved later.
-
The event is a call to action on two fronts: ensuring the safe use of near-Earth orbit and dealing with the dangerous escalation of anti-satellite technology(link is external).
-
Seeing America pulling back from the world and divided, China announced its plans to replace the United States as the most powerful nation on earth.
-
resentment of the middle class.
-
Young Americans who came of age in the 2000’s have never known a time of peace and tranquility.
-
five elements are required: (1) involvement of all the essential stakeholders (those that could make or break an agreement), (2) consensus definition of the problem, (3) sufficient common interests to generate a productive dialogue, (4) a shared commitment by the stakeholders to finding a solution, and (5) successful post-agreement implementation that stands the test of time.
-
The lack of a genuine partnership between the worlds of science and diplomacy to integrate multidisciplinary subject-matter
-
Today our diplomats are not trained in the scientific aspects of dealing with issues of global health, climate change, energy renewal, cyber threats, food and water resources, regional or global supply chains and outer space among others.
-
No global issue of significance today or for the foreseeable future will be solely national - allocating more of the diplomatic circle graph to the multilateral slice is both in our interest and more likely than ever to be the methodology of the future.
-
to engage with the larger issues looming just over the horizon.
-
is consumed with managing the moment, the immediate.
-
by engaging with others at the “early stages” of issues to keep us on the front lines of managing global trends.
-
one that recognizes the changing nature of the challenges we will inevitably face in the future – not just the problems we face now.
-
-
www.newyorker.com www.newyorker.com
-
The Luddites were not anti-technology; what they wanted was economic justice.
-
A.I. researchers are increasing the concentration of wealth to such extreme levels that the only way to avoid societal collapse is for the government to step in.
-
the only way to make things better is to make things worse.
-
What Žižek advocated for is an example of an idea in political philosophy known as accelerationism
-
it seems like a way for the people developing A.I. to pass the buck to the government.
-
A.I. assists capital at the expense of labor.
-
by hiring consultants, management can say that they were just following independent, expert advice.
-
we rely on metaphor,
-
-
fortune.com fortune.com
-
it triggers a mental shortcut in the minds of users that we call a “machine heuristic.” This shortcut is the belief that machines are accurate, objective, unbiased, infallible and so on.
An interesting conceput of machine heuristic.
-
Copyleft licensing allows for content to be used, reused or modified easily under the terms of a license – for example, open-source software.
-
But with self-driving cars, the engineers can never be sure how it will perform in novel situations.
-
that people treat computers as social beings when the machines show even the slightest hint of humanness, such as the use of conversational language.
-
Such beliefs build on “automation bias” or the tendency to let your guard down when machines are performing a task
-
-
carnegieendowment.org carnegieendowment.org
-
However, if not coupled with resources and pressure to actually perform this sector-specific adaptation, this approach runs the risk of resulting in no meaningfully binding regulation at all.
-
a unique blend of horizontal and vertical elements
-
sectoral regulators will be outmatched in their efforts to meaningfully constrain businesses applying AI.
-
If agencies do not coordinate to build common regulatory tools, they risk reinventing the wheel each time a new department is tasked with regulating a specific application of AI.
-
if industry-dominated standards bodies set weak standards for compliance, the regulation itself becomes weak.
-
Their legislative bodies will need the ability to amend or add to its main horizontal regulation in order to keep pace with the technology.
-