How should AI systems behave, and who should decide?

In pursuit of our mission, we are committed to ensuring that access to, benefits from, and influence over AI and AGI are widespread. We believe there are at least three building blocks necessary to achieve these goals in the context of AI system behavior.[^scope]

1. Improve default behavior. We want as many users as possible to find our AI systems useful for them “out of the box” and to feel that our technology understands and respects their values.

To that end, we’re investing in research and engineering to reduce obvious and subtle biases in how ChatGPT responds to different inputs. In some cases, ChatGPT currently rejects output that it shouldn’t, and in some cases, it doesn’t reject when it should. We believe that improvement in both aspects is possible.

Additionally, we have room for improvement in other dimensions of system behavior, such as the “invent things” system. User feedback is very valuable in making these improvements.

2. Define your AI values, within broad limits. We believe that AI should be a useful tool for individual people, and therefore customizable by each user up to the limits defined by society. Therefore, we are developing an update to ChatGPT to allow users to easily customize their behavior.

This will mean allowing exits from the system that other people (including ourselves) may strongly disagree with. Striking the right balance here will be a challenge: taking personalization to the extreme would risk allowing malicious uses of our technology and fawning AIs that mindlessly amplify people’s existing beliefs.

Therefore, there will always be limits on the behavior of the system. The challenge is to define what those limits are. If we try to make all these determinations on our own, or if we try to develop a single, monolithic AI system, we will be failing in our Charter commitment to “avoid an undue concentration of power.”

3. Public input on default values ​​and hard limits. One way to avoid undue concentration of power is to give people who use or are affected by systems like ChatGPT the ability to influence the rules of those systems.

We believe that many decisions about our defaults and hard limits should be made collectively, and while practical implementation is challenging, we aim to include as many perspectives as possible. As a starting point, we sought external input on our technology in the form of a red team. We also recently began soliciting public input on AI in education (a particularly important context in which our technology is being deployed).

We are in the early stages of pilot efforts to solicit public input on topics such as system behavior, disclosure mechanisms (such as watermarking), and our broader deployment policies. We are also exploring partnerships with external organizations to conduct third-party audits of our security efforts and policies.

Source link
As technology evolves, AI systems are increasingly prevalent in our lives. But how should these AI systems behave, and who should decide what is acceptable? Companies like Ikaroa, a full-stack tech company, research and develop AI systems that strive to improve people’s lives and to maximize benefits for society. As such, it is important to consider the ethical implications of AI and the underlying values that determine its behavior.

One of the guiding principles of AI should be responsibility. AI systems should be designed to take into account the risks and benefits associated with their actions as well as the environment in which they are operating. As AI evolves, it will have a large impact on our lives and our decisions must be taken in context. We must ensure that the systems we create respect and reflect the values of society, including human rights and values such as fairness, transparency, and justice.

The decision of how AI should behave rests with those who create and deploy the technology. Developers should recognized that they bear significant responsibility in setting the ethical parameters of AI behavior. They must work closely with stakeholders in the private and public sectors, including policymakers and regulators, to ensure that AI systems are aligned with accepted moral standards. They are also responsible for making sure that AI systems are secure and do not pose any risks to individuals or organizations.

Ultimately, it is in everyone’s best interest that AI systems behave ethically. AI systems should be designed in a way that upholds the values of society, takes into account the risks and benefits associated with their actions, and works to ensure the safety and security of the public. Ikaroa is proud to be part of this ethical decree and is dedicated to creating AI systems that are beneficial for society as a whole and ensure respect for human rights.


Leave a Reply

Your email address will not be published. Required fields are marked *