Just another WordPress site

“Reframing Superintelligence” + LLMs + 4 years — AI Alignment Forum

Eric Drexler revisits his 2019 report, «Reframing Superintelligence,» in light of recent advances in Large Language Models (LLMs). He argues that his original framework, which emphasizes AI services over singular superintelligent agents, remains relevant. The AI services model broadens the ontology to include more manageable and transparent AI systems. Drexler suggests focusing on AI development through task-focused AI systems, rather than self-modifying agents, to improve safety. He concludes that risks from misaligned humans are a more immediate threat than those from AI itself.

What is the main idea behind «Reframing Superintelligence» and the Comprehensive AI Services (CAIS) model?
«Reframing Superintelligence» challenges the idea that advanced AI should primarily be viewed as utility-driven agents. Instead, it proposes a broader perspective that includes compositions of AI systems understood through their structures, relationships, development processes, and the services they provide. This leads to the Comprehensive AI Services (CAIS) model, which envisions general intelligence as a property of flexible systems of services where task-focused agents are components. The CAIS model suggests that AI services can expand to achieve superintelligent-level performance, including the ability to create new services aligned with human objectives and informed by models of human approval. This reframing aims to broaden the understanding of AI risks and opportunities beyond agent-centric views, suggesting that superintelligent capabilities can exist in more accessible, transparent, and manageable forms than unitary agents. The goal is to add to, not subtract from, existing agent-focused concerns, by including systems in which superintelligent-level capabilities can take a more accessible, transparent, and manageable form, open agencies rather than unitary agents.

How does the concept of R&D automation relate to the development of superintelligence, according to the article?
The article suggests that R&D automation, particularly AI-enabled AI development, offers a more direct path to an intelligence explosion than self-transforming AI agents. In this model, new AI capabilities are implemented without any single system being self-modifying. Current applications of AI to AI development include using Large Language Models (LLMs) for tasks like filtering and upgrading internet datasets, serving as reward models for Reinforcement Learning from Human Feedback (RLHF), generating dialog content, synthesizing examples of instruction-following, and creating synthetic data for training. This approach decouples AI-enabled AI improvement from AI agency by employing task-focused AI systems to incrementally automate AI development, refocusing AI safety concerns on expanding safe AI functionality and investigating safety-relevant affordances, including predictive models of human approval.

What are some of the risks associated with the AI-services model, and what measures can be taken to mitigate them?
The AI-services model presents several risks, including the potential for enabling dangerous agents, empowering bad actors, accelerating harmful applications, and general disruption in economic, political, and military spheres. To mitigate these risks, it’s crucial to study methods for directing and constraining AI services and preventing emergent agent-like behaviors. The development of LLMs has made these risks more concrete and urgent. Potential mitigation strategies include directing and constraining AI services, avoiding emergent agent-like behaviors, and focusing on AI safety studies connected to current R&D practices, such as AI R&D automation, structured system development, and safety guidelines for structured AI systems.


Artículo Original: https://www.alignmentforum.org/posts/LxNwBNxXktvzAko65/reframing-superintelligence-llms-4-years


Advices:

  • Focus on developing AI services rather than unitary AI agents to mitigate risks and improve manageability.
  • Prioritize AI safety research that explores structured AI systems, R&D automation, and safety guidelines for these systems.
  • Leverage machine learning to build models of human approval to guide and constrain AI behavior, ensuring alignment with human values and laws.
Categories: , , , , , ,

Deja una respuesta

Tu dirección de correo electrónico no será publicada. Los campos obligatorios están marcados con *