Olle Häggström’s research while affiliated with Chalmers University of Technology and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (97)


Are Large Language Models Intelligent? Are Humans?
  • Conference Paper
  • Full-text available

August 2023

·

30 Reads

·

1 Citation

Olle Häggström
Download

Artificial General Intelligence and the Common Sense Argument

November 2022

·

46 Reads

·

2 Citations

It has sometimes been suggested that lack of common sense in present-day AI systems demonstrates that there is no need to be concerned about future scenarios where a powerful unaligned AI seizes control of the world from us humans. This argument is criticized, and the confusion is shown to stem partly from an overly one-sided focus on artificial general intelligence (AGI), along with the implicit and unwarranted assumption that AGI must be attained before any such drastic scenarios can be realized.


Philosophy and theory of artificial intelligence 2021

November 2022

·

3,364 Reads

·

15 Citations

·

·

·

[...]

·

This book gathers contributions from the fourth edition of the Conference on "Philosophy and Theory of Artificial Intelligence" (PT-AI), held on 27-28th of September 2021 at Chalmers University of Technology, in Gothenburg, Sweden. It covers topics at the interface between philosophy, cognitive science, ethics and computing. It discusses advanced theories fostering the understanding of human cognition, human autonomy, dignity and morality, and the development of corresponding artificial cognitive structures, analyzing important aspects of the relationship between humans and AI systems, including the ethics of AI. This book offers a thought-provoking snapshot of what is currently going on, and what are the main challenges, in the multidisciplinary field of the philosophy of artificial intelligence.



AI, orthogonality and the M\"uller-Cannon instrumental vs general intelligence distinction

September 2021

·

34 Reads

·

2 Citations

The by now standard argument put forth by Yudkowsky, Bostrom and others for why the possibility of a carelessly handled AI breakthrough poses an existential threat to humanity is shown through careful conceptual analysis to be very much alive and kicking, despite the suggestion in a recent paper by M\"uller and Cannon that the argument contains a flaw.


Aspects of Mind Uploading

January 2021

·

134 Reads

·

2 Citations

Cognitive Technologies

Mind uploading is the hypothetical future technology of transferring human minds to computer hardware using whole-brain emulation. After a brief review of the technological prospects for mind uploading, a range of philosophical and ethical aspects of the technology are reviewed. These include questions about whether uploads will have consciousness and whether uploading will preserve personal identity, as well as what impact on society a working uploading technology is likely to have and whether these impacts are desirable. The issue of whether we ought to move forward towards uploading technology remains as unclear as ever.


An AGI Modifying Its Utility Function in Violation of the Strong Orthogonality Thesis

December 2020

·

104 Reads

·

12 Citations

Philosophies

An artificial general intelligence (AGI) might have an instrumental drive to modify its utility function to improve its ability to cooperate, bargain, promise, threaten, and resist and engage in blackmail. Such an AGI would necessarily have a utility function that was at least partially observable and that was influenced by how other agents chose to interact with it. This instrumental drive would conflict with the strong orthogonality thesis since the modifications would be influenced by the AGI’s intelligence. AGIs in highly competitive environments might converge to having nearly the same utility function, one optimized to favorably influencing other agents through game theory. Nothing in our analysis weakens arguments concerning the risks of AGI.


An AGI Modifying Its Utility Function in Violation of the Orthogonality Thesis

January 2020

·

196 Reads

An artificial general intelligence (AGI) might have an instrumental drive to modify its utility function to improve its ability to cooperate, bargain, promise, threaten, and resist and engage in blackmail. Such an AGI would necessarily have a utility function that was at least partially observable and that was influenced by how other agents chose to interact with it. This instrumental drive would conflict with the orthogonality thesis since the modifications would be influenced by the AGI's intelligence. AGIs in highly competitive environments might converge to having nearly the same utility function, one optimized to favorably influencing other agents through game theory.


Figure 2 Historical world population estimates for 1950-2015, and future population projections out to 2100 at three prediction intervals, based on data from United Nations (2017b, 2017c)
Figure 5 Tentative probability distribution of post-catastrophe agriculture
Figure 7 Illustrative technological transformation and astronomical trajectories, including baseline and delayed (by n units of time) exponential growth trajectories for four trajectory types
Long-term trajectories of human civilization

February 2019

·

1,903 Reads

·

85 Citations

foresight

Purpose This paper aims to formalize long-term trajectories of human civilization as a scientific and ethical field of study. The long-term trajectory of human civilization can be defined as the path that human civilization takes during the entire future time period in which human civilization could continue to exist. Design/methodology/approach This paper focuses on four types of trajectories: status quo trajectories, in which human civilization persists in a state broadly similar to its current state into the distant future; catastrophe trajectories, in which one or more events cause significant harm to human civilization; technological transformation trajectories, in which radical technological breakthroughs put human civilization on a fundamentally different course; and astronomical trajectories, in which human civilization expands beyond its home planet and into the accessible portions of the cosmos. Findings Status quo trajectories appear unlikely to persist into the distant future, especially in light of long-term astronomical processes. Several catastrophe, technological transformation and astronomical trajectories appear possible. Originality/value Some current actions may be able to affect the long-term trajectory. Whether these actions should be pursued depends on a mix of empirical and ethical factors. For some ethical frameworks, these actions may be especially important to pursue.


Challenges to the Omohundro–Bostrom framework for AI motivations

October 2018

·

136 Reads

·

16 Citations

foresight

Purpose This paper aims to contribute to the futurology of a possible artificial intelligence (AI) breakthrough, by reexamining the Omohundro–Bostrom theory for instrumental vs final AI goals. Does that theory, along with its predictions for what a superintelligent AI would be motivated to do, hold water? Design/methodology/approach The standard tools of systematic reasoning and analytic philosophy are used to probe possible weaknesses of Omohundro–Bostrom theory from four different directions: self-referential contradictions, Tegmark’s physics challenge, moral realism and the messy case of human motivations. Findings The two cornerstones of Omohundro–Bostrom theory – the orthogonality thesis and the instrumental convergence thesis – are both open to various criticisms that question their validity and scope. These criticisms are however far from conclusive: while they do suggest that a reasonable amount of caution and epistemic humility is attached to predictions derived from the theory, further work will be needed to clarify its scope and to put it on more rigorous foundations. Originality/value The practical value of being able to predict AI goals and motivations under various circumstances cannot be overstated: the future of humanity may depend on it. Currently, the only framework available for making such predictions is Omohundro–Bostrom theory, and the value of the present paper is to demonstrate its tentative nature and the need for further scrutiny.


Citations (86)


... As such, this knowledge is distilled into the models, granting them an extremely broad and often quite deep 'understanding' [98]. While the true nature of their ability to reason is debated [27,46,91], they seem capable of combining knowledge and ideas together to produce new outputs [205]. This makes them well suited to interpreting the true meaning behind the tasks they are set, as well as being able to expand on observations and details by combining learned knowledge in order to achieve their goals. ...

Reference:

A Survey of Language-Based Communication in Robotics
Are Large Language Models Intelligent? Are Humans?

... Finally, and as I have argued at greater length elsewhere [4], the notion of "common sense" is perhaps better avoided altogether, as it is likely to confuse more than it enlightens. It tends to serve as a catch-all term for everything that humans still do better than AI, meaning that the phrase "AIs lack common sense" will almost tautologically continue to apply right until the moment when AI outperforms humanity at everything. ...

Artificial General Intelligence and the Common Sense Argument
  • Citing Chapter
  • November 2022

... It is not only that a copy of my mind would not be myself, but it is also that, very possibly, my own mind in another body would not be myself. This is at least what one would have to think about if we consider personal identity as something more than the possession of a mind and a mind as something more than a set of information or as a computer program (Häggström 2021;Hauskeller 2012;Schneider 2019). ...

Aspects of Mind Uploading
  • Citing Chapter
  • January 2021

Cognitive Technologies

... However, typical human devised security should be relatively trivial for an advanced AI to overcome-although there might be exceptions (Tegmark and Omohundro 2023), which might force the AI to resort to more extreme measures. As for differing purposes or utility functions as a hurdle to merger, Totschnig, and Miller and colleagues have independently published compelling arguments for why a self-determining AI will strategically modify its utility function (or purpose) (Miller et al. 2020;Totschnig 2019Totschnig , 2020. In the following subsections, I expand upon their arguments. ...

An AGI Modifying Its Utility Function in Violation of the Strong Orthogonality Thesis

Philosophies

... While many recent start-ups have been able to raise enormous amounts of capital buoyed by the hype surrounding the capabilities of the latest models (e.g., Stability AI, Cohere), it is unclear whether they will be able to continue to compete without partnering with larger firms (as OpenAI and Anthropic have, e.g., with Microsoft, and Alphabet and Amazon, respectively) given the race between wellcapitalised firms to acquire as many GPUs as possible. 12 It is unclear whether Nvidia will retain its dominance in the training of LLMs indefinitely, 13 but AMD appears to only be able to offer products that can efficiently be used to perform inference at this time (ORNL 2024; Marty et al. 2024). There is talk about trillion-dollar clusters (Hagey and Fitch 2024; Aschenbrenner 2024), but it is unclear whether -if 11 This is in essence why Intelligence Rising can generate relatively plausible narratives for a very complex global problem despite simulating the actions of very few stakeholders. ...

Long-term trajectories of human civilization

foresight

... One more principle from the AI control literature finds support in the historical record: the idea that intelligence is correlated with the development of a moral code. Chalmers (2016) calls this a 'Kantian' perspective; Häggström (2018) suggests that if moral realism and internalism are true, superintelligences will inevitably discover moral truths and find them compelling. The question of whether AI is subject to this kind of internally-generated moral pressure is hotly debated (Corabi 2017). ...

Challenges to the Omohundro–Bostrom framework for AI motivations
  • Citing Article
  • October 2018

foresight

... One can interpret a partial sum on a path as an height, which is updated by adding the slope of each visited edge, that has to be reversed when the edge is crossed in opposition with the standard orientation of the two axes. With this interpretation one can see some similarities with the model introduced in [12]. Loosely speaking, in this paper an environment of i.i.d. ...

Water transport on infinite graphs
  • Citing Article
  • March 2018