August 2023
·
30 Reads
·
1 Citation
This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.
August 2023
·
30 Reads
·
1 Citation
November 2022
·
46 Reads
·
2 Citations
It has sometimes been suggested that lack of common sense in present-day AI systems demonstrates that there is no need to be concerned about future scenarios where a powerful unaligned AI seizes control of the world from us humans. This argument is criticized, and the confusion is shown to stem partly from an overly one-sided focus on artificial general intelligence (AGI), along with the implicit and unwarranted assumption that AGI must be attained before any such drastic scenarios can be realized.
November 2022
·
3,364 Reads
·
15 Citations
This book gathers contributions from the fourth edition of the Conference on "Philosophy and Theory of Artificial Intelligence" (PT-AI), held on 27-28th of September 2021 at Chalmers University of Technology, in Gothenburg, Sweden. It covers topics at the interface between philosophy, cognitive science, ethics and computing. It discusses advanced theories fostering the understanding of human cognition, human autonomy, dignity and morality, and the development of corresponding artificial cognitive structures, analyzing important aspects of the relationship between humans and AI systems, including the ethics of AI. This book offers a thought-provoking snapshot of what is currently going on, and what are the main challenges, in the multidisciplinary field of the philosophy of artificial intelligence.
July 2022
·
3 Reads
The Mathematical Intelligencer
September 2021
·
34 Reads
·
2 Citations
The by now standard argument put forth by Yudkowsky, Bostrom and others for why the possibility of a carelessly handled AI breakthrough poses an existential threat to humanity is shown through careful conceptual analysis to be very much alive and kicking, despite the suggestion in a recent paper by M\"uller and Cannon that the argument contains a flaw.
January 2021
·
134 Reads
·
2 Citations
Cognitive Technologies
Mind uploading is the hypothetical future technology of transferring human minds to computer hardware using whole-brain emulation. After a brief review of the technological prospects for mind uploading, a range of philosophical and ethical aspects of the technology are reviewed. These include questions about whether uploads will have consciousness and whether uploading will preserve personal identity, as well as what impact on society a working uploading technology is likely to have and whether these impacts are desirable. The issue of whether we ought to move forward towards uploading technology remains as unclear as ever.
December 2020
·
104 Reads
·
12 Citations
Philosophies
An artificial general intelligence (AGI) might have an instrumental drive to modify its utility function to improve its ability to cooperate, bargain, promise, threaten, and resist and engage in blackmail. Such an AGI would necessarily have a utility function that was at least partially observable and that was influenced by how other agents chose to interact with it. This instrumental drive would conflict with the strong orthogonality thesis since the modifications would be influenced by the AGI’s intelligence. AGIs in highly competitive environments might converge to having nearly the same utility function, one optimized to favorably influencing other agents through game theory. Nothing in our analysis weakens arguments concerning the risks of AGI.
January 2020
·
196 Reads
An artificial general intelligence (AGI) might have an instrumental drive to modify its utility function to improve its ability to cooperate, bargain, promise, threaten, and resist and engage in blackmail. Such an AGI would necessarily have a utility function that was at least partially observable and that was influenced by how other agents chose to interact with it. This instrumental drive would conflict with the orthogonality thesis since the modifications would be influenced by the AGI's intelligence. AGIs in highly competitive environments might converge to having nearly the same utility function, one optimized to favorably influencing other agents through game theory.
February 2019
·
1,903 Reads
·
85 Citations
foresight
Purpose This paper aims to formalize long-term trajectories of human civilization as a scientific and ethical field of study. The long-term trajectory of human civilization can be defined as the path that human civilization takes during the entire future time period in which human civilization could continue to exist. Design/methodology/approach This paper focuses on four types of trajectories: status quo trajectories, in which human civilization persists in a state broadly similar to its current state into the distant future; catastrophe trajectories, in which one or more events cause significant harm to human civilization; technological transformation trajectories, in which radical technological breakthroughs put human civilization on a fundamentally different course; and astronomical trajectories, in which human civilization expands beyond its home planet and into the accessible portions of the cosmos. Findings Status quo trajectories appear unlikely to persist into the distant future, especially in light of long-term astronomical processes. Several catastrophe, technological transformation and astronomical trajectories appear possible. Originality/value Some current actions may be able to affect the long-term trajectory. Whether these actions should be pursued depends on a mix of empirical and ethical factors. For some ethical frameworks, these actions may be especially important to pursue.
October 2018
·
136 Reads
·
16 Citations
foresight
Purpose This paper aims to contribute to the futurology of a possible artificial intelligence (AI) breakthrough, by reexamining the Omohundro–Bostrom theory for instrumental vs final AI goals. Does that theory, along with its predictions for what a superintelligent AI would be motivated to do, hold water? Design/methodology/approach The standard tools of systematic reasoning and analytic philosophy are used to probe possible weaknesses of Omohundro–Bostrom theory from four different directions: self-referential contradictions, Tegmark’s physics challenge, moral realism and the messy case of human motivations. Findings The two cornerstones of Omohundro–Bostrom theory – the orthogonality thesis and the instrumental convergence thesis – are both open to various criticisms that question their validity and scope. These criticisms are however far from conclusive: while they do suggest that a reasonable amount of caution and epistemic humility is attached to predictions derived from the theory, further work will be needed to clarify its scope and to put it on more rigorous foundations. Originality/value The practical value of being able to predict AI goals and motivations under various circumstances cannot be overstated: the future of humanity may depend on it. Currently, the only framework available for making such predictions is Omohundro–Bostrom theory, and the value of the present paper is to demonstrate its tentative nature and the need for further scrutiny.
... As such, this knowledge is distilled into the models, granting them an extremely broad and often quite deep 'understanding' [98]. While the true nature of their ability to reason is debated [27,46,91], they seem capable of combining knowledge and ideas together to produce new outputs [205]. This makes them well suited to interpreting the true meaning behind the tasks they are set, as well as being able to expand on observations and details by combining learned knowledge in order to achieve their goals. ...
August 2023
... This reflection is often paired with considerations on the singularity hypothesis (cf. Dehaene, Lau, and Kouider 2021;Torrance 2011;Wolf 2021;Yampolskiy 2013). ...
November 2022
... Finally, and as I have argued at greater length elsewhere [4], the notion of "common sense" is perhaps better avoided altogether, as it is likely to confuse more than it enlightens. It tends to serve as a catch-all term for everything that humans still do better than AI, meaning that the phrase "AIs lack common sense" will almost tautologically continue to apply right until the moment when AI outperforms humanity at everything. ...
November 2022
... This point is echoed by Müller and Cannon (2022). While I cannot settle this issue here, I broadly accept Häggström's (2021) reply. 19 For a discussion of some possible countermeasures and their limitations, see Taylor et al. (2020). ...
September 2021
... It is not only that a copy of my mind would not be myself, but it is also that, very possibly, my own mind in another body would not be myself. This is at least what one would have to think about if we consider personal identity as something more than the possession of a mind and a mind as something more than a set of information or as a computer program (Häggström 2021;Hauskeller 2012;Schneider 2019). ...
January 2021
Cognitive Technologies
... However, typical human devised security should be relatively trivial for an advanced AI to overcome-although there might be exceptions (Tegmark and Omohundro 2023), which might force the AI to resort to more extreme measures. As for differing purposes or utility functions as a hurdle to merger, Totschnig, and Miller and colleagues have independently published compelling arguments for why a self-determining AI will strategically modify its utility function (or purpose) (Miller et al. 2020;Totschnig 2019Totschnig , 2020. In the following subsections, I expand upon their arguments. ...
December 2020
Philosophies
... While many recent start-ups have been able to raise enormous amounts of capital buoyed by the hype surrounding the capabilities of the latest models (e.g., Stability AI, Cohere), it is unclear whether they will be able to continue to compete without partnering with larger firms (as OpenAI and Anthropic have, e.g., with Microsoft, and Alphabet and Amazon, respectively) given the race between wellcapitalised firms to acquire as many GPUs as possible. 12 It is unclear whether Nvidia will retain its dominance in the training of LLMs indefinitely, 13 but AMD appears to only be able to offer products that can efficiently be used to perform inference at this time (ORNL 2024; Marty et al. 2024). There is talk about trillion-dollar clusters (Hagey and Fitch 2024; Aschenbrenner 2024), but it is unclear whether -if 11 This is in essence why Intelligence Rising can generate relatively plausible narratives for a very complex global problem despite simulating the actions of very few stakeholders. ...
February 2019
foresight
... One more principle from the AI control literature finds support in the historical record: the idea that intelligence is correlated with the development of a moral code. Chalmers (2016) calls this a 'Kantian' perspective; Häggström (2018) suggests that if moral realism and internalism are true, superintelligences will inevitably discover moral truths and find them compelling. The question of whether AI is subject to this kind of internally-generated moral pressure is hotly debated (Corabi 2017). ...
October 2018
foresight
... In this model, opinions are represented by real numbers, and when two agents interact, they move their opinions closer to each other, provided that their opinions are not too far apart. The model of edge-sharing also has applications in water resource management [4], [5]. ...
Reference:
Sharing tea on a graph
September 2018
Random Structures and Algorithms
... One can interpret a partial sum on a path as an height, which is updated by adding the slope of each visited edge, that has to be reversed when the edge is crossed in opposition with the standard orientation of the two axes. With this interpretation one can see some similarities with the model introduced in [12]. Loosely speaking, in this paper an environment of i.i.d. ...
March 2018