AGI and ASI might produce an extinction-level event that wipes out humanity. Not good. This is an existential risk of the ...