My attempt at clarifying a confusing topic
It appears to me that your definition of alignment is slightly askew. In all your examples the real issue is who the system is aligned TO.
It appears to me the systems in question are performing in perfect alignment to a system OTHER than your observer, implying alignment, in the sense you seem to be implying, is relative.
The first question to ask then, IMO, is whose desires and intentions the system was really designed to serve.
great points! we humans don't always align on many soft issues and today's guidelines may change with emerging evidence; how can we expect better from AI performance? can they successfully and unbiasedly navigate mess of human cognitive output that litter AI model's data feed? how do we lay the theory and causal inference as foundational principles of generative AI models? so many challenges to tackle, but exciting times for those who enjoy playing with hard questions.