POST-1 Artificial intelligence is to be considered a major facet in the frontier of technological innovation. Also and more commonly refereed as AI, artificial intelligence as defined by John McCarthy, a pioneer in the field is “the science and engineering of creating intelligent machines and software systems”. Although AI is not constrained within a biological framework, its intended purpose revolves around developing technological systems that can understand and emulate “human intelligence.” Even in these early stages of its development, AI has many cross-disciplinary applications some of which include, computer science, psychology, philosophy, neuroscience, and engineering to name a few. In efforts to further the progression of this field of study, a number of key questions must be sufficiently answered. For instance, what is intelligence, and is it able to be adequately transferred into inanimate objects for purposes of achieving certain intended objectives? Therein lies the need for philosophical analysis of what the possibilities are regarding AI, along with what foreseeable implications of these possibilities may result. Let us take a look at what some of these intended objectives are of AI development and what they require in the way of intelligence faculties. Intelligent behavior includes the autonomous exhibition of control, planning, scheduling, speech, handwriting, and facial recognition. All of these aforementioned displays of behavior are considered to be what we call subjective in application as well as interpretation. Subjectivity within this purview, is the nature of “cognitive” conceptualization as it is known in the “mind” as distinct from a thing in itself. Of course the concepts of cognition and the mind are unique aspects of the biological human reality, thereby not entirely characteristic of what occurs within mechanical processing. These concepts however, will have to do for now in efforts to effectively explain their intended relatedness to humanlike activity. The question becomes, to what extent can we legitimately automate the epistemology of subjective human consciousness in order to effectively produce truly intelligently processing machines? Essentially what is to be said from this is that in order to in affect create intellectual consciousness there must exist the human-like duality between the material systems (brain), in tandem with the mental process which exists within the intangible construct of the mind. Conscious acts are simply postulated to be interpretable somehow analogously to physical acts of measurement. Such accounts may provide fascinating scientific theory concerning the abstract relatedness to physical or material phenomenon. However, unless such detailed work leads beyond vague metaphors and analogies, they cannot yet represent scientific progress in the way of fabricating like-phenomenon within manufactured systems. In my discussions I will explore a number of various theories and concerns regarding the relative possibilities inherent in such scientific inquiry. There also lies a considerable degree of stock in the areas of quantum analysis in association with the possible distinctions between the physical constructs of material systems such as a brain and the causality of a mental state representing consciousness. Given that, this series of reports will cover a number of these positions pertaining to this inquiry; beginning with the initial approaches to measure intelligence with the Turing tests and like-applications to the more progressive quantum analysis of possibilities.