Changes

Jump to: navigation, search

FIVR

4 bytes added, 23:16, 28 September 2022
m
Text replacement - "artificial intelligence" to "artificial intelligence"
There have been progress in several areas that will converge to promote the development of a complete FIVR. Brain scan techniques will allow for the detection and analysis of thought patterns; full-body VR is being tried with the Virtuix Omni, a “virtual reality rig, which features 40 capacitive sensors in its base to track your every step and move your character inside a game.” This is combined with a headset for a richer experience in the virtual world <ref name=”9”></ref> <ref name=”10”> Flitman, S. S. (2000). Survey of brain imaging techniques with implications for nanomedicine. Eighth Foresight Conference on Molecular Nanotechnology, Bethesda, MD</ref> <ref name=”11”> Hamburguer, E. (2014). Full-body virtual reality is here, but try not to puke. Retrieved from http://www.theverge.com/2014/1/8/5289186/the-virtuix-omni-is-full-body-virtual-reality-but-try-not-to-puke</ref>. Another thing that needs improvement is computer processing speed, in order for a truly full immersive virtual experience to be generated <ref name=”8”></ref>.
While brain mapping studies allow for a greater understanding of the human nervous system, the development of alternative designs for computer chips - that are inspired by biological brains – will enhance [[artificial intelligence]], blurring the boundary between silicon and biological systems <ref name=”12”> Humphries, C. (2014). Brain mapping. Retrieved from https://www.technologyreview.com/s/526501/brain-mapping/</ref> <ref name=”13”> Hof, R. D. (2014). Neuromorphic chips: Microprocessors configured more like brains than traditional chips could soon make computers far more astute about what’s going on around them. Retrieved from https://www.technologyreview.com/s/526506/neuromorphic-chips/</ref>. Brain-computer interfaces keep evolving, and it was possible, in a 2013 study, for humans to control other animals with thoughts alone <ref name=”7”></ref> <ref name=”14”> Anthony, S. (2013). Harvard creates brain-to-brain interface, allows humans to control other animals with thoughts alone. Retrieved from http://www.extremetech.com/extreme/162678-harvard-creates-brain-to-brain-interface-allows-humans-to-control-other-animals-with-thoughts-alone</ref>. This demonstrated that human thought can be correctly interpreted by a computer and, in this case, used to control a rat’s brain. Also, the experiment was non-invasive for both the human and the rat involved in the study, using instead focused ultrasound to transmit the control signals <ref name=”7”></ref>.
In 2015, another experiment used an EEG device (electroencephalogram) and advanced software to detect human thought, making it possible – by placing electrodes on the head and legs of a paraplegic man – for the subject to walk for the first time in years. The signals from the patient’s brain were detected, interpreted, and sent to his legs, bypassing the damaged spinal cord. The U.S. Defense Advanced Research Project Agency (DARPA) is investing in studies to develop “a high-resolution, wide-bandwidth intracranial electrode array for recording and stimulating brain activity.” This would be a minimally invasive device that can be compared to a brain modem, and a possible step to achieve a full dive experience <ref name=”7”></ref>.

Navigation menu