WO2015047466A2 - Bi-phasic applications of real & imaginary separation, and reintegration in the time domain - Google Patents
Bi-phasic applications of real & imaginary separation, and reintegration in the time domain Download PDFInfo
- Publication number
- WO2015047466A2 WO2015047466A2 PCT/US2014/041126 US2014041126W WO2015047466A2 WO 2015047466 A2 WO2015047466 A2 WO 2015047466A2 US 2014041126 W US2014041126 W US 2014041126W WO 2015047466 A2 WO2015047466 A2 WO 2015047466A2
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- transform
- components
- user
- waveform
- real
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T11/00—2D [Two Dimensional] image generation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R5/00—Stereophonic arrangements
- H04R5/04—Circuit arrangements, e.g. for selective connection of amplifier inputs/outputs to loudspeakers, for loudspeaker detection, or for adaptation of settings to personal preferences or hearing impairments
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2420/00—Details of connection covered by H04R, not provided for in its groups
- H04R2420/01—Input selection or mixing for amplifiers or loudspeakers
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S1/00—Two-channel systems
- H04S1/002—Non-adaptive circuits, e.g. manually adjustable or static, for enhancing the sound image or the spatial distribution
Definitions
- Harmonic analysis involves the representation of functions or signals as superpositions of basic waves. Harmonic analysis has found application in a great many fields, including signal processing, quantum mechanics, and neuroscience.
- transforms and transformation are used herein to mean decomposition of a signal into a multiple components in the same or a different domain.
- a Hilbert transform converts a function in one domain into a function in the same domain.
- DTFT Discrete-time Fourier transform
- Transforms can be applied to time domains, spatial frequencies, and indeed to nearly any function domain.
- Various transforms are currently used for compression, filtering, frequency balancing, encryption, and for other purposes.
- transforms are usually mathematically based, transforms can also implemented in electronics, as for example, using a parallel pair of serially cascaded biquad filters.
- the inventive subject matter herein provides apparatus, systems and methods in which a signal is decomposed into different components using a transform, with the components then being separately presented to a person in a manner that produces a different cognitive experience than would have resulted from either (a) presentation of the original signal, or (b) presentation of a fully synthesized (inverse transformed) signal.
- Some embodiments of the inventive subject matter deal with audio signals.
- the real and imaginary components of a decomposed monaural audio signal can be presented separately to a listener through headphones, so that the sound of one component is rendered in one ear, and the sound of the other component is rendered in the other ear.
- the listener's brain will try to integrate the sounds in a manner that has been found to produce a much richer experience than if the listener had heard (a) only one of the components, or (b) a signal resulting from a synthesis (reverse transform) of the real and imaginary components.
- an electronic processor can generate a corresponding imaginary component from the real component, and then the two components can be rendered separately to the listener's left and right ears.
- real world testing demonstrates noticeable improvement in the cognitive experience of the listener.
- processors servers, services, interfaces, portals, platforms, or other systems formed from computing devices. It should be appreciated that the use of such terms is deemed to represent one or more computing devices having at least one processor configured to execute software instructions stored on a computer readable tangible, non-transitory medium.
- a server can include one or more computers operating as a web server, database server, or other type of computer server in a manner to fulfill described roles, responsibilities, or functions.
- aspects of the inventive subject matter can also be applied to video files or still images.
- separate viewing of the separate components can be accomplished with individual displays such as those on GoogleTM GlassTM, or by employing glasses or contacts that filter the superimposed components differently for the left and right eyes.
- components derived from transformation of the waves can be presented to the subject as separate sounds or images, and the subject producing the waves can learn to manipulate them using the component renderings as feedback.
- a subject can manipulate his/her thoughts to effectively collapse or otherwise modulate an auditory or visual representation of a wave function corresponding to biometric signal.
- the biometric signal comprises brain waves, this can be useful as part of an analog neurocomputer.
- Other biometrics beside brainwaves can also be used to provide feedback, however, including for example represent breathing or heart rates.
- components derived from transformation can be rendered to a user other than through the ears or eyes.
- the different components could be rendered to a user tactually, through skin vibrations applied to different sides of the body, or even different areas of the same side of the body, respectively.
- the different components could be rendered to a user though electrical or thermal stimulation to the skin or deeper tissues.
- the stimulation could apply the different components to different sides or areas of the body, respectively
- the different components could be rendered to the viewer in rapid sequence, e.g., a repeating sequence of the real component for l/50 th of a second, followed by imaginary component for 1/50 th of a second.
- the listener could control the time periods separately, along with duration of any resting periods. Such "flickering" could apply to auditory, visual, tactual, or any other renderings.
- the different components could be mixed in various percentages, preferably in near real time by the user, through operation of software.
- the sound rendered to the left ear could be 100% real, 90%> real and 10% imaginary, 80% real and 20% imaginary, etc., all the way to 100% imaginary.
- ranges herein is merely intended to serve as a shorthand method of referring individually to each separate value falling within the range.
- each individual intervening value is incorporated into the specification as if it were individually recited herein.
- numeric values set forth herein should be construed in light of the number of reported significant digits, and by applying ordinary rounding techniques.
- the term “near real time” as applied to a transform process means that rendering of at least one of the components produced by the process will tend to lag no more than five seconds from input of the signal being processed, more preferably no more than two seconds from input of the signal being processed, and most preferably no more than one second from input of the signal being processed.
- the term “near real time” means that rendering of at least one of the components lags no more than five seconds behind an operator-triggered manipulation, more preferably no more than two seconds behind an operator-triggered manipulation, and most preferably no more than one second behind an operator-triggered manipulation.
- the "different" cognitive experiences referred to herein are not necessarily desirable.
- the signal being processed is a breath, or background noise, for example, a listener might find the experience of renderings according to the inventive concepts herein to be better, worse, or substantially the same as the original sounds.
- a listener might find the experience of renderings according to the inventive concepts herein to be better, worse, or substantially the same as the original sounds.
- a video or other image rendering it may well be that separately presenting the different components could be very confusing or otherwise troubling to a viewer.
- Fig. 1 is a diagram of an embodiment in which harmonic analysis is used to process a monaural audio signal in a manner than can mimic a stereo listening experience.
- FIG. 2 is a diagram of an embodiment in which two-channel audio signals are used to improve upon a traditional stereo listening experience.
- FIG. 3 is a diagram of an embodiment in which an input source is used to feed both temporal and spectral reinforcement transforms.
- Fig. 4 is a collection of screen prints of various portacles.
- Fig.5A is a single torus created according to teachings herein.
- Fig.5B is a double torus created according to teachings herein.
- Fig.5C is a double torus torsion created according to teachings herein.
- Figs. 6A - 6D are graphic representations of toroidal reformations created according to teachings herein.
- Fig. 7 is a two-dimensional representation of a twelve-dimensional toroidal space based upon a person ' s breath.
- Figs. 8a, 8b, and 8c are real and imaginary components of biometric waves mapped to x and y axes of a matrix, respectively
- Fig. 9 is a collection of images, showing source, complex, real and imaginary components.
- Fig. 10 is a screen shot of the source image of Fig. 9, enhanced by a visual representation of a mental state.
- Fig. 11 is a screen print of an interface 100 of an instance of The Portacle ' Detailed Description
- inventive subject matter provides many example embodiments of the inventive subject matter. Although each embodiment represents a single combination of inventive elements, the inventive subject matter is considered to include all possible combinations of the disclosed elements. Thus if one embodiment comprises elements A, B, and C, and a second embodiment comprises elements B and D, then the inventive subject matter is also considered to include other remaining combinations of A, B, C, or D, even if not explicitly disclosed.
- Coupled to is intended to include both direct coupling (in which two elements that are coupled to each other contact each other) and indirect coupling (in which at least one additional element is located between the two elements). Therefore, the terms “coupled to” and “coupled with” are used synonymously.
- Figure 1 generally depicts an embodiment in which harmonic analysis is used to process a monaural signal in a manner than can mimic a stereo experience.
- a microphone, camera, neuro-headband, or other monaural (single channel) input 10 is decomposed using a Fourier transform into real and imaginary components by computer processor 20 operating suitable software.
- the real component is rendered at a first interface 32
- the imaginary component is rendered at a second interface 34.
- the interfaces 32, 34 comprise speakers
- the speakers are preferably separated by a suitable distance from each other so that a listener 40 can spatially distinguish sounds emanating from interfaces 32 relative to sounds emanating from speaker 34.
- speakers can be located in headphones, ear buds or otherwise placed adjacent the listener's ears, or could be placed at some distance from the listener, preferably arms length or farther.
- Speakers can also each independently include multiple diaphragms or other sound making apparatus, as for example a tweeter and a woofer.
- the interfaces 32, 34 comprise visual displays
- the displays are preferably located such that each eye sees only one of the interfaces. This can be accomplished using a device such as Google GlassTM that has two different display screens, but other options are also available.
- a desktop, laptop, tablet or even cell phone display screen could separately render images of the first and second components, and a barrier could be used so that only the left eye of the viewer sees the image of the first component, and only the right eye of the viewer sees the image of the second component.
- computer processor 20 can be configured to mix the first and second (usually real and imaginary) components under control of the listener or other operator, so that for example, the sounds, images or other output from the first interface 32 might comprise 80% first component and 20% second component, perhaps 60%> first component and 40%> second component, or perhaps 90%> first component minus 10%> second component. Indeed all possible combinations of the two components are contemplated, including those that vary over time.
- interfaces 32, 34 could utilize different modalities.
- interface 32 could be a speaker
- interface 34 could be a visual display, or perhaps a vibrating crystal.
- binaural phasing While not desiring to be limited to any particular theory or mode of operation, it is completed that the improved experience perceived by a listener using a system such as that shown in Figure 1 is due to a principle known as binaural phasing.
- binaural beats occur when two simple frequencies of beats are presented separately but concurrently to a listener. For example, when listening with one ear to 100 Hz beats, and with the other ear to 107 Hz beats, the user perceives beats at a third frequency of 7 Hz. Of course that third, perceived frequency is simply the difference between the two actual frequencies. Somehow, the brain "hears" it.
- Binaural phasing is different in that whereas both channels have the same frequencies, but those frequencies are in phase with each other. The brain is forced into resolving the differences of phase rather than frequency. One channel is carrying the real partials and the other is carrying the imaginary ones. Transforming From A Multiple Inputs
- Figure 2 expands upon the concepts of Figure 1 to situations where transforms are performed on more than one input.
- Figure 2 generally depicts an embodiment in which two-channel audio signals are used to improve upon a traditional stereo listening experience.
- the two-channel signals can be derived from any suitable source, including for example, left and right channels of a stereo player, a mixing board, or left and right microphones.
- computer processor 20 is configured to decompose the signals from inputs 62, 64 into real and imaginary components, which produced left real and imaginary components, and right real and imaginary components.
- either of the output channels could contain any ratio of the signals among these four discreet streams.
- what is routed to speaker 32 could be some amount of the left real component minus some amount of the right imaginary component
- what is routed to speaker 34 could be some amount of the right real component minus some amount of the left imaginary component.
- the amounts referred to in this paragraph can range from some very small percentage such as 0.1% to 100%.
- transforms are performed on multiple input channels, it is desirable but not essential that the same transform is used on the input of each of the channels. It is contemplated therefore in some embodiments to use a first transform on channel 1 , a different transform on channel 2, and still another transform on channel 3. Outputs of those transforms can still be rendered to a user as separate sound streams, images, vibration patterns, etc, or additively or subtractively combined in any desired ratios.
- Figure 3 expands upon the concepts of Figures 1 and 2 to situations in which an input source is used to feed both temporal and spectral reinforcement transforms.
- the resulting components are combined into a single stereo output pair (with a
- Temporal Mix control regulating the level of the temporal transform which is mixed into the output).
- the temporal transform takes the input source, and synthesizes a full complement of frequency partials rotated 90 degrees out of phase from the source (i.e. the difference between real & imaginary).
- stage 2 real (stage 2): 0.06338, .0.83714, 1. -0.83774,0.06338
- spectral reinforcement utilizes a variation of the spectral transform wherein individual real & Imaginary partial pairs are selectively attenuated or reinforced based upon their relationship to the whole.
- Three circular registers buffers
- the algorithm continuously monitors the magnitudes and phases of the real and imaginary pairs both (a) present at input and (b) within their accumulator loops using Cartesian to polar conversion.
- the partial of maximum amplitude within the present iteration triggers multiplication of an attenuation factor of 1/phi (i.e. the reciprocal of the Fibonacci constant) to that partial's feedback register entry.
- a microphone picks up the sounds of a user's breath or voice.
- the signal is processed in near real time, the noise removed, and the temporal signal split into real and imaginary components, and fed back to separate ears as described above.
- the spectral components of real and imaginary can fed back to separate ears sonically reinforcing those partials that are weak and need support to optimize the signal to its fullest potential.
- Figure 4 generally depicts screen prints of various portacles.
- Column A depicts a temporal chromaphase portacle (top) and a temporal magnitude portacle (bottom).
- Column B depicts a temporal chromaphase accumulator (top) and a temporal magnitude accumulator (bottom).
- Column C depicts a real chromaphase matrix (top) and a real magnitude matrix (bottom).
- Column D depicts a complex chromaphase matrix (top) and a complex magnitude matrix (bottom).
- Column E depicts an imaginary chromaphase matrix (top) and an imaginary magnitude matrix (bottom).
- components derived from transforms can be combined in myriad different ways. Such combinations can be visualized in various shapes and dimensions from a zero dimensional point to a line (1 dimension), circle (2 dimensions), sphere (3 dimensions) and torus (4 dimensions). Examples of toroidal representations of components produced by The PortacleTM software are shown in Figure 5A (single torus), Figure 5B (double torus) and Figure 5C (double torus torsion).
- a balanced real and imaginary dynamic flow represents as a torus. Under certain circumstances, however, the torus can undergo specific deformations. Below is a description of four of these deformations, which are collected in Figures 6A - 6D. There are many more.
- Composition refers to the completeness of connective vectors between the cells comprising the toroidal form. As composition decreases, gaps become evident in the interconnectedness of toroidal space.
- Extension refers to the normalized (full) magnitude of the toroidal form. As
- Transposition refers to toroidal shifts in phase relative to its origin. As transposition increases, the volume of the toroidal form moves above or below its origin, ultimately developing a double vortex thread (one of which is situated at origin) as it passes through vesica piscis into nested spheres.
- Convolution refers to the cross correlation of coordinates forming the toroidal manifold. As convolution increases, coordinates begin to wrap around in phase, progressively twisting the torus into the "knotted" form of a lemniscate.
- Figure 7 is a two-dimensional representation of a twelve-dimensional toroidal space based upon a person's breath. Each pixel represents one of the frequencies or partials. Its location is determined from the r:i (real imaginary) coefficients, the color (omitted from the application) is the phase (or relationship between the peaks and valleys of the wave) and the brightness represents the volume, amplitude or magnitude of the partial.
- points, lines, circles, ellipses, squares and other geometries have specific meanings. Different attracting zones scattered around the toroidal surface are indicative of traps and distractions that arise due to the point, line, circle, toroidal and strange attractors.
- the toroidal space is thus a map of the internal structure and can be interpreted as revealing holes in the pattern.
- providing a person with feedback based upon real or imaginary components of biometrics can provide entertainment, as for example where players try to manipulate the contours of the representations.
- Figures 8a, 8b, and 8c are the real and imaginary components of biometric waves mapped to x and y axes of a matrix, respectively. According to a preferred interpretation, these matrices include banding regions that correlate to psychological and mental states / conditions.
- Figure 8a is thought to indicate a preponderance of imaginary partials (vertical ellipse) that are centered in the present moment (center axis), outward focus (outer band predominance).
- Figure 8b is thought to indicate an extrospective state (outer banding), preponderance of real components (horizontal ellipse) and grounding at the core level (inner-banding).
- figure 8c is thought to indicate a balance between the real and imaginary components (circles instead of ellipses), lots of grounding in the core and self (central and middle bands).
- one or more physical controls i.e. the keyboard, a joystick or such
- the physical controls could optimally be used in tandem with the psiometric controls, the progression of game play being governed by a floating balance between the two.
- any n-dimensional array can be expressed as a series of waveforms.
- each horizontal row and vertical column of pixels forms a wave wherein pixel luminance corresponds to waveform sample amplitude.
- pixel luminance corresponds to waveform sample amplitude.
- the 256 rows and 256 columns would comprise a series of 512 waveforms. Applying a transform to those waveforms one can derive a complex (x+yi) image (bottom left), which matches the original, a real component image (bottom middle) and an imaginary component image (bottom right).
- An example is show in Figure 10.
- a person's brain can be used as an analog computer, with their body being used as an antennae in order to connect to the collective entangled mind of the planet.
- This concept is supported by the research of Dean Radin of the Noetic Society (IONS) and others regarding random number generators and their relationship to the collective mind and their belief that this is evidence of such.
- This research was recently reported in a book called Entangled Minds: Extrasensory Experiences in a Quantum Reality by Dean Radin.
- the co-inventors of this application have designed and built a device called the human neurocomputer.
- the neural interface hardware comprises two sensors, which are typically placed to either side equidistant to the center of the forehead, and a ground, which rests over the ear (although alternative configurations may be used depending upon the observing organism).
- the two sensors pick up the raw neurological and/(or) bioelectric signals emanating from the observer and send them to processing equipment at a sampling rate of 500Hz (along with supplementary signal analysis data processed by the interface).
- sampling is at 500Hz with the current technology, additional sample rates are contemplated, including faster rates that may become available in the future depending on the bioenergetic interface technology used.
- Preferred processing equipment comprises The PortacleTM Human NeurocomputerTM, which is a proprietary software program which is currently operable on a general purpose AppleTM computer. Operation on other computers is also contemplated.
- the raw data is routed to the neurocomputing algorithm within The Portacle, wherein its components of DC & GD (great diesis) through 250Hz (B# below middle-C) are separated into real & imaginary streams using the temporal transform.
- DC & GD greater diesis
- 250Hz B# below middle-C
- the user can select how much of the input signal becomes integrated into the portacles, matrices and spectral reinforcement using a Neural Mix control. With the right combination of settings, The Portacle can be run on bioelectric signals alone.
- a matched-mode signal encompassing a continuous range of all possible outcomes is sent to the observer as generated by the adaptive stochastic oscillator; the components of which may be apprehended as visual, auditory and vibrational stimuli, as well as sensitivity modulation (i.e. how much "attention" the neurocomputing algorithm is “paying” to the observer).
- This feedback loop continues until wave function collapse occurs. This takes advantage of a quantum mind-body problem, wherein an observer subject to a state vector encompassing multiple simultaneous possibilities will reduce these to a singular outcome consistent with their need for resolution in observational conflict.
- the Adaptive Stochastic Oscillator can be used as a stand-alone tool for transmitting highly targeted subsonic deterministic frequencies with frequency, phase and magnitude precision, as well as stochastic elements, to the human body.
- the first is one called Whole Body Vibration that deals with a device that one stands on to receive very low frequency vibrations that are applied to the entire body.
- the body oscillates up and down at specific frequencies that have been shown to have determinable results and that can target highly specific parts of the human body and mind.
- the second support comes from documented and declassified research that US Army undertook during the 60 's regarding the effect of low frequencies on soldiers under transit across rough water, land or air in different types of vehicles. This research pointed out specific frequencies that have harmful effects to humans.
- NLP Neuro-Linguistic Programming
- Figure 11 generally depicts a screen print of an interface 100 of an instance of The PortacleTM, through which a user can select and modify different display views (prilling).
- Radio buttons along the left 111 - 114 are used to select combinations of real or imaginary components for prilling.
- the first radio button 121 on the right selects the 1 st harmonic, which is in a particular instance the single torus of Figure 5 A.
- the second radio button 122 selects the 2 nd harmonic, which turns the single torus of Figure 5 A into the double torus of Figure 5B.
- Field 131 accepts a numeric input for phase differential
- field 132 accepts numeric input for ratio of torsion to rotation.
- the interface 100 should be interpreted as being driven by any suitable device, including a laptop, cell phone, tablet, phablet, etc, and using any suitable operating system. Of particular interest is an iPhoneTM or other device running a version of iOSTM.
Landscapes
- Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Acoustics & Sound (AREA)
- Signal Processing (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Complex Calculations (AREA)
- Transmission Systems Not Characterized By The Medium Used For Transmission (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
A signal is decomposed into different components using a transform, with the components then being separately presented to a person in a manner that produces a different cognitive experience than would have resulted from either (a) presentation of the original signal, or (b) presentation of a fully synthesized (inverse transformed) signal.
Description
BI-PHASIC APPLICATIONS OF REAL & IMAGINARY SEPARATION, AND
REINTEGRATION IN THE TIME DOMAIN
[0001] This application claims priority to US patent application ser. no. 13/910399, filed June 5, 2013 and US provisional application ser. no. 61/831540, filed June 5, 2013. This and all other referenced extrinsic materials are incorporated herein by reference in their entirety. Where a definition or use of a term in a reference that is incorporated by reference is inconsistent or contrary to the definition of that term provided herein, the definition of that term provided herein is deemed to be controlling.
Field of the Invention
[0002] The field of the invention is use of mathematical transforms in rendering of information. Background
[0003] The following description includes information that may be useful in understanding the present invention. It is not an admission that any of the information provided herein is prior art or relevant to the presently claimed invention, or that any publication specifically or implicitly referenced is prior art.
[0004] Harmonic analysis involves the representation of functions or signals as superpositions of basic waves. Harmonic analysis has found application in a great many fields, including signal processing, quantum mechanics, and neuroscience.
[0005] Fourier analysis is a subset of harmonic analysis, in which signals are decomposed into real and imaginary components using a transform. The terms "transforms" and "transformation" are used herein to mean decomposition of a signal into a multiple components in the same or a different domain. For example, a Hilbert transform converts a function in one domain into a function in the same domain. In contrast, a Fourier Series or a Discrete-time Fourier transform (DTFT) transform a time series into a frequency spectrum. Transforms can be applied to time domains, spatial frequencies, and indeed to nearly any function domain. Various transforms are currently used for compression, filtering, frequency balancing, encryption, and for other purposes. Although transforms are usually mathematically based, transforms can also
implemented in electronics, as for example, using a parallel pair of serially cascaded biquad filters.
[0006] Components produced by transformation can be processed separately, and synthesized (inverse-transformed) back together again. Transformed signals are not, however, always synthesized back to equivalent originals. MP3 compressed audio files, for example, contain only the real component of the original signal, not the imaginary component, and thus sustain significant loss in sound quality when being rendered to a listener. Additional losses can arise from the compression technology, resulting in sizzling, distortion, and flat, two dimensional sounds. Thus, there is a need to present such transformed audio files to a listener in a manner that is at least perceived to have a quality closer to that of the source of the original recording than a standard rendering.
[0007] In the case of video files, typical PEG compression applies a variant of a Fourier transformation (discrete cosine transform) to small square pieces of a digital image. The Fourier components of each square are rounded to lower arithmetic precision, and weak components are eliminated entirely, so that the remaining components can be stored very compactly. In normal image reconstruction, each image square is reassembled from the preserved approximate Fourier- transformed components, which are then synthesized to produce an approximation of the original image. Although rendering of a PEG-compressed file includes both components, current display technologies can cause blur and other distortions, due to inadequate pixel response time on LCD displays, resolution sampling methods, telecine processing by studios, and compression artifacts. These problems are especially pronounced with High Definition 4K and other large files. With the advent of LCD displays, motion blur has become even more of a problem due to sample-and- hold nature of the displays.
[0008] Several attempts have been made to resolve these distortions with respect to video files. ClearLCD™ and Clear Motion Rate™ technologies from Philips™ and Samsung™, for example, use a strobed backlight to reduce blurring. However, the existing solutions are limited to specific applications rather than being globally applicable. Thus, there is also a need to render compressed video files to a viewer in a manner that is at least perceived to have a quality closer to that of the source of the recording than a standard rendering.
[0009] With respect to biometrics, it is known to use brain waves to control physical or virtual objects, or to achieve a particular mental state, as for example a delta sleep state. Typically, this is accomplished by using the waves to trigger a beep, color on a display, movement of a mechanical arm, or other highly simplified indicia of a desired result. In so doing, a great deal of useful information about the subject's current psychology and cognition is eliminated. There is consequently a need to provide much more sophisticated feedback to a brain wave subject than is currently known.
[0010] Regardless of what type of signals are being processed (auditory, video, brain waves, etc) there is still a problem with the speed in which transforms and synthetic operations can be applied to complex signals, especially on a consumer device such as a laptop, tablet or cellphone. Presenting results to a listener, viewer, subject or other user even five, two, or one second after the signal is generated (or rendered from a data file) may be too slow to provide sufficient feedback to adequately manipulate the rendering in what appears to be a real-time fashion. Thus, there is a need for faster processing hardware and software to achieve the appearance of real- time operation and feedback.
Summary of The Invention
[0011] The inventive subject matter herein provides apparatus, systems and methods in which a signal is decomposed into different components using a transform, with the components then being separately presented to a person in a manner that produces a different cognitive experience than would have resulted from either (a) presentation of the original signal, or (b) presentation of a fully synthesized (inverse transformed) signal.
[0012] Some embodiments of the inventive subject matter deal with audio signals. For example, the real and imaginary components of a decomposed monaural audio signal can be presented separately to a listener through headphones, so that the sound of one component is rendered in one ear, and the sound of the other component is rendered in the other ear. Given the out of phase nature of the two components, the listener's brain will try to integrate the sounds in a manner that has been found to produce a much richer experience than if the listener had heard (a) only one of the components, or (b) a signal resulting from a synthesis (reverse transform) of the real and imaginary components.
[0013] Where the source includes only one of the components, as in an MP3 file that only includes the real component, an electronic processor can generate a corresponding imaginary component from the real component, and then the two components can be rendered separately to the listener's left and right ears. Here again, real world testing demonstrates noticeable improvement in the cognitive experience of the listener.
[0014] Throughout the discussion herein, numerous references will be made regarding processors, servers, services, interfaces, portals, platforms, or other systems formed from computing devices. It should be appreciated that the use of such terms is deemed to represent one or more computing devices having at least one processor configured to execute software instructions stored on a computer readable tangible, non-transitory medium. For example, a server can include one or more computers operating as a web server, database server, or other type of computer server in a manner to fulfill described roles, responsibilities, or functions.
[0015] Aspects of the inventive subject matter can also be applied to video files or still images. In those instances separate viewing of the separate components can be accomplished with individual displays such as those on Google™ Glass™, or by employing glasses or contacts that filter the superimposed components differently for the left and right eyes.
[0016] In the case of biometric signals, components derived from transformation of the waves can be presented to the subject as separate sounds or images, and the subject producing the waves can learn to manipulate them using the component renderings as feedback. In some embodiments, a subject can manipulate his/her thoughts to effectively collapse or otherwise modulate an auditory or visual representation of a wave function corresponding to biometric signal. Where the biometric signal comprises brain waves, this can be useful as part of an analog neurocomputer. Other biometrics beside brainwaves can also be used to provide feedback, however, including for example represent breathing or heart rates. [0017] It is still further contemplated that components derived from transformation can be rendered to a user other than through the ears or eyes. For example, the different components could be rendered to a user tactually, through skin vibrations applied to different sides of the body, or even different areas of the same side of the body, respectively. In other embodiments, the different components could be rendered to a user though electrical or thermal stimulation to
the skin or deeper tissues. Here again the stimulation could apply the different components to different sides or areas of the body, respectively
[0018] Additionally or alternatively to rendering the different components to different sides or areas of the body, it is contemplated that the different components could be rendered to the viewer in rapid sequence, e.g., a repeating sequence of the real component for l/50th of a second, followed by imaginary component for 1/50th of a second. Most preferably the listener could control the time periods separately, along with duration of any resting periods. Such "flickering" could apply to auditory, visual, tactual, or any other renderings.
[0019] It is also contemplated that the different components could be mixed in various percentages, preferably in near real time by the user, through operation of software. In the case of an audio file, for example, the sound rendered to the left ear could be 100% real, 90%> real and 10% imaginary, 80% real and 20% imaginary, etc., all the way to 100% imaginary. In interpreting these numbers it should be appreciated that the recitation of ranges herein is merely intended to serve as a shorthand method of referring individually to each separate value falling within the range. Thus, unless otherwise expressly indicated, each individual intervening value is incorporated into the specification as if it were individually recited herein. In addition, numeric values set forth herein should be construed in light of the number of reported significant digits, and by applying ordinary rounding techniques.
[0020] As used herein, the term "near real time" as applied to a transform process means that rendering of at least one of the components produced by the process will tend to lag no more than five seconds from input of the signal being processed, more preferably no more than two seconds from input of the signal being processed, and most preferably no more than one second from input of the signal being processed. Where an operator is manipulating mixing or other characteristics of rendered components, as for example in moving a control bar in an electronic interface, the term "near real time" means that rendering of at least one of the components lags no more than five seconds behind an operator-triggered manipulation, more preferably no more than two seconds behind an operator-triggered manipulation, and most preferably no more than one second behind an operator-triggered manipulation.
[0021] It is still further contemplated that the "different" cognitive experiences referred to herein are not necessarily desirable. Where the signal being processed is a breath, or background noise, for example, a listener might find the experience of renderings according to the inventive concepts herein to be better, worse, or substantially the same as the original sounds. In the case of a video or other image rendering, it may well be that separately presenting the different components could be very confusing or otherwise troubling to a viewer.
[0022] All methods described herein can be performed in any suitable order unless otherwise indicated herein or otherwise clearly contradicted by context. The use of any and all examples, or exemplary language (e.g. "such as") provided with respect to certain embodiments herein is intended merely to better illuminate the invention and does not pose a limitation on the scope of the invention otherwise claimed. No language in the specification should be construed as indicating any non-claimed element essential to the practice of the invention.
[0023] As used in the description herein and throughout the claims that follow, the meaning of "a," "an," and "the" includes plural reference unless the context clearly dictates otherwise. Also, as used in the description herein, the meaning of "in" includes "in" and "on" unless the context clearly dictates otherwise.
[0024] Groupings of alternative elements or embodiments of the invention disclosed herein are not to be construed as limitations. Each group member can be referred to and claimed individually or in any combination with other members of the group or other elements found herein. One or more members of a group can be included in, or deleted from, a group for reasons of convenience and/or patentability. When any such inclusion or deletion occurs, the
specification is herein deemed to contain the group as modified thus fulfilling the written description of all Markush groups used in the appended claims.
[0025] Various objects, features, aspects and advantages of the inventive subject matter will become more apparent from the following detailed description of preferred embodiments, along with the accompanying drawing figures in which like numerals represent like components.
Brief Description of The Drawings
[0026] Fig. 1 is a diagram of an embodiment in which harmonic analysis is used to process a monaural audio signal in a manner than can mimic a stereo listening experience.
[0027] Fig. 2 is a diagram of an embodiment in which two-channel audio signals are used to improve upon a traditional stereo listening experience.
[0028] Fig. 3 is a diagram of an embodiment in which an input source is used to feed both temporal and spectral reinforcement transforms.
[0029] Fig. 4 is a collection of screen prints of various portacles.
[0030] Fig.5A is a single torus created according to teachings herein.
[0031] Fig.5B is a double torus created according to teachings herein.
[0032] Fig.5C is a double torus torsion created according to teachings herein.
[0033] Figs. 6A - 6D are graphic representations of toroidal reformations created according to teachings herein.
[0034] Fig. 7 is a two-dimensional representation of a twelve-dimensional toroidal space based upon a person ' s breath.
[0035] Figs. 8a, 8b, and 8c are real and imaginary components of biometric waves mapped to x and y axes of a matrix, respectively
[0036] Fig. 9 is a collection of images, showing source, complex, real and imaginary components.
[0037] Fig. 10 is a screen shot of the source image of Fig. 9, enhanced by a visual representation of a mental state.
[0038] Fig. 11 is a screen print of an interface 100 of an instance of The Portacle'
Detailed Description
[0039] The following discussion provides many example embodiments of the inventive subject matter. Although each embodiment represents a single combination of inventive elements, the inventive subject matter is considered to include all possible combinations of the disclosed elements. Thus if one embodiment comprises elements A, B, and C, and a second embodiment comprises elements B and D, then the inventive subject matter is also considered to include other remaining combinations of A, B, C, or D, even if not explicitly disclosed.
[0040] As used herein, and unless the context dictates otherwise, the term "coupled to" is intended to include both direct coupling (in which two elements that are coupled to each other contact each other) and indirect coupling (in which at least one additional element is located between the two elements). Therefore, the terms "coupled to" and "coupled with" are used synonymously.
Transforming From A Single Input
[0041] Figure 1 generally depicts an embodiment in which harmonic analysis is used to process a monaural signal in a manner than can mimic a stereo experience. Here, a microphone, camera, neuro-headband, or other monaural (single channel) input 10 is decomposed using a Fourier transform into real and imaginary components by computer processor 20 operating suitable software. The real component is rendered at a first interface 32, and the imaginary component is rendered at a second interface 34. [0042] Where the interfaces 32, 34 comprise speakers, the speakers are preferably separated by a suitable distance from each other so that a listener 40 can spatially distinguish sounds emanating from interfaces 32 relative to sounds emanating from speaker 34. For example, speakers can be located in headphones, ear buds or otherwise placed adjacent the listener's ears, or could be placed at some distance from the listener, preferably arms length or farther. Speakers can also each independently include multiple diaphragms or other sound making apparatus, as for example a tweeter and a woofer.
[0043] Where the interfaces 32, 34 comprise visual displays, the displays are preferably located such that each eye sees only one of the interfaces. This can be accomplished using a device such
as Google Glass™ that has two different display screens, but other options are also available. For example, a desktop, laptop, tablet or even cell phone display screen could separately render images of the first and second components, and a barrier could be used so that only the left eye of the viewer sees the image of the first component, and only the right eye of the viewer sees the image of the second component.
[0044] Optionally, computer processor 20 can be configured to mix the first and second (usually real and imaginary) components under control of the listener or other operator, so that for example, the sounds, images or other output from the first interface 32 might comprise 80% first component and 20% second component, perhaps 60%> first component and 40%> second component, or perhaps 90%> first component minus 10%> second component. Indeed all possible combinations of the two components are contemplated, including those that vary over time.
[0045] Also, interfaces 32, 34 could utilize different modalities. For example, interface 32 could be a speaker, and interface 34 could be a visual display, or perhaps a vibrating crystal.
Binaural Phasing
[0046] While not desiring to be limited to any particular theory or mode of operation, it is completed that the improved experience perceived by a listener using a system such as that shown in Figure 1 is due to a principle known as binaural phasing. Previously known binaural beats occur when two simple frequencies of beats are presented separately but concurrently to a listener. For example, when listening with one ear to 100 Hz beats, and with the other ear to 107 Hz beats, the user perceives beats at a third frequency of 7 Hz. Of course that third, perceived frequency is simply the difference between the two actual frequencies. Somehow, the brain "hears" it.
[0047] Binaural phasing is different in that whereas both channels have the same frequencies, but those frequencies are in phase with each other. The brain is forced into resolving the differences of phase rather than frequency. One channel is carrying the real partials and the other is carrying the imaginary ones.
Transforming From A Multiple Inputs
[0048] Figure 2 expands upon the concepts of Figure 1 to situations where transforms are performed on more than one input. As representative of the general category of multiple inputs, Figure 2 generally depicts an embodiment in which two-channel audio signals are used to improve upon a traditional stereo listening experience. The two-channel signals can be derived from any suitable source, including for example, left and right channels of a stereo player, a mixing board, or left and right microphones. Here, computer processor 20 is configured to decompose the signals from inputs 62, 64 into real and imaginary components, which produced left real and imaginary components, and right real and imaginary components. [0049] It is contemplated that either of the output channels could contain any ratio of the signals among these four discreet streams. Thus, what is routed to speaker 32 could be some amount of the left real component minus some amount of the right imaginary component, and what is routed to speaker 34 could be some amount of the right real component minus some amount of the left imaginary component. The amounts referred to in this paragraph can range from some very small percentage such as 0.1% to 100%.
[0050] Where transforms are performed on multiple input channels, it is desirable but not essential that the same transform is used on the input of each of the channels. It is contemplated therefore in some embodiments to use a first transform on channel 1 , a different transform on channel 2, and still another transform on channel 3. Outputs of those transforms can still be rendered to a user as separate sound streams, images, vibration patterns, etc, or additively or subtractively combined in any desired ratios.
Temporal Transform With Spectral Reinforcement
[0051] Figure 3 expands upon the concepts of Figures 1 and 2 to situations in which an input source is used to feed both temporal and spectral reinforcement transforms. In this particular example, the resulting components are combined into a single stereo output pair (with a
Temporal Mix control regulating the level of the temporal transform which is mixed into the output).
[0052] The temporal transform takes the input source, and synthesizes a full complement of frequency partials rotated 90 degrees out of phase from the source (i.e. the difference between real & imaginary). As with the spectral transform, the temporal routes the discreet real & imaginary results to the left & right outputs, respectively. Its function may be approximated by two pairs of two-pole, two-zero biquad filters in series as implemented by the following equation: y[n] = aO * x[n] + al *x[n-l) + a2 * x[n-2] - bl * y[n-l] - b2 * y[n-2]
using the following coefficients for aO, al, a2, bl & b2: real (stage 1): 0.94657, - 1 ,94632, 1. -1.94632, 0.94657
real (stage 2): 0.06338, .0.83714, 1. -0.83774,0.06338
Imaginary (stage 1 } : .0.260502, 0.02569,1,0.02569, -0.260502
Imaginary {stage 2} : 0.870686, -1.8685, 1, .1.8685,0.870686
In this example, spectral reinforcement utilizes a variation of the spectral transform wherein individual real & Imaginary partial pairs are selectively attenuated or reinforced based upon their relationship to the whole. Three circular registers (buffers) are employed to hold the real, imaginary and feedback {attenuation} coefficients. The algorithm continuously monitors the magnitudes and phases of the real and imaginary pairs both (a) present at input and (b) within their accumulator loops using Cartesian to polar conversion. The partial of maximum amplitude within the present iteration triggers multiplication of an attenuation factor of 1/phi (i.e. the reciprocal of the Fibonacci constant) to that partial's feedback register entry. Conversely, the minimum phase of that iteration is mapped exponentially to the register indices, fully restoring the feedback register element corresponding to that phase to its maximum amplitude. During each period of iteration, the real & imaginary accumulators are multiplied by the updated feedback register to amplify or attenuate their contents accordingly. [0053] Whereas real and imaginary coefficient are unique to the spectral domain, preferred embodiments maintain these as separate and discrete components after transformation back into the time domain. Magnitudes (r) and phases (Θ) within the complex waveform can be calculated as shown in the following equations"
r = (x +y )
θ = atan2(y,x)
[0054] In a simple example, a microphone picks up the sounds of a user's breath or voice. The signal is processed in near real time, the noise removed, and the temporal signal split into real and imaginary components, and fed back to separate ears as described above. At the same time the spectral components of real and imaginary can fed back to separate ears sonically reinforcing those partials that are weak and need support to optimize the signal to its fullest potential.
Portacles
[0055] Real and imaginary components can be plotted visually using Cartesian or other coordinates. The resulting plots are referred to as "portacles" herein.
[0056] Figure 4 generally depicts screen prints of various portacles. Column A depicts a temporal chromaphase portacle (top) and a temporal magnitude portacle (bottom). Column B depicts a temporal chromaphase accumulator (top) and a temporal magnitude accumulator (bottom). Column C depicts a real chromaphase matrix (top) and a real magnitude matrix (bottom). Column D depicts a complex chromaphase matrix (top) and a complex magnitude matrix (bottom). Column E depicts an imaginary chromaphase matrix (top) and an imaginary magnitude matrix (bottom).
Toroidal Representations
[0057] It is well known in science that blasting a crystal with a beam of x-rays reveals the quantum latticework created by positioning of the atoms inside the crystal. In an analogous manner, one can mathematically "shine" a beam of coherent white light through a portacle and then a lens, which reduces the image to a zero dimensional point. One can then mathematically place the point inside of a torus, which is then torsioned to populate all surfaces with the real and imaginary coefficients. Toroidal space (area of a torus) is a preferred surface topology to graphically and visually show a dynamic energy field.
[0058] Surprisingly, we have found that if the portacle was created using breath, brainwaves or other biometrics, the colors and patterns present on the surface of such a multi-dimensional torus
can correspond more or less with intangible aspects of a person's personality. As of the filing of the current application, a 12 dimensional torus is thought to be optimal for that purpose.
[0059] As noted above, components derived from transforms can be combined in myriad different ways. Such combinations can be visualized in various shapes and dimensions from a zero dimensional point to a line (1 dimension), circle (2 dimensions), sphere (3 dimensions) and torus (4 dimensions). Examples of toroidal representations of components produced by The Portacle™ software are shown in Figure 5A (single torus), Figure 5B (double torus) and Figure 5C (double torus torsion).
[0060] A balanced real and imaginary dynamic flow represents as a torus. Under certain circumstances, however, the torus can undergo specific deformations. Below is a description of four of these deformations, which are collected in Figures 6A - 6D. There are many more. a. Composition refers to the completeness of connective vectors between the cells comprising the toroidal form. As composition decreases, gaps become evident in the interconnectedness of toroidal space. b. Extension refers to the normalized (full) magnitude of the toroidal form. As
extension decreases, the volume of the torus diminishes, thus limiting its inward and outward radiance. c. Transposition refers to toroidal shifts in phase relative to its origin. As transposition increases, the volume of the toroidal form moves above or below its origin, ultimately developing a double vortex thread (one of which is situated at origin) as it passes through vesica piscis into nested spheres. d. Convolution refers to the cross correlation of coordinates forming the toroidal manifold. As convolution increases, coordinates begin to wrap around in phase, progressively twisting the torus into the "knotted" form of a lemniscate.
[0061] Even taking as input a biometric as simple as a person's breath or speech can lead to useful information. Figure 7 is a two-dimensional representation of a twelve-dimensional toroidal space based upon a person's breath. Each pixel represents one of the frequencies or
partials. Its location is determined from the r:i (real imaginary) coefficients, the color (omitted from the application) is the phase (or relationship between the peaks and valleys of the wave) and the brightness represents the volume, amplitude or magnitude of the partial.
[0062] It is contemplated that points, lines, circles, ellipses, squares and other geometries have specific meanings. Different attracting zones scattered around the toroidal surface are indicative of traps and distractions that arise due to the point, line, circle, toroidal and strange attractors. The toroidal space is thus a map of the internal structure and can be interpreted as revealing holes in the pattern.
Games And Entertainment
[0063] At the very least, providing a person with feedback based upon real or imaginary components of biometrics can provide entertainment, as for example where players try to manipulate the contours of the representations.
[0064] Figures 8a, 8b, and 8c are the real and imaginary components of biometric waves mapped to x and y axes of a matrix, respectively. According to a preferred interpretation, these matrices include banding regions that correlate to psychological and mental states / conditions. For example, Figure 8a is thought to indicate a preponderance of imaginary partials (vertical ellipse) that are centered in the present moment (center axis), outward focus (outer band predominance). Along the same lines, Figure 8b is thought to indicate an extrospective state (outer banding), preponderance of real components (horizontal ellipse) and grounding at the core level (inner-banding). Still further along the same lines, figure 8c is thought to indicate a balance between the real and imaginary components (circles instead of ellipses), lots of grounding in the core and self (central and middle bands).
[0065] Games and other forms of entertainment can readily be developed where players try to alter their thoughts, breath, and so forth to control the matrix. Guidance for such games could advantageously be based upon some or all of the following experimentally observed correlations:
[0066] Control Value: (SELF) AWARENESS a. Complex Components {x+yi} : Introspection (i) / Extrospection (r)
b. +r / -i = "looking" outward, extroversion, first attention c. -r / +i = "looking" inward, introversion, meditation
d. +r / +i = ultimate presence, force of perception, second attention e. -r / -i = "nirvana" or the imperturbable stillness of being
[0067] Control Value: (CORE) ENERGETICS
a. Complex Components {x+yi} : Ktisis (i) / Energes (r) b. +r / -i - ["body" in action, work, effort, expending energy] c. -r / +i - [conflict, polarity, differential advancement]
d. +r / +i - [active pursuit, dedicated support, involvement] e. -r / -i - Allasso (transformation), ["body" at rest]
[0068] Control Value: (MANIFOLD) INTEGRATION
a. Complex Components {x+yi} : Singularity (i) / Connectivity (r) b. +r / -i - manifold focus, [togetherness]
c. -r / +i - individual focus, [aloneness / isolation]
d. +r / +i - [fundamental interconnectedness with all things] e. -r / -i - [bilateral disconnection], [therapeutic / positive] "undoing"
[0069] Control Value: (TEMPORAL) ORIENTATION
a. Complex Components {x+yi} : Past (i) / Future (r)
b. +r / -i - [coming from / looking toward future]
c. -r / +i - [coming from / looking toward past]
d. +r / +i - [rooted in present but formulating from past and future, "temporal convergence"] e. -r / -i - [rooted wholly in present, eliminating personal history]
[0070] It is also contemplated that one or more physical controls (i.e. the keyboard, a joystick or such) be used in tandem with psiometric information derived from the biometrics. Among other things, the physical controls could optimally be used in tandem with the psiometric controls, the progression of game play being governed by a floating balance between the two.
Image As Series Of Waveforms
[0071] Any n-dimensional array can be expressed as a series of waveforms. In the case of a photograph or other 2d array, each horizontal row and vertical column of pixels forms a wave wherein pixel luminance corresponds to waveform sample amplitude. In the top image of
Figure 9, for example, the 256 rows and 256 columns would comprise a series of 512 waveforms. Applying a transform to those waveforms one can derive a complex (x+yi) image (bottom left), which matches the original, a real component image (bottom middle) and an imaginary component image (bottom right).
[0072] One can also combine video or still images with visual representations of real or imaginary components arising from breathing, heart rate, brain waves or other biometrics. An example is show in Figure 10.
Analog NeuroComputer
[0073] It is contemplated that a person's brain (and possibly other living organisms) can be used as an analog computer, with their body being used as an antennae in order to connect to the collective entangled mind of the planet. This concept is supported by the research of Dean Radin of the Noetic Society (IONS) and others regarding random number generators and their relationship to the collective mind and their belief that this is evidence of such. This research was recently reported in a book called Entangled Minds: Extrasensory Experiences in a Quantum Reality by Dean Radin.
[0074] Regardless of the merits of that concept, the co-inventors of this application have designed and built a device called the human neurocomputer.
[0075] The neural interface hardware comprises two sensors, which are typically placed to either side equidistant to the center of the forehead, and a ground, which rests over the ear (although alternative configurations may be used depending upon the observing organism). The two sensors pick up the raw neurological and/(or) bioelectric signals emanating from the observer and send them to processing equipment at a sampling rate of 500Hz (along with supplementary signal analysis data processed by the interface). Although sampling is at 500Hz with the current technology, additional sample rates are contemplated, including faster rates that may become available in the future depending on the bioenergetic interface technology used.
[0076] Preferred processing equipment comprises The Portacle™ Human Neurocomputer™, which is a proprietary software program which is currently operable on a general purpose Apple™ computer. Operation on other computers is also contemplated. The raw data is routed to the neurocomputing algorithm within The Portacle, wherein its components of DC & GD (great diesis) through 250Hz (B# below middle-C) are separated into real & imaginary streams using the temporal transform. Although all of the data can be used in computation, in preferred embodiments the user can select how much of the input signal becomes integrated into the portacles, matrices and spectral reinforcement using a Neural Mix control. With the right combination of settings, The Portacle can be run on bioelectric signals alone. [0077] When analog processing is required by The Portacle, a matched-mode signal encompassing a continuous range of all possible outcomes is sent to the observer as generated by the adaptive stochastic oscillator; the components of which may be apprehended as visual, auditory and vibrational stimuli, as well as sensitivity modulation (i.e. how much "attention" the neurocomputing algorithm is "paying" to the observer). [0078] This feedback loop continues until wave function collapse occurs. This takes advantage of a quantum mind-body problem, wherein an observer subject to a state vector encompassing multiple simultaneous possibilities will reduce these to a singular outcome consistent with their need for resolution in observational conflict. It is this "spark of ktisis" that can assist in yielding an answer to whatever question is being posed.
[0079] Games and other forms of entertainment are also contemplated with respect to the Neurocomputer. For example, one can transpose the sound signal coming from ones brain via the Neurocomputer up five octaves into the piano range. The resulting melodies are extremely intriguing and at the very least highly entertaining. Other transposition coefficients could also be used that maintain the harmonic series integrity of the source. Thus, five octaves is just and example, and one could also perform other transpositions, including for example, four octaves, four octaves and a major fifth, etc. Thus, any harmonic transposition interval can be employed in the process.
Adaptive Stochastic Oscillator
[0080] The Adaptive Stochastic Oscillator can be used as a stand-alone tool for transmitting highly targeted subsonic deterministic frequencies with frequency, phase and magnitude precision, as well as stochastic elements, to the human body.
[0081] Two different sciences support this technology. The first is one called Whole Body Vibration that deals with a device that one stands on to receive very low frequency vibrations that are applied to the entire body. The body oscillates up and down at specific frequencies that have been shown to have determinable results and that can target highly specific parts of the human body and mind. There are lists of these helpful frequencies that are well known and accepted by those in the field.
[0082] The second support comes from documented and declassified research that US Army undertook during the 60 's regarding the effect of low frequencies on soldiers under transit across rough water, land or air in different types of vehicles. This research pointed out specific frequencies that have harmful effects to humans.
[0083] Combined together, these two provide a list of frequencies that can be fed back to a person to assist them in a number of different ways physically, mentally, emotionally and perhaps spiritually. Our ASO (adaptive stochastic oscillator) can be utilized as a high definition frequency generator, which can produce a subsonic frequency that is then applied to the human body via a proprietary transducer and amplifier capable of producing such a low frequency. This would be a new more sophisticated method of whole body vibration.
Sensory Resonance
[0084] Sensory resonance is described in the book Harmonic Law: The Science of Vibration by one of the co-inventors herein, Don Estes. The goal is to synchronize the visual, auditory and kinesthetic sensory mechanisms into one coherent experience, bringing about a state of profound relaxation and intense inspiration at the same time. Figure 11 is a chart that summarizes the theory of Sensory Resonance.
[0085] In contrast to the book's use of NLP (Neuro-Linguistic Programming) to seek sensory resonance, it is contemplated herein that one could use representations of the real and imaginary components derived from transforming biometric waves. In especially preferred embodiments, such representations can be delivered by multi-sensory platforms that allow the individual to see, hear and/or feel the collective feedback of many cells in the body at the same time.
Software Interface
[0086] Figure 11 generally depicts a screen print of an interface 100 of an instance of The Portacle™, through which a user can select and modify different display views (prilling). Radio buttons along the left 111 - 114 are used to select combinations of real or imaginary components for prilling. The first radio button 121 on the right selects the 1st harmonic, which is in a particular instance the single torus of Figure 5 A. The second radio button 122 selects the 2nd harmonic, which turns the single torus of Figure 5 A into the double torus of Figure 5B. Field 131 accepts a numeric input for phase differential, and field 132 accepts numeric input for ratio of torsion to rotation. The interface 100 should be interpreted as being driven by any suitable device, including a laptop, cell phone, tablet, phablet, etc, and using any suitable operating system. Of particular interest is an iPhone™ or other device running a version of iOS™.
[0087] By way of example, the following are prilling equations for Toroidal Spaces
space = matrix(x,y,z)
wherein... dimensional flags [optional]:
dl = 1st dimensional flag (0 to 1)
d2 = 2nd dimensional flag (0 to 1)
d3 = 3rd dimensional flag (0 to 1)
d4 = 4th dimensional flag (0 to 1)
variables:
h = harmonic (1 to ...)
p = phase (0 to 2pi)
t = torsion (0 to -2pi)
u = outer loop (-pi to pi)
v = inner loop (-pi to pi)
toroidal space:w = sin(p+h*v+t)...
[or substitute Id input vector flow for prill]
x = cos(v+t)*w
y = sin(v+t)*cos(u-(t*2))*w
z = sin(v+t)*sin(u-(t*2))*w with dimensional flags:
w = sin(p+h*v+t)*d4+(l-d4)
[or substitute (Id input vector flow)*d4+(l-d4) for prill]
x = (cos(v+t)*dl+(l- dl))*w
y = ((8ίη(ν+ί)^2+(1-ά2))*(ΰθ8(υ-(ί*2))^3+(1-ά3))^^2)
z = ((8ίη(ν+ί)^3+(1-ά3))*(8ίη(υ-(ί*2))^3+(1-ά3))^^3).
Filtering Of Noise From A Signal
[0088] It is possible to improve the accuracy of frequency analysis by matching the sampling rate to the unit of measurement. Since frequency is measured in Hertz or (cycles per second) and the standard unit of measurement is 1 cycle per one second in the first octave, the sampling rate should reduce to the number one at the first octave. However, this is rarely the case with modern analysis because this connection between sampling rate and unit of measurement is not known in science. The standard measurement utilizes a 32, 64, 128 or 256K sampling rate, which when reduced to the first octave results in 0.9765625 Hz., not 1.0. This means that standard measurements are off by 768 Hz. in the 15th octave. This ratio is 125 :128, the reciprocal of which is known in music theory as the great diesis.
[0089] The first harmonic law states that all harmonically related frequencies must be whole number ratios of each other. It thus turns out that that the noise inherent within any signal can be removed by deleting any real or imaginary partial that is not a whole number ratio of the fundamental frequency. [0090] It should be apparent to those skilled in the art that many more modifications besides those already described are possible without departing from the inventive concepts herein. The inventive subject matter, therefore, is not to be restricted except in the spirit of the appended claims. Moreover, in interpreting both the specification and the claims, all terms should be interpreted in the broadest possible manner consistent with the context. In particular, the terms
"comprises" and "comprising" should be interpreted as referring to elements, components, or steps in a non-exclusive manner, indicating that the referenced elements, components, or steps may be present, or utilized, or combined with other elements, components, or steps that are not expressly referenced. Where the specification claims refers to at least one of something selected from the group consisting of A, B, C .... and N, the text should be interpreted as requiring only one element from the group, not A plus N, or B plus N, etc.
Claims
1. A method of presenting information to a user, the method comprising:
applying a transform to a waveform to produce at least first and second independently manipulable components; and
presenting to perceptively different sense areas of the user, renderings of first and second streams that have unequal representations of the first and second components, respectively.
2. The method of claim 1, further comprising deriving the waveform from a digital
representation of an image.
3. The method of claim 1, further comprising deriving the waveform from a digital
representation of a sound.
4. The method of claim 1, further comprising deriving the waveform from a digital
representation of a brain wave.
5. The method of claim 1, wherein the waveform comprises a discretely and finitely sampled signal.
6. The method of claim 5, wherein the waveform comprises a compressed media file.
7. The method of claim 1, wherein the waveform comprises a time domain signal, and the step of applying the transform produces a spectral domain signal.
8. The method of claim 1, wherein the transform comprises a Fourier transform.
9. The method of claim 1, wherein the transform is selected from the list consisting of a Hilbert transform, a Cochlea transform, a Hadamard transform, a Walsh transform, a Haar transform, and a wavelet transform.
10. The method of claim 1, wherein the transform results from operation of first and second parallel, serially cascaded biquad filters.
11. The method of claim 1 , wherein the first and second independently manipulable components comprise real and imaginary portions, respectively.
12. The method of claim 1, further comprising effecting the steps of applying the transform and presenting the first and second streams in real time.
13. The method of claim 1, further comprising providing an interface through which the user can alter relative amounts of the first and second components in the first stream, (e.g., portacle)
14. The method of claim 13, wherein the interface allows the user to alter relative amounts of the first and second components in the second stream.
15. The method of claim 13, wherein the interface allows the user to subtract at least a portion of the second component from at least a portion of the first component in deriving the first stream.
16. The method of claim 13, wherein ratios between representations of the first and second components in the first stream are modulated by a time -varying modulation source.
17. The method of claim 1, further comprising providing an interface through which the user can visualize a graphic representation of at least one of (a) the first and second components and (b) the first and second streams.
18. The method of claim 17, wherein the graphic representation comprises a torus.
19. The method of claim 17, wherein the graphic representation comprises a deformed torus, wherein the deformation is selected from the group consisting of a composition, an extension, a transposition and a convolution.
20. The method of claim 17, wherein the waveform is derived from an image, and the graphic representation comprises a composite of at least a portion of the image and a representation of at least one of the first and second streams.
21. The method of claim 1, further comprising providing an interface through which the user can visualize a graphic representation of a correlation of a psychological status of at least one of (a) the first and second components and (b) the first and second streams.
22. The method of claim 21, wherein the psychological status of the user is selected from the list consisting of awareness, core energetics, integration, and time.
23. The method of claim 1 , further comprising deriving the waveform from first and second sensors coupled to the forehead of the user.
24. The method of claim 1, further comprising establishing a feedback loop between (a) the waveform sensed by the sensor and (b) an input to the user derived from the first and second streams.
25. The method of claim 24, further comprising training the user to consciously manipulate the waveform to produce a desired effect on the input.
26. The method of claim 24, wherein the desired effect is collapse of a wave function
corresponding to the waveform.
27. The method of claim 24, further comprising using the feedback loop as part of an analog neurocomputer.
28. A system for presenting data to a user, wherein at least a portion of the data can be reduced to a waveform, comprising:
an analyzer configured to apply a mathematical transform to the waveform to produce at least first and second independently manipulable components;
a mixer configured to produce first and second streams having unequal representations of the first and second components, respectively; and
wherein the system is capable of providing the first and second streams to the user in real time.
29. The system of claim 28, wherein the transform comprises a Fourier transform.
30. The system of claim 28, wherein the transform is selected from the list consisting of a Cochlea transform, a Hadamard transform, a Walsh transform, a Haar transform, a wavelet transform.
31. The system of claim 28, wherein the first and second independently manipulable components comprise real and imaginary portions, respectively.
32. The system of claim 28, further comprising a math chip dedicated to performing the transform.
33. The system of claim 28, wherein the analyzer and the mixer are implemented on a device having a telephony capability.
34. The system of claim 28, wherein the analyzer and the mixer are implemented on a portable device having a display through which the user can operate an interface to alter relative amounts of the first and second components in the first stream.
35. The method of claim 34, wherein the interface allows the user to alter relative amounts of the first and second components in the second stream.
36. The method of claim 34, wherein the interface allows the user to subtract at least a portion of the second component from at least a portion of the first component in deriving the first stream.
37. The system of claim 28, further comprising a sensor that detects a time varying biometric used to construct the waveform.
38. The system of claim 28, further comprising at least first and second effectors configured to present the first and second streams, respectively, to perceptively different sense areas of the user.
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201361831540P | 2013-06-05 | 2013-06-05 | |
US13/910,399 US8788557B2 (en) | 2012-12-11 | 2013-06-05 | Bi-phasic applications of real and imaginary separation, and reintegration in the time domain |
US13/910,399 | 2013-06-05 | ||
US61/831,540 | 2013-06-05 |
Publications (2)
Publication Number | Publication Date |
---|---|
WO2015047466A2 true WO2015047466A2 (en) | 2015-04-02 |
WO2015047466A3 WO2015047466A3 (en) | 2015-05-21 |
Family
ID=52744660
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US2014/041126 WO2015047466A2 (en) | 2013-06-05 | 2014-06-05 | Bi-phasic applications of real & imaginary separation, and reintegration in the time domain |
Country Status (1)
Country | Link |
---|---|
WO (1) | WO2015047466A2 (en) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109359633A (en) * | 2018-12-10 | 2019-02-19 | 西北工业大学 | Combined signal classification method based on Hilbert-Huang transform and wavelet ridge |
CN111721770A (en) * | 2020-06-12 | 2020-09-29 | 汕头大学 | Automatic crack detection method based on frequency division convolution |
CN112287811A (en) * | 2020-10-27 | 2021-01-29 | 广州番禺职业技术学院 | Domain self-adaption method based on HSIC and RKHS subspace learning |
WO2021144751A1 (en) * | 2020-01-15 | 2021-07-22 | TGR1.618 Limited | Method for bi-phasic separation and re-integration on mobile media devices |
GB2613248A (en) * | 2020-07-24 | 2023-05-31 | Tgr1 618 Ltd | Method and device for processing and providing audio information using bi-phasic separation and re-integration |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108761202B (en) * | 2018-05-04 | 2020-07-17 | 上海电力学院 | Harmonic detection method combining pole symmetric modal decomposition and Hilbert transform |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5331222A (en) * | 1993-04-29 | 1994-07-19 | University Of Maryland | Cochlear filter bank with switched-capacitor circuits |
JP3103793B2 (en) * | 1997-03-11 | 2000-10-30 | 株式会社モノリス | Object structure graph generation device, data conversion device and system that can use the same |
US6224549B1 (en) * | 1999-04-20 | 2001-05-01 | Nicolet Biomedical, Inc. | Medical signal monitoring and display |
US20040210159A1 (en) * | 2003-04-15 | 2004-10-21 | Osman Kibar | Determining a psychological state of a subject |
US7164899B2 (en) * | 2003-09-16 | 2007-01-16 | Microtune (Texas), L.P. | System and method for frequency translation with harmonic suppression using mixer stages |
-
2014
- 2014-06-05 WO PCT/US2014/041126 patent/WO2015047466A2/en active Application Filing
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109359633A (en) * | 2018-12-10 | 2019-02-19 | 西北工业大学 | Combined signal classification method based on Hilbert-Huang transform and wavelet ridge |
WO2021144751A1 (en) * | 2020-01-15 | 2021-07-22 | TGR1.618 Limited | Method for bi-phasic separation and re-integration on mobile media devices |
GB2591111B (en) * | 2020-01-15 | 2023-08-09 | Tgr1 618 Ltd | Method for bi-phasic separation and re-integration on mobile media devices |
CN111721770A (en) * | 2020-06-12 | 2020-09-29 | 汕头大学 | Automatic crack detection method based on frequency division convolution |
GB2613248A (en) * | 2020-07-24 | 2023-05-31 | Tgr1 618 Ltd | Method and device for processing and providing audio information using bi-phasic separation and re-integration |
CN112287811A (en) * | 2020-10-27 | 2021-01-29 | 广州番禺职业技术学院 | Domain self-adaption method based on HSIC and RKHS subspace learning |
Also Published As
Publication number | Publication date |
---|---|
WO2015047466A3 (en) | 2015-05-21 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2015047466A2 (en) | Bi-phasic applications of real & imaginary separation, and reintegration in the time domain | |
US20060252978A1 (en) | Biofeedback eyewear system | |
Wöstmann et al. | Spatiotemporal dynamics of auditory attention synchronize with speech | |
US9078065B2 (en) | System and method for displaying sound as vibrations | |
Oscari et al. | Substituting auditory for visual feedback to adapt to altered dynamic and kinematic environments during reaching | |
Hanneton et al. | The Vibe: a versatile vision-to-audition sensory substitution device | |
Yoo et al. | Consonance of vibrotactile chords | |
US11877975B2 (en) | Method and system for multimodal stimulation | |
US20180224936A1 (en) | Brain to brain communication system for social media | |
CN113301872A (en) | Apparatus and method for improving perceptibility by sound control | |
Wesslein et al. | When vision influences the invisible distractor: tactile response compatibility effects require vision. | |
US8788557B2 (en) | Bi-phasic applications of real and imaginary separation, and reintegration in the time domain | |
KR20210020314A (en) | Apparatus and method for evaluating cognitive response of comparative sounds | |
US20180210550A1 (en) | Apparatus and Method for Brain-to-Brain Communication | |
WO2011147015A1 (en) | System and method for displaying sound as vibrations | |
CN204498284U (en) | The quarter-phase application apparatus of real part and imaginary part segmentation and reformation in time domain | |
CN112969409B (en) | Pure audio disturbance training for cognitive disorder screening and treatment | |
Eaton et al. | BCMI systems for musical performance | |
Kim | A Basic Study on the Conversion of Sound into Color Image using both Pitch and Energy | |
CN115227264B (en) | Method for displaying brain wave signals by using Clarnisse graph | |
Stella | Auditory display of brain oscillatory activity with electroencephalography | |
US20240017166A1 (en) | Systems and methods for generating real-time directional haptic output | |
Riahi et al. | VBTones a visual to auditory device for the blind | |
EP4437403A1 (en) | Tactile representation of location characteristics and content in 3d | |
Park | Data-Driven Multivariate and Multiscale Methods for Brain Computer Interface |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 14848189 Country of ref document: EP Kind code of ref document: A2 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 14848189 Country of ref document: EP Kind code of ref document: A2 |