US11974012B1 - Modifying audio and video content based on user input - Google Patents
Modifying audio and video content based on user input Download PDFInfo
- Publication number
- US11974012B1 US11974012B1 US18/501,524 US202318501524A US11974012B1 US 11974012 B1 US11974012 B1 US 11974012B1 US 202318501524 A US202318501524 A US 202318501524A US 11974012 B1 US11974012 B1 US 11974012B1
- Authority
- US
- United States
- Prior art keywords
- video content
- images
- streaming video
- sounds
- specific types
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 230000000007 visual effect Effects 0.000 claims abstract description 27
- 230000003993 interaction Effects 0.000 claims abstract description 23
- 230000004048 modification Effects 0.000 claims abstract description 21
- 238000012986 modification Methods 0.000 claims abstract description 21
- 238000000034 method Methods 0.000 claims description 59
- 230000004044 response Effects 0.000 claims description 46
- 241001465754 Metazoa Species 0.000 claims description 40
- 238000001514 detection method Methods 0.000 claims description 25
- 238000012806 monitoring device Methods 0.000 claims description 15
- 230000015572 biosynthetic process Effects 0.000 claims description 9
- 238000003786 synthesis reaction Methods 0.000 claims description 9
- 238000003909 pattern recognition Methods 0.000 claims description 6
- 230000003190 augmentative effect Effects 0.000 claims description 5
- 238000013135 deep learning Methods 0.000 claims description 5
- 238000010801 machine learning Methods 0.000 claims description 5
- 238000001914 filtration Methods 0.000 claims description 4
- 238000010183 spectrum analysis Methods 0.000 claims description 4
- 206010011469 Crying Diseases 0.000 claims description 3
- 238000004458 analytical method Methods 0.000 claims description 2
- 230000005236 sound signal Effects 0.000 description 13
- 241000282472 Canis lupus familiaris Species 0.000 description 9
- 230000035882 stress Effects 0.000 description 9
- 230000002708 enhancing effect Effects 0.000 description 7
- 238000006243 chemical reaction Methods 0.000 description 5
- 238000013527 convolutional neural network Methods 0.000 description 4
- 230000009429 distress Effects 0.000 description 3
- 230000008569 process Effects 0.000 description 3
- 238000012545 processing Methods 0.000 description 3
- 238000001228 spectrum Methods 0.000 description 3
- 230000001629 suppression Effects 0.000 description 3
- 241000239290 Araneae Species 0.000 description 2
- 230000008030 elimination Effects 0.000 description 2
- 238000003379 elimination reaction Methods 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 208000028173 post-traumatic stress disease Diseases 0.000 description 2
- 241000282412 Homo Species 0.000 description 1
- 206010034912 Phobia Diseases 0.000 description 1
- 238000013528 artificial neural network Methods 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000003090 exacerbative effect Effects 0.000 description 1
- 230000008921 facial expression Effects 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 230000000977 initiatory effect Effects 0.000 description 1
- 230000002452 interceptive effect Effects 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 208000019899 phobic disease Diseases 0.000 description 1
- 230000035755 proliferation Effects 0.000 description 1
- 238000011897 real-time detection Methods 0.000 description 1
- 230000004043 responsiveness Effects 0.000 description 1
- 230000003595 spectral effect Effects 0.000 description 1
- 238000012706 support-vector machine Methods 0.000 description 1
- 238000012549 training Methods 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/439—Processing of audio elementary streams
- H04N21/4394—Processing of audio elementary streams involving operations for analysing the audio stream, e.g. detecting features or characteristics in audio streams
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/41—Structure of client; Structure of client peripherals
- H04N21/422—Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
- H04N21/42203—Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS] sound input device, e.g. microphone
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/41—Structure of client; Structure of client peripherals
- H04N21/422—Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
- H04N21/4223—Cameras
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/81—Monomedia components thereof
- H04N21/8106—Monomedia components thereof involving special audio data, e.g. different tracks for different languages
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/81—Monomedia components thereof
- H04N21/8146—Monomedia components thereof involving graphical data, e.g. 3D object, 2D graphics
- H04N21/8153—Monomedia components thereof involving graphical data, e.g. 3D object, 2D graphics comprising still images, e.g. texture, background image
Definitions
- the disclosure is related to techniques for modifying audio and video content.
- This disclosure includes techniques for modifying specific types of sounds and images in streaming video content in response to a user input. For example, a user may request modification or elimination of sounds and images, such as animal sounds or images of violence, which may cause distress or disturbance to the user.
- a system for managing audio and visual content in streaming video content includes a user interaction module configured to receive a user input to manage specific types of sounds and images in the streaming video content, a recognition module configured to recognize the specific types of sounds and images in the streaming video content based on the user input, and a modification module configured to provide a managed version of the streaming video content, wherein the recognized specific types of sounds and images are managed in the managed version of the streaming video content.
- this disclosure is directed towards a non-transitory computer-readable medium storing instructions that, when executed by a processor, cause the processor to perform a method for managing audio and visual content in streaming video content, the method comprising: receiving a user input to manage specific types of sounds and images in the streaming video content, recognizing the specific types of sounds and images in the streaming video content based on the user input, and providing a managed version of the streaming video content, wherein the recognized specific types of sounds and images are managed in the managed version of the streaming video content.
- FIG. 1 is a conceptual illustration of a system for modifying specific types of sounds and images in streaming video content in response to a user input.
- FIG. 2 is a flowchart illustrating example techniques for modifying specific types of sounds and images in streaming video content in response to a user input.
- This disclosure includes techniques for modifying specific types of sounds and images in streaming video content in response to a user input. For example, a user may request modification or elimination of sounds and images, such as animal sounds or images of violence, which may cause distress or disturbance to the user.
- the disclosed techniques allow a user to customize the audio and visual content of the video based on their personal preferences and requirements. This customization is achieved by managing specific types of sounds and images in the video content, particularly those that some users may find disruptive, annoying, or distressing.
- FIG. 1 is a conceptual illustration of a system 100 for modifying specific types of sounds and images in streaming video content 101 in response to a user input.
- the system 100 can manage animal sounds in the video content. Animal sounds, such as the barking of dogs or the chirping of birds, can potentially excite pets, wake sleeping children, or cause stress to individuals who may find these sounds disturbing. By providing the option to manage these sounds, the system 100 allows users to enjoy the video content without being disturbed by such sounds.
- the system 100 can also manage images in the video content. Certain images, such as images of violence or guns, can potentially cause distress or disturbance to viewers of the content. By providing the option to manage these images, the system 100 allows users to enjoy the video content without being subjected to such images.
- the system 100 provides a personalized and responsive viewing experience. Whether it's managing animal sounds, war sounds, gunshots, images of animals, war scenes, violence, or guns, the system 100 provides users with the flexibility to customize their viewing experience, making the viewing of streaming video content 101 a more enjoyable and personalized experience.
- the user input module 110 is configured to receive a user input from user interface 112 to manage specific types of sounds and images in the streaming video content 101 .
- the user input can include a selection of specified sounds and images to be managed, and specific levels of management for the sounds and images. For instance, a user may choose to completely suppress the sounds of dogs barking, or to reduce the volume of bird chirping by a specific percentage. This flexibility allows the user to customize their viewing experience according to their personal preferences and requirements.
- the user input module 110 can receive user input through a user interface 112 implemented on a variety of devices. These devices can include, but are not limited to, computer devices, mobile devices, and smart televisions. For instance, a user may input their preferences through a touch screen interface on a mobile device or tablet, a voice recognition system on a smart speaker or virtual assistant, a graphical user interface (GUI) on a computer or smart TV, or a physical control panel on a device such as a mixer or soundboard.
- GUI graphical user interface
- the user interaction module of the system 100 is designed to receive user input through a user interface 112 .
- This user interface 112 serves as the primary point of interaction between the user and the system 100 , enabling the user to specify their preferences for the management of sounds and images in the streaming video content 101 .
- the user interface 112 can include various interactive elements, such as drop-down menus, sliders, checkboxes, or text input fields, that allow the user to specify their preferences. It can also provide visual feedback, such as a preview of the managed sounds or a visual representation of the selected images, to assist the user in making their selection.
- the user interface 112 can be implemented on a variety of devices, each offering different capabilities and user experiences. These devices can include, but are not limited to, computer devices, mobile devices, smart televisions, virtual reality devices, augmented reality devices, gaming consoles, wearable devices, home automation systems, and vehicle infotainment systems. For instance, a user may input their preferences through a touch screen interface on a mobile device or tablet, a voice recognition system on a smart speaker or virtual assistant, a graphical user interface (GUI) on a computer or smart TV, or a physical control panel on a device such as a mixer or soundboard.
- GUI graphical user interface
- the user interface 112 is designed to be compatible with a wide range of devices, providing users with the flexibility to input their preferences through the device that is the easiest and the convenient for them.
- the user interface 112 can provide users with the option to save their preferences, allowing them to apply the same settings to future streaming video content 101 without having to input their preferences each time.
- the identification module 120 is configured to recognize the specific types of sounds and images in the streaming video content 101 based on the user input.
- the identification module 120 uses various techniques to analyze the audio and visual content of the video and detect the presence of the specified sounds and images. These techniques can include spectral analysis, machine learning algorithms, pattern recognition techniques, audio fingerprinting, and deep learning techniques.
- Spectral analysis involves examining the frequency spectrum of the audio signal to identify specific types of sounds. Each animal sound has a distinct frequency spectrum, which can be compared with a library of frequency spectra of known animal sounds. This comparison allows the identification module 120 to accurately identify the presence of specific types of animal sounds in the streaming video content 101 .
- Machine learning algorithms can be trained on a dataset of animal sounds, enabling them to recognize similar sounds in the streaming video content 101 .
- These algorithms can include supervised learning algorithms, such as support vector machines or neural networks, which are capable of learning complex patterns in the audio data and accurately identifying animal sounds.
- Pattern recognition techniques involve comparing the audio content of the streaming video with a library of animal sounds to find matches. By comparing the pattern of the audio signal with a library of patterns of known animal sounds, the identification module 120 can accurately identify the presence of specific types of animal sounds in the streaming video content 101 .
- Audio fingerprinting is a technique that involves creating a condensed digital summary, or “fingerprint,” of an audio signal. This fingerprint can then be compared with a database of fingerprints of known animal sounds. If the fingerprint of the audio signal matches the fingerprint of a known animal sound in the database, the identification module 120 can identify the presence of that animal sound in the streaming video content 101 .
- Deep learning techniques such as convolutional neural networks (CNNs)
- CNNs are particularly effective at identifying patterns in audio data, making them well-suited for the task of identifying animal sounds.
- the identification module 120 can learn to recognize the distinct features of each animal sound and accurately identify these sounds in the streaming video content 101 .
- the identification module 120 can accurately recognize specific types of sounds and images in the streaming video content 101 based on the user input. This recognition serves as a foundation for the subsequent management process, enabling the system 100 to manage the desired sounds and images in the streaming video content 101 .
- the modification module 130 is specifically designed to provide a managed version of the streaming video content 101 , wherein the recognized specific types of sounds and images are managed as per the user input. Specifically, the modification module 130 alters the desired sounds and images in the streaming video content 101 .
- the modification module 130 may apply any number of techniques to modify audio signals. These techniques include volume attenuation, noise cancellation, audio synthesis, pitch shifting, and time stretching. These techniques can also include suppression, replacement, and distortion of the sounds and images.
- Volume attenuation is a process of reducing the intensity of an audio signal. It is often used in audio mixing and sound engineering to balance the levels of different audio sources. Volume attenuation can be used to decrease the loudness of specific sounds in the streaming video content, such as animal sounds or loud noises, based on user preferences.
- Time stretching is a digital audio processing technique that allows the duration of an audio signal to be changed without affecting its pitch. This is achieved by altering the speed or tempo of the audio signal without changing its spectral characteristics.
- Time stretching is one of the techniques or algorithms that can be used to modify identified sounds in streaming video content. For instance, if a specific animal sound or a loud noise is identified in the video content, time stretching can be used to lengthen or shorten the duration of this sound in the modified version of the content. This can help to reduce the disruptive impact of the sound on the viewer, enhancing their viewing experience.
- Noise Cancellation This is a method used to reduce unwanted sounds by creating an “anti-noise” sound wave that interferes with the unwanted noise, effectively cancelling it out. Active noise cancellation is commonly used in headphones to reduce background noise.
- Audio Synthesis This is the electronic production of sound. It can involve creating new sounds or altering existing ones. Synthesizers, which can generate a wide range of sounds, are commonly used for audio synthesis. Audio synthesis can be used to create alternative sounds to replace specific types of sounds in the streaming video content, such as animal sounds or loud noises.
- Pitch Shifting This is a sound recording technique where the original pitch of a sound is raised or lowered. Effects units that utilize pitch shifting can transpose an audio signal up or down in pitch. Pitch shifting can be used to alter the pitch of specific sounds in the streaming video content, changing the way they are perceived by the viewer.
- Suppression involves reducing the amplitude of the audio signal corresponding to the recognized sounds. This effectively lowers the volume of these sounds in the streaming video content 101 , making them less noticeable or completely inaudible, depending on the degree of suppression applied. This technique can be particularly useful for users who wish to suppress the sounds of specific animals without completely removing them from the audio content.
- Replacement involves generating a new audio signal or visual content to replace the recognized sounds or images.
- the sounds of dogs barking could be replaced with the sounds of birds chirping, or the images of violence could be replaced with more peaceful scenes.
- the new audio signals or visual content can be selected from a library of sounds and images, or they can be synthesized on the fly using sound synthesis algorithms or image generation techniques. This technique can be particularly useful for users who wish to replace the sounds or images of specific elements with more pleasant or less disturbing sounds or images.
- Distortion involves changing the characteristics of the audio signal or visual content corresponding to the recognized sounds or images. This alters the quality of these sounds or images, making them sound or look different from their original form. This can be particularly useful for users who find the original quality of these sounds or images disturbing or annoying. The degree of distortion can be adjusted based on the user's preferences, allowing them to customize the quality of the sounds or images to their liking.
- the system 100 can be hosted on various types of systems, including, but not limited to, a cloud-based system, a local device system, a distributed network system, a peer-to-peer network system, an edge computing system, and a hybrid cloud-edge system.
- a cloud-based system a local device system
- a distributed network system a peer-to-peer network system
- an edge computing system a hybrid cloud-edge system.
- the disclosed techniques may be embodied in a non-transitory computer-readable medium storing instructions that, when executed by one or more processors, cause the processors to perform a method for managing audio and visual content.
- the user interaction module, the recognition module, and the modification module 130 are hosted on a cloud computing platform.
- the user inputs are received and processed in the cloud, and the managed streaming video content 101 is streamed back to the user's device over the internet.
- This system allows for remote access and management of the streaming video content 101 , and can scale to handle large volumes of video content and user inputs. However, it may require a stable and high-speed internet connection, and there may be concerns about data privacy and security in the cloud.
- the user interaction module, the recognition module, and the modification module 130 are hosted on the user's device, such as a computer, mobile device, or smart TV.
- the user inputs are received and processed on the device, and the managed streaming video content 101 is streamed directly on the device.
- This system allows for real-time management of the streaming video content 101 , and can provide a more personalized user experience. However, it may require more computational resources on the device, and the system 100 's performance may be limited by the device's capabilities.
- the user interaction module, the recognition module, and the modification module 130 are hosted on a network of servers.
- the user inputs are received and processed on the server that is closest to the user, and the managed streaming video content 101 is streamed back to the user's device over the network.
- This system allows for handling of user inputs and streaming video content 101 , and can provide a high-quality streaming experience for the user.
- it may require a complex network infrastructure, and there may be concerns about data privacy and security in the network.
- the user interaction module, the recognition module, and the modification module 130 are distributed across a network of user devices.
- Each device contributes its resources to the network, processing user inputs and streaming video content 101 in a decentralized manner.
- This system can provide high resilience and scalability, as the failure or overload of one device does not affect the overall performance of the system 100 . However, it may require a robust network protocol, and the system 100 's performance may be affected by the capabilities and availability of the user devices in the network.
- the user interaction module, the recognition module, and the modification module 130 are hosted on edge devices located close to the user's location. These edge devices can be routers, gateways, or other network devices that provide low latency access to the user.
- the user inputs are processed at the edge of the network, allowing for real-time management of the streaming video content 101 .
- the managed content is then streamed directly to the user's device, reducing the latency and bandwidth usage.
- a hybrid cloud-edge system some components of the system 100 are hosted on a cloud platform while others are hosted on edge devices.
- the user interaction module and modification module 130 could be hosted on the cloud for global accessibility and scalability, while the recognition module could be hosted on edge devices for real-time processing.
- This system combines the benefits of cloud and edge computing, providing a balance between scalability, latency, and bandwidth usage.
- system 100 may also managing images of animals in video content. Some users may find images of specific animals to be disturbing or distracting. For instance, a user may have a phobia of spiders and may wish to avoid seeing images of spiders in video content. In such cases, the system 100 can analyze the video content to detect the presence of images of specific animals based on user input, and then provide a managed version of the content where these images are blurred, obscured, or replaced with other images. This feature can enhance the user experience, allowing users to enjoy the video content without being subjected to objectionable images.
- War scenes can be distressing for some users, causing fear, exacerbating post-traumatic stress disorder (PTSD), or simply being too violent or graphic for some viewers.
- the system 100 can analyze the video content to detect the presence of war scenes based on user input, and then provide a managed version of the content where these scenes are blurred, obscured, or replaced with other scenes. This feature can enhance the user experience, allowing users to enjoy the video content without being subjected to distressing scenes.
- PTSD post-traumatic stress disorder
- a further variation of these techniques includes managing images of violence or guns in video content.
- Images of violence or guns can be disturbing for some users, particularly those who are sensitive to such content or those who are watching the content with young children.
- the system 100 can analyze the video content to detect the presence of images of violence or guns based on user input, and then provide a managed version of the content where these images are blurred, obscured, or replaced with other images. This feature can enhance the user experience, allowing users to enjoy the video content without being subjected to disturbing images.
- the system 100 may employs any of a variety of techniques to analyze the video content and detect the presence of specific types of images. These techniques can include image recognition techniques, machine learning algorithms, pattern recognition techniques, and deep learning techniques. Once the specific types of images have been detected, the modification module 130 modifies these images using various techniques, such as image blurring, image obscuring, or image replacement. The modified images are then combined with the original audio content or managed audio content of the video to provide a managed version of the video content that caters to the user's preferences and requirements.
- system 100 may further include a monitoring device 114 to observe reactions to the video content.
- Monitoring device 114 may include a camera or a microphone, for example.
- the monitoring device 114 is strategically configured to observe the users and the surrounding noise, thereby providing a comprehensive understanding of the user's environment and reactions to the streaming video content 101 .
- a camera When a camera is employed as the monitoring device 114 , it captures visual data that can be analyzed to detect user 115 responses to the streaming video content 101 . For instance, the camera may capture the physical reactions of a pet in response to animal sounds in the video content. Similarly, the camera can also capture visual cues of human stress, such as facial expressions or body language, in response to loud noises in the video content.
- the visual data captured by the camera provides a rich source of information that can be analyzed by the detection module 150 to identify specific responses to the audio and/or video elements in the streaming video content 101 .
- a microphone when a microphone is used as the monitoring device 114 , it records the surrounding noise and audio responses from the user or pets. For example, the microphone can detect a dog barking in response to animal sounds in the video content or a baby crying in response to loud noises. The audio data captured by the microphone provides a direct measure of the user's or pet's reactions to the streaming video content 101 , which can be analyzed by the detection module 150 to identify specific responses.
- the monitoring device 114 is configured to continuously capture data while the streaming video content 101 is being consumed. This continuous monitoring allows for real-time detection of ambient response, thereby enabling the system 100 to promptly react to any potentially disruptive elements in the video content. The data collected by the monitoring device 114 is then relayed to the detection module 150 for further analysis.
- the detection module 150 is designed to analyze the data received from the monitoring device 114 and detect specific responses to the audio and/or video elements in the streaming video content 101 .
- the detection module 150 is configured to identify a range of responses, from signs of human stress to indications of excited animals.
- the detection module 150 is configured to recognize specific sound patterns or visual cues associated with various responses. For instance, the detection module 150 may use sound pattern recognition techniques to identify a dog's bark or a baby's cry. Similarly, it may use visual cue recognition techniques to identify signs of human stress or excitement in response to the video content. These techniques ensure that the detection module 150 accurately identifies responses, thereby enabling the system 100 to react accordingly.
- the detection module 150 operates in real-time, analyzing the data received from the monitoring device 114 as it is captured. This real-time operation allows the system 100 to promptly detect ambient responses and react accordingly, thereby enhancing the user's viewing experience. Once a response is detected, the detection module 150 communicates with the query module 160 , initiating the next step in the system 100 's process of managing content in streaming video.
- the interaction between the detection module 150 and the other components of the system 100 is facilitated through a communication protocol.
- This protocol ensures that the detection of a response by the detection module 150 is communicated to the query module 160 .
- the seamless interaction between these components is integral to the system 100 's ability to effectively manage content in streaming video, thereby enhancing the user's viewing experience.
- the query module 160 is activated when the detection module 150 identifies a response to video content presented on display 102 .
- the primary function of the query module 160 is to interact with the user upon the detection of a response, providing the user with the option to manage the content in the streaming video content 101 .
- the query module 160 is configured to ask the user whether they would like to manage the content in the streaming video content 101 . This interaction is designed to provide the user with control over the content they consume, allowing them to filter out potentially disruptive elements. For instance, if the system 100 detects a dog barking in response to animal sounds in the video content, the query module 160 can interact with the user, asking them if they would like to manage the content. This could involve filtering out the animal sounds that caused the dog's response.
- the query module 160 can interact with the user, asking them if they would like to manage the content. This could involve filtering out the loud noises that caused the baby's response. This interaction is designed to provide the user with control over the content they consume, allowing them to filter out potentially disruptive elements.
- the interaction between the query module 160 and the user can take various forms, depending on the user's preferences and the nature of the detected ambient response.
- the query module 160 could present the user with a pop-up message on their screen, asking them if they would like to manage the content.
- the query module 160 could send the user a notification on their mobile device, asking them the same question.
- the specific form of interaction can be customized based on the user's preferences, thereby enhancing the user's control over the content they consume.
- the query module 160 interacts with the user upon the detection of a response. By asking the user whether they would like to manage the content in the streaming video content 101 , the query module 160 provides the user with control over the content they consume, allowing them to filter out potentially disruptive elements. This interaction enhances the user's viewing experience, making the viewing of streaming video content 101 a more enjoyable and personalized experience.
- the system 100 is designed to manage a variety of audio and visual elements within the streaming video content 101 . This includes, but is not limited to, animal sounds and loud noises such as gunfire, fights, or arguments.
- the system's ability to manage these types of content is integral to its purpose of enhancing the user's viewing experience.
- Animal sounds in the video content can cause excitement or stress in pets.
- the system 100 is designed to identify these sounds and manage them based on the user's preferences and the detected ambient responses. For example, if a dog barks in response to animal sounds in the video content, the system 100 can filter out these sounds, thereby preventing further excitement or stress in the pet.
- the detection module is configured to recognizing specific sound patterns associated with various animal sounds observed by monitoring device 114 .
- Loud noises in the video content can also cause stress in humans, particularly in young children.
- the system 100 is designed to identify these noises and manage them based on the user's preferences and the detected ambient responses. For instance, if a baby cries in response to loud noises in the video content, the system 100 can filter out these noises, thereby preventing further stress in the baby.
- the detection module 150 is configured distinguishing these noises from other sounds in the video content.
- the system 100 may be designed to operate in real-time, allowing for prompt reaction to potentially disruptive elements in the video content. This real-time operation is facilitated by the seamless interaction between the system 100 's components, particularly the monitoring device 114 , the detection module 150 , and the query module 160 .
- the monitoring device 114 captures the audio and visual data
- the detection module 150 analyzes this data and detects specific responses
- the query module 160 interacts with the user, providing them with the option to manage the content.
- This cohesive operation of the system 100 's components enables effective management of content in streaming video, thereby enhancing the user's viewing experience.
- the system 100 could be configured to automatically manage the content based on the detected ambient responses, without querying the user. For example, if the system 100 detects a dog barking in response to animal sounds in the video content, it could automatically filter out these sounds, thereby preventing further excitement or stress in the pet. This automatic content management could enhance the system 100 's responsiveness to potentially disruptive elements in the video content, thereby enhancing the user's viewing experience.
- FIG. 2 is a flowchart 200 illustrating example techniques for modifying specific types of sounds and images in streaming video content in response to a user input. For clarity, the techniques of FIG. 2 are described with respect to system 100 ( FIG. 1 ).
- a user device such as display 102 issues a request to access video content, such as streaming video content 101 (step 202 ).
- identification module 120 accesses the video content (step 204 ).
- identification module 120 recognizes the specific types of sounds and images in the streaming video content based on the user input. (step 206 ).
- Modification module 130 provides a managed version of the streaming video content to the user device (step 208 ).
- the managed version of the streaming video content includes altered content such that specific types of sounds and images are managed in the managed version of the streaming video content.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Computer Graphics (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
Abstract
A system for managing audio and visual content in streaming video content includes a user interaction module configured to receive a user input to manage specific types of sounds and images in the streaming video content, a recognition module configured to recognize the specific types of sounds and images in the streaming video content based on the user input, and a modification module configured to provide a managed version of the streaming video content, wherein the recognized specific types of sounds and images are managed in the managed version of the streaming video content.
Description
The disclosure is related to techniques for modifying audio and video content.
The advent of digital technology has revolutionized the way we consume media. In particular, the proliferation of streaming video content has provided users with unprecedented access to a vast array of multimedia content. This content ranges from movies and television shows to live broadcasts and user-generated videos. As the technology has evolved, so too have the tools and techniques for manipulating and customizing this content to suit individual user preferences.
This disclosure includes techniques for modifying specific types of sounds and images in streaming video content in response to a user input. For example, a user may request modification or elimination of sounds and images, such as animal sounds or images of violence, which may cause distress or disturbance to the user.
In one example, a system for managing audio and visual content in streaming video content includes a user interaction module configured to receive a user input to manage specific types of sounds and images in the streaming video content, a recognition module configured to recognize the specific types of sounds and images in the streaming video content based on the user input, and a modification module configured to provide a managed version of the streaming video content, wherein the recognized specific types of sounds and images are managed in the managed version of the streaming video content.
In another example, this disclosure is directed towards a non-transitory computer-readable medium storing instructions that, when executed by a processor, cause the processor to perform a method for managing audio and visual content in streaming video content, the method comprising: receiving a user input to manage specific types of sounds and images in the streaming video content, recognizing the specific types of sounds and images in the streaming video content based on the user input, and providing a managed version of the streaming video content, wherein the recognized specific types of sounds and images are managed in the managed version of the streaming video content.
This disclosure includes techniques for modifying specific types of sounds and images in streaming video content in response to a user input. For example, a user may request modification or elimination of sounds and images, such as animal sounds or images of violence, which may cause distress or disturbance to the user.
The disclosed techniques allow a user to customize the audio and visual content of the video based on their personal preferences and requirements. This customization is achieved by managing specific types of sounds and images in the video content, particularly those that some users may find disruptive, annoying, or distressing.
Similarly, the system 100 can also manage images in the video content. Certain images, such as images of violence or guns, can potentially cause distress or disturbance to viewers of the content. By providing the option to manage these images, the system 100 allows users to enjoy the video content without being subjected to such images.
The system 100 provides a personalized and responsive viewing experience. Whether it's managing animal sounds, war sounds, gunshots, images of animals, war scenes, violence, or guns, the system 100 provides users with the flexibility to customize their viewing experience, making the viewing of streaming video content 101 a more enjoyable and personalized experience.
User Input Module 110
The user input module 110 is configured to receive a user input from user interface 112 to manage specific types of sounds and images in the streaming video content 101. The user input can include a selection of specified sounds and images to be managed, and specific levels of management for the sounds and images. For instance, a user may choose to completely suppress the sounds of dogs barking, or to reduce the volume of bird chirping by a specific percentage. This flexibility allows the user to customize their viewing experience according to their personal preferences and requirements.
The user input module 110 can receive user input through a user interface 112 implemented on a variety of devices. These devices can include, but are not limited to, computer devices, mobile devices, and smart televisions. For instance, a user may input their preferences through a touch screen interface on a mobile device or tablet, a voice recognition system on a smart speaker or virtual assistant, a graphical user interface (GUI) on a computer or smart TV, or a physical control panel on a device such as a mixer or soundboard.
The user interaction module of the system 100 is designed to receive user input through a user interface 112. This user interface 112 serves as the primary point of interaction between the user and the system 100, enabling the user to specify their preferences for the management of sounds and images in the streaming video content 101. The user interface 112 can include various interactive elements, such as drop-down menus, sliders, checkboxes, or text input fields, that allow the user to specify their preferences. It can also provide visual feedback, such as a preview of the managed sounds or a visual representation of the selected images, to assist the user in making their selection.
The user interface 112 can be implemented on a variety of devices, each offering different capabilities and user experiences. These devices can include, but are not limited to, computer devices, mobile devices, smart televisions, virtual reality devices, augmented reality devices, gaming consoles, wearable devices, home automation systems, and vehicle infotainment systems. For instance, a user may input their preferences through a touch screen interface on a mobile device or tablet, a voice recognition system on a smart speaker or virtual assistant, a graphical user interface (GUI) on a computer or smart TV, or a physical control panel on a device such as a mixer or soundboard. The user interface 112 is designed to be compatible with a wide range of devices, providing users with the flexibility to input their preferences through the device that is the easiest and the convenient for them.
Furthermore, the user interface 112 can provide users with the option to save their preferences, allowing them to apply the same settings to future streaming video content 101 without having to input their preferences each time.
The identification module 120 is configured to recognize the specific types of sounds and images in the streaming video content 101 based on the user input. The identification module 120 uses various techniques to analyze the audio and visual content of the video and detect the presence of the specified sounds and images. These techniques can include spectral analysis, machine learning algorithms, pattern recognition techniques, audio fingerprinting, and deep learning techniques.
Spectral analysis involves examining the frequency spectrum of the audio signal to identify specific types of sounds. Each animal sound has a distinct frequency spectrum, which can be compared with a library of frequency spectra of known animal sounds. This comparison allows the identification module 120 to accurately identify the presence of specific types of animal sounds in the streaming video content 101.
Machine learning algorithms can be trained on a dataset of animal sounds, enabling them to recognize similar sounds in the streaming video content 101. These algorithms can include supervised learning algorithms, such as support vector machines or neural networks, which are capable of learning complex patterns in the audio data and accurately identifying animal sounds.
Pattern recognition techniques involve comparing the audio content of the streaming video with a library of animal sounds to find matches. By comparing the pattern of the audio signal with a library of patterns of known animal sounds, the identification module 120 can accurately identify the presence of specific types of animal sounds in the streaming video content 101.
Audio fingerprinting is a technique that involves creating a condensed digital summary, or “fingerprint,” of an audio signal. This fingerprint can then be compared with a database of fingerprints of known animal sounds. If the fingerprint of the audio signal matches the fingerprint of a known animal sound in the database, the identification module 120 can identify the presence of that animal sound in the streaming video content 101.
Deep learning techniques, such as convolutional neural networks (CNNs), can be used to identify animal sounds in streaming video content 101. CNNs are particularly effective at identifying patterns in audio data, making them well-suited for the task of identifying animal sounds. By training a CNN on a dataset of animal sounds, the identification module 120 can learn to recognize the distinct features of each animal sound and accurately identify these sounds in the streaming video content 101.
Through the use of these techniques, the identification module 120 can accurately recognize specific types of sounds and images in the streaming video content 101 based on the user input. This recognition serves as a foundation for the subsequent management process, enabling the system 100 to manage the desired sounds and images in the streaming video content 101.
The modification module 130 is specifically designed to provide a managed version of the streaming video content 101, wherein the recognized specific types of sounds and images are managed as per the user input. Specifically, the modification module 130 alters the desired sounds and images in the streaming video content 101. The modification module 130 may apply any number of techniques to modify audio signals. These techniques include volume attenuation, noise cancellation, audio synthesis, pitch shifting, and time stretching. These techniques can also include suppression, replacement, and distortion of the sounds and images.
Volume attenuation is a process of reducing the intensity of an audio signal. It is often used in audio mixing and sound engineering to balance the levels of different audio sources. Volume attenuation can be used to decrease the loudness of specific sounds in the streaming video content, such as animal sounds or loud noises, based on user preferences.
Time stretching is a digital audio processing technique that allows the duration of an audio signal to be changed without affecting its pitch. This is achieved by altering the speed or tempo of the audio signal without changing its spectral characteristics. Time stretching is one of the techniques or algorithms that can be used to modify identified sounds in streaming video content. For instance, if a specific animal sound or a loud noise is identified in the video content, time stretching can be used to lengthen or shorten the duration of this sound in the modified version of the content. This can help to reduce the disruptive impact of the sound on the viewer, enhancing their viewing experience.
Noise Cancellation: This is a method used to reduce unwanted sounds by creating an “anti-noise” sound wave that interferes with the unwanted noise, effectively cancelling it out. Active noise cancellation is commonly used in headphones to reduce background noise.
Audio Synthesis: This is the electronic production of sound. It can involve creating new sounds or altering existing ones. Synthesizers, which can generate a wide range of sounds, are commonly used for audio synthesis. Audio synthesis can be used to create alternative sounds to replace specific types of sounds in the streaming video content, such as animal sounds or loud noises.
Pitch Shifting: This is a sound recording technique where the original pitch of a sound is raised or lowered. Effects units that utilize pitch shifting can transpose an audio signal up or down in pitch. Pitch shifting can be used to alter the pitch of specific sounds in the streaming video content, changing the way they are perceived by the viewer.
Suppression involves reducing the amplitude of the audio signal corresponding to the recognized sounds. This effectively lowers the volume of these sounds in the streaming video content 101, making them less noticeable or completely inaudible, depending on the degree of suppression applied. This technique can be particularly useful for users who wish to suppress the sounds of specific animals without completely removing them from the audio content.
Replacement involves generating a new audio signal or visual content to replace the recognized sounds or images. For instance, the sounds of dogs barking could be replaced with the sounds of birds chirping, or the images of violence could be replaced with more peaceful scenes. The new audio signals or visual content can be selected from a library of sounds and images, or they can be synthesized on the fly using sound synthesis algorithms or image generation techniques. This technique can be particularly useful for users who wish to replace the sounds or images of specific elements with more pleasant or less disturbing sounds or images.
Distortion involves changing the characteristics of the audio signal or visual content corresponding to the recognized sounds or images. This alters the quality of these sounds or images, making them sound or look different from their original form. This can be particularly useful for users who find the original quality of these sounds or images disturbing or annoying. The degree of distortion can be adjusted based on the user's preferences, allowing them to customize the quality of the sounds or images to their liking.
The system 100 can be hosted on various types of systems, including, but not limited to, a cloud-based system, a local device system, a distributed network system, a peer-to-peer network system, an edge computing system, and a hybrid cloud-edge system. With any of these systems, the disclosed techniques may be embodied in a non-transitory computer-readable medium storing instructions that, when executed by one or more processors, cause the processors to perform a method for managing audio and visual content.
Cloud-Based System
In a cloud-based system, the user interaction module, the recognition module, and the modification module 130 are hosted on a cloud computing platform. The user inputs are received and processed in the cloud, and the managed streaming video content 101 is streamed back to the user's device over the internet. This system allows for remote access and management of the streaming video content 101, and can scale to handle large volumes of video content and user inputs. However, it may require a stable and high-speed internet connection, and there may be concerns about data privacy and security in the cloud.
Local Device System
In a local device system, the user interaction module, the recognition module, and the modification module 130 are hosted on the user's device, such as a computer, mobile device, or smart TV. The user inputs are received and processed on the device, and the managed streaming video content 101 is streamed directly on the device. This system allows for real-time management of the streaming video content 101, and can provide a more personalized user experience. However, it may require more computational resources on the device, and the system 100's performance may be limited by the device's capabilities.
Distributed Network System
In a distributed network system, the user interaction module, the recognition module, and the modification module 130 are hosted on a network of servers. The user inputs are received and processed on the server that is closest to the user, and the managed streaming video content 101 is streamed back to the user's device over the network. This system allows for handling of user inputs and streaming video content 101, and can provide a high-quality streaming experience for the user. However, it may require a complex network infrastructure, and there may be concerns about data privacy and security in the network.
Peer-to-Peer Network System
In a peer-to-peer network system, the user interaction module, the recognition module, and the modification module 130 are distributed across a network of user devices. Each device contributes its resources to the network, processing user inputs and streaming video content 101 in a decentralized manner. This system can provide high resilience and scalability, as the failure or overload of one device does not affect the overall performance of the system 100. However, it may require a robust network protocol, and the system 100's performance may be affected by the capabilities and availability of the user devices in the network.
Edge Computing System
In an edge computing system, the user interaction module, the recognition module, and the modification module 130 are hosted on edge devices located close to the user's location. These edge devices can be routers, gateways, or other network devices that provide low latency access to the user. The user inputs are processed at the edge of the network, allowing for real-time management of the streaming video content 101. The managed content is then streamed directly to the user's device, reducing the latency and bandwidth usage.
Hybrid Cloud-Edge System
In a hybrid cloud-edge system, some components of the system 100 are hosted on a cloud platform while others are hosted on edge devices. For instance, the user interaction module and modification module 130 could be hosted on the cloud for global accessibility and scalability, while the recognition module could be hosted on edge devices for real-time processing. This system combines the benefits of cloud and edge computing, providing a balance between scalability, latency, and bandwidth usage.
Image Filtering
In addition to the techniques for managing specific types of sounds in streaming video content 101, system 100 may also managing images of animals in video content. Some users may find images of specific animals to be disturbing or distracting. For instance, a user may have a phobia of spiders and may wish to avoid seeing images of spiders in video content. In such cases, the system 100 can analyze the video content to detect the presence of images of specific animals based on user input, and then provide a managed version of the content where these images are blurred, obscured, or replaced with other images. This feature can enhance the user experience, allowing users to enjoy the video content without being subjected to objectionable images.
Another variation of these techniques includes managing war scenes in video content. War scenes can be distressing for some users, causing fear, exacerbating post-traumatic stress disorder (PTSD), or simply being too violent or graphic for some viewers. In such cases, the system 100 can analyze the video content to detect the presence of war scenes based on user input, and then provide a managed version of the content where these scenes are blurred, obscured, or replaced with other scenes. This feature can enhance the user experience, allowing users to enjoy the video content without being subjected to distressing scenes.
A further variation of these techniques includes managing images of violence or guns in video content. Images of violence or guns can be disturbing for some users, particularly those who are sensitive to such content or those who are watching the content with young children. In such cases, the system 100 can analyze the video content to detect the presence of images of violence or guns based on user input, and then provide a managed version of the content where these images are blurred, obscured, or replaced with other images. This feature can enhance the user experience, allowing users to enjoy the video content without being subjected to disturbing images.
In all these variations, the system 100 may employs any of a variety of techniques to analyze the video content and detect the presence of specific types of images. These techniques can include image recognition techniques, machine learning algorithms, pattern recognition techniques, and deep learning techniques. Once the specific types of images have been detected, the modification module 130 modifies these images using various techniques, such as image blurring, image obscuring, or image replacement. The modified images are then combined with the original audio content or managed audio content of the video to provide a managed version of the video content that caters to the user's preferences and requirements.
In a variation of the techniques, system 100 may further include a monitoring device 114 to observe reactions to the video content. Monitoring device 114 may include a camera or a microphone, for example. The monitoring device 114 is strategically configured to observe the users and the surrounding noise, thereby providing a comprehensive understanding of the user's environment and reactions to the streaming video content 101.
When a camera is employed as the monitoring device 114, it captures visual data that can be analyzed to detect user 115 responses to the streaming video content 101. For instance, the camera may capture the physical reactions of a pet in response to animal sounds in the video content. Similarly, the camera can also capture visual cues of human stress, such as facial expressions or body language, in response to loud noises in the video content. The visual data captured by the camera provides a rich source of information that can be analyzed by the detection module 150 to identify specific responses to the audio and/or video elements in the streaming video content 101.
On the other hand, when a microphone is used as the monitoring device 114, it records the surrounding noise and audio responses from the user or pets. For example, the microphone can detect a dog barking in response to animal sounds in the video content or a baby crying in response to loud noises. The audio data captured by the microphone provides a direct measure of the user's or pet's reactions to the streaming video content 101, which can be analyzed by the detection module 150 to identify specific responses.
The monitoring device 114 is configured to continuously capture data while the streaming video content 101 is being consumed. This continuous monitoring allows for real-time detection of ambient response, thereby enabling the system 100 to promptly react to any potentially disruptive elements in the video content. The data collected by the monitoring device 114 is then relayed to the detection module 150 for further analysis.
The detection module 150 is designed to analyze the data received from the monitoring device 114 and detect specific responses to the audio and/or video elements in the streaming video content 101. The detection module 150 is configured to identify a range of responses, from signs of human stress to indications of excited animals.
The detection module 150 is configured to recognize specific sound patterns or visual cues associated with various responses. For instance, the detection module 150 may use sound pattern recognition techniques to identify a dog's bark or a baby's cry. Similarly, it may use visual cue recognition techniques to identify signs of human stress or excitement in response to the video content. These techniques ensure that the detection module 150 accurately identifies responses, thereby enabling the system 100 to react accordingly.
The detection module 150 operates in real-time, analyzing the data received from the monitoring device 114 as it is captured. This real-time operation allows the system 100 to promptly detect ambient responses and react accordingly, thereby enhancing the user's viewing experience. Once a response is detected, the detection module 150 communicates with the query module 160, initiating the next step in the system 100's process of managing content in streaming video.
The interaction between the detection module 150 and the other components of the system 100 is facilitated through a communication protocol. This protocol ensures that the detection of a response by the detection module 150 is communicated to the query module 160. The seamless interaction between these components is integral to the system 100's ability to effectively manage content in streaming video, thereby enhancing the user's viewing experience.
The query module 160 is activated when the detection module 150 identifies a response to video content presented on display 102. The primary function of the query module 160 is to interact with the user upon the detection of a response, providing the user with the option to manage the content in the streaming video content 101.
The query module 160 is configured to ask the user whether they would like to manage the content in the streaming video content 101. This interaction is designed to provide the user with control over the content they consume, allowing them to filter out potentially disruptive elements. For instance, if the system 100 detects a dog barking in response to animal sounds in the video content, the query module 160 can interact with the user, asking them if they would like to manage the content. This could involve filtering out the animal sounds that caused the dog's response.
Similarly, if the system 100 detects a baby crying in response to loud noises in the video content, the query module 160 can interact with the user, asking them if they would like to manage the content. This could involve filtering out the loud noises that caused the baby's response. This interaction is designed to provide the user with control over the content they consume, allowing them to filter out potentially disruptive elements.
The interaction between the query module 160 and the user can take various forms, depending on the user's preferences and the nature of the detected ambient response. For example, the query module 160 could present the user with a pop-up message on their screen, asking them if they would like to manage the content. Alternatively, the query module 160 could send the user a notification on their mobile device, asking them the same question. The specific form of interaction can be customized based on the user's preferences, thereby enhancing the user's control over the content they consume.
In summary, the query module 160 interacts with the user upon the detection of a response. By asking the user whether they would like to manage the content in the streaming video content 101, the query module 160 provides the user with control over the content they consume, allowing them to filter out potentially disruptive elements. This interaction enhances the user's viewing experience, making the viewing of streaming video content 101 a more enjoyable and personalized experience.
The system 100 is designed to manage a variety of audio and visual elements within the streaming video content 101. This includes, but is not limited to, animal sounds and loud noises such as gunfire, fights, or arguments. The system's ability to manage these types of content is integral to its purpose of enhancing the user's viewing experience.
Animal sounds in the video content, for instance, can cause excitement or stress in pets. The system 100 is designed to identify these sounds and manage them based on the user's preferences and the detected ambient responses. For example, if a dog barks in response to animal sounds in the video content, the system 100 can filter out these sounds, thereby preventing further excitement or stress in the pet. The detection module is configured to recognizing specific sound patterns associated with various animal sounds observed by monitoring device 114.
Loud noises in the video content, such as gunfire, fights, or arguments, can also cause stress in humans, particularly in young children. The system 100 is designed to identify these noises and manage them based on the user's preferences and the detected ambient responses. For instance, if a baby cries in response to loud noises in the video content, the system 100 can filter out these noises, thereby preventing further stress in the baby. The detection module 150 is configured distinguishing these noises from other sounds in the video content.
The system 100 may be designed to operate in real-time, allowing for prompt reaction to potentially disruptive elements in the video content. This real-time operation is facilitated by the seamless interaction between the system 100's components, particularly the monitoring device 114, the detection module 150, and the query module 160. The monitoring device 114 captures the audio and visual data, the detection module 150 analyzes this data and detects specific responses, and the query module 160 interacts with the user, providing them with the option to manage the content. This cohesive operation of the system 100's components enables effective management of content in streaming video, thereby enhancing the user's viewing experience.
In the same or different examples, the system 100 could be configured to automatically manage the content based on the detected ambient responses, without querying the user. For example, if the system 100 detects a dog barking in response to animal sounds in the video content, it could automatically filter out these sounds, thereby preventing further excitement or stress in the pet. This automatic content management could enhance the system 100's responsiveness to potentially disruptive elements in the video content, thereby enhancing the user's viewing experience.
First, a user device, such as display 102 issues a request to access video content, such as streaming video content 101 (step 202). In response to the request, identification module 120 accesses the video content (step 204). Based on a user input from user input module 110, identification module 120 recognizes the specific types of sounds and images in the streaming video content based on the user input. (step 206). Modification module 130, provides a managed version of the streaming video content to the user device (step 208). The managed version of the streaming video content includes altered content such that specific types of sounds and images are managed in the managed version of the streaming video content.
The specific techniques for modifying specific types of sounds and images in streaming video content in response to a user input, such as techniques embodied by system 100 are merely illustrative of the general inventive concepts included in this disclosure as defined by the following claims.
Claims (59)
1. A system for managing audio and visual content in streaming video content, the system comprising:
a user interaction module configured to receive a user input to manage specific types of sounds and images in the streaming video content;
a recognition module configured to recognize the specific types of sounds and images in the streaming video content based on the user input; and
a modification module configured to provide a managed version of the streaming video content, wherein the recognized specific types of sounds and images are managed in the managed version of the streaming video content,
wherein the management of the recognized sounds and images is achieved through techniques selected from a group consisting of:
volume attenuation;
noise cancellation;
audio synthesis;
pitch shifting; and
time stretching.
2. The system of claim 1 , wherein the user input includes a selection of specified sounds and images to be managed in the streaming video content.
3. The system of claim 1 , wherein the user input includes a selection of specific levels of management for the sounds and images.
4. The system of claim 1 , further comprising an analysis module configured to analyze the streaming video content to detect the specific types of sounds and images.
5. The system of claim 1 , further comprising a filtering module configured to filter other elements of a video signal based on a user selection.
6. The system of claim 1 , wherein the user interaction module, the recognition module, and the modification module are hosted on a cloud computing system, allowing for remote access and management of the streaming video content.
7. The system of claim 1 , wherein the management of the recognized sounds and images includes the volume attenuation.
8. The system of claim 1 , wherein the management of the recognized sounds and images includes the noise cancellation.
9. The system of claim 1 , wherein the management of the recognized sounds and images includes the audio synthesis.
10. The system of claim 1 , wherein the management of the recognized sounds and images includes the pitch shifting.
11. The system of claim 1 , wherein the management of the recognized sounds and images includes the time stretching.
12. A system for managing audio and visual content in streaming video content, the system comprising:
a user interaction module configured to receive a user input to manage specific types of sounds and images in the streaming video content;
a recognition module configured to recognize the specific types of sounds and images in the streaming video content based on the user input; and
a modification module configured to provide a managed version of the streaming video content, wherein the recognized specific types of sounds and images are managed in the managed version of the streaming video content,
wherein the user input is received through a user interface on a device selected from a group consisting of:
a computer device;
a mobile device;
a smart television;
a virtual reality device;
an augmented reality device;
a gaming console;
a wearable device;
a home automation system; and
a vehicle infotainment system.
13. The system of claim 12 , wherein the device is the computer device.
14. The system of claim 12 , wherein the device is the mobile device.
15. The system of claim 12 , wherein the device is the smart television.
16. The system of claim 12 , wherein the device is the virtual reality device.
17. The system of claim 12 , wherein the device is the augmented reality device.
18. The system of claim 12 , wherein the device is the gaming console.
19. The system of claim 12 , wherein the device is the wearable device.
20. The system of claim 12 , wherein the device is the home automation system.
21. The system of claim 12 , wherein the device is the vehicle infotainment system.
22. The system of claim 12 , wherein providing the managed version of the streaming video content includes blurring the specific types of images.
23. The system of claim 12 , wherein providing the managed version of the streaming video content includes obscuring the specific types of images.
24. The system of claim 12 , wherein providing the managed version of the streaming video content includes replacement of the specific types of images with new visual content.
25. A system for managing audio and visual content in streaming video content, the system comprising:
a user interaction module configured to receive a user input to manage specific types of sounds and images in the streaming video content;
a recognition module configured to recognize the specific types of sounds and images in the streaming video content based on the user input; and
a modification module configured to provide a managed version of the streaming video content, wherein the recognized specific types of sounds and images are managed in the managed version of the streaming video content,
wherein the recognition of sounds and images in the streaming video content is achieved through techniques selected from a group consisting of spectral analysis, machine learning algorithms, pattern recognition techniques, audio fingerprinting, and deep learning techniques.
26. The system of claim 25 , wherein the recognition of sounds and images in the streaming video content includes the spectral analysis.
27. The system of claim 25 , wherein the recognition of sounds and images in the streaming video content includes the machine learning algorithms.
28. The system of claim 25 , wherein the recognition of sounds and images in the streaming video content includes the pattern recognition techniques.
29. The system of claim 25 , wherein the recognition of sounds and images in the streaming video content includes the audio fingerprinting.
30. The system of claim 25 , wherein the recognition of sounds and images in the streaming video content includes the deep learning techniques.
31. The system of claim 25 , wherein providing the managed version of the streaming video content includes blurring the specific types of images.
32. The system of claim 25 , wherein providing the managed version of the streaming video content includes obscuring the specific types of images.
33. The system of claim 25 , wherein providing the managed version of the streaming video content includes replacement of the specific types of images with new visual content.
34. A system for managing audio and visual content in streaming video content, the system comprising:
a monitoring device configured to monitor users and noise;
a detection module configured to detect ambient responses to the streaming video content;
a query module configured to query the user whether they would like to manage content in the streaming video content in response to the detection of ambient response;
a user interaction module configured to receive a user input to manage specific types of sounds and images in the streaming video content;
a recognition module configured to recognize the specific types of sounds and images in the streaming video content based on the user input; and
a modification module configured to provide a managed version of the streaming video content, wherein the recognized specific types of sounds and images are managed in the managed version of the streaming video content.
35. The system of claim 34 , wherein the monitoring device is selected from a group consisting of a camera and a microphone.
36. The system of claim 34 , wherein the responses are selected from a group consisting of human stress and excited animals.
37. The system of claim 34 , wherein the detection module is further configured to detect a dog barking in response to animal sounds in the streaming video content.
38. The system of claim 34 , wherein the detection module is further configured to detect a baby crying in response to loud noises in the streaming video content.
39. A non-transitory computer-readable medium storing instructions that, when executed by a one or more processors, cause the processors to perform a method for managing audio and visual content in streaming video content, the method comprising:
receiving a user input to manage specific types of sounds and images in the streaming video content;
recognizing the specific types of sounds and images in the streaming video content based on the user input; and
providing a managed version of the streaming video content, wherein the recognized specific types of sounds and images are managed in the managed version of the streaming video content,
wherein the management of the recognized sounds and images is achieved through techniques selected from a group consisting of:
volume attenuation;
noise cancellation;
audio synthesis;
pitch shifting; and
time stretching.
40. The non-transitory computer-readable medium of claim 39 , wherein the user input includes a selection of specified sounds and images to be managed in the streaming video content.
41. The non-transitory computer-readable medium of claim 39 , wherein the user input includes a selection of specific levels of management for the sounds and images.
42. The non-transitory computer-readable medium of claim 39 , wherein the management of the recognized sounds and images includes the volume attenuation.
43. The non-transitory computer-readable medium of claim 39 , wherein the management of the recognized sounds and images includes the noise cancellation.
44. The non-transitory computer-readable medium of claim 39 , wherein the management of the recognized sounds and images includes the audio synthesis.
45. The non-transitory computer-readable medium of claim 39 , wherein the management of the recognized sounds and images includes the pitch shifting.
46. The non-transitory computer-readable medium of claim 39 , wherein the management of the recognized sounds and images includes the time stretching.
47. A non-transitory computer-readable medium storing instructions that, when executed by a one or more processors, cause the processors to perform a method for managing audio and visual content in streaming video content, the method comprising:
receiving a user input to manage specific types of sounds and images in the streaming video content;
recognizing the specific types of sounds and images in the streaming video content based on the user input; and
providing a managed version of the streaming video content, wherein the recognized specific types of sounds and images are managed in the managed version of the streaming video content,
wherein the user input is received through a user interface on a device selected from a group consisting of:
a computer device;
a mobile device;
a smart television;
a virtual reality device;
an augmented reality device;
a gaming console;
a wearable device;
a home automation system; and
a vehicle infotainment system.
48. The non-transitory computer-readable medium of claim 47 , wherein the device is the computer device.
49. The non-transitory computer-readable medium of claim 47 , wherein the device is the mobile device.
50. The non-transitory computer-readable medium of claim 47 , wherein the device is the smart television.
51. The non-transitory computer-readable medium of claim 47 , wherein the device is the virtual reality device.
52. The non-transitory computer-readable medium of claim 47 , wherein the device is the augmented reality device.
53. The non-transitory computer-readable medium of claim 47 , wherein the device is the gaming console.
54. The non-transitory computer-readable medium of claim 47 , wherein the device is the wearable device.
55. The non-transitory computer-readable medium of claim 47 , wherein the device is the home automation system.
56. The non-transitory computer-readable medium of claim 47 , wherein the device is the vehicle infotainment system.
57. The non-transitory computer-readable medium of claim 47 , wherein providing the managed version of the streaming video content includes blurring the specific types of images.
58. The non-transitory computer-readable medium of claim 47 , wherein providing the managed version of the streaming video content includes obscuring the specific types of images.
59. The non-transitory computer-readable medium of claim 47 , wherein providing the managed version of the streaming video content includes replacement of the specific types of images with new visual content.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US18/501,524 US11974012B1 (en) | 2023-11-03 | 2023-11-03 | Modifying audio and video content based on user input |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US18/501,524 US11974012B1 (en) | 2023-11-03 | 2023-11-03 | Modifying audio and video content based on user input |
Publications (1)
Publication Number | Publication Date |
---|---|
US11974012B1 true US11974012B1 (en) | 2024-04-30 |
Family
ID=90836043
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US18/501,524 Active US11974012B1 (en) | 2023-11-03 | 2023-11-03 | Modifying audio and video content based on user input |
Country Status (1)
Country | Link |
---|---|
US (1) | US11974012B1 (en) |
Citations (63)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030037335A1 (en) * | 2001-08-17 | 2003-02-20 | Jean-Marie Gatto | Interactive television devices and systems |
US20060056802A1 (en) * | 2004-08-17 | 2006-03-16 | Seo Kang S | Method and apparatus of reproducing data recorded on recording medium and local storage |
US20090310939A1 (en) * | 2008-06-12 | 2009-12-17 | Basson Sara H | Simulation method and system |
US7714878B2 (en) * | 2004-08-09 | 2010-05-11 | Nice Systems, Ltd. | Apparatus and method for multimedia content based manipulation |
US20100125353A1 (en) * | 2008-11-14 | 2010-05-20 | Marc Petit-Huguenin | Systems and methods for distributed conferencing |
US20110035034A1 (en) * | 2006-01-06 | 2011-02-10 | Google Inc. | Serving Media Articles with Altered Playback Speed |
US20120030699A1 (en) | 2010-08-01 | 2012-02-02 | Umesh Amin | Systems and methods for storing and rendering atleast an user preference based media content |
US20120159530A1 (en) | 2010-12-16 | 2012-06-21 | Cisco Technology, Inc. | Micro-Filtering of Streaming Entertainment Content Based on Parental Control Setting |
US8230343B2 (en) | 1999-03-29 | 2012-07-24 | Digitalsmiths, Inc. | Audio and video program recording, editing and playback systems using metadata |
US20140013228A1 (en) * | 2012-06-07 | 2014-01-09 | TapThere, Inc. | Remote Experience Interfaces, Systems and Methods |
US8639796B2 (en) * | 2004-12-16 | 2014-01-28 | Hewlett-Packard Development Company, L.P. | Monitoring the performance of a streaming media server using server-side and client-side measurements |
US20140043424A1 (en) * | 2012-08-09 | 2014-02-13 | Samsung Electronics Co., Ltd. | Video calling using a remote camera device to stream video to a local endpoint host acting as a proxy |
US20140229604A1 (en) * | 2013-02-12 | 2014-08-14 | Howard Leigh Pfeffer | Adaptive bit rate admission control of a shared resource |
US8879895B1 (en) * | 2009-03-28 | 2014-11-04 | Matrox Electronic Systems Ltd. | System and method for processing ancillary data associated with a video stream |
US20140358520A1 (en) | 2013-05-31 | 2014-12-04 | Thomson Licensing | Real-time online audio filtering |
US8922659B2 (en) * | 2008-06-03 | 2014-12-30 | Thales | Dynamically reconfigurable intelligent video surveillance system |
US20150070516A1 (en) | 2012-12-14 | 2015-03-12 | Biscotti Inc. | Automatic Content Filtering |
US20150201198A1 (en) * | 2014-01-15 | 2015-07-16 | Avigilon Corporation | Streaming multiple encodings encoded using different encoding parameters |
US20150215586A1 (en) * | 2014-01-29 | 2015-07-30 | Sensormatic Electronics, LLC | Selection and Display of Adaptive Rate Streams in Video Security System |
US9197974B1 (en) | 2012-01-06 | 2015-11-24 | Audience, Inc. | Directional audio capture adaptation based on alternative sensory input |
US20150341570A1 (en) * | 2014-05-21 | 2015-11-26 | Mersive Technologies, Inc. | Intelligent shared display infrastructure and associated methods |
US20160037217A1 (en) | 2014-02-18 | 2016-02-04 | Vidangel, Inc. | Curating Filters for Audiovisual Content |
US20160073138A1 (en) * | 2014-09-04 | 2016-03-10 | Comcast Cable Communications, Llc | User-Defined Content Streaming |
US9454993B1 (en) * | 2015-10-09 | 2016-09-27 | Sports Logic Group, LLC | System capable of integrating user-entered game event data with corresponding video data |
US20160285724A1 (en) * | 2015-03-27 | 2016-09-29 | Axis Ab | Method and devices for negotiating bandwidth in a peer-to-peer network |
US9716914B1 (en) | 2008-03-28 | 2017-07-25 | Rovi Guides, Inc. | Systems and methods for blocking selected commercials |
US20170264920A1 (en) | 2016-03-08 | 2017-09-14 | Echostar Technologies L.L.C. | Apparatus, systems and methods for control of sporting event presentation based on viewer engagement |
US9779554B2 (en) | 2015-04-10 | 2017-10-03 | Sony Interactive Entertainment Inc. | Filtering and parental control methods for restricting visual activity on a head mounted display |
US20170289624A1 (en) | 2016-04-01 | 2017-10-05 | Samsung ElectrĂ´nica da AmazĂ´nia Ltda. | Multimodal and real-time method for filtering sensitive media |
US20170295215A1 (en) | 2016-04-08 | 2017-10-12 | Microsoft Technology Licensing, Llc | Audience targeted filtering of content sections |
US9788777B1 (en) | 2013-08-12 | 2017-10-17 | The Neilsen Company (US), LLC | Methods and apparatus to identify a mood of media |
US9870798B2 (en) | 2013-05-02 | 2018-01-16 | FreshTake Media, Inc. | Interactive real-time video editor and recorder |
US20180091856A1 (en) | 2016-09-27 | 2018-03-29 | International Business Machines Corporation | Video censoring |
US20180349502A1 (en) | 2017-01-26 | 2018-12-06 | Rena Maycock | Data Content Filter |
US20180359477A1 (en) | 2012-03-05 | 2018-12-13 | Google Inc. | Distribution of video in multiple rating formats |
US20180376205A1 (en) | 2015-12-17 | 2018-12-27 | Thomson Licensing | Method and apparatus for remote parental control of content viewing in augmented reality settings |
US20190052471A1 (en) | 2017-08-10 | 2019-02-14 | Microsoft Technology Licensing, Llc | Personalized toxicity shield for multiuser virtual environments |
US10210884B2 (en) | 2013-08-21 | 2019-02-19 | Google Llc | Systems and methods facilitating selective removal of content from a mixed audio recording |
US20190130189A1 (en) | 2017-10-30 | 2019-05-02 | Qualcomm Incorporated | Suppressing duplicated bounding boxes from object detection in a video analytics system |
US10372991B1 (en) | 2018-04-03 | 2019-08-06 | Google Llc | Systems and methods that leverage deep learning to selectively store audiovisual content |
US20190306563A1 (en) * | 2016-12-21 | 2019-10-03 | Arris Enterprises Llc | Automatic activation of closed captioning for low volume periods |
US10440324B1 (en) | 2018-09-06 | 2019-10-08 | Amazon Technologies, Inc. | Altering undesirable communication data for communication sessions |
US10506237B1 (en) * | 2016-05-27 | 2019-12-10 | Google Llc | Methods and devices for dynamic adaptation of encoding bitrate for video streaming |
US20200077150A1 (en) | 2018-08-28 | 2020-03-05 | International Business Machines Corporation | Filtering Images of Live Stream Content |
US20200092610A1 (en) | 2018-09-19 | 2020-03-19 | International Business Machines Corporation | Dynamically providing customized versions of video content |
US10678828B2 (en) | 2016-01-03 | 2020-06-09 | Gracenote, Inc. | Model-based media classification service using sensed media noise characteristics |
US10924786B2 (en) * | 2019-05-08 | 2021-02-16 | Nanning Fugui Precision Industrial Co., Ltd. | Method for shaping video streams and set-up box using the method |
US10951902B2 (en) * | 2019-06-12 | 2021-03-16 | Rovi Guides, Inc. | Systems and methods for multiple bit rate content encoding |
US10990188B2 (en) | 2012-08-17 | 2021-04-27 | Flextronics Ap, Llc | Systems and methods for providing video on demand in an intelligent television |
US11205254B2 (en) | 2017-08-30 | 2021-12-21 | Pxlize, Llc | System and method for identifying and obscuring objectionable content |
US20220007075A1 (en) | 2019-06-27 | 2022-01-06 | Apple Inc. | Modifying Existing Content Based on Target Audience |
US11232686B2 (en) * | 2019-12-30 | 2022-01-25 | Axis Ab | Real-time deviation in video monitoring |
US20220038402A1 (en) | 2016-06-30 | 2022-02-03 | Snap Inc. | Automated content curation and communication |
US11336968B2 (en) | 2018-08-17 | 2022-05-17 | Samsung Electronics Co., Ltd. | Method and device for generating content |
US11407118B1 (en) * | 2018-12-10 | 2022-08-09 | Joseph E Augenbraun | Robot for performing dextrous tasks and related methods and systems |
US20220408131A1 (en) | 2021-06-22 | 2022-12-22 | Q Factor Holdings LLC | Image analysis system |
US11553251B2 (en) | 2014-06-20 | 2023-01-10 | Comcast Cable Communications, Llc | Content viewing tracking |
US11601721B2 (en) | 2018-06-04 | 2023-03-07 | JBF Interlude 2009 LTD | Interactive video dynamic adaptation and user profiling |
US20230076702A1 (en) * | 2021-08-31 | 2023-03-09 | JBF Interlude 2009 LTD | Shader-based dynamic video manipulation |
US11609868B1 (en) * | 2020-12-31 | 2023-03-21 | Waymo Llc | Control calibration timing to avoid memory write blackout period |
US20230115250A1 (en) * | 2021-09-27 | 2023-04-13 | Apple Inc. | User interfaces for providing live video |
US11695978B2 (en) * | 2018-07-05 | 2023-07-04 | Mux, Inc. | Methods for generating video-and audience-specific encoding ladders with audio and video just-in-time transcoding |
US11748055B2 (en) * | 2018-11-16 | 2023-09-05 | Roku, Inc. | Monitoring loudness level during media replacement event using shorter time constant |
-
2023
- 2023-11-03 US US18/501,524 patent/US11974012B1/en active Active
Patent Citations (65)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8230343B2 (en) | 1999-03-29 | 2012-07-24 | Digitalsmiths, Inc. | Audio and video program recording, editing and playback systems using metadata |
US20030037335A1 (en) * | 2001-08-17 | 2003-02-20 | Jean-Marie Gatto | Interactive television devices and systems |
US7714878B2 (en) * | 2004-08-09 | 2010-05-11 | Nice Systems, Ltd. | Apparatus and method for multimedia content based manipulation |
US20060056802A1 (en) * | 2004-08-17 | 2006-03-16 | Seo Kang S | Method and apparatus of reproducing data recorded on recording medium and local storage |
US8639796B2 (en) * | 2004-12-16 | 2014-01-28 | Hewlett-Packard Development Company, L.P. | Monitoring the performance of a streaming media server using server-side and client-side measurements |
US20110035034A1 (en) * | 2006-01-06 | 2011-02-10 | Google Inc. | Serving Media Articles with Altered Playback Speed |
US9716914B1 (en) | 2008-03-28 | 2017-07-25 | Rovi Guides, Inc. | Systems and methods for blocking selected commercials |
US8922659B2 (en) * | 2008-06-03 | 2014-12-30 | Thales | Dynamically reconfigurable intelligent video surveillance system |
US20090310939A1 (en) * | 2008-06-12 | 2009-12-17 | Basson Sara H | Simulation method and system |
US20100125353A1 (en) * | 2008-11-14 | 2010-05-20 | Marc Petit-Huguenin | Systems and methods for distributed conferencing |
US8879895B1 (en) * | 2009-03-28 | 2014-11-04 | Matrox Electronic Systems Ltd. | System and method for processing ancillary data associated with a video stream |
US20120030699A1 (en) | 2010-08-01 | 2012-02-02 | Umesh Amin | Systems and methods for storing and rendering atleast an user preference based media content |
US20120159530A1 (en) | 2010-12-16 | 2012-06-21 | Cisco Technology, Inc. | Micro-Filtering of Streaming Entertainment Content Based on Parental Control Setting |
US9197974B1 (en) | 2012-01-06 | 2015-11-24 | Audience, Inc. | Directional audio capture adaptation based on alternative sensory input |
US20180359477A1 (en) | 2012-03-05 | 2018-12-13 | Google Inc. | Distribution of video in multiple rating formats |
US20140013228A1 (en) * | 2012-06-07 | 2014-01-09 | TapThere, Inc. | Remote Experience Interfaces, Systems and Methods |
US20140043424A1 (en) * | 2012-08-09 | 2014-02-13 | Samsung Electronics Co., Ltd. | Video calling using a remote camera device to stream video to a local endpoint host acting as a proxy |
US10990188B2 (en) | 2012-08-17 | 2021-04-27 | Flextronics Ap, Llc | Systems and methods for providing video on demand in an intelligent television |
US20150070516A1 (en) | 2012-12-14 | 2015-03-12 | Biscotti Inc. | Automatic Content Filtering |
US20140229604A1 (en) * | 2013-02-12 | 2014-08-14 | Howard Leigh Pfeffer | Adaptive bit rate admission control of a shared resource |
US9870798B2 (en) | 2013-05-02 | 2018-01-16 | FreshTake Media, Inc. | Interactive real-time video editor and recorder |
US20140358520A1 (en) | 2013-05-31 | 2014-12-04 | Thomson Licensing | Real-time online audio filtering |
US9788777B1 (en) | 2013-08-12 | 2017-10-17 | The Neilsen Company (US), LLC | Methods and apparatus to identify a mood of media |
US10210884B2 (en) | 2013-08-21 | 2019-02-19 | Google Llc | Systems and methods facilitating selective removal of content from a mixed audio recording |
US20150201198A1 (en) * | 2014-01-15 | 2015-07-16 | Avigilon Corporation | Streaming multiple encodings encoded using different encoding parameters |
US20150215586A1 (en) * | 2014-01-29 | 2015-07-30 | Sensormatic Electronics, LLC | Selection and Display of Adaptive Rate Streams in Video Security System |
US20160037217A1 (en) | 2014-02-18 | 2016-02-04 | Vidangel, Inc. | Curating Filters for Audiovisual Content |
US20150341570A1 (en) * | 2014-05-21 | 2015-11-26 | Mersive Technologies, Inc. | Intelligent shared display infrastructure and associated methods |
US11553251B2 (en) | 2014-06-20 | 2023-01-10 | Comcast Cable Communications, Llc | Content viewing tracking |
US20160073138A1 (en) * | 2014-09-04 | 2016-03-10 | Comcast Cable Communications, Llc | User-Defined Content Streaming |
US20160285724A1 (en) * | 2015-03-27 | 2016-09-29 | Axis Ab | Method and devices for negotiating bandwidth in a peer-to-peer network |
US9779554B2 (en) | 2015-04-10 | 2017-10-03 | Sony Interactive Entertainment Inc. | Filtering and parental control methods for restricting visual activity on a head mounted display |
US9454993B1 (en) * | 2015-10-09 | 2016-09-27 | Sports Logic Group, LLC | System capable of integrating user-entered game event data with corresponding video data |
US20180376205A1 (en) | 2015-12-17 | 2018-12-27 | Thomson Licensing | Method and apparatus for remote parental control of content viewing in augmented reality settings |
US10678828B2 (en) | 2016-01-03 | 2020-06-09 | Gracenote, Inc. | Model-based media classification service using sensed media noise characteristics |
US20170264920A1 (en) | 2016-03-08 | 2017-09-14 | Echostar Technologies L.L.C. | Apparatus, systems and methods for control of sporting event presentation based on viewer engagement |
US20170289624A1 (en) | 2016-04-01 | 2017-10-05 | Samsung ElectrĂ´nica da AmazĂ´nia Ltda. | Multimodal and real-time method for filtering sensitive media |
US20170295215A1 (en) | 2016-04-08 | 2017-10-12 | Microsoft Technology Licensing, Llc | Audience targeted filtering of content sections |
US11082701B2 (en) * | 2016-05-27 | 2021-08-03 | Google Llc | Methods and devices for dynamic adaptation of encoding bitrate for video streaming |
US10506237B1 (en) * | 2016-05-27 | 2019-12-10 | Google Llc | Methods and devices for dynamic adaptation of encoding bitrate for video streaming |
US20220038402A1 (en) | 2016-06-30 | 2022-02-03 | Snap Inc. | Automated content curation and communication |
US20180091856A1 (en) | 2016-09-27 | 2018-03-29 | International Business Machines Corporation | Video censoring |
US20190306563A1 (en) * | 2016-12-21 | 2019-10-03 | Arris Enterprises Llc | Automatic activation of closed captioning for low volume periods |
US20180349502A1 (en) | 2017-01-26 | 2018-12-06 | Rena Maycock | Data Content Filter |
US20190052471A1 (en) | 2017-08-10 | 2019-02-14 | Microsoft Technology Licensing, Llc | Personalized toxicity shield for multiuser virtual environments |
US11205254B2 (en) | 2017-08-30 | 2021-12-21 | Pxlize, Llc | System and method for identifying and obscuring objectionable content |
US20190130189A1 (en) | 2017-10-30 | 2019-05-02 | Qualcomm Incorporated | Suppressing duplicated bounding boxes from object detection in a video analytics system |
US10372991B1 (en) | 2018-04-03 | 2019-08-06 | Google Llc | Systems and methods that leverage deep learning to selectively store audiovisual content |
US11601721B2 (en) | 2018-06-04 | 2023-03-07 | JBF Interlude 2009 LTD | Interactive video dynamic adaptation and user profiling |
US11695978B2 (en) * | 2018-07-05 | 2023-07-04 | Mux, Inc. | Methods for generating video-and audience-specific encoding ladders with audio and video just-in-time transcoding |
US11336968B2 (en) | 2018-08-17 | 2022-05-17 | Samsung Electronics Co., Ltd. | Method and device for generating content |
US20200077150A1 (en) | 2018-08-28 | 2020-03-05 | International Business Machines Corporation | Filtering Images of Live Stream Content |
US10440324B1 (en) | 2018-09-06 | 2019-10-08 | Amazon Technologies, Inc. | Altering undesirable communication data for communication sessions |
US11582420B1 (en) | 2018-09-06 | 2023-02-14 | Amazon Technologies, Inc. | Altering undesirable communication data for communication sessions |
US20200092610A1 (en) | 2018-09-19 | 2020-03-19 | International Business Machines Corporation | Dynamically providing customized versions of video content |
US11748055B2 (en) * | 2018-11-16 | 2023-09-05 | Roku, Inc. | Monitoring loudness level during media replacement event using shorter time constant |
US11407118B1 (en) * | 2018-12-10 | 2022-08-09 | Joseph E Augenbraun | Robot for performing dextrous tasks and related methods and systems |
US10924786B2 (en) * | 2019-05-08 | 2021-02-16 | Nanning Fugui Precision Industrial Co., Ltd. | Method for shaping video streams and set-up box using the method |
US10951902B2 (en) * | 2019-06-12 | 2021-03-16 | Rovi Guides, Inc. | Systems and methods for multiple bit rate content encoding |
US20220007075A1 (en) | 2019-06-27 | 2022-01-06 | Apple Inc. | Modifying Existing Content Based on Target Audience |
US11232686B2 (en) * | 2019-12-30 | 2022-01-25 | Axis Ab | Real-time deviation in video monitoring |
US11609868B1 (en) * | 2020-12-31 | 2023-03-21 | Waymo Llc | Control calibration timing to avoid memory write blackout period |
US20220408131A1 (en) | 2021-06-22 | 2022-12-22 | Q Factor Holdings LLC | Image analysis system |
US20230076702A1 (en) * | 2021-08-31 | 2023-03-09 | JBF Interlude 2009 LTD | Shader-based dynamic video manipulation |
US20230115250A1 (en) * | 2021-09-27 | 2023-04-13 | Apple Inc. | User interfaces for providing live video |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11538454B2 (en) | Systems and methods for presenting social network communications in audible form based on user engagement with a user device | |
US11863336B2 (en) | Dynamic virtual environment | |
JP6487367B2 (en) | Selective content presentation engine | |
US10826949B2 (en) | Distributed control of media content item during webcast | |
US10586536B2 (en) | Display device and operating method therefor | |
CN106605218B (en) | Method for collecting and processing computer user data during interaction with network-based content | |
US20150070516A1 (en) | Automatic Content Filtering | |
CN112449253B (en) | Interactive video generation | |
US20050223237A1 (en) | Emotion controlled system for processing multimedia data | |
US20230260514A1 (en) | Systems and methods for providing voice command recommendations | |
JP7436460B2 (en) | Neural network inference on protected data | |
WO2018125161A1 (en) | Aggregation of media effects | |
US10645464B2 (en) | Eyes free entertainment | |
US10864447B1 (en) | Highlight presentation interface in a game spectating system | |
CN112423081B (en) | Video data processing method, device and equipment and readable storage medium | |
US20220148614A1 (en) | Automatically Captioning Audible Parts of Content on a Computing Device | |
JP2012039550A (en) | Information processing device, information processing system, information processing method and program | |
CN109299326A (en) | Video recommendation method and device, system, electronic equipment and storage medium | |
US11974012B1 (en) | Modifying audio and video content based on user input | |
US12003821B2 (en) | Techniques for enhanced media experience | |
US12112758B2 (en) | Systems and methods for determining traits based on voice analysis | |
CN109948426A (en) | Application program method of adjustment, device, electronic equipment and storage medium | |
US20190332656A1 (en) | Adaptive interactive media method and system | |
WO2024233129A1 (en) | Real-time ai screening and auto-moderation of audio comments in a livestream | |
US11935557B2 (en) | Techniques for detecting and processing domain-specific terminology |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: MICROENTITY |
|
FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO SMALL (ORIGINAL EVENT CODE: SMAL); ENTITY STATUS OF PATENT OWNER: MICROENTITY Free format text: ENTITY STATUS SET TO MICRO (ORIGINAL EVENT CODE: MICR); ENTITY STATUS OF PATENT OWNER: MICROENTITY |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |