[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

US11974012B1 - Modifying audio and video content based on user input - Google Patents

Modifying audio and video content based on user input Download PDF

Info

Publication number
US11974012B1
US11974012B1 US18/501,524 US202318501524A US11974012B1 US 11974012 B1 US11974012 B1 US 11974012B1 US 202318501524 A US202318501524 A US 202318501524A US 11974012 B1 US11974012 B1 US 11974012B1
Authority
US
United States
Prior art keywords
video content
images
streaming video
sounds
specific types
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
US18/501,524
Inventor
Kim Norton Benton
Kimberly Ann Higgins
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Avtech Select LLC
Original Assignee
Avtech Select LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Avtech Select LLC filed Critical Avtech Select LLC
Priority to US18/501,524 priority Critical patent/US11974012B1/en
Assigned to AVTech Select LLC reassignment AVTech Select LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: BENTON, KIM NORTON, HIGGINS, KIMBERLY ANN
Application granted granted Critical
Publication of US11974012B1 publication Critical patent/US11974012B1/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/439Processing of audio elementary streams
    • H04N21/4394Processing of audio elementary streams involving operations for analysing the audio stream, e.g. detecting features or characteristics in audio streams
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/422Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
    • H04N21/42203Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS] sound input device, e.g. microphone
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/422Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
    • H04N21/4223Cameras
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/8106Monomedia components thereof involving special audio data, e.g. different tracks for different languages
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/8146Monomedia components thereof involving graphical data, e.g. 3D object, 2D graphics
    • H04N21/8153Monomedia components thereof involving graphical data, e.g. 3D object, 2D graphics comprising still images, e.g. texture, background image

Definitions

  • the disclosure is related to techniques for modifying audio and video content.
  • This disclosure includes techniques for modifying specific types of sounds and images in streaming video content in response to a user input. For example, a user may request modification or elimination of sounds and images, such as animal sounds or images of violence, which may cause distress or disturbance to the user.
  • a system for managing audio and visual content in streaming video content includes a user interaction module configured to receive a user input to manage specific types of sounds and images in the streaming video content, a recognition module configured to recognize the specific types of sounds and images in the streaming video content based on the user input, and a modification module configured to provide a managed version of the streaming video content, wherein the recognized specific types of sounds and images are managed in the managed version of the streaming video content.
  • this disclosure is directed towards a non-transitory computer-readable medium storing instructions that, when executed by a processor, cause the processor to perform a method for managing audio and visual content in streaming video content, the method comprising: receiving a user input to manage specific types of sounds and images in the streaming video content, recognizing the specific types of sounds and images in the streaming video content based on the user input, and providing a managed version of the streaming video content, wherein the recognized specific types of sounds and images are managed in the managed version of the streaming video content.
  • FIG. 1 is a conceptual illustration of a system for modifying specific types of sounds and images in streaming video content in response to a user input.
  • FIG. 2 is a flowchart illustrating example techniques for modifying specific types of sounds and images in streaming video content in response to a user input.
  • This disclosure includes techniques for modifying specific types of sounds and images in streaming video content in response to a user input. For example, a user may request modification or elimination of sounds and images, such as animal sounds or images of violence, which may cause distress or disturbance to the user.
  • the disclosed techniques allow a user to customize the audio and visual content of the video based on their personal preferences and requirements. This customization is achieved by managing specific types of sounds and images in the video content, particularly those that some users may find disruptive, annoying, or distressing.
  • FIG. 1 is a conceptual illustration of a system 100 for modifying specific types of sounds and images in streaming video content 101 in response to a user input.
  • the system 100 can manage animal sounds in the video content. Animal sounds, such as the barking of dogs or the chirping of birds, can potentially excite pets, wake sleeping children, or cause stress to individuals who may find these sounds disturbing. By providing the option to manage these sounds, the system 100 allows users to enjoy the video content without being disturbed by such sounds.
  • the system 100 can also manage images in the video content. Certain images, such as images of violence or guns, can potentially cause distress or disturbance to viewers of the content. By providing the option to manage these images, the system 100 allows users to enjoy the video content without being subjected to such images.
  • the system 100 provides a personalized and responsive viewing experience. Whether it's managing animal sounds, war sounds, gunshots, images of animals, war scenes, violence, or guns, the system 100 provides users with the flexibility to customize their viewing experience, making the viewing of streaming video content 101 a more enjoyable and personalized experience.
  • the user input module 110 is configured to receive a user input from user interface 112 to manage specific types of sounds and images in the streaming video content 101 .
  • the user input can include a selection of specified sounds and images to be managed, and specific levels of management for the sounds and images. For instance, a user may choose to completely suppress the sounds of dogs barking, or to reduce the volume of bird chirping by a specific percentage. This flexibility allows the user to customize their viewing experience according to their personal preferences and requirements.
  • the user input module 110 can receive user input through a user interface 112 implemented on a variety of devices. These devices can include, but are not limited to, computer devices, mobile devices, and smart televisions. For instance, a user may input their preferences through a touch screen interface on a mobile device or tablet, a voice recognition system on a smart speaker or virtual assistant, a graphical user interface (GUI) on a computer or smart TV, or a physical control panel on a device such as a mixer or soundboard.
  • GUI graphical user interface
  • the user interaction module of the system 100 is designed to receive user input through a user interface 112 .
  • This user interface 112 serves as the primary point of interaction between the user and the system 100 , enabling the user to specify their preferences for the management of sounds and images in the streaming video content 101 .
  • the user interface 112 can include various interactive elements, such as drop-down menus, sliders, checkboxes, or text input fields, that allow the user to specify their preferences. It can also provide visual feedback, such as a preview of the managed sounds or a visual representation of the selected images, to assist the user in making their selection.
  • the user interface 112 can be implemented on a variety of devices, each offering different capabilities and user experiences. These devices can include, but are not limited to, computer devices, mobile devices, smart televisions, virtual reality devices, augmented reality devices, gaming consoles, wearable devices, home automation systems, and vehicle infotainment systems. For instance, a user may input their preferences through a touch screen interface on a mobile device or tablet, a voice recognition system on a smart speaker or virtual assistant, a graphical user interface (GUI) on a computer or smart TV, or a physical control panel on a device such as a mixer or soundboard.
  • GUI graphical user interface
  • the user interface 112 is designed to be compatible with a wide range of devices, providing users with the flexibility to input their preferences through the device that is the easiest and the convenient for them.
  • the user interface 112 can provide users with the option to save their preferences, allowing them to apply the same settings to future streaming video content 101 without having to input their preferences each time.
  • the identification module 120 is configured to recognize the specific types of sounds and images in the streaming video content 101 based on the user input.
  • the identification module 120 uses various techniques to analyze the audio and visual content of the video and detect the presence of the specified sounds and images. These techniques can include spectral analysis, machine learning algorithms, pattern recognition techniques, audio fingerprinting, and deep learning techniques.
  • Spectral analysis involves examining the frequency spectrum of the audio signal to identify specific types of sounds. Each animal sound has a distinct frequency spectrum, which can be compared with a library of frequency spectra of known animal sounds. This comparison allows the identification module 120 to accurately identify the presence of specific types of animal sounds in the streaming video content 101 .
  • Machine learning algorithms can be trained on a dataset of animal sounds, enabling them to recognize similar sounds in the streaming video content 101 .
  • These algorithms can include supervised learning algorithms, such as support vector machines or neural networks, which are capable of learning complex patterns in the audio data and accurately identifying animal sounds.
  • Pattern recognition techniques involve comparing the audio content of the streaming video with a library of animal sounds to find matches. By comparing the pattern of the audio signal with a library of patterns of known animal sounds, the identification module 120 can accurately identify the presence of specific types of animal sounds in the streaming video content 101 .
  • Audio fingerprinting is a technique that involves creating a condensed digital summary, or “fingerprint,” of an audio signal. This fingerprint can then be compared with a database of fingerprints of known animal sounds. If the fingerprint of the audio signal matches the fingerprint of a known animal sound in the database, the identification module 120 can identify the presence of that animal sound in the streaming video content 101 .
  • Deep learning techniques such as convolutional neural networks (CNNs)
  • CNNs are particularly effective at identifying patterns in audio data, making them well-suited for the task of identifying animal sounds.
  • the identification module 120 can learn to recognize the distinct features of each animal sound and accurately identify these sounds in the streaming video content 101 .
  • the identification module 120 can accurately recognize specific types of sounds and images in the streaming video content 101 based on the user input. This recognition serves as a foundation for the subsequent management process, enabling the system 100 to manage the desired sounds and images in the streaming video content 101 .
  • the modification module 130 is specifically designed to provide a managed version of the streaming video content 101 , wherein the recognized specific types of sounds and images are managed as per the user input. Specifically, the modification module 130 alters the desired sounds and images in the streaming video content 101 .
  • the modification module 130 may apply any number of techniques to modify audio signals. These techniques include volume attenuation, noise cancellation, audio synthesis, pitch shifting, and time stretching. These techniques can also include suppression, replacement, and distortion of the sounds and images.
  • Volume attenuation is a process of reducing the intensity of an audio signal. It is often used in audio mixing and sound engineering to balance the levels of different audio sources. Volume attenuation can be used to decrease the loudness of specific sounds in the streaming video content, such as animal sounds or loud noises, based on user preferences.
  • Time stretching is a digital audio processing technique that allows the duration of an audio signal to be changed without affecting its pitch. This is achieved by altering the speed or tempo of the audio signal without changing its spectral characteristics.
  • Time stretching is one of the techniques or algorithms that can be used to modify identified sounds in streaming video content. For instance, if a specific animal sound or a loud noise is identified in the video content, time stretching can be used to lengthen or shorten the duration of this sound in the modified version of the content. This can help to reduce the disruptive impact of the sound on the viewer, enhancing their viewing experience.
  • Noise Cancellation This is a method used to reduce unwanted sounds by creating an “anti-noise” sound wave that interferes with the unwanted noise, effectively cancelling it out. Active noise cancellation is commonly used in headphones to reduce background noise.
  • Audio Synthesis This is the electronic production of sound. It can involve creating new sounds or altering existing ones. Synthesizers, which can generate a wide range of sounds, are commonly used for audio synthesis. Audio synthesis can be used to create alternative sounds to replace specific types of sounds in the streaming video content, such as animal sounds or loud noises.
  • Pitch Shifting This is a sound recording technique where the original pitch of a sound is raised or lowered. Effects units that utilize pitch shifting can transpose an audio signal up or down in pitch. Pitch shifting can be used to alter the pitch of specific sounds in the streaming video content, changing the way they are perceived by the viewer.
  • Suppression involves reducing the amplitude of the audio signal corresponding to the recognized sounds. This effectively lowers the volume of these sounds in the streaming video content 101 , making them less noticeable or completely inaudible, depending on the degree of suppression applied. This technique can be particularly useful for users who wish to suppress the sounds of specific animals without completely removing them from the audio content.
  • Replacement involves generating a new audio signal or visual content to replace the recognized sounds or images.
  • the sounds of dogs barking could be replaced with the sounds of birds chirping, or the images of violence could be replaced with more peaceful scenes.
  • the new audio signals or visual content can be selected from a library of sounds and images, or they can be synthesized on the fly using sound synthesis algorithms or image generation techniques. This technique can be particularly useful for users who wish to replace the sounds or images of specific elements with more pleasant or less disturbing sounds or images.
  • Distortion involves changing the characteristics of the audio signal or visual content corresponding to the recognized sounds or images. This alters the quality of these sounds or images, making them sound or look different from their original form. This can be particularly useful for users who find the original quality of these sounds or images disturbing or annoying. The degree of distortion can be adjusted based on the user's preferences, allowing them to customize the quality of the sounds or images to their liking.
  • the system 100 can be hosted on various types of systems, including, but not limited to, a cloud-based system, a local device system, a distributed network system, a peer-to-peer network system, an edge computing system, and a hybrid cloud-edge system.
  • a cloud-based system a local device system
  • a distributed network system a peer-to-peer network system
  • an edge computing system a hybrid cloud-edge system.
  • the disclosed techniques may be embodied in a non-transitory computer-readable medium storing instructions that, when executed by one or more processors, cause the processors to perform a method for managing audio and visual content.
  • the user interaction module, the recognition module, and the modification module 130 are hosted on a cloud computing platform.
  • the user inputs are received and processed in the cloud, and the managed streaming video content 101 is streamed back to the user's device over the internet.
  • This system allows for remote access and management of the streaming video content 101 , and can scale to handle large volumes of video content and user inputs. However, it may require a stable and high-speed internet connection, and there may be concerns about data privacy and security in the cloud.
  • the user interaction module, the recognition module, and the modification module 130 are hosted on the user's device, such as a computer, mobile device, or smart TV.
  • the user inputs are received and processed on the device, and the managed streaming video content 101 is streamed directly on the device.
  • This system allows for real-time management of the streaming video content 101 , and can provide a more personalized user experience. However, it may require more computational resources on the device, and the system 100 's performance may be limited by the device's capabilities.
  • the user interaction module, the recognition module, and the modification module 130 are hosted on a network of servers.
  • the user inputs are received and processed on the server that is closest to the user, and the managed streaming video content 101 is streamed back to the user's device over the network.
  • This system allows for handling of user inputs and streaming video content 101 , and can provide a high-quality streaming experience for the user.
  • it may require a complex network infrastructure, and there may be concerns about data privacy and security in the network.
  • the user interaction module, the recognition module, and the modification module 130 are distributed across a network of user devices.
  • Each device contributes its resources to the network, processing user inputs and streaming video content 101 in a decentralized manner.
  • This system can provide high resilience and scalability, as the failure or overload of one device does not affect the overall performance of the system 100 . However, it may require a robust network protocol, and the system 100 's performance may be affected by the capabilities and availability of the user devices in the network.
  • the user interaction module, the recognition module, and the modification module 130 are hosted on edge devices located close to the user's location. These edge devices can be routers, gateways, or other network devices that provide low latency access to the user.
  • the user inputs are processed at the edge of the network, allowing for real-time management of the streaming video content 101 .
  • the managed content is then streamed directly to the user's device, reducing the latency and bandwidth usage.
  • a hybrid cloud-edge system some components of the system 100 are hosted on a cloud platform while others are hosted on edge devices.
  • the user interaction module and modification module 130 could be hosted on the cloud for global accessibility and scalability, while the recognition module could be hosted on edge devices for real-time processing.
  • This system combines the benefits of cloud and edge computing, providing a balance between scalability, latency, and bandwidth usage.
  • system 100 may also managing images of animals in video content. Some users may find images of specific animals to be disturbing or distracting. For instance, a user may have a phobia of spiders and may wish to avoid seeing images of spiders in video content. In such cases, the system 100 can analyze the video content to detect the presence of images of specific animals based on user input, and then provide a managed version of the content where these images are blurred, obscured, or replaced with other images. This feature can enhance the user experience, allowing users to enjoy the video content without being subjected to objectionable images.
  • War scenes can be distressing for some users, causing fear, exacerbating post-traumatic stress disorder (PTSD), or simply being too violent or graphic for some viewers.
  • the system 100 can analyze the video content to detect the presence of war scenes based on user input, and then provide a managed version of the content where these scenes are blurred, obscured, or replaced with other scenes. This feature can enhance the user experience, allowing users to enjoy the video content without being subjected to distressing scenes.
  • PTSD post-traumatic stress disorder
  • a further variation of these techniques includes managing images of violence or guns in video content.
  • Images of violence or guns can be disturbing for some users, particularly those who are sensitive to such content or those who are watching the content with young children.
  • the system 100 can analyze the video content to detect the presence of images of violence or guns based on user input, and then provide a managed version of the content where these images are blurred, obscured, or replaced with other images. This feature can enhance the user experience, allowing users to enjoy the video content without being subjected to disturbing images.
  • the system 100 may employs any of a variety of techniques to analyze the video content and detect the presence of specific types of images. These techniques can include image recognition techniques, machine learning algorithms, pattern recognition techniques, and deep learning techniques. Once the specific types of images have been detected, the modification module 130 modifies these images using various techniques, such as image blurring, image obscuring, or image replacement. The modified images are then combined with the original audio content or managed audio content of the video to provide a managed version of the video content that caters to the user's preferences and requirements.
  • system 100 may further include a monitoring device 114 to observe reactions to the video content.
  • Monitoring device 114 may include a camera or a microphone, for example.
  • the monitoring device 114 is strategically configured to observe the users and the surrounding noise, thereby providing a comprehensive understanding of the user's environment and reactions to the streaming video content 101 .
  • a camera When a camera is employed as the monitoring device 114 , it captures visual data that can be analyzed to detect user 115 responses to the streaming video content 101 . For instance, the camera may capture the physical reactions of a pet in response to animal sounds in the video content. Similarly, the camera can also capture visual cues of human stress, such as facial expressions or body language, in response to loud noises in the video content.
  • the visual data captured by the camera provides a rich source of information that can be analyzed by the detection module 150 to identify specific responses to the audio and/or video elements in the streaming video content 101 .
  • a microphone when a microphone is used as the monitoring device 114 , it records the surrounding noise and audio responses from the user or pets. For example, the microphone can detect a dog barking in response to animal sounds in the video content or a baby crying in response to loud noises. The audio data captured by the microphone provides a direct measure of the user's or pet's reactions to the streaming video content 101 , which can be analyzed by the detection module 150 to identify specific responses.
  • the monitoring device 114 is configured to continuously capture data while the streaming video content 101 is being consumed. This continuous monitoring allows for real-time detection of ambient response, thereby enabling the system 100 to promptly react to any potentially disruptive elements in the video content. The data collected by the monitoring device 114 is then relayed to the detection module 150 for further analysis.
  • the detection module 150 is designed to analyze the data received from the monitoring device 114 and detect specific responses to the audio and/or video elements in the streaming video content 101 .
  • the detection module 150 is configured to identify a range of responses, from signs of human stress to indications of excited animals.
  • the detection module 150 is configured to recognize specific sound patterns or visual cues associated with various responses. For instance, the detection module 150 may use sound pattern recognition techniques to identify a dog's bark or a baby's cry. Similarly, it may use visual cue recognition techniques to identify signs of human stress or excitement in response to the video content. These techniques ensure that the detection module 150 accurately identifies responses, thereby enabling the system 100 to react accordingly.
  • the detection module 150 operates in real-time, analyzing the data received from the monitoring device 114 as it is captured. This real-time operation allows the system 100 to promptly detect ambient responses and react accordingly, thereby enhancing the user's viewing experience. Once a response is detected, the detection module 150 communicates with the query module 160 , initiating the next step in the system 100 's process of managing content in streaming video.
  • the interaction between the detection module 150 and the other components of the system 100 is facilitated through a communication protocol.
  • This protocol ensures that the detection of a response by the detection module 150 is communicated to the query module 160 .
  • the seamless interaction between these components is integral to the system 100 's ability to effectively manage content in streaming video, thereby enhancing the user's viewing experience.
  • the query module 160 is activated when the detection module 150 identifies a response to video content presented on display 102 .
  • the primary function of the query module 160 is to interact with the user upon the detection of a response, providing the user with the option to manage the content in the streaming video content 101 .
  • the query module 160 is configured to ask the user whether they would like to manage the content in the streaming video content 101 . This interaction is designed to provide the user with control over the content they consume, allowing them to filter out potentially disruptive elements. For instance, if the system 100 detects a dog barking in response to animal sounds in the video content, the query module 160 can interact with the user, asking them if they would like to manage the content. This could involve filtering out the animal sounds that caused the dog's response.
  • the query module 160 can interact with the user, asking them if they would like to manage the content. This could involve filtering out the loud noises that caused the baby's response. This interaction is designed to provide the user with control over the content they consume, allowing them to filter out potentially disruptive elements.
  • the interaction between the query module 160 and the user can take various forms, depending on the user's preferences and the nature of the detected ambient response.
  • the query module 160 could present the user with a pop-up message on their screen, asking them if they would like to manage the content.
  • the query module 160 could send the user a notification on their mobile device, asking them the same question.
  • the specific form of interaction can be customized based on the user's preferences, thereby enhancing the user's control over the content they consume.
  • the query module 160 interacts with the user upon the detection of a response. By asking the user whether they would like to manage the content in the streaming video content 101 , the query module 160 provides the user with control over the content they consume, allowing them to filter out potentially disruptive elements. This interaction enhances the user's viewing experience, making the viewing of streaming video content 101 a more enjoyable and personalized experience.
  • the system 100 is designed to manage a variety of audio and visual elements within the streaming video content 101 . This includes, but is not limited to, animal sounds and loud noises such as gunfire, fights, or arguments.
  • the system's ability to manage these types of content is integral to its purpose of enhancing the user's viewing experience.
  • Animal sounds in the video content can cause excitement or stress in pets.
  • the system 100 is designed to identify these sounds and manage them based on the user's preferences and the detected ambient responses. For example, if a dog barks in response to animal sounds in the video content, the system 100 can filter out these sounds, thereby preventing further excitement or stress in the pet.
  • the detection module is configured to recognizing specific sound patterns associated with various animal sounds observed by monitoring device 114 .
  • Loud noises in the video content can also cause stress in humans, particularly in young children.
  • the system 100 is designed to identify these noises and manage them based on the user's preferences and the detected ambient responses. For instance, if a baby cries in response to loud noises in the video content, the system 100 can filter out these noises, thereby preventing further stress in the baby.
  • the detection module 150 is configured distinguishing these noises from other sounds in the video content.
  • the system 100 may be designed to operate in real-time, allowing for prompt reaction to potentially disruptive elements in the video content. This real-time operation is facilitated by the seamless interaction between the system 100 's components, particularly the monitoring device 114 , the detection module 150 , and the query module 160 .
  • the monitoring device 114 captures the audio and visual data
  • the detection module 150 analyzes this data and detects specific responses
  • the query module 160 interacts with the user, providing them with the option to manage the content.
  • This cohesive operation of the system 100 's components enables effective management of content in streaming video, thereby enhancing the user's viewing experience.
  • the system 100 could be configured to automatically manage the content based on the detected ambient responses, without querying the user. For example, if the system 100 detects a dog barking in response to animal sounds in the video content, it could automatically filter out these sounds, thereby preventing further excitement or stress in the pet. This automatic content management could enhance the system 100 's responsiveness to potentially disruptive elements in the video content, thereby enhancing the user's viewing experience.
  • FIG. 2 is a flowchart 200 illustrating example techniques for modifying specific types of sounds and images in streaming video content in response to a user input. For clarity, the techniques of FIG. 2 are described with respect to system 100 ( FIG. 1 ).
  • a user device such as display 102 issues a request to access video content, such as streaming video content 101 (step 202 ).
  • identification module 120 accesses the video content (step 204 ).
  • identification module 120 recognizes the specific types of sounds and images in the streaming video content based on the user input. (step 206 ).
  • Modification module 130 provides a managed version of the streaming video content to the user device (step 208 ).
  • the managed version of the streaming video content includes altered content such that specific types of sounds and images are managed in the managed version of the streaming video content.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Computer Graphics (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

A system for managing audio and visual content in streaming video content includes a user interaction module configured to receive a user input to manage specific types of sounds and images in the streaming video content, a recognition module configured to recognize the specific types of sounds and images in the streaming video content based on the user input, and a modification module configured to provide a managed version of the streaming video content, wherein the recognized specific types of sounds and images are managed in the managed version of the streaming video content.

Description

TECHNICAL FIELD
The disclosure is related to techniques for modifying audio and video content.
BACKGROUND OF THE INVENTION
The advent of digital technology has revolutionized the way we consume media. In particular, the proliferation of streaming video content has provided users with unprecedented access to a vast array of multimedia content. This content ranges from movies and television shows to live broadcasts and user-generated videos. As the technology has evolved, so too have the tools and techniques for manipulating and customizing this content to suit individual user preferences.
SUMMARY OF THE INVENTION
This disclosure includes techniques for modifying specific types of sounds and images in streaming video content in response to a user input. For example, a user may request modification or elimination of sounds and images, such as animal sounds or images of violence, which may cause distress or disturbance to the user.
In one example, a system for managing audio and visual content in streaming video content includes a user interaction module configured to receive a user input to manage specific types of sounds and images in the streaming video content, a recognition module configured to recognize the specific types of sounds and images in the streaming video content based on the user input, and a modification module configured to provide a managed version of the streaming video content, wherein the recognized specific types of sounds and images are managed in the managed version of the streaming video content.
In another example, this disclosure is directed towards a non-transitory computer-readable medium storing instructions that, when executed by a processor, cause the processor to perform a method for managing audio and visual content in streaming video content, the method comprising: receiving a user input to manage specific types of sounds and images in the streaming video content, recognizing the specific types of sounds and images in the streaming video content based on the user input, and providing a managed version of the streaming video content, wherein the recognized specific types of sounds and images are managed in the managed version of the streaming video content.
BRIEF DESCRIPTION OF DRAWINGS
FIG. 1 is a conceptual illustration of a system for modifying specific types of sounds and images in streaming video content in response to a user input.
FIG. 2 is a flowchart illustrating example techniques for modifying specific types of sounds and images in streaming video content in response to a user input.
DETAILED DESCRIPTION OF THE INVENTION
This disclosure includes techniques for modifying specific types of sounds and images in streaming video content in response to a user input. For example, a user may request modification or elimination of sounds and images, such as animal sounds or images of violence, which may cause distress or disturbance to the user.
The disclosed techniques allow a user to customize the audio and visual content of the video based on their personal preferences and requirements. This customization is achieved by managing specific types of sounds and images in the video content, particularly those that some users may find disruptive, annoying, or distressing.
FIG. 1 is a conceptual illustration of a system 100 for modifying specific types of sounds and images in streaming video content 101 in response to a user input. For instance, the system 100 can manage animal sounds in the video content. Animal sounds, such as the barking of dogs or the chirping of birds, can potentially excite pets, wake sleeping children, or cause stress to individuals who may find these sounds disturbing. By providing the option to manage these sounds, the system 100 allows users to enjoy the video content without being disturbed by such sounds.
Similarly, the system 100 can also manage images in the video content. Certain images, such as images of violence or guns, can potentially cause distress or disturbance to viewers of the content. By providing the option to manage these images, the system 100 allows users to enjoy the video content without being subjected to such images.
The system 100 provides a personalized and responsive viewing experience. Whether it's managing animal sounds, war sounds, gunshots, images of animals, war scenes, violence, or guns, the system 100 provides users with the flexibility to customize their viewing experience, making the viewing of streaming video content 101 a more enjoyable and personalized experience.
User Input Module 110
The user input module 110 is configured to receive a user input from user interface 112 to manage specific types of sounds and images in the streaming video content 101. The user input can include a selection of specified sounds and images to be managed, and specific levels of management for the sounds and images. For instance, a user may choose to completely suppress the sounds of dogs barking, or to reduce the volume of bird chirping by a specific percentage. This flexibility allows the user to customize their viewing experience according to their personal preferences and requirements.
The user input module 110 can receive user input through a user interface 112 implemented on a variety of devices. These devices can include, but are not limited to, computer devices, mobile devices, and smart televisions. For instance, a user may input their preferences through a touch screen interface on a mobile device or tablet, a voice recognition system on a smart speaker or virtual assistant, a graphical user interface (GUI) on a computer or smart TV, or a physical control panel on a device such as a mixer or soundboard.
User Interface 112
The user interaction module of the system 100 is designed to receive user input through a user interface 112. This user interface 112 serves as the primary point of interaction between the user and the system 100, enabling the user to specify their preferences for the management of sounds and images in the streaming video content 101. The user interface 112 can include various interactive elements, such as drop-down menus, sliders, checkboxes, or text input fields, that allow the user to specify their preferences. It can also provide visual feedback, such as a preview of the managed sounds or a visual representation of the selected images, to assist the user in making their selection.
The user interface 112 can be implemented on a variety of devices, each offering different capabilities and user experiences. These devices can include, but are not limited to, computer devices, mobile devices, smart televisions, virtual reality devices, augmented reality devices, gaming consoles, wearable devices, home automation systems, and vehicle infotainment systems. For instance, a user may input their preferences through a touch screen interface on a mobile device or tablet, a voice recognition system on a smart speaker or virtual assistant, a graphical user interface (GUI) on a computer or smart TV, or a physical control panel on a device such as a mixer or soundboard. The user interface 112 is designed to be compatible with a wide range of devices, providing users with the flexibility to input their preferences through the device that is the easiest and the convenient for them.
Furthermore, the user interface 112 can provide users with the option to save their preferences, allowing them to apply the same settings to future streaming video content 101 without having to input their preferences each time.
Identification Module 120
The identification module 120 is configured to recognize the specific types of sounds and images in the streaming video content 101 based on the user input. The identification module 120 uses various techniques to analyze the audio and visual content of the video and detect the presence of the specified sounds and images. These techniques can include spectral analysis, machine learning algorithms, pattern recognition techniques, audio fingerprinting, and deep learning techniques.
Spectral analysis involves examining the frequency spectrum of the audio signal to identify specific types of sounds. Each animal sound has a distinct frequency spectrum, which can be compared with a library of frequency spectra of known animal sounds. This comparison allows the identification module 120 to accurately identify the presence of specific types of animal sounds in the streaming video content 101.
Machine learning algorithms can be trained on a dataset of animal sounds, enabling them to recognize similar sounds in the streaming video content 101. These algorithms can include supervised learning algorithms, such as support vector machines or neural networks, which are capable of learning complex patterns in the audio data and accurately identifying animal sounds.
Pattern recognition techniques involve comparing the audio content of the streaming video with a library of animal sounds to find matches. By comparing the pattern of the audio signal with a library of patterns of known animal sounds, the identification module 120 can accurately identify the presence of specific types of animal sounds in the streaming video content 101.
Audio fingerprinting is a technique that involves creating a condensed digital summary, or “fingerprint,” of an audio signal. This fingerprint can then be compared with a database of fingerprints of known animal sounds. If the fingerprint of the audio signal matches the fingerprint of a known animal sound in the database, the identification module 120 can identify the presence of that animal sound in the streaming video content 101.
Deep learning techniques, such as convolutional neural networks (CNNs), can be used to identify animal sounds in streaming video content 101. CNNs are particularly effective at identifying patterns in audio data, making them well-suited for the task of identifying animal sounds. By training a CNN on a dataset of animal sounds, the identification module 120 can learn to recognize the distinct features of each animal sound and accurately identify these sounds in the streaming video content 101.
Through the use of these techniques, the identification module 120 can accurately recognize specific types of sounds and images in the streaming video content 101 based on the user input. This recognition serves as a foundation for the subsequent management process, enabling the system 100 to manage the desired sounds and images in the streaming video content 101.
Modification Module 130
The modification module 130 is specifically designed to provide a managed version of the streaming video content 101, wherein the recognized specific types of sounds and images are managed as per the user input. Specifically, the modification module 130 alters the desired sounds and images in the streaming video content 101. The modification module 130 may apply any number of techniques to modify audio signals. These techniques include volume attenuation, noise cancellation, audio synthesis, pitch shifting, and time stretching. These techniques can also include suppression, replacement, and distortion of the sounds and images.
Volume attenuation is a process of reducing the intensity of an audio signal. It is often used in audio mixing and sound engineering to balance the levels of different audio sources. Volume attenuation can be used to decrease the loudness of specific sounds in the streaming video content, such as animal sounds or loud noises, based on user preferences.
Time stretching is a digital audio processing technique that allows the duration of an audio signal to be changed without affecting its pitch. This is achieved by altering the speed or tempo of the audio signal without changing its spectral characteristics. Time stretching is one of the techniques or algorithms that can be used to modify identified sounds in streaming video content. For instance, if a specific animal sound or a loud noise is identified in the video content, time stretching can be used to lengthen or shorten the duration of this sound in the modified version of the content. This can help to reduce the disruptive impact of the sound on the viewer, enhancing their viewing experience.
Noise Cancellation: This is a method used to reduce unwanted sounds by creating an “anti-noise” sound wave that interferes with the unwanted noise, effectively cancelling it out. Active noise cancellation is commonly used in headphones to reduce background noise.
Audio Synthesis: This is the electronic production of sound. It can involve creating new sounds or altering existing ones. Synthesizers, which can generate a wide range of sounds, are commonly used for audio synthesis. Audio synthesis can be used to create alternative sounds to replace specific types of sounds in the streaming video content, such as animal sounds or loud noises.
Pitch Shifting: This is a sound recording technique where the original pitch of a sound is raised or lowered. Effects units that utilize pitch shifting can transpose an audio signal up or down in pitch. Pitch shifting can be used to alter the pitch of specific sounds in the streaming video content, changing the way they are perceived by the viewer.
Suppression involves reducing the amplitude of the audio signal corresponding to the recognized sounds. This effectively lowers the volume of these sounds in the streaming video content 101, making them less noticeable or completely inaudible, depending on the degree of suppression applied. This technique can be particularly useful for users who wish to suppress the sounds of specific animals without completely removing them from the audio content.
Replacement involves generating a new audio signal or visual content to replace the recognized sounds or images. For instance, the sounds of dogs barking could be replaced with the sounds of birds chirping, or the images of violence could be replaced with more peaceful scenes. The new audio signals or visual content can be selected from a library of sounds and images, or they can be synthesized on the fly using sound synthesis algorithms or image generation techniques. This technique can be particularly useful for users who wish to replace the sounds or images of specific elements with more pleasant or less disturbing sounds or images.
Distortion involves changing the characteristics of the audio signal or visual content corresponding to the recognized sounds or images. This alters the quality of these sounds or images, making them sound or look different from their original form. This can be particularly useful for users who find the original quality of these sounds or images disturbing or annoying. The degree of distortion can be adjusted based on the user's preferences, allowing them to customize the quality of the sounds or images to their liking.
System 100 Hardware Examples
The system 100 can be hosted on various types of systems, including, but not limited to, a cloud-based system, a local device system, a distributed network system, a peer-to-peer network system, an edge computing system, and a hybrid cloud-edge system. With any of these systems, the disclosed techniques may be embodied in a non-transitory computer-readable medium storing instructions that, when executed by one or more processors, cause the processors to perform a method for managing audio and visual content.
Cloud-Based System
In a cloud-based system, the user interaction module, the recognition module, and the modification module 130 are hosted on a cloud computing platform. The user inputs are received and processed in the cloud, and the managed streaming video content 101 is streamed back to the user's device over the internet. This system allows for remote access and management of the streaming video content 101, and can scale to handle large volumes of video content and user inputs. However, it may require a stable and high-speed internet connection, and there may be concerns about data privacy and security in the cloud.
Local Device System
In a local device system, the user interaction module, the recognition module, and the modification module 130 are hosted on the user's device, such as a computer, mobile device, or smart TV. The user inputs are received and processed on the device, and the managed streaming video content 101 is streamed directly on the device. This system allows for real-time management of the streaming video content 101, and can provide a more personalized user experience. However, it may require more computational resources on the device, and the system 100's performance may be limited by the device's capabilities.
Distributed Network System
In a distributed network system, the user interaction module, the recognition module, and the modification module 130 are hosted on a network of servers. The user inputs are received and processed on the server that is closest to the user, and the managed streaming video content 101 is streamed back to the user's device over the network. This system allows for handling of user inputs and streaming video content 101, and can provide a high-quality streaming experience for the user. However, it may require a complex network infrastructure, and there may be concerns about data privacy and security in the network.
Peer-to-Peer Network System
In a peer-to-peer network system, the user interaction module, the recognition module, and the modification module 130 are distributed across a network of user devices. Each device contributes its resources to the network, processing user inputs and streaming video content 101 in a decentralized manner. This system can provide high resilience and scalability, as the failure or overload of one device does not affect the overall performance of the system 100. However, it may require a robust network protocol, and the system 100's performance may be affected by the capabilities and availability of the user devices in the network.
Edge Computing System
In an edge computing system, the user interaction module, the recognition module, and the modification module 130 are hosted on edge devices located close to the user's location. These edge devices can be routers, gateways, or other network devices that provide low latency access to the user. The user inputs are processed at the edge of the network, allowing for real-time management of the streaming video content 101. The managed content is then streamed directly to the user's device, reducing the latency and bandwidth usage.
Hybrid Cloud-Edge System
In a hybrid cloud-edge system, some components of the system 100 are hosted on a cloud platform while others are hosted on edge devices. For instance, the user interaction module and modification module 130 could be hosted on the cloud for global accessibility and scalability, while the recognition module could be hosted on edge devices for real-time processing. This system combines the benefits of cloud and edge computing, providing a balance between scalability, latency, and bandwidth usage.
Image Filtering
In addition to the techniques for managing specific types of sounds in streaming video content 101, system 100 may also managing images of animals in video content. Some users may find images of specific animals to be disturbing or distracting. For instance, a user may have a phobia of spiders and may wish to avoid seeing images of spiders in video content. In such cases, the system 100 can analyze the video content to detect the presence of images of specific animals based on user input, and then provide a managed version of the content where these images are blurred, obscured, or replaced with other images. This feature can enhance the user experience, allowing users to enjoy the video content without being subjected to objectionable images.
Another variation of these techniques includes managing war scenes in video content. War scenes can be distressing for some users, causing fear, exacerbating post-traumatic stress disorder (PTSD), or simply being too violent or graphic for some viewers. In such cases, the system 100 can analyze the video content to detect the presence of war scenes based on user input, and then provide a managed version of the content where these scenes are blurred, obscured, or replaced with other scenes. This feature can enhance the user experience, allowing users to enjoy the video content without being subjected to distressing scenes.
A further variation of these techniques includes managing images of violence or guns in video content. Images of violence or guns can be disturbing for some users, particularly those who are sensitive to such content or those who are watching the content with young children. In such cases, the system 100 can analyze the video content to detect the presence of images of violence or guns based on user input, and then provide a managed version of the content where these images are blurred, obscured, or replaced with other images. This feature can enhance the user experience, allowing users to enjoy the video content without being subjected to disturbing images.
In all these variations, the system 100 may employs any of a variety of techniques to analyze the video content and detect the presence of specific types of images. These techniques can include image recognition techniques, machine learning algorithms, pattern recognition techniques, and deep learning techniques. Once the specific types of images have been detected, the modification module 130 modifies these images using various techniques, such as image blurring, image obscuring, or image replacement. The modified images are then combined with the original audio content or managed audio content of the video to provide a managed version of the video content that caters to the user's preferences and requirements.
Monitoring Device 114
In a variation of the techniques, system 100 may further include a monitoring device 114 to observe reactions to the video content. Monitoring device 114 may include a camera or a microphone, for example. The monitoring device 114 is strategically configured to observe the users and the surrounding noise, thereby providing a comprehensive understanding of the user's environment and reactions to the streaming video content 101.
When a camera is employed as the monitoring device 114, it captures visual data that can be analyzed to detect user 115 responses to the streaming video content 101. For instance, the camera may capture the physical reactions of a pet in response to animal sounds in the video content. Similarly, the camera can also capture visual cues of human stress, such as facial expressions or body language, in response to loud noises in the video content. The visual data captured by the camera provides a rich source of information that can be analyzed by the detection module 150 to identify specific responses to the audio and/or video elements in the streaming video content 101.
On the other hand, when a microphone is used as the monitoring device 114, it records the surrounding noise and audio responses from the user or pets. For example, the microphone can detect a dog barking in response to animal sounds in the video content or a baby crying in response to loud noises. The audio data captured by the microphone provides a direct measure of the user's or pet's reactions to the streaming video content 101, which can be analyzed by the detection module 150 to identify specific responses.
The monitoring device 114 is configured to continuously capture data while the streaming video content 101 is being consumed. This continuous monitoring allows for real-time detection of ambient response, thereby enabling the system 100 to promptly react to any potentially disruptive elements in the video content. The data collected by the monitoring device 114 is then relayed to the detection module 150 for further analysis.
Detection Module 150
The detection module 150 is designed to analyze the data received from the monitoring device 114 and detect specific responses to the audio and/or video elements in the streaming video content 101. The detection module 150 is configured to identify a range of responses, from signs of human stress to indications of excited animals.
The detection module 150 is configured to recognize specific sound patterns or visual cues associated with various responses. For instance, the detection module 150 may use sound pattern recognition techniques to identify a dog's bark or a baby's cry. Similarly, it may use visual cue recognition techniques to identify signs of human stress or excitement in response to the video content. These techniques ensure that the detection module 150 accurately identifies responses, thereby enabling the system 100 to react accordingly.
The detection module 150 operates in real-time, analyzing the data received from the monitoring device 114 as it is captured. This real-time operation allows the system 100 to promptly detect ambient responses and react accordingly, thereby enhancing the user's viewing experience. Once a response is detected, the detection module 150 communicates with the query module 160, initiating the next step in the system 100's process of managing content in streaming video.
The interaction between the detection module 150 and the other components of the system 100 is facilitated through a communication protocol. This protocol ensures that the detection of a response by the detection module 150 is communicated to the query module 160. The seamless interaction between these components is integral to the system 100's ability to effectively manage content in streaming video, thereby enhancing the user's viewing experience.
Query Module 160
The query module 160 is activated when the detection module 150 identifies a response to video content presented on display 102. The primary function of the query module 160 is to interact with the user upon the detection of a response, providing the user with the option to manage the content in the streaming video content 101.
The query module 160 is configured to ask the user whether they would like to manage the content in the streaming video content 101. This interaction is designed to provide the user with control over the content they consume, allowing them to filter out potentially disruptive elements. For instance, if the system 100 detects a dog barking in response to animal sounds in the video content, the query module 160 can interact with the user, asking them if they would like to manage the content. This could involve filtering out the animal sounds that caused the dog's response.
Similarly, if the system 100 detects a baby crying in response to loud noises in the video content, the query module 160 can interact with the user, asking them if they would like to manage the content. This could involve filtering out the loud noises that caused the baby's response. This interaction is designed to provide the user with control over the content they consume, allowing them to filter out potentially disruptive elements.
The interaction between the query module 160 and the user can take various forms, depending on the user's preferences and the nature of the detected ambient response. For example, the query module 160 could present the user with a pop-up message on their screen, asking them if they would like to manage the content. Alternatively, the query module 160 could send the user a notification on their mobile device, asking them the same question. The specific form of interaction can be customized based on the user's preferences, thereby enhancing the user's control over the content they consume.
In summary, the query module 160 interacts with the user upon the detection of a response. By asking the user whether they would like to manage the content in the streaming video content 101, the query module 160 provides the user with control over the content they consume, allowing them to filter out potentially disruptive elements. This interaction enhances the user's viewing experience, making the viewing of streaming video content 101 a more enjoyable and personalized experience.
The system 100 is designed to manage a variety of audio and visual elements within the streaming video content 101. This includes, but is not limited to, animal sounds and loud noises such as gunfire, fights, or arguments. The system's ability to manage these types of content is integral to its purpose of enhancing the user's viewing experience.
Animal sounds in the video content, for instance, can cause excitement or stress in pets. The system 100 is designed to identify these sounds and manage them based on the user's preferences and the detected ambient responses. For example, if a dog barks in response to animal sounds in the video content, the system 100 can filter out these sounds, thereby preventing further excitement or stress in the pet. The detection module is configured to recognizing specific sound patterns associated with various animal sounds observed by monitoring device 114.
Loud noises in the video content, such as gunfire, fights, or arguments, can also cause stress in humans, particularly in young children. The system 100 is designed to identify these noises and manage them based on the user's preferences and the detected ambient responses. For instance, if a baby cries in response to loud noises in the video content, the system 100 can filter out these noises, thereby preventing further stress in the baby. The detection module 150 is configured distinguishing these noises from other sounds in the video content.
The system 100 may be designed to operate in real-time, allowing for prompt reaction to potentially disruptive elements in the video content. This real-time operation is facilitated by the seamless interaction between the system 100's components, particularly the monitoring device 114, the detection module 150, and the query module 160. The monitoring device 114 captures the audio and visual data, the detection module 150 analyzes this data and detects specific responses, and the query module 160 interacts with the user, providing them with the option to manage the content. This cohesive operation of the system 100's components enables effective management of content in streaming video, thereby enhancing the user's viewing experience.
In the same or different examples, the system 100 could be configured to automatically manage the content based on the detected ambient responses, without querying the user. For example, if the system 100 detects a dog barking in response to animal sounds in the video content, it could automatically filter out these sounds, thereby preventing further excitement or stress in the pet. This automatic content management could enhance the system 100's responsiveness to potentially disruptive elements in the video content, thereby enhancing the user's viewing experience.
FIG. 2 is a flowchart 200 illustrating example techniques for modifying specific types of sounds and images in streaming video content in response to a user input. For clarity, the techniques of FIG. 2 are described with respect to system 100 (FIG. 1 ).
First, a user device, such as display 102 issues a request to access video content, such as streaming video content 101 (step 202). In response to the request, identification module 120 accesses the video content (step 204). Based on a user input from user input module 110, identification module 120 recognizes the specific types of sounds and images in the streaming video content based on the user input. (step 206). Modification module 130, provides a managed version of the streaming video content to the user device (step 208). The managed version of the streaming video content includes altered content such that specific types of sounds and images are managed in the managed version of the streaming video content.
The specific techniques for modifying specific types of sounds and images in streaming video content in response to a user input, such as techniques embodied by system 100 are merely illustrative of the general inventive concepts included in this disclosure as defined by the following claims.

Claims (59)

What is claimed is:
1. A system for managing audio and visual content in streaming video content, the system comprising:
a user interaction module configured to receive a user input to manage specific types of sounds and images in the streaming video content;
a recognition module configured to recognize the specific types of sounds and images in the streaming video content based on the user input; and
a modification module configured to provide a managed version of the streaming video content, wherein the recognized specific types of sounds and images are managed in the managed version of the streaming video content,
wherein the management of the recognized sounds and images is achieved through techniques selected from a group consisting of:
volume attenuation;
noise cancellation;
audio synthesis;
pitch shifting; and
time stretching.
2. The system of claim 1, wherein the user input includes a selection of specified sounds and images to be managed in the streaming video content.
3. The system of claim 1, wherein the user input includes a selection of specific levels of management for the sounds and images.
4. The system of claim 1, further comprising an analysis module configured to analyze the streaming video content to detect the specific types of sounds and images.
5. The system of claim 1, further comprising a filtering module configured to filter other elements of a video signal based on a user selection.
6. The system of claim 1, wherein the user interaction module, the recognition module, and the modification module are hosted on a cloud computing system, allowing for remote access and management of the streaming video content.
7. The system of claim 1, wherein the management of the recognized sounds and images includes the volume attenuation.
8. The system of claim 1, wherein the management of the recognized sounds and images includes the noise cancellation.
9. The system of claim 1, wherein the management of the recognized sounds and images includes the audio synthesis.
10. The system of claim 1, wherein the management of the recognized sounds and images includes the pitch shifting.
11. The system of claim 1, wherein the management of the recognized sounds and images includes the time stretching.
12. A system for managing audio and visual content in streaming video content, the system comprising:
a user interaction module configured to receive a user input to manage specific types of sounds and images in the streaming video content;
a recognition module configured to recognize the specific types of sounds and images in the streaming video content based on the user input; and
a modification module configured to provide a managed version of the streaming video content, wherein the recognized specific types of sounds and images are managed in the managed version of the streaming video content,
wherein the user input is received through a user interface on a device selected from a group consisting of:
a computer device;
a mobile device;
a smart television;
a virtual reality device;
an augmented reality device;
a gaming console;
a wearable device;
a home automation system; and
a vehicle infotainment system.
13. The system of claim 12, wherein the device is the computer device.
14. The system of claim 12, wherein the device is the mobile device.
15. The system of claim 12, wherein the device is the smart television.
16. The system of claim 12, wherein the device is the virtual reality device.
17. The system of claim 12, wherein the device is the augmented reality device.
18. The system of claim 12, wherein the device is the gaming console.
19. The system of claim 12, wherein the device is the wearable device.
20. The system of claim 12, wherein the device is the home automation system.
21. The system of claim 12, wherein the device is the vehicle infotainment system.
22. The system of claim 12, wherein providing the managed version of the streaming video content includes blurring the specific types of images.
23. The system of claim 12, wherein providing the managed version of the streaming video content includes obscuring the specific types of images.
24. The system of claim 12, wherein providing the managed version of the streaming video content includes replacement of the specific types of images with new visual content.
25. A system for managing audio and visual content in streaming video content, the system comprising:
a user interaction module configured to receive a user input to manage specific types of sounds and images in the streaming video content;
a recognition module configured to recognize the specific types of sounds and images in the streaming video content based on the user input; and
a modification module configured to provide a managed version of the streaming video content, wherein the recognized specific types of sounds and images are managed in the managed version of the streaming video content,
wherein the recognition of sounds and images in the streaming video content is achieved through techniques selected from a group consisting of spectral analysis, machine learning algorithms, pattern recognition techniques, audio fingerprinting, and deep learning techniques.
26. The system of claim 25, wherein the recognition of sounds and images in the streaming video content includes the spectral analysis.
27. The system of claim 25, wherein the recognition of sounds and images in the streaming video content includes the machine learning algorithms.
28. The system of claim 25, wherein the recognition of sounds and images in the streaming video content includes the pattern recognition techniques.
29. The system of claim 25, wherein the recognition of sounds and images in the streaming video content includes the audio fingerprinting.
30. The system of claim 25, wherein the recognition of sounds and images in the streaming video content includes the deep learning techniques.
31. The system of claim 25, wherein providing the managed version of the streaming video content includes blurring the specific types of images.
32. The system of claim 25, wherein providing the managed version of the streaming video content includes obscuring the specific types of images.
33. The system of claim 25, wherein providing the managed version of the streaming video content includes replacement of the specific types of images with new visual content.
34. A system for managing audio and visual content in streaming video content, the system comprising:
a monitoring device configured to monitor users and noise;
a detection module configured to detect ambient responses to the streaming video content;
a query module configured to query the user whether they would like to manage content in the streaming video content in response to the detection of ambient response;
a user interaction module configured to receive a user input to manage specific types of sounds and images in the streaming video content;
a recognition module configured to recognize the specific types of sounds and images in the streaming video content based on the user input; and
a modification module configured to provide a managed version of the streaming video content, wherein the recognized specific types of sounds and images are managed in the managed version of the streaming video content.
35. The system of claim 34, wherein the monitoring device is selected from a group consisting of a camera and a microphone.
36. The system of claim 34, wherein the responses are selected from a group consisting of human stress and excited animals.
37. The system of claim 34, wherein the detection module is further configured to detect a dog barking in response to animal sounds in the streaming video content.
38. The system of claim 34, wherein the detection module is further configured to detect a baby crying in response to loud noises in the streaming video content.
39. A non-transitory computer-readable medium storing instructions that, when executed by a one or more processors, cause the processors to perform a method for managing audio and visual content in streaming video content, the method comprising:
receiving a user input to manage specific types of sounds and images in the streaming video content;
recognizing the specific types of sounds and images in the streaming video content based on the user input; and
providing a managed version of the streaming video content, wherein the recognized specific types of sounds and images are managed in the managed version of the streaming video content,
wherein the management of the recognized sounds and images is achieved through techniques selected from a group consisting of:
volume attenuation;
noise cancellation;
audio synthesis;
pitch shifting; and
time stretching.
40. The non-transitory computer-readable medium of claim 39, wherein the user input includes a selection of specified sounds and images to be managed in the streaming video content.
41. The non-transitory computer-readable medium of claim 39, wherein the user input includes a selection of specific levels of management for the sounds and images.
42. The non-transitory computer-readable medium of claim 39, wherein the management of the recognized sounds and images includes the volume attenuation.
43. The non-transitory computer-readable medium of claim 39, wherein the management of the recognized sounds and images includes the noise cancellation.
44. The non-transitory computer-readable medium of claim 39, wherein the management of the recognized sounds and images includes the audio synthesis.
45. The non-transitory computer-readable medium of claim 39, wherein the management of the recognized sounds and images includes the pitch shifting.
46. The non-transitory computer-readable medium of claim 39, wherein the management of the recognized sounds and images includes the time stretching.
47. A non-transitory computer-readable medium storing instructions that, when executed by a one or more processors, cause the processors to perform a method for managing audio and visual content in streaming video content, the method comprising:
receiving a user input to manage specific types of sounds and images in the streaming video content;
recognizing the specific types of sounds and images in the streaming video content based on the user input; and
providing a managed version of the streaming video content, wherein the recognized specific types of sounds and images are managed in the managed version of the streaming video content,
wherein the user input is received through a user interface on a device selected from a group consisting of:
a computer device;
a mobile device;
a smart television;
a virtual reality device;
an augmented reality device;
a gaming console;
a wearable device;
a home automation system; and
a vehicle infotainment system.
48. The non-transitory computer-readable medium of claim 47, wherein the device is the computer device.
49. The non-transitory computer-readable medium of claim 47, wherein the device is the mobile device.
50. The non-transitory computer-readable medium of claim 47, wherein the device is the smart television.
51. The non-transitory computer-readable medium of claim 47, wherein the device is the virtual reality device.
52. The non-transitory computer-readable medium of claim 47, wherein the device is the augmented reality device.
53. The non-transitory computer-readable medium of claim 47, wherein the device is the gaming console.
54. The non-transitory computer-readable medium of claim 47, wherein the device is the wearable device.
55. The non-transitory computer-readable medium of claim 47, wherein the device is the home automation system.
56. The non-transitory computer-readable medium of claim 47, wherein the device is the vehicle infotainment system.
57. The non-transitory computer-readable medium of claim 47, wherein providing the managed version of the streaming video content includes blurring the specific types of images.
58. The non-transitory computer-readable medium of claim 47, wherein providing the managed version of the streaming video content includes obscuring the specific types of images.
59. The non-transitory computer-readable medium of claim 47, wherein providing the managed version of the streaming video content includes replacement of the specific types of images with new visual content.
US18/501,524 2023-11-03 2023-11-03 Modifying audio and video content based on user input Active US11974012B1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US18/501,524 US11974012B1 (en) 2023-11-03 2023-11-03 Modifying audio and video content based on user input

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US18/501,524 US11974012B1 (en) 2023-11-03 2023-11-03 Modifying audio and video content based on user input

Publications (1)

Publication Number Publication Date
US11974012B1 true US11974012B1 (en) 2024-04-30

Family

ID=90836043

Family Applications (1)

Application Number Title Priority Date Filing Date
US18/501,524 Active US11974012B1 (en) 2023-11-03 2023-11-03 Modifying audio and video content based on user input

Country Status (1)

Country Link
US (1) US11974012B1 (en)

Citations (63)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030037335A1 (en) * 2001-08-17 2003-02-20 Jean-Marie Gatto Interactive television devices and systems
US20060056802A1 (en) * 2004-08-17 2006-03-16 Seo Kang S Method and apparatus of reproducing data recorded on recording medium and local storage
US20090310939A1 (en) * 2008-06-12 2009-12-17 Basson Sara H Simulation method and system
US7714878B2 (en) * 2004-08-09 2010-05-11 Nice Systems, Ltd. Apparatus and method for multimedia content based manipulation
US20100125353A1 (en) * 2008-11-14 2010-05-20 Marc Petit-Huguenin Systems and methods for distributed conferencing
US20110035034A1 (en) * 2006-01-06 2011-02-10 Google Inc. Serving Media Articles with Altered Playback Speed
US20120030699A1 (en) 2010-08-01 2012-02-02 Umesh Amin Systems and methods for storing and rendering atleast an user preference based media content
US20120159530A1 (en) 2010-12-16 2012-06-21 Cisco Technology, Inc. Micro-Filtering of Streaming Entertainment Content Based on Parental Control Setting
US8230343B2 (en) 1999-03-29 2012-07-24 Digitalsmiths, Inc. Audio and video program recording, editing and playback systems using metadata
US20140013228A1 (en) * 2012-06-07 2014-01-09 TapThere, Inc. Remote Experience Interfaces, Systems and Methods
US8639796B2 (en) * 2004-12-16 2014-01-28 Hewlett-Packard Development Company, L.P. Monitoring the performance of a streaming media server using server-side and client-side measurements
US20140043424A1 (en) * 2012-08-09 2014-02-13 Samsung Electronics Co., Ltd. Video calling using a remote camera device to stream video to a local endpoint host acting as a proxy
US20140229604A1 (en) * 2013-02-12 2014-08-14 Howard Leigh Pfeffer Adaptive bit rate admission control of a shared resource
US8879895B1 (en) * 2009-03-28 2014-11-04 Matrox Electronic Systems Ltd. System and method for processing ancillary data associated with a video stream
US20140358520A1 (en) 2013-05-31 2014-12-04 Thomson Licensing Real-time online audio filtering
US8922659B2 (en) * 2008-06-03 2014-12-30 Thales Dynamically reconfigurable intelligent video surveillance system
US20150070516A1 (en) 2012-12-14 2015-03-12 Biscotti Inc. Automatic Content Filtering
US20150201198A1 (en) * 2014-01-15 2015-07-16 Avigilon Corporation Streaming multiple encodings encoded using different encoding parameters
US20150215586A1 (en) * 2014-01-29 2015-07-30 Sensormatic Electronics, LLC Selection and Display of Adaptive Rate Streams in Video Security System
US9197974B1 (en) 2012-01-06 2015-11-24 Audience, Inc. Directional audio capture adaptation based on alternative sensory input
US20150341570A1 (en) * 2014-05-21 2015-11-26 Mersive Technologies, Inc. Intelligent shared display infrastructure and associated methods
US20160037217A1 (en) 2014-02-18 2016-02-04 Vidangel, Inc. Curating Filters for Audiovisual Content
US20160073138A1 (en) * 2014-09-04 2016-03-10 Comcast Cable Communications, Llc User-Defined Content Streaming
US9454993B1 (en) * 2015-10-09 2016-09-27 Sports Logic Group, LLC System capable of integrating user-entered game event data with corresponding video data
US20160285724A1 (en) * 2015-03-27 2016-09-29 Axis Ab Method and devices for negotiating bandwidth in a peer-to-peer network
US9716914B1 (en) 2008-03-28 2017-07-25 Rovi Guides, Inc. Systems and methods for blocking selected commercials
US20170264920A1 (en) 2016-03-08 2017-09-14 Echostar Technologies L.L.C. Apparatus, systems and methods for control of sporting event presentation based on viewer engagement
US9779554B2 (en) 2015-04-10 2017-10-03 Sony Interactive Entertainment Inc. Filtering and parental control methods for restricting visual activity on a head mounted display
US20170289624A1 (en) 2016-04-01 2017-10-05 Samsung ElectrĂ´nica da AmazĂ´nia Ltda. Multimodal and real-time method for filtering sensitive media
US20170295215A1 (en) 2016-04-08 2017-10-12 Microsoft Technology Licensing, Llc Audience targeted filtering of content sections
US9788777B1 (en) 2013-08-12 2017-10-17 The Neilsen Company (US), LLC Methods and apparatus to identify a mood of media
US9870798B2 (en) 2013-05-02 2018-01-16 FreshTake Media, Inc. Interactive real-time video editor and recorder
US20180091856A1 (en) 2016-09-27 2018-03-29 International Business Machines Corporation Video censoring
US20180349502A1 (en) 2017-01-26 2018-12-06 Rena Maycock Data Content Filter
US20180359477A1 (en) 2012-03-05 2018-12-13 Google Inc. Distribution of video in multiple rating formats
US20180376205A1 (en) 2015-12-17 2018-12-27 Thomson Licensing Method and apparatus for remote parental control of content viewing in augmented reality settings
US20190052471A1 (en) 2017-08-10 2019-02-14 Microsoft Technology Licensing, Llc Personalized toxicity shield for multiuser virtual environments
US10210884B2 (en) 2013-08-21 2019-02-19 Google Llc Systems and methods facilitating selective removal of content from a mixed audio recording
US20190130189A1 (en) 2017-10-30 2019-05-02 Qualcomm Incorporated Suppressing duplicated bounding boxes from object detection in a video analytics system
US10372991B1 (en) 2018-04-03 2019-08-06 Google Llc Systems and methods that leverage deep learning to selectively store audiovisual content
US20190306563A1 (en) * 2016-12-21 2019-10-03 Arris Enterprises Llc Automatic activation of closed captioning for low volume periods
US10440324B1 (en) 2018-09-06 2019-10-08 Amazon Technologies, Inc. Altering undesirable communication data for communication sessions
US10506237B1 (en) * 2016-05-27 2019-12-10 Google Llc Methods and devices for dynamic adaptation of encoding bitrate for video streaming
US20200077150A1 (en) 2018-08-28 2020-03-05 International Business Machines Corporation Filtering Images of Live Stream Content
US20200092610A1 (en) 2018-09-19 2020-03-19 International Business Machines Corporation Dynamically providing customized versions of video content
US10678828B2 (en) 2016-01-03 2020-06-09 Gracenote, Inc. Model-based media classification service using sensed media noise characteristics
US10924786B2 (en) * 2019-05-08 2021-02-16 Nanning Fugui Precision Industrial Co., Ltd. Method for shaping video streams and set-up box using the method
US10951902B2 (en) * 2019-06-12 2021-03-16 Rovi Guides, Inc. Systems and methods for multiple bit rate content encoding
US10990188B2 (en) 2012-08-17 2021-04-27 Flextronics Ap, Llc Systems and methods for providing video on demand in an intelligent television
US11205254B2 (en) 2017-08-30 2021-12-21 Pxlize, Llc System and method for identifying and obscuring objectionable content
US20220007075A1 (en) 2019-06-27 2022-01-06 Apple Inc. Modifying Existing Content Based on Target Audience
US11232686B2 (en) * 2019-12-30 2022-01-25 Axis Ab Real-time deviation in video monitoring
US20220038402A1 (en) 2016-06-30 2022-02-03 Snap Inc. Automated content curation and communication
US11336968B2 (en) 2018-08-17 2022-05-17 Samsung Electronics Co., Ltd. Method and device for generating content
US11407118B1 (en) * 2018-12-10 2022-08-09 Joseph E Augenbraun Robot for performing dextrous tasks and related methods and systems
US20220408131A1 (en) 2021-06-22 2022-12-22 Q Factor Holdings LLC Image analysis system
US11553251B2 (en) 2014-06-20 2023-01-10 Comcast Cable Communications, Llc Content viewing tracking
US11601721B2 (en) 2018-06-04 2023-03-07 JBF Interlude 2009 LTD Interactive video dynamic adaptation and user profiling
US20230076702A1 (en) * 2021-08-31 2023-03-09 JBF Interlude 2009 LTD Shader-based dynamic video manipulation
US11609868B1 (en) * 2020-12-31 2023-03-21 Waymo Llc Control calibration timing to avoid memory write blackout period
US20230115250A1 (en) * 2021-09-27 2023-04-13 Apple Inc. User interfaces for providing live video
US11695978B2 (en) * 2018-07-05 2023-07-04 Mux, Inc. Methods for generating video-and audience-specific encoding ladders with audio and video just-in-time transcoding
US11748055B2 (en) * 2018-11-16 2023-09-05 Roku, Inc. Monitoring loudness level during media replacement event using shorter time constant

Patent Citations (65)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8230343B2 (en) 1999-03-29 2012-07-24 Digitalsmiths, Inc. Audio and video program recording, editing and playback systems using metadata
US20030037335A1 (en) * 2001-08-17 2003-02-20 Jean-Marie Gatto Interactive television devices and systems
US7714878B2 (en) * 2004-08-09 2010-05-11 Nice Systems, Ltd. Apparatus and method for multimedia content based manipulation
US20060056802A1 (en) * 2004-08-17 2006-03-16 Seo Kang S Method and apparatus of reproducing data recorded on recording medium and local storage
US8639796B2 (en) * 2004-12-16 2014-01-28 Hewlett-Packard Development Company, L.P. Monitoring the performance of a streaming media server using server-side and client-side measurements
US20110035034A1 (en) * 2006-01-06 2011-02-10 Google Inc. Serving Media Articles with Altered Playback Speed
US9716914B1 (en) 2008-03-28 2017-07-25 Rovi Guides, Inc. Systems and methods for blocking selected commercials
US8922659B2 (en) * 2008-06-03 2014-12-30 Thales Dynamically reconfigurable intelligent video surveillance system
US20090310939A1 (en) * 2008-06-12 2009-12-17 Basson Sara H Simulation method and system
US20100125353A1 (en) * 2008-11-14 2010-05-20 Marc Petit-Huguenin Systems and methods for distributed conferencing
US8879895B1 (en) * 2009-03-28 2014-11-04 Matrox Electronic Systems Ltd. System and method for processing ancillary data associated with a video stream
US20120030699A1 (en) 2010-08-01 2012-02-02 Umesh Amin Systems and methods for storing and rendering atleast an user preference based media content
US20120159530A1 (en) 2010-12-16 2012-06-21 Cisco Technology, Inc. Micro-Filtering of Streaming Entertainment Content Based on Parental Control Setting
US9197974B1 (en) 2012-01-06 2015-11-24 Audience, Inc. Directional audio capture adaptation based on alternative sensory input
US20180359477A1 (en) 2012-03-05 2018-12-13 Google Inc. Distribution of video in multiple rating formats
US20140013228A1 (en) * 2012-06-07 2014-01-09 TapThere, Inc. Remote Experience Interfaces, Systems and Methods
US20140043424A1 (en) * 2012-08-09 2014-02-13 Samsung Electronics Co., Ltd. Video calling using a remote camera device to stream video to a local endpoint host acting as a proxy
US10990188B2 (en) 2012-08-17 2021-04-27 Flextronics Ap, Llc Systems and methods for providing video on demand in an intelligent television
US20150070516A1 (en) 2012-12-14 2015-03-12 Biscotti Inc. Automatic Content Filtering
US20140229604A1 (en) * 2013-02-12 2014-08-14 Howard Leigh Pfeffer Adaptive bit rate admission control of a shared resource
US9870798B2 (en) 2013-05-02 2018-01-16 FreshTake Media, Inc. Interactive real-time video editor and recorder
US20140358520A1 (en) 2013-05-31 2014-12-04 Thomson Licensing Real-time online audio filtering
US9788777B1 (en) 2013-08-12 2017-10-17 The Neilsen Company (US), LLC Methods and apparatus to identify a mood of media
US10210884B2 (en) 2013-08-21 2019-02-19 Google Llc Systems and methods facilitating selective removal of content from a mixed audio recording
US20150201198A1 (en) * 2014-01-15 2015-07-16 Avigilon Corporation Streaming multiple encodings encoded using different encoding parameters
US20150215586A1 (en) * 2014-01-29 2015-07-30 Sensormatic Electronics, LLC Selection and Display of Adaptive Rate Streams in Video Security System
US20160037217A1 (en) 2014-02-18 2016-02-04 Vidangel, Inc. Curating Filters for Audiovisual Content
US20150341570A1 (en) * 2014-05-21 2015-11-26 Mersive Technologies, Inc. Intelligent shared display infrastructure and associated methods
US11553251B2 (en) 2014-06-20 2023-01-10 Comcast Cable Communications, Llc Content viewing tracking
US20160073138A1 (en) * 2014-09-04 2016-03-10 Comcast Cable Communications, Llc User-Defined Content Streaming
US20160285724A1 (en) * 2015-03-27 2016-09-29 Axis Ab Method and devices for negotiating bandwidth in a peer-to-peer network
US9779554B2 (en) 2015-04-10 2017-10-03 Sony Interactive Entertainment Inc. Filtering and parental control methods for restricting visual activity on a head mounted display
US9454993B1 (en) * 2015-10-09 2016-09-27 Sports Logic Group, LLC System capable of integrating user-entered game event data with corresponding video data
US20180376205A1 (en) 2015-12-17 2018-12-27 Thomson Licensing Method and apparatus for remote parental control of content viewing in augmented reality settings
US10678828B2 (en) 2016-01-03 2020-06-09 Gracenote, Inc. Model-based media classification service using sensed media noise characteristics
US20170264920A1 (en) 2016-03-08 2017-09-14 Echostar Technologies L.L.C. Apparatus, systems and methods for control of sporting event presentation based on viewer engagement
US20170289624A1 (en) 2016-04-01 2017-10-05 Samsung ElectrĂ´nica da AmazĂ´nia Ltda. Multimodal and real-time method for filtering sensitive media
US20170295215A1 (en) 2016-04-08 2017-10-12 Microsoft Technology Licensing, Llc Audience targeted filtering of content sections
US11082701B2 (en) * 2016-05-27 2021-08-03 Google Llc Methods and devices for dynamic adaptation of encoding bitrate for video streaming
US10506237B1 (en) * 2016-05-27 2019-12-10 Google Llc Methods and devices for dynamic adaptation of encoding bitrate for video streaming
US20220038402A1 (en) 2016-06-30 2022-02-03 Snap Inc. Automated content curation and communication
US20180091856A1 (en) 2016-09-27 2018-03-29 International Business Machines Corporation Video censoring
US20190306563A1 (en) * 2016-12-21 2019-10-03 Arris Enterprises Llc Automatic activation of closed captioning for low volume periods
US20180349502A1 (en) 2017-01-26 2018-12-06 Rena Maycock Data Content Filter
US20190052471A1 (en) 2017-08-10 2019-02-14 Microsoft Technology Licensing, Llc Personalized toxicity shield for multiuser virtual environments
US11205254B2 (en) 2017-08-30 2021-12-21 Pxlize, Llc System and method for identifying and obscuring objectionable content
US20190130189A1 (en) 2017-10-30 2019-05-02 Qualcomm Incorporated Suppressing duplicated bounding boxes from object detection in a video analytics system
US10372991B1 (en) 2018-04-03 2019-08-06 Google Llc Systems and methods that leverage deep learning to selectively store audiovisual content
US11601721B2 (en) 2018-06-04 2023-03-07 JBF Interlude 2009 LTD Interactive video dynamic adaptation and user profiling
US11695978B2 (en) * 2018-07-05 2023-07-04 Mux, Inc. Methods for generating video-and audience-specific encoding ladders with audio and video just-in-time transcoding
US11336968B2 (en) 2018-08-17 2022-05-17 Samsung Electronics Co., Ltd. Method and device for generating content
US20200077150A1 (en) 2018-08-28 2020-03-05 International Business Machines Corporation Filtering Images of Live Stream Content
US10440324B1 (en) 2018-09-06 2019-10-08 Amazon Technologies, Inc. Altering undesirable communication data for communication sessions
US11582420B1 (en) 2018-09-06 2023-02-14 Amazon Technologies, Inc. Altering undesirable communication data for communication sessions
US20200092610A1 (en) 2018-09-19 2020-03-19 International Business Machines Corporation Dynamically providing customized versions of video content
US11748055B2 (en) * 2018-11-16 2023-09-05 Roku, Inc. Monitoring loudness level during media replacement event using shorter time constant
US11407118B1 (en) * 2018-12-10 2022-08-09 Joseph E Augenbraun Robot for performing dextrous tasks and related methods and systems
US10924786B2 (en) * 2019-05-08 2021-02-16 Nanning Fugui Precision Industrial Co., Ltd. Method for shaping video streams and set-up box using the method
US10951902B2 (en) * 2019-06-12 2021-03-16 Rovi Guides, Inc. Systems and methods for multiple bit rate content encoding
US20220007075A1 (en) 2019-06-27 2022-01-06 Apple Inc. Modifying Existing Content Based on Target Audience
US11232686B2 (en) * 2019-12-30 2022-01-25 Axis Ab Real-time deviation in video monitoring
US11609868B1 (en) * 2020-12-31 2023-03-21 Waymo Llc Control calibration timing to avoid memory write blackout period
US20220408131A1 (en) 2021-06-22 2022-12-22 Q Factor Holdings LLC Image analysis system
US20230076702A1 (en) * 2021-08-31 2023-03-09 JBF Interlude 2009 LTD Shader-based dynamic video manipulation
US20230115250A1 (en) * 2021-09-27 2023-04-13 Apple Inc. User interfaces for providing live video

Similar Documents

Publication Publication Date Title
US11538454B2 (en) Systems and methods for presenting social network communications in audible form based on user engagement with a user device
US11863336B2 (en) Dynamic virtual environment
JP6487367B2 (en) Selective content presentation engine
US10826949B2 (en) Distributed control of media content item during webcast
US10586536B2 (en) Display device and operating method therefor
CN106605218B (en) Method for collecting and processing computer user data during interaction with network-based content
US20150070516A1 (en) Automatic Content Filtering
CN112449253B (en) Interactive video generation
US20050223237A1 (en) Emotion controlled system for processing multimedia data
US20230260514A1 (en) Systems and methods for providing voice command recommendations
JP7436460B2 (en) Neural network inference on protected data
WO2018125161A1 (en) Aggregation of media effects
US10645464B2 (en) Eyes free entertainment
US10864447B1 (en) Highlight presentation interface in a game spectating system
CN112423081B (en) Video data processing method, device and equipment and readable storage medium
US20220148614A1 (en) Automatically Captioning Audible Parts of Content on a Computing Device
JP2012039550A (en) Information processing device, information processing system, information processing method and program
CN109299326A (en) Video recommendation method and device, system, electronic equipment and storage medium
US11974012B1 (en) Modifying audio and video content based on user input
US12003821B2 (en) Techniques for enhanced media experience
US12112758B2 (en) Systems and methods for determining traits based on voice analysis
CN109948426A (en) Application program method of adjustment, device, electronic equipment and storage medium
US20190332656A1 (en) Adaptive interactive media method and system
WO2024233129A1 (en) Real-time ai screening and auto-moderation of audio comments in a livestream
US11935557B2 (en) Techniques for detecting and processing domain-specific terminology

Legal Events

Date Code Title Description
FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: MICROENTITY

FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO SMALL (ORIGINAL EVENT CODE: SMAL); ENTITY STATUS OF PATENT OWNER: MICROENTITY

Free format text: ENTITY STATUS SET TO MICRO (ORIGINAL EVENT CODE: MICR); ENTITY STATUS OF PATENT OWNER: MICROENTITY

STCF Information on status: patent grant

Free format text: PATENTED CASE