US8634947B1 - System and method for identifying digital files - Google Patents
System and method for identifying digital files Download PDFInfo
- Publication number
- US8634947B1 US8634947B1 US12/908,045 US90804510A US8634947B1 US 8634947 B1 US8634947 B1 US 8634947B1 US 90804510 A US90804510 A US 90804510A US 8634947 B1 US8634947 B1 US 8634947B1
- Authority
- US
- United States
- Prior art keywords
- digital audio
- song
- name
- audio files
- fingerprint
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related, expires
Links
- 238000000034 method Methods 0.000 title claims abstract description 36
- 238000012545 processing Methods 0.000 claims description 15
- 238000007726 management method Methods 0.000 description 19
- 230000008901 benefit Effects 0.000 description 9
- 230000008569 process Effects 0.000 description 9
- 241000283220 Odobenus rosmarus Species 0.000 description 8
- 238000003860 storage Methods 0.000 description 8
- 238000007792 addition Methods 0.000 description 7
- 238000012986 modification Methods 0.000 description 6
- 230000004048 modification Effects 0.000 description 6
- 238000009826 distribution Methods 0.000 description 5
- 230000015654 memory Effects 0.000 description 5
- 238000004891 communication Methods 0.000 description 4
- 230000006835 compression Effects 0.000 description 4
- 238000007906 compression Methods 0.000 description 4
- 230000008707 rearrangement Effects 0.000 description 4
- 238000006467 substitution reaction Methods 0.000 description 4
- 230000003287 optical effect Effects 0.000 description 3
- 238000013500 data storage Methods 0.000 description 2
- 230000007423 decrease Effects 0.000 description 2
- 238000011161 development Methods 0.000 description 2
- 239000007858 starting material Substances 0.000 description 2
- 230000009471 action Effects 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 238000003491 array Methods 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 238000012937 correction Methods 0.000 description 1
- 238000012217 deletion Methods 0.000 description 1
- 230000037430 deletion Effects 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 230000002441 reversible effect Effects 0.000 description 1
- 238000010845 search algorithm Methods 0.000 description 1
- 235000019640 taste Nutrition 0.000 description 1
- 230000001131 transforming effect Effects 0.000 description 1
- 238000013519 translation Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/60—Information retrieval; Database structures therefor; File system structures therefor of audio data
- G06F16/68—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/686—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using information manually generated, e.g. tags, keywords, comments, title or artist information, time, location or usage information, user ratings
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/60—Information retrieval; Database structures therefor; File system structures therefor of audio data
- G06F16/68—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/683—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
Definitions
- the present disclosure relates generally to digital files, including digital audio or video files. More particularly, the present disclosure relates to the identification of digital files, including the identification of digital audio or video files. Even more specifically, the present disclosure relates to the use of digital fingerprints or metadata to identify digital audio or video files.
- certain of these organizational systems may attempt to identify and catalogue, or otherwise organize, digital audio files which are resident on a user's computer such that a user can access such a system to browse, organize, access or otherwise manage their digital music collection utilizing an interface provided by such a system.
- a user can access such a system to browse, organize, access or otherwise manage their digital music collection utilizing an interface provided by such a system.
- such systems can substantially automatically identify both the song and artist or an associated album for digital audio files stored at the user's computers such that these songs, artists or albums may be utilized to manage the user's collection.
- One method for identifying digital audio files is to utilize a reference database of known digital audio files such that a digital audio file of a user may be associated with one of the known digital audio files and the artist, song or album information associated with the known audio file may be associated with that digital audio file of the user.
- This type of reference database may be difficult to come by and, for a variety of reasons including the addition of new files, it can almost be assured that such a reference database will not include certain digital audio files. Additionally, in many cases users may be charged for the use of such databases.
- a fingerprint of a digital audio file may be utilized to find a set of fingerprints that are related to the fingerprint of the digital audio file it is desired to identify.
- a song, artist or album can be identified and assigned to the digital audio file it was desired to identify.
- the identified song, artist or album name may be associated with the digital audio files corresponding to each of the identified related acoustic fingerprints. In this manner, not only can digital audio files be identified but additionally, data associated with previously identified digital audio files may be iteratively refined by associating updated metadata with previously labeled digital audio files.
- a digital audio file may be associated with a song name utilizing a fingerprint associated with the digital audio file. Metadata including song names associated with one or more digital audio files may be obtained based on the fingerprint and the song names processed to determine a song name to associate with digital audio file. Furthermore, in certain embodiments the song name may also be associated with other digital audio files determined from the fingerprint.
- processing the metadata may comprise grouping the one or more song names into a set of groups, each group comprising the song names that are considered equivalent and determining the most common song name within the group with the highest number of entries.
- Embodiments of the systems and methods disclosed may provide a myriad number of advantages. For example, these systems and method may allow a user to better manage his music collection and provide a better user experience. Additionally, as digital music files may be consistently identified across users, music recommendations and comparisons may be more easily achieved. Furthermore, the consistent identification of music may enable the distribution of data related to the digital audio files from a central system or between and among users. Such data may include, for example, liner notes or image data including album artwork, concert photos or videos, etc. The data may also include a version of the digital audio file itself.
- FIG. 1 depicts one embodiment of a topology including a digital audio management system.
- FIG. 2 depicts one embodiment of a method for identifying digital audio files.
- One embodiment can include one or more computers communicatively coupled to a network.
- the computer can include a central processing unit (“CPU”), at least one read-only memory (“ROM”), at least one random access memory (“RAM”), at least one hard drive (“HD”), and one or more input/output (“I/O”) device(s).
- the I/O devices can include a keyboard, monitor, printer, electronic pointing device (such as a mouse, trackball, stylist, etc.), or the like.
- the computer has access to at least one database over the network.
- ROM, RAM, and HD are computer memories for storing computer instructions executable (in other which can be directly executed or made executable by, for example, compilation, translation, etc.) by the CPU.
- the term “computer-readable medium” is not limited to ROM, RAM, and HD and can include any type of data storage medium that can be read by a processor.
- a computer-readable medium may refer to a data cartridge, a data backup magnetic tape, a floppy diskette, a flash memory drive, an optical data storage drive, a CD-ROM, ROM, RAM, HD, or the like.
- the computer-executable instructions may be stored as software code components or modules on one or more computer readable media (such as non-volatile memories, volatile memories, DASD arrays, magnetic tapes, floppy diskettes, hard drives, optical storage devices, etc. or any other appropriate computer-readable medium or storage device).
- the computer-executable instructions may include lines of complied C++, Java, HTML, or any other programming or scripting code.
- the functions of the disclosed embodiments may be implemented on one computer or shared/distributed among two or more computers in or across a network. Communications between computers implementing embodiments can be accomplished using any electronic, optical, radio frequency signals, or other suitable methods and tools of communication in compliance with known network protocols.
- the terms “comprises,” “comprising,” “includes,” “including,” “has,” “having” or any other variation thereof, are intended to cover a non-exclusive inclusion.
- a process, process, article, or apparatus that comprises a list of elements is not necessarily limited only those elements but may include other elements not expressly listed or inherent to such process, process, article, or apparatus.
- “or” refers to an inclusive or and not to an exclusive or. For example, a condition A or B is satisfied by any one of the following: A is true (or present) and B is false (or not present), A is false (or not present) and B is true (or present), and both A and B are true (or present).
- any examples or illustrations given herein are not to be regarded in any way as restrictions on, limits to, or express definitions of, any term or terms with which they are utilized. Instead, these examples or illustrations are to be regarded as being described with respect to one particular embodiment and as illustrative only. Those of ordinary skill in the art will appreciate that any term or terms with which these examples or illustrations are utilized will encompass other embodiments which may or may not be given therewith or elsewhere in the specification and all such embodiments are intended to be included within the scope of that term or terms. Language designating such nonlimiting examples and illustrations includes, but is not limited to: “for example,” “for instance,” “e.g.”, “in one embodiment.”
- embodiments of the present invention may utilize fingerprints or metadata to identify digital audio files.
- a fingerprint of a digital audio file may be utilized to find a set of fingerprints that are related to the fingerprint of the digital audio file it is desired to identify.
- Using the metadata associated with each of the set of related fingerprints a song, artist or album can be identified and assigned to the digital audio file it was desired to identify.
- the identified song, artist or album name may be associated with the digital audio files corresponding to each of the identified related acoustic fingerprints.
- data associated with previously identified digital audio files may be iteratively refined by associating updated metadata with previously labeled digital audio files.
- Topology 100 comprises a set of entities including digital audio management system 120 (also referred to herein as the AG Entertainment system) which is coupled through network 170 to computing devices 110 (e.g. computer systems, personal data assistants, kiosks, etc.).
- the AG Entertainment system 120 may be executing a digital audio management application 190 (referred to as the AG Entertainment application and including one or more computer readable instructions) which may allow users at computing devices 110 to more easily manage their digital music collections utilizing data stored in data store 122 .
- users at computing devices 110 may have a collection of digital audio files 114 (e.g. mp3, mp4, FLAC, AAC, wav, TAK, TTA, AU, ogg, mpc, ra, divx, mpeg, avi, etc.) which are stored in a data store 112 (which may be any type of local or remote storage on which a user stores data).
- a data store 112 which may be any type of local or remote storage on which a user stores data.
- Each of these digital audio files may have associated metadata such as song title, artist, album, year, track, etc., stored in a metadata container such as that defined by the ID3 format or stored in MP4 Atoms, or other associated metadata which may stored or determined, such as a directory name, file name, file size, play time, etc.
- metadata container such as that defined by the ID3 format or stored in MP4 Atoms, or other associated metadata which may stored or determined, such as a directory name, file name, file size, play time, etc
- the AG Entertainment application 190 may provide just such a capability. More specifically, the AG Entertainment application 190 may obtain data (including metadata) associated with a user's digital music collection 114 using a client application 116 residing on the user's computing device 110 .
- the client application 116 residing on the user's computing device 110 may have a search module 117 that is configured to search the data store 112 associated with the user's computing device 110 for digital audio files; a metadata module 115 that is configured to obtain metadata (for example ID3 tags or other metadata associated with a digital file); and a fingerprint/hash module 119 that is configured to calculate a fingerprint of a digital audio file or one or more hashes associated with a digital audio file as will be described in more detail later.
- the client application 116 may thus be configured to communicate data to the digital audio management system 120 using interface module 192 . Such a communication may be in response to an instruction received by client application 116 from the digital audio management system or may be originated from client application 116 without prompting from the digital audio management system 120 .
- Data 124 related to the digital audio files 114 of the user may then be stored in data store 122 and utilized by the AG Entertainment application 190 to assist a user in managing his digital audio files 114 . While it should be noted that any suitable schema or architecture may be used for the storage of such data, it may be useful here to discuss one particular embodiment of a schema that may be used to store data for use by AG Entertainment application 190 in data store 122 .
- At least a portion of data store 122 may configured as a set of core tables stored in data store 122 .
- This core set of tables may include one or more artist tables, song tables, album tables, fullprints tables, identifiedprints tables or hashtoprint tables.
- Each row of an artist table may comprise an artistID and an artist name
- each row of a song table may comprise a songID, an artistID and a song name
- each row of an album table may comprise an albumID, an artistID and an album name.
- each artist name is unique, while in the songs table the artistID and song name of a row may be unique and in the album tables the artistID and album name of a row may be unique.
- Each row of a fullprints table may comprise a printID, an actual fingerprint taken from a digital audio file and the length in seconds of the digital audio from which the fingerprint was created. Accordingly, a row may be added to this table when a fingerprint is received that is not currently in the fullprints table.
- the printID may be an “auto increment” column, so that when a new row is inserted the printID may be automatically assigned as one greater than the printID of the last row of the table, the previously highest numbered printID, etc.
- each row may comprise a printID, a songID, a songConf (a confidence level associated with the songID), an albumID, an albumConf (a confidence level associated with the albumID).
- a row may be added to this table when a songID or an albumID is selected to be assigned to a fingerprint (as will be discussed in more detail later).
- a hashtoprint table may have rows comprising the hash of a digital audio file and an associated printID. Using this table, then, if a hash is received a matching row in a hashtoprint table may be located if it exists. From this row the associated printID can be obtained and used to look up information in other tables. For example, a printID may used with identifiedprints table to locate information such as songID, albumID, songConf and albumConf. From the songID a song name may be located, from an albumID an album name may be located, etc.
- Each row of a metadata table may comprise a printID, the clientID (associated with a user), the hash of the digital audio file, and the metadata associated with that digital audio file.
- each printID may be associated with multiple rows as different users may have the same (or similar) digital audio files associated with different metadata.
- the hash of the combination of the printID and clientID may be unique such that if one user has multiple copies of the same file, there may be only one entry for that digital audio file in the metadata tables.
- each digital audio file (which for example, has been processed using the client application on the user's computer) in a user's data store 112 there may be an entry in a metadata table, where the clientID corresponds to that user and the metadata of that entry comprises the metadata associated with that digital audio file in that user's data store 122 .
- fingerprints may be represented by a sequence of 892 integers, which is equivalent to 3568 bytes. To reduce storage costs, the fingerprints may be compressed before they are stored. While standard compression algorithms may be utilized they may not be ideal for this type of data. Thus, a fingerprint may be transformed before compressing it with gzip to increase gzip's efficiency.
- the fingerprint may be transformed by transforming the sequence into A, A XOR B, B XOR C, and then compressing it with gzip.
- This may improve the efficiency in that the XOR operation (exclusive OR) returns a number with a small number of bits set when performed on two similar numbers. This decreases the number of unique numbers in the sequence, which increases the gzip compression efficiency.
- the operation is reversible, so after uncompressing the data, the original sequence can be restored. Utilizing such an algorithm, the average length of a stored fingerprint may be around is 976 bytes, about 27% of the raw size.
- certain tables may be sharded (for example, into 64 or a different number of separate tables based on one or more criteria).
- the fullprint tables may be sharded by play time. All the fingerprints in a given table may correspond to digital audio files having the same range of times (for example, one table may comprise fingerprints associated with digital audio files having a play time from 300 seconds to 310 seconds, another table may comprise fingerprints associated with digital audio files having a play time from 311 seconds to 320 seconds, etc.).
- This configuration may be used advantageously with certain search algorithms or for certain types of searches.
- such a configuration may be useful when it is desired to search for fingerprints with a similar time to a particular fingerprint (in other words for fingerprints associated with digital audio files having a similar play time to the digital audio file associated with the particular fingerprint). So to find the fingerprint with a corresponding playtime that is within 10 seconds of a particular playtime, only tables containing fingerprints within 10 seconds of the particular playtime would need to be searched.
- printIDs are 8 byte integers generated by combining the table ID of a fingerprint and a sequence number.
- the print table ID may be stored in the most significant byte of the 8 byte ID. So for example, the first 2 printIDs in the second print table have the following IDs in hexadecimal: 200000000000000 and 200000000000001.
- Data store 122 may also comprise one or more user tables 126 .
- user tables 126 there may be one or more user tables corresponding to each user. These user tables may comprise one or more artist tables, song tables and album tables that comprise data associated only with the corresponding user.
- each user may have a set of tables that comprise the information utilized by the user's to identify his collections and the user's files can be managed specifically to each user using the tables corresponding to that user.
- a user may manage his digital music collection 114 .
- the management functionality of the AG Entertainment system 120 may be implemented, at least in part, using data 124 corresponding to the user's digital audio files 114 stored in the data store 122 or data 118 associated with the client application 116 stored in the user's data store 112 .
- management of a user's digital audio files may be accomplished using user tables corresponding to that user such that data specific to the user may be utilized to manage the user's digital audio files.
- the number of digital audio files 114 may be quite large and many of these files may be associated with incorrect identifiers.
- a user may wish to upload data on each of these digital audio files.
- This process may be time intensive and the prospect of such a task may dissuade a user from using the management functionality offered by the AG Entertainment system 120 .
- the AG Entertainment application 190 may comprise a song identification module 196 operable to identify digital audio files 114 of a user.
- the song identification module may utilize data 124 stored in data store 122 , where this data may comprise data obtained from a user's computing device 110 using the client application 116 , data determined by AG Entertainment application 190 from this obtained data or from other data, data obtained from a third party data source, data determined from third party data or any other type of data desired.
- This data as stored in the core tables of in the data store 122 can then be used to manage the user's digital music collection.
- a user may desire to play or otherwise manipulate his digital music collection.
- the digital audio management system 120 may provide an interface such as a web site or a mobile application (for example, a mobile phone application) with which the user can interact on utilizing the user's computing device 110 .
- the user may provide, or the digital audio management system may otherwise obtain, a clientID associated with the user.
- This clientID may be, for example, associated with a user's login credentials or identification.
- the digital audio management system 120 may obtain data on the user's digital music files from the data in the data store 122 , for example using the core tables of data 124 or the user tables 126 . Using this data a list of the user's digital music files may be presented to the user through the web site. This list may comprise, for example, the song names, artist names, album names, etc. of the digital music files stored in user's data store 112 . Each item of the list may be associated with a digital music file in the user's data store 112 , for example using a unique identifier.
- each item of the presented list may be a link associated with a URL comprising a unique identifier associated with the digital music file on the user's data store 112 .
- the link may be accessed. This link may be used to establish a relay between the digital audio management system 120 and the user's computer 110 such that the digital audio file on the user's data store 112 associated with the unique identifier may be played using the interface.
- FIG. 2 A flow diagram for one embodiment of method for identifying a digital audio file is depicted in FIG. 2 .
- a user may wish to manage a digital audio file (that may be referred to as the current digital audio file in the foregoing description of the embodiment of FIG. 2 ) on his data store 112 using the AG Entertainment system 120 .
- This situation may occur, for example, when a user initially registers with the AG Entertainment system 120 and installs the client application 116 on his system 110 , when the user adds another digital audio file 114 to his collection at a later point, when the user plays a particular digital audio file using client application 116 , or for some other reason entirely.
- the client application 116 on the user's computing device 110 may send data on the digital audio file to the AG Entertainment application 190 .
- the client application 116 transmits a hash of the digital audio file (for example, an md5 hash or the like) to the AG Entertainment application 190 which provides this data to the song identification module 196 .
- this hash may be on the order of 100 bytes or so.
- certain techniques may be utilized to increase the efficiency of the communication of this data. For example, hashes for multiple digital audio files may be batched and sent in single or multiple transmissions.
- step 212 it can be checked if the digital audio file associated with the received hash has been identified before by comparing the received hash value with hash values for digital audio files which have previously been registered. In one embodiment, such previous identification may have occurred if another user has registered the exact same file using a client application 116 . More specifically, the received hash may be compared to hashes stored in the data store 122 to determine if a match can be found. For example, the received hash may be compared to hashes stored in the hashtoprint table.
- the hash may be associated with a printID, where the printID may be associated with a songID and an albumID.
- the songID and the albumID may, in turn, be associated with a confidence level, as described with respect to the identifiedprints table discussed above.
- additional data associated with the digital audio file may be requested from the client application at step 217 .
- data including metadata
- data associated with the digital audio file (for example, the file name, file size, play time in seconds, and any id3 tags) may be requested from the client application 116 .
- This data, along with the previously received fingerprint and the printID associated with the previously received fingerprint will then be utilized as described beginning with step 226 below.
- a fingerprint associated with the digital audio file may be requested from the client application 116 .
- the client application 116 responds with data (including metadata) associated with the digital audio file (for example, the file name, file size, play time in seconds, and any id3 tags) and an acoustic fingerprint (also referred to as a fingerprint or a print and which refers to an algorithmically determined integer created from at least a portion of a digital file, regardless of whether the digital file is an audio or other type of digital file) of the digital audio file.
- a fingerprint comprises 892 integers, with each integer being 4 bytes in size.
- Fingerprints may be generated such that similar digital files will generate fingerprints that have small hamming distances, which is the number of bits that would need to be changed to convert one fingerprint into another.
- One embodiment of a fingerprinting algorithm that may be utilized is discussed in “Computer Vision for Music Identification” by Yan Ke et al published by the School of Computer Science at Carnegie Mellon, which is hereby fully incorporated by reference and that is documented on the website: http://www.yanke.org/research.htm. That particular paper describes a fingerprinting algorithm that uses 32 particular “filters” chosen from a much larger set to analyze audio for the purpose of matching audio samples after they have been played over a low-quality speaker and recorded using a low-quality microphone.
- the performance of the fingerprinting algorithm for grouping audio samples from the clients can be improved by using the techniques described in the paper to find the 32 filters that are best for identifying similar tracks after they have been encoded using different compression schemes (such as MP3 or AAC)
- client application 116 may select a particular range of a digital audio file to generate the fingerprint. While a 12 second portion of the digital audio file may be utilized and may result in desired lookup or search properties, longer songs tend to have less unique content at the beginning of the track. To address this, the offset into a song from where the 12 second portion begins may be increased based on the length of the play time of the digital audio file. Very short songs may utilize a 12 second portion beginning with an offset of 0 seconds while the digital audio files with the longest play times may utilize a 12 second portion beginning with an offset of 5 minutes.
- the length of silent regions at the beginning and end of the digital audio file may be subtracted from the overall play time of track. Silent regions are determined by looking at the sum of the audio power (a sum of squares of pulse code modulation (PCM) samples) over small, shifted ranges. Once the sum exceeds an empirically determined threshold, the silent region is considered finished.
- PCM pulse code modulation
- a match for the received fingerprint can be found. This matching may be performed by determining if an identical fingerprint associated with a substantially identical play time (as defined in the received metadata) can be located. As discussed above, the tables into which the fingerprint may be saved may be sharded by play time. Thus, to attempt to locate such a fingerprint the play time may be utilized to determine which table to search for the received fingerprint. Once this table is located a hash of the fingerprint may be calculated (for example, a hash comprising 4 byte integer or the like) to attempt to lookup one or more fingerprints in the table to compare to the received fingerprint. If any fingerprints are found based on the hash comparison, the received fingerprint may then be compared to the one or more fingerprints located through the hash comparison to see if the received fingerprint matches any of the one or more fingerprints located.
- a hash of the fingerprint may be calculated (for example, a hash comprising 4 byte integer or the like) to attempt to lookup one or more fingerprints in the table to compare to the received fingerprint. If any
- the fullprints table may be searched using a standard inverted index full text search, treating each unique integer (also called a key) in the fingerprint as a word.
- indexing all of the integers may be prohibitively expensive, so an optimization referred to as min-hashing may be applied in certain embodiments.
- min-hashing a subset of the keys of each fingerprint in the table may be selected and indexed.
- a search is then run against the index after using the same min-hashing algorithm to identify which keys in the fingerprint from the client should be used to search the index.
- the received fingerprint can be saved into data store 122 and assigned a unique identifier, known as a printID (for example, by using the auto-increment value assigned after creating a new row in the fullprints table).
- This assigned printID may also be associated with the originally received hash of the digital audio file at this point (for example, by creating a new row in the hashtoprint table).
- Such an identifier may, for example, be on the order of 8 bytes or so.
- the printID of the found match is retrieved at step 224 .
- the printID (assigned or retrieved) can be sent back to the client application 116 and stored on the data store 112 of that user's computer in association with the digital audio file from which the fingerprint was created such that generating the fingerprint (a relatively computationally expensive operation) for that digital audio file can be avoided in the future.
- the metadata received from the client application 116 may be saved and associated with the printID or a clientID corresponding to the user or user's computer from which the metadata was received, such that the metadata associated with a printID can be easily accessed and the information that may be utilized by AG Entertainment system 190 increased.
- This saving may be done, for example, by adding a row to the metadata, identifiedprints, artist, song or album tables.
- only one copy of metadata per hash per user may be saved so that a single client does not add multiple copies of metadata to the database if they register several copies of the same digital audio file.
- a set of related fingerprints (and printIDs and metadata associated with those printIDs) may be determined.
- a number of tables that are to be searched may be determined and similar to the search described above, an inverted index full text search with min-hashing may be used to locate a set of related fingerprints.
- This set of related fingerprints may be those fingerprints which were generated from digital audio files having a similar playing time (within some threshold time) to the digital audio file that is being identified.
- This set of related fingerprints may be further refined by determining which of this set of related fingerprints are within a certain “shifted hamming distance” of the fingerprint for the current digital audio file.
- Shifted hamming distances may be utilized instead of simple hamming distances, because many digital audio files have a slightly different start point. By checking for the shortest distance after shifting the fingerprints forward and back a number of bits (for example, by up to two hundred), better matches for similar digital audio files may be generated. It should be noted that simple or other types of hamming distances may also be utilized or the set of fingerprints may not be refined based on hamming distances.
- a set of related fingerprints for digital audio files that are acoustically similar to the original digital audio file have been determined.
- This set of related fingerprints and printIDs may, however, not be associated with the same song name, artist name or album name as the digital audio file associated with the current fingerprint.
- the song names, album names or artists associated with the set of related fingerprints may be grouped. More specifically, in one embodiment the metadata associated with each of the related fingerprints may be obtained. For example, the printID associated with each of the related fingerprints may be used to access the metadata tables and each row in the metadata tables that correspond to these printIDs may be accessed. For each printID then, there may be one or more associated song names, artist names and album names.
- the song name and artist name of the metadata may utilized to determine which combination of song name and artist name (e.g. “Rattled By the Rush” and “Pavement”) should be grouped together. Similarly, it can be determined which combination of album name and artist name (e.g. “Wowee Zowee” and “Pavement”) should be grouped together.
- loose matching rules may be utilized. Such algorithms for text matching may be known in the art and may be utilized to account for spelling mistakes, capitalization, the presence or lack thereof of articles, white space, etc. (for example, “The Clash” and “Clash” would be considered equivalent).
- each of these groups comprises all occurrences of a song name and artist name combination that occurs in the metadata associated with the related fingerprints that are to be considered equivalent.
- groups for combinations of album name and artist name Each of these groups comprises all occurrences of an album name and artist name combination that occurs in the metadata associated with the related fingerprints that are to be considered equivalent
- one or more associated confidence levels may be generated. More specifically, a confidence level may be generated for a combination of song name and artist name by counting the number of occurrences of the combination of song name and artist name that occurs most often and the number of occurrences of the combination of song name and artist name that occurs next most often and subtracting the two. In other words, the grouping of song name and artist name combinations with the largest number of entries and the second largest number of entries may be subtracted from one another.
- a confidence level may be generated for a combination of album name and artist name by counting the number of occurrences of the pair of album name and artist name that occur most often and the number of occurrences of the pair of album name and artist name that occurs next most often and subtracting the two.
- the grouping of album name and artist name combinations with the largest number of entries and the second largest number of entries may be subtracted from one another. For example, if the combination of “The Queen is Dead” and “The Smiths” occurs most often and occurs 7 times and the combination of “The Queen” and “Smiths” occurs 3 times the confidence level for the combination of “The Queen is Dead” and “The Smiths” may be 4.
- the thresholds to be applied to the confidence level for the combination of the song name and artist name may be different or the same as the threshold to be applied to the confidence level for the combination of the album name and artist name and that none, some or all of the data may be associated with a current fingerprint based on the outcome of this comparison. If both the confidence levels are below their respective thresholds, at step 236 the current fingerprint may be assigned a songID of zero and no further action may be taken. Alternatively, the printID associated with the received fingerprint and the digital audio file may be sent to the client application 116 where it is stored to allow the client to rapidly transmit information about the digital audio files 114 in the future as is discussed later.
- each grouping of combination of song name and artist name may comprise occurrences of song name and artist name that may or may not be exactly the same (for example, the metadata associated with each of these related fingerprints may not be the same) but are considered equivalent.
- the combination of song name and artist name that occurs most frequently within the grouping of song names and artist names with the highest confidence or the highest number of entries may be determined.
- the combination of album name and artist name that occurs most frequently within the grouping of album names and artist names with the highest confidence or the highest number of entries may also be determined.
- a “corrections” database may also be utilized to check to see if any of the variations of song name, artist name, album name, etc. occurring within a grouping are known misspellings or otherwise common mistakes such that these mistakes can be corrected for purposes of this determination. Using these corrected strings, then, the most common song name and artist name or album name and artist name may be selected from the grouping.
- Data may then be associated with the digital audio file at step 240 .
- the most commonly occurring combination of song name and artist name within the grouping of combination of song names and artist names with the highest confidence or more occurrences it can be attempted to locate the artist name (e.g. in the artist table) and obtain the artistID associated with the artist name. If no such artist name exists, an artistID may be created and associated with artist name. Using this artistID it can then be checked if the combination of the artistID and the song name exists (e.g. in the songs table). If the combination does not exist, the combination may be stored (e.g. by creating an entry in the songs table with a new songID), otherwise the songID corresponding to the combination is obtained.
- Associations may then be created for the identified current fingerprint associated with the digital audio file. More specifically, an entry in the identified prints table may be created using the printID assigned to the current fingerprint, the songID associated with the fingerprint, the confidence level generated for the song name, the albumID associated with the fingerprint and the confidence level generated for the album name.
- the songID associated with the current fingerprint may be associated with all of the printIDs corresponding to the related set of fingerprints and metadata determined at step 228 . For example, by altering entries for these related fingerprints in the identifiedprints table. If those fingerprints previously had a songID associated with them, their songID is changed and the previous songID is flagged for deletion if it has no other fingerprints associated with it.
- the songID or printID associated with the current fingerprint and the digital audio file is then sent to the client application 116 at step 242 where it may be stored to allow the client to rapidly transmit information about the digital audio files 114 in the future.
- a fingerprint may not have an associated songID if none of the refined set of metadata was of sufficiently high quality to identify the digital audio file when it was analyzed at step 232 .
- the original fingerprint may be assigned a songID based on an analysis of the newly received fingerprint (as discussed above).
- the original client application 116 comes online later, it will send the printID for the unidentified original fingerprint to the AG Entertainment Application 190 , which can then check if that printID has been assigned a song and send the associated songID back to the original client application 116 .
- the information in the data store corresponding to digital audio files 114 at user's computing systems 110 can be continually improved.
- a client may come online with incorrect metadata for a digital audio file, but once other clients come online with the correct, or more correct metadata, the original song identification will be deleted and the original fingerprint will be associated with the new song.
- embodiments of this invention eliminate the need for a master set of “known” fingerprints or an existing database of song and artist names. As clients send in more metadata and fingerprints prints, the database at AG Entertainment system 120 both grows and corrects the associated song and artist names.
- Print 1 has the following metadata (e.g. in a metadata table) associated with 3 different users (e.g. clientIDs):
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Library & Information Science (AREA)
- Multimedia (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
Description
-
- “I am the walrus” by “The Beatles”
- “Track 06” by “Unknown Artist”
- “I am the walrus” by “The Beatles!”
And Print 2 has the following metadata, from 1 user: - “I am the walrus” by “The Beatles”
-
- “I am the walrus” by “The Beatles”
- “I am the walrus” by “The Beatles!”
- “I am the walrus” by “The Beatles”
Group B: - “Track 06” by “Unknown Artist”
Here, Group A may be selected as the grouping with the most occurrences of the song name and artist name (3 occurrences versus 1 occurrence with respect to group B). The confidence level can then be calculated for Group A as 2, (the 1 occurrence of Group B subtracted from the 3 occurrences of Group A). Next, the most commonly occurring song name and artist name combination may be selected from within Group A. Here, “I am the walrus” by “The Beatles” may be selected since it occurs the most often (2 times). Thus, the song name and album name “I am the walrus” with confidence level 2 may be associated with the PrintID 2 (for example, in the identifiedprints table).
Claims (21)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US12/908,045 US8634947B1 (en) | 2009-10-21 | 2010-10-20 | System and method for identifying digital files |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US25355709P | 2009-10-21 | 2009-10-21 | |
US12/908,045 US8634947B1 (en) | 2009-10-21 | 2010-10-20 | System and method for identifying digital files |
Publications (1)
Publication Number | Publication Date |
---|---|
US8634947B1 true US8634947B1 (en) | 2014-01-21 |
Family
ID=49919294
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/908,045 Expired - Fee Related US8634947B1 (en) | 2009-10-21 | 2010-10-20 | System and method for identifying digital files |
Country Status (1)
Country | Link |
---|---|
US (1) | US8634947B1 (en) |
Cited By (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140032537A1 (en) * | 2012-07-30 | 2014-01-30 | Ajay Shekhawat | Apparatus, system, and method for music identification |
US20140032550A1 (en) * | 2012-07-25 | 2014-01-30 | Samsung Electronics Co., Ltd. | Method for managing data and an electronic device thereof |
US20140188835A1 (en) * | 2012-12-31 | 2014-07-03 | Via Technologies, Inc. | Search method, search system, and natural language comprehension system |
US20150213029A1 (en) * | 2013-03-13 | 2015-07-30 | Google Inc. | Using cross-matching between users and matching against reference data to facilitate content identification |
US9264519B1 (en) * | 2011-08-20 | 2016-02-16 | Datastax, Inc. | Embedding application services in a distributed datastore |
US9619560B1 (en) * | 2012-05-29 | 2017-04-11 | Soundhound, Inc. | System and methods for offline audio recognition |
US20170371963A1 (en) * | 2016-06-27 | 2017-12-28 | Facebook, Inc. | Systems and methods for identifying matching content |
CN107534420A (en) * | 2015-03-03 | 2018-01-02 | 欧朋德私人有限公司 | A kind of system, Edition Contains server, audio recording slave equipment and Edition Contains interface for the recording of distributable field performance predetermined audio, the audio content editor based on cloud and the distribution of the online content of audio track and associated metadata |
US10121165B1 (en) | 2011-05-10 | 2018-11-06 | Soundhound, Inc. | System and method for targeting content based on identified audio and multimedia |
CN110503961A (en) * | 2019-09-03 | 2019-11-26 | 北京字节跳动网络技术有限公司 | Audio identification methods, device, storage medium and electronic equipment |
CN110674337A (en) * | 2019-10-15 | 2020-01-10 | 盛唐威讯数媒科技(北京)有限公司 | Audio-video image-text recognition system |
US10963507B1 (en) * | 2020-09-01 | 2021-03-30 | Symphonic Distribution Inc. | System and method for music metadata reconstruction and audio fingerprint matching |
US11062693B1 (en) * | 2019-06-20 | 2021-07-13 | West Corporation | Silence calculator |
EP3855427A4 (en) * | 2018-09-18 | 2022-05-04 | AlphaTheta Corporation | Musical piece data comparison device, musical piece analysis data distribution server, musical piece data comparison program, and musical piece analysis data distribution program |
EP4137969A1 (en) * | 2021-08-16 | 2023-02-22 | Utopia Music AG | Apparatus, method and computer program code for processing an audio stream |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020049760A1 (en) * | 2000-06-16 | 2002-04-25 | Flycode, Inc. | Technique for accessing information in a peer-to-peer network |
US20040034650A1 (en) * | 2002-08-15 | 2004-02-19 | Microsoft Corporation | Media identifier registry |
US20050004941A1 (en) * | 2001-11-16 | 2005-01-06 | Maria Kalker Antonius Adrianus Cornelis | Fingerprint database updating method, client and server |
US20090164564A1 (en) * | 2005-07-01 | 2009-06-25 | Michael John Willis | System and methods for mobilizing web content |
US20090259690A1 (en) * | 2004-12-30 | 2009-10-15 | All Media Guide, Llc | Methods and apparatus for audio recognitiion |
-
2010
- 2010-10-20 US US12/908,045 patent/US8634947B1/en not_active Expired - Fee Related
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020049760A1 (en) * | 2000-06-16 | 2002-04-25 | Flycode, Inc. | Technique for accessing information in a peer-to-peer network |
US20050004941A1 (en) * | 2001-11-16 | 2005-01-06 | Maria Kalker Antonius Adrianus Cornelis | Fingerprint database updating method, client and server |
US20040034650A1 (en) * | 2002-08-15 | 2004-02-19 | Microsoft Corporation | Media identifier registry |
US20090259690A1 (en) * | 2004-12-30 | 2009-10-15 | All Media Guide, Llc | Methods and apparatus for audio recognitiion |
US20090164564A1 (en) * | 2005-07-01 | 2009-06-25 | Michael John Willis | System and methods for mobilizing web content |
Cited By (31)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10832287B2 (en) | 2011-05-10 | 2020-11-10 | Soundhound, Inc. | Promotional content targeting based on recognized audio |
US10121165B1 (en) | 2011-05-10 | 2018-11-06 | Soundhound, Inc. | System and method for targeting content based on identified audio and multimedia |
US12100023B2 (en) | 2011-05-10 | 2024-09-24 | Soundhound Ai Ip, Llc | Query-specific targeted ad delivery |
US9264519B1 (en) * | 2011-08-20 | 2016-02-16 | Datastax, Inc. | Embedding application services in a distributed datastore |
US9619560B1 (en) * | 2012-05-29 | 2017-04-11 | Soundhound, Inc. | System and methods for offline audio recognition |
US20140032550A1 (en) * | 2012-07-25 | 2014-01-30 | Samsung Electronics Co., Ltd. | Method for managing data and an electronic device thereof |
US9483507B2 (en) * | 2012-07-25 | 2016-11-01 | Samsung Electronics Co., Ltd. | Method for managing data and an electronic device thereof |
US20140032537A1 (en) * | 2012-07-30 | 2014-01-30 | Ajay Shekhawat | Apparatus, system, and method for music identification |
US9646000B2 (en) * | 2012-12-31 | 2017-05-09 | Via Technologies, Inc. | Search method, search system, and natural language comprehension system |
US9715524B1 (en) * | 2012-12-31 | 2017-07-25 | Via Technologies, Inc. | Natural language comprehension system |
US20140188835A1 (en) * | 2012-12-31 | 2014-07-03 | Via Technologies, Inc. | Search method, search system, and natural language comprehension system |
US20170177571A1 (en) * | 2013-03-13 | 2017-06-22 | Google Inc. | Using cross-matching between users and matching against reference data to facilitate content identification |
US9589027B2 (en) * | 2013-03-13 | 2017-03-07 | Google Inc. | Using cross-matching between users and matching against reference data to facilitate content identification |
US20230385326A1 (en) * | 2013-03-13 | 2023-11-30 | Google Llc | Using cross-matching between users and matching against reference data to facilitate content identification |
US20150213029A1 (en) * | 2013-03-13 | 2015-07-30 | Google Inc. | Using cross-matching between users and matching against reference data to facilitate content identification |
US10162824B2 (en) * | 2013-03-13 | 2018-12-25 | Google Llc | Using cross-matching between users and matching against reference data to facilitate content identification |
US11727045B2 (en) | 2013-03-13 | 2023-08-15 | Google Llc | Using cross-matching between users and matching against reference data to facilitate content identification |
US10990621B2 (en) | 2013-03-13 | 2021-04-27 | Google Llc | Using cross-matching between users and matching against reference data to facilitate content identification |
EP3266104A4 (en) * | 2015-03-03 | 2018-12-19 | Openhd Pty Ltd | A system, content editing server, audio recording slave device and content editing interface for distributed live performance scheduled audio recording, cloud-based audio content editing and online content distribution of audio track and associated metadata |
CN107534420A (en) * | 2015-03-03 | 2018-01-02 | 欧朋德私人有限公司 | A kind of system, Edition Contains server, audio recording slave equipment and Edition Contains interface for the recording of distributable field performance predetermined audio, the audio content editor based on cloud and the distribution of the online content of audio track and associated metadata |
US11030462B2 (en) | 2016-06-27 | 2021-06-08 | Facebook, Inc. | Systems and methods for storing content |
US20170371963A1 (en) * | 2016-06-27 | 2017-12-28 | Facebook, Inc. | Systems and methods for identifying matching content |
EP3855427A4 (en) * | 2018-09-18 | 2022-05-04 | AlphaTheta Corporation | Musical piece data comparison device, musical piece analysis data distribution server, musical piece data comparison program, and musical piece analysis data distribution program |
US11062693B1 (en) * | 2019-06-20 | 2021-07-13 | West Corporation | Silence calculator |
CN110503961B (en) * | 2019-09-03 | 2023-03-14 | 北京字节跳动网络技术有限公司 | Audio recognition method and device, storage medium and electronic equipment |
CN110503961A (en) * | 2019-09-03 | 2019-11-26 | 北京字节跳动网络技术有限公司 | Audio identification methods, device, storage medium and electronic equipment |
CN110674337A (en) * | 2019-10-15 | 2020-01-10 | 盛唐威讯数媒科技(北京)有限公司 | Audio-video image-text recognition system |
US10963507B1 (en) * | 2020-09-01 | 2021-03-30 | Symphonic Distribution Inc. | System and method for music metadata reconstruction and audio fingerprint matching |
US11068535B1 (en) | 2020-09-01 | 2021-07-20 | Symphonic Distribution Inc. | System and method for reconstructing music catalogs |
EP4137969A1 (en) * | 2021-08-16 | 2023-02-22 | Utopia Music AG | Apparatus, method and computer program code for processing an audio stream |
WO2023020999A1 (en) * | 2021-08-16 | 2023-02-23 | Utopia Music Ag | Apparatus, method and computer program code for processing an audio stream |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8634947B1 (en) | System and method for identifying digital files | |
US9542488B2 (en) | Associating audio tracks with video content | |
US8521759B2 (en) | Text-based fuzzy search | |
KR100678921B1 (en) | Method and apparatus for synchronizing multimedia contents with device which supports plural server environment | |
US8910044B1 (en) | Playlist incorporating tags | |
US9014832B2 (en) | Augmenting media content in a media sharing group | |
US9344759B2 (en) | Associating audio tracks of an album with video content | |
JP5005726B2 (en) | Managing media files from multiple sources | |
US8126997B2 (en) | Hot data management method based on hit counter | |
JP2007293855A5 (en) | ||
WO2020108441A1 (en) | Data storage method, device, computer apparatus, and readable storage medium | |
US9300986B2 (en) | Media system with canonical architecture for integrating media productions from different content providers | |
JP2009536418A5 (en) | ||
US20080065699A1 (en) | System and method for collection and synchronization of metadata to facilitate distributed content search | |
US9558272B2 (en) | Method of and a system for matching audio tracks using chromaprints with a fast candidate selection routine | |
US20130067237A1 (en) | Providing random access to archives with block maps | |
WO2021043088A1 (en) | File query method and device, and computer device and storage medium | |
US9881083B2 (en) | Method of and a system for indexing audio tracks using chromaprints | |
KR101713988B1 (en) | Method and apparatus for providing content sending metadata extracted from content | |
CN104123309A (en) | Method and system used for data management | |
EP3059706B1 (en) | Music on video site | |
US20140215544A1 (en) | Music on video site | |
CN103258052A (en) | Method for discovering related resources on eMule network | |
CN103870476A (en) | Retrieval method and device | |
CN113779025B (en) | Optimization method, system and application of classified data retrieval efficiency in block chain |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: AG ENTERTAINMENT, INC., WASHINGTON Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KLEINPETER, THOMAS;MERHEJ, MICHAEL;MODY, VIRAJ;REEL/FRAME:025604/0857 Effective date: 20101208 |
|
AS | Assignment |
Owner name: MERHEJ, MICHAEL, WASHINGTON Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:AG ENTERTAINMENT, INC.;REEL/FRAME:031663/0146 Effective date: 20131024 Owner name: KLEINPETER, THOMAS, WASHINGTON Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:AG ENTERTAINMENT, INC.;REEL/FRAME:031663/0146 Effective date: 20131024 |
|
FEPP | Fee payment procedure |
Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.) |
|
LAPS | Lapse for failure to pay maintenance fees |
Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.) |
|
STCH | Information on status: patent discontinuation |
Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362 |
|
FP | Lapsed due to failure to pay maintenance fee |
Effective date: 20180121 |