Patent application number | Description | Published |
20090070095 | MINING BILINGUAL DICTIONARIES FROM MONOLINGUAL WEB PAGES - Systems and methods for identifying translation pairs from web pages are provided. One disclosed method includes receiving monolingual web page data of a source language, and processing the web page data by detecting the occurrence of a predefined pattern in the web page data, and extracting a plurality of translation pair candidates. Each of the translation pair candidates may include a source language string and target language string. The method may further include determining whether each translation pair candidate is a valid transliteration. The method may also include, for each translation pair that is determined not to be a valid transliteration, determining whether each translation pair candidate is a valid translation. The method may further include adding each translation pair that is determined to be a valid translation or transliteration to a dictionary. | 03-12-2009 |
20090106173 | LIMITED-MEMORY QUASI-NEWTON OPTIMIZATION ALGORITHM FOR L1-REGULARIZED OBJECTIVES - An algorithm that employs modified methods developed for optimizing differential functions but which can also handle the special non-differentiabilities that occur with the L | 04-23-2009 |
20090125501 | RANKER SELECTION FOR STATISTICAL NATURAL LANGUAGE PROCESSING - Systems and methods for selecting a ranker for statistical natural language processing are provided. One disclosed system includes a computer program configured to be executed on a computing device, the computer program comprising a data store including reference performance data for a plurality of candidate rankers, the reference performance data being calculated based on a processing of test data by each of the plurality of candidate rankers. The system may further include a ranker selector configured to receive a statistical natural language processing task and a performance target, and determine a selected ranker from the plurality of candidate rankers based on the statistical natural language processing task, the performance target, and the reference performance data. | 05-14-2009 |
20090240486 | HMM ALIGNMENT FOR COMBINING TRANSLATION SYSTEMS - A computing system configured to produce an optimized translation hypothesis of text input into the computing system. The computing system includes a plurality of translation machines. Each of the translation machines is configured to produce their own translation hypothesis from the same text. An optimization machine is connected to the plurality of translation machines. The optimization machine is configured to receive the translation hypotheses from the translation machines. The optimization machine is further configured to align, word-to-word, the hypotheses in the plurality of hypotheses by using a hidden Markov model. | 09-24-2009 |
20090276414 | RANKING MODEL ADAPTATION FOR SEARCHING - Search results provided by a search engine (e.g., for the Internet) are improved and/or made more accurate by addressing the limited availability of human labeled training data for certain domains (e.g., languages other than English, within certain date ranges, corresponding to queries over a certain length, etc.). More particularly, a ranking model trained on in-domain data, for which a small amount of human labeled training data (e.g., query/URL pairs) is available (e.g., languages other than English) is adjusted based upon out-domain data, for which a large amount of human labeled training data (e.g., query/URL pairs) is available (e.g., English). Thus, even though the resulting adapted in-domain ranking model is used in the context of in-domain data (e.g., non-English) to provide search results, the search results are improved because they are influenced by an abundance of, albeit out-domain, human labeled training data. | 11-05-2009 |
20090326916 | UNSUPERVISED CHINESE WORD SEGMENTATION FOR STATISTICAL MACHINE TRANSLATION - Described is using a generative model in processing an unsegmented sentence into a segmented sentence. A segmenter includes the generative model, which given an unsegmented sentence (e.g., in Chinese) provides candidate segmented sentences to a probability-based decoder that selects the segmented sentence. For example, the segmented (e.g., Chinese-language) sentence may be provided to a statistical machine translator that outputs a translated (e.g., English-language) sentence. The generative model may include a word sub-model that generates hidden words using a word model, a spelling sub-model that generates characters from the hidden words, and an alignment sub-model that generates translated words and alignment data from the characters. The word sub-model may correspond to a unigram model having words and associated frequency data therein, and the alignment sub-model may correspond to a word aligned corpus having source sentence, translated target sentence pairings therein. Training is also described. | 12-31-2009 |
20100082510 | TRAINING A SEARCH RESULT RANKER WITH AUTOMATICALLY-GENERATED SAMPLES - A search result ranker may be trained with automatically-generated samples. In an example embodiment, user interests are inferred from user interactions with search results for a particular query so as to determine respective relevance scores associated with respective query-identifier pairs of the search results. Query-identifier-relevance score triplets are formulated from the respective relevance scores associated with the respective query-identifier pairs. The query-identifier-relevance score triplets are submitted as training samples to a search result ranker. The search result ranker is trained as a learning machine with multiple training samples of the query-identifier-relevance score triplets. | 04-01-2010 |
20100082582 | COMBINING LOG-BASED RANKERS AND DOCUMENT-BASED RANKERS FOR SEARCHING - Log-based rankers and document-based rankers may be combined for searching. In an example embodiment, there is a method for combining rankers to perform a search operation. A count of query instances in log data is ascertained based on a query. A search for the query is performed to produce a set of search results. The set of search results is ranked by relevance score with a document-based ranker and a log-based ranker using a weighting factor that is adapted responsive to the count of the query instances in the log data. | 04-01-2010 |
20100318531 | SMOOTHING CLICKTHROUGH DATA FOR WEB SEARCH RANKING - Described is a technology for using clickthrough data (e.g., based on data of a query log) in learning a ranking model that may be used in online ranking of search results. Clickthrough data, which is typically sparse (because many documents are often not clicked or rarely clicked), is processed/smoothed into smoothed clickthrough streams. The processing includes determining similar queries for a document with incomplete (insufficient) clickthrough data to provide expanded clickthrough data for that document, and/or by estimating at least one clickthrough feature for a document when that document has missing (e.g., no) clickthrough data. Similar queries may be determined by random walk clustering and/or session-based query analysis. Features extracted from the clickthrough streams may be used to provide a ranking model which may then be used in online ranking of documents that are located with respect to a query. | 12-16-2010 |
20110208735 | Learning Term Weights from the Query Click Field for Web Search - Described is a technology by which a term frequency function for web click data is machine learned from raw click features extracted from a query log or the like and training data. Also described is using combining the term frequency function with other functions/click features to learn a relevance function for use in ranking document relevance to a query. | 08-25-2011 |
20110295897 | QUERY CORRECTION PROBABILITY BASED ON QUERY-CORRECTION PAIRS - Query-correction pairs can be extracted from search log data. Each query-correction pair can include an original query and a follow-up query, where the follow-up query meets one or more criteria for being identified as a correction of the original query, such as an indication of user input indicating the follow-up query is a correction for the original query. The query-correction pairs can be segmented to identify bi-phrases in the query-correction pairs. Probabilities of corrections between the bi-phrases can be estimated based on frequencies of matches in the query-correction pairs. Identifications of the bi-phrases and representations of the probabilities of those bi-phrases can be stored in a probabilistic model data structure. | 12-01-2011 |
20120096042 | USER QUERY REFORMULATION USING RANDOM WALKS - There is provided a computer-implemented method for user query reformulation. A graph is created to represent a relationship between previous user query terms. The graph may represent the previous user searches in n-grams that correspond to nodes. A random walk analysis is performed to determine probabilities that various n-grams corresponding to nodes of the graph could be used to effectively alter a user search term. The probabilities represent a quantification of relationships between nodes of the graph. A determination may be made regarding whether to reformulate the user query based on a relationship between a user search term in the user query and a graphed search term represented by a node of the graph. The determination takes into account a relationship between the user search term and the graphed search term. | 04-19-2012 |
20120131031 | DEPENDENCY-BASED QUERY EXPANSION ALTERATION CANDIDATE SCORING - An alteration candidate for a query can be scored. The scoring may include computing one or more query-dependent feature scores and/or one or more intra-candidate dependent feature scores. The computation of the query-dependent feature score(s) can be based on dependencies to multiple query terms from each of one or more alteration terms (i.e., for each of the one or more alteration terms, there can be dependencies to multiple query terms that form at least a portion of the basis for the query-dependent feature score(s)). The computation of the intra-candidate dependent feature score(s) can be based on dependencies between different terms in the alteration candidate. A candidate score can be computed using the query dependent feature score(s) and/or the intra-candidate dependent feature score(s). Additionally, the candidate score can be used in determining whether to select the candidate to expand the query. If selected, the candidate can be used to expand the query. | 05-24-2012 |
20120150836 | TRAINING PARSERS TO APPROXIMATELY OPTIMIZE NDCG - A supervised technique uses relevance judgments to train a dependency parser such that it approximately optimizes Normalized Discounted Cumulative Gain (NDCG) in information retrieval. A weighted tree edit distance between the parse tree for a query and the parse tree for a document is added to a ranking function, where the edit distance weights are parameters from the parser. Using parser parameters in the ranking function enables approximate optimization of the parser's parameters for NDCG by adding some constraints to the objective function. | 06-14-2012 |
20120158621 | STRUCTURED CROSS-LINGUAL RELEVANCE FEEDBACK FOR ENHANCING SEARCH RESULTS - A “Cross-Lingual Unified Relevance Model” provides a feedback model that improves a machine-learned ranker for a language with few training resources, using feedback from a more complete ranker for a language that has more training resources. The model focuses on linguistically non-local queries, such as “world cup” (English language/U.S. market) and “copa mundial” (Spanish language/Mexican market), that have similar user intent in different languages and markets or regions, thus allowing the low-resource ranker to receive direct relevance feedback from the high-resource ranker. Among other things, the Cross-Lingual Unified Relevance Model differs from conventional relevancy-based techniques by incorporating both query- and document-level features. More specifically, the Cross-Lingual Unified Relevance Model generalizes existing cross-lingual feedback models, incorporating both query expansion and document re-ranking to further amplify the signal from the high-resource ranker to enable a learning to rank approach based on appropriately labeled training data. | 06-21-2012 |
20120203539 | SELECTION OF DOMAIN-ADAPTED TRANSLATION SUBCORPORA - Architecture that provides the capability to subselect the most relevant data from an out-domain corpus to use either in isolation or in combination conjunction with in-domain data. The architecture is a domain adaptation for machine translation that selects the most relevant sentences from a larger general-domain corpus of parallel translated sentences. The methods for selecting the data include monolingual cross-entropy measure, monolingual cross-entropy difference, bilingual cross entropy, and bilingual cross-entropy difference. A translation model is trained on both the in-domain data and an out-domain subset, and the models can be interpolated together to boost performance on in-domain translation tasks. | 08-09-2012 |
20120254217 | Enhanced Query Rewriting Through Click Log Analysis - Systems, methods, and computer media for identifying related strings for search query rewriting are provided. Session data for a user search query session in an accessed click log data is identified. It is determined whether a first additional search query in the session data is related to a first user search query based on at least one of: dwell time; a number of search result links clicked on; and similarity between web page titles or uniform resource locators (URLs). When related, the first additional search query is incorporated into a list of strings related to the first user search query. One or more supplemental strings that are related to the first user search query are also identified. The identified supplemental strings are also included in the list of strings related to the first user search query. | 10-04-2012 |
20120254218 | Enhanced Query Rewriting Through Statistical Machine Translation - Systems, methods, and computer media for identifying query rewriting replacement terms are provided. A list of related string pairs each comprising a first string and second string is received. The first string of each related string pair is a user search query extracted from user click log data. For one or more of the related string pairs, the string pair is provided as inputs to a statistical machine translation model. The model identifies one or more pairs of corresponding terms, each pair of corresponding terms including a first term from the first string and a second term from the second string. The model also calculates a probability of relatedness for each of the one or more pairs of corresponding terms. Term pairs whose calculated probability of relatedness exceeds a threshold are characterized as query term replacements and incorporated, along with the probability of relatedness, into a query rewriting candidate database. | 10-04-2012 |
20120296627 | UNIVERSAL TEXT INPUT - The universal text input technique described herein addresses the difficulties of typing text in various languages and scripts, and offers a unified solution, which combines character conversion, next word prediction, spelling correction and automatic script switching to make it extremely simple to type any language from any device. The technique provides a rich and seamless input experience in any language through a universal IME (input method editor). It allows a user to type in any script for any language using a regular qwerty keyboard via phonetic input and at the same time allows for auto-completion and spelling correction of words and phrases while typing. The technique also provides a modeless input that automatically turns on and off an input mode that changes between different types of script. | 11-22-2012 |
20130311504 | DEPENDENCY-BASED QUERY EXPANSION ALTERATION CANDIDATE SCORING - An alteration candidate for a query can be scored. The scoring may include computing one or more query-dependent feature scores and/or one or more intra-candidate dependent feature scores. The computation of the query-dependent feature score(s) can be based on dependencies to multiple query terms from each of one or more alteration terms (i.e., for each of the one or more alteration terms, there can be dependencies to multiple query terms that form at least a portion of the basis for the query-dependent feature score(s)). The computation of the intra-candidate dependent feature score(s) can be based on dependencies between different terms in the alteration candidate. A candidate score can be computed using the query dependent feature score(s) and/or the intra-candidate dependent feature score(s). Additionally, the candidate score can be used in determining whether to select the candidate to expand the query. If selected, the candidate can be used to expand the query. | 11-21-2013 |