%global _empty_manifest_terminate_build 0 Name: python-fast-autocomplete Version: 0.9.0 Release: 1 Summary: Fast Autocomplete using Directed Word Graph License: MIT URL: https://github.com/seperman/fast-autocomplete Source0: https://mirrors.nju.edu.cn/pypi/web/packages/26/f9/c8cec29088a25c13e8625e044dc55469475d4fd3bb6a5b441bc369733836/fast-autocomplete-0.9.0.tar.gz BuildArch: noarch Requires: python3-Levenshtein Requires: python3-pylev %description # Fast Autocomplete 0.9.0 [zepworks.com](https://zepworks.com) Fast autocomplete using Directed Word Graph (DWG) and Levenshtein Edit Distance. The results are cached via LFU (Least Frequently Used). # Why Read about why fast-autocomplete was built here: This library was written when we came to the conclusion that Elasticsearch's Autocomplete suggestor is not fast enough and doesn't do everything that we need: 1. Once we switched to Fast Autocomplete, our average latency went from 120ms to 30ms so an improvement of 3-4x in performance and errors went down to zero. 2. Elasticsearch's Autocomplete suggestor does not handle any sort of combination of the words you have put in. For example Fast Autocomplete can handle `2018 Toyota Camry in Los Angeles` when the words `2018`, `Toyota Camry`, `Los Angeles` are seperately fed into it. While Elasticsearch's autocomplete needs that whole sentence to be fed to it to show it in Autocomplete results. You might say: 1. Regarding #1: Yes, but you are using caching. Answer: shhh Yes, keep it quiet. We are also doing Levenshtein Edit distance using a C library so it improves there too. 2. Regarding #2: Cool. Answer: Ok, now we are talking. # How Read about how fast-autocomplete works here: In a nutshell, what the fast Autocomplete does is: 1. Populate the DWG with your words. 2. Follow the graph nodes letter by letter until it finds nodes that have words in them. 3. Continue after words are found on the graph until it reaches the leaf node. 4. Restart from the root node again until it reaches a letter that doesn't exist on the graph. 5. Depending on how much is left from the rest of the word, return all the descendant words from where it got stuck 6. Or run Levenshtein edit distance to find closes words to what is left and the continue from there. By doing so, it can tokenize a text such as: `2018 Toyota Camry in Los Angeles` into [`2018`, `toyota camry`, `in`, `los angeles`] And return Autocomplete results as you type. # Install `pip install fast-autocomplete` **Note: Fast Autocomplete only works with Python 3.6 and newer.** Are you still on Python 2? TIME TO UPGRADE. # Licence MIT # DWG The data structure we use in this library is called Dawg. DWG stands for Directed Word Graph. Here is an example DWG based on the "makes_models_short.csv" that is provided in the tests: ![dwg](tests/animation/short.gif) ![dwg](tests/AutoCompleteWithSynonymsShort_Graph.svg) # Usage First of all lets start from your data. The library leaves it up to you how to prepare your data. If you want to go straight to the factory function that lets you use the library in its easiest and most common case, skip all these and jump to the [sorting](#sorting) example. ## Example 1 ```py >>> from fast_autocomplete import AutoComplete >>> words = {'book': {}, 'burrito': {}, 'pizza': {}, 'pasta':{}} >>> autocomplete = AutoComplete(words=words) >>> autocomplete.search(word='b', max_cost=3, size=3) [['book'], ['burrito']] >>> autocomplete.search(word='bu', max_cost=3, size=3) [['burrito']] >>> autocomplete.search(word='barrito', max_cost=3, size=3) # mis-spelling [['burrito']] ``` Words is a dictionary and each word can have a context. For example the "count", how to display the word, some other context around the word etc. In this example words didn't have any context. ## Example 2 Imagine that we have a csv with the following content from vehicles' make and models: ```csv make,model acura,zdx alfa romeo,4c alfa romeo,4c coupe alfa romeo,giulia bmw,1 series bmw,2 series 2007,2007 2017,2017 2018,2018 ``` What we want to do is to convert this to a dictionary of words and their context. ```py import csv from fast_autocomplete.misc import read_csv_gen def get_words(path): csv_gen = read_csv_gen(path, csv_func=csv.DictReader) words = {} for line in csv_gen: make = line['make'] model = line['model'] if make != model: local_words = [model, '{} {}'.format(make, model)] while local_words: word = local_words.pop() if word not in words: words[word] = {} if make not in words: words[make] = {} return words ``` the `read_csv_gen` is just a helper function. You don't really need it. The whole point is that we are converting that csv to a dictionary that looks like this: ```py >>> words = get_words('path to the csv') >>> words {'acura zdx': {}, 'zdx': {}, 'acura': {}, 'alfa romeo 4c': {}, '4c': {}, 'alfa romeo': {}, 'alfa romeo 4c coupe': {}, '4c coupe': {}, 'alfa romeo giulia': {}, 'giulia': {}, 'bmw 1 series': {}, '1 series': {}, 'bmw': {}, 'bmw 2 series': {}, '2 series': {}, '2007': {}, '2017': {}, '2018': {}} ``` This is a dictionary of words to their context. We have decided that we don't want any context for the words in this example so all the contexts are empty. However generally you will want some context around the words for more complicated logics. The context is used to convert the words "keys" into their context which is the value of the key in the words dictionary. In addition to words, we usually want a dictionary of synonyms. Something like this: ```py synonyms = { "alfa romeo": ["alfa"], "bmw": ["beemer", "bimmer"], "mercedes-benz": ["mercedes", "benz"], "volkswagen": ["vw"] } ``` Note that synonyms are optional. Maybe in your use case you don't need synonyms. Now we can use the above to initialize Autocomplete ```py from fast_autocomplete import AutoComplete autocomplete = AutoComplete(words=words, synonyms=synonyms) ``` At this point, AutoComplete has created a [dwg](#DWG) structure. Now you can search! - word: the word to return autocomplete results for - max_cost: Maximum Levenshtein edit distance to be considered when calculating results - size: The max number of results to return ```py >>> autocomplete.search(word='2018 bmw 1', max_cost=3, size=3) [['2018', 'bmw'], ['2018', 'bmw 1 series']] ``` Now what if we pressed a by mistake then? It still works. No problem. ```py >>> autocomplete.search(word='2018 bmw 1a', max_cost=3, size=3) [['2018', 'bmw'], ['2018', 'bmw 1 series']] ``` Ok let's search for Alfa now: ```py >>> autocomplete.search(word='alfa', max_cost=3, size=3) [['alfa romeo'], ['alfa romeo 4c'], ['alfa romeo giulia']] ``` What if we don't know how to pronounce alfa and we type `alpha` ? ```py >>> autocomplete.search(word='alpha', max_cost=3, size=3) [['alfa romeo'], ['alfa romeo 4c'], ['alfa romeo giulia']] ``` It still works! Fast-Autocomplete makes sure the results make sense! Ok lets add the word `Los Angeles` there to the words: ```py >>> words['los angeles'] = {} >>> words['in'] = {} >>> autocomplete.search(word='2007 alfa in los', max_cost=3, size=3) [['2007', 'alfa romeo', 'in'], ['2007', 'alfa romeo', 'in', 'los angeles']] ``` So far we have not used the context. And this library leaves it up to you how to use the context. But basically if we giving a context to each one of those words, then the above response could easly be translated to the list of those contexts. ## context If our words dictionary was: ```py words = { 'in': {}, 'alfa romeo': {'type': 'make'}, '2007': {'type': 'year'}, 'los angeles': {'type': 'location'}, } ``` Then the `autocomplete.words` can be used to map the results into their context: ``` [['2007', 'alfa romeo', 'in'], ['2007', 'alfa romeo', 'in', 'los angeles']] converted to contexts: [[{'year': '2007'}, {'make': alfa romeo'}], [{'year': '2007'}, {'make': alfa romeo'}, {'location': 'los angeles'}]] ``` ## Sorting Most people who use Fast Autocomplete, want to control how results are sorted. If you don't control that, the results will be sorted based on the order that Autocomplete found the nodes in the graph that matched the criteria. The easiest way to sort is to give each item a count. **Fast AutoComplete will use the count to sort items that are partial matches.** For example: 1. Make a json file that is a dictionary of words to their context. The format of the file needs to be: ```json { word: [ context, display value, count ] } ``` An example is included in the [sample_words.json](tests/fixtures/sample_words.json) ```json { "acura rlx": [ { "model": "rlx", "make": "acura" }, "Acura RLX", 3132 ], "rlx": [ { "model": "rlx", "make": "acura" }, "Acura RLX", 3132 ], "acura": [ { "make": "acura" }, "Acura", 130123 ], ... } ``` You might be wondering why things are in this format. It is to save space when this json can become very big easily and the keys become repetitive. That's why we are using a list with predefined order of keys. For your use case for now you can leave the context and display values as None if you want. We will opensource other factory functions soon that will fully utilize those keys in the context. 2. Launch Autocomplete via the factory function: ```py from fast_autocomplete import autocomplete_factory content_files = { 'words': { 'filepath': path/to/sample_words.json, 'compress': True # means compress the graph data in memory } } autocomplete = autocomplete_factory(content_files=content_files) ``` 3. You can use Autocomplete and the results are ordered by count! ```py >>> autocomplete.search(word='acu') [['acura'], ['acura mdx'], ['acura rdx']] ``` 4. How do we use the context and display value now? Great question. You need to extend AutoComplete class to use these items. I will write a blog post about it. Here is a simple example without any extending: ```py >>> autocomplete.words['acura'] WordValue(context={'make': 'acura'}, display='Acura', count=130123, original_key=None) >>> autocomplete.words['acura'].display Acura ``` ### Change the sorting by updating counts Fast Autocomplete by default uses the "count" of the items to sort the items in the results. Think about these counts as a "guide" to Fast autocomplete so it can polish its results. Depending on whether or not Fast autocomplete finds exact matches to user's query, the counts will be used to refine the results. You can update the counts in an autocomplete object live. For example, in the [sample csv of car makes and models](tests/fixtures/makes_models_from_wikipedia.csv) we have: ```csv make,model,count Toyota,Aurion,6094 Toyota,Avalon,8803 Toyota,Avensis,1630 Toyota,Auris,4025 Toyota,Aygo,2115 ``` If we use the autocomplete to search: ```py >>> auto_complete = AutoComplete(words=WIKIPEDIA_WORDS, synonyms=SYNONYMS, full_stop_words=['bmw', 'alfa romeo']) >>> autocomplete.search(word='toyota a') [['toyota'], ['toyota avalon'], ['toyota aurion'], ['toyota auris']] ``` However as you can notice `toyota aygo` had the count of 2115 and thus it didn't make it to the top 3 results. We can set the count for `toyota aygo` to a higher number to boost it in the results using `update_count_of_word`. The `update_count_of_word` can change the count via setting the word's count directly or by offsetting its current value. ```py >>> auto_complete = AutoComplete(words=WIKIPEDIA_WORDS, synonyms=SYNONYMS, full_stop_words=['bmw', 'alfa romeo']) >>> auto_complete.update_count_of_word(word='toyota aygo', count=10000) 10000 ``` Now if we search: ```py >>> autocomplete.search(word='toyota a') [['toyota'], ['toyota aygo'], ['toyota avalon'], ['toyota aurion']] ``` We can double check the count of a node: ```py >>> autocomplete.get_count_of_word('toyota aygo') 10000 ``` Now let's use the offset to offset the current count of a different node: ```py >>> auto_complete.update_count_of_word(word='toyota aurion', offset=-6000) 94 ``` When we search, `toyota aurion` is not in the top 3 results anymore! ```py >>> autocomplete.search(word='toyota a') [['toyota'], ['toyota aygo'], ['toyota avalon'], ['toyota auris']] ``` ## Unicode By default this package only accepts ASCII lowercase letters, a-z. However you can pass the characters that you want to be acceptable via `valid_chars_for_string` for strings, and `valid_chars_for_integer` for numbers. For example here we tell Autocomplete to consider the Farsi alphabet characters for string characters. ```python AutoComplete( words=SHORT_WORDS_UNICODE, valid_chars_for_string='اآبپتثجچحخدذرزژسشصضطظعغفقکگلمنوهی') ``` If you want to pass other characters in addition to ASCII letters, such as punctuation marks, you need to set the `valid_chars_for_string` variable to include all of the characters you need. For example, the following code block sets ASCII letters a-z along with periods and apostrophes: ```python valid_chars = ".'" valid_chars += string.ascii_lowercase AutoComplete( words=WORDS_WITH_PUNCTUATION, valid_chars_for_string=valid_chars) ``` ## Draw This package can actually draw the dwgs as it is populating them or just once the dwg is populated for you! Here is the animation of populating the dwg with words from "makes_models_short.csv": ### Draw animation of dwg populating ```py from fast_autocomplete import AutoComplete, DrawGraphMixin class AutoCompleteDraw(DrawGraphMixin, AutoComplete): DRAW_POPULATION_ANIMATION = True DRAW_POPULATION_ANIMATION_PATH = 'animation/short_.svg' DRAW_POPULATION_ANIMATION_FILENO_PADDING = 6 autocomplete = AutoCompleteDraw(words=words, synonyms=synonyms) ``` As soon as you initialize the above AutoCompleteDraw class, it will populate the dwg and generate the animation! For an example of this code properly setup, take a look at the tests. In fact the animation in the [dwg](#dwg) section is generated the same way via unit tests! Note that if you have many words, the graph file will be big. Instead of drawing all frames as the dwg is being populated, you can just draw the final stage: ### Draw the final graph To draw just one graph that shows the final stage of the dwg, use the draw mixin and run the draw_graph function: ```py from fast_autocomplete import AutoComplete, DrawGraphMixin class AutoCompleteDraw(DrawGraphMixin, AutoComplete): pass autocomplete = AutoCompleteDraw(words=words, synonyms=synonyms) autocomplete.draw_graph('path to file') ``` ## Demo If you want to have a real-time interaction with Autocomplete results in your terminal, you can use the demo module: Just pass it an instance of the autocomplete and the search configs: ```py from fast_autocomplete import demo demo(autocomplete, max_cost=3, size=5) ``` # # Develop 1. Clone the repo 2. Make a virtualenv with Python 3.6 or newer 3. `pip install -r requirements-dev.txt` ## Run tests `pytest` We try to maintain high standard in code coverage. Currently the `dwg` module's coverage is around 99%! ## Releases We use bump2version to bump and tag releases. ```bash git checkout master && git pull bump2version {patch|minor|major} git push && git push --tags ``` # Authors - Autocomplete written by [Sep Dehpour](http://zepworks.com). - LFU Cache by [Shane Wang](https://medium.com/@epicshane) # Other ways of doing AutoComplete 1. Elastic search. Yes, Elasticsearch generally is a *better* Autocomplete solution than this library. I said generally. In our specific use case, we wanted Autocomplete to be faster than Elasticsearch and handle combination of words. Otherwise Elasticsearch would have been perfect. Behind the scene Elasticsearch uses Finite State Transducer (FST) in Lucene to achive AutoComplete. FST is more complicated than what we have used in fast-autocomplete. 2. If your autocomplete is supposed to return results based on a big blog of text (for example based on some book contents), then a better solution is to go with Markov chains and conditional probability. Yes, there is already a library out there for it! and it looks great. Disclaimer: we have not actually used it since it doesn't fit our specific use-case. # FAQ ## Why DWG DWG stands for Directed Word Graph. Originally we were using Trie-Tree structure. But soon it was obvious that some branches needed to merge back to other branches. Such as `beemer` and `bmw` branches both need to end in the same node since they are synonyms. Thus we used DWG. ## What are synonyms, clean synonyms and partial synonyms Synonyms are words that should produce the same results. - For example `beemer` and `bmw` should both give you `bmw`. - `alfa` and `alfa romeo` should both give you `alfa romeo` The synonyms get divided into 2 groups: 1. clean synonyms: The 2 words share little or no words. For example `beemer` vs. `bmw`. 2. partial synonyms: One of the 2 words is a substring of the other one. For example `alfa` and `alfa romeo` or `gm` vs. `gmc`. Internally these 2 types of synonyms are treated differently but as a user of the library, you don't need to really care about it. You just provide the synonyms dictionary via defining the `get_synonyms` method. ## Why do you have a whole subtree for partial synonyms Q: Partial synonym means the synonym is a part of the original word. Such as `alfa` is a partial synonym for `alfa romeo`. In that case you are inserting both `alfa` and `alfa romeo` in the dwg. `alfa` will have `alfa 4c` and `alpha romeo` will have `alfa romeo 4c` branches. Why not just have `alfa` branches to be `alfa romeo` and from there you will have automatically all the sub branches of `alfa romeo`. Answer: We use letters for edges. So `alfa` can have only one edge coming out of it that is space (` `). And that edge is going to a node that has sub-branches to `alfa romoe`, `alfa 4c` etc. It can't have a ` ` going to that node and another ` ` going to `alfa romeo`'s immediate child. That way when we are traversing the dwg for the input of `alfa 4` we get to the correct node. ## I put Toyota in the Dawg but when I type `toy`, it doesn't show up. Answer: If you put `Toyota` with capital T in the dwg, it expects the search word to start with capital T too. We suggest that you lower case everything before putting them in dwg. Fast-autocomplete does not automatically do that for you since it assumes the `words` dictionary is what you want to be put in the dwg. It is up to you to clean your own data before putting it in the dwg. %package -n python3-fast-autocomplete Summary: Fast Autocomplete using Directed Word Graph Provides: python-fast-autocomplete BuildRequires: python3-devel BuildRequires: python3-setuptools BuildRequires: python3-pip %description -n python3-fast-autocomplete # Fast Autocomplete 0.9.0 [zepworks.com](https://zepworks.com) Fast autocomplete using Directed Word Graph (DWG) and Levenshtein Edit Distance. The results are cached via LFU (Least Frequently Used). # Why Read about why fast-autocomplete was built here: This library was written when we came to the conclusion that Elasticsearch's Autocomplete suggestor is not fast enough and doesn't do everything that we need: 1. Once we switched to Fast Autocomplete, our average latency went from 120ms to 30ms so an improvement of 3-4x in performance and errors went down to zero. 2. Elasticsearch's Autocomplete suggestor does not handle any sort of combination of the words you have put in. For example Fast Autocomplete can handle `2018 Toyota Camry in Los Angeles` when the words `2018`, `Toyota Camry`, `Los Angeles` are seperately fed into it. While Elasticsearch's autocomplete needs that whole sentence to be fed to it to show it in Autocomplete results. You might say: 1. Regarding #1: Yes, but you are using caching. Answer: shhh Yes, keep it quiet. We are also doing Levenshtein Edit distance using a C library so it improves there too. 2. Regarding #2: Cool. Answer: Ok, now we are talking. # How Read about how fast-autocomplete works here: In a nutshell, what the fast Autocomplete does is: 1. Populate the DWG with your words. 2. Follow the graph nodes letter by letter until it finds nodes that have words in them. 3. Continue after words are found on the graph until it reaches the leaf node. 4. Restart from the root node again until it reaches a letter that doesn't exist on the graph. 5. Depending on how much is left from the rest of the word, return all the descendant words from where it got stuck 6. Or run Levenshtein edit distance to find closes words to what is left and the continue from there. By doing so, it can tokenize a text such as: `2018 Toyota Camry in Los Angeles` into [`2018`, `toyota camry`, `in`, `los angeles`] And return Autocomplete results as you type. # Install `pip install fast-autocomplete` **Note: Fast Autocomplete only works with Python 3.6 and newer.** Are you still on Python 2? TIME TO UPGRADE. # Licence MIT # DWG The data structure we use in this library is called Dawg. DWG stands for Directed Word Graph. Here is an example DWG based on the "makes_models_short.csv" that is provided in the tests: ![dwg](tests/animation/short.gif) ![dwg](tests/AutoCompleteWithSynonymsShort_Graph.svg) # Usage First of all lets start from your data. The library leaves it up to you how to prepare your data. If you want to go straight to the factory function that lets you use the library in its easiest and most common case, skip all these and jump to the [sorting](#sorting) example. ## Example 1 ```py >>> from fast_autocomplete import AutoComplete >>> words = {'book': {}, 'burrito': {}, 'pizza': {}, 'pasta':{}} >>> autocomplete = AutoComplete(words=words) >>> autocomplete.search(word='b', max_cost=3, size=3) [['book'], ['burrito']] >>> autocomplete.search(word='bu', max_cost=3, size=3) [['burrito']] >>> autocomplete.search(word='barrito', max_cost=3, size=3) # mis-spelling [['burrito']] ``` Words is a dictionary and each word can have a context. For example the "count", how to display the word, some other context around the word etc. In this example words didn't have any context. ## Example 2 Imagine that we have a csv with the following content from vehicles' make and models: ```csv make,model acura,zdx alfa romeo,4c alfa romeo,4c coupe alfa romeo,giulia bmw,1 series bmw,2 series 2007,2007 2017,2017 2018,2018 ``` What we want to do is to convert this to a dictionary of words and their context. ```py import csv from fast_autocomplete.misc import read_csv_gen def get_words(path): csv_gen = read_csv_gen(path, csv_func=csv.DictReader) words = {} for line in csv_gen: make = line['make'] model = line['model'] if make != model: local_words = [model, '{} {}'.format(make, model)] while local_words: word = local_words.pop() if word not in words: words[word] = {} if make not in words: words[make] = {} return words ``` the `read_csv_gen` is just a helper function. You don't really need it. The whole point is that we are converting that csv to a dictionary that looks like this: ```py >>> words = get_words('path to the csv') >>> words {'acura zdx': {}, 'zdx': {}, 'acura': {}, 'alfa romeo 4c': {}, '4c': {}, 'alfa romeo': {}, 'alfa romeo 4c coupe': {}, '4c coupe': {}, 'alfa romeo giulia': {}, 'giulia': {}, 'bmw 1 series': {}, '1 series': {}, 'bmw': {}, 'bmw 2 series': {}, '2 series': {}, '2007': {}, '2017': {}, '2018': {}} ``` This is a dictionary of words to their context. We have decided that we don't want any context for the words in this example so all the contexts are empty. However generally you will want some context around the words for more complicated logics. The context is used to convert the words "keys" into their context which is the value of the key in the words dictionary. In addition to words, we usually want a dictionary of synonyms. Something like this: ```py synonyms = { "alfa romeo": ["alfa"], "bmw": ["beemer", "bimmer"], "mercedes-benz": ["mercedes", "benz"], "volkswagen": ["vw"] } ``` Note that synonyms are optional. Maybe in your use case you don't need synonyms. Now we can use the above to initialize Autocomplete ```py from fast_autocomplete import AutoComplete autocomplete = AutoComplete(words=words, synonyms=synonyms) ``` At this point, AutoComplete has created a [dwg](#DWG) structure. Now you can search! - word: the word to return autocomplete results for - max_cost: Maximum Levenshtein edit distance to be considered when calculating results - size: The max number of results to return ```py >>> autocomplete.search(word='2018 bmw 1', max_cost=3, size=3) [['2018', 'bmw'], ['2018', 'bmw 1 series']] ``` Now what if we pressed a by mistake then? It still works. No problem. ```py >>> autocomplete.search(word='2018 bmw 1a', max_cost=3, size=3) [['2018', 'bmw'], ['2018', 'bmw 1 series']] ``` Ok let's search for Alfa now: ```py >>> autocomplete.search(word='alfa', max_cost=3, size=3) [['alfa romeo'], ['alfa romeo 4c'], ['alfa romeo giulia']] ``` What if we don't know how to pronounce alfa and we type `alpha` ? ```py >>> autocomplete.search(word='alpha', max_cost=3, size=3) [['alfa romeo'], ['alfa romeo 4c'], ['alfa romeo giulia']] ``` It still works! Fast-Autocomplete makes sure the results make sense! Ok lets add the word `Los Angeles` there to the words: ```py >>> words['los angeles'] = {} >>> words['in'] = {} >>> autocomplete.search(word='2007 alfa in los', max_cost=3, size=3) [['2007', 'alfa romeo', 'in'], ['2007', 'alfa romeo', 'in', 'los angeles']] ``` So far we have not used the context. And this library leaves it up to you how to use the context. But basically if we giving a context to each one of those words, then the above response could easly be translated to the list of those contexts. ## context If our words dictionary was: ```py words = { 'in': {}, 'alfa romeo': {'type': 'make'}, '2007': {'type': 'year'}, 'los angeles': {'type': 'location'}, } ``` Then the `autocomplete.words` can be used to map the results into their context: ``` [['2007', 'alfa romeo', 'in'], ['2007', 'alfa romeo', 'in', 'los angeles']] converted to contexts: [[{'year': '2007'}, {'make': alfa romeo'}], [{'year': '2007'}, {'make': alfa romeo'}, {'location': 'los angeles'}]] ``` ## Sorting Most people who use Fast Autocomplete, want to control how results are sorted. If you don't control that, the results will be sorted based on the order that Autocomplete found the nodes in the graph that matched the criteria. The easiest way to sort is to give each item a count. **Fast AutoComplete will use the count to sort items that are partial matches.** For example: 1. Make a json file that is a dictionary of words to their context. The format of the file needs to be: ```json { word: [ context, display value, count ] } ``` An example is included in the [sample_words.json](tests/fixtures/sample_words.json) ```json { "acura rlx": [ { "model": "rlx", "make": "acura" }, "Acura RLX", 3132 ], "rlx": [ { "model": "rlx", "make": "acura" }, "Acura RLX", 3132 ], "acura": [ { "make": "acura" }, "Acura", 130123 ], ... } ``` You might be wondering why things are in this format. It is to save space when this json can become very big easily and the keys become repetitive. That's why we are using a list with predefined order of keys. For your use case for now you can leave the context and display values as None if you want. We will opensource other factory functions soon that will fully utilize those keys in the context. 2. Launch Autocomplete via the factory function: ```py from fast_autocomplete import autocomplete_factory content_files = { 'words': { 'filepath': path/to/sample_words.json, 'compress': True # means compress the graph data in memory } } autocomplete = autocomplete_factory(content_files=content_files) ``` 3. You can use Autocomplete and the results are ordered by count! ```py >>> autocomplete.search(word='acu') [['acura'], ['acura mdx'], ['acura rdx']] ``` 4. How do we use the context and display value now? Great question. You need to extend AutoComplete class to use these items. I will write a blog post about it. Here is a simple example without any extending: ```py >>> autocomplete.words['acura'] WordValue(context={'make': 'acura'}, display='Acura', count=130123, original_key=None) >>> autocomplete.words['acura'].display Acura ``` ### Change the sorting by updating counts Fast Autocomplete by default uses the "count" of the items to sort the items in the results. Think about these counts as a "guide" to Fast autocomplete so it can polish its results. Depending on whether or not Fast autocomplete finds exact matches to user's query, the counts will be used to refine the results. You can update the counts in an autocomplete object live. For example, in the [sample csv of car makes and models](tests/fixtures/makes_models_from_wikipedia.csv) we have: ```csv make,model,count Toyota,Aurion,6094 Toyota,Avalon,8803 Toyota,Avensis,1630 Toyota,Auris,4025 Toyota,Aygo,2115 ``` If we use the autocomplete to search: ```py >>> auto_complete = AutoComplete(words=WIKIPEDIA_WORDS, synonyms=SYNONYMS, full_stop_words=['bmw', 'alfa romeo']) >>> autocomplete.search(word='toyota a') [['toyota'], ['toyota avalon'], ['toyota aurion'], ['toyota auris']] ``` However as you can notice `toyota aygo` had the count of 2115 and thus it didn't make it to the top 3 results. We can set the count for `toyota aygo` to a higher number to boost it in the results using `update_count_of_word`. The `update_count_of_word` can change the count via setting the word's count directly or by offsetting its current value. ```py >>> auto_complete = AutoComplete(words=WIKIPEDIA_WORDS, synonyms=SYNONYMS, full_stop_words=['bmw', 'alfa romeo']) >>> auto_complete.update_count_of_word(word='toyota aygo', count=10000) 10000 ``` Now if we search: ```py >>> autocomplete.search(word='toyota a') [['toyota'], ['toyota aygo'], ['toyota avalon'], ['toyota aurion']] ``` We can double check the count of a node: ```py >>> autocomplete.get_count_of_word('toyota aygo') 10000 ``` Now let's use the offset to offset the current count of a different node: ```py >>> auto_complete.update_count_of_word(word='toyota aurion', offset=-6000) 94 ``` When we search, `toyota aurion` is not in the top 3 results anymore! ```py >>> autocomplete.search(word='toyota a') [['toyota'], ['toyota aygo'], ['toyota avalon'], ['toyota auris']] ``` ## Unicode By default this package only accepts ASCII lowercase letters, a-z. However you can pass the characters that you want to be acceptable via `valid_chars_for_string` for strings, and `valid_chars_for_integer` for numbers. For example here we tell Autocomplete to consider the Farsi alphabet characters for string characters. ```python AutoComplete( words=SHORT_WORDS_UNICODE, valid_chars_for_string='اآبپتثجچحخدذرزژسشصضطظعغفقکگلمنوهی') ``` If you want to pass other characters in addition to ASCII letters, such as punctuation marks, you need to set the `valid_chars_for_string` variable to include all of the characters you need. For example, the following code block sets ASCII letters a-z along with periods and apostrophes: ```python valid_chars = ".'" valid_chars += string.ascii_lowercase AutoComplete( words=WORDS_WITH_PUNCTUATION, valid_chars_for_string=valid_chars) ``` ## Draw This package can actually draw the dwgs as it is populating them or just once the dwg is populated for you! Here is the animation of populating the dwg with words from "makes_models_short.csv": ### Draw animation of dwg populating ```py from fast_autocomplete import AutoComplete, DrawGraphMixin class AutoCompleteDraw(DrawGraphMixin, AutoComplete): DRAW_POPULATION_ANIMATION = True DRAW_POPULATION_ANIMATION_PATH = 'animation/short_.svg' DRAW_POPULATION_ANIMATION_FILENO_PADDING = 6 autocomplete = AutoCompleteDraw(words=words, synonyms=synonyms) ``` As soon as you initialize the above AutoCompleteDraw class, it will populate the dwg and generate the animation! For an example of this code properly setup, take a look at the tests. In fact the animation in the [dwg](#dwg) section is generated the same way via unit tests! Note that if you have many words, the graph file will be big. Instead of drawing all frames as the dwg is being populated, you can just draw the final stage: ### Draw the final graph To draw just one graph that shows the final stage of the dwg, use the draw mixin and run the draw_graph function: ```py from fast_autocomplete import AutoComplete, DrawGraphMixin class AutoCompleteDraw(DrawGraphMixin, AutoComplete): pass autocomplete = AutoCompleteDraw(words=words, synonyms=synonyms) autocomplete.draw_graph('path to file') ``` ## Demo If you want to have a real-time interaction with Autocomplete results in your terminal, you can use the demo module: Just pass it an instance of the autocomplete and the search configs: ```py from fast_autocomplete import demo demo(autocomplete, max_cost=3, size=5) ``` # # Develop 1. Clone the repo 2. Make a virtualenv with Python 3.6 or newer 3. `pip install -r requirements-dev.txt` ## Run tests `pytest` We try to maintain high standard in code coverage. Currently the `dwg` module's coverage is around 99%! ## Releases We use bump2version to bump and tag releases. ```bash git checkout master && git pull bump2version {patch|minor|major} git push && git push --tags ``` # Authors - Autocomplete written by [Sep Dehpour](http://zepworks.com). - LFU Cache by [Shane Wang](https://medium.com/@epicshane) # Other ways of doing AutoComplete 1. Elastic search. Yes, Elasticsearch generally is a *better* Autocomplete solution than this library. I said generally. In our specific use case, we wanted Autocomplete to be faster than Elasticsearch and handle combination of words. Otherwise Elasticsearch would have been perfect. Behind the scene Elasticsearch uses Finite State Transducer (FST) in Lucene to achive AutoComplete. FST is more complicated than what we have used in fast-autocomplete. 2. If your autocomplete is supposed to return results based on a big blog of text (for example based on some book contents), then a better solution is to go with Markov chains and conditional probability. Yes, there is already a library out there for it! and it looks great. Disclaimer: we have not actually used it since it doesn't fit our specific use-case. # FAQ ## Why DWG DWG stands for Directed Word Graph. Originally we were using Trie-Tree structure. But soon it was obvious that some branches needed to merge back to other branches. Such as `beemer` and `bmw` branches both need to end in the same node since they are synonyms. Thus we used DWG. ## What are synonyms, clean synonyms and partial synonyms Synonyms are words that should produce the same results. - For example `beemer` and `bmw` should both give you `bmw`. - `alfa` and `alfa romeo` should both give you `alfa romeo` The synonyms get divided into 2 groups: 1. clean synonyms: The 2 words share little or no words. For example `beemer` vs. `bmw`. 2. partial synonyms: One of the 2 words is a substring of the other one. For example `alfa` and `alfa romeo` or `gm` vs. `gmc`. Internally these 2 types of synonyms are treated differently but as a user of the library, you don't need to really care about it. You just provide the synonyms dictionary via defining the `get_synonyms` method. ## Why do you have a whole subtree for partial synonyms Q: Partial synonym means the synonym is a part of the original word. Such as `alfa` is a partial synonym for `alfa romeo`. In that case you are inserting both `alfa` and `alfa romeo` in the dwg. `alfa` will have `alfa 4c` and `alpha romeo` will have `alfa romeo 4c` branches. Why not just have `alfa` branches to be `alfa romeo` and from there you will have automatically all the sub branches of `alfa romeo`. Answer: We use letters for edges. So `alfa` can have only one edge coming out of it that is space (` `). And that edge is going to a node that has sub-branches to `alfa romoe`, `alfa 4c` etc. It can't have a ` ` going to that node and another ` ` going to `alfa romeo`'s immediate child. That way when we are traversing the dwg for the input of `alfa 4` we get to the correct node. ## I put Toyota in the Dawg but when I type `toy`, it doesn't show up. Answer: If you put `Toyota` with capital T in the dwg, it expects the search word to start with capital T too. We suggest that you lower case everything before putting them in dwg. Fast-autocomplete does not automatically do that for you since it assumes the `words` dictionary is what you want to be put in the dwg. It is up to you to clean your own data before putting it in the dwg. %package help Summary: Development documents and examples for fast-autocomplete Provides: python3-fast-autocomplete-doc %description help # Fast Autocomplete 0.9.0 [zepworks.com](https://zepworks.com) Fast autocomplete using Directed Word Graph (DWG) and Levenshtein Edit Distance. The results are cached via LFU (Least Frequently Used). # Why Read about why fast-autocomplete was built here: This library was written when we came to the conclusion that Elasticsearch's Autocomplete suggestor is not fast enough and doesn't do everything that we need: 1. Once we switched to Fast Autocomplete, our average latency went from 120ms to 30ms so an improvement of 3-4x in performance and errors went down to zero. 2. Elasticsearch's Autocomplete suggestor does not handle any sort of combination of the words you have put in. For example Fast Autocomplete can handle `2018 Toyota Camry in Los Angeles` when the words `2018`, `Toyota Camry`, `Los Angeles` are seperately fed into it. While Elasticsearch's autocomplete needs that whole sentence to be fed to it to show it in Autocomplete results. You might say: 1. Regarding #1: Yes, but you are using caching. Answer: shhh Yes, keep it quiet. We are also doing Levenshtein Edit distance using a C library so it improves there too. 2. Regarding #2: Cool. Answer: Ok, now we are talking. # How Read about how fast-autocomplete works here: In a nutshell, what the fast Autocomplete does is: 1. Populate the DWG with your words. 2. Follow the graph nodes letter by letter until it finds nodes that have words in them. 3. Continue after words are found on the graph until it reaches the leaf node. 4. Restart from the root node again until it reaches a letter that doesn't exist on the graph. 5. Depending on how much is left from the rest of the word, return all the descendant words from where it got stuck 6. Or run Levenshtein edit distance to find closes words to what is left and the continue from there. By doing so, it can tokenize a text such as: `2018 Toyota Camry in Los Angeles` into [`2018`, `toyota camry`, `in`, `los angeles`] And return Autocomplete results as you type. # Install `pip install fast-autocomplete` **Note: Fast Autocomplete only works with Python 3.6 and newer.** Are you still on Python 2? TIME TO UPGRADE. # Licence MIT # DWG The data structure we use in this library is called Dawg. DWG stands for Directed Word Graph. Here is an example DWG based on the "makes_models_short.csv" that is provided in the tests: ![dwg](tests/animation/short.gif) ![dwg](tests/AutoCompleteWithSynonymsShort_Graph.svg) # Usage First of all lets start from your data. The library leaves it up to you how to prepare your data. If you want to go straight to the factory function that lets you use the library in its easiest and most common case, skip all these and jump to the [sorting](#sorting) example. ## Example 1 ```py >>> from fast_autocomplete import AutoComplete >>> words = {'book': {}, 'burrito': {}, 'pizza': {}, 'pasta':{}} >>> autocomplete = AutoComplete(words=words) >>> autocomplete.search(word='b', max_cost=3, size=3) [['book'], ['burrito']] >>> autocomplete.search(word='bu', max_cost=3, size=3) [['burrito']] >>> autocomplete.search(word='barrito', max_cost=3, size=3) # mis-spelling [['burrito']] ``` Words is a dictionary and each word can have a context. For example the "count", how to display the word, some other context around the word etc. In this example words didn't have any context. ## Example 2 Imagine that we have a csv with the following content from vehicles' make and models: ```csv make,model acura,zdx alfa romeo,4c alfa romeo,4c coupe alfa romeo,giulia bmw,1 series bmw,2 series 2007,2007 2017,2017 2018,2018 ``` What we want to do is to convert this to a dictionary of words and their context. ```py import csv from fast_autocomplete.misc import read_csv_gen def get_words(path): csv_gen = read_csv_gen(path, csv_func=csv.DictReader) words = {} for line in csv_gen: make = line['make'] model = line['model'] if make != model: local_words = [model, '{} {}'.format(make, model)] while local_words: word = local_words.pop() if word not in words: words[word] = {} if make not in words: words[make] = {} return words ``` the `read_csv_gen` is just a helper function. You don't really need it. The whole point is that we are converting that csv to a dictionary that looks like this: ```py >>> words = get_words('path to the csv') >>> words {'acura zdx': {}, 'zdx': {}, 'acura': {}, 'alfa romeo 4c': {}, '4c': {}, 'alfa romeo': {}, 'alfa romeo 4c coupe': {}, '4c coupe': {}, 'alfa romeo giulia': {}, 'giulia': {}, 'bmw 1 series': {}, '1 series': {}, 'bmw': {}, 'bmw 2 series': {}, '2 series': {}, '2007': {}, '2017': {}, '2018': {}} ``` This is a dictionary of words to their context. We have decided that we don't want any context for the words in this example so all the contexts are empty. However generally you will want some context around the words for more complicated logics. The context is used to convert the words "keys" into their context which is the value of the key in the words dictionary. In addition to words, we usually want a dictionary of synonyms. Something like this: ```py synonyms = { "alfa romeo": ["alfa"], "bmw": ["beemer", "bimmer"], "mercedes-benz": ["mercedes", "benz"], "volkswagen": ["vw"] } ``` Note that synonyms are optional. Maybe in your use case you don't need synonyms. Now we can use the above to initialize Autocomplete ```py from fast_autocomplete import AutoComplete autocomplete = AutoComplete(words=words, synonyms=synonyms) ``` At this point, AutoComplete has created a [dwg](#DWG) structure. Now you can search! - word: the word to return autocomplete results for - max_cost: Maximum Levenshtein edit distance to be considered when calculating results - size: The max number of results to return ```py >>> autocomplete.search(word='2018 bmw 1', max_cost=3, size=3) [['2018', 'bmw'], ['2018', 'bmw 1 series']] ``` Now what if we pressed a by mistake then? It still works. No problem. ```py >>> autocomplete.search(word='2018 bmw 1a', max_cost=3, size=3) [['2018', 'bmw'], ['2018', 'bmw 1 series']] ``` Ok let's search for Alfa now: ```py >>> autocomplete.search(word='alfa', max_cost=3, size=3) [['alfa romeo'], ['alfa romeo 4c'], ['alfa romeo giulia']] ``` What if we don't know how to pronounce alfa and we type `alpha` ? ```py >>> autocomplete.search(word='alpha', max_cost=3, size=3) [['alfa romeo'], ['alfa romeo 4c'], ['alfa romeo giulia']] ``` It still works! Fast-Autocomplete makes sure the results make sense! Ok lets add the word `Los Angeles` there to the words: ```py >>> words['los angeles'] = {} >>> words['in'] = {} >>> autocomplete.search(word='2007 alfa in los', max_cost=3, size=3) [['2007', 'alfa romeo', 'in'], ['2007', 'alfa romeo', 'in', 'los angeles']] ``` So far we have not used the context. And this library leaves it up to you how to use the context. But basically if we giving a context to each one of those words, then the above response could easly be translated to the list of those contexts. ## context If our words dictionary was: ```py words = { 'in': {}, 'alfa romeo': {'type': 'make'}, '2007': {'type': 'year'}, 'los angeles': {'type': 'location'}, } ``` Then the `autocomplete.words` can be used to map the results into their context: ``` [['2007', 'alfa romeo', 'in'], ['2007', 'alfa romeo', 'in', 'los angeles']] converted to contexts: [[{'year': '2007'}, {'make': alfa romeo'}], [{'year': '2007'}, {'make': alfa romeo'}, {'location': 'los angeles'}]] ``` ## Sorting Most people who use Fast Autocomplete, want to control how results are sorted. If you don't control that, the results will be sorted based on the order that Autocomplete found the nodes in the graph that matched the criteria. The easiest way to sort is to give each item a count. **Fast AutoComplete will use the count to sort items that are partial matches.** For example: 1. Make a json file that is a dictionary of words to their context. The format of the file needs to be: ```json { word: [ context, display value, count ] } ``` An example is included in the [sample_words.json](tests/fixtures/sample_words.json) ```json { "acura rlx": [ { "model": "rlx", "make": "acura" }, "Acura RLX", 3132 ], "rlx": [ { "model": "rlx", "make": "acura" }, "Acura RLX", 3132 ], "acura": [ { "make": "acura" }, "Acura", 130123 ], ... } ``` You might be wondering why things are in this format. It is to save space when this json can become very big easily and the keys become repetitive. That's why we are using a list with predefined order of keys. For your use case for now you can leave the context and display values as None if you want. We will opensource other factory functions soon that will fully utilize those keys in the context. 2. Launch Autocomplete via the factory function: ```py from fast_autocomplete import autocomplete_factory content_files = { 'words': { 'filepath': path/to/sample_words.json, 'compress': True # means compress the graph data in memory } } autocomplete = autocomplete_factory(content_files=content_files) ``` 3. You can use Autocomplete and the results are ordered by count! ```py >>> autocomplete.search(word='acu') [['acura'], ['acura mdx'], ['acura rdx']] ``` 4. How do we use the context and display value now? Great question. You need to extend AutoComplete class to use these items. I will write a blog post about it. Here is a simple example without any extending: ```py >>> autocomplete.words['acura'] WordValue(context={'make': 'acura'}, display='Acura', count=130123, original_key=None) >>> autocomplete.words['acura'].display Acura ``` ### Change the sorting by updating counts Fast Autocomplete by default uses the "count" of the items to sort the items in the results. Think about these counts as a "guide" to Fast autocomplete so it can polish its results. Depending on whether or not Fast autocomplete finds exact matches to user's query, the counts will be used to refine the results. You can update the counts in an autocomplete object live. For example, in the [sample csv of car makes and models](tests/fixtures/makes_models_from_wikipedia.csv) we have: ```csv make,model,count Toyota,Aurion,6094 Toyota,Avalon,8803 Toyota,Avensis,1630 Toyota,Auris,4025 Toyota,Aygo,2115 ``` If we use the autocomplete to search: ```py >>> auto_complete = AutoComplete(words=WIKIPEDIA_WORDS, synonyms=SYNONYMS, full_stop_words=['bmw', 'alfa romeo']) >>> autocomplete.search(word='toyota a') [['toyota'], ['toyota avalon'], ['toyota aurion'], ['toyota auris']] ``` However as you can notice `toyota aygo` had the count of 2115 and thus it didn't make it to the top 3 results. We can set the count for `toyota aygo` to a higher number to boost it in the results using `update_count_of_word`. The `update_count_of_word` can change the count via setting the word's count directly or by offsetting its current value. ```py >>> auto_complete = AutoComplete(words=WIKIPEDIA_WORDS, synonyms=SYNONYMS, full_stop_words=['bmw', 'alfa romeo']) >>> auto_complete.update_count_of_word(word='toyota aygo', count=10000) 10000 ``` Now if we search: ```py >>> autocomplete.search(word='toyota a') [['toyota'], ['toyota aygo'], ['toyota avalon'], ['toyota aurion']] ``` We can double check the count of a node: ```py >>> autocomplete.get_count_of_word('toyota aygo') 10000 ``` Now let's use the offset to offset the current count of a different node: ```py >>> auto_complete.update_count_of_word(word='toyota aurion', offset=-6000) 94 ``` When we search, `toyota aurion` is not in the top 3 results anymore! ```py >>> autocomplete.search(word='toyota a') [['toyota'], ['toyota aygo'], ['toyota avalon'], ['toyota auris']] ``` ## Unicode By default this package only accepts ASCII lowercase letters, a-z. However you can pass the characters that you want to be acceptable via `valid_chars_for_string` for strings, and `valid_chars_for_integer` for numbers. For example here we tell Autocomplete to consider the Farsi alphabet characters for string characters. ```python AutoComplete( words=SHORT_WORDS_UNICODE, valid_chars_for_string='اآبپتثجچحخدذرزژسشصضطظعغفقکگلمنوهی') ``` If you want to pass other characters in addition to ASCII letters, such as punctuation marks, you need to set the `valid_chars_for_string` variable to include all of the characters you need. For example, the following code block sets ASCII letters a-z along with periods and apostrophes: ```python valid_chars = ".'" valid_chars += string.ascii_lowercase AutoComplete( words=WORDS_WITH_PUNCTUATION, valid_chars_for_string=valid_chars) ``` ## Draw This package can actually draw the dwgs as it is populating them or just once the dwg is populated for you! Here is the animation of populating the dwg with words from "makes_models_short.csv": ### Draw animation of dwg populating ```py from fast_autocomplete import AutoComplete, DrawGraphMixin class AutoCompleteDraw(DrawGraphMixin, AutoComplete): DRAW_POPULATION_ANIMATION = True DRAW_POPULATION_ANIMATION_PATH = 'animation/short_.svg' DRAW_POPULATION_ANIMATION_FILENO_PADDING = 6 autocomplete = AutoCompleteDraw(words=words, synonyms=synonyms) ``` As soon as you initialize the above AutoCompleteDraw class, it will populate the dwg and generate the animation! For an example of this code properly setup, take a look at the tests. In fact the animation in the [dwg](#dwg) section is generated the same way via unit tests! Note that if you have many words, the graph file will be big. Instead of drawing all frames as the dwg is being populated, you can just draw the final stage: ### Draw the final graph To draw just one graph that shows the final stage of the dwg, use the draw mixin and run the draw_graph function: ```py from fast_autocomplete import AutoComplete, DrawGraphMixin class AutoCompleteDraw(DrawGraphMixin, AutoComplete): pass autocomplete = AutoCompleteDraw(words=words, synonyms=synonyms) autocomplete.draw_graph('path to file') ``` ## Demo If you want to have a real-time interaction with Autocomplete results in your terminal, you can use the demo module: Just pass it an instance of the autocomplete and the search configs: ```py from fast_autocomplete import demo demo(autocomplete, max_cost=3, size=5) ``` # # Develop 1. Clone the repo 2. Make a virtualenv with Python 3.6 or newer 3. `pip install -r requirements-dev.txt` ## Run tests `pytest` We try to maintain high standard in code coverage. Currently the `dwg` module's coverage is around 99%! ## Releases We use bump2version to bump and tag releases. ```bash git checkout master && git pull bump2version {patch|minor|major} git push && git push --tags ``` # Authors - Autocomplete written by [Sep Dehpour](http://zepworks.com). - LFU Cache by [Shane Wang](https://medium.com/@epicshane) # Other ways of doing AutoComplete 1. Elastic search. Yes, Elasticsearch generally is a *better* Autocomplete solution than this library. I said generally. In our specific use case, we wanted Autocomplete to be faster than Elasticsearch and handle combination of words. Otherwise Elasticsearch would have been perfect. Behind the scene Elasticsearch uses Finite State Transducer (FST) in Lucene to achive AutoComplete. FST is more complicated than what we have used in fast-autocomplete. 2. If your autocomplete is supposed to return results based on a big blog of text (for example based on some book contents), then a better solution is to go with Markov chains and conditional probability. Yes, there is already a library out there for it! and it looks great. Disclaimer: we have not actually used it since it doesn't fit our specific use-case. # FAQ ## Why DWG DWG stands for Directed Word Graph. Originally we were using Trie-Tree structure. But soon it was obvious that some branches needed to merge back to other branches. Such as `beemer` and `bmw` branches both need to end in the same node since they are synonyms. Thus we used DWG. ## What are synonyms, clean synonyms and partial synonyms Synonyms are words that should produce the same results. - For example `beemer` and `bmw` should both give you `bmw`. - `alfa` and `alfa romeo` should both give you `alfa romeo` The synonyms get divided into 2 groups: 1. clean synonyms: The 2 words share little or no words. For example `beemer` vs. `bmw`. 2. partial synonyms: One of the 2 words is a substring of the other one. For example `alfa` and `alfa romeo` or `gm` vs. `gmc`. Internally these 2 types of synonyms are treated differently but as a user of the library, you don't need to really care about it. You just provide the synonyms dictionary via defining the `get_synonyms` method. ## Why do you have a whole subtree for partial synonyms Q: Partial synonym means the synonym is a part of the original word. Such as `alfa` is a partial synonym for `alfa romeo`. In that case you are inserting both `alfa` and `alfa romeo` in the dwg. `alfa` will have `alfa 4c` and `alpha romeo` will have `alfa romeo 4c` branches. Why not just have `alfa` branches to be `alfa romeo` and from there you will have automatically all the sub branches of `alfa romeo`. Answer: We use letters for edges. So `alfa` can have only one edge coming out of it that is space (` `). And that edge is going to a node that has sub-branches to `alfa romoe`, `alfa 4c` etc. It can't have a ` ` going to that node and another ` ` going to `alfa romeo`'s immediate child. That way when we are traversing the dwg for the input of `alfa 4` we get to the correct node. ## I put Toyota in the Dawg but when I type `toy`, it doesn't show up. Answer: If you put `Toyota` with capital T in the dwg, it expects the search word to start with capital T too. We suggest that you lower case everything before putting them in dwg. Fast-autocomplete does not automatically do that for you since it assumes the `words` dictionary is what you want to be put in the dwg. It is up to you to clean your own data before putting it in the dwg. %prep %autosetup -n fast-autocomplete-0.9.0 %build %py3_build %install %py3_install install -d -m755 %{buildroot}/%{_pkgdocdir} if [ -d doc ]; then cp -arf doc %{buildroot}/%{_pkgdocdir}; fi if [ -d docs ]; then cp -arf docs %{buildroot}/%{_pkgdocdir}; fi if [ -d example ]; then cp -arf example %{buildroot}/%{_pkgdocdir}; fi if [ -d examples ]; then cp -arf examples %{buildroot}/%{_pkgdocdir}; fi pushd %{buildroot} if [ -d usr/lib ]; then find usr/lib -type f -printf "/%h/%f\n" >> filelist.lst fi if [ -d usr/lib64 ]; then find usr/lib64 -type f -printf "/%h/%f\n" >> filelist.lst fi if [ -d usr/bin ]; then find usr/bin -type f -printf "/%h/%f\n" >> filelist.lst fi if [ -d usr/sbin ]; then find usr/sbin -type f -printf "/%h/%f\n" >> filelist.lst fi touch doclist.lst if [ -d usr/share/man ]; then find usr/share/man -type f -printf "/%h/%f.gz\n" >> doclist.lst fi popd mv %{buildroot}/filelist.lst . mv %{buildroot}/doclist.lst . %files -n python3-fast-autocomplete -f filelist.lst %dir %{python3_sitelib}/* %files help -f doclist.lst %{_docdir}/* %changelog * Wed May 10 2023 Python_Bot - 0.9.0-1 - Package Spec generated