Metadata-Version: 1.1
Name: google-ngram-downloader
Version: 3.0.1
Summary: The streaming access to the Google ngram data.
Home-page: https://github.com/dimazest/google-ngram-downloader
Author: Dmitrijs Milajevs
Author-email: dimazest@gmail.com
License: MIT license
Description: =========================
         Google ngram downloader
        =========================
        
        .. image:: https://travis-ci.org/dimazest/google-ngram-downloader.png?branch=master
            :target: https://travis-ci.org/dimazest/google-ngram-downloader
        
        .. image:: https://coveralls.io/repos/dimazest/google-ngram-downloader/badge.png?branch=master
            :target: https://coveralls.io/r/dimazest/google-ngram-downloader?branch=master
        
        .. image:: https://requires.io/github/dimazest/google-ngram-downloader/requirements.png?branch=master
           :target: https://requires.io/github/dimazest/google-ngram-downloader/requirements/?branch=master
           :alt: Requirements Status
        
        .. image:: https://pypip.in/v/google-ngram-downloader/badge.png
            :target: https://crate.io/packages/google-ngram-downloader/
            :alt: Latest PyPI version
        
        .. image:: https://pypip.in/d/google-ngram-downloader/badge.png
            :target: https://crate.io/packages/google-ngram-downloader/
            :alt: Number of PyPI downloads
        
        `The Google Books Ngram Viewer dataset`__ is a freely available resource under
        a `Creative Commons Attribution 3.0 Unported License`__ which provides ngram
        counts over books scanned by Google.
        
        __ http://storage.googleapis.com/books/ngrams/books/datasetsv2.html
        __ http://creativecommons.org/licenses/by/3.0/
        
        The data is so big, that storing it is almost impossible. However, sometimes
        you need an aggregate data over the dataset. For example to build a
        co-occurrence matrix.
        
        This package provides an iterator over the dataset stored at Google. It
        decompresses the data on the fly and provides you the access to the underlying
        data.
        
        
        Example use
        ===========
        
        >>> from google_ngram_downloader import readline_google_store
        >>>
        >>> fname, url, records = next(readline_google_store(ngram_len=5))
        >>> fname
        'googlebooks-eng-all-5gram-20120701-0.gz'
        >>> url
        'http://storage.googleapis.com/books/ngrams/books/googlebooks-eng-all-5gram-20120701-0.gz'
        >>> next(records)
        Record(ngram=u'0 " A most useful', year=1860, match_count=1, volume_count=1)
        
        
        Installation
        ============
        
        ::
        
            pip install google-ngram-downloader
        
        
        The command line tool
        =====================
        
        It also provides a simple command line tool to download the ngrams called
        `google-ngram-downloader`.
        
        
        
        Changes
        =======
        
        Version 3.0
        -----------
        
        * `download`, `readile` and `cooccurrence` subcommands.
        * `readline_google_store` transforms lines to `Record` in several processes.
        
Platform: UNKNOWN
Classifier: Development Status :: 3 - Alpha
Classifier: Intended Audience :: Developers
Classifier: Intended Audience :: Science/Research
Classifier: License :: OSI Approved :: MIT License
Classifier: Operating System :: POSIX
Classifier: Operating System :: Microsoft :: Windows
Classifier: Operating System :: MacOS :: MacOS X
Classifier: Topic :: Utilities
Classifier: Programming Language :: Python :: 2
Classifier: Programming Language :: Python :: 3
