Popularity
9.6
Growing
Activity
8.4
Declining
9,828
475
3,547

Description

Gensim is a Python library for topic modelling, document indexing and similarity retrieval with large corpora. Target audience is the natural language processing (NLP) and information retrieval (IR) community.

Code Quality Rank: L3
Programming language: Python
License: GNU Lesser General Public License v2.1 only
Latest version: v3.8.0

gensim alternatives and similar packages

Based on the "Machine Learning" category

Do you think we are missing an alternative of gensim or a related project?

Add another 'Machine Learning' Package

README

gensim – Topic Modelling in Python

Build Status GitHub release Conda-forge Build Wheel DOI Mailing List Gitter Follow

Gensim is a Python library for topic modelling, document indexing and similarity retrieval with large corpora. Target audience is the natural language processing (NLP) and information retrieval (IR) community.

Features

  • All algorithms are memory-independent w.r.t. the corpus size (can process input larger than RAM, streamed, out-of-core),
  • Intuitive interfaces
    • easy to plug in your own input corpus/datastream (trivial streaming API)
    • easy to extend with other Vector Space algorithms (trivial transformation API)
  • Efficient multicore implementations of popular algorithms, such as online Latent Semantic Analysis (LSA/LSI/SVD), Latent Dirichlet Allocation (LDA), Random Projections (RP), Hierarchical Dirichlet Process (HDP) or word2vec deep learning.
  • Distributed computing: can run Latent Semantic Analysis and Latent Dirichlet Allocation on a cluster of computers.
  • Extensive documentation and Jupyter Notebook tutorials.

If this feature list left you scratching your head, you can first read more about the Vector Space Model and unsupervised document analysis on Wikipedia.

Support

Ask open-ended or research questions on the Gensim Mailing List.

Raise bugs on Github but make sure you follow the issue template. Issues that are not bugs or fail to follow the issue template will be closed without inspection.

Installation

This software depends on NumPy and Scipy, two Python packages for scientific computing. You must have them installed prior to installing gensim.

It is also recommended you install a fast BLAS library before installing NumPy. This is optional, but using an optimized BLAS such as ATLAS or OpenBLAS is known to improve performance by as much as an order of magnitude. On OS X, NumPy picks up the BLAS that comes with it automatically, so you don’t need to do anything special.

The simple way to install gensim is:

pip install -U gensim

Or, if you have instead downloaded and unzipped the source tar.gz package, you’d run:

python setup.py test
python setup.py install

For alternative modes of installation (without root privileges, development installation, optional install features), see the documentation.

This version has been tested under Python 2.7, 3.5 and 3.6. Gensim’s github repo is hooked against Travis CI for automated testing on every commit push and pull request. Support for Python 2.6, 3.3 and 3.4 was dropped in gensim 1.0.0. Install gensim 0.13.4 if you must use Python 2.6, 3.3 or 3.4. Support for Python 2.5 was dropped in gensim 0.10.0; install gensim 0.9.1 if you must use Python 2.5).

How come gensim is so fast and memory efficient? Isn’t it pure Python, and isn’t Python slow and greedy?

Many scientific algorithms can be expressed in terms of large matrix operations (see the BLAS note above). Gensim taps into these low-level BLAS libraries, by means of its dependency on NumPy. So while gensim-the-top-level-code is pure Python, it actually executes highly optimized Fortran/C under the hood, including multithreading (if your BLAS is so configured).

Memory-wise, gensim makes heavy use of Python’s built-in generators and iterators for streamed data processing. Memory efficiency was one of gensim’s design goals, and is a central feature of gensim, rather than something bolted on as an afterthought.

Documentation


Adopters

Company Logo Industry Use of Gensim
RARE Technologies [rare](docs/src/readme_images/rare.png) ML & NLP consulting Creators of Gensim – this is us!
Amazon [amazon](docs/src/readme_images/amazon.png) Retail Document similarity.
National Institutes of Health [nih](docs/src/readme_images/nih.png) Health Processing grants and publications with word2vec.
Cisco Security [cisco](docs/src/readme_images/cisco.png) Security Large-scale fraud detection.
Mindseye [mindseye](docs/src/readme_images/mindseye.png) Legal Similarities in legal documents.
Channel 4 [channel4](docs/src/readme_images/channel4.png) Media Recommendation engine.
Talentpair [talent-pair](docs/src/readme_images/talent-pair.png) HR Candidate matching in high-touch recruiting.
Juju [juju](docs/src/readme_images/juju.png) HR Provide non-obvious related job suggestions.
Tailwind [tailwind](docs/src/readme_images/tailwind.png) Media Post interesting and relevant content to Pinterest.
Issuu [issuu](docs/src/readme_images/issuu.png) Media Gensim's LDA module lies at the very core of the analysis we perform on each uploaded publication to figure out what it's all about.
Search Metrics [search-metrics](docs/src/readme_images/search-metrics.png) Content Marketing Gensim word2vec used for entity disambiguation in Search Engine Optimisation.
12K Research [12k](docs/src/readme_images/12k.png) Media Document similarity analysis on media articles.
Stillwater Supercomputing [stillwater](docs/src/readme_images/stillwater.png) Hardware Document comprehension and association with word2vec.
SiteGround [siteground](docs/src/readme_images/siteground.png) Web hosting An ensemble search engine which uses different embeddings models and similarities, including word2vec, WMD, and LDA.
Capital One [capitalone](docs/src/readme_images/capitalone.png) Finance Topic modeling for customer complaints exploration.

Citing gensim

When citing gensim in academic papers and theses, please use this BibTeX entry:

@inproceedings{rehurek_lrec,
      title = {{Software Framework for Topic Modelling with Large Corpora}},
      author = {Radim {\v R}eh{\r u}{\v r}ek and Petr Sojka},
      booktitle = {{Proceedings of the LREC 2010 Workshop on New
           Challenges for NLP Frameworks}},
      pages = {45--50},
      year = 2010,
      month = May,
      day = 22,
      publisher = {ELRA},
      address = {Valletta, Malta},
      note={\url{http://is.muni.cz/publication/884893/en}},
      language={English}
}