Skip to content

Commit

Permalink
final release
Browse files Browse the repository at this point in the history
  • Loading branch information
huseinzol05 committed Aug 7, 2019
1 parent 521d010 commit 8819656
Show file tree
Hide file tree
Showing 215 changed files with 36,486 additions and 42,780 deletions.
14 changes: 6 additions & 8 deletions README.rst
Original file line number Diff line number Diff line change
Expand Up @@ -46,14 +46,12 @@ Features
- **Entities Recognition**

Latest state-of-art CRF deep learning and BERT models to do Naming Entity Recognition.

- **Language Detection**

using Multinomial, SGD, XGB, Fast-text N-grams deep learning to distinguish Malay, English, and Indonesian.
- **Normalizer**

using local Malaysia NLP researches to normalize any
bahasa texts.
using local Malaysia NLP researches to normalize any bahasa texts.
- **Num2Word**

Convert from numbers to cardinal or ordinal representation.
Expand Down Expand Up @@ -83,13 +81,13 @@ Features
From fine-tuning BERT, Attention-Recurrent model, and Self-Attention to build deep subjectivity analysis models.
- **Similarity**

Use deep LSTM siamese, deep Dilated CNN siamese, deep Self-Attention, siamese, Doc2Vec and BERT to build deep semantic similarity models.
Use deep Encoder, Doc2Vec and BERT to build deep semantic similarity models.
- **Summarization**

Using skip-thought and residual-network with attention state-of-art, LDA, LSA and Doc2Vec to give precise unsupervised summarization, and TextRank as scoring algorithm.
Using BERT, XLNET, skip-thought, LDA, LSA and Doc2Vec to give precise unsupervised summarization, and TextRank as scoring algorithm.
- **Topic Modelling**

Provide Transformer, LDA2Vec, LDA, NMF and LSA interface for easy topic modelling with topics visualization.
Provide Attention, LDA2Vec, LDA, NMF and LSA interface for easy topic modelling with topics visualization.
- **Toxicity Analysis**

From fine-tuning BERT, Attention-Recurrent model, Self-Attention to build deep toxicity analysis models.
Expand Down Expand Up @@ -123,12 +121,12 @@ If you use our software for research, please cite:
Acknowledgement
----------------

Thanks to `Im Big <https://www.facebook.com/imbigofficial/>`_, `LigBlou <https://www.facebook.com/ligblou>`_ and `Mesolitica <https://mesolitica.com/>`_ for sponsoring AWS and Google cloud to train Malaya models.
Thanks to `Im Big <https://www.facebook.com/imbigofficial/>`_, `LigBlou <https://www.facebook.com/ligblou>`_, `Mesolitica <https://mesolitica.com/>`_ and `KeyReply <https://www.keyreply.com/>`_ for sponsoring AWS and Google cloud to train Malaya models.

.. raw:: html

<a href="#readme">
<img alt="logo" width="30%" src="https://malaya-dataset.s3-ap-southeast-1.amazonaws.com/ligblou-mesolitica.png">
<img alt="logo" width="50%" src="https://malaya-dataset.s3-ap-southeast-1.amazonaws.com/ligblou-mesolitca-keyreply.png">
</a>

Contributing
Expand Down
Binary file modified accuracy/emotion-accuracy.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
9 changes: 5 additions & 4 deletions accuracy/emotion-template.js
Original file line number Diff line number Diff line change
Expand Up @@ -5,17 +5,18 @@ option = {
interval: 0,
rotate: 30
},
data: ['bahdanau','fast-text-char', 'luong', 'multinomial',
'self-attention', 'xgboost', 'BERT']
data: ['bahdanau','luong', 'multinomial',
'self-attention', 'xgboost', 'bert-multilanguage',
'bert-base','bert-small']
},
yAxis: {
type: 'value',
min:0.75,
min:0.71,
max:0.89
},
backgroundColor:'rgb(252,252,252)',
series: [{
data: [0.86, 0.82, 0.86, 0.76, 0.83, 0.82, 0.88],
data: [0.85, 0.85, 0.72, 0.83, 0.82, 0.87, 0.87, 0.87],
type: 'bar',
label: {
normal: {
Expand Down
799 changes: 509 additions & 290 deletions accuracy/models-accuracy.ipynb

Large diffs are not rendered by default.

Loading

0 comments on commit 8819656

Please sign in to comment.