You cannot select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
40ea2cae20 | 2 years ago | |
---|---|---|
exp.opacity | 2 years ago | |
exp.recommanded | 2 years ago | |
summa | 2 years ago | |
wikipage | 2 years ago | |
README.md | 2 years ago |
README.md
edited summa (textrank)
summa is a textrank python implementation (https://github.com/summanlp/textrank).
it was modified under summa/
, by adding an summa/edits.py
files to create two new function, to access the internal process steps of textrank:
scored_sentences
: gives the list of all the sentences with their score.similarity_graph
: gives the matrix of similarity of all the sentences in a text.
[EXP] opacity
For any wikipedia page, show the text content but where every sentences has an opacity inversely proportional to its TextRank score.
Meaning sentences considered as "relevant to be included in a summary of the article" becomes invisible; and what become visible is what would be considered as the "boring and redundant".
using
- edited summa (https://github.com/summanlp/textrank)
- wikipedia python module (https://pypi.org/project/wikipedia/)
to use
modify the variable wikipedia_page
in make.py
to whatever page then
cd exp.opacity
python3 make.py
technical notes
- headers opacities where manually recomputed has average of their section, this is justified because otherwise their break the flow of the document (their shortness seems to either put them nearly full black or white otherwise, independantly of how textrank rank the paragraphs in their associated sections)
- using the
.content
method of python wikipedia, we get plain text plus header in wikitext, but things like<p>
,<ul>
,<blockquote>
, etc all dissapeared. see if we want to craft a version using the.html
method of python wikipedia, but it becomes more complex because of sentence tokenisation, probably need an index to keep track of their original div nested location. - opacities were remapped to add contrast to their curves. still need to experiment with that to find some kind of nice compromise on both paper and screen ?
[EXP] recommanded
[EXP] custom similarity
technical note
- had to build a
similarity_graph
function to get the matrix of a text. - the computation of those numbers is made in the
_get_similarity
function insummarizer.py
, basically counting the words and dividing them by length of the sentence. the numbers can vary from approx 3.5 to 0 and are not symmetrized or normalized in any way. so it feels that we can input what we want lol - we want to input our own matrices, so we create a
set_graph_custom_edge_weights
andcustom_summarize
.