I try to retrieve an article from a public web site “cell” which is a science review on Biology.
Most of text has disappeared : I got one schema and list of references but the body of the article is missing. The URL is:
Is there anything I can try to improve ?
Thanks & best regards Jean-Louis
Normally this could be done withe the browser plugin you find at Push to Kindle Installation. Desktop browser only!
In this case, additional tweaks need to be made on P2K server side. Main Problem is, that I am only a normal user and have no direct access to P2K server. I try to make site depended configs for ‘wallabag’ and hope that they work for P2K as well, which does most times.
But your article is very, very long, so wallabag run into a timeout while trying to process. Could you provide me with an URL of a significant shorter article on cell.com?
Maybe I can look into that issue on the weekend.
Good afternoon HolgerAusB,
Thanks for your reply.
Yes indeed they are quite long. I have here some shorter ones having same issue:
Is there any way to tell “Push to Kindle”, to start working at a given word e.g. “Summary” and to stop at a given word e.g. “Acknowledgment” to improve result ?
What I have tried is to click on “pastepad” to open an editor, then copy/paste in it the content of the browser I want to see on my kindle. Then remove some silly text corresponding to references.
This works quite fine except that all diagrams & figures are lost.
Have a nice day.
With Best Regards Jean-Louis
Yes, that is the way it works. But not by choice of the user. The P2K server tries to predict the relevant parts without any site depended config. In about 50 percent of websites that works in a more or less usable way.
Most times a configfile delivers better results and sometimes we don’t get any result without that config.
Just to be clear: The decision as to what can and cannot be seen from an article is made on the server and always applies to all users and all products.
Removing the citations and cross-references would make the catch smaller, but scientists, journalists or book authors etc would probably miss them.
I’ll try to write a config at the weekend. But cell is using cloudflare to prevent crawlers to catch their content, which makes things difficult. No promisses.
Good morning HolgerAusB
Thanks for your reply.
I have read some examples in github such as “wsj.com.text” but this is not obvious!
When using “print edit we” extension of Firefox, I select the frame containing the body of the main text then click on “delete except” and all surrounding frames disappear in 2 clicks. Is there something like “unstrip” or “keep only” in the config files ?
Issue : when loading “push to kindle” to export this work, I am requested and forced to close the page and all manual editing is lost.
So I copy/paste to another page or save in HTML to relaunch…before using “Push to Kindle”.
Have a nice week…
With Best regards Jean-Louis
that’s the way it works. With
body: <xpath> we select the part of the page we want to have. And then strip unwanted elements within this body. Here you find a small introduction.
I just made a config for
cell.com. But as I wrote before: as ‘cell’ is using cloudflares anti-bot-service, you need to use the Pushtokindle browser plugin for a desktop browser. So try again.
Unfortunately the big article from your first post doesn’t contain any images - and I don’t know why. For the two shorter articles, P2K plugin fetches all images.
You will be facing some ugly newlines prior to every citation footnote. I could not manage to have these footnotes in the same line as the preceding text. I could remove the footnotes in total, but that would not help other users (students, scientists), who may need this, when archiving the article e.g. with wallabag.
Maybe @fivefilters has some ideas about the missing images from https://www.cell.com/cell/fulltext/S0092-8674(22)01377-0 or about the citation footnotes?
Thanks for your cell.com file.
I am indeed using “push to kindle” Firefox plugin and I get the images.
The high number of newlines each time there is a citation make it difficult to read.
So even if I agree that references are useful in general, having them this way is useless : I think nobody can use it like this. So I think it would be better to drop all of them, if you agree (and easy to implement).
With Best regards Jean-Louis
Thanks for the site config, @HolgerAusB! I had a look at the footnote issue and my suggestion would be to try string replacement to change the block element
<div> to an inline element
<span>. I haven’t tested this, but I’m thinking something like:
replace_string(<div class="dropBlock reference-citations"><sup>):<span><sup>
@fivefilters I already tried similar replacements, which I unfortunately deleted from my config.
The problem is, your and mine replacement only works for single footnotes but not if there are two and more comma-seperated footnotes 2, 3, 4
I tested that with wallabag and I think as an archive software, wallabag should keep these footnotes, the config has to work there. Main problem there, is that wallabag is auto-stripping span-elements.
Your replace-code is also cutting the article of first test_url after footnote 5 (in wallabag)
@fivefilters I found a dirty way for wallabag (self-hosters only) now. I set a custom.css in my wallabag:
display: inline; ## or ##
and then I added that class to the div via config-rule:
replace_string(<div class="dropBlock reference-citations">):<div class="dropBlock reference-citations nolinebreak">
I don’t know the difference between inline and inline-block but both versions worked here.
So, is there any class in P2K’s css which does not much more than ‘display: inline[-block]’ which we could inject by rule. Or could you set such a css-rule, please?