How can I open a url and output the processed DOM to a file?
Similar to TagSoup, JTidy, and JSoup I want to have a well formed version of a webpage as processed by firefox.
Do I need to make a plugin to hook into this?
Additional System Details
- Picasa plugin
- MozPlugger version 1.14.3, maintained by Louis Bavoil and Peter Leese, a fork of plugger written by Fredrik Hübinette.For documentation on how to configure mozplugger, check the man page. (type man mozplugger) Configuration file:/etc/mozpluggerrc Helper binary:mozplugger-helper Controller binary:mozplugger-controller Link launcher binary:mozplugger-linker Debug file:$HOME/tmp/mozdebug
- The Totem 2.32.0 plugin handles video and audio streams.
- DivX Web Player version 126.96.36.199
- The IcedTea-Web Plugin executes Java applets.
- Shockwave Flash 10.3 r183
- User Agent: Mozilla/5.0 (X11; Linux i686; rv:5.0) Gecko/20100101 Firefox/5.0
You can do that via the DOM Inspector > File > Save DOM as
That's great. Do you know how I can load a url and save the dom from the command line or an API ?
No. You will have to load the website (you can do that by adding the link to the command line) and launch the DOM Inspector.
Ya I saw that.. I found this python/qt/webkit thing to do it
And this project seems like it can make firefox do it: http://seleniumhq.org/