Question

WGET Spider Single Page

Asked by: stilliard
Hi all, im currently working on scripts that need to be able to crawl a single web page, and list out the page sizes and load times of each section / component.  E.g. CSS download times, images, javascript files.

It would need to be using the command line and i believe wget to be the best option for actually crawling all page components.

I can see how to use wget --spider to get the page, + all images, css etc from a page, but then it also crawls all internal site links on that page too, and in turn all the parts of that page, which is NOT what i need as i only want it to do this on a single page basis!

Heres what i have so far, but im sure there must be a wget setting for a single page and its components only.

wget -r -nd --spider -o links.txt -np -p http://www.mysite.co.uk/page.html

Cheers for any help.

This question has been solved and asker verified All Experts Exchange premium technology solutions are available to subscription members.

Subscribe now for full access to Experts Exchange and get

Instant Access to this Solution

  • Plus...
  • 30 Day FREE access, no risk, no obligation
  • Collaborate with the world's top tech experts
  • Unlimited access to our exclusive solution database
  • Never be left without tech help again

Subscribe Now

Asked On
2010-07-22 at 04:02:55ID: 26348367
Tags

unix

,

wget

,

spider

,

crawl

,

Linux

,

Unix Command Line

,

Linux Command Line

,

load time

,

page speed

,

page size

,

download size

Topics

Linux

,

Linux Programming

,

Unix Systems Programming

Participating Experts
2
Points
500
Comments
4

Trusted by hundreds of thousands everyday for fast, accurate and reliable tech support.

Companies like CAT, Honeywell, Chevron, Toyoto, TriStar Pictures, the U.S. Army, and Accenture

Your Complete Technology Resource

900+ topic areas (and counting)

Related Solutions

  1. Spider
    Does anyone know where I can get a free script or program that can list info it finds for a given URL or address? Eg. prompt> spider http://where.com/files and have it index and record what it finds at that address and whatever is ...
  2. Blocking a spider
    www.picsearch.com (spider4.picsearch.com ) I presume from the ip address 62.119.133.14 and various ranges is constantly indexing my site. It seems to visit every day and goes through hundreds of pages and I'm even tracking multiple vis...
  3. wget output
    Hello, I am using WGET to invoke a URL. My problem is that the URL is being invoked fine but its always creating a kinda log file every time the URL get invoked. Is there any flag to ignore the creation of these log files ? /usr/loca...
  4. wget problem
    Dear Unix users, I have implemented a restful web service (on a windows server machine - sorry) that accepts post requests and expects in the message body a data entry with actual data. This Webservice works fine when I use javascript/...
  5. wget -O
    Hi! I'm trying to write a bash script to download an XML file periodically. My command looks something like this: /usr/bin/wget http://www.example.com/?download=xml --output-document=file.xml This works fine, and stores the download...

Featured Articles

Get full access to the help you need.

Subscribe Now

30-day free trial. Register in 60 seconds.

360° Features

Go Beyond Just Q and A

Screenshot of Experts Exchange Knowledgebase

Got a question?

We've got the answer.

Experts Exchange has been collecting answers to technology questions since 1996…3 million and counting! If you have a question, chances are we already have your answer.

Screenshot of Experts Exchange Knowledgebase

Need individual assistance?

Our experts are ready to help.

If you can't find the exact answer you're looking for, ask our exclusive community of 50,000 experts. You’ll get a personalized answer from a trusted professional.

Screenshot of an Article

Want to learn from the best?

Read articles from industry experts.

Thousands of free tech tips, tricks, how-to’s and tutorials are available in our peer reviewed articles section. See for yourself how smart our experts are, no login required.

Screenshot of Experts Exchange Knowledgebase

Working on a long term project?

Store your work and research.

Save solutions to your questions, answers you’ve discovered through searching plus helpful articles in your personal knowledgebase for easy future access.

Try it out and discover for yourself.

Subscribe Now

30-day free trial. Register in 60 seconds.

The Latest Technology News and Tips

New Solutions Everyday

Because of our active community of experts, on average new questions receive their first comment in under 30 minutes.

Top Linux Questions

Stay ahead of the curve with Experts Exchange

Subscribe Now

30-day free trial. Register in 60 seconds.

Hear what our users are saying

"The time we save is the biggest benefit of Experts Exchange to Warner Bros. What could take multiple guys 2 hours or more each to find is accessed in around 15 minutes on Experts Exchange."

Mike Kapnisakis, Warner Bros.

"Our team likes having a resource that is more secure than just using Google and most experts using this service really know their stuff. It's nice to look here first versus using Google."

Dayna Sellner, Lockheed Martin

"Anytime that I've been stumped with a problem, 9 out of 10 times Experts Exchange has either the accepted solution or an open discussion of the potential solution to the problem."

Kenny Red, eBay Inc.

Members of the expert community talk about why the experience at Experts Exchange is different than what you will find anywhere else.

Try it out and discover for yourself.

Subscribe Now

30-day free trial. Register in 60 seconds.

Join the Community

Give a little. Get a lot.

Join the community of experts here and help other tech pros by answering question in your area of expertise. You can earn FREE access to all Experts Exchange's premium features and resources.

Join the Community

20121023-EE-VQP-FT-DP

3 Ways to Join

30-Day Free Trial

The Experts

98% positive feedback on 31,087 answers since March 2000. angeliii is a Microsoft Most Valuable Professional for his work with MS SQL Server & Development.

He has also proven his knowledge of Visual Basic Programming, PHP Scripting and Oracle Databases.

The Experts

97% positive feedback on 10,752 answers since July 2000. lrmoore has more than 18 years experience in the networking industry.

The six-time Microsoft MVPs specialties include firewalls, virtual private networking, and network management.

Testimonials

"...an excellent source for support... Kind of like having your very own IT dept." Electriciansnet

Testimonials

"I was apprehensive at signing up at first. However... it has already made my life as an IT administrator much easier." JaCrews

Testimonials

"WOW! You guys have great, active, and knowledgeable people on here." moore50

Business Clients

Business Clients

In the Press

"If you’ve got a question... Experts Exchange can supply an answer.”

In the Press

"...an invaluable aid for both IT professionals and those who require tech support."

In the Press

"where IT professionals provide quick answers on just about any topic"

Business Account Plans