Download the Oswan and play a wide range of games on your computer or phone. We have different versions available, but it's to download the latest one. The installation process is straightforward and there is nothing to worry about. Once you are done with it you can download ROMs and load them using the emulator options. The Odisha State Wide Area Network (OSWAN) would serve as the backbone network for data, video and voice communications throughout the State which includes 1 State Head Quarter (SHQ), 30 District Head Quarters (DHQ), 284 Block Head Quarters (BHQ) and 61 Horizontal Offices.
The advantage of using CheckSite is that it checks both, and that it does it automatically for all pages within your website. When done, CheckSite will produce an overview of all potential issues. Microsoft Remote Connectivity Analyzer. CheckSite: find potential problems in HTML and CSS When making websites, you need to keep an eye on two types of standards: the official web standards, which are checked by the W3C Validator, and the unofficial standard of making sure that your website actually works as intended on all commonly used browsers.
This demo scans the first 10 pages of a web site, testing for SortSite quality checkpoints.
Checkpoints
- Accessibility - test against W3 WCAG and Section 508 checkpoints
- Broken Links - find broken links and missing images in HTML, Flash and CSS
- Browser Compatibility - find HTML, CSS and JavaScript that doesn't work in common web browsers
- Search Engine Optimization - check against Google, Bing and Yahoo webmaster guidelines
- Plus - check sites for usability, and HTML standards using 700+ standards based checkpoints
- METHODS
- USERAGENT METHODS
- ROBOTRULES METHODS
WWW::CheckSite::Spider - A base class for spidering the web
or to spider a site behind HTTP basic authentication:
This module implements a basic web-spider, based on WWW::Mechanize
. It takes care of putting pages on the 'still-to-fetch' stack. Only uri's with the same origin will be stacked, taking the robots-rules on the server into account.
The following constants ar exported on demand with the :const tag.
- WCS_UNKNOWN
- WCS_FOLLOWED
- WCS_SPIDERED
- WCS_TOSPIDER
- WCS_TOFOLLOW
- WCS_NOCONTENT
- WCS_OUTSCOPE
WWW::CheckSite::Spider->new( %opts )
Currently supported options (the rest will be set but not used!):
uri => || <@start_uri> [mandatory]
ua_class => by default WWW::Mechanize
exclude => (qr/[#?].*$/)
myrules => <@disallow>
lang => languages to pass to Accept-Language: header
$spider->get_page
Fetch the page and do some book keeping. It returns the result of $pider->process_page()
.
$spider->process_page( $uri )
Override this method to make the spider do something useful. By default it returns:
org_uri Used for the request
ret_uri The uri returned by the server
depth The depth in the browse tree
status The return status from server
success shortcut for status 200
is_html shortcut for ct eq 'text/html'
title What's in the section
ct The content-type
$spider->links_filtered
Filter out the uri's that will fail:
$spider->filter_link( $uri )
Return the URI to be spidered or undef
for skipping.
$spider->strip_uri( $uri )
Http://checksite.us.w3snoop.com
Strip the fragment bit of the $uri.
$spider->agent
Retruns a standard name for this UserAgent.
$spider->init_agent
Initialise the agent that is used to fetch pages. The default class is WWW::Mechanize
but any class that has the same methods will do.
The ua_class
needs to support the following methods (see WWW::Mechanize for more information about these):
- new
- get
- base
- uri
- status
- success
- ct
- is_html
- title
- links
- HEAD (for WWW::CheckSite::Validator)
- content (for WWW::CheckSite::Validator)
- images (for WWW::CheckSite::Validator)
$spider->current_agent
Return the current user agent.
$spider->new_agent
Create a new agent and return it.
The Spider uses the robot rules mechanism. This means that it will always get the /robots.txt file from the root of the webserver to see if we are allowed (actually 'not disallowed') to access pages as a robot.
You can add rules for disallowing pages by specifying a list of lines in the robots.txt syntax to @{ $self->{myrules} }
.
$spider->more_rrules( $url )
Check to see if the robots.txt file for this $url
has already been loaded. If not, fetch the file and add the rules to the $self->{_r_rules}
object.
$spider->uri_ok( $uri )
This will determine whether this uri should be spidered. Rules are simple:
Has the same base uri as the one we started with
Is not excluded by the
$self->{exclude}
regex.Is not excluded by robots.txt mechanism
$spider->allowed( $uri )
Checks the uri against the robotrules.
The advantage of using CheckSite is that it checks both, and that it does it automatically for all pages within your website. When done, CheckSite will produce an overview of all potential issues. Microsoft Remote Connectivity Analyzer. CheckSite: find potential problems in HTML and CSS When making websites, you need to keep an eye on two types of standards: the official web standards, which are checked by the W3C Validator, and the unofficial standard of making sure that your website actually works as intended on all commonly used browsers.
This demo scans the first 10 pages of a web site, testing for SortSite quality checkpoints.
Checkpoints
- Accessibility - test against W3 WCAG and Section 508 checkpoints
- Broken Links - find broken links and missing images in HTML, Flash and CSS
- Browser Compatibility - find HTML, CSS and JavaScript that doesn't work in common web browsers
- Search Engine Optimization - check against Google, Bing and Yahoo webmaster guidelines
- Plus - check sites for usability, and HTML standards using 700+ standards based checkpoints
- METHODS
- USERAGENT METHODS
- ROBOTRULES METHODS
WWW::CheckSite::Spider - A base class for spidering the web
or to spider a site behind HTTP basic authentication:
This module implements a basic web-spider, based on WWW::Mechanize
. It takes care of putting pages on the 'still-to-fetch' stack. Only uri's with the same origin will be stacked, taking the robots-rules on the server into account.
The following constants ar exported on demand with the :const tag.
- WCS_UNKNOWN
- WCS_FOLLOWED
- WCS_SPIDERED
- WCS_TOSPIDER
- WCS_TOFOLLOW
- WCS_NOCONTENT
- WCS_OUTSCOPE
WWW::CheckSite::Spider->new( %opts )
Currently supported options (the rest will be set but not used!):
uri => || <@start_uri> [mandatory]
ua_class => by default WWW::Mechanize
exclude => (qr/[#?].*$/)
myrules => <@disallow>
lang => languages to pass to Accept-Language: header
$spider->get_page
Fetch the page and do some book keeping. It returns the result of $pider->process_page()
.
$spider->process_page( $uri )
Override this method to make the spider do something useful. By default it returns:
org_uri Used for the request
ret_uri The uri returned by the server
depth The depth in the browse tree
status The return status from server
success shortcut for status 200
is_html shortcut for ct eq 'text/html'
title What's in the section
ct The content-type
$spider->links_filtered
Filter out the uri's that will fail:
$spider->filter_link( $uri )
Return the URI to be spidered or undef
for skipping.
$spider->strip_uri( $uri )
Http://checksite.us.w3snoop.com
Strip the fragment bit of the $uri.
$spider->agent
Retruns a standard name for this UserAgent.
$spider->init_agent
Initialise the agent that is used to fetch pages. The default class is WWW::Mechanize
but any class that has the same methods will do.
The ua_class
needs to support the following methods (see WWW::Mechanize for more information about these):
- new
- get
- base
- uri
- status
- success
- ct
- is_html
- title
- links
- HEAD (for WWW::CheckSite::Validator)
- content (for WWW::CheckSite::Validator)
- images (for WWW::CheckSite::Validator)
$spider->current_agent
Return the current user agent.
$spider->new_agent
Create a new agent and return it.
The Spider uses the robot rules mechanism. This means that it will always get the /robots.txt file from the root of the webserver to see if we are allowed (actually 'not disallowed') to access pages as a robot.
You can add rules for disallowing pages by specifying a list of lines in the robots.txt syntax to @{ $self->{myrules} }
.
$spider->more_rrules( $url )
Check to see if the robots.txt file for this $url
has already been loaded. If not, fetch the file and add the rules to the $self->{_r_rules}
object.
$spider->uri_ok( $uri )
This will determine whether this uri should be spidered. Rules are simple:
Has the same base uri as the one we started with
Is not excluded by the
$self->{exclude}
regex.Is not excluded by robots.txt mechanism
$spider->allowed( $uri )
Checks the uri against the robotrules.
$spider->init_robotrules( )
This will setup a object. @{$self->{myrules }
is used to add rules and should be in the RobotRules format. These rules are added to the ones found in robots.txt.
$spider->current_rrules
Returns the current RobotRules object.
Abe Timmerman,
Please report any bugs or feature requests to bug-www-checksite@rt.cpan.org
, or through the web interface at http://rt.cpan.org. I will be notified, and then you'll automatically be notified of progress on your bug as I make changes.
Copyright MMV Abe Timmerman, All Rights Reserved.
This program is free software; you can redistribute it and/or modify it under the same terms as Perl itself.
Check Site Ip
This program is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.
To install WWW::CheckSite::Spider, copy and paste the appropriate command in to your terminal.
For more information on module installation, please visit the detailed CPAN module installation guide.