Archive

Archive for the 'HTML / CSS'

WordPress 2.5.1

April 28, 2008 No Comments

Release, as expected, after about a month of a version of correcting various bug fixes and security patches found after the long work for the 2.5 branch!

Usual upgrade procedure and usually link for downloading WordPress !

smileys

WordPress: get_links deprecated (also wp_get_links)

May 17, 2007 3 comments

Upgrading WordPress from 2.1.3 to 2.2 convinced that everything would be successful, and so it was ... or almost!

The theme that I used at the bottom of the page includes the insertion of my blogroll divided into two columns (respectively a category each).

The code that allowed this has always been get_links (arguments), and precisely this:

', '

',' ', FALSE,' name ', FALSE, 1,
-1, 1); ?>

but alas, using this old command get_links was created a strange space between each item and the other, thus making the output painful!

After smadonnato a bit 'in all languages ​​in search of information about on similar issue, I discovered that get_links as wp_get_links are considered deprecated or older, and should be used in their place a new command: wp_list_bookmarks (arguments).

Armed with good patience I am reading the documentation on the website of WordPress up to find that you have to use (in my case) this syntax:

category_name=0&title_li=&before= &after=

  • category_name=0&title_li=&before= &after=
  • &
    show_images = 1 & show_description = 0 & orderby = name '); ?>

    Rimando for more information about using wp_list_bookmarks the official documentation .

    Tag: Categories: HowTo , HTML / CSS , Wordpress Tags:

    Sitemap for better placement in search engines!

    May 15, 2007 5 comments

    Maybe not all webmasters are aware that the sitemap (XML, ROR or TXT) is a powerful tool that allows you to increase the presence of your site in search engines. Thanks to sitemap engine crawlers are easier site pages and index more correctly.

    A web tool that facilitates the creation of sitemap is definitely XML-Sitemaps.com that with a simple click allows you to create your own sitemap.

    The types of maps are available XML (for Google), Text (for Yahoo), ROR (for other search engines) and HTML (to be integrated within the site for easier navigation for users).

    However agrees, WordPress, to refer to a very interesting plugin that automatically regenerates its sitemap every time you change or insert a new article in the blog! This plugin is called Google Sitemaps .

    The output you can see it here (after applying a style sheet): sitemap.xml !

    The site will now be visited more frequently by search engines and thus gain more popularity!

    Tag: Categories: HowTo , HTML / CSS , Miscellaneous , Wordpress Tags:

    Robots.txt and bots not poke around more!

    May 15, 2007 4 comments

    You have just created a new website? Do not forget to include your robots.txt in your root directory.

    What is the robots.txt file?
    (I quote verbatim from Wikipedia)

    The robots.txt file contains settings used by the crawler (also called spiders) to apply restrictions analysis on the pages of a website.
    The web crawlers are automated software programmed to search and indexing periodicals. In the first phase of analysis of a website control the existence of the robots.txt file to apply any restrictions required by the site's webmaster.


    The fields available are two:

    1. User-Agent, the value of this field contains the name of the robot that must comply with the restrictions. With the character '*' the rule is applied to any robot;
    2. Disallow, the value of this field contains the pages of the site that should be excluded from the robot during indexing. You can specify a specific URL or set of URLs belonging to a pattern.


    HERE you can see an example of a file robots.txt

    An example of writing can be this:
    User-agent: Googlebot-Image
    Disallow: /

    where we tell the bot "Googlebot-Image", which scours the web in search of images, not to visit the whole root of our site ... and move on!

    Or again:
    User-agent: * # Applica a tutti i bot
    Disallow: /private_directory/ # Blocca /private_directory/
    Request-rate: 1/5 # Visita al massimo una pagina ogni 5 secondi
    Visit-time: 0700-0845 # Vista soo tra 7:00 AM e 8:45 AM UT (GMT)


    A face wrote the robots.txt file, simply save it to the root directory of the site and the bot will learn not to look around more in our site!

    And now not just check your spelling of the "code" of our robots.txt using the validator web .

    Tag: Categories: HowTo , HTML / CSS , Miscellaneous Tags:

    CSSEdit: style sheets in a few clicks!

    May 15, 2007 No Comments

    At the suggestion of Dade` I also tried this beautiful and handy software called CSSEdit .

    It is characterized by a innanziatutto graphica clean and very intuitive, but especially for its ability to give way to those who have a thorough knowledge of CSS to modify at all the details a stylesheet.

    You enter the url of the site you want to change and you find yourself immediately place before the CSS written ... then we can intervene with disarming simplicity of all that created and then continue the style sheet with sempici click! Is software that simplifies the work of the webmaster because each change is shown in real time in the preview screen (especially useful when working on the margin and on the various containers posizionamente "div")!

    Here's a screenshot of the program in action ...

    google_analytics
    [Click on the image to enlarge]


    Tag: Categories: Curiosity , HTML / CSS , Software Tags: