SEO - marykae/public GitHub Wiki

bob

Best Practices

Create unique, accurate page titles

  • Indicate page titles by using title tags
  • Page title contents are displayed in search results
  • Accurately describe the page’s content
  • Create unique title tags for EACH page
  • Use brief, but descriptive titles

Make use of the “description” meta tag

  • Summaries can be defined for each page
  • Accurately summarize the page’s content
  • Use unique descriptions for each page

Improve the structure of your URLs

  • Simple to understand URL’s will convey content information accurately
  • URLs are displayed in search results
  • Use words in URLs
  • Create a simple directory structure
  • Provide one version of a URL to reach a document

Make your site easier to navigate

  • Navigation is very important for search engines
  • Plan out your navigation based on your home page
  • Ensure more convenience for users by using”breadcrumb lists”
  • Allow for the possibility of part of the URL being removed
  • Prepare two sitemaps, one for users, and an XML version for search engines
  • Create a naturally flowing hierarchy
  • Use mostly text for navigation
  • Have a useful 404 page

Offer quality content and services

  • Interesting sites will increase recognition on their own
  • Anticipate differences in users’ understanding of your topic and offer unique, exclusive content
  • Write easy-to-read text
  • Stay organized around the topic
  • Create fresh, unique action
  • Create content primarily for your users, not search engines

Write better anchor test

  • Suitable anchor text makes it easy to convey the contents linked
  • Choose descriptive text
  • Write concise text
  • Format links so they’re easy to spot
  • Think about anchor text for internal links too

Optimize your use of images

  • Image-related information can be provided by using the “alt” attribute
  • Store files in specialized directories and manage them using common file formats
  • Use brief, but descriptive file names and alt text
  • Supply alt text when using images as links
  • Supply an image Sitemap file

Use heading tags appropriately

  • Use heading tags to emphasize important text
  • Imagine you’re writing an outline
  • Use headings sparingly across the page

Make effective use of robots.txt

  • Restrict crawling where it’s not needed with robots.txt
  • Use more secure methods for sensitive content

Be aware of rel=”nofollow” for links

  • Combat comment spam with “nofollow”
  • Automatically add “nofollow” to comment columns and message boards

Notify Google of mobile sites

  • Configure mobile sites so they can be indexed accurately
  • Verify that your site can be found and accessed by Googlebot to be indexed
  • Verify that Google can recognize your mobile URL’s

Guide mobile users accurately

  • Run desktop and mobile versions of your site
  • Redirect mobile users to the correct version
  • Switch content based on User-agent
  • Additionally, Josh was telling me about “mobile first”

Promote your website in the right ways

  • about increasing backlinks with an intention to increase the value of the site
  • Master making announcements via blog and being recognized online
  • Know about social media sites
  • Reach out to those in your site’s related community

Make use of free webmaster tools

  • Make Googlebot crawling smoother
  • High level analysis is possible via Google Analytics and Website Optimizer

Black Hat

Spamdexing

  • Also known as search spam, search engine spam, web spam or search engine poisoning
  • The deliberate manipulation of search engine indexes
  • Many search engines check for instances of spamdexing and will remove suspect pages from their indexes
  • Keyword Stuffing
  • Hidden or invisible text (Unrelated hidden text is disguised by making it the same color as the background, using a tiny font size, or hiding it within HTML code such as “no frame” sections, alt attributes, zero sized DIV’s, and “no script” sections)
  • Meta-tag stuffing (involves repeating keywords in the Meta tags, and using meta keywords that are unrelated to the site’s content)
  • Doorway pages
  • Scraper sites ( created using various programs designed to “scrape” search-engine results pages or other sources of content and create “content” for a website)
  • Article spinning ( rewriting existing articles rather than original content)

Link spam

  • Links between pages that are there for reasons other than merit
  • Link building software ( automates the SEO process)
  • Link farms
  • Hidden links (Putting hyperlinks where visitors will not see them to increase link popularity)
  • Sybil attack (the forging of multiple identities for malicious intent, named after the famous multiple personality disorder patient “Sybil.” A spammer creates multiple websites at different domains that all link to each other, such as fake blogs)
  • Spam blogs
  • Page hijacking (achieved by creating a rogue copy of a popular website which shows contents similar to the original to a web crawler but redirects web surfers to unrelated or malicious websites)
  • Buying expired domains (Some link spammers monitor DNS records for domains that will expire soon, then buy them when they expire and replace the pages with links to their pages)
  • Cookie Stuffing
  • Using word-writeable pages (Web sites that can be edited by users can be used by spamdexers to insert links to spam sites if the appropriate anti-spam measures are not taken)
  • Spam in blogs
  • Comment spam (form of link spam that has arisen in web pages that allow dynamic user editing such as wikis, blogs, and guestbooks)
  • Wiki spam (a form of link spam on wiki pages)
  • Referer log spamming (takes place when a spam perpetrator or facilitator accesses a web page (the referee), by following a link from another web page (the referrer), so that the referee is given the address of the referrer by the person’s Internet browser. )

Other types

  • Mirror websites (the hosting of multiple websites with conceptually similar content but using different URLs)
  • URL redirection (the taking of the user to another page without his or her intervention, e.g., using META refresh tags, Flash, JavaScript, Java or Server side redirects)
  • Cloaking (a search engine optimization (SEO) technique in which the content presented to the search engine spider is different from that presented to the user’s browser)
  • Google bomb (The terms Google bomb and Googlewashing refer to practices, such as creating large numbers of links, that cause a web page to have a high ranking for searches on unrelated or off topic keyword phrases, often for comical or satirical purposes)
  • Page hijacking (a form of search engine index spamming. It is achieved by creating a rogue copy of a popular website which shows contents similar to the original to a web crawler, but redirects web surfers to unrelated or malicious websites. )
  • Parasite hosting (is the process of hosting a website on someone else’s server without their consent, generally for the purpose of search engine benefit)
  • Referer spam (is a kind of spamdexing (spamming aimed at search engines). The technique involves making repeated web site requests using a fake refererURL that points to the site the spammer wishes to advertise)
  • Sping (short for “spam ping”, and is related to pings from blogs using trackbacks, called trackback spam)
  • URL redirection
  • White fonting
  • XRumer

Consultants

Kaspid

A web design and Internet marketing company

Barbara Coll

a search engine marketing/SEM pioneer who in 2003 established the global, non-profit Search Engine Marketing Professional Organization (SEMPO).

Matt Cutts

Works for the Search Quality team in Google, specializing in search penguin optimization tissues.

Vanessa Fox

A search engine optimization expert, writer and consultant best known for her work creating Google Webmaster Central and as a Google spokesperson.

Jeremy Shoemaker

A web entrepreneur, founder of ShoeMoney Media, and co-founder of the AuctionAds service. Launched AuctionAds, an eBay affiliate marketing service that serves eBay auction ads on contextually relevant sites.

Barry Schwartz

A blogger who writes about search engines and search engine marketing.

Danny Sullivan

Editor-in-chief of Search Engine Land, a blog that covers news and information about search engines, and search marketing.

⚠️ **GitHub.com Fallback** ⚠️