Seo CH1
Seo CH1
to how Google
works.
Introduction
• In its infancy, the Internet wasn’t what you think of when you use it now. In
fact, it was nothing like the web of interconnected sites that has become
one of the greatest business facilitators of our time.
• To find a specific file in that collection, users had to navigate through each
file. Sure, there were shortcuts. If you knew the right people — that would
be the people who knew the exact address of the file you were looking for
— you could go straight to the file. That’s assuming you knew exactly
what you were looking for.
History of Search Engines
•The whole process made finding files on the Internet a
difficult, time-consuming exercise in patience; but that was
before a student at McGill University in Montreal decided
there had to be an easier way.
•Search engine is a piece of software that uses algorithms to find and collect
information about web pages. The information collected is usually keywords or
phrases that are possible indicators of what is contained on the web page as a
whole, the URL of the page, the code that makes up the page, and links into and out
of the page. That information is then indexed and stored in a database.
•On the front end, the software has a user interface where users enter a search term
— a word or phrase — in an attempt to find specific information. When the user
clicks a search button, an algorithm then examines the information stored in the
back-end database and retrieves links to web pages that appear to match the search
term the user entered.
Anatomy of a Search Engine
1. Query interface
•The query interface is what most people
are familiar with, and it’s probably what
comes to mind when you hear the term
‘‘search engine.’’ The query interface is the
page, or user interface, that users see
when they navigate to a search engine to
enter a search term.
How about Yahoo (portal)
Anatomy of a Search Engine
2. Search engine results pages
•The other sides of the query interface, and the only other parts of a search
engine that’s visible to users, are the search engine results pages
(SERPs). This is the collection of pages that are returned with search
results after a user enters a search term or phrase and clicks the Search
button. This is also where you ultimately want to end up; and the higher
you are in the search results, the more traffic you can expect to generate
from search. Specifically, your goal is to end up on the first page of results
— in the top 10 or 20 results that are returned for a given search term or
phrase.
Anatomy of a Search Engine
2. Search engine results pages
•Based on previous exercise, What’s the first thing you do when the page
appears?
•There is no magic bullet or formula that will garner you those rankings
every time. Instead, it takes hard work and consistent effort to push your
site as high as possible in SERPs. At the risk of sounding repetitive, that’s
the information you’ll find moving forward. There’s a lot of it, though, and to
truly understand how to land good placement in SERPs, you really need to
understand how search engines work. There is much more to them than
what users see.
Anatomy of a Search Engine
3. Crawlers, spiders, and
robots
• The query interface and search results pages truly are the only parts of
a search engine that the user ever sees.(Front end)
• In fact, what’s in the back end is the most important part of the search
engine, and it’s what determines how you show up in the front end.
• Spiders, crawlers, and robots are programs that literally crawl around
the Web, cataloguing data so that it can be searched. In the most basic
sense, all three programs — crawlers, spiders, and robots — are
essentially the same. They all collect information about each and every
web URL.
Anatomy of a Search Engine
3.Crawlers, spiders, and robots
• As discussed in the previous slide, that back end of search engine consist
of 3 main parts. Search engine spiders follow links on the web to request
pages that are either not yet indexed or have been updated since they
were last indexed. These pages are crawled and are added to the search
engine index (also known as the catalogue). When you search using a
major search engine you are not actually searching the web, but are
searching a slightly outdated index of content which roughly
represents the content of the web. The third part of a search engine is
Robots which Perform spider and crawlers Actions.
Anatomy of a Search Engine
3. Crawlers, spiders, and
robots
• •In other words , the robot is able to find the site when the end user type
a word or phrase. This step is called query processor.
• •For more details please watch this video
• •https://www.youtube.com/watch?v=LVV_93mBfSU
Anatomy of a Search Engine
4. Databases
• Every search engine contains or is connected to a system of databases
where data about each URL on the Web (collected by crawlers, spiders,
or robots) is stored. These databases are massive storage areas that
contain multiple data points about each URL.
• The data might be arranged in any number of different ways and is
ranked according to a method of ranking and retrieval that is usually
proprietary to the company that owns the search engine.
• You’ve probably heard of the method of ranking called PageRank (for
Google) or even the more generic term quality scoring.
• This ranking or scoring determination is one of the most complex and
secretive parts of SEO. Why?
Anatomy of a Search Engine
4. Databases
• Keywords are still vitally important in web page ranking. However, they’re just
one of the dozens of considered elements, BUT Keywords are still the most
important elements of SEO.
• Simply put, everybody wants to be in Google. Google is fighting to keep its
search engine relevant and must constantly evolve to continue delivering
relevant results to users.
• This hasn't been without its challenges. Just like keyword stuffing, webmasters
eventually clued onto another way of gaming the system by having the most
'anchor text' pointing to the page. If you are unfamiliar with this term, anchor
text is the text in external links pointing to a page.
Anchor Text Example
Recent Google updates and how to
survive them.
• Anchor text created another loophole exploited by spammers. In many
cases, well-meaning marketers and business owners used this tactic to
achieve high rankings in the search results.
• Along came a new Google update in 2012, this time called 'Penguin'.
Google’s Penguin update punished sites with suspicious amounts of links
with exact matched anchor text pointing to a page, by completely delisting
sites from the search results.
• Many businesses that relied on search engine traffic lost all of their sales
literally overnight, just because Google believed sites with hundreds of links
containing just one phrase didn't acquire those links naturally
How to recover from Google
changes, or to prevent being
penalized by new updates
• There are many techniques you should consider it to increase your
website rank on SE. Google says :
• If you want to stay at the top of Google, never rely on one tactic.
• Always ensure your search engine strategies rely on SEO best
practices.
Authority, trust & relevance. Three
powerful SEO strategies explained.
• Fortunately, there are a handful of industry leaders who have figured it out, and
regularly publish their findings on the Internet. With these publications you can
get a working knowledge of what factors Google use to rank sites. These surveys
are typically updated every second year, but these factors don’t change often, so
you can use them to your advantage by knowing which areas to focus on.
Google’s top-10 ranking factors
• If your competitors’ pages have more of the above than yours, then it's likely they are
going to rank higher. If your pages have more of the above than competitors, then it
is likely you will beat them.
• The mentioned factors are from the Search Metrics Google Ranking Factors study
released in 2015
• •1.Word count.
• 2.Relevant keywords on page.
• 3.Responsive design.
• 4.User signals (click-through-rate, time-on-site, bounce-rate).
• 5.Domain SEO visibility (how strong the domain is in terms of links and authority).
• 6.Site speed.
• 7.Referring domains (number of sites linking to your site).
• 8.Keyword in internal links.
• 9.Content readability.
• 10.Number of images.
Google’s top-10 ranking factors
(2019)
• The "/robots.txt" file is a text file, with one or more records. Usually
contains a single record looking like this:
• User-agent: [Required, one or more per group] The name of a search
engine robot (web crawler software) that the rule applies to. This is the
first line for any rule. Most Google user-agent names are listed in the
Web Robots Database or in the Google list of user agents. Supports
the asterisk (*) wildcard for a path prefix, suffix, or entire string.
What to put in robots.txt
• # Example 1: Block only Googlebot
• User-agent: Googlebot
• Disallow: /
• # Example 2: Block Googlebot and Adsbot
• User-agent: Googlebot
• User-agent: AdsBot-Google
• Disallow: /
• # Example 3: Block all AdsBot crawlers
• User-agent: *
• Disallow: /
What to put in robots.txt
• Disallow: [At least one or more Disallow or Allow
entries per rule] A directory or page, relative to the root
domain, that should not be crawled by the user agent.