What is a web crawler and what function does it serve in a search engine quizlet?

  1. Science
  2. Computer Science
  3. Information Architecture

  • Flashcards

  • Learn

  • Test

  • Match

  • Flashcards

  • Learn

  • Test

  • Match

Terms in this set (28)

Web directory vs. Crawler based search engine

Crawler Search Engine= rely on sophisticated computer programs called "spiders," "crawlers," or "bots" that surf the Internet, locating webpages, links, and other content that are then stored in the SE's page repository.

Web Directory= categorized listings of webpages created and maintained by humans
- less irrelevant websites

What is the purpose of an index in a search engine?

An index helps search engines efficiently locate relevant pages containing keywords used in a search

Page-ranking method most commonly associated with Google's success

Certain algorithms

In some cases, a search engine may use hundreds of different criteria to determine which pages appear at the top of a SERP. Google, for instance, claims to use over 200 "clues" to determine how it ranks pages

Search Engine Optimization (SEO) vs. Pay Per Click (PPC)

SEO= to improve their website's organic listings on SERPs. No payments are made to the search engine service for organic search listings.

PPC= refers to paid search listings where advertisers pay search engines based on how many people click on the ads.

3 Different Real time search tools

Google Trends= will help you identify current and historical interest in the topic by reporting the volume of search activity over time.

Google Alerts= ) is an automated search tool for monitoring new Web content, news stories, videos, and blog posts about some topic

Twitter Search= You can leverage the crowd of over 650 million Twitter users to find information as well as gauge sentiment on a wide range of topics and issues in real time.

Website content based on Reputation or Popularity, Relevancy, User Satisfaction

Popularity --> based on how many times people click and visit the website
- Websites with good content will be more popular
- # of backlinks (external links that point back to a site)

Relevancy -->
- algorithms used
- Page titles and key words
- relevant phrases in text
- the proportion of relevant text to non relevant text
-Backlinks from relevant sites
- CTR=?
- Onpage factor=?

User Satisfaction -->
- Dwell time: Users that stay on a site longer are probably more satisfied.
- Site speed: Slow page loading time on websites reduces satisfaction.
- Reading level: Reading levels that are too high or too low frustrate users.
- Hacked sites, malware, spam reduce user satisfaction significantly.
- Web satisfaction surveys
- Barriers to content (making people register for things, pop up ads)

Backlinks

A backlinks is an external link that points back to a site

-Google hasn't been using them as much because other factors yield a better picture of relevant sites

black hat SEO tactics

Tricks search engines (like google) into thinking that a website has high quality content when it doesn't

- If google finds out, they are punished and put lower on the list, doesn't show up on SERPs at all

How do organizations evaluate their efficiency and effectiveness of their search engine optimization (SEO)

Google Analytics
-companies can determine how many people visit their site, what specific pages they visit

-Keyword Conversion Rates
- what keywords optimize results

Why is it important to have high quality SEO practices

When SEO practices are combined with valuable content, websites become easier to find in search engines but, more importantly, contribute to building brand awareness, positive attitudes toward the brand, and brand loyalty.

The difference between PPC and Organic listings

Paid advertisements
- receive preferential page placement
- different shading, labels or placement

Organic search results

4 Steps to creating a PPC

1. Set an overall budget for the campaign.
2. Create ads—most search engine ads are text only.
3. Select keywords associated with the campaign.
4. Set up billing account information.

What determines if an ad will appear on search page results (besides bid price)?

In addition to keywords and bid pricing for ads, search engines consider...

- Geographic location
- Time of Day
- Quality Score (related to users experience, ads that are more relevant cost less and run at the top)

How do webpage factors influence the effectiveness of PPC ads?

product ads link --> main page of website, or put item directly in cart

Four metrics used to evaluate the effectiveness of PPC advertising campaign

- Click through Rates (CTR)= useful for evaluating keyword selection

-Keyword Conversion= ?

-Cost of Customer Aquisition (CoCa)= total budget of campaign/ number of customers who purchased

- Return on Advertising Spending (ROAS)= Revenue/Cost

5 ways semantic technology is enhancing the search experience of users

-1. Related searches: (ex: "Did you mean...")
-2. Reference results: dictionary, wikipedia, maps, reviews, stock quotes
-3. Annotated: returns pages containing highlighting of search terms
-4. Full text similarity search: users can submit a full document to find similar content
-5. Sematic/sentatic annotations

(more listed on slides)

How do metadata tags facilitate more accurate search results

Data is tagged describing the nature of the information, where it came from, and how it is arranged

- makes the web a giant database that can be read and used by a variety of applications
- makes it possible to access all the information we can on the internet

3 evolutionary stages of the internet

1. The Initial Web
- pages or documents are hyperlinked making it easy to access information

2. The Social Web
- a web of applications

3. A Semantic Web
- documents with metadata tags allowing more information to be shared regardless of the original structure of the file

Context

defines the intent of the user; for example, trying to purchase music, to find a job, to share memories with friends and family

Personalization

refers to the user's personal characteristics that impact how relevant the content, commerce, and community are to an individual.

Vertical Search

as you have read, focuses on finding information in a particular content area, such as travel, finance, legal, and medical.

3 languages developed by the W3C and associated with the semantic web

The semantic Web utilizes additional languages that have been developed by the W3C. These include resource description framework (RDF), Web ontology language (OWL), and SPARQL protocol and RDF query language (SPARQL).

How is a recommendation engine different from a search engine

Search engine- customers find products through an active search, assuming customers know what they want and how to describe it when forming their search query.

Recommendation engines- proactively identify products that have a high probability of being something the consumer might want to buy.

Examples of ways recommendation engines are being used (besides e-commerce)

Netflix does recommendations of movies for customers similar to movies they already have watched.
Pandora creates recommendations or playlists based on song attributes.

Collaborative filtering and recommendation engines

- purchase history
- consumer behavior
- ratings
-reviews
- media use

Content based filtering, what does a company need to know about its products

Content-based filtering recommends products based on the product features of items the customer has interacted with in the past and the similarity to other products' features.

4 methods used by recommendation systems

- Content filtering
- Collaborative filtering
- Knowledge based system
- Demographic System

Hybrid recommendation systems collaborate all 4

Mixed, Cascade, and Compound Hybrids

-Mixed hybrid: Results from different recommenders are presented alongside of each other.

-Cascade hybrid: Recommenders are assigned a rank or priority. If a tie occurs (with two products assigned the same recommendation value), results from the lower-ranked systems are used to break ties from the higher-ranked systems.

-Compound hybrid: This approach combines results from two recommender systems from the same technique category (e.g., two collaborative filters), but uses different algorithms or calculation procedures.

Sets found in the same folder

MGSC 300 - Chapter 3

40 terms

jcbswartz8

MGSC 300 - Chapter 6

43 terms

maddisal2

MGSC 300 - Chapter 8

44 terms

maddisal2

MGSC 300 Chp 11

33 terms

maddisal2

Other sets by this creator

Word of Mouth & Post-Decision Processes

21 terms

elepetsos

Judgement and Decision Making - Consumer Behavior…

28 terms

elepetsos

Sales Management Midterm

65 terms

elepetsos

Reading Guide A Psychology of Sexuality David Fred…

22 terms

elepetsos

Verified questions

physics

"A capacitor has parallel plates of area $12 \mathrm{~cm}^2$ separated by $2.0 \mathrm{~mm}$. The space between the plates is filled with polystyrene. $(a)$ Find the maximum permissible voltage across the capacitor to avoid dielectric breakdown. $(b)$ When the voltage equals the value found in part $(a)$, find the surface charge density on the surface of the dielectric."

Verified answer

biology

Certain endocrine disorders, such as Cushing syndrome, can be caused by excessive secretion of a hormone (in this case, ACTH) by the pituitary gland or by a problem with the endocrine gland itself (in this case, the adrenals). If you determined the ACTH levels of a Cushing patient, how could you tell the difference between a pituitary problem and a primary adrenal problem?

Verified answer

chemistry

Write the electron configuration for $\ce{Ru^2+}$ ion.

Verified answer

physics

Construct a spreadsheet (or other numerical tool) that will reproduce Fig. 41-1, the graph of binding energy per nucleon (in $\mathrm{MeV}$ ) vs. the mass number $A$. Using Appendix F, calculate the binding energy per nucleon for the most stable isotope of each possible mass number $A \geq 2$. [The first few values will be for ${ }_1^2 \mathrm{H},{ }_2^3 \mathrm{He}$ (it is more stable than ${ }_1^3 \mathrm{H}$ ), ${ }_2^4 \mathrm{He},{ }_3^6 \mathrm{Li}$, and ${ }_3^7 \mathrm{Li}$ (since it is more stable than ${ }_4^7 \mathrm{Be}$ ).] To reduce the amount of data, for $A \geq 20$ plot only points for even values of $A$, and plot to a maximum of $A=142$

Verified answer

Recommended textbook solutions

What is a web crawler and what function does it serve in a search engine quizlet?

Information Technology Project Management: Providing Measurable Organizational Value

5th EditionJack T. Marchewka

346 solutions

What is a web crawler and what function does it serve in a search engine quizlet?

Introduction to Algorithms

3rd EditionCharles E. Leiserson, Clifford Stein, Ronald L. Rivest, Thomas H. Cormen

726 solutions

What is a web crawler and what function does it serve in a search engine quizlet?

Computer Organization and Design MIPS Edition: The Hardware/Software Interface

5th EditionDavid A. Patterson, John L. Hennessy

220 solutions

What is a web crawler and what function does it serve in a search engine quizlet?

Starting Out with C++ from Control Structures to Objects

8th EditionGodfrey Muganda, Judy Walters, Tony Gaddis

1,294 solutions

Other Quizlet sets

Rtirements

14 terms

Kfrancette

ACLAM BMBL

74 terms

emkatz86Plus

Nur-320 (Class 10: Nursing Care of the Normal Newb…

34 terms

Nisoni

Methods Quiz 1

50 terms

Mrs_LiebesmanPlus

What is a web crawler quizlet?

Tap the card to flip 👆 Definition. 1 / 34. A web crawler is a software program which browses the World Wide Web in a methodical and automated manner. It collects documents by recursively fetching links from a set of starting pages.

What is the role of web crawlers?

Web crawlers systematically browse webpages to learn what each page on the website is about, so this information can be indexed, updated and retrieved when a user makes a search query. Other websites use web crawling bots while updating their own web content.

What is a web crawler also called?

A web crawler, or spider, is a type of bot that is typically operated by search engines like Google and Bing. Their purpose is to index the content of websites all across the Internet so that those websites can appear in search engine results.

What is crawler based search engine?

Crawlers. These types of search engines use a "spider" or a "crawler" to search the Internet. The crawler digs through individual web pages, pulls out keywords and then adds the pages to the search engine's database. Google and Yahoo are examples of crawler search engines.