En webDelFuturo su página web es posicionada

en el mercado de su competencia

Fusionamos su estrategia web con redes sociales

más importantes para su negocio

Toda web es importante por lo que cada página 

tiene todos los beneficios y ventajas que ofrecemos

En webDelFuturo nos interesa presentar la información más actual y de importancia relacionada con los servicios que prestamos por lo que a continuación presentamos diferentes fuentes de consulta sobre diferentes temas de interés. Cada una de nuestra fuentes de información proviene de sitios web de alto renombre en el plano tecnológico actual. Si es de su interés podríamos incluir noticias de más fuentes de información de acuerdo a sus necesidades.

 

Google Webmaster Central Blog

Official news on crawling and indexing sites for the Google index.

Connect to job seekers with Google Search

July 20, 2017 update: Starting today, impressions and clicks stats for job listing pages and job details pages are available in the Search Analytics report in Search Console. Read more about how Jobs impressions and clicks are counted in the help centre. If you have questions, head to the webmaster forums.


At Google I/O this year, we announced Google for Jobs, a new company-wide initiative focused on helping both job seekers and employers, through collaboration with the job matching industry. One major part of this effort is launching an improved experience for job seekers on Google Search. We’re happy to announce this new experience is now open for all developers and site owners.
For queries with clear intent like [head of catering jobs in nyc] or [entry level jobs in DC], we’ll show a job listings preview, and each job can expand to display comprehensive details about the listing:
For employers or site owners with job content, this feature brings many benefits:
  • Prominent place in Search results: your postings are eligible to be displayed in the in the new job search feature on Google, featuring your logo, reviews, ratings, and job details.
  • More, motivated applicants: job seekers can filter by various criteria like location or job title, meaning you’re more likely to get applicants who are looking exactly for that job.
  • Increased chances of discovery and conversion: job seekers will have a new avenue to interact with your postings and click through to your site.

Get your job listings on Google

Implementation involves two steps:
  1. Mark up your job listings with Job Posting structured data.
  2. Submit a sitemap (or an RSS or Atom feed) with a <lastmod> date for each listing.

If you have more than 100,000 job postings or more than 10,000 changes per day, you can express interest to use the High Change Rate feature.
If you already publish your job openings on another site like LinkedIn, Monster, DirectEmployers, CareerBuilder, Glassdoor, and Facebook, they are eligible to appear in the feature as well.
Job search is an enriched search experience. We’ve created a dedicated guide to help you understand how Google ranking works for enriched search and practices for improving your presence

Keep track of how you’re doing and fix issues

There’s a suite of tools to help you with the implementation:

In the coming weeks, we’ll add new job listings filters in the Search Analytics report in Search Console, so you can track clicks and impressions for your listings.
As always, if you have questions, ask in the forums or find us on Twitter!

Posted by Nick Zakrasek, Product Manager
Author: Google Webmaster Central
Posted: June 20, 2017, 7:12 am

Making the Internet safer and faster: Introducing reCAPTCHA Android API

When we launched reCAPTCHA ten years ago, we had a simple goal: enable users to visit the sites they love without worrying about spam and abuse. Over the years, reCAPTCHA has changed quite a bit. It evolved from the distorted text to street numbers and names, then No CAPTCHA reCAPTCHA in 2014 and Invisible reCAPTCHA in March this year.

By now, more than a billion users have benefited from reCAPTCHA and we continue to work to refine our protections.

reCAPTCHA protects users wherever they may be online. As the use of mobile devices has grown rapidly, it’s important to keep the mobile applications and data safe. Today, on reCAPTCHA’s tenth birthday, we’re glad to announce the first reCAPTCHA Android API as part of Google Play Services.

With this API, reCAPTCHA can better tell human and bots apart to provide a streamlined user experience on mobile. It will use our newest Invisible reCAPTCHA technology, which runs risk analysis behind the scene and has enabled millions of human users to pass through with zero click everyday. Now mobile users can enjoy their apps without being interrupted, while still staying away from spam and abuse.

reCAPTCHA Android API is included with Google SafetyNet, which provides services like device attestation and safe browsing to protect mobile apps. Mobile developers can do both the device and user attestations in the same API to mitigate security risks of their apps more efficiently. This adds to the diversity of security protections on Android: Google Play Protect to monitor for potentially harmful applications, device encryption, and regular security updates. Please visit our site to learn more about how to integrate with the reCAPTCHA Android API, and keep an eye out for our iOS library.

The journey of reCAPTCHA continues: we’ll make the Internet safer and easier to use for everyone (except bots).


Posted by Wei Liu, Product Manager, reCAPTCHA
Author: Google Webmaster Central
Posted: June 8, 2017, 4:07 pm

Better Snippets for your Users

Before buying a book, people like to get a snapshot of how they’re about to spend a few hours reading. They'll take a look at the synopsis, the preface, or even the prologue just to get a sense about whether they'll like the book.
Search result snippets are much the same; they help people decide whether or not it makes sense to invest the time reading the page the snippet belongs to.  
The more descriptive and relevant a search result snippet is, the more likely that people will click through and be satisfied with the page they land on. Historically, snippets came from 3 places:
  1. The content of the page
  2. The meta description
  3. DMOZ listings
The content of the page is an obvious choice for result snippets, and  the content that can be extracted is often the most relevant to people’s queries. However, there are times when the content itself isn't the best source for a snippet. For instance, when someone searches for a publishing company for their book, the relevant homepages in the result set may contain only a few images describing the businesses and a logo, and maybe some links, none of which are particularly useful for a snippet.
The logical fallback in cases when the content of a page doesn't have much textual content for a search result snippet is the meta description. This should be short blurbs that describe accurately and precisely the content in a few words.
Finally, when a page doesn't have much textual content for snippet generation and the meta description is missing, unrelated to the page, or low quality, our fallback was DMOZ, also known as The Open Directory Project. For over 10 years, we relied on DMOZ for snippets because the quality of the DMOZ snippets were often much higher quality than those  provided by webmasters in their meta description, or were more descriptive than what the page provided.
With DMOZ now closed, we've stopped using its listings for snippeting, so it's a lot more important that webmasters provide good meta descriptions, if adding more content to the page is not an option.
What makes a good meta description?
Good meta descriptions are short blurbs that describe accurately the content of the page. They are like a pitch that convince the user that the page is exactly what they're looking for. For more tips, we have a handy help center article on the topic. Remember to make sure that both your desktop and your mobile pages include both a title and a meta description.
What are the most common problems with meta descriptions?
Because meta descriptions are usually visible only to search engines and other software, webmasters sometimes forget about them, leaving them completely empty. It's also common, for the same reason, that the same meta description is used across multiple (and sometimes many) pages. On the flip side, it's also relatively common that the description is completely off-topic, low quality, or outright spammy. These issues tarnish our users' search experience, so we prefer to ignore such meta descriptions.
Is there a character limit for meta descriptions?
There's no limit on how long a meta description can be, but the search result snippets are truncated as needed, typically to fit the device width.
What will happen with the "NOODP" robots directive?
With DMOZ (ODP) closed, we stopped relying on its data and thus the NOODP directive is already no-op.
Can I prevent Google from using the page contents as snippet?
You can prevent Google from generating snippets altogether by specifying the "nosnippet" robots directive. There's no way to prevent using page contents as snippet while allowing other sources.

As always, if you have questions, ask in the forums or find us on Twitter!
Posted by Gary, Search Team
Author: Google Webmaster Central
Posted: June 2, 2017, 9:00 am

A reminder about links in large-scale article campaigns

Lately we've seen an increase in spammy links contained in articles referred to as contributor posts, guest posts, partner posts, or syndicated posts. These articles are generally written by or in the name of one website, and published on a different one.

Google does not discourage these types of articles in the cases when they inform users, educate another site’s audience or bring awareness to your cause or company. However, what does violate Google's guidelines on link schemes is when the main intent is to build links in a large-scale way back to the author’s site. Below are factors that, when taken to an extreme, can indicate when an article is in violation of these guidelines:

  • Stuffing keyword-rich links to your site in your articles
  • Having the articles published across many different sites; alternatively, having a large number of articles on a few large, different sites
  • Using or hiring article writers that aren’t knowledgeable about the topics they’re writing on
  • Using the same or similar content across these articles; alternatively, duplicating the full content of articles found on your own site (in which case use of rel=”canonical”, in addition to rel=”nofollow”, is advised)

When Google detects that a website is publishing articles that contain spammy links, this may change Google's perception of the quality of the site and could affect its ranking. Sites accepting and publishing such articles should carefully vet them, asking questions like: Do I know this person? Does this person’s message fit with my site’s audience? Does the article contain useful content? If there are links of questionable intent in the article, has the author used rel=”nofollow” on them?


For websites creating articles made for links, Google takes action on this behavior because it’s bad for the Web as a whole. When link building comes first, the quality of the articles can suffer and create a bad experience for users. Also, webmasters generally prefer not to receive aggressive or repeated "Post my article!" requests, and we encourage such cases to be reported to our spam report form. And lastly, if a link is a form of endorsement, and you’re the one creating most of the endorsements for your own site, is this putting forth the best impression of your site? Our best advice in relation to link building is to focus on improving your site’s content and everything--including links--will follow (no pun intended).

Posted by the Google Webspam Team
Author: Google Webmaster Central
Posted: May 25, 2017, 5:23 pm

TechRadar: Mobile computing news

TechRadar UK latest feeds

Best free iPad games 2017

So you've got an iPad and now you have no cash left to buy any games. Have no fear, because there are plenty you can get for no money at all.
Author:
Posted: July 25, 2017, 9:41 am

The best laptops of 2017: top laptops for every need

In the market for a new laptop? Let us be your guide to the best laptops around.
Author:
Posted: July 24, 2017, 8:25 pm

A lesson on why accessibility features are so important

Technology is built for people, but not all technology works with all people. Here’s why it should.
Author:
Posted: July 24, 2017, 4:04 pm

Surface Book 2 release date, news and rumors

Could the Surface Book 2 arrive alongside the Windows 10 Fall Creators Update?
Author:
Posted: July 24, 2017, 2:47 pm

 



Formulario de contacto:

2 + 3 =

“En webDelFuturo le entregamos una web posicionada en el mercado de su competencia”

“No se conforme con webs que solo aparezcan en una tarjeta de presentación. Haga que su web sirva para las metas de su negocio”

“La web más las redes sociales crean nuevos espacios de negocios. Las aplicaciones móviles están para quedarse, no se quede fuera de la era digital”

Galo Hernández

CEO