The graphical user interface is presented (displayed) on the computer screen. It is the result of processed user input andRead more
Stormfront is a white nationalist, white supremacist, antisemitic, Holocaust-denialist, and neo-Nazi Internet forum, and the Web‘s first major racial hateRead more
Not to be confused with Jimmy Wales. Jimdo is a German website-builder and all-in-one hosting solution, designed to enable usersRead more
This is a comparison of file hosting services which are currently active. File hosting services are a particular kind ofRead more
Bluehost is a web hosting company owned by Endurance International Group. It is one of the 20 largest web hosts,Read more
I live in San Francisco, but I work an East Coast schedule to get a jump on the news day. So I’d already been at my desk for a couple of hours on Wednesday morning when I looked up and saw this:
As unsettling as it was to see the natural environment so transformed, I still got my work done. This is not to boast: I have a desk job and a working air filter. (People who make deliveries in the toxic air or are homeschooling their children while working from home during a global pandemic, however, impress the hell out of me.)
Not coincidentally, two of the Extra Crunch stories that ran since our Tuesday newsletter tie directly into what’s going on outside my window:
As this guest post predicted, a suboptimal attempt I made to track a delayed package using interactive voice response (IVR) indeed poisoned my customer experience, and;
Sheltering in place to avoid the novel coronavirus — and wildfire smoke — is fueling growth in the video-game industry, perhaps one factor in Unity Software Inc.’s plan to go public ahead of competitor Epic Games. In a two-part series, we looked at how the company has expanded beyond games and shared a detailed financial breakdown.
We covered a lot of ground this week, so scroll down or visit the recently redesigned Extra Crunch home page. If you’d like to receive this roundup via email each Tuesday and Friday, please click here.
Thanks very much for reading Extra Crunch; I hope you have a relaxing and safe weekend.
In a two-part series that ran on TechCrunch and Extra Crunch, former media columnist Eric Peckham returned to share his analysis of Unity Software Inc.’s S-1 filing.
Part one is a deep dive that explains how the company has grown beyond gaming to develop multiple revenue streams and where it’s headed.
For part two on Extra Crunch, he studied the company’s numbers to offer some context for its approximately $11 billion valuation.
As we’ve covered previously, the COVID-19 pandemic is making the world a lot smaller.
Investors who focus on their own backyards still have an advantage, but the ability to set up a quick coffee meeting with a promising investor is no longer one of them.
Even though some VCs are cutting first checks after Zoom calls, regional investors’ personal networks are still a trump card. Tourists will always rely on guide books, however, which is why we continue to survey investors around the world.
A Dealroom report issued this summer determined that 97 VC funds backed more than 1,600 funding rounds in Poland last year. With over 2,400 early- and late-stage startups and 400,000 engineers in the country, it’s easy to see why foreign investors are taking notice.
Editor-at-large Mike Butcher reached out to several investors who focus on Warsaw and Poland in general to learn more about the startups fueling their interest across fintech, gaming, security and other sectors:
- Bryony Cooper, managing partner, Arkley Brinc VC
- Anna Wnuk-Błażejczyk, investor relations manager, Experior.vc
- Rafał Roszak, investment director, YouNick Mint
- Michal Mroczkowski, partner, Market One Capital
- Marcus Erken, partner, Sunfish Partners
- Borys Musielak, partner, SMOK Ventures
- Mathias Åsberg, partner, Nextgrid
- Kuba Dudek, SpeedUp Venture Capital Group
- Marcin Laczynski, partner, Next Road Ventures
- Michał Rokosz, partner, Inovo Venture Partners
We’ll run the conclusion of his survey next Tuesday.
Even for fledgling startups, creating a robust customer service channel — or at least one that doesn’t annoy people — is a reliable way to keep users in the sales funnel.
Using AI and automation is fine, but now that consumers have grown used to asking phones and smart speakers to predict the weather and read recipe instructions, their expectations are higher than ever.
If you’re trying to figure out what people want from hyper-personalized customer experiences and how you can operationalize AI to give them what they’re after, start here.
For today’s edition of The Exchange, Natasha Mascarenhas joined Alex Wilhelm to examine how the pandemic-fueled surge of interest in edtech is manifesting on the funding front.
The numbers suggest that funding will far surpass the sector’s high-water mark set in 2018, so the duo studied the numbers through August 31, which included a number of mega-rounds that exceeded $100 million.
“Now the challenge for the sector will be keeping its growth alive in 2021, showing investors that their 2020 bets were not merely wagers made during a single, overheated year,” they conclude.
The odds are low that someone’s going to enter my home and steal my belongings. I still lock my door when I leave the house, however, and my valuables are insured. I’m an optimist, not a fool.
Similarly: Is your startup’s cybersecurity strategy based on optimism, or do you have an actual response plan in case of a data breach?
Security reporter Zack Whittaker has seen some shambolic reactions to security lapses, which is why he turned in a post-mortem about a corporation that got it right.
“Once in a while, a company’s response almost makes up for the daily deluge of hypocrisy, obfuscation and downright lies,” says Zack.
There’s a lot of buzz about special purpose acquisition companies these days.
Used-car marketplace Shift announced its SPAC in June 2020, and is on track to complete the process in the next few months, so co-founder/co-CEO George Arison wrote an Extra Crunch guest post to share what he has learned.
Step one: “If you go the SPAC route, you’ll need to become an expert at financial engineering.”
I am a software engineer and have been looking at job postings in the U.S. I’ve heard from my friends about J-1 Visa Training or J-1 Research.
What is a J-1 status? What are the requirements to qualify? Do I need to find a U.S. employer willing to sponsor me before I apply for one? Can I get a visa? How long could I stay?
— Determined in Delhi
While we count down to the September 23 premiere of NYSE: PLTR, Danny Crichton looked at the “robust secondary market” that has allowed some investors to acquire shares early.
“Given the number of people involved and the number of shares bought and sold over the past 18 months, we can get some insight regarding how insiders perceive Palantir’s value,” he writes.
Zack Whittaker interviewed Bugcrowd CTO, founder and chairman Casey Ellis about the best practices he recommends for creating a startup culture that takes security seriously.
“It’s an everyone problem,” said Ellis, who encouraged founders to promote the notion of “productive paranoia.”
Now that the threat envelope includes everyone from marketing to engineering, employees need to “internalize the fact that bad stuff can and does happen if you do it wrong,” Ellis said.
RollerCoaster Tycoon 3: Complete Edition arrives on both Switch and PC on September 24. The complete edition comes with the Soaked! and Wild! expansion packs, so you’re good to go on all the bonus content to make some ripping rides.
The draw on Switch is clearly to build the ultimate coaster anywhere, anytime, but PC players get the bonus of having widescreen mode and 1080p. Take a peek at a new trailer showcasing the roaring roller coaster ride below!
Google, Google, Google
For well over a decade Google has dominated search to where most stories in the search sphere were about Google or something on the periphery.
In 2019 Google generated $134.81 billion in ad revenues.
When Verizon bought core Yahoo three years ago the final purchase price was $4.48 billion. That amount was to own their finance vertical, news vertical, web portal, homepage, email & web search. It also included a variety of other services like Tumblr.
Part of what keeps Google so dominant in search is their brand awareness. That is also augmented by distribution as defaults in Chrome and Android. Then when it comes to buying search distribution from other players like Mozilla Firefox, Opera or Apple’s Safari they can outbid everyone else as they are much better at monetizing tier 2 markets and emerging markets than other search companies are since they have such strong ad depth. Even if Bing gave a 100% revshare to Apple they still could not compete with Google in most markets in terms of search monetization.
Apple as a Huge Search Traffic Driver
In 2019 Google paid just under £1.2 billion in default payments for UK search traffic. Most of that went to Apple. Historically when Google broke out their search revenues by region typically the US was around 45% to 46% of search ad revenue & the UK was around 11% to 12%, so it is likely Google is spending north of $10 billion a year to be the default search provider on Apple devices:
Apple submitted that search engines do not pay Apple for the right to be set as the primary default search engine on its devices. However, our assessment is that Google does pay to be the primary default on Apple devices. The agreement between Google and Apple states that Google will be the default web search provider and the same agreement states that Google will pay Apple a specified share of search advertising revenues. We also note that Google does not pay compensation to any partners that set Google Search as a secondary option. This further suggests that Google’s payment to Apple is in return for Apple setting Google as the primary default.
Apple is glad to cash those checks & let Google handle the core algorithmic search function in the web browser, but Apple also auto-completes many searches from within the address bar via various features like website history, top hit, news, Siri suggested website, suggested sites, etc.
A Unique Voice in Search
The nice thing about Apple powering some of those search auto-complete results themselves is their results are not simply a re-hash of the Google search results so they can add a unique voice to the search marketplace where if your site isn’t doing as well in Google it could still be promoted by Apple based on other factors.
Apple users generally have plenty of disposable personal income and a tendency to dispose of much of it, so if you are an Android user it is probably worth having an Apple device to see what they are recommending for core terms in your client’s markets. If you want to see recommendations for a particular country you may need to have a specialized router targeted to that country or use a web proxy or VPN.
Most users likely conduct full search queries and click through to listings from the Google search result page, but over time the search autocomplete feature that recommends previously viewed websites and other sites likely picks up incremental share of voice.
A friend of mine from the UK runs a local site and the following shows how the Apple ecosystem drove nearly 2/3 of his website traffic.
His website is only a couple years old, so it doesn’t get a ton of traffic from other sources yet. As of now his site does not have great Google rankings, but even if it did the boost by the Apple recommendations still provides a tailwind of free distribution and awareness (for however long it lasts).
For topics covered in news or repeat navigational searches Apple likely sends a lot of direct visits via their URL auto-completion features, but they do not use the feature broadly into the tail of search across other verticals, so it is a limited set of searches that ultimately benefit from the shortcuts.
Apple Search Ranking Factors
Apple Search may take the following into account when ranking web search results:
- Aggregated user engagement with search results
- Relevancy and matching of search terms to webpage topics and content
- Number and quality of links from other pages on the web
- User location based signals (approximate data)
- Webpage design characteristics
I have seen some country-code TLDs do well in their local markets in spite of not necessarily being associated with large brands. Sites which do not rank well in Google can still end up in the mix provided the user experience is clean, the site is useful and it is easy for Apple to associate the site with a related keyword.
Panda-like Quality Updates
Markets like news change every day as the news changes, but I think Apple also does some Panda-like updates roughly quarterly where they do a broad refresh of what they recommend generally. As part of those updates sites which were once recommended can end up seeing the recommendation go away (especially if user experience declined since the initial recommendation via an ad heavy layout or similar) while other sites that have good engagement metrics get recommended on related searches.
A friend had a website they sort of forgot that was recommended by Apple. That site saw a big jump on July 9, 2018 then it slid back in early August that year, likely after the testing data showed it wasn’t as good as some other site Apple recommended. They noticed the spike in traffic & improved the site a bit. In early October it was widely recommended once again. That lasted until May of 2019 when it fell off a cliff once more. They had monetized the site with a somewhat spammy ad network & the recommendation mostly went away.
The recommendations happen as the person types and they may be different for searches where there is a space between keywords and the word is ran together. It is also worth noting Apple will typically recommend the www. version of a site over the m. version of a site for sites that offer both, so it makes sense to ensure if you used separate URLs that the www version also uses a responsive website design.
Indirect Impact on Google
While the Apple search shortcuts bypass Google search & thus do not create direct user signals to impact Google search, people who own an iPhone then search on a Windows computer at work or a Windows laptop at home might remember the site they liked from their iPhone and search for it once more, giving the site some awareness that could indirectly bleed over into impacting Google’s search rankings.
Apple could also eventually roll out their own fully featured search engine.
Links = Rank
Old Google (pre-Panda) was to some degree largely the following: links = rank.
Once you had enough links to a site you could literally pour content into a site like water and have the domain’s aggregate link authority help anything on that site rank well quickly.
As much as PageRank was hyped & important, having a diverse range of linking domains and keyword-focused anchor text were important.
Brand = Rank
After Vince then Panda a site’s brand awareness (or, rather, ranking signals that might best simulate it) were folded into the ability to rank well.
Panda considered factors beyond links & when it first rolled out it would clip anything on a particular domain or subdomain. Some sites like HubPages shifted their content into subdomains by users. And some aggressive spammers would rotate their entire site onto different subdomains repeatedly each time a Panda update happened. That allowed those sites to immediately recover from the first couple Panda updates, but eventually Google closed off that loophole.
Any signal which gets relied on eventually gets abused intentionally or unintentionally. And over time it leads to a “sameness” of the result set unless other signals are used:
Google is absolute garbage for searching anything related to a product. If I’m trying to learn something invariably I am required to search another source like Reddit through Google. For example, I became introduced to the concept of weighted blankets and was intrigued. So I Google “why use a weighted blanket” and “weighted blanket benefits”. Just by virtue of the word “weighted blanket” being in the search I got pages and pages of nothing but ads trying to sell them, and zero meaningful discourse on why I would use one
Getting More Granular
Over time as Google got more refined with Panda broad-based sites outside of the news vertical often fell on tough times unless they were dedicated to some specific media format or had a lot of user engagement metrics like a strong social network site. That is a big part of why the New York Times sold About.com for less than they paid for it & after IAC bought it they broke it down into a variety of sites like: Verywell (health), the Spruce (home decor), the Balance (personal finance), Lifewire (technology), Tripsavvy (travel) and ThoughtCo (education & self-improvement).
Penguin further clipped aggressive anchor text built on low quality links. When the Penguin update rolled out Google also rolled out an on-page spam classifier to further obfuscate the update. And the Penguin update was sandwiched by Panda updates on either side, making it hard for people to reverse engineer any signal out of weekly winners and losers lists from services that aggregate massive amounts of keyword rank tracking data.
So much of the link graph has been decimated that Google reversed their stance on nofollow to where in March 1st of this year they started treating it as a hint versus a directive for ranking purposes. Many mainstream media websites were overusing nofollow or not citing sources at all, so this additional layer of obfuscation on Google’s part will allow them to find more signal in that noise.
May 4, 2020 Algo Update
On May 4th Google rolled out another major core update.
Later today, we are releasing a broad core algorithm update, as we do several times per year. It is called the May 2020 Core Update. Our guidance about such updates remains as we’ve covered before. Please see this blog post for more about that:https://t.co/e5ZQUAlt0G— Google SearchLiaison (@searchliaison) May 4, 2020
I saw some sites which had their rankings suppressed for years see a big jump. But many things changed at once.
On some political search queries which were primarily classified as being news related Google is trying to limit political blowback by showing official sites and data scraped from official sites instead of putting news front & center.
“Google’s pretty much made it explicit that they’re not going to propagate news sites when it comes to election related queries and you scroll and you get a giant election widget in your phone and it shows you all the different data on the primary results and then you go down, you find Wikipedia, you find other like historical references, and before you even get to a single news article, it’s pretty crazy how Google’s changed the way that the SERP is intended.”
That change reflects the permanent change to the news media ecosystem brought on by the web.
The Internet commoditized the distribution of facts. The “news” media responded by pivoting wholesale into opinions and entertainment.— Naval (@naval) May 26, 2016
A blog post by Lily Ray from Path Interactive used Sistrix data to show many of the sites which saw high volatility were in the healthcare vertical & other your money, your life (YMYL) categories.
One of the more interesting pieces of feedback on the update was from Rank Ranger, where they looked at particular pages that jumped or fell hard on the update. They noticed sites that put ads or ad-like content front and center may have seen sharp falls on some of those big money pages which were aggressively monetized:
Seeing this all but cements the notion (in my mind at least) that Google did not want content unrelated to the main purpose of the page to appear above the fold to the exclusion of the page’s main content! Now for the second wrinkle in my theory…. A lot of the pages being swapped out for new ones did not use the above-indicated format where a series of “navigation boxes” dominated the page above the fold.
The above shift had a big impact on some sites which are worth serious money. Intuit paid over $7 billion to acquire Credit Karma, but their credit card affiliate pages recently slid hard.
Credit Karma lost 40% traffic from May core update. That’s insane, they do major TV ads and likely pay millions in SEO expenses. Think about that folks. Your site isn’t safe. Google changes what they want radically with every update, while telling us nothing!— SEOwner (@tehseowner) May 14, 2020
The above sort of shift reflects Google getting more granular with their algorithms. Early Panda was all or nothing. Then it started to have different levels of impact throughout different portions of a site.
Brand was sort of a band aid or a rising tide that lifted all (branded) boats. Now we are seeing Google get more granular with their algorithms where a strong brand might not be enough if they view the monetization as being excessive. That same focus on page layout can have a more adverse impact on small niche websites.
One of my old legacy clients had a site which was primarily monetized by the Amazon affiliate program. About a month ago Amazon chopped affiliate commissions in half & then the aggressive ad placement caused search traffic to the site to get chopped in half when rankings slid on this update.
Their site has been trending down over the past couple years largely due to neglect as it was always a small side project. They recently improved some of the content about a month or so ago and that ended up leading to a bit of a boost, but then this update came. As long as that ad placement doesn’t change the declines are likely to continue.
They just recently removed that ad unit, but that meant another drop in income as until there is another big algo update they’re likely to stay at around half search traffic. So now they have a half of a half of a half. Good thing the site did not have any full time employees or they’d be among the millions of newly unemployed. That experience though really reflects how websites can be almost like debt levered companies in terms of going under virtually overnight. Who can have revenue slide around 88% and then take increase investment in the property using the remaining 12% while they wait for the site to be rescored for a quarter year or more?
“If you have been negatively impacted by a core update, you (mostly) cannot see recovery from that until another core update. In addition, you will only see recovery if you significantly improve the site over the long-term. If you haven’t done enough to improve the site overall, you might have to wait several updates to see an increase as you keep improving the site. And since core updates are typically separated by 3-4 months, that means you might need to wait a while.”
Almost nobody can afford to do that unless the site is just a side project.
Google could choose to run major updates more frequently, allowing sites to recover more quickly, but they gain economic benefit in defunding SEO investments & adding opportunity cost to aggressive SEO strategies by ensuring ranking declines on major updates last a season or more.
Choosing a Strategy vs Letting Things Come at You
They probably should have lowered their ad density when they did those other upgrades. If they had they likely would have seen rankings at worst flat or likely up as some other competing sites fell. Instead they are rolling with a half of a half of a half on the revenue front. Glenn Gabe preaches the importance of fixing all the problems you can find rather than just fixing one or two things and hoping it is enough. If you have a site which is on the edge you sort of have to consider the trade offs between various approaches to monetization.
- monetize it lightly and hope the site does well for many years
- monetize it slightly aggressively while using the extra income to further improve the site elsewhere and ensure you have enough to get by any lean months
- aggressively monetize the shortly after a major ranking update if it was previously lightly monetized & then hope to sell it off a month or two later before the next major algorithm update clips it again
Outcomes will depend partly on timing and luck, but consciously choosing a strategy is likely to yield better returns than doing a bit of mix-n-match while having your head buried in the sand.
Reading the Algo Updates
You can spend 50 or 100 hours reading blog posts about the update and learn precisely nothing in the process if you do not know which authors are bullshitting and which authors are writing about the correct signals.
But how do you know who knows what they are talking about?
It is more than a bit tricky as the people who know the most often do not have any economic advantage in writing specifics about the update. If you primarily monetize your own websites, then the ignorance of the broader market is a big part of your competitive advantage.
Making things even trickier, the less you know the more likely Google would be to trust you with sending official messaging through you. If you syndicate their messaging without questioning it, you get a treat – more exclusives. If you question their messaging in a way that undermines their goals, you’d quickly become persona non grata – something cNet learned many years ago when they published Eric Schmidt’s address.
It would be unlikely you’d see the following sort of Tweet from say Blue Hat SEO or Fantomaster or such.
I asked Gary about E-A-T. He said it’s largely based on links and mentions on authoritative sites. i.e. if the Washington post mentions you, that’s good.
To be able to read the algorithms well you have to have some market sectors and keyword groups you know well. Passively collecting an archive of historical data makes the big changes stand out quickly.
Everyone who depends on SEO to make a living should subscribe to an online rank tracking service or run something like Serposcope locally to track at least a dozen or two dozen keywords. If you track rankings locally it makes sense to use a set of web proxies and run the queries slowly through each so you don’t get blocked.
You should track at least a diverse range to get a true sense of the algorithmic changes.
- a couple different industries
- a couple different geographic markets (or at least some local-intent vs national-intent terms within a country)
- some head, midtail and longtail keywords
- sites of different size, age & brand awareness within a particular market
Some tools make it easy to quickly add or remove graphing of anything which moved big and is in the top 50 or 100 results, which can help you quickly find outliers. And some tools also make it easy to compare their rankings over time. As updates develop you’ll often see multiple sites making big moves at the same time & if you know a lot about the keyword, the market & the sites you can get a good idea of what might have been likely to change to cause those shifts.
Once you see someone mention outliers most people miss that align with what you see in a data set, your level of confidence increases and you can spend more time trying to unravel what signals changed.
I’ve read influential industry writers mention that links were heavily discounted on this update. I have also read Tweets like this one which could potentially indicate the opposite.
If I had little to no data, I wouldn’t be able to get any signal out of that range of opinions. I’d sort of be stuck at “who knows.”
By having my own data I track I can quickly figure out which message is more inline with what I saw in my subset of data & form a more solid hypothesis.
No Single Smoking Gun
As Glenn Gabe is fond of saying, sites that tank usually have multiple major issues.
Google rolls out major updates infrequently enough that they can sandwich a couple different aspects into major updates at the same time in order to make it harder to reverse engineer updates. So it does help to read widely with an open mind and imagine what signal shifts could cause the sorts of ranking shifts you are seeing.
Sometimes site level data is more than enough to figure out what changed, but as the above Credit Karma example showed sometimes you need to get far more granular and look at page-level data to form a solid hypothesis.
As the World Changes, the Web Also Changes
About 15 years ago online dating was seen as a weird niche for recluses who perhaps typically repulsed real people in person. Now there are all sorts of niche specialty dating sites including a variety of DTF type apps. What was once weird & absurd had over time become normal.
The COVID-19 scare is going to cause lasting shifts in consumer behavior that accelerate the movement of commerce online. A decade of change will happen in a year or two across many markets.
Telemedicine will grow quickly. Facebook is adding commerce featured directly onto their platform through partnering with Shopify. Spotify is spending big money to buy exclusives rights to distribute widely followed podcasters like Joe Rogan. Uber recently offered to acquire GrubHub. Google and Apple will continue adding financing features to their mobile devices. Movie theaters have lost much of their appeal.
Tons of offline “value” businesses ended up having no value after months of revenue disappearing while large outstanding debts accumulated interest. There is a belief that some of those brands will have strong latent brand value that carries over online, but if they were weak even when the offline stores acting like interactive billboards subsidized consumer awareness of their brands then as those stores close the consumer awareness & loyalty from in-person interactions will also dry up. A shell of a company rebuilt around the Toys R’ Us brand is unlikely to beat out Amazon’s parallel offering or a company which still runs stores offline.
Big box retailers like Target & Walmart are growing their online sales at hundreds of percent year over year.
There will be waves of bankruptcies, dramatic shifts in commercial real estate prices (already reflected in plunging REIT prices), and more people working remotely (shifting residential real estate demand from the urban core back out into suburbs).
People who work remote are easier to hire and easier to fire. Those who keep leveling up their skills will eventually get rewarded while those who don’t will rotate jobs every year or two. The lack of stability will increase demand for education, though much of that incremental demand will be around new technologies and specific sectors – certificates or informal training programs instead of degrees.
More and more activities will become normal online activities.
The University of California has about a half-million students & in the fall semester they are going to try to have most of those classes happen online. How much usage data does Google gain as thousands of institutions put more and more of their infrastructure and service online?
Colleges have to convince students for the next year that a remote education is worth every bit as much as an in-person one, and then pivot back before students actually start believing it.
It’s like only being able to sell your competitor’s product for a year.— Naval (@naval) May 6, 2020
A lot of B & C level schools are going to go under as the like-vs-like comparison gets easier. Back when I ran a membership site here a college paid us to have students gain access to our membership area of the site. As online education gets normalized many unofficial trade-related sites will look more economically attractive on a relative basis.
If core institutions of the state deliver most of their services online, then other companies can be expected to follow. When big cities publish lists of crimes they will not respond to during economic downturns they are effectively subsidizing more crime. That in turn makes moving to somewhere a bit more rural & cheaper make sense, particularly when you no longer need to live near your employer.
The most important implication of this permanent WFH movement are state income taxes.
The warm, sunny states with affordable housing and zero taxes will see an influx of educated, rich workers. States will need to cut taxes to keep up.
The biggest loser in this is CA.— Chamath Palihapitiya (@chamath) May 21, 2020